Kafka Integration
by brasil-scrapers
This integration facilitates the processing of data from an Apify dataset and its delivery to a Kafka topic. It is configured via a schema that includ...
Opens on Apify.com
About Kafka Integration
This integration facilitates the processing of data from an Apify dataset and its delivery to a Kafka topic. It is configured via a schema that includes details about the dataset, Kafka configuration, and batch processing size.
What does this actor do?
Kafka Integration is a web scraping and automation tool available on the Apify platform. It's designed to help you extract data and automate tasks efficiently in the cloud.
Key Features
- Cloud-based execution - no local setup required
- Scalable infrastructure for large-scale operations
- API access for integration with your applications
- Built-in proxy rotation and anti-blocking measures
- Scheduled runs and webhooks for automation
How to Use
- Click "Try This Actor" to open it on Apify
- Create a free Apify account if you don't have one
- Configure the input parameters as needed
- Run the actor and download your results
Documentation
Apify to Kafka Integration ## Overview This integration facilitates the processing of data from an Apify dataset and its delivery to a Kafka topic. It is configured via a schema that includes details about the dataset, Kafka configuration, and batch processing size. ## Input Schema The input schema defines the configuration required to run the integration effectively. Below are the details of the schema and its properties: ### Schema Details - Title: Apify to Kafka Input Schema - Type: Object - Schema Version: 1 ### Properties #### 1. Default Dataset ID - Title: Default Dataset ID - Type: String - Description: The ID of the Apify dataset to process. - Default: default - Editor: Textfield #### 2. Kafka Configuration - Title: Kafka Configuration - Type: Object - Description: Configuration settings for Kafka connection. - Editor: JSON - Default: json { "clientId": "apify-kafka-producer", "brokers": ["localhost:9092"], "topic": "test-topic", "ssl": false } - Properties: - Client ID: - Title: Client ID - Type: String - Description: Kafka client identifier. - Default: apify-kafka-producer - Brokers: - Title: Brokers - Type: Array of strings - Description: Array of Kafka broker addresses. - Default: ["localhost:9092"] - Topic: - Title: Topic - Type: String - Description: Kafka topic name. - Default: test-topic - SSL: - Title: SSL - Type: Boolean - Description: Enable/disable SSL connection. - Default: false - SASL (Optional): - Title: SASL - Type: Object - Description: SASL configuration for Kafka connection. - Properties: - Username: - Title: Username - Type: String - Description: Kafka SASL username. - Password: - Title: Password - Type: String - Description: Kafka SASL password. - Mechanism: - Title: Mechanism - Type: String - Description: Kafka SASL mechanism. - Default: plain #### 3. Batch Size - Title: Batch Size - Type: Integer - Description: Number of messages to process in each batch. - Default: 2 - Minimum: 1 ### Required Properties - defaultDatasetId - kafkaConfig ## Example Configuration json { "defaultDatasetId": "my-dataset-id", "kafkaConfig": { "clientId": "my-kafka-client", "brokers": ["kafka-broker1:9092", "kafka-broker2:9092"], "topic": "my-kafka-topic", "ssl": true, "sasl": { "username": "my-username", "password": "my-password", "mechanism": "plain" } }, "batchSize": 5 } ## How to Use 1. Define the Input: Provide the necessary configurations in JSON format as per the schema. 2. Run the Integration: Pass the configuration to the Apify actor or script responsible for processing and delivering the dataset to Kafka. 3. Monitor Outputs: Check your Kafka topic for incoming messages based on the processed dataset. ## Notes - Ensure the Kafka broker addresses are reachable from the environment where the integration is executed. - SASL configuration is optional but required for secure Kafka connections. - Modify the batch size according to the volume of data and processing capacity.
Categories
Common Use Cases
Market Research
Gather competitive intelligence and market data
Lead Generation
Extract contact information for sales outreach
Price Monitoring
Track competitor pricing and product changes
Content Aggregation
Collect and organize content from multiple sources
Ready to Get Started?
Try Kafka Integration now on Apify. Free tier available with no credit card required.
Start Free TrialActor Information
- Developer
- brasil-scrapers
- Pricing
- Paid
- Total Runs
- 21
- Active Users
- 3
Related Actors
Tecdoc Car Parts
by making-data-meaningful
OpenRouter - Unified LLM Interface for ChatGPT, Claude, Gemini
by xyzzy
Google Sheets Import & Export
by lukaskrivka
Send Email
by apify
Apify provides a cloud platform for web scraping, data extraction, and automation. Build and run web scrapers in the cloud.
Learn more about ApifyNeed Professional Help?
Couldn't solve your problem? Hire a verified specialist on Fiverr to get it done quickly and professionally.
Trusted by millions | Money-back guarantee | 24/7 Support