Kafka Integration

Kafka Integration

by brasil-scrapers

This integration facilitates the processing of data from an Apify dataset and its delivery to a Kafka topic. It is configured via a schema that includ...

21 runs
3 users
Try This Actor

Opens on Apify.com

About Kafka Integration

This integration facilitates the processing of data from an Apify dataset and its delivery to a Kafka topic. It is configured via a schema that includes details about the dataset, Kafka configuration, and batch processing size.

What does this actor do?

Kafka Integration is a web scraping and automation tool available on the Apify platform. It's designed to help you extract data and automate tasks efficiently in the cloud.

Key Features

  • Cloud-based execution - no local setup required
  • Scalable infrastructure for large-scale operations
  • API access for integration with your applications
  • Built-in proxy rotation and anti-blocking measures
  • Scheduled runs and webhooks for automation

How to Use

  1. Click "Try This Actor" to open it on Apify
  2. Create a free Apify account if you don't have one
  3. Configure the input parameters as needed
  4. Run the actor and download your results

Documentation

Apify to Kafka Integration ## Overview This integration facilitates the processing of data from an Apify dataset and its delivery to a Kafka topic. It is configured via a schema that includes details about the dataset, Kafka configuration, and batch processing size. ## Input Schema The input schema defines the configuration required to run the integration effectively. Below are the details of the schema and its properties: ### Schema Details - Title: Apify to Kafka Input Schema - Type: Object - Schema Version: 1 ### Properties #### 1. Default Dataset ID - Title: Default Dataset ID - Type: String - Description: The ID of the Apify dataset to process. - Default: default - Editor: Textfield #### 2. Kafka Configuration - Title: Kafka Configuration - Type: Object - Description: Configuration settings for Kafka connection. - Editor: JSON - Default: json { "clientId": "apify-kafka-producer", "brokers": ["localhost:9092"], "topic": "test-topic", "ssl": false } - Properties: - Client ID: - Title: Client ID - Type: String - Description: Kafka client identifier. - Default: apify-kafka-producer - Brokers: - Title: Brokers - Type: Array of strings - Description: Array of Kafka broker addresses. - Default: ["localhost:9092"] - Topic: - Title: Topic - Type: String - Description: Kafka topic name. - Default: test-topic - SSL: - Title: SSL - Type: Boolean - Description: Enable/disable SSL connection. - Default: false - SASL (Optional): - Title: SASL - Type: Object - Description: SASL configuration for Kafka connection. - Properties: - Username: - Title: Username - Type: String - Description: Kafka SASL username. - Password: - Title: Password - Type: String - Description: Kafka SASL password. - Mechanism: - Title: Mechanism - Type: String - Description: Kafka SASL mechanism. - Default: plain #### 3. Batch Size - Title: Batch Size - Type: Integer - Description: Number of messages to process in each batch. - Default: 2 - Minimum: 1 ### Required Properties - defaultDatasetId - kafkaConfig ## Example Configuration json { "defaultDatasetId": "my-dataset-id", "kafkaConfig": { "clientId": "my-kafka-client", "brokers": ["kafka-broker1:9092", "kafka-broker2:9092"], "topic": "my-kafka-topic", "ssl": true, "sasl": { "username": "my-username", "password": "my-password", "mechanism": "plain" } }, "batchSize": 5 } ## How to Use 1. Define the Input: Provide the necessary configurations in JSON format as per the schema. 2. Run the Integration: Pass the configuration to the Apify actor or script responsible for processing and delivering the dataset to Kafka. 3. Monitor Outputs: Check your Kafka topic for incoming messages based on the processed dataset. ## Notes - Ensure the Kafka broker addresses are reachable from the environment where the integration is executed. - SASL configuration is optional but required for secure Kafka connections. - Modify the batch size according to the volume of data and processing capacity.

Common Use Cases

Market Research

Gather competitive intelligence and market data

Lead Generation

Extract contact information for sales outreach

Price Monitoring

Track competitor pricing and product changes

Content Aggregation

Collect and organize content from multiple sources

Ready to Get Started?

Try Kafka Integration now on Apify. Free tier available with no credit card required.

Start Free Trial

Actor Information

Developer
brasil-scrapers
Pricing
Paid
Total Runs
21
Active Users
3
Apify Platform

Apify provides a cloud platform for web scraping, data extraction, and automation. Build and run web scrapers in the cloud.

Learn more about Apify

Need Professional Help?

Couldn't solve your problem? Hire a verified specialist on Fiverr to get it done quickly and professionally.

Find a Specialist

Trusted by millions | Money-back guarantee | 24/7 Support