Keboola Uploader

Keboola Uploader

by hlidac-shopu

Reliable uploader of Apify Datasets to Keboola Connection (aka KBC). Integration-ready. Supports development of non-profit service https://www.hlidacs...

110 runs
3 users
Try This Actor

Opens on Apify.com

About Keboola Uploader

Reliable uploader of Apify Datasets to Keboola Connection (aka KBC). Integration-ready. Supports development of non-profit service https://www.hlidacshopu.cz/

What does this actor do?

Keboola Uploader is a web scraping and automation tool available on the Apify platform. It's designed to help you extract data and automate tasks efficiently in the cloud.

Key Features

  • Cloud-based execution - no local setup required
  • Scalable infrastructure for large-scale operations
  • API access for integration with your applications
  • Built-in proxy rotation and anti-blocking measures
  • Scheduled runs and webhooks for automation

How to Use

  1. Click "Try This Actor" to open it on Apify
  2. Create a free Apify account if you don't have one
  3. Configure the input parameters as needed
  4. Run the actor and download your results

Documentation

Keboola Uploader Reliable uploader of Apify Datasets to Keboola Connection. We are using Storage API Importer with optimal defaults. This actor is helpful in workflows or for ad-hoc data uploads. This actor is generalisation of our custom-made uploaders for many of our projects. It uses minimum dependencies and optimizes for speed and reliability. - gracefully handles migrations - implements retry policy for failed uploads - supports Actor Integration - allows to fine tune the batch size for you optimal usage of resources Your Apify Dataset will be split into batches, converted to CSV and uploaded with gzip compression enabled. You should choose the batchSize according to the nature of you data. Primitive properties from your Dataset will be 1:1 mapped to CSV table. Complex properties (arrays and objects) will be serialized to JSON, so you can use Snowflake support for JSON in your transformations. ## Inputs ### Apify Dataset ID datasetId ID of Apify Dataset that should be uploaded to Keboola. When you use this actor in Integrations workflow, this field is optional. Default Dataset of previous actor in the flow will be used. ### Keboola stack name keboolaStack Hostname of your Keboola stack import endpoint. See Keboola documentations for more details. Default is import.keboola.com for AWS US-East region. You can alternatively set KEBOOLA_STACK environment variable instead. Current multi-tenant stacks are: | region | hostname | | ---------------- | --------------------------------------- | | US Virginia AWS | import.keboola.com | | US Virginia GCP | import.us-east4.gcp.keboola.com | | EU Frankfurt AWS | import.eu-central-1.keboola.com | | EU Ireland Azure | import.north-europe.azure.keboola.com | | EU Frankfurt GCP | import.europe-west3.gcp.keboola.com | If you are a single tenant user then your hostname is in format import.CUSTOMER_NAME.keboola.com. ### Keboola API Key keboolaStorageApiKey Your API Key to a Keboola project where you want to upload the data. You should generate a new API key just for this actor with limited rights to write-only for destination bucket. You can alternatively set KEBOOLA_STORAGE_API_KEY environment variable instead. ### Keboola bucket bucket Name of the destination Keboola bucket. eg. in.c-apify ### Keboola table table Name of the destination Keboola table. eg. scrape_results ### Keboola table headers headers Array of header names of destination Keboola table. You can use this to select subset of properties to result table or to reorder the columns - the order of headers is preserved in result table. You can leave it blank if your Dataset items have all properties always specified (without undefined values). In this case properties of the first Dataset item are used. Our recommendation is to be explicit to prevent unexpected data loss. ### Batch size batchSize Size of the batch to upload. Dataset will be split into more batches if it has more items that this number. Batches will be uploaded sequentially. Choose the batch size according to the nature of you data and parallelization of you process. Generally speaking, Keboola Importer works best if you send less frequent bigger portions (dozens of MB gzipped) of data. On the other side you are constrained by the Actor size. You can easily hit OOM condition when this number is too high. ### Incremental load incremental When enabled, imported data will be added to the existing table. When disabled, table will be truncated - all existing data will be deleted from the table. Default is enabled (true). ## SDK You can also use this actor from yours custom actors via @hckr_/apify-keboola npm package. ### Install shell npm install @hckr_/apify-keboola ### Usage You can use free (pay per usage) variant: javascript import { uploadToKeboola } from "@hckr_/apify-keboola"; Or you can support our work on https://www.hlidacshopu.cz/ with the paid option: javascript import { uploadToKeboola } from "@hckr_/apify-keboola/hlidac-shopu";

Common Use Cases

Market Research

Gather competitive intelligence and market data

Lead Generation

Extract contact information for sales outreach

Price Monitoring

Track competitor pricing and product changes

Content Aggregation

Collect and organize content from multiple sources

Ready to Get Started?

Try Keboola Uploader now on Apify. Free tier available with no credit card required.

Start Free Trial

Actor Information

Developer
hlidac-shopu
Pricing
Paid
Total Runs
110
Active Users
3
Apify Platform

Apify provides a cloud platform for web scraping, data extraction, and automation. Build and run web scrapers in the cloud.

Learn more about Apify

Need Professional Help?

Couldn't solve your problem? Hire a verified specialist on Fiverr to get it done quickly and professionally.

Find a Specialist

Trusted by millions | Money-back guarantee | 24/7 Support