Job Listings Aggregator Pro

by assertive_analogy

Aggregate listings from LinkedIn, Indeed, RemoteOK & more into one feed. Find Python, tech, and remote jobs faster with smart deduplication and filters.

55 runs
20 users
Try This Actor

Opens on Apify.com

About Job Listings Aggregator Pro

Tired of bouncing between ten different tabs just to see what's out there? I built this because I was spending more time checking job boards than actually applying. The Job Listings Aggregator Pro pulls listings from the major sites—LinkedIn, Indeed, RemoteOK, Dice, and a handful of others—into a single, clean feed. It's tuned specifically for Python and tech roles, with a real focus on remote opportunities. The magic is in how it handles the data. It runs a smart deduplication pass so you're not seeing the same job from three different sources, and the keyword filters let you cut out the noise immediately. You get results fast, without the usual scraping headaches of setting up proxies or parsing weird HTML structures. It just works. I use it to get a daily digest of what's new in my field without the manual grind. If you're on the hunt, this consolidates the first, most tedious step into a one-click process, giving you back hours to tailor your CV and write cover letters instead of mindlessly refreshing pages.

What does this actor do?

Job Listings Aggregator Pro is a web scraping and automation tool available on the Apify platform. It's designed to help you extract data and automate tasks efficiently in the cloud.

Key Features

  • Cloud-based execution - no local setup required
  • Scalable infrastructure for large-scale operations
  • API access for integration with your applications
  • Built-in proxy rotation and anti-blocking measures
  • Scheduled runs and webhooks for automation

How to Use

  1. Click "Try This Actor" to open it on Apify
  2. Create a free Apify account if you don't have one
  3. Configure the input parameters as needed
  4. Run the actor and download your results

Documentation

Job Listings Aggregator

A Python application that crawls multiple public job boards, normalizes the listings, and provides a unified interface for searching and managing them. It's designed for automation and can be scheduled to run daily.

Key Features

  • Multi-source scraping: Fetches jobs from sites like RemoteOK and We Work Remotely. The modular architecture makes it easy to add new job boards.
  • Data processing: Automatically normalizes and deduplicates job listings from different sources into a consistent format.
  • Flexible storage: Saves aggregated jobs to either JSON files or a SQLite database.
  • Advanced search & filtering: Filter jobs by keywords, location (including "remote"), job type, and company.
  • Export results: Export filtered job lists to CSV or JSON files.
  • CLI & API: Use via a command-line interface or import directly into your Python code.
  • Automated scheduling: Set up daily, automated scraping runs.

How to Use

Installation & Setup

  1. Clone the repository and install dependencies:
    bash git clone <repository-url> cd job-listings-aggregator pip install -r requirements.txt
  2. (Optional) Install Playwright browsers for JavaScript-heavy sites:
    bash playwright install
  3. Configure the app by editing config.py. Key settings include STORAGE_TYPE ('json' or 'sqlite'), MAX_JOBS_PER_BOARD, and scheduler options.

Basic Commands (CLI)

Scrape Jobs:

# Scrape all enabled job boards
python main.py scrape

# Scrape with specific keywords
python main.py scrape -k python -k remote

# Limit the number of jobs per board
python main.py scrape -m 50

Search & Filter Jobs:

# Search by keyword
python main.py search -k python

# Filter for remote jobs
python main.py search --remote

# Complex filter (keyword, location, company)
python main.py search -k "data scientist" -l remote --company google

# Export results
python main.py search -k python --export results.csv

Manage Automation:

# Start/stop the daily scheduler
python main.py schedule --start
python main.py schedule --stop

# Run a scraping job immediately
python main.py schedule --run-now

Python API

You can also use the aggregator directly in your scripts:

from main import JobAggregator

app = JobAggregator()
jobs = app.scrape_all_jobs(keywords=['python', 'remote'])
app.save_jobs(jobs)

remote_jobs = app.search_jobs(remote_only=True, keyword='developer')
for job in remote_jobs[:5]:
    print(f"{job.title} at {job.company}")

Input / Output

  • Input: The tool takes its input from configured job board websites. You control the process via CLI commands (keywords, limits) or the configuration file (storage type, scheduling).
  • Processing: It scrapes, normalizes fields (title, company, location, link), and deduplicates listings.
  • Output: Jobs are saved to your chosen storage backend (JSON file or SQLite DB). Search results can be viewed in the terminal or exported to CSV/JSON files.

Common Use Cases

Market Research

Gather competitive intelligence and market data

Lead Generation

Extract contact information for sales outreach

Price Monitoring

Track competitor pricing and product changes

Content Aggregation

Collect and organize content from multiple sources

Ready to Get Started?

Try Job Listings Aggregator Pro now on Apify. Free tier available with no credit card required.

Start Free Trial

Actor Information

Developer
assertive_analogy
Pricing
Paid
Total Runs
55
Active Users
20
Apify Platform

Apify provides a cloud platform for web scraping, data extraction, and automation. Build and run web scrapers in the cloud.

Learn more about Apify

Need Professional Help?

Couldn't solve your problem? Hire a verified specialist on Fiverr to get it done quickly and professionally.

Find a Specialist

Trusted by millions | Money-back guarantee | 24/7 Support