Job Listings Aggregator Pro
by assertive_analogy
Aggregate listings from LinkedIn, Indeed, RemoteOK & more into one feed. Find Python, tech, and remote jobs faster with smart deduplication and filters.
Opens on Apify.com
About Job Listings Aggregator Pro
Tired of bouncing between ten different tabs just to see what's out there? I built this because I was spending more time checking job boards than actually applying. The Job Listings Aggregator Pro pulls listings from the major sites—LinkedIn, Indeed, RemoteOK, Dice, and a handful of others—into a single, clean feed. It's tuned specifically for Python and tech roles, with a real focus on remote opportunities. The magic is in how it handles the data. It runs a smart deduplication pass so you're not seeing the same job from three different sources, and the keyword filters let you cut out the noise immediately. You get results fast, without the usual scraping headaches of setting up proxies or parsing weird HTML structures. It just works. I use it to get a daily digest of what's new in my field without the manual grind. If you're on the hunt, this consolidates the first, most tedious step into a one-click process, giving you back hours to tailor your CV and write cover letters instead of mindlessly refreshing pages.
What does this actor do?
Job Listings Aggregator Pro is a web scraping and automation tool available on the Apify platform. It's designed to help you extract data and automate tasks efficiently in the cloud.
Key Features
- Cloud-based execution - no local setup required
- Scalable infrastructure for large-scale operations
- API access for integration with your applications
- Built-in proxy rotation and anti-blocking measures
- Scheduled runs and webhooks for automation
How to Use
- Click "Try This Actor" to open it on Apify
- Create a free Apify account if you don't have one
- Configure the input parameters as needed
- Run the actor and download your results
Documentation
Job Listings Aggregator
A Python application that crawls multiple public job boards, normalizes the listings, and provides a unified interface for searching and managing them. It's designed for automation and can be scheduled to run daily.
Key Features
- Multi-source scraping: Fetches jobs from sites like RemoteOK and We Work Remotely. The modular architecture makes it easy to add new job boards.
- Data processing: Automatically normalizes and deduplicates job listings from different sources into a consistent format.
- Flexible storage: Saves aggregated jobs to either JSON files or a SQLite database.
- Advanced search & filtering: Filter jobs by keywords, location (including "remote"), job type, and company.
- Export results: Export filtered job lists to CSV or JSON files.
- CLI & API: Use via a command-line interface or import directly into your Python code.
- Automated scheduling: Set up daily, automated scraping runs.
How to Use
Installation & Setup
- Clone the repository and install dependencies:
bash git clone <repository-url> cd job-listings-aggregator pip install -r requirements.txt - (Optional) Install Playwright browsers for JavaScript-heavy sites:
bash playwright install - Configure the app by editing
config.py. Key settings includeSTORAGE_TYPE('json'or'sqlite'),MAX_JOBS_PER_BOARD, and scheduler options.
Basic Commands (CLI)
Scrape Jobs:
# Scrape all enabled job boards
python main.py scrape
# Scrape with specific keywords
python main.py scrape -k python -k remote
# Limit the number of jobs per board
python main.py scrape -m 50
Search & Filter Jobs:
# Search by keyword
python main.py search -k python
# Filter for remote jobs
python main.py search --remote
# Complex filter (keyword, location, company)
python main.py search -k "data scientist" -l remote --company google
# Export results
python main.py search -k python --export results.csv
Manage Automation:
# Start/stop the daily scheduler
python main.py schedule --start
python main.py schedule --stop
# Run a scraping job immediately
python main.py schedule --run-now
Python API
You can also use the aggregator directly in your scripts:
from main import JobAggregator
app = JobAggregator()
jobs = app.scrape_all_jobs(keywords=['python', 'remote'])
app.save_jobs(jobs)
remote_jobs = app.search_jobs(remote_only=True, keyword='developer')
for job in remote_jobs[:5]:
print(f"{job.title} at {job.company}")
Input / Output
- Input: The tool takes its input from configured job board websites. You control the process via CLI commands (keywords, limits) or the configuration file (storage type, scheduling).
- Processing: It scrapes, normalizes fields (title, company, location, link), and deduplicates listings.
- Output: Jobs are saved to your chosen storage backend (JSON file or SQLite DB). Search results can be viewed in the terminal or exported to CSV/JSON files.
Categories
Common Use Cases
Market Research
Gather competitive intelligence and market data
Lead Generation
Extract contact information for sales outreach
Price Monitoring
Track competitor pricing and product changes
Content Aggregation
Collect and organize content from multiple sources
Ready to Get Started?
Try Job Listings Aggregator Pro now on Apify. Free tier available with no credit card required.
Start Free TrialActor Information
- Developer
- assertive_analogy
- Pricing
- Paid
- Total Runs
- 55
- Active Users
- 20
Related Actors
Company Employees Scraper
by build_matrix
🔥 LinkedIn Jobs Scraper
by bebity
Linkedin Company Detail (No Cookies)
by apimaestro
Linkedin Profile Details Batch Scraper + EMAIL (No Cookies)
by apimaestro
Apify provides a cloud platform for web scraping, data extraction, and automation. Build and run web scrapers in the cloud.
Learn more about ApifyNeed Professional Help?
Couldn't solve your problem? Hire a verified specialist on Fiverr to get it done quickly and professionally.
Trusted by millions | Money-back guarantee | 24/7 Support