Reddit Scraper Lite
by trudax
Need to scrape Reddit for data but don't want the hassle of logins, rate limits, or complex setups? Reddit Scraper Lite is built for that. It’s a stra...
Opens on Apify.com
About Reddit Scraper Lite
Need to scrape Reddit for data but don't want the hassle of logins, rate limits, or complex setups? Reddit Scraper Lite is built for that. It’s a straightforward, pay-per-result scraper that lets you crawl posts, comments, subreddits, and user profiles without needing an account. You control the scope. Just tell it how many posts or items you need, and it handles the rest, pulling all the data into a clean dataset. It exports in multiple formats, so you can get straight to analysis in your preferred tool. I use it for gathering public sentiment on products, tracking discussions in specific communities, or collecting datasets for training simple models. It’s especially handy for market research, academic projects, or keeping an eye on brand mentions. The best part is the pricing model—you only pay for what you successfully scrape. No subscriptions, no guessing. If you need a no-fuss way to get structured data from Reddit’s public pages, this scraper gets the job done.
What does this actor do?
Reddit Scraper Lite is a web scraping and automation tool available on the Apify platform. It's designed to help you extract data and automate tasks efficiently in the cloud.
Key Features
- Cloud-based execution - no local setup required
- Scalable infrastructure for large-scale operations
- API access for integration with your applications
- Built-in proxy rotation and anti-blocking measures
- Scheduled runs and webhooks for automation
How to Use
- Click "Try This Actor" to open it on Apify
- Create a free Apify account if you don't have one
- Configure the input parameters as needed
- Run the actor and download your results
Documentation
What does Reddit Scraper do? Our unofficial Reddit API will get data from Reddit with no limitations or authentication. It enables you to extract posts and comments together with some user info without login. It is built on top of Apify SDK, and you can run it both on the Apify platform. Reddit Scraper allows you to: - scrape subreddits (communities) with top posts - scrape Reddit posts with title and text, username, number of comments, votes, media elements. - get Reddit comments, timestamps, points, usernames, post and comment URLs. - scrape user details, their most recent posts and comments. - sort scraped data by categories of Relevance, Hot, Top, and New. - scrape data using a specific URL or by keyword. ## What data can I extract from Reddit?
| 📌 Popular subreddits | 🔍 Subreddit details |
| 📋 Subreddit name | 👥 Number of members |
| 🌐 Community URL | 📚 Category |
| 📌 Reddit posts | 💬 Reddit comments |
| 📃 Title and text | ⏱ Timestamps |
| 👤 Username | 🔗 Post and comment URLs |
| 👍 Votes | 📷 Media elements |
| 👤 User details | 📄 Recent posts and comments |
search URL as a parameter for startUrls, it will only scrape for posts. If you want to search for communities and users, use the search field or a specific URL instead. ### How to scrape Reddit by search term - Search Term or searches - the keywords you want to search via the Reddit's search engine. You can keep one field or add as many as you want. Don't use this field if you're using the startUrls parameter. ### Advanced parameters - Include NSFW content or includeNSFW - Toggle to include or exclude NSFW (Not Safe For Work) content from your search results. - Search for posts or searchPosts - If enabled, the scraper will search for posts matching the keyword. - Search for comments or searchComments - If enabled, the scraper will search for comments matching the keyword. - Search for communities or searchCommunities - If enabled, the scraper will search for communities matching the keyword. - Search for users or searchUsers - If enabled, the scraper will search for users matching the keyword. - Skip comments or skipComments - If enabled, the scraper will not extract comments when scraping posts. - Skip user posts or skipUserPosts - If enabled, the scraper will not extract user posts when scraping user profiles. - Skip community or skipCommunity - If enabled, the scraper will not extract community details when scraping communities. - Debug Mode or debugMode - If enabled, the scraper will output detailed logs to help with debugging. - Page scroll timeout or scrollTimeout - Set the timeout in seconds for how long the page should scroll down to load new items. - Ignore start URLs or ignoreStartUrls - If enabled, the scraper will ignore the startUrls field. This is mainly used as a fix for integrations like Zapier. - Sort search or sort - will sort search results by Relevance, Hot, Top, New or most amount of Comments. - Filter by date or time - will filter the search by the last hour, day, week, month or year. Only available if you're scraping Posts. - Post date limit or postDateLimit - Use this value to only retrieve posts published after a specific date. When this parameter is used, the 'Sort search' (sort) parameter will automatically be set to 'New' to ensure chronological consistency. To see the full list of parameters, their default values, and how to set the values of your own, head over to Input Schema tab. ### Input example This is an example of how your input will look like if you decide to scrape all Reddit communities that contain the keyword parrot. Results will be sorted by the newest first. json { "maxItems": 10, "maxPostCount": 10, "maxComments": 10, "maxCommunitiesCount": 10, "maxUserCount": 10, "maxLeaderBoardItems": 10, "scrollTimeout": 40, "proxy": { "useApifyProxy": true }, "searches": ["parrots"], "type": "community", "sort": "new", "time": "all" } ## Results The output from scraping Reddit is stored in the dataset. Each post, comment, user or community is stored as an item inside the dataset. After the run is finished, you can download the scraped data onto your computer or export to any web app in various data formats (JSON, CSV, XML, RSS, HTML Table). Here's a few examples of the outputs you can get for different types of inputs: ### 📝 Example Reddit post json { "id": "t3_144w7sn", "parsedId": "144w7sn", "url": "https://www.reddit.com/r/HonkaiStarRail/comments/144w7sn/my_luckiest_10x_pull_yet/", "username": "YourKingLives", "title": "My Luckiest 10x Pull Yet", "communityName": "r/HonkaiStarRail", "parsedCommunityName": "HonkaiStarRail", "body": "URL: https://i.redd.it/yod3okjkgx4b1.jpg\nThumbnail: https://b.thumbs.redditmedia.com/lm9KxS4laQWgx4uOoioM3N7-tBK3GLPrxb9da2hGtjs.jpg\nImages:\n\thttps://preview.redd.it/yod3okjkgx4b1.jpg?auto=webp&v=enabled&s=be5faf0250e19138b82c7bbe5e7406fa46da4e73\n", "html": null, "numberOfComments": 0, "upVotes": 1, "isVideo": false, "isAd": false, "over18": false, "createdAt": "2023-06-09T05:23:15.000Z", "scrapedAt": "2023-06-09T05:23:28.409Z", "dataType": "post" }, ### 💬 Example Reddit comment json { "id": "t1_jnhqrgg", "parsedId": "jnhqrgg", "url": "https://www.reddit.com/r/NewsWithJingjing/comments/144v5c3/theres_no_flag_large_enough/jnhqrgg/", "parentId": "t3_144v5c3", "username": "smokecat20", "category": "NewsWithJingjing", "communityName": "r/NewsWithJingjing", "body": "A true patriot.", "createdAt": "2023-06-09T05:00:00.000Z", "scrapedAt": "2023-06-09T05:23:32.025Z", "upVotes": 3, "numberOfreplies": 0, "html": "<div class=\"md\"><p>A true patriot.</p>\n</div>", "dataType": "comment" } ### 👥 Example Reddit community json { "id": "2qlhq", "name": "t5_2qlhq", "title": "Pizza", "headerImage": "https://b.thumbs.redditmedia.com/jq9ytPEOecwd5bmGIvNQzjTPE9hdd0kB9XGa--wq55A.png", "description": "The home of pizza on reddit. An educational community devoted to the art of pizza making.", "over18": false, "createdAt": "2008-08-26T00:03:48.000Z", "scrapedAt": "2023-06-09T05:16:55.443Z", "numberOfMembers": 569724, "url": "https://www.reddit.com/r/Pizza/", "dataType": "community" } ### 👤 Example Reddit user json { "id": "c3h2qmv", "url": "https://www.reddit.com/user/jancurn/", "username": "jancurn", "userIcon": "https://www.redditstatic.com/avatars/defaults/v2/avatar_default_7.png", "postKarma": 4, "commentKarma": 10, "description": "", "over18": false, "createdAt": "2018-09-10T15:13:39.000Z", "scrapedAt": "2023-06-09T05:21:14.409Z", "dataType": "user" } ## Only need a few Reddit results? Use our super fast dedicated Reddit Scraper Lite if you want to scrape Reddit data on a smaller scale. Just enter one or more Reddit URLs or keywords and click to scrape. ## Notes for developers ### Limiting results with maxItems If you need to limit the scope of your search, you can do that by setting the max number of posts you want to scrape inside a community or user. You can also set a limit to the number of comments for each post. You can limit the number of communities and the number of leaderboards by using the following parameters: json { "maxItems": 100, "maxPostCount": 50, "maxComments": 10, "maxCommunitiesCount": 5, "maxUserCount": 5, "maxLeaderBoardsItems": 5 } You can also set maxItems to prevent a very long run of the Actor. This parameter will stop your scraper when it reaches the number of results you've indicated, so you need to be careful not to trim your results. See the Input Schema tab for the full list of the ways to restrict Reddit Scraper using these parameters: maxItems, maxPostCount, maxComments, maxCommunitiesCount, maxLeaderBoardItems ## FAQ ### Is Reddit scraping legal? While scraping publicly available data from Reddit is generally allowed, it's important to comply with Reddit's terms of service and respect the site's usage policies. It's recommended to use the scraper responsibly, avoid excessive requests, and ensure that the scraped data is used in compliance with applicable laws and regulations. You can read more about compliance with ToS in our blogpost. ### Can I use Reddit API to scrape Reddit? The Reddit API is currently free. However, Reddit has specific API rules, and free access to the API will be restricted in the future due to concerns over data usage. In addition, Reddit API has been recently announced to become a paid service. Using a Reddit web scraper such as this one as an API has advantages such over the official one such as not requiring authentication, special authorization for commercial use, or registration for a token. ### How can I scrape Reddit comments? Reddit Scraper allows scraping specific parts of Reddit, including comments. You can extract posts and comments along with user information, such as timestamps, number of votes, usernames, post URL, and comment URLs. This enables you to gather comprehensive commment data from subreddits and Reddit users. ### Is it necessary to use cookies for accessing logged-in content when scraping Reddit? No, it is not required. As of May 2023, Reddit maintains its data publicly accessible and does not enforce a login barrier. ### Do you need proxies for scraping Reddit? It is highly recommended. Subreddits are open for access and do not require a login to retrieve information. Typically, using proxies is necessary to ensure successful Reddit scraping. While some results can be obtained with datacenter proxies, residential proxies are preferred for Reddit scraping. Fortunately, our Free plan offers a trial of Apify Proxy, which get you started. ### Can I export or import scraped Reddit data using API? Yes. The Apify API gives you programmatic access to the Apify platform. The API is organized around RESTful HTTP endpoints that enable you to manage, schedule, and run any Apify Actor, including this one. The API also lets you access any datasets, monitor Actor performance, fetch results, create and update versions, and more. To access the API using Node.js, use the apify-client NPM package. To access the API using Python, use the apify-client PyPi package. Check out the Apify API reference docs for full details or click on the API tab for code examples. ### What is the difference between Reddit Scraper and Reddit Scraper Lite? The functionality of both actor are the same. The difference is that the Lite version charges based on the number of results it extracts and it has a custom proxy configuration integrated with the actor that is cheaper than then the Residential proxies used by Apify. The Reddit Scraper version allows you to use either one of Apify`s proxies or a custom one.
Categories
Common Use Cases
Market Research
Gather competitive intelligence and market data
Lead Generation
Extract contact information for sales outreach
Price Monitoring
Track competitor pricing and product changes
Content Aggregation
Collect and organize content from multiple sources
Ready to Get Started?
Try Reddit Scraper Lite now on Apify. Free tier available with no credit card required.
Start Free TrialActor Information
- Developer
- trudax
- Pricing
- Paid
- Total Runs
- 2,388,075
- Active Users
- 11,533
Related Actors
🏯 Tweet Scraper V2 - X / Twitter Scraper
by apidojo
Instagram Scraper
by apify
TikTok Scraper
by clockworks
Instagram Profile Scraper
by apify
Apify provides a cloud platform for web scraping, data extraction, and automation. Build and run web scrapers in the cloud.
Learn more about ApifyNeed Professional Help?
Couldn't solve your problem? Hire a verified specialist on Fiverr to get it done quickly and professionally.
Trusted by millions | Money-back guarantee | 24/7 Support