Extract millions of posts and memes from Imgur. Crawl and scrape descriptions, number of views, favorites, upvotes, downvotes, score, comment details, post creator, and all other deep-level details. You can specify search terms, filters, tags, list pages, and much more! Extremely fast, no limits!
Since Imgur doesn't provide a good and free API, this actor should help you to retrieve data from it.
The Imgur data scraper supports the following features:
Search anything - You can search any keyword with any sorting or filtering option. Provide the options you require and retrieve everything in a blazing-fast manner!
Scrape tags - Looking for posts of a specific tag? No problem. Provide the URL and the Imgur Scraper will retrieve everything in no time.
Scrape posts of any user - You can retrieve all the posts of any user without any limits.
Scrape comments - If you are looking for the comments that have been added by the users, you are in the right place. Within the powerful algorithm of Imgur Scraper, you can extract all the comments (and replies) directly!
Scrape posts - Scrape very detailed information for each of the posts that you'd like to get.
This scraper is under active development. If you have any feature requests you can create an issue from here.
The input of this scraper should be JSON containing the list of pages on Imgur that should be visited. Required fields are:
search
: (Optional) (String) Keyword that you want to search on Imgur.
startUrls
: (Optional) (Array) List of Imgur URLs. You should only provide post detail, search, user, or tag URLs.
includeComments
: (Optional) (Boolean) This will add all the comments that Imgur provides into the post objects. Please keep in mind that the time and resources the actor uses will increase proportionally to the number of comments.
endPage
: (Optional) (Number) Final number of page that you want to scrape. The default is Infinite
. This applies to all search
requests and startUrls
individually.
maxItems
: (Optional) (Number) You can limit scraped items. This should be useful when you search through the big lists or search results.
proxy
: (Required) (Proxy Object) Proxy configuration.
extendOutputFunction
: (Optional) (String) Function that takes a JQuery handle ($) as an argument and returns an object with data.
customMapFunction
: (Optional) (String) Function that takes each object's handle as an argument and returns the object with executing the function.
This solution requires the use of Proxy servers, either your own proxy servers or you can use Apify Proxy.
When you want to scrape over a specific list URL, just copy and paste the link as one of the startUrl.
If you would like to scrape only the first page of a list then put the link for the page and have the endPage
as 1.
With the last approach that is explained above you can also fetch any interval of pages. If you provide the 5th page of a list and define the endPage
parameter as 6 then you'll have the 5th and 6th pages only.
The actor is optimized to run blazing fast and scrape as many items as possible. Therefore, it forefronts all the detailed requests. If the actor doesn't block very often it'll scrape 100 listings in 2 minutes with ~0.025-0.03 compute units.
1{ 2 "startUrls": [ 3 "https://imgur.com/user/IsNice", 4 "https://imgur.com/t/burrito", 5 "https://imgur.com/gallery/ILtkgRA", 6 "https://imgur.com/search?q=nice%20word" 7 ], 8 "search": "super meme", 9 "includeComments": false, 10 "proxy": { 11 "useApifyProxy": true 12 }, 13 "endPage": 5, 14 "maxItems": 10 15}
During the run, the actor will output messages letting you know what is going on. Each message always contains a short label specifying which page from the provided list is currently specified. When items are loaded from the page, you should see a message about this event with a loaded item count and total item count for each page.
If you provide incorrect input to the actor, it will immediately stop with a failure state and output an explanation of what is wrong.
During the run, the actor stores results into a dataset. Each item is a separate item in the dataset.
You can manage the results in any language (Python, PHP, Node JS/NPM). See the FAQ or our API reference to learn more about getting results from this Imgur actor.
The structure of each item in Imgur looks like this:
1{ 2 "type": "post", 3 "id": "JBTJqu2", 4 "accountId": "22440270", 5 "title": "holiday", 6 "description": "", 7 "numberOfViews": 8578, 8 "numberOfUpvotes": 23, 9 "numberOfDownvotes": 5, 10 "numberOfPoints": 18, 11 "numberOfImages": 1, 12 "numberOfComments": 6, 13 "numberOfFavorites": 1, 14 "virality": 3475.255272505103, 15 "score": 18.5435, 16 "isInMostViral": false, 17 "isAlbum": true, 18 "isMature": false, 19 "coverId": "dK9p4A1", 20 "createdAt": "2019-07-13T03:20:56Z", 21 "updatedAt": null, 22 "url": "https://imgur.com/gallery/JBTJqu2", 23 "platform": "api", 24 "account": { 25 "id": "22440270", 26 "username": "IsNice", 27 "avatarUrl": "https://i.imgur.com/YLvWS5K_d.png?maxwidth=290&fidelity=grand", 28 "createdAt": "2015-07-19T10:05:29Z" 29 }, 30 "tags": [ 31 "storytime", 32 "funny", 33 "awesome" 34 ], 35 "media": [ 36 { 37 "mime_type": "video/mp4", 38 "url": "https://i.imgur.com/dK9p4A1.mp4", 39 "ext": "mp4", 40 "width": 960, 41 "height": 540, 42 "size": 87699, 43 "title": "", 44 "description": "", 45 "isAnimated": true, 46 "isLooping": true, 47 "duration": 9, 48 "has_sound": false 49 } 50 ], 51 "comments": [ 52 { 53 "id": 1681576587, 54 "parent_id": 0, 55 "comment": "https://youtu.be/q-qqrGtlHkg", 56 "account_id": 3708825, 57 "post_id": "JBTJqu2", 58 "upvote_count": 2, 59 "downvote_count": 0, 60 "point_count": 2, 61 "vote": null, 62 "platform_id": 4, 63 "platform": "android", 64 "created_at": "2019-07-13T03:45:15Z", 65 "updated_at": "2019-07-13T04:05:31Z", 66 "deleted_at": null, 67 "next": null, 68 "comments": [ 69 { 70 "id": 1681616815, 71 "parent_id": 1681576587, 72 "comment": "http://i.imgur.com/D5veJQj.gif", 73 "account_id": 85975675, 74 "post_id": "JBTJqu2", 75 "upvote_count": 2, 76 "downvote_count": 0, 77 "point_count": 2, 78 "vote": null, 79 "platform_id": 4, 80 "platform": "android", 81 "created_at": "2019-07-13T05:28:37Z", 82 "updated_at": "2019-07-13T18:44:25Z", 83 "deleted_at": null, 84 "next": null, 85 "comments": [] 86 } 87 ] 88 } 89 ] 90}
Please visit us through epctex.com to see all the products that are available for you. If you are looking for any custom integration or so, please reach out to us through the chat box in epctex.com. In need of support? business@epctex.com is at your service.
Yes, if you're scraping publicly available data for personal or internal use. Always review Websute's Terms of Service before large-scale use or redistribution.
No. This is a no-code tool — just enter a job title, location, and run the scraper directly from your dashboard or Apify actor page.
It extracts job titles, companies, salaries (if available), descriptions, locations, and post dates. You can export all of it to Excel or JSON.
Yes, you can scrape multiple pages and refine by job title, location, keyword, or more depending on the input settings you use.
You can use the Try Now button on this page to go to the scraper. You’ll be guided to input a search term and get structured results. No setup needed!