wget searching and downloading images on a website ?
Linux - SoftwareThis forum is for Software issues.
Having a problem installing a new program? Want to know which application is best for the job? Post your question in this forum.
Welcome to LinuxQuestions.org, a friendly and active Linux Community.
You are currently viewing LQ as a guest. By joining our community you will have the ability to post topics, receive our newsletter, use the advanced search, subscribe to threads and access many other special features. Registration is quick, simple and absolutely free. Join our community today!
Note that registered members see fewer ads, and ContentLink is completely disabled once you log in.
wget how to specify a numeric range and download images from a website ?
I was needing images on a server which are on the server but i have to access them manually.
It would be for example www.xyz.com/collection/846_02_900_x.jpg
When i change to /collection/847_02_900_x.jpg (just one number higher) there is a picture as well, and i need all the pictures in the collection folder.. However, if i just go to www.xyz.com/collection there are no images, it just says the there was an error... So without me having to sit 3 hours and do it all manually, going from 846_02_900_x.jpg 847_02_900_x.jpg 848_02_900_x.jpg etc.. i was wondering if wget can download the pictures somehow?
When i try it with wget -m -p -k http://www.xyz.com/collection all i get is HTTP request sent, awaiting response... 404 Not Found
And of course when i do wget -m -p -k www.xyz.com/collection/846_02_900_x.jpg it download only that image. The question is if there is a possibility to set it to increase certain values automatically such as 846, and download accordingly, without changing any other numbers ?
come on.. scraping multiple images in this fashion is the fastest way to grow your Pr0n collection..
But seriously though, all kidding aside, have you looked at httrack ? it's another option you could try.
Description: Copy websites to your computer (Offline browser)
HTTrack is an offline browser utility, allowing you to download a World Wide website from the Internet to a local directory, building
recursively all directories, getting html, images, and other files from the server to your computer.
HTTrack arranges the original site's relative link-structure.