LinuxQuestions.org
Download your favorite Linux distribution at LQ ISO.
Go Back   LinuxQuestions.org > Forums > Linux Forums > Linux - Newbie
User Name
Password
Linux - Newbie This Linux forum is for members that are new to Linux.
Just starting out and have a question? If it is not in the man pages or the how-to's this is the place!

Notices


Reply
  Search this Thread
Old 01-22-2008, 12:58 PM   #1
clwhitt
LQ Newbie
 
Registered: Apr 2004
Location: Washougal, WA
Distribution: Ubuntu Hardy Heron (8.04)
Posts: 19

Rep: Reputation: 0
import entire web site


What I would like to do is import an entire web site onto my local computer so that I can work with the web site with Bluefish while off line (these are my web sites).
I've done it once, but for the life of me can't remember how I did it. I'm pretty sure is was a command line process that got it done.

Thanks,
Chuck
Ubuntu Gutsy Gibbon
 
Old 01-22-2008, 01:03 PM   #2
David the H.
Bash Guru
 
Registered: Jun 2004
Location: Osaka, Japan
Distribution: Debian sid + kde 3.5 & 4.4
Posts: 6,823

Rep: Reputation: 1958Reputation: 1958Reputation: 1958Reputation: 1958Reputation: 1958Reputation: 1958Reputation: 1958Reputation: 1958Reputation: 1958Reputation: 1958Reputation: 1958
I suggest httrack. It's a dedicated cli site downloader that's very powerful and flexible. There's also a web interface available for it.

wget is also able to mirror whole sites, but I think it has trouble with certain directory structures or something. At least I've seen posts by people who've had problems before.
 
Old 01-22-2008, 01:07 PM   #3
pljvaldez
LQ Guru
 
Registered: Dec 2005
Location: Somewhere on the String
Distribution: Debian Wheezy (x86)
Posts: 6,094

Rep: Reputation: 271Reputation: 271Reputation: 271
A quick google turned up these:

http://linuxreviews.org/quicktips/wget/
http://www.httrack.com/
 
Old 01-22-2008, 01:20 PM   #4
clwhitt
LQ Newbie
 
Registered: Apr 2004
Location: Washougal, WA
Distribution: Ubuntu Hardy Heron (8.04)
Posts: 19

Original Poster
Rep: Reputation: 0
import entire web site

I'll look at httrack and see how that works. I think though, I might have used wget to do the job before (it rings bells), though I just can't remember exactly how I did it.
pljvaldez - what search phrase did you use? I consider myself pretty good at using Google, but spent a couple of frustrating hours trying to find any useful information that was not related to FrontPage.

Chuck
 
Old 01-22-2008, 01:27 PM   #5
pljvaldez
LQ Guru
 
Registered: Dec 2005
Location: Somewhere on the String
Distribution: Debian Wheezy (x86)
Posts: 6,094

Rep: Reputation: 271Reputation: 271Reputation: 271
I used "linux download entire website".

The wget article was first, and the httrack one was third.
 
Old 01-22-2008, 01:44 PM   #6
clwhitt
LQ Newbie
 
Registered: Apr 2004
Location: Washougal, WA
Distribution: Ubuntu Hardy Heron (8.04)
Posts: 19

Original Poster
Rep: Reputation: 0
import entire web site

I tried that without the "entire" and with "web site" as separate words, and all I got was a bunch of hits on downloading Live CD's. I tried "import" rather than "download", and a whole bunch of other variations too. Dang it, I was so close!
I did use wget before. Following these messages, and knowing what I was looking for, I checked my terminal command line and saw that I had used wget with the recursive switch (-r) to download the website.
I'm still going to read up on httrack to see how it works. BTW, the link to wget you posted was an interesting read in how it can be manipulated to fool other websites that put up obstacles to downloading a website.

Thanks for you help,
Chuck
 
Old 01-23-2008, 06:53 PM   #7
clwhitt
LQ Newbie
 
Registered: Apr 2004
Location: Washougal, WA
Distribution: Ubuntu Hardy Heron (8.04)
Posts: 19

Original Poster
Rep: Reputation: 0
import entire web site

As a follow-up here, I was able to download my web site quickly, completely and easily with wget -r http://www.obabytheboat.com. WebHTTrack on the other hand spent several hours parsing the site and did not get it all before I lost patience. I haven't spent any more time with it yet to figure out where the fault lies. I also have not tried the cli version yet, so I may get to that just for kicks. I'll try to remember to post a report on what I find out.
Thanks everyone for their help.
Chuck
 
  


Reply


Thread Tools Search this Thread
Search this Thread:

Advanced Search

Posting Rules
You may not post new threads
You may not post replies
You may not post attachments
You may not edit your posts

BB code is On
Smilies are On
[IMG] code is Off
HTML code is Off



Similar Threads
Thread Thread Starter Forum Replies Last Post
PHP password protect entire web directory newuser455 Programming 1 06-10-2006 01:22 PM
get Squidguard to block entire countries web sites? Possible? Pcghost Linux - Software 3 09-19-2004 02:25 PM
import site failed hartb47 Linux - Newbie 3 06-18-2004 05:10 AM
how can i donload and view entire web site offline !! hitesh_linux Linux - Software 8 03-08-2004 10:32 AM
up2date problem: 'import site" failed docli Linux - Newbie 1 12-05-2003 09:40 PM


All times are GMT -5. The time now is 02:37 AM.

Main Menu
Advertisement
My LQ
Write for LQ
LinuxQuestions.org is looking for people interested in writing Editorials, Articles, Reviews, and more. If you'd like to contribute content, let us know.
Main Menu
Syndicate
RSS1  Latest Threads
RSS1  LQ News
Twitter: @linuxquestions
Facebook: linuxquestions Google+: linuxquestions
Open Source Consulting | Domain Registration