LinuxQuestions.org

LinuxQuestions.org (/questions/)
-   Linux - Newbie (https://www.linuxquestions.org/questions/linux-newbie-8/)
-   -   wget problems (https://www.linuxquestions.org/questions/linux-newbie-8/wget-problems-801816/)

linf 04-13-2010 02:40 PM

wget problems
 
hi not sure if this is the right place to ask this question but here goes...
im currently having problems mirroring a web site with wget (or any other program for that mater). I have used wget many times before without any problem so i dont think its anything i am doing wrong. All that happens is it only grabs the first page (index.html).
the command i use in this case is:

Code:

wget -r -m "name of site"
any help would be much appreciated.

rweaver 04-13-2010 02:52 PM

If name of the site is say... "LinuxQuestions" it's going to fail, that's not a url. You're not providing any details on what the problem actually is or errors or well anything other than its not working so its pretty hard to help you....

linf 04-13-2010 03:06 PM

thanks for the quick reply.
I understand its not exactly a linux related question although i was hoping that i would get some help. i am not sure what other information to give other than the url to the site (greenroomswirral.co.uk). if you need any other information then i am happy to provide it.

harryhaller 04-13-2010 03:09 PM

What is needed is that you copy/paste the exact command that you entered as well as any messages from wget.

pixellany 04-13-2010 03:20 PM

Tried it from here---It seems to be downloading many different pages (I stopped it before it finished)

linf 04-13-2010 03:44 PM

Code:

wget -r -m www.greenroomswirral.co.uk
wget gives no errors it just stops at the first page (index.html). i have never seen it do this before with any other site.

this is the output

Code:

lina@lina-laptop:~/Desktop/new$ wget -r -m www.greenroomswirral.co.uk
--2010-04-13 21:19:43--  http://www.greenroomswirral.co.uk/
Resolving www.greenroomswirral.co.uk... 79.170.44.108
Connecting to www.greenroomswirral.co.uk|79.170.44.108|:80... connected.
HTTP request sent, awaiting response... 200 OK
Length: unspecified [text/html]
Last-modified header missing -- time-stamps turned off.
--2010-04-13 21:19:43--  http://www.greenroomswirral.co.uk/
Connecting to www.greenroomswirral.co.uk|79.170.44.108|:80... connected.
HTTP request sent, awaiting response... 200 OK
Length: unspecified [text/html]
Saving to: `www.greenroomswirral.co.uk/index.html'

    [  <=>                                  ] 14,137      45.9K/s  in 0.3s   

2010-04-13 21:19:44 (45.9 KB/s) - `www.greenroomswirral.co.uk/index.html' saved [14137]

FINISHED --2010-04-13 21:19:44--
Downloaded: 1 files, 14K in 0.3s (45.9 KB/s)
lina@lina-laptop:~/Desktop/new$

Quote:

Tried it from here---It seems to be downloading many different pages (I stopped it before it finished)
i have tried this on a few different computer with different networks with the same result so i would be interested in the commands you have used

pixellany 04-13-2010 05:10 PM

I did it without the "www"---and got many many pages.

WITH the "www", I get your result.

I have NO CLUE why the difference

linf 04-13-2010 06:09 PM

That works perfect though i am not sure why...(would be interested to know why though).
Thanks very much for all you help and patience.


All times are GMT -5. The time now is 07:34 PM.