Verified:

119

Member
145

May 17th 2011, 6:21:14

I'm trying to archive a website/blog and all of its contents. I tried several different methods but all of them didn't turn out quite as I was hoping, and the wget man is quite long and extensive.

Each time I tried to wget site it just downloads the index page. The site pages are setup as so:

http://www.site.com
http://www.site.com/page/1
http://www.site.com/page/2

Each page has a post that should also be downloaded. I don't need to downloads the links to external sites.. just the pages on that domain.

So far I have tried

wget -m http://www.site.com
wget -r -l inf -k -E -p -nc http://www.site.com

Anyone know the proper options that I should use to get this working right?

Edited By: 119 on May 17th 2011, 6:27:32
Back To Thread
See Subsequent Edit