Idle Banter For non SV and non bike related chat (and the odd bit of humour - but if any post isn't suitable it'll get deleted real quick).![]() |
![]() |
|
Thread Tools |
![]() |
#1 |
Member
Mega Poster
Join Date: Mar 2005
Location: Llanwrtyd Wells Powys
Posts: 1,146
|
![]()
I want to download the contents of a website so i can view it all offline, as its a right chore to keep reloading each time.
The contents of the website are a huge collecion of .pdf files, themselves referred to by a master .pdf file. The root or master .pdf file is here: http://www.hants.gov.uk/rh/row/maps/definitive-maps.pdf I've tried a website download tool (HTTrack website copier) but either I'm not driving it right or it cant see the individual map files on the site. What I'd like to end up with is a directory with the root page as listed above that will link and load all the map squares as it does online, that I can also burn to DVD and take along to my meetings. any ideas? EDIT: I've managed it, very crudely. Used excel to create every possible xxxx.pdf weblink combination, then pasted that into a download program. It will take a while, but should work I would still be interested in a 'proper' way of achieving it though Last edited by tigersaw; 29-08-10 at 01:27 AM. |
![]() |
![]() |
![]() |
#2 |
Member
Mega Poster
Join Date: Mar 2004
Location: Not in Yorkshire. (Thank God)
Posts: 4,116
|
![]()
wget is the tool of choice for downloading web sites/pages, using the recursive mode switch. As it will parse the root page and get any pages linked, even follow links to other sites etc. I use it all the time when diagnosing site problems etc. One simple command and suddenly you have all a sites content.
__________________
Not Grumpy, opinionated. |
![]() |
![]() |
![]() |
#3 |
Member
Mega Poster
Join Date: Mar 2004
Location: Not in Yorkshire. (Thank God)
Posts: 4,116
|
![]()
Just re read your original post, I hadn't realised your trying to follow a .pdf trail, good luck. there may be a tool. or you could start with the wget source and add .pdf parsing as I don't think it would manage that as it was designed to parse html.
If it is your site, I would be inclined to write a bit of php to tree walk your directory and offer every file as a link. Then use this php page as the wget target.
__________________
Not Grumpy, opinionated. Last edited by timwilky; 29-08-10 at 11:02 AM. |
![]() |
![]() |
![]() |
#4 |
Member
Mega Poster
Join Date: Mar 2005
Location: Llanwrtyd Wells Powys
Posts: 1,146
|
![]()
Thanks for your thoughts - I did succeed in the end, using HTTrack. I used excel to produce a huge txt file of all possible combinations, pointed HTTrack at it and off it went, downloading all the valid ones.
It took all night, now I have all the files (1Gb), but the master page obviously does not point to my cached versions, it still looks out to the www, but I'm happy - I can now take a stored copy of all the rights of way with me to meetings etc. |
![]() |
![]() |
![]() |
|
|
![]() |
||||
Thread | Thread Starter | Forum | Replies | Last Post |
MP3 downloading | pencil shavings | Idle Banter | 15 | 22-05-08 03:56 PM |
downloading from youtube | Shinsei Jutsu | Idle Banter | 3 | 15-05-07 11:57 AM |
downloading games | hovis | Idle Banter | 9 | 23-10-06 01:26 PM |
Downloading clips | nickj | Idle Banter | 8 | 28-09-06 07:11 PM |
downloading movies | hovis | Idle Banter | 14 | 20-09-06 08:03 AM |