Forum Home
Press F1
 
Thread ID: 61722 2005-09-14 11:21:00 How to check a websites total content. John W (523) Press F1
Post ID Timestamp Content User
388359 2005-09-14 11:21:00 Hi there

If I click onto a Website, short of clicking onto every link, is there a way I can view the number of pages it contains and how they are structured?

Thanks ...............John in Mosgiel.
John W (523)
388360 2005-09-14 12:46:00 I can think of only two ways:
1. Google. Use this search term: site:sitename. It should list all pages and subpages it knows about. Note that dynamically-generated content can complicate matters - you will need an operator like -php to exclude php generated pages for example.
2. Use a browser that can list links, such as the text-only browser lynx. Runs on all *nix and DOS.

I have vague memory of an app that downloads all pages of a site in the background whilst you are viewing a page, but I can't remember the name.

If you have directory access to the site (unlikely if it is not yours) then you can see it.
vinref (6194)
388361 2005-09-14 13:01:00 I have vague memory of an app that downloads all pages of a site in the background whilst you are viewing a page, but I can't remember the name.
One of the more popular ones I know of (I dont use it) is HTTrack (http://www.httrack.com/). Basically 'backsup' a website recursively.
sal (67)
388362 2005-09-14 13:10:00 Good one sal. That's the one (I think...). vinref (6194)
388363 2005-09-14 17:00:00 One of the more popular ones I know of (I dont use it) is HTTrack (http://www.httrack.com/). Basically 'backsup' a website recursively.
I also got that advice from someone here on PressF1 a while back. It's goodly.

Mozzila/Netscape users can also right-click and View Page Info to get a lot of info about the page that's currently viewed.
Greg (193)
1