| Forum Home | ||||
| Press F1 | ||||
| Thread ID: 61722 | 2005-09-14 11:21:00 | How to check a websites total content. | John W (523) | Press F1 |
| Post ID | Timestamp | Content | User | ||
| 388359 | 2005-09-14 11:21:00 | Hi there If I click onto a Website, short of clicking onto every link, is there a way I can view the number of pages it contains and how they are structured? Thanks ...............John in Mosgiel. |
John W (523) | ||
| 388360 | 2005-09-14 12:46:00 | I can think of only two ways: 1. Google. Use this search term: site:sitename. It should list all pages and subpages it knows about. Note that dynamically-generated content can complicate matters - you will need an operator like -php to exclude php generated pages for example. 2. Use a browser that can list links, such as the text-only browser lynx. Runs on all *nix and DOS. I have vague memory of an app that downloads all pages of a site in the background whilst you are viewing a page, but I can't remember the name. If you have directory access to the site (unlikely if it is not yours) then you can see it. |
vinref (6194) | ||
| 388361 | 2005-09-14 13:01:00 | I have vague memory of an app that downloads all pages of a site in the background whilst you are viewing a page, but I can't remember the name. One of the more popular ones I know of (I dont use it) is HTTrack (http://www.httrack.com/). Basically 'backsup' a website recursively. |
sal (67) | ||
| 388362 | 2005-09-14 13:10:00 | Good one sal. That's the one (I think...). | vinref (6194) | ||
| 388363 | 2005-09-14 17:00:00 | One of the more popular ones I know of (I dont use it) is HTTrack (http://www.httrack.com/). Basically 'backsup' a website recursively. I also got that advice from someone here on PressF1 a while back. It's goodly. Mozzila/Netscape users can also right-click and View Page Info to get a lot of info about the page that's currently viewed. |
Greg (193) | ||
| 1 | |||||