|This page was nominated for deletion on 10 March 2013 (UTC). The result of the discussion was keep.|
|WikiProject Computing / Software||(Rated Stub-class)|
Please review my English
Please review my English!!!--Alex brollo 01:42, 23 May 2006 (UTC)
This article may read slightly like an advertisement, but there is nothing to be gained by the makers of HTTrack by you using the software since it's GNU licensed.
It doesn't explain what "recursively" means or what a "chron job" is. So it's a pretty obscure advert.
The wording of the 'Bugs' Section is strange and looks like original research.
'It can be used for small websites only which can be downloaded all at once' - well, what does that mean? I've personally used the update feature successfully.
'HTTrack is unable to interpret the HTTP header "Content-Type" correctly'. Maybe true, but this this seems to be the design of the program, and hence not a bug.
Note from the author: httrack is definitely able to interpret the HTTP header "Content-Type" correctly ; you can test this by mirroring these two URls:
httrack "http://www.httrack.com/test_type2/get.htm?1234" "http://www.httrack.com/test_type/fake_image.gif"
Will correctly produce "fake_image.html" and "get81dc.gif"
And regarding the update issue, this was currently only reproduced because of a corrupted index, and/or server-side problems (dynamic pages not handling update requests correctly)
Lack of citable sources
Note from the author of HTTrack (xroche): I do not know whether it may be considered as citable source (and I won't discuss on the opportunity to keep the page - due to obvious conflicts of interest!), but several studies and/or pages are citing (sometimes very briefly - the main topic is not generally httrack itself) httrack:
- Tools and Techniques for Harvesting the World Wide Web - Jennifer L. Marill, Andrew Boyko, Michael Ashenfelder, and Laura Graham - Library of Congress
- Library Of Congress - NDSA: Web Archiving Survey Report - Digital Preservation
- File Data Dictionary for Preservation Metadata - Library of Congress
- Managing Electronic Records (SAA Digital Archives Specialist Program)
- Archivematica: Creating a Comprehensive Digital Preservation System
- Report on the Crawl and Harvest of the Whole Australian Web Domain Undertaken during June and July 2005 - National Library of Australia
- Evaluation of Open Source Spidering Technology - David Kellogg June 14, 2004
- Collecting and preserving the World Wide Web - A feasibility study undertaken for the JISC and Wellcome Trust - Michael Day - UKOLN, University of Bath
And brief citations from the following sources:
- Penetration Tester's Open Source Toolkit - Jeremy Faircloth (ISBN 978-1597496278)
- How to Start A Home-Based Web Design Business, 4th - Jim Smith (ISBN 978-0762759552)
- Placing the Suspect Behind the Keyboard: Using Digital Forensics and Investigative Techniques to Identify Cybercrime Suspects - Brett Shavers (ISBN 978-1597499859)
- Malware Forensic Field Guide for Windows Systems: Digital Forensics Field Guides - Cameron H. Malin, James M. Aquilina, Eoghan Casey (ISBN 978-1597494724)
Other "online-only" citations:
- Resource List - archives.org - National archives
- Personal Digital Archiving Day Kit - The Library of Congress
Again, these may not be considered as sources, depending on what is considered as a valid source. And the cited samples are neither exhaustive, nor necessarily perfectly representative. Xavier Roche (talk) 08:19, 7 March 2013 (UTC)