1/30/2024 0 Comments A1 website download and httrack![]() J *parse Java Classes (j0 don't parse, bitmask: |1 parse default, |2 don't parse. U check document type if unknown (cgi,asp.) (u0 don't check, * u1 check but /, u2 check always) (-check-type) %T links conversion to UTF-8 (-utf8-conversion)īN accept cookies in cookies.txt (0=do not accept,* 1=accept) (-cookies) %p preserve html files 'as is' (identical to '-K4 -%F ""') (-preserve) X *purge old files after update (X0 keep delete) (-purge-old) O *generate output html file in case of error (404.) (o0 don't generate) (-generate-errors) %q *include query string for local files (useless, for information purpose only) (%q0 don't include) (-include-query-string) %x do not include any password for external password protected websites (%x0 include) (-disable-passwords) ![]() X replace external html links by error pages (-replace-external) ) (K0 *relative link, K absolute links, K4 original links, K3 absolute URI links, K5 transparent proxy link) (-keep-links) LN long names (L1 *long names / L0 8-3 conversion / L2 ISO9660 compatible) (-long-names) %M generate a RFC MIME-encapsulated full-archive (.mht) (-mime-html) %D cached delayed type check, don't wait for remote type during updates, to speedup them (%D0 wait, * %D1 don't wait) (-cached-delayed-type-check) %N delayed type check, don't make any link test but wait for files download to start instead (experimental) (%N0 don't use, %N1 use for unknown extensions, * %N2 always use) Or user defined structure (-N "%h%p/%n%q.%t") NN structure type (0 *original structure, 1+: see below) (-structure) %S add all scan rules located in this text file (one scan rule per line) (-urllist ) %L add all URL located in this text file (one URL per line) (-list ) ![]() T test all URLs (even forbidden ones) (-test) N get non-html files 'near' an html file (ex: an image located outside) (-near) %P *extended parsing, attempt to parse all links, even in unknown tags or Javascript (%P0 don't use) (-extended-parsing) HN host is abandoned if: 0=never, 1=timeout, 2=slow, 3=timeout or slow (-host-control) JN traffic jam control, minimum transfert rate (bytes/seconds) tolerated for a link (-min-rate) RN number of retries, in case of timeout or non-fatal errors (*R1) (-retries) TN timeout, number of seconds after a non-responding link is shutdown (-timeout) GN pause transfer if N bytes reached, and wait until lock file is deleted (-max-pause)ĬN number of multiple connections (*c8) (-sockets) %cN maximum number of connections/seconds (*%c10) (-connection-per-second) MN maximum overall size that can be uploaded/scanned (-max-size)ĮN maximum mirror time in seconds (60=1 minute, 3600=1 hour) (-max-time)ĪN maximum transfer rate in bytes/seconds (1000=1KB/s max) (-max-rate) MN,N2 maximum file length for non html (N) and html (N2) MN maximum file length for a non-html file (-max-files) %eN set the external links depth to N (* %e0) (-ext-depth) RN set the mirror depth to N (* r9999) (-depth) %b use this local hostname to make/send requests (-%b hostname) (-bind ) %f *use proxy for ftp (f0 don't use) (-httpproxy-ftp) P proxy use (-P proxy:port or -P user: :port) (-proxy ) Y mirror ALL links located in the first level pages (mirror links) (-mirrorlinks) I continue an interrupted mirror using the cache (-continue) G just get files (saved in the current directory) (-get-files) W mirror web sites, semi-automatic (asks questions) (-mirror-wizard) O path for mirror/logfiles+cache (-O path_mirror) (-path ) With options listed below: (* is the default value) Offline browser : copy websites to a local directory :~# httrack -h
0 Comments
Leave a Reply. |
AuthorWrite something about yourself. No need to be fancy, just an overview. ArchivesCategories |