katana
A next-generation crawling and spidering framework.
Name:katana
Category:Discovery
Publisher:trickest-mhmdiaa
Created:11/7/2022
Container:
quay.io/trickest/katana:v1.1.1-patch-2
Output Type:
License:Unknown
Source:View Source
Parameters
-u
target url-list
target url list-debug
display debug output-delay
request delay between each request in seconds-depth
maximum depth to crawl (default 3)-field
field to display in output (url,path,fqdn,rdn,rurl,qurl,qpath,file,ufile,key,value,kv,dir,udir)-jsonl
write output in JSONL(ines) format-proxy
http/socks5 proxy to use-retry
number of times to retry the request (default 1)-config
path to the katana configuration file-resume
resume scan using resume.cfg-silent
display output only-exclude
exclude host matching specified filter ('cdn', 'private-ips', cidr, ip, regex)-headers
custom header/cookie to include in request-jsluice
enable jsluice parsing in javascript file (memory intensive)-timeout
time to wait for request in seconds (default 10)-verbose
display verbose output-headless
enable headless hybrid crawling (experimental)-js-crawl
enable endpoint parsing / crawling in javascript file-no-scope
disables host based default scope-omit-raw
omit raw requests/responses from jsonl output-strategy
Visit strategy (depth-first, breadth-first) (default depth-first)-omit-body
omit response body from jsonl output-resolvers
list of custom resolver (file or comma separated)-no-sandbox
start headless chrome in --no-sandbox mode-rate-limit
maximum requests to send per second (default 150)-concurrency
number of concurrent fetchers to use (default 10)-crawl-scope
in scope url regex to be followed by crawler-field-scope
pre-defined scope field (dn,rdn,fqdn) or custom regex (e.g., '(company-staging.io|company.com)') (default rdn)-form-config
path to custom form configuration file-known-files
enable crawling of known files (all,robotstxt,sitemapxml)-match-regex
regex to match on output url-parallelism
number of concurrent inputs to process (default 10)-store-field
field to store in per-host output (url,path,fqdn,rdn,rurl,qurl,qpath,file,ufile,key,value,kv,dir,udir)-field-config
path to custom field configuration file-filter-regex
regex to filter on output url-headers
custom header/cookie to include in request-health-check
run diagnostic check up-no-incognito
start headless chrome without incognito mode-crawl-duration
maximum duration to crawl the target for (s, m, h, d) (default s)-xhr-extraction
extract xhr request url,method in jsonl output-chrome-data-dir
path to store chrome browser data-crawl-out-scope
out of scope url regex to be excluded by crawler-extension-match
match output for given extension (eg, -em php,html,js)-form-extraction
extract form, input, textarea & select elements in jsonl output-match-condition
match response with dsl based condition-tls-impersonate
enable experimental client hello (ja3) tls randomization-crawl-scope
in scope url regex to be followed by crawler-extension-filter
filter output for given extension (eg, -ef png,css)-filter-condition
filter response with dsl based condition-headless-options
start headless chrome with additional options-match-regex
list of regex to match on output url-disable-redirects
disable following redirects (default false)-display-out-scope
display external endpoint from scoped crawling-filter-regex
list of regex to filter on output url-max-response-size
maximum response size to read (default 9223372036854775807)-rate-limit-minute
maximum number of requests to send per minute-automatic-form-fill
enable automatic form filling (experimental)-ignore-query-params
Ignore crawling same path with different query-param values-crawl-out-scope
out of scope url regex to be excluded by crawler