gospider
Fast web spider written in Go
Details
Category: Discovery
Publisher: trickest
Created Date: 6/23/2021
Container: quay.io/trickest/gospider:a4244c8
Source URL: https://github.com/jaeles-project/gospider
Parameters
raw
boolean
Command:
--raw
- Turn on rawproxy
string
Command:
-p
- Proxy (Ex: http://127.0.0.1:8080)length
boolean
Command:
--length
- Turn on lengthverbose
boolean
Command:
--verbose
- Turn on verbosedebug-mode
boolean
Command:
--debug
- Turn on debug modeuser-agent
string
Command:
-u
- User Agent to use (web: random web user-agent, mobi: random mobile user-agent)json-output
boolean
Command:
--json
- Enable JSON outputrandom-delay
string
Command:
-K
- RandomDelay is the extra randomized duration to wait added to Delay before creating a new request (second)3rd-party-url
boolean
Command:
-a
- Find URLs from 3rd party (Archive.org, CommonCrawl.org, VirusTotal.com, AlienVault.com)cookie-to-use
string
Command:
--cookie
- Cookie to use (testA=a; testB=b)filter-length
boolean
Command:
--filter-length
- Turn on length filterheader-to-use
string
Command:
--header
- Header to use (Use multiple flag to set multiple header)request-delay
string
Command:
-k
- Delay is the duration to wait before creating a new request to the matching domains (second)site-to-crawl
string
requiredCommand:
-s
- Site to crawlcrawl-robot-txt
boolean
Command:
--robots
- Try to crawl robots.txt (default true)request-timeout
string
Command:
-m
- Request timeout (second) (default 10)disable-redirect
boolean
Command:
--no-redirect
- Disable redirectwhitelist-domain
string
Command:
--whitelist-domain
- Whitelist Domaincrawl-sitemap-xml
boolean
Command:
--sitemap
- Try to crawl sitemap.xmlhtml-only-content
boolean
Command:
--base
- Disable all and only use HTML contentnumber-of-threads
string
Command:
-t
- Number of threads (Run sites in parallel) (default 1)enable-link-finder
boolean
Command:
--js
- Enable linkfinder in javascript file (default true)include-subdomains
boolean
Command:
--subs
- Include subdomainsblacklist-url-regex
string
Command:
--blacklist
- Blacklist URL Regexmax-recursion-depth
string
Command:
-d
- MaxDepth limits the recursion depth of visited URLs. (Set it to 0 for infinite recursion) (default 1)sites-list-to-crawl
file
requiredCommand:
-S
- Site list to crawlwhitelist-url-regex
string
Command:
--whitelist
- Whitelist URL Regexinclude-3rd-party-urls
boolean
Command:
-r
- Also include other-source's urls (still crawl and request)number-of-concurent-req
string
Command:
-c
- The number of the maximum allowed concurrent requests of the matching domains (default 5)include-3rd-party-subdomains
boolean
Command:
-w
- Include subdomains crawled from 3rd party. Default is main domainload-headers-and-strings-from-burp
string
Command:
--burp
- Load headers and cookie from burp raw http request