Take a list of domains, crawl urls and scan for endpoints, secrets, api keys, file extensions, tokens and more...
Coded with 💙 by edoardottt.
Share on Twitter!
Preview • Install • Get Started • Examples • Contributing • License
You need Go.
-
Linux
git clone https://github.com/edoardottt/cariddi.gitcd cariddigo getmake linux(to install)make unlinux(to uninstall)
Or in one line:
git clone https://github.com/edoardottt/cariddi.git; cd cariddi; go get; make linux -
Windows (executable works only in cariddi folder.)
git clone https://github.com/edoardottt/cariddi.gitcd cariddigo get.\make.bat windows(to install).\make.bat unwindows(to uninstall)
cariddi -h prints the help in the command line.
Usage of cariddi:
-c int
Concurrency level. (default 20)
-cache
Use the .cariddi_cache folder as cache.
-d int
Delay between a page crawled and another.
-e Hunt for juicy endpoints.
-ef string
Use an external file (txt, one per line) to use custom parameters for endpoints hunting.
-examples
Print the examples.
-ext int
Hunt for juicy file extensions. Integer from 1(juicy) to 7(not juicy).
-h Print the help.
-i string
Ignore the URL containing at least one of the elements of this array.
-it string
Ignore the URL containing at least one of the lines of this file.
-oh string
Write the output into an HTML file.
-ot string
Write the output into a TXT file.
-plain
Print only the results.
-s Hunt for secrets.
-sf string
Use an external file (txt, one per line) to use custom regexes for secrets hunting.
-t int
Set timeout for the requests. (default 10)
-version
Print the version.
-
cariddi -version(Print the version) -
cariddi -h(Print the help) -
cariddi -examples(Print the examples) -
cat urls | cariddi -s(Hunt for secrets) -
cat urls | cariddi -d 2(2 seconds between a page crawled and another) -
cat urls | cariddi -c 200(Set the concurrency level to 200) -
cat urls | cariddi -e(Hunt for juicy endpoints) -
cat urls | cariddi -plain(Print only useful things) -
cat urls | cariddi -ot target_name(Results in txt file) -
cat urls | cariddi -oh target_name(Results in html file) -
cat urls | cariddi -ext 2(Hunt for juicy (level 2 of 7) files) -
cat urls | cariddi -e -ef endpoints_file(Hunt for custom endpoints) -
cat urls | cariddi -s -sf secrets_file(Hunt for custom secrets) -
cat urls | cariddi -i forum,blog,community,open(Ignore urls containing these words) -
cat urls | cariddi -it ignore_file(Ignore urls containing at least one line in the input file) -
cat urls | cariddi -cache(Use the .cariddi_cache folder as cache) -
cat urls | cariddi -t 5(Set the timeout for the requests) -
For Windows use
powershell.exe -Command "cat urls | .\cariddi.exe"
Just open an issue/pull request. See also CONTRIBUTING.md and CODE OF CONDUCT.md
Help me building this!
A special thanks to:
To do:
-
Tests (😂)
-
Tor support
-
Proxy support
-
Cookie support
-
Ignore specific types of urls
-
Plain output (print only results)
-
HTML output
-
Output color
-
Endpoints (parameters) scan
-
Secrets scan
-
Extensions scan
-
TXT output
This repository is under GNU General Public License v3.0.
edoardoottavianelli.it to contact me.