Browse Source

katana: add page (#12619)

* katana: add page

---------

Co-authored-by: Juri Dispan <juri.dispan@posteo.net>
Co-authored-by: K.B.Dharun Krishna <kbdharunkrishna@gmail.com>
Co-authored-by: Vítor Henrique <87824454+vitorhcl@users.noreply.github.com>
Fazle Arefin 11 months ago
parent
commit
33bb6e4404
1 changed files with 29 additions and 0 deletions
  1. 29 0
      pages/common/katana.md

+ 29 - 0
pages/common/katana.md

@@ -0,0 +1,29 @@
+# katana
+
+> A fast crawler focused on execution in automation pipelines offering both headless and non-headless crawling.
+> See also: `gau`, `scrapy`, `waymore`.
+> More information: <https://github.com/projectdiscovery/katana>.
+
+- Crawl a list of URLs:
+
+`katana -list {{https://example.com,https://google.com,...}}`
+
+- Crawl a [u]RL using headless mode using Chromium:
+
+`katana -u {{https://example.com}} -headless`
+
+- Use [p]a[s]sive sources (Wayback Machine, Common Crawl, and AlienVault) for URL discovery:
+
+`cat {{example.com}} | katana -passive`
+
+- Pass requests through a proxy (http/socks5) and use custom [H]eaders from a file:
+
+`katana -proxy {{http://127.0.0.1:8080}} -headers {{path/to/headers.txt}} -u {{https://example.com}}`
+
+- Specify the crawling [s]trategy, [d]epth of subdirectories to crawl, and rate limiting (requests per second):
+
+`katana -strategy {{depth-first|breadth-first}} -depth {{value}} -rate-limit {{value}} -u {{https://example.com}}`
+
+- Crawl a list of domains, each for a specific amount of seconds, and write results to an [o]utput file:
+
+`cat {{path/to/domains.txt}} | katana -crawl-duration {{value}} -output {{path/to/output.txt}}`