98bd85e0c4 | 4 years ago | |
---|---|---|
.github/workflows | 4 years ago | |
build | 4 years ago | |
cmd | 5 years ago | |
deployments | 4 years ago | |
internal | 4 years ago | |
pkg/proto | 4 years ago | |
scripts | 4 years ago | |
.gitignore | 5 years ago | |
LICENSE | 5 years ago | |
README.md | 4 years ago | |
go.mod | 5 years ago | |
go.sum | 5 years ago |
README.md
Trandoshan dark web crawler
This repository is a complete rewrite of the Trandoshan dark web crawler. Everything has been written inside a single Git repository to ease maintenance.
Why a rewrite?
The first version of Trandoshan (available here) is working great but not really professional, the code start to be a mess, hard to manage since split in multiple repositories, etc..
I have therefore decided to create & maintain the project in this specific directory, where all process code will be available (as a Go module).
How build the crawler
Since the docker image are not available yet, one must run the following script in order to build the crawler fully.
./scripts/build.sh
How to start the crawler
Execute the /scripts/start.sh
and wait for all containers to start.
Note
Ensure you have at least 3GB of memory as the Elasticsearch stack docker will require 2GB.
How to start the crawling process
Since the API is explosed on localhost:15005, one can use it to start the crawling process:
feeder --api-uri http://localhost:15005 --url https://www.facebookcorewwwi.onion
this will 'force' the API to publish given URL in crawling queue.
How to access the Kibana UI
Now head out to http://localhost:15004
You will need to create an index pattern named 'resources', and when it asks for the time field, choose 'time'.