A distributed crawler built for my master thesis. Goal is to collect data and pre-process it from e.g. GitHub.
Node.js >= v10
Grakn.ai - 1.8.4
Docker (optional)
Docker-Compose (optional)
npm i
GITHUBACCOUNTS='[{"username": "name", "token": "00000"}]' TYPE=master node app.js
node app.js
docker-compose up --build
Please look at config.js for a better overview and understanding.
- PORT - (def) '8000'
- TYPE - (def) 'client' || 'master'
- GRAKNURI - (def) 'localhost:48555'
- SENTRY - (def) ''
- MASTER_SOCKET - (def) 'http://localhost:8000' (Socket of the Master)
- GITHUBACCOUNTS - (def) [] - example '[{"username": "name", "token": "00000"},...]'