Project Description: DigiVigi is a 'DNIF Open Source' project which simply tries exhibiting a "How To?" process of analyzing real-time data inside DNIF from start to finish.
DNIF - Open Big Data Analytics Platform (Free Forever Version)
- Virtual Box
- JetBrains: PyCharm Community Edition
- Ubuntu 16.04 or above
- Docker
- Postman
- AlwaysUp (Trial Version)
The execution of project is carried out in two process-phase. Its just a procedural way based on the diagram from issue #1
PROCESS 1: Refer Issue #1
Stage 1:
- Select & understand data-set from a domain of interest.
Stage 2:
- Understand DNIF platform and its capabilities limited to project scope.
- Here's a link to DNIF's complete documentation: https://dnif.it/docs/
Stage 3:
- Get the static data-set inside DNIF platform by following the guidelines mentioned on the website.
- One can use a ready made dataset(json, csv or excel) or create one by scripting it, like writing a web scrapper code.
Stage 4:
- Perform analytics, create dashboards, build alerts(for which one will have to configure SMTP in their container)
- Here's how to configure SMTP: https://dnif.it/docs/guides/tutorials/configuring-smtp-in-docker.html
NOTE:Process 1 will ensure you get a good grasp of executing the project on a more fundametal level, before moving on to advanced level.
PROCESS 2: Refer Issue #1
Stage 1:
- Select & understand data-set from a domain of interest. Dataset has to be dynamic this time. Meaning it can update over a period of min(s), hour(s), day(s), week(s) ... The selection should be careful because in a later stage this might count when it comes down to your machine processing speed (CPU, RAM, Disk Space) - - - Check out the pre-requisites on https://dnif.it/docs/guides/getting-started/prerequisites.html
Stage 2:
- Understand DNIF platform and its capabilities limited to project scope.
- Here's a link to DNIF's complete documentation: https://dnif.it/docs/
Stage 3:
- Fetch the data continuously from the source selected in stage one and store/ update it in a file (ex; mydata.csv)
- For this one can write a code to do so or using an existing one from the repo.
- This file will get updated continuously depending on your scheduler/cron job frequency or any other way of your choice.
Stage 4:
- The data captured needs to be fed to DNIF for it to be analyzed and played with.
- Use DNIF API which does this job for you and use the existing unified code from the repo - under Process 2 folder; a file called "SourceToDnif.py".
Stage5:
- Perform analytics, create dashboards, build alerts(for which one will have to configure SMTP in their container)
- Here's how to configure SMTP: https://dnif.it/docs/guides/tutorials/configuring-smtp-in-docker.html
Process 1:
Process 2:
Additional Credits to: SOC18-Genesis
Webiron Feeds: It provides a comprehensive managed security service that will keep your web servers safe from harm. Webiron's intelligent technology is designed to immediately detect, block and prevent automated bot and malware attacks.
Key Metrics Abuse e-mail feed contains a log of our abuse reports and status of the issue reported. This feed is filterable by e-mail address, IP address, or ASN number. This is the master feed for the Twitter “bad abuse” feed and is pulled from live data.
Fields Descriptions:
Field | Description |
---|---|
Log Entry Type | Contains the action. This is either, report sent, report opened, report or if the host has replied with a resolved statement. |
Log Time | Time action was done. |
Attacker IP | The IP reported for issues (lookup link forwards to IP lookup page). The “IP” link filters the feed by the IP while the “lookup” provides more detailed information on the IP |
Logged E-Mails | These are either a list of e-mail addresses reported to for the attacker IP or the address that responded to a resolved or opened event. Clicking on an e-mail will filter the feed by that e-mail address. |
Log Message | The list of issues reported or an action message. |
Deliverable | Was the e-mail accepted by the host? |
Days Unresolved | The number of days the issue since the issue was reported to the host. |
Incidents Reported | The number of incidents reported. Some bots use thousands of nodes rather than heavier concentrations from fewer hosts. The damages are the same however. |
There's only one thing you'll need to do. Click on the "clone or download" button and get that ZIP file.
Here's a Video Link to get going with Installation Part https://youtu.be/ddpfh5sHMtA
THANKS FOR VISITING