Giter Site home page Giter Site logo

ant-louis / netbert-search Goto Github PK

View Code? Open in Web Editor NEW
3.0 2.0 0.0 1.5 MB

⌨️ A neural search engine for RFC documents that uses NetBERT.

License: MIT License

Makefile 0.80% Dockerfile 1.59% Shell 4.38% Python 83.31% HTML 9.92%
nlp search bert domain-specific-language computer-network information-retrieval

netbert-search's Introduction

License: MIT

NetBERT Search Engine

This repository provides the implementation of a neural search engine for RFC documents that use a pre-trained NetBERT model for information retrieval.

Table of contents

  1. System Architecture
  2. Setup
    1. Requirements
    2. Download a pre-trained NetBERT model
    3. Launch the Docker containers
  3. Index creation
    1. Download RFC data
    2. Clean and process data
    3. Convert data in proper format
    4. Create index
    5. Create documents
    6. Index documents
  4. Let's search!

1. System architecture

System architecture

2. Setup

Requirements

The following section lists the requirements in order to start running the project.

This project is based on Docker containers, so ensure to have Docker and docker-compose installed on your machine. In addition, your machine should dispose from a working version of Python 3.6 as well as the following packages:

These libraries can be installed automatically by running the following command in the code/ repository:

pip install -r requirements.txt

Download a pre-trained NetBERT model

You can download the pre-trained NetBERT model with 🤗 transformers as follows:

from transformers import AutoTokenizer, AutoModel

# Load model.
tokenizer = AutoTokenizer.from_pretrained("antoiloui/netbert")
model = AutoModel.from_pretrained("antoiloui/netbert")

# Save it.
model.save_pretrained("path/to/local/folder")
tokenizer.save_pretrained("path/to/local/folder")

Then, you need to convert the 🤗 model checkpoint to the checkpoint format from original BERT repository:

bash convert_model_checkpoint.sh $path/to/local/folder

This will create a new 'tensorflow' repository where you should change the file names that way in order for the docker containers to run properly:

  • netbert_config.json -> bert_config.json
  • netbert.ckpt.data-00000-of-00001 -> bert_model.ckpt.data-00000-of-00001
  • netbert.ckpt.index -> bert_model.ckpt.index
  • netbert.ckpt.meta -> bert_model.ckpt.meta

Finally, set the following environnement variable with the path of the folder containing the convert NetBERT checkpoint:

export PATH_MODEL=$path/to/local/folder/tensorflow

Launch the Docker containers

In order to run the containers, run the following command:

make install

3. Index creation

Go into the 'index_creation' repository:

cd index_creation/

Download RFC data

bash download_data.sh $OUT_DIR

Clean and process data

bash clean_data.sh $DATA_DIR

Convert data in proper format

bash convert_data_format.sh $DATA_DIR

Create index

You can use the create index API to add a new index to an Elasticsearch cluster. When creating an index, you can specify the following:

  • Settings for the index
  • Mappings for fields in the index
  • Index aliases

For example, if you want to create rfcsearch index with title, text and text_vector fields, you can create the index by the following command:

$ bash create_index.sh

# index.json
{
  "settings": {
    "number_of_shards": 2,
    "number_of_replicas": 1
  },
  "mappings": {
    "dynamic": "true",
    "_source": {
      "enabled": "true"
    },
    "properties": {
      "title": {
        "type": "text"
      },
      "text": {
        "type": "text"
      },
      "text_vector": {
        "type": "dense_vector",
        "dims": 768
      }
    }
  }
}

NB: The dims value of text_vector must need to match the dims of a pretrained BERT model.

Create documents

Once you created an index, you’re ready to index some document. The point here is to convert your document into a vector using BERT. The resulting vector is stored in the text_vector field. Let`s convert your data into a JSON document:

bash create_documents.sh $DATA_DIR $DATA_FILE

# $DATA_FILE=data/example.csv
"Title","Text"
"rfc1 - Host Software","Somewhat independently, Gerard DeLoche of UCLA has been working on the HOST-IMP interface."
"rfc153 - SRI ARC-NIC status","The specifications of DEL are under discussion. The following diagrams show the sequence of actions."
"rfc354 - File Transfer Protocol","The links have the following primitive characteristics. They are always functioning and there are always 32 of them."
"rfcxxx - Lorem Ipsum","Lorem Ipsum"
...

After finishing the script, you get a JSON document as follows:

# documents.json
{"_op_type": "index", "_index": "rfcsearch", "text": "lorem ipsum", "title": "lorem ipsum", "text_vector": [...]}
{"_op_type": "index", "_index": "rfcsearch", "text": "lorem ipsum", "title": "lorem ipsum", "text_vector": [...]}
{"_op_type": "index", "_index": "rfcsearch", "text": "lorem ipsum", "title": "lorem ipsum", "text_vector": [...]}
...

Index documents

After converting your data into a JSON, you can adds a JSON document to the specified index and makes it searchable:

bash index_documents.sh $DATA_DIR

4. Let's search!

Open your browser and go to http://127.0.0.1:5000.

Example


  • Credits: This project was inspired by Hironsan.

Recommend Projects

  • React photo React

    A declarative, efficient, and flexible JavaScript library for building user interfaces.

  • Vue.js photo Vue.js

    🖖 Vue.js is a progressive, incrementally-adoptable JavaScript framework for building UI on the web.

  • Typescript photo Typescript

    TypeScript is a superset of JavaScript that compiles to clean JavaScript output.

  • TensorFlow photo TensorFlow

    An Open Source Machine Learning Framework for Everyone

  • Django photo Django

    The Web framework for perfectionists with deadlines.

  • D3 photo D3

    Bring data to life with SVG, Canvas and HTML. 📊📈🎉

Recommend Topics

  • javascript

    JavaScript (JS) is a lightweight interpreted programming language with first-class functions.

  • web

    Some thing interesting about web. New door for the world.

  • server

    A server is a program made to process requests and deliver data to clients.

  • Machine learning

    Machine learning is a way of modeling and interpreting data that allows a piece of software to respond intelligently.

  • Game

    Some thing interesting about game, make everyone happy.

Recommend Org

  • Facebook photo Facebook

    We are working to build community through open source technology. NB: members must have two-factor auth.

  • Microsoft photo Microsoft

    Open source projects and samples from Microsoft.

  • Google photo Google

    Google ❤️ Open Source for everyone.

  • D3 photo D3

    Data-Driven Documents codes.