Giter Site home page Giter Site logo

Comments (10)

albertz avatar albertz commented on May 25, 2024

The vocabulary created by the apply_bpe script is the same as for the pretrained model.

from returnn-experiments.

Slyne avatar Slyne commented on May 25, 2024

from returnn-experiments.

albertz avatar albertz commented on May 25, 2024

I'm but sure what you mean by librispeech-lm-norm.txt.
You just use the vocabulary (vocab_file) which you created via full-setup-attention.
You don't need anything else.
You can use the configs here to do recognition with your attention model together with a LM (e.g. the pretrained model).

from returnn-experiments.

Slyne avatar Slyne commented on May 25, 2024

from returnn-experiments.

albertz avatar albertz commented on May 25, 2024

I don't quite understand. If you use a pretrained model, you don't need to train it. The config you are referring to is for training (which you don't need, because it is already trained). For recognition, you need a different config, which I linked already.
The vocabulary which you pasted is the correct one.

from returnn-experiments.

Slyne avatar Slyne commented on May 25, 2024

from returnn-experiments.

Slyne avatar Slyne commented on May 25, 2024

And the config file seems to be a training configuration. You can notice that the task="train" in the configuration file.
image

from returnn-experiments.

albertz avatar albertz commented on May 25, 2024

Ah, I thought that you want to use the pretrained model.
If you want to train it yourself, then yes, you need that config (btw, yes, there is task='train', but still in many cases, the config file is used for both training + inference; any option there can be overwritten via command line).
I just checked and you are right, the vocab used for the LmDataset has a different format. But I think it should be straight forward to convert from one to the other. Or to add support for the other format. Actually I wonder why we have that at all. I uploaded that vocab file here.
The train files (data_files in config) are generated somehow from the LibriSpeech LM training data. We might have done some post processing / normalization. @kazuki-irie can give you more details on that. Actually, maybe he can just upload his normalization scripts also here.

from returnn-experiments.

Slyne avatar Slyne commented on May 25, 2024

from returnn-experiments.

albertz avatar albertz commented on May 25, 2024

No problem. I extended the Readme a bit (here).
Maybe @kazuki-irie can later add some of the post processing scripts.
Closing this now.

from returnn-experiments.

Related Issues (20)

Recommend Projects

  • React photo React

    A declarative, efficient, and flexible JavaScript library for building user interfaces.

  • Vue.js photo Vue.js

    🖖 Vue.js is a progressive, incrementally-adoptable JavaScript framework for building UI on the web.

  • Typescript photo Typescript

    TypeScript is a superset of JavaScript that compiles to clean JavaScript output.

  • TensorFlow photo TensorFlow

    An Open Source Machine Learning Framework for Everyone

  • Django photo Django

    The Web framework for perfectionists with deadlines.

  • D3 photo D3

    Bring data to life with SVG, Canvas and HTML. 📊📈🎉

Recommend Topics

  • javascript

    JavaScript (JS) is a lightweight interpreted programming language with first-class functions.

  • web

    Some thing interesting about web. New door for the world.

  • server

    A server is a program made to process requests and deliver data to clients.

  • Machine learning

    Machine learning is a way of modeling and interpreting data that allows a piece of software to respond intelligently.

  • Game

    Some thing interesting about game, make everyone happy.

Recommend Org

  • Facebook photo Facebook

    We are working to build community through open source technology. NB: members must have two-factor auth.

  • Microsoft photo Microsoft

    Open source projects and samples from Microsoft.

  • Google photo Google

    Google ❤️ Open Source for everyone.

  • D3 photo D3

    Data-Driven Documents codes.