Comments (3)
We already have variations of that, where we also play around with scheduling of SpecAugment.
E.g. see Switchboard base2.conv2l.specaug4a.
@papar22 and @ZhouW321 also have some more variations, which we will upload soon to the repo.
Note that the random_mask
in that config already runs multiple times (which is stochastically sampled). That are the options min_num
and max_num
. If you want that the mask is always runs exactly 3 times, just set min_num=3, max_num=3
.
Yes sure, you can play around with learning rate warm-up as well. My experience however is that increasing usually will not help.
Reducing the LR decay helps when you want to increase your overall training time, i.e. train more epochs. And training longer usually helps. When you look at this original SpecAugment paper, you will see that they effectively train much longer than we do.
from returnn-experiments.
Thanks for your answer, Albert!
I am sorry for a possibly naive question, but in the config example you mention above, the newbob_learning_rate_decay
is 0.7
.
My understanding is: LR_epoch_t+1
=decay
*LR_epoch_t
. So if I am starting from a baseline model trained for 12.5 epochs using newbob_learning_rate_decay
= 0.9
, and I want to train another model for say 25 epochs, I should increase newbob_learning_rate_decay
to say 0.95
instead of reducing it to 0.7
, right?
from returnn-experiments.
Yes sure.
from returnn-experiments.
Related Issues (20)
- local attention with unidirectional lstm not converging HOT 5
- Implement a unidirectional variant of local attention HOT 10
- Loading a saved Returnn model from its .meta file HOT 16
- query regarding LM data preprocessing HOT 2
- Reusing parameters inside rec layer HOT 5
- Training Configuration for TEDLIUMv2 HOT 3
- Question about 2020-rnn-transducer HOT 16
- 2018-asr-attention/librispeech/attention/exp3.ctc.lm.config: target 'bpe' unknown HOT 3
- Question about 2018-asr-librispeech dev = get_dataset("dev", subset=3000) HOT 2
- loss nan and cost nan while running my own corpus using librispeech sets HOT 10
- Hierarchical layer name not captured correctly
- Problem with retrieving source layer from a hierarchical definition
- Multi Stage Training
- Questions on librispeech transformer lm HOT 10
- Transducer error in GetFilteredScoreOp HOT 4
- Big files in repo HOT 5
- Git commit/push rule to not allow big files HOT 3
- Could you please provide a script that could run lsh-attention for translation? HOT 4
- Assert Error when running 2022-lsh-attention HOT 7
Recommend Projects
-
React
A declarative, efficient, and flexible JavaScript library for building user interfaces.
-
Vue.js
🖖 Vue.js is a progressive, incrementally-adoptable JavaScript framework for building UI on the web.
-
Typescript
TypeScript is a superset of JavaScript that compiles to clean JavaScript output.
-
TensorFlow
An Open Source Machine Learning Framework for Everyone
-
Django
The Web framework for perfectionists with deadlines.
-
Laravel
A PHP framework for web artisans
-
D3
Bring data to life with SVG, Canvas and HTML. 📊📈🎉
-
Recommend Topics
-
javascript
JavaScript (JS) is a lightweight interpreted programming language with first-class functions.
-
web
Some thing interesting about web. New door for the world.
-
server
A server is a program made to process requests and deliver data to clients.
-
Machine learning
Machine learning is a way of modeling and interpreting data that allows a piece of software to respond intelligently.
-
Visualization
Some thing interesting about visualization, use data art
-
Game
Some thing interesting about game, make everyone happy.
Recommend Org
-
Facebook
We are working to build community through open source technology. NB: members must have two-factor auth.
-
Microsoft
Open source projects and samples from Microsoft.
-
Google
Google ❤️ Open Source for everyone.
-
Alibaba
Alibaba Open Source for everyone
-
D3
Data-Driven Documents codes.
-
Tencent
China tencent open source team.
from returnn-experiments.