Comments (6)
No, you are right.
The actual code just saves the checkpoint when the model finishes the training: https://github.com/ml5js/training-lstm/blob/master/train.py#L168
We should add support to save the model by default every N amount of iterations.
Here's the original code doing that: https://github.com/sherjilozair/char-rnn-tensorflow/blob/master/train.py#L133
We removed it at some point and now it makes sense to put it back
from training-charrnn.
Yes, it is possible. Here are the arguments you need to pass to the script:
parser.add_argument('--init_from', type=str, default=None,
help="""continue training from saved model at this path. Path must contain files saved by previous training process:
'config.pkl' : configuration;
'chars_vocab.pkl' : vocabulary definitions;
'checkpoint' : paths to model file(s) (created by tf).
Note: this file contains absolute paths, be careful when moving files around;
'model.ckpt-*' : file(s) with model definition (created by tf)
""")
See https://github.com/ml5js/training-lstm/blob/master/train.py#L67
from training-charrnn.
Oh cool, thanks! The file model.ckpt isn't created, so I'm a little worried about whether it actually picks up from where it left off. It does say that it's loading the preprocessed files, but then the time taken to train the model is no different from doing it all over again.
I assume this is because this process is usually meant for re-training models rather than shifting from computer to computer like what I'm wanting to do?
from training-charrnn.
Good spot!
There's no easy way to put it back in right? My knowledge of python is limited, but we would have to rewrite that entire last bit to make it work, right?
from training-charrnn.
Has this ability been implemented (or re-implemented)? It would be very useful, in the spirit of making machine learning accessible, to have the ability to start, stop, and resume training, given the enormous amounts of time training can take. For example, I was trying this just now on an admittedly pretty large input corpus (around 300MB), and if I understand how it's working, it's going to take several weeks of computer time to finish on the fasted computer I could find today. Obviously, finding a good GPU or maybe renting some cloud computing time would speed that up, but ¯_(ツ)_/¯
from training-charrnn.
Hello so I tried to figure something out with my rudimentary python, but I can't seem to get anywhere. Some error with the args
somewhere.
Any luck with this?
from training-charrnn.
Related Issues (13)
- README clarification on hyperparameters HOT 2
- tf-core.esm.js:17 Uncaught TypeError: Y is not a function
- Trining on big files 25+ MB gets killed HOT 1
- TensorFlow 2.0 compatibility? HOT 3
- Training "Cache"?
- does training-charRNN use GPU or CPU
- typo in about HOT 1
- How to make it learn to rhyme?
- Training on word level HOT 2
- UnicodeDecodeError HOT 1
- Add training resources / tutorials HOT 5
- Uncaught TypeError: ml5.LSTMGenerator is not a function HOT 4
Recommend Projects
-
React
A declarative, efficient, and flexible JavaScript library for building user interfaces.
-
Vue.js
🖖 Vue.js is a progressive, incrementally-adoptable JavaScript framework for building UI on the web.
-
Typescript
TypeScript is a superset of JavaScript that compiles to clean JavaScript output.
-
TensorFlow
An Open Source Machine Learning Framework for Everyone
-
Django
The Web framework for perfectionists with deadlines.
-
Laravel
A PHP framework for web artisans
-
D3
Bring data to life with SVG, Canvas and HTML. 📊📈🎉
-
Recommend Topics
-
javascript
JavaScript (JS) is a lightweight interpreted programming language with first-class functions.
-
web
Some thing interesting about web. New door for the world.
-
server
A server is a program made to process requests and deliver data to clients.
-
Machine learning
Machine learning is a way of modeling and interpreting data that allows a piece of software to respond intelligently.
-
Visualization
Some thing interesting about visualization, use data art
-
Game
Some thing interesting about game, make everyone happy.
Recommend Org
-
Facebook
We are working to build community through open source technology. NB: members must have two-factor auth.
-
Microsoft
Open source projects and samples from Microsoft.
-
Google
Google ❤️ Open Source for everyone.
-
Alibaba
Alibaba Open Source for everyone
-
D3
Data-Driven Documents codes.
-
Tencent
China tencent open source team.
from training-charrnn.