Comments (12)
How many train data do u use for crnn?
from 2019-ccf-bdci-ocr-mczj-ocr-identificationidelement.
How many train data do u use for crnn?
it generates a new dataset(25000 data) randomly each 3000 epoch
from 2019-ccf-bdci-ocr-mczj-ocr-identificationidelement.
Hello, in this project, we have 6032 characters, and i generate 6 million train data,with batch_size=128,each epoch cost about 3 hours(I trained about 15 epoch,then stopped,so the total train time is about 2 days....2 weeks seems too long...), use an TITAN Xp.
I guess the train time depends on the number of characters, but the train time is too long...Maybe you should check your code. You can use command nvidia-smi to check GPU-Util.
By the way,is it convenient to reveal what your task is? I'm curious why there are so many characters to recognize...
from 2019-ccf-bdci-ocr-mczj-ocr-identificationidelement.
Hello, in this project, we have 6032 characters, and i generate 6 million train data,with batch_size=128,each epoch cost about 3 hours(I trained about 15 epoch,then stopped,so the total train time is about 2 days....2 weeks seems too long...), use an TITAN Xp.
I guess the train time depends on the number of characters, but the train time is too long...Maybe you should check your code. You can use command nvidia-smi to check GPU-Util.
By the way,is it convenient to reveal what your task is? I'm curious why there are so many characters to recognize...
I saw your config set: __C.TRAIN.EPOCHS = 580000,so this is your iterations numbers or epoch numbers?
AND i am working on passport OCR.
from 2019-ccf-bdci-ocr-mczj-ocr-identificationidelement.
Hello, in this project, we have 6032 characters, and i generate 6 million train data,with batch_size=128,each epoch cost about 3 hours(I trained about 15 epoch,then stopped,so the total train time is about 2 days....2 weeks seems too long...), use an TITAN Xp.
I guess the train time depends on the number of characters, but the train time is too long...Maybe you should check your code. You can use command nvidia-smi to check GPU-Util.
By the way,is it convenient to reveal what your task is? I'm curious why there are so many characters to recognize...
In what i m training ,there are 5032 characters, 13million train data,batch_size=128,each epoch cost 107 hours, really 107 hours. use an 2080ti. Our task seems pretty simmilar, and i wonder why mine is so slow, Memory-Usage | GPU-Util=10956MiB / 10989MiB | 56%. GPU-Util seems lower ,i am not using a tfrecord for last train version with 8 million train data, the speed using tfrecord didn't speed up much.
is there any clue? appreciate lot...
from 2019-ccf-bdci-ocr-mczj-ocr-identificationidelement.
How many train data do u use for crnn?
it generates a new dataset(25000 data) randomly each 3000 epoch
__C.TRAIN.EPOCHS = 580000, seems to be iterations number .
25000train data? it 's to small to accomplish a recognition task with 20838 characters. I wonder you are not using a tfrecord like me?
as I know, the number classification as 20838 is too big ,i wonder whether the model could do it(convergent).
from 2019-ccf-bdci-ocr-mczj-ocr-identificationidelement.
How many train data do u use for crnn?
it generates a new dataset(25000 data) randomly each 3000 epoch
__C.TRAIN.EPOCHS = 580000, seems to be iterations number .
25000train data? it 's to small to accomplish a recognition task with 20838 characters. I wonder you are not using a tfrecord like me?
as I know, the number classification as 20838 is too big ,i wonder whether the model could do it(convergent).
no, the number of total training data is far more than 25000,i just change another dataset every 3000 epoch(i think it should be iteration, but epoch in author's source code), and change another dataset means generating another 25000 data randomly.
from 2019-ccf-bdci-ocr-mczj-ocr-identificationidelement.
How many train data do u use for crnn?
it generates a new dataset(25000 data) randomly each 3000 epoch
__C.TRAIN.EPOCHS = 580000, seems to be iterations number .
25000train data? it 's to small to accomplish a recognition task with 20838 characters. I wonder you are not using a tfrecord like me?
as I know, the number classification as 20838 is too big ,i wonder whether the model could do it(convergent).no, the number of total training data is far more than 25000,i just change another dataset every 3000 epoch(i think it should be iteration, but epoch in author's source code), and change another dataset means generating another 25000 data randomly.
i see. then you train data should be 75 million. a new way of training to me ,thank you for explain.
then our training speed seems differ not much^_^
from 2019-ccf-bdci-ocr-mczj-ocr-identificationidelement.
How many train data do u use for crnn?
it generates a new dataset(25000 data) randomly each 3000 epoch
__C.TRAIN.EPOCHS = 580000, seems to be iterations number .
25000train data? it 's to small to accomplish a recognition task with 20838 characters. I wonder you are not using a tfrecord like me?
as I know, the number classification as 20838 is too big ,i wonder whether the model could do it(convergent).
no, the number of total training data is far more than 25000,i just change another dataset every 3000 epoch(i think it should be iteration, but epoch in author's source code), and change another dataset means generating another 25000 data randomly.
How many train data do u use for crnn?
it generates a new dataset(25000 data) randomly each 3000 epoch
__C.TRAIN.EPOCHS = 580000, seems to be iterations number .
25000train data? it 's to small to accomplish a recognition task with 20838 characters. I wonder you are not using a tfrecord like me?
as I know, the number classification as 20838 is too big ,i wonder whether the model could do it(convergent).
How many train data do u use for crnn?
it generates a new dataset(25000 data) randomly each 3000 epoch
__C.TRAIN.EPOCHS = 580000, seems to be iterations number .
25000train data? it 's to small to accomplish a recognition task with 20838 characters. I wonder you are not using a tfrecord like me?
as I know, the number classification as 20838 is too big ,i wonder whether the model could do it(convergent).
no, the number of total training data is far more than 25000,i just change another dataset every 3000 epoch(i think it should be iteration, but epoch in author's source code), and change another dataset means generating another 25000 data randomly.i see. then you train data should be 75 million. a new way of training to me ,thank you for explain.
then our training speed seems differ not much^_^
may i ask you how high your val set accuracy can reach?
from 2019-ccf-bdci-ocr-mczj-ocr-identificationidelement.
How many train data do u use for crnn?
it generates a new dataset(25000 data) randomly each 3000 epoch
__C.TRAIN.EPOCHS = 580000, seems to be iterations number .
25000train data? it 's to small to accomplish a recognition task with 20838 characters. I wonder you are not using a tfrecord like me?
as I know, the number classification as 20838 is too big ,i wonder whether the model could do it(convergent).no, the number of total training data is far more than 25000,i just change another dataset every 3000 epoch(i think it should be iteration, but epoch in author's source code), and change another dataset means generating another 25000 data randomly.
How many train data do u use for crnn?
it generates a new dataset(25000 data) randomly each 3000 epoch
__C.TRAIN.EPOCHS = 580000, seems to be iterations number .
25000train data? it 's to small to accomplish a recognition task with 20838 characters. I wonder you are not using a tfrecord like me?
as I know, the number classification as 20838 is too big ,i wonder whether the model could do it(convergent).How many train data do u use for crnn?
it generates a new dataset(25000 data) randomly each 3000 epoch
__C.TRAIN.EPOCHS = 580000, seems to be iterations number .
25000train data? it 's to small to accomplish a recognition task with 20838 characters. I wonder you are not using a tfrecord like me?
as I know, the number classification as 20838 is too big ,i wonder whether the model could do it(convergent).
no, the number of total training data is far more than 25000,i just change another dataset every 3000 epoch(i think it should be iteration, but epoch in author's source code), and change another dataset means generating another 25000 data randomly.i see. then you train data should be 75 million. a new way of training to me ,thank you for explain.
then our training speed seems differ not much^_^may i ask you how high your val set accuracy can reach?
i use LER(label error rate), last version my loss reach [0.9,1.8], LER is 1%. i think this is not acurrate to Measure my model. cause there is no real scence data in my val set.
from 2019-ccf-bdci-ocr-mczj-ocr-identificationidelement.
label error rate
My method is really hard to converge....
from 2019-ccf-bdci-ocr-mczj-ocr-identificationidelement.
Hello, in this project, we have 6032 characters, and i generate 6 million train data,with batch_size=128,each epoch cost about 3 hours(I trained about 15 epoch,then stopped,so the total train time is about 2 days....2 weeks seems too long...), use an TITAN Xp.
I guess the train time depends on the number of characters, but the train time is too long...Maybe you should check your code. You can use command nvidia-smi to check GPU-Util.
By the way,is it convenient to reveal what your task is? I'm curious why there are so many characters to recognize...I saw your config set: __C.TRAIN.EPOCHS = 580000,so this is your iterations numbers or epoch numbers?
AND i am working on passport OCR.
hello, 580000 is iterations; My guess is that data processing takes up too much time,gpu is waiting for data.
from 2019-ccf-bdci-ocr-mczj-ocr-identificationidelement.
Related Issues (20)
- test error HOT 1
- test model HOT 7
- there are something wrong in test_data_dst_path, exit... HOT 3
- CGAN data HOT 1
- char_map.json文件生成方式 HOT 1
- Some question on Pix2Pix HOT 2
- 百度网盘的分享无效了 HOT 1
- 百度网盘分享失效,跪求留个联系方式,感激不尽,QQ 83656313 HOT 2
- 百度云盘分享失效了,跪求,跪求 ,感谢恩师,QQ 83656313 HOT 8
- 关于tensorflow-gpu版本 HOT 4
- repitle_address_extract.json HOT 2
- 百度网盘的链接失效了,还有大佬手上有资源吗,很关键!!有报酬!!!!QQ767618141
- recognise model HOT 1
- 关于模型和数据下载链接失效了 HOT 2
- 已加网盘好友,十分谢谢分享 HOT 6
- 为什么项目clone不下来 HOT 1
- 请问参考文献CGAN和参考代码CycleGAN不一样是为什么?
- 这个可以适用单张身份证图片吗,我试了下会报错 HOT 1
- bug
- 您好,数据集的标签文件csv中的身份证号被科学计数给约了,怎么处理呢? HOT 1
Recommend Projects
-
React
A declarative, efficient, and flexible JavaScript library for building user interfaces.
-
Vue.js
🖖 Vue.js is a progressive, incrementally-adoptable JavaScript framework for building UI on the web.
-
Typescript
TypeScript is a superset of JavaScript that compiles to clean JavaScript output.
-
TensorFlow
An Open Source Machine Learning Framework for Everyone
-
Django
The Web framework for perfectionists with deadlines.
-
Laravel
A PHP framework for web artisans
-
D3
Bring data to life with SVG, Canvas and HTML. 📊📈🎉
-
Recommend Topics
-
javascript
JavaScript (JS) is a lightweight interpreted programming language with first-class functions.
-
web
Some thing interesting about web. New door for the world.
-
server
A server is a program made to process requests and deliver data to clients.
-
Machine learning
Machine learning is a way of modeling and interpreting data that allows a piece of software to respond intelligently.
-
Visualization
Some thing interesting about visualization, use data art
-
Game
Some thing interesting about game, make everyone happy.
Recommend Org
-
Facebook
We are working to build community through open source technology. NB: members must have two-factor auth.
-
Microsoft
Open source projects and samples from Microsoft.
-
Google
Google ❤️ Open Source for everyone.
-
Alibaba
Alibaba Open Source for everyone
-
D3
Data-Driven Documents codes.
-
Tencent
China tencent open source team.
from 2019-ccf-bdci-ocr-mczj-ocr-identificationidelement.