haibuo1981 / cuda-convnet2 Goto Github PK
View Code? Open in Web Editor NEWAutomatically exported from code.google.com/p/cuda-convnet2
License: Apache License 2.0
Automatically exported from code.google.com/p/cuda-convnet2
License: Apache License 2.0
What steps will reproduce the problem?
1. Simply add a cost.sum2 layer into the layer definition
2. Run it
3.
What is the expected output? What do you see instead?
It crashed and said:
python: src/nvmatrix.cu:738: bool NVMatrix::resize(int, int, bool): Assertion
`_ownsData || (_numElements == numRows * numCols && isContiguous())' failed.
Error signal 6:
What version of the product are you using? On what operating system?
Latest cuda-convnet2 + Titan + CUDA 5.5 + Ubuntu 12.04
Please provide any additional information below.
To reproduce the problem, download the attached layer definition (reg.cfg and
reg-params.cfg) and test it with command:
python convnet.py --data-path=. --save-path=./tmp --test-range=1
--train-range=1 --layer-def=layers/reg-ori.cfg
--layer-params=layers/reg-params.cfg --data-provider=dummy-labeled-1 --gpu=0
It seems that getAct() of the sum2 layer will produce a 0*128 matrix and thus
cause the error.
Original issue reported on code.google.com by [email protected]
on 8 Aug 2014 at 12:42
Attachments:
What steps will reproduce the problem?
1. building project dumps a lot of deprecated warnings due to NPY api version
2.
3.
What is the expected output? What do you see instead?
What version of the product are you using? On what operating system?
Please provide any additional information below.
added some workarounds to remove the messages in my cloned version:
alexpark-numpyclean
Original issue reported on code.google.com by [email protected]
on 26 Nov 2014 at 9:58
Nvidia cards don't allow textures bigger than 512MB. Because this code uses
texture memory, this imposes a limit on the sizes of various buffers. For
example if your layer has too many filters (such that its output size exceeds
512MB), the code will crash.
TODO: add non-texture-using routines to bypass this.
Original issue reported on code.google.com by [email protected]
on 25 Jul 2014 at 1:28
What steps will reproduce the problem?
1. Follow the steps in Compiling, Data, and TrainingExample for ILSVRC2012
2. I used same parameters for running convnet except --train-freq=10
What is the expected output? What do you see instead?
Keep training with correct weights. (Top-5 error reaches 0.7 before 13th epochs)
After 14th epochs, conv1 weights and biases become nan and top-5 error becomes
0.99...
What version of the product are you using? On what operating system?
latest version of cuda-convnet2, Ubuntu 14.04
Please provide any additional information below.
Original issue reported on code.google.com by [email protected]
on 1 Dec 2014 at 1:07
What steps will reproduce the problem?
1. Compile Success -> Batch Generation Success -> Fail to run convnet.py
2.
3.
What is the expected output? What do you see instead?
src/nvmatrix.cu(394): getLastCudaError() CUDA error : kSetupCurand: Kernel
execution failed : (8) invalid device function.
What version of the product are you using? On what operating system?
"cuda-convnet2-c67ec1220aca" with cuda5.5/python2.7.3
Please provide any additional information below.
There's no problem with cuda-convnet(convnet1) code, but with cuda-convnet2,
this error occurred.
Original issue reported on code.google.com by [email protected]
on 1 Aug 2014 at 10:09
What steps will reproduce the problem?
1. Add an element-wise sum layer to your config file.
2. Specify exactly same inputs to the 'inputs' parameter
3. Specify coeffs=1, -1
What is the expected output? What do you see instead?
With exactly same inputs and the coeffs specified as 1 and -1, it is expected
that the output of the layer produce 0.0
Instead we see a non-zero output. Also changing the coeffs values does not seem
to have any effect.
What version of the product are you using? On what operating system?
cuda-convnet2 on ubuntu-linux
Please provide any additional information below.
Original issue reported on code.google.com by [email protected]
on 29 Oct 2014 at 11:17
I have some problems running this code on 8 gpus. It crashed at the line:
assert(same.size() == 3); in reducepipeline.cu
What steps will reproduce the problem?
1. get 8 k40 gpu, install them in 2 PCI buses. 4 for each.
2. train with 512 mini batch, data parallelism.
Original issue reported on code.google.com by [email protected]
on 3 Oct 2014 at 5:04
What steps will reproduce the problem?
1.
2.
3.
What is the expected output? What do you see instead?
What version of the product are you using? On what operating system?
Please provide any additional information below.
Not exactly a bug, but if I want to see predictions with shownet (python
shownet --show-preds=probs), the script loads all batches before showing me
predictions from the test batch.
If have many GBs of data for training, the script takes a lot of time before I
can see test the case predictions.
Original issue reported on code.google.com by [email protected]
on 18 Aug 2014 at 3:42
It looks like the RC for CUDA 6.5 is out:
https://developer.nvidia.com/cuda-toolkit
Original issue reported on code.google.com by [email protected]
on 31 Jul 2014 at 8:12
Will the code run good on GTX 770, 780 and 780Ti GPU?
Thanks.
Original issue reported on code.google.com by [email protected]
on 13 Oct 2014 at 12:02
What steps will reproduce the problem?
1. I can reproduce it if I am luck
2.
3.
What is the expected output? What do you see instead?
What version of the product are you using? On what operating system?
state of the art
Please provide any additional information below.
this is because the tye of cudaTextureObject_t is not a pointer
Original issue reported on code.google.com by [email protected]
on 30 Jul 2014 at 2:28
What steps will reproduce the problem?
1. train a model
2. multiview test the model and --test-out=1
3.
What is the expected output? What do you see instead?
probs matrix of multiview tested result.
All zero matrix
What version of the product are you using? On what operating system?
latest version
Please provide any additional information below.
Is --test-out function not yet developed? Since I saw the part of writing probs
matrix is commented. Or is there any other simple way to save the multiview
test predictions? thanks
Original issue reported on code.google.com by [email protected]
on 13 Aug 2014 at 6:45
When i try to run training example of cuda-convnet2 i get this error :
src/nvmatrix.cu(394) : getLastCudaError() CUDA error : kSetupCurand: Kernel
execution failed : (8) invalid device function .
I have GTX 980 on my machine and it has compute capability 5.2
I tried to modify makefiles in cudaconv3 & cudaconvnet & nvmatrix like this and
to add 52 instead of 50 tooand i stil have same error.
GENCODE_SM35 := -gencode arch=compute_35,code=sm_35
GENCODE_FLAGS := $(GENCODE_SM35)
to
GENCODE_SM35 := -gencode arch=compute_35,code=sm_35
GENCODE_SM50 := -gencode arch=compute_50,code=sm_50
GENCODE_FLAGS := $(GENCODE_SM50)
Original issue reported on code.google.com by [email protected]
on 11 Apr 2015 at 9:21
Attachments:
What steps will reproduce the problem?
1. There are more than 2 data layers
2. Use BinomialCrossEntropyCostLayer cost layer
3.
What is the expected output? What do you see instead?
The (output) dimension of third data layer is 1700.
Without adding "start=0 end=1700" in the layer definition file for the third
layer, the program will crash,
The error info is
src/../../cudaconv3/include/../../nvmatrix/include/nvmatrix.cuh:376: void
NVMatrix::applyBinary(Op, NVMatrix&, NVMatrix&, CUstream_st*) [with Op =
BinomialCrossEntOperator]: Assertion `this->isSameDims(b)' failed.
Then I add the following lines in layer.cu
int numCases = labels.getLeadingDim(); //line 2108 in layer.cu
printf("%d %d=====\n\n",probs.getNumRows(), probs.getNumCols());
printf("%d %d=====\n\n",labels.getNumRows(),labels.getNumCols());
The size of labels is (0, 1024), and the size of probs is (1700,1024).
After adding start=0 end=1700, the size will be correct, but I got the
following error,
CUDA error at src/../include/memory.cuh:272 code=2(cudaErrorMemoryAllocation) "cudaMalloc(data, size)"
What version of the product are you using? On what operating system?
Cuda5.5, CentOS6.5
Please provide any additional information below.
Original issue reported on code.google.com by [email protected]
on 26 Aug 2014 at 3:19
When running the convnet.py, I encountered the following error:
“1.1 (0.00%)...cannot allocate memory for thread-local data: ABORT”.
I have no idea about this.
By the way, I used cuda-convnet2 and the codes are running on Red Hat.
Original issue reported on code.google.com by [email protected]
on 30 Nov 2014 at 4:21
A declarative, efficient, and flexible JavaScript library for building user interfaces.
🖖 Vue.js is a progressive, incrementally-adoptable JavaScript framework for building UI on the web.
TypeScript is a superset of JavaScript that compiles to clean JavaScript output.
An Open Source Machine Learning Framework for Everyone
The Web framework for perfectionists with deadlines.
A PHP framework for web artisans
Bring data to life with SVG, Canvas and HTML. 📊📈🎉
JavaScript (JS) is a lightweight interpreted programming language with first-class functions.
Some thing interesting about web. New door for the world.
A server is a program made to process requests and deliver data to clients.
Machine learning is a way of modeling and interpreting data that allows a piece of software to respond intelligently.
Some thing interesting about visualization, use data art
Some thing interesting about game, make everyone happy.
We are working to build community through open source technology. NB: members must have two-factor auth.
Open source projects and samples from Microsoft.
Google ❤️ Open Source for everyone.
Alibaba Open Source for everyone
Data-Driven Documents codes.
China tencent open source team.