Comments (4)
finally I got it work:
function zerograd() 0 end
@zerograd zerograd()
import Base.convert
for t in (Float32, Float64)
@eval @primitive convert(T::Type{$t},x),dy zerograd() T(dy)
end
Wondering is there better solutions.
from autograd.jl.
@ylxdzsw Your solution works, here are some suggestions for improvement:
- Read the comments in src/core.jl (section 6) to see how gradients are actually defined.
- Your case is closest to 6.4.
- In your solution, instead of zerograd() you can simply use 0.
- Instead of T(dy), using convert(T,dy) may be better.
from autograd.jl.
I experimented with defining a general convert
gradient in AutoGrad. Unfortunately Julia keeps calling convert
in all sorts of unexpected places and so far I could not get it to work without breaking AutoGrad. However defining it for specific types like you have done is ok. There is an example for KnetArray / Array conversion in the latest Knet release. It shows an alternative to your solution that does not use the @primitive
macro and leaves the gradient wrt first arg undefined.
from autograd.jl.
Thanks, that solution looks nice.
from autograd.jl.
Related Issues (20)
- grad doesn't work on trivial identity HOT 6
- Copy of param is value of param HOT 3
- bug in gradient with broadcast HOT 3
- cannot take derivative of a scalar function HOT 6
- Atom canβt evaluate a line which the regular REPL can
- Inside `@diff`, shouldn't `similar(Result{Array})` return `Result{Array}`? HOT 1
- mapslices and Param wrapped arrays
- Extend UngetIndex to multiple indices, values for efficiency HOT 1
- Autograd.Sparse type causes regression HOT 10
- getindex gradient collection bug
- Cannot `convert` an object of type UnitRange{Int64} to an object of type Colon
- Multiple functions sharing some gradients HOT 1
- Problems with grad after grad HOT 2
- Alkan bias add problem? HOT 1
- TagBot trigger issue HOT 2
- how to calculate gradient without calculating function outputs
- keyword args in broadcasted funcs not supported
- Param(x) as AbstractArray? HOT 1
- Incorrect derivatives when new arrays are constructed in computation
- AutoGrad and CUDA kernels HOT 1
Recommend Projects
-
React
A declarative, efficient, and flexible JavaScript library for building user interfaces.
-
Vue.js
π Vue.js is a progressive, incrementally-adoptable JavaScript framework for building UI on the web.
-
Typescript
TypeScript is a superset of JavaScript that compiles to clean JavaScript output.
-
TensorFlow
An Open Source Machine Learning Framework for Everyone
-
Django
The Web framework for perfectionists with deadlines.
-
Laravel
A PHP framework for web artisans
-
D3
Bring data to life with SVG, Canvas and HTML. πππ
-
Recommend Topics
-
javascript
JavaScript (JS) is a lightweight interpreted programming language with first-class functions.
-
web
Some thing interesting about web. New door for the world.
-
server
A server is a program made to process requests and deliver data to clients.
-
Machine learning
Machine learning is a way of modeling and interpreting data that allows a piece of software to respond intelligently.
-
Visualization
Some thing interesting about visualization, use data art
-
Game
Some thing interesting about game, make everyone happy.
Recommend Org
-
Facebook
We are working to build community through open source technology. NB: members must have two-factor auth.
-
Microsoft
Open source projects and samples from Microsoft.
-
Google
Google β€οΈ Open Source for everyone.
-
Alibaba
Alibaba Open Source for everyone
-
D3
Data-Driven Documents codes.
-
Tencent
China tencent open source team.
from autograd.jl.