Giter Site home page Giter Site logo

applied-deep-learning's People

Contributors

maziarraissi avatar

Stargazers

 avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar

Watchers

 avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar

applied-deep-learning's Issues

EfficentNet Lecture Video Question

Hello Dr. Maziar Raissi, I am a PhD student in Machine Learning and sincerely appreciate you posting these videos online as they are very helpful. I have a small question that is not explained well in the paper and the authors skip over it. In the mnasnet paper, they use an empirical observation that doubling the Latency will increase accuracy by 5%: " for instance, we empirically observed doubling the latency usually brings about 5% relative accuracy gain." and therefore the authors solve for the omega parameter (setting alpha equal to beta) and obtain -0.07 from this empirical observation. However, this observation is only for LATENCY and not for flops or memory as the parameter they were optimizing was LATENCY. In the efficientnet paper the authors describe that they optimized flops instead of latency with an objective function (reward) as the following formula: ACC(m)*[FLOPS(m)/T]^w . With this reward function they use the same value for omega for controlling the trade off between accuracy and flops: "Specifically, we use the same search space as (Tan et al., 2019), and use ACC(m)ร—[FLOP S(m)/T]^w as the optimization goal, where ACC(m) and F LOPS(m) denote the accuracy and FLOPS of model m, T is the target FLOPS and w=-0.07 is a hyperparameter for controlling the trade-off between accuracy and FLOPS. Here, they use w=-0.07 which is based on the empirical observation from the mnasnet paper that doubling the LATENCY will increase accuracy by 5%. This observation is only for latency and not flops though. So my question is why they can make this assumption that w=-0.07 for flops as well unless this is some sort of guess for the w hyperparameter? Thanks, let me know if you would like me to elaborate on my confusion.

Newer Models

Hello Dr. Raissi,

I really like your videos! Wondering if you will be reviewing any newer papers/models from NLP side of things for example GPT-4, PaLM, or LLama2?

Thanks,

Karl Gardner

Missing Lecture Videos

Thank you for this incredible resource!
Will the missing lecture videos (for example the ones under advanced topics in part 1) ever be posted?

Source code of these wonderful slides

Thanks a lot for sharing these awesome slides. I was surprised that in one page how nicely you put everything in an organized manner. Is it possible to share the source code of the slides? Or at Atleast one so that we can also make something similar.

Recommend Projects

  • React photo React

    A declarative, efficient, and flexible JavaScript library for building user interfaces.

  • Vue.js photo Vue.js

    ๐Ÿ–– Vue.js is a progressive, incrementally-adoptable JavaScript framework for building UI on the web.

  • Typescript photo Typescript

    TypeScript is a superset of JavaScript that compiles to clean JavaScript output.

  • TensorFlow photo TensorFlow

    An Open Source Machine Learning Framework for Everyone

  • Django photo Django

    The Web framework for perfectionists with deadlines.

  • D3 photo D3

    Bring data to life with SVG, Canvas and HTML. ๐Ÿ“Š๐Ÿ“ˆ๐ŸŽ‰

Recommend Topics

  • javascript

    JavaScript (JS) is a lightweight interpreted programming language with first-class functions.

  • web

    Some thing interesting about web. New door for the world.

  • server

    A server is a program made to process requests and deliver data to clients.

  • Machine learning

    Machine learning is a way of modeling and interpreting data that allows a piece of software to respond intelligently.

  • Game

    Some thing interesting about game, make everyone happy.

Recommend Org

  • Facebook photo Facebook

    We are working to build community through open source technology. NB: members must have two-factor auth.

  • Microsoft photo Microsoft

    Open source projects and samples from Microsoft.

  • Google photo Google

    Google โค๏ธ Open Source for everyone.

  • D3 photo D3

    Data-Driven Documents codes.