Topic: model-compression Goto Github
Some thing interesting about model-compression
Some thing interesting about model-compression
model-compression,Infrastructures™ for Machine Learning Training/Inference in Production.
User: 1duo
model-compression,micronet, a model compression and deploy lib. compression: 1、quantization: quantization-aware-training(QAT), High-Bit(>2b)(DoReFa/Quantization and Training of Neural Networks for Efficient Integer-Arithmetic-Only Inference)、Low-Bit(≤2b)/Ternary and Binary(TWN/BNN/XNOR-Net); post-training-quantization(PTQ), 8-bit(tensorrt); 2、 pruning: normal、regular and group convolutional channel pruning; 3、 group convolution structure; 4、batch-normalization fuse for quantization. deploy: tensorrt, fp32/fp16/int8(ptq-calibration)、op-adapt(upsample)、dynamic_shape
User: 666dzy666
model-compression,Pytorch implementation of various Knowledge Distillation (KD) methods.
User: aberhu
model-compression,TinyNeuralNetwork is an efficient and easy-to-use deep learning model compression framework.
Organization: alibaba
model-compression,Awesome machine learning model compression research papers, tools, and learning material.
User: cedrickchee
model-compression,(CVPR 2021, Oral) Dynamic Slimmable Network
User: changlin31
model-compression,Papers for deep neural network compression and acceleration
User: chester256
model-compression,Lightweight and Scalable framework that combines mainstream algorithms of Click-Through-Rate prediction based computational DAG, philosophy of Parameter Server and Ring-AllReduce collective communication.
User: cnkuangshi
model-compression,SlimSAM: 0.1% Data Makes Segment Anything Slim
User: czg1225
model-compression,Awesome Knowledge Distillation
User: dkozlov
model-compression,Channel Pruning for Accelerating Very Deep Neural Networks (ICCV'17)
User: ethanhe42
Home Page: https://yihui-he.github.io/blog/channel-pruning-for-accelerating-very-deep-neural-networks
model-compression,Awesome Knowledge-Distillation. 分类整理的知识蒸馏paper(2014-2021)。
User: flhonker
model-compression,A list of high-quality (newest) AutoML works and lightweight models including 1.) Neural Architecture Search, 2.) Lightweight Structures, 3.) Model Compression, Quantization and Acceleration, 4.) Hyperparameter Optimization, 5.) Automated Feature Engineering.
User: guan-yuan
model-compression,A PyTorch implementation for exploring deep and shallow knowledge distillation (KD) experiments with flexibility
User: haitongli
model-compression,A curated list of neural network pruning resources.
User: he-y
model-compression,Filter Pruning via Geometric Median for Deep Convolutional Neural Networks Acceleration (CVPR 2019 Oral)
User: he-y
Home Page: https://arxiv.org/abs/1811.00250
model-compression,Soft Filter Pruning for Accelerating Deep Convolutional Neural Networks
User: he-y
Home Page: https://arxiv.org/abs/1808.06866
model-compression,[CVPR 2024] DeepCache: Accelerating Diffusion Models for Free
User: horseee
Home Page: https://horseee.github.io/Diffusion_DeepCache/
model-compression,Knowledge distillation in text classification with pytorch. 知识蒸馏,中文文本分类,教师模型BERT、XLNET,学生模型biLSTM。
User: hoytta0
model-compression,A list of papers, docs, codes about model quantization. This repo is aimed to provide the info for model quantization research, we are continuously improving the project. Welcome to PR the works (papers, repositories) that are missed by the repo.
User: htqin
model-compression,Efficient AI Backbones including GhostNet, TNT and MLP, developed by Huawei Noah's Ark Lab.
Organization: huawei-noah
model-compression,Efficient computing methods developed by Huawei Noah's Ark Lab
Organization: huawei-noah
model-compression,Pretrained language model and its related optimization techniques developed by Huawei Noah's Ark Lab.
Organization: huawei-noah
model-compression,[CVPR2020] GhostNet: More Features from Cheap Operations
User: iamhankai
Home Page: https://arxiv.org/abs/1911.11907
model-compression,⛵️The official PyTorch implementation for "BERT-of-Theseus: Compressing BERT by Progressive Module Replacing" (EMNLP 2020).
User: jetrunner
model-compression,[ICML'21 Oral] I-BERT: Integer-only BERT Quantization
User: kssteven418
Home Page: https://arxiv.org/abs/2101.01321
model-compression,knowledge distillation papers
User: lhyfst
model-compression,Deep Face Model Compression
User: liuziwei7
Home Page: http://personal.ie.cuhk.edu.hk/~lz013/projects/MobileID.html
model-compression,Accelerate your Neural Architecture Search (NAS) through fast, reproducible and modular research.
Organization: microsoft
Home Page: https://microsoft.github.io/archai
model-compression,NLP DNN Toolkit - Building Your NLP DNN Models Like Playing Lego
Organization: microsoft
model-compression,An open source AutoML toolkit for automate machine learning lifecycle, including feature engineering, neural architecture search, model compression and hyper-parameter tuning.
Organization: microsoft
Home Page: https://nni.readthedocs.io
model-compression,Collection of recent methods on (deep) neural network compression and acceleration.
User: mingsun-tse
model-compression,[ECCV 2018] AMC: AutoML for Model Compression and Acceleration on Mobile Devices
Organization: mit-han-lab
Home Page: https://arxiv.org/abs/1802.03494
model-compression,yolov3 by pytorch
User: peterisfar
model-compression,The Truth Is In There: Improving Reasoning in Language Models with Layer-Selective Rank Reduction
User: pratyushasharma
Home Page: https://pratyushasharma.github.io/laser/
model-compression,A Pytorch Knowledge Distillation library for benchmarking and extending works in the domains of Knowledge Distillation, Pruning, and Quantization.
Organization: sforaidl
Home Page: https://kd-lib.readthedocs.io/
model-compression,Yolov5 distillation training | Yolov5知识蒸馏训练,支持训练自己的数据
User: sharpiless
model-compression,KVQuant: Towards 10 Million Context Length LLM Inference with KV Cache Quantization
Organization: squeezeailab
Home Page: https://arxiv.org/abs/2401.18079
model-compression,SqueezeLLM: Dense-and-Sparse Quantization
Organization: squeezeailab
Home Page: https://arxiv.org/abs/2306.07629
model-compression,[CVPR 2024 Highlight] Logit Standardization in Knowledge Distillation
User: sunshangquan
model-compression,An Automatic Model Compression (AutoMC) framework for developing smaller and faster AI applications.
Organization: tencent
Home Page: https://pocketflow.github.io
model-compression,A toolkit to optimize ML models for deployment for Keras and TensorFlow, including quantization and pruning.
Organization: tensorflow
Home Page: https://www.tensorflow.org/model_optimization
model-compression,针对pytorch模型的自动化模型结构分析和修改工具集,包含自动分析模型结构的模型压缩算法库
Organization: thu-mig
model-compression,OTOv1-v3, NeurIPS, ICLR, TMLR, DNN Training, Compression, Structured Pruning, Erasing Operators, CNN, Diffusion, LLM
User: tianyic
model-compression,[CVPR 2023] Towards Any Structural Pruning; LLMs / SAM / Diffusion / Transformers / YOLOv8 / CNNs
User: vainf
Home Page: https://arxiv.org/abs/2301.12900
model-compression,Java interface for fastText
User: vinhkhuc
model-compression,[ICCV 2023] Q-Diffusion: Quantizing Diffusion Models.
User: xiuyu-li
Home Page: https://xiuyuli.com/qdiffusion/
model-compression,Code for "Co-Evolutionary Compression for Unpaired Image Translation" (ICCV 2019), "SCOP: Scientific Control for Reliable Neural Network Pruning" (NeurIPS 2020) and “Manifold Regularized Dynamic Network Pruning” (CVPR 2021).
User: yehuitang
model-compression,List of papers related to neural network quantization in recent AI conferences and journals.
User: zhen-dong
model-compression,Quantization library for PyTorch. Support low-precision and mixed-precision quantization, with hardware implementation through TVM.
User: zhen-dong
A declarative, efficient, and flexible JavaScript library for building user interfaces.
🖖 Vue.js is a progressive, incrementally-adoptable JavaScript framework for building UI on the web.
TypeScript is a superset of JavaScript that compiles to clean JavaScript output.
An Open Source Machine Learning Framework for Everyone
The Web framework for perfectionists with deadlines.
A PHP framework for web artisans
Bring data to life with SVG, Canvas and HTML. 📊📈🎉
JavaScript (JS) is a lightweight interpreted programming language with first-class functions.
Some thing interesting about web. New door for the world.
A server is a program made to process requests and deliver data to clients.
Machine learning is a way of modeling and interpreting data that allows a piece of software to respond intelligently.
Some thing interesting about visualization, use data art
Some thing interesting about game, make everyone happy.
We are working to build community through open source technology. NB: members must have two-factor auth.
Open source projects and samples from Microsoft.
Google ❤️ Open Source for everyone.
Alibaba Open Source for everyone
Data-Driven Documents codes.
China tencent open source team.