Topic: vit Goto Github
Some thing interesting about vit
Some thing interesting about vit
vit,Summary of Transformer applications for computer vision tasks.
User: aiprogrammer
vit,:robot: PaddleViT: State-of-the-art Visual Transformer and MLP Models for PaddlePaddle 2.0+
User: br-idl
Home Page: https://github.com/BR-IDL/PaddleViT
vit,A PyTorch implementation of "MobileViT: Light-weight, General-purpose, and Mobile-friendly Vision Transformer"
User: chinhsuanwu
Home Page: https://arxiv.org/abs/2110.02178
vit,An ultimately comprehensive paper list of Vision Transformer/Attention, including papers, codes, and related websites
User: cmhungsteve
vit,Vision Transformer Pruning
User: cydia2018
vit,An unofficial implementation of TubeViT in "Rethinking Video ViTs: Sparse Video Tubes for Joint Image and Video Learning"
User: daniel-code
vit,📖A small curated list of Awesome SD/DiT/ViT/Diffusion Inference with Distributed/Caching/Sampling: DistriFusion, PipeFusion, AsyncDiff, DeepCache, Block Caching etc.
User: deftruth
Home Page: https://github.com/DefTruth/Awesome-SD-Distributed-Inference
vit,i. A practical application of Transformer (ViT) on 2-D physiological signal (EEG) classification tasks. Also could be tried with EMG, EOG, ECG, etc. ii. Including the attention of spatial dimension (channel attention) and *temporal dimension*. iii. Common spatial pattern (CSP), an efficient feature enhancement method, realized with Python.
User: eeyhsong
vit,Continuous Augmented Positional Embeddings (CAPE) implementation for PyTorch
User: gcambara
vit,An Image is Worth 16x16 Words: Transformers for Image Recognition at Scale
User: gupta-abhay
Home Page: https://arxiv.org/abs/2010.11929
vit,[CVPR 2021] Official PyTorch implementation for Transformer Interpretability Beyond Attention Visualization, a novel method to visualize classifications by Transformer based networks.
User: hila-chefer
vit,Training ImageNet / CIFAR models with sota strategies and fancy techniques such as ViT, KD, Rep, etc.
User: hunto
vit,The code repo contains multiple code reproduction processes of various SOTA deep learning algorithms
User: hzcirving
vit,reproduction of semantic segmentation using masked autoencoder (mae)
User: implus
vit,An unofficial implementation of ViTPose [Y. Xu et al., 2022]
User: jaehyunnn
vit,Jittor Image Models is a library for pulling together a wide variety of SOTA deep learning models in the Jittor framework.
User: jittor-image-models
vit,Vision Transformer using TensorFlow 2.0
User: kamalkraj
Home Page: https://openreview.net/forum?id=YicbFdNTTy
vit,My implementation of "Patch n’ Pack: NaViT, a Vision Transformer for any Aspect Ratio and Resolution"
User: kyegomez
Home Page: https://discord.gg/qUtxnK2NMf
vit,Open source implementation of "Vision Transformers Need Registers"
User: kyegomez
Home Page: https://discord.gg/qUtxnK2NMf
vit,pix2tex: Using a ViT to convert images of equations into LaTeX code.
User: lukas-blecher
Home Page: https://lukas-blecher.github.io/LaTeX-OCR/
vit,Official Code of Paper "Reversible Column Networks" "RevColv2"
Organization: megvii-research
vit,Vision Transformer explanation and implementation with PyTorch
User: nerminnuraydogan
vit,Code for the paper "Revisiting Adversarial Training for ImageNet: Architectures, Training and Generalization across Threat Models"
User: nmndeep
vit,Open-source evaluation toolkit of large vision-language models (LVLMs), support ~100 VLMs, 30+ benchmarks
Organization: open-compass
Home Page: https://huggingface.co/spaces/opencompass/open_vlm_leaderboard
vit,PASSL包含 SimCLR,MoCo v1/v2,BYOL,CLIP,PixPro,simsiam, SwAV, BEiT,MAE 等图像自监督算法以及 Vision Transformer,DEiT,Swin Transformer,CvT,T2T-ViT,MLP-Mixer,XCiT,ConvNeXt,PVTv2 等基础视觉算法
Organization: paddlepaddle
vit,Paddle Large Scale Classification Tools,supports ArcFace, CosFace, PartialFC, Data Parallel + Model Parallel. Model includes ResNet, ViT, Swin, DeiT, CaiT, FaceViT, MoCo, MAE, ConvMAE, CAE.
Organization: paddlepaddle
vit,A PyTorch implementation of CNN+Vision Transformer for hyperspectral image classification
User: purbayankar
vit,HugsVision is a easy to use huggingface wrapper for state-of-the-art computer vision
User: qanastek
Home Page: https://pypi.org/project/hugsvision/
vit,This code repository contains the code used for my "Optimizing Memory Usage for Training LLMs and Vision Transformers in PyTorch" blog post.
User: rasbt
Home Page: https://lightning.ai/pages/community/tutorial/pytorch-memory-vit-llm/
vit,A fast, easy-to-use, production-ready inference server for computer vision supporting deployment of many popular model architectures and fine-tuned models.
Organization: roboflow
Home Page: https://inference.roboflow.com
vit,From scratch, simple and easy to understand PyTorch implementation of Vision Transformer (ViT) for small datasets like MNIST, FashionMNIST, SVHN and CIFAR10 with detailed steps.
User: s-chh
vit,Implementation of transformers based architecture in PyTorch.
User: shivamrajsharma
vit,A hub for innovation through web development projects
Organization: ssitvit
Home Page: https://codecanvas.ieeessitvit.com/
vit,Official PyTorch implementation of our ECCV 2022 paper "Sliced Recursive Transformer"
User: szq0214
vit,Flexible Python library providing building blocks (layers) for reproducible Transformers research (Tensorflow ✅, Pytorch 🔜, and Jax 🔜)
Organization: tensorops
vit,Towhee is a framework that is dedicated to making neural data processing pipelines simple and fast.
Organization: towhee-io
Home Page: https://towhee.io
vit,Code of TVT: Transferable Vision Transformer for Unsupervised Domain Adaptation, WACV 2023
Organization: uta-smile
vit,Extract video features from raw videos using multiple GPUs. We support RAFT flow frames as well as S3D, I3D, R(2+1)D, VGGish, CLIP, and TIMM models.
User: v-iashin
Home Page: https://v-iashin.github.io/video_features
vit,FFCS course registration made hassle free for VITians. Search courses and visualize the timetable on the go!
User: vatz88
Home Page: https://ffcsonthego.vatz88.in
vit,🚀 React application framework inspired by UmiJS / 类 UmiJS 的 React 应用框架
Organization: vitjs
vit,[MedIA Journal] An ultimately comprehensive paper list of Vision Transformer/Attention, including papers, codes, and related websites
Organization: xmindflow
vit,A paper list of some recent Transformer-based CV works.
User: yangzhangcst
vit,Mimix: A Text Generation Tool and Pretrained Chinese Models
User: yaoxiaoyuan
vit,ICCV2021, Tokens-to-Token ViT: Training Vision Transformers from Scratch on ImageNet
Organization: yitu-opensource
vit,[ECCV 2024] Pytorch code for our ECCV'24 paper NeRF-MAE: Masked AutoEncoders for Self-Supervised 3D Representation Learning for Neural Radiance Fields
User: zubair-irshad
Home Page: https://nerf-mae.github.io/
vit,A ViT based transformer applied on multi-channel time-series EEG data for motor imagery classification
User: zwcolin
A declarative, efficient, and flexible JavaScript library for building user interfaces.
🖖 Vue.js is a progressive, incrementally-adoptable JavaScript framework for building UI on the web.
TypeScript is a superset of JavaScript that compiles to clean JavaScript output.
An Open Source Machine Learning Framework for Everyone
The Web framework for perfectionists with deadlines.
A PHP framework for web artisans
Bring data to life with SVG, Canvas and HTML. 📊📈🎉
JavaScript (JS) is a lightweight interpreted programming language with first-class functions.
Some thing interesting about web. New door for the world.
A server is a program made to process requests and deliver data to clients.
Machine learning is a way of modeling and interpreting data that allows a piece of software to respond intelligently.
Some thing interesting about visualization, use data art
Some thing interesting about game, make everyone happy.
We are working to build community through open source technology. NB: members must have two-factor auth.
Open source projects and samples from Microsoft.
Google ❤️ Open Source for everyone.
Alibaba Open Source for everyone
Data-Driven Documents codes.
China tencent open source team.