Topic: inference Goto Github
Some thing interesting about inference
Some thing interesting about inference
inference,On-device Speech Recognition for Apple Silicon
Organization: argmaxinc
Home Page: https://takeargmax.com/blog/whisperkit
inference,An easy-to-use LLMs quantization package with user-friendly apis, based on GPTQ algorithm.
Organization: autogptq
inference,Example 📓 Jupyter notebooks that demonstrate how to build, train, and deploy machine learning models using 🧠 Amazon SageMaker.
Organization: aws
Home Page: https://sagemaker-examples.readthedocs.io
inference,Multi Model Server is a tool for serving neural net models for inference
Organization: awslabs
inference,LightSeq: A High Performance Library for Sequence Processing and Generation
Organization: bytedance
inference,DELTA is a deep learning based natural language and speech processing platform.
Organization: delta-ml
Home Page: https://delta-didi.readthedocs.io/
inference,Hello AI World guide to deploying deep-learning inference networks and deep vision primitives with TensorRT and NVIDIA Jetson.
User: dusty-nv
Home Page: https://developer.nvidia.com/embedded/twodaystoademo
inference,Deploy a ML inference service on a budget in less than 10 lines of code.
Organization: ebhy
inference,Efficient, scalable and enterprise-grade CPU/GPU inference server for 🤗 Hugging Face transformer models 🚀
Organization: els-rd
Home Page: https://els-rd.github.io/transformer-deploy/
inference,Package for causal inference in graphs and in the pairwise settings. Tools for graph structure recovery and dependencies are included.
Organization: fentechsolutions
Home Page: https://fentechsolutions.github.io/CausalDiscoveryToolbox/html/index.html
inference,Runtime type system for IO decoding/encoding
User: gcanti
Home Page: https://gcanti.github.io/io-ts/
inference,Port of OpenAI's Whisper model in C/C++
User: ggerganov
inference,Cross-platform, customizable ML solutions for live and streaming media.
Organization: google-ai-edge
Home Page: https://mediapipe.dev
inference,High-efficiency floating-point neural network inference operators for mobile, server, and Web
Organization: google
inference,🎨 The exhaustive Pattern Matching library for TypeScript, with smart type inference.
User: gvergnaud
inference,Making large AI models cheaper, faster and more accessible
Organization: hpcaitech
Home Page: https://www.colossalai.org
inference,Utilities to use the Hugging Face Hub API
Organization: huggingface
Home Page: https://hf.co/docs/huggingface.js
inference,🚀 Accelerate training and inference of 🤗 Transformers and 🤗 Diffusers with easy to use hardware optimization tools
Organization: huggingface
Home Page: https://huggingface.co/docs/optimum/main/
inference,Large Language Model Text Generation Inference
Organization: huggingface
Home Page: http://hf.co/docs/text-generation-inference
inference,Pytorch-Named-Entity-Recognition-with-BERT
User: kamalkraj
inference, 💎1MB lightweight face detection model (1MB轻量级人脸检测模型)
User: linzaer
inference,Acceleration package for neural networks on multi-core CPUs
User: maratyszcza
inference,AICI: Prompts as (Wasm) Programs
Organization: microsoft
inference,DeepSpeed is a deep learning optimization library that makes distributed training and inference easy, efficient, and effective.
Organization: microsoft
Home Page: https://www.deepspeed.ai/
inference,MII makes low-latency and high-throughput inference possible, powered by DeepSpeed.
Organization: microsoft
inference,Sparsity-aware deep learning inference runtime for CPUs
Organization: neuralmagic
Home Page: https://neuralmagic.com/deepsparse/
inference,An easy to use PyTorch to TensorRT converter
Organization: nvidia-ai-iot
inference,NVIDIA® TensorRT™ is an SDK for high-performance deep learning inference on NVIDIA GPUs. This repository contains the open source components of TensorRT.
Organization: nvidia
Home Page: https://developer.nvidia.com/tensorrt
inference,Fast inference engine for Transformer models
Organization: opennmt
Home Page: https://opennmt.net/CTranslate2
inference,Pre-trained Deep Learning models and demos (high quality and extremely fast)
Organization: openvinotoolkit
Home Page: https://docs.openvino.ai/latest/model_zoo.html
inference,OpenVINO™ is an open-source toolkit for optimizing and deploying AI inference
Organization: openvinotoolkit
Home Page: https://docs.openvino.ai
inference,📚 Jupyter notebook tutorials for OpenVINO™
Organization: openvinotoolkit
inference,Python Library for learning (Structure and Parameter), inference (Probabilistic and Causal), and simulations in Bayesian Networks.
Organization: pgmpy
Home Page: https://pgmpy.org/
inference,A fast, easy-to-use, production-ready inference server for computer vision supporting deployment of many popular model architectures and fine-tuned models.
Organization: roboflow
Home Page: https://inference.roboflow.com
inference,Superduper: Bring AI to your database! Integrate AI models and workflows with your database to implement custom AI applications, without moving your data. Including streaming inference, scalable model hosting, training and vector search.
Organization: superduper-io
Home Page: https://superduper.io
inference,Faster Whisper transcription with CTranslate2
Organization: systran
inference,Inference Llama 2 in one file of pure 🔥
User: tairov
Home Page: https://www.modular.com/blog/community-spotlight-how-i-built-llama2-by-aydyn-tairov
inference,ncnn is a high-performance neural network inference framework optimized for the mobile platform
Organization: tencent
inference,TNN: developed by Tencent Youtu Lab and Guangying Lab, a uniform deep learning inference framework for mobile、desktop and server. TNN is distinguished by several outstanding features, including its cross-platform capability, high performance, model compression and code pruning. Based on ncnn and Rapidnet, TNN further strengthens the support and performance optimization for mobile devices, and also draws on the advantages of good extensibility and high performance from existed open source efforts. TNN has been deployed in multiple Apps from Tencent, such as Mobile QQ, Weishi, Pitu, etc. Contributions are welcome to work in collaborative with us and make TNN a better framework.
Organization: tencent
inference,a fast and user-friendly runtime for transformer inference (Bert, Albert, GPT2, Decoders, etc) on CPU and GPU.
Organization: tencent
inference,cube studio开源云原生一站式机器学习/深度学习/大模型AI平台,支持sso登录,多租户,大数据平台对接,notebook在线开发,拖拉拽任务流pipeline编排,多机多卡分布式训练,超参搜索,推理服务VGPU,边缘计算,serverless,标注平台,自动化标注,数据集管理,大模型微调,vllm大模型推理,llmops,私有知识库,AI模型应用商店,支持模型一键开发/推理/微调,支持国产cpu/gpu/npu芯片,支持RDMA,支持pytorch/tf/mxnet/deepspeed/paddle/colossalai/horovod/spark/ray/volcano分布式
Organization: tencentmusic
inference,TensorFlow template application for deep learning
User: tobegit3hub
inference,The Triton Inference Server provides an optimized cloud and edge inferencing solution.
Organization: triton-inference-server
Home Page: https://docs.nvidia.com/deeplearning/triton-inference-server/user-guide/docs/index.html
inference,Adversarial Robustness Toolbox (ART) - Python Library for Machine Learning Security - Evasion, Poisoning, Extraction, Inference - Red and Blue Teams
Organization: trusted-ai
Home Page: https://adversarial-robustness-toolbox.readthedocs.io/en/latest/
inference,The challenge projects for Inferencing machine learning models on iOS
User: tucan9389
inference,TypeDB: one giant leap for databases
Organization: typedb
Home Page: https://typedb.com
inference,A uniform interface to run deep learning models from multiple frameworks
Organization: uber
Home Page: https://neuropod.ai
inference,A high-throughput and memory-efficient inference and serving engine for LLMs
Organization: vllm-project
Home Page: https://docs.vllm.ai
inference,Replace OpenAI GPT with another LLM in your app by changing a single line of code. Xinference gives you the freedom to use any LLM you need. With Xinference, you're empowered to run inference with any open-source language models, speech recognition models, and multimodal models, whether in the cloud, on-premises, or even on your laptop.
Organization: xorbitsai
Home Page: https://inference.readthedocs.io
inference,校招、秋招、春招、实习好项目!带你从零实现一个高性能的深度学习推理库,支持大模型 llama2 、Unet、Yolov5、Resnet等模型的推理。Implement a high-performance deep learning inference library step by step
User: zjhellofss
A declarative, efficient, and flexible JavaScript library for building user interfaces.
🖖 Vue.js is a progressive, incrementally-adoptable JavaScript framework for building UI on the web.
TypeScript is a superset of JavaScript that compiles to clean JavaScript output.
An Open Source Machine Learning Framework for Everyone
The Web framework for perfectionists with deadlines.
A PHP framework for web artisans
Bring data to life with SVG, Canvas and HTML. 📊📈🎉
JavaScript (JS) is a lightweight interpreted programming language with first-class functions.
Some thing interesting about web. New door for the world.
A server is a program made to process requests and deliver data to clients.
Machine learning is a way of modeling and interpreting data that allows a piece of software to respond intelligently.
Some thing interesting about visualization, use data art
Some thing interesting about game, make everyone happy.
We are working to build community through open source technology. NB: members must have two-factor auth.
Open source projects and samples from Microsoft.
Google ❤️ Open Source for everyone.
Alibaba Open Source for everyone
Data-Driven Documents codes.
China tencent open source team.