Statistics for topic inference
RepositoryStats tracks 595,858 Github repositories, of these 309 are tagged with the inference topic. The most common primary language for repositories using this topic is Python (133). Other languages include: C++ (59), Jupyter Notebook (27), Rust (11), TypeScript (11)
Stargazers over time for topic inference
Most starred repositories for topic inference (view more)
Trending repositories for topic inference (view more)
A high-throughput and memory-efficient inference and serving engine for LLMs
Cross-platform, customizable ML solutions for live and streaming media.
DeepSpeed is a deep learning optimization library that makes distributed training and inference easy, efficient, and effective.
A project that optimizes OWL-ViT for real-time inference with NVIDIA TensorRT.
Python library for YOLO small object detection and instance segmentation
Neural Network-Boosted Importance Nested Sampling for Bayesian Statistics
A high-throughput and memory-efficient inference and serving engine for LLMs
DeepSpeed is a deep learning optimization library that makes distributed training and inference easy, efficient, and effective.
SGLang is a fast serving framework for large language models and vision language models.
The Qualcomm® AI Hub apps are a collection of state-of-the-art machine learning models optimized for performance (latency, memory etc.) and ready to deploy on Qualcomm® devices.
Python library for YOLO small object detection and instance segmentation
A high-throughput and memory-efficient inference and serving engine for LLMs
SGLang is a fast serving framework for large language models and vision language models.
The Qualcomm® AI Hub apps are a collection of state-of-the-art machine learning models optimized for performance (latency, memory etc.) and ready to deploy on Qualcomm® devices.
SGLang is a fast serving framework for large language models and vision language models.
Any model. Any hardware. Zero compromise. Built with @ziglang / @openxla / MLIR / @bazelbuild
A high-throughput and memory-efficient inference and serving engine for LLMs
SGLang is a fast serving framework for large language models and vision language models.
DeepSpeed is a deep learning optimization library that makes distributed training and inference easy, efficient, and effective.
SGLang is a fast serving framework for large language models and vision language models.