Statistics for topic model-compression
RepositoryStats tracks 595,856 Github repositories, of these 108 are tagged with the model-compression topic. The most common primary language for repositories using this topic is Python (78).
Stargazers over time for topic model-compression
Most starred repositories for topic model-compression (view more)
Trending repositories for topic model-compression (view more)
Awesome Knowledge-Distillation. 分类整理的知识蒸馏paper(2014-2021)。
Official Repo for SparseLLM: Global Pruning of LLMs (NeurIPS 2024)
Official Repo for SparseLLM: Global Pruning of LLMs (NeurIPS 2024)
Awesome Knowledge-Distillation. 分类整理的知识蒸馏paper(2014-2021)。
Channel Pruning for Accelerating Very Deep Neural Networks (ICCV'17)
Awesome Knowledge-Distillation. 分类整理的知识蒸馏paper(2014-2021)。
A list of papers, docs, codes about model quantization. This repo is aimed to provide the info for model quantization research, we are continuously improving the project. Welcome to PR the works (pape...
Official Repo for SparseLLM: Global Pruning of LLMs (NeurIPS 2024)
Official Code for "SVD-LLM: Truncation-aware Singular Value Decomposition for Large Language Model Compression"
The official implementation of the paper <MoA: Mixture of Sparse Attention for Automatic Large Language Model Compression>
Gather research papers, corresponding codes (if having), reading notes and any other related materials about Hot🔥🔥🔥 fields in Computer Vision based on Deep Learning.
A list of papers, docs, codes about model quantization. This repo is aimed to provide the info for model quantization research, we are continuously improving the project. Welcome to PR the works (pape...
Efficient AI Backbones including GhostNet, TNT and MLP, developed by Huawei Noah's Ark Lab.
Awesome Knowledge-Distillation. 分类整理的知识蒸馏paper(2014-2021)。
Official Repo for SparseLLM: Global Pruning of LLMs (NeurIPS 2024)
😎 A curated list of tensor decomposition resources for model compression.
Official Code for "SVD-LLM: Truncation-aware Singular Value Decomposition for Large Language Model Compression"
The official implementation of the paper <MoA: Mixture of Sparse Attention for Automatic Large Language Model Compression>
[CVPR 2024 Highlight] Logit Standardization in Knowledge Distillation
[NeurIPS 2024] KVQuant: Towards 10 Million Context Length LLM Inference with KV Cache Quantization
The official implementation of the paper "What Matters in Transformers? Not All Attention is Needed".
An open source AutoML toolkit for automate machine learning lifecycle, including feature engineering, neural architecture search, model compression and hyper-parameter tuning.
A list of papers, docs, codes about model quantization. This repo is aimed to provide the info for model quantization research, we are continuously improving the project. Welcome to PR the works (pape...
[CVPR 2024] DeepCache: Accelerating Diffusion Models for Free
Efficient AI Backbones including GhostNet, TNT and MLP, developed by Huawei Noah's Ark Lab.
The Truth Is In There: Improving Reasoning in Language Models with Layer-Selective Rank Reduction
Official Code for "SVD-LLM: Truncation-aware Singular Value Decomposition for Large Language Model Compression"
a collection of computer vision projects&tools. 计算机视觉方向项目和工具集合。