Trending repositories for topic model-compression
[CVPR 2023] Towards Any Structural Pruning; LLMs / SAM / Diffusion / Transformers / YOLOv8 / CNNs
Filter Pruning via Geometric Median for Deep Convolutional Neural Networks Acceleration (CVPR 2019 Oral)
[CVPR 2024 Highlight] Logit Standardization in Knowledge Distillation
A PyTorch implementation for exploring deep and shallow knowledge distillation (KD) experiments with flexibility
KVQuant: Towards 10 Million Context Length LLM Inference with KV Cache Quantization
List of papers related to neural network quantization in recent AI conferences and journals.
Efficient computing methods developed by Huawei Noah's Ark Lab
A list of papers, docs, codes about model quantization. This repo is aimed to provide the info for model quantization research, we are continuously improving the project. Welcome to PR the works (pape...
Pretrained language model and its related optimization techniques developed by Huawei Noah's Ark Lab.
Efficient AI Backbones including GhostNet, TNT and MLP, developed by Huawei Noah's Ark Lab.
Pytorch implementation of various Knowledge Distillation (KD) methods.
Awesome Knowledge-Distillation. 分类整理的知识蒸馏paper(2014-2021)。
An open source AutoML toolkit for automate machine learning lifecycle, including feature engineering, neural architecture search, model compression and hyper-parameter tuning.
[CVPR 2024 Highlight] Logit Standardization in Knowledge Distillation
Filter Pruning via Geometric Median for Deep Convolutional Neural Networks Acceleration (CVPR 2019 Oral)
KVQuant: Towards 10 Million Context Length LLM Inference with KV Cache Quantization
List of papers related to neural network quantization in recent AI conferences and journals.
[CVPR 2023] Towards Any Structural Pruning; LLMs / SAM / Diffusion / Transformers / YOLOv8 / CNNs
A PyTorch implementation for exploring deep and shallow knowledge distillation (KD) experiments with flexibility
Efficient computing methods developed by Huawei Noah's Ark Lab
[CVPR 2024] DeepCache: Accelerating Diffusion Models for Free
A list of papers, docs, codes about model quantization. This repo is aimed to provide the info for model quantization research, we are continuously improving the project. Welcome to PR the works (pape...
Pretrained language model and its related optimization techniques developed by Huawei Noah's Ark Lab.
Pytorch implementation of various Knowledge Distillation (KD) methods.
Efficient AI Backbones including GhostNet, TNT and MLP, developed by Huawei Noah's Ark Lab.
Awesome Knowledge-Distillation. 分类整理的知识蒸馏paper(2014-2021)。
An open source AutoML toolkit for automate machine learning lifecycle, including feature engineering, neural architecture search, model compression and hyper-parameter tuning.
[CVPR 2023] Towards Any Structural Pruning; LLMs / SAM / Diffusion / Transformers / YOLOv8 / CNNs
Filter Pruning via Geometric Median for Deep Convolutional Neural Networks Acceleration (CVPR 2019 Oral)
[CVPR 2024 Highlight] Logit Standardization in Knowledge Distillation
List of papers related to neural network quantization in recent AI conferences and journals.
A PyTorch implementation for exploring deep and shallow knowledge distillation (KD) experiments with flexibility
Pretrained language model and its related optimization techniques developed by Huawei Noah's Ark Lab.
Efficient AI Backbones including GhostNet, TNT and MLP, developed by Huawei Noah's Ark Lab.
Efficient computing methods developed by Huawei Noah's Ark Lab
Pytorch implementation of various Knowledge Distillation (KD) methods.
A list of papers, docs, codes about model quantization. This repo is aimed to provide the info for model quantization research, we are continuously improving the project. Welcome to PR the works (pape...
Collection of recent methods on (deep) neural network compression and acceleration.
Awesome Knowledge-Distillation. 分类整理的知识蒸馏paper(2014-2021)。
[CVPR 2024 Highlight] Logit Standardization in Knowledge Distillation
Filter Pruning via Geometric Median for Deep Convolutional Neural Networks Acceleration (CVPR 2019 Oral)
List of papers related to neural network quantization in recent AI conferences and journals.
KVQuant: Towards 10 Million Context Length LLM Inference with KV Cache Quantization
[CVPR 2024] DeepCache: Accelerating Diffusion Models for Free
[CVPR 2023] Towards Any Structural Pruning; LLMs / SAM / Diffusion / Transformers / YOLOv8 / CNNs
The Truth Is In There: Improving Reasoning in Language Models with Layer-Selective Rank Reduction
Efficient computing methods developed by Huawei Noah's Ark Lab
Collection of recent methods on (deep) neural network compression and acceleration.
A PyTorch implementation for exploring deep and shallow knowledge distillation (KD) experiments with flexibility
Pytorch implementation of various Knowledge Distillation (KD) methods.
A list of papers, docs, codes about model quantization. This repo is aimed to provide the info for model quantization research, we are continuously improving the project. Welcome to PR the works (pape...
Pretrained language model and its related optimization techniques developed by Huawei Noah's Ark Lab.
[CVPR 2023] Towards Any Structural Pruning; LLMs / SAM / Diffusion / Transformers / YOLOv8 / CNNs
An open source AutoML toolkit for automate machine learning lifecycle, including feature engineering, neural architecture search, model compression and hyper-parameter tuning.
Efficient AI Backbones including GhostNet, TNT and MLP, developed by Huawei Noah's Ark Lab.
A list of papers, docs, codes about model quantization. This repo is aimed to provide the info for model quantization research, we are continuously improving the project. Welcome to PR the works (pape...
[CVPR 2024] DeepCache: Accelerating Diffusion Models for Free
KVQuant: Towards 10 Million Context Length LLM Inference with KV Cache Quantization
[CVPR 2024 Highlight] Logit Standardization in Knowledge Distillation
Filter Pruning via Geometric Median for Deep Convolutional Neural Networks Acceleration (CVPR 2019 Oral)
Awesome Knowledge-Distillation. 分类整理的知识蒸馏paper(2014-2021)。
List of papers related to neural network quantization in recent AI conferences and journals.
Pytorch implementation of various Knowledge Distillation (KD) methods.
A PyTorch implementation for exploring deep and shallow knowledge distillation (KD) experiments with flexibility
Collection of recent methods on (deep) neural network compression and acceleration.
Pretrained language model and its related optimization techniques developed by Huawei Noah's Ark Lab.
KVQuant: Towards 10 Million Context Length LLM Inference with KV Cache Quantization
[CVPR 2024 Highlight] Logit Standardization in Knowledge Distillation
List of papers related to neural network quantization in recent AI conferences and journals.
A list of papers, docs, codes about efficient AIGC. This repo is aimed to provide the info for efficient AIGC research, including language and vision, we are continuously improving the project. Welcom...
[CVPR 2024] DeepCache: Accelerating Diffusion Models for Free
a collection of computer vision projects&tools. 计算机视觉方向项目和工具集合。
[ICML 2023] UPop: Unified and Progressive Pruning for Compressing Vision-Language Transformers.
Filter Pruning via Geometric Median for Deep Convolutional Neural Networks Acceleration (CVPR 2019 Oral)
A toolkit to boost the productivity of machine learning engineers.
Gather research papers, corresponding codes (if having), reading notes and any other related materials about Hot🔥🔥🔥 fields in Computer Vision based on Deep Learning.
[CVPR 2023] Towards Any Structural Pruning; LLMs / SAM / Diffusion / Transformers / YOLOv8 / CNNs
Resources of our survey paper "Enabling AI on Edges: Techniques, Applications and Challenges"
The Truth Is In There: Improving Reasoning in Language Models with Layer-Selective Rank Reduction
[CVPR 2024 Highlight] Logit Standardization in Knowledge Distillation
KVQuant: Towards 10 Million Context Length LLM Inference with KV Cache Quantization
[CVPR 2023] Towards Any Structural Pruning; LLMs / SAM / Diffusion / Transformers / YOLOv8 / CNNs
An open source AutoML toolkit for automate machine learning lifecycle, including feature engineering, neural architecture search, model compression and hyper-parameter tuning.
[CVPR 2024] DeepCache: Accelerating Diffusion Models for Free
A list of papers, docs, codes about model quantization. This repo is aimed to provide the info for model quantization research, we are continuously improving the project. Welcome to PR the works (pape...
Efficient AI Backbones including GhostNet, TNT and MLP, developed by Huawei Noah's Ark Lab.
Efficient computing methods developed by Huawei Noah's Ark Lab
The Truth Is In There: Improving Reasoning in Language Models with Layer-Selective Rank Reduction
List of papers related to neural network quantization in recent AI conferences and journals.
Pytorch implementation of various Knowledge Distillation (KD) methods.
Awesome Knowledge-Distillation. 分类整理的知识蒸馏paper(2014-2021)。
[CVPR 2024 Highlight] Logit Standardization in Knowledge Distillation
KVQuant: Towards 10 Million Context Length LLM Inference with KV Cache Quantization
Pretrained language model and its related optimization techniques developed by Huawei Noah's Ark Lab.
A PyTorch implementation for exploring deep and shallow knowledge distillation (KD) experiments with flexibility
micronet, a model compression and deploy lib. compression: 1、quantization: quantization-aware-training(QAT), High-Bit(>2b)(DoReFa/Quantization and Training of Neural Networks for Efficient Integer-Ari...
[CVPR 2024] DeepCache: Accelerating Diffusion Models for Free
The Truth Is In There: Improving Reasoning in Language Models with Layer-Selective Rank Reduction
This is a collection of our research on efficient AI, covering hardware-aware NAS and model compression.
List of papers related to neural network quantization in recent AI conferences and journals.
This project is the official implementation of 'Basic Binary Convolution Unit for Binarized Image Restoration Network', ICLR2023
[ACL 2023] Code for paper “Tailoring Instructions to Student’s Learning Levels Boosts Knowledge Distillation”(https://arxiv.org/abs/2305.09651)
Resources of our survey paper "Enabling AI on Edges: Techniques, Applications and Challenges"
This project is the official implementation of our accepted ICML 2023 paper BiBench: Benchmarking and Analyzing Network Binarization.
a collection of computer vision projects&tools. 计算机视觉方向项目和工具集合。
[CVPR 2023] Towards Any Structural Pruning; LLMs / SAM / Diffusion / Transformers / YOLOv8 / CNNs
Efficient computing methods developed by Huawei Noah's Ark Lab
Gather research papers, corresponding codes (if having), reading notes and any other related materials about Hot🔥🔥🔥 fields in Computer Vision based on Deep Learning.
A list of papers, docs, codes about model quantization. This repo is aimed to provide the info for model quantization research, we are continuously improving the project. Welcome to PR the works (pape...
OTOv1-v3, NeurIPS, ICLR, TMLR, DNN Training, Compression, Structured Pruning, Erasing Operators, CNN, Diffusion, LLM