Trending repositories for topic transformers
18 Lessons, Get Started Building with Generative AI 🔗 https://microsoft.github.io/generative-ai-for-beginners/
🧑🏫 60 Implementations/tutorials of deep learning papers with side-by-side notes 📝; including transformers (original, xl, switch, feedback, vit, ...), optimizers (adam, adabelief, sophia, ...), gan...
This repository contains demos I made with the Transformers library by HuggingFace.
Official PyTorch implementation of the CVPR 2024 paper: State Space Models for Event Cameras.
:mag: LLM orchestration framework to build customizable, production-ready LLM applications. Connect components (models, vector DBs, file converters) to pipelines or agents that can interact with your ...
An Easy-to-use, Scalable and High-performance RLHF Framework (Support 70B+ full tuning & LoRA & Mixtral & KTO)
Lumina-T2X is a unified framework for Text to Any Modality Generation
Robust recipes to align language models with human and AI preferences
State-of-the-art Machine Learning for the web. Run 🤗 Transformers directly in your browser, with no need for a server!
Multi-LoRA inference server that scales to 1000s of fine-tuned LLMs
Chronos: Pretrained (Language) Models for Probabilistic Time Series Forecasting
💡 All-in-one open-source embeddings database for semantic search, LLM orchestration and language model workflows
Chinese version of CLIP which achieves Chinese cross-modal retrieval and representation generation.
Official PyTorch implementation of the CVPR 2024 paper: State Space Models for Event Cameras.
BabyGPT: Build Your Own GPT Large Language Model from Scratch Pre-Training Generative Transformer Models: Building GPT from Scratch with a Step-by-Step Guide to Generative AI in PyTorch and Python
⚡️SwanLab: your ML experiment notebook. 你的AI实验笔记本,跟踪与可视化你的机器学习全流程
Code for "Can We Scale Transformers to Predict Parameters of Diverse ImageNet Models?" [ICML 2023]
MOMENT: A Family of Open Time-series Foundation Models
Sparse and discrete interpretability tool for neural networks
An Easy-to-use, Scalable and High-performance RLHF Framework (Support 70B+ full tuning & LoRA & Mixtral & KTO)
🔥🔥🔥Official Codebase of "DiT-3D: Exploring Plain Diffusion Transformers for 3D Shape Generation"
Lumina-T2X is a unified framework for Text to Any Modality Generation
A Simplified PyTorch Implementation of Vision Transformer (ViT)
An AI-powered natural language & reverse Image Search Engine powered by CLIP & qdrant.
AI-First Process Automation with Large ([Language (LLMs) / Action (LAMs) / Multimodal (LMMs)] / Visual Language (VLMs)) Models
Transformer models from BERT to GPT-4, environments from Hugging Face to OpenAI. Fine-tuning, training, and prompt engineering examples. A bonus section with ChatGPT, GPT-3.5-turbo, GPT-4, and DALL-E ...
Unified Training of Universal Time Series Forecasting Transformers
Unofficial PyTorch/🤗Transformers(Gemma/Llama3) implementation of Leave No Context Behind: Efficient Infinite Context Transformers with Infini-attention
Implementation of Hourglass Transformer, in Pytorch, from Google and OpenAI
18 Lessons, Get Started Building with Generative AI 🔗 https://microsoft.github.io/generative-ai-for-beginners/
This repository contains demos I made with the Transformers library by HuggingFace.
🧑🏫 60 Implementations/tutorials of deep learning papers with side-by-side notes 📝; including transformers (original, xl, switch, feedback, vit, ...), optimizers (adam, adabelief, sophia, ...), gan...
:mag: LLM orchestration framework to build customizable, production-ready LLM applications. Connect components (models, vector DBs, file converters) to pipelines or agents that can interact with your ...
A Framework of Small-scale Large Multimodal Models
An Easy-to-use, Scalable and High-performance RLHF Framework (Support 70B+ full tuning & LoRA & Mixtral & KTO)
Swift native on-device speech recognition with Whisper for Apple Silicon
Implementation of Vision Transformer, a simple way to achieve SOTA in vision classification with only a single transformer encoder, in Pytorch
State-of-the-art Machine Learning for the web. Run 🤗 Transformers directly in your browser, with no need for a server!
Lumina-T2X is a unified framework for Text to Any Modality Generation
AI-First Process Automation with Large ([Language (LLMs) / Action (LAMs) / Multimodal (LMMs)] / Visual Language (VLMs)) Models
Robust recipes to align language models with human and AI preferences
OpenChat: Advancing Open-source Language Models with Imperfect Data
Official PyTorch implementation of the CVPR 2024 paper: State Space Models for Event Cameras.
💡 All-in-one open-source embeddings database for semantic search, LLM orchestration and language model workflows
Official PyTorch implementation of the CVPR 2024 paper: State Space Models for Event Cameras.
A Framework of Small-scale Large Multimodal Models
An AI-powered natural language & reverse Image Search Engine powered by CLIP & qdrant.
AI-First Process Automation with Large ([Language (LLMs) / Action (LAMs) / Multimodal (LMMs)] / Visual Language (VLMs)) Models
⚡️SwanLab: your ML experiment notebook. 你的AI实验笔记本,跟踪与可视化你的机器学习全流程
An Easy-to-use, Scalable and High-performance RLHF Framework (Support 70B+ full tuning & LoRA & Mixtral & KTO)
MOMENT: A Family of Open Time-series Foundation Models
BabyGPT: Build Your Own GPT Large Language Model from Scratch Pre-Training Generative Transformer Models: Building GPT from Scratch with a Step-by-Step Guide to Generative AI in PyTorch and Python
Segment Anything Model for large-scale, vectorized road network extraction from aerial imagery. CVPRW 2024
Lumina-T2X is a unified framework for Text to Any Modality Generation
Transformer models from BERT to GPT-4, environments from Hugging Face to OpenAI. Fine-tuning, training, and prompt engineering examples. A bonus section with ChatGPT, GPT-3.5-turbo, GPT-4, and DALL-E ...
Unified Training of Universal Time Series Forecasting Transformers
[CVPR 2022] Joint hand motion and interaction hotspots prediction from egocentric videos
Swift native on-device speech recognition with Whisper for Apple Silicon
Explore a comprehensive collection of resources, tutorials, papers, tools, and best practices for fine-tuning Large Language Models (LLMs). Perfect for ML practitioners and researchers!
Code for "Can We Scale Transformers to Predict Parameters of Diverse ImageNet Models?" [ICML 2023]
The PyTorch implementation of Generative Pre-trained Transformers (GPTs) using Kolmogorov-Arnold Networks (KANs) for language modeling
MOMENT: A Family of Open Time-series Foundation Models
Exploration into the proposed "Self Reasoning Tokens" by Felipe Bonetto
18 Lessons, Get Started Building with Generative AI 🔗 https://microsoft.github.io/generative-ai-for-beginners/
Lumina-T2X is a unified framework for Text to Any Modality Generation
🧑🏫 60 Implementations/tutorials of deep learning papers with side-by-side notes 📝; including transformers (original, xl, switch, feedback, vit, ...), optimizers (adam, adabelief, sophia, ...), gan...
The PyTorch implementation of Generative Pre-trained Transformers (GPTs) using Kolmogorov-Arnold Networks (KANs) for language modeling
This repository contains demos I made with the Transformers library by HuggingFace.
:mag: LLM orchestration framework to build customizable, production-ready LLM applications. Connect components (models, vector DBs, file converters) to pipelines or agents that can interact with your ...
[GPT beats diffusion🔥] [scaling laws in visual generation📈] Official impl. of "Visual Autoregressive Modeling: Scalable Image Generation via Next-Scale Prediction". An *ultra-simple, user-friendly y...
State-of-the-art Machine Learning for the web. Run 🤗 Transformers directly in your browser, with no need for a server!
A curriculum for learning about foundation models, from scratch to the frontier
Implementation of Vision Transformer, a simple way to achieve SOTA in vision classification with only a single transformer encoder, in Pytorch
An Easy-to-use, Scalable and High-performance RLHF Framework (Support 70B+ full tuning & LoRA & Mixtral & KTO)
Chronos: Pretrained (Language) Models for Probabilistic Time Series Forecasting
LLM-PowerHouse: Unleash LLMs' potential through curated tutorials, best practices, and ready-to-use code for custom training and inferencing.
Swift native on-device speech recognition with Whisper for Apple Silicon
MOMENT: A Family of Open Time-series Foundation Models
Lumina-T2X is a unified framework for Text to Any Modality Generation
Exploration into the proposed "Self Reasoning Tokens" by Felipe Bonetto
Official PyTorch implementation of the CVPR 2024 paper: State Space Models for Event Cameras.
Code for NOLA, an implementation of "nola: Compressing LoRA using Linear Combination of Random Basis"
A Faster LayoutReader Model based on LayoutLMv3, Sort OCR bboxes to reading order.
A curriculum for learning about foundation models, from scratch to the frontier
LLM-PowerHouse: Unleash LLMs' potential through curated tutorials, best practices, and ready-to-use code for custom training and inferencing.
A Framework of Small-scale Large Multimodal Models
This the repo for the paper tiltled "AgileFormer: Spatially Agile Transformer UNet for Medical Image Segmentation"
An AI-powered natural language & reverse Image Search Engine powered by CLIP & qdrant.
Segment Anything Model for large-scale, vectorized road network extraction from aerial imagery. CVPRW 2024
Implementation of the LongRoPE: Extending LLM Context Window Beyond 2 Million Tokens Paper
AI-First Process Automation with Large ([Language (LLMs) / Action (LAMs) / Multimodal (LMMs)] / Visual Language (VLMs)) Models
[CVPR'24] Multimodal Pathway: Improve Transformers with Irrelevant Data from Other Modalities
A library for easily merging multiple LLM experts, and efficiently train the merged LLM.
Transformers PHP is a toolkit for PHP developers to add machine learning magic to their projects easily.
Unified Training of Universal Time Series Forecasting Transformers
18 Lessons, Get Started Building with Generative AI 🔗 https://microsoft.github.io/generative-ai-for-beginners/
Robust recipes to align language models with human and AI preferences
[GPT beats diffusion🔥] [scaling laws in visual generation📈] Official impl. of "Visual Autoregressive Modeling: Scalable Image Generation via Next-Scale Prediction". An *ultra-simple, user-friendly y...
🦖 𝗟𝗲𝗮𝗿𝗻 about 𝗟𝗟𝗠𝘀, 𝗟𝗟𝗠𝗢𝗽𝘀, and 𝘃𝗲𝗰𝘁𝗼𝗿 𝗗𝗕𝘀 for free by designing, training, and deploying a real-time financial advisor LLM system ~ 𝘴𝘰𝘶𝘳𝘤𝘦 𝘤𝘰𝘥𝘦 + 𝘷𝘪𝘥𝘦𝘰 & 𝘳𝘦�...
Swift native on-device speech recognition with Whisper for Apple Silicon
Chronos: Pretrained (Language) Models for Probabilistic Time Series Forecasting
Multi-LoRA inference server that scales to 1000s of fine-tuned LLMs
An Easy-to-use, Scalable and High-performance RLHF Framework (Support 70B+ full tuning & LoRA & Mixtral & KTO)
Implementation of the training framework proposed in Self-Rewarding Language Model, from MetaAI
Lumina-T2X is a unified framework for Text to Any Modality Generation
Lag-Llama: Towards Foundation Models for Probabilistic Time Series Forecasting
[ICLR'24 spotlight] Chinese and English Multimodal Large Model Series (Chat and Paint) | 基于CPM基础模型的中英双语多模态大模型系列
Running Llama 2 and other Open-Source LLMs on CPU Inference Locally for Document Q&A
Official implementation of paper "MiniGPT-5: Interleaved Vision-and-Language Generation via Generative Vokens"
DataDreamer: Prompt. Generate Synthetic Data. Train & Align Models. 🤖💤
18 Lessons, Get Started Building with Generative AI 🔗 https://microsoft.github.io/generative-ai-for-beginners/
🧑🏫 60 Implementations/tutorials of deep learning papers with side-by-side notes 📝; including transformers (original, xl, switch, feedback, vit, ...), optimizers (adam, adabelief, sophia, ...), gan...
🤗 PEFT: State-of-the-art Parameter-Efficient Fine-Tuning.
State-of-the-art Machine Learning for the web. Run 🤗 Transformers directly in your browser, with no need for a server!
:mag: LLM orchestration framework to build customizable, production-ready LLM applications. Connect components (models, vector DBs, file converters) to pipelines or agents that can interact with your ...
OpenChat: Advancing Open-source Language Models with Imperfect Data
🔮 SuperDuperDB: Bring AI to your database! Build, deploy and manage any AI application directly with your existing data infrastructure, without moving your data. Including streaming inference, scalab...
Implementation of Vision Transformer, a simple way to achieve SOTA in vision classification with only a single transformer encoder, in Pytorch
Robust recipes to align language models with human and AI preferences
RWKV is an RNN with transformer-level LLM performance. It can be directly trained like a GPT (parallelizable). So it's combining the best of RNN and transformer - great performance, fast inference, sa...
[GPT beats diffusion🔥] [scaling laws in visual generation📈] Official impl. of "Visual Autoregressive Modeling: Scalable Image Generation via Next-Scale Prediction". An *ultra-simple, user-friendly y...
An easy-to-use LLMs quantization package with user-friendly apis, based on GPTQ algorithm.
This repository contains demos I made with the Transformers library by HuggingFace.
💡 All-in-one open-source embeddings database for semantic search, LLM orchestration and language model workflows
LightGlue: Local Feature Matching at Light Speed (ICCV 2023)
Fine-tuning ChatGLM-6B with PEFT | 基于 PEFT 的高效 ChatGLM 微调
18 Lessons, Get Started Building with Generative AI 🔗 https://microsoft.github.io/generative-ai-for-beginners/
🔮 SuperDuperDB: Bring AI to your database! Build, deploy and manage any AI application directly with your existing data infrastructure, without moving your data. Including streaming inference, scalab...
An Easy-to-use, Scalable and High-performance RLHF Framework (Support 70B+ full tuning & LoRA & Mixtral & KTO)
OpenChat: Advancing Open-source Language Models with Imperfect Data
🤖 A PyTorch library of curated Transformer models and their composable components
Extend existing LLMs way beyond the original training length with constant memory usage, without retraining
Official implementation of paper "MiniGPT-5: Interleaved Vision-and-Language Generation via Generative Vokens"
A library for easily merging multiple LLM experts, and efficiently train the merged LLM.
A novel implementation of fusing ViT with Mamba into a fast, agile, and high performance Multi-Modal Model. Powered by Zeta, the simplest AI framework ever.
🔥🔥🔥Official Codebase of "DiT-3D: Exploring Plain Diffusion Transformers for 3D Shape Generation"
Implementation of Band Split Roformer, SOTA Attention network for music source separation out of ByteDance AI Labs
The Truth Is In There: Improving Reasoning in Language Models with Layer-Selective Rank Reduction
Explorations into some recent techniques surrounding speculative decoding