Trending repositories for topic transformers
Unified Efficient Fine-Tuning of 100+ LLMs (ACL 2024)
🧑🏫 60+ Implementations/tutorials of deep learning papers with side-by-side notes 📝; including transformers (original, xl, switch, feedback, vit, ...), optimizers (adam, adabelief, sophia, ...), ga...
An Easy-to-use, Scalable and High-performance RLHF Framework (70B+ PPO Full Tuning & Iterative DPO & LoRA & RingAttention & RFT)
A collection of 🤗 Transformers.js demos and example applications
[NeurIPS 2024 Oral][GPT beats diffusion🔥] [scaling laws in visual generation📈] Official impl. of "Visual Autoregressive Modeling: Scalable Image Generation via Next-Scale Prediction". An *ultra-simp...
21 Lessons, Get Started Building with Generative AI 🔗 https://microsoft.github.io/generative-ai-for-beginners/
AI orchestration framework to build customizable, production-ready LLM applications. Connect components (models, vector DBs, file converters) to pipelines or agents that can interact with your data. W...
State-of-the-art Machine Learning for the web. Run 🤗 Transformers directly in your browser, with no need for a server!
Implementation of Vision Transformer, a simple way to achieve SOTA in vision classification with only a single transformer encoder, in Pytorch
💡 All-in-one open-source embeddings database for semantic search, LLM orchestration and language model workflows
RWKV is an RNN with transformer-level LLM performance. It can be directly trained like a GPT (parallelizable). So it's combining the best of RNN and transformer - great performance, fast inference, sa...
Nexa SDK is a comprehensive toolkit for supporting GGML and ONNX models. It supports text generation, image generation, vision-language models (VLM), Audio Language Model, auto-speech-recognition (ASR...
Infinity ∞ : Scaling Bitwise AutoRegressive Modeling for High-Resolution Image Synthesis
Code to train and evaluate Neural Attention Memory Models to obtain universally-applicable memory systems for transformers.
A collection of 🤗 Transformers.js demos and example applications
Implementation of GotenNet, new SOTA 3d equivariant transformer, in Pytorch
Code to train and evaluate Neural Attention Memory Models to obtain universally-applicable memory systems for transformers.
Implementation of 🥥 Coconut, Chain of Continuous Thought, in Pytorch
Infinity ∞ : Scaling Bitwise AutoRegressive Modeling for High-Resolution Image Synthesis
📚 Text Classification with LoRA (Low-Rank Adaptation) of Language Models - Efficiently fine-tune large language models for text classification tasks using the Stanford Sentiment Treebank (SST-2) data...
The official code for "TEMPO: Prompt-based Generative Pre-trained Transformer for Time Series Forecasting (ICLR 2024)". TEMPO is one of the very first open source Time Series Foundation Models for fo...
Synthesizing Graphics Programs for Scientific Figures and Sketches with TikZ
An Easy-to-use, Scalable and High-performance RLHF Framework (70B+ PPO Full Tuning & Iterative DPO & LoRA & RingAttention & RFT)
A project that optimizes OWL-ViT for real-time inference with NVIDIA TensorRT.
⚡️SwanLab: your ML experiment notebook. 你的AI实验笔记本,日志记录与可视化AI训练全流程。
SELFormer: Molecular Representation Learning via SELFIES Language Models
A performant high-throughput CPU-based API for Meta's No Language Left Behind (NLLB) using CTranslate2, hosted on Hugging Face Spaces.
Unified Efficient Fine-Tuning of 100+ LLMs (ACL 2024)
An Easy-to-use, Scalable and High-performance RLHF Framework (70B+ PPO Full Tuning & Iterative DPO & LoRA & RingAttention & RFT)
[NeurIPS 2024 Oral][GPT beats diffusion🔥] [scaling laws in visual generation📈] Official impl. of "Visual Autoregressive Modeling: Scalable Image Generation via Next-Scale Prediction". An *ultra-simp...
🧑🏫 60+ Implementations/tutorials of deep learning papers with side-by-side notes 📝; including transformers (original, xl, switch, feedback, vit, ...), optimizers (adam, adabelief, sophia, ...), ga...
Code to train and evaluate Neural Attention Memory Models to obtain universally-applicable memory systems for transformers.
21 Lessons, Get Started Building with Generative AI 🔗 https://microsoft.github.io/generative-ai-for-beginners/
Nexa SDK is a comprehensive toolkit for supporting GGML and ONNX models. It supports text generation, image generation, vision-language models (VLM), Audio Language Model, auto-speech-recognition (ASR...
A collection of 🤗 Transformers.js demos and example applications
AI orchestration framework to build customizable, production-ready LLM applications. Connect components (models, vector DBs, file converters) to pipelines or agents that can interact with your data. W...
Implementation of Vision Transformer, a simple way to achieve SOTA in vision classification with only a single transformer encoder, in Pytorch
State-of-the-art Machine Learning for the web. Run 🤗 Transformers directly in your browser, with no need for a server!
💡 All-in-one open-source embeddings database for semantic search, LLM orchestration and language model workflows
Infinity ∞ : Scaling Bitwise AutoRegressive Modeling for High-Resolution Image Synthesis
RWKV is an RNN with transformer-level LLM performance. It can be directly trained like a GPT (parallelizable). So it's combining the best of RNN and transformer - great performance, fast inference, sa...
This repository contains demos I made with the Transformers library by HuggingFace.
Code to train and evaluate Neural Attention Memory Models to obtain universally-applicable memory systems for transformers.
Implementation of GotenNet, new SOTA 3d equivariant transformer, in Pytorch
Implementation of 🥥 Coconut, Chain of Continuous Thought, in Pytorch
A collection of 🤗 Transformers.js demos and example applications
Infinity ∞ : Scaling Bitwise AutoRegressive Modeling for High-Resolution Image Synthesis
Run local LLM from Huggingface in React-Native or Expo using onnxruntime.
Run Vision LLMs, TTS and STT APIs. Website and API for https://text-generator.io
An Easy-to-use, Scalable and High-performance RLHF Framework (70B+ PPO Full Tuning & Iterative DPO & LoRA & RingAttention & RFT)
Implementation of π₀, the robotic foundation model architecture proposed by Physical Intelligence
📚 Text Classification with LoRA (Low-Rank Adaptation) of Language Models - Efficiently fine-tune large language models for text classification tasks using the Stanford Sentiment Treebank (SST-2) data...
Official Implementation of SegFormer3D: an Efficient Transformer for 3D Medical Image Segmentation (CVPRW 2024)
Repository for CARTE: Context-Aware Representation of Table Entries
CoDe: Collaborative Decoding Makes Visual Auto-Regressive Modeling Efficient
This repo contains my work & The code base for this Deep Learning Specialization offered by deeplearning.AI
The official code for "TEMPO: Prompt-based Generative Pre-trained Transformer for Time Series Forecasting (ICLR 2024)". TEMPO is one of the very first open source Time Series Foundation Models for fo...
Infinity ∞ : Scaling Bitwise AutoRegressive Modeling for High-Resolution Image Synthesis
Implementation of 🥥 Coconut, Chain of Continuous Thought, in Pytorch
Implementation of GotenNet, new SOTA 3d equivariant transformer, in Pytorch
[NeurIPS 2024 Oral][GPT beats diffusion🔥] [scaling laws in visual generation📈] Official impl. of "Visual Autoregressive Modeling: Scalable Image Generation via Next-Scale Prediction". An *ultra-simp...
Unified Efficient Fine-Tuning of 100+ LLMs (ACL 2024)
Nexa SDK is a comprehensive toolkit for supporting GGML and ONNX models. It supports text generation, image generation, vision-language models (VLM), Audio Language Model, auto-speech-recognition (ASR...
🧑🏫 60+ Implementations/tutorials of deep learning papers with side-by-side notes 📝; including transformers (original, xl, switch, feedback, vit, ...), optimizers (adam, adabelief, sophia, ...), ga...
21 Lessons, Get Started Building with Generative AI 🔗 https://microsoft.github.io/generative-ai-for-beginners/
An Easy-to-use, Scalable and High-performance RLHF Framework (70B+ PPO Full Tuning & Iterative DPO & LoRA & RingAttention & RFT)
Implementation of Vision Transformer, a simple way to achieve SOTA in vision classification with only a single transformer encoder, in Pytorch
AI orchestration framework to build customizable, production-ready LLM applications. Connect components (models, vector DBs, file converters) to pipelines or agents that can interact with your data. W...
Infinity ∞ : Scaling Bitwise AutoRegressive Modeling for High-Resolution Image Synthesis
💡 All-in-one open-source embeddings database for semantic search, LLM orchestration and language model workflows
State-of-the-art Machine Learning for the web. Run 🤗 Transformers directly in your browser, with no need for a server!
Code to train and evaluate Neural Attention Memory Models to obtain universally-applicable memory systems for transformers.
This repository contains demos I made with the Transformers library by HuggingFace.
Implementation of GotenNet, new SOTA 3d equivariant transformer, in Pytorch
Code to train and evaluate Neural Attention Memory Models to obtain universally-applicable memory systems for transformers.
Implementation of 🥥 Coconut, Chain of Continuous Thought, in Pytorch
This repository contains MAPF-GPT, a deep learning-based model for solving MAPF problems. Trained with imitation learning on trajectories produced by LaCAM, it generates collision-free paths under par...
Run local LLM from Huggingface in React-Native or Expo using onnxruntime.
Implementation of π₀, the robotic foundation model architecture proposed by Physical Intelligence
[NeurIPS 2024 Oral][GPT beats diffusion🔥] [scaling laws in visual generation📈] Official impl. of "Visual Autoregressive Modeling: Scalable Image Generation via Next-Scale Prediction". An *ultra-simp...
A collection of 🤗 Transformers.js demos and example applications
Explore a comprehensive collection of resources, tutorials, papers, tools, and best practices for fine-tuning Large Language Models (LLMs). Perfect for ML practitioners and researchers!
A Faster LayoutReader Model based on LayoutLMv3, Sort OCR bboxes to reading order.
Nexa SDK is a comprehensive toolkit for supporting GGML and ONNX models. It supports text generation, image generation, vision-language models (VLM), Audio Language Model, auto-speech-recognition (ASR...
[NeurIPS 2024 Oral][GPT beats diffusion🔥] [scaling laws in visual generation📈] Official impl. of "Visual Autoregressive Modeling: Scalable Image Generation via Next-Scale Prediction". An *ultra-simp...
Nexa SDK is a comprehensive toolkit for supporting GGML and ONNX models. It supports text generation, image generation, vision-language models (VLM), Audio Language Model, auto-speech-recognition (ASR...
Chronos: Pretrained Models for Probabilistic Time Series Forecasting
Lumina-T2X is a unified framework for Text to Any Modality Generation
From anything to mesh like human artists. Official impl. of "MeshAnything: Artist-Created Mesh Generation with Autoregressive Transformers"
Implementation of the training framework proposed in Self-Rewarding Language Model, from MetaAI
Implementation of Alphafold 3 from Google Deepmind in Pytorch
Lag-Llama: Towards Foundation Models for Probabilistic Time Series Forecasting
A curriculum for learning about foundation models, from scratch to the frontier
Unified Training of Universal Time Series Forecasting Transformers
Official PyTorch Implementation of MambaVision: A Hybrid Mamba-Transformer Vision Backbone
Pytorch implementation of Transfusion, "Predict the Next Token and Diffuse Images with One Multi-Modal Model", from MetaAI
The PyTorch implementation of Generative Pre-trained Transformers (GPTs) using Kolmogorov-Arnold Networks (KANs) for language modeling
From anything to mesh like human artists. Official impl. of "MeshAnything V2: Artist-Created Mesh Generation With Adjacent Mesh Tokenization"
21 Lessons, Get Started Building with Generative AI 🔗 https://microsoft.github.io/generative-ai-for-beginners/
Unified Efficient Fine-Tuning of 100+ LLMs (ACL 2024)
🧑🏫 60+ Implementations/tutorials of deep learning papers with side-by-side notes 📝; including transformers (original, xl, switch, feedback, vit, ...), optimizers (adam, adabelief, sophia, ...), ga...
State-of-the-art Machine Learning for the web. Run 🤗 Transformers directly in your browser, with no need for a server!
Open source real-time translation app for Android that runs locally
[NeurIPS 2024 Oral][GPT beats diffusion🔥] [scaling laws in visual generation📈] Official impl. of "Visual Autoregressive Modeling: Scalable Image Generation via Next-Scale Prediction". An *ultra-simp...
AI orchestration framework to build customizable, production-ready LLM applications. Connect components (models, vector DBs, file converters) to pipelines or agents that can interact with your data. W...
🤗 PEFT: State-of-the-art Parameter-Efficient Fine-Tuning.
Nexa SDK is a comprehensive toolkit for supporting GGML and ONNX models. It supports text generation, image generation, vision-language models (VLM), Audio Language Model, auto-speech-recognition (ASR...
Implementation of Vision Transformer, a simple way to achieve SOTA in vision classification with only a single transformer encoder, in Pytorch
💡 All-in-one open-source embeddings database for semantic search, LLM orchestration and language model workflows
This repository contains demos I made with the Transformers library by HuggingFace.
Chronos: Pretrained Models for Probabilistic Time Series Forecasting
An Easy-to-use, Scalable and High-performance RLHF Framework (70B+ PPO Full Tuning & Iterative DPO & LoRA & RingAttention & RFT)
OpenVINO™ is an open-source toolkit for optimizing and deploying AI inference
RWKV is an RNN with transformer-level LLM performance. It can be directly trained like a GPT (parallelizable). So it's combining the best of RNN and transformer - great performance, fast inference, sa...
A MLX port of FLUX based on the Huggingface Diffusers implementation.
From anything to mesh like human artists. Official impl. of "MeshAnything: Artist-Created Mesh Generation with Autoregressive Transformers"
A collection of 🤗 Transformers.js demos and example applications
From anything to mesh like human artists. Official impl. of "MeshAnything V2: Artist-Created Mesh Generation With Adjacent Mesh Tokenization"
This repository is a curated collection of the most exciting and influential CVPR 2024 papers. 🔥 [Paper + Code + Demo]
A library for easily merging multiple LLM experts, and efficiently train the merged LLM.
【三年面试五年模拟】AI算法工程师面试秘籍。涵盖AIGC、传统深度学习、自动驾驶、机器学习、计算机视觉、自然语言处理、SLAM、具身智能、元宇宙、AGI等AI行业面试笔试经验与干货知识。
Open source real-time translation app for Android that runs locally
⚡️SwanLab: your ML experiment notebook. 你的AI实验笔记本,日志记录与可视化AI训练全流程。
Chronos: Pretrained Models for Probabilistic Time Series Forecasting
A novel implementation of fusing ViT with Mamba into a fast, agile, and high performance Multi-Modal Model. Powered by Zeta, the simplest AI framework ever.
The Truth Is In There: Improving Reasoning in Language Models with Layer-Selective Rank Reduction
MOMENT: A Family of Open Time-series Foundation Models
Explore a comprehensive collection of resources, tutorials, papers, tools, and best practices for fine-tuning Large Language Models (LLMs). Perfect for ML practitioners and researchers!
Pytorch implementation of Transfusion, "Predict the Next Token and Diffuse Images with One Multi-Modal Model", from MetaAI
Here lies the resources and topics necessary for the role of Data Scientist and Machine Learning
A curriculum for learning about foundation models, from scratch to the frontier