Trending repositories for topic transformers
Nexa SDK is a comprehensive toolkit for supporting ONNX and GGML models. It supports text generation, image generation, vision-language models (VLM), auto-speech-recognition (ASR), and text-to-speech ...
Unified Efficient Fine-Tuning of 100+ LLMs (ACL 2024)
🧑🏫 60+ Implementations/tutorials of deep learning papers with side-by-side notes 📝; including transformers (original, xl, switch, feedback, vit, ...), optimizers (adam, adabelief, sophia, ...), ga...
AI orchestration framework to build customizable, production-ready LLM applications. Connect components (models, vector DBs, file converters) to pipelines or agents that can interact with your data. W...
21 Lessons, Get Started Building with Generative AI 🔗 https://microsoft.github.io/generative-ai-for-beginners/
💡 All-in-one open-source embeddings database for semantic search, LLM orchestration and language model workflows
An Easy-to-use, Scalable and High-performance RLHF Framework (70B+ PPO Full Tuning & Iterative DPO & LoRA & RingAttention)
Implementation of Vision Transformer, a simple way to achieve SOTA in vision classification with only a single transformer encoder, in Pytorch
Implementation of π₀, the robotic foundation model architecture proposed by Physical Intelligence
State-of-the-art Machine Learning for the web. Run 🤗 Transformers directly in your browser, with no need for a server!
Chronos: Pretrained (Language) Models for Probabilistic Time Series Forecasting
An easy-to-use LLMs quantization package with user-friendly apis, based on GPTQ algorithm.
Pytorch implementation of Transfusion, "Predict the Next Token and Diffuse Images with One Multi-Modal Model", from MetaAI
[NeurIPS 2024 Oral][GPT beats diffusion🔥] [scaling laws in visual generation📈] Official impl. of "Visual Autoregressive Modeling: Scalable Image Generation via Next-Scale Prediction". An *ultra-simp...
RWKV is an RNN with transformer-level LLM performance. It can be directly trained like a GPT (parallelizable). So it's combining the best of RNN and transformer - great performance, fast inference, sa...
Implementation of π₀, the robotic foundation model architecture proposed by Physical Intelligence
Python package implementing transformers for pre processing steps for machine learning.
A Faster LayoutReader Model based on LayoutLMv3, Sort OCR bboxes to reading order.
Nexa SDK is a comprehensive toolkit for supporting ONNX and GGML models. It supports text generation, image generation, vision-language models (VLM), auto-speech-recognition (ASR), and text-to-speech ...
Code for "LayerSkip: Enabling Early Exit Inference and Self-Speculative Decoding", ACL 2024
This is the official implementation of the work presented at CVPR 2024, titled Multiple View Geometry Transformers for 3D Human Pose Estimation (MVGFormer).
The official code for "TEMPO: Prompt-based Generative Pre-trained Transformer for Time Series Forecasting (ICLR 2024)". TEMPO is one of the very first open source Time Series Foundation Models for fo...
Pytorch implementation of Transfusion, "Predict the Next Token and Diffuse Images with One Multi-Modal Model", from MetaAI
Retrieve, Read and LinK: Fast and Accurate Entity Linking and Relation Extraction on an Academic Budget (ACL 2024)
Quick implementation of nGPT, learning entirely on the hypersphere, from NvidiaAI
An Easy-to-use, Scalable and High-performance RLHF Framework (70B+ PPO Full Tuning & Iterative DPO & LoRA & RingAttention)
Working collection of papers, repos and models of transformer based language models trained or tuned for the Chemical domain, from natural language to chemical modeling and property prediction
State-of-the-art CLIP/SigLIP embedding models finetuned for the fashion domain. +57% increase in evaluation metrics vs FashionCLIP 2.0.
Official Implementation of SegFormer3D: an Efficient Transformer for 3D Medical Image Segmentation (CVPRW 2024)
Nexa SDK is a comprehensive toolkit for supporting ONNX and GGML models. It supports text generation, image generation, vision-language models (VLM), auto-speech-recognition (ASR), and text-to-speech ...
Unified Efficient Fine-Tuning of 100+ LLMs (ACL 2024)
🧑🏫 60+ Implementations/tutorials of deep learning papers with side-by-side notes 📝; including transformers (original, xl, switch, feedback, vit, ...), optimizers (adam, adabelief, sophia, ...), ga...
21 Lessons, Get Started Building with Generative AI 🔗 https://microsoft.github.io/generative-ai-for-beginners/
Implementation of Vision Transformer, a simple way to achieve SOTA in vision classification with only a single transformer encoder, in Pytorch
AI orchestration framework to build customizable, production-ready LLM applications. Connect components (models, vector DBs, file converters) to pipelines or agents that can interact with your data. W...
An Easy-to-use, Scalable and High-performance RLHF Framework (70B+ PPO Full Tuning & Iterative DPO & LoRA & RingAttention)
💡 All-in-one open-source embeddings database for semantic search, LLM orchestration and language model workflows
State-of-the-art Machine Learning for the web. Run 🤗 Transformers directly in your browser, with no need for a server!
A MLX port of FLUX based on the Huggingface Diffusers implementation.
Open source real-time translation app for Android that runs locally
Implementation of π₀, the robotic foundation model architecture proposed by Physical Intelligence
A minimal TensorFlow.js re-implementation of Karpathy's minGPT (Generative Pre-trained Transformer). The GPT model itself is <300 lines of code.
Nexa SDK is a comprehensive toolkit for supporting ONNX and GGML models. It supports text generation, image generation, vision-language models (VLM), auto-speech-recognition (ASR), and text-to-speech ...
A Faster LayoutReader Model based on LayoutLMv3, Sort OCR bboxes to reading order.
Code for "LayerSkip: Enabling Early Exit Inference and Self-Speculative Decoding", ACL 2024
Python package implementing transformers for pre processing steps for machine learning.
🔴 VERY LARGE AI TOOL LIST! 🔴 ⭐An extensive, ever-growing collection of AI tools and resources covering a wide range of applications in artificial intelligence and machine learning. ⭐ From educationa...
The official code for "TEMPO: Prompt-based Generative Pre-trained Transformer for Time Series Forecasting (ICLR 2024)". TEMPO is one of the very first open source Time Series Foundation Models for fo...
🔥 Aurora Series: A more efficient multimodal large language model series for video.
A MLX port of FLUX based on the Huggingface Diffusers implementation.
A radically simple, reliable, and high performance template to enable you to quickly get set up building multi-agent applications
Code for the paper "Aligning LLM Agents by Learning Latent Preference from User Edits".
Implementation of π₀, the robotic foundation model architecture proposed by Physical Intelligence
Implementation of the proposed Spline-Based Transformer from Disney Research
Implementation of LVSM, SOTA Large View Synthesis with Minimal 3d Inductive Bias, from Adobe Research
A minimal TensorFlow.js re-implementation of Karpathy's minGPT (Generative Pre-trained Transformer). The GPT model itself is <300 lines of code.
Unified Efficient Fine-Tuning of 100+ LLMs (ACL 2024)
Nexa SDK is a comprehensive toolkit for supporting ONNX and GGML models. It supports text generation, image generation, vision-language models (VLM), auto-speech-recognition (ASR), and text-to-speech ...
🧑🏫 60+ Implementations/tutorials of deep learning papers with side-by-side notes 📝; including transformers (original, xl, switch, feedback, vit, ...), optimizers (adam, adabelief, sophia, ...), ga...
21 Lessons, Get Started Building with Generative AI 🔗 https://microsoft.github.io/generative-ai-for-beginners/
State-of-the-art Machine Learning for the web. Run 🤗 Transformers directly in your browser, with no need for a server!
AI orchestration framework to build customizable, production-ready LLM applications. Connect components (models, vector DBs, file converters) to pipelines or agents that can interact with your data. W...
Implementation of Vision Transformer, a simple way to achieve SOTA in vision classification with only a single transformer encoder, in Pytorch
💡 All-in-one open-source embeddings database for semantic search, LLM orchestration and language model workflows
An Easy-to-use, Scalable and High-performance RLHF Framework (70B+ PPO Full Tuning & Iterative DPO & LoRA & RingAttention)
A MLX port of FLUX based on the Huggingface Diffusers implementation.
This repository contains demos I made with the Transformers library by HuggingFace.
Code for "LayerSkip: Enabling Early Exit Inference and Self-Speculative Decoding", ACL 2024
Implementation of π₀, the robotic foundation model architecture proposed by Physical Intelligence
Implementation of LVSM, SOTA Large View Synthesis with Minimal 3d Inductive Bias, from Adobe Research
Code for "LayerSkip: Enabling Early Exit Inference and Self-Speculative Decoding", ACL 2024
A minimal TensorFlow.js re-implementation of Karpathy's minGPT (Generative Pre-trained Transformer). The GPT model itself is <300 lines of code.
Implementation of the proposed Spline-Based Transformer from Disney Research
Quick implementation of nGPT, learning entirely on the hypersphere, from NvidiaAI
Nexa SDK is a comprehensive toolkit for supporting ONNX and GGML models. It supports text generation, image generation, vision-language models (VLM), auto-speech-recognition (ASR), and text-to-speech ...
🔥 Aurora Series: A more efficient multimodal large language model series for video.
🔴 VERY LARGE AI TOOL LIST! 🔴 ⭐An extensive, ever-growing collection of AI tools and resources covering a wide range of applications in artificial intelligence and machine learning. ⭐ From educationa...
Official implementation for "UniST: A Prompt-Empowered Universal Model for Urban Spatio-Temporal Prediction" (KDD 2024)
An open source implementation of LFMs from Liquid AI: Liquid Foundation Models
A MLX port of FLUX based on the Huggingface Diffusers implementation.
(Windows/Linux/MacOS) Local WebUI with neural network models (Text, Image, Video, 3D, Audio) on python (Gradio interface). Translated on 3 languages
[NeurIPS 2024 Oral][GPT beats diffusion🔥] [scaling laws in visual generation📈] Official impl. of "Visual Autoregressive Modeling: Scalable Image Generation via Next-Scale Prediction". An *ultra-simp...
Nexa SDK is a comprehensive toolkit for supporting ONNX and GGML models. It supports text generation, image generation, vision-language models (VLM), auto-speech-recognition (ASR), and text-to-speech ...
Chronos: Pretrained (Language) Models for Probabilistic Time Series Forecasting
Lumina-T2X is a unified framework for Text to Any Modality Generation
From anything to mesh like human artists. Official impl. of "MeshAnything: Artist-Created Mesh Generation with Autoregressive Transformers"
streamline the fine-tuning process for multimodal models: PaliGemma, Florence-2, and Qwen2-VL
Implementation of the training framework proposed in Self-Rewarding Language Model, from MetaAI
Lag-Llama: Towards Foundation Models for Probabilistic Time Series Forecasting
A curriculum for learning about foundation models, from scratch to the frontier
Unified Training of Universal Time Series Forecasting Transformers
Official PyTorch Implementation of MambaVision: A Hybrid Mamba-Transformer Vision Backbone
Implementation of MeshGPT, SOTA Mesh generation using Attention, in Pytorch
Pytorch implementation of Transfusion, "Predict the Next Token and Diffuse Images with One Multi-Modal Model", from MetaAI
The PyTorch implementation of Generative Pre-trained Transformers (GPTs) using Kolmogorov-Arnold Networks (KANs) for language modeling
This repository is a curated collection of the most exciting and influential CVPR 2024 papers. 🔥 [Paper + Code + Demo]
21 Lessons, Get Started Building with Generative AI 🔗 https://microsoft.github.io/generative-ai-for-beginners/
Unified Efficient Fine-Tuning of 100+ LLMs (ACL 2024)
🧑🏫 60+ Implementations/tutorials of deep learning papers with side-by-side notes 📝; including transformers (original, xl, switch, feedback, vit, ...), optimizers (adam, adabelief, sophia, ...), ga...
State-of-the-art Machine Learning for the web. Run 🤗 Transformers directly in your browser, with no need for a server!
Open source real-time translation app for Android that runs locally
AI orchestration framework to build customizable, production-ready LLM applications. Connect components (models, vector DBs, file converters) to pipelines or agents that can interact with your data. W...
🤗 PEFT: State-of-the-art Parameter-Efficient Fine-Tuning.
Implementation of Vision Transformer, a simple way to achieve SOTA in vision classification with only a single transformer encoder, in Pytorch
Superduper: Build end-to-end AI applications and agent workflows on your existing data infrastructure and preferred tools - without migrating your data.
[NeurIPS 2024 Oral][GPT beats diffusion🔥] [scaling laws in visual generation📈] Official impl. of "Visual Autoregressive Modeling: Scalable Image Generation via Next-Scale Prediction". An *ultra-simp...
💡 All-in-one open-source embeddings database for semantic search, LLM orchestration and language model workflows
Nexa SDK is a comprehensive toolkit for supporting ONNX and GGML models. It supports text generation, image generation, vision-language models (VLM), auto-speech-recognition (ASR), and text-to-speech ...
This repository contains demos I made with the Transformers library by HuggingFace.
🦖 𝗟𝗲𝗮𝗿𝗻 about 𝗟𝗟𝗠𝘀, 𝗟𝗟𝗠𝗢𝗽𝘀, and 𝘃𝗲𝗰𝘁𝗼𝗿 𝗗𝗕𝘀 for free by designing, training, and deploying a real-time financial advisor LLM system ~ 𝘴𝘰𝘶𝘳𝘤𝘦 𝘤𝘰𝘥𝘦 + 𝘷𝘪𝘥𝘦𝘰 & 𝘳𝘦�...
Chronos: Pretrained (Language) Models for Probabilistic Time Series Forecasting
RWKV is an RNN with transformer-level LLM performance. It can be directly trained like a GPT (parallelizable). So it's combining the best of RNN and transformer - great performance, fast inference, sa...
A MLX port of FLUX based on the Huggingface Diffusers implementation.
【三年面试五年模拟】算法工程师秘籍。涵盖AIGC、传统深度学习、自动驾驶、机器学习、计算机视觉、自然语言处理、SLAM、具身智能、元宇宙、AGI等AI行业面试笔试经验与干货知识。
From anything to mesh like human artists. Official impl. of "MeshAnything: Artist-Created Mesh Generation with Autoregressive Transformers"
streamline the fine-tuning process for multimodal models: PaliGemma, Florence-2, and Qwen2-VL
A collection of 🤗 Transformers.js demos and example applications
This repository is a curated collection of the most exciting and influential CVPR 2024 papers. 🔥 [Paper + Code + Demo]
From anything to mesh like human artists. Official impl. of "MeshAnything V2: Artist-Created Mesh Generation With Adjacent Mesh Tokenization"
RESTai is an AIaaS (AI as a Service) open-source platform. Built on top of LlamaIndex & Langchain. Supports any public LLM supported by LlamaIndex and any local LLM suported by Ollama/vLLM/etc. Precis...
A library for easily merging multiple LLM experts, and efficiently train the merged LLM.
Open source real-time translation app for Android that runs locally
A novel implementation of fusing ViT with Mamba into a fast, agile, and high performance Multi-Modal Model. Powered by Zeta, the simplest AI framework ever.
Chronos: Pretrained (Language) Models for Probabilistic Time Series Forecasting
The Truth Is In There: Improving Reasoning in Language Models with Layer-Selective Rank Reduction
MOMENT: A Family of Open Time-series Foundation Models
Here lies the resources and topics necessary for the role of Data Scientist and Machine Learning
Pytorch implementation of Transfusion, "Predict the Next Token and Diffuse Images with One Multi-Modal Model", from MetaAI
A curriculum for learning about foundation models, from scratch to the frontier
PyTorch Implementation of Jamba: "Jamba: A Hybrid Transformer-Mamba Language Model"
Implementation of the training framework proposed in Self-Rewarding Language Model, from MetaAI