Trending repositories for topic transformers
Unified Efficient Fine-Tuning of 100+ LLMs (ACL 2024)
Nexa SDK is a comprehensive toolkit for supporting ONNX and GGML models. It supports text generation, image generation, vision-language models (VLM), auto-speech-recognition (ASR), and text-to-speech ...
21 Lessons, Get Started Building with Generative AI 🔗 https://microsoft.github.io/generative-ai-for-beginners/
🧑🏫 60+ Implementations/tutorials of deep learning papers with side-by-side notes 📝; including transformers (original, xl, switch, feedback, vit, ...), optimizers (adam, adabelief, sophia, ...), ga...
:mag: AI orchestration framework to build customizable, production-ready LLM applications. Connect components (models, vector DBs, file converters) to pipelines or agents that can interact with your d...
A MLX port of FLUX based on the Huggingface Diffusers implementation.
State-of-the-art Machine Learning for the web. Run 🤗 Transformers directly in your browser, with no need for a server!
Implementation of Vision Transformer, a simple way to achieve SOTA in vision classification with only a single transformer encoder, in Pytorch
Official implementation for "UniST: A Prompt-Empowered Universal Model for Urban Spatio-Temporal Prediction" (KDD 2024)
This repository contains demos I made with the Transformers library by HuggingFace.
💡 All-in-one open-source embeddings database for semantic search, LLM orchestration and language model workflows
Implementation of π₀, the robotic foundation model architecture proposed by Physical Intelligence
Semantic segmentation models with 500+ pretrained convolutional and transformer-based backbones.
[NeurIPS 2024 Oral][GPT beats diffusion🔥] [scaling laws in visual generation📈] Official impl. of "Visual Autoregressive Modeling: Scalable Image Generation via Next-Scale Prediction". An *ultra-simp...
Open-source offline translation library written in Python
Implementation of π₀, the robotic foundation model architecture proposed by Physical Intelligence
Official implementation for "UniST: A Prompt-Empowered Universal Model for Urban Spatio-Temporal Prediction" (KDD 2024)
A MLX port of FLUX based on the Huggingface Diffusers implementation.
Nexa SDK is a comprehensive toolkit for supporting ONNX and GGML models. It supports text generation, image generation, vision-language models (VLM), auto-speech-recognition (ASR), and text-to-speech ...
STB-VMM: Swin Transformer Based Video Motion Magnification (official repository)
[NeurIPS 2023] This is the official code for the paper "TPSR: Transformer-based Planning for Symbolic Regression"
An open source implementation of LFMs from Liquid AI: Liquid Foundation Models
Quick implementation of nGPT, learning entirely on the hypersphere, from NvidiaAI
Implementation of the proposed Spline-Based Transformer from Disney Research
A Faster LayoutReader Model based on LayoutLMv3, Sort OCR bboxes to reading order.
A performant high-throughput CPU-based API for Meta's No Language Left Behind (NLLB) using CTranslate2, hosted on Hugging Face Spaces.
🔥🔥🔥Official Codebase of "DiT-3D: Exploring Plain Diffusion Transformers for 3D Shape Generation"
This repo is the homebase of a community driven course on Computer Vision with Neural Networks. Feel free to join us on the Hugging Face discord: hf.co/join/discord
Official implementation of our ICLR 2023 paper "Crossformer: Transformer Utilizing Cross-Dimension Dependency for Multivariate Time Series Forecasting"
pretrained BERT model for cyber security text, learned CyberSecurity Knowledge
Implementation of π₀, the robotic foundation model architecture proposed by Physical Intelligence
Unified Efficient Fine-Tuning of 100+ LLMs (ACL 2024)
Nexa SDK is a comprehensive toolkit for supporting ONNX and GGML models. It supports text generation, image generation, vision-language models (VLM), auto-speech-recognition (ASR), and text-to-speech ...
21 Lessons, Get Started Building with Generative AI 🔗 https://microsoft.github.io/generative-ai-for-beginners/
🧑🏫 60+ Implementations/tutorials of deep learning papers with side-by-side notes 📝; including transformers (original, xl, switch, feedback, vit, ...), optimizers (adam, adabelief, sophia, ...), ga...
Implementation of π₀, the robotic foundation model architecture proposed by Physical Intelligence
:mag: AI orchestration framework to build customizable, production-ready LLM applications. Connect components (models, vector DBs, file converters) to pipelines or agents that can interact with your d...
A MLX port of FLUX based on the Huggingface Diffusers implementation.
State-of-the-art Machine Learning for the web. Run 🤗 Transformers directly in your browser, with no need for a server!
Implementation of Vision Transformer, a simple way to achieve SOTA in vision classification with only a single transformer encoder, in Pytorch
Official implementation for "UniST: A Prompt-Empowered Universal Model for Urban Spatio-Temporal Prediction" (KDD 2024)
This repository contains demos I made with the Transformers library by HuggingFace.
💡 All-in-one open-source embeddings database for semantic search, LLM orchestration and language model workflows
Semantic segmentation models with 500+ pretrained convolutional and transformer-based backbones.
[NeurIPS 2024 Oral][GPT beats diffusion🔥] [scaling laws in visual generation📈] Official impl. of "Visual Autoregressive Modeling: Scalable Image Generation via Next-Scale Prediction". An *ultra-simp...
Open-source offline translation library written in Python
Implementation of π₀, the robotic foundation model architecture proposed by Physical Intelligence
Official implementation for "UniST: A Prompt-Empowered Universal Model for Urban Spatio-Temporal Prediction" (KDD 2024)
A MLX port of FLUX based on the Huggingface Diffusers implementation.
Interpretable Pre-Trained Transformers for Heart Time-Series Data
Nexa SDK is a comprehensive toolkit for supporting ONNX and GGML models. It supports text generation, image generation, vision-language models (VLM), auto-speech-recognition (ASR), and text-to-speech ...
STB-VMM: Swin Transformer Based Video Motion Magnification (official repository)
[NeurIPS 2023] This is the official code for the paper "TPSR: Transformer-based Planning for Symbolic Regression"
An open source implementation of LFMs from Liquid AI: Liquid Foundation Models
Quick implementation of nGPT, learning entirely on the hypersphere, from NvidiaAI
Implementation of the proposed Spline-Based Transformer from Disney Research
A Faster LayoutReader Model based on LayoutLMv3, Sort OCR bboxes to reading order.
A performant high-throughput CPU-based API for Meta's No Language Left Behind (NLLB) using CTranslate2, hosted on Hugging Face Spaces.
🔥🔥🔥Official Codebase of "DiT-3D: Exploring Plain Diffusion Transformers for 3D Shape Generation"
Official implementation of our ICLR 2023 paper "Crossformer: Transformer Utilizing Cross-Dimension Dependency for Multivariate Time Series Forecasting"
pretrained BERT model for cyber security text, learned CyberSecurity Knowledge
Quick implementation of nGPT, learning entirely on the hypersphere, from NvidiaAI
AI-powered universal search for all your personal data, tailored just for you. Goal:The world's first product with "edge-side LLMs + consumer data localization" as its core development direction.
Implementation of the proposed Spline-Based Transformer from Disney Research
Implementation of LVSM, SOTA Large View Synthesis with Minimal 3d Inductive Bias, from Adobe Research
Implementation of π₀, the robotic foundation model architecture proposed by Physical Intelligence
21 Lessons, Get Started Building with Generative AI 🔗 https://microsoft.github.io/generative-ai-for-beginners/
Unified Efficient Fine-Tuning of 100+ LLMs (ACL 2024)
🧑🏫 60+ Implementations/tutorials of deep learning papers with side-by-side notes 📝; including transformers (original, xl, switch, feedback, vit, ...), optimizers (adam, adabelief, sophia, ...), ga...
Nexa SDK is a comprehensive toolkit for supporting ONNX and GGML models. It supports text generation, image generation, vision-language models (VLM), auto-speech-recognition (ASR), and text-to-speech ...
State-of-the-art Machine Learning for the web. Run 🤗 Transformers directly in your browser, with no need for a server!
:mag: AI orchestration framework to build customizable, production-ready LLM applications. Connect components (models, vector DBs, file converters) to pipelines or agents that can interact with your d...
Implementation of Vision Transformer, a simple way to achieve SOTA in vision classification with only a single transformer encoder, in Pytorch
💡 All-in-one open-source embeddings database for semantic search, LLM orchestration and language model workflows
An Easy-to-use, Scalable and High-performance RLHF Framework (70B+ PPO Full Tuning & Iterative DPO & LoRA & RingAttention)
This repository contains demos I made with the Transformers library by HuggingFace.
Quick implementation of nGPT, learning entirely on the hypersphere, from NvidiaAI
Chatbot for documentation, that allows you to chat with your data. Privately deployable, provides AI knowledge sharing and integrates knowledge into your AI workflow
OpenVINO™ is an open-source toolkit for optimizing and deploying AI inference
Code for "LayerSkip: Enabling Early Exit Inference and Self-Speculative Decoding", ACL 2024
Quick implementation of nGPT, learning entirely on the hypersphere, from NvidiaAI
Implementation of LVSM, SOTA Large View Synthesis with Minimal 3d Inductive Bias, from Adobe Research
AI-powered universal search for all your personal data, tailored just for you. Goal:The world's first product with "edge-side LLMs + consumer data localization" as its core development direction.
Implementation of π₀, the robotic foundation model architecture proposed by Physical Intelligence
🔥 Aurora Series: A more efficient multimodal large language model series for video.
Implementation of the proposed Spline-Based Transformer from Disney Research
An open source implementation of LFMs from Liquid AI: Liquid Foundation Models
Nexa SDK is a comprehensive toolkit for supporting ONNX and GGML models. It supports text generation, image generation, vision-language models (VLM), auto-speech-recognition (ASR), and text-to-speech ...
This paper list focuses on the theoretical and empirical analysis of language models, especially large language models (LLMs). The papers in this list investigate the learning behavior, generalization...
(Windows/Linux/MacOS) Local WebUI with neural network models (Text, Image, Video, 3D, Audio) on python (Gradio interface). Translated on 3 languages
VoiceRestore: Flow-Matching Transformers for Universal Speech Restoration
Official implementation for "UniST: A Prompt-Empowered Universal Model for Urban Spatio-Temporal Prediction" (KDD 2024)
Open Source Application for Advanced LLM Engineering: interact, train, fine-tune, and evaluate large language models on your own computer.
[NeurIPS 2024 Oral][GPT beats diffusion🔥] [scaling laws in visual generation📈] Official impl. of "Visual Autoregressive Modeling: Scalable Image Generation via Next-Scale Prediction". An *ultra-simp...
Nexa SDK is a comprehensive toolkit for supporting ONNX and GGML models. It supports text generation, image generation, vision-language models (VLM), auto-speech-recognition (ASR), and text-to-speech ...
Chronos: Pretrained (Language) Models for Probabilistic Time Series Forecasting
Lumina-T2X is a unified framework for Text to Any Modality Generation
From anything to mesh like human artists. Official impl. of "MeshAnything: Artist-Created Mesh Generation with Autoregressive Transformers"
streamline the fine-tuning process for multimodal models: PaliGemma, Florence-2, and Qwen2-VL
Implementation of the training framework proposed in Self-Rewarding Language Model, from MetaAI
Lag-Llama: Towards Foundation Models for Probabilistic Time Series Forecasting
A curriculum for learning about foundation models, from scratch to the frontier
Unified Training of Universal Time Series Forecasting Transformers
Official PyTorch Implementation of MambaVision: A Hybrid Mamba-Transformer Vision Backbone
Implementation of MeshGPT, SOTA Mesh generation using Attention, in Pytorch
The PyTorch implementation of Generative Pre-trained Transformers (GPTs) using Kolmogorov-Arnold Networks (KANs) for language modeling
Pytorch implementation of Transfusion, "Predict the Next Token and Diffuse Images with One Multi-Modal Model", from MetaAI
This repository is a curated collection of the most exciting and influential CVPR 2024 papers. 🔥 [Paper + Code + Demo]
From anything to mesh like human artists. Official impl. of "MeshAnything V2: Artist-Created Mesh Generation With Adjacent Mesh Tokenization"
21 Lessons, Get Started Building with Generative AI 🔗 https://microsoft.github.io/generative-ai-for-beginners/
Unified Efficient Fine-Tuning of 100+ LLMs (ACL 2024)
🧑🏫 60+ Implementations/tutorials of deep learning papers with side-by-side notes 📝; including transformers (original, xl, switch, feedback, vit, ...), optimizers (adam, adabelief, sophia, ...), ga...
State-of-the-art Machine Learning for the web. Run 🤗 Transformers directly in your browser, with no need for a server!
Open source real-time translation app for Android that runs locally
:mag: AI orchestration framework to build customizable, production-ready LLM applications. Connect components (models, vector DBs, file converters) to pipelines or agents that can interact with your d...
🤗 PEFT: State-of-the-art Parameter-Efficient Fine-Tuning.
Implementation of Vision Transformer, a simple way to achieve SOTA in vision classification with only a single transformer encoder, in Pytorch
Superduper: build end-2-end AI applications and templates using your existing data infrastructure and tools of choice
[NeurIPS 2024 Oral][GPT beats diffusion🔥] [scaling laws in visual generation📈] Official impl. of "Visual Autoregressive Modeling: Scalable Image Generation via Next-Scale Prediction". An *ultra-simp...
💡 All-in-one open-source embeddings database for semantic search, LLM orchestration and language model workflows
Robust recipes to align language models with human and AI preferences
This repository contains demos I made with the Transformers library by HuggingFace.
Nexa SDK is a comprehensive toolkit for supporting ONNX and GGML models. It supports text generation, image generation, vision-language models (VLM), auto-speech-recognition (ASR), and text-to-speech ...
OpenChat: Advancing Open-source Language Models with Imperfect Data
🦖 𝗟𝗲𝗮𝗿𝗻 about 𝗟𝗟𝗠𝘀, 𝗟𝗟𝗠𝗢𝗽𝘀, and 𝘃𝗲𝗰𝘁𝗼𝗿 𝗗𝗕𝘀 for free by designing, training, and deploying a real-time financial advisor LLM system ~ 𝘴𝘰𝘶𝘳𝘤𝘦 𝘤𝘰𝘥𝘦 + 𝘷𝘪𝘥𝘦𝘰 & 𝘳𝘦�...
A MLX port of FLUX based on the Huggingface Diffusers implementation.
From anything to mesh like human artists. Official impl. of "MeshAnything: Artist-Created Mesh Generation with Autoregressive Transformers"
streamline the fine-tuning process for multimodal models: PaliGemma, Florence-2, and Qwen2-VL
A collection of 🤗 Transformers.js demos and example applications
This repository is a curated collection of the most exciting and influential CVPR 2024 papers. 🔥 [Paper + Code + Demo]
From anything to mesh like human artists. Official impl. of "MeshAnything V2: Artist-Created Mesh Generation With Adjacent Mesh Tokenization"
Multi-LoRA inference server that scales to 1000s of fine-tuned LLMs
A library for easily merging multiple LLM experts, and efficiently train the merged LLM.
Open source real-time translation app for Android that runs locally
A novel implementation of fusing ViT with Mamba into a fast, agile, and high performance Multi-Modal Model. Powered by Zeta, the simplest AI framework ever.
The Truth Is In There: Improving Reasoning in Language Models with Layer-Selective Rank Reduction
Chronos: Pretrained (Language) Models for Probabilistic Time Series Forecasting
MOMENT: A Family of Open Time-series Foundation Models
Here lies the resources and topics necessary for the role of Data Scientist and Machine Learning
Pytorch implementation of Transfusion, "Predict the Next Token and Diffuse Images with One Multi-Modal Model", from MetaAI
A curriculum for learning about foundation models, from scratch to the frontier
Implementation of the training framework proposed in Self-Rewarding Language Model, from MetaAI
PyTorch Implementation of Jamba: "Jamba: A Hybrid Transformer-Mamba Language Model"
A Framework of Small-scale Large Multimodal Models