Statistics for topic llama
RepositoryStats tracks 579,129 Github repositories, of these 515 are tagged with the llama topic. The most common primary language for repositories using this topic is Python (271). Other languages include: Jupyter Notebook (42), TypeScript (30), C++ (24), Go (18), JavaScript (18), Rust (16)
Stargazers over time for topic llama
Most starred repositories for topic llama (view more)
Trending repositories for topic llama (view more)
Get up and running with Llama 3.2, Mistral, Gemma 2, and other large language models.
Scripts for fine-tuning Meta Llama with composable FSDP & PEFT methods to cover single/multi-node GPUs. Supports default & custom datasets for applications such as summarization and Q&A. Supporting a ...
Finetune Llama 3.2, Mistral, Phi, Qwen & Gemma LLMs 2-5x faster with 80% less memory
Dabarqus is a stand alone application that implements a complete RAG solution.
🦙 echoOLlama: A real-time voice AI platform powered by local LLMs. Features WebSocket streaming, voice interactions, and OpenAI API compatibility. Built with FastAPI, Redis, and PostgreSQL. Perfect f...
🦙 echoOLlama: A real-time voice AI platform powered by local LLMs. Features WebSocket streaming, voice interactions, and OpenAI API compatibility. Built with FastAPI, Redis, and PostgreSQL. Perfect f...
Get up and running with Llama 3.2, Mistral, Gemma 2, and other large language models.
Scripts for fine-tuning Meta Llama with composable FSDP & PEFT methods to cover single/multi-node GPUs. Supports default & custom datasets for applications such as summarization and Q&A. Supporting a ...
Finetune Llama 3.2, Mistral, Phi, Qwen & Gemma LLMs 2-5x faster with 80% less memory
🦙 echoOLlama: A real-time voice AI platform powered by local LLMs. Features WebSocket streaming, voice interactions, and OpenAI API compatibility. Built with FastAPI, Redis, and PostgreSQL. Perfect f...
Dabarqus is a stand alone application that implements a complete RAG solution.
📋 NotebookMLX - An Open Source version of NotebookLM (Ported NotebookLlama)
Open source AI analyst powered by E2B. Analyze your CSV files with Llama 3.1 and create interactive charts.
📋 NotebookMLX - An Open Source version of NotebookLM (Ported NotebookLlama)
The official code of the paper "Deciphering Cross-Modal Alignment in Large Vision-Language Models with Modality Integration Rate".
🦙 echoOLlama: A real-time voice AI platform powered by local LLMs. Features WebSocket streaming, voice interactions, and OpenAI API compatibility. Built with FastAPI, Redis, and PostgreSQL. Perfect f...
Get up and running with Llama 3.2, Mistral, Gemma 2, and other large language models.
Scripts for fine-tuning Meta Llama with composable FSDP & PEFT methods to cover single/multi-node GPUs. Supports default & custom datasets for applications such as summarization and Q&A. Supporting a ...
Unified Efficient Fine-Tuning of 100+ LLMs (ACL 2024)
A high-throughput and memory-efficient inference and serving engine for LLMs
The official code of the paper "Deciphering Cross-Modal Alignment in Large Vision-Language Models with Modality Integration Rate".
🦙 echoOLlama: A real-time voice AI platform powered by local LLMs. Features WebSocket streaming, voice interactions, and OpenAI API compatibility. Built with FastAPI, Redis, and PostgreSQL. Perfect f...
LangCommand is a local inference command-line tool that transforms natural language descriptions into shell commands.
📋 NotebookMLX - An Open Source version of NotebookLM (Ported NotebookLlama)
Finetune Llama 3.2, Mistral, Phi, Qwen & Gemma LLMs 2-5x faster with 80% less memory
High-speed Large Language Model Serving on PCs with Consumer-grade GPUs
Private & local AI personal knowledge management app for high entropy thinkers.
SGLang is a fast serving framework for large language models and vision language models.
Llama3、Llama3.1 中文仓库(随书籍撰写中... 各种网友及厂商微调、魔改版本有趣权重 & 训练、推理、评测、部署教程视频 & 文档)
Get up and running with Llama 3.2, Mistral, Gemma 2, and other large language models.
Unified Efficient Fine-Tuning of 100+ LLMs (ACL 2024)
A high-throughput and memory-efficient inference and serving engine for LLMs
Finetune Llama 3.2, Mistral, Phi, Qwen & Gemma LLMs 2-5x faster with 80% less memory
Finetune Llama 3.2, Mistral, Phi, Qwen & Gemma LLMs 2-5x faster with 80% less memory
PyTorch native quantization and sparsity for training and inference
中文羊驼大模型三期项目 (Chinese Llama-3 LLMs) developed from Meta Llama 3
SGLang is a fast serving framework for large language models and vision language models.