Statistics for topic llama
RepositoryStats tracks 584,797 Github repositories, of these 521 are tagged with the llama topic. The most common primary language for repositories using this topic is Python (272). Other languages include: Jupyter Notebook (43), TypeScript (31), C++ (23), Go (20), JavaScript (19), Rust (16)
Stargazers over time for topic llama
Most starred repositories for topic llama (view more)
Trending repositories for topic llama (view more)
Get up and running with Llama 3.2, Mistral, Gemma 2, and other large language models.
A high-throughput and memory-efficient inference and serving engine for LLMs
Unified Efficient Fine-Tuning of 100+ LLMs (ACL 2024)
:robot: The free, Open Source alternative to OpenAI, Claude and others. Self-hosted and local-first. Drop-in replacement for OpenAI, running on consumer-grade hardware. No GPU required. Runs gguf, tr...
Dabarqus is a stand alone application that implements a complete RAG solution.
Open source AI analyst powered by E2B. Analyze your CSV files with Llama 3.1 and create interactive charts.
This repository provides programs to build Retrieval Augmented Generation (RAG) code for Generative AI with LlamaIndex, Deep Lake, and Pinecone leveraging the power of OpenAI and Hugging Face models f...
Get up and running with Llama 3.2, Mistral, Gemma 2, and other large language models.
:robot: The free, Open Source alternative to OpenAI, Claude and others. Self-hosted and local-first. Drop-in replacement for OpenAI, running on consumer-grade hardware. No GPU required. Runs gguf, tr...
A high-throughput and memory-efficient inference and serving engine for LLMs
Unified Efficient Fine-Tuning of 100+ LLMs (ACL 2024)
TensorZero creates a feedback loop for optimizing LLM applications — turning production data into smarter, faster, and cheaper models.
Chat with AI large language models running natively in your browser. Enjoy private, server-free, seamless AI conversations.
Open source AI analyst powered by E2B. Analyze your CSV files with Llama 3.1 and create interactive charts.
📋 NotebookMLX - An Open Source version of NotebookLM (Ported NotebookLlama)
🦙 echoOLlama: A real-time voice AI platform powered by local LLMs. Features WebSocket streaming, voice interactions, and OpenAI API compatibility. Built with FastAPI, Redis, and PostgreSQL. Perfect f...
Open source AI analyst powered by E2B. Analyze your CSV files with Llama 3.1 and create interactive charts.
Get up and running with Llama 3.2, Mistral, Gemma 2, and other large language models.
Scripts for fine-tuning Meta Llama with composable FSDP & PEFT methods to cover single/multi-node GPUs. Supports default & custom datasets for applications such as summarization and Q&A. Supporting a ...
:robot: The free, Open Source alternative to OpenAI, Claude and others. Self-hosted and local-first. Drop-in replacement for OpenAI, running on consumer-grade hardware. No GPU required. Runs gguf, tr...
Unified Efficient Fine-Tuning of 100+ LLMs (ACL 2024)
A high-throughput and memory-efficient inference and serving engine for LLMs
🦙 echoOLlama: A real-time voice AI platform powered by local LLMs. Features WebSocket streaming, voice interactions, and OpenAI API compatibility. Built with FastAPI, Redis, and PostgreSQL. Perfect f...
LangCommand is a local inference command-line tool that transforms natural language descriptions into shell commands.
📋 NotebookMLX - An Open Source version of NotebookLM (Ported NotebookLlama)
TensorZero creates a feedback loop for optimizing LLM applications — turning production data into smarter, faster, and cheaper models.
Finetune Llama 3.2, Mistral, Phi, Qwen 2.5 & Gemma LLMs 2-5x faster with 80% less memory
High-speed Large Language Model Serving on PCs with Consumer-grade GPUs
Private & local AI personal knowledge management app for high entropy people.
SGLang is a fast serving framework for large language models and vision language models.
Llama3、Llama3.1 中文仓库(随书籍撰写中... 各种网友及厂商微调、魔改版本有趣权重 & 训练、推理、评测、部署教程视频 & 文档)
Get up and running with Llama 3.2, Mistral, Gemma 2, and other large language models.
Unified Efficient Fine-Tuning of 100+ LLMs (ACL 2024)
A high-throughput and memory-efficient inference and serving engine for LLMs
Finetune Llama 3.2, Mistral, Phi, Qwen 2.5 & Gemma LLMs 2-5x faster with 80% less memory
Finetune Llama 3.2, Mistral, Phi, Qwen 2.5 & Gemma LLMs 2-5x faster with 80% less memory
中文羊驼大模型三期项目 (Chinese Llama-3 LLMs) developed from Meta Llama 3
SGLang is a fast serving framework for large language models and vision language models.
[ECCV2024] Grounded Multimodal Large Language Model with Localized Visual Tokenization