11 results found Sort:

438
2.8k
apache-2.0
33
MedicalGPT: Training Your Own Medical GPT Model with ChatGPT Training Pipeline. 训练医疗大模型,实现了包括增量预训练(PT)、有监督微调(SFT)、RLHF、DPO、ORPO。
Created 2023-06-02
492 commits to main branch, last one 2 days ago
184
1.9k
apache-2.0
18
ms-swift: Use PEFT or Full-parameter to finetune 250+ LLMs or 35+ MLLMs. (Qwen2, GLM4, Internlm2, Yi, Llama3, Llava, Deepseek, Baichuan2...)
Created 2023-08-01
621 commits to main branch, last one 20 hours ago
31
604
apache-2.0
6
A library with extensible implementations of DPO, KTO, PPO, ORPO, and other human-aware loss functions (HALOs).
Created 2023-12-03
111 commits to main branch, last one 13 days ago
59
541
apache-2.0
9
Easy and Efficient Finetuning LLMs. (Supported LLama, LLama2, LLama3, Qwen, Baichuan, GLM , Falcon) 大模型高效量化训练+部署.
Created 2023-05-25
479 commits to main branch, last one 6 days ago
tensorflow를 사용하여 텍스트 전처리부터, Topic Models, BERT, GPT, LLM과 같은 최신 모델의 다운스트림 태스크들을 정리한 Deep Learning NLP 저장소입니다.
Created 2021-12-30
250 commits to main branch, last one 3 months ago
20
160
mit
4
SiLLM simplifies the process of training and running Large Language Models (LLMs) on Apple Silicon by leveraging the MLX framework.
Created 2024-01-11
298 commits to main branch, last one 11 days ago
Notus is a collection of fine-tuned LLMs using SFT, DPO, SFT+DPO, and/or any other RLHF techniques, while always keeping a data-first approach
Created 2023-11-16
52 commits to main branch, last one 6 months ago
Technical anaysis library for .NET
Created 2016-06-30
64 commits to master branch, last one 3 years ago
CodeUltraFeedback: aligning large language models to coding preferences
Created 2024-01-25
50 commits to main branch, last one 13 days ago
0
54
unknown
3
Step-aware Preference Optimization: Aligning Preference with Denoising Performance at Each Step
Created 2024-05-26
4 commits to main branch, last one a day ago
1
49
apache-2.0
3
A RLHF Infrastructure for Vision-Language Models
Created 2023-12-27
5 commits to main branch, last one a day ago