11 results found Sort:
- Filter by Primary Language:
- Python (9)
- C# (1)
- Jupyter Notebook (1)
- +
MedicalGPT: Training Your Own Medical GPT Model with ChatGPT Training Pipeline. 训练医疗大模型,实现了包括增量预训练(PT)、有监督微调(SFT)、RLHF、DPO、ORPO。
Created
2023-06-02
492 commits to main branch, last one 2 days ago
ms-swift: Use PEFT or Full-parameter to finetune 250+ LLMs or 35+ MLLMs. (Qwen2, GLM4, Internlm2, Yi, Llama3, Llava, Deepseek, Baichuan2...)
Created
2023-08-01
621 commits to main branch, last one 20 hours ago
A library with extensible implementations of DPO, KTO, PPO, ORPO, and other human-aware loss functions (HALOs).
Created
2023-12-03
111 commits to main branch, last one 13 days ago
Easy and Efficient Finetuning LLMs. (Supported LLama, LLama2, LLama3, Qwen, Baichuan, GLM , Falcon) 大模型高效量化训练+部署.
Created
2023-05-25
479 commits to main branch, last one 6 days ago
tensorflow를 사용하여 텍스트 전처리부터, Topic Models, BERT, GPT, LLM과 같은 최신 모델의 다운스트림 태스크들을 정리한 Deep Learning NLP 저장소입니다.
Created
2021-12-30
250 commits to main branch, last one 3 months ago
SiLLM simplifies the process of training and running Large Language Models (LLMs) on Apple Silicon by leveraging the MLX framework.
Created
2024-01-11
298 commits to main branch, last one 11 days ago
Notus is a collection of fine-tuned LLMs using SFT, DPO, SFT+DPO, and/or any other RLHF techniques, while always keeping a data-first approach
Created
2023-11-16
52 commits to main branch, last one 6 months ago
Technical anaysis library for .NET
Created
2016-06-30
64 commits to master branch, last one 3 years ago
CodeUltraFeedback: aligning large language models to coding preferences
Created
2024-01-25
50 commits to main branch, last one 13 days ago
Step-aware Preference Optimization: Aligning Preference with Denoising Performance at Each Step
Created
2024-05-26
4 commits to main branch, last one a day ago