8 results found Sort:

Guide: Finetune GPT2-XL (1.5 Billion Parameters) and finetune GPT-NEO (2.7 B) on a single GPU with Huggingface Transformers using DeepSpeed
Created 2021-03-26
118 commits to main branch, last one about a year ago
SkyCode是一个多语言开源编程大模型,采用GPT3模型结构,支持Java, JavaScript, C, C++, Python, Go, shell等多种主流编程语言,并能理解中文注释。模型可以对代码进行补全,拥有强大解题能力,使您从编程中解放出来,专心于解决更重要的问题。| SkyCode is an open source programming model, which adopts...
Created 2022-12-14
19 commits to main branch, last one about a year ago
Salesforce codegen with web server
Created 2022-10-24
91 commits to main branch, last one 10 months ago
16
162
mit
8
QuickAI is a Python library that makes it extremely easy to experiment with state-of-the-art Machine Learning models.
Created 2021-04-09
222 commits to main branch, last one 16 days ago
A repository to run gpt-j-6b on low vram machines (4.2 gb minimum vram for 2000 token context, 3.5 gb for 1000 token context). Model loading takes 12gb free ram.
Created 2021-06-22
21 commits to main branch, last one 2 years ago
Fine-tuning 6-Billion GPT-J (& other models) with LoRA and 8-bit compression
Created 2022-03-02
11 commits to main branch, last one about a year ago
Colab notebooks to run a basic AI Dungeon clone using gpt-neo-2.7B
Created 2021-04-17
68 commits to main branch, last one 2 years ago
A basic ui for running gpt neo 2.7B on low vram (3 gb Vram minimum)
Created 2021-05-25
9 commits to main branch, last one 3 years ago