2 results found Sort:

📖A curated list of Awesome LLM/VLM Inference Papers with codes: WINT8/4, Flash-Attention, Paged-Attention, MLA, PP/TP/SP/CP/EP Parallelism, Prefix Cache, Chunked-Prefill, etc. 🎉🎉
Created 2023-08-27
458 commits to main branch, last one 17 days ago
135
1.8k
mit
57
[Paper List] Papers integrating knowledge graphs (KGs) and large language models (LLMs)
Created 2023-06-12
172 commits to main branch, last one 14 days ago