1 result found Sort:

50
641
apache-2.0
15
FP16xINT4 LLM inference kernel that can achieve near-ideal ~4x speedups up to medium batchsizes of 16-32 tokens.
Created 2024-01-17
14 commits to master branch, last one 3 months ago