3 results found Sort:
:bust_in_silhouette: Multi-Armed Bandit Algorithms Library (MAB) :cop:
Created
2019-01-24
69 commits to master branch, last one 2 years ago
🌾 OAT: Online AlignmenT for LLMs
Created
2024-10-15
22 commits to main branch, last one 23 hours ago
Library for multi-armed bandit selection strategies, including efficient deterministic implementations of Thompson sampling and epsilon-greedy.
Created
2021-02-18
20 commits to main branch, last one 3 years ago