14 results found Sort:

Loki: Open-source solution designed to automate the process of verifying factuality
Created 2024-03-25
47 commits to main branch, last one 26 days ago
29
571
unknown
15
✨✨Woodpecker: Hallucination Correction for Multimodal Large Language Models. The first work to correct hallucinations in MLLMs.
Created 2023-09-26
106 commits to main branch, last one 11 days ago
Awesome-LLM-Robustness: a curated list of Uncertainty, Reliability and Robustness in Large Language Models
Created 2023-03-20
150 commits to main branch, last one 10 days ago
14
230
bsd-3-clause
11
[ICLR'24] Mitigating Hallucination in Large Multi-Modal Models via Robust Instruction Tuning
Created 2023-06-15
366 commits to main branch, last one 3 months ago
17
220
apache-2.0
8
RefChecker provides automatic checking pipeline and benchmark dataset for detecting fine-grained hallucinations generated by Large Language Models.
Created 2023-12-04
69 commits to main branch, last one 4 days ago
3
200
bsd-3-clause
4
[CVPR'24] HallusionBench: You See What You Think? Or You Think What You See? An Image-Context Reasoning Benchmark Challenging for GPT-4V(ision), LLaVA-1.5, and Other Multi-modality Models
Created 2023-10-22
132 commits to main branch, last one 3 months ago
17
166
apache-2.0
11
[ACL 2024] Benchmarking the Hallucination of Chinese Large Language Models via Unconstrained Generation
Created 2023-11-06
130 commits to main branch, last one about a month ago
😎 up-to-date & curated list of awesome LMM hallucinations papers, methods & resources.
Created 2023-10-11
57 commits to main branch, last one 3 months ago
3
75
gpl-3.0
4
Code for ACL 2024 paper "TruthX: Alleviating Hallucinations by Editing Large Language Models in Truthful Space"
Created 2024-02-27
18 commits to main branch, last one 3 months ago
[IJCAI 2024] FactCHD: Benchmarking Fact-Conflicting Hallucination Detection
Created 2023-04-05
22 commits to main branch, last one 2 months ago
2
65
apache-2.0
2
This is the official repo for Debiasing Large Visual Language Models, including a Post-Hoc debias method and Visual Debias Decoding strategy.
Created 2024-01-23
7 commits to main branch, last one 3 months ago
Code & Data for our Paper "Alleviating Hallucinations of Large Language Models through Induced Hallucinations"
Created 2023-12-23
10 commits to main branch, last one 4 months ago
Official repo for the paper PHUDGE: Phi-3 as Scalable Judge. Evaluate your LLMs with or without custom rubric, reference answer, absolute, relative and much more. It contains a list of all the availab...
Created 2024-05-11
29 commits to main branch, last one 28 days ago
"Enhancing LLM Factual Accuracy with RAG to Counter Hallucinations: A Case Study on Domain-Specific Queries in Private Knowledge-Bases" by Jiarui Li and Ye Yuan and Zehua Zhang
Created 2024-02-17
205 commits to main branch, last one 3 months ago