14 results found Sort:
- Filter by Primary Language:
- Python (10)
- HTML (1)
- Jupyter Notebook (1)
- +
Loki: Open-source solution designed to automate the process of verifying factuality
Created
2024-03-25
47 commits to main branch, last one 26 days ago
✨✨Woodpecker: Hallucination Correction for Multimodal Large Language Models. The first work to correct hallucinations in MLLMs.
Created
2023-09-26
106 commits to main branch, last one 11 days ago
Awesome-LLM-Robustness: a curated list of Uncertainty, Reliability and Robustness in Large Language Models
Created
2023-03-20
150 commits to main branch, last one 10 days ago
[ICLR'24] Mitigating Hallucination in Large Multi-Modal Models via Robust Instruction Tuning
Created
2023-06-15
366 commits to main branch, last one 3 months ago
RefChecker provides automatic checking pipeline and benchmark dataset for detecting fine-grained hallucinations generated by Large Language Models.
Created
2023-12-04
69 commits to main branch, last one 4 days ago
[CVPR'24] HallusionBench: You See What You Think? Or You Think What You See? An Image-Context Reasoning Benchmark Challenging for GPT-4V(ision), LLaVA-1.5, and Other Multi-modality Models
Created
2023-10-22
132 commits to main branch, last one 3 months ago
[ACL 2024] Benchmarking the Hallucination of Chinese Large Language Models via Unconstrained Generation
Created
2023-11-06
130 commits to main branch, last one about a month ago
😎 up-to-date & curated list of awesome LMM hallucinations papers, methods & resources.
Created
2023-10-11
57 commits to main branch, last one 3 months ago
Code for ACL 2024 paper "TruthX: Alleviating Hallucinations by Editing Large Language Models in Truthful Space"
Created
2024-02-27
18 commits to main branch, last one 3 months ago
[IJCAI 2024] FactCHD: Benchmarking Fact-Conflicting Hallucination Detection
Created
2023-04-05
22 commits to main branch, last one 2 months ago
This is the official repo for Debiasing Large Visual Language Models, including a Post-Hoc debias method and Visual Debias Decoding strategy.
Created
2024-01-23
7 commits to main branch, last one 3 months ago
Code & Data for our Paper "Alleviating Hallucinations of Large Language Models through Induced Hallucinations"
Created
2023-12-23
10 commits to main branch, last one 4 months ago
Official repo for the paper PHUDGE: Phi-3 as Scalable Judge. Evaluate your LLMs with or without custom rubric, reference answer, absolute, relative and much more. It contains a list of all the availab...
Created
2024-05-11
29 commits to main branch, last one 28 days ago
"Enhancing LLM Factual Accuracy with RAG to Counter Hallucinations: A Case Study on Domain-Specific Queries in Private Knowledge-Bases" by Jiarui Li and Ye Yuan and Zehua Zhang
Created
2024-02-17
205 commits to main branch, last one 3 months ago