[ACL 2024] User-friendly evaluation framework: Eval Suite & Benchmarks: UHGEval, HaluEval, HalluQA, etc.
-
Updated
Nov 12, 2024 - Python
[ACL 2024] User-friendly evaluation framework: Eval Suite & Benchmarks: UHGEval, HaluEval, HalluQA, etc.
up-to-date curated list of state-of-the-art Large vision language models hallucinations research work, papers & resources
🔎Official code for our paper: "VL-Uncertainty: Detecting Hallucination in Large Vision-Language Model via Uncertainty Estimation".
Dataset Generation and Pre-processing Scripts for the Research titled: Leveraging the Domain Adaptation of Retrieval Augmented Generation (RAG) Models in Conversational AI for Enhanced Customer Service
This project integrates business rules management systems (BRMS) and a RAG, to offer an automated text generation solution, applicable in different contexts and significantly reducing LLM hallucinations. It's a complete architecture available in a chatBot and fully scalable according to needs
Add a description, image, and links to the hallucination-evaluation topic page so that developers can more easily learn about it.
To associate your repository with the hallucination-evaluation topic, visit your repo's landing page and select "manage topics."