An open API service providing repository metadata for many open source software ecosystems.

Topic: "hallucination"

Libr-AI/OpenFactVerification

Loki: Open-source solution designed to automate the process of verifying factuality

Language: Python - Size: 42.4 MB - Last synced at: 7 days ago - Pushed at: 8 months ago - Stars: 1,075 - Forks: 50

jxzhangjhu/Awesome-LLM-Uncertainty-Reliability-Robustness

Awesome-LLM-Robustness: a curated list of Uncertainty, Reliability and Robustness in Large Language Models

Size: 193 KB - Last synced at: 15 days ago - Pushed at: 15 days ago - Stars: 751 - Forks: 50

VITA-MLLM/Woodpecker

✨✨Woodpecker: Hallucination Correction for Multimodal Large Language Models

Language: Python - Size: 21.2 MB - Last synced at: 14 days ago - Pushed at: 5 months ago - Stars: 634 - Forks: 30

cvs-health/uqlm

UQLM: Uncertainty Quantification for Language Models, is a Python package for UQ-based LLM hallucination detection

Language: Python - Size: 11.6 MB - Last synced at: 2 days ago - Pushed at: 3 days ago - Stars: 625 - Forks: 51

amazon-science/RefChecker

RefChecker provides automatic checking pipeline and benchmark dataset for detecting fine-grained hallucinations generated by Large Language Models.

Language: Python - Size: 2.75 MB - Last synced at: 6 days ago - Pushed at: 20 days ago - Stars: 370 - Forks: 42

FuxiaoLiu/LRV-Instruction

[ICLR'24] Mitigating Hallucination in Large Multi-Modal Models via Robust Instruction Tuning

Language: Python - Size: 23.9 MB - Last synced at: 8 days ago - Pushed at: about 1 year ago - Stars: 279 - Forks: 13

tianyi-lab/HallusionBench

[CVPR'24] HallusionBench: You See What You Think? Or You Think What You See? An Image-Context Reasoning Benchmark Challenging for GPT-4V(ision), LLaVA-1.5, and Other Multi-modality Models

Language: Python - Size: 11.1 MB - Last synced at: 4 months ago - Pushed at: 7 months ago - Stars: 270 - Forks: 8

IAAR-Shanghai/ICSFSurvey

Explore concepts like Self-Correct, Self-Refine, Self-Improve, Self-Contradict, Self-Play, and Self-Knowledge, alongside o1-like reasoning elevation🍓 and hallucination alleviation🍄.

Language: Jupyter Notebook - Size: 5.02 MB - Last synced at: 13 days ago - Pushed at: 6 months ago - Stars: 168 - Forks: 5

IAAR-Shanghai/UHGEval

[ACL 2024] User-friendly evaluation framework: Eval Suite & Benchmarks: UHGEval, HaluEval, HalluQA, etc.

Language: Python - Size: 65.1 MB - Last synced at: 3 days ago - Pushed at: 7 months ago - Stars: 167 - Forks: 17

xieyuquanxx/awesome-Large-MultiModal-Hallucination 📦

😎 curated list of awesome LMM hallucinations papers, methods & resources.

Size: 66.4 KB - Last synced at: 24 days ago - Pushed at: about 1 year ago - Stars: 149 - Forks: 14

ictnlp/TruthX

Code for ACL 2024 paper "TruthX: Alleviating Hallucinations by Editing Large Language Models in Truthful Space"

Language: Python - Size: 31.6 MB - Last synced at: 3 days ago - Pushed at: about 1 year ago - Stars: 147 - Forks: 6

NishilBalar/Awesome-LVLM-Hallucination

up-to-date curated list of state-of-the-art Large vision language models hallucinations research work, papers & resources

Size: 189 KB - Last synced at: 26 days ago - Pushed at: 26 days ago - Stars: 125 - Forks: 6

zjunlp/FactCHD

[IJCAI 2024] FactCHD: Benchmarking Fact-Conflicting Hallucination Detection

Language: Python - Size: 8.42 MB - Last synced at: 29 days ago - Pushed at: about 1 year ago - Stars: 87 - Forks: 2

zjunlp/Deco

[ICLR 2025] MLLM can see? Dynamic Correction Decoding for Hallucination Mitigation

Language: Python - Size: 17.6 MB - Last synced at: 3 days ago - Pushed at: 6 months ago - Stars: 82 - Forks: 7

yfzhang114/LLaVA-Align

This is the official repo for Debiasing Large Visual Language Models, including a Post-Hoc debias method and Visual Debias Decoding strategy.

Language: Python - Size: 64.9 MB - Last synced at: about 2 months ago - Pushed at: 3 months ago - Stars: 77 - Forks: 2

HillZhang1999/ICD

Code & Data for our Paper "Alleviating Hallucinations of Large Language Models through Induced Hallucinations"

Language: Python - Size: 14.3 MB - Last synced at: about 1 year ago - Pushed at: over 1 year ago - Stars: 50 - Forks: 4

deshwalmahesh/PHUDGE

Official repo for the paper PHUDGE: Phi-3 as Scalable Judge. Evaluate your LLMs with or without custom rubric, reference answer, absolute, relative and much more. It contains a list of all the available tool, methods, repo, code etc to detect hallucination, LLM evaluation, grading and much more.

Language: Jupyter Notebook - Size: 13.1 MB - Last synced at: 24 days ago - Pushed at: 11 months ago - Stars: 49 - Forks: 7

zjunlp/KnowledgeCircuits

Knowledge Circuits in Pretrained Transformers

Language: Python - Size: 5.93 MB - Last synced at: 9 months ago - Pushed at: 9 months ago - Stars: 46 - Forks: 1

anlp-team/LTI_Neural_Navigator

"Enhancing LLM Factual Accuracy with RAG to Counter Hallucinations: A Case Study on Domain-Specific Queries in Private Knowledge-Bases" by Jiarui Li and Ye Yuan and Zehua Zhang

Language: HTML - Size: 32.3 MB - Last synced at: 8 days ago - Pushed at: about 1 year ago - Stars: 44 - Forks: 3

taco-group/Re-Align

A novel alignment framework that leverages image retrieval to mitigate hallucinations in Vision Language Models.

Language: Python - Size: 18.6 MB - Last synced at: about 1 month ago - Pushed at: about 1 month ago - Stars: 40 - Forks: 1

dmis-lab/OLAPH

OLAPH: Improving Factuality in Biomedical Long-form Question Answering

Language: Python - Size: 161 MB - Last synced at: about 2 months ago - Pushed at: 9 months ago - Stars: 39 - Forks: 4

AmourWaltz/Reliable-LLM

Language: JavaScript - Size: 14.8 MB - Last synced at: 9 months ago - Pushed at: 9 months ago - Stars: 35 - Forks: 1

345ishaan/DenseLidarNet

Language: Jupyter Notebook - Size: 125 MB - Last synced at: about 2 years ago - Pushed at: over 6 years ago - Stars: 35 - Forks: 8

dobriban/Principles-of-AI-LLMs

Materials for the course Principles of AI: LLMs at UPenn (Stat 9911, Spring 2025). LLM architectures, training paradigms (pre- and post-training, alignment), test-time computation, reasoning, safety and robustness (jailbreaking, oversight, uncertainty), representations, interpretability (circuits), etc.

Size: 238 MB - Last synced at: 3 days ago - Pushed at: 4 days ago - Stars: 33 - Forks: 2

Ruiyang-061X/VL-Uncertainty

🔎Official code for our paper: "VL-Uncertainty: Detecting Hallucination in Large Vision-Language Model via Uncertainty Estimation".

Language: Python - Size: 7.12 MB - Last synced at: about 2 months ago - Pushed at: 3 months ago - Stars: 31 - Forks: 2

zjunlp/EasyDetect

[ACL 2024] An Easy-to-use Hallucination Detection Framework for LLMs.

Language: Python - Size: 11.5 MB - Last synced at: 3 months ago - Pushed at: 3 months ago - Stars: 30 - Forks: 1

kortex-labs/korrect

Framework for hallucination detection and correction in LLMs

Language: Python - Size: 314 KB - Last synced at: over 1 year ago - Pushed at: over 1 year ago - Stars: 30 - Forks: 2

kereva-dev/kereva-scanner

Code scanner to check for issues in prompts and LLM calls

Language: Python - Size: 7.12 MB - Last synced at: 2 months ago - Pushed at: 2 months ago - Stars: 29 - Forks: 2

germain-hug/NeurHal

Visual Correspondence Hallucination: Towards Geometric Reasoning (Under Review)

Size: 817 KB - Last synced at: 10 months ago - Pushed at: over 2 years ago - Stars: 29 - Forks: 0

project-miracl/nomiracl

NoMIRACL: A multilingual hallucination evaluation dataset to evaluate LLM robustness in RAG against first-stage retrieval errors on 18 languages.

Language: Python - Size: 42.1 MB - Last synced at: 29 days ago - Pushed at: 6 months ago - Stars: 24 - Forks: 4

KejiaZhang-Robust/VAP

Poison as Cure: Visual Noise for Mitigating Object Hallucinations in LVMs

Language: Python - Size: 33.5 MB - Last synced at: about 1 month ago - Pushed at: about 1 month ago - Stars: 22 - Forks: 0

ahnjaewoo/timechara

🧙🏻Code and benchmark for our Findings of ACL 2024 paper - "TimeChara: Evaluating Point-in-Time Character Hallucination of Role-Playing Large Language Models"

Language: Python - Size: 869 KB - Last synced at: 6 months ago - Pushed at: 6 months ago - Stars: 19 - Forks: 0

XinYuANU/FaceAttr

CVPR2018 Face Super-resolution with supplementary Attributes

Language: Lua - Size: 2.62 MB - Last synced at: over 1 year ago - Pushed at: almost 7 years ago - Stars: 18 - Forks: 7

baixianghuang/HalluEditBench

Can Knowledge Editing Really Correct Hallucinations? (ICLR 2025)

Language: Python - Size: 91.6 MB - Last synced at: 16 days ago - Pushed at: 16 days ago - Stars: 15 - Forks: 2

The-Martyr/Awesome-Modality-Priors-in-MLLMs

Latest Advances on Modality Priors in Multimodal Large Language Models

Size: 76.2 KB - Last synced at: about 1 month ago - Pushed at: about 1 month ago - Stars: 15 - Forks: 1

jiazhen-code/PhD

[CVPR25] A ChatGPT-Prompted Visual hallucination Evaluation Dataset, featuring over 100,000 data samples and four advanced evaluation modes. The dataset includes extensive contextual descriptions, counterintuitive images, and clear indicators of hallucination items.

Size: 28 MB - Last synced at: about 2 months ago - Pushed at: about 2 months ago - Stars: 14 - Forks: 0

zjunlp/NLPCC2024_RegulatingLLM

[NLPCC 2024] Shared Task 10: Regulating Large Language Models

Size: 2.06 MB - Last synced at: 3 months ago - Pushed at: 12 months ago - Stars: 13 - Forks: 2

fanqiwan/KCA Fork of 18907305772/KCA

Knowledge Verification to Nip Hallucination in the Bud

Language: Python - Size: 4.84 MB - Last synced at: about 1 year ago - Pushed at: about 1 year ago - Stars: 13 - Forks: 0

Sreyan88/VDGD

Code for ICLR 2025 Paper: Visual Description Grounding Reduces Hallucinations and Boosts Reasoning in LVLMs

Language: Python - Size: 16.6 MB - Last synced at: 29 days ago - Pushed at: 29 days ago - Stars: 11 - Forks: 2

sled-group/3D-GRAND

Official Implementation of 3D-GRAND: Towards Better Grounding and Less Hallucination for 3D-LLMs

Size: 26 MB - Last synced at: 12 months ago - Pushed at: 12 months ago - Stars: 10 - Forks: 0

yg211/explainable-metrics

An explainable sentence similarity measurement

Language: Jupyter Notebook - Size: 1.47 MB - Last synced at: about 2 years ago - Pushed at: about 4 years ago - Stars: 9 - Forks: 1

wisecubeai/pythia

Open source AI hallucination monitoring

Language: Python - Size: 1.06 MB - Last synced at: 3 months ago - Pushed at: 3 months ago - Stars: 6 - Forks: 0

edshkim98/LocalDiffusion-Hallucination

Official code for 'Tackling Structural Hallucination in Image Translation with Local Diffusion' (ECCV'24 Oral)

Language: Python - Size: 33.2 MB - Last synced at: 9 months ago - Pushed at: 9 months ago - Stars: 6 - Forks: 0

aimonlabs/hallucination-detection-model

HalluciNot: Hallucination Detection Through Context and Common Knowledge Verification

Language: Python - Size: 279 KB - Last synced at: 17 days ago - Pushed at: about 1 month ago - Stars: 5 - Forks: 0

hanmenghan/Skip-n

This repository contains the code of our paper 'Skip \n: A simple method to reduce hallucination in Large Vision-Language Models'.

Language: Python - Size: 301 KB - Last synced at: over 1 year ago - Pushed at: over 1 year ago - Stars: 4 - Forks: 0

KevinBian107/RPLH

Developing various frameworks for multi-agent collaborations systems

Language: Python - Size: 36.6 MB - Last synced at: 2 months ago - Pushed at: 2 months ago - Stars: 3 - Forks: 1

Wind-Flipped/DyG-Hallucination

Cocktail dynamic graph prompting technique in LLM for hallucination

Language: Python - Size: 1.3 MB - Last synced at: 3 months ago - Pushed at: 3 months ago - Stars: 3 - Forks: 0

qqplot/dcpmi

[NAACL24] Official Implementation of Mitigating Hallucination in Abstractive Summarization with Domain-Conditional Mutual Information

Language: Python - Size: 24.7 MB - Last synced at: 10 months ago - Pushed at: 10 months ago - Stars: 2 - Forks: 1

ritesh-modi/embedding-hallucinations

This repo shows how foundational model hallucinates and how we can fix such hallucinations using fine-tuning them

Language: Python - Size: 474 KB - Last synced at: about 2 months ago - Pushed at: about 2 months ago - Stars: 1 - Forks: 0

robertbenson/docker_openai_custom_weather_demo

openai function calling demo that gets customizable weather information

Language: Python - Size: 760 KB - Last synced at: 12 months ago - Pushed at: 12 months ago - Stars: 1 - Forks: 0

18907305772/KCA

Knowledge Verification to Nip Hallucination in the Bud

Language: Python - Size: 4.84 MB - Last synced at: about 1 year ago - Pushed at: about 1 year ago - Stars: 1 - Forks: 1

wangtz19/DecodingStrategy

Unofficial implementations for optimized decoding strategies of large language models

Language: Jupyter Notebook - Size: 14.6 KB - Last synced at: 3 days ago - Pushed at: over 1 year ago - Stars: 1 - Forks: 0

aimpowered/stuttered-speech-benchmark

Data and code used to produce result in InterSpeech 2025 paper "J-j-j-just Stutter: Benchmarking Whisper’s Performance Disparities on Different Stuttering Patterns"

Language: Jupyter Notebook - Size: 176 MB - Last synced at: 5 days ago - Pushed at: 5 days ago - Stars: 0 - Forks: 0

rafay123321/embedding-hallucinations

This repo shows how foundational model hallucinates and how we can fix such hallucinations using fine-tuning them

Language: Python - Size: 476 KB - Last synced at: 6 days ago - Pushed at: 6 days ago - Stars: 0 - Forks: 0

SALT-Research/SHALLOW

SHALLOW, the first hallucination benchmark for ASR models

Language: Python - Size: 3.42 MB - Last synced at: 13 days ago - Pushed at: 13 days ago - Stars: 0 - Forks: 0

S18-Niloy/ChatBot_langchain_CoVe_Demo

Language: Jupyter Notebook - Size: 110 KB - Last synced at: 14 days ago - Pushed at: about 1 month ago - Stars: 0 - Forks: 0

DIDSR/sfrc_evals_hallucinations

sFRC: To identify fakes in medical images reconstructed using AI

Language: Python - Size: 137 MB - Last synced at: 23 days ago - Pushed at: about 1 month ago - Stars: 0 - Forks: 0

kou-saki/Talking-to-an-AI-That-Hallucinates

A user-driven exploration into why inference-based AI hallucinates—and how to talk with it more meaningfully.

Size: 35.2 KB - Last synced at: about 2 months ago - Pushed at: about 2 months ago - Stars: 0 - Forks: 0

kou-saki/i-asked-it-to-forget

I Asked It to Forget, but It Didn't — A Case of Miscommunication Between AI and Humans

Size: 25.4 KB - Last synced at: about 2 months ago - Pushed at: about 2 months ago - Stars: 0 - Forks: 0

zer0int/LLMorse

Talk Morse code to multimodal LLM using your voice. Beep-boop!

Language: Python - Size: 66.4 KB - Last synced at: 3 months ago - Pushed at: 7 months ago - Stars: 0 - Forks: 0

amarquaye/atlas-chrome

Chrome extension for the ATLAS project.

Language: JavaScript - Size: 346 KB - Last synced at: 3 months ago - Pushed at: 9 months ago - Stars: 0 - Forks: 0

amarquaye/atlas

🔢Hallucination detector for Large Language Models.

Language: Jupyter Notebook - Size: 1.86 MB - Last synced at: 3 months ago - Pushed at: 9 months ago - Stars: 0 - Forks: 0

weijiaheng/CHALE

Controlled HALlucination-Evaluation (CHALE) Question-Answering Dataset

Language: Python - Size: 1.38 MB - Last synced at: about 1 year ago - Pushed at: about 1 year ago - Stars: 0 - Forks: 0

robertbenson/openai_assistant_code_interpreter

openai assistant using code interpreter

Language: Python - Size: 1.95 KB - Last synced at: about 1 year ago - Pushed at: about 1 year ago - Stars: 0 - Forks: 0

kenza-ily/QuantHaLL

QuantHaLL: Quantifying Hallucination in machine translation for Low-resource Languages

Language: Jupyter Notebook - Size: 71.2 MB - Last synced at: about 1 year ago - Pushed at: about 1 year ago - Stars: 0 - Forks: 0

CrackedResearcher/LLMVerify

Verify outputs generated by LLMs backed with real time data

Language: Python - Size: 10.7 KB - Last synced at: 16 days ago - Pushed at: about 1 year ago - Stars: 0 - Forks: 0

vr25/lrec-coling-hallucination-tutorial

LREC-COLING 2024 Tutorial

Language: JavaScript - Size: 80 MB - Last synced at: about 1 year ago - Pushed at: about 1 year ago - Stars: 0 - Forks: 0

SingularityLabs-ai/truthgpt-on-google-extension-mini

hallucination free LLM - TruthGPT for Google extension is a version of TruthGPT (developed by Labs) which integrates TruthGPT with Google search results.

Size: 1.95 KB - Last synced at: over 1 year ago - Pushed at: over 1 year ago - Stars: 0 - Forks: 0

VerseMetaVerse/GPT

Hallucinate - GPT - LLM - AI Chat - OpenAI - Sam Altman info

Size: 1.95 KB - Last synced at: 2 months ago - Pushed at: over 1 year ago - Stars: 0 - Forks: 0

Arena-Rosnav/lflh

Language: Python - Size: 69.7 MB - Last synced at: about 2 years ago - Pushed at: over 2 years ago - Stars: 0 - Forks: 0

Related Topics
llm 22 large-language-models 21 hallucination-detection 9 chatgpt 9 ai 8 llms 7 gpt-4 7 multimodal-large-language-models 6 natural-language-processing 6 evaluation 6 benchmark 6 openai 6 multimodal 5 mllm 5 hallucinations 4 llama 4 large-vision-language-models 4 hallucination-evaluation 4 artificial-intelligence 4 machine-learning 4 safety 4 decoding 4 nlp 4 hallucination-mitigation 4 prompt-engineering 4 dataset 4 language-model 3 uncertainty-quantification 3 uncertainty-estimation 3 rag 3 vision-language-model 3 reasoning 3 retrieval-augmented-generation 3 factuality 3 llm-evaluation 3 knowledge-editing 3 gpt 3 chain-of-thought 2 rlhf 2 lmm 2 interpretability 2 in-context-learning 2 inference-based-ai 2 deep-learning 2 model-editing 2 generative-ai 2 lvlm 2 claude 2 embedding-models 2 embedding-vectors 2 fine-tuning 2 msmarco 2 sentence-embeddings 2 asr 2 alignment 2 ai-evaluation 2 multi-modal 2 pytorch 2 3d 2 python 2 supervised-finetuning 2 gpt-3 2 transformers 2 reliability 2 explainable-ai 2 embeddings 2 robustness 2 uncertainty 2 knowledge 2 llama2 2 inference 2 llava 2 miscommunication 1 feedback-collection 1 custom-dataset 1 speech-recognition 1 speechllm 1 mistral 1 truthfulness 1 dialogue 1 ai-behaviour 1 prompt-analysis 1 ai-errors 1 context-persistence 1 representation 1 chatgpt4 1 llm-inference 1 ai-human-interaction 1 ai-research 1 ai-interpretability 1 inverse-problems 1 fake-detection 1 vqa 1 subsampling 1 vision-and-language 1 superresolution 1 vision 1 vicuna 1 object-detection 1 iclr2024 1