# How to measure hallucination
[INSIDE: LLMS’ INTERNAL STATES RETAIN THE POWER OF HALLUCINATION DETECTION](https://openreview.net/pdf?id=Zj12nzlQbz)
- ICLR2024


[A Mathematical Investigation of Hallucination and Creativity in GPT Models](https://www.mdpi.com/2227-7390/11/10/2320)

[Measuring and Reducing LLM Hallucination without Gold-Standard Answers via Expertise-Weighting](https://arxiv.org/pdf/2402.10412.pdf)


[KCTS: Knowledge-Constrained Tree Search Decoding with
Token-Level Hallucination Detection](https://aclanthology.org/2023.emnlp-main.867.pdf)
- EMNLP2023

[AutoHall: Automated Hallucination Dataset Generation for Large Language Models](https://arxiv.org/pdf/2310.00259.pdf)

[HaluEval: A Large-Scale Hallucination Evaluation Benchmar for Large Language Models](https://arxiv.org/pdf/2305.11747.pdf)
[Semantic Consistency for Assuring Reliability of Large Language Models](https://arxiv.org/pdf/2308.09138.pdf)