Hallucination in large language models (LLMs) refers to generating unfaithful, fabricated, or nonsensical content not grounded in provided context or world knowledge. The focus is on extrinsic hallucination, emphasizing the need for LLMs to produce factual content and acknowledge when they lack knowledge. Causes of

27m read timeFrom lilianweng.github.io
Post cover image
Table of contents
What Causes Hallucinations? #Hallucination Detection #Anti-Hallucination Methods #Appendix: Evaluation Benchmarks #Citation #References #

Sort: