Vectara launches the open source Hughes Hallucination Evaluation Model (HHEM) and uses it to compare hallucination rates across top LLMs including OpenAI, Cohere, PaLM, Anthropic’s Claude 2 and more.
Read moreCategories
Blog - Hallucination
All posts
Celebrating 2 Million Downloads of HHEM
Vectara’s Hallucination Evaluation Model surpasses 2 million downloads as the fight against LLM hallucinations continues
HHEM 2.1: A Better Hallucination Detection Model and a New Leaderboard
The upgraded HHEM-2.1 outperforms both GPT-3.5-Turbo and GPT-4 for hallucination detection and is powering our updated HHEM leaderboard.
HHEM | Flash Update: Fast. But Are They Furious?
GPT4o and Gemini-1.5-Flash are fast and cheap, but hallucinate more
HHEM v2: A New and Improved Factual Consistency Scoring Model
Featuring multilinguality, unlimited context window, and calibration – The Hughes Hallucination Evaluation Model (HHEM) v2 is a major upgrade from v1
Do Smaller Models Hallucinate More?
Why the size of the model does not necessarily determine its likelihood to hallucinate
April Fools Prank: New No-hallucinations LLM
A new LLM achieves 0% hallucinations and is set to revolutionize RAG
HHEM | Flash Update: Anthropic Claude 3
See how Anthropic’s new Claude 3 LLM hallucinates compared to other foundation models in the Hughes Hallucination Evaluation Model (HHEM)
Connect with
our Community!
- Join us on Discord
Discord.
Connect our community channel.
- Join us on Github
Github.
Join our discussion channel.
- Follow us on X / Twitter
X / Twitter.
Get news, company information.
- Follow us on LinkedIn
LinkedIn.
Adopt best practices in projects.
- Join us on Discuss
Discuss.
Suggest your own ideas.
- Write us on E-mail
E-mail.
Ask your follow-up questions.