Introducing the Hallucinations Leaderboard
Generative AI often produces inaccurate or misleading content, leading to potential misinformation. To address this issue, Huggingface has launched the Hallucinations Leaderboard, an initiative aimed at evaluating open source Large Language Models (LLMs) and their tendency to generate hallucinated content. The leaderboard assesses LLMs based on two categories of hallucinations: factuality and faithfulness. Factuality refers to when generated content contradicts real-world facts, while faithfulness occurs when the content deviates from user instructions or established context. The leaderboard utilizes EleutherAI’s Language Model Evaluation Harness to conduct comprehensive evaluations across various tasks, providing an overall performance score for each model.
The Least “Crazy” Models
Preliminary results from the Hallucinations Leaderboard indicate that models like Meow (Based on Solar), Stability AI’s Stable Beluga, and Meta’s LlaMA-2 exhibit fewer hallucinations and are considered among the best. However, certain models based on Mistral LLMs perform exceptionally well in specific tests. It’s important to consider the nature of each user’s task when selecting a model. The higher average score on the leaderboard reflects a lower propensity for hallucinations, indicating greater accuracy and reliability in generating content aligned with factual information and user input.
Limitations and Future Scope
While the Hallucinations Leaderboard provides a comprehensive evaluation of open-source models, closed-source models have not undergone this rigorous testing yet. Commercial models may have proprietary restrictions that prevent their inclusion in the leaderboard’s scoring system. It is crucial to continue developing LLMs towards more accurate and faithful language generation in order to mitigate the risks associated with AI-generated content.
Hot Take: Ensuring Reliable AI Language Generation
AI-generated content has the potential to be highly transformative, but it also poses risks in terms of misinformation. The introduction of the Hallucinations Leaderboard by Huggingface is a commendable step towards evaluating and improving the reliability of open source Large Language Models. By identifying models with lower tendencies for hallucinations, this initiative aims to aid researchers and engineers in selecting more accurate and faithful language generation models. However, it is essential to consider specific tasks and limitations when choosing a model, as well as the need for further evaluation of closed-source models.