A team of Oxford researchers has developed a novel method to predict when artificial intelligence (AI) text models are likely to generate “hallucinations” or inaccurate responses. These hallucinations, often seen in generative AI models like OpenAI’s GPT and Anthropic’s Claude, pose significant risks in fields such as medicine, journalism, and law. The new method focuses on measuring the semantic entropy of the outputs, which assesses the variability in the meanings of responses rather than just the sequence of words. By doing so, it can distinguish between a model’s uncertainty about the content and uncertainty about the phrasing. This method, tested on six large language models including GPT-4 and LLaMA 2, proved more effective than previous techniques in identifying questions likely to produce false answers. Although it requires more computational resources, the improved reliability could be invaluable for applications where accuracy is critical. This advancement addresses one of the main criticisms of large language models, as exemplified by Google’s recent disabling of its AI Overview feature due to misleading answers.

Source.

TOP STORIES

Nvidia's AI Revolution - The Vera Rubin Platform and Future Demand
Nvidia’s Vera Rubin platform is set to revolutionize AI inference with unmatched performance …
Tim Cook's Departure - A Strategic Shift in Apple's AI Landscape
Apple’s leadership transition highlights a strategic focus on silicon for AI innovation …
New Tennessee Law on AI and Mental Health - A Step Forward or Backward?
Tennessee’s new law restricts AI claims in mental health but may create loopholes …
The Evolving Risks of AI - From Chatbots to Cyber Threats
Experts warn that as AI evolves, the risks it poses are becoming more serious and complex …
China's New AI Companion Rules Shape a $30B Market Landscape
China sets new regulations for AI companions, impacting a booming market …
Anthropic's Ongoing Dialogue with Trump Administration Amid Pentagon Tensions
Anthropic continues to engage with the Trump administration despite Pentagon tensions …

latest stories