












OpenAI details CriticGPT, a model based on GPT-4 to catch errors in ChatGPT's code output, assisting human trainers in assessing and spotting errors (@willknight / Wired) https://t.co/8Sh6QwF0SM 📫 Subscribe: https://t.co/OyWeKSRpIM https://t.co/wXACZeWYBj
$MSFT | OpenAI Introduces CriticGPT to Enhance Code Accuracy OpenAI has trained a model, CriticGPT, based on GPT-4, to catch errors in ChatGPT's code output. The company is beginning work to integrate CriticGPT-like models into their RLHF labeling pipeline, providing explicit AI…
OpenAI is applying RLHF to GPT-4 using CriticGPT CriticGPT helps trainers to write more comprehensive critiques than they do without help while producing fewer hallucinations than critiques from the model alone. What about GPT-4o code? https://t.co/0EPzJNWBBp https://t.co/14fwtrWqe7

Researchers have developed an algorithm to detect when AI systems, particularly large language models (LLMs) such as ChatGPT and Gemini, are hallucinating. This new method, based on calculating semantic entropy, measures the uncertainty in meaning to improve the reliability of AI outputs. The approach, known as Semantic Entropy Probes (SEPs), involves repeated LLM generation and consistency checks. Oxford researchers are among those advancing this field, highlighting the importance of effective detection and evaluation of AI hallucinations. Researchers also suggest that chatbots policing each other can correct some hallucinations. Additionally, OpenAI has introduced CriticGPT, a model based on GPT-4, to catch errors in ChatGPT's code output, enhancing the accuracy of AI-generated content. The University of Oxford has contributed significantly to this research, emphasizing the robustness and cost-effectiveness of this detection method.