"...even state-of-the-art models still produce logical mistakes, often called hallucinations. Mitigating hallucinations is a critical step towards building aligned AGI." https://openai.com/research/improving-mathematical-reasoning-with-process-supervision #AI #ArtificialIntelligence #Maths #Logic
Source: https://twitter.com/AlexKontorovich/status/1664272620370694145