I’m currently building an AI tutoring app, and I asked it to teach me about the Carnot Cycle – something which I’m vaguely familiar with, but never felt like I truly understood. I came away very impressed by its explanatory abilities. But when I later tried to get it to do Olympiad math problems – something that I have extensive past experience with, and is explicitly trained for by OpenAI + Anthropic – I could see that it was spouting absolute nonsense, despite various claims that frontier models are capable of handling math olympiad problems.
I think what is most surprising to me about AI is how frequently I go from having a seemingly good, trust-building interaction with the AI, to having another incoherent trust-destroying interaction in a different domain. Maybe what’s happening is that LLMs talk so smoothly that they trigger some underlying ‘deference to authority’ instinct in humans – the textual equivalent of charisma. Whatever it is, it’s hard to ignore the effect even when you know it’s happening.
— “LLM Shibboleths determine AI effectiveness,” Brian Kihoon Lee
Discover more from Fluid Imagination
Subscribe to get the latest posts sent to your email.