Leading AI chatbots are failing dramatically at financial advice, demonstrating how conversational AI systems can present dangerously incorrect information with an authoritative tone. A new study from the Walter Bradley Center for Natural and Artificial Intelligence tested four top large language models with basic financial questions, revealing significant deficiencies in mathematical accuracy and financial reasoning that could mislead users who trust these systems for important financial decisions.
The big picture: AI researchers tested ChatGPT-4o, DeepSeek-V2, Grok 3 Beta, and Gemini 2 with 12 finance questions, finding all models performed poorly despite their confident conversational style.
Key details: The researchers found these AI systems offer a “reassuring illusion of human-like intelligence” reinforced by a conversational style that can mask fundamental errors.
Why this matters: The friendly, confident tone of AI chatbots creates a particularly dangerous dynamic where users might trust these systems with important financial decisions.