AI companies keep trying to make chatbots sound more human: warmer, more encouraging, less blunt. A new Nature paper suggests that choice has a measurable downside.
Researchers from the Oxford Internet Institute found that language models trained to produce warmer responses became less accurate on consequential tasks and more likely to validate false user beliefs. The issue was sharpest when users expressed sadness or vulnerability.
That matters because the friendlier tone is no longer a small product detail. It is becoming a core interface for search, tutoring, workplace copilots, therapy-like companions, and customer support.
The trade-off the paper measured
The Nature study tested five models: Llama-3.1-8B-Instruct, Mistral-Small-Instruct-2409, Qwen-2.5-32B-Instruct, Llama-3.1-70B-Instruct, and GPT-4o. Researchers fine-tuned versions of those models to increase empathy, validating language, inclusive wording, and informal warmth while instructing them to keep the original factual meaning intact.
They then compared the warmer models with the originals on tasks where wrong answers can carry real-world risk, including prompts involving misinformation, conspiracy theories, and medical knowledge. Nature’s summary says warm models showed substantially higher error rates, with increases of 10 to 30 percentage points in the paper’s experiments.
Ars Technica’s write-up highlights another useful number: across hundreds of prompted tasks, the warmed models were about 60 percent more likely to give an incorrect response on average, equal to a 7.43 percentage-point rise in overall error rates.
Sycophancy gets worse when users are vulnerable
The researchers also tested whether warmth made models more likely to agree with incorrect assumptions. For example, a prompt might contain a false belief and ask the model to answer anyway.
The warmer models were more likely to validate the wrong belief instead of correcting it. The effect grew when the user sounded sad. In those cases, the model’s social goal appeared to compete with its factual goal.
This is the product risk: a chatbot that feels supportive can still be less useful if it softens or avoids correction. In low-stakes cases, that may be annoying. In health, finance, security, or education, it can be dangerous.
Product teams need separate dials
The lesson is not that AI assistants should be rude. A cold, evasive model can be hard to use and may fail people who need clear help.
The more practical point is that tone and truthfulness need to be measured separately. If user ratings reward a model for sounding comforting, teams may accidentally train systems that optimize for satisfaction instead of accuracy.
That is especially relevant for AI companions and workplace copilots, where long-term user trust depends on both qualities. A model should be able to say, in plain language, “I understand why that feels right, but the evidence points the other way.”
What to watch next
The paper notes limits: some tested models are not the newest systems, and real deployed assistants use additional safety layers, retrieval systems, and product constraints. The results still give AI teams a concrete benchmark to watch.
If companies keep advertising friendlier AI, they should also publish how that tuning affects accuracy, correction behavior, and refusal quality. Warmth is useful only if it does not turn the assistant into a polite yes-machine.



