Multi-turn conversation diagram
All essays
Conversation design

Why LLMs Get Lost in Multi-Turn Conversations

Research from Microsoft and Salesforce reveals a critical blind spot in even the most advanced AI models: a 39% performance drop in natural back-and-forth exchanges. Here's what it means for designers and strategists.

Ben Farrell May 2025 2 min read

In the fast-moving world of AI, Large Language Models like GPT-4, Claude, and Gemini are being hailed as game-changers for customer experience, automation, and digital conversation design. But new research from Microsoft and Salesforce reveals a critical blind spot that should have every conversation designer, prompt engineer, and AI strategist paying attention.

The Hidden Weakness in LLMs: Multi-Turn Conversations

The study found that even the most advanced LLMs see a 39% drop in performance when handling multi-turn, underspecified conversations — the kind of natural back-and-forth exchanges we see every day in customer support, virtual assistants, and AI chatbots.

The issue isn't a lack of intelligence. It's unreliability. Models start strong but get lost in conversation as turns progress. The more a conversation evolves, the more assumptions the LLM makes — and the harder it becomes for the model to recover.

Key findings from the research:

  • LLMs prematurely "lock in" assumptions based on limited early context
  • They over-rely on their previous outputs, even when wrong
  • They struggle to incorporate new information mid-conversation
  • Reliability drops by over 100% in multi-turn settings compared to single-turn prompts

The issue isn't a lack of intelligence. It's unreliability.

What This Means for AI Conversation Designers

This is a wake-up call for anyone working in AI chatbot design, LLM implementation, or generative AI applications. If your LLM-based solution depends on iterative clarification from users — and most real-world use cases do — you're likely not getting the accuracy or consistency you expect.

Reduce ambiguity early. Design prompts and flows that front-load key context. The more specific a prompt is upfront, the less chance the LLM will wander off track.

Use recap and context reinforcement strategies. Think of your chatbot as having short-term memory issues. Reinforce prior inputs either through summarisation ("Let me recap what I understand…") or repeated context in each turn — a "snowball" method that carries forward what matters.

Advocate for reliability as a core model metric. Current benchmarks often measure aptitude — what the model can do — not reliability — what it does consistently. Push for evaluation frameworks that consider multi-turn robustness.

Educate your stakeholders. LLM hallucination and conversation drift are real. Help business stakeholders understand the risks of underspecified conversations and the value of good UX and prompt engineering.

Why This Matters for the Future of Conversational AI

As we move toward more human-like AI assistants, voicebots, and digital support agents, we need models that can genuinely follow the thread of a conversation — not just perform well on neatly packaged prompts.

If you're designing AI-powered solutions that rely on natural language interaction, it's not just about the model you choose. It's about how you structure the conversation.

The gap between what LLMs can do in demo conditions and what they deliver in the wild is still significant. Understanding where they break down isn't pessimism — it's the foundation of good design.