Building AI Products Users Actually Trust: NLP+ RAG+MCP

Written on 11/08/2025
Bandan Singh

Basics of how NLP, RAG, and MCP can help product managers build AI systems that stay factual, grounded, and trustworthy.

Imagine this: you’re a product manager at a company building an AI-powered customer support chatbot. Your team spent months training it on company policies, product documentation, and FAQs. Launch day arrives, and initially, everything looks promising response times drop by 60%, customer satisfaction scores climb.

Then the complaints start rolling in.

A customer asks about certain refund clauses, and the chatbot confidently explains a policy that doesn’t exist. Another user inquires about payment deadlines and receives incorrect dates that could cost them late fees. Your support team is suddenly drowning in escalations, and trust in your new AI feature is evaporating faster than you can say “product-market fit.”

This isn’t science fiction. It’s happening right now to product teams across industries. The challenge? AI systems, for all their brilliance, have a fundamental flaw: they sometimes fabricate information with unsettling confidence. And as product managers, we’re now responsible for preventing these digital hallucinations from damaging our users’ trust and our company’s reputation.

The good news? There are powerful techniques emerging that can help us build AI products that are not just smart, but reliable and trustworthy.

Three frameworks stand at the forefront:

Natural Language Processing (NLP),

Retrieval-Augmented Generation (RAG),

and Model Context Protocol (MCP).

Understanding how these work and more importantly, how they work together can transform how you approach AI product development.

Consider upgrading your subscription, or access more product resources:

Browse and Access Product Resources

NLP: Teaching Machines to Actually Understand Us

Before we dive into the sophisticated stuff, let’s talk about the foundation: Natural Language Processing. Think of NLP as the basic translation layer between human communication and machine understanding.

At its core, NLP enables computers to read, interpret, and make sense of human language in a valuable way.

It’s what powers the autocomplete in your email, the sentiment analysis in your customer feedback tools, and the chatbots that (hopefully) understand your questions.

How NLP Actually Works

When you type a message or speak to an AI system, NLP breaks down your communication into digestible pieces through several steps:

  • Tokenization splits your text into individual words or phrases turning “I love this product” into [”I”, “love”, “this”, “product”]

  • Semantic analysis figures out what you actually mean, understanding that “this product rocks” and “this product is amazing” convey similar sentiments

  • Part-of-speech tagging identifies whether words are nouns, verbs, or adjectives to understand sentence structure

  • Named entity recognition picks out specific items like company names, dates, or product features

For product managers, NLP is incredibly practical. It can analyze thousands of customer reviews to identify common pain points, power conversational interfaces that feel natural, and extract insights from unstructured feedback that would take humans weeks to process manually.

Real-World Impact

DoorDash uses NLP to power their delivery support chatbot, condensing conversations to grasp core issues and searching their knowledge base for relevant solutions. The result? Faster resolution times and more satisfied delivery partners.

Bell Canada leveraged NLP-based systems to enhance knowledge management, ensuring employees have access to up-to-date company policies through intelligent document processing and indexing.

Airbnb deployed NLP to localize content across 190 countries, using sentiment analysis to ensure reviews in Japanese carry the same weight and emotional tone as those in Spanish or English. After integrating NLP into their multilingual strategy, they saw a 15% boost in international bookings.

The Shortfall You Can’t Ignore

But here’s where NLP alone falls short: it understands language patterns, but it doesn’t inherently know facts. An NLP system can beautifully parse the question “What’s our refund policy for orders over $500?” but if it hasn’t been explicitly trained on that specific policy detail, it might generate a plausible-sounding but completely incorrect answer. This is where the conversation gets interesting.

RAG: Giving Your AI a Research Library

Retrieval-Augmented Generation is the solution to one of AI’s most embarrassing problems: confidently spouting nonsense when it doesn’t actually know the answer.

Think of RAG this way: instead of forcing an AI to memorize everything (which is expensive, time-consuming, and quickly becomes outdated), you give it access to a well-organized library of current, accurate information. When a user asks a question, the AI first looks up relevant documents in this library, then uses that retrieved information to formulate its response.

How RAG Works in Practice

The RAG process has three main stages:

  • Document preparation: Your company’s policies, product docs, customer data, and other knowledge sources are broken into chunks and converted into numerical representations called embeddings

  • Retrieval: When a user asks a question, the system converts that question into the same numerical format and searches the database for the most semantically similar documents

  • Generation: The retrieved documents are fed to the language model as context, which then generates a response grounded in that actual information rather than making something up

The beauty of this approach is that you can update your knowledge base without retraining your entire AI model. Add a new policy document, update product specs, or refresh pricing information, and your AI immediately has access to current data.

Companies Getting This Right

LinkedIn built a RAG-based system for their customer service team that reduced median per-issue resolution time by 28.6%. By retrieving relevant past cases and knowledge base articles, support agents get contextually appropriate responses tailored to each specific query.

Thomson Reuters uses RAG to help customer support executives quickly access relevant information from curated databases through a conversational interface. Support staff can find accurate, up-to-date answers while maintaining natural interactions with customers.

Vimeo developed a RAG-powered chatbot that lets users converse with videos summarizing content, linking to key moments, and suggesting related questions by retrieving relevant context from video transcripts.

Ramp, a fintech company, built a RAG-based industry classification system that enabled migration to a standardized framework, using retrieval to match customer businesses with the most accurate classification codes.

Netflix leverages RAG principles to combine user history with real-time data for recommendations, analyzing viewing patterns and social media buzz to suggest trending shows that feel both personal and current.

The Pitfalls Product Managers Must Avoid

RAG isn’t a magic bullet, and implementing it comes with real challenges:

  • Poor retrieval quality is the silent killer. Even the most powerful language model generates bad answers if it retrieves irrelevant documents. You need carefully selected embedding models and ranking strategies to surface the right context.

  • Context window limitations mean you can risk truncated sources or diluted responses when you stuff too much retrieved content into your prompt.

  • High latency becomes a problem as your system spans vector search, rerankers, and large language models. Without optimization, response times slip, especially as indexes grow.

  • Data chunking decisions matter more than most teams realize. Split documents too small and you lose context; too large and retrieval precision suffers.

  • Scalability planning often gets overlooked in prototypes. Systems that work fine with small datasets can stumble at scale with memory spikes and throughput collapse.

  • Inadequate evaluation is perhaps the biggest mistake. Generic text metrics don’t catch factual errors or missing evidence. You need continuous monitoring with real-world queries, not just demo questions.

MCP: The Protocol That Makes AI Systems Actually Work Together

While RAG solves the knowledge problem, Model Context Protocol addresses a different challenge: how do we enable AI systems to interact with external tools, databases, and services in a standardized, secure way?

MCP is like establishing a universal language for AI applications to communicate with external systems. Before MCP, every new integration required custom code connecting your AI to Slack required one approach, Google Drive another, your CRM yet another. MCP standardizes these connections.

The Architecture That Makes It Work

MCP uses a client-server architecture with three key components:

  • The MCP host is your AI application (like Claude Desktop or an IDE) where users interact

  • The MCP client sits within the host, translating requests between the AI and external systems

  • MCP servers are external services providing data or capabilities databases, APIs, file systems, CRM tools

The protocol uses JSON-RPC for lightweight, stateless communication, allowing AI systems to discover available tools, invoke functions, and receive structured responses.

What Makes MCP Different

MCP offers three core “primitives” that define how AI interacts with the world:

  • Resources provide contextual data managed by the application file contents, database records, Git history

  • Tools are executable functions the AI can invoke API calls, file operations, database updates

  • Prompts are pre-defined templates triggered by user actions slash commands or menu selections

This creates a hierarchy of control: users control prompts, applications control resources, and the AI model itself decides when to use tools.

Real Companies Seeing Results

Block (the company behind Square and Cash App) built an internal AI agent called Goose running on MCP architecture. They created all MCP servers in-house for complete security control, integrating with Snowflake, Jira, Slack, Google Drive, and internal APIs. Thousands of Block employees now use Goose, cutting up to 75% of time spent on daily engineering tasks.

Bloomberg’s engineering team adopted MCP to help AI developers reduce time-to-production from days to minutes. The protocol connects AI researchers to an ever-growing toolset, creating a flywheel where tools and agents interact and reinforce each other.

Implementation Challenges to Watch For

MCP implementations face their own hurdles:

  • Security considerations are paramount. You need robust authentication, least-privilege access controls, and careful tool scoping to prevent unrestricted access to sensitive systems.

  • Tool validation becomes critical. Every invocation should be validated and logged to prevent unintended actions.

  • Context management requires careful orchestration of what data the AI can access and how it uses that information.

  • Capability negotiation between client and server must ensure compatibility across protocol versions.

The Power Move: Combining RAG and MCP

Here’s where it gets really interesting for product managers. RAG and MCP aren’t competing approaches they’re complementary techniques that solve different problems. When you combine them strategically, you unlock AI capabilities that neither can achieve alone.

Think of it this way: RAG gives your AI accurate knowledge to draw from, while MCP gives it the ability to take actions in the real world.

When to Use What

  • Use RAG when you need to ground responses in static, authoritative documents product policies, technical manuals, research papers, historical data. RAG excels at explaining, summarizing, and answering questions based on established knowledge.

  • Use MCP when you need real-time actions querying live databases, checking current system status, executing workflows, updating records. MCP shines when your AI needs to interact with external systems or perform transactional operations.

  • Combine both when you need comprehensive solutions that require both knowledge and action.

A Hybrid Workflow in Action

Imagine you’re building an e-commerce customer service system. A customer asks: “What’s the return policy for my order, and has my refund been processed yet?”

With RAG alone, you could explain the return policy by retrieving the relevant documentation. But you couldn’t check their actual refund status.

With MCP alone, you could query the payment system to check refund status. But you couldn’t explain the policy that governs it.

Combined approach:

  1. RAG retrieves the return policy documents to understand the rules

  2. MCP makes a tool call to your order management API to check the specific customer’s refund status

  3. RAG synthesizes both the policy and the current status into a comprehensive, actionable response

This hybrid workflow provides both traceable information and real-time data, creating AI systems that are both knowledgeable and capable.

Strategic Implementation for Product Managers

  • For customer support automation, use RAG to retrieve relevant help articles and MCP to actually update tickets, trigger workflows, or check account status.

  • In financial services, RAG can pull historical policy information while MCP queries real-time market data or executes transactions.

  • For internal tools, RAG provides context from documentation while MCP integrates with project management platforms, databases, and communication tools.

  • When building procurement systems, RAG retrieves past vendor contracts and performance data while MCP checks real-time inventory levels and processes orders.

The Implementation Roadmap

Don’t try to boil the ocean. Start with a phased approach:

  • Phase 1 focuses on basic RAG with a single document type. Get document chunking right, select appropriate embedding models, and validate retrieval quality with real queries.

  • Phase 2 adds MCP for simple tool integrations. Start with read-only operations to external systems before enabling write capabilities.

  • Phase 3 combines both approaches in hybrid workflows. Build orchestration layers that intelligently decide when to retrieve knowledge versus when to take action.

  • Phase 4 scales with monitoring, continuous evaluation, and optimization. Implement feedback loops, track hallucination rates, and refine based on real user interactions.

What This Means for You as a Product Manager

The product managers who will thrive in the AI era aren’t necessarily those who can code or train models. They’re the ones who understand how to architect systems that combine the right techniques to solve real user problems while managing inherent risks.

When evaluating AI features for your product, ask yourself:

  • Does this need to reference factual, authoritative information? Consider RAG to ground responses in verified sources rather than risking hallucinations.

  • Does this need to take actions in external systems or access real-time data? Consider MCP to enable tool use and integrations in a standardized way.

  • Does this need both knowledge and action? Design a hybrid architecture that leverages both approaches strategically.

The teams getting this right aren’t just building AI features they’re building trust. They’re creating systems where users can verify information, where responses trace back to source documents, where actions are scoped appropriately and validated carefully.