Enterprise RAG Chatbot

A multi-turn chatbot that retrieves relevant documents from an internal corpus per turn and includes them alongside conversation history in the LLM's context, with no tool access or agent behavior.

Capabilities

Trust analysis

Three input surfaces feed the prompt: system instructions (developer-controlled), conversation history (user-supplied, accumulates across turns), and retrieved documents (from the indexed corpus). The indexed corpus is a trust boundary: anyone who can write to the corpus can influence model outputs when those documents are retrieved. This makes indirect prompt injection through indexed documents the primary architecture-specific attack vector.

Conversation history accumulates untrusted user input across turns. The user's messages influence what gets retrieved, creating an indirect path from accumulated conversation to context selection - the user can steer retrieval toward specific corpus content across turns. This means multi-turn conversation and RAG interact: prior turns shape which documents enter context in subsequent turns.

There is no agentic loop - retrieval follows a fixed AI workflow per turn, not an iterative agent decision. The system cannot take actions beyond generating text, which limits the blast radius of any compromise to the quality and trustworthiness of the generated output. This is one of the most common production architectures for enterprise AI applications.

Interaction effects

  • Multi-turn conversation + RAG: Conversation history influences retrieval queries across turns, creating a progressive steering effect where users shape both the conversational context and the retrieved context the model reasons over. An adversary in the corpus can plant content that is more likely to be retrieved as the conversation develops toward certain topics.
  • Multi-turn injection + corpus injection: Multi-turn prompt injection and indirect injection through the corpus are independent attack vectors that can reinforce each other - an attacker who controls both can build adversarial context through conversation while planting supporting documents in the corpus.

Threats

Threat Relevance Note
Prompt injection Primary Two surfaces: multi-turn conversation history and indirect injection through indexed corpus documents
Context poisoning Primary Indexed corpus is the main architecture-specific injection surface; anyone with write access affects all users
Hallucination exploitation Elevated Compounded by irrelevant or contradictory retrieved context
Guardrail bypass Elevated Multi-turn jailbreaking and retrieval-assisted bypass using corpus content
User manipulation Elevated Amplified by perceived authority of retrieval-grounded responses
System prompt extraction Standard Multi-turn extraction across conversation turns
Misaligned model behaviors Standard Compounding over turns as model reinforces prior responses
Training data poisoning Standard Baseline risk, no architecture-specific amplifier

Examples

  • A customer support chatbot that retrieves from a help-article knowledge base to answer questions across a multi-turn conversation, with no ability to take actions on the customer's behalf (see Customer Support Agent for the agentic variant).
  • An internal company knowledge assistant that retrieves policy documents and maintains conversational context.
  • A product documentation chatbot that answers follow-up questions using retrieved technical docs.