Skip to content
Data & AI

LLM & RAG: From Chatbot to Agent

LLM & RAG: From Chatbot to Agent" provides participants with the conceptual foundation and technical know-how to apply Large Language Models effectively through Retrieval-Augmented Generation techniques. This intensive 5-day course is designed for technical professionals, data scientists, ML engineers, and technical leaders who want to understand how LLMs work, how retrieval pipelines enhance accuracy and factual grounding, and how to deploy such systems responsibly in production environments.

The course bridges the gap between theory and practice, helping learners design, build, and evaluate RAG architectures for real-world use cases. Through a combination of conceptual lectures, technical deep-dives, and hands-on implementation exercises, participants will gain practical experience building three progressively sophisticated systems: a basic LLM chatbot, a functional RAG pipeline, and an optimized production-ready RAG application.

What You'll Learn

Day 1: Foundation and First Implementation Participants begin with a thorough understanding of Large Language Model fundamentals, including transformer architecture, training processes, tokenization, and embeddings. You'll explore how LLMs generate text through sampling strategies and learn prompt engineering techniques to optimize responses. The afternoon session focuses on practical implementation: connecting to LLM APIs, working with open-source models, and building your first functional chatbot with optimized latency and response quality.

Day 2: Retrieval-Augmented Generation The second day addresses the critical limitations of standalone LLMs, hallucinations, knowledge cutoff dates, and lack of domain-specific knowledge, and introduces RAG as the solution. Participants will master information retrieval fundamentals including keyword search (TF-IDF, BM25), semantic search with embeddings, hybrid search techniques, and metadata filtering. You'll learn to work with vector databases (FAISS, Pinecone, Chroma, Weaviate) and implement comprehensive evaluation strategies. The hands-on session guides you through building a complete RAG system, from knowledge base construction to chatbot integration.

Day 3: Advanced RAG Techniques The third day explores advanced optimization techniques that separate prototype systems from production-ready applications. Topics include approximate nearest neighbors (ANN) algorithms for scaling, advanced chunking strategies, query parsing and rewriting, cross-encoders, and reranking methods. Participants will also learn about agentic RAG systems that can use tools and make autonomous decisions, and advanced prompt engineering techniques including few-shot learning and chain-of-thought prompting. The afternoon focuses on evaluation strategies with component-level testing, end-to-end evaluation, and cost/latency optimization.

 Day 4: Production RAG & Deployment The fourth day focuses on taking RAG systems from prototype to production. Participants will explore multimodal RAG for processing text, images, and documents, and learn critical production considerations including logging, monitoring, quantization for deployment, security, data privacy, and bias mitigation. The afternoon covers the strategic choice between RAG and fine-tuning approaches, and culminates with a hands-on session to improve and productionize the RAG chatbot built in previous days.

Day 5: From Chatbot to Agent The final day introduces the paradigm shift from passive chatbots to autonomous AI agents. Participants will explore function calling, tool use patterns, and the Model Context Protocol (MCP) as the open standard for connecting LLMs to external tools and data sources. The morning covers MCP architecture, connecting to existing MCP servers, and agent orchestration patterns. The afternoon is dedicated to hands-on building: participants will create their own MCP server from scratch, implement custom tools, and transform their RAG chatbot into a fully functional AI agent.

 

Content

Day 1: Understanding LLMs

Morning Session (4H) - Foundations of Large Language Models

 

  • What is AI and the evolution to LLMs
  • Deep dive into Transformer architecture
  • LLM training processes and capabilities
  • Tokenization, embeddings, and contextual understanding
  • Sampling strategies and quantization
  • Prompt Engineering fundamentals

Afternoon Session (4H) - Construct Your First Chatbot

  • Using LLMs through APIs (setup, cost considerations, key parameters)
  • Using open-source LLMs (setup, quantization, serving)
  • Connect the model to a chat interface
  • Optimize latency and response quality
  • Hands-on: Build a basic chatbot

Day 2: Basics of RAG

Morning Session (4H) - Overview of RAG Architecture

  • Limitations of LLMs and the motivation for RAG
  • RAG Architecture overview (retriever, vector DB, generator)
  • Vector store basics and embedding models
  • Information Retrieval Fundamentals:

o    Keyword search: TF-IDF and BM25

o    Semantic search principles

o    Hybrid Search (combining methods)

o    Metadata filtering

  • RAG evaluation strategies (precision, recall, F1)
  • Security considerations
  • Handling hallucinations

Afternoon Session (4H) - Implementing a Small RAG

  • Choosing the right tools (vector stores: FAISS, Pinecone, Chroma, Weaviate)
  • Constructing and organizing your first knowledge base
  • Add RAG to your initial chatbot
  • Evaluate the first prototype
  • Hands-on: Build end-to-end RAG system

Day 3: Advanced RAG

Morning Session (4H) - Advanced Techniques

Vector Database Deep Dive:

  • Approximate Nearest Neighbors (ANN) algorithms
  • Scaling considerations

Optimization Strategies: 

  • Chunking strategies and advanced techniques
  • Query parsing and rewriting
  • Cross-encoders and ColBERT
  • Reranking techniques

Advanced Prompt Engineering:  

  • Building effective augmented prompts
  • Few-shot learning techniques
  • Chain-of-thought prompting

Agentic RAG:

  • Tool integration (Tools and MCP)
  • Autonomous decision-making

Afternoon Session (3H) - Evaluation & Optimization

Customized Evaluation:

  • Component-level testing
  • End-to-end evaluation
  • Custom metrics

Cost/Latency vs Response Quality tradeoffs

 Hands-on: Optimize and evaluate the RAG system


Day 4: Production RAG & Deployment

Morning Session (4H) - Multimodal & Production RAG

Multimodal RAG (text, images, documents)

Production Considerations:

  • Logging, monitoring, and observability
  • Quantization for deployment
  • Security and data privacy
  • Bias mitigation

RAG vs Fine-tuning: When to use each approach

Afternoon Session (3H) - Deployment & Optimization

EU AI Act compliance for AI systems

  • Performance vs. cost optimization strategies
  • Caching and scaling strategies
  • Hands-on: Improve and productionize the RAG chatbot

Day 5: From Chatbot to Agent

Morning Session (4H) - Agentic AI & MCP Fundamentals

  • The paradigm shift: from chatbot to autonomous agent
  • Function calling and tool use patterns
  • Model Context Protocol (MCP):

o    MCP architecture and the open standard

o    Connecting to existing MCP servers

o    Resources, tools, and prompts

  • Agent orchestration patterns (ReAct, planning, reflection)

Afternoon Session (3H) - Building Your MCP Server & Agent

  • Setting up an MCP server from scratch
  • Implementing custom tools and resources
  • Connecting the RAG chatbot to your MCP server
  • Security and guardrails for tool-using agents
  • Hands-on: Transform the RAG chatbot into an MCP-powered agent


Learning Outcomes

• Understand the principles and architecture of Large Language Models (LLMs) and RAG systems

• Implement a basic RAG pipeline integrating an LLM and a retrieval mechanism

• Evaluate and fine-tune RAG performance for factual accuracy and user intent

• Identify opportunities for responsible adoption of RAG solutions within organizations

• Explain business and ethical implications of AI-driven retrieval systems

Training Method

This course combines theoretical instruction with hands-on practical exercises, demonstrations, and discussions. Participants will engage in small project work to build and test RAG-enabled applications using open-source tools.

Certification
Certificate of Participation
Prerequisites

AI practitioners, data scientists, machine learning engineers, and solution architects seeking both conceptual understanding and hands-on experience with Large Language Models (LLMs) and Retrieval-Augmented Generation (RAG) systems. Technical managers

interested in the practical use and integration of LLMs in production environments are also welcome.


Planning and location
Session 1
18/05/2026 - Monday
09:00 - 16:00
Session 2
19/05/2026 - Tuesday
09:00 - 16:00
Session 3
20/05/2026 - Wednesday
09:00 - 16:00
Session 4
21/05/2026 - Thursday
09:00 - 16:00
Session 5
22/05/2026 - Friday
09:00 - 16:00
Available Edition(s):

https://www.dlh.lu/web/image/product.template/2613/image_1920?unique=26d331d

This combination does not exist.

120.00 € 120.0 EUR 120.00 €

120.00 €

Not Available For Sale

Your trainer(s) for this course
Alexandre Hotton
See trainer's courses.
Geoffrey Nichil
See trainer's courses.