RAG (Retrieval-Augmented Generation)
Overview
Architecture
User Query
│
▼
┌─────────────────── LlmTask.executeTask() ───────────────────┐
│ │
│ 1. Extract user input from conversation memory │
│ 2. RagContextProvider.retrieveContext() │
│ ├── WorkflowTraversal.discoverConfigs() → find RAG steps │
│ ├── Match KBs (explicit refs or auto-discover all) │
│ ├── EmbeddingModelFactory → cached embedding model │
│ ├── EmbeddingStoreFactory → cached vector store │
│ ├── EmbeddingStoreContentRetriever → similarity search │
│ └── Store audit trace in conversation memory │
│ 3. Inject context: systemMessage += "## Relevant Context" │
│ 4. Build chat messages and call LLM │
│ │
└──────────────────────────────────────────────────────────────┘Configuration
RagConfiguration (Knowledge Base)
Field
Default
Description
LLM Task RAG Configuration
Option 1: Explicit Knowledge Base References
Option 2: Auto-Discovery
Option 3: httpCall RAG (Phase 8c-0)
REST API
Configuration Management
Method
Path
Description
Document Ingestion
Method
Path
Description
Observability
Memory Key
Content
Embedding Providers
Provider
Default Model
Required Parameters
Notes
Vector Stores
Store Type
Required Parameters
Notes
Status
Future Enhancements
Last updated
Was this helpful?