Knowledge Base#
The NeMo Guardrails toolkit supports using a set of documents as context for generating bot responses through Retrieval-Augmented Generation (RAG). This guide explains how to configure and use the knowledge base folder.
Overview#
By default, an LLMRails instance supports using documents as context for generating responses. To include documents as part of your knowledge base, place them in the kb folder inside your configuration folder:
.
├── config
│ ├── config.yml
│ ├── kb
│ │ ├── file_1.md
│ │ ├── file_2.md
│ │ └── ...
│ └── rails
│ └── ...
Note
Currently, only the Markdown format is supported.
Document Structure#
Documents in the knowledge base kb folder are automatically processed and indexed for retrieval. The system:
Splits documents into topic chunks based on markdown headers.
Uses the configured embedding model to create vector representations of each chunk.
Stores the embeddings for efficient similarity search.
Example Document#
# Employee Handbook
## Time Off Policy
Employees are eligible for the following time off:
* Vacation: 20 days per year, accrued monthly.
* Sick leave: 15 days per year, accrued monthly.
* Personal days: 5 days per year, accrued monthly.
## Holiday Schedule
Paid holidays include:
* New Year's Day
* Memorial Day
* Independence Day
* Thanksgiving Day
* Christmas Day
Retrieval Process#
When a user query is received, the system:
Computes embeddings for the user query using the configured embedding model.
Performs similarity search against the indexed document chunks.
Retrieves the most relevant chunks based on similarity scores.
Makes the retrieved chunks available as
$relevant_chunksin the context.Uses these chunks as additional context when generating the bot response.
Configuration#
The knowledge base functionality is automatically enabled when documents are present in the kb folder. You can customize the behavior using the knowledge_base section in your config.yml:
knowledge_base:
folder: "kb" # Default folder name
embedding_search_provider:
name: "default"
parameters: {}
Configuration Options#
Option |
Description |
Default |
|---|---|---|
|
The folder from which documents should be loaded |
|
|
The name of the embedding search provider |
|
|
Provider-specific parameters |
|
Embedding Model Configuration#
The knowledge base uses the embedding model configured in the models section of your config.yml:
models:
- type: main
engine: openai
model: gpt-4
- type: embeddings
engine: openai
model: text-embedding-ada-002
For more details on embedding model configuration, refer to Model Configuration.
Alternative Knowledge Base Methods#
There are three ways to configure a knowledge base:
1. Using the kb Folder (Default)#
Place markdown files in the kb folder as described above. This is the simplest approach for static document collections.
2. Using Custom retrieve_relevant_chunks Action#
Implement a custom action to retrieve chunks from external sources:
from nemoguardrails.actions import action
@action()
async def retrieve_relevant_chunks(context: dict, llm: BaseLLM):
"""Custom retrieval from external knowledge base."""
user_message = context.get("last_user_message")
# Implement custom retrieval logic
# For example, query an external vector database
chunks = await query_external_kb(user_message)
return chunks
3. Using Custom EmbeddingSearchProvider#
For advanced use cases, implement a custom embedding search provider:
from nemoguardrails.embeddings.index import EmbeddingsIndex
class CustomEmbeddingSearchProvider(EmbeddingsIndex):
"""Custom embedding search provider."""
async def add_item(self, item: IndexItem):
# Custom indexing logic
pass
async def search(self, text: str, max_results: int) -> List[IndexItem]:
# Custom search logic
pass
For more details, refer to Embedding Search Providers.
Passing Context Directly#
You can also pass relevant context directly when making a generate call:
response = rails.generate(messages=[
{
"role": "context",
"content": {
"relevant_chunks": """
Employees are eligible for the following time off:
* Vacation: 20 days per year, accrued monthly.
* Sick leave: 15 days per year, accrued monthly.
"""
}
},
{
"role": "user",
"content": "How many vacation days do I have per year?"
}
])
Using Knowledge Base in Colang Flows#
You can reference the retrieved chunks in your Colang flows:
import core
import llm
flow main
activate llm continuation
user asked question
$chunks = ..."Summarize the relevant information from the knowledge base."
bot say $chunks
flow user asked question
user said "what" or user said "how" or user said "tell me"
define flow answer question
user ask question
# Use the retrieved knowledge base chunks to answer
bot respond with knowledge
Best Practices#
Organize documents logically: Use clear markdown headers to structure your documents. The system chunks documents based on headers.
Keep chunks focused: Each section should cover a single topic for better retrieval accuracy.
Use descriptive headers: Headers help the system understand the content of each chunk.
Test retrieval quality: Verify that the system retrieves relevant chunks for common user queries.
Monitor embedding model: Ensure your embedding model is appropriate for your document content and user queries.
Complete Example#
Here’s a complete example configuration with a knowledge base:
Directory structure:
.
├── config
│ ├── config.yml
│ ├── kb
│ │ └── company_policy.md
│ └── rails
│ └── main.co
config.yml:
models:
- type: main
engine: openai
model: gpt-4
- type: embeddings
engine: openai
model: text-embedding-ada-002
instructions:
- type: general
content: |
You are a helpful HR assistant. Answer questions based on the
company policy documents provided.
knowledge_base:
folder: "kb"
kb/company_policy.md:
# Company Policy
## Vacation Policy
All full-time employees receive 20 days of paid vacation per year.
Vacation days accrue monthly at a rate of 1.67 days per month.
## Sick Leave
Employees receive 15 days of paid sick leave per year.
Unused sick days do not carry over to the next year.