Configure environment variables for ROCm-RAG#
2026-04-28
3 min read time
You can configure both extraction and retrieval parameters by setting environment variables for the Docker container in ROCm-RAG installation. There are three ways to set environment variables:
Start with default.env as a base.
Modify the variables as needed and provide the
.envfile when running the container:docker run --env-file <your env file> ...
Set variables individually when starting the container:
docker run -e VAR1=value1 -e VAR2=value2 ...
Export variables inside the container when running in interactive mode:
export VAR1=value1
export VAR2=value2
Environment variable reference#
The following tables list the configurable environment variables for ROCm-RAG.
Workspace and storage variables
Variable |
Description |
|---|---|
|
ROCm-RAG workspace directory |
|
Directory to save page-level hash |
|
File to save list of scraped URLs |
Extraction parameters
Variable |
Description |
|---|---|
|
Extraction RAG framework ( |
|
Haystack pipeline server port |
|
LangGraph server port |
|
Embedder model |
|
Embedder API base URL |
|
Embedder API port |
|
Embedder model max token limit |
|
Weaviate DB API base URL |
|
Weaviate DB API port |
|
Weaviate classname |
|
Wait time for vector DB server to be ready |
|
Wait time for embedder server to be ready |
|
Tensor parallelism for embedder |
|
List of visible GPUs when deploying embedder model |
|
Start URL for scraping |
|
List of supported URL extensions to scrape |
|
List of regex filters for selecting valid pages to scrape |
|
List of regex filters for identifying pages that require human verification |
|
List of regex filters for identifying not found pages |
|
Enable limit on the maximum number of pages to scrape |
|
Maximum number of pages to scrape |
|
Maximum number of tokens for SemanticChunkMerger |
|
Similarity threshold for SemanticChunkMerger to merge |
Retrieval parameters
Variable |
Description |
|---|---|
|
Retrieval RAG framework ( |
|
Deploy example LLM inference server inside this Docker |
|
LLM API base URL |
|
LLM API port |
|
LLM model |
|
Tensor parallelism |
|
Visible GPUs, for example, LLM |
|
Certainty threshold for retrieval |
|
Top K retrieved documents for Haystack retrieval pipeline |
|
Top K retrieved documents for LangGraph retrieval pipeline |