This document explains the main configuration options available in ClawRAG.
All configuration is done through the .env file. Here are the most important settings:
PORT: External port (default: 8080)DOCS_DIR: Directory for document ingestion (default: ./data/docs)
LLM_PROVIDER: Choose between ollama, openai, anthropic, gemini, openai_compatible (default: ollama)LLM_MODEL: The model to use (default: llama3:latest)
EMBEDDING_PROVIDER: Usually matches LLM_PROVIDER (default: ollama)EMBEDDING_MODEL: Embedding model (default: nomic-embed-text)
CHUNK_SIZE: Size of text chunks (default: 512)CHUNK_OVERLAP: Overlap between chunks (default: 128)INGEST_BATCH_SIZE: Batch size for ingestion (default: 10)
DEBUG: Enable debug logging (default: false)LOG_LEVEL: Logging level (default: INFO)
Uses the built-in Ollama container. No additional configuration needed.
Set:
LLM_PROVIDER=openai_compatible
OPENAI_BASE_URL=http://host.docker.internal:1234/v1
Set the appropriate API key in the docker-compose.yml file or environment.
- The context window is limited to 8192 tokens to prevent OOM errors
- For 8GB VRAM systems, use smaller models like llama3.2 (3b) or llama3 (8b)
- Smaller
CHUNK_SIZEvalues provide more granular search results - Larger
CHUNK_OVERLAPvalues help maintain context across chunks - Adjust based on your document types and search requirements
- If ingestion fails, check that your
DOCS_DIRis correctly mounted - If search returns no results, verify that embeddings are working properly
- For LLM connectivity issues, ensure the provider is properly configured
Enable DEBUG=true to get more detailed logs for troubleshooting.