-
Notifications
You must be signed in to change notification settings - Fork 20
Expand file tree
/
Copy path.env.example
More file actions
90 lines (70 loc) · 3.68 KB
/
.env.example
File metadata and controls
90 lines (70 loc) · 3.68 KB
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
UI_LANG="zh"
#######################################################################
################### #####################
################### Model Setting #####################
################### #####################
#######################################################################
HF_ENDPOINT=https://hf-mirror.com
### Chat model
API_KEY=xxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxx
LLM_MODEL="qwen3-coder-plus"
LLM_BASE_URL="https://dashscope.aliyuncs.com/compatible-mode/v1"
# =============================================================================
# Embedding Model Configuration
# =============================================================================
# EMBEDDED_TYPE: Embedding model type, available options:
# - default: Use built-in pyseekdb all-MiniLM-L6-v2 model (no additional config needed)
# - ollama: Use Ollama embedding service (requires EMBEDDED_API_KEY, EMBEDDED_LLM_MODEL, EMBEDDED_LLM_BASE_URL)
# - openai_embedding: Use OpenAI-compatible embedding API (requires EMBEDDED_API_KEY, EMBEDDED_LLM_MODEL, EMBEDDED_LLM_BASE_URL)
EMBEDDED_TYPE=default
# Vector embedding dimension (must match your embedding model's output dimension)
EMBEDDED_DIMENSION=384
# EMBEDDED_API_KEY: API key for embedding service
# - Required for: ollama, openai_embedding
# - Not required for: default
EMBEDDED_API_KEY=
# EMBEDDED_LLM_MODEL: Embedding model name
# - For ollama: model name (e.g., nomic-embed-text)
# - For openai_embedding: model name (e.g., text-embedding-3-small)
EMBEDDED_LLM_MODEL=
# EMBEDDED_LLM_BASE_URL: Base URL for embedding service
# - For ollama: Ollama server URL (e.g., http://localhost:11434)
# - For openai_embedding: OpenAI API base URL (e.g., https://api.openai.com/v1)
EMBEDDED_LLM_BASE_URL=
#######################################################################
################### #####################
################### Database Setting #####################
################### #####################
#######################################################################
# Whether to reuse the current database
# When set to true, will reuse existing database connection
# When set to false, will start download dockers, at this time:
# if DB_STORE is seekdb, DB_USER must be root
# if DB_STORE is oceanbase, DB_USER must be root@test
REUSE_CURRENT_DB=false
# Use what kind of docker, seekdb's docker or oceanbase-ce's docker
# Options: seekdb, oceanbase
# seekdb: If REUSE_CURRENT_DB is false, download seekdb docker
# oceanbase: If REUSE_CURRENT_DB is false, download oceanbase-ce docker
DB_STORE=seekdb
# Database Setting, please change as your environment.
DB_HOST="127.0.0.1"
DB_PORT="2881"
#if REUSE_CURRENT_DB=false and DB_STORE=seekdb, DB_USER must be root
#if REUSE_CURRENT_DB=false and DB_STORE=oceanbase, DB_USER must be root@test
DB_USER="root"
# If database use OceanBase, the DB_USER will contain tenant's name
# DB_USER="root@test"
DB_PASSWORD="oblab"
DB_NAME="test"
#######################################################################
################### #####################
################### RAG Parser Setting #####################
################### #####################
#######################################################################
# Maximum chunk size for text splitting (in characters)
MAX_CHUNK_SIZE=4096
# Limit the number of documents to process (0 means no limit)
LIMIT=0
# Patterns to skip when processing documents (comma-separated, e.g., "*.log,*.tmp")
SKIP_PATTERNS=""