You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
# export no_proxy="localhost,127.0.0.1,${HOST_IP}" # Add other hosts if necessary
62
+
# export no_proxy="localhost,127.0.0.1,${host_ip}" # Add other hosts if necessary
63
63
source ../../../set_env.sh
64
64
```
65
65
66
-
_Note: The compose file might read additional variables from a `.env` file or expect them defined elsewhere. Ensure all required variables like ports (`LLM_SERVICE_PORT`, `MEGA_SERVICE_PORT`, etc.) are setif not using defaults from the compose file._
66
+
_Note: The compose file might read additional variables from set_env.sh. Ensure all required variables like ports (`LLM_SERVICE_PORT`, `MEGA_SERVICE_PORT`, etc.) are setif not using defaults from the compose file._
|`HOST_IP`| External IP address of the host machine. **Required.**|`your_external_ip_address`|
117
+
|`host_ip`| External IP address of the host machine. **Required.**|`your_external_ip_address`|
107
118
|`HUGGINGFACEHUB_API_TOKEN`| Your Hugging Face Hub token for model access. **Required.**|`your_huggingface_token`|
108
119
|`LLM_MODEL_ID`| Hugging Face model ID for the CodeGen LLM (used by TGI/vLLM service). Configured within `compose.yaml` environment. |`Qwen/Qwen2.5-Coder-7B-Instruct`|
109
120
|`EMBEDDING_MODEL_ID`| Hugging Face model ID for the embedding model (used by TEI service). Configured within `compose.yaml` environment. |`BAAI/bge-base-en-v1.5`|
110
121
|`LLM_ENDPOINT`| Internal URL forthe LLM serving endpoint (used by `codegen-llm-server`). Configuredin`compose.yaml`. |`http://codegen-tgi-server:80/generate` or `http://codegen-vllm-server:8000/v1/chat/completions`|
|`DATAPREP_ENDPOINT`| Internal URL forthe Data Preparation service. Configuredin`compose.yaml`. |`http://codegen-dataprep-server:80/dataprep`|
113
-
|`BACKEND_SERVICE_ENDPOINT`| External URL for the CodeGen Gateway (MegaService). Derived from `HOST_IP` and port `7778`. |`http://${HOST_IP}:7778/v1/codegen`|
124
+
|`BACKEND_SERVICE_ENDPOINT`| External URL for the CodeGen Gateway (MegaService). Derived from `host_ip` and port `7778`. |`http://${host_ip}:7778/v1/codegen`|
@@ -189,15 +200,15 @@ _(Port `8080` is the default host mapping for `codegen-gradio-ui-server`)_
189
200
190
201
1. Modify `compose.yaml`: Comment out the `codegen-gradio-ui-server` service and uncomment/add the `codegen-xeon-ui-server` (Svelte) service definition, ensuring the port mapping is correct (e.g., `"- 5173:5173"`).
191
202
2. Restart Docker Compose: `docker compose --profile <profile_name> up -d`
192
-
3. Access: `http://{HOST_IP}:5173` (or the host port you mapped).
203
+
3. Access: `http://{host_ip}:5173` (or the host port you mapped).
1. Modify `compose.yaml`: Comment out the default UI service and uncomment/add the `codegen-xeon-react-ui-server` definition, ensuring correct port mapping (e.g., `"- 5174:80"`).
199
210
2. Restart Docker Compose: `docker compose --profile <profile_name> up -d`
200
-
3. Access: `http://{HOST_IP}:5174` (or the host port you mapped).
211
+
3. Access: `http://{host_ip}:5174` (or the host port you mapped).
2. **Configure:** Set the "Service URL"in the extension settings to your CodeGen backend endpoint: `http://${HOST_IP}:7778/v1/codegen` (use the correct port if changed).
221
+
2. **Configure:** Set the "Service URL"in the extension settings to your CodeGen backend endpoint: `http://${host_ip}:7778/v1/codegen` (use the correct port if changed).
- **Inline Suggestion:** Type a comment describing the code you want (e.g., `# Python function to read a file`) and waitfor suggestions.
@@ -218,7 +229,7 @@ Users can interact with the backend service using the `Neural Copilot` VS Code e
218
229
## Troubleshooting
219
230
220
231
- **Model Download Issues:** Check `HUGGINGFACEHUB_API_TOKEN`. Ensure internet connectivity or correct proxy settings. Check logs of `tgi-service`/`vllm-service` and `tei-embedding-server`. Gated models need prior Hugging Face access.
221
-
- **Connection Errors:** Verify `HOST_IP` is correct and accessible. Check `docker ps`for port mappings. Ensure `no_proxy` includes `HOST_IP`if using a proxy. Check logs of the service failing to connect (e.g., `codegen-backend-server` logs if it can't reach `codegen-llm-server`).
232
+
- **Connection Errors:** Verify `host_ip` is correct and accessible. Check `docker ps`for port mappings. Ensure `no_proxy` includes `host_ip`if using a proxy. Check logs of the service failing to connect (e.g., `codegen-backend-server` logs if it can't reach `codegen-llm-server`).
222
233
- **"Container name is in use"**: Stop existing containers (`docker compose down`) or change `container_name` in `compose.yaml`.
223
234
- **Resource Issues:** CodeGen models can be memory-intensive. Monitor host RAM usage. Increase Docker resources if needed.
# export no_proxy="localhost,127.0.0.1,${HOST_IP}" # Add other hosts if necessary
63
+
# export no_proxy="localhost,127.0.0.1,${host_ip}" # Add other hosts if necessary
64
64
source ../../../set_env.sh
65
65
```
66
66
67
-
_Note: Ensure all required variables like ports (`LLM_SERVICE_PORT`, `MEGA_SERVICE_PORT`, etc.) are setif not using defaults from the compose file._
67
+
_Note: The compose file might read additional variables from set_env.sh. Ensure all required variables like ports (`LLM_SERVICE_PORT`, `MEGA_SERVICE_PORT`, etc.) are setif not using defaults from the compose file._
|`HOST_IP`| External IP address of the host machine. **Required.**|`your_external_ip_address`|
120
+
|`host_ip`| External IP address of the host machine. **Required.**|`your_external_ip_address`|
110
121
|`HUGGINGFACEHUB_API_TOKEN`| Your Hugging Face Hub token for model access. **Required.**|`your_huggingface_token`|
111
-
|`LLM_MODEL_ID`| Hugging Face model ID for the CodeGen LLM (used by TGI/vLLM service). Configured within `compose.yaml` environment. |`Qwen/Qwen2.5-Coder-7B-Instruct`|
122
+
|`LLM_MODEL_ID`| Hugging Face model ID for the CodeGen LLM (used by TGI/vLLM service). Configured within `compose.yaml` environment. |`Qwen/Qwen2.5-Coder-32B-Instruct`|
112
123
|`EMBEDDING_MODEL_ID`| Hugging Face model ID for the embedding model (used by TEI service). Configured within `compose.yaml` environment. |`BAAI/bge-base-en-v1.5`|
113
124
|`LLM_ENDPOINT`| Internal URL forthe LLM serving endpoint (used by `codegen-llm-server`). Configuredin`compose.yaml`. |`http://codegen-tgi-server:80/generate` or `http://codegen-vllm-server:8000/v1/chat/completions`|
|`DATAPREP_ENDPOINT`| Internal URL forthe Data Preparation service. Configuredin`compose.yaml`. |`http://codegen-dataprep-server:80/dataprep`|
116
-
|`BACKEND_SERVICE_ENDPOINT`| External URL for the CodeGen Gateway (MegaService). Derived from `HOST_IP` and port `7778`. |`http://${HOST_IP}:7778/v1/codegen`|
127
+
|`BACKEND_SERVICE_ENDPOINT`| External URL for the CodeGen Gateway (MegaService). Derived from `host_ip` and port `7778`. |`http://${host_ip}:7778/v1/codegen`|
1. Modify `compose.yaml`: Swap Gradio service for Svelte (`codegen-gaudi-ui-server`), check port map (e.g., `5173:5173`).
208
219
2. Restart: `docker compose --profile <profile_name> up -d`
209
-
3. Access: `http://{HOST_IP}:5173`
220
+
3. Access: `http://{host_ip}:5173`
210
221
211
222
### React UI (Optional)
212
223
213
224
1. Modify `compose.yaml`: Swap Gradio service for React (`codegen-gaudi-react-ui-server`), check port map (e.g., `5174:80`).
214
225
2. Restart: `docker compose --profile <profile_name> up -d`
215
-
3. Access: `http://{HOST_IP}:5174`
226
+
3. Access: `http://{host_ip}:5174`
216
227
217
228
### VS Code Extension (Optional)
218
229
219
-
Use the `Neural Copilot` extension configured with the CodeGen backend URL: `http://${HOST_IP}:7778/v1/codegen`. (See Xeon README for detailed setup screenshots).
230
+
Use the `Neural Copilot` extension configured with the CodeGen backend URL: `http://${host_ip}:7778/v1/codegen`. (See Xeon README for detailed setup screenshots).
220
231
221
232
## Troubleshooting
222
233
@@ -226,7 +237,7 @@ Use the `Neural Copilot` extension configured with the CodeGen backend URL: `htt
226
237
- Verify `runtime: habana` and volume mounts in`compose.yaml`.
227
238
- Gaudi initialization can take significant time and memory. Monitor resource usage.
228
239
- **Model Download Issues:** Check `HUGGINGFACEHUB_API_TOKEN`, internet access, proxy settings. Check LLM service logs.
229
-
- **Connection Errors:** Verify `HOST_IP`, ports, and proxy settings. Use `docker ps` and check service logs.
240
+
- **Connection Errors:** Verify `host_ip`, ports, and proxy settings. Use `docker ps` and check service logs.
0 commit comments