diff --git a/docs/2.build/4.web3-apps/ai/ai-assistant.md b/docs/2.build/4.web3-apps/ai/ai-assistant.md
index 4232c6faec..0a51676914 100644
--- a/docs/2.build/4.web3-apps/ai/ai-assistant.md
+++ b/docs/2.build/4.web3-apps/ai/ai-assistant.md
@@ -62,7 +62,7 @@ Before starting, make sure you have the following tools installed:
### AI Model
-In this tutorial we will be using the [NousResearch/Hermes-2-Pro-Mistral-7B-GGUF](https://huggingface.co/squeeze-ai-lab/TinyAgent-1.1B-GGUF) model, which is hosted on [Hugging Face](https://huggingface.co/login).
+In this tutorial we will be using the [NousResearch/Hermes-2-Pro-Llama-3-8B-GGUF](https://huggingface.co/NousResearch/Hermes-2-Pro-Llama-3-8B-GGUF) model, which is hosted on [Hugging Face](https://huggingface.co/login).
```sh
# Install the Hugging Face library
@@ -72,23 +72,23 @@ pip install huggingface_hub
huggingface-cli login
# get the model from Hugging Face
-huggingface-cli download huggingface-cli download NousResearch/Hermes-2-Pro-Mistral-7B-GGUF Hermes-2-Pro-Mistral-7B.Q3_K_M.gguf --local-dir model
+huggingface-cli download NousResearch/Hermes-2-Pro-Llama-3-8B-GGUF Hermes-2-Pro-Llama-3-8B-Q4_K_M.gguf --local-dir model
```
-:::info Q3_K_M Model
+:::info Q4_K_M Model
-The `Q3_K_M` model is one of the smallest models available (`Q2` is smaller), it is used in this tutorial to reduce the time and resources needed to run the AI agent
+We use the small `Q4_K_M` model to reduce the time and resources needed to run the AI agent
:::
### Execute the Model
-You should now have a folder named `./model` with the [GGUF file](https://huggingface.co/docs/hub/en/gguf) `./model/Hermes-2-Pro-Mistral-7B.Q3_K_M.gguf`, lets use `llama.cpp` to run it.
+You should now have a folder named `./model` with the [GGUF file](https://huggingface.co/docs/hub/en/gguf) `./model/Hermes-2-Pro-Llama-3-8B-Q4_K_M.gguf`, lets use `llama.cpp` to run it.
```bash
# run the model with llama.cpp
-llama-server -m ./model/Hermes-2-Pro-Mistral-7B.Q3_K_M.gguf
+llama-server -m ./model/Hermes-2-Pro-Llama-3-8B-Q4_K_M.gguf
```
Open your browser at `http://localhost:8080`, if you see an interface similar to this one you are ready to go 🚀
@@ -100,9 +100,6 @@ You can use a different model with llama.cpp if you wish! Just make sure it supp
:::
-
-
---
### Project Setup