Skip to content

Commit 8634727

Browse files
committed
docs: Fill the INSTALL_AND_USAGE docs
1 parent 3a7a61c commit 8634727

File tree

3 files changed

+91
-9
lines changed

3 files changed

+91
-9
lines changed

6-bonus-superlinked-rag/README.md

Lines changed: 6 additions & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -46,7 +46,9 @@ make local-start-superlinked
4646
```
4747

4848
> [!IMPORTANT]
49-
> Before starting, make sure that you have your `.env` file filled with everything required to run the system.
49+
> Before starting, ensure you have your `.env` file filled with everything required to run the system.
50+
>
51+
> For more details on setting up the local infrastructure, you can check out the course's main [INSTALL_AND_USAGE](https://github.com/decodingml/llm-twin-course/blob/main/INSTALL_AND_USAGE.md) document.
5052
5153
To stop the local infrastructure, run:
5254
```shell
@@ -71,3 +73,6 @@ To query the vector DB, run:
7173
```shell
7274
make ... # TO BE ADDED
7375
```
76+
77+
> [!IMPORTANT]
78+
> You can check out the main [INSTALL_AND_USAGE](https://github.com/decodingml/llm-twin-course/blob/main/INSTALL_AND_USAGE.md) document of the course for more details on an end-to-end flow.

INSTALL_AND_USAGE.md

Lines changed: 70 additions & 7 deletions
Original file line numberDiff line numberDiff line change
@@ -18,13 +18,17 @@ Go to the root of the repository, copy our `.env.example` file and fill it with
1818
cp .env.example .env
1919
```
2020

21-
## Run
21+
## Supported commands
2222

23-
### Check all possible `Make` commands
23+
We will use `GNU Make` to install and run our application.
24+
25+
To see all our supported commands, run the following:
2426
```shell
2527
make help
2628
```
2729

30+
## Set up the infrastructure
31+
2832
### Spin up the infrastructure
2933

3034
Now, the whole infrastructure can be spun up using a simple Make command:
@@ -33,18 +37,77 @@ Now, the whole infrastructure can be spun up using a simple Make command:
3337
make local-start
3438
```
3539

36-
Behind the scenes it will build and run all the Docker images defined in the [docker-compose.yml[(https://github.com/decodingml/llm-twin-course/blob/main/docker-compose.yml)] file.
40+
Behind the scenes it will build and run all the Docker images defined in the [docker-compose.yml](https://github.com/decodingml/llm-twin-course/blob/main/docker-compose.yml) file.
3741

42+
## Read this before starting 🚨
3843

3944
> [!CAUTION]
40-
> For `Mongo`` to work with multiple replicas on MacOs or linux systems you have to add the following lines of code to `/etc/hosts`:
45+
> For `Mongo` to work with multiple replicas (as we use it in our Docker setup) on `macOS` or `Linux` systems, you have to add the following lines of code to `/etc/hosts`:
4146
>
4247
> ```
4348
> 127.0.0.1 mongo1
4449
> 127.0.0.1 mongo2
4550
> 127.0.0.1 mongo3
4651
> ```
4752
>
48-
> From what we know on Windows, it works out-of-the-box.
49-
>
50-
>
53+
> From what we know, on `Windows`, it `works out-of-the-box`.
54+
55+
> [!WARNING]
56+
> For `arm` users (e.g., `M1/M2/M3 macOS devices`), go to your Docker desktop application and enable `Use Rosetta for x86_64/amd64 emulation on Apple Silicon` from the Settings. There is a checkbox you have to check.
57+
> Otherwise, your Docker containers will crash.
58+
59+
### Tear down the infrastructure
60+
61+
Run the following `Make` command to tear down all your docker containers:
62+
63+
```shell
64+
make local-stop
65+
```
66+
67+
## Run an end-to-end flow
68+
69+
Now that we have configured our credentials and started our infrastructure let's look at how to run an end-to-end flow of the LLM Twin application.
70+
71+
> [!IMPORTANT]
72+
> Note that we won't go into the details of the system here. To fully understand it, check out our free article series, which explains everything step-by-step: [LLM Twin articles series](https://medium.com/decodingml/llm-twin-course/home).
73+
74+
### Step 1: Crawlers
75+
76+
Trigger the crawler to collect data and add it to the MongoDB:
77+
78+
```shell
79+
make local-test-github
80+
# or make local-test-medium
81+
```
82+
83+
After the data is added to Mongo, the CDC component will be triggered, which will populate the RabbitMQ with the event.
84+
85+
### Step 2: Feature engineering & Vector DB
86+
87+
Check that the feature pipeline works and the vector DB is successfully populated.
88+
89+
To check the `feature pipeline`, check the logs of the `llm-twin-bytewax` Docker container by running:
90+
```shell
91+
docker logs llm-twin-bytewax
92+
```
93+
You should see logs reflecting the cleaning, chunking, and embedding operations (without any errors, of course).
94+
95+
To check that the Qdrant `vector DB` is populated successfully, go to its dashboard at localhost:6333/dashboard. There, you should see the repositories or article collections created and populated.
96+
97+
> [!NOTE]
98+
> If using the cloud version of Qdrant, go to your Qdrant account and cluster to see the same thing as in the local dashboard.
99+
100+
### Step 3: RAG retrieval step
101+
102+
Now that we have some data in our vector DB, let's test out the RAG retriever:
103+
104+
### Step 4: Generate the instruct dataset
105+
106+
107+
### Step 5: Fine-tuning
108+
109+
110+
### Step 6: Inference
111+
112+
113+

README.md

Lines changed: 15 additions & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -1,6 +1,6 @@
11
<div align="center">
22
<h2>LLM Twin Course: Building Your Production-Ready AI Replica</h2>
3-
<h1>An End-to-End Framework for Production-Ready LLM & RAG Systems by Building Your LLM Twin</h1>
3+
<h1>Learn to build an end-to-end system for production-ready LLM & RAG systems by building your LLM Twin</h1>
44
<h3>From data gathering to productionizing LLMs using LLMOps good practices.</h3>
55
<i>by <a href="https://github.com/iusztinpaul">Paul Iusztin</a>, <a href="https://github.com/alexandruvesa">Alexandru Vesa</a> and <a href="https://github.com/Joywalker">Alexandru Razvant</a></i>
66
</div>
@@ -142,7 +142,21 @@ If you plan to run the code while reading it, you have to know that we use sever
142142
11. [Build a scalable RAG ingestion pipeline using 74.3% less code](https://medium.com/decodingml/build-a-scalable-rag-ingestion-pipeline-using-74-3-less-code-ac50095100d6)
143143
12. Refactoring the 4 Advanced RAG Algorithms using Superlinked
144144

145+
## Install & Usage
146+
147+
To understand how to install and run the LLM Twin code, go to the [INSTALL_AND_USAGE](https://github.com/decodingml/llm-twin-course/blob/main/INSTALL_AND_USAGE.md) dedicated document.
148+
149+
> [!NOTE]
150+
> Even though you can run everything solely using the [INSTALL_AND_USAGE](https://github.com/decodingml/llm-twin-course/blob/main/INSTALL_AND_USAGE.md) dedicated document, we recommend that you read the articles to understand the LLM Twin system and design choices fully.
151+
152+
### Bonus Superlinked series
153+
154+
The bonus Superlinked series has an extra dedicated [README](https://github.com/decodingml/llm-twin-course/blob/main/6-bonus-superlinked-rag/README.md) that you can access under the [6-bonus-superlinked-rag](https://github.com/decodingml/llm-twin-course/tree/main/6-bonus-superlinked-rag) directory.
155+
156+
Here we explain all the changes made to the code to run it with the improved RAG layer powered by [Superlinked](https://rebrand.ly/superlinked-github).
157+
145158
## Meet your teachers!
159+
146160
The course is created under the [Decoding ML](https://decodingml.substack.com/) umbrella by:
147161

148162
<table>

0 commit comments

Comments
 (0)