You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
Copy file name to clipboardExpand all lines: 6-bonus-superlinked-rag/README.md
+6-1Lines changed: 6 additions & 1 deletion
Display the source diff
Display the rich diff
Original file line number
Diff line number
Diff line change
@@ -46,7 +46,9 @@ make local-start-superlinked
46
46
```
47
47
48
48
> [!IMPORTANT]
49
-
> Before starting, make sure that you have your `.env` file filled with everything required to run the system.
49
+
> Before starting, ensure you have your `.env` file filled with everything required to run the system.
50
+
>
51
+
> For more details on setting up the local infrastructure, you can check out the course's main [INSTALL_AND_USAGE](https://github.com/decodingml/llm-twin-course/blob/main/INSTALL_AND_USAGE.md) document.
50
52
51
53
To stop the local infrastructure, run:
52
54
```shell
@@ -71,3 +73,6 @@ To query the vector DB, run:
71
73
```shell
72
74
make ... # TO BE ADDED
73
75
```
76
+
77
+
> [!IMPORTANT]
78
+
> You can check out the main [INSTALL_AND_USAGE](https://github.com/decodingml/llm-twin-course/blob/main/INSTALL_AND_USAGE.md) document of the course for more details on an end-to-end flow.
Copy file name to clipboardExpand all lines: INSTALL_AND_USAGE.md
+70-7Lines changed: 70 additions & 7 deletions
Display the source diff
Display the rich diff
Original file line number
Diff line number
Diff line change
@@ -18,13 +18,17 @@ Go to the root of the repository, copy our `.env.example` file and fill it with
18
18
cp .env.example .env
19
19
```
20
20
21
-
## Run
21
+
## Supported commands
22
22
23
-
### Check all possible `Make` commands
23
+
We will use `GNU Make` to install and run our application.
24
+
25
+
To see all our supported commands, run the following:
24
26
```shell
25
27
make help
26
28
```
27
29
30
+
## Set up the infrastructure
31
+
28
32
### Spin up the infrastructure
29
33
30
34
Now, the whole infrastructure can be spun up using a simple Make command:
@@ -33,18 +37,77 @@ Now, the whole infrastructure can be spun up using a simple Make command:
33
37
make local-start
34
38
```
35
39
36
-
Behind the scenes it will build and run all the Docker images defined in the [docker-compose.yml[(https://github.com/decodingml/llm-twin-course/blob/main/docker-compose.yml)] file.
40
+
Behind the scenes it will build and run all the Docker images defined in the [docker-compose.yml](https://github.com/decodingml/llm-twin-course/blob/main/docker-compose.yml) file.
37
41
42
+
## Read this before starting 🚨
38
43
39
44
> [!CAUTION]
40
-
> For `Mongo`` to work with multiple replicas on MacOs or linux systems you have to add the following lines of code to `/etc/hosts`:
45
+
> For `Mongo` to work with multiple replicas (as we use it in our Docker setup) on `macOS` or `Linux` systems, you have to add the following lines of code to `/etc/hosts`:
41
46
>
42
47
> ```
43
48
> 127.0.0.1 mongo1
44
49
> 127.0.0.1 mongo2
45
50
> 127.0.0.1 mongo3
46
51
> ```
47
52
>
48
-
> From what we know on Windows, it works out-of-the-box.
49
-
>
50
-
>
53
+
> From what we know, on `Windows`, it `works out-of-the-box`.
54
+
55
+
> [!WARNING]
56
+
> For `arm` users (e.g., `M1/M2/M3 macOS devices`), go to your Docker desktop application and enable `Use Rosetta for x86_64/amd64 emulation on Apple Silicon` from the Settings. There is a checkbox you have to check.
57
+
> Otherwise, your Docker containers will crash.
58
+
59
+
### Tear down the infrastructure
60
+
61
+
Run the following `Make` command to tear down all your docker containers:
62
+
63
+
```shell
64
+
make local-stop
65
+
```
66
+
67
+
## Run an end-to-end flow
68
+
69
+
Now that we have configured our credentials and started our infrastructure let's look at how to run an end-to-end flow of the LLM Twin application.
70
+
71
+
> [!IMPORTANT]
72
+
> Note that we won't go into the details of the system here. To fully understand it, check out our free article series, which explains everything step-by-step: [LLM Twin articles series](https://medium.com/decodingml/llm-twin-course/home).
73
+
74
+
### Step 1: Crawlers
75
+
76
+
Trigger the crawler to collect data and add it to the MongoDB:
77
+
78
+
```shell
79
+
make local-test-github
80
+
# or make local-test-medium
81
+
```
82
+
83
+
After the data is added to Mongo, the CDC component will be triggered, which will populate the RabbitMQ with the event.
84
+
85
+
### Step 2: Feature engineering & Vector DB
86
+
87
+
Check that the feature pipeline works and the vector DB is successfully populated.
88
+
89
+
To check the `feature pipeline`, check the logs of the `llm-twin-bytewax` Docker container by running:
90
+
```shell
91
+
docker logs llm-twin-bytewax
92
+
```
93
+
You should see logs reflecting the cleaning, chunking, and embedding operations (without any errors, of course).
94
+
95
+
To check that the Qdrant `vector DB` is populated successfully, go to its dashboard at localhost:6333/dashboard. There, you should see the repositories or article collections created and populated.
96
+
97
+
> [!NOTE]
98
+
> If using the cloud version of Qdrant, go to your Qdrant account and cluster to see the same thing as in the local dashboard.
99
+
100
+
### Step 3: RAG retrieval step
101
+
102
+
Now that we have some data in our vector DB, let's test out the RAG retriever:
@@ -142,7 +142,21 @@ If you plan to run the code while reading it, you have to know that we use sever
142
142
11.[Build a scalable RAG ingestion pipeline using 74.3% less code](https://medium.com/decodingml/build-a-scalable-rag-ingestion-pipeline-using-74-3-less-code-ac50095100d6)
143
143
12. Refactoring the 4 Advanced RAG Algorithms using Superlinked
144
144
145
+
## Install & Usage
146
+
147
+
To understand how to install and run the LLM Twin code, go to the [INSTALL_AND_USAGE](https://github.com/decodingml/llm-twin-course/blob/main/INSTALL_AND_USAGE.md) dedicated document.
148
+
149
+
> [!NOTE]
150
+
> Even though you can run everything solely using the [INSTALL_AND_USAGE](https://github.com/decodingml/llm-twin-course/blob/main/INSTALL_AND_USAGE.md) dedicated document, we recommend that you read the articles to understand the LLM Twin system and design choices fully.
151
+
152
+
### Bonus Superlinked series
153
+
154
+
The bonus Superlinked series has an extra dedicated [README](https://github.com/decodingml/llm-twin-course/blob/main/6-bonus-superlinked-rag/README.md) that you can access under the [6-bonus-superlinked-rag](https://github.com/decodingml/llm-twin-course/tree/main/6-bonus-superlinked-rag) directory.
155
+
156
+
Here we explain all the changes made to the code to run it with the improved RAG layer powered by [Superlinked](https://rebrand.ly/superlinked-github).
157
+
145
158
## Meet your teachers!
159
+
146
160
The course is created under the [Decoding ML](https://decodingml.substack.com/) umbrella by:
0 commit comments