You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
Copy file name to clipboardexpand all lines: README.md
+51-15
Original file line number
Diff line number
Diff line change
@@ -1,36 +1,43 @@
1
1
# LevelRAG: Enhancing Retrieval-Augmented Generation with Multi-hop Logic Planning over Rewriting Augmented Searchers
2
2
3
-
Source code for paper "LevelRAG: Enhancing Retrieval-Augmented Generation with Multi-hop Logic Planning over Rewriting Augmented Searchers".
3
+
Source code for paper **"LevelRAG: Enhancing Retrieval-Augmented Generation with Multi-hop Logic Planning over Rewriting Augmented Searchers"**.
4
4
5
5
## Overview
6
6
**LevelRAG** is a two-stage retrieval-augmented generation (RAG) framework that incorporates multi-hop logic planning and hybrid retrieval to enhance both completeness and accuracy of the retrieval process. The first stage involves a high-level searcher that decomposing the user query into atomic sub-queries. The second stage utilizes multiple low-level searchers to retrieve the most relevant documents for each sub-query, which are then used to generate the final answer. In each low-level searcher, large language models (LLMs) are employed to refine the atomic queries to better fit the corresponding retriever.
7
7
8
8
<center>
9
-
<imgsrc="./assets/levelrag.png"width="80%">
9
+
<imgsrc="./assets/LevelRAG.png"width="80%">
10
10
</center>
11
11
12
12
13
13
## Running LevelRAG
14
14
15
15
### Prepare the Environment
16
-
Our code is based on [FlexRAG](https://github.com/ictnlp/flexrag). Please follow the instruction to install FlexRAG:
16
+
Our code is based on [FlexRAG](https://github.com/ictnlp/flexrag) project. Please follow the instruction to install FlexRAG:
17
17
```bash
18
-
pip install flexrag
18
+
pip install flexrag==0.1.11
19
+
```
20
+
21
+
Download the source code of this project:
22
+
```bash
23
+
git clone https://github.com/ictnlp/LevelRAG
19
24
```
20
25
21
26
### Prepare the Retriever
22
-
Before running the LevelRAG, preparing the retriever is necessary. LevelRAG employs three kind of retrievers in total, naming `dense retriever`, `sparse retriever`, and `web retriever`, respectively. Except for the `web retriever`, which does not require index construction, both the dense retriever and the sparse retriever need to prepare the index first. In our experiments, we use the wikipedia corpus provided by [Atlas](https://github.com/facebookresearch/atlas). You can download the corpus by running the following command:
27
+
Before running the LevelRAG, preparing the retriever is necessary. LevelRAG employs three kind of retrievers in total, naming `DenseRetriever`, `ElasticRetriever`, and `WebRetriever`, respectively. Except for the `WebRetriever`, which does not require index construction, both the `DenseRetriever` and the `ElasticRetriever` need to prepare the index first. In our experiments, we use the wikipedia corpus provided by [Atlas](https://github.com/facebookresearch/atlas). You can download the corpus by running the following command:
> You need to setup the elasticsearch server first. You can follow the instruction [here](https://www.elastic.co/guide/en/elasticsearch/reference/current/install-elasticsearch.html) to install the elasticsearch server.
67
+
> Before building the `ElasticRetriever`, you need to setup the elasticsearch server. You can follow the instruction [here](https://www.elastic.co/guide/en/elasticsearch/reference/current/install-elasticsearch.html) to install the elasticsearch server.
68
+
69
+
Using the `WebRetriever` does not require index construction. However, you need to prepare the Bing Search API_KEY in advance. You can visit [Bing Search API](https://www.microsoft.com/en-us/bing/apis) to get the API_KEY.
70
+
71
+
### Prepare the Genereator
72
+
LevelRAG uses the `Qwen2-7B-Instruct` as the generator. You can deploy the generator using `vllm` by running the following command:
73
+
74
+
```bash
75
+
python -m vllm.entrypoints.openai.api_server \
76
+
--model Qwen2-7B-Instruct \
77
+
--gpu-memory-utilization 0.95 \
78
+
--tensor-parallel-size 4 \
79
+
--port 8000 \
80
+
--host 0.0.0.0 \
81
+
--trust-remote-code
82
+
```
83
+
84
+
This command will deploy the `Qwen2-7B-Instruct` using 4 GPUs. You can adjust the `--tensor-parallel-size` and the `--gpu-memory-utilization` according to your own GPU configuration.
85
+
86
+
87
+
### Run the LevelRAG Evaluation Script
88
+
After preparing the retriever, you can run the LevelRAG by running the scripts in the `scripts` folder. Before running the scripts, make sure you have substituted the placeholder variables in the scripts with the correct values.
59
89
90
+
- LEVELRAG_PATH: The path to the LevelRAG repository.
91
+
- DENSE_PATH: The path to the `DenseRetriever`.
92
+
- BING_KEY: The Bing Search API_KEY.
60
93
61
-
### Run the LevelRAG
62
-
After preparing the retriever, you can run the LevelRAG by running the scripts in the `scripts` folder.
94
+
### Run the LevelRAG GUI Demo
95
+
We also provide a GUI demo for LevelRAG. You can run the GUI demo by running the `run_highlevel_gui.sh` script in the `scripts` folder. In the GUI, you can input the query and view the output of LevelRAG.
63
96
97
+
## Running the Simple LevelRAG
98
+
If you think building the retriever is too complicated, you can run the simple version of LevelRAG by running the
99
+
`run_simple.sh` script in the `scripts` folder. The simple version of LevelRAG only uses the `DenseRetriever` and does not require the `WebRetriever` and the `ElasticRetriever`. Thanks to the good multi-hop problem decomposition and sub-query adaptivity optimization, LevelRAG can achieve good performance even with a single retriever, and the running speed is faster. You can also run the `run_simple_gui.sh` script to start the GUI application of the simple version of LevelRAG.
64
100
65
101
## Citation
66
102
If you find our work useful, please consider citing our paper:
0 commit comments