-
Notifications
You must be signed in to change notification settings - Fork 0
/
Copy pathindex_cpp.py
56 lines (44 loc) · 1.75 KB
/
index_cpp.py
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
from llama_cpp import Llama
# conda activate llama
# https://github.com/abetlen/llama-cpp-python
def get_llama_model(model_name = "models\mistral-7b-openorca.Q4_0.gguf"):
# Load the LLaMA model
model = Llama(model_path= model_name,main_gpu=0,
n_gpu_layers=40, n_ctx=512)
return model
def llama_chat(prompt):
llm = get_llama_model()
out = llm.create_completion(f"""<|im_start|>system
You are a helpful chatbot.
<|im_end|>
<|im_start|>user
{prompt}?<|im_end|>
<|im_start|>assistant""", max_tokens=1000, stop=["<|im_end|>"], stream=True)
return out
def send_response(output):
robot_brain=''
for token in output:
robot_brain += token["choices"][0]["text"]
return robot_brain
#create Llama model
'''
prompt = ''
prompt = str(input('Enter the prompt: '))
while True:
if prompt != 'quit':
try:
output = llama_chat(prompt)
print('=========================================================================================================================================================================')
print('User: '+prompt+'\n')
print('System: ')
for token in output:
print(token["choices"][0]["text"], end='', flush=True)
print('=========================================================================================================================================================================')
except Exception as e:
print("Error occurred: ", e)
prompt = str(input('Enter another prompt: '))
else:
print('User: '+prompt+'\n')
print('System: Chatbot is shutdown')
break
'''