25
25
def get_llm (data ):
26
26
logger .info ("llm data:" )
27
27
logger .info (data ["client_settings" ]["ll_model" ]["model" ])
28
+ model_full = data ["client_settings" ]["ll_model" ]["model" ]
29
+ _ , prefix , model = model_full .partition ('/' )
28
30
llm = {}
29
31
models_by_id = {m ["id" ]: m for m in data .get ("model_configs" , [])}
30
- llm_config = models_by_id .get (data [ "client_settings" ][ "ll_model" ][ " model" ] )
32
+ llm_config = models_by_id .get (model )
31
33
logger .info (llm_config )
32
34
provider = llm_config ["provider" ]
33
35
url = llm_config ["api_base" ]
34
- api_key = llm_config ["api_key" ]
35
- model = data [ "client_settings" ][ "ll_model" ][ "model" ]
36
+ api_key = llm_config ["api_key" ]
37
+
36
38
logger .info (f"CHAT_MODEL: { model } { provider } { url } { api_key } " )
37
39
if provider == "ollama" :
38
40
# Initialize the LLM
@@ -41,21 +43,22 @@ def get_llm(data):
41
43
elif provider == "openai" :
42
44
llm = ChatOpenAI (model = model , api_key = api_key )
43
45
logger .info ("OpenAI LLM created" )
44
- elif provider == "openai_compatible " :
46
+ elif provider == "hosted_vllm " :
45
47
llm = ChatOpenAI (model = model , api_key = api_key ,base_url = url )
46
- logger .info ("OpenAI compatible LLM created" )
48
+ logger .info ("hosted_vllm compatible LLM created" )
47
49
return llm
48
50
49
51
50
52
def get_embeddings (data ):
51
53
embeddings = {}
52
54
logger .info ("getting embeddings.." )
53
- model = data ["client_settings" ]["vector_search" ]["model" ]
55
+ model_full = data ["client_settings" ]["vector_search" ]["model" ]
56
+ _ , prefix , model = model_full .partition ('/' )
54
57
logger .info (f"embedding model: { model } " )
55
58
models_by_id = {m ["id" ]: m for m in data .get ("model_configs" , [])}
56
59
model_params = models_by_id .get (model )
57
60
provider = model_params ["provider" ]
58
- url = model_params ["url " ]
61
+ url = model_params ["api_base " ]
59
62
api_key = model_params ["api_key" ]
60
63
61
64
logger .info (f"Embeddings Model: { model } { provider } { url } { api_key } " )
@@ -68,7 +71,7 @@ def get_embeddings(data):
68
71
logger .info ("OpenAI embeddings connection successful" )
69
72
elif (provider == "hosted_vllm" ):
70
73
embeddings = OpenAIEmbeddings (model = model , api_key = api_key ,base_url = url ,check_embedding_ctx_length = False )
71
- logger .info ("OpenAI compatible embeddings connection successful" )
74
+ logger .info ("hosted_vllm compatible embeddings connection successful" )
72
75
73
76
return embeddings
74
77
0 commit comments