Skip to content

Commit 6cd5f67

Browse files
committed
LangChain MCP template fix
1 parent 90a6130 commit 6cd5f67

File tree

2 files changed

+14
-11
lines changed

2 files changed

+14
-11
lines changed

src/client/content/config/tabs/settings.py

Lines changed: 3 additions & 3 deletions
Original file line numberDiff line numberDiff line change
@@ -160,8 +160,8 @@ def spring_ai_conf_check(ll_model: dict, embed_model: dict) -> str:
160160
ll_provider = ll_model.get("provider", "")
161161
embed_provider = embed_model.get("provider", "")
162162
logger.info(f"llm chat:{ll_provider} - embeddings:{embed_provider}")
163-
if all("openai_compatible" in p for p in (ll_provider, embed_provider)):
164-
return "openai_compatible"
163+
if all("hosted_vllm" in p for p in (ll_provider, embed_provider)):
164+
return "hosted_vllm"
165165
if all("openai" in p for p in (ll_provider, embed_provider)):
166166
return "openai"
167167
if all("ollama" in p for p in (ll_provider, embed_provider)):
@@ -364,7 +364,7 @@ def display_settings():
364364
disabled=spring_ai_conf == "hybrid",
365365
)
366366
with col_centre:
367-
if (spring_ai_conf != "openai_compatible"):
367+
if (spring_ai_conf != "hosted_vllm"):
368368
st.download_button(
369369
label="Download SpringAI",
370370
data=spring_ai_zip(spring_ai_conf, ll_config, embed_config), # Generate zip on the fly

src/client/mcp/rag/optimizer_utils/config.py

Lines changed: 11 additions & 8 deletions
Original file line numberDiff line numberDiff line change
@@ -25,14 +25,16 @@
2525
def get_llm(data):
2626
logger.info("llm data:")
2727
logger.info(data["client_settings"]["ll_model"]["model"])
28+
model_full = data["client_settings"]["ll_model"]["model"]
29+
_, prefix, model = model_full.partition('/')
2830
llm = {}
2931
models_by_id = {m["id"]: m for m in data.get("model_configs", [])}
30-
llm_config= models_by_id.get(data["client_settings"]["ll_model"]["model"])
32+
llm_config= models_by_id.get(model)
3133
logger.info(llm_config)
3234
provider = llm_config["provider"]
3335
url = llm_config["api_base"]
34-
api_key = llm_config["api_key"]
35-
model = data["client_settings"]["ll_model"]["model"]
36+
api_key = llm_config["api_key"]
37+
3638
logger.info(f"CHAT_MODEL: {model} {provider} {url} {api_key}")
3739
if provider == "ollama":
3840
# Initialize the LLM
@@ -41,21 +43,22 @@ def get_llm(data):
4143
elif provider == "openai":
4244
llm = ChatOpenAI(model=model, api_key=api_key)
4345
logger.info("OpenAI LLM created")
44-
elif provider =="openai_compatible":
46+
elif provider =="hosted_vllm":
4547
llm = ChatOpenAI(model=model, api_key=api_key,base_url=url)
46-
logger.info("OpenAI compatible LLM created")
48+
logger.info("hosted_vllm compatible LLM created")
4749
return llm
4850

4951

5052
def get_embeddings(data):
5153
embeddings = {}
5254
logger.info("getting embeddings..")
53-
model = data["client_settings"]["vector_search"]["model"]
55+
model_full = data["client_settings"]["vector_search"]["model"]
56+
_, prefix, model = model_full.partition('/')
5457
logger.info(f"embedding model: {model}")
5558
models_by_id = {m["id"]: m for m in data.get("model_configs", [])}
5659
model_params= models_by_id.get(model)
5760
provider = model_params["provider"]
58-
url = model_params["url"]
61+
url = model_params["api_base"]
5962
api_key = model_params["api_key"]
6063

6164
logger.info(f"Embeddings Model: {model} {provider} {url} {api_key}")
@@ -68,7 +71,7 @@ def get_embeddings(data):
6871
logger.info("OpenAI embeddings connection successful")
6972
elif (provider == "hosted_vllm"):
7073
embeddings = OpenAIEmbeddings(model=model, api_key=api_key,base_url=url,check_embedding_ctx_length=False)
71-
logger.info("OpenAI compatible embeddings connection successful")
74+
logger.info("hosted_vllm compatible embeddings connection successful")
7275

7376
return embeddings
7477

0 commit comments

Comments
 (0)