You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
Copy file name to clipboardExpand all lines: bin/lib/local-inference.js
+12-4Lines changed: 12 additions & 4 deletions
Original file line number
Diff line number
Diff line change
@@ -70,7 +70,8 @@ function validateLocalProvider(provider, runCapture) {
70
70
case"ollama-local":
71
71
return{
72
72
ok: false,
73
-
message: "Local Ollama was selected, but nothing is responding on http://localhost:11434.",
73
+
message:
74
+
"Local Ollama was selected, but nothing is responding on http://localhost:11434.",
74
75
};
75
76
default:
76
77
return{ok: false,message: "The selected local inference provider is unavailable."};
@@ -101,7 +102,10 @@ function validateLocalProvider(provider, runCapture) {
101
102
"Local Ollama is responding on localhost, but containers cannot reach http://host.openshell.internal:11434. Ensure Ollama listens on 0.0.0.0:11434 instead of 127.0.0.1 so sandboxes can reach it.",
102
103
};
103
104
default:
104
-
return{ok: false,message: "The selected local inference provider is unavailable from containers."};
105
+
return{
106
+
ok: false,
107
+
message: "The selected local inference provider is unavailable from containers.",
108
+
};
105
109
}
106
110
}
107
111
@@ -127,7 +131,9 @@ function parseOllamaTags(output) {
0 commit comments