diff --git a/build/build_number.txt b/build/build_number.txt index cdffbbc..f2c1eee 100644 --- a/build/build_number.txt +++ b/build/build_number.txt @@ -1 +1 @@ -166 +167 diff --git a/src/services/assistant.ts b/src/services/assistant.ts index e3a7999..fd5e0b0 100644 --- a/src/services/assistant.ts +++ b/src/services/assistant.ts @@ -180,13 +180,10 @@ export default class { if (error.name !== 'AbortError') { if (error.status === 401 || error.message.includes('401') || error.message.toLowerCase().includes('apikey')) { message.setText('You need to enter your API key in the Models tab of Settings in order to chat.') - callback?.call(null, { done: true }) } else if (error.status === 400 && (error.message.includes('credit') || error.message.includes('balance'))) { message.setText('Sorry, it seems you have run out of credits. Check the balance of your LLM provider account.') - callback?.call(null, { done: true }) } else if (message.content === '') { message.setText('Sorry, I could not generate text for that prompt.') - callback?.call(null, { done: true }) } else { message.appendText({ text: '\n\nSorry, I am not able to continue here.', done: true }) } diff --git a/tests/unit/assistant.test.ts b/tests/unit/assistant.test.ts index 4f8ffb6..6d9cf46 100644 --- a/tests/unit/assistant.test.ts +++ b/tests/unit/assistant.test.ts @@ -27,16 +27,14 @@ let assistant: Assistant = null const prompt = async (prompt: string, opts: LlmCompletionOpts = {}): Promise => { // callback - let done = false let content = '' const callback = (chunk: LlmStream) => { - done = chunk?.done content += chunk?.text || '' } // call and wait await assistant.prompt(prompt, { ...opts, save: false }, callback) - await vi.waitUntil(async () => done) + await vi.waitUntil(async () => !assistant.chat.lastMessage().transient) // return return content