Skip to content

Commit 60c6995

Browse files
[autofix.ci] apply automated fixes
1 parent 52d587a commit 60c6995

File tree

6 files changed

+86
-19
lines changed

6 files changed

+86
-19
lines changed
Lines changed: 10 additions & 7 deletions
Original file line numberDiff line numberDiff line change
@@ -1,9 +1,12 @@
1-
import { EmbeddingProvider } from "src/common/embedding-provider";
1+
import { EmbeddingProvider } from 'src/common/embedding-provider';
22

33
describe('Model Provider Test', () => {
4-
let embProvider = EmbeddingProvider.getInstance();
5-
it('should generate a response from the model provider', async () => {
6-
let res = await embProvider.generateEmbResponse("Your text string goes here", "text-embedding-3-small");
7-
console.log(res);
8-
});
9-
});
4+
const embProvider = EmbeddingProvider.getInstance();
5+
it('should generate a response from the model provider', async () => {
6+
const res = await embProvider.generateEmbResponse(
7+
'Your text string goes here',
8+
'text-embedding-3-small',
9+
);
10+
console.log(res);
11+
});
12+
});

backend/src/config/config-loader.ts

Lines changed: 0 additions & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -118,7 +118,6 @@ export class ConfigLoader {
118118
}
119119

120120
this.logger.log(ConfigLoader.config);
121-
122121
}
123122

124123
get<T>(path?: string): T {

backend/src/embedding/__tests__/loadAllEmbModels.spec.ts

Lines changed: 0 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -30,8 +30,6 @@ describe('testing embedding provider', () => {
3030
documents,
3131
);
3232
}, 6000000);
33-
34-
3533
});
3634

3735
afterAll(() => {

llm-server/src/emb-provider.ts

Lines changed: 7 additions & 5 deletions
Original file line numberDiff line numberDiff line change
@@ -1,6 +1,6 @@
11
import { Response } from 'express';
22
import { openAIEmbProvider } from './embedding/openai-embedding-provider';
3-
import { LlamaModelProvider } from './model/llama-model-provider'; // 如果支持Llama模型
3+
import { LlamaModelProvider } from './model/llama-model-provider'; // 如果支持Llama模型
44
import { Logger } from '@nestjs/common';
55
import {
66
ModelProviderType,
@@ -43,9 +43,9 @@ export class EmbeddingModelProvider {
4343
private createModelProvider(type: ModelProviderType): EmbeddingProvider {
4444
switch (type) {
4545
case 'openai':
46-
return new openAIEmbProvider({apiKey: process.env.OPEN_API_KEY});
46+
return new openAIEmbProvider({ apiKey: process.env.OPEN_API_KEY });
4747
// case 'llama':
48-
//
48+
//
4949
// // return new LlamaModelProvider();
5050
default:
5151
throw new Error(`Unsupported embedding model provider type: ${type}`);
@@ -66,7 +66,7 @@ export class EmbeddingModelProvider {
6666
}
6767

6868
async generateEmbeddingResponse(
69-
params: GenerateMessageParams,
69+
params: GenerateMessageParams,
7070
res: Response,
7171
): Promise<void> {
7272
this.ensureInitialized();
@@ -100,7 +100,9 @@ export class EmbeddingModelProvider {
100100

101101
private ensureInitialized(): void {
102102
if (!this.initialized) {
103-
throw new Error('Embedding provider not initialized. Call initialize() first.');
103+
throw new Error(
104+
'Embedding provider not initialized. Call initialize() first.',
105+
);
104106
}
105107
}
106108

llm-server/src/embedding/emb-provider.ts

Lines changed: 4 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -3,7 +3,9 @@ import { GenerateMessageParams } from '../types';
33

44
export interface EmbeddingProvider {
55
initialize(): Promise<void>;
6-
generateEmbResponse(params: GenerateMessageParams,
7-
res: Response,): Promise<void>;
6+
generateEmbResponse(
7+
params: GenerateMessageParams,
8+
res: Response,
9+
): Promise<void>;
810
getEmbList(res: Response): Promise<void>;
911
}

pnpm-lock.yaml

Lines changed: 65 additions & 2 deletions
Some generated files are not rendered by default. Learn more about customizing how changed files appear on GitHub.

0 commit comments

Comments
 (0)