feat: minimize token usage when testing model (#5905)

This commit is contained in:
Wang Jiyuan 2025-05-12 18:19:03 +08:00 committed by GitHub
parent 84b205cac2
commit 55c78382e4
No known key found for this signature in database
GPG Key ID: B5690EEEBB952194
4 changed files with 9 additions and 4 deletions

View File

@ -678,7 +678,7 @@ export default class AnthropicProvider extends BaseProvider {
const body = {
model: model.id,
messages: [{ role: 'user' as const, content: 'hi' }],
max_tokens: 100,
max_tokens: 2, // api文档写的 x>1
stream
}

View File

@ -916,7 +916,7 @@ export default class GeminiProvider extends BaseProvider {
model: model.id,
contents: [{ role: 'user', parts: [{ text: 'hi' }] }],
config: {
maxOutputTokens: 100
maxOutputTokens: 1
}
})
if (isEmpty(result.text)) {
@ -927,7 +927,7 @@ export default class GeminiProvider extends BaseProvider {
model: model.id,
contents: [{ role: 'user', parts: [{ text: 'hi' }] }],
config: {
maxOutputTokens: 100
maxOutputTokens: 1
}
})
// 等待整个流式响应结束

View File

@ -1112,6 +1112,9 @@ export default class OpenAICompatibleProvider extends BaseOpenAiProvider {
const body = {
model: model.id,
messages: [{ role: 'user', content: 'hi' }],
max_completion_tokens: 1, // openAI
max_tokens: 1, // openAI deprecated 但大部分OpenAI兼容的提供商继续用这个头
enable_thinking: false, // qwen3
stream
}

View File

@ -1026,6 +1026,7 @@ export abstract class BaseOpenAiProvider extends BaseProvider {
const response = await this.sdk.responses.create({
model: model.id,
input: [{ role: 'user', content: 'hi' }],
max_output_tokens: 1,
stream: true
})
let hasContent = false
@ -1042,7 +1043,8 @@ export abstract class BaseOpenAiProvider extends BaseProvider {
const response = await this.sdk.responses.create({
model: model.id,
input: [{ role: 'user', content: 'hi' }],
stream: false
stream: false,
max_output_tokens: 1
})
if (!response.output_text) {
throw new Error('Empty response')