From 30ee6fba02c4fb2b789a0473a8cc181492e2c84a Mon Sep 17 00:00:00 2001 From: icarus Date: Wed, 6 Aug 2025 16:56:33 +0800 Subject: [PATCH] =?UTF-8?q?refactor:=20=E9=87=8D=E5=91=BD=E5=90=8D?= =?UTF-8?q?=E5=87=BD=E6=95=B0isSupportedFlexServiceTier=E4=B8=BAisSupportF?= =?UTF-8?q?lexServiceTierModel?= MIME-Version: 1.0 Content-Type: text/plain; charset=UTF-8 Content-Transfer-Encoding: 8bit 统一函数命名风格,提高代码可读性 --- src/renderer/src/aiCore/clients/BaseApiClient.ts | 6 +++--- src/renderer/src/config/models.ts | 2 +- src/renderer/src/pages/home/Tabs/SettingsTab.tsx | 6 +++--- 3 files changed, 7 insertions(+), 7 deletions(-) diff --git a/src/renderer/src/aiCore/clients/BaseApiClient.ts b/src/renderer/src/aiCore/clients/BaseApiClient.ts index 9bb0f92789..a88cb2a628 100644 --- a/src/renderer/src/aiCore/clients/BaseApiClient.ts +++ b/src/renderer/src/aiCore/clients/BaseApiClient.ts @@ -3,7 +3,7 @@ import { isFunctionCallingModel, isNotSupportTemperatureAndTopP, isOpenAIModel, - isSupportedFlexServiceTier + isSupportFlexServiceTierModel } from '@renderer/config/models' import { REFERENCE_PROMPT } from '@renderer/config/prompts' import { getLMStudioKeepAliveTime } from '@renderer/hooks/useLMStudio' @@ -210,7 +210,7 @@ export abstract class BaseApiClient< let serviceTier = 'auto' as OpenAIServiceTier if (openAI && openAI?.serviceTier === 'flex') { - if (isSupportedFlexServiceTier(model)) { + if (isSupportFlexServiceTierModel(model)) { serviceTier = 'flex' } else { serviceTier = 'auto' @@ -223,7 +223,7 @@ export abstract class BaseApiClient< } protected getTimeout(model: Model) { - if (isSupportedFlexServiceTier(model)) { + if (isSupportFlexServiceTierModel(model)) { return 15 * 1000 * 60 } return defaultTimeout diff --git a/src/renderer/src/config/models.ts b/src/renderer/src/config/models.ts index eebcaada43..bba2b8da2e 100644 --- a/src/renderer/src/config/models.ts +++ b/src/renderer/src/config/models.ts @@ -2548,7 +2548,7 @@ export function isOpenAIModel(model: Model): boolean { return model.id.includes('gpt') || isOpenAIReasoningModel(model) } -export function isSupportedFlexServiceTier(model: Model): boolean { +export function isSupportFlexServiceTierModel(model: Model): boolean { if (!model) { return false } diff --git a/src/renderer/src/pages/home/Tabs/SettingsTab.tsx b/src/renderer/src/pages/home/Tabs/SettingsTab.tsx index 7f0d91602c..a01763ede3 100644 --- a/src/renderer/src/pages/home/Tabs/SettingsTab.tsx +++ b/src/renderer/src/pages/home/Tabs/SettingsTab.tsx @@ -5,8 +5,8 @@ import Selector from '@renderer/components/Selector' import { DEFAULT_CONTEXTCOUNT, DEFAULT_MAX_TOKENS, DEFAULT_TEMPERATURE } from '@renderer/config/constant' import { isOpenAIModel, - isSupportedFlexServiceTier, - isSupportedReasoningEffortOpenAIModel + isSupportedReasoningEffortOpenAIModel, + isSupportFlexServiceTierModel } from '@renderer/config/models' import { translateLanguageOptions } from '@renderer/config/translate' import { useCodeStyle } from '@renderer/context/CodeStyleProvider' @@ -172,7 +172,7 @@ const SettingsTab: FC = (props) => { isSupportedReasoningEffortOpenAIModel(model) && !model.id.includes('o1-pro') && (provider.type === 'openai-response' || provider.id === 'aihubmix') - const isOpenAIFlexServiceTier = isSupportedFlexServiceTier(model) + const isOpenAIFlexServiceTier = isSupportFlexServiceTierModel(model) return (