From e1f902c203bc3ef8bcdd8b91469e220552389eaa Mon Sep 17 00:00:00 2001 From: FlyLoongZ <60963681+FlyLoongZ@users.noreply.github.com> Date: Mon, 11 Aug 2025 16:03:30 +0800 Subject: [PATCH] Rename apiCustomParams to customBody --- src/apis/trans.js | 30 +++++++++++++++--------------- src/config/i18n.js | 6 +++--- src/config/index.js | 10 +++++----- src/views/Options/Apis.js | 6 +++--- 4 files changed, 26 insertions(+), 26 deletions(-) diff --git a/src/apis/trans.js b/src/apis/trans.js index 62b70b9..ba5b4f8 100644 --- a/src/apis/trans.js +++ b/src/apis/trans.js @@ -240,7 +240,7 @@ const genOpenAI = ({ model, temperature, maxTokens, - apiCustomParams, + customBody, }) => { // 兼容历史上作为systemPrompt的prompt,如果prompt中不包含带翻译文本,则添加文本到prompt末尾 // if (!prompt.includes(INPUT_PLACE_TEXT)) { @@ -255,7 +255,7 @@ const genOpenAI = ({ .replaceAll(INPUT_PLACE_TO, to) .replaceAll(INPUT_PLACE_TEXT, text); - apiCustomParams = JSON.parse("{" + apiCustomParams + "}"); + customBody = JSON.parse("{" + customBody + "}"); const data = { model, @@ -271,7 +271,7 @@ const genOpenAI = ({ ], temperature, max_completion_tokens: maxTokens, - ...apiCustomParams, + ...customBody, }; const init = { @@ -298,7 +298,7 @@ const genGemini = ({ model, temperature, maxTokens, - apiCustomParams, + customBody, }) => { url = url .replaceAll(INPUT_PLACE_MODEL, model) @@ -312,7 +312,7 @@ const genGemini = ({ .replaceAll(INPUT_PLACE_TO, to) .replaceAll(INPUT_PLACE_TEXT, text); - apiCustomParams = JSON.parse("{" + apiCustomParams + "}"); + customBody = JSON.parse("{" + customBody + "}"); const data = { system_instruction: { @@ -332,7 +332,7 @@ const genGemini = ({ // topP: 0.8, // topK: 10, }, - ...apiCustomParams, + ...customBody, }; const init = { @@ -357,7 +357,7 @@ const genGemini2 = ({ model, temperature, maxTokens, - apiCustomParams, + customBody, }) => { systemPrompt = systemPrompt .replaceAll(INPUT_PLACE_FROM, from) @@ -368,7 +368,7 @@ const genGemini2 = ({ .replaceAll(INPUT_PLACE_TO, to) .replaceAll(INPUT_PLACE_TEXT, text); - apiCustomParams = JSON.parse("{" + apiCustomParams + "}"); + customBody = JSON.parse("{" + customBody + "}"); const data = { model, @@ -384,7 +384,7 @@ const genGemini2 = ({ ], temperature, max_tokens: maxTokens, - ...apiCustomParams, + ...customBody, }; const init = { @@ -410,7 +410,7 @@ const genClaude = ({ model, temperature, maxTokens, - apiCustomParams, + customBody, }) => { systemPrompt = systemPrompt .replaceAll(INPUT_PLACE_FROM, from) @@ -421,7 +421,7 @@ const genClaude = ({ .replaceAll(INPUT_PLACE_TO, to) .replaceAll(INPUT_PLACE_TEXT, text); - apiCustomParams = JSON.parse("{" + apiCustomParams + "}"); + customBody = JSON.parse("{" + customBody + "}"); const data = { model, @@ -434,7 +434,7 @@ const genClaude = ({ ], temperature, max_tokens: maxTokens, - ...apiCustomParams, + ...customBody, }; const init = { @@ -460,7 +460,7 @@ const genOllama = ({ systemPrompt, userPrompt, model, - apiCustomParams, + customBody, }) => { systemPrompt = systemPrompt .replaceAll(INPUT_PLACE_FROM, from) @@ -471,7 +471,7 @@ const genOllama = ({ .replaceAll(INPUT_PLACE_TO, to) .replaceAll(INPUT_PLACE_TEXT, text); - apiCustomParams = JSON.parse("{" + apiCustomParams + "}"); + customBody = JSON.parse("{" + customBody + "}"); const data = { model, @@ -479,7 +479,7 @@ const genOllama = ({ prompt: userPrompt, think: think, stream: false, - ...apiCustomParams, + ...customBody, }; const init = { diff --git a/src/config/i18n.js b/src/config/i18n.js index 487e1c7..fb05652 100644 --- a/src/config/i18n.js +++ b/src/config/i18n.js @@ -228,11 +228,11 @@ export const I18N = { zh: `请求超时时间 (5000-30000ms)`, en: `Request Timeout Time (5000-30000ms)`, }, - api_custom_params: { - zh: `API自定义参数`, + custom_body: { + zh: `自定义Body参数`, en: `API Custom Params`, }, - api_custom_params_help: { + custom_body_help: { zh: `使用JSON格式,例如 "top_p": 0.7`, en: `Use JSON format, for example "top_p": 0.7`, }, diff --git a/src/config/index.js b/src/config/index.js index 1273ebc..184535c 100644 --- a/src/config/index.js +++ b/src/config/index.js @@ -563,7 +563,7 @@ const defaultOpenaiApi = { model: "gpt-4", systemPrompt: `You are a professional, authentic machine translation engine.`, userPrompt: `Translate the following source text from ${INPUT_PLACE_FROM} to ${INPUT_PLACE_TO}. Output translation directly without any additional text.\n\nSource Text: ${INPUT_PLACE_TEXT}\n\nTranslated Text:`, - apiCustomParams: "", + customBody: "", temperature: 0, maxTokens: 256, fetchLimit: 1, @@ -578,7 +578,7 @@ const defaultOllamaApi = { model: "llama3.1", systemPrompt: `You are a professional, authentic machine translation engine.`, userPrompt: `Translate the following source text from ${INPUT_PLACE_FROM} to ${INPUT_PLACE_TO}. Output translation directly without any additional text.\n\nSource Text: ${INPUT_PLACE_TEXT}\n\nTranslated Text:`, - apiCustomParams: "", + customBody: "", think: false, thinkIgnore: `qwen3,deepseek-r1`, fetchLimit: 1, @@ -679,7 +679,7 @@ export const DEFAULT_TRANS_APIS = { model: "gemini-2.5-flash", systemPrompt: `You are a professional, authentic machine translation engine.`, userPrompt: `Translate the following source text from ${INPUT_PLACE_FROM} to ${INPUT_PLACE_TO}. Output translation directly without any additional text.\n\nSource Text: ${INPUT_PLACE_TEXT}\n\nTranslated Text:`, - apiCustomParams: "", + customBody: "", temperature: 0, maxTokens: 2048, fetchLimit: 1, @@ -694,7 +694,7 @@ export const DEFAULT_TRANS_APIS = { model: "gemini-2.0-flash", systemPrompt: `You are a professional, authentic machine translation engine.`, userPrompt: `Translate the following source text from ${INPUT_PLACE_FROM} to ${INPUT_PLACE_TO}. Output translation directly without any additional text.\n\nSource Text: ${INPUT_PLACE_TEXT}\n\nTranslated Text:`, - apiCustomParams: "", + customBody: "", temperature: 0, maxTokens: 2048, fetchLimit: 1, @@ -709,7 +709,7 @@ export const DEFAULT_TRANS_APIS = { model: "claude-3-haiku-20240307", systemPrompt: `You are a professional, authentic machine translation engine.`, userPrompt: `Translate the following source text from ${INPUT_PLACE_FROM} to ${INPUT_PLACE_TO}. Output translation directly without any additional text.\n\nSource Text: ${INPUT_PLACE_TEXT}\n\nTranslated Text:`, - apiCustomParams: "", + customBody: "", temperature: 0, maxTokens: 1024, fetchLimit: 1, diff --git a/src/views/Options/Apis.js b/src/views/Options/Apis.js index 1bc3b88..f64ebaf 100644 --- a/src/views/Options/Apis.js +++ b/src/views/Options/Apis.js @@ -125,7 +125,7 @@ function ApiFields({ translator, api, updateApi, resetApi }) { model = "", systemPrompt = "", userPrompt = "", - apiCustomParams = "", + customBody = "", think = false, thinkIgnore = "", fetchLimit = DEFAULT_FETCH_LIMIT, @@ -278,8 +278,8 @@ function ApiFields({ translator, api, updateApi, resetApi }) {