From d14d90e624032771098e9142c053f633ca8555c3 Mon Sep 17 00:00:00 2001 From: armorbreak001 Date: Wed, 15 Apr 2026 12:39:45 +0800 Subject: [PATCH] fix: set default maxRetries to 0 for OpenAI endpoint to prevent retry delays (fixes #12547) --- packages/api/src/endpoints/openai/llm.ts | 4 ++++ 1 file changed, 4 insertions(+) diff --git a/packages/api/src/endpoints/openai/llm.ts b/packages/api/src/endpoints/openai/llm.ts index a89f6fce44..819ae381f0 100644 --- a/packages/api/src/endpoints/openai/llm.ts +++ b/packages/api/src/endpoints/openai/llm.ts @@ -159,6 +159,10 @@ export function getOpenAILLMConfig({ { streaming, model: modelOptions.model ?? '', + /** Default to 0 retries to avoid long delays from LangChain's + * exponential backoff (up to ~2 min with default maxRetries=6). + * Can be overridden via modelOptions or customParams.defaultParams. */ + maxRetries: 0, }, modelOptions, ) as Partial & Partial & Partial;