From 5361f88381b377389d7c0b2154048bf54178a9b8 Mon Sep 17 00:00:00 2001 From: Kirigaya <1193466151@qq.com> Date: Wed, 4 Jun 2025 22:22:07 +0800 Subject: [PATCH] release openmcp-sdk 0.0.5 --- service/src/llm/llm.service.ts | 22 +++++++++++----------- 1 file changed, 11 insertions(+), 11 deletions(-) diff --git a/service/src/llm/llm.service.ts b/service/src/llm/llm.service.ts index 4e9c43c..3539332 100644 --- a/service/src/llm/llm.service.ts +++ b/service/src/llm/llm.service.ts @@ -59,18 +59,18 @@ export async function streamingChatCompletion( await postProcessMessages(messages); - // 使用表格渲染请求参数 - requestTable.push( - ['Model', model], - ['Base URL', baseURL || 'Default'], - ['Temperature', temperature], - ['Tools Count', tools.length], - ['Parallel Tool Calls', parallelToolCalls], - ['Proxy Server', proxyServer || 'No Proxy'] - ); + // // 使用表格渲染请求参数 + // requestTable.push( + // ['Model', model], + // ['Base URL', baseURL || 'Default'], + // ['Temperature', temperature], + // ['Tools Count', tools.length], + // ['Parallel Tool Calls', parallelToolCalls], + // ['Proxy Server', proxyServer || 'No Proxy'] + // ); - console.log('\nOpenAI Request Parameters:'); - console.log(requestTable.toString()); + // console.log('\nOpenAI Request Parameters:'); + // console.log(requestTable.toString()); const stream = await client.chat.completions.create({ model,