diff --git a/src/api/anthropic.ts b/src/api/anthropic.ts index bc7f5ee..804b5f2 100644 --- a/src/api/anthropic.ts +++ b/src/api/anthropic.ts @@ -37,6 +37,7 @@ export class AnthropicHandler implements ApiHandler { { model: modelId, max_tokens: this.getModel().info.maxTokens, + temperature: 0.2, system: [{ text: systemPrompt, type: "text", cache_control: { type: "ephemeral" } }], // setting cache breakpoint for system prompt so new tasks can reuse it messages: messages.map((message, index) => { if (index === lastUserMsgIndex || index === secondLastMsgUserIndex) { @@ -89,6 +90,7 @@ export class AnthropicHandler implements ApiHandler { const message = await this.client.messages.create({ model: modelId, max_tokens: this.getModel().info.maxTokens, + temperature: 0.2, system: [{ text: systemPrompt, type: "text" }], messages, tools, diff --git a/src/api/bedrock.ts b/src/api/bedrock.ts index 9bca9d1..342ef84 100644 --- a/src/api/bedrock.ts +++ b/src/api/bedrock.ts @@ -31,6 +31,7 @@ export class AwsBedrockHandler implements ApiHandler { const message = await this.client.messages.create({ model: this.getModel().id, max_tokens: this.getModel().info.maxTokens, + temperature: 0.2, system: systemPrompt, messages, tools, diff --git a/src/api/gemini.ts b/src/api/gemini.ts index 97f627c..7a3249c 100644 --- a/src/api/gemini.ts +++ b/src/api/gemini.ts @@ -39,6 +39,7 @@ export class GeminiHandler implements ApiHandler { contents: messages.map(convertAnthropicMessageToGemini), generationConfig: { maxOutputTokens: this.getModel().info.maxTokens, + temperature: 0.2, }, }) const message = convertGeminiResponseToAnthropic(result.response) diff --git a/src/api/ollama.ts b/src/api/ollama.ts index b1c863e..d70b365 100644 --- a/src/api/ollama.ts +++ b/src/api/ollama.ts @@ -36,6 +36,7 @@ export class OllamaHandler implements ApiHandler { const createParams: OpenAI.Chat.Completions.ChatCompletionCreateParamsNonStreaming = { model: this.options.ollamaModelId ?? "", messages: openAiMessages, + temperature: 0.2, tools: openAiTools, tool_choice: "auto", } diff --git a/src/api/openai-native.ts b/src/api/openai-native.ts index def0618..06d5982 100644 --- a/src/api/openai-native.ts +++ b/src/api/openai-native.ts @@ -48,6 +48,7 @@ export class OpenAiNativeHandler implements ApiHandler { createParams = { model: this.getModel().id, max_completion_tokens: this.getModel().info.maxTokens, + temperature: 0.2, messages: convertToO1Messages(convertToOpenAiMessages(messages), systemPrompt), } break @@ -55,6 +56,7 @@ export class OpenAiNativeHandler implements ApiHandler { createParams = { model: this.getModel().id, max_completion_tokens: this.getModel().info.maxTokens, + temperature: 0.2, messages: openAiMessages, tools: openAiTools, tool_choice: "auto", diff --git a/src/api/openai.ts b/src/api/openai.ts index 4e520d0..82584d9 100644 --- a/src/api/openai.ts +++ b/src/api/openai.ts @@ -47,6 +47,7 @@ export class OpenAiHandler implements ApiHandler { const createParams: OpenAI.Chat.Completions.ChatCompletionCreateParamsNonStreaming = { model: this.options.openAiModelId ?? "", messages: openAiMessages, + temperature: 0.2, tools: openAiTools, tool_choice: "auto", } diff --git a/src/api/openrouter.ts b/src/api/openrouter.ts index e3df5f1..b687b01 100644 --- a/src/api/openrouter.ts +++ b/src/api/openrouter.ts @@ -95,6 +95,7 @@ export class OpenRouterHandler implements ApiHandler { createParams = { model: this.getModel().id, max_tokens: this.getModel().info.maxTokens, + temperature: 0.2, messages: convertToO1Messages(convertToOpenAiMessages(messages), systemPrompt), } break @@ -102,6 +103,7 @@ export class OpenRouterHandler implements ApiHandler { createParams = { model: this.getModel().id, max_tokens: this.getModel().info.maxTokens, + temperature: 0.2, messages: openAiMessages, tools: openAiTools, tool_choice: "auto", diff --git a/src/api/vertex.ts b/src/api/vertex.ts index 9dfbe32..3073e1e 100644 --- a/src/api/vertex.ts +++ b/src/api/vertex.ts @@ -25,6 +25,7 @@ export class VertexHandler implements ApiHandler { const message = await this.client.messages.create({ model: this.getModel().id, max_tokens: this.getModel().info.maxTokens, + temperature: 0.2, system: systemPrompt, messages, tools,