mirror of
https://github.com/pacnpal/Roo-Code.git
synced 2025-12-20 04:11:10 -05:00
79 lines
2.3 KiB
TypeScript
79 lines
2.3 KiB
TypeScript
import { Anthropic } from "@anthropic-ai/sdk"
|
|
import { GoogleGenerativeAI } from "@google/generative-ai"
|
|
import { ApiHandler, SingleCompletionHandler } from "../"
|
|
import { ApiHandlerOptions, geminiDefaultModelId, GeminiModelId, geminiModels, ModelInfo } from "../../shared/api"
|
|
import { convertAnthropicMessageToGemini } from "../transform/gemini-format"
|
|
import { ApiStream } from "../transform/stream"
|
|
|
|
export class GeminiHandler implements ApiHandler, SingleCompletionHandler {
|
|
private options: ApiHandlerOptions
|
|
private client: GoogleGenerativeAI
|
|
|
|
constructor(options: ApiHandlerOptions) {
|
|
if (!options.geminiApiKey) {
|
|
throw new Error("API key is required for Google Gemini")
|
|
}
|
|
this.options = options
|
|
this.client = new GoogleGenerativeAI(options.geminiApiKey)
|
|
}
|
|
|
|
async *createMessage(systemPrompt: string, messages: Anthropic.Messages.MessageParam[]): ApiStream {
|
|
const model = this.client.getGenerativeModel({
|
|
model: this.getModel().id,
|
|
systemInstruction: systemPrompt,
|
|
})
|
|
const result = await model.generateContentStream({
|
|
contents: messages.map(convertAnthropicMessageToGemini),
|
|
generationConfig: {
|
|
// maxOutputTokens: this.getModel().info.maxTokens,
|
|
temperature: 0,
|
|
},
|
|
})
|
|
|
|
for await (const chunk of result.stream) {
|
|
yield {
|
|
type: "text",
|
|
text: chunk.text(),
|
|
}
|
|
}
|
|
|
|
const response = await result.response
|
|
yield {
|
|
type: "usage",
|
|
inputTokens: response.usageMetadata?.promptTokenCount ?? 0,
|
|
outputTokens: response.usageMetadata?.candidatesTokenCount ?? 0,
|
|
}
|
|
}
|
|
|
|
getModel(): { id: GeminiModelId; info: ModelInfo } {
|
|
const modelId = this.options.apiModelId
|
|
if (modelId && modelId in geminiModels) {
|
|
const id = modelId as GeminiModelId
|
|
return { id, info: geminiModels[id] }
|
|
}
|
|
return { id: geminiDefaultModelId, info: geminiModels[geminiDefaultModelId] }
|
|
}
|
|
|
|
async completePrompt(prompt: string): Promise<string> {
|
|
try {
|
|
const model = this.client.getGenerativeModel({
|
|
model: this.getModel().id,
|
|
})
|
|
|
|
const result = await model.generateContent({
|
|
contents: [{ role: "user", parts: [{ text: prompt }] }],
|
|
generationConfig: {
|
|
temperature: 0,
|
|
},
|
|
})
|
|
|
|
return result.response.text()
|
|
} catch (error) {
|
|
if (error instanceof Error) {
|
|
throw new Error(`Gemini completion error: ${error.message}`)
|
|
}
|
|
throw error
|
|
}
|
|
}
|
|
}
|