Add prompt caching to openrouter; remove o1 since it doesn't support tool use

This commit is contained in:
Saoud Rizwan
2024-09-13 00:49:04 -04:00
parent c29fdaa520
commit dc617a92a9
5 changed files with 115 additions and 63 deletions

View File

@@ -70,14 +70,6 @@ export const anthropicModels = {
cacheWritesPrice: 18.75,
cacheReadsPrice: 1.5,
},
"claude-3-sonnet-20240229": {
maxTokens: 4096,
contextWindow: 200_000,
supportsImages: true,
supportsPromptCache: false,
inputPrice: 3.0,
outputPrice: 15.0,
},
"claude-3-haiku-20240307": {
maxTokens: 4096,
contextWindow: 200_000,
@@ -111,14 +103,6 @@ export const bedrockModels = {
inputPrice: 15.0,
outputPrice: 75.0,
},
"anthropic.claude-3-sonnet-20240229-v1:0": {
maxTokens: 4096,
contextWindow: 200_000,
supportsImages: true,
supportsPromptCache: false,
inputPrice: 3.0,
outputPrice: 15.0,
},
"anthropic.claude-3-haiku-20240307-v1:0": {
maxTokens: 4096,
contextWindow: 200_000,
@@ -138,34 +122,49 @@ export const openRouterModels = {
maxTokens: 8192,
contextWindow: 200_000,
supportsImages: true,
supportsPromptCache: false,
supportsPromptCache: true,
inputPrice: 3.0,
outputPrice: 15.0,
cacheWritesPrice: 3.75,
cacheReadsPrice: 0.3,
},
"anthropic/claude-3-opus:beta": {
maxTokens: 4096,
contextWindow: 200_000,
supportsImages: true,
supportsPromptCache: false,
supportsPromptCache: true,
inputPrice: 15,
outputPrice: 75,
},
"anthropic/claude-3-sonnet:beta": {
maxTokens: 4096,
contextWindow: 200_000,
supportsImages: true,
supportsPromptCache: false,
inputPrice: 3,
outputPrice: 15,
cacheWritesPrice: 18.75,
cacheReadsPrice: 1.5,
},
"anthropic/claude-3-haiku:beta": {
maxTokens: 4096,
contextWindow: 200_000,
supportsImages: true,
supportsPromptCache: false,
supportsPromptCache: true,
inputPrice: 0.25,
outputPrice: 1.25,
cacheWritesPrice: 0.3,
cacheReadsPrice: 0.03,
},
// Doesn't support tool use (yet)
// "openai/o1-preview": {
// maxTokens: 32_768,
// contextWindow: 128_000,
// supportsImages: true,
// supportsPromptCache: false,
// inputPrice: 15,
// outputPrice: 60,
// },
// "openai/o1-mini": {
// maxTokens: 65_536,
// contextWindow: 128_000,
// supportsImages: true,
// supportsPromptCache: false,
// inputPrice: 3,
// outputPrice: 12,
// },
"openai/gpt-4o-2024-08-06": {
maxTokens: 16384,
contextWindow: 128_000,
@@ -294,14 +293,6 @@ export const vertexModels = {
inputPrice: 15.0,
outputPrice: 75.0,
},
"claude-3-sonnet@20240229": {
maxTokens: 4096,
contextWindow: 200_000,
supportsImages: true,
supportsPromptCache: false,
inputPrice: 3.0,
outputPrice: 15.0,
},
"claude-3-haiku@20240307": {
maxTokens: 4096,
contextWindow: 200_000,
@@ -347,24 +338,25 @@ export const geminiModels = {
// OpenAI Native
// https://openai.com/api/pricing/
export type OpenAiNativeModelId = keyof typeof openAiNativeModels
export const openAiNativeDefaultModelId: OpenAiNativeModelId = "o1-preview"
export const openAiNativeDefaultModelId: OpenAiNativeModelId = "gpt-4o"
export const openAiNativeModels = {
"o1-preview": {
maxTokens: 32_768,
contextWindow: 128_000,
supportsImages: true,
supportsPromptCache: false,
inputPrice: 15,
outputPrice: 60,
},
"o1-mini": {
maxTokens: 65_536,
contextWindow: 128_000,
supportsImages: true,
supportsPromptCache: false,
inputPrice: 3,
outputPrice: 12,
},
// don't support tool use yet
// "o1-preview": {
// maxTokens: 32_768,
// contextWindow: 128_000,
// supportsImages: true,
// supportsPromptCache: false,
// inputPrice: 15,
// outputPrice: 60,
// },
// "o1-mini": {
// maxTokens: 65_536,
// contextWindow: 128_000,
// supportsImages: true,
// supportsPromptCache: false,
// inputPrice: 3,
// outputPrice: 12,
// },
"gpt-4o": {
maxTokens: 4_096,
contextWindow: 128_000,