From 83a0c273ddbf5436639bbc8a52b538784b6875b0 Mon Sep 17 00:00:00 2001 From: Justin Torre Date: Mon, 13 May 2024 17:58:15 -0700 Subject: [PATCH] fixed cost calc (#1915) --- costs/__tests__/ensureOnlyOne.ts | 4 ++-- costs/src/providers/openai/index.ts | 4 ++-- valhalla/jawn/src/packages/cost/providers/openai/index.ts | 4 ++-- web/packages/cost/providers/mappings.ts | 2 +- web/packages/cost/providers/openai/index.ts | 4 ++-- worker/src/packages/cost/providers/openai/index.ts | 4 ++-- 6 files changed, 11 insertions(+), 11 deletions(-) diff --git a/costs/__tests__/ensureOnlyOne.ts b/costs/__tests__/ensureOnlyOne.ts index 54be3a5fb7..75354f7fd8 100644 --- a/costs/__tests__/ensureOnlyOne.ts +++ b/costs/__tests__/ensureOnlyOne.ts @@ -283,7 +283,7 @@ WHEN (request_response_log.model = 'gpt-4-32k-0613') THEN 60000 * request_respon WHEN (request_response_log.model = 'gpt-4-0125-preview') THEN 10000 * request_response_log.prompt_tokens + 30000 * request_response_log.completion_tokens WHEN (request_response_log.model = 'gpt-4-1106-preview') THEN 10000 * request_response_log.prompt_tokens + 30000 * request_response_log.completion_tokens WHEN (request_response_log.model = 'gpt-4-1106-vision-preview') THEN 10000 * request_response_log.prompt_tokens + 30000 * request_response_log.completion_tokens -WHEN (request_response_log.model = 'gpt-4o') THEN 15000 * request_response_log.prompt_tokens + 30000 * request_response_log.completion_tokens +WHEN (request_response_log.model = 'gpt-4o') THEN 5000 * request_response_log.prompt_tokens + 15000 * request_response_log.completion_tokens WHEN (request_response_log.model = 'gpt-4o-2024-05-13') THEN 15000 * request_response_log.prompt_tokens + 30000 * request_response_log.completion_tokens WHEN (request_response_log.model = 'gpt-3.5-turbo-0613') THEN 1500 * request_response_log.prompt_tokens + 2000 * request_response_log.completion_tokens WHEN (request_response_log.model = 'gpt-35-turbo-16k') THEN 3000 * request_response_log.prompt_tokens + 4000 * request_response_log.completion_tokens @@ -584,7 +584,7 @@ WHEN (request_response_log.model = 'gpt-4-32k-0613') THEN 60000 * request_respon WHEN (request_response_log.model = 'gpt-4-0125-preview') THEN 10000 * request_response_log.prompt_tokens + 30000 * request_response_log.completion_tokens WHEN (request_response_log.model = 'gpt-4-1106-preview') THEN 10000 * request_response_log.prompt_tokens + 30000 * request_response_log.completion_tokens WHEN (request_response_log.model = 'gpt-4-1106-vision-preview') THEN 10000 * request_response_log.prompt_tokens + 30000 * request_response_log.completion_tokens -WHEN (request_response_log.model = 'gpt-4o') THEN 15000 * request_response_log.prompt_tokens + 30000 * request_response_log.completion_tokens +WHEN (request_response_log.model = 'gpt-4o') THEN 5000 * request_response_log.prompt_tokens + 15000 * request_response_log.completion_tokens WHEN (request_response_log.model = 'gpt-4o-2024-05-13') THEN 15000 * request_response_log.prompt_tokens + 30000 * request_response_log.completion_tokens WHEN (request_response_log.model = 'gpt-3.5-turbo-0613') THEN 1500 * request_response_log.prompt_tokens + 2000 * request_response_log.completion_tokens WHEN (request_response_log.model = 'gpt-35-turbo-16k') THEN 3000 * request_response_log.prompt_tokens + 4000 * request_response_log.completion_tokens diff --git a/costs/src/providers/openai/index.ts b/costs/src/providers/openai/index.ts index fe397b954f..e2ac328fd5 100644 --- a/costs/src/providers/openai/index.ts +++ b/costs/src/providers/openai/index.ts @@ -266,8 +266,8 @@ export const costs: ModelRow[] = [ value: "gpt-4o", }, cost: { - prompt_token: 0.000015, - completion_token: 0.00003, + prompt_token: 0.000005, + completion_token: 0.000015, }, showInPlayground: true, }, diff --git a/valhalla/jawn/src/packages/cost/providers/openai/index.ts b/valhalla/jawn/src/packages/cost/providers/openai/index.ts index fe397b954f..e2ac328fd5 100644 --- a/valhalla/jawn/src/packages/cost/providers/openai/index.ts +++ b/valhalla/jawn/src/packages/cost/providers/openai/index.ts @@ -266,8 +266,8 @@ export const costs: ModelRow[] = [ value: "gpt-4o", }, cost: { - prompt_token: 0.000015, - completion_token: 0.00003, + prompt_token: 0.000005, + completion_token: 0.000015, }, showInPlayground: true, }, diff --git a/web/packages/cost/providers/mappings.ts b/web/packages/cost/providers/mappings.ts index 0dd08c3fc5..3065d9b393 100644 --- a/web/packages/cost/providers/mappings.ts +++ b/web/packages/cost/providers/mappings.ts @@ -169,7 +169,7 @@ export const playgroundModels: { })); }) .flat() - .filter((model) => model !== undefined && model.provider !== "AZURE") as { + .filter((model) => model !== undefined) as { name: string; provider: ProviderName; }[]) ?? []; diff --git a/web/packages/cost/providers/openai/index.ts b/web/packages/cost/providers/openai/index.ts index fe397b954f..e2ac328fd5 100644 --- a/web/packages/cost/providers/openai/index.ts +++ b/web/packages/cost/providers/openai/index.ts @@ -266,8 +266,8 @@ export const costs: ModelRow[] = [ value: "gpt-4o", }, cost: { - prompt_token: 0.000015, - completion_token: 0.00003, + prompt_token: 0.000005, + completion_token: 0.000015, }, showInPlayground: true, }, diff --git a/worker/src/packages/cost/providers/openai/index.ts b/worker/src/packages/cost/providers/openai/index.ts index fe397b954f..e2ac328fd5 100644 --- a/worker/src/packages/cost/providers/openai/index.ts +++ b/worker/src/packages/cost/providers/openai/index.ts @@ -266,8 +266,8 @@ export const costs: ModelRow[] = [ value: "gpt-4o", }, cost: { - prompt_token: 0.000015, - completion_token: 0.00003, + prompt_token: 0.000005, + completion_token: 0.000015, }, showInPlayground: true, },