Skip to content

Commit

Permalink
Update TogetherAI costs (#2380)
Browse files Browse the repository at this point in the history
* update groq costs

* update togetherai costs
  • Loading branch information
Anewryzm authored Aug 5, 2024
1 parent 99cb9ec commit b308ff0
Show file tree
Hide file tree
Showing 5 changed files with 287 additions and 0 deletions.
7 changes: 7 additions & 0 deletions costs/__tests__/ensureOnlyOne.ts
Original file line number Diff line number Diff line change
Expand Up @@ -375,6 +375,13 @@ WHEN (request_response_log.model = 'meta-llama/Llama-3-8b-chat-hf') THEN 200 * r
WHEN (request_response_log.model = 'NousResearch/Nous-Hermes-llama-2-7b') THEN 200 * request_response_log.prompt_tokens + 200 * request_response_log.completion_tokens
WHEN (request_response_log.model = 'NousResearch/Nous-Hermes-Llama2-13b') THEN 225 * request_response_log.prompt_tokens + 225 * request_response_log.completion_tokens
WHEN (request_response_log.model = 'togethercomputer/Llama-2-7B-32K-Instruct') THEN 200 * request_response_log.prompt_tokens + 200 * request_response_log.completion_tokens
WHEN (request_response_log.model = 'meta-llama/Meta-Llama-3.1-70B-Instruct-Turbo') THEN 880 * request_response_log.prompt_tokens + 880 * request_response_log.completion_tokens
WHEN (request_response_log.model = 'meta-llama/Meta-Llama-3.1-8B-Instruct-Turbo') THEN 180 * request_response_log.prompt_tokens + 180 * request_response_log.completion_tokens
WHEN (request_response_log.model = 'meta-llama/Meta-Llama-3.1-405B-Instruct-Turbo') THEN 5000 * request_response_log.prompt_tokens + 5000 * request_response_log.completion_tokens
WHEN (request_response_log.model = 'meta-llama/Meta-Llama-3-70B-Instruct-Turbo') THEN 880 * request_response_log.prompt_tokens + 880 * request_response_log.completion_tokens
WHEN (request_response_log.model = 'meta-llama/Meta-Llama-3-8B-Instruct-Turbo') THEN 180 * request_response_log.prompt_tokens + 180 * request_response_log.completion_tokens
WHEN (request_response_log.model = 'meta-llama/Meta-Llama-3-70B-Instruct-Lite') THEN 540 * request_response_log.prompt_tokens + 540 * request_response_log.completion_tokens
WHEN (request_response_log.model = 'meta-llama/Meta-Llama-3-8B-Instruct-Lite') THEN 100 * request_response_log.prompt_tokens + 100 * request_response_log.completion_tokens
WHEN (request_response_log.model = 'zero-one-ai/Yi-34B') THEN 800 * request_response_log.prompt_tokens + 800 * request_response_log.completion_tokens
WHEN (request_response_log.model = 'zero-one-ai/Yi-6B') THEN 200 * request_response_log.prompt_tokens + 200 * request_response_log.completion_tokens
WHEN (request_response_log.model = 'google/gemma-2b') THEN 100 * request_response_log.prompt_tokens + 100 * request_response_log.completion_tokens
Expand Down
70 changes: 70 additions & 0 deletions costs/src/providers/togetherai/chat/llama.ts
Original file line number Diff line number Diff line change
Expand Up @@ -137,4 +137,74 @@ export const costs: ModelRow[] = [
completion_token: 0.0000002,
},
},
{
model: {
operator: "equals",
value: "meta-llama/Meta-Llama-3.1-70B-Instruct-Turbo",
},
cost: {
prompt_token: 0.00000088,
completion_token: 0.00000088,
},
},
{
model: {
operator: "equals",
value: "meta-llama/Meta-Llama-3.1-8B-Instruct-Turbo",
},
cost: {
prompt_token: 0.00000018,
completion_token: 0.00000018,
},
},
{
model: {
operator: "equals",
value: "meta-llama/Meta-Llama-3.1-405B-Instruct-Turbo",
},
cost: {
prompt_token: 0.000005,
completion_token: 0.000005,
},
},
{
model: {
operator: "equals",
value: "meta-llama/Meta-Llama-3-70B-Instruct-Turbo",
},
cost: {
prompt_token: 0.00000088,
completion_token: 0.00000088,
},
},
{
model: {
operator: "equals",
value: "meta-llama/Meta-Llama-3-8B-Instruct-Turbo",
},
cost: {
prompt_token: 0.00000018,
completion_token: 0.00000018,
},
},
{
model: {
operator: "equals",
value: "meta-llama/Meta-Llama-3-70B-Instruct-Lite",
},
cost: {
prompt_token: 0.00000054,
completion_token: 0.00000054,
},
},
{
model: {
operator: "equals",
value: "meta-llama/Meta-Llama-3-8B-Instruct-Lite",
},
cost: {
prompt_token: 0.0000001,
completion_token: 0.0000001,
},
},
];
70 changes: 70 additions & 0 deletions valhalla/jawn/src/packages/cost/providers/togetherai/chat/llama.ts
Original file line number Diff line number Diff line change
Expand Up @@ -137,4 +137,74 @@ export const costs: ModelRow[] = [
completion_token: 0.0000002,
},
},
{
model: {
operator: "equals",
value: "meta-llama/Meta-Llama-3.1-70B-Instruct-Turbo",
},
cost: {
prompt_token: 0.00000088,
completion_token: 0.00000088,
},
},
{
model: {
operator: "equals",
value: "meta-llama/Meta-Llama-3.1-8B-Instruct-Turbo",
},
cost: {
prompt_token: 0.00000018,
completion_token: 0.00000018,
},
},
{
model: {
operator: "equals",
value: "meta-llama/Meta-Llama-3.1-405B-Instruct-Turbo",
},
cost: {
prompt_token: 0.000005,
completion_token: 0.000005,
},
},
{
model: {
operator: "equals",
value: "meta-llama/Meta-Llama-3-70B-Instruct-Turbo",
},
cost: {
prompt_token: 0.00000088,
completion_token: 0.00000088,
},
},
{
model: {
operator: "equals",
value: "meta-llama/Meta-Llama-3-8B-Instruct-Turbo",
},
cost: {
prompt_token: 0.00000018,
completion_token: 0.00000018,
},
},
{
model: {
operator: "equals",
value: "meta-llama/Meta-Llama-3-70B-Instruct-Lite",
},
cost: {
prompt_token: 0.00000054,
completion_token: 0.00000054,
},
},
{
model: {
operator: "equals",
value: "meta-llama/Meta-Llama-3-8B-Instruct-Lite",
},
cost: {
prompt_token: 0.0000001,
completion_token: 0.0000001,
},
},
];
70 changes: 70 additions & 0 deletions web/packages/cost/providers/togetherai/chat/llama.ts
Original file line number Diff line number Diff line change
Expand Up @@ -137,4 +137,74 @@ export const costs: ModelRow[] = [
completion_token: 0.0000002,
},
},
{
model: {
operator: "equals",
value: "meta-llama/Meta-Llama-3.1-70B-Instruct-Turbo",
},
cost: {
prompt_token: 0.00000088,
completion_token: 0.00000088,
},
},
{
model: {
operator: "equals",
value: "meta-llama/Meta-Llama-3.1-8B-Instruct-Turbo",
},
cost: {
prompt_token: 0.00000018,
completion_token: 0.00000018,
},
},
{
model: {
operator: "equals",
value: "meta-llama/Meta-Llama-3.1-405B-Instruct-Turbo",
},
cost: {
prompt_token: 0.000005,
completion_token: 0.000005,
},
},
{
model: {
operator: "equals",
value: "meta-llama/Meta-Llama-3-70B-Instruct-Turbo",
},
cost: {
prompt_token: 0.00000088,
completion_token: 0.00000088,
},
},
{
model: {
operator: "equals",
value: "meta-llama/Meta-Llama-3-8B-Instruct-Turbo",
},
cost: {
prompt_token: 0.00000018,
completion_token: 0.00000018,
},
},
{
model: {
operator: "equals",
value: "meta-llama/Meta-Llama-3-70B-Instruct-Lite",
},
cost: {
prompt_token: 0.00000054,
completion_token: 0.00000054,
},
},
{
model: {
operator: "equals",
value: "meta-llama/Meta-Llama-3-8B-Instruct-Lite",
},
cost: {
prompt_token: 0.0000001,
completion_token: 0.0000001,
},
},
];
70 changes: 70 additions & 0 deletions worker/src/packages/cost/providers/togetherai/chat/llama.ts
Original file line number Diff line number Diff line change
Expand Up @@ -137,4 +137,74 @@ export const costs: ModelRow[] = [
completion_token: 0.0000002,
},
},
{
model: {
operator: "equals",
value: "meta-llama/Meta-Llama-3.1-70B-Instruct-Turbo",
},
cost: {
prompt_token: 0.00000088,
completion_token: 0.00000088,
},
},
{
model: {
operator: "equals",
value: "meta-llama/Meta-Llama-3.1-8B-Instruct-Turbo",
},
cost: {
prompt_token: 0.00000018,
completion_token: 0.00000018,
},
},
{
model: {
operator: "equals",
value: "meta-llama/Meta-Llama-3.1-405B-Instruct-Turbo",
},
cost: {
prompt_token: 0.000005,
completion_token: 0.000005,
},
},
{
model: {
operator: "equals",
value: "meta-llama/Meta-Llama-3-70B-Instruct-Turbo",
},
cost: {
prompt_token: 0.00000088,
completion_token: 0.00000088,
},
},
{
model: {
operator: "equals",
value: "meta-llama/Meta-Llama-3-8B-Instruct-Turbo",
},
cost: {
prompt_token: 0.00000018,
completion_token: 0.00000018,
},
},
{
model: {
operator: "equals",
value: "meta-llama/Meta-Llama-3-70B-Instruct-Lite",
},
cost: {
prompt_token: 0.00000054,
completion_token: 0.00000054,
},
},
{
model: {
operator: "equals",
value: "meta-llama/Meta-Llama-3-8B-Instruct-Lite",
},
cost: {
prompt_token: 0.0000001,
completion_token: 0.0000001,
},
},
];

0 comments on commit b308ff0

Please sign in to comment.