From d711cc1f097d9cc26e8af8dcb09a113115fa7360 Mon Sep 17 00:00:00 2001 From: PostHog Bot <69588470+posthog-bot@users.noreply.github.com> Date: Wed, 12 Nov 2025 11:20:44 +0100 Subject: [PATCH] chore(llma): Update LLM costs (#41333) Co-authored-by: Radu-Raicea <15457029+Radu-Raicea@users.noreply.github.com> --- .../ai-costs/providers/canonical-providers.ts | 5 +- .../ai-costs/providers/llm-costs.json | 62 ++++++++++--------- 2 files changed, 38 insertions(+), 29 deletions(-) diff --git a/plugin-server/src/ingestion/ai-costs/providers/canonical-providers.ts b/plugin-server/src/ingestion/ai-costs/providers/canonical-providers.ts index 0644f2c900..5ddf4479c0 100644 --- a/plugin-server/src/ingestion/ai-costs/providers/canonical-providers.ts +++ b/plugin-server/src/ingestion/ai-costs/providers/canonical-providers.ts @@ -1,5 +1,5 @@ // Auto-generated from OpenRouter API - Do not edit manually -// Generated at: 2025-11-11 10:03:37 UTC +// Generated at: 2025-11-12 10:03:39 UTC export type CanonicalProvider = | 'default' @@ -20,6 +20,7 @@ export type CanonicalProvider = | 'chutes' | 'chutes-bf16' | 'chutes-fp8' + | 'chutes-int4' | 'cirrascale' | 'clarifai-fp4' | 'cloudflare' @@ -41,6 +42,7 @@ export type CanonicalProvider = | 'friendli' | 'gmicloud-bf16' | 'gmicloud-fp8' + | 'gmicloud-int4' | 'google-ai-studio' | 'google-vertex' | 'google-vertex-europe' @@ -60,6 +62,7 @@ export type CanonicalProvider = | 'mancer-int4' | 'meta-fp8' | 'minimax' + | 'minimax-fp8' | 'mistral' | 'modelrun' | 'moonshotai' diff --git a/plugin-server/src/ingestion/ai-costs/providers/llm-costs.json b/plugin-server/src/ingestion/ai-costs/providers/llm-costs.json index f528eb2e4e..4f3b971f32 100644 --- a/plugin-server/src/ingestion/ai-costs/providers/llm-costs.json +++ b/plugin-server/src/ingestion/ai-costs/providers/llm-costs.json @@ -154,16 +154,12 @@ "model": "alpindale/goliath-120b", "cost": { "default": { - "prompt_token": 0.000004, - "completion_token": 0.0000055 + "prompt_token": 0.000006, + "completion_token": 0.000008 }, "mancer-int4": { "prompt_token": 0.000006, "completion_token": 0.000008 - }, - "nextbit-int4": { - "prompt_token": 0.000004, - "completion_token": 0.0000055 } } }, @@ -2923,8 +2919,8 @@ "model": "minimax/minimax-m2", "cost": { "default": { - "prompt_token": 2.55e-7, - "completion_token": 0.00000102 + "prompt_token": 2.1e-7, + "completion_token": 8.4e-7 }, "atlas-cloud-fp8": { "prompt_token": 2.8e-7, @@ -2946,13 +2942,13 @@ "prompt_token": 3e-7, "completion_token": 1.2e-7 }, - "minimax": { + "minimax-fp8": { "prompt_token": 2.55e-7, "completion_token": 0.00000102 }, "novita-fp8": { - "prompt_token": 3e-7, - "completion_token": 0.0000012 + "prompt_token": 2.1e-7, + "completion_token": 8.4e-7 }, "siliconflow-fp8": { "prompt_token": 3e-7, @@ -2960,19 +2956,6 @@ } } }, - { - "model": "minimax/minimax-m2:free", - "cost": { - "default": { - "prompt_token": 0, - "completion_token": 0 - }, - "minimax": { - "prompt_token": 0, - "completion_token": 0 - } - } - }, { "model": "mistralai/codestral-2501", "cost": { @@ -3662,9 +3645,20 @@ "model": "moonshotai/kimi-k2-thinking", "cost": { "default": { + "prompt_token": 5.5e-7, + "completion_token": 0.00000225 + }, + "chutes-int4": { + "prompt_token": 5.5e-7, + "completion_token": 0.00000225 + }, + "fireworks": { "prompt_token": 6e-7, - "completion_token": 0.0000025, - "cache_read_token": 1.5e-7 + "completion_token": 0.0000025 + }, + "gmicloud-int4": { + "prompt_token": 8e-7, + "completion_token": 0.0000012 }, "moonshotai-int4": { "prompt_token": 6e-7, @@ -3684,6 +3678,10 @@ "parasail-int4": { "prompt_token": 6e-7, "completion_token": 0.0000025 + }, + "together": { + "prompt_token": 0.0000012, + "completion_token": 0.000004 } } }, @@ -6151,6 +6149,10 @@ "prompt_token": 1.8e-7, "completion_token": 6.9e-7 }, + "nextbit-bf16": { + "prompt_token": 1.2e-7, + "completion_token": 7e-7 + }, "novita-fp8": { "prompt_token": 8e-8, "completion_token": 5e-7 @@ -6365,6 +6367,10 @@ "prompt_token": 3e-7, "completion_token": 5e-7 }, + "nextbit-bf16": { + "prompt_token": 3e-7, + "completion_token": 5e-7 + }, "parasail-bf16": { "prompt_token": 3e-7, "completion_token": 5e-7 @@ -6650,8 +6656,8 @@ "completion_token": 0.0000016 }, "mancer-fp8": { - "prompt_token": 3.5e-7, - "completion_token": 0.000002 + "prompt_token": 4.5e-7, + "completion_token": 0.0000015 }, "nebius-fp8": { "prompt_token": 6e-7,