From b40ca1df2c8de867af81af4d8fb8e6ccf3a82380 Mon Sep 17 00:00:00 2001 From: Daniel Salazar Date: Tue, 9 Dec 2025 17:00:01 -0800 Subject: [PATCH] Revert: change to ai model names (#2123) * Revert "fix: prevent loot qp if failing ai models? (#2122)" This reverts commit 8d52cefe476a5401ddb9d65573417953f86e736f. * Revert "feat: rename ai models to match norm (#2121)" This reverts commit 9da8aa08493d9d85b1406f68a09caf1d1f7a65ab. --- .../ClaudeProvider/ClaudeProvider.ts | 4 +- .../chat/providers/ClaudeProvider/models.ts | 38 ++++++----- .../DeepSeekProvider/DeepSeekProvider.ts | 2 +- .../chat/providers/DeepSeekProvider/models.ts | 8 +-- .../GeminiProvider/GeminiChatProvider.ts | 2 +- .../chat/providers/GeminiProvider/models.ts | 18 ++---- .../GroqAiProvider/GroqAIProvider.ts | 2 +- .../chat/providers/GroqAiProvider/models.ts | 54 +++++----------- .../MistralAiProvider/MistralAiProvider.ts | 2 +- .../providers/MistralAiProvider/models.ts | 50 +++++---------- .../OpenAiProvider/OpenAiChatProvider.ts | 2 +- .../chat/providers/OpenAiProvider/models.ts | 63 +++++++------------ .../OpenRouterProvider/OpenRouterProvider.ts | 2 +- .../chat/providers/XAIProvider/XAIProvider.ts | 2 +- .../ai/chat/providers/XAIProvider/models.ts | 27 +++----- 15 files changed, 102 insertions(+), 174 deletions(-) diff --git a/src/backend/src/services/ai/chat/providers/ClaudeProvider/ClaudeProvider.ts b/src/backend/src/services/ai/chat/providers/ClaudeProvider/ClaudeProvider.ts index 59e349263..24bce2374 100644 --- a/src/backend/src/services/ai/chat/providers/ClaudeProvider/ClaudeProvider.ts +++ b/src/backend/src/services/ai/chat/providers/ClaudeProvider/ClaudeProvider.ts @@ -55,7 +55,7 @@ export class ClaudeProvider implements IChatProvider { }); } getDefaultModel () { - return 'claude/claude-haiku-4-5-20251001'; + return 'claude-haiku-4-5-20251001'; } async list () { @@ -318,4 +318,4 @@ export class ClaudeProvider implements IChatProvider { checkModeration (_text: string): ReturnType { throw new Error('CheckModeration Not provided.'); } -} +} \ No newline at end of file diff --git a/src/backend/src/services/ai/chat/providers/ClaudeProvider/models.ts b/src/backend/src/services/ai/chat/providers/ClaudeProvider/models.ts index 6d55176c3..5dfc8281f 100644 --- a/src/backend/src/services/ai/chat/providers/ClaudeProvider/models.ts +++ b/src/backend/src/services/ai/chat/providers/ClaudeProvider/models.ts @@ -2,8 +2,8 @@ import { IChatModel } from '../types'; export const CLAUDE_MODELS: IChatModel[] = [ { - id: 'claude/claude-opus-4-5-20251101', - aliases: ['claude-opus-4-5-20251101', 'claude-opus-4-5-latest', 'claude-opus-4-5', 'claude-opus-4.5'], + id: 'claude-opus-4-5-20251101', + aliases: ['claude-opus-4-5-latest', 'claude-opus-4-5', 'claude-opus-4.5'], name: 'Claude Opus 4.5', costs_currency: 'usd-cents', input_cost_key: 'input_tokens', @@ -20,8 +20,8 @@ export const CLAUDE_MODELS: IChatModel[] = [ max_tokens: 64000, }, { - id: 'claude/claude-haiku-4-5-20251001', - aliases: ['claude-haiku-4-5-20251001', 'claude-haiku-4.5', 'claude-haiku-4-5', 'claude-4-5-haiku'], + id: 'claude-haiku-4-5-20251001', + aliases: ['claude-haiku-4.5', 'claude-haiku-4-5', 'claude-4-5-haiku'], name: 'Claude Haiku 4.5', costs_currency: 'usd-cents', input_cost_key: 'input_tokens', @@ -38,8 +38,8 @@ export const CLAUDE_MODELS: IChatModel[] = [ max_tokens: 64000, }, { - id: 'claude/claude-sonnet-4-5-20250929', - aliases: ['claude-sonnet-4-5-20250929', 'claude-sonnet-4.5', 'claude-sonnet-4-5'], + id: 'claude-sonnet-4-5-20250929', + aliases: ['claude-sonnet-4.5', 'claude-sonnet-4-5'], name: 'Claude Sonnet 4.5', costs_currency: 'usd-cents', input_cost_key: 'input_tokens', @@ -56,8 +56,8 @@ export const CLAUDE_MODELS: IChatModel[] = [ max_tokens: 64000, }, { - id: 'claude/claude-opus-4-1-20250805', - aliases: ['claude-opus-4-1-20250805', 'claude-opus-4-1'], + id: 'claude-opus-4-1-20250805', + aliases: ['claude-opus-4-1'], name: 'Claude Opus 4.1', costs_currency: 'usd-cents', input_cost_key: 'input_tokens', @@ -74,8 +74,8 @@ export const CLAUDE_MODELS: IChatModel[] = [ max_tokens: 32000, }, { - id: 'claude/claude-opus-4-20250514', - aliases: ['claude-opus-4-20250514', 'claude-opus-4', 'claude-opus-4-latest'], + id: 'claude-opus-4-20250514', + aliases: ['claude-opus-4', 'claude-opus-4-latest'], name: 'Claude Opus 4', costs_currency: 'usd-cents', input_cost_key: 'input_tokens', @@ -92,8 +92,8 @@ export const CLAUDE_MODELS: IChatModel[] = [ max_tokens: 32000, }, { - id: 'claude/claude-sonnet-4-20250514', - aliases: ['claude-sonnet-4-20250514', 'claude-sonnet-4', 'claude-sonnet-4-latest'], + id: 'claude-sonnet-4-20250514', + aliases: ['claude-sonnet-4', 'claude-sonnet-4-latest'], name: 'Claude Sonnet 4', costs_currency: 'usd-cents', input_cost_key: 'input_tokens', @@ -110,8 +110,8 @@ export const CLAUDE_MODELS: IChatModel[] = [ max_tokens: 64000, }, { - id: 'claude/claude-3-7-sonnet-20250219', - aliases: ['claude-3-7-sonnet-20250219', 'claude-3-7-sonnet-latest'], + id: 'claude-3-7-sonnet-20250219', + aliases: ['claude-3-7-sonnet-latest'], succeeded_by: 'claude-sonnet-4-20250514', costs_currency: 'usd-cents', input_cost_key: 'input_tokens', @@ -128,9 +128,9 @@ export const CLAUDE_MODELS: IChatModel[] = [ max_tokens: 8192, }, { - id: 'claude/claude-3-5-sonnet-20241022', + id: 'claude-3-5-sonnet-20241022', name: 'Claude 3.5 Sonnet', - aliases: ['claude-3-5-sonnet-20241022', 'claude-3-5-sonnet-latest'], + aliases: ['claude-3-5-sonnet-latest'], costs_currency: 'usd-cents', input_cost_key: 'input_tokens', output_cost_key: 'output_tokens', @@ -148,9 +148,8 @@ export const CLAUDE_MODELS: IChatModel[] = [ max_tokens: 8192, }, { - id: 'claude/claude-3-5-sonnet-20240620', + id: 'claude-3-5-sonnet-20240620', succeeded_by: 'claude-3-5-sonnet-20241022', - aliases: ['claude-3-5-sonnet-20240620'], costs_currency: 'usd-cents', input_cost_key: 'input_tokens', output_cost_key: 'output_tokens', @@ -166,8 +165,7 @@ export const CLAUDE_MODELS: IChatModel[] = [ max_tokens: 8192, }, { - id: 'claude/claude-3-haiku-20240307', - aliases: ['claude-3-haiku-20240307'], + id: 'claude-3-haiku-20240307', costs_currency: 'usd-cents', input_cost_key: 'input_tokens', output_cost_key: 'output_tokens', diff --git a/src/backend/src/services/ai/chat/providers/DeepSeekProvider/DeepSeekProvider.ts b/src/backend/src/services/ai/chat/providers/DeepSeekProvider/DeepSeekProvider.ts index 5396e915b..7c1624761 100644 --- a/src/backend/src/services/ai/chat/providers/DeepSeekProvider/DeepSeekProvider.ts +++ b/src/backend/src/services/ai/chat/providers/DeepSeekProvider/DeepSeekProvider.ts @@ -40,7 +40,7 @@ export class DeepSeekProvider implements IChatProvider { } getDefaultModel () { - return 'deepseek/deepseek-chat'; + return 'deepseek-chat'; } models () { diff --git a/src/backend/src/services/ai/chat/providers/DeepSeekProvider/models.ts b/src/backend/src/services/ai/chat/providers/DeepSeekProvider/models.ts index 95f7112f1..f93e3bbf5 100644 --- a/src/backend/src/services/ai/chat/providers/DeepSeekProvider/models.ts +++ b/src/backend/src/services/ai/chat/providers/DeepSeekProvider/models.ts @@ -2,9 +2,9 @@ import { IChatModel } from '../types.js'; export const DEEPSEEK_MODELS: IChatModel[] = [ { - id: 'deepseek/deepseek-chat', + id: 'deepseek-chat', name: 'DeepSeek Chat', - aliases: ['deepseek-chat'], + aliases: [], context: 128000, costs_currency: 'usd-cents', input_cost_key: 'prompt_tokens', @@ -18,9 +18,9 @@ export const DEEPSEEK_MODELS: IChatModel[] = [ max_tokens: 8000, }, { - id: 'deepseek/deepseek-reasoner', + id: 'deepseek-reasoner', name: 'DeepSeek Reasoner', - aliases: ['deepseek-reasoner'], + aliases: [], context: 128000, costs_currency: 'usd-cents', input_cost_key: 'prompt_tokens', diff --git a/src/backend/src/services/ai/chat/providers/GeminiProvider/GeminiChatProvider.ts b/src/backend/src/services/ai/chat/providers/GeminiProvider/GeminiChatProvider.ts index bbb0e7ce8..ce92192b6 100644 --- a/src/backend/src/services/ai/chat/providers/GeminiProvider/GeminiChatProvider.ts +++ b/src/backend/src/services/ai/chat/providers/GeminiProvider/GeminiChatProvider.ts @@ -13,7 +13,7 @@ export class GeminiChatProvider implements IChatProvider { meteringService: MeteringService; openai: OpenAI; - defaultModel = 'gemini/gemini-2.5-flash'; + defaultModel = 'gemini-2.5-flash'; constructor ( meteringService: MeteringService, config: { apiKey: string }) { diff --git a/src/backend/src/services/ai/chat/providers/GeminiProvider/models.ts b/src/backend/src/services/ai/chat/providers/GeminiProvider/models.ts index 9335439a0..ed7c70e8c 100644 --- a/src/backend/src/services/ai/chat/providers/GeminiProvider/models.ts +++ b/src/backend/src/services/ai/chat/providers/GeminiProvider/models.ts @@ -2,9 +2,8 @@ import { IChatModel } from '../types'; export const GEMINI_MODELS: IChatModel[] = [ { - id: 'gemini/gemini-2.0-flash', + id: 'gemini-2.0-flash', name: 'Gemini 2.0 Flash', - aliases: ['gemini-2.0-flash'], context: 131072, costs_currency: 'usd-cents', input_cost_key: 'prompt_tokens', @@ -19,9 +18,8 @@ export const GEMINI_MODELS: IChatModel[] = [ max_tokens: 8192, }, { - id: 'gemini/gemini-2.0-flash-lite', + id: 'gemini-2.0-flash-lite', name: 'Gemini 2.0 Flash-Lite', - aliases: ['gemini-2.0-flash-lite'], context: 1_048_576, costs_currency: 'usd-cents', input_cost_key: 'prompt_tokens', @@ -34,9 +32,8 @@ export const GEMINI_MODELS: IChatModel[] = [ max_tokens: 8192, }, { - id: 'gemini/gemini-2.5-flash', + id: 'gemini-2.5-flash', name: 'Gemini 2.5 Flash', - aliases: ['gemini-2.5-flash'], context: 1_048_576, costs_currency: 'usd-cents', input_cost_key: 'prompt_tokens', @@ -50,9 +47,8 @@ export const GEMINI_MODELS: IChatModel[] = [ max_tokens: 65536, }, { - id: 'gemini/gemini-2.5-flash-lite', + id: 'gemini-2.5-flash-lite', name: 'Gemini 2.5 Flash-Lite', - aliases: ['gemini-2.5-flash-lite'], context: 1_048_576, costs_currency: 'usd-cents', input_cost_key: 'prompt_tokens', @@ -66,9 +62,8 @@ export const GEMINI_MODELS: IChatModel[] = [ max_tokens: 65536, }, { - id: 'gemini/gemini-2.5-pro', + id: 'gemini-2.5-pro', name: 'Gemini 2.5 Pro', - aliases: ['gemini-2.5-pro'], context: 1_048_576, costs_currency: 'usd-cents', input_cost_key: 'prompt_tokens', @@ -82,9 +77,8 @@ export const GEMINI_MODELS: IChatModel[] = [ max_tokens: 200_000, }, { - id: 'gemini/gemini-3-pro-preview', + id: 'gemini-3-pro-preview', name: 'Gemini 3 Pro', - aliases: ['gemini-3-pro-preview'], context: 1_048_576, costs_currency: 'usd-cents', input_cost_key: 'prompt_tokens', diff --git a/src/backend/src/services/ai/chat/providers/GroqAiProvider/GroqAIProvider.ts b/src/backend/src/services/ai/chat/providers/GroqAiProvider/GroqAIProvider.ts index 05f03dd1e..309b4e2ac 100644 --- a/src/backend/src/services/ai/chat/providers/GroqAiProvider/GroqAIProvider.ts +++ b/src/backend/src/services/ai/chat/providers/GroqAiProvider/GroqAIProvider.ts @@ -39,7 +39,7 @@ export class GroqAIProvider implements IChatProvider { } getDefaultModel () { - return 'groq/llama-3.1-8b-instant'; + return 'llama-3.1-8b-instant'; } models () { diff --git a/src/backend/src/services/ai/chat/providers/GroqAiProvider/models.ts b/src/backend/src/services/ai/chat/providers/GroqAiProvider/models.ts index 79e3d0a9d..edecb95bb 100644 --- a/src/backend/src/services/ai/chat/providers/GroqAiProvider/models.ts +++ b/src/backend/src/services/ai/chat/providers/GroqAiProvider/models.ts @@ -6,7 +6,6 @@ const makeModel = ({ context, input, output, - aliases, max_tokens, }: { id: string; @@ -14,12 +13,10 @@ const makeModel = ({ context?: number; input: number; output: number; - aliases?: string[]; max_tokens?: number; }): IChatModel => ({ id, name, - aliases, context, costs_currency: 'usd-cents', input_cost_key: 'prompt_tokens', @@ -35,56 +32,49 @@ const makeModel = ({ export const GROQ_MODELS: IChatModel[] = [ makeModel({ - id: 'groq/gemma2-9b-it', - aliases: ['gemma2-9b-it'], + id: 'gemma2-9b-it', name: 'Gemma 2 9B 8k', context: 8192, input: 20, output: 20, }), makeModel({ - id: 'groq/gemma-7b-it', - aliases: ['gemma-7b-it'], + id: 'gemma-7b-it', name: 'Gemma 7B 8k Instruct', context: 8192, input: 7, output: 7, }), makeModel({ - id: 'groq/llama3-groq-70b-8192-tool-use-preview', - aliases: ['llama3-groq-70b-8192-tool-use-preview'], + id: 'llama3-groq-70b-8192-tool-use-preview', name: 'Llama 3 Groq 70B Tool Use Preview 8k', context: 8192, input: 89, output: 89, }), makeModel({ - id: 'groq/llama3-groq-8b-8192-tool-use-preview', - aliases: ['llama3-groq-8b-8192-tool-use-preview'], + id: 'llama3-groq-8b-8192-tool-use-preview', name: 'Llama 3 Groq 8B Tool Use Preview 8k', context: 8192, input: 19, output: 19, }), makeModel({ - id: 'groq/llama-3.1-70b-versatile', - aliases: ['llama-3.1-70b-versatile'], + id: 'llama-3.1-70b-versatile', name: 'Llama 3.1 70B Versatile 128k', context: 128000, input: 59, output: 79, }), makeModel({ - id: 'groq/llama-3.1-70b-specdec', - aliases: ['llama-3.1-70b-specdec'], + id: 'llama-3.1-70b-specdec', name: 'Llama 3.1 8B Instant 128k', context: 128000, input: 59, output: 99, }), makeModel({ - id: 'groq/llama-3.1-8b-instant', - aliases: ['llama-3.1-8b-instant'], + id: 'llama-3.1-8b-instant', name: 'Llama 3.1 8B Instant 128k', context: 131072, input: 5, @@ -92,8 +82,7 @@ export const GROQ_MODELS: IChatModel[] = [ max_tokens: 131072, }), makeModel({ - id: 'groq/meta-llama/llama-guard-4-12b', - aliases: ['meta-llama/llama-guard-4-12b'], + id: 'meta-llama/llama-guard-4-12b', name: 'Llama Guard 4 12B', context: 131072, input: 20, @@ -101,8 +90,7 @@ export const GROQ_MODELS: IChatModel[] = [ max_tokens: 1024, }), makeModel({ - id: 'groq/meta-llama/llama-prompt-guard-2-86m', - aliases: ['meta-llama/llama-prompt-guard-2-86m'], + id: 'meta-llama/llama-prompt-guard-2-86m', name: 'Prompt Guard 2 86M', context: 512, input: 4, @@ -110,64 +98,56 @@ export const GROQ_MODELS: IChatModel[] = [ max_tokens: 512, }), makeModel({ - id: 'groq/llama-3.2-1b-preview', - aliases: ['llama-3.2-1b-preview'], + id: 'llama-3.2-1b-preview', name: 'Llama 3.2 1B (Preview) 8k', context: 128000, input: 4, output: 4, }), makeModel({ - id: 'groq/llama-3.2-3b-preview', - aliases: ['llama-3.2-3b-preview'], + id: 'llama-3.2-3b-preview', name: 'Llama 3.2 3B (Preview) 8k', context: 128000, input: 6, output: 6, }), makeModel({ - id: 'groq/llama-3.2-11b-vision-preview', - aliases: ['llama-3.2-11b-vision-preview'], + id: 'llama-3.2-11b-vision-preview', name: 'Llama 3.2 11B Vision 8k (Preview)', context: 8000, input: 18, output: 18, }), makeModel({ - id: 'groq/llama-3.2-90b-vision-preview', - aliases: ['llama-3.2-90b-vision-preview'], + id: 'llama-3.2-90b-vision-preview', name: 'Llama 3.2 90B Vision 8k (Preview)', context: 8000, input: 90, output: 90, }), makeModel({ - id: 'groq/llama3-70b-8192', - aliases: ['llama3-70b-8192'], + id: 'llama3-70b-8192', name: 'Llama 3 70B 8k', context: 8192, input: 59, output: 79, }), makeModel({ - id: 'groq/llama3-8b-8192', - aliases: ['llama3-8b-8192'], + id: 'llama3-8b-8192', name: 'Llama 3 8B 8k', context: 8192, input: 5, output: 8, }), makeModel({ - id: 'groq/mixtral-8x7b-32768', - aliases: ['mixtral-8x7b-32768'], + id: 'mixtral-8x7b-32768', name: 'Mixtral 8x7B Instruct 32k', context: 32768, input: 24, output: 24, }), makeModel({ - id: 'groq/llama-guard-3-8b', - aliases: ['llama-guard-3-8b'], + id: 'llama-guard-3-8b', name: 'Llama Guard 3 8B 8k', context: 8192, input: 20, diff --git a/src/backend/src/services/ai/chat/providers/MistralAiProvider/MistralAiProvider.ts b/src/backend/src/services/ai/chat/providers/MistralAiProvider/MistralAiProvider.ts index bcd1471a0..f676a7a42 100644 --- a/src/backend/src/services/ai/chat/providers/MistralAiProvider/MistralAiProvider.ts +++ b/src/backend/src/services/ai/chat/providers/MistralAiProvider/MistralAiProvider.ts @@ -38,7 +38,7 @@ export class MistralAIProvider implements IChatProvider { } getDefaultModel () { - return 'mistral/mistral-small-2506'; + return 'mistral-small-2506'; } async models () { diff --git a/src/backend/src/services/ai/chat/providers/MistralAiProvider/models.ts b/src/backend/src/services/ai/chat/providers/MistralAiProvider/models.ts index f86b20247..54b58f942 100644 --- a/src/backend/src/services/ai/chat/providers/MistralAiProvider/models.ts +++ b/src/backend/src/services/ai/chat/providers/MistralAiProvider/models.ts @@ -2,10 +2,9 @@ import { IChatModel } from '../types'; export const MISTRAL_MODELS: IChatModel[] = [ { - id: 'mistral/mistral-medium-2508', + id: 'mistral-medium-2508', name: 'mistral-medium-2508', aliases: [ - 'mistral-medium-2508', 'mistral-medium-latest', 'mistral-medium', ], @@ -22,10 +21,9 @@ export const MISTRAL_MODELS: IChatModel[] = [ }, }, { - id: 'mistral/open-mistral-7b', + id: 'open-mistral-7b', name: 'open-mistral-7b', aliases: [ - 'open-mistral-7b', 'mistral-tiny', 'mistral-tiny-2312', ], @@ -42,10 +40,9 @@ export const MISTRAL_MODELS: IChatModel[] = [ }, }, { - id: 'mistral/open-mistral-nemo', + id: 'open-mistral-nemo', name: 'open-mistral-nemo', aliases: [ - 'open-mistral-nemo', 'open-mistral-nemo-2407', 'mistral-tiny-2407', 'mistral-tiny-latest', @@ -63,10 +60,9 @@ export const MISTRAL_MODELS: IChatModel[] = [ }, }, { - id: 'mistral/pixtral-large-2411', + id: 'pixtral-large-2411', name: 'pixtral-large-2411', aliases: [ - 'pixtral-large-2411', 'pixtral-large-latest', 'mistral-large-pixtral-2411', ], @@ -83,10 +79,9 @@ export const MISTRAL_MODELS: IChatModel[] = [ }, }, { - id: 'mistral/codestral-2508', + id: 'codestral-2508', name: 'codestral-2508', aliases: [ - 'codestral-2508', 'codestral-latest', ], max_tokens: 256000, @@ -102,10 +97,9 @@ export const MISTRAL_MODELS: IChatModel[] = [ }, }, { - id: 'mistral/devstral-small-2507', + id: 'devstral-small-2507', name: 'devstral-small-2507', aliases: [ - 'devstral-small-2507', 'devstral-small-latest', ], max_tokens: 131072, @@ -122,10 +116,9 @@ export const MISTRAL_MODELS: IChatModel[] = [ }, }, { - id: 'mistral/devstral-medium-2507', + id: 'devstral-medium-2507', name: 'devstral-medium-2507', aliases: [ - 'devstral-medium-2507', 'devstral-medium-latest', ], max_tokens: 131072, @@ -142,10 +135,9 @@ export const MISTRAL_MODELS: IChatModel[] = [ }, }, { - id: 'mistral/mistral-small-2506', + id: 'mistral-small-2506', name: 'mistral-small-2506', aliases: [ - 'mistral-small-2506', 'mistral-small-latest', ], max_tokens: 131072, @@ -161,10 +153,9 @@ export const MISTRAL_MODELS: IChatModel[] = [ }, }, { - id: 'mistral/magistral-medium-2509', + id: 'magistral-medium-2509', name: 'magistral-medium-2509', aliases: [ - 'magistral-medium-2509', 'magistral-medium-latest', ], max_tokens: 131072, @@ -180,10 +171,9 @@ export const MISTRAL_MODELS: IChatModel[] = [ }, }, { - id: 'mistral/magistral-small-2509', + id: 'magistral-small-2509', name: 'magistral-small-2509', aliases: [ - 'magistral-small-2509', 'magistral-small-latest', ], max_tokens: 131072, @@ -199,10 +189,9 @@ export const MISTRAL_MODELS: IChatModel[] = [ }, }, { - id: 'mistral/voxtral-mini-2507', + id: 'voxtral-mini-2507', name: 'voxtral-mini-2507', aliases: [ - 'voxtral-mini-2507', 'voxtral-mini-latest', ], max_tokens: 32768, @@ -218,10 +207,9 @@ export const MISTRAL_MODELS: IChatModel[] = [ }, }, { - id: 'mistral/voxtral-small-2507', + id: 'voxtral-small-2507', name: 'voxtral-small-2507', aliases: [ - 'voxtral-small-2507', 'voxtral-small-latest', ], max_tokens: 32768, @@ -237,10 +225,9 @@ export const MISTRAL_MODELS: IChatModel[] = [ }, }, { - id: 'mistral/mistral-large-latest', + id: 'mistral-large-latest', name: 'mistral-large-2512', aliases: [ - 'mistral-large-latest', 'mistral-large-2512', ], max_tokens: 262144, @@ -256,10 +243,9 @@ export const MISTRAL_MODELS: IChatModel[] = [ }, }, { - id: 'mistral/ministral-3b-2512', + id: 'ministral-3b-2512', name: 'ministral-3b-2512', aliases: [ - 'ministral-3b-2512', 'ministral-3b-latest', ], max_tokens: 131072, @@ -275,10 +261,9 @@ export const MISTRAL_MODELS: IChatModel[] = [ }, }, { - id: 'mistral/ministral-8b-2512', + id: 'ministral-8b-2512', name: 'ministral-8b-2512', aliases: [ - 'ministral-8b-2512', 'ministral-8b-latest', ], max_tokens: 262144, @@ -294,10 +279,9 @@ export const MISTRAL_MODELS: IChatModel[] = [ }, }, { - id: 'mistral/ministral-14b-2512', + id: 'ministral-14b-2512', name: 'ministral-14b-2512', aliases: [ - 'ministral-14b-2512', 'ministral-14b-latest', ], max_tokens: 262144, @@ -312,4 +296,4 @@ export const MISTRAL_MODELS: IChatModel[] = [ completion_tokens: 20, }, }, -]; +]; \ No newline at end of file diff --git a/src/backend/src/services/ai/chat/providers/OpenAiProvider/OpenAiChatProvider.ts b/src/backend/src/services/ai/chat/providers/OpenAiProvider/OpenAiChatProvider.ts index b01a217f9..6e72bee6d 100644 --- a/src/backend/src/services/ai/chat/providers/OpenAiProvider/OpenAiChatProvider.ts +++ b/src/backend/src/services/ai/chat/providers/OpenAiProvider/OpenAiChatProvider.ts @@ -50,7 +50,7 @@ export class OpenAiChatProvider implements IChatProvider { */ #openAi: OpenAI; - #defaultModel = 'openai/gpt-5-nano'; + #defaultModel = 'gpt-5-nano'; #meteringService: MeteringService; diff --git a/src/backend/src/services/ai/chat/providers/OpenAiProvider/models.ts b/src/backend/src/services/ai/chat/providers/OpenAiProvider/models.ts index 022d47093..36aeb1650 100644 --- a/src/backend/src/services/ai/chat/providers/OpenAiProvider/models.ts +++ b/src/backend/src/services/ai/chat/providers/OpenAiProvider/models.ts @@ -4,8 +4,7 @@ import { IChatModel } from '../types'; export const OPEN_AI_MODELS: IChatModel[] = [ { - id: 'openai/gpt-5.1', - aliases: ['gpt-5.1'], + id: 'gpt-5.1', costs_currency: 'usd-cents', input_cost_key: 'prompt_tokens', output_cost_key: 'completion_tokens', @@ -18,8 +17,7 @@ export const OPEN_AI_MODELS: IChatModel[] = [ max_tokens: 128000, }, { - id: 'openai/gpt-5.1-codex', - aliases: ['gpt-5.1-codex'], + id: 'gpt-5.1-codex', costs_currency: 'usd-cents', input_cost_key: 'prompt_tokens', output_cost_key: 'completion_tokens', @@ -32,8 +30,7 @@ export const OPEN_AI_MODELS: IChatModel[] = [ max_tokens: 128000, }, { - id: 'openai/gpt-5.1-codex-mini', - aliases: ['gpt-5.1-codex-mini'], + id: 'gpt-5.1-codex-mini', costs_currency: 'usd-cents', input_cost_key: 'prompt_tokens', output_cost_key: 'completion_tokens', @@ -46,8 +43,7 @@ export const OPEN_AI_MODELS: IChatModel[] = [ max_tokens: 128000, }, { - id: 'openai/gpt-5.1-chat-latest', - aliases: ['gpt-5.1-chat-latest'], + id: 'gpt-5.1-chat-latest', costs_currency: 'usd-cents', input_cost_key: 'prompt_tokens', output_cost_key: 'completion_tokens', @@ -60,8 +56,8 @@ export const OPEN_AI_MODELS: IChatModel[] = [ max_tokens: 16384, }, { - id: 'openai/gpt-5-2025-08-07', - aliases: ['gpt-5-2025-08-07', 'gpt-5'], + id: 'gpt-5-2025-08-07', + aliases: ['gpt-5'], costs_currency: 'usd-cents', input_cost_key: 'prompt_tokens', output_cost_key: 'completion_tokens', @@ -74,8 +70,8 @@ export const OPEN_AI_MODELS: IChatModel[] = [ max_tokens: 128000, }, { - id: 'openai/gpt-5-mini-2025-08-07', - aliases: ['gpt-5-mini-2025-08-07', 'gpt-5-mini'], + id: 'gpt-5-mini-2025-08-07', + aliases: ['gpt-5-mini'], costs_currency: 'usd-cents', input_cost_key: 'prompt_tokens', output_cost_key: 'completion_tokens', @@ -88,8 +84,8 @@ export const OPEN_AI_MODELS: IChatModel[] = [ max_tokens: 128000, }, { - id: 'openai/gpt-5-nano-2025-08-07', - aliases: ['gpt-5-nano-2025-08-07', 'gpt-5-nano'], + id: 'gpt-5-nano-2025-08-07', + aliases: ['gpt-5-nano'], costs_currency: 'usd-cents', input_cost_key: 'prompt_tokens', output_cost_key: 'completion_tokens', @@ -102,8 +98,7 @@ export const OPEN_AI_MODELS: IChatModel[] = [ max_tokens: 128000, }, { - id: 'openai/gpt-5-chat-latest', - aliases: ['gpt-5-chat-latest'], + id: 'gpt-5-chat-latest', costs_currency: 'usd-cents', input_cost_key: 'prompt_tokens', output_cost_key: 'completion_tokens', @@ -116,8 +111,7 @@ export const OPEN_AI_MODELS: IChatModel[] = [ max_tokens: 16384, }, { - id: 'openai/gpt-4o', - aliases: ['gpt-4o'], + id: 'gpt-4o', costs_currency: 'usd-cents', input_cost_key: 'prompt_tokens', output_cost_key: 'completion_tokens', @@ -130,8 +124,7 @@ export const OPEN_AI_MODELS: IChatModel[] = [ max_tokens: 16384, }, { - id: 'openai/gpt-4o-mini', - aliases: ['gpt-4o-mini'], + id: 'gpt-4o-mini', max_tokens: 16384, costs_currency: 'usd-cents', input_cost_key: 'prompt_tokens', @@ -144,8 +137,7 @@ export const OPEN_AI_MODELS: IChatModel[] = [ }, }, { - id: 'openai/o1', - aliases: ['o1'], + id: 'o1', costs_currency: 'usd-cents', input_cost_key: 'prompt_tokens', output_cost_key: 'completion_tokens', @@ -158,8 +150,7 @@ export const OPEN_AI_MODELS: IChatModel[] = [ max_tokens: 100000, }, { - id: 'openai/o1-mini', - aliases: ['o1-mini'], + id: 'o1-mini', costs_currency: 'usd-cents', input_cost_key: 'prompt_tokens', output_cost_key: 'completion_tokens', @@ -171,8 +162,7 @@ export const OPEN_AI_MODELS: IChatModel[] = [ max_tokens: 65536, }, { - id: 'openai/o1-pro', - aliases: ['o1-pro'], + id: 'o1-pro', costs_currency: 'usd-cents', input_cost_key: 'prompt_tokens', output_cost_key: 'completion_tokens', @@ -184,8 +174,7 @@ export const OPEN_AI_MODELS: IChatModel[] = [ max_tokens: 100000, }, { - id: 'openai/o3', - aliases: ['o3'], + id: 'o3', costs_currency: 'usd-cents', input_cost_key: 'prompt_tokens', output_cost_key: 'completion_tokens', @@ -198,8 +187,7 @@ export const OPEN_AI_MODELS: IChatModel[] = [ max_tokens: 100000, }, { - id: 'openai/o3-mini', - aliases: ['o3-mini'], + id: 'o3-mini', costs_currency: 'usd-cents', input_cost_key: 'prompt_tokens', output_cost_key: 'completion_tokens', @@ -212,8 +200,7 @@ export const OPEN_AI_MODELS: IChatModel[] = [ max_tokens: 100000, }, { - id: 'openai/o4-mini', - aliases: ['o4-mini'], + id: 'o4-mini', costs_currency: 'usd-cents', input_cost_key: 'prompt_tokens', output_cost_key: 'completion_tokens', @@ -225,8 +212,7 @@ export const OPEN_AI_MODELS: IChatModel[] = [ max_tokens: 100000, }, { - id: 'openai/gpt-4.1', - aliases: ['gpt-4.1'], + id: 'gpt-4.1', costs_currency: 'usd-cents', input_cost_key: 'prompt_tokens', output_cost_key: 'completion_tokens', @@ -239,8 +225,7 @@ export const OPEN_AI_MODELS: IChatModel[] = [ max_tokens: 32768, }, { - id: 'openai/gpt-4.1-mini', - aliases: ['gpt-4.1-mini'], + id: 'gpt-4.1-mini', costs_currency: 'usd-cents', input_cost_key: 'prompt_tokens', output_cost_key: 'completion_tokens', @@ -253,8 +238,7 @@ export const OPEN_AI_MODELS: IChatModel[] = [ max_tokens: 32768, }, { - id: 'openai/gpt-4.1-nano', - aliases: ['gpt-4.1-nano'], + id: 'gpt-4.1-nano', costs_currency: 'usd-cents', input_cost_key: 'prompt_tokens', output_cost_key: 'completion_tokens', @@ -267,8 +251,7 @@ export const OPEN_AI_MODELS: IChatModel[] = [ max_tokens: 32768, }, { - id: 'openai/gpt-4.5-preview', - aliases: ['gpt-4.5-preview'], + id: 'gpt-4.5-preview', costs_currency: 'usd-cents', input_cost_key: 'prompt_tokens', output_cost_key: 'completion_tokens', diff --git a/src/backend/src/services/ai/chat/providers/OpenRouterProvider/OpenRouterProvider.ts b/src/backend/src/services/ai/chat/providers/OpenRouterProvider/OpenRouterProvider.ts index 660e1e37a..73e32b727 100644 --- a/src/backend/src/services/ai/chat/providers/OpenRouterProvider/OpenRouterProvider.ts +++ b/src/backend/src/services/ai/chat/providers/OpenRouterProvider/OpenRouterProvider.ts @@ -138,7 +138,7 @@ export class OpenRouterProvider implements IChatProvider { coerced_models.push({ id: `openrouter:${model.id}`, name: `${model.name} (OpenRouter)`, - aliases: [model.id, `openrouter/${model.id}`, model.id.split('/').slice(1).join('/')], + aliases: [model.id, model.name, `openrouter/${model.id}`, model.id.split('/').slice(1).join('/')], max_tokens: model.top_provider.max_completion_tokens, costs_currency: 'usd-cents', input_cost_key: 'prompt', diff --git a/src/backend/src/services/ai/chat/providers/XAIProvider/XAIProvider.ts b/src/backend/src/services/ai/chat/providers/XAIProvider/XAIProvider.ts index 0935d45b2..97327cc1e 100644 --- a/src/backend/src/services/ai/chat/providers/XAIProvider/XAIProvider.ts +++ b/src/backend/src/services/ai/chat/providers/XAIProvider/XAIProvider.ts @@ -39,7 +39,7 @@ export class XAIProvider implements IChatProvider { } getDefaultModel () { - return 'xai/grok-beta'; + return 'grok-beta'; } models () { diff --git a/src/backend/src/services/ai/chat/providers/XAIProvider/models.ts b/src/backend/src/services/ai/chat/providers/XAIProvider/models.ts index 69dc6d165..a77f0153d 100644 --- a/src/backend/src/services/ai/chat/providers/XAIProvider/models.ts +++ b/src/backend/src/services/ai/chat/providers/XAIProvider/models.ts @@ -6,18 +6,15 @@ const makeModel = ({ context, input, output, - aliases, }: { id: string; name: string; context: number; input: number; output: number; - aliases?: string[]; }): IChatModel => ({ id, name, - aliases, context, costs_currency: 'usd-cents', input_cost_key: 'prompt_tokens', @@ -32,64 +29,56 @@ const makeModel = ({ export const XAI_MODELS: IChatModel[] = [ makeModel({ - id: 'xai/grok-beta', - aliases: ['grok-beta'], + id: 'grok-beta', name: 'Grok Beta', context: 131072, input: 500, output: 1500, }), makeModel({ - id: 'xai/grok-vision-beta', - aliases: ['grok-vision-beta'], + id: 'grok-vision-beta', name: 'Grok Vision Beta', context: 8192, input: 500, output: 1500, }), makeModel({ - id: 'xai/grok-3', - aliases: ['grok-3'], + id: 'grok-3', name: 'Grok 3', context: 131072, input: 300, output: 1500, }), makeModel({ - id: 'xai/grok-3-fast', - aliases: ['grok-3-fast'], + id: 'grok-3-fast', name: 'Grok 3 Fast', context: 131072, input: 500, output: 2500, }), makeModel({ - id: 'xai/grok-3-mini', - aliases: ['grok-3-mini'], + id: 'grok-3-mini', name: 'Grok 3 Mini', context: 131072, input: 30, output: 50, }), makeModel({ - id: 'xai/grok-3-mini-fast', - aliases: ['grok-3-mini-fast'], + id: 'grok-3-mini-fast', name: 'Grok 3 Mini Fast', context: 131072, input: 60, output: 400, }), makeModel({ - id: 'xai/grok-2-vision', - aliases: ['grok-2-vision'], + id: 'grok-2-vision', name: 'Grok 2 Vision', context: 8192, input: 200, output: 1000, }), makeModel({ - id: 'xai/grok-2', - aliases: ['grok-2'], + id: 'grok-2', name: 'Grok 2', context: 131072, input: 200,