Skip to content
Open
Show file tree
Hide file tree
Changes from all commits
Commits
File filter

Filter by extension

Filter by extension

Conversations
Failed to load comments.
Loading
Jump to
Jump to file
Failed to load files.
Loading
Diff view
Diff view
1 change: 1 addition & 0 deletions .env.example
Original file line number Diff line number Diff line change
Expand Up @@ -6,6 +6,7 @@ ANTHROPIC_API_KEY=dummy_anthropic_key
FIREWORKS_API_KEY=dummy_fireworks_key
CANOPYWAVE_API_KEY=dummy_canopywave_key
SILICONFLOW_API_KEY=dummy_siliconflow_key
MINIMAX_API_KEY=dummy_minimax_key

# Database & Server
DATABASE_URL=postgresql://manicode_user_local:secretpassword_local@localhost:5432/manicode_db_local
Expand Down
5 changes: 5 additions & 0 deletions common/src/constants/model-config.ts
Original file line number Diff line number Diff line change
Expand Up @@ -6,6 +6,7 @@ export const ALLOWED_MODEL_PREFIXES = [
'openai',
'google',
'x-ai',
'minimax',
] as const

export const costModes = [
Expand Down Expand Up @@ -47,6 +48,8 @@ export const openrouterModels = {
openrouter_gemini2_5_flash_thinking:
'google/gemini-2.5-flash-preview:thinking',
openrouter_grok_4: 'x-ai/grok-4-07-09',
openrouter_minimax_m2_5: 'minimax/minimax-m2.5',
openrouter_minimax_m2_5_highspeed: 'minimax/minimax-m2.5-highspeed',
} as const
export type openrouterModel =
(typeof openrouterModels)[keyof typeof openrouterModels]
Expand Down Expand Up @@ -171,6 +174,7 @@ export const providerDomains = {
openai: 'chatgpt.com',
deepseek: 'deepseek.com',
xai: 'x.ai',
minimax: 'minimax.io',
} as const

export function getLogoForModel(modelName: string): string | undefined {
Expand All @@ -182,6 +186,7 @@ export function getLogoForModel(modelName: string): string | undefined {
domain = providerDomains.deepseek
else if (modelName.includes('claude')) domain = providerDomains.anthropic
else if (modelName.includes('grok')) domain = providerDomains.xai
else if (modelName.includes('minimax')) domain = providerDomains.minimax

return domain
? `https://www.google.com/s2/favicons?domain=${domain}&sz=256`
Expand Down
2 changes: 2 additions & 0 deletions packages/internal/src/env-schema.ts
Original file line number Diff line number Diff line change
Expand Up @@ -9,6 +9,7 @@ export const serverEnvSchema = clientEnvSchema.extend({
FIREWORKS_API_KEY: z.string().min(1),
CANOPYWAVE_API_KEY: z.string().min(1).optional(),
SILICONFLOW_API_KEY: z.string().min(1).optional(),
MINIMAX_API_KEY: z.string().min(1).optional(),
LINKUP_API_KEY: z.string().min(1),
CONTEXT7_API_KEY: z.string().optional(),
GRAVITY_API_KEY: z.string().min(1),
Expand Down Expand Up @@ -54,6 +55,7 @@ export const serverProcessEnv: ServerInput = {
FIREWORKS_API_KEY: process.env.FIREWORKS_API_KEY,
CANOPYWAVE_API_KEY: process.env.CANOPYWAVE_API_KEY,
SILICONFLOW_API_KEY: process.env.SILICONFLOW_API_KEY,
MINIMAX_API_KEY: process.env.MINIMAX_API_KEY,
LINKUP_API_KEY: process.env.LINKUP_API_KEY,
CONTEXT7_API_KEY: process.env.CONTEXT7_API_KEY,
GRAVITY_API_KEY: process.env.GRAVITY_API_KEY,
Expand Down
53 changes: 44 additions & 9 deletions web/src/app/api/v1/chat/completions/_post.ts
Original file line number Diff line number Diff line change
Expand Up @@ -53,6 +53,12 @@ import {
handleSiliconFlowStream,
isSiliconFlowModel,
} from '@/llm-api/siliconflow'
import {
MiniMaxError,
handleMiniMaxNonStream,
handleMiniMaxStream,
isMiniMaxModel,
} from '@/llm-api/minimax'
import {
handleOpenAINonStream,
OPENAI_SUPPORTED_MODELS,
Expand Down Expand Up @@ -366,11 +372,22 @@ export async function postChatCompletions(params: {
// Handle streaming vs non-streaming
try {
if (bodyStream) {
// Streaming request — route to SiliconFlow/CanopyWave/Fireworks for supported models
// Streaming request — route to MiniMax/SiliconFlow/CanopyWave/Fireworks for supported models
const useMiniMax = isMiniMaxModel(typedBody.model)
const useSiliconFlow = false // isSiliconFlowModel(typedBody.model)
const useCanopyWave = false // isCanopyWaveModel(typedBody.model)
const useFireworks = isFireworksModel(typedBody.model)
const stream = useSiliconFlow
const useFireworks = !useMiniMax && isFireworksModel(typedBody.model)
const stream = useMiniMax
? await handleMiniMaxStream({
body: typedBody,
userId,
stripeCustomerId,
agentId,
fetch,
logger,
insertMessageBigquery,
})
: useSiliconFlow
? await handleSiliconFlowStream({
body: typedBody,
userId,
Expand Down Expand Up @@ -430,12 +447,13 @@ export async function postChatCompletions(params: {
},
})
} else {
// Non-streaming request — route to SiliconFlow/CanopyWave/Fireworks for supported models
// Non-streaming request — route to MiniMax/SiliconFlow/CanopyWave/Fireworks for supported models
// TEMPORARILY DISABLED: route through OpenRouter
const model = typedBody.model
const useMiniMaxDirect = isMiniMaxModel(model)
const useSiliconFlow = false // isSiliconFlowModel(model)
const useCanopyWave = false // isCanopyWaveModel(model)
const useFireworks = isFireworksModel(model)
const useFireworks = !useMiniMaxDirect && isFireworksModel(model)
const modelParts = model.split('/')
const shortModelName = modelParts.length > 1 ? modelParts[1] : model
const isOpenAIDirectModel =
Expand All @@ -446,7 +464,17 @@ export async function postChatCompletions(params: {
const shouldUseOpenAIEndpoint =
isOpenAIDirectModel && typedBody.codebuff_metadata?.n !== undefined

const nonStreamRequest = useSiliconFlow
const nonStreamRequest = useMiniMaxDirect
? handleMiniMaxNonStream({
body: typedBody,
userId,
stripeCustomerId,
agentId,
fetch,
logger,
insertMessageBigquery,
})
: useSiliconFlow
? handleSiliconFlowNonStream({
body: typedBody,
userId,
Expand Down Expand Up @@ -528,10 +556,14 @@ export async function postChatCompletions(params: {
if (error instanceof SiliconFlowError) {
siliconflowError = error
}
let minimaxError: MiniMaxError | undefined
if (error instanceof MiniMaxError) {
minimaxError = error
}

// Log detailed error information for debugging
const errorDetails = openrouterError?.toJSON()
const providerLabel = siliconflowError ? 'SiliconFlow' : canopywaveError ? 'CanopyWave' : fireworksError ? 'Fireworks' : 'OpenRouter'
const providerLabel = minimaxError ? 'MiniMax' : siliconflowError ? 'SiliconFlow' : canopywaveError ? 'CanopyWave' : fireworksError ? 'Fireworks' : 'OpenRouter'
logger.error(
{
error: getErrorObject(error),
Expand All @@ -545,8 +577,8 @@ export async function postChatCompletions(params: {
? typedBody.messages.length
: 0,
messages: typedBody.messages,
providerStatusCode: (openrouterError ?? fireworksError ?? canopywaveError ?? siliconflowError)?.statusCode,
providerStatusText: (openrouterError ?? fireworksError ?? canopywaveError ?? siliconflowError)?.statusText,
providerStatusCode: (openrouterError ?? fireworksError ?? canopywaveError ?? siliconflowError ?? minimaxError)?.statusCode,
providerStatusText: (openrouterError ?? fireworksError ?? canopywaveError ?? siliconflowError ?? minimaxError)?.statusText,
openrouterErrorCode: errorDetails?.error?.code,
openrouterErrorType: errorDetails?.error?.type,
openrouterErrorMessage: errorDetails?.error?.message,
Expand Down Expand Up @@ -580,6 +612,9 @@ export async function postChatCompletions(params: {
if (error instanceof SiliconFlowError) {
return NextResponse.json(error.toJSON(), { status: error.statusCode })
}
if (error instanceof MiniMaxError) {
return NextResponse.json(error.toJSON(), { status: error.statusCode })
}

return NextResponse.json(
{ error: 'Failed to process request' },
Expand Down
Loading