Files
bolt-diy/app/lib/modules/llm/providers/google.ts
Stijnus df242a7935 feat: add Moonshot AI (Kimi) provider and update xAI Grok models (#1953)
- Add comprehensive Moonshot AI provider with 11 models including:
  * Legacy moonshot-v1 series (8k, 32k, 128k context)
  * Latest Kimi K2 models (K2 Preview, Turbo, Thinking)
  * Vision-enabled models for multimodal capabilities
  * Auto-selecting model variants

- Update xAI provider with latest Grok models:
  * Add Grok 4 (256K context) and Grok 4 (07-09) variant
  * Add Grok 3 Mini Beta and Mini Fast Beta variants
  * Update context limits to match actual model capabilities
  * Remove outdated grok-beta and grok-2-1212 models

- Add MOONSHOT_API_KEY to environment configuration
- Register Moonshot provider in service status monitoring
- Full OpenAI-compatible API integration via api.moonshot.ai
- Fix TypeScript errors in GitHub provider

🤖 Generated with [Claude Code](https://claude.ai/code)

Co-authored-by: Claude <noreply@anthropic.com>
2025-08-31 18:54:14 +02:00

148 lines
4.8 KiB
TypeScript

import { BaseProvider } from '~/lib/modules/llm/base-provider';
import type { ModelInfo } from '~/lib/modules/llm/types';
import type { IProviderSetting } from '~/types/model';
import type { LanguageModelV1 } from 'ai';
import { createGoogleGenerativeAI } from '@ai-sdk/google';
export default class GoogleProvider extends BaseProvider {
name = 'Google';
getApiKeyLink = 'https://aistudio.google.com/app/apikey';
config = {
apiTokenKey: 'GOOGLE_GENERATIVE_AI_API_KEY',
};
staticModels: ModelInfo[] = [
/*
* Essential fallback models - only the most reliable/stable ones
* Gemini 1.5 Pro: 2M context, 8K output limit (verified from API docs)
*/
{
name: 'gemini-1.5-pro',
label: 'Gemini 1.5 Pro',
provider: 'Google',
maxTokenAllowed: 2000000,
maxCompletionTokens: 8192,
},
// Gemini 1.5 Flash: 1M context, 8K output limit, fast and cost-effective
{
name: 'gemini-1.5-flash',
label: 'Gemini 1.5 Flash',
provider: 'Google',
maxTokenAllowed: 1000000,
maxCompletionTokens: 8192,
},
];
async getDynamicModels(
apiKeys?: Record<string, string>,
settings?: IProviderSetting,
serverEnv?: Record<string, string>,
): Promise<ModelInfo[]> {
const { apiKey } = this.getProviderBaseUrlAndKey({
apiKeys,
providerSettings: settings,
serverEnv: serverEnv as any,
defaultBaseUrlKey: '',
defaultApiTokenKey: 'GOOGLE_GENERATIVE_AI_API_KEY',
});
if (!apiKey) {
throw `Missing Api Key configuration for ${this.name} provider`;
}
const response = await fetch(`https://generativelanguage.googleapis.com/v1beta/models?key=${apiKey}`, {
headers: {
['Content-Type']: 'application/json',
},
});
if (!response.ok) {
throw new Error(`Failed to fetch models from Google API: ${response.status} ${response.statusText}`);
}
const res = (await response.json()) as any;
if (!res.models || !Array.isArray(res.models)) {
throw new Error('Invalid response format from Google API');
}
// Filter out models with very low token limits and experimental/unstable models
const data = res.models.filter((model: any) => {
const hasGoodTokenLimit = (model.outputTokenLimit || 0) > 8000;
const isStable = !model.name.includes('exp') || model.name.includes('flash-exp');
return hasGoodTokenLimit && isStable;
});
return data.map((m: any) => {
const modelName = m.name.replace('models/', '');
// Get accurate context window from Google API
let contextWindow = 32000; // default fallback
if (m.inputTokenLimit && m.outputTokenLimit) {
// Use the input limit as the primary context window (typically larger)
contextWindow = m.inputTokenLimit;
} else if (modelName.includes('gemini-1.5-pro')) {
contextWindow = 2000000; // Gemini 1.5 Pro has 2M context
} else if (modelName.includes('gemini-1.5-flash')) {
contextWindow = 1000000; // Gemini 1.5 Flash has 1M context
} else if (modelName.includes('gemini-2.0-flash')) {
contextWindow = 1000000; // Gemini 2.0 Flash has 1M context
} else if (modelName.includes('gemini-pro')) {
contextWindow = 32000; // Gemini Pro has 32k context
} else if (modelName.includes('gemini-flash')) {
contextWindow = 32000; // Gemini Flash has 32k context
}
// Cap at reasonable limits to prevent issues
const maxAllowed = 2000000; // 2M tokens max
const finalContext = Math.min(contextWindow, maxAllowed);
// Get completion token limit from Google API
let completionTokens = 8192; // default fallback (Gemini 1.5 standard limit)
if (m.outputTokenLimit && m.outputTokenLimit > 0) {
completionTokens = Math.min(m.outputTokenLimit, 128000); // Use API value, cap at reasonable limit
}
return {
name: modelName,
label: `${m.displayName} (${finalContext >= 1000000 ? Math.floor(finalContext / 1000000) + 'M' : Math.floor(finalContext / 1000) + 'k'} context)`,
provider: this.name,
maxTokenAllowed: finalContext,
maxCompletionTokens: completionTokens,
};
});
}
getModelInstance(options: {
model: string;
serverEnv: any;
apiKeys?: Record<string, string>;
providerSettings?: Record<string, IProviderSetting>;
}): LanguageModelV1 {
const { model, serverEnv, apiKeys, providerSettings } = options;
const { apiKey } = this.getProviderBaseUrlAndKey({
apiKeys,
providerSettings: providerSettings?.[this.name],
serverEnv: serverEnv as any,
defaultBaseUrlKey: '',
defaultApiTokenKey: 'GOOGLE_GENERATIVE_AI_API_KEY',
});
if (!apiKey) {
throw new Error(`Missing API key for ${this.name} provider`);
}
const google = createGoogleGenerativeAI({
apiKey,
});
return google(model);
}
}