- Add comprehensive Moonshot AI provider with 11 models including:
* Legacy moonshot-v1 series (8k, 32k, 128k context)
* Latest Kimi K2 models (K2 Preview, Turbo, Thinking)
* Vision-enabled models for multimodal capabilities
* Auto-selecting model variants
- Update xAI provider with latest Grok models:
* Add Grok 4 (256K context) and Grok 4 (07-09) variant
* Add Grok 3 Mini Beta and Mini Fast Beta variants
* Update context limits to match actual model capabilities
* Remove outdated grok-beta and grok-2-1212 models
- Add MOONSHOT_API_KEY to environment configuration
- Register Moonshot provider in service status monitoring
- Full OpenAI-compatible API integration via api.moonshot.ai
- Fix TypeScript errors in GitHub provider
🤖 Generated with [Claude Code](https://claude.ai/code)
Co-authored-by: Claude <noreply@anthropic.com>
* Update LLM providers and constants
- Updated constants in app/lib/.server/llm/constants.ts
- Modified stream-text functionality in app/lib/.server/llm/stream-text.ts
- Updated Anthropic provider in app/lib/modules/llm/providers/anthropic.ts
- Modified GitHub provider in app/lib/modules/llm/providers/github.ts
- Updated Google provider in app/lib/modules/llm/providers/google.ts
- Modified OpenAI provider in app/lib/modules/llm/providers/openai.ts
- Updated LLM types in app/lib/modules/llm/types.ts
- Modified API route in app/routes/api.llmcall.ts
* Fix maxCompletionTokens Implementation for All Providers
- Cohere: Added maxCompletionTokens: 4000 to all 10 static models
- DeepSeek: Added maxCompletionTokens: 8192 to all 3 static models
- Groq: Added maxCompletionTokens: 8192 to both static models
- Mistral: Added maxCompletionTokens: 8192 to all 9 static models
- Together: Added maxCompletionTokens: 8192 to both static models
- Groq: Fixed getDynamicModels to include maxCompletionTokens: 8192
- Together: Fixed getDynamicModels to include maxCompletionTokens: 8192
- OpenAI: Fixed getDynamicModels with proper logic for reasoning models (o1: 16384, o1-mini: 8192) and standard models
- Updated constants in app/lib/.server/llm/constants.ts
- Modified stream-text functionality in app/lib/.server/llm/stream-text.ts
- Updated Anthropic provider in app/lib/modules/llm/providers/anthropic.ts
- Modified GitHub provider in app/lib/modules/llm/providers/github.ts
- Updated Google provider in app/lib/modules/llm/providers/google.ts
- Modified OpenAI provider in app/lib/modules/llm/providers/openai.ts
- Updated LLM types in app/lib/modules/llm/types.ts
- Modified API route in app/routes/api.llmcall.ts
* fix: updated logger and model caching
* usage token stream issue fix
* minor changes
* updated starter template change to fix the app title
* starter template bigfix
* fixed hydretion errors and raw logs
* removed raw log
* made auto select template false by default
* more cleaner logs and updated logic to call dynamicModels only if not found in static models
* updated starter template instructions
* browser console log improved for firefox
* provider icons fix icons