fix: maxCompletionTokens Implementation for All Providers (#1938)
* Update LLM providers and constants - Updated constants in app/lib/.server/llm/constants.ts - Modified stream-text functionality in app/lib/.server/llm/stream-text.ts - Updated Anthropic provider in app/lib/modules/llm/providers/anthropic.ts - Modified GitHub provider in app/lib/modules/llm/providers/github.ts - Updated Google provider in app/lib/modules/llm/providers/google.ts - Modified OpenAI provider in app/lib/modules/llm/providers/openai.ts - Updated LLM types in app/lib/modules/llm/types.ts - Modified API route in app/routes/api.llmcall.ts * Fix maxCompletionTokens Implementation for All Providers - Cohere: Added maxCompletionTokens: 4000 to all 10 static models - DeepSeek: Added maxCompletionTokens: 8192 to all 3 static models - Groq: Added maxCompletionTokens: 8192 to both static models - Mistral: Added maxCompletionTokens: 8192 to all 9 static models - Together: Added maxCompletionTokens: 8192 to both static models - Groq: Fixed getDynamicModels to include maxCompletionTokens: 8192 - Together: Fixed getDynamicModels to include maxCompletionTokens: 8192 - OpenAI: Fixed getDynamicModels with proper logic for reasoning models (o1: 16384, o1-mini: 8192) and standard models
This commit is contained in:
@@ -17,7 +17,13 @@ export default class GroqProvider extends BaseProvider {
|
||||
* Essential fallback models - only the most stable/reliable ones
|
||||
* Llama 3.1 8B: 128k context, fast and efficient
|
||||
*/
|
||||
{ name: 'llama-3.1-8b-instant', label: 'Llama 3.1 8B', provider: 'Groq', maxTokenAllowed: 128000 },
|
||||
{
|
||||
name: 'llama-3.1-8b-instant',
|
||||
label: 'Llama 3.1 8B',
|
||||
provider: 'Groq',
|
||||
maxTokenAllowed: 128000,
|
||||
maxCompletionTokens: 8192,
|
||||
},
|
||||
|
||||
// Llama 3.3 70B: 128k context, most capable model
|
||||
{
|
||||
@@ -25,6 +31,7 @@ export default class GroqProvider extends BaseProvider {
|
||||
label: 'Llama 3.3 70B',
|
||||
provider: 'Groq',
|
||||
maxTokenAllowed: 128000,
|
||||
maxCompletionTokens: 8192,
|
||||
},
|
||||
];
|
||||
|
||||
@@ -62,6 +69,7 @@ export default class GroqProvider extends BaseProvider {
|
||||
label: `${m.id} - context ${m.context_window ? Math.floor(m.context_window / 1000) + 'k' : 'N/A'} [ by ${m.owned_by}]`,
|
||||
provider: this.name,
|
||||
maxTokenAllowed: Math.min(m.context_window || 8192, 16384),
|
||||
maxCompletionTokens: 8192,
|
||||
}));
|
||||
}
|
||||
|
||||
|
||||
Reference in New Issue
Block a user