update groq
This commit is contained in:
parent
812aa76376
commit
e49265c9c1
@ -78,11 +78,6 @@ class LLMManager implements LLMManagerInterface {
|
||||
this.anthropicProvider = new AnthropicProvider(settings.anthropicProvider.apiKey)
|
||||
this.googleProvider = new GeminiProvider(settings.googleProvider.apiKey)
|
||||
this.groqProvider = new GroqProvider(settings.groqProvider.apiKey)
|
||||
console.log('GrokProvider',
|
||||
settings.grokProvider.apiKey,
|
||||
settings.grokProvider.baseUrl,
|
||||
settings.grokProvider.useCustomUrl
|
||||
)
|
||||
this.grokProvider = new OpenAICompatibleProvider(settings.grokProvider.apiKey,
|
||||
settings.grokProvider.baseUrl && settings.grokProvider.useCustomUrl ?
|
||||
settings.grokProvider.baseUrl
|
||||
|
||||
@ -1034,6 +1034,29 @@ export const siliconFlowEmbeddingModels = {
|
||||
export type GroqModelId = keyof typeof groqModels
|
||||
export const groqDefaultModelId: GroqModelId = "llama-3.3-70b-versatile"
|
||||
export const groqModels = {
|
||||
"meta-llama/llama-4-scout-17b-16e-instruct": {
|
||||
maxTokens: 8192,
|
||||
contextWindow: 131072,
|
||||
supportsImages: false,
|
||||
supportsPromptCache: false,
|
||||
inputPrice: 0,
|
||||
outputPrice: 0,
|
||||
},
|
||||
"meta-llama/llama-4-maverick-17b-128e-instruct": {
|
||||
maxTokens: 8192,
|
||||
contextWindow: 131072,
|
||||
supportsImages: false,
|
||||
supportsPromptCache: false,
|
||||
inputPrice: 0,
|
||||
outputPrice: 0,
|
||||
},
|
||||
"qwen-qwq-32b": {
|
||||
maxTokens: 8192,
|
||||
contextWindow: 131072,
|
||||
supportsImages: false,
|
||||
supportsPromptCache: false,
|
||||
inputPrice: 0,
|
||||
},
|
||||
"llama-3.2-1b-preview": {
|
||||
maxTokens: 4096,
|
||||
contextWindow: 8192,
|
||||
|
||||
Loading…
x
Reference in New Issue
Block a user