Fix max_tokens limit for moonshotai/kimi-k2-instruct on Groq#5740
Merged
Fix max_tokens limit for moonshotai/kimi-k2-instruct on Groq#5740
Conversation
…6384 - Updated maxTokens from 8192 to 16384 for moonshotai/kimi-k2-instruct model - Fixes Groq API error: "max_tokens must be less than or equal to 16384" - Aligns with Groq API limits for this specific model Fixes #5739
mrubens
approved these changes
Jul 15, 2025
fxcl
added a commit
to Anabaai/Roo-Cline
that referenced
this pull request
Jul 16, 2025
* main: fix: Resolve confusing auto-approve checkbox states (RooCodeInc#5602) fix: prevent empty mode names from being saved (RooCodeInc#5766) (RooCodeInc#5794) Format time in ISO 8601 (RooCodeInc#5793) fix: resolve DirectoryScanner memory leak and improve file limit handling (RooCodeInc#5785) Fix settings dirty check (RooCodeInc#5779) feat: increase Ollama API timeout values and extract as constants (RooCodeInc#5778) fix: Exclude Terraform and Terragrunt cache directories from checkpoints (RooCodeInc#4601) (RooCodeInc#5750) Move less commonly used provider settings into an advanced dropdown (RooCodeInc#5762) feat: Add configurable error & repetition limit with unified control (RooCodeInc#5654) (RooCodeInc#5752) list-files must include at least the first-level directory contents (RooCodeInc#5303) Update evals repo link (RooCodeInc#5758) Feature/vertex ai model name conversion (RooCodeInc#5728) fix(litellm): handle baseurl with paths correctly (RooCodeInc#5697) Add telemetry for todos (RooCodeInc#5746) feat: add undo functionality for enhance prompt feature (fixes RooCodeInc#5741) (RooCodeInc#5742) Fix max_tokens limit for moonshotai/kimi-k2-instruct on Groq (RooCodeInc#5740) Changeset version bump (RooCodeInc#5735) Add changeset for v3.23.12 patch release (RooCodeInc#5734) Update the max-token calculation in model-params to use the shared logic (RooCodeInc#5720) Changeset version bump (RooCodeInc#5719) chore: add changeset for v3.23.11 patch release (RooCodeInc#5718) Add Kimi K2 model and better support (RooCodeInc#5717) Fix: Remove invalid skip-checkout parameter from GitHub Actions workflows (RooCodeInc#5676) feat: add Cmd+Shift+. keyboard shortcut for previous mode switching (RooCodeInc#5695) Changeset version bump (RooCodeInc#5708) chore: add changeset for v3.23.10 patch release (RooCodeInc#5707) Add padding to the index model options (RooCodeInc#5706) fix: prioritize built-in model dimensions over custom dimensions (RooCodeInc#5705) Update CHANGELOG.md Changeset version bump (RooCodeInc#5702) chore: add changeset for v3.23.9 patch release (RooCodeInc#5701) Tweaks to command timeout error (RooCodeInc#5700) Update contributors list (RooCodeInc#5639) feat: enable Claude Code provider to run natively on Windows (RooCodeInc#5615) feat: Add configurable timeout for command execution (RooCodeInc#5668) feat: add gemini-embedding-001 model to code-index service (RooCodeInc#5698) fix: resolve vector dimension mismatch error when switching embedding models (RooCodeInc#5616) (RooCodeInc#5617) fix: [5424] return the cwd in the exec tool's response so that the model is not lost after subsequent calls (RooCodeInc#5667) Changeset version bump (RooCodeInc#5670) chore: add changeset for v3.23.8 patch release (RooCodeInc#5669)
This file contains hidden or bidirectional Unicode text that may be interpreted or compiled differently than what appears below. To review, open the file in an editor that reveals hidden Unicode characters.
Learn more about bidirectional Unicode characters
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
Add this suggestion to a batch that can be applied as a single commit.This suggestion is invalid because no changes were made to the code.Suggestions cannot be applied while the pull request is closed.Suggestions cannot be applied while viewing a subset of changes.Only one suggestion per line can be applied in a batch.Add this suggestion to a batch that can be applied as a single commit.Applying suggestions on deleted lines is not supported.You must change the existing code in this line in order to create a valid suggestion.Outdated suggestions cannot be applied.This suggestion has been applied or marked resolved.Suggestions cannot be applied from pending reviews.Suggestions cannot be applied on multi-line comments.Suggestions cannot be applied while the pull request is queued to merge.Suggestion cannot be applied right now. Please check back later.
Summary
This PR fixes the max_tokens error for the
moonshotai/kimi-k2-instructmodel when using Groq as the provider.Problem
Users were encountering the following error when using the
moonshotai/kimi-k2-instructmodel through Groq:Root Cause
The model configuration had
maxTokensset to 8192, but Groq API allows up to 16,384 tokens for this specific model. The error occurred because Roo Code was setting max_tokens to a value that exceeded Groq's API limits for this model.Solution
Updated the
maxTokensvalue formoonshotai/kimi-k2-instructfrom 8192 to 16384 in the Groq provider configuration to align with Groq's API limits.Changes
packages/types/src/providers/groq.ts:maxTokensfrom 8192 to 16384 formoonshotai/kimi-k2-instructTesting
Verification
The fix ensures that the max_tokens parameter sent to Groq API will be within the allowed limit of 16,384 tokens for the
moonshotai/kimi-k2-instructmodel.Fixes #5739
Important
Fixes
max_tokenserror formoonshotai/kimi-k2-instructon Groq by updatingmaxTokensto 16384 ingroq.ts.max_tokenserror formoonshotai/kimi-k2-instructmodel on Groq by updatingmaxTokensto 16384.maxTokensfrom 8192 to 16384 ingroq.tsformoonshotai/kimi-k2-instruct.This description was created by
for ea4860a. You can customize this summary. It will automatically update as commits are pushed.