Skip to content

feat(google-vertex): update model YAMLs [bot]#486

Open
harshiv-26 wants to merge 1 commit intomainfrom
bot/update-google-vertex-20260330-071703
Open

feat(google-vertex): update model YAMLs [bot]#486
harshiv-26 wants to merge 1 commit intomainfrom
bot/update-google-vertex-20260330-071703

Conversation

@harshiv-26
Copy link
Copy Markdown
Collaborator

@harshiv-26 harshiv-26 commented Mar 30, 2026

Auto-generated by poc-agent for provider google-vertex.


Note

Medium Risk
Primarily metadata updates, but changes to status/isDeprecated and supported modalities/features can alter which Vertex models are selectable and how requests are formed.

Overview
Updates Google Vertex model YAML metadata across Anthropic/Claude, Gemini, Imagen, DeepSeek, Llama, Mistral, Qwen, Moonshot, OpenAI, and Z.ai entries.

The diff adds/standardizes lifecycle fields (e.g., status: active|deprecated|preview, plus some isDeprecated and new deprecationDates) and adjusts declared capabilities (new/removed features, modality tweaks, embedding outputs, and a few parameter/cost field additions such as video token pricing and tool_use_system_prompt_tokens).

Written by Cursor Bugbot for commit e24fa5a. This will update automatically on new commits. Configure here.

Copy link
Copy Markdown

@cursor cursor bot left a comment

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Cursor Bugbot has reviewed your changes and found 2 potential issues.

Fix All in Cursor

Bugbot Autofix is OFF. To automatically fix reported issues with cloud agents, enable autofix in the Cursor dashboard.

limits:
context_window: 128000
max_input_tokens: 128000
max_output_tokens: 128000
Copy link
Copy Markdown

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Output token limits removed from codestral-2 model

Medium Severity

The max_output_tokens and max_tokens fields were removed from the codestral-2 limits, leaving only context_window and max_input_tokens. The versioned variant codestral-2@001.yaml still retains both at 128000. This means the unversioned model no longer declares any output token limit, which could cause incorrect request validation or cost calculations for consumers of this data.

Additional Locations (1)
Fix in Cursor Fix in Web

- cache_read_input_audio_token_cost: 1.e-7
cache_read_input_token_cost: 5.e-8
input_cost_per_audio_token: 0.000001
- input_cost_per_audio_token: 0.000001
Copy link
Copy Markdown

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Cache cost data removed but caching feature retained

Medium Severity

All cache_read_input_token_cost and cache_read_input_audio_token_cost fields were removed from every region's cost entry, but prompt_caching is still listed in features. Comparable models like gemini-2.5-flash retain both the feature and the corresponding cache cost fields. This inconsistency could cause incorrect cost calculations when prompt caching is used with this model.

Additional Locations (1)
Fix in Cursor Fix in Web

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment

Labels

None yet

Projects

None yet

Development

Successfully merging this pull request may close these issues.

1 participant