Skip to content

Update STT model to large-v3-turbo#267

Open
denics wants to merge 1 commit intoOpenVoiceOS:devfrom
denics:patch-1
Open

Update STT model to large-v3-turbo#267
denics wants to merge 1 commit intoOpenVoiceOS:devfrom
denics:patch-1

Conversation

@denics
Copy link
Copy Markdown

@denics denics commented Apr 13, 2026

This solve #266

Summary by CodeRabbit

  • Chores
    • Updated FasterWhisper speech-to-text plugin model configuration.

@coderabbitai
Copy link
Copy Markdown
Contributor

coderabbitai Bot commented Apr 13, 2026

📝 Walkthrough

Walkthrough

A configuration value in the FasterWhisper STT plugin was updated, changing the model identifier from whisper-large-v3-turbo to large-v3-turbo in the GPU-recommended configuration file.

Changes

Cohort / File(s) Summary
GPU Configuration Update
ovos_config/recommends/gpu/en-us.conf
Updated FasterWhisper STT plugin model name from whisper-large-v3-turbo to large-v3-turbo; all other parameters remain unchanged.

Estimated code review effort

🎯 1 (Trivial) | ⏱️ ~2 minutes

Poem

🐰 A whisper grows shorter, the name now so bright,
From whisper-large down to just large, what a sight!
The model still mighty, just lighter to speak,
Configuration tweaks make the config less sleek,
Hops of efficiency, hopping all night! 🌙

🚥 Pre-merge checks | ✅ 3
✅ Passed checks (3 passed)
Check name Status Explanation
Description Check ✅ Passed Check skipped - CodeRabbit’s high-level summary is enabled.
Title check ✅ Passed The title 'Update STT model to large-v3-turbo' directly and clearly describes the main change: updating the FasterWhisper STT plugin model configuration from 'whisper-large-v3-turbo' to 'large-v3-turbo'.
Docstring Coverage ✅ Passed No functions found in the changed files to evaluate docstring coverage. Skipping docstring coverage check.

✏️ Tip: You can configure your own custom pre-merge checks in the settings.

✨ Finishing Touches
🧪 Generate unit tests (beta)
  • Create PR with unit tests

Thanks for using CodeRabbit! It's free for OSS, and your support helps us grow. If you like it, consider giving us a shout-out.

❤️ Share

Comment @coderabbitai help to get the list of available commands and usage tips.

Copy link
Copy Markdown
Contributor

@coderabbitai coderabbitai Bot left a comment

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Actionable comments posted: 1

🤖 Prompt for all review comments with AI agents
Verify each finding against the current code and only fix it if needed.

Inline comments:
In `@ovos_config/recommends/gpu/en-us.conf`:
- Line 6: The model identifier value for the "model" key is inconsistent:
en-us.conf uses "large-v3-turbo" while other GPU language configs use
"whisper-large-v3-turbo"; decide which canonical identifier to use (either the
bare CTranslate2 target "large-v3-turbo" or the explicit
"whisper-large-v3-turbo"), then update the divergent files so they all use the
same string; specifically locate the "model" entries in the per-language GPU
config files (e.g., en-us.conf, nl-nl.conf, it-it.conf, de-de.conf, fr-fr.conf,
da-dk.conf) and replace the non-canonical value so all files match, and run any
config validation or integration tests to ensure FasterWhisper/CTranslate2
loading still works with the chosen identifier.
🪄 Autofix (Beta)

Fix all unresolved CodeRabbit comments on this PR:

  • Push a commit to this branch (recommended)
  • Create a new PR with the fixes

ℹ️ Review info
⚙️ Run configuration

Configuration used: defaults

Review profile: CHILL

Plan: Pro

Run ID: 3d3c114b-8322-4a2e-85ff-fabd495c29d7

📥 Commits

Reviewing files that changed from the base of the PR and between 0fbc5b1 and ddfafbc.

📒 Files selected for processing (1)
  • ovos_config/recommends/gpu/en-us.conf

"fallback_module": "",
"ovos-stt-plugin-fasterwhisper": {
"model": "whisper-large-v3-turbo",
"model": "large-v3-turbo",
Copy link
Copy Markdown
Contributor

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

⚠️ Potential issue | 🟠 Major

🧩 Analysis chain

🏁 Script executed:

#!/bin/bash
# Description: Search for FasterWhisper model configurations and verify consistency

# Search for all FasterWhisper model configurations
echo "=== Searching for ovos-stt-plugin-fasterwhisper model configurations ==="
rg -n --type=json -A2 -B2 'ovos-stt-plugin-fasterwhisper'

echo -e "\n=== Searching for whisper model references ==="
rg -n 'whisper.*turbo|turbo.*whisper|large-v3-turbo' --glob='*.conf' --glob='*.json'

echo -e "\n=== Checking other GPU/STT config files ==="
fd -e conf -e json . ovos_config/recommends --exec echo "File: {}" \; --exec cat {}

Repository: OpenVoiceOS/ovos-config

Length of output: 41075


🌐 Web query:

ovos-stt-plugin-fasterwhisper model identifier format large-v3-turbo

💡 Result:

The model identifier format for ovos-stt-plugin-fasterwhisper is a simple string name like "large-v3" for standard Whisper models, or a full Hugging Face repository ID/path for custom models (e.g., "projecte-aina/faster-whisper-large-v3-ca-3catparla"). Supported standard model identifiers include: 'tiny.en', 'tiny', 'base.en', 'base', 'small.en', 'small', 'medium.en', 'medium', 'large-v1', 'large-v2', 'large-v3', 'large', 'distil-large-v2', 'distil-medium.en', 'distil-small.en', 'distil-large-v3'. "large-v3-turbo" is not listed among the standard identifiers in the official plugin documentation. However, since the plugin uses faster-whisper (which supports CTranslate2-converted Whisper models), "large-v3-turbo" can be used directly if referring to a compatible converted model. Community configurations explicitly use "whisper-large-v3-turbo" or "large-v3-turbo" as the model string, and faster-whisper loads models like "dropbox-dash/faster-whisper-large-v3-turbo" (the CTranslate2 conversion of openai/whisper-large-v3-turbo). Thus, the format allows "large-v3-turbo" as a valid identifier, likely mapping to the turbo variant.

Citations:


Inconsistency: en-us.conf differs from all other language variants in model identifier format.

The model identifier changed from whisper-large-v3-turbo to large-v3-turbo, but five other GPU language configs (nl-nl, it-it, de-de, fr-fr, da-dk) still use whisper-large-v3-turbo. While the web search confirms both formats may work with FasterWhisper's support for CTranslate2-converted models, this cross-file inconsistency suggests either an incomplete migration across all language files or a language-specific fix for issue #266. Verify that either all GPU language configs should be updated to large-v3-turbo, or confirm that en-us requires a distinct format.

🤖 Prompt for AI Agents
Verify each finding against the current code and only fix it if needed.

In `@ovos_config/recommends/gpu/en-us.conf` at line 6, The model identifier value
for the "model" key is inconsistent: en-us.conf uses "large-v3-turbo" while
other GPU language configs use "whisper-large-v3-turbo"; decide which canonical
identifier to use (either the bare CTranslate2 target "large-v3-turbo" or the
explicit "whisper-large-v3-turbo"), then update the divergent files so they all
use the same string; specifically locate the "model" entries in the per-language
GPU config files (e.g., en-us.conf, nl-nl.conf, it-it.conf, de-de.conf,
fr-fr.conf, da-dk.conf) and replace the non-canonical value so all files match,
and run any config validation or integration tests to ensure
FasterWhisper/CTranslate2 loading still works with the chosen identifier.

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment

Labels

None yet

Projects

None yet

Development

Successfully merging this pull request may close these issues.

1 participant