From 61c139fa7de7110cb729615c26f23e913b80e6ff Mon Sep 17 00:00:00 2001 From: "LocalAI [bot]" <139863280+localai-bot@users.noreply.github.com> Date: Thu, 5 Mar 2026 09:51:47 +0100 Subject: [PATCH] feat: Rename 'Whisper' model type to 'STT' in UI (#8785) * feat: Rename 'Whisper' model type to 'STT' in UI - Updated models.html: Changed 'Whisper' filter button to 'STT' - Updated talk.html: Changed 'Whisper Model' to 'STT Model' - Updated backends.html: Changed 'Whisper' to 'STT' - Updated talk.js: Renamed getWhisperModel() to getSTTModel(), sendAudioToWhisper() to sendAudioToSTT(), and whisperModelSelect to sttModelSelect This change makes the UI more consistent with the model category naming, where all speech-to-text models (including Whisper, Parakeet, Moonshine, WhisperX, etc.) are grouped under the 'STT' (Speech-to-Text) category. Fixes #8776 Signed-off-by: team-coding-agent-1 * Rename whisperModelSelect to sttModelSelect in talk.html As requested by maintainer mudler in PR review, replacing all whisperModelSelect occurrences with sttModelSelect since the model type was renamed from Whisper to STT. Signed-off-by: LocalAI [bot] --------- Signed-off-by: team-coding-agent-1 Signed-off-by: LocalAI [bot] Co-authored-by: team-coding-agent-1 Co-authored-by: LocalAI [bot] --- core/http/static/talk.js | 12 ++++++------ core/http/views/backends.html | 2 +- core/http/views/models.html | 4 ++-- core/http/views/talk.html | 8 ++++---- 4 files changed, 13 insertions(+), 13 deletions(-) diff --git a/core/http/static/talk.js b/core/http/static/talk.js index 3fafa0a04..f8a2f838e 100644 --- a/core/http/static/talk.js +++ b/core/http/static/talk.js @@ -13,8 +13,8 @@ function getModel() { return document.getElementById('modelSelect').value; } -function getWhisperModel() { - return document.getElementById('whisperModelSelect').value; +function getSTTModel() { + return document.getElementById('sttModelSelect').value; } function getTTSModel() { @@ -73,7 +73,7 @@ function stopRecording() { document.getElementById("loader").style.display = "block"; const audioBlob = new Blob(audioChunks, { type: 'audio/webm' }); document.getElementById("statustext").textContent = "Processing audio..."; - const transcript = await sendAudioToWhisper(audioBlob); + const transcript = await sendAudioToSTT(audioBlob); console.log("Transcript:", transcript); document.getElementById("statustext").textContent = "Seems you said: " + transcript+ ". Generating response..."; const responseText = await sendTextToChatGPT(transcript); @@ -95,10 +95,10 @@ function stopRecording() { }; } -async function sendAudioToWhisper(audioBlob) { +async function sendAudioToSTT(audioBlob) { const formData = new FormData(); formData.append('file', audioBlob); - formData.append('model', getWhisperModel()); + formData.append('model', getSTTModel()); const response = await fetch('v1/audio/transcriptions', { method: 'POST', @@ -106,7 +106,7 @@ async function sendAudioToWhisper(audioBlob) { }); const result = await response.json(); - console.log("Whisper result:", result) + console.log("STT result:", result) return result.text; } diff --git a/core/http/views/backends.html b/core/http/views/backends.html index ba22dc4fb..f63c4131e 100644 --- a/core/http/views/backends.html +++ b/core/http/views/backends.html @@ -186,7 +186,7 @@ -