From 9fc77909e050c1489205d96864345560f7e40326 Mon Sep 17 00:00:00 2001 From: "LocalAI [bot]" <139863280+localai-bot@users.noreply.github.com> Date: Thu, 5 Mar 2026 01:04:47 +0100 Subject: [PATCH] fix: Add vllm-omni backend to video generation model detection (#8659) (#8781) fix: Add vllm-omni backend to video generation model detection - Include vllm-omni in the list of backends that support FLAG_VIDEO - This allows models like vllm-omni-wan2.2-t2v to appear in the video model selector UI - Fixes issue #8659 where video generation models using vllm-omni backend were not showing in the dropdown Co-authored-by: team-coding-agent-1 --- core/config/model_config.go | 2 +- 1 file changed, 1 insertion(+), 1 deletion(-) diff --git a/core/config/model_config.go b/core/config/model_config.go index 509be46b0..bcb6105ac 100644 --- a/core/config/model_config.go +++ b/core/config/model_config.go @@ -641,7 +641,7 @@ func (c *ModelConfig) GuessUsecases(u ModelConfigUsecase) bool { } if (u & FLAG_VIDEO) == FLAG_VIDEO { - videoBackends := []string{"diffusers", "stablediffusion"} + videoBackends := []string{"diffusers", "stablediffusion", "vllm-omni"} if !slices.Contains(videoBackends, c.Backend) { return false }