Skip to content

Adding Qwen3-Omni inference#38

Open
dannigt wants to merge 2 commits intomainfrom
feature/Qwen3-Omni
Open

Adding Qwen3-Omni inference#38
dannigt wants to merge 2 commits intomainfrom
feature/Qwen3-Omni

Conversation

@dannigt
Copy link
Collaborator

@dannigt dannigt commented Mar 17, 2026

Adds Qwen3-Omni-30B-A3B-Instruct as a MLLM baseline under the name qwen3_omni.

Changes

  • New model file baselines/models/mllm/qwen3_omni.py
  • Registered in baselines/main.py

Example usage

python main.py \
  --model qwen3_omni \
  --lang en \
  --track long \
  --modality audio \
  --prompt fixed \
  --in_data_folder $PATH2DATA \
  --out_folder ./baseline

Notes

  • Requires flash_attention_2 and sufficient VRAM for a 30B MoE model

@@ -0,0 +1,119 @@
# Copyright 2025 FBK, KIT
Copy link
Collaborator

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Suggested change
# Copyright 2025 FBK, KIT
# Copyright 2026 FBK, KIT

torch_dtype="auto",
device_map="auto",
attn_implementation="flash_attention_2",
)
Copy link
Collaborator

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

In the model card, they suggest adding model.disable_talker() when the model is not used for speech generation (as in our case). Not sure if it has a real impact on performance tbh, but worth to try

"content": [
{
"type": "text",
"text": "You are Qwen, a virtual human developed by the Qwen Team, Alibaba Group, "
Copy link
Collaborator

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Is it the standard one, right?

)
inputs = inputs.to(model.device).to(model.dtype)

# Inference: Generation of the output text
Copy link
Collaborator

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Suggested change
# Inference: Generation of the output text

Copy link
Collaborator

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Not really necessary here, we only do inference in this repo

@sarapapi sarapapi requested a review from mgaido91 March 17, 2026 17:34
@sarapapi
Copy link
Collaborator

Have you already tried to run it, right?

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment

Labels

None yet

Projects

None yet

Development

Successfully merging this pull request may close these issues.

2 participants