transformers @ git+https://github.com/huggingface/transformers@3a1ead0aabed473eafe527915eea8c197d424356 qwen-omni-utils[decord] soundfile torch gradio torchvision torchaudio accelerate # flash-attn https://github.com/Dao-AILab/flash-attention/releases/download/v2.7.4.post1/flash_attn-2.7.4.post1+cu12torch2.2cxx11abiTRUE-cp312-cp312-linux_x86_64.whl