Issue for flash attention 2.0 support

#20
by xiongsy - opened

“_supports_flash_attn_2 = True” is required for QWenPreTrainedModel when using flash attention 2.0,
image.png
or you will get an error
image.png

Qwen org

this model use custom code and you should not use it as other transformers models. in addition, this model will not be updated. please use qwen2

jklj077 changed discussion status to closed

Sign up or log in to comment