|
152 | 152 | Qwen2Model, |
153 | 153 | Qwen2RMSNorm, |
154 | 154 | ) |
| 155 | +from transformers.models.qwen2_5_vl.modeling_qwen2_5_vl import ( |
| 156 | + Qwen2_5_VisionTransformerPretrainedModel, |
| 157 | + Qwen2_5_VLAttention, |
| 158 | + Qwen2_5_VLDecoderLayer, |
| 159 | + Qwen2_5_VLForConditionalGeneration, |
| 160 | + Qwen2_5_VLModel, |
| 161 | + Qwen2_5_VLTextModel, |
| 162 | + Qwen2_5_VLVisionAttention, |
| 163 | +) |
| 164 | +from transformers.models.qwen2_5_vl.modeling_qwen2_5_vl import ( |
| 165 | + Qwen2RMSNorm as Qwen2_5RMSNorm, |
| 166 | +) |
155 | 167 | from transformers.models.qwen3.modeling_qwen3 import ( |
156 | 168 | Qwen3Attention, |
157 | 169 | Qwen3DecoderLayer, |
|
356 | 368 | QEffQwen2ForCausalLM, |
357 | 369 | QEffQwen2Model, |
358 | 370 | ) |
| 371 | +from QEfficient.transformers.models.qwen2_5_vl.modeling_qwen2_5_vl import ( |
| 372 | + QEffQwen2_5_VisionTransformerPretrainedModel, |
| 373 | + QEffQwen2_5_VLAttention, |
| 374 | + QEffQwen2_5_VLDecoderLayer, |
| 375 | + QEffQwen2_5_VLModel, |
| 376 | + QEffQwen2_5_VLTextModel, |
| 377 | + QEffQwen2_5_VLVisionAttention, |
| 378 | + QEffQwen_2_5_vl_ForConditionalGeneration, |
| 379 | +) |
359 | 380 | from QEfficient.transformers.models.qwen3.modeling_qwen3 import ( |
360 | 381 | QEffQwen3Attention, |
361 | 382 | QEffQwen3DecoderLayer, |
@@ -404,6 +425,7 @@ class CustomOpsTransform(ModuleMappingTransform): |
404 | 425 | Phi3RMSNorm: CustomRMSNormAIC, |
405 | 426 | Qwen2RMSNorm: CustomRMSNormAIC, |
406 | 427 | Qwen3RMSNorm: CustomRMSNormAIC, |
| 428 | + Qwen2_5RMSNorm: CustomRMSNormAIC, |
407 | 429 | MllamaTextRMSNorm: CustomRMSNormAIC, |
408 | 430 | GraniteRMSNorm: CustomRMSNormAIC, |
409 | 431 | PixtralRMSNorm: CustomRMSNormAIC, |
@@ -544,6 +566,14 @@ class KVCacheTransform(ModuleMappingTransform): |
544 | 566 | Qwen3DecoderLayer: QEffQwen3DecoderLayer, |
545 | 567 | Qwen3Model: QEffQwen3Model, |
546 | 568 | Qwen3ForCausalLM: QEffQwen3ForCausalLM, |
| 569 | + # Qwen2.5 VL |
| 570 | + Qwen2_5_VLForConditionalGeneration: QEffQwen_2_5_vl_ForConditionalGeneration, |
| 571 | + Qwen2_5_VLModel: QEffQwen2_5_VLModel, |
| 572 | + Qwen2_5_VLAttention: QEffQwen2_5_VLAttention, |
| 573 | + Qwen2_5_VLDecoderLayer: QEffQwen2_5_VLDecoderLayer, |
| 574 | + Qwen2_5_VisionTransformerPretrainedModel: QEffQwen2_5_VisionTransformerPretrainedModel, |
| 575 | + Qwen2_5_VLVisionAttention: QEffQwen2_5_VLVisionAttention, |
| 576 | + Qwen2_5_VLTextModel: QEffQwen2_5_VLTextModel, |
547 | 577 | # Starcoder2 |
548 | 578 | Starcoder2Attention: QEffStarcoder2Attention, |
549 | 579 | Starcoder2DecoderLayer: QEFFStarcoder2DecoderLayer, |
|
0 commit comments