From 8108871fdd00095d3ec1def640cb686ee417b484 Mon Sep 17 00:00:00 2001 From: Mitch Lewis Date: Tue, 13 Jan 2026 17:43:27 -0700 Subject: [PATCH] Revert "[Bugfix] Support compile for Transformers multimodal (#23095)" This reverts commit 0e3bb543f064eb416bca4f6f3013efa3830b12f7. --- vllm/model_executor/models/transformers.py | 7 ------- 1 file changed, 7 deletions(-) diff --git a/vllm/model_executor/models/transformers.py b/vllm/model_executor/models/transformers.py index ed9d6c0ab4ce..712667b1e274 100644 --- a/vllm/model_executor/models/transformers.py +++ b/vllm/model_executor/models/transformers.py @@ -709,13 +709,6 @@ def _can_concat(x: list[torch.Tensor]): MultiModalProcessor, info=MultiModalProcessingInfo, dummy_inputs=MultiModalDummyInputsBuilder) -@support_torch_compile( - dynamic_arg_dims={ - "input_ids": 0, - "positions": -1, - "intermediate_tensors": 0, - "inputs_embeds": 0, - }) # set `positions` to last dim to support Qwen-mrope class TransformersForMultimodalLM(TransformersForCausalLM, SupportsMultiModal): # Backwards compatibility for prev released models. State dicts back then # had different formats and cannot be loaded with `AutoModel` mapping as is