|
41 | 41 | from vllm.model_executor.layers.rotary_embedding import get_rope
|
42 | 42 | from vllm.model_executor.model_loader.utils import initialize_model
|
43 | 43 | from vllm.model_executor.model_loader.weight_utils import default_weight_loader
|
| 44 | +from vllm.model_executor.models.module_mapping import MultiModelKeys |
44 | 45 | from vllm.model_executor.sampling_metadata import SamplingMetadata
|
45 | 46 | from vllm.multimodal import MULTIMODAL_REGISTRY
|
46 | 47 | from vllm.multimodal.inputs import (MultiModalDataDict, MultiModalFieldConfig,
|
|
54 | 55 | from vllm.multimodal.utils import run_dp_sharded_vision_model
|
55 | 56 | from vllm.sequence import IntermediateTensors
|
56 | 57 |
|
57 |
| -from .interfaces import MultiModalEmbeddings, SupportsMultiModal, SupportsPP, SupportsLoRA |
| 58 | +from .interfaces import (MultiModalEmbeddings, SupportsLoRA, |
| 59 | + SupportsMultiModal, SupportsPP) |
58 | 60 | from .llama4 import Llama4ForCausalLM
|
59 | 61 | from .utils import (AutoWeightsLoader, flatten_bn, maybe_prefix,
|
60 | 62 | merge_multimodal_embeddings)
|
61 |
| -from vllm.model_executor.models.module_mapping import MultiModelKeys |
| 63 | + |
62 | 64 |
|
63 | 65 | class Llama4ImagePatchInputs(TypedDict):
|
64 | 66 | type: Literal["pixel_values"]
|
@@ -711,9 +713,8 @@ def get_dummy_mm_data(
|
711 | 713 | info=Mllama4ProcessingInfo,
|
712 | 714 | dummy_inputs=Mllama4DummyInputsBuilder,
|
713 | 715 | )
|
714 |
| -class Llama4ForConditionalGeneration( |
715 |
| - nn.Module, SupportsMultiModal, SupportsPP, SupportsLoRA |
716 |
| -): |
| 716 | +class Llama4ForConditionalGeneration(nn.Module, SupportsMultiModal, SupportsPP, |
| 717 | + SupportsLoRA): |
717 | 718 | packed_modules_mapping = {
|
718 | 719 | "qkv_proj": ["q_proj", "k_proj", "v_proj"],
|
719 | 720 | }
|
|
0 commit comments