vllm.model_executor.models.opencua ¶
Inference-only OpenCUA-7B model compatible with HuggingFace weights.
OpenCUADummyInputsBuilder ¶
Bases: Qwen2VLDummyInputsBuilder
Source code in vllm/model_executor/models/opencua.py
get_dummy_text ¶
OpenCUAForConditionalGeneration ¶
Bases: Qwen2_5_VLForConditionalGeneration
Source code in vllm/model_executor/models/opencua.py
hf_to_vllm_mapper class-attribute instance-attribute ¶
hf_to_vllm_mapper = WeightsMapper(
orig_to_new_prefix={
"model.language_model.": "language_model.model.",
"model.visual.": "visual.",
"vision_tower.": "visual.",
"lm_head.": "language_model.lm_head.",
"model.": "language_model.model.",
}
)
is_multimodal_pruning_enabled instance-attribute ¶
language_model instance-attribute ¶
language_model = init_vllm_registered_model(
vllm_config=vllm_config,
hf_config=text_config,
prefix=maybe_prefix(prefix, "language_model"),
architectures=["Qwen2ForCausalLM"],
)
make_empty_intermediate_tensors instance-attribute ¶
multimodal_cpu_fields class-attribute instance-attribute ¶
packed_modules_mapping class-attribute instance-attribute ¶
packed_modules_mapping = {
"qkv_proj": ["q_proj", "k_proj", "v_proj"],
"gate_up_proj": ["gate_proj", "up_proj"],
}
visual instance-attribute ¶
visual = Qwen2_5_VisionTransformer(
vision_config=vision_config,
norm_eps=getattr(config, "rms_norm_eps", 1e-06),
quant_config=quant_config,
prefix=maybe_prefix(prefix, "visual"),
use_data_parallel=use_data_parallel,
attn_backend_override=attn_backend_override,
)
__init__ ¶
__init__(*, vllm_config: VllmConfig, prefix: str = '')
Source code in vllm/model_executor/models/opencua.py
get_placeholder_str classmethod ¶
OpenCUAMultiModalProcessor ¶
Bases: BaseMultiModalProcessor[OpenCUAProcessingInfo]
Source code in vllm/model_executor/models/opencua.py
_get_data_parser ¶
_get_data_parser() -> MultiModalDataParser
_get_mm_fields_config ¶
_get_mm_fields_config(
hf_inputs: BatchFeature,
hf_processor_mm_kwargs: Mapping[str, object],
) -> Mapping[str, MultiModalFieldConfig]
Source code in vllm/model_executor/models/opencua.py
_get_prompt_updates ¶
_get_prompt_updates(
mm_items: MultiModalDataItems,
hf_processor_mm_kwargs: Mapping[str, Any],
out_mm_kwargs: MultiModalKwargs,
) -> Sequence[PromptUpdate]
Source code in vllm/model_executor/models/opencua.py
OpenCUAProcessingInfo ¶
Bases: Qwen2VLProcessingInfo
Source code in vllm/model_executor/models/opencua.py
OpenCUAProcessor ¶
Bases: Qwen2VLProcessor
Source code in vllm/model_executor/models/opencua.py
__call__ ¶
Source code in vllm/model_executor/models/opencua.py
__init__ ¶
__init__(
vision_config: dict, tokenizer: AnyTokenizer, **kwargs
)