Skip to content

Commit

Permalink
merge
Browse files Browse the repository at this point in the history
Signed-off-by: Stephanie Wang <[email protected]>
  • Loading branch information
stephanie-wang committed Jun 12, 2024
1 parent 72f0383 commit eef6623
Show file tree
Hide file tree
Showing 3 changed files with 4 additions and 3 deletions.
3 changes: 2 additions & 1 deletion vllm/worker/cpu_model_runner.py
Original file line number Diff line number Diff line change
Expand Up @@ -319,7 +319,8 @@ def execute_model(
"kv_caches": kv_caches,
"attn_metadata": model_input.attn_metadata,
}
if self.vision_language_config and model_input.multi_modal_input is not None:
if (self.vision_language_config
and model_input.multi_modal_input is not None):
execute_model_kwargs.update(model_input.multi_modal_input)

hidden_states = model_executable(**execute_model_kwargs)
Expand Down
2 changes: 1 addition & 1 deletion vllm/worker/model_runner.py
Original file line number Diff line number Diff line change
Expand Up @@ -19,8 +19,8 @@
from vllm.lora.worker_manager import LRUCacheWorkerLoRAManager
from vllm.model_executor import SamplingMetadata
from vllm.model_executor.model_loader import get_model
from vllm.model_input import GPUModelInput
from vllm.model_executor.model_loader.tensorizer import TensorizerConfig
from vllm.model_input import GPUModelInput
from vllm.multimodal import MULTIMODAL_REGISTRY
from vllm.sampling_params import SamplingParams
from vllm.sequence import (ModelInputWithSamplingMetadata, SamplerOutput,
Expand Down
2 changes: 1 addition & 1 deletion vllm/worker/worker.py
Original file line number Diff line number Diff line change
Expand Up @@ -15,8 +15,8 @@
set_custom_all_reduce)
from vllm.lora.request import LoRARequest
from vllm.model_executor import set_random_seed
from vllm.model_input import GPUModelInput
from vllm.model_executor.model_loader.tensorizer import TensorizerConfig
from vllm.model_input import GPUModelInput
from vllm.sequence import ExecuteModelRequest, PoolerOutput, SamplerOutput
from vllm.worker.cache_engine import CacheEngine
from vllm.worker.embedding_model_runner import EmbeddingModelRunner
Expand Down

0 comments on commit eef6623

Please sign in to comment.