@@ -924,8 +924,8 @@ def get_num_layers_by_block_type(
924924 layers_block_type_value = getattr (self .hf_config ,
925925 "layers_block_type" , None )
926926 if layers_block_type_value is None :
927- raise ValueError ("The model is an hybrid without a"
928- "layers_block_type in the hf_config,"
927+ raise ValueError ("The model is an hybrid without a "
928+ "layers_block_type in the hf_config, "
929929 "cannot determine the num of "
930930 f"{ block_type .value } layers" )
931931
@@ -2518,7 +2518,7 @@ def _get_and_verify_dtype(
25182518
25192519 if current_platform .is_hpu () and config_dtype == torch .float16 :
25202520 logger .info (
2521- "For HPU, we cast models to bfloat16 instead of"
2521+ "For HPU, we cast models to bfloat16 instead of "
25222522 "using float16 by default. Please specify `dtype` if you "
25232523 "want to use float16." )
25242524 torch_dtype = torch .bfloat16
@@ -2734,7 +2734,7 @@ def __post_init__(self):
27342734 backend = self .guided_decoding_backend ).backend_name
27352735 if backend not in valid_guided_backends :
27362736 raise ValueError (f"Invalid guided_decoding_backend '{ backend } ,"
2737- f"must be one of { valid_guided_backends } " )
2737+ f" must be one of { valid_guided_backends } " )
27382738
27392739
27402740@dataclass
@@ -3010,7 +3010,7 @@ def uuid(self):
30103010 def model_post_init (self , __context : Any ) -> None :
30113011 if not self .enable_reshape and self .enable_fusion :
30123012 logger .warning_once (
3013- "Fusion enabled but reshape elimination disabled."
3013+ "Fusion enabled but reshape elimination disabled. "
30143014 "RMSNorm + quant (fp8) fusion might not work" )
30153015
30163016 pass_config : PassConfig = Field (default_factory = PassConfig )
@@ -3565,7 +3565,7 @@ def set_current_vllm_config(vllm_config: VllmConfig, check_compile=False):
35653565 logger .warning (
35663566 "`torch.compile` is turned on, but the model %s"
35673567 " does not support it. Please open an issue on GitHub"
3568- "if you want it to be supported." ,
3568+ " if you want it to be supported." ,
35693569 vllm_config .model_config .model )
35703570 _current_vllm_config = old_vllm_config
35713571
0 commit comments