|
7 | 7 | import ray |
8 | 8 | from prometheus_client import REGISTRY |
9 | 9 |
|
| 10 | +import vllm.envs as envs |
10 | 11 | from vllm import EngineArgs, LLMEngine |
11 | 12 | from vllm.distributed import cleanup_dist_env_and_memory |
12 | 13 | from vllm.engine.arg_utils import AsyncEngineArgs |
@@ -141,8 +142,10 @@ def test_metric_set_tag_model_name(vllm_runner, model: str, dtype: str, |
141 | 142 | stat_logger = vllm_model.model.llm_engine.stat_loggers['prometheus'] |
142 | 143 | metrics_tag_content = stat_logger.labels["model_name"] |
143 | 144 |
|
| 145 | + if envs.VLLM_CI_USE_S3: |
| 146 | + model = f"{MODEL_WEIGHTS_S3_BUCKET}/{model}" |
144 | 147 | if served_model_name is None or served_model_name == []: |
145 | | - assert metrics_tag_content == f"{MODEL_WEIGHTS_S3_BUCKET}/{model}", ( |
| 148 | + assert metrics_tag_content == model, ( |
146 | 149 | f"Metrics tag model_name is wrong! expect: {model!r}\n" |
147 | 150 | f"actual: {metrics_tag_content!r}") |
148 | 151 | else: |
@@ -215,8 +218,9 @@ def test_engine_log_metrics_regression( |
215 | 218 | while engine.has_unfinished_requests(): |
216 | 219 | engine.step() |
217 | 220 |
|
218 | | - assert_metrics(f"{MODEL_WEIGHTS_S3_BUCKET}/{model}", engine, |
219 | | - disable_log_stats, len(example_prompts)) |
| 221 | + if envs.VLLM_CI_USE_S3: |
| 222 | + model = f"{MODEL_WEIGHTS_S3_BUCKET}/{model}" |
| 223 | + assert_metrics(model, engine, disable_log_stats, len(example_prompts)) |
220 | 224 |
|
221 | 225 |
|
222 | 226 | @pytest.mark.parametrize("model", MODELS) |
|
0 commit comments