diff --git a/setup.py b/setup.py index a80cb81bc8..f42cf7d192 100644 --- a/setup.py +++ b/setup.py @@ -54,7 +54,7 @@ def get_console_scripts() -> List[str]: "gptq": ["optimum>=1.17.0", "auto-gptq>=0.5.0"], "awq": ["autoawq"], "aqlm": ["aqlm[gpu]>=1.1.0"], - "vllm": ["vllm>=0.4.3,<=0.6.0"], + "vllm": ["vllm>=0.4.3,<=0.6.2"], "galore": ["galore-torch"], "badam": ["badam>=1.2.1"], "adam-mini": ["adam-mini"], diff --git a/src/llamafactory/hparams/parser.py b/src/llamafactory/hparams/parser.py index 38ac87fe90..68a5d2f62f 100644 --- a/src/llamafactory/hparams/parser.py +++ b/src/llamafactory/hparams/parser.py @@ -123,7 +123,7 @@ def _check_extra_dependencies( require_version("mixture-of-depth>=1.1.6", "To fix: pip install mixture-of-depth>=1.1.6") if model_args.infer_backend == "vllm": - require_version("vllm>=0.4.3,<=0.6.0", "To fix: pip install vllm>=0.4.3,<=0.6.0") + require_version("vllm>=0.4.3,<=0.6.2", "To fix: pip install vllm>=0.4.3,<=0.6.2") if finetuning_args.use_galore: require_version("galore_torch", "To fix: pip install galore_torch")