You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
Traceback (most recent call last):
2024-08-01T21:29:17.880522621Z File "/src/handler.py", line 6, in <module>
2024-08-01T21:29:17.880527641Z vllm_engine = vLLMEngine()
2024-08-01T21:29:17.880533331Z File "/src/engine.py", line 25, in __init__
2024-08-01T21:29:17.880543011Z self.llm = self._initialize_llm() if engine is None else engine
2024-08-01T21:29:17.880549201Z File "/src/engine.py", line 111, in _initialize_llm
2024-08-01T21:29:17.880554071Z raise e
2024-08-01T21:29:17.880559761Z File "/src/engine.py", line 105, in _initialize_llm
2024-08-01T21:29:17.880582981Z engine = AsyncLLMEngine.from_engine_args(AsyncEngineArgs(**self.config))
2024-08-01T21:29:17.880598581Z File "/usr/local/lib/python3.10/dist-packages/vllm/engine/async_llm_engine.py", line 346, in from_engine_args
2024-08-01T21:29:17.880669700Z engine_config = engine_args.create_engine_config()
2024-08-01T21:29:17.880679090Z File "/usr/local/lib/python3.10/dist-packages/vllm/engine/arg_utils.py", line 520, in create_engine_config
2024-08-01T21:29:17.880798769Z model_config = ModelConfig(
2024-08-01T21:29:17.880806039Z File "/usr/local/lib/python3.10/dist-packages/vllm/config.py", line 119, in __init__
2024-08-01T21:29:17.880814579Z self.hf_config = get_config(self.model, trust_remote_code, revision,
2024-08-01T21:29:17.880909728Z File "/usr/local/lib/python3.10/dist-packages/vllm/transformers_utils/config.py", line 38, in get_config
2024-08-01T21:29:17.880929687Z raise e
2024-08-01T21:29:17.880935207Z File "/usr/local/lib/python3.10/dist-packages/vllm/transformers_utils/config.py", line 23, in get_config
2024-08-01T21:29:17.881092116Z config = AutoConfig.from_pretrained(
2024-08-01T21:29:17.881125706Z File "/usr/local/lib/python3.10/dist-packages/transformers/models/auto/configuration_auto.py", line 958, in from_pretrained
2024-08-01T21:29:17.881143606Z return config_class.from_dict(config_dict, **unused_kwargs)
2024-08-01T21:29:17.881151795Z File "/usr/local/lib/python3.10/dist-packages/transformers/configuration_utils.py", line 768, in from_dict
2024-08-01T21:29:17.881304244Z config = cls(**config_dict)
2024-08-01T21:29:17.881319394Z File "/usr/local/lib/python3.10/dist-packages/transformers/models/llama/configuration_llama.py", line 161, in __init__
2024-08-01T21:29:17.881325854Z self._rope_scaling_validation()
2024-08-01T21:29:17.881334464Z File "/usr/local/lib/python3.10/dist-packages/transformers/models/llama/configuration_llama.py", line 182, in _rope_scaling_validation
2024-08-01T21:29:17.881384983Z raise ValueError(
2024-08-01T21:29:17.881394323Z ValueError: rope_scaling must be a dictionary with two fields, type and factor, got {'factor': 8.0, 'high_freq_factor': 4.0, 'low_freq_factor': 1.0, 'original_max_position_embeddings': 8192, 'rope_type': 'llama3'}
As far as I know, this issue is fixed simply by upgrading Transformers (https://huggingface.co/meta-llama/Meta-Llama-3.1-8B-Instruct/discussions/15)
The text was updated successfully, but these errors were encountered: