runtime error

Exit code: 1. Reason: s.py", line 1417, in check_torch_load_is_safe raise ValueError( ValueError: Due to a serious vulnerability issue in `torch.load`, even with `weights_only=True`, we now require users to upgrade torch to at least v2.6 in order to use the function. This version restriction does not apply when loading files with safetensors. See the vulnerability report here https://nvd.nist.gov/vuln/detail/CVE-2025-32434 while loading with MarianMTModel, an error is thrown: Traceback (most recent call last): File "/usr/local/lib/python3.10/site-packages/transformers/pipelines/base.py", line 292, in infer_framework_load_model model = model_class.from_pretrained(model, **kwargs) File "/usr/local/lib/python3.10/site-packages/transformers/modeling_utils.py", line 309, in _wrapper return func(*args, **kwargs) File "/usr/local/lib/python3.10/site-packages/transformers/modeling_utils.py", line 4573, in from_pretrained ) = cls._load_pretrained_model( File "/usr/local/lib/python3.10/site-packages/transformers/modeling_utils.py", line 4832, in _load_pretrained_model load_state_dict(checkpoint_files[0], map_location="meta", weights_only=weights_only).keys() File "/usr/local/lib/python3.10/site-packages/transformers/modeling_utils.py", line 553, in load_state_dict check_torch_load_is_safe() File "/usr/local/lib/python3.10/site-packages/transformers/utils/import_utils.py", line 1417, in check_torch_load_is_safe raise ValueError( ValueError: Due to a serious vulnerability issue in `torch.load`, even with `weights_only=True`, we now require users to upgrade torch to at least v2.6 in order to use the function. This version restriction does not apply when loading files with safetensors. See the vulnerability report here https://nvd.nist.gov/vuln/detail/CVE-2025-32434 model.safetensors: 0%| | 0.00/312M [00:00<?, ?B/s] model.safetensors: 63%|β–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–Ž | 197M/312M [00:01<00:00, 187MB/s] model.safetensors: 100%|β–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–‰| 312M/312M [00:01<00:00, 267MB/s]

Container logs:

Fetching error logs...