runtime error
Exit code: 1. Reason: n: Traceback (most recent call last): File "/home/user/app/app.py", line 19, in <module> pipe = pipeline("text-generation", "microsoft/Phi-3-mini-4k-instruct", torch_dtype=torch.bfloat16, device_map="auto") File "/usr/local/lib/python3.10/site-packages/transformers/pipelines/__init__.py", line 895, in pipeline framework, model = infer_framework_load_model( File "/usr/local/lib/python3.10/site-packages/transformers/pipelines/base.py", line 283, in infer_framework_load_model model = model_class.from_pretrained(model, **kwargs) File "/usr/local/lib/python3.10/site-packages/transformers/models/auto/auto_factory.py", line 564, in from_pretrained return model_class.from_pretrained( File "/usr/local/lib/python3.10/site-packages/transformers/modeling_utils.py", line 3681, in from_pretrained resolved_archive_file, sharded_metadata = get_checkpoint_shard_files( File "/usr/local/lib/python3.10/site-packages/transformers/utils/hub.py", line 1079, in get_checkpoint_shard_files cached_filename = cached_file( File "/usr/local/lib/python3.10/site-packages/transformers/utils/hub.py", line 402, in cached_file resolved_file = hf_hub_download( File "/usr/local/lib/python3.10/site-packages/huggingface_hub/utils/_validators.py", line 114, in _inner_fn return fn(*args, **kwargs) File "/usr/local/lib/python3.10/site-packages/huggingface_hub/file_download.py", line 1221, in hf_hub_download return _hf_hub_download_to_cache_dir( File "/usr/local/lib/python3.10/site-packages/huggingface_hub/file_download.py", line 1367, in _hf_hub_download_to_cache_dir _download_to_tmp_and_move( File "/usr/local/lib/python3.10/site-packages/huggingface_hub/file_download.py", line 1884, in _download_to_tmp_and_move http_get( File "/usr/local/lib/python3.10/site-packages/huggingface_hub/file_download.py", line 524, in http_get raise RuntimeError( RuntimeError: An error occurred while downloading using `hf_transfer`. Consider disabling HF_HUB_ENABLE_HF_TRANSFER for better error handling.
Container logs:
Fetching error logs...