ariG23498 HF Staff commited on
Commit
1b6fa87
·
verified ·
1 Parent(s): 2c50684

Upload meta-llama_Llama-3.1-8B-Instruct_1.txt with huggingface_hub

Browse files
meta-llama_Llama-3.1-8B-Instruct_1.txt CHANGED
@@ -1 +1,106 @@
1
- Everything was good in meta-llama_Llama-3.1-8B-Instruct_1.txt
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ```CODE:
2
+ # Use a pipeline as a high-level helper
3
+ from transformers import pipeline
4
+
5
+ pipe = pipeline("text-generation", model="meta-llama/Llama-3.1-8B-Instruct")
6
+ messages = [
7
+ {"role": "user", "content": "Who are you?"},
8
+ ]
9
+ pipe(messages)
10
+ ```
11
+
12
+ ERROR:
13
+ Traceback (most recent call last):
14
+ File "/tmp/.cache/uv/environments-v2/8c467f4d75ae1bd9/lib/python3.13/site-packages/huggingface_hub/utils/_http.py", line 720, in hf_raise_for_status
15
+ response.raise_for_status()
16
+ ~~~~~~~~~~~~~~~~~~~~~~~~~^^
17
+ File "/tmp/.cache/uv/environments-v2/8c467f4d75ae1bd9/lib/python3.13/site-packages/httpx/_models.py", line 829, in raise_for_status
18
+ raise HTTPStatusError(message, request=request, response=self)
19
+ httpx.HTTPStatusError: Client error '403 Forbidden' for url 'https://huggingface.co/meta-llama/Llama-3.1-8B-Instruct/resolve/main/config.json'
20
+ For more information check: https://developer.mozilla.org/en-US/docs/Web/HTTP/Status/403
21
+
22
+ The above exception was the direct cause of the following exception:
23
+
24
+ Traceback (most recent call last):
25
+ File "/tmp/.cache/uv/environments-v2/8c467f4d75ae1bd9/lib/python3.13/site-packages/huggingface_hub/file_download.py", line 1649, in _get_metadata_or_catch_error
26
+ metadata = get_hf_file_metadata(
27
+ url=url,
28
+ ...<4 lines>...
29
+ retry_on_errors=retry_on_errors,
30
+ )
31
+ File "/tmp/.cache/uv/environments-v2/8c467f4d75ae1bd9/lib/python3.13/site-packages/huggingface_hub/utils/_validators.py", line 89, in _inner_fn
32
+ return fn(*args, **kwargs)
33
+ File "/tmp/.cache/uv/environments-v2/8c467f4d75ae1bd9/lib/python3.13/site-packages/huggingface_hub/file_download.py", line 1572, in get_hf_file_metadata
34
+ response = _httpx_follow_relative_redirects_with_backoff(
35
+ method="HEAD", url=url, headers=hf_headers, timeout=timeout, retry_on_errors=retry_on_errors
36
+ )
37
+ File "/tmp/.cache/uv/environments-v2/8c467f4d75ae1bd9/lib/python3.13/site-packages/huggingface_hub/utils/_http.py", line 651, in _httpx_follow_relative_redirects_with_backoff
38
+ hf_raise_for_status(response)
39
+ ~~~~~~~~~~~~~~~~~~~^^^^^^^^^^
40
+ File "/tmp/.cache/uv/environments-v2/8c467f4d75ae1bd9/lib/python3.13/site-packages/huggingface_hub/utils/_http.py", line 802, in hf_raise_for_status
41
+ raise _format(HfHubHTTPError, message, response) from e
42
+ huggingface_hub.errors.HfHubHTTPError: (Request ID: Root=1-69a93029-61586cef0bd9991262c36762;807e3d68-cba8-49cd-bb00-b752b86c6328)
43
+
44
+ 403 Forbidden: Please enable access to public gated repositories in your fine-grained token settings to view this repository..
45
+ Cannot access content at: https://huggingface.co/meta-llama/Llama-3.1-8B-Instruct/resolve/main/config.json.
46
+ Make sure your token has the correct permissions.
47
+
48
+ The above exception was the direct cause of the following exception:
49
+
50
+ Traceback (most recent call last):
51
+ File "/tmp/.cache/uv/environments-v2/8c467f4d75ae1bd9/lib/python3.13/site-packages/transformers/utils/hub.py", line 420, in cached_files
52
+ hf_hub_download(
53
+ ~~~~~~~~~~~~~~~^
54
+ path_or_repo_id,
55
+ ^^^^^^^^^^^^^^^^
56
+ ...<9 lines>...
57
+ local_files_only=local_files_only,
58
+ ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
59
+ )
60
+ ^
61
+ File "/tmp/.cache/uv/environments-v2/8c467f4d75ae1bd9/lib/python3.13/site-packages/huggingface_hub/utils/_validators.py", line 89, in _inner_fn
62
+ return fn(*args, **kwargs)
63
+ File "/tmp/.cache/uv/environments-v2/8c467f4d75ae1bd9/lib/python3.13/site-packages/huggingface_hub/file_download.py", line 982, in hf_hub_download
64
+ return _hf_hub_download_to_cache_dir(
65
+ # Destination
66
+ ...<15 lines>...
67
+ dry_run=dry_run,
68
+ )
69
+ File "/tmp/.cache/uv/environments-v2/8c467f4d75ae1bd9/lib/python3.13/site-packages/huggingface_hub/file_download.py", line 1133, in _hf_hub_download_to_cache_dir
70
+ _raise_on_head_call_error(head_call_error, force_download, local_files_only)
71
+ ~~~~~~~~~~~~~~~~~~~~~~~~~^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
72
+ File "/tmp/.cache/uv/environments-v2/8c467f4d75ae1bd9/lib/python3.13/site-packages/huggingface_hub/file_download.py", line 1765, in _raise_on_head_call_error
73
+ raise LocalEntryNotFoundError(
74
+ ...<3 lines>...
75
+ ) from head_call_error
76
+ huggingface_hub.errors.LocalEntryNotFoundError: An error happened while trying to locate the file on the Hub and we cannot find the requested files in the local cache. Please check your connection and try again or make sure your Internet connection is on.
77
+
78
+ The above exception was the direct cause of the following exception:
79
+
80
+ Traceback (most recent call last):
81
+ File "/tmp/meta-llama_Llama-3.1-8B-Instruct_1BGdThD.py", line 26, in <module>
82
+ pipe = pipeline("text-generation", model="meta-llama/Llama-3.1-8B-Instruct")
83
+ File "/tmp/.cache/uv/environments-v2/8c467f4d75ae1bd9/lib/python3.13/site-packages/transformers/pipelines/__init__.py", line 699, in pipeline
84
+ config = AutoConfig.from_pretrained(
85
+ model, _from_pipeline=task, code_revision=code_revision, **hub_kwargs, **model_kwargs
86
+ )
87
+ File "/tmp/.cache/uv/environments-v2/8c467f4d75ae1bd9/lib/python3.13/site-packages/transformers/models/auto/configuration_auto.py", line 1429, in from_pretrained
88
+ config_dict, unused_kwargs = PreTrainedConfig.get_config_dict(pretrained_model_name_or_path, **kwargs)
89
+ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
90
+ File "/tmp/.cache/uv/environments-v2/8c467f4d75ae1bd9/lib/python3.13/site-packages/transformers/configuration_utils.py", line 574, in get_config_dict
91
+ config_dict, kwargs = cls._get_config_dict(pretrained_model_name_or_path, **kwargs)
92
+ ~~~~~~~~~~~~~~~~~~~~^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
93
+ File "/tmp/.cache/uv/environments-v2/8c467f4d75ae1bd9/lib/python3.13/site-packages/transformers/configuration_utils.py", line 629, in _get_config_dict
94
+ resolved_config_file = cached_file(
95
+ pretrained_model_name_or_path,
96
+ ...<9 lines>...
97
+ _commit_hash=commit_hash,
98
+ )
99
+ File "/tmp/.cache/uv/environments-v2/8c467f4d75ae1bd9/lib/python3.13/site-packages/transformers/utils/hub.py", line 277, in cached_file
100
+ file = cached_files(path_or_repo_id=path_or_repo_id, filenames=[filename], **kwargs)
101
+ File "/tmp/.cache/uv/environments-v2/8c467f4d75ae1bd9/lib/python3.13/site-packages/transformers/utils/hub.py", line 494, in cached_files
102
+ raise OSError(
103
+ ...<3 lines>...
104
+ ) from e
105
+ OSError: We couldn't connect to 'https://huggingface.co' to load the files, and couldn't find them in the cached files.
106
+ Check your internet connection or see how to run the library in offline mode at 'https://huggingface.co/docs/transformers/installation#offline-mode'.