ariG23498 HF Staff commited on
Commit
45925f4
·
verified ·
1 Parent(s): 361368e

Upload meta-llama_Llama-3.1-8B-Instruct_2.txt with huggingface_hub

Browse files
meta-llama_Llama-3.1-8B-Instruct_2.txt CHANGED
@@ -1 +1,209 @@
1
- Everything was good in meta-llama_Llama-3.1-8B-Instruct_2.txt
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ```CODE:
2
+ # Load model directly
3
+ from transformers import AutoTokenizer, AutoModelForCausalLM
4
+
5
+ tokenizer = AutoTokenizer.from_pretrained("meta-llama/Llama-3.1-8B-Instruct")
6
+ model = AutoModelForCausalLM.from_pretrained("meta-llama/Llama-3.1-8B-Instruct")
7
+ messages = [
8
+ {"role": "user", "content": "Who are you?"},
9
+ ]
10
+ inputs = tokenizer.apply_chat_template(
11
+ messages,
12
+ add_generation_prompt=True,
13
+ tokenize=True,
14
+ return_dict=True,
15
+ return_tensors="pt",
16
+ ).to(model.device)
17
+
18
+ outputs = model.generate(**inputs, max_new_tokens=40)
19
+ print(tokenizer.decode(outputs[0][inputs["input_ids"].shape[-1]:]))
20
+ ```
21
+
22
+ ERROR:
23
+ Traceback (most recent call last):
24
+ File "/tmp/.cache/uv/environments-v2/0f9915e46e2946d2/lib/python3.13/site-packages/huggingface_hub/utils/_http.py", line 720, in hf_raise_for_status
25
+ response.raise_for_status()
26
+ ~~~~~~~~~~~~~~~~~~~~~~~~~^^
27
+ File "/tmp/.cache/uv/environments-v2/0f9915e46e2946d2/lib/python3.13/site-packages/httpx/_models.py", line 829, in raise_for_status
28
+ raise HTTPStatusError(message, request=request, response=self)
29
+ httpx.HTTPStatusError: Client error '403 Forbidden' for url 'https://huggingface.co/meta-llama/Llama-3.1-8B-Instruct/resolve/main/config.json'
30
+ For more information check: https://developer.mozilla.org/en-US/docs/Web/HTTP/Status/403
31
+
32
+ The above exception was the direct cause of the following exception:
33
+
34
+ Traceback (most recent call last):
35
+ File "/tmp/.cache/uv/environments-v2/0f9915e46e2946d2/lib/python3.13/site-packages/huggingface_hub/file_download.py", line 1649, in _get_metadata_or_catch_error
36
+ metadata = get_hf_file_metadata(
37
+ url=url,
38
+ ...<4 lines>...
39
+ retry_on_errors=retry_on_errors,
40
+ )
41
+ File "/tmp/.cache/uv/environments-v2/0f9915e46e2946d2/lib/python3.13/site-packages/huggingface_hub/utils/_validators.py", line 89, in _inner_fn
42
+ return fn(*args, **kwargs)
43
+ File "/tmp/.cache/uv/environments-v2/0f9915e46e2946d2/lib/python3.13/site-packages/huggingface_hub/file_download.py", line 1572, in get_hf_file_metadata
44
+ response = _httpx_follow_relative_redirects_with_backoff(
45
+ method="HEAD", url=url, headers=hf_headers, timeout=timeout, retry_on_errors=retry_on_errors
46
+ )
47
+ File "/tmp/.cache/uv/environments-v2/0f9915e46e2946d2/lib/python3.13/site-packages/huggingface_hub/utils/_http.py", line 651, in _httpx_follow_relative_redirects_with_backoff
48
+ hf_raise_for_status(response)
49
+ ~~~~~~~~~~~~~~~~~~~^^^^^^^^^^
50
+ File "/tmp/.cache/uv/environments-v2/0f9915e46e2946d2/lib/python3.13/site-packages/huggingface_hub/utils/_http.py", line 802, in hf_raise_for_status
51
+ raise _format(HfHubHTTPError, message, response) from e
52
+ huggingface_hub.errors.HfHubHTTPError: (Request ID: Root=1-69a93026-312b2fa736f01f512b02d019;1a0d0376-d25c-4ba3-873e-7dab282f374c)
53
+
54
+ 403 Forbidden: Please enable access to public gated repositories in your fine-grained token settings to view this repository..
55
+ Cannot access content at: https://huggingface.co/meta-llama/Llama-3.1-8B-Instruct/resolve/main/config.json.
56
+ Make sure your token has the correct permissions.
57
+
58
+ The above exception was the direct cause of the following exception:
59
+
60
+ Traceback (most recent call last):
61
+ File "/tmp/.cache/uv/environments-v2/0f9915e46e2946d2/lib/python3.13/site-packages/transformers/utils/hub.py", line 420, in cached_files
62
+ hf_hub_download(
63
+ ~~~~~~~~~~~~~~~^
64
+ path_or_repo_id,
65
+ ^^^^^^^^^^^^^^^^
66
+ ...<9 lines>...
67
+ local_files_only=local_files_only,
68
+ ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
69
+ )
70
+ ^
71
+ File "/tmp/.cache/uv/environments-v2/0f9915e46e2946d2/lib/python3.13/site-packages/huggingface_hub/utils/_validators.py", line 89, in _inner_fn
72
+ return fn(*args, **kwargs)
73
+ File "/tmp/.cache/uv/environments-v2/0f9915e46e2946d2/lib/python3.13/site-packages/huggingface_hub/file_download.py", line 982, in hf_hub_download
74
+ return _hf_hub_download_to_cache_dir(
75
+ # Destination
76
+ ...<15 lines>...
77
+ dry_run=dry_run,
78
+ )
79
+ File "/tmp/.cache/uv/environments-v2/0f9915e46e2946d2/lib/python3.13/site-packages/huggingface_hub/file_download.py", line 1133, in _hf_hub_download_to_cache_dir
80
+ _raise_on_head_call_error(head_call_error, force_download, local_files_only)
81
+ ~~~~~~~~~~~~~~~~~~~~~~~~~^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
82
+ File "/tmp/.cache/uv/environments-v2/0f9915e46e2946d2/lib/python3.13/site-packages/huggingface_hub/file_download.py", line 1765, in _raise_on_head_call_error
83
+ raise LocalEntryNotFoundError(
84
+ ...<3 lines>...
85
+ ) from head_call_error
86
+ huggingface_hub.errors.LocalEntryNotFoundError: An error happened while trying to locate the file on the Hub and we cannot find the requested files in the local cache. Please check your connection and try again or make sure your Internet connection is on.
87
+
88
+ The above exception was the direct cause of the following exception:
89
+
90
+ Traceback (most recent call last):
91
+ File "/tmp/.cache/uv/environments-v2/0f9915e46e2946d2/lib/python3.13/site-packages/transformers/models/auto/tokenization_auto.py", line 654, in from_pretrained
92
+ config = AutoConfig.from_pretrained(
93
+ pretrained_model_name_or_path, trust_remote_code=trust_remote_code, **kwargs
94
+ )
95
+ File "/tmp/.cache/uv/environments-v2/0f9915e46e2946d2/lib/python3.13/site-packages/transformers/models/auto/configuration_auto.py", line 1429, in from_pretrained
96
+ config_dict, unused_kwargs = PreTrainedConfig.get_config_dict(pretrained_model_name_or_path, **kwargs)
97
+ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
98
+ File "/tmp/.cache/uv/environments-v2/0f9915e46e2946d2/lib/python3.13/site-packages/transformers/configuration_utils.py", line 574, in get_config_dict
99
+ config_dict, kwargs = cls._get_config_dict(pretrained_model_name_or_path, **kwargs)
100
+ ~~~~~~~~~~~~~~~~~~~~^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
101
+ File "/tmp/.cache/uv/environments-v2/0f9915e46e2946d2/lib/python3.13/site-packages/transformers/configuration_utils.py", line 629, in _get_config_dict
102
+ resolved_config_file = cached_file(
103
+ pretrained_model_name_or_path,
104
+ ...<9 lines>...
105
+ _commit_hash=commit_hash,
106
+ )
107
+ File "/tmp/.cache/uv/environments-v2/0f9915e46e2946d2/lib/python3.13/site-packages/transformers/utils/hub.py", line 277, in cached_file
108
+ file = cached_files(path_or_repo_id=path_or_repo_id, filenames=[filename], **kwargs)
109
+ File "/tmp/.cache/uv/environments-v2/0f9915e46e2946d2/lib/python3.13/site-packages/transformers/utils/hub.py", line 494, in cached_files
110
+ raise OSError(
111
+ ...<3 lines>...
112
+ ) from e
113
+ OSError: We couldn't connect to 'https://huggingface.co' to load the files, and couldn't find them in the cached files.
114
+ Check your internet connection or see how to run the library in offline mode at 'https://huggingface.co/docs/transformers/installation#offline-mode'.
115
+
116
+ During handling of the above exception, another exception occurred:
117
+
118
+ Traceback (most recent call last):
119
+ File "/tmp/.cache/uv/environments-v2/0f9915e46e2946d2/lib/python3.13/site-packages/huggingface_hub/utils/_http.py", line 720, in hf_raise_for_status
120
+ response.raise_for_status()
121
+ ~~~~~~~~~~~~~~~~~~~~~~~~~^^
122
+ File "/tmp/.cache/uv/environments-v2/0f9915e46e2946d2/lib/python3.13/site-packages/httpx/_models.py", line 829, in raise_for_status
123
+ raise HTTPStatusError(message, request=request, response=self)
124
+ httpx.HTTPStatusError: Client error '403 Forbidden' for url 'https://huggingface.co/meta-llama/Llama-3.1-8B-Instruct/resolve/main/config.json'
125
+ For more information check: https://developer.mozilla.org/en-US/docs/Web/HTTP/Status/403
126
+
127
+ The above exception was the direct cause of the following exception:
128
+
129
+ Traceback (most recent call last):
130
+ File "/tmp/.cache/uv/environments-v2/0f9915e46e2946d2/lib/python3.13/site-packages/huggingface_hub/file_download.py", line 1649, in _get_metadata_or_catch_error
131
+ metadata = get_hf_file_metadata(
132
+ url=url,
133
+ ...<4 lines>...
134
+ retry_on_errors=retry_on_errors,
135
+ )
136
+ File "/tmp/.cache/uv/environments-v2/0f9915e46e2946d2/lib/python3.13/site-packages/huggingface_hub/utils/_validators.py", line 89, in _inner_fn
137
+ return fn(*args, **kwargs)
138
+ File "/tmp/.cache/uv/environments-v2/0f9915e46e2946d2/lib/python3.13/site-packages/huggingface_hub/file_download.py", line 1572, in get_hf_file_metadata
139
+ response = _httpx_follow_relative_redirects_with_backoff(
140
+ method="HEAD", url=url, headers=hf_headers, timeout=timeout, retry_on_errors=retry_on_errors
141
+ )
142
+ File "/tmp/.cache/uv/environments-v2/0f9915e46e2946d2/lib/python3.13/site-packages/huggingface_hub/utils/_http.py", line 651, in _httpx_follow_relative_redirects_with_backoff
143
+ hf_raise_for_status(response)
144
+ ~~~~~~~~~~~~~~~~~~~^^^^^^^^^^
145
+ File "/tmp/.cache/uv/environments-v2/0f9915e46e2946d2/lib/python3.13/site-packages/huggingface_hub/utils/_http.py", line 802, in hf_raise_for_status
146
+ raise _format(HfHubHTTPError, message, response) from e
147
+ huggingface_hub.errors.HfHubHTTPError: (Request ID: Root=1-69a93026-5e9d2e0945896e42298f6889;22ecdf28-8550-4389-870a-da547418fb1f)
148
+
149
+ 403 Forbidden: Please enable access to public gated repositories in your fine-grained token settings to view this repository..
150
+ Cannot access content at: https://huggingface.co/meta-llama/Llama-3.1-8B-Instruct/resolve/main/config.json.
151
+ Make sure your token has the correct permissions.
152
+
153
+ The above exception was the direct cause of the following exception:
154
+
155
+ Traceback (most recent call last):
156
+ File "/tmp/.cache/uv/environments-v2/0f9915e46e2946d2/lib/python3.13/site-packages/transformers/utils/hub.py", line 420, in cached_files
157
+ hf_hub_download(
158
+ ~~~~~~~~~~~~~~~^
159
+ path_or_repo_id,
160
+ ^^^^^^^^^^^^^^^^
161
+ ...<9 lines>...
162
+ local_files_only=local_files_only,
163
+ ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
164
+ )
165
+ ^
166
+ File "/tmp/.cache/uv/environments-v2/0f9915e46e2946d2/lib/python3.13/site-packages/huggingface_hub/utils/_validators.py", line 89, in _inner_fn
167
+ return fn(*args, **kwargs)
168
+ File "/tmp/.cache/uv/environments-v2/0f9915e46e2946d2/lib/python3.13/site-packages/huggingface_hub/file_download.py", line 982, in hf_hub_download
169
+ return _hf_hub_download_to_cache_dir(
170
+ # Destination
171
+ ...<15 lines>...
172
+ dry_run=dry_run,
173
+ )
174
+ File "/tmp/.cache/uv/environments-v2/0f9915e46e2946d2/lib/python3.13/site-packages/huggingface_hub/file_download.py", line 1133, in _hf_hub_download_to_cache_dir
175
+ _raise_on_head_call_error(head_call_error, force_download, local_files_only)
176
+ ~~~~~~~~~~~~~~~~~~~~~~~~~^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
177
+ File "/tmp/.cache/uv/environments-v2/0f9915e46e2946d2/lib/python3.13/site-packages/huggingface_hub/file_download.py", line 1765, in _raise_on_head_call_error
178
+ raise LocalEntryNotFoundError(
179
+ ...<3 lines>...
180
+ ) from head_call_error
181
+ huggingface_hub.errors.LocalEntryNotFoundError: An error happened while trying to locate the file on the Hub and we cannot find the requested files in the local cache. Please check your connection and try again or make sure your Internet connection is on.
182
+
183
+ The above exception was the direct cause of the following exception:
184
+
185
+ Traceback (most recent call last):
186
+ File "/tmp/meta-llama_Llama-3.1-8B-Instruct_2lcvj9V.py", line 26, in <module>
187
+ tokenizer = AutoTokenizer.from_pretrained("meta-llama/Llama-3.1-8B-Instruct")
188
+ File "/tmp/.cache/uv/environments-v2/0f9915e46e2946d2/lib/python3.13/site-packages/transformers/models/auto/tokenization_auto.py", line 658, in from_pretrained
189
+ config = PreTrainedConfig.from_pretrained(pretrained_model_name_or_path, **kwargs)
190
+ File "/tmp/.cache/uv/environments-v2/0f9915e46e2946d2/lib/python3.13/site-packages/transformers/configuration_utils.py", line 533, in from_pretrained
191
+ config_dict, kwargs = cls.get_config_dict(pretrained_model_name_or_path, **kwargs)
192
+ ~~~~~~~~~~~~~~~~~~~^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
193
+ File "/tmp/.cache/uv/environments-v2/0f9915e46e2946d2/lib/python3.13/site-packages/transformers/configuration_utils.py", line 574, in get_config_dict
194
+ config_dict, kwargs = cls._get_config_dict(pretrained_model_name_or_path, **kwargs)
195
+ ~~~~~~~~~~~~~~~~~~~~^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
196
+ File "/tmp/.cache/uv/environments-v2/0f9915e46e2946d2/lib/python3.13/site-packages/transformers/configuration_utils.py", line 629, in _get_config_dict
197
+ resolved_config_file = cached_file(
198
+ pretrained_model_name_or_path,
199
+ ...<9 lines>...
200
+ _commit_hash=commit_hash,
201
+ )
202
+ File "/tmp/.cache/uv/environments-v2/0f9915e46e2946d2/lib/python3.13/site-packages/transformers/utils/hub.py", line 277, in cached_file
203
+ file = cached_files(path_or_repo_id=path_or_repo_id, filenames=[filename], **kwargs)
204
+ File "/tmp/.cache/uv/environments-v2/0f9915e46e2946d2/lib/python3.13/site-packages/transformers/utils/hub.py", line 494, in cached_files
205
+ raise OSError(
206
+ ...<3 lines>...
207
+ ) from e
208
+ OSError: We couldn't connect to 'https://huggingface.co' to load the files, and couldn't find them in the cached files.
209
+ Check your internet connection or see how to run the library in offline mode at 'https://huggingface.co/docs/transformers/installation#offline-mode'.