ariG23498 HF Staff commited on
Commit
e5657a1
·
verified ·
1 Parent(s): 73b015d

Upload black-forest-labs_FLUX.2-dev_1.txt with huggingface_hub

Browse files
Files changed (1) hide show
  1. black-forest-labs_FLUX.2-dev_1.txt +65 -45
black-forest-labs_FLUX.2-dev_1.txt CHANGED
@@ -14,54 +14,74 @@ image = pipe(image=input_image, prompt=prompt).images[0]
14
 
15
  ERROR:
16
  Traceback (most recent call last):
17
- File "/tmp/black-forest-labs_FLUX.2-dev_135Owr3.py", line 28, in <module>
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
18
  pipe = DiffusionPipeline.from_pretrained("black-forest-labs/FLUX.2-dev", dtype=torch.bfloat16, device_map="cuda")
19
  File "/tmp/.cache/uv/environments-v2/d49ad6c613615895/lib/python3.13/site-packages/huggingface_hub/utils/_validators.py", line 89, in _inner_fn
20
  return fn(*args, **kwargs)
21
- File "/tmp/.cache/uv/environments-v2/d49ad6c613615895/lib/python3.13/site-packages/diffusers/pipelines/pipeline_utils.py", line 1021, in from_pretrained
22
- loaded_sub_model = load_sub_model(
23
- library_name=library_name,
24
- ...<21 lines>...
25
- quantization_config=quantization_config,
26
  )
27
- File "/tmp/.cache/uv/environments-v2/d49ad6c613615895/lib/python3.13/site-packages/diffusers/pipelines/pipeline_loading_utils.py", line 876, in load_sub_model
28
- loaded_sub_model = load_method(os.path.join(cached_folder, name), **loading_kwargs)
29
- File "/tmp/.cache/uv/environments-v2/d49ad6c613615895/lib/python3.13/site-packages/transformers/modeling_utils.py", line 4109, in from_pretrained
30
- load_info = cls._load_pretrained_model(model, state_dict, checkpoint_files, load_config)
31
- File "/tmp/.cache/uv/environments-v2/d49ad6c613615895/lib/python3.13/site-packages/transformers/modeling_utils.py", line 4231, in _load_pretrained_model
32
- convert_and_load_state_dict_in_model(
33
- ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^
34
- model=model,
35
- ^^^^^^^^^^^^
36
- ...<4 lines>...
37
- disk_offload_index=disk_offload_index,
38
- ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
39
  )
40
- ^
41
- File "/tmp/.cache/uv/environments-v2/d49ad6c613615895/lib/python3.13/site-packages/transformers/core_model_loading.py", line 1217, in convert_and_load_state_dict_in_model
42
- realized_value, conversion_errors = mapping.convert(
43
- ~~~~~~~~~~~~~~~^
44
- first_param_name,
45
- ^^^^^^^^^^^^^^^^^
46
- ...<4 lines>...
47
- conversion_errors=conversion_errors,
48
- ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
49
  )
50
- ^
51
- File "/tmp/.cache/uv/environments-v2/d49ad6c613615895/lib/python3.13/site-packages/transformers/core_model_loading.py", line 696, in convert
52
- collected_tensors = self.materialize_tensors()
53
- File "/tmp/.cache/uv/environments-v2/d49ad6c613615895/lib/python3.13/site-packages/transformers/core_model_loading.py", line 671, in materialize_tensors
54
- tensors = [future.result() for future in tensors]
55
- ~~~~~~~~~~~~~^^
56
- File "/usr/lib/python3.13/concurrent/futures/_base.py", line 456, in result
57
- return self.__get_result()
58
- ~~~~~~~~~~~~~~~~~^^
59
- File "/usr/lib/python3.13/concurrent/futures/_base.py", line 401, in __get_result
60
- raise self._exception
61
- File "/usr/lib/python3.13/concurrent/futures/thread.py", line 59, in run
62
- result = self.fn(*self.args, **self.kwargs)
63
- File "/tmp/.cache/uv/environments-v2/d49ad6c613615895/lib/python3.13/site-packages/transformers/core_model_loading.py", line 818, in _job
64
- return _materialize_copy(tensor, device, dtype)
65
- File "/tmp/.cache/uv/environments-v2/d49ad6c613615895/lib/python3.13/site-packages/transformers/core_model_loading.py", line 807, in _materialize_copy
66
- tensor = tensor.to(device=device, dtype=dtype)
67
- torch.OutOfMemoryError: CUDA out of memory. Tried to allocate 320.00 MiB. GPU 0 has a total capacity of 22.03 GiB of which 287.12 MiB is free. Including non-PyTorch memory, this process has 21.75 GiB memory in use. Of the allocated memory 21.45 GiB is allocated by PyTorch, and 120.15 MiB is reserved by PyTorch but unallocated. If reserved but unallocated memory is large try setting PYTORCH_ALLOC_CONF=expandable_segments:True to avoid fragmentation. See documentation for Memory Management (https://pytorch.org/docs/stable/notes/cuda.html#environment-variables)
 
14
 
15
  ERROR:
16
  Traceback (most recent call last):
17
+ File "/tmp/.cache/uv/environments-v2/d49ad6c613615895/lib/python3.13/site-packages/huggingface_hub/utils/_http.py", line 720, in hf_raise_for_status
18
+ response.raise_for_status()
19
+ ~~~~~~~~~~~~~~~~~~~~~~~~~^^
20
+ File "/tmp/.cache/uv/environments-v2/d49ad6c613615895/lib/python3.13/site-packages/httpx/_models.py", line 829, in raise_for_status
21
+ raise HTTPStatusError(message, request=request, response=self)
22
+ httpx.HTTPStatusError: Client error '403 Forbidden' for url 'https://huggingface.co/black-forest-labs/FLUX.2-dev/resolve/main/model_index.json'
23
+ For more information check: https://developer.mozilla.org/en-US/docs/Web/HTTP/Status/403
24
+
25
+ The above exception was the direct cause of the following exception:
26
+
27
+ Traceback (most recent call last):
28
+ File "/tmp/.cache/uv/environments-v2/d49ad6c613615895/lib/python3.13/site-packages/huggingface_hub/file_download.py", line 1649, in _get_metadata_or_catch_error
29
+ metadata = get_hf_file_metadata(
30
+ url=url,
31
+ ...<4 lines>...
32
+ retry_on_errors=retry_on_errors,
33
+ )
34
+ File "/tmp/.cache/uv/environments-v2/d49ad6c613615895/lib/python3.13/site-packages/huggingface_hub/utils/_validators.py", line 89, in _inner_fn
35
+ return fn(*args, **kwargs)
36
+ File "/tmp/.cache/uv/environments-v2/d49ad6c613615895/lib/python3.13/site-packages/huggingface_hub/file_download.py", line 1572, in get_hf_file_metadata
37
+ response = _httpx_follow_relative_redirects_with_backoff(
38
+ method="HEAD", url=url, headers=hf_headers, timeout=timeout, retry_on_errors=retry_on_errors
39
+ )
40
+ File "/tmp/.cache/uv/environments-v2/d49ad6c613615895/lib/python3.13/site-packages/huggingface_hub/utils/_http.py", line 651, in _httpx_follow_relative_redirects_with_backoff
41
+ hf_raise_for_status(response)
42
+ ~~~~~~~~~~~~~~~~~~~^^^^^^^^^^
43
+ File "/tmp/.cache/uv/environments-v2/d49ad6c613615895/lib/python3.13/site-packages/huggingface_hub/utils/_http.py", line 802, in hf_raise_for_status
44
+ raise _format(HfHubHTTPError, message, response) from e
45
+ huggingface_hub.errors.HfHubHTTPError: (Request ID: Root=1-69a93020-388c1b4f1049842007fcbb1d;e24ec4fe-35e4-4124-baad-7f28da2f5bf7)
46
+
47
+ 403 Forbidden: Please enable access to public gated repositories in your fine-grained token settings to view this repository..
48
+ Cannot access content at: https://huggingface.co/black-forest-labs/FLUX.2-dev/resolve/main/model_index.json.
49
+ Make sure your token has the correct permissions.
50
+
51
+ The above exception was the direct cause of the following exception:
52
+
53
+ Traceback (most recent call last):
54
+ File "/tmp/black-forest-labs_FLUX.2-dev_1mRGmho.py", line 28, in <module>
55
  pipe = DiffusionPipeline.from_pretrained("black-forest-labs/FLUX.2-dev", dtype=torch.bfloat16, device_map="cuda")
56
  File "/tmp/.cache/uv/environments-v2/d49ad6c613615895/lib/python3.13/site-packages/huggingface_hub/utils/_validators.py", line 89, in _inner_fn
57
  return fn(*args, **kwargs)
58
+ File "/tmp/.cache/uv/environments-v2/d49ad6c613615895/lib/python3.13/site-packages/diffusers/pipelines/pipeline_utils.py", line 829, in from_pretrained
59
+ cached_folder = cls.download(
60
+ pretrained_model_name_or_path,
61
+ ...<14 lines>...
62
+ **kwargs,
63
  )
64
+ File "/tmp/.cache/uv/environments-v2/d49ad6c613615895/lib/python3.13/site-packages/huggingface_hub/utils/_validators.py", line 89, in _inner_fn
65
+ return fn(*args, **kwargs)
66
+ File "/tmp/.cache/uv/environments-v2/d49ad6c613615895/lib/python3.13/site-packages/diffusers/pipelines/pipeline_utils.py", line 1605, in download
67
+ config_file = hf_hub_download(
68
+ pretrained_model_name,
69
+ ...<5 lines>...
70
+ token=token,
 
 
 
 
 
71
  )
72
+ File "/tmp/.cache/uv/environments-v2/d49ad6c613615895/lib/python3.13/site-packages/huggingface_hub/utils/_validators.py", line 89, in _inner_fn
73
+ return fn(*args, **kwargs)
74
+ File "/tmp/.cache/uv/environments-v2/d49ad6c613615895/lib/python3.13/site-packages/huggingface_hub/file_download.py", line 982, in hf_hub_download
75
+ return _hf_hub_download_to_cache_dir(
76
+ # Destination
77
+ ...<15 lines>...
78
+ dry_run=dry_run,
 
 
79
  )
80
+ File "/tmp/.cache/uv/environments-v2/d49ad6c613615895/lib/python3.13/site-packages/huggingface_hub/file_download.py", line 1133, in _hf_hub_download_to_cache_dir
81
+ _raise_on_head_call_error(head_call_error, force_download, local_files_only)
82
+ ~~~~~~~~~~~~~~~~~~~~~~~~~^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
83
+ File "/tmp/.cache/uv/environments-v2/d49ad6c613615895/lib/python3.13/site-packages/huggingface_hub/file_download.py", line 1765, in _raise_on_head_call_error
84
+ raise LocalEntryNotFoundError(
85
+ ...<3 lines>...
86
+ ) from head_call_error
87
+ huggingface_hub.errors.LocalEntryNotFoundError: An error happened while trying to locate the file on the Hub and we cannot find the requested files in the local cache. Please check your connection and try again or make sure your Internet connection is on.