ariG23498 HF Staff commited on
Commit
933af3a
·
verified ·
1 Parent(s): 1f97f34

Upload black-forest-labs_FLUX.1-dev_1.txt with huggingface_hub

Browse files
Files changed (1) hide show
  1. black-forest-labs_FLUX.1-dev_1.txt +64 -23
black-forest-labs_FLUX.1-dev_1.txt CHANGED
@@ -11,33 +11,74 @@ image = pipe(prompt).images[0]
11
 
12
  ERROR:
13
  Traceback (most recent call last):
14
- File "/tmp/black-forest-labs_FLUX.1-dev_1uzjER9.py", line 27, in <module>
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
15
  pipe = DiffusionPipeline.from_pretrained("black-forest-labs/FLUX.1-dev", dtype=torch.bfloat16, device_map="cuda")
16
  File "/tmp/.cache/uv/environments-v2/b90b3a1935bc74f7/lib/python3.13/site-packages/huggingface_hub/utils/_validators.py", line 89, in _inner_fn
17
  return fn(*args, **kwargs)
18
- File "/tmp/.cache/uv/environments-v2/b90b3a1935bc74f7/lib/python3.13/site-packages/diffusers/pipelines/pipeline_utils.py", line 1021, in from_pretrained
19
- loaded_sub_model = load_sub_model(
20
- library_name=library_name,
21
- ...<21 lines>...
22
- quantization_config=quantization_config,
 
 
 
 
 
 
 
 
23
  )
24
- File "/tmp/.cache/uv/environments-v2/b90b3a1935bc74f7/lib/python3.13/site-packages/diffusers/pipelines/pipeline_loading_utils.py", line 876, in load_sub_model
25
- loaded_sub_model = load_method(os.path.join(cached_folder, name), **loading_kwargs)
26
  File "/tmp/.cache/uv/environments-v2/b90b3a1935bc74f7/lib/python3.13/site-packages/huggingface_hub/utils/_validators.py", line 89, in _inner_fn
27
  return fn(*args, **kwargs)
28
- File "/tmp/.cache/uv/environments-v2/b90b3a1935bc74f7/lib/python3.13/site-packages/diffusers/models/modeling_utils.py", line 1296, in from_pretrained
29
- ) = cls._load_pretrained_model(
30
- ~~~~~~~~~~~~~~~~~~~~~~~~~~^
31
- model,
32
- ^^^^^^
33
- ...<13 lines>...
34
- is_parallel_loading_enabled=is_parallel_loading_enabled,
35
- ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
36
  )
37
- ^
38
- File "/tmp/.cache/uv/environments-v2/b90b3a1935bc74f7/lib/python3.13/site-packages/diffusers/models/modeling_utils.py", line 1635, in _load_pretrained_model
39
- _caching_allocator_warmup(model, expanded_device_map, dtype, hf_quantizer)
40
- ~~~~~~~~~~~~~~~~~~~~~~~~~^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
41
- File "/tmp/.cache/uv/environments-v2/b90b3a1935bc74f7/lib/python3.13/site-packages/diffusers/models/model_loading_utils.py", line 751, in _caching_allocator_warmup
42
- _ = torch.empty(warmup_elems, dtype=dtype, device=device, requires_grad=False)
43
- torch.OutOfMemoryError: CUDA out of memory. Tried to allocate 22.17 GiB. GPU 0 has a total capacity of 22.03 GiB of which 10.54 GiB is free. Including non-PyTorch memory, this process has 11.49 GiB memory in use. Of the allocated memory 11.29 GiB is allocated by PyTorch, and 13.27 MiB is reserved by PyTorch but unallocated. If reserved but unallocated memory is large try setting PYTORCH_ALLOC_CONF=expandable_segments:True to avoid fragmentation. See documentation for Memory Management (https://pytorch.org/docs/stable/notes/cuda.html#environment-variables)
 
 
11
 
12
  ERROR:
13
  Traceback (most recent call last):
14
+ File "/tmp/.cache/uv/environments-v2/b90b3a1935bc74f7/lib/python3.13/site-packages/huggingface_hub/utils/_http.py", line 720, in hf_raise_for_status
15
+ response.raise_for_status()
16
+ ~~~~~~~~~~~~~~~~~~~~~~~~~^^
17
+ File "/tmp/.cache/uv/environments-v2/b90b3a1935bc74f7/lib/python3.13/site-packages/httpx/_models.py", line 829, in raise_for_status
18
+ raise HTTPStatusError(message, request=request, response=self)
19
+ httpx.HTTPStatusError: Client error '403 Forbidden' for url 'https://huggingface.co/black-forest-labs/FLUX.1-dev/resolve/main/model_index.json'
20
+ For more information check: https://developer.mozilla.org/en-US/docs/Web/HTTP/Status/403
21
+
22
+ The above exception was the direct cause of the following exception:
23
+
24
+ Traceback (most recent call last):
25
+ File "/tmp/.cache/uv/environments-v2/b90b3a1935bc74f7/lib/python3.13/site-packages/huggingface_hub/file_download.py", line 1649, in _get_metadata_or_catch_error
26
+ metadata = get_hf_file_metadata(
27
+ url=url,
28
+ ...<4 lines>...
29
+ retry_on_errors=retry_on_errors,
30
+ )
31
+ File "/tmp/.cache/uv/environments-v2/b90b3a1935bc74f7/lib/python3.13/site-packages/huggingface_hub/utils/_validators.py", line 89, in _inner_fn
32
+ return fn(*args, **kwargs)
33
+ File "/tmp/.cache/uv/environments-v2/b90b3a1935bc74f7/lib/python3.13/site-packages/huggingface_hub/file_download.py", line 1572, in get_hf_file_metadata
34
+ response = _httpx_follow_relative_redirects_with_backoff(
35
+ method="HEAD", url=url, headers=hf_headers, timeout=timeout, retry_on_errors=retry_on_errors
36
+ )
37
+ File "/tmp/.cache/uv/environments-v2/b90b3a1935bc74f7/lib/python3.13/site-packages/huggingface_hub/utils/_http.py", line 651, in _httpx_follow_relative_redirects_with_backoff
38
+ hf_raise_for_status(response)
39
+ ~~~~~~~~~~~~~~~~~~~^^^^^^^^^^
40
+ File "/tmp/.cache/uv/environments-v2/b90b3a1935bc74f7/lib/python3.13/site-packages/huggingface_hub/utils/_http.py", line 802, in hf_raise_for_status
41
+ raise _format(HfHubHTTPError, message, response) from e
42
+ huggingface_hub.errors.HfHubHTTPError: (Request ID: Root=1-69a93021-18fc60e2507df191227a5262;f0c0c735-e9b9-4d65-9e74-d04544e4ed50)
43
+
44
+ 403 Forbidden: Please enable access to public gated repositories in your fine-grained token settings to view this repository..
45
+ Cannot access content at: https://huggingface.co/black-forest-labs/FLUX.1-dev/resolve/main/model_index.json.
46
+ Make sure your token has the correct permissions.
47
+
48
+ The above exception was the direct cause of the following exception:
49
+
50
+ Traceback (most recent call last):
51
+ File "/tmp/black-forest-labs_FLUX.1-dev_19Wh1Cv.py", line 27, in <module>
52
  pipe = DiffusionPipeline.from_pretrained("black-forest-labs/FLUX.1-dev", dtype=torch.bfloat16, device_map="cuda")
53
  File "/tmp/.cache/uv/environments-v2/b90b3a1935bc74f7/lib/python3.13/site-packages/huggingface_hub/utils/_validators.py", line 89, in _inner_fn
54
  return fn(*args, **kwargs)
55
+ File "/tmp/.cache/uv/environments-v2/b90b3a1935bc74f7/lib/python3.13/site-packages/diffusers/pipelines/pipeline_utils.py", line 829, in from_pretrained
56
+ cached_folder = cls.download(
57
+ pretrained_model_name_or_path,
58
+ ...<14 lines>...
59
+ **kwargs,
60
+ )
61
+ File "/tmp/.cache/uv/environments-v2/b90b3a1935bc74f7/lib/python3.13/site-packages/huggingface_hub/utils/_validators.py", line 89, in _inner_fn
62
+ return fn(*args, **kwargs)
63
+ File "/tmp/.cache/uv/environments-v2/b90b3a1935bc74f7/lib/python3.13/site-packages/diffusers/pipelines/pipeline_utils.py", line 1605, in download
64
+ config_file = hf_hub_download(
65
+ pretrained_model_name,
66
+ ...<5 lines>...
67
+ token=token,
68
  )
 
 
69
  File "/tmp/.cache/uv/environments-v2/b90b3a1935bc74f7/lib/python3.13/site-packages/huggingface_hub/utils/_validators.py", line 89, in _inner_fn
70
  return fn(*args, **kwargs)
71
+ File "/tmp/.cache/uv/environments-v2/b90b3a1935bc74f7/lib/python3.13/site-packages/huggingface_hub/file_download.py", line 982, in hf_hub_download
72
+ return _hf_hub_download_to_cache_dir(
73
+ # Destination
74
+ ...<15 lines>...
75
+ dry_run=dry_run,
 
 
 
76
  )
77
+ File "/tmp/.cache/uv/environments-v2/b90b3a1935bc74f7/lib/python3.13/site-packages/huggingface_hub/file_download.py", line 1133, in _hf_hub_download_to_cache_dir
78
+ _raise_on_head_call_error(head_call_error, force_download, local_files_only)
79
+ ~~~~~~~~~~~~~~~~~~~~~~~~~^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
80
+ File "/tmp/.cache/uv/environments-v2/b90b3a1935bc74f7/lib/python3.13/site-packages/huggingface_hub/file_download.py", line 1765, in _raise_on_head_call_error
81
+ raise LocalEntryNotFoundError(
82
+ ...<3 lines>...
83
+ ) from head_call_error
84
+ huggingface_hub.errors.LocalEntryNotFoundError: An error happened while trying to locate the file on the Hub and we cannot find the requested files in the local cache. Please check your connection and try again or make sure your Internet connection is on.