Buckets:
| import{s as Xe,o as Ae,n as Ne}from"../chunks/scheduler.8c3d61f6.js";import{S as Ee,i as Oe,g as r,s as n,r as h,A as He,h as a,f as i,c as s,j as B,u as _,x as m,k as F,y as t,a as c,v as b,d as v,t as x,w as D}from"../chunks/index.da70eac4.js";import{T as Re}from"../chunks/Tip.1d9b8c37.js";import{D as de}from"../chunks/Docstring.ee4b6913.js";import{H as Pe,E as Ue}from"../chunks/EditOnGithub.1e64e623.js";function Ve(Y){let f,S="This is an experimental pipeline and is likely to change in the future.";return{c(){f=r("p"),f.textContent=S},l(g){f=a(g,"P",{"data-svelte-h":!0}),m(f)!=="svelte-5fv3pw"&&(f.textContent=S)},m(g,A){c(g,f,A)},p:Ne,d(g){g&&i(f)}}}function qe(Y){let f,S,g,A,k,J,$,Le='<a href="https://github.com/crowsonkb/k-diffusion" rel="nofollow">k-diffusion</a> is a popular library created by <a href="https://github.com/crowsonkb/" rel="nofollow">Katherine Crowson</a>. We provide <code>StableDiffusionKDiffusionPipeline</code> and <code>StableDiffusionXLKDiffusionPipeline</code> that allow you to run Stable DIffusion with samplers from k-diffusion.',Q,C,we='Note that most the samplers from k-diffusion are implemented in Diffusers and we recommend using existing schedulers. You can find a mapping between k-diffusion samplers and schedulers in Diffusers <a href="https://huggingface.co/docs/diffusers/api/schedulers/overview" rel="nofollow">here</a>',Z,T,ee,d,K,fe,N,Se="Pipeline for text-to-image generation using Stable Diffusion.",le,E,ke=`This model inherits from <a href="/docs/diffusers/main/en/api/pipelines/overview#diffusers.DiffusionPipeline">DiffusionPipeline</a>. Check the superclass documentation for the generic methods the | |
| library implements for all the pipelines (such as downloading or saving, running on a particular device, etc.)`,pe,O,$e="The pipeline also inherits the following loading methods:",ce,H,Ce='<li><a href="/docs/diffusers/main/en/api/loaders/textual_inversion#diffusers.loaders.TextualInversionLoaderMixin.load_textual_inversion">load_textual_inversion()</a> for loading textual inversion embeddings</li> <li><a href="/docs/diffusers/main/en/api/loaders/lora#diffusers.loaders.StableDiffusionLoraLoaderMixin.load_lora_weights">load_lora_weights()</a> for loading LoRA weights</li> <li><a href="/docs/diffusers/main/en/api/loaders/lora#diffusers.loaders.StableDiffusionLoraLoaderMixin.save_lora_weights">save_lora_weights()</a> for saving LoRA weights</li>',me,P,ue,L,M,ge,R,Te="Encodes the prompt into text encoder hidden states.",oe,I,te,l,y,he,U,Ke="Pipeline for text-to-image generation using Stable Diffusion XL and k-diffusion.",_e,V,Me=`This model inherits from <a href="/docs/diffusers/main/en/api/pipelines/overview#diffusers.DiffusionPipeline">DiffusionPipeline</a>. Check the superclass documentation for the generic methods the | |
| library implements for all the pipelines (such as downloading or saving, running on a particular device, etc.)`,be,q,Ie="The pipeline also inherits the following loading methods:",ve,j,ye='<li><a href="/docs/diffusers/main/en/api/loaders/textual_inversion#diffusers.loaders.TextualInversionLoaderMixin.load_textual_inversion">load_textual_inversion()</a> for loading textual inversion embeddings</li> <li><a href="/docs/diffusers/main/en/api/loaders/single_file#diffusers.loaders.FromSingleFileMixin.from_single_file">from_single_file()</a> for loading <code>.ckpt</code> files</li> <li><a href="/docs/diffusers/main/en/api/loaders/lora#diffusers.loaders.StableDiffusionXLLoraLoaderMixin.load_lora_weights">load_lora_weights()</a> for loading LoRA weights</li> <li><a href="/docs/diffusers/main/en/api/loaders/lora#diffusers.loaders.StableDiffusionXLLoraLoaderMixin.save_lora_weights">save_lora_weights()</a> for saving LoRA weights</li> <li><a href="/docs/diffusers/main/en/api/loaders/ip_adapter#diffusers.loaders.IPAdapterMixin.load_ip_adapter">load_ip_adapter()</a> for loading IP Adapters</li>',xe,w,z,De,W,ze="Encodes the prompt into text encoder hidden states.",ie,X,ne,G,se;return k=new Pe({props:{title:"K-Diffusion",local:"k-diffusion",headingTag:"h1"}}),T=new Pe({props:{title:"StableDiffusionKDiffusionPipeline",local:"diffusers.StableDiffusionKDiffusionPipeline",headingTag:"h2"}}),K=new de({props:{name:"class diffusers.StableDiffusionKDiffusionPipeline",anchor:"diffusers.StableDiffusionKDiffusionPipeline",parameters:[{name:"vae",val:""},{name:"text_encoder",val:""},{name:"tokenizer",val:""},{name:"unet",val:""},{name:"scheduler",val:""},{name:"safety_checker",val:""},{name:"feature_extractor",val:""},{name:"requires_safety_checker",val:": bool = True"}],parametersDescription:[{anchor:"diffusers.StableDiffusionKDiffusionPipeline.vae",description:`<strong>vae</strong> (<a href="/docs/diffusers/main/en/api/models/autoencoderkl#diffusers.AutoencoderKL">AutoencoderKL</a>) — | |
| Variational Auto-Encoder (VAE) Model to encode and decode images to and from latent representations.`,name:"vae"},{anchor:"diffusers.StableDiffusionKDiffusionPipeline.text_encoder",description:`<strong>text_encoder</strong> (<code>CLIPTextModel</code>) — | |
| Frozen text-encoder. Stable Diffusion uses the text portion of | |
| <a href="https://huggingface.co/docs/transformers/model_doc/clip#transformers.CLIPTextModel" rel="nofollow">CLIP</a>, specifically | |
| the <a href="https://huggingface.co/openai/clip-vit-large-patch14" rel="nofollow">clip-vit-large-patch14</a> variant.`,name:"text_encoder"},{anchor:"diffusers.StableDiffusionKDiffusionPipeline.tokenizer",description:`<strong>tokenizer</strong> (<code>CLIPTokenizer</code>) — | |
| Tokenizer of class | |
| <a href="https://huggingface.co/docs/transformers/v4.21.0/en/model_doc/clip#transformers.CLIPTokenizer" rel="nofollow">CLIPTokenizer</a>.`,name:"tokenizer"},{anchor:"diffusers.StableDiffusionKDiffusionPipeline.unet",description:'<strong>unet</strong> (<a href="/docs/diffusers/main/en/api/models/unet2d-cond#diffusers.UNet2DConditionModel">UNet2DConditionModel</a>) — Conditional U-Net architecture to denoise the encoded image latents.',name:"unet"},{anchor:"diffusers.StableDiffusionKDiffusionPipeline.scheduler",description:`<strong>scheduler</strong> (<a href="/docs/diffusers/main/en/api/schedulers/overview#diffusers.SchedulerMixin">SchedulerMixin</a>) — | |
| A scheduler to be used in combination with <code>unet</code> to denoise the encoded image latents. Can be one of | |
| <a href="/docs/diffusers/main/en/api/schedulers/ddim#diffusers.DDIMScheduler">DDIMScheduler</a>, <a href="/docs/diffusers/main/en/api/schedulers/lms_discrete#diffusers.LMSDiscreteScheduler">LMSDiscreteScheduler</a>, or <a href="/docs/diffusers/main/en/api/schedulers/pndm#diffusers.PNDMScheduler">PNDMScheduler</a>.`,name:"scheduler"},{anchor:"diffusers.StableDiffusionKDiffusionPipeline.safety_checker",description:`<strong>safety_checker</strong> (<code>StableDiffusionSafetyChecker</code>) — | |
| Classification module that estimates whether generated images could be considered offensive or harmful. | |
| Please, refer to the <a href="https://huggingface.co/runwayml/stable-diffusion-v1-5" rel="nofollow">model card</a> for details.`,name:"safety_checker"},{anchor:"diffusers.StableDiffusionKDiffusionPipeline.feature_extractor",description:`<strong>feature_extractor</strong> (<code>CLIPImageProcessor</code>) — | |
| Model that extracts features from generated images to be used as inputs for the <code>safety_checker</code>.`,name:"feature_extractor"}],source:"https://github.com/huggingface/diffusers/blob/main/src/diffusers/pipelines/stable_diffusion_k_diffusion/pipeline_stable_diffusion_k_diffusion.py#L50"}}),P=new Re({props:{warning:!0,$$slots:{default:[Ve]},$$scope:{ctx:Y}}}),M=new de({props:{name:"encode_prompt",anchor:"diffusers.StableDiffusionKDiffusionPipeline.encode_prompt",parameters:[{name:"prompt",val:""},{name:"device",val:""},{name:"num_images_per_prompt",val:""},{name:"do_classifier_free_guidance",val:""},{name:"negative_prompt",val:" = None"},{name:"prompt_embeds",val:": Optional = None"},{name:"negative_prompt_embeds",val:": Optional = None"},{name:"lora_scale",val:": Optional = None"},{name:"clip_skip",val:": Optional = None"}],parametersDescription:[{anchor:"diffusers.StableDiffusionKDiffusionPipeline.encode_prompt.prompt",description:`<strong>prompt</strong> (<code>str</code> or <code>List[str]</code>, <em>optional</em>) — | |
| prompt to be encoded | |
| device — (<code>torch.device</code>): | |
| torch device`,name:"prompt"},{anchor:"diffusers.StableDiffusionKDiffusionPipeline.encode_prompt.num_images_per_prompt",description:`<strong>num_images_per_prompt</strong> (<code>int</code>) — | |
| number of images that should be generated per prompt`,name:"num_images_per_prompt"},{anchor:"diffusers.StableDiffusionKDiffusionPipeline.encode_prompt.do_classifier_free_guidance",description:`<strong>do_classifier_free_guidance</strong> (<code>bool</code>) — | |
| whether to use classifier free guidance or not`,name:"do_classifier_free_guidance"},{anchor:"diffusers.StableDiffusionKDiffusionPipeline.encode_prompt.negative_prompt",description:`<strong>negative_prompt</strong> (<code>str</code> or <code>List[str]</code>, <em>optional</em>) — | |
| The prompt or prompts not to guide the image generation. If not defined, one has to pass | |
| <code>negative_prompt_embeds</code> instead. Ignored when not using guidance (i.e., ignored if <code>guidance_scale</code> is | |
| less than <code>1</code>).`,name:"negative_prompt"},{anchor:"diffusers.StableDiffusionKDiffusionPipeline.encode_prompt.prompt_embeds",description:`<strong>prompt_embeds</strong> (<code>torch.Tensor</code>, <em>optional</em>) — | |
| Pre-generated text embeddings. Can be used to easily tweak text inputs, <em>e.g.</em> prompt weighting. If not | |
| provided, text embeddings will be generated from <code>prompt</code> input argument.`,name:"prompt_embeds"},{anchor:"diffusers.StableDiffusionKDiffusionPipeline.encode_prompt.negative_prompt_embeds",description:`<strong>negative_prompt_embeds</strong> (<code>torch.Tensor</code>, <em>optional</em>) — | |
| Pre-generated negative text embeddings. Can be used to easily tweak text inputs, <em>e.g.</em> prompt | |
| weighting. If not provided, negative_prompt_embeds will be generated from <code>negative_prompt</code> input | |
| argument.`,name:"negative_prompt_embeds"},{anchor:"diffusers.StableDiffusionKDiffusionPipeline.encode_prompt.lora_scale",description:`<strong>lora_scale</strong> (<code>float</code>, <em>optional</em>) — | |
| A LoRA scale that will be applied to all LoRA layers of the text encoder if LoRA layers are loaded.`,name:"lora_scale"},{anchor:"diffusers.StableDiffusionKDiffusionPipeline.encode_prompt.clip_skip",description:`<strong>clip_skip</strong> (<code>int</code>, <em>optional</em>) — | |
| Number of layers to be skipped from CLIP while computing the prompt embeddings. A value of 1 means that | |
| the output of the pre-final layer will be used for computing the prompt embeddings.`,name:"clip_skip"}],source:"https://github.com/huggingface/diffusers/blob/main/src/diffusers/pipelines/stable_diffusion_k_diffusion/pipeline_stable_diffusion_k_diffusion.py#L183"}}),I=new Pe({props:{title:"StableDiffusionXLKDiffusionPipeline",local:"diffusers.StableDiffusionXLKDiffusionPipeline",headingTag:"h2"}}),y=new de({props:{name:"class diffusers.StableDiffusionXLKDiffusionPipeline",anchor:"diffusers.StableDiffusionXLKDiffusionPipeline",parameters:[{name:"vae",val:": AutoencoderKL"},{name:"text_encoder",val:": CLIPTextModel"},{name:"text_encoder_2",val:": CLIPTextModelWithProjection"},{name:"tokenizer",val:": CLIPTokenizer"},{name:"tokenizer_2",val:": CLIPTokenizer"},{name:"unet",val:": UNet2DConditionModel"},{name:"scheduler",val:": KarrasDiffusionSchedulers"},{name:"force_zeros_for_empty_prompt",val:": bool = True"}],parametersDescription:[{anchor:"diffusers.StableDiffusionXLKDiffusionPipeline.vae",description:`<strong>vae</strong> (<a href="/docs/diffusers/main/en/api/models/autoencoderkl#diffusers.AutoencoderKL">AutoencoderKL</a>) — | |
| Variational Auto-Encoder (VAE) Model to encode and decode images to and from latent representations.`,name:"vae"},{anchor:"diffusers.StableDiffusionXLKDiffusionPipeline.text_encoder",description:`<strong>text_encoder</strong> (<code>CLIPTextModel</code>) — | |
| Frozen text-encoder. Stable Diffusion XL uses the text portion of | |
| <a href="https://huggingface.co/docs/transformers/model_doc/clip#transformers.CLIPTextModel" rel="nofollow">CLIP</a>, specifically | |
| the <a href="https://huggingface.co/openai/clip-vit-large-patch14" rel="nofollow">clip-vit-large-patch14</a> variant.`,name:"text_encoder"},{anchor:"diffusers.StableDiffusionXLKDiffusionPipeline.text_encoder_2",description:`<strong>text_encoder_2</strong> (<code> CLIPTextModelWithProjection</code>) — | |
| Second frozen text-encoder. Stable Diffusion XL uses the text and pool portion of | |
| <a href="https://huggingface.co/docs/transformers/model_doc/clip#transformers.CLIPTextModelWithProjection" rel="nofollow">CLIP</a>, | |
| specifically the | |
| <a href="https://huggingface.co/laion/CLIP-ViT-bigG-14-laion2B-39B-b160k" rel="nofollow">laion/CLIP-ViT-bigG-14-laion2B-39B-b160k</a> | |
| variant.`,name:"text_encoder_2"},{anchor:"diffusers.StableDiffusionXLKDiffusionPipeline.tokenizer",description:`<strong>tokenizer</strong> (<code>CLIPTokenizer</code>) — | |
| Tokenizer of class | |
| <a href="https://huggingface.co/docs/transformers/v4.21.0/en/model_doc/clip#transformers.CLIPTokenizer" rel="nofollow">CLIPTokenizer</a>.`,name:"tokenizer"},{anchor:"diffusers.StableDiffusionXLKDiffusionPipeline.tokenizer_2",description:`<strong>tokenizer_2</strong> (<code>CLIPTokenizer</code>) — | |
| Second Tokenizer of class | |
| <a href="https://huggingface.co/docs/transformers/v4.21.0/en/model_doc/clip#transformers.CLIPTokenizer" rel="nofollow">CLIPTokenizer</a>.`,name:"tokenizer_2"},{anchor:"diffusers.StableDiffusionXLKDiffusionPipeline.unet",description:'<strong>unet</strong> (<a href="/docs/diffusers/main/en/api/models/unet2d-cond#diffusers.UNet2DConditionModel">UNet2DConditionModel</a>) — Conditional U-Net architecture to denoise the encoded image latents.',name:"unet"},{anchor:"diffusers.StableDiffusionXLKDiffusionPipeline.scheduler",description:`<strong>scheduler</strong> (<a href="/docs/diffusers/main/en/api/schedulers/overview#diffusers.SchedulerMixin">SchedulerMixin</a>) — | |
| A scheduler to be used in combination with <code>unet</code> to denoise the encoded image latents. Can be one of | |
| <a href="/docs/diffusers/main/en/api/schedulers/ddim#diffusers.DDIMScheduler">DDIMScheduler</a>, <a href="/docs/diffusers/main/en/api/schedulers/lms_discrete#diffusers.LMSDiscreteScheduler">LMSDiscreteScheduler</a>, or <a href="/docs/diffusers/main/en/api/schedulers/pndm#diffusers.PNDMScheduler">PNDMScheduler</a>.`,name:"scheduler"},{anchor:"diffusers.StableDiffusionXLKDiffusionPipeline.force_zeros_for_empty_prompt",description:`<strong>force_zeros_for_empty_prompt</strong> (<code>bool</code>, <em>optional</em>, defaults to <code>"True"</code>) — | |
| Whether the negative prompt embeddings shall be forced to always be set to 0. Also see the config of | |
| <code>stabilityai/stable-diffusion-xl-base-1-0</code>.`,name:"force_zeros_for_empty_prompt"}],source:"https://github.com/huggingface/diffusers/blob/main/src/diffusers/pipelines/stable_diffusion_k_diffusion/pipeline_stable_diffusion_xl_k_diffusion.py#L90"}}),z=new de({props:{name:"encode_prompt",anchor:"diffusers.StableDiffusionXLKDiffusionPipeline.encode_prompt",parameters:[{name:"prompt",val:": str"},{name:"prompt_2",val:": Optional = None"},{name:"device",val:": Optional = None"},{name:"num_images_per_prompt",val:": int = 1"},{name:"do_classifier_free_guidance",val:": bool = True"},{name:"negative_prompt",val:": Optional = None"},{name:"negative_prompt_2",val:": Optional = None"},{name:"prompt_embeds",val:": Optional = None"},{name:"negative_prompt_embeds",val:": Optional = None"},{name:"pooled_prompt_embeds",val:": Optional = None"},{name:"negative_pooled_prompt_embeds",val:": Optional = None"},{name:"lora_scale",val:": Optional = None"},{name:"clip_skip",val:": Optional = None"}],parametersDescription:[{anchor:"diffusers.StableDiffusionXLKDiffusionPipeline.encode_prompt.prompt",description:`<strong>prompt</strong> (<code>str</code> or <code>List[str]</code>, <em>optional</em>) — | |
| prompt to be encoded`,name:"prompt"},{anchor:"diffusers.StableDiffusionXLKDiffusionPipeline.encode_prompt.prompt_2",description:`<strong>prompt_2</strong> (<code>str</code> or <code>List[str]</code>, <em>optional</em>) — | |
| The prompt or prompts to be sent to the <code>tokenizer_2</code> and <code>text_encoder_2</code>. If not defined, <code>prompt</code> is | |
| used in both text-encoders | |
| device — (<code>torch.device</code>): | |
| torch device`,name:"prompt_2"},{anchor:"diffusers.StableDiffusionXLKDiffusionPipeline.encode_prompt.num_images_per_prompt",description:`<strong>num_images_per_prompt</strong> (<code>int</code>) — | |
| number of images that should be generated per prompt`,name:"num_images_per_prompt"},{anchor:"diffusers.StableDiffusionXLKDiffusionPipeline.encode_prompt.do_classifier_free_guidance",description:`<strong>do_classifier_free_guidance</strong> (<code>bool</code>) — | |
| whether to use classifier free guidance or not`,name:"do_classifier_free_guidance"},{anchor:"diffusers.StableDiffusionXLKDiffusionPipeline.encode_prompt.negative_prompt",description:`<strong>negative_prompt</strong> (<code>str</code> or <code>List[str]</code>, <em>optional</em>) — | |
| The prompt or prompts not to guide the image generation. If not defined, one has to pass | |
| <code>negative_prompt_embeds</code> instead. Ignored when not using guidance (i.e., ignored if <code>guidance_scale</code> is | |
| less than <code>1</code>).`,name:"negative_prompt"},{anchor:"diffusers.StableDiffusionXLKDiffusionPipeline.encode_prompt.negative_prompt_2",description:`<strong>negative_prompt_2</strong> (<code>str</code> or <code>List[str]</code>, <em>optional</em>) — | |
| The prompt or prompts not to guide the image generation to be sent to <code>tokenizer_2</code> and | |
| <code>text_encoder_2</code>. If not defined, <code>negative_prompt</code> is used in both text-encoders`,name:"negative_prompt_2"},{anchor:"diffusers.StableDiffusionXLKDiffusionPipeline.encode_prompt.prompt_embeds",description:`<strong>prompt_embeds</strong> (<code>torch.Tensor</code>, <em>optional</em>) — | |
| Pre-generated text embeddings. Can be used to easily tweak text inputs, <em>e.g.</em> prompt weighting. If not | |
| provided, text embeddings will be generated from <code>prompt</code> input argument.`,name:"prompt_embeds"},{anchor:"diffusers.StableDiffusionXLKDiffusionPipeline.encode_prompt.negative_prompt_embeds",description:`<strong>negative_prompt_embeds</strong> (<code>torch.Tensor</code>, <em>optional</em>) — | |
| Pre-generated negative text embeddings. Can be used to easily tweak text inputs, <em>e.g.</em> prompt | |
| weighting. If not provided, negative_prompt_embeds will be generated from <code>negative_prompt</code> input | |
| argument.`,name:"negative_prompt_embeds"},{anchor:"diffusers.StableDiffusionXLKDiffusionPipeline.encode_prompt.pooled_prompt_embeds",description:`<strong>pooled_prompt_embeds</strong> (<code>torch.Tensor</code>, <em>optional</em>) — | |
| Pre-generated pooled text embeddings. Can be used to easily tweak text inputs, <em>e.g.</em> prompt weighting. | |
| If not provided, pooled text embeddings will be generated from <code>prompt</code> input argument.`,name:"pooled_prompt_embeds"},{anchor:"diffusers.StableDiffusionXLKDiffusionPipeline.encode_prompt.negative_pooled_prompt_embeds",description:`<strong>negative_pooled_prompt_embeds</strong> (<code>torch.Tensor</code>, <em>optional</em>) — | |
| Pre-generated negative pooled text embeddings. Can be used to easily tweak text inputs, <em>e.g.</em> prompt | |
| weighting. If not provided, pooled negative_prompt_embeds will be generated from <code>negative_prompt</code> | |
| input argument.`,name:"negative_pooled_prompt_embeds"},{anchor:"diffusers.StableDiffusionXLKDiffusionPipeline.encode_prompt.lora_scale",description:`<strong>lora_scale</strong> (<code>float</code>, <em>optional</em>) — | |
| A lora scale that will be applied to all LoRA layers of the text encoder if LoRA layers are loaded.`,name:"lora_scale"},{anchor:"diffusers.StableDiffusionXLKDiffusionPipeline.encode_prompt.clip_skip",description:`<strong>clip_skip</strong> (<code>int</code>, <em>optional</em>) — | |
| Number of layers to be skipped from CLIP while computing the prompt embeddings. A value of 1 means that | |
| the output of the pre-final layer will be used for computing the prompt embeddings.`,name:"clip_skip"}],source:"https://github.com/huggingface/diffusers/blob/main/src/diffusers/pipelines/stable_diffusion_k_diffusion/pipeline_stable_diffusion_xl_k_diffusion.py#L199"}}),X=new Ue({props:{source:"https://github.com/huggingface/diffusers/blob/main/docs/source/en/api/pipelines/stable_diffusion/k_diffusion.md"}}),{c(){f=r("meta"),S=n(),g=r("p"),A=n(),h(k.$$.fragment),J=n(),$=r("p"),$.innerHTML=Le,Q=n(),C=r("p"),C.innerHTML=we,Z=n(),h(T.$$.fragment),ee=n(),d=r("div"),h(K.$$.fragment),fe=n(),N=r("p"),N.textContent=Se,le=n(),E=r("p"),E.innerHTML=ke,pe=n(),O=r("p"),O.textContent=$e,ce=n(),H=r("ul"),H.innerHTML=Ce,me=n(),h(P.$$.fragment),ue=n(),L=r("div"),h(M.$$.fragment),ge=n(),R=r("p"),R.textContent=Te,oe=n(),h(I.$$.fragment),te=n(),l=r("div"),h(y.$$.fragment),he=n(),U=r("p"),U.textContent=Ke,_e=n(),V=r("p"),V.innerHTML=Me,be=n(),q=r("p"),q.textContent=Ie,ve=n(),j=r("ul"),j.innerHTML=ye,xe=n(),w=r("div"),h(z.$$.fragment),De=n(),W=r("p"),W.textContent=ze,ie=n(),h(X.$$.fragment),ne=n(),G=r("p"),this.h()},l(e){const o=He("svelte-u9bgzb",document.head);f=a(o,"META",{name:!0,content:!0}),o.forEach(i),S=s(e),g=a(e,"P",{}),B(g).forEach(i),A=s(e),_(k.$$.fragment,e),J=s(e),$=a(e,"P",{"data-svelte-h":!0}),m($)!=="svelte-1u0qckb"&&($.innerHTML=Le),Q=s(e),C=a(e,"P",{"data-svelte-h":!0}),m(C)!=="svelte-kxbfc9"&&(C.innerHTML=we),Z=s(e),_(T.$$.fragment,e),ee=s(e),d=a(e,"DIV",{class:!0});var p=B(d);_(K.$$.fragment,p),fe=s(p),N=a(p,"P",{"data-svelte-h":!0}),m(N)!=="svelte-35p306"&&(N.textContent=Se),le=s(p),E=a(p,"P",{"data-svelte-h":!0}),m(E)!=="svelte-wu1jmh"&&(E.innerHTML=ke),pe=s(p),O=a(p,"P",{"data-svelte-h":!0}),m(O)!=="svelte-14s6m4u"&&(O.textContent=$e),ce=s(p),H=a(p,"UL",{"data-svelte-h":!0}),m(H)!=="svelte-txtwv6"&&(H.innerHTML=Ce),me=s(p),_(P.$$.fragment,p),ue=s(p),L=a(p,"DIV",{class:!0});var re=B(L);_(M.$$.fragment,re),ge=s(re),R=a(re,"P",{"data-svelte-h":!0}),m(R)!=="svelte-16q0ax1"&&(R.textContent=Te),re.forEach(i),p.forEach(i),oe=s(e),_(I.$$.fragment,e),te=s(e),l=a(e,"DIV",{class:!0});var u=B(l);_(y.$$.fragment,u),he=s(u),U=a(u,"P",{"data-svelte-h":!0}),m(U)!=="svelte-3dg7oo"&&(U.textContent=Ke),_e=s(u),V=a(u,"P",{"data-svelte-h":!0}),m(V)!=="svelte-wu1jmh"&&(V.innerHTML=Me),be=s(u),q=a(u,"P",{"data-svelte-h":!0}),m(q)!=="svelte-14s6m4u"&&(q.textContent=Ie),ve=s(u),j=a(u,"UL",{"data-svelte-h":!0}),m(j)!=="svelte-9wc0dt"&&(j.innerHTML=ye),xe=s(u),w=a(u,"DIV",{class:!0});var ae=B(w);_(z.$$.fragment,ae),De=s(ae),W=a(ae,"P",{"data-svelte-h":!0}),m(W)!=="svelte-16q0ax1"&&(W.textContent=ze),ae.forEach(i),u.forEach(i),ie=s(e),_(X.$$.fragment,e),ne=s(e),G=a(e,"P",{}),B(G).forEach(i),this.h()},h(){F(f,"name","hf:doc:metadata"),F(f,"content",je),F(L,"class","docstring border-l-2 border-t-2 pl-4 pt-3.5 border-gray-100 rounded-tl-xl mb-6 mt-8"),F(d,"class","docstring border-l-2 border-t-2 pl-4 pt-3.5 border-gray-100 rounded-tl-xl mb-6 mt-8"),F(w,"class","docstring border-l-2 border-t-2 pl-4 pt-3.5 border-gray-100 rounded-tl-xl mb-6 mt-8"),F(l,"class","docstring border-l-2 border-t-2 pl-4 pt-3.5 border-gray-100 rounded-tl-xl mb-6 mt-8")},m(e,o){t(document.head,f),c(e,S,o),c(e,g,o),c(e,A,o),b(k,e,o),c(e,J,o),c(e,$,o),c(e,Q,o),c(e,C,o),c(e,Z,o),b(T,e,o),c(e,ee,o),c(e,d,o),b(K,d,null),t(d,fe),t(d,N),t(d,le),t(d,E),t(d,pe),t(d,O),t(d,ce),t(d,H),t(d,me),b(P,d,null),t(d,ue),t(d,L),b(M,L,null),t(L,ge),t(L,R),c(e,oe,o),b(I,e,o),c(e,te,o),c(e,l,o),b(y,l,null),t(l,he),t(l,U),t(l,_e),t(l,V),t(l,be),t(l,q),t(l,ve),t(l,j),t(l,xe),t(l,w),b(z,w,null),t(w,De),t(w,W),c(e,ie,o),b(X,e,o),c(e,ne,o),c(e,G,o),se=!0},p(e,[o]){const p={};o&2&&(p.$$scope={dirty:o,ctx:e}),P.$set(p)},i(e){se||(v(k.$$.fragment,e),v(T.$$.fragment,e),v(K.$$.fragment,e),v(P.$$.fragment,e),v(M.$$.fragment,e),v(I.$$.fragment,e),v(y.$$.fragment,e),v(z.$$.fragment,e),v(X.$$.fragment,e),se=!0)},o(e){x(k.$$.fragment,e),x(T.$$.fragment,e),x(K.$$.fragment,e),x(P.$$.fragment,e),x(M.$$.fragment,e),x(I.$$.fragment,e),x(y.$$.fragment,e),x(z.$$.fragment,e),x(X.$$.fragment,e),se=!1},d(e){e&&(i(S),i(g),i(A),i(J),i($),i(Q),i(C),i(Z),i(ee),i(d),i(oe),i(te),i(l),i(ie),i(ne),i(G)),i(f),D(k,e),D(T,e),D(K),D(P),D(M),D(I,e),D(y),D(z),D(X,e)}}}const je='{"title":"K-Diffusion","local":"k-diffusion","sections":[{"title":"StableDiffusionKDiffusionPipeline","local":"diffusers.StableDiffusionKDiffusionPipeline","sections":[],"depth":2},{"title":"StableDiffusionXLKDiffusionPipeline","local":"diffusers.StableDiffusionXLKDiffusionPipeline","sections":[],"depth":2}],"depth":1}';function We(Y){return Ae(()=>{new URLSearchParams(window.location.search).get("fw")}),[]}class Qe extends Ee{constructor(f){super(),Oe(this,f,We,qe,Xe,{})}}export{Qe as component}; | |
Xet Storage Details
- Size:
- 25.3 kB
- Xet hash:
- 72973b30d589a54114e9ce545d4c43791b79494d7d8533bfbc9e0ccbd780e5a8
·
Xet efficiently stores files, intelligently splitting them into unique chunks and accelerating uploads and downloads. More info.