Buckets:
| import{s as th,o as ah,n as D}from"../chunks/scheduler.8c3d61f6.js";import{S as nh,i as sh,g as n,s as r,r as p,A as ih,h as s,f as i,c as t,j as v,u as m,x as f,k as w,y as o,a as x,v as _,d as u,t as h,w as g}from"../chunks/index.da70eac4.js";import{T as S}from"../chunks/Tip.1d9b8c37.js";import{D as $}from"../chunks/Docstring.9419aa1d.js";import{C as gl}from"../chunks/CodeBlock.a9c4becf.js";import{E as hl}from"../chunks/ExampleCodeBlock.1b2603c3.js";import{H as J,E as dh}from"../chunks/getInferenceSnippets.39110341.js";function lh(y){let a,b='To learn more about how to load LoRA weights, see the <a href="../../using-diffusers/loading_adapters#lora">LoRA</a> loading guide.';return{c(){a=n("p"),a.innerHTML=b},l(l){a=s(l,"P",{"data-svelte-h":!0}),f(a)!=="svelte-1fw6lx1"&&(a.innerHTML=b)},m(l,c){x(l,a,c)},p:D,d(l){l&&i(a)}}}function ch(y){let a,b="We support loading A1111 formatted LoRA checkpoints in a limited capacity.",l,c,T="This function is experimental and might change in the future.";return{c(){a=n("p"),a.textContent=b,l=r(),c=n("p"),c.textContent=T},l(d){a=s(d,"P",{"data-svelte-h":!0}),f(a)!=="svelte-15l1sdn"&&(a.textContent=b),l=t(d),c=s(d,"P",{"data-svelte-h":!0}),f(c)!=="svelte-3fufvn"&&(c.textContent=T)},m(d,M){x(d,a,M),x(d,l,M),x(d,c,M)},p:D,d(d){d&&(i(a),i(l),i(c))}}}function fh(y){let a,b="We support loading A1111 formatted LoRA checkpoints in a limited capacity.",l,c,T="This function is experimental and might change in the future.";return{c(){a=n("p"),a.textContent=b,l=r(),c=n("p"),c.textContent=T},l(d){a=s(d,"P",{"data-svelte-h":!0}),f(a)!=="svelte-15l1sdn"&&(a.textContent=b),l=t(d),c=s(d,"P",{"data-svelte-h":!0}),f(c)!=="svelte-3fufvn"&&(c.textContent=T)},m(d,M){x(d,a,M),x(d,l,M),x(d,c,M)},p:D,d(d){d&&(i(a),i(l),i(c))}}}function ph(y){let a,b="We support loading A1111 formatted LoRA checkpoints in a limited capacity.",l,c,T="This function is experimental and might change in the future.";return{c(){a=n("p"),a.textContent=b,l=r(),c=n("p"),c.textContent=T},l(d){a=s(d,"P",{"data-svelte-h":!0}),f(a)!=="svelte-15l1sdn"&&(a.textContent=b),l=t(d),c=s(d,"P",{"data-svelte-h":!0}),f(c)!=="svelte-3fufvn"&&(c.textContent=T)},m(d,M){x(d,a,M),x(d,l,M),x(d,c,M)},p:D,d(d){d&&(i(a),i(l),i(c))}}}function mh(y){let a,b="This is an experimental API.";return{c(){a=n("p"),a.textContent=b},l(l){a=s(l,"P",{"data-svelte-h":!0}),f(a)!=="svelte-8w79b9"&&(a.textContent=b)},m(l,c){x(l,a,c)},p:D,d(l){l&&i(a)}}}function _h(y){let a,b="We support loading A1111 formatted LoRA checkpoints in a limited capacity.",l,c,T="This function is experimental and might change in the future.";return{c(){a=n("p"),a.textContent=b,l=r(),c=n("p"),c.textContent=T},l(d){a=s(d,"P",{"data-svelte-h":!0}),f(a)!=="svelte-15l1sdn"&&(a.textContent=b),l=t(d),c=s(d,"P",{"data-svelte-h":!0}),f(c)!=="svelte-3fufvn"&&(c.textContent=T)},m(d,M){x(d,a,M),x(d,l,M),x(d,c,M)},p:D,d(d){d&&(i(a),i(l),i(c))}}}function uh(y){let a,b="This is an experimental API.";return{c(){a=n("p"),a.textContent=b},l(l){a=s(l,"P",{"data-svelte-h":!0}),f(a)!=="svelte-8w79b9"&&(a.textContent=b)},m(l,c){x(l,a,c)},p:D,d(l){l&&i(a)}}}function hh(y){let a,b="Examples:",l,c,T;return c=new gl({props:{code:"JTIzJTIwQXNzdW1pbmclMjAlNjBwaXBlbGluZSU2MCUyMGlzJTIwYWxyZWFkeSUyMGxvYWRlZCUyMHdpdGglMjB0aGUlMjBMb1JBJTIwcGFyYW1ldGVycy4lMEFwaXBlbGluZS51bmxvYWRfbG9yYV93ZWlnaHRzKCklMEEuLi4=",highlighted:'<span class="hljs-meta">>>> </span><span class="hljs-comment"># Assuming `pipeline` is already loaded with the LoRA parameters.</span>\n<span class="hljs-meta">>>> </span>pipeline.unload_lora_weights()\n<span class="hljs-meta">>>> </span>...',wrap:!1}}),{c(){a=n("p"),a.textContent=b,l=r(),p(c.$$.fragment)},l(d){a=s(d,"P",{"data-svelte-h":!0}),f(a)!=="svelte-kvfsh7"&&(a.textContent=b),l=t(d),m(c.$$.fragment,d)},m(d,M){x(d,a,M),x(d,l,M),_(c,d,M),T=!0},p:D,i(d){T||(u(c.$$.fragment,d),T=!0)},o(d){h(c.$$.fragment,d),T=!1},d(d){d&&(i(a),i(l)),g(c,d)}}}function gh(y){let a,b="We support loading A1111 formatted LoRA checkpoints in a limited capacity.",l,c,T="This function is experimental and might change in the future.";return{c(){a=n("p"),a.textContent=b,l=r(),c=n("p"),c.textContent=T},l(d){a=s(d,"P",{"data-svelte-h":!0}),f(a)!=="svelte-15l1sdn"&&(a.textContent=b),l=t(d),c=s(d,"P",{"data-svelte-h":!0}),f(c)!=="svelte-3fufvn"&&(c.textContent=T)},m(d,M){x(d,a,M),x(d,l,M),x(d,c,M)},p:D,d(d){d&&(i(a),i(l),i(c))}}}function Lh(y){let a,b="This is an experimental API.";return{c(){a=n("p"),a.textContent=b},l(l){a=s(l,"P",{"data-svelte-h":!0}),f(a)!=="svelte-8w79b9"&&(a.textContent=b)},m(l,c){x(l,a,c)},p:D,d(l){l&&i(a)}}}function xh(y){let a,b="We support loading A1111 formatted LoRA checkpoints in a limited capacity.",l,c,T="This function is experimental and might change in the future.";return{c(){a=n("p"),a.textContent=b,l=r(),c=n("p"),c.textContent=T},l(d){a=s(d,"P",{"data-svelte-h":!0}),f(a)!=="svelte-15l1sdn"&&(a.textContent=b),l=t(d),c=s(d,"P",{"data-svelte-h":!0}),f(c)!=="svelte-3fufvn"&&(c.textContent=T)},m(d,M){x(d,a,M),x(d,l,M),x(d,c,M)},p:D,d(d){d&&(i(a),i(l),i(c))}}}function bh(y){let a,b="This is an experimental API.";return{c(){a=n("p"),a.textContent=b},l(l){a=s(l,"P",{"data-svelte-h":!0}),f(a)!=="svelte-8w79b9"&&(a.textContent=b)},m(l,c){x(l,a,c)},p:D,d(l){l&&i(a)}}}function vh(y){let a,b="We support loading A1111 formatted LoRA checkpoints in a limited capacity.",l,c,T="This function is experimental and might change in the future.";return{c(){a=n("p"),a.textContent=b,l=r(),c=n("p"),c.textContent=T},l(d){a=s(d,"P",{"data-svelte-h":!0}),f(a)!=="svelte-15l1sdn"&&(a.textContent=b),l=t(d),c=s(d,"P",{"data-svelte-h":!0}),f(c)!=="svelte-3fufvn"&&(c.textContent=T)},m(d,M){x(d,a,M),x(d,l,M),x(d,c,M)},p:D,d(d){d&&(i(a),i(l),i(c))}}}function wh(y){let a,b="This is an experimental API.";return{c(){a=n("p"),a.textContent=b},l(l){a=s(l,"P",{"data-svelte-h":!0}),f(a)!=="svelte-8w79b9"&&(a.textContent=b)},m(l,c){x(l,a,c)},p:D,d(l){l&&i(a)}}}function $h(y){let a,b="We support loading A1111 formatted LoRA checkpoints in a limited capacity.",l,c,T="This function is experimental and might change in the future.";return{c(){a=n("p"),a.textContent=b,l=r(),c=n("p"),c.textContent=T},l(d){a=s(d,"P",{"data-svelte-h":!0}),f(a)!=="svelte-15l1sdn"&&(a.textContent=b),l=t(d),c=s(d,"P",{"data-svelte-h":!0}),f(c)!=="svelte-3fufvn"&&(c.textContent=T)},m(d,M){x(d,a,M),x(d,l,M),x(d,c,M)},p:D,d(d){d&&(i(a),i(l),i(c))}}}function Mh(y){let a,b="This is an experimental API.";return{c(){a=n("p"),a.textContent=b},l(l){a=s(l,"P",{"data-svelte-h":!0}),f(a)!=="svelte-8w79b9"&&(a.textContent=b)},m(l,c){x(l,a,c)},p:D,d(l){l&&i(a)}}}function yh(y){let a,b="We support loading A1111 formatted LoRA checkpoints in a limited capacity.",l,c,T="This function is experimental and might change in the future.";return{c(){a=n("p"),a.textContent=b,l=r(),c=n("p"),c.textContent=T},l(d){a=s(d,"P",{"data-svelte-h":!0}),f(a)!=="svelte-15l1sdn"&&(a.textContent=b),l=t(d),c=s(d,"P",{"data-svelte-h":!0}),f(c)!=="svelte-3fufvn"&&(c.textContent=T)},m(d,M){x(d,a,M),x(d,l,M),x(d,c,M)},p:D,d(d){d&&(i(a),i(l),i(c))}}}function Th(y){let a,b="This is an experimental API.";return{c(){a=n("p"),a.textContent=b},l(l){a=s(l,"P",{"data-svelte-h":!0}),f(a)!=="svelte-8w79b9"&&(a.textContent=b)},m(l,c){x(l,a,c)},p:D,d(l){l&&i(a)}}}function Dh(y){let a,b="We support loading original format HunyuanVideo LoRA checkpoints.",l,c,T="This function is experimental and might change in the future.";return{c(){a=n("p"),a.textContent=b,l=r(),c=n("p"),c.textContent=T},l(d){a=s(d,"P",{"data-svelte-h":!0}),f(a)!=="svelte-gyrs6h"&&(a.textContent=b),l=t(d),c=s(d,"P",{"data-svelte-h":!0}),f(c)!=="svelte-3fufvn"&&(c.textContent=T)},m(d,M){x(d,a,M),x(d,l,M),x(d,c,M)},p:D,d(d){d&&(i(a),i(l),i(c))}}}function Sh(y){let a,b="This is an experimental API.";return{c(){a=n("p"),a.textContent=b},l(l){a=s(l,"P",{"data-svelte-h":!0}),f(a)!=="svelte-8w79b9"&&(a.textContent=b)},m(l,c){x(l,a,c)},p:D,d(l){l&&i(a)}}}function Ch(y){let a,b="We support loading A1111 formatted LoRA checkpoints in a limited capacity.",l,c,T="This function is experimental and might change in the future.";return{c(){a=n("p"),a.textContent=b,l=r(),c=n("p"),c.textContent=T},l(d){a=s(d,"P",{"data-svelte-h":!0}),f(a)!=="svelte-15l1sdn"&&(a.textContent=b),l=t(d),c=s(d,"P",{"data-svelte-h":!0}),f(c)!=="svelte-3fufvn"&&(c.textContent=T)},m(d,M){x(d,a,M),x(d,l,M),x(d,c,M)},p:D,d(d){d&&(i(a),i(l),i(c))}}}function Ah(y){let a,b="This is an experimental API.";return{c(){a=n("p"),a.textContent=b},l(l){a=s(l,"P",{"data-svelte-h":!0}),f(a)!=="svelte-8w79b9"&&(a.textContent=b)},m(l,c){x(l,a,c)},p:D,d(l){l&&i(a)}}}function kh(y){let a,b="We support loading A1111 formatted LoRA checkpoints in a limited capacity.",l,c,T="This function is experimental and might change in the future.";return{c(){a=n("p"),a.textContent=b,l=r(),c=n("p"),c.textContent=T},l(d){a=s(d,"P",{"data-svelte-h":!0}),f(a)!=="svelte-15l1sdn"&&(a.textContent=b),l=t(d),c=s(d,"P",{"data-svelte-h":!0}),f(c)!=="svelte-3fufvn"&&(c.textContent=T)},m(d,M){x(d,a,M),x(d,l,M),x(d,c,M)},p:D,d(d){d&&(i(a),i(l),i(c))}}}function Ph(y){let a,b="This is an experimental API.";return{c(){a=n("p"),a.textContent=b},l(l){a=s(l,"P",{"data-svelte-h":!0}),f(a)!=="svelte-8w79b9"&&(a.textContent=b)},m(l,c){x(l,a,c)},p:D,d(l){l&&i(a)}}}function Rh(y){let a,b="We support loading A1111 formatted LoRA checkpoints in a limited capacity.",l,c,T="This function is experimental and might change in the future.";return{c(){a=n("p"),a.textContent=b,l=r(),c=n("p"),c.textContent=T},l(d){a=s(d,"P",{"data-svelte-h":!0}),f(a)!=="svelte-15l1sdn"&&(a.textContent=b),l=t(d),c=s(d,"P",{"data-svelte-h":!0}),f(c)!=="svelte-3fufvn"&&(c.textContent=T)},m(d,M){x(d,a,M),x(d,l,M),x(d,c,M)},p:D,d(d){d&&(i(a),i(l),i(c))}}}function Hh(y){let a,b="This is an experimental API.";return{c(){a=n("p"),a.textContent=b},l(l){a=s(l,"P",{"data-svelte-h":!0}),f(a)!=="svelte-8w79b9"&&(a.textContent=b)},m(l,c){x(l,a,c)},p:D,d(l){l&&i(a)}}}function Ih(y){let a,b="We support loading A1111 formatted LoRA checkpoints in a limited capacity.",l,c,T="This function is experimental and might change in the future.";return{c(){a=n("p"),a.textContent=b,l=r(),c=n("p"),c.textContent=T},l(d){a=s(d,"P",{"data-svelte-h":!0}),f(a)!=="svelte-15l1sdn"&&(a.textContent=b),l=t(d),c=s(d,"P",{"data-svelte-h":!0}),f(c)!=="svelte-3fufvn"&&(c.textContent=T)},m(d,M){x(d,a,M),x(d,l,M),x(d,c,M)},p:D,d(d){d&&(i(a),i(l),i(c))}}}function Vh(y){let a,b="This is an experimental API.";return{c(){a=n("p"),a.textContent=b},l(l){a=s(l,"P",{"data-svelte-h":!0}),f(a)!=="svelte-8w79b9"&&(a.textContent=b)},m(l,c){x(l,a,c)},p:D,d(l){l&&i(a)}}}function Wh(y){let a,b="This is an experimental API.";return{c(){a=n("p"),a.textContent=b},l(l){a=s(l,"P",{"data-svelte-h":!0}),f(a)!=="svelte-8w79b9"&&(a.textContent=b)},m(l,c){x(l,a,c)},p:D,d(l){l&&i(a)}}}function Fh(y){let a,b="Example:",l,c,T;return c=new gl({props:{code:"ZnJvbSUyMGRpZmZ1c2VycyUyMGltcG9ydCUyMERpZmZ1c2lvblBpcGVsaW5lJTBBaW1wb3J0JTIwdG9yY2glMEElMEFwaXBlbGluZSUyMCUzRCUyMERpZmZ1c2lvblBpcGVsaW5lLmZyb21fcHJldHJhaW5lZCglMEElMjAlMjAlMjAlMjAlMjJzdGFiaWxpdHlhaSUyRnN0YWJsZS1kaWZmdXNpb24teGwtYmFzZS0xLjAlMjIlMkMlMjB0b3JjaF9kdHlwZSUzRHRvcmNoLmZsb2F0MTYlMEEpLnRvKCUyMmN1ZGElMjIpJTBBcGlwZWxpbmUubG9hZF9sb3JhX3dlaWdodHMoJTIybmVyaWpzJTJGcGl4ZWwtYXJ0LXhsJTIyJTJDJTIwd2VpZ2h0X25hbWUlM0QlMjJwaXhlbC1hcnQteGwuc2FmZXRlbnNvcnMlMjIlMkMlMjBhZGFwdGVyX25hbWUlM0QlMjJwaXhlbCUyMiklMEFwaXBlbGluZS5mdXNlX2xvcmEobG9yYV9zY2FsZSUzRDAuNyk=",highlighted:`<span class="hljs-keyword">from</span> diffusers <span class="hljs-keyword">import</span> DiffusionPipeline | |
| <span class="hljs-keyword">import</span> torch | |
| pipeline = DiffusionPipeline.from_pretrained( | |
| <span class="hljs-string">"stabilityai/stable-diffusion-xl-base-1.0"</span>, torch_dtype=torch.float16 | |
| ).to(<span class="hljs-string">"cuda"</span>) | |
| pipeline.load_lora_weights(<span class="hljs-string">"nerijs/pixel-art-xl"</span>, weight_name=<span class="hljs-string">"pixel-art-xl.safetensors"</span>, adapter_name=<span class="hljs-string">"pixel"</span>) | |
| pipeline.fuse_lora(lora_scale=<span class="hljs-number">0.7</span>)`,wrap:!1}}),{c(){a=n("p"),a.textContent=b,l=r(),p(c.$$.fragment)},l(d){a=s(d,"P",{"data-svelte-h":!0}),f(a)!=="svelte-11lpom8"&&(a.textContent=b),l=t(d),m(c.$$.fragment,d)},m(d,M){x(d,a,M),x(d,l,M),_(c,d,M),T=!0},p:D,i(d){T||(u(c.$$.fragment,d),T=!0)},o(d){h(c.$$.fragment,d),T=!1},d(d){d&&(i(a),i(l)),g(c,d)}}}function Uh(y){let a,b="Example:",l,c,T;return c=new gl({props:{code:"ZnJvbSUyMGRpZmZ1c2VycyUyMGltcG9ydCUyMERpZmZ1c2lvblBpcGVsaW5lJTBBJTBBcGlwZWxpbmUlMjAlM0QlMjBEaWZmdXNpb25QaXBlbGluZS5mcm9tX3ByZXRyYWluZWQoJTBBJTIwJTIwJTIwJTIwJTIyc3RhYmlsaXR5YWklMkZzdGFibGUtZGlmZnVzaW9uLXhsLWJhc2UtMS4wJTIyJTJDJTBBKS50byglMjJjdWRhJTIyKSUwQXBpcGVsaW5lLmxvYWRfbG9yYV93ZWlnaHRzKCUyMkNpcm9OMjAyMiUyRnRveS1mYWNlJTIyJTJDJTIwd2VpZ2h0X25hbWUlM0QlMjJ0b3lfZmFjZV9zZHhsLnNhZmV0ZW5zb3JzJTIyJTJDJTIwYWRhcHRlcl9uYW1lJTNEJTIydG95JTIyKSUwQXBpcGVsaW5lLmdldF9hY3RpdmVfYWRhcHRlcnMoKQ==",highlighted:`<span class="hljs-keyword">from</span> diffusers <span class="hljs-keyword">import</span> DiffusionPipeline | |
| pipeline = DiffusionPipeline.from_pretrained( | |
| <span class="hljs-string">"stabilityai/stable-diffusion-xl-base-1.0"</span>, | |
| ).to(<span class="hljs-string">"cuda"</span>) | |
| pipeline.load_lora_weights(<span class="hljs-string">"CiroN2022/toy-face"</span>, weight_name=<span class="hljs-string">"toy_face_sdxl.safetensors"</span>, adapter_name=<span class="hljs-string">"toy"</span>) | |
| pipeline.get_active_adapters()`,wrap:!1}}),{c(){a=n("p"),a.textContent=b,l=r(),p(c.$$.fragment)},l(d){a=s(d,"P",{"data-svelte-h":!0}),f(a)!=="svelte-11lpom8"&&(a.textContent=b),l=t(d),m(c.$$.fragment,d)},m(d,M){x(d,a,M),x(d,l,M),_(c,d,M),T=!0},p:D,i(d){T||(u(c.$$.fragment,d),T=!0)},o(d){h(c.$$.fragment,d),T=!1},d(d){d&&(i(a),i(l)),g(c,d)}}}function Eh(y){let a,b="This is an experimental API.";return{c(){a=n("p"),a.textContent=b},l(l){a=s(l,"P",{"data-svelte-h":!0}),f(a)!=="svelte-8w79b9"&&(a.textContent=b)},m(l,c){x(l,a,c)},p:D,d(l){l&&i(a)}}}function Nh(y){let a,b="Examples:",l,c,T;return c=new gl({props:{code:"JTIzJTIwQXNzdW1pbmclMjAlNjBwaXBlbGluZSU2MCUyMGlzJTIwYWxyZWFkeSUyMGxvYWRlZCUyMHdpdGglMjB0aGUlMjBMb1JBJTIwcGFyYW1ldGVycy4lMEFwaXBlbGluZS51bmxvYWRfbG9yYV93ZWlnaHRzKCklMEEuLi4=",highlighted:'<span class="hljs-meta">>>> </span><span class="hljs-comment"># Assuming `pipeline` is already loaded with the LoRA parameters.</span>\n<span class="hljs-meta">>>> </span>pipeline.unload_lora_weights()\n<span class="hljs-meta">>>> </span>...',wrap:!1}}),{c(){a=n("p"),a.textContent=b,l=r(),p(c.$$.fragment)},l(d){a=s(d,"P",{"data-svelte-h":!0}),f(a)!=="svelte-kvfsh7"&&(a.textContent=b),l=t(d),m(c.$$.fragment,d)},m(d,M){x(d,a,M),x(d,l,M),_(c,d,M),T=!0},p:D,i(d){T||(u(c.$$.fragment,d),T=!0)},o(d){h(c.$$.fragment,d),T=!1},d(d){d&&(i(a),i(l)),g(c,d)}}}function Xh(y){let a,b="We support loading A1111 formatted LoRA checkpoints in a limited capacity.",l,c,T="This function is experimental and might change in the future.";return{c(){a=n("p"),a.textContent=b,l=r(),c=n("p"),c.textContent=T},l(d){a=s(d,"P",{"data-svelte-h":!0}),f(a)!=="svelte-15l1sdn"&&(a.textContent=b),l=t(d),c=s(d,"P",{"data-svelte-h":!0}),f(c)!=="svelte-3fufvn"&&(c.textContent=T)},m(d,M){x(d,a,M),x(d,l,M),x(d,c,M)},p:D,d(d){d&&(i(a),i(l),i(c))}}}function qh(y){let a,b="This is an experimental API.";return{c(){a=n("p"),a.textContent=b},l(l){a=s(l,"P",{"data-svelte-h":!0}),f(a)!=="svelte-8w79b9"&&(a.textContent=b)},m(l,c){x(l,a,c)},p:D,d(l){l&&i(a)}}}function zh(y){let a,b,l,c,T,d,M,qm='LoRA is a fast and lightweight training method that inserts and trains a significantly smaller number of parameters instead of all the model parameters. This produces a smaller file (~100 MBs) and makes it easier to quickly train a model to learn a new concept. LoRA weights are typically loaded into the denoiser, text encoder or both. The denoiser usually corresponds to a UNet (<a href="/docs/diffusers/pr_11477/en/api/models/unet2d-cond#diffusers.UNet2DConditionModel">UNet2DConditionModel</a>, for example) or a Transformer (<a href="/docs/diffusers/pr_11477/en/api/models/sd3_transformer2d#diffusers.SD3Transformer2DModel">SD3Transformer2DModel</a>, for example). There are several classes for loading LoRA weights:',Qi,Wr,zm='<li><code>StableDiffusionLoraLoaderMixin</code> provides functions for loading and unloading, fusing and unfusing, enabling and disabling, and more functions for managing LoRA weights. This class can be used with any model.</li> <li><code>StableDiffusionXLLoraLoaderMixin</code> is a <a href="../../api/pipelines/stable_diffusion/stable_diffusion_xl">Stable Diffusion (SDXL)</a> version of the <code>StableDiffusionLoraLoaderMixin</code> class for loading and saving LoRA weights. It can only be used with the SDXL model.</li> <li><code>SD3LoraLoaderMixin</code> provides similar functions for <a href="https://huggingface.co/blog/sd3" rel="nofollow">Stable Diffusion 3</a>.</li> <li><code>FluxLoraLoaderMixin</code> provides similar functions for <a href="https://huggingface.co/docs/diffusers/main/en/api/pipelines/flux" rel="nofollow">Flux</a>.</li> <li><code>CogVideoXLoraLoaderMixin</code> provides similar functions for <a href="https://huggingface.co/docs/diffusers/main/en/api/pipelines/cogvideox" rel="nofollow">CogVideoX</a>.</li> <li><code>Mochi1LoraLoaderMixin</code> provides similar functions for <a href="https://huggingface.co/docs/diffusers/main/en/api/pipelines/mochi" rel="nofollow">Mochi</a>.</li> <li><code>AuraFlowLoraLoaderMixin</code> provides similar functions for <a href="https://huggingface.co/fal/AuraFlow" rel="nofollow">AuraFlow</a>.</li> <li><code>LTXVideoLoraLoaderMixin</code> provides similar functions for <a href="https://huggingface.co/docs/diffusers/main/en/api/pipelines/ltx_video" rel="nofollow">LTX-Video</a>.</li> <li><code>SanaLoraLoaderMixin</code> provides similar functions for <a href="https://huggingface.co/docs/diffusers/main/en/api/pipelines/sana" rel="nofollow">Sana</a>.</li> <li><code>HunyuanVideoLoraLoaderMixin</code> provides similar functions for <a href="https://huggingface.co/docs/diffusers/main/en/api/pipelines/hunyuan_video" rel="nofollow">HunyuanVideo</a>.</li> <li><code>Lumina2LoraLoaderMixin</code> provides similar functions for <a href="https://huggingface.co/docs/diffusers/main/en/api/pipelines/lumina2" rel="nofollow">Lumina2</a>.</li> <li><code>WanLoraLoaderMixin</code> provides similar functions for <a href="https://huggingface.co/docs/diffusers/main/en/api/pipelines/wan" rel="nofollow">Wan</a>.</li> <li><code>CogView4LoraLoaderMixin</code> provides similar functions for <a href="https://huggingface.co/docs/diffusers/main/en/api/pipelines/cogview4" rel="nofollow">CogView4</a>.</li> <li><code>AmusedLoraLoaderMixin</code> is for the <a href="/docs/diffusers/pr_11477/en/api/pipelines/amused#diffusers.AmusedPipeline">AmusedPipeline</a>.</li> <li><code>HiDreamImageLoraLoaderMixin</code> provides similar functions for <a href="https://huggingface.co/docs/diffusers/main/en/api/pipelines/hidream" rel="nofollow">HiDream Image</a></li> <li><code>LoraBaseMixin</code> provides a base class with several utility methods to fuse, unfuse, unload, LoRAs and more.</li>',Ki,so,ed,Fr,od,I,Ur,Ll,un,Bm=`Load LoRA layers into Stable Diffusion <a href="/docs/diffusers/pr_11477/en/api/models/unet2d-cond#diffusers.UNet2DConditionModel">UNet2DConditionModel</a> and | |
| <a href="https://huggingface.co/docs/transformers/model_doc/clip#transformers.CLIPTextModel" rel="nofollow"><code>CLIPTextModel</code></a>.`,xl,io,Er,bl,hn,jm="This will load the LoRA layers specified in <code>state_dict</code> into <code>text_encoder</code>",vl,lo,Nr,wl,gn,Gm="This will load the LoRA layers specified in <code>state_dict</code> into <code>unet</code>.",$l,K,Xr,Ml,Ln,Jm=`Load LoRA weights specified in <code>pretrained_model_name_or_path_or_dict</code> into <code>self.unet</code> and | |
| <code>self.text_encoder</code>.`,yl,xn,Zm="All kwargs are forwarded to <code>self.lora_state_dict</code>.",Tl,bn,Om=`See <a href="/docs/diffusers/pr_11477/en/api/loaders/lora#diffusers.loaders.StableDiffusionLoraLoaderMixin.lora_state_dict">lora_state_dict()</a> for more details on how the state dict is | |
| loaded.`,Dl,vn,Ym=`See <a href="/docs/diffusers/pr_11477/en/api/loaders/lora#diffusers.loaders.StableDiffusionLoraLoaderMixin.load_lora_into_unet">load_lora_into_unet()</a> for more details on how the state dict is | |
| loaded into <code>self.unet</code>.`,Sl,wn,Qm=`See <a href="/docs/diffusers/pr_11477/en/api/loaders/lora#diffusers.loaders.StableDiffusionLoraLoaderMixin.load_lora_into_text_encoder">load_lora_into_text_encoder()</a> for more details on how the state | |
| dict is loaded into <code>self.text_encoder</code>.`,Cl,ge,qr,Al,$n,Km="Return state dict for lora weights and the network alphas.",kl,co,Pl,fo,zr,Rl,Mn,e_="Save the LoRA parameters corresponding to the UNet and text encoder.",rd,Br,td,V,jr,Hl,yn,o_=`Load LoRA layers into Stable Diffusion XL <a href="/docs/diffusers/pr_11477/en/api/models/unet2d-cond#diffusers.UNet2DConditionModel">UNet2DConditionModel</a>, | |
| <a href="https://huggingface.co/docs/transformers/model_doc/clip#transformers.CLIPTextModel" rel="nofollow"><code>CLIPTextModel</code></a>, and | |
| <a href="https://huggingface.co/docs/transformers/model_doc/clip#transformers.CLIPTextModelWithProjection" rel="nofollow"><code>CLIPTextModelWithProjection</code></a>.`,Il,po,Gr,Vl,Tn,r_="This will load the LoRA layers specified in <code>state_dict</code> into <code>text_encoder</code>",Wl,mo,Jr,Fl,Dn,t_="This will load the LoRA layers specified in <code>state_dict</code> into <code>unet</code>.",Ul,ee,Zr,El,Sn,a_=`Load LoRA weights specified in <code>pretrained_model_name_or_path_or_dict</code> into <code>self.unet</code> and | |
| <code>self.text_encoder</code>.`,Nl,Cn,n_="All kwargs are forwarded to <code>self.lora_state_dict</code>.",Xl,An,s_=`See <a href="/docs/diffusers/pr_11477/en/api/loaders/lora#diffusers.loaders.StableDiffusionLoraLoaderMixin.lora_state_dict">lora_state_dict()</a> for more details on how the state dict is | |
| loaded.`,ql,kn,i_=`See <a href="/docs/diffusers/pr_11477/en/api/loaders/lora#diffusers.loaders.StableDiffusionLoraLoaderMixin.load_lora_into_unet">load_lora_into_unet()</a> for more details on how the state dict is | |
| loaded into <code>self.unet</code>.`,zl,Pn,d_=`See <a href="/docs/diffusers/pr_11477/en/api/loaders/lora#diffusers.loaders.StableDiffusionLoraLoaderMixin.load_lora_into_text_encoder">load_lora_into_text_encoder()</a> for more details on how the state | |
| dict is loaded into <code>self.text_encoder</code>.`,Bl,Le,Or,jl,Rn,l_="Return state dict for lora weights and the network alphas.",Gl,_o,Jl,uo,Yr,Zl,Hn,c_="Save the LoRA parameters corresponding to the UNet and text encoder.",ad,Qr,nd,k,Kr,Ol,In,f_=`Load LoRA layers into <a href="/docs/diffusers/pr_11477/en/api/models/sd3_transformer2d#diffusers.SD3Transformer2DModel">SD3Transformer2DModel</a>, | |
| <a href="https://huggingface.co/docs/transformers/model_doc/clip#transformers.CLIPTextModel" rel="nofollow"><code>CLIPTextModel</code></a>, and | |
| <a href="https://huggingface.co/docs/transformers/model_doc/clip#transformers.CLIPTextModelWithProjection" rel="nofollow"><code>CLIPTextModelWithProjection</code></a>.`,Yl,Vn,p_='Specific to <a href="/docs/diffusers/pr_11477/en/api/pipelines/stable_diffusion/stable_diffusion_3#diffusers.StableDiffusion3Pipeline">StableDiffusion3Pipeline</a>.',Ql,ho,et,Kl,Wn,m_="This will load the LoRA layers specified in <code>state_dict</code> into <code>text_encoder</code>",ec,go,ot,oc,Fn,__="This will load the LoRA layers specified in <code>state_dict</code> into <code>transformer</code>.",rc,te,rt,tc,Un,u_=`Load LoRA weights specified in <code>pretrained_model_name_or_path_or_dict</code> into <code>self.unet</code> and | |
| <code>self.text_encoder</code>.`,ac,En,h_="All kwargs are forwarded to <code>self.lora_state_dict</code>.",nc,Nn,g_=`See <a href="/docs/diffusers/pr_11477/en/api/loaders/lora#diffusers.loaders.StableDiffusionLoraLoaderMixin.lora_state_dict">lora_state_dict()</a> for more details on how the state dict is | |
| loaded.`,sc,Xn,L_=`See <code>~loaders.StableDiffusionLoraLoaderMixin.load_lora_into_transformer</code> for more details on how the state | |
| dict is loaded into <code>self.transformer</code>.`,ic,xe,tt,dc,qn,x_="Return state dict for lora weights and the network alphas.",lc,Lo,cc,xo,at,fc,zn,b_="Save the LoRA parameters corresponding to the UNet and text encoder.",pc,be,nt,mc,Bn,v_=`Reverses the effect of | |
| <a href="https://huggingface.co/docs/diffusers/main/en/api/loaders#diffusers.loaders.LoraBaseMixin.fuse_lora" rel="nofollow"><code>pipe.fuse_lora()</code></a>.`,_c,bo,sd,st,id,C,it,uc,jn,w_=`Load LoRA layers into <a href="/docs/diffusers/pr_11477/en/api/models/flux_transformer#diffusers.FluxTransformer2DModel">FluxTransformer2DModel</a>, | |
| <a href="https://huggingface.co/docs/transformers/model_doc/clip#transformers.CLIPTextModel" rel="nofollow"><code>CLIPTextModel</code></a>.`,hc,Gn,$_='Specific to <a href="/docs/diffusers/pr_11477/en/api/pipelines/stable_diffusion/stable_diffusion_3#diffusers.StableDiffusion3Pipeline">StableDiffusion3Pipeline</a>.',gc,vo,dt,Lc,Jn,M_="This will load the LoRA layers specified in <code>state_dict</code> into <code>text_encoder</code>",xc,wo,lt,bc,Zn,y_="This will load the LoRA layers specified in <code>state_dict</code> into <code>transformer</code>.",vc,ae,ct,wc,On,T_=`Load LoRA weights specified in <code>pretrained_model_name_or_path_or_dict</code> into <code>self.transformer</code> and | |
| <code>self.text_encoder</code>.`,$c,Yn,D_="All kwargs are forwarded to <code>self.lora_state_dict</code>.",Mc,Qn,S_=`See <a href="/docs/diffusers/pr_11477/en/api/loaders/lora#diffusers.loaders.StableDiffusionLoraLoaderMixin.lora_state_dict">lora_state_dict()</a> for more details on how the state dict is | |
| loaded.`,yc,Kn,C_=`See <code>~loaders.StableDiffusionLoraLoaderMixin.load_lora_into_transformer</code> for more details on how the state | |
| dict is loaded into <code>self.transformer</code>.`,Tc,ve,ft,Dc,es,A_="Return state dict for lora weights and the network alphas.",Sc,$o,Cc,Mo,pt,Ac,os,k_="Save the LoRA parameters corresponding to the UNet and text encoder.",kc,we,mt,Pc,rs,P_=`Reverses the effect of | |
| <a href="https://huggingface.co/docs/diffusers/main/en/api/loaders#diffusers.loaders.LoraBaseMixin.fuse_lora" rel="nofollow"><code>pipe.fuse_lora()</code></a>.`,Rc,yo,Hc,$e,_t,Ic,ts,R_="Unloads the LoRA parameters.",Vc,To,dd,ut,ld,W,ht,Wc,as,H_='Load LoRA layers into <a href="/docs/diffusers/pr_11477/en/api/models/cogvideox_transformer3d#diffusers.CogVideoXTransformer3DModel">CogVideoXTransformer3DModel</a>. Specific to <a href="/docs/diffusers/pr_11477/en/api/pipelines/cogvideox#diffusers.CogVideoXPipeline">CogVideoXPipeline</a>.',Fc,Do,gt,Uc,ns,I_="This will load the LoRA layers specified in <code>state_dict</code> into <code>transformer</code>.",Ec,So,Lt,Nc,ss,V_=`Load LoRA weights specified in <code>pretrained_model_name_or_path_or_dict</code> into <code>self.transformer</code> and | |
| <code>self.text_encoder</code>. All kwargs are forwarded to <code>self.lora_state_dict</code>. See | |
| <a href="/docs/diffusers/pr_11477/en/api/loaders/lora#diffusers.loaders.StableDiffusionLoraLoaderMixin.lora_state_dict">lora_state_dict()</a> for more details on how the state dict is loaded. | |
| See <code>~loaders.StableDiffusionLoraLoaderMixin.load_lora_into_transformer</code> for more details on how the state | |
| dict is loaded into <code>self.transformer</code>.`,Xc,Me,xt,qc,is,W_="Return state dict for lora weights and the network alphas.",zc,Co,Bc,Ao,bt,jc,ds,F_="Save the LoRA parameters corresponding to the UNet and text encoder.",Gc,ye,vt,Jc,ls,U_=`Reverses the effect of | |
| <a href="https://huggingface.co/docs/diffusers/main/en/api/loaders#diffusers.loaders.LoraBaseMixin.fuse_lora" rel="nofollow"><code>pipe.fuse_lora()</code></a>.`,Zc,ko,cd,wt,fd,F,$t,Oc,cs,E_='Load LoRA layers into <a href="/docs/diffusers/pr_11477/en/api/models/mochi_transformer3d#diffusers.MochiTransformer3DModel">MochiTransformer3DModel</a>. Specific to <a href="/docs/diffusers/pr_11477/en/api/pipelines/mochi#diffusers.MochiPipeline">MochiPipeline</a>.',Yc,Po,Mt,Qc,fs,N_="This will load the LoRA layers specified in <code>state_dict</code> into <code>transformer</code>.",Kc,Ro,yt,ef,ps,X_=`Load LoRA weights specified in <code>pretrained_model_name_or_path_or_dict</code> into <code>self.transformer</code> and | |
| <code>self.text_encoder</code>. All kwargs are forwarded to <code>self.lora_state_dict</code>. See | |
| <a href="/docs/diffusers/pr_11477/en/api/loaders/lora#diffusers.loaders.StableDiffusionLoraLoaderMixin.lora_state_dict">lora_state_dict()</a> for more details on how the state dict is loaded. | |
| See <code>~loaders.StableDiffusionLoraLoaderMixin.load_lora_into_transformer</code> for more details on how the state | |
| dict is loaded into <code>self.transformer</code>.`,of,Te,Tt,rf,ms,q_="Return state dict for lora weights and the network alphas.",tf,Ho,af,Io,Dt,nf,_s,z_="Save the LoRA parameters corresponding to the UNet and text encoder.",sf,De,St,df,us,B_=`Reverses the effect of | |
| <a href="https://huggingface.co/docs/diffusers/main/en/api/loaders#diffusers.loaders.LoraBaseMixin.fuse_lora" rel="nofollow"><code>pipe.fuse_lora()</code></a>.`,lf,Vo,pd,Ct,md,U,At,cf,hs,j_='Load LoRA layers into <a href="/docs/diffusers/pr_11477/en/api/models/aura_flow_transformer2d#diffusers.AuraFlowTransformer2DModel">AuraFlowTransformer2DModel</a> Specific to <a href="/docs/diffusers/pr_11477/en/api/pipelines/aura_flow#diffusers.AuraFlowPipeline">AuraFlowPipeline</a>.',ff,Wo,kt,pf,gs,G_="This will load the LoRA layers specified in <code>state_dict</code> into <code>transformer</code>.",mf,Fo,Pt,_f,Ls,J_=`Load LoRA weights specified in <code>pretrained_model_name_or_path_or_dict</code> into <code>self.transformer</code> and | |
| <code>self.text_encoder</code>. All kwargs are forwarded to <code>self.lora_state_dict</code>. See | |
| <a href="/docs/diffusers/pr_11477/en/api/loaders/lora#diffusers.loaders.StableDiffusionLoraLoaderMixin.lora_state_dict">lora_state_dict()</a> for more details on how the state dict is loaded. | |
| See <code>~loaders.StableDiffusionLoraLoaderMixin.load_lora_into_transformer</code> for more details on how the state | |
| dict is loaded into <code>self.transformer</code>.`,uf,Se,Rt,hf,xs,Z_="Return state dict for lora weights and the network alphas.",gf,Uo,Lf,Eo,Ht,xf,bs,O_="Save the LoRA parameters corresponding to the UNet and text encoder.",bf,Ce,It,vf,vs,Y_=`Reverses the effect of | |
| <a href="https://huggingface.co/docs/diffusers/main/en/api/loaders#diffusers.loaders.LoraBaseMixin.fuse_lora" rel="nofollow"><code>pipe.fuse_lora()</code></a>.`,wf,No,_d,Vt,ud,E,Wt,$f,ws,Q_='Load LoRA layers into <a href="/docs/diffusers/pr_11477/en/api/models/ltx_video_transformer3d#diffusers.LTXVideoTransformer3DModel">LTXVideoTransformer3DModel</a>. Specific to <a href="/docs/diffusers/pr_11477/en/api/pipelines/ltx_video#diffusers.LTXPipeline">LTXPipeline</a>.',Mf,Xo,Ft,yf,$s,K_="This will load the LoRA layers specified in <code>state_dict</code> into <code>transformer</code>.",Tf,qo,Ut,Df,Ms,eu=`Load LoRA weights specified in <code>pretrained_model_name_or_path_or_dict</code> into <code>self.transformer</code> and | |
| <code>self.text_encoder</code>. All kwargs are forwarded to <code>self.lora_state_dict</code>. See | |
| <a href="/docs/diffusers/pr_11477/en/api/loaders/lora#diffusers.loaders.StableDiffusionLoraLoaderMixin.lora_state_dict">lora_state_dict()</a> for more details on how the state dict is loaded. | |
| See <code>~loaders.StableDiffusionLoraLoaderMixin.load_lora_into_transformer</code> for more details on how the state | |
| dict is loaded into <code>self.transformer</code>.`,Sf,Ae,Et,Cf,ys,ou="Return state dict for lora weights and the network alphas.",Af,zo,kf,Bo,Nt,Pf,Ts,ru="Save the LoRA parameters corresponding to the UNet and text encoder.",Rf,ke,Xt,Hf,Ds,tu=`Reverses the effect of | |
| <a href="https://huggingface.co/docs/diffusers/main/en/api/loaders#diffusers.loaders.LoraBaseMixin.fuse_lora" rel="nofollow"><code>pipe.fuse_lora()</code></a>.`,If,jo,hd,qt,gd,N,zt,Vf,Ss,au='Load LoRA layers into <a href="/docs/diffusers/pr_11477/en/api/models/sana_transformer2d#diffusers.SanaTransformer2DModel">SanaTransformer2DModel</a>. Specific to <a href="/docs/diffusers/pr_11477/en/api/pipelines/sana#diffusers.SanaPipeline">SanaPipeline</a>.',Wf,Go,Bt,Ff,Cs,nu="This will load the LoRA layers specified in <code>state_dict</code> into <code>transformer</code>.",Uf,Jo,jt,Ef,As,su=`Load LoRA weights specified in <code>pretrained_model_name_or_path_or_dict</code> into <code>self.transformer</code> and | |
| <code>self.text_encoder</code>. All kwargs are forwarded to <code>self.lora_state_dict</code>. See | |
| <a href="/docs/diffusers/pr_11477/en/api/loaders/lora#diffusers.loaders.StableDiffusionLoraLoaderMixin.lora_state_dict">lora_state_dict()</a> for more details on how the state dict is loaded. | |
| See <code>~loaders.StableDiffusionLoraLoaderMixin.load_lora_into_transformer</code> for more details on how the state | |
| dict is loaded into <code>self.transformer</code>.`,Nf,Pe,Gt,Xf,ks,iu="Return state dict for lora weights and the network alphas.",qf,Zo,zf,Oo,Jt,Bf,Ps,du="Save the LoRA parameters corresponding to the UNet and text encoder.",jf,Re,Zt,Gf,Rs,lu=`Reverses the effect of | |
| <a href="https://huggingface.co/docs/diffusers/main/en/api/loaders#diffusers.loaders.LoraBaseMixin.fuse_lora" rel="nofollow"><code>pipe.fuse_lora()</code></a>.`,Jf,Yo,Ld,Ot,xd,X,Yt,Zf,Hs,cu='Load LoRA layers into <a href="/docs/diffusers/pr_11477/en/api/models/hunyuan_video_transformer_3d#diffusers.HunyuanVideoTransformer3DModel">HunyuanVideoTransformer3DModel</a>. Specific to <a href="/docs/diffusers/pr_11477/en/api/pipelines/hunyuan_video#diffusers.HunyuanVideoPipeline">HunyuanVideoPipeline</a>.',Of,Qo,Qt,Yf,Is,fu="This will load the LoRA layers specified in <code>state_dict</code> into <code>transformer</code>.",Qf,Ko,Kt,Kf,Vs,pu=`Load LoRA weights specified in <code>pretrained_model_name_or_path_or_dict</code> into <code>self.transformer</code> and | |
| <code>self.text_encoder</code>. All kwargs are forwarded to <code>self.lora_state_dict</code>. See | |
| <a href="/docs/diffusers/pr_11477/en/api/loaders/lora#diffusers.loaders.StableDiffusionLoraLoaderMixin.lora_state_dict">lora_state_dict()</a> for more details on how the state dict is loaded. | |
| See <code>~loaders.StableDiffusionLoraLoaderMixin.load_lora_into_transformer</code> for more details on how the state | |
| dict is loaded into <code>self.transformer</code>.`,ep,He,ea,op,Ws,mu="Return state dict for lora weights and the network alphas.",rp,er,tp,or,oa,ap,Fs,_u="Save the LoRA parameters corresponding to the UNet and text encoder.",np,Ie,ra,sp,Us,uu=`Reverses the effect of | |
| <a href="https://huggingface.co/docs/diffusers/main/en/api/loaders#diffusers.loaders.LoraBaseMixin.fuse_lora" rel="nofollow"><code>pipe.fuse_lora()</code></a>.`,ip,rr,bd,ta,vd,q,aa,dp,Es,hu='Load LoRA layers into <a href="/docs/diffusers/pr_11477/en/api/models/lumina2_transformer2d#diffusers.Lumina2Transformer2DModel">Lumina2Transformer2DModel</a>. Specific to <code>Lumina2Text2ImgPipeline</code>.',lp,tr,na,cp,Ns,gu="This will load the LoRA layers specified in <code>state_dict</code> into <code>transformer</code>.",fp,ar,sa,pp,Xs,Lu=`Load LoRA weights specified in <code>pretrained_model_name_or_path_or_dict</code> into <code>self.transformer</code> and | |
| <code>self.text_encoder</code>. All kwargs are forwarded to <code>self.lora_state_dict</code>. See | |
| <a href="/docs/diffusers/pr_11477/en/api/loaders/lora#diffusers.loaders.StableDiffusionLoraLoaderMixin.lora_state_dict">lora_state_dict()</a> for more details on how the state dict is loaded. | |
| See <code>~loaders.StableDiffusionLoraLoaderMixin.load_lora_into_transformer</code> for more details on how the state | |
| dict is loaded into <code>self.transformer</code>.`,mp,Ve,ia,_p,qs,xu="Return state dict for lora weights and the network alphas.",up,nr,hp,sr,da,gp,zs,bu="Save the LoRA parameters corresponding to the UNet and text encoder.",Lp,We,la,xp,Bs,vu=`Reverses the effect of | |
| <a href="https://huggingface.co/docs/diffusers/main/en/api/loaders#diffusers.loaders.LoraBaseMixin.fuse_lora" rel="nofollow"><code>pipe.fuse_lora()</code></a>.`,bp,ir,wd,ca,$d,z,fa,vp,js,wu='Load LoRA layers into <a href="/docs/diffusers/pr_11477/en/api/models/wan_transformer_3d#diffusers.WanTransformer3DModel">WanTransformer3DModel</a>. Specific to <a href="/docs/diffusers/pr_11477/en/api/pipelines/cogview4#diffusers.CogView4Pipeline">CogView4Pipeline</a>.',wp,dr,pa,$p,Gs,$u="This will load the LoRA layers specified in <code>state_dict</code> into <code>transformer</code>.",Mp,lr,ma,yp,Js,Mu=`Load LoRA weights specified in <code>pretrained_model_name_or_path_or_dict</code> into <code>self.transformer</code> and | |
| <code>self.text_encoder</code>. All kwargs are forwarded to <code>self.lora_state_dict</code>. See | |
| <a href="/docs/diffusers/pr_11477/en/api/loaders/lora#diffusers.loaders.StableDiffusionLoraLoaderMixin.lora_state_dict">lora_state_dict()</a> for more details on how the state dict is loaded. | |
| See <code>~loaders.StableDiffusionLoraLoaderMixin.load_lora_into_transformer</code> for more details on how the state | |
| dict is loaded into <code>self.transformer</code>.`,Tp,Fe,_a,Dp,Zs,yu="Return state dict for lora weights and the network alphas.",Sp,cr,Cp,fr,ua,Ap,Os,Tu="Save the LoRA parameters corresponding to the UNet and text encoder.",kp,Ue,ha,Pp,Ys,Du=`Reverses the effect of | |
| <a href="https://huggingface.co/docs/diffusers/main/en/api/loaders#diffusers.loaders.LoraBaseMixin.fuse_lora" rel="nofollow"><code>pipe.fuse_lora()</code></a>.`,Rp,pr,Md,ga,yd,B,La,Hp,Qs,Su='Load LoRA layers into <a href="/docs/diffusers/pr_11477/en/api/models/wan_transformer_3d#diffusers.WanTransformer3DModel">WanTransformer3DModel</a>. Specific to <a href="/docs/diffusers/pr_11477/en/api/pipelines/wan#diffusers.WanPipeline">WanPipeline</a> and <code>[WanImageToVideoPipeline</code>].',Ip,mr,xa,Vp,Ks,Cu="This will load the LoRA layers specified in <code>state_dict</code> into <code>transformer</code>.",Wp,_r,ba,Fp,ei,Au=`Load LoRA weights specified in <code>pretrained_model_name_or_path_or_dict</code> into <code>self.transformer</code> and | |
| <code>self.text_encoder</code>. All kwargs are forwarded to <code>self.lora_state_dict</code>. See | |
| <a href="/docs/diffusers/pr_11477/en/api/loaders/lora#diffusers.loaders.StableDiffusionLoraLoaderMixin.lora_state_dict">lora_state_dict()</a> for more details on how the state dict is loaded. | |
| See <code>~loaders.StableDiffusionLoraLoaderMixin.load_lora_into_transformer</code> for more details on how the state | |
| dict is loaded into <code>self.transformer</code>.`,Up,Ee,va,Ep,oi,ku="Return state dict for lora weights and the network alphas.",Np,ur,Xp,hr,wa,qp,ri,Pu="Save the LoRA parameters corresponding to the UNet and text encoder.",zp,Ne,$a,Bp,ti,Ru=`Reverses the effect of | |
| <a href="https://huggingface.co/docs/diffusers/main/en/api/loaders#diffusers.loaders.LoraBaseMixin.fuse_lora" rel="nofollow"><code>pipe.fuse_lora()</code></a>.`,jp,gr,Td,Ma,Dd,Oe,ya,Gp,Lr,Ta,Jp,ai,Hu="This will load the LoRA layers specified in <code>state_dict</code> into <code>transformer</code>.",Sd,Da,Cd,j,Sa,Zp,ni,Iu='Load LoRA layers into <a href="/docs/diffusers/pr_11477/en/api/models/hidream_image_transformer#diffusers.HiDreamImageTransformer2DModel">HiDreamImageTransformer2DModel</a>. Specific to <a href="/docs/diffusers/pr_11477/en/api/pipelines/hidream#diffusers.HiDreamImagePipeline">HiDreamImagePipeline</a>.',Op,xr,Ca,Yp,si,Vu="This will load the LoRA layers specified in <code>state_dict</code> into <code>transformer</code>.",Qp,br,Aa,Kp,ii,Wu=`Load LoRA weights specified in <code>pretrained_model_name_or_path_or_dict</code> into <code>self.transformer</code> and | |
| <code>self.text_encoder</code>. All kwargs are forwarded to <code>self.lora_state_dict</code>. See | |
| <a href="/docs/diffusers/pr_11477/en/api/loaders/lora#diffusers.loaders.StableDiffusionLoraLoaderMixin.lora_state_dict">lora_state_dict()</a> for more details on how the state dict is loaded. | |
| See <code>~loaders.StableDiffusionLoraLoaderMixin.load_lora_into_transformer</code> for more details on how the state | |
| dict is loaded into <code>self.transformer</code>.`,em,Xe,ka,om,di,Fu="Return state dict for lora weights and the network alphas.",rm,vr,tm,wr,Pa,am,li,Uu="Save the LoRA parameters corresponding to the UNet and text encoder.",nm,qe,Ra,sm,ci,Eu=`Reverses the effect of | |
| <a href="https://huggingface.co/docs/diffusers/main/en/api/loaders#diffusers.loaders.LoraBaseMixin.fuse_lora" rel="nofollow"><code>pipe.fuse_lora()</code></a>.`,im,$r,Ad,Ha,kd,A,Ia,dm,fi,Nu="Utility class for handling LoRAs.",lm,pi,Va,cm,ze,Wa,fm,mi,Xu="Enables the possibility to hotswap LoRA adapters.",pm,_i,qu=`Calling this method is only required when hotswapping adapters and if the model is compiled or if the ranks of | |
| the loaded adapters differ.`,mm,he,Fa,_m,ui,zu="Fuses the LoRA parameters into the original parameters of the corresponding blocks.",um,Mr,hm,yr,gm,Be,Ua,Lm,hi,Bu="Gets the list of the current active adapters.",xm,Tr,bm,Dr,Ea,vm,gi,ju="Gets the current list of all available adapters in the pipeline.",wm,Sr,Na,$m,Li,Gu=`Moves the LoRAs listed in <code>adapter_names</code> to a target device. Useful for offloading the LoRA to the CPU in case | |
| you want to load multiple adapters and free some GPU memory.`,Mm,je,Xa,ym,xi,Ju=`Reverses the effect of | |
| <a href="https://huggingface.co/docs/diffusers/main/en/api/loaders#diffusers.loaders.LoraBaseMixin.fuse_lora" rel="nofollow"><code>pipe.fuse_lora()</code></a>.`,Tm,Cr,Dm,Ge,qa,Sm,bi,Zu="Unloads the LoRA parameters.",Cm,Ar,Pd,za,Rd,G,Ba,Am,vi,Ou='Load LoRA layers into <a href="/docs/diffusers/pr_11477/en/api/models/wan_transformer_3d#diffusers.WanTransformer3DModel">WanTransformer3DModel</a>. Specific to <a href="/docs/diffusers/pr_11477/en/api/pipelines/wan#diffusers.WanPipeline">WanPipeline</a> and <code>[WanImageToVideoPipeline</code>].',km,kr,ja,Pm,wi,Yu="This will load the LoRA layers specified in <code>state_dict</code> into <code>transformer</code>.",Rm,Pr,Ga,Hm,$i,Qu=`Load LoRA weights specified in <code>pretrained_model_name_or_path_or_dict</code> into <code>self.transformer</code> and | |
| <code>self.text_encoder</code>. All kwargs are forwarded to <code>self.lora_state_dict</code>. See | |
| <a href="/docs/diffusers/pr_11477/en/api/loaders/lora#diffusers.loaders.StableDiffusionLoraLoaderMixin.lora_state_dict">lora_state_dict()</a> for more details on how the state dict is loaded. | |
| See <code>~loaders.StableDiffusionLoraLoaderMixin.load_lora_into_transformer</code> for more details on how the state | |
| dict is loaded into <code>self.transformer</code>.`,Im,Je,Ja,Vm,Mi,Ku="Return state dict for lora weights and the network alphas.",Wm,Rr,Fm,Hr,Za,Um,yi,eh="Save the LoRA parameters corresponding to the UNet and text encoder.",Em,Ze,Oa,Nm,Ti,oh=`Reverses the effect of | |
| <a href="https://huggingface.co/docs/diffusers/main/en/api/loaders#diffusers.loaders.LoraBaseMixin.fuse_lora" rel="nofollow"><code>pipe.fuse_lora()</code></a>.`,Xm,Ir,Hd,Ya,Id,Yi,Vd;return T=new J({props:{title:"LoRA",local:"lora",headingTag:"h1"}}),so=new S({props:{$$slots:{default:[lh]},$$scope:{ctx:y}}}),Fr=new J({props:{title:"StableDiffusionLoraLoaderMixin",local:"diffusers.loaders.StableDiffusionLoraLoaderMixin",headingTag:"h2"}}),Ur=new $({props:{name:"class diffusers.loaders.StableDiffusionLoraLoaderMixin",anchor:"diffusers.loaders.StableDiffusionLoraLoaderMixin",parameters:[],source:"https://github.com/huggingface/diffusers/blob/vr_11477/src/diffusers/loaders/lora_pipeline.py#L120"}}),Er=new $({props:{name:"load_lora_into_text_encoder",anchor:"diffusers.loaders.StableDiffusionLoraLoaderMixin.load_lora_into_text_encoder",parameters:[{name:"state_dict",val:""},{name:"network_alphas",val:""},{name:"text_encoder",val:""},{name:"prefix",val:" = None"},{name:"lora_scale",val:" = 1.0"},{name:"adapter_name",val:" = None"},{name:"_pipeline",val:" = None"},{name:"low_cpu_mem_usage",val:" = False"},{name:"hotswap",val:": bool = False"}],parametersDescription:[{anchor:"diffusers.loaders.StableDiffusionLoraLoaderMixin.load_lora_into_text_encoder.state_dict",description:`<strong>state_dict</strong> (<code>dict</code>) — | |
| A standard state dict containing the lora layer parameters. The key should be prefixed with an | |
| additional <code>text_encoder</code> to distinguish between unet lora layers.`,name:"state_dict"},{anchor:"diffusers.loaders.StableDiffusionLoraLoaderMixin.load_lora_into_text_encoder.network_alphas",description:`<strong>network_alphas</strong> (<code>Dict[str, float]</code>) — | |
| The value of the network alpha used for stable learning and preventing underflow. This value has the | |
| same meaning as the <code>--network_alpha</code> option in the kohya-ss trainer script. Refer to <a href="https://github.com/darkstorm2150/sd-scripts/blob/main/docs/train_network_README-en.md#execute-learning" rel="nofollow">this | |
| link</a>.`,name:"network_alphas"},{anchor:"diffusers.loaders.StableDiffusionLoraLoaderMixin.load_lora_into_text_encoder.text_encoder",description:`<strong>text_encoder</strong> (<code>CLIPTextModel</code>) — | |
| The text encoder model to load the LoRA layers into.`,name:"text_encoder"},{anchor:"diffusers.loaders.StableDiffusionLoraLoaderMixin.load_lora_into_text_encoder.prefix",description:`<strong>prefix</strong> (<code>str</code>) — | |
| Expected prefix of the <code>text_encoder</code> in the <code>state_dict</code>.`,name:"prefix"},{anchor:"diffusers.loaders.StableDiffusionLoraLoaderMixin.load_lora_into_text_encoder.lora_scale",description:`<strong>lora_scale</strong> (<code>float</code>) — | |
| How much to scale the output of the lora linear layer before it is added with the output of the regular | |
| lora layer.`,name:"lora_scale"},{anchor:"diffusers.loaders.StableDiffusionLoraLoaderMixin.load_lora_into_text_encoder.adapter_name",description:`<strong>adapter_name</strong> (<code>str</code>, <em>optional</em>) — | |
| Adapter name to be used for referencing the loaded adapter model. If not specified, it will use | |
| <code>default_{i}</code> where i is the total number of adapters being loaded.`,name:"adapter_name"},{anchor:"diffusers.loaders.StableDiffusionLoraLoaderMixin.load_lora_into_text_encoder.low_cpu_mem_usage",description:`<strong>low_cpu_mem_usage</strong> (<code>bool</code>, <em>optional</em>) — | |
| Speed up model loading by only loading the pretrained LoRA weights and not initializing the random | |
| weights.`,name:"low_cpu_mem_usage"},{anchor:"diffusers.loaders.StableDiffusionLoraLoaderMixin.load_lora_into_text_encoder.hotswap",description:`<strong>hotswap</strong> (<code>bool</code>, <em>optional</em>) — | |
| See <a href="/docs/diffusers/pr_11477/en/api/loaders/lora#diffusers.loaders.StableDiffusionLoraLoaderMixin.load_lora_weights">load_lora_weights()</a>.`,name:"hotswap"}],source:"https://github.com/huggingface/diffusers/blob/vr_11477/src/diffusers/loaders/lora_pipeline.py#L399"}}),Nr=new $({props:{name:"load_lora_into_unet",anchor:"diffusers.loaders.StableDiffusionLoraLoaderMixin.load_lora_into_unet",parameters:[{name:"state_dict",val:""},{name:"network_alphas",val:""},{name:"unet",val:""},{name:"adapter_name",val:" = None"},{name:"_pipeline",val:" = None"},{name:"low_cpu_mem_usage",val:" = False"},{name:"hotswap",val:": bool = False"}],parametersDescription:[{anchor:"diffusers.loaders.StableDiffusionLoraLoaderMixin.load_lora_into_unet.state_dict",description:`<strong>state_dict</strong> (<code>dict</code>) — | |
| A standard state dict containing the lora layer parameters. The keys can either be indexed directly | |
| into the unet or prefixed with an additional <code>unet</code> which can be used to distinguish between text | |
| encoder lora layers.`,name:"state_dict"},{anchor:"diffusers.loaders.StableDiffusionLoraLoaderMixin.load_lora_into_unet.network_alphas",description:`<strong>network_alphas</strong> (<code>Dict[str, float]</code>) — | |
| The value of the network alpha used for stable learning and preventing underflow. This value has the | |
| same meaning as the <code>--network_alpha</code> option in the kohya-ss trainer script. Refer to <a href="https://github.com/darkstorm2150/sd-scripts/blob/main/docs/train_network_README-en.md#execute-learning" rel="nofollow">this | |
| link</a>.`,name:"network_alphas"},{anchor:"diffusers.loaders.StableDiffusionLoraLoaderMixin.load_lora_into_unet.unet",description:`<strong>unet</strong> (<code>UNet2DConditionModel</code>) — | |
| The UNet model to load the LoRA layers into.`,name:"unet"},{anchor:"diffusers.loaders.StableDiffusionLoraLoaderMixin.load_lora_into_unet.adapter_name",description:`<strong>adapter_name</strong> (<code>str</code>, <em>optional</em>) — | |
| Adapter name to be used for referencing the loaded adapter model. If not specified, it will use | |
| <code>default_{i}</code> where i is the total number of adapters being loaded.`,name:"adapter_name"},{anchor:"diffusers.loaders.StableDiffusionLoraLoaderMixin.load_lora_into_unet.low_cpu_mem_usage",description:`<strong>low_cpu_mem_usage</strong> (<code>bool</code>, <em>optional</em>) — | |
| Speed up model loading only loading the pretrained LoRA weights and not initializing the random | |
| weights.`,name:"low_cpu_mem_usage"},{anchor:"diffusers.loaders.StableDiffusionLoraLoaderMixin.load_lora_into_unet.hotswap",description:`<strong>hotswap</strong> (<code>bool</code>, <em>optional</em>) — | |
| See <a href="/docs/diffusers/pr_11477/en/api/loaders/lora#diffusers.loaders.StableDiffusionLoraLoaderMixin.load_lora_weights">load_lora_weights()</a>.`,name:"hotswap"}],source:"https://github.com/huggingface/diffusers/blob/vr_11477/src/diffusers/loaders/lora_pipeline.py#L343"}}),Xr=new $({props:{name:"load_lora_weights",anchor:"diffusers.loaders.StableDiffusionLoraLoaderMixin.load_lora_weights",parameters:[{name:"pretrained_model_name_or_path_or_dict",val:": typing.Union[str, typing.Dict[str, torch.Tensor]]"},{name:"adapter_name",val:": typing.Optional[str] = None"},{name:"hotswap",val:": bool = False"},{name:"**kwargs",val:""}],parametersDescription:[{anchor:"diffusers.loaders.StableDiffusionLoraLoaderMixin.load_lora_weights.pretrained_model_name_or_path_or_dict",description:`<strong>pretrained_model_name_or_path_or_dict</strong> (<code>str</code> or <code>os.PathLike</code> or <code>dict</code>) — | |
| See <a href="/docs/diffusers/pr_11477/en/api/loaders/lora#diffusers.loaders.StableDiffusionLoraLoaderMixin.lora_state_dict">lora_state_dict()</a>.`,name:"pretrained_model_name_or_path_or_dict"},{anchor:"diffusers.loaders.StableDiffusionLoraLoaderMixin.load_lora_weights.adapter_name",description:`<strong>adapter_name</strong> (<code>str</code>, <em>optional</em>) — | |
| Adapter name to be used for referencing the loaded adapter model. If not specified, it will use | |
| <code>default_{i}</code> where i is the total number of adapters being loaded.`,name:"adapter_name"},{anchor:"diffusers.loaders.StableDiffusionLoraLoaderMixin.load_lora_weights.low_cpu_mem_usage",description:`<strong>low_cpu_mem_usage</strong> (<code>bool</code>, <em>optional</em>) — | |
| Speed up model loading by only loading the pretrained LoRA weights and not initializing the random | |
| weights.`,name:"low_cpu_mem_usage"},{anchor:"diffusers.loaders.StableDiffusionLoraLoaderMixin.load_lora_weights.hotswap",description:`<strong>hotswap</strong> (<code>bool</code>, <em>optional</em>) — | |
| Defaults to <code>False</code>. Whether to substitute an existing (LoRA) adapter with the newly loaded adapter | |
| in-place. This means that, instead of loading an additional adapter, this will take the existing | |
| adapter weights and replace them with the weights of the new adapter. This can be faster and more | |
| memory efficient. However, the main advantage of hotswapping is that when the model is compiled with | |
| torch.compile, loading the new adapter does not require recompilation of the model. When using | |
| hotswapping, the passed <code>adapter_name</code> should be the name of an already loaded adapter.</p> | |
| <p>If the new adapter and the old adapter have different ranks and/or LoRA alphas (i.e. scaling), you need | |
| to call an additional method before loading the adapter:`,name:"hotswap"}],source:"https://github.com/huggingface/diffusers/blob/vr_11477/src/diffusers/loaders/lora_pipeline.py#L130"}}),qr=new $({props:{name:"lora_state_dict",anchor:"diffusers.loaders.StableDiffusionLoraLoaderMixin.lora_state_dict",parameters:[{name:"pretrained_model_name_or_path_or_dict",val:": typing.Union[str, typing.Dict[str, torch.Tensor]]"},{name:"**kwargs",val:""}],parametersDescription:[{anchor:"diffusers.loaders.StableDiffusionLoraLoaderMixin.lora_state_dict.pretrained_model_name_or_path_or_dict",description:`<strong>pretrained_model_name_or_path_or_dict</strong> (<code>str</code> or <code>os.PathLike</code> or <code>dict</code>) — | |
| Can be either:</p> | |
| <ul> | |
| <li>A string, the <em>model id</em> (for example <code>google/ddpm-celebahq-256</code>) of a pretrained model hosted on | |
| the Hub.</li> | |
| <li>A path to a <em>directory</em> (for example <code>./my_model_directory</code>) containing the model weights saved | |
| with <a href="/docs/diffusers/pr_11477/en/api/models/overview#diffusers.ModelMixin.save_pretrained">ModelMixin.save_pretrained()</a>.</li> | |
| <li>A <a href="https://pytorch.org/tutorials/beginner/saving_loading_models.html#what-is-a-state-dict" rel="nofollow">torch state | |
| dict</a>.</li> | |
| </ul>`,name:"pretrained_model_name_or_path_or_dict"},{anchor:"diffusers.loaders.StableDiffusionLoraLoaderMixin.lora_state_dict.cache_dir",description:`<strong>cache_dir</strong> (<code>Union[str, os.PathLike]</code>, <em>optional</em>) — | |
| Path to a directory where a downloaded pretrained model configuration is cached if the standard cache | |
| is not used.`,name:"cache_dir"},{anchor:"diffusers.loaders.StableDiffusionLoraLoaderMixin.lora_state_dict.force_download",description:`<strong>force_download</strong> (<code>bool</code>, <em>optional</em>, defaults to <code>False</code>) — | |
| Whether or not to force the (re-)download of the model weights and configuration files, overriding the | |
| cached versions if they exist.`,name:"force_download"},{anchor:"diffusers.loaders.StableDiffusionLoraLoaderMixin.lora_state_dict.proxies",description:`<strong>proxies</strong> (<code>Dict[str, str]</code>, <em>optional</em>) — | |
| A dictionary of proxy servers to use by protocol or endpoint, for example, <code>{'http': 'foo.bar:3128', 'http://hostname': 'foo.bar:4012'}</code>. The proxies are used on each request.`,name:"proxies"},{anchor:"diffusers.loaders.StableDiffusionLoraLoaderMixin.lora_state_dict.local_files_only",description:`<strong>local_files_only</strong> (<code>bool</code>, <em>optional</em>, defaults to <code>False</code>) — | |
| Whether to only load local model weights and configuration files or not. If set to <code>True</code>, the model | |
| won’t be downloaded from the Hub.`,name:"local_files_only"},{anchor:"diffusers.loaders.StableDiffusionLoraLoaderMixin.lora_state_dict.token",description:`<strong>token</strong> (<code>str</code> or <em>bool</em>, <em>optional</em>) — | |
| The token to use as HTTP bearer authorization for remote files. If <code>True</code>, the token generated from | |
| <code>diffusers-cli login</code> (stored in <code>~/.huggingface</code>) is used.`,name:"token"},{anchor:"diffusers.loaders.StableDiffusionLoraLoaderMixin.lora_state_dict.revision",description:`<strong>revision</strong> (<code>str</code>, <em>optional</em>, defaults to <code>"main"</code>) — | |
| The specific model version to use. It can be a branch name, a tag name, a commit id, or any identifier | |
| allowed by Git.`,name:"revision"},{anchor:"diffusers.loaders.StableDiffusionLoraLoaderMixin.lora_state_dict.subfolder",description:`<strong>subfolder</strong> (<code>str</code>, <em>optional</em>, defaults to <code>""</code>) — | |
| The subfolder location of a model file within a larger model repository on the Hub or locally.`,name:"subfolder"},{anchor:"diffusers.loaders.StableDiffusionLoraLoaderMixin.lora_state_dict.weight_name",description:`<strong>weight_name</strong> (<code>str</code>, <em>optional</em>, defaults to None) — | |
| Name of the serialized state dict file.`,name:"weight_name"}],source:"https://github.com/huggingface/diffusers/blob/vr_11477/src/diffusers/loaders/lora_pipeline.py#L228"}}),co=new S({props:{warning:!0,$$slots:{default:[ch]},$$scope:{ctx:y}}}),zr=new $({props:{name:"save_lora_weights",anchor:"diffusers.loaders.StableDiffusionLoraLoaderMixin.save_lora_weights",parameters:[{name:"save_directory",val:": typing.Union[str, os.PathLike]"},{name:"unet_lora_layers",val:": typing.Dict[str, typing.Union[torch.nn.modules.module.Module, torch.Tensor]] = None"},{name:"text_encoder_lora_layers",val:": typing.Dict[str, torch.nn.modules.module.Module] = None"},{name:"is_main_process",val:": bool = True"},{name:"weight_name",val:": str = None"},{name:"save_function",val:": typing.Callable = None"},{name:"safe_serialization",val:": bool = True"}],parametersDescription:[{anchor:"diffusers.loaders.StableDiffusionLoraLoaderMixin.save_lora_weights.save_directory",description:`<strong>save_directory</strong> (<code>str</code> or <code>os.PathLike</code>) — | |
| Directory to save LoRA parameters to. Will be created if it doesn’t exist.`,name:"save_directory"},{anchor:"diffusers.loaders.StableDiffusionLoraLoaderMixin.save_lora_weights.unet_lora_layers",description:`<strong>unet_lora_layers</strong> (<code>Dict[str, torch.nn.Module]</code> or <code>Dict[str, torch.Tensor]</code>) — | |
| State dict of the LoRA layers corresponding to the <code>unet</code>.`,name:"unet_lora_layers"},{anchor:"diffusers.loaders.StableDiffusionLoraLoaderMixin.save_lora_weights.text_encoder_lora_layers",description:`<strong>text_encoder_lora_layers</strong> (<code>Dict[str, torch.nn.Module]</code> or <code>Dict[str, torch.Tensor]</code>) — | |
| State dict of the LoRA layers corresponding to the <code>text_encoder</code>. Must explicitly pass the text | |
| encoder LoRA state dict because it comes from 🤗 Transformers.`,name:"text_encoder_lora_layers"},{anchor:"diffusers.loaders.StableDiffusionLoraLoaderMixin.save_lora_weights.is_main_process",description:`<strong>is_main_process</strong> (<code>bool</code>, <em>optional</em>, defaults to <code>True</code>) — | |
| Whether the process calling this is the main process or not. Useful during distributed training and you | |
| need to call this function on all processes. In this case, set <code>is_main_process=True</code> only on the main | |
| process to avoid race conditions.`,name:"is_main_process"},{anchor:"diffusers.loaders.StableDiffusionLoraLoaderMixin.save_lora_weights.save_function",description:`<strong>save_function</strong> (<code>Callable</code>) — | |
| The function to use to save the state dictionary. Useful during distributed training when you need to | |
| replace <code>torch.save</code> with another method. Can be configured with the environment variable | |
| <code>DIFFUSERS_SAVE_MODE</code>.`,name:"save_function"},{anchor:"diffusers.loaders.StableDiffusionLoraLoaderMixin.save_lora_weights.safe_serialization",description:`<strong>safe_serialization</strong> (<code>bool</code>, <em>optional</em>, defaults to <code>True</code>) — | |
| Whether to save the model using <code>safetensors</code> or the traditional PyTorch way with <code>pickle</code>.`,name:"safe_serialization"}],source:"https://github.com/huggingface/diffusers/blob/vr_11477/src/diffusers/loaders/lora_pipeline.py#L452"}}),Br=new J({props:{title:"StableDiffusionXLLoraLoaderMixin",local:"diffusers.loaders.StableDiffusionXLLoraLoaderMixin",headingTag:"h2"}}),jr=new $({props:{name:"class diffusers.loaders.StableDiffusionXLLoraLoaderMixin",anchor:"diffusers.loaders.StableDiffusionXLLoraLoaderMixin",parameters:[],source:"https://github.com/huggingface/diffusers/blob/vr_11477/src/diffusers/loaders/lora_pipeline.py#L574"}}),Gr=new $({props:{name:"load_lora_into_text_encoder",anchor:"diffusers.loaders.StableDiffusionXLLoraLoaderMixin.load_lora_into_text_encoder",parameters:[{name:"state_dict",val:""},{name:"network_alphas",val:""},{name:"text_encoder",val:""},{name:"prefix",val:" = None"},{name:"lora_scale",val:" = 1.0"},{name:"adapter_name",val:" = None"},{name:"_pipeline",val:" = None"},{name:"low_cpu_mem_usage",val:" = False"},{name:"hotswap",val:": bool = False"}],parametersDescription:[{anchor:"diffusers.loaders.StableDiffusionXLLoraLoaderMixin.load_lora_into_text_encoder.state_dict",description:`<strong>state_dict</strong> (<code>dict</code>) — | |
| A standard state dict containing the lora layer parameters. The key should be prefixed with an | |
| additional <code>text_encoder</code> to distinguish between unet lora layers.`,name:"state_dict"},{anchor:"diffusers.loaders.StableDiffusionXLLoraLoaderMixin.load_lora_into_text_encoder.network_alphas",description:`<strong>network_alphas</strong> (<code>Dict[str, float]</code>) — | |
| The value of the network alpha used for stable learning and preventing underflow. This value has the | |
| same meaning as the <code>--network_alpha</code> option in the kohya-ss trainer script. Refer to <a href="https://github.com/darkstorm2150/sd-scripts/blob/main/docs/train_network_README-en.md#execute-learning" rel="nofollow">this | |
| link</a>.`,name:"network_alphas"},{anchor:"diffusers.loaders.StableDiffusionXLLoraLoaderMixin.load_lora_into_text_encoder.text_encoder",description:`<strong>text_encoder</strong> (<code>CLIPTextModel</code>) — | |
| The text encoder model to load the LoRA layers into.`,name:"text_encoder"},{anchor:"diffusers.loaders.StableDiffusionXLLoraLoaderMixin.load_lora_into_text_encoder.prefix",description:`<strong>prefix</strong> (<code>str</code>) — | |
| Expected prefix of the <code>text_encoder</code> in the <code>state_dict</code>.`,name:"prefix"},{anchor:"diffusers.loaders.StableDiffusionXLLoraLoaderMixin.load_lora_into_text_encoder.lora_scale",description:`<strong>lora_scale</strong> (<code>float</code>) — | |
| How much to scale the output of the lora linear layer before it is added with the output of the regular | |
| lora layer.`,name:"lora_scale"},{anchor:"diffusers.loaders.StableDiffusionXLLoraLoaderMixin.load_lora_into_text_encoder.adapter_name",description:`<strong>adapter_name</strong> (<code>str</code>, <em>optional</em>) — | |
| Adapter name to be used for referencing the loaded adapter model. If not specified, it will use | |
| <code>default_{i}</code> where i is the total number of adapters being loaded.`,name:"adapter_name"},{anchor:"diffusers.loaders.StableDiffusionXLLoraLoaderMixin.load_lora_into_text_encoder.low_cpu_mem_usage",description:`<strong>low_cpu_mem_usage</strong> (<code>bool</code>, <em>optional</em>) — | |
| Speed up model loading by only loading the pretrained LoRA weights and not initializing the random | |
| weights.`,name:"low_cpu_mem_usage"},{anchor:"diffusers.loaders.StableDiffusionXLLoraLoaderMixin.load_lora_into_text_encoder.hotswap",description:`<strong>hotswap</strong> (<code>bool</code>, <em>optional</em>) — | |
| See <a href="/docs/diffusers/pr_11477/en/api/loaders/lora#diffusers.loaders.StableDiffusionLoraLoaderMixin.load_lora_weights">load_lora_weights()</a>.`,name:"hotswap"}],source:"https://github.com/huggingface/diffusers/blob/vr_11477/src/diffusers/loaders/lora_pipeline.py#L854"}}),Jr=new $({props:{name:"load_lora_into_unet",anchor:"diffusers.loaders.StableDiffusionXLLoraLoaderMixin.load_lora_into_unet",parameters:[{name:"state_dict",val:""},{name:"network_alphas",val:""},{name:"unet",val:""},{name:"adapter_name",val:" = None"},{name:"_pipeline",val:" = None"},{name:"low_cpu_mem_usage",val:" = False"},{name:"hotswap",val:": bool = False"}],parametersDescription:[{anchor:"diffusers.loaders.StableDiffusionXLLoraLoaderMixin.load_lora_into_unet.state_dict",description:`<strong>state_dict</strong> (<code>dict</code>) — | |
| A standard state dict containing the lora layer parameters. The keys can either be indexed directly | |
| into the unet or prefixed with an additional <code>unet</code> which can be used to distinguish between text | |
| encoder lora layers.`,name:"state_dict"},{anchor:"diffusers.loaders.StableDiffusionXLLoraLoaderMixin.load_lora_into_unet.network_alphas",description:`<strong>network_alphas</strong> (<code>Dict[str, float]</code>) — | |
| The value of the network alpha used for stable learning and preventing underflow. This value has the | |
| same meaning as the <code>--network_alpha</code> option in the kohya-ss trainer script. Refer to <a href="https://github.com/darkstorm2150/sd-scripts/blob/main/docs/train_network_README-en.md#execute-learning" rel="nofollow">this | |
| link</a>.`,name:"network_alphas"},{anchor:"diffusers.loaders.StableDiffusionXLLoraLoaderMixin.load_lora_into_unet.unet",description:`<strong>unet</strong> (<code>UNet2DConditionModel</code>) — | |
| The UNet model to load the LoRA layers into.`,name:"unet"},{anchor:"diffusers.loaders.StableDiffusionXLLoraLoaderMixin.load_lora_into_unet.adapter_name",description:`<strong>adapter_name</strong> (<code>str</code>, <em>optional</em>) — | |
| Adapter name to be used for referencing the loaded adapter model. If not specified, it will use | |
| <code>default_{i}</code> where i is the total number of adapters being loaded.`,name:"adapter_name"},{anchor:"diffusers.loaders.StableDiffusionXLLoraLoaderMixin.load_lora_into_unet.low_cpu_mem_usage",description:`<strong>low_cpu_mem_usage</strong> (<code>bool</code>, <em>optional</em>) — | |
| Speed up model loading only loading the pretrained LoRA weights and not initializing the random | |
| weights.`,name:"low_cpu_mem_usage"},{anchor:"diffusers.loaders.StableDiffusionXLLoraLoaderMixin.load_lora_into_unet.hotswap",description:`<strong>hotswap</strong> (<code>bool</code>, <em>optional</em>) — | |
| See <a href="/docs/diffusers/pr_11477/en/api/loaders/lora#diffusers.loaders.StableDiffusionLoraLoaderMixin.load_lora_weights">load_lora_weights()</a>.`,name:"hotswap"}],source:"https://github.com/huggingface/diffusers/blob/vr_11477/src/diffusers/loaders/lora_pipeline.py#L797"}}),Zr=new $({props:{name:"load_lora_weights",anchor:"diffusers.loaders.StableDiffusionXLLoraLoaderMixin.load_lora_weights",parameters:[{name:"pretrained_model_name_or_path_or_dict",val:": typing.Union[str, typing.Dict[str, torch.Tensor]]"},{name:"adapter_name",val:": typing.Optional[str] = None"},{name:"hotswap",val:": bool = False"},{name:"**kwargs",val:""}],parametersDescription:[{anchor:"diffusers.loaders.StableDiffusionXLLoraLoaderMixin.load_lora_weights.pretrained_model_name_or_path_or_dict",description:`<strong>pretrained_model_name_or_path_or_dict</strong> (<code>str</code> or <code>os.PathLike</code> or <code>dict</code>) — | |
| See <a href="/docs/diffusers/pr_11477/en/api/loaders/lora#diffusers.loaders.StableDiffusionLoraLoaderMixin.lora_state_dict">lora_state_dict()</a>.`,name:"pretrained_model_name_or_path_or_dict"},{anchor:"diffusers.loaders.StableDiffusionXLLoraLoaderMixin.load_lora_weights.adapter_name",description:`<strong>adapter_name</strong> (<code>str</code>, <em>optional</em>) — | |
| Adapter name to be used for referencing the loaded adapter model. If not specified, it will use | |
| <code>default_{i}</code> where i is the total number of adapters being loaded.`,name:"adapter_name"},{anchor:"diffusers.loaders.StableDiffusionXLLoraLoaderMixin.load_lora_weights.low_cpu_mem_usage",description:`<strong>low_cpu_mem_usage</strong> (<code>bool</code>, <em>optional</em>) — | |
| Speed up model loading by only loading the pretrained LoRA weights and not initializing the random | |
| weights.`,name:"low_cpu_mem_usage"},{anchor:"diffusers.loaders.StableDiffusionXLLoraLoaderMixin.load_lora_weights.hotswap",description:`<strong>hotswap</strong> (<code>bool</code>, <em>optional</em>) — | |
| See <a href="/docs/diffusers/pr_11477/en/api/loaders/lora#diffusers.loaders.StableDiffusionLoraLoaderMixin.load_lora_weights">load_lora_weights()</a>.`,name:"hotswap"},{anchor:"diffusers.loaders.StableDiffusionXLLoraLoaderMixin.load_lora_weights.kwargs",description:`<strong>kwargs</strong> (<code>dict</code>, <em>optional</em>) — | |
| See <a href="/docs/diffusers/pr_11477/en/api/loaders/lora#diffusers.loaders.StableDiffusionLoraLoaderMixin.lora_state_dict">lora_state_dict()</a>.`,name:"kwargs"}],source:"https://github.com/huggingface/diffusers/blob/vr_11477/src/diffusers/loaders/lora_pipeline.py#L585"}}),Or=new $({props:{name:"lora_state_dict",anchor:"diffusers.loaders.StableDiffusionXLLoraLoaderMixin.lora_state_dict",parameters:[{name:"pretrained_model_name_or_path_or_dict",val:": typing.Union[str, typing.Dict[str, torch.Tensor]]"},{name:"**kwargs",val:""}],parametersDescription:[{anchor:"diffusers.loaders.StableDiffusionXLLoraLoaderMixin.lora_state_dict.pretrained_model_name_or_path_or_dict",description:`<strong>pretrained_model_name_or_path_or_dict</strong> (<code>str</code> or <code>os.PathLike</code> or <code>dict</code>) — | |
| Can be either:</p> | |
| <ul> | |
| <li>A string, the <em>model id</em> (for example <code>google/ddpm-celebahq-256</code>) of a pretrained model hosted on | |
| the Hub.</li> | |
| <li>A path to a <em>directory</em> (for example <code>./my_model_directory</code>) containing the model weights saved | |
| with <a href="/docs/diffusers/pr_11477/en/api/models/overview#diffusers.ModelMixin.save_pretrained">ModelMixin.save_pretrained()</a>.</li> | |
| <li>A <a href="https://pytorch.org/tutorials/beginner/saving_loading_models.html#what-is-a-state-dict" rel="nofollow">torch state | |
| dict</a>.</li> | |
| </ul>`,name:"pretrained_model_name_or_path_or_dict"},{anchor:"diffusers.loaders.StableDiffusionXLLoraLoaderMixin.lora_state_dict.cache_dir",description:`<strong>cache_dir</strong> (<code>Union[str, os.PathLike]</code>, <em>optional</em>) — | |
| Path to a directory where a downloaded pretrained model configuration is cached if the standard cache | |
| is not used.`,name:"cache_dir"},{anchor:"diffusers.loaders.StableDiffusionXLLoraLoaderMixin.lora_state_dict.force_download",description:`<strong>force_download</strong> (<code>bool</code>, <em>optional</em>, defaults to <code>False</code>) — | |
| Whether or not to force the (re-)download of the model weights and configuration files, overriding the | |
| cached versions if they exist.`,name:"force_download"},{anchor:"diffusers.loaders.StableDiffusionXLLoraLoaderMixin.lora_state_dict.proxies",description:`<strong>proxies</strong> (<code>Dict[str, str]</code>, <em>optional</em>) — | |
| A dictionary of proxy servers to use by protocol or endpoint, for example, <code>{'http': 'foo.bar:3128', 'http://hostname': 'foo.bar:4012'}</code>. The proxies are used on each request.`,name:"proxies"},{anchor:"diffusers.loaders.StableDiffusionXLLoraLoaderMixin.lora_state_dict.local_files_only",description:`<strong>local_files_only</strong> (<code>bool</code>, <em>optional</em>, defaults to <code>False</code>) — | |
| Whether to only load local model weights and configuration files or not. If set to <code>True</code>, the model | |
| won’t be downloaded from the Hub.`,name:"local_files_only"},{anchor:"diffusers.loaders.StableDiffusionXLLoraLoaderMixin.lora_state_dict.token",description:`<strong>token</strong> (<code>str</code> or <em>bool</em>, <em>optional</em>) — | |
| The token to use as HTTP bearer authorization for remote files. If <code>True</code>, the token generated from | |
| <code>diffusers-cli login</code> (stored in <code>~/.huggingface</code>) is used.`,name:"token"},{anchor:"diffusers.loaders.StableDiffusionXLLoraLoaderMixin.lora_state_dict.revision",description:`<strong>revision</strong> (<code>str</code>, <em>optional</em>, defaults to <code>"main"</code>) — | |
| The specific model version to use. It can be a branch name, a tag name, a commit id, or any identifier | |
| allowed by Git.`,name:"revision"},{anchor:"diffusers.loaders.StableDiffusionXLLoraLoaderMixin.lora_state_dict.subfolder",description:`<strong>subfolder</strong> (<code>str</code>, <em>optional</em>, defaults to <code>""</code>) — | |
| The subfolder location of a model file within a larger model repository on the Hub or locally.`,name:"subfolder"},{anchor:"diffusers.loaders.StableDiffusionXLLoraLoaderMixin.lora_state_dict.weight_name",description:`<strong>weight_name</strong> (<code>str</code>, <em>optional</em>, defaults to None) — | |
| Name of the serialized state dict file.`,name:"weight_name"}],source:"https://github.com/huggingface/diffusers/blob/vr_11477/src/diffusers/loaders/lora_pipeline.py#L681"}}),_o=new S({props:{warning:!0,$$slots:{default:[fh]},$$scope:{ctx:y}}}),Yr=new $({props:{name:"save_lora_weights",anchor:"diffusers.loaders.StableDiffusionXLLoraLoaderMixin.save_lora_weights",parameters:[{name:"save_directory",val:": typing.Union[str, os.PathLike]"},{name:"unet_lora_layers",val:": typing.Dict[str, typing.Union[torch.nn.modules.module.Module, torch.Tensor]] = None"},{name:"text_encoder_lora_layers",val:": typing.Dict[str, typing.Union[torch.nn.modules.module.Module, torch.Tensor]] = None"},{name:"text_encoder_2_lora_layers",val:": typing.Dict[str, typing.Union[torch.nn.modules.module.Module, torch.Tensor]] = None"},{name:"is_main_process",val:": bool = True"},{name:"weight_name",val:": str = None"},{name:"save_function",val:": typing.Callable = None"},{name:"safe_serialization",val:": bool = True"}],parametersDescription:[{anchor:"diffusers.loaders.StableDiffusionXLLoraLoaderMixin.save_lora_weights.save_directory",description:`<strong>save_directory</strong> (<code>str</code> or <code>os.PathLike</code>) — | |
| Directory to save LoRA parameters to. Will be created if it doesn’t exist.`,name:"save_directory"},{anchor:"diffusers.loaders.StableDiffusionXLLoraLoaderMixin.save_lora_weights.unet_lora_layers",description:`<strong>unet_lora_layers</strong> (<code>Dict[str, torch.nn.Module]</code> or <code>Dict[str, torch.Tensor]</code>) — | |
| State dict of the LoRA layers corresponding to the <code>unet</code>.`,name:"unet_lora_layers"},{anchor:"diffusers.loaders.StableDiffusionXLLoraLoaderMixin.save_lora_weights.text_encoder_lora_layers",description:`<strong>text_encoder_lora_layers</strong> (<code>Dict[str, torch.nn.Module]</code> or <code>Dict[str, torch.Tensor]</code>) — | |
| State dict of the LoRA layers corresponding to the <code>text_encoder</code>. Must explicitly pass the text | |
| encoder LoRA state dict because it comes from 🤗 Transformers.`,name:"text_encoder_lora_layers"},{anchor:"diffusers.loaders.StableDiffusionXLLoraLoaderMixin.save_lora_weights.text_encoder_2_lora_layers",description:`<strong>text_encoder_2_lora_layers</strong> (<code>Dict[str, torch.nn.Module]</code> or <code>Dict[str, torch.Tensor]</code>) — | |
| State dict of the LoRA layers corresponding to the <code>text_encoder_2</code>. Must explicitly pass the text | |
| encoder LoRA state dict because it comes from 🤗 Transformers.`,name:"text_encoder_2_lora_layers"},{anchor:"diffusers.loaders.StableDiffusionXLLoraLoaderMixin.save_lora_weights.is_main_process",description:`<strong>is_main_process</strong> (<code>bool</code>, <em>optional</em>, defaults to <code>True</code>) — | |
| Whether the process calling this is the main process or not. Useful during distributed training and you | |
| need to call this function on all processes. In this case, set <code>is_main_process=True</code> only on the main | |
| process to avoid race conditions.`,name:"is_main_process"},{anchor:"diffusers.loaders.StableDiffusionXLLoraLoaderMixin.save_lora_weights.save_function",description:`<strong>save_function</strong> (<code>Callable</code>) — | |
| The function to use to save the state dictionary. Useful during distributed training when you need to | |
| replace <code>torch.save</code> with another method. Can be configured with the environment variable | |
| <code>DIFFUSERS_SAVE_MODE</code>.`,name:"save_function"},{anchor:"diffusers.loaders.StableDiffusionXLLoraLoaderMixin.save_lora_weights.safe_serialization",description:`<strong>safe_serialization</strong> (<code>bool</code>, <em>optional</em>, defaults to <code>True</code>) — | |
| Whether to save the model using <code>safetensors</code> or the traditional PyTorch way with <code>pickle</code>.`,name:"safe_serialization"}],source:"https://github.com/huggingface/diffusers/blob/vr_11477/src/diffusers/loaders/lora_pipeline.py#L908"}}),Qr=new J({props:{title:"SD3LoraLoaderMixin",local:"diffusers.loaders.SD3LoraLoaderMixin",headingTag:"h2"}}),Kr=new $({props:{name:"class diffusers.loaders.SD3LoraLoaderMixin",anchor:"diffusers.loaders.SD3LoraLoaderMixin",parameters:[],source:"https://github.com/huggingface/diffusers/blob/vr_11477/src/diffusers/loaders/lora_pipeline.py#L1038"}}),et=new $({props:{name:"load_lora_into_text_encoder",anchor:"diffusers.loaders.SD3LoraLoaderMixin.load_lora_into_text_encoder",parameters:[{name:"state_dict",val:""},{name:"network_alphas",val:""},{name:"text_encoder",val:""},{name:"prefix",val:" = None"},{name:"lora_scale",val:" = 1.0"},{name:"adapter_name",val:" = None"},{name:"_pipeline",val:" = None"},{name:"low_cpu_mem_usage",val:" = False"},{name:"hotswap",val:": bool = False"}],parametersDescription:[{anchor:"diffusers.loaders.SD3LoraLoaderMixin.load_lora_into_text_encoder.state_dict",description:`<strong>state_dict</strong> (<code>dict</code>) — | |
| A standard state dict containing the lora layer parameters. The key should be prefixed with an | |
| additional <code>text_encoder</code> to distinguish between unet lora layers.`,name:"state_dict"},{anchor:"diffusers.loaders.SD3LoraLoaderMixin.load_lora_into_text_encoder.network_alphas",description:`<strong>network_alphas</strong> (<code>Dict[str, float]</code>) — | |
| The value of the network alpha used for stable learning and preventing underflow. This value has the | |
| same meaning as the <code>--network_alpha</code> option in the kohya-ss trainer script. Refer to <a href="https://github.com/darkstorm2150/sd-scripts/blob/main/docs/train_network_README-en.md#execute-learning" rel="nofollow">this | |
| link</a>.`,name:"network_alphas"},{anchor:"diffusers.loaders.SD3LoraLoaderMixin.load_lora_into_text_encoder.text_encoder",description:`<strong>text_encoder</strong> (<code>CLIPTextModel</code>) — | |
| The text encoder model to load the LoRA layers into.`,name:"text_encoder"},{anchor:"diffusers.loaders.SD3LoraLoaderMixin.load_lora_into_text_encoder.prefix",description:`<strong>prefix</strong> (<code>str</code>) — | |
| Expected prefix of the <code>text_encoder</code> in the <code>state_dict</code>.`,name:"prefix"},{anchor:"diffusers.loaders.SD3LoraLoaderMixin.load_lora_into_text_encoder.lora_scale",description:`<strong>lora_scale</strong> (<code>float</code>) — | |
| How much to scale the output of the lora linear layer before it is added with the output of the regular | |
| lora layer.`,name:"lora_scale"},{anchor:"diffusers.loaders.SD3LoraLoaderMixin.load_lora_into_text_encoder.adapter_name",description:`<strong>adapter_name</strong> (<code>str</code>, <em>optional</em>) — | |
| Adapter name to be used for referencing the loaded adapter model. If not specified, it will use | |
| <code>default_{i}</code> where i is the total number of adapters being loaded.`,name:"adapter_name"},{anchor:"diffusers.loaders.SD3LoraLoaderMixin.load_lora_into_text_encoder.low_cpu_mem_usage",description:`<strong>low_cpu_mem_usage</strong> (<code>bool</code>, <em>optional</em>) — | |
| Speed up model loading by only loading the pretrained LoRA weights and not initializing the random | |
| weights.`,name:"low_cpu_mem_usage"},{anchor:"diffusers.loaders.SD3LoraLoaderMixin.load_lora_into_text_encoder.hotswap",description:`<strong>hotswap</strong> (<code>bool</code>, <em>optional</em>) — | |
| See <a href="/docs/diffusers/pr_11477/en/api/loaders/lora#diffusers.loaders.StableDiffusionLoraLoaderMixin.load_lora_weights">load_lora_weights()</a>.`,name:"hotswap"}],source:"https://github.com/huggingface/diffusers/blob/vr_11477/src/diffusers/loaders/lora_pipeline.py#L1271"}}),ot=new $({props:{name:"load_lora_into_transformer",anchor:"diffusers.loaders.SD3LoraLoaderMixin.load_lora_into_transformer",parameters:[{name:"state_dict",val:""},{name:"transformer",val:""},{name:"adapter_name",val:" = None"},{name:"_pipeline",val:" = None"},{name:"low_cpu_mem_usage",val:" = False"},{name:"hotswap",val:": bool = False"}],parametersDescription:[{anchor:"diffusers.loaders.SD3LoraLoaderMixin.load_lora_into_transformer.state_dict",description:`<strong>state_dict</strong> (<code>dict</code>) — | |
| A standard state dict containing the lora layer parameters. The keys can either be indexed directly | |
| into the unet or prefixed with an additional <code>unet</code> which can be used to distinguish between text | |
| encoder lora layers.`,name:"state_dict"},{anchor:"diffusers.loaders.SD3LoraLoaderMixin.load_lora_into_transformer.transformer",description:`<strong>transformer</strong> (<code>SD3Transformer2DModel</code>) — | |
| The Transformer model to load the LoRA layers into.`,name:"transformer"},{anchor:"diffusers.loaders.SD3LoraLoaderMixin.load_lora_into_transformer.adapter_name",description:`<strong>adapter_name</strong> (<code>str</code>, <em>optional</em>) — | |
| Adapter name to be used for referencing the loaded adapter model. If not specified, it will use | |
| <code>default_{i}</code> where i is the total number of adapters being loaded.`,name:"adapter_name"},{anchor:"diffusers.loaders.SD3LoraLoaderMixin.load_lora_into_transformer.low_cpu_mem_usage",description:`<strong>low_cpu_mem_usage</strong> (<code>bool</code>, <em>optional</em>) — | |
| Speed up model loading by only loading the pretrained LoRA weights and not initializing the random | |
| weights.`,name:"low_cpu_mem_usage"},{anchor:"diffusers.loaders.SD3LoraLoaderMixin.load_lora_into_transformer.hotswap",description:`<strong>hotswap</strong> (<code>bool</code>, <em>optional</em>) — | |
| See <a href="/docs/diffusers/pr_11477/en/api/loaders/lora#diffusers.loaders.StableDiffusionLoraLoaderMixin.load_lora_weights">load_lora_weights()</a>.`,name:"hotswap"}],source:"https://github.com/huggingface/diffusers/blob/vr_11477/src/diffusers/loaders/lora_pipeline.py#L1232"}}),rt=new $({props:{name:"load_lora_weights",anchor:"diffusers.loaders.SD3LoraLoaderMixin.load_lora_weights",parameters:[{name:"pretrained_model_name_or_path_or_dict",val:": typing.Union[str, typing.Dict[str, torch.Tensor]]"},{name:"adapter_name",val:" = None"},{name:"hotswap",val:": bool = False"},{name:"**kwargs",val:""}],parametersDescription:[{anchor:"diffusers.loaders.SD3LoraLoaderMixin.load_lora_weights.pretrained_model_name_or_path_or_dict",description:`<strong>pretrained_model_name_or_path_or_dict</strong> (<code>str</code> or <code>os.PathLike</code> or <code>dict</code>) — | |
| See <a href="/docs/diffusers/pr_11477/en/api/loaders/lora#diffusers.loaders.StableDiffusionLoraLoaderMixin.lora_state_dict">lora_state_dict()</a>.`,name:"pretrained_model_name_or_path_or_dict"},{anchor:"diffusers.loaders.SD3LoraLoaderMixin.load_lora_weights.adapter_name",description:`<strong>adapter_name</strong> (<code>str</code>, <em>optional</em>) — | |
| Adapter name to be used for referencing the loaded adapter model. If not specified, it will use | |
| <code>default_{i}</code> where i is the total number of adapters being loaded.`,name:"adapter_name"},{anchor:"diffusers.loaders.SD3LoraLoaderMixin.load_lora_weights.low_cpu_mem_usage",description:`<strong>low_cpu_mem_usage</strong> (<code>bool</code>, <em>optional</em>) — | |
| Speed up model loading by only loading the pretrained LoRA weights and not initializing the random | |
| weights.`,name:"low_cpu_mem_usage"},{anchor:"diffusers.loaders.SD3LoraLoaderMixin.load_lora_weights.hotswap",description:`<strong>hotswap</strong> (<code>bool</code>, <em>optional</em>) — | |
| See <a href="/docs/diffusers/pr_11477/en/api/loaders/lora#diffusers.loaders.StableDiffusionLoraLoaderMixin.load_lora_weights">load_lora_weights()</a>.`,name:"hotswap"},{anchor:"diffusers.loaders.SD3LoraLoaderMixin.load_lora_weights.kwargs",description:`<strong>kwargs</strong> (<code>dict</code>, <em>optional</em>) — | |
| See <a href="/docs/diffusers/pr_11477/en/api/loaders/lora#diffusers.loaders.StableDiffusionLoraLoaderMixin.lora_state_dict">lora_state_dict()</a>.`,name:"kwargs"}],source:"https://github.com/huggingface/diffusers/blob/vr_11477/src/diffusers/loaders/lora_pipeline.py#L1148"}}),tt=new $({props:{name:"lora_state_dict",anchor:"diffusers.loaders.SD3LoraLoaderMixin.lora_state_dict",parameters:[{name:"pretrained_model_name_or_path_or_dict",val:": typing.Union[str, typing.Dict[str, torch.Tensor]]"},{name:"**kwargs",val:""}],parametersDescription:[{anchor:"diffusers.loaders.SD3LoraLoaderMixin.lora_state_dict.pretrained_model_name_or_path_or_dict",description:`<strong>pretrained_model_name_or_path_or_dict</strong> (<code>str</code> or <code>os.PathLike</code> or <code>dict</code>) — | |
| Can be either:</p> | |
| <ul> | |
| <li>A string, the <em>model id</em> (for example <code>google/ddpm-celebahq-256</code>) of a pretrained model hosted on | |
| the Hub.</li> | |
| <li>A path to a <em>directory</em> (for example <code>./my_model_directory</code>) containing the model weights saved | |
| with <a href="/docs/diffusers/pr_11477/en/api/models/overview#diffusers.ModelMixin.save_pretrained">ModelMixin.save_pretrained()</a>.</li> | |
| <li>A <a href="https://pytorch.org/tutorials/beginner/saving_loading_models.html#what-is-a-state-dict" rel="nofollow">torch state | |
| dict</a>.</li> | |
| </ul>`,name:"pretrained_model_name_or_path_or_dict"},{anchor:"diffusers.loaders.SD3LoraLoaderMixin.lora_state_dict.cache_dir",description:`<strong>cache_dir</strong> (<code>Union[str, os.PathLike]</code>, <em>optional</em>) — | |
| Path to a directory where a downloaded pretrained model configuration is cached if the standard cache | |
| is not used.`,name:"cache_dir"},{anchor:"diffusers.loaders.SD3LoraLoaderMixin.lora_state_dict.force_download",description:`<strong>force_download</strong> (<code>bool</code>, <em>optional</em>, defaults to <code>False</code>) — | |
| Whether or not to force the (re-)download of the model weights and configuration files, overriding the | |
| cached versions if they exist.`,name:"force_download"},{anchor:"diffusers.loaders.SD3LoraLoaderMixin.lora_state_dict.proxies",description:`<strong>proxies</strong> (<code>Dict[str, str]</code>, <em>optional</em>) — | |
| A dictionary of proxy servers to use by protocol or endpoint, for example, <code>{'http': 'foo.bar:3128', 'http://hostname': 'foo.bar:4012'}</code>. The proxies are used on each request.`,name:"proxies"},{anchor:"diffusers.loaders.SD3LoraLoaderMixin.lora_state_dict.local_files_only",description:`<strong>local_files_only</strong> (<code>bool</code>, <em>optional</em>, defaults to <code>False</code>) — | |
| Whether to only load local model weights and configuration files or not. If set to <code>True</code>, the model | |
| won’t be downloaded from the Hub.`,name:"local_files_only"},{anchor:"diffusers.loaders.SD3LoraLoaderMixin.lora_state_dict.token",description:`<strong>token</strong> (<code>str</code> or <em>bool</em>, <em>optional</em>) — | |
| The token to use as HTTP bearer authorization for remote files. If <code>True</code>, the token generated from | |
| <code>diffusers-cli login</code> (stored in <code>~/.huggingface</code>) is used.`,name:"token"},{anchor:"diffusers.loaders.SD3LoraLoaderMixin.lora_state_dict.revision",description:`<strong>revision</strong> (<code>str</code>, <em>optional</em>, defaults to <code>"main"</code>) — | |
| The specific model version to use. It can be a branch name, a tag name, a commit id, or any identifier | |
| allowed by Git.`,name:"revision"},{anchor:"diffusers.loaders.SD3LoraLoaderMixin.lora_state_dict.subfolder",description:`<strong>subfolder</strong> (<code>str</code>, <em>optional</em>, defaults to <code>""</code>) — | |
| The subfolder location of a model file within a larger model repository on the Hub or locally.`,name:"subfolder"}],source:"https://github.com/huggingface/diffusers/blob/vr_11477/src/diffusers/loaders/lora_pipeline.py#L1051"}}),Lo=new S({props:{warning:!0,$$slots:{default:[ph]},$$scope:{ctx:y}}}),at=new $({props:{name:"save_lora_weights",anchor:"diffusers.loaders.SD3LoraLoaderMixin.save_lora_weights",parameters:[{name:"save_directory",val:": typing.Union[str, os.PathLike]"},{name:"transformer_lora_layers",val:": typing.Dict[str, typing.Union[torch.nn.modules.module.Module, torch.Tensor]] = None"},{name:"text_encoder_lora_layers",val:": typing.Dict[str, typing.Union[torch.nn.modules.module.Module, torch.Tensor]] = None"},{name:"text_encoder_2_lora_layers",val:": typing.Dict[str, typing.Union[torch.nn.modules.module.Module, torch.Tensor]] = None"},{name:"is_main_process",val:": bool = True"},{name:"weight_name",val:": str = None"},{name:"save_function",val:": typing.Callable = None"},{name:"safe_serialization",val:": bool = True"}],parametersDescription:[{anchor:"diffusers.loaders.SD3LoraLoaderMixin.save_lora_weights.save_directory",description:`<strong>save_directory</strong> (<code>str</code> or <code>os.PathLike</code>) — | |
| Directory to save LoRA parameters to. Will be created if it doesn’t exist.`,name:"save_directory"},{anchor:"diffusers.loaders.SD3LoraLoaderMixin.save_lora_weights.transformer_lora_layers",description:`<strong>transformer_lora_layers</strong> (<code>Dict[str, torch.nn.Module]</code> or <code>Dict[str, torch.Tensor]</code>) — | |
| State dict of the LoRA layers corresponding to the <code>transformer</code>.`,name:"transformer_lora_layers"},{anchor:"diffusers.loaders.SD3LoraLoaderMixin.save_lora_weights.text_encoder_lora_layers",description:`<strong>text_encoder_lora_layers</strong> (<code>Dict[str, torch.nn.Module]</code> or <code>Dict[str, torch.Tensor]</code>) — | |
| State dict of the LoRA layers corresponding to the <code>text_encoder</code>. Must explicitly pass the text | |
| encoder LoRA state dict because it comes from 🤗 Transformers.`,name:"text_encoder_lora_layers"},{anchor:"diffusers.loaders.SD3LoraLoaderMixin.save_lora_weights.text_encoder_2_lora_layers",description:`<strong>text_encoder_2_lora_layers</strong> (<code>Dict[str, torch.nn.Module]</code> or <code>Dict[str, torch.Tensor]</code>) — | |
| State dict of the LoRA layers corresponding to the <code>text_encoder_2</code>. Must explicitly pass the text | |
| encoder LoRA state dict because it comes from 🤗 Transformers.`,name:"text_encoder_2_lora_layers"},{anchor:"diffusers.loaders.SD3LoraLoaderMixin.save_lora_weights.is_main_process",description:`<strong>is_main_process</strong> (<code>bool</code>, <em>optional</em>, defaults to <code>True</code>) — | |
| Whether the process calling this is the main process or not. Useful during distributed training and you | |
| need to call this function on all processes. In this case, set <code>is_main_process=True</code> only on the main | |
| process to avoid race conditions.`,name:"is_main_process"},{anchor:"diffusers.loaders.SD3LoraLoaderMixin.save_lora_weights.save_function",description:`<strong>save_function</strong> (<code>Callable</code>) — | |
| The function to use to save the state dictionary. Useful during distributed training when you need to | |
| replace <code>torch.save</code> with another method. Can be configured with the environment variable | |
| <code>DIFFUSERS_SAVE_MODE</code>.`,name:"save_function"},{anchor:"diffusers.loaders.SD3LoraLoaderMixin.save_lora_weights.safe_serialization",description:`<strong>safe_serialization</strong> (<code>bool</code>, <em>optional</em>, defaults to <code>True</code>) — | |
| Whether to save the model using <code>safetensors</code> or the traditional PyTorch way with <code>pickle</code>.`,name:"safe_serialization"}],source:"https://github.com/huggingface/diffusers/blob/vr_11477/src/diffusers/loaders/lora_pipeline.py#L1325"}}),nt=new $({props:{name:"unfuse_lora",anchor:"diffusers.loaders.SD3LoraLoaderMixin.unfuse_lora",parameters:[{name:"components",val:": typing.List[str] = ['transformer', 'text_encoder', 'text_encoder_2']"},{name:"**kwargs",val:""}],parametersDescription:[{anchor:"diffusers.loaders.SD3LoraLoaderMixin.unfuse_lora.components",description:"<strong>components</strong> (<code>List[str]</code>) — List of LoRA-injectable components to unfuse LoRA from.",name:"components"},{anchor:"diffusers.loaders.SD3LoraLoaderMixin.unfuse_lora.unfuse_transformer",description:"<strong>unfuse_transformer</strong> (<code>bool</code>, defaults to <code>True</code>) — Whether to unfuse the UNet LoRA parameters.",name:"unfuse_transformer"},{anchor:"diffusers.loaders.SD3LoraLoaderMixin.unfuse_lora.unfuse_text_encoder",description:`<strong>unfuse_text_encoder</strong> (<code>bool</code>, defaults to <code>True</code>) — | |
| Whether to unfuse the text encoder LoRA parameters. If the text encoder wasn’t monkey-patched with the | |
| LoRA parameters then it won’t have any effect.`,name:"unfuse_text_encoder"}],source:"https://github.com/huggingface/diffusers/blob/vr_11477/src/diffusers/loaders/lora_pipeline.py#L1437"}}),bo=new S({props:{warning:!0,$$slots:{default:[mh]},$$scope:{ctx:y}}}),st=new J({props:{title:"FluxLoraLoaderMixin",local:"diffusers.loaders.FluxLoraLoaderMixin",headingTag:"h2"}}),it=new $({props:{name:"class diffusers.loaders.FluxLoraLoaderMixin",anchor:"diffusers.loaders.FluxLoraLoaderMixin",parameters:[],source:"https://github.com/huggingface/diffusers/blob/vr_11477/src/diffusers/loaders/lora_pipeline.py#L1777"}}),dt=new $({props:{name:"load_lora_into_text_encoder",anchor:"diffusers.loaders.FluxLoraLoaderMixin.load_lora_into_text_encoder",parameters:[{name:"state_dict",val:""},{name:"network_alphas",val:""},{name:"text_encoder",val:""},{name:"prefix",val:" = None"},{name:"lora_scale",val:" = 1.0"},{name:"adapter_name",val:" = None"},{name:"_pipeline",val:" = None"},{name:"low_cpu_mem_usage",val:" = False"},{name:"hotswap",val:": bool = False"}],parametersDescription:[{anchor:"diffusers.loaders.FluxLoraLoaderMixin.load_lora_into_text_encoder.state_dict",description:`<strong>state_dict</strong> (<code>dict</code>) — | |
| A standard state dict containing the lora layer parameters. The key should be prefixed with an | |
| additional <code>text_encoder</code> to distinguish between unet lora layers.`,name:"state_dict"},{anchor:"diffusers.loaders.FluxLoraLoaderMixin.load_lora_into_text_encoder.network_alphas",description:`<strong>network_alphas</strong> (<code>Dict[str, float]</code>) — | |
| The value of the network alpha used for stable learning and preventing underflow. This value has the | |
| same meaning as the <code>--network_alpha</code> option in the kohya-ss trainer script. Refer to <a href="https://github.com/darkstorm2150/sd-scripts/blob/main/docs/train_network_README-en.md#execute-learning" rel="nofollow">this | |
| link</a>.`,name:"network_alphas"},{anchor:"diffusers.loaders.FluxLoraLoaderMixin.load_lora_into_text_encoder.text_encoder",description:`<strong>text_encoder</strong> (<code>CLIPTextModel</code>) — | |
| The text encoder model to load the LoRA layers into.`,name:"text_encoder"},{anchor:"diffusers.loaders.FluxLoraLoaderMixin.load_lora_into_text_encoder.prefix",description:`<strong>prefix</strong> (<code>str</code>) — | |
| Expected prefix of the <code>text_encoder</code> in the <code>state_dict</code>.`,name:"prefix"},{anchor:"diffusers.loaders.FluxLoraLoaderMixin.load_lora_into_text_encoder.lora_scale",description:`<strong>lora_scale</strong> (<code>float</code>) — | |
| How much to scale the output of the lora linear layer before it is added with the output of the regular | |
| lora layer.`,name:"lora_scale"},{anchor:"diffusers.loaders.FluxLoraLoaderMixin.load_lora_into_text_encoder.adapter_name",description:`<strong>adapter_name</strong> (<code>str</code>, <em>optional</em>) — | |
| Adapter name to be used for referencing the loaded adapter model. If not specified, it will use | |
| <code>default_{i}</code> where i is the total number of adapters being loaded.`,name:"adapter_name"},{anchor:"diffusers.loaders.FluxLoraLoaderMixin.load_lora_into_text_encoder.low_cpu_mem_usage",description:`<strong>low_cpu_mem_usage</strong> (<code>bool</code>, <em>optional</em>) — | |
| Speed up model loading by only loading the pretrained LoRA weights and not initializing the random | |
| weights.`,name:"low_cpu_mem_usage"},{anchor:"diffusers.loaders.FluxLoraLoaderMixin.load_lora_into_text_encoder.hotswap",description:`<strong>hotswap</strong> (<code>bool</code>, <em>optional</em>) — | |
| See <a href="/docs/diffusers/pr_11477/en/api/loaders/lora#diffusers.loaders.StableDiffusionLoraLoaderMixin.load_lora_weights">load_lora_weights()</a>.`,name:"hotswap"}],source:"https://github.com/huggingface/diffusers/blob/vr_11477/src/diffusers/loaders/lora_pipeline.py#L2150"}}),lt=new $({props:{name:"load_lora_into_transformer",anchor:"diffusers.loaders.FluxLoraLoaderMixin.load_lora_into_transformer",parameters:[{name:"state_dict",val:""},{name:"network_alphas",val:""},{name:"transformer",val:""},{name:"adapter_name",val:" = None"},{name:"_pipeline",val:" = None"},{name:"low_cpu_mem_usage",val:" = False"},{name:"hotswap",val:": bool = False"}],parametersDescription:[{anchor:"diffusers.loaders.FluxLoraLoaderMixin.load_lora_into_transformer.state_dict",description:`<strong>state_dict</strong> (<code>dict</code>) — | |
| A standard state dict containing the lora layer parameters. The keys can either be indexed directly | |
| into the unet or prefixed with an additional <code>unet</code> which can be used to distinguish between text | |
| encoder lora layers.`,name:"state_dict"},{anchor:"diffusers.loaders.FluxLoraLoaderMixin.load_lora_into_transformer.network_alphas",description:`<strong>network_alphas</strong> (<code>Dict[str, float]</code>) — | |
| The value of the network alpha used for stable learning and preventing underflow. This value has the | |
| same meaning as the <code>--network_alpha</code> option in the kohya-ss trainer script. Refer to <a href="https://github.com/darkstorm2150/sd-scripts/blob/main/docs/train_network_README-en.md#execute-learning" rel="nofollow">this | |
| link</a>.`,name:"network_alphas"},{anchor:"diffusers.loaders.FluxLoraLoaderMixin.load_lora_into_transformer.transformer",description:`<strong>transformer</strong> (<code>FluxTransformer2DModel</code>) — | |
| The Transformer model to load the LoRA layers into.`,name:"transformer"},{anchor:"diffusers.loaders.FluxLoraLoaderMixin.load_lora_into_transformer.adapter_name",description:`<strong>adapter_name</strong> (<code>str</code>, <em>optional</em>) — | |
| Adapter name to be used for referencing the loaded adapter model. If not specified, it will use | |
| <code>default_{i}</code> where i is the total number of adapters being loaded.`,name:"adapter_name"},{anchor:"diffusers.loaders.FluxLoraLoaderMixin.load_lora_into_transformer.low_cpu_mem_usage",description:`<strong>low_cpu_mem_usage</strong> (<code>bool</code>, <em>optional</em>) — | |
| Speed up model loading by only loading the pretrained LoRA weights and not initializing the random | |
| weights.`,name:"low_cpu_mem_usage"},{anchor:"diffusers.loaders.FluxLoraLoaderMixin.load_lora_into_transformer.hotswap",description:`<strong>hotswap</strong> (<code>bool</code>, <em>optional</em>) — | |
| See <a href="/docs/diffusers/pr_11477/en/api/loaders/lora#diffusers.loaders.StableDiffusionLoraLoaderMixin.load_lora_weights">load_lora_weights()</a>.`,name:"hotswap"}],source:"https://github.com/huggingface/diffusers/blob/vr_11477/src/diffusers/loaders/lora_pipeline.py#L2046"}}),ct=new $({props:{name:"load_lora_weights",anchor:"diffusers.loaders.FluxLoraLoaderMixin.load_lora_weights",parameters:[{name:"pretrained_model_name_or_path_or_dict",val:": typing.Union[str, typing.Dict[str, torch.Tensor]]"},{name:"adapter_name",val:": typing.Optional[str] = None"},{name:"hotswap",val:": bool = False"},{name:"**kwargs",val:""}],parametersDescription:[{anchor:"diffusers.loaders.FluxLoraLoaderMixin.load_lora_weights.pretrained_model_name_or_path_or_dict",description:`<strong>pretrained_model_name_or_path_or_dict</strong> (<code>str</code> or <code>os.PathLike</code> or <code>dict</code>) — | |
| See <a href="/docs/diffusers/pr_11477/en/api/loaders/lora#diffusers.loaders.StableDiffusionLoraLoaderMixin.lora_state_dict">lora_state_dict()</a>.`,name:"pretrained_model_name_or_path_or_dict"},{anchor:"diffusers.loaders.FluxLoraLoaderMixin.load_lora_weights.adapter_name",description:`<strong>adapter_name</strong> (<code>str</code>, <em>optional</em>) — | |
| Adapter name to be used for referencing the loaded adapter model. If not specified, it will use | |
| <code>default_{i}</code> where i is the total number of adapters being loaded.`,name:"adapter_name"},{anchor:"diffusers.loaders.FluxLoraLoaderMixin.load_lora_weights.low_cpu_mem_usage",description:`<strong>low_cpu_mem_usage</strong> (<code>bool</code>, <em>optional</em>) — | |
| \`Speed up model loading by only loading the pretrained LoRA weights and not initializing the random | |
| weights.`,name:"low_cpu_mem_usage"},{anchor:"diffusers.loaders.FluxLoraLoaderMixin.load_lora_weights.hotswap",description:`<strong>hotswap</strong> (<code>bool</code>, <em>optional</em>) — | |
| See <a href="/docs/diffusers/pr_11477/en/api/loaders/lora#diffusers.loaders.StableDiffusionLoraLoaderMixin.load_lora_weights">load_lora_weights()</a>.`,name:"hotswap"},{anchor:"diffusers.loaders.FluxLoraLoaderMixin.load_lora_weights.kwargs",description:`<strong>kwargs</strong> (<code>dict</code>, <em>optional</em>) — | |
| See <a href="/docs/diffusers/pr_11477/en/api/loaders/lora#diffusers.loaders.StableDiffusionLoraLoaderMixin.lora_state_dict">lora_state_dict()</a>.`,name:"kwargs"}],source:"https://github.com/huggingface/diffusers/blob/vr_11477/src/diffusers/loaders/lora_pipeline.py#L1924"}}),ft=new $({props:{name:"lora_state_dict",anchor:"diffusers.loaders.FluxLoraLoaderMixin.lora_state_dict",parameters:[{name:"pretrained_model_name_or_path_or_dict",val:": typing.Union[str, typing.Dict[str, torch.Tensor]]"},{name:"return_alphas",val:": bool = False"},{name:"**kwargs",val:""}],parametersDescription:[{anchor:"diffusers.loaders.FluxLoraLoaderMixin.lora_state_dict.pretrained_model_name_or_path_or_dict",description:`<strong>pretrained_model_name_or_path_or_dict</strong> (<code>str</code> or <code>os.PathLike</code> or <code>dict</code>) — | |
| Can be either:</p> | |
| <ul> | |
| <li>A string, the <em>model id</em> (for example <code>google/ddpm-celebahq-256</code>) of a pretrained model hosted on | |
| the Hub.</li> | |
| <li>A path to a <em>directory</em> (for example <code>./my_model_directory</code>) containing the model weights saved | |
| with <a href="/docs/diffusers/pr_11477/en/api/models/overview#diffusers.ModelMixin.save_pretrained">ModelMixin.save_pretrained()</a>.</li> | |
| <li>A <a href="https://pytorch.org/tutorials/beginner/saving_loading_models.html#what-is-a-state-dict" rel="nofollow">torch state | |
| dict</a>.</li> | |
| </ul>`,name:"pretrained_model_name_or_path_or_dict"},{anchor:"diffusers.loaders.FluxLoraLoaderMixin.lora_state_dict.cache_dir",description:`<strong>cache_dir</strong> (<code>Union[str, os.PathLike]</code>, <em>optional</em>) — | |
| Path to a directory where a downloaded pretrained model configuration is cached if the standard cache | |
| is not used.`,name:"cache_dir"},{anchor:"diffusers.loaders.FluxLoraLoaderMixin.lora_state_dict.force_download",description:`<strong>force_download</strong> (<code>bool</code>, <em>optional</em>, defaults to <code>False</code>) — | |
| Whether or not to force the (re-)download of the model weights and configuration files, overriding the | |
| cached versions if they exist.`,name:"force_download"},{anchor:"diffusers.loaders.FluxLoraLoaderMixin.lora_state_dict.proxies",description:`<strong>proxies</strong> (<code>Dict[str, str]</code>, <em>optional</em>) — | |
| A dictionary of proxy servers to use by protocol or endpoint, for example, <code>{'http': 'foo.bar:3128', 'http://hostname': 'foo.bar:4012'}</code>. The proxies are used on each request.`,name:"proxies"},{anchor:"diffusers.loaders.FluxLoraLoaderMixin.lora_state_dict.local_files_only",description:`<strong>local_files_only</strong> (<code>bool</code>, <em>optional</em>, defaults to <code>False</code>) — | |
| Whether to only load local model weights and configuration files or not. If set to <code>True</code>, the model | |
| won’t be downloaded from the Hub.`,name:"local_files_only"},{anchor:"diffusers.loaders.FluxLoraLoaderMixin.lora_state_dict.token",description:`<strong>token</strong> (<code>str</code> or <em>bool</em>, <em>optional</em>) — | |
| The token to use as HTTP bearer authorization for remote files. If <code>True</code>, the token generated from | |
| <code>diffusers-cli login</code> (stored in <code>~/.huggingface</code>) is used.`,name:"token"},{anchor:"diffusers.loaders.FluxLoraLoaderMixin.lora_state_dict.revision",description:`<strong>revision</strong> (<code>str</code>, <em>optional</em>, defaults to <code>"main"</code>) — | |
| The specific model version to use. It can be a branch name, a tag name, a commit id, or any identifier | |
| allowed by Git.`,name:"revision"},{anchor:"diffusers.loaders.FluxLoraLoaderMixin.lora_state_dict.subfolder",description:`<strong>subfolder</strong> (<code>str</code>, <em>optional</em>, defaults to <code>""</code>) — | |
| The subfolder location of a model file within a larger model repository on the Hub or locally.`,name:"subfolder"}],source:"https://github.com/huggingface/diffusers/blob/vr_11477/src/diffusers/loaders/lora_pipeline.py#L1790"}}),$o=new S({props:{warning:!0,$$slots:{default:[_h]},$$scope:{ctx:y}}}),pt=new $({props:{name:"save_lora_weights",anchor:"diffusers.loaders.FluxLoraLoaderMixin.save_lora_weights",parameters:[{name:"save_directory",val:": typing.Union[str, os.PathLike]"},{name:"transformer_lora_layers",val:": typing.Dict[str, typing.Union[torch.nn.modules.module.Module, torch.Tensor]] = None"},{name:"text_encoder_lora_layers",val:": typing.Dict[str, torch.nn.modules.module.Module] = None"},{name:"is_main_process",val:": bool = True"},{name:"weight_name",val:": str = None"},{name:"save_function",val:": typing.Callable = None"},{name:"safe_serialization",val:": bool = True"}],parametersDescription:[{anchor:"diffusers.loaders.FluxLoraLoaderMixin.save_lora_weights.save_directory",description:`<strong>save_directory</strong> (<code>str</code> or <code>os.PathLike</code>) — | |
| Directory to save LoRA parameters to. Will be created if it doesn’t exist.`,name:"save_directory"},{anchor:"diffusers.loaders.FluxLoraLoaderMixin.save_lora_weights.transformer_lora_layers",description:`<strong>transformer_lora_layers</strong> (<code>Dict[str, torch.nn.Module]</code> or <code>Dict[str, torch.Tensor]</code>) — | |
| State dict of the LoRA layers corresponding to the <code>transformer</code>.`,name:"transformer_lora_layers"},{anchor:"diffusers.loaders.FluxLoraLoaderMixin.save_lora_weights.text_encoder_lora_layers",description:`<strong>text_encoder_lora_layers</strong> (<code>Dict[str, torch.nn.Module]</code> or <code>Dict[str, torch.Tensor]</code>) — | |
| State dict of the LoRA layers corresponding to the <code>text_encoder</code>. Must explicitly pass the text | |
| encoder LoRA state dict because it comes from 🤗 Transformers.`,name:"text_encoder_lora_layers"},{anchor:"diffusers.loaders.FluxLoraLoaderMixin.save_lora_weights.is_main_process",description:`<strong>is_main_process</strong> (<code>bool</code>, <em>optional</em>, defaults to <code>True</code>) — | |
| Whether the process calling this is the main process or not. Useful during distributed training and you | |
| need to call this function on all processes. In this case, set <code>is_main_process=True</code> only on the main | |
| process to avoid race conditions.`,name:"is_main_process"},{anchor:"diffusers.loaders.FluxLoraLoaderMixin.save_lora_weights.save_function",description:`<strong>save_function</strong> (<code>Callable</code>) — | |
| The function to use to save the state dictionary. Useful during distributed training when you need to | |
| replace <code>torch.save</code> with another method. Can be configured with the environment variable | |
| <code>DIFFUSERS_SAVE_MODE</code>.`,name:"save_function"},{anchor:"diffusers.loaders.FluxLoraLoaderMixin.save_lora_weights.safe_serialization",description:`<strong>safe_serialization</strong> (<code>bool</code>, <em>optional</em>, defaults to <code>True</code>) — | |
| Whether to save the model using <code>safetensors</code> or the traditional PyTorch way with <code>pickle</code>.`,name:"safe_serialization"}],source:"https://github.com/huggingface/diffusers/blob/vr_11477/src/diffusers/loaders/lora_pipeline.py#L2204"}}),mt=new $({props:{name:"unfuse_lora",anchor:"diffusers.loaders.FluxLoraLoaderMixin.unfuse_lora",parameters:[{name:"components",val:": typing.List[str] = ['transformer', 'text_encoder']"},{name:"**kwargs",val:""}],parametersDescription:[{anchor:"diffusers.loaders.FluxLoraLoaderMixin.unfuse_lora.components",description:"<strong>components</strong> (<code>List[str]</code>) — List of LoRA-injectable components to unfuse LoRA from.",name:"components"}],source:"https://github.com/huggingface/diffusers/blob/vr_11477/src/diffusers/loaders/lora_pipeline.py#L2319"}}),yo=new S({props:{warning:!0,$$slots:{default:[uh]},$$scope:{ctx:y}}}),_t=new $({props:{name:"unload_lora_weights",anchor:"diffusers.loaders.FluxLoraLoaderMixin.unload_lora_weights",parameters:[{name:"reset_to_overwritten_params",val:" = False"}],parametersDescription:[{anchor:"diffusers.loaders.FluxLoraLoaderMixin.unload_lora_weights.reset_to_overwritten_params",description:`<strong>reset_to_overwritten_params</strong> (<code>bool</code>, defaults to <code>False</code>) — Whether to reset the LoRA-loaded modules | |
| to their original params. Refer to the <a href="https://huggingface.co/docs/diffusers/main/en/api/pipelines/flux" rel="nofollow">Flux | |
| documentation</a> to learn more.`,name:"reset_to_overwritten_params"}],source:"https://github.com/huggingface/diffusers/blob/vr_11477/src/diffusers/loaders/lora_pipeline.py#L2340"}}),To=new hl({props:{anchor:"diffusers.loaders.FluxLoraLoaderMixin.unload_lora_weights.example",$$slots:{default:[hh]},$$scope:{ctx:y}}}),ut=new J({props:{title:"CogVideoXLoraLoaderMixin",local:"diffusers.loaders.CogVideoXLoraLoaderMixin",headingTag:"h2"}}),ht=new $({props:{name:"class diffusers.loaders.CogVideoXLoraLoaderMixin",anchor:"diffusers.loaders.CogVideoXLoraLoaderMixin",parameters:[],source:"https://github.com/huggingface/diffusers/blob/vr_11477/src/diffusers/loaders/lora_pipeline.py#L2776"}}),gt=new $({props:{name:"load_lora_into_transformer",anchor:"diffusers.loaders.CogVideoXLoraLoaderMixin.load_lora_into_transformer",parameters:[{name:"state_dict",val:""},{name:"transformer",val:""},{name:"adapter_name",val:" = None"},{name:"_pipeline",val:" = None"},{name:"low_cpu_mem_usage",val:" = False"},{name:"hotswap",val:": bool = False"}],parametersDescription:[{anchor:"diffusers.loaders.CogVideoXLoraLoaderMixin.load_lora_into_transformer.state_dict",description:`<strong>state_dict</strong> (<code>dict</code>) — | |
| A standard state dict containing the lora layer parameters. The keys can either be indexed directly | |
| into the unet or prefixed with an additional <code>unet</code> which can be used to distinguish between text | |
| encoder lora layers.`,name:"state_dict"},{anchor:"diffusers.loaders.CogVideoXLoraLoaderMixin.load_lora_into_transformer.transformer",description:`<strong>transformer</strong> (<code>CogVideoXTransformer3DModel</code>) — | |
| The Transformer model to load the LoRA layers into.`,name:"transformer"},{anchor:"diffusers.loaders.CogVideoXLoraLoaderMixin.load_lora_into_transformer.adapter_name",description:`<strong>adapter_name</strong> (<code>str</code>, <em>optional</em>) — | |
| Adapter name to be used for referencing the loaded adapter model. If not specified, it will use | |
| <code>default_{i}</code> where i is the total number of adapters being loaded.`,name:"adapter_name"},{anchor:"diffusers.loaders.CogVideoXLoraLoaderMixin.load_lora_into_transformer.low_cpu_mem_usage",description:`<strong>low_cpu_mem_usage</strong> (<code>bool</code>, <em>optional</em>) — | |
| Speed up model loading by only loading the pretrained LoRA weights and not initializing the random | |
| weights.`,name:"low_cpu_mem_usage"},{anchor:"diffusers.loaders.CogVideoXLoraLoaderMixin.load_lora_into_transformer.hotswap",description:`<strong>hotswap</strong> (<code>bool</code>, <em>optional</em>) — | |
| See <a href="/docs/diffusers/pr_11477/en/api/loaders/lora#diffusers.loaders.StableDiffusionLoraLoaderMixin.load_lora_weights">load_lora_weights()</a>.`,name:"hotswap"}],source:"https://github.com/huggingface/diffusers/blob/vr_11477/src/diffusers/loaders/lora_pipeline.py#L2939"}}),Lt=new $({props:{name:"load_lora_weights",anchor:"diffusers.loaders.CogVideoXLoraLoaderMixin.load_lora_weights",parameters:[{name:"pretrained_model_name_or_path_or_dict",val:": typing.Union[str, typing.Dict[str, torch.Tensor]]"},{name:"adapter_name",val:": typing.Optional[str] = None"},{name:"hotswap",val:": bool = False"},{name:"**kwargs",val:""}],parametersDescription:[{anchor:"diffusers.loaders.CogVideoXLoraLoaderMixin.load_lora_weights.pretrained_model_name_or_path_or_dict",description:`<strong>pretrained_model_name_or_path_or_dict</strong> (<code>str</code> or <code>os.PathLike</code> or <code>dict</code>) — | |
| See <a href="/docs/diffusers/pr_11477/en/api/loaders/lora#diffusers.loaders.StableDiffusionLoraLoaderMixin.lora_state_dict">lora_state_dict()</a>.`,name:"pretrained_model_name_or_path_or_dict"},{anchor:"diffusers.loaders.CogVideoXLoraLoaderMixin.load_lora_weights.adapter_name",description:`<strong>adapter_name</strong> (<code>str</code>, <em>optional</em>) — | |
| Adapter name to be used for referencing the loaded adapter model. If not specified, it will use | |
| <code>default_{i}</code> where i is the total number of adapters being loaded.`,name:"adapter_name"},{anchor:"diffusers.loaders.CogVideoXLoraLoaderMixin.load_lora_weights.low_cpu_mem_usage",description:`<strong>low_cpu_mem_usage</strong> (<code>bool</code>, <em>optional</em>) — | |
| Speed up model loading by only loading the pretrained LoRA weights and not initializing the random | |
| weights.`,name:"low_cpu_mem_usage"},{anchor:"diffusers.loaders.CogVideoXLoraLoaderMixin.load_lora_weights.hotswap",description:`<strong>hotswap</strong> (<code>bool</code>, <em>optional</em>) — | |
| See <a href="/docs/diffusers/pr_11477/en/api/loaders/lora#diffusers.loaders.StableDiffusionLoraLoaderMixin.load_lora_weights">load_lora_weights()</a>.`,name:"hotswap"},{anchor:"diffusers.loaders.CogVideoXLoraLoaderMixin.load_lora_weights.kwargs",description:`<strong>kwargs</strong> (<code>dict</code>, <em>optional</em>) — | |
| See <a href="/docs/diffusers/pr_11477/en/api/loaders/lora#diffusers.loaders.StableDiffusionLoraLoaderMixin.lora_state_dict">lora_state_dict()</a>.`,name:"kwargs"}],source:"https://github.com/huggingface/diffusers/blob/vr_11477/src/diffusers/loaders/lora_pipeline.py#L2882"}}),xt=new $({props:{name:"lora_state_dict",anchor:"diffusers.loaders.CogVideoXLoraLoaderMixin.lora_state_dict",parameters:[{name:"pretrained_model_name_or_path_or_dict",val:": typing.Union[str, typing.Dict[str, torch.Tensor]]"},{name:"**kwargs",val:""}],parametersDescription:[{anchor:"diffusers.loaders.CogVideoXLoraLoaderMixin.lora_state_dict.pretrained_model_name_or_path_or_dict",description:`<strong>pretrained_model_name_or_path_or_dict</strong> (<code>str</code> or <code>os.PathLike</code> or <code>dict</code>) — | |
| Can be either:</p> | |
| <ul> | |
| <li>A string, the <em>model id</em> (for example <code>google/ddpm-celebahq-256</code>) of a pretrained model hosted on | |
| the Hub.</li> | |
| <li>A path to a <em>directory</em> (for example <code>./my_model_directory</code>) containing the model weights saved | |
| with <a href="/docs/diffusers/pr_11477/en/api/models/overview#diffusers.ModelMixin.save_pretrained">ModelMixin.save_pretrained()</a>.</li> | |
| <li>A <a href="https://pytorch.org/tutorials/beginner/saving_loading_models.html#what-is-a-state-dict" rel="nofollow">torch state | |
| dict</a>.</li> | |
| </ul>`,name:"pretrained_model_name_or_path_or_dict"},{anchor:"diffusers.loaders.CogVideoXLoraLoaderMixin.lora_state_dict.cache_dir",description:`<strong>cache_dir</strong> (<code>Union[str, os.PathLike]</code>, <em>optional</em>) — | |
| Path to a directory where a downloaded pretrained model configuration is cached if the standard cache | |
| is not used.`,name:"cache_dir"},{anchor:"diffusers.loaders.CogVideoXLoraLoaderMixin.lora_state_dict.force_download",description:`<strong>force_download</strong> (<code>bool</code>, <em>optional</em>, defaults to <code>False</code>) — | |
| Whether or not to force the (re-)download of the model weights and configuration files, overriding the | |
| cached versions if they exist.`,name:"force_download"},{anchor:"diffusers.loaders.CogVideoXLoraLoaderMixin.lora_state_dict.proxies",description:`<strong>proxies</strong> (<code>Dict[str, str]</code>, <em>optional</em>) — | |
| A dictionary of proxy servers to use by protocol or endpoint, for example, <code>{'http': 'foo.bar:3128', 'http://hostname': 'foo.bar:4012'}</code>. The proxies are used on each request.`,name:"proxies"},{anchor:"diffusers.loaders.CogVideoXLoraLoaderMixin.lora_state_dict.local_files_only",description:`<strong>local_files_only</strong> (<code>bool</code>, <em>optional</em>, defaults to <code>False</code>) — | |
| Whether to only load local model weights and configuration files or not. If set to <code>True</code>, the model | |
| won’t be downloaded from the Hub.`,name:"local_files_only"},{anchor:"diffusers.loaders.CogVideoXLoraLoaderMixin.lora_state_dict.token",description:`<strong>token</strong> (<code>str</code> or <em>bool</em>, <em>optional</em>) — | |
| The token to use as HTTP bearer authorization for remote files. If <code>True</code>, the token generated from | |
| <code>diffusers-cli login</code> (stored in <code>~/.huggingface</code>) is used.`,name:"token"},{anchor:"diffusers.loaders.CogVideoXLoraLoaderMixin.lora_state_dict.revision",description:`<strong>revision</strong> (<code>str</code>, <em>optional</em>, defaults to <code>"main"</code>) — | |
| The specific model version to use. It can be a branch name, a tag name, a commit id, or any identifier | |
| allowed by Git.`,name:"revision"},{anchor:"diffusers.loaders.CogVideoXLoraLoaderMixin.lora_state_dict.subfolder",description:`<strong>subfolder</strong> (<code>str</code>, <em>optional</em>, defaults to <code>""</code>) — | |
| The subfolder location of a model file within a larger model repository on the Hub or locally.`,name:"subfolder"}],source:"https://github.com/huggingface/diffusers/blob/vr_11477/src/diffusers/loaders/lora_pipeline.py#L2784"}}),Co=new S({props:{warning:!0,$$slots:{default:[gh]},$$scope:{ctx:y}}}),bt=new $({props:{name:"save_lora_weights",anchor:"diffusers.loaders.CogVideoXLoraLoaderMixin.save_lora_weights",parameters:[{name:"save_directory",val:": typing.Union[str, os.PathLike]"},{name:"transformer_lora_layers",val:": typing.Dict[str, typing.Union[torch.nn.modules.module.Module, torch.Tensor]] = None"},{name:"is_main_process",val:": bool = True"},{name:"weight_name",val:": str = None"},{name:"save_function",val:": typing.Callable = None"},{name:"safe_serialization",val:": bool = True"}],parametersDescription:[{anchor:"diffusers.loaders.CogVideoXLoraLoaderMixin.save_lora_weights.save_directory",description:`<strong>save_directory</strong> (<code>str</code> or <code>os.PathLike</code>) — | |
| Directory to save LoRA parameters to. Will be created if it doesn’t exist.`,name:"save_directory"},{anchor:"diffusers.loaders.CogVideoXLoraLoaderMixin.save_lora_weights.transformer_lora_layers",description:`<strong>transformer_lora_layers</strong> (<code>Dict[str, torch.nn.Module]</code> or <code>Dict[str, torch.Tensor]</code>) — | |
| State dict of the LoRA layers corresponding to the <code>transformer</code>.`,name:"transformer_lora_layers"},{anchor:"diffusers.loaders.CogVideoXLoraLoaderMixin.save_lora_weights.is_main_process",description:`<strong>is_main_process</strong> (<code>bool</code>, <em>optional</em>, defaults to <code>True</code>) — | |
| Whether the process calling this is the main process or not. Useful during distributed training and you | |
| need to call this function on all processes. In this case, set <code>is_main_process=True</code> only on the main | |
| process to avoid race conditions.`,name:"is_main_process"},{anchor:"diffusers.loaders.CogVideoXLoraLoaderMixin.save_lora_weights.save_function",description:`<strong>save_function</strong> (<code>Callable</code>) — | |
| The function to use to save the state dictionary. Useful during distributed training when you need to | |
| replace <code>torch.save</code> with another method. Can be configured with the environment variable | |
| <code>DIFFUSERS_SAVE_MODE</code>.`,name:"save_function"},{anchor:"diffusers.loaders.CogVideoXLoraLoaderMixin.save_lora_weights.safe_serialization",description:`<strong>safe_serialization</strong> (<code>bool</code>, <em>optional</em>, defaults to <code>True</code>) — | |
| Whether to save the model using <code>safetensors</code> or the traditional PyTorch way with <code>pickle</code>.`,name:"safe_serialization"}],source:"https://github.com/huggingface/diffusers/blob/vr_11477/src/diffusers/loaders/lora_pipeline.py#L2979"}}),vt=new $({props:{name:"unfuse_lora",anchor:"diffusers.loaders.CogVideoXLoraLoaderMixin.unfuse_lora",parameters:[{name:"components",val:": typing.List[str] = ['transformer']"},{name:"**kwargs",val:""}],parametersDescription:[{anchor:"diffusers.loaders.CogVideoXLoraLoaderMixin.unfuse_lora.components",description:"<strong>components</strong> (<code>List[str]</code>) — List of LoRA-injectable components to unfuse LoRA from.",name:"components"},{anchor:"diffusers.loaders.CogVideoXLoraLoaderMixin.unfuse_lora.unfuse_transformer",description:"<strong>unfuse_transformer</strong> (<code>bool</code>, defaults to <code>True</code>) — Whether to unfuse the UNet LoRA parameters.",name:"unfuse_transformer"}],source:"https://github.com/huggingface/diffusers/blob/vr_11477/src/diffusers/loaders/lora_pipeline.py#L3074"}}),ko=new S({props:{warning:!0,$$slots:{default:[Lh]},$$scope:{ctx:y}}}),wt=new J({props:{title:"Mochi1LoraLoaderMixin",local:"diffusers.loaders.Mochi1LoraLoaderMixin",headingTag:"h2"}}),$t=new $({props:{name:"class diffusers.loaders.Mochi1LoraLoaderMixin",anchor:"diffusers.loaders.Mochi1LoraLoaderMixin",parameters:[],source:"https://github.com/huggingface/diffusers/blob/vr_11477/src/diffusers/loaders/lora_pipeline.py#L3092"}}),Mt=new $({props:{name:"load_lora_into_transformer",anchor:"diffusers.loaders.Mochi1LoraLoaderMixin.load_lora_into_transformer",parameters:[{name:"state_dict",val:""},{name:"transformer",val:""},{name:"adapter_name",val:" = None"},{name:"_pipeline",val:" = None"},{name:"low_cpu_mem_usage",val:" = False"},{name:"hotswap",val:": bool = False"}],parametersDescription:[{anchor:"diffusers.loaders.Mochi1LoraLoaderMixin.load_lora_into_transformer.state_dict",description:`<strong>state_dict</strong> (<code>dict</code>) — | |
| A standard state dict containing the lora layer parameters. The keys can either be indexed directly | |
| into the unet or prefixed with an additional <code>unet</code> which can be used to distinguish between text | |
| encoder lora layers.`,name:"state_dict"},{anchor:"diffusers.loaders.Mochi1LoraLoaderMixin.load_lora_into_transformer.transformer",description:`<strong>transformer</strong> (<code>MochiTransformer3DModel</code>) — | |
| The Transformer model to load the LoRA layers into.`,name:"transformer"},{anchor:"diffusers.loaders.Mochi1LoraLoaderMixin.load_lora_into_transformer.adapter_name",description:`<strong>adapter_name</strong> (<code>str</code>, <em>optional</em>) — | |
| Adapter name to be used for referencing the loaded adapter model. If not specified, it will use | |
| <code>default_{i}</code> where i is the total number of adapters being loaded.`,name:"adapter_name"},{anchor:"diffusers.loaders.Mochi1LoraLoaderMixin.load_lora_into_transformer.low_cpu_mem_usage",description:`<strong>low_cpu_mem_usage</strong> (<code>bool</code>, <em>optional</em>) — | |
| Speed up model loading by only loading the pretrained LoRA weights and not initializing the random | |
| weights.`,name:"low_cpu_mem_usage"},{anchor:"diffusers.loaders.Mochi1LoraLoaderMixin.load_lora_into_transformer.hotswap",description:`<strong>hotswap</strong> (<code>bool</code>, <em>optional</em>) — | |
| See <a href="/docs/diffusers/pr_11477/en/api/loaders/lora#diffusers.loaders.StableDiffusionLoraLoaderMixin.load_lora_weights">load_lora_weights()</a>.`,name:"hotswap"}],source:"https://github.com/huggingface/diffusers/blob/vr_11477/src/diffusers/loaders/lora_pipeline.py#L3256"}}),yt=new $({props:{name:"load_lora_weights",anchor:"diffusers.loaders.Mochi1LoraLoaderMixin.load_lora_weights",parameters:[{name:"pretrained_model_name_or_path_or_dict",val:": typing.Union[str, typing.Dict[str, torch.Tensor]]"},{name:"adapter_name",val:": typing.Optional[str] = None"},{name:"hotswap",val:": bool = False"},{name:"**kwargs",val:""}],parametersDescription:[{anchor:"diffusers.loaders.Mochi1LoraLoaderMixin.load_lora_weights.pretrained_model_name_or_path_or_dict",description:`<strong>pretrained_model_name_or_path_or_dict</strong> (<code>str</code> or <code>os.PathLike</code> or <code>dict</code>) — | |
| See <a href="/docs/diffusers/pr_11477/en/api/loaders/lora#diffusers.loaders.StableDiffusionLoraLoaderMixin.lora_state_dict">lora_state_dict()</a>.`,name:"pretrained_model_name_or_path_or_dict"},{anchor:"diffusers.loaders.Mochi1LoraLoaderMixin.load_lora_weights.adapter_name",description:`<strong>adapter_name</strong> (<code>str</code>, <em>optional</em>) — | |
| Adapter name to be used for referencing the loaded adapter model. If not specified, it will use | |
| <code>default_{i}</code> where i is the total number of adapters being loaded.`,name:"adapter_name"},{anchor:"diffusers.loaders.Mochi1LoraLoaderMixin.load_lora_weights.low_cpu_mem_usage",description:`<strong>low_cpu_mem_usage</strong> (<code>bool</code>, <em>optional</em>) — | |
| Speed up model loading by only loading the pretrained LoRA weights and not initializing the random | |
| weights.`,name:"low_cpu_mem_usage"},{anchor:"diffusers.loaders.Mochi1LoraLoaderMixin.load_lora_weights.hotswap",description:`<strong>hotswap</strong> (<code>bool</code>, <em>optional</em>) — | |
| See <a href="/docs/diffusers/pr_11477/en/api/loaders/lora#diffusers.loaders.StableDiffusionLoraLoaderMixin.load_lora_weights">load_lora_weights()</a>.`,name:"hotswap"},{anchor:"diffusers.loaders.Mochi1LoraLoaderMixin.load_lora_weights.kwargs",description:`<strong>kwargs</strong> (<code>dict</code>, <em>optional</em>) — | |
| See <a href="/docs/diffusers/pr_11477/en/api/loaders/lora#diffusers.loaders.StableDiffusionLoraLoaderMixin.lora_state_dict">lora_state_dict()</a>.`,name:"kwargs"}],source:"https://github.com/huggingface/diffusers/blob/vr_11477/src/diffusers/loaders/lora_pipeline.py#L3199"}}),Tt=new $({props:{name:"lora_state_dict",anchor:"diffusers.loaders.Mochi1LoraLoaderMixin.lora_state_dict",parameters:[{name:"pretrained_model_name_or_path_or_dict",val:": typing.Union[str, typing.Dict[str, torch.Tensor]]"},{name:"**kwargs",val:""}],parametersDescription:[{anchor:"diffusers.loaders.Mochi1LoraLoaderMixin.lora_state_dict.pretrained_model_name_or_path_or_dict",description:`<strong>pretrained_model_name_or_path_or_dict</strong> (<code>str</code> or <code>os.PathLike</code> or <code>dict</code>) — | |
| Can be either:</p> | |
| <ul> | |
| <li>A string, the <em>model id</em> (for example <code>google/ddpm-celebahq-256</code>) of a pretrained model hosted on | |
| the Hub.</li> | |
| <li>A path to a <em>directory</em> (for example <code>./my_model_directory</code>) containing the model weights saved | |
| with <a href="/docs/diffusers/pr_11477/en/api/models/overview#diffusers.ModelMixin.save_pretrained">ModelMixin.save_pretrained()</a>.</li> | |
| <li>A <a href="https://pytorch.org/tutorials/beginner/saving_loading_models.html#what-is-a-state-dict" rel="nofollow">torch state | |
| dict</a>.</li> | |
| </ul>`,name:"pretrained_model_name_or_path_or_dict"},{anchor:"diffusers.loaders.Mochi1LoraLoaderMixin.lora_state_dict.cache_dir",description:`<strong>cache_dir</strong> (<code>Union[str, os.PathLike]</code>, <em>optional</em>) — | |
| Path to a directory where a downloaded pretrained model configuration is cached if the standard cache | |
| is not used.`,name:"cache_dir"},{anchor:"diffusers.loaders.Mochi1LoraLoaderMixin.lora_state_dict.force_download",description:`<strong>force_download</strong> (<code>bool</code>, <em>optional</em>, defaults to <code>False</code>) — | |
| Whether or not to force the (re-)download of the model weights and configuration files, overriding the | |
| cached versions if they exist.`,name:"force_download"},{anchor:"diffusers.loaders.Mochi1LoraLoaderMixin.lora_state_dict.proxies",description:`<strong>proxies</strong> (<code>Dict[str, str]</code>, <em>optional</em>) — | |
| A dictionary of proxy servers to use by protocol or endpoint, for example, <code>{'http': 'foo.bar:3128', 'http://hostname': 'foo.bar:4012'}</code>. The proxies are used on each request.`,name:"proxies"},{anchor:"diffusers.loaders.Mochi1LoraLoaderMixin.lora_state_dict.local_files_only",description:`<strong>local_files_only</strong> (<code>bool</code>, <em>optional</em>, defaults to <code>False</code>) — | |
| Whether to only load local model weights and configuration files or not. If set to <code>True</code>, the model | |
| won’t be downloaded from the Hub.`,name:"local_files_only"},{anchor:"diffusers.loaders.Mochi1LoraLoaderMixin.lora_state_dict.token",description:`<strong>token</strong> (<code>str</code> or <em>bool</em>, <em>optional</em>) — | |
| The token to use as HTTP bearer authorization for remote files. If <code>True</code>, the token generated from | |
| <code>diffusers-cli login</code> (stored in <code>~/.huggingface</code>) is used.`,name:"token"},{anchor:"diffusers.loaders.Mochi1LoraLoaderMixin.lora_state_dict.revision",description:`<strong>revision</strong> (<code>str</code>, <em>optional</em>, defaults to <code>"main"</code>) — | |
| The specific model version to use. It can be a branch name, a tag name, a commit id, or any identifier | |
| allowed by Git.`,name:"revision"},{anchor:"diffusers.loaders.Mochi1LoraLoaderMixin.lora_state_dict.subfolder",description:`<strong>subfolder</strong> (<code>str</code>, <em>optional</em>, defaults to <code>""</code>) — | |
| The subfolder location of a model file within a larger model repository on the Hub or locally.`,name:"subfolder"}],source:"https://github.com/huggingface/diffusers/blob/vr_11477/src/diffusers/loaders/lora_pipeline.py#L3100"}}),Ho=new S({props:{warning:!0,$$slots:{default:[xh]},$$scope:{ctx:y}}}),Dt=new $({props:{name:"save_lora_weights",anchor:"diffusers.loaders.Mochi1LoraLoaderMixin.save_lora_weights",parameters:[{name:"save_directory",val:": typing.Union[str, os.PathLike]"},{name:"transformer_lora_layers",val:": typing.Dict[str, typing.Union[torch.nn.modules.module.Module, torch.Tensor]] = None"},{name:"is_main_process",val:": bool = True"},{name:"weight_name",val:": str = None"},{name:"save_function",val:": typing.Callable = None"},{name:"safe_serialization",val:": bool = True"}],parametersDescription:[{anchor:"diffusers.loaders.Mochi1LoraLoaderMixin.save_lora_weights.save_directory",description:`<strong>save_directory</strong> (<code>str</code> or <code>os.PathLike</code>) — | |
| Directory to save LoRA parameters to. Will be created if it doesn’t exist.`,name:"save_directory"},{anchor:"diffusers.loaders.Mochi1LoraLoaderMixin.save_lora_weights.transformer_lora_layers",description:`<strong>transformer_lora_layers</strong> (<code>Dict[str, torch.nn.Module]</code> or <code>Dict[str, torch.Tensor]</code>) — | |
| State dict of the LoRA layers corresponding to the <code>transformer</code>.`,name:"transformer_lora_layers"},{anchor:"diffusers.loaders.Mochi1LoraLoaderMixin.save_lora_weights.is_main_process",description:`<strong>is_main_process</strong> (<code>bool</code>, <em>optional</em>, defaults to <code>True</code>) — | |
| Whether the process calling this is the main process or not. Useful during distributed training and you | |
| need to call this function on all processes. In this case, set <code>is_main_process=True</code> only on the main | |
| process to avoid race conditions.`,name:"is_main_process"},{anchor:"diffusers.loaders.Mochi1LoraLoaderMixin.save_lora_weights.save_function",description:`<strong>save_function</strong> (<code>Callable</code>) — | |
| The function to use to save the state dictionary. Useful during distributed training when you need to | |
| replace <code>torch.save</code> with another method. Can be configured with the environment variable | |
| <code>DIFFUSERS_SAVE_MODE</code>.`,name:"save_function"},{anchor:"diffusers.loaders.Mochi1LoraLoaderMixin.save_lora_weights.safe_serialization",description:`<strong>safe_serialization</strong> (<code>bool</code>, <em>optional</em>, defaults to <code>True</code>) — | |
| Whether to save the model using <code>safetensors</code> or the traditional PyTorch way with <code>pickle</code>.`,name:"safe_serialization"}],source:"https://github.com/huggingface/diffusers/blob/vr_11477/src/diffusers/loaders/lora_pipeline.py#L3296"}}),St=new $({props:{name:"unfuse_lora",anchor:"diffusers.loaders.Mochi1LoraLoaderMixin.unfuse_lora",parameters:[{name:"components",val:": typing.List[str] = ['transformer']"},{name:"**kwargs",val:""}],parametersDescription:[{anchor:"diffusers.loaders.Mochi1LoraLoaderMixin.unfuse_lora.components",description:"<strong>components</strong> (<code>List[str]</code>) — List of LoRA-injectable components to unfuse LoRA from.",name:"components"},{anchor:"diffusers.loaders.Mochi1LoraLoaderMixin.unfuse_lora.unfuse_transformer",description:"<strong>unfuse_transformer</strong> (<code>bool</code>, defaults to <code>True</code>) — Whether to unfuse the UNet LoRA parameters.",name:"unfuse_transformer"}],source:"https://github.com/huggingface/diffusers/blob/vr_11477/src/diffusers/loaders/lora_pipeline.py#L3393"}}),Vo=new S({props:{warning:!0,$$slots:{default:[bh]},$$scope:{ctx:y}}}),Ct=new J({props:{title:"AuraFlowLoraLoaderMixin",local:"diffusers.loaders.AuraFlowLoraLoaderMixin",headingTag:"h2"}}),At=new $({props:{name:"class diffusers.loaders.AuraFlowLoraLoaderMixin",anchor:"diffusers.loaders.AuraFlowLoraLoaderMixin",parameters:[],source:"https://github.com/huggingface/diffusers/blob/vr_11477/src/diffusers/loaders/lora_pipeline.py#L1458"}}),kt=new $({props:{name:"load_lora_into_transformer",anchor:"diffusers.loaders.AuraFlowLoraLoaderMixin.load_lora_into_transformer",parameters:[{name:"state_dict",val:""},{name:"transformer",val:""},{name:"adapter_name",val:" = None"},{name:"_pipeline",val:" = None"},{name:"low_cpu_mem_usage",val:" = False"},{name:"hotswap",val:": bool = False"}],parametersDescription:[{anchor:"diffusers.loaders.AuraFlowLoraLoaderMixin.load_lora_into_transformer.state_dict",description:`<strong>state_dict</strong> (<code>dict</code>) — | |
| A standard state dict containing the lora layer parameters. The keys can either be indexed directly | |
| into the unet or prefixed with an additional <code>unet</code> which can be used to distinguish between text | |
| encoder lora layers.`,name:"state_dict"},{anchor:"diffusers.loaders.AuraFlowLoraLoaderMixin.load_lora_into_transformer.transformer",description:`<strong>transformer</strong> (<code>AuraFlowTransformer2DModel</code>) — | |
| The Transformer model to load the LoRA layers into.`,name:"transformer"},{anchor:"diffusers.loaders.AuraFlowLoraLoaderMixin.load_lora_into_transformer.adapter_name",description:`<strong>adapter_name</strong> (<code>str</code>, <em>optional</em>) — | |
| Adapter name to be used for referencing the loaded adapter model. If not specified, it will use | |
| <code>default_{i}</code> where i is the total number of adapters being loaded.`,name:"adapter_name"},{anchor:"diffusers.loaders.AuraFlowLoraLoaderMixin.load_lora_into_transformer.low_cpu_mem_usage",description:`<strong>low_cpu_mem_usage</strong> (<code>bool</code>, <em>optional</em>) — | |
| Speed up model loading by only loading the pretrained LoRA weights and not initializing the random | |
| weights.`,name:"low_cpu_mem_usage"},{anchor:"diffusers.loaders.AuraFlowLoraLoaderMixin.load_lora_into_transformer.hotswap",description:`<strong>hotswap</strong> (<code>bool</code>, <em>optional</em>) — | |
| See <a href="/docs/diffusers/pr_11477/en/api/loaders/lora#diffusers.loaders.StableDiffusionLoraLoaderMixin.load_lora_weights">load_lora_weights()</a>.`,name:"hotswap"}],source:"https://github.com/huggingface/diffusers/blob/vr_11477/src/diffusers/loaders/lora_pipeline.py#L1622"}}),Pt=new $({props:{name:"load_lora_weights",anchor:"diffusers.loaders.AuraFlowLoraLoaderMixin.load_lora_weights",parameters:[{name:"pretrained_model_name_or_path_or_dict",val:": typing.Union[str, typing.Dict[str, torch.Tensor]]"},{name:"adapter_name",val:": typing.Optional[str] = None"},{name:"hotswap",val:": bool = False"},{name:"**kwargs",val:""}],parametersDescription:[{anchor:"diffusers.loaders.AuraFlowLoraLoaderMixin.load_lora_weights.pretrained_model_name_or_path_or_dict",description:`<strong>pretrained_model_name_or_path_or_dict</strong> (<code>str</code> or <code>os.PathLike</code> or <code>dict</code>) — | |
| See <a href="/docs/diffusers/pr_11477/en/api/loaders/lora#diffusers.loaders.StableDiffusionLoraLoaderMixin.lora_state_dict">lora_state_dict()</a>.`,name:"pretrained_model_name_or_path_or_dict"},{anchor:"diffusers.loaders.AuraFlowLoraLoaderMixin.load_lora_weights.adapter_name",description:`<strong>adapter_name</strong> (<code>str</code>, <em>optional</em>) — | |
| Adapter name to be used for referencing the loaded adapter model. If not specified, it will use | |
| <code>default_{i}</code> where i is the total number of adapters being loaded.`,name:"adapter_name"},{anchor:"diffusers.loaders.AuraFlowLoraLoaderMixin.load_lora_weights.low_cpu_mem_usage",description:`<strong>low_cpu_mem_usage</strong> (<code>bool</code>, <em>optional</em>) — | |
| Speed up model loading by only loading the pretrained LoRA weights and not initializing the random | |
| weights.`,name:"low_cpu_mem_usage"},{anchor:"diffusers.loaders.AuraFlowLoraLoaderMixin.load_lora_weights.hotswap",description:`<strong>hotswap</strong> (<code>bool</code>, <em>optional</em>) — | |
| See <a href="/docs/diffusers/pr_11477/en/api/loaders/lora#diffusers.loaders.StableDiffusionLoraLoaderMixin.load_lora_weights">load_lora_weights()</a>.`,name:"hotswap"},{anchor:"diffusers.loaders.AuraFlowLoraLoaderMixin.load_lora_weights.kwargs",description:`<strong>kwargs</strong> (<code>dict</code>, <em>optional</em>) — | |
| See <a href="/docs/diffusers/pr_11477/en/api/loaders/lora#diffusers.loaders.StableDiffusionLoraLoaderMixin.lora_state_dict">lora_state_dict()</a>.`,name:"kwargs"}],source:"https://github.com/huggingface/diffusers/blob/vr_11477/src/diffusers/loaders/lora_pipeline.py#L1565"}}),Rt=new $({props:{name:"lora_state_dict",anchor:"diffusers.loaders.AuraFlowLoraLoaderMixin.lora_state_dict",parameters:[{name:"pretrained_model_name_or_path_or_dict",val:": typing.Union[str, typing.Dict[str, torch.Tensor]]"},{name:"**kwargs",val:""}],parametersDescription:[{anchor:"diffusers.loaders.AuraFlowLoraLoaderMixin.lora_state_dict.pretrained_model_name_or_path_or_dict",description:`<strong>pretrained_model_name_or_path_or_dict</strong> (<code>str</code> or <code>os.PathLike</code> or <code>dict</code>) — | |
| Can be either:</p> | |
| <ul> | |
| <li>A string, the <em>model id</em> (for example <code>google/ddpm-celebahq-256</code>) of a pretrained model hosted on | |
| the Hub.</li> | |
| <li>A path to a <em>directory</em> (for example <code>./my_model_directory</code>) containing the model weights saved | |
| with <a href="/docs/diffusers/pr_11477/en/api/models/overview#diffusers.ModelMixin.save_pretrained">ModelMixin.save_pretrained()</a>.</li> | |
| <li>A <a href="https://pytorch.org/tutorials/beginner/saving_loading_models.html#what-is-a-state-dict" rel="nofollow">torch state | |
| dict</a>.</li> | |
| </ul>`,name:"pretrained_model_name_or_path_or_dict"},{anchor:"diffusers.loaders.AuraFlowLoraLoaderMixin.lora_state_dict.cache_dir",description:`<strong>cache_dir</strong> (<code>Union[str, os.PathLike]</code>, <em>optional</em>) — | |
| Path to a directory where a downloaded pretrained model configuration is cached if the standard cache | |
| is not used.`,name:"cache_dir"},{anchor:"diffusers.loaders.AuraFlowLoraLoaderMixin.lora_state_dict.force_download",description:`<strong>force_download</strong> (<code>bool</code>, <em>optional</em>, defaults to <code>False</code>) — | |
| Whether or not to force the (re-)download of the model weights and configuration files, overriding the | |
| cached versions if they exist.`,name:"force_download"},{anchor:"diffusers.loaders.AuraFlowLoraLoaderMixin.lora_state_dict.proxies",description:`<strong>proxies</strong> (<code>Dict[str, str]</code>, <em>optional</em>) — | |
| A dictionary of proxy servers to use by protocol or endpoint, for example, <code>{'http': 'foo.bar:3128', 'http://hostname': 'foo.bar:4012'}</code>. The proxies are used on each request.`,name:"proxies"},{anchor:"diffusers.loaders.AuraFlowLoraLoaderMixin.lora_state_dict.local_files_only",description:`<strong>local_files_only</strong> (<code>bool</code>, <em>optional</em>, defaults to <code>False</code>) — | |
| Whether to only load local model weights and configuration files or not. If set to <code>True</code>, the model | |
| won’t be downloaded from the Hub.`,name:"local_files_only"},{anchor:"diffusers.loaders.AuraFlowLoraLoaderMixin.lora_state_dict.token",description:`<strong>token</strong> (<code>str</code> or <em>bool</em>, <em>optional</em>) — | |
| The token to use as HTTP bearer authorization for remote files. If <code>True</code>, the token generated from | |
| <code>diffusers-cli login</code> (stored in <code>~/.huggingface</code>) is used.`,name:"token"},{anchor:"diffusers.loaders.AuraFlowLoraLoaderMixin.lora_state_dict.revision",description:`<strong>revision</strong> (<code>str</code>, <em>optional</em>, defaults to <code>"main"</code>) — | |
| The specific model version to use. It can be a branch name, a tag name, a commit id, or any identifier | |
| allowed by Git.`,name:"revision"},{anchor:"diffusers.loaders.AuraFlowLoraLoaderMixin.lora_state_dict.subfolder",description:`<strong>subfolder</strong> (<code>str</code>, <em>optional</em>, defaults to <code>""</code>) — | |
| The subfolder location of a model file within a larger model repository on the Hub or locally.`,name:"subfolder"}],source:"https://github.com/huggingface/diffusers/blob/vr_11477/src/diffusers/loaders/lora_pipeline.py#L1466"}}),Uo=new S({props:{warning:!0,$$slots:{default:[vh]},$$scope:{ctx:y}}}),Ht=new $({props:{name:"save_lora_weights",anchor:"diffusers.loaders.AuraFlowLoraLoaderMixin.save_lora_weights",parameters:[{name:"save_directory",val:": typing.Union[str, os.PathLike]"},{name:"transformer_lora_layers",val:": typing.Dict[str, typing.Union[torch.nn.modules.module.Module, torch.Tensor]] = None"},{name:"is_main_process",val:": bool = True"},{name:"weight_name",val:": str = None"},{name:"save_function",val:": typing.Callable = None"},{name:"safe_serialization",val:": bool = True"}],parametersDescription:[{anchor:"diffusers.loaders.AuraFlowLoraLoaderMixin.save_lora_weights.save_directory",description:`<strong>save_directory</strong> (<code>str</code> or <code>os.PathLike</code>) — | |
| Directory to save LoRA parameters to. Will be created if it doesn’t exist.`,name:"save_directory"},{anchor:"diffusers.loaders.AuraFlowLoraLoaderMixin.save_lora_weights.transformer_lora_layers",description:`<strong>transformer_lora_layers</strong> (<code>Dict[str, torch.nn.Module]</code> or <code>Dict[str, torch.Tensor]</code>) — | |
| State dict of the LoRA layers corresponding to the <code>transformer</code>.`,name:"transformer_lora_layers"},{anchor:"diffusers.loaders.AuraFlowLoraLoaderMixin.save_lora_weights.is_main_process",description:`<strong>is_main_process</strong> (<code>bool</code>, <em>optional</em>, defaults to <code>True</code>) — | |
| Whether the process calling this is the main process or not. Useful during distributed training and you | |
| need to call this function on all processes. In this case, set <code>is_main_process=True</code> only on the main | |
| process to avoid race conditions.`,name:"is_main_process"},{anchor:"diffusers.loaders.AuraFlowLoraLoaderMixin.save_lora_weights.save_function",description:`<strong>save_function</strong> (<code>Callable</code>) — | |
| The function to use to save the state dictionary. Useful during distributed training when you need to | |
| replace <code>torch.save</code> with another method. Can be configured with the environment variable | |
| <code>DIFFUSERS_SAVE_MODE</code>.`,name:"save_function"},{anchor:"diffusers.loaders.AuraFlowLoraLoaderMixin.save_lora_weights.safe_serialization",description:`<strong>safe_serialization</strong> (<code>bool</code>, <em>optional</em>, defaults to <code>True</code>) — | |
| Whether to save the model using <code>safetensors</code> or the traditional PyTorch way with <code>pickle</code>.`,name:"safe_serialization"}],source:"https://github.com/huggingface/diffusers/blob/vr_11477/src/diffusers/loaders/lora_pipeline.py#L1662"}}),It=new $({props:{name:"unfuse_lora",anchor:"diffusers.loaders.AuraFlowLoraLoaderMixin.unfuse_lora",parameters:[{name:"components",val:": typing.List[str] = ['transformer', 'text_encoder']"},{name:"**kwargs",val:""}],parametersDescription:[{anchor:"diffusers.loaders.AuraFlowLoraLoaderMixin.unfuse_lora.components",description:"<strong>components</strong> (<code>List[str]</code>) — List of LoRA-injectable components to unfuse LoRA from.",name:"components"},{anchor:"diffusers.loaders.AuraFlowLoraLoaderMixin.unfuse_lora.unfuse_transformer",description:"<strong>unfuse_transformer</strong> (<code>bool</code>, defaults to <code>True</code>) — Whether to unfuse the UNet LoRA parameters.",name:"unfuse_transformer"}],source:"https://github.com/huggingface/diffusers/blob/vr_11477/src/diffusers/loaders/lora_pipeline.py#L1759"}}),No=new S({props:{warning:!0,$$slots:{default:[wh]},$$scope:{ctx:y}}}),Vt=new J({props:{title:"LTXVideoLoraLoaderMixin",local:"diffusers.loaders.LTXVideoLoraLoaderMixin",headingTag:"h2"}}),Wt=new $({props:{name:"class diffusers.loaders.LTXVideoLoraLoaderMixin",anchor:"diffusers.loaders.LTXVideoLoraLoaderMixin",parameters:[],source:"https://github.com/huggingface/diffusers/blob/vr_11477/src/diffusers/loaders/lora_pipeline.py#L3411"}}),Ft=new $({props:{name:"load_lora_into_transformer",anchor:"diffusers.loaders.LTXVideoLoraLoaderMixin.load_lora_into_transformer",parameters:[{name:"state_dict",val:""},{name:"transformer",val:""},{name:"adapter_name",val:" = None"},{name:"_pipeline",val:" = None"},{name:"low_cpu_mem_usage",val:" = False"},{name:"hotswap",val:": bool = False"}],parametersDescription:[{anchor:"diffusers.loaders.LTXVideoLoraLoaderMixin.load_lora_into_transformer.state_dict",description:`<strong>state_dict</strong> (<code>dict</code>) — | |
| A standard state dict containing the lora layer parameters. The keys can either be indexed directly | |
| into the unet or prefixed with an additional <code>unet</code> which can be used to distinguish between text | |
| encoder lora layers.`,name:"state_dict"},{anchor:"diffusers.loaders.LTXVideoLoraLoaderMixin.load_lora_into_transformer.transformer",description:`<strong>transformer</strong> (<code>LTXVideoTransformer3DModel</code>) — | |
| The Transformer model to load the LoRA layers into.`,name:"transformer"},{anchor:"diffusers.loaders.LTXVideoLoraLoaderMixin.load_lora_into_transformer.adapter_name",description:`<strong>adapter_name</strong> (<code>str</code>, <em>optional</em>) — | |
| Adapter name to be used for referencing the loaded adapter model. If not specified, it will use | |
| <code>default_{i}</code> where i is the total number of adapters being loaded.`,name:"adapter_name"},{anchor:"diffusers.loaders.LTXVideoLoraLoaderMixin.load_lora_into_transformer.low_cpu_mem_usage",description:`<strong>low_cpu_mem_usage</strong> (<code>bool</code>, <em>optional</em>) — | |
| Speed up model loading by only loading the pretrained LoRA weights and not initializing the random | |
| weights.`,name:"low_cpu_mem_usage"},{anchor:"diffusers.loaders.LTXVideoLoraLoaderMixin.load_lora_into_transformer.hotswap",description:`<strong>hotswap</strong> (<code>bool</code>, <em>optional</em>) — | |
| See <a href="/docs/diffusers/pr_11477/en/api/loaders/lora#diffusers.loaders.StableDiffusionLoraLoaderMixin.load_lora_weights">load_lora_weights()</a>.`,name:"hotswap"}],source:"https://github.com/huggingface/diffusers/blob/vr_11477/src/diffusers/loaders/lora_pipeline.py#L3578"}}),Ut=new $({props:{name:"load_lora_weights",anchor:"diffusers.loaders.LTXVideoLoraLoaderMixin.load_lora_weights",parameters:[{name:"pretrained_model_name_or_path_or_dict",val:": typing.Union[str, typing.Dict[str, torch.Tensor]]"},{name:"adapter_name",val:": typing.Optional[str] = None"},{name:"hotswap",val:": bool = False"},{name:"**kwargs",val:""}],parametersDescription:[{anchor:"diffusers.loaders.LTXVideoLoraLoaderMixin.load_lora_weights.pretrained_model_name_or_path_or_dict",description:`<strong>pretrained_model_name_or_path_or_dict</strong> (<code>str</code> or <code>os.PathLike</code> or <code>dict</code>) — | |
| See <a href="/docs/diffusers/pr_11477/en/api/loaders/lora#diffusers.loaders.StableDiffusionLoraLoaderMixin.lora_state_dict">lora_state_dict()</a>.`,name:"pretrained_model_name_or_path_or_dict"},{anchor:"diffusers.loaders.LTXVideoLoraLoaderMixin.load_lora_weights.adapter_name",description:`<strong>adapter_name</strong> (<code>str</code>, <em>optional</em>) — | |
| Adapter name to be used for referencing the loaded adapter model. If not specified, it will use | |
| <code>default_{i}</code> where i is the total number of adapters being loaded.`,name:"adapter_name"},{anchor:"diffusers.loaders.LTXVideoLoraLoaderMixin.load_lora_weights.low_cpu_mem_usage",description:`<strong>low_cpu_mem_usage</strong> (<code>bool</code>, <em>optional</em>) — | |
| Speed up model loading by only loading the pretrained LoRA weights and not initializing the random | |
| weights.`,name:"low_cpu_mem_usage"},{anchor:"diffusers.loaders.LTXVideoLoraLoaderMixin.load_lora_weights.hotswap",description:`<strong>hotswap</strong> (<code>bool</code>, <em>optional</em>) — | |
| See <a href="/docs/diffusers/pr_11477/en/api/loaders/lora#diffusers.loaders.StableDiffusionLoraLoaderMixin.load_lora_weights">load_lora_weights()</a>.`,name:"hotswap"},{anchor:"diffusers.loaders.LTXVideoLoraLoaderMixin.load_lora_weights.kwargs",description:`<strong>kwargs</strong> (<code>dict</code>, <em>optional</em>) — | |
| See <a href="/docs/diffusers/pr_11477/en/api/loaders/lora#diffusers.loaders.StableDiffusionLoraLoaderMixin.lora_state_dict">lora_state_dict()</a>.`,name:"kwargs"}],source:"https://github.com/huggingface/diffusers/blob/vr_11477/src/diffusers/loaders/lora_pipeline.py#L3521"}}),Et=new $({props:{name:"lora_state_dict",anchor:"diffusers.loaders.LTXVideoLoraLoaderMixin.lora_state_dict",parameters:[{name:"pretrained_model_name_or_path_or_dict",val:": typing.Union[str, typing.Dict[str, torch.Tensor]]"},{name:"**kwargs",val:""}],parametersDescription:[{anchor:"diffusers.loaders.LTXVideoLoraLoaderMixin.lora_state_dict.pretrained_model_name_or_path_or_dict",description:`<strong>pretrained_model_name_or_path_or_dict</strong> (<code>str</code> or <code>os.PathLike</code> or <code>dict</code>) — | |
| Can be either:</p> | |
| <ul> | |
| <li>A string, the <em>model id</em> (for example <code>google/ddpm-celebahq-256</code>) of a pretrained model hosted on | |
| the Hub.</li> | |
| <li>A path to a <em>directory</em> (for example <code>./my_model_directory</code>) containing the model weights saved | |
| with <a href="/docs/diffusers/pr_11477/en/api/models/overview#diffusers.ModelMixin.save_pretrained">ModelMixin.save_pretrained()</a>.</li> | |
| <li>A <a href="https://pytorch.org/tutorials/beginner/saving_loading_models.html#what-is-a-state-dict" rel="nofollow">torch state | |
| dict</a>.</li> | |
| </ul>`,name:"pretrained_model_name_or_path_or_dict"},{anchor:"diffusers.loaders.LTXVideoLoraLoaderMixin.lora_state_dict.cache_dir",description:`<strong>cache_dir</strong> (<code>Union[str, os.PathLike]</code>, <em>optional</em>) — | |
| Path to a directory where a downloaded pretrained model configuration is cached if the standard cache | |
| is not used.`,name:"cache_dir"},{anchor:"diffusers.loaders.LTXVideoLoraLoaderMixin.lora_state_dict.force_download",description:`<strong>force_download</strong> (<code>bool</code>, <em>optional</em>, defaults to <code>False</code>) — | |
| Whether or not to force the (re-)download of the model weights and configuration files, overriding the | |
| cached versions if they exist.`,name:"force_download"},{anchor:"diffusers.loaders.LTXVideoLoraLoaderMixin.lora_state_dict.proxies",description:`<strong>proxies</strong> (<code>Dict[str, str]</code>, <em>optional</em>) — | |
| A dictionary of proxy servers to use by protocol or endpoint, for example, <code>{'http': 'foo.bar:3128', 'http://hostname': 'foo.bar:4012'}</code>. The proxies are used on each request.`,name:"proxies"},{anchor:"diffusers.loaders.LTXVideoLoraLoaderMixin.lora_state_dict.local_files_only",description:`<strong>local_files_only</strong> (<code>bool</code>, <em>optional</em>, defaults to <code>False</code>) — | |
| Whether to only load local model weights and configuration files or not. If set to <code>True</code>, the model | |
| won’t be downloaded from the Hub.`,name:"local_files_only"},{anchor:"diffusers.loaders.LTXVideoLoraLoaderMixin.lora_state_dict.token",description:`<strong>token</strong> (<code>str</code> or <em>bool</em>, <em>optional</em>) — | |
| The token to use as HTTP bearer authorization for remote files. If <code>True</code>, the token generated from | |
| <code>diffusers-cli login</code> (stored in <code>~/.huggingface</code>) is used.`,name:"token"},{anchor:"diffusers.loaders.LTXVideoLoraLoaderMixin.lora_state_dict.revision",description:`<strong>revision</strong> (<code>str</code>, <em>optional</em>, defaults to <code>"main"</code>) — | |
| The specific model version to use. It can be a branch name, a tag name, a commit id, or any identifier | |
| allowed by Git.`,name:"revision"},{anchor:"diffusers.loaders.LTXVideoLoraLoaderMixin.lora_state_dict.subfolder",description:`<strong>subfolder</strong> (<code>str</code>, <em>optional</em>, defaults to <code>""</code>) — | |
| The subfolder location of a model file within a larger model repository on the Hub or locally.`,name:"subfolder"}],source:"https://github.com/huggingface/diffusers/blob/vr_11477/src/diffusers/loaders/lora_pipeline.py#L3419"}}),zo=new S({props:{warning:!0,$$slots:{default:[$h]},$$scope:{ctx:y}}}),Nt=new $({props:{name:"save_lora_weights",anchor:"diffusers.loaders.LTXVideoLoraLoaderMixin.save_lora_weights",parameters:[{name:"save_directory",val:": typing.Union[str, os.PathLike]"},{name:"transformer_lora_layers",val:": typing.Dict[str, typing.Union[torch.nn.modules.module.Module, torch.Tensor]] = None"},{name:"is_main_process",val:": bool = True"},{name:"weight_name",val:": str = None"},{name:"save_function",val:": typing.Callable = None"},{name:"safe_serialization",val:": bool = True"}],parametersDescription:[{anchor:"diffusers.loaders.LTXVideoLoraLoaderMixin.save_lora_weights.save_directory",description:`<strong>save_directory</strong> (<code>str</code> or <code>os.PathLike</code>) — | |
| Directory to save LoRA parameters to. Will be created if it doesn’t exist.`,name:"save_directory"},{anchor:"diffusers.loaders.LTXVideoLoraLoaderMixin.save_lora_weights.transformer_lora_layers",description:`<strong>transformer_lora_layers</strong> (<code>Dict[str, torch.nn.Module]</code> or <code>Dict[str, torch.Tensor]</code>) — | |
| State dict of the LoRA layers corresponding to the <code>transformer</code>.`,name:"transformer_lora_layers"},{anchor:"diffusers.loaders.LTXVideoLoraLoaderMixin.save_lora_weights.is_main_process",description:`<strong>is_main_process</strong> (<code>bool</code>, <em>optional</em>, defaults to <code>True</code>) — | |
| Whether the process calling this is the main process or not. Useful during distributed training and you | |
| need to call this function on all processes. In this case, set <code>is_main_process=True</code> only on the main | |
| process to avoid race conditions.`,name:"is_main_process"},{anchor:"diffusers.loaders.LTXVideoLoraLoaderMixin.save_lora_weights.save_function",description:`<strong>save_function</strong> (<code>Callable</code>) — | |
| The function to use to save the state dictionary. Useful during distributed training when you need to | |
| replace <code>torch.save</code> with another method. Can be configured with the environment variable | |
| <code>DIFFUSERS_SAVE_MODE</code>.`,name:"save_function"},{anchor:"diffusers.loaders.LTXVideoLoraLoaderMixin.save_lora_weights.safe_serialization",description:`<strong>safe_serialization</strong> (<code>bool</code>, <em>optional</em>, defaults to <code>True</code>) — | |
| Whether to save the model using <code>safetensors</code> or the traditional PyTorch way with <code>pickle</code>.`,name:"safe_serialization"}],source:"https://github.com/huggingface/diffusers/blob/vr_11477/src/diffusers/loaders/lora_pipeline.py#L3618"}}),Xt=new $({props:{name:"unfuse_lora",anchor:"diffusers.loaders.LTXVideoLoraLoaderMixin.unfuse_lora",parameters:[{name:"components",val:": typing.List[str] = ['transformer']"},{name:"**kwargs",val:""}],parametersDescription:[{anchor:"diffusers.loaders.LTXVideoLoraLoaderMixin.unfuse_lora.components",description:"<strong>components</strong> (<code>List[str]</code>) — List of LoRA-injectable components to unfuse LoRA from.",name:"components"},{anchor:"diffusers.loaders.LTXVideoLoraLoaderMixin.unfuse_lora.unfuse_transformer",description:"<strong>unfuse_transformer</strong> (<code>bool</code>, defaults to <code>True</code>) — Whether to unfuse the UNet LoRA parameters.",name:"unfuse_transformer"}],source:"https://github.com/huggingface/diffusers/blob/vr_11477/src/diffusers/loaders/lora_pipeline.py#L3715"}}),jo=new S({props:{warning:!0,$$slots:{default:[Mh]},$$scope:{ctx:y}}}),qt=new J({props:{title:"SanaLoraLoaderMixin",local:"diffusers.loaders.SanaLoraLoaderMixin",headingTag:"h2"}}),zt=new $({props:{name:"class diffusers.loaders.SanaLoraLoaderMixin",anchor:"diffusers.loaders.SanaLoraLoaderMixin",parameters:[],source:"https://github.com/huggingface/diffusers/blob/vr_11477/src/diffusers/loaders/lora_pipeline.py#L3733"}}),Bt=new $({props:{name:"load_lora_into_transformer",anchor:"diffusers.loaders.SanaLoraLoaderMixin.load_lora_into_transformer",parameters:[{name:"state_dict",val:""},{name:"transformer",val:""},{name:"adapter_name",val:" = None"},{name:"_pipeline",val:" = None"},{name:"low_cpu_mem_usage",val:" = False"},{name:"hotswap",val:": bool = False"}],parametersDescription:[{anchor:"diffusers.loaders.SanaLoraLoaderMixin.load_lora_into_transformer.state_dict",description:`<strong>state_dict</strong> (<code>dict</code>) — | |
| A standard state dict containing the lora layer parameters. The keys can either be indexed directly | |
| into the unet or prefixed with an additional <code>unet</code> which can be used to distinguish between text | |
| encoder lora layers.`,name:"state_dict"},{anchor:"diffusers.loaders.SanaLoraLoaderMixin.load_lora_into_transformer.transformer",description:`<strong>transformer</strong> (<code>SanaTransformer2DModel</code>) — | |
| The Transformer model to load the LoRA layers into.`,name:"transformer"},{anchor:"diffusers.loaders.SanaLoraLoaderMixin.load_lora_into_transformer.adapter_name",description:`<strong>adapter_name</strong> (<code>str</code>, <em>optional</em>) — | |
| Adapter name to be used for referencing the loaded adapter model. If not specified, it will use | |
| <code>default_{i}</code> where i is the total number of adapters being loaded.`,name:"adapter_name"},{anchor:"diffusers.loaders.SanaLoraLoaderMixin.load_lora_into_transformer.low_cpu_mem_usage",description:`<strong>low_cpu_mem_usage</strong> (<code>bool</code>, <em>optional</em>) — | |
| Speed up model loading by only loading the pretrained LoRA weights and not initializing the random | |
| weights.`,name:"low_cpu_mem_usage"},{anchor:"diffusers.loaders.SanaLoraLoaderMixin.load_lora_into_transformer.hotswap",description:`<strong>hotswap</strong> (<code>bool</code>, <em>optional</em>) — | |
| See <a href="/docs/diffusers/pr_11477/en/api/loaders/lora#diffusers.loaders.StableDiffusionLoraLoaderMixin.load_lora_weights">load_lora_weights()</a>.`,name:"hotswap"}],source:"https://github.com/huggingface/diffusers/blob/vr_11477/src/diffusers/loaders/lora_pipeline.py#L3897"}}),jt=new $({props:{name:"load_lora_weights",anchor:"diffusers.loaders.SanaLoraLoaderMixin.load_lora_weights",parameters:[{name:"pretrained_model_name_or_path_or_dict",val:": typing.Union[str, typing.Dict[str, torch.Tensor]]"},{name:"adapter_name",val:": typing.Optional[str] = None"},{name:"hotswap",val:": bool = False"},{name:"**kwargs",val:""}],parametersDescription:[{anchor:"diffusers.loaders.SanaLoraLoaderMixin.load_lora_weights.pretrained_model_name_or_path_or_dict",description:`<strong>pretrained_model_name_or_path_or_dict</strong> (<code>str</code> or <code>os.PathLike</code> or <code>dict</code>) — | |
| See <a href="/docs/diffusers/pr_11477/en/api/loaders/lora#diffusers.loaders.StableDiffusionLoraLoaderMixin.lora_state_dict">lora_state_dict()</a>.`,name:"pretrained_model_name_or_path_or_dict"},{anchor:"diffusers.loaders.SanaLoraLoaderMixin.load_lora_weights.adapter_name",description:`<strong>adapter_name</strong> (<code>str</code>, <em>optional</em>) — | |
| Adapter name to be used for referencing the loaded adapter model. If not specified, it will use | |
| <code>default_{i}</code> where i is the total number of adapters being loaded.`,name:"adapter_name"},{anchor:"diffusers.loaders.SanaLoraLoaderMixin.load_lora_weights.low_cpu_mem_usage",description:`<strong>low_cpu_mem_usage</strong> (<code>bool</code>, <em>optional</em>) — | |
| Speed up model loading by only loading the pretrained LoRA weights and not initializing the random | |
| weights.`,name:"low_cpu_mem_usage"},{anchor:"diffusers.loaders.SanaLoraLoaderMixin.load_lora_weights.hotswap",description:`<strong>hotswap</strong> (<code>bool</code>, <em>optional</em>) — | |
| See <a href="/docs/diffusers/pr_11477/en/api/loaders/lora#diffusers.loaders.StableDiffusionLoraLoaderMixin.load_lora_weights">load_lora_weights()</a>.`,name:"hotswap"},{anchor:"diffusers.loaders.SanaLoraLoaderMixin.load_lora_weights.kwargs",description:`<strong>kwargs</strong> (<code>dict</code>, <em>optional</em>) — | |
| See <a href="/docs/diffusers/pr_11477/en/api/loaders/lora#diffusers.loaders.StableDiffusionLoraLoaderMixin.lora_state_dict">lora_state_dict()</a>.`,name:"kwargs"}],source:"https://github.com/huggingface/diffusers/blob/vr_11477/src/diffusers/loaders/lora_pipeline.py#L3840"}}),Gt=new $({props:{name:"lora_state_dict",anchor:"diffusers.loaders.SanaLoraLoaderMixin.lora_state_dict",parameters:[{name:"pretrained_model_name_or_path_or_dict",val:": typing.Union[str, typing.Dict[str, torch.Tensor]]"},{name:"**kwargs",val:""}],parametersDescription:[{anchor:"diffusers.loaders.SanaLoraLoaderMixin.lora_state_dict.pretrained_model_name_or_path_or_dict",description:`<strong>pretrained_model_name_or_path_or_dict</strong> (<code>str</code> or <code>os.PathLike</code> or <code>dict</code>) — | |
| Can be either:</p> | |
| <ul> | |
| <li>A string, the <em>model id</em> (for example <code>google/ddpm-celebahq-256</code>) of a pretrained model hosted on | |
| the Hub.</li> | |
| <li>A path to a <em>directory</em> (for example <code>./my_model_directory</code>) containing the model weights saved | |
| with <a href="/docs/diffusers/pr_11477/en/api/models/overview#diffusers.ModelMixin.save_pretrained">ModelMixin.save_pretrained()</a>.</li> | |
| <li>A <a href="https://pytorch.org/tutorials/beginner/saving_loading_models.html#what-is-a-state-dict" rel="nofollow">torch state | |
| dict</a>.</li> | |
| </ul>`,name:"pretrained_model_name_or_path_or_dict"},{anchor:"diffusers.loaders.SanaLoraLoaderMixin.lora_state_dict.cache_dir",description:`<strong>cache_dir</strong> (<code>Union[str, os.PathLike]</code>, <em>optional</em>) — | |
| Path to a directory where a downloaded pretrained model configuration is cached if the standard cache | |
| is not used.`,name:"cache_dir"},{anchor:"diffusers.loaders.SanaLoraLoaderMixin.lora_state_dict.force_download",description:`<strong>force_download</strong> (<code>bool</code>, <em>optional</em>, defaults to <code>False</code>) — | |
| Whether or not to force the (re-)download of the model weights and configuration files, overriding the | |
| cached versions if they exist.`,name:"force_download"},{anchor:"diffusers.loaders.SanaLoraLoaderMixin.lora_state_dict.proxies",description:`<strong>proxies</strong> (<code>Dict[str, str]</code>, <em>optional</em>) — | |
| A dictionary of proxy servers to use by protocol or endpoint, for example, <code>{'http': 'foo.bar:3128', 'http://hostname': 'foo.bar:4012'}</code>. The proxies are used on each request.`,name:"proxies"},{anchor:"diffusers.loaders.SanaLoraLoaderMixin.lora_state_dict.local_files_only",description:`<strong>local_files_only</strong> (<code>bool</code>, <em>optional</em>, defaults to <code>False</code>) — | |
| Whether to only load local model weights and configuration files or not. If set to <code>True</code>, the model | |
| won’t be downloaded from the Hub.`,name:"local_files_only"},{anchor:"diffusers.loaders.SanaLoraLoaderMixin.lora_state_dict.token",description:`<strong>token</strong> (<code>str</code> or <em>bool</em>, <em>optional</em>) — | |
| The token to use as HTTP bearer authorization for remote files. If <code>True</code>, the token generated from | |
| <code>diffusers-cli login</code> (stored in <code>~/.huggingface</code>) is used.`,name:"token"},{anchor:"diffusers.loaders.SanaLoraLoaderMixin.lora_state_dict.revision",description:`<strong>revision</strong> (<code>str</code>, <em>optional</em>, defaults to <code>"main"</code>) — | |
| The specific model version to use. It can be a branch name, a tag name, a commit id, or any identifier | |
| allowed by Git.`,name:"revision"},{anchor:"diffusers.loaders.SanaLoraLoaderMixin.lora_state_dict.subfolder",description:`<strong>subfolder</strong> (<code>str</code>, <em>optional</em>, defaults to <code>""</code>) — | |
| The subfolder location of a model file within a larger model repository on the Hub or locally.`,name:"subfolder"}],source:"https://github.com/huggingface/diffusers/blob/vr_11477/src/diffusers/loaders/lora_pipeline.py#L3741"}}),Zo=new S({props:{warning:!0,$$slots:{default:[yh]},$$scope:{ctx:y}}}),Jt=new $({props:{name:"save_lora_weights",anchor:"diffusers.loaders.SanaLoraLoaderMixin.save_lora_weights",parameters:[{name:"save_directory",val:": typing.Union[str, os.PathLike]"},{name:"transformer_lora_layers",val:": typing.Dict[str, typing.Union[torch.nn.modules.module.Module, torch.Tensor]] = None"},{name:"is_main_process",val:": bool = True"},{name:"weight_name",val:": str = None"},{name:"save_function",val:": typing.Callable = None"},{name:"safe_serialization",val:": bool = True"}],parametersDescription:[{anchor:"diffusers.loaders.SanaLoraLoaderMixin.save_lora_weights.save_directory",description:`<strong>save_directory</strong> (<code>str</code> or <code>os.PathLike</code>) — | |
| Directory to save LoRA parameters to. Will be created if it doesn’t exist.`,name:"save_directory"},{anchor:"diffusers.loaders.SanaLoraLoaderMixin.save_lora_weights.transformer_lora_layers",description:`<strong>transformer_lora_layers</strong> (<code>Dict[str, torch.nn.Module]</code> or <code>Dict[str, torch.Tensor]</code>) — | |
| State dict of the LoRA layers corresponding to the <code>transformer</code>.`,name:"transformer_lora_layers"},{anchor:"diffusers.loaders.SanaLoraLoaderMixin.save_lora_weights.is_main_process",description:`<strong>is_main_process</strong> (<code>bool</code>, <em>optional</em>, defaults to <code>True</code>) — | |
| Whether the process calling this is the main process or not. Useful during distributed training and you | |
| need to call this function on all processes. In this case, set <code>is_main_process=True</code> only on the main | |
| process to avoid race conditions.`,name:"is_main_process"},{anchor:"diffusers.loaders.SanaLoraLoaderMixin.save_lora_weights.save_function",description:`<strong>save_function</strong> (<code>Callable</code>) — | |
| The function to use to save the state dictionary. Useful during distributed training when you need to | |
| replace <code>torch.save</code> with another method. Can be configured with the environment variable | |
| <code>DIFFUSERS_SAVE_MODE</code>.`,name:"save_function"},{anchor:"diffusers.loaders.SanaLoraLoaderMixin.save_lora_weights.safe_serialization",description:`<strong>safe_serialization</strong> (<code>bool</code>, <em>optional</em>, defaults to <code>True</code>) — | |
| Whether to save the model using <code>safetensors</code> or the traditional PyTorch way with <code>pickle</code>.`,name:"safe_serialization"}],source:"https://github.com/huggingface/diffusers/blob/vr_11477/src/diffusers/loaders/lora_pipeline.py#L3937"}}),Zt=new $({props:{name:"unfuse_lora",anchor:"diffusers.loaders.SanaLoraLoaderMixin.unfuse_lora",parameters:[{name:"components",val:": typing.List[str] = ['transformer']"},{name:"**kwargs",val:""}],parametersDescription:[{anchor:"diffusers.loaders.SanaLoraLoaderMixin.unfuse_lora.components",description:"<strong>components</strong> (<code>List[str]</code>) — List of LoRA-injectable components to unfuse LoRA from.",name:"components"},{anchor:"diffusers.loaders.SanaLoraLoaderMixin.unfuse_lora.unfuse_transformer",description:"<strong>unfuse_transformer</strong> (<code>bool</code>, defaults to <code>True</code>) — Whether to unfuse the UNet LoRA parameters.",name:"unfuse_transformer"}],source:"https://github.com/huggingface/diffusers/blob/vr_11477/src/diffusers/loaders/lora_pipeline.py#L4034"}}),Yo=new S({props:{warning:!0,$$slots:{default:[Th]},$$scope:{ctx:y}}}),Ot=new J({props:{title:"HunyuanVideoLoraLoaderMixin",local:"diffusers.loaders.HunyuanVideoLoraLoaderMixin",headingTag:"h2"}}),Yt=new $({props:{name:"class diffusers.loaders.HunyuanVideoLoraLoaderMixin",anchor:"diffusers.loaders.HunyuanVideoLoraLoaderMixin",parameters:[],source:"https://github.com/huggingface/diffusers/blob/vr_11477/src/diffusers/loaders/lora_pipeline.py#L4052"}}),Qt=new $({props:{name:"load_lora_into_transformer",anchor:"diffusers.loaders.HunyuanVideoLoraLoaderMixin.load_lora_into_transformer",parameters:[{name:"state_dict",val:""},{name:"transformer",val:""},{name:"adapter_name",val:" = None"},{name:"_pipeline",val:" = None"},{name:"low_cpu_mem_usage",val:" = False"},{name:"hotswap",val:": bool = False"}],parametersDescription:[{anchor:"diffusers.loaders.HunyuanVideoLoraLoaderMixin.load_lora_into_transformer.state_dict",description:`<strong>state_dict</strong> (<code>dict</code>) — | |
| A standard state dict containing the lora layer parameters. The keys can either be indexed directly | |
| into the unet or prefixed with an additional <code>unet</code> which can be used to distinguish between text | |
| encoder lora layers.`,name:"state_dict"},{anchor:"diffusers.loaders.HunyuanVideoLoraLoaderMixin.load_lora_into_transformer.transformer",description:`<strong>transformer</strong> (<code>HunyuanVideoTransformer3DModel</code>) — | |
| The Transformer model to load the LoRA layers into.`,name:"transformer"},{anchor:"diffusers.loaders.HunyuanVideoLoraLoaderMixin.load_lora_into_transformer.adapter_name",description:`<strong>adapter_name</strong> (<code>str</code>, <em>optional</em>) — | |
| Adapter name to be used for referencing the loaded adapter model. If not specified, it will use | |
| <code>default_{i}</code> where i is the total number of adapters being loaded.`,name:"adapter_name"},{anchor:"diffusers.loaders.HunyuanVideoLoraLoaderMixin.load_lora_into_transformer.low_cpu_mem_usage",description:`<strong>low_cpu_mem_usage</strong> (<code>bool</code>, <em>optional</em>) — | |
| Speed up model loading by only loading the pretrained LoRA weights and not initializing the random | |
| weights.`,name:"low_cpu_mem_usage"},{anchor:"diffusers.loaders.HunyuanVideoLoraLoaderMixin.load_lora_into_transformer.hotswap",description:`<strong>hotswap</strong> (<code>bool</code>, <em>optional</em>) — | |
| See <a href="/docs/diffusers/pr_11477/en/api/loaders/lora#diffusers.loaders.StableDiffusionLoraLoaderMixin.load_lora_weights">load_lora_weights()</a>.`,name:"hotswap"}],source:"https://github.com/huggingface/diffusers/blob/vr_11477/src/diffusers/loaders/lora_pipeline.py#L4219"}}),Kt=new $({props:{name:"load_lora_weights",anchor:"diffusers.loaders.HunyuanVideoLoraLoaderMixin.load_lora_weights",parameters:[{name:"pretrained_model_name_or_path_or_dict",val:": typing.Union[str, typing.Dict[str, torch.Tensor]]"},{name:"adapter_name",val:": typing.Optional[str] = None"},{name:"hotswap",val:": bool = False"},{name:"**kwargs",val:""}],parametersDescription:[{anchor:"diffusers.loaders.HunyuanVideoLoraLoaderMixin.load_lora_weights.pretrained_model_name_or_path_or_dict",description:`<strong>pretrained_model_name_or_path_or_dict</strong> (<code>str</code> or <code>os.PathLike</code> or <code>dict</code>) — | |
| See <a href="/docs/diffusers/pr_11477/en/api/loaders/lora#diffusers.loaders.StableDiffusionLoraLoaderMixin.lora_state_dict">lora_state_dict()</a>.`,name:"pretrained_model_name_or_path_or_dict"},{anchor:"diffusers.loaders.HunyuanVideoLoraLoaderMixin.load_lora_weights.adapter_name",description:`<strong>adapter_name</strong> (<code>str</code>, <em>optional</em>) — | |
| Adapter name to be used for referencing the loaded adapter model. If not specified, it will use | |
| <code>default_{i}</code> where i is the total number of adapters being loaded.`,name:"adapter_name"},{anchor:"diffusers.loaders.HunyuanVideoLoraLoaderMixin.load_lora_weights.low_cpu_mem_usage",description:`<strong>low_cpu_mem_usage</strong> (<code>bool</code>, <em>optional</em>) — | |
| Speed up model loading by only loading the pretrained LoRA weights and not initializing the random | |
| weights.`,name:"low_cpu_mem_usage"},{anchor:"diffusers.loaders.HunyuanVideoLoraLoaderMixin.load_lora_weights.hotswap",description:`<strong>hotswap</strong> (<code>bool</code>, <em>optional</em>) — | |
| See <a href="/docs/diffusers/pr_11477/en/api/loaders/lora#diffusers.loaders.StableDiffusionLoraLoaderMixin.load_lora_weights">load_lora_weights()</a>.`,name:"hotswap"},{anchor:"diffusers.loaders.HunyuanVideoLoraLoaderMixin.load_lora_weights.kwargs",description:`<strong>kwargs</strong> (<code>dict</code>, <em>optional</em>) — | |
| See <a href="/docs/diffusers/pr_11477/en/api/loaders/lora#diffusers.loaders.StableDiffusionLoraLoaderMixin.lora_state_dict">lora_state_dict()</a>.`,name:"kwargs"}],source:"https://github.com/huggingface/diffusers/blob/vr_11477/src/diffusers/loaders/lora_pipeline.py#L4162"}}),ea=new $({props:{name:"lora_state_dict",anchor:"diffusers.loaders.HunyuanVideoLoraLoaderMixin.lora_state_dict",parameters:[{name:"pretrained_model_name_or_path_or_dict",val:": typing.Union[str, typing.Dict[str, torch.Tensor]]"},{name:"**kwargs",val:""}],parametersDescription:[{anchor:"diffusers.loaders.HunyuanVideoLoraLoaderMixin.lora_state_dict.pretrained_model_name_or_path_or_dict",description:`<strong>pretrained_model_name_or_path_or_dict</strong> (<code>str</code> or <code>os.PathLike</code> or <code>dict</code>) — | |
| Can be either:</p> | |
| <ul> | |
| <li>A string, the <em>model id</em> (for example <code>google/ddpm-celebahq-256</code>) of a pretrained model hosted on | |
| the Hub.</li> | |
| <li>A path to a <em>directory</em> (for example <code>./my_model_directory</code>) containing the model weights saved | |
| with <a href="/docs/diffusers/pr_11477/en/api/models/overview#diffusers.ModelMixin.save_pretrained">ModelMixin.save_pretrained()</a>.</li> | |
| <li>A <a href="https://pytorch.org/tutorials/beginner/saving_loading_models.html#what-is-a-state-dict" rel="nofollow">torch state | |
| dict</a>.</li> | |
| </ul>`,name:"pretrained_model_name_or_path_or_dict"},{anchor:"diffusers.loaders.HunyuanVideoLoraLoaderMixin.lora_state_dict.cache_dir",description:`<strong>cache_dir</strong> (<code>Union[str, os.PathLike]</code>, <em>optional</em>) — | |
| Path to a directory where a downloaded pretrained model configuration is cached if the standard cache | |
| is not used.`,name:"cache_dir"},{anchor:"diffusers.loaders.HunyuanVideoLoraLoaderMixin.lora_state_dict.force_download",description:`<strong>force_download</strong> (<code>bool</code>, <em>optional</em>, defaults to <code>False</code>) — | |
| Whether or not to force the (re-)download of the model weights and configuration files, overriding the | |
| cached versions if they exist.`,name:"force_download"},{anchor:"diffusers.loaders.HunyuanVideoLoraLoaderMixin.lora_state_dict.proxies",description:`<strong>proxies</strong> (<code>Dict[str, str]</code>, <em>optional</em>) — | |
| A dictionary of proxy servers to use by protocol or endpoint, for example, <code>{'http': 'foo.bar:3128', 'http://hostname': 'foo.bar:4012'}</code>. The proxies are used on each request.`,name:"proxies"},{anchor:"diffusers.loaders.HunyuanVideoLoraLoaderMixin.lora_state_dict.local_files_only",description:`<strong>local_files_only</strong> (<code>bool</code>, <em>optional</em>, defaults to <code>False</code>) — | |
| Whether to only load local model weights and configuration files or not. If set to <code>True</code>, the model | |
| won’t be downloaded from the Hub.`,name:"local_files_only"},{anchor:"diffusers.loaders.HunyuanVideoLoraLoaderMixin.lora_state_dict.token",description:`<strong>token</strong> (<code>str</code> or <em>bool</em>, <em>optional</em>) — | |
| The token to use as HTTP bearer authorization for remote files. If <code>True</code>, the token generated from | |
| <code>diffusers-cli login</code> (stored in <code>~/.huggingface</code>) is used.`,name:"token"},{anchor:"diffusers.loaders.HunyuanVideoLoraLoaderMixin.lora_state_dict.revision",description:`<strong>revision</strong> (<code>str</code>, <em>optional</em>, defaults to <code>"main"</code>) — | |
| The specific model version to use. It can be a branch name, a tag name, a commit id, or any identifier | |
| allowed by Git.`,name:"revision"},{anchor:"diffusers.loaders.HunyuanVideoLoraLoaderMixin.lora_state_dict.subfolder",description:`<strong>subfolder</strong> (<code>str</code>, <em>optional</em>, defaults to <code>""</code>) — | |
| The subfolder location of a model file within a larger model repository on the Hub or locally.`,name:"subfolder"}],source:"https://github.com/huggingface/diffusers/blob/vr_11477/src/diffusers/loaders/lora_pipeline.py#L4060"}}),er=new S({props:{warning:!0,$$slots:{default:[Dh]},$$scope:{ctx:y}}}),oa=new $({props:{name:"save_lora_weights",anchor:"diffusers.loaders.HunyuanVideoLoraLoaderMixin.save_lora_weights",parameters:[{name:"save_directory",val:": typing.Union[str, os.PathLike]"},{name:"transformer_lora_layers",val:": typing.Dict[str, typing.Union[torch.nn.modules.module.Module, torch.Tensor]] = None"},{name:"is_main_process",val:": bool = True"},{name:"weight_name",val:": str = None"},{name:"save_function",val:": typing.Callable = None"},{name:"safe_serialization",val:": bool = True"}],parametersDescription:[{anchor:"diffusers.loaders.HunyuanVideoLoraLoaderMixin.save_lora_weights.save_directory",description:`<strong>save_directory</strong> (<code>str</code> or <code>os.PathLike</code>) — | |
| Directory to save LoRA parameters to. Will be created if it doesn’t exist.`,name:"save_directory"},{anchor:"diffusers.loaders.HunyuanVideoLoraLoaderMixin.save_lora_weights.transformer_lora_layers",description:`<strong>transformer_lora_layers</strong> (<code>Dict[str, torch.nn.Module]</code> or <code>Dict[str, torch.Tensor]</code>) — | |
| State dict of the LoRA layers corresponding to the <code>transformer</code>.`,name:"transformer_lora_layers"},{anchor:"diffusers.loaders.HunyuanVideoLoraLoaderMixin.save_lora_weights.is_main_process",description:`<strong>is_main_process</strong> (<code>bool</code>, <em>optional</em>, defaults to <code>True</code>) — | |
| Whether the process calling this is the main process or not. Useful during distributed training and you | |
| need to call this function on all processes. In this case, set <code>is_main_process=True</code> only on the main | |
| process to avoid race conditions.`,name:"is_main_process"},{anchor:"diffusers.loaders.HunyuanVideoLoraLoaderMixin.save_lora_weights.save_function",description:`<strong>save_function</strong> (<code>Callable</code>) — | |
| The function to use to save the state dictionary. Useful during distributed training when you need to | |
| replace <code>torch.save</code> with another method. Can be configured with the environment variable | |
| <code>DIFFUSERS_SAVE_MODE</code>.`,name:"save_function"},{anchor:"diffusers.loaders.HunyuanVideoLoraLoaderMixin.save_lora_weights.safe_serialization",description:`<strong>safe_serialization</strong> (<code>bool</code>, <em>optional</em>, defaults to <code>True</code>) — | |
| Whether to save the model using <code>safetensors</code> or the traditional PyTorch way with <code>pickle</code>.`,name:"safe_serialization"}],source:"https://github.com/huggingface/diffusers/blob/vr_11477/src/diffusers/loaders/lora_pipeline.py#L4259"}}),ra=new $({props:{name:"unfuse_lora",anchor:"diffusers.loaders.HunyuanVideoLoraLoaderMixin.unfuse_lora",parameters:[{name:"components",val:": typing.List[str] = ['transformer']"},{name:"**kwargs",val:""}],parametersDescription:[{anchor:"diffusers.loaders.HunyuanVideoLoraLoaderMixin.unfuse_lora.components",description:"<strong>components</strong> (<code>List[str]</code>) — List of LoRA-injectable components to unfuse LoRA from.",name:"components"},{anchor:"diffusers.loaders.HunyuanVideoLoraLoaderMixin.unfuse_lora.unfuse_transformer",description:"<strong>unfuse_transformer</strong> (<code>bool</code>, defaults to <code>True</code>) — Whether to unfuse the UNet LoRA parameters.",name:"unfuse_transformer"}],source:"https://github.com/huggingface/diffusers/blob/vr_11477/src/diffusers/loaders/lora_pipeline.py#L4356"}}),rr=new S({props:{warning:!0,$$slots:{default:[Sh]},$$scope:{ctx:y}}}),ta=new J({props:{title:"Lumina2LoraLoaderMixin",local:"diffusers.loaders.Lumina2LoraLoaderMixin",headingTag:"h2"}}),aa=new $({props:{name:"class diffusers.loaders.Lumina2LoraLoaderMixin",anchor:"diffusers.loaders.Lumina2LoraLoaderMixin",parameters:[],source:"https://github.com/huggingface/diffusers/blob/vr_11477/src/diffusers/loaders/lora_pipeline.py#L4374"}}),na=new $({props:{name:"load_lora_into_transformer",anchor:"diffusers.loaders.Lumina2LoraLoaderMixin.load_lora_into_transformer",parameters:[{name:"state_dict",val:""},{name:"transformer",val:""},{name:"adapter_name",val:" = None"},{name:"_pipeline",val:" = None"},{name:"low_cpu_mem_usage",val:" = False"},{name:"hotswap",val:": bool = False"}],parametersDescription:[{anchor:"diffusers.loaders.Lumina2LoraLoaderMixin.load_lora_into_transformer.state_dict",description:`<strong>state_dict</strong> (<code>dict</code>) — | |
| A standard state dict containing the lora layer parameters. The keys can either be indexed directly | |
| into the unet or prefixed with an additional <code>unet</code> which can be used to distinguish between text | |
| encoder lora layers.`,name:"state_dict"},{anchor:"diffusers.loaders.Lumina2LoraLoaderMixin.load_lora_into_transformer.transformer",description:`<strong>transformer</strong> (<code>Lumina2Transformer2DModel</code>) — | |
| The Transformer model to load the LoRA layers into.`,name:"transformer"},{anchor:"diffusers.loaders.Lumina2LoraLoaderMixin.load_lora_into_transformer.adapter_name",description:`<strong>adapter_name</strong> (<code>str</code>, <em>optional</em>) — | |
| Adapter name to be used for referencing the loaded adapter model. If not specified, it will use | |
| <code>default_{i}</code> where i is the total number of adapters being loaded.`,name:"adapter_name"},{anchor:"diffusers.loaders.Lumina2LoraLoaderMixin.load_lora_into_transformer.low_cpu_mem_usage",description:`<strong>low_cpu_mem_usage</strong> (<code>bool</code>, <em>optional</em>) — | |
| Speed up model loading by only loading the pretrained LoRA weights and not initializing the random | |
| weights.`,name:"low_cpu_mem_usage"},{anchor:"diffusers.loaders.Lumina2LoraLoaderMixin.load_lora_into_transformer.hotswap",description:`<strong>hotswap</strong> (<code>bool</code>, <em>optional</em>) — | |
| See <a href="/docs/diffusers/pr_11477/en/api/loaders/lora#diffusers.loaders.StableDiffusionLoraLoaderMixin.load_lora_weights">load_lora_weights()</a>.`,name:"hotswap"}],source:"https://github.com/huggingface/diffusers/blob/vr_11477/src/diffusers/loaders/lora_pipeline.py#L4542"}}),sa=new $({props:{name:"load_lora_weights",anchor:"diffusers.loaders.Lumina2LoraLoaderMixin.load_lora_weights",parameters:[{name:"pretrained_model_name_or_path_or_dict",val:": typing.Union[str, typing.Dict[str, torch.Tensor]]"},{name:"adapter_name",val:": typing.Optional[str] = None"},{name:"hotswap",val:": bool = False"},{name:"**kwargs",val:""}],parametersDescription:[{anchor:"diffusers.loaders.Lumina2LoraLoaderMixin.load_lora_weights.pretrained_model_name_or_path_or_dict",description:`<strong>pretrained_model_name_or_path_or_dict</strong> (<code>str</code> or <code>os.PathLike</code> or <code>dict</code>) — | |
| See <a href="/docs/diffusers/pr_11477/en/api/loaders/lora#diffusers.loaders.StableDiffusionLoraLoaderMixin.lora_state_dict">lora_state_dict()</a>.`,name:"pretrained_model_name_or_path_or_dict"},{anchor:"diffusers.loaders.Lumina2LoraLoaderMixin.load_lora_weights.adapter_name",description:`<strong>adapter_name</strong> (<code>str</code>, <em>optional</em>) — | |
| Adapter name to be used for referencing the loaded adapter model. If not specified, it will use | |
| <code>default_{i}</code> where i is the total number of adapters being loaded.`,name:"adapter_name"},{anchor:"diffusers.loaders.Lumina2LoraLoaderMixin.load_lora_weights.low_cpu_mem_usage",description:`<strong>low_cpu_mem_usage</strong> (<code>bool</code>, <em>optional</em>) — | |
| Speed up model loading by only loading the pretrained LoRA weights and not initializing the random | |
| weights.`,name:"low_cpu_mem_usage"},{anchor:"diffusers.loaders.Lumina2LoraLoaderMixin.load_lora_weights.hotswap",description:`<strong>hotswap</strong> (<code>bool</code>, <em>optional</em>) — | |
| See <a href="/docs/diffusers/pr_11477/en/api/loaders/lora#diffusers.loaders.StableDiffusionLoraLoaderMixin.load_lora_weights">load_lora_weights()</a>.`,name:"hotswap"},{anchor:"diffusers.loaders.Lumina2LoraLoaderMixin.load_lora_weights.kwargs",description:`<strong>kwargs</strong> (<code>dict</code>, <em>optional</em>) — | |
| See <a href="/docs/diffusers/pr_11477/en/api/loaders/lora#diffusers.loaders.StableDiffusionLoraLoaderMixin.lora_state_dict">lora_state_dict()</a>.`,name:"kwargs"}],source:"https://github.com/huggingface/diffusers/blob/vr_11477/src/diffusers/loaders/lora_pipeline.py#L4485"}}),ia=new $({props:{name:"lora_state_dict",anchor:"diffusers.loaders.Lumina2LoraLoaderMixin.lora_state_dict",parameters:[{name:"pretrained_model_name_or_path_or_dict",val:": typing.Union[str, typing.Dict[str, torch.Tensor]]"},{name:"**kwargs",val:""}],parametersDescription:[{anchor:"diffusers.loaders.Lumina2LoraLoaderMixin.lora_state_dict.pretrained_model_name_or_path_or_dict",description:`<strong>pretrained_model_name_or_path_or_dict</strong> (<code>str</code> or <code>os.PathLike</code> or <code>dict</code>) — | |
| Can be either:</p> | |
| <ul> | |
| <li>A string, the <em>model id</em> (for example <code>google/ddpm-celebahq-256</code>) of a pretrained model hosted on | |
| the Hub.</li> | |
| <li>A path to a <em>directory</em> (for example <code>./my_model_directory</code>) containing the model weights saved | |
| with <a href="/docs/diffusers/pr_11477/en/api/models/overview#diffusers.ModelMixin.save_pretrained">ModelMixin.save_pretrained()</a>.</li> | |
| <li>A <a href="https://pytorch.org/tutorials/beginner/saving_loading_models.html#what-is-a-state-dict" rel="nofollow">torch state | |
| dict</a>.</li> | |
| </ul>`,name:"pretrained_model_name_or_path_or_dict"},{anchor:"diffusers.loaders.Lumina2LoraLoaderMixin.lora_state_dict.cache_dir",description:`<strong>cache_dir</strong> (<code>Union[str, os.PathLike]</code>, <em>optional</em>) — | |
| Path to a directory where a downloaded pretrained model configuration is cached if the standard cache | |
| is not used.`,name:"cache_dir"},{anchor:"diffusers.loaders.Lumina2LoraLoaderMixin.lora_state_dict.force_download",description:`<strong>force_download</strong> (<code>bool</code>, <em>optional</em>, defaults to <code>False</code>) — | |
| Whether or not to force the (re-)download of the model weights and configuration files, overriding the | |
| cached versions if they exist.`,name:"force_download"},{anchor:"diffusers.loaders.Lumina2LoraLoaderMixin.lora_state_dict.proxies",description:`<strong>proxies</strong> (<code>Dict[str, str]</code>, <em>optional</em>) — | |
| A dictionary of proxy servers to use by protocol or endpoint, for example, <code>{'http': 'foo.bar:3128', 'http://hostname': 'foo.bar:4012'}</code>. The proxies are used on each request.`,name:"proxies"},{anchor:"diffusers.loaders.Lumina2LoraLoaderMixin.lora_state_dict.local_files_only",description:`<strong>local_files_only</strong> (<code>bool</code>, <em>optional</em>, defaults to <code>False</code>) — | |
| Whether to only load local model weights and configuration files or not. If set to <code>True</code>, the model | |
| won’t be downloaded from the Hub.`,name:"local_files_only"},{anchor:"diffusers.loaders.Lumina2LoraLoaderMixin.lora_state_dict.token",description:`<strong>token</strong> (<code>str</code> or <em>bool</em>, <em>optional</em>) — | |
| The token to use as HTTP bearer authorization for remote files. If <code>True</code>, the token generated from | |
| <code>diffusers-cli login</code> (stored in <code>~/.huggingface</code>) is used.`,name:"token"},{anchor:"diffusers.loaders.Lumina2LoraLoaderMixin.lora_state_dict.revision",description:`<strong>revision</strong> (<code>str</code>, <em>optional</em>, defaults to <code>"main"</code>) — | |
| The specific model version to use. It can be a branch name, a tag name, a commit id, or any identifier | |
| allowed by Git.`,name:"revision"},{anchor:"diffusers.loaders.Lumina2LoraLoaderMixin.lora_state_dict.subfolder",description:`<strong>subfolder</strong> (<code>str</code>, <em>optional</em>, defaults to <code>""</code>) — | |
| The subfolder location of a model file within a larger model repository on the Hub or locally.`,name:"subfolder"}],source:"https://github.com/huggingface/diffusers/blob/vr_11477/src/diffusers/loaders/lora_pipeline.py#L4382"}}),nr=new S({props:{warning:!0,$$slots:{default:[Ch]},$$scope:{ctx:y}}}),da=new $({props:{name:"save_lora_weights",anchor:"diffusers.loaders.Lumina2LoraLoaderMixin.save_lora_weights",parameters:[{name:"save_directory",val:": typing.Union[str, os.PathLike]"},{name:"transformer_lora_layers",val:": typing.Dict[str, typing.Union[torch.nn.modules.module.Module, torch.Tensor]] = None"},{name:"is_main_process",val:": bool = True"},{name:"weight_name",val:": str = None"},{name:"save_function",val:": typing.Callable = None"},{name:"safe_serialization",val:": bool = True"}],parametersDescription:[{anchor:"diffusers.loaders.Lumina2LoraLoaderMixin.save_lora_weights.save_directory",description:`<strong>save_directory</strong> (<code>str</code> or <code>os.PathLike</code>) — | |
| Directory to save LoRA parameters to. Will be created if it doesn’t exist.`,name:"save_directory"},{anchor:"diffusers.loaders.Lumina2LoraLoaderMixin.save_lora_weights.transformer_lora_layers",description:`<strong>transformer_lora_layers</strong> (<code>Dict[str, torch.nn.Module]</code> or <code>Dict[str, torch.Tensor]</code>) — | |
| State dict of the LoRA layers corresponding to the <code>transformer</code>.`,name:"transformer_lora_layers"},{anchor:"diffusers.loaders.Lumina2LoraLoaderMixin.save_lora_weights.is_main_process",description:`<strong>is_main_process</strong> (<code>bool</code>, <em>optional</em>, defaults to <code>True</code>) — | |
| Whether the process calling this is the main process or not. Useful during distributed training and you | |
| need to call this function on all processes. In this case, set <code>is_main_process=True</code> only on the main | |
| process to avoid race conditions.`,name:"is_main_process"},{anchor:"diffusers.loaders.Lumina2LoraLoaderMixin.save_lora_weights.save_function",description:`<strong>save_function</strong> (<code>Callable</code>) — | |
| The function to use to save the state dictionary. Useful during distributed training when you need to | |
| replace <code>torch.save</code> with another method. Can be configured with the environment variable | |
| <code>DIFFUSERS_SAVE_MODE</code>.`,name:"save_function"},{anchor:"diffusers.loaders.Lumina2LoraLoaderMixin.save_lora_weights.safe_serialization",description:`<strong>safe_serialization</strong> (<code>bool</code>, <em>optional</em>, defaults to <code>True</code>) — | |
| Whether to save the model using <code>safetensors</code> or the traditional PyTorch way with <code>pickle</code>.`,name:"safe_serialization"}],source:"https://github.com/huggingface/diffusers/blob/vr_11477/src/diffusers/loaders/lora_pipeline.py#L4582"}}),la=new $({props:{name:"unfuse_lora",anchor:"diffusers.loaders.Lumina2LoraLoaderMixin.unfuse_lora",parameters:[{name:"components",val:": typing.List[str] = ['transformer']"},{name:"**kwargs",val:""}],parametersDescription:[{anchor:"diffusers.loaders.Lumina2LoraLoaderMixin.unfuse_lora.components",description:"<strong>components</strong> (<code>List[str]</code>) — List of LoRA-injectable components to unfuse LoRA from.",name:"components"},{anchor:"diffusers.loaders.Lumina2LoraLoaderMixin.unfuse_lora.unfuse_transformer",description:"<strong>unfuse_transformer</strong> (<code>bool</code>, defaults to <code>True</code>) — Whether to unfuse the UNet LoRA parameters.",name:"unfuse_transformer"}],source:"https://github.com/huggingface/diffusers/blob/vr_11477/src/diffusers/loaders/lora_pipeline.py#L4679"}}),ir=new S({props:{warning:!0,$$slots:{default:[Ah]},$$scope:{ctx:y}}}),ca=new J({props:{title:"CogView4LoraLoaderMixin",local:"diffusers.loaders.CogView4LoraLoaderMixin",headingTag:"h2"}}),fa=new $({props:{name:"class diffusers.loaders.CogView4LoraLoaderMixin",anchor:"diffusers.loaders.CogView4LoraLoaderMixin",parameters:[],source:"https://github.com/huggingface/diffusers/blob/vr_11477/src/diffusers/loaders/lora_pipeline.py#L5070"}}),pa=new $({props:{name:"load_lora_into_transformer",anchor:"diffusers.loaders.CogView4LoraLoaderMixin.load_lora_into_transformer",parameters:[{name:"state_dict",val:""},{name:"transformer",val:""},{name:"adapter_name",val:" = None"},{name:"_pipeline",val:" = None"},{name:"low_cpu_mem_usage",val:" = False"},{name:"hotswap",val:": bool = False"}],parametersDescription:[{anchor:"diffusers.loaders.CogView4LoraLoaderMixin.load_lora_into_transformer.state_dict",description:`<strong>state_dict</strong> (<code>dict</code>) — | |
| A standard state dict containing the lora layer parameters. The keys can either be indexed directly | |
| into the unet or prefixed with an additional <code>unet</code> which can be used to distinguish between text | |
| encoder lora layers.`,name:"state_dict"},{anchor:"diffusers.loaders.CogView4LoraLoaderMixin.load_lora_into_transformer.transformer",description:`<strong>transformer</strong> (<code>CogView4Transformer2DModel</code>) — | |
| The Transformer model to load the LoRA layers into.`,name:"transformer"},{anchor:"diffusers.loaders.CogView4LoraLoaderMixin.load_lora_into_transformer.adapter_name",description:`<strong>adapter_name</strong> (<code>str</code>, <em>optional</em>) — | |
| Adapter name to be used for referencing the loaded adapter model. If not specified, it will use | |
| <code>default_{i}</code> where i is the total number of adapters being loaded.`,name:"adapter_name"},{anchor:"diffusers.loaders.CogView4LoraLoaderMixin.load_lora_into_transformer.low_cpu_mem_usage",description:`<strong>low_cpu_mem_usage</strong> (<code>bool</code>, <em>optional</em>) — | |
| Speed up model loading by only loading the pretrained LoRA weights and not initializing the random | |
| weights.`,name:"low_cpu_mem_usage"},{anchor:"diffusers.loaders.CogView4LoraLoaderMixin.load_lora_into_transformer.hotswap",description:`<strong>hotswap</strong> (<code>bool</code>, <em>optional</em>) — | |
| See <a href="/docs/diffusers/pr_11477/en/api/loaders/lora#diffusers.loaders.StableDiffusionLoraLoaderMixin.load_lora_weights">load_lora_weights()</a>.`,name:"hotswap"}],source:"https://github.com/huggingface/diffusers/blob/vr_11477/src/diffusers/loaders/lora_pipeline.py#L5234"}}),ma=new $({props:{name:"load_lora_weights",anchor:"diffusers.loaders.CogView4LoraLoaderMixin.load_lora_weights",parameters:[{name:"pretrained_model_name_or_path_or_dict",val:": typing.Union[str, typing.Dict[str, torch.Tensor]]"},{name:"adapter_name",val:": typing.Optional[str] = None"},{name:"hotswap",val:": bool = False"},{name:"**kwargs",val:""}],parametersDescription:[{anchor:"diffusers.loaders.CogView4LoraLoaderMixin.load_lora_weights.pretrained_model_name_or_path_or_dict",description:`<strong>pretrained_model_name_or_path_or_dict</strong> (<code>str</code> or <code>os.PathLike</code> or <code>dict</code>) — | |
| See <a href="/docs/diffusers/pr_11477/en/api/loaders/lora#diffusers.loaders.StableDiffusionLoraLoaderMixin.lora_state_dict">lora_state_dict()</a>.`,name:"pretrained_model_name_or_path_or_dict"},{anchor:"diffusers.loaders.CogView4LoraLoaderMixin.load_lora_weights.adapter_name",description:`<strong>adapter_name</strong> (<code>str</code>, <em>optional</em>) — | |
| Adapter name to be used for referencing the loaded adapter model. If not specified, it will use | |
| <code>default_{i}</code> where i is the total number of adapters being loaded.`,name:"adapter_name"},{anchor:"diffusers.loaders.CogView4LoraLoaderMixin.load_lora_weights.low_cpu_mem_usage",description:`<strong>low_cpu_mem_usage</strong> (<code>bool</code>, <em>optional</em>) — | |
| Speed up model loading by only loading the pretrained LoRA weights and not initializing the random | |
| weights.`,name:"low_cpu_mem_usage"},{anchor:"diffusers.loaders.CogView4LoraLoaderMixin.load_lora_weights.hotswap",description:`<strong>hotswap</strong> (<code>bool</code>, <em>optional</em>) — | |
| See <a href="/docs/diffusers/pr_11477/en/api/loaders/lora#diffusers.loaders.StableDiffusionLoraLoaderMixin.load_lora_weights">load_lora_weights()</a>.`,name:"hotswap"},{anchor:"diffusers.loaders.CogView4LoraLoaderMixin.load_lora_weights.kwargs",description:`<strong>kwargs</strong> (<code>dict</code>, <em>optional</em>) — | |
| See <a href="/docs/diffusers/pr_11477/en/api/loaders/lora#diffusers.loaders.StableDiffusionLoraLoaderMixin.lora_state_dict">lora_state_dict()</a>.`,name:"kwargs"}],source:"https://github.com/huggingface/diffusers/blob/vr_11477/src/diffusers/loaders/lora_pipeline.py#L5177"}}),_a=new $({props:{name:"lora_state_dict",anchor:"diffusers.loaders.CogView4LoraLoaderMixin.lora_state_dict",parameters:[{name:"pretrained_model_name_or_path_or_dict",val:": typing.Union[str, typing.Dict[str, torch.Tensor]]"},{name:"**kwargs",val:""}],parametersDescription:[{anchor:"diffusers.loaders.CogView4LoraLoaderMixin.lora_state_dict.pretrained_model_name_or_path_or_dict",description:`<strong>pretrained_model_name_or_path_or_dict</strong> (<code>str</code> or <code>os.PathLike</code> or <code>dict</code>) — | |
| Can be either:</p> | |
| <ul> | |
| <li>A string, the <em>model id</em> (for example <code>google/ddpm-celebahq-256</code>) of a pretrained model hosted on | |
| the Hub.</li> | |
| <li>A path to a <em>directory</em> (for example <code>./my_model_directory</code>) containing the model weights saved | |
| with <a href="/docs/diffusers/pr_11477/en/api/models/overview#diffusers.ModelMixin.save_pretrained">ModelMixin.save_pretrained()</a>.</li> | |
| <li>A <a href="https://pytorch.org/tutorials/beginner/saving_loading_models.html#what-is-a-state-dict" rel="nofollow">torch state | |
| dict</a>.</li> | |
| </ul>`,name:"pretrained_model_name_or_path_or_dict"},{anchor:"diffusers.loaders.CogView4LoraLoaderMixin.lora_state_dict.cache_dir",description:`<strong>cache_dir</strong> (<code>Union[str, os.PathLike]</code>, <em>optional</em>) — | |
| Path to a directory where a downloaded pretrained model configuration is cached if the standard cache | |
| is not used.`,name:"cache_dir"},{anchor:"diffusers.loaders.CogView4LoraLoaderMixin.lora_state_dict.force_download",description:`<strong>force_download</strong> (<code>bool</code>, <em>optional</em>, defaults to <code>False</code>) — | |
| Whether or not to force the (re-)download of the model weights and configuration files, overriding the | |
| cached versions if they exist.`,name:"force_download"},{anchor:"diffusers.loaders.CogView4LoraLoaderMixin.lora_state_dict.proxies",description:`<strong>proxies</strong> (<code>Dict[str, str]</code>, <em>optional</em>) — | |
| A dictionary of proxy servers to use by protocol or endpoint, for example, <code>{'http': 'foo.bar:3128', 'http://hostname': 'foo.bar:4012'}</code>. The proxies are used on each request.`,name:"proxies"},{anchor:"diffusers.loaders.CogView4LoraLoaderMixin.lora_state_dict.local_files_only",description:`<strong>local_files_only</strong> (<code>bool</code>, <em>optional</em>, defaults to <code>False</code>) — | |
| Whether to only load local model weights and configuration files or not. If set to <code>True</code>, the model | |
| won’t be downloaded from the Hub.`,name:"local_files_only"},{anchor:"diffusers.loaders.CogView4LoraLoaderMixin.lora_state_dict.token",description:`<strong>token</strong> (<code>str</code> or <em>bool</em>, <em>optional</em>) — | |
| The token to use as HTTP bearer authorization for remote files. If <code>True</code>, the token generated from | |
| <code>diffusers-cli login</code> (stored in <code>~/.huggingface</code>) is used.`,name:"token"},{anchor:"diffusers.loaders.CogView4LoraLoaderMixin.lora_state_dict.revision",description:`<strong>revision</strong> (<code>str</code>, <em>optional</em>, defaults to <code>"main"</code>) — | |
| The specific model version to use. It can be a branch name, a tag name, a commit id, or any identifier | |
| allowed by Git.`,name:"revision"},{anchor:"diffusers.loaders.CogView4LoraLoaderMixin.lora_state_dict.subfolder",description:`<strong>subfolder</strong> (<code>str</code>, <em>optional</em>, defaults to <code>""</code>) — | |
| The subfolder location of a model file within a larger model repository on the Hub or locally.`,name:"subfolder"}],source:"https://github.com/huggingface/diffusers/blob/vr_11477/src/diffusers/loaders/lora_pipeline.py#L5078"}}),cr=new S({props:{warning:!0,$$slots:{default:[kh]},$$scope:{ctx:y}}}),ua=new $({props:{name:"save_lora_weights",anchor:"diffusers.loaders.CogView4LoraLoaderMixin.save_lora_weights",parameters:[{name:"save_directory",val:": typing.Union[str, os.PathLike]"},{name:"transformer_lora_layers",val:": typing.Dict[str, typing.Union[torch.nn.modules.module.Module, torch.Tensor]] = None"},{name:"is_main_process",val:": bool = True"},{name:"weight_name",val:": str = None"},{name:"save_function",val:": typing.Callable = None"},{name:"safe_serialization",val:": bool = True"}],parametersDescription:[{anchor:"diffusers.loaders.CogView4LoraLoaderMixin.save_lora_weights.save_directory",description:`<strong>save_directory</strong> (<code>str</code> or <code>os.PathLike</code>) — | |
| Directory to save LoRA parameters to. Will be created if it doesn’t exist.`,name:"save_directory"},{anchor:"diffusers.loaders.CogView4LoraLoaderMixin.save_lora_weights.transformer_lora_layers",description:`<strong>transformer_lora_layers</strong> (<code>Dict[str, torch.nn.Module]</code> or <code>Dict[str, torch.Tensor]</code>) — | |
| State dict of the LoRA layers corresponding to the <code>transformer</code>.`,name:"transformer_lora_layers"},{anchor:"diffusers.loaders.CogView4LoraLoaderMixin.save_lora_weights.is_main_process",description:`<strong>is_main_process</strong> (<code>bool</code>, <em>optional</em>, defaults to <code>True</code>) — | |
| Whether the process calling this is the main process or not. Useful during distributed training and you | |
| need to call this function on all processes. In this case, set <code>is_main_process=True</code> only on the main | |
| process to avoid race conditions.`,name:"is_main_process"},{anchor:"diffusers.loaders.CogView4LoraLoaderMixin.save_lora_weights.save_function",description:`<strong>save_function</strong> (<code>Callable</code>) — | |
| The function to use to save the state dictionary. Useful during distributed training when you need to | |
| replace <code>torch.save</code> with another method. Can be configured with the environment variable | |
| <code>DIFFUSERS_SAVE_MODE</code>.`,name:"save_function"},{anchor:"diffusers.loaders.CogView4LoraLoaderMixin.save_lora_weights.safe_serialization",description:`<strong>safe_serialization</strong> (<code>bool</code>, <em>optional</em>, defaults to <code>True</code>) — | |
| Whether to save the model using <code>safetensors</code> or the traditional PyTorch way with <code>pickle</code>.`,name:"safe_serialization"}],source:"https://github.com/huggingface/diffusers/blob/vr_11477/src/diffusers/loaders/lora_pipeline.py#L5274"}}),ha=new $({props:{name:"unfuse_lora",anchor:"diffusers.loaders.CogView4LoraLoaderMixin.unfuse_lora",parameters:[{name:"components",val:": typing.List[str] = ['transformer']"},{name:"**kwargs",val:""}],parametersDescription:[{anchor:"diffusers.loaders.CogView4LoraLoaderMixin.unfuse_lora.components",description:"<strong>components</strong> (<code>List[str]</code>) — List of LoRA-injectable components to unfuse LoRA from.",name:"components"},{anchor:"diffusers.loaders.CogView4LoraLoaderMixin.unfuse_lora.unfuse_transformer",description:"<strong>unfuse_transformer</strong> (<code>bool</code>, defaults to <code>True</code>) — Whether to unfuse the UNet LoRA parameters.",name:"unfuse_transformer"}],source:"https://github.com/huggingface/diffusers/blob/vr_11477/src/diffusers/loaders/lora_pipeline.py#L5371"}}),pr=new S({props:{warning:!0,$$slots:{default:[Ph]},$$scope:{ctx:y}}}),ga=new J({props:{title:"WanLoraLoaderMixin",local:"diffusers.loaders.WanLoraLoaderMixin",headingTag:"h2"}}),La=new $({props:{name:"class diffusers.loaders.WanLoraLoaderMixin",anchor:"diffusers.loaders.WanLoraLoaderMixin",parameters:[],source:"https://github.com/huggingface/diffusers/blob/vr_11477/src/diffusers/loaders/lora_pipeline.py#L4697"}}),xa=new $({props:{name:"load_lora_into_transformer",anchor:"diffusers.loaders.WanLoraLoaderMixin.load_lora_into_transformer",parameters:[{name:"state_dict",val:""},{name:"transformer",val:""},{name:"adapter_name",val:" = None"},{name:"_pipeline",val:" = None"},{name:"low_cpu_mem_usage",val:" = False"},{name:"hotswap",val:": bool = False"}],parametersDescription:[{anchor:"diffusers.loaders.WanLoraLoaderMixin.load_lora_into_transformer.state_dict",description:`<strong>state_dict</strong> (<code>dict</code>) — | |
| A standard state dict containing the lora layer parameters. The keys can either be indexed directly | |
| into the unet or prefixed with an additional <code>unet</code> which can be used to distinguish between text | |
| encoder lora layers.`,name:"state_dict"},{anchor:"diffusers.loaders.WanLoraLoaderMixin.load_lora_into_transformer.transformer",description:`<strong>transformer</strong> (<code>WanTransformer3DModel</code>) — | |
| The Transformer model to load the LoRA layers into.`,name:"transformer"},{anchor:"diffusers.loaders.WanLoraLoaderMixin.load_lora_into_transformer.adapter_name",description:`<strong>adapter_name</strong> (<code>str</code>, <em>optional</em>) — | |
| Adapter name to be used for referencing the loaded adapter model. If not specified, it will use | |
| <code>default_{i}</code> where i is the total number of adapters being loaded.`,name:"adapter_name"},{anchor:"diffusers.loaders.WanLoraLoaderMixin.load_lora_into_transformer.low_cpu_mem_usage",description:`<strong>low_cpu_mem_usage</strong> (<code>bool</code>, <em>optional</em>) — | |
| Speed up model loading by only loading the pretrained LoRA weights and not initializing the random | |
| weights.`,name:"low_cpu_mem_usage"},{anchor:"diffusers.loaders.WanLoraLoaderMixin.load_lora_into_transformer.hotswap",description:`<strong>hotswap</strong> (<code>bool</code>, <em>optional</em>) — | |
| See <a href="/docs/diffusers/pr_11477/en/api/loaders/lora#diffusers.loaders.StableDiffusionLoraLoaderMixin.load_lora_weights">load_lora_weights()</a>.`,name:"hotswap"}],source:"https://github.com/huggingface/diffusers/blob/vr_11477/src/diffusers/loaders/lora_pipeline.py#L4915"}}),ba=new $({props:{name:"load_lora_weights",anchor:"diffusers.loaders.WanLoraLoaderMixin.load_lora_weights",parameters:[{name:"pretrained_model_name_or_path_or_dict",val:": typing.Union[str, typing.Dict[str, torch.Tensor]]"},{name:"adapter_name",val:": typing.Optional[str] = None"},{name:"hotswap",val:": bool = False"},{name:"**kwargs",val:""}],parametersDescription:[{anchor:"diffusers.loaders.WanLoraLoaderMixin.load_lora_weights.pretrained_model_name_or_path_or_dict",description:`<strong>pretrained_model_name_or_path_or_dict</strong> (<code>str</code> or <code>os.PathLike</code> or <code>dict</code>) — | |
| See <a href="/docs/diffusers/pr_11477/en/api/loaders/lora#diffusers.loaders.StableDiffusionLoraLoaderMixin.lora_state_dict">lora_state_dict()</a>.`,name:"pretrained_model_name_or_path_or_dict"},{anchor:"diffusers.loaders.WanLoraLoaderMixin.load_lora_weights.adapter_name",description:`<strong>adapter_name</strong> (<code>str</code>, <em>optional</em>) — | |
| Adapter name to be used for referencing the loaded adapter model. If not specified, it will use | |
| <code>default_{i}</code> where i is the total number of adapters being loaded.`,name:"adapter_name"},{anchor:"diffusers.loaders.WanLoraLoaderMixin.load_lora_weights.low_cpu_mem_usage",description:`<strong>low_cpu_mem_usage</strong> (<code>bool</code>, <em>optional</em>) — | |
| Speed up model loading by only loading the pretrained LoRA weights and not initializing the random | |
| weights.`,name:"low_cpu_mem_usage"},{anchor:"diffusers.loaders.WanLoraLoaderMixin.load_lora_weights.hotswap",description:`<strong>hotswap</strong> (<code>bool</code>, <em>optional</em>) — | |
| See <a href="/docs/diffusers/pr_11477/en/api/loaders/lora#diffusers.loaders.StableDiffusionLoraLoaderMixin.load_lora_weights">load_lora_weights()</a>.`,name:"hotswap"},{anchor:"diffusers.loaders.WanLoraLoaderMixin.load_lora_weights.kwargs",description:`<strong>kwargs</strong> (<code>dict</code>, <em>optional</em>) — | |
| See <a href="/docs/diffusers/pr_11477/en/api/loaders/lora#diffusers.loaders.StableDiffusionLoraLoaderMixin.lora_state_dict">lora_state_dict()</a>.`,name:"kwargs"}],source:"https://github.com/huggingface/diffusers/blob/vr_11477/src/diffusers/loaders/lora_pipeline.py#L4854"}}),va=new $({props:{name:"lora_state_dict",anchor:"diffusers.loaders.WanLoraLoaderMixin.lora_state_dict",parameters:[{name:"pretrained_model_name_or_path_or_dict",val:": typing.Union[str, typing.Dict[str, torch.Tensor]]"},{name:"**kwargs",val:""}],parametersDescription:[{anchor:"diffusers.loaders.WanLoraLoaderMixin.lora_state_dict.pretrained_model_name_or_path_or_dict",description:`<strong>pretrained_model_name_or_path_or_dict</strong> (<code>str</code> or <code>os.PathLike</code> or <code>dict</code>) — | |
| Can be either:</p> | |
| <ul> | |
| <li>A string, the <em>model id</em> (for example <code>google/ddpm-celebahq-256</code>) of a pretrained model hosted on | |
| the Hub.</li> | |
| <li>A path to a <em>directory</em> (for example <code>./my_model_directory</code>) containing the model weights saved | |
| with <a href="/docs/diffusers/pr_11477/en/api/models/overview#diffusers.ModelMixin.save_pretrained">ModelMixin.save_pretrained()</a>.</li> | |
| <li>A <a href="https://pytorch.org/tutorials/beginner/saving_loading_models.html#what-is-a-state-dict" rel="nofollow">torch state | |
| dict</a>.</li> | |
| </ul>`,name:"pretrained_model_name_or_path_or_dict"},{anchor:"diffusers.loaders.WanLoraLoaderMixin.lora_state_dict.cache_dir",description:`<strong>cache_dir</strong> (<code>Union[str, os.PathLike]</code>, <em>optional</em>) — | |
| Path to a directory where a downloaded pretrained model configuration is cached if the standard cache | |
| is not used.`,name:"cache_dir"},{anchor:"diffusers.loaders.WanLoraLoaderMixin.lora_state_dict.force_download",description:`<strong>force_download</strong> (<code>bool</code>, <em>optional</em>, defaults to <code>False</code>) — | |
| Whether or not to force the (re-)download of the model weights and configuration files, overriding the | |
| cached versions if they exist.`,name:"force_download"},{anchor:"diffusers.loaders.WanLoraLoaderMixin.lora_state_dict.proxies",description:`<strong>proxies</strong> (<code>Dict[str, str]</code>, <em>optional</em>) — | |
| A dictionary of proxy servers to use by protocol or endpoint, for example, <code>{'http': 'foo.bar:3128', 'http://hostname': 'foo.bar:4012'}</code>. The proxies are used on each request.`,name:"proxies"},{anchor:"diffusers.loaders.WanLoraLoaderMixin.lora_state_dict.local_files_only",description:`<strong>local_files_only</strong> (<code>bool</code>, <em>optional</em>, defaults to <code>False</code>) — | |
| Whether to only load local model weights and configuration files or not. If set to <code>True</code>, the model | |
| won’t be downloaded from the Hub.`,name:"local_files_only"},{anchor:"diffusers.loaders.WanLoraLoaderMixin.lora_state_dict.token",description:`<strong>token</strong> (<code>str</code> or <em>bool</em>, <em>optional</em>) — | |
| The token to use as HTTP bearer authorization for remote files. If <code>True</code>, the token generated from | |
| <code>diffusers-cli login</code> (stored in <code>~/.huggingface</code>) is used.`,name:"token"},{anchor:"diffusers.loaders.WanLoraLoaderMixin.lora_state_dict.revision",description:`<strong>revision</strong> (<code>str</code>, <em>optional</em>, defaults to <code>"main"</code>) — | |
| The specific model version to use. It can be a branch name, a tag name, a commit id, or any identifier | |
| allowed by Git.`,name:"revision"},{anchor:"diffusers.loaders.WanLoraLoaderMixin.lora_state_dict.subfolder",description:`<strong>subfolder</strong> (<code>str</code>, <em>optional</em>, defaults to <code>""</code>) — | |
| The subfolder location of a model file within a larger model repository on the Hub or locally.`,name:"subfolder"}],source:"https://github.com/huggingface/diffusers/blob/vr_11477/src/diffusers/loaders/lora_pipeline.py#L4705"}}),ur=new S({props:{warning:!0,$$slots:{default:[Rh]},$$scope:{ctx:y}}}),wa=new $({props:{name:"save_lora_weights",anchor:"diffusers.loaders.WanLoraLoaderMixin.save_lora_weights",parameters:[{name:"save_directory",val:": typing.Union[str, os.PathLike]"},{name:"transformer_lora_layers",val:": typing.Dict[str, typing.Union[torch.nn.modules.module.Module, torch.Tensor]] = None"},{name:"is_main_process",val:": bool = True"},{name:"weight_name",val:": str = None"},{name:"save_function",val:": typing.Callable = None"},{name:"safe_serialization",val:": bool = True"}],parametersDescription:[{anchor:"diffusers.loaders.WanLoraLoaderMixin.save_lora_weights.save_directory",description:`<strong>save_directory</strong> (<code>str</code> or <code>os.PathLike</code>) — | |
| Directory to save LoRA parameters to. Will be created if it doesn’t exist.`,name:"save_directory"},{anchor:"diffusers.loaders.WanLoraLoaderMixin.save_lora_weights.transformer_lora_layers",description:`<strong>transformer_lora_layers</strong> (<code>Dict[str, torch.nn.Module]</code> or <code>Dict[str, torch.Tensor]</code>) — | |
| State dict of the LoRA layers corresponding to the <code>transformer</code>.`,name:"transformer_lora_layers"},{anchor:"diffusers.loaders.WanLoraLoaderMixin.save_lora_weights.is_main_process",description:`<strong>is_main_process</strong> (<code>bool</code>, <em>optional</em>, defaults to <code>True</code>) — | |
| Whether the process calling this is the main process or not. Useful during distributed training and you | |
| need to call this function on all processes. In this case, set <code>is_main_process=True</code> only on the main | |
| process to avoid race conditions.`,name:"is_main_process"},{anchor:"diffusers.loaders.WanLoraLoaderMixin.save_lora_weights.save_function",description:`<strong>save_function</strong> (<code>Callable</code>) — | |
| The function to use to save the state dictionary. Useful during distributed training when you need to | |
| replace <code>torch.save</code> with another method. Can be configured with the environment variable | |
| <code>DIFFUSERS_SAVE_MODE</code>.`,name:"save_function"},{anchor:"diffusers.loaders.WanLoraLoaderMixin.save_lora_weights.safe_serialization",description:`<strong>safe_serialization</strong> (<code>bool</code>, <em>optional</em>, defaults to <code>True</code>) — | |
| Whether to save the model using <code>safetensors</code> or the traditional PyTorch way with <code>pickle</code>.`,name:"safe_serialization"}],source:"https://github.com/huggingface/diffusers/blob/vr_11477/src/diffusers/loaders/lora_pipeline.py#L4955"}}),$a=new $({props:{name:"unfuse_lora",anchor:"diffusers.loaders.WanLoraLoaderMixin.unfuse_lora",parameters:[{name:"components",val:": typing.List[str] = ['transformer']"},{name:"**kwargs",val:""}],parametersDescription:[{anchor:"diffusers.loaders.WanLoraLoaderMixin.unfuse_lora.components",description:"<strong>components</strong> (<code>List[str]</code>) — List of LoRA-injectable components to unfuse LoRA from.",name:"components"},{anchor:"diffusers.loaders.WanLoraLoaderMixin.unfuse_lora.unfuse_transformer",description:"<strong>unfuse_transformer</strong> (<code>bool</code>, defaults to <code>True</code>) — Whether to unfuse the UNet LoRA parameters.",name:"unfuse_transformer"}],source:"https://github.com/huggingface/diffusers/blob/vr_11477/src/diffusers/loaders/lora_pipeline.py#L5052"}}),gr=new S({props:{warning:!0,$$slots:{default:[Hh]},$$scope:{ctx:y}}}),Ma=new J({props:{title:"AmusedLoraLoaderMixin",local:"diffusers.loaders.AmusedLoraLoaderMixin",headingTag:"h2"}}),ya=new $({props:{name:"class diffusers.loaders.AmusedLoraLoaderMixin",anchor:"diffusers.loaders.AmusedLoraLoaderMixin",parameters:[],source:"https://github.com/huggingface/diffusers/blob/vr_11477/src/diffusers/loaders/lora_pipeline.py#L2611"}}),Ta=new $({props:{name:"load_lora_into_transformer",anchor:"diffusers.loaders.AmusedLoraLoaderMixin.load_lora_into_transformer",parameters:[{name:"state_dict",val:""},{name:"network_alphas",val:""},{name:"transformer",val:""},{name:"adapter_name",val:" = None"},{name:"_pipeline",val:" = None"},{name:"low_cpu_mem_usage",val:" = False"},{name:"hotswap",val:": bool = False"}],parametersDescription:[{anchor:"diffusers.loaders.AmusedLoraLoaderMixin.load_lora_into_transformer.state_dict",description:`<strong>state_dict</strong> (<code>dict</code>) — | |
| A standard state dict containing the lora layer parameters. The keys can either be indexed directly | |
| into the unet or prefixed with an additional <code>unet</code> which can be used to distinguish between text | |
| encoder lora layers.`,name:"state_dict"},{anchor:"diffusers.loaders.AmusedLoraLoaderMixin.load_lora_into_transformer.network_alphas",description:`<strong>network_alphas</strong> (<code>Dict[str, float]</code>) — | |
| The value of the network alpha used for stable learning and preventing underflow. This value has the | |
| same meaning as the <code>--network_alpha</code> option in the kohya-ss trainer script. Refer to <a href="https://github.com/darkstorm2150/sd-scripts/blob/main/docs/train_network_README-en.md#execute-learning" rel="nofollow">this | |
| link</a>.`,name:"network_alphas"},{anchor:"diffusers.loaders.AmusedLoraLoaderMixin.load_lora_into_transformer.transformer",description:`<strong>transformer</strong> (<code>UVit2DModel</code>) — | |
| The Transformer model to load the LoRA layers into.`,name:"transformer"},{anchor:"diffusers.loaders.AmusedLoraLoaderMixin.load_lora_into_transformer.adapter_name",description:`<strong>adapter_name</strong> (<code>str</code>, <em>optional</em>) — | |
| Adapter name to be used for referencing the loaded adapter model. If not specified, it will use | |
| <code>default_{i}</code> where i is the total number of adapters being loaded.`,name:"adapter_name"},{anchor:"diffusers.loaders.AmusedLoraLoaderMixin.load_lora_into_transformer.low_cpu_mem_usage",description:`<strong>low_cpu_mem_usage</strong> (<code>bool</code>, <em>optional</em>) — | |
| Speed up model loading by only loading the pretrained LoRA weights and not initializing the random | |
| weights.`,name:"low_cpu_mem_usage"},{anchor:"diffusers.loaders.AmusedLoraLoaderMixin.load_lora_into_transformer.hotswap",description:`<strong>hotswap</strong> (<code>bool</code>, <em>optional</em>) — | |
| See <a href="/docs/diffusers/pr_11477/en/api/loaders/lora#diffusers.loaders.StableDiffusionLoraLoaderMixin.load_lora_weights">load_lora_weights()</a>.`,name:"hotswap"}],source:"https://github.com/huggingface/diffusers/blob/vr_11477/src/diffusers/loaders/lora_pipeline.py#L2616"}}),Da=new J({props:{title:"HiDreamImageLoraLoaderMixin",local:"diffusers.loaders.HiDreamImageLoraLoaderMixin",headingTag:"h2"}}),Sa=new $({props:{name:"class diffusers.loaders.HiDreamImageLoraLoaderMixin",anchor:"diffusers.loaders.HiDreamImageLoraLoaderMixin",parameters:[],source:"https://github.com/huggingface/diffusers/blob/vr_11477/src/diffusers/loaders/lora_pipeline.py#L5389"}}),Ca=new $({props:{name:"load_lora_into_transformer",anchor:"diffusers.loaders.HiDreamImageLoraLoaderMixin.load_lora_into_transformer",parameters:[{name:"state_dict",val:""},{name:"transformer",val:""},{name:"adapter_name",val:" = None"},{name:"_pipeline",val:" = None"},{name:"low_cpu_mem_usage",val:" = False"},{name:"hotswap",val:": bool = False"}],parametersDescription:[{anchor:"diffusers.loaders.HiDreamImageLoraLoaderMixin.load_lora_into_transformer.state_dict",description:`<strong>state_dict</strong> (<code>dict</code>) — | |
| A standard state dict containing the lora layer parameters. The keys can either be indexed directly | |
| into the unet or prefixed with an additional <code>unet</code> which can be used to distinguish between text | |
| encoder lora layers.`,name:"state_dict"},{anchor:"diffusers.loaders.HiDreamImageLoraLoaderMixin.load_lora_into_transformer.transformer",description:`<strong>transformer</strong> (<code>HiDreamImageTransformer2DModel</code>) — | |
| The Transformer model to load the LoRA layers into.`,name:"transformer"},{anchor:"diffusers.loaders.HiDreamImageLoraLoaderMixin.load_lora_into_transformer.adapter_name",description:`<strong>adapter_name</strong> (<code>str</code>, <em>optional</em>) — | |
| Adapter name to be used for referencing the loaded adapter model. If not specified, it will use | |
| <code>default_{i}</code> where i is the total number of adapters being loaded.`,name:"adapter_name"},{anchor:"diffusers.loaders.HiDreamImageLoraLoaderMixin.load_lora_into_transformer.low_cpu_mem_usage",description:`<strong>low_cpu_mem_usage</strong> (<code>bool</code>, <em>optional</em>) — | |
| Speed up model loading by only loading the pretrained LoRA weights and not initializing the random | |
| weights.`,name:"low_cpu_mem_usage"},{anchor:"diffusers.loaders.HiDreamImageLoraLoaderMixin.load_lora_into_transformer.hotswap",description:`<strong>hotswap</strong> (<code>bool</code>, <em>optional</em>) — | |
| See <a href="/docs/diffusers/pr_11477/en/api/loaders/lora#diffusers.loaders.StableDiffusionLoraLoaderMixin.load_lora_weights">load_lora_weights()</a>.`,name:"hotswap"}],source:"https://github.com/huggingface/diffusers/blob/vr_11477/src/diffusers/loaders/lora_pipeline.py#L5556"}}),Aa=new $({props:{name:"load_lora_weights",anchor:"diffusers.loaders.HiDreamImageLoraLoaderMixin.load_lora_weights",parameters:[{name:"pretrained_model_name_or_path_or_dict",val:": typing.Union[str, typing.Dict[str, torch.Tensor]]"},{name:"adapter_name",val:": typing.Optional[str] = None"},{name:"hotswap",val:": bool = False"},{name:"**kwargs",val:""}],parametersDescription:[{anchor:"diffusers.loaders.HiDreamImageLoraLoaderMixin.load_lora_weights.pretrained_model_name_or_path_or_dict",description:`<strong>pretrained_model_name_or_path_or_dict</strong> (<code>str</code> or <code>os.PathLike</code> or <code>dict</code>) — | |
| See <a href="/docs/diffusers/pr_11477/en/api/loaders/lora#diffusers.loaders.StableDiffusionLoraLoaderMixin.lora_state_dict">lora_state_dict()</a>.`,name:"pretrained_model_name_or_path_or_dict"},{anchor:"diffusers.loaders.HiDreamImageLoraLoaderMixin.load_lora_weights.adapter_name",description:`<strong>adapter_name</strong> (<code>str</code>, <em>optional</em>) — | |
| Adapter name to be used for referencing the loaded adapter model. If not specified, it will use | |
| <code>default_{i}</code> where i is the total number of adapters being loaded.`,name:"adapter_name"},{anchor:"diffusers.loaders.HiDreamImageLoraLoaderMixin.load_lora_weights.low_cpu_mem_usage",description:`<strong>low_cpu_mem_usage</strong> (<code>bool</code>, <em>optional</em>) — | |
| Speed up model loading by only loading the pretrained LoRA weights and not initializing the random | |
| weights.`,name:"low_cpu_mem_usage"},{anchor:"diffusers.loaders.HiDreamImageLoraLoaderMixin.load_lora_weights.hotswap",description:`<strong>hotswap</strong> (<code>bool</code>, <em>optional</em>) — | |
| See <a href="/docs/diffusers/pr_11477/en/api/loaders/lora#diffusers.loaders.StableDiffusionLoraLoaderMixin.load_lora_weights">load_lora_weights()</a>.`,name:"hotswap"},{anchor:"diffusers.loaders.HiDreamImageLoraLoaderMixin.load_lora_weights.kwargs",description:`<strong>kwargs</strong> (<code>dict</code>, <em>optional</em>) — | |
| See <a href="/docs/diffusers/pr_11477/en/api/loaders/lora#diffusers.loaders.StableDiffusionLoraLoaderMixin.lora_state_dict">lora_state_dict()</a>.`,name:"kwargs"}],source:"https://github.com/huggingface/diffusers/blob/vr_11477/src/diffusers/loaders/lora_pipeline.py#L5499"}}),ka=new $({props:{name:"lora_state_dict",anchor:"diffusers.loaders.HiDreamImageLoraLoaderMixin.lora_state_dict",parameters:[{name:"pretrained_model_name_or_path_or_dict",val:": typing.Union[str, typing.Dict[str, torch.Tensor]]"},{name:"**kwargs",val:""}],parametersDescription:[{anchor:"diffusers.loaders.HiDreamImageLoraLoaderMixin.lora_state_dict.pretrained_model_name_or_path_or_dict",description:`<strong>pretrained_model_name_or_path_or_dict</strong> (<code>str</code> or <code>os.PathLike</code> or <code>dict</code>) — | |
| Can be either:</p> | |
| <ul> | |
| <li>A string, the <em>model id</em> (for example <code>google/ddpm-celebahq-256</code>) of a pretrained model hosted on | |
| the Hub.</li> | |
| <li>A path to a <em>directory</em> (for example <code>./my_model_directory</code>) containing the model weights saved | |
| with <a href="/docs/diffusers/pr_11477/en/api/models/overview#diffusers.ModelMixin.save_pretrained">ModelMixin.save_pretrained()</a>.</li> | |
| <li>A <a href="https://pytorch.org/tutorials/beginner/saving_loading_models.html#what-is-a-state-dict" rel="nofollow">torch state | |
| dict</a>.</li> | |
| </ul>`,name:"pretrained_model_name_or_path_or_dict"},{anchor:"diffusers.loaders.HiDreamImageLoraLoaderMixin.lora_state_dict.cache_dir",description:`<strong>cache_dir</strong> (<code>Union[str, os.PathLike]</code>, <em>optional</em>) — | |
| Path to a directory where a downloaded pretrained model configuration is cached if the standard cache | |
| is not used.`,name:"cache_dir"},{anchor:"diffusers.loaders.HiDreamImageLoraLoaderMixin.lora_state_dict.force_download",description:`<strong>force_download</strong> (<code>bool</code>, <em>optional</em>, defaults to <code>False</code>) — | |
| Whether or not to force the (re-)download of the model weights and configuration files, overriding the | |
| cached versions if they exist.`,name:"force_download"},{anchor:"diffusers.loaders.HiDreamImageLoraLoaderMixin.lora_state_dict.proxies",description:`<strong>proxies</strong> (<code>Dict[str, str]</code>, <em>optional</em>) — | |
| A dictionary of proxy servers to use by protocol or endpoint, for example, <code>{'http': 'foo.bar:3128', 'http://hostname': 'foo.bar:4012'}</code>. The proxies are used on each request.`,name:"proxies"},{anchor:"diffusers.loaders.HiDreamImageLoraLoaderMixin.lora_state_dict.local_files_only",description:`<strong>local_files_only</strong> (<code>bool</code>, <em>optional</em>, defaults to <code>False</code>) — | |
| Whether to only load local model weights and configuration files or not. If set to <code>True</code>, the model | |
| won’t be downloaded from the Hub.`,name:"local_files_only"},{anchor:"diffusers.loaders.HiDreamImageLoraLoaderMixin.lora_state_dict.token",description:`<strong>token</strong> (<code>str</code> or <em>bool</em>, <em>optional</em>) — | |
| The token to use as HTTP bearer authorization for remote files. If <code>True</code>, the token generated from | |
| <code>diffusers-cli login</code> (stored in <code>~/.huggingface</code>) is used.`,name:"token"},{anchor:"diffusers.loaders.HiDreamImageLoraLoaderMixin.lora_state_dict.revision",description:`<strong>revision</strong> (<code>str</code>, <em>optional</em>, defaults to <code>"main"</code>) — | |
| The specific model version to use. It can be a branch name, a tag name, a commit id, or any identifier | |
| allowed by Git.`,name:"revision"},{anchor:"diffusers.loaders.HiDreamImageLoraLoaderMixin.lora_state_dict.subfolder",description:`<strong>subfolder</strong> (<code>str</code>, <em>optional</em>, defaults to <code>""</code>) — | |
| The subfolder location of a model file within a larger model repository on the Hub or locally.`,name:"subfolder"}],source:"https://github.com/huggingface/diffusers/blob/vr_11477/src/diffusers/loaders/lora_pipeline.py#L5397"}}),vr=new S({props:{warning:!0,$$slots:{default:[Ih]},$$scope:{ctx:y}}}),Pa=new $({props:{name:"save_lora_weights",anchor:"diffusers.loaders.HiDreamImageLoraLoaderMixin.save_lora_weights",parameters:[{name:"save_directory",val:": typing.Union[str, os.PathLike]"},{name:"transformer_lora_layers",val:": typing.Dict[str, typing.Union[torch.nn.modules.module.Module, torch.Tensor]] = None"},{name:"is_main_process",val:": bool = True"},{name:"weight_name",val:": str = None"},{name:"save_function",val:": typing.Callable = None"},{name:"safe_serialization",val:": bool = True"}],parametersDescription:[{anchor:"diffusers.loaders.HiDreamImageLoraLoaderMixin.save_lora_weights.save_directory",description:`<strong>save_directory</strong> (<code>str</code> or <code>os.PathLike</code>) — | |
| Directory to save LoRA parameters to. Will be created if it doesn’t exist.`,name:"save_directory"},{anchor:"diffusers.loaders.HiDreamImageLoraLoaderMixin.save_lora_weights.transformer_lora_layers",description:`<strong>transformer_lora_layers</strong> (<code>Dict[str, torch.nn.Module]</code> or <code>Dict[str, torch.Tensor]</code>) — | |
| State dict of the LoRA layers corresponding to the <code>transformer</code>.`,name:"transformer_lora_layers"},{anchor:"diffusers.loaders.HiDreamImageLoraLoaderMixin.save_lora_weights.is_main_process",description:`<strong>is_main_process</strong> (<code>bool</code>, <em>optional</em>, defaults to <code>True</code>) — | |
| Whether the process calling this is the main process or not. Useful during distributed training and you | |
| need to call this function on all processes. In this case, set <code>is_main_process=True</code> only on the main | |
| process to avoid race conditions.`,name:"is_main_process"},{anchor:"diffusers.loaders.HiDreamImageLoraLoaderMixin.save_lora_weights.save_function",description:`<strong>save_function</strong> (<code>Callable</code>) — | |
| The function to use to save the state dictionary. Useful during distributed training when you need to | |
| replace <code>torch.save</code> with another method. Can be configured with the environment variable | |
| <code>DIFFUSERS_SAVE_MODE</code>.`,name:"save_function"},{anchor:"diffusers.loaders.HiDreamImageLoraLoaderMixin.save_lora_weights.safe_serialization",description:`<strong>safe_serialization</strong> (<code>bool</code>, <em>optional</em>, defaults to <code>True</code>) — | |
| Whether to save the model using <code>safetensors</code> or the traditional PyTorch way with <code>pickle</code>.`,name:"safe_serialization"}],source:"https://github.com/huggingface/diffusers/blob/vr_11477/src/diffusers/loaders/lora_pipeline.py#L5596"}}),Ra=new $({props:{name:"unfuse_lora",anchor:"diffusers.loaders.HiDreamImageLoraLoaderMixin.unfuse_lora",parameters:[{name:"components",val:": typing.List[str] = ['transformer']"},{name:"**kwargs",val:""}],parametersDescription:[{anchor:"diffusers.loaders.HiDreamImageLoraLoaderMixin.unfuse_lora.components",description:"<strong>components</strong> (<code>List[str]</code>) — List of LoRA-injectable components to unfuse LoRA from.",name:"components"},{anchor:"diffusers.loaders.HiDreamImageLoraLoaderMixin.unfuse_lora.unfuse_transformer",description:"<strong>unfuse_transformer</strong> (<code>bool</code>, defaults to <code>True</code>) — Whether to unfuse the UNet LoRA parameters.",name:"unfuse_transformer"}],source:"https://github.com/huggingface/diffusers/blob/vr_11477/src/diffusers/loaders/lora_pipeline.py#L5693"}}),$r=new S({props:{warning:!0,$$slots:{default:[Vh]},$$scope:{ctx:y}}}),Ha=new J({props:{title:"LoraBaseMixin",local:"diffusers.loaders.lora_base.LoraBaseMixin",headingTag:"h2"}}),Ia=new $({props:{name:"class diffusers.loaders.lora_base.LoraBaseMixin",anchor:"diffusers.loaders.lora_base.LoraBaseMixin",parameters:[],source:"https://github.com/huggingface/diffusers/blob/vr_11477/src/diffusers/loaders/lora_base.py#L464"}}),Va=new $({props:{name:"delete_adapters",anchor:"diffusers.loaders.lora_base.LoraBaseMixin.delete_adapters",parameters:[{name:"adapter_names",val:": typing.Union[typing.List[str], str]"}],parametersDescription:[{anchor:"diffusers.loaders.lora_base.LoraBaseMixin.delete_adapters.Deletes",description:`<strong>Deletes</strong> the LoRA layers of <code>adapter_name</code> for the unet and text-encoder(s). — | |
| adapter_names (<code>Union[List[str], str]</code>): | |
| The names of the adapter to delete. Can be a single string or a list of strings`,name:"Deletes"}],source:"https://github.com/huggingface/diffusers/blob/vr_11477/src/diffusers/loaders/lora_base.py#L783"}}),Wa=new $({props:{name:"enable_lora_hotswap",anchor:"diffusers.loaders.lora_base.LoraBaseMixin.enable_lora_hotswap",parameters:[{name:"**kwargs",val:""}],parametersDescription:[{anchor:"diffusers.loaders.lora_base.LoraBaseMixin.enable_lora_hotswap.target_rank",description:`<strong>target_rank</strong> (<code>int</code>) — | |
| The highest rank among all the adapters that will be loaded.`,name:"target_rank"},{anchor:"diffusers.loaders.lora_base.LoraBaseMixin.enable_lora_hotswap.check_compiled",description:`<strong>check_compiled</strong> (<code>str</code>, <em>optional</em>, defaults to <code>"error"</code>) — | |
| How to handle the case when the model is already compiled, which should generally be avoided. The | |
| options are:<ul> | |
| <li>“error” (default): raise an error</li> | |
| <li>“warn”: issue a warning</li> | |
| <li>“ignore”: do nothing</li> | |
| </ul>`,name:"check_compiled"}],source:"https://github.com/huggingface/diffusers/blob/vr_11477/src/diffusers/loaders/lora_base.py#L935"}}),Fa=new $({props:{name:"fuse_lora",anchor:"diffusers.loaders.lora_base.LoraBaseMixin.fuse_lora",parameters:[{name:"components",val:": typing.List[str] = []"},{name:"lora_scale",val:": float = 1.0"},{name:"safe_fusing",val:": bool = False"},{name:"adapter_names",val:": typing.Optional[typing.List[str]] = None"},{name:"**kwargs",val:""}],parametersDescription:[{anchor:"diffusers.loaders.lora_base.LoraBaseMixin.fuse_lora.components",description:"<strong>components</strong> — (<code>List[str]</code>): List of LoRA-injectable components to fuse the LoRAs into.",name:"components"},{anchor:"diffusers.loaders.lora_base.LoraBaseMixin.fuse_lora.lora_scale",description:`<strong>lora_scale</strong> (<code>float</code>, defaults to 1.0) — | |
| Controls how much to influence the outputs with the LoRA parameters.`,name:"lora_scale"},{anchor:"diffusers.loaders.lora_base.LoraBaseMixin.fuse_lora.safe_fusing",description:`<strong>safe_fusing</strong> (<code>bool</code>, defaults to <code>False</code>) — | |
| Whether to check fused weights for NaN values before fusing and if values are NaN not fusing them.`,name:"safe_fusing"},{anchor:"diffusers.loaders.lora_base.LoraBaseMixin.fuse_lora.adapter_names",description:`<strong>adapter_names</strong> (<code>List[str]</code>, <em>optional</em>) — | |
| Adapter names to be used for fusing. If nothing is passed, all active adapters will be fused.`,name:"adapter_names"}],source:"https://github.com/huggingface/diffusers/blob/vr_11477/src/diffusers/loaders/lora_base.py#L531"}}),Mr=new S({props:{warning:!0,$$slots:{default:[Wh]},$$scope:{ctx:y}}}),yr=new hl({props:{anchor:"diffusers.loaders.lora_base.LoraBaseMixin.fuse_lora.example",$$slots:{default:[Fh]},$$scope:{ctx:y}}}),Ua=new $({props:{name:"get_active_adapters",anchor:"diffusers.loaders.lora_base.LoraBaseMixin.get_active_adapters",parameters:[],source:"https://github.com/huggingface/diffusers/blob/vr_11477/src/diffusers/loaders/lora_base.py#L805"}}),Tr=new hl({props:{anchor:"diffusers.loaders.lora_base.LoraBaseMixin.get_active_adapters.example",$$slots:{default:[Uh]},$$scope:{ctx:y}}}),Ea=new $({props:{name:"get_list_adapters",anchor:"diffusers.loaders.lora_base.LoraBaseMixin.get_list_adapters",parameters:[],source:"https://github.com/huggingface/diffusers/blob/vr_11477/src/diffusers/loaders/lora_base.py#L838"}}),Na=new $({props:{name:"set_lora_device",anchor:"diffusers.loaders.lora_base.LoraBaseMixin.set_lora_device",parameters:[{name:"adapter_names",val:": typing.List[str]"},{name:"device",val:": typing.Union[torch.device, str, int]"}],parametersDescription:[{anchor:"diffusers.loaders.lora_base.LoraBaseMixin.set_lora_device.adapter_names",description:`<strong>adapter_names</strong> (<code>List[str]</code>) — | |
| List of adapters to send device to.`,name:"adapter_names"},{anchor:"diffusers.loaders.lora_base.LoraBaseMixin.set_lora_device.device",description:`<strong>device</strong> (<code>Union[torch.device, str, int]</code>) — | |
| Device to send the adapters to. Can be either a torch device, a str or an integer.`,name:"device"}],source:"https://github.com/huggingface/diffusers/blob/vr_11477/src/diffusers/loaders/lora_base.py#L860"}}),Xa=new $({props:{name:"unfuse_lora",anchor:"diffusers.loaders.lora_base.LoraBaseMixin.unfuse_lora",parameters:[{name:"components",val:": typing.List[str] = []"},{name:"**kwargs",val:""}],parametersDescription:[{anchor:"diffusers.loaders.lora_base.LoraBaseMixin.unfuse_lora.components",description:"<strong>components</strong> (<code>List[str]</code>) — List of LoRA-injectable components to unfuse LoRA from.",name:"components"},{anchor:"diffusers.loaders.lora_base.LoraBaseMixin.unfuse_lora.unfuse_unet",description:"<strong>unfuse_unet</strong> (<code>bool</code>, defaults to <code>True</code>) — Whether to unfuse the UNet LoRA parameters.",name:"unfuse_unet"},{anchor:"diffusers.loaders.lora_base.LoraBaseMixin.unfuse_lora.unfuse_text_encoder",description:`<strong>unfuse_text_encoder</strong> (<code>bool</code>, defaults to <code>True</code>) — | |
| Whether to unfuse the text encoder LoRA parameters. If the text encoder wasn’t monkey-patched with the | |
| LoRA parameters then it won’t have any effect.`,name:"unfuse_text_encoder"}],source:"https://github.com/huggingface/diffusers/blob/vr_11477/src/diffusers/loaders/lora_base.py#L621"}}),Cr=new S({props:{warning:!0,$$slots:{default:[Eh]},$$scope:{ctx:y}}}),qa=new $({props:{name:"unload_lora_weights",anchor:"diffusers.loaders.lora_base.LoraBaseMixin.unload_lora_weights",parameters:[],source:"https://github.com/huggingface/diffusers/blob/vr_11477/src/diffusers/loaders/lora_base.py#L508"}}),Ar=new hl({props:{anchor:"diffusers.loaders.lora_base.LoraBaseMixin.unload_lora_weights.example",$$slots:{default:[Nh]},$$scope:{ctx:y}}}),za=new J({props:{title:"WanLoraLoaderMixin",local:"diffusers.loaders.WanLoraLoaderMixin",headingTag:"h2"}}),Ba=new $({props:{name:"class diffusers.loaders.WanLoraLoaderMixin",anchor:"diffusers.loaders.WanLoraLoaderMixin",parameters:[],source:"https://github.com/huggingface/diffusers/blob/vr_11477/src/diffusers/loaders/lora_pipeline.py#L4697"}}),ja=new $({props:{name:"load_lora_into_transformer",anchor:"diffusers.loaders.WanLoraLoaderMixin.load_lora_into_transformer",parameters:[{name:"state_dict",val:""},{name:"transformer",val:""},{name:"adapter_name",val:" = None"},{name:"_pipeline",val:" = None"},{name:"low_cpu_mem_usage",val:" = False"},{name:"hotswap",val:": bool = False"}],parametersDescription:[{anchor:"diffusers.loaders.WanLoraLoaderMixin.load_lora_into_transformer.state_dict",description:`<strong>state_dict</strong> (<code>dict</code>) — | |
| A standard state dict containing the lora layer parameters. The keys can either be indexed directly | |
| into the unet or prefixed with an additional <code>unet</code> which can be used to distinguish between text | |
| encoder lora layers.`,name:"state_dict"},{anchor:"diffusers.loaders.WanLoraLoaderMixin.load_lora_into_transformer.transformer",description:`<strong>transformer</strong> (<code>WanTransformer3DModel</code>) — | |
| The Transformer model to load the LoRA layers into.`,name:"transformer"},{anchor:"diffusers.loaders.WanLoraLoaderMixin.load_lora_into_transformer.adapter_name",description:`<strong>adapter_name</strong> (<code>str</code>, <em>optional</em>) — | |
| Adapter name to be used for referencing the loaded adapter model. If not specified, it will use | |
| <code>default_{i}</code> where i is the total number of adapters being loaded.`,name:"adapter_name"},{anchor:"diffusers.loaders.WanLoraLoaderMixin.load_lora_into_transformer.low_cpu_mem_usage",description:`<strong>low_cpu_mem_usage</strong> (<code>bool</code>, <em>optional</em>) — | |
| Speed up model loading by only loading the pretrained LoRA weights and not initializing the random | |
| weights.`,name:"low_cpu_mem_usage"},{anchor:"diffusers.loaders.WanLoraLoaderMixin.load_lora_into_transformer.hotswap",description:`<strong>hotswap</strong> (<code>bool</code>, <em>optional</em>) — | |
| See <a href="/docs/diffusers/pr_11477/en/api/loaders/lora#diffusers.loaders.StableDiffusionLoraLoaderMixin.load_lora_weights">load_lora_weights()</a>.`,name:"hotswap"}],source:"https://github.com/huggingface/diffusers/blob/vr_11477/src/diffusers/loaders/lora_pipeline.py#L4915"}}),Ga=new $({props:{name:"load_lora_weights",anchor:"diffusers.loaders.WanLoraLoaderMixin.load_lora_weights",parameters:[{name:"pretrained_model_name_or_path_or_dict",val:": typing.Union[str, typing.Dict[str, torch.Tensor]]"},{name:"adapter_name",val:": typing.Optional[str] = None"},{name:"hotswap",val:": bool = False"},{name:"**kwargs",val:""}],parametersDescription:[{anchor:"diffusers.loaders.WanLoraLoaderMixin.load_lora_weights.pretrained_model_name_or_path_or_dict",description:`<strong>pretrained_model_name_or_path_or_dict</strong> (<code>str</code> or <code>os.PathLike</code> or <code>dict</code>) — | |
| See <a href="/docs/diffusers/pr_11477/en/api/loaders/lora#diffusers.loaders.StableDiffusionLoraLoaderMixin.lora_state_dict">lora_state_dict()</a>.`,name:"pretrained_model_name_or_path_or_dict"},{anchor:"diffusers.loaders.WanLoraLoaderMixin.load_lora_weights.adapter_name",description:`<strong>adapter_name</strong> (<code>str</code>, <em>optional</em>) — | |
| Adapter name to be used for referencing the loaded adapter model. If not specified, it will use | |
| <code>default_{i}</code> where i is the total number of adapters being loaded.`,name:"adapter_name"},{anchor:"diffusers.loaders.WanLoraLoaderMixin.load_lora_weights.low_cpu_mem_usage",description:`<strong>low_cpu_mem_usage</strong> (<code>bool</code>, <em>optional</em>) — | |
| Speed up model loading by only loading the pretrained LoRA weights and not initializing the random | |
| weights.`,name:"low_cpu_mem_usage"},{anchor:"diffusers.loaders.WanLoraLoaderMixin.load_lora_weights.hotswap",description:`<strong>hotswap</strong> (<code>bool</code>, <em>optional</em>) — | |
| See <a href="/docs/diffusers/pr_11477/en/api/loaders/lora#diffusers.loaders.StableDiffusionLoraLoaderMixin.load_lora_weights">load_lora_weights()</a>.`,name:"hotswap"},{anchor:"diffusers.loaders.WanLoraLoaderMixin.load_lora_weights.kwargs",description:`<strong>kwargs</strong> (<code>dict</code>, <em>optional</em>) — | |
| See <a href="/docs/diffusers/pr_11477/en/api/loaders/lora#diffusers.loaders.StableDiffusionLoraLoaderMixin.lora_state_dict">lora_state_dict()</a>.`,name:"kwargs"}],source:"https://github.com/huggingface/diffusers/blob/vr_11477/src/diffusers/loaders/lora_pipeline.py#L4854"}}),Ja=new $({props:{name:"lora_state_dict",anchor:"diffusers.loaders.WanLoraLoaderMixin.lora_state_dict",parameters:[{name:"pretrained_model_name_or_path_or_dict",val:": typing.Union[str, typing.Dict[str, torch.Tensor]]"},{name:"**kwargs",val:""}],parametersDescription:[{anchor:"diffusers.loaders.WanLoraLoaderMixin.lora_state_dict.pretrained_model_name_or_path_or_dict",description:`<strong>pretrained_model_name_or_path_or_dict</strong> (<code>str</code> or <code>os.PathLike</code> or <code>dict</code>) — | |
| Can be either:</p> | |
| <ul> | |
| <li>A string, the <em>model id</em> (for example <code>google/ddpm-celebahq-256</code>) of a pretrained model hosted on | |
| the Hub.</li> | |
| <li>A path to a <em>directory</em> (for example <code>./my_model_directory</code>) containing the model weights saved | |
| with <a href="/docs/diffusers/pr_11477/en/api/models/overview#diffusers.ModelMixin.save_pretrained">ModelMixin.save_pretrained()</a>.</li> | |
| <li>A <a href="https://pytorch.org/tutorials/beginner/saving_loading_models.html#what-is-a-state-dict" rel="nofollow">torch state | |
| dict</a>.</li> | |
| </ul>`,name:"pretrained_model_name_or_path_or_dict"},{anchor:"diffusers.loaders.WanLoraLoaderMixin.lora_state_dict.cache_dir",description:`<strong>cache_dir</strong> (<code>Union[str, os.PathLike]</code>, <em>optional</em>) — | |
| Path to a directory where a downloaded pretrained model configuration is cached if the standard cache | |
| is not used.`,name:"cache_dir"},{anchor:"diffusers.loaders.WanLoraLoaderMixin.lora_state_dict.force_download",description:`<strong>force_download</strong> (<code>bool</code>, <em>optional</em>, defaults to <code>False</code>) — | |
| Whether or not to force the (re-)download of the model weights and configuration files, overriding the | |
| cached versions if they exist.`,name:"force_download"},{anchor:"diffusers.loaders.WanLoraLoaderMixin.lora_state_dict.proxies",description:`<strong>proxies</strong> (<code>Dict[str, str]</code>, <em>optional</em>) — | |
| A dictionary of proxy servers to use by protocol or endpoint, for example, <code>{'http': 'foo.bar:3128', 'http://hostname': 'foo.bar:4012'}</code>. The proxies are used on each request.`,name:"proxies"},{anchor:"diffusers.loaders.WanLoraLoaderMixin.lora_state_dict.local_files_only",description:`<strong>local_files_only</strong> (<code>bool</code>, <em>optional</em>, defaults to <code>False</code>) — | |
| Whether to only load local model weights and configuration files or not. If set to <code>True</code>, the model | |
| won’t be downloaded from the Hub.`,name:"local_files_only"},{anchor:"diffusers.loaders.WanLoraLoaderMixin.lora_state_dict.token",description:`<strong>token</strong> (<code>str</code> or <em>bool</em>, <em>optional</em>) — | |
| The token to use as HTTP bearer authorization for remote files. If <code>True</code>, the token generated from | |
| <code>diffusers-cli login</code> (stored in <code>~/.huggingface</code>) is used.`,name:"token"},{anchor:"diffusers.loaders.WanLoraLoaderMixin.lora_state_dict.revision",description:`<strong>revision</strong> (<code>str</code>, <em>optional</em>, defaults to <code>"main"</code>) — | |
| The specific model version to use. It can be a branch name, a tag name, a commit id, or any identifier | |
| allowed by Git.`,name:"revision"},{anchor:"diffusers.loaders.WanLoraLoaderMixin.lora_state_dict.subfolder",description:`<strong>subfolder</strong> (<code>str</code>, <em>optional</em>, defaults to <code>""</code>) — | |
| The subfolder location of a model file within a larger model repository on the Hub or locally.`,name:"subfolder"}],source:"https://github.com/huggingface/diffusers/blob/vr_11477/src/diffusers/loaders/lora_pipeline.py#L4705"}}),Rr=new S({props:{warning:!0,$$slots:{default:[Xh]},$$scope:{ctx:y}}}),Za=new $({props:{name:"save_lora_weights",anchor:"diffusers.loaders.WanLoraLoaderMixin.save_lora_weights",parameters:[{name:"save_directory",val:": typing.Union[str, os.PathLike]"},{name:"transformer_lora_layers",val:": typing.Dict[str, typing.Union[torch.nn.modules.module.Module, torch.Tensor]] = None"},{name:"is_main_process",val:": bool = True"},{name:"weight_name",val:": str = None"},{name:"save_function",val:": typing.Callable = None"},{name:"safe_serialization",val:": bool = True"}],parametersDescription:[{anchor:"diffusers.loaders.WanLoraLoaderMixin.save_lora_weights.save_directory",description:`<strong>save_directory</strong> (<code>str</code> or <code>os.PathLike</code>) — | |
| Directory to save LoRA parameters to. Will be created if it doesn’t exist.`,name:"save_directory"},{anchor:"diffusers.loaders.WanLoraLoaderMixin.save_lora_weights.transformer_lora_layers",description:`<strong>transformer_lora_layers</strong> (<code>Dict[str, torch.nn.Module]</code> or <code>Dict[str, torch.Tensor]</code>) — | |
| State dict of the LoRA layers corresponding to the <code>transformer</code>.`,name:"transformer_lora_layers"},{anchor:"diffusers.loaders.WanLoraLoaderMixin.save_lora_weights.is_main_process",description:`<strong>is_main_process</strong> (<code>bool</code>, <em>optional</em>, defaults to <code>True</code>) — | |
| Whether the process calling this is the main process or not. Useful during distributed training and you | |
| need to call this function on all processes. In this case, set <code>is_main_process=True</code> only on the main | |
| process to avoid race conditions.`,name:"is_main_process"},{anchor:"diffusers.loaders.WanLoraLoaderMixin.save_lora_weights.save_function",description:`<strong>save_function</strong> (<code>Callable</code>) — | |
| The function to use to save the state dictionary. Useful during distributed training when you need to | |
| replace <code>torch.save</code> with another method. Can be configured with the environment variable | |
| <code>DIFFUSERS_SAVE_MODE</code>.`,name:"save_function"},{anchor:"diffusers.loaders.WanLoraLoaderMixin.save_lora_weights.safe_serialization",description:`<strong>safe_serialization</strong> (<code>bool</code>, <em>optional</em>, defaults to <code>True</code>) — | |
| Whether to save the model using <code>safetensors</code> or the traditional PyTorch way with <code>pickle</code>.`,name:"safe_serialization"}],source:"https://github.com/huggingface/diffusers/blob/vr_11477/src/diffusers/loaders/lora_pipeline.py#L4955"}}),Oa=new $({props:{name:"unfuse_lora",anchor:"diffusers.loaders.WanLoraLoaderMixin.unfuse_lora",parameters:[{name:"components",val:": typing.List[str] = ['transformer']"},{name:"**kwargs",val:""}],parametersDescription:[{anchor:"diffusers.loaders.WanLoraLoaderMixin.unfuse_lora.components",description:"<strong>components</strong> (<code>List[str]</code>) — List of LoRA-injectable components to unfuse LoRA from.",name:"components"},{anchor:"diffusers.loaders.WanLoraLoaderMixin.unfuse_lora.unfuse_transformer",description:"<strong>unfuse_transformer</strong> (<code>bool</code>, defaults to <code>True</code>) — Whether to unfuse the UNet LoRA parameters.",name:"unfuse_transformer"}],source:"https://github.com/huggingface/diffusers/blob/vr_11477/src/diffusers/loaders/lora_pipeline.py#L5052"}}),Ir=new S({props:{warning:!0,$$slots:{default:[qh]},$$scope:{ctx:y}}}),Ya=new dh({props:{source:"https://github.com/huggingface/diffusers/blob/main/docs/source/en/api/loaders/lora.md"}}),{c(){a=n("meta"),b=r(),l=n("p"),c=r(),p(T.$$.fragment),d=r(),M=n("p"),M.innerHTML=qm,Qi=r(),Wr=n("ul"),Wr.innerHTML=zm,Ki=r(),p(so.$$.fragment),ed=r(),p(Fr.$$.fragment),od=r(),I=n("div"),p(Ur.$$.fragment),Ll=r(),un=n("p"),un.innerHTML=Bm,xl=r(),io=n("div"),p(Er.$$.fragment),bl=r(),hn=n("p"),hn.innerHTML=jm,vl=r(),lo=n("div"),p(Nr.$$.fragment),wl=r(),gn=n("p"),gn.innerHTML=Gm,$l=r(),K=n("div"),p(Xr.$$.fragment),Ml=r(),Ln=n("p"),Ln.innerHTML=Jm,yl=r(),xn=n("p"),xn.innerHTML=Zm,Tl=r(),bn=n("p"),bn.innerHTML=Om,Dl=r(),vn=n("p"),vn.innerHTML=Ym,Sl=r(),wn=n("p"),wn.innerHTML=Qm,Cl=r(),ge=n("div"),p(qr.$$.fragment),Al=r(),$n=n("p"),$n.textContent=Km,kl=r(),p(co.$$.fragment),Pl=r(),fo=n("div"),p(zr.$$.fragment),Rl=r(),Mn=n("p"),Mn.textContent=e_,rd=r(),p(Br.$$.fragment),td=r(),V=n("div"),p(jr.$$.fragment),Hl=r(),yn=n("p"),yn.innerHTML=o_,Il=r(),po=n("div"),p(Gr.$$.fragment),Vl=r(),Tn=n("p"),Tn.innerHTML=r_,Wl=r(),mo=n("div"),p(Jr.$$.fragment),Fl=r(),Dn=n("p"),Dn.innerHTML=t_,Ul=r(),ee=n("div"),p(Zr.$$.fragment),El=r(),Sn=n("p"),Sn.innerHTML=a_,Nl=r(),Cn=n("p"),Cn.innerHTML=n_,Xl=r(),An=n("p"),An.innerHTML=s_,ql=r(),kn=n("p"),kn.innerHTML=i_,zl=r(),Pn=n("p"),Pn.innerHTML=d_,Bl=r(),Le=n("div"),p(Or.$$.fragment),jl=r(),Rn=n("p"),Rn.textContent=l_,Gl=r(),p(_o.$$.fragment),Jl=r(),uo=n("div"),p(Yr.$$.fragment),Zl=r(),Hn=n("p"),Hn.textContent=c_,ad=r(),p(Qr.$$.fragment),nd=r(),k=n("div"),p(Kr.$$.fragment),Ol=r(),In=n("p"),In.innerHTML=f_,Yl=r(),Vn=n("p"),Vn.innerHTML=p_,Ql=r(),ho=n("div"),p(et.$$.fragment),Kl=r(),Wn=n("p"),Wn.innerHTML=m_,ec=r(),go=n("div"),p(ot.$$.fragment),oc=r(),Fn=n("p"),Fn.innerHTML=__,rc=r(),te=n("div"),p(rt.$$.fragment),tc=r(),Un=n("p"),Un.innerHTML=u_,ac=r(),En=n("p"),En.innerHTML=h_,nc=r(),Nn=n("p"),Nn.innerHTML=g_,sc=r(),Xn=n("p"),Xn.innerHTML=L_,ic=r(),xe=n("div"),p(tt.$$.fragment),dc=r(),qn=n("p"),qn.textContent=x_,lc=r(),p(Lo.$$.fragment),cc=r(),xo=n("div"),p(at.$$.fragment),fc=r(),zn=n("p"),zn.textContent=b_,pc=r(),be=n("div"),p(nt.$$.fragment),mc=r(),Bn=n("p"),Bn.innerHTML=v_,_c=r(),p(bo.$$.fragment),sd=r(),p(st.$$.fragment),id=r(),C=n("div"),p(it.$$.fragment),uc=r(),jn=n("p"),jn.innerHTML=w_,hc=r(),Gn=n("p"),Gn.innerHTML=$_,gc=r(),vo=n("div"),p(dt.$$.fragment),Lc=r(),Jn=n("p"),Jn.innerHTML=M_,xc=r(),wo=n("div"),p(lt.$$.fragment),bc=r(),Zn=n("p"),Zn.innerHTML=y_,vc=r(),ae=n("div"),p(ct.$$.fragment),wc=r(),On=n("p"),On.innerHTML=T_,$c=r(),Yn=n("p"),Yn.innerHTML=D_,Mc=r(),Qn=n("p"),Qn.innerHTML=S_,yc=r(),Kn=n("p"),Kn.innerHTML=C_,Tc=r(),ve=n("div"),p(ft.$$.fragment),Dc=r(),es=n("p"),es.textContent=A_,Sc=r(),p($o.$$.fragment),Cc=r(),Mo=n("div"),p(pt.$$.fragment),Ac=r(),os=n("p"),os.textContent=k_,kc=r(),we=n("div"),p(mt.$$.fragment),Pc=r(),rs=n("p"),rs.innerHTML=P_,Rc=r(),p(yo.$$.fragment),Hc=r(),$e=n("div"),p(_t.$$.fragment),Ic=r(),ts=n("p"),ts.textContent=R_,Vc=r(),p(To.$$.fragment),dd=r(),p(ut.$$.fragment),ld=r(),W=n("div"),p(ht.$$.fragment),Wc=r(),as=n("p"),as.innerHTML=H_,Fc=r(),Do=n("div"),p(gt.$$.fragment),Uc=r(),ns=n("p"),ns.innerHTML=I_,Ec=r(),So=n("div"),p(Lt.$$.fragment),Nc=r(),ss=n("p"),ss.innerHTML=V_,Xc=r(),Me=n("div"),p(xt.$$.fragment),qc=r(),is=n("p"),is.textContent=W_,zc=r(),p(Co.$$.fragment),Bc=r(),Ao=n("div"),p(bt.$$.fragment),jc=r(),ds=n("p"),ds.textContent=F_,Gc=r(),ye=n("div"),p(vt.$$.fragment),Jc=r(),ls=n("p"),ls.innerHTML=U_,Zc=r(),p(ko.$$.fragment),cd=r(),p(wt.$$.fragment),fd=r(),F=n("div"),p($t.$$.fragment),Oc=r(),cs=n("p"),cs.innerHTML=E_,Yc=r(),Po=n("div"),p(Mt.$$.fragment),Qc=r(),fs=n("p"),fs.innerHTML=N_,Kc=r(),Ro=n("div"),p(yt.$$.fragment),ef=r(),ps=n("p"),ps.innerHTML=X_,of=r(),Te=n("div"),p(Tt.$$.fragment),rf=r(),ms=n("p"),ms.textContent=q_,tf=r(),p(Ho.$$.fragment),af=r(),Io=n("div"),p(Dt.$$.fragment),nf=r(),_s=n("p"),_s.textContent=z_,sf=r(),De=n("div"),p(St.$$.fragment),df=r(),us=n("p"),us.innerHTML=B_,lf=r(),p(Vo.$$.fragment),pd=r(),p(Ct.$$.fragment),md=r(),U=n("div"),p(At.$$.fragment),cf=r(),hs=n("p"),hs.innerHTML=j_,ff=r(),Wo=n("div"),p(kt.$$.fragment),pf=r(),gs=n("p"),gs.innerHTML=G_,mf=r(),Fo=n("div"),p(Pt.$$.fragment),_f=r(),Ls=n("p"),Ls.innerHTML=J_,uf=r(),Se=n("div"),p(Rt.$$.fragment),hf=r(),xs=n("p"),xs.textContent=Z_,gf=r(),p(Uo.$$.fragment),Lf=r(),Eo=n("div"),p(Ht.$$.fragment),xf=r(),bs=n("p"),bs.textContent=O_,bf=r(),Ce=n("div"),p(It.$$.fragment),vf=r(),vs=n("p"),vs.innerHTML=Y_,wf=r(),p(No.$$.fragment),_d=r(),p(Vt.$$.fragment),ud=r(),E=n("div"),p(Wt.$$.fragment),$f=r(),ws=n("p"),ws.innerHTML=Q_,Mf=r(),Xo=n("div"),p(Ft.$$.fragment),yf=r(),$s=n("p"),$s.innerHTML=K_,Tf=r(),qo=n("div"),p(Ut.$$.fragment),Df=r(),Ms=n("p"),Ms.innerHTML=eu,Sf=r(),Ae=n("div"),p(Et.$$.fragment),Cf=r(),ys=n("p"),ys.textContent=ou,Af=r(),p(zo.$$.fragment),kf=r(),Bo=n("div"),p(Nt.$$.fragment),Pf=r(),Ts=n("p"),Ts.textContent=ru,Rf=r(),ke=n("div"),p(Xt.$$.fragment),Hf=r(),Ds=n("p"),Ds.innerHTML=tu,If=r(),p(jo.$$.fragment),hd=r(),p(qt.$$.fragment),gd=r(),N=n("div"),p(zt.$$.fragment),Vf=r(),Ss=n("p"),Ss.innerHTML=au,Wf=r(),Go=n("div"),p(Bt.$$.fragment),Ff=r(),Cs=n("p"),Cs.innerHTML=nu,Uf=r(),Jo=n("div"),p(jt.$$.fragment),Ef=r(),As=n("p"),As.innerHTML=su,Nf=r(),Pe=n("div"),p(Gt.$$.fragment),Xf=r(),ks=n("p"),ks.textContent=iu,qf=r(),p(Zo.$$.fragment),zf=r(),Oo=n("div"),p(Jt.$$.fragment),Bf=r(),Ps=n("p"),Ps.textContent=du,jf=r(),Re=n("div"),p(Zt.$$.fragment),Gf=r(),Rs=n("p"),Rs.innerHTML=lu,Jf=r(),p(Yo.$$.fragment),Ld=r(),p(Ot.$$.fragment),xd=r(),X=n("div"),p(Yt.$$.fragment),Zf=r(),Hs=n("p"),Hs.innerHTML=cu,Of=r(),Qo=n("div"),p(Qt.$$.fragment),Yf=r(),Is=n("p"),Is.innerHTML=fu,Qf=r(),Ko=n("div"),p(Kt.$$.fragment),Kf=r(),Vs=n("p"),Vs.innerHTML=pu,ep=r(),He=n("div"),p(ea.$$.fragment),op=r(),Ws=n("p"),Ws.textContent=mu,rp=r(),p(er.$$.fragment),tp=r(),or=n("div"),p(oa.$$.fragment),ap=r(),Fs=n("p"),Fs.textContent=_u,np=r(),Ie=n("div"),p(ra.$$.fragment),sp=r(),Us=n("p"),Us.innerHTML=uu,ip=r(),p(rr.$$.fragment),bd=r(),p(ta.$$.fragment),vd=r(),q=n("div"),p(aa.$$.fragment),dp=r(),Es=n("p"),Es.innerHTML=hu,lp=r(),tr=n("div"),p(na.$$.fragment),cp=r(),Ns=n("p"),Ns.innerHTML=gu,fp=r(),ar=n("div"),p(sa.$$.fragment),pp=r(),Xs=n("p"),Xs.innerHTML=Lu,mp=r(),Ve=n("div"),p(ia.$$.fragment),_p=r(),qs=n("p"),qs.textContent=xu,up=r(),p(nr.$$.fragment),hp=r(),sr=n("div"),p(da.$$.fragment),gp=r(),zs=n("p"),zs.textContent=bu,Lp=r(),We=n("div"),p(la.$$.fragment),xp=r(),Bs=n("p"),Bs.innerHTML=vu,bp=r(),p(ir.$$.fragment),wd=r(),p(ca.$$.fragment),$d=r(),z=n("div"),p(fa.$$.fragment),vp=r(),js=n("p"),js.innerHTML=wu,wp=r(),dr=n("div"),p(pa.$$.fragment),$p=r(),Gs=n("p"),Gs.innerHTML=$u,Mp=r(),lr=n("div"),p(ma.$$.fragment),yp=r(),Js=n("p"),Js.innerHTML=Mu,Tp=r(),Fe=n("div"),p(_a.$$.fragment),Dp=r(),Zs=n("p"),Zs.textContent=yu,Sp=r(),p(cr.$$.fragment),Cp=r(),fr=n("div"),p(ua.$$.fragment),Ap=r(),Os=n("p"),Os.textContent=Tu,kp=r(),Ue=n("div"),p(ha.$$.fragment),Pp=r(),Ys=n("p"),Ys.innerHTML=Du,Rp=r(),p(pr.$$.fragment),Md=r(),p(ga.$$.fragment),yd=r(),B=n("div"),p(La.$$.fragment),Hp=r(),Qs=n("p"),Qs.innerHTML=Su,Ip=r(),mr=n("div"),p(xa.$$.fragment),Vp=r(),Ks=n("p"),Ks.innerHTML=Cu,Wp=r(),_r=n("div"),p(ba.$$.fragment),Fp=r(),ei=n("p"),ei.innerHTML=Au,Up=r(),Ee=n("div"),p(va.$$.fragment),Ep=r(),oi=n("p"),oi.textContent=ku,Np=r(),p(ur.$$.fragment),Xp=r(),hr=n("div"),p(wa.$$.fragment),qp=r(),ri=n("p"),ri.textContent=Pu,zp=r(),Ne=n("div"),p($a.$$.fragment),Bp=r(),ti=n("p"),ti.innerHTML=Ru,jp=r(),p(gr.$$.fragment),Td=r(),p(Ma.$$.fragment),Dd=r(),Oe=n("div"),p(ya.$$.fragment),Gp=r(),Lr=n("div"),p(Ta.$$.fragment),Jp=r(),ai=n("p"),ai.innerHTML=Hu,Sd=r(),p(Da.$$.fragment),Cd=r(),j=n("div"),p(Sa.$$.fragment),Zp=r(),ni=n("p"),ni.innerHTML=Iu,Op=r(),xr=n("div"),p(Ca.$$.fragment),Yp=r(),si=n("p"),si.innerHTML=Vu,Qp=r(),br=n("div"),p(Aa.$$.fragment),Kp=r(),ii=n("p"),ii.innerHTML=Wu,em=r(),Xe=n("div"),p(ka.$$.fragment),om=r(),di=n("p"),di.textContent=Fu,rm=r(),p(vr.$$.fragment),tm=r(),wr=n("div"),p(Pa.$$.fragment),am=r(),li=n("p"),li.textContent=Uu,nm=r(),qe=n("div"),p(Ra.$$.fragment),sm=r(),ci=n("p"),ci.innerHTML=Eu,im=r(),p($r.$$.fragment),Ad=r(),p(Ha.$$.fragment),kd=r(),A=n("div"),p(Ia.$$.fragment),dm=r(),fi=n("p"),fi.textContent=Nu,lm=r(),pi=n("div"),p(Va.$$.fragment),cm=r(),ze=n("div"),p(Wa.$$.fragment),fm=r(),mi=n("p"),mi.textContent=Xu,pm=r(),_i=n("p"),_i.textContent=qu,mm=r(),he=n("div"),p(Fa.$$.fragment),_m=r(),ui=n("p"),ui.textContent=zu,um=r(),p(Mr.$$.fragment),hm=r(),p(yr.$$.fragment),gm=r(),Be=n("div"),p(Ua.$$.fragment),Lm=r(),hi=n("p"),hi.textContent=Bu,xm=r(),p(Tr.$$.fragment),bm=r(),Dr=n("div"),p(Ea.$$.fragment),vm=r(),gi=n("p"),gi.textContent=ju,wm=r(),Sr=n("div"),p(Na.$$.fragment),$m=r(),Li=n("p"),Li.innerHTML=Gu,Mm=r(),je=n("div"),p(Xa.$$.fragment),ym=r(),xi=n("p"),xi.innerHTML=Ju,Tm=r(),p(Cr.$$.fragment),Dm=r(),Ge=n("div"),p(qa.$$.fragment),Sm=r(),bi=n("p"),bi.textContent=Zu,Cm=r(),p(Ar.$$.fragment),Pd=r(),p(za.$$.fragment),Rd=r(),G=n("div"),p(Ba.$$.fragment),Am=r(),vi=n("p"),vi.innerHTML=Ou,km=r(),kr=n("div"),p(ja.$$.fragment),Pm=r(),wi=n("p"),wi.innerHTML=Yu,Rm=r(),Pr=n("div"),p(Ga.$$.fragment),Hm=r(),$i=n("p"),$i.innerHTML=Qu,Im=r(),Je=n("div"),p(Ja.$$.fragment),Vm=r(),Mi=n("p"),Mi.textContent=Ku,Wm=r(),p(Rr.$$.fragment),Fm=r(),Hr=n("div"),p(Za.$$.fragment),Um=r(),yi=n("p"),yi.textContent=eh,Em=r(),Ze=n("div"),p(Oa.$$.fragment),Nm=r(),Ti=n("p"),Ti.innerHTML=oh,Xm=r(),p(Ir.$$.fragment),Hd=r(),p(Ya.$$.fragment),Id=r(),Yi=n("p"),this.h()},l(e){const L=ih("svelte-u9bgzb",document.head);a=s(L,"META",{name:!0,content:!0}),L.forEach(i),b=t(e),l=s(e,"P",{}),v(l).forEach(i),c=t(e),m(T.$$.fragment,e),d=t(e),M=s(e,"P",{"data-svelte-h":!0}),f(M)!=="svelte-k6qohx"&&(M.innerHTML=qm),Qi=t(e),Wr=s(e,"UL",{"data-svelte-h":!0}),f(Wr)!=="svelte-15ou0ah"&&(Wr.innerHTML=zm),Ki=t(e),m(so.$$.fragment,e),ed=t(e),m(Fr.$$.fragment,e),od=t(e),I=s(e,"DIV",{class:!0});var Z=v(I);m(Ur.$$.fragment,Z),Ll=t(Z),un=s(Z,"P",{"data-svelte-h":!0}),f(un)!=="svelte-1ua21r1"&&(un.innerHTML=Bm),xl=t(Z),io=s(Z,"DIV",{class:!0});var Qa=v(io);m(Er.$$.fragment,Qa),bl=t(Qa),hn=s(Qa,"P",{"data-svelte-h":!0}),f(hn)!=="svelte-1062ci4"&&(hn.innerHTML=jm),Qa.forEach(i),vl=t(Z),lo=s(Z,"DIV",{class:!0});var Ka=v(lo);m(Nr.$$.fragment,Ka),wl=t(Ka),gn=s(Ka,"P",{"data-svelte-h":!0}),f(gn)!=="svelte-u3q4so"&&(gn.innerHTML=Gm),Ka.forEach(i),$l=t(Z),K=s(Z,"DIV",{class:!0});var oe=v(K);m(Xr.$$.fragment,oe),Ml=t(oe),Ln=s(oe,"P",{"data-svelte-h":!0}),f(Ln)!=="svelte-vs7s0z"&&(Ln.innerHTML=Jm),yl=t(oe),xn=s(oe,"P",{"data-svelte-h":!0}),f(xn)!=="svelte-15b960v"&&(xn.innerHTML=Zm),Tl=t(oe),bn=s(oe,"P",{"data-svelte-h":!0}),f(bn)!=="svelte-d80i8q"&&(bn.innerHTML=Om),Dl=t(oe),vn=s(oe,"P",{"data-svelte-h":!0}),f(vn)!=="svelte-1ge10rl"&&(vn.innerHTML=Ym),Sl=t(oe),wn=s(oe,"P",{"data-svelte-h":!0}),f(wn)!=="svelte-1iu9wlz"&&(wn.innerHTML=Qm),oe.forEach(i),Cl=t(Z),ge=s(Z,"DIV",{class:!0});var Ye=v(ge);m(qr.$$.fragment,Ye),Al=t(Ye),$n=s(Ye,"P",{"data-svelte-h":!0}),f($n)!=="svelte-flusvq"&&($n.textContent=Km),kl=t(Ye),m(co.$$.fragment,Ye),Ye.forEach(i),Pl=t(Z),fo=s(Z,"DIV",{class:!0});var en=v(fo);m(zr.$$.fragment,en),Rl=t(en),Mn=s(en,"P",{"data-svelte-h":!0}),f(Mn)!=="svelte-1ufq5ot"&&(Mn.textContent=e_),en.forEach(i),Z.forEach(i),rd=t(e),m(Br.$$.fragment,e),td=t(e),V=s(e,"DIV",{class:!0});var O=v(V);m(jr.$$.fragment,O),Hl=t(O),yn=s(O,"P",{"data-svelte-h":!0}),f(yn)!=="svelte-ree59i"&&(yn.innerHTML=o_),Il=t(O),po=s(O,"DIV",{class:!0});var on=v(po);m(Gr.$$.fragment,on),Vl=t(on),Tn=s(on,"P",{"data-svelte-h":!0}),f(Tn)!=="svelte-1062ci4"&&(Tn.innerHTML=r_),on.forEach(i),Wl=t(O),mo=s(O,"DIV",{class:!0});var rn=v(mo);m(Jr.$$.fragment,rn),Fl=t(rn),Dn=s(rn,"P",{"data-svelte-h":!0}),f(Dn)!=="svelte-u3q4so"&&(Dn.innerHTML=t_),rn.forEach(i),Ul=t(O),ee=s(O,"DIV",{class:!0});var re=v(ee);m(Zr.$$.fragment,re),El=t(re),Sn=s(re,"P",{"data-svelte-h":!0}),f(Sn)!=="svelte-vs7s0z"&&(Sn.innerHTML=a_),Nl=t(re),Cn=s(re,"P",{"data-svelte-h":!0}),f(Cn)!=="svelte-15b960v"&&(Cn.innerHTML=n_),Xl=t(re),An=s(re,"P",{"data-svelte-h":!0}),f(An)!=="svelte-d80i8q"&&(An.innerHTML=s_),ql=t(re),kn=s(re,"P",{"data-svelte-h":!0}),f(kn)!=="svelte-1ge10rl"&&(kn.innerHTML=i_),zl=t(re),Pn=s(re,"P",{"data-svelte-h":!0}),f(Pn)!=="svelte-1iu9wlz"&&(Pn.innerHTML=d_),re.forEach(i),Bl=t(O),Le=s(O,"DIV",{class:!0});var Qe=v(Le);m(Or.$$.fragment,Qe),jl=t(Qe),Rn=s(Qe,"P",{"data-svelte-h":!0}),f(Rn)!=="svelte-flusvq"&&(Rn.textContent=l_),Gl=t(Qe),m(_o.$$.fragment,Qe),Qe.forEach(i),Jl=t(O),uo=s(O,"DIV",{class:!0});var tn=v(uo);m(Yr.$$.fragment,tn),Zl=t(tn),Hn=s(tn,"P",{"data-svelte-h":!0}),f(Hn)!=="svelte-1ufq5ot"&&(Hn.textContent=c_),tn.forEach(i),O.forEach(i),ad=t(e),m(Qr.$$.fragment,e),nd=t(e),k=s(e,"DIV",{class:!0});var R=v(k);m(Kr.$$.fragment,R),Ol=t(R),In=s(R,"P",{"data-svelte-h":!0}),f(In)!=="svelte-1uqumi3"&&(In.innerHTML=f_),Yl=t(R),Vn=s(R,"P",{"data-svelte-h":!0}),f(Vn)!=="svelte-1dbjlx1"&&(Vn.innerHTML=p_),Ql=t(R),ho=s(R,"DIV",{class:!0});var an=v(ho);m(et.$$.fragment,an),Kl=t(an),Wn=s(an,"P",{"data-svelte-h":!0}),f(Wn)!=="svelte-1062ci4"&&(Wn.innerHTML=m_),an.forEach(i),ec=t(R),go=s(R,"DIV",{class:!0});var nn=v(go);m(ot.$$.fragment,nn),oc=t(nn),Fn=s(nn,"P",{"data-svelte-h":!0}),f(Fn)!=="svelte-1lgbsz7"&&(Fn.innerHTML=__),nn.forEach(i),rc=t(R),te=s(R,"DIV",{class:!0});var _e=v(te);m(rt.$$.fragment,_e),tc=t(_e),Un=s(_e,"P",{"data-svelte-h":!0}),f(Un)!=="svelte-vs7s0z"&&(Un.innerHTML=u_),ac=t(_e),En=s(_e,"P",{"data-svelte-h":!0}),f(En)!=="svelte-15b960v"&&(En.innerHTML=h_),nc=t(_e),Nn=s(_e,"P",{"data-svelte-h":!0}),f(Nn)!=="svelte-d80i8q"&&(Nn.innerHTML=g_),sc=t(_e),Xn=s(_e,"P",{"data-svelte-h":!0}),f(Xn)!=="svelte-1ukghd4"&&(Xn.innerHTML=L_),_e.forEach(i),ic=t(R),xe=s(R,"DIV",{class:!0});var Ke=v(xe);m(tt.$$.fragment,Ke),dc=t(Ke),qn=s(Ke,"P",{"data-svelte-h":!0}),f(qn)!=="svelte-flusvq"&&(qn.textContent=x_),lc=t(Ke),m(Lo.$$.fragment,Ke),Ke.forEach(i),cc=t(R),xo=s(R,"DIV",{class:!0});var sn=v(xo);m(at.$$.fragment,sn),fc=t(sn),zn=s(sn,"P",{"data-svelte-h":!0}),f(zn)!=="svelte-1ufq5ot"&&(zn.textContent=b_),sn.forEach(i),pc=t(R),be=s(R,"DIV",{class:!0});var eo=v(be);m(nt.$$.fragment,eo),mc=t(eo),Bn=s(eo,"P",{"data-svelte-h":!0}),f(Bn)!=="svelte-ioswce"&&(Bn.innerHTML=v_),_c=t(eo),m(bo.$$.fragment,eo),eo.forEach(i),R.forEach(i),sd=t(e),m(st.$$.fragment,e),id=t(e),C=s(e,"DIV",{class:!0});var P=v(C);m(it.$$.fragment,P),uc=t(P),jn=s(P,"P",{"data-svelte-h":!0}),f(jn)!=="svelte-oktj5e"&&(jn.innerHTML=w_),hc=t(P),Gn=s(P,"P",{"data-svelte-h":!0}),f(Gn)!=="svelte-1dbjlx1"&&(Gn.innerHTML=$_),gc=t(P),vo=s(P,"DIV",{class:!0});var dn=v(vo);m(dt.$$.fragment,dn),Lc=t(dn),Jn=s(dn,"P",{"data-svelte-h":!0}),f(Jn)!=="svelte-1062ci4"&&(Jn.innerHTML=M_),dn.forEach(i),xc=t(P),wo=s(P,"DIV",{class:!0});var ln=v(wo);m(lt.$$.fragment,ln),bc=t(ln),Zn=s(ln,"P",{"data-svelte-h":!0}),f(Zn)!=="svelte-1lgbsz7"&&(Zn.innerHTML=y_),ln.forEach(i),vc=t(P),ae=s(P,"DIV",{class:!0});var ue=v(ae);m(ct.$$.fragment,ue),wc=t(ue),On=s(ue,"P",{"data-svelte-h":!0}),f(On)!=="svelte-178gcly"&&(On.innerHTML=T_),$c=t(ue),Yn=s(ue,"P",{"data-svelte-h":!0}),f(Yn)!=="svelte-15b960v"&&(Yn.innerHTML=D_),Mc=t(ue),Qn=s(ue,"P",{"data-svelte-h":!0}),f(Qn)!=="svelte-d80i8q"&&(Qn.innerHTML=S_),yc=t(ue),Kn=s(ue,"P",{"data-svelte-h":!0}),f(Kn)!=="svelte-1ukghd4"&&(Kn.innerHTML=C_),ue.forEach(i),Tc=t(P),ve=s(P,"DIV",{class:!0});var oo=v(ve);m(ft.$$.fragment,oo),Dc=t(oo),es=s(oo,"P",{"data-svelte-h":!0}),f(es)!=="svelte-flusvq"&&(es.textContent=A_),Sc=t(oo),m($o.$$.fragment,oo),oo.forEach(i),Cc=t(P),Mo=s(P,"DIV",{class:!0});var cn=v(Mo);m(pt.$$.fragment,cn),Ac=t(cn),os=s(cn,"P",{"data-svelte-h":!0}),f(os)!=="svelte-1ufq5ot"&&(os.textContent=k_),cn.forEach(i),kc=t(P),we=s(P,"DIV",{class:!0});var ro=v(we);m(mt.$$.fragment,ro),Pc=t(ro),rs=s(ro,"P",{"data-svelte-h":!0}),f(rs)!=="svelte-ioswce"&&(rs.innerHTML=P_),Rc=t(ro),m(yo.$$.fragment,ro),ro.forEach(i),Hc=t(P),$e=s(P,"DIV",{class:!0});var to=v($e);m(_t.$$.fragment,to),Ic=t(to),ts=s(to,"P",{"data-svelte-h":!0}),f(ts)!=="svelte-119cgd9"&&(ts.textContent=R_),Vc=t(to),m(To.$$.fragment,to),to.forEach(i),P.forEach(i),dd=t(e),m(ut.$$.fragment,e),ld=t(e),W=s(e,"DIV",{class:!0});var Y=v(W);m(ht.$$.fragment,Y),Wc=t(Y),as=s(Y,"P",{"data-svelte-h":!0}),f(as)!=="svelte-19msb7v"&&(as.innerHTML=H_),Fc=t(Y),Do=s(Y,"DIV",{class:!0});var fn=v(Do);m(gt.$$.fragment,fn),Uc=t(fn),ns=s(fn,"P",{"data-svelte-h":!0}),f(ns)!=="svelte-1lgbsz7"&&(ns.innerHTML=I_),fn.forEach(i),Ec=t(Y),So=s(Y,"DIV",{class:!0});var pn=v(So);m(Lt.$$.fragment,pn),Nc=t(pn),ss=s(pn,"P",{"data-svelte-h":!0}),f(ss)!=="svelte-1l29c27"&&(ss.innerHTML=V_),pn.forEach(i),Xc=t(Y),Me=s(Y,"DIV",{class:!0});var ao=v(Me);m(xt.$$.fragment,ao),qc=t(ao),is=s(ao,"P",{"data-svelte-h":!0}),f(is)!=="svelte-flusvq"&&(is.textContent=W_),zc=t(ao),m(Co.$$.fragment,ao),ao.forEach(i),Bc=t(Y),Ao=s(Y,"DIV",{class:!0});var mn=v(Ao);m(bt.$$.fragment,mn),jc=t(mn),ds=s(mn,"P",{"data-svelte-h":!0}),f(ds)!=="svelte-1ufq5ot"&&(ds.textContent=F_),mn.forEach(i),Gc=t(Y),ye=s(Y,"DIV",{class:!0});var no=v(ye);m(vt.$$.fragment,no),Jc=t(no),ls=s(no,"P",{"data-svelte-h":!0}),f(ls)!=="svelte-ioswce"&&(ls.innerHTML=U_),Zc=t(no),m(ko.$$.fragment,no),no.forEach(i),Y.forEach(i),cd=t(e),m(wt.$$.fragment,e),fd=t(e),F=s(e,"DIV",{class:!0});var Q=v(F);m($t.$$.fragment,Q),Oc=t(Q),cs=s(Q,"P",{"data-svelte-h":!0}),f(cs)!=="svelte-1vcxpyf"&&(cs.innerHTML=E_),Yc=t(Q),Po=s(Q,"DIV",{class:!0});var _n=v(Po);m(Mt.$$.fragment,_n),Qc=t(_n),fs=s(_n,"P",{"data-svelte-h":!0}),f(fs)!=="svelte-1lgbsz7"&&(fs.innerHTML=N_),_n.forEach(i),Kc=t(Q),Ro=s(Q,"DIV",{class:!0});var Wd=v(Ro);m(yt.$$.fragment,Wd),ef=t(Wd),ps=s(Wd,"P",{"data-svelte-h":!0}),f(ps)!=="svelte-1l29c27"&&(ps.innerHTML=X_),Wd.forEach(i),of=t(Q),Te=s(Q,"DIV",{class:!0});var Di=v(Te);m(Tt.$$.fragment,Di),rf=t(Di),ms=s(Di,"P",{"data-svelte-h":!0}),f(ms)!=="svelte-flusvq"&&(ms.textContent=q_),tf=t(Di),m(Ho.$$.fragment,Di),Di.forEach(i),af=t(Q),Io=s(Q,"DIV",{class:!0});var Fd=v(Io);m(Dt.$$.fragment,Fd),nf=t(Fd),_s=s(Fd,"P",{"data-svelte-h":!0}),f(_s)!=="svelte-1ufq5ot"&&(_s.textContent=z_),Fd.forEach(i),sf=t(Q),De=s(Q,"DIV",{class:!0});var Si=v(De);m(St.$$.fragment,Si),df=t(Si),us=s(Si,"P",{"data-svelte-h":!0}),f(us)!=="svelte-ioswce"&&(us.innerHTML=B_),lf=t(Si),m(Vo.$$.fragment,Si),Si.forEach(i),Q.forEach(i),pd=t(e),m(Ct.$$.fragment,e),md=t(e),U=s(e,"DIV",{class:!0});var ne=v(U);m(At.$$.fragment,ne),cf=t(ne),hs=s(ne,"P",{"data-svelte-h":!0}),f(hs)!=="svelte-19pa8s"&&(hs.innerHTML=j_),ff=t(ne),Wo=s(ne,"DIV",{class:!0});var Ud=v(Wo);m(kt.$$.fragment,Ud),pf=t(Ud),gs=s(Ud,"P",{"data-svelte-h":!0}),f(gs)!=="svelte-1lgbsz7"&&(gs.innerHTML=G_),Ud.forEach(i),mf=t(ne),Fo=s(ne,"DIV",{class:!0});var Ed=v(Fo);m(Pt.$$.fragment,Ed),_f=t(Ed),Ls=s(Ed,"P",{"data-svelte-h":!0}),f(Ls)!=="svelte-1l29c27"&&(Ls.innerHTML=J_),Ed.forEach(i),uf=t(ne),Se=s(ne,"DIV",{class:!0});var Ci=v(Se);m(Rt.$$.fragment,Ci),hf=t(Ci),xs=s(Ci,"P",{"data-svelte-h":!0}),f(xs)!=="svelte-flusvq"&&(xs.textContent=Z_),gf=t(Ci),m(Uo.$$.fragment,Ci),Ci.forEach(i),Lf=t(ne),Eo=s(ne,"DIV",{class:!0});var Nd=v(Eo);m(Ht.$$.fragment,Nd),xf=t(Nd),bs=s(Nd,"P",{"data-svelte-h":!0}),f(bs)!=="svelte-1ufq5ot"&&(bs.textContent=O_),Nd.forEach(i),bf=t(ne),Ce=s(ne,"DIV",{class:!0});var Ai=v(Ce);m(It.$$.fragment,Ai),vf=t(Ai),vs=s(Ai,"P",{"data-svelte-h":!0}),f(vs)!=="svelte-ioswce"&&(vs.innerHTML=Y_),wf=t(Ai),m(No.$$.fragment,Ai),Ai.forEach(i),ne.forEach(i),_d=t(e),m(Vt.$$.fragment,e),ud=t(e),E=s(e,"DIV",{class:!0});var se=v(E);m(Wt.$$.fragment,se),$f=t(se),ws=s(se,"P",{"data-svelte-h":!0}),f(ws)!=="svelte-419qa7"&&(ws.innerHTML=Q_),Mf=t(se),Xo=s(se,"DIV",{class:!0});var Xd=v(Xo);m(Ft.$$.fragment,Xd),yf=t(Xd),$s=s(Xd,"P",{"data-svelte-h":!0}),f($s)!=="svelte-1lgbsz7"&&($s.innerHTML=K_),Xd.forEach(i),Tf=t(se),qo=s(se,"DIV",{class:!0});var qd=v(qo);m(Ut.$$.fragment,qd),Df=t(qd),Ms=s(qd,"P",{"data-svelte-h":!0}),f(Ms)!=="svelte-1l29c27"&&(Ms.innerHTML=eu),qd.forEach(i),Sf=t(se),Ae=s(se,"DIV",{class:!0});var ki=v(Ae);m(Et.$$.fragment,ki),Cf=t(ki),ys=s(ki,"P",{"data-svelte-h":!0}),f(ys)!=="svelte-flusvq"&&(ys.textContent=ou),Af=t(ki),m(zo.$$.fragment,ki),ki.forEach(i),kf=t(se),Bo=s(se,"DIV",{class:!0});var zd=v(Bo);m(Nt.$$.fragment,zd),Pf=t(zd),Ts=s(zd,"P",{"data-svelte-h":!0}),f(Ts)!=="svelte-1ufq5ot"&&(Ts.textContent=ru),zd.forEach(i),Rf=t(se),ke=s(se,"DIV",{class:!0});var Pi=v(ke);m(Xt.$$.fragment,Pi),Hf=t(Pi),Ds=s(Pi,"P",{"data-svelte-h":!0}),f(Ds)!=="svelte-ioswce"&&(Ds.innerHTML=tu),If=t(Pi),m(jo.$$.fragment,Pi),Pi.forEach(i),se.forEach(i),hd=t(e),m(qt.$$.fragment,e),gd=t(e),N=s(e,"DIV",{class:!0});var ie=v(N);m(zt.$$.fragment,ie),Vf=t(ie),Ss=s(ie,"P",{"data-svelte-h":!0}),f(Ss)!=="svelte-18mhu58"&&(Ss.innerHTML=au),Wf=t(ie),Go=s(ie,"DIV",{class:!0});var Bd=v(Go);m(Bt.$$.fragment,Bd),Ff=t(Bd),Cs=s(Bd,"P",{"data-svelte-h":!0}),f(Cs)!=="svelte-1lgbsz7"&&(Cs.innerHTML=nu),Bd.forEach(i),Uf=t(ie),Jo=s(ie,"DIV",{class:!0});var jd=v(Jo);m(jt.$$.fragment,jd),Ef=t(jd),As=s(jd,"P",{"data-svelte-h":!0}),f(As)!=="svelte-1l29c27"&&(As.innerHTML=su),jd.forEach(i),Nf=t(ie),Pe=s(ie,"DIV",{class:!0});var Ri=v(Pe);m(Gt.$$.fragment,Ri),Xf=t(Ri),ks=s(Ri,"P",{"data-svelte-h":!0}),f(ks)!=="svelte-flusvq"&&(ks.textContent=iu),qf=t(Ri),m(Zo.$$.fragment,Ri),Ri.forEach(i),zf=t(ie),Oo=s(ie,"DIV",{class:!0});var Gd=v(Oo);m(Jt.$$.fragment,Gd),Bf=t(Gd),Ps=s(Gd,"P",{"data-svelte-h":!0}),f(Ps)!=="svelte-1ufq5ot"&&(Ps.textContent=du),Gd.forEach(i),jf=t(ie),Re=s(ie,"DIV",{class:!0});var Hi=v(Re);m(Zt.$$.fragment,Hi),Gf=t(Hi),Rs=s(Hi,"P",{"data-svelte-h":!0}),f(Rs)!=="svelte-ioswce"&&(Rs.innerHTML=lu),Jf=t(Hi),m(Yo.$$.fragment,Hi),Hi.forEach(i),ie.forEach(i),Ld=t(e),m(Ot.$$.fragment,e),xd=t(e),X=s(e,"DIV",{class:!0});var de=v(X);m(Yt.$$.fragment,de),Zf=t(de),Hs=s(de,"P",{"data-svelte-h":!0}),f(Hs)!=="svelte-4uev2g"&&(Hs.innerHTML=cu),Of=t(de),Qo=s(de,"DIV",{class:!0});var Jd=v(Qo);m(Qt.$$.fragment,Jd),Yf=t(Jd),Is=s(Jd,"P",{"data-svelte-h":!0}),f(Is)!=="svelte-1lgbsz7"&&(Is.innerHTML=fu),Jd.forEach(i),Qf=t(de),Ko=s(de,"DIV",{class:!0});var Zd=v(Ko);m(Kt.$$.fragment,Zd),Kf=t(Zd),Vs=s(Zd,"P",{"data-svelte-h":!0}),f(Vs)!=="svelte-1l29c27"&&(Vs.innerHTML=pu),Zd.forEach(i),ep=t(de),He=s(de,"DIV",{class:!0});var Ii=v(He);m(ea.$$.fragment,Ii),op=t(Ii),Ws=s(Ii,"P",{"data-svelte-h":!0}),f(Ws)!=="svelte-flusvq"&&(Ws.textContent=mu),rp=t(Ii),m(er.$$.fragment,Ii),Ii.forEach(i),tp=t(de),or=s(de,"DIV",{class:!0});var Od=v(or);m(oa.$$.fragment,Od),ap=t(Od),Fs=s(Od,"P",{"data-svelte-h":!0}),f(Fs)!=="svelte-1ufq5ot"&&(Fs.textContent=_u),Od.forEach(i),np=t(de),Ie=s(de,"DIV",{class:!0});var Vi=v(Ie);m(ra.$$.fragment,Vi),sp=t(Vi),Us=s(Vi,"P",{"data-svelte-h":!0}),f(Us)!=="svelte-ioswce"&&(Us.innerHTML=uu),ip=t(Vi),m(rr.$$.fragment,Vi),Vi.forEach(i),de.forEach(i),bd=t(e),m(ta.$$.fragment,e),vd=t(e),q=s(e,"DIV",{class:!0});var le=v(q);m(aa.$$.fragment,le),dp=t(le),Es=s(le,"P",{"data-svelte-h":!0}),f(Es)!=="svelte-1kyn2fc"&&(Es.innerHTML=hu),lp=t(le),tr=s(le,"DIV",{class:!0});var Yd=v(tr);m(na.$$.fragment,Yd),cp=t(Yd),Ns=s(Yd,"P",{"data-svelte-h":!0}),f(Ns)!=="svelte-1lgbsz7"&&(Ns.innerHTML=gu),Yd.forEach(i),fp=t(le),ar=s(le,"DIV",{class:!0});var Qd=v(ar);m(sa.$$.fragment,Qd),pp=t(Qd),Xs=s(Qd,"P",{"data-svelte-h":!0}),f(Xs)!=="svelte-1l29c27"&&(Xs.innerHTML=Lu),Qd.forEach(i),mp=t(le),Ve=s(le,"DIV",{class:!0});var Wi=v(Ve);m(ia.$$.fragment,Wi),_p=t(Wi),qs=s(Wi,"P",{"data-svelte-h":!0}),f(qs)!=="svelte-flusvq"&&(qs.textContent=xu),up=t(Wi),m(nr.$$.fragment,Wi),Wi.forEach(i),hp=t(le),sr=s(le,"DIV",{class:!0});var Kd=v(sr);m(da.$$.fragment,Kd),gp=t(Kd),zs=s(Kd,"P",{"data-svelte-h":!0}),f(zs)!=="svelte-1ufq5ot"&&(zs.textContent=bu),Kd.forEach(i),Lp=t(le),We=s(le,"DIV",{class:!0});var Fi=v(We);m(la.$$.fragment,Fi),xp=t(Fi),Bs=s(Fi,"P",{"data-svelte-h":!0}),f(Bs)!=="svelte-ioswce"&&(Bs.innerHTML=vu),bp=t(Fi),m(ir.$$.fragment,Fi),Fi.forEach(i),le.forEach(i),wd=t(e),m(ca.$$.fragment,e),$d=t(e),z=s(e,"DIV",{class:!0});var ce=v(z);m(fa.$$.fragment,ce),vp=t(ce),js=s(ce,"P",{"data-svelte-h":!0}),f(js)!=="svelte-1flrsvi"&&(js.innerHTML=wu),wp=t(ce),dr=s(ce,"DIV",{class:!0});var el=v(dr);m(pa.$$.fragment,el),$p=t(el),Gs=s(el,"P",{"data-svelte-h":!0}),f(Gs)!=="svelte-1lgbsz7"&&(Gs.innerHTML=$u),el.forEach(i),Mp=t(ce),lr=s(ce,"DIV",{class:!0});var ol=v(lr);m(ma.$$.fragment,ol),yp=t(ol),Js=s(ol,"P",{"data-svelte-h":!0}),f(Js)!=="svelte-1l29c27"&&(Js.innerHTML=Mu),ol.forEach(i),Tp=t(ce),Fe=s(ce,"DIV",{class:!0});var Ui=v(Fe);m(_a.$$.fragment,Ui),Dp=t(Ui),Zs=s(Ui,"P",{"data-svelte-h":!0}),f(Zs)!=="svelte-flusvq"&&(Zs.textContent=yu),Sp=t(Ui),m(cr.$$.fragment,Ui),Ui.forEach(i),Cp=t(ce),fr=s(ce,"DIV",{class:!0});var rl=v(fr);m(ua.$$.fragment,rl),Ap=t(rl),Os=s(rl,"P",{"data-svelte-h":!0}),f(Os)!=="svelte-1ufq5ot"&&(Os.textContent=Tu),rl.forEach(i),kp=t(ce),Ue=s(ce,"DIV",{class:!0});var Ei=v(Ue);m(ha.$$.fragment,Ei),Pp=t(Ei),Ys=s(Ei,"P",{"data-svelte-h":!0}),f(Ys)!=="svelte-ioswce"&&(Ys.innerHTML=Du),Rp=t(Ei),m(pr.$$.fragment,Ei),Ei.forEach(i),ce.forEach(i),Md=t(e),m(ga.$$.fragment,e),yd=t(e),B=s(e,"DIV",{class:!0});var fe=v(B);m(La.$$.fragment,fe),Hp=t(fe),Qs=s(fe,"P",{"data-svelte-h":!0}),f(Qs)!=="svelte-k340nx"&&(Qs.innerHTML=Su),Ip=t(fe),mr=s(fe,"DIV",{class:!0});var tl=v(mr);m(xa.$$.fragment,tl),Vp=t(tl),Ks=s(tl,"P",{"data-svelte-h":!0}),f(Ks)!=="svelte-1lgbsz7"&&(Ks.innerHTML=Cu),tl.forEach(i),Wp=t(fe),_r=s(fe,"DIV",{class:!0});var al=v(_r);m(ba.$$.fragment,al),Fp=t(al),ei=s(al,"P",{"data-svelte-h":!0}),f(ei)!=="svelte-1l29c27"&&(ei.innerHTML=Au),al.forEach(i),Up=t(fe),Ee=s(fe,"DIV",{class:!0});var Ni=v(Ee);m(va.$$.fragment,Ni),Ep=t(Ni),oi=s(Ni,"P",{"data-svelte-h":!0}),f(oi)!=="svelte-flusvq"&&(oi.textContent=ku),Np=t(Ni),m(ur.$$.fragment,Ni),Ni.forEach(i),Xp=t(fe),hr=s(fe,"DIV",{class:!0});var nl=v(hr);m(wa.$$.fragment,nl),qp=t(nl),ri=s(nl,"P",{"data-svelte-h":!0}),f(ri)!=="svelte-1ufq5ot"&&(ri.textContent=Pu),nl.forEach(i),zp=t(fe),Ne=s(fe,"DIV",{class:!0});var Xi=v(Ne);m($a.$$.fragment,Xi),Bp=t(Xi),ti=s(Xi,"P",{"data-svelte-h":!0}),f(ti)!=="svelte-ioswce"&&(ti.innerHTML=Ru),jp=t(Xi),m(gr.$$.fragment,Xi),Xi.forEach(i),fe.forEach(i),Td=t(e),m(Ma.$$.fragment,e),Dd=t(e),Oe=s(e,"DIV",{class:!0});var sl=v(Oe);m(ya.$$.fragment,sl),Gp=t(sl),Lr=s(sl,"DIV",{class:!0});var il=v(Lr);m(Ta.$$.fragment,il),Jp=t(il),ai=s(il,"P",{"data-svelte-h":!0}),f(ai)!=="svelte-1lgbsz7"&&(ai.innerHTML=Hu),il.forEach(i),sl.forEach(i),Sd=t(e),m(Da.$$.fragment,e),Cd=t(e),j=s(e,"DIV",{class:!0});var pe=v(j);m(Sa.$$.fragment,pe),Zp=t(pe),ni=s(pe,"P",{"data-svelte-h":!0}),f(ni)!=="svelte-s6gvfy"&&(ni.innerHTML=Iu),Op=t(pe),xr=s(pe,"DIV",{class:!0});var dl=v(xr);m(Ca.$$.fragment,dl),Yp=t(dl),si=s(dl,"P",{"data-svelte-h":!0}),f(si)!=="svelte-1lgbsz7"&&(si.innerHTML=Vu),dl.forEach(i),Qp=t(pe),br=s(pe,"DIV",{class:!0});var ll=v(br);m(Aa.$$.fragment,ll),Kp=t(ll),ii=s(ll,"P",{"data-svelte-h":!0}),f(ii)!=="svelte-1l29c27"&&(ii.innerHTML=Wu),ll.forEach(i),em=t(pe),Xe=s(pe,"DIV",{class:!0});var qi=v(Xe);m(ka.$$.fragment,qi),om=t(qi),di=s(qi,"P",{"data-svelte-h":!0}),f(di)!=="svelte-flusvq"&&(di.textContent=Fu),rm=t(qi),m(vr.$$.fragment,qi),qi.forEach(i),tm=t(pe),wr=s(pe,"DIV",{class:!0});var cl=v(wr);m(Pa.$$.fragment,cl),am=t(cl),li=s(cl,"P",{"data-svelte-h":!0}),f(li)!=="svelte-1ufq5ot"&&(li.textContent=Uu),cl.forEach(i),nm=t(pe),qe=s(pe,"DIV",{class:!0});var zi=v(qe);m(Ra.$$.fragment,zi),sm=t(zi),ci=s(zi,"P",{"data-svelte-h":!0}),f(ci)!=="svelte-ioswce"&&(ci.innerHTML=Eu),im=t(zi),m($r.$$.fragment,zi),zi.forEach(i),pe.forEach(i),Ad=t(e),m(Ha.$$.fragment,e),kd=t(e),A=s(e,"DIV",{class:!0});var H=v(A);m(Ia.$$.fragment,H),dm=t(H),fi=s(H,"P",{"data-svelte-h":!0}),f(fi)!=="svelte-1q4bbx"&&(fi.textContent=Nu),lm=t(H),pi=s(H,"DIV",{class:!0});var rh=v(pi);m(Va.$$.fragment,rh),rh.forEach(i),cm=t(H),ze=s(H,"DIV",{class:!0});var Bi=v(ze);m(Wa.$$.fragment,Bi),fm=t(Bi),mi=s(Bi,"P",{"data-svelte-h":!0}),f(mi)!=="svelte-w22v11"&&(mi.textContent=Xu),pm=t(Bi),_i=s(Bi,"P",{"data-svelte-h":!0}),f(_i)!=="svelte-zdadya"&&(_i.textContent=qu),Bi.forEach(i),mm=t(H),he=s(H,"DIV",{class:!0});var Vr=v(he);m(Fa.$$.fragment,Vr),_m=t(Vr),ui=s(Vr,"P",{"data-svelte-h":!0}),f(ui)!=="svelte-1nr2dy0"&&(ui.textContent=zu),um=t(Vr),m(Mr.$$.fragment,Vr),hm=t(Vr),m(yr.$$.fragment,Vr),Vr.forEach(i),gm=t(H),Be=s(H,"DIV",{class:!0});var ji=v(Be);m(Ua.$$.fragment,ji),Lm=t(ji),hi=s(ji,"P",{"data-svelte-h":!0}),f(hi)!=="svelte-h0os0v"&&(hi.textContent=Bu),xm=t(ji),m(Tr.$$.fragment,ji),ji.forEach(i),bm=t(H),Dr=s(H,"DIV",{class:!0});var fl=v(Dr);m(Ea.$$.fragment,fl),vm=t(fl),gi=s(fl,"P",{"data-svelte-h":!0}),f(gi)!=="svelte-1825k9e"&&(gi.textContent=ju),fl.forEach(i),wm=t(H),Sr=s(H,"DIV",{class:!0});var pl=v(Sr);m(Na.$$.fragment,pl),$m=t(pl),Li=s(pl,"P",{"data-svelte-h":!0}),f(Li)!=="svelte-rvubqa"&&(Li.innerHTML=Gu),pl.forEach(i),Mm=t(H),je=s(H,"DIV",{class:!0});var Gi=v(je);m(Xa.$$.fragment,Gi),ym=t(Gi),xi=s(Gi,"P",{"data-svelte-h":!0}),f(xi)!=="svelte-ioswce"&&(xi.innerHTML=Ju),Tm=t(Gi),m(Cr.$$.fragment,Gi),Gi.forEach(i),Dm=t(H),Ge=s(H,"DIV",{class:!0});var Ji=v(Ge);m(qa.$$.fragment,Ji),Sm=t(Ji),bi=s(Ji,"P",{"data-svelte-h":!0}),f(bi)!=="svelte-119cgd9"&&(bi.textContent=Zu),Cm=t(Ji),m(Ar.$$.fragment,Ji),Ji.forEach(i),H.forEach(i),Pd=t(e),m(za.$$.fragment,e),Rd=t(e),G=s(e,"DIV",{class:!0});var me=v(G);m(Ba.$$.fragment,me),Am=t(me),vi=s(me,"P",{"data-svelte-h":!0}),f(vi)!=="svelte-k340nx"&&(vi.innerHTML=Ou),km=t(me),kr=s(me,"DIV",{class:!0});var ml=v(kr);m(ja.$$.fragment,ml),Pm=t(ml),wi=s(ml,"P",{"data-svelte-h":!0}),f(wi)!=="svelte-1lgbsz7"&&(wi.innerHTML=Yu),ml.forEach(i),Rm=t(me),Pr=s(me,"DIV",{class:!0});var _l=v(Pr);m(Ga.$$.fragment,_l),Hm=t(_l),$i=s(_l,"P",{"data-svelte-h":!0}),f($i)!=="svelte-1l29c27"&&($i.innerHTML=Qu),_l.forEach(i),Im=t(me),Je=s(me,"DIV",{class:!0});var Zi=v(Je);m(Ja.$$.fragment,Zi),Vm=t(Zi),Mi=s(Zi,"P",{"data-svelte-h":!0}),f(Mi)!=="svelte-flusvq"&&(Mi.textContent=Ku),Wm=t(Zi),m(Rr.$$.fragment,Zi),Zi.forEach(i),Fm=t(me),Hr=s(me,"DIV",{class:!0});var ul=v(Hr);m(Za.$$.fragment,ul),Um=t(ul),yi=s(ul,"P",{"data-svelte-h":!0}),f(yi)!=="svelte-1ufq5ot"&&(yi.textContent=eh),ul.forEach(i),Em=t(me),Ze=s(me,"DIV",{class:!0});var Oi=v(Ze);m(Oa.$$.fragment,Oi),Nm=t(Oi),Ti=s(Oi,"P",{"data-svelte-h":!0}),f(Ti)!=="svelte-ioswce"&&(Ti.innerHTML=oh),Xm=t(Oi),m(Ir.$$.fragment,Oi),Oi.forEach(i),me.forEach(i),Hd=t(e),m(Ya.$$.fragment,e),Id=t(e),Yi=s(e,"P",{}),v(Yi).forEach(i),this.h()},h(){w(a,"name","hf:doc:metadata"),w(a,"content",Bh),w(io,"class","docstring border-l-2 border-t-2 pl-4 pt-3.5 border-gray-100 rounded-tl-xl mb-6 mt-8"),w(lo,"class","docstring border-l-2 border-t-2 pl-4 pt-3.5 border-gray-100 rounded-tl-xl mb-6 mt-8"),w(K,"class","docstring border-l-2 border-t-2 pl-4 pt-3.5 border-gray-100 rounded-tl-xl mb-6 mt-8"),w(ge,"class","docstring border-l-2 border-t-2 pl-4 pt-3.5 border-gray-100 rounded-tl-xl mb-6 mt-8"),w(fo,"class","docstring border-l-2 border-t-2 pl-4 pt-3.5 border-gray-100 rounded-tl-xl mb-6 mt-8"),w(I,"class","docstring border-l-2 border-t-2 pl-4 pt-3.5 border-gray-100 rounded-tl-xl mb-6 mt-8"),w(po,"class","docstring border-l-2 border-t-2 pl-4 pt-3.5 border-gray-100 rounded-tl-xl mb-6 mt-8"),w(mo,"class","docstring border-l-2 border-t-2 pl-4 pt-3.5 border-gray-100 rounded-tl-xl mb-6 mt-8"),w(ee,"class","docstring border-l-2 border-t-2 pl-4 pt-3.5 border-gray-100 rounded-tl-xl mb-6 mt-8"),w(Le,"class","docstring border-l-2 border-t-2 pl-4 pt-3.5 border-gray-100 rounded-tl-xl mb-6 mt-8"),w(uo,"class","docstring border-l-2 border-t-2 pl-4 pt-3.5 border-gray-100 rounded-tl-xl mb-6 mt-8"),w(V,"class","docstring border-l-2 border-t-2 pl-4 pt-3.5 border-gray-100 rounded-tl-xl mb-6 mt-8"),w(ho,"class","docstring border-l-2 border-t-2 pl-4 pt-3.5 border-gray-100 rounded-tl-xl mb-6 mt-8"),w(go,"class","docstring border-l-2 border-t-2 pl-4 pt-3.5 border-gray-100 rounded-tl-xl mb-6 mt-8"),w(te,"class","docstring border-l-2 border-t-2 pl-4 pt-3.5 border-gray-100 rounded-tl-xl mb-6 mt-8"),w(xe,"class","docstring border-l-2 border-t-2 pl-4 pt-3.5 border-gray-100 rounded-tl-xl mb-6 mt-8"),w(xo,"class","docstring border-l-2 border-t-2 pl-4 pt-3.5 border-gray-100 rounded-tl-xl mb-6 mt-8"),w(be,"class","docstring border-l-2 border-t-2 pl-4 pt-3.5 border-gray-100 rounded-tl-xl mb-6 mt-8"),w(k,"class","docstring border-l-2 border-t-2 pl-4 pt-3.5 border-gray-100 rounded-tl-xl mb-6 mt-8"),w(vo,"class","docstring border-l-2 border-t-2 pl-4 pt-3.5 border-gray-100 rounded-tl-xl mb-6 mt-8"),w(wo,"class","docstring border-l-2 border-t-2 pl-4 pt-3.5 border-gray-100 rounded-tl-xl mb-6 mt-8"),w(ae,"class","docstring border-l-2 border-t-2 pl-4 pt-3.5 border-gray-100 rounded-tl-xl mb-6 mt-8"),w(ve,"class","docstring border-l-2 border-t-2 pl-4 pt-3.5 border-gray-100 rounded-tl-xl mb-6 mt-8"),w(Mo,"class","docstring border-l-2 border-t-2 pl-4 pt-3.5 border-gray-100 rounded-tl-xl mb-6 mt-8"),w(we,"class","docstring border-l-2 border-t-2 pl-4 pt-3.5 border-gray-100 rounded-tl-xl mb-6 mt-8"),w($e,"class","docstring border-l-2 border-t-2 pl-4 pt-3.5 border-gray-100 rounded-tl-xl mb-6 mt-8"),w(C,"class","docstring border-l-2 border-t-2 pl-4 pt-3.5 border-gray-100 rounded-tl-xl mb-6 mt-8"),w(Do,"class","docstring border-l-2 border-t-2 pl-4 pt-3.5 border-gray-100 rounded-tl-xl mb-6 mt-8"),w(So,"class","docstring border-l-2 border-t-2 pl-4 pt-3.5 border-gray-100 rounded-tl-xl mb-6 mt-8"),w(Me,"class","docstring border-l-2 border-t-2 pl-4 pt-3.5 border-gray-100 rounded-tl-xl mb-6 mt-8"),w(Ao,"class","docstring border-l-2 border-t-2 pl-4 pt-3.5 border-gray-100 rounded-tl-xl mb-6 mt-8"),w(ye,"class","docstring border-l-2 border-t-2 pl-4 pt-3.5 border-gray-100 rounded-tl-xl mb-6 mt-8"),w(W,"class","docstring border-l-2 border-t-2 pl-4 pt-3.5 border-gray-100 rounded-tl-xl mb-6 mt-8"),w(Po,"class","docstring border-l-2 border-t-2 pl-4 pt-3.5 border-gray-100 rounded-tl-xl mb-6 mt-8"),w(Ro,"class","docstring border-l-2 border-t-2 pl-4 pt-3.5 border-gray-100 rounded-tl-xl mb-6 mt-8"),w(Te,"class","docstring border-l-2 border-t-2 pl-4 pt-3.5 border-gray-100 rounded-tl-xl mb-6 mt-8"),w(Io,"class","docstring border-l-2 border-t-2 pl-4 pt-3.5 border-gray-100 rounded-tl-xl mb-6 mt-8"),w(De,"class","docstring border-l-2 border-t-2 pl-4 pt-3.5 border-gray-100 rounded-tl-xl mb-6 mt-8"),w(F,"class","docstring border-l-2 border-t-2 pl-4 pt-3.5 border-gray-100 rounded-tl-xl mb-6 mt-8"),w(Wo,"class","docstring border-l-2 border-t-2 pl-4 pt-3.5 border-gray-100 rounded-tl-xl mb-6 mt-8"),w(Fo,"class","docstring border-l-2 border-t-2 pl-4 pt-3.5 border-gray-100 rounded-tl-xl mb-6 mt-8"),w(Se,"class","docstring border-l-2 border-t-2 pl-4 pt-3.5 border-gray-100 rounded-tl-xl mb-6 mt-8"),w(Eo,"class","docstring border-l-2 border-t-2 pl-4 pt-3.5 border-gray-100 rounded-tl-xl mb-6 mt-8"),w(Ce,"class","docstring border-l-2 border-t-2 pl-4 pt-3.5 border-gray-100 rounded-tl-xl mb-6 mt-8"),w(U,"class","docstring border-l-2 border-t-2 pl-4 pt-3.5 border-gray-100 rounded-tl-xl mb-6 mt-8"),w(Xo,"class","docstring border-l-2 border-t-2 pl-4 pt-3.5 border-gray-100 rounded-tl-xl mb-6 mt-8"),w(qo,"class","docstring border-l-2 border-t-2 pl-4 pt-3.5 border-gray-100 rounded-tl-xl mb-6 mt-8"),w(Ae,"class","docstring border-l-2 border-t-2 pl-4 pt-3.5 border-gray-100 rounded-tl-xl mb-6 mt-8"),w(Bo,"class","docstring border-l-2 border-t-2 pl-4 pt-3.5 border-gray-100 rounded-tl-xl mb-6 mt-8"),w(ke,"class","docstring border-l-2 border-t-2 pl-4 pt-3.5 border-gray-100 rounded-tl-xl mb-6 mt-8"),w(E,"class","docstring border-l-2 border-t-2 pl-4 pt-3.5 border-gray-100 rounded-tl-xl mb-6 mt-8"),w(Go,"class","docstring border-l-2 border-t-2 pl-4 pt-3.5 border-gray-100 rounded-tl-xl mb-6 mt-8"),w(Jo,"class","docstring border-l-2 border-t-2 pl-4 pt-3.5 border-gray-100 rounded-tl-xl mb-6 mt-8"),w(Pe,"class","docstring border-l-2 border-t-2 pl-4 pt-3.5 border-gray-100 rounded-tl-xl mb-6 mt-8"),w(Oo,"class","docstring border-l-2 border-t-2 pl-4 pt-3.5 border-gray-100 rounded-tl-xl mb-6 mt-8"),w(Re,"class","docstring border-l-2 border-t-2 pl-4 pt-3.5 border-gray-100 rounded-tl-xl mb-6 mt-8"),w(N,"class","docstring border-l-2 border-t-2 pl-4 pt-3.5 border-gray-100 rounded-tl-xl mb-6 mt-8"),w(Qo,"class","docstring border-l-2 border-t-2 pl-4 pt-3.5 border-gray-100 rounded-tl-xl mb-6 mt-8"),w(Ko,"class","docstring border-l-2 border-t-2 pl-4 pt-3.5 border-gray-100 rounded-tl-xl mb-6 mt-8"),w(He,"class","docstring border-l-2 border-t-2 pl-4 pt-3.5 border-gray-100 rounded-tl-xl mb-6 mt-8"),w(or,"class","docstring border-l-2 border-t-2 pl-4 pt-3.5 border-gray-100 rounded-tl-xl mb-6 mt-8"),w(Ie,"class","docstring border-l-2 border-t-2 pl-4 pt-3.5 border-gray-100 rounded-tl-xl mb-6 mt-8"),w(X,"class","docstring border-l-2 border-t-2 pl-4 pt-3.5 border-gray-100 rounded-tl-xl mb-6 mt-8"),w(tr,"class","docstring border-l-2 border-t-2 pl-4 pt-3.5 border-gray-100 rounded-tl-xl mb-6 mt-8"),w(ar,"class","docstring border-l-2 border-t-2 pl-4 pt-3.5 border-gray-100 rounded-tl-xl mb-6 mt-8"),w(Ve,"class","docstring border-l-2 border-t-2 pl-4 pt-3.5 border-gray-100 rounded-tl-xl mb-6 mt-8"),w(sr,"class","docstring border-l-2 border-t-2 pl-4 pt-3.5 border-gray-100 rounded-tl-xl mb-6 mt-8"),w(We,"class","docstring border-l-2 border-t-2 pl-4 pt-3.5 border-gray-100 rounded-tl-xl mb-6 mt-8"),w(q,"class","docstring border-l-2 border-t-2 pl-4 pt-3.5 border-gray-100 rounded-tl-xl mb-6 mt-8"),w(dr,"class","docstring border-l-2 border-t-2 pl-4 pt-3.5 border-gray-100 rounded-tl-xl mb-6 mt-8"),w(lr,"class","docstring border-l-2 border-t-2 pl-4 pt-3.5 border-gray-100 rounded-tl-xl mb-6 mt-8"),w(Fe,"class","docstring border-l-2 border-t-2 pl-4 pt-3.5 border-gray-100 rounded-tl-xl mb-6 mt-8"),w(fr,"class","docstring border-l-2 border-t-2 pl-4 pt-3.5 border-gray-100 rounded-tl-xl mb-6 mt-8"),w(Ue,"class","docstring border-l-2 border-t-2 pl-4 pt-3.5 border-gray-100 rounded-tl-xl mb-6 mt-8"),w(z,"class","docstring border-l-2 border-t-2 pl-4 pt-3.5 border-gray-100 rounded-tl-xl mb-6 mt-8"),w(mr,"class","docstring border-l-2 border-t-2 pl-4 pt-3.5 border-gray-100 rounded-tl-xl mb-6 mt-8"),w(_r,"class","docstring border-l-2 border-t-2 pl-4 pt-3.5 border-gray-100 rounded-tl-xl mb-6 mt-8"),w(Ee,"class","docstring border-l-2 border-t-2 pl-4 pt-3.5 border-gray-100 rounded-tl-xl mb-6 mt-8"),w(hr,"class","docstring border-l-2 border-t-2 pl-4 pt-3.5 border-gray-100 rounded-tl-xl mb-6 mt-8"),w(Ne,"class","docstring border-l-2 border-t-2 pl-4 pt-3.5 border-gray-100 rounded-tl-xl mb-6 mt-8"),w(B,"class","docstring border-l-2 border-t-2 pl-4 pt-3.5 border-gray-100 rounded-tl-xl mb-6 mt-8"),w(Lr,"class","docstring border-l-2 border-t-2 pl-4 pt-3.5 border-gray-100 rounded-tl-xl mb-6 mt-8"),w(Oe,"class","docstring border-l-2 border-t-2 pl-4 pt-3.5 border-gray-100 rounded-tl-xl mb-6 mt-8"),w(xr,"class","docstring border-l-2 border-t-2 pl-4 pt-3.5 border-gray-100 rounded-tl-xl mb-6 mt-8"),w(br,"class","docstring border-l-2 border-t-2 pl-4 pt-3.5 border-gray-100 rounded-tl-xl mb-6 mt-8"),w(Xe,"class","docstring border-l-2 border-t-2 pl-4 pt-3.5 border-gray-100 rounded-tl-xl mb-6 mt-8"),w(wr,"class","docstring border-l-2 border-t-2 pl-4 pt-3.5 border-gray-100 rounded-tl-xl mb-6 mt-8"),w(qe,"class","docstring border-l-2 border-t-2 pl-4 pt-3.5 border-gray-100 rounded-tl-xl mb-6 mt-8"),w(j,"class","docstring border-l-2 border-t-2 pl-4 pt-3.5 border-gray-100 rounded-tl-xl mb-6 mt-8"),w(pi,"class","docstring border-l-2 border-t-2 pl-4 pt-3.5 border-gray-100 rounded-tl-xl mb-6 mt-8"),w(ze,"class","docstring border-l-2 border-t-2 pl-4 pt-3.5 border-gray-100 rounded-tl-xl mb-6 mt-8"),w(he,"class","docstring border-l-2 border-t-2 pl-4 pt-3.5 border-gray-100 rounded-tl-xl mb-6 mt-8"),w(Be,"class","docstring border-l-2 border-t-2 pl-4 pt-3.5 border-gray-100 rounded-tl-xl mb-6 mt-8"),w(Dr,"class","docstring border-l-2 border-t-2 pl-4 pt-3.5 border-gray-100 rounded-tl-xl mb-6 mt-8"),w(Sr,"class","docstring border-l-2 border-t-2 pl-4 pt-3.5 border-gray-100 rounded-tl-xl mb-6 mt-8"),w(je,"class","docstring border-l-2 border-t-2 pl-4 pt-3.5 border-gray-100 rounded-tl-xl mb-6 mt-8"),w(Ge,"class","docstring border-l-2 border-t-2 pl-4 pt-3.5 border-gray-100 rounded-tl-xl mb-6 mt-8"),w(A,"class","docstring border-l-2 border-t-2 pl-4 pt-3.5 border-gray-100 rounded-tl-xl mb-6 mt-8"),w(kr,"class","docstring border-l-2 border-t-2 pl-4 pt-3.5 border-gray-100 rounded-tl-xl mb-6 mt-8"),w(Pr,"class","docstring border-l-2 border-t-2 pl-4 pt-3.5 border-gray-100 rounded-tl-xl mb-6 mt-8"),w(Je,"class","docstring border-l-2 border-t-2 pl-4 pt-3.5 border-gray-100 rounded-tl-xl mb-6 mt-8"),w(Hr,"class","docstring border-l-2 border-t-2 pl-4 pt-3.5 border-gray-100 rounded-tl-xl mb-6 mt-8"),w(Ze,"class","docstring border-l-2 border-t-2 pl-4 pt-3.5 border-gray-100 rounded-tl-xl mb-6 mt-8"),w(G,"class","docstring border-l-2 border-t-2 pl-4 pt-3.5 border-gray-100 rounded-tl-xl mb-6 mt-8")},m(e,L){o(document.head,a),x(e,b,L),x(e,l,L),x(e,c,L),_(T,e,L),x(e,d,L),x(e,M,L),x(e,Qi,L),x(e,Wr,L),x(e,Ki,L),_(so,e,L),x(e,ed,L),_(Fr,e,L),x(e,od,L),x(e,I,L),_(Ur,I,null),o(I,Ll),o(I,un),o(I,xl),o(I,io),_(Er,io,null),o(io,bl),o(io,hn),o(I,vl),o(I,lo),_(Nr,lo,null),o(lo,wl),o(lo,gn),o(I,$l),o(I,K),_(Xr,K,null),o(K,Ml),o(K,Ln),o(K,yl),o(K,xn),o(K,Tl),o(K,bn),o(K,Dl),o(K,vn),o(K,Sl),o(K,wn),o(I,Cl),o(I,ge),_(qr,ge,null),o(ge,Al),o(ge,$n),o(ge,kl),_(co,ge,null),o(I,Pl),o(I,fo),_(zr,fo,null),o(fo,Rl),o(fo,Mn),x(e,rd,L),_(Br,e,L),x(e,td,L),x(e,V,L),_(jr,V,null),o(V,Hl),o(V,yn),o(V,Il),o(V,po),_(Gr,po,null),o(po,Vl),o(po,Tn),o(V,Wl),o(V,mo),_(Jr,mo,null),o(mo,Fl),o(mo,Dn),o(V,Ul),o(V,ee),_(Zr,ee,null),o(ee,El),o(ee,Sn),o(ee,Nl),o(ee,Cn),o(ee,Xl),o(ee,An),o(ee,ql),o(ee,kn),o(ee,zl),o(ee,Pn),o(V,Bl),o(V,Le),_(Or,Le,null),o(Le,jl),o(Le,Rn),o(Le,Gl),_(_o,Le,null),o(V,Jl),o(V,uo),_(Yr,uo,null),o(uo,Zl),o(uo,Hn),x(e,ad,L),_(Qr,e,L),x(e,nd,L),x(e,k,L),_(Kr,k,null),o(k,Ol),o(k,In),o(k,Yl),o(k,Vn),o(k,Ql),o(k,ho),_(et,ho,null),o(ho,Kl),o(ho,Wn),o(k,ec),o(k,go),_(ot,go,null),o(go,oc),o(go,Fn),o(k,rc),o(k,te),_(rt,te,null),o(te,tc),o(te,Un),o(te,ac),o(te,En),o(te,nc),o(te,Nn),o(te,sc),o(te,Xn),o(k,ic),o(k,xe),_(tt,xe,null),o(xe,dc),o(xe,qn),o(xe,lc),_(Lo,xe,null),o(k,cc),o(k,xo),_(at,xo,null),o(xo,fc),o(xo,zn),o(k,pc),o(k,be),_(nt,be,null),o(be,mc),o(be,Bn),o(be,_c),_(bo,be,null),x(e,sd,L),_(st,e,L),x(e,id,L),x(e,C,L),_(it,C,null),o(C,uc),o(C,jn),o(C,hc),o(C,Gn),o(C,gc),o(C,vo),_(dt,vo,null),o(vo,Lc),o(vo,Jn),o(C,xc),o(C,wo),_(lt,wo,null),o(wo,bc),o(wo,Zn),o(C,vc),o(C,ae),_(ct,ae,null),o(ae,wc),o(ae,On),o(ae,$c),o(ae,Yn),o(ae,Mc),o(ae,Qn),o(ae,yc),o(ae,Kn),o(C,Tc),o(C,ve),_(ft,ve,null),o(ve,Dc),o(ve,es),o(ve,Sc),_($o,ve,null),o(C,Cc),o(C,Mo),_(pt,Mo,null),o(Mo,Ac),o(Mo,os),o(C,kc),o(C,we),_(mt,we,null),o(we,Pc),o(we,rs),o(we,Rc),_(yo,we,null),o(C,Hc),o(C,$e),_(_t,$e,null),o($e,Ic),o($e,ts),o($e,Vc),_(To,$e,null),x(e,dd,L),_(ut,e,L),x(e,ld,L),x(e,W,L),_(ht,W,null),o(W,Wc),o(W,as),o(W,Fc),o(W,Do),_(gt,Do,null),o(Do,Uc),o(Do,ns),o(W,Ec),o(W,So),_(Lt,So,null),o(So,Nc),o(So,ss),o(W,Xc),o(W,Me),_(xt,Me,null),o(Me,qc),o(Me,is),o(Me,zc),_(Co,Me,null),o(W,Bc),o(W,Ao),_(bt,Ao,null),o(Ao,jc),o(Ao,ds),o(W,Gc),o(W,ye),_(vt,ye,null),o(ye,Jc),o(ye,ls),o(ye,Zc),_(ko,ye,null),x(e,cd,L),_(wt,e,L),x(e,fd,L),x(e,F,L),_($t,F,null),o(F,Oc),o(F,cs),o(F,Yc),o(F,Po),_(Mt,Po,null),o(Po,Qc),o(Po,fs),o(F,Kc),o(F,Ro),_(yt,Ro,null),o(Ro,ef),o(Ro,ps),o(F,of),o(F,Te),_(Tt,Te,null),o(Te,rf),o(Te,ms),o(Te,tf),_(Ho,Te,null),o(F,af),o(F,Io),_(Dt,Io,null),o(Io,nf),o(Io,_s),o(F,sf),o(F,De),_(St,De,null),o(De,df),o(De,us),o(De,lf),_(Vo,De,null),x(e,pd,L),_(Ct,e,L),x(e,md,L),x(e,U,L),_(At,U,null),o(U,cf),o(U,hs),o(U,ff),o(U,Wo),_(kt,Wo,null),o(Wo,pf),o(Wo,gs),o(U,mf),o(U,Fo),_(Pt,Fo,null),o(Fo,_f),o(Fo,Ls),o(U,uf),o(U,Se),_(Rt,Se,null),o(Se,hf),o(Se,xs),o(Se,gf),_(Uo,Se,null),o(U,Lf),o(U,Eo),_(Ht,Eo,null),o(Eo,xf),o(Eo,bs),o(U,bf),o(U,Ce),_(It,Ce,null),o(Ce,vf),o(Ce,vs),o(Ce,wf),_(No,Ce,null),x(e,_d,L),_(Vt,e,L),x(e,ud,L),x(e,E,L),_(Wt,E,null),o(E,$f),o(E,ws),o(E,Mf),o(E,Xo),_(Ft,Xo,null),o(Xo,yf),o(Xo,$s),o(E,Tf),o(E,qo),_(Ut,qo,null),o(qo,Df),o(qo,Ms),o(E,Sf),o(E,Ae),_(Et,Ae,null),o(Ae,Cf),o(Ae,ys),o(Ae,Af),_(zo,Ae,null),o(E,kf),o(E,Bo),_(Nt,Bo,null),o(Bo,Pf),o(Bo,Ts),o(E,Rf),o(E,ke),_(Xt,ke,null),o(ke,Hf),o(ke,Ds),o(ke,If),_(jo,ke,null),x(e,hd,L),_(qt,e,L),x(e,gd,L),x(e,N,L),_(zt,N,null),o(N,Vf),o(N,Ss),o(N,Wf),o(N,Go),_(Bt,Go,null),o(Go,Ff),o(Go,Cs),o(N,Uf),o(N,Jo),_(jt,Jo,null),o(Jo,Ef),o(Jo,As),o(N,Nf),o(N,Pe),_(Gt,Pe,null),o(Pe,Xf),o(Pe,ks),o(Pe,qf),_(Zo,Pe,null),o(N,zf),o(N,Oo),_(Jt,Oo,null),o(Oo,Bf),o(Oo,Ps),o(N,jf),o(N,Re),_(Zt,Re,null),o(Re,Gf),o(Re,Rs),o(Re,Jf),_(Yo,Re,null),x(e,Ld,L),_(Ot,e,L),x(e,xd,L),x(e,X,L),_(Yt,X,null),o(X,Zf),o(X,Hs),o(X,Of),o(X,Qo),_(Qt,Qo,null),o(Qo,Yf),o(Qo,Is),o(X,Qf),o(X,Ko),_(Kt,Ko,null),o(Ko,Kf),o(Ko,Vs),o(X,ep),o(X,He),_(ea,He,null),o(He,op),o(He,Ws),o(He,rp),_(er,He,null),o(X,tp),o(X,or),_(oa,or,null),o(or,ap),o(or,Fs),o(X,np),o(X,Ie),_(ra,Ie,null),o(Ie,sp),o(Ie,Us),o(Ie,ip),_(rr,Ie,null),x(e,bd,L),_(ta,e,L),x(e,vd,L),x(e,q,L),_(aa,q,null),o(q,dp),o(q,Es),o(q,lp),o(q,tr),_(na,tr,null),o(tr,cp),o(tr,Ns),o(q,fp),o(q,ar),_(sa,ar,null),o(ar,pp),o(ar,Xs),o(q,mp),o(q,Ve),_(ia,Ve,null),o(Ve,_p),o(Ve,qs),o(Ve,up),_(nr,Ve,null),o(q,hp),o(q,sr),_(da,sr,null),o(sr,gp),o(sr,zs),o(q,Lp),o(q,We),_(la,We,null),o(We,xp),o(We,Bs),o(We,bp),_(ir,We,null),x(e,wd,L),_(ca,e,L),x(e,$d,L),x(e,z,L),_(fa,z,null),o(z,vp),o(z,js),o(z,wp),o(z,dr),_(pa,dr,null),o(dr,$p),o(dr,Gs),o(z,Mp),o(z,lr),_(ma,lr,null),o(lr,yp),o(lr,Js),o(z,Tp),o(z,Fe),_(_a,Fe,null),o(Fe,Dp),o(Fe,Zs),o(Fe,Sp),_(cr,Fe,null),o(z,Cp),o(z,fr),_(ua,fr,null),o(fr,Ap),o(fr,Os),o(z,kp),o(z,Ue),_(ha,Ue,null),o(Ue,Pp),o(Ue,Ys),o(Ue,Rp),_(pr,Ue,null),x(e,Md,L),_(ga,e,L),x(e,yd,L),x(e,B,L),_(La,B,null),o(B,Hp),o(B,Qs),o(B,Ip),o(B,mr),_(xa,mr,null),o(mr,Vp),o(mr,Ks),o(B,Wp),o(B,_r),_(ba,_r,null),o(_r,Fp),o(_r,ei),o(B,Up),o(B,Ee),_(va,Ee,null),o(Ee,Ep),o(Ee,oi),o(Ee,Np),_(ur,Ee,null),o(B,Xp),o(B,hr),_(wa,hr,null),o(hr,qp),o(hr,ri),o(B,zp),o(B,Ne),_($a,Ne,null),o(Ne,Bp),o(Ne,ti),o(Ne,jp),_(gr,Ne,null),x(e,Td,L),_(Ma,e,L),x(e,Dd,L),x(e,Oe,L),_(ya,Oe,null),o(Oe,Gp),o(Oe,Lr),_(Ta,Lr,null),o(Lr,Jp),o(Lr,ai),x(e,Sd,L),_(Da,e,L),x(e,Cd,L),x(e,j,L),_(Sa,j,null),o(j,Zp),o(j,ni),o(j,Op),o(j,xr),_(Ca,xr,null),o(xr,Yp),o(xr,si),o(j,Qp),o(j,br),_(Aa,br,null),o(br,Kp),o(br,ii),o(j,em),o(j,Xe),_(ka,Xe,null),o(Xe,om),o(Xe,di),o(Xe,rm),_(vr,Xe,null),o(j,tm),o(j,wr),_(Pa,wr,null),o(wr,am),o(wr,li),o(j,nm),o(j,qe),_(Ra,qe,null),o(qe,sm),o(qe,ci),o(qe,im),_($r,qe,null),x(e,Ad,L),_(Ha,e,L),x(e,kd,L),x(e,A,L),_(Ia,A,null),o(A,dm),o(A,fi),o(A,lm),o(A,pi),_(Va,pi,null),o(A,cm),o(A,ze),_(Wa,ze,null),o(ze,fm),o(ze,mi),o(ze,pm),o(ze,_i),o(A,mm),o(A,he),_(Fa,he,null),o(he,_m),o(he,ui),o(he,um),_(Mr,he,null),o(he,hm),_(yr,he,null),o(A,gm),o(A,Be),_(Ua,Be,null),o(Be,Lm),o(Be,hi),o(Be,xm),_(Tr,Be,null),o(A,bm),o(A,Dr),_(Ea,Dr,null),o(Dr,vm),o(Dr,gi),o(A,wm),o(A,Sr),_(Na,Sr,null),o(Sr,$m),o(Sr,Li),o(A,Mm),o(A,je),_(Xa,je,null),o(je,ym),o(je,xi),o(je,Tm),_(Cr,je,null),o(A,Dm),o(A,Ge),_(qa,Ge,null),o(Ge,Sm),o(Ge,bi),o(Ge,Cm),_(Ar,Ge,null),x(e,Pd,L),_(za,e,L),x(e,Rd,L),x(e,G,L),_(Ba,G,null),o(G,Am),o(G,vi),o(G,km),o(G,kr),_(ja,kr,null),o(kr,Pm),o(kr,wi),o(G,Rm),o(G,Pr),_(Ga,Pr,null),o(Pr,Hm),o(Pr,$i),o(G,Im),o(G,Je),_(Ja,Je,null),o(Je,Vm),o(Je,Mi),o(Je,Wm),_(Rr,Je,null),o(G,Fm),o(G,Hr),_(Za,Hr,null),o(Hr,Um),o(Hr,yi),o(G,Em),o(G,Ze),_(Oa,Ze,null),o(Ze,Nm),o(Ze,Ti),o(Ze,Xm),_(Ir,Ze,null),x(e,Hd,L),_(Ya,e,L),x(e,Id,L),x(e,Yi,L),Vd=!0},p(e,[L]){const Z={};L&2&&(Z.$$scope={dirty:L,ctx:e}),so.$set(Z);const Qa={};L&2&&(Qa.$$scope={dirty:L,ctx:e}),co.$set(Qa);const Ka={};L&2&&(Ka.$$scope={dirty:L,ctx:e}),_o.$set(Ka);const oe={};L&2&&(oe.$$scope={dirty:L,ctx:e}),Lo.$set(oe);const Ye={};L&2&&(Ye.$$scope={dirty:L,ctx:e}),bo.$set(Ye);const en={};L&2&&(en.$$scope={dirty:L,ctx:e}),$o.$set(en);const O={};L&2&&(O.$$scope={dirty:L,ctx:e}),yo.$set(O);const on={};L&2&&(on.$$scope={dirty:L,ctx:e}),To.$set(on);const rn={};L&2&&(rn.$$scope={dirty:L,ctx:e}),Co.$set(rn);const re={};L&2&&(re.$$scope={dirty:L,ctx:e}),ko.$set(re);const Qe={};L&2&&(Qe.$$scope={dirty:L,ctx:e}),Ho.$set(Qe);const tn={};L&2&&(tn.$$scope={dirty:L,ctx:e}),Vo.$set(tn);const R={};L&2&&(R.$$scope={dirty:L,ctx:e}),Uo.$set(R);const an={};L&2&&(an.$$scope={dirty:L,ctx:e}),No.$set(an);const nn={};L&2&&(nn.$$scope={dirty:L,ctx:e}),zo.$set(nn);const _e={};L&2&&(_e.$$scope={dirty:L,ctx:e}),jo.$set(_e);const Ke={};L&2&&(Ke.$$scope={dirty:L,ctx:e}),Zo.$set(Ke);const sn={};L&2&&(sn.$$scope={dirty:L,ctx:e}),Yo.$set(sn);const eo={};L&2&&(eo.$$scope={dirty:L,ctx:e}),er.$set(eo);const P={};L&2&&(P.$$scope={dirty:L,ctx:e}),rr.$set(P);const dn={};L&2&&(dn.$$scope={dirty:L,ctx:e}),nr.$set(dn);const ln={};L&2&&(ln.$$scope={dirty:L,ctx:e}),ir.$set(ln);const ue={};L&2&&(ue.$$scope={dirty:L,ctx:e}),cr.$set(ue);const oo={};L&2&&(oo.$$scope={dirty:L,ctx:e}),pr.$set(oo);const cn={};L&2&&(cn.$$scope={dirty:L,ctx:e}),ur.$set(cn);const ro={};L&2&&(ro.$$scope={dirty:L,ctx:e}),gr.$set(ro);const to={};L&2&&(to.$$scope={dirty:L,ctx:e}),vr.$set(to);const Y={};L&2&&(Y.$$scope={dirty:L,ctx:e}),$r.$set(Y);const fn={};L&2&&(fn.$$scope={dirty:L,ctx:e}),Mr.$set(fn);const pn={};L&2&&(pn.$$scope={dirty:L,ctx:e}),yr.$set(pn);const ao={};L&2&&(ao.$$scope={dirty:L,ctx:e}),Tr.$set(ao);const mn={};L&2&&(mn.$$scope={dirty:L,ctx:e}),Cr.$set(mn);const no={};L&2&&(no.$$scope={dirty:L,ctx:e}),Ar.$set(no);const Q={};L&2&&(Q.$$scope={dirty:L,ctx:e}),Rr.$set(Q);const _n={};L&2&&(_n.$$scope={dirty:L,ctx:e}),Ir.$set(_n)},i(e){Vd||(u(T.$$.fragment,e),u(so.$$.fragment,e),u(Fr.$$.fragment,e),u(Ur.$$.fragment,e),u(Er.$$.fragment,e),u(Nr.$$.fragment,e),u(Xr.$$.fragment,e),u(qr.$$.fragment,e),u(co.$$.fragment,e),u(zr.$$.fragment,e),u(Br.$$.fragment,e),u(jr.$$.fragment,e),u(Gr.$$.fragment,e),u(Jr.$$.fragment,e),u(Zr.$$.fragment,e),u(Or.$$.fragment,e),u(_o.$$.fragment,e),u(Yr.$$.fragment,e),u(Qr.$$.fragment,e),u(Kr.$$.fragment,e),u(et.$$.fragment,e),u(ot.$$.fragment,e),u(rt.$$.fragment,e),u(tt.$$.fragment,e),u(Lo.$$.fragment,e),u(at.$$.fragment,e),u(nt.$$.fragment,e),u(bo.$$.fragment,e),u(st.$$.fragment,e),u(it.$$.fragment,e),u(dt.$$.fragment,e),u(lt.$$.fragment,e),u(ct.$$.fragment,e),u(ft.$$.fragment,e),u($o.$$.fragment,e),u(pt.$$.fragment,e),u(mt.$$.fragment,e),u(yo.$$.fragment,e),u(_t.$$.fragment,e),u(To.$$.fragment,e),u(ut.$$.fragment,e),u(ht.$$.fragment,e),u(gt.$$.fragment,e),u(Lt.$$.fragment,e),u(xt.$$.fragment,e),u(Co.$$.fragment,e),u(bt.$$.fragment,e),u(vt.$$.fragment,e),u(ko.$$.fragment,e),u(wt.$$.fragment,e),u($t.$$.fragment,e),u(Mt.$$.fragment,e),u(yt.$$.fragment,e),u(Tt.$$.fragment,e),u(Ho.$$.fragment,e),u(Dt.$$.fragment,e),u(St.$$.fragment,e),u(Vo.$$.fragment,e),u(Ct.$$.fragment,e),u(At.$$.fragment,e),u(kt.$$.fragment,e),u(Pt.$$.fragment,e),u(Rt.$$.fragment,e),u(Uo.$$.fragment,e),u(Ht.$$.fragment,e),u(It.$$.fragment,e),u(No.$$.fragment,e),u(Vt.$$.fragment,e),u(Wt.$$.fragment,e),u(Ft.$$.fragment,e),u(Ut.$$.fragment,e),u(Et.$$.fragment,e),u(zo.$$.fragment,e),u(Nt.$$.fragment,e),u(Xt.$$.fragment,e),u(jo.$$.fragment,e),u(qt.$$.fragment,e),u(zt.$$.fragment,e),u(Bt.$$.fragment,e),u(jt.$$.fragment,e),u(Gt.$$.fragment,e),u(Zo.$$.fragment,e),u(Jt.$$.fragment,e),u(Zt.$$.fragment,e),u(Yo.$$.fragment,e),u(Ot.$$.fragment,e),u(Yt.$$.fragment,e),u(Qt.$$.fragment,e),u(Kt.$$.fragment,e),u(ea.$$.fragment,e),u(er.$$.fragment,e),u(oa.$$.fragment,e),u(ra.$$.fragment,e),u(rr.$$.fragment,e),u(ta.$$.fragment,e),u(aa.$$.fragment,e),u(na.$$.fragment,e),u(sa.$$.fragment,e),u(ia.$$.fragment,e),u(nr.$$.fragment,e),u(da.$$.fragment,e),u(la.$$.fragment,e),u(ir.$$.fragment,e),u(ca.$$.fragment,e),u(fa.$$.fragment,e),u(pa.$$.fragment,e),u(ma.$$.fragment,e),u(_a.$$.fragment,e),u(cr.$$.fragment,e),u(ua.$$.fragment,e),u(ha.$$.fragment,e),u(pr.$$.fragment,e),u(ga.$$.fragment,e),u(La.$$.fragment,e),u(xa.$$.fragment,e),u(ba.$$.fragment,e),u(va.$$.fragment,e),u(ur.$$.fragment,e),u(wa.$$.fragment,e),u($a.$$.fragment,e),u(gr.$$.fragment,e),u(Ma.$$.fragment,e),u(ya.$$.fragment,e),u(Ta.$$.fragment,e),u(Da.$$.fragment,e),u(Sa.$$.fragment,e),u(Ca.$$.fragment,e),u(Aa.$$.fragment,e),u(ka.$$.fragment,e),u(vr.$$.fragment,e),u(Pa.$$.fragment,e),u(Ra.$$.fragment,e),u($r.$$.fragment,e),u(Ha.$$.fragment,e),u(Ia.$$.fragment,e),u(Va.$$.fragment,e),u(Wa.$$.fragment,e),u(Fa.$$.fragment,e),u(Mr.$$.fragment,e),u(yr.$$.fragment,e),u(Ua.$$.fragment,e),u(Tr.$$.fragment,e),u(Ea.$$.fragment,e),u(Na.$$.fragment,e),u(Xa.$$.fragment,e),u(Cr.$$.fragment,e),u(qa.$$.fragment,e),u(Ar.$$.fragment,e),u(za.$$.fragment,e),u(Ba.$$.fragment,e),u(ja.$$.fragment,e),u(Ga.$$.fragment,e),u(Ja.$$.fragment,e),u(Rr.$$.fragment,e),u(Za.$$.fragment,e),u(Oa.$$.fragment,e),u(Ir.$$.fragment,e),u(Ya.$$.fragment,e),Vd=!0)},o(e){h(T.$$.fragment,e),h(so.$$.fragment,e),h(Fr.$$.fragment,e),h(Ur.$$.fragment,e),h(Er.$$.fragment,e),h(Nr.$$.fragment,e),h(Xr.$$.fragment,e),h(qr.$$.fragment,e),h(co.$$.fragment,e),h(zr.$$.fragment,e),h(Br.$$.fragment,e),h(jr.$$.fragment,e),h(Gr.$$.fragment,e),h(Jr.$$.fragment,e),h(Zr.$$.fragment,e),h(Or.$$.fragment,e),h(_o.$$.fragment,e),h(Yr.$$.fragment,e),h(Qr.$$.fragment,e),h(Kr.$$.fragment,e),h(et.$$.fragment,e),h(ot.$$.fragment,e),h(rt.$$.fragment,e),h(tt.$$.fragment,e),h(Lo.$$.fragment,e),h(at.$$.fragment,e),h(nt.$$.fragment,e),h(bo.$$.fragment,e),h(st.$$.fragment,e),h(it.$$.fragment,e),h(dt.$$.fragment,e),h(lt.$$.fragment,e),h(ct.$$.fragment,e),h(ft.$$.fragment,e),h($o.$$.fragment,e),h(pt.$$.fragment,e),h(mt.$$.fragment,e),h(yo.$$.fragment,e),h(_t.$$.fragment,e),h(To.$$.fragment,e),h(ut.$$.fragment,e),h(ht.$$.fragment,e),h(gt.$$.fragment,e),h(Lt.$$.fragment,e),h(xt.$$.fragment,e),h(Co.$$.fragment,e),h(bt.$$.fragment,e),h(vt.$$.fragment,e),h(ko.$$.fragment,e),h(wt.$$.fragment,e),h($t.$$.fragment,e),h(Mt.$$.fragment,e),h(yt.$$.fragment,e),h(Tt.$$.fragment,e),h(Ho.$$.fragment,e),h(Dt.$$.fragment,e),h(St.$$.fragment,e),h(Vo.$$.fragment,e),h(Ct.$$.fragment,e),h(At.$$.fragment,e),h(kt.$$.fragment,e),h(Pt.$$.fragment,e),h(Rt.$$.fragment,e),h(Uo.$$.fragment,e),h(Ht.$$.fragment,e),h(It.$$.fragment,e),h(No.$$.fragment,e),h(Vt.$$.fragment,e),h(Wt.$$.fragment,e),h(Ft.$$.fragment,e),h(Ut.$$.fragment,e),h(Et.$$.fragment,e),h(zo.$$.fragment,e),h(Nt.$$.fragment,e),h(Xt.$$.fragment,e),h(jo.$$.fragment,e),h(qt.$$.fragment,e),h(zt.$$.fragment,e),h(Bt.$$.fragment,e),h(jt.$$.fragment,e),h(Gt.$$.fragment,e),h(Zo.$$.fragment,e),h(Jt.$$.fragment,e),h(Zt.$$.fragment,e),h(Yo.$$.fragment,e),h(Ot.$$.fragment,e),h(Yt.$$.fragment,e),h(Qt.$$.fragment,e),h(Kt.$$.fragment,e),h(ea.$$.fragment,e),h(er.$$.fragment,e),h(oa.$$.fragment,e),h(ra.$$.fragment,e),h(rr.$$.fragment,e),h(ta.$$.fragment,e),h(aa.$$.fragment,e),h(na.$$.fragment,e),h(sa.$$.fragment,e),h(ia.$$.fragment,e),h(nr.$$.fragment,e),h(da.$$.fragment,e),h(la.$$.fragment,e),h(ir.$$.fragment,e),h(ca.$$.fragment,e),h(fa.$$.fragment,e),h(pa.$$.fragment,e),h(ma.$$.fragment,e),h(_a.$$.fragment,e),h(cr.$$.fragment,e),h(ua.$$.fragment,e),h(ha.$$.fragment,e),h(pr.$$.fragment,e),h(ga.$$.fragment,e),h(La.$$.fragment,e),h(xa.$$.fragment,e),h(ba.$$.fragment,e),h(va.$$.fragment,e),h(ur.$$.fragment,e),h(wa.$$.fragment,e),h($a.$$.fragment,e),h(gr.$$.fragment,e),h(Ma.$$.fragment,e),h(ya.$$.fragment,e),h(Ta.$$.fragment,e),h(Da.$$.fragment,e),h(Sa.$$.fragment,e),h(Ca.$$.fragment,e),h(Aa.$$.fragment,e),h(ka.$$.fragment,e),h(vr.$$.fragment,e),h(Pa.$$.fragment,e),h(Ra.$$.fragment,e),h($r.$$.fragment,e),h(Ha.$$.fragment,e),h(Ia.$$.fragment,e),h(Va.$$.fragment,e),h(Wa.$$.fragment,e),h(Fa.$$.fragment,e),h(Mr.$$.fragment,e),h(yr.$$.fragment,e),h(Ua.$$.fragment,e),h(Tr.$$.fragment,e),h(Ea.$$.fragment,e),h(Na.$$.fragment,e),h(Xa.$$.fragment,e),h(Cr.$$.fragment,e),h(qa.$$.fragment,e),h(Ar.$$.fragment,e),h(za.$$.fragment,e),h(Ba.$$.fragment,e),h(ja.$$.fragment,e),h(Ga.$$.fragment,e),h(Ja.$$.fragment,e),h(Rr.$$.fragment,e),h(Za.$$.fragment,e),h(Oa.$$.fragment,e),h(Ir.$$.fragment,e),h(Ya.$$.fragment,e),Vd=!1},d(e){e&&(i(b),i(l),i(c),i(d),i(M),i(Qi),i(Wr),i(Ki),i(ed),i(od),i(I),i(rd),i(td),i(V),i(ad),i(nd),i(k),i(sd),i(id),i(C),i(dd),i(ld),i(W),i(cd),i(fd),i(F),i(pd),i(md),i(U),i(_d),i(ud),i(E),i(hd),i(gd),i(N),i(Ld),i(xd),i(X),i(bd),i(vd),i(q),i(wd),i($d),i(z),i(Md),i(yd),i(B),i(Td),i(Dd),i(Oe),i(Sd),i(Cd),i(j),i(Ad),i(kd),i(A),i(Pd),i(Rd),i(G),i(Hd),i(Id),i(Yi)),i(a),g(T,e),g(so,e),g(Fr,e),g(Ur),g(Er),g(Nr),g(Xr),g(qr),g(co),g(zr),g(Br,e),g(jr),g(Gr),g(Jr),g(Zr),g(Or),g(_o),g(Yr),g(Qr,e),g(Kr),g(et),g(ot),g(rt),g(tt),g(Lo),g(at),g(nt),g(bo),g(st,e),g(it),g(dt),g(lt),g(ct),g(ft),g($o),g(pt),g(mt),g(yo),g(_t),g(To),g(ut,e),g(ht),g(gt),g(Lt),g(xt),g(Co),g(bt),g(vt),g(ko),g(wt,e),g($t),g(Mt),g(yt),g(Tt),g(Ho),g(Dt),g(St),g(Vo),g(Ct,e),g(At),g(kt),g(Pt),g(Rt),g(Uo),g(Ht),g(It),g(No),g(Vt,e),g(Wt),g(Ft),g(Ut),g(Et),g(zo),g(Nt),g(Xt),g(jo),g(qt,e),g(zt),g(Bt),g(jt),g(Gt),g(Zo),g(Jt),g(Zt),g(Yo),g(Ot,e),g(Yt),g(Qt),g(Kt),g(ea),g(er),g(oa),g(ra),g(rr),g(ta,e),g(aa),g(na),g(sa),g(ia),g(nr),g(da),g(la),g(ir),g(ca,e),g(fa),g(pa),g(ma),g(_a),g(cr),g(ua),g(ha),g(pr),g(ga,e),g(La),g(xa),g(ba),g(va),g(ur),g(wa),g($a),g(gr),g(Ma,e),g(ya),g(Ta),g(Da,e),g(Sa),g(Ca),g(Aa),g(ka),g(vr),g(Pa),g(Ra),g($r),g(Ha,e),g(Ia),g(Va),g(Wa),g(Fa),g(Mr),g(yr),g(Ua),g(Tr),g(Ea),g(Na),g(Xa),g(Cr),g(qa),g(Ar),g(za,e),g(Ba),g(ja),g(Ga),g(Ja),g(Rr),g(Za),g(Oa),g(Ir),g(Ya,e)}}}const Bh='{"title":"LoRA","local":"lora","sections":[{"title":"StableDiffusionLoraLoaderMixin","local":"diffusers.loaders.StableDiffusionLoraLoaderMixin","sections":[],"depth":2},{"title":"StableDiffusionXLLoraLoaderMixin","local":"diffusers.loaders.StableDiffusionXLLoraLoaderMixin","sections":[],"depth":2},{"title":"SD3LoraLoaderMixin","local":"diffusers.loaders.SD3LoraLoaderMixin","sections":[],"depth":2},{"title":"FluxLoraLoaderMixin","local":"diffusers.loaders.FluxLoraLoaderMixin","sections":[],"depth":2},{"title":"CogVideoXLoraLoaderMixin","local":"diffusers.loaders.CogVideoXLoraLoaderMixin","sections":[],"depth":2},{"title":"Mochi1LoraLoaderMixin","local":"diffusers.loaders.Mochi1LoraLoaderMixin","sections":[],"depth":2},{"title":"AuraFlowLoraLoaderMixin","local":"diffusers.loaders.AuraFlowLoraLoaderMixin","sections":[],"depth":2},{"title":"LTXVideoLoraLoaderMixin","local":"diffusers.loaders.LTXVideoLoraLoaderMixin","sections":[],"depth":2},{"title":"SanaLoraLoaderMixin","local":"diffusers.loaders.SanaLoraLoaderMixin","sections":[],"depth":2},{"title":"HunyuanVideoLoraLoaderMixin","local":"diffusers.loaders.HunyuanVideoLoraLoaderMixin","sections":[],"depth":2},{"title":"Lumina2LoraLoaderMixin","local":"diffusers.loaders.Lumina2LoraLoaderMixin","sections":[],"depth":2},{"title":"CogView4LoraLoaderMixin","local":"diffusers.loaders.CogView4LoraLoaderMixin","sections":[],"depth":2},{"title":"WanLoraLoaderMixin","local":"diffusers.loaders.WanLoraLoaderMixin","sections":[],"depth":2},{"title":"AmusedLoraLoaderMixin","local":"diffusers.loaders.AmusedLoraLoaderMixin","sections":[],"depth":2},{"title":"HiDreamImageLoraLoaderMixin","local":"diffusers.loaders.HiDreamImageLoraLoaderMixin","sections":[],"depth":2},{"title":"LoraBaseMixin","local":"diffusers.loaders.lora_base.LoraBaseMixin","sections":[],"depth":2},{"title":"WanLoraLoaderMixin","local":"diffusers.loaders.WanLoraLoaderMixin","sections":[],"depth":2}],"depth":1}';function jh(y){return ah(()=>{new URLSearchParams(window.location.search).get("fw")}),[]}class eg extends nh{constructor(a){super(),sh(this,a,jh,zh,th,{})}}export{eg as component}; | |
Xet Storage Details
- Size:
- 313 kB
- Xet hash:
- 213d35293c3802d2a68bf751c02750ddc0acbd1128ff91bfc3a5f8782d341427
·
Xet efficiently stores files, intelligently splitting them into unique chunks and accelerating uploads and downloads. More info.