Buckets:

rtrm's picture
download
raw
17.1 kB
import{s as Qe,n as Re,o as Fe}from"../chunks/scheduler.94020406.js";import{S as Ie,i as Be,g as a,s as i,r as m,E as Ne,h as o,f as l,c as n,j as xe,u as d,x as p,k as Ze,y as Se,a as s,v as c,d as r,t as f,w as u}from"../chunks/index.a08c8d92.js";import{C as Y}from"../chunks/CodeBlock.b23cf525.js";import{D as Ye}from"../chunks/DocNotebookDropdown.d8a25975.js";import{H as Me,E as Ee}from"../chunks/EditOnGithub.b1bceb47.js";function ze(Je){let b,E,N,z,h,D,M,H,Z,ye='<a href="https://huggingface.co/papers/2311.15127" rel="nofollow">Stable Video Diffusion (SVD)</a>은 입력 이미지에 맞춰 2~4초 분량의 고해상도(576x1024) 비디오를 생성할 수 있는 강력한 image-to-video 생성 모델입니다.',L,J,we="이 가이드에서는 SVD를 사용하여 이미지에서 짧은 동영상을 생성하는 방법을 설명합니다.",q,y,Te="시작하기 전에 다음 라이브러리가 설치되어 있는지 확인하세요:",P,w,K,T,ve='이 모델에는 <a href="https://huggingface.co/stabilityai/stable-video-diffusion-img2vid" rel="nofollow">SVD</a>와 <a href="https://huggingface.co/stabilityai/stable-video-diffusion-img2vid-xt" rel="nofollow">SVD-XT</a> 두 가지 종류가 있습니다. SVD 체크포인트는 14개의 프레임을 생성하도록 학습되었고, SVD-XT 체크포인트는 25개의 프레임을 생성하도록 파인튜닝되었습니다.',A,v,Ve="이 가이드에서는 SVD-XT 체크포인트를 사용합니다.",O,V,ee,g,Ue='<div><img class="rounded-xl" src="https://huggingface.co/datasets/huggingface/documentation-images/resolve/main/diffusers/svd/rocket.png"/> <figcaption class="mt-2 text-center text-sm text-gray-500">&quot;source image of a rocket&quot;</figcaption></div> <div><img class="rounded-xl" src="https://huggingface.co/datasets/huggingface/documentation-images/resolve/main/diffusers/svd/output_rocket.gif"/> <figcaption class="mt-2 text-center text-sm text-gray-500">&quot;generated video from source image&quot;</figcaption></div>',te,U,le,C,Ce='UNet을 <a href="../optimization/torch2.0#torchcompile">컴파일</a>하면 메모리 사용량이 살짝 증가하지만, 20~25%의 속도 향상을 얻을 수 있습니다.',se,k,ie,_,ne,W,ke="비디오 생성은 기본적으로 배치 크기가 큰 text-to-image 생성과 유사하게 ‘num_frames’를 한 번에 생성하기 때문에 메모리 사용량이 매우 높습니다. 메모리 사용량을 줄이기 위해 추론 속도와 메모리 사용량을 절충하는 여러 가지 옵션이 있습니다:",ae,$,_e="<li>모델 오프로링 활성화: 파이프라인의 각 구성 요소가 더 이상 필요하지 않을 때 CPU로 오프로드됩니다.</li> <li>Feed-forward chunking 활성화: feed-forward 레이어가 배치 크기가 큰 단일 feed-forward를 실행하는 대신 루프로 반복해서 실행됩니다.</li> <li><code>decode_chunk_size</code> 감소: VAE가 프레임들을 한꺼번에 디코딩하는 대신 chunk 단위로 디코딩합니다. <code>decode_chunk_size=1</code>을 설정하면 한 번에 한 프레임씩 디코딩하고 최소한의 메모리만 사용하지만(GPU 메모리에 따라 이 값을 조정하는 것이 좋습니다), 동영상에 약간의 깜박임이 발생할 수 있습니다.</li>",oe,j,pe,X,We="이러한 모든 방법들을 사용하면 메모리 사용량이 8GAM VRAM보다 적을 것입니다.",me,G,de,x,$e="Stable Diffusion Video는 또한 이미지 conditoning 외에도 micro-conditioning을 허용하므로 생성된 비디오를 더 잘 제어할 수 있습니다:",ce,Q,je="<li><code>fps</code>: 생성된 비디오의 초당 프레임 수입니다.</li> <li><code>motion_bucket_id</code>: 생성된 동영상에 사용할 모션 버킷 아이디입니다. 생성된 동영상의 모션을 제어하는 데 사용할 수 있습니다. 모션 버킷 아이디를 늘리면 생성되는 동영상의 모션이 증가합니다.</li> <li><code>noise_aug_strength</code>: Conditioning 이미지에 추가되는 노이즈의 양입니다. 값이 클수록 비디오가 conditioning 이미지와 덜 유사해집니다. 이 값을 높이면 생성된 비디오의 움직임도 증가합니다.</li>",re,R,Xe="예를 들어, 모션이 더 많은 동영상을 생성하려면 <code>motion_bucket_id</code> 및 <code>noise_aug_strength</code> micro-conditioning 파라미터를 사용합니다:",fe,F,ue,I,Ge='<img src="https://huggingface.co/datasets/huggingface/documentation-images/resolve/main/diffusers/svd/output_rocket_with_conditions.gif"/>',be,B,ge,S,he;return h=new Me({props:{title:"Stable Video Diffusion",local:"stable-video-diffusion",headingTag:"h1"}}),M=new Ye({props:{classNames:"absolute z-10 right-0 top-0",options:[{label:"Mixed",value:"https://colab.research.google.com/github/huggingface/notebooks/blob/main/diffusers_doc/ko/svd.ipynb"},{label:"PyTorch",value:"https://colab.research.google.com/github/huggingface/notebooks/blob/main/diffusers_doc/ko/pytorch/svd.ipynb"},{label:"TensorFlow",value:"https://colab.research.google.com/github/huggingface/notebooks/blob/main/diffusers_doc/ko/tensorflow/svd.ipynb"},{label:"Mixed",value:"https://studiolab.sagemaker.aws/import/github/huggingface/notebooks/blob/main/diffusers_doc/ko/svd.ipynb"},{label:"PyTorch",value:"https://studiolab.sagemaker.aws/import/github/huggingface/notebooks/blob/main/diffusers_doc/ko/pytorch/svd.ipynb"},{label:"TensorFlow",value:"https://studiolab.sagemaker.aws/import/github/huggingface/notebooks/blob/main/diffusers_doc/ko/tensorflow/svd.ipynb"}]}}),w=new Y({props:{code:"IXBpcCUyMGluc3RhbGwlMjAtcSUyMC1VJTIwZGlmZnVzZXJzJTIwdHJhbnNmb3JtZXJzJTIwYWNjZWxlcmF0ZQ==",highlighted:"!pip install -q -U diffusers transformers accelerate",wrap:!1}}),V=new Y({props:{code:"aW1wb3J0JTIwdG9yY2glMEElMEFmcm9tJTIwZGlmZnVzZXJzJTIwaW1wb3J0JTIwU3RhYmxlVmlkZW9EaWZmdXNpb25QaXBlbGluZSUwQWZyb20lMjBkaWZmdXNlcnMudXRpbHMlMjBpbXBvcnQlMjBsb2FkX2ltYWdlJTJDJTIwZXhwb3J0X3RvX3ZpZGVvJTBBJTBBcGlwZSUyMCUzRCUyMFN0YWJsZVZpZGVvRGlmZnVzaW9uUGlwZWxpbmUuZnJvbV9wcmV0cmFpbmVkKCUwQSUyMCUyMCUyMCUyMCUyMnN0YWJpbGl0eWFpJTJGc3RhYmxlLXZpZGVvLWRpZmZ1c2lvbi1pbWcydmlkLXh0JTIyJTJDJTIwdG9yY2hfZHR5cGUlM0R0b3JjaC5mbG9hdDE2JTJDJTIwdmFyaWFudCUzRCUyMmZwMTYlMjIlMEEpJTBBcGlwZS5lbmFibGVfbW9kZWxfY3B1X29mZmxvYWQoKSUwQSUwQSUyMyUyMENvbmRpdGlvbmluZyUyMCVFQyU5RCVCNCVFQiVBRiVCOCVFQyVBNyU4MCUyMCVFQiVCNiU4OCVFQiU5RiVBQyVFQyU5OCVBNCVFQSVCOCVCMCUwQWltYWdlJTIwJTNEJTIwbG9hZF9pbWFnZSglMjJodHRwcyUzQSUyRiUyRmh1Z2dpbmdmYWNlLmNvJTJGZGF0YXNldHMlMkZodWdnaW5nZmFjZSUyRmRvY3VtZW50YXRpb24taW1hZ2VzJTJGcmVzb2x2ZSUyRm1haW4lMkZkaWZmdXNlcnMlMkZzdmQlMkZyb2NrZXQucG5nJTIyKSUwQWltYWdlJTIwJTNEJTIwaW1hZ2UucmVzaXplKCgxMDI0JTJDJTIwNTc2KSklMEElMEFnZW5lcmF0b3IlMjAlM0QlMjB0b3JjaC5tYW51YWxfc2VlZCg0MiklMEFmcmFtZXMlMjAlM0QlMjBwaXBlKGltYWdlJTJDJTIwZGVjb2RlX2NodW5rX3NpemUlM0Q4JTJDJTIwZ2VuZXJhdG9yJTNEZ2VuZXJhdG9yKS5mcmFtZXMlNUIwJTVEJTBBJTBBZXhwb3J0X3RvX3ZpZGVvKGZyYW1lcyUyQyUyMCUyMmdlbmVyYXRlZC5tcDQlMjIlMkMlMjBmcHMlM0Q3KQ==",highlighted:`<span class="hljs-keyword">import</span> torch
<span class="hljs-keyword">from</span> diffusers <span class="hljs-keyword">import</span> StableVideoDiffusionPipeline
<span class="hljs-keyword">from</span> diffusers.utils <span class="hljs-keyword">import</span> load_image, export_to_video
pipe = StableVideoDiffusionPipeline.from_pretrained(
<span class="hljs-string">&quot;stabilityai/stable-video-diffusion-img2vid-xt&quot;</span>, torch_dtype=torch.float16, variant=<span class="hljs-string">&quot;fp16&quot;</span>
)
pipe.enable_model_cpu_offload()
<span class="hljs-comment"># Conditioning 이미지 불러오기</span>
image = load_image(<span class="hljs-string">&quot;https://huggingface.co/datasets/huggingface/documentation-images/resolve/main/diffusers/svd/rocket.png&quot;</span>)
image = image.resize((<span class="hljs-number">1024</span>, <span class="hljs-number">576</span>))
generator = torch.manual_seed(<span class="hljs-number">42</span>)
frames = pipe(image, decode_chunk_size=<span class="hljs-number">8</span>, generator=generator).frames[<span class="hljs-number">0</span>]
export_to_video(frames, <span class="hljs-string">&quot;generated.mp4&quot;</span>, fps=<span class="hljs-number">7</span>)`,wrap:!1}}),U=new Me({props:{title:"torch.compile",local:"torchcompile",headingTag:"h2"}}),k=new Y({props:{code:"LSUyMHBpcGUuZW5hYmxlX21vZGVsX2NwdV9vZmZsb2FkKCklMEElMkIlMjBwaXBlLnRvKCUyMmN1ZGElMjIpJTBBJTJCJTIwcGlwZS51bmV0JTIwJTNEJTIwdG9yY2guY29tcGlsZShwaXBlLnVuZXQlMkMlMjBtb2RlJTNEJTIycmVkdWNlLW92ZXJoZWFkJTIyJTJDJTIwZnVsbGdyYXBoJTNEVHJ1ZSk=",highlighted:`<span class="hljs-deletion">- pipe.enable_model_cpu_offload()</span>
<span class="hljs-addition">+ pipe.to(&quot;cuda&quot;)</span>
<span class="hljs-addition">+ pipe.unet = torch.compile(pipe.unet, mode=&quot;reduce-overhead&quot;, fullgraph=True)</span>`,wrap:!1}}),_=new Me({props:{title:"메모리 사용량 줄이기",local:"메모리-사용량-줄이기",headingTag:"h2"}}),j=new Y({props:{code:"LSUyMHBpcGUuZW5hYmxlX21vZGVsX2NwdV9vZmZsb2FkKCklMEEtJTIwZnJhbWVzJTIwJTNEJTIwcGlwZShpbWFnZSUyQyUyMGRlY29kZV9jaHVua19zaXplJTNEOCUyQyUyMGdlbmVyYXRvciUzRGdlbmVyYXRvcikuZnJhbWVzJTVCMCU1RCUwQSUyQiUyMHBpcGUuZW5hYmxlX21vZGVsX2NwdV9vZmZsb2FkKCklMEElMkIlMjBwaXBlLnVuZXQuZW5hYmxlX2ZvcndhcmRfY2h1bmtpbmcoKSUwQSUyQiUyMGZyYW1lcyUyMCUzRCUyMHBpcGUoaW1hZ2UlMkMlMjBkZWNvZGVfY2h1bmtfc2l6ZSUzRDIlMkMlMjBnZW5lcmF0b3IlM0RnZW5lcmF0b3IlMkMlMjBudW1fZnJhbWVzJTNEMjUpLmZyYW1lcyU1QjAlNUQ=",highlighted:`<span class="hljs-deletion">- pipe.enable_model_cpu_offload()</span>
<span class="hljs-deletion">- frames = pipe(image, decode_chunk_size=8, generator=generator).frames[0]</span>
<span class="hljs-addition">+ pipe.enable_model_cpu_offload()</span>
<span class="hljs-addition">+ pipe.unet.enable_forward_chunking()</span>
<span class="hljs-addition">+ frames = pipe(image, decode_chunk_size=2, generator=generator, num_frames=25).frames[0]</span>`,wrap:!1}}),G=new Me({props:{title:"Micro-conditioning",local:"micro-conditioning",headingTag:"h2"}}),F=new Y({props:{code:"aW1wb3J0JTIwdG9yY2glMEElMEFmcm9tJTIwZGlmZnVzZXJzJTIwaW1wb3J0JTIwU3RhYmxlVmlkZW9EaWZmdXNpb25QaXBlbGluZSUwQWZyb20lMjBkaWZmdXNlcnMudXRpbHMlMjBpbXBvcnQlMjBsb2FkX2ltYWdlJTJDJTIwZXhwb3J0X3RvX3ZpZGVvJTBBJTBBcGlwZSUyMCUzRCUyMFN0YWJsZVZpZGVvRGlmZnVzaW9uUGlwZWxpbmUuZnJvbV9wcmV0cmFpbmVkKCUwQSUyMCUyMCUyMnN0YWJpbGl0eWFpJTJGc3RhYmxlLXZpZGVvLWRpZmZ1c2lvbi1pbWcydmlkLXh0JTIyJTJDJTIwdG9yY2hfZHR5cGUlM0R0b3JjaC5mbG9hdDE2JTJDJTIwdmFyaWFudCUzRCUyMmZwMTYlMjIlMEEpJTBBcGlwZS5lbmFibGVfbW9kZWxfY3B1X29mZmxvYWQoKSUwQSUwQSUyMyUyMENvbmRpdGlvbmluZyUyMCVFQyU5RCVCNCVFQiVBRiVCOCVFQyVBNyU4MCUyMCVFQiVCNiU4OCVFQiU5RiVBQyVFQyU5OCVBNCVFQSVCOCVCMCUwQWltYWdlJTIwJTNEJTIwbG9hZF9pbWFnZSglMjJodHRwcyUzQSUyRiUyRmh1Z2dpbmdmYWNlLmNvJTJGZGF0YXNldHMlMkZodWdnaW5nZmFjZSUyRmRvY3VtZW50YXRpb24taW1hZ2VzJTJGcmVzb2x2ZSUyRm1haW4lMkZkaWZmdXNlcnMlMkZzdmQlMkZyb2NrZXQucG5nJTIyKSUwQWltYWdlJTIwJTNEJTIwaW1hZ2UucmVzaXplKCgxMDI0JTJDJTIwNTc2KSklMEElMEFnZW5lcmF0b3IlMjAlM0QlMjB0b3JjaC5tYW51YWxfc2VlZCg0MiklMEFmcmFtZXMlMjAlM0QlMjBwaXBlKGltYWdlJTJDJTIwZGVjb2RlX2NodW5rX3NpemUlM0Q4JTJDJTIwZ2VuZXJhdG9yJTNEZ2VuZXJhdG9yJTJDJTIwbW90aW9uX2J1Y2tldF9pZCUzRDE4MCUyQyUyMG5vaXNlX2F1Z19zdHJlbmd0aCUzRDAuMSkuZnJhbWVzJTVCMCU1RCUwQWV4cG9ydF90b192aWRlbyhmcmFtZXMlMkMlMjAlMjJnZW5lcmF0ZWQubXA0JTIyJTJDJTIwZnBzJTNENyk=",highlighted:`<span class="hljs-keyword">import</span> torch
<span class="hljs-keyword">from</span> diffusers <span class="hljs-keyword">import</span> StableVideoDiffusionPipeline
<span class="hljs-keyword">from</span> diffusers.utils <span class="hljs-keyword">import</span> load_image, export_to_video
pipe = StableVideoDiffusionPipeline.from_pretrained(
<span class="hljs-string">&quot;stabilityai/stable-video-diffusion-img2vid-xt&quot;</span>, torch_dtype=torch.float16, variant=<span class="hljs-string">&quot;fp16&quot;</span>
)
pipe.enable_model_cpu_offload()
<span class="hljs-comment"># Conditioning 이미지 불러오기</span>
image = load_image(<span class="hljs-string">&quot;https://huggingface.co/datasets/huggingface/documentation-images/resolve/main/diffusers/svd/rocket.png&quot;</span>)
image = image.resize((<span class="hljs-number">1024</span>, <span class="hljs-number">576</span>))
generator = torch.manual_seed(<span class="hljs-number">42</span>)
frames = pipe(image, decode_chunk_size=<span class="hljs-number">8</span>, generator=generator, motion_bucket_id=<span class="hljs-number">180</span>, noise_aug_strength=<span class="hljs-number">0.1</span>).frames[<span class="hljs-number">0</span>]
export_to_video(frames, <span class="hljs-string">&quot;generated.mp4&quot;</span>, fps=<span class="hljs-number">7</span>)`,wrap:!1}}),B=new Ee({props:{source:"https://github.com/huggingface/diffusers/blob/main/docs/source/ko/using-diffusers/svd.md"}}),{c(){b=a("meta"),E=i(),N=a("p"),z=i(),m(h.$$.fragment),D=i(),m(M.$$.fragment),H=i(),Z=a("p"),Z.innerHTML=ye,L=i(),J=a("p"),J.textContent=we,q=i(),y=a("p"),y.textContent=Te,P=i(),m(w.$$.fragment),K=i(),T=a("p"),T.innerHTML=ve,A=i(),v=a("p"),v.textContent=Ve,O=i(),m(V.$$.fragment),ee=i(),g=a("div"),g.innerHTML=Ue,te=i(),m(U.$$.fragment),le=i(),C=a("p"),C.innerHTML=Ce,se=i(),m(k.$$.fragment),ie=i(),m(_.$$.fragment),ne=i(),W=a("p"),W.textContent=ke,ae=i(),$=a("ul"),$.innerHTML=_e,oe=i(),m(j.$$.fragment),pe=i(),X=a("p"),X.textContent=We,me=i(),m(G.$$.fragment),de=i(),x=a("p"),x.textContent=$e,ce=i(),Q=a("ul"),Q.innerHTML=je,re=i(),R=a("p"),R.innerHTML=Xe,fe=i(),m(F.$$.fragment),ue=i(),I=a("p"),I.innerHTML=Ge,be=i(),m(B.$$.fragment),ge=i(),S=a("p"),this.h()},l(e){const t=Ne("svelte-u9bgzb",document.head);b=o(t,"META",{name:!0,content:!0}),t.forEach(l),E=n(e),N=o(e,"P",{}),xe(N).forEach(l),z=n(e),d(h.$$.fragment,e),D=n(e),d(M.$$.fragment,e),H=n(e),Z=o(e,"P",{"data-svelte-h":!0}),p(Z)!=="svelte-1echk5x"&&(Z.innerHTML=ye),L=n(e),J=o(e,"P",{"data-svelte-h":!0}),p(J)!=="svelte-qrtvxi"&&(J.textContent=we),q=n(e),y=o(e,"P",{"data-svelte-h":!0}),p(y)!=="svelte-1dydg7a"&&(y.textContent=Te),P=n(e),d(w.$$.fragment,e),K=n(e),T=o(e,"P",{"data-svelte-h":!0}),p(T)!=="svelte-15oqpnn"&&(T.innerHTML=ve),A=n(e),v=o(e,"P",{"data-svelte-h":!0}),p(v)!=="svelte-1stbntq"&&(v.textContent=Ve),O=n(e),d(V.$$.fragment,e),ee=n(e),g=o(e,"DIV",{class:!0,"data-svelte-h":!0}),p(g)!=="svelte-25rdw"&&(g.innerHTML=Ue),te=n(e),d(U.$$.fragment,e),le=n(e),C=o(e,"P",{"data-svelte-h":!0}),p(C)!=="svelte-1d7iwu3"&&(C.innerHTML=Ce),se=n(e),d(k.$$.fragment,e),ie=n(e),d(_.$$.fragment,e),ne=n(e),W=o(e,"P",{"data-svelte-h":!0}),p(W)!=="svelte-ic28rl"&&(W.textContent=ke),ae=n(e),$=o(e,"UL",{"data-svelte-h":!0}),p($)!=="svelte-lce2t7"&&($.innerHTML=_e),oe=n(e),d(j.$$.fragment,e),pe=n(e),X=o(e,"P",{"data-svelte-h":!0}),p(X)!=="svelte-1v9bzrn"&&(X.textContent=We),me=n(e),d(G.$$.fragment,e),de=n(e),x=o(e,"P",{"data-svelte-h":!0}),p(x)!=="svelte-12yo38n"&&(x.textContent=$e),ce=n(e),Q=o(e,"UL",{"data-svelte-h":!0}),p(Q)!=="svelte-18zyz4a"&&(Q.innerHTML=je),re=n(e),R=o(e,"P",{"data-svelte-h":!0}),p(R)!=="svelte-1sx9h8v"&&(R.innerHTML=Xe),fe=n(e),d(F.$$.fragment,e),ue=n(e),I=o(e,"P",{"data-svelte-h":!0}),p(I)!=="svelte-19qaph4"&&(I.innerHTML=Ge),be=n(e),d(B.$$.fragment,e),ge=n(e),S=o(e,"P",{}),xe(S).forEach(l),this.h()},h(){Ze(b,"name","hf:doc:metadata"),Ze(b,"content",De),Ze(g,"class","flex gap-4")},m(e,t){Se(document.head,b),s(e,E,t),s(e,N,t),s(e,z,t),c(h,e,t),s(e,D,t),c(M,e,t),s(e,H,t),s(e,Z,t),s(e,L,t),s(e,J,t),s(e,q,t),s(e,y,t),s(e,P,t),c(w,e,t),s(e,K,t),s(e,T,t),s(e,A,t),s(e,v,t),s(e,O,t),c(V,e,t),s(e,ee,t),s(e,g,t),s(e,te,t),c(U,e,t),s(e,le,t),s(e,C,t),s(e,se,t),c(k,e,t),s(e,ie,t),c(_,e,t),s(e,ne,t),s(e,W,t),s(e,ae,t),s(e,$,t),s(e,oe,t),c(j,e,t),s(e,pe,t),s(e,X,t),s(e,me,t),c(G,e,t),s(e,de,t),s(e,x,t),s(e,ce,t),s(e,Q,t),s(e,re,t),s(e,R,t),s(e,fe,t),c(F,e,t),s(e,ue,t),s(e,I,t),s(e,be,t),c(B,e,t),s(e,ge,t),s(e,S,t),he=!0},p:Re,i(e){he||(r(h.$$.fragment,e),r(M.$$.fragment,e),r(w.$$.fragment,e),r(V.$$.fragment,e),r(U.$$.fragment,e),r(k.$$.fragment,e),r(_.$$.fragment,e),r(j.$$.fragment,e),r(G.$$.fragment,e),r(F.$$.fragment,e),r(B.$$.fragment,e),he=!0)},o(e){f(h.$$.fragment,e),f(M.$$.fragment,e),f(w.$$.fragment,e),f(V.$$.fragment,e),f(U.$$.fragment,e),f(k.$$.fragment,e),f(_.$$.fragment,e),f(j.$$.fragment,e),f(G.$$.fragment,e),f(F.$$.fragment,e),f(B.$$.fragment,e),he=!1},d(e){e&&(l(E),l(N),l(z),l(D),l(H),l(Z),l(L),l(J),l(q),l(y),l(P),l(K),l(T),l(A),l(v),l(O),l(ee),l(g),l(te),l(le),l(C),l(se),l(ie),l(ne),l(W),l(ae),l($),l(oe),l(pe),l(X),l(me),l(de),l(x),l(ce),l(Q),l(re),l(R),l(fe),l(ue),l(I),l(be),l(ge),l(S)),l(b),u(h,e),u(M,e),u(w,e),u(V,e),u(U,e),u(k,e),u(_,e),u(j,e),u(G,e),u(F,e),u(B,e)}}}const De='{"title":"Stable Video Diffusion","local":"stable-video-diffusion","sections":[{"title":"torch.compile","local":"torchcompile","sections":[],"depth":2},{"title":"메모리 사용량 줄이기","local":"메모리-사용량-줄이기","sections":[],"depth":2},{"title":"Micro-conditioning","local":"micro-conditioning","sections":[],"depth":2}],"depth":1}';function He(Je){return Fe(()=>{new URLSearchParams(window.location.search).get("fw")}),[]}class Oe extends Ie{constructor(b){super(),Be(this,b,He,ze,Qe,{})}}export{Oe as component};

Xet Storage Details

Size:
17.1 kB
·
Xet hash:
31ac8450fca588d344bc1c1bb847f74a0e3bae179e4b433df9d975c3deb43d2f

Xet efficiently stores files, intelligently splitting them into unique chunks and accelerating uploads and downloads. More info.