Buckets:
| import{s as bt,o as St,n as Mt}from"../chunks/scheduler.8c3d61f6.js";import{S as xt,i as $t,g as i,s as r,r as m,A as yt,h as d,f as s,c as n,j as b,u as f,x as u,k as S,y as t,a as c,v as h,d as g,t as _,w as v}from"../chunks/index.da70eac4.js";import{T as Dt}from"../chunks/Tip.1d9b8c37.js";import{D as $}from"../chunks/Docstring.6b390b9a.js";import{H as ze,E as Pt}from"../chunks/EditOnGithub.1e64e623.js";function Tt(fe){let p,I=`The algorithm and model type are decoupled. You can use either DPMSolver or DPMSolver++ for both noise | |
| prediction and data prediction models.`;return{c(){p=i("p"),p.textContent=I},l(M){p=d(M,"P",{"data-svelte-h":!0}),u(p)!=="svelte-95n5s"&&(p.textContent=I)},m(M,ee){c(M,p,ee)},p:Mt,d(M){M&&s(p)}}}function wt(fe){let p,I,M,ee,q,he,N,ot='<code>DPMSolverMultistepScheduler</code> is a multistep scheduler from <a href="https://huggingface.co/papers/2206.00927" rel="nofollow">DPM-Solver: A Fast ODE Solver for Diffusion Probabilistic Model Sampling in Around 10 Steps</a> and <a href="https://huggingface.co/papers/2211.01095" rel="nofollow">DPM-Solver++: Fast Solver for Guided Sampling of Diffusion Probabilistic Models</a> by Cheng Lu, Yuhao Zhou, Fan Bao, Jianfei Chen, Chongxuan Li, and Jun Zhu.',ge,z,rt=`DPMSolver (and the improved version DPMSolver++) is a fast dedicated high-order solver for diffusion ODEs with convergence order guarantee. Empirically, DPMSolver sampling with only 20 steps can generate high-quality | |
| samples, and it can generate quite good samples even in 10 steps.`,_e,F,ve,k,nt="It is recommended to set <code>solver_order</code> to 2 for guide sampling, and <code>solver_order=3</code> for unconditional sampling.",be,H,it=`Dynamic thresholding from <a href="https://huggingface.co/papers/2205.11487" rel="nofollow">Imagen</a> is supported, and for pixel-space | |
| diffusion models, you can set both <code>algorithm_type="dpmsolver++"</code> and <code>thresholding=True</code> to use the dynamic | |
| thresholding. This thresholding method is unsuitable for latent-space diffusion models such as | |
| Stable Diffusion.`,Se,A,dt="The SDE variant of DPMSolver and DPM-Solver++ is also supported, but only for the first and second-order solvers. This is a fast SDE solver for the reverse diffusion SDE. It is recommended to use the second-order <code>sde-dpmsolver++</code>.",Me,V,xe,a,W,Fe,te,at="<code>DPMSolverMultistepScheduler</code> is a fast dedicated high-order solver for diffusion ODEs.",ke,se,lt=`This model inherits from <a href="/docs/diffusers/pr_10312/en/api/schedulers/overview#diffusers.SchedulerMixin">SchedulerMixin</a> and <a href="/docs/diffusers/pr_10312/en/api/configuration#diffusers.ConfigMixin">ConfigMixin</a>. Check the superclass documentation for the generic | |
| methods the library implements for all schedulers such as loading and saving.`,He,x,R,Ae,oe,ct=`Convert the model output to the corresponding type the DPMSolver/DPMSolver++ algorithm needs. DPM-Solver is | |
| designed to discretize an integral of the noise prediction model, and DPM-Solver++ is designed to discretize an | |
| integral of the data prediction model.`,Ve,D,We,P,U,Re,re,pt="One step for the first-order DPMSolver (equivalent to DDIM).",Ue,T,j,je,ne,ut="One step for the second-order multistep DPMSolver.",Ge,w,G,Be,ie,mt="One step for the third-order multistep DPMSolver.",Ye,C,B,Je,de,ft=`Ensures interchangeability with schedulers that need to scale the denoising model input depending on the | |
| current timestep.`,Ze,L,Y,Ke,ae,ht="Sets the begin index for the scheduler. This function should be run from pipeline before the inference.",Qe,O,J,Xe,le,gt="Sets the discrete timesteps used for the diffusion chain (to be run before inference).",et,E,Z,tt,ce,_t=`Predict the sample from the previous timestep by reversing the SDE. This function propagates the sample with | |
| the multistep DPMSolver.`,$e,K,ye,y,Q,st,pe,vt="Base class for the output of a scheduler’s <code>step</code> function.",De,X,Pe,me,Te;return q=new ze({props:{title:"DPMSolverMultistepScheduler",local:"dpmsolvermultistepscheduler",headingTag:"h1"}}),F=new ze({props:{title:"Tips",local:"tips",headingTag:"h2"}}),V=new ze({props:{title:"DPMSolverMultistepScheduler",local:"diffusers.DPMSolverMultistepScheduler",headingTag:"h2"}}),W=new $({props:{name:"class diffusers.DPMSolverMultistepScheduler",anchor:"diffusers.DPMSolverMultistepScheduler",parameters:[{name:"num_train_timesteps",val:": int = 1000"},{name:"beta_start",val:": float = 0.0001"},{name:"beta_end",val:": float = 0.02"},{name:"beta_schedule",val:": str = 'linear'"},{name:"trained_betas",val:": typing.Union[numpy.ndarray, typing.List[float], NoneType] = None"},{name:"solver_order",val:": int = 2"},{name:"prediction_type",val:": str = 'epsilon'"},{name:"thresholding",val:": bool = False"},{name:"dynamic_thresholding_ratio",val:": float = 0.995"},{name:"sample_max_value",val:": float = 1.0"},{name:"algorithm_type",val:": str = 'dpmsolver++'"},{name:"solver_type",val:": str = 'midpoint'"},{name:"lower_order_final",val:": bool = True"},{name:"euler_at_final",val:": bool = False"},{name:"use_karras_sigmas",val:": typing.Optional[bool] = False"},{name:"use_exponential_sigmas",val:": typing.Optional[bool] = False"},{name:"use_beta_sigmas",val:": typing.Optional[bool] = False"},{name:"use_lu_lambdas",val:": typing.Optional[bool] = False"},{name:"use_flow_sigmas",val:": typing.Optional[bool] = False"},{name:"flow_shift",val:": typing.Optional[float] = 1.0"},{name:"final_sigmas_type",val:": typing.Optional[str] = 'zero'"},{name:"lambda_min_clipped",val:": float = -inf"},{name:"variance_type",val:": typing.Optional[str] = None"},{name:"timestep_spacing",val:": str = 'linspace'"},{name:"steps_offset",val:": int = 0"},{name:"rescale_betas_zero_snr",val:": bool = False"}],parametersDescription:[{anchor:"diffusers.DPMSolverMultistepScheduler.num_train_timesteps",description:`<strong>num_train_timesteps</strong> (<code>int</code>, defaults to 1000) — | |
| The number of diffusion steps to train the model.`,name:"num_train_timesteps"},{anchor:"diffusers.DPMSolverMultistepScheduler.beta_start",description:`<strong>beta_start</strong> (<code>float</code>, defaults to 0.0001) — | |
| The starting <code>beta</code> value of inference.`,name:"beta_start"},{anchor:"diffusers.DPMSolverMultistepScheduler.beta_end",description:`<strong>beta_end</strong> (<code>float</code>, defaults to 0.02) — | |
| The final <code>beta</code> value.`,name:"beta_end"},{anchor:"diffusers.DPMSolverMultistepScheduler.beta_schedule",description:`<strong>beta_schedule</strong> (<code>str</code>, defaults to <code>"linear"</code>) — | |
| The beta schedule, a mapping from a beta range to a sequence of betas for stepping the model. Choose from | |
| <code>linear</code>, <code>scaled_linear</code>, or <code>squaredcos_cap_v2</code>.`,name:"beta_schedule"},{anchor:"diffusers.DPMSolverMultistepScheduler.trained_betas",description:`<strong>trained_betas</strong> (<code>np.ndarray</code>, <em>optional</em>) — | |
| Pass an array of betas directly to the constructor to bypass <code>beta_start</code> and <code>beta_end</code>.`,name:"trained_betas"},{anchor:"diffusers.DPMSolverMultistepScheduler.solver_order",description:`<strong>solver_order</strong> (<code>int</code>, defaults to 2) — | |
| The DPMSolver order which can be <code>1</code> or <code>2</code> or <code>3</code>. It is recommended to use <code>solver_order=2</code> for guided | |
| sampling, and <code>solver_order=3</code> for unconditional sampling.`,name:"solver_order"},{anchor:"diffusers.DPMSolverMultistepScheduler.prediction_type",description:`<strong>prediction_type</strong> (<code>str</code>, defaults to <code>epsilon</code>, <em>optional</em>) — | |
| Prediction type of the scheduler function; can be <code>epsilon</code> (predicts the noise of the diffusion process), | |
| <code>sample</code> (directly predicts the noisy sample<code>) or </code>v_prediction\` (see section 2.4 of <a href="https://imagen.research.google/video/paper.pdf" rel="nofollow">Imagen | |
| Video</a> paper).`,name:"prediction_type"},{anchor:"diffusers.DPMSolverMultistepScheduler.thresholding",description:`<strong>thresholding</strong> (<code>bool</code>, defaults to <code>False</code>) — | |
| Whether to use the “dynamic thresholding” method. This is unsuitable for latent-space diffusion models such | |
| as Stable Diffusion.`,name:"thresholding"},{anchor:"diffusers.DPMSolverMultistepScheduler.dynamic_thresholding_ratio",description:`<strong>dynamic_thresholding_ratio</strong> (<code>float</code>, defaults to 0.995) — | |
| The ratio for the dynamic thresholding method. Valid only when <code>thresholding=True</code>.`,name:"dynamic_thresholding_ratio"},{anchor:"diffusers.DPMSolverMultistepScheduler.sample_max_value",description:`<strong>sample_max_value</strong> (<code>float</code>, defaults to 1.0) — | |
| The threshold value for dynamic thresholding. Valid only when <code>thresholding=True</code> and | |
| <code>algorithm_type="dpmsolver++"</code>.`,name:"sample_max_value"},{anchor:"diffusers.DPMSolverMultistepScheduler.algorithm_type",description:`<strong>algorithm_type</strong> (<code>str</code>, defaults to <code>dpmsolver++</code>) — | |
| Algorithm type for the solver; can be <code>dpmsolver</code>, <code>dpmsolver++</code>, <code>sde-dpmsolver</code> or <code>sde-dpmsolver++</code>. The | |
| <code>dpmsolver</code> type implements the algorithms in the <a href="https://huggingface.co/papers/2206.00927" rel="nofollow">DPMSolver</a> | |
| paper, and the <code>dpmsolver++</code> type implements the algorithms in the | |
| <a href="https://huggingface.co/papers/2211.01095" rel="nofollow">DPMSolver++</a> paper. It is recommended to use <code>dpmsolver++</code> or | |
| <code>sde-dpmsolver++</code> with <code>solver_order=2</code> for guided sampling like in Stable Diffusion.`,name:"algorithm_type"},{anchor:"diffusers.DPMSolverMultistepScheduler.solver_type",description:`<strong>solver_type</strong> (<code>str</code>, defaults to <code>midpoint</code>) — | |
| Solver type for the second-order solver; can be <code>midpoint</code> or <code>heun</code>. The solver type slightly affects the | |
| sample quality, especially for a small number of steps. It is recommended to use <code>midpoint</code> solvers.`,name:"solver_type"},{anchor:"diffusers.DPMSolverMultistepScheduler.lower_order_final",description:`<strong>lower_order_final</strong> (<code>bool</code>, defaults to <code>True</code>) — | |
| Whether to use lower-order solvers in the final steps. Only valid for < 15 inference steps. This can | |
| stabilize the sampling of DPMSolver for steps < 15, especially for steps <= 10.`,name:"lower_order_final"},{anchor:"diffusers.DPMSolverMultistepScheduler.euler_at_final",description:`<strong>euler_at_final</strong> (<code>bool</code>, defaults to <code>False</code>) — | |
| Whether to use Euler’s method in the final step. It is a trade-off between numerical stability and detail | |
| richness. This can stabilize the sampling of the SDE variant of DPMSolver for small number of inference | |
| steps, but sometimes may result in blurring.`,name:"euler_at_final"},{anchor:"diffusers.DPMSolverMultistepScheduler.use_karras_sigmas",description:`<strong>use_karras_sigmas</strong> (<code>bool</code>, <em>optional</em>, defaults to <code>False</code>) — | |
| Whether to use Karras sigmas for step sizes in the noise schedule during the sampling process. If <code>True</code>, | |
| the sigmas are determined according to a sequence of noise levels {σi}.`,name:"use_karras_sigmas"},{anchor:"diffusers.DPMSolverMultistepScheduler.use_exponential_sigmas",description:`<strong>use_exponential_sigmas</strong> (<code>bool</code>, <em>optional</em>, defaults to <code>False</code>) — | |
| Whether to use exponential sigmas for step sizes in the noise schedule during the sampling process.`,name:"use_exponential_sigmas"},{anchor:"diffusers.DPMSolverMultistepScheduler.use_beta_sigmas",description:`<strong>use_beta_sigmas</strong> (<code>bool</code>, <em>optional</em>, defaults to <code>False</code>) — | |
| Whether to use beta sigmas for step sizes in the noise schedule during the sampling process. Refer to <a href="https://huggingface.co/papers/2407.12173" rel="nofollow">Beta | |
| Sampling is All You Need</a> for more information.`,name:"use_beta_sigmas"},{anchor:"diffusers.DPMSolverMultistepScheduler.use_lu_lambdas",description:`<strong>use_lu_lambdas</strong> (<code>bool</code>, <em>optional</em>, defaults to <code>False</code>) — | |
| Whether to use the uniform-logSNR for step sizes proposed by Lu’s DPM-Solver in the noise schedule during | |
| the sampling process. If <code>True</code>, the sigmas and time steps are determined according to a sequence of | |
| <code>lambda(t)</code>.`,name:"use_lu_lambdas"},{anchor:"diffusers.DPMSolverMultistepScheduler.final_sigmas_type",description:`<strong>final_sigmas_type</strong> (<code>str</code>, defaults to <code>"zero"</code>) — | |
| The final <code>sigma</code> value for the noise schedule during the sampling process. If <code>"sigma_min"</code>, the final | |
| sigma is the same as the last sigma in the training schedule. If <code>zero</code>, the final sigma is set to 0.`,name:"final_sigmas_type"},{anchor:"diffusers.DPMSolverMultistepScheduler.lambda_min_clipped",description:`<strong>lambda_min_clipped</strong> (<code>float</code>, defaults to <code>-inf</code>) — | |
| Clipping threshold for the minimum value of <code>lambda(t)</code> for numerical stability. This is critical for the | |
| cosine (<code>squaredcos_cap_v2</code>) noise schedule.`,name:"lambda_min_clipped"},{anchor:"diffusers.DPMSolverMultistepScheduler.variance_type",description:`<strong>variance_type</strong> (<code>str</code>, <em>optional</em>) — | |
| Set to “learned” or “learned_range” for diffusion models that predict variance. If set, the model’s output | |
| contains the predicted Gaussian variance.`,name:"variance_type"},{anchor:"diffusers.DPMSolverMultistepScheduler.timestep_spacing",description:`<strong>timestep_spacing</strong> (<code>str</code>, defaults to <code>"linspace"</code>) — | |
| The way the timesteps should be scaled. Refer to Table 2 of the <a href="https://huggingface.co/papers/2305.08891" rel="nofollow">Common Diffusion Noise Schedules and | |
| Sample Steps are Flawed</a> for more information.`,name:"timestep_spacing"},{anchor:"diffusers.DPMSolverMultistepScheduler.steps_offset",description:`<strong>steps_offset</strong> (<code>int</code>, defaults to 0) — | |
| An offset added to the inference steps, as required by some model families.`,name:"steps_offset"},{anchor:"diffusers.DPMSolverMultistepScheduler.rescale_betas_zero_snr",description:`<strong>rescale_betas_zero_snr</strong> (<code>bool</code>, defaults to <code>False</code>) — | |
| Whether to rescale the betas to have zero terminal SNR. This enables the model to generate very bright and | |
| dark samples instead of limiting it to samples with medium brightness. Loosely related to | |
| <a href="https://github.com/huggingface/diffusers/blob/74fd735eb073eb1d774b1ab4154a0876eb82f055/examples/dreambooth/train_dreambooth.py#L506" rel="nofollow"><code>--offset_noise</code></a>.`,name:"rescale_betas_zero_snr"}],source:"https://github.com/huggingface/diffusers/blob/vr_10312/src/diffusers/schedulers/scheduling_dpmsolver_multistep.py#L115"}}),R=new $({props:{name:"convert_model_output",anchor:"diffusers.DPMSolverMultistepScheduler.convert_model_output",parameters:[{name:"model_output",val:": Tensor"},{name:"*args",val:""},{name:"sample",val:": Tensor = None"},{name:"**kwargs",val:""}],parametersDescription:[{anchor:"diffusers.DPMSolverMultistepScheduler.convert_model_output.model_output",description:`<strong>model_output</strong> (<code>torch.Tensor</code>) — | |
| The direct output from the learned diffusion model.`,name:"model_output"},{anchor:"diffusers.DPMSolverMultistepScheduler.convert_model_output.sample",description:`<strong>sample</strong> (<code>torch.Tensor</code>) — | |
| A current instance of a sample created by the diffusion process.`,name:"sample"}],source:"https://github.com/huggingface/diffusers/blob/vr_10312/src/diffusers/schedulers/scheduling_dpmsolver_multistep.py#L607",returnDescription:`<script context="module">export const metadata = 'undefined';<\/script> | |
| <p>The converted model output.</p> | |
| `,returnType:`<script context="module">export const metadata = 'undefined';<\/script> | |
| <p><code>torch.Tensor</code></p> | |
| `}}),D=new Dt({props:{$$slots:{default:[Tt]},$$scope:{ctx:fe}}}),U=new $({props:{name:"dpm_solver_first_order_update",anchor:"diffusers.DPMSolverMultistepScheduler.dpm_solver_first_order_update",parameters:[{name:"model_output",val:": Tensor"},{name:"*args",val:""},{name:"sample",val:": Tensor = None"},{name:"noise",val:": typing.Optional[torch.Tensor] = None"},{name:"**kwargs",val:""}],parametersDescription:[{anchor:"diffusers.DPMSolverMultistepScheduler.dpm_solver_first_order_update.model_output",description:`<strong>model_output</strong> (<code>torch.Tensor</code>) — | |
| The direct output from the learned diffusion model.`,name:"model_output"},{anchor:"diffusers.DPMSolverMultistepScheduler.dpm_solver_first_order_update.sample",description:`<strong>sample</strong> (<code>torch.Tensor</code>) — | |
| A current instance of a sample created by the diffusion process.`,name:"sample"}],source:"https://github.com/huggingface/diffusers/blob/vr_10312/src/diffusers/schedulers/scheduling_dpmsolver_multistep.py#L709",returnDescription:`<script context="module">export const metadata = 'undefined';<\/script> | |
| <p>The sample tensor at the previous timestep.</p> | |
| `,returnType:`<script context="module">export const metadata = 'undefined';<\/script> | |
| <p><code>torch.Tensor</code></p> | |
| `}}),j=new $({props:{name:"multistep_dpm_solver_second_order_update",anchor:"diffusers.DPMSolverMultistepScheduler.multistep_dpm_solver_second_order_update",parameters:[{name:"model_output_list",val:": typing.List[torch.Tensor]"},{name:"*args",val:""},{name:"sample",val:": Tensor = None"},{name:"noise",val:": typing.Optional[torch.Tensor] = None"},{name:"**kwargs",val:""}],parametersDescription:[{anchor:"diffusers.DPMSolverMultistepScheduler.multistep_dpm_solver_second_order_update.model_output_list",description:`<strong>model_output_list</strong> (<code>List[torch.Tensor]</code>) — | |
| The direct outputs from learned diffusion model at current and latter timesteps.`,name:"model_output_list"},{anchor:"diffusers.DPMSolverMultistepScheduler.multistep_dpm_solver_second_order_update.sample",description:`<strong>sample</strong> (<code>torch.Tensor</code>) — | |
| A current instance of a sample created by the diffusion process.`,name:"sample"}],source:"https://github.com/huggingface/diffusers/blob/vr_10312/src/diffusers/schedulers/scheduling_dpmsolver_multistep.py#L778",returnDescription:`<script context="module">export const metadata = 'undefined';<\/script> | |
| <p>The sample tensor at the previous timestep.</p> | |
| `,returnType:`<script context="module">export const metadata = 'undefined';<\/script> | |
| <p><code>torch.Tensor</code></p> | |
| `}}),G=new $({props:{name:"multistep_dpm_solver_third_order_update",anchor:"diffusers.DPMSolverMultistepScheduler.multistep_dpm_solver_third_order_update",parameters:[{name:"model_output_list",val:": typing.List[torch.Tensor]"},{name:"*args",val:""},{name:"sample",val:": Tensor = None"},{name:"noise",val:": typing.Optional[torch.Tensor] = None"},{name:"**kwargs",val:""}],parametersDescription:[{anchor:"diffusers.DPMSolverMultistepScheduler.multistep_dpm_solver_third_order_update.model_output_list",description:`<strong>model_output_list</strong> (<code>List[torch.Tensor]</code>) — | |
| The direct outputs from learned diffusion model at current and latter timesteps.`,name:"model_output_list"},{anchor:"diffusers.DPMSolverMultistepScheduler.multistep_dpm_solver_third_order_update.sample",description:`<strong>sample</strong> (<code>torch.Tensor</code>) — | |
| A current instance of a sample created by diffusion process.`,name:"sample"}],source:"https://github.com/huggingface/diffusers/blob/vr_10312/src/diffusers/schedulers/scheduling_dpmsolver_multistep.py#L901",returnDescription:`<script context="module">export const metadata = 'undefined';<\/script> | |
| <p>The sample tensor at the previous timestep.</p> | |
| `,returnType:`<script context="module">export const metadata = 'undefined';<\/script> | |
| <p><code>torch.Tensor</code></p> | |
| `}}),B=new $({props:{name:"scale_model_input",anchor:"diffusers.DPMSolverMultistepScheduler.scale_model_input",parameters:[{name:"sample",val:": Tensor"},{name:"*args",val:""},{name:"**kwargs",val:""}],parametersDescription:[{anchor:"diffusers.DPMSolverMultistepScheduler.scale_model_input.sample",description:`<strong>sample</strong> (<code>torch.Tensor</code>) — | |
| The input sample.`,name:"sample"}],source:"https://github.com/huggingface/diffusers/blob/vr_10312/src/diffusers/schedulers/scheduling_dpmsolver_multistep.py#L1116",returnDescription:`<script context="module">export const metadata = 'undefined';<\/script> | |
| <p>A scaled input sample.</p> | |
| `,returnType:`<script context="module">export const metadata = 'undefined';<\/script> | |
| <p><code>torch.Tensor</code></p> | |
| `}}),Y=new $({props:{name:"set_begin_index",anchor:"diffusers.DPMSolverMultistepScheduler.set_begin_index",parameters:[{name:"begin_index",val:": int = 0"}],parametersDescription:[{anchor:"diffusers.DPMSolverMultistepScheduler.set_begin_index.begin_index",description:`<strong>begin_index</strong> (<code>int</code>) — | |
| The begin index for the scheduler.`,name:"begin_index"}],source:"https://github.com/huggingface/diffusers/blob/vr_10312/src/diffusers/schedulers/scheduling_dpmsolver_multistep.py#L315"}}),J=new $({props:{name:"set_timesteps",anchor:"diffusers.DPMSolverMultistepScheduler.set_timesteps",parameters:[{name:"num_inference_steps",val:": int = None"},{name:"device",val:": typing.Union[str, torch.device] = None"},{name:"timesteps",val:": typing.Optional[typing.List[int]] = None"}],parametersDescription:[{anchor:"diffusers.DPMSolverMultistepScheduler.set_timesteps.num_inference_steps",description:`<strong>num_inference_steps</strong> (<code>int</code>) — | |
| The number of diffusion steps used when generating samples with a pre-trained model.`,name:"num_inference_steps"},{anchor:"diffusers.DPMSolverMultistepScheduler.set_timesteps.device",description:`<strong>device</strong> (<code>str</code> or <code>torch.device</code>, <em>optional</em>) — | |
| The device to which the timesteps should be moved to. If <code>None</code>, the timesteps are not moved.`,name:"device"},{anchor:"diffusers.DPMSolverMultistepScheduler.set_timesteps.timesteps",description:`<strong>timesteps</strong> (<code>List[int]</code>, <em>optional</em>) — | |
| Custom timesteps used to support arbitrary timesteps schedule. If <code>None</code>, timesteps will be generated | |
| based on the <code>timestep_spacing</code> attribute. If <code>timesteps</code> is passed, <code>num_inference_steps</code> and <code>sigmas</code> | |
| must be <code>None</code>, and <code>timestep_spacing</code> attribute will be ignored.`,name:"timesteps"}],source:"https://github.com/huggingface/diffusers/blob/vr_10312/src/diffusers/schedulers/scheduling_dpmsolver_multistep.py#L325"}}),Z=new $({props:{name:"step",anchor:"diffusers.DPMSolverMultistepScheduler.step",parameters:[{name:"model_output",val:": Tensor"},{name:"timestep",val:": typing.Union[int, torch.Tensor]"},{name:"sample",val:": Tensor"},{name:"generator",val:" = None"},{name:"variance_noise",val:": typing.Optional[torch.Tensor] = None"},{name:"return_dict",val:": bool = True"}],parametersDescription:[{anchor:"diffusers.DPMSolverMultistepScheduler.step.model_output",description:`<strong>model_output</strong> (<code>torch.Tensor</code>) — | |
| The direct output from learned diffusion model.`,name:"model_output"},{anchor:"diffusers.DPMSolverMultistepScheduler.step.timestep",description:`<strong>timestep</strong> (<code>int</code>) — | |
| The current discrete timestep in the diffusion chain.`,name:"timestep"},{anchor:"diffusers.DPMSolverMultistepScheduler.step.sample",description:`<strong>sample</strong> (<code>torch.Tensor</code>) — | |
| A current instance of a sample created by the diffusion process.`,name:"sample"},{anchor:"diffusers.DPMSolverMultistepScheduler.step.generator",description:`<strong>generator</strong> (<code>torch.Generator</code>, <em>optional</em>) — | |
| A random number generator.`,name:"generator"},{anchor:"diffusers.DPMSolverMultistepScheduler.step.variance_noise",description:`<strong>variance_noise</strong> (<code>torch.Tensor</code>) — | |
| Alternative to generating noise with <code>generator</code> by directly providing the noise for the variance | |
| itself. Useful for methods such as <code>LEdits++</code>.`,name:"variance_noise"},{anchor:"diffusers.DPMSolverMultistepScheduler.step.return_dict",description:`<strong>return_dict</strong> (<code>bool</code>) — | |
| Whether or not to return a <a href="/docs/diffusers/pr_10312/en/api/schedulers/multistep_dpm_solver#diffusers.schedulers.scheduling_utils.SchedulerOutput">SchedulerOutput</a> or <code>tuple</code>.`,name:"return_dict"}],source:"https://github.com/huggingface/diffusers/blob/vr_10312/src/diffusers/schedulers/scheduling_dpmsolver_multistep.py#L1027",returnDescription:`<script context="module">export const metadata = 'undefined';<\/script> | |
| <p>If return_dict is <code>True</code>, <a | |
| href="/docs/diffusers/pr_10312/en/api/schedulers/multistep_dpm_solver#diffusers.schedulers.scheduling_utils.SchedulerOutput" | |
| >SchedulerOutput</a> is returned, otherwise a | |
| tuple is returned where the first element is the sample tensor.</p> | |
| `,returnType:`<script context="module">export const metadata = 'undefined';<\/script> | |
| <p><a | |
| href="/docs/diffusers/pr_10312/en/api/schedulers/multistep_dpm_solver#diffusers.schedulers.scheduling_utils.SchedulerOutput" | |
| >SchedulerOutput</a> or <code>tuple</code></p> | |
| `}}),K=new ze({props:{title:"SchedulerOutput",local:"diffusers.schedulers.scheduling_utils.SchedulerOutput",headingTag:"h2"}}),Q=new $({props:{name:"class diffusers.schedulers.scheduling_utils.SchedulerOutput",anchor:"diffusers.schedulers.scheduling_utils.SchedulerOutput",parameters:[{name:"prev_sample",val:": Tensor"}],parametersDescription:[{anchor:"diffusers.schedulers.scheduling_utils.SchedulerOutput.prev_sample",description:`<strong>prev_sample</strong> (<code>torch.Tensor</code> of shape <code>(batch_size, num_channels, height, width)</code> for images) — | |
| Computed sample <code>(x_{t-1})</code> of previous timestep. <code>prev_sample</code> should be used as next model input in the | |
| denoising loop.`,name:"prev_sample"}],source:"https://github.com/huggingface/diffusers/blob/vr_10312/src/diffusers/schedulers/scheduling_utils.py#L60"}}),X=new Pt({props:{source:"https://github.com/huggingface/diffusers/blob/main/docs/source/en/api/schedulers/multistep_dpm_solver.md"}}),{c(){p=i("meta"),I=r(),M=i("p"),ee=r(),m(q.$$.fragment),he=r(),N=i("p"),N.innerHTML=ot,ge=r(),z=i("p"),z.textContent=rt,_e=r(),m(F.$$.fragment),ve=r(),k=i("p"),k.innerHTML=nt,be=r(),H=i("p"),H.innerHTML=it,Se=r(),A=i("p"),A.innerHTML=dt,Me=r(),m(V.$$.fragment),xe=r(),a=i("div"),m(W.$$.fragment),Fe=r(),te=i("p"),te.innerHTML=at,ke=r(),se=i("p"),se.innerHTML=lt,He=r(),x=i("div"),m(R.$$.fragment),Ae=r(),oe=i("p"),oe.textContent=ct,Ve=r(),m(D.$$.fragment),We=r(),P=i("div"),m(U.$$.fragment),Re=r(),re=i("p"),re.textContent=pt,Ue=r(),T=i("div"),m(j.$$.fragment),je=r(),ne=i("p"),ne.textContent=ut,Ge=r(),w=i("div"),m(G.$$.fragment),Be=r(),ie=i("p"),ie.textContent=mt,Ye=r(),C=i("div"),m(B.$$.fragment),Je=r(),de=i("p"),de.textContent=ft,Ze=r(),L=i("div"),m(Y.$$.fragment),Ke=r(),ae=i("p"),ae.textContent=ht,Qe=r(),O=i("div"),m(J.$$.fragment),Xe=r(),le=i("p"),le.textContent=gt,et=r(),E=i("div"),m(Z.$$.fragment),tt=r(),ce=i("p"),ce.textContent=_t,$e=r(),m(K.$$.fragment),ye=r(),y=i("div"),m(Q.$$.fragment),st=r(),pe=i("p"),pe.innerHTML=vt,De=r(),m(X.$$.fragment),Pe=r(),me=i("p"),this.h()},l(e){const o=yt("svelte-u9bgzb",document.head);p=d(o,"META",{name:!0,content:!0}),o.forEach(s),I=n(e),M=d(e,"P",{}),b(M).forEach(s),ee=n(e),f(q.$$.fragment,e),he=n(e),N=d(e,"P",{"data-svelte-h":!0}),u(N)!=="svelte-194x5ku"&&(N.innerHTML=ot),ge=n(e),z=d(e,"P",{"data-svelte-h":!0}),u(z)!=="svelte-l6rv24"&&(z.textContent=rt),_e=n(e),f(F.$$.fragment,e),ve=n(e),k=d(e,"P",{"data-svelte-h":!0}),u(k)!=="svelte-pt2glg"&&(k.innerHTML=nt),be=n(e),H=d(e,"P",{"data-svelte-h":!0}),u(H)!=="svelte-1ng0gyl"&&(H.innerHTML=it),Se=n(e),A=d(e,"P",{"data-svelte-h":!0}),u(A)!=="svelte-18n28n2"&&(A.innerHTML=dt),Me=n(e),f(V.$$.fragment,e),xe=n(e),a=d(e,"DIV",{class:!0});var l=b(a);f(W.$$.fragment,l),Fe=n(l),te=d(l,"P",{"data-svelte-h":!0}),u(te)!=="svelte-12v1qo2"&&(te.innerHTML=at),ke=n(l),se=d(l,"P",{"data-svelte-h":!0}),u(se)!=="svelte-1oywc2v"&&(se.innerHTML=lt),He=n(l),x=d(l,"DIV",{class:!0});var ue=b(x);f(R.$$.fragment,ue),Ae=n(ue),oe=d(ue,"P",{"data-svelte-h":!0}),u(oe)!=="svelte-1st02i8"&&(oe.textContent=ct),Ve=n(ue),f(D.$$.fragment,ue),ue.forEach(s),We=n(l),P=d(l,"DIV",{class:!0});var we=b(P);f(U.$$.fragment,we),Re=n(we),re=d(we,"P",{"data-svelte-h":!0}),u(re)!=="svelte-4nw43d"&&(re.textContent=pt),we.forEach(s),Ue=n(l),T=d(l,"DIV",{class:!0});var Ce=b(T);f(j.$$.fragment,Ce),je=n(Ce),ne=d(Ce,"P",{"data-svelte-h":!0}),u(ne)!=="svelte-1d6o6zu"&&(ne.textContent=ut),Ce.forEach(s),Ge=n(l),w=d(l,"DIV",{class:!0});var Le=b(w);f(G.$$.fragment,Le),Be=n(Le),ie=d(Le,"P",{"data-svelte-h":!0}),u(ie)!=="svelte-17nne8p"&&(ie.textContent=mt),Le.forEach(s),Ye=n(l),C=d(l,"DIV",{class:!0});var Oe=b(C);f(B.$$.fragment,Oe),Je=n(Oe),de=d(Oe,"P",{"data-svelte-h":!0}),u(de)!=="svelte-1rkfgpx"&&(de.textContent=ft),Oe.forEach(s),Ze=n(l),L=d(l,"DIV",{class:!0});var Ee=b(L);f(Y.$$.fragment,Ee),Ke=n(Ee),ae=d(Ee,"P",{"data-svelte-h":!0}),u(ae)!=="svelte-1k141rk"&&(ae.textContent=ht),Ee.forEach(s),Qe=n(l),O=d(l,"DIV",{class:!0});var Ie=b(O);f(J.$$.fragment,Ie),Xe=n(Ie),le=d(Ie,"P",{"data-svelte-h":!0}),u(le)!=="svelte-1vzm9q"&&(le.textContent=gt),Ie.forEach(s),et=n(l),E=d(l,"DIV",{class:!0});var qe=b(E);f(Z.$$.fragment,qe),tt=n(qe),ce=d(qe,"P",{"data-svelte-h":!0}),u(ce)!=="svelte-7jbjgx"&&(ce.textContent=_t),qe.forEach(s),l.forEach(s),$e=n(e),f(K.$$.fragment,e),ye=n(e),y=d(e,"DIV",{class:!0});var Ne=b(y);f(Q.$$.fragment,Ne),st=n(Ne),pe=d(Ne,"P",{"data-svelte-h":!0}),u(pe)!=="svelte-6ojmkw"&&(pe.innerHTML=vt),Ne.forEach(s),De=n(e),f(X.$$.fragment,e),Pe=n(e),me=d(e,"P",{}),b(me).forEach(s),this.h()},h(){S(p,"name","hf:doc:metadata"),S(p,"content",Ct),S(x,"class","docstring border-l-2 border-t-2 pl-4 pt-3.5 border-gray-100 rounded-tl-xl mb-6 mt-8"),S(P,"class","docstring border-l-2 border-t-2 pl-4 pt-3.5 border-gray-100 rounded-tl-xl mb-6 mt-8"),S(T,"class","docstring border-l-2 border-t-2 pl-4 pt-3.5 border-gray-100 rounded-tl-xl mb-6 mt-8"),S(w,"class","docstring border-l-2 border-t-2 pl-4 pt-3.5 border-gray-100 rounded-tl-xl mb-6 mt-8"),S(C,"class","docstring border-l-2 border-t-2 pl-4 pt-3.5 border-gray-100 rounded-tl-xl mb-6 mt-8"),S(L,"class","docstring border-l-2 border-t-2 pl-4 pt-3.5 border-gray-100 rounded-tl-xl mb-6 mt-8"),S(O,"class","docstring border-l-2 border-t-2 pl-4 pt-3.5 border-gray-100 rounded-tl-xl mb-6 mt-8"),S(E,"class","docstring border-l-2 border-t-2 pl-4 pt-3.5 border-gray-100 rounded-tl-xl mb-6 mt-8"),S(a,"class","docstring border-l-2 border-t-2 pl-4 pt-3.5 border-gray-100 rounded-tl-xl mb-6 mt-8"),S(y,"class","docstring border-l-2 border-t-2 pl-4 pt-3.5 border-gray-100 rounded-tl-xl mb-6 mt-8")},m(e,o){t(document.head,p),c(e,I,o),c(e,M,o),c(e,ee,o),h(q,e,o),c(e,he,o),c(e,N,o),c(e,ge,o),c(e,z,o),c(e,_e,o),h(F,e,o),c(e,ve,o),c(e,k,o),c(e,be,o),c(e,H,o),c(e,Se,o),c(e,A,o),c(e,Me,o),h(V,e,o),c(e,xe,o),c(e,a,o),h(W,a,null),t(a,Fe),t(a,te),t(a,ke),t(a,se),t(a,He),t(a,x),h(R,x,null),t(x,Ae),t(x,oe),t(x,Ve),h(D,x,null),t(a,We),t(a,P),h(U,P,null),t(P,Re),t(P,re),t(a,Ue),t(a,T),h(j,T,null),t(T,je),t(T,ne),t(a,Ge),t(a,w),h(G,w,null),t(w,Be),t(w,ie),t(a,Ye),t(a,C),h(B,C,null),t(C,Je),t(C,de),t(a,Ze),t(a,L),h(Y,L,null),t(L,Ke),t(L,ae),t(a,Qe),t(a,O),h(J,O,null),t(O,Xe),t(O,le),t(a,et),t(a,E),h(Z,E,null),t(E,tt),t(E,ce),c(e,$e,o),h(K,e,o),c(e,ye,o),c(e,y,o),h(Q,y,null),t(y,st),t(y,pe),c(e,De,o),h(X,e,o),c(e,Pe,o),c(e,me,o),Te=!0},p(e,[o]){const l={};o&2&&(l.$$scope={dirty:o,ctx:e}),D.$set(l)},i(e){Te||(g(q.$$.fragment,e),g(F.$$.fragment,e),g(V.$$.fragment,e),g(W.$$.fragment,e),g(R.$$.fragment,e),g(D.$$.fragment,e),g(U.$$.fragment,e),g(j.$$.fragment,e),g(G.$$.fragment,e),g(B.$$.fragment,e),g(Y.$$.fragment,e),g(J.$$.fragment,e),g(Z.$$.fragment,e),g(K.$$.fragment,e),g(Q.$$.fragment,e),g(X.$$.fragment,e),Te=!0)},o(e){_(q.$$.fragment,e),_(F.$$.fragment,e),_(V.$$.fragment,e),_(W.$$.fragment,e),_(R.$$.fragment,e),_(D.$$.fragment,e),_(U.$$.fragment,e),_(j.$$.fragment,e),_(G.$$.fragment,e),_(B.$$.fragment,e),_(Y.$$.fragment,e),_(J.$$.fragment,e),_(Z.$$.fragment,e),_(K.$$.fragment,e),_(Q.$$.fragment,e),_(X.$$.fragment,e),Te=!1},d(e){e&&(s(I),s(M),s(ee),s(he),s(N),s(ge),s(z),s(_e),s(ve),s(k),s(be),s(H),s(Se),s(A),s(Me),s(xe),s(a),s($e),s(ye),s(y),s(De),s(Pe),s(me)),s(p),v(q,e),v(F,e),v(V,e),v(W),v(R),v(D),v(U),v(j),v(G),v(B),v(Y),v(J),v(Z),v(K,e),v(Q),v(X,e)}}}const Ct='{"title":"DPMSolverMultistepScheduler","local":"dpmsolvermultistepscheduler","sections":[{"title":"Tips","local":"tips","sections":[],"depth":2},{"title":"DPMSolverMultistepScheduler","local":"diffusers.DPMSolverMultistepScheduler","sections":[],"depth":2},{"title":"SchedulerOutput","local":"diffusers.schedulers.scheduling_utils.SchedulerOutput","sections":[],"depth":2}],"depth":1}';function Lt(fe){return St(()=>{new URLSearchParams(window.location.search).get("fw")}),[]}class zt extends xt{constructor(p){super(),$t(this,p,Lt,wt,bt,{})}}export{zt as component}; | |
Xet Storage Details
- Size:
- 33.9 kB
- Xet hash:
- 01b226575df36ac455917afae4523194ddc38bc19b00d87cccc49a99c771cb45
·
Xet efficiently stores files, intelligently splitting them into unique chunks and accelerating uploads and downloads. More info.