Buckets:

hf-doc-build/doc / diffusers /main /en /api /models /autoencoder_kl_kvae.md
HuggingFaceDocBuilder's picture
|
download
raw
3 kB
# AutoencoderKLKVAE
The 2D variational autoencoder (VAE) model with KL loss.
The model can be loaded with the following code snippet.
```python
import torch
from diffusers import AutoencoderKLKVAE
vae = AutoencoderKLKVAE.from_pretrained("kandinskylab/KVAE-2D-1.0", subfolder="diffusers", torch_dtype=torch.bfloat16)
```
## AutoencoderKLKVAE[[diffusers.AutoencoderKLKVAE]]
#### diffusers.AutoencoderKLKVAE[[diffusers.AutoencoderKLKVAE]]
[Source](https://github.com/huggingface/diffusers/blob/main/src/diffusers/models/autoencoders/autoencoder_kl_kvae.py#L521)
A VAE model with KL loss for encoding images into latents and decoding latent representations into images.
This model inherits from [ModelMixin](/docs/diffusers/main/en/api/models/overview#diffusers.ModelMixin). Check the superclass documentation for its generic methods implemented for
all models (such as downloading or saving).
wrapperdiffusers.AutoencoderKLKVAE.decodehttps://github.com/huggingface/diffusers/blob/main/src/diffusers/utils/accelerate_utils.py#L43[{"name": "*args", "val": ""}, {"name": "**kwargs", "val": ""}]
**Parameters:**
in_channels (int, *optional*, defaults to 3) : Number of channels in the input image.
channels (int, *optional*, defaults to 128) : The base number of channels in multiresolution blocks.
num_enc_blocks (int, *optional*, defaults to 2) : The number of Resnet blocks in encoder multiresolution layers.
num_dec_blocks (int, *optional*, defaults to 2) : The number of Resnet blocks in decoder multiresolution layers.
z_channels (int, *optional*, defaults to 16) : Number of channels in the latent space.
double_z (`bool`, *optional*, defaults to `True`) : Whether to double the number of output channels of encoder.
ch_mult (`Tuple[int, ...]`, *optional*, default to `(1, 2, 4, 8)`) : The channel multipliers in multiresolution blocks.
sample_size (`int`, *optional*, defaults to `1024`) : Sample input size.
#### forward[[diffusers.AutoencoderKLKVAE.forward]]
[Source](https://github.com/huggingface/diffusers/blob/main/src/diffusers/models/autoencoders/autoencoder_kl_kvae.py#L776)
**Parameters:**
sample (`torch.Tensor`) : Input sample.
sample_posterior (`bool`, *optional*, defaults to `False`) : Whether to sample from the posterior.
return_dict (`bool`, *optional*, defaults to `True`) : Whether or not to return a `DecoderOutput` instead of a plain tuple.
#### tiled_decode[[diffusers.AutoencoderKLKVAE.tiled_decode]]
[Source](https://github.com/huggingface/diffusers/blob/main/src/diffusers/models/autoencoders/autoencoder_kl_kvae.py#L729)
Decode a batch of images using a tiled decoder.
**Parameters:**
z (`torch.Tensor`) : Input batch of latent vectors.
return_dict (`bool`, *optional*, defaults to `True`) : Whether or not to return a `~models.vae.DecoderOutput` instead of a plain tuple.
**Returns:**
``~models.vae.DecoderOutput` or `tuple``
If return_dict is True, a `~models.vae.DecoderOutput` is returned, otherwise a plain `tuple` is
returned.

Xet Storage Details

Size:
3 kB
·
Xet hash:
1d3c81109fda8b7bb93106029d75ebfbe4d02ab8a6edb2833cae7ead78a87c3a

Xet efficiently stores files, intelligently splitting them into unique chunks and accelerating uploads and downloads. More info.