File size: 3,510 Bytes
2a4c86a
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
import argparse
from pathlib import Path
import sys

import torch

SCRIPT_DIR = Path(__file__).resolve().parent
if str(SCRIPT_DIR) not in sys.path:
    sys.path.insert(0, str(SCRIPT_DIR))

from jit_diffusers import JiTPipeline


def parse_args() -> argparse.Namespace:
    parser = argparse.ArgumentParser(description="Run single-image JiT diffusers inference.")
    parser.add_argument("--model_path", type=str, required=True, help="Path to converted diffusers model directory.")
    parser.add_argument("--output_path", type=str, required=True, help="Path to save output PNG image.")
    parser.add_argument("--class_label", type=int, default=207, help="ImageNet class id for conditional generation.")
    parser.add_argument("--seed", type=int, default=42, help="Random seed.")
    parser.add_argument("--steps", type=int, default=50, help="Number of ODE sampling steps.")
    parser.add_argument("--cfg", type=float, default=2.9, help="Classifier-free guidance scale.")
    parser.add_argument("--interval_min", type=float, default=0.1, help="CFG interval min.")
    parser.add_argument("--interval_max", type=float, default=1.0, help="CFG interval max.")
    parser.add_argument("--noise_scale", type=float, default=2.0, help="Initial Gaussian noise scale.")
    parser.add_argument("--t_eps", type=float, default=5e-2, help="Small epsilon for timestep denominator.")
    parser.add_argument(
        "--device",
        type=str,
        default="auto",
        choices=["auto", "cuda", "cpu"],
        help="Inference device.",
    )
    parser.add_argument(
        "--dtype",
        type=str,
        default="bf16",
        choices=["bf16", "fp32"],
        help="Inference dtype. Defaults to bf16 on CUDA.",
    )
    parser.add_argument(
        "--solver",
        type=str,
        default="scheduler",
        choices=["scheduler", "heun", "euler"],
        help="Sampling solver. Use scheduler to keep pipeline default.",
    )
    return parser.parse_args()


def resolve_device(name: str) -> torch.device:
    if name == "auto":
        return torch.device("cuda" if torch.cuda.is_available() else "cpu")
    return torch.device(name)


def resolve_dtype(name: str, device: torch.device) -> torch.dtype:
    if name == "bf16":
        return torch.bfloat16 if device.type == "cuda" else torch.float32
    return torch.float32


def main() -> None:
    args = parse_args()
    device = resolve_device(args.device)
    dtype = resolve_dtype(args.dtype, device)
    if device.type == "cuda":
        torch.set_float32_matmul_precision("high")

    pipe = JiTPipeline.from_pretrained(args.model_path).to(device)
    pipe.transformer = pipe.transformer.to(device=device, dtype=dtype)
    pipe.transformer.eval()
    sampling_method = None if args.solver == "scheduler" else args.solver

    generator = torch.Generator(device=device).manual_seed(args.seed)
    output = pipe(
        class_labels=[args.class_label],
        num_inference_steps=args.steps,
        guidance_scale=args.cfg,
        guidance_interval_min=args.interval_min,
        guidance_interval_max=args.interval_max,
        noise_scale=args.noise_scale,
        t_eps=args.t_eps,
        sampling_method=sampling_method,
        generator=generator,
        output_type="pil",
    )
    image = output.images[0]

    output_path = Path(args.output_path)
    output_path.parent.mkdir(parents=True, exist_ok=True)
    image.save(output_path)
    print(f"Saved image to: {output_path}")


if __name__ == "__main__":
    main()