Update app.py
Browse files
app.py
CHANGED
|
@@ -357,10 +357,6 @@ print(f"Checkpoint: {checkpoint_path}")
|
|
| 357 |
print(f"Spatial upsampler: {spatial_upsampler_path}")
|
| 358 |
print(f"Gemma root: {gemma_root}")
|
| 359 |
|
| 360 |
-
del pipeline
|
| 361 |
-
gc.collect()
|
| 362 |
-
torch.cuda.empty_cache()
|
| 363 |
-
|
| 364 |
# Initialize pipeline WITH text encoder and optional audio support
|
| 365 |
# ---- Replace block (pipeline init) lines 275-281 ----
|
| 366 |
pipeline = LTX23NegativePromptTwoStagePipeline(
|
|
@@ -491,6 +487,10 @@ def apply_prepared_lora_config_to_pipeline():
|
|
| 491 |
print("[LoRA] Prepared LoRA config already active; skipping.")
|
| 492 |
return True
|
| 493 |
|
|
|
|
|
|
|
|
|
|
|
|
|
| 494 |
pipeline = LTX23NegativePromptTwoStagePipeline(
|
| 495 |
checkpoint_path=str(checkpoint_path),
|
| 496 |
spatial_upsampler_path=str(spatial_upsampler_path),
|
|
|
|
| 357 |
print(f"Spatial upsampler: {spatial_upsampler_path}")
|
| 358 |
print(f"Gemma root: {gemma_root}")
|
| 359 |
|
|
|
|
|
|
|
|
|
|
|
|
|
| 360 |
# Initialize pipeline WITH text encoder and optional audio support
|
| 361 |
# ---- Replace block (pipeline init) lines 275-281 ----
|
| 362 |
pipeline = LTX23NegativePromptTwoStagePipeline(
|
|
|
|
| 487 |
print("[LoRA] Prepared LoRA config already active; skipping.")
|
| 488 |
return True
|
| 489 |
|
| 490 |
+
del pipeline
|
| 491 |
+
gc.collect()
|
| 492 |
+
torch.cuda.empty_cache()
|
| 493 |
+
|
| 494 |
pipeline = LTX23NegativePromptTwoStagePipeline(
|
| 495 |
checkpoint_path=str(checkpoint_path),
|
| 496 |
spatial_upsampler_path=str(spatial_upsampler_path),
|