diff --git a/ControlNeXt-SVD-v2-Training/models/controlnext_vid_svd.py b/ControlNeXt-SVD-v2-Training/models/controlnext_vid_svd.py index f1fd620..e759aed 100644 --- a/ControlNeXt-SVD-v2-Training/models/controlnext_vid_svd.py +++ b/ControlNeXt-SVD-v2-Training/models/controlnext_vid_svd.py @@ -9,7 +9,7 @@ from diffusers.models.resnet import Downsample2D, ResnetBlock2D -class ControlNeXtSDVModel(ModelMixin, ConfigMixin): +class ControlNeXtSVDModel(ModelMixin, ConfigMixin): _supports_gradient_checkpointing = True @register_to_config diff --git a/ControlNeXt-SVD-v2-Training/pipeline/pipeline_stable_video_diffusion_controlnext.py b/ControlNeXt-SVD-v2-Training/pipeline/pipeline_stable_video_diffusion_controlnext.py index d0c1e7c..214d142 100644 --- a/ControlNeXt-SVD-v2-Training/pipeline/pipeline_stable_video_diffusion_controlnext.py +++ b/ControlNeXt-SVD-v2-Training/pipeline/pipeline_stable_video_diffusion_controlnext.py @@ -20,7 +20,7 @@ import PIL.Image import torch from transformers import CLIPImageProcessor, CLIPVisionModelWithProjection -from models.controlnext_vid_svd import ControlNeXtSDVModel +from models.controlnext_vid_svd import ControlNeXtSVDModel from diffusers.image_processor import VaeImageProcessor from diffusers.models import AutoencoderKLTemporalDecoder, UNetSpatioTemporalConditionModel @@ -125,7 +125,7 @@ def __init__( vae: AutoencoderKLTemporalDecoder, image_encoder: CLIPVisionModelWithProjection, unet: UNetSpatioTemporalConditionControlNeXtModel, - controlnext: ControlNeXtSDVModel, + controlnext: ControlNeXtSVDModel, scheduler: EulerDiscreteScheduler, feature_extractor: CLIPImageProcessor, ): diff --git a/ControlNeXt-SVD-v2-Training/train_svd.py b/ControlNeXt-SVD-v2-Training/train_svd.py index 2ba806d..5345cf5 100644 --- a/ControlNeXt-SVD-v2-Training/train_svd.py +++ b/ControlNeXt-SVD-v2-Training/train_svd.py @@ -59,7 +59,7 @@ from utils.vid_dataset import UBCFashion from models.unet_spatio_temporal_condition_controlnext import UNetSpatioTemporalConditionControlNeXtModel from pipeline.pipeline_stable_video_diffusion_controlnext import StableVideoDiffusionPipelineControlNeXt -from models.controlnext_vid_svd import ControlNeXtSDVModel +from models.controlnext_vid_svd import ControlNeXtSVDModel import torch.nn as nn import pdb from diffusers.utils.torch_utils import randn_tensor @@ -920,7 +920,7 @@ def main(): ) logger.info("Initializing controlnext weights from unet") - controlnext = ControlNeXtSDVModel() + controlnext = ControlNeXtSVDModel() if args.controlnet_model_name_or_path: logger.info("Loading existing controlnet weights") diff --git a/ControlNeXt-SVD-v2/models/controlnext_vid_svd.py b/ControlNeXt-SVD-v2/models/controlnext_vid_svd.py index c390277..663d79f 100644 --- a/ControlNeXt-SVD-v2/models/controlnext_vid_svd.py +++ b/ControlNeXt-SVD-v2/models/controlnext_vid_svd.py @@ -9,7 +9,7 @@ from diffusers.models.resnet import Downsample2D, ResnetBlock2D -class ControlNeXtSDVModel(ModelMixin, ConfigMixin): +class ControlNeXtSVDModel(ModelMixin, ConfigMixin): _supports_gradient_checkpointing = True @register_to_config diff --git a/ControlNeXt-SVD-v2/pipeline/pipeline_stable_video_diffusion_controlnext.py b/ControlNeXt-SVD-v2/pipeline/pipeline_stable_video_diffusion_controlnext.py index 03466e2..e45711a 100644 --- a/ControlNeXt-SVD-v2/pipeline/pipeline_stable_video_diffusion_controlnext.py +++ b/ControlNeXt-SVD-v2/pipeline/pipeline_stable_video_diffusion_controlnext.py @@ -20,7 +20,7 @@ import PIL.Image import torch from transformers import CLIPImageProcessor, CLIPVisionModelWithProjection -from models.controlnext_vid_svd import ControlNeXtSDVModel +from models.controlnext_vid_svd import ControlNeXtSVDModel from diffusers.image_processor import VaeImageProcessor from diffusers.models import AutoencoderKLTemporalDecoder, UNetSpatioTemporalConditionModel @@ -125,7 +125,7 @@ def __init__( vae: AutoencoderKLTemporalDecoder, image_encoder: CLIPVisionModelWithProjection, unet: UNetSpatioTemporalConditionControlNeXtModel, - controlnext: ControlNeXtSDVModel, + controlnext: ControlNeXtSVDModel, scheduler: EulerDiscreteScheduler, feature_extractor: CLIPImageProcessor, ): diff --git a/ControlNeXt-SVD-v2/run_controlnext.py b/ControlNeXt-SVD-v2/run_controlnext.py index 3f6a8f2..a41a828 100644 --- a/ControlNeXt-SVD-v2/run_controlnext.py +++ b/ControlNeXt-SVD-v2/run_controlnext.py @@ -3,7 +3,7 @@ import numpy as np from PIL import Image from pipeline.pipeline_stable_video_diffusion_controlnext import StableVideoDiffusionPipelineControlNeXt -from models.controlnext_vid_svd import ControlNeXtSDVModel +from models.controlnext_vid_svd import ControlNeXtSVDModel from models.unet_spatio_temporal_condition_controlnext import UNetSpatioTemporalConditionControlNeXtModel from transformers import CLIPVisionModelWithProjection import re @@ -221,7 +221,7 @@ def load_tensor(tensor_path): subfolder="unet", low_cpu_mem_usage=True, ) - controlnext = ControlNeXtSDVModel() + controlnext = ControlNeXtSVDModel() controlnext.load_state_dict(load_tensor(args.controlnext_path)) unet.load_state_dict(load_tensor(args.unet_path), strict=False) @@ -279,4 +279,4 @@ def load_tensor(tensor_path): final_result, validation_control_images[:num_frames], args.output_dir, - fps=fps) \ No newline at end of file + fps=fps) diff --git a/ControlNeXt-SVD/models/controlnext_vid_svd.py b/ControlNeXt-SVD/models/controlnext_vid_svd.py index 6b72e64..e915d8b 100644 --- a/ControlNeXt-SVD/models/controlnext_vid_svd.py +++ b/ControlNeXt-SVD/models/controlnext_vid_svd.py @@ -337,7 +337,7 @@ def forward( -class ControlNeXtSDVModel(ModelMixin, ConfigMixin): +class ControlNeXtSVDModel(ModelMixin, ConfigMixin): _supports_gradient_checkpointing = True diff --git a/ControlNeXt-SVD/pipeline/pipeline_stable_video_diffusion_controlnext.py b/ControlNeXt-SVD/pipeline/pipeline_stable_video_diffusion_controlnext.py index 01d2bda..5b001a8 100644 --- a/ControlNeXt-SVD/pipeline/pipeline_stable_video_diffusion_controlnext.py +++ b/ControlNeXt-SVD/pipeline/pipeline_stable_video_diffusion_controlnext.py @@ -20,7 +20,7 @@ import PIL.Image import torch from transformers import CLIPImageProcessor, CLIPVisionModelWithProjection -from models.controlnext_vid_svd import ControlNeXtSDVModel +from models.controlnext_vid_svd import ControlNeXtSVDModel from diffusers.image_processor import VaeImageProcessor from diffusers.models import AutoencoderKLTemporalDecoder, UNetSpatioTemporalConditionModel @@ -125,7 +125,7 @@ def __init__( vae: AutoencoderKLTemporalDecoder, image_encoder: CLIPVisionModelWithProjection, unet: UNetSpatioTemporalConditionControlNeXtModel, - controlnext: ControlNeXtSDVModel, + controlnext: ControlNeXtSVDModel, scheduler: EulerDiscreteScheduler, feature_extractor: CLIPImageProcessor, ): diff --git a/ControlNeXt-SVD/run_controlnext.py b/ControlNeXt-SVD/run_controlnext.py index d85dd22..a3bfb0e 100644 --- a/ControlNeXt-SVD/run_controlnext.py +++ b/ControlNeXt-SVD/run_controlnext.py @@ -3,7 +3,7 @@ import numpy as np from PIL import Image from pipeline.pipeline_stable_video_diffusion_controlnext import StableVideoDiffusionPipelineControlNeXt -from models.controlnext_vid_svd import ControlNeXtSDVModel +from models.controlnext_vid_svd import ControlNeXtSVDModel from models.unet_spatio_temporal_condition_controlnext import UNetSpatioTemporalConditionControlNeXtModel from transformers import CLIPVisionModelWithProjection import re @@ -202,7 +202,7 @@ def load_tensor(tensor_path): low_cpu_mem_usage=True, variant="fp16", ) - controlnext = ControlNeXtSDVModel() + controlnext = ControlNeXtSVDModel() controlnext.load_state_dict(load_tensor(args.controlnext_path)) unet.load_state_dict(load_tensor(args.unet_path), strict=False)