|
|
|
|
|
import imp |
|
import numpy as np |
|
import torch |
|
import random |
|
from PIL import Image, ImageDraw, ImageFont |
|
import copy |
|
from typing import Optional, Union, Tuple, List, Callable, Dict, Any |
|
from tqdm.notebook import tqdm |
|
from diffusers.utils import BaseOutput, logging |
|
from diffusers.models.embeddings import TimestepEmbedding, Timesteps |
|
from diffusers.models.unet_2d_blocks import ( |
|
CrossAttnDownBlock2D, |
|
CrossAttnUpBlock2D, |
|
DownBlock2D, |
|
UNetMidBlock2DCrossAttn, |
|
UpBlock2D, |
|
get_down_block, |
|
get_up_block, |
|
) |
|
from diffusers.models.unet_2d_condition import UNet2DConditionOutput |
|
from copy import deepcopy |
|
import json |
|
|
|
import inspect |
|
import os |
|
import warnings |
|
from typing import Any, Callable, Dict, List, Optional, Tuple, Union |
|
|
|
import numpy as np |
|
import PIL.Image |
|
import torch |
|
import torch.nn.functional as F |
|
from transformers import CLIPImageProcessor, CLIPTextModel, CLIPTokenizer |
|
|
|
from diffusers.image_processor import VaeImageProcessor |
|
from diffusers.loaders import LoraLoaderMixin, TextualInversionLoaderMixin |
|
from diffusers.models import AutoencoderKL, ControlNetModel, UNet2DConditionModel |
|
from diffusers.schedulers import KarrasDiffusionSchedulers |
|
from diffusers.utils.torch_utils import is_compiled_module |
|
|
|
from diffusers.pipelines.stable_diffusion import StableDiffusionPipelineOutput |
|
from diffusers.pipelines.stable_diffusion.safety_checker import StableDiffusionSafetyChecker |
|
from diffusers.pipelines.controlnet.multicontrolnet import MultiControlNetModel |
|
from tqdm import tqdm |
|
import time |
|
|
|
from dataclasses import dataclass |
|
from typing import Any, Dict, List, Optional, Tuple, Union |
|
|
|
import torch |
|
import torch.nn as nn |
|
import torch.utils.checkpoint |
|
|
|
from diffusers.models.embeddings import ( |
|
GaussianFourierProjection, |
|
ImageHintTimeEmbedding, |
|
ImageProjection, |
|
ImageTimeEmbedding, |
|
TextImageProjection, |
|
TextImageTimeEmbedding, |
|
TextTimeEmbedding, |
|
TimestepEmbedding, |
|
Timesteps, |
|
) |
|
from diffusers.models.modeling_utils import ModelMixin |
|
from diffusers.models.unet_2d_blocks import ( |
|
UNetMidBlock2D, |
|
UNetMidBlock2DCrossAttn, |
|
UNetMidBlock2DSimpleCrossAttn, |
|
get_down_block, |
|
get_up_block, |
|
) |
|
|
|
from diffusers.image_processor import PipelineImageInput, VaeImageProcessor |
|
from diffusers.loaders import ( |
|
FromSingleFileMixin, |
|
IPAdapterMixin, |
|
StableDiffusionXLLoraLoaderMixin, |
|
TextualInversionLoaderMixin, |
|
) |
|
from diffusers.models import AutoencoderKL, ImageProjection, UNet2DConditionModel |
|
from diffusers.models.attention_processor import ( |
|
AttnProcessor2_0, |
|
FusedAttnProcessor2_0, |
|
XFormersAttnProcessor, |
|
) |
|
from diffusers.models.lora import adjust_lora_scale_text_encoder |
|
from diffusers.schedulers import KarrasDiffusionSchedulers |
|
from diffusers.utils import ( |
|
USE_PEFT_BACKEND, |
|
deprecate, |
|
is_invisible_watermark_available, |
|
is_torch_xla_available, |
|
logging, |
|
replace_example_docstring, |
|
scale_lora_layers, |
|
unscale_lora_layers, |
|
) |
|
from diffusers.utils.torch_utils import randn_tensor |
|
|
|
from diffusers.pipelines.stable_diffusion_xl.pipeline_output import StableDiffusionXLPipelineOutput |
|
|
|
from typing import Any, Dict, List |
|
|
|
from diffusers.configuration_utils import ConfigMixin, register_to_config |
|
from diffusers.utils import CONFIG_NAME |
|
|
|
|
|
class PipelineCallback(ConfigMixin): |
|
|
|
config_name = CONFIG_NAME |
|
|
|
@register_to_config |
|
def __init__(self, cutoff_step_ratio=1.0, cutoff_step_index=None): |
|
super().__init__() |
|
|
|
if (cutoff_step_ratio is None and cutoff_step_index is None) or ( |
|
cutoff_step_ratio is not None and cutoff_step_index is not None |
|
): |
|
raise ValueError("Either cutoff_step_ratio or cutoff_step_index should be provided, not both or none.") |
|
|
|
if cutoff_step_ratio is not None and ( |
|
not isinstance(cutoff_step_ratio, float) or not (0.0 <= cutoff_step_ratio <= 1.0) |
|
): |
|
raise ValueError("cutoff_step_ratio must be a float between 0.0 and 1.0.") |
|
|
|
@property |
|
def tensor_inputs(self) -> List[str]: |
|
raise NotImplementedError(f"You need to set the attribute `tensor_inputs` for {self.__class__}") |
|
|
|
def callback_fn(self, pipeline, step_index, timesteps, callback_kwargs) -> Dict[str, Any]: |
|
raise NotImplementedError(f"You need to implement the method `callback_fn` for {self.__class__}") |
|
|
|
def __call__(self, pipeline, step_index, timestep, callback_kwargs) -> Dict[str, Any]: |
|
return self.callback_fn(pipeline, step_index, timestep, callback_kwargs) |
|
|
|
|
|
class MultiPipelineCallbacks: |
|
|
|
def __init__(self, callbacks: List[PipelineCallback]): |
|
self.callbacks = callbacks |
|
|
|
@property |
|
def tensor_inputs(self) -> List[str]: |
|
return [input for callback in self.callbacks for input in callback.tensor_inputs] |
|
|
|
def __call__(self, pipeline, step_index, timestep, callback_kwargs) -> Dict[str, Any]: |
|
""" |
|
Calls all the callbacks in order with the given arguments and returns the final callback_kwargs. |
|
""" |
|
for callback in self.callbacks: |
|
callback_kwargs = callback(pipeline, step_index, timestep, callback_kwargs) |
|
|
|
return callback_kwargs |
|
|
|
def seed_everything(seed): |
|
torch.manual_seed(seed) |
|
torch.cuda.manual_seed(seed) |
|
random.seed(seed) |
|
np.random.seed(seed) |
|
|
|
def get_promptls(prompt_path): |
|
with open(prompt_path) as f: |
|
prompt_ls = json.load(f) |
|
prompt_ls = [prompt['caption'].replace('/','_') for prompt in prompt_ls] |
|
return prompt_ls |
|
|
|
|
|
|
|
def warpped_feature(sample, step): |
|
""" |
|
sample: batch_size*dim*h*w, uncond: 0 - batch_size//2, cond: batch_size//2 - batch_size |
|
step: timestep span |
|
""" |
|
bs, dim, h, w = sample.shape |
|
|
|
|
|
|
|
|
|
if sample.shape[0] >= 2: |
|
uncond_fea, cond_fea = sample.chunk(2, dim=0) |
|
uncond_fea = uncond_fea.repeat(step,1,1,1) |
|
cond_fea = cond_fea.repeat(step,1,1,1) |
|
else: |
|
uncond_fea = sample.repeat(step,1,1,1) |
|
cond_fea = sample.repeat(step,1,1,1) |
|
print("Warning: sample has only one chunk. Proceeding with only uncond_fea") |
|
|
|
return torch.cat([uncond_fea, cond_fea]) |
|
|
|
def warpped_skip_feature(block_samples, step): |
|
down_block_res_samples = [] |
|
|
|
for sample in block_samples: |
|
sample_expand = warpped_feature(sample, step) |
|
down_block_res_samples.append(sample_expand) |
|
return tuple(down_block_res_samples) |
|
|
|
def warpped_text_emb(text_emb, step): |
|
""" |
|
text_emb: batch_size*77*768, uncond: 0 - batch_size//2, cond: batch_size//2 - batch_size |
|
step: timestep span |
|
""" |
|
bs, token_len, dim = text_emb.shape |
|
if text_emb.shape[0] >= 2: |
|
uncond_fea, cond_fea = text_emb.chunk(2) |
|
uncond_fea = uncond_fea.repeat(step,1,1) |
|
cond_fea = cond_fea.repeat(step,1,1) |
|
else: |
|
uncond_fea, cond_fea = text_emb, text_emb |
|
uncond_fea = uncond_fea.repeat(step,1,1) |
|
cond_fea = cond_fea.repeat(step,1,1) |
|
return torch.cat([uncond_fea, cond_fea]) |
|
|
|
def warpped_timestep(timesteps, bs): |
|
""" |
|
timestpes: list, such as [981, 961, 941] |
|
""" |
|
semi_bs = bs//2 |
|
ts = [] |
|
for timestep in timesteps: |
|
timestep = timestep[None] |
|
texp = timestep.expand(semi_bs) |
|
ts.append(texp) |
|
timesteps = torch.cat(ts) |
|
return timesteps.repeat(2,1).reshape(-1) |
|
|
|
def rescale_noise_cfg(noise_cfg, noise_pred_text, guidance_rescale=0.0): |
|
""" |
|
Rescale `noise_cfg` according to `guidance_rescale`. Based on findings of [Common Diffusion Noise Schedules and |
|
Sample Steps are Flawed](https://arxiv.org/pdf/2305.08891.pdf). See Section 3.4 |
|
""" |
|
std_text = noise_pred_text.std(dim=list(range(1, noise_pred_text.ndim)), keepdim=True) |
|
std_cfg = noise_cfg.std(dim=list(range(1, noise_cfg.ndim)), keepdim=True) |
|
|
|
noise_pred_rescaled = noise_cfg * (std_text / std_cfg) |
|
|
|
noise_cfg = guidance_rescale * noise_pred_rescaled + (1 - guidance_rescale) * noise_cfg |
|
return noise_cfg |
|
|
|
def register_parallel_pipeline_orig(pipe, mod='50ls4'): |
|
def new_call(self): |
|
@torch.no_grad() |
|
def call( |
|
prompt: Union[str, List[str]] = None, |
|
prompt_2: Optional[Union[str, List[str]]] = None, |
|
height: Optional[int] = None, |
|
width: Optional[int] = None, |
|
num_inference_steps: int = 50, |
|
timesteps: List[int] = None, |
|
sigmas: List[float] = None, |
|
denoising_end: Optional[float] = None, |
|
guidance_scale: float = 5.0, |
|
negative_prompt: Optional[Union[str, List[str]]] = None, |
|
negative_prompt_2: Optional[Union[str, List[str]]] = None, |
|
num_images_per_prompt: Optional[int] = 1, |
|
eta: float = 0.0, |
|
generator: Optional[Union[torch.Generator, List[torch.Generator]]] = None, |
|
latents: Optional[torch.Tensor] = None, |
|
prompt_embeds: Optional[torch.Tensor] = None, |
|
negative_prompt_embeds: Optional[torch.Tensor] = None, |
|
pooled_prompt_embeds: Optional[torch.Tensor] = None, |
|
negative_pooled_prompt_embeds: Optional[torch.Tensor] = None, |
|
ip_adapter_image: Optional[PipelineImageInput] = None, |
|
ip_adapter_image_embeds: Optional[List[torch.Tensor]] = None, |
|
output_type: Optional[str] = "pil", |
|
return_dict: bool = True, |
|
cross_attention_kwargs: Optional[Dict[str, Any]] = None, |
|
guidance_rescale: float = 0.0, |
|
original_size: Optional[Tuple[int, int]] = None, |
|
crops_coords_top_left: Tuple[int, int] = (0, 0), |
|
target_size: Optional[Tuple[int, int]] = None, |
|
negative_original_size: Optional[Tuple[int, int]] = None, |
|
negative_crops_coords_top_left: Tuple[int, int] = (0, 0), |
|
negative_target_size: Optional[Tuple[int, int]] = None, |
|
clip_skip: Optional[int] = None, |
|
callback_on_step_end: Optional[ |
|
Union[Callable[[int, int, Dict], None], PipelineCallback, MultiPipelineCallbacks] |
|
] = None, |
|
callback_on_step_end_tensor_inputs: List[str] = ["latents"], |
|
**kwargs, |
|
): |
|
|
|
callback = kwargs.pop("callback", None) |
|
callback_steps = kwargs.pop("callback_steps", None) |
|
|
|
if callback is not None: |
|
deprecate( |
|
"callback", |
|
"1.0.0", |
|
"Passing `callback` as an input argument to `__call__` is deprecated, consider use `callback_on_step_end`", |
|
) |
|
if callback_steps is not None: |
|
deprecate( |
|
"callback_steps", |
|
"1.0.0", |
|
"Passing `callback_steps` as an input argument to `__call__` is deprecated, consider use `callback_on_step_end`", |
|
) |
|
|
|
if isinstance(callback_on_step_end, (PipelineCallback, MultiPipelineCallbacks)): |
|
callback_on_step_end_tensor_inputs = callback_on_step_end.tensor_inputs |
|
|
|
|
|
height = height or self.default_sample_size * self.vae_scale_factor |
|
width = width or self.default_sample_size * self.vae_scale_factor |
|
original_size = original_size or (height, width) |
|
target_size = target_size or (height, width) |
|
|
|
|
|
self.check_inputs( |
|
prompt, |
|
prompt_2, |
|
height, |
|
width, |
|
callback_steps, |
|
negative_prompt, |
|
negative_prompt_2, |
|
prompt_embeds, |
|
negative_prompt_embeds, |
|
pooled_prompt_embeds, |
|
negative_pooled_prompt_embeds, |
|
ip_adapter_image, |
|
ip_adapter_image_embeds, |
|
callback_on_step_end_tensor_inputs, |
|
) |
|
|
|
self._guidance_scale = guidance_scale |
|
self._guidance_rescale = guidance_rescale |
|
self._clip_skip = clip_skip |
|
self._cross_attention_kwargs = cross_attention_kwargs |
|
self._denoising_end = denoising_end |
|
self._interrupt = False |
|
|
|
|
|
if prompt is not None and isinstance(prompt, str): |
|
batch_size = 1 |
|
elif prompt is not None and isinstance(prompt, list): |
|
batch_size = len(prompt) |
|
else: |
|
batch_size = prompt_embeds.shape[0] |
|
|
|
device = self._execution_device |
|
|
|
|
|
lora_scale = ( |
|
self.cross_attention_kwargs.get("scale", None) if self.cross_attention_kwargs is not None else None |
|
) |
|
|
|
( |
|
prompt_embeds, |
|
negative_prompt_embeds, |
|
pooled_prompt_embeds, |
|
negative_pooled_prompt_embeds, |
|
) = self.encode_prompt( |
|
prompt=prompt, |
|
prompt_2=prompt_2, |
|
device=device, |
|
num_images_per_prompt=num_images_per_prompt, |
|
do_classifier_free_guidance=self.do_classifier_free_guidance, |
|
negative_prompt=negative_prompt, |
|
negative_prompt_2=negative_prompt_2, |
|
prompt_embeds=prompt_embeds, |
|
negative_prompt_embeds=negative_prompt_embeds, |
|
pooled_prompt_embeds=pooled_prompt_embeds, |
|
negative_pooled_prompt_embeds=negative_pooled_prompt_embeds, |
|
lora_scale=lora_scale, |
|
clip_skip=self.clip_skip, |
|
) |
|
|
|
if self.do_classifier_free_guidance: |
|
prompt_embeds = torch.cat([negative_prompt_embeds, prompt_embeds], dim=0) |
|
pooled_prompt_embeds = torch.cat([negative_pooled_prompt_embeds, pooled_prompt_embeds], dim=0) |
|
|
|
|
|
|
|
self.scheduler.set_timesteps(num_inference_steps, device=device) |
|
timesteps = self.scheduler.timesteps |
|
|
|
|
|
num_channels_latents = self.unet.config.in_channels |
|
latents = self.prepare_latents( |
|
batch_size * num_images_per_prompt, |
|
num_channels_latents, |
|
height, |
|
width, |
|
prompt_embeds.dtype, |
|
device, |
|
generator, |
|
latents, |
|
) |
|
|
|
|
|
extra_step_kwargs = self.prepare_extra_step_kwargs(generator, eta) |
|
|
|
|
|
timestep_cond = None |
|
if self.unet.config.time_cond_proj_dim is not None: |
|
guidance_scale_tensor = torch.tensor(self.guidance_scale - 1).repeat(batch_size * num_images_per_prompt) |
|
timestep_cond = self.get_guidance_scale_embedding( |
|
guidance_scale_tensor, embedding_dim=self.unet.config.time_cond_proj_dim |
|
).to(device=device, dtype=latents.dtype) |
|
|
|
|
|
init_latents = latents.detach().clone() |
|
all_steps = len(self.scheduler.timesteps) |
|
curr_step = 0 |
|
if mod == '50ls': |
|
cond = lambda timestep: timestep in [0, 1, 2, 3, 5, 10, 15, 25, 35] |
|
elif isinstance(mod, int): |
|
cond = lambda timestep: timestep % mod == 0 |
|
else: |
|
raise Exception("Currently not supported, But you can modify the code to customize the keytime") |
|
|
|
|
|
add_time_ids = self._get_add_time_ids( |
|
original_size, |
|
crops_coords_top_left, |
|
target_size, |
|
dtype=prompt_embeds.dtype, |
|
text_encoder_projection_dim=pooled_prompt_embeds.shape[-1] if self.text_encoder_2 is None else self.text_encoder_2.config.projection_dim, |
|
) |
|
if negative_original_size is not None and negative_target_size is not None: |
|
negative_add_time_ids = self._get_add_time_ids( |
|
negative_original_size, |
|
negative_crops_coords_top_left, |
|
negative_target_size, |
|
dtype=prompt_embeds.dtype, |
|
text_encoder_projection_dim=pooled_prompt_embeds.shape[-1] if self.text_encoder_2 is None else self.text_encoder_2.config.projection_dim, |
|
) |
|
else: |
|
negative_add_time_ids = add_time_ids |
|
|
|
if self.do_classifier_free_guidance: |
|
add_time_ids = torch.cat([negative_add_time_ids, add_time_ids], dim=0) |
|
|
|
add_text_embeds = pooled_prompt_embeds.to(device) |
|
add_time_ids = add_time_ids.to(device).repeat(batch_size * num_images_per_prompt, 1) |
|
|
|
|
|
while curr_step < all_steps: |
|
register_time(self.unet, curr_step) |
|
time_ls = [timesteps[curr_step]] |
|
curr_step += 1 |
|
while not cond(curr_step) and curr_step < all_steps: |
|
time_ls.append(timesteps[curr_step]) |
|
curr_step += 1 |
|
|
|
latent_model_input = torch.cat([latents] * 2) if self.do_classifier_free_guidance else latents |
|
|
|
|
|
added_cond_kwargs = {"text_embeds": add_text_embeds, "time_ids": add_time_ids} |
|
if ip_adapter_image is not None or ip_adapter_image_embeds is not None: |
|
image_embeds = self.prepare_ip_adapter_image_embeds( |
|
ip_adapter_image, |
|
ip_adapter_image_embeds, |
|
device, |
|
batch_size * num_images_per_prompt, |
|
self.do_classifier_free_guidance, |
|
) |
|
added_cond_kwargs["image_embeds"] = image_embeds |
|
|
|
noise_pred = self.unet( |
|
latent_model_input, |
|
time_ls, |
|
encoder_hidden_states=prompt_embeds, |
|
timestep_cond=timestep_cond, |
|
cross_attention_kwargs=self.cross_attention_kwargs, |
|
added_cond_kwargs=added_cond_kwargs, |
|
return_dict=False, |
|
)[0] |
|
|
|
if self.do_classifier_free_guidance: |
|
noise_pred_uncond, noise_pred_text = noise_pred.chunk(2) |
|
noise_pred = noise_pred_uncond + self.guidance_scale * (noise_pred_text - noise_pred_uncond) |
|
|
|
if self.do_classifier_free_guidance and self.guidance_rescale > 0.0: |
|
noise_pred = rescale_noise_cfg(noise_pred, noise_pred_text, guidance_rescale=self.guidance_rescale) |
|
|
|
bs = noise_pred.shape[0] |
|
bs_perstep = bs // len(time_ls) |
|
denoised_latent = latents |
|
for i, timestep in enumerate(time_ls): |
|
if timestep / 1000 < 0.5: |
|
denoised_latent = denoised_latent + 0.003 * init_latents |
|
curr_noise = noise_pred[i * bs_perstep : (i + 1) * bs_perstep] |
|
denoised_latent = self.scheduler.step( |
|
curr_noise, timestep, denoised_latent, **extra_step_kwargs, return_dict=False |
|
)[0] |
|
latents = denoised_latent |
|
|
|
|
|
if not output_type == "latent": |
|
needs_upcasting = self.vae.dtype == torch.float16 and self.vae.config.force_upcast |
|
if needs_upcasting: |
|
self.upcast_vae() |
|
latents = latents.to(next(iter(self.vae.post_quant_conv.parameters())).dtype) |
|
elif latents.dtype != self.vae.dtype: |
|
if torch.backends.mps.is_available(): |
|
self.vae = self.vae.to(latents.dtype) |
|
|
|
has_latents_mean = hasattr(self.vae.config, "latents_mean") and self.vae.config.latents_mean is not None |
|
has_latents_std = hasattr(self.vae.config, "latents_std") and self.vae.config.latents_std is not None |
|
if has_latents_mean and has_latents_std: |
|
latents_mean = ( |
|
torch.tensor(self.vae.config.latents_mean).view(1, 4, 1, 1).to(latents.device, latents.dtype) |
|
) |
|
latents_std = ( |
|
torch.tensor(self.vae.config.latents_std).view(1, 4, 1, 1).to(latents.device, latents.dtype) |
|
) |
|
latents = latents * latents_std / self.vae.config.scaling_factor + latents_mean |
|
else: |
|
latents = latents / self.vae.config.scaling_factor |
|
|
|
image = self.vae.decode(latents, return_dict=False)[0] |
|
|
|
if needs_upcasting: |
|
self.vae.to(dtype=torch.float16) |
|
else: |
|
image = latents |
|
|
|
if not output_type == "latent": |
|
if self.watermark is not None: |
|
image = self.watermark.apply_watermark(image) |
|
|
|
image = self.image_processor.postprocess(image, output_type=output_type) |
|
|
|
self.maybe_free_model_hooks() |
|
|
|
if not return_dict: |
|
return (image,) |
|
|
|
return StableDiffusionXLPipelineOutput(images=image) |
|
|
|
return call |
|
|
|
pipe.call = new_call(pipe) |
|
|
|
def register_faster_forward(model, mod = '50ls4'): |
|
def faster_forward(self): |
|
def forward( |
|
sample: torch.FloatTensor, |
|
timestep: Union[torch.Tensor, float, int], |
|
encoder_hidden_states: torch.Tensor, |
|
class_labels: Optional[torch.Tensor] = None, |
|
timestep_cond: Optional[torch.Tensor] = None, |
|
attention_mask: Optional[torch.Tensor] = None, |
|
cross_attention_kwargs: Optional[Dict[str, Any]] = None, |
|
added_cond_kwargs: Optional[Dict[str, torch.Tensor]] = None, |
|
down_block_additional_residuals: Optional[Tuple[torch.Tensor]] = None, |
|
mid_block_additional_residual: Optional[torch.Tensor] = None, |
|
down_intrablock_additional_residuals: Optional[Tuple[torch.Tensor]] = None, |
|
return_dict: bool = True, |
|
) -> Union[UNet2DConditionOutput, Tuple]: |
|
r""" |
|
Args: |
|
sample (`torch.FloatTensor`): (batch, channel, height, width) noisy inputs tensor |
|
timestep (`torch.FloatTensor` or `float` or `int`): (batch) timesteps |
|
encoder_hidden_states (`torch.FloatTensor`): (batch, sequence_length, feature_dim) encoder hidden states |
|
return_dict (`bool`, *optional*, defaults to `True`): |
|
Whether or not to return a [`models.unet_2d_condition.UNet2DConditionOutput`] instead of a plain tuple. |
|
cross_attention_kwargs (`dict`, *optional*): |
|
A kwargs dictionary that if specified is passed along to the `AttentionProcessor` as defined under |
|
`self.processor` in |
|
[diffusers.cross_attention](https://github.com/huggingface/diffusers/blob/main/src/diffusers/models/cross_attention.py). |
|
|
|
Returns: |
|
[`~models.unet_2d_condition.UNet2DConditionOutput`] or `tuple`: |
|
[`~models.unet_2d_condition.UNet2DConditionOutput`] if `return_dict` is True, otherwise a `tuple`. When |
|
returning a tuple, the first element is the sample tensor. |
|
""" |
|
|
|
|
|
|
|
|
|
default_overall_up_factor = 2**self.num_upsamplers |
|
|
|
|
|
forward_upsample_size = False |
|
upsample_size = None |
|
|
|
if any(s % default_overall_up_factor != 0 for s in sample.shape[-2:]): |
|
print("Forward upsample size to force interpolation output size.") |
|
forward_upsample_size = True |
|
|
|
|
|
if attention_mask is not None: |
|
attention_mask = (1 - attention_mask.to(sample.dtype)) * -10000.0 |
|
attention_mask = attention_mask.unsqueeze(1) |
|
|
|
|
|
if self.config.center_input_sample: |
|
sample = 2 * sample - 1.0 |
|
|
|
|
|
step = 1 |
|
|
|
|
|
|
|
|
|
|
|
|
|
if not torch.is_tensor(timesteps) and (not isinstance(timesteps,list)): |
|
|
|
|
|
is_mps = sample.device.type == "mps" |
|
if isinstance(timestep, float): |
|
dtype = torch.float32 if is_mps else torch.float64 |
|
else: |
|
dtype = torch.int32 if is_mps else torch.int64 |
|
timesteps = torch.tensor([timesteps], dtype=dtype, device=sample.device) |
|
elif (not isinstance(timesteps,list)) and len(timesteps.shape) == 0: |
|
timesteps = timesteps[None].to(sample.device) |
|
|
|
if (not isinstance(timesteps,list)) and len(timesteps.shape) == 1: |
|
|
|
timesteps = timesteps.expand(sample.shape[0]) |
|
elif isinstance(timesteps, list): |
|
|
|
timesteps = warpped_timestep(timesteps, sample.shape[0]).to(sample.device) |
|
t_emb = self.time_proj(timesteps) |
|
|
|
|
|
|
|
|
|
t_emb = t_emb.to(dtype=self.dtype) |
|
|
|
emb = self.time_embedding(t_emb, timestep_cond) |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
if self.class_embedding is not None: |
|
if class_labels is None: |
|
raise ValueError("class_labels should be provided when num_class_embeds > 0") |
|
|
|
if self.config.class_embed_type == "timestep": |
|
class_labels = self.time_proj(class_labels) |
|
|
|
|
|
|
|
class_labels = class_labels.to(dtype=sample.dtype) |
|
|
|
class_emb = self.class_embedding(class_labels).to(dtype=self.dtype) |
|
|
|
if self.config.class_embeddings_concat: |
|
emb = torch.cat([emb, class_emb], dim=-1) |
|
else: |
|
emb = emb + class_emb |
|
|
|
if self.config.addition_embed_type == "text": |
|
aug_emb = self.add_embedding(encoder_hidden_states) |
|
emb = emb + aug_emb |
|
|
|
|
|
|
|
if self.time_embed_act is not None: |
|
emb = self.time_embed_act(emb) |
|
|
|
if self.encoder_hid_proj is not None: |
|
print("Encoder hid projections are not none") |
|
encoder_hidden_states = self.encoder_hid_proj(encoder_hidden_states) |
|
|
|
|
|
order = self.order |
|
|
|
ipow = int(np.sqrt(9 + 8*order)) |
|
cond = order in [0, 1, 2, 3, 5, 10, 15, 25, 35] |
|
if isinstance(mod, int): |
|
cond = order % mod == 0 |
|
elif mod == "pro": |
|
cond = ipow * ipow == (9 + 8 * order) |
|
elif mod == "50ls": |
|
cond = order in [0, 1, 2, 3, 5, 10, 15, 25, 35] |
|
elif mod == "50ls2": |
|
cond = order in [0, 10, 11, 12, 15, 20, 25, 30,35,45] |
|
elif mod == "50ls3": |
|
cond = order in [0, 20, 25, 30,35,45,46,47,48,49] |
|
elif mod == "50ls4": |
|
cond = order in [0, 9, 13, 14, 15, 28, 29, 32, 36,45] |
|
elif mod == "100ls": |
|
cond = order > 85 or order < 10 or order % 5 == 0 |
|
elif mod == "75ls": |
|
cond = order > 65 or order < 10 or order % 5 == 0 |
|
elif mod == "s2": |
|
cond = order < 20 or order > 40 or order % 2 == 0 |
|
|
|
if cond: |
|
|
|
|
|
sample = self.conv_in(sample) |
|
|
|
|
|
down_block_res_samples = (sample,) |
|
print(f"Shape of sample at the start of downsample: {sample.shape}") |
|
for i, downsample_block in enumerate(self.down_blocks): |
|
print(f"\n--- Downsample Block {i+1} ---") |
|
print(f" Input shape: {sample.shape}") |
|
if hasattr(downsample_block, "has_cross_attention") and downsample_block.has_cross_attention: |
|
print("has cross attention") |
|
additional_residuals = {} |
|
if down_intrablock_additional_residuals is not None and len(down_intrablock_additional_residuals) > 0: |
|
additional_residuals["additional_residuals"] = down_intrablock_additional_residuals.pop(0) |
|
print(f" Additional Residuals shape: {additional_residuals['additional_residuals'].shape}") |
|
|
|
sample, res_samples = downsample_block( |
|
hidden_states=sample, |
|
temb=emb, |
|
encoder_hidden_states=encoder_hidden_states, |
|
attention_mask=attention_mask, |
|
cross_attention_kwargs=cross_attention_kwargs, |
|
**additional_residuals |
|
) |
|
else: |
|
sample, res_samples = downsample_block(hidden_states=sample, temb=emb) |
|
|
|
|
|
print(f" Output shape: {sample.shape}") |
|
for j, res_sample in enumerate(res_samples): |
|
print(f" Res sample {j+1} shape: {res_sample.shape}") |
|
down_block_res_samples += res_samples |
|
|
|
|
|
|
|
if down_block_additional_residuals is not None: |
|
new_down_block_res_samples = () |
|
|
|
for down_block_res_sample, down_block_additional_residual in zip( |
|
down_block_res_samples, down_block_additional_residuals |
|
): |
|
down_block_res_sample = down_block_res_sample + down_block_additional_residual |
|
new_down_block_res_samples += (down_block_res_sample,) |
|
|
|
down_block_res_samples = new_down_block_res_samples |
|
|
|
|
|
if down_block_additional_residuals is not None: |
|
new_down_block_res_samples = () |
|
for down_block_res_sample, down_block_additional_residual in zip( |
|
down_block_res_samples, down_block_additional_residuals |
|
): |
|
down_block_res_sample = down_block_res_sample + down_block_additional_residual |
|
new_down_block_res_samples += (down_block_res_sample,) |
|
down_block_res_samples = new_down_block_res_samples |
|
|
|
|
|
if self.mid_block is not None: |
|
if hasattr(self.mid_block, "has_cross_attention") and self.mid_block.has_cross_attention: |
|
sample = self.mid_block( |
|
sample, |
|
emb, |
|
encoder_hidden_states=encoder_hidden_states, |
|
attention_mask=attention_mask, |
|
cross_attention_kwargs=cross_attention_kwargs, |
|
) |
|
else: |
|
sample = self.mid_block(sample, emb) |
|
|
|
if down_intrablock_additional_residuals is not None and len(down_intrablock_additional_residuals) > 0 and sample.shape == down_intrablock_additional_residuals[0].shape: |
|
sample += down_intrablock_additional_residuals.pop(0) |
|
|
|
if mid_block_additional_residual is not None: |
|
sample = sample + mid_block_additional_residual |
|
|
|
|
|
|
|
setattr(self, 'skip_feature', deepcopy(down_block_res_samples)) |
|
setattr(self, 'toup_feature', sample.detach().clone()) |
|
|
|
|
|
|
|
|
|
|
|
if isinstance(timestep, list): |
|
|
|
timesteps = warpped_timestep(timestep, sample.shape[0]).to(sample.device) |
|
t_emb = self.time_proj(timesteps) |
|
|
|
|
|
|
|
|
|
t_emb = t_emb.to(dtype=self.dtype) |
|
|
|
emb = self.time_embedding(t_emb, timestep_cond) |
|
|
|
down_block_res_samples = warpped_skip_feature(down_block_res_samples, step) |
|
sample = warpped_feature(sample, step) |
|
encoder_hidden_states = warpped_text_emb(encoder_hidden_states, step) |
|
|
|
|
|
else: |
|
down_block_res_samples = self.skip_feature |
|
sample = self.toup_feature |
|
|
|
|
|
down_block_res_samples = warpped_skip_feature(down_block_res_samples, step) |
|
sample = warpped_feature(sample, step) |
|
encoder_hidden_states = warpped_text_emb(encoder_hidden_states, step) |
|
|
|
|
|
|
|
for i, upsample_block in enumerate(self.up_blocks): |
|
is_final_block = i == len(self.up_blocks) - 1 |
|
|
|
res_samples = down_block_res_samples[-len(upsample_block.resnets) :] |
|
down_block_res_samples = down_block_res_samples[: -len(upsample_block.resnets)] |
|
|
|
|
|
|
|
if not is_final_block and forward_upsample_size: |
|
upsample_size = down_block_res_samples[-1].shape[2:] |
|
|
|
if hasattr(upsample_block, "has_cross_attention") and upsample_block.has_cross_attention: |
|
sample = upsample_block( |
|
hidden_states=sample, |
|
temb=emb, |
|
res_hidden_states_tuple=res_samples, |
|
encoder_hidden_states=encoder_hidden_states, |
|
cross_attention_kwargs=cross_attention_kwargs, |
|
upsample_size=upsample_size, |
|
attention_mask=attention_mask, |
|
) |
|
else: |
|
sample = upsample_block( |
|
hidden_states=sample, temb=emb, res_hidden_states_tuple=res_samples, upsample_size=upsample_size |
|
) |
|
|
|
|
|
if self.conv_norm_out: |
|
sample = self.conv_norm_out(sample) |
|
sample = self.conv_act(sample) |
|
sample = self.conv_out(sample) |
|
|
|
if not return_dict: |
|
return (sample,) |
|
|
|
return UNet2DConditionOutput(sample=sample) |
|
return forward |
|
if model.__class__.__name__ == 'UNet2DConditionModel': |
|
model.forward = faster_forward(model) |
|
|
|
def register_faster_orig_forward(model, mod = '50ls4'): |
|
def faster_orig_forward(self): |
|
def forward( |
|
sample: torch.FloatTensor, |
|
timestep: Union[torch.Tensor, float, int], |
|
encoder_hidden_states: torch.Tensor, |
|
class_labels: Optional[torch.Tensor] = None, |
|
timestep_cond: Optional[torch.Tensor] = None, |
|
attention_mask: Optional[torch.Tensor] = None, |
|
cross_attention_kwargs: Optional[Dict[str, Any]] = None, |
|
added_cond_kwargs: Optional[Dict[str, torch.Tensor]] = None, |
|
down_block_additional_residuals: Optional[Tuple[torch.Tensor]] = None, |
|
mid_block_additional_residual: Optional[torch.Tensor] = None, |
|
down_intrablock_additional_residuals: Optional[Tuple[torch.Tensor]] = None, |
|
encoder_attention_mask: Optional[torch.Tensor] = None, |
|
return_dict: bool = True, |
|
) -> Union[UNet2DConditionOutput, Tuple]: |
|
|
|
default_overall_up_factor = 2**self.num_upsamplers |
|
|
|
|
|
forward_upsample_size = False |
|
upsample_size = None |
|
|
|
for dim in sample.shape[-2:]: |
|
if dim % default_overall_up_factor != 0: |
|
|
|
forward_upsample_size = True |
|
break |
|
|
|
if attention_mask is not None: |
|
attention_mask = (1 - attention_mask.to(sample.dtype)) * -10000.0 |
|
attention_mask = attention_mask.unsqueeze(1) |
|
|
|
|
|
if encoder_attention_mask is not None: |
|
encoder_attention_mask = (1 - encoder_attention_mask.to(sample.dtype)) * -10000.0 |
|
encoder_attention_mask = encoder_attention_mask.unsqueeze(1) |
|
|
|
|
|
if self.config.center_input_sample: |
|
sample = 2 * sample - 1.0 |
|
|
|
|
|
timesteps = timestep |
|
if not torch.is_tensor(timesteps): |
|
is_mps = sample.device.type == "mps" |
|
if isinstance(timestep, float): |
|
dtype = torch.float32 if is_mps else torch.float64 |
|
else: |
|
dtype = torch.int32 if is_mps else torch.int64 |
|
timesteps = torch.tensor([timesteps], dtype=dtype, device=sample.device) |
|
elif len(timesteps.shape) == 0: |
|
timesteps = timesteps[None].to(sample.device) |
|
|
|
|
|
timesteps = timesteps.expand(sample.shape[0]) |
|
|
|
t_emb = self.time_proj(timesteps) |
|
|
|
t_emb = t_emb.to(dtype=sample.dtype) |
|
|
|
emb = self.time_embedding(t_emb, timestep_cond) |
|
aug_emb = None |
|
|
|
if self.class_embedding is not None: |
|
if class_labels is None: |
|
raise ValueError("class_labels should be provided when num_class_embeds > 0") |
|
|
|
if self.config.class_embed_type == "timestep": |
|
class_labels = self.time_proj(class_labels) |
|
class_labels = class_labels.to(dtype=sample.dtype) |
|
|
|
class_emb = self.class_embedding(class_labels).to(dtype=sample.dtype) |
|
|
|
if self.config.class_embeddings_concat: |
|
emb = torch.cat([emb, class_emb], dim=-1) |
|
else: |
|
emb = emb + class_emb |
|
|
|
if self.config.addition_embed_type == "text": |
|
aug_emb = self.add_embedding(encoder_hidden_states) |
|
elif self.config.addition_embed_type == "text_image": |
|
|
|
if "image_embeds" not in added_cond_kwargs: |
|
raise ValueError( |
|
f"{self.__class__} has the config param `addition_embed_type` set to 'text_image' which requires the keyword argument `image_embeds` to be passed in `added_cond_kwargs`" |
|
) |
|
|
|
image_embs = added_cond_kwargs.get("image_embeds") |
|
text_embs = added_cond_kwargs.get("text_embeds", encoder_hidden_states) |
|
aug_emb = self.add_embedding(text_embs, image_embs) |
|
elif self.config.addition_embed_type == "text_time": |
|
|
|
if "text_embeds" not in added_cond_kwargs: |
|
raise ValueError( |
|
f"{self.__class__} has the config param `addition_embed_type` set to 'text_time' which requires the keyword argument `text_embeds` to be passed in `added_cond_kwargs`" |
|
) |
|
text_embeds = added_cond_kwargs.get("text_embeds") |
|
if "time_ids" not in added_cond_kwargs: |
|
raise ValueError( |
|
f"{self.__class__} has the config param `addition_embed_type` set to 'text_time' which requires the keyword argument `time_ids` to be passed in `added_cond_kwargs`" |
|
) |
|
time_ids = added_cond_kwargs.get("time_ids") |
|
time_embeds = self.add_time_proj(time_ids.flatten()) |
|
time_embeds = time_embeds.reshape((text_embeds.shape[0], -1)) |
|
add_embeds = torch.concat([text_embeds, time_embeds], dim=-1) |
|
add_embeds = add_embeds.to(emb.dtype) |
|
aug_emb = self.add_embedding(add_embeds) |
|
elif self.config.addition_embed_type == "image": |
|
|
|
if "image_embeds" not in added_cond_kwargs: |
|
raise ValueError( |
|
f"{self.__class__} has the config param `addition_embed_type` set to 'image' which requires the keyword argument `image_embeds` to be passed in `added_cond_kwargs`" |
|
) |
|
image_embs = added_cond_kwargs.get("image_embeds") |
|
aug_emb = self.add_embedding(image_embs) |
|
elif self.config.addition_embed_type == "image_hint": |
|
|
|
if "image_embeds" not in added_cond_kwargs or "hint" not in added_cond_kwargs: |
|
raise ValueError( |
|
f"{self.__class__} has the config param `addition_embed_type` set to 'image_hint' which requires the keyword arguments `image_embeds` and `hint` to be passed in `added_cond_kwargs`" |
|
) |
|
image_embs = added_cond_kwargs.get("image_embeds") |
|
hint = added_cond_kwargs.get("hint") |
|
aug_emb, hint = self.add_embedding(image_embs, hint) |
|
sample = torch.cat([sample, hint], dim=1) |
|
|
|
emb = emb + aug_emb if aug_emb is not None else emb |
|
|
|
if self.time_embed_act is not None: |
|
emb = self.time_embed_act(emb) |
|
|
|
if self.encoder_hid_proj is not None and self.config.encoder_hid_dim_type == "text_proj": |
|
encoder_hidden_states = self.encoder_hid_proj(encoder_hidden_states) |
|
elif self.encoder_hid_proj is not None and self.config.encoder_hid_dim_type == "text_image_proj": |
|
|
|
if "image_embeds" not in added_cond_kwargs: |
|
raise ValueError( |
|
f"{self.__class__} has the config param `encoder_hid_dim_type` set to 'text_image_proj' which requires the keyword argument `image_embeds` to be passed in `added_conditions`" |
|
) |
|
|
|
image_embeds = added_cond_kwargs.get("image_embeds") |
|
encoder_hidden_states = self.encoder_hid_proj(encoder_hidden_states, image_embeds) |
|
elif self.encoder_hid_proj is not None and self.config.encoder_hid_dim_type == "image_proj": |
|
|
|
if "image_embeds" not in added_cond_kwargs: |
|
raise ValueError( |
|
f"{self.__class__} has the config param `encoder_hid_dim_type` set to 'image_proj' which requires the keyword argument `image_embeds` to be passed in `added_conditions`" |
|
) |
|
image_embeds = added_cond_kwargs.get("image_embeds") |
|
encoder_hidden_states = self.encoder_hid_proj(image_embeds) |
|
elif self.encoder_hid_proj is not None and self.config.encoder_hid_dim_type == "ip_image_proj": |
|
if "image_embeds" not in added_cond_kwargs: |
|
raise ValueError( |
|
f"{self.__class__} has the config param `encoder_hid_dim_type` set to 'ip_image_proj' which requires the keyword argument `image_embeds` to be passed in `added_conditions`" |
|
) |
|
image_embeds = added_cond_kwargs.get("image_embeds") |
|
image_embeds = self.encoder_hid_proj(image_embeds).to(encoder_hidden_states.dtype) |
|
encoder_hidden_states = torch.cat([encoder_hidden_states, image_embeds], dim=1) |
|
|
|
|
|
order = model.order |
|
|
|
ipow = int(np.sqrt(9 + 8*order)) |
|
cond = order in [0, 1, 2, 3, 5, 10, 15, 25, 35] |
|
if isinstance(mod, int): |
|
cond = order % mod == 0 |
|
elif mod == "pro": |
|
cond = ipow * ipow == (9 + 8 * order) |
|
elif mod == "50ls": |
|
cond = order in [0, 1, 2, 3, 5, 10, 15, 25, 35] |
|
elif mod == "50ls2": |
|
cond = order in [0, 10, 11, 12, 15, 20, 25, 30,35,45] |
|
elif mod == "50ls3": |
|
cond = order in [0, 20, 25, 30,35,45,46,47,48,49] |
|
elif mod == "50ls4": |
|
cond = order in [0, 9, 13, 14, 15, 28, 29, 32, 36,45] |
|
elif mod == "100ls": |
|
cond = order > 85 or order < 10 or order % 5 == 0 |
|
elif mod == "75ls": |
|
cond = order > 65 or order < 10 or order % 5 == 0 |
|
elif mod == "s2": |
|
cond = order < 20 or order > 40 or order % 2 == 0 |
|
|
|
if cond: |
|
|
|
sample = self.conv_in(sample) |
|
|
|
|
|
if cross_attention_kwargs is not None and cross_attention_kwargs.get("gligen", None) is not None: |
|
cross_attention_kwargs = cross_attention_kwargs.copy() |
|
gligen_args = cross_attention_kwargs.pop("gligen") |
|
cross_attention_kwargs["gligen"] = {"objs": self.position_net(**gligen_args)} |
|
|
|
|
|
lora_scale = cross_attention_kwargs.get("scale", 1.0) if cross_attention_kwargs is not None else 1.0 |
|
if USE_PEFT_BACKEND: |
|
|
|
scale_lora_layers(self, lora_scale) |
|
|
|
is_controlnet = mid_block_additional_residual is not None and down_block_additional_residuals is not None |
|
|
|
is_adapter = down_intrablock_additional_residuals is not None |
|
|
|
|
|
|
|
if not is_adapter and mid_block_additional_residual is None and down_block_additional_residuals is not None: |
|
deprecate( |
|
"T2I should not use down_block_additional_residuals", |
|
"1.3.0", |
|
"Passing intrablock residual connections with `down_block_additional_residuals` is deprecated \ |
|
and will be removed in diffusers 1.3.0. `down_block_additional_residuals` should only be used \ |
|
for ControlNet. Please make sure use `down_intrablock_additional_residuals` instead. ", |
|
standard_warn=False, |
|
) |
|
down_intrablock_additional_residuals = down_block_additional_residuals |
|
is_adapter = True |
|
|
|
down_block_res_samples = (sample,) |
|
for downsample_block in self.down_blocks: |
|
if hasattr(downsample_block, "has_cross_attention") and downsample_block.has_cross_attention: |
|
|
|
additional_residuals = {} |
|
if is_adapter and len(down_intrablock_additional_residuals) > 0: |
|
additional_residuals["additional_residuals"] = down_intrablock_additional_residuals.pop(0) |
|
|
|
sample, res_samples = downsample_block( |
|
hidden_states=sample, |
|
temb=emb, |
|
encoder_hidden_states=encoder_hidden_states, |
|
attention_mask=attention_mask, |
|
cross_attention_kwargs=cross_attention_kwargs, |
|
encoder_attention_mask=encoder_attention_mask, |
|
**additional_residuals, |
|
) |
|
else: |
|
sample, res_samples = downsample_block(hidden_states=sample, temb=emb, scale=lora_scale) |
|
if is_adapter and len(down_intrablock_additional_residuals) > 0: |
|
sample += down_intrablock_additional_residuals.pop(0) |
|
|
|
down_block_res_samples += res_samples |
|
|
|
if is_controlnet: |
|
new_down_block_res_samples = () |
|
|
|
for down_block_res_sample, down_block_additional_residual in zip( |
|
down_block_res_samples, down_block_additional_residuals |
|
): |
|
down_block_res_sample = down_block_res_sample + down_block_additional_residual |
|
new_down_block_res_samples = new_down_block_res_samples + (down_block_res_sample,) |
|
|
|
down_block_res_samples = new_down_block_res_samples |
|
|
|
|
|
if self.mid_block is not None: |
|
if hasattr(self.mid_block, "has_cross_attention") and self.mid_block.has_cross_attention: |
|
sample = self.mid_block( |
|
sample, |
|
emb, |
|
encoder_hidden_states=encoder_hidden_states, |
|
attention_mask=attention_mask, |
|
cross_attention_kwargs=cross_attention_kwargs, |
|
encoder_attention_mask=encoder_attention_mask, |
|
) |
|
else: |
|
sample = self.mid_block(sample, emb) |
|
|
|
|
|
if ( |
|
is_adapter |
|
and len(down_intrablock_additional_residuals) > 0 |
|
and sample.shape == down_intrablock_additional_residuals[0].shape |
|
): |
|
sample += down_intrablock_additional_residuals.pop(0) |
|
|
|
if is_controlnet: |
|
sample = sample + mid_block_additional_residual |
|
|
|
|
|
|
|
setattr(self, 'skip_feature', deepcopy(down_block_res_samples)) |
|
setattr(self, 'toup_feature', sample.detach().clone()) |
|
|
|
|
|
|
|
|
|
|
|
if isinstance(timestep, list): |
|
step = len(timestep) |
|
|
|
timesteps = warpped_timestep(timestep, sample.shape[0]).to(sample.device) |
|
t_emb = self.time_proj(timesteps) |
|
|
|
|
|
|
|
|
|
t_emb = t_emb.to(dtype=self.dtype) |
|
|
|
emb = self.time_embedding(t_emb, timestep_cond) |
|
else: |
|
step = 1 |
|
down_block_res_samples = warpped_skip_feature(down_block_res_samples, step) |
|
sample = warpped_feature(sample, step) |
|
encoder_hidden_states = warpped_text_emb(encoder_hidden_states, step) |
|
|
|
|
|
else: |
|
step = 1 |
|
down_block_res_samples = self.skip_feature |
|
sample = self.toup_feature |
|
|
|
|
|
down_block_res_samples = warpped_skip_feature(down_block_res_samples, step) |
|
sample = warpped_feature(sample, step) |
|
encoder_hidden_states = warpped_text_emb(encoder_hidden_states, step) |
|
|
|
|
|
|
|
for i, upsample_block in enumerate(self.up_blocks): |
|
is_final_block = i == len(self.up_blocks) - 1 |
|
|
|
res_samples = down_block_res_samples[-len(upsample_block.resnets) :] |
|
down_block_res_samples = down_block_res_samples[: -len(upsample_block.resnets)] |
|
|
|
|
|
|
|
if not is_final_block and forward_upsample_size: |
|
upsample_size = down_block_res_samples[-1].shape[2:] |
|
|
|
if hasattr(upsample_block, "has_cross_attention") and upsample_block.has_cross_attention: |
|
sample = upsample_block( |
|
hidden_states=sample, |
|
temb=emb, |
|
res_hidden_states_tuple=res_samples, |
|
encoder_hidden_states=encoder_hidden_states, |
|
cross_attention_kwargs=cross_attention_kwargs, |
|
upsample_size=upsample_size, |
|
attention_mask=attention_mask, |
|
encoder_attention_mask=encoder_attention_mask, |
|
) |
|
else: |
|
lora_scale = cross_attention_kwargs.get("scale", 1.0) if cross_attention_kwargs is not None else 1.0 |
|
sample = upsample_block( |
|
hidden_states=sample, |
|
temb=emb, |
|
res_hidden_states_tuple=res_samples, |
|
upsample_size=upsample_size, |
|
scale=lora_scale, |
|
) |
|
|
|
|
|
if self.conv_norm_out: |
|
sample = self.conv_norm_out(sample) |
|
sample = self.conv_act(sample) |
|
sample = self.conv_out(sample) |
|
|
|
if USE_PEFT_BACKEND: |
|
|
|
unscale_lora_layers(self, lora_scale) |
|
|
|
if not return_dict: |
|
return (sample,) |
|
|
|
return UNet2DConditionOutput(sample=sample) |
|
return forward |
|
if model.__class__.__name__ == 'UNet2DConditionModel': |
|
model.forward = faster_orig_forward(model) |
|
|
|
|
|
def register_time(unet, t): |
|
setattr(unet, 'order', t) |
|
|
|
|