Skip to content

Commit 96f872c

Browse files
SHSH
authored andcommitted
fix bug bring in diffuser 0.24.0
1 parent 219fef6 commit 96f872c

File tree

3 files changed

+29
-10
lines changed

3 files changed

+29
-10
lines changed

inference.sh

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -1,2 +1,2 @@
11

2-
python src/diffusers_sample.py --delta_ckpt logs/wooden_pot/delta.bin --ckpt "/data/home/chensh/data/huggingface_model/stable-diffusion-xl-base-1.0" --prompt "<new1> cat playing with a ball"
2+
CUDA_VISIBLE_DEVICES=0 python src/diffusers_sample.py --delta_ckpt logs/wooden_pot/delta.bin --sdxl --ckpt "/data/home/chensh/data/huggingface_model/stable-diffusion-xl-base-1.0" --prompt "<new1> cat playing with a ball"

src/diffusers_model_pipeline.py

Lines changed: 23 additions & 8 deletions
Original file line numberDiff line numberDiff line change
@@ -212,7 +212,8 @@
212212
# limitations under the License.
213213
from typing import Callable, Optional
214214
import torch
215-
from transformers import CLIPFeatureExtractor, CLIPTextModel, CLIPTokenizer, CLIPTextModelWithProjection
215+
from transformers import CLIPFeatureExtractor, CLIPTextModel, CLIPTokenizer, CLIPTextModelWithProjection, \
216+
CLIPVisionModelWithProjection, CLIPImageProcessor
216217
from accelerate.logging import get_logger
217218

218219
from diffusers.models import AutoencoderKL, UNet2DConditionModel
@@ -551,22 +552,36 @@ def __init__(
551552
tokenizer_2: CLIPTokenizer,
552553
unet: UNet2DConditionModel,
553554
scheduler: KarrasDiffusionSchedulers,
555+
image_encoder: CLIPVisionModelWithProjection = None,
556+
feature_extractor: CLIPImageProcessor = None,
554557
force_zeros_for_empty_prompt: bool = True,
555558
add_watermarker: Optional[bool] = None,
556559
modifier_token: list = [],
557560
modifier_token_id: list = [],
558561
modifier_token_id_2: list = []
559562
):
560-
super().__init__(vae=vae,
561-
text_encoder=text_encoder,
562-
text_encoder_2=text_encoder_2,
563-
tokenizer=tokenizer,
564-
tokenizer_2=tokenizer_2,
565-
unet=unet,
566-
scheduler=scheduler,
563+
super().__init__(vae,
564+
text_encoder,
565+
text_encoder_2,
566+
tokenizer,
567+
tokenizer_2,
568+
unet,
569+
scheduler,
570+
image_encoder=image_encoder,
571+
feature_extractor=feature_extractor,
567572
force_zeros_for_empty_prompt=force_zeros_for_empty_prompt,
568573
add_watermarker=add_watermarker,
569574
)
575+
# super().__init__(vae,
576+
# text_encoder,
577+
# text_encoder_2,
578+
# tokenizer,
579+
# tokenizer_2,
580+
# unet,
581+
# scheduler,
582+
# force_zeros_for_empty_prompt,
583+
# add_watermarker,
584+
# )
570585

571586
# change attn class
572587
self.modifier_token = modifier_token

src/diffusers_sample.py

Lines changed: 5 additions & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -10,6 +10,7 @@
1010
import numpy as np
1111
import torch
1212
from PIL import Image
13+
from diffusers import StableDiffusionXLPipeline
1314

1415
sys.path.append('./')
1516
from src.diffusers_model_pipeline import CustomDiffusionPipeline, CustomDiffusionXLPipeline
@@ -18,7 +19,10 @@
1819
def sample(ckpt, delta_ckpt, from_file, prompt, compress, batch_size, freeze_model, sdxl=False):
1920
model_id = ckpt
2021
if sdxl:
21-
pipe = CustomDiffusionXLPipeline.from_pretrained(model_id, torch_dtype=torch.float16).to("cuda")
22+
pipe = CustomDiffusionXLPipeline.from_pretrained(model_id, torch_dtype=torch.float16)
23+
print(pipe.components)
24+
# pipe = StableDiffusionXLPipeline.from_pretrained(model_id, torch_dtype=torch.float16)
25+
pipe = pipe.to("cuda")
2226
else:
2327
pipe = CustomDiffusionPipeline.from_pretrained(model_id, torch_dtype=torch.float16).to("cuda")
2428
pipe.load_model(delta_ckpt, compress)

0 commit comments

Comments
 (0)