Diffusers not pulling the fp16 model?

#5
by jsmidt - opened

Using the example, on the model page:

 from diffusers import AuraFlowPipeline
 import torch 

 pipeline = AuraFlowPipeline.from_pretrained(
     "fal/AuraFlow-v0.3",
     torch_dtype=torch.float16,
     variant="fp16",
 ).to("cuda")

downloaded the normal safetensors file and not the fp16.safetensors file. This also happened when I updated to the to the latest diffusers on git as the model page suggests.

After downloading the non fp16 transformer files, running the pipeline generated this warning:

 A mixture of fp16 and non-fp16 filenames will be loaded.
 Loaded fp16 filenames:
 [vae/diffusion_pytorch_model.fp16.safetensors, text_encoder/model.fp16.safetensors]
 Loaded non-fp16 filenames:
 [transformer/diffusion_pytorch_model-00001-of-00003.safetensors, transformer/diffusion_pytorch_model-00002-of-00003.safetensors, transformer/diffusion_pytorch_model-00003-of-00003.safetensors
 If this behavior is not expected, please check your folder structure.

Any ideas what the problem might be or how to fix this? Thanks.

diffusion_pytorch_model.safetensors.fp16.index.json should be renamed to
diffusion_pytorch_model.fp16.safetensors.index.json

i wasnt able to use variant="fp16" until i renamed it (found the hard way, because i deleted the fp32 files)

Sign up or log in to comment