--- license: mit base_model: - stabilityai/stable-diffusion-xl-base-1.0 --- - Requires a custom training notebook that will be provided soon. - Distilling SDXL using T5 attention masking for the sake of teaching SDXL; CLIP_L and CLIP_G to expect the T5 attention mask. - Additional finetuning required, additional interpolation required, addistional distillation required for full cohesion. - Ongoing training effort interpolating the T5 into SDXL using teacher/student process. - -config = { - "epochs": 10, - "batch_size": 64, - "learning_rate": 1e-6, # Lower learning rate for stability - "save_interval_steps": 10, # Save checkpoint every 10 training steps - "test_save_interval_steps": 10, # Save test images every 10 training steps - "checkpoint_dir": "./checkpoints", # Full diffusers checkpoint folder - "compact_model_dir": "./compact_model", # For final compact model (not used for caching) - "baseline_test_dir": "./baseline_test", # For baseline images & captions - "cache_dir": "./cache", # Folder for caching T5 outputs and teacher features - "num_generated_captions": 128, # Number of captions to generate for training - "model_id": "stabilityai/stable-diffusion-xl-base-1.0", - "model_name": "my_interpolative_distillation", # Folder name for checkpoints - "seed": 420, - "device": torch.device("cuda:0") if torch.cuda.is_available() else torch.device("cpu"), - "inference_steps": 50, - "height": 1024, - "width": 1024, - "guidance_scale": 7.5, - "inference_interval": 10, - "max_caption_length": 512, - # Batch size for teacher feature caching (set very low to reduce VRAM usage) - "cache_teacher_batch_size": 64, -} -