Update README.md
Browse files
README.md
CHANGED
@@ -1,3 +1,70 @@
|
|
1 |
-
---
|
2 |
-
license: cc-by-nd-4.0
|
3 |
-
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
---
|
2 |
+
license: cc-by-nd-4.0
|
3 |
+
language:
|
4 |
+
- en
|
5 |
+
library_name: diffusers
|
6 |
+
pipeline_tag: text-to-image
|
7 |
+
tags:
|
8 |
+
- art
|
9 |
+
- reatistic
|
10 |
+
- text-image-generator
|
11 |
+
- stable-diffusion
|
12 |
+
---
|
13 |
+
# Fine-Tuning Stable Diffusion with Realistic Vision V2.0
|
14 |
+
|
15 |
+
## Overview
|
16 |
+
|
17 |
+
This repository contains a fine-tuned version of the **Realistic Vision V2.0** model, a powerful variant of the Stable Diffusion model, tailored for generating high-quality, realistic images from text prompts. The fine-tuning process was conducted on a custom dataset to improve the model's performance in specific domains.
|
18 |
+
|
19 |
+
## Features of Realistic Vision V2.0
|
20 |
+
|
21 |
+
- **High-Quality Image Generation**: Produces detailed and realistic images that closely adhere to the provided text prompts.
|
22 |
+
- **Enhanced Detail Preservation**: Maintains fine details in the generated images, making it suitable for applications requiring high fidelity.
|
23 |
+
- **Versatile Output**: Capable of generating a wide range of visual styles based on varying prompts, from artistic to photorealistic images.
|
24 |
+
- **Optimized Inference**: Efficient performance on modern GPUs, with customizable parameters like inference steps and guidance scale to balance speed and quality.
|
25 |
+
|
26 |
+
## Why Use Realistic Vision V2.0?
|
27 |
+
|
28 |
+
- **Superior Realism**: Compared to earlier versions, Realistic Vision V2.0 has been fine-tuned to enhance the realism of generated images, making it ideal for applications in media, design, and content creation.
|
29 |
+
- **Customizable Outputs**: The model allows users to fine-tune parameters to match their specific needs, whether they are looking for highly accurate or more creative and abstract images.
|
30 |
+
- **Proven Performance**: Backed by the robust Stable Diffusion framework, Realistic Vision V2.0 leverages state-of-the-art techniques in diffusion models to deliver consistent, high-quality results.
|
31 |
+
|
32 |
+
## Using the Pretrained Model
|
33 |
+
|
34 |
+
The fine-tuned model is available on Hugging Face and can be easily accessed and utilized:
|
35 |
+
|
36 |
+
### 1. Installation
|
37 |
+
|
38 |
+
First, install the necessary libraries:
|
39 |
+
|
40 |
+
pip install torch torchvision diffusers accelerate huggingface_hub
|
41 |
+
|
42 |
+
### 2. Access the Model
|
43 |
+
#### You can load and use the model in your Python environment as follows:
|
44 |
+
from diffusers import StableDiffusionPipeline
|
45 |
+
import torch
|
46 |
+
|
47 |
+
#### Load the fine-tuned model
|
48 |
+
|
49 |
+
model_id = "majid230/Realistic_Vision_V2.0"
|
50 |
+
|
51 |
+
pipe = StableDiffusionPipeline.from_pretrained(model_id, torch_dtype=torch.float32)
|
52 |
+
|
53 |
+
pipe = pipe.to("cuda" if torch.cuda.is_available() else "cpu")
|
54 |
+
|
55 |
+
#### Generate an image from a prompt
|
56 |
+
|
57 |
+
prompt = "A futuristic cityscape at sunset"
|
58 |
+
|
59 |
+
image = pipe(prompt, num_inference_steps=50, guidance_scale=7.5).images[0]
|
60 |
+
|
61 |
+
#### Save or display the image
|
62 |
+
|
63 |
+
image.save("generated_image.png")
|
64 |
+
|
65 |
+
image.show()
|
66 |
+
## 3.Customization
|
67 |
+
num_inference_steps: Adjust this parameter to control the number of steps the model takes during image generation. More steps typically yield higher-quality images.
|
68 |
+
guidance_scale: Modify this to control how closely the generated image follows the prompt. Higher values make the image more prompt-specific, while lower values allow for more creative interpretations.
|
69 |
+
## Acknowledgment
|
70 |
+
This project was generously supported and provided by Machine Learning 1 Pvt Ltd. The fine-tuning and further development were carried out by Majid Hanif.
|