File size: 2,321 Bytes
d71a0a7
67b7886
 
 
 
 
 
b62fe87
 
f4c878f
b62fe87
f4c878f
79df340
 
 
 
67b7886
d71a0a7
67b7886
78fad51
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
2b04258
78fad51
2b04258
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
---
pipeline_tag: text-to-image
widget:
- text: >-
    Three cow grazing in a bay window
  output:
    url: cow.png 

- text: >-
    Super Closeup Portrait, action shot, Profoundly dark whitish meadow, glass flowers, Stains, space grunge style, Jeanne d'Arc wearing White Olive green used styled Cotton frock, Wielding thin silver sword, Sci-fi vibe, dirty, noisy, Vintage monk style, very detailed, hd
  output:
    url: girl.png 
- text: >-
    A pile of bananas
  output:
    url: banana.png 
license: gpl-3.0 
---
<Gallery />
## BitDiffusionV0.1

This is the initial version of the image model trained on the Bittensor network within subnet 17. It's not expected for this model to perform as well as MidJourney V6 at the moment. However, it does generate better images than base SDXL model.

**Trained on the dataset of Subnet 19 Vision.**

## Settings for BitDiffusionV0.1

Use these settings for the best results with BitDiffusionV0.1:

CFG Scale: Use a CFG scale of 8

Steps: 40 to 60 steps

Sampler: DPM++ 2M SDE

Scheduler: Karras

Resolution: 1024x1024

## Use it with 🧨 diffusers
```python
import torch
from diffusers import (
    StableDiffusionXLPipeline, 
    KDPM2AncestralDiscreteScheduler,
    AutoencoderKL
)

# Load VAE component
vae = AutoencoderKL.from_pretrained(
    "madebyollin/sdxl-vae-fp16-fix", 
    torch_dtype=torch.float16
)

# Configure the pipeline
pipe = StableDiffusionXLPipeline.from_pretrained(
    "PlixAI/BitDiffusionV0.1", 
    vae=vae,
    torch_dtype=torch.float16
)
pipe.scheduler = KDPM2AncestralDiscreteScheduler.from_config(pipe.scheduler.config)
pipe.to('cuda')

# Define prompts and generate image
prompt = "black fluffy gorgeous dangerous cat animal creature, large orange eyes, big fluffy ears, piercing gaze, full moon, dark ambiance, best quality, extremely detailed"
negative_prompt = "nsfw, bad quality, bad anatomy, worst quality, low quality, low resolutions, extra fingers, blur, blurry, ugly, wrongs proportions, watermark, image artifacts, lowres, ugly, jpeg artifacts, deformed, noisy image"

image = pipe(
    prompt, 
    negative_prompt=negative_prompt, 
    width=1024,
    height=1024,
    guidance_scale=7.5,
    num_inference_steps=50
).images[0]
```

Training Subnet : https://github.com/PlixML/pixel

Data Subnet : https://github.com/namoray/vision