matrixglitch
commited on
Commit
•
d57378f
1
Parent(s):
e465bc1
Update README.md
Browse files
README.md
CHANGED
@@ -12,5 +12,5 @@ this finetune is based on llama3 8b and operates under the same license
|
|
12 |
the model aims to convert non-natural language prompts to natural language automatically while retaining the general idea of the prompt because SD3 performs poorly on non-natural language prompts.
|
13 |
|
14 |
this model was trained for 1500 steps on an RTX 3090 for 2.5 hours, training any longer gave a deminishing result because of the low batch size.
|
15 |
-
the dataset has 90k+ original prompts (non-natural language questions) and
|
16 |
it has not been trained on the whole dataset because of lack of compute but the results are already amazing.
|
|
|
12 |
the model aims to convert non-natural language prompts to natural language automatically while retaining the general idea of the prompt because SD3 performs poorly on non-natural language prompts.
|
13 |
|
14 |
this model was trained for 1500 steps on an RTX 3090 for 2.5 hours, training any longer gave a deminishing result because of the low batch size.
|
15 |
+
the dataset has 90k+ original prompts (non-natural language questions) and 220k modified prompts (natural language answers).
|
16 |
it has not been trained on the whole dataset because of lack of compute but the results are already amazing.
|