Text Generation
Transformers
Safetensors
mistral
Mistral_Star
Mistral_Quiet
Mistral
Mixtral
Question-Answer
Token-Classification
Sequence-Classification
SpydazWeb-AI
chemistry
biology
legal
code
climate
medical
text-generation-inference
Not-For-All-Audiences
chain-of-thought
tree-of-knowledge
forest-of-thoughts
visual-spacial-sketchpad
alpha-mind
knowledge-graph
entity-detection
encyclopedia
wikipedia
stack-exchange
Reddit
Cyber-series
MegaMind
Cybertron
SpydazWeb
Spydaz
LCARS
star-trek
mega-transformers
Mulit-Mega-Merge
Multi-Lingual
Afro-Centric
African-Model
Ancient-One
Inference Endpoints
Update README.md
Browse files
README.md
CHANGED
@@ -192,6 +192,12 @@ the model can also generate markdown charts with mermaid.
|
|
192 |
* Chain of thoughts methods
|
193 |
* One shot / Multi shot prompting tasks
|
194 |
|
|
|
|
|
|
|
|
|
|
|
|
|
195 |
|
196 |
# Higher Settigs in generations to allow for sparse and indepth training of information: METHODS OF TRAINING
|
197 |
* highly trained on multiple facets and content ! - i use 1000topK, 0.42topK, 0.2Temp :hence drawing on a large results pool and selecting from a higher likelyhood of truth ..with a low temp to allow for accuracy of results .. to allow for more expression just up temp tiny... as i train my role play in deep to be a part of the models core !
|
|
|
192 |
* Chain of thoughts methods
|
193 |
* One shot / Multi shot prompting tasks
|
194 |
|
195 |
+
## Training Paradigms:: ( 1 year of training 365 merges later (all with the preious parents ))
|
196 |
+
|
197 |
+
### GENETIC MERGES TO PRESERVE THE PAST TASKS AND SUPER FINE TUNING
|
198 |
+
* A series of merges (Genetic algorithm Merge) - multiple merge targets chose to create the x+y+z models : utilizing the current series and previous series of models :
|
199 |
+
* hence the LCARS/ARCHIVE models have been remerged into the world archive models : hence the prompt used was to enable the model to learn books and recall books and pasages: as well as recount the historys of mankind:
|
200 |
+
|
201 |
|
202 |
# Higher Settigs in generations to allow for sparse and indepth training of information: METHODS OF TRAINING
|
203 |
* highly trained on multiple facets and content ! - i use 1000topK, 0.42topK, 0.2Temp :hence drawing on a large results pool and selecting from a higher likelyhood of truth ..with a low temp to allow for accuracy of results .. to allow for more expression just up temp tiny... as i train my role play in deep to be a part of the models core !
|