Upload README.md with huggingface_hub
Browse files
README.md
ADDED
@@ -0,0 +1,170 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
|
2 |
+
---
|
3 |
+
|
4 |
+
language: en
|
5 |
+
license: mit
|
6 |
+
tags:
|
7 |
+
- chain-of-thought
|
8 |
+
- structured-response
|
9 |
+
- causal-lm
|
10 |
+
- text-generation
|
11 |
+
datasets:
|
12 |
+
- diverse
|
13 |
+
pipeline_tag: text-generation
|
14 |
+
model_name: state-0
|
15 |
+
library_name: transformers
|
16 |
+
metrics:
|
17 |
+
- accuracy
|
18 |
+
- character
|
19 |
+
inference: true
|
20 |
+
|
21 |
+
---
|
22 |
+
|
23 |
+
[](https://hf.co/QuantFactory)
|
24 |
+
|
25 |
+
|
26 |
+
# QuantFactory/state-0-GGUF
|
27 |
+
This is quantized version of [Exthalpy/state-0](https://huggingface.co/Exthalpy/state-0) created using llama.cpp
|
28 |
+
|
29 |
+
# Original Model Card
|
30 |
+
|
31 |
+
|
32 |
+
|
33 |
+
|
34 |
+
|
35 |
+
# State-0: A chain-of-thoughts-based 8B alternative to GPT-o1
|
36 |
+
|
37 |
+
[](https://colab.research.google.com/drive/124hfluZIrtVeZ-gWJEz6C_6nhfFpUBhY?usp=sharing)
|
38 |
+
|
39 |
+
[](https://exthalpy.com/2024/09/18/introducing-state-0-exthalpys-advanced-chain-of-thought-ai-model-on-hugging-face/)
|
40 |
+
|
41 |
+
|
42 |
+
## Model Card
|
43 |
+
|
44 |
+
- **Model Name**: State-0
|
45 |
+
- **Version**: 1.0
|
46 |
+
- **Author**: Udit Akhouri
|
47 |
+
- **Hugging Face Model Page**: [Exthalpy/state-0](https://huggingface.co/Exthalpy/state-0/)
|
48 |
+
- **Architecture**: 8b core parameters with an additional 40 million parameters
|
49 |
+
- **Training Data**: Diverse datasets across various domains
|
50 |
+
- **Capabilities**: Chain-of-thought reasoning, Socratic instincts, in-depth and structured responses
|
51 |
+
- **Competitive Benchmark**: Capable of matching and surpassing the reasoning ability of GPT-4o1
|
52 |
+
- **Applications**: Educational tools, research, analytical problem-solving, and more
|
53 |
+
- **License**: MIT License
|
54 |
+
|
55 |
+
## Abstract
|
56 |
+
|
57 |
+
State-0 is a novel chain-of-thought language model, designed to emulate structured human-like reasoning in its responses. Inspired from the robust architecture of Llama 3.1 8b and enhanced with over 40 million additional parameters, State-0 achieves a significant leap in cognitive capabilities. It incorporates "Socratic instincts" to dissect complex queries methodically and arrive at well-rounded conclusions. Competing with the reasoning prowess of GPT-4o1, State-0 not only provides accurate answers but also elucidates the logical pathways leading to those answers, making it a powerful tool for applications requiring in-depth analysis and clarity.
|
58 |
+
|
59 |
+
## 1. Introduction
|
60 |
+
|
61 |
+
The field of natural language processing (NLP) has been significantly advanced by large language models (LLMs) capable of generating human-like text. However, most LLMs still lack the ability to break down complex queries into multiple facets, analyze them, and synthesize a comprehensive answer. State-0 addresses this limitation by employing a chain-of-thought reasoning mechanism combined with Socratic instincts. This paper introduces the architecture, training, and capabilities of State-0, demonstrating its competitive edge against models like GPT-4o1 in structured thinking and problem-solving.
|
62 |
+
|
63 |
+
## 2. Model Architecture
|
64 |
+
|
65 |
+
State-0, fundamentally inspired by Llama 3.1 8b, is augmented with over 40 million chain-of-thoughts and socratic instinct parameters. These parameters are meticulously trained to enhance the model's ability to reason, question, and deduce, drawing from vast datasets encompassing diverse fields of knowledge.
|
66 |
+
|
67 |
+
### 2.1 Enhancements Over Llama 3.1 8b
|
68 |
+
|
69 |
+
- **Additional Parameters**: State-0 incorporates 40 million additional parameters specifically fine-tuned to improve logical reasoning and analytical depth.
|
70 |
+
- **Chain-of-Thought Mechanism**: The model leverages a multi-step process for breaking down queries into manageable components, similar to human analytical thinking.
|
71 |
+
- **Socratic Instincts**: Inspired by the Socratic method, State-0 probes different aspects of a question, leading to a comprehensive and reasoned conclusion.
|
72 |
+
|
73 |
+
## 3. Model Capabilities
|
74 |
+
|
75 |
+
### 3.1 Chain-of-Thought Reasoning
|
76 |
+
|
77 |
+
State-0 excels at decomposing complex questions into simpler elements. It addresses these components individually and synthesizes the answers into a coherent response.
|
78 |
+
|
79 |
+
**Example**:
|
80 |
+
**Prompt**: Is milk a good source of protein?
|
81 |
+
|
82 |
+
**Response**:
|
83 |
+
1. **Definition**: Milk is a dairy product produced by mammals, including cows, goats, and sheep.
|
84 |
+
2. **Protein Content**: Milk is a rich source of protein, providing essential amino acids.
|
85 |
+
3. **Importance**: Protein is necessary for muscle growth, tissue repair, and immune function.
|
86 |
+
**Conclusion**: Milk is a valuable source of protein and contributes to overall health.
|
87 |
+
|
88 |
+
### 3.2 Competing with GPT-4o1
|
89 |
+
|
90 |
+
State-0 demonstrates competitive performance in terms of analytical depth and reasoning, often surpassing models like GPT-4o1 in its ability to provide contextually relevant and logically sound answers.
|
91 |
+
|
92 |
+
## 4. Getting Started
|
93 |
+
|
94 |
+
State-0 is available for use via the Hugging Face `transformers` library. This section outlines the installation and usage process for integrating State-0 into your projects.
|
95 |
+
|
96 |
+
### 4.1 Installation
|
97 |
+
|
98 |
+
Ensure you have the `transformers` library installed:
|
99 |
+
|
100 |
+
```bash
|
101 |
+
pip install transformers
|
102 |
+
```
|
103 |
+
|
104 |
+
### 4.2 Usage
|
105 |
+
|
106 |
+
#### High-Level Pipeline
|
107 |
+
|
108 |
+
State-0 can be easily used with the high-level pipeline API for text generation:
|
109 |
+
|
110 |
+
```python
|
111 |
+
from transformers import pipeline
|
112 |
+
|
113 |
+
pipe = pipeline("text-generation", model="uditakhouri/state-0")
|
114 |
+
response = pipe("Is milk a good source of protein?")
|
115 |
+
print(response)
|
116 |
+
```
|
117 |
+
|
118 |
+
#### Direct Model Loading
|
119 |
+
|
120 |
+
For more control, State-0 can be loaded directly using the following code:
|
121 |
+
|
122 |
+
```python
|
123 |
+
from transformers import AutoTokenizer, AutoModelForCausalLM
|
124 |
+
|
125 |
+
tokenizer = AutoTokenizer.from_pretrained("uditakhouri/state-0")
|
126 |
+
model = AutoModelForCausalLM.from_pretrained("uditakhouri/state-0")
|
127 |
+
|
128 |
+
input_text = "Is milk a good source of protein?"
|
129 |
+
input_ids = tokenizer.encode(input_text, return_tensors="pt")
|
130 |
+
|
131 |
+
output = model.generate(input_ids, max_length=100)
|
132 |
+
response = tokenizer.decode(output[0], skip_special_tokens=True)
|
133 |
+
print(response)
|
134 |
+
```
|
135 |
+
|
136 |
+
## 5. Training Details
|
137 |
+
|
138 |
+
State-0 was trained using a diverse set of datasets, fine-tuned to enhance its reasoning and conversational abilities. The training process focused on:
|
139 |
+
- Reinforcement Learning from Human Feedback (RLHF) for nuanced responses.
|
140 |
+
- Incorporating various fields of knowledge, from basic concepts to complex theories, to create a versatile reasoning engine.
|
141 |
+
|
142 |
+
## 6. Socratic Instincts
|
143 |
+
|
144 |
+
Inspired by the Socratic method, State-0 is designed to think through different scenarios and perspectives before arriving at an answer. This is achieved through:
|
145 |
+
- **Multi-Step Processing**: Breaking down a question into smaller parts, analyzing each component, and then synthesizing an answer.
|
146 |
+
- **Self-Interrogation**: The model internally queries different aspects of a topic, ensuring a balanced and well-thought-out response.
|
147 |
+
|
148 |
+
## 7. Evaluation and Results
|
149 |
+
|
150 |
+
State-0 has been rigorously tested against existing models like GPT-4o1, showing a high level of competence in chain-of-thought reasoning. It provides not only accurate answers but also the logical pathway leading to those answers, setting a new benchmark in LLM reasoning.
|
151 |
+
|
152 |
+
## 8. Conclusion
|
153 |
+
|
154 |
+
State-0 represents a significant advancement in the field of NLP by integrating chain-of-thought reasoning and Socratic instincts into its framework. With its enhanced parameters and structured analytical capabilities, State-0 is a formidable model for applications that demand a deep and reasoned understanding of complex queries.
|
155 |
+
|
156 |
+
## 9. Future Work
|
157 |
+
|
158 |
+
Future versions of State-0 aim to further enhance its reasoning capabilities by incorporating more advanced cognitive models and expanding its knowledge base.
|
159 |
+
|
160 |
+
## 10. License
|
161 |
+
|
162 |
+
State-0 is released under the MIT License.
|
163 |
+
|
164 |
+
## 11. References
|
165 |
+
|
166 |
+
For a complete list of references and further reading, please visit the model's page on [Hugging Face](https://huggingface.co/uditakhouri/state-0).
|
167 |
+
|
168 |
+
## 12. Contact
|
169 |
+
|
170 |
+
For inquiries, collaborations, or further information, please contact Udit Akhouri.
|