aashish1904 commited on
Commit
a8f1134
·
verified ·
1 Parent(s): bee465d

Upload README.md with huggingface_hub

Browse files
Files changed (1) hide show
  1. README.md +170 -0
README.md ADDED
@@ -0,0 +1,170 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+
2
+ ---
3
+
4
+ language: en
5
+ license: mit
6
+ tags:
7
+ - chain-of-thought
8
+ - structured-response
9
+ - causal-lm
10
+ - text-generation
11
+ datasets:
12
+ - diverse
13
+ pipeline_tag: text-generation
14
+ model_name: state-0
15
+ library_name: transformers
16
+ metrics:
17
+ - accuracy
18
+ - character
19
+ inference: true
20
+
21
+ ---
22
+
23
+ [![QuantFactory Banner](https://lh7-rt.googleusercontent.com/docsz/AD_4nXeiuCm7c8lEwEJuRey9kiVZsRn2W-b4pWlu3-X534V3YmVuVc2ZL-NXg2RkzSOOS2JXGHutDuyyNAUtdJI65jGTo8jT9Y99tMi4H4MqL44Uc5QKG77B0d6-JfIkZHFaUA71-RtjyYZWVIhqsNZcx8-OMaA?key=xt3VSDoCbmTY7o-cwwOFwQ)](https://hf.co/QuantFactory)
24
+
25
+
26
+ # QuantFactory/state-0-GGUF
27
+ This is quantized version of [Exthalpy/state-0](https://huggingface.co/Exthalpy/state-0) created using llama.cpp
28
+
29
+ # Original Model Card
30
+
31
+
32
+
33
+
34
+
35
+ # State-0: A chain-of-thoughts-based 8B alternative to GPT-o1
36
+
37
+ [![Open in Colab](https://colab.research.google.com/assets/colab-badge.svg)](https://colab.research.google.com/drive/124hfluZIrtVeZ-gWJEz6C_6nhfFpUBhY?usp=sharing)
38
+
39
+ [![Read Release Note](https://img.shields.io/badge/Read-Release%20Note-brightgreen)](https://exthalpy.com/2024/09/18/introducing-state-0-exthalpys-advanced-chain-of-thought-ai-model-on-hugging-face/)
40
+
41
+
42
+ ## Model Card
43
+
44
+ - **Model Name**: State-0
45
+ - **Version**: 1.0
46
+ - **Author**: Udit Akhouri
47
+ - **Hugging Face Model Page**: [Exthalpy/state-0](https://huggingface.co/Exthalpy/state-0/)
48
+ - **Architecture**: 8b core parameters with an additional 40 million parameters
49
+ - **Training Data**: Diverse datasets across various domains
50
+ - **Capabilities**: Chain-of-thought reasoning, Socratic instincts, in-depth and structured responses
51
+ - **Competitive Benchmark**: Capable of matching and surpassing the reasoning ability of GPT-4o1
52
+ - **Applications**: Educational tools, research, analytical problem-solving, and more
53
+ - **License**: MIT License
54
+
55
+ ## Abstract
56
+
57
+ State-0 is a novel chain-of-thought language model, designed to emulate structured human-like reasoning in its responses. Inspired from the robust architecture of Llama 3.1 8b and enhanced with over 40 million additional parameters, State-0 achieves a significant leap in cognitive capabilities. It incorporates "Socratic instincts" to dissect complex queries methodically and arrive at well-rounded conclusions. Competing with the reasoning prowess of GPT-4o1, State-0 not only provides accurate answers but also elucidates the logical pathways leading to those answers, making it a powerful tool for applications requiring in-depth analysis and clarity.
58
+
59
+ ## 1. Introduction
60
+
61
+ The field of natural language processing (NLP) has been significantly advanced by large language models (LLMs) capable of generating human-like text. However, most LLMs still lack the ability to break down complex queries into multiple facets, analyze them, and synthesize a comprehensive answer. State-0 addresses this limitation by employing a chain-of-thought reasoning mechanism combined with Socratic instincts. This paper introduces the architecture, training, and capabilities of State-0, demonstrating its competitive edge against models like GPT-4o1 in structured thinking and problem-solving.
62
+
63
+ ## 2. Model Architecture
64
+
65
+ State-0, fundamentally inspired by Llama 3.1 8b, is augmented with over 40 million chain-of-thoughts and socratic instinct parameters. These parameters are meticulously trained to enhance the model's ability to reason, question, and deduce, drawing from vast datasets encompassing diverse fields of knowledge.
66
+
67
+ ### 2.1 Enhancements Over Llama 3.1 8b
68
+
69
+ - **Additional Parameters**: State-0 incorporates 40 million additional parameters specifically fine-tuned to improve logical reasoning and analytical depth.
70
+ - **Chain-of-Thought Mechanism**: The model leverages a multi-step process for breaking down queries into manageable components, similar to human analytical thinking.
71
+ - **Socratic Instincts**: Inspired by the Socratic method, State-0 probes different aspects of a question, leading to a comprehensive and reasoned conclusion.
72
+
73
+ ## 3. Model Capabilities
74
+
75
+ ### 3.1 Chain-of-Thought Reasoning
76
+
77
+ State-0 excels at decomposing complex questions into simpler elements. It addresses these components individually and synthesizes the answers into a coherent response.
78
+
79
+ **Example**:
80
+ **Prompt**: Is milk a good source of protein?
81
+
82
+ **Response**:
83
+ 1. **Definition**: Milk is a dairy product produced by mammals, including cows, goats, and sheep.
84
+ 2. **Protein Content**: Milk is a rich source of protein, providing essential amino acids.
85
+ 3. **Importance**: Protein is necessary for muscle growth, tissue repair, and immune function.
86
+ **Conclusion**: Milk is a valuable source of protein and contributes to overall health.
87
+
88
+ ### 3.2 Competing with GPT-4o1
89
+
90
+ State-0 demonstrates competitive performance in terms of analytical depth and reasoning, often surpassing models like GPT-4o1 in its ability to provide contextually relevant and logically sound answers.
91
+
92
+ ## 4. Getting Started
93
+
94
+ State-0 is available for use via the Hugging Face `transformers` library. This section outlines the installation and usage process for integrating State-0 into your projects.
95
+
96
+ ### 4.1 Installation
97
+
98
+ Ensure you have the `transformers` library installed:
99
+
100
+ ```bash
101
+ pip install transformers
102
+ ```
103
+
104
+ ### 4.2 Usage
105
+
106
+ #### High-Level Pipeline
107
+
108
+ State-0 can be easily used with the high-level pipeline API for text generation:
109
+
110
+ ```python
111
+ from transformers import pipeline
112
+
113
+ pipe = pipeline("text-generation", model="uditakhouri/state-0")
114
+ response = pipe("Is milk a good source of protein?")
115
+ print(response)
116
+ ```
117
+
118
+ #### Direct Model Loading
119
+
120
+ For more control, State-0 can be loaded directly using the following code:
121
+
122
+ ```python
123
+ from transformers import AutoTokenizer, AutoModelForCausalLM
124
+
125
+ tokenizer = AutoTokenizer.from_pretrained("uditakhouri/state-0")
126
+ model = AutoModelForCausalLM.from_pretrained("uditakhouri/state-0")
127
+
128
+ input_text = "Is milk a good source of protein?"
129
+ input_ids = tokenizer.encode(input_text, return_tensors="pt")
130
+
131
+ output = model.generate(input_ids, max_length=100)
132
+ response = tokenizer.decode(output[0], skip_special_tokens=True)
133
+ print(response)
134
+ ```
135
+
136
+ ## 5. Training Details
137
+
138
+ State-0 was trained using a diverse set of datasets, fine-tuned to enhance its reasoning and conversational abilities. The training process focused on:
139
+ - Reinforcement Learning from Human Feedback (RLHF) for nuanced responses.
140
+ - Incorporating various fields of knowledge, from basic concepts to complex theories, to create a versatile reasoning engine.
141
+
142
+ ## 6. Socratic Instincts
143
+
144
+ Inspired by the Socratic method, State-0 is designed to think through different scenarios and perspectives before arriving at an answer. This is achieved through:
145
+ - **Multi-Step Processing**: Breaking down a question into smaller parts, analyzing each component, and then synthesizing an answer.
146
+ - **Self-Interrogation**: The model internally queries different aspects of a topic, ensuring a balanced and well-thought-out response.
147
+
148
+ ## 7. Evaluation and Results
149
+
150
+ State-0 has been rigorously tested against existing models like GPT-4o1, showing a high level of competence in chain-of-thought reasoning. It provides not only accurate answers but also the logical pathway leading to those answers, setting a new benchmark in LLM reasoning.
151
+
152
+ ## 8. Conclusion
153
+
154
+ State-0 represents a significant advancement in the field of NLP by integrating chain-of-thought reasoning and Socratic instincts into its framework. With its enhanced parameters and structured analytical capabilities, State-0 is a formidable model for applications that demand a deep and reasoned understanding of complex queries.
155
+
156
+ ## 9. Future Work
157
+
158
+ Future versions of State-0 aim to further enhance its reasoning capabilities by incorporating more advanced cognitive models and expanding its knowledge base.
159
+
160
+ ## 10. License
161
+
162
+ State-0 is released under the MIT License.
163
+
164
+ ## 11. References
165
+
166
+ For a complete list of references and further reading, please visit the model's page on [Hugging Face](https://huggingface.co/uditakhouri/state-0).
167
+
168
+ ## 12. Contact
169
+
170
+ For inquiries, collaborations, or further information, please contact Udit Akhouri.