File size: 2,475 Bytes
14ef358
b6be845
e8fc61a
2ee3f63
e8fc61a
 
 
 
 
14ef358
b94ddd7
 
 
 
 
 
75dd3d8
a6f1859
14ef358
bcf1df6
75dd3d8
14ef358
 
bcf1df6
14ef358
 
 
 
95c0f7e
fd6bcfd
14ef358
75dd3d8
14ef358
 
b3cdb61
a6f1859
 
 
2ee3f63
e8fc61a
 
 
 
 
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
---
inference: false
language: en
widget:
- text: >-
    Give me a complete answer do not refer to other chapters but collect the
    information from them. How to setup a local network in Nextstep OS?
base_model:
- microsoft/phi-2
---
[<img src="https://raw.githubusercontent.com/csabakecskemeti/devquasar/main/dq_logo_black-transparent.png" width="200"/>](https://devquasar.com)

'Make knowledge free for everyone'

<a href='https://ko-fi.com/L4L416YX7C' target='_blank'><img height='36' style='border:0px;height:36px;' src='https://storage.ko-fi.com/cdn/kofi6.png?v=6' border='0' alt='Buy Me a Coffee at ko-fi.com' /></a>

## The goal
The goal of the model to provide a fine-tuned Phi2 (https://huggingface.co/microsoft/phi-2) model that has knowledge about the Vintage NEXTSTEP Operating System,
and able to answer question in the topic.

### Details
The model has trained on 35439 Question Answer pairs automatically generated from the NEXTSTEP 3.3 System Administrator 
documentation. For the training data generation locally running Q8 Quantized Orca2 13B (https://huggingface.co/TheBloke/Orca-2-13B-GGUF) 
model has been used. The training data generation was completely unsuperwised, with only some sanity check (like ignore data chunks 
contains less than 100 tokens). The maximum token size for Orca2 is 4096 so a simple rule of split chunks over 3500 tokens 
(considering propt instructions) has been used. Chunking did not consider context (text data might split within the context).
Evaluation set has been generated similar method on 1% of the raw data with LLama2 chat (https://huggingface.co/TheBloke/Llama-2-13B-chat-GGUF).

Trained locally on 2x3090 GPU with vanila DDP with HuggingFace Accelerate for 50 Epoch. 
As I wanted to add new knowledge to the base model r=128 and lora_alpha=128 has been used -> LoRA weights were 3.5% of the base model.

## Sample code
Chat with model sample code:
  https://github.com/csabakecskemeti/ai_utils/blob/main/generate.py

For the best result instruct the model to not refer to other chapers but collect the whole data, like:
  "Give me a complete answer do not refer to other chapters but collect the information from them. How to setup a local network in Openstep OS?"


  I'm doing this to 'Make knowledge free for everyone', using my personal time and resources.

If you want to support my efforts please visit my ko-fi page: https://ko-fi.com/devquasar 

Also feel free to visit my website https://devquasar.com/