adlumal commited on
Commit
f624145
1 Parent(s): 06d695d

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +29 -0
README.md CHANGED
@@ -7,3 +7,32 @@ tags:
7
  - legal
8
  - australia
9
  ---
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
7
  - legal
8
  - australia
9
  ---
10
+ # AusLegalQA
11
+
12
+ AusLegalQA is a fine-tune of [Mistral-8x7B-Instruct-0.1](https://huggingface.co/mistralai/Mixtral-8x7B-Instruct-v0.1) using PEFT techniques, trained on the [Open Australian Legal Corpus](https://huggingface.co/datasets/umarbutler/open-australian-legal-corpus).
13
+
14
+ The model achieved an eval loss of 1.1391 on a subset of 100 prompts and answers from the original dataset.
15
+
16
+ The model was trained with the following hyperparameters for 3 epochs. The epoch with the lowest eval loss was selected (coinciding with end of epoch 2) and the resulting qLoRA (4 bits) was merged into the base model.
17
+ | Hyperparameter | Value |
18
+ | --- | --- |
19
+ | Sequence length | 1024 |
20
+ | Epochs | 2 |
21
+ | Optimiser | AdamW |
22
+ | Learning rate | 1e-4 |
23
+ | Learning rate scheduler | Cosine |
24
+ | Batch size | 1 |
25
+ | Weight decay | 0.01 |
26
+ | Warmup ratio | 0.05 |
27
+ | LoRA rank | 64 |
28
+ | LoRA alpha | 128 |
29
+ | LoRA dropout | 0.1 |
30
+ | LoRA target | q_proj,v_proj |
31
+ | NEFTune alpha | 5 |
32
+ | Flash Attention | on |
33
+
34
+ ## Strengths
35
+ The model is strong at summarisation and short-form answers with the key details. It is more likely to provide responses which assume the user is located in Australia. Ideal use-case is in a LLamaIndex/LangChain environment.
36
+
37
+ ## Limitations
38
+ Just as the base model it does not have any moderation mechanisms.