namratanwani
commited on
Commit
•
7088ce0
1
Parent(s):
8e69de9
Update README.md
Browse files
README.md
CHANGED
@@ -2,16 +2,30 @@
|
|
2 |
library_name: transformers
|
3 |
tags:
|
4 |
- unsloth
|
|
|
|
|
|
|
|
|
|
|
5 |
---
|
6 |
|
7 |
# Model Card for Model ID
|
8 |
|
9 |
-
|
|
|
|
|
|
|
|
|
10 |
|
11 |
|
12 |
|
13 |
## Model Details
|
14 |
|
|
|
|
|
|
|
|
|
|
|
15 |
### Model Description
|
16 |
|
17 |
<!-- Provide a longer summary of what this model is. -->
|
@@ -93,8 +107,33 @@ Use the code below to get started with the model.
|
|
93 |
|
94 |
#### Training Hyperparameters
|
95 |
|
96 |
-
- **Training regime:**
|
97 |
-
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
98 |
#### Speeds, Sizes, Times [optional]
|
99 |
|
100 |
<!-- This section provides information about throughput, start/end time, checkpoint size if relevant, etc. -->
|
@@ -145,9 +184,9 @@ Use the code below to get started with the model.
|
|
145 |
|
146 |
Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700).
|
147 |
|
148 |
-
- **Hardware Type:**
|
149 |
- **Hours used:** [More Information Needed]
|
150 |
-
- **Cloud Provider:**
|
151 |
- **Compute Region:** [More Information Needed]
|
152 |
- **Carbon Emitted:** [More Information Needed]
|
153 |
|
@@ -171,7 +210,7 @@ Carbon emissions can be estimated using the [Machine Learning Impact calculator]
|
|
171 |
|
172 |
## Citation [optional]
|
173 |
|
174 |
-
|
175 |
|
176 |
**BibTeX:**
|
177 |
|
|
|
2 |
library_name: transformers
|
3 |
tags:
|
4 |
- unsloth
|
5 |
+
- information-extraction
|
6 |
+
- llama3
|
7 |
+
- finetuning
|
8 |
+
language:
|
9 |
+
- en
|
10 |
---
|
11 |
|
12 |
# Model Card for Model ID
|
13 |
|
14 |
+
The model extracts the triplet from the given text.
|
15 |
+
Example:
|
16 |
+
Input: "Nie Haisheng, born on October 13, 1964, worked as a fighter pilot."
|
17 |
+
Output: {'mtriple_set': [['Nie_Haisheng | birthDate | 1964-10-13',
|
18 |
+
'Nie_Haisheng | occupation | Fighter_pilot']]}
|
19 |
|
20 |
|
21 |
|
22 |
## Model Details
|
23 |
|
24 |
+
Base Model: Llama 3 - 8B
|
25 |
+
Qunatisation: 4 bit
|
26 |
+
LoRA rank: 16
|
27 |
+
|
28 |
+
|
29 |
### Model Description
|
30 |
|
31 |
<!-- Provide a longer summary of what this model is. -->
|
|
|
107 |
|
108 |
#### Training Hyperparameters
|
109 |
|
110 |
+
- **Training regime:**
|
111 |
+
|
112 |
+
max_seq_length = 2000
|
113 |
+
trainer = SFTTrainer(
|
114 |
+
model = model,
|
115 |
+
tokenizer = tokenizer,
|
116 |
+
train_dataset = train,
|
117 |
+
dataset_text_field = "text",
|
118 |
+
max_seq_length = max_seq_length,
|
119 |
+
dataset_num_proc = 2,
|
120 |
+
packing = False, # Can make training 5x faster for short sequences.
|
121 |
+
args = TrainingArguments(
|
122 |
+
per_device_train_batch_size = 2,
|
123 |
+
gradient_accumulation_steps = 4,
|
124 |
+
warmup_steps = 5,
|
125 |
+
max_steps = 50,
|
126 |
+
learning_rate = 2e-4,
|
127 |
+
fp16 = not is_bfloat16_supported(),
|
128 |
+
bf16 = is_bfloat16_supported(),
|
129 |
+
logging_steps = 1,
|
130 |
+
optim = "adamw_8bit",
|
131 |
+
weight_decay = 0.01,
|
132 |
+
lr_scheduler_type = "linear",
|
133 |
+
seed = 3407,
|
134 |
+
output_dir = "outputs",
|
135 |
+
),
|
136 |
+
)
|
137 |
#### Speeds, Sizes, Times [optional]
|
138 |
|
139 |
<!-- This section provides information about throughput, start/end time, checkpoint size if relevant, etc. -->
|
|
|
184 |
|
185 |
Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700).
|
186 |
|
187 |
+
- **Hardware Type:** 1 T4 GPU, RAM: 16 GB
|
188 |
- **Hours used:** [More Information Needed]
|
189 |
+
- **Cloud Provider:** Google CoLab
|
190 |
- **Compute Region:** [More Information Needed]
|
191 |
- **Carbon Emitted:** [More Information Needed]
|
192 |
|
|
|
210 |
|
211 |
## Citation [optional]
|
212 |
|
213 |
+
https://colab.research.google.com/drive/135ced7oHytdxu3N2DNe1Z0kqjyYIkDXp?usp=sharing#scrollTo=kR3gIAX-SM2q
|
214 |
|
215 |
**BibTeX:**
|
216 |
|