namratanwani commited on
Commit
7088ce0
1 Parent(s): 8e69de9

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +45 -6
README.md CHANGED
@@ -2,16 +2,30 @@
2
  library_name: transformers
3
  tags:
4
  - unsloth
 
 
 
 
 
5
  ---
6
 
7
  # Model Card for Model ID
8
 
9
- <!-- Provide a quick summary of what the model is/does. -->
 
 
 
 
10
 
11
 
12
 
13
  ## Model Details
14
 
 
 
 
 
 
15
  ### Model Description
16
 
17
  <!-- Provide a longer summary of what this model is. -->
@@ -93,8 +107,33 @@ Use the code below to get started with the model.
93
 
94
  #### Training Hyperparameters
95
 
96
- - **Training regime:** [More Information Needed] <!--fp32, fp16 mixed precision, bf16 mixed precision, bf16 non-mixed precision, fp16 non-mixed precision, fp8 mixed precision -->
97
-
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
98
  #### Speeds, Sizes, Times [optional]
99
 
100
  <!-- This section provides information about throughput, start/end time, checkpoint size if relevant, etc. -->
@@ -145,9 +184,9 @@ Use the code below to get started with the model.
145
 
146
  Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700).
147
 
148
- - **Hardware Type:** [More Information Needed]
149
  - **Hours used:** [More Information Needed]
150
- - **Cloud Provider:** [More Information Needed]
151
  - **Compute Region:** [More Information Needed]
152
  - **Carbon Emitted:** [More Information Needed]
153
 
@@ -171,7 +210,7 @@ Carbon emissions can be estimated using the [Machine Learning Impact calculator]
171
 
172
  ## Citation [optional]
173
 
174
- <!-- If there is a paper or blog post introducing the model, the APA and Bibtex information for that should go in this section. -->
175
 
176
  **BibTeX:**
177
 
 
2
  library_name: transformers
3
  tags:
4
  - unsloth
5
+ - information-extraction
6
+ - llama3
7
+ - finetuning
8
+ language:
9
+ - en
10
  ---
11
 
12
  # Model Card for Model ID
13
 
14
+ The model extracts the triplet from the given text.
15
+ Example:
16
+ Input: "Nie Haisheng, born on October 13, 1964, worked as a fighter pilot."
17
+ Output: {'mtriple_set': [['Nie_Haisheng | birthDate | 1964-10-13',
18
+ 'Nie_Haisheng | occupation | Fighter_pilot']]}
19
 
20
 
21
 
22
  ## Model Details
23
 
24
+ Base Model: Llama 3 - 8B
25
+ Qunatisation: 4 bit
26
+ LoRA rank: 16
27
+
28
+
29
  ### Model Description
30
 
31
  <!-- Provide a longer summary of what this model is. -->
 
107
 
108
  #### Training Hyperparameters
109
 
110
+ - **Training regime:**
111
+
112
+ max_seq_length = 2000
113
+ trainer = SFTTrainer(
114
+ model = model,
115
+ tokenizer = tokenizer,
116
+ train_dataset = train,
117
+ dataset_text_field = "text",
118
+ max_seq_length = max_seq_length,
119
+ dataset_num_proc = 2,
120
+ packing = False, # Can make training 5x faster for short sequences.
121
+ args = TrainingArguments(
122
+ per_device_train_batch_size = 2,
123
+ gradient_accumulation_steps = 4,
124
+ warmup_steps = 5,
125
+ max_steps = 50,
126
+ learning_rate = 2e-4,
127
+ fp16 = not is_bfloat16_supported(),
128
+ bf16 = is_bfloat16_supported(),
129
+ logging_steps = 1,
130
+ optim = "adamw_8bit",
131
+ weight_decay = 0.01,
132
+ lr_scheduler_type = "linear",
133
+ seed = 3407,
134
+ output_dir = "outputs",
135
+ ),
136
+ )
137
  #### Speeds, Sizes, Times [optional]
138
 
139
  <!-- This section provides information about throughput, start/end time, checkpoint size if relevant, etc. -->
 
184
 
185
  Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700).
186
 
187
+ - **Hardware Type:** 1 T4 GPU, RAM: 16 GB
188
  - **Hours used:** [More Information Needed]
189
+ - **Cloud Provider:** Google CoLab
190
  - **Compute Region:** [More Information Needed]
191
  - **Carbon Emitted:** [More Information Needed]
192
 
 
210
 
211
  ## Citation [optional]
212
 
213
+ https://colab.research.google.com/drive/135ced7oHytdxu3N2DNe1Z0kqjyYIkDXp?usp=sharing#scrollTo=kR3gIAX-SM2q
214
 
215
  **BibTeX:**
216