Mahalingam commited on
Commit
c8a6e0d
1 Parent(s): 2a65213

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +76 -1
README.md CHANGED
@@ -132,4 +132,79 @@ model-index:
132
  name: gen_len
133
  verified: true
134
  verifyToken: eyJhbGciOiJFZERTQSIsInR5cCI6IkpXVCJ9.eyJoYXNoIjoiOTdhNGYxYWRlNTU0MzAxMWU1NzNmMTBjMmY3NzkzODAyYTMzZWYyZmNiMTViMzNmYTE0ZmFmNDdhMzQwMmJkNyIsInZlcnNpb24iOjF9.8lm84JtbCh-diuNQ01oXK6P8vV9CPyA8y-7D9o_OHb9Vk3pNEFM1jMSZVdEG9wFuMpWL3ARbXLadEPQB5HN8AQ
135
- ---
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
132
  name: gen_len
133
  verified: true
134
  verifyToken: eyJhbGciOiJFZERTQSIsInR5cCI6IkpXVCJ9.eyJoYXNoIjoiOTdhNGYxYWRlNTU0MzAxMWU1NzNmMTBjMmY3NzkzODAyYTMzZWYyZmNiMTViMzNmYTE0ZmFmNDdhMzQwMmJkNyIsInZlcnNpb24iOjF9.8lm84JtbCh-diuNQ01oXK6P8vV9CPyA8y-7D9o_OHb9Vk3pNEFM1jMSZVdEG9wFuMpWL3ARbXLadEPQB5HN8AQ
135
+ ---
136
+
137
+ ## `distilbart-cnn-12-6-samsum`
138
+
139
+ This model was trained using Amazon SageMaker and the new Hugging Face Deep Learning container.
140
+
141
+ For more information look at:
142
+ - [🤗 Transformers Documentation: Amazon SageMaker](https://huggingface.co/transformers/sagemaker.html)
143
+ - [Example Notebooks](https://github.com/huggingface/notebooks/tree/master/sagemaker)
144
+ - [Amazon SageMaker documentation for Hugging Face](https://docs.aws.amazon.com/sagemaker/latest/dg/hugging-face.html)
145
+ - [Python SDK SageMaker documentation for Hugging Face](https://sagemaker.readthedocs.io/en/stable/frameworks/huggingface/index.html)
146
+ - [Deep Learning Container](https://github.com/aws/deep-learning-containers/blob/master/available_images.md#huggingface-training-containers)
147
+
148
+ ## Hyperparameters
149
+ ```json
150
+ {
151
+ "dataset_name": "samsum",
152
+ "do_eval": true,
153
+ "do_train": true,
154
+ "fp16": true,
155
+ "learning_rate": 5e-05,
156
+ "model_name_or_path": "sshleifer/distilbart-cnn-12-6",
157
+ "num_train_epochs": 3,
158
+ "output_dir": "/opt/ml/model",
159
+ "per_device_eval_batch_size": 8,
160
+ "per_device_train_batch_size": 8,
161
+ "seed": 7
162
+ }
163
+ ```
164
+
165
+ ## Train results
166
+
167
+ | key | value |
168
+ | --- | ----- |
169
+ | epoch | 3.0 |
170
+ | init_mem_cpu_alloc_delta | 180338 |
171
+ | init_mem_cpu_peaked_delta | 18282 |
172
+ | init_mem_gpu_alloc_delta | 1222242816 |
173
+ | init_mem_gpu_peaked_delta | 0 |
174
+ | train_mem_cpu_alloc_delta | 6971403 |
175
+ | train_mem_cpu_peaked_delta | 640733 |
176
+ | train_mem_gpu_alloc_delta | 4910897664 |
177
+ | train_mem_gpu_peaked_delta | 23331969536 |
178
+ | train_runtime | 155.2034 |
179
+ | train_samples | 14732 |
180
+ | train_samples_per_second | 2.242 |
181
+
182
+ ## Eval results
183
+
184
+ | key | value |
185
+ | --- | ----- |
186
+ | epoch | 3.0 |
187
+ | eval_loss | 1.4209576845169067 |
188
+ | eval_mem_cpu_alloc_delta | 868003 |
189
+ | eval_mem_cpu_peaked_delta | 18250 |
190
+ | eval_mem_gpu_alloc_delta | 0 |
191
+ | eval_mem_gpu_peaked_delta | 328244736 |
192
+ | eval_runtime | 0.6088 |
193
+ | eval_samples | 818 |
194
+ | eval_samples_per_second | 1343.647 |
195
+
196
+
197
+ ## Usage
198
+ ```python
199
+ from transformers import pipeline
200
+ summarizer = pipeline("summarization", model="philschmid/distilbart-cnn-12-6-samsum")
201
+
202
+ conversation = '''Jeff: Can I train a 🤗 Transformers model on Amazon SageMaker?
203
+ Philipp: Sure you can use the new Hugging Face Deep Learning Container.
204
+ Jeff: ok.
205
+ Jeff: and how can I get started?
206
+ Jeff: where can I find documentation?
207
+ Philipp: ok, ok you can find everything here. https://huggingface.co/blog/the-partnership-amazon-sagemaker-and-hugging-face
208
+ '''
209
+ nlp(conversation)
210
+ ```