Update README.md
Browse files
README.md
CHANGED
@@ -127,7 +127,7 @@ from datasets import load_dataset
|
|
127 |
|
128 |
# config
|
129 |
model_id = "kotoba-tech/kotoba-whisper-v1.0"
|
130 |
-
torch_dtype = torch.
|
131 |
device = "cuda:0" if torch.cuda.is_available() else "cpu"
|
132 |
model_kwargs = {"attn_implementation": "sdpa"} if torch.cuda.is_available() else {}
|
133 |
generate_kwargs = {"language": "japanese", "task": "transcribe"}
|
@@ -189,7 +189,7 @@ from datasets import load_dataset
|
|
189 |
|
190 |
# config
|
191 |
model_id = "kotoba-tech/kotoba-whisper-v1.0"
|
192 |
-
torch_dtype = torch.
|
193 |
device = "cuda:0" if torch.cuda.is_available() else "cpu"
|
194 |
model_kwargs = {"attn_implementation": "sdpa"} if torch.cuda.is_available() else {}
|
195 |
generate_kwargs = {"language": "japanese", "task": "transcribe"}
|
@@ -225,7 +225,7 @@ from datasets import load_dataset
|
|
225 |
|
226 |
# config
|
227 |
model_id = "kotoba-tech/kotoba-whisper-v1.0"
|
228 |
-
torch_dtype = torch.
|
229 |
device = "cuda:0" if torch.cuda.is_available() else "cpu"
|
230 |
model_kwargs = {"attn_implementation": "sdpa"} if torch.cuda.is_available() else {}
|
231 |
generate_kwargs = {"language": "japanese", "task": "transcribe"}
|
@@ -303,7 +303,7 @@ from evaluate import load
|
|
303 |
|
304 |
# model config
|
305 |
model_id = "kotoba-tech/kotoba-whisper-v1.0"
|
306 |
-
torch_dtype = torch.
|
307 |
device = "cuda:0" if torch.cuda.is_available() else "cpu"
|
308 |
model_kwargs = {"attn_implementation": "sdpa"} if torch.cuda.is_available() else {}
|
309 |
generate_kwargs = {"language": "japanese", "task": "transcribe"}
|
|
|
127 |
|
128 |
# config
|
129 |
model_id = "kotoba-tech/kotoba-whisper-v1.0"
|
130 |
+
torch_dtype = torch.bfloat16 if torch.cuda.is_available() else torch.float32
|
131 |
device = "cuda:0" if torch.cuda.is_available() else "cpu"
|
132 |
model_kwargs = {"attn_implementation": "sdpa"} if torch.cuda.is_available() else {}
|
133 |
generate_kwargs = {"language": "japanese", "task": "transcribe"}
|
|
|
189 |
|
190 |
# config
|
191 |
model_id = "kotoba-tech/kotoba-whisper-v1.0"
|
192 |
+
torch_dtype = torch.bfloat16 if torch.cuda.is_available() else torch.float32
|
193 |
device = "cuda:0" if torch.cuda.is_available() else "cpu"
|
194 |
model_kwargs = {"attn_implementation": "sdpa"} if torch.cuda.is_available() else {}
|
195 |
generate_kwargs = {"language": "japanese", "task": "transcribe"}
|
|
|
225 |
|
226 |
# config
|
227 |
model_id = "kotoba-tech/kotoba-whisper-v1.0"
|
228 |
+
torch_dtype = torch.bfloat16 if torch.cuda.is_available() else torch.float32
|
229 |
device = "cuda:0" if torch.cuda.is_available() else "cpu"
|
230 |
model_kwargs = {"attn_implementation": "sdpa"} if torch.cuda.is_available() else {}
|
231 |
generate_kwargs = {"language": "japanese", "task": "transcribe"}
|
|
|
303 |
|
304 |
# model config
|
305 |
model_id = "kotoba-tech/kotoba-whisper-v1.0"
|
306 |
+
torch_dtype = torch.bfloat16 if torch.cuda.is_available() else torch.float32
|
307 |
device = "cuda:0" if torch.cuda.is_available() else "cpu"
|
308 |
model_kwargs = {"attn_implementation": "sdpa"} if torch.cuda.is_available() else {}
|
309 |
generate_kwargs = {"language": "japanese", "task": "transcribe"}
|