Update README.md
Browse files
README.md
CHANGED
@@ -26,15 +26,13 @@ widget:
|
|
26 |
- src: https://datasets-server.huggingface.co/assets/google/fleurs/--/nb_no/train/4/audio/audio.mp3
|
27 |
example_title: FLEURS sample 2
|
28 |
---
|
29 |
-
# Finetuned Verbatim model.
|
30 |
|
31 |
-
This model is trained 200 additional steps on top of the model below. This makes it outputting only text in lowercase and without punctation. It is also considerably more verbatim, and will not make any attempt at correcting grammatical errors in the text
|
32 |
|
33 |
-
# NB-Whisper Small
|
34 |
|
35 |
**IMPORTANT:** These models are currently Release Candidates. We are in the final stages of testing. If everything proceeds smoothly, we plan to officially release the models later this month.
|
36 |
|
37 |
-
Introducing the **_Norwegian NB-Whisper Small
|
38 |
|
39 |
| Model Size | Parameters | Model |
|
40 |
|------------|------------|------------|
|
@@ -96,7 +94,7 @@ After this is done, you should be able to run this in Python:
|
|
96 |
from transformers import pipeline
|
97 |
|
98 |
# Load the model
|
99 |
-
asr = pipeline("automatic-speech-recognition", "NbAiLabBeta/nb-whisper-
|
100 |
|
101 |
#transcribe
|
102 |
asr("king.mp3", generate_kwargs={'task': 'transcribe', 'language': 'no'})
|
@@ -252,7 +250,7 @@ wget -N https://github.com/NbAiLab/nb-whisper/raw/main/audio/knuthamsun.mp3
|
|
252 |
pip uninstall whisperx && pip install git+https://github.com/m-bain/whisperx.git@8540ff5985fceee764acbed94f656063d7f56540
|
253 |
|
254 |
# Transcribe the test file. All transcripts will end up in the directory of the mp3-file
|
255 |
-
whisperx knuthamsun.mp3 --model NbAiLabBeta/nb-whisper-
|
256 |
|
257 |
```
|
258 |
|
@@ -284,7 +282,7 @@ Using these models without adequate risk assessment and mitigation could be cons
|
|
284 |
The model was trained using Jax/Flax and converted to PyTorch, Tensorflow, whisper.cpp, and ONXX formats. These are available under `Files and versions`. We welcome requests for conversion to other formats. All training code and scripts are released under the Apache License 2.0 in the GitHub repository [nb-whisper](https://github.com/NbAiLab/nb-whisper/).
|
285 |
|
286 |
## Citation & Contributors
|
287 |
-
The NB-Whisper Small
|
288 |
|
289 |
## Disclaimer
|
290 |
|
|
|
26 |
- src: https://datasets-server.huggingface.co/assets/google/fleurs/--/nb_no/train/4/audio/audio.mp3
|
27 |
example_title: FLEURS sample 2
|
28 |
---
|
|
|
29 |
|
|
|
30 |
|
31 |
+
# NB-Whisper Small (Release Candidate)
|
32 |
|
33 |
**IMPORTANT:** These models are currently Release Candidates. We are in the final stages of testing. If everything proceeds smoothly, we plan to officially release the models later this month.
|
34 |
|
35 |
+
Introducing the **_Norwegian NB-Whisper Small model_**, proudly developed by the National Library of Norway. NB-Whisper is a cutting-edge series of models designed for automatic speech recognition (ASR) and speech translation. These models are based on the work of [OpenAI's Whisper](https://arxiv.org/abs/2212.04356). Each model in the series has been trained for 250,000 steps, utilizing a diverse dataset of 8 million samples. These samples consist of aligned audio clips, each 30 seconds long, culminating in a staggering 66,000 hours of speech. For an in-depth understanding of our training methodology and dataset composition, keep an eye out for our upcoming article.
|
36 |
|
37 |
| Model Size | Parameters | Model |
|
38 |
|------------|------------|------------|
|
|
|
94 |
from transformers import pipeline
|
95 |
|
96 |
# Load the model
|
97 |
+
asr = pipeline("automatic-speech-recognition", "NbAiLabBeta/nb-whisper-small")
|
98 |
|
99 |
#transcribe
|
100 |
asr("king.mp3", generate_kwargs={'task': 'transcribe', 'language': 'no'})
|
|
|
250 |
pip uninstall whisperx && pip install git+https://github.com/m-bain/whisperx.git@8540ff5985fceee764acbed94f656063d7f56540
|
251 |
|
252 |
# Transcribe the test file. All transcripts will end up in the directory of the mp3-file
|
253 |
+
whisperx knuthamsun.mp3 --model NbAiLabBeta/nb-whisper-small --language no --diarize
|
254 |
|
255 |
```
|
256 |
|
|
|
282 |
The model was trained using Jax/Flax and converted to PyTorch, Tensorflow, whisper.cpp, and ONXX formats. These are available under `Files and versions`. We welcome requests for conversion to other formats. All training code and scripts are released under the Apache License 2.0 in the GitHub repository [nb-whisper](https://github.com/NbAiLab/nb-whisper/).
|
283 |
|
284 |
## Citation & Contributors
|
285 |
+
The NB-Whisper Small model is a product of the NoSTram project led by Per Egil Kummervold ([@pere](https://huggingface.co/pere)) at the National Library of Norway. Key contributors include Javier de la Rosa ([@versae](https://huggingface.co/versae)), Freddy Wetjen ([@freddyw](https://huggingface.co/freddyw)), and Rolv-Arild Braaten ([@Rolv-Arild](https://huggingface.co/Rolv-Arild)). NB AI-Lab, under the direction of Svein Arne Brygfjeld ([@Brygfjeld](https://huggingface.co/Brygfjeld)), supported the project's successful completion. A detailed paper on our process and findings is forthcoming.
|
286 |
|
287 |
## Disclaimer
|
288 |
|