Commit
•
c7bea3d
1
Parent(s):
796a56b
candle
Browse files
README.md
CHANGED
@@ -229,18 +229,6 @@ model = AutoModelForSpeechSeq2Seq.from_pretrained(model_id, torch_dtype=torch_dt
|
|
229 |
+ model = model.to_bettertransformer()
|
230 |
```
|
231 |
|
232 |
-
### 8bit & 4bit Quantization
|
233 |
-
|
234 |
-
Coming soon ...
|
235 |
-
|
236 |
-
### Candle
|
237 |
-
|
238 |
-
Coming soon ...
|
239 |
-
|
240 |
-
### Whisper.cpp
|
241 |
-
|
242 |
-
Coming soon ...
|
243 |
-
|
244 |
### Running Whisper in `openai-whisper`
|
245 |
|
246 |
To use the model in the original Whisper format, first ensure you have the [`openai-whisper`](https://pypi.org/project/openai-whisper/) package installed:
|
@@ -290,6 +278,44 @@ let output = await transcriber(url);
|
|
290 |
See the [docs](https://huggingface.co/docs/transformers.js/api/pipelines#module_pipelines.AutomaticSpeechRecognitionPipeline) for more information.
|
291 |
|
292 |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
293 |
## Model Details
|
294 |
|
295 |
Distil-Whisper inherits the encoder-decoder architecture from Whisper. The encoder maps a sequence of speech vector
|
|
|
229 |
+ model = model.to_bettertransformer()
|
230 |
```
|
231 |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
232 |
### Running Whisper in `openai-whisper`
|
233 |
|
234 |
To use the model in the original Whisper format, first ensure you have the [`openai-whisper`](https://pypi.org/project/openai-whisper/) package installed:
|
|
|
278 |
See the [docs](https://huggingface.co/docs/transformers.js/api/pipelines#module_pipelines.AutomaticSpeechRecognitionPipeline) for more information.
|
279 |
|
280 |
|
281 |
+
|
282 |
+
### Candle
|
283 |
+
|
284 |
+
Through an integration with Hugging Face [Candle](https://github.com/huggingface/candle/tree/main) 🕯️, Distil-Whisper is
|
285 |
+
now available in the Rust library 🦀
|
286 |
+
|
287 |
+
Benefit from:
|
288 |
+
* Optimised CPU backend with optional MKL support for x86 and Accelerate for Macs
|
289 |
+
* CUDA backend for efficiently running on GPUs, multiple GPU distribution via NCCL
|
290 |
+
* WASM support: run Distil-Whisper in a browser
|
291 |
+
|
292 |
+
Steps for getting started:
|
293 |
+
1. Install [`candle-core`](https://github.com/huggingface/candle/tree/main/candle-core) as explained [here](https://huggingface.github.io/candle/guide/installation.html)
|
294 |
+
2. Clone the `candle` repository locally:
|
295 |
+
```
|
296 |
+
git clone https://github.com/huggingface/candle.git
|
297 |
+
```
|
298 |
+
3. Enter the example directory for [Whisper](https://github.com/huggingface/candle/tree/main/candle-examples/examples/whisper):
|
299 |
+
```
|
300 |
+
cd candle/candle-examples/examples/whisper
|
301 |
+
```
|
302 |
+
4. Run an example:
|
303 |
+
```
|
304 |
+
cargo run --example whisper --release -- --model distil-medium.en
|
305 |
+
```
|
306 |
+
5. To specify your own audio file, add the `--input` flag:
|
307 |
+
```
|
308 |
+
cargo run --example whisper --release -- --model distil-medium.en --input audio.wav
|
309 |
+
```
|
310 |
+
|
311 |
+
### 8bit & 4bit Quantization
|
312 |
+
|
313 |
+
Coming soon ...
|
314 |
+
|
315 |
+
### Whisper.cpp
|
316 |
+
|
317 |
+
Coming soon ...
|
318 |
+
|
319 |
## Model Details
|
320 |
|
321 |
Distil-Whisper inherits the encoder-decoder architecture from Whisper. The encoder maps a sequence of speech vector
|