Update README.md
Browse files
README.md
CHANGED
|
@@ -17,24 +17,23 @@ tags:
|
|
| 17 |
- audioprocessing
|
| 18 |
- transformer
|
| 19 |
---
|
| 20 |
-
#
|
| 21 |
|
| 22 |
<!-- Provide a quick summary of what the model is/does. -->
|
| 23 |
|
| 24 |
-
Conformer is a popular Transformer based speech recognition network, suitable for embedded devices. This repository contains FP32 trained weights and the associated tokenizer for an implementation of Conformer. We also include exported quantized program with ExecuTorch, quantized for the ExecuTorch Ethos-U backend allowing an easy deployment on SoCs with an Arm® Ethos™-U NPU.
|
| 25 |
## Model Details
|
| 26 |
|
| 27 |
### Model Description
|
| 28 |
|
| 29 |
Conformer is a popular Neural Network for speech recognition. This repository contains trained weights for the Conformer implementation in https://github.com/sooftware/conformer/
|
| 30 |
|
| 31 |
-
|
| 32 |
- **Developed by:** Arm
|
| 33 |
- **Model type:** Transformer
|
| 34 |
- **Language(s) (NLP):** English
|
| 35 |
- **License:** BigScience OpenRAIL-M v1.1
|
| 36 |
|
| 37 |
-
### Model Sources
|
| 38 |
|
| 39 |
<!-- Provide the basic links for the model. -->
|
| 40 |
|
|
@@ -116,7 +115,7 @@ We used the LibriSpeech 960h dataset. The dataset is composed of 460h of clean a
|
|
| 116 |
If you want to train the Conformer model from scratch, you can do so by following the instructions in https://github.com/Arm-Examples/ML-examples/tree/main/pytorch-conformer-train-quantize/training
|
| 117 |
We used an AWS g5.24xlarge instance to train the NN.
|
| 118 |
|
| 119 |
-
#### Preprocessing
|
| 120 |
|
| 121 |
We first train a tokenizer on the Librispeech dataset. The tokenizer converts labels into tokens. For example, in English, it is very common to have 's at the end of words, the tokenizer will identify that patten and have a dedicated token for the 's combination.
|
| 122 |
The code to obtain the tokenizer is available in https://github.com/Arm-Examples/ML-examples/blob/main/pytorch-conformer-train-quantize/training/build_sp_128_librispeech.py . The trained tokenizer is also available in the Hugging Face repository.
|
|
@@ -136,12 +135,6 @@ recommended by the paper and 512 FFTs.
|
|
| 136 |
- **Warmup-epochs:** 2.0
|
| 137 |
|
| 138 |
|
| 139 |
-
|
| 140 |
-
|
| 141 |
-
<!-- This section describes the evaluation protocols and provides the results. -->
|
| 142 |
-
|
| 143 |
-
### Testing Data, Factors & Metrics
|
| 144 |
-
|
| 145 |
-
#### Testing Data
|
| 146 |
|
| 147 |
-
We test the model on the LibriSpeech `test-clean` dataset and obtain 7% Word Error Rate.
|
|
|
|
| 17 |
- audioprocessing
|
| 18 |
- transformer
|
| 19 |
---
|
| 20 |
+
# Arm ExecuTorch Conformer
|
| 21 |
|
| 22 |
<!-- Provide a quick summary of what the model is/does. -->
|
| 23 |
|
| 24 |
+
Conformer is a popular Transformer based speech recognition network, suitable for embedded devices. This repository contains an example of FP32 trained weights and the associated tokenizer for an implementation of Conformer. We also include exported quantized program with ExecuTorch, quantized for the ExecuTorch Ethos-U backend allowing an easy deployment on SoCs with an Arm® Ethos™-U NPU.
|
| 25 |
## Model Details
|
| 26 |
|
| 27 |
### Model Description
|
| 28 |
|
| 29 |
Conformer is a popular Neural Network for speech recognition. This repository contains trained weights for the Conformer implementation in https://github.com/sooftware/conformer/
|
| 30 |
|
|
|
|
| 31 |
- **Developed by:** Arm
|
| 32 |
- **Model type:** Transformer
|
| 33 |
- **Language(s) (NLP):** English
|
| 34 |
- **License:** BigScience OpenRAIL-M v1.1
|
| 35 |
|
| 36 |
+
### Model Sources
|
| 37 |
|
| 38 |
<!-- Provide the basic links for the model. -->
|
| 39 |
|
|
|
|
| 115 |
If you want to train the Conformer model from scratch, you can do so by following the instructions in https://github.com/Arm-Examples/ML-examples/tree/main/pytorch-conformer-train-quantize/training
|
| 116 |
We used an AWS g5.24xlarge instance to train the NN.
|
| 117 |
|
| 118 |
+
#### Preprocessing
|
| 119 |
|
| 120 |
We first train a tokenizer on the Librispeech dataset. The tokenizer converts labels into tokens. For example, in English, it is very common to have 's at the end of words, the tokenizer will identify that patten and have a dedicated token for the 's combination.
|
| 121 |
The code to obtain the tokenizer is available in https://github.com/Arm-Examples/ML-examples/blob/main/pytorch-conformer-train-quantize/training/build_sp_128_librispeech.py . The trained tokenizer is also available in the Hugging Face repository.
|
|
|
|
| 135 |
- **Warmup-epochs:** 2.0
|
| 136 |
|
| 137 |
|
| 138 |
+
### Testing Data
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 139 |
|
| 140 |
+
We test the model on the LibriSpeech `test-clean` dataset and obtain 7% Word Error Rate. The accuracy of the model may be improved through training with additional datasets, and through data augmentation techniques such as time slicing.
|