8bitkick commited on
Commit
ebd4025
·
verified ·
1 Parent(s): e54e045

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +6 -13
README.md CHANGED
@@ -17,24 +17,23 @@ tags:
17
  - audioprocessing
18
  - transformer
19
  ---
20
- # Model Card for Model ID
21
 
22
  <!-- Provide a quick summary of what the model is/does. -->
23
 
24
- Conformer is a popular Transformer based speech recognition network, suitable for embedded devices. This repository contains FP32 trained weights and the associated tokenizer for an implementation of Conformer. We also include exported quantized program with ExecuTorch, quantized for the ExecuTorch Ethos-U backend allowing an easy deployment on SoCs with an Arm® Ethos™-U NPU.
25
  ## Model Details
26
 
27
  ### Model Description
28
 
29
  Conformer is a popular Neural Network for speech recognition. This repository contains trained weights for the Conformer implementation in https://github.com/sooftware/conformer/
30
 
31
-
32
  - **Developed by:** Arm
33
  - **Model type:** Transformer
34
  - **Language(s) (NLP):** English
35
  - **License:** BigScience OpenRAIL-M v1.1
36
 
37
- ### Model Sources [optional]
38
 
39
  <!-- Provide the basic links for the model. -->
40
 
@@ -116,7 +115,7 @@ We used the LibriSpeech 960h dataset. The dataset is composed of 460h of clean a
116
  If you want to train the Conformer model from scratch, you can do so by following the instructions in https://github.com/Arm-Examples/ML-examples/tree/main/pytorch-conformer-train-quantize/training
117
  We used an AWS g5.24xlarge instance to train the NN.
118
 
119
- #### Preprocessing [optional]
120
 
121
  We first train a tokenizer on the Librispeech dataset. The tokenizer converts labels into tokens. For example, in English, it is very common to have 's at the end of words, the tokenizer will identify that patten and have a dedicated token for the 's combination.
122
  The code to obtain the tokenizer is available in https://github.com/Arm-Examples/ML-examples/blob/main/pytorch-conformer-train-quantize/training/build_sp_128_librispeech.py . The trained tokenizer is also available in the Hugging Face repository.
@@ -136,12 +135,6 @@ recommended by the paper and 512 FFTs.
136
  - **Warmup-epochs:** 2.0
137
 
138
 
139
- ## Evaluation
140
-
141
- <!-- This section describes the evaluation protocols and provides the results. -->
142
-
143
- ### Testing Data, Factors & Metrics
144
-
145
- #### Testing Data
146
 
147
- We test the model on the LibriSpeech `test-clean` dataset and obtain 7% Word Error Rate.
 
17
  - audioprocessing
18
  - transformer
19
  ---
20
+ # Arm ExecuTorch Conformer
21
 
22
  <!-- Provide a quick summary of what the model is/does. -->
23
 
24
+ Conformer is a popular Transformer based speech recognition network, suitable for embedded devices. This repository contains an example of FP32 trained weights and the associated tokenizer for an implementation of Conformer. We also include exported quantized program with ExecuTorch, quantized for the ExecuTorch Ethos-U backend allowing an easy deployment on SoCs with an Arm® Ethos™-U NPU.
25
  ## Model Details
26
 
27
  ### Model Description
28
 
29
  Conformer is a popular Neural Network for speech recognition. This repository contains trained weights for the Conformer implementation in https://github.com/sooftware/conformer/
30
 
 
31
  - **Developed by:** Arm
32
  - **Model type:** Transformer
33
  - **Language(s) (NLP):** English
34
  - **License:** BigScience OpenRAIL-M v1.1
35
 
36
+ ### Model Sources
37
 
38
  <!-- Provide the basic links for the model. -->
39
 
 
115
  If you want to train the Conformer model from scratch, you can do so by following the instructions in https://github.com/Arm-Examples/ML-examples/tree/main/pytorch-conformer-train-quantize/training
116
  We used an AWS g5.24xlarge instance to train the NN.
117
 
118
+ #### Preprocessing
119
 
120
  We first train a tokenizer on the Librispeech dataset. The tokenizer converts labels into tokens. For example, in English, it is very common to have 's at the end of words, the tokenizer will identify that patten and have a dedicated token for the 's combination.
121
  The code to obtain the tokenizer is available in https://github.com/Arm-Examples/ML-examples/blob/main/pytorch-conformer-train-quantize/training/build_sp_128_librispeech.py . The trained tokenizer is also available in the Hugging Face repository.
 
135
  - **Warmup-epochs:** 2.0
136
 
137
 
138
+ ### Testing Data
 
 
 
 
 
 
139
 
140
+ We test the model on the LibriSpeech `test-clean` dataset and obtain 7% Word Error Rate. The accuracy of the model may be improved through training with additional datasets, and through data augmentation techniques such as time slicing.