Skip to content

Commit 0c19911

Browse files
committed
docs: update conformer transducer/ctc results
1 parent edcb7db commit 0c19911

File tree

10 files changed

+81
-96
lines changed

10 files changed

+81
-96
lines changed
Lines changed: 28 additions & 74 deletions
Original file line numberDiff line numberDiff line change
@@ -1,23 +1,7 @@
1-
**Table of Contents**
2-
- [LibriSpeech](#librispeech)
1+
- [\[English\] LibriSpeech](#english-librispeech)
32
- [I. Small + SentencePiece 256](#i-small--sentencepiece-256)
4-
- [Training](#training)
5-
- [1. Epoch Loss](#1-epoch-loss)
6-
- [2. Batch Loss](#2-batch-loss)
7-
- [3. Learning Rate](#3-learning-rate)
8-
- [Pretrained Model](#pretrained-model)
9-
- [Results](#results)
10-
- [VietBud500](#vietbud500)
11-
- [I. Small + SentencePiece 256](#i-small--sentencepiece-256-1)
12-
- [Training](#training-1)
13-
- [1. Epoch Loss](#1-epoch-loss-1)
14-
- [2. Batch Loss](#2-batch-loss-1)
15-
- [3. Learning Rate](#3-learning-rate-1)
16-
- [Pretrained Model](#pretrained-model-1)
17-
- [Results](#results-1)
183

19-
20-
# LibriSpeech
4+
# [English] LibriSpeech
215

226
## I. Small + SentencePiece 256
237

@@ -30,75 +14,45 @@
3014
| Global Batch Size | 8 * 4 * 8 = 256 (as 4 TPUs, 8 Gradient Accumulation Steps) |
3115
| Max Epochs | 450 |
3216

17+
**Config:**
3318

34-
### Training
35-
36-
#### 1. Epoch Loss
37-
38-
![Epoch Loss](./figs/)
39-
40-
#### 2. Batch Loss
41-
42-
![Batch Loss](./figs/)
43-
44-
#### 3. Learning Rate
45-
46-
![Learning Rate](./figs/)
47-
48-
### Pretrained Model
49-
50-
[Link]()
51-
52-
### Results
53-
54-
55-
```json
56-
[
57-
{
58-
"epoch": 157,
59-
"test-clean": {
60-
},
61-
"test-other": {
62-
}
63-
}
64-
]
19+
```jinja2
20+
{% import "examples/datasets/librispeech/sentencepiece/sp.256.yml.j2" as decoder_config with context %}
21+
{{decoder_config}}
22+
{% import "examples/models/ctc/conformer/small.yml.j2" as config with context %}
23+
{{config}}
6524
```
6625

67-
# VietBud500
26+
**Results:**
6827

69-
## I. Small + SentencePiece 256
28+
| Epoch | Dataset | decoding | wer | cer | mer | wil | wip |
29+
| :---- | :--------- | :------- | :-------- | :-------- | :-------- | :------- | :------- |
30+
| 170 | test-clean | greedy | 0.0967171 | 0.031954 | 0.0958403 | 0.168307 | 0.831693 |
31+
| 170 | test-other | greedy | 0.201612 | 0.0812955 | 0.197415 | 0.330207 | 0.669793 |
32+
33+
<!--
34+
## II. Small + Streaming + SentencePiece 256
7035
7136
| Category | Description |
7237
| :---------------- | :--------------------------------------------------------- |
73-
| Config | [small.yml.j2](../../small.yml.j2) |
38+
| Config | [small-streaming.yml.j2](../../small-streaming.yml.j2) |
7439
| Tensorflow | **2.18.0** |
7540
| Device | Google Cloud TPUs v4-8 |
7641
| Mixed Precision | strict |
7742
| Global Batch Size | 8 * 4 * 8 = 256 (as 4 TPUs, 8 Gradient Accumulation Steps) |
7843
| Max Epochs | 450 |
7944
80-
### Training
81-
82-
#### 1. Epoch Loss
83-
84-
![Epoch Loss](./figs/)
85-
86-
#### 2. Batch Loss
45+
**Config:**
8746
88-
![Batch Loss](./figs/)
89-
90-
#### 3. Learning Rate
91-
92-
![Learning Rate](./figs/)
93-
94-
### Pretrained Model
95-
96-
[Link]()
97-
98-
### Results
47+
```jinja2
48+
{% import "examples/datasets/librispeech/sentencepiece/sp.256.yml.j2" as decoder_config with context %}
49+
{{decoder_config}}
50+
{% import "examples/models/ctc/conformer/small-streaming.yml.j2" as config with context %}
51+
{{config}}
52+
```
9953
100-
```json
101-
[
54+
**Results:**
10255
103-
]
104-
```
56+
| Epoch | Dataset | decoding | wer | cer | mer | wil | wip |
57+
| :---- | :------ | :------- | :--- | :--- | :--- | :--- | :--- |
58+
-->

examples/models/transducer/conformer/results/sentencepiece/README.md

Lines changed: 53 additions & 17 deletions
Original file line numberDiff line numberDiff line change
@@ -1,12 +1,8 @@
1-
**Table of Contents**
21
- [\[English\] LibriSpeech](#english-librispeech)
32
- [I. Small + SentencePiece 1k](#i-small--sentencepiece-1k)
4-
- [Config](#config)
5-
- [Results](#results)
6-
- [\[Vietnamese\] VietBud500](#vietnamese-vietbud500)
73
- [II. Small + Streaming + SentencePiece 1k](#ii-small--streaming--sentencepiece-1k)
8-
- [Config](#config-1)
9-
- [Results](#results-1)
4+
- [\[Vietnamese\] VietBud500](#vietnamese-vietbud500)
5+
- [I. Small + Streaming + SentencePiece 1k](#i-small--streaming--sentencepiece-1k)
106

117
<!-- ----------------------------------------------------- EN ------------------------------------------------------ -->
128

@@ -23,7 +19,7 @@
2319
| Global Batch Size | 4 * 4 * 8 = 128 (as 4 TPUs, 8 Gradient Accumulation Steps) |
2420
| Max Epochs | 300 |
2521

26-
### Config
22+
**Config:**
2723

2824
```jinja2
2925
{% import "examples/datasets/librispeech/sentencepiece/sp.yml.j2" as decoder_config with context %}
@@ -32,18 +28,45 @@
3228
{{config}}
3329
```
3430

35-
### Results
31+
**Results:**
3632

3733
| Epoch | Dataset | decoding | wer | cer | mer | wil | wip |
3834
| :---- | :--------- | :------- | :------- | :------- | :------- | :------- | :------- |
3935
| 157 | test-clean | greedy | 0.062918 | 0.025361 | 0.062527 | 0.109992 | 0.890007 |
4036
| 157 | test-other | greedy | 0.142616 | 0.066839 | 0.140610 | 0.239201 | 0.760798 |
4137

38+
## II. Small + Streaming + SentencePiece 1k
39+
40+
| Category | Description |
41+
| :---------------- | :--------------------------------------------------------- |
42+
| Config | [small-streaming.yml.j2](../../small-streaming.yml.j2) |
43+
| Tensorflow | **2.18.0** |
44+
| Device | Google Cloud TPUs v4-8 |
45+
| Mixed Precision | strict |
46+
| Global Batch Size | 4 * 4 * 8 = 128 (as 4 TPUs, 8 Gradient Accumulation Steps) |
47+
| Max Epochs | 300 |
48+
49+
**Config:**
50+
51+
```jinja2
52+
{% import "examples/datasets/librispeech/sentencepiece/sp.yml.j2" as decoder_config with context %}
53+
{{decoder_config}}
54+
{% import "examples/models/transducer/conformer/small-streaming.yml.j2" as config with context %}
55+
{{config}}
56+
```
57+
58+
**Results:**
59+
60+
| Epoch | Dataset | decoding | wer | cer | mer | wil | wip |
61+
| :---- | :--------- | :------- | :------- | :-------- | :------- | :------- | :------- |
62+
| 45 | test-clean | greedy | 0.110564 | 0.0460022 | 0.109064 | 0.186109 | 0.813891 |
63+
| 45 | test-other | greedy | 0.267772 | 0.139369 | 0.260952 | 0.417361 | 0.582639 |
64+
4265
<!-- ----------------------------------------------------- VN ------------------------------------------------------ -->
4366

4467
# [Vietnamese] VietBud500
4568

46-
## II. Small + Streaming + SentencePiece 1k
69+
## I. Small + Streaming + SentencePiece 1k
4770

4871
| Category | Description |
4972
| :---------------- | :--------------------------------------------------------- |
@@ -54,7 +77,7 @@
5477
| Global Batch Size | 8 * 4 * 8 = 256 (as 4 TPUs, 8 Gradient Accumulation Steps) |
5578
| Max Epochs | 300 |
5679

57-
### Config
80+
**Config:**
5881

5982
```jinja2
6083
{% import "examples/datasets/vietbud500/sentencepiece/sp.yml.j2" as decoder_config with context %}
@@ -63,13 +86,26 @@
6386
{{config}}
6487
```
6588

66-
### Results
67-
68-
| Training | Image |
69-
| :------------ | :-------------------------------------------------------------- |
70-
| Epoch Loss | ![Epoch Loss](./figs/vietbud500-small-streaming-epoch-loss.svg) |
71-
| Batch Loss | ![Batch Loss](./figs/vietbud500-small-streaming-batch-loss.svg) |
72-
| Learning Rate | ![Learning Rate](./figs/vietbud500-small-streaming-lr.svg) |
89+
**Tensorboard:**
90+
91+
<table>
92+
<tr>
93+
<td align="center">
94+
<img src="./figs/vietbud500-small-streaming-epoch-loss.jpg" width="200px"><br>
95+
<sub><strong>Epoch Loss</strong></sub>
96+
</td>
97+
<td align="center">
98+
<img src="./figs/vietbud500-small-streaming-batch-loss.jpg" width="200px"><br>
99+
<sub><strong>Batch Loss</strong></sub>
100+
</td>
101+
<td align="center">
102+
<img src="./figs/vietbud500-small-streaming-lr.jpg " width="200px"><br>
103+
<sub><strong>Learning Rate</strong></sub>
104+
</td>
105+
</tr>
106+
</table>
107+
108+
**Results:**
73109

74110
| Epoch | decoding | wer | cer | mer | wil | wip |
75111
| :---- | :------- | :------- | :------- | :------ | :------- | :------- |

examples/models/transducer/conformer/results/sentencepiece/figs/vietbud500-small-epoch-loss.svg

Lines changed: 0 additions & 1 deletion
This file was deleted.

examples/models/transducer/conformer/results/sentencepiece/figs/vietbud500-small-lr.svg

Lines changed: 0 additions & 1 deletion
This file was deleted.
62 KB
Loading

examples/models/transducer/conformer/results/sentencepiece/figs/vietbud500-small-streaming-batch-loss.svg

Lines changed: 0 additions & 1 deletion
This file was deleted.
72.5 KB
Loading

examples/models/transducer/conformer/results/sentencepiece/figs/vietbud500-small-streaming-epoch-loss.svg

Lines changed: 0 additions & 1 deletion
This file was deleted.
63 KB
Loading

examples/models/transducer/conformer/results/sentencepiece/figs/vietbud500-small-streaming-lr.svg

Lines changed: 0 additions & 1 deletion
This file was deleted.

0 commit comments

Comments
 (0)