forked from espnet/espnet
-
Notifications
You must be signed in to change notification settings - Fork 0
Commit
This commit does not belong to any branch on this repository, and may belong to a fork outside of the repository.
- Loading branch information
Showing
460 changed files
with
17,875 additions
and
630 deletions.
There are no files selected for viewing
This file contains bidirectional Unicode text that may be interpreted or compiled differently than what appears below. To review, open the file in an editor that reveals hidden Unicode characters.
Learn more about bidirectional Unicode characters
Original file line number | Diff line number | Diff line change |
---|---|---|
@@ -0,0 +1,48 @@ | ||
name: docker-builder | ||
|
||
on: | ||
pull_request: | ||
types: [closed] | ||
branches: | ||
- master | ||
paths: | ||
- 'tools/**' | ||
- setup.py | ||
|
||
jobs: | ||
docker: | ||
runs-on: ubuntu-latest | ||
if: github.event.pull_request.merged == true | ||
steps: | ||
- uses: actions/checkout@v2 | ||
|
||
- name: Set up QEMU | ||
uses: docker/setup-qemu-action@v1 | ||
|
||
- name: Set up Docker Buildx | ||
uses: docker/setup-buildx-action@v1 | ||
|
||
- name: Login to DockerHub | ||
uses: docker/login-action@v1 | ||
with: | ||
username: ${{ secrets.DOCKERHUB_USERNAME }} | ||
password: ${{ secrets.DOCKERHUB_TOKEN }} | ||
|
||
- name: Build and push CPU container | ||
run: | | ||
cd docker | ||
docker build --build-arg FROM_TAG=runtime-latest \ | ||
-f prebuilt/devel.dockerfile \ | ||
--target devel \ | ||
-t espnet/espnet:cpu-latest . | ||
docker push espnet/espnet:cpu-latest | ||
- name: Build and push GPU container | ||
run: | | ||
cd docker | ||
docker build --build-arg FROM_TAG=cuda-latest \ | ||
--build-arg CUDA_VER=11.1 \ | ||
-f prebuilt/devel.dockerfile \ | ||
--target devel \ | ||
-t espnet/espnet:gpu-latest . | ||
docker push espnet/espnet:gpu-latest |
This file contains bidirectional Unicode text that may be interpreted or compiled differently than what appears below. To review, open the file in an editor that reveals hidden Unicode characters.
Learn more about bidirectional Unicode characters
This file contains bidirectional Unicode text that may be interpreted or compiled differently than what appears below. To review, open the file in an editor that reveals hidden Unicode characters.
Learn more about bidirectional Unicode characters
This file contains bidirectional Unicode text that may be interpreted or compiled differently than what appears below. To review, open the file in an editor that reveals hidden Unicode characters.
Learn more about bidirectional Unicode characters
Empty file.
File renamed without changes.
This file contains bidirectional Unicode text that may be interpreted or compiled differently than what appears below. To review, open the file in an editor that reveals hidden Unicode characters.
Learn more about bidirectional Unicode characters
Original file line number | Diff line number | Diff line change |
---|---|---|
@@ -0,0 +1 @@ | ||
tuning/decode_pytorch_transformer.yaml |
File renamed without changes.
File renamed without changes.
This file contains bidirectional Unicode text that may be interpreted or compiled differently than what appears below. To review, open the file in an editor that reveals hidden Unicode characters.
Learn more about bidirectional Unicode characters
Original file line number | Diff line number | Diff line change |
---|---|---|
@@ -0,0 +1,8 @@ | ||
# rnnlm related | ||
layer: 2 | ||
unit: 650 | ||
opt: sgd # or adam | ||
batchsize: 64 # batch size in LM training | ||
epoch: 20 # if the data size is large, we can reduce this | ||
patience: 3 | ||
maxlen: 100 # if sentence length > lm_maxlen, lm_batchsize is automatically reduced |
File renamed without changes.
File renamed without changes.
File renamed without changes.
This file contains bidirectional Unicode text that may be interpreted or compiled differently than what appears below. To review, open the file in an editor that reveals hidden Unicode characters.
Learn more about bidirectional Unicode characters
Original file line number | Diff line number | Diff line change |
---|---|---|
@@ -0,0 +1,16 @@ | ||
process: | ||
# these three processes are a.k.a. SpecAugument | ||
- type: "time_warp" | ||
max_time_warp: 5 | ||
inplace: true | ||
mode: "PIL" | ||
- type: "freq_mask" | ||
F: 30 | ||
n_mask: 2 | ||
inplace: true | ||
replace_with_zero: false | ||
- type: "time_mask" | ||
T: 40 | ||
n_mask: 2 | ||
inplace: true | ||
replace_with_zero: false |
This file contains bidirectional Unicode text that may be interpreted or compiled differently than what appears below. To review, open the file in an editor that reveals hidden Unicode characters.
Learn more about bidirectional Unicode characters
Original file line number | Diff line number | Diff line change |
---|---|---|
@@ -0,0 +1 @@ | ||
tuning/train_pytorch_conformer_kernel15.yaml |
8 changes: 8 additions & 0 deletions
8
egs/aesrc2020/asr1/conf/tuning/decode_pytorch_transformer.yaml
This file contains bidirectional Unicode text that may be interpreted or compiled differently than what appears below. To review, open the file in an editor that reveals hidden Unicode characters.
Learn more about bidirectional Unicode characters
Original file line number | Diff line number | Diff line change |
---|---|---|
@@ -0,0 +1,8 @@ | ||
batchsize: 0 | ||
beam-size: 10 | ||
penalty: 0.0 | ||
maxlenratio: 0.0 | ||
minlenratio: 0.0 | ||
ctc-weight: 0.5 | ||
lm-weight: 0.3 | ||
ngram-weight: 0.3 |
This file contains bidirectional Unicode text that may be interpreted or compiled differently than what appears below. To review, open the file in an editor that reveals hidden Unicode characters.
Learn more about bidirectional Unicode characters
Original file line number | Diff line number | Diff line change |
---|---|---|
@@ -0,0 +1,6 @@ | ||
beam-size: 20 | ||
penalty: 0.0 | ||
maxlenratio: 0.0 | ||
minlenratio: 0.0 | ||
ctc-weight: 0.6 | ||
lm-weight: 0.3 |
47 changes: 47 additions & 0 deletions
47
egs/aesrc2020/asr1/conf/tuning/train_pytorch_conformer_kernel15.yaml
This file contains bidirectional Unicode text that may be interpreted or compiled differently than what appears below. To review, open the file in an editor that reveals hidden Unicode characters.
Learn more about bidirectional Unicode characters
Original file line number | Diff line number | Diff line change |
---|---|---|
@@ -0,0 +1,47 @@ | ||
# network architecture | ||
# encoder related | ||
elayers: 12 | ||
eunits: 2048 | ||
# decoder related | ||
dlayers: 6 | ||
dunits: 2048 | ||
# attention related | ||
adim: 256 | ||
aheads: 4 | ||
|
||
# hybrid CTC/attention | ||
mtlalpha: 0.3 | ||
|
||
# label smoothing | ||
lsm-weight: 0.1 | ||
|
||
# minibatch related | ||
batch-size: 32 | ||
maxlen-in: 512 # if input length > maxlen-in, batchsize is automatically reduced | ||
maxlen-out: 150 # if output length > maxlen-out, batchsize is automatically reduced | ||
|
||
# optimization related | ||
sortagrad: 0 # Feed samples from shortest to longest ; -1: enabled for all epochs, 0: disabled, other: enabled for 'other' epochs | ||
opt: noam | ||
accum-grad: 2 | ||
grad-clip: 5 | ||
patience: 0 | ||
epochs: 50 | ||
dropout-rate: 0.1 | ||
|
||
# transformer specific setting | ||
backend: pytorch | ||
model-module: "espnet.nets.pytorch_backend.e2e_asr_conformer:E2E" | ||
transformer-input-layer: conv2d # encoder architecture type | ||
transformer-lr: 1.0 | ||
transformer-warmup-steps: 25000 | ||
transformer-attn-dropout-rate: 0.0 | ||
transformer-length-normalized-loss: false | ||
transformer-init: pytorch | ||
|
||
# conformer specific setting | ||
transformer-encoder-pos-enc-layer-type: rel_pos | ||
transformer-encoder-selfattn-layer-type: rel_selfattn | ||
macaron-style: true | ||
use-cnn-module: true | ||
cnn-module-kernel: 15 |
47 changes: 47 additions & 0 deletions
47
egs/aesrc2020/asr1/conf/tuning/train_pytorch_conformer_kernel31.yaml
This file contains bidirectional Unicode text that may be interpreted or compiled differently than what appears below. To review, open the file in an editor that reveals hidden Unicode characters.
Learn more about bidirectional Unicode characters
Original file line number | Diff line number | Diff line change |
---|---|---|
@@ -0,0 +1,47 @@ | ||
# network architecture | ||
# encoder related | ||
elayers: 12 | ||
eunits: 2048 | ||
# decoder related | ||
dlayers: 6 | ||
dunits: 2048 | ||
# attention related | ||
adim: 256 | ||
aheads: 4 | ||
|
||
# hybrid CTC/attention | ||
mtlalpha: 0.3 | ||
|
||
# label smoothing | ||
lsm-weight: 0.1 | ||
|
||
# minibatch related | ||
batch-size: 32 | ||
maxlen-in: 512 # if input length > maxlen-in, batchsize is automatically reduced | ||
maxlen-out: 150 # if output length > maxlen-out, batchsize is automatically reduced | ||
|
||
# optimization related | ||
sortagrad: 0 # Feed samples from shortest to longest ; -1: enabled for all epochs, 0: disabled, other: enabled for 'other' epochs | ||
opt: noam | ||
accum-grad: 2 | ||
grad-clip: 5 | ||
patience: 0 | ||
epochs: 50 | ||
dropout-rate: 0.1 | ||
|
||
# transformer specific setting | ||
backend: pytorch | ||
model-module: "espnet.nets.pytorch_backend.e2e_asr_conformer:E2E" | ||
transformer-input-layer: conv2d # encoder architecture type | ||
transformer-lr: 1.0 | ||
transformer-warmup-steps: 25000 | ||
transformer-attn-dropout-rate: 0.0 | ||
transformer-length-normalized-loss: false | ||
transformer-init: pytorch | ||
|
||
# conformer specific setting | ||
transformer-encoder-pos-enc-layer-type: rel_pos | ||
transformer-encoder-selfattn-layer-type: rel_selfattn | ||
macaron-style: true | ||
use-cnn-module: true | ||
cnn-module-kernel: 31 |
Oops, something went wrong.