Skip to content

Commit

Permalink
updated batch sizes.
Browse files Browse the repository at this point in the history
Signed-off-by: Vahid <[email protected]>
  • Loading branch information
VahidooX committed Apr 19, 2023
1 parent 82326a1 commit 84608ab
Show file tree
Hide file tree
Showing 4 changed files with 10 additions and 11 deletions.
2 changes: 1 addition & 1 deletion examples/asr/conf/conformer/conformer_ctc_bpe.yaml
Original file line number Diff line number Diff line change
Expand Up @@ -25,7 +25,7 @@
# The checkpoint of the large model trained on LibriSpeech with this recipe can be found here: https://ngc.nvidia.com/catalog/models/nvidia:nemo:stt_en_conformer_ctc_large_ls

# We suggest to use trainer.precision=bf16 for GPUs which support it otherwise trainer.precision=16 is recommended.
# Using bf16 or 16 would makes it possible to double the batch size and speedup training/inference. If fp16 is not stable and model diverges after some epochs, you may use fp32.
# Using bf16 or 16 would make it possible to double the batch size and speedup training/inference. If fp16 is not stable and model diverges after some epochs, you may use fp32.
# Here are the suggested batch size per GPU for each precision and memory sizes:
# +-----------+------------+------------+
# | Precision | GPU Memory | Batch Size |
Expand Down
12 changes: 6 additions & 6 deletions examples/asr/conf/conformer/conformer_transducer_bpe.yaml
Original file line number Diff line number Diff line change
Expand Up @@ -31,13 +31,13 @@
# +-----------+------------+------------+
# | Precision | GPU Memory | Batch Size |
# +===========+============+============+
# | 32 | 16GB | 4 |
# | | 32GB | 8 |
# | | 80GB | 16 |
# | 32 | 16GB | 8 |
# | | 32GB | 16 |
# | | 80GB | 32 |
# +-----------+------------+------------+
# | 16 or | 16GB | 8 |
# | bf16 | 32GB | 16 |
# | | 80GB | 23 |
# | 16 or | 16GB | 16 |
# | bf16 | 32GB | 32 |
# | | 80GB | 64 |
# +-----------+------------+------------+
# Note: They are based on the assumption of max_duration of 20. If you have longer or shorter max_duration, then batch sizes may need to get updated accordingly.

Expand Down
4 changes: 2 additions & 2 deletions examples/asr/conf/fastconformer/fast-conformer_ctc_bpe.yaml
Original file line number Diff line number Diff line change
Expand Up @@ -3,15 +3,15 @@
# You may find more info about FastConformer here: https://docs.nvidia.com/deeplearning/nemo/user-guide/docs/en/stable/asr/models.html#fast-conformer

# We suggest to use trainer.precision=bf16 for GPUs which support it otherwise trainer.precision=16 is recommended.
# Using bf16 or 16 would makes it possible to double the batch size and speedup training/inference. If fp16 is not stable and model diverges after some epochs, you may use fp32.
# Using bf16 or 16 would make it possible to double the batch size and speedup training/inference. If fp16 is not stable and model diverges after some epochs, you may use fp32.
# Here are the suggested batch size per GPU for each precision and memory sizes:

# +-----------+------------+------------+
# | Precision | GPU Memory | Batch Size |
# +===========+============+============+
# | 32 | 16GB | 16 |
# | | 32GB | 32 |
# | | 80GB | 128 |
# | | 80GB | 64 |
# +-----------+------------+------------+
# | fp16 or | 16GB | 32 |
# | bf16 | 32GB | 64 |
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -2,9 +2,8 @@

# You may find more info about FastConformer here: https://docs.nvidia.com/deeplearning/nemo/user-guide/docs/en/stable/asr/models.html#fast-conformer


# We suggest to use trainer.precision=bf16 for GPUs which support it otherwise trainer.precision=16 is recommended.
# Using bf16 or 16 would makes it possible to double the batch size and speedup training/inference. If fp16 is not stable and model diverges after some epochs, you may use fp32.
# Using bf16 or 16 would make it possible to double the batch size and speedup training/inference. If fp16 is not stable and model diverges after some epochs, you may use fp32.
# Here are the suggested batch size per GPU for each precision and memory sizes:

# +-----------+------------+------------+
Expand Down

0 comments on commit 84608ab

Please sign in to comment.