diff --git a/docs/source/nlp/nemo_megatron/gpt/gpt_training.rst b/docs/source/nlp/nemo_megatron/gpt/gpt_training.rst index 807dce64e86e..4c0a09b7f6ea 100644 --- a/docs/source/nlp/nemo_megatron/gpt/gpt_training.rst +++ b/docs/source/nlp/nemo_megatron/gpt/gpt_training.rst @@ -70,7 +70,7 @@ After this is done (will take a while), you'll have two files: ```spm_32k_wiki.m **Step 4: Convert training data into memory map format** -This format makes trainig more efficient, especially with many nodes and GPUs. This step will also tokenize data using tokenizer model from Step 3. +This format makes training more efficient, especially with many nodes and GPUs. This step will also tokenize data using tokenizer model from Step 3. *Option 1:* Using HuggingFace GPT2 tokenizer files.