File tree 1 file changed +1
-3
lines changed
nemo/collections/llm/gpt/model
1 file changed +1
-3
lines changed Original file line number Diff line number Diff line change 20
20
import torch .distributed
21
21
from megatron .core .inference .model_inference_wrappers .gpt .gpt_inference_wrapper import GPTInferenceWrapper
22
22
from megatron .core .inference .model_inference_wrappers .inference_wrapper_config import InferenceWrapperConfig
23
+ from megatron .core .models .gpt .gpt_model import GPTModel as MCoreGPTModel
23
24
from megatron .core .optimizer import OptimizerConfig
24
25
from megatron .core .transformer .spec_utils import ModuleSpec
25
26
from megatron .core .transformer .transformer_config import TransformerConfig
44
45
_grad_accum_fusion_available = False
45
46
46
47
if TYPE_CHECKING :
47
- from megatron .core .models .gpt .gpt_model import GPTModel as MCoreGPTModel
48
-
49
48
from nemo .collections .common .tokenizers .tokenizer_spec import TokenizerSpec
50
49
51
50
@@ -189,7 +188,6 @@ def configure_model(self, tokenizer) -> "MCoreGPTModel":
189
188
) % vp_size == 0 , "Make sure the number of model chunks is the same across all pipeline stages."
190
189
191
190
from megatron .core import parallel_state
192
- from megatron .core .models .gpt .gpt_model import GPTModel as MCoreGPTModel
193
191
194
192
transformer_layer_spec = self .transformer_layer_spec
195
193
if not isinstance (transformer_layer_spec , ModuleSpec ):
You can’t perform that action at this time.
0 commit comments