Popular repositories Loading
-
Megatron-LM
Megatron-LM PublicForked from NVIDIA/Megatron-LM
Ongoing research training transformer models at scale
Python
-
DeepSpeed
DeepSpeed PublicForked from microsoft/DeepSpeed
DeepSpeed is a deep learning optimization library that makes distributed training and inference easy, efficient, and effective.
Python
-
transformers
transformers PublicForked from huggingface/transformers
🤗 Transformers: State-of-the-art Machine Learning for Pytorch, TensorFlow, and JAX.
Python
-
accelerate
accelerate PublicForked from huggingface/accelerate
🚀 A simple way to launch, train, and use PyTorch models on almost any device and distributed configuration, automatic mixed precision (including fp8), and easy-to-configure FSDP and DeepSpeed support
Python
If the problem persists, check the GitHub status page or contact support.