Megatron-LM
M
Megatron LM
Overview :
Megatron-LM is a powerful large-scale Transformer model developed by NVIDIA's Applied Deep Learning Research team. It is used in continuous research on training Transformer language models at scale. We utilize mixed precision, efficient model parallelism and data parallelism, along with the pre-training of multi-node Transformer models such as GPT, BERT, and T5.
Target Users :
Suitable for research and practice scenarios related to training large-scale language models.
Total Visits: 474.6M
Top Region: US(18.64%)
Website Views : 54.9K
Features
Efficient training of large-scale language models
Support for model parallelism and data parallelism
Support for Transformer models such as GPT, BERT, and T5
AIbase
Empowering the Future, Your AI Solution Knowledge Base
© 2025AIbase