Top Repositories
🤗 Transformers: State-of-the-art Machine Learning for Pytorch, TensorFlow, and JAX.
DeepSpeed is a deep learning optimization library that makes distributed training and inference easy, efficient, and effective.
Tensors and Dynamic neural networks in Python with strong GPU acceleration
Example code for AWS Neuron SDK developers building inference and training applications
LLM training code for MosaicML foundation models
Tutel MoE: An Optimized Mixture-of-Experts Implementation
Repositories
11DeepSpeed is a deep learning optimization library that makes distributed training and inference easy, efficient, and effective.
🤗 Transformers: State-of-the-art Machine Learning for Pytorch, TensorFlow, and JAX.
Tensors and Dynamic neural networks in Python with strong GPU acceleration
Example code for AWS Neuron SDK developers building inference and training applications
LLM training code for MosaicML foundation models
Tutel MoE: An Optimized Mixture-of-Experts Implementation
Fast and flexible reference benchmarks
Ongoing research training transformer language models at scale, including: BERT & GPT-2
Mistral: A strong, northwesterly wind: Framework for transparent and accessible large-scale language model training, built with Hugging Face 🤗 Transformers.
🤗 The largest hub of ready-to-use datasets for ML models with fast, easy-to-use and efficient data manipulation tools
Composing methods for ML training efficiency