deepspeed
https://github.com/microsoft/deepspeed
Python
DeepSpeed is a deep learning optimization library that makes distributed training easy, efficient, and effective.
Triage Issues!
When you volunteer to triage issues, you'll receive an email each day with a link to an open issue that needs help in this project. You'll also receive instructions on how to triage issues.
Triage Docs!
Receive a documented method or class from your favorite GitHub repos in your inbox every day. If you're really pro, receive undocumented methods or classes and supercharge your commit history.
Python not yet supported9 Subscribers
Add a CodeTriage badge to deepspeed
Help out
- Issues
- [WARNING] [cpu_adam.py:84:__init__] FP16 params for CPUAdam may not work on AMD CPUs
- deepspeed memory
- [Question] Why is the call to allocate_workspace() using only the first input as size parameters?
- [Question] Deepspeed Transformer Kernel benchmarks
- [Question] bf16 suppot DeepSpeedCPUAdamOptimizer and performance degradation with multiple instances
- [REQUEST] Mixed dtype for model parameters
- where is the gradient
- [REQUEST] LLaMA-30B can not support 8K sequence length
- [REQUEST] torch.compile + DeepSpeed
- [Question] Difference between launch_bias_add_transform_0213 and launch_apply_rotary_pos_emb in ds_softmax_context
- Docs
- Python not yet supported