transformers
https://github.com/huggingface/transformers
Python
🤗 Transformers: State-of-the-art Machine Learning for Pytorch, TensorFlow, and JAX.
Triage Issues!
When you volunteer to triage issues, you'll receive an email each day with a link to an open issue that needs help in this project. You'll also receive instructions on how to triage issues.
Triage Docs!
Receive a documented method or class from your favorite GitHub repos in your inbox every day. If you're really pro, receive undocumented methods or classes and supercharge your commit history.
Python not yet supported22 Subscribers
Add a CodeTriage badge to transformers
Help out
- Issues
- error when using PPO in Gemma
- Whisper Translation on low resource languages
- Cache in different devices when use split model with dispatch_model() function and model.generate()
- [WIP] Remove the need to contribute to inits when contributing a model
- ReadTimeOutError with from_pretrained for some model checkpoints only
- ChatGLM3-6b测试模型时报错AttributeError: can't set attribute
- AutoModal how to enable TP for extremly large models?
- mistralai/Mixtral-8x7B-v0.1 bfloat16 much slower than FP32 on Intel EMR CPU
- Community contribution: enable dynamic resolution input for more vision models.
- i cannot find the code that transformers trainer model_wrapped by deepspeed , i can find the theory about model_wrapped was wraped by DDP(Deepspeed(transformer model )) ,but i only find the code transformers model wrapped by ddp, where is the deepspeed wrapped ? thanks ^-^
- Docs
- Python not yet supported