transformers
https://github.com/huggingface/transformers
Python
🤗 Transformers: State-of-the-art Machine Learning for Pytorch, TensorFlow, and JAX.
Triage Issues!
When you volunteer to triage issues, you'll receive an email each day with a link to an open issue that needs help in this project. You'll also receive instructions on how to triage issues.
Triage Docs!
Receive a documented method or class from your favorite GitHub repos in your inbox every day. If you're really pro, receive undocumented methods or classes and supercharge your commit history.
Python not yet supported35 Subscribers
View all SubscribersAdd a CodeTriage badge to transformers
Help out
- Issues
- Support context parallel training with ring-flash-attention
- Whisper Beam Search doesn't work
- Exception raised with trainer + `accelerate launch` FSDP + large gradient accumulation steps + small dataset
- The examples in the examples directory are mostly for fine-tuning pre-trained models?how to trian from scratch
- Support to use adam_mini from installing directly
- VLMs: `patch_size` -> `num_image_tokens` in processing
- Add French translation of task_summary and tasks_explained
- fix flash attention comment
- A Trainer subclass for Decoder-Only LM with generation in evaluate()
- The same situation as #31377 occurred when using Qwen/Qwen2-VL-7B-Instruct
- Docs
- Python not yet supported