transformers
https://github.com/huggingface/transformers
Python
🤗 Transformers: State-of-the-art Machine Learning for Pytorch, TensorFlow, and JAX.
Triage Issues!
When you volunteer to triage issues, you'll receive an email each day with a link to an open issue that needs help in this project. You'll also receive instructions on how to triage issues.
Triage Docs!
Receive a documented method or class from your favorite GitHub repos in your inbox every day. If you're really pro, receive undocumented methods or classes and supercharge your commit history.
Python not yet supported35 Subscribers
View all SubscribersAdd a CodeTriage badge to transformers
Help out
- Issues
- Run nightly ci test new runner
- fix llava next incorrect padding size
- removed redundant creation of causal mask when attention mask is already 4D
- Fix `inputs` deprecation warning in automatic_speech_recognition.py
- The attention mask is not set and cannot be inferred from input because pad token is same as eos token. As a consequence, you may observe unexpected behavior. Please pass your input's `attention_mask` to obtain reliable results.
- add `private` parameter to trainer.push_to_hub
- Clarify what kwargs can be accepted by "AutoModelForCausalLM.from_pretrained()"
- Regression in tokenizer loading
- Support context parallel training with ring-flash-attention
- Whisper Beam Search doesn't work
- Docs
- Python not yet supported