transformers
https://github.com/huggingface/transformers
Python
🤗 Transformers: State-of-the-art Machine Learning for Pytorch, TensorFlow, and JAX.
Triage Issues!
When you volunteer to triage issues, you'll receive an email each day with a link to an open issue that needs help in this project. You'll also receive instructions on how to triage issues.
Triage Docs!
Receive a documented method or class from your favorite GitHub repos in your inbox every day. If you're really pro, receive undocumented methods or classes and supercharge your commit history.
Python not yet supported35 Subscribers
View all SubscribersAdd a CodeTriage badge to transformers
Help out
- Issues
- removed redundant creation of causal mask when attention mask is already 4D
- Fix `inputs` deprecation warning in automatic_speech_recognition.py
- The attention mask is not set and cannot be inferred from input because pad token is same as eos token. As a consequence, you may observe unexpected behavior. Please pass your input's `attention_mask` to obtain reliable results.
- add `private` parameter to trainer.push_to_hub
- Clarify what kwargs can be accepted by "AutoModelForCausalLM.from_pretrained()"
- Regression in tokenizer loading
- Support context parallel training with ring-flash-attention
- Whisper Beam Search doesn't work
- Exception raised with trainer + `accelerate launch` FSDP + large gradient accumulation steps + small dataset
- The examples in the examples directory are mostly for fine-tuning pre-trained models?how to trian from scratch
- Docs
- Python not yet supported