-
Notifications
You must be signed in to change notification settings - Fork 28.3k
Issues: huggingface/transformers
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Author
Label
Projects
Milestones
Assignee
Sort
Issues list
Support Flex Attention for encoder only models (XLMRoberta, ModernBERT etc...)
Feature request
Request for a new feature
#36697
opened Mar 13, 2025 by
ccdv-ai
AttributeError: 'Gemma3Config' object has no attribute 'vocab_size'
bug
#36683
opened Mar 12, 2025 by
jumelet
4 tasks
Transformers 4.49.0 breaks nvdiffrast plugin loading
bug
#36676
opened Mar 12, 2025 by
drake7707
4 tasks
NotImplementedError: aten::_log_softmax_backward_data with SparseCUDA backend
bug
#36674
opened Mar 12, 2025 by
rangehow
2 of 4 tasks
The parameter 'text' may be None as the comments says, there is a confuse.
#36667
opened Mar 12, 2025 by
ralgond
[FEAT] [non-CUDA]: Support alternative implementation for Request for a new feature
constraints.positive_definite.check
Feature request
#36660
opened Mar 12, 2025 by
tjtanaa
Qwen2 MoE manual Request for a new feature
head_dim
Feature request
#36659
opened Mar 12, 2025 by
yunju63
Add seed to data collator classes
Feature request
Request for a new feature
#36655
opened Mar 12, 2025 by
capemox
AutoModel from_pretrained does not recursively download relative imports
bug
#36653
opened Mar 12, 2025 by
yair-schiff
Marian RNN conversion support
Feature request
Request for a new feature
#36651
opened Mar 11, 2025 by
FricoRico
Is it correct that the repetition penalty is applied to the input_ids encompassing all inputs and outputs, rather than solely on the generated tokens?
bug
#36642
opened Mar 11, 2025 by
Ostrichpie818
3 of 4 tasks
[Feature Request]: refactor _update_causal_mask to a public utility
Feature request
Request for a new feature
#36640
opened Mar 11, 2025 by
Irvingwangjr
[BUG] Batch inference DDP + zero stage 3 = inference code hangs
#36638
opened Mar 11, 2025 by
ShengYun-Peng
output_hidden_states
only return part of hidden_state when setting device_map="auto"
bug
#36636
opened Mar 10, 2025 by
sangyx
2 of 4 tasks
Add Magma from Microsoft to Transformers
New model
#36629
opened Mar 10, 2025 by
VladOS95-cyber
2 tasks done
Unable to use converted Llama 3.3 instruct model
bug
#36628
opened Mar 10, 2025 by
SRGAnalytics-MD
4 tasks
[deepspeed] any plans for deepspeed-domino?
Feature request
Request for a new feature
#36624
opened Mar 10, 2025 by
SoundProvider
Can not use flash-attention and flash-varlen-attention on Ascend NPU
#36618
opened Mar 9, 2025 by
FightingZhen
Previous Next
ProTip!
Add no:assignee to see everything that’s not assigned.