-
Notifications
You must be signed in to change notification settings - Fork 26.2k
Issues: huggingface/transformers
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Author
Label
Projects
Milestones
Assignee
Sort
Issues list
tokenizer
save_pretrained
can not handle non-string value in dtype
bug
#33304
opened Sep 4, 2024 by
jiaweihhuang
2 of 4 tasks
Decoder and cross-attention shape is different when obtained by model.generate() and model()
bug
#33296
opened Sep 4, 2024 by
cgr71ii
1 of 4 tasks
Mask2FormerImageProcessor - fails to process multichannel image
bug
#33295
opened Sep 4, 2024 by
maciej-adamiak
2 of 4 tasks
"Qwen2-VL FP16 inference results in errors or gibberish output."
bug
#33294
opened Sep 4, 2024 by
GeLee-Q
2 of 4 tasks
Is it possible to make wasm support all models in huggingface?
Feature request
Request for a new feature
#33293
opened Sep 4, 2024 by
wa008
oom when using adafactor optimizer in deepspeed
bug
#33290
opened Sep 4, 2024 by
zhangvia
1 of 4 tasks
'DepthEstimationPipeline' object has no attribute 'image_size' when num_workers > 0
bug
#33288
opened Sep 4, 2024 by
rsamf
2 of 4 tasks
transformers 4.44.2 doesn't work with torch.compile and torch.export on T5 generate()
bug
#33283
opened Sep 3, 2024 by
yiming0416
2 of 4 tasks
TypeError: MistralForSequenceClassification.forward() got an unexpected keyword argument 'token_type_ids'
bug
#33280
opened Sep 3, 2024 by
edchengg
2 of 4 tasks
Custom pipeline in remote repo cannot load custom model from remote repo.
bug
#33272
opened Sep 3, 2024 by
ljw20180420
4 tasks
Static KV cache status: How to use it? Does it work for all models?
#33270
opened Sep 3, 2024 by
oobabooga
prepare_fa2_from_position_ids error in training with batch_size > 1
bug
#33268
opened Sep 2, 2024 by
meliksahturker
2 of 4 tasks
Add Log-Softmax Temperature Option to transformers.Seq2SeqTrainingArguments for CE Loss
Feature request
Request for a new feature
#33267
opened Sep 2, 2024 by
oswaldoludwig
Community contribution: Adding GGUF support for more architectures
Feature request
Request for a new feature
Good Second Issue
Issues that are more difficult to do than "Good First" issues - give it a try if you want!
#33260
opened Sep 2, 2024 by
SunMarc
3 of 7 tasks
cannot import name 'ShardedDDPOption' from 'transformers.trainer'
bug
#33242
opened Sep 2, 2024 by
nishitanand
2 of 4 tasks
when l use the model generate method within @tf.function, it encounterd a mistake
bug
#33241
opened Sep 2, 2024 by
HelloWorldU
2 of 4 tasks
M2M100Tokenizer vocabulary size is not equal to the m2m embedding_size for the "facebook/m2m100_418M" model.
bug
#33240
opened Sep 1, 2024 by
GerrySant
1 of 4 tasks
RobertaTokenizer has incorrect character offsets
bug
#33237
opened Aug 31, 2024 by
Craigacp
2 of 4 tasks
When will Transformers version 4.45.0 be released?
bug
#33236
opened Aug 31, 2024 by
hahmad2008
4 tasks
'CohereModel' object has no attribute '_prune_heads'
bug
#33235
opened Aug 31, 2024 by
mnauf
4 tasks
[Falcon Mamba] Unexpected model output with use_cache=False and model.train()
bug
#33234
opened Aug 31, 2024 by
jploski
2 of 4 tasks
Previous Next
ProTip!
Add no:assignee to see everything that’s not assigned.