Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

merge with master #5

Open
wants to merge 9,112 commits into
base: master
Choose a base branch
from
Open
Show file tree
Hide file tree
Changes from all commits
Commits
Show all changes
9112 commits
Select commit Hold shift + click to select a range
89be34c
Fix SegformerForImageClassification (#15895)
NielsRogge Mar 2, 2022
3d22428
Update delete-dev-doc job to match build-dev-doc (#15891)
sgugger Mar 2, 2022
39249c9
Fix doc links in release utils (#15903)
sgugger Mar 2, 2022
4cd7ed4
Fix a TF Vision Encoder Decoder test (#15896)
ydshieh Mar 3, 2022
439de3f
[Fix link in pipeline doc] (#15906)
patrickvonplaten Mar 3, 2022
7b3bd1f
Fix and improve REALM fine-tuning (#15297)
qqaatw Mar 3, 2022
3c4fbc6
Freeze FlaxWav2Vec2 Feature Encoder (#15873)
sanchit-gandhi Mar 3, 2022
b693cbf
The tests were not updated after the addition of `torch.diag` (#15890)
Narsil Mar 3, 2022
6cbfa7b
[Doctests] Fix ignore bug and add more doc tests (#15911)
patrickvonplaten Mar 3, 2022
79d28e8
v4.18.0.dev.0
sgugger Mar 3, 2022
3822e4a
Enabling MaskFormer in pipelines (#15917)
Narsil Mar 3, 2022
7c45fe7
Mark slow tests as slow
sgugger Mar 3, 2022
742273a
fix for the output from post_process_panoptic_segmentation (#15916)
FrancescoSaverioZuppichini Mar 3, 2022
9251427
Add vision models to doc tests (#15905)
NielsRogge Mar 3, 2022
c0281fe
Fix #15898 (#15928)
davidleonfdez Mar 3, 2022
a7df656
Update README.md (#15926)
patrickvonplaten Mar 3, 2022
a6e3b17
Re-enabling all fast pipeline tests. (#15924)
Narsil Mar 4, 2022
b714748
Update README.md
sanchit-gandhi Mar 4, 2022
6b104c5
Support CLIPTokenizerFast for CLIPProcessor (#15913)
cosmoquester Mar 4, 2022
7ade7c1
Updating the slow tests: (#15893)
Narsil Mar 4, 2022
89c7d9c
Making MaskFormerForInstanceSegmentation. (#15934)
Narsil Mar 4, 2022
01485ce
Add missing support for Flax XLM-RoBERTa (#15900)
versae Mar 4, 2022
10b7698
[FlaxT5 Example] fix flax t5 example pretraining (#15835)
patrickvonplaten Mar 4, 2022
f0aacc1
Do not change the output from tuple to list - to match PT's version (…
ydshieh Mar 4, 2022
040c11f
Tests for MaskFormerFeatureExtractor's post_process*** methods (#15929)
FrancescoSaverioZuppichini Mar 4, 2022
5c6f57e
Constrained Beam Search [*With* Disjunctive Decoding] (#15761)
cwkeam Mar 4, 2022
e8efaec
Move dependency to call method (#15941)
NielsRogge Mar 4, 2022
9932ee4
made MaskFormerModelTest faster (#15942)
FrancescoSaverioZuppichini Mar 4, 2022
ef9c3ca
[Bug Fix] Beam search example in docs fails & a fix (integrating `max…
cwkeam Mar 7, 2022
60b81df
remove re-defination of FlaxWav2Vec2ForCTCModule (#15965)
patil-suraj Mar 7, 2022
544fd98
Support modern list type hints in HfArgumentParser (#15951)
konstantinjdobler Mar 7, 2022
1a62b25
Backprop Test for Freeze FlaxWav2Vec2 Feature Encoder (#15938)
sanchit-gandhi Mar 7, 2022
2596f95
Fix Embedding Module Bug in Flax Models (#15920)
sanchit-gandhi Mar 7, 2022
e9fa7cd
Make is_thing_map in Feature Extractor post_process_panoptic_segmenta…
FrancescoSaverioZuppichini Mar 7, 2022
c87cfd6
Better error message when inputs are empty
sgugger Mar 7, 2022
38cc350
Update training scripts docs (#15931)
stevhliu Mar 7, 2022
8b9ae45
Set scale_embedding to False in some TF tests (#15952)
ydshieh Mar 7, 2022
9879a1d
Fix LayoutLMv2 test (#15939)
NielsRogge Mar 8, 2022
b19f3e6
[Tests] Fix ViTMAE integration test (#15949)
NielsRogge Mar 8, 2022
ea07064
Returning outputs only when asked for for MaskFormer. (#15936)
Narsil Mar 8, 2022
91fb62d
Speedup training by using numpy instead of jnp for batch shuffling (#…
yhavinga Mar 8, 2022
f5a080d
Do a pull in case docs were updated during build (#15922)
sgugger Mar 8, 2022
7298330
Fix TFEncoderDecoderModelTest - Pytorch device (#15979)
ydshieh Mar 8, 2022
ab2f8d1
add hf hub to env version command (#15981)
patrickvonplaten Mar 8, 2022
62d8476
Update TF multiple choice example (#15868)
gante Mar 8, 2022
70203b5
TF generate refactor - past without encoder outputs (#15944)
gante Mar 8, 2022
5b7dcc7
Seed _get_train_sampler's generator with arg seed to improve reproduc…
dlwh Mar 8, 2022
f4e4ad3
Add `ForInstanceSegmentation` models to `image-segmentation` pipeline…
Narsil Mar 9, 2022
c1aaa43
[Doctests] Move doctests to new GPU & Fix bugs (#15969)
patrickvonplaten Mar 9, 2022
3ea0469
Removed an outdated check about hdf5_version (#16011)
ydshieh Mar 9, 2022
e7f34cc
Swag example: Update doc format (#16014)
gante Mar 9, 2022
e493a3a
Fix github actions comment (#16009)
LysandreJik Mar 9, 2022
cec89e1
Simplify release utils (#15921)
sgugger Mar 9, 2022
38bce1d
Make `pos` optional to avoid crashing `PerceiverModel` operation (#15…
basilevh Mar 9, 2022
1e8f379
done (#16012)
FrancescoSaverioZuppichini Mar 9, 2022
8feede2
Fix broken code blocks in README.md (#15967)
upura Mar 9, 2022
b7fa1e3
Use tiny models for get_pretrained_model in TFEncoderDecoderModelTest…
ydshieh Mar 9, 2022
50dd314
Add ONNX export for ViT (#15658)
lewtun Mar 9, 2022
b256f35
Add FlaxBartForCausalLM (#15995)
sanchit-gandhi Mar 9, 2022
a69e185
add doctests for bart like seq2seq models (#15987)
patil-suraj Mar 9, 2022
65f9653
Fix warning message in ElectraForCausalLM (#16023)
pbelevich Mar 9, 2022
fde9018
Freeze Feature Encoder in FlaxSpeechEncoderDecoder (#15997)
sanchit-gandhi Mar 10, 2022
6c9010e
Update README.md
sanchit-gandhi Mar 10, 2022
0835119
Add Document Image Transformer (DiT) (#15984)
NielsRogge Mar 10, 2022
0951d31
Fix dependency error message in ServeCommand (#16033)
andstor Mar 10, 2022
6ce11c2
[Docs] Improve PyTorch, Flax generate API (#15988)
patrickvonplaten Mar 10, 2022
8d83ebd
[Tests] Add attentions_option to ModelTesterMixin (#15909)
NielsRogge Mar 10, 2022
b2a1c99
[README] fix url for Preprocessing tutorial (#16042)
patil-suraj Mar 10, 2022
1da84ae
Fix Bug in Flax-Speech-Encoder-Decoder Test (#16041)
sanchit-gandhi Mar 10, 2022
2f463ef
Fix TFDebertaV2ConvLayer in TFDebertaV2Model (#16031)
ydshieh Mar 10, 2022
1059139
Build the doc in a seperate folder then move it (#16020)
sgugger Mar 10, 2022
1959799
Don't compute metrics in LM examples on TPU (#16029)
sgugger Mar 10, 2022
b7018ab
TF: Unpack model inputs through a decorator (#15907)
gante Mar 10, 2022
741e493
Fix Bug in Flax Seq2Seq Models (#16021)
sanchit-gandhi Mar 10, 2022
e66743e
DeBERTa/DeBERTa-v2/SEW Support for torch 1.11 (#16043)
LysandreJik Mar 10, 2022
ba21001
support new marian models (#15831)
patil-suraj Mar 10, 2022
6b09328
Fix duplicate arguments passed to dummy inputs in ONNX export (#16045)
lewtun Mar 10, 2022
96ac754
updating fine-tune classifier documentation (#16063)
davidsbatista Mar 10, 2022
b6bdb94
Fix a TF test name (LayoutLMModelTest) (#16061)
ydshieh Mar 11, 2022
f5741bc
Move QDQBert in just PyTorch block (#16062)
sgugger Mar 11, 2022
5b369dc
Remove assertion over possible activation functions in DistilBERT (#1…
mfuntowicz Mar 11, 2022
0868fde
Fix torch-scatter version (#16072)
LysandreJik Mar 11, 2022
ecf989c
Trigger doc build
sgugger Mar 11, 2022
f7708e1
Force default brnahc name via the config
sgugger Mar 11, 2022
bb69d15
Add type annotations for BERT and copies (#16074)
Rocketknight1 Mar 11, 2022
5d2fed2
Adding type hints for TFRoBERTa (#16057)
Rocketknight1 Mar 11, 2022
7e00247
check for key 'torch.dtype' in nested dicts in config (#16065)
feifang24 Mar 11, 2022
322c853
Run daily test without time-out at least once (#16077)
patrickvonplaten Mar 11, 2022
9442b3c
Add soft length regulation for sequence generation (#15245)
kevinpl07 Mar 11, 2022
5b4c97d
Update troubleshoot guide (#16001)
stevhliu Mar 11, 2022
7f3d444
add type annotations for ImageGPT (#16088)
johnnv1 Mar 11, 2022
a01fe4c
Rebuild deepspeed (#16081)
LysandreJik Mar 11, 2022
eaed689
Add missing type hints for all flavors of RoBERTa PyTorch models. (#1…
chainyo Mar 11, 2022
cb5e50c
[Fix doc example] FSMT (#16085)
ydshieh Mar 11, 2022
ae2dd42
Audio/vision task guides (#15808)
stevhliu Mar 11, 2022
c1f209d
[ZeRO] Fixes issue with embedding resize (#16093)
jeffra Mar 11, 2022
580dd87
[Deepspeed] add support for bf16 mode (#14569)
stas00 Mar 12, 2022
3e9d0f7
Change unpacking of TF Bart inputs (#16094)
osanseviero Mar 12, 2022
9042dfe
add unpack_inputs decorator to mbart (#16097)
Abdelrhman-Hosny Mar 12, 2022
62b05b6
Add type annotations for segformer classes (#16099)
p-mishra1 Mar 12, 2022
8416206
apply unpack_input decorator to ViT model (#16102)
johnnv1 Mar 12, 2022
65cf33e
Add type hints to XLM model (PyTorch) (#16108)
jbrry Mar 12, 2022
20ab158
Add missing type hints for all flavors of LayoutLMv2 PyTorch models. …
chainyo Mar 13, 2022
6e1e88f
Add TFCamembertForCausalLM and ONNX integration test (#16073)
lewtun Mar 14, 2022
802984a
Fix and document Zero Shot Image Classification (#16079)
osanseviero Mar 14, 2022
2de99e6
Fix Loading of Flax(Speech)EncoderDecoderModel kwargs from PreTrained…
sanchit-gandhi Mar 14, 2022
9e9f6b8
Update convert_marian_to_pytorch.py (#16124)
jorgtied Mar 14, 2022
923c35b
Make TF pt-tf equivalence test more aggressive (#15839)
ydshieh Mar 14, 2022
5dbf36b
Fix ProphetNetTokenizer (#16082)
ydshieh Mar 14, 2022
e3645fd
Change unpacking of TF mobilebert inputs to use decorator (#16110)
vumichien Mar 14, 2022
f284aa3
steps strategy fix for PushtoHubCallback (#16138)
merveenoyan Mar 14, 2022
9fd584e
Add copied from statements and fix prefix (#16119)
NielsRogge Mar 14, 2022
efd6e9a
Spanish translation of the file training.mdx (#16047)
yharyarias Mar 14, 2022
1d43933
Added missing type hints (#16103)
kamalkraj Mar 14, 2022
72ae06b
Added missing type hints - V1 and V2 (#16105)
kamalkraj Mar 14, 2022
c8c8c11
[Fix doc example] Fix checkpoint name in docstring example in Speech2…
ydshieh Mar 14, 2022
d02bd4f
Better input variable naming for OpenAI (TF) (#16129)
bhavika Mar 14, 2022
97e32b7
Improve model variable naming - CLIP [TF] (#16128)
bhavika Mar 14, 2022
3f8360a
Add type hints for TFDistilBert (#16107)
PepijnBoers Mar 14, 2022
37a9fc4
Choose framework for ONNX export (#16018)
michaelbenayoun Mar 14, 2022
6c2f3ed
Add type hints for Luke in PyTorch (#16111)
bhavika Mar 14, 2022
5493c10
Add type hints for PoolFormer in Pytorch (#16121)
sooperset Mar 14, 2022
6db8693
Add type hints for SqueezeBert PyTorch (#16126)
Tegzes Mar 14, 2022
10cf1ff
Added missing type hints - ELECTRA TF (#16104)
kamalkraj Mar 14, 2022
c1000e7
Dcoker images runtime -> devel (#16141)
LysandreJik Mar 14, 2022
f86235a
Add type annotations for CLIP (torch) (#16059) (#16106)
jacobdineen Mar 14, 2022
0dcdfe8
Add type hints for FNet PyTorch (#16123)
wpan03 Mar 14, 2022
e109edf
Use `HF_ENDPOINT` for custom endpoints (#16139)
sgugger Mar 14, 2022
3779325
update albert with tf decorator (#16147)
infinite-Joy Mar 14, 2022
6458236
TF Electra - clearer model variable naming (#16143)
kamalkraj Mar 14, 2022
e3008c6
[WIP] Resnet (#15770)
FrancescoSaverioZuppichini Mar 14, 2022
8f3ea7a
Add type hints for GPTNeo PyTorch (#16127)
Tegzes Mar 14, 2022
0a05720
Visual Attention Network (VAN) (#16027)
FrancescoSaverioZuppichini Mar 15, 2022
a7aca42
Improve Swin for VisionEncoderDecoder (#16070)
NielsRogge Mar 15, 2022
5a386fb
Make transformers.utils.fx. _SUPPORTED_MODELS unique (#16015)
pbelevich Mar 15, 2022
5664d27
Shift responsibilities a bit (#16154)
patrickvonplaten Mar 15, 2022
cd1ffb4
typo "conaining" -> "containing" (#16132)
marxav Mar 15, 2022
5771344
Configurable Relative Position Max. Distance (#16155)
agemagician Mar 15, 2022
daa4944
Added spanish translation of quicktour.mdx (#16158)
duedme Mar 15, 2022
8bfd2fb
Use templates (#16142)
sgugger Mar 15, 2022
bcaf566
[Fix doc example] Fix first example for the custom_datasets tutorial …
MarkusSagen Mar 15, 2022
e5bc438
[Fix doc example] Fix 2 PyTorch Vilt docstring examples (#16076)
ydshieh Mar 15, 2022
cd4c5c9
TF XLA greedy generation (#15786)
Rocketknight1 Mar 15, 2022
0d7322c
TF clearer model variable naming: pegasus (#16152)
kamalkraj Mar 15, 2022
611d3a0
Change unpacking of TF inputs: layoutlm, mpnet, rag, and roformer (#1…
vumichien Mar 15, 2022
c1c17bd
update transformer XL with tf decorator (#16166)
infinite-Joy Mar 15, 2022
4a353ca
added type hints to yoso (#16163)
mowafess Mar 15, 2022
4f4e5dd
Framework split (#16030)
sgugger Mar 15, 2022
a78565b
[MT5Config] add relative_attention_max_distance in config (#16170)
patil-suraj Mar 15, 2022
79465ac
TF clearer model variable naming: Tapas (#16145)
kamalkraj Mar 15, 2022
366c18f
TF clearer model variable naming: Deberta (#16146)
kamalkraj Mar 15, 2022
a23a7c0
Add flaubert types (#16118)
chainyo Mar 15, 2022
015de6f
TF clearer model variable naming: xlnet (#16150)
kamalkraj Mar 15, 2022
16399d6
Add type annotations for Perceiver (#16174)
jcmc00 Mar 15, 2022
88f7c56
Added type hints for Reformer (#16175)
Tegzes Mar 15, 2022
ea05d67
Fix some Flax models' `hidden_states` (#16167)
ydshieh Mar 15, 2022
db4dd44
Trigger doc build
sgugger Mar 15, 2022
99fd3eb
Add the XTREME-S fine-tuning example (#15985)
anton-l Mar 15, 2022
c2dc89b
[Xtreme-S] fix some namings (#16183)
patrickvonplaten Mar 16, 2022
ee27b3d
Replace all deprecated `jax.ops` operations with jnp's `at` (#16078)
sanchit-gandhi Mar 16, 2022
0f35cda
TF clearer model variable naming: funnel (#16178)
utkusaglm Mar 16, 2022
8cc925a
TF clearer model variable naming: blenderbot (#16192)
utkusaglm Mar 16, 2022
d35e0c6
Minor fixes to XTREME-S (#16193)
anton-l Mar 16, 2022
f06c2c2
TF unpack_input decorator for convnext (#16181)
johko Mar 16, 2022
0b8b061
clearer model variable naming: blenderbot_small (#16194)
utkusaglm Mar 16, 2022
5bdf331
Adding type hints for Distilbert (#16090)
johnryan465 Mar 16, 2022
36f8c42
ResNet: update modules names (#16196)
FrancescoSaverioZuppichini Mar 16, 2022
09013ef
Update step name (#16189)
ydshieh Mar 16, 2022
1909945
Fix loading CLIPVisionConfig and CLIPTextConfig (#16198)
patil-suraj Mar 16, 2022
204c54d
TF: add beam search tests (#16202)
gante Mar 16, 2022
667b823
Swin support for any input size (#15986)
FrancescoSaverioZuppichini Mar 16, 2022
2410d0f
Fix generation min length (#16206)
patrickvonplaten Mar 16, 2022
99e2982
Add/type annotations/model vision (#16151)
johnnv1 Mar 16, 2022
76c74b3
VAN: update modules names (#16201)
FrancescoSaverioZuppichini Mar 17, 2022
73f0a5d
Fixes Loss for TransfoXL when using Trainer API v2 (#16140)
LysandreJik Mar 17, 2022
03c14a5
[Tests] Fix DiT test (#16218)
NielsRogge Mar 17, 2022
25b8f9a
Fix FlaxRoFormerClassificationHead activation (#16168)
ydshieh Mar 17, 2022
e5101c2
Fix typo (#16208)
daysm Mar 17, 2022
e1da89c
Fix reproducibility in Training for PyTorch 1.11 (#16209)
sgugger Mar 17, 2022
7e0d04b
Fix readmes (#16217)
qqaatw Mar 17, 2022
d9b8d1a
update test (#16219)
FrancescoSaverioZuppichini Mar 17, 2022
abd503d
TF - Adding Unpack Decorator For DPR model (#16212)
forsc Mar 17, 2022
5a6b3cc
Skip equivalence test for TransfoXL (#16224)
LysandreJik Mar 17, 2022
8481ece
Fix Type Hint of Nan/Inf Logging Filter Arg (#16227)
Sophylax Mar 17, 2022
93d3fd8
remove jax.ops.index (#16220)
patil-suraj Mar 17, 2022
81643ed
Support PEP 563 for HfArgumentParser (#15795)
function2-llx Mar 17, 2022
1c1e377
TF - add unpack_inputs decorator for marian (#16226)
johko Mar 17, 2022
b6e06c8
fix(flax): generate with logits processor/warper (#16231)
borisdayma Mar 17, 2022
632ff3c
[FlaxSpeechEncoderDecoderModel] Skip from_encoder_decoder_pretrained …
patil-suraj Mar 17, 2022
8a96b0f
[Generate Docs] Correct docs (#16133)
patrickvonplaten Mar 17, 2022
47cccb5
[Deepspeed] non-HF Trainer doc update (#16238)
stas00 Mar 17, 2022
12d1f07
integrations: mlflow: skip start_run() if a run is already active and…
ktzsh Mar 17, 2022
ec4e421
Update expected slices for pillow > 9 (#16117)
NielsRogge Mar 18, 2022
ecb4662
Attention mask is important in the case of batching... (#16222)
Narsil Mar 18, 2022
cb2b027
Change assertion to warning when passing past_key_value to T5 encoder…
ZhaofengWu Mar 18, 2022
a627196
Override _pad in LEDTokenizer to deal with global_attention_mask (#15…
ydshieh Mar 18, 2022
ddbc9ae
Update XLM with TF decorator (#16247)
louisowen6 Mar 18, 2022
5709a20
Add unpack_inputs decorator for ctrl (#16242)
johko Mar 18, 2022
b25b92a
update jax version and re-enable some tests (#16254)
patil-suraj Mar 18, 2022
fdc2e64
added cbs to notebooks, made copy-paste error fix in generation_utils…
cwkeam Mar 18, 2022
c03b6e4
value check for typical sampling (#16165)
cimeister Mar 18, 2022
d481b64
Make Flax pt-flax equivalence test more aggressive (#15841)
ydshieh Mar 18, 2022
75c666b
Aggressive PT/TF equivalence test on PT side (#16250)
ydshieh Mar 18, 2022
277fc2c
Update flaubert with tf decorator (#16258)
Tegzes Mar 18, 2022
ffc319e
Fix links in guides (#16182)
stevhliu Mar 18, 2022
8d74207
Small fixes to the documentation (#16180)
sgugger Mar 18, 2022
f466936
Add has_attentions to TFModelTesterMixin as done on PyTorch side (#16…
ydshieh Mar 19, 2022
0a5ef03
Make `add-new-model-like` work in an env without all frameworks (#16239)
sgugger Mar 21, 2022
319cbbe
Deberta v2 code simplification (#15732)
guillaume-be Mar 21, 2022
c1af180
Add Slack notification support for doc tests (#16253)
patrickvonplaten Mar 21, 2022
c36b856
Framework split for Spanish version of doc quicktour.mdx (#16215)
omarespejel Mar 21, 2022
8bbd413
removed the 'optional' string (#16266)
dinesh-GDK Mar 21, 2022
4ecb022
Draft a guide with our code quirks for new models (#16237)
sgugger Mar 21, 2022
f393868
Fixed Error Raised Due to Wrongly Accessing Training Sample (#16115)
aflah02 Mar 21, 2022
641e5f3
Fix XGLM cross attention (#16290)
patil-suraj Mar 21, 2022
abf3cc7
Fix a typo (add a coma) (#16291)
PolarisRisingWar Mar 21, 2022
c6f7ea1
Add type hints to xlnet (#16214)
mowafess Mar 21, 2022
3f0f75e
Remove disclaimer from Longformer docs (#16296)
gchhablani Mar 21, 2022
2afe9cd
Add argument "cache_dir" for transformers.onnx (#16284)
happyXia Mar 21, 2022
460f36d
Add type hints transfoxl (#16267)
jcmc00 Mar 21, 2022
d50f62f
added type hints for BART model (#16270)
robotjellyzone Mar 21, 2022
9ad77af
test (#16294)
FrancescoSaverioZuppichini Mar 21, 2022
250b478
GPT2 TensorFlow Type Hints (#16261)
cakiki Mar 21, 2022
94be424
Added type hints for PyTorch T5 model (#16257)
yhl48 Mar 21, 2022
f0c00d8
Fix Marian conversion script (#16300)
patil-suraj Mar 21, 2022
fbb4543
[SegFormer] Remove unused attributes (#16285)
NielsRogge Mar 21, 2022
5a42bb4
Update troubleshoot with more content (#16243)
stevhliu Mar 21, 2022
4b27748
fix last element in hidden_states for XGLM (#16301)
ydshieh Mar 21, 2022
c77092a
[FlaxGPTJ] Fix bug in rotary embeddings (#16298)
patil-suraj Mar 21, 2022
7643b1c
Added type hints to PyTorch Longformer models (#16244)
johnnygreco Mar 21, 2022
6f1727d
Fix Seq2SeqTrainingArguments docs (#16295)
gchhablani Mar 21, 2022
e226a24
[xtreme-s] Update Minds14 results (#16241)
anton-l Mar 21, 2022
96cd5bc
added type hints for blenderbot and blenderbot_small (#16307)
IvanLauLinTiong Mar 21, 2022
f9387c9
Update Makefile Phonies (#16306)
gchhablani Mar 21, 2022
9fef668
TF - update (vision_)encoder_decoder past variable (#16260)
gante Mar 21, 2022
0aac9ba
Add Flaubert OnnxConfig to Transformers (#16279)
chainyo Mar 21, 2022
df32b5d
TFLongformer: Add missing type hints and unpack inputs decorator (#16…
johnnygreco Mar 21, 2022
0c55d47
Add GLPN (#16199)
NielsRogge Mar 22, 2022
7865f4d
add xglm conversion script (#16305)
patil-suraj Mar 22, 2022
7cc2c9c
Fix bugs of s2t fairseq model converting (#15593)
beomseok-lee Mar 22, 2022
deb61e5
Add type hints for Pegasus (#16324)
Tegzes Mar 22, 2022
7b262b9
Funnel type hints (#16323)
Mar 22, 2022
87a9af5
Add type hints for ProphetNet PyTorch (#16272)
Tegzes Mar 22, 2022
a2379b9
[GLPN] Improve docs (#16331)
NielsRogge Mar 22, 2022
d49f8d3
Added type hints for Pytorch Marian calls (#16200)
clefourrier Mar 22, 2022
c30798e
done (#16340)
FrancescoSaverioZuppichini Mar 22, 2022
ec3aace
Add type annotations for Rembert/Splinter and copies (#16338)
jacobdineen Mar 22, 2022
4f6c938
[Bug template] Shift responsibilities for long-range (#16344)
patrickvonplaten Mar 22, 2022
62cbd84
Fix code repetition in serialization guide (#16346)
osanseviero Mar 22, 2022
7732148
Adopt framework-specific blocks for content (#16342)
stevhliu Mar 22, 2022
File filter

Filter by extension

Filter by extension


Conversations
Failed to load comments.
Loading
Jump to
The table of contents is too big for display.
Diff view
Diff view
  •  
  •  
  •  
7 changes: 7 additions & 0 deletions .circleci/TROUBLESHOOT.md
Original file line number Diff line number Diff line change
@@ -0,0 +1,7 @@
# Troubleshooting

This is a document explaining how to deal with various issues on Circle-CI. The entries may include actually solutions or pointers to Issues that cover those.

## Circle CI

* pytest worker runs out of resident RAM and gets killed by `cgroups`: https://github.com/huggingface/transformers/issues/11408
1,010 changes: 1,010 additions & 0 deletions .circleci/config.yml

Large diffs are not rendered by default.

12 changes: 12 additions & 0 deletions .coveragerc
Original file line number Diff line number Diff line change
@@ -0,0 +1,12 @@
[run]
source=transformers
omit =
# skip convertion scripts from testing for now
*/convert_*
*/__main__.py
[report]
exclude_lines =
pragma: no cover
raise
except
register_parameter
3 changes: 3 additions & 0 deletions .gitattributes
Original file line number Diff line number Diff line change
@@ -0,0 +1,3 @@
*.py eol=lf
*.rst eol=lf
*.md eol=lf
22 changes: 22 additions & 0 deletions .github/ISSUE_TEMPLATE/---new-benchmark.md
Original file line number Diff line number Diff line change
@@ -0,0 +1,22 @@
---
name: "\U0001F5A5 New benchmark"
about: Benchmark a part of this library and share your results
title: "[Benchmark]"
labels: ''
assignees: ''

---

# 🖥 Benchmarking `transformers`

## Benchmark

Which part of `transformers` did you benchmark?

## Set-up

What did you run your benchmarks on? Please include details, such as: CPU, GPU? If using multiple GPUs, which parallelization did you use?

## Results

Put your results here!
20 changes: 20 additions & 0 deletions .github/ISSUE_TEMPLATE/--new-model-addition.md
Original file line number Diff line number Diff line change
@@ -0,0 +1,20 @@
---
name: "\U0001F31F New model addition"
about: Submit a proposal/request to implement a new Transformer-based model
title: ''
labels: New model
assignees: ''

---

# 🌟 New model addition

## Model description

<!-- Important information -->

## Open source status

* [ ] the model implementation is available: (give details)
* [ ] the model weights are available: (give details)
* [ ] who are the authors: (mention them, if possible by @gh-username)
107 changes: 107 additions & 0 deletions .github/ISSUE_TEMPLATE/bug-report.md
Original file line number Diff line number Diff line change
@@ -0,0 +1,107 @@
---
name: "\U0001F41B Bug Report"
about: Submit a bug report to help us improve transformers
title: ''
labels: ''
assignees: ''

---


## Environment info
<!-- You can run the command `transformers-cli env` and copy-and-paste its output below.
Don't forget to fill out the missing fields in that output! -->

- `transformers` version:
- Platform:
- Python version:
- PyTorch version (GPU?):
- Tensorflow version (GPU?):
- Using GPU in script?:
- Using distributed or parallel set-up in script?:

### Who can help
<!-- Your issue will be replied to more quickly if you can figure out the right person to tag with @
If you know how to use git blame, that is the easiest way, otherwise, here is a rough guide of **who to tag**.
Please tag fewer than 3 people.

Models:

- ALBERT, BERT, XLM, DeBERTa, DeBERTa-v2, ELECTRA, MobileBert, SqueezeBert: @LysandreJik
- T5, Pegasus, EncoderDecoder: @patrickvonplaten
- Blenderbot, MBART, BART, Marian, Pegasus: @patil-suraj
- Reformer, TransfoXL, XLNet, FNet: @patrickvonplaten
- Longformer, BigBird: @ydshieh
- FSMT: @stas00
- Funnel: @sgugger
- GPT-2, GPT: @patil-suraj, @patrickvonplaten, @LysandreJik
- RAG, DPR: @patrickvonplaten, @lhoestq
- TensorFlow: @Rocketknight1
- JAX/Flax: @patil-suraj
- TAPAS, LayoutLM, LayoutLMv2, LUKE, ViT, BEiT, DEiT, DETR, CANINE: @NielsRogge
- GPT-Neo, GPT-J, CLIP: @patil-suraj
- Wav2Vec2, HuBERT, SpeechEncoderDecoder, UniSpeech, UniSpeechSAT, SEW, SEW-D, Speech2Text: @patrickvonplaten, @anton-l

If the model isn't in the list, ping @LysandreJik who will redirect you to the correct contributor.

Library:

- Benchmarks: @patrickvonplaten
- Deepspeed: @stas00
- Ray/raytune: @richardliaw, @amogkam
- Text generation: @patrickvonplaten @narsil
- Tokenizers: @SaulLu
- Trainer: @sgugger
- Pipelines: @Narsil
- Speech: @patrickvonplaten, @anton-l
- Vision: @NielsRogge, @sgugger

Documentation: @sgugger

Model hub:

- for issues with a model, report at https://discuss.huggingface.co/ and tag the model's creator.

HF projects:

- datasets: [different repo](https://github.com/huggingface/datasets)
- rust tokenizers: [different repo](https://github.com/huggingface/tokenizers)

Examples:

- maintained examples (not research project or legacy): @sgugger, @patil-suraj

For research projetcs, please ping the contributor directly. For example, on the following projects:

- research_projects/bert-loses-patience: @JetRunner
- research_projects/distillation: @VictorSanh

-->

## Information

Model I am using (Bert, XLNet ...):

The problem arises when using:
* [ ] the official example scripts: (give details below)
* [ ] my own modified scripts: (give details below)

The tasks I am working on is:
* [ ] an official GLUE/SQUaD task: (give the name)
* [ ] my own task or dataset: (give details below)

## To reproduce

Steps to reproduce the behavior:

1.
2.
3.

<!-- If you have code snippets, error messages, stack traces please provide them here as well.
Important! Use code tags to correctly format your code. See https://help.github.com/en/github/writing-on-github/creating-and-highlighting-code-blocks#syntax-highlighting
Do not use screenshots, as they are hard to read and (more importantly) don't allow others to copy-and-paste your code.-->

## Expected behavior

<!-- A clear and concise description of what you would expect to happen. -->
25 changes: 25 additions & 0 deletions .github/ISSUE_TEMPLATE/feature-request.md
Original file line number Diff line number Diff line change
@@ -0,0 +1,25 @@
---
name: "\U0001F680 Feature request"
about: Submit a proposal/request for a new transformers feature
title: ''
labels: ''
assignees: ''

---

# 🚀 Feature request

<!-- A clear and concise description of the feature proposal.
Please provide a link to the paper and code in case they exist. -->

## Motivation

<!-- Please outline the motivation for the proposal. Is your feature request
related to a problem? e.g., I'm always frustrated when [...]. If this is related
to another GitHub issue, please link here too. -->

## Your contribution

<!-- Is there any way that you could help, e.g. by submitting a PR?
Make sure to read the CONTRIBUTING.MD readme:
https://github.com/huggingface/transformers/blob/master/CONTRIBUTING.md -->
58 changes: 58 additions & 0 deletions .github/ISSUE_TEMPLATE/migration.md
Original file line number Diff line number Diff line change
@@ -0,0 +1,58 @@
---
name: "\U0001F4DA Migration from pytorch-pretrained-bert or pytorch-transformers"
about: Report a problem when migrating from pytorch-pretrained-bert or pytorch-transformers
to transformers
title: ''
labels: Migration
assignees: ''

---

# 📚 Migration

## Information

<!-- Important information -->

Model I am using (Bert, XLNet ...):

Language I am using the model on (English, Chinese ...):

The problem arises when using:
* [ ] the official example scripts: (give details below)
* [ ] my own modified scripts: (give details below)

The tasks I am working on is:
* [ ] an official GLUE/SQUaD task: (give the name)
* [ ] my own task or dataset: (give details below)

## Details

<!-- A clear and concise description of the migration issue.
If you have code snippets, please provide it here as well.
Important! Use code tags to correctly format your code. See https://help.github.com/en/github/writing-on-github/creating-and-highlighting-code-blocks#syntax-highlighting
Do not use screenshots, as they are hard to read and (more importantly) don't allow others to copy-and-paste your code.
-->

## Environment info
<!-- You can run the command `python transformers-cli env` and copy-and-paste its output below.
Don't forget to fill out the missing fields in that output! -->

- `transformers` version:
- Platform:
- Python version:
- PyTorch version (GPU?):
- Tensorflow version (GPU?):
- Using GPU in script?:
- Using distributed or parallel set-up in script?:

<!-- IMPORTANT: which version of the former library do you use? -->
* `pytorch-transformers` or `pytorch-pretrained-bert` version (or branch):


## Checklist

- [ ] I have read the migration guide in the readme.
([pytorch-transformers](https://github.com/huggingface/transformers#migrating-from-pytorch-transformers-to-transformers);
[pytorch-pretrained-bert](https://github.com/huggingface/transformers#migrating-from-pytorch-pretrained-bert-to-transformers))
- [ ] I checked if a related official extension example runs on my machine.
26 changes: 26 additions & 0 deletions .github/ISSUE_TEMPLATE/question-help.md
Original file line number Diff line number Diff line change
@@ -0,0 +1,26 @@
---
name: "❓ Questions & Help"
about: Post your general questions on the Hugging Face forum: https://discuss.huggingface.co/
title: ''
labels: ''
assignees: ''

---

# ❓ Questions & Help

<!-- The GitHub issue tracker is primarly intended for bugs, feature requests,
new models, benchmarks, and migration questions. For all other questions,
we direct you to the Hugging Face forum: https://discuss.huggingface.co/ .
-->

## Details

<!-- Description of your issue -->

<!-- You should first ask your question on the forum, and only if
you didn't get an answer after a few days ask it here on GitHub. -->

**A link to original question on the forum**:

<!-- Your issue will be closed if you don't fill this part. -->
74 changes: 74 additions & 0 deletions .github/PULL_REQUEST_TEMPLATE.md
Original file line number Diff line number Diff line change
@@ -0,0 +1,74 @@
# What does this PR do?

<!--
Congratulations! You've made it this far! You're not quite done yet though.

Once merged, your PR is going to appear in the release notes with the title you set, so make sure it's a great title that fully reflects the extent of your awesome contribution.

Then, please replace this with a description of the change and which issue is fixed (if applicable). Please also include relevant motivation and context. List any dependencies (if any) that are required for this change.

Once you're done, someone will review your PR shortly (see the section "Who can review?" below to tag some potential reviewers). They may suggest changes to make the code even better. If no one reviewed your PR after a week has passed, don't hesitate to post a new comment @-mentioning the same persons---sometimes notifications get lost.
-->

<!-- Remove if not applicable -->

Fixes # (issue)


## Before submitting
- [ ] This PR fixes a typo or improves the docs (you can dismiss the other checks if that's the case).
- [ ] Did you read the [contributor guideline](https://github.com/huggingface/transformers/blob/master/CONTRIBUTING.md#start-contributing-pull-requests),
Pull Request section?
- [ ] Was this discussed/approved via a Github issue or the [forum](https://discuss.huggingface.co/)? Please add a link
to it if that's the case.
- [ ] Did you make sure to update the documentation with your changes? Here are the
[documentation guidelines](https://github.com/huggingface/transformers/tree/master/docs), and
[here are tips on formatting docstrings](https://github.com/huggingface/transformers/tree/master/docs#writing-source-documentation).
- [ ] Did you write any new necessary tests?


## Who can review?

Anyone in the community is free to review the PR once the tests have passed. Feel free to tag
members/contributors who may be interested in your PR.

<!-- Your PR will be replied to more quickly if you can figure out the right person to tag with @

If you know how to use git blame, that is the easiest way, otherwise, here is a rough guide of **who to tag**.
Please tag fewer than 3 people.

Models:

- albert, bert, xlm: @LysandreJik
- blenderbot, bart, marian, pegasus, encoderdecoder, t5: @patrickvonplaten, @patil-suraj
- longformer, reformer, transfoxl, xlnet: @patrickvonplaten
- fsmt: @stas00
- funnel: @sgugger
- gpt2: @patrickvonplaten, @LysandreJik
- rag: @patrickvonplaten, @lhoestq
- tensorflow: @LysandreJik

Library:

- benchmarks: @patrickvonplaten
- deepspeed: @stas00
- ray/raytune: @richardliaw, @amogkam
- text generation: @patrickvonplaten
- tokenizers: @n1t0, @LysandreJik
- trainer: @sgugger
- pipelines: @LysandreJik

Documentation: @sgugger

HF projects:

- datasets: [different repo](https://github.com/huggingface/datasets)
- rust tokenizers: [different repo](https://github.com/huggingface/tokenizers)

Examples:

- maintained examples (not research project or legacy): @sgugger, @patil-suraj
- research_projects/bert-loses-patience: @JetRunner
- research_projects/distillation: @VictorSanh

-->
1 change: 1 addition & 0 deletions .github/conda/build.sh
Original file line number Diff line number Diff line change
@@ -0,0 +1 @@
$PYTHON setup.py install # Python command to install the script.
Loading