transformers/tests/bert
Sanchit Gandhi cd9274d010
[FlaxBert] Add ForCausalLM (#16995)
* [FlaxBert] Add ForCausalLM

* make style

* fix output attentions

* Add RobertaForCausalLM

* remove comment

* fix fx-to-pt model loading

* remove comment

* add modeling tests

* add enc-dec model tests

* add big_bird

* add electra

* make style

* make repo-consitency

* add to docs

* remove roberta test

* quality

* amend cookiecutter

* fix attention_mask bug in flax bert model tester

* tighten pt-fx thresholds to 1e-5

* add 'copied from' statements

* amend 'copied from' statements

* amend 'copied from' statements

* quality
2022-05-03 11:26:19 +02:00
..
__init__.py
test_modeling_bert.py
test_modeling_flax_bert.py [FlaxBert] Add ForCausalLM (#16995) 2022-05-03 11:26:19 +02:00
test_modeling_tf_bert.py Use random_attention_mask for TF tests (#16517) 2022-04-01 16:53:07 +02:00
test_tokenization_bert.py