transformers/src/transformers
Younes Belkada 78a93d17c0
[GPTNeo] Fix gradient checkpointing bug (#21733)
* fix bug

* forward contrib credits from discussions

* change logic

---------

Co-authored-by: edbeeching <edbeeching@users.noreply.github.com>
2023-02-23 09:48:19 +01:00
..
benchmark Apply ruff flake8-comprehensions (#21694) 2023-02-22 09:14:54 +01:00
commands Update quality tooling for formatting (#21480) 2023-02-06 18:10:56 -05:00
data Cleanup quality (#21493) 2023-02-07 12:27:31 -05:00
generation Apply ruff flake8-comprehensions (#21694) 2023-02-22 09:14:54 +01:00
models [GPTNeo] Fix gradient checkpointing bug (#21733) 2023-02-23 09:48:19 +01:00
onnx Apply ruff flake8-comprehensions (#21694) 2023-02-22 09:14:54 +01:00
pipelines Apply ruff flake8-comprehensions (#21694) 2023-02-22 09:14:54 +01:00
sagemaker Cleanup quality (#21493) 2023-02-07 12:27:31 -05:00
utils Apply ruff flake8-comprehensions (#21694) 2023-02-22 09:14:54 +01:00
__init__.py Add WhisperTokenizerFast (#21222) 2023-02-21 06:58:54 +01:00
activations.py Add the GeLU activation from pytorch with the tanh approximation (#21345) 2023-02-02 09:33:04 -05:00
activations_tf.py
audio_utils.py [CLAP] Add CLAP to the library (#21370) 2023-02-16 20:59:27 +01:00
configuration_utils.py Apply ruff flake8-comprehensions (#21694) 2023-02-22 09:14:54 +01:00
convert_graph_to_onnx.py Update quality tooling for formatting (#21480) 2023-02-06 18:10:56 -05:00
convert_pytorch_checkpoint_to_tf2.py Update quality tooling for formatting (#21480) 2023-02-06 18:10:56 -05:00
convert_slow_tokenizer.py Add WhisperTokenizerFast (#21222) 2023-02-21 06:58:54 +01:00
convert_slow_tokenizers_checkpoints_to_fast.py
convert_tf_hub_seq_to_seq_bert_to_pytorch.py
debug_utils.py
deepspeed.py Apply ruff flake8-comprehensions (#21694) 2023-02-22 09:14:54 +01:00
dependency_versions_check.py
dependency_versions_table.py Sort deps alphabetically 2023-02-16 13:27:27 -05:00
dynamic_module_utils.py Fix get_class_in_module (#21709) 2023-02-21 09:39:15 +01:00
feature_extraction_sequence_utils.py Apply ruff flake8-comprehensions (#21694) 2023-02-22 09:14:54 +01:00
feature_extraction_utils.py Little cleanup: let huggingface_hub manage token retrieval (#21333) 2023-01-27 12:09:49 -05:00
file_utils.py Cleanup quality (#21493) 2023-02-07 12:27:31 -05:00
generation_flax_utils.py
generation_tf_utils.py
generation_utils.py
hf_argparser.py fix typo in comment (#21088) 2023-01-11 17:51:41 +01:00
image_processing_utils.py Make ImageProcessorMixin compatible with subfolder kwarg (#21725) 2023-02-23 09:28:18 +01:00
image_transforms.py Move convert_to_rgb to image_transforms module (#20784) 2022-12-15 18:47:04 +00:00
image_utils.py Apply ruff flake8-comprehensions (#21694) 2023-02-22 09:14:54 +01:00
integrations.py Apply ruff flake8-comprehensions (#21694) 2023-02-22 09:14:54 +01:00
keras_callbacks.py Fix to KerasMetricCallback when the model returns unstructured output (#21727) 2023-02-22 13:15:14 +00:00
modelcard.py Apply ruff flake8-comprehensions (#21694) 2023-02-22 09:14:54 +01:00
modeling_flax_outputs.py
modeling_flax_pytorch_utils.py Apply ruff flake8-comprehensions (#21694) 2023-02-22 09:14:54 +01:00
modeling_flax_utils.py Apply ruff flake8-comprehensions (#21694) 2023-02-22 09:14:54 +01:00
modeling_outputs.py add GPTSAN model (reopen) (#21291) 2023-02-20 11:25:27 +01:00
modeling_tf_outputs.py
modeling_tf_pytorch_utils.py Apply ruff flake8-comprehensions (#21694) 2023-02-22 09:14:54 +01:00
modeling_tf_utils.py Apply ruff flake8-comprehensions (#21694) 2023-02-22 09:14:54 +01:00
modeling_utils.py Apply ruff flake8-comprehensions (#21694) 2023-02-22 09:14:54 +01:00
optimization.py Apply ruff flake8-comprehensions (#21694) 2023-02-22 09:14:54 +01:00
optimization_tf.py Apply ruff flake8-comprehensions (#21694) 2023-02-22 09:14:54 +01:00
processing_utils.py refactor: Make direct_transformers_import util (#21652) 2023-02-16 11:32:32 -05:00
pytorch_utils.py [Past CI] 🔥 Leave Past CI failures in the past 🔥 (#20861) 2022-12-27 18:37:25 +01:00
testing_utils.py Fix env. variable type issue in testing (#21609) 2023-02-13 20:53:26 +01:00
tf_utils.py
tokenization_utils.py Apply ruff flake8-comprehensions (#21694) 2023-02-22 09:14:54 +01:00
tokenization_utils_base.py Apply ruff flake8-comprehensions (#21694) 2023-02-22 09:14:54 +01:00
tokenization_utils_fast.py Apply ruff flake8-comprehensions (#21694) 2023-02-22 09:14:54 +01:00
trainer.py Apply ruff flake8-comprehensions (#21694) 2023-02-22 09:14:54 +01:00
trainer_callback.py
trainer_pt_utils.py Apply ruff flake8-comprehensions (#21694) 2023-02-22 09:14:54 +01:00
trainer_seq2seq.py Update quality tooling for formatting (#21480) 2023-02-06 18:10:56 -05:00
trainer_tf.py Update quality tooling for formatting (#21480) 2023-02-06 18:10:56 -05:00
trainer_utils.py Apply ruff flake8-comprehensions (#21694) 2023-02-22 09:14:54 +01:00
training_args.py Respect documentation on passive log level (#21700) 2023-02-22 09:39:18 +01:00
training_args_seq2seq.py
training_args_tf.py Replaces xxx_required with requires_backends (#20715) 2022-12-14 14:38:44 +00:00