mirror of
https://github.com/saymrwulf/transformers.git
synced 2026-05-14 20:58:08 +00:00
[CI] Don't test apex (#6021)
This commit is contained in:
parent
a884b7fa38
commit
c69ea5efc4
2 changed files with 2 additions and 3 deletions
|
|
@ -55,9 +55,8 @@ def test_train_mbart_cc25_enro_script():
|
|||
if CUDA_AVAILABLE:
|
||||
gpus = 1 # torch.cuda.device_count()
|
||||
else:
|
||||
bash_script = bash_script.replace("--fp16", "")
|
||||
gpus = 0
|
||||
|
||||
bash_script = bash_script.replace("--fp16", "")
|
||||
testargs = (
|
||||
["finetune.py"]
|
||||
+ bash_script.split()
|
||||
|
|
|
|||
|
|
@ -43,7 +43,7 @@ CHEAP_ARGS = {
|
|||
"student_decoder_layers": 1,
|
||||
"val_check_interval": 1.0,
|
||||
"output_dir": "",
|
||||
"fp16": CUDA_AVAILABLE,
|
||||
"fp16": False, # TODO(SS): set this to CUDA_AVAILABLE if ci installs apex or start using native amp
|
||||
"no_teacher": False,
|
||||
"fp16_opt_level": "O1",
|
||||
"gpus": 1 if CUDA_AVAILABLE else 0,
|
||||
|
|
|
|||
Loading…
Reference in a new issue