mirror of
https://github.com/huggingface/transformers.git
synced 2025-07-29 17:22:25 +06:00
[CI] Don't test apex (#6021)
This commit is contained in:
parent
a884b7fa38
commit
c69ea5efc4
@ -55,9 +55,8 @@ def test_train_mbart_cc25_enro_script():
|
||||
if CUDA_AVAILABLE:
|
||||
gpus = 1 # torch.cuda.device_count()
|
||||
else:
|
||||
bash_script = bash_script.replace("--fp16", "")
|
||||
gpus = 0
|
||||
|
||||
bash_script = bash_script.replace("--fp16", "")
|
||||
testargs = (
|
||||
["finetune.py"]
|
||||
+ bash_script.split()
|
||||
|
@ -43,7 +43,7 @@ CHEAP_ARGS = {
|
||||
"student_decoder_layers": 1,
|
||||
"val_check_interval": 1.0,
|
||||
"output_dir": "",
|
||||
"fp16": CUDA_AVAILABLE,
|
||||
"fp16": False, # TODO(SS): set this to CUDA_AVAILABLE if ci installs apex or start using native amp
|
||||
"no_teacher": False,
|
||||
"fp16_opt_level": "O1",
|
||||
"gpus": 1 if CUDA_AVAILABLE else 0,
|
||||
|
Loading…
Reference in New Issue
Block a user