mirror of
https://github.com/saymrwulf/pytorch.git
synced 2026-05-14 20:57:59 +00:00
[torch.compile][ci] Flaky models in CI (similar to DISABLED_TEST) (#128715)
These models are really flaky. I went into the CI machine and ran the model many times, sometime it fails, sometimes it passes. Even Pytorch-eager results change from run to run, so the accuracy comparison is fundamentally broken/non-deterministic. I am hitting these issues more frequently in inlining work. There is nothing wrong with inlining, I think these models are on the edge of already-broken accuracy measurement, and inlining is just pushing it in more broken direction. Pull Request resolved: https://github.com/pytorch/pytorch/pull/128715 Approved by: https://github.com/eellison
This commit is contained in:
parent
2e5366fbc0
commit
9c77332116
1 changed files with 15 additions and 0 deletions
|
|
@ -6,6 +6,14 @@ import textwrap
|
|||
import pandas as pd
|
||||
|
||||
|
||||
# Hack to have something similar to DISABLED_TEST. These models are flaky.
|
||||
|
||||
flaky_models = {
|
||||
"yolov3",
|
||||
"gluon_inception_v3",
|
||||
}
|
||||
|
||||
|
||||
def get_field(csv, model_name: str, field: str):
|
||||
try:
|
||||
return csv.loc[csv["name"] == model_name][field].item()
|
||||
|
|
@ -25,6 +33,13 @@ def check_accuracy(actual_csv, expected_csv, expected_filename):
|
|||
status = "PASS" if expected_accuracy == "pass" else "XFAIL"
|
||||
print(f"{model:34} {status}")
|
||||
continue
|
||||
elif model in flaky_models:
|
||||
if accuracy == "pass":
|
||||
# model passed but marked xfailed
|
||||
status = "PASS_BUT_FLAKY:"
|
||||
else:
|
||||
# model failed but marked passe
|
||||
status = "FAIL_BUT_FLAKY:"
|
||||
elif accuracy != "pass":
|
||||
status = "FAIL:"
|
||||
failed.append(model)
|
||||
|
|
|
|||
Loading…
Reference in a new issue