Skip to content

Commit

Permalink
Merge pull request #425 from Xilinx/matthias.fix_feature
Browse files Browse the repository at this point in the history
Fix xfail on feature branch
  • Loading branch information
mgehre-amd authored Dec 17, 2024
2 parents 245c924 + 16fa61f commit 4af5bb8
Show file tree
Hide file tree
Showing 2 changed files with 8 additions and 17 deletions.
2 changes: 1 addition & 1 deletion .github/workflows/ci.yml
Original file line number Diff line number Diff line change
Expand Up @@ -19,7 +19,7 @@ jobs:
build-test-linux:
strategy:
# AMD: Disable fail-fast to see whether failures are different between stable & nightly
# fail-fast: true
fail-fast: false
matrix:
torch-version: [nightly, stable]
name: Build and Test (Linux, torch-${{ matrix.torch-version }}, assertions)
Expand Down
23 changes: 7 additions & 16 deletions projects/pt1/e2e_testing/xfail_sets.py
Original file line number Diff line number Diff line change
Expand Up @@ -36,13 +36,6 @@
"PowIntIntModule_basic",
}

if torch_version_for_comparison() < version.parse("2.5.0.dev"):
# AttributeError: '_OpNamespace' 'aten' object has no attribute '_safe_softmax'
LINALG_XFAIL_SET = LINALG_XFAIL_SET | {
"SafeSoftmaxModule_basic",
"SafeSoftmaxNonNoneDtypeModule_basic",
}

if torch_version_for_comparison() < version.parse("2.5.0.dev"):
LINALG_XFAIL_SET = LINALG_XFAIL_SET | {
# Error: 'torch.aten.scaled_dot_product_attention' op expected 8 operands, but found 7
Expand Down Expand Up @@ -2390,6 +2383,9 @@
TOSA_PASS_SET
| {
### Tests additionally passing in make_fx_tosa
"AdaptiveAvgPool1dStaticEvenMultiple_basic",
"ScaledDotProductAttentionBoolMaskModule_basic",
"ScaledDotProductAttentionDifferentDynamicCausalModule_basic",
"ArgminIntModule_basic",
"ArgminIntModule_multiple_mins",
"ArgminModule_basic",
Expand Down Expand Up @@ -2444,14 +2440,14 @@
"MaxPool1dStaticModule_basic",
"AdaptiveAvgPool1dNonUnitOutputSizeStaticModule_basic",
"AdaptiveAvgPool1dUnitOutputSizeStaticModule_basic",
"AdaptiveAvgPool1dStaticEvenMultiple_basic",
"CosineSimilarityModule_basic",
"NativeGroupNormBackwardModule_basic",
"ReduceFrobeniusNormKeepDimModule_basic",
"ReduceFrobeniusNormModule_basic",
"SliceWholeTensorModule_basic",
"TensorFloatModule_basic",
"TensorIntModule_basic",
"RepeatInterleaveSelfIntModule_basic",
"AdaptiveAvgPool1dNonUnitOutputSizeStaticModule_basic",
"AdaptiveAvgPool1dUnitOutputSizeStaticModule_basic",
"TorchPrimLoopForLikeTensorArgModule_basic",
Expand All @@ -2469,21 +2465,22 @@
"NormalizeModule_basic",
"ReduceFrobeniusNormKeepDimModule_basic",
"ReduceFrobeniusNormModule_basic",
"ScaledDotProductAttentionBoolMaskModule_basic",
"SliceEndSleStartStaticModule_basic",
"ViewSizeDimFollowedByCollapsedOnesModule_basic",
"ViewSizeDimFollowedByExpandedOnesModule_basic",
"ViewSizeDimLedAndFollowedByCollapsedOnesModule_basic",
"ViewSizeDimLedByCollapsedOnesModule_basic",
"ViewSizeFromOtherTensor_basic",
"RepeatInterleaveSelfIntModule_basic",
"RenormModuleFloat32NegativeDim_basic",
"RenormModuleFloat32_basic",
}
) - {
### Test failing in make_fx_tosa but not in tosa
# Dynamic shape, has extra unsupported broadcast ops
"Matmul_3d",
# Unimplemented operator 'aten._index_put_impl_.hacked_twin'
"IndexPutImpl1DFloatNonAccumulateModule_basic",
"IndexPutImpl1DIntNonAccumulateModule_basic",
# RuntimeError: The size of tensor a (7) must match the size of tensor b (3) at non-singleton dimension 1
"Add_Module_basic",
"Conv2dBiasNoPaddingModule_basic",
Expand All @@ -2494,8 +2491,6 @@
"ElementwisePreluModule_basic",
"ElementwisePreluStaticModule_basic",
"ElementwiseLogSigmoidModule_basic",
"IndexPutImpl1DFloatNonAccumulateModule_basic",
"IndexPutImpl1DIntNonAccumulateModule_basic",
# It appears that you're trying to get value out of a tracing tensor
# failed to legalize operation 'torch.aten.rrelu_with_noise'
"ElementwiseRreluEvalModule_basic",
Expand Down Expand Up @@ -2524,15 +2519,11 @@
MAKE_FX_TOSA_PASS_SET = MAKE_FX_TOSA_PASS_SET | {
"ScaledDotProductAttentionBoolMaskModule_basic",
"ScaledDotProductAttentionDifferentModule_basic",
"ScaledDotProductAttentionDifferentDynamicCausalModule_basic",
"ScaledDotProductAttentionMaskModule_basic",
"ScaledDotProductAttentionSameModule_basic",
}

if torch_version_for_comparison() > version.parse("2.6.0.dev"):
MAKE_FX_TOSA_PASS_SET = MAKE_FX_TOSA_PASS_SET | {
"ScaledDotProductAttentionDifferentDynamicCausalModule_basic",
}
MAKE_FX_TOSA_PASS_SET = MAKE_FX_TOSA_PASS_SET - {
"ChunkListUnpackUneven_Module_basic",
"ChunkListUnpack_Module_basic",
Expand Down

0 comments on commit 4af5bb8

Please sign in to comment.