Skip to content

Commit

Permalink
Merge branch 'feature/backport_ea1_ops' into bump_to_aa7e77ee
Browse files Browse the repository at this point in the history
  • Loading branch information
mgehre-amd authored Dec 17, 2024
2 parents 534d8c8 + 9a76fb9 commit 46cbbd0
Show file tree
Hide file tree
Showing 4 changed files with 9 additions and 20 deletions.
2 changes: 1 addition & 1 deletion .github/workflows/ci.yml
Original file line number Diff line number Diff line change
Expand Up @@ -19,7 +19,7 @@ jobs:
build-test-linux:
strategy:
# AMD: Disable fail-fast to see whether failures are different between stable & nightly
# fail-fast: true
fail-fast: false
matrix:
torch-version: [nightly, stable]
name: Build and Test (Linux, torch-${{ matrix.torch-version }}, assertions)
Expand Down
21 changes: 5 additions & 16 deletions projects/pt1/e2e_testing/xfail_sets.py
Original file line number Diff line number Diff line change
Expand Up @@ -36,13 +36,6 @@
"PowIntIntModule_basic",
}

if torch_version_for_comparison() < version.parse("2.5.0.dev"):
# AttributeError: '_OpNamespace' 'aten' object has no attribute '_safe_softmax'
LINALG_XFAIL_SET = LINALG_XFAIL_SET | {
"SafeSoftmaxModule_basic",
"SafeSoftmaxNonNoneDtypeModule_basic",
}

if torch_version_for_comparison() < version.parse("2.5.0.dev"):
LINALG_XFAIL_SET = LINALG_XFAIL_SET | {
# Error: 'torch.aten.scaled_dot_product_attention' op expected 8 operands, but found 7
Expand Down Expand Up @@ -2461,6 +2454,7 @@
| {
### Tests additionally passing in make_fx_tosa
"AdaptiveAvgPool1dStaticLargerOutput_basic",
"AdaptiveAvgPool1dStaticEvenMultiple_basic",
"ScaledDotProductAttentionBoolMaskModule_basic",
"ScaledDotProductAttentionDifferentDynamicCausalModule_basic",
"ArgminIntModule_basic",
Expand Down Expand Up @@ -2517,14 +2511,14 @@
"MaxPool1dStaticModule_basic",
"AdaptiveAvgPool1dNonUnitOutputSizeStaticModule_basic",
"AdaptiveAvgPool1dUnitOutputSizeStaticModule_basic",
"AdaptiveAvgPool1dStaticEvenMultiple_basic",
"CosineSimilarityModule_basic",
"NativeGroupNormBackwardModule_basic",
"ReduceFrobeniusNormKeepDimModule_basic",
"ReduceFrobeniusNormModule_basic",
"SliceWholeTensorModule_basic",
"TensorFloatModule_basic",
"TensorIntModule_basic",
"RepeatInterleaveSelfIntModule_basic",
"AdaptiveAvgPool1dNonUnitOutputSizeStaticModule_basic",
"AdaptiveAvgPool1dUnitOutputSizeStaticModule_basic",
"TorchPrimLoopForLikeTensorArgModule_basic",
Expand All @@ -2542,21 +2536,22 @@
"NormalizeModule_basic",
"ReduceFrobeniusNormKeepDimModule_basic",
"ReduceFrobeniusNormModule_basic",
"ScaledDotProductAttentionBoolMaskModule_basic",
"SliceEndSleStartStaticModule_basic",
"ViewSizeDimFollowedByCollapsedOnesModule_basic",
"ViewSizeDimFollowedByExpandedOnesModule_basic",
"ViewSizeDimLedAndFollowedByCollapsedOnesModule_basic",
"ViewSizeDimLedByCollapsedOnesModule_basic",
"ViewSizeFromOtherTensor_basic",
"RepeatInterleaveSelfIntModule_basic",
"RenormModuleFloat32NegativeDim_basic",
"RenormModuleFloat32_basic",
}
) - {
### Test failing in make_fx_tosa but not in tosa
# Dynamic shape, has extra unsupported broadcast ops
"Matmul_3d",
# Unimplemented operator 'aten._index_put_impl_.hacked_twin'
"IndexPutImpl1DFloatNonAccumulateModule_basic",
"IndexPutImpl1DIntNonAccumulateModule_basic",
# RuntimeError: The size of tensor a (7) must match the size of tensor b (3) at non-singleton dimension 1
"Add_Module_basic",
"Conv2dBiasNoPaddingModule_basic",
Expand All @@ -2567,8 +2562,6 @@
"ElementwisePreluModule_basic",
"ElementwisePreluStaticModule_basic",
"ElementwiseLogSigmoidModule_basic",
"IndexPutImpl1DFloatNonAccumulateModule_basic",
"IndexPutImpl1DIntNonAccumulateModule_basic",
# It appears that you're trying to get value out of a tracing tensor
# failed to legalize operation 'torch.aten.rrelu_with_noise'
"ElementwiseRreluEvalModule_basic",
Expand Down Expand Up @@ -2596,15 +2589,11 @@
MAKE_FX_TOSA_PASS_SET = MAKE_FX_TOSA_PASS_SET | {
"ScaledDotProductAttentionBoolMaskModule_basic",
"ScaledDotProductAttentionDifferentModule_basic",
"ScaledDotProductAttentionDifferentDynamicCausalModule_basic",
"ScaledDotProductAttentionMaskModule_basic",
"ScaledDotProductAttentionSameModule_basic",
}

if torch_version_for_comparison() > version.parse("2.6.0.dev"):
MAKE_FX_TOSA_PASS_SET = MAKE_FX_TOSA_PASS_SET | {
"ScaledDotProductAttentionDifferentDynamicCausalModule_basic",
}
MAKE_FX_TOSA_PASS_SET = MAKE_FX_TOSA_PASS_SET - {
"ChunkListUnpackUneven_Module_basic",
"ChunkListUnpack_Module_basic",
Expand Down
4 changes: 2 additions & 2 deletions stable-requirements.txt
Original file line number Diff line number Diff line change
@@ -1,3 +1,3 @@
--index-url https://download.pytorch.org/whl/cpu
torch==2.3.1+cpu
torchvision==0.18.1+cpu
torch==2.5.1+cpu
torchvision==0.20.1+cpu

0 comments on commit 46cbbd0

Please sign in to comment.