@@ -747,7 +747,7 @@ def func(arg: torch.Tensor) -> torch.Tensor:
747
747
assert "= torch.ops._c10d_functional.wait_tensor.default" not in code
748
748
749
749
# Test aoti
750
- AOTIRunnerUtil .run ("cuda" , func , (arg ,))
750
+ AOTIRunnerUtil .run (func , (arg ,))
751
751
torch .cuda .synchronize ()
752
752
753
753
@unittest .skipIf (not HAS_GPU , "Inductor+gpu needs triton and recent GPU arch" )
@@ -793,7 +793,7 @@ def func(args: list[torch.Tensor]) -> torch.Tensor:
793
793
assert "= torch.ops._c10d_functional.wait_tensor.default" not in code
794
794
795
795
# Test aoti
796
- out = AOTIRunnerUtil .run ("cuda" , func , (args ,)) # noqa: F841
796
+ out = AOTIRunnerUtil .run (func , (args ,)) # noqa: F841
797
797
torch .cuda .synchronize ()
798
798
799
799
@unittest .skipIf (not HAS_GPU , "Inductor+gpu needs triton and recent GPU arch" )
@@ -905,7 +905,7 @@ def func(arg: torch.Tensor) -> torch.Tensor:
905
905
assert "= torch.ops._c10d_functional.wait_tensor.default" not in code
906
906
907
907
# Test aoti
908
- AOTIRunnerUtil .run ("cuda" , func , (arg ,))
908
+ AOTIRunnerUtil .run (func , (arg ,))
909
909
torch .cuda .synchronize ()
910
910
911
911
@unittest .skipIf (not HAS_GPU , "Inductor+gpu needs triton and recent GPU arch" )
@@ -939,7 +939,7 @@ def func(args: list[torch.Tensor]) -> torch.Tensor:
939
939
)
940
940
941
941
# Test aoti
942
- out = AOTIRunnerUtil .run ("cuda" , func , (args ,)) # noqa: F841
942
+ out = AOTIRunnerUtil .run (func , (args ,)) # noqa: F841
943
943
torch .cuda .synchronize ()
944
944
945
945
@unittest .skipIf (not HAS_GPU , "This is a GPU test!" )
@@ -961,7 +961,7 @@ def func(arg: torch.Tensor) -> torch.Tensor:
961
961
)
962
962
963
963
# Test aoti
964
- AOTIRunnerUtil .run ("cuda" , func , (arg ,))
964
+ AOTIRunnerUtil .run (func , (arg ,))
965
965
torch .cuda .synchronize ()
966
966
967
967
@unittest .skipIf (not HAS_GPU , "Inductor+gpu needs triton and recent GPU arch" )
@@ -987,7 +987,7 @@ def func(arg: torch.Tensor) -> torch.Tensor:
987
987
)
988
988
989
989
# Test aoti
990
- AOTIRunnerUtil .run ("cuda" , func , (arg ,))
990
+ AOTIRunnerUtil .run (func , (arg ,))
991
991
torch .cuda .synchronize ()
992
992
993
993
@unittest .skipIf (not HAS_GPU , "Inductor+gpu needs triton and recent GPU arch" )
@@ -1023,7 +1023,7 @@ def func(args: list[torch.Tensor]) -> torch.Tensor:
1023
1023
)
1024
1024
1025
1025
# Test aoti
1026
- AOTIRunnerUtil .run ("cuda" , func , (args ,))
1026
+ AOTIRunnerUtil .run (func , (args ,))
1027
1027
torch .cuda .synchronize ()
1028
1028
1029
1029
@unittest .skipIf (not HAS_GPU , "Inductor+gpu needs triton and recent GPU arch" )
@@ -1108,7 +1108,7 @@ def func(arg: torch.Tensor) -> torch.Tensor:
1108
1108
)
1109
1109
1110
1110
# Test aoti
1111
- AOTIRunnerUtil .run ("cuda" , func , (arg ,))
1111
+ AOTIRunnerUtil .run (func , (arg ,))
1112
1112
torch .cuda .synchronize ()
1113
1113
1114
1114
@unittest .skipIf (not HAS_GPU , "Inductor+gpu needs triton and recent GPU arch" )
0 commit comments