@@ -970,8 +970,8 @@ func @subtensor_insert() {
970
970
%2 = hal.interface.binding.subspan @io ::@ret0 [%c0 ] : !flow.dispatch.tensor <writeonly :?x ?xi32 >
971
971
%3 = flow.dispatch.tensor.load %0 , offsets = [], sizes = [], strides = [] : !flow.dispatch.tensor <readonly :?x ?xi32 > -> tensor <?x?xi32 >
972
972
%4 = flow.dispatch.tensor.load %1 , offsets = [], sizes = [], strides = [] : !flow.dispatch.tensor <readonly :?x ?xi32 > -> tensor <?x?xi32 >
973
- %5 = memref .dim %3 , %c0 : tensor <?x?xi32 >
974
- %6 = memref .dim %3 , %c1 : tensor <?x?xi32 >
973
+ %5 = tensor .dim %3 , %c0 : tensor <?x?xi32 >
974
+ %6 = tensor .dim %3 , %c1 : tensor <?x?xi32 >
975
975
%7 = tensor.insert_slice %3 into %4 [3 , 4 ] [%5 , %6 ] [1 , 1 ] : tensor <?x?xi32 > into tensor <?x?xi32 >
976
976
flow.dispatch.tensor.store %7 , %2 , offsets = [], sizes = [], strides = [] : tensor <?x?xi32 > -> !flow.dispatch.tensor <writeonly :?x ?xi32 >
977
977
return
@@ -1118,8 +1118,8 @@ func @gather() {
1118
1118
%2 = hal.interface.binding.subspan @io ::@ret0 [%c0 ] : !flow.dispatch.tensor <writeonly :?x ?xf32 >
1119
1119
%4 = flow.dispatch.tensor.load %0 , offsets = [], sizes = [], strides = []: !flow.dispatch.tensor <readonly :?x ?xf32 > -> tensor <?x?xf32 >
1120
1120
%5 = flow.dispatch.tensor.load %1 , offsets = [], sizes = [], strides = [] : !flow.dispatch.tensor <readonly :?xi32 > -> tensor <?xi32 >
1121
- %d0 = memref .dim %5 , %c0 : tensor <?xi32 >
1122
- %d1 = memref .dim %4 , %c1 : tensor <?x?xf32 >
1121
+ %d0 = tensor .dim %5 , %c0 : tensor <?xi32 >
1122
+ %d1 = tensor .dim %4 , %c1 : tensor <?x?xf32 >
1123
1123
%3 = linalg.init_tensor [%d0 , %d1 ] : tensor <?x?xf32 >
1124
1124
%7 = linalg.generic {index ing_maps = [affine_map <(d0 , d1 ) -> (d0 )>, affine_map <(d0 , d1 ) -> (d0 , d1 )>], iterator_types = [" parallel" , " parallel" ]} ins (%5 : tensor <?xi32 >) outs (%3 : tensor <?x?xf32 >) {
1125
1125
^bb0 ( %arg2: i32 , %arg3: f32 ): // no predecessors
@@ -1203,8 +1203,8 @@ func @read_only_subtensor() {
1203
1203
%workgroup_count_y = hal.interface.workgroup.count [1 ] : index
1204
1204
%5 = affine.apply affine_map <()[s0 , s1 ] -> (s0 * s1 )>()[%workgroup_id_y , %workgroup_size_y ]
1205
1205
%6 = affine.apply affine_map <()[s0 , s1 ] -> (s0 * s1 )>()[%workgroup_count_y , %workgroup_size_y ]
1206
- %dim0 = memref .dim %2 , %c0 : tensor <?x?xf32 >
1207
- %dim1 = memref .dim %2 , %c1 : tensor <?x?xf32 >
1206
+ %dim0 = tensor .dim %2 , %c0 : tensor <?x?xf32 >
1207
+ %dim1 = tensor .dim %2 , %c1 : tensor <?x?xf32 >
1208
1208
scf.for %arg0 = %5 to %dim0 step %6 {
1209
1209
%7 = affine.apply affine_map <()[s0 , s1 ] -> (s0 * s1 )>()[%workgroup_id_x , %workgroup_size_x ]
1210
1210
%8 = affine.apply affine_map <()[s0 , s1 ] -> (s0 * s1 )>()[%workgroup_count_x , %workgroup_size_x ]
@@ -1263,7 +1263,7 @@ func @reshape_read_only() {
1263
1263
%2 = flow.dispatch.tensor.load %0 , offsets = [], sizes = [], strides = [] : !flow.dispatch.tensor <readonly :?x ?xf32 > -> tensor <?x?xf32 >
1264
1264
%3 = linalg.tensor_collapse_shape %2 [[0 , 1 ]]
1265
1265
: tensor <?x?xf32 > into tensor <?xf32 >
1266
- %4 = memref .dim %3 , %c0 : tensor <?xf32 >
1266
+ %4 = tensor .dim %3 , %c0 : tensor <?xf32 >
1267
1267
%5 = linalg.init_tensor [%4 ] : tensor <?xf32 >
1268
1268
%6 = linalg.generic {
1269
1269
indexing_maps = [affine_map <(d0 ) -> (d0 )>, affine_map <(d0 ) -> (d0 )>],
@@ -1496,8 +1496,8 @@ func @rank_reduced_subtensor_insert() {
1496
1496
%1 = hal.interface.binding.subspan @io ::@ret0 [%c0 ] : !flow.dispatch.tensor <readwrite :?x ?x ?xf32 >
1497
1497
%2 = flow.dispatch.tensor.load %0 , offsets = [], sizes = [], strides = [] : !flow.dispatch.tensor <readonly :?x ?xf32 > -> tensor <?x?xf32 >
1498
1498
%3 = flow.dispatch.tensor.load %1 , offsets = [], sizes = [], strides = [] : !flow.dispatch.tensor <readwrite :?x ?x ?xf32 > -> tensor <?x?x?xf32 >
1499
- %4 = memref .dim %3 , %c1 : tensor <?x?x?xf32 >
1500
- %5 = memref .dim %3 , %c2 : tensor <?x?x?xf32 >
1499
+ %4 = tensor .dim %3 , %c1 : tensor <?x?x?xf32 >
1500
+ %5 = tensor .dim %3 , %c2 : tensor <?x?x?xf32 >
1501
1501
%6 = tensor.insert_slice %2 into %3 [0 , 0 , 0 ] [1 , %4 , %5 ] [1 , 1 , 1 ] : tensor <?x?xf32 > into tensor <?x?x?xf32 >
1502
1502
flow.dispatch.tensor.store %6 , %1 , offsets = [], sizes = [], strides = [] : tensor <?x?x?xf32 > -> !flow.dispatch.tensor <readwrite :?x ?x ?xf32 >
1503
1503
return
@@ -2035,9 +2035,9 @@ module {
2035
2035
%18 = tensor.extract_slice %8 [%arg2 , %arg6 ] [%16 , %17 ] [1 , 1 ] : tensor <?x144 xf32 > to tensor <?x?xf32 >
2036
2036
%19 = affine.min #map5 (%arg4 )
2037
2037
%20 = tensor.extract_slice %10 [%arg6 , %arg4 ] [%17 , %19 ] [1 , 1 ] : tensor <144 x?xf32 > to tensor <?x?xf32 >
2038
- %21 = memref .dim %arg7 , %c0 : tensor <?x?xf32 >
2038
+ %21 = tensor .dim %arg7 , %c0 : tensor <?x?xf32 >
2039
2039
%22 = affine.min #map6 (%21 , %arg2 )
2040
- %23 = memref .dim %arg7 , %c1 : tensor <?x?xf32 >
2040
+ %23 = tensor .dim %arg7 , %c1 : tensor <?x?xf32 >
2041
2041
%24 = affine.min #map6 (%23 , %arg4 )
2042
2042
%25 = tensor.extract_slice %arg7 [%arg2 , %arg4 ] [%22 , %24 ] [1 , 1 ] : tensor <?x?xf32 > to tensor <?x?xf32 >
2043
2043
%26 = linalg.matmul {__internal_linalg_transform__ = " workgroup_l1_tile" , lowering.config = #config1 } ins (%18 , %20 : tensor <?x?xf32 >, tensor <?x?xf32 >) outs (%25 : tensor <?x?xf32 >) -> tensor <?x?xf32 >
@@ -2140,9 +2140,9 @@ module {
2140
2140
%18 = tensor.extract_slice %8 [%arg2 , %arg6 ] [%16 , %17 ] [1 , 1 ] : tensor <?x144 xf32 > to tensor <?x?xf32 >
2141
2141
%19 = affine.min #map5 (%arg4 )
2142
2142
%20 = tensor.extract_slice %10 [%arg6 , %arg4 ] [%17 , %19 ] [1 , 1 ] : tensor <144 x?xf32 > to tensor <?x?xf32 >
2143
- %21 = memref .dim %arg7 , %c0 : tensor <?x?xf32 >
2143
+ %21 = tensor .dim %arg7 , %c0 : tensor <?x?xf32 >
2144
2144
%22 = affine.min #map6 (%21 , %arg2 )
2145
- %23 = memref .dim %arg7 , %c1 : tensor <?x?xf32 >
2145
+ %23 = tensor .dim %arg7 , %c1 : tensor <?x?xf32 >
2146
2146
%24 = affine.min #map6 (%23 , %arg4 )
2147
2147
%25 = tensor.extract_slice %arg7 [%arg2 , %arg4 ] [%22 , %24 ] [1 , 1 ] : tensor <?x?xf32 > to tensor <?x?xf32 >
2148
2148
%26 = linalg.matmul {__internal_linalg_transform__ = " workgroup_l1_tile" , lowering.config = #config1 } ins (%18 , %20 : tensor <?x?xf32 >, tensor <?x?xf32 >) outs (%25 : tensor <?x?xf32 >) -> tensor <?x?xf32 >
@@ -2241,9 +2241,9 @@ module {
2241
2241
%18 = tensor.extract_slice %8 [%arg2 , %arg6 ] [%16 , %17 ] [1 , 1 ] : tensor <?x144 xf32 > to tensor <?x?xf32 >
2242
2242
%19 = affine.min #map5 (%arg4 )
2243
2243
%20 = tensor.extract_slice %10 [%arg6 , %arg4 ] [%17 , %19 ] [1 , 1 ] : tensor <144 x?xf32 > to tensor <?x?xf32 >
2244
- %21 = memref .dim %arg7 , %c0 : tensor <?x?xf32 >
2244
+ %21 = tensor .dim %arg7 , %c0 : tensor <?x?xf32 >
2245
2245
%22 = affine.min #map6 (%21 , %arg2 )
2246
- %23 = memref .dim %arg7 , %c1 : tensor <?x?xf32 >
2246
+ %23 = tensor .dim %arg7 , %c1 : tensor <?x?xf32 >
2247
2247
%24 = affine.min #map6 (%23 , %arg4 )
2248
2248
%25 = tensor.extract_slice %arg7 [%arg2 , %arg4 ] [%22 , %24 ] [1 , 1 ] : tensor <?x?xf32 > to tensor <?x?xf32 >
2249
2249
%26 = linalg.matmul {__internal_linalg_transform__ = " workgroup_l1_tile" , lowering.config = #config1 } ins (%18 , %20 : tensor <?x?xf32 >, tensor <?x?xf32 >) outs (%25 : tensor <?x?xf32 >) -> tensor <?x?xf32 >
0 commit comments