/external/llvm-project/mlir/test/Dialect/Linalg/ |
D | fusion-tensor.mlir | 4 #map0 = affine_map<(d0, d1) -> (d0, d1)> 9 …%0 = linalg.generic {indexing_maps = [#map0, #map0, #map0], iterator_types = ["parallel", "paralle… 17 …%2 = linalg.generic {indexing_maps = [#map0, #map0, #map0], iterator_types = ["parallel", "paralle… 38 #map0 = affine_map<(d0, d1) -> (d0, d1)> 44 …%0 = linalg.generic {indexing_maps = [#map0, #map1, #map0], iterator_types = ["parallel", "paralle… 52 …%2 = linalg.generic {indexing_maps = [#map0, #map0, #map0], iterator_types = ["parallel", "paralle… 65 #map0 = affine_map<(d0, d1) -> (d0, d1)> 71 …%0 = linalg.generic {indexing_maps = [#map0, #map1, #map0], iterator_types = ["parallel", "paralle… 79 …%2 = linalg.generic {indexing_maps = [#map1, #map0, #map0], iterator_types = ["parallel", "paralle… 92 #map0 = affine_map<(d0, d1) -> (d0, d1)> [all …]
|
D | reshape_linearization_fusion.mlir | 7 #map0 = affine_map<(d0, d1, d2, d3) -> (d0, d1, d2, d3)> 17 indexing_maps = [#map0, #map0, #map0], 38 #map0 = affine_map<(d0, d1, d2, d3) -> (d0, d1, d2, d3)> 44 indexing_maps = [#map0, #map0, #map0], 64 #map0 = affine_map<(d0, d1, d2, d3) -> (d0, d1, d2, d3)> 70 indexing_maps = [#map0, #map0, #map0], 92 #map0 = affine_map<(d0, d1, d2, d3) -> (d0, d1, d2, d3)> 100 indexing_maps = [#map0, #map0], 122 #map0 = affine_map<(d0, d1, d2, d3) -> (d0, d1, d2, d3)> 126 indexing_maps = [#map0, #map0], [all …]
|
D | fold-unit-trip-loops.mlir | 32 #map0 = affine_map<(i, j) -> (i, j)> 33 #access = [#map0, #map0] 57 #map0 = affine_map<(i, j) -> (i, j)> 58 #access = [#map0, #map0]
|
D | bufferize.mlir | 3 #map0 = affine_map<(d0) -> (d0)> 27 indexing_maps = [#map0, #map0], 40 #map0 = affine_map<(d0) -> (d0)> 51 indexing_maps = [#map0, #map0, #map0], 63 #map0 = affine_map<(d0) -> (d0)> 75 indexing_maps = [#map0, #map0, #map0], 94 // CHECK: #map0 = affine_map<(d0, d1) -> (d0, d1)> 105 // CHECK: linalg.generic {indexing_maps = [#map0, #map0, #map1]
|
D | canonicalize-duplicate-inputs.mlir | 26 #map0 = affine_map<(d0, d1) -> (d0, d1)> 34 …%0 = linalg.generic {indexing_maps = [#map0, #map1, #map0], iterator_types = ["parallel", "paralle… 47 #map0 = affine_map<(d0, d1) -> (d0, d1)> 57 …%0 = linalg.generic {indexing_maps = [#map0, #map1, #map0, #map0], iterator_types = ["parallel", "…
|
D | reshape_fusion.mlir | 3 #map0 = affine_map<(d0, d1, d2) -> (d2, d0, d1)> 14 indexing_maps = [#map0, #map1, #map1], 46 #map0 = affine_map<(d0, d1) -> (d0, d1)> 52 indexing_maps = [#map0, #map0, #map0], 131 #map0 = affine_map<(d0, d1) -> (d0, d1)> 139 indexing_maps = [#map0, #map0, #map0], 196 #map0 = affine_map<(d0, d1, d2) -> (d2, d0, d1)> 207 indexing_maps = [#map0, #map1, #map1], 244 #map0 = affine_map<(d0, d1) -> (d0, d1)> 250 indexing_maps = [#map0, #map0, #map0], [all …]
|
D | parallel-loops.mlir | 3 #map0 = affine_map<(d0, d1) -> (d0, d1)> 8 indexing_maps = [#map0, #map0, #map0],
|
D | standard.mlir | 3 // CHECK-DAG: #[[$map0:.*]] = affine_map<(d0)[s0] -> (d0 + s0)> 20 // CHECK-SAME: %[[arg0:[a-zA-z0-9]*]]: memref<?xf32, #[[$map0]]>, 21 // CHECK-SAME: %[[arg1:[a-zA-z0-9]*]]: memref<?xf32, #[[$map0]]>, 24 // CHECK-SAME: memref<?xf32, #[[$map0]]> to memref<?xf32, #[[$map6]]> 26 // CHECK-SAME: memref<?xf32, #[[$map0]]> to memref<?xf32, #[[$map6]]>
|
D | drop-unit-extent-dims.mlir | 80 #map0 = affine_map<(i, j) -> (i, j)> 81 #access = [#map0, #map0] 106 #map0 = affine_map<(i, j) -> (i, j)> 107 #access = [#map0, #map0]
|
D | fusion.mlir | 278 #map0 = affine_map<(d0) -> (d0 + 2)> 305 %3 = affine.apply #map0(%arg5) 420 #map0 = affine_map<(d0) -> (d0 + 2)> 447 %3 = affine.apply #map0(%arg5) 595 #map0 = affine_map<(d0, d1) -> (d0)> 606 indexing_maps = [#map0, #map1], 664 #map0 = affine_map<(d0, d1, d2) -> (d0, d1 - d2)> 690 %15 = affine.min #map0(%c2, %c1, %arg3) 692 %17 = affine.min #map0(%16, %c4, %arg4) 696 %21 = affine.min #map0(%c2, %c1, %arg3) [all …]
|
D | tile-parallel-reduce.mlir | 47 #map0 = affine_map<(d0, d1, d2) -> (d0, d1, d2)> 50 #accesses = [#map0, #map1, #map2]
|
/external/llvm-project/mlir/test/Transforms/ |
D | normalize-memrefs-ops.mlir | 12 #map0 = affine_map<(d0, d1, d2, d3) -> (d0, d1, d2 floordiv 32, d3 floordiv 64, d2 mod 32, d3 mod 6… 18 func @test_norm(%arg0 : memref<1x16x14x14xf32, #map0>) -> () { 19 %0 = alloc() : memref<1x16x14x14xf32, #map0> 20 "test.op_norm"(%arg0, %0) : (memref<1x16x14x14xf32, #map0>, memref<1x16x14x14xf32, #map0>) -> () 21 dealloc %0 : memref<1x16x14x14xf32, #map0> 33 func @test_nonnorm(%arg0 : memref<1x16x14x14xf32, #map0>) -> () { 34 %0 = alloc() : memref<1x16x14x14xf32, #map0> 35 …"test.op_nonnorm"(%arg0, %0) : (memref<1x16x14x14xf32, #map0>, memref<1x16x14x14xf32, #map0>) -> () 36 dealloc %0 : memref<1x16x14x14xf32, #map0> 49 %0 = alloc() : memref<1x16x14x14xf32, #map0> [all …]
|
D | copy-removal.mlir | 163 #map0 = affine_map<(d0) -> (d0)> 172 indexing_maps = [#map0, #map0], 239 #map0 = affine_map<(d0) -> (d0)> 251 indexing_maps = [#map0, #map0], 270 #map0 = affine_map<(d0) -> (d0)> 277 indexing_maps = [#map0, #map0], 286 indexing_maps = [#map0, #map0],
|
D | cse.mlir | 4 #map0 = affine_map<(d0) -> (d0 mod 2)> 23 %0 = affine.apply #map0(%c0) 24 %1 = affine.apply #map0(%c1)
|
/external/tensorflow/tensorflow/compiler/mlir/hlo/tests/ |
D | lhlo-fuse-linalg.mlir | 5 #map0 = affine_map<(d0, d1) -> (d0, d1)> 6 #pointwise_2d_trait = {indexing_maps = [#map0, #map0, #map0], 138 #map0 = affine_map<(d0, d1, d2, d3) -> (d0, d1, d2, d3)> 139 #pointwise_4d_trait = {indexing_maps = [#map0, #map0, #map0], 198 #map0 = affine_map<(d0, d1) -> (d0, d1)> 199 #pointwise_2d_trait = {indexing_maps = [#map0, #map0, #map0],
|
/external/llvm-project/mlir/test/Conversion/SCFToGPU/ |
D | parallel_loop.mlir | 200 #map0 = affine_map<(d0, d1)[s0, s1] -> (d0 * s1 + s0 + d1)> 206 …func @sum(%arg0: memref<?x?xf32, #map0>, %arg1: memref<?x?xf32, #map0>, %arg2: memref<?x?xf32, #ma… 211 %0 = dim %arg0, %c0 : memref<?x?xf32, #map0> 212 %1 = dim %arg0, %c1 : memref<?x?xf32, #map0> 214 %2 = dim %arg0, %c0 : memref<?x?xf32, #map0> 217 %4 = dim %arg0, %c1 : memref<?x?xf32, #map0> 219 …%6 = std.subview %arg0[%arg3, %arg4][%squared_min, %5][%c1, %c1] : memref<?x?xf32, #map0> to memre… 220 %7 = dim %arg1, %c0 : memref<?x?xf32, #map0> 222 %9 = dim %arg1, %c1 : memref<?x?xf32, #map0> 224 …%11 = std.subview %arg1[%arg3, %arg4][%8, %10][%c1, %c1] : memref<?x?xf32, #map0> to memref<?x?xf3… [all …]
|
/external/llvm-project/mlir/test/Dialect/SCF/ |
D | parallel-loop-specialization.mlir | 3 #map0 = affine_map<()[s0, s1] -> (1024, s0 - s1)> 12 %b0 = affine.min #map0()[%d0, %outer_i0] 29 // CHECK: [[VAL_10:%.*]] = affine.min #map0(){{\[}}[[VAL_8]], [[VAL_0]]]
|
D | for-loop-specialization.mlir | 3 #map0 = affine_map<()[s0, s1] -> (1024, s0 - s1)> 11 %b0 = affine.min #map0()[%d0, %outer]
|
/external/llvm-project/mlir/integration_test/Dialect/Vector/CPU/ |
D | test-transfer-to-loops.mlir | 6 #map0 = affine_map<(d0, d1) -> (d1, d0)> 57 …%1 = vector.transfer_read %0[%c1, %c1], %cst {permutation_map = #map0} : memref<?x?xf32>, vector<5… 67 …vector.transfer_write %1, %0[%c0, %c0] {permutation_map = #map0} : vector<5x5xf32>, memref<?x?xf32> 90 …%4 = vector.transfer_read %0[%c2, %c3], %cst {permutation_map = #map0} : memref<?x?xf32>, vector<5…
|
/external/llvm-project/mlir/integration_test/Dialect/Async/CPU/ |
D | microbench-linalg-async-parallel-for.mlir | 31 #map0 = affine_map<(d0, d1) -> (d0, d1)> 37 indexing_maps = [#map0, #map0, #map0],
|
/external/llvm-project/mlir/test/Dialect/Vector/ |
D | vector-transforms.mlir | 96 // CHECK-NEXT: %[[R1S00:.*]] = vector.contract {indexing_maps = [#map0, #map1, #map2], iterator_typ… 102 // CHECK-NEXT: %[[R2S00:.*]] = vector.contract {indexing_maps = [#map0, #map1, #map2], iterator_typ… 108 // CHECK-NEXT: %[[R3S00:.*]] = vector.contract {indexing_maps = [#map0, #map1, #map2], iterator_typ… 115 // CHECK-NEXT: %[[R1S02:.*]] = vector.contract {indexing_maps = [#map0, #map1, #map2], iterator_typ… 119 // CHECK-NEXT: %[[R2S02:.*]] = vector.contract {indexing_maps = [#map0, #map1, #map2], iterator_typ… 123 // CHECK-NEXT: %[[R3S02:.*]] = vector.contract {indexing_maps = [#map0, #map1, #map2], iterator_typ… 130 // CHECK-NEXT: %[[R1S20:.*]] = vector.contract {indexing_maps = [#map0, #map1, #map2], iterator_typ… 134 // CHECK-NEXT: %[[R2S20:.*]] = vector.contract {indexing_maps = [#map0, #map1, #map2], iterator_ty… 138 // CHECK-NEXT: %[[R3S20:.*]] = vector.contract {indexing_maps = [#map0, #map1, #map2], iterator_ty… 143 // CHECK-NEXT: %[[R1S22:.*]] = vector.contract {indexing_maps = [#map0, #map1, #map2], iterator_typ… [all …]
|
/external/llvm-project/mlir/docs/Dialects/ |
D | Linalg.md | 102 #map0 = affine_map<(d0) -> (d0 * 2 + 1)> 104 func @example(%arg0: memref<?xf32>, %arg1: memref<?xvector<4xf32>, #map0>) { 110 %2 = load %arg1[%arg2] : memref<?xvector<4xf32>, #map0> 112 store %3, %arg1[%arg2] : memref<?xvector<4xf32>, #map0> 182 #map0 = affine_map<(d0, d1) -> (d0 * 2 + d1 * 2)> 184 func @example(%arg0: memref<8x?xf32, #map0>, %arg1: memref<?xvector<4xf32>>) { 188 %0 = dim %arg0, %c1 : memref<8x?xf32, #map0> 191 %1 = load %arg0[%arg3, %arg2] : memref<8x?xf32, #map0> 374 #map0 = affine_map<(d0, d1)[s0, s1, s2] -> (d0 * s1 + s0 + d1 * s2)> 377 %0 = memref_cast %arg0 : memref<?x?xf32> to memref<?x?xf32, #map0> [all …]
|
/external/llvm-project/mlir/test/Conversion/LinalgToVector/ |
D | linalg-to-vector.mlir | 3 // CHECK-DAG: #[[$map0:.*]] = affine_map<(d0)[s0] -> (1, -d0 + s0)> 35 // CHECK: %[[v9:.*]] = affine.min #[[$map0]](%[[arg3]])[%[[v1]]]
|
/external/llvm-project/mlir/test/Dialect/Affine/ |
D | invalid.mlir | 68 #map0 = affine_map<(d0)[s0] -> (d0 + s0)> 73 affine.for %n0 = #map0(%i0)[%i0] to 7 { 81 #map0 = affine_map<(d0)[s0] -> (d0 + s0)> 86 affine.for %n0 = 0 to #map0(%i0)[%i0] {
|
/external/llvm-project/mlir/test/IR/ |
D | invalid.mlir | 44 #map0 = affine_map<(d0, d1) -> (d0, d1)> 46 func @memrefs(memref<2x4xi8, #map0, #map8>) // expected-error {{undefined symbol alias id 'map8'}} 50 #map0 = affine_map<(d0, d1) -> (d0, d1)> 51 func @memrefs(memref<2x4xi8, #map0, 1, 2>) // expected-error {{multiple memory spaces specified in … 55 #map0 = affine_map<(d0, d1) -> (d0, d1)> 58 func @memrefs(memref<2x4xi8, #map0, 1, #map1>) // expected-error {{expected memory space to be last… 64 #map0 = affine_map<(d0, d1) -> (d0, d1)> 65 func @memrefs(memref<42xi8, #map0>) // expected-error {{memref affine map dimension mismatch}} 69 #map0 = affine_map<(d0, d1) -> (d0, d1)> 71 func @memrefs(memref<42x42xi8, #map0, #map1>) // expected-error {{memref affine map dimension misma… [all …]
|