diff --git a/mlir/lib/Dialect/Linalg/Transforms/FusionOnTensors.cpp b/mlir/lib/Dialect/Linalg/Transforms/FusionOnTensors.cpp --- a/mlir/lib/Dialect/Linalg/Transforms/FusionOnTensors.cpp +++ b/mlir/lib/Dialect/Linalg/Transforms/FusionOnTensors.cpp @@ -145,11 +145,14 @@ fusedBlock->getArguments().take_front(numFusedOpIndices)); mapper.map(std::get<0>(it), newIndex); } - // 2b. Replace the producer index operations by index operations placed in the - // fused block using the `consumerToProducerLoopsMap` to map the index spaces. + // 2b. Add an index operation for every fused loop dimension and use the + // `consumerToProducerLoopsMap` to replace the producer index operations. unsigned numFusedOpLoops = std::max(producer.getNumLoops(), consumer.getNumLoops()); if (producer.hasIndexSemantics()) { + // Add an index operation for every fused loop dimension. + OpBuilder::InsertionGuard guard(rewriter); + rewriter.setInsertionPointToStart(&producerBlock); SmallVector fusedIndices; fusedIndices.reserve(numFusedOpLoops); llvm::transform(llvm::seq(0, numFusedOpLoops), @@ -158,12 +161,14 @@ }); for (IndexOp indexOp : llvm::make_early_inc_range(producerBlock.getOps())) { + // Skip index operations having no use, especially the 'fusedIndices'. + if (indexOp->use_empty()) + continue; + // Use a mapping of the fused indices to replace the index operation. + rewriter.setInsertionPoint(indexOp); Value newIndex = rewriter.create( producer.getLoc(), consumerToProducerLoopsMap.getSubMap(indexOp.dim()), fusedIndices); - // Replace the producer index operation by the index value computed in the - // fused block. All remaining operations in the producer block are later - // on cloned to the fused block. rewriter.replaceOp(indexOp, newIndex); } } diff --git a/mlir/test/Dialect/Linalg/fusion-tensor.mlir b/mlir/test/Dialect/Linalg/fusion-tensor.mlir --- a/mlir/test/Dialect/Linalg/fusion-tensor.mlir +++ b/mlir/test/Dialect/Linalg/fusion-tensor.mlir @@ -462,8 +462,7 @@ // ----- #map0 = affine_map<(d0, d1) -> (d0, d1)> -func @indexed_producer_consumer_fusion(%arg0: tensor, - %arg1: tensor) -> tensor { +func @indexed_producer_consumer_fusion(%arg0: tensor) -> tensor { %c0 = constant 0 : index %c1 = constant 1 : index %0 = memref.dim %arg0, %c0 : tensor @@ -486,7 +485,7 @@ %4 = linalg.generic { indexing_maps = [#map0, #map0, #map0], iterator_types = ["parallel", "parallel"] } - ins(%3, %arg1 : tensor, tensor) + ins(%3, %arg0 : tensor, tensor) outs(%2 : tensor) { ^bb0(%arg2: i32, %arg3: i32, %arg4: i32): // no predecessors %10 = addi %arg2, %arg3 : i32 @@ -497,7 +496,7 @@ // CHECK-DAG: #[[$MAP0:.*]] = affine_map<(d0, d1) -> (d0, d1)> // CHECK-LABEL: func @indexed_producer_consumer_fusion // CHECK: linalg.generic -// CHECK-SAME: indexing_maps = [#[[$MAP0]], #[[$MAP0]], #[[$MAP0]]] +// CHECK-SAME: indexing_maps = [#[[$MAP0]], #[[$MAP0]]] // CHECK: ^{{[a-zA-Z0-9_]*}} // CHECK-SAME: %[[ARG0:[a-zA-Z0-9_]*]]: i32 // CHECK-SAME: %[[ARG1:[a-zA-Z0-9_]*]]: i32 @@ -507,7 +506,7 @@ // CHECK: %[[SUB_OPERAND:.+]] = index_cast %[[IDX1]] : index to i32 // CHECK: %[[VAL1:.+]] = addi %[[ARG0]], %[[ADD_OPERAND]] : i32 // CHECK: %[[VAL2:.+]] = subi %[[VAL1]], %[[SUB_OPERAND]] : i32 -// CHECK: %[[VAL3:.+]] = addi %[[VAL2]], %[[ARG1]] : i32 +// CHECK: %[[VAL3:.+]] = addi %[[VAL2]], %[[ARG0]] : i32 // CHECK: linalg.yield %[[VAL3]] : i32 // CHECK-NOT: linalg.generic