let results = (outs AnyMemRef);
let extraClassDeclaration = [{
- /// The result of a memref_shape_cast is always a memref.
+ /// The result of a view is always a memref.
MemRefType getType() { return getResult()->getType().cast<MemRefType>(); }
- /// Returns the dynamic offset for this shape cast operation if specified.
+ /// Returns the dynamic offset for this view operation if specified.
/// Returns nullptr if no dynamic offset was specified.
Value *getDynamicOffset();
return getDynamicOffset() == nullptr ? 1 : 2;
}
- /// Returns the dynamic sizes for this shape cast operation.
+ /// Returns the dynamic sizes for this view operation.
operand_range getDynamicSizes() {
return {operand_begin() + getDynamicSizesOperandStart(), operand_end()};
}
let hasCanonicalizer = 1;
}
+def SubViewOp : Std_Op<"subview", [SameVariadicOperandSize]> {
+ let summary = "memref subview operation";
+ let description = [{
+ The "subview" operation converts a memref type to another memref type
+ which represents a reduced-size view of the original memref as specified by
+ the operation's offsets, sizes and strides arguments.
+
+ The SubView operation supports the following arguments:
+ *) Memref: the "base" memref on which to create a "view" memref.
+ *) Offsets: memref-rank number of dynamic offsets into the "base" memref at
+ which to create the "view" memref.
+ *) Sizes: memref-rank dynamic size operands which specify the dynamic sizes
+ of the result "view" memref type.
+ *) Strides: memref-rank number of dynamic strides which are applied
+ multiplicatively to the base memref strides in each dimension.
+
+ Example 1:
+
+ %0 = alloc() : memref<64x4xf32, (d0, d1) -> (d0 * 4 + d1)>
+
+ // Create a sub-view of "base" memref '%0' with offset arguments '%c0',
+ // dynamic sizes for each dimension, and stride arguments '%c1'.
+ %1 = subview %0[%c0, %c0][%size0, %size1][%c1, %c1]
+ : memref<64x4xf32, (d0, d1) -> (d0 * 4 + d1) > to
+ memref<?x?xf32, (d0, d1)[s0, s1] -> (d0 * s1 + d1 + s0)>
+
+ Example 2:
+
+ %0 = alloc() : memref<8x16x4xf32, (d0, d1, d1) -> (d0 * 64 + d1 * 4 + d2)>
+
+ // Create a sub-view of "base" memref '%0' with dynamic offsets, sizes,
+ // and strides.
+ // Note that dynamic offsets are represented by the linearized dynamic
+ // offset symbol 's0' in the subview memref layout map, and that the
+ // dynamic strides operands, after being applied to the base memref
+ // strides in each dimension, are represented in the view memref layout
+ // map as symbols 's1', 's2' and 's3'.
+ %1 = subview %0[%i, %j, %k][%size0, %size1, %size2][%x, %y, %z]
+ : memref<8x16x4xf32, (d0, d1, d2) -> (d0 * 64 + d1 * 4 + d2)>
+ memref<?x?x?xf32,
+ (d0, d1, d2)[s0, s1, s2, s3] -> (d0 * s1 + d1 * s2 + d2 * s3 + s0)>
+ }
+ }];
+
+ let arguments = (ins AnyMemRef:$source, Variadic<Index>:$offsets,
+ Variadic<Index>:$sizes, Variadic<Index>:$strides);
+ let results = (outs AnyMemRef);
+
+ let extraClassDeclaration = [{
+ /// The result of a subview is always a memref.
+ MemRefType getType() { return getResult()->getType().cast<MemRefType>(); }
+
+ /// Returns the dynamic offsets for this subview operation.
+ operand_range getDynamicOffsets() {
+ return {operand_begin() + 1, operand_begin() + 1 + getType().getRank()};
+ }
+
+ /// Returns the operand starting position of the size operands.
+ unsigned getSizeOperandsStart() { return 1 + getType().getRank(); }
+
+ /// Returns the dynamic sizes for this subview operation if specified.
+ operand_range getDynamicSizes() {
+ return {operand_begin() + getSizeOperandsStart(),
+ operand_begin() + getSizeOperandsStart() + getType().getRank()};
+ }
+
+ /// Returns the operand starting position of the size operands.
+ unsigned getStrideOperandsStart() { return 1 + 2 * getType().getRank(); }
+
+ /// Returns the dynamic strides for this subview operation if specified.
+ operand_range getDynamicStrides() {
+ return {operand_begin() + getStrideOperandsStart(),
+ operand_begin() + getStrideOperandsStart() + getType().getRank()};
+ }
+ }];
+
+ // TODO(andydavis) Add canonicalizer.
+}
+
def XOrOp : IntArithmeticOp<"xor", [Commutative]> {
let summary = "integer binary xor";
let hasFolder = 1;
result.addAttributes(attrs);
}
-static void print(OpAsmPrinter &p, SubViewOp op) {
+static void print(OpAsmPrinter &p, mlir::linalg::SubViewOp op) {
p << op.getOperationName() << " " << *op.getOperand(0) << "[";
auto ranges = op.getRanges();
- interleaveComma(ranges, p, [&p](const SubViewOp::Range &i) {
+ interleaveComma(ranges, p, [&p](const mlir::linalg::SubViewOp::Range &i) {
p << *i.min << ", " << *i.max << ", " << *i.step;
});
p << "]";
return success();
}
-llvm::raw_ostream &mlir::linalg::operator<<(llvm::raw_ostream &os,
- SubViewOp::Range &range) {
+llvm::raw_ostream &
+mlir::linalg::operator<<(llvm::raw_ostream &os,
+ mlir::linalg::SubViewOp::Range &range) {
return os << "range " << *range.min << ":" << *range.max << ":"
<< *range.step;
}
// a subset of the original loop ranges of `op`.
// This is achieved by applying the `loopToOperandRangesMaps` permutation maps
// to the `loopRanges` in order to obtain view ranges.
-static LinalgOp cloneWithLoopRanges(OpBuilder &b, Location loc, LinalgOp op,
- ArrayRef<SubViewOp::Range> loopRanges) {
+static LinalgOp
+cloneWithLoopRanges(OpBuilder &b, Location loc, LinalgOp op,
+ ArrayRef<mlir::linalg::SubViewOp::Range> loopRanges) {
auto maps = loopToOperandRangesMaps(op);
SmallVector<Value *, 8> clonedViews;
clonedViews.reserve(op.getNumInputsAndOutputs());
auto map = maps[idx];
LLVM_DEBUG(dbgs() << "map: " << map << "\n");
Value *view = en.value();
- SmallVector<SubViewOp::Range, 8> viewRanges(map.getNumResults());
+ SmallVector<mlir::linalg::SubViewOp::Range, 8> viewRanges(
+ map.getNumResults());
for (auto en2 : llvm::enumerate(map.getResults())) {
unsigned d = en2.index();
// loopToOperandRangesMaps are permutations-only.
subViewOperands.push_back(r.max);
subViewOperands.push_back(r.step);
}
- clonedViews.push_back(b.create<SubViewOp>(loc, view, subViewOperands));
+ clonedViews.push_back(
+ b.create<mlir::linalg::SubViewOp>(loc, view, subViewOperands));
}
auto operands = getAssumedNonViewOperands(op);
clonedViews.append(operands.begin(), operands.end());
static LinalgOp fuse(Value *producedView, LinalgOp producer, LinalgOp consumer,
unsigned consumerIdx, unsigned producerIdx,
OperationFolder *folder) {
- auto subView = dyn_cast_or_null<SubViewOp>(
+ auto subView = dyn_cast_or_null<mlir::linalg::SubViewOp>(
consumer.getInput(consumerIdx)->getDefiningOp());
auto slice = dyn_cast_or_null<SliceOp>(
consumer.getInput(consumerIdx)->getDefiningOp());
unsigned nPar = producer.getNumParallelLoops();
unsigned nRed = producer.getNumReductionLoops();
unsigned nWin = producer.getNumWindowLoops();
- SmallVector<SubViewOp::Range, 8> loopRanges(nPar + nRed + nWin);
+ SmallVector<mlir::linalg::SubViewOp::Range, 8> loopRanges(nPar + nRed + nWin);
// Iterate over dimensions identified by the producer map for `producerIdx`.
// This defines a subset of the loop ranges that we need to complete later.
<< "existing LoopRange: " << loopRanges[i] << "\n");
else {
auto viewDim = getViewDefiningLoopRange(producer, i);
- loopRanges[i] = SubViewOp::Range{constant_index(folder, 0),
- dim(viewDim.view, viewDim.dimension),
- constant_index(folder, 1)};
+ loopRanges[i] = mlir::linalg::SubViewOp::Range{
+ constant_index(folder, 0), dim(viewDim.view, viewDim.dimension),
+ constant_index(folder, 1)};
LLVM_DEBUG(llvm::dbgs() << "new LoopRange: " << loopRanges[i] << "\n");
}
}
// Must be a subview or a slice to guarantee there are loops we can fuse
// into.
- auto subView = dyn_cast_or_null<SubViewOp>(consumedView->getDefiningOp());
+ auto subView = dyn_cast_or_null<mlir::linalg::SubViewOp>(
+ consumedView->getDefiningOp());
auto slice = dyn_cast_or_null<SliceOp>(consumedView->getDefiningOp());
if (!subView && !slice) {
LLVM_DEBUG(dbgs() << "\nNot fusable (not a subview or slice)");
/// A non-conversion rewrite pattern kicks in to convert SubViewOp into RangeOps
/// and SliceOps.
-class SubViewOpConversion : public OpRewritePattern<SubViewOp> {
+class SubViewOpConversion : public OpRewritePattern<mlir::linalg::SubViewOp> {
public:
- using OpRewritePattern<SubViewOp>::OpRewritePattern;
+ using OpRewritePattern<mlir::linalg::SubViewOp>::OpRewritePattern;
- PatternMatchResult matchAndRewrite(SubViewOp op,
+ PatternMatchResult matchAndRewrite(mlir::linalg::SubViewOp op,
PatternRewriter &rewriter) const override {
auto *view = op.getView();
SmallVector<Value *, 8> ranges;
// boundary tiles. For now this is done with an unconditional `fill` op followed
// by a partial `copy` op.
static PromotionInfo promoteFullTileBuffer(OpBuilder &b, Location loc,
- SubViewOp subView,
+ mlir::linalg::SubViewOp subView,
bool dynamicBuffers,
OperationFolder *folder) {
auto zero = constant_index(folder, 0);
res.reserve(subViews.size());
DenseMap<Value *, PromotionInfo> promotionInfoMap;
for (auto *v : subViews) {
- SubViewOp subView = cast<SubViewOp>(v->getDefiningOp());
+ mlir::linalg::SubViewOp subView =
+ cast<mlir::linalg::SubViewOp>(v->getDefiningOp());
auto viewType = subView.getViewType();
// TODO(ntv): support more cases than just float.
if (!viewType.getElementType().isa<FloatType>())
}
for (auto *v : subViews) {
- SubViewOp subView = cast<SubViewOp>(v->getDefiningOp());
+ mlir::linalg::SubViewOp subView =
+ cast<mlir::linalg::SubViewOp>(v->getDefiningOp());
auto info = promotionInfoMap.find(v);
if (info == promotionInfoMap.end())
continue;
auto info = promotionInfoMap.find(v);
if (info == promotionInfoMap.end())
continue;
- copy(cast<SubViewOp>(v->getDefiningOp()), info->second.partialLocalView);
+ copy(cast<mlir::linalg::SubViewOp>(v->getDefiningOp()),
+ info->second.partialLocalView);
}
return res;
}
// nothing.
SetVector<Value *> subViews;
for (auto it : op.getInputsAndOutputs())
- if (auto sv = dyn_cast_or_null<SubViewOp>(it->getDefiningOp()))
+ if (auto sv =
+ dyn_cast_or_null<mlir::linalg::SubViewOp>(it->getDefiningOp()))
subViews.insert(sv);
if (!subViews.empty()) {
promoteSubViewOperands(op, subViews, dynamicBuffers, &folder);
// avoiding affine map manipulations.
// The returned ranges correspond to the loop ranges, in the proper order, that
// are tiled and for which new loops will be created.
-static SmallVector<SubViewOp::Range, 4>
+static SmallVector<mlir::linalg::SubViewOp::Range, 4>
makeTiledLoopRanges(OpBuilder &b, Location loc, AffineMap map,
ArrayRef<Value *> allViewSizes,
ArrayRef<Value *> allTileSizes, OperationFolder *folder) {
}
// Create a new range with the applied tile sizes.
- SmallVector<SubViewOp::Range, 4> res;
+ SmallVector<mlir::linalg::SubViewOp::Range, 4> res;
for (unsigned idx = 0, e = tileSizes.size(); idx < e; ++idx) {
- res.push_back(SubViewOp::Range{constant_index(folder, 0), viewSizes[idx],
- tileSizes[idx]});
+ res.push_back(mlir::linalg::SubViewOp::Range{
+ constant_index(folder, 0), viewSizes[idx], tileSizes[idx]});
}
return res;
}
}
// Construct a new subview for the tile.
- SmallVector<SubViewOp::Range, 4> subViewRangeOperands;
+ SmallVector<mlir::linalg::SubViewOp::Range, 4> subViewRangeOperands;
subViewRangeOperands.reserve(rank * 3);
for (unsigned r = 0; r < rank; ++r) {
if (!isTiled(map.getSubMap({r}), tileSizes)) {
- subViewRangeOperands.push_back(
- SubViewOp::Range{constant_index(folder, 0), dim(view, r),
- constant_index(folder, 1)});
+ subViewRangeOperands.push_back(mlir::linalg::SubViewOp::Range{
+ constant_index(folder, 0), dim(view, r),
+ constant_index(folder, 1)});
continue;
}
// Tiling creates a new slice at the proper index, the slice step is 1
// (i.e. the slice view does not subsample, stepping occurs in the loop).
subViewRangeOperands.push_back(
- SubViewOp::Range{min, max, constant_index(folder, 1)});
+ mlir::linalg::SubViewOp::Range{min, max, constant_index(folder, 1)});
}
SmallVector<Value *, 12> subViewOperands;
subViewOperands.reserve(subViewRangeOperands.size() * 3);
subViewOperands.push_back(r.max);
subViewOperands.push_back(r.step);
}
- res.push_back(b.create<SubViewOp>(loc, view, subViewOperands));
+ res.push_back(
+ b.create<mlir::linalg::SubViewOp>(loc, view, subViewOperands));
}
// Traverse the mins/maxes and erase those that don't have uses left.
enter(body, /*prev=*/1);
}
-mlir::edsc::LoopRangeBuilder::LoopRangeBuilder(ValueHandle *iv,
- SubViewOp::Range range) {
+mlir::edsc::LoopRangeBuilder::LoopRangeBuilder(
+ ValueHandle *iv, mlir::linalg::SubViewOp::Range range) {
auto forOp =
OperationHandle::createOp<ForOp>(range.min, range.max, range.step);
*iv = ValueHandle(forOp.getInductionVar());
}
mlir::edsc::LoopNestRangeBuilder::LoopNestRangeBuilder(
- ArrayRef<ValueHandle *> ivs, ArrayRef<SubViewOp::Range> ranges) {
+ ArrayRef<ValueHandle *> ivs,
+ ArrayRef<mlir::linalg::SubViewOp::Range> ranges) {
loops.reserve(ranges.size());
for (unsigned i = 0, e = ranges.size(); i < e; ++i) {
loops.emplace_back(ivs[i], ranges[i]);
return nullptr;
}
+static LogicalResult verifyDynamicStrides(MemRefType memrefType,
+ ArrayRef<int64_t> strides) {
+ ArrayRef<int64_t> shape = memrefType.getShape();
+ unsigned rank = memrefType.getRank();
+ assert(rank == strides.size());
+ bool dynamicStrides = false;
+ for (int i = rank - 2; i >= 0; --i) {
+ // If size at dim 'i + 1' is dynamic, set the 'dynamicStrides' flag.
+ if (ShapedType::isDynamic(shape[i + 1]))
+ dynamicStrides = true;
+ // If stride at dim 'i' is not dynamic, return error.
+ if (dynamicStrides && strides[i] != MemRefType::getDynamicStrideOrOffset())
+ return failure();
+ }
+ return success();
+}
+
static LogicalResult verify(ViewOp op) {
auto baseType = op.getOperand(0)->getType().cast<MemRefType>();
auto viewType = op.getResult()->getType().cast<MemRefType>();
"type ")
<< baseType << " and view memref type " << viewType;
- // Verify that the result memref type has a strided layout map. is strided
+ // Verify that the result memref type has a strided layout map.
int64_t offset;
llvm::SmallVector<int64_t, 4> strides;
if (failed(getStridesAndOffset(viewType, strides, offset)))
// Verify dynamic strides symbols were added to correct dimensions based
// on dynamic sizes.
- ArrayRef<int64_t> viewShape = viewType.getShape();
- unsigned viewRank = viewType.getRank();
- assert(viewRank == strides.size());
- bool dynamicStrides = false;
- for (int i = viewRank - 2; i >= 0; --i) {
- // If size at dim 'i + 1' is dynamic, set the 'dynamicStrides' flag.
- if (ShapedType::isDynamic(viewShape[i + 1]))
- dynamicStrides = true;
- // If stride at dim 'i' is not dynamic, return error.
- if (dynamicStrides && strides[i] != MemRefType::getDynamicStrideOrOffset())
- return op.emitError("incorrect dynamic strides in view memref type ")
- << viewType;
- }
-
+ if (failed(verifyDynamicStrides(viewType, strides)))
+ return op.emitError("incorrect dynamic strides in view memref type ")
+ << viewType;
return success();
}
}
//===----------------------------------------------------------------------===//
+// SubViewOp
+//===----------------------------------------------------------------------===//
+
+static ParseResult parseSubViewOp(OpAsmParser &parser, OperationState &result) {
+ OpAsmParser::OperandType srcInfo;
+ SmallVector<OpAsmParser::OperandType, 4> offsetsInfo;
+ SmallVector<OpAsmParser::OperandType, 4> sizesInfo;
+ SmallVector<OpAsmParser::OperandType, 4> stridesInfo;
+ auto indexType = parser.getBuilder().getIndexType();
+ Type srcType, dstType;
+ return failure(
+ parser.parseOperand(srcInfo) ||
+ parser.parseOperandList(offsetsInfo, OpAsmParser::Delimiter::Square) ||
+ parser.parseOperandList(sizesInfo, OpAsmParser::Delimiter::Square) ||
+ parser.parseOperandList(stridesInfo, OpAsmParser::Delimiter::Square) ||
+ parser.parseOptionalAttrDict(result.attributes) ||
+ parser.parseColonType(srcType) ||
+ parser.resolveOperand(srcInfo, srcType, result.operands) ||
+ parser.resolveOperands(offsetsInfo, indexType, result.operands) ||
+ parser.resolveOperands(sizesInfo, indexType, result.operands) ||
+ parser.resolveOperands(stridesInfo, indexType, result.operands) ||
+ parser.parseKeywordType("to", dstType) ||
+ parser.addTypeToList(dstType, result.types));
+}
+
+static void print(OpAsmPrinter &p, SubViewOp op) {
+ p << op.getOperationName() << ' ' << *op.getOperand(0) << '[';
+ p.printOperands(op.getDynamicOffsets());
+ p << "][";
+ p.printOperands(op.getDynamicSizes());
+ p << "][";
+ p.printOperands(op.getDynamicStrides());
+ p << ']';
+ p.printOptionalAttrDict(op.getAttrs());
+ p << " : " << op.getOperand(0)->getType() << " to " << op.getType();
+}
+
+static LogicalResult verify(SubViewOp op) {
+ auto baseType = op.getOperand(0)->getType().cast<MemRefType>();
+ auto subViewType = op.getResult()->getType().cast<MemRefType>();
+
+ // The base memref and the view memref should be in the same memory space.
+ if (baseType.getMemorySpace() != subViewType.getMemorySpace())
+ return op.emitError("different memory spaces specified for base memref "
+ "type ")
+ << baseType << " and subview memref type " << subViewType;
+
+ // Verify that the base memref type has a strided layout map.
+ int64_t baseOffset;
+ llvm::SmallVector<int64_t, 4> baseStrides;
+ if (failed(getStridesAndOffset(baseType, baseStrides, baseOffset)))
+ return op.emitError("base type ") << subViewType << " is not strided";
+
+ // Verify that the result memref type has a strided layout map.
+ int64_t subViewOffset;
+ llvm::SmallVector<int64_t, 4> subViewStrides;
+ if (failed(getStridesAndOffset(subViewType, subViewStrides, subViewOffset)))
+ return op.emitError("result type ") << subViewType << " is not strided";
+
+ unsigned memrefOperandCount = 1;
+ unsigned numDynamicOffsets = llvm::size(op.getDynamicOffsets());
+ unsigned numDynamicSizes = llvm::size(op.getDynamicSizes());
+ unsigned numDynamicStrides = llvm::size(op.getDynamicStrides());
+
+ // Verify that we have the correct number of operands for the result type.
+ if (op.getNumOperands() != memrefOperandCount + numDynamicOffsets +
+ numDynamicSizes + numDynamicStrides)
+ return op.emitError("incorrect number of operands for type ")
+ << subViewType;
+
+ // Verify that the subview layout map has a dynamic offset.
+ if (subViewOffset != MemRefType::getDynamicStrideOrOffset())
+ return op.emitError("subview memref layout map must specify a dynamic "
+ "offset for type ")
+ << subViewType;
+
+ // Verify dynamic strides symbols were added to correct dimensions based
+ // on dynamic sizes.
+ if (failed(verifyDynamicStrides(subViewType, subViewStrides)))
+ return op.emitError("incorrect dynamic strides in view memref type ")
+ << subViewType;
+ return success();
+}
+
+//===----------------------------------------------------------------------===//
// ZeroExtendIOp
//===----------------------------------------------------------------------===//
// CHECK-DAG: #[[VIEW_MAP2:map[0-9]+]] = (d0, d1)[s0, s1] -> (d0 * s1 + d1 + s0)
// CHECK-DAG: #[[VIEW_MAP3:map[0-9]+]] = (d0, d1)[s0] -> (d0 * s0 + d1)
+// CHECK-DAG: #[[BASE_MAP0:map[0-9]+]] = (d0, d1, d2) -> (d0 * 64 + d1 * 4 + d2)
+// CHECK-DAG: #[[SUBVIEW_MAP0:map[0-9]+]] = (d0, d1, d2)[s0, s1, s2, s3] -> (d0 * s1 + d1 * s2 + d2 * s3 + s0)
+
+// CHECK-DAG: #[[BASE_MAP1:map[0-9]+]] = (d0)[s0] -> (d0 + s0)
+// CHECK-DAG: #[[SUBVIEW_MAP1:map[0-9]+]] = (d0)[s0, s1] -> (d0 * s1 + s0)
+
+// CHECK-DAG: #[[BASE_MAP2:map[0-9]+]] = (d0, d1) -> (d0 * 22 + d1)
+// CHECK-DAG: #[[SUBVIEW_MAP2:map[0-9]+]] = (d0, d1)[s0, s1, s2] -> (d0 * s1 + d1 * s2 + s0)
+
// CHECK-LABEL: func @func_with_ops(%arg0: f32) {
func @func_with_ops(f32) {
^bb0(%a : f32):
return
}
+// CHECK-LABEL: func @memref_subview(%arg0
+func @memref_subview(%arg0 : index, %arg1 : index, %arg2 : index) {
+ %c0 = constant 0 : index
+ %c1 = constant 1 : index
+
+ //%2 = alloc() : memref<64xf32, (d0) -> (d0)>
+
+ %0 = alloc() : memref<8x16x4xf32, (d0, d1, d2) -> (d0 * 64 + d1 * 4 + d2)>
+ // CHECK: std.subview %0[%c0, %c0, %c0][%arg0, %arg1, %arg2][%c1, %c1, %c1] : memref<8x16x4xf32, #[[BASE_MAP0]]> to memref<?x?x?xf32, #[[SUBVIEW_MAP0]]>
+ %1 = subview %0[%c0, %c0, %c0][%arg0, %arg1, %arg2][%c1, %c1, %c1]
+ : memref<8x16x4xf32, (d0, d1, d2) -> (d0 * 64 + d1 * 4 + d2)> to
+ memref<?x?x?xf32,
+ (d0, d1, d2)[s0, s1, s2, s3] -> (d0 * s1 + d1 * s2 + d2 * s3 + s0)>
+
+ %2 = alloc()[%arg2] : memref<64xf32, (d0)[s0] -> (d0 + s0)>
+ // CHECK: std.subview %2[%c1][%arg0][%c1] : memref<64xf32, #[[BASE_MAP1]]> to memref<?xf32, #[[SUBVIEW_MAP1]]>
+ %3 = subview %2[%c1][%arg0][%c1]
+ : memref<64xf32, (d0)[s0] -> (d0 + s0)> to
+ memref<?xf32, (d0)[s0, s1] -> (d0 * s1 + s0)>
+
+ %4 = alloc() : memref<64x22xf32, (d0, d1) -> (d0 * 22 + d1)>
+ // CHECK: std.subview %4[%c0, %c1][%arg0, %arg1][%c1, %c0] : memref<64x22xf32, #[[BASE_MAP2]]> to memref<?x?xf32, #[[SUBVIEW_MAP2]]>
+ %5 = subview %4[%c0, %c1][%arg0, %arg1][%c1, %c0]
+ : memref<64x22xf32, (d0, d1) -> (d0 * 22 + d1)> to
+ memref<?x?xf32, (d0, d1)[s0, s1, s2] -> (d0 * s1 + d1 * s2 + s0)>
+ return
+}
+
// CHECK-LABEL: func @test_dimop(%arg0
func @test_dimop(%arg0: tensor<4x4x?xf32>) {
// CHECK: %0 = dim %arg0, 2 : tensor<4x4x?xf32>
return
}
+// -----
+
+func @invalid_subview(%arg0 : index, %arg1 : index, %arg2 : index) {
+ %0 = alloc() : memref<8x16x4xf32, (d0, d1, d2) -> (d0 * 64 + d1 * 4 + d2), 2>
+ // expected-error@+1 {{different memory spaces}}
+ %1 = subview %0[][%arg2][]
+ : memref<8x16x4xf32, (d0, d1, d2) -> (d0 * 64 + d1 * 4 + d2), 2> to
+ memref<8x?x4xf32, (d0, d1, d2)[s0] -> (d0 * s0 + d1 * 4 + d2)>
+ return
+}
+
+// -----
+
+func @invalid_subview(%arg0 : index, %arg1 : index, %arg2 : index) {
+ %0 = alloc() : memref<8x16x4xf32, (d0, d1, d2) -> (d0 * 64 + d1 * 4 + d2)>
+ // expected-error@+1 {{is not strided}}
+ %1 = subview %0[][%arg2][]
+ : memref<8x16x4xf32, (d0, d1, d2) -> (d0 * 64 + d1 * 4 + d2)> to
+ memref<8x?x4xf32, (d0, d1, d2)[s0] -> (d0 + s0, d1, d2)>
+ return
+}
+
+// -----
+
+func @invalid_subview(%arg0 : index, %arg1 : index, %arg2 : index) {
+ %0 = alloc() : memref<8x16x4xf32, (d0, d1, d2) -> (d0 + d1, d1 + d2, d2)>
+ // expected-error@+1 {{is not strided}}
+ %1 = subview %0[][%arg2][]
+ : memref<8x16x4xf32, (d0, d1, d2) -> (d0 + d1, d1 + d2, d2)> to
+ memref<8x?x4xf32, (d0, d1, d2)[s0] -> (d0 * s0 + d1 * 4 + d2)>
+ return
+}
+
+// -----
+
+func @invalid_subview(%arg0 : index, %arg1 : index, %arg2 : index) {
+ %0 = alloc() : memref<8x16x4xf32, (d0, d1, d2) -> (d0 * 64 + d1 * 4 + d2)>
+ // expected-error@+1 {{incorrect number of operands for type}}
+ %1 = subview %0[%arg0, %arg1][%arg2][]
+ : memref<8x16x4xf32, (d0, d1, d2) -> (d0 * 64 + d1 * 4 + d2)> to
+ memref<8x?x4xf32, (d0, d1, d2)[s0] -> (d0 * s0 + d1 * 4 + d2)>
+ return
+}
+
+// -----
+
+func @invalid_subview(%arg0 : index, %arg1 : index, %arg2 : index) {
+ %0 = alloc() : memref<8x16x4xf32, (d0, d1, d2) -> (d0 * 64 + d1 * 4 + d2)>
+ // expected-error@+1 {{incorrect dynamic strides in view memref type}}
+ %1 = subview %0[%arg0, %arg1, %arg2][%arg0, %arg1, %arg2][%arg0, %arg1, %arg2]
+ : memref<8x16x4xf32, (d0, d1, d2) -> (d0 * 64 + d1 * 4 + d2)> to
+ memref<?x?x4xf32, (d0, d1, d2)[s0] -> (d0 * 64 + d1 * 4 + d2 + s0)>
+ return
+}
+
+// -----
+
+func @invalid_subview(%arg0 : index, %arg1 : index, %arg2 : index) {
+ %0 = alloc() : memref<8x16x4xf32, (d0, d1, d2) -> (d0 * 64 + d1 * 4 + d2)>
+ %c0 = constant 0 : index
+ %c1 = constant 1 : index
+ // expected-error@+1 {{subview memref layout map must specify a dynamic offset}}
+ %1 = subview %0[%c0, %c0, %c0][%arg0, %arg1, %arg2][%c1, %c1, %c1]
+ : memref<8x16x4xf32, (d0, d1, d2) -> (d0 * 64 + d1 * 4 + d2)> to
+ memref<?x?x?xf32, (d0, d1, d2)[s0, s1, s2] -> (d0 * s0 + d1 * s1 + d2 * s2)>
+ return
+}