break;
}
default:
- if (match(Cmp, m_ICmp(Pred, m_V, m_Value(A)))) {
+ const APInt *Offset = nullptr;
+ if (match(Cmp, m_ICmp(Pred, m_CombineOr(m_V, m_Add(m_V, m_APInt(Offset))),
+ m_Value(A)))) {
KnownBits RHSKnown = computeKnownBits(A, Depth + 1, QueryNoAC);
ConstantRange RHSRange =
ConstantRange::fromKnownBits(RHSKnown, Cmp->isSigned());
ConstantRange LHSRange =
ConstantRange::makeAllowedICmpRegion(Pred, RHSRange);
+ if (Offset)
+ LHSRange = LHSRange.sub(*Offset);
Known = Known.unionWith(LHSRange.toKnownBits());
}
break;
define i1 @nonnull1(ptr %a) {
; CHECK-LABEL: @nonnull1(
-; CHECK-NEXT: [[LOAD:%.*]] = load ptr, ptr [[A:%.*]], align 8, !nonnull !6, !noundef !6
+; CHECK-NEXT: [[LOAD:%.*]] = load ptr, ptr [[A:%.*]], align 8, !nonnull [[META6:![0-9]+]], !noundef [[META6]]
; CHECK-NEXT: tail call void @escape(ptr nonnull [[LOAD]])
; CHECK-NEXT: ret i1 false
;
; CHECK-NEXT: [[ADD:%.*]] = add i32 [[X:%.*]], -16
; CHECK-NEXT: [[CMP:%.*]] = icmp ult i32 [[ADD]], 15
; CHECK-NEXT: call void @llvm.assume(i1 [[CMP]])
-; CHECK-NEXT: [[RES:%.*]] = and i32 [[X]], -16
-; CHECK-NEXT: ret i32 [[RES]]
+; CHECK-NEXT: ret i32 16
;
%add = add i32 %x, -16
%cmp = icmp ult i32 %add, 15
; CHECK-NEXT: [[ADD:%.*]] = add i32 [[X:%.*]], -16
; CHECK-NEXT: [[CMP:%.*]] = icmp ult i32 [[ADD]], 17
; CHECK-NEXT: call void @llvm.assume(i1 [[CMP]])
-; CHECK-NEXT: [[RES:%.*]] = and i32 [[X]], -16
+; CHECK-NEXT: [[RES:%.*]] = and i32 [[X]], 48
; CHECK-NEXT: ret i32 [[RES]]
;
%add = add i32 %x, -16
; CHECK-NEXT: [[ADD:%.*]] = add i32 [[X:%.*]], -16
; CHECK-NEXT: [[CMP:%.*]] = icmp ult i32 [[ADD]], 17
; CHECK-NEXT: call void @llvm.assume(i1 [[CMP]])
-; CHECK-NEXT: [[RES:%.*]] = and i32 [[X]], -32
+; CHECK-NEXT: [[RES:%.*]] = and i32 [[X]], 32
; CHECK-NEXT: ret i32 [[RES]]
;
%add = add i32 %x, -16
; CHECK-NEXT: [[ADD:%.*]] = add i32 [[X:%.*]], -16
; CHECK-NEXT: [[CMP:%.*]] = icmp ult i32 [[ADD]], 17
; CHECK-NEXT: call void @llvm.assume(i1 [[CMP]])
-; CHECK-NEXT: [[RES:%.*]] = and i32 [[X]], -64
-; CHECK-NEXT: ret i32 [[RES]]
+; CHECK-NEXT: ret i32 0
;
%add = add i32 %x, -16
%cmp = icmp ult i32 %add, 17
; CHECK-NEXT: [[ADD:%.*]] = add i32 [[X:%.*]], -15
; CHECK-NEXT: [[CMP:%.*]] = icmp ult i32 [[ADD]], 16
; CHECK-NEXT: call void @llvm.assume(i1 [[CMP]])
-; CHECK-NEXT: [[RES:%.*]] = and i32 [[X]], -16
+; CHECK-NEXT: [[RES:%.*]] = and i32 [[X]], 16
; CHECK-NEXT: ret i32 [[RES]]
;
%add = add i32 %x, -15
; CHECK-NEXT: [[ADD:%.*]] = add i32 [[X:%.*]], -15
; CHECK-NEXT: [[CMP:%.*]] = icmp ult i32 [[ADD]], 16
; CHECK-NEXT: call void @llvm.assume(i1 [[CMP]])
-; CHECK-NEXT: [[RES:%.*]] = and i32 [[X]], -32
-; CHECK-NEXT: ret i32 [[RES]]
+; CHECK-NEXT: ret i32 0
;
%add = add i32 %x, -15
%cmp = icmp ult i32 %add, 16