From: Roman Lebedev Date: Sat, 10 Aug 2019 19:29:03 +0000 (+0000) Subject: [NFC][InstCombine] Tests for shift amount reassociation in bittest with truncated... X-Git-Tag: llvmorg-11-init~12189 X-Git-Url: http://review.tizen.org/git/?a=commitdiff_plain;h=404e978f2721eb77894881d60a6175f4018182a2;p=platform%2Fupstream%2Fllvm.git [NFC][InstCombine] Tests for shift amount reassociation in bittest with truncated shl (PR42399) trunc-of-shl: https://rise4fun.com/Alive/zGx https://rise4fun.com/Alive/sl0L I.e. no extra legality check needed. https://bugs.llvm.org/show_bug.cgi?id=42399 llvm-svn: 368520 --- diff --git a/llvm/test/Transforms/InstCombine/shift-amount-reassociation-in-bittest-with-truncation-shl.ll b/llvm/test/Transforms/InstCombine/shift-amount-reassociation-in-bittest-with-truncation-shl.ll new file mode 100644 index 0000000..7213616 --- /dev/null +++ b/llvm/test/Transforms/InstCombine/shift-amount-reassociation-in-bittest-with-truncation-shl.ll @@ -0,0 +1,475 @@ +; NOTE: Assertions have been autogenerated by utils/update_test_checks.py +; RUN: opt %s -instcombine -S | FileCheck %s + +; Given pattern: +; icmp eq/ne (and ((x shift Q), (y oppositeshift K))), 0 +; we should move shifts to the same hand of 'and', i.e. e.g. rewrite as +; icmp eq/ne (and (((x shift Q) shift K), y)), 0 +; We are only interested in opposite logical shifts here. + +;------------------------------------------------------------------------------- +; Basic scalar tests +;------------------------------------------------------------------------------- + +define i1 @t0_const_after_fold_lshr_shl_ne(i32 %x, i64 %y, i32 %len) { +; CHECK-LABEL: @t0_const_after_fold_lshr_shl_ne( +; CHECK-NEXT: [[T0:%.*]] = sub i32 32, [[LEN:%.*]] +; CHECK-NEXT: [[T1:%.*]] = lshr i32 [[X:%.*]], [[T0]] +; CHECK-NEXT: [[T2:%.*]] = add i32 [[LEN]], -1 +; CHECK-NEXT: [[T2_WIDE:%.*]] = zext i32 [[T2]] to i64 +; CHECK-NEXT: [[T3:%.*]] = shl i64 [[Y:%.*]], [[T2_WIDE]] +; CHECK-NEXT: [[T3_TRUNC:%.*]] = trunc i64 [[T3]] to i32 +; CHECK-NEXT: [[T4:%.*]] = and i32 [[T1]], [[T3_TRUNC]] +; CHECK-NEXT: [[T5:%.*]] = icmp ne i32 [[T4]], 0 +; CHECK-NEXT: ret i1 [[T5]] +; + %t0 = sub i32 32, %len + %t1 = lshr i32 %x, %t0 + %t2 = add i32 %len, -1 + %t2_wide = zext i32 %t2 to i64 + %t3 = shl i64 %y, %t2_wide + %t3_trunc = trunc i64 %t3 to i32 + %t4 = and i32 %t1, %t3_trunc + %t5 = icmp ne i32 %t4, 0 + ret i1 %t5 +} + +;------------------------------------------------------------------------------- +; Very basic vector tests +;------------------------------------------------------------------------------- + +define <2 x i1> @t1_vec_splat(<2 x i32> %x, <2 x i64> %y, <2 x i32> %len) { +; CHECK-LABEL: @t1_vec_splat( +; CHECK-NEXT: [[T0:%.*]] = sub <2 x i32> , [[LEN:%.*]] +; CHECK-NEXT: [[T1:%.*]] = lshr <2 x i32> [[X:%.*]], [[T0]] +; CHECK-NEXT: [[T2:%.*]] = add <2 x i32> [[LEN]], +; CHECK-NEXT: [[T2_WIDE:%.*]] = zext <2 x i32> [[T2]] to <2 x i64> +; CHECK-NEXT: [[T3:%.*]] = shl <2 x i64> [[Y:%.*]], [[T2_WIDE]] +; CHECK-NEXT: [[T3_TRUNC:%.*]] = trunc <2 x i64> [[T3]] to <2 x i32> +; CHECK-NEXT: [[T4:%.*]] = and <2 x i32> [[T1]], [[T3_TRUNC]] +; CHECK-NEXT: [[T5:%.*]] = icmp ne <2 x i32> [[T4]], zeroinitializer +; CHECK-NEXT: ret <2 x i1> [[T5]] +; + %t0 = sub <2 x i32> , %len + %t1 = lshr <2 x i32> %x, %t0 + %t2 = add <2 x i32> %len, + %t2_wide = zext <2 x i32> %t2 to <2 x i64> + %t3 = shl <2 x i64> %y, %t2_wide + %t3_trunc = trunc <2 x i64> %t3 to <2 x i32> + %t4 = and <2 x i32> %t1, %t3_trunc + %t5 = icmp ne <2 x i32> %t4, + ret <2 x i1> %t5 +} + +define <2 x i1> @t2_vec_nonsplat(<2 x i32> %x, <2 x i64> %y, <2 x i32> %len) { +; CHECK-LABEL: @t2_vec_nonsplat( +; CHECK-NEXT: [[T0:%.*]] = sub <2 x i32> , [[LEN:%.*]] +; CHECK-NEXT: [[T1:%.*]] = lshr <2 x i32> [[X:%.*]], [[T0]] +; CHECK-NEXT: [[T2:%.*]] = add <2 x i32> [[LEN]], +; CHECK-NEXT: [[T2_WIDE:%.*]] = zext <2 x i32> [[T2]] to <2 x i64> +; CHECK-NEXT: [[T3:%.*]] = shl <2 x i64> [[Y:%.*]], [[T2_WIDE]] +; CHECK-NEXT: [[T3_TRUNC:%.*]] = trunc <2 x i64> [[T3]] to <2 x i32> +; CHECK-NEXT: [[T4:%.*]] = and <2 x i32> [[T1]], [[T3_TRUNC]] +; CHECK-NEXT: [[T5:%.*]] = icmp ne <2 x i32> [[T4]], zeroinitializer +; CHECK-NEXT: ret <2 x i1> [[T5]] +; + %t0 = sub <2 x i32> , %len + %t1 = lshr <2 x i32> %x, %t0 + %t2 = add <2 x i32> %len, + %t2_wide = zext <2 x i32> %t2 to <2 x i64> + %t3 = shl <2 x i64> %y, %t2_wide + %t3_trunc = trunc <2 x i64> %t3 to <2 x i32> + %t4 = and <2 x i32> %t1, %t3_trunc + %t5 = icmp ne <2 x i32> %t4, + ret <2 x i1> %t5 +} + +;------------------------------------------------------------------------------- +; Commutativity tests +;------------------------------------------------------------------------------- + +declare i32 @gen32() +declare i64 @gen64() + +; While 'and' is commutative, the 'trunc' *always* seems to be getting +; canonicalized to the RHS, it does not seem possible to prevent that. + +;------------------------------------------------------------------------------- +; One-use tests +;------------------------------------------------------------------------------- + +declare void @use32(i32) +declare void @use64(i64) + +; Nope, everything has extra uses. +define i1 @t3_oneuse0(i32 %x, i64 %y, i32 %len) { +; CHECK-LABEL: @t3_oneuse0( +; CHECK-NEXT: [[T0:%.*]] = sub i32 32, [[LEN:%.*]] +; CHECK-NEXT: call void @use32(i32 [[T0]]) +; CHECK-NEXT: [[T1:%.*]] = lshr i32 [[X:%.*]], [[T0]] +; CHECK-NEXT: call void @use32(i32 [[T1]]) +; CHECK-NEXT: [[T2:%.*]] = add i32 [[LEN]], -1 +; CHECK-NEXT: call void @use32(i32 [[T2]]) +; CHECK-NEXT: [[T2_WIDE:%.*]] = zext i32 [[T2]] to i64 +; CHECK-NEXT: call void @use64(i64 [[T2_WIDE]]) +; CHECK-NEXT: [[T3:%.*]] = shl i64 [[Y:%.*]], [[T2_WIDE]] +; CHECK-NEXT: call void @use64(i64 [[T3]]) +; CHECK-NEXT: [[T3_TRUNC:%.*]] = trunc i64 [[T3]] to i32 +; CHECK-NEXT: call void @use32(i32 [[T3_TRUNC]]) +; CHECK-NEXT: [[T4:%.*]] = and i32 [[T1]], [[T3_TRUNC]] +; CHECK-NEXT: call void @use32(i32 [[T4]]) +; CHECK-NEXT: [[T5:%.*]] = icmp ne i32 [[T4]], 0 +; CHECK-NEXT: ret i1 [[T5]] +; + %t0 = sub i32 32, %len + call void @use32(i32 %t0) + %t1 = lshr i32 %x, %t0 + call void @use32(i32 %t1) + %t2 = add i32 %len, -1 + call void @use32(i32 %t2) + %t2_wide = zext i32 %t2 to i64 + call void @use64(i64 %t2_wide) + %t3 = shl i64 %y, %t2_wide + call void @use64(i64 %t3) + %t3_trunc = trunc i64 %t3 to i32 + call void @use32(i32 %t3_trunc) + %t4 = and i32 %t1, %t3_trunc + call void @use32(i32 %t4) + %t5 = icmp ne i32 %t4, 0 + ret i1 %t5 +} + +; Nope, still too much extra uses. +define i1 @t4_oneuse1(i32 %x, i64 %y, i32 %len) { +; CHECK-LABEL: @t4_oneuse1( +; CHECK-NEXT: [[T0:%.*]] = sub i32 32, [[LEN:%.*]] +; CHECK-NEXT: call void @use32(i32 [[T0]]) +; CHECK-NEXT: [[T1:%.*]] = lshr i32 [[X:%.*]], [[T0]] +; CHECK-NEXT: call void @use32(i32 [[T1]]) +; CHECK-NEXT: [[T2:%.*]] = add i32 [[LEN]], -1 +; CHECK-NEXT: call void @use32(i32 [[T2]]) +; CHECK-NEXT: [[T2_WIDE:%.*]] = zext i32 [[T2]] to i64 +; CHECK-NEXT: call void @use64(i64 [[T2_WIDE]]) +; CHECK-NEXT: [[T3:%.*]] = shl i64 [[Y:%.*]], [[T2_WIDE]] +; CHECK-NEXT: call void @use64(i64 [[T3]]) +; CHECK-NEXT: [[T3_TRUNC:%.*]] = trunc i64 [[T3]] to i32 +; CHECK-NEXT: call void @use32(i32 [[T3_TRUNC]]) +; CHECK-NEXT: [[T4:%.*]] = and i32 [[T1]], [[T3_TRUNC]] +; CHECK-NEXT: [[T5:%.*]] = icmp ne i32 [[T4]], 0 +; CHECK-NEXT: ret i1 [[T5]] +; + %t0 = sub i32 32, %len + call void @use32(i32 %t0) + %t1 = lshr i32 %x, %t0 + call void @use32(i32 %t1) + %t2 = add i32 %len, -1 + call void @use32(i32 %t2) + %t2_wide = zext i32 %t2 to i64 + call void @use64(i64 %t2_wide) + %t3 = shl i64 %y, %t2_wide + call void @use64(i64 %t3) + %t3_trunc = trunc i64 %t3 to i32 + call void @use32(i32 %t3_trunc) + %t4 = and i32 %t1, %t3_trunc ; no extra uses + %t5 = icmp ne i32 %t4, 0 + ret i1 %t5 +} + +; Still too much extra uses. +define i1 @t5_oneuse2(i32 %x, i64 %y, i32 %len) { +; CHECK-LABEL: @t5_oneuse2( +; CHECK-NEXT: [[T0:%.*]] = sub i32 32, [[LEN:%.*]] +; CHECK-NEXT: call void @use32(i32 [[T0]]) +; CHECK-NEXT: [[T1:%.*]] = lshr i32 [[X:%.*]], [[T0]] +; CHECK-NEXT: [[T2:%.*]] = add i32 [[LEN]], -1 +; CHECK-NEXT: call void @use32(i32 [[T2]]) +; CHECK-NEXT: [[T2_WIDE:%.*]] = zext i32 [[T2]] to i64 +; CHECK-NEXT: call void @use64(i64 [[T2_WIDE]]) +; CHECK-NEXT: [[T3:%.*]] = shl i64 [[Y:%.*]], [[T2_WIDE]] +; CHECK-NEXT: call void @use64(i64 [[T3]]) +; CHECK-NEXT: [[T3_TRUNC:%.*]] = trunc i64 [[T3]] to i32 +; CHECK-NEXT: call void @use32(i32 [[T3_TRUNC]]) +; CHECK-NEXT: [[T4:%.*]] = and i32 [[T1]], [[T3_TRUNC]] +; CHECK-NEXT: [[T5:%.*]] = icmp ne i32 [[T4]], 0 +; CHECK-NEXT: ret i1 [[T5]] +; + %t0 = sub i32 32, %len + call void @use32(i32 %t0) + %t1 = lshr i32 %x, %t0 ; no extra uses + %t2 = add i32 %len, -1 + call void @use32(i32 %t2) + %t2_wide = zext i32 %t2 to i64 + call void @use64(i64 %t2_wide) + %t3 = shl i64 %y, %t2_wide + call void @use64(i64 %t3) + %t3_trunc = trunc i64 %t3 to i32 + call void @use32(i32 %t3_trunc) + %t4 = and i32 %t1, %t3_trunc ; no extra uses + %t5 = icmp ne i32 %t4, 0 + ret i1 %t5 +} + +; Ok, trunc has no extra uses. +define i1 @t6_oneuse3(i32 %x, i64 %y, i32 %len) { +; CHECK-LABEL: @t6_oneuse3( +; CHECK-NEXT: [[T0:%.*]] = sub i32 32, [[LEN:%.*]] +; CHECK-NEXT: call void @use32(i32 [[T0]]) +; CHECK-NEXT: [[T1:%.*]] = lshr i32 [[X:%.*]], [[T0]] +; CHECK-NEXT: [[T2:%.*]] = add i32 [[LEN]], -1 +; CHECK-NEXT: call void @use32(i32 [[T2]]) +; CHECK-NEXT: [[T2_WIDE:%.*]] = zext i32 [[T2]] to i64 +; CHECK-NEXT: call void @use64(i64 [[T2_WIDE]]) +; CHECK-NEXT: [[T3:%.*]] = shl i64 [[Y:%.*]], [[T2_WIDE]] +; CHECK-NEXT: call void @use64(i64 [[T3]]) +; CHECK-NEXT: [[T3_TRUNC:%.*]] = trunc i64 [[T3]] to i32 +; CHECK-NEXT: [[T4:%.*]] = and i32 [[T1]], [[T3_TRUNC]] +; CHECK-NEXT: [[T5:%.*]] = icmp ne i32 [[T4]], 0 +; CHECK-NEXT: ret i1 [[T5]] +; + %t0 = sub i32 32, %len + call void @use32(i32 %t0) + %t1 = lshr i32 %x, %t0 ; no extra uses + %t2 = add i32 %len, -1 + call void @use32(i32 %t2) + %t2_wide = zext i32 %t2 to i64 + call void @use64(i64 %t2_wide) + %t3 = shl i64 %y, %t2_wide + call void @use64(i64 %t3) + %t3_trunc = trunc i64 %t3 to i32 ; no extra uses + %t4 = and i32 %t1, %t3_trunc ; no extra uses + %t5 = icmp ne i32 %t4, 0 + ret i1 %t5 +} + +; Ok, shift amount of non-truncated shift has no extra uses; +define i1 @t7_oneuse4(i32 %x, i64 %y, i32 %len) { +; CHECK-LABEL: @t7_oneuse4( +; CHECK-NEXT: [[T0:%.*]] = sub i32 32, [[LEN:%.*]] +; CHECK-NEXT: [[T1:%.*]] = lshr i32 [[X:%.*]], [[T0]] +; CHECK-NEXT: [[T2:%.*]] = add i32 [[LEN]], -1 +; CHECK-NEXT: call void @use32(i32 [[T2]]) +; CHECK-NEXT: [[T2_WIDE:%.*]] = zext i32 [[T2]] to i64 +; CHECK-NEXT: call void @use64(i64 [[T2_WIDE]]) +; CHECK-NEXT: [[T3:%.*]] = shl i64 [[Y:%.*]], [[T2_WIDE]] +; CHECK-NEXT: call void @use64(i64 [[T3]]) +; CHECK-NEXT: [[T3_TRUNC:%.*]] = trunc i64 [[T3]] to i32 +; CHECK-NEXT: call void @use32(i32 [[T3_TRUNC]]) +; CHECK-NEXT: [[T4:%.*]] = and i32 [[T1]], [[T3_TRUNC]] +; CHECK-NEXT: [[T5:%.*]] = icmp ne i32 [[T4]], 0 +; CHECK-NEXT: ret i1 [[T5]] +; + %t0 = sub i32 32, %len ; no extra uses + %t1 = lshr i32 %x, %t0 ; no extra uses + %t2 = add i32 %len, -1 + call void @use32(i32 %t2) + %t2_wide = zext i32 %t2 to i64 + call void @use64(i64 %t2_wide) + %t3 = shl i64 %y, %t2_wide + call void @use64(i64 %t3) + %t3_trunc = trunc i64 %t3 to i32 + call void @use32(i32 %t3_trunc) + %t4 = and i32 %t1, %t3_trunc ; no extra uses + %t5 = icmp ne i32 %t4, 0 + ret i1 %t5 +} + +; Ok, non-truncated shift is of constant; +define i1 @t8_oneuse5(i32 %x, i64 %y, i32 %len) { +; CHECK-LABEL: @t8_oneuse5( +; CHECK-NEXT: [[T0:%.*]] = sub i32 32, [[LEN:%.*]] +; CHECK-NEXT: call void @use32(i32 [[T0]]) +; CHECK-NEXT: [[T1:%.*]] = lshr i32 -52543054, [[T0]] +; CHECK-NEXT: call void @use32(i32 [[T1]]) +; CHECK-NEXT: [[T2:%.*]] = add i32 [[LEN]], -1 +; CHECK-NEXT: call void @use32(i32 [[T2]]) +; CHECK-NEXT: [[T2_WIDE:%.*]] = zext i32 [[T2]] to i64 +; CHECK-NEXT: call void @use64(i64 [[T2_WIDE]]) +; CHECK-NEXT: [[T3:%.*]] = shl i64 [[Y:%.*]], [[T2_WIDE]] +; CHECK-NEXT: call void @use64(i64 [[T3]]) +; CHECK-NEXT: [[T3_TRUNC:%.*]] = trunc i64 [[T3]] to i32 +; CHECK-NEXT: call void @use32(i32 [[T3_TRUNC]]) +; CHECK-NEXT: [[T4:%.*]] = and i32 [[T1]], [[T3_TRUNC]] +; CHECK-NEXT: [[T5:%.*]] = icmp ne i32 [[T4]], 0 +; CHECK-NEXT: ret i1 [[T5]] +; + %t0 = sub i32 32, %len + call void @use32(i32 %t0) + %t1 = lshr i32 4242424242, %t0 ; shift-of-constant + call void @use32(i32 %t1) + %t2 = add i32 %len, -1 + call void @use32(i32 %t2) + %t2_wide = zext i32 %t2 to i64 + call void @use64(i64 %t2_wide) + %t3 = shl i64 %y, %t2_wide + call void @use64(i64 %t3) + %t3_trunc = trunc i64 %t3 to i32 + call void @use32(i32 %t3_trunc) + %t4 = and i32 %t1, %t3_trunc ; no extra uses + %t5 = icmp ne i32 %t4, 0 + ret i1 %t5 +} + +; Ok, truncated shift is of constant; +define i1 @t9_oneuse5(i32 %x, i64 %y, i32 %len) { +; CHECK-LABEL: @t9_oneuse5( +; CHECK-NEXT: [[T0:%.*]] = sub i32 32, [[LEN:%.*]] +; CHECK-NEXT: call void @use32(i32 [[T0]]) +; CHECK-NEXT: [[T1:%.*]] = lshr i32 [[X:%.*]], [[T0]] +; CHECK-NEXT: call void @use32(i32 [[T1]]) +; CHECK-NEXT: [[T2:%.*]] = add i32 [[LEN]], -1 +; CHECK-NEXT: call void @use32(i32 [[T2]]) +; CHECK-NEXT: [[T2_WIDE:%.*]] = zext i32 [[T2]] to i64 +; CHECK-NEXT: call void @use64(i64 [[T2_WIDE]]) +; CHECK-NEXT: [[T3:%.*]] = shl i64 4242424242, [[T2_WIDE]] +; CHECK-NEXT: call void @use64(i64 [[T3]]) +; CHECK-NEXT: [[T3_TRUNC:%.*]] = trunc i64 [[T3]] to i32 +; CHECK-NEXT: call void @use32(i32 [[T3_TRUNC]]) +; CHECK-NEXT: [[T4:%.*]] = and i32 [[T1]], [[T3_TRUNC]] +; CHECK-NEXT: [[T5:%.*]] = icmp ne i32 [[T4]], 0 +; CHECK-NEXT: ret i1 [[T5]] +; + %t0 = sub i32 32, %len + call void @use32(i32 %t0) + %t1 = lshr i32 %x, %t0 ; shift-of-constant + call void @use32(i32 %t1) + %t2 = add i32 %len, -1 + call void @use32(i32 %t2) + %t2_wide = zext i32 %t2 to i64 + call void @use64(i64 %t2_wide) + %t3 = shl i64 4242424242, %t2_wide + call void @use64(i64 %t3) + %t3_trunc = trunc i64 %t3 to i32 + call void @use32(i32 %t3_trunc) + %t4 = and i32 %t1, %t3_trunc ; no extra uses + %t5 = icmp ne i32 %t4, 0 + ret i1 %t5 +} + +;------------------------------------------------------------------------------- +; Commutativity with extra uses +;------------------------------------------------------------------------------- + +; While 'and' is commutative, the 'trunc' *always* seems to be getting +; canonicalized to the RHS, it does not seem possible to prevent that. + +; + +; Constant shift amounts + +define i1 @t10_constants(i32 %x, i64 %y) { +; CHECK-LABEL: @t10_constants( +; CHECK-NEXT: [[Y_TR:%.*]] = trunc i64 [[Y:%.*]] to i32 +; CHECK-NEXT: [[TMP1:%.*]] = lshr i32 [[X:%.*]], 26 +; CHECK-NEXT: [[TMP2:%.*]] = and i32 [[TMP1]], [[Y_TR]] +; CHECK-NEXT: [[TMP3:%.*]] = icmp ne i32 [[TMP2]], 0 +; CHECK-NEXT: ret i1 [[TMP3]] +; + %t0 = lshr i32 %x, 12 + %t1 = shl i64 %y, 14 + %t1_trunc = trunc i64 %t1 to i32 + %t2 = and i32 %t0, %t1_trunc + %t3 = icmp ne i32 %t2, 0 + ret i1 %t3 +} + +define <2 x i1> @t11_constants_vec_splat(<2 x i32> %x, <2 x i64> %y) { +; CHECK-LABEL: @t11_constants_vec_splat( +; CHECK-NEXT: [[T0:%.*]] = lshr <2 x i32> [[X:%.*]], +; CHECK-NEXT: [[T1:%.*]] = shl <2 x i64> [[Y:%.*]], +; CHECK-NEXT: [[T1_TRUNC:%.*]] = trunc <2 x i64> [[T1]] to <2 x i32> +; CHECK-NEXT: [[T2:%.*]] = and <2 x i32> [[T0]], [[T1_TRUNC]] +; CHECK-NEXT: [[T3:%.*]] = icmp ne <2 x i32> [[T2]], zeroinitializer +; CHECK-NEXT: ret <2 x i1> [[T3]] +; + %t0 = lshr <2 x i32> %x, + %t1 = shl <2 x i64> %y, + %t1_trunc = trunc <2 x i64> %t1 to <2 x i32> + %t2 = and <2 x i32> %t0, %t1_trunc + %t3 = icmp ne <2 x i32> %t2, + ret <2 x i1> %t3 +} +define <2 x i1> @t12_constants_vec_nonsplat(<2 x i32> %x, <2 x i64> %y) { +; CHECK-LABEL: @t12_constants_vec_nonsplat( +; CHECK-NEXT: [[T0:%.*]] = lshr <2 x i32> [[X:%.*]], +; CHECK-NEXT: [[T1:%.*]] = shl <2 x i64> [[Y:%.*]], +; CHECK-NEXT: [[T1_TRUNC:%.*]] = trunc <2 x i64> [[T1]] to <2 x i32> +; CHECK-NEXT: [[T2:%.*]] = and <2 x i32> [[T0]], [[T1_TRUNC]] +; CHECK-NEXT: [[T3:%.*]] = icmp ne <2 x i32> [[T2]], zeroinitializer +; CHECK-NEXT: ret <2 x i1> [[T3]] +; + %t0 = lshr <2 x i32> %x, + %t1 = shl <2 x i64> %y, + %t1_trunc = trunc <2 x i64> %t1 to <2 x i32> + %t2 = and <2 x i32> %t0, %t1_trunc + %t3 = icmp ne <2 x i32> %t2, + ret <2 x i1> %t3 +} + +;------------------------------------------------------------------------------- +; Negative tests +;------------------------------------------------------------------------------- + +define i1 @n13_overshift(i32 %x, i64 %y, i32 %len) { +; CHECK-LABEL: @n13_overshift( +; CHECK-NEXT: [[T0:%.*]] = sub i32 32, [[LEN:%.*]] +; CHECK-NEXT: [[T1:%.*]] = lshr i32 [[X:%.*]], [[T0]] +; CHECK-NEXT: [[T2:%.*]] = add i32 [[LEN]], 1 +; CHECK-NEXT: [[T2_WIDE:%.*]] = zext i32 [[T2]] to i64 +; CHECK-NEXT: [[T3:%.*]] = shl i64 [[Y:%.*]], [[T2_WIDE]] +; CHECK-NEXT: [[T3_TRUNC:%.*]] = trunc i64 [[T3]] to i32 +; CHECK-NEXT: [[T4:%.*]] = and i32 [[T1]], [[T3_TRUNC]] +; CHECK-NEXT: [[T5:%.*]] = icmp ne i32 [[T4]], 0 +; CHECK-NEXT: ret i1 [[T5]] +; + %t0 = sub i32 32, %len + %t1 = lshr i32 %x, %t0 + %t2 = add i32 %len, 1 ; too much + %t2_wide = zext i32 %t2 to i64 + %t3 = shl i64 %y, %t2_wide + %t3_trunc = trunc i64 %t3 to i32 + %t4 = and i32 %t1, %t3_trunc + %t5 = icmp ne i32 %t4, 0 + ret i1 %t5 +} + +define i1 @n14_trunc_of_lshr(i64 %x, i32 %y, i32 %len) { +; CHECK-LABEL: @n14_trunc_of_lshr( +; CHECK-NEXT: [[T0:%.*]] = sub i32 32, [[LEN:%.*]] +; CHECK-NEXT: [[T0_WIDE:%.*]] = zext i32 [[T0]] to i64 +; CHECK-NEXT: [[T1:%.*]] = lshr i64 [[X:%.*]], [[T0_WIDE]] +; CHECK-NEXT: [[T1_TRUNC:%.*]] = trunc i64 [[T1]] to i32 +; CHECK-NEXT: [[T2:%.*]] = add i32 [[LEN]], -1 +; CHECK-NEXT: [[T3:%.*]] = shl i32 [[Y:%.*]], [[T2]] +; CHECK-NEXT: [[T4:%.*]] = and i32 [[T3]], [[T1_TRUNC]] +; CHECK-NEXT: [[T5:%.*]] = icmp ne i32 [[T4]], 0 +; CHECK-NEXT: ret i1 [[T5]] +; + %t0 = sub i32 32, %len + %t0_wide = zext i32 %t0 to i64 + %t1 = lshr i64 %x, %t0_wide + %t1_trunc = trunc i64 %t1 to i32 + %t2 = add i32 %len, -1 + %t3 = shl i32 %y, %t2 + %t4 = and i32 %t1_trunc, %t3 + %t5 = icmp ne i32 %t4, 0 + ret i1 %t5 +} + +; Completely variable shift amounts + +define i1 @n15_variable_shamts(i32 %x, i64 %y, i32 %shamt0, i64 %shamt1) { +; CHECK-LABEL: @n15_variable_shamts( +; CHECK-NEXT: [[T0:%.*]] = lshr i32 [[X:%.*]], [[SHAMT0:%.*]] +; CHECK-NEXT: [[T1:%.*]] = shl i64 [[Y:%.*]], [[SHAMT1:%.*]] +; CHECK-NEXT: [[T1_TRUNC:%.*]] = trunc i64 [[T1]] to i32 +; CHECK-NEXT: [[T2:%.*]] = and i32 [[T0]], [[T1_TRUNC]] +; CHECK-NEXT: [[T3:%.*]] = icmp ne i32 [[T2]], 0 +; CHECK-NEXT: ret i1 [[T3]] +; + %t0 = lshr i32 %x, %shamt0 + %t1 = shl i64 %y, %shamt1 + %t1_trunc = trunc i64 %t1 to i32 + %t2 = and i32 %t1_trunc, %t0 + %t3 = icmp ne i32 %t2, 0 + ret i1 %t3 +}