def : AtomicPat<int_loongarch_masked_atomicrmw_sub_i64,
PseudoMaskedAtomicLoadSub32>;
def : Pat<(atomic_load_nand_64 GPR:$rj, GPR:$rk),
- (PseudoAtomicLoadNand64 GPR:$rk, GPR:$rj)>;
+ (PseudoAtomicLoadNand64 GPR:$rj, GPR:$rk)>;
def : AtomicPat<int_loongarch_masked_atomicrmw_nand_i64,
PseudoMaskedAtomicLoadNand32>;
def : Pat<(atomic_load_add_32 GPR:$rj, GPR:$rk),
} // Predicates = [IsLA64]
def : Pat<(atomic_load_nand_32 GPR:$rj, GPR:$rk),
- (PseudoAtomicLoadNand32 GPR:$rk, GPR:$rj)>;
+ (PseudoAtomicLoadNand32 GPR:$rj, GPR:$rk)>;
let Predicates = [IsLA32] in {
def : AtomicPat<int_loongarch_masked_atomicrmw_xchg_i32,
; LA32: # %bb.0:
; LA32-NEXT: .LBB14_1: # =>This Inner Loop Header: Depth=1
; LA32-NEXT: dbar 0
-; LA32-NEXT: ll.w $a2, $a1, 0
-; LA32-NEXT: and $a3, $a2, $a0
+; LA32-NEXT: ll.w $a2, $a0, 0
+; LA32-NEXT: and $a3, $a2, $a1
; LA32-NEXT: nor $a3, $a3, $zero
-; LA32-NEXT: sc.w $a3, $a1, 0
+; LA32-NEXT: sc.w $a3, $a0, 0
; LA32-NEXT: beqz $a3, .LBB14_1
; LA32-NEXT: # %bb.2:
; LA32-NEXT: move $a0, $a2
; LA64: # %bb.0:
; LA64-NEXT: .LBB14_1: # =>This Inner Loop Header: Depth=1
; LA64-NEXT: dbar 0
-; LA64-NEXT: ll.w $a2, $a1, 0
-; LA64-NEXT: and $a3, $a2, $a0
+; LA64-NEXT: ll.w $a2, $a0, 0
+; LA64-NEXT: and $a3, $a2, $a1
; LA64-NEXT: nor $a3, $a3, $zero
-; LA64-NEXT: sc.w $a3, $a1, 0
+; LA64-NEXT: sc.w $a3, $a0, 0
; LA64-NEXT: beqz $a3, .LBB14_1
; LA64-NEXT: # %bb.2:
; LA64-NEXT: move $a0, $a2
; LA64: # %bb.0:
; LA64-NEXT: .LBB15_1: # =>This Inner Loop Header: Depth=1
; LA64-NEXT: dbar 0
-; LA64-NEXT: ll.d $a2, $a1, 0
-; LA64-NEXT: and $a3, $a2, $a0
+; LA64-NEXT: ll.d $a2, $a0, 0
+; LA64-NEXT: and $a3, $a2, $a1
; LA64-NEXT: nor $a3, $a3, $zero
-; LA64-NEXT: sc.d $a3, $a1, 0
+; LA64-NEXT: sc.d $a3, $a0, 0
; LA64-NEXT: beqz $a3, .LBB15_1
; LA64-NEXT: # %bb.2:
; LA64-NEXT: move $a0, $a2