llvm/llvm/test/CodeGen/LoongArch/atomicrmw-uinc-udec-wrap.ll

; NOTE: Assertions have been autogenerated by utils/update_llc_test_checks.py
; RUN: llc --mtriple=loongarch64 -mattr=+d < %s | FileCheck --check-prefix=LA64 %s

define i8 @atomicrmw_uinc_wrap_i8(ptr %ptr, i8 %val) {
; LA64-LABEL: atomicrmw_uinc_wrap_i8:
; LA64:       # %bb.0:
; LA64-NEXT:    slli.d $a3, $a0, 3
; LA64-NEXT:    bstrins.d $a0, $zero, 1, 0
; LA64-NEXT:    andi $a2, $a3, 24
; LA64-NEXT:    ori $a5, $zero, 255
; LA64-NEXT:    ld.w $a4, $a0, 0
; LA64-NEXT:    sll.w $a3, $a5, $a3
; LA64-NEXT:    nor $a3, $a3, $zero
; LA64-NEXT:    andi $a1, $a1, 255
; LA64-NEXT:    .p2align 4, , 16
; LA64-NEXT:  .LBB0_1: # %atomicrmw.start
; LA64-NEXT:    # =>This Loop Header: Depth=1
; LA64-NEXT:    # Child Loop BB0_3 Depth 2
; LA64-NEXT:    move $a5, $a4
; LA64-NEXT:    srl.w $a4, $a4, $a2
; LA64-NEXT:    andi $a6, $a4, 255
; LA64-NEXT:    addi.d $a4, $a4, 1
; LA64-NEXT:    sltu $a6, $a6, $a1
; LA64-NEXT:    xori $a6, $a6, 1
; LA64-NEXT:    masknez $a4, $a4, $a6
; LA64-NEXT:    andi $a4, $a4, 255
; LA64-NEXT:    sll.w $a4, $a4, $a2
; LA64-NEXT:    and $a6, $a5, $a3
; LA64-NEXT:    or $a6, $a6, $a4
; LA64-NEXT:  .LBB0_3: # %atomicrmw.start
; LA64-NEXT:    # Parent Loop BB0_1 Depth=1
; LA64-NEXT:    # => This Inner Loop Header: Depth=2
; LA64-NEXT:    ll.w $a4, $a0, 0
; LA64-NEXT:    bne $a4, $a5, .LBB0_5
; LA64-NEXT:  # %bb.4: # %atomicrmw.start
; LA64-NEXT:    # in Loop: Header=BB0_3 Depth=2
; LA64-NEXT:    move $a7, $a6
; LA64-NEXT:    sc.w $a7, $a0, 0
; LA64-NEXT:    beqz $a7, .LBB0_3
; LA64-NEXT:    b .LBB0_6
; LA64-NEXT:  .LBB0_5: # %atomicrmw.start
; LA64-NEXT:    # in Loop: Header=BB0_1 Depth=1
; LA64-NEXT:    dbar 20
; LA64-NEXT:  .LBB0_6: # %atomicrmw.start
; LA64-NEXT:    # in Loop: Header=BB0_1 Depth=1
; LA64-NEXT:    bne $a4, $a5, .LBB0_1
; LA64-NEXT:  # %bb.2: # %atomicrmw.end
; LA64-NEXT:    srl.w $a0, $a4, $a2
; LA64-NEXT:    ret
  %result = atomicrmw uinc_wrap ptr %ptr, i8 %val seq_cst
  ret i8 %result
}

define i16 @atomicrmw_uinc_wrap_i16(ptr %ptr, i16 %val) {
; LA64-LABEL: atomicrmw_uinc_wrap_i16:
; LA64:       # %bb.0:
; LA64-NEXT:    slli.d $a3, $a0, 3
; LA64-NEXT:    bstrins.d $a0, $zero, 1, 0
; LA64-NEXT:    andi $a2, $a3, 24
; LA64-NEXT:    lu12i.w $a4, 15
; LA64-NEXT:    ori $a5, $a4, 4095
; LA64-NEXT:    ld.w $a4, $a0, 0
; LA64-NEXT:    sll.w $a3, $a5, $a3
; LA64-NEXT:    nor $a3, $a3, $zero
; LA64-NEXT:    bstrpick.d $a1, $a1, 15, 0
; LA64-NEXT:    .p2align 4, , 16
; LA64-NEXT:  .LBB1_1: # %atomicrmw.start
; LA64-NEXT:    # =>This Loop Header: Depth=1
; LA64-NEXT:    # Child Loop BB1_3 Depth 2
; LA64-NEXT:    move $a5, $a4
; LA64-NEXT:    srl.w $a4, $a4, $a2
; LA64-NEXT:    bstrpick.d $a6, $a4, 15, 0
; LA64-NEXT:    addi.d $a4, $a4, 1
; LA64-NEXT:    sltu $a6, $a6, $a1
; LA64-NEXT:    xori $a6, $a6, 1
; LA64-NEXT:    masknez $a4, $a4, $a6
; LA64-NEXT:    bstrpick.d $a4, $a4, 15, 0
; LA64-NEXT:    sll.w $a4, $a4, $a2
; LA64-NEXT:    and $a6, $a5, $a3
; LA64-NEXT:    or $a6, $a6, $a4
; LA64-NEXT:  .LBB1_3: # %atomicrmw.start
; LA64-NEXT:    # Parent Loop BB1_1 Depth=1
; LA64-NEXT:    # => This Inner Loop Header: Depth=2
; LA64-NEXT:    ll.w $a4, $a0, 0
; LA64-NEXT:    bne $a4, $a5, .LBB1_5
; LA64-NEXT:  # %bb.4: # %atomicrmw.start
; LA64-NEXT:    # in Loop: Header=BB1_3 Depth=2
; LA64-NEXT:    move $a7, $a6
; LA64-NEXT:    sc.w $a7, $a0, 0
; LA64-NEXT:    beqz $a7, .LBB1_3
; LA64-NEXT:    b .LBB1_6
; LA64-NEXT:  .LBB1_5: # %atomicrmw.start
; LA64-NEXT:    # in Loop: Header=BB1_1 Depth=1
; LA64-NEXT:    dbar 20
; LA64-NEXT:  .LBB1_6: # %atomicrmw.start
; LA64-NEXT:    # in Loop: Header=BB1_1 Depth=1
; LA64-NEXT:    bne $a4, $a5, .LBB1_1
; LA64-NEXT:  # %bb.2: # %atomicrmw.end
; LA64-NEXT:    srl.w $a0, $a4, $a2
; LA64-NEXT:    ret
  %result = atomicrmw uinc_wrap ptr %ptr, i16 %val seq_cst
  ret i16 %result
}

define i32 @atomicrmw_uinc_wrap_i32(ptr %ptr, i32 %val) {
; LA64-LABEL: atomicrmw_uinc_wrap_i32:
; LA64:       # %bb.0:
; LA64-NEXT:    ld.w $a2, $a0, 0
; LA64-NEXT:    addi.w $a1, $a1, 0
; LA64-NEXT:    .p2align 4, , 16
; LA64-NEXT:  .LBB2_1: # %atomicrmw.start
; LA64-NEXT:    # =>This Loop Header: Depth=1
; LA64-NEXT:    # Child Loop BB2_3 Depth 2
; LA64-NEXT:    move $a3, $a2
; LA64-NEXT:    addi.w $a2, $a2, 1
; LA64-NEXT:    sltu $a4, $a3, $a1
; LA64-NEXT:    xori $a4, $a4, 1
; LA64-NEXT:    masknez $a4, $a2, $a4
; LA64-NEXT:  .LBB2_3: # %atomicrmw.start
; LA64-NEXT:    # Parent Loop BB2_1 Depth=1
; LA64-NEXT:    # => This Inner Loop Header: Depth=2
; LA64-NEXT:    ll.w $a2, $a0, 0
; LA64-NEXT:    bne $a2, $a3, .LBB2_5
; LA64-NEXT:  # %bb.4: # %atomicrmw.start
; LA64-NEXT:    # in Loop: Header=BB2_3 Depth=2
; LA64-NEXT:    move $a5, $a4
; LA64-NEXT:    sc.w $a5, $a0, 0
; LA64-NEXT:    beqz $a5, .LBB2_3
; LA64-NEXT:    b .LBB2_6
; LA64-NEXT:  .LBB2_5: # %atomicrmw.start
; LA64-NEXT:    # in Loop: Header=BB2_1 Depth=1
; LA64-NEXT:    dbar 20
; LA64-NEXT:  .LBB2_6: # %atomicrmw.start
; LA64-NEXT:    # in Loop: Header=BB2_1 Depth=1
; LA64-NEXT:    bne $a2, $a3, .LBB2_1
; LA64-NEXT:  # %bb.2: # %atomicrmw.end
; LA64-NEXT:    move $a0, $a2
; LA64-NEXT:    ret
  %result = atomicrmw uinc_wrap ptr %ptr, i32 %val seq_cst
  ret i32 %result
}

define i64 @atomicrmw_uinc_wrap_i64(ptr %ptr, i64 %val) {
; LA64-LABEL: atomicrmw_uinc_wrap_i64:
; LA64:       # %bb.0:
; LA64-NEXT:    ld.d $a2, $a0, 0
; LA64-NEXT:    .p2align 4, , 16
; LA64-NEXT:  .LBB3_1: # %atomicrmw.start
; LA64-NEXT:    # =>This Loop Header: Depth=1
; LA64-NEXT:    # Child Loop BB3_3 Depth 2
; LA64-NEXT:    move $a3, $a2
; LA64-NEXT:    addi.d $a2, $a2, 1
; LA64-NEXT:    sltu $a4, $a3, $a1
; LA64-NEXT:    xori $a4, $a4, 1
; LA64-NEXT:    masknez $a4, $a2, $a4
; LA64-NEXT:  .LBB3_3: # %atomicrmw.start
; LA64-NEXT:    # Parent Loop BB3_1 Depth=1
; LA64-NEXT:    # => This Inner Loop Header: Depth=2
; LA64-NEXT:    ll.d $a2, $a0, 0
; LA64-NEXT:    bne $a2, $a3, .LBB3_5
; LA64-NEXT:  # %bb.4: # %atomicrmw.start
; LA64-NEXT:    # in Loop: Header=BB3_3 Depth=2
; LA64-NEXT:    move $a5, $a4
; LA64-NEXT:    sc.d $a5, $a0, 0
; LA64-NEXT:    beqz $a5, .LBB3_3
; LA64-NEXT:    b .LBB3_6
; LA64-NEXT:  .LBB3_5: # %atomicrmw.start
; LA64-NEXT:    # in Loop: Header=BB3_1 Depth=1
; LA64-NEXT:    dbar 20
; LA64-NEXT:  .LBB3_6: # %atomicrmw.start
; LA64-NEXT:    # in Loop: Header=BB3_1 Depth=1
; LA64-NEXT:    bne $a2, $a3, .LBB3_1
; LA64-NEXT:  # %bb.2: # %atomicrmw.end
; LA64-NEXT:    move $a0, $a2
; LA64-NEXT:    ret
  %result = atomicrmw uinc_wrap ptr %ptr, i64 %val seq_cst
  ret i64 %result
}

define i8 @atomicrmw_udec_wrap_i8(ptr %ptr, i8 %val) {
; LA64-LABEL: atomicrmw_udec_wrap_i8:
; LA64:       # %bb.0:
; LA64-NEXT:    slli.d $a3, $a0, 3
; LA64-NEXT:    bstrins.d $a0, $zero, 1, 0
; LA64-NEXT:    andi $a2, $a3, 24
; LA64-NEXT:    ori $a4, $zero, 255
; LA64-NEXT:    ld.w $a5, $a0, 0
; LA64-NEXT:    sll.w $a3, $a4, $a3
; LA64-NEXT:    nor $a3, $a3, $zero
; LA64-NEXT:    andi $a4, $a1, 255
; LA64-NEXT:    .p2align 4, , 16
; LA64-NEXT:  .LBB4_1: # %atomicrmw.start
; LA64-NEXT:    # =>This Loop Header: Depth=1
; LA64-NEXT:    # Child Loop BB4_3 Depth 2
; LA64-NEXT:    move $a6, $a5
; LA64-NEXT:    srl.w $a5, $a5, $a2
; LA64-NEXT:    andi $a7, $a5, 255
; LA64-NEXT:    addi.d $a5, $a5, -1
; LA64-NEXT:    sltui $t0, $a7, 1
; LA64-NEXT:    sltu $a7, $a4, $a7
; LA64-NEXT:    masknez $a5, $a5, $a7
; LA64-NEXT:    maskeqz $a7, $a1, $a7
; LA64-NEXT:    or $a5, $a7, $a5
; LA64-NEXT:    masknez $a5, $a5, $t0
; LA64-NEXT:    maskeqz $a7, $a1, $t0
; LA64-NEXT:    or $a5, $a7, $a5
; LA64-NEXT:    andi $a5, $a5, 255
; LA64-NEXT:    sll.w $a5, $a5, $a2
; LA64-NEXT:    and $a7, $a6, $a3
; LA64-NEXT:    or $a7, $a7, $a5
; LA64-NEXT:  .LBB4_3: # %atomicrmw.start
; LA64-NEXT:    # Parent Loop BB4_1 Depth=1
; LA64-NEXT:    # => This Inner Loop Header: Depth=2
; LA64-NEXT:    ll.w $a5, $a0, 0
; LA64-NEXT:    bne $a5, $a6, .LBB4_5
; LA64-NEXT:  # %bb.4: # %atomicrmw.start
; LA64-NEXT:    # in Loop: Header=BB4_3 Depth=2
; LA64-NEXT:    move $t0, $a7
; LA64-NEXT:    sc.w $t0, $a0, 0
; LA64-NEXT:    beqz $t0, .LBB4_3
; LA64-NEXT:    b .LBB4_6
; LA64-NEXT:  .LBB4_5: # %atomicrmw.start
; LA64-NEXT:    # in Loop: Header=BB4_1 Depth=1
; LA64-NEXT:    dbar 20
; LA64-NEXT:  .LBB4_6: # %atomicrmw.start
; LA64-NEXT:    # in Loop: Header=BB4_1 Depth=1
; LA64-NEXT:    bne $a5, $a6, .LBB4_1
; LA64-NEXT:  # %bb.2: # %atomicrmw.end
; LA64-NEXT:    srl.w $a0, $a5, $a2
; LA64-NEXT:    ret
  %result = atomicrmw udec_wrap ptr %ptr, i8 %val seq_cst
  ret i8 %result
}

define i16 @atomicrmw_udec_wrap_i16(ptr %ptr, i16 %val) {
; LA64-LABEL: atomicrmw_udec_wrap_i16:
; LA64:       # %bb.0:
; LA64-NEXT:    slli.d $a3, $a0, 3
; LA64-NEXT:    bstrins.d $a0, $zero, 1, 0
; LA64-NEXT:    andi $a2, $a3, 24
; LA64-NEXT:    lu12i.w $a4, 15
; LA64-NEXT:    ori $a4, $a4, 4095
; LA64-NEXT:    ld.w $a5, $a0, 0
; LA64-NEXT:    sll.w $a3, $a4, $a3
; LA64-NEXT:    nor $a3, $a3, $zero
; LA64-NEXT:    bstrpick.d $a4, $a1, 15, 0
; LA64-NEXT:    .p2align 4, , 16
; LA64-NEXT:  .LBB5_1: # %atomicrmw.start
; LA64-NEXT:    # =>This Loop Header: Depth=1
; LA64-NEXT:    # Child Loop BB5_3 Depth 2
; LA64-NEXT:    move $a6, $a5
; LA64-NEXT:    srl.w $a5, $a5, $a2
; LA64-NEXT:    bstrpick.d $a7, $a5, 15, 0
; LA64-NEXT:    addi.d $a5, $a5, -1
; LA64-NEXT:    sltui $t0, $a7, 1
; LA64-NEXT:    sltu $a7, $a4, $a7
; LA64-NEXT:    masknez $a5, $a5, $a7
; LA64-NEXT:    maskeqz $a7, $a1, $a7
; LA64-NEXT:    or $a5, $a7, $a5
; LA64-NEXT:    masknez $a5, $a5, $t0
; LA64-NEXT:    maskeqz $a7, $a1, $t0
; LA64-NEXT:    or $a5, $a7, $a5
; LA64-NEXT:    bstrpick.d $a5, $a5, 15, 0
; LA64-NEXT:    sll.w $a5, $a5, $a2
; LA64-NEXT:    and $a7, $a6, $a3
; LA64-NEXT:    or $a7, $a7, $a5
; LA64-NEXT:  .LBB5_3: # %atomicrmw.start
; LA64-NEXT:    # Parent Loop BB5_1 Depth=1
; LA64-NEXT:    # => This Inner Loop Header: Depth=2
; LA64-NEXT:    ll.w $a5, $a0, 0
; LA64-NEXT:    bne $a5, $a6, .LBB5_5
; LA64-NEXT:  # %bb.4: # %atomicrmw.start
; LA64-NEXT:    # in Loop: Header=BB5_3 Depth=2
; LA64-NEXT:    move $t0, $a7
; LA64-NEXT:    sc.w $t0, $a0, 0
; LA64-NEXT:    beqz $t0, .LBB5_3
; LA64-NEXT:    b .LBB5_6
; LA64-NEXT:  .LBB5_5: # %atomicrmw.start
; LA64-NEXT:    # in Loop: Header=BB5_1 Depth=1
; LA64-NEXT:    dbar 20
; LA64-NEXT:  .LBB5_6: # %atomicrmw.start
; LA64-NEXT:    # in Loop: Header=BB5_1 Depth=1
; LA64-NEXT:    bne $a5, $a6, .LBB5_1
; LA64-NEXT:  # %bb.2: # %atomicrmw.end
; LA64-NEXT:    srl.w $a0, $a5, $a2
; LA64-NEXT:    ret
  %result = atomicrmw udec_wrap ptr %ptr, i16 %val seq_cst
  ret i16 %result
}

define i32 @atomicrmw_udec_wrap_i32(ptr %ptr, i32 %val) {
; LA64-LABEL: atomicrmw_udec_wrap_i32:
; LA64:       # %bb.0:
; LA64-NEXT:    ld.w $a2, $a0, 0
; LA64-NEXT:    addi.w $a3, $a1, 0
; LA64-NEXT:    .p2align 4, , 16
; LA64-NEXT:  .LBB6_1: # %atomicrmw.start
; LA64-NEXT:    # =>This Loop Header: Depth=1
; LA64-NEXT:    # Child Loop BB6_3 Depth 2
; LA64-NEXT:    move $a4, $a2
; LA64-NEXT:    addi.w $a2, $a2, -1
; LA64-NEXT:    sltui $a5, $a4, 1
; LA64-NEXT:    sltu $a6, $a3, $a4
; LA64-NEXT:    masknez $a2, $a2, $a6
; LA64-NEXT:    maskeqz $a6, $a1, $a6
; LA64-NEXT:    or $a2, $a6, $a2
; LA64-NEXT:    masknez $a2, $a2, $a5
; LA64-NEXT:    maskeqz $a5, $a1, $a5
; LA64-NEXT:    or $a5, $a5, $a2
; LA64-NEXT:  .LBB6_3: # %atomicrmw.start
; LA64-NEXT:    # Parent Loop BB6_1 Depth=1
; LA64-NEXT:    # => This Inner Loop Header: Depth=2
; LA64-NEXT:    ll.w $a2, $a0, 0
; LA64-NEXT:    bne $a2, $a4, .LBB6_5
; LA64-NEXT:  # %bb.4: # %atomicrmw.start
; LA64-NEXT:    # in Loop: Header=BB6_3 Depth=2
; LA64-NEXT:    move $a6, $a5
; LA64-NEXT:    sc.w $a6, $a0, 0
; LA64-NEXT:    beqz $a6, .LBB6_3
; LA64-NEXT:    b .LBB6_6
; LA64-NEXT:  .LBB6_5: # %atomicrmw.start
; LA64-NEXT:    # in Loop: Header=BB6_1 Depth=1
; LA64-NEXT:    dbar 20
; LA64-NEXT:  .LBB6_6: # %atomicrmw.start
; LA64-NEXT:    # in Loop: Header=BB6_1 Depth=1
; LA64-NEXT:    bne $a2, $a4, .LBB6_1
; LA64-NEXT:  # %bb.2: # %atomicrmw.end
; LA64-NEXT:    move $a0, $a2
; LA64-NEXT:    ret
  %result = atomicrmw udec_wrap ptr %ptr, i32 %val seq_cst
  ret i32 %result
}

define i64 @atomicrmw_udec_wrap_i64(ptr %ptr, i64 %val) {
; LA64-LABEL: atomicrmw_udec_wrap_i64:
; LA64:       # %bb.0:
; LA64-NEXT:    ld.d $a2, $a0, 0
; LA64-NEXT:    .p2align 4, , 16
; LA64-NEXT:  .LBB7_1: # %atomicrmw.start
; LA64-NEXT:    # =>This Loop Header: Depth=1
; LA64-NEXT:    # Child Loop BB7_3 Depth 2
; LA64-NEXT:    move $a3, $a2
; LA64-NEXT:    addi.d $a2, $a2, -1
; LA64-NEXT:    sltui $a4, $a3, 1
; LA64-NEXT:    sltu $a5, $a1, $a3
; LA64-NEXT:    masknez $a2, $a2, $a5
; LA64-NEXT:    maskeqz $a5, $a1, $a5
; LA64-NEXT:    or $a2, $a5, $a2
; LA64-NEXT:    masknez $a2, $a2, $a4
; LA64-NEXT:    maskeqz $a4, $a1, $a4
; LA64-NEXT:    or $a4, $a4, $a2
; LA64-NEXT:  .LBB7_3: # %atomicrmw.start
; LA64-NEXT:    # Parent Loop BB7_1 Depth=1
; LA64-NEXT:    # => This Inner Loop Header: Depth=2
; LA64-NEXT:    ll.d $a2, $a0, 0
; LA64-NEXT:    bne $a2, $a3, .LBB7_5
; LA64-NEXT:  # %bb.4: # %atomicrmw.start
; LA64-NEXT:    # in Loop: Header=BB7_3 Depth=2
; LA64-NEXT:    move $a5, $a4
; LA64-NEXT:    sc.d $a5, $a0, 0
; LA64-NEXT:    beqz $a5, .LBB7_3
; LA64-NEXT:    b .LBB7_6
; LA64-NEXT:  .LBB7_5: # %atomicrmw.start
; LA64-NEXT:    # in Loop: Header=BB7_1 Depth=1
; LA64-NEXT:    dbar 20
; LA64-NEXT:  .LBB7_6: # %atomicrmw.start
; LA64-NEXT:    # in Loop: Header=BB7_1 Depth=1
; LA64-NEXT:    bne $a2, $a3, .LBB7_1
; LA64-NEXT:  # %bb.2: # %atomicrmw.end
; LA64-NEXT:    move $a0, $a2
; LA64-NEXT:    ret
  %result = atomicrmw udec_wrap ptr %ptr, i64 %val seq_cst
  ret i64 %result
}