;; Machine Description for shared bits common to IWMMXT and Neon.
;; Copyright (C) 2006-2022 Free Software Foundation, Inc.
;; Written by CodeSourcery.
;;
;; This file is part of GCC.
;;
;; GCC is free software; you can redistribute it and/or modify it
;; under the terms of the GNU General Public License as published by
;; the Free Software Foundation; either version 3, or (at your option)
;; any later version.
;;
;; GCC is distributed in the hope that it will be useful, but
;; WITHOUT ANY WARRANTY; without even the implied warranty of
;; MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU
;; General Public License for more details.
;;
;; You should have received a copy of the GNU General Public License
;; along with GCC; see the file COPYING3. If not see
;; .
;; Vector Moves
(define_expand "mov"
[(set (match_operand:VNIM1 0 "nonimmediate_operand")
(match_operand:VNIM1 1 "general_operand"))]
"TARGET_NEON
|| (TARGET_REALLY_IWMMXT && VALID_IWMMXT_REG_MODE (mode))
|| (TARGET_HAVE_MVE && VALID_MVE_SI_MODE (mode))
|| (TARGET_HAVE_MVE_FLOAT && VALID_MVE_SF_MODE (mode))"
{
gcc_checking_assert (aligned_operand (operands[0], mode));
gcc_checking_assert (aligned_operand (operands[1], mode));
if (can_create_pseudo_p ())
{
if (!REG_P (operands[0]))
operands[1] = force_reg (mode, operands[1]);
else if ((TARGET_NEON || TARGET_HAVE_MVE || TARGET_HAVE_MVE_FLOAT)
&& (CONSTANT_P (operands[1])))
{
operands[1] = neon_make_constant (operands[1]);
gcc_assert (operands[1] != NULL_RTX);
}
}
})
(define_expand "mov"
[(set (match_operand:VNINOTM1 0 "nonimmediate_operand")
(match_operand:VNINOTM1 1 "general_operand"))]
"TARGET_NEON
|| (TARGET_REALLY_IWMMXT && VALID_IWMMXT_REG_MODE (mode))"
{
gcc_checking_assert (aligned_operand (operands[0], mode));
gcc_checking_assert (aligned_operand (operands[1], mode));
if (can_create_pseudo_p ())
{
if (!REG_P (operands[0]))
operands[1] = force_reg (mode, operands[1]);
else if (TARGET_NEON && CONSTANT_P (operands[1]))
{
operands[1] = neon_make_constant (operands[1]);
gcc_assert (operands[1] != NULL_RTX);
}
}
})
(define_expand "movv8hf"
[(set (match_operand:V8HF 0 "s_register_operand")
(match_operand:V8HF 1 "s_register_operand"))]
"TARGET_NEON || TARGET_HAVE_MVE_FLOAT"
{
gcc_checking_assert (aligned_operand (operands[0], E_V8HFmode));
gcc_checking_assert (aligned_operand (operands[1], E_V8HFmode));
if (can_create_pseudo_p ())
{
if (!REG_P (operands[0]))
operands[1] = force_reg (E_V8HFmode, operands[1]);
else if (TARGET_HAVE_MVE_FLOAT && CONSTANT_P (operands[1]))
{
operands[1] = neon_make_constant (operands[1]);
gcc_assert (operands[1] != NULL_RTX);
}
}
})
;; Vector arithmetic. Expanders are blank, then unnamed insns implement
;; patterns separately for Neon, IWMMXT and MVE.
(define_expand "add3"
[(set (match_operand:VDQ 0 "s_register_operand")
(plus:VDQ (match_operand:VDQ 1 "s_register_operand")
(match_operand:VDQ 2 "s_register_operand")))]
"ARM_HAVE__ARITH"
)
(define_expand "sub3"
[(set (match_operand:VDQ 0 "s_register_operand")
(minus:VDQ (match_operand:VDQ 1 "s_register_operand")
(match_operand:VDQ 2 "s_register_operand")))]
"ARM_HAVE__ARITH"
)
(define_expand "mul3"
[(set (match_operand:VDQWH 0 "s_register_operand")
(mult:VDQWH (match_operand:VDQWH 1 "s_register_operand")
(match_operand:VDQWH 2 "s_register_operand")))]
"ARM_HAVE__ARITH
&& (!TARGET_REALLY_IWMMXT
|| mode == V4HImode
|| mode == V2SImode)"
)
(define_expand "smin3"
[(set (match_operand:VALLW 0 "s_register_operand")
(smin:VALLW (match_operand:VALLW 1 "s_register_operand")
(match_operand:VALLW 2 "s_register_operand")))]
"ARM_HAVE__ARITH"
)
(define_expand "umin3"
[(set (match_operand:VINTW 0 "s_register_operand")
(umin:VINTW (match_operand:VINTW 1 "s_register_operand")
(match_operand:VINTW 2 "s_register_operand")))]
"ARM_HAVE__ARITH"
)
(define_expand "smax3"
[(set (match_operand:VALLW 0 "s_register_operand")
(smax:VALLW (match_operand:VALLW 1 "s_register_operand")
(match_operand:VALLW 2 "s_register_operand")))]
"ARM_HAVE__ARITH"
)
(define_expand "umax3"
[(set (match_operand:VINTW 0 "s_register_operand")
(umax:VINTW (match_operand:VINTW 1 "s_register_operand")
(match_operand:VINTW 2 "s_register_operand")))]
"ARM_HAVE__ARITH"
)
(define_expand "vec_perm"
[(match_operand:VE 0 "s_register_operand")
(match_operand:VE 1 "s_register_operand")
(match_operand:VE 2 "s_register_operand")
(match_operand:VE 3 "s_register_operand")]
"TARGET_NEON && !BYTES_BIG_ENDIAN"
{
arm_expand_vec_perm (operands[0], operands[1], operands[2], operands[3]);
DONE;
})
(define_expand "vec_extract"
[(match_operand: 0 "nonimmediate_operand")
(match_operand:VQX_NOBF 1 "s_register_operand")
(match_operand:SI 2 "immediate_operand")]
"TARGET_NEON || TARGET_HAVE_MVE"
{
if (TARGET_NEON)
emit_insn (gen_neon_vec_extract (operands[0], operands[1],
operands[2]));
else if (TARGET_HAVE_MVE)
emit_insn (gen_mve_vec_extract (operands[0], operands[1],
operands[2]));
else
gcc_unreachable ();
DONE;
})
(define_expand "vec_set"
[(match_operand:VQX_NOBF 0 "s_register_operand" "")
(match_operand: 1 "s_register_operand" "")
(match_operand:SI 2 "immediate_operand" "")]
"TARGET_NEON || TARGET_HAVE_MVE"
{
HOST_WIDE_INT elem = HOST_WIDE_INT_1 << INTVAL (operands[2]);
if (TARGET_NEON)
emit_insn (gen_vec_set_internal (operands[0], operands[1],
GEN_INT (elem), operands[0]));
else
emit_insn (gen_mve_vec_set_internal (operands[0], operands[1],
GEN_INT (elem), operands[0]));
DONE;
})
(define_expand "and3"
[(set (match_operand:VDQ 0 "s_register_operand" "")
(and:VDQ (match_operand:VDQ 1 "s_register_operand" "")
(match_operand:VDQ 2 "neon_inv_logic_op2" "")))]
"ARM_HAVE__ARITH"
)
(define_expand "ior3"
[(set (match_operand:VDQ 0 "s_register_operand" "")
(ior:VDQ (match_operand:VDQ 1 "s_register_operand" "")
(match_operand:VDQ 2 "neon_logic_op2" "")))]
"ARM_HAVE__ARITH"
)
(define_expand "xor3"
[(set (match_operand:VDQ 0 "s_register_operand" "")
(xor:VDQ (match_operand:VDQ 1 "s_register_operand" "")
(match_operand:VDQ 2 "s_register_operand" "")))]
"ARM_HAVE__ARITH"
)
(define_expand "one_cmpl2"
[(set (match_operand:VDQ 0 "s_register_operand")
(not:VDQ (match_operand:VDQ 1 "s_register_operand")))]
"ARM_HAVE__ARITH && !TARGET_REALLY_IWMMXT"
)
(define_expand "2"
[(set (match_operand:VDQWH 0 "s_register_operand" "")
(ABSNEG:VDQWH (match_operand:VDQWH 1 "s_register_operand" "")))]
"ARM_HAVE__ARITH && !TARGET_REALLY_IWMMXT"
)
(define_expand "cadd3"
[(set (match_operand:VF 0 "register_operand")
(unspec:VF [(match_operand:VF 1 "register_operand")
(match_operand:VF 2 "register_operand")]
VCADD))]
"(TARGET_COMPLEX || (TARGET_HAVE_MVE && TARGET_HAVE_MVE_FLOAT
&& ARM_HAVE__ARITH)) && !BYTES_BIG_ENDIAN"
)
;; The complex mul operations always need to expand to two instructions.
;; The first operation does half the computation and the second does the
;; remainder. Because of this, expand early.
(define_expand "cmul3"
[(set (match_operand:VQ_HSF 0 "register_operand")
(unspec:VQ_HSF [(match_operand:VQ_HSF 1 "register_operand")
(match_operand:VQ_HSF 2 "register_operand")]
VCMUL_OP))]
"(TARGET_COMPLEX || (TARGET_HAVE_MVE && TARGET_HAVE_MVE_FLOAT))
&& !BYTES_BIG_ENDIAN"
{
rtx res1 = gen_reg_rtx (mode);
if (TARGET_COMPLEX)
{
rtx tmp = force_reg (mode, CONST0_RTX (mode));
emit_insn (gen_arm_vcmla (res1, tmp,
operands[2], operands[1]));
}
else
emit_insn (gen_arm_vcmla (res1, CONST0_RTX (mode),
operands[2], operands[1]));
emit_insn (gen_arm_vcmla (operands[0], res1,
operands[2], operands[1]));
DONE;
})
(define_expand "arm_vcmla"
[(set (match_operand:VF 0 "register_operand")
(plus:VF (match_operand:VF 1 "register_operand")
(unspec:VF [(match_operand:VF 2 "register_operand")
(match_operand:VF 3 "register_operand")]
VCMLA)))]
"(TARGET_COMPLEX || (TARGET_HAVE_MVE && TARGET_HAVE_MVE_FLOAT
&& ARM_HAVE__ARITH)) && !BYTES_BIG_ENDIAN"
)
;; The complex mla/mls operations always need to expand to two instructions.
;; The first operation does half the computation and the second does the
;; remainder. Because of this, expand early.
(define_expand "cml4"
[(set (match_operand:VF 0 "register_operand")
(plus:VF (unspec:VF [(match_operand:VF 1 "register_operand")
(match_operand:VF 2 "register_operand")]
VCMLA_OP)
(match_operand:VF 3 "register_operand")))]
"(TARGET_COMPLEX || (TARGET_HAVE_MVE && TARGET_HAVE_MVE_FLOAT
&& ARM_HAVE__ARITH)) && !BYTES_BIG_ENDIAN"
{
rtx tmp = gen_reg_rtx (mode);
emit_insn (gen_arm_vcmla (tmp, operands[3],
operands[2], operands[1]));
emit_insn (gen_arm_vcmla (operands[0], tmp,
operands[2], operands[1]));
DONE;
})
(define_expand "@movmisalign"
[(set (match_operand:VDQ 0 "nonimmediate_operand")
(unspec:VDQ [(match_operand:VDQ 1 "general_operand")]
UNSPEC_MISALIGNED_ACCESS))]
"ARM_HAVE__LDST && !BYTES_BIG_ENDIAN
&& unaligned_access && !TARGET_REALLY_IWMMXT"
{
rtx *memloc;
bool for_store = false;
/* This pattern is not permitted to fail during expansion: if both arguments
are non-registers (e.g. memory := constant, which can be created by the
auto-vectorizer), force operand 1 into a register. */
if (!s_register_operand (operands[0], mode)
&& !s_register_operand (operands[1], mode))
operands[1] = force_reg (mode, operands[1]);
if (s_register_operand (operands[0], mode))
memloc = &operands[1];
else
{
memloc = &operands[0];
for_store = true;
}
/* For MVE, vector loads/stores must be aligned to the element size. If the
alignment is less than that convert the load/store to a suitable mode. */
if (TARGET_HAVE_MVE
&& (MEM_ALIGN (*memloc)
< GET_MODE_ALIGNMENT (GET_MODE_INNER (mode))))
{
scalar_mode new_smode;
switch (MEM_ALIGN (*memloc))
{
case 64:
case 32:
new_smode = SImode;
break;
case 16:
new_smode = HImode;
break;
default:
new_smode = QImode;
break;
}
machine_mode new_mode
= mode_for_vector (new_smode,
GET_MODE_SIZE (mode)
/ GET_MODE_SIZE (new_smode)).require ();
rtx new_mem = adjust_address (*memloc, new_mode, 0);
if (!for_store)
{
rtx reg = gen_reg_rtx (new_mode);
emit_insn (gen_movmisalign (new_mode, reg, new_mem));
emit_move_insn (operands[0], gen_lowpart (mode, reg));
DONE;
}
emit_insn (gen_movmisalign (new_mode, new_mem,
gen_lowpart (new_mode, operands[1])));
DONE;
}
/* Legitimize address. */
if ((TARGET_HAVE_MVE
&& !mve_vector_mem_operand (mode, XEXP (*memloc, 0), false))
|| (!TARGET_HAVE_MVE
&& !neon_vector_mem_operand (*memloc, 2, false)))
{
rtx new_mem
= replace_equiv_address (*memloc,
force_reg (Pmode, XEXP (*memloc, 0)),
false);
gcc_assert (MEM_ALIGN (new_mem) == MEM_ALIGN (*memloc));
*memloc = new_mem;
}
})
(define_insn "mve_vshlq_"
[(set (match_operand:VDQIW 0 "s_register_operand" "=w,w")
(unspec:VDQIW [(match_operand:VDQIW 1 "s_register_operand" "w,w")
(match_operand:VDQIW 2 "imm_lshift_or_reg_neon" "w,Ds")]
VSHLQ))]
"ARM_HAVE__ARITH && !TARGET_REALLY_IWMMXT"
"@
vshl.%#\t%0, %1, %2
* return neon_output_shift_immediate (\"vshl\", 'i', &operands[2], mode, VALID_NEON_QREG_MODE (mode), true);"
[(set_attr "type" "neon_shift_reg, neon_shift_imm")]
)
(define_expand "vashl3"
[(set (match_operand:VDQIW 0 "s_register_operand" "")
(ashift:VDQIW (match_operand:VDQIW 1 "s_register_operand" "")
(match_operand:VDQIW 2 "imm_lshift_or_reg_neon" "")))]
"ARM_HAVE__ARITH && !TARGET_REALLY_IWMMXT"
{
emit_insn (gen_mve_vshlq_u (operands[0], operands[1], operands[2]));
DONE;
})
;; When operand 2 is an immediate, use the normal expansion to match
;; gen_vashr3_imm for Neon and gen_mve_vshrq_n_s_imm for
;; MVE.
(define_expand "vashr3"
[(set (match_operand:VDQIW 0 "s_register_operand")
(ashiftrt:VDQIW (match_operand:VDQIW 1 "s_register_operand")
(match_operand:VDQIW 2 "imm_rshift_or_reg_neon")))]
"ARM_HAVE__ARITH && !TARGET_REALLY_IWMMXT"
{
if (s_register_operand (operands[2], mode))
{
rtx neg = gen_reg_rtx (mode);
emit_insn (gen_neg2 (neg, operands[2]));
emit_insn (gen_mve_vshlq_s (operands[0], operands[1], neg));
DONE;
}
})
;; When operand 2 is an immediate, use the normal expansion to match
;; gen_vashr3_imm for Neon and gen_mve_vshrq_n_u_imm for
;; MVE.
(define_expand "vlshr3"
[(set (match_operand:VDQIW 0 "s_register_operand")
(lshiftrt:VDQIW (match_operand:VDQIW 1 "s_register_operand")
(match_operand:VDQIW 2 "imm_rshift_or_reg_neon")))]
"ARM_HAVE__ARITH && !TARGET_REALLY_IWMMXT"
{
if (s_register_operand (operands[2], mode))
{
rtx neg = gen_reg_rtx (mode);
emit_insn (gen_neg2 (neg, operands[2]));
emit_insn (gen_mve_vshlq_u (operands[0], operands[1], neg));
DONE;
}
})
;; Conditional instructions. These are comparisons with conditional moves for
;; vectors. They perform the assignment:
;;
;; Vop0 = (Vop4 Vop5) ? Vop1 : Vop2;
;;
;; where op3 is <, <=, ==, !=, >= or >. Operations are performed
;; element-wise.
(define_expand "vcond"
[(set (match_operand:VDQWH 0 "s_register_operand")
(if_then_else:VDQWH
(match_operator 3 "comparison_operator"
[(match_operand:VDQWH 4 "s_register_operand")
(match_operand:VDQWH 5 "reg_or_zero_operand")])
(match_operand:VDQWH 1 "s_register_operand")
(match_operand:VDQWH 2 "s_register_operand")))]
"ARM_HAVE__ARITH
&& !TARGET_REALLY_IWMMXT
&& (! || flag_unsafe_math_optimizations)"
{
arm_expand_vcond (operands, mode);
DONE;
})
(define_expand "vcond"
[(set (match_operand: 0 "s_register_operand")
(if_then_else:
(match_operator 3 "comparison_operator"
[(match_operand:V32 4 "s_register_operand")
(match_operand:V32 5 "reg_or_zero_operand")])
(match_operand: 1 "s_register_operand")
(match_operand: 2 "s_register_operand")))]
"ARM_HAVE__ARITH
&& !TARGET_REALLY_IWMMXT
&& (! || flag_unsafe_math_optimizations)"
{
arm_expand_vcond (operands, mode);
DONE;
})
(define_expand "vcond"
[(set (match_operand: 0 "s_register_operand")
(if_then_else:
(match_operator 3 "comparison_operator"
[(match_operand:V16 4 "s_register_operand")
(match_operand:V16 5 "reg_or_zero_operand")])
(match_operand: 1 "s_register_operand")
(match_operand: 2 "s_register_operand")))]
"ARM_HAVE__ARITH
&& !TARGET_REALLY_IWMMXT
&& (! || flag_unsafe_math_optimizations)"
{
arm_expand_vcond (operands, mode);
DONE;
})
(define_expand "vcondu"
[(set (match_operand:VDQW 0 "s_register_operand")
(if_then_else:VDQW
(match_operator 3 "arm_comparison_operator"
[(match_operand: 4 "s_register_operand")
(match_operand: 5 "reg_or_zero_operand")])
(match_operand:VDQW 1 "s_register_operand")
(match_operand:VDQW 2 "s_register_operand")))]
"ARM_HAVE__ARITH
&& !TARGET_REALLY_IWMMXT"
{
arm_expand_vcond (operands, mode);
DONE;
})
(define_expand "vec_load_lanesoi"
[(set (match_operand:OI 0 "s_register_operand")
(unspec:OI [(match_operand:OI 1 "neon_struct_operand")
(unspec:VQ2 [(const_int 0)] UNSPEC_VSTRUCTDUMMY)]
UNSPEC_VLD2))]
"TARGET_NEON || TARGET_HAVE_MVE"
{
if (TARGET_NEON)
emit_insn (gen_neon_vld2 (operands[0], operands[1]));
else
emit_insn (gen_mve_vld2q (operands[0], operands[1]));
DONE;
})
(define_expand "vec_store_lanesoi"
[(set (match_operand:OI 0 "neon_struct_operand")
(unspec:OI [(match_operand:OI 1 "s_register_operand")
(unspec:VQ2 [(const_int 0)] UNSPEC_VSTRUCTDUMMY)]
UNSPEC_VST2))]
"TARGET_NEON || TARGET_HAVE_MVE"
{
if (TARGET_NEON)
emit_insn (gen_neon_vst2 (operands[0], operands[1]));
else
emit_insn (gen_mve_vst2q (operands[0], operands[1]));
DONE;
})
(define_expand "vec_load_lanesxi"
[(match_operand:XI 0 "s_register_operand")
(match_operand:XI 1 "neon_struct_operand")
(unspec:VQ2 [(const_int 0)] UNSPEC_VSTRUCTDUMMY)]
"TARGET_NEON || TARGET_HAVE_MVE"
{
if (TARGET_NEON)
emit_insn (gen_neon_vld4 (operands[0], operands[1]));
else
emit_insn (gen_mve_vld4q (operands[0], operands[1]));
DONE;
})
(define_expand "vec_store_lanesxi"
[(match_operand:XI 0 "neon_struct_operand")
(match_operand:XI 1 "s_register_operand")
(unspec:VQ2 [(const_int 0)] UNSPEC_VSTRUCTDUMMY)]
"TARGET_NEON || TARGET_HAVE_MVE"
{
if (TARGET_NEON)
emit_insn (gen_neon_vst4 (operands[0], operands[1]));
else
emit_insn (gen_mve_vst4q (operands[0], operands[1]));
DONE;
})
(define_expand "reduc_plus_scal_"
[(match_operand: 0 "nonimmediate_operand")
(match_operand:VQ 1 "s_register_operand")]
"ARM_HAVE__ARITH
&& !(TARGET_HAVE_MVE && FLOAT_MODE_P (mode))
&& !BYTES_BIG_ENDIAN"
{
if (TARGET_NEON)
{
rtx step1 = gen_reg_rtx (mode);
emit_insn (gen_quad_halves_plus (step1, operands[1]));
emit_insn (gen_reduc_plus_scal_ (operands[0], step1));
}
else
{
/* vaddv generates a 32 bits accumulator. */
rtx op0 = gen_reg_rtx (SImode);
emit_insn (gen_mve_vaddvq (VADDVQ_S, mode, op0, operands[1]));
emit_move_insn (operands[0], gen_lowpart (mode, op0));
}
DONE;
})
(define_expand "avg3_floor"
[(match_operand:MVE_2 0 "s_register_operand")
(match_operand:MVE_2 1 "s_register_operand")
(match_operand:MVE_2 2 "s_register_operand")]
"ARM_HAVE__ARITH"
{
if (TARGET_HAVE_MVE)
emit_insn (gen_mve_vhaddq (VHADDQ_S, mode,
operands[0], operands[1], operands[2]));
else
emit_insn (gen_neon_vhadd (UNSPEC_VHADD_S, UNSPEC_VHADD_S, mode,
operands[0], operands[1], operands[2]));
DONE;
})
(define_expand "uavg3_floor"
[(match_operand:MVE_2 0 "s_register_operand")
(match_operand:MVE_2 1 "s_register_operand")
(match_operand:MVE_2 2 "s_register_operand")]
"ARM_HAVE__ARITH"
{
if (TARGET_HAVE_MVE)
emit_insn (gen_mve_vhaddq (VHADDQ_U, mode,
operands[0], operands[1], operands[2]));
else
emit_insn (gen_neon_vhadd (UNSPEC_VHADD_U, UNSPEC_VHADD_U, mode,
operands[0], operands[1], operands[2]));
DONE;
})
(define_expand "avg3_ceil"
[(match_operand:MVE_2 0 "s_register_operand")
(match_operand:MVE_2 1 "s_register_operand")
(match_operand:MVE_2 2 "s_register_operand")]
"ARM_HAVE__ARITH"
{
if (TARGET_HAVE_MVE)
emit_insn (gen_mve_vrhaddq (VRHADDQ_S, mode,
operands[0], operands[1], operands[2]));
else
emit_insn (gen_neon_vhadd (UNSPEC_VRHADD_S, UNSPEC_VRHADD_S, mode,
operands[0], operands[1], operands[2]));
DONE;
})
(define_expand "uavg3_ceil"
[(match_operand:MVE_2 0 "s_register_operand")
(match_operand:MVE_2 1 "s_register_operand")
(match_operand:MVE_2 2 "s_register_operand")]
"ARM_HAVE__ARITH"
{
if (TARGET_HAVE_MVE)
emit_insn (gen_mve_vrhaddq (VRHADDQ_U, mode,
operands[0], operands[1], operands[2]));
else
emit_insn (gen_neon_vhadd (UNSPEC_VRHADD_U, UNSPEC_VRHADD_U, mode,
operands[0], operands[1], operands[2]));
DONE;
})
(define_expand "clz2"
[(set (match_operand:VDQIW 0 "s_register_operand")
(clz:VDQIW (match_operand:VDQIW 1 "s_register_operand")))]
"ARM_HAVE__ARITH
&& !TARGET_REALLY_IWMMXT"
)
(define_expand "vec_init"
[(match_operand:VDQX 0 "s_register_operand")
(match_operand 1 "" "")]
"TARGET_NEON || (TARGET_HAVE_MVE && VALID_MVE_MODE (mode))"
{
neon_expand_vector_init (operands[0], operands[1]);
DONE;
})