;; Machine description of the Adaptiva epiphany cpu for GNU C compiler
;; Copyright (C) 1994-2022 Free Software Foundation, Inc.
;; Contributed by Embecosm on behalf of Adapteva, Inc.
;; This file is part of GCC.
;; GCC is free software; you can redistribute it and/or modify
;; it under the terms of the GNU General Public License as published by
;; the Free Software Foundation; either version 3, or (at your option)
;; any later version.
;; GCC is distributed in the hope that it will be useful,
;; but WITHOUT ANY WARRANTY; without even the implied warranty of
;; MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
;; GNU General Public License for more details.
;; You should have received a copy of the GNU General Public License
;; along with GCC; see the file COPYING3. If not see
;; .
;; See file "rtl.def" for documentation on define_insn, match_*, et. al.
(define_constants
[(GPR_0 0)
(GPR_1 1)
(GPR_FP 11)
(GPR_IP 12)
(GPR_SP 13)
(GPR_LR 14)
(GPR_16 16)
(GPR_18 18)
(GPR_20 20)
(ARG_POINTER_REGNUM 64)
(FRAME_POINTER_REGNUM 65)
(CC_REGNUM 66) ;; 66 or 17
(CCFP_REGNUM 67) ;; 67 or 18
(CONFIG_REGNUM 68)
(STATUS_REGNUM 69)
(LC_REGNUM 70)
(LS_REGNUM 71)
(LE_REGNUM 72)
(IRET_REGNUM 73)
(FP_NEAREST_REGNUM 74)
(FP_TRUNCATE_REGNUM 75)
(FP_ANYFP_REGNUM 76)
(UNKNOWN_REGNUM 77) ; used for addsi3_r and friends
; We represent the return address as an unspec rather than a reg.
; If we used a reg, we could use register elimination, but eliminating
; to GPR_LR would make the latter visible to dataflow, thus making it
; harder to determine when it must be saved.
(UNSPEC_RETURN_ADDR 0)
(UNSPEC_FP_MODE 1)
(UNSPECV_GID 0)
(UNSPECV_GIE 1)])
;; Insn type. Used to default other attribute values.
(define_attr "type"
"move,load,store,cmove,unary,compare,shift,mul,uncond_branch,branch,call,fp,fp_int,v2fp,misc,sfunc,fp_sfunc,flow"
(const_string "misc"))
;; Length (in # bytes)
(define_attr "length" "" (const_int 4))
;; The length here is the length of a single asm.
(define_asm_attributes
[(set_attr "length" "4")
(set_attr "type" "misc")])
;; pipeline model; so far we have only one.
(define_attr "pipe_model" "epiphany" (const_string "epiphany"))
(define_attr "rounding" "trunc,nearest"
(cond [(ne (symbol_ref "TARGET_ROUND_NEAREST") (const_int 0))
(const_string "nearest")]
(const_string "trunc")))
(define_attr "fp_mode" "round_unknown,round_nearest,round_trunc,int,caller,none"
(cond [(eq_attr "type" "fp,v2fp,fp_sfunc")
(symbol_ref "(enum attr_fp_mode) epiphany_normal_fp_rounding")
(eq_attr "type" "call")
(symbol_ref "(enum attr_fp_mode) epiphany_normal_fp_mode")
(eq_attr "type" "fp_int")
(const_string "int")]
(const_string "none")))
(include "epiphany-sched.md")
(include "predicates.md")
(include "constraints.md")
;; modes that are held in a single register, and hence, a word.
(define_mode_iterator WMODE [SI SF HI QI V2HI V4QI])
(define_mode_iterator WMODE2 [SI SF HI QI V2HI V4QI])
;; modes that are held in a two single registers
(define_mode_iterator DWMODE [DI DF V2SI V2SF V4HI V8QI])
;; Double-word mode made up of two single-word mode values.
(define_mode_iterator DWV2MODE [V2SI V2SF])
(define_mode_attr vmode_part [(V2SI "si") (V2SF "sf")])
(define_mode_attr vmode_PART [(V2SI "SI") (V2SF "SF")])
(define_mode_attr vmode_fp_type [(V2SI "fp_int") (V2SF "fp")])
(define_mode_attr vmode_ccmode [(V2SI "CC") (V2SF "CC_FP")])
(define_mode_attr vmode_cc [(V2SI "CC_REGNUM") (V2SF "CCFP_REGNUM")])
;; Move instructions.
(define_expand "mov"
[(set (match_operand:WMODE 0 "general_operand" "")
(match_operand:WMODE 1 "general_operand" ""))]
""
{
if (mode == V4QImode || mode == V2HImode)
{
operands[0] = simplify_gen_subreg (SImode, operands[0], mode, 0);
operands[1] = simplify_gen_subreg (SImode, operands[1], mode, 0);
emit_insn (gen_movsi (operands[0], operands[1]));
DONE;
}
if (GET_CODE (operands[0]) == MEM)
operands[1] = force_reg (mode, operands[1]);
if (mode == SImode
&& (operands[1] == frame_pointer_rtx || operands[1] == arg_pointer_rtx))
{
rtx reg = operands[0];
if (!REG_P (reg))
reg = gen_reg_rtx (SImode);
emit_insn (gen_move_frame (reg, operands[1]));
operands[1] = reg;
if (operands[0] == reg)
DONE;
}
})
(define_insn "*movqi_insn"
[(set (match_operand:QI 0 "move_dest_operand" "=Rcs, r, r,r,m")
(match_operand:QI 1 "move_src_operand" "Rcs,rU16,Cal,m,r"))]
;; ??? Needed?
"gpr_operand (operands[0], QImode)
|| gpr_operand (operands[1], QImode)"
"@
mov %0,%1
mov %0,%1
mov %0,%1
ldrb %0,%1
strb %1,%0"
[(set_attr "type" "move,move,move,load,store")])
(define_insn_and_split "*movhi_insn"
[(set (match_operand:HI 0 "move_dest_operand" "=r, r,r,m")
(match_operand:HI 1 "move_src_operand""rU16,Cal,m,r"))]
"gpr_operand (operands[0], HImode)
|| gpr_operand (operands[1], HImode)"
"@
mov %0,%1
mov %0,%%low(%1); %1
ldrh %0,%c1
strh %1,%c0"
"reload_completed && CONSTANT_P (operands[1])
&& !satisfies_constraint_U16 (operands[1]) && TARGET_SPLIT_LOHI"
[(set (match_dup 2) (match_dup 3))]
"operands[2] = simplify_gen_subreg (SImode, operands[0], HImode, 0);
operands[3] = simplify_gen_subreg (SImode, operands[1], HImode, 0);"
[(set_attr "type" "move,move,load,store")])
;; We use a special pattern for a move from the frame pointer to
;; show the flag clobber that is needed when this move is changed
;; to an add by register elimination.
;; ??? A pseudo register might be equivalent to a function invariant,
;; and thus placed by reload into reg_equiv_invariant; if the pseudo
;; does not get a hard register, we then end up with the function
;; invariant in its place, i.e. an unexpected clobber of the flags
;; register.
;;
;; N.B. operand 1 is an operand so that reload will perform elimination.
;;
;; The post-reload pattern recognition and splitting is done in frame_move_1.
(define_insn "move_frame"
[(set (match_operand:SI 0 "gpr_operand" "=r")
(match_operand:SI 1 "register_operand" "r"))
(clobber (reg:CC CC_REGNUM))]
"operands[1] == frame_pointer_rtx || operands[1] == arg_pointer_rtx"
"#")
(define_insn "movsi_high"
[(set (match_operand:SI 0 "gpr_operand" "+r")
(ior:SI (and:SI (match_dup 0) (const_int 65535))
(high:SI (match_operand:SI 1 "move_src_operand" "i"))))]
""
"movt %0, %%high(%1)"
[(set_attr "type" "move")
(set_attr "length" "4")])
(define_insn "movsi_lo_sum"
[(set (match_operand:SI 0 "gpr_operand" "=r")
(lo_sum:SI (const_int 0)
(match_operand:SI 1 "move_src_operand" "i")))]
""
"mov %0, %%low(%1)"
[(set_attr "type" "move")
(set_attr "length" "4")])
(define_insn_and_split "*movsi_insn"
[(set (match_operand:SI 0 "move_dest_operand"
"= r, r, r, r, r, r, m, r, Rct")
(match_operand:SI 1 "move_src_operand"
"rU16Rra,Cm1,Cl1,Cr1,Cal,mSra,rRra,Rct,r"))]
"gpr_operand (operands[0], SImode)
|| gpr_operand (operands[1], SImode)
|| satisfies_constraint_Sra (operands[1])"
{
switch (which_alternative)
{
case 0: return "mov %0,%1";
case 1: return "add %0,%-,(1+%1)";
case 2: operands[1] = GEN_INT (exact_log2 (-INTVAL (operands[1])));
return "lsl %0,%-,%1";
case 3: operands[1] = GEN_INT (32 - exact_log2 (INTVAL (operands[1]) + 1));
return "lsr %0,%-,%1";
case 4: return "mov %0,%%low(%1)\;movt %0,%%high(%1) ; %1";
case 5: return "ldr %0,%C1";
case 6: return "str %1,%C0";
case 7: return "movfs %0,%1";
case 8: return "movts %0,%1";
default: gcc_unreachable ();
}
}
"reload_completed && CONSTANT_P (operands[1])
&& !satisfies_constraint_U16 (operands[1])
&& !satisfies_constraint_Cm1 (operands[1])
&& !satisfies_constraint_Cl1 (operands[1])
&& !satisfies_constraint_Cr1 (operands[1])
&& TARGET_SPLIT_LOHI"
[(match_dup 2) (match_dup 3)]
"operands[2] = gen_movsi_lo_sum (operands[0], operands[1]);
operands[3] = gen_movsi_high (operands[0], operands[1]);"
[(set_attr "type" "move,misc,misc,misc,move,load,store,flow,flow")
(set_attr "length" "4,4,4,4,8,4,4,4,4")])
(define_split
[(set (match_operand:SI 0 "nonimmediate_operand")
(unspec:SI [(const_int 0)] UNSPEC_RETURN_ADDR))]
"reload_completed && !MACHINE_FUNCTION (cfun)->lr_clobbered"
[(set (match_dup 0) (reg:SI GPR_LR))])
(define_split
[(set (match_operand:SI 0 "gpr_operand")
(unspec:SI [(const_int 0)] UNSPEC_RETURN_ADDR))]
"reload_completed"
[(set (match_dup 0) (match_dup 1))]
{
emit_insn (gen_reload_insi_ra (operands[0], operands[1]));
DONE;
})
(define_expand "reload_insi_ra"
[(set (match_operand:SI 0 "gpr_operand" "r") (match_operand:SI 1 "" "Sra"))]
""
{
rtx addr
= (frame_pointer_needed ? hard_frame_pointer_rtx : stack_pointer_rtx);
if (!MACHINE_FUNCTION (cfun)->lr_slot_known)
{
start_sequence ();
epiphany_expand_prologue ();
if (!MACHINE_FUNCTION (cfun)->lr_slot_known)
epiphany_expand_epilogue (0);
end_sequence ();
gcc_assert (MACHINE_FUNCTION (cfun)->lr_slot_known);
}
addr = plus_constant (Pmode, addr, MACHINE_FUNCTION (cfun)->lr_slot_offset);
operands[1] = gen_frame_mem (SImode, addr);
})
;; If the frame pointer elimination offset is zero, we'll use this pattern.
;; Note that the splitter can accept any gpr in operands[1]; this is
;; necessary, (e.g. for compile/20021015-1.c -O0,)
;; because when register elimination cannot be done with the constant
;; as an immediate operand of the add instruction, reload will resort to
;; loading the constant into a reload register, using gen_add2_insn to add
;; the stack pointer, and then use the reload register as new source in
;; the move_frame pattern.
(define_insn_and_split "*move_frame_1"
[(set (match_operand:SI 0 "gpr_operand" "=r")
(match_operand:SI 1 "gpr_operand" "r"))
(clobber (reg:CC CC_REGNUM))]
"(reload_in_progress || reload_completed)
&& (operands[1] == stack_pointer_rtx
|| operands[1] == hard_frame_pointer_rtx)"
"#"
"reload_in_progress || reload_completed"
[(set (match_dup 0) (match_dup 1))])
(define_expand "mov"
[(set (match_operand:DWMODE 0 "general_operand" "")
(match_operand:DWMODE 1 "general_operand" ""))]
""
"
{
if (GET_MODE_CLASS (mode) == MODE_VECTOR_INT
|| GET_MODE_CLASS (mode) == MODE_VECTOR_FLOAT)
{
if (epiphany_vect_align == 4 && TARGET_SPLIT_VECMOVE_EARLY)
{
rtx o0l, o0h, o1l, o1h;
o0l = simplify_gen_subreg (SImode, operands[0], mode, 0);
o0h = simplify_gen_subreg (SImode, operands[0], mode,
UNITS_PER_WORD);
o1l = simplify_gen_subreg (SImode, operands[1], mode, 0);
o1h = simplify_gen_subreg (SImode, operands[1], mode,
UNITS_PER_WORD);
if (reg_overlap_mentioned_p (o0l, o1h))
{
emit_move_insn (o0h, o1h);
emit_move_insn (o0l, o1l);
}
else
{
emit_move_insn (o0l, o1l);
emit_move_insn (o0h, o1h);
}
DONE;
}
/* lower_subreg has a tendency to muck up vectorized code.
To protect the wide memory accesses, we must use same-size
subregs. */
if (epiphany_vect_align != 4 /* == 8 */
&& !reload_in_progress
&& (GET_CODE (operands[0]) == MEM || GET_CODE (operands[1]) == MEM)
&& !misaligned_operand (operands[1], mode)
&& (GET_CODE (operands[0]) != SUBREG
|| (GET_MODE_SIZE (GET_MODE (SUBREG_REG (operands[0])))
!= GET_MODE_SIZE (mode)
&& GET_CODE (operands[1]) != SUBREG)))
{
operands[0]
= simplify_gen_subreg (DImode, operands[0], mode, 0);
operands[1]
= simplify_gen_subreg (DImode, operands[1], mode, 0);
emit_insn (gen_movdi (operands[0], operands[1]));
DONE;
}
}
/* Everything except mem = const or mem = mem can be done easily. */
if (GET_CODE (operands[0]) == MEM)
operands[1] = force_reg (mode, operands[1]);
}")
(define_insn_and_split "*mov_insn"
[(set (match_operand:DWMODE 0 "move_dest_operand" "=r, r,r,m")
(match_operand:DWMODE 1 "move_double_src_operand" "r,CalE,m,r"))]
"(gpr_operand (operands[0], mode)
|| gpr_operand (operands[1], mode))"
"@
#
#
ldrd %0,%X1
strd %1,%X0"
"reload_completed
&& (((!MEM_P (operands[0]) || misaligned_operand (operands[0], mode))
&& (!MEM_P (operands[1])
|| misaligned_operand (operands[1], mode)))
|| epiphany_vect_align == 4)"
[(set (match_dup 2) (match_dup 3))
(set (match_dup 4) (match_dup 5))]
{
int word0 = 0, word1 = UNITS_PER_WORD;
if (post_modify_operand (operands[0], mode)
|| post_modify_operand (operands[1], mode))
word0 = UNITS_PER_WORD, word1 = 0;
operands[2] = simplify_gen_subreg (SImode, operands[0], mode, word0);
operands[3] = simplify_gen_subreg (SImode, operands[1], mode, word0);
operands[4] = simplify_gen_subreg (SImode, operands[0], mode, word1);
operands[5] = simplify_gen_subreg (SImode, operands[1], mode, word1);
if (post_modify_operand (operands[0], mode))
operands[2]
= change_address (operands[2], VOIDmode,
plus_constant (Pmode, XEXP (XEXP (operands[0], 0), 0),
UNITS_PER_WORD));
if (post_modify_operand (operands[1], mode))
operands[3]
= change_address (operands[3], VOIDmode,
plus_constant (Pmode, XEXP (XEXP (operands[1], 0), 0),
UNITS_PER_WORD));
}
[(set_attr "type" "move,move,load,store")
(set_attr "length" "8,16,4,4")])
(define_insn_and_split "*movsf_insn"
[(set (match_operand:SF 0 "move_dest_operand" "=r,r,r,m")
(match_operand:SF 1 "move_src_operand" "r,E,m,r"))]
"gpr_operand (operands[0], SFmode)
|| gpr_operand (operands[1], SFmode)"
"@
mov %0,%1
mov %0,%%low(%1)\;movt %0,%%high(%1) ; %1
ldr %0,%C1
str %1,%C0"
"reload_completed && CONSTANT_P (operands[1]) && TARGET_SPLIT_LOHI"
[(set (match_dup 2) (match_dup 3))]
"operands[2] = simplify_gen_subreg (SImode, operands[0], SFmode, 0);
operands[3] = simplify_gen_subreg (SImode, operands[1], SFmode, 0);"
[(set_attr "type" "move,move,load,store")
(set_attr "length" "4,8,4,4")])
(define_expand "addsi3"
[(set (match_operand:SI 0 "add_reg_operand" "")
(plus:SI (match_operand:SI 1 "add_reg_operand" "")
(match_operand:SI 2 "add_operand" "")))]
""
"
{
if (reload_in_progress || reload_completed)
emit_insn (gen_addsi3_r (operands[0], operands[1], operands[2]));
else if (TARGET_FP_IARITH && add_reg_operand (operands[2], SImode))
emit_insn (gen_iadd (operands[0], operands[1], operands[2]));
else
emit_insn (gen_addsi3_i (operands[0], operands[1], operands[2]));
DONE;
}")
; The default case of epiphany_print_operand emits IMMEDIATE_PREFIX
; where appropriate; however, 'n' is processed by output_asm_insn
; which doesn't, so we have to explicitly emit the '# in the
; r/r/CnL output template alternative.
(define_insn "addsi3_i"
[(set (match_operand:SI 0 "add_reg_operand" "=r,r")
(plus:SI (match_operand:SI 1 "add_reg_operand" "%r,r")
(match_operand:SI 2 "add_operand" "rL,CnL")))
(clobber (reg:CC CC_REGNUM))]
""
"@
add %0,%1,%2
sub %0,%1,#%n2"
[(set_attr "type" "misc")])
; We use a clobber of UNKNOWN_REGNUM here so that the peephole optimizers
; can identify the unresolved flags clobber problem, and also to
; avoid unwanted matches.
;
; At -O0 / -O1 we don't peephole all instances away. We could get better
; debug unwinding through the emitted code if we added a splitter.
(define_insn "addsi3_r"
[(set (match_operand:SI 0 "gpr_operand" "=r")
(plus:SI (match_operand:SI 1 "gpr_operand" "%r")
(match_operand:SI 2 "nonmemory_operand" "rCar")))
(clobber (reg:CC UNKNOWN_REGNUM))]
"reload_in_progress || reload_completed"
{
int scratch = (0x17
^ (true_regnum (operands[0]) & 1)
^ (true_regnum (operands[1]) & 2)
^ (true_regnum (operands[2]) & 4));
asm_fprintf (asm_out_file, "\tstr r%d,[sp,#0]\n", scratch);
asm_fprintf (asm_out_file, "\tmovfs r%d,status\n", scratch);
output_asm_insn ("add %0,%1,%2", operands);
asm_fprintf (asm_out_file, "\tmovts status,r%d\n", scratch);
asm_fprintf (asm_out_file, "\tldr r%d,[sp,#0]\n", scratch);
return "";
}
[(set_attr "length" "20")
(set_attr "type" "misc")])
;; reload uses gen_addsi2 because it doesn't understand the need for
;; the clobber.
(define_peephole2
[(set (match_operand:SI 0 "gpr_operand" "")
(match_operand:SI 1 "const_int_operand" ""))
(parallel [(set (match_dup 0)
(plus:SI (match_dup 0)
(match_operand:SI 2 "gpr_operand")))
(clobber (reg:CC UNKNOWN_REGNUM))])]
"satisfies_constraint_L (operands[1])
|| ((operands[2] == stack_pointer_rtx
|| (operands[2] == hard_frame_pointer_rtx && frame_pointer_needed))
&& !peep2_regno_dead_p (2, CC_REGNUM)
&& satisfies_constraint_Car (operands[1]))"
[(parallel [(set (match_dup 0)
(plus:SI (match_dup 2) (match_dup 1)))
(clobber (reg:CC UNKNOWN_REGNUM))])]
;; FIXME:
;; need this patch: http://gcc.gnu.org/ml/gcc-patches/2011-10/msg02819.html
;; "peep2_rescan = true;"
)
(define_peephole2
[(match_parallel 5 ""
[(set (match_operand 3 "cc_operand" "") (match_operand 4 "" ""))])
(parallel [(set (match_operand:SI 0 "gpr_operand" "")
(plus:SI (match_operand:SI 1 "gpr_operand" "")
(match_operand:SI 2 "nonmemory_operand" "")))
(clobber (reg:CC UNKNOWN_REGNUM))])]
"REGNO (operands[3]) == CC_REGNUM
&& (gpr_operand (operands[2], SImode)
|| satisfies_constraint_L (operands[2]))
&& !reg_overlap_mentioned_p (operands[0], operands[5])
&& !reg_set_p (operands[1], operands[5])
&& !reg_set_p (operands[2], operands[5])"
[(parallel [(set (match_operand:SI 0 "gpr_operand" "")
(plus:SI (match_operand:SI 1 "gpr_operand" "")
(match_operand:SI 2 "nonmemory_operand" "")))
(clobber (reg:CC CC_REGNUM))])
(match_dup 5)]
"")
(define_peephole2
[(parallel [(set (match_operand:SI 0 "gpr_operand" "")
(plus:SI (match_operand:SI 1 "gpr_operand" "")
(match_operand:SI 2 "nonmemory_operand" "")))
(clobber (reg:CC UNKNOWN_REGNUM))])]
"peep2_regno_dead_p (1, CC_REGNUM)
&& (gpr_operand (operands[2], SImode)
|| satisfies_constraint_L (operands[2]))"
[(parallel [(set (match_operand:SI 0 "gpr_operand" "")
(plus:SI (match_operand:SI 1 "gpr_operand" "")
(match_operand:SI 2 "nonmemory_operand" "")))
(clobber (reg:CC CC_REGNUM))])]
"")
(define_peephole2
[(parallel [(set (match_operand:SI 0 "gpr_operand" "")
(plus:SI (reg:SI GPR_SP)
(match_operand:SI 1 "nonmemory_operand" "")))
(clobber (reg:CC UNKNOWN_REGNUM))])]
"(REG_P (operands[1]) && !reg_overlap_mentioned_p (operands[0], operands[1]))
|| RTX_OK_FOR_OFFSET_P (mode, operands[1])"
[(set (match_dup 0) (reg:SI GPR_SP))
(set (mem:WMODE (post_modify (match_dup 0)
(plus:SI (match_dup 0) (match_dup 1))))
(reg:WMODE GPR_SP))]
"")
(define_peephole2
[(parallel [(set (match_operand:SI 0 "gpr_operand" "")
(plus:SI (reg:SI GPR_FP)
(match_operand:SI 1 "nonmemory_operand" "")))
(clobber (reg:CC UNKNOWN_REGNUM))])
(match_scratch:WMODE 2 "r")]
"frame_pointer_needed
&& ((REG_P (operands[1])
&& !reg_overlap_mentioned_p (operands[0], operands[1]))
|| RTX_OK_FOR_OFFSET_P (mode, operands[1]))"
[(set (match_dup 0) (reg:SI GPR_FP))
(set (match_dup 2)
(mem:WMODE (post_modify (match_dup 0)
(plus:SI (match_dup 0) (match_dup 1)))))]
"")
(define_expand "subsi3"
[(set (match_operand:SI 0 "gpr_operand" "")
(plus:SI (match_operand:SI 1 "add_reg_operand" "")
(match_operand:SI 2 "arith_operand" "")))]
""
"
{
gcc_assert (!reload_in_progress && !reload_completed);
if (TARGET_FP_IARITH)
emit_insn (gen_isub (operands[0], operands[1], operands[2]));
else
emit_insn (gen_subsi3_i (operands[0], operands[1], operands[2]));
DONE;
}")
(define_insn "subsi3_i"
[(set (match_operand:SI 0 "gpr_operand" "=r")
(minus:SI (match_operand:SI 1 "add_reg_operand" "r")
(match_operand:SI 2 "arith_operand" "rL")))
(clobber (reg:CC CC_REGNUM))]
""
"sub %0,%1,%2"
[(set_attr "type" "misc")])
; After mode-switching, floating point operations, fp_sfuncs and calls
; must exhibit the use of the control register, lest the setting of the
; control register could be deleted or moved. OTOH a use of a hard register
; greatly counfounds optimizers like the rtl loop optimizers or combine.
; Therefore, we put an extra pass immediately after the mode switching pass
; that inserts the USEs of the control registers, and sets a flag in struct
; machine_function that float_operation can henceforth only match with that
; USE.
;; Addition
(define_expand "addsf3"
[(parallel
[(set (match_operand:SF 0 "gpr_operand" "")
(plus:SF (match_operand:SF 1 "gpr_operand" "")
(match_operand:SF 2 "gpr_operand" "")))
(clobber (reg:CC_FP CCFP_REGNUM))])])
(define_insn "*addsf3_i"
[(match_parallel 3 "float_operation"
[(set (match_operand:SF 0 "gpr_operand" "=r")
(plus:SF (match_operand:SF 1 "gpr_operand" "%r")
(match_operand:SF 2 "gpr_operand" "r")))
(clobber (reg:CC_FP CCFP_REGNUM))])]
""
"fadd %0,%1,%2"
[(set_attr "type" "fp")])
;; Subtraction
(define_expand "subsf3"
[(parallel
[(set (match_operand:SF 0 "gpr_operand" "")
(minus:SF (match_operand:SF 1 "gpr_operand" "")
(match_operand:SF 2 "gpr_operand" "")))
(clobber (reg:CC_FP CCFP_REGNUM))])])
(define_insn "*subsf3_i"
[(match_parallel 3 "float_operation"
[(set (match_operand:SF 0 "gpr_operand" "=r")
(minus:SF (match_operand:SF 1 "gpr_operand" "r")
(match_operand:SF 2 "gpr_operand" "r")))
(clobber (reg:CC_FP CCFP_REGNUM))])]
""
"fsub %0,%1,%2"
[(set_attr "type" "fp")])
(define_expand "subsf3_f"
[(parallel
[(set (reg:CC_FP CCFP_REGNUM)
(compare:CC_FP (match_operand:SF 1 "gpr_operand" "r")
(match_operand:SF 2 "gpr_operand" "r")))
(set (match_operand:SF 0 "gpr_operand" "=r")
(minus:SF (match_dup 1) (match_dup 2)))])]
"!TARGET_SOFT_CMPSF")
(define_insn "*subsf3_f_i"
[(match_parallel 3 "float_operation"
[(set (reg:CC_FP CCFP_REGNUM)
(compare:CC_FP (match_operand:SF 1 "gpr_operand" "r")
(match_operand:SF 2 "gpr_operand" "r")))
(set (match_operand:SF 0 "gpr_operand" "=r")
(minus:SF (match_dup 1) (match_dup 2)))])]
"!TARGET_SOFT_CMPSF"
"fsub %0,%1,%2"
[(set_attr "type" "fp")])
; There is an fabs instruction, but it has longer latency.
(define_expand "abssf2"
[(set (match_operand:SF 0 "gpr_operand" "")
(abs:SF (match_operand:SF 1 "gpr_operand" "")))]
""
"
{
rtx op1 = copy_to_mode_reg (SImode, simplify_gen_subreg (SImode, operands[1],
SFmode, 0));
rtx op0 = simplify_gen_subreg (SImode, operands[0], SFmode, 0);
emit_insn (gen_ashlsi3 (op1, op1, const1_rtx));
emit_insn (gen_lshrsi3 (op0, op1, const1_rtx));
DONE;
}")
;; Multiplication
(define_expand "mulsf3"
[(parallel
[(set (match_operand:SF 0 "gpr_operand" "")
(mult:SF (match_operand:SF 1 "gpr_operand" "")
(match_operand:SF 2 "gpr_operand" "")))
(clobber (reg:CC_FP CCFP_REGNUM))])])
(define_insn "*mulsf3_i"
[(match_parallel 3 "float_operation"
[(set (match_operand:SF 0 "gpr_operand" "=r")
(mult:SF (match_operand:SF 1 "gpr_operand" "%r")
(match_operand:SF 2 "gpr_operand" "r")))
(clobber (reg:CC_FP CCFP_REGNUM))])]
""
"fmul %0,%1,%2"
[(set_attr "type" "fp")])
;; Division
(define_expand "divsf3"
[(set (match_operand:SF 0 "gpr_operand" "")
(div:SF (match_operand:SF 1 "gpr_operand" "")
(match_operand:SF 2 "gpr_operand" "")))]
"flag_reciprocal_math"
{
rtx one = CONST1_RTX (SFmode);
rtx dst = operands[0];
if (rtx_equal_p (dst, operands[1]))
{
emit_move_insn (dst, one);
DONE;
}
else if (!register_operand (dst, SFmode) && can_create_pseudo_p ())
dst = gen_reg_rtx (SFmode);
emit_insn (gen_recipsf2 (dst, one, operands[2],
sfunc_symbol (\"__fast_recipsf2\")));
emit_insn (gen_mulsf3 (operands[0], operands[1], dst));
DONE;
})
;; Before reload, keep the hard reg usage to clobbers so that the loop
;; optimizers can more easily move this insn.
;; It would be nicer to use a constraint for a GPR_0 - only register class,
;; but sched1 can still cause trouble then, and there is no guarantee of
;; better register allocations.
;; Neither is there when using the opposite strategy - putting explicit
;; hard register references into pre-reload rtl.
(define_expand "recipsf2"
[(parallel
[(set (match_operand:SF 0 "gpr_operand" "")
(div:SF (match_operand:SF 1 "const_float_1_operand" "")
(match_operand:SF 2 "move_src_operand" "")))
(use (match_operand:SI 3 "move_src_operand" ""))
(clobber (reg:SF 0))
(clobber (reg:SI 1))
(clobber (reg:SF GPR_IP))
(clobber (reg:DI GPR_16))
(clobber (reg:DI GPR_18))
(clobber (reg:SI GPR_20))
(clobber (reg:SI GPR_LR))
(clobber (reg:CC CC_REGNUM))
(clobber (reg:CC_FP CCFP_REGNUM))])])
(define_insn_and_split "*recipsf2_1"
[(match_parallel 4 "float_operation"
[(set (match_operand:SF 0 "gpr_operand" "=r,r")
(div:SF (match_operand:SF 1 "const_float_1_operand" "")
(match_operand:SF 2 "move_src_operand" "rU16m,rU16mCal")))
(use (match_operand:SI 3 "move_src_operand" "rU16m,rU16mCal"))
(clobber (reg:SF 0))
(clobber (reg:SI 1))
(clobber (reg:SF GPR_IP))
(clobber (reg:DI GPR_16))
(clobber (reg:DI GPR_18))
(clobber (reg:SI GPR_20))
(clobber (reg:SI GPR_LR))
(clobber (reg:CC CC_REGNUM))
(clobber (reg:CC_FP CCFP_REGNUM))])]
"flag_reciprocal_math"
"#"
"&& reload_completed"
[(set (reg:SI 1) (match_dup 3))
(set (reg:SF 0) (match_dup 2))
(parallel
[(set (reg:SF 0)
(div:SF (match_dup 1)
(reg:SF 0)))
(use (reg:SI 1))
(clobber (reg:SI GPR_IP))
(clobber (reg:DI GPR_16))
(clobber (reg:DI GPR_18))
(clobber (reg:SI GPR_20))
(clobber (reg:SI GPR_LR))
(clobber (reg:CC CC_REGNUM))
(clobber (reg:CC_FP CCFP_REGNUM))
(match_dup 5)
(match_dup 6)])
(set (match_dup 0) (reg:SF 0))]
"operands[5] = XVECEXP (operands[4], 0, XVECLEN (operands[4], 0) - 2);
operands[6] = XVECEXP (operands[4], 0, XVECLEN (operands[4], 0) - 1);"
[(set_attr "type" "fp_sfunc")
(set_attr "length" "16,24")])
(define_insn "*recipsf2_2"
[(match_parallel 1 "float_operation"
[(set (reg:SF 0)
(div:SF (match_operand:SF 0 "const_float_1_operand" "")
(reg:SF 0)))
(use (reg:SI 1))
(clobber (reg:SI GPR_IP))
(clobber (reg:DI GPR_16))
(clobber (reg:DI GPR_18))
(clobber (reg:SI GPR_20))
(clobber (reg:SI GPR_LR))
(clobber (reg:CC CC_REGNUM))
(clobber (reg:CC_FP CCFP_REGNUM))])]
"flag_reciprocal_math"
"jalr r1"
[(set_attr "type" "fp_sfunc")])
;; Fused multiply-add
(define_expand "fmasf4"
[(parallel
[(set (match_operand:SF 0 "gpr_operand" "")
(fma:SF (match_operand:SF 1 "gpr_operand" "")
(match_operand:SF 2 "gpr_operand" "")
(match_operand:SF 3 "gpr_operand" "")))
(clobber (reg:CC_FP CCFP_REGNUM))])]
"")
; The multiply operands are commutative, but since they have the
; same constraints, there is no point in telling reload about this.
(define_insn "*fmadd"
[(match_parallel 4 "float_operation"
[(set (match_operand:SF 0 "gpr_operand" "=r")
(fma:SF (match_operand:SF 1 "gpr_operand" "r")
(match_operand:SF 2 "gpr_operand" "r")
(match_operand:SF 3 "gpr_operand" "0")))
(clobber (reg:CC_FP CCFP_REGNUM))])]
""
"fmadd %0,%1,%2"
[(set_attr "type" "fp")])
; Once vetorization consistently works for this port, should check
; if the fmadd / fmsub patterns still serve a purpose. With the
; introduction of fma / fnma handling by the SSA optimizers,
; at least scalars should be handled by these optimizers, would
; have to see how well they do on vectors from auto-vectorization.
;
; combiner pattern, also used by vector combiner pattern
(define_expand "maddsf"
[(parallel
[(set (match_operand:SF 0 "gpr_operand" "=r")
(plus:SF (mult:SF (match_operand:SF 1 "gpr_operand" "r")
(match_operand:SF 2 "gpr_operand" "r"))
(match_operand:SF 3 "gpr_operand" "0")))
(clobber (reg:CC_FP CCFP_REGNUM))])]
"TARGET_FUSED_MADD")
(define_insn "*maddsf_combine"
[(match_parallel 4 "float_operation"
[(set (match_operand:SF 0 "gpr_operand" "=r")
(plus:SF (mult:SF (match_operand:SF 1 "gpr_operand" "r")
(match_operand:SF 2 "gpr_operand" "r"))
(match_operand:SF 3 "gpr_operand" "0")))
(clobber (reg:CC_FP CCFP_REGNUM))])]
"TARGET_FUSED_MADD"
"fmadd %0,%1,%2"
[(set_attr "type" "fp")])
;; Fused multiply-sub
(define_expand "fnmasf4"
[(parallel
[(set (match_operand:SF 0 "gpr_operand" "")
(fma:SF (neg:SF (match_operand:SF 1 "gpr_operand" ""))
(match_operand:SF 2 "gpr_operand" "")
(match_operand:SF 3 "gpr_operand" "")))
(clobber (reg:CC_FP CCFP_REGNUM))])]
"")
(define_insn "*fmsub"
[(match_parallel 4 "float_operation"
[(set (match_operand:SF 0 "gpr_operand" "=r")
(fma:SF (neg:SF (match_operand:SF 1 "gpr_operand" "r"))
(match_operand:SF 2 "gpr_operand" "r")
(match_operand:SF 3 "gpr_operand" "0")))
(clobber (reg:CC_FP CCFP_REGNUM))])]
""
"fmsub %0,%1,%2"
[(set_attr "type" "fp")])
(define_insn "*fmsub_combine"
[(match_parallel 4 "float_operation"
[(set (match_operand:SF 0 "gpr_operand" "=r")
(minus:SF (match_operand:SF 3 "gpr_operand" "0")
(mult:SF (match_operand:SF 1 "gpr_operand" "r")
(match_operand:SF 2 "gpr_operand" "r"))))
(clobber (reg:CC_FP CCFP_REGNUM))])]
"TARGET_FUSED_MADD"
"fmsub %0,%1,%2"
[(set_attr "type" "fp")])
;; float / integer conversions
(define_expand "floatsisf2"
[(parallel
[(set (match_operand:SF 0 "gpr_operand" "")
(float:SF (match_operand:SI 1 "gpr_operand" "")))
(clobber (reg:CC_FP CCFP_REGNUM))])])
(define_insn "*floatsisf2_i"
[(match_parallel 2 "float_operation"
[(set (match_operand:SF 0 "gpr_operand" "=r")
(float:SF (match_operand:SI 1 "gpr_operand" "r")))
(clobber (reg:CC_FP CCFP_REGNUM))])]
""
"float %0, %1"
[(set_attr "type" "fp")])
(define_expand "floatsisf2_cmp"
[(parallel
[(set (reg:CC_FP CCFP_REGNUM)
(compare:CC_FP (float:SF (match_operand:SF 1 "gpr_operand" "r"))
(match_dup 2)))
(set (match_operand:SF 0 "gpr_operand" "=r")
(float:SF (match_dup 1)))])]
""
"operands[2] = CONST0_RTX (SFmode);")
(define_insn "*floatsisf2_cmp_i"
[(match_parallel 3 "float_operation"
[(set (reg:CC_FP CCFP_REGNUM)
(compare:CC_FP (float:SF (match_operand:SF 1 "gpr_operand" "r"))
(match_operand:SF 2 "const0_operand" "")))
(set (match_operand:SF 0 "gpr_operand" "=r")
(float:SF (match_dup 1)))])]
""
"float %0, %1"
[(set_attr "type" "fp")])
(define_expand "floatunssisf2"
[(set (match_operand:SF 0 "gpr_operand" "")
(float:SF (match_operand:SI 1 "gpr_operand" "")))]
"epiphany_normal_fp_rounding == /*FP_MODE_ROUND_TRUNC*/ 2"
{
rtx cst = force_reg (SImode, gen_int_mode (0xb0800000, SImode));
rtx tmp = gen_reg_rtx (SImode);
rtx cmp = gen_rtx_GTU (VOIDmode, gen_rtx_REG (CCmode, CC_REGNUM), const0_rtx);
if (reg_overlap_mentioned_p (operands[0], operands[1]))
operands[1] = copy_to_mode_reg (SImode, operands[1]);
emit_insn (gen_floatsisf2 (operands[0], operands[1]));
emit_insn (gen_ashrsi3 (tmp, operands[1], GEN_INT (8)));
emit_insn (gen_sub_f (tmp, tmp, cst));
emit_insn (gen_movsfcc (operands[0], cmp,
simplify_gen_subreg (SFmode, tmp, SImode, 0),
operands[0]));
DONE;
})
(define_expand "fix_truncsfsi2"
[(parallel
[(set (match_operand:SI 0 "gpr_operand" "")
(fix:SI (match_operand:SF 1 "gpr_operand" "")))
(clobber (reg:CC_FP CCFP_REGNUM))])])
(define_insn "*fix_truncsfsi2_i"
[(match_parallel 2 "float_operation"
[(set (match_operand:SI 0 "gpr_operand" "=r")
(fix:SI (match_operand:SF 1 "gpr_operand" "r")))
(clobber (reg:CC_FP CCFP_REGNUM))])]
""
"fix %0, %1"
[(set_attr "type" "fp")
(set (attr "fp_mode")
(cond [(match_test "TARGET_MAY_ROUND_FOR_TRUNC")
(const_string "round_unknown")]
(const_string "round_trunc")))])
(define_expand "fixuns_truncsfsi2"
[(set (match_operand:SI 0 "gpr_operand" "")
(unsigned_fix:SI (match_operand:SF 1 "gpr_operand" "")))]
""
{
if (reg_overlap_mentioned_p (operands[0], operands[1]))
operands[1] = copy_to_mode_reg (SImode, operands[1]);
if (TARGET_SOFT_CMPSF || optimize_function_for_speed_p (cfun))
{
rtx op1si;
/* By toggling what it to be bit31 before the shift, we get a chance to
use a short movt insn. */
rtx bit31 = force_reg (SImode, GEN_INT (0x800000));
rtx tmp = gen_reg_rtx (SImode);
rtx limit = force_reg (SImode, gen_int_mode (0x4f000000, SImode));
rtx cmp
= gen_rtx_GE (VOIDmode, gen_rtx_REG (CCmode, CC_REGNUM), const0_rtx);
op1si = simplify_gen_subreg (SImode, operands[1], SFmode, 0);
emit_insn (gen_fix_truncsfsi2 (operands[0], operands[1]));
emit_insn (gen_subsi3_i (tmp, op1si, bit31));
emit_insn (gen_ashlsi3 (tmp, tmp, GEN_INT (8)));
emit_insn (gen_cmpsi_cc_insn (op1si, limit));
emit_insn (gen_movsicc (operands[0], cmp, tmp, operands[0]));
}
else
{
REAL_VALUE_TYPE offset;
rtx limit;
rtx tmp = gen_reg_rtx (SFmode);
rtx_code_label *label = gen_label_rtx ();
rtx bit31;
rtx cc1 = gen_rtx_REG (CC_FPmode, CCFP_REGNUM);
rtx cmp = gen_rtx_LT (VOIDmode, cc1, CONST0_RTX (SFmode));
real_2expN (&offset, 31, SFmode);
limit = const_double_from_real_value (offset, SFmode);
limit = force_reg (SFmode, limit);
emit_insn (gen_fix_truncsfsi2 (operands[0], operands[1]));
emit_insn (gen_subsf3_f (tmp, operands[1], limit));
emit_jump_insn (gen_branch_insn (label, cmp, cc1));
bit31 = force_reg (SImode, gen_int_mode (0x80000000, SImode));
emit_insn (gen_fix_truncsfsi2 (operands[0], tmp));
emit_insn (gen_xorsi3 (operands[0], operands[0], bit31));
emit_label (label);
}
DONE;
})
(define_expand "iadd"
[(parallel
[(set (match_operand:SF 0 "gpr_operand" "")
(plus:SI (match_operand:SF 1 "gpr_operand" "")
(match_operand:SF 2 "gpr_operand" "")))
(clobber (reg:CC_FP CCFP_REGNUM))])])
(define_insn "*iadd_i"
[(match_parallel 3 "float_operation"
[(set (match_operand:SI 0 "gpr_operand" "=r")
(plus:SI (match_operand:SI 1 "gpr_operand" "%r")
(match_operand:SI 2 "gpr_operand" "r")))
(clobber (reg:CC_FP CCFP_REGNUM))])]
""
"iadd %0, %1, %2"
[(set_attr "type" "fp_int")])
(define_expand "isub"
[(parallel
[(set (match_operand:SF 0 "gpr_operand" "")
(minus:SI (match_operand:SF 1 "gpr_operand" "")
(match_operand:SF 2 "gpr_operand" "")))
(clobber (reg:CC_FP CCFP_REGNUM))])])
(define_insn "*isub_i"
[(match_parallel 3 "float_operation"
[(set (match_operand:SI 0 "gpr_operand" "=r")
(minus:SI (match_operand:SI 1 "gpr_operand" "r")
(match_operand:SI 2 "gpr_operand" "r")))
(clobber (reg:CC_FP CCFP_REGNUM))])]
""
"isub %0, %1, %2"
[(set_attr "type" "fp_int")])
; Try to figure out if we over-committed the FPU, and if so, move
; some insns back over to the integer pipe.
; The peephole optimizer 'consumes' the insns that are explicitly
; mentioned. We do not want the preceding insn reconsidered, but
; we do want that for the following one, so that if we have a run
; of five fpu users, two of them get changed. Therefore, we
; use next_active_insn to look at the 'following' insn. That should
; exist, because peephole2 runs after reload, and there has to be
; a return after an fp_int insn.
; ??? However, we cannot even ordinarily match the preceding insn;
; there is some bug in the generators such that then it leaves out
; the check for PARALLEL before the length check for the then-second
; main insn. Observed when compiling compatibility-atomic-c++0x.cc
; from libstdc++-v3.
(define_peephole2
[(match_parallel 3 "float_operation"
[(set (match_operand:SI 0 "gpr_operand" "")
(match_operator:SI 4 "addsub_operator"
[(match_operand:SI 1 "gpr_operand" "")
(match_operand:SI 2 "gpr_operand" "")]))
(clobber (reg:CC_FP CCFP_REGNUM))])]
"get_attr_sched_use_fpu (prev_active_insn (peep2_next_insn (0)))
&& peep2_regno_dead_p (1, CC_REGNUM)
&& get_attr_sched_use_fpu (next_active_insn (peep2_next_insn (0)))"
[(parallel [(set (match_dup 0) (match_dup 4))
(clobber (reg:CC CC_REGNUM))])]
)
(define_peephole2
[(match_parallel 3 "float_operation"
[(set (match_operand:SI 0 "gpr_operand" "")
(mult:SI
(match_operand:SI 1 "gpr_operand" "")
(match_operand:SI 2 "gpr_operand" "")))
(clobber (reg:CC_FP CCFP_REGNUM))])]
"prev_active_insn (peep2_next_insn (0))
&& get_attr_sched_use_fpu (prev_active_insn (peep2_next_insn (0)))
&& peep2_regno_dead_p (1, CC_REGNUM)
&& get_attr_sched_use_fpu (next_active_insn (peep2_next_insn (0)))
&& find_reg_note (insn, REG_EQUAL, NULL_RTX) != NULL_RTX
&& GET_CODE (XEXP (find_reg_note (insn, REG_EQUAL, NULL_RTX), 0)) == MULT
&& CONST_INT_P (XEXP (XEXP (find_reg_note (insn, REG_EQUAL, NULL_RTX), 0),
1))"
[(parallel [(set (match_dup 0) (ashift:SI (match_dup 1) (match_dup 4)))
(clobber (reg:CC CC_REGNUM))])]
{
operands[4]
= XEXP (XEXP (find_reg_note (curr_insn, REG_EQUAL, NULL_RTX), 0), 1);
})
(define_expand "mulsi3"
[(parallel
[(set (match_operand:SI 0 "gpr_operand" "")
(mult:SI (match_operand:SI 1 "gpr_operand" "")
(match_operand:SI 2 "gpr_operand" "")))
(clobber (reg:CC_FP CCFP_REGNUM))])])
(define_insn "*imul"
[(match_parallel 3 "float_operation"
[(set (match_operand:SI 0 "gpr_operand" "=r")
(mult:SI (match_operand:SI 1 "gpr_operand" "%r")
(match_operand:SI 2 "gpr_operand" "r")))
(clobber (reg:CC_FP CCFP_REGNUM))])]
""
"imul %0, %1, %2"
[(set_attr "type" "fp_int")])
; combiner pattern, also used by vector combiner pattern
(define_expand "maddsi"
[(parallel
[(set (match_operand:SI 0 "gpr_operand" "=r")
(plus:SI (mult:SI (match_operand:SI 1 "gpr_operand" "r")
(match_operand:SI 2 "gpr_operand" "r"))
(match_operand:SI 3 "gpr_operand" "0")))
(clobber (reg:CC_FP CCFP_REGNUM))])]
"")
(define_insn "*maddsi_combine"
[(match_parallel 4 "float_operation"
[(set (match_operand:SI 0 "gpr_operand" "=r")
(plus:SI (mult:SI (match_operand:SI 1 "gpr_operand" "r")
(match_operand:SI 2 "gpr_operand" "r"))
(match_operand:SI 3 "gpr_operand" "0")))
(clobber (reg:CC_FP CCFP_REGNUM))])]
""
"imadd %0, %1, %2"
[(set_attr "type" "fp_int")])
(define_insn "*imsub"
[(match_parallel 4 "float_operation"
[(set (match_operand:SI 0 "gpr_operand" "=r")
(minus:SI (match_operand:SI 3 "gpr_operand" "0")
(mult:SI (match_operand:SI 1 "gpr_operand" "r")
(match_operand:SI 2 "gpr_operand" "r"))))
(clobber (reg:CC_FP CCFP_REGNUM))])]
""
"imsub %0, %1, %2"
[(set_attr "type" "fp_int")])
(define_expand "divsi3"
[(parallel
[(set (match_operand:SI 0 "move_dest_operand" "")
(div:SI (match_operand:SI 1 "move_src_operand" "")
(match_operand:SI 2 "move_src_operand" "")))
(use (match_dup 3))
(clobber (reg:SI 0))
(clobber (reg:SI 1))
(clobber (reg:SI GPR_IP))
(clobber (reg:DI GPR_16))
(clobber (reg:DI GPR_18))
(clobber (reg:SI GPR_20))
(clobber (reg:SI GPR_LR))
(clobber (reg:CC CC_REGNUM))
(clobber (reg:CC_FP CCFP_REGNUM))])]
""
"operands[3] = sfunc_symbol (\"__divsi3\");")
;; Before reload, keep the hard reg usage to clobbers so that the loop
;; optimizers can more easily move this insn.
(define_insn_and_split "*divsi3_1"
[(match_parallel 4 "float_operation"
[(set (match_operand:SI 0 "move_dest_operand" "=r,r")
(div:SI (match_operand:SI 1 "move_src_operand" "rU16m,rU16mCal")
(match_operand:SI 2 "move_src_operand" "rU16m,rU16mCal")))
(use (match_operand:SI 3 "call_address_operand" "Csy,r"))
(clobber (reg:SI 0))
(clobber (reg:SI 1))
(clobber (reg:SI GPR_IP))
(clobber (reg:DI GPR_16))
(clobber (reg:DI GPR_18))
(clobber (reg:SI GPR_20))
(clobber (reg:SI GPR_LR))
(clobber (reg:CC CC_REGNUM))
(clobber (reg:CC_FP CCFP_REGNUM))])]
""
"#"
"&& reload_completed"
[(set (reg:SI 0) (match_dup 1))
(set (reg:SI 1) (match_dup 2))
(parallel
[(set (reg:SI 0) (div:SI (reg:SI 0) (reg:SI 1)))
(use (match_dup 3))
(clobber (reg:SI 1))
(clobber (reg:SI GPR_IP))
(clobber (reg:DI GPR_16))
(clobber (reg:DI GPR_18))
(clobber (reg:SI GPR_20))
(clobber (reg:SI GPR_LR))
(clobber (reg:CC CC_REGNUM))
(clobber (reg:CC_FP CCFP_REGNUM))
(match_dup 5)
(match_dup 6)])
(set (match_dup 0) (reg:SI 0))]
"operands[5] = XVECEXP (operands[4], 0, XVECLEN (operands[4], 0) - 2);
operands[6] = XVECEXP (operands[4], 0, XVECLEN (operands[4], 0) - 1);"
[(set_attr "type" "fp_sfunc")
(set_attr "length" "16,24")])
(define_insn "*divsi3_2"
[(match_parallel 1 "float_operation"
[(set (reg:SI 0) (div:SI (reg:SI 0) (reg:SI 1)))
(use (match_operand:SI 0 "call_address_operand" "Csy,r"))
(clobber (reg:SI 1))
(clobber (reg:SI GPR_IP))
(clobber (reg:DI GPR_16))
(clobber (reg:DI GPR_18))
(clobber (reg:SI GPR_20))
(clobber (reg:SI GPR_LR))
(clobber (reg:CC CC_REGNUM))
(clobber (reg:CC_FP CCFP_REGNUM))])]
""
"%f0"
[(set_attr "type" "fp_sfunc")])
(define_expand "udivsi3"
[(parallel
[(set (match_operand:SI 0 "move_dest_operand" "")
(udiv:SI (match_operand:SI 1 "move_src_operand" "")
(match_operand:SI 2 "move_src_operand" "")))
(use (match_dup 3))
(clobber (reg:SI 0))
(clobber (reg:SI 1))
(clobber (reg:SI GPR_IP))
(clobber (reg:DI GPR_16))
(clobber (reg:SI GPR_18))
(clobber (reg:SI GPR_LR))
(clobber (reg:CC CC_REGNUM))
(clobber (reg:CC_FP CCFP_REGNUM))])]
""
"operands[3] = sfunc_symbol (\"__udivsi3\");")
;; Before reload, keep the hard reg usage to clobbers so that the loop
;; optimizers can more easily move this insn.
(define_insn_and_split "*udivsi3_1"
[(match_parallel 4 "float_operation"
[(set (match_operand:SI 0 "move_dest_operand" "=r,r")
(udiv:SI (match_operand:SI 1 "move_src_operand" "rU16m,rU16mCal")
(match_operand:SI 2 "move_src_operand" "rU16m,rU16mCal")))
(use (match_operand:SI 3 "call_address_operand" "Csy,r"))
(clobber (reg:SI 0))
(clobber (reg:SI 1))
(clobber (reg:SI GPR_IP))
(clobber (reg:DI GPR_16))
(clobber (reg:SI GPR_18))
(clobber (reg:SI GPR_LR))
(clobber (reg:CC CC_REGNUM))
(clobber (reg:CC_FP CCFP_REGNUM))])]
""
"#"
"&& reload_completed"
[(set (reg:SI 0) (match_dup 1))
(set (reg:SI 1) (match_dup 2))
(parallel
[(set (reg:SI 0) (udiv:SI (reg:SI 0) (reg:SI 1)))
(use (match_dup 3))
(clobber (reg:SI 1))
(clobber (reg:SI GPR_IP))
(clobber (reg:DI GPR_16))
(clobber (reg:SI GPR_18))
(clobber (reg:SI GPR_LR))
(clobber (reg:CC CC_REGNUM))
(clobber (reg:CC_FP CCFP_REGNUM))
(match_dup 5)
(match_dup 6)])
(set (match_dup 0) (reg:SI 0))]
"operands[5] = XVECEXP (operands[4], 0, XVECLEN (operands[4], 0) - 2);
operands[6] = XVECEXP (operands[4], 0, XVECLEN (operands[4], 0) - 1);"
[(set_attr "type" "fp_sfunc")
(set_attr "length" "16,24")])
(define_insn "*udivsi3_2"
[(match_parallel 1 "float_operation"
[(set (reg:SI 0) (udiv:SI (reg:SI 0) (reg:SI 1)))
(use (match_operand:SI 0 "call_address_operand" "Csy,r"))
(clobber (reg:SI 1))
(clobber (reg:SI GPR_IP))
(clobber (reg:DI GPR_16))
(clobber (reg:SI GPR_18))
(clobber (reg:SI GPR_LR))
(clobber (reg:CC CC_REGNUM))
(clobber (reg:CC_FP CCFP_REGNUM))])]
""
"%f0"
[(set_attr "type" "fp_sfunc")])
(define_expand "modsi3"
[(parallel
[(set (match_operand:SI 0 "move_dest_operand" "")
(mod:SI (match_operand:SI 1 "move_src_operand" "")
(match_operand:SI 2 "move_src_operand" "")))
(use (match_dup 3))
(clobber (reg:SI 0))
(clobber (reg:SI 1))
(clobber (reg:SI 2))
(clobber (reg:SI GPR_IP))
(clobber (reg:DI GPR_16))
(clobber (reg:DI GPR_18))
(clobber (reg:SI GPR_LR))
(clobber (reg:CC CC_REGNUM))
(clobber (reg:CC_FP CCFP_REGNUM))])]
""
"operands[3] = sfunc_symbol (\"__modsi3\");")
;; Before reload, keep the hard reg usage to clobbers so that the loop
;; optimizers can more easily move this insn.
(define_insn_and_split "*modsi3_1"
[(match_parallel 4 "float_operation"
[(set (match_operand:SI 0 "move_dest_operand" "=r,r")
(mod:SI (match_operand:SI 1 "move_src_operand" "rU16m,rU16mCal")
(match_operand:SI 2 "move_src_operand" "rU16m,rU16mCal")))
(use (match_operand:SI 3 "call_address_operand" "Csy,r"))
(clobber (reg:SI 0))
(clobber (reg:SI 1))
(clobber (reg:SI 2))
(clobber (reg:SI GPR_IP))
(clobber (reg:DI GPR_16))
(clobber (reg:DI GPR_18))
(clobber (reg:SI GPR_LR))
(clobber (reg:CC CC_REGNUM))
(clobber (reg:CC_FP CCFP_REGNUM))])]
""
"#"
"&& reload_completed"
[(set (reg:SI 0) (match_dup 1))
(set (reg:SI 1) (match_dup 2))
(parallel
[(set (reg:SI 0) (mod:SI (reg:SI 0) (reg:SI 1)))
(use (match_dup 3))
(clobber (reg:SI 2))
(clobber (reg:SI GPR_IP))
(clobber (reg:DI GPR_16))
(clobber (reg:DI GPR_18))
(clobber (reg:SI GPR_LR))
(clobber (reg:CC CC_REGNUM))
(clobber (reg:CC_FP CCFP_REGNUM))
(match_dup 5)
(match_dup 6)])
(set (match_dup 0) (reg:SI 0))]
"operands[5] = XVECEXP (operands[4], 0, XVECLEN (operands[4], 0) - 2);
operands[6] = XVECEXP (operands[4], 0, XVECLEN (operands[4], 0) - 1);"
[(set_attr "type" "fp_sfunc")
(set_attr "length" "16,24")])
(define_insn "*modsi3_2"
[(match_parallel 1 "float_operation"
[(set (reg:SI 0) (mod:SI (reg:SI 0) (reg:SI 1)))
(use (match_operand:SI 0 "call_address_operand" "Csy,r"))
(clobber (reg:SI 2))
(clobber (reg:SI GPR_IP))
(clobber (reg:DI GPR_16))
(clobber (reg:DI GPR_18))
(clobber (reg:SI GPR_LR))
(clobber (reg:CC CC_REGNUM))
(clobber (reg:CC_FP CCFP_REGNUM))])]
""
"%f0"
[(set_attr "type" "fp_sfunc")])
(define_expand "umodsi3"
[(parallel
[(set (match_operand:SI 0 "move_dest_operand" "")
(umod:SI (match_operand:SI 1 "move_src_operand" "")
(match_operand:SI 2 "move_src_operand" "")))
(use (match_dup 3))
(clobber (reg:SI 0))
(clobber (reg:SI 1))
(clobber (reg:SI 2))
(clobber (reg:SI GPR_IP))
(clobber (reg:DI GPR_16))
(clobber (reg:SI GPR_LR))
(clobber (reg:CC CC_REGNUM))
(clobber (reg:CC_FP CCFP_REGNUM))])]
""
"operands[3] = sfunc_symbol (\"__umodsi3\");")
;; Before reload, keep the hard reg usage to clobbers so that the loop
;; optimizers can more easily move this insn.
(define_insn_and_split "*umodsi3_1"
[(match_parallel 4 "float_operation"
[(set (match_operand:SI 0 "move_dest_operand" "=r,r")
(umod:SI (match_operand:SI 1 "move_src_operand" "rU16m,rU16mCal")
(match_operand:SI 2 "move_src_operand" "rU16m,rU16mCal")))
(use (match_operand:SI 3 "call_address_operand" "Csy,r"))
(clobber (reg:SI 0))
(clobber (reg:SI 1))
(clobber (reg:SI 2))
(clobber (reg:SI GPR_IP))
(clobber (reg:DI GPR_16))
(clobber (reg:SI GPR_LR))
(clobber (reg:CC CC_REGNUM))
(clobber (reg:CC_FP CCFP_REGNUM))])]
""
"#"
"&& reload_completed"
[(set (reg:SI 0) (match_dup 1))
(set (reg:SI 1) (match_dup 2))
(parallel
[(set (reg:SI 0) (umod:SI (reg:SI 0) (reg:SI 1)))
(use (match_dup 3))
(clobber (reg:SI 2))
(clobber (reg:SI GPR_IP))
(clobber (reg:DI GPR_16))
(clobber (reg:SI GPR_LR))
(clobber (reg:CC CC_REGNUM))
(clobber (reg:CC_FP CCFP_REGNUM))
(match_dup 5)
(match_dup 6)])
(set (match_dup 0) (reg:SI 0))]
"operands[5] = XVECEXP (operands[4], 0, XVECLEN (operands[4], 0) - 2);
operands[6] = XVECEXP (operands[4], 0, XVECLEN (operands[4], 0) - 1);"
[(set_attr "type" "fp_sfunc")
(set_attr "length" "16,24")])
(define_insn "*umodsi3_2"
[(match_parallel 1 "float_operation"
[(set (reg:SI 0) (umod:SI (reg:SI 0) (reg:SI 1)))
(use (match_operand:SI 0 "call_address_operand" "Csy,r"))
(clobber (reg:SI 2))
(clobber (reg:SI GPR_IP))
(clobber (reg:DI GPR_16))
(clobber (reg:SI GPR_LR))
(clobber (reg:CC CC_REGNUM))
(clobber (reg:CC_FP CCFP_REGNUM))])]
""
"%f0"
[(set_attr "type" "fp_sfunc")])
; Disable interrupts.
; Any earlier values read from CONFIG_REGNUM are out of date, since interrupts
; might have changed settings that we do not want to mess with.
(define_insn "gid"
[(set (reg:SI CONFIG_REGNUM)
(unspec_volatile:SI [(const_int 0)] UNSPECV_GID))]
""
"gid"
[(set_attr "type" "flow")])
; Enable interrupts.
; Present CONTROL_REGNUM here to make sure it is live before the
; actual uses in floating point insns / calls are inserted.
; FWIW, interrupts also do mind what is in the control register.
(define_insn "gie"
[(unspec_volatile [(reg:SI CONFIG_REGNUM)] UNSPECV_GIE)]
""
"gie"
[(set_attr "type" "flow")])
; Floating point instructions require manipulating the control register.
; Manipulating the control register needs arithmetic.
; Arithmetic clobbers flags.
; The flags are in the status register, which also contains the alternate
; flag and the interrupt enable/disable bits.
; saving/restoring status and mixing up the order with gid/gie could
; lead to disaster.
; Usually, saving/restoring the status is unnecessary, and will be optimized
; away. But when we really need it, we must make sure that we don't change
; anything but the flags.
; N.B.: We could make the constant easier to load by inverting it, but
; then we'd need to clobber the saved value - and that would make optimizing
; away unneeded saves/restores harder / less likely.
(define_expand "movcc"
[(parallel [(set (match_operand:CC 0 "cc_move_operand" "")
(match_operand:CC 1 "cc_move_operand" ""))
(use (match_dup 2))
(clobber (match_scratch:SI 3 "=X, &r"))])]
""
"operands[2] = gen_int_mode (~0x10f0, SImode);")
(define_insn "*movcc_i"
[(set (match_operand:CC 0 "cc_move_operand" "=r,Rcc")
(match_operand:CC 1 "cc_move_operand" "Rcc, r"))
(use (match_operand:SI 2 "nonmemory_operand" "X, r"))
(clobber (match_scratch:SI 3 "=X, &r"))]
""
"@
movfs %0,status
movfs %3,status\;eor %3,%3,%1\;and %3,%3,%2\;eor %3,%3,%1\;movts status,%3"
[(set_attr "type" "flow")
(set_attr "length" "20,4")])
(define_insn_and_split "save_config"
[(set (match_operand:SI 0 "gpr_operand" "=r") (reg:SI CONFIG_REGNUM))
(use (reg:SI FP_NEAREST_REGNUM))
(use (reg:SI FP_TRUNCATE_REGNUM))
(use (reg:SI FP_ANYFP_REGNUM))]
""
"#"
"reload_completed"
[(set (match_dup 0) (reg:SI CONFIG_REGNUM))])
(define_insn_and_split "set_fp_mode"
[(set (reg:SI FP_NEAREST_REGNUM)
(match_operand:SI 0 "set_fp_mode_operand" "rCfm"))
(set (reg:SI FP_TRUNCATE_REGNUM) (match_dup 0))
(set (reg:SI FP_ANYFP_REGNUM)
(match_operand:SI 1 "set_fp_mode_operand" "rCfm"))
(use (match_operand:SI 2 "gpr_operand" "r"))
(clobber (reg:CC CC_REGNUM))
(clobber (match_scratch:SI 3 "=&r"))]
""
"#"
"reload_completed || !rtx_equal_p (operands[0], operands[1])"
[(const_int 0)]
{
if (!reload_completed)
emit_note (NOTE_INSN_DELETED);
else
epiphany_expand_set_fp_mode (operands);
DONE;
})
;; Boolean instructions.
;;
;; We don't define the DImode versions as expand_binop does a good enough job.
(define_insn "andsi3"
[(set (match_operand:SI 0 "gpr_operand" "=r")
(and:SI (match_operand:SI 1 "gpr_operand" "r")
(match_operand:SI 2 "gpr_operand" "r")))
(clobber (reg:CC CC_REGNUM))]
""
"and %0,%1,%2")
(define_insn "iorsi3"
[(set (match_operand:SI 0 "gpr_operand" "=r")
(ior:SI (match_operand:SI 1 "gpr_operand" "r")
(match_operand:SI 2 "gpr_operand" "r")))
(clobber (reg:CC CC_REGNUM))]
""
"orr %0,%1,%2")
(define_insn "xorsi3"
[(set (match_operand:SI 0 "gpr_operand" "=r")
(xor:SI (match_operand:SI 1 "gpr_operand" "r")
(match_operand:SI 2 "gpr_operand" "r")))
(clobber (reg:CC CC_REGNUM))]
""
"eor %0,%1,%2")
(define_expand "one_cmplsi2"
[(set (match_operand:SI 0 "gpr_operand" "")
(xor:SI (match_operand:SI 1 "gpr_operand" "")
(match_dup 2)))]
""
{
if (epiphany_m1reg >= 0)
emit_insn (gen_one_cmplsi2_i (operands[0], operands[1]));
else
emit_insn (gen_xorsi3 (operands[0], operands[1],
force_reg (SImode, GEN_INT (-1))));
DONE;
})
; Note that folding this pattern into the xorsi3 pattern would make combine
; less effective.
(define_insn "one_cmplsi2_i"
[(set (match_operand:SI 0 "gpr_operand" "=r")
(not:SI (match_operand:SI 1 "gpr_operand" "r")))
(clobber (reg:CC CC_REGNUM))]
"epiphany_m1reg >= 0"
"eor %0,%1,%-")
;; Shift instructions.
;; In principle we could support arbitrary symbolic values as shift constant
;; (truncating the value appropriately), but that would require a suitable
;; relocation and assembler & linker support.
(define_insn "ashrsi3"
[(set (match_operand:SI 0 "gpr_operand" "=r,r")
(ashiftrt:SI (match_operand:SI 1 "gpr_operand" "r,r")
(match_operand:SI 2 "arith_operand" "r,K")))
(clobber (reg:CC CC_REGNUM))]
""
"asr %0,%1,%2"
[(set_attr "length" "4")
(set_attr "type" "shift")])
(define_insn "ashrsi3_tst"
[(set (reg:CC CC_REGNUM)
(compare:CC
(ashiftrt:SI (match_operand:SI 1 "gpr_operand" "r,r")
(match_operand:SI 2 "arith_operand" "r,K"))
(const_int 0)))
(set (match_operand:SI 0 "gpr_operand" "=r,r")
(ashiftrt:SI (match_dup 1) (match_dup 2)))]
""
"asr %0,%1,%2"
[(set_attr "length" "4")
(set_attr "type" "shift")])
;; Logical Shift Right
(define_insn "lshrsi3"
[(set (match_operand:SI 0 "gpr_operand" "=r,r")
(lshiftrt:SI (match_operand:SI 1 "gpr_operand" "r,r")
(match_operand:SI 2 "arith_operand" "r,K")))
(clobber (reg:CC CC_REGNUM))]
""
"lsr %0,%1,%2"
[(set_attr "length" "4")
(set_attr "type" "shift")])
(define_insn "lshrsi3_tst"
[(set (reg:CC CC_REGNUM)
(compare:CC
(lshiftrt:SI (match_operand:SI 1 "gpr_operand" "r,r")
(match_operand:SI 2 "arith_operand" "r,K"))
(const_int 0)))
(set (match_operand:SI 0 "gpr_operand" "=r,r")
(lshiftrt:SI (match_dup 1) (match_dup 2)))]
""
"lsr %0,%1,%2"
[(set_attr "length" "4")
(set_attr "type" "shift")])
;; Logical/Arithmetic Shift Left
(define_insn "ashlsi3"
[(set (match_operand:SI 0 "gpr_operand" "=r,r")
(ashift:SI (match_operand:SI 1 "gpr_operand" "r,r")
(match_operand:SI 2 "arith_operand" "r,K")))
(clobber (reg:CC CC_REGNUM))]
""
"lsl %0,%1,%2"
[(set_attr "length" "4")
(set_attr "type" "shift")])
(define_insn "*ashlsi_btst"
[(set (reg:CC_N_NE CC_REGNUM)
(compare:CC_N_NE
(zero_extract:SI (match_operand:SI 1 "gpr_operand" "r")
(const_int 1)
(match_operand 2 "const_int_operand" "K"))
(const_int 0)))
(clobber (match_scratch:SI 0 "=r"))]
""
{
rtx xop[3];
xop[0] = operands[0];
xop[1] = operands[1];
xop[2] = GEN_INT (31-INTVAL (operands[2]));
output_asm_insn ("lsl %0,%1,%2", xop);
return "";
})
;; zero extensions
(define_insn_and_split "zero_extendqisi2"
[(set (match_operand:SI 0 "register_operand" "=r,r")
(zero_extend:SI (match_operand:QI 1 "nonimmediate_operand" "r,m")))
(clobber (reg:CC CC_REGNUM))]
""
"@
#
ldrb %0,%1"
"reload_completed
? true_regnum (operands[1]) >= 0
: REG_P (operands[1]) && REGNO (operands[1]) < FIRST_PSEUDO_REGISTER"
[(parallel [(set (match_dup 0) (ashift:SI (match_dup 2) (const_int 24)))
(clobber (reg:CC CC_REGNUM))])
(parallel [(set (match_dup 0) (lshiftrt:SI (match_dup 0) (const_int 24)))
(clobber (reg:CC CC_REGNUM))])]
"operands[2] = simplify_gen_subreg (SImode, operands[1], QImode, 0);")
(define_insn "zero_extendhisi2"
[(set (match_operand:SI 0 "register_operand" "=r,r")
(zero_extend:SI (match_operand:HI 1 "nonimmediate_operand" "0,m")))]
""
"@
movt %0, 0
ldrh %0,%c1")
;; Compare instructions.
(define_insn "cmpsi_cc_insn"
[(set (reg:CC CC_REGNUM)
(compare:CC (match_operand:SI 0 "add_reg_operand" "r,r")
(match_operand:SI 1 "arith_operand" "r,L")))
(clobber (match_scratch:SI 2 "=r,r"))]
""
"sub %2,%0,%1"
[(set_attr "type" "compare")])
(define_insn "sub_f"
[(set (reg:CC CC_REGNUM)
(compare:CC (match_operand:SI 1 "gpr_operand" "r,r")
(match_operand:SI 2 "arith_operand" "r,L")))
(set (match_operand:SI 0 "gpr_operand" "=r,r")
(minus:SI (match_dup 1) (match_dup 2)))]
""
"sub %0,%1,%2"
[(set_attr "type" "compare")])
(define_insn "*sub_f_add_imm"
[(set (reg:CC CC_REGNUM)
(compare:CC (match_operand:SI 1 "gpr_operand" "r")
(match_operand:SI 2 "arith_int_operand" "L")))
(set (match_operand:SI 0 "gpr_operand" "=r")
(plus:SI (match_dup 1) (match_operand:SI 3 "const_int_operand" "CnL")))]
"INTVAL (operands[2]) == -INTVAL (operands[3])"
"sub %0,%1,%2"
[(set_attr "type" "compare")])
(define_expand "abssi2"
[(set (match_dup 2) (const_int 0))
(parallel [(set (reg:CC CC_REGNUM)
(compare:CC (match_dup 2)
(match_operand:SI 1 "nonmemory_operand" "")))
(set (match_dup 3)
(minus:SI (match_dup 2) (match_dup 1)))])
(set (match_operand:SI 0 "gpr_operand" "=r")
(if_then_else:SI (gt:SI (reg:CC CC_REGNUM) (const_int 0))
(match_dup 3)
(match_dup 1)))]
"TARGET_CMOVE"
"operands[2] = gen_reg_rtx (SImode); operands[3] = gen_reg_rtx (SImode);")
(define_insn "*add_c"
[(set (reg:CC_C_LTU CC_REGNUM)
(compare:CC_C_LTU
(plus:SI (match_operand:SI 1 "gpr_operand" "%r,r")
(match_operand:SI 2 "arith_operand" "r,L"))
(match_dup 1)))
(set (match_operand:SI 0 "gpr_operand" "=r,r")
(plus:SI (match_dup 1) (match_dup 2)))]
""
"add %0,%1,%2"
[(set_attr "type" "compare")])
(define_insn "*add_c_rev"
[(set (reg:CC_C_LTU CC_REGNUM)
(compare:CC_C_LTU
(plus:SI (match_operand:SI 1 "gpr_operand" "%r,r")
(match_operand:SI 2 "arith_operand" "r,L"))
(match_dup 1)))
(set (match_operand:SI 0 "gpr_operand" "=r,r")
(plus:SI (match_dup 2) (match_dup 1)))]
""
"add %0,%1,%2"
[(set_attr "type" "compare")])
(define_insn "*sub_c"
[(set (reg:CC_C_GTU CC_REGNUM)
(compare:CC_C_GTU
(minus:SI (match_operand:SI 1 "gpr_operand" "r,r")
(match_operand:SI 2 "arith_operand" "r,L"))
(match_dup 1)))
(set (match_operand:SI 0 "gpr_operand" "=r,r")
(minus:SI (match_dup 1) (match_dup 2)))]
""
"sub %0,%1,%2"
[(set_attr "type" "compare")])
(define_insn "*sub_c_void"
[(set (reg:CC_C_GTU CC_REGNUM)
(compare:CC_C_GTU
(minus:SI (match_operand:SI 1 "gpr_operand" "r,r")
(match_operand:SI 2 "arith_operand" "r,L"))
(match_dup 1)))
(clobber (match_scratch:SI 0 "=r,r"))]
""
"sub %0,%1,%2"
[(set_attr "type" "compare")])
(define_code_iterator logical_op
[and ior xor])
(define_code_attr op_mnc
[(plus "add") (minus "sub") (and "and") (ior "orr") (xor "eor")])
(define_insn "*_f"
[(set (reg:CC CC_REGNUM)
(compare:CC (logical_op:SI (match_operand:SI 1 "gpr_operand" "%r")
(match_operand:SI 2 "gpr_operand" "r"))
(const_int 0)))
(set (match_operand:SI 0 "gpr_operand" "=r")
(logical_op:SI (match_dup 1) (match_dup 2)))]
""
" %0,%1,%2"
[(set_attr "type" "compare")])
(define_insn_and_split "*mov_f"
[(set (reg:CC CC_REGNUM)
(compare:CC (match_operand:SI 1 "gpr_operand" "r") (const_int 0)))
(set (match_operand:SI 0 "gpr_operand" "=r") (match_dup 1))]
""
"#"
"reload_completed"
[(parallel
[(set (reg:CC CC_REGNUM)
(compare:CC (and:SI (match_dup 1) (match_dup 1)) (const_int 0)))
(set (match_operand:SI 0 "gpr_operand" "=r")
(and:SI (match_dup 1) (match_dup 1)))])]
""
[(set_attr "type" "compare")])
(define_peephole2
[(parallel
[(set (match_operand:SI 0 "gpr_operand")
(logical_op:SI (match_operand:SI 1 "gpr_operand")
(match_operand:SI 2 "gpr_operand")))
(clobber (reg:CC CC_REGNUM))])
(parallel
[(set (reg:CC CC_REGNUM)
(compare:CC (and:SI (match_dup 0) (match_dup 0)) (const_int 0)))
(set (match_operand:SI 3 "gpr_operand")
(and:SI (match_dup 0) (match_dup 0)))])]
"peep2_reg_dead_p (2, operands[0])"
[(parallel
[(set (reg:CC CC_REGNUM)
(compare:CC (logical_op:SI (match_dup 1) (match_dup 2))
(const_int 0)))
(set (match_dup 3) (logical_op:SI (match_dup 1) (match_dup 2)))])])
(define_peephole2
[(parallel
[(set (match_operand:SI 0 "gpr_operand")
(logical_op:SI (match_operand:SI 1 "gpr_operand")
(match_operand:SI 2 "gpr_operand")))
(clobber (reg:CC CC_REGNUM))])
(parallel
[(set (reg:CC CC_REGNUM)
(compare:CC (and:SI (match_dup 0) (match_dup 0)) (const_int 0)))
(set (match_operand:SI 3 "gpr_operand")
(and:SI (match_dup 0) (match_dup 0)))])]
"peep2_reg_dead_p (2, operands[3])"
[(parallel
[(set (reg:CC CC_REGNUM)
(compare:CC (logical_op:SI (match_dup 1) (match_dup 2))
(const_int 0)))
(set (match_dup 0) (logical_op:SI (match_dup 1) (match_dup 2)))])])
(define_peephole2
[(parallel
[(set (match_operand:SI 0 "gpr_operand")
(logical_op:SI (match_operand:SI 1 "gpr_operand")
(match_operand:SI 2 "gpr_operand")))
(clobber (reg:CC CC_REGNUM))])
(parallel
[(set (reg:CC CC_REGNUM)
(compare:CC (match_dup 0) (const_int 0)))
(clobber (match_operand:SI 3 "gpr_operand"))])]
""
[(parallel
[(set (reg:CC CC_REGNUM)
(compare:CC (logical_op:SI (match_dup 1) (match_dup 2))
(const_int 0)))
(set (match_dup 0) (logical_op:SI (match_dup 1) (match_dup 2)))])])
(define_expand "cstoresi4"
[(parallel
[(set (reg:CC CC_REGNUM)
(match_operand:SI 1 "comparison_operator"))
(match_operand:SI 2 "" "")])
(set (match_dup 0) (match_operand:SI 3 "arith_operand" ""))
(set (match_operand:SI 0 "gpr_operand" "=r")
(if_then_else:SI (match_dup 4) (match_dup 5) (match_dup 0)))]
""
{
enum rtx_code o2_code = GET_CODE (operands[2]);
enum rtx_code cmp_code = GET_CODE (operands[1]);
if ((o2_code == AND || o2_code == IOR || o2_code == XOR)
&& operands[3] == const0_rtx)
{
operands[2] = copy_rtx(operands[2]);
XEXP (operands[2], 0) = force_reg (SImode, XEXP (operands[2], 0));
XEXP (operands[2], 1) = force_reg (SImode, XEXP (operands[2], 1));
}
else
operands[2] = force_reg (SImode, operands[2]);
operands[1] = gen_rtx_COMPARE (CCmode, operands[2], operands[3]);
if (cmp_code != NE)
{
operands[2] = gen_rtx_CLOBBER (VOIDmode, gen_rtx_SCRATCH (SImode));
operands[3] = const0_rtx;
}
else
{
if (operands[3] != const0_rtx)
operands[2] = gen_rtx_MINUS (SImode, operands[2], operands[3]);
operands[2] = gen_rtx_SET (operands[0], operands[2]);
operands[3] = operands[0];
}
operands[4] = gen_rtx_fmt_ee (cmp_code, SImode,
gen_rtx_REG (CCmode, CC_REGNUM), const0_rtx);
operands[5] = force_reg (SImode, GEN_INT (STORE_FLAG_VALUE));
})
; floating point comparisons
(define_insn "*cmpsf_cc_insn"
[(match_parallel 3 "float_operation"
[(set (reg:CC_FP CCFP_REGNUM)
(compare:CC_FP (match_operand:SF 0 "gpr_operand" "r")
(match_operand:SF 1 "gpr_operand" "r")))
(clobber (match_scratch:SF 2 "=r"))])]
"!TARGET_SOFT_CMPSF"
"fsub %2,%0,%1"
[(set_attr "type" "fp")
(set_attr "fp_mode" "round_unknown")])
;; ??? do we have to relax the operand0 predicate to immediate_operand
;; to allow the rtl loop optimizer to generate comparisons? OTOH
;; we want call_address_operand to enforce valid operands so that
;; combine won't do silly things, allowing instruction scheduling to do
;; a proper job.
(define_insn "*cmpsf_eq"
[(set (reg:CC_FP_EQ CC_REGNUM) (compare:CC_FP_EQ (reg:SF 0) (reg:SF 1)))
(use (match_operand:SI 0 "call_address_operand" "Csy,r"))
(clobber (reg:SI GPR_IP))
(clobber (reg:SI GPR_LR))]
"TARGET_SOFT_CMPSF"
"%f0"
[(set_attr "type" "sfunc")])
(define_insn "*cmpsf_gte"
[(set (reg:CC_FP_GTE CC_REGNUM) (compare:CC_FP_GTE (reg:SF 0) (reg:SF 1)))
(use (match_operand:SI 0 "call_address_operand" "Csy,r"))
(clobber (reg:SI GPR_IP))
(clobber (reg:SI GPR_LR))]
"TARGET_SOFT_CMPSF"
"%f0"
[(set_attr "type" "sfunc")])
(define_insn "*cmpsf_ord"
[(set (reg:CC_FP_ORD CC_REGNUM) (compare:CC_FP_ORD (reg:SF 0) (reg:SF 1)))
(use (match_operand:SI 0 "call_address_operand" "Csy,r"))
(clobber (reg:SI GPR_IP))
(clobber (reg:SI GPR_16))
(clobber (reg:SI GPR_LR))]
""
"%f0"
[(set_attr "type" "sfunc")])
(define_insn "*cmpsf_uneq"
[(set (reg:CC_FP_UNEQ CC_REGNUM) (compare:CC_FP_UNEQ (reg:SF 0) (reg:SF 1)))
(use (match_operand:SI 0 "call_address_operand" "Csy,r"))
(clobber (reg:SI GPR_IP))
(clobber (reg:SI GPR_16))
(clobber (reg:SI GPR_LR))]
"TARGET_SOFT_CMPSF"
"%f0"
[(set_attr "type" "sfunc")])
;; conditional moves
(define_expand "movcc"
[(set (match_operand:WMODE 0 "gpr_operand" "")
(if_then_else:WMODE (match_operand 1 "comparison_operator" "")
(match_operand:WMODE 2 "gpr_operand" "")
(match_operand:WMODE 3 "gpr_operand" "")))]
"TARGET_CMOVE"
{
rtx cmp_op0 = XEXP (operands[1], 0);
rtx cmp_op1 = XEXP (operands[1], 1);
machine_mode cmp_in_mode;
enum rtx_code code = GET_CODE (operands[1]);
cmp_in_mode = GET_MODE (cmp_op0);
if (cmp_in_mode == VOIDmode)
cmp_in_mode = GET_MODE (cmp_op1);
if (cmp_in_mode == VOIDmode)
cmp_in_mode = SImode;
/* If the operands are a better match when reversed, swap them now.
This allows combine to see the proper comparison codes. */
if (rtx_equal_p (operands[0], operands[2])
&& !rtx_equal_p (operands[0], operands[3]))
{
rtx tmp = operands[2]; operands[2] = operands[3]; operands[3] = tmp;
code = (FLOAT_MODE_P (GET_MODE (cmp_op0)) && !flag_finite_math_only
? reverse_condition_maybe_unordered (code)
: reverse_condition (code));
}
if (proper_comparison_operator (operands[1], VOIDmode))
operands[1] = gen_rtx_fmt_ee (code, cmp_in_mode, cmp_op0, cmp_op1);
else
{
if (!currently_expanding_to_rtl)
{
/* ??? It would seem safest to FAIL here, but that would defeat
the purpose of having an if-conversion pass; its logic currently
assumes that the backend should be safe to insert condition code
setting instructions, as the same condition codes were presumably
set by the if-conversion input code. */
}
/* What mode to give as first operand to gen_compare_reg here is
debatable. VOIDmode would be minimalist; telling gen_compare_reg
to use the mode of CC_REGNUM (or putting it on the comparison
operator afterwards) is also a logical choice. OTOH, by using
mode, we have mode combine opportunities with flag setting
operations - if we get some. */
operands[1]
= gen_compare_reg (mode, code, cmp_in_mode, cmp_op0, cmp_op1);
if (!operands[1])
FAIL;
}
})
(define_insn "*movcc_insn"
[(set (match_operand:WMODE 0 "gpr_operand" "=r")
(if_then_else:WMODE (match_operator 3 "proper_comparison_operator"
[(match_operand 4 "cc_operand") (const_int 0)])
(match_operand:WMODE 1 "gpr_operand" "r")
(match_operand:WMODE 2 "gpr_operand" "0")))]
"TARGET_CMOVE"
"mov%d3 %0,%1"
[(set_attr "type" "cmove")])
(define_peephole2
[(parallel [(set (match_operand:WMODE 0 "gpr_operand" "")
(match_operand:WMODE 1 "" ""))
(clobber (match_operand 8 "cc_operand"))])
(match_operand 2 "" "")
(set (match_operand:WMODE2 3 "gpr_operand" "")
(match_operand:WMODE2 9 "gpr_operand" ""))
(set (match_dup 3)
(if_then_else:WMODE2 (match_operator 5 "proper_comparison_operator"
[(match_operand 6 "cc_operand")
(match_operand 7 "const0_operand")])
(match_operand:WMODE2 4 "nonmemory_operand" "")
(match_dup 3)))]
"REGNO (operands[0]) == REGNO (operands[9])
&& peep2_reg_dead_p (3, operands[0])
&& !reg_set_p (operands[0], operands[2])
&& !reg_set_p (operands[3], operands[2])
&& !reg_overlap_mentioned_p (operands[3], operands[2])"
[(parallel [(set (match_dup 10) (match_dup 1))
(clobber (match_dup 8))])
(match_dup 2)
(set (match_dup 3)
(if_then_else:WMODE2 (match_dup 5) (match_dup 4) (match_dup 3)))]
{
operands[10] = simplify_gen_subreg (mode, operands[3],
mode, 0);
replace_rtx (operands[2], operands[9], operands[3], true);
replace_rtx (operands[2], operands[0], operands[10], true);
gcc_assert (!reg_overlap_mentioned_p (operands[0], operands[2]));
})
(define_peephole2
[(parallel [(set (match_operand 6 "cc_operand") (match_operand 2 "" ""))
(set (match_operand:WMODE 0 "gpr_operand" "")
(match_operand:WMODE 1 "" ""))])
(set (match_operand:WMODE2 3 "gpr_operand" "")
(match_operand:WMODE2 4 "gpr_operand"))
(set (match_dup 3)
(if_then_else:WMODE2 (match_operator 5 "proper_comparison_operator"
[(match_dup 6)
(match_operand:WMODE 7 "const0_operand")])
(match_operand:WMODE2 8 "gpr_operand")
(match_dup 3)))]
"REGNO (operands[0]) == REGNO (operands[8])
&& REVERSIBLE_CC_MODE (GET_MODE (operands[6]))
&& peep2_reg_dead_p (3, operands[6])
&& peep2_reg_dead_p (3, operands[0])
&& !reg_overlap_mentioned_p (operands[4], operands[3])"
[(parallel [(set (match_dup 6) (match_dup 2))
(set (match_dup 9) (match_dup 1))])
(set (match_dup 3)
(if_then_else:WMODE2 (match_dup 5) (match_dup 4) (match_dup 3)))]
"
{
operands[5]
= gen_rtx_fmt_ee (REVERSE_CONDITION (GET_CODE (operands[5]),
GET_MODE (operands[6])),
GET_MODE (operands[5]), operands[6], operands[7]);
operands[9] = simplify_gen_subreg (mode, operands[3],
mode, 0);
}")
;; These control RTL generation for conditional jump insns
;; To signal to can_compare_p that the cbranchs?4 patterns work,
;; they must allow const0_rtx for both comparison operands
(define_expand "cbranchsi4"
[(set (reg CC_REGNUM)
(compare (match_operand:SI 1 "add_operand" "")
(match_operand:SI 2 "arith_operand" "")))
(set (pc)
(if_then_else
(match_operator 0 "ordered_comparison_operator" [(reg CC_REGNUM)
(const_int 0)])
(label_ref (match_operand 3 "" ""))
(pc)))]
""
{
rtx cmp = gen_compare_reg (VOIDmode, GET_CODE (operands[0]), SImode,
operands[1], operands[2]);
emit_jump_insn (gen_branch_insn (operands[3], cmp, XEXP (cmp, 0)));
DONE;
})
(define_expand "cbranchsf4"
[(set (reg CC_REGNUM)
(compare (match_operand:SF 1 "arith_operand" "")
(match_operand:SF 2 "arith_operand" "")))
(set (pc)
(if_then_else
(match_operator 0 "comparison_operator" [(reg CC_REGNUM)
(const_int 0)])
(label_ref (match_operand 3 "" ""))
(pc)))]
""
{
rtx cmp = gen_compare_reg (VOIDmode, GET_CODE (operands[0]), SFmode,
operands[1], operands[2]);
emit_jump_insn (gen_branch_insn (operands[3], cmp, XEXP (cmp, 0)));
DONE;
})
;; Now match both normal and inverted jump.
(define_insn "branch_insn"
[(set (pc)
(if_then_else (match_operator 1 "proper_comparison_operator"
[(match_operand 2 "cc_operand")
(const_int 0)])
(label_ref (match_operand 0 "" ""))
(pc)))]
""
"b%d1 %l0"
[(set_attr "type" "branch")])
(define_insn "*rev_branch_insn"
[(set (pc)
(if_then_else (match_operator 1 "proper_comparison_operator"
[(reg CC_REGNUM) (const_int 0)])
(pc)
(label_ref (match_operand 0 "" ""))))]
""
"b%D1 %l0"
[(set_attr "type" "branch")])
;; Unconditional and other jump instructions.
(define_insn "jump"
[(set (pc) (label_ref (match_operand 0 "" "")))]
""
"b %l0"
[(set_attr "type" "uncond_branch")])
(define_insn "indirect_jump"
[(set (pc) (match_operand:SI 0 "gpr_operand" "r"))]
""
"jr %0"
[(set_attr "type" "uncond_branch")])
(define_expand "tablejump"
[(parallel [(set (pc) (match_operand:SI 0 "gpr_operand" ""))
(use (label_ref (match_operand 1 "" "")))])]
""
{
/* In PIC mode, the table entries are stored PC relative.
Convert the relative address to an absolute address. */
if (flag_pic)
{
rtx op1 = gen_rtx_LABEL_REF (Pmode, operands[1]);
operands[0] = expand_simple_binop (Pmode, PLUS, operands[0],
op1, NULL_RTX, 0, OPTAB_DIRECT);
}
})
(define_insn "*tablejump_internal"
[(set (pc) (match_operand:SI 0 "gpr_operand" "r"))
(use (label_ref (match_operand 1 "" "")))]
""
"jr %0;"
[(set_attr "type" "uncond_branch")])
(define_insn "*tablejump_hi_internal"
[(set (pc) (match_operand:HI 0 "gpr_operand" "r"))
(use (label_ref (match_operand 1 "" "")))]
"optimize_size && TARGET_SMALL16"
"jr %0;"
[(set_attr "type" "uncond_branch")])
(define_expand "call"
;; operands[1] is stack_size_rtx
;; operands[2] is next_arg_register
[(parallel [(call (match_operand:SI 0 "call_operand" "")
(match_operand 1 "" ""))
(clobber (reg:SI GPR_LR))])]
""
{
bool target_uninterruptible = epiphany_call_uninterruptible_p (operands[0]);
if (!call_operand (operands[1], VOIDmode))
operands[0]
= change_address (operands[0], VOIDmode,
copy_to_mode_reg (Pmode, XEXP (operands[0], 0)));
if (epiphany_uninterruptible_p (current_function_decl)
!= target_uninterruptible)
{
emit_insn (target_uninterruptible ? gen_gid (): gen_gie ());
emit_call_insn
(gen_rtx_PARALLEL
(VOIDmode,
gen_rtvec (2, gen_rtx_CALL (VOIDmode, operands[0], operands[1]),
gen_rtx_CLOBBER (VOIDmode,
gen_rtx_REG (SImode, GPR_LR)))));
emit_insn (target_uninterruptible ? gen_gie (): gen_gid ());
DONE;
}
})
(define_insn "*call_i"
[(match_parallel 2 "float_operation"
[(call (mem:SI (match_operand:SI 0 "call_address_operand" "Csy,r"))
(match_operand 1 "" ""))
(clobber (reg:SI GPR_LR))])]
""
"%f0"
[(set_attr "type" "call")])
(define_expand "sibcall"
;; operands[1] is stack_size_rtx
;; operands[2] is next_arg_register
[(parallel [(call (match_operand:SI 0 "call_operand" "")
(match_operand 1 "" ""))
(return)])]
""
{
bool target_uninterruptible = epiphany_call_uninterruptible_p (operands[0]);
if (!call_operand (operands[1], VOIDmode))
operands[0]
= change_address (operands[0], VOIDmode,
copy_to_mode_reg (Pmode, XEXP (operands[0], 0)));
if (epiphany_uninterruptible_p (current_function_decl)
!= target_uninterruptible)
{
emit_insn (target_uninterruptible ? gen_gid (): gen_gie ());
emit_call_insn
(gen_rtx_PARALLEL
(VOIDmode,
gen_rtvec (2, gen_rtx_CALL (VOIDmode, operands[0], operands[1]),
ret_rtx)));
emit_insn (target_uninterruptible ? gen_gie (): gen_gid ());
DONE;
}
})
(define_insn "*sibcall_i"
[(call (mem:SI (match_operand:SI 0 "call_address_operand" "Csy,Rsc"))
(match_operand 1 "" ""))
(return)]
""
"@
b %0
jr %0"
[(set_attr "type" "call")])
(define_expand "call_value"
;; operand 2 is stack_size_rtx
;; operand 3 is next_arg_register
[(parallel [(set (match_operand 0 "gpr_operand" "=r")
(call (match_operand:SI 1 "call_operand" "")
(match_operand 2 "" "")))
(clobber (reg:SI GPR_LR))])]
""
{
bool target_uninterruptible = epiphany_call_uninterruptible_p (operands[1]);
if (!call_operand (operands[1], VOIDmode))
operands[1]
= change_address (operands[1], VOIDmode,
copy_to_mode_reg (Pmode, XEXP (operands[1], 0)));
if (epiphany_uninterruptible_p (current_function_decl)
!= target_uninterruptible)
{
emit_insn (target_uninterruptible ? gen_gid (): gen_gie ());
emit_call_insn
(gen_rtx_PARALLEL
(VOIDmode,
gen_rtvec (2, gen_rtx_SET
(operands[0],
gen_rtx_CALL (VOIDmode, operands[1], operands[2])),
gen_rtx_CLOBBER (VOIDmode,
gen_rtx_REG (SImode, GPR_LR)))));
emit_insn (target_uninterruptible ? gen_gie (): gen_gid ());
DONE;
}
})
(define_insn "*call_value_i"
[(match_parallel 3 "float_operation"
[(set (match_operand 0 "gpr_operand" "=r,r")
(call (mem:SI (match_operand:SI 1 "call_address_operand" "Csy,r"))
(match_operand 2 "" "")))
(clobber (reg:SI GPR_LR))])]
""
"%f1"
[(set_attr "type" "call")
(set_attr "length" "4")])
(define_expand "sibcall_value"
;; operand 2 is stack_size_rtx
;; operand 3 is next_arg_register
[(parallel [(set (match_operand 0 "gpr_operand" "=r")
(call (match_operand:SI 1 "call_operand" "")
(match_operand 2 "" "")))
(return)])]
""
{
bool target_uninterruptible = epiphany_call_uninterruptible_p (operands[1]);
if (!call_operand (operands[1], VOIDmode))
operands[1]
= change_address (operands[1], VOIDmode,
copy_to_mode_reg (Pmode, XEXP (operands[1], 0)));
if (epiphany_uninterruptible_p (current_function_decl)
!= target_uninterruptible)
{
emit_insn (target_uninterruptible ? gen_gid (): gen_gie ());
emit_call_insn
(gen_rtx_PARALLEL
(VOIDmode,
gen_rtvec (2, gen_rtx_SET
(operands[0],
gen_rtx_CALL (VOIDmode, operands[1], operands[2])),
ret_rtx)));
emit_insn (target_uninterruptible ? gen_gie (): gen_gid ());
DONE;
}
})
(define_insn "*sibcall_value_i"
[(set (match_operand 0 "gpr_operand" "=r,r")
(call (mem:SI (match_operand:SI 1 "call_address_operand" "Csy,Rsc"))
(match_operand 2 "" "")))
(return)]
""
"@
b %1
jr %1"
[(set_attr "type" "call")
(set_attr "length" "4")])
(define_expand "prologue"
[(pc)]
""
{
epiphany_expand_prologue ();
DONE;
})
(define_expand "epilogue"
[(pc)]
""
{
epiphany_expand_epilogue (0);
DONE;
})
(define_expand "sibcall_epilogue"
[(pc)]
""
{
epiphany_expand_epilogue (1);
DONE;
})
; Since the demise of REG_N_SETS, it is no longer possible to find out
; in the prologue / epilogue expanders how many times lr is set.
; Using df_regs_ever_live_p to decide if lr needs saving means that
; any explicit use of lr will cause it to be saved; hence we cannot
; represent the blink use in return / sibcall instructions themselves, and
; instead have to show it in EPILOGUE_USES.
(define_insn "return_i"
[(return)]
"reload_completed"
"rts"
[(set_attr "type" "uncond_branch")])
(define_insn "return_internal_interrupt"
[(return)
(unspec_volatile [(const_int 0)] 1)]
""
"rti"
[(set_attr "type" "uncond_branch")])
(define_insn "stack_adjust_add"
[(set (reg:SI GPR_SP)
(plus:SI (reg:SI GPR_SP) (match_operand:SI 0 "arith_operand" "rL")))
(clobber (reg:CC CC_REGNUM))
(clobber (reg:SI STATUS_REGNUM))
(clobber (match_operand:BLK 1 "memclob_operand" "=X"))]
"reload_completed"
"add sp,sp,%0")
(define_insn "stack_adjust_mov"
[(set (reg:SI GPR_SP) (reg:SI GPR_FP))
(clobber (match_operand:BLK 0 "memory_operand" "=m"))]
"reload_completed"
"mov sp,fp"
[(set_attr "type" "move")])
(define_insn "stack_adjust_str"
[(set (match_operand 0 "stacktop_operand" "=m")
(match_operand 1 "any_gpr_operand" "r"))
(set (reg:SI GPR_SP)
(plus:SI (reg:SI GPR_SP) (match_operand:SI 2 "nonmemory_operand" "rn")))
(clobber (match_operand:BLK 3 "memclob_operand" "=X"))]
"reload_completed"
{
return (GET_MODE_SIZE (GET_MODE (operands[0])) <= 4
? \"str %1,%0,%C2\" : \"strd %1,%0,%X2\");
}
[(set_attr "type" "store")])
(define_insn "stack_adjust_ldr"
[(set (match_operand:SI 0 "gpr_operand" "=r")
(match_operand:SI 1 "stacktop_operand" "m"))
(set (reg:SI GPR_SP)
(plus:SI (reg:SI GPR_SP) (match_operand:SI 2 "nonmemory_operand" "rn")))
(clobber (match_operand:BLK 3 "memory_operand" "=m"))]
"reload_completed"
"ldr %0,%1,%C2"
[(set_attr "type" "load")])
;; Define some fake vector operations so that the vectorizer is happy to use
;; 64 bit loads/stores.
(define_expand "vec_unpacks_lo_v4hi"
[(match_operand:V2SI 0 "gpr_operand")
(match_operand:V4HI 1 "gpr_operand")]
""
{
rtx in = simplify_gen_subreg (SImode, operands[1], V4HImode, 0);
rtx outl = simplify_gen_subreg (SImode, operands[0], V2SImode, 0);
rtx outh
= simplify_gen_subreg (SImode, operands[0], V2SImode, UNITS_PER_WORD);
if (reg_overlap_mentioned_p (outl, in))
in = copy_to_mode_reg (SImode, in);
emit_insn (gen_ashlsi3 (outl, in, GEN_INT (16)));
emit_insn (gen_ashrsi3 (outl, outl, GEN_INT (16)));
emit_insn (gen_ashrsi3 (outh, in, GEN_INT (16)));
DONE;
})
(define_expand "vec_unpacks_hi_v4hi"
[(match_operand:V2SI 0 "gpr_operand")
(match_operand:V4HI 1 "gpr_operand")]
""
{
rtx in = simplify_gen_subreg (SImode, operands[1], V4HImode, UNITS_PER_WORD);
rtx outl = simplify_gen_subreg (SImode, operands[0], V2SImode, 0);
rtx outh
= simplify_gen_subreg (SImode, operands[0], V2SImode, UNITS_PER_WORD);
if (reg_overlap_mentioned_p (outl, in))
in = copy_to_mode_reg (SImode, in);
emit_insn (gen_ashlsi3 (outl, in, GEN_INT (16)));
emit_insn (gen_ashrsi3 (outl, outl, GEN_INT (16)));
emit_insn (gen_ashrsi3 (outh, in, GEN_INT (16)));
DONE;
})
(define_code_iterator addsub [plus minus])
(define_code_iterator alu_binop
[plus minus and ior xor])
(define_code_attr insn_opname
[(plus "add") (minus "sub") (mult "mul") (div "div")
(and "and") (ior "ior") (xor "xor")])
; The addsi3 / subsi3 do checks that we don't want when splitting V2SImode
; operations into two SImode operations.
(define_code_attr si_pattern_suffix
[(plus "_i") (minus "_i") (and "") (ior "") (xor "")])
; You might think that this would work better as a define_expand, but
; again lower_subreg pessimizes the code if it sees indiviudual operations.
; We need to keep inputs and outputs as register pairs if we want to
; get sensible register allocation for double-word load and store operations.
(define_insn_and_split "v2si3"
[(set (match_operand:V2SI 0 "gpr_operand" "=r")
(alu_binop:V2SI (match_operand:V2SI 1 "gpr_operand" "r")
(match_operand:V2SI 2 "gpr_operand" "r")))
(clobber (reg:CC CC_REGNUM))]
""
"#"
"reload_completed || (epiphany_vect_align == 4 && TARGET_SPLIT_VECMOVE_EARLY)"
[(const_int 0)]
{
rtx o0l, o0h, o1l, o1h, o2l, o2h;
o0l = simplify_gen_subreg (SImode, operands[0], V2SImode, 0);
o0h = simplify_gen_subreg (SImode, operands[0], V2SImode, UNITS_PER_WORD);
o1l = simplify_gen_subreg (SImode, operands[1], V2SImode, 0);
o1h = simplify_gen_subreg (SImode, operands[1], V2SImode, UNITS_PER_WORD);
o2l = simplify_gen_subreg (SImode, operands[2], V2SImode, 0);
o2h = simplify_gen_subreg (SImode, operands[2], V2SImode, UNITS_PER_WORD);
if (reg_overlap_mentioned_p (o0l, o1h))
o1h = copy_to_mode_reg (SImode, o1h);
if (reg_overlap_mentioned_p (o0l, o2h))
o2h = copy_to_mode_reg (SImode, o2h);
emit_insn (gen_si3 (o0l, o1l, o2l));
emit_insn (gen_si3 (o0h, o1h, o2h));
DONE;
}
[(set_attr "length" "8")])
(define_expand "v2sf3"
[(parallel
[(set (match_operand:V2SF 0 "gpr_operand" "")
(addsub:V2SF (match_operand:V2SF 1 "gpr_operand" "")
(match_operand:V2SF 2 "gpr_operand" "")))
(clobber (reg:CC_FP CCFP_REGNUM))])])
(define_insn_and_split "v2sf3_i"
[(match_parallel 3 "float_operation"
[(set (match_operand:V2SF 0 "gpr_operand" "=r")
(addsub:V2SF (match_operand:V2SF 1 "gpr_operand" "r")
(match_operand:V2SF 2 "gpr_operand" "r")))
(clobber (reg:CC_FP CCFP_REGNUM))])]
""
"#"
"reload_completed || (epiphany_vect_align == 4 && TARGET_SPLIT_VECMOVE_EARLY)"
[(parallel
[(set (match_dup 4) (addsub:SF (match_dup 5) (match_dup 6)))
(clobber (reg:CC_FP CCFP_REGNUM))
(match_dup 10)
(match_dup 11)])
(parallel
[(set (match_dup 7) (addsub:SF (match_dup 8) (match_dup 9)))
(clobber (reg:CC_FP CCFP_REGNUM))
(match_dup 10)
(match_dup 11)])]
{
operands[4] = simplify_gen_subreg (SFmode, operands[0], V2SFmode, 0);
operands[5] = simplify_gen_subreg (SFmode, operands[1], V2SFmode, 0);
operands[6] = simplify_gen_subreg (SFmode, operands[2], V2SFmode, 0);
operands[7]
= simplify_gen_subreg (SFmode, operands[0], V2SFmode, UNITS_PER_WORD);
operands[8]
= simplify_gen_subreg (SFmode, operands[1], V2SFmode, UNITS_PER_WORD);
operands[9]
= simplify_gen_subreg (SFmode, operands[2], V2SFmode, UNITS_PER_WORD);
if (!reload_completed)
{
if (reg_overlap_mentioned_p (operands[4], operands[8]))
operands[8] = copy_to_mode_reg (SFmode, operands[8]);
if (reg_overlap_mentioned_p (operands[4], operands[9]))
operands[9] = copy_to_mode_reg (SFmode, operands[9]);
emit_insn (gen_sf3 (operands[4], operands[5], operands[6]));
emit_insn (gen_sf3 (operands[7], operands[8], operands[9]));
DONE;
}
gcc_assert (!reg_overlap_mentioned_p (operands[4], operands[8]));
gcc_assert (!reg_overlap_mentioned_p (operands[4], operands[9]));
operands[10] = XVECEXP (operands[3], 0, XVECLEN (operands[3], 0) - 2);
operands[11] = XVECEXP (operands[3], 0, XVECLEN (operands[3], 0) - 1);
}
[(set_attr "length" "8")
(set_attr "type" "v2fp")])
(define_expand "ashlv2si3"
[(parallel
[(set (match_operand:V2SI 0 "gpr_operand" "")
(ashift:V2SI (match_operand:V2SI 1 "gpr_operand" "")
(match_operand:SI 2 "general_operand")))
(use (match_dup 3))
(clobber (reg:CC_FP CCFP_REGNUM))])]
""
{
if (const_int_operand (operands[2], VOIDmode))
operands[3]
= copy_to_mode_reg (SImode, GEN_INT (1 << INTVAL (operands[2])));
else
{
int o, i;
rtx xop[2], last_out = pc_rtx;
for (o = 0; o <= UNITS_PER_WORD; o += UNITS_PER_WORD)
{
for (i = 0; i < 2; i++)
{
xop[i]
= (i == 2 ? operands[2]
: simplify_gen_subreg (SImode, operands[i], V2SImode, o));
gcc_assert (!reg_overlap_mentioned_p (last_out, xop[i])
/* ??? reg_overlap_mentioned_p doesn't understand
about multi-word SUBREGs. */
|| (GET_CODE (last_out) == SUBREG
&& GET_CODE (xop[i]) == SUBREG
&& SUBREG_REG (last_out) == SUBREG_REG (xop[i])
&& ((SUBREG_BYTE (last_out) & -UNITS_PER_WORD)
!= (SUBREG_BYTE (xop[i]) & -UNITS_PER_WORD))));
}
emit_insn (gen_ashlsi3 (xop[0], xop[1], operands[2]));
last_out = xop[0];
}
DONE;
}
})
(define_insn_and_split "*ashlv2si3_i"
[(match_parallel 3 "float_operation"
[(set (match_operand:V2SI 0 "gpr_operand" "=&r,*1*2")
(ashift:V2SI (match_operand:V2SI 1 "gpr_operand" "r,r")
(match_operand 2 "const_int_operand" "n,n")))
(use (match_operand:SI 4 "gpr_operand" "r,r"))
(clobber (reg:CC_FP CCFP_REGNUM))])]
""
"#"
"reload_completed"
[(parallel
[(set (match_dup 5) (mult:SI (match_dup 6) (match_dup 4)))
(clobber (reg:CC_FP CCFP_REGNUM))
(match_dup 9)
(match_dup 10)])
(parallel
[(set (match_dup 7) (mult:SI (match_dup 8) (match_dup 4)))
(clobber (reg:CC_FP CCFP_REGNUM))
(match_dup 9)
(match_dup 10)])]
{
operands[5] = simplify_gen_subreg (SImode, operands[0], V2SImode, 0);
operands[6] = simplify_gen_subreg (SImode, operands[1], V2SImode, 0);
operands[7] = simplify_gen_subreg (SImode, operands[0],
V2SImode, UNITS_PER_WORD);
operands[8] = simplify_gen_subreg (SImode, operands[1],
V2SImode, UNITS_PER_WORD);
gcc_assert (!reg_overlap_mentioned_p (operands[5], operands[8]));
gcc_assert (!reg_overlap_mentioned_p (operands[5], operands[4]));
operands[9] = XVECEXP (operands[3], 0, XVECLEN (operands[3], 0) - 2);
operands[10] = XVECEXP (operands[3], 0, XVECLEN (operands[3], 0) - 1);
rtx insn
= (gen_rtx_PARALLEL
(VOIDmode,
gen_rtvec
(4,
gen_rtx_SET (operands[5],
gen_rtx_MULT (SImode, operands[6], operands[4])),
gen_rtx_CLOBBER (VOIDmode, gen_rtx_REG (CC_FPmode, CCFP_REGNUM)),
operands[9], operands[10])));
insn = emit_insn (insn);
add_reg_note (insn, REG_EQUAL,
gen_rtx_ASHIFT (SImode, operands[6], operands[2]));
insn
= (gen_rtx_PARALLEL
(VOIDmode,
gen_rtvec
(4,
gen_rtx_SET (operands[7],
gen_rtx_MULT (SImode, operands[8], operands[4])),
gen_rtx_CLOBBER (VOIDmode, gen_rtx_REG (CC_FPmode, CCFP_REGNUM)),
operands[9], operands[10])));
insn = emit_insn (insn);
add_reg_note (insn, REG_EQUAL,
gen_rtx_ASHIFT (SImode, operands[7], operands[2]));
DONE;
}
[(set_attr "length" "8")
(set_attr "type" "fp_int")])
(define_expand "mul3"
[(parallel
[(set (match_operand:DWV2MODE 0 "gpr_operand" "")
(mult:DWV2MODE (match_operand:DWV2MODE 1 "gpr_operand" "")
(match_operand:DWV2MODE 2 "gpr_operand" "")))
(clobber (reg:CC_FP CCFP_REGNUM))])])
(define_insn_and_split "mul3_i"
[(match_parallel 3 "float_operation"
[(set (match_operand:DWV2MODE 0 "gpr_operand" "=r")
(mult:DWV2MODE (match_operand:DWV2MODE 1 "gpr_operand" "r")
(match_operand:DWV2MODE 2 "gpr_operand" "r")))
(clobber (reg:CC_FP CCFP_REGNUM))])]
""
"#"
"reload_completed || (epiphany_vect_align == 4 && TARGET_SPLIT_VECMOVE_EARLY)"
[(parallel
[(set (match_dup 4) (mult: (match_dup 5) (match_dup 6)))
(clobber (reg:CC_FP CCFP_REGNUM))
(match_dup 10)
(match_dup 11)])
(parallel
[(set (match_dup 7) (mult: (match_dup 8) (match_dup 9)))
(clobber (reg:CC_FP CCFP_REGNUM))
(match_dup 10)
(match_dup 11)])]
{
operands[4]
= simplify_gen_subreg (mode, operands[0], mode, 0);
operands[5]
= simplify_gen_subreg (mode, operands[1], mode, 0);
operands[6]
= simplify_gen_subreg (mode, operands[2], mode, 0);
operands[7] = simplify_gen_subreg (mode, operands[0],
mode, UNITS_PER_WORD);
operands[8] = simplify_gen_subreg (mode, operands[1],
mode, UNITS_PER_WORD);
operands[9] = simplify_gen_subreg (mode, operands[2],
mode, UNITS_PER_WORD);
if (!reload_completed)
{
if (reg_overlap_mentioned_p (operands[4], operands[8]))
operands[8] = copy_to_mode_reg (mode, operands[8]);
if (reg_overlap_mentioned_p (operands[4], operands[9]))
operands[9] = copy_to_mode_reg (mode, operands[9]);
emit_insn (gen_mul3 (operands[4], operands[5], operands[6]));
emit_insn (gen_mul3 (operands[7], operands[8], operands[9]));
DONE;
}
gcc_assert (!reg_overlap_mentioned_p (operands[4], operands[8]));
gcc_assert (!reg_overlap_mentioned_p (operands[4], operands[9]));
operands[10] = XVECEXP (operands[3], 0, XVECLEN (operands[3], 0) - 2);
operands[11] = XVECEXP (operands[3], 0, XVECLEN (operands[3], 0) - 1);
}
[(set_attr "length" "8")
(set_attr "type" "")])
(define_insn_and_split "*fmadd_combine"
[(match_parallel 4 "float_operation"
[(set (match_operand:DWV2MODE 0 "gpr_operand" "=r")
(plus:DWV2MODE (mult:
(match_operand: 1 "gpr_operand" "r")
(match_operand: 2 "gpr_operand" "r"))
(match_operand: 3 "gpr_operand" "0")))
(clobber (reg:CC_FP CCFP_REGNUM))])]
"TARGET_FUSED_MADD || mode == V2SImode"
"#"
"reload_completed || (epiphany_vect_align == 4 && TARGET_SPLIT_VECMOVE_EARLY)"
[(parallel
[(set (match_dup 5)
(plus: (mult: (match_dup 6) (match_dup 7))
(match_dup 8)))
(clobber (reg:CC_FP CCFP_REGNUM))
(match_dup 13)
(match_dup 14)])
(parallel
[(set (match_dup 9)
(plus: (mult: (match_dup 10) (match_dup 11))
(match_dup 12)))
(clobber (reg:CC_FP CCFP_REGNUM))
(match_dup 13)
(match_dup 14)])]
{
operands[5]
= simplify_gen_subreg (mode, operands[0], mode, 0);
operands[6]
= simplify_gen_subreg (mode, operands[1], mode, 0);
operands[7]
= simplify_gen_subreg (mode, operands[2], mode, 0);
operands[8]
= simplify_gen_subreg (mode, operands[3], mode, 0);
operands[9] = simplify_gen_subreg (mode, operands[0],
mode, UNITS_PER_WORD);
operands[10] = simplify_gen_subreg (mode, operands[1],
mode, UNITS_PER_WORD);
operands[11] = simplify_gen_subreg (mode, operands[2],
mode, UNITS_PER_WORD);
operands[12] = simplify_gen_subreg (mode, operands[3],
mode, UNITS_PER_WORD);
if (!reload_completed)
{
if (reg_overlap_mentioned_p (operands[5], operands[10]))
operands[10] = copy_to_mode_reg (mode, operands[10]);
if (reg_overlap_mentioned_p (operands[5], operands[11]))
operands[11] = copy_to_mode_reg (mode, operands[11]);
if (reg_overlap_mentioned_p (operands[5], operands[12]))
operands[12] = copy_to_mode_reg (mode, operands[12]);
emit_insn (gen_madd (operands[5], operands[6], operands[7],
operands[8]));
emit_insn (gen_madd (operands[9], operands[10], operands[11],
operands[12]));
DONE;
}
gcc_assert (!reg_overlap_mentioned_p (operands[5], operands[10]));
gcc_assert (!reg_overlap_mentioned_p (operands[5], operands[11]));
gcc_assert (!reg_overlap_mentioned_p (operands[5], operands[12]));
operands[13] = XVECEXP (operands[4], 0, XVECLEN (operands[4], 0) - 2);
operands[14] = XVECEXP (operands[4], 0, XVECLEN (operands[4], 0) - 1);
}
[(set_attr "length" "8")
(set_attr "type" "")])
(define_expand "vec_set"
[(match_operand:DWV2MODE 0 "register_operand")
(match_operand: 1 "register_operand")
(match_operand 2 "const_int_operand" "")]
""
{
operands[0]
= simplify_gen_subreg (mode, operands[0], mode,
UNITS_PER_WORD * INTVAL (operands[2]));
emit_move_insn (operands[0], operands[1]);
DONE;
})
(define_expand "movmisalign"
[(set (match_operand:DWV2MODE 0 "nonimmediate_operand" "")
(match_operand:DWV2MODE 1 "general_operand" ""))]
""
{
rtx op00, op01, op10, op11;
op00 = simplify_gen_subreg (mode, operands[0], mode, 0);
op01 = simplify_gen_subreg (mode, operands[0], mode,
UNITS_PER_WORD);
op10 = simplify_gen_subreg (mode, operands[1], mode, 0);
op11 = simplify_gen_subreg (mode, operands[1], mode,
UNITS_PER_WORD);
emit_move_insn (op00, op10);
emit_move_insn (op01, op11);
DONE;
})
(define_insn "nop"
[(const_int 0)]
""
"nop"
[(set_attr "type" "flow")])