;; GCC machine description for CRIS cpu cores.
;; Copyright (C) 1998-2013 Free Software Foundation, Inc.
;; Contributed by Axis Communications.
;; This file is part of GCC.
;;
;; GCC is free software; you can redistribute it and/or modify
;; it under the terms of the GNU General Public License as published by
;; the Free Software Foundation; either version 3, or (at your option)
;; any later version.
;;
;; GCC is distributed in the hope that it will be useful,
;; but WITHOUT ANY WARRANTY; without even the implied warranty of
;; MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
;; GNU General Public License for more details.
;;
;; You should have received a copy of the GNU General Public License
;; along with GCC; see the file COPYING3. If not see
;; .
;; The original PO technology requires these to be ordered by speed,
;; so that assigner will pick the fastest.
;; See files "md.texi" and "rtl.def" for documentation on define_insn,
;; match_*, et. al.
;;
;; The function cris_notice_update_cc in cris.c handles condition code
;; updates for most instructions, helped by the "cc" attribute.
;; There are several instructions that are orthogonal in size, and seems
;; they could be matched by a single pattern without a specified size
;; for the operand that is orthogonal. However, this did not work on
;; gcc-2.7.2 (and probably not on gcc-2.8.1), relating to that when a
;; constant is substituted into an operand, the actual mode must be
;; deduced from the pattern. There is reasonable hope that that has been
;; fixed, so FIXME: try again.
;; You will notice that three-operand alternatives ("=r", "r", "!To")
;; are marked with a "!" constraint modifier to avoid being reloaded
;; into. This is because gcc would otherwise prefer to use the constant
;; pool and its offsettable address instead of reloading to an
;; ("=r", "0", "i") alternative. Also, the constant-pool support was not
;; only suboptimal but also buggy in 2.7.2, ??? maybe only in 2.6.3.
;; All insns that look like (set (...) (plus (...) (reg:SI 8)))
;; get problems when reloading r8 (frame pointer) to r14 + offs (stack
;; pointer). Thus the instructions that get into trouble have specific
;; checks against matching frame_pointer_rtx.
;; ??? But it should be re-checked for gcc > 2.7.2
;; FIXME: This changed some time ago (from 2000-03-16) for gcc-2.9x.
;; FIXME: When PIC, all [rX=rY+S] could be enabled to match
;; [rX=gotless_symbol].
;; The movsi for a gotless symbol could be split (post reload).
(define_c_enum ""
[
;; PLT reference from call expansion: operand 0 is the address,
;; the mode is VOIDmode. Always wrapped in CONST.
;; The value is relative to the GOT.
CRIS_UNSPEC_PLT_GOTREL
;; PLT reference from call expansion: operand 0 is the address,
;; the mode is VOIDmode. Always wrapped in CONST.
;; The value is relative to the PC. It's arch-dependent whether
;; the offset counts from the start or the end of the current item.
CRIS_UNSPEC_PLT_PCREL
;; The address of the global offset table as a source operand.
CRIS_UNSPEC_GOT
;; The offset from the global offset table to the operand.
CRIS_UNSPEC_GOTREL
;; The PC-relative offset to the operand. It's arch-dependent whether
;; the offset counts from the start or the end of the current item.
CRIS_UNSPEC_PCREL
;; The index into the global offset table of a symbol, while
;; also generating a GOT entry for the symbol.
CRIS_UNSPEC_GOTREAD
;; Similar to CRIS_UNSPEC_GOTREAD, but also generating a PLT entry.
CRIS_UNSPEC_PLTGOTREAD
;; Condition for v32 casesi jump, since it needs to have if_then_else
;; form with register as one branch and default label as other.
;; Operand 0 is const_int 0.
CRIS_UNSPEC_CASESI
;; Stack frame deallocation barrier.
CRIS_UNSPEC_FRAME_DEALLOC
;; Swap all 32 bits of the operand; 31 <=> 0, 30 <=> 1...
CRIS_UNSPEC_SWAP_BITS
])
;; Register numbers.
(define_constants
[(CRIS_GOT_REGNUM 0)
(CRIS_STATIC_CHAIN_REGNUM 7)
(CRIS_FP_REGNUM 8)
(CRIS_SP_REGNUM 14)
(CRIS_ACR_REGNUM 15)
(CRIS_SRP_REGNUM 16)
(CRIS_MOF_REGNUM 17)
(CRIS_AP_REGNUM 18)
(CRIS_CC0_REGNUM 19)]
)
;; We need an attribute to define whether an instruction can be put in
;; a branch-delay slot or not, and whether it has a delay slot.
;;
;; Branches and return instructions have a delay slot, and cannot
;; themselves be put in a delay slot. This has changed *for short
;; branches only* between architecture variants, but the possible win
;; is presumed negligible compared to the added complexity of the machine
;; description: one would have to add always-correct infrastructure to
;; distinguish short branches.
;;
;; Whether an instruction can be put in a delay slot depends on the
;; instruction (all short instructions except jumps and branches)
;; and the addressing mode (must not be prefixed or referring to pc).
;; In short, any "slottable" instruction must be 16 bit and not refer
;; to pc, or alter it.
;;
;; The possible values are "yes", "no", "has_slot", "has_return_slot"
;; and "has_call_slot".
;; Yes/no tells whether the insn is slottable or not. Has_call_slot means
;; that the insn is a call insn, which for CRIS v32 has a delay-slot.
;; Of special concern is that no RTX_FRAME_RELATED insn must go in that
;; call delay slot, as it's located in the address *after* the call insn,
;; and the unwind machinery doesn't know about delay slots.
;; Has_slot means that the insn is a branch insn (which are
;; not considered slottable since that is generally true). Having the
;; seemingly illogical value "has_slot" means we do not have to add
;; another attribute just to say that an insn has a delay-slot, since it
;; also infers that it is not slottable. Better names for the attribute
;; were found to be longer and not add readability to the machine
;; description.
;; Has_return_slot is similar, for the return insn.
;;
;; The default that is defined here for this attribute is "no", not
;; slottable, not having a delay-slot, so there's no need to worry about
;; it being wrong for non-branch and return instructions.
;; The default could depend on the kind of insn and the addressing
;; mode, but that would need more attributes and hairier, more error
;; prone code.
;;
;; There is an extra memory constraint, 'Q', which recognizes an indirect
;; register. The constraints 'Q' and '>' together match all possible
;; memory operands that are slottable.
;; For other operands, you need to check if it has a valid "slottable"
;; quick-immediate operand, where the particular signedness-variation
;; may match the constraints 'I' or 'J'.), and include it in the
;; constraint pattern for the slottable pattern. An alternative using
;; only "r" constraints is most often slottable.
(define_attr "slottable" "no,yes,has_slot,has_return_slot,has_call_slot"
(const_string "no"))
;; We also need attributes to sanely determine the condition code
;; state. See cris_notice_update_cc for how this is used.
(define_attr "cc" "none,clobber,normal,noov32,rev" (const_string "normal"))
;; At the moment, this attribute is just used to help bb-reorder do its
;; work; the default 0 doesn't help it. Many insns have other lengths,
;; though none are shorter.
(define_attr "length" "" (const_int 2))
;; A branch has one delay-slot. The instruction in the
;; delay-slot is always executed, independent of whether the branch is
;; taken or not. Note that besides setting "slottable" to "has_slot",
;; there also has to be a "%#" at the end of a "delayed" instruction
;; output pattern (for "jump" this means "ba %l0%#"), so print_operand can
;; catch it and print a "nop" if necessary. This method was stolen from
;; sparc.md.
(define_delay (eq_attr "slottable" "has_slot")
[(eq_attr "slottable" "yes") (nil) (nil)])
;; We can't put prologue insns in call-insn delay-slots when
;; DWARF2 unwind info is emitted, because the unwinder matches the
;; address after the insn. It must see the return address of a call at
;; a position at least *one byte after* the insn, or it'll think that
;; the insn hasn't been executed. If the insn is in a delay-slot of a
;; call, it's just *exactly* after the insn.
(define_delay (eq_attr "slottable" "has_call_slot")
[(and (eq_attr "slottable" "yes")
(ior (not (match_test "RTX_FRAME_RELATED_P (insn)"))
(not (match_test "flag_exceptions"))))
(nil) (nil)])
;; The insn in the return insn slot must not be the
;; return-address-register restore. FIXME: Use has_slot and express
;; as a parallel with a use of the return-address-register (currently
;; only SRP). However, this requires an amount of fixing tests for
;; naked RETURN in middle-end.
(define_delay (eq_attr "slottable" "has_return_slot")
[(and (eq_attr "slottable" "yes")
(not (match_test "dead_or_set_regno_p (insn, CRIS_SRP_REGNUM)")))
(nil) (nil)])
;; Iterator definitions.
;; For the "usual" pattern size alternatives.
(define_mode_iterator BWD [SI HI QI])
(define_mode_iterator WD [SI HI])
(define_mode_iterator BW [HI QI])
(define_mode_attr S [(SI "HI") (HI "QI")])
(define_mode_attr s [(SI "hi") (HI "qi")])
(define_mode_attr m [(SI ".d") (HI ".w") (QI ".b")])
(define_mode_attr mm [(SI ".w") (HI ".b")])
(define_mode_attr nbitsm1 [(SI "31") (HI "15") (QI "7")])
;; For the sign_extend+zero_extend variants.
(define_code_iterator szext [sign_extend zero_extend])
(define_code_attr u [(sign_extend "") (zero_extend "u")])
(define_code_attr su [(sign_extend "s") (zero_extend "u")])
;; For the shift variants.
(define_code_iterator shift [ashiftrt lshiftrt ashift])
(define_code_iterator shiftrt [ashiftrt lshiftrt])
(define_code_attr shlr [(ashiftrt "ashr") (lshiftrt "lshr") (ashift "ashl")])
(define_code_attr slr [(ashiftrt "asr") (lshiftrt "lsr") (ashift "lsl")])
(define_code_iterator ncond [eq ne gtu ltu geu leu])
(define_code_iterator ocond [gt le])
(define_code_iterator rcond [lt ge])
(define_code_attr CC [(eq "eq") (ne "ne") (gt "gt") (gtu "hi") (lt "lt")
(ltu "lo") (ge "ge") (geu "hs") (le "le") (leu "ls")])
(define_code_attr rCC [(eq "ne") (ne "eq") (gt "le") (gtu "ls") (lt "ge")
(ltu "hs") (ge "lt") (geu "lo") (le "gt") (leu "hi")])
(define_code_attr oCC [(lt "mi") (ge "pl")])
(define_code_attr roCC [(lt "pl") (ge "mi")])
;; Operand and operator predicates.
(include "predicates.md")
(include "constraints.md")
;; Test insns.
;; No test insns with side-effect on the mem addressing.
;;
;; See note on cmp-insns with side-effects (or lack of them)
;; Normal named test patterns from SI on.
(define_insn "*tstsi"
[(set (cc0)
(compare (match_operand:SI 0 "nonimmediate_operand" "r,Q>,m")
(const_int 0)))]
""
{
if (which_alternative == 0 && TARGET_V32)
return "cmpq 0,%0";
return "test.d %0";
}
[(set_attr "slottable" "yes,yes,no")])
(define_insn "*tst_cmp"
[(set (cc0)
(compare (match_operand:BW 0 "nonimmediate_operand" "r,Q>,m")
(const_int 0)))]
"cris_cc0_user_requires_cmp (insn)"
"@
cmp 0,%0
test %0
test %0"
[(set_attr "slottable" "no,yes,no")])
(define_insn "*tst_non_cmp"
[(set (cc0)
(compare (match_operand:BW 0 "nonimmediate_operand" "r,Q>,m")
(const_int 0)))]
"!cris_cc0_user_requires_cmp (insn)"
"@
move %0,%0
test %0
test %0"
[(set_attr "slottable" "yes,yes,no")
(set_attr "cc" "noov32,*,*")])
;; It seems that the position of the sign-bit and the fact that 0.0 is
;; all 0-bits would make "tstsf" a straight-forward implementation;
;; either "test.d" it for positive/negative or "btstq 30,r" it for
;; zeroness.
;;
;; FIXME: Do that some time; check next_cc0_user to determine if
;; zero or negative is tested for.
;; Compare insns.
;; We could optimize the sizes of the immediate operands for various
;; cases, but that is not worth it because of the very little usage of
;; DImode for anything else but a structure/block-mode. Just do the
;; obvious stuff for the straight-forward constraint letters.
(define_insn "*cmpdi_non_v32"
[(set (cc0)
(compare (match_operand:DI 0 "nonimmediate_operand" "rm,r,r,r,r,r,r,o")
(match_operand:DI 1 "general_operand" "M,Kc,I,P,n,r,o,r")))]
"!TARGET_V32"
"@
test.d %M0\;ax\;test.d %H0
cmpq %1,%M0\;ax\;cmpq 0,%H0
cmpq %1,%M0\;ax\;cmpq -1,%H0
cmp%e1.%z1 %1,%M0\;ax\;cmpq %H1,%H0
cmp.d %M1,%M0\;ax\;cmp.d %H1,%H0
cmp.d %M1,%M0\;ax\;cmp.d %H1,%H0
cmp.d %M1,%M0\;ax\;cmp.d %H1,%H0
cmp.d %M0,%M1\;ax\;cmp.d %H0,%H1")
(define_insn "*cmpdi_v32"
[(set (cc0)
(compare (match_operand:DI 0 "register_operand" "r,r,r,r,r")
(match_operand:DI 1 "nonmemory_operand" "Kc,I,P,n,r")))]
"TARGET_V32"
"@
cmpq %1,%M0\;ax\;cmpq 0,%H0
cmpq %1,%M0\;ax\;cmpq -1,%H0
cmp%e1.%z1 %1,%M0\;ax\;cmpq %H1,%H0
cmp.d %M1,%M0\;ax\;cmp.d %H1,%H0
cmp.d %M1,%M0\;ax\;cmp.d %H1,%H0")
;; Note that compare insns with side effect addressing mode (e.g.):
;;
;; cmp.S [rx=ry+i],rz;
;; cmp.S [%3=%1+%2],%0
;;
;; are *not* usable for gcc since the reloader *does not accept*
;; cc0-changing insns with side-effects other than setting the condition
;; codes. The reason is that the reload stage *may* cause another insn to
;; be output after the main instruction, in turn invalidating cc0 for the
;; insn using the test. (This does not apply to the CRIS case, since a
;; reload for output -- move to memory -- does not change the condition
;; code. Unfortunately we have no way to describe that at the moment. I
;; think code would improve being in the order of one percent faster.
;; We have cmps and cmpu (compare reg w. sign/zero extended mem).
;; These are mostly useful for compares in SImode, using 8 or 16-bit
;; constants, but sometimes gcc will find its way to use it for other
;; (memory) operands. Avoid side-effect patterns, though (see above).
(define_insn "*cmp_ext"
[(set (cc0)
(compare
(match_operand:SI 0 "register_operand" "r,r")
(match_operator:SI 2 "cris_extend_operator"
[(match_operand:BW 1 "memory_operand" "Q>,m")])))]
""
"cmp%e2 %1,%0"
[(set_attr "slottable" "yes,no")])
;; Swap operands; it seems the canonical look (if any) is not enforced.
;;
;; FIXME: Investigate that.
(define_insn "*cmp_swapext"
[(set (cc0)
(compare
(match_operator:SI 2 "cris_extend_operator"
[(match_operand:BW 0 "memory_operand" "Q>,m")])
(match_operand:SI 1 "register_operand" "r,r")))]
""
"cmp%e2 %0,%1"
[(set_attr "slottable" "yes,no")
(set_attr "cc" "rev")])
;; The "normal" compare patterns, from SI on. Special-cases with zero
;; are covered above.
(define_insn "*cmpsi"
[(set (cc0)
(compare
(match_operand:SI 0 "nonimmediate_operand" "r,r,r, Q>,r,r,m")
(match_operand:SI 1 "general_operand" "I,r,Q>,r, P,g,r")))]
""
"@
cmpq %1,%0
cmp.d %1,%0
cmp.d %1,%0
cmp.d %0,%1
cmp%e1.%z1 %1,%0
cmp.d %1,%0
cmp.d %0,%1"
[(set_attr "slottable" "yes,yes,yes,yes,no,no,no")
(set_attr "cc" "normal,normal,normal,rev,normal,normal,rev")])
(define_insn "*cmp"
[(set (cc0)
(compare (match_operand:BW 0 "nonimmediate_operand" "r,r, Q>,r,m")
(match_operand:BW 1 "general_operand" "r,Q>,r, g,r")))]
""
"@
cmp %1,%0
cmp %1,%0
cmp %0,%1
cmp %1,%0
cmp %0,%1"
[(set_attr "slottable" "yes,yes,yes,no,no")
(set_attr "cc" "normal,normal,rev,normal,rev")])
;; Pattern matching the BTST insn.
;; It is useful for "if (i & val)" constructs, where val is an exact
;; power of 2, or if val + 1 is a power of two, where we check for a bunch
;; of zeros starting at bit 0).
;; SImode. This mode is the only one needed, since gcc automatically
;; extends subregs for lower-size modes. FIXME: Add testcase.
(define_insn "*btst"
[(set (cc0)
(compare
(zero_extract:SI
(match_operand:SI 0 "nonmemory_operand" "r, r,r, r,r, r,Kp")
(match_operand:SI 1 "const_int_operand" "Kc,n,Kc,n,Kc,n,n")
(match_operand:SI 2 "nonmemory_operand" "M, M,Kc,n,r, r,r"))
(const_int 0)))]
;; Either it is a single bit, or consecutive ones starting at 0.
;; The btst ones depend on stuff in NOTICE_UPDATE_CC.
"CONST_INT_P (operands[1])
&& (operands[1] == const1_rtx || operands[2] == const0_rtx)
&& (REG_S_P (operands[0])
|| (operands[1] == const1_rtx
&& REG_S_P (operands[2])
&& CONST_INT_P (operands[0])
&& exact_log2 (INTVAL (operands[0])) >= 0))
&& !TARGET_CCINIT"
;; The next-to-last "&&" condition above should be caught by some kind of
;; canonicalization in gcc, but we can easily help with it here.
;; It results from expressions of the type
;; "power_of_2_value & (1 << y)".
;;
;; Since there may be codes with tests in on bits (in constant position)
;; beyond the size of a word, handle that by assuming those bits are 0.
;; GCC should handle that, but it's a matter of easily-added belts while
;; having suspenders.
"@
btstq (%1-1),%0
cmpq 0,%0
btstq %2,%0
clearf nz
btst %2,%0
clearf nz
cmpq %p0,%2"
[(set_attr "slottable" "yes")
(set_attr "cc" "noov32")])
;; Move insns.
;; The whole mandatory movdi family is here; expander, "anonymous"
;; recognizer and splitter. We're forced to have a movdi pattern,
;; although GCC should be able to split it up itself. Normally it can,
;; but if other insns have DI operands (as is the case here), reload
;; must be able to generate or match a movdi. many testcases fail at
;; -O3 or -fssa if we don't have this. FIXME: Fix GCC... See
;; .
;; However, a patch from Richard Kenner (similar to the cause of
;; discussion at the URL above), indicates otherwise. See
;; .
;; The truth has IMO is not been decided yet, so check from time to
;; time by disabling the movdi patterns.
;; To appease testcase gcc.c-torture/execute/920501-2.c (and others) at
;; -O0, we need a movdi as a temporary measure. Here's how things fail:
;; A cmpdi RTX needs reloading (global):
;; (insn 185 326 186 (set (cc0)
;; (compare (mem/f:DI (reg/v:SI 22) 0)
;; (const_int 1 [0x1]))) 4 {cmpdi} (nil)
;; (nil))
;; Now, reg 22 is reloaded for input address, and the mem is also moved
;; out of the instruction (into a register), since one of the operands
;; must be a register. Reg 22 is reloaded (into reg 10), and the mem is
;; moved out and synthesized in SImode parts (reg 9, reg 10 - should be ok
;; wrt. overlap). The bad things happen with the synthesis in
;; emit_move_insn_1; the location where to substitute reg 10 is lost into
;; two new RTX:es, both still having reg 22. Later on, the left-over reg
;; 22 is recognized to have an equivalent in memory which is substituted
;; straight in, and we end up with an unrecognizable insn:
;; (insn 325 324 326 (set (reg:SI 9 r9)
;; (mem/f:SI (mem:SI (plus:SI (reg:SI 8 r8)
;; (const_int -84 [0xffffffac])) 0) 0)) -1 (nil)
;; (nil))
;; which is the first part of the reloaded synthesized "movdi".
;; The right thing would be to add equivalent replacement locations for
;; insn with pseudos that need more reloading. The question is where.
(define_expand "movdi"
[(set (match_operand:DI 0 "nonimmediate_operand" "")
(match_operand:DI 1 "general_operand" ""))]
""
{
if (MEM_P (operands[0])
&& operands[1] != const0_rtx
&& (!TARGET_V32 || (!REG_P (operands[1]) && can_create_pseudo_p ())))
operands[1] = copy_to_mode_reg (DImode, operands[1]);
/* Some other ports (as of 2001-09-10 for example mcore and romp) also
prefer to split up constants early, like this. The testcase in
gcc.c-torture/execute/961213-1.c shows that CSE2 gets confused by the
resulting subreg sets when using the construct from mcore (as of FSF
CVS, version -r 1.5), and it believes that the high part (the last one
emitted) is the final value. */
if ((CONST_INT_P (operands[1]) || GET_CODE (operands[1]) == CONST_DOUBLE)
&& ! reload_completed
&& ! reload_in_progress)
{
rtx insns;
rtx op0 = operands[0];
rtx op1 = operands[1];
start_sequence ();
emit_move_insn (operand_subword (op0, 0, 1, DImode),
operand_subword (op1, 0, 1, DImode));
emit_move_insn (operand_subword (op0, 1, 1, DImode),
operand_subword (op1, 1, 1, DImode));
insns = get_insns ();
end_sequence ();
emit_insn (insns);
DONE;
}
})
(define_insn_and_split "*movdi_insn_non_v32"
[(set (match_operand:DI 0 "nonimmediate_operand" "=r,rx,m")
(match_operand:DI 1 "general_operand" "rx,g,rxM"))]
"(register_operand (operands[0], DImode)
|| register_operand (operands[1], DImode)
|| operands[1] == const0_rtx)
&& !TARGET_V32"
"#"
"&& reload_completed"
[(match_dup 2)]
"operands[2] = cris_split_movdx (operands);")
;; Overlapping (but non-identical) source memory address and destination
;; register would be a compiler bug, so we don't have to specify that.
(define_insn "*movdi_v32"
[(set
(match_operand:DI 0 "nonimmediate_operand" "=r,rx,&r,>, m,r,x,m")
(match_operand:DI 1 "general_operand" "rxi,r>,m, rx,r,m,m,x"))]
"TARGET_V32"
{
switch (which_alternative)
{
/* FIXME: 1) Use autoincrement where possible. 2) Have peephole2,
particularly for cases where the address register is dead. */
case 5:
if (REGNO (operands[0]) == REGNO (XEXP (operands[1], 0)))
return "addq 4,%L1\;move.d %1,%H0\;subq 4,%L1\;move.d %1,%M0";
gcc_assert (REGNO (operands[0]) + 1 == REGNO (XEXP (operands[1], 0)));
return "move.d [%L1+],%M0\;move.d [%L1],%H0";
case 2:
/* We could do away with the addq if we knew the address-register
isn't ACR. If we knew the address-register is dead, we could do
away with the subq too. */
return "move.d [%L1],%M0\;addq 4,%L1\;move.d [%L1],%H0\;subq 4,%L1";
case 4:
return "move.d %M1,[%L0]\;addq 4,%L0\;move.d %H1,[%L0]\;subq 4,%L0";
case 6:
return "move [%L1],%M0\;addq 4,%L1\;move [%L1],%H0\;subq 4,%L1";
case 7:
return "move %M1,[%L0]\;addq 4,%L0\;move %H1,[%L0]\;subq 4,%L0";
default:
return "#";
}
}
;; The non-split cases clobber cc0 because of their adds and subs.
;; Beware that NOTICE_UPDATE_CC is called before the forced split happens.
[(set_attr "cc" "*,*,clobber,*,clobber,clobber,*,*")])
;; Much like "*movdi_insn_non_v32". Overlapping registers and constants
;; is handled so much better in cris_split_movdx.
(define_split
[(set (match_operand:DI 0 "nonimmediate_operand" "")
(match_operand:DI 1 "general_operand" ""))]
"TARGET_V32
&& reload_completed
&& (!MEM_P (operands[0]) || !REG_P (XEXP (operands[0], 0)))
&& (!MEM_P (operands[1]) || !REG_P (XEXP (operands[1], 0)))"
[(match_dup 2)]
"operands[2] = cris_split_movdx (operands);")
;; Side-effect patterns for move.S1 [rx=ry+rx.S2],rw
;; and move.S1 [rx=ry+i],rz
;; Then movs.S1 and movu.S1 for both modes.
;;
;; move.S1 [rx=ry+rz.S],rw avoiding when rx is ry, or rw is rx
;; FIXME: These could have anonymous mode for operand 0.
;; FIXME: Special registers' alternatives too.
(define_insn "*mov_side_biap"
[(set (match_operand:BW 0 "register_operand" "=r,r")
(mem:BW (plus:SI
(mult:SI (match_operand:SI 1 "register_operand" "r,r")
(match_operand:SI 2 "const_int_operand" "n,n"))
(match_operand:SI 3 "register_operand" "r,r"))))
(set (match_operand:SI 4 "register_operand" "=*3,r")
(plus:SI (mult:SI (match_dup 1)
(match_dup 2))
(match_dup 3)))]
"cris_side_effect_mode_ok (MULT, operands, 4, 3, 1, 2, 0)"
"@
#
move [%4=%3+%1%T2],%0")
(define_insn "*mov_sidesisf_biap"
[(set (match_operand 0 "register_operand" "=r,r,x,x")
(mem (plus:SI
(mult:SI (match_operand:SI 1 "register_operand" "r,r,r,r")
(match_operand:SI 2 "const_int_operand" "n,n,n,n"))
(match_operand:SI 3 "register_operand" "r,r,r,r"))))
(set (match_operand:SI 4 "register_operand" "=*3,r,*3,r")
(plus:SI (mult:SI (match_dup 1)
(match_dup 2))
(match_dup 3)))]
"GET_MODE_SIZE (GET_MODE (operands[0])) == UNITS_PER_WORD
&& cris_side_effect_mode_ok (MULT, operands, 4, 3, 1, 2, 0)"
"@
#
move.%s0 [%4=%3+%1%T2],%0
#
move [%4=%3+%1%T2],%0")
;; move.S1 [rx=ry+i],rz
;; avoiding move.S1 [ry=ry+i],rz
;; and move.S1 [rz=ry+i],rz
;; Note that "i" is allowed to be a register.
(define_insn "*mov_side"
[(set (match_operand:BW 0 "register_operand" "=r,r,r,r,r")
(mem:BW
(plus:SI (match_operand:SI 1 "cris_bdap_operand" "%r,r,r,R,R")
(match_operand:SI 2 "cris_bdap_operand" "r>Rn,r,>Rn,r,r"))))
(set (match_operand:SI 3 "register_operand" "=*1,r,r,*2,r")
(plus:SI (match_dup 1)
(match_dup 2)))]
"cris_side_effect_mode_ok (PLUS, operands, 3, 1, 2, -1, 0)"
{
if ((which_alternative == 0 || which_alternative == 3)
&& (!CONST_INT_P (operands[2])
|| INTVAL (operands[2]) > 127
|| INTVAL (operands[2]) < -128
|| satisfies_constraint_N (operands[2])
|| satisfies_constraint_J (operands[2])))
return "#";
if (which_alternative == 4)
return "move [%3=%2%S1],%0";
return "move [%3=%1%S2],%0";
})
(define_insn "*mov_sidesisf"
[(set (match_operand 0 "register_operand" "=r,r,r,x,x,x,r,r,x,x")
(mem
(plus:SI
(match_operand:SI 1 "cris_bdap_operand" "%r,r,r,r,r,r,R,R,R,R")
(match_operand:SI 2 "cris_bdap_operand" "r>Rn,r,>Rn,r>Rn,r,>Rn,r,r,r,r"))))
(set (match_operand:SI 3 "register_operand" "=*1,r,r,*1,r,r,*2,r,*2,r")
(plus:SI (match_dup 1)
(match_dup 2)))]
"GET_MODE_SIZE (GET_MODE (operands[0])) == UNITS_PER_WORD
&& cris_side_effect_mode_ok (PLUS, operands, 3, 1, 2, -1, 0)"
{
if ((which_alternative == 0
|| which_alternative == 3
|| which_alternative == 6
|| which_alternative == 8)
&& (!CONST_INT_P (operands[2])
|| INTVAL (operands[2]) > 127
|| INTVAL (operands[2]) < -128
|| satisfies_constraint_N (operands[2])
|| satisfies_constraint_J (operands[2])))
return "#";
if (which_alternative < 3)
return "move.%s0 [%3=%1%S2],%0";
if (which_alternative == 7)
return "move.%s0 [%3=%2%S1],%0";
if (which_alternative == 9)
return "move [%3=%2%S1],%0";
return "move [%3=%1%S2],%0";
})
;; Other way around; move to memory.
;; Note that the condition (which for side-effect patterns is usually a
;; call to cris_side_effect_mode_ok), isn't consulted for register
;; allocation preferences -- constraints is the method for that. The
;; drawback is that we can't exclude register allocation to cause
;; "move.s rw,[rx=ry+rz.S]" when rw==rx without also excluding rx==ry or
;; rx==rz if we use an earlyclobber modifier for the constraint for rx.
;; Instead of that, we recognize and split the cases where dangerous
;; register combinations are spotted: where a register is set in the
;; side-effect, and used in the main insn. We don't handle the case where
;; the set in the main insn overlaps the set in the side-effect; that case
;; must be handled in gcc. We handle just the case where the set in the
;; side-effect overlaps the input operand of the main insn (i.e. just
;; moves to memory).
;;
;; move.s rz,[ry=rx+rw.S]
(define_insn "*mov_side_biap_mem"
[(set (mem:BW (plus:SI
(mult:SI (match_operand:SI 0 "register_operand" "r,r,r")
(match_operand:SI 1 "const_int_operand" "n,n,n"))
(match_operand:SI 2 "register_operand" "r,r,r")))
(match_operand:BW 3 "register_operand" "r,r,r"))
(set (match_operand:SI 4 "register_operand" "=*2,!3,r")
(plus:SI (mult:SI (match_dup 0)
(match_dup 1))
(match_dup 2)))]
"cris_side_effect_mode_ok (MULT, operands, 4, 2, 0, 1, 3)"
"@
#
#
move %3,[%4=%2+%0%T1]")
(define_insn "*mov_sidesisf_biap_mem"
[(set (mem (plus:SI
(mult:SI (match_operand:SI 0 "register_operand" "r,r,r,r,r,r")
(match_operand:SI 1 "const_int_operand" "n,n,n,n,n,n"))
(match_operand:SI 2 "register_operand" "r,r,r,r,r,r")))
(match_operand 3 "register_operand" "r,r,r,x,x,x"))
(set (match_operand:SI 4 "register_operand" "=*2,!3,r,*2,!3,r")
(plus:SI (mult:SI (match_dup 0)
(match_dup 1))
(match_dup 2)))]
"GET_MODE_SIZE (GET_MODE (operands[3])) == UNITS_PER_WORD
&& cris_side_effect_mode_ok (MULT, operands, 4, 2, 0, 1, 3)"
"@
#
#
move.%s3 %3,[%4=%2+%0%T1]
#
#
move %3,[%4=%2+%0%T1]")
;; Split for the case above where we're out of luck with register
;; allocation (again, the condition isn't checked for that), and we end up
;; with the set in the side-effect getting the same register as the input
;; register.
(define_split
[(parallel
[(set (match_operator
6 "cris_mem_op"
[(plus:SI
(mult:SI (match_operand:SI 0 "register_operand" "")
(match_operand:SI 1 "const_int_operand" ""))
(match_operand:SI 2 "register_operand" ""))])
(match_operand 3 "register_operand" ""))
(set (match_operand:SI 4 "cris_nonsp_register_operand" "")
(plus:SI (mult:SI (match_dup 0)
(match_dup 1))
(match_dup 2)))])]
"reload_completed && reg_overlap_mentioned_p (operands[4], operands[3])"
[(set (match_dup 5) (match_dup 3))
(set (match_dup 4) (match_dup 2))
(set (match_dup 4)
(plus:SI (mult:SI (match_dup 0)
(match_dup 1))
(match_dup 4)))]
"operands[5]
= replace_equiv_address (operands[6],
gen_rtx_PLUS (SImode,
gen_rtx_MULT (SImode,
operands[0],
operands[1]),
operands[2]));")
;; move.s rx,[ry=rz+i]
;; FIXME: These could have anonymous mode for operand 2.
;; QImode
(define_insn "*mov_side_mem"
[(set (mem:BW
(plus:SI (match_operand:SI 0 "cris_bdap_operand" "%r,r,r,r,R,R,R")
(match_operand:SI 1 "cris_bdap_operand" "r>Rn,r>Rn,r,>Rn,r,r,r")))
(match_operand:BW 2 "register_operand" "r,r,r,r,r,r,r"))
(set (match_operand:SI 3 "register_operand" "=*0,!*2,r,r,*1,!*2,r")
(plus:SI (match_dup 0)
(match_dup 1)))]
"cris_side_effect_mode_ok (PLUS, operands, 3, 0, 1, -1, 2)"
{
if ((which_alternative == 0 || which_alternative == 4)
&& (!CONST_INT_P (operands[1])
|| INTVAL (operands[1]) > 127
|| INTVAL (operands[1]) < -128
|| satisfies_constraint_N (operands[1])
|| satisfies_constraint_J (operands[1])))
return "#";
if (which_alternative == 1 || which_alternative == 5)
return "#";
if (which_alternative == 6)
return "move.%s2 %2,[%3=%1%S0]";
return "move %2,[%3=%0%S1]";
})
;; SImode
(define_insn "*mov_sidesisf_mem"
[(set (mem
(plus:SI
(match_operand:SI
0 "cris_bdap_operand"
"%r, r, r,r, r, r,r, R,R, R,R, R")
(match_operand:SI
1 "cris_bdap_operand"
"r>Rn,r>Rn,r,>Rn,r>Rn,r,>Rn,r,r, r,r, r")))
(match_operand 2 "register_operand"
"r, r, r,r, x, x,x, r,r, r,x, x"))
(set (match_operand:SI 3 "register_operand"
"=*0,!2, r,r, *0, r,r, *1,!*2,r,*1,r")
(plus:SI (match_dup 0)
(match_dup 1)))]
"GET_MODE_SIZE (GET_MODE (operands[2])) == UNITS_PER_WORD
&& cris_side_effect_mode_ok (PLUS, operands, 3, 0, 1, -1, 2)"
{
if ((which_alternative == 0 || which_alternative == 4)
&& (!CONST_INT_P (operands[1])
|| INTVAL (operands[1]) > 127
|| INTVAL (operands[1]) < -128
|| satisfies_constraint_N (operands[1])
|| satisfies_constraint_J (operands[1])))
return "#";
if (which_alternative == 1
|| which_alternative == 7
|| which_alternative == 8
|| which_alternative == 10)
return "#";
if (which_alternative < 4)
return "move.%s2 %2,[%3=%0%S1]";
if (which_alternative == 9)
return "move.%s2 %2,[%3=%1%S0]";
if (which_alternative == 11)
return "move %2,[%3=%1%S0]";
return "move %2,[%3=%0%S1]";
})
;; Like the biap case, a split where the set in the side-effect gets the
;; same register as the input register to the main insn, since the
;; condition isn't checked at register allocation.
(define_split
[(parallel
[(set (match_operator
4 "cris_mem_op"
[(plus:SI
(match_operand:SI 0 "cris_bdap_operand" "")
(match_operand:SI 1 "cris_bdap_operand" ""))])
(match_operand 2 "register_operand" ""))
(set (match_operand:SI 3 "cris_nonsp_register_operand" "")
(plus:SI (match_dup 0) (match_dup 1)))])]
"reload_completed && reg_overlap_mentioned_p (operands[3], operands[2])"
[(set (match_dup 4) (match_dup 2))
(set (match_dup 3) (match_dup 0))
(set (match_dup 3) (plus:SI (match_dup 3) (match_dup 1)))]
"")
;; Clear memory side-effect patterns. It is hard to get to the mode if
;; the MEM was anonymous, so there will be one for each mode.
;; clear.[bwd] [ry=rx+rw.s2]
(define_insn "*clear_side_biap"
[(set (mem:BWD (plus:SI
(mult:SI (match_operand:SI 0 "register_operand" "r,r")
(match_operand:SI 1 "const_int_operand" "n,n"))
(match_operand:SI 2 "register_operand" "r,r")))
(const_int 0))
(set (match_operand:SI 3 "register_operand" "=*2,r")
(plus:SI (mult:SI (match_dup 0)
(match_dup 1))
(match_dup 2)))]
"cris_side_effect_mode_ok (MULT, operands, 3, 2, 0, 1, -1)"
"@
#
clear [%3=%2+%0%T1]")
;; clear.[bwd] [ry=rz+i]
(define_insn "*clear_side"
[(set (mem:BWD
(plus:SI (match_operand:SI 0 "cris_bdap_operand" "%r,r,r,R,R")
(match_operand:SI 1 "cris_bdap_operand" "r>Rn,r,>Rn,r,r")))
(const_int 0))
(set (match_operand:SI 2 "register_operand" "=*0,r,r,*1,r")
(plus:SI (match_dup 0)
(match_dup 1)))]
"cris_side_effect_mode_ok (PLUS, operands, 2, 0, 1, -1, -1)"
{
if ((which_alternative == 0 || which_alternative == 3)
&& (!CONST_INT_P (operands[1])
|| INTVAL (operands[1]) > 127
|| INTVAL (operands[1]) < -128
|| satisfies_constraint_N (operands[1])
|| satisfies_constraint_J (operands[1])))
return "#";
if (which_alternative == 4)
return "clear [%2=%1%S0]";
return "clear [%2=%0%S1]";
})
;; Normal move patterns from SI on.
(define_expand "movsi"
[(set
(match_operand:SI 0 "nonimmediate_operand" "")
(match_operand:SI 1 "cris_general_operand_or_symbol" ""))]
""
{
/* If the output goes to a MEM, make sure we have zero or a register as
input. */
if (MEM_P (operands[0])
&& ! REG_S_P (operands[1])
&& operands[1] != const0_rtx
&& can_create_pseudo_p ())
operands[1] = force_reg (SImode, operands[1]);
/* If we're generating PIC and have an incoming symbol, validize it to a
general operand or something that will match a special pattern.
FIXME: Do we *have* to recognize anything that would normally be a
valid symbol? Can we exclude global PIC addresses with an added
offset? */
if (flag_pic
&& CONSTANT_ADDRESS_P (operands[1])
&& !cris_valid_pic_const (operands[1], false))
{
enum cris_pic_symbol_type t = cris_pic_symbol_type_of (operands[1]);
gcc_assert (t != cris_no_symbol);
if (! REG_S_P (operands[0]))
{
/* We must have a register as destination for what we're about to
do, and for the patterns we generate. */
CRIS_ASSERT (can_create_pseudo_p ());
operands[1] = force_reg (SImode, operands[1]);
}
else
{
/* FIXME: add a REG_EQUAL (or is it REG_EQUIV) note to the
destination register for the symbol. It might not be
worth it. Measure. */
crtl->uses_pic_offset_table = 1;
if (t == cris_rel_symbol)
{
/* Change a "move.d sym(+offs),rN" into (allocate register rM)
for pre-v32:
"move.d (const (plus (unspec [sym]
CRIS_UNSPEC_GOTREL) offs)),rM" "add.d rPIC,rM,rN"
and for v32:
"move.d (const (plus (unspec [sym]
CRIS_UNSPEC_PCREL) offs)),rN". */
rtx tem, rm, rn = operands[0];
rtx sym = GET_CODE (operands[1]) != CONST
? operands[1] : get_related_value (operands[1]);
HOST_WIDE_INT offs = get_integer_term (operands[1]);
gcc_assert (can_create_pseudo_p ());
if (TARGET_V32)
{
tem = gen_rtx_UNSPEC (Pmode, gen_rtvec (1, sym),
CRIS_UNSPEC_PCREL);
if (offs != 0)
tem = plus_constant (Pmode, tem, offs);
rm = rn;
emit_move_insn (rm, gen_rtx_CONST (Pmode, tem));
}
else
{
/* We still uses GOT-relative addressing for
pre-v32. */
crtl->uses_pic_offset_table = 1;
tem = gen_rtx_UNSPEC (Pmode, gen_rtvec (1, sym),
CRIS_UNSPEC_GOTREL);
if (offs != 0)
tem = plus_constant (Pmode, tem, offs);
rm = gen_reg_rtx (Pmode);
emit_move_insn (rm, gen_rtx_CONST (Pmode, tem));
if (expand_binop (Pmode, add_optab, rm, pic_offset_table_rtx,
rn, 0, OPTAB_LIB_WIDEN) != rn)
internal_error ("expand_binop failed in movsi gotrel");
}
DONE;
}
else if (t == cris_got_symbol)
{
/* Change a "move.d sym,rN" into (allocate register rM, rO)
"move.d (const (unspec [sym] CRIS_UNSPEC_GOTREAD)),rM"
"add.d rPIC,rM,rO", "move.d [rO],rN" with
the memory access marked as read-only. */
rtx tem, mem, rm, ro, rn = operands[0];
gcc_assert (can_create_pseudo_p ());
tem = gen_rtx_UNSPEC (Pmode, gen_rtvec (1, operands[1]),
CRIS_UNSPEC_GOTREAD);
rm = gen_reg_rtx (Pmode);
emit_move_insn (rm, gen_rtx_CONST (Pmode, tem));
ro = gen_reg_rtx (Pmode);
if (expand_binop (Pmode, add_optab, rm, pic_offset_table_rtx,
ro, 0, OPTAB_LIB_WIDEN) != ro)
internal_error ("expand_binop failed in movsi got");
mem = gen_rtx_MEM (Pmode, ro);
/* This MEM doesn't alias anything. Whether it
aliases other same symbols is unimportant. */
set_mem_alias_set (mem, new_alias_set ());
MEM_NOTRAP_P (mem) = 1;
/* We can set the GOT memory read of a non-called symbol
to readonly, but not that of a call symbol, as those
are subject to lazy evaluation and usually have the value
changed from the first call to the second (but
constant thereafter). */
MEM_READONLY_P (mem) = 1;
emit_move_insn (rn, mem);
DONE;
}
else
{
/* We get here when we have to change something that would
be recognizable if it wasn't PIC. A ``sym'' is ok for
PIC symbols both with and without a GOT entry. And ``sym
+ offset'' is ok for local symbols, so the only thing it
could be, is a global symbol with an offset. Check and
abort if not. */
rtx reg = gen_reg_rtx (Pmode);
rtx sym = get_related_value (operands[1]);
HOST_WIDE_INT offs = get_integer_term (operands[1]);
gcc_assert (can_create_pseudo_p ()
&& t == cris_got_symbol_needing_fixup
&& sym != NULL_RTX && offs != 0);
emit_move_insn (reg, sym);
if (expand_binop (SImode, add_optab, reg,
GEN_INT (offs), operands[0], 0,
OPTAB_LIB_WIDEN) != operands[0])
internal_error ("expand_binop failed in movsi got+offs");
DONE;
}
}
}
})
(define_insn "*movsi_got_load"
[(set (reg:SI CRIS_GOT_REGNUM) (unspec:SI [(const_int 0)] CRIS_UNSPEC_GOT))]
"flag_pic"
{
return TARGET_V32
? "lapc _GLOBAL_OFFSET_TABLE_,%:"
: "move.d $pc,%:\;sub.d .:GOTOFF,%:";
}
[(set_attr "cc" "clobber")])
(define_insn "*movsi_internal"
[(set
(match_operand:SI 0 "nonimmediate_operand"
"=r,r, r,Q>,r,Q>,g,r,r, r,g,rQ>,x, m,x")
(match_operand:SI 1 "cris_general_operand_or_pic_source"
"r,Q>,M,M, I,r, M,n,!S,g,r,x, rQ>,x,gi"))]
;; Note that we prefer not to use the S alternative (if for some reason
;; it competes with others) above, but g matches S.
""
{
/* Better to have c-switch here; it is worth it to optimize the size of
move insns. The alternative would be to try to find more constraint
letters. FIXME: Check again. It seems this could shrink a bit. */
switch (which_alternative)
{
case 9:
if (TARGET_V32)
{
if (!flag_pic
&& (GET_CODE (operands[1]) == SYMBOL_REF
|| GET_CODE (operands[1]) == LABEL_REF
|| GET_CODE (operands[1]) == CONST))
{
/* FIXME: Express this through (set_attr cc none) instead,
since we can't express the ``none'' at this point. FIXME:
Use lapc for everything except const_int and when next cc0
user would want the flag setting. */
CC_STATUS_INIT;
return "lapc %1,%0";
}
if (flag_pic == 1
&& GET_CODE (operands[1]) == CONST
&& GET_CODE (XEXP (operands[1], 0)) == UNSPEC
&& XINT (XEXP (operands[1], 0), 1) == CRIS_UNSPEC_GOTREAD)
return "movu.w %1,%0";
}
/* FALLTHROUGH */
case 0:
case 1:
case 5:
case 10:
return "move.d %1,%0";
case 11:
case 12:
case 13:
case 14:
return "move %d1,%0";
case 2:
case 3:
case 6:
return "clear.d %0";
/* Constants -32..31 except 0. */
case 4:
return "moveq %1,%0";
/* We can win a little on constants -32768..-33, 32..65535. */
case 7:
if (INTVAL (operands[1]) > 0 && INTVAL (operands[1]) < 65536)
{
if (INTVAL (operands[1]) < 256)
return "movu.b %1,%0";
return "movu.w %1,%0";
}
else if (INTVAL (operands[1]) >= -32768 && INTVAL (operands[1]) < 32768)
{
if (INTVAL (operands[1]) >= -128 && INTVAL (operands[1]) < 128)
return "movs.b %1,%0";
return "movs.w %1,%0";
}
return "move.d %1,%0";
case 8:
{
rtx tem = operands[1];
gcc_assert (GET_CODE (tem) == CONST);
tem = XEXP (tem, 0);
if (GET_CODE (tem) == PLUS
&& GET_CODE (XEXP (tem, 0)) == UNSPEC
&& (XINT (XEXP (tem, 0), 1) == CRIS_UNSPEC_GOTREL
|| XINT (XEXP (tem, 0), 1) == CRIS_UNSPEC_PCREL)
&& CONST_INT_P (XEXP (tem, 1)))
tem = XEXP (tem, 0);
gcc_assert (GET_CODE (tem) == UNSPEC);
switch (XINT (tem, 1))
{
case CRIS_UNSPEC_GOTREAD:
case CRIS_UNSPEC_PLTGOTREAD:
/* Using sign-extend mostly to be consistent with the
indexed addressing mode. */
if (flag_pic == 1)
return "movs.w %1,%0";
return "move.d %1,%0";
case CRIS_UNSPEC_GOTREL:
case CRIS_UNSPEC_PLT_GOTREL:
gcc_assert (!TARGET_V32);
return "move.d %1,%0";
case CRIS_UNSPEC_PCREL:
case CRIS_UNSPEC_PLT_PCREL:
gcc_assert (TARGET_V32);
return "lapc %1,%0";
default:
gcc_unreachable ();
}
}
default:
return "BOGUS: %1 to %0";
}
}
[(set_attr "slottable" "yes,yes,yes,yes,yes,yes,no,no,no,no,no,yes,yes,no,no")
(set_attr "cc" "*,*,*,*,*,*,*,*,*,*,*,none,none,none,none")])
;; Extend operations with side-effect from mem to register, using
;; MOVS/MOVU. These are from mem to register only.
;;
;; [rx=ry+rz.S]
;;
;; QImode to HImode
;;
;; FIXME: Can we omit extend to HImode, since GCC should truncate for
;; HImode by itself? Perhaps use only anonymous modes?
(define_insn "*ext_sideqihi_biap"
[(set (match_operand:HI 0 "register_operand" "=r,r")
(match_operator:HI
5 "cris_extend_operator"
[(mem:QI (plus:SI
(mult:SI (match_operand:SI 1 "register_operand" "r,r")
(match_operand:SI 2 "const_int_operand" "n,n"))
(match_operand:SI 3 "register_operand" "r,r")))]))
(set (match_operand:SI 4 "register_operand" "=*3,r")
(plus:SI (mult:SI (match_dup 1)
(match_dup 2))
(match_dup 3)))]
"cris_side_effect_mode_ok (MULT, operands, 4, 3, 1, 2, 0)"
"@
#
mov%e5.%m5 [%4=%3+%1%T2],%0")
(define_insn "*ext_sidesi_biap"
[(set (match_operand:SI 0 "register_operand" "=r,r")
(match_operator:SI
5 "cris_extend_operator"
[(mem:BW (plus:SI
(mult:SI (match_operand:SI 1 "register_operand" "r,r")
(match_operand:SI 2 "const_int_operand" "n,n"))
(match_operand:SI 3 "register_operand" "r,r")))]))
(set (match_operand:SI 4 "register_operand" "=*3,r")
(plus:SI (mult:SI (match_dup 1)
(match_dup 2))
(match_dup 3)))]
"cris_side_effect_mode_ok (MULT, operands, 4, 3, 1, 2, 0)"
"@
#
mov%e5 [%4=%3+%1%T2],%0")
;; Same but [rx=ry+i]
;; QImode to HImode
(define_insn "*ext_sideqihi"
[(set (match_operand:HI 0 "register_operand" "=r,r,r,r,r")
(match_operator:HI
4 "cris_extend_operator"
[(mem:QI (plus:SI
(match_operand:SI 1 "cris_bdap_operand" "%r,r,r,R,R")
(match_operand:SI 2 "cris_bdap_operand" "r>Rn,r,>Rn,r,r")))]))
(set (match_operand:SI 3 "register_operand" "=*1,r,r,*2,r")
(plus:SI (match_dup 1)
(match_dup 2)))]
"cris_side_effect_mode_ok (PLUS, operands, 3, 1, 2, -1, 0)"
{
if ((which_alternative == 0 || which_alternative == 3)
&& (!CONST_INT_P (operands[2])
|| INTVAL (operands[2]) > 127
|| INTVAL (operands[2]) < -128
|| satisfies_constraint_N (operands[2])
|| satisfies_constraint_J (operands[2])))
return "#";
if (which_alternative == 4)
return "mov%e4.%m4 [%3=%2%S1],%0";
return "mov%e4.%m4 [%3=%1%S2],%0";
})
(define_insn "*ext_sidesi"
[(set (match_operand:SI 0 "register_operand" "=r,r,r,r,r")
(match_operator:SI
4 "cris_extend_operator"
[(mem:BW (plus:SI
(match_operand:SI 1 "cris_bdap_operand" "%r,r,r,R,R")
(match_operand:SI 2 "cris_bdap_operand" "r>Rn,r,>Rn,r,r")))]))
(set (match_operand:SI 3 "register_operand" "=*1,r,r,*2,r")
(plus:SI (match_dup 1)
(match_dup 2)))]
"cris_side_effect_mode_ok (PLUS, operands, 3, 1, 2, -1, 0)"
{
if ((which_alternative == 0 || which_alternative == 3)
&& (!CONST_INT_P (operands[2])
|| INTVAL (operands[2]) > 127
|| INTVAL (operands[2]) < -128
|| satisfies_constraint_N (operands[2])
|| satisfies_constraint_J (operands[2])))
return "#";
if (which_alternative == 4)
return "mov%e4 [%3=%2%S1],%0";
return "mov%e4 [%3=%1%S2],%0";
})
;; FIXME: See movsi.
(define_insn "movhi"
[(set
(match_operand:HI 0 "nonimmediate_operand" "=r,r, r,Q>,r,Q>,r,r,r,g,g,r,r,x")
(match_operand:HI 1 "general_operand" "r,Q>,M,M, I,r, L,O,n,M,r,g,x,r"))]
""
{
switch (which_alternative)
{
case 0:
case 1:
case 5:
case 10:
case 11:
return "move.w %1,%0";
case 12:
case 13:
return "move %1,%0";
case 2:
case 3:
case 9:
return "clear.w %0";
case 4:
return "moveq %1,%0";
case 6:
case 8:
if (INTVAL (operands[1]) < 256 && INTVAL (operands[1]) >= -128)
{
if (INTVAL (operands[1]) > 0)
return "movu.b %1,%0";
return "movs.b %1,%0";
}
return "move.w %1,%0";
case 7:
return "movEq %b1,%0";
default:
return "BOGUS: %1 to %0";
}
}
[(set_attr "slottable" "yes,yes,yes,yes,yes,yes,no,yes,no,no,no,no,yes,yes")
(set_attr "cc" "*,*,none,none,*,none,*,clobber,*,none,none,*,none,none")])
(define_insn "movstricthi"
[(set
(strict_low_part
(match_operand:HI 0 "nonimmediate_operand" "+r,r, r,Q>,Q>,g,r,g"))
(match_operand:HI 1 "general_operand" "r,Q>,M,M, r, M,g,r"))]
""
"@
move.w %1,%0
move.w %1,%0
clear.w %0
clear.w %0
move.w %1,%0
clear.w %0
move.w %1,%0
move.w %1,%0"
[(set_attr "slottable" "yes,yes,yes,yes,yes,no,no,no")])
(define_expand "reload_in"
[(set (match_operand:BW 2 "register_operand" "=r")
(match_operand:BW 1 "memory_operand" "m"))
(set (match_operand:BW 0 "register_operand" "=x")
(match_dup 2))]
""
"")
(define_expand "reload_out"
[(set (match_operand:BW 2 "register_operand" "=&r")
(match_operand:BW 1 "register_operand" "x"))
(set (match_operand:BW 0 "memory_operand" "=m")
(match_dup 2))]
""
"")
(define_insn "movqi"
[(set (match_operand:QI 0 "nonimmediate_operand" "=r,Q>,r, r,Q>,r,g,g,r,r,r,x")
(match_operand:QI 1 "general_operand" "r,r, Q>,M,M, I,M,r,O,g,x,r"))]
""
"@
move.b %1,%0
move.b %1,%0
move.b %1,%0
clear.b %0
clear.b %0
moveq %1,%0
clear.b %0
move.b %1,%0
moveq %b1,%0
move.b %1,%0
move %1,%0
move %1,%0"
[(set_attr "slottable" "yes,yes,yes,yes,yes,yes,no,no,yes,no,yes,yes")
(set_attr "cc" "*,*,*,*,*,*,*,*,clobber,*,none,none")])
(define_insn "movstrictqi"
[(set (strict_low_part
(match_operand:QI 0 "nonimmediate_operand" "+r,Q>,r, r,Q>,g,g,r"))
(match_operand:QI 1 "general_operand" "r,r, Q>,M,M, M,r,g"))]
""
"@
move.b %1,%0
move.b %1,%0
move.b %1,%0
clear.b %0
clear.b %0
clear.b %0
move.b %1,%0
move.b %1,%0"
[(set_attr "slottable" "yes,yes,yes,yes,yes,no,no,no")])
;; The valid "quick" bit-patterns are, except for 0.0, denormalized
;; values REALLY close to 0, and some NaN:s (I think; their exponent is
;; all ones); the worthwhile one is "0.0".
;; It will use clear, so we know ALL types of immediate 0 never change cc.
(define_insn "movsf"
[(set (match_operand:SF 0 "nonimmediate_operand" "=r,Q>,r, r,Q>,g,g,r,r,x,Q>,m,x, x")
(match_operand:SF 1 "general_operand" "r,r, Q>,G,G, G,r,g,x,r,x, x,Q>,g"))]
""
"@
move.d %1,%0
move.d %1,%0
move.d %1,%0
clear.d %0
clear.d %0
clear.d %0
move.d %1,%0
move.d %1,%0
move %1,%0
move %1,%0
move %1,%0
move %1,%0
move %1,%0
move %1,%0"
[(set_attr "slottable" "yes,yes,yes,yes,yes,no,no,no,yes,yes,yes,no,yes,no")])
;; Movem patterns. Primarily for use in function prologue and epilogue.
;; The V32 variants have an ordering matching the expectations of the
;; standard names "load_multiple" and "store_multiple"; pre-v32 movem
;; store R0 in the highest memory location.
(define_expand "load_multiple"
[(match_operand:SI 0 "register_operand" "")
(match_operand:SI 1 "memory_operand" "")
(match_operand:SI 2 "const_int_operand" "")]
"TARGET_V32"
{
rtx indreg;
/* Apparently the predicate isn't checked, so we need to do so
manually. Once happened for libstdc++-v3 locale_facets.tcc. */
if (!MEM_P (operands[1]))
FAIL;
indreg = XEXP (operands[1], 0);
if (GET_CODE (indreg) == POST_INC)
indreg = XEXP (indreg, 0);
if (!REG_P (indreg)
|| GET_CODE (operands[2]) != CONST_INT
|| !REG_P (operands[0])
|| REGNO (operands[0]) != 0
|| INTVAL (operands[2]) > CRIS_SP_REGNUM
|| (int) REGNO (indreg) < INTVAL (operands[2]))
FAIL;
gcc_unreachable ();
emit_insn (cris_gen_movem_load (operands[1], operands[2], 0));
DONE;
})
(define_expand "store_multiple"
[(match_operand:SI 0 "memory_operand" "")
(match_operand:SI 1 "register_operand" "")
(match_operand:SI 2 "const_int_operand" "")]
"TARGET_V32"
{
rtx indreg;
/* See load_multiple. */
if (!MEM_P (operands[0]))
FAIL;
indreg = XEXP (operands[0], 0);
if (GET_CODE (indreg) == POST_INC)
indreg = XEXP (indreg, 0);
if (!REG_P (indreg)
|| GET_CODE (operands[2]) != CONST_INT
|| !REG_P (operands[1])
|| REGNO (operands[1]) != 0
|| INTVAL (operands[2]) > CRIS_SP_REGNUM
|| (int) REGNO (indreg) < INTVAL (operands[2]))
FAIL;
gcc_unreachable ();
cris_emit_movem_store (operands[0], operands[2], 0, false);
DONE;
})
(define_insn "*cris_load_multiple"
[(match_parallel 0 "cris_load_multiple_op"
[(set (match_operand:SI 1 "register_operand" "=r,r")
(match_operand:SI 2 "memory_operand" "Q,m"))])]
""
"movem %O0,%o0"
[(set_attr "cc" "none")
(set_attr "slottable" "yes,no")
;; Not true, but setting the length to 0 causes return sequences (ret
;; movem) to have the cost they had when (return) included the movem
;; and reduces the performance penalty taken for needing to emit an
;; epilogue (in turn copied by bb-reorder) instead of return patterns.
;; FIXME: temporary change until all insn lengths are correctly
;; described. FIXME: have better target control over bb-reorder.
(set_attr "length" "0")])
(define_insn "*cris_store_multiple"
[(match_parallel 0 "cris_store_multiple_op"
[(set (match_operand:SI 2 "memory_operand" "=Q,m")
(match_operand:SI 1 "register_operand" "r,r"))])]
""
"movem %o0,%O0"
[(set_attr "cc" "none")
(set_attr "slottable" "yes,no")])
;; Sign- and zero-extend insns with standard names.
;; Those for integer source operand are ordered with the widest source
;; type first.
;; Sign-extend.
(define_insn "extendsidi2"
[(set (match_operand:DI 0 "register_operand" "=r")
(sign_extend:DI (match_operand:SI 1 "general_operand" "g")))]
""
"move.d %1,%M0\;smi %H0\;neg.d %H0,%H0")
(define_insn "extenddi2"
[(set (match_operand:DI 0 "register_operand" "=r")
(sign_extend:DI (match_operand:BW 1 "general_operand" "g")))]
""
"movs %1,%M0\;smi %H0\;neg.d %H0,%H0")
(define_insn "extendsi2"
[(set (match_operand:SI 0 "register_operand" "=r,r,r")
(sign_extend:SI (match_operand:BW 1 "general_operand" "r,Q>,g")))]
""
"movs %1,%0"
[(set_attr "slottable" "yes,yes,no")])
;; To do a byte->word extension, extend to dword, except that the top half
;; of the register will be clobbered. FIXME: Perhaps this is not needed.
(define_insn "extendqihi2"
[(set (match_operand:HI 0 "register_operand" "=r,r,r")
(sign_extend:HI (match_operand:QI 1 "general_operand" "r,Q>,g")))]
""
"movs.b %1,%0"
[(set_attr "slottable" "yes,yes,no")])
;; Zero-extend. The DImode ones are synthesized by gcc, so we don't
;; specify them here.
(define_insn "zero_extendsi2"
[(set (match_operand:SI 0 "register_operand" "=r,r,r")
(zero_extend:SI
(match_operand:BW 1 "nonimmediate_operand" "r,Q>,m")))]
""
"movu %1,%0"
[(set_attr "slottable" "yes,yes,no")])
;; Same comment as sign-extend QImode to HImode above applies.
(define_insn "zero_extendqihi2"
[(set (match_operand:HI 0 "register_operand" "=r,r,r")
(zero_extend:HI
(match_operand:QI 1 "nonimmediate_operand" "r,Q>,m")))]
""
"movu.b %1,%0"
[(set_attr "slottable" "yes,yes,no")])
;; All kinds of arithmetic and logical instructions.
;;
;; First, anonymous patterns to match addressing modes with
;; side-effects.
;;
;; op.S [rx=ry+I],rz; (add, sub, or, and, bound).
;;
;; [rx=ry+rz.S]
(define_insn "*op_side_biap"
[(set (match_operand:BWD 0 "register_operand" "=r,r")
(match_operator:BWD
6 "cris_orthogonal_operator"
[(match_operand:BWD 1 "register_operand" "0,0")
(mem:BWD (plus:SI
(mult:SI (match_operand:SI 2 "register_operand" "r,r")
(match_operand:SI 3 "const_int_operand" "n,n"))
(match_operand:SI 4 "register_operand" "r,r")))]))
(set (match_operand:SI 5 "register_operand" "=*4,r")
(plus:SI (mult:SI (match_dup 2)
(match_dup 3))
(match_dup 4)))]
"cris_side_effect_mode_ok (MULT, operands, 5, 4, 2, 3, 0)"
"@
#
%x6 [%5=%4+%2%T3],%0")
;; [rx=ry+i] ([%4=%2+%3])
(define_insn "*op_side"
[(set (match_operand:BWD 0 "register_operand" "=r,r,r,r,r")
(match_operator:BWD
5 "cris_orthogonal_operator"
[(match_operand:BWD 1 "register_operand" "0,0,0,0,0")
(mem:BWD (plus:SI
(match_operand:SI 2 "cris_bdap_operand" "%r,r,r,R,R")
(match_operand:SI 3 "cris_bdap_operand" "r>Rn,r,>Rn,r,r")))]))
(set (match_operand:SI 4 "register_operand" "=*2,r,r,*3,r")
(plus:SI (match_dup 2)
(match_dup 3)))]
"cris_side_effect_mode_ok (PLUS, operands, 4, 2, 3, -1, 0)"
{
if ((which_alternative == 0 || which_alternative == 3)
&& (!CONST_INT_P (operands[3])
|| INTVAL (operands[3]) > 127
|| INTVAL (operands[3]) < -128
|| satisfies_constraint_N (operands[3])
|| satisfies_constraint_J (operands[3])))
return "#";
if (which_alternative == 4)
return "%x5.%s0 [%4=%3%S2],%0";
return "%x5 [%4=%2%S3],%0";
})
;; To match all cases for commutative operations we may have to have the
;; following pattern for add, or & and. I do not know really, but it does
;; not break anything.
;;
;; FIXME: This really ought to be checked.
;;
;; op.S [rx=ry+I],rz;
;;
;; [rx=ry+rz.S]
(define_insn "*op_swap_side_biap"
[(set (match_operand:BWD 0 "register_operand" "=r,r")
(match_operator:BWD
6 "cris_commutative_orth_op"
[(mem:BWD (plus:SI
(mult:SI (match_operand:SI 2 "register_operand" "r,r")
(match_operand:SI 3 "const_int_operand" "n,n"))
(match_operand:SI 4 "register_operand" "r,r")))
(match_operand:BWD 1 "register_operand" "0,0")]))
(set (match_operand:SI 5 "register_operand" "=*4,r")
(plus:SI (mult:SI (match_dup 2)
(match_dup 3))
(match_dup 4)))]
"cris_side_effect_mode_ok (MULT, operands, 5, 4, 2, 3, 0)"
"@
#
%x6 [%5=%4+%2%T3],%0")
;; [rx=ry+i] ([%4=%2+%3])
;; FIXME: These could have anonymous mode for operand 0.
;; QImode
(define_insn "*op_swap_side"
[(set (match_operand:BWD 0 "register_operand" "=r,r,r,r,r")
(match_operator:BWD
5 "cris_commutative_orth_op"
[(mem:BWD
(plus:SI (match_operand:SI 2 "cris_bdap_operand" "%r,r,r,R,R")
(match_operand:SI 3 "cris_bdap_operand" "r>Rn,r,>Rn,r,r")))
(match_operand:BWD 1 "register_operand" "0,0,0,0,0")]))
(set (match_operand:SI 4 "register_operand" "=*2,r,r,*3,r")
(plus:SI (match_dup 2)
(match_dup 3)))]
"cris_side_effect_mode_ok (PLUS, operands, 4, 2, 3, -1, 0)"
{
if ((which_alternative == 0 || which_alternative == 3)
&& (!CONST_INT_P (operands[3])
|| INTVAL (operands[3]) > 127
|| INTVAL (operands[3]) < -128
|| satisfies_constraint_N (operands[3])
|| satisfies_constraint_J (operands[3])))
return "#";
if (which_alternative == 4)
return "%x5 [%4=%3%S2],%0";
return "%x5 [%4=%2%S3],%0";
})
;; Add operations, standard names.
;; Note that for the 'P' constraint, the high part can be -1 or 0. We
;; output the insn through the 'A' output modifier as "adds.w" and "addq",
;; respectively.
(define_expand "adddi3"
[(set (match_operand:DI 0 "register_operand")
(plus:DI (match_operand:DI 1 "register_operand")
(match_operand:DI 2 "general_operand")))]
""
{
if (MEM_P (operands[2]) && TARGET_V32)
operands[2] = force_reg (DImode, operands[2]);
})
(define_insn "*adddi3_non_v32"
[(set (match_operand:DI 0 "register_operand" "=r,r,r,&r,&r")
(plus:DI (match_operand:DI 1 "register_operand" "%0,0,0,0,r")
(match_operand:DI 2 "general_operand" "J,N,P,g,!To")))]
"!TARGET_V32"
"@
addq %2,%M0\;ax\;addq 0,%H0
subq %n2,%M0\;ax\;subq 0,%H0
add%e2.%z2 %2,%M0\;ax\;%A2 %H2,%H0
add.d %M2,%M0\;ax\;add.d %H2,%H0
add.d %M2,%M1,%M0\;ax\;add.d %H2,%H1,%H0")
; It seems no use allowing a memory operand for this one, because we'd
; need a scratch register for incrementing the address.
(define_insn "*adddi3_v32"
[(set (match_operand:DI 0 "register_operand" "=r,r,r,r,r")
(plus:DI (match_operand:DI 1 "register_operand" "%0,0,0,0,0")
(match_operand:DI 2 "nonmemory_operand" "J,N,P,r,n")))]
"TARGET_V32"
"@
addq %2,%M0\;addc 0,%H0
subq %n2,%M0\;ax\;subq 0,%H0
add%e2.%z2 %2,%M0\;addc %H2,%H0
add.d %M2,%M0\;addc %H2,%H0
add.d %M2,%M0\;addc %H2,%H0")
(define_expand "add3"
[(set (match_operand:BWD 0 "register_operand")
(plus:BWD
(match_operand:BWD 1 "register_operand")
(match_operand:BWD 2 "general_operand")))]
""
"")
(define_insn "*addsi3_non_v32"
[(set (match_operand:SI 0 "register_operand" "=r,r, r,r,r,r, r,r, r")
(plus:SI
(match_operand:SI 1 "register_operand" "%0,0, 0,0,0,0, 0,r, r")
(match_operand:SI 2 "general_operand" "r,Q>,J,N,n,!S,g,!To,0")))]
;; The last constraint is due to that after reload, the '%' is not
;; honored, and canonicalization doesn't care about keeping the same
;; register as in destination. This will happen after insn splitting.
;; gcc <= 2.7.2. FIXME: Check for gcc-2.9x
"!TARGET_V32"
{
switch (which_alternative)
{
case 0:
case 1:
return "add.d %2,%0";
case 2:
return "addq %2,%0";
case 3:
return "subq %n2,%0";
case 4:
/* 'Known value', but not in -63..63.
Check if addu/subu may be used. */
if (INTVAL (operands[2]) > 0)
{
if (INTVAL (operands[2]) < 256)
return "addu.b %2,%0";
if (INTVAL (operands[2]) < 65536)
return "addu.w %2,%0";
}
else
{
if (INTVAL (operands[2]) >= -255)
return "subu.b %n2,%0";
if (INTVAL (operands[2]) >= -65535)
return "subu.w %n2,%0";
}
return "add.d %2,%0";
case 5:
{
rtx tem = operands[2];
gcc_assert (GET_CODE (tem) == CONST);
tem = XEXP (tem, 0);
if (GET_CODE (tem) == PLUS
&& GET_CODE (XEXP (tem, 0)) == UNSPEC
/* We don't allow CRIS_UNSPEC_PCREL here; we can't have a
pc-relative operand in an add insn. */
&& XINT (XEXP (tem, 0), 1) == CRIS_UNSPEC_GOTREL
&& CONST_INT_P (XEXP (tem, 1)))
tem = XEXP (tem, 0);
gcc_assert (GET_CODE (tem) == UNSPEC);
switch (XINT (tem, 1))
{
case CRIS_UNSPEC_GOTREAD:
case CRIS_UNSPEC_PLTGOTREAD:
/* Using sign-extend mostly to be consistent with the
indexed addressing mode. */
if (flag_pic == 1)
return "adds.w %2,%0";
return "add.d %2,%0";
case CRIS_UNSPEC_PLT_GOTREL:
case CRIS_UNSPEC_GOTREL:
return "add.d %2,%0";
default:
gcc_unreachable ();
}
}
case 6:
return "add%u2 %2,%0";
case 7:
return "add.d %2,%1,%0";
case 8:
return "add.d %1,%0";
default:
return "BOGUS addsi %2+%1 to %0";
}
}
[(set_attr "slottable" "yes,yes,yes,yes,no,no,no,no,yes")])
; FIXME: Check what's best: having the three-operand ACR alternative
; before or after the corresponding-operand2 alternative. Check for
; *all* insns. FIXME: constant constraint letter for -128..127.
(define_insn "*addsi3_v32"
[(set (match_operand:SI 0 "register_operand" "=r,!a,r,!a, r,r,!a,r,!a,r,r,r,!a")
(plus:SI
(match_operand:SI 1 "register_operand" "%0,r, 0, r, 0,0,r, 0,r, 0,0,0,r")
(match_operand:SI 2 "general_operand" "r, r, Q>,Q>,J,N,NJ,L,L, P,n,g,g")))]
"TARGET_V32"
"@
add.d %2,%0
addi %2.b,%1,%0
add.d %2,%0
addo.d %2,%1,%0
addq %2,%0
subq %n2,%0
addoq %2,%1,%0
adds.w %2,%0
addo %2,%1,%0
addu.w %2,%0
add.d %2,%0
add%u2 %2,%0
addo.%Z2 %2,%1,%0"
[(set_attr "slottable" "yes,yes,yes,yes,yes,yes,yes,no,no,no,no,no,no")
(set_attr "cc" "*,none,*,none,*,*,none,*,none,*,*,*,none")])
(define_insn "*addhi3_non_v32"
[(set (match_operand:HI 0 "register_operand" "=r,r, r,r,r,r")
(plus:HI (match_operand:HI 1 "register_operand" "%0,0, 0,0,0,r")
(match_operand:HI 2 "general_operand" "r,Q>,J,N,g,!To")))]
"!TARGET_V32"
"@
add.w %2,%0
add.w %2,%0
addq %2,%0
subq %n2,%0
add.w %2,%0
add.w %2,%1,%0"
[(set_attr "slottable" "yes,yes,yes,yes,no,no")
(set_attr "cc" "normal,normal,clobber,clobber,normal,normal")])
(define_insn "*addhi3_v32"
[(set (match_operand:HI 0 "register_operand" "=r, !a,r,!a, r,r,!a,r,!a")
(plus:HI
(match_operand:HI 1 "register_operand" "%0,r, 0, r, 0,0,r, 0,r")
(match_operand:HI 2 "general_operand" "r, r, Q>,Q>,J,N,NJ,g,g")))]
"TARGET_V32"
"@
add.w %2,%0
addi %2.b,%1,%0
add.w %2,%0
addo.w %2,%1,%0
addq %2,%0
subq %n2,%0
addoq %2,%1,%0
add.w %2,%0
addo.w %2,%1,%0"
[(set_attr "slottable" "yes,yes,yes,yes,yes,yes,yes,no,no")
(set_attr "cc" "*,none,*,none,clobber,clobber,none,*,none")])
(define_insn "*addqi3_non_v32"
[(set (match_operand:QI 0 "register_operand" "=r,r, r,r,r,r,r")
(plus:QI (match_operand:QI 1 "register_operand" "%0,0, 0,0,0,0,r")
(match_operand:QI 2 "general_operand" "r,Q>,J,N,O,g,!To")))]
"!TARGET_V32"
"@
add.b %2,%0
add.b %2,%0
addq %2,%0
subq %n2,%0
subQ -%b2,%0
add.b %2,%0
add.b %2,%1,%0"
[(set_attr "slottable" "yes,yes,yes,yes,yes,no,no")
(set_attr "cc" "normal,normal,clobber,clobber,clobber,normal,normal")])
(define_insn "*addqi3_v32"
[(set (match_operand:QI 0 "register_operand" "=r,!a,r,!a, r,r,!a,r,r,!a")
(plus:QI
(match_operand:QI 1 "register_operand" "%0,r, 0, r, 0,0,r, 0,0,r")
(match_operand:QI 2 "general_operand" "r,r, Q>,Q>,J,N,NJ,O,g,g")))]
"TARGET_V32"
"@
add.b %2,%0
addi %2.b,%1,%0
add.b %2,%0
addo.b %2,%1,%0
addq %2,%0
subq %n2,%0
addoq %2,%1,%0
subQ -%b2,%0
add.b %2,%0
addo.b %2,%1,%0"
[(set_attr "slottable" "yes,yes,yes,yes,yes,yes,yes,yes,no,no")
(set_attr "cc" "*,none,*,none,clobber,clobber,none,clobber,*,none")])
;; Subtract.
;;
;; Note that because of insn canonicalization these will *seldom* but
;; rarely be used with a known constant as an operand.
;; Note that for the 'P' constraint, the high part can be -1 or 0. We
;; output the insn through the 'D' output modifier as "subs.w" and "subq",
;; respectively.
(define_expand "subdi3"
[(set (match_operand:DI 0 "register_operand")
(minus:DI (match_operand:DI 1 "register_operand")
(match_operand:DI 2 "general_operand")))]
""
{
if (TARGET_V32 && MEM_P (operands[2]))
operands[2] = force_reg (DImode, operands[2]);
})
(define_insn "*subdi3_non_v32"
[(set (match_operand:DI 0 "register_operand" "=r,r,r,&r,&r")
(minus:DI (match_operand:DI 1 "register_operand" "0,0,0,0,r")
(match_operand:DI 2 "general_operand" "J,N,P,g,!To")))]
"!TARGET_V32"
"@
subq %2,%M0\;ax\;subq 0,%H0
addq %n2,%M0\;ax\;addq 0,%H0
sub%e2.%z2 %2,%M0\;ax\;%D2 %H2,%H0
sub.d %M2,%M0\;ax\;sub.d %H2,%H0
sub.d %M2,%M1,%M0\;ax\;sub.d %H2,%H1,%H0")
(define_insn "*subdi3_v32"
[(set (match_operand:DI 0 "register_operand" "=r,r,r,&r")
(minus:DI (match_operand:DI 1 "register_operand" "0,0,0,0")
(match_operand:DI 2 "nonmemory_operand" "J,N,P,r")))]
"TARGET_V32"
"@
subq %2,%M0\;ax\;subq 0,%H0
addq %n2,%M0\;ax\;addq 0,%H0
sub%e2.%z2 %2,%M0\;ax\;%D2 %H2,%H0
sub.d %M2,%M0\;ax\;sub.d %H2,%H0")
(define_expand "sub3"
[(set (match_operand:BWD 0 "register_operand")
(minus:BWD
(match_operand:BWD 1 "register_operand")
(match_operand:BWD 2 "general_operand")))]
""
"")
(define_insn "*subsi3_non_v32"
[(set (match_operand:SI 0 "register_operand" "=r,r, r,r,r,r,r,r")
(minus:SI
(match_operand:SI 1 "register_operand" "0,0, 0,0,0,0,0,r")
(match_operand:SI 2 "general_operand" "r,Q>,J,N,P,n,g,!To")))]
"!TARGET_V32"
;; This does not do the optimal: "addu.w 65535,r0" when %2 is negative.
;; But then again, %2 should not be negative.
"@
sub.d %2,%0
sub.d %2,%0
subq %2,%0
addq %n2,%0
sub%e2.%z2 %2,%0
sub.d %2,%0
sub.d %2,%0
sub.d %2,%1,%0"
[(set_attr "slottable" "yes,yes,yes,yes,no,no,no,no")])
(define_insn "*subsi3_v32"
[(set (match_operand:SI 0 "register_operand" "=r,r,r,r,r,r,r")
(minus:SI
(match_operand:SI 1 "register_operand" "0,0,0,0,0,0,0")
(match_operand:SI 2 "general_operand" "r,Q>,J,N,P,n,g")))]
"TARGET_V32"
"@
sub.d %2,%0
sub.d %2,%0
subq %2,%0
addq %n2,%0
sub%e2.%z2 %2,%0
sub.d %2,%0
sub.d %2,%0"
[(set_attr "slottable" "yes,yes,yes,yes,no,no,no")])
(define_insn "*sub3_nonv32"
[(set (match_operand:BW 0 "register_operand" "=r,r, r,r,r,r")
(minus:BW (match_operand:BW 1 "register_operand" "0,0, 0,0,0,r")
(match_operand:BW 2 "general_operand" "r,Q>,J,N,g,!To")))]
"!TARGET_V32"
"@
sub %2,%0
sub %2,%0
subq %2,%0
addq %n2,%0
sub %2,%0
sub %2,%1,%0"
[(set_attr "slottable" "yes,yes,yes,yes,no,no")
(set_attr "cc" "normal,normal,clobber,clobber,normal,normal")])
(define_insn "*sub3_v32"
[(set (match_operand:BW 0 "register_operand" "=r,r,r,r,r")
(minus:BW (match_operand:BW 1 "register_operand" "0,0,0,0,0")
(match_operand:BW 2 "general_operand" "r,Q>,J,N,g")))]
"TARGET_V32"
"@
sub %2,%0
sub %2,%0
subq %2,%0
addq %n2,%0
sub %2,%0"
[(set_attr "slottable" "yes,yes,yes,yes,no")
(set_attr "cc" "normal,normal,clobber,clobber,normal")])
;; CRIS has some add/sub-with-sign/zero-extend instructions.
;; Although these perform sign/zero-extension to SImode, they are
;; equally applicable for the HImode case.
;; FIXME: Check; GCC should handle the widening.
;; Note that these must be located after the normal add/sub patterns,
;; so not to get constants into any less specific operands.
;;
;; Extend with add/sub and side-effect.
;;
;; ADDS/SUBS/ADDU/SUBU and BOUND, which needs a check for zero_extend
;;
;; adds/subs/addu/subu bound [rx=ry+rz.S]
;; QImode to HImode
;; FIXME: GCC should widen.
(define_insn "*extopqihi_side_biap"
[(set (match_operand:HI 0 "register_operand" "=r,r")
(match_operator:HI
6 "cris_additive_operand_extend_operator"
[(match_operand:HI 1 "register_operand" "0,0")
(match_operator:HI
7 "cris_extend_operator"
[(mem:QI (plus:SI
(mult:SI (match_operand:SI 2 "register_operand" "r,r")
(match_operand:SI 3 "const_int_operand" "n,n"))
(match_operand:SI 4 "register_operand" "r,r")))])]))
(set (match_operand:SI 5 "register_operand" "=*4,r")
(plus:SI (mult:SI (match_dup 2)
(match_dup 3))
(match_dup 4)))]
"cris_side_effect_mode_ok (MULT, operands, 5, 4, 2, 3, 0)"
"@
#
%x6%e7.%m7 [%5=%4+%2%T3],%0")
(define_insn "*extopsi_side_biap"
[(set (match_operand:SI 0 "register_operand" "=r,r")
(match_operator:SI
6 "cris_operand_extend_operator"
[(match_operand:SI 1 "register_operand" "0,0")
(match_operator:SI
7 "cris_extend_operator"
[(mem:BW (plus:SI
(mult:SI (match_operand:SI 2 "register_operand" "r,r")
(match_operand:SI 3 "const_int_operand" "n,n"))
(match_operand:SI 4 "register_operand" "r,r")))])]))
(set (match_operand:SI 5 "register_operand" "=*4,r")
(plus:SI (mult:SI (match_dup 2)
(match_dup 3))
(match_dup 4)))]
"(GET_CODE (operands[6]) != UMIN || GET_CODE (operands[7]) == ZERO_EXTEND)
&& cris_side_effect_mode_ok (MULT, operands, 5, 4, 2, 3, 0)"
"@
#
%x6%e7 [%5=%4+%2%T3],%0")
;; [rx=ry+i]
;; QImode to HImode
(define_insn "*extopqihi_side"
[(set (match_operand:HI 0 "register_operand" "=r,r,r,r,r")
(match_operator:HI
5 "cris_additive_operand_extend_operator"
[(match_operand:HI 1 "register_operand" "0,0,0,0,0")
(match_operator:HI
6 "cris_extend_operator"
[(mem:QI
(plus:SI (match_operand:SI 2 "cris_bdap_operand" "%r,r,r,R,R")
(match_operand:SI 3 "cris_bdap_operand" "r>Rn,r,>Rn,r,r")
))])]))
(set (match_operand:SI 4 "register_operand" "=*2,r,r,*3,r")
(plus:SI (match_dup 2)
(match_dup 3)))]
"cris_side_effect_mode_ok (PLUS, operands, 4, 2, 3, -1, 0)"
{
if ((which_alternative == 0 || which_alternative == 3)
&& (!CONST_INT_P (operands[3])
|| INTVAL (operands[3]) > 127
|| INTVAL (operands[3]) < -128
|| satisfies_constraint_N (operands[3])
|| satisfies_constraint_J (operands[3])))
return "#";
if (which_alternative == 4)
return "%x5%E6.%m6 [%4=%3%S2],%0";
return "%x5%E6.%m6 [%4=%2%S3],%0";
})
(define_insn "*extopsi_side"
[(set (match_operand:SI 0 "register_operand" "=r,r,r,r,r")
(match_operator:SI
5 "cris_operand_extend_operator"
[(match_operand:SI 1 "register_operand" "0,0,0,0,0")
(match_operator:SI
6 "cris_extend_operator"
[(mem:BW
(plus:SI (match_operand:SI 2 "cris_bdap_operand" "%r,r,r,R,R")
(match_operand:SI 3 "cris_bdap_operand" "r>Rn,r,>Rn,r,r")
))])]))
(set (match_operand:SI 4 "register_operand" "=*2,r,r,*3,r")
(plus:SI (match_dup 2)
(match_dup 3)))]
"(GET_CODE (operands[5]) != UMIN || GET_CODE (operands[6]) == ZERO_EXTEND)
&& cris_side_effect_mode_ok (PLUS, operands, 4, 2, 3, -1, 0)"
{
if ((which_alternative == 0 || which_alternative == 3)
&& (!CONST_INT_P (operands[3])
|| INTVAL (operands[3]) > 127
|| INTVAL (operands[3]) < -128
|| satisfies_constraint_N (operands[3])
|| satisfies_constraint_J (operands[3])))
return "#";
if (which_alternative == 4)
return "%x5%E6 [%4=%3%S2],%0";
return "%x5%E6 [%4=%2%S3],%0";
})
;; As with op.S we may have to add special pattern to match commuted
;; operands to adds/addu and bound
;;
;; adds/addu/bound [rx=ry+rz.S]
;; QImode to HImode
;; FIXME: GCC should widen.
(define_insn "*extopqihi_swap_side_biap"
[(set (match_operand:HI 0 "register_operand" "=r,r")
(plus:HI
(match_operator:HI
6 "cris_extend_operator"
[(mem:QI (plus:SI
(mult:SI (match_operand:SI 2 "register_operand" "r,r")
(match_operand:SI 3 "const_int_operand" "n,n"))
(match_operand:SI 4 "register_operand" "r,r")))])
(match_operand:HI 1 "register_operand" "0,0")))
(set (match_operand:SI 5 "register_operand" "=*4,r")
(plus:SI (mult:SI (match_dup 2)
(match_dup 3))
(match_dup 4)))]
"cris_side_effect_mode_ok (MULT, operands, 5, 4, 2, 3, 0)"
"@
#
add%e6.b [%5=%4+%2%T3],%0")
(define_insn "*extopsi_swap_side_biap"
[(set (match_operand:SI 0 "register_operand" "=r,r")
(match_operator:SI
7 "cris_plus_or_bound_operator"
[(match_operator:SI
6 "cris_extend_operator"
[(mem:BW (plus:SI
(mult:SI (match_operand:SI 2 "register_operand" "r,r")
(match_operand:SI 3 "const_int_operand" "n,n"))
(match_operand:SI 4 "register_operand" "r,r")))])
(match_operand:SI 1 "register_operand" "0,0")]))
(set (match_operand:SI 5 "register_operand" "=*4,r")
(plus:SI (mult:SI (match_dup 2)
(match_dup 3))
(match_dup 4)))]
"(GET_CODE (operands[7]) != UMIN || GET_CODE (operands[6]) == ZERO_EXTEND)
&& cris_side_effect_mode_ok (MULT, operands, 5, 4, 2, 3, 0)"
"@
#
%x7%E6 [%5=%4+%2%T3],%0")
;; [rx=ry+i]
;; FIXME: GCC should widen.
;; QImode to HImode
(define_insn "*extopqihi_swap_side"
[(set (match_operand:HI 0 "register_operand" "=r,r,r,r,r")
(plus:HI
(match_operator:HI
5 "cris_extend_operator"
[(mem:QI (plus:SI
(match_operand:SI 2 "cris_bdap_operand" "%r,r,r,R,R")
(match_operand:SI 3 "cris_bdap_operand" "r>Rn,r,>Rn,r,r")))])
(match_operand:HI 1 "register_operand" "0,0,0,0,0")))
(set (match_operand:SI 4 "register_operand" "=*2,r,r,*3,r")
(plus:SI (match_dup 2)
(match_dup 3)))]
"cris_side_effect_mode_ok (PLUS, operands, 4, 2, 3, -1, 0)"
{
if ((which_alternative == 0 || which_alternative == 3)
&& (!CONST_INT_P (operands[3])
|| INTVAL (operands[3]) > 127
|| INTVAL (operands[3]) < -128
|| satisfies_constraint_N (operands[3])
|| satisfies_constraint_J (operands[3])))
return "#";
if (which_alternative == 4)
return "add%e5.b [%4=%3%S2],%0";
return "add%e5.b [%4=%2%S3],%0";
})
(define_insn "*extopsi_swap_side"
[(set (match_operand:SI 0 "register_operand" "=r,r,r,r,r")
(match_operator:SI
6 "cris_plus_or_bound_operator"
[(match_operator:SI
5 "cris_extend_operator"
[(mem:BW (plus:SI
(match_operand:SI 2 "cris_bdap_operand" "%r,r,r,R,R")
(match_operand:SI 3 "cris_bdap_operand" "r>Rn,r,>Rn,r,r")))])
(match_operand:SI 1 "register_operand" "0,0,0,0,0")]))
(set (match_operand:SI 4 "register_operand" "=*2,r,r,*3,r")
(plus:SI (match_dup 2)
(match_dup 3)))]
"(GET_CODE (operands[6]) != UMIN || GET_CODE (operands[5]) == ZERO_EXTEND)
&& cris_side_effect_mode_ok (PLUS, operands, 4, 2, 3, -1, 0)"
{
if ((which_alternative == 0 || which_alternative == 3)
&& (!CONST_INT_P (operands[3])
|| INTVAL (operands[3]) > 127
|| INTVAL (operands[3]) < -128
|| satisfies_constraint_N (operands[3])
|| satisfies_constraint_J (operands[3])))
return "#";
if (which_alternative == 4)
return \"%x6%E5.%m5 [%4=%3%S2],%0\";
return "%x6%E5 [%4=%2%S3],%0";
})
;; Extend versions (zero/sign) of normal add/sub (no side-effects).
;; QImode to HImode
;; FIXME: GCC should widen.
(define_insn "*extopqihi_non_v32"
[(set (match_operand:HI 0 "register_operand" "=r,r,r,r")
(match_operator:HI
3 "cris_additive_operand_extend_operator"
[(match_operand:HI 1 "register_operand" "0,0,0,r")
(match_operator:HI
4 "cris_extend_operator"
[(match_operand:QI 2 "nonimmediate_operand" "r,Q>,m,!To")])]))]
"!TARGET_V32 && GET_MODE_SIZE (GET_MODE (operands[0])) <= UNITS_PER_WORD
&& (operands[1] != frame_pointer_rtx || GET_CODE (operands[3]) != PLUS)"
"@
%x3%E4.%m4 %2,%0
%x3%E4.%m4 %2,%0
%x3%E4.%m4 %2,%0
%x3%E4.%m4 %2,%1,%0"
[(set_attr "slottable" "yes,yes,no,no")
(set_attr "cc" "clobber")])
(define_insn "*extopqihi_v32"
[(set (match_operand:HI 0 "register_operand" "=r,r")
(match_operator:HI
3 "cris_additive_operand_extend_operator"
[(match_operand:HI 1 "register_operand" "0,0")
(match_operator:HI
4 "cris_extend_operator"
[(match_operand:QI 2 "nonimmediate_operand" "r,m")])]))]
"TARGET_V32"
"%x3%e4.%m4 %2,%0"
[(set_attr "slottable" "yes")
(set_attr "cc" "clobber")])
;; QImode to SImode
(define_insn "*extopsi_non_v32"
[(set (match_operand:SI 0 "register_operand" "=r,r,r,r")
(match_operator:SI
3 "cris_operand_extend_operator"
[(match_operand:SI 1 "register_operand" "0,0,0,r")
(match_operator:SI
4 "cris_extend_operator"
[(match_operand:BW 2 "nonimmediate_operand" "r,Q>,m,!To")])]))]
"!TARGET_V32
&& (GET_CODE (operands[3]) != UMIN || GET_CODE (operands[4]) == ZERO_EXTEND)
&& GET_MODE_SIZE (GET_MODE (operands[0])) <= UNITS_PER_WORD
&& (operands[1] != frame_pointer_rtx || GET_CODE (operands[3]) != PLUS)"
"@
%x3%E4 %2,%0
%x3%E4 %2,%0
%x3%E4 %2,%0
%x3%E4 %2,%1,%0"
[(set_attr "slottable" "yes,yes,no,no")])
(define_insn "*extopsi_v32"
[(set (match_operand:SI 0 "register_operand" "=r,r")
(match_operator:SI
3 "cris_additive_operand_extend_operator"
[(match_operand:SI 1 "register_operand" "0,0")
(match_operator:SI
4 "cris_extend_operator"
[(match_operand:BW 2 "nonimmediate_operand" "r,m")])]))]
"TARGET_V32"
"%x3%e4.%m4 %2,%0"
[(set_attr "slottable" "yes")])
;; As with the side-effect patterns, may have to have swapped operands for add.
;; For commutative operands, these are the canonical forms.
;; QImode to HImode
(define_insn "*addxqihi_swap_non_v32"
[(set (match_operand:HI 0 "register_operand" "=r,r,r,r")
(plus:HI
(match_operator:HI
3 "cris_extend_operator"
[(match_operand:QI 2 "nonimmediate_operand" "r,Q>,m,!To")])
(match_operand:HI 1 "register_operand" "0,0,0,r")))]
"!TARGET_V32 && operands[1] != frame_pointer_rtx"
"@
add%e3.b %2,%0
add%e3.b %2,%0
add%e3.b %2,%0
add%e3.b %2,%1,%0"
[(set_attr "slottable" "yes,yes,no,no")
(set_attr "cc" "clobber")])
;; A case for v32, to catch the "addo" insn in addition to "adds". We
;; only care to match the canonical form; there should be no other.
(define_insn "*addsbw_v32"
[(set (match_operand:HI 0 "register_operand" "=r,r,!a")
(plus:HI
(sign_extend:HI
(match_operand:QI 2 "nonimmediate_operand" "r,m,m"))
(match_operand:HI 1 "register_operand" "0,0,r")))]
"TARGET_V32"
"@
adds.b %2,%0
adds.b %2,%0
addo.b %2,%1,%0"
[(set_attr "slottable" "yes")
(set_attr "cc" "clobber,clobber,none")])
(define_insn "*addubw_v32"
[(set (match_operand:HI 0 "register_operand" "=r,r")
(plus:HI
(zero_extend:HI
(match_operand:QI 2 "nonimmediate_operand" "r,m"))
(match_operand:HI 1 "register_operand" "0,0")))]
"TARGET_V32"
"addu.b %2,%0"
[(set_attr "slottable" "yes")
(set_attr "cc" "clobber")])
(define_insn "*extopsi_swap_non_v32"
[(set (match_operand:SI 0 "register_operand" "=r,r,r,r")
(match_operator:SI
4 "cris_plus_or_bound_operator"
[(match_operator:SI
3 "cris_extend_operator"
[(match_operand:BW 2 "nonimmediate_operand" "r,Q>,m,!To")])
(match_operand:SI 1 "register_operand" "0,0,0,r")]))]
"!TARGET_V32
&& (GET_CODE (operands[4]) != UMIN || GET_CODE (operands[3]) == ZERO_EXTEND)
&& operands[1] != frame_pointer_rtx"
"@
%x4%E3 %2,%0
%x4%E3 %2,%0
%x4%E3 %2,%0
%x4%E3 %2,%1,%0"
[(set_attr "slottable" "yes,yes,no,no")])
(define_insn "*adds_v32"
[(set (match_operand:SI 0 "register_operand" "=r,r,!a")
(plus:SI
(sign_extend:SI
(match_operand:BW 2 "nonimmediate_operand" "r,m,m"))
(match_operand:SI 1 "register_operand" "0,0,r")))]
"TARGET_V32"
"@
adds %2,%0
adds %2,%0
addo %2,%1,%0"
[(set_attr "slottable" "yes")
(set_attr "cc" "*,*,none")])
(define_insn "*addu_v32"
[(set (match_operand:SI 0 "register_operand" "=r,r")
(plus:SI
(zero_extend:SI
(match_operand:BW 2 "nonimmediate_operand" "r,m"))
(match_operand:SI 1 "register_operand" "0,0")))]
"TARGET_V32 && operands[1] != frame_pointer_rtx"
"addu %2,%0"
[(set_attr "slottable" "yes")])
(define_insn "*bound_v32"
[(set (match_operand:SI 0 "register_operand" "=r")
(umin:SI
(zero_extend:SI
(match_operand:BW 2 "register_operand" "r"))
(match_operand:SI 1 "register_operand" "0")))]
"TARGET_V32 && operands[1] != frame_pointer_rtx"
"bound %2,%0"
[(set_attr "slottable" "yes")])
;; This is the special case when we use what corresponds to the
;; instruction above in "casesi". Do *not* change it to use the generic
;; pattern and "REG 15" as pc; I did that and it led to madness and
;; maintenance problems: Instead of (as imagined) recognizing and removing
;; or replacing this pattern with something simpler, other variant
;; patterns were recognized or combined, including some prefix variants
;; where the value in pc is not that of the next instruction (which means
;; this instruction actually *is* special and *should* be marked as such).
;; When switching from the "generic pattern match" approach to this simpler
;; approach, there were insignificant differences in gcc, ipps and
;; product code, somehow due to scratching reload behind the ear or
;; something. Testcase "gcc" looked .01% slower and 4 bytes bigger;
;; product code became .001% smaller but "looked better". The testcase
;; "ipps" was just different at register allocation).
;;
;; Assumptions in the jump optimizer forces us to use IF_THEN_ELSE in this
;; pattern with the default-label as the else, with the "if" being
;; index-is-less-than the max number of cases plus one. The default-label
;; is attached to the end of the case-table at time of output.
(define_insn "*casesi_adds_w"
[(set (pc)
(if_then_else
(ltu (match_operand:SI 0 "register_operand" "r")
(match_operand:SI 1 "const_int_operand" "n"))
(plus:SI (sign_extend:SI
(mem:HI
(plus:SI (mult:SI (match_dup 0) (const_int 2))
(pc))))
(pc))
(label_ref (match_operand 2 "" ""))))
(use (label_ref (match_operand 3 "" "")))]
"!TARGET_V32 && operands[0] != frame_pointer_rtx"
"adds.w [$pc+%0.w],$pc"
[(set_attr "cc" "clobber")])
;; For V32, we just have a jump, but we need to mark the table as used,
;; and the jump insn must have the if_then_else form expected by core
;; GCC. Since we don't want to prolong the lifetime of the original
;; index value, we compare against "unspec 0". It's a pity we have to
;; jump through to get the default label in place and to keep the jump
;; table around. FIXME: Look into it some time.
(define_insn "*casesi_jump_v32"
[(set (pc)
(if_then_else
(ltu (unspec [(const_int 0)] CRIS_UNSPEC_CASESI)
(match_operand:SI 0 "const_int_operand" "n"))
(match_operand:SI 1 "register_operand" "r")
(label_ref (match_operand 2 "" ""))))
(use (label_ref (match_operand 3 "" "")))]
"TARGET_V32"
"jump %1%#"
[(set_attr "cc" "clobber")
(set_attr "slottable" "has_slot")])
;; Multiply instructions.
;; Sometimes powers of 2 (which are normally canonicalized to a
;; left-shift) appear here, as a result of address reloading.
;; As a special, for values 3 and 5, we can match with an addi, so add those.
;;
;; FIXME: This may be unnecessary now.
;; Explicitly named for convenience of having a gen_... function.
(define_insn "addi_mul"
[(set (match_operand:SI 0 "register_operand" "=r")
(mult:SI
(match_operand:SI 1 "register_operand" "%0")
(match_operand:SI 2 "const_int_operand" "n")))]
"operands[0] != frame_pointer_rtx
&& operands[1] != frame_pointer_rtx
&& CONST_INT_P (operands[2])
&& (INTVAL (operands[2]) == 2
|| INTVAL (operands[2]) == 4 || INTVAL (operands[2]) == 3
|| INTVAL (operands[2]) == 5)"
{
if (INTVAL (operands[2]) == 2)
return "lslq 1,%0";
else if (INTVAL (operands[2]) == 4)
return "lslq 2,%0";
else if (INTVAL (operands[2]) == 3)
return "addi %0.w,%0";
else if (INTVAL (operands[2]) == 5)
return "addi %0.d,%0";
return "BAD: adr_mulsi: %0=%1*%2";
}
[(set_attr "slottable" "yes")
;; No flags are changed if this insn is "addi", but it does not seem
;; worth the trouble to distinguish that to the lslq cases.
(set_attr "cc" "clobber")])
;; The addi insn as it is normally used.
;; Make the ACR alternative taste bad enough to not choose it as a
;; preference to avoid spilling problems (unwind-dw2-fde.c at build).
;; FIXME: Revisit for new register allocator.
(define_insn "*addi"
[(set (match_operand:SI 0 "register_operand" "=r,!a")
(plus:SI
(mult:SI (match_operand:SI 2 "register_operand" "r,r")
(match_operand:SI 3 "const_int_operand" "n,n"))
(match_operand:SI 1 "register_operand" "0,r")))]
"operands[0] != frame_pointer_rtx
&& operands[1] != frame_pointer_rtx
&& CONST_INT_P (operands[3])
&& (INTVAL (operands[3]) == 1
|| INTVAL (operands[3]) == 2 || INTVAL (operands[3]) == 4)"
"@
addi %2%T3,%0
addi %2%T3,%1,%0"
[(set_attr "slottable" "yes")
(set_attr "cc" "none")])
;; The mstep instruction. Probably not useful by itself; it's to
;; non-linear wrt. the other insns. We used to expand to it, so at least
;; it's correct.
(define_insn "mstep_shift"
[(set (match_operand:SI 0 "register_operand" "=r")
(if_then_else:SI
(lt:SI (cc0) (const_int 0))
(plus:SI (ashift:SI (match_operand:SI 1 "register_operand" "0")
(const_int 1))
(match_operand:SI 2 "register_operand" "r"))
(ashift:SI (match_operand:SI 3 "register_operand" "0")
(const_int 1))))]
"!TARGET_V32"
"mstep %2,%0"
[(set_attr "slottable" "yes")])
;; When illegitimate addresses are legitimized, sometimes gcc forgets
;; to canonicalize the multiplications.
;;
;; FIXME: Check gcc > 2.7.2, remove and possibly fix in gcc.
(define_insn "mstep_mul"
[(set (match_operand:SI 0 "register_operand" "=r")
(if_then_else:SI
(lt:SI (cc0) (const_int 0))
(plus:SI (mult:SI (match_operand:SI 1 "register_operand" "0")
(const_int 2))
(match_operand:SI 2 "register_operand" "r"))
(mult:SI (match_operand:SI 3 "register_operand" "0")
(const_int 2))))]
"!TARGET_V32
&& operands[0] != frame_pointer_rtx
&& operands[1] != frame_pointer_rtx
&& operands[2] != frame_pointer_rtx
&& operands[3] != frame_pointer_rtx"
"mstep %2,%0"
[(set_attr "slottable" "yes")])
(define_insn "mul3"
[(set (match_operand:WD 0 "register_operand" "=r")
(mult:WD
(szext:WD (match_operand: 1 "register_operand" "%0"))
(szext:WD (match_operand: 2 "register_operand" "r"))))
(clobber (match_scratch:SI 3 "=h"))]
"TARGET_HAS_MUL_INSNS"
"%!mul %2,%0"
[(set (attr "slottable")
(if_then_else (match_test "TARGET_MUL_BUG")
(const_string "no")
(const_string "yes")))
;; For umuls.[bwd] it's just N unusable here, but let's be safe.
;; For muls.b, this really extends to SImode, so cc should be
;; considered clobbered.
;; For muls.w, it's just N unusable here, but let's be safe.
(set_attr "cc" "clobber")])
;; Note that gcc does not make use of such a thing as umulqisi3. It gets
;; confused and will erroneously use it instead of umulhisi3, failing (at
;; least) gcc.c-torture/execute/arith-rand.c at all optimization levels.
;; Inspection of optab code shows that there must be only one widening
;; multiplication per mode widened to.
(define_insn "mulsi3"
[(set (match_operand:SI 0 "register_operand" "=r")
(mult:SI (match_operand:SI 1 "register_operand" "%0")
(match_operand:SI 2 "register_operand" "r")))
(clobber (match_scratch:SI 3 "=h"))]
"TARGET_HAS_MUL_INSNS"
"%!muls.d %2,%0"
[(set (attr "slottable")
(if_then_else (match_test "TARGET_MUL_BUG")
(const_string "no")
(const_string "yes")))
;; Just N unusable here, but let's be safe.
(set_attr "cc" "clobber")])
;; A few multiply variations.
;; When needed, we can get the high 32 bits from the overflow
;; register. We don't care to split and optimize these.
;;
;; Note that cc0 is still valid after the move-from-overflow-register
;; insn; no special precaution need to be taken in cris_notice_update_cc.
(define_insn "mulsidi3"
[(set (match_operand:DI 0 "register_operand" "=r")
(mult:DI
(szext:DI (match_operand:SI 1 "register_operand" "%0"))
(szext:DI (match_operand:SI 2 "register_operand" "r"))))
(clobber (match_scratch:SI 3 "=h"))]
"TARGET_HAS_MUL_INSNS"
"%!mul.d %2,%M0\;move $mof,%H0")
;; These two patterns may be expressible by other means, perhaps by making
;; [u]?mulsidi3 a define_expand.
;; Due to register allocation braindamage, the clobber 1,2 alternatives
;; cause a move into the clobbered register *before* the insn, then
;; after the insn, mof is moved too, rather than the clobber assigned
;; the last mof target. This became apparent when making MOF and SRP
;; visible registers, with the necessary tweak to smulsi3_highpart.
;; Because these patterns are used in division by constants, that damage
;; is visible (ipps regression tests). Therefore the last two
;; alternatives, "helping" reload to avoid an unnecessary move, but
;; punished by force of one "?". Check code from "int d (int a) {return
;; a / 1000;}" and unsigned. FIXME: Comment above was for 3.2, revisit.
(define_insn "mulsi3_highpart"
[(set (match_operand:SI 0 "register_operand" "=h,h,?r,?r")
(truncate:SI
(lshiftrt:DI
(mult:DI
(szext:DI (match_operand:SI 1 "register_operand" "r,r,0,r"))
(szext:DI (match_operand:SI 2 "register_operand" "r,r,r,0")))
(const_int 32))))
(clobber (match_scratch:SI 3 "=1,2,h,h"))]
"TARGET_HAS_MUL_INSNS"
"@
%!mul.d %2,%1
%!mul.d %1,%2
%!mul.d %2,%1\;move $mof,%0
%!mul.d %1,%2\;move $mof,%0"
[(set_attr "slottable" "yes,yes,no,no")
(set_attr "cc" "clobber")])
;; Divide and modulus instructions. CRIS only has a step instruction.
(define_insn "dstep_shift"
[(set (match_operand:SI 0 "register_operand" "=r")
(if_then_else:SI
(geu:SI (ashift:SI (match_operand:SI 1 "register_operand" "0")
(const_int 1))
(match_operand:SI 2 "register_operand" "r"))
(minus:SI (ashift:SI (match_operand:SI 3 "register_operand" "0")
(const_int 1))
(match_operand:SI 4 "register_operand" "2"))
(ashift:SI (match_operand:SI 5 "register_operand" "0")
(const_int 1))))]
""
"dstep %2,%0"
[(set_attr "slottable" "yes")
(set_attr "cc" "noov32")])
;; Here's a variant with mult instead of ashift.
;;
;; FIXME: This should be investigated. Which one matches through combination?
(define_insn "dstep_mul"
[(set (match_operand:SI 0 "register_operand" "=r")
(if_then_else:SI
(geu:SI (mult:SI (match_operand:SI 1 "register_operand" "0")
(const_int 2))
(match_operand:SI 2 "register_operand" "r"))
(minus:SI (mult:SI (match_operand:SI 3 "register_operand" "0")
(const_int 2))
(match_operand:SI 4 "register_operand" "2"))
(mult:SI (match_operand:SI 5 "register_operand" "0")
(const_int 2))))]
"operands[0] != frame_pointer_rtx
&& operands[1] != frame_pointer_rtx
&& operands[2] != frame_pointer_rtx
&& operands[3] != frame_pointer_rtx"
"dstep %2,%0"
[(set_attr "slottable" "yes")
(set_attr "cc" "noov32")])
;; Logical operators.
;; Bitwise "and".
;; There is no use in defining "anddi3", because gcc can expand this by
;; itself, and make reasonable code without interference.
;; If the first operand is memory or a register and is the same as the
;; second operand, and the third operand is -256 or -65536, we can use
;; CLEAR instead. Or, if the first operand is a register, and the third
;; operand is 255 or 65535, we can zero_extend.
;; GCC isn't smart enough to recognize these cases (yet), and they seem
;; to be common enough to be worthwhile.
;; FIXME: This should be made obsolete.
(define_expand "andsi3"
[(set (match_operand:SI 0 "nonimmediate_operand" "")
(and:SI (match_operand:SI 1 "nonimmediate_operand" "")
(match_operand:SI 2 "general_operand" "")))]
""
{
if (! (CONST_INT_P (operands[2])
&& (((INTVAL (operands[2]) == -256
|| INTVAL (operands[2]) == -65536)
&& rtx_equal_p (operands[1], operands[0]))
|| ((INTVAL (operands[2]) == 255
|| INTVAL (operands[2]) == 65535)
&& REG_P (operands[0])))))
{
/* Make intermediate steps if operand0 is not a register or
operand1 is not a register, and hope that the reload pass will
make something useful out of it. Note that the operands are
*not* canonicalized. For the moment, I chicken out on this,
because all or most ports do not describe 'and' with
canonicalized operands, and I seem to remember magic in reload,
checking that operand1 has constraint '%0', in which case
operand0 and operand1 must have similar predicates.
FIXME: Investigate. */
rtx reg0 = REG_P (operands[0]) ? operands[0] : gen_reg_rtx (SImode);
rtx reg1 = operands[1];
if (! REG_P (reg1))
{
emit_move_insn (reg0, reg1);
reg1 = reg0;
}
emit_insn (gen_rtx_SET (SImode, reg0,
gen_rtx_AND (SImode, reg1, operands[2])));
/* Make sure we get the right *final* destination. */
if (! REG_P (operands[0]))
emit_move_insn (operands[0], reg0);
DONE;
}
})
;; Some special cases of andsi3.
(define_insn "*andsi_movu"
[(set (match_operand:SI 0 "register_operand" "=r,r,r")
(and:SI (match_operand:SI 1 "nonimmediate_operand" "%r,Q,To")
(match_operand:SI 2 "const_int_operand" "n,n,n")))]
"(INTVAL (operands[2]) == 255 || INTVAL (operands[2]) == 65535)
&& !side_effects_p (operands[1])"
"movu.%z2 %1,%0"
[(set_attr "slottable" "yes,yes,no")])
(define_insn "*andsi_clear"
[(set (match_operand:SI 0 "nonimmediate_operand" "=r,r,Q,Q,To,To")
(and:SI (match_operand:SI 1 "nonimmediate_operand" "%0,0,0,0,0,0")
(match_operand:SI 2 "const_int_operand" "P,n,P,n,P,n")))]
"(INTVAL (operands[2]) == -65536 || INTVAL (operands[2]) == -256)
&& !side_effects_p (operands[0])"
"@
cLear.b %0
cLear.w %0
cLear.b %0
cLear.w %0
cLear.b %0
cLear.w %0"
[(set_attr "slottable" "yes,yes,yes,yes,no,no")
(set_attr "cc" "none")])
;; This is a catch-all pattern, taking care of everything that was not
;; matched in the insns above.
;;
;; Sidenote: the tightening from "nonimmediate_operand" to
;; "register_operand" for operand 1 actually increased the register
;; pressure (worse code). That will hopefully change with an
;; improved reload pass.
(define_insn "*expanded_andsi_non_v32"
[(set (match_operand:SI 0 "register_operand" "=r,r,r, r,r")
(and:SI (match_operand:SI 1 "register_operand" "%0,0,0, 0,r")
(match_operand:SI 2 "general_operand" "I,r,Q>,g,!To")))]
"!TARGET_V32"
"@
andq %2,%0
and.d %2,%0
and.d %2,%0
and.d %2,%0
and.d %2,%1,%0"
[(set_attr "slottable" "yes,yes,yes,no,no")])
(define_insn "*expanded_andsi_v32"
[(set (match_operand:SI 0 "register_operand" "=r,r,r,r")
(and:SI (match_operand:SI 1 "register_operand" "%0,0,0,0")
(match_operand:SI 2 "general_operand" "I,r,Q>,g")))]
"TARGET_V32"
"@
andq %2,%0
and.d %2,%0
and.d %2,%0
and.d %2,%0"
[(set_attr "slottable" "yes,yes,yes,no")
(set_attr "cc" "noov32")])
;; For both QI and HI we may use the quick patterns. This results in
;; useless condition codes, but that is used rarely enough for it to
;; normally be a win (could check ahead for use of cc0, but seems to be
;; more pain than win).
;; FIXME: See note for andsi3
(define_expand "andhi3"
[(set (match_operand:HI 0 "nonimmediate_operand" "")
(and:HI (match_operand:HI 1 "nonimmediate_operand" "")
(match_operand:HI 2 "general_operand" "")))]
""
{
if (! (CONST_INT_P (operands[2])
&& (((INTVAL (operands[2]) == -256
|| INTVAL (operands[2]) == 65280)
&& rtx_equal_p (operands[1], operands[0]))
|| (INTVAL (operands[2]) == 255
&& REG_P (operands[0])))))
{
/* See comment for andsi3. */
rtx reg0 = REG_P (operands[0]) ? operands[0] : gen_reg_rtx (HImode);
rtx reg1 = operands[1];
if (! REG_P (reg1))
{
emit_move_insn (reg0, reg1);
reg1 = reg0;
}
emit_insn (gen_rtx_SET (HImode, reg0,
gen_rtx_AND (HImode, reg1, operands[2])));
/* Make sure we get the right destination. */
if (! REG_P (operands[0]))
emit_move_insn (operands[0], reg0);
DONE;
}
})
;; Some fast andhi3 special cases.
(define_insn "*andhi_movu"
[(set (match_operand:HI 0 "register_operand" "=r,r,r")
(and:HI (match_operand:HI 1 "nonimmediate_operand" "r,Q,To")
(const_int 255)))]
"!side_effects_p (operands[1])"
"mOvu.b %1,%0"
[(set_attr "slottable" "yes,yes,no")])
(define_insn "*andhi_clear"
[(set (match_operand:HI 0 "nonimmediate_operand" "=r,Q,To")
(and:HI (match_operand:HI 1 "nonimmediate_operand" "0,0,0")
(const_int -256)))]
"!side_effects_p (operands[0])"
"cLear.b %0"
[(set_attr "slottable" "yes,yes,no")
(set_attr "cc" "none")])
;; Catch-all andhi3 pattern.
(define_insn "*expanded_andhi_non_v32"
[(set (match_operand:HI 0 "register_operand" "=r,r,r, r,r,r,r")
(and:HI (match_operand:HI 1 "register_operand" "%0,0,0, 0,0,0,r")
(match_operand:HI 2 "general_operand" "I,r,Q>,L,O,g,!To")))]
;; Sidenote: the tightening from "general_operand" to
;; "register_operand" for operand 1 actually increased the register
;; pressure (worse code). That will hopefully change with an
;; improved reload pass.
"!TARGET_V32"
"@
andq %2,%0
and.w %2,%0
and.w %2,%0
and.w %2,%0
anDq %b2,%0
and.w %2,%0
and.w %2,%1,%0"
[(set_attr "slottable" "yes,yes,yes,no,yes,no,no")
(set_attr "cc" "clobber,normal,normal,normal,clobber,normal,normal")])
(define_insn "*expanded_andhi_v32"
[(set (match_operand:HI 0 "register_operand" "=r,r,r,r,r,r")
(and:HI (match_operand:HI 1 "register_operand" "%0,0,0,0,0,0")
(match_operand:HI 2 "general_operand" "I,r,Q>,L,O,g")))]
"TARGET_V32"
"@
andq %2,%0
and.w %2,%0
and.w %2,%0
and.w %2,%0
anDq %b2,%0
and.w %2,%0"
[(set_attr "slottable" "yes,yes,yes,no,yes,no")
(set_attr "cc" "clobber,noov32,noov32,noov32,clobber,noov32")])
;; A strict_low_part pattern.
;; Note the use of (match_dup 0) for the first operand of the operation
;; here. Reload can't handle an operand pair where one is read-write
;; and must match a read, like in:
;; (insn 80 79 81 4
;; (set (strict_low_part
;; (subreg:QI (reg/v:SI 0 r0 [orig:36 data ] [36]) 0))
;; (and:QI
;; (subreg:QI (reg:SI 15 acr [orig:27 D.7531 ] [27]) 0)
;; (const_int -64 [0xf..fc0]))) x.c:126 147 {*andqi_lowpart_v32}
;; (nil))
;; In theory, it could reload this as a movstrictqi of the register
;; operand at the and:QI to the destination register and change the
;; and:QI operand to the same as the read-write output operand and the
;; result would be recognized, but it doesn't recognize that's a valid
;; reload for a strict_low_part-destination; it just sees a "+" at the
;; destination constraints. Better than adding complexity to reload is
;; to follow the lead of m68k (see comment that begins with "These insns
;; must use MATCH_DUP") since prehistoric times and make it just a
;; match_dup. FIXME: a sanity-check in gen* to refuse an insn with
;; input-constraints matching input-output-constraints, e.g. "+r" <- "0".
(define_insn "*andhi_lowpart_non_v32"
[(set (strict_low_part
(match_operand:HI 0 "register_operand" "+r,r,r"))
(and:HI (match_dup 0)
(match_operand:HI 1 "general_operand" "r,Q>,g")))]
"!TARGET_V32"
"@
and.w %1,%0
and.w %1,%0
and.w %1,%0"
[(set_attr "slottable" "yes,yes,no")])
(define_insn "*andhi_lowpart_v32"
[(set (strict_low_part
(match_operand:HI 0 "register_operand" "+r,r,r"))
(and:HI (match_dup 0)
(match_operand:HI 1 "general_operand" "r,Q>,g")))]
"TARGET_V32"
"@
and.w %1,%0
and.w %1,%0
and.w %1,%0"
[(set_attr "slottable" "yes,yes,no")
(set_attr "cc" "noov32")])
(define_expand "andqi3"
[(set (match_operand:QI 0 "register_operand")
(and:QI (match_operand:QI 1 "register_operand")
(match_operand:QI 2 "general_operand")))]
""
"")
(define_insn "*andqi3_non_v32"
[(set (match_operand:QI 0 "register_operand" "=r,r,r, r,r,r")
(and:QI (match_operand:QI 1 "register_operand" "%0,0,0, 0,0,r")
(match_operand:QI 2 "general_operand" "I,r,Q>,O,g,!To")))]
"!TARGET_V32"
"@
andq %2,%0
and.b %2,%0
and.b %2,%0
andQ %b2,%0
and.b %2,%0
and.b %2,%1,%0"
[(set_attr "slottable" "yes,yes,yes,yes,no,no")
(set_attr "cc" "clobber,normal,normal,clobber,normal,normal")])
(define_insn "*andqi3_v32"
[(set (match_operand:QI 0 "register_operand" "=r,r,r,r,r")
(and:QI (match_operand:QI 1 "register_operand" "%0,0,0,0,0")
(match_operand:QI 2 "general_operand" "I,r,Q>,O,g")))]
"TARGET_V32"
"@
andq %2,%0
and.b %2,%0
and.b %2,%0
andQ %b2,%0
and.b %2,%0"
[(set_attr "slottable" "yes,yes,yes,yes,no")
(set_attr "cc" "clobber,noov32,noov32,clobber,noov32")])
(define_insn "*andqi_lowpart_non_v32"
[(set (strict_low_part
(match_operand:QI 0 "register_operand" "+r,r,r"))
(and:QI (match_dup 0)
(match_operand:QI 1 "general_operand" "r,Q>,g")))]
"!TARGET_V32"
"@
and.b %1,%0
and.b %1,%0
and.b %1,%0"
[(set_attr "slottable" "yes,yes,no")])
(define_insn "*andqi_lowpart_v32"
[(set (strict_low_part
(match_operand:QI 0 "register_operand" "+r,r,r"))
(and:QI (match_dup 0)
(match_operand:QI 1 "general_operand" "r,Q>,g")))]
"TARGET_V32"
"@
and.b %1,%0
and.b %1,%0
and.b %1,%0"
[(set_attr "slottable" "yes,yes,no")
(set_attr "cc" "noov32")])
;; Bitwise or.
;; Same comment as anddi3 applies here - no need for such a pattern.
;; It seems there's no need to jump through hoops to get good code such as
;; with andsi3.
(define_expand "ior3"
[(set (match_operand:BWD 0 "register_operand")
(ior:BWD (match_operand:BWD 1 "register_operand")
(match_operand:BWD 2 "general_operand")))]
""
"")
(define_insn "*iorsi3_non_v32"
[(set (match_operand:SI 0 "register_operand" "=r,r,r, r,r,r")
(ior:SI (match_operand:SI 1 "register_operand" "%0,0,0, 0,0,r")
(match_operand:SI 2 "general_operand" "I, r,Q>,n,g,!To")))]
"!TARGET_V32"
"@
orq %2,%0
or.d %2,%0
or.d %2,%0
oR.%s2 %2,%0
or.d %2,%0
or.d %2,%1,%0"
[(set_attr "slottable" "yes,yes,yes,no,no,no")
(set_attr "cc" "normal,normal,normal,clobber,normal,normal")])
(define_insn "*iorsi3_v32"
[(set (match_operand:SI 0 "register_operand" "=r,r,r,r,r")
(ior:SI (match_operand:SI 1 "register_operand" "%0,0,0,0,0")
(match_operand:SI 2 "general_operand" "I,r,Q>,n,g")))]
"TARGET_V32"
"@
orq %2,%0
or.d %2,%0
or.d %2,%0
oR.%s2 %2,%0
or.d %2,%0"
[(set_attr "slottable" "yes,yes,yes,no,no")
(set_attr "cc" "noov32,noov32,noov32,clobber,noov32")])
(define_insn "*iorhi3_non_v32"
[(set (match_operand:HI 0 "register_operand" "=r,r,r, r,r,r,r")
(ior:HI (match_operand:HI 1 "register_operand" "%0,0,0, 0,0,0,r")
(match_operand:HI 2 "general_operand" "I,r,Q>,L,O,g,!To")))]
"!TARGET_V32"
"@
orq %2,%0
or.w %2,%0
or.w %2,%0
or.w %2,%0
oRq %b2,%0
or.w %2,%0
or.w %2,%1,%0"
[(set_attr "slottable" "yes,yes,yes,no,yes,no,no")
(set_attr "cc" "clobber,normal,normal,normal,clobber,normal,normal")])
(define_insn "*iorhi3_v32"
[(set (match_operand:HI 0 "register_operand" "=r,r,r,r,r,r")
(ior:HI (match_operand:HI 1 "register_operand" "%0,0,0,0,0,0")
(match_operand:HI 2 "general_operand" "I,r,Q>,L,O,g")))]
"TARGET_V32"
"@
orq %2,%0
or.w %2,%0
or.w %2,%0
or.w %2,%0
oRq %b2,%0
or.w %2,%0"
[(set_attr "slottable" "yes,yes,yes,no,yes,no")
(set_attr "cc" "clobber,noov32,noov32,noov32,clobber,noov32")])
(define_insn "*iorqi3_non_v32"
[(set (match_operand:QI 0 "register_operand" "=r,r,r, r,r,r")
(ior:QI (match_operand:QI 1 "register_operand" "%0,0,0, 0,0,r")
(match_operand:QI 2 "general_operand" "I,r,Q>,O,g,!To")))]
"!TARGET_V32"
"@
orq %2,%0
or.b %2,%0
or.b %2,%0
orQ %b2,%0
or.b %2,%0
or.b %2,%1,%0"
[(set_attr "slottable" "yes,yes,yes,yes,no,no")
(set_attr "cc" "clobber,normal,normal,clobber,normal,normal")])
(define_insn "*iorqi3_v32"
[(set (match_operand:QI 0 "register_operand" "=r,r,r,r,r")
(ior:QI (match_operand:QI 1 "register_operand" "%0,0,0,0,0")
(match_operand:QI 2 "general_operand" "I,r,Q>,O,g")))]
"TARGET_V32"
"@
orq %2,%0
or.b %2,%0
or.b %2,%0
orQ %b2,%0
or.b %2,%0"
[(set_attr "slottable" "yes,yes,yes,yes,no")
(set_attr "cc" "clobber,noov32,noov32,clobber,noov32")])
;; Exclusive-or
;; See comment about "anddi3" for xordi3 - no need for such a pattern.
;; FIXME: Do we really need the shorter variants?
(define_insn "xorsi3"
[(set (match_operand:SI 0 "register_operand" "=r")
(xor:SI (match_operand:SI 1 "register_operand" "%0")
(match_operand:SI 2 "register_operand" "r")))]
""
"xor %2,%0"
[(set_attr "slottable" "yes")
(set_attr "cc" "noov32")])
(define_insn "xor3"
[(set (match_operand:BW 0 "register_operand" "=r")
(xor:BW (match_operand:BW 1 "register_operand" "%0")
(match_operand:BW 2 "register_operand" "r")))]
""
"xor %2,%0"
[(set_attr "slottable" "yes")
(set_attr "cc" "clobber")])
;; Negation insns.
;; Questionable use, here mostly as a (slightly usable) define_expand
;; example.
(define_expand "negsf2"
[(set (match_dup 2)
(match_dup 3))
(parallel [(set (match_operand:SF 0 "register_operand" "=r")
(neg:SF (match_operand:SF 1
"register_operand" "0")))
(use (match_dup 2))])]
""
{
operands[2] = gen_reg_rtx (SImode);
operands[3] = GEN_INT (1 << 31);
})
(define_insn "*expanded_negsf2"
[(set (match_operand:SF 0 "register_operand" "=r")
(neg:SF (match_operand:SF 1 "register_operand" "0")))
(use (match_operand:SI 2 "register_operand" "r"))]
""
"xor %2,%0"
[(set_attr "slottable" "yes")])
;; No "negdi2" although we could make one up that may be faster than
;; the one in libgcc.
(define_insn "neg2"
[(set (match_operand:BWD 0 "register_operand" "=r")
(neg:BWD (match_operand:BWD 1 "register_operand" "r")))]
""
"neg %1,%0"
[(set_attr "slottable" "yes")])
;; One-complements.
;; See comment on anddi3 - no need for a DImode pattern.
;; See also xor comment.
(define_insn "one_cmplsi2"
[(set (match_operand:SI 0 "register_operand" "=r")
(not:SI (match_operand:SI 1 "register_operand" "0")))]
""
"not %0"
[(set_attr "slottable" "yes")
(set_attr "cc" "noov32")])
(define_insn "one_cmpl2"
[(set (match_operand:BW 0 "register_operand" "=r")
(not:BW (match_operand:BW 1 "register_operand" "0")))]
""
"not %0"
[(set_attr "slottable" "yes")
(set_attr "cc" "clobber")])
;; Arithmetic/Logical shift right (and SI left).
(define_insn "si3"
[(set (match_operand:SI 0 "register_operand" "=r")
(shift:SI (match_operand:SI 1 "register_operand" "0")
(match_operand:SI 2 "nonmemory_operand" "Kcr")))]
""
{
if (REG_S_P (operands[2]))
return ".d %2,%0";
return "q %2,%0";
}
[(set_attr "slottable" "yes")
(set_attr "cc" "noov32")])
;; Since gcc gets lost, and forgets to zero-extend the source (or mask
;; the destination) when it changes shifts of lower modes into SImode,
;; it is better to make these expands an anonymous patterns instead of
;; the more correct define_insns. This occurs when gcc thinks that is
;; is better to widen to SImode and use immediate shift count.
;; FIXME: Is this legacy or still true for gcc >= 2.7.2?
;; FIXME: Can't parametrize sign_extend and zero_extend (before
;; mentioning "shiftrt"), so we need two patterns.
(define_expand "ashr3"
[(set (match_dup 3)
(sign_extend:SI (match_operand:BW 1 "nonimmediate_operand" "")))
(set (match_dup 4)
(zero_extend:SI (match_operand:BW 2 "nonimmediate_operand" "")))
(set (match_dup 5) (ashiftrt:SI (match_dup 3) (match_dup 4)))
(set (match_operand:BW 0 "general_operand" "")
(subreg:BW (match_dup 5) 0))]
""
{
int i;
for (i = 3; i < 6; i++)
operands[i] = gen_reg_rtx (SImode);
})
(define_expand "lshr3"
[(set (match_dup 3)
(zero_extend:SI (match_operand:BW 1 "nonimmediate_operand" "")))
(set (match_dup 4)
(zero_extend:SI (match_operand:BW 2 "nonimmediate_operand" "")))
(set (match_dup 5) (lshiftrt:SI (match_dup 3) (match_dup 4)))
(set (match_operand:BW 0 "general_operand" "")
(subreg:BW (match_dup 5) 0))]
""
{
int i;
for (i = 3; i < 6; i++)
operands[i] = gen_reg_rtx (SImode);
})
(define_insn "*expanded_"
[(set (match_operand:BW 0 "register_operand" "=r")
(shiftrt:BW (match_operand:BW 1 "register_operand" "0")
(match_operand:BW 2 "register_operand" "r")))]
""
" %2,%0"
[(set_attr "slottable" "yes")
(set_attr "cc" "noov32")])
(define_insn "*_lowpart"
[(set (strict_low_part (match_operand:BW 0 "register_operand" "+r"))
(shiftrt:BW (match_dup 0)
(match_operand:BW 1 "register_operand" "r")))]
""
" %1,%0"
[(set_attr "slottable" "yes")
(set_attr "cc" "noov32")])
;; Arithmetic/logical shift left.
;; For narrower modes than SI, we can use lslq although it makes cc
;; unusable. The win is that we do not have to reload the shift-count
;; into a register.
(define_insn "ashl3"
[(set (match_operand:BW 0 "register_operand" "=r,r")
(ashift:BW (match_operand:BW 1 "register_operand" "0,0")
(match_operand:BW 2 "nonmemory_operand" "r,Kc")))]
""
{
return
(CONST_INT_P (operands[2]) && INTVAL (operands[2]) > )
? "moveq 0,%0"
: (CONSTANT_P (operands[2])
? "lslq %2,%0" : "lsl %2,%0");
}
[(set_attr "slottable" "yes")
(set_attr "cc" "noov32,clobber")])
;; A strict_low_part matcher.
(define_insn "*ashl_lowpart"
[(set (strict_low_part (match_operand:BW 0 "register_operand" "+r"))
(ashift:BW (match_dup 0)
(match_operand:HI 1 "register_operand" "r")))]
""
"lsl %1,%0"
[(set_attr "slottable" "yes")
(set_attr "cc" "noov32")])
;; Various strange insns that gcc likes.
;; Fortunately, it is simple to construct an abssf (although it may not
;; be very much used in practice).
(define_insn "abssf2"
[(set (match_operand:SF 0 "register_operand" "=r")
(abs:SF (match_operand:SF 1 "register_operand" "0")))]
""
"lslq 1,%0\;lsrq 1,%0")
(define_insn "abssi2"
[(set (match_operand:SI 0 "register_operand" "=r")
(abs:SI (match_operand:SI 1 "register_operand" "r")))]
""
"abs %1,%0"
[(set_attr "slottable" "yes")
(set_attr "cc" "noov32")])
;; FIXME: GCC should be able to do these expansions itself.
(define_expand "abs2"
[(set (match_dup 2)
(sign_extend:SI (match_operand:BW 1 "general_operand" "")))
(set (match_dup 3) (abs:SI (match_dup 2)))
(set (match_operand:BW 0 "register_operand" "")
(subreg:BW (match_dup 3) 0))]
""
"operands[2] = gen_reg_rtx (SImode); operands[3] = gen_reg_rtx (SImode);")
(define_insn "clzsi2"
[(set (match_operand:SI 0 "register_operand" "=r")
(clz:SI (match_operand:SI 1 "register_operand" "r")))]
"TARGET_HAS_LZ"
"lz %1,%0"
[(set_attr "slottable" "yes")
(set_attr "cc" "noov32")])
(define_insn "bswapsi2"
[(set (match_operand:SI 0 "register_operand" "=r")
(bswap:SI (match_operand:SI 1 "register_operand" "0")))]
"TARGET_HAS_SWAP"
"swapwb %0"
[(set_attr "slottable" "yes")
(set_attr "cc" "noov32")])
;; This instruction swaps all bits in a register.
;; That means that the most significant bit is put in the place
;; of the least significant bit, and so on.
(define_insn "cris_swap_bits"
[(set (match_operand:SI 0 "register_operand" "=r")
(unspec:SI [(match_operand:SI 1 "register_operand" "0")]
CRIS_UNSPEC_SWAP_BITS))]
"TARGET_HAS_SWAP"
"swapwbr %0"
[(set_attr "slottable" "yes")
(set_attr "cc" "noov32")])
;; Implement ctz using two instructions, one for bit swap and one for clz.
;; Defines a scratch register to avoid clobbering input.
(define_expand "ctzsi2"
[(set (match_dup 2)
(match_operand:SI 1 "register_operand"))
(set (match_dup 2)
(unspec:SI [(match_dup 2)] CRIS_UNSPEC_SWAP_BITS))
(set (match_operand:SI 0 "register_operand")
(clz:SI (match_dup 2)))]
"TARGET_HAS_LZ && TARGET_HAS_SWAP"
"operands[2] = gen_reg_rtx (SImode);")
;; Bound-insn. Defined to be the same as an unsigned minimum, which is an
;; operation supported by gcc. Used in casesi, but used now and then in
;; normal code too.
(define_expand "uminsi3"
[(set (match_operand:SI 0 "register_operand" "")
(umin:SI (match_operand:SI 1 "register_operand" "")
(match_operand:SI 2 "general_operand" "")))]
""
{
if (MEM_P (operands[2]) && TARGET_V32)
operands[2] = force_reg (SImode, operands[2]);
})
(define_insn "*uminsi3_non_v32"
[(set (match_operand:SI 0 "register_operand" "=r,r, r,r")
(umin:SI (match_operand:SI 1 "register_operand" "%0,0, 0,r")
(match_operand:SI 2 "general_operand" "r,Q>,g,!To")))]
"!TARGET_V32"
{
if (CONST_INT_P (operands[2]))
{
/* Constant operands are zero-extended, so only 32-bit operands
may be negative. */
if (INTVAL (operands[2]) >= 0)
{
if (INTVAL (operands[2]) < 256)
return "bound.b %2,%0";
if (INTVAL (operands[2]) < 65536)
return "bound.w %2,%0";
}
}
else if (which_alternative == 3)
return "bound.d %2,%1,%0";
return "bound.d %2,%0";
}
[(set_attr "slottable" "yes,yes,no,no")])
(define_insn "*uminsi3_v32"
[(set (match_operand:SI 0 "register_operand" "=r,r")
(umin:SI (match_operand:SI 1 "register_operand" "%0,0")
(match_operand:SI 2 "nonmemory_operand" "r,i")))]
"TARGET_V32"
{
if (GET_CODE (operands[2]) == CONST_INT)
{
/* Constant operands are zero-extended, so only 32-bit operands
may be negative. */
if (INTVAL (operands[2]) >= 0)
{
if (INTVAL (operands[2]) < 256)
return "bound.b %2,%0";
if (INTVAL (operands[2]) < 65536)
return "bound.w %2,%0";
}
}
return "bound.d %2,%0";
}
[(set_attr "slottable" "yes,no")])
;; Jump and branch insns.
(define_insn "jump"
[(set (pc)
(label_ref (match_operand 0 "" "")))]
""
"ba %l0%#"
[(set_attr "slottable" "has_slot")])
;; Testcase gcc.c-torture/compile/991213-3.c fails if we allow a constant
;; here, since the insn is not recognized as an indirect jump by
;; jmp_uses_reg_or_mem used by computed_jump_p. Perhaps it is a kludge to
;; change from general_operand to nonimmediate_operand (at least the docs
;; should be changed), but then again the pattern is called indirect_jump.
(define_expand "indirect_jump"
[(set (pc) (match_operand:SI 0 "nonimmediate_operand"))]
""
{
if (TARGET_V32 && MEM_P (operands[0]))
operands[0] = force_reg (SImode, operands[0]);
})
(define_insn "*indirect_jump_non_v32"
[(set (pc) (match_operand:SI 0 "nonimmediate_operand" "rm"))]
"!TARGET_V32"
"jump %0")
(define_insn "*indirect_jump_v32"
[(set (pc) (match_operand:SI 0 "register_operand" "r"))]
"TARGET_V32"
"jump %0%#"
[(set_attr "slottable" "has_slot")])
;; Return insn. Used whenever the epilogue is very simple; if it is only
;; a single ret or jump [sp+]. No allocated stack space or saved
;; registers are allowed.
;; Note that for this pattern, although named, it is ok to check the
;; context of the insn in the test, not only compiler switches.
(define_expand "return"
[(return)]
"cris_simple_epilogue ()"
"cris_expand_return (cris_return_address_on_stack ()); DONE;")
(define_insn "*return_expanded"
[(return)]
""
{
return cris_return_address_on_stack_for_return ()
? "jump [$sp+]" : "ret%#";
}
[(set (attr "slottable")
(if_then_else
(match_test "cris_return_address_on_stack_for_return ()")
(const_string "no")
(const_string "has_return_slot")))])
(define_expand "prologue"
[(const_int 0)]
"TARGET_PROLOGUE_EPILOGUE"
"cris_expand_prologue (); DONE;")
;; Note that the (return) from the expander itself is always the last
;; insn in the epilogue.
(define_expand "epilogue"
[(const_int 0)]
"TARGET_PROLOGUE_EPILOGUE"
"cris_expand_epilogue (); DONE;")
;; Conditional branches.
(define_expand "cbranch4"
[(set (cc0) (compare
(match_operand:BWD 1 "nonimmediate_operand")
(match_operand:BWD 2 "general_operand")))
(set (pc)
(if_then_else (match_operator 0 "ordered_comparison_operator"
[(cc0) (const_int 0)])
(label_ref (match_operand 3 "" ""))
(pc)))]
""
"")
(define_expand "cbranchdi4"
[(set (cc0)
(compare (match_operand:DI 1 "nonimmediate_operand" "")
(match_operand:DI 2 "general_operand" "")))
(set (pc)
(if_then_else (match_operator 0 "ordered_comparison_operator"
[(cc0) (const_int 0)])
(label_ref (match_operand 3 "" ""))
(pc)))]
""
{
if (TARGET_V32 && !REG_P (operands[1]))
operands[1] = force_reg (DImode, operands[1]);
if (TARGET_V32 && MEM_P (operands[2]))
operands[2] = force_reg (DImode, operands[2]);
})
;; We suffer from the same overflow-bit-gets-in-the-way problem as
;; e.g. m68k, so we have to check if overflow bit is set on all "signed"
;; conditions.
(define_insn "b"
[(set (pc)
(if_then_else (ncond (cc0)
(const_int 0))
(label_ref (match_operand 0 "" ""))
(pc)))]
""
"b %l0%#"
[(set_attr "slottable" "has_slot")])
(define_insn "b"
[(set (pc)
(if_then_else (ocond (cc0)
(const_int 0))
(label_ref (match_operand 0 "" ""))
(pc)))]
""
{
return
(cc_prev_status.flags & CC_NO_OVERFLOW)
? 0 : "b %l0%#";
}
[(set_attr "slottable" "has_slot")])
(define_insn "b"
[(set (pc)
(if_then_else (rcond (cc0)
(const_int 0))
(label_ref (match_operand 0 "" ""))
(pc)))]
""
{
return
(cc_prev_status.flags & CC_NO_OVERFLOW)
? "b %l0%#" : "b %l0%#";
}
[(set_attr "slottable" "has_slot")])
;; Reversed anonymous patterns to the ones above, as mandated.
(define_insn "*b_reversed"
[(set (pc)
(if_then_else (ncond (cc0)
(const_int 0))
(pc)
(label_ref (match_operand 0 "" ""))))]
""
"b %l0%#"
[(set_attr "slottable" "has_slot")])
(define_insn "*b_reversed"
[(set (pc)
(if_then_else (ocond (cc0)
(const_int 0))
(pc)
(label_ref (match_operand 0 "" ""))))]
""
{
return
(cc_prev_status.flags & CC_NO_OVERFLOW)
? 0 : "b %l0%#";
}
[(set_attr "slottable" "has_slot")])
(define_insn "*b_reversed"
[(set (pc)
(if_then_else (rcond (cc0)
(const_int 0))
(pc)
(label_ref (match_operand 0 "" ""))))]
""
{
return
(cc_prev_status.flags & CC_NO_OVERFLOW)
? "b %l0%#" : "b %l0%#";
}
[(set_attr "slottable" "has_slot")])
;; Set on condition: sCC.
(define_expand "cstoredi4"
[(set (cc0) (compare
(match_operand:DI 2 "nonimmediate_operand")
(match_operand:DI 3 "general_operand")))
(set (match_operand:SI 0 "register_operand")
(match_operator:SI 1 "ordered_comparison_operator"
[(cc0) (const_int 0)]))]
""
{
if (TARGET_V32 && !REG_P (operands[2]))
operands[2] = force_reg (DImode, operands[2]);
if (TARGET_V32 && MEM_P (operands[3]))
operands[3] = force_reg (DImode, operands[3]);
})
(define_expand "cstore4"
[(set (cc0) (compare
(match_operand:BWD 2 "nonimmediate_operand")
(match_operand:BWD 3 "general_operand")))
(set (match_operand:SI 0 "register_operand")
(match_operator:SI 1 "ordered_comparison_operator"
[(cc0) (const_int 0)]))]
""
"")
;; Like bCC, we have to check the overflow bit for
;; signed conditions.
(define_insn "s"
[(set (match_operand:SI 0 "register_operand" "=r")
(ncond:SI (cc0) (const_int 0)))]
""
"s %0"
[(set_attr "slottable" "yes")
(set_attr "cc" "none")])
(define_insn "s"
[(set (match_operand:SI 0 "register_operand" "=r")
(rcond:SI (cc0) (const_int 0)))]
""
{
return
(cc_prev_status.flags & CC_NO_OVERFLOW)
? "s %0" : "s %0";
}
[(set_attr "slottable" "yes")
(set_attr "cc" "none")])
(define_insn "s"
[(set (match_operand:SI 0 "register_operand" "=r")
(ocond:SI (cc0) (const_int 0)))]
""
{
return
(cc_prev_status.flags & CC_NO_OVERFLOW)
? 0 : "s %0";
}
[(set_attr "slottable" "yes")
(set_attr "cc" "none")])
;; Call insns.
;; We need to make these patterns "expand", since the real operand is
;; hidden in a (mem:QI ) inside operand[0] (call_value: operand[1]),
;; and cannot be checked if it were a "normal" pattern.
;; Note that "call" and "call_value" are *always* called with a
;; mem-operand for operand 0 and 1 respective. What happens for combined
;; instructions is a different issue.
(define_expand "call"
[(parallel [(call (match_operand:QI 0 "cris_mem_call_operand" "")
(match_operand 1 "general_operand" ""))
(clobber (reg:SI CRIS_SRP_REGNUM))])]
""
{
gcc_assert (MEM_P (operands[0]));
if (flag_pic)
cris_expand_pic_call_address (&operands[0]);
})
;; Accept *anything* as operand 1. Accept operands for operand 0 in
;; order of preference.
(define_insn "*expanded_call_non_v32"
[(call (mem:QI (match_operand:SI 0 "general_operand" "r,Q>,g"))
(match_operand 1 "" ""))
(clobber (reg:SI CRIS_SRP_REGNUM))]
"!TARGET_V32"
"jsr %0")
(define_insn "*expanded_call_v32"
[(call
(mem:QI
(match_operand:SI 0 "cris_nonmemory_operand_or_callable_symbol" "n,r,U,i"))
(match_operand 1 "" ""))
(clobber (reg:SI CRIS_SRP_REGNUM))]
"TARGET_V32"
"@
jsr %0%#
jsr %0%#
bsr %0%#
bsr %0%#"
[(set_attr "slottable" "has_call_slot")])
;; Parallel when calculating and reusing address of indirect pointer
;; with simple offset. (Makes most sense with PIC.) It looks a bit
;; wrong not to have the clobber last, but that's the way combine
;; generates it (except it doesn' look into the *inner* mem, so this
;; just matches a peephole2). FIXME: investigate that.
(define_insn "*expanded_call_side"
[(call (mem:QI
(mem:SI
(plus:SI (match_operand:SI 0 "cris_bdap_operand" "%r, r,r")
(match_operand:SI 1 "cris_bdap_operand" "r>Rn,r,>Rn"))))
(match_operand 2 "" ""))
(clobber (reg:SI CRIS_SRP_REGNUM))
(set (match_operand:SI 3 "register_operand" "=*0,r,r")
(plus:SI (match_dup 0)
(match_dup 1)))]
"!TARGET_AVOID_GOTPLT && !TARGET_V32"
"jsr [%3=%0%S1]")
(define_expand "call_value"
[(parallel [(set (match_operand 0 "" "")
(call (match_operand:QI 1 "cris_mem_call_operand" "")
(match_operand 2 "" "")))
(clobber (reg:SI CRIS_SRP_REGNUM))])]
""
{
gcc_assert (MEM_P (operands[1]));
if (flag_pic)
cris_expand_pic_call_address (&operands[1]);
})
;; Accept *anything* as operand 2. The validity other than "general" of
;; operand 0 will be checked elsewhere. Accept operands for operand 1 in
;; order of preference (Q includes r, but r is shorter, faster).
;; We also accept a PLT symbol. We output it as [rPIC+sym:GOTPLT] rather
;; than requiring getting rPIC + sym:PLT into a register.
(define_insn "*expanded_call_value_non_v32"
[(set (match_operand 0 "nonimmediate_operand" "=g,g,g")
(call (mem:QI (match_operand:SI 1 "general_operand" "r,Q>,g"))
(match_operand 2 "" "")))
(clobber (reg:SI CRIS_SRP_REGNUM))]
"!TARGET_V32"
"Jsr %1"
[(set_attr "cc" "clobber")])
;; See similar call special-case.
(define_insn "*expanded_call_value_side"
[(set (match_operand 0 "nonimmediate_operand" "=g,g,g")
(call
(mem:QI
(mem:SI
(plus:SI (match_operand:SI 1 "cris_bdap_operand" "%r, r,r")
(match_operand:SI 2 "cris_bdap_operand" "r>Rn,r,>Rn"))))
(match_operand 3 "" "")))
(clobber (reg:SI CRIS_SRP_REGNUM))
(set (match_operand:SI 4 "register_operand" "=*1,r,r")
(plus:SI (match_dup 1)
(match_dup 2)))]
"!TARGET_AVOID_GOTPLT && !TARGET_V32"
"Jsr [%4=%1%S2]"
[(set_attr "cc" "clobber")])
(define_insn "*expanded_call_value_v32"
[(set
(match_operand 0 "nonimmediate_operand" "=g,g,g,g")
(call
(mem:QI
(match_operand:SI 1 "cris_nonmemory_operand_or_callable_symbol" "n,r,U,i"))
(match_operand 2 "" "")))
(clobber (reg:SI 16))]
"TARGET_V32"
"@
Jsr %1%#
Jsr %1%#
Bsr %1%#
Bsr %1%#"
[(set_attr "cc" "clobber")
(set_attr "slottable" "has_call_slot")])
;; Used in debugging. No use for the direct pattern; unfilled
;; delayed-branches are taken care of by other means.
(define_insn "nop"
[(const_int 0)]
""
"nop"
[(set_attr "cc" "none")])
;; Same as the gdb trap breakpoint, will cause a SIGTRAP for
;; cris-linux* and crisv32-linux*, as intended. Will work in
;; freestanding environments with sufficient framework.
(define_insn "trap"
[(trap_if (const_int 1) (const_int 8))]
"TARGET_TRAP_USING_BREAK8"
"break 8")
;; We need to stop accesses to the stack after the memory is
;; deallocated. Unfortunately, reorg doesn't look at naked clobbers,
;; e.g. (insn ... (clobber (mem:BLK (stack_pointer_rtx)))) and we don't
;; want to use a naked (unspec_volatile) as that would stop any
;; scheduling in the epilogue. Hence we model it as a "real" insn that
;; sets the memory in an unspecified manner. FIXME: Unfortunately it
;; still has the effect of an unspec_volatile.
(define_insn "cris_frame_deallocated_barrier"
[(set (mem:BLK (reg:SI CRIS_SP_REGNUM))
(unspec:BLK [(const_int 0)] CRIS_UNSPEC_FRAME_DEALLOC))]
""
""
[(set_attr "length" "0")])
;; We expand on casesi so we can use "bound" and "add offset fetched from
;; a table to pc" (adds.w [pc+%0.w],pc).
;; Note: if you change the "parallel" (or add anything after it) in
;; this expansion, you must change the macro ASM_OUTPUT_CASE_END
;; accordingly, to add the default case at the end of the jump-table.
(define_expand "cris_casesi_non_v32"
[(set (match_dup 5) (match_operand:SI 0 "general_operand" ""))
(set (match_dup 6)
(minus:SI (match_dup 5)
(match_operand:SI 1 "const_int_operand" "n")))
(set (match_dup 7)
(umin:SI (match_dup 6)
(match_operand:SI 2 "const_int_operand" "n")))
(parallel
[(set (pc)
(if_then_else
(ltu (match_dup 7) (match_dup 2))
(plus:SI (sign_extend:SI
(mem:HI
(plus:SI (mult:SI (match_dup 7) (const_int 2))
(pc))))
(pc))
(label_ref (match_operand 4 "" ""))))
(use (label_ref (match_operand 3 "" "")))])]
""
{
operands[2] = plus_constant (SImode, operands[2], 1);
operands[5] = gen_reg_rtx (SImode);
operands[6] = gen_reg_rtx (SImode);
operands[7] = gen_reg_rtx (SImode);
})
;; FIXME: Check effect of not JUMP_TABLES_IN_TEXT_SECTION.
(define_expand "cris_casesi_v32"
[(set (match_dup 5) (match_operand:SI 0 "general_operand"))
(set (match_dup 6)
(minus:SI (match_dup 5)
(match_operand:SI 1 "const_int_operand")))
(set (match_dup 7)
(umin:SI (match_dup 6)
(match_operand:SI 2 "const_int_operand")))
(set (match_dup 8) (match_dup 11))
(set (match_dup 9)
(plus:SI (mult:SI (match_dup 7) (const_int 2))
(match_dup 8)))
(set (match_dup 10)
(plus:SI (sign_extend:SI (mem:HI (match_dup 9)))
(match_dup 9)))
(parallel
[(set (pc)
(if_then_else
(ltu (unspec [(const_int 0)] CRIS_UNSPEC_CASESI) (match_dup 2))
(match_dup 10)
(label_ref (match_operand 4 "" ""))))
(use (label_ref (match_dup 3)))])]
"TARGET_V32"
{
int i;
rtx xlabel = gen_rtx_LABEL_REF (VOIDmode, operands[3]);
for (i = 5; i <= 10; i++)
operands[i] = gen_reg_rtx (SImode);
operands[2] = plus_constant (SImode, operands[2], 1);
/* Don't forget to decorate labels too, for PIC. */
operands[11] = flag_pic
? gen_rtx_CONST (Pmode,
gen_rtx_UNSPEC (Pmode, gen_rtvec (1, xlabel),
CRIS_UNSPEC_PCREL))
: xlabel;
})
(define_expand "casesi"
[(match_operand:SI 0 "general_operand")
(match_operand:SI 1 "const_int_operand")
(match_operand:SI 2 "const_int_operand")
(match_operand 3 "" "")
(match_operand 4 "" "")]
""
{
if (TARGET_V32)
emit_insn (gen_cris_casesi_v32 (operands[0], operands[1], operands[2],
operands[3], operands[4]));
else
emit_insn (gen_cris_casesi_non_v32 (operands[0], operands[1], operands[2],
operands[3], operands[4]));
DONE;
})
;; Split-patterns. Some of them have modes unspecified. This
;; should always be ok; if for no other reason sparc.md has it as
;; well.
;;
;; When register_operand is specified for an operand, we can get a
;; subreg as well (Axis-990331), so don't just assume that REG_P is true
;; for a register_operand and that REGNO can be used as is. It is best to
;; guard with REG_P, unless it is worth it to adjust for the subreg case.
;; op [rx + 0],ry,rz
;; The index to rx is optimized into zero, and gone.
;; First, recognize bound [rx],ry,rz; where [rx] is zero-extended,
;; and add/sub [rx],ry,rz, with zero or sign-extend on [rx].
;; Split this into:
;; move ry,rz
;; op [rx],rz
;; Lose if rz=ry or rx=rz.
;; Call this op-extend-split.
;; Do not match for V32; the addo and addi shouldn't be split
;; up.
(define_split
[(set (match_operand 0 "cris_nonsp_register_operand" "")
(match_operator
4 "cris_operand_extend_operator"
[(match_operand 1 "register_operand" "")
(match_operator
3 "cris_extend_operator"
[(match_operand 2 "memory_operand" "")])]))]
"!TARGET_V32
&& REG_P (operands[0])
&& REG_P (operands[1])
&& REGNO (operands[1]) != REGNO (operands[0])
&& GET_MODE_SIZE (GET_MODE (operands[0])) <= UNITS_PER_WORD
&& REG_P (XEXP (operands[2], 0))
&& REGNO (XEXP (operands[2], 0)) != REGNO (operands[0])"
[(set (match_dup 0)
(match_dup 1))
(set (match_dup 0)
(match_op_dup
4 [(match_dup 0)
(match_op_dup 3 [(match_dup 2)])]))]
"")
;; As op-extend-split, but recognize and split op [rz],ry,rz into
;; ext [rz],rz
;; op ry,rz
;; Do this for plus or bound only, being commutative operations, since we
;; have swapped the operands.
;; Call this op-extend-split-rx=rz
(define_split
[(set (match_operand 0 "cris_nonsp_register_operand" "")
(match_operator
4 "cris_plus_or_bound_operator"
[(match_operand 1 "register_operand" "")
(match_operator
3 "cris_extend_operator"
[(match_operand 2 "memory_operand" "")])]))]
"!TARGET_V32
&& REG_P (operands[0])
&& REG_P (operands[1])
&& REGNO (operands[1]) != REGNO (operands[0])
&& GET_MODE_SIZE (GET_MODE (operands[0])) <= UNITS_PER_WORD
&& REG_P (XEXP (operands[2], 0))
&& REGNO (XEXP (operands[2], 0)) == REGNO (operands[0])"
[(set (match_dup 0)
(match_op_dup 3 [(match_dup 2)]))
(set (match_dup 0)
(match_op_dup
4 [(match_dup 0)
(match_dup 1)]))]
"")
;; As the op-extend-split, but swapped operands, and only for
;; plus or bound, being the commutative extend-operators. FIXME: Why is
;; this needed? Is it?
;; Call this op-extend-split-swapped
(define_split
[(set (match_operand 0 "cris_nonsp_register_operand" "")
(match_operator
4 "cris_plus_or_bound_operator"
[(match_operator
3 "cris_extend_operator"
[(match_operand 2 "memory_operand" "")])
(match_operand 1 "register_operand" "")]))]
"!TARGET_V32
&& REG_P (operands[0])
&& REG_P (operands[1])
&& REGNO (operands[1]) != REGNO (operands[0])
&& GET_MODE_SIZE (GET_MODE (operands[0])) <= UNITS_PER_WORD
&& REG_P (XEXP (operands[2], 0))
&& REGNO (XEXP (operands[2], 0)) != REGNO (operands[0])"
[(set (match_dup 0)
(match_dup 1))
(set (match_dup 0)
(match_op_dup
4 [(match_dup 0)
(match_op_dup 3 [(match_dup 2)])]))]
"")
;; As op-extend-split-rx=rz, but swapped operands, only for plus or
;; bound. Call this op-extend-split-swapped-rx=rz.
(define_split
[(set (match_operand 0 "cris_nonsp_register_operand" "")
(match_operator
4 "cris_plus_or_bound_operator"
[(match_operator
3 "cris_extend_operator"
[(match_operand 2 "memory_operand" "")])
(match_operand 1 "register_operand" "")]))]
"!TARGET_V32
&& REG_P (operands[0])
&& REG_P (operands[1])
&& REGNO (operands[1]) != REGNO (operands[0])
&& GET_MODE_SIZE (GET_MODE (operands[0])) <= UNITS_PER_WORD
&& REG_P (XEXP (operands[2], 0))
&& REGNO (XEXP (operands[2], 0)) == REGNO (operands[0])"
[(set (match_dup 0)
(match_op_dup 3 [(match_dup 2)]))
(set (match_dup 0)
(match_op_dup
4 [(match_dup 0)
(match_dup 1)]))]
"")
;; As op-extend-split, but the mem operand is not extended.
;;
;; op [rx],ry,rz changed into
;; move ry,rz
;; op [rx],rz
;; lose if ry=rz or rx=rz
;; Call this op-extend.
(define_split
[(set (match_operand 0 "cris_nonsp_register_operand" "")
(match_operator
3 "cris_orthogonal_operator"
[(match_operand 1 "register_operand" "")
(match_operand 2 "memory_operand" "")]))]
"!TARGET_V32
&& REG_P (operands[0])
&& REG_P (operands[1])
&& REGNO (operands[1]) != REGNO (operands[0])
&& GET_MODE_SIZE (GET_MODE (operands[0])) <= UNITS_PER_WORD
&& REG_P (XEXP (operands[2], 0))
&& REGNO (XEXP (operands[2], 0)) != REGNO (operands[0])"
[(set (match_dup 0)
(match_dup 1))
(set (match_dup 0)
(match_op_dup
3 [(match_dup 0)
(match_dup 2)]))]
"")
;; As op-extend-split-rx=rz, non-extended.
;; Call this op-split-rx=rz
(define_split
[(set (match_operand 0 "cris_nonsp_register_operand" "")
(match_operator
3 "cris_commutative_orth_op"
[(match_operand 2 "memory_operand" "")
(match_operand 1 "register_operand" "")]))]
"!TARGET_V32
&& REG_P (operands[0])
&& REG_P (operands[1])
&& REGNO (operands[1]) != REGNO (operands[0])
&& GET_MODE_SIZE (GET_MODE (operands[0])) <= UNITS_PER_WORD
&& REG_P (XEXP (operands[2], 0))
&& REGNO (XEXP (operands[2], 0)) != REGNO (operands[0])"
[(set (match_dup 0)
(match_dup 1))
(set (match_dup 0)
(match_op_dup
3 [(match_dup 0)
(match_dup 2)]))]
"")
;; As op-extend-split-swapped, nonextended.
;; Call this op-split-swapped.
(define_split
[(set (match_operand 0 "cris_nonsp_register_operand" "")
(match_operator
3 "cris_commutative_orth_op"
[(match_operand 1 "register_operand" "")
(match_operand 2 "memory_operand" "")]))]
"!TARGET_V32
&& REG_P (operands[0]) && REG_P (operands[1])
&& REGNO (operands[1]) != REGNO (operands[0])
&& GET_MODE_SIZE (GET_MODE (operands[0])) <= UNITS_PER_WORD
&& REG_P (XEXP (operands[2], 0))
&& REGNO (XEXP (operands[2], 0)) == REGNO (operands[0])"
[(set (match_dup 0)
(match_dup 2))
(set (match_dup 0)
(match_op_dup
3 [(match_dup 0)
(match_dup 1)]))]
"")
;; As op-extend-split-swapped-rx=rz, non-extended.
;; Call this op-split-swapped-rx=rz.
(define_split
[(set (match_operand 0 "cris_nonsp_register_operand" "")
(match_operator
3 "cris_orthogonal_operator"
[(match_operand 2 "memory_operand" "")
(match_operand 1 "register_operand" "")]))]
"!TARGET_V32
&& REG_P (operands[0]) && REG_P (operands[1])
&& REGNO (operands[1]) != REGNO (operands[0])
&& GET_MODE_SIZE (GET_MODE (operands[0])) <= UNITS_PER_WORD
&& REG_P (XEXP (operands[2], 0))
&& REGNO (XEXP (operands[2], 0)) == REGNO (operands[0])"
[(set (match_dup 0)
(match_dup 2))
(set (match_dup 0)
(match_op_dup
3 [(match_dup 0)
(match_dup 1)]))]
"")
(include "sync.md")
;; Splits for all cases in side-effect insns where (possibly after reload
;; and register allocation) rx and ry in [rx=ry+i] are equal.
;; move.S1 [rx=rx+rz.S2],ry
(define_split
[(parallel
[(set (match_operand 0 "register_operand" "")
(match_operator
6 "cris_mem_op"
[(plus:SI
(mult:SI (match_operand:SI 1 "register_operand" "")
(match_operand:SI 2 "const_int_operand" ""))
(match_operand:SI 3 "register_operand" ""))]))
(set (match_operand:SI 4 "register_operand" "")
(plus:SI (mult:SI (match_dup 1)
(match_dup 2))
(match_dup 3)))])]
"REG_P (operands[3]) && REG_P (operands[4])
&& REGNO (operands[3]) == REGNO (operands[4])"
[(set (match_dup 4) (plus:SI (mult:SI (match_dup 1) (match_dup 2))
(match_dup 3)))
(set (match_dup 0) (match_dup 5))]
"operands[5] = replace_equiv_address (operands[6], operands[3]);")
;; move.S1 [rx=rx+i],ry
(define_split
[(parallel
[(set (match_operand 0 "register_operand" "")
(match_operator
5 "cris_mem_op"
[(plus:SI (match_operand:SI 1 "cris_bdap_operand" "")
(match_operand:SI 2 "cris_bdap_operand" ""))]))
(set (match_operand:SI 3 "register_operand" "")
(plus:SI (match_dup 1)
(match_dup 2)))])]
"(rtx_equal_p (operands[3], operands[1])
|| rtx_equal_p (operands[3], operands[2]))"
[(set (match_dup 3) (plus:SI (match_dup 1) (match_dup 2)))
(set (match_dup 0) (match_dup 4))]
{
operands[4] = replace_equiv_address (operands[5], operands[3]);
cris_order_for_addsi3 (operands, 1);
})
;; move.S1 ry,[rx=rx+rz.S2]
(define_split
[(parallel
[(set (match_operator
6 "cris_mem_op"
[(plus:SI
(mult:SI (match_operand:SI 0 "register_operand" "")
(match_operand:SI 1 "const_int_operand" ""))
(match_operand:SI 2 "register_operand" ""))])
(match_operand 3 "register_operand" ""))
(set (match_operand:SI 4 "register_operand" "")
(plus:SI (mult:SI (match_dup 0)
(match_dup 1))
(match_dup 2)))])]
"REG_P (operands[2]) && REG_P (operands[4])
&& REGNO (operands[4]) == REGNO (operands[2])"
[(set (match_dup 4) (plus:SI (mult:SI (match_dup 0) (match_dup 1))
(match_dup 2)))
(set (match_dup 5) (match_dup 3))]
"operands[5] = replace_equiv_address (operands[6], operands[4]);")
;; move.S1 ry,[rx=rx+i]
(define_split
[(parallel
[(set (match_operator
6 "cris_mem_op"
[(plus:SI (match_operand:SI 0 "cris_bdap_operand" "")
(match_operand:SI 1 "cris_bdap_operand" ""))])
(match_operand 2 "register_operand" ""))
(set (match_operand:SI 3 "register_operand" "")
(plus:SI (match_dup 0)
(match_dup 1)))])]
"(rtx_equal_p (operands[3], operands[0])
|| rtx_equal_p (operands[3], operands[1]))"
[(set (match_dup 3) (plus:SI (match_dup 0) (match_dup 1)))
(set (match_dup 5) (match_dup 2))]
{
operands[5] = replace_equiv_address (operands[6], operands[3]);
cris_order_for_addsi3 (operands, 0);
})
;; clear.[bwd] [rx=rx+rz.S2]
(define_split
[(parallel
[(set (mem:BWD (plus:SI
(mult:SI (match_operand:SI 0 "register_operand" "")
(match_operand:SI 1 "const_int_operand" ""))
(match_operand:SI 2 "register_operand" "")))
(const_int 0))
(set (match_operand:SI 3 "register_operand" "")
(plus:SI (mult:SI (match_dup 0)
(match_dup 1))
(match_dup 2)))])]
"REG_P (operands[2]) && REG_P (operands[3])
&& REGNO (operands[3]) == REGNO (operands[2])"
[(set (match_dup 3) (plus:SI (mult:SI (match_dup 0) (match_dup 1))
(match_dup 2)))
(set (mem:BWD (match_dup 3)) (const_int 0))]
"")
;; clear.[bwd] [rx=rx+i]
(define_split
[(parallel
[(set (mem:BWD
(plus:SI (match_operand:SI 0 "cris_bdap_operand" "")
(match_operand:SI 1 "cris_bdap_operand" "")))
(const_int 0))
(set (match_operand:SI 2 "register_operand" "")
(plus:SI (match_dup 0)
(match_dup 1)))])]
"(rtx_equal_p (operands[0], operands[2])
|| rtx_equal_p (operands[2], operands[1]))"
[(set (match_dup 2) (plus:SI (match_dup 0) (match_dup 1)))
(set (mem:BWD (match_dup 2)) (const_int 0))]
"cris_order_for_addsi3 (operands, 0);")
;; mov(s|u).S1 [rx=rx+rz.S2],ry
(define_split
[(parallel
[(set (match_operand 0 "register_operand" "")
(match_operator
5 "cris_extend_operator"
[(mem (plus:SI
(mult:SI (match_operand:SI 1 "register_operand" "")
(match_operand:SI 2 "const_int_operand" ""))
(match_operand:SI 3 "register_operand" "")))]))
(set (match_operand:SI 4 "register_operand" "")
(plus:SI (mult:SI (match_dup 1)
(match_dup 2))
(match_dup 3)))])]
"REG_P (operands[3])
&& REG_P (operands[4])
&& REGNO (operands[3]) == REGNO (operands[4])"
[(set (match_dup 4) (plus:SI (mult:SI (match_dup 1) (match_dup 2))
(match_dup 3)))
(set (match_dup 0) (match_op_dup 5 [(match_dup 6)]))]
"operands[6] = replace_equiv_address (XEXP (operands[5], 0), operands[4]);")
;; mov(s|u).S1 [rx=rx+i],ry
(define_split
[(parallel
[(set (match_operand 0 "register_operand" "")
(match_operator
4 "cris_extend_operator"
[(mem (plus:SI
(match_operand:SI 1 "cris_bdap_operand" "")
(match_operand:SI 2 "cris_bdap_operand" "")))]))
(set (match_operand:SI 3 "register_operand" "")
(plus:SI (match_dup 1)
(match_dup 2)))])]
"(rtx_equal_p (operands[1], operands[3])
|| rtx_equal_p (operands[2], operands[3]))"
[(set (match_dup 3) (plus:SI (match_dup 1) (match_dup 2)))
(set (match_dup 0) (match_op_dup 4 [(match_dup 5)]))]
{
operands[5] = replace_equiv_address (XEXP (operands[4], 0), operands[3]);
cris_order_for_addsi3 (operands, 1);
})
;; op.S1 [rx=rx+i],ry
(define_split
[(parallel
[(set (match_operand 0 "register_operand" "")
(match_operator
5 "cris_orthogonal_operator"
[(match_operand 1 "register_operand" "")
(mem (plus:SI
(match_operand:SI 2 "cris_bdap_operand" "")
(match_operand:SI 3 "cris_bdap_operand" "")))]))
(set (match_operand:SI 4 "register_operand" "")
(plus:SI (match_dup 2)
(match_dup 3)))])]
"(rtx_equal_p (operands[4], operands[2])
|| rtx_equal_p (operands[4], operands[3]))"
[(set (match_dup 4) (plus:SI (match_dup 2) (match_dup 3)))
(set (match_dup 0) (match_op_dup 5 [(match_dup 1) (match_dup 6)]))]
{
operands[6] = replace_equiv_address (XEXP (operands[5], 1), operands[4]);
cris_order_for_addsi3 (operands, 2);
})
;; op.S1 [rx=rx+rz.S2],ry
(define_split
[(parallel
[(set (match_operand 0 "register_operand" "")
(match_operator
6 "cris_orthogonal_operator"
[(match_operand 1 "register_operand" "")
(mem (plus:SI
(mult:SI (match_operand:SI 2 "register_operand" "")
(match_operand:SI 3 "const_int_operand" ""))
(match_operand:SI 4 "register_operand" "")))]))
(set (match_operand:SI 5 "register_operand" "")
(plus:SI (mult:SI (match_dup 2)
(match_dup 3))
(match_dup 4)))])]
"REG_P (operands[4])
&& REG_P (operands[5])
&& REGNO (operands[5]) == REGNO (operands[4])"
[(set (match_dup 5) (plus:SI (mult:SI (match_dup 2) (match_dup 3))
(match_dup 4)))
(set (match_dup 0) (match_op_dup 6 [(match_dup 1) (match_dup 7)]))]
"operands[7] = replace_equiv_address (XEXP (operands[6], 1), operands[5]);")
;; op.S1 [rx=rx+rz.S2],ry (swapped)
(define_split
[(parallel
[(set (match_operand 0 "register_operand" "")
(match_operator
6 "cris_commutative_orth_op"
[(mem (plus:SI
(mult:SI (match_operand:SI 2 "register_operand" "")
(match_operand:SI 3 "const_int_operand" ""))
(match_operand:SI 4 "register_operand" "")))
(match_operand 1 "register_operand" "")]))
(set (match_operand:SI 5 "register_operand" "")
(plus:SI (mult:SI (match_dup 2)
(match_dup 3))
(match_dup 4)))])]
"REG_P (operands[4])
&& REG_P (operands[5])
&& REGNO (operands[5]) == REGNO (operands[4])"
[(set (match_dup 5) (plus:SI (mult:SI (match_dup 2) (match_dup 3))
(match_dup 4)))
(set (match_dup 0) (match_op_dup 6 [(match_dup 7) (match_dup 1)]))]
"operands[7] = replace_equiv_address (XEXP (operands[6], 0), operands[5]);")
;; op.S1 [rx=rx+i],ry (swapped)
(define_split
[(parallel
[(set (match_operand 0 "register_operand" "")
(match_operator
5 "cris_commutative_orth_op"
[(mem
(plus:SI (match_operand:SI 2 "cris_bdap_operand" "")
(match_operand:SI 3 "cris_bdap_operand" "")))
(match_operand 1 "register_operand" "")]))
(set (match_operand:SI 4 "register_operand" "")
(plus:SI (match_dup 2)
(match_dup 3)))])]
"(rtx_equal_p (operands[4], operands[2])
|| rtx_equal_p (operands[4], operands[3]))"
[(set (match_dup 4) (plus:SI (match_dup 2) (match_dup 3)))
(set (match_dup 0) (match_op_dup 5 [(match_dup 6) (match_dup 1)]))]
{
operands[6] = replace_equiv_address (XEXP (operands[5], 0), operands[4]);
cris_order_for_addsi3 (operands, 2);
})
;; op(s|u).S1 [rx=rx+rz.S2],ry
(define_split
[(parallel
[(set (match_operand 0 "register_operand" "")
(match_operator
6 "cris_operand_extend_operator"
[(match_operand 1 "register_operand" "")
(match_operator
7 "cris_extend_operator"
[(mem (plus:SI
(mult:SI (match_operand:SI 2 "register_operand" "")
(match_operand:SI 3 "const_int_operand" ""))
(match_operand:SI 4 "register_operand" "")))])]))
(set (match_operand:SI 5 "register_operand" "")
(plus:SI (mult:SI (match_dup 2)
(match_dup 3))
(match_dup 4)))])]
"REG_P (operands[4])
&& REG_P (operands[5])
&& REGNO (operands[5]) == REGNO (operands[4])"
[(set (match_dup 5) (plus:SI (mult:SI (match_dup 2) (match_dup 3))
(match_dup 4)))
(set (match_dup 0) (match_op_dup 6 [(match_dup 1) (match_dup 8)]))]
"operands[8] = gen_rtx_fmt_e (GET_CODE (operands[7]), GET_MODE (operands[7]),
replace_equiv_address (XEXP (operands[7], 0),
operands[5]));")
;; op(s|u).S1 [rx=rx+i],ry
(define_split
[(parallel
[(set (match_operand 0 "register_operand" "")
(match_operator
5 "cris_operand_extend_operator"
[(match_operand 1 "register_operand" "")
(match_operator
6 "cris_extend_operator"
[(mem
(plus:SI (match_operand:SI 2 "cris_bdap_operand" "")
(match_operand:SI 3 "cris_bdap_operand" "")
))])]))
(set (match_operand:SI 4 "register_operand" "")
(plus:SI (match_dup 2)
(match_dup 3)))])]
"(rtx_equal_p (operands[4], operands[2])
|| rtx_equal_p (operands[4], operands[3]))"
[(set (match_dup 4) (plus:SI (match_dup 2) (match_dup 3)))
(set (match_dup 0) (match_op_dup 5 [(match_dup 1) (match_dup 7)]))]
{
operands[7] = gen_rtx_fmt_e (GET_CODE (operands[6]), GET_MODE (operands[6]),
replace_equiv_address (XEXP (operands[6], 0),
operands[4]));
cris_order_for_addsi3 (operands, 2);
})
;; op(s|u).S1 [rx=rx+rz.S2],ry (swapped, plus or bound)
(define_split
[(parallel
[(set (match_operand 0 "register_operand" "")
(match_operator
7 "cris_plus_or_bound_operator"
[(match_operator
6 "cris_extend_operator"
[(mem (plus:SI
(mult:SI (match_operand:SI 2 "register_operand" "")
(match_operand:SI 3 "const_int_operand" ""))
(match_operand:SI 4 "register_operand" "")))])
(match_operand 1 "register_operand" "")]))
(set (match_operand:SI 5 "register_operand" "")
(plus:SI (mult:SI (match_dup 2)
(match_dup 3))
(match_dup 4)))])]
"REG_P (operands[4]) && REG_P (operands[5])
&& REGNO (operands[5]) == REGNO (operands[4])"
[(set (match_dup 5) (plus:SI (mult:SI (match_dup 2) (match_dup 3))
(match_dup 4)))
(set (match_dup 0) (match_op_dup 6 [(match_dup 8) (match_dup 1)]))]
"operands[8] = gen_rtx_fmt_e (GET_CODE (operands[6]), GET_MODE (operands[6]),
replace_equiv_address (XEXP (operands[6], 0),
operands[5]));")
;; op(s|u).S1 [rx=rx+i],ry (swapped, plus or bound)
(define_split
[(parallel
[(set (match_operand 0 "register_operand" "")
(match_operator
6 "cris_plus_or_bound_operator"
[(match_operator
5 "cris_extend_operator"
[(mem (plus:SI
(match_operand:SI 2 "cris_bdap_operand" "")
(match_operand:SI 3 "cris_bdap_operand" "")))])
(match_operand 1 "register_operand" "")]))
(set (match_operand:SI 4 "register_operand" "")
(plus:SI (match_dup 2)
(match_dup 3)))])]
"(rtx_equal_p (operands[4], operands[2])
|| rtx_equal_p (operands[4], operands[3]))"
[(set (match_dup 4) (plus:SI (match_dup 2) (match_dup 3)))
(set (match_dup 0) (match_op_dup 6 [(match_dup 7) (match_dup 1)]))]
{
operands[7] = gen_rtx_fmt_e (GET_CODE (operands[5]), GET_MODE (operands[5]),
replace_equiv_address (XEXP (operands[5], 0),
operands[4]));
cris_order_for_addsi3 (operands, 2);
})
;; Splits for addressing prefixes that have no side-effects, so we can
;; fill a delay slot. Never split if we lose something, though.
;; If we have a
;; move [indirect_ref],rx
;; where indirect ref = {const, [r+], [r]}, it costs as much as
;; move indirect_ref,rx
;; move [rx],rx
;; Take care not to allow indirect_ref = register.
;; We're not allowed to generate copies of registers with different mode
;; until after reload; copying pseudos upsets reload. CVS as of
;; 2001-08-24, unwind-dw2-fde.c, _Unwind_Find_FDE ICE in
;; cselib_invalidate_regno. Also, don't do this for the stack-pointer,
;; as we don't want it set temporarily to an invalid value.
(define_split ; indir_to_reg_split
[(set (match_operand 0 "cris_nonsp_register_operand" "")
(match_operand 1 "indirect_operand" ""))]
"reload_completed
&& REG_P (operands[0])
&& GET_MODE_SIZE (GET_MODE (operands[0])) <= UNITS_PER_WORD
&& (MEM_P (XEXP (operands[1], 0)) || CONSTANT_P (XEXP (operands[1], 0)))
&& REGNO (operands[0]) < CRIS_LAST_GENERAL_REGISTER"
[(set (match_dup 2) (match_dup 4))
(set (match_dup 0) (match_dup 3))]
"operands[2] = gen_rtx_REG (Pmode, REGNO (operands[0]));
operands[3] = replace_equiv_address (operands[1], operands[2]);
operands[4] = XEXP (operands[1], 0);")
;; As the above, but MOVS and MOVU.
(define_split
[(set (match_operand 0 "cris_nonsp_register_operand" "")
(match_operator
4 "cris_extend_operator"
[(match_operand 1 "indirect_operand" "")]))]
"reload_completed
&& REG_P (operands[0])
&& GET_MODE_SIZE (GET_MODE (operands[0])) <= UNITS_PER_WORD
&& (MEM_P (XEXP (operands[1], 0))
|| CONSTANT_P (XEXP (operands[1], 0)))"
[(set (match_dup 2) (match_dup 5))
(set (match_dup 0) (match_op_dup 4 [(match_dup 3)]))]
"operands[2] = gen_rtx_REG (Pmode, REGNO (operands[0]));
operands[3] = replace_equiv_address (XEXP (operands[4], 0), operands[2]);
operands[5] = XEXP (operands[1], 0);")
;; Various peephole optimizations.
;;
;; Watch out: when you exchange one set of instructions for another, the
;; condition codes setting must be the same, or you have to CC_INIT or
;; whatever is appropriate, in the pattern before you emit the
;; assembly text. This is best done here, not in cris_notice_update_cc,
;; to keep changes local to their cause.
;;
;; Do not add patterns that you do not know will be matched.
;; Please also add a self-contained testcase.
;; We have trouble with and:s and shifts. Maybe something is broken in
;; gcc? Or it could just be that bit-field insn expansion is a bit
;; suboptimal when not having extzv insns.
;; Testcase for the following four peepholes: gcc.dg/cris-peep2-xsrand.c
(define_peephole2 ; asrandb (peephole casesi+31)
[(set (match_operand:SI 0 "register_operand" "")
(ashiftrt:SI (match_dup 0)
(match_operand:SI 1 "const_int_operand" "")))
(set (match_dup 0)
(and:SI (match_dup 0)
(match_operand 2 "const_int_operand" "")))]
"INTVAL (operands[2]) > 31
&& INTVAL (operands[2]) < 255
&& INTVAL (operands[1]) > 23
/* Check that the and-operation enables us to use logical-shift. */
&& (INTVAL (operands[2])
& ((HOST_WIDE_INT) -1 << (32 - INTVAL (operands[1])))) == 0"
[(set (match_dup 0) (lshiftrt:SI (match_dup 0) (match_dup 1)))
(set (match_dup 3) (and:QI (match_dup 3) (match_dup 4)))]
;; FIXME: CC0 is valid except for the M bit.
{
operands[3] = gen_rtx_REG (QImode, REGNO (operands[0]));
operands[4] = GEN_INT (trunc_int_for_mode (INTVAL (operands[2]), QImode));
})
(define_peephole2 ; asrandw (peephole casesi+32)
[(set (match_operand:SI 0 "register_operand" "")
(ashiftrt:SI (match_dup 0)
(match_operand:SI 1 "const_int_operand" "")))
(set (match_dup 0)
(and:SI (match_dup 0) (match_operand 2 "const_int_operand" "")))]
"INTVAL (operands[2]) > 31
&& INTVAL (operands[2]) < 65535
&& INTVAL (operands[2]) != 255
&& INTVAL (operands[1]) > 15
/* Check that the and-operation enables us to use logical-shift. */
&& (INTVAL (operands[2])
& ((HOST_WIDE_INT) -1 << (32 - INTVAL (operands[1])))) == 0"
[(set (match_dup 0) (lshiftrt:SI (match_dup 0) (match_dup 1)))
(set (match_dup 3) (and:HI (match_dup 3) (match_dup 4)))]
;; FIXME: CC0 is valid except for the M bit.
{
operands[3] = gen_rtx_REG (HImode, REGNO (operands[0]));
operands[4] = GEN_INT (trunc_int_for_mode (INTVAL (operands[2]), HImode));
})
(define_peephole2 ; lsrandb (peephole casesi+33)
[(set (match_operand:SI 0 "register_operand" "")
(lshiftrt:SI (match_dup 0)
(match_operand:SI 1 "const_int_operand" "")))
(set (match_dup 0)
(and:SI (match_dup 0) (match_operand 2 "const_int_operand" "")))]
"INTVAL (operands[2]) > 31
&& INTVAL (operands[2]) < 255
&& INTVAL (operands[1]) > 23"
[(set (match_dup 0) (lshiftrt:SI (match_dup 0) (match_dup 1)))
(set (match_dup 3) (and:QI (match_dup 3) (match_dup 4)))]
;; FIXME: CC0 is valid except for the M bit.
{
operands[3] = gen_rtx_REG (QImode, REGNO (operands[0]));
operands[4] = GEN_INT (trunc_int_for_mode (INTVAL (operands[2]), QImode));
})
(define_peephole2 ; lsrandw (peephole casesi+34)
[(set (match_operand:SI 0 "register_operand" "")
(lshiftrt:SI (match_dup 0)
(match_operand:SI 1 "const_int_operand" "")))
(set (match_dup 0)
(and:SI (match_dup 0) (match_operand 2 "const_int_operand" "")))]
"INTVAL (operands[2]) > 31 && INTVAL (operands[2]) < 65535
&& INTVAL (operands[2]) != 255
&& INTVAL (operands[1]) > 15"
[(set (match_dup 0) (lshiftrt:SI (match_dup 0) (match_dup 1)))
(set (match_dup 3) (and:HI (match_dup 3) (match_dup 4)))]
;; FIXME: CC0 is valid except for the M bit.
{
operands[3] = gen_rtx_REG (HImode, REGNO (operands[0]));
operands[4] = GEN_INT (trunc_int_for_mode (INTVAL (operands[2]), HImode));
})
;; Change
;; add.d n,rx
;; move [rx],ry
;; into
;; move [rx=rx+n],ry
;; when -128 <= n <= 127.
;; This will reduce the size of the assembler code for n = [-128..127],
;; and speed up accordingly. Don't match if the previous insn is
;; (set rx rz) because that combination is matched by another peephole.
;; No stable test-case.
(define_peephole2 ; moversideqi (peephole casesi+35)
[(set (match_operand:SI 0 "register_operand" "")
(plus:SI (match_operand:SI 1 "register_operand" "")
(match_operand:SI 2 "const_int_operand" "")))
(set (match_operand 3 "register_operand" "")
(match_operator 4 "cris_mem_op" [(match_dup 0)]))]
"GET_MODE_SIZE (GET_MODE (operands[4])) <= UNITS_PER_WORD
&& REGNO (operands[3]) != REGNO (operands[0])
&& (cris_base_p (operands[1], true) || cris_base_p (operands[2], true))
&& !satisfies_constraint_J (operands[2])
&& !satisfies_constraint_N (operands[2])
&& (INTVAL (operands[2]) >= -128 && INTVAL (operands[2]) < 128)
&& TARGET_SIDE_EFFECT_PREFIXES"
[(parallel
[(set (match_dup 3) (match_dup 5))
(set (match_dup 0) (plus:SI (match_dup 1) (match_dup 2)))])]
;; Checking the previous insn is a bit too awkward for the condition.
{
rtx prev = prev_nonnote_insn (curr_insn);
if (prev != NULL_RTX)
{
rtx set = single_set (prev);
if (set != NULL_RTX
&& REG_S_P (SET_DEST (set))
&& REGNO (SET_DEST (set)) == REGNO (operands[0])
&& REG_S_P (SET_SRC (set)))
FAIL;
}
operands[5]
= replace_equiv_address (operands[4],
gen_rtx_PLUS (SImode,
operands[1], operands[2]));
})
;; Vice versa: move ry,[rx=rx+n]
(define_peephole2 ; movemsideqi (peephole casesi+36)
[(set (match_operand:SI 0 "register_operand" "")
(plus:SI (match_operand:SI 1 "register_operand" "")
(match_operand:SI 2 "const_int_operand" "")))
(set (match_operator 3 "cris_mem_op" [(match_dup 0)])
(match_operand 4 "register_operand" ""))]
"GET_MODE_SIZE (GET_MODE (operands[4])) <= UNITS_PER_WORD
&& REGNO (operands[4]) != REGNO (operands[0])
&& (cris_base_p (operands[1], true) || cris_base_p (operands[2], true))
&& !satisfies_constraint_J (operands[2])
&& !satisfies_constraint_N (operands[2])
&& (INTVAL (operands[2]) >= -128 && INTVAL (operands[2]) < 128)
&& TARGET_SIDE_EFFECT_PREFIXES"
[(parallel
[(set (match_dup 5) (match_dup 4))
(set (match_dup 0) (plus:SI (match_dup 1) (match_dup 2)))])]
"operands[5]
= replace_equiv_address (operands[3],
gen_rtx_PLUS (SImode,
operands[1], operands[2]));")
;; As above, change:
;; add.d n,rx
;; op.d [rx],ry
;; into:
;; op.d [rx=rx+n],ry
;; Saves when n = [-128..127].
;;
;; Splitting and joining combinations for side-effect modes are slightly
;; out of hand. They probably will not save the time they take typing in,
;; not to mention the bugs that creep in. FIXME: Get rid of as many of
;; the splits and peepholes as possible.
;; No stable test-case.
(define_peephole2 ; mover2side (peephole casesi+37)
[(set (match_operand:SI 0 "register_operand" "")
(plus:SI (match_operand:SI 1 "register_operand" "")
(match_operand:SI 2 "const_int_operand" "")))
(set (match_operand 3 "register_operand" "")
(match_operator 4 "cris_orthogonal_operator"
[(match_dup 3)
(match_operator
5 "cris_mem_op" [(match_dup 0)])]))]
;; FIXME: What about DFmode?
;; Change to GET_MODE_SIZE (GET_MODE (operands[3])) <= UNITS_PER_WORD?
"GET_MODE (operands[3]) != DImode
&& REGNO (operands[0]) != REGNO (operands[3])
&& !satisfies_constraint_J (operands[2])
&& !satisfies_constraint_N (operands[2])
&& INTVAL (operands[2]) >= -128
&& INTVAL (operands[2]) <= 127
&& TARGET_SIDE_EFFECT_PREFIXES"
[(parallel
[(set (match_dup 3) (match_op_dup 4 [(match_dup 3) (match_dup 6)]))
(set (match_dup 0) (plus:SI (match_dup 1) (match_dup 2)))])]
"operands[6]
= replace_equiv_address (operands[5],
gen_rtx_PLUS (SImode,
operands[1], operands[2]));")
;; Sometimes, for some reason the pattern
;; move x,rx
;; add y,rx
;; move [rx],rz
;; will occur. Solve this, and likewise for to-memory.
;; No stable test-case.
(define_peephole2 ; moverside (peephole casesi+38)
[(set (match_operand:SI 0 "register_operand" "")
(match_operand:SI 1 "cris_bdap_biap_operand" ""))
(set (match_dup 0)
(plus:SI (match_operand:SI 2 "cris_bdap_biap_operand" "")
(match_operand:SI 3 "cris_bdap_biap_operand" "")))
(set (match_operand 4 "register_operand" "")
(match_operator 5 "cris_mem_op" [(match_dup 0)]))]
"(rtx_equal_p (operands[2], operands[0])
|| rtx_equal_p (operands[3], operands[0]))
&& cris_side_effect_mode_ok (PLUS, operands, 0,
(REG_S_P (operands[1])
? 1
: (rtx_equal_p (operands[2], operands[0])
? 3 : 2)),
(! REG_S_P (operands[1])
? 1
: (rtx_equal_p (operands[2], operands[0])
? 3 : 2)),
-1, 4)"
[(parallel
[(set (match_dup 4) (match_dup 6))
(set (match_dup 0) (plus:SI (match_dup 7) (match_dup 8)))])]
{
rtx otherop
= rtx_equal_p (operands[2], operands[0]) ? operands[3] : operands[2];
/* Make sure we have canonical RTX so we match the insn pattern -
not a constant in the first operand. We also require the order
(plus reg mem) to match the final pattern. */
if (CONSTANT_P (otherop) || MEM_P (otherop))
{
operands[7] = operands[1];
operands[8] = otherop;
}
else
{
operands[7] = otherop;
operands[8] = operands[1];
}
operands[6]
= replace_equiv_address (operands[5],
gen_rtx_PLUS (SImode,
operands[7], operands[8]));
})
;; As above but to memory.
;; FIXME: Split movemside and moverside into variants and prune
;; the ones that don't trig.
;; No stable test-case.
(define_peephole2 ; movemside (peephole casesi+39)
[(set (match_operand:SI 0 "register_operand" "")
(match_operand:SI 1 "cris_bdap_biap_operand" ""))
(set (match_dup 0)
(plus:SI (match_operand:SI 2 "cris_bdap_biap_operand" "")
(match_operand:SI 3 "cris_bdap_biap_operand" "")))
(set (match_operator 4 "cris_mem_op" [(match_dup 0)])
(match_operand 5 "register_operand" ""))]
"(rtx_equal_p (operands[2], operands[0])
|| rtx_equal_p (operands[3], operands[0]))
&& cris_side_effect_mode_ok (PLUS, operands, 0,
(REG_S_P (operands[1])
? 1
: (rtx_equal_p (operands[2], operands[0])
? 3 : 2)),
(! REG_S_P (operands[1])
? 1
: (rtx_equal_p (operands[2], operands[0])
? 3 : 2)),
-1, 5)"
[(parallel
[(set (match_dup 6) (match_dup 5))
(set (match_dup 0) (plus:SI (match_dup 7) (match_dup 8)))])]
{
rtx otherop
= rtx_equal_p (operands[2], operands[0]) ? operands[3] : operands[2];
/* Make sure we have canonical RTX so we match the insn pattern -
not a constant in the first operand. We also require the order
(plus reg mem) to match the final pattern. */
if (CONSTANT_P (otherop) || MEM_P (otherop))
{
operands[7] = operands[1];
operands[8] = otherop;
}
else
{
operands[7] = otherop;
operands[8] = operands[1];
}
operands[6]
= replace_equiv_address (operands[4],
gen_rtx_PLUS (SImode,
operands[7], operands[8]));
})
;; Another spotted bad code:
;; move rx,ry
;; move [ry],ry
;; No stable test-case.
(define_peephole2 ; movei (peephole casesi+42)
[(set (match_operand:SI 0 "register_operand" "")
(match_operand:SI 1 "register_operand" ""))
(set (match_operand 2 "register_operand" "")
(match_operator 3 "cris_mem_op" [(match_dup 0)]))]
"REGNO (operands[0]) == REGNO (operands[2])
&& (REGNO_REG_CLASS (REGNO (operands[0]))
== REGNO_REG_CLASS (REGNO (operands[1])))
&& GET_MODE_SIZE (GET_MODE (operands[2])) <= UNITS_PER_WORD"
[(set (match_dup 2) (match_dup 4))]
"operands[4] = replace_equiv_address (operands[3], operands[1]);")
;; move.d [r10+16],r9
;; and.d r12,r9
;; change to
;; and.d [r10+16],r12,r9
;; With generalization of the operation, the size and the addressing mode.
;; This seems to be the result of a quirk in register allocation
;; missing the three-operand cases when having different predicates.
;; Maybe that it matters that it is a commutative operation.
;; This pattern helps that situation, but there's still the increased
;; register pressure.
;; Note that adding the noncommutative variant did not show any matches
;; in ipps and cc1, so it's not here.
;; No stable test-case.
(define_peephole2 ; op3 (peephole casesi+44)
[(set (match_operand 0 "register_operand" "")
(match_operator
6 "cris_mem_op"
[(plus:SI
(match_operand:SI 1 "cris_bdap_biap_operand" "")
(match_operand:SI 2 "cris_bdap_biap_operand" ""))]))
(set (match_dup 0)
(match_operator
5 "cris_commutative_orth_op"
[(match_operand 3 "register_operand" "")
(match_operand 4 "register_operand" "")]))]
"(rtx_equal_p (operands[3], operands[0])
|| rtx_equal_p (operands[4], operands[0]))
&& ! rtx_equal_p (operands[3], operands[4])
&& (REG_S_P (operands[1]) || REG_S_P (operands[2]))
&& GET_MODE_SIZE (GET_MODE (operands[0])) <= UNITS_PER_WORD"
[(set (match_dup 0) (match_op_dup 5 [(match_dup 7) (match_dup 6)]))]
"operands[7]
= rtx_equal_p (operands[3], operands[0]) ? operands[4] : operands[3];")
;; There seems to be no other way to make GCC (including 4.8/trunk at
;; r186932) optimally reload an instruction that looks like
;; and.d reg_or_mem,const_32__65535,other_reg
;; where other_reg is the destination.
;; It should be:
;; movu.[bw] reg_or_mem,reg_32
;; and.[bw] trunc_int_for_mode([bw], const_32__65535),reg_32 ;; or andq
;; but it turns into:
;; move.d reg_or_mem,reg_32
;; and.d const_32__65535,reg_32
;; Fix it with these two peephole2's.
;; Testcases: gcc.dg/cris-peep2-andu1.c gcc.dg/cris-peep2-andu2.c
(define_peephole2 ; andu (casesi+45)
[(set (match_operand:SI 0 "register_operand" "")
(match_operand:SI 1 "nonimmediate_operand" ""))
(set (match_operand:SI 2 "register_operand" "")
(and:SI (match_dup 0)
(match_operand:SI 3 "const_int_operand" "")))]
;; Since the size of the memory access could be made different here,
;; don't do this for a mem-volatile access.
"REGNO (operands[2]) == REGNO (operands[0])
&& INTVAL (operands[3]) <= 65535 && INTVAL (operands[3]) >= 0
&& !satisfies_constraint_I (operands[3])
&& !side_effects_p (operands[1])
&& (!REG_P (operands[1])
|| REGNO (operands[1]) <= CRIS_LAST_GENERAL_REGISTER)"
;; FIXME: CC0 valid except for M (i.e. CC_NOT_NEGATIVE).
[(set (match_dup 0) (match_dup 4))
(set (match_dup 5) (match_dup 6))]
{
enum machine_mode zmode = INTVAL (operands[3]) <= 255 ? QImode : HImode;
enum machine_mode amode
= satisfies_constraint_O (operands[3]) ? SImode : zmode;
rtx op1
= (REG_S_P (operands[1])
? gen_rtx_REG (zmode, REGNO (operands[1]))
: adjust_address (operands[1], zmode, 0));
operands[4]
= gen_rtx_ZERO_EXTEND (SImode, op1);
operands[5] = gen_rtx_REG (amode, REGNO (operands[0]));
operands[6]
= gen_rtx_AND (amode, gen_rtx_REG (amode, REGNO (operands[0])),
GEN_INT (trunc_int_for_mode (INTVAL (operands[3]),
amode == SImode
? QImode : amode)));
})
;; Since r186861, gcc.dg/cris-peep2-andu2.c trigs this pattern, with which
;; we fix up e.g.:
;; movu.b 254,$r9.
;; and.d $r10,$r9
;; into:
;; movu.b $r10,$r9
;; andq -2,$r9.
;; Only do this for values fitting the quick immediate operand.
(define_peephole2 ; andqu (casesi+46)
[(set (match_operand:SI 0 "register_operand")
(match_operand:SI 1 "const_int_operand"))
(set (match_dup 0)
(and:SI (match_dup 0) (match_operand:SI 2 "nonimmediate_operand")))]
;; Since the size of the memory access will be made different here,
;; don't do this for a volatile access or a post-incremented address.
"satisfies_constraint_O (operands[1])
&& !side_effects_p (operands[2])
&& !reg_overlap_mentioned_p (operands[0], operands[2])"
[(set (match_dup 0) (match_dup 3))
(set (match_dup 0) (and:SI (match_dup 0) (match_dup 4)))]
{
enum machine_mode zmode = INTVAL (operands[2]) <= 255 ? QImode : HImode;
rtx op1
= (REG_S_P (operands[2])
? gen_rtx_REG (zmode, REGNO (operands[2]))
: adjust_address (operands[2], zmode, 0));
operands[3] = gen_rtx_ZERO_EXTEND (SImode, op1);
operands[4] = GEN_INT (trunc_int_for_mode (INTVAL (operands[1]), QImode));
})
;; Try and avoid GOTPLT reads escaping a call: transform them into
;; PLT. Curiously (but thankfully), peepholes for instructions
;; *without side-effects* that just feed a call (or call_value) are
;; not matched neither in a build or test-suite, so those patterns are
;; omitted.
;; A "normal" move where we don't check the consumer.
(define_peephole2 ; gotplt-to-plt
[(set
(match_operand:SI 0 "register_operand" "")
(match_operator:SI
1 "cris_mem_op"
[(plus:SI
(reg:SI CRIS_GOT_REGNUM)
(const:SI
(unspec:SI [(match_operand:SI 2 "cris_general_operand_or_symbol" "")]
CRIS_UNSPEC_PLTGOTREAD)))]))]
"flag_pic
&& cris_valid_pic_const (XEXP (XEXP (operands[1], 0), 1), true)
&& REGNO_REG_CLASS (REGNO (operands[0])) == REGNO_REG_CLASS (0)"
[(set (match_dup 0) (const:SI (unspec:SI [(match_dup 2)] CRIS_UNSPEC_PLT_GOTREL)))
(set (match_dup 0) (plus:SI (match_dup 0) (reg:SI CRIS_GOT_REGNUM)))]
"")
;; And one set with a side-effect getting the PLTGOT offset.
;; First call and call_value variants.
(define_peephole2 ; gotplt-to-plt-side-call
[(parallel
[(set
(match_operand:SI 0 "register_operand" "")
(match_operator:SI
1 "cris_mem_op"
[(plus:SI
(reg:SI CRIS_GOT_REGNUM)
(const:SI
(unspec:SI [(match_operand:SI
2 "cris_general_operand_or_symbol" "")]
CRIS_UNSPEC_PLTGOTREAD)))]))
(set (match_operand:SI 3 "register_operand" "")
(plus:SI (reg:SI CRIS_GOT_REGNUM)
(const:SI
(unspec:SI [(match_dup 2)] CRIS_UNSPEC_PLTGOTREAD))))])
(parallel [(call (mem:QI (match_dup 0))
(match_operand 4 "" ""))
(clobber (reg:SI CRIS_SRP_REGNUM))])]
"flag_pic
&& cris_valid_pic_const (XEXP (XEXP (operands[1], 0), 1), true)
&& peep2_reg_dead_p (2, operands[0])"
[(parallel [(call (mem:QI (match_dup 1))
(match_dup 4))
(clobber (reg:SI CRIS_SRP_REGNUM))
(set (match_dup 3)
(plus:SI (reg:SI CRIS_GOT_REGNUM)
(const:SI
(unspec:SI [(match_dup 2)]
CRIS_UNSPEC_PLTGOTREAD))))])]
"")
(define_peephole2 ; gotplt-to-plt-side-call-value
[(parallel
[(set
(match_operand:SI 0 "register_operand" "")
(match_operator:SI
1 "cris_mem_op"
[(plus:SI
(reg:SI CRIS_GOT_REGNUM)
(const:SI
(unspec:SI [(match_operand:SI
2 "cris_general_operand_or_symbol" "")]
CRIS_UNSPEC_PLTGOTREAD)))]))
(set (match_operand:SI 3 "register_operand" "")
(plus:SI (reg:SI CRIS_GOT_REGNUM)
(const:SI
(unspec:SI [(match_dup 2)] CRIS_UNSPEC_PLTGOTREAD))))])
(parallel [(set (match_operand 5 "" "")
(call (mem:QI (match_dup 0))
(match_operand 4 "" "")))
(clobber (reg:SI CRIS_SRP_REGNUM))])]
"flag_pic
&& cris_valid_pic_const (XEXP (XEXP (operands[1], 0), 1), true)
&& peep2_reg_dead_p (2, operands[0])"
[(parallel [(set (match_dup 5)
(call (mem:QI (match_dup 1))
(match_dup 4)))
(clobber (reg:SI CRIS_SRP_REGNUM))
(set (match_dup 3)
(plus:SI (reg:SI CRIS_GOT_REGNUM)
(const:SI
(unspec:SI [(match_dup 2)]
CRIS_UNSPEC_PLTGOTREAD))))])]
"")
(define_peephole2 ; gotplt-to-plt-side
[(parallel
[(set
(match_operand:SI 0 "register_operand" "")
(match_operator:SI
1 "cris_mem_op"
[(plus:SI
(reg:SI CRIS_GOT_REGNUM)
(const:SI
(unspec:SI [(match_operand:SI
2 "cris_general_operand_or_symbol" "")]
CRIS_UNSPEC_PLTGOTREAD)))]))
(set (match_operand:SI 3 "register_operand" "")
(plus:SI (reg:SI CRIS_GOT_REGNUM)
(const:SI
(unspec:SI [(match_dup 2)] CRIS_UNSPEC_PLTGOTREAD))))])]
"flag_pic
&& cris_valid_pic_const (XEXP (XEXP (operands[1], 0), 1), true)
&& REGNO_REG_CLASS (REGNO (operands[0])) == REGNO_REG_CLASS (0)"
[(set (match_dup 3)
(const:SI (unspec:SI [(match_dup 2)] CRIS_UNSPEC_PLTGOTREAD)))
(set (match_dup 3) (plus:SI (match_dup 3) (reg:SI CRIS_GOT_REGNUM)))
(set (match_dup 0)
(const:SI (unspec:SI [(match_dup 2)] CRIS_UNSPEC_PLT_GOTREL)))
(set (match_dup 0) (plus:SI (match_dup 0) (reg:SI CRIS_GOT_REGNUM)))]
"")
;; Local variables:
;; mode:emacs-lisp
;; comment-start: ";; "
;; eval: (set-syntax-table (copy-sequence (syntax-table)))
;; eval: (modify-syntax-entry ?[ "(]")
;; eval: (modify-syntax-entry ?] ")[")
;; eval: (modify-syntax-entry ?{ "(}")
;; eval: (modify-syntax-entry ?} "){")
;; eval: (setq indent-tabs-mode t)
;; End: