aboutsummaryrefslogtreecommitdiff
diff options
context:
space:
mode:
authorYvan Roux <yvan.roux@linaro.org>2015-08-23 22:04:00 +0200
committerLinaro Code Review <review@review.linaro.org>2015-08-27 17:28:31 +0000
commit360a5bda57277ed6fbf064893484c651459999c1 (patch)
tree1ed9027f22540d8cf24d09b71a1defeffa3e2ba9
parent80b77eee8430cbb8182cbbf6b2c6199c3d6d3ca7 (diff)
gcc/
Backport from trunk r222624, r222625. 2015-04-30 Kyrylo Tkachov <kyrylo.tkachov@arm.com> * config/aarch64/aarch64.c (aarch64_shift_p): New function. (aarch64_rtx_mult_cost): Update comment to reflect that it also handles combined arithmetic-shift ops. Properly handle all shift and extend operations that can occur in combination with PLUS/MINUS. Rename maybe_fma to compound_p. (aarch64_rtx_costs): Use aarch64_shift_p when costing compound arithmetic and shift operations. Change-Id: I41566559c54f3d47e6cbeba38591281ec4deabaf
-rw-r--r--gcc/config/aarch64/aarch64.c54
1 files changed, 40 insertions, 14 deletions
diff --git a/gcc/config/aarch64/aarch64.c b/gcc/config/aarch64/aarch64.c
index e4010233bab..3b17a5793d0 100644
--- a/gcc/config/aarch64/aarch64.c
+++ b/gcc/config/aarch64/aarch64.c
@@ -5181,9 +5181,18 @@ aarch64_strip_extend (rtx x)
return x;
}
+/* Return true iff CODE is a shift supported in combination
+ with arithmetic instructions. */
+
+static bool
+aarch64_shift_p (enum rtx_code code)
+{
+ return code == ASHIFT || code == ASHIFTRT || code == LSHIFTRT;
+}
+
/* Helper function for rtx cost calculation. Calculate the cost of
- a MULT, which may be part of a multiply-accumulate rtx. Return
- the calculated cost of the expression, recursing manually in to
+ a MULT or ASHIFT, which may be part of a compound PLUS/MINUS rtx.
+ Return the calculated cost of the expression, recursing manually in to
operands where needed. */
static int
@@ -5193,7 +5202,7 @@ aarch64_rtx_mult_cost (rtx x, int code, int outer, bool speed)
const struct cpu_cost_table *extra_cost
= aarch64_tune_params->insn_extra_cost;
int cost = 0;
- bool maybe_fma = (outer == PLUS || outer == MINUS);
+ bool compound_p = (outer == PLUS || outer == MINUS);
machine_mode mode = GET_MODE (x);
gcc_checking_assert (code == MULT);
@@ -5208,18 +5217,35 @@ aarch64_rtx_mult_cost (rtx x, int code, int outer, bool speed)
if (GET_MODE_CLASS (mode) == MODE_INT)
{
/* The multiply will be canonicalized as a shift, cost it as such. */
- if (CONST_INT_P (op1)
- && exact_log2 (INTVAL (op1)) > 0)
+ if (aarch64_shift_p (GET_CODE (x))
+ || (CONST_INT_P (op1)
+ && exact_log2 (INTVAL (op1)) > 0))
{
+ bool is_extend = GET_CODE (op0) == ZERO_EXTEND
+ || GET_CODE (op0) == SIGN_EXTEND;
if (speed)
{
- if (maybe_fma)
- /* ADD (shifted register). */
- cost += extra_cost->alu.arith_shift;
+ if (compound_p)
+ {
+ if (REG_P (op1))
+ /* ARITH + shift-by-register. */
+ cost += extra_cost->alu.arith_shift_reg;
+ else if (is_extend)
+ /* ARITH + extended register. We don't have a cost field
+ for ARITH+EXTEND+SHIFT, so use extend_arith here. */
+ cost += extra_cost->alu.extend_arith;
+ else
+ /* ARITH + shift-by-immediate. */
+ cost += extra_cost->alu.arith_shift;
+ }
else
/* LSL (immediate). */
- cost += extra_cost->alu.shift;
+ cost += extra_cost->alu.shift;
+
}
+ /* Strip extends as we will have costed them in the case above. */
+ if (is_extend)
+ op0 = aarch64_strip_extend (op0);
cost += rtx_cost (op0, GET_CODE (op0), 0, speed);
@@ -5237,7 +5263,7 @@ aarch64_rtx_mult_cost (rtx x, int code, int outer, bool speed)
if (speed)
{
- if (maybe_fma)
+ if (compound_p)
/* MADD/SMADDL/UMADDL. */
cost += extra_cost->mult[0].extend_add;
else
@@ -5255,7 +5281,7 @@ aarch64_rtx_mult_cost (rtx x, int code, int outer, bool speed)
if (speed)
{
- if (maybe_fma)
+ if (compound_p)
/* MADD. */
cost += extra_cost->mult[mode == DImode].add;
else
@@ -5276,7 +5302,7 @@ aarch64_rtx_mult_cost (rtx x, int code, int outer, bool speed)
if (GET_CODE (op1) == NEG)
op1 = XEXP (op1, 0);
- if (maybe_fma)
+ if (compound_p)
/* FMADD/FNMADD/FNMSUB/FMSUB. */
cost += extra_cost->fp[mode == DFmode].fma;
else
@@ -5916,7 +5942,7 @@ cost_minus:
/* Cost this as an FMA-alike operation. */
if ((GET_CODE (new_op1) == MULT
- || GET_CODE (new_op1) == ASHIFT)
+ || aarch64_shift_p (GET_CODE (new_op1)))
&& code != COMPARE)
{
*cost += aarch64_rtx_mult_cost (new_op1, MULT,
@@ -5987,7 +6013,7 @@ cost_plus:
new_op0 = aarch64_strip_extend (op0);
if (GET_CODE (new_op0) == MULT
- || GET_CODE (new_op0) == ASHIFT)
+ || aarch64_shift_p (GET_CODE (new_op0)))
{
*cost += aarch64_rtx_mult_cost (new_op0, MULT, PLUS,
speed);