/* Memory address lowering and addressing mode selection.
- Copyright (C) 2004 Free Software Foundation, Inc.
+ Copyright (C) 2004, 2006, 2007, 2008, 2009 Free Software Foundation, Inc.
This file is part of GCC.
GCC is free software; you can redistribute it and/or modify it
under the terms of the GNU General Public License as published by the
-Free Software Foundation; either version 2, or (at your option) any
+Free Software Foundation; either version 3, or (at your option) any
later version.
GCC is distributed in the hope that it will be useful, but WITHOUT
for more details.
You should have received a copy of the GNU General Public License
-along with GCC; see the file COPYING. If not, write to the Free
-Software Foundation, 59 Temple Place - Suite 330, Boston, MA
-02111-1307, USA. */
+along with GCC; see the file COPYING3. If not see
+<http://www.gnu.org/licenses/>. */
/* Utility functions for manipulation with TARGET_MEM_REFs -- tree expressions
that directly map to addressing modes of the target. */
#include "recog.h"
#include "expr.h"
#include "ggc.h"
+#include "tree-affine.h"
/* TODO -- handling of symbols (according to Richard Hendersons
comments, http://gcc.gnu.org/ml/gcc-patches/2005-04/msg00949.html):
/* A "template" for memory address, used to determine whether the address is
valid for mode. */
-struct mem_addr_template GTY (())
-{
+struct GTY (()) mem_addr_template {
rtx ref; /* The template. */
rtx * GTY ((skip)) step_p; /* The point in template where the step should be
filled in. */
if (base)
{
if (*addr)
- *addr = gen_rtx_PLUS (Pmode, *addr, base);
+ *addr = simplify_gen_binary (PLUS, Pmode, base, *addr);
else
*addr = base;
}
act_elem = symbol;
if (offset)
{
- act_elem = gen_rtx_CONST (Pmode,
- gen_rtx_PLUS (Pmode, act_elem, offset));
+ act_elem = gen_rtx_PLUS (Pmode, act_elem, offset);
+
if (offset_p)
- *offset_p = &XEXP (XEXP (act_elem, 0), 1);
+ *offset_p = &XEXP (act_elem, 1);
+
+ if (GET_CODE (symbol) == SYMBOL_REF
+ || GET_CODE (symbol) == LABEL_REF
+ || GET_CODE (symbol) == CONST)
+ act_elem = gen_rtx_CONST (Pmode, act_elem);
}
if (*addr)
tree
tree_mem_ref_addr (tree type, tree mem_ref)
{
- tree addr = NULL_TREE;
+ tree addr;
tree act_elem;
tree step = TMR_STEP (mem_ref), offset = TMR_OFFSET (mem_ref);
+ tree sym = TMR_SYMBOL (mem_ref), base = TMR_BASE (mem_ref);
+ tree addr_base = NULL_TREE, addr_off = NULL_TREE;
+
+ if (sym)
+ addr_base = fold_convert (type, build_addr (sym, current_function_decl));
+ else if (base && POINTER_TYPE_P (TREE_TYPE (base)))
+ {
+ addr_base = fold_convert (type, base);
+ base = NULL_TREE;
+ }
act_elem = TMR_INDEX (mem_ref);
if (act_elem)
{
- act_elem = fold_convert (type, act_elem);
-
if (step)
- act_elem = fold_build2 (MULT_EXPR, type, act_elem,
- fold_convert (type, step));
- addr = act_elem;
+ act_elem = fold_build2 (MULT_EXPR, sizetype, act_elem, step);
+ addr_off = act_elem;
}
- act_elem = TMR_BASE (mem_ref);
+ act_elem = base;
if (act_elem)
{
- act_elem = fold_convert (type, act_elem);
-
- if (addr)
- addr = fold_build2 (PLUS_EXPR, type, addr, act_elem);
+ if (addr_off)
+ addr_off = fold_build2 (PLUS_EXPR, sizetype, addr_off, act_elem);
else
- addr = act_elem;
+ addr_off = act_elem;
}
- act_elem = TMR_SYMBOL (mem_ref);
- if (act_elem)
+ if (offset && !integer_zerop (offset))
{
- act_elem = fold_convert (type, build_addr (act_elem,
- current_function_decl));
- if (addr)
- addr = fold_build2 (PLUS_EXPR, type, addr, act_elem);
+ if (addr_off)
+ addr_off = fold_build2 (PLUS_EXPR, sizetype, addr_off, offset);
else
- addr = act_elem;
+ addr_off = offset;
}
- if (!zero_p (offset))
+ if (addr_off)
{
- act_elem = fold_convert (type, offset);
-
- if (addr)
- addr = fold_build2 (PLUS_EXPR, type, addr, act_elem);
+ if (addr_base)
+ addr = fold_build2 (POINTER_PLUS_EXPR, type, addr_base, addr_off);
else
- addr = act_elem;
+ addr = fold_convert (type, addr_off);
}
-
- if (!addr)
+ else if (addr_base)
+ addr = addr_base;
+ else
addr = build_int_cst (type, 0);
return addr;
if (addr->step && integer_onep (addr->step))
addr->step = NULL_TREE;
- if (addr->offset && zero_p (addr->offset))
+ if (addr->offset && integer_zerop (addr->offset))
addr->offset = NULL_TREE;
- return build7 (TARGET_MEM_REF, type,
+ return build6 (TARGET_MEM_REF, type,
addr->symbol, addr->base, addr->index,
- addr->step, addr->offset, NULL, NULL);
+ addr->step, addr->offset, NULL);
}
/* Returns true if OBJ is an object whose address is a link time constant. */
{
return (TREE_CODE (obj) == VAR_DECL
&& (TREE_STATIC (obj)
- || DECL_EXTERNAL (obj)));
+ || DECL_EXTERNAL (obj))
+ && ! DECL_DLLIMPORT_P (obj));
}
-/* Adds COEF * ELT to PARTS. TYPE is the type of the address we
- construct. */
+/* If ADDR contains an address of object that is a link time constant,
+ move it to PARTS->symbol. */
static void
-add_to_parts (struct mem_address *parts, tree type, tree elt,
- unsigned HOST_WIDE_INT coef)
+move_fixed_address_to_symbol (struct mem_address *parts, aff_tree *addr)
{
- /* Check if this is a symbol. */
- if (!parts->symbol
- && coef == 1
- && TREE_CODE (elt) == ADDR_EXPR
- && fixed_address_object_p (TREE_OPERAND (elt, 0)))
+ unsigned i;
+ tree val = NULL_TREE;
+
+ for (i = 0; i < addr->n; i++)
{
- parts->symbol = TREE_OPERAND (elt, 0);
- return;
+ if (!double_int_one_p (addr->elts[i].coef))
+ continue;
+
+ val = addr->elts[i].val;
+ if (TREE_CODE (val) == ADDR_EXPR
+ && fixed_address_object_p (TREE_OPERAND (val, 0)))
+ break;
}
- if (coef != 1)
- elt = fold_build2 (MULT_EXPR, type, fold_convert (type, elt),
- build_int_cst_type (type, coef));
- else
- elt = fold_convert (type, elt);
+ if (i == addr->n)
+ return;
- if (!parts->base)
+ parts->symbol = TREE_OPERAND (val, 0);
+ aff_combination_remove_elt (addr, i);
+}
+
+/* If ADDR contains an address of a dereferenced pointer, move it to
+ PARTS->base. */
+
+static void
+move_pointer_to_base (struct mem_address *parts, aff_tree *addr)
+{
+ unsigned i;
+ tree val = NULL_TREE;
+
+ for (i = 0; i < addr->n; i++)
{
- parts->base = elt;
- return;
+ if (!double_int_one_p (addr->elts[i].coef))
+ continue;
+
+ val = addr->elts[i].val;
+ if (POINTER_TYPE_P (TREE_TYPE (val)))
+ break;
}
+ if (i == addr->n)
+ return;
+
+ parts->base = val;
+ aff_combination_remove_elt (addr, i);
+}
+
+/* Adds ELT to PARTS. */
+
+static void
+add_to_parts (struct mem_address *parts, tree elt)
+{
+ tree type;
+
if (!parts->index)
{
- parts->index = elt;
+ parts->index = fold_convert (sizetype, elt);
+ return;
+ }
+
+ if (!parts->base)
+ {
+ parts->base = elt;
return;
}
/* Add ELT to base. */
- parts->base = fold_build2 (PLUS_EXPR, type, parts->base, elt);
+ type = TREE_TYPE (parts->base);
+ if (POINTER_TYPE_P (type))
+ parts->base = fold_build2 (POINTER_PLUS_EXPR, type,
+ parts->base,
+ fold_convert (sizetype, elt));
+ else
+ parts->base = fold_build2 (PLUS_EXPR, type,
+ parts->base, elt);
}
/* Finds the most expensive multiplication in ADDR that can be
expressed in an addressing mode and move the corresponding
- element(s) to PARTS. TYPE is the type of the address we
- construct. */
+ element(s) to PARTS. */
static void
-most_expensive_mult_to_index (struct mem_address *parts, tree type,
- struct affine_tree_combination *addr)
+most_expensive_mult_to_index (struct mem_address *parts, aff_tree *addr,
+ bool speed)
{
- unsigned HOST_WIDE_INT best_mult = 0;
+ HOST_WIDE_INT coef;
+ double_int best_mult, amult, amult_neg;
unsigned best_mult_cost = 0, acost;
tree mult_elt = NULL_TREE, elt;
unsigned i, j;
+ enum tree_code op_code;
+ best_mult = double_int_zero;
for (i = 0; i < addr->n; i++)
{
- if (addr->coefs[i] == 1
- || !multiplier_allowed_in_address_p (addr->coefs[i]))
+ if (!double_int_fits_in_shwi_p (addr->elts[i].coef))
+ continue;
+
+ /* FIXME: Should use the correct memory mode rather than Pmode. */
+
+ coef = double_int_to_shwi (addr->elts[i].coef);
+ if (coef == 1
+ || !multiplier_allowed_in_address_p (coef, Pmode))
continue;
-
- acost = multiply_by_cost (addr->coefs[i], Pmode);
+
+ acost = multiply_by_cost (coef, Pmode, speed);
if (acost > best_mult_cost)
{
best_mult_cost = acost;
- best_mult = addr->coefs[i];
+ best_mult = addr->elts[i].coef;
}
}
- if (!best_mult)
+ if (!best_mult_cost)
return;
+ /* Collect elements multiplied by best_mult. */
for (i = j = 0; i < addr->n; i++)
{
- if (addr->coefs[i] != best_mult)
+ amult = addr->elts[i].coef;
+ amult_neg = double_int_ext_for_comb (double_int_neg (amult), addr);
+
+ if (double_int_equal_p (amult, best_mult))
+ op_code = PLUS_EXPR;
+ else if (double_int_equal_p (amult_neg, best_mult))
+ op_code = MINUS_EXPR;
+ else
{
- addr->coefs[j] = addr->coefs[i];
addr->elts[j] = addr->elts[i];
j++;
continue;
}
- elt = fold_convert (type, addr->elts[i]);
- if (!mult_elt)
+ elt = fold_convert (sizetype, addr->elts[i].val);
+ if (mult_elt)
+ mult_elt = fold_build2 (op_code, sizetype, mult_elt, elt);
+ else if (op_code == PLUS_EXPR)
mult_elt = elt;
else
- mult_elt = fold_build2 (PLUS_EXPR, type, mult_elt, elt);
+ mult_elt = fold_build1 (NEGATE_EXPR, sizetype, elt);
}
addr->n = j;
-
+
parts->index = mult_elt;
- parts->step = build_int_cst_type (type, best_mult);
+ parts->step = double_int_to_tree (sizetype, best_mult);
}
/* Splits address ADDR into PARTS.
to PARTS. Some architectures do not support anything but single
register in address, possibly with a small integer offset; while
create_mem_ref will simplify the address to an acceptable shape
- later, it would be a small bit more efficient to know that asking
- for complicated addressing modes is useless. */
+ later, it would be more efficient to know that asking for complicated
+ addressing modes is useless. */
static void
-addr_to_parts (struct affine_tree_combination *addr, tree type,
- struct mem_address *parts)
+addr_to_parts (aff_tree *addr, struct mem_address *parts, bool speed)
{
+ tree part;
unsigned i;
parts->symbol = NULL_TREE;
parts->index = NULL_TREE;
parts->step = NULL_TREE;
- if (addr->offset)
- parts->offset = build_int_cst_type (type, addr->offset);
+ if (!double_int_zero_p (addr->offset))
+ parts->offset = double_int_to_tree (sizetype, addr->offset);
else
parts->offset = NULL_TREE;
+ /* Try to find a symbol. */
+ move_fixed_address_to_symbol (parts, addr);
+
/* First move the most expensive feasible multiplication
to index. */
- most_expensive_mult_to_index (parts, type, addr);
+ most_expensive_mult_to_index (parts, addr, speed);
+
+ /* Try to find a base of the reference. Since at the moment
+ there is no reliable way how to distinguish between pointer and its
+ offset, this is just a guess. */
+ if (!parts->symbol)
+ move_pointer_to_base (parts, addr);
/* Then try to process the remaining elements. */
for (i = 0; i < addr->n; i++)
- add_to_parts (parts, type, addr->elts[i], addr->coefs[i]);
+ {
+ part = fold_convert (sizetype, addr->elts[i].val);
+ if (!double_int_one_p (addr->elts[i].coef))
+ part = fold_build2 (MULT_EXPR, sizetype, part,
+ double_int_to_tree (sizetype, addr->elts[i].coef));
+ add_to_parts (parts, part);
+ }
if (addr->rest)
- add_to_parts (parts, type, addr->rest, 1);
+ add_to_parts (parts, fold_convert (sizetype, addr->rest));
}
/* Force the PARTS to register. */
static void
-gimplify_mem_ref_parts (block_stmt_iterator *bsi, struct mem_address *parts)
+gimplify_mem_ref_parts (gimple_stmt_iterator *gsi, struct mem_address *parts)
{
if (parts->base)
- parts->base = force_gimple_operand_bsi (bsi, parts->base,
- true, NULL_TREE);
+ parts->base = force_gimple_operand_gsi (gsi, parts->base,
+ true, NULL_TREE,
+ true, GSI_SAME_STMT);
if (parts->index)
- parts->index = force_gimple_operand_bsi (bsi, parts->index,
- true, NULL_TREE);
+ parts->index = force_gimple_operand_gsi (gsi, parts->index,
+ true, NULL_TREE,
+ true, GSI_SAME_STMT);
}
/* Creates and returns a TARGET_MEM_REF for address ADDR. If necessary
- computations are emitted in front of BSI. TYPE is the mode
+ computations are emitted in front of GSI. TYPE is the mode
of created memory reference. */
tree
-create_mem_ref (block_stmt_iterator *bsi, tree type,
- struct affine_tree_combination *addr)
+create_mem_ref (gimple_stmt_iterator *gsi, tree type, aff_tree *addr,
+ bool speed)
{
tree mem_ref, tmp;
- tree addr_type = build_pointer_type (type);
+ tree atype;
struct mem_address parts;
- addr_to_parts (addr, addr_type, &parts);
- gimplify_mem_ref_parts (bsi, &parts);
+ addr_to_parts (addr, &parts, speed);
+ gimplify_mem_ref_parts (gsi, &parts);
mem_ref = create_mem_ref_raw (type, &parts);
if (mem_ref)
return mem_ref;
{
/* Move the multiplication to index. */
gcc_assert (parts.index);
- parts.index = force_gimple_operand_bsi (bsi,
- build2 (MULT_EXPR, addr_type,
- parts.index, parts.step),
- true, NULL_TREE);
+ parts.index = force_gimple_operand_gsi (gsi,
+ fold_build2 (MULT_EXPR, sizetype,
+ parts.index, parts.step),
+ true, NULL_TREE, true, GSI_SAME_STMT);
parts.step = NULL_TREE;
mem_ref = create_mem_ref_raw (type, &parts);
if (parts.symbol)
{
tmp = build_addr (parts.symbol, current_function_decl);
+ gcc_assert (is_gimple_val (tmp));
/* Add the symbol to base, eventually forcing it to register. */
if (parts.base)
- parts.base = force_gimple_operand_bsi (bsi,
- build2 (PLUS_EXPR, addr_type,
- parts.base, tmp),
- true, NULL_TREE);
+ {
+ gcc_assert (useless_type_conversion_p
+ (sizetype, TREE_TYPE (parts.base)));
+
+ if (parts.index)
+ {
+ atype = TREE_TYPE (tmp);
+ parts.base = force_gimple_operand_gsi (gsi,
+ fold_build2 (POINTER_PLUS_EXPR, atype,
+ tmp,
+ fold_convert (sizetype, parts.base)),
+ true, NULL_TREE, true, GSI_SAME_STMT);
+ }
+ else
+ {
+ parts.index = parts.base;
+ parts.base = tmp;
+ }
+ }
else
parts.base = tmp;
parts.symbol = NULL_TREE;
return mem_ref;
}
- if (parts.base)
+ if (parts.index)
{
- /* Add base to index. */
- if (parts.index)
- parts.index = force_gimple_operand_bsi (bsi,
- build2 (PLUS_EXPR, addr_type,
- parts.base,
- parts.index),
- true, NULL_TREE);
+ /* Add index to base. */
+ if (parts.base)
+ {
+ atype = TREE_TYPE (parts.base);
+ parts.base = force_gimple_operand_gsi (gsi,
+ fold_build2 (POINTER_PLUS_EXPR, atype,
+ parts.base,
+ parts.index),
+ true, NULL_TREE, true, GSI_SAME_STMT);
+ }
else
- parts.index = parts.base;
- parts.base = NULL_TREE;
+ parts.base = parts.index;
+ parts.index = NULL_TREE;
mem_ref = create_mem_ref_raw (type, &parts);
if (mem_ref)
if (parts.offset && !integer_zerop (parts.offset))
{
- /* Try adding offset to index. */
- if (parts.index)
- parts.index = force_gimple_operand_bsi (bsi,
- build2 (PLUS_EXPR, addr_type,
- parts.index,
- parts.offset),
- true, NULL_TREE);
+ /* Try adding offset to base. */
+ if (parts.base)
+ {
+ atype = TREE_TYPE (parts.base);
+ parts.base = force_gimple_operand_gsi (gsi,
+ fold_build2 (POINTER_PLUS_EXPR, atype,
+ parts.base,
+ fold_convert (sizetype, parts.offset)),
+ true, NULL_TREE, true, GSI_SAME_STMT);
+ }
else
- parts.index = parts.offset, bsi;
+ parts.base = parts.offset;
parts.offset = NULL_TREE;
(only a register). If we cannot create such a memory reference,
something is really wrong. */
gcc_assert (parts.symbol == NULL_TREE);
- gcc_assert (parts.base == NULL_TREE);
+ gcc_assert (parts.index == NULL_TREE);
gcc_assert (!parts.step || integer_onep (parts.step));
gcc_assert (!parts.offset || integer_zerop (parts.offset));
gcc_unreachable ();
void
copy_mem_ref_info (tree to, tree from)
{
- /* Copy the annotation, to preserve the aliasing information. */
- TMR_TAG (to) = TMR_TAG (from);
-
/* And the info about the original reference. */
TMR_ORIGINAL (to) = TMR_ORIGINAL (from);
}
if (addr.base && TREE_CODE (addr.base) == INTEGER_CST)
{
if (addr.offset)
- addr.offset = fold_binary_to_constant (PLUS_EXPR, ptr_type_node,
- addr.offset, addr.base);
+ addr.offset = fold_binary_to_constant (PLUS_EXPR, sizetype,
+ addr.offset,
+ fold_convert (sizetype, addr.base));
else
addr.offset = addr.base;
off = addr.index;
if (addr.step)
{
- off = fold_binary_to_constant (MULT_EXPR, ptr_type_node,
+ off = fold_binary_to_constant (MULT_EXPR, sizetype,
off, addr.step);
addr.step = NULL_TREE;
}
if (addr.offset)
{
- addr.offset = fold_binary_to_constant (PLUS_EXPR, ptr_type_node,
+ addr.offset = fold_binary_to_constant (PLUS_EXPR, sizetype,
addr.offset, off);
}
else