diff options
-rw-r--r-- | gcc/ChangeLog | 9 | ||||
-rw-r--r-- | gcc/fold-const.c | 113 | ||||
-rw-r--r-- | gcc/testsuite/ChangeLog | 3 | ||||
-rw-r--r-- | gcc/testsuite/gcc.dg/fold-rotate-1.c | 74 |
4 files changed, 199 insertions, 0 deletions
diff --git a/gcc/ChangeLog b/gcc/ChangeLog index 2d405bc..4e16b73 100644 --- a/gcc/ChangeLog +++ b/gcc/ChangeLog @@ -1,5 +1,14 @@ 2007-12-03 Jakub Jelinek <jakub@redhat.com> + PR middle-end/29749 + * fold-const.c (fold_binary) <case BIT_AND_EXPR>: Optimize + (X << C1) & C2 into (X << C1) & (C2 | ((1 << C1) - 1)) + and (X >> C1) & C2 into (X >> C1) & (C2 | ~((type) -1 >> C1)). + (fold_binary) <case LSHIFT_EXPR, case RSHIFT_EXPR>: Optimize + (X & C2) << C1 into (X << C1) & (C2 << C1) and + (X & C2) >> C1 into (X >> C1) & (C2 >> C1) if that allows further + optimizations. + PR tree-optimization/33453 * tree-data-ref.c (split_constant_offset): Use POINTER_PLUS_EXPR for pointer addition. diff --git a/gcc/fold-const.c b/gcc/fold-const.c index 1c03f30..1502994 100644 --- a/gcc/fold-const.c +++ b/gcc/fold-const.c @@ -10973,6 +10973,100 @@ fold_binary (enum tree_code code, tree type, tree op0, tree op1) return build_int_cst (type, residue & low); } + /* Fold (X << C1) & C2 into (X << C1) & (C2 | ((1 << C1) - 1)) + (X >> C1) & C2 into (X >> C1) & (C2 | ~((type) -1 >> C1)) + if the new mask might be further optimized. */ + if ((TREE_CODE (arg0) == LSHIFT_EXPR + || TREE_CODE (arg0) == RSHIFT_EXPR) + && host_integerp (TREE_OPERAND (arg0, 1), 1) + && host_integerp (arg1, TYPE_UNSIGNED (TREE_TYPE (arg1))) + && tree_low_cst (TREE_OPERAND (arg0, 1), 1) + < TYPE_PRECISION (TREE_TYPE (arg0)) + && TYPE_PRECISION (TREE_TYPE (arg0)) <= HOST_BITS_PER_WIDE_INT + && tree_low_cst (TREE_OPERAND (arg0, 1), 1) > 0) + { + unsigned int shiftc = tree_low_cst (TREE_OPERAND (arg0, 1), 1); + unsigned HOST_WIDE_INT mask + = tree_low_cst (arg1, TYPE_UNSIGNED (TREE_TYPE (arg1))); + unsigned HOST_WIDE_INT newmask, zerobits = 0; + tree shift_type = TREE_TYPE (arg0); + + if (TREE_CODE (arg0) == LSHIFT_EXPR) + zerobits = ((((unsigned HOST_WIDE_INT) 1) << shiftc) - 1); + else if (TREE_CODE (arg0) == RSHIFT_EXPR + && TYPE_PRECISION (TREE_TYPE (arg0)) + == GET_MODE_BITSIZE (TYPE_MODE (TREE_TYPE (arg0)))) + { + unsigned int prec = TYPE_PRECISION (TREE_TYPE (arg0)); + tree arg00 = TREE_OPERAND (arg0, 0); + /* See if more bits can be proven as zero because of + zero extension. */ + if (TREE_CODE (arg00) == NOP_EXPR + && TYPE_UNSIGNED (TREE_TYPE (TREE_OPERAND (arg00, 0)))) + { + tree inner_type = TREE_TYPE (TREE_OPERAND (arg00, 0)); + if (TYPE_PRECISION (inner_type) + == GET_MODE_BITSIZE (TYPE_MODE (inner_type)) + && TYPE_PRECISION (inner_type) < prec) + { + prec = TYPE_PRECISION (inner_type); + /* See if we can shorten the right shift. */ + if (shiftc < prec) + shift_type = inner_type; + } + } + zerobits = ~(unsigned HOST_WIDE_INT) 0; + zerobits >>= HOST_BITS_PER_WIDE_INT - shiftc; + zerobits <<= prec - shiftc; + /* For arithmetic shift if sign bit could be set, zerobits + can contain actually sign bits, so no transformation is + possible, unless MASK masks them all away. In that + case the shift needs to be converted into logical shift. */ + if (!TYPE_UNSIGNED (TREE_TYPE (arg0)) + && prec == TYPE_PRECISION (TREE_TYPE (arg0))) + { + if ((mask & zerobits) == 0) + shift_type = unsigned_type_for (TREE_TYPE (arg0)); + else + zerobits = 0; + } + } + + /* ((X << 16) & 0xff00) is (X, 0). */ + if ((mask & zerobits) == mask) + return omit_one_operand (type, build_int_cst (type, 0), arg0); + + newmask = mask | zerobits; + if (newmask != mask && (newmask & (newmask + 1)) == 0) + { + unsigned int prec; + + /* Only do the transformation if NEWMASK is some integer + mode's mask. */ + for (prec = BITS_PER_UNIT; + prec < HOST_BITS_PER_WIDE_INT; prec <<= 1) + if (newmask == (((unsigned HOST_WIDE_INT) 1) << prec) - 1) + break; + if (prec < HOST_BITS_PER_WIDE_INT + || newmask == ~(unsigned HOST_WIDE_INT) 0) + { + if (shift_type != TREE_TYPE (arg0)) + { + tem = fold_build2 (TREE_CODE (arg0), shift_type, + fold_convert (shift_type, + TREE_OPERAND (arg0, 0)), + TREE_OPERAND (arg0, 1)); + tem = fold_convert (type, tem); + } + else + tem = op0; + return fold_build2 (BIT_AND_EXPR, type, tem, + build_int_cst_type (TREE_TYPE (op1), + newmask)); + } + } + } + goto associate; case RDIV_EXPR: @@ -11526,6 +11620,25 @@ fold_binary (enum tree_code code, tree type, tree op0, tree op1) == (unsigned int) GET_MODE_BITSIZE (TYPE_MODE (type)))) return TREE_OPERAND (arg0, 0); + /* Fold (X & C2) << C1 into (X << C1) & (C2 << C1) + (X & C2) >> C1 into (X >> C1) & (C2 >> C1) + if the latter can be further optimized. */ + if ((code == LSHIFT_EXPR || code == RSHIFT_EXPR) + && TREE_CODE (arg0) == BIT_AND_EXPR + && TREE_CODE (arg1) == INTEGER_CST + && TREE_CODE (TREE_OPERAND (arg0, 1)) == INTEGER_CST) + { + tree mask = fold_build2 (code, type, + fold_convert (type, TREE_OPERAND (arg0, 1)), + arg1); + tree shift = fold_build2 (code, type, + fold_convert (type, TREE_OPERAND (arg0, 0)), + arg1); + tem = fold_binary (BIT_AND_EXPR, type, shift, mask); + if (tem) + return tem; + } + return NULL_TREE; case MIN_EXPR: diff --git a/gcc/testsuite/ChangeLog b/gcc/testsuite/ChangeLog index dc8bb82..11a07de 100644 --- a/gcc/testsuite/ChangeLog +++ b/gcc/testsuite/ChangeLog @@ -1,5 +1,8 @@ 2007-12-03 Jakub Jelinek <jakub@redhat.com> + PR middle-end/29749 + * gcc.dg/fold-rotate-1.c: New test. + PR tree-optimization/33453 * gcc.c-torture/compile/20071203-1.c: New test. diff --git a/gcc/testsuite/gcc.dg/fold-rotate-1.c b/gcc/testsuite/gcc.dg/fold-rotate-1.c new file mode 100644 index 0000000..c04447fb --- /dev/null +++ b/gcc/testsuite/gcc.dg/fold-rotate-1.c @@ -0,0 +1,74 @@ +/* PR middle-end/29749 */ +/* { dg-do compile } */ +/* { dg-options "-O2 -fdump-tree-original" } */ + +#if __SCHAR_MAX__ == 127 + +unsigned char +e1 (unsigned char a) +{ + return a >> 5 | a << 3; +} + +unsigned char +e2 (unsigned char a) +{ + return (a & 0xe0) >> 5 | (a & 0x1f) << 3; +} + +unsigned char +e3 (unsigned char a) +{ + return ((a >> 5) & 0x07) | ((a << 3) & 0xf8); +} + +#endif + +#if __SHRT_MAX__ == 32767 + +unsigned short +f1 (unsigned short a) +{ + return a >> 8 | a << 8; +} + +unsigned short +f2 (unsigned short a) +{ + return (a & 0xff00) >> 8 | (a & 0x00ff) << 8; +} + +unsigned short +f3 (unsigned short a) +{ + return ((a >> 8) & 0x00ff) | ((a << 8) & 0xff00); +} + +#endif + +#if __INT_MAX__ == 2147483647 + +unsigned int +g1 (unsigned int a) +{ + return a >> 24 | a << 8; +} + +unsigned int +g2 (unsigned int a) +{ + return (a & 0xff000000) >> 24 | (a & 0x00ffffff) << 8; +} + +unsigned int +g3 (unsigned int a) +{ + return ((a >> 24) & 0x000000ff) | ((a << 8) & 0xffffff00U); +} + +#endif + +int i; + +/* { dg-final { scan-tree-dump-times "&" 0 "original" } } */ +/* { dg-final { cleanup-tree-dump "original" } } */ |