diff options
Diffstat (limited to 'gcc')
-rw-r--r-- | gcc/config/aarch64/aarch64-builtins.c | 35 | ||||
-rw-r--r-- | gcc/testsuite/gcc.target/aarch64/lane-bound-1.c | 14 | ||||
-rw-r--r-- | gcc/testsuite/gcc.target/aarch64/lane-bound-2.c | 10 |
3 files changed, 59 insertions, 0 deletions
diff --git a/gcc/config/aarch64/aarch64-builtins.c b/gcc/config/aarch64/aarch64-builtins.c index eef9fc0..119f67d 100644 --- a/gcc/config/aarch64/aarch64-builtins.c +++ b/gcc/config/aarch64/aarch64-builtins.c @@ -29,6 +29,7 @@ #include "rtl.h" #include "tree.h" #include "gimple.h" +#include "ssa.h" #include "memmodel.h" #include "tm_p.h" #include "expmed.h" @@ -2333,6 +2334,27 @@ aarch64_general_builtin_rsqrt (unsigned int fn) return NULL_TREE; } +/* Return true if the lane check can be removed as there is no + error going to be emitted. */ +static bool +aarch64_fold_builtin_lane_check (tree arg0, tree arg1, tree arg2) +{ + if (TREE_CODE (arg0) != INTEGER_CST) + return false; + if (TREE_CODE (arg1) != INTEGER_CST) + return false; + if (TREE_CODE (arg2) != INTEGER_CST) + return false; + + auto totalsize = wi::to_widest (arg0); + auto elementsize = wi::to_widest (arg1); + if (totalsize == 0 || elementsize == 0) + return false; + auto lane = wi::to_widest (arg2); + auto high = wi::udiv_trunc (totalsize, elementsize); + return wi::ltu_p (lane, high); +} + #undef VAR1 #define VAR1(T, N, MAP, FLAG, A) \ case AARCH64_SIMD_BUILTIN_##T##_##N##A: @@ -2353,6 +2375,11 @@ aarch64_general_fold_builtin (unsigned int fcode, tree type, VAR1 (UNOP, floatv4si, 2, ALL, v4sf) VAR1 (UNOP, floatv2di, 2, ALL, v2df) return fold_build1 (FLOAT_EXPR, type, args[0]); + case AARCH64_SIMD_BUILTIN_LANE_CHECK: + gcc_assert (n_args == 3); + if (aarch64_fold_builtin_lane_check (args[0], args[1], args[2])) + return void_node; + break; default: break; } @@ -2440,6 +2467,14 @@ aarch64_general_gimple_fold_builtin (unsigned int fcode, gcall *stmt) } break; } + case AARCH64_SIMD_BUILTIN_LANE_CHECK: + if (aarch64_fold_builtin_lane_check (args[0], args[1], args[2])) + { + unlink_stmt_vdef (stmt); + release_defs (stmt); + new_stmt = gimple_build_nop (); + } + break; default: break; } diff --git a/gcc/testsuite/gcc.target/aarch64/lane-bound-1.c b/gcc/testsuite/gcc.target/aarch64/lane-bound-1.c new file mode 100644 index 0000000..bbbe679 --- /dev/null +++ b/gcc/testsuite/gcc.target/aarch64/lane-bound-1.c @@ -0,0 +1,14 @@ +/* { dg-do compile } */ +/* { dg-options "-O2 -fdump-tree-optimized" } */ +#include <arm_neon.h> + +void +f (float32x4_t **ptr) +{ + float32x4_t res = vsetq_lane_f32 (0.0f, **ptr, 0); + **ptr = res; +} +/* GCC should be able to remove the call to "__builtin_aarch64_im_lane_boundsi" + and optimize out the second load from *ptr. */ +/* { dg-final { scan-tree-dump-times "__builtin_aarch64_im_lane_boundsi" 0 "optimized" } } */ +/* { dg-final { scan-tree-dump-times " = \\\*ptr_" 1 "optimized" } } */ diff --git a/gcc/testsuite/gcc.target/aarch64/lane-bound-2.c b/gcc/testsuite/gcc.target/aarch64/lane-bound-2.c new file mode 100644 index 0000000..923c946 --- /dev/null +++ b/gcc/testsuite/gcc.target/aarch64/lane-bound-2.c @@ -0,0 +1,10 @@ +/* { dg-do compile } */ +/* { dg-options "-O2 -fdump-tree-original" } */ +void +f (void) +{ + __builtin_aarch64_im_lane_boundsi (16, 4, 0); + __builtin_aarch64_im_lane_boundsi (8, 8, 0); +} +/* GCC should be able to optimize these out before gimplification. */ +/* { dg-final { scan-tree-dump-times "__builtin_aarch64_im_lane_boundsi" 0 "original" } } */ |