aboutsummaryrefslogtreecommitdiff
path: root/gcc/rtlanal.c
diff options
context:
space:
mode:
authorMartin Liska <mliska@suse.cz>2022-01-14 16:56:44 +0100
committerMartin Liska <mliska@suse.cz>2022-01-17 22:12:04 +0100
commit5c69acb32329d49e58c26fa41ae74229a52b9106 (patch)
treeddb05f9d73afb6f998457d2ac4b720e3b3b60483 /gcc/rtlanal.c
parent490e23032baaece71f2ec09fa1805064b150fbc2 (diff)
downloadgcc-5c69acb32329d49e58c26fa41ae74229a52b9106.zip
gcc-5c69acb32329d49e58c26fa41ae74229a52b9106.tar.gz
gcc-5c69acb32329d49e58c26fa41ae74229a52b9106.tar.bz2
Rename .c files to .cc files.
gcc/ada/ChangeLog: * adadecode.c: Moved to... * adadecode.cc: ...here. * affinity.c: Moved to... * affinity.cc: ...here. * argv-lynxos178-raven-cert.c: Moved to... * argv-lynxos178-raven-cert.cc: ...here. * argv.c: Moved to... * argv.cc: ...here. * aux-io.c: Moved to... * aux-io.cc: ...here. * cio.c: Moved to... * cio.cc: ...here. * cstreams.c: Moved to... * cstreams.cc: ...here. * env.c: Moved to... * env.cc: ...here. * exit.c: Moved to... * exit.cc: ...here. * expect.c: Moved to... * expect.cc: ...here. * final.c: Moved to... * final.cc: ...here. * gcc-interface/cuintp.c: Moved to... * gcc-interface/cuintp.cc: ...here. * gcc-interface/decl.c: Moved to... * gcc-interface/decl.cc: ...here. * gcc-interface/misc.c: Moved to... * gcc-interface/misc.cc: ...here. * gcc-interface/targtyps.c: Moved to... * gcc-interface/targtyps.cc: ...here. * gcc-interface/trans.c: Moved to... * gcc-interface/trans.cc: ...here. * gcc-interface/utils.c: Moved to... * gcc-interface/utils.cc: ...here. * gcc-interface/utils2.c: Moved to... * gcc-interface/utils2.cc: ...here. * init.c: Moved to... * init.cc: ...here. * initialize.c: Moved to... * initialize.cc: ...here. * libgnarl/thread.c: Moved to... * libgnarl/thread.cc: ...here. * link.c: Moved to... * link.cc: ...here. * locales.c: Moved to... * locales.cc: ...here. * mkdir.c: Moved to... * mkdir.cc: ...here. * raise.c: Moved to... * raise.cc: ...here. * rtfinal.c: Moved to... * rtfinal.cc: ...here. * rtinit.c: Moved to... * rtinit.cc: ...here. * seh_init.c: Moved to... * seh_init.cc: ...here. * sigtramp-armdroid.c: Moved to... * sigtramp-armdroid.cc: ...here. * sigtramp-ios.c: Moved to... * sigtramp-ios.cc: ...here. * sigtramp-qnx.c: Moved to... * sigtramp-qnx.cc: ...here. * sigtramp-vxworks.c: Moved to... * sigtramp-vxworks.cc: ...here. * socket.c: Moved to... * socket.cc: ...here. * tracebak.c: Moved to... * tracebak.cc: ...here. * version.c: Moved to... * version.cc: ...here. * vx_stack_info.c: Moved to... * vx_stack_info.cc: ...here. gcc/ChangeLog: * adjust-alignment.c: Moved to... * adjust-alignment.cc: ...here. * alias.c: Moved to... * alias.cc: ...here. * alloc-pool.c: Moved to... * alloc-pool.cc: ...here. * asan.c: Moved to... * asan.cc: ...here. * attribs.c: Moved to... * attribs.cc: ...here. * auto-inc-dec.c: Moved to... * auto-inc-dec.cc: ...here. * auto-profile.c: Moved to... * auto-profile.cc: ...here. * bb-reorder.c: Moved to... * bb-reorder.cc: ...here. * bitmap.c: Moved to... * bitmap.cc: ...here. * btfout.c: Moved to... * btfout.cc: ...here. * builtins.c: Moved to... * builtins.cc: ...here. * caller-save.c: Moved to... * caller-save.cc: ...here. * calls.c: Moved to... * calls.cc: ...here. * ccmp.c: Moved to... * ccmp.cc: ...here. * cfg.c: Moved to... * cfg.cc: ...here. * cfganal.c: Moved to... * cfganal.cc: ...here. * cfgbuild.c: Moved to... * cfgbuild.cc: ...here. * cfgcleanup.c: Moved to... * cfgcleanup.cc: ...here. * cfgexpand.c: Moved to... * cfgexpand.cc: ...here. * cfghooks.c: Moved to... * cfghooks.cc: ...here. * cfgloop.c: Moved to... * cfgloop.cc: ...here. * cfgloopanal.c: Moved to... * cfgloopanal.cc: ...here. * cfgloopmanip.c: Moved to... * cfgloopmanip.cc: ...here. * cfgrtl.c: Moved to... * cfgrtl.cc: ...here. * cgraph.c: Moved to... * cgraph.cc: ...here. * cgraphbuild.c: Moved to... * cgraphbuild.cc: ...here. * cgraphclones.c: Moved to... * cgraphclones.cc: ...here. * cgraphunit.c: Moved to... * cgraphunit.cc: ...here. * collect-utils.c: Moved to... * collect-utils.cc: ...here. * collect2-aix.c: Moved to... * collect2-aix.cc: ...here. * collect2.c: Moved to... * collect2.cc: ...here. * combine-stack-adj.c: Moved to... * combine-stack-adj.cc: ...here. * combine.c: Moved to... * combine.cc: ...here. * common/common-targhooks.c: Moved to... * common/common-targhooks.cc: ...here. * common/config/aarch64/aarch64-common.c: Moved to... * common/config/aarch64/aarch64-common.cc: ...here. * common/config/alpha/alpha-common.c: Moved to... * common/config/alpha/alpha-common.cc: ...here. * common/config/arc/arc-common.c: Moved to... * common/config/arc/arc-common.cc: ...here. * common/config/arm/arm-common.c: Moved to... * common/config/arm/arm-common.cc: ...here. * common/config/avr/avr-common.c: Moved to... * common/config/avr/avr-common.cc: ...here. * common/config/bfin/bfin-common.c: Moved to... * common/config/bfin/bfin-common.cc: ...here. * common/config/bpf/bpf-common.c: Moved to... * common/config/bpf/bpf-common.cc: ...here. * common/config/c6x/c6x-common.c: Moved to... * common/config/c6x/c6x-common.cc: ...here. * common/config/cr16/cr16-common.c: Moved to... * common/config/cr16/cr16-common.cc: ...here. * common/config/cris/cris-common.c: Moved to... * common/config/cris/cris-common.cc: ...here. * common/config/csky/csky-common.c: Moved to... * common/config/csky/csky-common.cc: ...here. * common/config/default-common.c: Moved to... * common/config/default-common.cc: ...here. * common/config/epiphany/epiphany-common.c: Moved to... * common/config/epiphany/epiphany-common.cc: ...here. * common/config/fr30/fr30-common.c: Moved to... * common/config/fr30/fr30-common.cc: ...here. * common/config/frv/frv-common.c: Moved to... * common/config/frv/frv-common.cc: ...here. * common/config/gcn/gcn-common.c: Moved to... * common/config/gcn/gcn-common.cc: ...here. * common/config/h8300/h8300-common.c: Moved to... * common/config/h8300/h8300-common.cc: ...here. * common/config/i386/i386-common.c: Moved to... * common/config/i386/i386-common.cc: ...here. * common/config/ia64/ia64-common.c: Moved to... * common/config/ia64/ia64-common.cc: ...here. * common/config/iq2000/iq2000-common.c: Moved to... * common/config/iq2000/iq2000-common.cc: ...here. * common/config/lm32/lm32-common.c: Moved to... * common/config/lm32/lm32-common.cc: ...here. * common/config/m32r/m32r-common.c: Moved to... * common/config/m32r/m32r-common.cc: ...here. * common/config/m68k/m68k-common.c: Moved to... * common/config/m68k/m68k-common.cc: ...here. * common/config/mcore/mcore-common.c: Moved to... * common/config/mcore/mcore-common.cc: ...here. * common/config/microblaze/microblaze-common.c: Moved to... * common/config/microblaze/microblaze-common.cc: ...here. * common/config/mips/mips-common.c: Moved to... * common/config/mips/mips-common.cc: ...here. * common/config/mmix/mmix-common.c: Moved to... * common/config/mmix/mmix-common.cc: ...here. * common/config/mn10300/mn10300-common.c: Moved to... * common/config/mn10300/mn10300-common.cc: ...here. * common/config/msp430/msp430-common.c: Moved to... * common/config/msp430/msp430-common.cc: ...here. * common/config/nds32/nds32-common.c: Moved to... * common/config/nds32/nds32-common.cc: ...here. * common/config/nios2/nios2-common.c: Moved to... * common/config/nios2/nios2-common.cc: ...here. * common/config/nvptx/nvptx-common.c: Moved to... * common/config/nvptx/nvptx-common.cc: ...here. * common/config/or1k/or1k-common.c: Moved to... * common/config/or1k/or1k-common.cc: ...here. * common/config/pa/pa-common.c: Moved to... * common/config/pa/pa-common.cc: ...here. * common/config/pdp11/pdp11-common.c: Moved to... * common/config/pdp11/pdp11-common.cc: ...here. * common/config/pru/pru-common.c: Moved to... * common/config/pru/pru-common.cc: ...here. * common/config/riscv/riscv-common.c: Moved to... * common/config/riscv/riscv-common.cc: ...here. * common/config/rs6000/rs6000-common.c: Moved to... * common/config/rs6000/rs6000-common.cc: ...here. * common/config/rx/rx-common.c: Moved to... * common/config/rx/rx-common.cc: ...here. * common/config/s390/s390-common.c: Moved to... * common/config/s390/s390-common.cc: ...here. * common/config/sh/sh-common.c: Moved to... * common/config/sh/sh-common.cc: ...here. * common/config/sparc/sparc-common.c: Moved to... * common/config/sparc/sparc-common.cc: ...here. * common/config/tilegx/tilegx-common.c: Moved to... * common/config/tilegx/tilegx-common.cc: ...here. * common/config/tilepro/tilepro-common.c: Moved to... * common/config/tilepro/tilepro-common.cc: ...here. * common/config/v850/v850-common.c: Moved to... * common/config/v850/v850-common.cc: ...here. * common/config/vax/vax-common.c: Moved to... * common/config/vax/vax-common.cc: ...here. * common/config/visium/visium-common.c: Moved to... * common/config/visium/visium-common.cc: ...here. * common/config/xstormy16/xstormy16-common.c: Moved to... * common/config/xstormy16/xstormy16-common.cc: ...here. * common/config/xtensa/xtensa-common.c: Moved to... * common/config/xtensa/xtensa-common.cc: ...here. * compare-elim.c: Moved to... * compare-elim.cc: ...here. * config/aarch64/aarch64-bti-insert.c: Moved to... * config/aarch64/aarch64-bti-insert.cc: ...here. * config/aarch64/aarch64-builtins.c: Moved to... * config/aarch64/aarch64-builtins.cc: ...here. * config/aarch64/aarch64-c.c: Moved to... * config/aarch64/aarch64-c.cc: ...here. * config/aarch64/aarch64-d.c: Moved to... * config/aarch64/aarch64-d.cc: ...here. * config/aarch64/aarch64.c: Moved to... * config/aarch64/aarch64.cc: ...here. * config/aarch64/cortex-a57-fma-steering.c: Moved to... * config/aarch64/cortex-a57-fma-steering.cc: ...here. * config/aarch64/driver-aarch64.c: Moved to... * config/aarch64/driver-aarch64.cc: ...here. * config/aarch64/falkor-tag-collision-avoidance.c: Moved to... * config/aarch64/falkor-tag-collision-avoidance.cc: ...here. * config/aarch64/host-aarch64-darwin.c: Moved to... * config/aarch64/host-aarch64-darwin.cc: ...here. * config/alpha/alpha.c: Moved to... * config/alpha/alpha.cc: ...here. * config/alpha/driver-alpha.c: Moved to... * config/alpha/driver-alpha.cc: ...here. * config/arc/arc-c.c: Moved to... * config/arc/arc-c.cc: ...here. * config/arc/arc.c: Moved to... * config/arc/arc.cc: ...here. * config/arc/driver-arc.c: Moved to... * config/arc/driver-arc.cc: ...here. * config/arm/aarch-common.c: Moved to... * config/arm/aarch-common.cc: ...here. * config/arm/arm-builtins.c: Moved to... * config/arm/arm-builtins.cc: ...here. * config/arm/arm-c.c: Moved to... * config/arm/arm-c.cc: ...here. * config/arm/arm-d.c: Moved to... * config/arm/arm-d.cc: ...here. * config/arm/arm.c: Moved to... * config/arm/arm.cc: ...here. * config/arm/driver-arm.c: Moved to... * config/arm/driver-arm.cc: ...here. * config/avr/avr-c.c: Moved to... * config/avr/avr-c.cc: ...here. * config/avr/avr-devices.c: Moved to... * config/avr/avr-devices.cc: ...here. * config/avr/avr-log.c: Moved to... * config/avr/avr-log.cc: ...here. * config/avr/avr.c: Moved to... * config/avr/avr.cc: ...here. * config/avr/driver-avr.c: Moved to... * config/avr/driver-avr.cc: ...here. * config/avr/gen-avr-mmcu-specs.c: Moved to... * config/avr/gen-avr-mmcu-specs.cc: ...here. * config/avr/gen-avr-mmcu-texi.c: Moved to... * config/avr/gen-avr-mmcu-texi.cc: ...here. * config/bfin/bfin.c: Moved to... * config/bfin/bfin.cc: ...here. * config/bpf/bpf.c: Moved to... * config/bpf/bpf.cc: ...here. * config/bpf/coreout.c: Moved to... * config/bpf/coreout.cc: ...here. * config/c6x/c6x.c: Moved to... * config/c6x/c6x.cc: ...here. * config/cr16/cr16.c: Moved to... * config/cr16/cr16.cc: ...here. * config/cris/cris.c: Moved to... * config/cris/cris.cc: ...here. * config/csky/csky.c: Moved to... * config/csky/csky.cc: ...here. * config/darwin-c.c: Moved to... * config/darwin-c.cc: ...here. * config/darwin-d.c: Moved to... * config/darwin-d.cc: ...here. * config/darwin-driver.c: Moved to... * config/darwin-driver.cc: ...here. * config/darwin-f.c: Moved to... * config/darwin-f.cc: ...here. * config/darwin.c: Moved to... * config/darwin.cc: ...here. * config/default-c.c: Moved to... * config/default-c.cc: ...here. * config/default-d.c: Moved to... * config/default-d.cc: ...here. * config/dragonfly-d.c: Moved to... * config/dragonfly-d.cc: ...here. * config/epiphany/epiphany.c: Moved to... * config/epiphany/epiphany.cc: ...here. * config/epiphany/mode-switch-use.c: Moved to... * config/epiphany/mode-switch-use.cc: ...here. * config/epiphany/resolve-sw-modes.c: Moved to... * config/epiphany/resolve-sw-modes.cc: ...here. * config/fr30/fr30.c: Moved to... * config/fr30/fr30.cc: ...here. * config/freebsd-d.c: Moved to... * config/freebsd-d.cc: ...here. * config/frv/frv.c: Moved to... * config/frv/frv.cc: ...here. * config/ft32/ft32.c: Moved to... * config/ft32/ft32.cc: ...here. * config/gcn/driver-gcn.c: Moved to... * config/gcn/driver-gcn.cc: ...here. * config/gcn/gcn-run.c: Moved to... * config/gcn/gcn-run.cc: ...here. * config/gcn/gcn-tree.c: Moved to... * config/gcn/gcn-tree.cc: ...here. * config/gcn/gcn.c: Moved to... * config/gcn/gcn.cc: ...here. * config/gcn/mkoffload.c: Moved to... * config/gcn/mkoffload.cc: ...here. * config/glibc-c.c: Moved to... * config/glibc-c.cc: ...here. * config/glibc-d.c: Moved to... * config/glibc-d.cc: ...here. * config/h8300/h8300.c: Moved to... * config/h8300/h8300.cc: ...here. * config/host-darwin.c: Moved to... * config/host-darwin.cc: ...here. * config/host-hpux.c: Moved to... * config/host-hpux.cc: ...here. * config/host-linux.c: Moved to... * config/host-linux.cc: ...here. * config/host-netbsd.c: Moved to... * config/host-netbsd.cc: ...here. * config/host-openbsd.c: Moved to... * config/host-openbsd.cc: ...here. * config/host-solaris.c: Moved to... * config/host-solaris.cc: ...here. * config/i386/djgpp.c: Moved to... * config/i386/djgpp.cc: ...here. * config/i386/driver-i386.c: Moved to... * config/i386/driver-i386.cc: ...here. * config/i386/driver-mingw32.c: Moved to... * config/i386/driver-mingw32.cc: ...here. * config/i386/gnu-property.c: Moved to... * config/i386/gnu-property.cc: ...here. * config/i386/host-cygwin.c: Moved to... * config/i386/host-cygwin.cc: ...here. * config/i386/host-i386-darwin.c: Moved to... * config/i386/host-i386-darwin.cc: ...here. * config/i386/host-mingw32.c: Moved to... * config/i386/host-mingw32.cc: ...here. * config/i386/i386-builtins.c: Moved to... * config/i386/i386-builtins.cc: ...here. * config/i386/i386-c.c: Moved to... * config/i386/i386-c.cc: ...here. * config/i386/i386-d.c: Moved to... * config/i386/i386-d.cc: ...here. * config/i386/i386-expand.c: Moved to... * config/i386/i386-expand.cc: ...here. * config/i386/i386-features.c: Moved to... * config/i386/i386-features.cc: ...here. * config/i386/i386-options.c: Moved to... * config/i386/i386-options.cc: ...here. * config/i386/i386.c: Moved to... * config/i386/i386.cc: ...here. * config/i386/intelmic-mkoffload.c: Moved to... * config/i386/intelmic-mkoffload.cc: ...here. * config/i386/msformat-c.c: Moved to... * config/i386/msformat-c.cc: ...here. * config/i386/winnt-cxx.c: Moved to... * config/i386/winnt-cxx.cc: ...here. * config/i386/winnt-d.c: Moved to... * config/i386/winnt-d.cc: ...here. * config/i386/winnt-stubs.c: Moved to... * config/i386/winnt-stubs.cc: ...here. * config/i386/winnt.c: Moved to... * config/i386/winnt.cc: ...here. * config/i386/x86-tune-sched-atom.c: Moved to... * config/i386/x86-tune-sched-atom.cc: ...here. * config/i386/x86-tune-sched-bd.c: Moved to... * config/i386/x86-tune-sched-bd.cc: ...here. * config/i386/x86-tune-sched-core.c: Moved to... * config/i386/x86-tune-sched-core.cc: ...here. * config/i386/x86-tune-sched.c: Moved to... * config/i386/x86-tune-sched.cc: ...here. * config/ia64/ia64-c.c: Moved to... * config/ia64/ia64-c.cc: ...here. * config/ia64/ia64.c: Moved to... * config/ia64/ia64.cc: ...here. * config/iq2000/iq2000.c: Moved to... * config/iq2000/iq2000.cc: ...here. * config/linux.c: Moved to... * config/linux.cc: ...here. * config/lm32/lm32.c: Moved to... * config/lm32/lm32.cc: ...here. * config/m32c/m32c-pragma.c: Moved to... * config/m32c/m32c-pragma.cc: ...here. * config/m32c/m32c.c: Moved to... * config/m32c/m32c.cc: ...here. * config/m32r/m32r.c: Moved to... * config/m32r/m32r.cc: ...here. * config/m68k/m68k.c: Moved to... * config/m68k/m68k.cc: ...here. * config/mcore/mcore.c: Moved to... * config/mcore/mcore.cc: ...here. * config/microblaze/microblaze-c.c: Moved to... * config/microblaze/microblaze-c.cc: ...here. * config/microblaze/microblaze.c: Moved to... * config/microblaze/microblaze.cc: ...here. * config/mips/driver-native.c: Moved to... * config/mips/driver-native.cc: ...here. * config/mips/frame-header-opt.c: Moved to... * config/mips/frame-header-opt.cc: ...here. * config/mips/mips-d.c: Moved to... * config/mips/mips-d.cc: ...here. * config/mips/mips.c: Moved to... * config/mips/mips.cc: ...here. * config/mmix/mmix.c: Moved to... * config/mmix/mmix.cc: ...here. * config/mn10300/mn10300.c: Moved to... * config/mn10300/mn10300.cc: ...here. * config/moxie/moxie.c: Moved to... * config/moxie/moxie.cc: ...here. * config/msp430/driver-msp430.c: Moved to... * config/msp430/driver-msp430.cc: ...here. * config/msp430/msp430-c.c: Moved to... * config/msp430/msp430-c.cc: ...here. * config/msp430/msp430-devices.c: Moved to... * config/msp430/msp430-devices.cc: ...here. * config/msp430/msp430.c: Moved to... * config/msp430/msp430.cc: ...here. * config/nds32/nds32-cost.c: Moved to... * config/nds32/nds32-cost.cc: ...here. * config/nds32/nds32-fp-as-gp.c: Moved to... * config/nds32/nds32-fp-as-gp.cc: ...here. * config/nds32/nds32-intrinsic.c: Moved to... * config/nds32/nds32-intrinsic.cc: ...here. * config/nds32/nds32-isr.c: Moved to... * config/nds32/nds32-isr.cc: ...here. * config/nds32/nds32-md-auxiliary.c: Moved to... * config/nds32/nds32-md-auxiliary.cc: ...here. * config/nds32/nds32-memory-manipulation.c: Moved to... * config/nds32/nds32-memory-manipulation.cc: ...here. * config/nds32/nds32-pipelines-auxiliary.c: Moved to... * config/nds32/nds32-pipelines-auxiliary.cc: ...here. * config/nds32/nds32-predicates.c: Moved to... * config/nds32/nds32-predicates.cc: ...here. * config/nds32/nds32-relax-opt.c: Moved to... * config/nds32/nds32-relax-opt.cc: ...here. * config/nds32/nds32-utils.c: Moved to... * config/nds32/nds32-utils.cc: ...here. * config/nds32/nds32.c: Moved to... * config/nds32/nds32.cc: ...here. * config/netbsd-d.c: Moved to... * config/netbsd-d.cc: ...here. * config/netbsd.c: Moved to... * config/netbsd.cc: ...here. * config/nios2/nios2.c: Moved to... * config/nios2/nios2.cc: ...here. * config/nvptx/mkoffload.c: Moved to... * config/nvptx/mkoffload.cc: ...here. * config/nvptx/nvptx-c.c: Moved to... * config/nvptx/nvptx-c.cc: ...here. * config/nvptx/nvptx.c: Moved to... * config/nvptx/nvptx.cc: ...here. * config/openbsd-d.c: Moved to... * config/openbsd-d.cc: ...here. * config/or1k/or1k.c: Moved to... * config/or1k/or1k.cc: ...here. * config/pa/pa-d.c: Moved to... * config/pa/pa-d.cc: ...here. * config/pa/pa.c: Moved to... * config/pa/pa.cc: ...here. * config/pdp11/pdp11.c: Moved to... * config/pdp11/pdp11.cc: ...here. * config/pru/pru-passes.c: Moved to... * config/pru/pru-passes.cc: ...here. * config/pru/pru-pragma.c: Moved to... * config/pru/pru-pragma.cc: ...here. * config/pru/pru.c: Moved to... * config/pru/pru.cc: ...here. * config/riscv/riscv-builtins.c: Moved to... * config/riscv/riscv-builtins.cc: ...here. * config/riscv/riscv-c.c: Moved to... * config/riscv/riscv-c.cc: ...here. * config/riscv/riscv-d.c: Moved to... * config/riscv/riscv-d.cc: ...here. * config/riscv/riscv-shorten-memrefs.c: Moved to... * config/riscv/riscv-shorten-memrefs.cc: ...here. * config/riscv/riscv-sr.c: Moved to... * config/riscv/riscv-sr.cc: ...here. * config/riscv/riscv.c: Moved to... * config/riscv/riscv.cc: ...here. * config/rl78/rl78-c.c: Moved to... * config/rl78/rl78-c.cc: ...here. * config/rl78/rl78.c: Moved to... * config/rl78/rl78.cc: ...here. * config/rs6000/driver-rs6000.c: Moved to... * config/rs6000/driver-rs6000.cc: ...here. * config/rs6000/host-darwin.c: Moved to... * config/rs6000/host-darwin.cc: ...here. * config/rs6000/host-ppc64-darwin.c: Moved to... * config/rs6000/host-ppc64-darwin.cc: ...here. * config/rs6000/rbtree.c: Moved to... * config/rs6000/rbtree.cc: ...here. * config/rs6000/rs6000-c.c: Moved to... * config/rs6000/rs6000-c.cc: ...here. * config/rs6000/rs6000-call.c: Moved to... * config/rs6000/rs6000-call.cc: ...here. * config/rs6000/rs6000-d.c: Moved to... * config/rs6000/rs6000-d.cc: ...here. * config/rs6000/rs6000-gen-builtins.c: Moved to... * config/rs6000/rs6000-gen-builtins.cc: ...here. * config/rs6000/rs6000-linux.c: Moved to... * config/rs6000/rs6000-linux.cc: ...here. * config/rs6000/rs6000-logue.c: Moved to... * config/rs6000/rs6000-logue.cc: ...here. * config/rs6000/rs6000-p8swap.c: Moved to... * config/rs6000/rs6000-p8swap.cc: ...here. * config/rs6000/rs6000-pcrel-opt.c: Moved to... * config/rs6000/rs6000-pcrel-opt.cc: ...here. * config/rs6000/rs6000-string.c: Moved to... * config/rs6000/rs6000-string.cc: ...here. * config/rs6000/rs6000.c: Moved to... * config/rs6000/rs6000.cc: ...here. * config/rx/rx.c: Moved to... * config/rx/rx.cc: ...here. * config/s390/driver-native.c: Moved to... * config/s390/driver-native.cc: ...here. * config/s390/s390-c.c: Moved to... * config/s390/s390-c.cc: ...here. * config/s390/s390-d.c: Moved to... * config/s390/s390-d.cc: ...here. * config/s390/s390.c: Moved to... * config/s390/s390.cc: ...here. * config/sh/divtab-sh4-300.c: Moved to... * config/sh/divtab-sh4-300.cc: ...here. * config/sh/divtab-sh4.c: Moved to... * config/sh/divtab-sh4.cc: ...here. * config/sh/divtab.c: Moved to... * config/sh/divtab.cc: ...here. * config/sh/sh-c.c: Moved to... * config/sh/sh-c.cc: ...here. * config/sh/sh.c: Moved to... * config/sh/sh.cc: ...here. * config/sol2-c.c: Moved to... * config/sol2-c.cc: ...here. * config/sol2-cxx.c: Moved to... * config/sol2-cxx.cc: ...here. * config/sol2-d.c: Moved to... * config/sol2-d.cc: ...here. * config/sol2-stubs.c: Moved to... * config/sol2-stubs.cc: ...here. * config/sol2.c: Moved to... * config/sol2.cc: ...here. * config/sparc/driver-sparc.c: Moved to... * config/sparc/driver-sparc.cc: ...here. * config/sparc/sparc-c.c: Moved to... * config/sparc/sparc-c.cc: ...here. * config/sparc/sparc-d.c: Moved to... * config/sparc/sparc-d.cc: ...here. * config/sparc/sparc.c: Moved to... * config/sparc/sparc.cc: ...here. * config/stormy16/stormy16.c: Moved to... * config/stormy16/stormy16.cc: ...here. * config/tilegx/mul-tables.c: Moved to... * config/tilegx/mul-tables.cc: ...here. * config/tilegx/tilegx-c.c: Moved to... * config/tilegx/tilegx-c.cc: ...here. * config/tilegx/tilegx.c: Moved to... * config/tilegx/tilegx.cc: ...here. * config/tilepro/mul-tables.c: Moved to... * config/tilepro/mul-tables.cc: ...here. * config/tilepro/tilepro-c.c: Moved to... * config/tilepro/tilepro-c.cc: ...here. * config/tilepro/tilepro.c: Moved to... * config/tilepro/tilepro.cc: ...here. * config/v850/v850-c.c: Moved to... * config/v850/v850-c.cc: ...here. * config/v850/v850.c: Moved to... * config/v850/v850.cc: ...here. * config/vax/vax.c: Moved to... * config/vax/vax.cc: ...here. * config/visium/visium.c: Moved to... * config/visium/visium.cc: ...here. * config/vms/vms-c.c: Moved to... * config/vms/vms-c.cc: ...here. * config/vms/vms-f.c: Moved to... * config/vms/vms-f.cc: ...here. * config/vms/vms.c: Moved to... * config/vms/vms.cc: ...here. * config/vxworks-c.c: Moved to... * config/vxworks-c.cc: ...here. * config/vxworks.c: Moved to... * config/vxworks.cc: ...here. * config/winnt-c.c: Moved to... * config/winnt-c.cc: ...here. * config/xtensa/xtensa.c: Moved to... * config/xtensa/xtensa.cc: ...here. * context.c: Moved to... * context.cc: ...here. * convert.c: Moved to... * convert.cc: ...here. * coverage.c: Moved to... * coverage.cc: ...here. * cppbuiltin.c: Moved to... * cppbuiltin.cc: ...here. * cppdefault.c: Moved to... * cppdefault.cc: ...here. * cprop.c: Moved to... * cprop.cc: ...here. * cse.c: Moved to... * cse.cc: ...here. * cselib.c: Moved to... * cselib.cc: ...here. * ctfc.c: Moved to... * ctfc.cc: ...here. * ctfout.c: Moved to... * ctfout.cc: ...here. * data-streamer-in.c: Moved to... * data-streamer-in.cc: ...here. * data-streamer-out.c: Moved to... * data-streamer-out.cc: ...here. * data-streamer.c: Moved to... * data-streamer.cc: ...here. * dbgcnt.c: Moved to... * dbgcnt.cc: ...here. * dbxout.c: Moved to... * dbxout.cc: ...here. * dce.c: Moved to... * dce.cc: ...here. * ddg.c: Moved to... * ddg.cc: ...here. * debug.c: Moved to... * debug.cc: ...here. * df-core.c: Moved to... * df-core.cc: ...here. * df-problems.c: Moved to... * df-problems.cc: ...here. * df-scan.c: Moved to... * df-scan.cc: ...here. * dfp.c: Moved to... * dfp.cc: ...here. * diagnostic-color.c: Moved to... * diagnostic-color.cc: ...here. * diagnostic-show-locus.c: Moved to... * diagnostic-show-locus.cc: ...here. * diagnostic-spec.c: Moved to... * diagnostic-spec.cc: ...here. * diagnostic.c: Moved to... * diagnostic.cc: ...here. * dojump.c: Moved to... * dojump.cc: ...here. * dominance.c: Moved to... * dominance.cc: ...here. * domwalk.c: Moved to... * domwalk.cc: ...here. * double-int.c: Moved to... * double-int.cc: ...here. * dse.c: Moved to... * dse.cc: ...here. * dumpfile.c: Moved to... * dumpfile.cc: ...here. * dwarf2asm.c: Moved to... * dwarf2asm.cc: ...here. * dwarf2cfi.c: Moved to... * dwarf2cfi.cc: ...here. * dwarf2ctf.c: Moved to... * dwarf2ctf.cc: ...here. * dwarf2out.c: Moved to... * dwarf2out.cc: ...here. * early-remat.c: Moved to... * early-remat.cc: ...here. * edit-context.c: Moved to... * edit-context.cc: ...here. * emit-rtl.c: Moved to... * emit-rtl.cc: ...here. * errors.c: Moved to... * errors.cc: ...here. * et-forest.c: Moved to... * et-forest.cc: ...here. * except.c: Moved to... * except.cc: ...here. * explow.c: Moved to... * explow.cc: ...here. * expmed.c: Moved to... * expmed.cc: ...here. * expr.c: Moved to... * expr.cc: ...here. * fibonacci_heap.c: Moved to... * fibonacci_heap.cc: ...here. * file-find.c: Moved to... * file-find.cc: ...here. * file-prefix-map.c: Moved to... * file-prefix-map.cc: ...here. * final.c: Moved to... * final.cc: ...here. * fixed-value.c: Moved to... * fixed-value.cc: ...here. * fold-const-call.c: Moved to... * fold-const-call.cc: ...here. * fold-const.c: Moved to... * fold-const.cc: ...here. * fp-test.c: Moved to... * fp-test.cc: ...here. * function-tests.c: Moved to... * function-tests.cc: ...here. * function.c: Moved to... * function.cc: ...here. * fwprop.c: Moved to... * fwprop.cc: ...here. * gcc-ar.c: Moved to... * gcc-ar.cc: ...here. * gcc-main.c: Moved to... * gcc-main.cc: ...here. * gcc-rich-location.c: Moved to... * gcc-rich-location.cc: ...here. * gcc.c: Moved to... * gcc.cc: ...here. * gcov-dump.c: Moved to... * gcov-dump.cc: ...here. * gcov-io.c: Moved to... * gcov-io.cc: ...here. * gcov-tool.c: Moved to... * gcov-tool.cc: ...here. * gcov.c: Moved to... * gcov.cc: ...here. * gcse-common.c: Moved to... * gcse-common.cc: ...here. * gcse.c: Moved to... * gcse.cc: ...here. * genattr-common.c: Moved to... * genattr-common.cc: ...here. * genattr.c: Moved to... * genattr.cc: ...here. * genattrtab.c: Moved to... * genattrtab.cc: ...here. * genautomata.c: Moved to... * genautomata.cc: ...here. * gencfn-macros.c: Moved to... * gencfn-macros.cc: ...here. * gencheck.c: Moved to... * gencheck.cc: ...here. * genchecksum.c: Moved to... * genchecksum.cc: ...here. * gencodes.c: Moved to... * gencodes.cc: ...here. * genconditions.c: Moved to... * genconditions.cc: ...here. * genconfig.c: Moved to... * genconfig.cc: ...here. * genconstants.c: Moved to... * genconstants.cc: ...here. * genemit.c: Moved to... * genemit.cc: ...here. * genenums.c: Moved to... * genenums.cc: ...here. * generic-match-head.c: Moved to... * generic-match-head.cc: ...here. * genextract.c: Moved to... * genextract.cc: ...here. * genflags.c: Moved to... * genflags.cc: ...here. * gengenrtl.c: Moved to... * gengenrtl.cc: ...here. * gengtype-parse.c: Moved to... * gengtype-parse.cc: ...here. * gengtype-state.c: Moved to... * gengtype-state.cc: ...here. * gengtype.c: Moved to... * gengtype.cc: ...here. * genhooks.c: Moved to... * genhooks.cc: ...here. * genmatch.c: Moved to... * genmatch.cc: ...here. * genmddeps.c: Moved to... * genmddeps.cc: ...here. * genmddump.c: Moved to... * genmddump.cc: ...here. * genmodes.c: Moved to... * genmodes.cc: ...here. * genopinit.c: Moved to... * genopinit.cc: ...here. * genoutput.c: Moved to... * genoutput.cc: ...here. * genpeep.c: Moved to... * genpeep.cc: ...here. * genpreds.c: Moved to... * genpreds.cc: ...here. * genrecog.c: Moved to... * genrecog.cc: ...here. * gensupport.c: Moved to... * gensupport.cc: ...here. * gentarget-def.c: Moved to... * gentarget-def.cc: ...here. * genversion.c: Moved to... * genversion.cc: ...here. * ggc-common.c: Moved to... * ggc-common.cc: ...here. * ggc-none.c: Moved to... * ggc-none.cc: ...here. * ggc-page.c: Moved to... * ggc-page.cc: ...here. * ggc-tests.c: Moved to... * ggc-tests.cc: ...here. * gimple-builder.c: Moved to... * gimple-builder.cc: ...here. * gimple-expr.c: Moved to... * gimple-expr.cc: ...here. * gimple-fold.c: Moved to... * gimple-fold.cc: ...here. * gimple-iterator.c: Moved to... * gimple-iterator.cc: ...here. * gimple-laddress.c: Moved to... * gimple-laddress.cc: ...here. * gimple-loop-jam.c: Moved to... * gimple-loop-jam.cc: ...here. * gimple-low.c: Moved to... * gimple-low.cc: ...here. * gimple-match-head.c: Moved to... * gimple-match-head.cc: ...here. * gimple-pretty-print.c: Moved to... * gimple-pretty-print.cc: ...here. * gimple-ssa-backprop.c: Moved to... * gimple-ssa-backprop.cc: ...here. * gimple-ssa-evrp-analyze.c: Moved to... * gimple-ssa-evrp-analyze.cc: ...here. * gimple-ssa-evrp.c: Moved to... * gimple-ssa-evrp.cc: ...here. * gimple-ssa-isolate-paths.c: Moved to... * gimple-ssa-isolate-paths.cc: ...here. * gimple-ssa-nonnull-compare.c: Moved to... * gimple-ssa-nonnull-compare.cc: ...here. * gimple-ssa-split-paths.c: Moved to... * gimple-ssa-split-paths.cc: ...here. * gimple-ssa-sprintf.c: Moved to... * gimple-ssa-sprintf.cc: ...here. * gimple-ssa-store-merging.c: Moved to... * gimple-ssa-store-merging.cc: ...here. * gimple-ssa-strength-reduction.c: Moved to... * gimple-ssa-strength-reduction.cc: ...here. * gimple-ssa-warn-alloca.c: Moved to... * gimple-ssa-warn-alloca.cc: ...here. * gimple-ssa-warn-restrict.c: Moved to... * gimple-ssa-warn-restrict.cc: ...here. * gimple-streamer-in.c: Moved to... * gimple-streamer-in.cc: ...here. * gimple-streamer-out.c: Moved to... * gimple-streamer-out.cc: ...here. * gimple-walk.c: Moved to... * gimple-walk.cc: ...here. * gimple-warn-recursion.c: Moved to... * gimple-warn-recursion.cc: ...here. * gimple.c: Moved to... * gimple.cc: ...here. * gimplify-me.c: Moved to... * gimplify-me.cc: ...here. * gimplify.c: Moved to... * gimplify.cc: ...here. * godump.c: Moved to... * godump.cc: ...here. * graph.c: Moved to... * graph.cc: ...here. * graphds.c: Moved to... * graphds.cc: ...here. * graphite-dependences.c: Moved to... * graphite-dependences.cc: ...here. * graphite-isl-ast-to-gimple.c: Moved to... * graphite-isl-ast-to-gimple.cc: ...here. * graphite-optimize-isl.c: Moved to... * graphite-optimize-isl.cc: ...here. * graphite-poly.c: Moved to... * graphite-poly.cc: ...here. * graphite-scop-detection.c: Moved to... * graphite-scop-detection.cc: ...here. * graphite-sese-to-poly.c: Moved to... * graphite-sese-to-poly.cc: ...here. * graphite.c: Moved to... * graphite.cc: ...here. * haifa-sched.c: Moved to... * haifa-sched.cc: ...here. * hash-map-tests.c: Moved to... * hash-map-tests.cc: ...here. * hash-set-tests.c: Moved to... * hash-set-tests.cc: ...here. * hash-table.c: Moved to... * hash-table.cc: ...here. * hooks.c: Moved to... * hooks.cc: ...here. * host-default.c: Moved to... * host-default.cc: ...here. * hw-doloop.c: Moved to... * hw-doloop.cc: ...here. * hwint.c: Moved to... * hwint.cc: ...here. * ifcvt.c: Moved to... * ifcvt.cc: ...here. * inchash.c: Moved to... * inchash.cc: ...here. * incpath.c: Moved to... * incpath.cc: ...here. * init-regs.c: Moved to... * init-regs.cc: ...here. * input.c: Moved to... * input.cc: ...here. * internal-fn.c: Moved to... * internal-fn.cc: ...here. * intl.c: Moved to... * intl.cc: ...here. * ipa-comdats.c: Moved to... * ipa-comdats.cc: ...here. * ipa-cp.c: Moved to... * ipa-cp.cc: ...here. * ipa-devirt.c: Moved to... * ipa-devirt.cc: ...here. * ipa-fnsummary.c: Moved to... * ipa-fnsummary.cc: ...here. * ipa-icf-gimple.c: Moved to... * ipa-icf-gimple.cc: ...here. * ipa-icf.c: Moved to... * ipa-icf.cc: ...here. * ipa-inline-analysis.c: Moved to... * ipa-inline-analysis.cc: ...here. * ipa-inline-transform.c: Moved to... * ipa-inline-transform.cc: ...here. * ipa-inline.c: Moved to... * ipa-inline.cc: ...here. * ipa-modref-tree.c: Moved to... * ipa-modref-tree.cc: ...here. * ipa-modref.c: Moved to... * ipa-modref.cc: ...here. * ipa-param-manipulation.c: Moved to... * ipa-param-manipulation.cc: ...here. * ipa-polymorphic-call.c: Moved to... * ipa-polymorphic-call.cc: ...here. * ipa-predicate.c: Moved to... * ipa-predicate.cc: ...here. * ipa-profile.c: Moved to... * ipa-profile.cc: ...here. * ipa-prop.c: Moved to... * ipa-prop.cc: ...here. * ipa-pure-const.c: Moved to... * ipa-pure-const.cc: ...here. * ipa-ref.c: Moved to... * ipa-ref.cc: ...here. * ipa-reference.c: Moved to... * ipa-reference.cc: ...here. * ipa-split.c: Moved to... * ipa-split.cc: ...here. * ipa-sra.c: Moved to... * ipa-sra.cc: ...here. * ipa-utils.c: Moved to... * ipa-utils.cc: ...here. * ipa-visibility.c: Moved to... * ipa-visibility.cc: ...here. * ipa.c: Moved to... * ipa.cc: ...here. * ira-build.c: Moved to... * ira-build.cc: ...here. * ira-color.c: Moved to... * ira-color.cc: ...here. * ira-conflicts.c: Moved to... * ira-conflicts.cc: ...here. * ira-costs.c: Moved to... * ira-costs.cc: ...here. * ira-emit.c: Moved to... * ira-emit.cc: ...here. * ira-lives.c: Moved to... * ira-lives.cc: ...here. * ira.c: Moved to... * ira.cc: ...here. * jump.c: Moved to... * jump.cc: ...here. * langhooks.c: Moved to... * langhooks.cc: ...here. * lcm.c: Moved to... * lcm.cc: ...here. * lists.c: Moved to... * lists.cc: ...here. * loop-doloop.c: Moved to... * loop-doloop.cc: ...here. * loop-init.c: Moved to... * loop-init.cc: ...here. * loop-invariant.c: Moved to... * loop-invariant.cc: ...here. * loop-iv.c: Moved to... * loop-iv.cc: ...here. * loop-unroll.c: Moved to... * loop-unroll.cc: ...here. * lower-subreg.c: Moved to... * lower-subreg.cc: ...here. * lra-assigns.c: Moved to... * lra-assigns.cc: ...here. * lra-coalesce.c: Moved to... * lra-coalesce.cc: ...here. * lra-constraints.c: Moved to... * lra-constraints.cc: ...here. * lra-eliminations.c: Moved to... * lra-eliminations.cc: ...here. * lra-lives.c: Moved to... * lra-lives.cc: ...here. * lra-remat.c: Moved to... * lra-remat.cc: ...here. * lra-spills.c: Moved to... * lra-spills.cc: ...here. * lra.c: Moved to... * lra.cc: ...here. * lto-cgraph.c: Moved to... * lto-cgraph.cc: ...here. * lto-compress.c: Moved to... * lto-compress.cc: ...here. * lto-opts.c: Moved to... * lto-opts.cc: ...here. * lto-section-in.c: Moved to... * lto-section-in.cc: ...here. * lto-section-out.c: Moved to... * lto-section-out.cc: ...here. * lto-streamer-in.c: Moved to... * lto-streamer-in.cc: ...here. * lto-streamer-out.c: Moved to... * lto-streamer-out.cc: ...here. * lto-streamer.c: Moved to... * lto-streamer.cc: ...here. * lto-wrapper.c: Moved to... * lto-wrapper.cc: ...here. * main.c: Moved to... * main.cc: ...here. * mcf.c: Moved to... * mcf.cc: ...here. * mode-switching.c: Moved to... * mode-switching.cc: ...here. * modulo-sched.c: Moved to... * modulo-sched.cc: ...here. * multiple_target.c: Moved to... * multiple_target.cc: ...here. * omp-expand.c: Moved to... * omp-expand.cc: ...here. * omp-general.c: Moved to... * omp-general.cc: ...here. * omp-low.c: Moved to... * omp-low.cc: ...here. * omp-offload.c: Moved to... * omp-offload.cc: ...here. * omp-simd-clone.c: Moved to... * omp-simd-clone.cc: ...here. * opt-suggestions.c: Moved to... * opt-suggestions.cc: ...here. * optabs-libfuncs.c: Moved to... * optabs-libfuncs.cc: ...here. * optabs-query.c: Moved to... * optabs-query.cc: ...here. * optabs-tree.c: Moved to... * optabs-tree.cc: ...here. * optabs.c: Moved to... * optabs.cc: ...here. * opts-common.c: Moved to... * opts-common.cc: ...here. * opts-global.c: Moved to... * opts-global.cc: ...here. * opts.c: Moved to... * opts.cc: ...here. * passes.c: Moved to... * passes.cc: ...here. * plugin.c: Moved to... * plugin.cc: ...here. * postreload-gcse.c: Moved to... * postreload-gcse.cc: ...here. * postreload.c: Moved to... * postreload.cc: ...here. * predict.c: Moved to... * predict.cc: ...here. * prefix.c: Moved to... * prefix.cc: ...here. * pretty-print.c: Moved to... * pretty-print.cc: ...here. * print-rtl-function.c: Moved to... * print-rtl-function.cc: ...here. * print-rtl.c: Moved to... * print-rtl.cc: ...here. * print-tree.c: Moved to... * print-tree.cc: ...here. * profile-count.c: Moved to... * profile-count.cc: ...here. * profile.c: Moved to... * profile.cc: ...here. * read-md.c: Moved to... * read-md.cc: ...here. * read-rtl-function.c: Moved to... * read-rtl-function.cc: ...here. * read-rtl.c: Moved to... * read-rtl.cc: ...here. * real.c: Moved to... * real.cc: ...here. * realmpfr.c: Moved to... * realmpfr.cc: ...here. * recog.c: Moved to... * recog.cc: ...here. * ree.c: Moved to... * ree.cc: ...here. * reg-stack.c: Moved to... * reg-stack.cc: ...here. * regcprop.c: Moved to... * regcprop.cc: ...here. * reginfo.c: Moved to... * reginfo.cc: ...here. * regrename.c: Moved to... * regrename.cc: ...here. * regstat.c: Moved to... * regstat.cc: ...here. * reload.c: Moved to... * reload.cc: ...here. * reload1.c: Moved to... * reload1.cc: ...here. * reorg.c: Moved to... * reorg.cc: ...here. * resource.c: Moved to... * resource.cc: ...here. * rtl-error.c: Moved to... * rtl-error.cc: ...here. * rtl-tests.c: Moved to... * rtl-tests.cc: ...here. * rtl.c: Moved to... * rtl.cc: ...here. * rtlanal.c: Moved to... * rtlanal.cc: ...here. * rtlhash.c: Moved to... * rtlhash.cc: ...here. * rtlhooks.c: Moved to... * rtlhooks.cc: ...here. * rtx-vector-builder.c: Moved to... * rtx-vector-builder.cc: ...here. * run-rtl-passes.c: Moved to... * run-rtl-passes.cc: ...here. * sancov.c: Moved to... * sancov.cc: ...here. * sanopt.c: Moved to... * sanopt.cc: ...here. * sbitmap.c: Moved to... * sbitmap.cc: ...here. * sched-deps.c: Moved to... * sched-deps.cc: ...here. * sched-ebb.c: Moved to... * sched-ebb.cc: ...here. * sched-rgn.c: Moved to... * sched-rgn.cc: ...here. * sel-sched-dump.c: Moved to... * sel-sched-dump.cc: ...here. * sel-sched-ir.c: Moved to... * sel-sched-ir.cc: ...here. * sel-sched.c: Moved to... * sel-sched.cc: ...here. * selftest-diagnostic.c: Moved to... * selftest-diagnostic.cc: ...here. * selftest-rtl.c: Moved to... * selftest-rtl.cc: ...here. * selftest-run-tests.c: Moved to... * selftest-run-tests.cc: ...here. * selftest.c: Moved to... * selftest.cc: ...here. * sese.c: Moved to... * sese.cc: ...here. * shrink-wrap.c: Moved to... * shrink-wrap.cc: ...here. * simplify-rtx.c: Moved to... * simplify-rtx.cc: ...here. * sparseset.c: Moved to... * sparseset.cc: ...here. * spellcheck-tree.c: Moved to... * spellcheck-tree.cc: ...here. * spellcheck.c: Moved to... * spellcheck.cc: ...here. * sreal.c: Moved to... * sreal.cc: ...here. * stack-ptr-mod.c: Moved to... * stack-ptr-mod.cc: ...here. * statistics.c: Moved to... * statistics.cc: ...here. * stmt.c: Moved to... * stmt.cc: ...here. * stor-layout.c: Moved to... * stor-layout.cc: ...here. * store-motion.c: Moved to... * store-motion.cc: ...here. * streamer-hooks.c: Moved to... * streamer-hooks.cc: ...here. * stringpool.c: Moved to... * stringpool.cc: ...here. * substring-locations.c: Moved to... * substring-locations.cc: ...here. * symtab.c: Moved to... * symtab.cc: ...here. * target-globals.c: Moved to... * target-globals.cc: ...here. * targhooks.c: Moved to... * targhooks.cc: ...here. * timevar.c: Moved to... * timevar.cc: ...here. * toplev.c: Moved to... * toplev.cc: ...here. * tracer.c: Moved to... * tracer.cc: ...here. * trans-mem.c: Moved to... * trans-mem.cc: ...here. * tree-affine.c: Moved to... * tree-affine.cc: ...here. * tree-call-cdce.c: Moved to... * tree-call-cdce.cc: ...here. * tree-cfg.c: Moved to... * tree-cfg.cc: ...here. * tree-cfgcleanup.c: Moved to... * tree-cfgcleanup.cc: ...here. * tree-chrec.c: Moved to... * tree-chrec.cc: ...here. * tree-complex.c: Moved to... * tree-complex.cc: ...here. * tree-data-ref.c: Moved to... * tree-data-ref.cc: ...here. * tree-dfa.c: Moved to... * tree-dfa.cc: ...here. * tree-diagnostic.c: Moved to... * tree-diagnostic.cc: ...here. * tree-dump.c: Moved to... * tree-dump.cc: ...here. * tree-eh.c: Moved to... * tree-eh.cc: ...here. * tree-emutls.c: Moved to... * tree-emutls.cc: ...here. * tree-if-conv.c: Moved to... * tree-if-conv.cc: ...here. * tree-inline.c: Moved to... * tree-inline.cc: ...here. * tree-into-ssa.c: Moved to... * tree-into-ssa.cc: ...here. * tree-iterator.c: Moved to... * tree-iterator.cc: ...here. * tree-loop-distribution.c: Moved to... * tree-loop-distribution.cc: ...here. * tree-nested.c: Moved to... * tree-nested.cc: ...here. * tree-nrv.c: Moved to... * tree-nrv.cc: ...here. * tree-object-size.c: Moved to... * tree-object-size.cc: ...here. * tree-outof-ssa.c: Moved to... * tree-outof-ssa.cc: ...here. * tree-parloops.c: Moved to... * tree-parloops.cc: ...here. * tree-phinodes.c: Moved to... * tree-phinodes.cc: ...here. * tree-predcom.c: Moved to... * tree-predcom.cc: ...here. * tree-pretty-print.c: Moved to... * tree-pretty-print.cc: ...here. * tree-profile.c: Moved to... * tree-profile.cc: ...here. * tree-scalar-evolution.c: Moved to... * tree-scalar-evolution.cc: ...here. * tree-sra.c: Moved to... * tree-sra.cc: ...here. * tree-ssa-address.c: Moved to... * tree-ssa-address.cc: ...here. * tree-ssa-alias.c: Moved to... * tree-ssa-alias.cc: ...here. * tree-ssa-ccp.c: Moved to... * tree-ssa-ccp.cc: ...here. * tree-ssa-coalesce.c: Moved to... * tree-ssa-coalesce.cc: ...here. * tree-ssa-copy.c: Moved to... * tree-ssa-copy.cc: ...here. * tree-ssa-dce.c: Moved to... * tree-ssa-dce.cc: ...here. * tree-ssa-dom.c: Moved to... * tree-ssa-dom.cc: ...here. * tree-ssa-dse.c: Moved to... * tree-ssa-dse.cc: ...here. * tree-ssa-forwprop.c: Moved to... * tree-ssa-forwprop.cc: ...here. * tree-ssa-ifcombine.c: Moved to... * tree-ssa-ifcombine.cc: ...here. * tree-ssa-live.c: Moved to... * tree-ssa-live.cc: ...here. * tree-ssa-loop-ch.c: Moved to... * tree-ssa-loop-ch.cc: ...here. * tree-ssa-loop-im.c: Moved to... * tree-ssa-loop-im.cc: ...here. * tree-ssa-loop-ivcanon.c: Moved to... * tree-ssa-loop-ivcanon.cc: ...here. * tree-ssa-loop-ivopts.c: Moved to... * tree-ssa-loop-ivopts.cc: ...here. * tree-ssa-loop-manip.c: Moved to... * tree-ssa-loop-manip.cc: ...here. * tree-ssa-loop-niter.c: Moved to... * tree-ssa-loop-niter.cc: ...here. * tree-ssa-loop-prefetch.c: Moved to... * tree-ssa-loop-prefetch.cc: ...here. * tree-ssa-loop-split.c: Moved to... * tree-ssa-loop-split.cc: ...here. * tree-ssa-loop-unswitch.c: Moved to... * tree-ssa-loop-unswitch.cc: ...here. * tree-ssa-loop.c: Moved to... * tree-ssa-loop.cc: ...here. * tree-ssa-math-opts.c: Moved to... * tree-ssa-math-opts.cc: ...here. * tree-ssa-operands.c: Moved to... * tree-ssa-operands.cc: ...here. * tree-ssa-phiopt.c: Moved to... * tree-ssa-phiopt.cc: ...here. * tree-ssa-phiprop.c: Moved to... * tree-ssa-phiprop.cc: ...here. * tree-ssa-pre.c: Moved to... * tree-ssa-pre.cc: ...here. * tree-ssa-propagate.c: Moved to... * tree-ssa-propagate.cc: ...here. * tree-ssa-reassoc.c: Moved to... * tree-ssa-reassoc.cc: ...here. * tree-ssa-sccvn.c: Moved to... * tree-ssa-sccvn.cc: ...here. * tree-ssa-scopedtables.c: Moved to... * tree-ssa-scopedtables.cc: ...here. * tree-ssa-sink.c: Moved to... * tree-ssa-sink.cc: ...here. * tree-ssa-strlen.c: Moved to... * tree-ssa-strlen.cc: ...here. * tree-ssa-structalias.c: Moved to... * tree-ssa-structalias.cc: ...here. * tree-ssa-tail-merge.c: Moved to... * tree-ssa-tail-merge.cc: ...here. * tree-ssa-ter.c: Moved to... * tree-ssa-ter.cc: ...here. * tree-ssa-threadbackward.c: Moved to... * tree-ssa-threadbackward.cc: ...here. * tree-ssa-threadedge.c: Moved to... * tree-ssa-threadedge.cc: ...here. * tree-ssa-threadupdate.c: Moved to... * tree-ssa-threadupdate.cc: ...here. * tree-ssa-uncprop.c: Moved to... * tree-ssa-uncprop.cc: ...here. * tree-ssa-uninit.c: Moved to... * tree-ssa-uninit.cc: ...here. * tree-ssa.c: Moved to... * tree-ssa.cc: ...here. * tree-ssanames.c: Moved to... * tree-ssanames.cc: ...here. * tree-stdarg.c: Moved to... * tree-stdarg.cc: ...here. * tree-streamer-in.c: Moved to... * tree-streamer-in.cc: ...here. * tree-streamer-out.c: Moved to... * tree-streamer-out.cc: ...here. * tree-streamer.c: Moved to... * tree-streamer.cc: ...here. * tree-switch-conversion.c: Moved to... * tree-switch-conversion.cc: ...here. * tree-tailcall.c: Moved to... * tree-tailcall.cc: ...here. * tree-vect-data-refs.c: Moved to... * tree-vect-data-refs.cc: ...here. * tree-vect-generic.c: Moved to... * tree-vect-generic.cc: ...here. * tree-vect-loop-manip.c: Moved to... * tree-vect-loop-manip.cc: ...here. * tree-vect-loop.c: Moved to... * tree-vect-loop.cc: ...here. * tree-vect-patterns.c: Moved to... * tree-vect-patterns.cc: ...here. * tree-vect-slp-patterns.c: Moved to... * tree-vect-slp-patterns.cc: ...here. * tree-vect-slp.c: Moved to... * tree-vect-slp.cc: ...here. * tree-vect-stmts.c: Moved to... * tree-vect-stmts.cc: ...here. * tree-vector-builder.c: Moved to... * tree-vector-builder.cc: ...here. * tree-vectorizer.c: Moved to... * tree-vectorizer.cc: ...here. * tree-vrp.c: Moved to... * tree-vrp.cc: ...here. * tree.c: Moved to... * tree.cc: ...here. * tsan.c: Moved to... * tsan.cc: ...here. * typed-splay-tree.c: Moved to... * typed-splay-tree.cc: ...here. * ubsan.c: Moved to... * ubsan.cc: ...here. * valtrack.c: Moved to... * valtrack.cc: ...here. * value-prof.c: Moved to... * value-prof.cc: ...here. * var-tracking.c: Moved to... * var-tracking.cc: ...here. * varasm.c: Moved to... * varasm.cc: ...here. * varpool.c: Moved to... * varpool.cc: ...here. * vec-perm-indices.c: Moved to... * vec-perm-indices.cc: ...here. * vec.c: Moved to... * vec.cc: ...here. * vmsdbgout.c: Moved to... * vmsdbgout.cc: ...here. * vr-values.c: Moved to... * vr-values.cc: ...here. * vtable-verify.c: Moved to... * vtable-verify.cc: ...here. * web.c: Moved to... * web.cc: ...here. * xcoffout.c: Moved to... * xcoffout.cc: ...here. gcc/c-family/ChangeLog: * c-ada-spec.c: Moved to... * c-ada-spec.cc: ...here. * c-attribs.c: Moved to... * c-attribs.cc: ...here. * c-common.c: Moved to... * c-common.cc: ...here. * c-cppbuiltin.c: Moved to... * c-cppbuiltin.cc: ...here. * c-dump.c: Moved to... * c-dump.cc: ...here. * c-format.c: Moved to... * c-format.cc: ...here. * c-gimplify.c: Moved to... * c-gimplify.cc: ...here. * c-indentation.c: Moved to... * c-indentation.cc: ...here. * c-lex.c: Moved to... * c-lex.cc: ...here. * c-omp.c: Moved to... * c-omp.cc: ...here. * c-opts.c: Moved to... * c-opts.cc: ...here. * c-pch.c: Moved to... * c-pch.cc: ...here. * c-ppoutput.c: Moved to... * c-ppoutput.cc: ...here. * c-pragma.c: Moved to... * c-pragma.cc: ...here. * c-pretty-print.c: Moved to... * c-pretty-print.cc: ...here. * c-semantics.c: Moved to... * c-semantics.cc: ...here. * c-ubsan.c: Moved to... * c-ubsan.cc: ...here. * c-warn.c: Moved to... * c-warn.cc: ...here. * cppspec.c: Moved to... * cppspec.cc: ...here. * stub-objc.c: Moved to... * stub-objc.cc: ...here. gcc/c/ChangeLog: * c-aux-info.c: Moved to... * c-aux-info.cc: ...here. * c-convert.c: Moved to... * c-convert.cc: ...here. * c-decl.c: Moved to... * c-decl.cc: ...here. * c-errors.c: Moved to... * c-errors.cc: ...here. * c-fold.c: Moved to... * c-fold.cc: ...here. * c-lang.c: Moved to... * c-lang.cc: ...here. * c-objc-common.c: Moved to... * c-objc-common.cc: ...here. * c-parser.c: Moved to... * c-parser.cc: ...here. * c-typeck.c: Moved to... * c-typeck.cc: ...here. * gccspec.c: Moved to... * gccspec.cc: ...here. * gimple-parser.c: Moved to... * gimple-parser.cc: ...here. gcc/cp/ChangeLog: * call.c: Moved to... * call.cc: ...here. * class.c: Moved to... * class.cc: ...here. * constexpr.c: Moved to... * constexpr.cc: ...here. * cp-gimplify.c: Moved to... * cp-gimplify.cc: ...here. * cp-lang.c: Moved to... * cp-lang.cc: ...here. * cp-objcp-common.c: Moved to... * cp-objcp-common.cc: ...here. * cp-ubsan.c: Moved to... * cp-ubsan.cc: ...here. * cvt.c: Moved to... * cvt.cc: ...here. * cxx-pretty-print.c: Moved to... * cxx-pretty-print.cc: ...here. * decl.c: Moved to... * decl.cc: ...here. * decl2.c: Moved to... * decl2.cc: ...here. * dump.c: Moved to... * dump.cc: ...here. * error.c: Moved to... * error.cc: ...here. * except.c: Moved to... * except.cc: ...here. * expr.c: Moved to... * expr.cc: ...here. * friend.c: Moved to... * friend.cc: ...here. * g++spec.c: Moved to... * g++spec.cc: ...here. * init.c: Moved to... * init.cc: ...here. * lambda.c: Moved to... * lambda.cc: ...here. * lex.c: Moved to... * lex.cc: ...here. * mangle.c: Moved to... * mangle.cc: ...here. * method.c: Moved to... * method.cc: ...here. * name-lookup.c: Moved to... * name-lookup.cc: ...here. * optimize.c: Moved to... * optimize.cc: ...here. * parser.c: Moved to... * parser.cc: ...here. * pt.c: Moved to... * pt.cc: ...here. * ptree.c: Moved to... * ptree.cc: ...here. * rtti.c: Moved to... * rtti.cc: ...here. * search.c: Moved to... * search.cc: ...here. * semantics.c: Moved to... * semantics.cc: ...here. * tree.c: Moved to... * tree.cc: ...here. * typeck.c: Moved to... * typeck.cc: ...here. * typeck2.c: Moved to... * typeck2.cc: ...here. * vtable-class-hierarchy.c: Moved to... * vtable-class-hierarchy.cc: ...here. gcc/fortran/ChangeLog: * arith.c: Moved to... * arith.cc: ...here. * array.c: Moved to... * array.cc: ...here. * bbt.c: Moved to... * bbt.cc: ...here. * check.c: Moved to... * check.cc: ...here. * class.c: Moved to... * class.cc: ...here. * constructor.c: Moved to... * constructor.cc: ...here. * convert.c: Moved to... * convert.cc: ...here. * cpp.c: Moved to... * cpp.cc: ...here. * data.c: Moved to... * data.cc: ...here. * decl.c: Moved to... * decl.cc: ...here. * dependency.c: Moved to... * dependency.cc: ...here. * dump-parse-tree.c: Moved to... * dump-parse-tree.cc: ...here. * error.c: Moved to... * error.cc: ...here. * expr.c: Moved to... * expr.cc: ...here. * f95-lang.c: Moved to... * f95-lang.cc: ...here. * frontend-passes.c: Moved to... * frontend-passes.cc: ...here. * gfortranspec.c: Moved to... * gfortranspec.cc: ...here. * interface.c: Moved to... * interface.cc: ...here. * intrinsic.c: Moved to... * intrinsic.cc: ...here. * io.c: Moved to... * io.cc: ...here. * iresolve.c: Moved to... * iresolve.cc: ...here. * match.c: Moved to... * match.cc: ...here. * matchexp.c: Moved to... * matchexp.cc: ...here. * misc.c: Moved to... * misc.cc: ...here. * module.c: Moved to... * module.cc: ...here. * openmp.c: Moved to... * openmp.cc: ...here. * options.c: Moved to... * options.cc: ...here. * parse.c: Moved to... * parse.cc: ...here. * primary.c: Moved to... * primary.cc: ...here. * resolve.c: Moved to... * resolve.cc: ...here. * scanner.c: Moved to... * scanner.cc: ...here. * simplify.c: Moved to... * simplify.cc: ...here. * st.c: Moved to... * st.cc: ...here. * symbol.c: Moved to... * symbol.cc: ...here. * target-memory.c: Moved to... * target-memory.cc: ...here. * trans-array.c: Moved to... * trans-array.cc: ...here. * trans-common.c: Moved to... * trans-common.cc: ...here. * trans-const.c: Moved to... * trans-const.cc: ...here. * trans-decl.c: Moved to... * trans-decl.cc: ...here. * trans-expr.c: Moved to... * trans-expr.cc: ...here. * trans-intrinsic.c: Moved to... * trans-intrinsic.cc: ...here. * trans-io.c: Moved to... * trans-io.cc: ...here. * trans-openmp.c: Moved to... * trans-openmp.cc: ...here. * trans-stmt.c: Moved to... * trans-stmt.cc: ...here. * trans-types.c: Moved to... * trans-types.cc: ...here. * trans.c: Moved to... * trans.cc: ...here. gcc/go/ChangeLog: * go-backend.c: Moved to... * go-backend.cc: ...here. * go-lang.c: Moved to... * go-lang.cc: ...here. * gospec.c: Moved to... * gospec.cc: ...here. gcc/jit/ChangeLog: * dummy-frontend.c: Moved to... * dummy-frontend.cc: ...here. * jit-builtins.c: Moved to... * jit-builtins.cc: ...here. * jit-logging.c: Moved to... * jit-logging.cc: ...here. * jit-playback.c: Moved to... * jit-playback.cc: ...here. * jit-recording.c: Moved to... * jit-recording.cc: ...here. * jit-result.c: Moved to... * jit-result.cc: ...here. * jit-spec.c: Moved to... * jit-spec.cc: ...here. * jit-tempdir.c: Moved to... * jit-tempdir.cc: ...here. * jit-w32.c: Moved to... * jit-w32.cc: ...here. * libgccjit.c: Moved to... * libgccjit.cc: ...here. gcc/lto/ChangeLog: * common.c: Moved to... * common.cc: ...here. * lto-common.c: Moved to... * lto-common.cc: ...here. * lto-dump.c: Moved to... * lto-dump.cc: ...here. * lto-lang.c: Moved to... * lto-lang.cc: ...here. * lto-object.c: Moved to... * lto-object.cc: ...here. * lto-partition.c: Moved to... * lto-partition.cc: ...here. * lto-symtab.c: Moved to... * lto-symtab.cc: ...here. * lto.c: Moved to... * lto.cc: ...here. gcc/objc/ChangeLog: * objc-act.c: Moved to... * objc-act.cc: ...here. * objc-encoding.c: Moved to... * objc-encoding.cc: ...here. * objc-gnu-runtime-abi-01.c: Moved to... * objc-gnu-runtime-abi-01.cc: ...here. * objc-lang.c: Moved to... * objc-lang.cc: ...here. * objc-map.c: Moved to... * objc-map.cc: ...here. * objc-next-runtime-abi-01.c: Moved to... * objc-next-runtime-abi-01.cc: ...here. * objc-next-runtime-abi-02.c: Moved to... * objc-next-runtime-abi-02.cc: ...here. * objc-runtime-shared-support.c: Moved to... * objc-runtime-shared-support.cc: ...here. gcc/objcp/ChangeLog: * objcp-decl.c: Moved to... * objcp-decl.cc: ...here. * objcp-lang.c: Moved to... * objcp-lang.cc: ...here. libcpp/ChangeLog: * charset.c: Moved to... * charset.cc: ...here. * directives.c: Moved to... * directives.cc: ...here. * errors.c: Moved to... * errors.cc: ...here. * expr.c: Moved to... * expr.cc: ...here. * files.c: Moved to... * files.cc: ...here. * identifiers.c: Moved to... * identifiers.cc: ...here. * init.c: Moved to... * init.cc: ...here. * lex.c: Moved to... * lex.cc: ...here. * line-map.c: Moved to... * line-map.cc: ...here. * macro.c: Moved to... * macro.cc: ...here. * makeucnid.c: Moved to... * makeucnid.cc: ...here. * mkdeps.c: Moved to... * mkdeps.cc: ...here. * pch.c: Moved to... * pch.cc: ...here. * symtab.c: Moved to... * symtab.cc: ...here. * traditional.c: Moved to... * traditional.cc: ...here.
Diffstat (limited to 'gcc/rtlanal.c')
-rw-r--r--gcc/rtlanal.c6992
1 files changed, 0 insertions, 6992 deletions
diff --git a/gcc/rtlanal.c b/gcc/rtlanal.c
deleted file mode 100644
index 69e7f39..0000000
--- a/gcc/rtlanal.c
+++ /dev/null
@@ -1,6992 +0,0 @@
-/* Analyze RTL for GNU compiler.
- Copyright (C) 1987-2022 Free Software Foundation, Inc.
-
-This file is part of GCC.
-
-GCC is free software; you can redistribute it and/or modify it under
-the terms of the GNU General Public License as published by the Free
-Software Foundation; either version 3, or (at your option) any later
-version.
-
-GCC is distributed in the hope that it will be useful, but WITHOUT ANY
-WARRANTY; without even the implied warranty of MERCHANTABILITY or
-FITNESS FOR A PARTICULAR PURPOSE. See the GNU General Public License
-for more details.
-
-You should have received a copy of the GNU General Public License
-along with GCC; see the file COPYING3. If not see
-<http://www.gnu.org/licenses/>. */
-
-
-#include "config.h"
-#include "system.h"
-#include "coretypes.h"
-#include "backend.h"
-#include "target.h"
-#include "rtl.h"
-#include "rtlanal.h"
-#include "tree.h"
-#include "predict.h"
-#include "df.h"
-#include "memmodel.h"
-#include "tm_p.h"
-#include "insn-config.h"
-#include "regs.h"
-#include "emit-rtl.h" /* FIXME: Can go away once crtl is moved to rtl.h. */
-#include "recog.h"
-#include "addresses.h"
-#include "rtl-iter.h"
-#include "hard-reg-set.h"
-#include "function-abi.h"
-
-/* Forward declarations */
-static void set_of_1 (rtx, const_rtx, void *);
-static bool covers_regno_p (const_rtx, unsigned int);
-static bool covers_regno_no_parallel_p (const_rtx, unsigned int);
-static int computed_jump_p_1 (const_rtx);
-static void parms_set (rtx, const_rtx, void *);
-
-static unsigned HOST_WIDE_INT cached_nonzero_bits (const_rtx, scalar_int_mode,
- const_rtx, machine_mode,
- unsigned HOST_WIDE_INT);
-static unsigned HOST_WIDE_INT nonzero_bits1 (const_rtx, scalar_int_mode,
- const_rtx, machine_mode,
- unsigned HOST_WIDE_INT);
-static unsigned int cached_num_sign_bit_copies (const_rtx, scalar_int_mode,
- const_rtx, machine_mode,
- unsigned int);
-static unsigned int num_sign_bit_copies1 (const_rtx, scalar_int_mode,
- const_rtx, machine_mode,
- unsigned int);
-
-rtx_subrtx_bound_info rtx_all_subrtx_bounds[NUM_RTX_CODE];
-rtx_subrtx_bound_info rtx_nonconst_subrtx_bounds[NUM_RTX_CODE];
-
-/* Truncation narrows the mode from SOURCE mode to DESTINATION mode.
- If TARGET_MODE_REP_EXTENDED (DESTINATION, DESTINATION_REP) is
- SIGN_EXTEND then while narrowing we also have to enforce the
- representation and sign-extend the value to mode DESTINATION_REP.
-
- If the value is already sign-extended to DESTINATION_REP mode we
- can just switch to DESTINATION mode on it. For each pair of
- integral modes SOURCE and DESTINATION, when truncating from SOURCE
- to DESTINATION, NUM_SIGN_BIT_COPIES_IN_REP[SOURCE][DESTINATION]
- contains the number of high-order bits in SOURCE that have to be
- copies of the sign-bit so that we can do this mode-switch to
- DESTINATION. */
-
-static unsigned int
-num_sign_bit_copies_in_rep[MAX_MODE_INT + 1][MAX_MODE_INT + 1];
-
-/* Store X into index I of ARRAY. ARRAY is known to have at least I
- elements. Return the new base of ARRAY. */
-
-template <typename T>
-typename T::value_type *
-generic_subrtx_iterator <T>::add_single_to_queue (array_type &array,
- value_type *base,
- size_t i, value_type x)
-{
- if (base == array.stack)
- {
- if (i < LOCAL_ELEMS)
- {
- base[i] = x;
- return base;
- }
- gcc_checking_assert (i == LOCAL_ELEMS);
- /* A previous iteration might also have moved from the stack to the
- heap, in which case the heap array will already be big enough. */
- if (vec_safe_length (array.heap) <= i)
- vec_safe_grow (array.heap, i + 1, true);
- base = array.heap->address ();
- memcpy (base, array.stack, sizeof (array.stack));
- base[LOCAL_ELEMS] = x;
- return base;
- }
- unsigned int length = array.heap->length ();
- if (length > i)
- {
- gcc_checking_assert (base == array.heap->address ());
- base[i] = x;
- return base;
- }
- else
- {
- gcc_checking_assert (i == length);
- vec_safe_push (array.heap, x);
- return array.heap->address ();
- }
-}
-
-/* Add the subrtxes of X to worklist ARRAY, starting at END. Return the
- number of elements added to the worklist. */
-
-template <typename T>
-size_t
-generic_subrtx_iterator <T>::add_subrtxes_to_queue (array_type &array,
- value_type *base,
- size_t end, rtx_type x)
-{
- enum rtx_code code = GET_CODE (x);
- const char *format = GET_RTX_FORMAT (code);
- size_t orig_end = end;
- if (__builtin_expect (INSN_P (x), false))
- {
- /* Put the pattern at the top of the queue, since that's what
- we're likely to want most. It also allows for the SEQUENCE
- code below. */
- for (int i = GET_RTX_LENGTH (GET_CODE (x)) - 1; i >= 0; --i)
- if (format[i] == 'e')
- {
- value_type subx = T::get_value (x->u.fld[i].rt_rtx);
- if (__builtin_expect (end < LOCAL_ELEMS, true))
- base[end++] = subx;
- else
- base = add_single_to_queue (array, base, end++, subx);
- }
- }
- else
- for (int i = 0; format[i]; ++i)
- if (format[i] == 'e')
- {
- value_type subx = T::get_value (x->u.fld[i].rt_rtx);
- if (__builtin_expect (end < LOCAL_ELEMS, true))
- base[end++] = subx;
- else
- base = add_single_to_queue (array, base, end++, subx);
- }
- else if (format[i] == 'E')
- {
- unsigned int length = GET_NUM_ELEM (x->u.fld[i].rt_rtvec);
- rtx *vec = x->u.fld[i].rt_rtvec->elem;
- if (__builtin_expect (end + length <= LOCAL_ELEMS, true))
- for (unsigned int j = 0; j < length; j++)
- base[end++] = T::get_value (vec[j]);
- else
- for (unsigned int j = 0; j < length; j++)
- base = add_single_to_queue (array, base, end++,
- T::get_value (vec[j]));
- if (code == SEQUENCE && end == length)
- /* If the subrtxes of the sequence fill the entire array then
- we know that no other parts of a containing insn are queued.
- The caller is therefore iterating over the sequence as a
- PATTERN (...), so we also want the patterns of the
- subinstructions. */
- for (unsigned int j = 0; j < length; j++)
- {
- typename T::rtx_type x = T::get_rtx (base[j]);
- if (INSN_P (x))
- base[j] = T::get_value (PATTERN (x));
- }
- }
- return end - orig_end;
-}
-
-template <typename T>
-void
-generic_subrtx_iterator <T>::free_array (array_type &array)
-{
- vec_free (array.heap);
-}
-
-template <typename T>
-const size_t generic_subrtx_iterator <T>::LOCAL_ELEMS;
-
-template class generic_subrtx_iterator <const_rtx_accessor>;
-template class generic_subrtx_iterator <rtx_var_accessor>;
-template class generic_subrtx_iterator <rtx_ptr_accessor>;
-
-/* Return 1 if the value of X is unstable
- (would be different at a different point in the program).
- The frame pointer, arg pointer, etc. are considered stable
- (within one function) and so is anything marked `unchanging'. */
-
-int
-rtx_unstable_p (const_rtx x)
-{
- const RTX_CODE code = GET_CODE (x);
- int i;
- const char *fmt;
-
- switch (code)
- {
- case MEM:
- return !MEM_READONLY_P (x) || rtx_unstable_p (XEXP (x, 0));
-
- case CONST:
- CASE_CONST_ANY:
- case SYMBOL_REF:
- case LABEL_REF:
- return 0;
-
- case REG:
- /* As in rtx_varies_p, we have to use the actual rtx, not reg number. */
- if (x == frame_pointer_rtx || x == hard_frame_pointer_rtx
- /* The arg pointer varies if it is not a fixed register. */
- || (x == arg_pointer_rtx && fixed_regs[ARG_POINTER_REGNUM]))
- return 0;
- /* ??? When call-clobbered, the value is stable modulo the restore
- that must happen after a call. This currently screws up local-alloc
- into believing that the restore is not needed. */
- if (!PIC_OFFSET_TABLE_REG_CALL_CLOBBERED && x == pic_offset_table_rtx)
- return 0;
- return 1;
-
- case ASM_OPERANDS:
- if (MEM_VOLATILE_P (x))
- return 1;
-
- /* Fall through. */
-
- default:
- break;
- }
-
- fmt = GET_RTX_FORMAT (code);
- for (i = GET_RTX_LENGTH (code) - 1; i >= 0; i--)
- if (fmt[i] == 'e')
- {
- if (rtx_unstable_p (XEXP (x, i)))
- return 1;
- }
- else if (fmt[i] == 'E')
- {
- int j;
- for (j = 0; j < XVECLEN (x, i); j++)
- if (rtx_unstable_p (XVECEXP (x, i, j)))
- return 1;
- }
-
- return 0;
-}
-
-/* Return 1 if X has a value that can vary even between two
- executions of the program. 0 means X can be compared reliably
- against certain constants or near-constants.
- FOR_ALIAS is nonzero if we are called from alias analysis; if it is
- zero, we are slightly more conservative.
- The frame pointer and the arg pointer are considered constant. */
-
-bool
-rtx_varies_p (const_rtx x, bool for_alias)
-{
- RTX_CODE code;
- int i;
- const char *fmt;
-
- if (!x)
- return 0;
-
- code = GET_CODE (x);
- switch (code)
- {
- case MEM:
- return !MEM_READONLY_P (x) || rtx_varies_p (XEXP (x, 0), for_alias);
-
- case CONST:
- CASE_CONST_ANY:
- case SYMBOL_REF:
- case LABEL_REF:
- return 0;
-
- case REG:
- /* Note that we have to test for the actual rtx used for the frame
- and arg pointers and not just the register number in case we have
- eliminated the frame and/or arg pointer and are using it
- for pseudos. */
- if (x == frame_pointer_rtx || x == hard_frame_pointer_rtx
- /* The arg pointer varies if it is not a fixed register. */
- || (x == arg_pointer_rtx && fixed_regs[ARG_POINTER_REGNUM]))
- return 0;
- if (x == pic_offset_table_rtx
- /* ??? When call-clobbered, the value is stable modulo the restore
- that must happen after a call. This currently screws up
- local-alloc into believing that the restore is not needed, so we
- must return 0 only if we are called from alias analysis. */
- && (!PIC_OFFSET_TABLE_REG_CALL_CLOBBERED || for_alias))
- return 0;
- return 1;
-
- case LO_SUM:
- /* The operand 0 of a LO_SUM is considered constant
- (in fact it is related specifically to operand 1)
- during alias analysis. */
- return (! for_alias && rtx_varies_p (XEXP (x, 0), for_alias))
- || rtx_varies_p (XEXP (x, 1), for_alias);
-
- case ASM_OPERANDS:
- if (MEM_VOLATILE_P (x))
- return 1;
-
- /* Fall through. */
-
- default:
- break;
- }
-
- fmt = GET_RTX_FORMAT (code);
- for (i = GET_RTX_LENGTH (code) - 1; i >= 0; i--)
- if (fmt[i] == 'e')
- {
- if (rtx_varies_p (XEXP (x, i), for_alias))
- return 1;
- }
- else if (fmt[i] == 'E')
- {
- int j;
- for (j = 0; j < XVECLEN (x, i); j++)
- if (rtx_varies_p (XVECEXP (x, i, j), for_alias))
- return 1;
- }
-
- return 0;
-}
-
-/* Compute an approximation for the offset between the register
- FROM and TO for the current function, as it was at the start
- of the routine. */
-
-static poly_int64
-get_initial_register_offset (int from, int to)
-{
- static const struct elim_table_t
- {
- const int from;
- const int to;
- } table[] = ELIMINABLE_REGS;
- poly_int64 offset1, offset2;
- unsigned int i, j;
-
- if (to == from)
- return 0;
-
- /* It is not safe to call INITIAL_ELIMINATION_OFFSET before the epilogue
- is completed, but we need to give at least an estimate for the stack
- pointer based on the frame size. */
- if (!epilogue_completed)
- {
- offset1 = crtl->outgoing_args_size + get_frame_size ();
-#if !STACK_GROWS_DOWNWARD
- offset1 = - offset1;
-#endif
- if (to == STACK_POINTER_REGNUM)
- return offset1;
- else if (from == STACK_POINTER_REGNUM)
- return - offset1;
- else
- return 0;
- }
-
- for (i = 0; i < ARRAY_SIZE (table); i++)
- if (table[i].from == from)
- {
- if (table[i].to == to)
- {
- INITIAL_ELIMINATION_OFFSET (table[i].from, table[i].to,
- offset1);
- return offset1;
- }
- for (j = 0; j < ARRAY_SIZE (table); j++)
- {
- if (table[j].to == to
- && table[j].from == table[i].to)
- {
- INITIAL_ELIMINATION_OFFSET (table[i].from, table[i].to,
- offset1);
- INITIAL_ELIMINATION_OFFSET (table[j].from, table[j].to,
- offset2);
- return offset1 + offset2;
- }
- if (table[j].from == to
- && table[j].to == table[i].to)
- {
- INITIAL_ELIMINATION_OFFSET (table[i].from, table[i].to,
- offset1);
- INITIAL_ELIMINATION_OFFSET (table[j].from, table[j].to,
- offset2);
- return offset1 - offset2;
- }
- }
- }
- else if (table[i].to == from)
- {
- if (table[i].from == to)
- {
- INITIAL_ELIMINATION_OFFSET (table[i].from, table[i].to,
- offset1);
- return - offset1;
- }
- for (j = 0; j < ARRAY_SIZE (table); j++)
- {
- if (table[j].to == to
- && table[j].from == table[i].from)
- {
- INITIAL_ELIMINATION_OFFSET (table[i].from, table[i].to,
- offset1);
- INITIAL_ELIMINATION_OFFSET (table[j].from, table[j].to,
- offset2);
- return - offset1 + offset2;
- }
- if (table[j].from == to
- && table[j].to == table[i].from)
- {
- INITIAL_ELIMINATION_OFFSET (table[i].from, table[i].to,
- offset1);
- INITIAL_ELIMINATION_OFFSET (table[j].from, table[j].to,
- offset2);
- return - offset1 - offset2;
- }
- }
- }
-
- /* If the requested register combination was not found,
- try a different more simple combination. */
- if (from == ARG_POINTER_REGNUM)
- return get_initial_register_offset (HARD_FRAME_POINTER_REGNUM, to);
- else if (to == ARG_POINTER_REGNUM)
- return get_initial_register_offset (from, HARD_FRAME_POINTER_REGNUM);
- else if (from == HARD_FRAME_POINTER_REGNUM)
- return get_initial_register_offset (FRAME_POINTER_REGNUM, to);
- else if (to == HARD_FRAME_POINTER_REGNUM)
- return get_initial_register_offset (from, FRAME_POINTER_REGNUM);
- else
- return 0;
-}
-
-/* Return nonzero if the use of X+OFFSET as an address in a MEM with SIZE
- bytes can cause a trap. MODE is the mode of the MEM (not that of X) and
- UNALIGNED_MEMS controls whether nonzero is returned for unaligned memory
- references on strict alignment machines. */
-
-static int
-rtx_addr_can_trap_p_1 (const_rtx x, poly_int64 offset, poly_int64 size,
- machine_mode mode, bool unaligned_mems)
-{
- enum rtx_code code = GET_CODE (x);
- gcc_checking_assert (mode == BLKmode
- || mode == VOIDmode
- || known_size_p (size));
- poly_int64 const_x1;
-
- /* The offset must be a multiple of the mode size if we are considering
- unaligned memory references on strict alignment machines. */
- if (STRICT_ALIGNMENT
- && unaligned_mems
- && mode != BLKmode
- && mode != VOIDmode)
- {
- poly_int64 actual_offset = offset;
-
-#ifdef SPARC_STACK_BOUNDARY_HACK
- /* ??? The SPARC port may claim a STACK_BOUNDARY higher than
- the real alignment of %sp. However, when it does this, the
- alignment of %sp+STACK_POINTER_OFFSET is STACK_BOUNDARY. */
- if (SPARC_STACK_BOUNDARY_HACK
- && (x == stack_pointer_rtx || x == hard_frame_pointer_rtx))
- actual_offset -= STACK_POINTER_OFFSET;
-#endif
-
- if (!multiple_p (actual_offset, GET_MODE_SIZE (mode)))
- return 1;
- }
-
- switch (code)
- {
- case SYMBOL_REF:
- if (SYMBOL_REF_WEAK (x))
- return 1;
- if (!CONSTANT_POOL_ADDRESS_P (x) && !SYMBOL_REF_FUNCTION_P (x))
- {
- tree decl;
- poly_int64 decl_size;
-
- if (maybe_lt (offset, 0))
- return 1;
- if (!known_size_p (size))
- return maybe_ne (offset, 0);
-
- /* If the size of the access or of the symbol is unknown,
- assume the worst. */
- decl = SYMBOL_REF_DECL (x);
-
- /* Else check that the access is in bounds. TODO: restructure
- expr_size/tree_expr_size/int_expr_size and just use the latter. */
- if (!decl)
- decl_size = -1;
- else if (DECL_P (decl) && DECL_SIZE_UNIT (decl))
- {
- if (!poly_int_tree_p (DECL_SIZE_UNIT (decl), &decl_size))
- decl_size = -1;
- }
- else if (TREE_CODE (decl) == STRING_CST)
- decl_size = TREE_STRING_LENGTH (decl);
- else if (TYPE_SIZE_UNIT (TREE_TYPE (decl)))
- decl_size = int_size_in_bytes (TREE_TYPE (decl));
- else
- decl_size = -1;
-
- return (!known_size_p (decl_size) || known_eq (decl_size, 0)
- ? maybe_ne (offset, 0)
- : !known_subrange_p (offset, size, 0, decl_size));
- }
-
- return 0;
-
- case LABEL_REF:
- return 0;
-
- case REG:
- /* Stack references are assumed not to trap, but we need to deal with
- nonsensical offsets. */
- if (x == frame_pointer_rtx || x == hard_frame_pointer_rtx
- || x == stack_pointer_rtx
- /* The arg pointer varies if it is not a fixed register. */
- || (x == arg_pointer_rtx && fixed_regs[ARG_POINTER_REGNUM]))
- {
-#ifdef RED_ZONE_SIZE
- poly_int64 red_zone_size = RED_ZONE_SIZE;
-#else
- poly_int64 red_zone_size = 0;
-#endif
- poly_int64 stack_boundary = PREFERRED_STACK_BOUNDARY / BITS_PER_UNIT;
- poly_int64 low_bound, high_bound;
-
- if (!known_size_p (size))
- return 1;
-
- if (x == frame_pointer_rtx)
- {
- if (FRAME_GROWS_DOWNWARD)
- {
- high_bound = targetm.starting_frame_offset ();
- low_bound = high_bound - get_frame_size ();
- }
- else
- {
- low_bound = targetm.starting_frame_offset ();
- high_bound = low_bound + get_frame_size ();
- }
- }
- else if (x == hard_frame_pointer_rtx)
- {
- poly_int64 sp_offset
- = get_initial_register_offset (STACK_POINTER_REGNUM,
- HARD_FRAME_POINTER_REGNUM);
- poly_int64 ap_offset
- = get_initial_register_offset (ARG_POINTER_REGNUM,
- HARD_FRAME_POINTER_REGNUM);
-
-#if STACK_GROWS_DOWNWARD
- low_bound = sp_offset - red_zone_size - stack_boundary;
- high_bound = ap_offset
- + FIRST_PARM_OFFSET (current_function_decl)
-#if !ARGS_GROW_DOWNWARD
- + crtl->args.size
-#endif
- + stack_boundary;
-#else
- high_bound = sp_offset + red_zone_size + stack_boundary;
- low_bound = ap_offset
- + FIRST_PARM_OFFSET (current_function_decl)
-#if ARGS_GROW_DOWNWARD
- - crtl->args.size
-#endif
- - stack_boundary;
-#endif
- }
- else if (x == stack_pointer_rtx)
- {
- poly_int64 ap_offset
- = get_initial_register_offset (ARG_POINTER_REGNUM,
- STACK_POINTER_REGNUM);
-
-#if STACK_GROWS_DOWNWARD
- low_bound = - red_zone_size - stack_boundary;
- high_bound = ap_offset
- + FIRST_PARM_OFFSET (current_function_decl)
-#if !ARGS_GROW_DOWNWARD
- + crtl->args.size
-#endif
- + stack_boundary;
-#else
- high_bound = red_zone_size + stack_boundary;
- low_bound = ap_offset
- + FIRST_PARM_OFFSET (current_function_decl)
-#if ARGS_GROW_DOWNWARD
- - crtl->args.size
-#endif
- - stack_boundary;
-#endif
- }
- else
- {
- /* We assume that accesses are safe to at least the
- next stack boundary.
- Examples are varargs and __builtin_return_address. */
-#if ARGS_GROW_DOWNWARD
- high_bound = FIRST_PARM_OFFSET (current_function_decl)
- + stack_boundary;
- low_bound = FIRST_PARM_OFFSET (current_function_decl)
- - crtl->args.size - stack_boundary;
-#else
- low_bound = FIRST_PARM_OFFSET (current_function_decl)
- - stack_boundary;
- high_bound = FIRST_PARM_OFFSET (current_function_decl)
- + crtl->args.size + stack_boundary;
-#endif
- }
-
- if (known_ge (offset, low_bound)
- && known_le (offset, high_bound - size))
- return 0;
- return 1;
- }
- /* All of the virtual frame registers are stack references. */
- if (REGNO (x) >= FIRST_VIRTUAL_REGISTER
- && REGNO (x) <= LAST_VIRTUAL_REGISTER)
- return 0;
- return 1;
-
- case CONST:
- return rtx_addr_can_trap_p_1 (XEXP (x, 0), offset, size,
- mode, unaligned_mems);
-
- case PLUS:
- /* An address is assumed not to trap if:
- - it is the pic register plus a const unspec without offset. */
- if (XEXP (x, 0) == pic_offset_table_rtx
- && GET_CODE (XEXP (x, 1)) == CONST
- && GET_CODE (XEXP (XEXP (x, 1), 0)) == UNSPEC
- && known_eq (offset, 0))
- return 0;
-
- /* - or it is an address that can't trap plus a constant integer. */
- if (poly_int_rtx_p (XEXP (x, 1), &const_x1)
- && !rtx_addr_can_trap_p_1 (XEXP (x, 0), offset + const_x1,
- size, mode, unaligned_mems))
- return 0;
-
- return 1;
-
- case LO_SUM:
- case PRE_MODIFY:
- return rtx_addr_can_trap_p_1 (XEXP (x, 1), offset, size,
- mode, unaligned_mems);
-
- case PRE_DEC:
- case PRE_INC:
- case POST_DEC:
- case POST_INC:
- case POST_MODIFY:
- return rtx_addr_can_trap_p_1 (XEXP (x, 0), offset, size,
- mode, unaligned_mems);
-
- default:
- break;
- }
-
- /* If it isn't one of the case above, it can cause a trap. */
- return 1;
-}
-
-/* Return nonzero if the use of X as an address in a MEM can cause a trap. */
-
-int
-rtx_addr_can_trap_p (const_rtx x)
-{
- return rtx_addr_can_trap_p_1 (x, 0, -1, BLKmode, false);
-}
-
-/* Return true if X contains a MEM subrtx. */
-
-bool
-contains_mem_rtx_p (rtx x)
-{
- subrtx_iterator::array_type array;
- FOR_EACH_SUBRTX (iter, array, x, ALL)
- if (MEM_P (*iter))
- return true;
-
- return false;
-}
-
-/* Return true if X is an address that is known to not be zero. */
-
-bool
-nonzero_address_p (const_rtx x)
-{
- const enum rtx_code code = GET_CODE (x);
-
- switch (code)
- {
- case SYMBOL_REF:
- return flag_delete_null_pointer_checks && !SYMBOL_REF_WEAK (x);
-
- case LABEL_REF:
- return true;
-
- case REG:
- /* As in rtx_varies_p, we have to use the actual rtx, not reg number. */
- if (x == frame_pointer_rtx || x == hard_frame_pointer_rtx
- || x == stack_pointer_rtx
- || (x == arg_pointer_rtx && fixed_regs[ARG_POINTER_REGNUM]))
- return true;
- /* All of the virtual frame registers are stack references. */
- if (REGNO (x) >= FIRST_VIRTUAL_REGISTER
- && REGNO (x) <= LAST_VIRTUAL_REGISTER)
- return true;
- return false;
-
- case CONST:
- return nonzero_address_p (XEXP (x, 0));
-
- case PLUS:
- /* Handle PIC references. */
- if (XEXP (x, 0) == pic_offset_table_rtx
- && CONSTANT_P (XEXP (x, 1)))
- return true;
- return false;
-
- case PRE_MODIFY:
- /* Similar to the above; allow positive offsets. Further, since
- auto-inc is only allowed in memories, the register must be a
- pointer. */
- if (CONST_INT_P (XEXP (x, 1))
- && INTVAL (XEXP (x, 1)) > 0)
- return true;
- return nonzero_address_p (XEXP (x, 0));
-
- case PRE_INC:
- /* Similarly. Further, the offset is always positive. */
- return true;
-
- case PRE_DEC:
- case POST_DEC:
- case POST_INC:
- case POST_MODIFY:
- return nonzero_address_p (XEXP (x, 0));
-
- case LO_SUM:
- return nonzero_address_p (XEXP (x, 1));
-
- default:
- break;
- }
-
- /* If it isn't one of the case above, might be zero. */
- return false;
-}
-
-/* Return 1 if X refers to a memory location whose address
- cannot be compared reliably with constant addresses,
- or if X refers to a BLKmode memory object.
- FOR_ALIAS is nonzero if we are called from alias analysis; if it is
- zero, we are slightly more conservative. */
-
-bool
-rtx_addr_varies_p (const_rtx x, bool for_alias)
-{
- enum rtx_code code;
- int i;
- const char *fmt;
-
- if (x == 0)
- return 0;
-
- code = GET_CODE (x);
- if (code == MEM)
- return GET_MODE (x) == BLKmode || rtx_varies_p (XEXP (x, 0), for_alias);
-
- fmt = GET_RTX_FORMAT (code);
- for (i = GET_RTX_LENGTH (code) - 1; i >= 0; i--)
- if (fmt[i] == 'e')
- {
- if (rtx_addr_varies_p (XEXP (x, i), for_alias))
- return 1;
- }
- else if (fmt[i] == 'E')
- {
- int j;
- for (j = 0; j < XVECLEN (x, i); j++)
- if (rtx_addr_varies_p (XVECEXP (x, i, j), for_alias))
- return 1;
- }
- return 0;
-}
-
-/* Return the CALL in X if there is one. */
-
-rtx
-get_call_rtx_from (const rtx_insn *insn)
-{
- rtx x = PATTERN (insn);
- if (GET_CODE (x) == PARALLEL)
- x = XVECEXP (x, 0, 0);
- if (GET_CODE (x) == SET)
- x = SET_SRC (x);
- if (GET_CODE (x) == CALL && MEM_P (XEXP (x, 0)))
- return x;
- return NULL_RTX;
-}
-
-/* Get the declaration of the function called by INSN. */
-
-tree
-get_call_fndecl (const rtx_insn *insn)
-{
- rtx note, datum;
-
- note = find_reg_note (insn, REG_CALL_DECL, NULL_RTX);
- if (note == NULL_RTX)
- return NULL_TREE;
-
- datum = XEXP (note, 0);
- if (datum != NULL_RTX)
- return SYMBOL_REF_DECL (datum);
-
- return NULL_TREE;
-}
-
-/* Return the value of the integer term in X, if one is apparent;
- otherwise return 0.
- Only obvious integer terms are detected.
- This is used in cse.c with the `related_value' field. */
-
-HOST_WIDE_INT
-get_integer_term (const_rtx x)
-{
- if (GET_CODE (x) == CONST)
- x = XEXP (x, 0);
-
- if (GET_CODE (x) == MINUS
- && CONST_INT_P (XEXP (x, 1)))
- return - INTVAL (XEXP (x, 1));
- if (GET_CODE (x) == PLUS
- && CONST_INT_P (XEXP (x, 1)))
- return INTVAL (XEXP (x, 1));
- return 0;
-}
-
-/* If X is a constant, return the value sans apparent integer term;
- otherwise return 0.
- Only obvious integer terms are detected. */
-
-rtx
-get_related_value (const_rtx x)
-{
- if (GET_CODE (x) != CONST)
- return 0;
- x = XEXP (x, 0);
- if (GET_CODE (x) == PLUS
- && CONST_INT_P (XEXP (x, 1)))
- return XEXP (x, 0);
- else if (GET_CODE (x) == MINUS
- && CONST_INT_P (XEXP (x, 1)))
- return XEXP (x, 0);
- return 0;
-}
-
-/* Return true if SYMBOL is a SYMBOL_REF and OFFSET + SYMBOL points
- to somewhere in the same object or object_block as SYMBOL. */
-
-bool
-offset_within_block_p (const_rtx symbol, HOST_WIDE_INT offset)
-{
- tree decl;
-
- if (GET_CODE (symbol) != SYMBOL_REF)
- return false;
-
- if (offset == 0)
- return true;
-
- if (offset > 0)
- {
- if (CONSTANT_POOL_ADDRESS_P (symbol)
- && offset < (int) GET_MODE_SIZE (get_pool_mode (symbol)))
- return true;
-
- decl = SYMBOL_REF_DECL (symbol);
- if (decl && offset < int_size_in_bytes (TREE_TYPE (decl)))
- return true;
- }
-
- if (SYMBOL_REF_HAS_BLOCK_INFO_P (symbol)
- && SYMBOL_REF_BLOCK (symbol)
- && SYMBOL_REF_BLOCK_OFFSET (symbol) >= 0
- && ((unsigned HOST_WIDE_INT) offset + SYMBOL_REF_BLOCK_OFFSET (symbol)
- < (unsigned HOST_WIDE_INT) SYMBOL_REF_BLOCK (symbol)->size))
- return true;
-
- return false;
-}
-
-/* Split X into a base and a constant offset, storing them in *BASE_OUT
- and *OFFSET_OUT respectively. */
-
-void
-split_const (rtx x, rtx *base_out, rtx *offset_out)
-{
- if (GET_CODE (x) == CONST)
- {
- x = XEXP (x, 0);
- if (GET_CODE (x) == PLUS && CONST_INT_P (XEXP (x, 1)))
- {
- *base_out = XEXP (x, 0);
- *offset_out = XEXP (x, 1);
- return;
- }
- }
- *base_out = x;
- *offset_out = const0_rtx;
-}
-
-/* Express integer value X as some value Y plus a polynomial offset,
- where Y is either const0_rtx, X or something within X (as opposed
- to a new rtx). Return the Y and store the offset in *OFFSET_OUT. */
-
-rtx
-strip_offset (rtx x, poly_int64_pod *offset_out)
-{
- rtx base = const0_rtx;
- rtx test = x;
- if (GET_CODE (test) == CONST)
- test = XEXP (test, 0);
- if (GET_CODE (test) == PLUS)
- {
- base = XEXP (test, 0);
- test = XEXP (test, 1);
- }
- if (poly_int_rtx_p (test, offset_out))
- return base;
- *offset_out = 0;
- return x;
-}
-
-/* Return the argument size in REG_ARGS_SIZE note X. */
-
-poly_int64
-get_args_size (const_rtx x)
-{
- gcc_checking_assert (REG_NOTE_KIND (x) == REG_ARGS_SIZE);
- return rtx_to_poly_int64 (XEXP (x, 0));
-}
-
-/* Return the number of places FIND appears within X. If COUNT_DEST is
- zero, we do not count occurrences inside the destination of a SET. */
-
-int
-count_occurrences (const_rtx x, const_rtx find, int count_dest)
-{
- int i, j;
- enum rtx_code code;
- const char *format_ptr;
- int count;
-
- if (x == find)
- return 1;
-
- code = GET_CODE (x);
-
- switch (code)
- {
- case REG:
- CASE_CONST_ANY:
- case SYMBOL_REF:
- case CODE_LABEL:
- case PC:
- return 0;
-
- case EXPR_LIST:
- count = count_occurrences (XEXP (x, 0), find, count_dest);
- if (XEXP (x, 1))
- count += count_occurrences (XEXP (x, 1), find, count_dest);
- return count;
-
- case MEM:
- if (MEM_P (find) && rtx_equal_p (x, find))
- return 1;
- break;
-
- case SET:
- if (SET_DEST (x) == find && ! count_dest)
- return count_occurrences (SET_SRC (x), find, count_dest);
- break;
-
- default:
- break;
- }
-
- format_ptr = GET_RTX_FORMAT (code);
- count = 0;
-
- for (i = 0; i < GET_RTX_LENGTH (code); i++)
- {
- switch (*format_ptr++)
- {
- case 'e':
- count += count_occurrences (XEXP (x, i), find, count_dest);
- break;
-
- case 'E':
- for (j = 0; j < XVECLEN (x, i); j++)
- count += count_occurrences (XVECEXP (x, i, j), find, count_dest);
- break;
- }
- }
- return count;
-}
-
-
-/* Return TRUE if OP is a register or subreg of a register that
- holds an unsigned quantity. Otherwise, return FALSE. */
-
-bool
-unsigned_reg_p (rtx op)
-{
- if (REG_P (op)
- && REG_EXPR (op)
- && TYPE_UNSIGNED (TREE_TYPE (REG_EXPR (op))))
- return true;
-
- if (GET_CODE (op) == SUBREG
- && SUBREG_PROMOTED_SIGN (op))
- return true;
-
- return false;
-}
-
-
-/* Nonzero if register REG appears somewhere within IN.
- Also works if REG is not a register; in this case it checks
- for a subexpression of IN that is Lisp "equal" to REG. */
-
-int
-reg_mentioned_p (const_rtx reg, const_rtx in)
-{
- const char *fmt;
- int i;
- enum rtx_code code;
-
- if (in == 0)
- return 0;
-
- if (reg == in)
- return 1;
-
- if (GET_CODE (in) == LABEL_REF)
- return reg == label_ref_label (in);
-
- code = GET_CODE (in);
-
- switch (code)
- {
- /* Compare registers by number. */
- case REG:
- return REG_P (reg) && REGNO (in) == REGNO (reg);
-
- /* These codes have no constituent expressions
- and are unique. */
- case SCRATCH:
- case PC:
- return 0;
-
- CASE_CONST_ANY:
- /* These are kept unique for a given value. */
- return 0;
-
- default:
- break;
- }
-
- if (GET_CODE (reg) == code && rtx_equal_p (reg, in))
- return 1;
-
- fmt = GET_RTX_FORMAT (code);
-
- for (i = GET_RTX_LENGTH (code) - 1; i >= 0; i--)
- {
- if (fmt[i] == 'E')
- {
- int j;
- for (j = XVECLEN (in, i) - 1; j >= 0; j--)
- if (reg_mentioned_p (reg, XVECEXP (in, i, j)))
- return 1;
- }
- else if (fmt[i] == 'e'
- && reg_mentioned_p (reg, XEXP (in, i)))
- return 1;
- }
- return 0;
-}
-
-/* Return 1 if in between BEG and END, exclusive of BEG and END, there is
- no CODE_LABEL insn. */
-
-int
-no_labels_between_p (const rtx_insn *beg, const rtx_insn *end)
-{
- rtx_insn *p;
- if (beg == end)
- return 0;
- for (p = NEXT_INSN (beg); p != end; p = NEXT_INSN (p))
- if (LABEL_P (p))
- return 0;
- return 1;
-}
-
-/* Nonzero if register REG is used in an insn between
- FROM_INSN and TO_INSN (exclusive of those two). */
-
-int
-reg_used_between_p (const_rtx reg, const rtx_insn *from_insn,
- const rtx_insn *to_insn)
-{
- rtx_insn *insn;
-
- if (from_insn == to_insn)
- return 0;
-
- for (insn = NEXT_INSN (from_insn); insn != to_insn; insn = NEXT_INSN (insn))
- if (NONDEBUG_INSN_P (insn)
- && (reg_overlap_mentioned_p (reg, PATTERN (insn))
- || (CALL_P (insn) && find_reg_fusage (insn, USE, reg))))
- return 1;
- return 0;
-}
-
-/* Nonzero if the old value of X, a register, is referenced in BODY. If X
- is entirely replaced by a new value and the only use is as a SET_DEST,
- we do not consider it a reference. */
-
-int
-reg_referenced_p (const_rtx x, const_rtx body)
-{
- int i;
-
- switch (GET_CODE (body))
- {
- case SET:
- if (reg_overlap_mentioned_p (x, SET_SRC (body)))
- return 1;
-
- /* If the destination is anything other than PC, a REG or a SUBREG
- of a REG that occupies all of the REG, the insn references X if
- it is mentioned in the destination. */
- if (GET_CODE (SET_DEST (body)) != PC
- && !REG_P (SET_DEST (body))
- && ! (GET_CODE (SET_DEST (body)) == SUBREG
- && REG_P (SUBREG_REG (SET_DEST (body)))
- && !read_modify_subreg_p (SET_DEST (body)))
- && reg_overlap_mentioned_p (x, SET_DEST (body)))
- return 1;
- return 0;
-
- case ASM_OPERANDS:
- for (i = ASM_OPERANDS_INPUT_LENGTH (body) - 1; i >= 0; i--)
- if (reg_overlap_mentioned_p (x, ASM_OPERANDS_INPUT (body, i)))
- return 1;
- return 0;
-
- case CALL:
- case USE:
- case IF_THEN_ELSE:
- return reg_overlap_mentioned_p (x, body);
-
- case TRAP_IF:
- return reg_overlap_mentioned_p (x, TRAP_CONDITION (body));
-
- case PREFETCH:
- return reg_overlap_mentioned_p (x, XEXP (body, 0));
-
- case UNSPEC:
- case UNSPEC_VOLATILE:
- for (i = XVECLEN (body, 0) - 1; i >= 0; i--)
- if (reg_overlap_mentioned_p (x, XVECEXP (body, 0, i)))
- return 1;
- return 0;
-
- case PARALLEL:
- for (i = XVECLEN (body, 0) - 1; i >= 0; i--)
- if (reg_referenced_p (x, XVECEXP (body, 0, i)))
- return 1;
- return 0;
-
- case CLOBBER:
- if (MEM_P (XEXP (body, 0)))
- if (reg_overlap_mentioned_p (x, XEXP (XEXP (body, 0), 0)))
- return 1;
- return 0;
-
- case COND_EXEC:
- if (reg_overlap_mentioned_p (x, COND_EXEC_TEST (body)))
- return 1;
- return reg_referenced_p (x, COND_EXEC_CODE (body));
-
- default:
- return 0;
- }
-}
-
-/* Nonzero if register REG is set or clobbered in an insn between
- FROM_INSN and TO_INSN (exclusive of those two). */
-
-int
-reg_set_between_p (const_rtx reg, const rtx_insn *from_insn,
- const rtx_insn *to_insn)
-{
- const rtx_insn *insn;
-
- if (from_insn == to_insn)
- return 0;
-
- for (insn = NEXT_INSN (from_insn); insn != to_insn; insn = NEXT_INSN (insn))
- if (INSN_P (insn) && reg_set_p (reg, insn))
- return 1;
- return 0;
-}
-
-/* Return true if REG is set or clobbered inside INSN. */
-
-int
-reg_set_p (const_rtx reg, const_rtx insn)
-{
- /* After delay slot handling, call and branch insns might be in a
- sequence. Check all the elements there. */
- if (INSN_P (insn) && GET_CODE (PATTERN (insn)) == SEQUENCE)
- {
- for (int i = 0; i < XVECLEN (PATTERN (insn), 0); ++i)
- if (reg_set_p (reg, XVECEXP (PATTERN (insn), 0, i)))
- return true;
-
- return false;
- }
-
- /* We can be passed an insn or part of one. If we are passed an insn,
- check if a side-effect of the insn clobbers REG. */
- if (INSN_P (insn)
- && (FIND_REG_INC_NOTE (insn, reg)
- || (CALL_P (insn)
- && ((REG_P (reg)
- && REGNO (reg) < FIRST_PSEUDO_REGISTER
- && (insn_callee_abi (as_a<const rtx_insn *> (insn))
- .clobbers_reg_p (GET_MODE (reg), REGNO (reg))))
- || MEM_P (reg)
- || find_reg_fusage (insn, CLOBBER, reg)))))
- return true;
-
- /* There are no REG_INC notes for SP autoinc. */
- if (reg == stack_pointer_rtx && INSN_P (insn))
- {
- subrtx_var_iterator::array_type array;
- FOR_EACH_SUBRTX_VAR (iter, array, PATTERN (insn), NONCONST)
- {
- rtx mem = *iter;
- if (mem
- && MEM_P (mem)
- && GET_RTX_CLASS (GET_CODE (XEXP (mem, 0))) == RTX_AUTOINC)
- {
- if (XEXP (XEXP (mem, 0), 0) == stack_pointer_rtx)
- return true;
- iter.skip_subrtxes ();
- }
- }
- }
-
- return set_of (reg, insn) != NULL_RTX;
-}
-
-/* Similar to reg_set_between_p, but check all registers in X. Return 0
- only if none of them are modified between START and END. Return 1 if
- X contains a MEM; this routine does use memory aliasing. */
-
-int
-modified_between_p (const_rtx x, const rtx_insn *start, const rtx_insn *end)
-{
- const enum rtx_code code = GET_CODE (x);
- const char *fmt;
- int i, j;
- rtx_insn *insn;
-
- if (start == end)
- return 0;
-
- switch (code)
- {
- CASE_CONST_ANY:
- case CONST:
- case SYMBOL_REF:
- case LABEL_REF:
- return 0;
-
- case PC:
- return 1;
-
- case MEM:
- if (modified_between_p (XEXP (x, 0), start, end))
- return 1;
- if (MEM_READONLY_P (x))
- return 0;
- for (insn = NEXT_INSN (start); insn != end; insn = NEXT_INSN (insn))
- if (memory_modified_in_insn_p (x, insn))
- return 1;
- return 0;
-
- case REG:
- return reg_set_between_p (x, start, end);
-
- default:
- break;
- }
-
- fmt = GET_RTX_FORMAT (code);
- for (i = GET_RTX_LENGTH (code) - 1; i >= 0; i--)
- {
- if (fmt[i] == 'e' && modified_between_p (XEXP (x, i), start, end))
- return 1;
-
- else if (fmt[i] == 'E')
- for (j = XVECLEN (x, i) - 1; j >= 0; j--)
- if (modified_between_p (XVECEXP (x, i, j), start, end))
- return 1;
- }
-
- return 0;
-}
-
-/* Similar to reg_set_p, but check all registers in X. Return 0 only if none
- of them are modified in INSN. Return 1 if X contains a MEM; this routine
- does use memory aliasing. */
-
-int
-modified_in_p (const_rtx x, const_rtx insn)
-{
- const enum rtx_code code = GET_CODE (x);
- const char *fmt;
- int i, j;
-
- switch (code)
- {
- CASE_CONST_ANY:
- case CONST:
- case SYMBOL_REF:
- case LABEL_REF:
- return 0;
-
- case PC:
- return 1;
-
- case MEM:
- if (modified_in_p (XEXP (x, 0), insn))
- return 1;
- if (MEM_READONLY_P (x))
- return 0;
- if (memory_modified_in_insn_p (x, insn))
- return 1;
- return 0;
-
- case REG:
- return reg_set_p (x, insn);
-
- default:
- break;
- }
-
- fmt = GET_RTX_FORMAT (code);
- for (i = GET_RTX_LENGTH (code) - 1; i >= 0; i--)
- {
- if (fmt[i] == 'e' && modified_in_p (XEXP (x, i), insn))
- return 1;
-
- else if (fmt[i] == 'E')
- for (j = XVECLEN (x, i) - 1; j >= 0; j--)
- if (modified_in_p (XVECEXP (x, i, j), insn))
- return 1;
- }
-
- return 0;
-}
-
-/* Return true if X is a SUBREG and if storing a value to X would
- preserve some of its SUBREG_REG. For example, on a normal 32-bit
- target, using a SUBREG to store to one half of a DImode REG would
- preserve the other half. */
-
-bool
-read_modify_subreg_p (const_rtx x)
-{
- if (GET_CODE (x) != SUBREG)
- return false;
- poly_uint64 isize = GET_MODE_SIZE (GET_MODE (SUBREG_REG (x)));
- poly_uint64 osize = GET_MODE_SIZE (GET_MODE (x));
- poly_uint64 regsize = REGMODE_NATURAL_SIZE (GET_MODE (SUBREG_REG (x)));
- /* The inner and outer modes of a subreg must be ordered, so that we
- can tell whether they're paradoxical or partial. */
- gcc_checking_assert (ordered_p (isize, osize));
- return (maybe_gt (isize, osize) && maybe_gt (isize, regsize));
-}
-
-/* Helper function for set_of. */
-struct set_of_data
- {
- const_rtx found;
- const_rtx pat;
- };
-
-static void
-set_of_1 (rtx x, const_rtx pat, void *data1)
-{
- struct set_of_data *const data = (struct set_of_data *) (data1);
- if (rtx_equal_p (x, data->pat)
- || (!MEM_P (x) && reg_overlap_mentioned_p (data->pat, x)))
- data->found = pat;
-}
-
-/* Give an INSN, return a SET or CLOBBER expression that does modify PAT
- (either directly or via STRICT_LOW_PART and similar modifiers). */
-const_rtx
-set_of (const_rtx pat, const_rtx insn)
-{
- struct set_of_data data;
- data.found = NULL_RTX;
- data.pat = pat;
- note_pattern_stores (INSN_P (insn) ? PATTERN (insn) : insn, set_of_1, &data);
- return data.found;
-}
-
-/* Check whether instruction pattern PAT contains a SET with the following
- properties:
-
- - the SET is executed unconditionally; and
- - either:
- - the destination of the SET is a REG that contains REGNO; or
- - both:
- - the destination of the SET is a SUBREG of such a REG; and
- - writing to the subreg clobbers all of the SUBREG_REG
- (in other words, read_modify_subreg_p is false).
-
- If PAT does have a SET like that, return the set, otherwise return null.
-
- This is intended to be an alternative to single_set for passes that
- can handle patterns with multiple_sets. */
-rtx
-simple_regno_set (rtx pat, unsigned int regno)
-{
- if (GET_CODE (pat) == PARALLEL)
- {
- int last = XVECLEN (pat, 0) - 1;
- for (int i = 0; i < last; ++i)
- if (rtx set = simple_regno_set (XVECEXP (pat, 0, i), regno))
- return set;
-
- pat = XVECEXP (pat, 0, last);
- }
-
- if (GET_CODE (pat) == SET
- && covers_regno_no_parallel_p (SET_DEST (pat), regno))
- return pat;
-
- return nullptr;
-}
-
-/* Add all hard register in X to *PSET. */
-void
-find_all_hard_regs (const_rtx x, HARD_REG_SET *pset)
-{
- subrtx_iterator::array_type array;
- FOR_EACH_SUBRTX (iter, array, x, NONCONST)
- {
- const_rtx x = *iter;
- if (REG_P (x) && REGNO (x) < FIRST_PSEUDO_REGISTER)
- add_to_hard_reg_set (pset, GET_MODE (x), REGNO (x));
- }
-}
-
-/* This function, called through note_stores, collects sets and
- clobbers of hard registers in a HARD_REG_SET, which is pointed to
- by DATA. */
-void
-record_hard_reg_sets (rtx x, const_rtx pat ATTRIBUTE_UNUSED, void *data)
-{
- HARD_REG_SET *pset = (HARD_REG_SET *)data;
- if (REG_P (x) && HARD_REGISTER_P (x))
- add_to_hard_reg_set (pset, GET_MODE (x), REGNO (x));
-}
-
-/* Examine INSN, and compute the set of hard registers written by it.
- Store it in *PSET. Should only be called after reload.
-
- IMPLICIT is true if we should include registers that are fully-clobbered
- by calls. This should be used with caution, since it doesn't include
- partially-clobbered registers. */
-void
-find_all_hard_reg_sets (const rtx_insn *insn, HARD_REG_SET *pset, bool implicit)
-{
- rtx link;
-
- CLEAR_HARD_REG_SET (*pset);
- note_stores (insn, record_hard_reg_sets, pset);
- if (CALL_P (insn) && implicit)
- *pset |= insn_callee_abi (insn).full_reg_clobbers ();
- for (link = REG_NOTES (insn); link; link = XEXP (link, 1))
- if (REG_NOTE_KIND (link) == REG_INC)
- record_hard_reg_sets (XEXP (link, 0), NULL, pset);
-}
-
-/* Like record_hard_reg_sets, but called through note_uses. */
-void
-record_hard_reg_uses (rtx *px, void *data)
-{
- find_all_hard_regs (*px, (HARD_REG_SET *) data);
-}
-
-/* Given an INSN, return a SET expression if this insn has only a single SET.
- It may also have CLOBBERs, USEs, or SET whose output
- will not be used, which we ignore. */
-
-rtx
-single_set_2 (const rtx_insn *insn, const_rtx pat)
-{
- rtx set = NULL;
- int set_verified = 1;
- int i;
-
- if (GET_CODE (pat) == PARALLEL)
- {
- for (i = 0; i < XVECLEN (pat, 0); i++)
- {
- rtx sub = XVECEXP (pat, 0, i);
- switch (GET_CODE (sub))
- {
- case USE:
- case CLOBBER:
- break;
-
- case SET:
- /* We can consider insns having multiple sets, where all
- but one are dead as single set insns. In common case
- only single set is present in the pattern so we want
- to avoid checking for REG_UNUSED notes unless necessary.
-
- When we reach set first time, we just expect this is
- the single set we are looking for and only when more
- sets are found in the insn, we check them. */
- if (!set_verified)
- {
- if (find_reg_note (insn, REG_UNUSED, SET_DEST (set))
- && !side_effects_p (set))
- set = NULL;
- else
- set_verified = 1;
- }
- if (!set)
- set = sub, set_verified = 0;
- else if (!find_reg_note (insn, REG_UNUSED, SET_DEST (sub))
- || side_effects_p (sub))
- return NULL_RTX;
- break;
-
- default:
- return NULL_RTX;
- }
- }
- }
- return set;
-}
-
-/* Given an INSN, return nonzero if it has more than one SET, else return
- zero. */
-
-int
-multiple_sets (const_rtx insn)
-{
- int found;
- int i;
-
- /* INSN must be an insn. */
- if (! INSN_P (insn))
- return 0;
-
- /* Only a PARALLEL can have multiple SETs. */
- if (GET_CODE (PATTERN (insn)) == PARALLEL)
- {
- for (i = 0, found = 0; i < XVECLEN (PATTERN (insn), 0); i++)
- if (GET_CODE (XVECEXP (PATTERN (insn), 0, i)) == SET)
- {
- /* If we have already found a SET, then return now. */
- if (found)
- return 1;
- else
- found = 1;
- }
- }
-
- /* Either zero or one SET. */
- return 0;
-}
-
-/* Return nonzero if the destination of SET equals the source
- and there are no side effects. */
-
-int
-set_noop_p (const_rtx set)
-{
- rtx src = SET_SRC (set);
- rtx dst = SET_DEST (set);
-
- if (dst == pc_rtx && src == pc_rtx)
- return 1;
-
- if (MEM_P (dst) && MEM_P (src))
- return rtx_equal_p (dst, src) && !side_effects_p (dst);
-
- if (GET_CODE (dst) == ZERO_EXTRACT)
- return rtx_equal_p (XEXP (dst, 0), src)
- && !BITS_BIG_ENDIAN && XEXP (dst, 2) == const0_rtx
- && !side_effects_p (src);
-
- if (GET_CODE (dst) == STRICT_LOW_PART)
- dst = XEXP (dst, 0);
-
- if (GET_CODE (src) == SUBREG && GET_CODE (dst) == SUBREG)
- {
- if (maybe_ne (SUBREG_BYTE (src), SUBREG_BYTE (dst)))
- return 0;
- src = SUBREG_REG (src);
- dst = SUBREG_REG (dst);
- if (GET_MODE (src) != GET_MODE (dst))
- /* It is hard to tell whether subregs refer to the same bits, so act
- conservatively and return 0. */
- return 0;
- }
-
- /* It is a NOOP if destination overlaps with selected src vector
- elements. */
- if (GET_CODE (src) == VEC_SELECT
- && REG_P (XEXP (src, 0)) && REG_P (dst)
- && HARD_REGISTER_P (XEXP (src, 0))
- && HARD_REGISTER_P (dst))
- {
- int i;
- rtx par = XEXP (src, 1);
- rtx src0 = XEXP (src, 0);
- poly_int64 c0;
- if (!poly_int_rtx_p (XVECEXP (par, 0, 0), &c0))
- return 0;
- poly_int64 offset = GET_MODE_UNIT_SIZE (GET_MODE (src0)) * c0;
-
- for (i = 1; i < XVECLEN (par, 0); i++)
- {
- poly_int64 c0i;
- if (!poly_int_rtx_p (XVECEXP (par, 0, i), &c0i)
- || maybe_ne (c0i, c0 + i))
- return 0;
- }
- return
- REG_CAN_CHANGE_MODE_P (REGNO (dst), GET_MODE (src0), GET_MODE (dst))
- && simplify_subreg_regno (REGNO (src0), GET_MODE (src0),
- offset, GET_MODE (dst)) == (int) REGNO (dst);
- }
-
- return (REG_P (src) && REG_P (dst)
- && REGNO (src) == REGNO (dst));
-}
-
-/* Return nonzero if an insn consists only of SETs, each of which only sets a
- value to itself. */
-
-int
-noop_move_p (const rtx_insn *insn)
-{
- rtx pat = PATTERN (insn);
-
- if (INSN_CODE (insn) == NOOP_MOVE_INSN_CODE)
- return 1;
-
- /* Check the code to be executed for COND_EXEC. */
- if (GET_CODE (pat) == COND_EXEC)
- pat = COND_EXEC_CODE (pat);
-
- if (GET_CODE (pat) == SET && set_noop_p (pat))
- return 1;
-
- if (GET_CODE (pat) == PARALLEL)
- {
- int i;
- /* If nothing but SETs of registers to themselves,
- this insn can also be deleted. */
- for (i = 0; i < XVECLEN (pat, 0); i++)
- {
- rtx tem = XVECEXP (pat, 0, i);
-
- if (GET_CODE (tem) == USE || GET_CODE (tem) == CLOBBER)
- continue;
-
- if (GET_CODE (tem) != SET || ! set_noop_p (tem))
- return 0;
- }
-
- return 1;
- }
- return 0;
-}
-
-
-/* Return nonzero if register in range [REGNO, ENDREGNO)
- appears either explicitly or implicitly in X
- other than being stored into.
-
- References contained within the substructure at LOC do not count.
- LOC may be zero, meaning don't ignore anything. */
-
-bool
-refers_to_regno_p (unsigned int regno, unsigned int endregno, const_rtx x,
- rtx *loc)
-{
- int i;
- unsigned int x_regno;
- RTX_CODE code;
- const char *fmt;
-
- repeat:
- /* The contents of a REG_NONNEG note is always zero, so we must come here
- upon repeat in case the last REG_NOTE is a REG_NONNEG note. */
- if (x == 0)
- return false;
-
- code = GET_CODE (x);
-
- switch (code)
- {
- case REG:
- x_regno = REGNO (x);
-
- /* If we modifying the stack, frame, or argument pointer, it will
- clobber a virtual register. In fact, we could be more precise,
- but it isn't worth it. */
- if ((x_regno == STACK_POINTER_REGNUM
- || (FRAME_POINTER_REGNUM != ARG_POINTER_REGNUM
- && x_regno == ARG_POINTER_REGNUM)
- || x_regno == FRAME_POINTER_REGNUM)
- && regno >= FIRST_VIRTUAL_REGISTER && regno <= LAST_VIRTUAL_REGISTER)
- return true;
-
- return endregno > x_regno && regno < END_REGNO (x);
-
- case SUBREG:
- /* If this is a SUBREG of a hard reg, we can see exactly which
- registers are being modified. Otherwise, handle normally. */
- if (REG_P (SUBREG_REG (x))
- && REGNO (SUBREG_REG (x)) < FIRST_PSEUDO_REGISTER)
- {
- unsigned int inner_regno = subreg_regno (x);
- unsigned int inner_endregno
- = inner_regno + (inner_regno < FIRST_PSEUDO_REGISTER
- ? subreg_nregs (x) : 1);
-
- return endregno > inner_regno && regno < inner_endregno;
- }
- break;
-
- case CLOBBER:
- case SET:
- if (&SET_DEST (x) != loc
- /* Note setting a SUBREG counts as referring to the REG it is in for
- a pseudo but not for hard registers since we can
- treat each word individually. */
- && ((GET_CODE (SET_DEST (x)) == SUBREG
- && loc != &SUBREG_REG (SET_DEST (x))
- && REG_P (SUBREG_REG (SET_DEST (x)))
- && REGNO (SUBREG_REG (SET_DEST (x))) >= FIRST_PSEUDO_REGISTER
- && refers_to_regno_p (regno, endregno,
- SUBREG_REG (SET_DEST (x)), loc))
- || (!REG_P (SET_DEST (x))
- && refers_to_regno_p (regno, endregno, SET_DEST (x), loc))))
- return true;
-
- if (code == CLOBBER || loc == &SET_SRC (x))
- return false;
- x = SET_SRC (x);
- goto repeat;
-
- default:
- break;
- }
-
- /* X does not match, so try its subexpressions. */
-
- fmt = GET_RTX_FORMAT (code);
- for (i = GET_RTX_LENGTH (code) - 1; i >= 0; i--)
- {
- if (fmt[i] == 'e' && loc != &XEXP (x, i))
- {
- if (i == 0)
- {
- x = XEXP (x, 0);
- goto repeat;
- }
- else
- if (refers_to_regno_p (regno, endregno, XEXP (x, i), loc))
- return true;
- }
- else if (fmt[i] == 'E')
- {
- int j;
- for (j = XVECLEN (x, i) - 1; j >= 0; j--)
- if (loc != &XVECEXP (x, i, j)
- && refers_to_regno_p (regno, endregno, XVECEXP (x, i, j), loc))
- return true;
- }
- }
- return false;
-}
-
-/* Nonzero if modifying X will affect IN. If X is a register or a SUBREG,
- we check if any register number in X conflicts with the relevant register
- numbers. If X is a constant, return 0. If X is a MEM, return 1 iff IN
- contains a MEM (we don't bother checking for memory addresses that can't
- conflict because we expect this to be a rare case. */
-
-int
-reg_overlap_mentioned_p (const_rtx x, const_rtx in)
-{
- unsigned int regno, endregno;
-
- /* If either argument is a constant, then modifying X cannot
- affect IN. Here we look at IN, we can profitably combine
- CONSTANT_P (x) with the switch statement below. */
- if (CONSTANT_P (in))
- return 0;
-
- recurse:
- switch (GET_CODE (x))
- {
- case CLOBBER:
- case STRICT_LOW_PART:
- case ZERO_EXTRACT:
- case SIGN_EXTRACT:
- /* Overly conservative. */
- x = XEXP (x, 0);
- goto recurse;
-
- case SUBREG:
- regno = REGNO (SUBREG_REG (x));
- if (regno < FIRST_PSEUDO_REGISTER)
- regno = subreg_regno (x);
- endregno = regno + (regno < FIRST_PSEUDO_REGISTER
- ? subreg_nregs (x) : 1);
- goto do_reg;
-
- case REG:
- regno = REGNO (x);
- endregno = END_REGNO (x);
- do_reg:
- return refers_to_regno_p (regno, endregno, in, (rtx*) 0);
-
- case MEM:
- {
- const char *fmt;
- int i;
-
- if (MEM_P (in))
- return 1;
-
- fmt = GET_RTX_FORMAT (GET_CODE (in));
- for (i = GET_RTX_LENGTH (GET_CODE (in)) - 1; i >= 0; i--)
- if (fmt[i] == 'e')
- {
- if (reg_overlap_mentioned_p (x, XEXP (in, i)))
- return 1;
- }
- else if (fmt[i] == 'E')
- {
- int j;
- for (j = XVECLEN (in, i) - 1; j >= 0; --j)
- if (reg_overlap_mentioned_p (x, XVECEXP (in, i, j)))
- return 1;
- }
-
- return 0;
- }
-
- case SCRATCH:
- case PC:
- return reg_mentioned_p (x, in);
-
- case PARALLEL:
- {
- int i;
-
- /* If any register in here refers to it we return true. */
- for (i = XVECLEN (x, 0) - 1; i >= 0; i--)
- if (XEXP (XVECEXP (x, 0, i), 0) != 0
- && reg_overlap_mentioned_p (XEXP (XVECEXP (x, 0, i), 0), in))
- return 1;
- return 0;
- }
-
- default:
- gcc_assert (CONSTANT_P (x));
- return 0;
- }
-}
-
-/* Call FUN on each register or MEM that is stored into or clobbered by X.
- (X would be the pattern of an insn). DATA is an arbitrary pointer,
- ignored by note_stores, but passed to FUN.
-
- FUN receives three arguments:
- 1. the REG, MEM or PC being stored in or clobbered,
- 2. the SET or CLOBBER rtx that does the store,
- 3. the pointer DATA provided to note_stores.
-
- If the item being stored in or clobbered is a SUBREG of a hard register,
- the SUBREG will be passed. */
-
-void
-note_pattern_stores (const_rtx x,
- void (*fun) (rtx, const_rtx, void *), void *data)
-{
- int i;
-
- if (GET_CODE (x) == COND_EXEC)
- x = COND_EXEC_CODE (x);
-
- if (GET_CODE (x) == SET || GET_CODE (x) == CLOBBER)
- {
- rtx dest = SET_DEST (x);
-
- while ((GET_CODE (dest) == SUBREG
- && (!REG_P (SUBREG_REG (dest))
- || REGNO (SUBREG_REG (dest)) >= FIRST_PSEUDO_REGISTER))
- || GET_CODE (dest) == ZERO_EXTRACT
- || GET_CODE (dest) == STRICT_LOW_PART)
- dest = XEXP (dest, 0);
-
- /* If we have a PARALLEL, SET_DEST is a list of EXPR_LIST expressions,
- each of whose first operand is a register. */
- if (GET_CODE (dest) == PARALLEL)
- {
- for (i = XVECLEN (dest, 0) - 1; i >= 0; i--)
- if (XEXP (XVECEXP (dest, 0, i), 0) != 0)
- (*fun) (XEXP (XVECEXP (dest, 0, i), 0), x, data);
- }
- else
- (*fun) (dest, x, data);
- }
-
- else if (GET_CODE (x) == PARALLEL)
- for (i = XVECLEN (x, 0) - 1; i >= 0; i--)
- note_pattern_stores (XVECEXP (x, 0, i), fun, data);
-}
-
-/* Same, but for an instruction. If the instruction is a call, include
- any CLOBBERs in its CALL_INSN_FUNCTION_USAGE. */
-
-void
-note_stores (const rtx_insn *insn,
- void (*fun) (rtx, const_rtx, void *), void *data)
-{
- if (CALL_P (insn))
- for (rtx link = CALL_INSN_FUNCTION_USAGE (insn);
- link; link = XEXP (link, 1))
- if (GET_CODE (XEXP (link, 0)) == CLOBBER)
- note_pattern_stores (XEXP (link, 0), fun, data);
- note_pattern_stores (PATTERN (insn), fun, data);
-}
-
-/* Like notes_stores, but call FUN for each expression that is being
- referenced in PBODY, a pointer to the PATTERN of an insn. We only call
- FUN for each expression, not any interior subexpressions. FUN receives a
- pointer to the expression and the DATA passed to this function.
-
- Note that this is not quite the same test as that done in reg_referenced_p
- since that considers something as being referenced if it is being
- partially set, while we do not. */
-
-void
-note_uses (rtx *pbody, void (*fun) (rtx *, void *), void *data)
-{
- rtx body = *pbody;
- int i;
-
- switch (GET_CODE (body))
- {
- case COND_EXEC:
- (*fun) (&COND_EXEC_TEST (body), data);
- note_uses (&COND_EXEC_CODE (body), fun, data);
- return;
-
- case PARALLEL:
- for (i = XVECLEN (body, 0) - 1; i >= 0; i--)
- note_uses (&XVECEXP (body, 0, i), fun, data);
- return;
-
- case SEQUENCE:
- for (i = XVECLEN (body, 0) - 1; i >= 0; i--)
- note_uses (&PATTERN (XVECEXP (body, 0, i)), fun, data);
- return;
-
- case USE:
- (*fun) (&XEXP (body, 0), data);
- return;
-
- case ASM_OPERANDS:
- for (i = ASM_OPERANDS_INPUT_LENGTH (body) - 1; i >= 0; i--)
- (*fun) (&ASM_OPERANDS_INPUT (body, i), data);
- return;
-
- case TRAP_IF:
- (*fun) (&TRAP_CONDITION (body), data);
- return;
-
- case PREFETCH:
- (*fun) (&XEXP (body, 0), data);
- return;
-
- case UNSPEC:
- case UNSPEC_VOLATILE:
- for (i = XVECLEN (body, 0) - 1; i >= 0; i--)
- (*fun) (&XVECEXP (body, 0, i), data);
- return;
-
- case CLOBBER:
- if (MEM_P (XEXP (body, 0)))
- (*fun) (&XEXP (XEXP (body, 0), 0), data);
- return;
-
- case SET:
- {
- rtx dest = SET_DEST (body);
-
- /* For sets we replace everything in source plus registers in memory
- expression in store and operands of a ZERO_EXTRACT. */
- (*fun) (&SET_SRC (body), data);
-
- if (GET_CODE (dest) == ZERO_EXTRACT)
- {
- (*fun) (&XEXP (dest, 1), data);
- (*fun) (&XEXP (dest, 2), data);
- }
-
- while (GET_CODE (dest) == SUBREG || GET_CODE (dest) == STRICT_LOW_PART)
- dest = XEXP (dest, 0);
-
- if (MEM_P (dest))
- (*fun) (&XEXP (dest, 0), data);
- }
- return;
-
- default:
- /* All the other possibilities never store. */
- (*fun) (pbody, data);
- return;
- }
-}
-
-/* Try to add a description of REG X to this object, stopping once
- the REF_END limit has been reached. FLAGS is a bitmask of
- rtx_obj_reference flags that describe the context. */
-
-void
-rtx_properties::try_to_add_reg (const_rtx x, unsigned int flags)
-{
- if (REG_NREGS (x) != 1)
- flags |= rtx_obj_flags::IS_MULTIREG;
- machine_mode mode = GET_MODE (x);
- unsigned int start_regno = REGNO (x);
- unsigned int end_regno = END_REGNO (x);
- for (unsigned int regno = start_regno; regno < end_regno; ++regno)
- if (ref_iter != ref_end)
- *ref_iter++ = rtx_obj_reference (regno, flags, mode,
- regno - start_regno);
-}
-
-/* Add a description of destination X to this object. FLAGS is a bitmask
- of rtx_obj_reference flags that describe the context.
-
- This routine accepts all rtxes that can legitimately appear in a
- SET_DEST. */
-
-void
-rtx_properties::try_to_add_dest (const_rtx x, unsigned int flags)
-{
- /* If we have a PARALLEL, SET_DEST is a list of EXPR_LIST expressions,
- each of whose first operand is a register. */
- if (__builtin_expect (GET_CODE (x) == PARALLEL, 0))
- {
- for (int i = XVECLEN (x, 0) - 1; i >= 0; --i)
- if (rtx dest = XEXP (XVECEXP (x, 0, i), 0))
- try_to_add_dest (dest, flags);
- return;
- }
-
- unsigned int base_flags = flags & rtx_obj_flags::STICKY_FLAGS;
- flags |= rtx_obj_flags::IS_WRITE;
- for (;;)
- if (GET_CODE (x) == ZERO_EXTRACT)
- {
- try_to_add_src (XEXP (x, 1), base_flags);
- try_to_add_src (XEXP (x, 2), base_flags);
- flags |= rtx_obj_flags::IS_READ;
- x = XEXP (x, 0);
- }
- else if (GET_CODE (x) == STRICT_LOW_PART)
- {
- flags |= rtx_obj_flags::IS_READ;
- x = XEXP (x, 0);
- }
- else if (GET_CODE (x) == SUBREG)
- {
- flags |= rtx_obj_flags::IN_SUBREG;
- if (read_modify_subreg_p (x))
- flags |= rtx_obj_flags::IS_READ;
- x = SUBREG_REG (x);
- }
- else
- break;
-
- if (MEM_P (x))
- {
- if (ref_iter != ref_end)
- *ref_iter++ = rtx_obj_reference (MEM_REGNO, flags, GET_MODE (x));
-
- unsigned int addr_flags = base_flags | rtx_obj_flags::IN_MEM_STORE;
- if (flags & rtx_obj_flags::IS_READ)
- addr_flags |= rtx_obj_flags::IN_MEM_LOAD;
- try_to_add_src (XEXP (x, 0), addr_flags);
- return;
- }
-
- if (__builtin_expect (REG_P (x), 1))
- {
- /* We want to keep sp alive everywhere - by making all
- writes to sp also use sp. */
- if (REGNO (x) == STACK_POINTER_REGNUM)
- flags |= rtx_obj_flags::IS_READ;
- try_to_add_reg (x, flags);
- return;
- }
-}
-
-/* Try to add a description of source X to this object, stopping once
- the REF_END limit has been reached. FLAGS is a bitmask of
- rtx_obj_reference flags that describe the context.
-
- This routine accepts all rtxes that can legitimately appear in a SET_SRC. */
-
-void
-rtx_properties::try_to_add_src (const_rtx x, unsigned int flags)
-{
- unsigned int base_flags = flags & rtx_obj_flags::STICKY_FLAGS;
- subrtx_iterator::array_type array;
- FOR_EACH_SUBRTX (iter, array, x, NONCONST)
- {
- const_rtx x = *iter;
- rtx_code code = GET_CODE (x);
- if (code == REG)
- try_to_add_reg (x, flags | rtx_obj_flags::IS_READ);
- else if (code == MEM)
- {
- if (MEM_VOLATILE_P (x))
- has_volatile_refs = true;
-
- if (!MEM_READONLY_P (x) && ref_iter != ref_end)
- {
- auto mem_flags = flags | rtx_obj_flags::IS_READ;
- *ref_iter++ = rtx_obj_reference (MEM_REGNO, mem_flags,
- GET_MODE (x));
- }
-
- try_to_add_src (XEXP (x, 0),
- base_flags | rtx_obj_flags::IN_MEM_LOAD);
- iter.skip_subrtxes ();
- }
- else if (code == SUBREG)
- {
- try_to_add_src (SUBREG_REG (x), flags | rtx_obj_flags::IN_SUBREG);
- iter.skip_subrtxes ();
- }
- else if (code == UNSPEC_VOLATILE)
- has_volatile_refs = true;
- else if (code == ASM_INPUT || code == ASM_OPERANDS)
- {
- has_asm = true;
- if (MEM_VOLATILE_P (x))
- has_volatile_refs = true;
- }
- else if (code == PRE_INC
- || code == PRE_DEC
- || code == POST_INC
- || code == POST_DEC
- || code == PRE_MODIFY
- || code == POST_MODIFY)
- {
- has_pre_post_modify = true;
-
- unsigned int addr_flags = (base_flags
- | rtx_obj_flags::IS_PRE_POST_MODIFY
- | rtx_obj_flags::IS_READ);
- try_to_add_dest (XEXP (x, 0), addr_flags);
- if (code == PRE_MODIFY || code == POST_MODIFY)
- iter.substitute (XEXP (XEXP (x, 1), 1));
- else
- iter.skip_subrtxes ();
- }
- else if (code == CALL)
- has_call = true;
- }
-}
-
-/* Try to add a description of instruction pattern PAT to this object,
- stopping once the REF_END limit has been reached. */
-
-void
-rtx_properties::try_to_add_pattern (const_rtx pat)
-{
- switch (GET_CODE (pat))
- {
- case COND_EXEC:
- try_to_add_src (COND_EXEC_TEST (pat));
- try_to_add_pattern (COND_EXEC_CODE (pat));
- break;
-
- case PARALLEL:
- {
- int last = XVECLEN (pat, 0) - 1;
- for (int i = 0; i < last; ++i)
- try_to_add_pattern (XVECEXP (pat, 0, i));
- try_to_add_pattern (XVECEXP (pat, 0, last));
- break;
- }
-
- case ASM_OPERANDS:
- for (int i = 0, len = ASM_OPERANDS_INPUT_LENGTH (pat); i < len; ++i)
- try_to_add_src (ASM_OPERANDS_INPUT (pat, i));
- break;
-
- case CLOBBER:
- try_to_add_dest (XEXP (pat, 0), rtx_obj_flags::IS_CLOBBER);
- break;
-
- case SET:
- try_to_add_dest (SET_DEST (pat));
- try_to_add_src (SET_SRC (pat));
- break;
-
- default:
- /* All the other possibilities never store and can use a normal
- rtx walk. This includes:
-
- - USE
- - TRAP_IF
- - PREFETCH
- - UNSPEC
- - UNSPEC_VOLATILE. */
- try_to_add_src (pat);
- break;
- }
-}
-
-/* Try to add a description of INSN to this object, stopping once
- the REF_END limit has been reached. INCLUDE_NOTES is true if the
- description should include REG_EQUAL and REG_EQUIV notes; all such
- references will then be marked with rtx_obj_flags::IN_NOTE.
-
- For calls, this description includes all accesses in
- CALL_INSN_FUNCTION_USAGE. It also include all implicit accesses
- to global registers by the target function. However, it does not
- include clobbers performed by the target function; callers that want
- this information should instead use the function_abi interface. */
-
-void
-rtx_properties::try_to_add_insn (const rtx_insn *insn, bool include_notes)
-{
- if (CALL_P (insn))
- {
- /* Non-const functions can read from global registers. Impure
- functions can also set them.
-
- Adding the global registers first removes a situation in which
- a fixed-form clobber of register R could come before a real set
- of register R. */
- if (!hard_reg_set_empty_p (global_reg_set)
- && !RTL_CONST_CALL_P (insn))
- {
- unsigned int flags = rtx_obj_flags::IS_READ;
- if (!RTL_PURE_CALL_P (insn))
- flags |= rtx_obj_flags::IS_WRITE;
- for (unsigned int regno = 0; regno < FIRST_PSEUDO_REGISTER; ++regno)
- /* As a special case, the stack pointer is invariant across calls
- even if it has been marked global; see the corresponding
- handling in df_get_call_refs. */
- if (regno != STACK_POINTER_REGNUM
- && global_regs[regno]
- && ref_iter != ref_end)
- *ref_iter++ = rtx_obj_reference (regno, flags,
- reg_raw_mode[regno], 0);
- }
- /* Untyped calls implicitly set all function value registers.
- Again, we add them first in case the main pattern contains
- a fixed-form clobber. */
- if (find_reg_note (insn, REG_UNTYPED_CALL, NULL_RTX))
- for (unsigned int regno = 0; regno < FIRST_PSEUDO_REGISTER; ++regno)
- if (targetm.calls.function_value_regno_p (regno)
- && ref_iter != ref_end)
- *ref_iter++ = rtx_obj_reference (regno, rtx_obj_flags::IS_WRITE,
- reg_raw_mode[regno], 0);
- if (ref_iter != ref_end && !RTL_CONST_CALL_P (insn))
- {
- auto mem_flags = rtx_obj_flags::IS_READ;
- if (!RTL_PURE_CALL_P (insn))
- mem_flags |= rtx_obj_flags::IS_WRITE;
- *ref_iter++ = rtx_obj_reference (MEM_REGNO, mem_flags, BLKmode);
- }
- try_to_add_pattern (PATTERN (insn));
- for (rtx link = CALL_INSN_FUNCTION_USAGE (insn); link;
- link = XEXP (link, 1))
- {
- rtx x = XEXP (link, 0);
- if (GET_CODE (x) == CLOBBER)
- try_to_add_dest (XEXP (x, 0), rtx_obj_flags::IS_CLOBBER);
- else if (GET_CODE (x) == USE)
- try_to_add_src (XEXP (x, 0));
- }
- }
- else
- try_to_add_pattern (PATTERN (insn));
-
- if (include_notes)
- for (rtx note = REG_NOTES (insn); note; note = XEXP (note, 1))
- if (REG_NOTE_KIND (note) == REG_EQUAL
- || REG_NOTE_KIND (note) == REG_EQUIV)
- try_to_add_note (XEXP (note, 0));
-}
-
-/* Grow the storage by a bit while keeping the contents of the first
- START elements. */
-
-void
-vec_rtx_properties_base::grow (ptrdiff_t start)
-{
- /* The same heuristic that vec uses. */
- ptrdiff_t new_elems = (ref_end - ref_begin) * 3 / 2;
- if (ref_begin == m_storage)
- {
- ref_begin = XNEWVEC (rtx_obj_reference, new_elems);
- if (start)
- memcpy (ref_begin, m_storage, start * sizeof (rtx_obj_reference));
- }
- else
- ref_begin = reinterpret_cast<rtx_obj_reference *>
- (xrealloc (ref_begin, new_elems * sizeof (rtx_obj_reference)));
- ref_iter = ref_begin + start;
- ref_end = ref_begin + new_elems;
-}
-
-/* Return nonzero if X's old contents don't survive after INSN.
- This will be true if X is a register and X dies in INSN or because
- INSN entirely sets X.
-
- "Entirely set" means set directly and not through a SUBREG, or
- ZERO_EXTRACT, so no trace of the old contents remains.
- Likewise, REG_INC does not count.
-
- REG may be a hard or pseudo reg. Renumbering is not taken into account,
- but for this use that makes no difference, since regs don't overlap
- during their lifetimes. Therefore, this function may be used
- at any time after deaths have been computed.
-
- If REG is a hard reg that occupies multiple machine registers, this
- function will only return 1 if each of those registers will be replaced
- by INSN. */
-
-int
-dead_or_set_p (const rtx_insn *insn, const_rtx x)
-{
- unsigned int regno, end_regno;
- unsigned int i;
-
- gcc_assert (REG_P (x));
-
- regno = REGNO (x);
- end_regno = END_REGNO (x);
- for (i = regno; i < end_regno; i++)
- if (! dead_or_set_regno_p (insn, i))
- return 0;
-
- return 1;
-}
-
-/* Return TRUE iff DEST is a register or subreg of a register, is a
- complete rather than read-modify-write destination, and contains
- register TEST_REGNO. */
-
-static bool
-covers_regno_no_parallel_p (const_rtx dest, unsigned int test_regno)
-{
- unsigned int regno, endregno;
-
- if (GET_CODE (dest) == SUBREG && !read_modify_subreg_p (dest))
- dest = SUBREG_REG (dest);
-
- if (!REG_P (dest))
- return false;
-
- regno = REGNO (dest);
- endregno = END_REGNO (dest);
- return (test_regno >= regno && test_regno < endregno);
-}
-
-/* Like covers_regno_no_parallel_p, but also handles PARALLELs where
- any member matches the covers_regno_no_parallel_p criteria. */
-
-static bool
-covers_regno_p (const_rtx dest, unsigned int test_regno)
-{
- if (GET_CODE (dest) == PARALLEL)
- {
- /* Some targets place small structures in registers for return
- values of functions, and those registers are wrapped in
- PARALLELs that we may see as the destination of a SET. */
- int i;
-
- for (i = XVECLEN (dest, 0) - 1; i >= 0; i--)
- {
- rtx inner = XEXP (XVECEXP (dest, 0, i), 0);
- if (inner != NULL_RTX
- && covers_regno_no_parallel_p (inner, test_regno))
- return true;
- }
-
- return false;
- }
- else
- return covers_regno_no_parallel_p (dest, test_regno);
-}
-
-/* Utility function for dead_or_set_p to check an individual register. */
-
-int
-dead_or_set_regno_p (const rtx_insn *insn, unsigned int test_regno)
-{
- const_rtx pattern;
-
- /* See if there is a death note for something that includes TEST_REGNO. */
- if (find_regno_note (insn, REG_DEAD, test_regno))
- return 1;
-
- if (CALL_P (insn)
- && find_regno_fusage (insn, CLOBBER, test_regno))
- return 1;
-
- pattern = PATTERN (insn);
-
- /* If a COND_EXEC is not executed, the value survives. */
- if (GET_CODE (pattern) == COND_EXEC)
- return 0;
-
- if (GET_CODE (pattern) == SET || GET_CODE (pattern) == CLOBBER)
- return covers_regno_p (SET_DEST (pattern), test_regno);
- else if (GET_CODE (pattern) == PARALLEL)
- {
- int i;
-
- for (i = XVECLEN (pattern, 0) - 1; i >= 0; i--)
- {
- rtx body = XVECEXP (pattern, 0, i);
-
- if (GET_CODE (body) == COND_EXEC)
- body = COND_EXEC_CODE (body);
-
- if ((GET_CODE (body) == SET || GET_CODE (body) == CLOBBER)
- && covers_regno_p (SET_DEST (body), test_regno))
- return 1;
- }
- }
-
- return 0;
-}
-
-/* Return the reg-note of kind KIND in insn INSN, if there is one.
- If DATUM is nonzero, look for one whose datum is DATUM. */
-
-rtx
-find_reg_note (const_rtx insn, enum reg_note kind, const_rtx datum)
-{
- rtx link;
-
- gcc_checking_assert (insn);
-
- /* Ignore anything that is not an INSN, JUMP_INSN or CALL_INSN. */
- if (! INSN_P (insn))
- return 0;
- if (datum == 0)
- {
- for (link = REG_NOTES (insn); link; link = XEXP (link, 1))
- if (REG_NOTE_KIND (link) == kind)
- return link;
- return 0;
- }
-
- for (link = REG_NOTES (insn); link; link = XEXP (link, 1))
- if (REG_NOTE_KIND (link) == kind && datum == XEXP (link, 0))
- return link;
- return 0;
-}
-
-/* Return the reg-note of kind KIND in insn INSN which applies to register
- number REGNO, if any. Return 0 if there is no such reg-note. Note that
- the REGNO of this NOTE need not be REGNO if REGNO is a hard register;
- it might be the case that the note overlaps REGNO. */
-
-rtx
-find_regno_note (const_rtx insn, enum reg_note kind, unsigned int regno)
-{
- rtx link;
-
- /* Ignore anything that is not an INSN, JUMP_INSN or CALL_INSN. */
- if (! INSN_P (insn))
- return 0;
-
- for (link = REG_NOTES (insn); link; link = XEXP (link, 1))
- if (REG_NOTE_KIND (link) == kind
- /* Verify that it is a register, so that scratch and MEM won't cause a
- problem here. */
- && REG_P (XEXP (link, 0))
- && REGNO (XEXP (link, 0)) <= regno
- && END_REGNO (XEXP (link, 0)) > regno)
- return link;
- return 0;
-}
-
-/* Return a REG_EQUIV or REG_EQUAL note if insn has only a single set and
- has such a note. */
-
-rtx
-find_reg_equal_equiv_note (const_rtx insn)
-{
- rtx link;
-
- if (!INSN_P (insn))
- return 0;
-
- for (link = REG_NOTES (insn); link; link = XEXP (link, 1))
- if (REG_NOTE_KIND (link) == REG_EQUAL
- || REG_NOTE_KIND (link) == REG_EQUIV)
- {
- /* FIXME: We should never have REG_EQUAL/REG_EQUIV notes on
- insns that have multiple sets. Checking single_set to
- make sure of this is not the proper check, as explained
- in the comment in set_unique_reg_note.
-
- This should be changed into an assert. */
- if (GET_CODE (PATTERN (insn)) == PARALLEL && multiple_sets (insn))
- return 0;
- return link;
- }
- return NULL;
-}
-
-/* Check whether INSN is a single_set whose source is known to be
- equivalent to a constant. Return that constant if so, otherwise
- return null. */
-
-rtx
-find_constant_src (const rtx_insn *insn)
-{
- rtx note, set, x;
-
- set = single_set (insn);
- if (set)
- {
- x = avoid_constant_pool_reference (SET_SRC (set));
- if (CONSTANT_P (x))
- return x;
- }
-
- note = find_reg_equal_equiv_note (insn);
- if (note && CONSTANT_P (XEXP (note, 0)))
- return XEXP (note, 0);
-
- return NULL_RTX;
-}
-
-/* Return true if DATUM, or any overlap of DATUM, of kind CODE is found
- in the CALL_INSN_FUNCTION_USAGE information of INSN. */
-
-int
-find_reg_fusage (const_rtx insn, enum rtx_code code, const_rtx datum)
-{
- /* If it's not a CALL_INSN, it can't possibly have a
- CALL_INSN_FUNCTION_USAGE field, so don't bother checking. */
- if (!CALL_P (insn))
- return 0;
-
- gcc_assert (datum);
-
- if (!REG_P (datum))
- {
- rtx link;
-
- for (link = CALL_INSN_FUNCTION_USAGE (insn);
- link;
- link = XEXP (link, 1))
- if (GET_CODE (XEXP (link, 0)) == code
- && rtx_equal_p (datum, XEXP (XEXP (link, 0), 0)))
- return 1;
- }
- else
- {
- unsigned int regno = REGNO (datum);
-
- /* CALL_INSN_FUNCTION_USAGE information cannot contain references
- to pseudo registers, so don't bother checking. */
-
- if (regno < FIRST_PSEUDO_REGISTER)
- {
- unsigned int end_regno = END_REGNO (datum);
- unsigned int i;
-
- for (i = regno; i < end_regno; i++)
- if (find_regno_fusage (insn, code, i))
- return 1;
- }
- }
-
- return 0;
-}
-
-/* Return true if REGNO, or any overlap of REGNO, of kind CODE is found
- in the CALL_INSN_FUNCTION_USAGE information of INSN. */
-
-int
-find_regno_fusage (const_rtx insn, enum rtx_code code, unsigned int regno)
-{
- rtx link;
-
- /* CALL_INSN_FUNCTION_USAGE information cannot contain references
- to pseudo registers, so don't bother checking. */
-
- if (regno >= FIRST_PSEUDO_REGISTER
- || !CALL_P (insn) )
- return 0;
-
- for (link = CALL_INSN_FUNCTION_USAGE (insn); link; link = XEXP (link, 1))
- {
- rtx op, reg;
-
- if (GET_CODE (op = XEXP (link, 0)) == code
- && REG_P (reg = XEXP (op, 0))
- && REGNO (reg) <= regno
- && END_REGNO (reg) > regno)
- return 1;
- }
-
- return 0;
-}
-
-
-/* Return true if KIND is an integer REG_NOTE. */
-
-static bool
-int_reg_note_p (enum reg_note kind)
-{
- return kind == REG_BR_PROB;
-}
-
-/* Allocate a register note with kind KIND and datum DATUM. LIST is
- stored as the pointer to the next register note. */
-
-rtx
-alloc_reg_note (enum reg_note kind, rtx datum, rtx list)
-{
- rtx note;
-
- gcc_checking_assert (!int_reg_note_p (kind));
- switch (kind)
- {
- case REG_LABEL_TARGET:
- case REG_LABEL_OPERAND:
- case REG_TM:
- /* These types of register notes use an INSN_LIST rather than an
- EXPR_LIST, so that copying is done right and dumps look
- better. */
- note = alloc_INSN_LIST (datum, list);
- PUT_REG_NOTE_KIND (note, kind);
- break;
-
- default:
- note = alloc_EXPR_LIST (kind, datum, list);
- break;
- }
-
- return note;
-}
-
-/* Add register note with kind KIND and datum DATUM to INSN. */
-
-void
-add_reg_note (rtx insn, enum reg_note kind, rtx datum)
-{
- REG_NOTES (insn) = alloc_reg_note (kind, datum, REG_NOTES (insn));
-}
-
-/* Add an integer register note with kind KIND and datum DATUM to INSN. */
-
-void
-add_int_reg_note (rtx_insn *insn, enum reg_note kind, int datum)
-{
- gcc_checking_assert (int_reg_note_p (kind));
- REG_NOTES (insn) = gen_rtx_INT_LIST ((machine_mode) kind,
- datum, REG_NOTES (insn));
-}
-
-/* Add a REG_ARGS_SIZE note to INSN with value VALUE. */
-
-void
-add_args_size_note (rtx_insn *insn, poly_int64 value)
-{
- gcc_checking_assert (!find_reg_note (insn, REG_ARGS_SIZE, NULL_RTX));
- add_reg_note (insn, REG_ARGS_SIZE, gen_int_mode (value, Pmode));
-}
-
-/* Add a register note like NOTE to INSN. */
-
-void
-add_shallow_copy_of_reg_note (rtx_insn *insn, rtx note)
-{
- if (GET_CODE (note) == INT_LIST)
- add_int_reg_note (insn, REG_NOTE_KIND (note), XINT (note, 0));
- else
- add_reg_note (insn, REG_NOTE_KIND (note), XEXP (note, 0));
-}
-
-/* Duplicate NOTE and return the copy. */
-rtx
-duplicate_reg_note (rtx note)
-{
- reg_note kind = REG_NOTE_KIND (note);
-
- if (GET_CODE (note) == INT_LIST)
- return gen_rtx_INT_LIST ((machine_mode) kind, XINT (note, 0), NULL_RTX);
- else if (GET_CODE (note) == EXPR_LIST)
- return alloc_reg_note (kind, copy_insn_1 (XEXP (note, 0)), NULL_RTX);
- else
- return alloc_reg_note (kind, XEXP (note, 0), NULL_RTX);
-}
-
-/* Remove register note NOTE from the REG_NOTES of INSN. */
-
-void
-remove_note (rtx_insn *insn, const_rtx note)
-{
- rtx link;
-
- if (note == NULL_RTX)
- return;
-
- if (REG_NOTES (insn) == note)
- REG_NOTES (insn) = XEXP (note, 1);
- else
- for (link = REG_NOTES (insn); link; link = XEXP (link, 1))
- if (XEXP (link, 1) == note)
- {
- XEXP (link, 1) = XEXP (note, 1);
- break;
- }
-
- switch (REG_NOTE_KIND (note))
- {
- case REG_EQUAL:
- case REG_EQUIV:
- df_notes_rescan (insn);
- break;
- default:
- break;
- }
-}
-
-/* Remove REG_EQUAL and/or REG_EQUIV notes if INSN has such notes.
- If NO_RESCAN is false and any notes were removed, call
- df_notes_rescan. Return true if any note has been removed. */
-
-bool
-remove_reg_equal_equiv_notes (rtx_insn *insn, bool no_rescan)
-{
- rtx *loc;
- bool ret = false;
-
- loc = &REG_NOTES (insn);
- while (*loc)
- {
- enum reg_note kind = REG_NOTE_KIND (*loc);
- if (kind == REG_EQUAL || kind == REG_EQUIV)
- {
- *loc = XEXP (*loc, 1);
- ret = true;
- }
- else
- loc = &XEXP (*loc, 1);
- }
- if (ret && !no_rescan)
- df_notes_rescan (insn);
- return ret;
-}
-
-/* Remove all REG_EQUAL and REG_EQUIV notes referring to REGNO. */
-
-void
-remove_reg_equal_equiv_notes_for_regno (unsigned int regno)
-{
- df_ref eq_use;
-
- if (!df)
- return;
-
- /* This loop is a little tricky. We cannot just go down the chain because
- it is being modified by some actions in the loop. So we just iterate
- over the head. We plan to drain the list anyway. */
- while ((eq_use = DF_REG_EQ_USE_CHAIN (regno)) != NULL)
- {
- rtx_insn *insn = DF_REF_INSN (eq_use);
- rtx note = find_reg_equal_equiv_note (insn);
-
- /* This assert is generally triggered when someone deletes a REG_EQUAL
- or REG_EQUIV note by hacking the list manually rather than calling
- remove_note. */
- gcc_assert (note);
-
- remove_note (insn, note);
- }
-}
-
-/* Search LISTP (an EXPR_LIST) for an entry whose first operand is NODE and
- return 1 if it is found. A simple equality test is used to determine if
- NODE matches. */
-
-bool
-in_insn_list_p (const rtx_insn_list *listp, const rtx_insn *node)
-{
- const_rtx x;
-
- for (x = listp; x; x = XEXP (x, 1))
- if (node == XEXP (x, 0))
- return true;
-
- return false;
-}
-
-/* Search LISTP (an EXPR_LIST) for an entry whose first operand is NODE and
- remove that entry from the list if it is found.
-
- A simple equality test is used to determine if NODE matches. */
-
-void
-remove_node_from_expr_list (const_rtx node, rtx_expr_list **listp)
-{
- rtx_expr_list *temp = *listp;
- rtx_expr_list *prev = NULL;
-
- while (temp)
- {
- if (node == temp->element ())
- {
- /* Splice the node out of the list. */
- if (prev)
- XEXP (prev, 1) = temp->next ();
- else
- *listp = temp->next ();
-
- return;
- }
-
- prev = temp;
- temp = temp->next ();
- }
-}
-
-/* Search LISTP (an INSN_LIST) for an entry whose first operand is NODE and
- remove that entry from the list if it is found.
-
- A simple equality test is used to determine if NODE matches. */
-
-void
-remove_node_from_insn_list (const rtx_insn *node, rtx_insn_list **listp)
-{
- rtx_insn_list *temp = *listp;
- rtx_insn_list *prev = NULL;
-
- while (temp)
- {
- if (node == temp->insn ())
- {
- /* Splice the node out of the list. */
- if (prev)
- XEXP (prev, 1) = temp->next ();
- else
- *listp = temp->next ();
-
- return;
- }
-
- prev = temp;
- temp = temp->next ();
- }
-}
-
-/* Nonzero if X contains any volatile instructions. These are instructions
- which may cause unpredictable machine state instructions, and thus no
- instructions or register uses should be moved or combined across them.
- This includes only volatile asms and UNSPEC_VOLATILE instructions. */
-
-int
-volatile_insn_p (const_rtx x)
-{
- const RTX_CODE code = GET_CODE (x);
- switch (code)
- {
- case LABEL_REF:
- case SYMBOL_REF:
- case CONST:
- CASE_CONST_ANY:
- case PC:
- case REG:
- case SCRATCH:
- case CLOBBER:
- case ADDR_VEC:
- case ADDR_DIFF_VEC:
- case CALL:
- case MEM:
- return 0;
-
- case UNSPEC_VOLATILE:
- return 1;
-
- case ASM_INPUT:
- case ASM_OPERANDS:
- if (MEM_VOLATILE_P (x))
- return 1;
-
- default:
- break;
- }
-
- /* Recursively scan the operands of this expression. */
-
- {
- const char *const fmt = GET_RTX_FORMAT (code);
- int i;
-
- for (i = GET_RTX_LENGTH (code) - 1; i >= 0; i--)
- {
- if (fmt[i] == 'e')
- {
- if (volatile_insn_p (XEXP (x, i)))
- return 1;
- }
- else if (fmt[i] == 'E')
- {
- int j;
- for (j = 0; j < XVECLEN (x, i); j++)
- if (volatile_insn_p (XVECEXP (x, i, j)))
- return 1;
- }
- }
- }
- return 0;
-}
-
-/* Nonzero if X contains any volatile memory references
- UNSPEC_VOLATILE operations or volatile ASM_OPERANDS expressions. */
-
-int
-volatile_refs_p (const_rtx x)
-{
- const RTX_CODE code = GET_CODE (x);
- switch (code)
- {
- case LABEL_REF:
- case SYMBOL_REF:
- case CONST:
- CASE_CONST_ANY:
- case PC:
- case REG:
- case SCRATCH:
- case CLOBBER:
- case ADDR_VEC:
- case ADDR_DIFF_VEC:
- return 0;
-
- case UNSPEC_VOLATILE:
- return 1;
-
- case MEM:
- case ASM_INPUT:
- case ASM_OPERANDS:
- if (MEM_VOLATILE_P (x))
- return 1;
-
- default:
- break;
- }
-
- /* Recursively scan the operands of this expression. */
-
- {
- const char *const fmt = GET_RTX_FORMAT (code);
- int i;
-
- for (i = GET_RTX_LENGTH (code) - 1; i >= 0; i--)
- {
- if (fmt[i] == 'e')
- {
- if (volatile_refs_p (XEXP (x, i)))
- return 1;
- }
- else if (fmt[i] == 'E')
- {
- int j;
- for (j = 0; j < XVECLEN (x, i); j++)
- if (volatile_refs_p (XVECEXP (x, i, j)))
- return 1;
- }
- }
- }
- return 0;
-}
-
-/* Similar to above, except that it also rejects register pre- and post-
- incrementing. */
-
-int
-side_effects_p (const_rtx x)
-{
- const RTX_CODE code = GET_CODE (x);
- switch (code)
- {
- case LABEL_REF:
- case SYMBOL_REF:
- case CONST:
- CASE_CONST_ANY:
- case PC:
- case REG:
- case SCRATCH:
- case ADDR_VEC:
- case ADDR_DIFF_VEC:
- case VAR_LOCATION:
- return 0;
-
- case CLOBBER:
- /* Reject CLOBBER with a non-VOID mode. These are made by combine.c
- when some combination can't be done. If we see one, don't think
- that we can simplify the expression. */
- return (GET_MODE (x) != VOIDmode);
-
- case PRE_INC:
- case PRE_DEC:
- case POST_INC:
- case POST_DEC:
- case PRE_MODIFY:
- case POST_MODIFY:
- case CALL:
- case UNSPEC_VOLATILE:
- return 1;
-
- case MEM:
- case ASM_INPUT:
- case ASM_OPERANDS:
- if (MEM_VOLATILE_P (x))
- return 1;
-
- default:
- break;
- }
-
- /* Recursively scan the operands of this expression. */
-
- {
- const char *fmt = GET_RTX_FORMAT (code);
- int i;
-
- for (i = GET_RTX_LENGTH (code) - 1; i >= 0; i--)
- {
- if (fmt[i] == 'e')
- {
- if (side_effects_p (XEXP (x, i)))
- return 1;
- }
- else if (fmt[i] == 'E')
- {
- int j;
- for (j = 0; j < XVECLEN (x, i); j++)
- if (side_effects_p (XVECEXP (x, i, j)))
- return 1;
- }
- }
- }
- return 0;
-}
-
-/* Return nonzero if evaluating rtx X might cause a trap.
- FLAGS controls how to consider MEMs. A nonzero means the context
- of the access may have changed from the original, such that the
- address may have become invalid. */
-
-int
-may_trap_p_1 (const_rtx x, unsigned flags)
-{
- int i;
- enum rtx_code code;
- const char *fmt;
-
- /* We make no distinction currently, but this function is part of
- the internal target-hooks ABI so we keep the parameter as
- "unsigned flags". */
- bool code_changed = flags != 0;
-
- if (x == 0)
- return 0;
- code = GET_CODE (x);
- switch (code)
- {
- /* Handle these cases quickly. */
- CASE_CONST_ANY:
- case SYMBOL_REF:
- case LABEL_REF:
- case CONST:
- case PC:
- case REG:
- case SCRATCH:
- return 0;
-
- case UNSPEC:
- return targetm.unspec_may_trap_p (x, flags);
-
- case UNSPEC_VOLATILE:
- case ASM_INPUT:
- case TRAP_IF:
- return 1;
-
- case ASM_OPERANDS:
- return MEM_VOLATILE_P (x);
-
- /* Memory ref can trap unless it's a static var or a stack slot. */
- case MEM:
- /* Recognize specific pattern of stack checking probes. */
- if (flag_stack_check
- && MEM_VOLATILE_P (x)
- && XEXP (x, 0) == stack_pointer_rtx)
- return 1;
- if (/* MEM_NOTRAP_P only relates to the actual position of the memory
- reference; moving it out of context such as when moving code
- when optimizing, might cause its address to become invalid. */
- code_changed
- || !MEM_NOTRAP_P (x))
- {
- poly_int64 size = MEM_SIZE_KNOWN_P (x) ? MEM_SIZE (x) : -1;
- return rtx_addr_can_trap_p_1 (XEXP (x, 0), 0, size,
- GET_MODE (x), code_changed);
- }
-
- return 0;
-
- /* Division by a non-constant might trap. */
- case DIV:
- case MOD:
- case UDIV:
- case UMOD:
- if (HONOR_SNANS (x))
- return 1;
- if (FLOAT_MODE_P (GET_MODE (x)))
- return flag_trapping_math;
- if (!CONSTANT_P (XEXP (x, 1)) || (XEXP (x, 1) == const0_rtx))
- return 1;
- if (GET_CODE (XEXP (x, 1)) == CONST_VECTOR)
- {
- /* For CONST_VECTOR, return 1 if any element is or might be zero. */
- unsigned int n_elts;
- rtx op = XEXP (x, 1);
- if (!GET_MODE_NUNITS (GET_MODE (op)).is_constant (&n_elts))
- {
- if (!CONST_VECTOR_DUPLICATE_P (op))
- return 1;
- for (unsigned i = 0; i < (unsigned int) XVECLEN (op, 0); i++)
- if (CONST_VECTOR_ENCODED_ELT (op, i) == const0_rtx)
- return 1;
- }
- else
- for (unsigned i = 0; i < n_elts; i++)
- if (CONST_VECTOR_ELT (op, i) == const0_rtx)
- return 1;
- }
- break;
-
- case EXPR_LIST:
- /* An EXPR_LIST is used to represent a function call. This
- certainly may trap. */
- return 1;
-
- case GE:
- case GT:
- case LE:
- case LT:
- case LTGT:
- case COMPARE:
- /* Some floating point comparisons may trap. */
- if (!flag_trapping_math)
- break;
- /* ??? There is no machine independent way to check for tests that trap
- when COMPARE is used, though many targets do make this distinction.
- For instance, sparc uses CCFPE for compares which generate exceptions
- and CCFP for compares which do not generate exceptions. */
- if (HONOR_NANS (x))
- return 1;
- /* But often the compare has some CC mode, so check operand
- modes as well. */
- if (HONOR_NANS (XEXP (x, 0))
- || HONOR_NANS (XEXP (x, 1)))
- return 1;
- break;
-
- case EQ:
- case NE:
- if (HONOR_SNANS (x))
- return 1;
- /* Often comparison is CC mode, so check operand modes. */
- if (HONOR_SNANS (XEXP (x, 0))
- || HONOR_SNANS (XEXP (x, 1)))
- return 1;
- break;
-
- case FIX:
- case UNSIGNED_FIX:
- /* Conversion of floating point might trap. */
- if (flag_trapping_math && HONOR_NANS (XEXP (x, 0)))
- return 1;
- break;
-
- case NEG:
- case ABS:
- case SUBREG:
- case VEC_MERGE:
- case VEC_SELECT:
- case VEC_CONCAT:
- case VEC_DUPLICATE:
- /* These operations don't trap even with floating point. */
- break;
-
- default:
- /* Any floating arithmetic may trap. */
- if (FLOAT_MODE_P (GET_MODE (x)) && flag_trapping_math)
- return 1;
- }
-
- fmt = GET_RTX_FORMAT (code);
- for (i = GET_RTX_LENGTH (code) - 1; i >= 0; i--)
- {
- if (fmt[i] == 'e')
- {
- if (may_trap_p_1 (XEXP (x, i), flags))
- return 1;
- }
- else if (fmt[i] == 'E')
- {
- int j;
- for (j = 0; j < XVECLEN (x, i); j++)
- if (may_trap_p_1 (XVECEXP (x, i, j), flags))
- return 1;
- }
- }
- return 0;
-}
-
-/* Return nonzero if evaluating rtx X might cause a trap. */
-
-int
-may_trap_p (const_rtx x)
-{
- return may_trap_p_1 (x, 0);
-}
-
-/* Same as above, but additionally return nonzero if evaluating rtx X might
- cause a fault. We define a fault for the purpose of this function as a
- erroneous execution condition that cannot be encountered during the normal
- execution of a valid program; the typical example is an unaligned memory
- access on a strict alignment machine. The compiler guarantees that it
- doesn't generate code that will fault from a valid program, but this
- guarantee doesn't mean anything for individual instructions. Consider
- the following example:
-
- struct S { int d; union { char *cp; int *ip; }; };
-
- int foo(struct S *s)
- {
- if (s->d == 1)
- return *s->ip;
- else
- return *s->cp;
- }
-
- on a strict alignment machine. In a valid program, foo will never be
- invoked on a structure for which d is equal to 1 and the underlying
- unique field of the union not aligned on a 4-byte boundary, but the
- expression *s->ip might cause a fault if considered individually.
-
- At the RTL level, potentially problematic expressions will almost always
- verify may_trap_p; for example, the above dereference can be emitted as
- (mem:SI (reg:P)) and this expression is may_trap_p for a generic register.
- However, suppose that foo is inlined in a caller that causes s->cp to
- point to a local character variable and guarantees that s->d is not set
- to 1; foo may have been effectively translated into pseudo-RTL as:
-
- if ((reg:SI) == 1)
- (set (reg:SI) (mem:SI (%fp - 7)))
- else
- (set (reg:QI) (mem:QI (%fp - 7)))
-
- Now (mem:SI (%fp - 7)) is considered as not may_trap_p since it is a
- memory reference to a stack slot, but it will certainly cause a fault
- on a strict alignment machine. */
-
-int
-may_trap_or_fault_p (const_rtx x)
-{
- return may_trap_p_1 (x, 1);
-}
-
-/* Replace any occurrence of FROM in X with TO. The function does
- not enter into CONST_DOUBLE for the replace.
-
- Note that copying is not done so X must not be shared unless all copies
- are to be modified.
-
- ALL_REGS is true if we want to replace all REGs equal to FROM, not just
- those pointer-equal ones. */
-
-rtx
-replace_rtx (rtx x, rtx from, rtx to, bool all_regs)
-{
- int i, j;
- const char *fmt;
-
- if (x == from)
- return to;
-
- /* Allow this function to make replacements in EXPR_LISTs. */
- if (x == 0)
- return 0;
-
- if (all_regs
- && REG_P (x)
- && REG_P (from)
- && REGNO (x) == REGNO (from))
- {
- gcc_assert (GET_MODE (x) == GET_MODE (from));
- return to;
- }
- else if (GET_CODE (x) == SUBREG)
- {
- rtx new_rtx = replace_rtx (SUBREG_REG (x), from, to, all_regs);
-
- if (CONST_INT_P (new_rtx))
- {
- x = simplify_subreg (GET_MODE (x), new_rtx,
- GET_MODE (SUBREG_REG (x)),
- SUBREG_BYTE (x));
- gcc_assert (x);
- }
- else
- SUBREG_REG (x) = new_rtx;
-
- return x;
- }
- else if (GET_CODE (x) == ZERO_EXTEND)
- {
- rtx new_rtx = replace_rtx (XEXP (x, 0), from, to, all_regs);
-
- if (CONST_INT_P (new_rtx))
- {
- x = simplify_unary_operation (ZERO_EXTEND, GET_MODE (x),
- new_rtx, GET_MODE (XEXP (x, 0)));
- gcc_assert (x);
- }
- else
- XEXP (x, 0) = new_rtx;
-
- return x;
- }
-
- fmt = GET_RTX_FORMAT (GET_CODE (x));
- for (i = GET_RTX_LENGTH (GET_CODE (x)) - 1; i >= 0; i--)
- {
- if (fmt[i] == 'e')
- XEXP (x, i) = replace_rtx (XEXP (x, i), from, to, all_regs);
- else if (fmt[i] == 'E')
- for (j = XVECLEN (x, i) - 1; j >= 0; j--)
- XVECEXP (x, i, j) = replace_rtx (XVECEXP (x, i, j),
- from, to, all_regs);
- }
-
- return x;
-}
-
-/* Replace occurrences of the OLD_LABEL in *LOC with NEW_LABEL. Also track
- the change in LABEL_NUSES if UPDATE_LABEL_NUSES. */
-
-void
-replace_label (rtx *loc, rtx old_label, rtx new_label, bool update_label_nuses)
-{
- /* Handle jump tables specially, since ADDR_{DIFF_,}VECs can be long. */
- rtx x = *loc;
- if (JUMP_TABLE_DATA_P (x))
- {
- x = PATTERN (x);
- rtvec vec = XVEC (x, GET_CODE (x) == ADDR_DIFF_VEC);
- int len = GET_NUM_ELEM (vec);
- for (int i = 0; i < len; ++i)
- {
- rtx ref = RTVEC_ELT (vec, i);
- if (XEXP (ref, 0) == old_label)
- {
- XEXP (ref, 0) = new_label;
- if (update_label_nuses)
- {
- ++LABEL_NUSES (new_label);
- --LABEL_NUSES (old_label);
- }
- }
- }
- return;
- }
-
- /* If this is a JUMP_INSN, then we also need to fix the JUMP_LABEL
- field. This is not handled by the iterator because it doesn't
- handle unprinted ('0') fields. */
- if (JUMP_P (x) && JUMP_LABEL (x) == old_label)
- JUMP_LABEL (x) = new_label;
-
- subrtx_ptr_iterator::array_type array;
- FOR_EACH_SUBRTX_PTR (iter, array, loc, ALL)
- {
- rtx *loc = *iter;
- if (rtx x = *loc)
- {
- if (GET_CODE (x) == SYMBOL_REF
- && CONSTANT_POOL_ADDRESS_P (x))
- {
- rtx c = get_pool_constant (x);
- if (rtx_referenced_p (old_label, c))
- {
- /* Create a copy of constant C; replace the label inside
- but do not update LABEL_NUSES because uses in constant pool
- are not counted. */
- rtx new_c = copy_rtx (c);
- replace_label (&new_c, old_label, new_label, false);
-
- /* Add the new constant NEW_C to constant pool and replace
- the old reference to constant by new reference. */
- rtx new_mem = force_const_mem (get_pool_mode (x), new_c);
- *loc = replace_rtx (x, x, XEXP (new_mem, 0));
- }
- }
-
- if ((GET_CODE (x) == LABEL_REF
- || GET_CODE (x) == INSN_LIST)
- && XEXP (x, 0) == old_label)
- {
- XEXP (x, 0) = new_label;
- if (update_label_nuses)
- {
- ++LABEL_NUSES (new_label);
- --LABEL_NUSES (old_label);
- }
- }
- }
- }
-}
-
-void
-replace_label_in_insn (rtx_insn *insn, rtx_insn *old_label,
- rtx_insn *new_label, bool update_label_nuses)
-{
- rtx insn_as_rtx = insn;
- replace_label (&insn_as_rtx, old_label, new_label, update_label_nuses);
- gcc_checking_assert (insn_as_rtx == insn);
-}
-
-/* Return true if X is referenced in BODY. */
-
-bool
-rtx_referenced_p (const_rtx x, const_rtx body)
-{
- subrtx_iterator::array_type array;
- FOR_EACH_SUBRTX (iter, array, body, ALL)
- if (const_rtx y = *iter)
- {
- /* Check if a label_ref Y refers to label X. */
- if (GET_CODE (y) == LABEL_REF
- && LABEL_P (x)
- && label_ref_label (y) == x)
- return true;
-
- if (rtx_equal_p (x, y))
- return true;
-
- /* If Y is a reference to pool constant traverse the constant. */
- if (GET_CODE (y) == SYMBOL_REF
- && CONSTANT_POOL_ADDRESS_P (y))
- iter.substitute (get_pool_constant (y));
- }
- return false;
-}
-
-/* If INSN is a tablejump return true and store the label (before jump table) to
- *LABELP and the jump table to *TABLEP. LABELP and TABLEP may be NULL. */
-
-bool
-tablejump_p (const rtx_insn *insn, rtx_insn **labelp,
- rtx_jump_table_data **tablep)
-{
- if (!JUMP_P (insn))
- return false;
-
- rtx target = JUMP_LABEL (insn);
- if (target == NULL_RTX || ANY_RETURN_P (target))
- return false;
-
- rtx_insn *label = as_a<rtx_insn *> (target);
- rtx_insn *table = next_insn (label);
- if (table == NULL_RTX || !JUMP_TABLE_DATA_P (table))
- return false;
-
- if (labelp)
- *labelp = label;
- if (tablep)
- *tablep = as_a <rtx_jump_table_data *> (table);
- return true;
-}
-
-/* For INSN known to satisfy tablejump_p, determine if it actually is a
- CASESI. Return the insn pattern if so, NULL_RTX otherwise. */
-
-rtx
-tablejump_casesi_pattern (const rtx_insn *insn)
-{
- rtx tmp;
-
- if ((tmp = single_set (insn)) != NULL
- && SET_DEST (tmp) == pc_rtx
- && GET_CODE (SET_SRC (tmp)) == IF_THEN_ELSE
- && GET_CODE (XEXP (SET_SRC (tmp), 2)) == LABEL_REF)
- return tmp;
-
- return NULL_RTX;
-}
-
-/* A subroutine of computed_jump_p, return 1 if X contains a REG or MEM or
- constant that is not in the constant pool and not in the condition
- of an IF_THEN_ELSE. */
-
-static int
-computed_jump_p_1 (const_rtx x)
-{
- const enum rtx_code code = GET_CODE (x);
- int i, j;
- const char *fmt;
-
- switch (code)
- {
- case LABEL_REF:
- case PC:
- return 0;
-
- case CONST:
- CASE_CONST_ANY:
- case SYMBOL_REF:
- case REG:
- return 1;
-
- case MEM:
- return ! (GET_CODE (XEXP (x, 0)) == SYMBOL_REF
- && CONSTANT_POOL_ADDRESS_P (XEXP (x, 0)));
-
- case IF_THEN_ELSE:
- return (computed_jump_p_1 (XEXP (x, 1))
- || computed_jump_p_1 (XEXP (x, 2)));
-
- default:
- break;
- }
-
- fmt = GET_RTX_FORMAT (code);
- for (i = GET_RTX_LENGTH (code) - 1; i >= 0; i--)
- {
- if (fmt[i] == 'e'
- && computed_jump_p_1 (XEXP (x, i)))
- return 1;
-
- else if (fmt[i] == 'E')
- for (j = 0; j < XVECLEN (x, i); j++)
- if (computed_jump_p_1 (XVECEXP (x, i, j)))
- return 1;
- }
-
- return 0;
-}
-
-/* Return nonzero if INSN is an indirect jump (aka computed jump).
-
- Tablejumps and casesi insns are not considered indirect jumps;
- we can recognize them by a (use (label_ref)). */
-
-int
-computed_jump_p (const rtx_insn *insn)
-{
- int i;
- if (JUMP_P (insn))
- {
- rtx pat = PATTERN (insn);
-
- /* If we have a JUMP_LABEL set, we're not a computed jump. */
- if (JUMP_LABEL (insn) != NULL)
- return 0;
-
- if (GET_CODE (pat) == PARALLEL)
- {
- int len = XVECLEN (pat, 0);
- int has_use_labelref = 0;
-
- for (i = len - 1; i >= 0; i--)
- if (GET_CODE (XVECEXP (pat, 0, i)) == USE
- && (GET_CODE (XEXP (XVECEXP (pat, 0, i), 0))
- == LABEL_REF))
- {
- has_use_labelref = 1;
- break;
- }
-
- if (! has_use_labelref)
- for (i = len - 1; i >= 0; i--)
- if (GET_CODE (XVECEXP (pat, 0, i)) == SET
- && SET_DEST (XVECEXP (pat, 0, i)) == pc_rtx
- && computed_jump_p_1 (SET_SRC (XVECEXP (pat, 0, i))))
- return 1;
- }
- else if (GET_CODE (pat) == SET
- && SET_DEST (pat) == pc_rtx
- && computed_jump_p_1 (SET_SRC (pat)))
- return 1;
- }
- return 0;
-}
-
-
-
-/* MEM has a PRE/POST-INC/DEC/MODIFY address X. Extract the operands of
- the equivalent add insn and pass the result to FN, using DATA as the
- final argument. */
-
-static int
-for_each_inc_dec_find_inc_dec (rtx mem, for_each_inc_dec_fn fn, void *data)
-{
- rtx x = XEXP (mem, 0);
- switch (GET_CODE (x))
- {
- case PRE_INC:
- case POST_INC:
- {
- poly_int64 size = GET_MODE_SIZE (GET_MODE (mem));
- rtx r1 = XEXP (x, 0);
- rtx c = gen_int_mode (size, GET_MODE (r1));
- return fn (mem, x, r1, r1, c, data);
- }
-
- case PRE_DEC:
- case POST_DEC:
- {
- poly_int64 size = GET_MODE_SIZE (GET_MODE (mem));
- rtx r1 = XEXP (x, 0);
- rtx c = gen_int_mode (-size, GET_MODE (r1));
- return fn (mem, x, r1, r1, c, data);
- }
-
- case PRE_MODIFY:
- case POST_MODIFY:
- {
- rtx r1 = XEXP (x, 0);
- rtx add = XEXP (x, 1);
- return fn (mem, x, r1, add, NULL, data);
- }
-
- default:
- gcc_unreachable ();
- }
-}
-
-/* Traverse *LOC looking for MEMs that have autoinc addresses.
- For each such autoinc operation found, call FN, passing it
- the innermost enclosing MEM, the operation itself, the RTX modified
- by the operation, two RTXs (the second may be NULL) that, once
- added, represent the value to be held by the modified RTX
- afterwards, and DATA. FN is to return 0 to continue the
- traversal or any other value to have it returned to the caller of
- for_each_inc_dec. */
-
-int
-for_each_inc_dec (rtx x,
- for_each_inc_dec_fn fn,
- void *data)
-{
- subrtx_var_iterator::array_type array;
- FOR_EACH_SUBRTX_VAR (iter, array, x, NONCONST)
- {
- rtx mem = *iter;
- if (mem
- && MEM_P (mem)
- && GET_RTX_CLASS (GET_CODE (XEXP (mem, 0))) == RTX_AUTOINC)
- {
- int res = for_each_inc_dec_find_inc_dec (mem, fn, data);
- if (res != 0)
- return res;
- iter.skip_subrtxes ();
- }
- }
- return 0;
-}
-
-
-/* Searches X for any reference to REGNO, returning the rtx of the
- reference found if any. Otherwise, returns NULL_RTX. */
-
-rtx
-regno_use_in (unsigned int regno, rtx x)
-{
- const char *fmt;
- int i, j;
- rtx tem;
-
- if (REG_P (x) && REGNO (x) == regno)
- return x;
-
- fmt = GET_RTX_FORMAT (GET_CODE (x));
- for (i = GET_RTX_LENGTH (GET_CODE (x)) - 1; i >= 0; i--)
- {
- if (fmt[i] == 'e')
- {
- if ((tem = regno_use_in (regno, XEXP (x, i))))
- return tem;
- }
- else if (fmt[i] == 'E')
- for (j = XVECLEN (x, i) - 1; j >= 0; j--)
- if ((tem = regno_use_in (regno , XVECEXP (x, i, j))))
- return tem;
- }
-
- return NULL_RTX;
-}
-
-/* Return a value indicating whether OP, an operand of a commutative
- operation, is preferred as the first or second operand. The more
- positive the value, the stronger the preference for being the first
- operand. */
-
-int
-commutative_operand_precedence (rtx op)
-{
- enum rtx_code code = GET_CODE (op);
-
- /* Constants always become the second operand. Prefer "nice" constants. */
- if (code == CONST_INT)
- return -10;
- if (code == CONST_WIDE_INT)
- return -9;
- if (code == CONST_POLY_INT)
- return -8;
- if (code == CONST_DOUBLE)
- return -8;
- if (code == CONST_FIXED)
- return -8;
- op = avoid_constant_pool_reference (op);
- code = GET_CODE (op);
-
- switch (GET_RTX_CLASS (code))
- {
- case RTX_CONST_OBJ:
- if (code == CONST_INT)
- return -7;
- if (code == CONST_WIDE_INT)
- return -6;
- if (code == CONST_POLY_INT)
- return -5;
- if (code == CONST_DOUBLE)
- return -5;
- if (code == CONST_FIXED)
- return -5;
- return -4;
-
- case RTX_EXTRA:
- /* SUBREGs of objects should come second. */
- if (code == SUBREG && OBJECT_P (SUBREG_REG (op)))
- return -3;
- return 0;
-
- case RTX_OBJ:
- /* Complex expressions should be the first, so decrease priority
- of objects. Prefer pointer objects over non pointer objects. */
- if ((REG_P (op) && REG_POINTER (op))
- || (MEM_P (op) && MEM_POINTER (op)))
- return -1;
- return -2;
-
- case RTX_COMM_ARITH:
- /* Prefer operands that are themselves commutative to be first.
- This helps to make things linear. In particular,
- (and (and (reg) (reg)) (not (reg))) is canonical. */
- return 4;
-
- case RTX_BIN_ARITH:
- /* If only one operand is a binary expression, it will be the first
- operand. In particular, (plus (minus (reg) (reg)) (neg (reg)))
- is canonical, although it will usually be further simplified. */
- return 2;
-
- case RTX_UNARY:
- /* Then prefer NEG and NOT. */
- if (code == NEG || code == NOT)
- return 1;
- /* FALLTHRU */
-
- default:
- return 0;
- }
-}
-
-/* Return 1 iff it is necessary to swap operands of commutative operation
- in order to canonicalize expression. */
-
-bool
-swap_commutative_operands_p (rtx x, rtx y)
-{
- return (commutative_operand_precedence (x)
- < commutative_operand_precedence (y));
-}
-
-/* Return 1 if X is an autoincrement side effect and the register is
- not the stack pointer. */
-int
-auto_inc_p (const_rtx x)
-{
- switch (GET_CODE (x))
- {
- case PRE_INC:
- case POST_INC:
- case PRE_DEC:
- case POST_DEC:
- case PRE_MODIFY:
- case POST_MODIFY:
- /* There are no REG_INC notes for SP. */
- if (XEXP (x, 0) != stack_pointer_rtx)
- return 1;
- default:
- break;
- }
- return 0;
-}
-
-/* Return nonzero if IN contains a piece of rtl that has the address LOC. */
-int
-loc_mentioned_in_p (rtx *loc, const_rtx in)
-{
- enum rtx_code code;
- const char *fmt;
- int i, j;
-
- if (!in)
- return 0;
-
- code = GET_CODE (in);
- fmt = GET_RTX_FORMAT (code);
- for (i = GET_RTX_LENGTH (code) - 1; i >= 0; i--)
- {
- if (fmt[i] == 'e')
- {
- if (loc == &XEXP (in, i) || loc_mentioned_in_p (loc, XEXP (in, i)))
- return 1;
- }
- else if (fmt[i] == 'E')
- for (j = XVECLEN (in, i) - 1; j >= 0; j--)
- if (loc == &XVECEXP (in, i, j)
- || loc_mentioned_in_p (loc, XVECEXP (in, i, j)))
- return 1;
- }
- return 0;
-}
-
-/* Reinterpret a subreg as a bit extraction from an integer and return
- the position of the least significant bit of the extracted value.
- In other words, if the extraction were performed as a shift right
- and mask, return the number of bits to shift right.
-
- The outer value of the subreg has OUTER_BYTES bytes and starts at
- byte offset SUBREG_BYTE within an inner value of INNER_BYTES bytes. */
-
-poly_uint64
-subreg_size_lsb (poly_uint64 outer_bytes,
- poly_uint64 inner_bytes,
- poly_uint64 subreg_byte)
-{
- poly_uint64 subreg_end, trailing_bytes, byte_pos;
-
- /* A paradoxical subreg begins at bit position 0. */
- gcc_checking_assert (ordered_p (outer_bytes, inner_bytes));
- if (maybe_gt (outer_bytes, inner_bytes))
- {
- gcc_checking_assert (known_eq (subreg_byte, 0U));
- return 0;
- }
-
- subreg_end = subreg_byte + outer_bytes;
- trailing_bytes = inner_bytes - subreg_end;
- if (WORDS_BIG_ENDIAN && BYTES_BIG_ENDIAN)
- byte_pos = trailing_bytes;
- else if (!WORDS_BIG_ENDIAN && !BYTES_BIG_ENDIAN)
- byte_pos = subreg_byte;
- else
- {
- /* When bytes and words have opposite endianness, we must be able
- to split offsets into words and bytes at compile time. */
- poly_uint64 leading_word_part
- = force_align_down (subreg_byte, UNITS_PER_WORD);
- poly_uint64 trailing_word_part
- = force_align_down (trailing_bytes, UNITS_PER_WORD);
- /* If the subreg crosses a word boundary ensure that
- it also begins and ends on a word boundary. */
- gcc_assert (known_le (subreg_end - leading_word_part,
- (unsigned int) UNITS_PER_WORD)
- || (known_eq (leading_word_part, subreg_byte)
- && known_eq (trailing_word_part, trailing_bytes)));
- if (WORDS_BIG_ENDIAN)
- byte_pos = trailing_word_part + (subreg_byte - leading_word_part);
- else
- byte_pos = leading_word_part + (trailing_bytes - trailing_word_part);
- }
-
- return byte_pos * BITS_PER_UNIT;
-}
-
-/* Given a subreg X, return the bit offset where the subreg begins
- (counting from the least significant bit of the reg). */
-
-poly_uint64
-subreg_lsb (const_rtx x)
-{
- return subreg_lsb_1 (GET_MODE (x), GET_MODE (SUBREG_REG (x)),
- SUBREG_BYTE (x));
-}
-
-/* Return the subreg byte offset for a subreg whose outer value has
- OUTER_BYTES bytes, whose inner value has INNER_BYTES bytes, and where
- there are LSB_SHIFT *bits* between the lsb of the outer value and the
- lsb of the inner value. This is the inverse of the calculation
- performed by subreg_lsb_1 (which converts byte offsets to bit shifts). */
-
-poly_uint64
-subreg_size_offset_from_lsb (poly_uint64 outer_bytes, poly_uint64 inner_bytes,
- poly_uint64 lsb_shift)
-{
- /* A paradoxical subreg begins at bit position 0. */
- gcc_checking_assert (ordered_p (outer_bytes, inner_bytes));
- if (maybe_gt (outer_bytes, inner_bytes))
- {
- gcc_checking_assert (known_eq (lsb_shift, 0U));
- return 0;
- }
-
- poly_uint64 lower_bytes = exact_div (lsb_shift, BITS_PER_UNIT);
- poly_uint64 upper_bytes = inner_bytes - (lower_bytes + outer_bytes);
- if (WORDS_BIG_ENDIAN && BYTES_BIG_ENDIAN)
- return upper_bytes;
- else if (!WORDS_BIG_ENDIAN && !BYTES_BIG_ENDIAN)
- return lower_bytes;
- else
- {
- /* When bytes and words have opposite endianness, we must be able
- to split offsets into words and bytes at compile time. */
- poly_uint64 lower_word_part = force_align_down (lower_bytes,
- UNITS_PER_WORD);
- poly_uint64 upper_word_part = force_align_down (upper_bytes,
- UNITS_PER_WORD);
- if (WORDS_BIG_ENDIAN)
- return upper_word_part + (lower_bytes - lower_word_part);
- else
- return lower_word_part + (upper_bytes - upper_word_part);
- }
-}
-
-/* Fill in information about a subreg of a hard register.
- xregno - A regno of an inner hard subreg_reg (or what will become one).
- xmode - The mode of xregno.
- offset - The byte offset.
- ymode - The mode of a top level SUBREG (or what may become one).
- info - Pointer to structure to fill in.
-
- Rather than considering one particular inner register (and thus one
- particular "outer" register) in isolation, this function really uses
- XREGNO as a model for a sequence of isomorphic hard registers. Thus the
- function does not check whether adding INFO->offset to XREGNO gives
- a valid hard register; even if INFO->offset + XREGNO is out of range,
- there might be another register of the same type that is in range.
- Likewise it doesn't check whether targetm.hard_regno_mode_ok accepts
- the new register, since that can depend on things like whether the final
- register number is even or odd. Callers that want to check whether
- this particular subreg can be replaced by a simple (reg ...) should
- use simplify_subreg_regno. */
-
-void
-subreg_get_info (unsigned int xregno, machine_mode xmode,
- poly_uint64 offset, machine_mode ymode,
- struct subreg_info *info)
-{
- unsigned int nregs_xmode, nregs_ymode;
-
- gcc_assert (xregno < FIRST_PSEUDO_REGISTER);
-
- poly_uint64 xsize = GET_MODE_SIZE (xmode);
- poly_uint64 ysize = GET_MODE_SIZE (ymode);
-
- bool rknown = false;
-
- /* If the register representation of a non-scalar mode has holes in it,
- we expect the scalar units to be concatenated together, with the holes
- distributed evenly among the scalar units. Each scalar unit must occupy
- at least one register. */
- if (HARD_REGNO_NREGS_HAS_PADDING (xregno, xmode))
- {
- /* As a consequence, we must be dealing with a constant number of
- scalars, and thus a constant offset and number of units. */
- HOST_WIDE_INT coffset = offset.to_constant ();
- HOST_WIDE_INT cysize = ysize.to_constant ();
- nregs_xmode = HARD_REGNO_NREGS_WITH_PADDING (xregno, xmode);
- unsigned int nunits = GET_MODE_NUNITS (xmode).to_constant ();
- scalar_mode xmode_unit = GET_MODE_INNER (xmode);
- gcc_assert (HARD_REGNO_NREGS_HAS_PADDING (xregno, xmode_unit));
- gcc_assert (nregs_xmode
- == (nunits
- * HARD_REGNO_NREGS_WITH_PADDING (xregno, xmode_unit)));
- gcc_assert (hard_regno_nregs (xregno, xmode)
- == hard_regno_nregs (xregno, xmode_unit) * nunits);
-
- /* You can only ask for a SUBREG of a value with holes in the middle
- if you don't cross the holes. (Such a SUBREG should be done by
- picking a different register class, or doing it in memory if
- necessary.) An example of a value with holes is XCmode on 32-bit
- x86 with -m128bit-long-double; it's represented in 6 32-bit registers,
- 3 for each part, but in memory it's two 128-bit parts.
- Padding is assumed to be at the end (not necessarily the 'high part')
- of each unit. */
- if ((coffset / GET_MODE_SIZE (xmode_unit) + 1 < nunits)
- && (coffset / GET_MODE_SIZE (xmode_unit)
- != ((coffset + cysize - 1) / GET_MODE_SIZE (xmode_unit))))
- {
- info->representable_p = false;
- rknown = true;
- }
- }
- else
- nregs_xmode = hard_regno_nregs (xregno, xmode);
-
- nregs_ymode = hard_regno_nregs (xregno, ymode);
-
- /* Subreg sizes must be ordered, so that we can tell whether they are
- partial, paradoxical or complete. */
- gcc_checking_assert (ordered_p (xsize, ysize));
-
- /* Paradoxical subregs are otherwise valid. */
- if (!rknown && known_eq (offset, 0U) && maybe_gt (ysize, xsize))
- {
- info->representable_p = true;
- /* If this is a big endian paradoxical subreg, which uses more
- actual hard registers than the original register, we must
- return a negative offset so that we find the proper highpart
- of the register.
-
- We assume that the ordering of registers within a multi-register
- value has a consistent endianness: if bytes and register words
- have different endianness, the hard registers that make up a
- multi-register value must be at least word-sized. */
- if (REG_WORDS_BIG_ENDIAN)
- info->offset = (int) nregs_xmode - (int) nregs_ymode;
- else
- info->offset = 0;
- info->nregs = nregs_ymode;
- return;
- }
-
- /* If registers store different numbers of bits in the different
- modes, we cannot generally form this subreg. */
- poly_uint64 regsize_xmode, regsize_ymode;
- if (!HARD_REGNO_NREGS_HAS_PADDING (xregno, xmode)
- && !HARD_REGNO_NREGS_HAS_PADDING (xregno, ymode)
- && multiple_p (xsize, nregs_xmode, &regsize_xmode)
- && multiple_p (ysize, nregs_ymode, &regsize_ymode))
- {
- if (!rknown
- && ((nregs_ymode > 1 && maybe_gt (regsize_xmode, regsize_ymode))
- || (nregs_xmode > 1 && maybe_gt (regsize_ymode, regsize_xmode))))
- {
- info->representable_p = false;
- if (!can_div_away_from_zero_p (ysize, regsize_xmode, &info->nregs)
- || !can_div_trunc_p (offset, regsize_xmode, &info->offset))
- /* Checked by validate_subreg. We must know at compile time
- which inner registers are being accessed. */
- gcc_unreachable ();
- return;
- }
- /* It's not valid to extract a subreg of mode YMODE at OFFSET that
- would go outside of XMODE. */
- if (!rknown && maybe_gt (ysize + offset, xsize))
- {
- info->representable_p = false;
- info->nregs = nregs_ymode;
- if (!can_div_trunc_p (offset, regsize_xmode, &info->offset))
- /* Checked by validate_subreg. We must know at compile time
- which inner registers are being accessed. */
- gcc_unreachable ();
- return;
- }
- /* Quick exit for the simple and common case of extracting whole
- subregisters from a multiregister value. */
- /* ??? It would be better to integrate this into the code below,
- if we can generalize the concept enough and figure out how
- odd-sized modes can coexist with the other weird cases we support. */
- HOST_WIDE_INT count;
- if (!rknown
- && WORDS_BIG_ENDIAN == REG_WORDS_BIG_ENDIAN
- && known_eq (regsize_xmode, regsize_ymode)
- && constant_multiple_p (offset, regsize_ymode, &count))
- {
- info->representable_p = true;
- info->nregs = nregs_ymode;
- info->offset = count;
- gcc_assert (info->offset + info->nregs <= (int) nregs_xmode);
- return;
- }
- }
-
- /* Lowpart subregs are otherwise valid. */
- if (!rknown && known_eq (offset, subreg_lowpart_offset (ymode, xmode)))
- {
- info->representable_p = true;
- rknown = true;
-
- if (known_eq (offset, 0U) || nregs_xmode == nregs_ymode)
- {
- info->offset = 0;
- info->nregs = nregs_ymode;
- return;
- }
- }
-
- /* Set NUM_BLOCKS to the number of independently-representable YMODE
- values there are in (reg:XMODE XREGNO). We can view the register
- as consisting of this number of independent "blocks", where each
- block occupies NREGS_YMODE registers and contains exactly one
- representable YMODE value. */
- gcc_assert ((nregs_xmode % nregs_ymode) == 0);
- unsigned int num_blocks = nregs_xmode / nregs_ymode;
-
- /* Calculate the number of bytes in each block. This must always
- be exact, otherwise we don't know how to verify the constraint.
- These conditions may be relaxed but subreg_regno_offset would
- need to be redesigned. */
- poly_uint64 bytes_per_block = exact_div (xsize, num_blocks);
-
- /* Get the number of the first block that contains the subreg and the byte
- offset of the subreg from the start of that block. */
- unsigned int block_number;
- poly_uint64 subblock_offset;
- if (!can_div_trunc_p (offset, bytes_per_block, &block_number,
- &subblock_offset))
- /* Checked by validate_subreg. We must know at compile time which
- inner registers are being accessed. */
- gcc_unreachable ();
-
- if (!rknown)
- {
- /* Only the lowpart of each block is representable. */
- info->representable_p
- = known_eq (subblock_offset,
- subreg_size_lowpart_offset (ysize, bytes_per_block));
- rknown = true;
- }
-
- /* We assume that the ordering of registers within a multi-register
- value has a consistent endianness: if bytes and register words
- have different endianness, the hard registers that make up a
- multi-register value must be at least word-sized. */
- if (WORDS_BIG_ENDIAN != REG_WORDS_BIG_ENDIAN)
- /* The block number we calculated above followed memory endianness.
- Convert it to register endianness by counting back from the end.
- (Note that, because of the assumption above, each block must be
- at least word-sized.) */
- info->offset = (num_blocks - block_number - 1) * nregs_ymode;
- else
- info->offset = block_number * nregs_ymode;
- info->nregs = nregs_ymode;
-}
-
-/* This function returns the regno offset of a subreg expression.
- xregno - A regno of an inner hard subreg_reg (or what will become one).
- xmode - The mode of xregno.
- offset - The byte offset.
- ymode - The mode of a top level SUBREG (or what may become one).
- RETURN - The regno offset which would be used. */
-unsigned int
-subreg_regno_offset (unsigned int xregno, machine_mode xmode,
- poly_uint64 offset, machine_mode ymode)
-{
- struct subreg_info info;
- subreg_get_info (xregno, xmode, offset, ymode, &info);
- return info.offset;
-}
-
-/* This function returns true when the offset is representable via
- subreg_offset in the given regno.
- xregno - A regno of an inner hard subreg_reg (or what will become one).
- xmode - The mode of xregno.
- offset - The byte offset.
- ymode - The mode of a top level SUBREG (or what may become one).
- RETURN - Whether the offset is representable. */
-bool
-subreg_offset_representable_p (unsigned int xregno, machine_mode xmode,
- poly_uint64 offset, machine_mode ymode)
-{
- struct subreg_info info;
- subreg_get_info (xregno, xmode, offset, ymode, &info);
- return info.representable_p;
-}
-
-/* Return the number of a YMODE register to which
-
- (subreg:YMODE (reg:XMODE XREGNO) OFFSET)
-
- can be simplified. Return -1 if the subreg can't be simplified.
-
- XREGNO is a hard register number. */
-
-int
-simplify_subreg_regno (unsigned int xregno, machine_mode xmode,
- poly_uint64 offset, machine_mode ymode)
-{
- struct subreg_info info;
- unsigned int yregno;
-
- /* Give the backend a chance to disallow the mode change. */
- if (GET_MODE_CLASS (xmode) != MODE_COMPLEX_INT
- && GET_MODE_CLASS (xmode) != MODE_COMPLEX_FLOAT
- && !REG_CAN_CHANGE_MODE_P (xregno, xmode, ymode))
- return -1;
-
- /* We shouldn't simplify stack-related registers. */
- if ((!reload_completed || frame_pointer_needed)
- && xregno == FRAME_POINTER_REGNUM)
- return -1;
-
- if (FRAME_POINTER_REGNUM != ARG_POINTER_REGNUM
- && xregno == ARG_POINTER_REGNUM)
- return -1;
-
- if (xregno == STACK_POINTER_REGNUM
- /* We should convert hard stack register in LRA if it is
- possible. */
- && ! lra_in_progress)
- return -1;
-
- /* Try to get the register offset. */
- subreg_get_info (xregno, xmode, offset, ymode, &info);
- if (!info.representable_p)
- return -1;
-
- /* Make sure that the offsetted register value is in range. */
- yregno = xregno + info.offset;
- if (!HARD_REGISTER_NUM_P (yregno))
- return -1;
-
- /* See whether (reg:YMODE YREGNO) is valid.
-
- ??? We allow invalid registers if (reg:XMODE XREGNO) is also invalid.
- This is a kludge to work around how complex FP arguments are passed
- on IA-64 and should be fixed. See PR target/49226. */
- if (!targetm.hard_regno_mode_ok (yregno, ymode)
- && targetm.hard_regno_mode_ok (xregno, xmode))
- return -1;
-
- return (int) yregno;
-}
-
-/* A wrapper around simplify_subreg_regno that uses subreg_lowpart_offset
- (xmode, ymode) as the offset. */
-
-int
-lowpart_subreg_regno (unsigned int regno, machine_mode xmode,
- machine_mode ymode)
-{
- poly_uint64 offset = subreg_lowpart_offset (xmode, ymode);
- return simplify_subreg_regno (regno, xmode, offset, ymode);
-}
-
-/* Return the final regno that a subreg expression refers to. */
-unsigned int
-subreg_regno (const_rtx x)
-{
- unsigned int ret;
- rtx subreg = SUBREG_REG (x);
- int regno = REGNO (subreg);
-
- ret = regno + subreg_regno_offset (regno,
- GET_MODE (subreg),
- SUBREG_BYTE (x),
- GET_MODE (x));
- return ret;
-
-}
-
-/* Return the number of registers that a subreg expression refers
- to. */
-unsigned int
-subreg_nregs (const_rtx x)
-{
- return subreg_nregs_with_regno (REGNO (SUBREG_REG (x)), x);
-}
-
-/* Return the number of registers that a subreg REG with REGNO
- expression refers to. This is a copy of the rtlanal.c:subreg_nregs
- changed so that the regno can be passed in. */
-
-unsigned int
-subreg_nregs_with_regno (unsigned int regno, const_rtx x)
-{
- struct subreg_info info;
- rtx subreg = SUBREG_REG (x);
-
- subreg_get_info (regno, GET_MODE (subreg), SUBREG_BYTE (x), GET_MODE (x),
- &info);
- return info.nregs;
-}
-
-struct parms_set_data
-{
- int nregs;
- HARD_REG_SET regs;
-};
-
-/* Helper function for noticing stores to parameter registers. */
-static void
-parms_set (rtx x, const_rtx pat ATTRIBUTE_UNUSED, void *data)
-{
- struct parms_set_data *const d = (struct parms_set_data *) data;
- if (REG_P (x) && REGNO (x) < FIRST_PSEUDO_REGISTER
- && TEST_HARD_REG_BIT (d->regs, REGNO (x)))
- {
- CLEAR_HARD_REG_BIT (d->regs, REGNO (x));
- d->nregs--;
- }
-}
-
-/* Look backward for first parameter to be loaded.
- Note that loads of all parameters will not necessarily be
- found if CSE has eliminated some of them (e.g., an argument
- to the outer function is passed down as a parameter).
- Do not skip BOUNDARY. */
-rtx_insn *
-find_first_parameter_load (rtx_insn *call_insn, rtx_insn *boundary)
-{
- struct parms_set_data parm;
- rtx p;
- rtx_insn *before, *first_set;
-
- /* Since different machines initialize their parameter registers
- in different orders, assume nothing. Collect the set of all
- parameter registers. */
- CLEAR_HARD_REG_SET (parm.regs);
- parm.nregs = 0;
- for (p = CALL_INSN_FUNCTION_USAGE (call_insn); p; p = XEXP (p, 1))
- if (GET_CODE (XEXP (p, 0)) == USE
- && REG_P (XEXP (XEXP (p, 0), 0))
- && !STATIC_CHAIN_REG_P (XEXP (XEXP (p, 0), 0)))
- {
- gcc_assert (REGNO (XEXP (XEXP (p, 0), 0)) < FIRST_PSEUDO_REGISTER);
-
- /* We only care about registers which can hold function
- arguments. */
- if (!FUNCTION_ARG_REGNO_P (REGNO (XEXP (XEXP (p, 0), 0))))
- continue;
-
- SET_HARD_REG_BIT (parm.regs, REGNO (XEXP (XEXP (p, 0), 0)));
- parm.nregs++;
- }
- before = call_insn;
- first_set = call_insn;
-
- /* Search backward for the first set of a register in this set. */
- while (parm.nregs && before != boundary)
- {
- before = PREV_INSN (before);
-
- /* It is possible that some loads got CSEed from one call to
- another. Stop in that case. */
- if (CALL_P (before))
- break;
-
- /* Our caller needs either ensure that we will find all sets
- (in case code has not been optimized yet), or take care
- for possible labels in a way by setting boundary to preceding
- CODE_LABEL. */
- if (LABEL_P (before))
- {
- gcc_assert (before == boundary);
- break;
- }
-
- if (INSN_P (before))
- {
- int nregs_old = parm.nregs;
- note_stores (before, parms_set, &parm);
- /* If we found something that did not set a parameter reg,
- we're done. Do not keep going, as that might result
- in hoisting an insn before the setting of a pseudo
- that is used by the hoisted insn. */
- if (nregs_old != parm.nregs)
- first_set = before;
- else
- break;
- }
- }
- return first_set;
-}
-
-/* Return true if we should avoid inserting code between INSN and preceding
- call instruction. */
-
-bool
-keep_with_call_p (const rtx_insn *insn)
-{
- rtx set;
-
- if (INSN_P (insn) && (set = single_set (insn)) != NULL)
- {
- if (REG_P (SET_DEST (set))
- && REGNO (SET_DEST (set)) < FIRST_PSEUDO_REGISTER
- && fixed_regs[REGNO (SET_DEST (set))]
- && general_operand (SET_SRC (set), VOIDmode))
- return true;
- if (REG_P (SET_SRC (set))
- && targetm.calls.function_value_regno_p (REGNO (SET_SRC (set)))
- && REG_P (SET_DEST (set))
- && REGNO (SET_DEST (set)) >= FIRST_PSEUDO_REGISTER)
- return true;
- /* There may be a stack pop just after the call and before the store
- of the return register. Search for the actual store when deciding
- if we can break or not. */
- if (SET_DEST (set) == stack_pointer_rtx)
- {
- /* This CONST_CAST is okay because next_nonnote_insn just
- returns its argument and we assign it to a const_rtx
- variable. */
- const rtx_insn *i2
- = next_nonnote_insn (const_cast<rtx_insn *> (insn));
- if (i2 && keep_with_call_p (i2))
- return true;
- }
- }
- return false;
-}
-
-/* Return true if LABEL is a target of JUMP_INSN. This applies only
- to non-complex jumps. That is, direct unconditional, conditional,
- and tablejumps, but not computed jumps or returns. It also does
- not apply to the fallthru case of a conditional jump. */
-
-bool
-label_is_jump_target_p (const_rtx label, const rtx_insn *jump_insn)
-{
- rtx tmp = JUMP_LABEL (jump_insn);
- rtx_jump_table_data *table;
-
- if (label == tmp)
- return true;
-
- if (tablejump_p (jump_insn, NULL, &table))
- {
- rtvec vec = table->get_labels ();
- int i, veclen = GET_NUM_ELEM (vec);
-
- for (i = 0; i < veclen; ++i)
- if (XEXP (RTVEC_ELT (vec, i), 0) == label)
- return true;
- }
-
- if (find_reg_note (jump_insn, REG_LABEL_TARGET, label))
- return true;
-
- return false;
-}
-
-
-/* Return an estimate of the cost of computing rtx X.
- One use is in cse, to decide which expression to keep in the hash table.
- Another is in rtl generation, to pick the cheapest way to multiply.
- Other uses like the latter are expected in the future.
-
- X appears as operand OPNO in an expression with code OUTER_CODE.
- SPEED specifies whether costs optimized for speed or size should
- be returned. */
-
-int
-rtx_cost (rtx x, machine_mode mode, enum rtx_code outer_code,
- int opno, bool speed)
-{
- int i, j;
- enum rtx_code code;
- const char *fmt;
- int total;
- int factor;
- unsigned mode_size;
-
- if (x == 0)
- return 0;
-
- if (GET_CODE (x) == SET)
- /* A SET doesn't have a mode, so let's look at the SET_DEST to get
- the mode for the factor. */
- mode = GET_MODE (SET_DEST (x));
- else if (GET_MODE (x) != VOIDmode)
- mode = GET_MODE (x);
-
- mode_size = estimated_poly_value (GET_MODE_SIZE (mode));
-
- /* A size N times larger than UNITS_PER_WORD likely needs N times as
- many insns, taking N times as long. */
- factor = mode_size > UNITS_PER_WORD ? mode_size / UNITS_PER_WORD : 1;
-
- /* Compute the default costs of certain things.
- Note that targetm.rtx_costs can override the defaults. */
-
- code = GET_CODE (x);
- switch (code)
- {
- case MULT:
- /* Multiplication has time-complexity O(N*N), where N is the
- number of units (translated from digits) when using
- schoolbook long multiplication. */
- total = factor * factor * COSTS_N_INSNS (5);
- break;
- case DIV:
- case UDIV:
- case MOD:
- case UMOD:
- /* Similarly, complexity for schoolbook long division. */
- total = factor * factor * COSTS_N_INSNS (7);
- break;
- case USE:
- /* Used in combine.c as a marker. */
- total = 0;
- break;
- default:
- total = factor * COSTS_N_INSNS (1);
- }
-
- switch (code)
- {
- case REG:
- return 0;
-
- case SUBREG:
- total = 0;
- /* If we can't tie these modes, make this expensive. The larger
- the mode, the more expensive it is. */
- if (!targetm.modes_tieable_p (mode, GET_MODE (SUBREG_REG (x))))
- return COSTS_N_INSNS (2 + factor);
- break;
-
- case TRUNCATE:
- if (targetm.modes_tieable_p (mode, GET_MODE (XEXP (x, 0))))
- {
- total = 0;
- break;
- }
- /* FALLTHRU */
- default:
- if (targetm.rtx_costs (x, mode, outer_code, opno, &total, speed))
- return total;
- break;
- }
-
- /* Sum the costs of the sub-rtx's, plus cost of this operation,
- which is already in total. */
-
- fmt = GET_RTX_FORMAT (code);
- for (i = GET_RTX_LENGTH (code) - 1; i >= 0; i--)
- if (fmt[i] == 'e')
- total += rtx_cost (XEXP (x, i), mode, code, i, speed);
- else if (fmt[i] == 'E')
- for (j = 0; j < XVECLEN (x, i); j++)
- total += rtx_cost (XVECEXP (x, i, j), mode, code, i, speed);
-
- return total;
-}
-
-/* Fill in the structure C with information about both speed and size rtx
- costs for X, which is operand OPNO in an expression with code OUTER. */
-
-void
-get_full_rtx_cost (rtx x, machine_mode mode, enum rtx_code outer, int opno,
- struct full_rtx_costs *c)
-{
- c->speed = rtx_cost (x, mode, outer, opno, true);
- c->size = rtx_cost (x, mode, outer, opno, false);
-}
-
-
-/* Return cost of address expression X.
- Expect that X is properly formed address reference.
-
- SPEED parameter specify whether costs optimized for speed or size should
- be returned. */
-
-int
-address_cost (rtx x, machine_mode mode, addr_space_t as, bool speed)
-{
- /* We may be asked for cost of various unusual addresses, such as operands
- of push instruction. It is not worthwhile to complicate writing
- of the target hook by such cases. */
-
- if (!memory_address_addr_space_p (mode, x, as))
- return 1000;
-
- return targetm.address_cost (x, mode, as, speed);
-}
-
-/* If the target doesn't override, compute the cost as with arithmetic. */
-
-int
-default_address_cost (rtx x, machine_mode, addr_space_t, bool speed)
-{
- return rtx_cost (x, Pmode, MEM, 0, speed);
-}
-
-
-unsigned HOST_WIDE_INT
-nonzero_bits (const_rtx x, machine_mode mode)
-{
- if (mode == VOIDmode)
- mode = GET_MODE (x);
- scalar_int_mode int_mode;
- if (!is_a <scalar_int_mode> (mode, &int_mode))
- return GET_MODE_MASK (mode);
- return cached_nonzero_bits (x, int_mode, NULL_RTX, VOIDmode, 0);
-}
-
-unsigned int
-num_sign_bit_copies (const_rtx x, machine_mode mode)
-{
- if (mode == VOIDmode)
- mode = GET_MODE (x);
- scalar_int_mode int_mode;
- if (!is_a <scalar_int_mode> (mode, &int_mode))
- return 1;
- return cached_num_sign_bit_copies (x, int_mode, NULL_RTX, VOIDmode, 0);
-}
-
-/* Return true if nonzero_bits1 might recurse into both operands
- of X. */
-
-static inline bool
-nonzero_bits_binary_arith_p (const_rtx x)
-{
- if (!ARITHMETIC_P (x))
- return false;
- switch (GET_CODE (x))
- {
- case AND:
- case XOR:
- case IOR:
- case UMIN:
- case UMAX:
- case SMIN:
- case SMAX:
- case PLUS:
- case MINUS:
- case MULT:
- case DIV:
- case UDIV:
- case MOD:
- case UMOD:
- return true;
- default:
- return false;
- }
-}
-
-/* The function cached_nonzero_bits is a wrapper around nonzero_bits1.
- It avoids exponential behavior in nonzero_bits1 when X has
- identical subexpressions on the first or the second level. */
-
-static unsigned HOST_WIDE_INT
-cached_nonzero_bits (const_rtx x, scalar_int_mode mode, const_rtx known_x,
- machine_mode known_mode,
- unsigned HOST_WIDE_INT known_ret)
-{
- if (x == known_x && mode == known_mode)
- return known_ret;
-
- /* Try to find identical subexpressions. If found call
- nonzero_bits1 on X with the subexpressions as KNOWN_X and the
- precomputed value for the subexpression as KNOWN_RET. */
-
- if (nonzero_bits_binary_arith_p (x))
- {
- rtx x0 = XEXP (x, 0);
- rtx x1 = XEXP (x, 1);
-
- /* Check the first level. */
- if (x0 == x1)
- return nonzero_bits1 (x, mode, x0, mode,
- cached_nonzero_bits (x0, mode, known_x,
- known_mode, known_ret));
-
- /* Check the second level. */
- if (nonzero_bits_binary_arith_p (x0)
- && (x1 == XEXP (x0, 0) || x1 == XEXP (x0, 1)))
- return nonzero_bits1 (x, mode, x1, mode,
- cached_nonzero_bits (x1, mode, known_x,
- known_mode, known_ret));
-
- if (nonzero_bits_binary_arith_p (x1)
- && (x0 == XEXP (x1, 0) || x0 == XEXP (x1, 1)))
- return nonzero_bits1 (x, mode, x0, mode,
- cached_nonzero_bits (x0, mode, known_x,
- known_mode, known_ret));
- }
-
- return nonzero_bits1 (x, mode, known_x, known_mode, known_ret);
-}
-
-/* We let num_sign_bit_copies recur into nonzero_bits as that is useful.
- We don't let nonzero_bits recur into num_sign_bit_copies, because that
- is less useful. We can't allow both, because that results in exponential
- run time recursion. There is a nullstone testcase that triggered
- this. This macro avoids accidental uses of num_sign_bit_copies. */
-#define cached_num_sign_bit_copies sorry_i_am_preventing_exponential_behavior
-
-/* Given an expression, X, compute which bits in X can be nonzero.
- We don't care about bits outside of those defined in MODE.
-
- For most X this is simply GET_MODE_MASK (GET_MODE (X)), but if X is
- an arithmetic operation, we can do better. */
-
-static unsigned HOST_WIDE_INT
-nonzero_bits1 (const_rtx x, scalar_int_mode mode, const_rtx known_x,
- machine_mode known_mode,
- unsigned HOST_WIDE_INT known_ret)
-{
- unsigned HOST_WIDE_INT nonzero = GET_MODE_MASK (mode);
- unsigned HOST_WIDE_INT inner_nz;
- enum rtx_code code = GET_CODE (x);
- machine_mode inner_mode;
- unsigned int inner_width;
- scalar_int_mode xmode;
-
- unsigned int mode_width = GET_MODE_PRECISION (mode);
-
- if (CONST_INT_P (x))
- {
- if (SHORT_IMMEDIATES_SIGN_EXTEND
- && INTVAL (x) > 0
- && mode_width < BITS_PER_WORD
- && (UINTVAL (x) & (HOST_WIDE_INT_1U << (mode_width - 1))) != 0)
- return UINTVAL (x) | (HOST_WIDE_INT_M1U << mode_width);
-
- return UINTVAL (x);
- }
-
- if (!is_a <scalar_int_mode> (GET_MODE (x), &xmode))
- return nonzero;
- unsigned int xmode_width = GET_MODE_PRECISION (xmode);
-
- /* If X is wider than MODE, use its mode instead. */
- if (xmode_width > mode_width)
- {
- mode = xmode;
- nonzero = GET_MODE_MASK (mode);
- mode_width = xmode_width;
- }
-
- if (mode_width > HOST_BITS_PER_WIDE_INT)
- /* Our only callers in this case look for single bit values. So
- just return the mode mask. Those tests will then be false. */
- return nonzero;
-
- /* If MODE is wider than X, but both are a single word for both the host
- and target machines, we can compute this from which bits of the object
- might be nonzero in its own mode, taking into account the fact that, on
- CISC machines, accessing an object in a wider mode generally causes the
- high-order bits to become undefined, so they are not known to be zero.
- We extend this reasoning to RISC machines for operations that might not
- operate on the full registers. */
- if (mode_width > xmode_width
- && xmode_width <= BITS_PER_WORD
- && xmode_width <= HOST_BITS_PER_WIDE_INT
- && !(WORD_REGISTER_OPERATIONS && word_register_operation_p (x)))
- {
- nonzero &= cached_nonzero_bits (x, xmode,
- known_x, known_mode, known_ret);
- nonzero |= GET_MODE_MASK (mode) & ~GET_MODE_MASK (xmode);
- return nonzero;
- }
-
- /* Please keep nonzero_bits_binary_arith_p above in sync with
- the code in the switch below. */
- switch (code)
- {
- case REG:
-#if defined(POINTERS_EXTEND_UNSIGNED)
- /* If pointers extend unsigned and this is a pointer in Pmode, say that
- all the bits above ptr_mode are known to be zero. */
- /* As we do not know which address space the pointer is referring to,
- we can do this only if the target does not support different pointer
- or address modes depending on the address space. */
- if (target_default_pointer_address_modes_p ()
- && POINTERS_EXTEND_UNSIGNED
- && xmode == Pmode
- && REG_POINTER (x)
- && !targetm.have_ptr_extend ())
- nonzero &= GET_MODE_MASK (ptr_mode);
-#endif
-
- /* Include declared information about alignment of pointers. */
- /* ??? We don't properly preserve REG_POINTER changes across
- pointer-to-integer casts, so we can't trust it except for
- things that we know must be pointers. See execute/960116-1.c. */
- if ((x == stack_pointer_rtx
- || x == frame_pointer_rtx
- || x == arg_pointer_rtx)
- && REGNO_POINTER_ALIGN (REGNO (x)))
- {
- unsigned HOST_WIDE_INT alignment
- = REGNO_POINTER_ALIGN (REGNO (x)) / BITS_PER_UNIT;
-
-#ifdef PUSH_ROUNDING
- /* If PUSH_ROUNDING is defined, it is possible for the
- stack to be momentarily aligned only to that amount,
- so we pick the least alignment. */
- if (x == stack_pointer_rtx && targetm.calls.push_argument (0))
- {
- poly_uint64 rounded_1 = PUSH_ROUNDING (poly_int64 (1));
- alignment = MIN (known_alignment (rounded_1), alignment);
- }
-#endif
-
- nonzero &= ~(alignment - 1);
- }
-
- {
- unsigned HOST_WIDE_INT nonzero_for_hook = nonzero;
- rtx new_rtx = rtl_hooks.reg_nonzero_bits (x, xmode, mode,
- &nonzero_for_hook);
-
- if (new_rtx)
- nonzero_for_hook &= cached_nonzero_bits (new_rtx, mode, known_x,
- known_mode, known_ret);
-
- return nonzero_for_hook;
- }
-
- case MEM:
- /* In many, if not most, RISC machines, reading a byte from memory
- zeros the rest of the register. Noticing that fact saves a lot
- of extra zero-extends. */
- if (load_extend_op (xmode) == ZERO_EXTEND)
- nonzero &= GET_MODE_MASK (xmode);
- break;
-
- case EQ: case NE:
- case UNEQ: case LTGT:
- case GT: case GTU: case UNGT:
- case LT: case LTU: case UNLT:
- case GE: case GEU: case UNGE:
- case LE: case LEU: case UNLE:
- case UNORDERED: case ORDERED:
- /* If this produces an integer result, we know which bits are set.
- Code here used to clear bits outside the mode of X, but that is
- now done above. */
- /* Mind that MODE is the mode the caller wants to look at this
- operation in, and not the actual operation mode. We can wind
- up with (subreg:DI (gt:V4HI x y)), and we don't have anything
- that describes the results of a vector compare. */
- if (GET_MODE_CLASS (xmode) == MODE_INT
- && mode_width <= HOST_BITS_PER_WIDE_INT)
- nonzero = STORE_FLAG_VALUE;
- break;
-
- case NEG:
-#if 0
- /* Disabled to avoid exponential mutual recursion between nonzero_bits
- and num_sign_bit_copies. */
- if (num_sign_bit_copies (XEXP (x, 0), xmode) == xmode_width)
- nonzero = 1;
-#endif
-
- if (xmode_width < mode_width)
- nonzero |= (GET_MODE_MASK (mode) & ~GET_MODE_MASK (xmode));
- break;
-
- case ABS:
-#if 0
- /* Disabled to avoid exponential mutual recursion between nonzero_bits
- and num_sign_bit_copies. */
- if (num_sign_bit_copies (XEXP (x, 0), xmode) == xmode_width)
- nonzero = 1;
-#endif
- break;
-
- case TRUNCATE:
- nonzero &= (cached_nonzero_bits (XEXP (x, 0), mode,
- known_x, known_mode, known_ret)
- & GET_MODE_MASK (mode));
- break;
-
- case ZERO_EXTEND:
- nonzero &= cached_nonzero_bits (XEXP (x, 0), mode,
- known_x, known_mode, known_ret);
- if (GET_MODE (XEXP (x, 0)) != VOIDmode)
- nonzero &= GET_MODE_MASK (GET_MODE (XEXP (x, 0)));
- break;
-
- case SIGN_EXTEND:
- /* If the sign bit is known clear, this is the same as ZERO_EXTEND.
- Otherwise, show all the bits in the outer mode but not the inner
- may be nonzero. */
- inner_nz = cached_nonzero_bits (XEXP (x, 0), mode,
- known_x, known_mode, known_ret);
- if (GET_MODE (XEXP (x, 0)) != VOIDmode)
- {
- inner_nz &= GET_MODE_MASK (GET_MODE (XEXP (x, 0)));
- if (val_signbit_known_set_p (GET_MODE (XEXP (x, 0)), inner_nz))
- inner_nz |= (GET_MODE_MASK (mode)
- & ~GET_MODE_MASK (GET_MODE (XEXP (x, 0))));
- }
-
- nonzero &= inner_nz;
- break;
-
- case AND:
- nonzero &= cached_nonzero_bits (XEXP (x, 0), mode,
- known_x, known_mode, known_ret)
- & cached_nonzero_bits (XEXP (x, 1), mode,
- known_x, known_mode, known_ret);
- break;
-
- case XOR: case IOR:
- case UMIN: case UMAX: case SMIN: case SMAX:
- {
- unsigned HOST_WIDE_INT nonzero0
- = cached_nonzero_bits (XEXP (x, 0), mode,
- known_x, known_mode, known_ret);
-
- /* Don't call nonzero_bits for the second time if it cannot change
- anything. */
- if ((nonzero & nonzero0) != nonzero)
- nonzero &= nonzero0
- | cached_nonzero_bits (XEXP (x, 1), mode,
- known_x, known_mode, known_ret);
- }
- break;
-
- case PLUS: case MINUS:
- case MULT:
- case DIV: case UDIV:
- case MOD: case UMOD:
- /* We can apply the rules of arithmetic to compute the number of
- high- and low-order zero bits of these operations. We start by
- computing the width (position of the highest-order nonzero bit)
- and the number of low-order zero bits for each value. */
- {
- unsigned HOST_WIDE_INT nz0
- = cached_nonzero_bits (XEXP (x, 0), mode,
- known_x, known_mode, known_ret);
- unsigned HOST_WIDE_INT nz1
- = cached_nonzero_bits (XEXP (x, 1), mode,
- known_x, known_mode, known_ret);
- int sign_index = xmode_width - 1;
- int width0 = floor_log2 (nz0) + 1;
- int width1 = floor_log2 (nz1) + 1;
- int low0 = ctz_or_zero (nz0);
- int low1 = ctz_or_zero (nz1);
- unsigned HOST_WIDE_INT op0_maybe_minusp
- = nz0 & (HOST_WIDE_INT_1U << sign_index);
- unsigned HOST_WIDE_INT op1_maybe_minusp
- = nz1 & (HOST_WIDE_INT_1U << sign_index);
- unsigned int result_width = mode_width;
- int result_low = 0;
-
- switch (code)
- {
- case PLUS:
- result_width = MAX (width0, width1) + 1;
- result_low = MIN (low0, low1);
- break;
- case MINUS:
- result_low = MIN (low0, low1);
- break;
- case MULT:
- result_width = width0 + width1;
- result_low = low0 + low1;
- break;
- case DIV:
- if (width1 == 0)
- break;
- if (!op0_maybe_minusp && !op1_maybe_minusp)
- result_width = width0;
- break;
- case UDIV:
- if (width1 == 0)
- break;
- result_width = width0;
- break;
- case MOD:
- if (width1 == 0)
- break;
- if (!op0_maybe_minusp && !op1_maybe_minusp)
- result_width = MIN (width0, width1);
- result_low = MIN (low0, low1);
- break;
- case UMOD:
- if (width1 == 0)
- break;
- result_width = MIN (width0, width1);
- result_low = MIN (low0, low1);
- break;
- default:
- gcc_unreachable ();
- }
-
- /* Note that mode_width <= HOST_BITS_PER_WIDE_INT, see above. */
- if (result_width < mode_width)
- nonzero &= (HOST_WIDE_INT_1U << result_width) - 1;
-
- if (result_low > 0)
- {
- if (result_low < HOST_BITS_PER_WIDE_INT)
- nonzero &= ~((HOST_WIDE_INT_1U << result_low) - 1);
- else
- nonzero = 0;
- }
- }
- break;
-
- case ZERO_EXTRACT:
- if (CONST_INT_P (XEXP (x, 1))
- && INTVAL (XEXP (x, 1)) < HOST_BITS_PER_WIDE_INT)
- nonzero &= (HOST_WIDE_INT_1U << INTVAL (XEXP (x, 1))) - 1;
- break;
-
- case SUBREG:
- /* If this is a SUBREG formed for a promoted variable that has
- been zero-extended, we know that at least the high-order bits
- are zero, though others might be too. */
- if (SUBREG_PROMOTED_VAR_P (x) && SUBREG_PROMOTED_UNSIGNED_P (x))
- nonzero = GET_MODE_MASK (xmode)
- & cached_nonzero_bits (SUBREG_REG (x), xmode,
- known_x, known_mode, known_ret);
-
- /* If the inner mode is a single word for both the host and target
- machines, we can compute this from which bits of the inner
- object might be nonzero. */
- inner_mode = GET_MODE (SUBREG_REG (x));
- if (GET_MODE_PRECISION (inner_mode).is_constant (&inner_width)
- && inner_width <= BITS_PER_WORD
- && inner_width <= HOST_BITS_PER_WIDE_INT)
- {
- nonzero &= cached_nonzero_bits (SUBREG_REG (x), mode,
- known_x, known_mode, known_ret);
-
- /* On a typical CISC machine, accessing an object in a wider mode
- causes the high-order bits to become undefined. So they are
- not known to be zero.
-
- On a typical RISC machine, we only have to worry about the way
- loads are extended. Otherwise, if we get a reload for the inner
- part, it may be loaded from the stack, and then we may lose all
- the zero bits that existed before the store to the stack. */
- rtx_code extend_op;
- if ((!WORD_REGISTER_OPERATIONS
- || ((extend_op = load_extend_op (inner_mode)) == SIGN_EXTEND
- ? val_signbit_known_set_p (inner_mode, nonzero)
- : extend_op != ZERO_EXTEND)
- || !MEM_P (SUBREG_REG (x)))
- && xmode_width > inner_width)
- nonzero
- |= (GET_MODE_MASK (GET_MODE (x)) & ~GET_MODE_MASK (inner_mode));
- }
- break;
-
- case ASHIFT:
- case ASHIFTRT:
- case LSHIFTRT:
- case ROTATE:
- case ROTATERT:
- /* The nonzero bits are in two classes: any bits within MODE
- that aren't in xmode are always significant. The rest of the
- nonzero bits are those that are significant in the operand of
- the shift when shifted the appropriate number of bits. This
- shows that high-order bits are cleared by the right shift and
- low-order bits by left shifts. */
- if (CONST_INT_P (XEXP (x, 1))
- && INTVAL (XEXP (x, 1)) >= 0
- && INTVAL (XEXP (x, 1)) < HOST_BITS_PER_WIDE_INT
- && INTVAL (XEXP (x, 1)) < xmode_width)
- {
- int count = INTVAL (XEXP (x, 1));
- unsigned HOST_WIDE_INT mode_mask = GET_MODE_MASK (xmode);
- unsigned HOST_WIDE_INT op_nonzero
- = cached_nonzero_bits (XEXP (x, 0), mode,
- known_x, known_mode, known_ret);
- unsigned HOST_WIDE_INT inner = op_nonzero & mode_mask;
- unsigned HOST_WIDE_INT outer = 0;
-
- if (mode_width > xmode_width)
- outer = (op_nonzero & nonzero & ~mode_mask);
-
- switch (code)
- {
- case ASHIFT:
- inner <<= count;
- break;
-
- case LSHIFTRT:
- inner >>= count;
- break;
-
- case ASHIFTRT:
- inner >>= count;
-
- /* If the sign bit may have been nonzero before the shift, we
- need to mark all the places it could have been copied to
- by the shift as possibly nonzero. */
- if (inner & (HOST_WIDE_INT_1U << (xmode_width - 1 - count)))
- inner |= (((HOST_WIDE_INT_1U << count) - 1)
- << (xmode_width - count));
- break;
-
- case ROTATE:
- inner = (inner << (count % xmode_width)
- | (inner >> (xmode_width - (count % xmode_width))))
- & mode_mask;
- break;
-
- case ROTATERT:
- inner = (inner >> (count % xmode_width)
- | (inner << (xmode_width - (count % xmode_width))))
- & mode_mask;
- break;
-
- default:
- gcc_unreachable ();
- }
-
- nonzero &= (outer | inner);
- }
- break;
-
- case FFS:
- case POPCOUNT:
- /* This is at most the number of bits in the mode. */
- nonzero = ((unsigned HOST_WIDE_INT) 2 << (floor_log2 (mode_width))) - 1;
- break;
-
- case CLZ:
- /* If CLZ has a known value at zero, then the nonzero bits are
- that value, plus the number of bits in the mode minus one. */
- if (CLZ_DEFINED_VALUE_AT_ZERO (mode, nonzero))
- nonzero
- |= (HOST_WIDE_INT_1U << (floor_log2 (mode_width))) - 1;
- else
- nonzero = -1;
- break;
-
- case CTZ:
- /* If CTZ has a known value at zero, then the nonzero bits are
- that value, plus the number of bits in the mode minus one. */
- if (CTZ_DEFINED_VALUE_AT_ZERO (mode, nonzero))
- nonzero
- |= (HOST_WIDE_INT_1U << (floor_log2 (mode_width))) - 1;
- else
- nonzero = -1;
- break;
-
- case CLRSB:
- /* This is at most the number of bits in the mode minus 1. */
- nonzero = (HOST_WIDE_INT_1U << (floor_log2 (mode_width))) - 1;
- break;
-
- case PARITY:
- nonzero = 1;
- break;
-
- case IF_THEN_ELSE:
- {
- unsigned HOST_WIDE_INT nonzero_true
- = cached_nonzero_bits (XEXP (x, 1), mode,
- known_x, known_mode, known_ret);
-
- /* Don't call nonzero_bits for the second time if it cannot change
- anything. */
- if ((nonzero & nonzero_true) != nonzero)
- nonzero &= nonzero_true
- | cached_nonzero_bits (XEXP (x, 2), mode,
- known_x, known_mode, known_ret);
- }
- break;
-
- default:
- break;
- }
-
- return nonzero;
-}
-
-/* See the macro definition above. */
-#undef cached_num_sign_bit_copies
-
-
-/* Return true if num_sign_bit_copies1 might recurse into both operands
- of X. */
-
-static inline bool
-num_sign_bit_copies_binary_arith_p (const_rtx x)
-{
- if (!ARITHMETIC_P (x))
- return false;
- switch (GET_CODE (x))
- {
- case IOR:
- case AND:
- case XOR:
- case SMIN:
- case SMAX:
- case UMIN:
- case UMAX:
- case PLUS:
- case MINUS:
- case MULT:
- return true;
- default:
- return false;
- }
-}
-
-/* The function cached_num_sign_bit_copies is a wrapper around
- num_sign_bit_copies1. It avoids exponential behavior in
- num_sign_bit_copies1 when X has identical subexpressions on the
- first or the second level. */
-
-static unsigned int
-cached_num_sign_bit_copies (const_rtx x, scalar_int_mode mode,
- const_rtx known_x, machine_mode known_mode,
- unsigned int known_ret)
-{
- if (x == known_x && mode == known_mode)
- return known_ret;
-
- /* Try to find identical subexpressions. If found call
- num_sign_bit_copies1 on X with the subexpressions as KNOWN_X and
- the precomputed value for the subexpression as KNOWN_RET. */
-
- if (num_sign_bit_copies_binary_arith_p (x))
- {
- rtx x0 = XEXP (x, 0);
- rtx x1 = XEXP (x, 1);
-
- /* Check the first level. */
- if (x0 == x1)
- return
- num_sign_bit_copies1 (x, mode, x0, mode,
- cached_num_sign_bit_copies (x0, mode, known_x,
- known_mode,
- known_ret));
-
- /* Check the second level. */
- if (num_sign_bit_copies_binary_arith_p (x0)
- && (x1 == XEXP (x0, 0) || x1 == XEXP (x0, 1)))
- return
- num_sign_bit_copies1 (x, mode, x1, mode,
- cached_num_sign_bit_copies (x1, mode, known_x,
- known_mode,
- known_ret));
-
- if (num_sign_bit_copies_binary_arith_p (x1)
- && (x0 == XEXP (x1, 0) || x0 == XEXP (x1, 1)))
- return
- num_sign_bit_copies1 (x, mode, x0, mode,
- cached_num_sign_bit_copies (x0, mode, known_x,
- known_mode,
- known_ret));
- }
-
- return num_sign_bit_copies1 (x, mode, known_x, known_mode, known_ret);
-}
-
-/* Return the number of bits at the high-order end of X that are known to
- be equal to the sign bit. X will be used in mode MODE. The returned
- value will always be between 1 and the number of bits in MODE. */
-
-static unsigned int
-num_sign_bit_copies1 (const_rtx x, scalar_int_mode mode, const_rtx known_x,
- machine_mode known_mode,
- unsigned int known_ret)
-{
- enum rtx_code code = GET_CODE (x);
- unsigned int bitwidth = GET_MODE_PRECISION (mode);
- int num0, num1, result;
- unsigned HOST_WIDE_INT nonzero;
-
- if (CONST_INT_P (x))
- {
- /* If the constant is negative, take its 1's complement and remask.
- Then see how many zero bits we have. */
- nonzero = UINTVAL (x) & GET_MODE_MASK (mode);
- if (bitwidth <= HOST_BITS_PER_WIDE_INT
- && (nonzero & (HOST_WIDE_INT_1U << (bitwidth - 1))) != 0)
- nonzero = (~nonzero) & GET_MODE_MASK (mode);
-
- return (nonzero == 0 ? bitwidth : bitwidth - floor_log2 (nonzero) - 1);
- }
-
- scalar_int_mode xmode, inner_mode;
- if (!is_a <scalar_int_mode> (GET_MODE (x), &xmode))
- return 1;
-
- unsigned int xmode_width = GET_MODE_PRECISION (xmode);
-
- /* For a smaller mode, just ignore the high bits. */
- if (bitwidth < xmode_width)
- {
- num0 = cached_num_sign_bit_copies (x, xmode,
- known_x, known_mode, known_ret);
- return MAX (1, num0 - (int) (xmode_width - bitwidth));
- }
-
- if (bitwidth > xmode_width)
- {
- /* If this machine does not do all register operations on the entire
- register and MODE is wider than the mode of X, we can say nothing
- at all about the high-order bits. We extend this reasoning to RISC
- machines for operations that might not operate on full registers. */
- if (!(WORD_REGISTER_OPERATIONS && word_register_operation_p (x)))
- return 1;
-
- /* Likewise on machines that do, if the mode of the object is smaller
- than a word and loads of that size don't sign extend, we can say
- nothing about the high order bits. */
- if (xmode_width < BITS_PER_WORD
- && load_extend_op (xmode) != SIGN_EXTEND)
- return 1;
- }
-
- /* Please keep num_sign_bit_copies_binary_arith_p above in sync with
- the code in the switch below. */
- switch (code)
- {
- case REG:
-
-#if defined(POINTERS_EXTEND_UNSIGNED)
- /* If pointers extend signed and this is a pointer in Pmode, say that
- all the bits above ptr_mode are known to be sign bit copies. */
- /* As we do not know which address space the pointer is referring to,
- we can do this only if the target does not support different pointer
- or address modes depending on the address space. */
- if (target_default_pointer_address_modes_p ()
- && ! POINTERS_EXTEND_UNSIGNED && xmode == Pmode
- && mode == Pmode && REG_POINTER (x)
- && !targetm.have_ptr_extend ())
- return GET_MODE_PRECISION (Pmode) - GET_MODE_PRECISION (ptr_mode) + 1;
-#endif
-
- {
- unsigned int copies_for_hook = 1, copies = 1;
- rtx new_rtx = rtl_hooks.reg_num_sign_bit_copies (x, xmode, mode,
- &copies_for_hook);
-
- if (new_rtx)
- copies = cached_num_sign_bit_copies (new_rtx, mode, known_x,
- known_mode, known_ret);
-
- if (copies > 1 || copies_for_hook > 1)
- return MAX (copies, copies_for_hook);
-
- /* Else, use nonzero_bits to guess num_sign_bit_copies (see below). */
- }
- break;
-
- case MEM:
- /* Some RISC machines sign-extend all loads of smaller than a word. */
- if (load_extend_op (xmode) == SIGN_EXTEND)
- return MAX (1, ((int) bitwidth - (int) xmode_width + 1));
- break;
-
- case SUBREG:
- /* If this is a SUBREG for a promoted object that is sign-extended
- and we are looking at it in a wider mode, we know that at least the
- high-order bits are known to be sign bit copies. */
-
- if (SUBREG_PROMOTED_VAR_P (x) && SUBREG_PROMOTED_SIGNED_P (x))
- {
- num0 = cached_num_sign_bit_copies (SUBREG_REG (x), mode,
- known_x, known_mode, known_ret);
- return MAX ((int) bitwidth - (int) xmode_width + 1, num0);
- }
-
- if (is_a <scalar_int_mode> (GET_MODE (SUBREG_REG (x)), &inner_mode))
- {
- /* For a smaller object, just ignore the high bits. */
- if (bitwidth <= GET_MODE_PRECISION (inner_mode))
- {
- num0 = cached_num_sign_bit_copies (SUBREG_REG (x), inner_mode,
- known_x, known_mode,
- known_ret);
- return MAX (1, num0 - (int) (GET_MODE_PRECISION (inner_mode)
- - bitwidth));
- }
-
- /* For paradoxical SUBREGs on machines where all register operations
- affect the entire register, just look inside. Note that we are
- passing MODE to the recursive call, so the number of sign bit
- copies will remain relative to that mode, not the inner mode.
-
- This works only if loads sign extend. Otherwise, if we get a
- reload for the inner part, it may be loaded from the stack, and
- then we lose all sign bit copies that existed before the store
- to the stack. */
- if (WORD_REGISTER_OPERATIONS
- && load_extend_op (inner_mode) == SIGN_EXTEND
- && paradoxical_subreg_p (x)
- && MEM_P (SUBREG_REG (x)))
- return cached_num_sign_bit_copies (SUBREG_REG (x), mode,
- known_x, known_mode, known_ret);
- }
- break;
-
- case SIGN_EXTRACT:
- if (CONST_INT_P (XEXP (x, 1)))
- return MAX (1, (int) bitwidth - INTVAL (XEXP (x, 1)));
- break;
-
- case SIGN_EXTEND:
- if (is_a <scalar_int_mode> (GET_MODE (XEXP (x, 0)), &inner_mode))
- return (bitwidth - GET_MODE_PRECISION (inner_mode)
- + cached_num_sign_bit_copies (XEXP (x, 0), inner_mode,
- known_x, known_mode, known_ret));
- break;
-
- case TRUNCATE:
- /* For a smaller object, just ignore the high bits. */
- inner_mode = as_a <scalar_int_mode> (GET_MODE (XEXP (x, 0)));
- num0 = cached_num_sign_bit_copies (XEXP (x, 0), inner_mode,
- known_x, known_mode, known_ret);
- return MAX (1, (num0 - (int) (GET_MODE_PRECISION (inner_mode)
- - bitwidth)));
-
- case NOT:
- return cached_num_sign_bit_copies (XEXP (x, 0), mode,
- known_x, known_mode, known_ret);
-
- case ROTATE: case ROTATERT:
- /* If we are rotating left by a number of bits less than the number
- of sign bit copies, we can just subtract that amount from the
- number. */
- if (CONST_INT_P (XEXP (x, 1))
- && INTVAL (XEXP (x, 1)) >= 0
- && INTVAL (XEXP (x, 1)) < (int) bitwidth)
- {
- num0 = cached_num_sign_bit_copies (XEXP (x, 0), mode,
- known_x, known_mode, known_ret);
- return MAX (1, num0 - (code == ROTATE ? INTVAL (XEXP (x, 1))
- : (int) bitwidth - INTVAL (XEXP (x, 1))));
- }
- break;
-
- case NEG:
- /* In general, this subtracts one sign bit copy. But if the value
- is known to be positive, the number of sign bit copies is the
- same as that of the input. Finally, if the input has just one bit
- that might be nonzero, all the bits are copies of the sign bit. */
- num0 = cached_num_sign_bit_copies (XEXP (x, 0), mode,
- known_x, known_mode, known_ret);
- if (bitwidth > HOST_BITS_PER_WIDE_INT)
- return num0 > 1 ? num0 - 1 : 1;
-
- nonzero = nonzero_bits (XEXP (x, 0), mode);
- if (nonzero == 1)
- return bitwidth;
-
- if (num0 > 1
- && ((HOST_WIDE_INT_1U << (bitwidth - 1)) & nonzero))
- num0--;
-
- return num0;
-
- case IOR: case AND: case XOR:
- case SMIN: case SMAX: case UMIN: case UMAX:
- /* Logical operations will preserve the number of sign-bit copies.
- MIN and MAX operations always return one of the operands. */
- num0 = cached_num_sign_bit_copies (XEXP (x, 0), mode,
- known_x, known_mode, known_ret);
- num1 = cached_num_sign_bit_copies (XEXP (x, 1), mode,
- known_x, known_mode, known_ret);
-
- /* If num1 is clearing some of the top bits then regardless of
- the other term, we are guaranteed to have at least that many
- high-order zero bits. */
- if (code == AND
- && num1 > 1
- && bitwidth <= HOST_BITS_PER_WIDE_INT
- && CONST_INT_P (XEXP (x, 1))
- && (UINTVAL (XEXP (x, 1))
- & (HOST_WIDE_INT_1U << (bitwidth - 1))) == 0)
- return num1;
-
- /* Similarly for IOR when setting high-order bits. */
- if (code == IOR
- && num1 > 1
- && bitwidth <= HOST_BITS_PER_WIDE_INT
- && CONST_INT_P (XEXP (x, 1))
- && (UINTVAL (XEXP (x, 1))
- & (HOST_WIDE_INT_1U << (bitwidth - 1))) != 0)
- return num1;
-
- return MIN (num0, num1);
-
- case PLUS: case MINUS:
- /* For addition and subtraction, we can have a 1-bit carry. However,
- if we are subtracting 1 from a positive number, there will not
- be such a carry. Furthermore, if the positive number is known to
- be 0 or 1, we know the result is either -1 or 0. */
-
- if (code == PLUS && XEXP (x, 1) == constm1_rtx
- && bitwidth <= HOST_BITS_PER_WIDE_INT)
- {
- nonzero = nonzero_bits (XEXP (x, 0), mode);
- if (((HOST_WIDE_INT_1U << (bitwidth - 1)) & nonzero) == 0)
- return (nonzero == 1 || nonzero == 0 ? bitwidth
- : bitwidth - floor_log2 (nonzero) - 1);
- }
-
- num0 = cached_num_sign_bit_copies (XEXP (x, 0), mode,
- known_x, known_mode, known_ret);
- num1 = cached_num_sign_bit_copies (XEXP (x, 1), mode,
- known_x, known_mode, known_ret);
- result = MAX (1, MIN (num0, num1) - 1);
-
- return result;
-
- case MULT:
- /* The number of bits of the product is the sum of the number of
- bits of both terms. However, unless one of the terms if known
- to be positive, we must allow for an additional bit since negating
- a negative number can remove one sign bit copy. */
-
- num0 = cached_num_sign_bit_copies (XEXP (x, 0), mode,
- known_x, known_mode, known_ret);
- num1 = cached_num_sign_bit_copies (XEXP (x, 1), mode,
- known_x, known_mode, known_ret);
-
- result = bitwidth - (bitwidth - num0) - (bitwidth - num1);
- if (result > 0
- && (bitwidth > HOST_BITS_PER_WIDE_INT
- || (((nonzero_bits (XEXP (x, 0), mode)
- & (HOST_WIDE_INT_1U << (bitwidth - 1))) != 0)
- && ((nonzero_bits (XEXP (x, 1), mode)
- & (HOST_WIDE_INT_1U << (bitwidth - 1)))
- != 0))))
- result--;
-
- return MAX (1, result);
-
- case UDIV:
- /* The result must be <= the first operand. If the first operand
- has the high bit set, we know nothing about the number of sign
- bit copies. */
- if (bitwidth > HOST_BITS_PER_WIDE_INT)
- return 1;
- else if ((nonzero_bits (XEXP (x, 0), mode)
- & (HOST_WIDE_INT_1U << (bitwidth - 1))) != 0)
- return 1;
- else
- return cached_num_sign_bit_copies (XEXP (x, 0), mode,
- known_x, known_mode, known_ret);
-
- case UMOD:
- /* The result must be <= the second operand. If the second operand
- has (or just might have) the high bit set, we know nothing about
- the number of sign bit copies. */
- if (bitwidth > HOST_BITS_PER_WIDE_INT)
- return 1;
- else if ((nonzero_bits (XEXP (x, 1), mode)
- & (HOST_WIDE_INT_1U << (bitwidth - 1))) != 0)
- return 1;
- else
- return cached_num_sign_bit_copies (XEXP (x, 1), mode,
- known_x, known_mode, known_ret);
-
- case DIV:
- /* Similar to unsigned division, except that we have to worry about
- the case where the divisor is negative, in which case we have
- to add 1. */
- result = cached_num_sign_bit_copies (XEXP (x, 0), mode,
- known_x, known_mode, known_ret);
- if (result > 1
- && (bitwidth > HOST_BITS_PER_WIDE_INT
- || (nonzero_bits (XEXP (x, 1), mode)
- & (HOST_WIDE_INT_1U << (bitwidth - 1))) != 0))
- result--;
-
- return result;
-
- case MOD:
- result = cached_num_sign_bit_copies (XEXP (x, 1), mode,
- known_x, known_mode, known_ret);
- if (result > 1
- && (bitwidth > HOST_BITS_PER_WIDE_INT
- || (nonzero_bits (XEXP (x, 1), mode)
- & (HOST_WIDE_INT_1U << (bitwidth - 1))) != 0))
- result--;
-
- return result;
-
- case ASHIFTRT:
- /* Shifts by a constant add to the number of bits equal to the
- sign bit. */
- num0 = cached_num_sign_bit_copies (XEXP (x, 0), mode,
- known_x, known_mode, known_ret);
- if (CONST_INT_P (XEXP (x, 1))
- && INTVAL (XEXP (x, 1)) > 0
- && INTVAL (XEXP (x, 1)) < xmode_width)
- num0 = MIN ((int) bitwidth, num0 + INTVAL (XEXP (x, 1)));
-
- return num0;
-
- case ASHIFT:
- /* Left shifts destroy copies. */
- if (!CONST_INT_P (XEXP (x, 1))
- || INTVAL (XEXP (x, 1)) < 0
- || INTVAL (XEXP (x, 1)) >= (int) bitwidth
- || INTVAL (XEXP (x, 1)) >= xmode_width)
- return 1;
-
- num0 = cached_num_sign_bit_copies (XEXP (x, 0), mode,
- known_x, known_mode, known_ret);
- return MAX (1, num0 - INTVAL (XEXP (x, 1)));
-
- case IF_THEN_ELSE:
- num0 = cached_num_sign_bit_copies (XEXP (x, 1), mode,
- known_x, known_mode, known_ret);
- num1 = cached_num_sign_bit_copies (XEXP (x, 2), mode,
- known_x, known_mode, known_ret);
- return MIN (num0, num1);
-
- case EQ: case NE: case GE: case GT: case LE: case LT:
- case UNEQ: case LTGT: case UNGE: case UNGT: case UNLE: case UNLT:
- case GEU: case GTU: case LEU: case LTU:
- case UNORDERED: case ORDERED:
- /* If the constant is negative, take its 1's complement and remask.
- Then see how many zero bits we have. */
- nonzero = STORE_FLAG_VALUE;
- if (bitwidth <= HOST_BITS_PER_WIDE_INT
- && (nonzero & (HOST_WIDE_INT_1U << (bitwidth - 1))) != 0)
- nonzero = (~nonzero) & GET_MODE_MASK (mode);
-
- return (nonzero == 0 ? bitwidth : bitwidth - floor_log2 (nonzero) - 1);
-
- default:
- break;
- }
-
- /* If we haven't been able to figure it out by one of the above rules,
- see if some of the high-order bits are known to be zero. If so,
- count those bits and return one less than that amount. If we can't
- safely compute the mask for this mode, always return BITWIDTH. */
-
- bitwidth = GET_MODE_PRECISION (mode);
- if (bitwidth > HOST_BITS_PER_WIDE_INT)
- return 1;
-
- nonzero = nonzero_bits (x, mode);
- return nonzero & (HOST_WIDE_INT_1U << (bitwidth - 1))
- ? 1 : bitwidth - floor_log2 (nonzero) - 1;
-}
-
-/* Calculate the rtx_cost of a single instruction pattern. A return value of
- zero indicates an instruction pattern without a known cost. */
-
-int
-pattern_cost (rtx pat, bool speed)
-{
- int i, cost;
- rtx set;
-
- /* Extract the single set rtx from the instruction pattern. We
- can't use single_set since we only have the pattern. We also
- consider PARALLELs of a normal set and a single comparison. In
- that case we use the cost of the non-comparison SET operation,
- which is most-likely to be the real cost of this operation. */
- if (GET_CODE (pat) == SET)
- set = pat;
- else if (GET_CODE (pat) == PARALLEL)
- {
- set = NULL_RTX;
- rtx comparison = NULL_RTX;
-
- for (i = 0; i < XVECLEN (pat, 0); i++)
- {
- rtx x = XVECEXP (pat, 0, i);
- if (GET_CODE (x) == SET)
- {
- if (GET_CODE (SET_SRC (x)) == COMPARE)
- {
- if (comparison)
- return 0;
- comparison = x;
- }
- else
- {
- if (set)
- return 0;
- set = x;
- }
- }
- }
-
- if (!set && comparison)
- set = comparison;
-
- if (!set)
- return 0;
- }
- else
- return 0;
-
- cost = set_src_cost (SET_SRC (set), GET_MODE (SET_DEST (set)), speed);
- return cost > 0 ? cost : COSTS_N_INSNS (1);
-}
-
-/* Calculate the cost of a single instruction. A return value of zero
- indicates an instruction pattern without a known cost. */
-
-int
-insn_cost (rtx_insn *insn, bool speed)
-{
- if (targetm.insn_cost)
- return targetm.insn_cost (insn, speed);
-
- return pattern_cost (PATTERN (insn), speed);
-}
-
-/* Returns estimate on cost of computing SEQ. */
-
-unsigned
-seq_cost (const rtx_insn *seq, bool speed)
-{
- unsigned cost = 0;
- rtx set;
-
- for (; seq; seq = NEXT_INSN (seq))
- {
- set = single_set (seq);
- if (set)
- cost += set_rtx_cost (set, speed);
- else if (NONDEBUG_INSN_P (seq))
- {
- int this_cost = insn_cost (CONST_CAST_RTX_INSN (seq), speed);
- if (this_cost > 0)
- cost += this_cost;
- else
- cost++;
- }
- }
-
- return cost;
-}
-
-/* Given an insn INSN and condition COND, return the condition in a
- canonical form to simplify testing by callers. Specifically:
-
- (1) The code will always be a comparison operation (EQ, NE, GT, etc.).
- (2) Both operands will be machine operands.
- (3) If an operand is a constant, it will be the second operand.
- (4) (LE x const) will be replaced with (LT x <const+1>) and similarly
- for GE, GEU, and LEU.
-
- If the condition cannot be understood, or is an inequality floating-point
- comparison which needs to be reversed, 0 will be returned.
-
- If REVERSE is nonzero, then reverse the condition prior to canonizing it.
-
- If EARLIEST is nonzero, it is a pointer to a place where the earliest
- insn used in locating the condition was found. If a replacement test
- of the condition is desired, it should be placed in front of that
- insn and we will be sure that the inputs are still valid.
-
- If WANT_REG is nonzero, we wish the condition to be relative to that
- register, if possible. Therefore, do not canonicalize the condition
- further. If ALLOW_CC_MODE is nonzero, allow the condition returned
- to be a compare to a CC mode register.
-
- If VALID_AT_INSN_P, the condition must be valid at both *EARLIEST
- and at INSN. */
-
-rtx
-canonicalize_condition (rtx_insn *insn, rtx cond, int reverse,
- rtx_insn **earliest,
- rtx want_reg, int allow_cc_mode, int valid_at_insn_p)
-{
- enum rtx_code code;
- rtx_insn *prev = insn;
- const_rtx set;
- rtx tem;
- rtx op0, op1;
- int reverse_code = 0;
- machine_mode mode;
- basic_block bb = BLOCK_FOR_INSN (insn);
-
- code = GET_CODE (cond);
- mode = GET_MODE (cond);
- op0 = XEXP (cond, 0);
- op1 = XEXP (cond, 1);
-
- if (reverse)
- code = reversed_comparison_code (cond, insn);
- if (code == UNKNOWN)
- return 0;
-
- if (earliest)
- *earliest = insn;
-
- /* If we are comparing a register with zero, see if the register is set
- in the previous insn to a COMPARE or a comparison operation. Perform
- the same tests as a function of STORE_FLAG_VALUE as find_comparison_args
- in cse.c */
-
- while ((GET_RTX_CLASS (code) == RTX_COMPARE
- || GET_RTX_CLASS (code) == RTX_COMM_COMPARE)
- && op1 == CONST0_RTX (GET_MODE (op0))
- && op0 != want_reg)
- {
- /* Set nonzero when we find something of interest. */
- rtx x = 0;
-
- /* If this is a COMPARE, pick up the two things being compared. */
- if (GET_CODE (op0) == COMPARE)
- {
- op1 = XEXP (op0, 1);
- op0 = XEXP (op0, 0);
- continue;
- }
- else if (!REG_P (op0))
- break;
-
- /* Go back to the previous insn. Stop if it is not an INSN. We also
- stop if it isn't a single set or if it has a REG_INC note because
- we don't want to bother dealing with it. */
-
- prev = prev_nonnote_nondebug_insn (prev);
-
- if (prev == 0
- || !NONJUMP_INSN_P (prev)
- || FIND_REG_INC_NOTE (prev, NULL_RTX)
- /* In cfglayout mode, there do not have to be labels at the
- beginning of a block, or jumps at the end, so the previous
- conditions would not stop us when we reach bb boundary. */
- || BLOCK_FOR_INSN (prev) != bb)
- break;
-
- set = set_of (op0, prev);
-
- if (set
- && (GET_CODE (set) != SET
- || !rtx_equal_p (SET_DEST (set), op0)))
- break;
-
- /* If this is setting OP0, get what it sets it to if it looks
- relevant. */
- if (set)
- {
- machine_mode inner_mode = GET_MODE (SET_DEST (set));
-#ifdef FLOAT_STORE_FLAG_VALUE
- REAL_VALUE_TYPE fsfv;
-#endif
-
- /* ??? We may not combine comparisons done in a CCmode with
- comparisons not done in a CCmode. This is to aid targets
- like Alpha that have an IEEE compliant EQ instruction, and
- a non-IEEE compliant BEQ instruction. The use of CCmode is
- actually artificial, simply to prevent the combination, but
- should not affect other platforms.
-
- However, we must allow VOIDmode comparisons to match either
- CCmode or non-CCmode comparison, because some ports have
- modeless comparisons inside branch patterns.
-
- ??? This mode check should perhaps look more like the mode check
- in simplify_comparison in combine. */
- if (((GET_MODE_CLASS (mode) == MODE_CC)
- != (GET_MODE_CLASS (inner_mode) == MODE_CC))
- && mode != VOIDmode
- && inner_mode != VOIDmode)
- break;
- if (GET_CODE (SET_SRC (set)) == COMPARE
- || (((code == NE
- || (code == LT
- && val_signbit_known_set_p (inner_mode,
- STORE_FLAG_VALUE))
-#ifdef FLOAT_STORE_FLAG_VALUE
- || (code == LT
- && SCALAR_FLOAT_MODE_P (inner_mode)
- && (fsfv = FLOAT_STORE_FLAG_VALUE (inner_mode),
- REAL_VALUE_NEGATIVE (fsfv)))
-#endif
- ))
- && COMPARISON_P (SET_SRC (set))))
- x = SET_SRC (set);
- else if (((code == EQ
- || (code == GE
- && val_signbit_known_set_p (inner_mode,
- STORE_FLAG_VALUE))
-#ifdef FLOAT_STORE_FLAG_VALUE
- || (code == GE
- && SCALAR_FLOAT_MODE_P (inner_mode)
- && (fsfv = FLOAT_STORE_FLAG_VALUE (inner_mode),
- REAL_VALUE_NEGATIVE (fsfv)))
-#endif
- ))
- && COMPARISON_P (SET_SRC (set)))
- {
- reverse_code = 1;
- x = SET_SRC (set);
- }
- else if ((code == EQ || code == NE)
- && GET_CODE (SET_SRC (set)) == XOR)
- /* Handle sequences like:
-
- (set op0 (xor X Y))
- ...(eq|ne op0 (const_int 0))...
-
- in which case:
-
- (eq op0 (const_int 0)) reduces to (eq X Y)
- (ne op0 (const_int 0)) reduces to (ne X Y)
-
- This is the form used by MIPS16, for example. */
- x = SET_SRC (set);
- else
- break;
- }
-
- else if (reg_set_p (op0, prev))
- /* If this sets OP0, but not directly, we have to give up. */
- break;
-
- if (x)
- {
- /* If the caller is expecting the condition to be valid at INSN,
- make sure X doesn't change before INSN. */
- if (valid_at_insn_p)
- if (modified_in_p (x, prev) || modified_between_p (x, prev, insn))
- break;
- if (COMPARISON_P (x))
- code = GET_CODE (x);
- if (reverse_code)
- {
- code = reversed_comparison_code (x, prev);
- if (code == UNKNOWN)
- return 0;
- reverse_code = 0;
- }
-
- op0 = XEXP (x, 0), op1 = XEXP (x, 1);
- if (earliest)
- *earliest = prev;
- }
- }
-
- /* If constant is first, put it last. */
- if (CONSTANT_P (op0))
- code = swap_condition (code), tem = op0, op0 = op1, op1 = tem;
-
- /* If OP0 is the result of a comparison, we weren't able to find what
- was really being compared, so fail. */
- if (!allow_cc_mode
- && GET_MODE_CLASS (GET_MODE (op0)) == MODE_CC)
- return 0;
-
- /* Canonicalize any ordered comparison with integers involving equality
- if we can do computations in the relevant mode and we do not
- overflow. */
-
- scalar_int_mode op0_mode;
- if (CONST_INT_P (op1)
- && is_a <scalar_int_mode> (GET_MODE (op0), &op0_mode)
- && GET_MODE_PRECISION (op0_mode) <= HOST_BITS_PER_WIDE_INT)
- {
- HOST_WIDE_INT const_val = INTVAL (op1);
- unsigned HOST_WIDE_INT uconst_val = const_val;
- unsigned HOST_WIDE_INT max_val
- = (unsigned HOST_WIDE_INT) GET_MODE_MASK (op0_mode);
-
- switch (code)
- {
- case LE:
- if ((unsigned HOST_WIDE_INT) const_val != max_val >> 1)
- code = LT, op1 = gen_int_mode (const_val + 1, op0_mode);
- break;
-
- /* When cross-compiling, const_val might be sign-extended from
- BITS_PER_WORD to HOST_BITS_PER_WIDE_INT */
- case GE:
- if ((const_val & max_val)
- != (HOST_WIDE_INT_1U << (GET_MODE_PRECISION (op0_mode) - 1)))
- code = GT, op1 = gen_int_mode (const_val - 1, op0_mode);
- break;
-
- case LEU:
- if (uconst_val < max_val)
- code = LTU, op1 = gen_int_mode (uconst_val + 1, op0_mode);
- break;
-
- case GEU:
- if (uconst_val != 0)
- code = GTU, op1 = gen_int_mode (uconst_val - 1, op0_mode);
- break;
-
- default:
- break;
- }
- }
-
- /* We promised to return a comparison. */
- rtx ret = gen_rtx_fmt_ee (code, VOIDmode, op0, op1);
- if (COMPARISON_P (ret))
- return ret;
- return 0;
-}
-
-/* Given a jump insn JUMP, return the condition that will cause it to branch
- to its JUMP_LABEL. If the condition cannot be understood, or is an
- inequality floating-point comparison which needs to be reversed, 0 will
- be returned.
-
- If EARLIEST is nonzero, it is a pointer to a place where the earliest
- insn used in locating the condition was found. If a replacement test
- of the condition is desired, it should be placed in front of that
- insn and we will be sure that the inputs are still valid. If EARLIEST
- is null, the returned condition will be valid at INSN.
-
- If ALLOW_CC_MODE is nonzero, allow the condition returned to be a
- compare CC mode register.
-
- VALID_AT_INSN_P is the same as for canonicalize_condition. */
-
-rtx
-get_condition (rtx_insn *jump, rtx_insn **earliest, int allow_cc_mode,
- int valid_at_insn_p)
-{
- rtx cond;
- int reverse;
- rtx set;
-
- /* If this is not a standard conditional jump, we can't parse it. */
- if (!JUMP_P (jump)
- || ! any_condjump_p (jump))
- return 0;
- set = pc_set (jump);
-
- cond = XEXP (SET_SRC (set), 0);
-
- /* If this branches to JUMP_LABEL when the condition is false, reverse
- the condition. */
- reverse
- = GET_CODE (XEXP (SET_SRC (set), 2)) == LABEL_REF
- && label_ref_label (XEXP (SET_SRC (set), 2)) == JUMP_LABEL (jump);
-
- return canonicalize_condition (jump, cond, reverse, earliest, NULL_RTX,
- allow_cc_mode, valid_at_insn_p);
-}
-
-/* Initialize the table NUM_SIGN_BIT_COPIES_IN_REP based on
- TARGET_MODE_REP_EXTENDED.
-
- Note that we assume that the property of
- TARGET_MODE_REP_EXTENDED(B, C) is sticky to the integral modes
- narrower than mode B. I.e., if A is a mode narrower than B then in
- order to be able to operate on it in mode B, mode A needs to
- satisfy the requirements set by the representation of mode B. */
-
-static void
-init_num_sign_bit_copies_in_rep (void)
-{
- opt_scalar_int_mode in_mode_iter;
- scalar_int_mode mode;
-
- FOR_EACH_MODE_IN_CLASS (in_mode_iter, MODE_INT)
- FOR_EACH_MODE_UNTIL (mode, in_mode_iter.require ())
- {
- scalar_int_mode in_mode = in_mode_iter.require ();
- scalar_int_mode i;
-
- /* Currently, it is assumed that TARGET_MODE_REP_EXTENDED
- extends to the next widest mode. */
- gcc_assert (targetm.mode_rep_extended (mode, in_mode) == UNKNOWN
- || GET_MODE_WIDER_MODE (mode).require () == in_mode);
-
- /* We are in in_mode. Count how many bits outside of mode
- have to be copies of the sign-bit. */
- FOR_EACH_MODE (i, mode, in_mode)
- {
- /* This must always exist (for the last iteration it will be
- IN_MODE). */
- scalar_int_mode wider = GET_MODE_WIDER_MODE (i).require ();
-
- if (targetm.mode_rep_extended (i, wider) == SIGN_EXTEND
- /* We can only check sign-bit copies starting from the
- top-bit. In order to be able to check the bits we
- have already seen we pretend that subsequent bits
- have to be sign-bit copies too. */
- || num_sign_bit_copies_in_rep [in_mode][mode])
- num_sign_bit_copies_in_rep [in_mode][mode]
- += GET_MODE_PRECISION (wider) - GET_MODE_PRECISION (i);
- }
- }
-}
-
-/* Suppose that truncation from the machine mode of X to MODE is not a
- no-op. See if there is anything special about X so that we can
- assume it already contains a truncated value of MODE. */
-
-bool
-truncated_to_mode (machine_mode mode, const_rtx x)
-{
- /* This register has already been used in MODE without explicit
- truncation. */
- if (REG_P (x) && rtl_hooks.reg_truncated_to_mode (mode, x))
- return true;
-
- /* See if we already satisfy the requirements of MODE. If yes we
- can just switch to MODE. */
- if (num_sign_bit_copies_in_rep[GET_MODE (x)][mode]
- && (num_sign_bit_copies (x, GET_MODE (x))
- >= num_sign_bit_copies_in_rep[GET_MODE (x)][mode] + 1))
- return true;
-
- return false;
-}
-
-/* Return true if RTX code CODE has a single sequence of zero or more
- "e" operands and no rtvec operands. Initialize its rtx_all_subrtx_bounds
- entry in that case. */
-
-static bool
-setup_reg_subrtx_bounds (unsigned int code)
-{
- const char *format = GET_RTX_FORMAT ((enum rtx_code) code);
- unsigned int i = 0;
- for (; format[i] != 'e'; ++i)
- {
- if (!format[i])
- /* No subrtxes. Leave start and count as 0. */
- return true;
- if (format[i] == 'E' || format[i] == 'V')
- return false;
- }
-
- /* Record the sequence of 'e's. */
- rtx_all_subrtx_bounds[code].start = i;
- do
- ++i;
- while (format[i] == 'e');
- rtx_all_subrtx_bounds[code].count = i - rtx_all_subrtx_bounds[code].start;
- /* rtl-iter.h relies on this. */
- gcc_checking_assert (rtx_all_subrtx_bounds[code].count <= 3);
-
- for (; format[i]; ++i)
- if (format[i] == 'E' || format[i] == 'V' || format[i] == 'e')
- return false;
-
- return true;
-}
-
-/* Initialize rtx_all_subrtx_bounds. */
-void
-init_rtlanal (void)
-{
- int i;
- for (i = 0; i < NUM_RTX_CODE; i++)
- {
- if (!setup_reg_subrtx_bounds (i))
- rtx_all_subrtx_bounds[i].count = UCHAR_MAX;
- if (GET_RTX_CLASS (i) != RTX_CONST_OBJ)
- rtx_nonconst_subrtx_bounds[i] = rtx_all_subrtx_bounds[i];
- }
-
- init_num_sign_bit_copies_in_rep ();
-}
-
-/* Check whether this is a constant pool constant. */
-bool
-constant_pool_constant_p (rtx x)
-{
- x = avoid_constant_pool_reference (x);
- return CONST_DOUBLE_P (x);
-}
-
-/* If M is a bitmask that selects a field of low-order bits within an item but
- not the entire word, return the length of the field. Return -1 otherwise.
- M is used in machine mode MODE. */
-
-int
-low_bitmask_len (machine_mode mode, unsigned HOST_WIDE_INT m)
-{
- if (mode != VOIDmode)
- {
- if (!HWI_COMPUTABLE_MODE_P (mode))
- return -1;
- m &= GET_MODE_MASK (mode);
- }
-
- return exact_log2 (m + 1);
-}
-
-/* Return the mode of MEM's address. */
-
-scalar_int_mode
-get_address_mode (rtx mem)
-{
- machine_mode mode;
-
- gcc_assert (MEM_P (mem));
- mode = GET_MODE (XEXP (mem, 0));
- if (mode != VOIDmode)
- return as_a <scalar_int_mode> (mode);
- return targetm.addr_space.address_mode (MEM_ADDR_SPACE (mem));
-}
-
-/* Split up a CONST_DOUBLE or integer constant rtx
- into two rtx's for single words,
- storing in *FIRST the word that comes first in memory in the target
- and in *SECOND the other.
-
- TODO: This function needs to be rewritten to work on any size
- integer. */
-
-void
-split_double (rtx value, rtx *first, rtx *second)
-{
- if (CONST_INT_P (value))
- {
- if (HOST_BITS_PER_WIDE_INT >= (2 * BITS_PER_WORD))
- {
- /* In this case the CONST_INT holds both target words.
- Extract the bits from it into two word-sized pieces.
- Sign extend each half to HOST_WIDE_INT. */
- unsigned HOST_WIDE_INT low, high;
- unsigned HOST_WIDE_INT mask, sign_bit, sign_extend;
- unsigned bits_per_word = BITS_PER_WORD;
-
- /* Set sign_bit to the most significant bit of a word. */
- sign_bit = 1;
- sign_bit <<= bits_per_word - 1;
-
- /* Set mask so that all bits of the word are set. We could
- have used 1 << BITS_PER_WORD instead of basing the
- calculation on sign_bit. However, on machines where
- HOST_BITS_PER_WIDE_INT == BITS_PER_WORD, it could cause a
- compiler warning, even though the code would never be
- executed. */
- mask = sign_bit << 1;
- mask--;
-
- /* Set sign_extend as any remaining bits. */
- sign_extend = ~mask;
-
- /* Pick the lower word and sign-extend it. */
- low = INTVAL (value);
- low &= mask;
- if (low & sign_bit)
- low |= sign_extend;
-
- /* Pick the higher word, shifted to the least significant
- bits, and sign-extend it. */
- high = INTVAL (value);
- high >>= bits_per_word - 1;
- high >>= 1;
- high &= mask;
- if (high & sign_bit)
- high |= sign_extend;
-
- /* Store the words in the target machine order. */
- if (WORDS_BIG_ENDIAN)
- {
- *first = GEN_INT (high);
- *second = GEN_INT (low);
- }
- else
- {
- *first = GEN_INT (low);
- *second = GEN_INT (high);
- }
- }
- else
- {
- /* The rule for using CONST_INT for a wider mode
- is that we regard the value as signed.
- So sign-extend it. */
- rtx high = (INTVAL (value) < 0 ? constm1_rtx : const0_rtx);
- if (WORDS_BIG_ENDIAN)
- {
- *first = high;
- *second = value;
- }
- else
- {
- *first = value;
- *second = high;
- }
- }
- }
- else if (GET_CODE (value) == CONST_WIDE_INT)
- {
- /* All of this is scary code and needs to be converted to
- properly work with any size integer. */
- gcc_assert (CONST_WIDE_INT_NUNITS (value) == 2);
- if (WORDS_BIG_ENDIAN)
- {
- *first = GEN_INT (CONST_WIDE_INT_ELT (value, 1));
- *second = GEN_INT (CONST_WIDE_INT_ELT (value, 0));
- }
- else
- {
- *first = GEN_INT (CONST_WIDE_INT_ELT (value, 0));
- *second = GEN_INT (CONST_WIDE_INT_ELT (value, 1));
- }
- }
- else if (!CONST_DOUBLE_P (value))
- {
- if (WORDS_BIG_ENDIAN)
- {
- *first = const0_rtx;
- *second = value;
- }
- else
- {
- *first = value;
- *second = const0_rtx;
- }
- }
- else if (GET_MODE (value) == VOIDmode
- /* This is the old way we did CONST_DOUBLE integers. */
- || GET_MODE_CLASS (GET_MODE (value)) == MODE_INT)
- {
- /* In an integer, the words are defined as most and least significant.
- So order them by the target's convention. */
- if (WORDS_BIG_ENDIAN)
- {
- *first = GEN_INT (CONST_DOUBLE_HIGH (value));
- *second = GEN_INT (CONST_DOUBLE_LOW (value));
- }
- else
- {
- *first = GEN_INT (CONST_DOUBLE_LOW (value));
- *second = GEN_INT (CONST_DOUBLE_HIGH (value));
- }
- }
- else
- {
- long l[2];
-
- /* Note, this converts the REAL_VALUE_TYPE to the target's
- format, splits up the floating point double and outputs
- exactly 32 bits of it into each of l[0] and l[1] --
- not necessarily BITS_PER_WORD bits. */
- REAL_VALUE_TO_TARGET_DOUBLE (*CONST_DOUBLE_REAL_VALUE (value), l);
-
- /* If 32 bits is an entire word for the target, but not for the host,
- then sign-extend on the host so that the number will look the same
- way on the host that it would on the target. See for instance
- simplify_unary_operation. The #if is needed to avoid compiler
- warnings. */
-
-#if HOST_BITS_PER_LONG > 32
- if (BITS_PER_WORD < HOST_BITS_PER_LONG && BITS_PER_WORD == 32)
- {
- if (l[0] & ((long) 1 << 31))
- l[0] |= ((unsigned long) (-1) << 32);
- if (l[1] & ((long) 1 << 31))
- l[1] |= ((unsigned long) (-1) << 32);
- }
-#endif
-
- *first = GEN_INT (l[0]);
- *second = GEN_INT (l[1]);
- }
-}
-
-/* Return true if X is a sign_extract or zero_extract from the least
- significant bit. */
-
-static bool
-lsb_bitfield_op_p (rtx x)
-{
- if (GET_RTX_CLASS (GET_CODE (x)) == RTX_BITFIELD_OPS)
- {
- machine_mode mode = GET_MODE (XEXP (x, 0));
- HOST_WIDE_INT len = INTVAL (XEXP (x, 1));
- HOST_WIDE_INT pos = INTVAL (XEXP (x, 2));
- poly_int64 remaining_bits = GET_MODE_PRECISION (mode) - len;
-
- return known_eq (pos, BITS_BIG_ENDIAN ? remaining_bits : 0);
- }
- return false;
-}
-
-/* Strip outer address "mutations" from LOC and return a pointer to the
- inner value. If OUTER_CODE is nonnull, store the code of the innermost
- stripped expression there.
-
- "Mutations" either convert between modes or apply some kind of
- extension, truncation or alignment. */
-
-rtx *
-strip_address_mutations (rtx *loc, enum rtx_code *outer_code)
-{
- for (;;)
- {
- enum rtx_code code = GET_CODE (*loc);
- if (GET_RTX_CLASS (code) == RTX_UNARY)
- /* Things like SIGN_EXTEND, ZERO_EXTEND and TRUNCATE can be
- used to convert between pointer sizes. */
- loc = &XEXP (*loc, 0);
- else if (lsb_bitfield_op_p (*loc))
- /* A [SIGN|ZERO]_EXTRACT from the least significant bit effectively
- acts as a combined truncation and extension. */
- loc = &XEXP (*loc, 0);
- else if (code == AND && CONST_INT_P (XEXP (*loc, 1)))
- /* (and ... (const_int -X)) is used to align to X bytes. */
- loc = &XEXP (*loc, 0);
- else if (code == SUBREG
- && !OBJECT_P (SUBREG_REG (*loc))
- && subreg_lowpart_p (*loc))
- /* (subreg (operator ...) ...) inside and is used for mode
- conversion too. */
- loc = &SUBREG_REG (*loc);
- else
- return loc;
- if (outer_code)
- *outer_code = code;
- }
-}
-
-/* Return true if CODE applies some kind of scale. The scaled value is
- is the first operand and the scale is the second. */
-
-static bool
-binary_scale_code_p (enum rtx_code code)
-{
- return (code == MULT
- || code == ASHIFT
- /* Needed by ARM targets. */
- || code == ASHIFTRT
- || code == LSHIFTRT
- || code == ROTATE
- || code == ROTATERT);
-}
-
-/* If *INNER can be interpreted as a base, return a pointer to the inner term
- (see address_info). Return null otherwise. */
-
-static rtx *
-get_base_term (rtx *inner)
-{
- if (GET_CODE (*inner) == LO_SUM)
- inner = strip_address_mutations (&XEXP (*inner, 0));
- if (REG_P (*inner)
- || MEM_P (*inner)
- || GET_CODE (*inner) == SUBREG
- || GET_CODE (*inner) == SCRATCH)
- return inner;
- return 0;
-}
-
-/* If *INNER can be interpreted as an index, return a pointer to the inner term
- (see address_info). Return null otherwise. */
-
-static rtx *
-get_index_term (rtx *inner)
-{
- /* At present, only constant scales are allowed. */
- if (binary_scale_code_p (GET_CODE (*inner)) && CONSTANT_P (XEXP (*inner, 1)))
- inner = strip_address_mutations (&XEXP (*inner, 0));
- if (REG_P (*inner)
- || MEM_P (*inner)
- || GET_CODE (*inner) == SUBREG
- || GET_CODE (*inner) == SCRATCH)
- return inner;
- return 0;
-}
-
-/* Set the segment part of address INFO to LOC, given that INNER is the
- unmutated value. */
-
-static void
-set_address_segment (struct address_info *info, rtx *loc, rtx *inner)
-{
- gcc_assert (!info->segment);
- info->segment = loc;
- info->segment_term = inner;
-}
-
-/* Set the base part of address INFO to LOC, given that INNER is the
- unmutated value. */
-
-static void
-set_address_base (struct address_info *info, rtx *loc, rtx *inner)
-{
- gcc_assert (!info->base);
- info->base = loc;
- info->base_term = inner;
-}
-
-/* Set the index part of address INFO to LOC, given that INNER is the
- unmutated value. */
-
-static void
-set_address_index (struct address_info *info, rtx *loc, rtx *inner)
-{
- gcc_assert (!info->index);
- info->index = loc;
- info->index_term = inner;
-}
-
-/* Set the displacement part of address INFO to LOC, given that INNER
- is the constant term. */
-
-static void
-set_address_disp (struct address_info *info, rtx *loc, rtx *inner)
-{
- gcc_assert (!info->disp);
- info->disp = loc;
- info->disp_term = inner;
-}
-
-/* INFO->INNER describes a {PRE,POST}_{INC,DEC} address. Set up the
- rest of INFO accordingly. */
-
-static void
-decompose_incdec_address (struct address_info *info)
-{
- info->autoinc_p = true;
-
- rtx *base = &XEXP (*info->inner, 0);
- set_address_base (info, base, base);
- gcc_checking_assert (info->base == info->base_term);
-
- /* These addresses are only valid when the size of the addressed
- value is known. */
- gcc_checking_assert (info->mode != VOIDmode);
-}
-
-/* INFO->INNER describes a {PRE,POST}_MODIFY address. Set up the rest
- of INFO accordingly. */
-
-static void
-decompose_automod_address (struct address_info *info)
-{
- info->autoinc_p = true;
-
- rtx *base = &XEXP (*info->inner, 0);
- set_address_base (info, base, base);
- gcc_checking_assert (info->base == info->base_term);
-
- rtx plus = XEXP (*info->inner, 1);
- gcc_assert (GET_CODE (plus) == PLUS);
-
- info->base_term2 = &XEXP (plus, 0);
- gcc_checking_assert (rtx_equal_p (*info->base_term, *info->base_term2));
-
- rtx *step = &XEXP (plus, 1);
- rtx *inner_step = strip_address_mutations (step);
- if (CONSTANT_P (*inner_step))
- set_address_disp (info, step, inner_step);
- else
- set_address_index (info, step, inner_step);
-}
-
-/* Treat *LOC as a tree of PLUS operands and store pointers to the summed
- values in [PTR, END). Return a pointer to the end of the used array. */
-
-static rtx **
-extract_plus_operands (rtx *loc, rtx **ptr, rtx **end)
-{
- rtx x = *loc;
- if (GET_CODE (x) == PLUS)
- {
- ptr = extract_plus_operands (&XEXP (x, 0), ptr, end);
- ptr = extract_plus_operands (&XEXP (x, 1), ptr, end);
- }
- else
- {
- gcc_assert (ptr != end);
- *ptr++ = loc;
- }
- return ptr;
-}
-
-/* Evaluate the likelihood of X being a base or index value, returning
- positive if it is likely to be a base, negative if it is likely to be
- an index, and 0 if we can't tell. Make the magnitude of the return
- value reflect the amount of confidence we have in the answer.
-
- MODE, AS, OUTER_CODE and INDEX_CODE are as for ok_for_base_p_1. */
-
-static int
-baseness (rtx x, machine_mode mode, addr_space_t as,
- enum rtx_code outer_code, enum rtx_code index_code)
-{
- /* Believe *_POINTER unless the address shape requires otherwise. */
- if (REG_P (x) && REG_POINTER (x))
- return 2;
- if (MEM_P (x) && MEM_POINTER (x))
- return 2;
-
- if (REG_P (x) && HARD_REGISTER_P (x))
- {
- /* X is a hard register. If it only fits one of the base
- or index classes, choose that interpretation. */
- int regno = REGNO (x);
- bool base_p = ok_for_base_p_1 (regno, mode, as, outer_code, index_code);
- bool index_p = REGNO_OK_FOR_INDEX_P (regno);
- if (base_p != index_p)
- return base_p ? 1 : -1;
- }
- return 0;
-}
-
-/* INFO->INNER describes a normal, non-automodified address.
- Fill in the rest of INFO accordingly. */
-
-static void
-decompose_normal_address (struct address_info *info)
-{
- /* Treat the address as the sum of up to four values. */
- rtx *ops[4];
- size_t n_ops = extract_plus_operands (info->inner, ops,
- ops + ARRAY_SIZE (ops)) - ops;
-
- /* If there is more than one component, any base component is in a PLUS. */
- if (n_ops > 1)
- info->base_outer_code = PLUS;
-
- /* Try to classify each sum operand now. Leave those that could be
- either a base or an index in OPS. */
- rtx *inner_ops[4];
- size_t out = 0;
- for (size_t in = 0; in < n_ops; ++in)
- {
- rtx *loc = ops[in];
- rtx *inner = strip_address_mutations (loc);
- if (CONSTANT_P (*inner))
- set_address_disp (info, loc, inner);
- else if (GET_CODE (*inner) == UNSPEC)
- set_address_segment (info, loc, inner);
- else
- {
- /* The only other possibilities are a base or an index. */
- rtx *base_term = get_base_term (inner);
- rtx *index_term = get_index_term (inner);
- gcc_assert (base_term || index_term);
- if (!base_term)
- set_address_index (info, loc, index_term);
- else if (!index_term)
- set_address_base (info, loc, base_term);
- else
- {
- gcc_assert (base_term == index_term);
- ops[out] = loc;
- inner_ops[out] = base_term;
- ++out;
- }
- }
- }
-
- /* Classify the remaining OPS members as bases and indexes. */
- if (out == 1)
- {
- /* If we haven't seen a base or an index yet, assume that this is
- the base. If we were confident that another term was the base
- or index, treat the remaining operand as the other kind. */
- if (!info->base)
- set_address_base (info, ops[0], inner_ops[0]);
- else
- set_address_index (info, ops[0], inner_ops[0]);
- }
- else if (out == 2)
- {
- /* In the event of a tie, assume the base comes first. */
- if (baseness (*inner_ops[0], info->mode, info->as, PLUS,
- GET_CODE (*ops[1]))
- >= baseness (*inner_ops[1], info->mode, info->as, PLUS,
- GET_CODE (*ops[0])))
- {
- set_address_base (info, ops[0], inner_ops[0]);
- set_address_index (info, ops[1], inner_ops[1]);
- }
- else
- {
- set_address_base (info, ops[1], inner_ops[1]);
- set_address_index (info, ops[0], inner_ops[0]);
- }
- }
- else
- gcc_assert (out == 0);
-}
-
-/* Describe address *LOC in *INFO. MODE is the mode of the addressed value,
- or VOIDmode if not known. AS is the address space associated with LOC.
- OUTER_CODE is MEM if *LOC is a MEM address and ADDRESS otherwise. */
-
-void
-decompose_address (struct address_info *info, rtx *loc, machine_mode mode,
- addr_space_t as, enum rtx_code outer_code)
-{
- memset (info, 0, sizeof (*info));
- info->mode = mode;
- info->as = as;
- info->addr_outer_code = outer_code;
- info->outer = loc;
- info->inner = strip_address_mutations (loc, &outer_code);
- info->base_outer_code = outer_code;
- switch (GET_CODE (*info->inner))
- {
- case PRE_DEC:
- case PRE_INC:
- case POST_DEC:
- case POST_INC:
- decompose_incdec_address (info);
- break;
-
- case PRE_MODIFY:
- case POST_MODIFY:
- decompose_automod_address (info);
- break;
-
- default:
- decompose_normal_address (info);
- break;
- }
-}
-
-/* Describe address operand LOC in INFO. */
-
-void
-decompose_lea_address (struct address_info *info, rtx *loc)
-{
- decompose_address (info, loc, VOIDmode, ADDR_SPACE_GENERIC, ADDRESS);
-}
-
-/* Describe the address of MEM X in INFO. */
-
-void
-decompose_mem_address (struct address_info *info, rtx x)
-{
- gcc_assert (MEM_P (x));
- decompose_address (info, &XEXP (x, 0), GET_MODE (x),
- MEM_ADDR_SPACE (x), MEM);
-}
-
-/* Update INFO after a change to the address it describes. */
-
-void
-update_address (struct address_info *info)
-{
- decompose_address (info, info->outer, info->mode, info->as,
- info->addr_outer_code);
-}
-
-/* Return the scale applied to *INFO->INDEX_TERM, or 0 if the index is
- more complicated than that. */
-
-HOST_WIDE_INT
-get_index_scale (const struct address_info *info)
-{
- rtx index = *info->index;
- if (GET_CODE (index) == MULT
- && CONST_INT_P (XEXP (index, 1))
- && info->index_term == &XEXP (index, 0))
- return INTVAL (XEXP (index, 1));
-
- if (GET_CODE (index) == ASHIFT
- && CONST_INT_P (XEXP (index, 1))
- && info->index_term == &XEXP (index, 0))
- return HOST_WIDE_INT_1 << INTVAL (XEXP (index, 1));
-
- if (info->index == info->index_term)
- return 1;
-
- return 0;
-}
-
-/* Return the "index code" of INFO, in the form required by
- ok_for_base_p_1. */
-
-enum rtx_code
-get_index_code (const struct address_info *info)
-{
- if (info->index)
- return GET_CODE (*info->index);
-
- if (info->disp)
- return GET_CODE (*info->disp);
-
- return SCRATCH;
-}
-
-/* Return true if RTL X contains a SYMBOL_REF. */
-
-bool
-contains_symbol_ref_p (const_rtx x)
-{
- subrtx_iterator::array_type array;
- FOR_EACH_SUBRTX (iter, array, x, ALL)
- if (SYMBOL_REF_P (*iter))
- return true;
-
- return false;
-}
-
-/* Return true if RTL X contains a SYMBOL_REF or LABEL_REF. */
-
-bool
-contains_symbolic_reference_p (const_rtx x)
-{
- subrtx_iterator::array_type array;
- FOR_EACH_SUBRTX (iter, array, x, ALL)
- if (SYMBOL_REF_P (*iter) || GET_CODE (*iter) == LABEL_REF)
- return true;
-
- return false;
-}
-
-/* Return true if RTL X contains a constant pool address. */
-
-bool
-contains_constant_pool_address_p (const_rtx x)
-{
- subrtx_iterator::array_type array;
- FOR_EACH_SUBRTX (iter, array, x, ALL)
- if (SYMBOL_REF_P (*iter) && CONSTANT_POOL_ADDRESS_P (*iter))
- return true;
-
- return false;
-}
-
-
-/* Return true if X contains a thread-local symbol. */
-
-bool
-tls_referenced_p (const_rtx x)
-{
- if (!targetm.have_tls)
- return false;
-
- subrtx_iterator::array_type array;
- FOR_EACH_SUBRTX (iter, array, x, ALL)
- if (GET_CODE (*iter) == SYMBOL_REF && SYMBOL_REF_TLS_MODEL (*iter) != 0)
- return true;
- return false;
-}
-
-/* Process recursively X of INSN and add REG_INC notes if necessary. */
-void
-add_auto_inc_notes (rtx_insn *insn, rtx x)
-{
- enum rtx_code code = GET_CODE (x);
- const char *fmt;
- int i, j;
-
- if (code == MEM && auto_inc_p (XEXP (x, 0)))
- {
- add_reg_note (insn, REG_INC, XEXP (XEXP (x, 0), 0));
- return;
- }
-
- /* Scan all X sub-expressions. */
- fmt = GET_RTX_FORMAT (code);
- for (i = GET_RTX_LENGTH (code) - 1; i >= 0; i--)
- {
- if (fmt[i] == 'e')
- add_auto_inc_notes (insn, XEXP (x, i));
- else if (fmt[i] == 'E')
- for (j = XVECLEN (x, i) - 1; j >= 0; j--)
- add_auto_inc_notes (insn, XVECEXP (x, i, j));
- }
-}
-
-/* Return true if X is register asm. */
-
-bool
-register_asm_p (const_rtx x)
-{
- return (REG_P (x)
- && REG_EXPR (x) != NULL_TREE
- && HAS_DECL_ASSEMBLER_NAME_P (REG_EXPR (x))
- && DECL_ASSEMBLER_NAME_SET_P (REG_EXPR (x))
- && DECL_REGISTER (REG_EXPR (x)));
-}
-
-/* Return true if, for all OP of mode OP_MODE:
-
- (vec_select:RESULT_MODE OP SEL)
-
- is equivalent to the highpart RESULT_MODE of OP. */
-
-bool
-vec_series_highpart_p (machine_mode result_mode, machine_mode op_mode, rtx sel)
-{
- int nunits;
- if (GET_MODE_NUNITS (op_mode).is_constant (&nunits)
- && targetm.can_change_mode_class (op_mode, result_mode, ALL_REGS))
- {
- int offset = BYTES_BIG_ENDIAN ? 0 : nunits - XVECLEN (sel, 0);
- return rtvec_series_p (XVEC (sel, 0), offset);
- }
- return false;
-}
-
-/* Return true if, for all OP of mode OP_MODE:
-
- (vec_select:RESULT_MODE OP SEL)
-
- is equivalent to the lowpart RESULT_MODE of OP. */
-
-bool
-vec_series_lowpart_p (machine_mode result_mode, machine_mode op_mode, rtx sel)
-{
- int nunits;
- if (GET_MODE_NUNITS (op_mode).is_constant (&nunits)
- && targetm.can_change_mode_class (op_mode, result_mode, ALL_REGS))
- {
- int offset = BYTES_BIG_ENDIAN ? nunits - XVECLEN (sel, 0) : 0;
- return rtvec_series_p (XVEC (sel, 0), offset);
- }
- return false;
-}