diff options
author | Martin Liska <mliska@suse.cz> | 2022-01-14 16:56:44 +0100 |
---|---|---|
committer | Martin Liska <mliska@suse.cz> | 2022-01-17 22:12:04 +0100 |
commit | 5c69acb32329d49e58c26fa41ae74229a52b9106 (patch) | |
tree | ddb05f9d73afb6f998457d2ac4b720e3b3b60483 /gcc/optabs.c | |
parent | 490e23032baaece71f2ec09fa1805064b150fbc2 (diff) | |
download | gcc-5c69acb32329d49e58c26fa41ae74229a52b9106.zip gcc-5c69acb32329d49e58c26fa41ae74229a52b9106.tar.gz gcc-5c69acb32329d49e58c26fa41ae74229a52b9106.tar.bz2 |
Rename .c files to .cc files.
gcc/ada/ChangeLog:
* adadecode.c: Moved to...
* adadecode.cc: ...here.
* affinity.c: Moved to...
* affinity.cc: ...here.
* argv-lynxos178-raven-cert.c: Moved to...
* argv-lynxos178-raven-cert.cc: ...here.
* argv.c: Moved to...
* argv.cc: ...here.
* aux-io.c: Moved to...
* aux-io.cc: ...here.
* cio.c: Moved to...
* cio.cc: ...here.
* cstreams.c: Moved to...
* cstreams.cc: ...here.
* env.c: Moved to...
* env.cc: ...here.
* exit.c: Moved to...
* exit.cc: ...here.
* expect.c: Moved to...
* expect.cc: ...here.
* final.c: Moved to...
* final.cc: ...here.
* gcc-interface/cuintp.c: Moved to...
* gcc-interface/cuintp.cc: ...here.
* gcc-interface/decl.c: Moved to...
* gcc-interface/decl.cc: ...here.
* gcc-interface/misc.c: Moved to...
* gcc-interface/misc.cc: ...here.
* gcc-interface/targtyps.c: Moved to...
* gcc-interface/targtyps.cc: ...here.
* gcc-interface/trans.c: Moved to...
* gcc-interface/trans.cc: ...here.
* gcc-interface/utils.c: Moved to...
* gcc-interface/utils.cc: ...here.
* gcc-interface/utils2.c: Moved to...
* gcc-interface/utils2.cc: ...here.
* init.c: Moved to...
* init.cc: ...here.
* initialize.c: Moved to...
* initialize.cc: ...here.
* libgnarl/thread.c: Moved to...
* libgnarl/thread.cc: ...here.
* link.c: Moved to...
* link.cc: ...here.
* locales.c: Moved to...
* locales.cc: ...here.
* mkdir.c: Moved to...
* mkdir.cc: ...here.
* raise.c: Moved to...
* raise.cc: ...here.
* rtfinal.c: Moved to...
* rtfinal.cc: ...here.
* rtinit.c: Moved to...
* rtinit.cc: ...here.
* seh_init.c: Moved to...
* seh_init.cc: ...here.
* sigtramp-armdroid.c: Moved to...
* sigtramp-armdroid.cc: ...here.
* sigtramp-ios.c: Moved to...
* sigtramp-ios.cc: ...here.
* sigtramp-qnx.c: Moved to...
* sigtramp-qnx.cc: ...here.
* sigtramp-vxworks.c: Moved to...
* sigtramp-vxworks.cc: ...here.
* socket.c: Moved to...
* socket.cc: ...here.
* tracebak.c: Moved to...
* tracebak.cc: ...here.
* version.c: Moved to...
* version.cc: ...here.
* vx_stack_info.c: Moved to...
* vx_stack_info.cc: ...here.
gcc/ChangeLog:
* adjust-alignment.c: Moved to...
* adjust-alignment.cc: ...here.
* alias.c: Moved to...
* alias.cc: ...here.
* alloc-pool.c: Moved to...
* alloc-pool.cc: ...here.
* asan.c: Moved to...
* asan.cc: ...here.
* attribs.c: Moved to...
* attribs.cc: ...here.
* auto-inc-dec.c: Moved to...
* auto-inc-dec.cc: ...here.
* auto-profile.c: Moved to...
* auto-profile.cc: ...here.
* bb-reorder.c: Moved to...
* bb-reorder.cc: ...here.
* bitmap.c: Moved to...
* bitmap.cc: ...here.
* btfout.c: Moved to...
* btfout.cc: ...here.
* builtins.c: Moved to...
* builtins.cc: ...here.
* caller-save.c: Moved to...
* caller-save.cc: ...here.
* calls.c: Moved to...
* calls.cc: ...here.
* ccmp.c: Moved to...
* ccmp.cc: ...here.
* cfg.c: Moved to...
* cfg.cc: ...here.
* cfganal.c: Moved to...
* cfganal.cc: ...here.
* cfgbuild.c: Moved to...
* cfgbuild.cc: ...here.
* cfgcleanup.c: Moved to...
* cfgcleanup.cc: ...here.
* cfgexpand.c: Moved to...
* cfgexpand.cc: ...here.
* cfghooks.c: Moved to...
* cfghooks.cc: ...here.
* cfgloop.c: Moved to...
* cfgloop.cc: ...here.
* cfgloopanal.c: Moved to...
* cfgloopanal.cc: ...here.
* cfgloopmanip.c: Moved to...
* cfgloopmanip.cc: ...here.
* cfgrtl.c: Moved to...
* cfgrtl.cc: ...here.
* cgraph.c: Moved to...
* cgraph.cc: ...here.
* cgraphbuild.c: Moved to...
* cgraphbuild.cc: ...here.
* cgraphclones.c: Moved to...
* cgraphclones.cc: ...here.
* cgraphunit.c: Moved to...
* cgraphunit.cc: ...here.
* collect-utils.c: Moved to...
* collect-utils.cc: ...here.
* collect2-aix.c: Moved to...
* collect2-aix.cc: ...here.
* collect2.c: Moved to...
* collect2.cc: ...here.
* combine-stack-adj.c: Moved to...
* combine-stack-adj.cc: ...here.
* combine.c: Moved to...
* combine.cc: ...here.
* common/common-targhooks.c: Moved to...
* common/common-targhooks.cc: ...here.
* common/config/aarch64/aarch64-common.c: Moved to...
* common/config/aarch64/aarch64-common.cc: ...here.
* common/config/alpha/alpha-common.c: Moved to...
* common/config/alpha/alpha-common.cc: ...here.
* common/config/arc/arc-common.c: Moved to...
* common/config/arc/arc-common.cc: ...here.
* common/config/arm/arm-common.c: Moved to...
* common/config/arm/arm-common.cc: ...here.
* common/config/avr/avr-common.c: Moved to...
* common/config/avr/avr-common.cc: ...here.
* common/config/bfin/bfin-common.c: Moved to...
* common/config/bfin/bfin-common.cc: ...here.
* common/config/bpf/bpf-common.c: Moved to...
* common/config/bpf/bpf-common.cc: ...here.
* common/config/c6x/c6x-common.c: Moved to...
* common/config/c6x/c6x-common.cc: ...here.
* common/config/cr16/cr16-common.c: Moved to...
* common/config/cr16/cr16-common.cc: ...here.
* common/config/cris/cris-common.c: Moved to...
* common/config/cris/cris-common.cc: ...here.
* common/config/csky/csky-common.c: Moved to...
* common/config/csky/csky-common.cc: ...here.
* common/config/default-common.c: Moved to...
* common/config/default-common.cc: ...here.
* common/config/epiphany/epiphany-common.c: Moved to...
* common/config/epiphany/epiphany-common.cc: ...here.
* common/config/fr30/fr30-common.c: Moved to...
* common/config/fr30/fr30-common.cc: ...here.
* common/config/frv/frv-common.c: Moved to...
* common/config/frv/frv-common.cc: ...here.
* common/config/gcn/gcn-common.c: Moved to...
* common/config/gcn/gcn-common.cc: ...here.
* common/config/h8300/h8300-common.c: Moved to...
* common/config/h8300/h8300-common.cc: ...here.
* common/config/i386/i386-common.c: Moved to...
* common/config/i386/i386-common.cc: ...here.
* common/config/ia64/ia64-common.c: Moved to...
* common/config/ia64/ia64-common.cc: ...here.
* common/config/iq2000/iq2000-common.c: Moved to...
* common/config/iq2000/iq2000-common.cc: ...here.
* common/config/lm32/lm32-common.c: Moved to...
* common/config/lm32/lm32-common.cc: ...here.
* common/config/m32r/m32r-common.c: Moved to...
* common/config/m32r/m32r-common.cc: ...here.
* common/config/m68k/m68k-common.c: Moved to...
* common/config/m68k/m68k-common.cc: ...here.
* common/config/mcore/mcore-common.c: Moved to...
* common/config/mcore/mcore-common.cc: ...here.
* common/config/microblaze/microblaze-common.c: Moved to...
* common/config/microblaze/microblaze-common.cc: ...here.
* common/config/mips/mips-common.c: Moved to...
* common/config/mips/mips-common.cc: ...here.
* common/config/mmix/mmix-common.c: Moved to...
* common/config/mmix/mmix-common.cc: ...here.
* common/config/mn10300/mn10300-common.c: Moved to...
* common/config/mn10300/mn10300-common.cc: ...here.
* common/config/msp430/msp430-common.c: Moved to...
* common/config/msp430/msp430-common.cc: ...here.
* common/config/nds32/nds32-common.c: Moved to...
* common/config/nds32/nds32-common.cc: ...here.
* common/config/nios2/nios2-common.c: Moved to...
* common/config/nios2/nios2-common.cc: ...here.
* common/config/nvptx/nvptx-common.c: Moved to...
* common/config/nvptx/nvptx-common.cc: ...here.
* common/config/or1k/or1k-common.c: Moved to...
* common/config/or1k/or1k-common.cc: ...here.
* common/config/pa/pa-common.c: Moved to...
* common/config/pa/pa-common.cc: ...here.
* common/config/pdp11/pdp11-common.c: Moved to...
* common/config/pdp11/pdp11-common.cc: ...here.
* common/config/pru/pru-common.c: Moved to...
* common/config/pru/pru-common.cc: ...here.
* common/config/riscv/riscv-common.c: Moved to...
* common/config/riscv/riscv-common.cc: ...here.
* common/config/rs6000/rs6000-common.c: Moved to...
* common/config/rs6000/rs6000-common.cc: ...here.
* common/config/rx/rx-common.c: Moved to...
* common/config/rx/rx-common.cc: ...here.
* common/config/s390/s390-common.c: Moved to...
* common/config/s390/s390-common.cc: ...here.
* common/config/sh/sh-common.c: Moved to...
* common/config/sh/sh-common.cc: ...here.
* common/config/sparc/sparc-common.c: Moved to...
* common/config/sparc/sparc-common.cc: ...here.
* common/config/tilegx/tilegx-common.c: Moved to...
* common/config/tilegx/tilegx-common.cc: ...here.
* common/config/tilepro/tilepro-common.c: Moved to...
* common/config/tilepro/tilepro-common.cc: ...here.
* common/config/v850/v850-common.c: Moved to...
* common/config/v850/v850-common.cc: ...here.
* common/config/vax/vax-common.c: Moved to...
* common/config/vax/vax-common.cc: ...here.
* common/config/visium/visium-common.c: Moved to...
* common/config/visium/visium-common.cc: ...here.
* common/config/xstormy16/xstormy16-common.c: Moved to...
* common/config/xstormy16/xstormy16-common.cc: ...here.
* common/config/xtensa/xtensa-common.c: Moved to...
* common/config/xtensa/xtensa-common.cc: ...here.
* compare-elim.c: Moved to...
* compare-elim.cc: ...here.
* config/aarch64/aarch64-bti-insert.c: Moved to...
* config/aarch64/aarch64-bti-insert.cc: ...here.
* config/aarch64/aarch64-builtins.c: Moved to...
* config/aarch64/aarch64-builtins.cc: ...here.
* config/aarch64/aarch64-c.c: Moved to...
* config/aarch64/aarch64-c.cc: ...here.
* config/aarch64/aarch64-d.c: Moved to...
* config/aarch64/aarch64-d.cc: ...here.
* config/aarch64/aarch64.c: Moved to...
* config/aarch64/aarch64.cc: ...here.
* config/aarch64/cortex-a57-fma-steering.c: Moved to...
* config/aarch64/cortex-a57-fma-steering.cc: ...here.
* config/aarch64/driver-aarch64.c: Moved to...
* config/aarch64/driver-aarch64.cc: ...here.
* config/aarch64/falkor-tag-collision-avoidance.c: Moved to...
* config/aarch64/falkor-tag-collision-avoidance.cc: ...here.
* config/aarch64/host-aarch64-darwin.c: Moved to...
* config/aarch64/host-aarch64-darwin.cc: ...here.
* config/alpha/alpha.c: Moved to...
* config/alpha/alpha.cc: ...here.
* config/alpha/driver-alpha.c: Moved to...
* config/alpha/driver-alpha.cc: ...here.
* config/arc/arc-c.c: Moved to...
* config/arc/arc-c.cc: ...here.
* config/arc/arc.c: Moved to...
* config/arc/arc.cc: ...here.
* config/arc/driver-arc.c: Moved to...
* config/arc/driver-arc.cc: ...here.
* config/arm/aarch-common.c: Moved to...
* config/arm/aarch-common.cc: ...here.
* config/arm/arm-builtins.c: Moved to...
* config/arm/arm-builtins.cc: ...here.
* config/arm/arm-c.c: Moved to...
* config/arm/arm-c.cc: ...here.
* config/arm/arm-d.c: Moved to...
* config/arm/arm-d.cc: ...here.
* config/arm/arm.c: Moved to...
* config/arm/arm.cc: ...here.
* config/arm/driver-arm.c: Moved to...
* config/arm/driver-arm.cc: ...here.
* config/avr/avr-c.c: Moved to...
* config/avr/avr-c.cc: ...here.
* config/avr/avr-devices.c: Moved to...
* config/avr/avr-devices.cc: ...here.
* config/avr/avr-log.c: Moved to...
* config/avr/avr-log.cc: ...here.
* config/avr/avr.c: Moved to...
* config/avr/avr.cc: ...here.
* config/avr/driver-avr.c: Moved to...
* config/avr/driver-avr.cc: ...here.
* config/avr/gen-avr-mmcu-specs.c: Moved to...
* config/avr/gen-avr-mmcu-specs.cc: ...here.
* config/avr/gen-avr-mmcu-texi.c: Moved to...
* config/avr/gen-avr-mmcu-texi.cc: ...here.
* config/bfin/bfin.c: Moved to...
* config/bfin/bfin.cc: ...here.
* config/bpf/bpf.c: Moved to...
* config/bpf/bpf.cc: ...here.
* config/bpf/coreout.c: Moved to...
* config/bpf/coreout.cc: ...here.
* config/c6x/c6x.c: Moved to...
* config/c6x/c6x.cc: ...here.
* config/cr16/cr16.c: Moved to...
* config/cr16/cr16.cc: ...here.
* config/cris/cris.c: Moved to...
* config/cris/cris.cc: ...here.
* config/csky/csky.c: Moved to...
* config/csky/csky.cc: ...here.
* config/darwin-c.c: Moved to...
* config/darwin-c.cc: ...here.
* config/darwin-d.c: Moved to...
* config/darwin-d.cc: ...here.
* config/darwin-driver.c: Moved to...
* config/darwin-driver.cc: ...here.
* config/darwin-f.c: Moved to...
* config/darwin-f.cc: ...here.
* config/darwin.c: Moved to...
* config/darwin.cc: ...here.
* config/default-c.c: Moved to...
* config/default-c.cc: ...here.
* config/default-d.c: Moved to...
* config/default-d.cc: ...here.
* config/dragonfly-d.c: Moved to...
* config/dragonfly-d.cc: ...here.
* config/epiphany/epiphany.c: Moved to...
* config/epiphany/epiphany.cc: ...here.
* config/epiphany/mode-switch-use.c: Moved to...
* config/epiphany/mode-switch-use.cc: ...here.
* config/epiphany/resolve-sw-modes.c: Moved to...
* config/epiphany/resolve-sw-modes.cc: ...here.
* config/fr30/fr30.c: Moved to...
* config/fr30/fr30.cc: ...here.
* config/freebsd-d.c: Moved to...
* config/freebsd-d.cc: ...here.
* config/frv/frv.c: Moved to...
* config/frv/frv.cc: ...here.
* config/ft32/ft32.c: Moved to...
* config/ft32/ft32.cc: ...here.
* config/gcn/driver-gcn.c: Moved to...
* config/gcn/driver-gcn.cc: ...here.
* config/gcn/gcn-run.c: Moved to...
* config/gcn/gcn-run.cc: ...here.
* config/gcn/gcn-tree.c: Moved to...
* config/gcn/gcn-tree.cc: ...here.
* config/gcn/gcn.c: Moved to...
* config/gcn/gcn.cc: ...here.
* config/gcn/mkoffload.c: Moved to...
* config/gcn/mkoffload.cc: ...here.
* config/glibc-c.c: Moved to...
* config/glibc-c.cc: ...here.
* config/glibc-d.c: Moved to...
* config/glibc-d.cc: ...here.
* config/h8300/h8300.c: Moved to...
* config/h8300/h8300.cc: ...here.
* config/host-darwin.c: Moved to...
* config/host-darwin.cc: ...here.
* config/host-hpux.c: Moved to...
* config/host-hpux.cc: ...here.
* config/host-linux.c: Moved to...
* config/host-linux.cc: ...here.
* config/host-netbsd.c: Moved to...
* config/host-netbsd.cc: ...here.
* config/host-openbsd.c: Moved to...
* config/host-openbsd.cc: ...here.
* config/host-solaris.c: Moved to...
* config/host-solaris.cc: ...here.
* config/i386/djgpp.c: Moved to...
* config/i386/djgpp.cc: ...here.
* config/i386/driver-i386.c: Moved to...
* config/i386/driver-i386.cc: ...here.
* config/i386/driver-mingw32.c: Moved to...
* config/i386/driver-mingw32.cc: ...here.
* config/i386/gnu-property.c: Moved to...
* config/i386/gnu-property.cc: ...here.
* config/i386/host-cygwin.c: Moved to...
* config/i386/host-cygwin.cc: ...here.
* config/i386/host-i386-darwin.c: Moved to...
* config/i386/host-i386-darwin.cc: ...here.
* config/i386/host-mingw32.c: Moved to...
* config/i386/host-mingw32.cc: ...here.
* config/i386/i386-builtins.c: Moved to...
* config/i386/i386-builtins.cc: ...here.
* config/i386/i386-c.c: Moved to...
* config/i386/i386-c.cc: ...here.
* config/i386/i386-d.c: Moved to...
* config/i386/i386-d.cc: ...here.
* config/i386/i386-expand.c: Moved to...
* config/i386/i386-expand.cc: ...here.
* config/i386/i386-features.c: Moved to...
* config/i386/i386-features.cc: ...here.
* config/i386/i386-options.c: Moved to...
* config/i386/i386-options.cc: ...here.
* config/i386/i386.c: Moved to...
* config/i386/i386.cc: ...here.
* config/i386/intelmic-mkoffload.c: Moved to...
* config/i386/intelmic-mkoffload.cc: ...here.
* config/i386/msformat-c.c: Moved to...
* config/i386/msformat-c.cc: ...here.
* config/i386/winnt-cxx.c: Moved to...
* config/i386/winnt-cxx.cc: ...here.
* config/i386/winnt-d.c: Moved to...
* config/i386/winnt-d.cc: ...here.
* config/i386/winnt-stubs.c: Moved to...
* config/i386/winnt-stubs.cc: ...here.
* config/i386/winnt.c: Moved to...
* config/i386/winnt.cc: ...here.
* config/i386/x86-tune-sched-atom.c: Moved to...
* config/i386/x86-tune-sched-atom.cc: ...here.
* config/i386/x86-tune-sched-bd.c: Moved to...
* config/i386/x86-tune-sched-bd.cc: ...here.
* config/i386/x86-tune-sched-core.c: Moved to...
* config/i386/x86-tune-sched-core.cc: ...here.
* config/i386/x86-tune-sched.c: Moved to...
* config/i386/x86-tune-sched.cc: ...here.
* config/ia64/ia64-c.c: Moved to...
* config/ia64/ia64-c.cc: ...here.
* config/ia64/ia64.c: Moved to...
* config/ia64/ia64.cc: ...here.
* config/iq2000/iq2000.c: Moved to...
* config/iq2000/iq2000.cc: ...here.
* config/linux.c: Moved to...
* config/linux.cc: ...here.
* config/lm32/lm32.c: Moved to...
* config/lm32/lm32.cc: ...here.
* config/m32c/m32c-pragma.c: Moved to...
* config/m32c/m32c-pragma.cc: ...here.
* config/m32c/m32c.c: Moved to...
* config/m32c/m32c.cc: ...here.
* config/m32r/m32r.c: Moved to...
* config/m32r/m32r.cc: ...here.
* config/m68k/m68k.c: Moved to...
* config/m68k/m68k.cc: ...here.
* config/mcore/mcore.c: Moved to...
* config/mcore/mcore.cc: ...here.
* config/microblaze/microblaze-c.c: Moved to...
* config/microblaze/microblaze-c.cc: ...here.
* config/microblaze/microblaze.c: Moved to...
* config/microblaze/microblaze.cc: ...here.
* config/mips/driver-native.c: Moved to...
* config/mips/driver-native.cc: ...here.
* config/mips/frame-header-opt.c: Moved to...
* config/mips/frame-header-opt.cc: ...here.
* config/mips/mips-d.c: Moved to...
* config/mips/mips-d.cc: ...here.
* config/mips/mips.c: Moved to...
* config/mips/mips.cc: ...here.
* config/mmix/mmix.c: Moved to...
* config/mmix/mmix.cc: ...here.
* config/mn10300/mn10300.c: Moved to...
* config/mn10300/mn10300.cc: ...here.
* config/moxie/moxie.c: Moved to...
* config/moxie/moxie.cc: ...here.
* config/msp430/driver-msp430.c: Moved to...
* config/msp430/driver-msp430.cc: ...here.
* config/msp430/msp430-c.c: Moved to...
* config/msp430/msp430-c.cc: ...here.
* config/msp430/msp430-devices.c: Moved to...
* config/msp430/msp430-devices.cc: ...here.
* config/msp430/msp430.c: Moved to...
* config/msp430/msp430.cc: ...here.
* config/nds32/nds32-cost.c: Moved to...
* config/nds32/nds32-cost.cc: ...here.
* config/nds32/nds32-fp-as-gp.c: Moved to...
* config/nds32/nds32-fp-as-gp.cc: ...here.
* config/nds32/nds32-intrinsic.c: Moved to...
* config/nds32/nds32-intrinsic.cc: ...here.
* config/nds32/nds32-isr.c: Moved to...
* config/nds32/nds32-isr.cc: ...here.
* config/nds32/nds32-md-auxiliary.c: Moved to...
* config/nds32/nds32-md-auxiliary.cc: ...here.
* config/nds32/nds32-memory-manipulation.c: Moved to...
* config/nds32/nds32-memory-manipulation.cc: ...here.
* config/nds32/nds32-pipelines-auxiliary.c: Moved to...
* config/nds32/nds32-pipelines-auxiliary.cc: ...here.
* config/nds32/nds32-predicates.c: Moved to...
* config/nds32/nds32-predicates.cc: ...here.
* config/nds32/nds32-relax-opt.c: Moved to...
* config/nds32/nds32-relax-opt.cc: ...here.
* config/nds32/nds32-utils.c: Moved to...
* config/nds32/nds32-utils.cc: ...here.
* config/nds32/nds32.c: Moved to...
* config/nds32/nds32.cc: ...here.
* config/netbsd-d.c: Moved to...
* config/netbsd-d.cc: ...here.
* config/netbsd.c: Moved to...
* config/netbsd.cc: ...here.
* config/nios2/nios2.c: Moved to...
* config/nios2/nios2.cc: ...here.
* config/nvptx/mkoffload.c: Moved to...
* config/nvptx/mkoffload.cc: ...here.
* config/nvptx/nvptx-c.c: Moved to...
* config/nvptx/nvptx-c.cc: ...here.
* config/nvptx/nvptx.c: Moved to...
* config/nvptx/nvptx.cc: ...here.
* config/openbsd-d.c: Moved to...
* config/openbsd-d.cc: ...here.
* config/or1k/or1k.c: Moved to...
* config/or1k/or1k.cc: ...here.
* config/pa/pa-d.c: Moved to...
* config/pa/pa-d.cc: ...here.
* config/pa/pa.c: Moved to...
* config/pa/pa.cc: ...here.
* config/pdp11/pdp11.c: Moved to...
* config/pdp11/pdp11.cc: ...here.
* config/pru/pru-passes.c: Moved to...
* config/pru/pru-passes.cc: ...here.
* config/pru/pru-pragma.c: Moved to...
* config/pru/pru-pragma.cc: ...here.
* config/pru/pru.c: Moved to...
* config/pru/pru.cc: ...here.
* config/riscv/riscv-builtins.c: Moved to...
* config/riscv/riscv-builtins.cc: ...here.
* config/riscv/riscv-c.c: Moved to...
* config/riscv/riscv-c.cc: ...here.
* config/riscv/riscv-d.c: Moved to...
* config/riscv/riscv-d.cc: ...here.
* config/riscv/riscv-shorten-memrefs.c: Moved to...
* config/riscv/riscv-shorten-memrefs.cc: ...here.
* config/riscv/riscv-sr.c: Moved to...
* config/riscv/riscv-sr.cc: ...here.
* config/riscv/riscv.c: Moved to...
* config/riscv/riscv.cc: ...here.
* config/rl78/rl78-c.c: Moved to...
* config/rl78/rl78-c.cc: ...here.
* config/rl78/rl78.c: Moved to...
* config/rl78/rl78.cc: ...here.
* config/rs6000/driver-rs6000.c: Moved to...
* config/rs6000/driver-rs6000.cc: ...here.
* config/rs6000/host-darwin.c: Moved to...
* config/rs6000/host-darwin.cc: ...here.
* config/rs6000/host-ppc64-darwin.c: Moved to...
* config/rs6000/host-ppc64-darwin.cc: ...here.
* config/rs6000/rbtree.c: Moved to...
* config/rs6000/rbtree.cc: ...here.
* config/rs6000/rs6000-c.c: Moved to...
* config/rs6000/rs6000-c.cc: ...here.
* config/rs6000/rs6000-call.c: Moved to...
* config/rs6000/rs6000-call.cc: ...here.
* config/rs6000/rs6000-d.c: Moved to...
* config/rs6000/rs6000-d.cc: ...here.
* config/rs6000/rs6000-gen-builtins.c: Moved to...
* config/rs6000/rs6000-gen-builtins.cc: ...here.
* config/rs6000/rs6000-linux.c: Moved to...
* config/rs6000/rs6000-linux.cc: ...here.
* config/rs6000/rs6000-logue.c: Moved to...
* config/rs6000/rs6000-logue.cc: ...here.
* config/rs6000/rs6000-p8swap.c: Moved to...
* config/rs6000/rs6000-p8swap.cc: ...here.
* config/rs6000/rs6000-pcrel-opt.c: Moved to...
* config/rs6000/rs6000-pcrel-opt.cc: ...here.
* config/rs6000/rs6000-string.c: Moved to...
* config/rs6000/rs6000-string.cc: ...here.
* config/rs6000/rs6000.c: Moved to...
* config/rs6000/rs6000.cc: ...here.
* config/rx/rx.c: Moved to...
* config/rx/rx.cc: ...here.
* config/s390/driver-native.c: Moved to...
* config/s390/driver-native.cc: ...here.
* config/s390/s390-c.c: Moved to...
* config/s390/s390-c.cc: ...here.
* config/s390/s390-d.c: Moved to...
* config/s390/s390-d.cc: ...here.
* config/s390/s390.c: Moved to...
* config/s390/s390.cc: ...here.
* config/sh/divtab-sh4-300.c: Moved to...
* config/sh/divtab-sh4-300.cc: ...here.
* config/sh/divtab-sh4.c: Moved to...
* config/sh/divtab-sh4.cc: ...here.
* config/sh/divtab.c: Moved to...
* config/sh/divtab.cc: ...here.
* config/sh/sh-c.c: Moved to...
* config/sh/sh-c.cc: ...here.
* config/sh/sh.c: Moved to...
* config/sh/sh.cc: ...here.
* config/sol2-c.c: Moved to...
* config/sol2-c.cc: ...here.
* config/sol2-cxx.c: Moved to...
* config/sol2-cxx.cc: ...here.
* config/sol2-d.c: Moved to...
* config/sol2-d.cc: ...here.
* config/sol2-stubs.c: Moved to...
* config/sol2-stubs.cc: ...here.
* config/sol2.c: Moved to...
* config/sol2.cc: ...here.
* config/sparc/driver-sparc.c: Moved to...
* config/sparc/driver-sparc.cc: ...here.
* config/sparc/sparc-c.c: Moved to...
* config/sparc/sparc-c.cc: ...here.
* config/sparc/sparc-d.c: Moved to...
* config/sparc/sparc-d.cc: ...here.
* config/sparc/sparc.c: Moved to...
* config/sparc/sparc.cc: ...here.
* config/stormy16/stormy16.c: Moved to...
* config/stormy16/stormy16.cc: ...here.
* config/tilegx/mul-tables.c: Moved to...
* config/tilegx/mul-tables.cc: ...here.
* config/tilegx/tilegx-c.c: Moved to...
* config/tilegx/tilegx-c.cc: ...here.
* config/tilegx/tilegx.c: Moved to...
* config/tilegx/tilegx.cc: ...here.
* config/tilepro/mul-tables.c: Moved to...
* config/tilepro/mul-tables.cc: ...here.
* config/tilepro/tilepro-c.c: Moved to...
* config/tilepro/tilepro-c.cc: ...here.
* config/tilepro/tilepro.c: Moved to...
* config/tilepro/tilepro.cc: ...here.
* config/v850/v850-c.c: Moved to...
* config/v850/v850-c.cc: ...here.
* config/v850/v850.c: Moved to...
* config/v850/v850.cc: ...here.
* config/vax/vax.c: Moved to...
* config/vax/vax.cc: ...here.
* config/visium/visium.c: Moved to...
* config/visium/visium.cc: ...here.
* config/vms/vms-c.c: Moved to...
* config/vms/vms-c.cc: ...here.
* config/vms/vms-f.c: Moved to...
* config/vms/vms-f.cc: ...here.
* config/vms/vms.c: Moved to...
* config/vms/vms.cc: ...here.
* config/vxworks-c.c: Moved to...
* config/vxworks-c.cc: ...here.
* config/vxworks.c: Moved to...
* config/vxworks.cc: ...here.
* config/winnt-c.c: Moved to...
* config/winnt-c.cc: ...here.
* config/xtensa/xtensa.c: Moved to...
* config/xtensa/xtensa.cc: ...here.
* context.c: Moved to...
* context.cc: ...here.
* convert.c: Moved to...
* convert.cc: ...here.
* coverage.c: Moved to...
* coverage.cc: ...here.
* cppbuiltin.c: Moved to...
* cppbuiltin.cc: ...here.
* cppdefault.c: Moved to...
* cppdefault.cc: ...here.
* cprop.c: Moved to...
* cprop.cc: ...here.
* cse.c: Moved to...
* cse.cc: ...here.
* cselib.c: Moved to...
* cselib.cc: ...here.
* ctfc.c: Moved to...
* ctfc.cc: ...here.
* ctfout.c: Moved to...
* ctfout.cc: ...here.
* data-streamer-in.c: Moved to...
* data-streamer-in.cc: ...here.
* data-streamer-out.c: Moved to...
* data-streamer-out.cc: ...here.
* data-streamer.c: Moved to...
* data-streamer.cc: ...here.
* dbgcnt.c: Moved to...
* dbgcnt.cc: ...here.
* dbxout.c: Moved to...
* dbxout.cc: ...here.
* dce.c: Moved to...
* dce.cc: ...here.
* ddg.c: Moved to...
* ddg.cc: ...here.
* debug.c: Moved to...
* debug.cc: ...here.
* df-core.c: Moved to...
* df-core.cc: ...here.
* df-problems.c: Moved to...
* df-problems.cc: ...here.
* df-scan.c: Moved to...
* df-scan.cc: ...here.
* dfp.c: Moved to...
* dfp.cc: ...here.
* diagnostic-color.c: Moved to...
* diagnostic-color.cc: ...here.
* diagnostic-show-locus.c: Moved to...
* diagnostic-show-locus.cc: ...here.
* diagnostic-spec.c: Moved to...
* diagnostic-spec.cc: ...here.
* diagnostic.c: Moved to...
* diagnostic.cc: ...here.
* dojump.c: Moved to...
* dojump.cc: ...here.
* dominance.c: Moved to...
* dominance.cc: ...here.
* domwalk.c: Moved to...
* domwalk.cc: ...here.
* double-int.c: Moved to...
* double-int.cc: ...here.
* dse.c: Moved to...
* dse.cc: ...here.
* dumpfile.c: Moved to...
* dumpfile.cc: ...here.
* dwarf2asm.c: Moved to...
* dwarf2asm.cc: ...here.
* dwarf2cfi.c: Moved to...
* dwarf2cfi.cc: ...here.
* dwarf2ctf.c: Moved to...
* dwarf2ctf.cc: ...here.
* dwarf2out.c: Moved to...
* dwarf2out.cc: ...here.
* early-remat.c: Moved to...
* early-remat.cc: ...here.
* edit-context.c: Moved to...
* edit-context.cc: ...here.
* emit-rtl.c: Moved to...
* emit-rtl.cc: ...here.
* errors.c: Moved to...
* errors.cc: ...here.
* et-forest.c: Moved to...
* et-forest.cc: ...here.
* except.c: Moved to...
* except.cc: ...here.
* explow.c: Moved to...
* explow.cc: ...here.
* expmed.c: Moved to...
* expmed.cc: ...here.
* expr.c: Moved to...
* expr.cc: ...here.
* fibonacci_heap.c: Moved to...
* fibonacci_heap.cc: ...here.
* file-find.c: Moved to...
* file-find.cc: ...here.
* file-prefix-map.c: Moved to...
* file-prefix-map.cc: ...here.
* final.c: Moved to...
* final.cc: ...here.
* fixed-value.c: Moved to...
* fixed-value.cc: ...here.
* fold-const-call.c: Moved to...
* fold-const-call.cc: ...here.
* fold-const.c: Moved to...
* fold-const.cc: ...here.
* fp-test.c: Moved to...
* fp-test.cc: ...here.
* function-tests.c: Moved to...
* function-tests.cc: ...here.
* function.c: Moved to...
* function.cc: ...here.
* fwprop.c: Moved to...
* fwprop.cc: ...here.
* gcc-ar.c: Moved to...
* gcc-ar.cc: ...here.
* gcc-main.c: Moved to...
* gcc-main.cc: ...here.
* gcc-rich-location.c: Moved to...
* gcc-rich-location.cc: ...here.
* gcc.c: Moved to...
* gcc.cc: ...here.
* gcov-dump.c: Moved to...
* gcov-dump.cc: ...here.
* gcov-io.c: Moved to...
* gcov-io.cc: ...here.
* gcov-tool.c: Moved to...
* gcov-tool.cc: ...here.
* gcov.c: Moved to...
* gcov.cc: ...here.
* gcse-common.c: Moved to...
* gcse-common.cc: ...here.
* gcse.c: Moved to...
* gcse.cc: ...here.
* genattr-common.c: Moved to...
* genattr-common.cc: ...here.
* genattr.c: Moved to...
* genattr.cc: ...here.
* genattrtab.c: Moved to...
* genattrtab.cc: ...here.
* genautomata.c: Moved to...
* genautomata.cc: ...here.
* gencfn-macros.c: Moved to...
* gencfn-macros.cc: ...here.
* gencheck.c: Moved to...
* gencheck.cc: ...here.
* genchecksum.c: Moved to...
* genchecksum.cc: ...here.
* gencodes.c: Moved to...
* gencodes.cc: ...here.
* genconditions.c: Moved to...
* genconditions.cc: ...here.
* genconfig.c: Moved to...
* genconfig.cc: ...here.
* genconstants.c: Moved to...
* genconstants.cc: ...here.
* genemit.c: Moved to...
* genemit.cc: ...here.
* genenums.c: Moved to...
* genenums.cc: ...here.
* generic-match-head.c: Moved to...
* generic-match-head.cc: ...here.
* genextract.c: Moved to...
* genextract.cc: ...here.
* genflags.c: Moved to...
* genflags.cc: ...here.
* gengenrtl.c: Moved to...
* gengenrtl.cc: ...here.
* gengtype-parse.c: Moved to...
* gengtype-parse.cc: ...here.
* gengtype-state.c: Moved to...
* gengtype-state.cc: ...here.
* gengtype.c: Moved to...
* gengtype.cc: ...here.
* genhooks.c: Moved to...
* genhooks.cc: ...here.
* genmatch.c: Moved to...
* genmatch.cc: ...here.
* genmddeps.c: Moved to...
* genmddeps.cc: ...here.
* genmddump.c: Moved to...
* genmddump.cc: ...here.
* genmodes.c: Moved to...
* genmodes.cc: ...here.
* genopinit.c: Moved to...
* genopinit.cc: ...here.
* genoutput.c: Moved to...
* genoutput.cc: ...here.
* genpeep.c: Moved to...
* genpeep.cc: ...here.
* genpreds.c: Moved to...
* genpreds.cc: ...here.
* genrecog.c: Moved to...
* genrecog.cc: ...here.
* gensupport.c: Moved to...
* gensupport.cc: ...here.
* gentarget-def.c: Moved to...
* gentarget-def.cc: ...here.
* genversion.c: Moved to...
* genversion.cc: ...here.
* ggc-common.c: Moved to...
* ggc-common.cc: ...here.
* ggc-none.c: Moved to...
* ggc-none.cc: ...here.
* ggc-page.c: Moved to...
* ggc-page.cc: ...here.
* ggc-tests.c: Moved to...
* ggc-tests.cc: ...here.
* gimple-builder.c: Moved to...
* gimple-builder.cc: ...here.
* gimple-expr.c: Moved to...
* gimple-expr.cc: ...here.
* gimple-fold.c: Moved to...
* gimple-fold.cc: ...here.
* gimple-iterator.c: Moved to...
* gimple-iterator.cc: ...here.
* gimple-laddress.c: Moved to...
* gimple-laddress.cc: ...here.
* gimple-loop-jam.c: Moved to...
* gimple-loop-jam.cc: ...here.
* gimple-low.c: Moved to...
* gimple-low.cc: ...here.
* gimple-match-head.c: Moved to...
* gimple-match-head.cc: ...here.
* gimple-pretty-print.c: Moved to...
* gimple-pretty-print.cc: ...here.
* gimple-ssa-backprop.c: Moved to...
* gimple-ssa-backprop.cc: ...here.
* gimple-ssa-evrp-analyze.c: Moved to...
* gimple-ssa-evrp-analyze.cc: ...here.
* gimple-ssa-evrp.c: Moved to...
* gimple-ssa-evrp.cc: ...here.
* gimple-ssa-isolate-paths.c: Moved to...
* gimple-ssa-isolate-paths.cc: ...here.
* gimple-ssa-nonnull-compare.c: Moved to...
* gimple-ssa-nonnull-compare.cc: ...here.
* gimple-ssa-split-paths.c: Moved to...
* gimple-ssa-split-paths.cc: ...here.
* gimple-ssa-sprintf.c: Moved to...
* gimple-ssa-sprintf.cc: ...here.
* gimple-ssa-store-merging.c: Moved to...
* gimple-ssa-store-merging.cc: ...here.
* gimple-ssa-strength-reduction.c: Moved to...
* gimple-ssa-strength-reduction.cc: ...here.
* gimple-ssa-warn-alloca.c: Moved to...
* gimple-ssa-warn-alloca.cc: ...here.
* gimple-ssa-warn-restrict.c: Moved to...
* gimple-ssa-warn-restrict.cc: ...here.
* gimple-streamer-in.c: Moved to...
* gimple-streamer-in.cc: ...here.
* gimple-streamer-out.c: Moved to...
* gimple-streamer-out.cc: ...here.
* gimple-walk.c: Moved to...
* gimple-walk.cc: ...here.
* gimple-warn-recursion.c: Moved to...
* gimple-warn-recursion.cc: ...here.
* gimple.c: Moved to...
* gimple.cc: ...here.
* gimplify-me.c: Moved to...
* gimplify-me.cc: ...here.
* gimplify.c: Moved to...
* gimplify.cc: ...here.
* godump.c: Moved to...
* godump.cc: ...here.
* graph.c: Moved to...
* graph.cc: ...here.
* graphds.c: Moved to...
* graphds.cc: ...here.
* graphite-dependences.c: Moved to...
* graphite-dependences.cc: ...here.
* graphite-isl-ast-to-gimple.c: Moved to...
* graphite-isl-ast-to-gimple.cc: ...here.
* graphite-optimize-isl.c: Moved to...
* graphite-optimize-isl.cc: ...here.
* graphite-poly.c: Moved to...
* graphite-poly.cc: ...here.
* graphite-scop-detection.c: Moved to...
* graphite-scop-detection.cc: ...here.
* graphite-sese-to-poly.c: Moved to...
* graphite-sese-to-poly.cc: ...here.
* graphite.c: Moved to...
* graphite.cc: ...here.
* haifa-sched.c: Moved to...
* haifa-sched.cc: ...here.
* hash-map-tests.c: Moved to...
* hash-map-tests.cc: ...here.
* hash-set-tests.c: Moved to...
* hash-set-tests.cc: ...here.
* hash-table.c: Moved to...
* hash-table.cc: ...here.
* hooks.c: Moved to...
* hooks.cc: ...here.
* host-default.c: Moved to...
* host-default.cc: ...here.
* hw-doloop.c: Moved to...
* hw-doloop.cc: ...here.
* hwint.c: Moved to...
* hwint.cc: ...here.
* ifcvt.c: Moved to...
* ifcvt.cc: ...here.
* inchash.c: Moved to...
* inchash.cc: ...here.
* incpath.c: Moved to...
* incpath.cc: ...here.
* init-regs.c: Moved to...
* init-regs.cc: ...here.
* input.c: Moved to...
* input.cc: ...here.
* internal-fn.c: Moved to...
* internal-fn.cc: ...here.
* intl.c: Moved to...
* intl.cc: ...here.
* ipa-comdats.c: Moved to...
* ipa-comdats.cc: ...here.
* ipa-cp.c: Moved to...
* ipa-cp.cc: ...here.
* ipa-devirt.c: Moved to...
* ipa-devirt.cc: ...here.
* ipa-fnsummary.c: Moved to...
* ipa-fnsummary.cc: ...here.
* ipa-icf-gimple.c: Moved to...
* ipa-icf-gimple.cc: ...here.
* ipa-icf.c: Moved to...
* ipa-icf.cc: ...here.
* ipa-inline-analysis.c: Moved to...
* ipa-inline-analysis.cc: ...here.
* ipa-inline-transform.c: Moved to...
* ipa-inline-transform.cc: ...here.
* ipa-inline.c: Moved to...
* ipa-inline.cc: ...here.
* ipa-modref-tree.c: Moved to...
* ipa-modref-tree.cc: ...here.
* ipa-modref.c: Moved to...
* ipa-modref.cc: ...here.
* ipa-param-manipulation.c: Moved to...
* ipa-param-manipulation.cc: ...here.
* ipa-polymorphic-call.c: Moved to...
* ipa-polymorphic-call.cc: ...here.
* ipa-predicate.c: Moved to...
* ipa-predicate.cc: ...here.
* ipa-profile.c: Moved to...
* ipa-profile.cc: ...here.
* ipa-prop.c: Moved to...
* ipa-prop.cc: ...here.
* ipa-pure-const.c: Moved to...
* ipa-pure-const.cc: ...here.
* ipa-ref.c: Moved to...
* ipa-ref.cc: ...here.
* ipa-reference.c: Moved to...
* ipa-reference.cc: ...here.
* ipa-split.c: Moved to...
* ipa-split.cc: ...here.
* ipa-sra.c: Moved to...
* ipa-sra.cc: ...here.
* ipa-utils.c: Moved to...
* ipa-utils.cc: ...here.
* ipa-visibility.c: Moved to...
* ipa-visibility.cc: ...here.
* ipa.c: Moved to...
* ipa.cc: ...here.
* ira-build.c: Moved to...
* ira-build.cc: ...here.
* ira-color.c: Moved to...
* ira-color.cc: ...here.
* ira-conflicts.c: Moved to...
* ira-conflicts.cc: ...here.
* ira-costs.c: Moved to...
* ira-costs.cc: ...here.
* ira-emit.c: Moved to...
* ira-emit.cc: ...here.
* ira-lives.c: Moved to...
* ira-lives.cc: ...here.
* ira.c: Moved to...
* ira.cc: ...here.
* jump.c: Moved to...
* jump.cc: ...here.
* langhooks.c: Moved to...
* langhooks.cc: ...here.
* lcm.c: Moved to...
* lcm.cc: ...here.
* lists.c: Moved to...
* lists.cc: ...here.
* loop-doloop.c: Moved to...
* loop-doloop.cc: ...here.
* loop-init.c: Moved to...
* loop-init.cc: ...here.
* loop-invariant.c: Moved to...
* loop-invariant.cc: ...here.
* loop-iv.c: Moved to...
* loop-iv.cc: ...here.
* loop-unroll.c: Moved to...
* loop-unroll.cc: ...here.
* lower-subreg.c: Moved to...
* lower-subreg.cc: ...here.
* lra-assigns.c: Moved to...
* lra-assigns.cc: ...here.
* lra-coalesce.c: Moved to...
* lra-coalesce.cc: ...here.
* lra-constraints.c: Moved to...
* lra-constraints.cc: ...here.
* lra-eliminations.c: Moved to...
* lra-eliminations.cc: ...here.
* lra-lives.c: Moved to...
* lra-lives.cc: ...here.
* lra-remat.c: Moved to...
* lra-remat.cc: ...here.
* lra-spills.c: Moved to...
* lra-spills.cc: ...here.
* lra.c: Moved to...
* lra.cc: ...here.
* lto-cgraph.c: Moved to...
* lto-cgraph.cc: ...here.
* lto-compress.c: Moved to...
* lto-compress.cc: ...here.
* lto-opts.c: Moved to...
* lto-opts.cc: ...here.
* lto-section-in.c: Moved to...
* lto-section-in.cc: ...here.
* lto-section-out.c: Moved to...
* lto-section-out.cc: ...here.
* lto-streamer-in.c: Moved to...
* lto-streamer-in.cc: ...here.
* lto-streamer-out.c: Moved to...
* lto-streamer-out.cc: ...here.
* lto-streamer.c: Moved to...
* lto-streamer.cc: ...here.
* lto-wrapper.c: Moved to...
* lto-wrapper.cc: ...here.
* main.c: Moved to...
* main.cc: ...here.
* mcf.c: Moved to...
* mcf.cc: ...here.
* mode-switching.c: Moved to...
* mode-switching.cc: ...here.
* modulo-sched.c: Moved to...
* modulo-sched.cc: ...here.
* multiple_target.c: Moved to...
* multiple_target.cc: ...here.
* omp-expand.c: Moved to...
* omp-expand.cc: ...here.
* omp-general.c: Moved to...
* omp-general.cc: ...here.
* omp-low.c: Moved to...
* omp-low.cc: ...here.
* omp-offload.c: Moved to...
* omp-offload.cc: ...here.
* omp-simd-clone.c: Moved to...
* omp-simd-clone.cc: ...here.
* opt-suggestions.c: Moved to...
* opt-suggestions.cc: ...here.
* optabs-libfuncs.c: Moved to...
* optabs-libfuncs.cc: ...here.
* optabs-query.c: Moved to...
* optabs-query.cc: ...here.
* optabs-tree.c: Moved to...
* optabs-tree.cc: ...here.
* optabs.c: Moved to...
* optabs.cc: ...here.
* opts-common.c: Moved to...
* opts-common.cc: ...here.
* opts-global.c: Moved to...
* opts-global.cc: ...here.
* opts.c: Moved to...
* opts.cc: ...here.
* passes.c: Moved to...
* passes.cc: ...here.
* plugin.c: Moved to...
* plugin.cc: ...here.
* postreload-gcse.c: Moved to...
* postreload-gcse.cc: ...here.
* postreload.c: Moved to...
* postreload.cc: ...here.
* predict.c: Moved to...
* predict.cc: ...here.
* prefix.c: Moved to...
* prefix.cc: ...here.
* pretty-print.c: Moved to...
* pretty-print.cc: ...here.
* print-rtl-function.c: Moved to...
* print-rtl-function.cc: ...here.
* print-rtl.c: Moved to...
* print-rtl.cc: ...here.
* print-tree.c: Moved to...
* print-tree.cc: ...here.
* profile-count.c: Moved to...
* profile-count.cc: ...here.
* profile.c: Moved to...
* profile.cc: ...here.
* read-md.c: Moved to...
* read-md.cc: ...here.
* read-rtl-function.c: Moved to...
* read-rtl-function.cc: ...here.
* read-rtl.c: Moved to...
* read-rtl.cc: ...here.
* real.c: Moved to...
* real.cc: ...here.
* realmpfr.c: Moved to...
* realmpfr.cc: ...here.
* recog.c: Moved to...
* recog.cc: ...here.
* ree.c: Moved to...
* ree.cc: ...here.
* reg-stack.c: Moved to...
* reg-stack.cc: ...here.
* regcprop.c: Moved to...
* regcprop.cc: ...here.
* reginfo.c: Moved to...
* reginfo.cc: ...here.
* regrename.c: Moved to...
* regrename.cc: ...here.
* regstat.c: Moved to...
* regstat.cc: ...here.
* reload.c: Moved to...
* reload.cc: ...here.
* reload1.c: Moved to...
* reload1.cc: ...here.
* reorg.c: Moved to...
* reorg.cc: ...here.
* resource.c: Moved to...
* resource.cc: ...here.
* rtl-error.c: Moved to...
* rtl-error.cc: ...here.
* rtl-tests.c: Moved to...
* rtl-tests.cc: ...here.
* rtl.c: Moved to...
* rtl.cc: ...here.
* rtlanal.c: Moved to...
* rtlanal.cc: ...here.
* rtlhash.c: Moved to...
* rtlhash.cc: ...here.
* rtlhooks.c: Moved to...
* rtlhooks.cc: ...here.
* rtx-vector-builder.c: Moved to...
* rtx-vector-builder.cc: ...here.
* run-rtl-passes.c: Moved to...
* run-rtl-passes.cc: ...here.
* sancov.c: Moved to...
* sancov.cc: ...here.
* sanopt.c: Moved to...
* sanopt.cc: ...here.
* sbitmap.c: Moved to...
* sbitmap.cc: ...here.
* sched-deps.c: Moved to...
* sched-deps.cc: ...here.
* sched-ebb.c: Moved to...
* sched-ebb.cc: ...here.
* sched-rgn.c: Moved to...
* sched-rgn.cc: ...here.
* sel-sched-dump.c: Moved to...
* sel-sched-dump.cc: ...here.
* sel-sched-ir.c: Moved to...
* sel-sched-ir.cc: ...here.
* sel-sched.c: Moved to...
* sel-sched.cc: ...here.
* selftest-diagnostic.c: Moved to...
* selftest-diagnostic.cc: ...here.
* selftest-rtl.c: Moved to...
* selftest-rtl.cc: ...here.
* selftest-run-tests.c: Moved to...
* selftest-run-tests.cc: ...here.
* selftest.c: Moved to...
* selftest.cc: ...here.
* sese.c: Moved to...
* sese.cc: ...here.
* shrink-wrap.c: Moved to...
* shrink-wrap.cc: ...here.
* simplify-rtx.c: Moved to...
* simplify-rtx.cc: ...here.
* sparseset.c: Moved to...
* sparseset.cc: ...here.
* spellcheck-tree.c: Moved to...
* spellcheck-tree.cc: ...here.
* spellcheck.c: Moved to...
* spellcheck.cc: ...here.
* sreal.c: Moved to...
* sreal.cc: ...here.
* stack-ptr-mod.c: Moved to...
* stack-ptr-mod.cc: ...here.
* statistics.c: Moved to...
* statistics.cc: ...here.
* stmt.c: Moved to...
* stmt.cc: ...here.
* stor-layout.c: Moved to...
* stor-layout.cc: ...here.
* store-motion.c: Moved to...
* store-motion.cc: ...here.
* streamer-hooks.c: Moved to...
* streamer-hooks.cc: ...here.
* stringpool.c: Moved to...
* stringpool.cc: ...here.
* substring-locations.c: Moved to...
* substring-locations.cc: ...here.
* symtab.c: Moved to...
* symtab.cc: ...here.
* target-globals.c: Moved to...
* target-globals.cc: ...here.
* targhooks.c: Moved to...
* targhooks.cc: ...here.
* timevar.c: Moved to...
* timevar.cc: ...here.
* toplev.c: Moved to...
* toplev.cc: ...here.
* tracer.c: Moved to...
* tracer.cc: ...here.
* trans-mem.c: Moved to...
* trans-mem.cc: ...here.
* tree-affine.c: Moved to...
* tree-affine.cc: ...here.
* tree-call-cdce.c: Moved to...
* tree-call-cdce.cc: ...here.
* tree-cfg.c: Moved to...
* tree-cfg.cc: ...here.
* tree-cfgcleanup.c: Moved to...
* tree-cfgcleanup.cc: ...here.
* tree-chrec.c: Moved to...
* tree-chrec.cc: ...here.
* tree-complex.c: Moved to...
* tree-complex.cc: ...here.
* tree-data-ref.c: Moved to...
* tree-data-ref.cc: ...here.
* tree-dfa.c: Moved to...
* tree-dfa.cc: ...here.
* tree-diagnostic.c: Moved to...
* tree-diagnostic.cc: ...here.
* tree-dump.c: Moved to...
* tree-dump.cc: ...here.
* tree-eh.c: Moved to...
* tree-eh.cc: ...here.
* tree-emutls.c: Moved to...
* tree-emutls.cc: ...here.
* tree-if-conv.c: Moved to...
* tree-if-conv.cc: ...here.
* tree-inline.c: Moved to...
* tree-inline.cc: ...here.
* tree-into-ssa.c: Moved to...
* tree-into-ssa.cc: ...here.
* tree-iterator.c: Moved to...
* tree-iterator.cc: ...here.
* tree-loop-distribution.c: Moved to...
* tree-loop-distribution.cc: ...here.
* tree-nested.c: Moved to...
* tree-nested.cc: ...here.
* tree-nrv.c: Moved to...
* tree-nrv.cc: ...here.
* tree-object-size.c: Moved to...
* tree-object-size.cc: ...here.
* tree-outof-ssa.c: Moved to...
* tree-outof-ssa.cc: ...here.
* tree-parloops.c: Moved to...
* tree-parloops.cc: ...here.
* tree-phinodes.c: Moved to...
* tree-phinodes.cc: ...here.
* tree-predcom.c: Moved to...
* tree-predcom.cc: ...here.
* tree-pretty-print.c: Moved to...
* tree-pretty-print.cc: ...here.
* tree-profile.c: Moved to...
* tree-profile.cc: ...here.
* tree-scalar-evolution.c: Moved to...
* tree-scalar-evolution.cc: ...here.
* tree-sra.c: Moved to...
* tree-sra.cc: ...here.
* tree-ssa-address.c: Moved to...
* tree-ssa-address.cc: ...here.
* tree-ssa-alias.c: Moved to...
* tree-ssa-alias.cc: ...here.
* tree-ssa-ccp.c: Moved to...
* tree-ssa-ccp.cc: ...here.
* tree-ssa-coalesce.c: Moved to...
* tree-ssa-coalesce.cc: ...here.
* tree-ssa-copy.c: Moved to...
* tree-ssa-copy.cc: ...here.
* tree-ssa-dce.c: Moved to...
* tree-ssa-dce.cc: ...here.
* tree-ssa-dom.c: Moved to...
* tree-ssa-dom.cc: ...here.
* tree-ssa-dse.c: Moved to...
* tree-ssa-dse.cc: ...here.
* tree-ssa-forwprop.c: Moved to...
* tree-ssa-forwprop.cc: ...here.
* tree-ssa-ifcombine.c: Moved to...
* tree-ssa-ifcombine.cc: ...here.
* tree-ssa-live.c: Moved to...
* tree-ssa-live.cc: ...here.
* tree-ssa-loop-ch.c: Moved to...
* tree-ssa-loop-ch.cc: ...here.
* tree-ssa-loop-im.c: Moved to...
* tree-ssa-loop-im.cc: ...here.
* tree-ssa-loop-ivcanon.c: Moved to...
* tree-ssa-loop-ivcanon.cc: ...here.
* tree-ssa-loop-ivopts.c: Moved to...
* tree-ssa-loop-ivopts.cc: ...here.
* tree-ssa-loop-manip.c: Moved to...
* tree-ssa-loop-manip.cc: ...here.
* tree-ssa-loop-niter.c: Moved to...
* tree-ssa-loop-niter.cc: ...here.
* tree-ssa-loop-prefetch.c: Moved to...
* tree-ssa-loop-prefetch.cc: ...here.
* tree-ssa-loop-split.c: Moved to...
* tree-ssa-loop-split.cc: ...here.
* tree-ssa-loop-unswitch.c: Moved to...
* tree-ssa-loop-unswitch.cc: ...here.
* tree-ssa-loop.c: Moved to...
* tree-ssa-loop.cc: ...here.
* tree-ssa-math-opts.c: Moved to...
* tree-ssa-math-opts.cc: ...here.
* tree-ssa-operands.c: Moved to...
* tree-ssa-operands.cc: ...here.
* tree-ssa-phiopt.c: Moved to...
* tree-ssa-phiopt.cc: ...here.
* tree-ssa-phiprop.c: Moved to...
* tree-ssa-phiprop.cc: ...here.
* tree-ssa-pre.c: Moved to...
* tree-ssa-pre.cc: ...here.
* tree-ssa-propagate.c: Moved to...
* tree-ssa-propagate.cc: ...here.
* tree-ssa-reassoc.c: Moved to...
* tree-ssa-reassoc.cc: ...here.
* tree-ssa-sccvn.c: Moved to...
* tree-ssa-sccvn.cc: ...here.
* tree-ssa-scopedtables.c: Moved to...
* tree-ssa-scopedtables.cc: ...here.
* tree-ssa-sink.c: Moved to...
* tree-ssa-sink.cc: ...here.
* tree-ssa-strlen.c: Moved to...
* tree-ssa-strlen.cc: ...here.
* tree-ssa-structalias.c: Moved to...
* tree-ssa-structalias.cc: ...here.
* tree-ssa-tail-merge.c: Moved to...
* tree-ssa-tail-merge.cc: ...here.
* tree-ssa-ter.c: Moved to...
* tree-ssa-ter.cc: ...here.
* tree-ssa-threadbackward.c: Moved to...
* tree-ssa-threadbackward.cc: ...here.
* tree-ssa-threadedge.c: Moved to...
* tree-ssa-threadedge.cc: ...here.
* tree-ssa-threadupdate.c: Moved to...
* tree-ssa-threadupdate.cc: ...here.
* tree-ssa-uncprop.c: Moved to...
* tree-ssa-uncprop.cc: ...here.
* tree-ssa-uninit.c: Moved to...
* tree-ssa-uninit.cc: ...here.
* tree-ssa.c: Moved to...
* tree-ssa.cc: ...here.
* tree-ssanames.c: Moved to...
* tree-ssanames.cc: ...here.
* tree-stdarg.c: Moved to...
* tree-stdarg.cc: ...here.
* tree-streamer-in.c: Moved to...
* tree-streamer-in.cc: ...here.
* tree-streamer-out.c: Moved to...
* tree-streamer-out.cc: ...here.
* tree-streamer.c: Moved to...
* tree-streamer.cc: ...here.
* tree-switch-conversion.c: Moved to...
* tree-switch-conversion.cc: ...here.
* tree-tailcall.c: Moved to...
* tree-tailcall.cc: ...here.
* tree-vect-data-refs.c: Moved to...
* tree-vect-data-refs.cc: ...here.
* tree-vect-generic.c: Moved to...
* tree-vect-generic.cc: ...here.
* tree-vect-loop-manip.c: Moved to...
* tree-vect-loop-manip.cc: ...here.
* tree-vect-loop.c: Moved to...
* tree-vect-loop.cc: ...here.
* tree-vect-patterns.c: Moved to...
* tree-vect-patterns.cc: ...here.
* tree-vect-slp-patterns.c: Moved to...
* tree-vect-slp-patterns.cc: ...here.
* tree-vect-slp.c: Moved to...
* tree-vect-slp.cc: ...here.
* tree-vect-stmts.c: Moved to...
* tree-vect-stmts.cc: ...here.
* tree-vector-builder.c: Moved to...
* tree-vector-builder.cc: ...here.
* tree-vectorizer.c: Moved to...
* tree-vectorizer.cc: ...here.
* tree-vrp.c: Moved to...
* tree-vrp.cc: ...here.
* tree.c: Moved to...
* tree.cc: ...here.
* tsan.c: Moved to...
* tsan.cc: ...here.
* typed-splay-tree.c: Moved to...
* typed-splay-tree.cc: ...here.
* ubsan.c: Moved to...
* ubsan.cc: ...here.
* valtrack.c: Moved to...
* valtrack.cc: ...here.
* value-prof.c: Moved to...
* value-prof.cc: ...here.
* var-tracking.c: Moved to...
* var-tracking.cc: ...here.
* varasm.c: Moved to...
* varasm.cc: ...here.
* varpool.c: Moved to...
* varpool.cc: ...here.
* vec-perm-indices.c: Moved to...
* vec-perm-indices.cc: ...here.
* vec.c: Moved to...
* vec.cc: ...here.
* vmsdbgout.c: Moved to...
* vmsdbgout.cc: ...here.
* vr-values.c: Moved to...
* vr-values.cc: ...here.
* vtable-verify.c: Moved to...
* vtable-verify.cc: ...here.
* web.c: Moved to...
* web.cc: ...here.
* xcoffout.c: Moved to...
* xcoffout.cc: ...here.
gcc/c-family/ChangeLog:
* c-ada-spec.c: Moved to...
* c-ada-spec.cc: ...here.
* c-attribs.c: Moved to...
* c-attribs.cc: ...here.
* c-common.c: Moved to...
* c-common.cc: ...here.
* c-cppbuiltin.c: Moved to...
* c-cppbuiltin.cc: ...here.
* c-dump.c: Moved to...
* c-dump.cc: ...here.
* c-format.c: Moved to...
* c-format.cc: ...here.
* c-gimplify.c: Moved to...
* c-gimplify.cc: ...here.
* c-indentation.c: Moved to...
* c-indentation.cc: ...here.
* c-lex.c: Moved to...
* c-lex.cc: ...here.
* c-omp.c: Moved to...
* c-omp.cc: ...here.
* c-opts.c: Moved to...
* c-opts.cc: ...here.
* c-pch.c: Moved to...
* c-pch.cc: ...here.
* c-ppoutput.c: Moved to...
* c-ppoutput.cc: ...here.
* c-pragma.c: Moved to...
* c-pragma.cc: ...here.
* c-pretty-print.c: Moved to...
* c-pretty-print.cc: ...here.
* c-semantics.c: Moved to...
* c-semantics.cc: ...here.
* c-ubsan.c: Moved to...
* c-ubsan.cc: ...here.
* c-warn.c: Moved to...
* c-warn.cc: ...here.
* cppspec.c: Moved to...
* cppspec.cc: ...here.
* stub-objc.c: Moved to...
* stub-objc.cc: ...here.
gcc/c/ChangeLog:
* c-aux-info.c: Moved to...
* c-aux-info.cc: ...here.
* c-convert.c: Moved to...
* c-convert.cc: ...here.
* c-decl.c: Moved to...
* c-decl.cc: ...here.
* c-errors.c: Moved to...
* c-errors.cc: ...here.
* c-fold.c: Moved to...
* c-fold.cc: ...here.
* c-lang.c: Moved to...
* c-lang.cc: ...here.
* c-objc-common.c: Moved to...
* c-objc-common.cc: ...here.
* c-parser.c: Moved to...
* c-parser.cc: ...here.
* c-typeck.c: Moved to...
* c-typeck.cc: ...here.
* gccspec.c: Moved to...
* gccspec.cc: ...here.
* gimple-parser.c: Moved to...
* gimple-parser.cc: ...here.
gcc/cp/ChangeLog:
* call.c: Moved to...
* call.cc: ...here.
* class.c: Moved to...
* class.cc: ...here.
* constexpr.c: Moved to...
* constexpr.cc: ...here.
* cp-gimplify.c: Moved to...
* cp-gimplify.cc: ...here.
* cp-lang.c: Moved to...
* cp-lang.cc: ...here.
* cp-objcp-common.c: Moved to...
* cp-objcp-common.cc: ...here.
* cp-ubsan.c: Moved to...
* cp-ubsan.cc: ...here.
* cvt.c: Moved to...
* cvt.cc: ...here.
* cxx-pretty-print.c: Moved to...
* cxx-pretty-print.cc: ...here.
* decl.c: Moved to...
* decl.cc: ...here.
* decl2.c: Moved to...
* decl2.cc: ...here.
* dump.c: Moved to...
* dump.cc: ...here.
* error.c: Moved to...
* error.cc: ...here.
* except.c: Moved to...
* except.cc: ...here.
* expr.c: Moved to...
* expr.cc: ...here.
* friend.c: Moved to...
* friend.cc: ...here.
* g++spec.c: Moved to...
* g++spec.cc: ...here.
* init.c: Moved to...
* init.cc: ...here.
* lambda.c: Moved to...
* lambda.cc: ...here.
* lex.c: Moved to...
* lex.cc: ...here.
* mangle.c: Moved to...
* mangle.cc: ...here.
* method.c: Moved to...
* method.cc: ...here.
* name-lookup.c: Moved to...
* name-lookup.cc: ...here.
* optimize.c: Moved to...
* optimize.cc: ...here.
* parser.c: Moved to...
* parser.cc: ...here.
* pt.c: Moved to...
* pt.cc: ...here.
* ptree.c: Moved to...
* ptree.cc: ...here.
* rtti.c: Moved to...
* rtti.cc: ...here.
* search.c: Moved to...
* search.cc: ...here.
* semantics.c: Moved to...
* semantics.cc: ...here.
* tree.c: Moved to...
* tree.cc: ...here.
* typeck.c: Moved to...
* typeck.cc: ...here.
* typeck2.c: Moved to...
* typeck2.cc: ...here.
* vtable-class-hierarchy.c: Moved to...
* vtable-class-hierarchy.cc: ...here.
gcc/fortran/ChangeLog:
* arith.c: Moved to...
* arith.cc: ...here.
* array.c: Moved to...
* array.cc: ...here.
* bbt.c: Moved to...
* bbt.cc: ...here.
* check.c: Moved to...
* check.cc: ...here.
* class.c: Moved to...
* class.cc: ...here.
* constructor.c: Moved to...
* constructor.cc: ...here.
* convert.c: Moved to...
* convert.cc: ...here.
* cpp.c: Moved to...
* cpp.cc: ...here.
* data.c: Moved to...
* data.cc: ...here.
* decl.c: Moved to...
* decl.cc: ...here.
* dependency.c: Moved to...
* dependency.cc: ...here.
* dump-parse-tree.c: Moved to...
* dump-parse-tree.cc: ...here.
* error.c: Moved to...
* error.cc: ...here.
* expr.c: Moved to...
* expr.cc: ...here.
* f95-lang.c: Moved to...
* f95-lang.cc: ...here.
* frontend-passes.c: Moved to...
* frontend-passes.cc: ...here.
* gfortranspec.c: Moved to...
* gfortranspec.cc: ...here.
* interface.c: Moved to...
* interface.cc: ...here.
* intrinsic.c: Moved to...
* intrinsic.cc: ...here.
* io.c: Moved to...
* io.cc: ...here.
* iresolve.c: Moved to...
* iresolve.cc: ...here.
* match.c: Moved to...
* match.cc: ...here.
* matchexp.c: Moved to...
* matchexp.cc: ...here.
* misc.c: Moved to...
* misc.cc: ...here.
* module.c: Moved to...
* module.cc: ...here.
* openmp.c: Moved to...
* openmp.cc: ...here.
* options.c: Moved to...
* options.cc: ...here.
* parse.c: Moved to...
* parse.cc: ...here.
* primary.c: Moved to...
* primary.cc: ...here.
* resolve.c: Moved to...
* resolve.cc: ...here.
* scanner.c: Moved to...
* scanner.cc: ...here.
* simplify.c: Moved to...
* simplify.cc: ...here.
* st.c: Moved to...
* st.cc: ...here.
* symbol.c: Moved to...
* symbol.cc: ...here.
* target-memory.c: Moved to...
* target-memory.cc: ...here.
* trans-array.c: Moved to...
* trans-array.cc: ...here.
* trans-common.c: Moved to...
* trans-common.cc: ...here.
* trans-const.c: Moved to...
* trans-const.cc: ...here.
* trans-decl.c: Moved to...
* trans-decl.cc: ...here.
* trans-expr.c: Moved to...
* trans-expr.cc: ...here.
* trans-intrinsic.c: Moved to...
* trans-intrinsic.cc: ...here.
* trans-io.c: Moved to...
* trans-io.cc: ...here.
* trans-openmp.c: Moved to...
* trans-openmp.cc: ...here.
* trans-stmt.c: Moved to...
* trans-stmt.cc: ...here.
* trans-types.c: Moved to...
* trans-types.cc: ...here.
* trans.c: Moved to...
* trans.cc: ...here.
gcc/go/ChangeLog:
* go-backend.c: Moved to...
* go-backend.cc: ...here.
* go-lang.c: Moved to...
* go-lang.cc: ...here.
* gospec.c: Moved to...
* gospec.cc: ...here.
gcc/jit/ChangeLog:
* dummy-frontend.c: Moved to...
* dummy-frontend.cc: ...here.
* jit-builtins.c: Moved to...
* jit-builtins.cc: ...here.
* jit-logging.c: Moved to...
* jit-logging.cc: ...here.
* jit-playback.c: Moved to...
* jit-playback.cc: ...here.
* jit-recording.c: Moved to...
* jit-recording.cc: ...here.
* jit-result.c: Moved to...
* jit-result.cc: ...here.
* jit-spec.c: Moved to...
* jit-spec.cc: ...here.
* jit-tempdir.c: Moved to...
* jit-tempdir.cc: ...here.
* jit-w32.c: Moved to...
* jit-w32.cc: ...here.
* libgccjit.c: Moved to...
* libgccjit.cc: ...here.
gcc/lto/ChangeLog:
* common.c: Moved to...
* common.cc: ...here.
* lto-common.c: Moved to...
* lto-common.cc: ...here.
* lto-dump.c: Moved to...
* lto-dump.cc: ...here.
* lto-lang.c: Moved to...
* lto-lang.cc: ...here.
* lto-object.c: Moved to...
* lto-object.cc: ...here.
* lto-partition.c: Moved to...
* lto-partition.cc: ...here.
* lto-symtab.c: Moved to...
* lto-symtab.cc: ...here.
* lto.c: Moved to...
* lto.cc: ...here.
gcc/objc/ChangeLog:
* objc-act.c: Moved to...
* objc-act.cc: ...here.
* objc-encoding.c: Moved to...
* objc-encoding.cc: ...here.
* objc-gnu-runtime-abi-01.c: Moved to...
* objc-gnu-runtime-abi-01.cc: ...here.
* objc-lang.c: Moved to...
* objc-lang.cc: ...here.
* objc-map.c: Moved to...
* objc-map.cc: ...here.
* objc-next-runtime-abi-01.c: Moved to...
* objc-next-runtime-abi-01.cc: ...here.
* objc-next-runtime-abi-02.c: Moved to...
* objc-next-runtime-abi-02.cc: ...here.
* objc-runtime-shared-support.c: Moved to...
* objc-runtime-shared-support.cc: ...here.
gcc/objcp/ChangeLog:
* objcp-decl.c: Moved to...
* objcp-decl.cc: ...here.
* objcp-lang.c: Moved to...
* objcp-lang.cc: ...here.
libcpp/ChangeLog:
* charset.c: Moved to...
* charset.cc: ...here.
* directives.c: Moved to...
* directives.cc: ...here.
* errors.c: Moved to...
* errors.cc: ...here.
* expr.c: Moved to...
* expr.cc: ...here.
* files.c: Moved to...
* files.cc: ...here.
* identifiers.c: Moved to...
* identifiers.cc: ...here.
* init.c: Moved to...
* init.cc: ...here.
* lex.c: Moved to...
* lex.cc: ...here.
* line-map.c: Moved to...
* line-map.cc: ...here.
* macro.c: Moved to...
* macro.cc: ...here.
* makeucnid.c: Moved to...
* makeucnid.cc: ...here.
* mkdeps.c: Moved to...
* mkdeps.cc: ...here.
* pch.c: Moved to...
* pch.cc: ...here.
* symtab.c: Moved to...
* symtab.cc: ...here.
* traditional.c: Moved to...
* traditional.cc: ...here.
Diffstat (limited to 'gcc/optabs.c')
-rw-r--r-- | gcc/optabs.c | 7966 |
1 files changed, 0 insertions, 7966 deletions
diff --git a/gcc/optabs.c b/gcc/optabs.c deleted file mode 100644 index 5f759d5..0000000 --- a/gcc/optabs.c +++ /dev/null @@ -1,7966 +0,0 @@ -/* Expand the basic unary and binary arithmetic operations, for GNU compiler. - Copyright (C) 1987-2022 Free Software Foundation, Inc. - -This file is part of GCC. - -GCC is free software; you can redistribute it and/or modify it under -the terms of the GNU General Public License as published by the Free -Software Foundation; either version 3, or (at your option) any later -version. - -GCC is distributed in the hope that it will be useful, but WITHOUT ANY -WARRANTY; without even the implied warranty of MERCHANTABILITY or -FITNESS FOR A PARTICULAR PURPOSE. See the GNU General Public License -for more details. - -You should have received a copy of the GNU General Public License -along with GCC; see the file COPYING3. If not see -<http://www.gnu.org/licenses/>. */ - - -#include "config.h" -#include "system.h" -#include "coretypes.h" -#include "backend.h" -#include "target.h" -#include "rtl.h" -#include "tree.h" -#include "memmodel.h" -#include "predict.h" -#include "tm_p.h" -#include "optabs.h" -#include "expmed.h" -#include "emit-rtl.h" -#include "recog.h" -#include "diagnostic-core.h" -#include "rtx-vector-builder.h" - -/* Include insn-config.h before expr.h so that HAVE_conditional_move - is properly defined. */ -#include "stor-layout.h" -#include "except.h" -#include "dojump.h" -#include "explow.h" -#include "expr.h" -#include "optabs-tree.h" -#include "libfuncs.h" -#include "internal-fn.h" -#include "langhooks.h" - -static void prepare_float_lib_cmp (rtx, rtx, enum rtx_code, rtx *, - machine_mode *); -static rtx expand_unop_direct (machine_mode, optab, rtx, rtx, int); -static void emit_libcall_block_1 (rtx_insn *, rtx, rtx, rtx, bool); - -/* Debug facility for use in GDB. */ -void debug_optab_libfuncs (void); - -/* Add a REG_EQUAL note to the last insn in INSNS. TARGET is being set to - the result of operation CODE applied to OP0 (and OP1 if it is a binary - operation). OP0_MODE is OP0's mode. - - If the last insn does not set TARGET, don't do anything, but return 1. - - If the last insn or a previous insn sets TARGET and TARGET is one of OP0 - or OP1, don't add the REG_EQUAL note but return 0. Our caller can then - try again, ensuring that TARGET is not one of the operands. */ - -static int -add_equal_note (rtx_insn *insns, rtx target, enum rtx_code code, rtx op0, - rtx op1, machine_mode op0_mode) -{ - rtx_insn *last_insn; - rtx set; - rtx note; - - gcc_assert (insns && INSN_P (insns) && NEXT_INSN (insns)); - - if (GET_RTX_CLASS (code) != RTX_COMM_ARITH - && GET_RTX_CLASS (code) != RTX_BIN_ARITH - && GET_RTX_CLASS (code) != RTX_COMM_COMPARE - && GET_RTX_CLASS (code) != RTX_COMPARE - && GET_RTX_CLASS (code) != RTX_UNARY) - return 1; - - if (GET_CODE (target) == ZERO_EXTRACT) - return 1; - - for (last_insn = insns; - NEXT_INSN (last_insn) != NULL_RTX; - last_insn = NEXT_INSN (last_insn)) - ; - - /* If TARGET is in OP0 or OP1, punt. We'd end up with a note referencing - a value changing in the insn, so the note would be invalid for CSE. */ - if (reg_overlap_mentioned_p (target, op0) - || (op1 && reg_overlap_mentioned_p (target, op1))) - { - if (MEM_P (target) - && (rtx_equal_p (target, op0) - || (op1 && rtx_equal_p (target, op1)))) - { - /* For MEM target, with MEM = MEM op X, prefer no REG_EQUAL note - over expanding it as temp = MEM op X, MEM = temp. If the target - supports MEM = MEM op X instructions, it is sometimes too hard - to reconstruct that form later, especially if X is also a memory, - and due to multiple occurrences of addresses the address might - be forced into register unnecessarily. - Note that not emitting the REG_EQUIV note might inhibit - CSE in some cases. */ - set = single_set (last_insn); - if (set - && GET_CODE (SET_SRC (set)) == code - && MEM_P (SET_DEST (set)) - && (rtx_equal_p (SET_DEST (set), XEXP (SET_SRC (set), 0)) - || (op1 && rtx_equal_p (SET_DEST (set), - XEXP (SET_SRC (set), 1))))) - return 1; - } - return 0; - } - - set = set_for_reg_notes (last_insn); - if (set == NULL_RTX) - return 1; - - if (! rtx_equal_p (SET_DEST (set), target) - /* For a STRICT_LOW_PART, the REG_NOTE applies to what is inside it. */ - && (GET_CODE (SET_DEST (set)) != STRICT_LOW_PART - || ! rtx_equal_p (XEXP (SET_DEST (set), 0), target))) - return 1; - - if (GET_RTX_CLASS (code) == RTX_UNARY) - switch (code) - { - case FFS: - case CLZ: - case CTZ: - case CLRSB: - case POPCOUNT: - case PARITY: - case BSWAP: - if (op0_mode != VOIDmode && GET_MODE (target) != op0_mode) - { - note = gen_rtx_fmt_e (code, op0_mode, copy_rtx (op0)); - if (GET_MODE_UNIT_SIZE (op0_mode) - > GET_MODE_UNIT_SIZE (GET_MODE (target))) - note = simplify_gen_unary (TRUNCATE, GET_MODE (target), - note, op0_mode); - else - note = simplify_gen_unary (ZERO_EXTEND, GET_MODE (target), - note, op0_mode); - break; - } - /* FALLTHRU */ - default: - note = gen_rtx_fmt_e (code, GET_MODE (target), copy_rtx (op0)); - break; - } - else - note = gen_rtx_fmt_ee (code, GET_MODE (target), copy_rtx (op0), copy_rtx (op1)); - - set_unique_reg_note (last_insn, REG_EQUAL, note); - - return 1; -} - -/* Given two input operands, OP0 and OP1, determine what the correct from_mode - for a widening operation would be. In most cases this would be OP0, but if - that's a constant it'll be VOIDmode, which isn't useful. */ - -static machine_mode -widened_mode (machine_mode to_mode, rtx op0, rtx op1) -{ - machine_mode m0 = GET_MODE (op0); - machine_mode m1 = GET_MODE (op1); - machine_mode result; - - if (m0 == VOIDmode && m1 == VOIDmode) - return to_mode; - else if (m0 == VOIDmode || GET_MODE_UNIT_SIZE (m0) < GET_MODE_UNIT_SIZE (m1)) - result = m1; - else - result = m0; - - if (GET_MODE_UNIT_SIZE (result) > GET_MODE_UNIT_SIZE (to_mode)) - return to_mode; - - return result; -} - -/* Widen OP to MODE and return the rtx for the widened operand. UNSIGNEDP - says whether OP is signed or unsigned. NO_EXTEND is nonzero if we need - not actually do a sign-extend or zero-extend, but can leave the - higher-order bits of the result rtx undefined, for example, in the case - of logical operations, but not right shifts. */ - -static rtx -widen_operand (rtx op, machine_mode mode, machine_mode oldmode, - int unsignedp, int no_extend) -{ - rtx result; - scalar_int_mode int_mode; - - /* If we don't have to extend and this is a constant, return it. */ - if (no_extend && GET_MODE (op) == VOIDmode) - return op; - - /* If we must extend do so. If OP is a SUBREG for a promoted object, also - extend since it will be more efficient to do so unless the signedness of - a promoted object differs from our extension. */ - if (! no_extend - || !is_a <scalar_int_mode> (mode, &int_mode) - || (GET_CODE (op) == SUBREG && SUBREG_PROMOTED_VAR_P (op) - && SUBREG_CHECK_PROMOTED_SIGN (op, unsignedp))) - return convert_modes (mode, oldmode, op, unsignedp); - - /* If MODE is no wider than a single word, we return a lowpart or paradoxical - SUBREG. */ - if (GET_MODE_SIZE (int_mode) <= UNITS_PER_WORD) - return gen_lowpart (int_mode, force_reg (GET_MODE (op), op)); - - /* Otherwise, get an object of MODE, clobber it, and set the low-order - part to OP. */ - - result = gen_reg_rtx (int_mode); - emit_clobber (result); - emit_move_insn (gen_lowpart (GET_MODE (op), result), op); - return result; -} - -/* Expand vector widening operations. - - There are two different classes of operations handled here: - 1) Operations whose result is wider than all the arguments to the operation. - Examples: VEC_UNPACK_HI/LO_EXPR, VEC_WIDEN_MULT_HI/LO_EXPR - In this case OP0 and optionally OP1 would be initialized, - but WIDE_OP wouldn't (not relevant for this case). - 2) Operations whose result is of the same size as the last argument to the - operation, but wider than all the other arguments to the operation. - Examples: WIDEN_SUM_EXPR, VEC_DOT_PROD_EXPR. - In the case WIDE_OP, OP0 and optionally OP1 would be initialized. - - E.g, when called to expand the following operations, this is how - the arguments will be initialized: - nops OP0 OP1 WIDE_OP - widening-sum 2 oprnd0 - oprnd1 - widening-dot-product 3 oprnd0 oprnd1 oprnd2 - widening-mult 2 oprnd0 oprnd1 - - type-promotion (vec-unpack) 1 oprnd0 - - */ - -rtx -expand_widen_pattern_expr (sepops ops, rtx op0, rtx op1, rtx wide_op, - rtx target, int unsignedp) -{ - class expand_operand eops[4]; - tree oprnd0, oprnd1, oprnd2; - machine_mode wmode = VOIDmode, tmode0, tmode1 = VOIDmode; - optab widen_pattern_optab; - enum insn_code icode; - int nops = TREE_CODE_LENGTH (ops->code); - int op; - bool sbool = false; - - oprnd0 = ops->op0; - if (nops >= 2) - oprnd1 = ops->op1; - if (nops >= 3) - oprnd2 = ops->op2; - - tmode0 = TYPE_MODE (TREE_TYPE (oprnd0)); - if (ops->code == VEC_UNPACK_FIX_TRUNC_HI_EXPR - || ops->code == VEC_UNPACK_FIX_TRUNC_LO_EXPR) - /* The sign is from the result type rather than operand's type - for these ops. */ - widen_pattern_optab - = optab_for_tree_code (ops->code, ops->type, optab_default); - else if ((ops->code == VEC_UNPACK_HI_EXPR - || ops->code == VEC_UNPACK_LO_EXPR) - && VECTOR_BOOLEAN_TYPE_P (ops->type) - && VECTOR_BOOLEAN_TYPE_P (TREE_TYPE (oprnd0)) - && TYPE_MODE (ops->type) == TYPE_MODE (TREE_TYPE (oprnd0)) - && SCALAR_INT_MODE_P (TYPE_MODE (ops->type))) - { - /* For VEC_UNPACK_{LO,HI}_EXPR if the mode of op0 and result is - the same scalar mode for VECTOR_BOOLEAN_TYPE_P vectors, use - vec_unpacks_sbool_{lo,hi}_optab, so that we can pass in - the pattern number of elements in the wider vector. */ - widen_pattern_optab - = (ops->code == VEC_UNPACK_HI_EXPR - ? vec_unpacks_sbool_hi_optab : vec_unpacks_sbool_lo_optab); - sbool = true; - } - else if (ops->code == DOT_PROD_EXPR) - { - enum optab_subtype subtype = optab_default; - signop sign1 = TYPE_SIGN (TREE_TYPE (oprnd0)); - signop sign2 = TYPE_SIGN (TREE_TYPE (oprnd1)); - if (sign1 == sign2) - ; - else if (sign1 == SIGNED && sign2 == UNSIGNED) - { - subtype = optab_vector_mixed_sign; - /* Same as optab_vector_mixed_sign but flip the operands. */ - std::swap (op0, op1); - } - else if (sign1 == UNSIGNED && sign2 == SIGNED) - subtype = optab_vector_mixed_sign; - else - gcc_unreachable (); - - widen_pattern_optab - = optab_for_tree_code (ops->code, TREE_TYPE (oprnd0), subtype); - } - else - widen_pattern_optab - = optab_for_tree_code (ops->code, TREE_TYPE (oprnd0), optab_default); - if (ops->code == WIDEN_MULT_PLUS_EXPR - || ops->code == WIDEN_MULT_MINUS_EXPR) - icode = find_widening_optab_handler (widen_pattern_optab, - TYPE_MODE (TREE_TYPE (ops->op2)), - tmode0); - else - icode = optab_handler (widen_pattern_optab, tmode0); - gcc_assert (icode != CODE_FOR_nothing); - - if (nops >= 2) - tmode1 = TYPE_MODE (TREE_TYPE (oprnd1)); - else if (sbool) - { - nops = 2; - op1 = GEN_INT (TYPE_VECTOR_SUBPARTS (TREE_TYPE (oprnd0)).to_constant ()); - tmode1 = tmode0; - } - - /* The last operand is of a wider mode than the rest of the operands. */ - if (nops == 2) - wmode = tmode1; - else if (nops == 3) - { - gcc_assert (tmode1 == tmode0); - gcc_assert (op1); - wmode = TYPE_MODE (TREE_TYPE (oprnd2)); - } - - op = 0; - create_output_operand (&eops[op++], target, TYPE_MODE (ops->type)); - create_convert_operand_from (&eops[op++], op0, tmode0, unsignedp); - if (op1) - create_convert_operand_from (&eops[op++], op1, tmode1, unsignedp); - if (wide_op) - create_convert_operand_from (&eops[op++], wide_op, wmode, unsignedp); - expand_insn (icode, op, eops); - return eops[0].value; -} - -/* Generate code to perform an operation specified by TERNARY_OPTAB - on operands OP0, OP1 and OP2, with result having machine-mode MODE. - - UNSIGNEDP is for the case where we have to widen the operands - to perform the operation. It says to use zero-extension. - - If TARGET is nonzero, the value - is generated there, if it is convenient to do so. - In all cases an rtx is returned for the locus of the value; - this may or may not be TARGET. */ - -rtx -expand_ternary_op (machine_mode mode, optab ternary_optab, rtx op0, - rtx op1, rtx op2, rtx target, int unsignedp) -{ - class expand_operand ops[4]; - enum insn_code icode = optab_handler (ternary_optab, mode); - - gcc_assert (optab_handler (ternary_optab, mode) != CODE_FOR_nothing); - - create_output_operand (&ops[0], target, mode); - create_convert_operand_from (&ops[1], op0, mode, unsignedp); - create_convert_operand_from (&ops[2], op1, mode, unsignedp); - create_convert_operand_from (&ops[3], op2, mode, unsignedp); - expand_insn (icode, 4, ops); - return ops[0].value; -} - - -/* Like expand_binop, but return a constant rtx if the result can be - calculated at compile time. The arguments and return value are - otherwise the same as for expand_binop. */ - -rtx -simplify_expand_binop (machine_mode mode, optab binoptab, - rtx op0, rtx op1, rtx target, int unsignedp, - enum optab_methods methods) -{ - if (CONSTANT_P (op0) && CONSTANT_P (op1)) - { - rtx x = simplify_binary_operation (optab_to_code (binoptab), - mode, op0, op1); - if (x) - return x; - } - - return expand_binop (mode, binoptab, op0, op1, target, unsignedp, methods); -} - -/* Like simplify_expand_binop, but always put the result in TARGET. - Return true if the expansion succeeded. */ - -bool -force_expand_binop (machine_mode mode, optab binoptab, - rtx op0, rtx op1, rtx target, int unsignedp, - enum optab_methods methods) -{ - rtx x = simplify_expand_binop (mode, binoptab, op0, op1, - target, unsignedp, methods); - if (x == 0) - return false; - if (x != target) - emit_move_insn (target, x); - return true; -} - -/* Create a new vector value in VMODE with all elements set to OP. The - mode of OP must be the element mode of VMODE. If OP is a constant, - then the return value will be a constant. */ - -rtx -expand_vector_broadcast (machine_mode vmode, rtx op) -{ - int n; - rtvec vec; - - gcc_checking_assert (VECTOR_MODE_P (vmode)); - - if (valid_for_const_vector_p (vmode, op)) - return gen_const_vec_duplicate (vmode, op); - - insn_code icode = optab_handler (vec_duplicate_optab, vmode); - if (icode != CODE_FOR_nothing) - { - class expand_operand ops[2]; - create_output_operand (&ops[0], NULL_RTX, vmode); - create_input_operand (&ops[1], op, GET_MODE (op)); - expand_insn (icode, 2, ops); - return ops[0].value; - } - - if (!GET_MODE_NUNITS (vmode).is_constant (&n)) - return NULL; - - /* ??? If the target doesn't have a vec_init, then we have no easy way - of performing this operation. Most of this sort of generic support - is hidden away in the vector lowering support in gimple. */ - icode = convert_optab_handler (vec_init_optab, vmode, - GET_MODE_INNER (vmode)); - if (icode == CODE_FOR_nothing) - return NULL; - - vec = rtvec_alloc (n); - for (int i = 0; i < n; ++i) - RTVEC_ELT (vec, i) = op; - rtx ret = gen_reg_rtx (vmode); - emit_insn (GEN_FCN (icode) (ret, gen_rtx_PARALLEL (vmode, vec))); - - return ret; -} - -/* This subroutine of expand_doubleword_shift handles the cases in which - the effective shift value is >= BITS_PER_WORD. The arguments and return - value are the same as for the parent routine, except that SUPERWORD_OP1 - is the shift count to use when shifting OUTOF_INPUT into INTO_TARGET. - INTO_TARGET may be null if the caller has decided to calculate it. */ - -static bool -expand_superword_shift (optab binoptab, rtx outof_input, rtx superword_op1, - rtx outof_target, rtx into_target, - int unsignedp, enum optab_methods methods) -{ - if (into_target != 0) - if (!force_expand_binop (word_mode, binoptab, outof_input, superword_op1, - into_target, unsignedp, methods)) - return false; - - if (outof_target != 0) - { - /* For a signed right shift, we must fill OUTOF_TARGET with copies - of the sign bit, otherwise we must fill it with zeros. */ - if (binoptab != ashr_optab) - emit_move_insn (outof_target, CONST0_RTX (word_mode)); - else - if (!force_expand_binop (word_mode, binoptab, outof_input, - gen_int_shift_amount (word_mode, - BITS_PER_WORD - 1), - outof_target, unsignedp, methods)) - return false; - } - return true; -} - -/* This subroutine of expand_doubleword_shift handles the cases in which - the effective shift value is < BITS_PER_WORD. The arguments and return - value are the same as for the parent routine. */ - -static bool -expand_subword_shift (scalar_int_mode op1_mode, optab binoptab, - rtx outof_input, rtx into_input, rtx op1, - rtx outof_target, rtx into_target, - int unsignedp, enum optab_methods methods, - unsigned HOST_WIDE_INT shift_mask) -{ - optab reverse_unsigned_shift, unsigned_shift; - rtx tmp, carries; - - reverse_unsigned_shift = (binoptab == ashl_optab ? lshr_optab : ashl_optab); - unsigned_shift = (binoptab == ashl_optab ? ashl_optab : lshr_optab); - - /* The low OP1 bits of INTO_TARGET come from the high bits of OUTOF_INPUT. - We therefore need to shift OUTOF_INPUT by (BITS_PER_WORD - OP1) bits in - the opposite direction to BINOPTAB. */ - if (CONSTANT_P (op1) || shift_mask >= BITS_PER_WORD) - { - carries = outof_input; - tmp = immed_wide_int_const (wi::shwi (BITS_PER_WORD, - op1_mode), op1_mode); - tmp = simplify_expand_binop (op1_mode, sub_optab, tmp, op1, - 0, true, methods); - } - else - { - /* We must avoid shifting by BITS_PER_WORD bits since that is either - the same as a zero shift (if shift_mask == BITS_PER_WORD - 1) or - has unknown behavior. Do a single shift first, then shift by the - remainder. It's OK to use ~OP1 as the remainder if shift counts - are truncated to the mode size. */ - carries = expand_binop (word_mode, reverse_unsigned_shift, - outof_input, const1_rtx, 0, unsignedp, methods); - if (shift_mask == BITS_PER_WORD - 1) - { - tmp = immed_wide_int_const - (wi::minus_one (GET_MODE_PRECISION (op1_mode)), op1_mode); - tmp = simplify_expand_binop (op1_mode, xor_optab, op1, tmp, - 0, true, methods); - } - else - { - tmp = immed_wide_int_const (wi::shwi (BITS_PER_WORD - 1, - op1_mode), op1_mode); - tmp = simplify_expand_binop (op1_mode, sub_optab, tmp, op1, - 0, true, methods); - } - } - if (tmp == 0 || carries == 0) - return false; - carries = expand_binop (word_mode, reverse_unsigned_shift, - carries, tmp, 0, unsignedp, methods); - if (carries == 0) - return false; - - /* Shift INTO_INPUT logically by OP1. This is the last use of INTO_INPUT - so the result can go directly into INTO_TARGET if convenient. */ - tmp = expand_binop (word_mode, unsigned_shift, into_input, op1, - into_target, unsignedp, methods); - if (tmp == 0) - return false; - - /* Now OR in the bits carried over from OUTOF_INPUT. */ - if (!force_expand_binop (word_mode, ior_optab, tmp, carries, - into_target, unsignedp, methods)) - return false; - - /* Use a standard word_mode shift for the out-of half. */ - if (outof_target != 0) - if (!force_expand_binop (word_mode, binoptab, outof_input, op1, - outof_target, unsignedp, methods)) - return false; - - return true; -} - - -/* Try implementing expand_doubleword_shift using conditional moves. - The shift is by < BITS_PER_WORD if (CMP_CODE CMP1 CMP2) is true, - otherwise it is by >= BITS_PER_WORD. SUBWORD_OP1 and SUPERWORD_OP1 - are the shift counts to use in the former and latter case. All other - arguments are the same as the parent routine. */ - -static bool -expand_doubleword_shift_condmove (scalar_int_mode op1_mode, optab binoptab, - enum rtx_code cmp_code, rtx cmp1, rtx cmp2, - rtx outof_input, rtx into_input, - rtx subword_op1, rtx superword_op1, - rtx outof_target, rtx into_target, - int unsignedp, enum optab_methods methods, - unsigned HOST_WIDE_INT shift_mask) -{ - rtx outof_superword, into_superword; - - /* Put the superword version of the output into OUTOF_SUPERWORD and - INTO_SUPERWORD. */ - outof_superword = outof_target != 0 ? gen_reg_rtx (word_mode) : 0; - if (outof_target != 0 && subword_op1 == superword_op1) - { - /* The value INTO_TARGET >> SUBWORD_OP1, which we later store in - OUTOF_TARGET, is the same as the value of INTO_SUPERWORD. */ - into_superword = outof_target; - if (!expand_superword_shift (binoptab, outof_input, superword_op1, - outof_superword, 0, unsignedp, methods)) - return false; - } - else - { - into_superword = gen_reg_rtx (word_mode); - if (!expand_superword_shift (binoptab, outof_input, superword_op1, - outof_superword, into_superword, - unsignedp, methods)) - return false; - } - - /* Put the subword version directly in OUTOF_TARGET and INTO_TARGET. */ - if (!expand_subword_shift (op1_mode, binoptab, - outof_input, into_input, subword_op1, - outof_target, into_target, - unsignedp, methods, shift_mask)) - return false; - - /* Select between them. Do the INTO half first because INTO_SUPERWORD - might be the current value of OUTOF_TARGET. */ - if (!emit_conditional_move (into_target, cmp_code, cmp1, cmp2, op1_mode, - into_target, into_superword, word_mode, false)) - return false; - - if (outof_target != 0) - if (!emit_conditional_move (outof_target, cmp_code, cmp1, cmp2, op1_mode, - outof_target, outof_superword, - word_mode, false)) - return false; - - return true; -} - -/* Expand a doubleword shift (ashl, ashr or lshr) using word-mode shifts. - OUTOF_INPUT and INTO_INPUT are the two word-sized halves of the first - input operand; the shift moves bits in the direction OUTOF_INPUT-> - INTO_TARGET. OUTOF_TARGET and INTO_TARGET are the equivalent words - of the target. OP1 is the shift count and OP1_MODE is its mode. - If OP1 is constant, it will have been truncated as appropriate - and is known to be nonzero. - - If SHIFT_MASK is zero, the result of word shifts is undefined when the - shift count is outside the range [0, BITS_PER_WORD). This routine must - avoid generating such shifts for OP1s in the range [0, BITS_PER_WORD * 2). - - If SHIFT_MASK is nonzero, all word-mode shift counts are effectively - masked by it and shifts in the range [BITS_PER_WORD, SHIFT_MASK) will - fill with zeros or sign bits as appropriate. - - If SHIFT_MASK is BITS_PER_WORD - 1, this routine will synthesize - a doubleword shift whose equivalent mask is BITS_PER_WORD * 2 - 1. - Doing this preserves semantics required by SHIFT_COUNT_TRUNCATED. - In all other cases, shifts by values outside [0, BITS_PER_UNIT * 2) - are undefined. - - BINOPTAB, UNSIGNEDP and METHODS are as for expand_binop. This function - may not use INTO_INPUT after modifying INTO_TARGET, and similarly for - OUTOF_INPUT and OUTOF_TARGET. OUTOF_TARGET can be null if the parent - function wants to calculate it itself. - - Return true if the shift could be successfully synthesized. */ - -static bool -expand_doubleword_shift (scalar_int_mode op1_mode, optab binoptab, - rtx outof_input, rtx into_input, rtx op1, - rtx outof_target, rtx into_target, - int unsignedp, enum optab_methods methods, - unsigned HOST_WIDE_INT shift_mask) -{ - rtx superword_op1, tmp, cmp1, cmp2; - enum rtx_code cmp_code; - - /* See if word-mode shifts by BITS_PER_WORD...BITS_PER_WORD * 2 - 1 will - fill the result with sign or zero bits as appropriate. If so, the value - of OUTOF_TARGET will always be (SHIFT OUTOF_INPUT OP1). Recursively call - this routine to calculate INTO_TARGET (which depends on both OUTOF_INPUT - and INTO_INPUT), then emit code to set up OUTOF_TARGET. - - This isn't worthwhile for constant shifts since the optimizers will - cope better with in-range shift counts. */ - if (shift_mask >= BITS_PER_WORD - && outof_target != 0 - && !CONSTANT_P (op1)) - { - if (!expand_doubleword_shift (op1_mode, binoptab, - outof_input, into_input, op1, - 0, into_target, - unsignedp, methods, shift_mask)) - return false; - if (!force_expand_binop (word_mode, binoptab, outof_input, op1, - outof_target, unsignedp, methods)) - return false; - return true; - } - - /* Set CMP_CODE, CMP1 and CMP2 so that the rtx (CMP_CODE CMP1 CMP2) - is true when the effective shift value is less than BITS_PER_WORD. - Set SUPERWORD_OP1 to the shift count that should be used to shift - OUTOF_INPUT into INTO_TARGET when the condition is false. */ - tmp = immed_wide_int_const (wi::shwi (BITS_PER_WORD, op1_mode), op1_mode); - if (!CONSTANT_P (op1) && shift_mask == BITS_PER_WORD - 1) - { - /* Set CMP1 to OP1 & BITS_PER_WORD. The result is zero iff OP1 - is a subword shift count. */ - cmp1 = simplify_expand_binop (op1_mode, and_optab, op1, tmp, - 0, true, methods); - cmp2 = CONST0_RTX (op1_mode); - cmp_code = EQ; - superword_op1 = op1; - } - else - { - /* Set CMP1 to OP1 - BITS_PER_WORD. */ - cmp1 = simplify_expand_binop (op1_mode, sub_optab, op1, tmp, - 0, true, methods); - cmp2 = CONST0_RTX (op1_mode); - cmp_code = LT; - superword_op1 = cmp1; - } - if (cmp1 == 0) - return false; - - /* If we can compute the condition at compile time, pick the - appropriate subroutine. */ - tmp = simplify_relational_operation (cmp_code, SImode, op1_mode, cmp1, cmp2); - if (tmp != 0 && CONST_INT_P (tmp)) - { - if (tmp == const0_rtx) - return expand_superword_shift (binoptab, outof_input, superword_op1, - outof_target, into_target, - unsignedp, methods); - else - return expand_subword_shift (op1_mode, binoptab, - outof_input, into_input, op1, - outof_target, into_target, - unsignedp, methods, shift_mask); - } - - /* Try using conditional moves to generate straight-line code. */ - if (HAVE_conditional_move) - { - rtx_insn *start = get_last_insn (); - if (expand_doubleword_shift_condmove (op1_mode, binoptab, - cmp_code, cmp1, cmp2, - outof_input, into_input, - op1, superword_op1, - outof_target, into_target, - unsignedp, methods, shift_mask)) - return true; - delete_insns_since (start); - } - - /* As a last resort, use branches to select the correct alternative. */ - rtx_code_label *subword_label = gen_label_rtx (); - rtx_code_label *done_label = gen_label_rtx (); - - NO_DEFER_POP; - do_compare_rtx_and_jump (cmp1, cmp2, cmp_code, false, op1_mode, - 0, 0, subword_label, - profile_probability::uninitialized ()); - OK_DEFER_POP; - - if (!expand_superword_shift (binoptab, outof_input, superword_op1, - outof_target, into_target, - unsignedp, methods)) - return false; - - emit_jump_insn (targetm.gen_jump (done_label)); - emit_barrier (); - emit_label (subword_label); - - if (!expand_subword_shift (op1_mode, binoptab, - outof_input, into_input, op1, - outof_target, into_target, - unsignedp, methods, shift_mask)) - return false; - - emit_label (done_label); - return true; -} - -/* Subroutine of expand_binop. Perform a double word multiplication of - operands OP0 and OP1 both of mode MODE, which is exactly twice as wide - as the target's word_mode. This function return NULL_RTX if anything - goes wrong, in which case it may have already emitted instructions - which need to be deleted. - - If we want to multiply two two-word values and have normal and widening - multiplies of single-word values, we can do this with three smaller - multiplications. - - The multiplication proceeds as follows: - _______________________ - [__op0_high_|__op0_low__] - _______________________ - * [__op1_high_|__op1_low__] - _______________________________________________ - _______________________ - (1) [__op0_low__*__op1_low__] - _______________________ - (2a) [__op0_low__*__op1_high_] - _______________________ - (2b) [__op0_high_*__op1_low__] - _______________________ - (3) [__op0_high_*__op1_high_] - - - This gives a 4-word result. Since we are only interested in the - lower 2 words, partial result (3) and the upper words of (2a) and - (2b) don't need to be calculated. Hence (2a) and (2b) can be - calculated using non-widening multiplication. - - (1), however, needs to be calculated with an unsigned widening - multiplication. If this operation is not directly supported we - try using a signed widening multiplication and adjust the result. - This adjustment works as follows: - - If both operands are positive then no adjustment is needed. - - If the operands have different signs, for example op0_low < 0 and - op1_low >= 0, the instruction treats the most significant bit of - op0_low as a sign bit instead of a bit with significance - 2**(BITS_PER_WORD-1), i.e. the instruction multiplies op1_low - with 2**BITS_PER_WORD - op0_low, and two's complements the - result. Conclusion: We need to add op1_low * 2**BITS_PER_WORD to - the result. - - Similarly, if both operands are negative, we need to add - (op0_low + op1_low) * 2**BITS_PER_WORD. - - We use a trick to adjust quickly. We logically shift op0_low right - (op1_low) BITS_PER_WORD-1 steps to get 0 or 1, and add this to - op0_high (op1_high) before it is used to calculate 2b (2a). If no - logical shift exists, we do an arithmetic right shift and subtract - the 0 or -1. */ - -static rtx -expand_doubleword_mult (machine_mode mode, rtx op0, rtx op1, rtx target, - bool umulp, enum optab_methods methods) -{ - int low = (WORDS_BIG_ENDIAN ? 1 : 0); - int high = (WORDS_BIG_ENDIAN ? 0 : 1); - rtx wordm1 = (umulp ? NULL_RTX - : gen_int_shift_amount (word_mode, BITS_PER_WORD - 1)); - rtx product, adjust, product_high, temp; - - rtx op0_high = operand_subword_force (op0, high, mode); - rtx op0_low = operand_subword_force (op0, low, mode); - rtx op1_high = operand_subword_force (op1, high, mode); - rtx op1_low = operand_subword_force (op1, low, mode); - - /* If we're using an unsigned multiply to directly compute the product - of the low-order words of the operands and perform any required - adjustments of the operands, we begin by trying two more multiplications - and then computing the appropriate sum. - - We have checked above that the required addition is provided. - Full-word addition will normally always succeed, especially if - it is provided at all, so we don't worry about its failure. The - multiplication may well fail, however, so we do handle that. */ - - if (!umulp) - { - /* ??? This could be done with emit_store_flag where available. */ - temp = expand_binop (word_mode, lshr_optab, op0_low, wordm1, - NULL_RTX, 1, methods); - if (temp) - op0_high = expand_binop (word_mode, add_optab, op0_high, temp, - NULL_RTX, 0, OPTAB_DIRECT); - else - { - temp = expand_binop (word_mode, ashr_optab, op0_low, wordm1, - NULL_RTX, 0, methods); - if (!temp) - return NULL_RTX; - op0_high = expand_binop (word_mode, sub_optab, op0_high, temp, - NULL_RTX, 0, OPTAB_DIRECT); - } - - if (!op0_high) - return NULL_RTX; - } - - adjust = expand_binop (word_mode, smul_optab, op0_high, op1_low, - NULL_RTX, 0, OPTAB_DIRECT); - if (!adjust) - return NULL_RTX; - - /* OP0_HIGH should now be dead. */ - - if (!umulp) - { - /* ??? This could be done with emit_store_flag where available. */ - temp = expand_binop (word_mode, lshr_optab, op1_low, wordm1, - NULL_RTX, 1, methods); - if (temp) - op1_high = expand_binop (word_mode, add_optab, op1_high, temp, - NULL_RTX, 0, OPTAB_DIRECT); - else - { - temp = expand_binop (word_mode, ashr_optab, op1_low, wordm1, - NULL_RTX, 0, methods); - if (!temp) - return NULL_RTX; - op1_high = expand_binop (word_mode, sub_optab, op1_high, temp, - NULL_RTX, 0, OPTAB_DIRECT); - } - - if (!op1_high) - return NULL_RTX; - } - - temp = expand_binop (word_mode, smul_optab, op1_high, op0_low, - NULL_RTX, 0, OPTAB_DIRECT); - if (!temp) - return NULL_RTX; - - /* OP1_HIGH should now be dead. */ - - adjust = expand_binop (word_mode, add_optab, adjust, temp, - NULL_RTX, 0, OPTAB_DIRECT); - - if (target && !REG_P (target)) - target = NULL_RTX; - - /* *_widen_optab needs to determine operand mode, make sure at least - one operand has non-VOID mode. */ - if (GET_MODE (op0_low) == VOIDmode && GET_MODE (op1_low) == VOIDmode) - op0_low = force_reg (word_mode, op0_low); - - if (umulp) - product = expand_binop (mode, umul_widen_optab, op0_low, op1_low, - target, 1, OPTAB_DIRECT); - else - product = expand_binop (mode, smul_widen_optab, op0_low, op1_low, - target, 1, OPTAB_DIRECT); - - if (!product) - return NULL_RTX; - - product_high = operand_subword (product, high, 1, mode); - adjust = expand_binop (word_mode, add_optab, product_high, adjust, - NULL_RTX, 0, OPTAB_DIRECT); - emit_move_insn (product_high, adjust); - return product; -} - -/* Subroutine of expand_binop. Optimize unsigned double-word OP0 % OP1 for - constant OP1. If for some bit in [BITS_PER_WORD / 2, BITS_PER_WORD] range - (prefer higher bits) ((1w << bit) % OP1) == 1, then the modulo can be - computed in word-mode as ((OP0 & (bit - 1)) + ((OP0 >> bit) & (bit - 1)) - + (OP0 >> (2 * bit))) % OP1. Whether we need to sum 2, 3 or 4 values - depends on the bit value, if 2, then carry from the addition needs to be - added too, i.e. like: - sum += __builtin_add_overflow (low, high, &sum) - - Optimize signed double-word OP0 % OP1 similarly, just apply some correction - factor to the sum before doing unsigned remainder, in the form of - sum += (((signed) OP0 >> (2 * BITS_PER_WORD - 1)) & const); - then perform unsigned - remainder = sum % OP1; - and finally - remainder += ((signed) OP0 >> (2 * BITS_PER_WORD - 1)) & (1 - OP1); */ - -static rtx -expand_doubleword_mod (machine_mode mode, rtx op0, rtx op1, bool unsignedp) -{ - if (INTVAL (op1) <= 1 || (INTVAL (op1) & 1) == 0) - return NULL_RTX; - - rtx_insn *last = get_last_insn (); - for (int bit = BITS_PER_WORD; bit >= BITS_PER_WORD / 2; bit--) - { - wide_int w = wi::shifted_mask (bit, 1, false, 2 * BITS_PER_WORD); - if (wi::ne_p (wi::umod_trunc (w, INTVAL (op1)), 1)) - continue; - rtx sum = NULL_RTX, mask = NULL_RTX; - if (bit == BITS_PER_WORD) - { - /* For signed modulo we need to add correction to the sum - and that might again overflow. */ - if (!unsignedp) - continue; - if (optab_handler (uaddv4_optab, word_mode) == CODE_FOR_nothing) - continue; - tree wtype = lang_hooks.types.type_for_mode (word_mode, 1); - if (wtype == NULL_TREE) - continue; - tree ctype = build_complex_type (wtype); - if (TYPE_MODE (ctype) != GET_MODE_COMPLEX_MODE (word_mode)) - continue; - machine_mode cmode = TYPE_MODE (ctype); - rtx op00 = operand_subword_force (op0, 0, mode); - rtx op01 = operand_subword_force (op0, 1, mode); - rtx cres = gen_rtx_CONCAT (cmode, gen_reg_rtx (word_mode), - gen_reg_rtx (word_mode)); - tree lhs = make_tree (ctype, cres); - tree arg0 = make_tree (wtype, op00); - tree arg1 = make_tree (wtype, op01); - expand_addsub_overflow (UNKNOWN_LOCATION, PLUS_EXPR, lhs, arg0, - arg1, true, true, true, false, NULL); - sum = expand_simple_binop (word_mode, PLUS, XEXP (cres, 0), - XEXP (cres, 1), NULL_RTX, 1, - OPTAB_DIRECT); - if (sum == NULL_RTX) - return NULL_RTX; - } - else - { - /* Code below uses GEN_INT, so we need the masks to be representable - in HOST_WIDE_INTs. */ - if (bit >= HOST_BITS_PER_WIDE_INT) - continue; - /* If op0 is e.g. -1 or -2 unsigned, then the 2 additions might - overflow. Consider 64-bit -1ULL for word size 32, if we add - 0x7fffffffU + 0x7fffffffU + 3U, it wraps around to 1. */ - if (bit == BITS_PER_WORD - 1) - continue; - - int count = (2 * BITS_PER_WORD + bit - 1) / bit; - rtx sum_corr = NULL_RTX; - - if (!unsignedp) - { - /* For signed modulo, compute it as unsigned modulo of - sum with a correction added to it if OP0 is negative, - such that the result can be computed as unsigned - remainder + ((OP1 >> (2 * BITS_PER_WORD - 1)) & (1 - OP1). */ - w = wi::min_value (2 * BITS_PER_WORD, SIGNED); - wide_int wmod1 = wi::umod_trunc (w, INTVAL (op1)); - wide_int wmod2 = wi::smod_trunc (w, INTVAL (op1)); - /* wmod2 == -wmod1. */ - wmod2 = wmod2 + (INTVAL (op1) - 1); - if (wi::ne_p (wmod1, wmod2)) - { - wide_int wcorr = wmod2 - wmod1; - if (wi::neg_p (w)) - wcorr = wcorr + INTVAL (op1); - /* Now verify if the count sums can't overflow, and punt - if they could. */ - w = wi::mask (bit, false, 2 * BITS_PER_WORD); - w = w * (count - 1); - w = w + wi::mask (2 * BITS_PER_WORD - (count - 1) * bit, - false, 2 * BITS_PER_WORD); - w = w + wcorr; - w = wi::lrshift (w, BITS_PER_WORD); - if (wi::ne_p (w, 0)) - continue; - - mask = operand_subword_force (op0, WORDS_BIG_ENDIAN ? 0 : 1, - mode); - mask = expand_simple_binop (word_mode, ASHIFTRT, mask, - GEN_INT (BITS_PER_WORD - 1), - NULL_RTX, 0, OPTAB_DIRECT); - if (mask == NULL_RTX) - return NULL_RTX; - sum_corr = immed_wide_int_const (wcorr, word_mode); - sum_corr = expand_simple_binop (word_mode, AND, mask, - sum_corr, NULL_RTX, 1, - OPTAB_DIRECT); - if (sum_corr == NULL_RTX) - return NULL_RTX; - } - } - - for (int i = 0; i < count; i++) - { - rtx v = op0; - if (i) - v = expand_simple_binop (mode, LSHIFTRT, v, GEN_INT (i * bit), - NULL_RTX, 1, OPTAB_DIRECT); - if (v == NULL_RTX) - return NULL_RTX; - v = lowpart_subreg (word_mode, v, mode); - if (v == NULL_RTX) - return NULL_RTX; - if (i != count - 1) - v = expand_simple_binop (word_mode, AND, v, - GEN_INT ((HOST_WIDE_INT_1U << bit) - - 1), NULL_RTX, 1, - OPTAB_DIRECT); - if (v == NULL_RTX) - return NULL_RTX; - if (sum == NULL_RTX) - sum = v; - else - sum = expand_simple_binop (word_mode, PLUS, sum, v, NULL_RTX, - 1, OPTAB_DIRECT); - if (sum == NULL_RTX) - return NULL_RTX; - } - if (sum_corr) - { - sum = expand_simple_binop (word_mode, PLUS, sum, sum_corr, - NULL_RTX, 1, OPTAB_DIRECT); - if (sum == NULL_RTX) - return NULL_RTX; - } - } - rtx remainder = expand_divmod (1, TRUNC_MOD_EXPR, word_mode, sum, - gen_int_mode (INTVAL (op1), word_mode), - NULL_RTX, 1, OPTAB_DIRECT); - if (remainder == NULL_RTX) - return NULL_RTX; - - if (!unsignedp) - { - if (mask == NULL_RTX) - { - mask = operand_subword_force (op0, WORDS_BIG_ENDIAN ? 0 : 1, - mode); - mask = expand_simple_binop (word_mode, ASHIFTRT, mask, - GEN_INT (BITS_PER_WORD - 1), - NULL_RTX, 0, OPTAB_DIRECT); - if (mask == NULL_RTX) - return NULL_RTX; - } - mask = expand_simple_binop (word_mode, AND, mask, - gen_int_mode (1 - INTVAL (op1), - word_mode), - NULL_RTX, 1, OPTAB_DIRECT); - if (mask == NULL_RTX) - return NULL_RTX; - remainder = expand_simple_binop (word_mode, PLUS, remainder, - mask, NULL_RTX, 1, OPTAB_DIRECT); - if (remainder == NULL_RTX) - return NULL_RTX; - } - - remainder = convert_modes (mode, word_mode, remainder, unsignedp); - /* Punt if we need any library calls. */ - if (last) - last = NEXT_INSN (last); - else - last = get_insns (); - for (; last; last = NEXT_INSN (last)) - if (CALL_P (last)) - return NULL_RTX; - return remainder; - } - return NULL_RTX; -} - -/* Similarly to the above function, but compute both quotient and remainder. - Quotient can be computed from the remainder as: - rem = op0 % op1; // Handled using expand_doubleword_mod - quot = (op0 - rem) * inv; // inv is multiplicative inverse of op1 modulo - // 2 * BITS_PER_WORD - - We can also handle cases where op1 is a multiple of power of two constant - and constant handled by expand_doubleword_mod. - op11 = 1 << __builtin_ctz (op1); - op12 = op1 / op11; - rem1 = op0 % op12; // Handled using expand_doubleword_mod - quot1 = (op0 - rem1) * inv; // inv is multiplicative inverse of op12 modulo - // 2 * BITS_PER_WORD - rem = (quot1 % op11) * op12 + rem1; - quot = quot1 / op11; */ - -rtx -expand_doubleword_divmod (machine_mode mode, rtx op0, rtx op1, rtx *rem, - bool unsignedp) -{ - *rem = NULL_RTX; - - /* Negative dividend should have been optimized into positive, - similarly modulo by 1 and modulo by power of two is optimized - differently too. */ - if (INTVAL (op1) <= 1 || pow2p_hwi (INTVAL (op1))) - return NULL_RTX; - - rtx op11 = const1_rtx; - rtx op12 = op1; - if ((INTVAL (op1) & 1) == 0) - { - int bit = ctz_hwi (INTVAL (op1)); - op11 = GEN_INT (HOST_WIDE_INT_1 << bit); - op12 = GEN_INT (INTVAL (op1) >> bit); - } - - rtx rem1 = expand_doubleword_mod (mode, op0, op12, unsignedp); - if (rem1 == NULL_RTX) - return NULL_RTX; - - int prec = 2 * BITS_PER_WORD; - wide_int a = wide_int::from (INTVAL (op12), prec + 1, UNSIGNED); - wide_int b = wi::shifted_mask (prec, 1, false, prec + 1); - wide_int m = wide_int::from (wi::mod_inv (a, b), prec, UNSIGNED); - rtx inv = immed_wide_int_const (m, mode); - - rtx_insn *last = get_last_insn (); - rtx quot1 = expand_simple_binop (mode, MINUS, op0, rem1, - NULL_RTX, unsignedp, OPTAB_DIRECT); - if (quot1 == NULL_RTX) - return NULL_RTX; - - quot1 = expand_simple_binop (mode, MULT, quot1, inv, - NULL_RTX, unsignedp, OPTAB_DIRECT); - if (quot1 == NULL_RTX) - return NULL_RTX; - - if (op11 != const1_rtx) - { - rtx rem2 = expand_divmod (1, TRUNC_MOD_EXPR, mode, quot1, op11, - NULL_RTX, unsignedp, OPTAB_DIRECT); - if (rem2 == NULL_RTX) - return NULL_RTX; - - rem2 = expand_simple_binop (mode, MULT, rem2, op12, NULL_RTX, - unsignedp, OPTAB_DIRECT); - if (rem2 == NULL_RTX) - return NULL_RTX; - - rem2 = expand_simple_binop (mode, PLUS, rem2, rem1, NULL_RTX, - unsignedp, OPTAB_DIRECT); - if (rem2 == NULL_RTX) - return NULL_RTX; - - rtx quot2 = expand_divmod (0, TRUNC_DIV_EXPR, mode, quot1, op11, - NULL_RTX, unsignedp, OPTAB_DIRECT); - if (quot2 == NULL_RTX) - return NULL_RTX; - - rem1 = rem2; - quot1 = quot2; - } - - /* Punt if we need any library calls. */ - if (last) - last = NEXT_INSN (last); - else - last = get_insns (); - for (; last; last = NEXT_INSN (last)) - if (CALL_P (last)) - return NULL_RTX; - - *rem = rem1; - return quot1; -} - -/* Wrapper around expand_binop which takes an rtx code to specify - the operation to perform, not an optab pointer. All other - arguments are the same. */ -rtx -expand_simple_binop (machine_mode mode, enum rtx_code code, rtx op0, - rtx op1, rtx target, int unsignedp, - enum optab_methods methods) -{ - optab binop = code_to_optab (code); - gcc_assert (binop); - - return expand_binop (mode, binop, op0, op1, target, unsignedp, methods); -} - -/* Return whether OP0 and OP1 should be swapped when expanding a commutative - binop. Order them according to commutative_operand_precedence and, if - possible, try to put TARGET or a pseudo first. */ -static bool -swap_commutative_operands_with_target (rtx target, rtx op0, rtx op1) -{ - int op0_prec = commutative_operand_precedence (op0); - int op1_prec = commutative_operand_precedence (op1); - - if (op0_prec < op1_prec) - return true; - - if (op0_prec > op1_prec) - return false; - - /* With equal precedence, both orders are ok, but it is better if the - first operand is TARGET, or if both TARGET and OP0 are pseudos. */ - if (target == 0 || REG_P (target)) - return (REG_P (op1) && !REG_P (op0)) || target == op1; - else - return rtx_equal_p (op1, target); -} - -/* Return true if BINOPTAB implements a shift operation. */ - -static bool -shift_optab_p (optab binoptab) -{ - switch (optab_to_code (binoptab)) - { - case ASHIFT: - case SS_ASHIFT: - case US_ASHIFT: - case ASHIFTRT: - case LSHIFTRT: - case ROTATE: - case ROTATERT: - return true; - - default: - return false; - } -} - -/* Return true if BINOPTAB implements a commutative binary operation. */ - -static bool -commutative_optab_p (optab binoptab) -{ - return (GET_RTX_CLASS (optab_to_code (binoptab)) == RTX_COMM_ARITH - || binoptab == smul_widen_optab - || binoptab == umul_widen_optab - || binoptab == smul_highpart_optab - || binoptab == umul_highpart_optab); -} - -/* X is to be used in mode MODE as operand OPN to BINOPTAB. If we're - optimizing, and if the operand is a constant that costs more than - 1 instruction, force the constant into a register and return that - register. Return X otherwise. UNSIGNEDP says whether X is unsigned. */ - -static rtx -avoid_expensive_constant (machine_mode mode, optab binoptab, - int opn, rtx x, bool unsignedp) -{ - bool speed = optimize_insn_for_speed_p (); - - if (mode != VOIDmode - && optimize - && CONSTANT_P (x) - && (rtx_cost (x, mode, optab_to_code (binoptab), opn, speed) - > set_src_cost (x, mode, speed))) - { - if (CONST_INT_P (x)) - { - HOST_WIDE_INT intval = trunc_int_for_mode (INTVAL (x), mode); - if (intval != INTVAL (x)) - x = GEN_INT (intval); - } - else - x = convert_modes (mode, VOIDmode, x, unsignedp); - x = force_reg (mode, x); - } - return x; -} - -/* Helper function for expand_binop: handle the case where there - is an insn ICODE that directly implements the indicated operation. - Returns null if this is not possible. */ -static rtx -expand_binop_directly (enum insn_code icode, machine_mode mode, optab binoptab, - rtx op0, rtx op1, - rtx target, int unsignedp, enum optab_methods methods, - rtx_insn *last) -{ - machine_mode xmode0 = insn_data[(int) icode].operand[1].mode; - machine_mode xmode1 = insn_data[(int) icode].operand[2].mode; - machine_mode mode0, mode1, tmp_mode; - class expand_operand ops[3]; - bool commutative_p; - rtx_insn *pat; - rtx xop0 = op0, xop1 = op1; - bool canonicalize_op1 = false; - - /* If it is a commutative operator and the modes would match - if we would swap the operands, we can save the conversions. */ - commutative_p = commutative_optab_p (binoptab); - if (commutative_p - && GET_MODE (xop0) != xmode0 && GET_MODE (xop1) != xmode1 - && GET_MODE (xop0) == xmode1 && GET_MODE (xop1) == xmode0) - std::swap (xop0, xop1); - - /* If we are optimizing, force expensive constants into a register. */ - xop0 = avoid_expensive_constant (xmode0, binoptab, 0, xop0, unsignedp); - if (!shift_optab_p (binoptab)) - xop1 = avoid_expensive_constant (xmode1, binoptab, 1, xop1, unsignedp); - else - /* Shifts and rotates often use a different mode for op1 from op0; - for VOIDmode constants we don't know the mode, so force it - to be canonicalized using convert_modes. */ - canonicalize_op1 = true; - - /* In case the insn wants input operands in modes different from - those of the actual operands, convert the operands. It would - seem that we don't need to convert CONST_INTs, but we do, so - that they're properly zero-extended, sign-extended or truncated - for their mode. */ - - mode0 = GET_MODE (xop0) != VOIDmode ? GET_MODE (xop0) : mode; - if (xmode0 != VOIDmode && xmode0 != mode0) - { - xop0 = convert_modes (xmode0, mode0, xop0, unsignedp); - mode0 = xmode0; - } - - mode1 = ((GET_MODE (xop1) != VOIDmode || canonicalize_op1) - ? GET_MODE (xop1) : mode); - if (xmode1 != VOIDmode && xmode1 != mode1) - { - xop1 = convert_modes (xmode1, mode1, xop1, unsignedp); - mode1 = xmode1; - } - - /* If operation is commutative, - try to make the first operand a register. - Even better, try to make it the same as the target. - Also try to make the last operand a constant. */ - if (commutative_p - && swap_commutative_operands_with_target (target, xop0, xop1)) - std::swap (xop0, xop1); - - /* Now, if insn's predicates don't allow our operands, put them into - pseudo regs. */ - - if (binoptab == vec_pack_trunc_optab - || binoptab == vec_pack_usat_optab - || binoptab == vec_pack_ssat_optab - || binoptab == vec_pack_ufix_trunc_optab - || binoptab == vec_pack_sfix_trunc_optab - || binoptab == vec_packu_float_optab - || binoptab == vec_packs_float_optab) - { - /* The mode of the result is different then the mode of the - arguments. */ - tmp_mode = insn_data[(int) icode].operand[0].mode; - if (VECTOR_MODE_P (mode) - && maybe_ne (GET_MODE_NUNITS (tmp_mode), 2 * GET_MODE_NUNITS (mode))) - { - delete_insns_since (last); - return NULL_RTX; - } - } - else - tmp_mode = mode; - - create_output_operand (&ops[0], target, tmp_mode); - create_input_operand (&ops[1], xop0, mode0); - create_input_operand (&ops[2], xop1, mode1); - pat = maybe_gen_insn (icode, 3, ops); - if (pat) - { - /* If PAT is composed of more than one insn, try to add an appropriate - REG_EQUAL note to it. If we can't because TEMP conflicts with an - operand, call expand_binop again, this time without a target. */ - if (INSN_P (pat) && NEXT_INSN (pat) != NULL_RTX - && ! add_equal_note (pat, ops[0].value, - optab_to_code (binoptab), - ops[1].value, ops[2].value, mode0)) - { - delete_insns_since (last); - return expand_binop (mode, binoptab, op0, op1, NULL_RTX, - unsignedp, methods); - } - - emit_insn (pat); - return ops[0].value; - } - delete_insns_since (last); - return NULL_RTX; -} - -/* Generate code to perform an operation specified by BINOPTAB - on operands OP0 and OP1, with result having machine-mode MODE. - - UNSIGNEDP is for the case where we have to widen the operands - to perform the operation. It says to use zero-extension. - - If TARGET is nonzero, the value - is generated there, if it is convenient to do so. - In all cases an rtx is returned for the locus of the value; - this may or may not be TARGET. */ - -rtx -expand_binop (machine_mode mode, optab binoptab, rtx op0, rtx op1, - rtx target, int unsignedp, enum optab_methods methods) -{ - enum optab_methods next_methods - = (methods == OPTAB_LIB || methods == OPTAB_LIB_WIDEN - ? OPTAB_WIDEN : methods); - enum mode_class mclass; - enum insn_code icode; - machine_mode wider_mode; - scalar_int_mode int_mode; - rtx libfunc; - rtx temp; - rtx_insn *entry_last = get_last_insn (); - rtx_insn *last; - - mclass = GET_MODE_CLASS (mode); - - /* If subtracting an integer constant, convert this into an addition of - the negated constant. */ - - if (binoptab == sub_optab && CONST_INT_P (op1)) - { - op1 = negate_rtx (mode, op1); - binoptab = add_optab; - } - /* For shifts, constant invalid op1 might be expanded from different - mode than MODE. As those are invalid, force them to a register - to avoid further problems during expansion. */ - else if (CONST_INT_P (op1) - && shift_optab_p (binoptab) - && UINTVAL (op1) >= GET_MODE_BITSIZE (GET_MODE_INNER (mode))) - { - op1 = gen_int_mode (INTVAL (op1), GET_MODE_INNER (mode)); - op1 = force_reg (GET_MODE_INNER (mode), op1); - } - - /* Record where to delete back to if we backtrack. */ - last = get_last_insn (); - - /* If we can do it with a three-operand insn, do so. */ - - if (methods != OPTAB_MUST_WIDEN) - { - if (convert_optab_p (binoptab)) - { - machine_mode from_mode = widened_mode (mode, op0, op1); - icode = find_widening_optab_handler (binoptab, mode, from_mode); - } - else - icode = optab_handler (binoptab, mode); - if (icode != CODE_FOR_nothing) - { - temp = expand_binop_directly (icode, mode, binoptab, op0, op1, - target, unsignedp, methods, last); - if (temp) - return temp; - } - } - - /* If we were trying to rotate, and that didn't work, try rotating - the other direction before falling back to shifts and bitwise-or. */ - if (((binoptab == rotl_optab - && (icode = optab_handler (rotr_optab, mode)) != CODE_FOR_nothing) - || (binoptab == rotr_optab - && (icode = optab_handler (rotl_optab, mode)) != CODE_FOR_nothing)) - && is_int_mode (mode, &int_mode)) - { - optab otheroptab = (binoptab == rotl_optab ? rotr_optab : rotl_optab); - rtx newop1; - unsigned int bits = GET_MODE_PRECISION (int_mode); - - if (CONST_INT_P (op1)) - newop1 = gen_int_shift_amount (int_mode, bits - INTVAL (op1)); - else if (targetm.shift_truncation_mask (int_mode) == bits - 1) - newop1 = negate_rtx (GET_MODE (op1), op1); - else - newop1 = expand_binop (GET_MODE (op1), sub_optab, - gen_int_mode (bits, GET_MODE (op1)), op1, - NULL_RTX, unsignedp, OPTAB_DIRECT); - - temp = expand_binop_directly (icode, int_mode, otheroptab, op0, newop1, - target, unsignedp, methods, last); - if (temp) - return temp; - } - - /* If this is a multiply, see if we can do a widening operation that - takes operands of this mode and makes a wider mode. */ - - if (binoptab == smul_optab - && GET_MODE_2XWIDER_MODE (mode).exists (&wider_mode) - && (convert_optab_handler ((unsignedp - ? umul_widen_optab - : smul_widen_optab), - wider_mode, mode) != CODE_FOR_nothing)) - { - /* *_widen_optab needs to determine operand mode, make sure at least - one operand has non-VOID mode. */ - if (GET_MODE (op0) == VOIDmode && GET_MODE (op1) == VOIDmode) - op0 = force_reg (mode, op0); - temp = expand_binop (wider_mode, - unsignedp ? umul_widen_optab : smul_widen_optab, - op0, op1, NULL_RTX, unsignedp, OPTAB_DIRECT); - - if (temp != 0) - { - if (GET_MODE_CLASS (mode) == MODE_INT - && TRULY_NOOP_TRUNCATION_MODES_P (mode, GET_MODE (temp))) - return gen_lowpart (mode, temp); - else - return convert_to_mode (mode, temp, unsignedp); - } - } - - /* If this is a vector shift by a scalar, see if we can do a vector - shift by a vector. If so, broadcast the scalar into a vector. */ - if (mclass == MODE_VECTOR_INT) - { - optab otheroptab = unknown_optab; - - if (binoptab == ashl_optab) - otheroptab = vashl_optab; - else if (binoptab == ashr_optab) - otheroptab = vashr_optab; - else if (binoptab == lshr_optab) - otheroptab = vlshr_optab; - else if (binoptab == rotl_optab) - otheroptab = vrotl_optab; - else if (binoptab == rotr_optab) - otheroptab = vrotr_optab; - - if (otheroptab - && (icode = optab_handler (otheroptab, mode)) != CODE_FOR_nothing) - { - /* The scalar may have been extended to be too wide. Truncate - it back to the proper size to fit in the broadcast vector. */ - scalar_mode inner_mode = GET_MODE_INNER (mode); - if (!CONST_INT_P (op1) - && (GET_MODE_BITSIZE (as_a <scalar_int_mode> (GET_MODE (op1))) - > GET_MODE_BITSIZE (inner_mode))) - op1 = force_reg (inner_mode, - simplify_gen_unary (TRUNCATE, inner_mode, op1, - GET_MODE (op1))); - rtx vop1 = expand_vector_broadcast (mode, op1); - if (vop1) - { - temp = expand_binop_directly (icode, mode, otheroptab, op0, vop1, - target, unsignedp, methods, last); - if (temp) - return temp; - } - } - } - - /* Look for a wider mode of the same class for which we think we - can open-code the operation. Check for a widening multiply at the - wider mode as well. */ - - if (CLASS_HAS_WIDER_MODES_P (mclass) - && methods != OPTAB_DIRECT && methods != OPTAB_LIB) - FOR_EACH_WIDER_MODE (wider_mode, mode) - { - machine_mode next_mode; - if (optab_handler (binoptab, wider_mode) != CODE_FOR_nothing - || (binoptab == smul_optab - && GET_MODE_WIDER_MODE (wider_mode).exists (&next_mode) - && (find_widening_optab_handler ((unsignedp - ? umul_widen_optab - : smul_widen_optab), - next_mode, mode) - != CODE_FOR_nothing))) - { - rtx xop0 = op0, xop1 = op1; - int no_extend = 0; - - /* For certain integer operations, we need not actually extend - the narrow operands, as long as we will truncate - the results to the same narrowness. */ - - if ((binoptab == ior_optab || binoptab == and_optab - || binoptab == xor_optab - || binoptab == add_optab || binoptab == sub_optab - || binoptab == smul_optab || binoptab == ashl_optab) - && mclass == MODE_INT) - { - no_extend = 1; - xop0 = avoid_expensive_constant (mode, binoptab, 0, - xop0, unsignedp); - if (binoptab != ashl_optab) - xop1 = avoid_expensive_constant (mode, binoptab, 1, - xop1, unsignedp); - } - - xop0 = widen_operand (xop0, wider_mode, mode, unsignedp, no_extend); - - /* The second operand of a shift must always be extended. */ - xop1 = widen_operand (xop1, wider_mode, mode, unsignedp, - no_extend && binoptab != ashl_optab); - - temp = expand_binop (wider_mode, binoptab, xop0, xop1, NULL_RTX, - unsignedp, OPTAB_DIRECT); - if (temp) - { - if (mclass != MODE_INT - || !TRULY_NOOP_TRUNCATION_MODES_P (mode, wider_mode)) - { - if (target == 0) - target = gen_reg_rtx (mode); - convert_move (target, temp, 0); - return target; - } - else - return gen_lowpart (mode, temp); - } - else - delete_insns_since (last); - } - } - - /* If operation is commutative, - try to make the first operand a register. - Even better, try to make it the same as the target. - Also try to make the last operand a constant. */ - if (commutative_optab_p (binoptab) - && swap_commutative_operands_with_target (target, op0, op1)) - std::swap (op0, op1); - - /* These can be done a word at a time. */ - if ((binoptab == and_optab || binoptab == ior_optab || binoptab == xor_optab) - && is_int_mode (mode, &int_mode) - && GET_MODE_SIZE (int_mode) > UNITS_PER_WORD - && optab_handler (binoptab, word_mode) != CODE_FOR_nothing) - { - int i; - rtx_insn *insns; - - /* If TARGET is the same as one of the operands, the REG_EQUAL note - won't be accurate, so use a new target. */ - if (target == 0 - || target == op0 - || target == op1 - || reg_overlap_mentioned_p (target, op0) - || reg_overlap_mentioned_p (target, op1) - || !valid_multiword_target_p (target)) - target = gen_reg_rtx (int_mode); - - start_sequence (); - - /* Do the actual arithmetic. */ - machine_mode op0_mode = GET_MODE (op0); - machine_mode op1_mode = GET_MODE (op1); - if (op0_mode == VOIDmode) - op0_mode = int_mode; - if (op1_mode == VOIDmode) - op1_mode = int_mode; - for (i = 0; i < GET_MODE_BITSIZE (int_mode) / BITS_PER_WORD; i++) - { - rtx target_piece = operand_subword (target, i, 1, int_mode); - rtx x = expand_binop (word_mode, binoptab, - operand_subword_force (op0, i, op0_mode), - operand_subword_force (op1, i, op1_mode), - target_piece, unsignedp, next_methods); - - if (x == 0) - break; - - if (target_piece != x) - emit_move_insn (target_piece, x); - } - - insns = get_insns (); - end_sequence (); - - if (i == GET_MODE_BITSIZE (int_mode) / BITS_PER_WORD) - { - emit_insn (insns); - return target; - } - } - - /* Synthesize double word shifts from single word shifts. */ - if ((binoptab == lshr_optab || binoptab == ashl_optab - || binoptab == ashr_optab) - && is_int_mode (mode, &int_mode) - && (CONST_INT_P (op1) || optimize_insn_for_speed_p ()) - && GET_MODE_SIZE (int_mode) == 2 * UNITS_PER_WORD - && GET_MODE_PRECISION (int_mode) == GET_MODE_BITSIZE (int_mode) - && optab_handler (binoptab, word_mode) != CODE_FOR_nothing - && optab_handler (ashl_optab, word_mode) != CODE_FOR_nothing - && optab_handler (lshr_optab, word_mode) != CODE_FOR_nothing) - { - unsigned HOST_WIDE_INT shift_mask, double_shift_mask; - scalar_int_mode op1_mode; - - double_shift_mask = targetm.shift_truncation_mask (int_mode); - shift_mask = targetm.shift_truncation_mask (word_mode); - op1_mode = (GET_MODE (op1) != VOIDmode - ? as_a <scalar_int_mode> (GET_MODE (op1)) - : word_mode); - - /* Apply the truncation to constant shifts. */ - if (double_shift_mask > 0 && CONST_INT_P (op1)) - op1 = gen_int_mode (INTVAL (op1) & double_shift_mask, op1_mode); - - if (op1 == CONST0_RTX (op1_mode)) - return op0; - - /* Make sure that this is a combination that expand_doubleword_shift - can handle. See the comments there for details. */ - if (double_shift_mask == 0 - || (shift_mask == BITS_PER_WORD - 1 - && double_shift_mask == BITS_PER_WORD * 2 - 1)) - { - rtx_insn *insns; - rtx into_target, outof_target; - rtx into_input, outof_input; - int left_shift, outof_word; - - /* If TARGET is the same as one of the operands, the REG_EQUAL note - won't be accurate, so use a new target. */ - if (target == 0 - || target == op0 - || target == op1 - || reg_overlap_mentioned_p (target, op0) - || reg_overlap_mentioned_p (target, op1) - || !valid_multiword_target_p (target)) - target = gen_reg_rtx (int_mode); - - start_sequence (); - - /* OUTOF_* is the word we are shifting bits away from, and - INTO_* is the word that we are shifting bits towards, thus - they differ depending on the direction of the shift and - WORDS_BIG_ENDIAN. */ - - left_shift = binoptab == ashl_optab; - outof_word = left_shift ^ ! WORDS_BIG_ENDIAN; - - outof_target = operand_subword (target, outof_word, 1, int_mode); - into_target = operand_subword (target, 1 - outof_word, 1, int_mode); - - outof_input = operand_subword_force (op0, outof_word, int_mode); - into_input = operand_subword_force (op0, 1 - outof_word, int_mode); - - if (expand_doubleword_shift (op1_mode, binoptab, - outof_input, into_input, op1, - outof_target, into_target, - unsignedp, next_methods, shift_mask)) - { - insns = get_insns (); - end_sequence (); - - emit_insn (insns); - return target; - } - end_sequence (); - } - } - - /* Synthesize double word rotates from single word shifts. */ - if ((binoptab == rotl_optab || binoptab == rotr_optab) - && is_int_mode (mode, &int_mode) - && CONST_INT_P (op1) - && GET_MODE_PRECISION (int_mode) == 2 * BITS_PER_WORD - && optab_handler (ashl_optab, word_mode) != CODE_FOR_nothing - && optab_handler (lshr_optab, word_mode) != CODE_FOR_nothing) - { - rtx_insn *insns; - rtx into_target, outof_target; - rtx into_input, outof_input; - rtx inter; - int shift_count, left_shift, outof_word; - - /* If TARGET is the same as one of the operands, the REG_EQUAL note - won't be accurate, so use a new target. Do this also if target is not - a REG, first because having a register instead may open optimization - opportunities, and second because if target and op0 happen to be MEMs - designating the same location, we would risk clobbering it too early - in the code sequence we generate below. */ - if (target == 0 - || target == op0 - || target == op1 - || !REG_P (target) - || reg_overlap_mentioned_p (target, op0) - || reg_overlap_mentioned_p (target, op1) - || !valid_multiword_target_p (target)) - target = gen_reg_rtx (int_mode); - - start_sequence (); - - shift_count = INTVAL (op1); - - /* OUTOF_* is the word we are shifting bits away from, and - INTO_* is the word that we are shifting bits towards, thus - they differ depending on the direction of the shift and - WORDS_BIG_ENDIAN. */ - - left_shift = (binoptab == rotl_optab); - outof_word = left_shift ^ ! WORDS_BIG_ENDIAN; - - outof_target = operand_subword (target, outof_word, 1, int_mode); - into_target = operand_subword (target, 1 - outof_word, 1, int_mode); - - outof_input = operand_subword_force (op0, outof_word, int_mode); - into_input = operand_subword_force (op0, 1 - outof_word, int_mode); - - if (shift_count == BITS_PER_WORD) - { - /* This is just a word swap. */ - emit_move_insn (outof_target, into_input); - emit_move_insn (into_target, outof_input); - inter = const0_rtx; - } - else - { - rtx into_temp1, into_temp2, outof_temp1, outof_temp2; - HOST_WIDE_INT first_shift_count, second_shift_count; - optab reverse_unsigned_shift, unsigned_shift; - - reverse_unsigned_shift = (left_shift ^ (shift_count < BITS_PER_WORD) - ? lshr_optab : ashl_optab); - - unsigned_shift = (left_shift ^ (shift_count < BITS_PER_WORD) - ? ashl_optab : lshr_optab); - - if (shift_count > BITS_PER_WORD) - { - first_shift_count = shift_count - BITS_PER_WORD; - second_shift_count = 2 * BITS_PER_WORD - shift_count; - } - else - { - first_shift_count = BITS_PER_WORD - shift_count; - second_shift_count = shift_count; - } - rtx first_shift_count_rtx - = gen_int_shift_amount (word_mode, first_shift_count); - rtx second_shift_count_rtx - = gen_int_shift_amount (word_mode, second_shift_count); - - into_temp1 = expand_binop (word_mode, unsigned_shift, - outof_input, first_shift_count_rtx, - NULL_RTX, unsignedp, next_methods); - into_temp2 = expand_binop (word_mode, reverse_unsigned_shift, - into_input, second_shift_count_rtx, - NULL_RTX, unsignedp, next_methods); - - if (into_temp1 != 0 && into_temp2 != 0) - inter = expand_binop (word_mode, ior_optab, into_temp1, into_temp2, - into_target, unsignedp, next_methods); - else - inter = 0; - - if (inter != 0 && inter != into_target) - emit_move_insn (into_target, inter); - - outof_temp1 = expand_binop (word_mode, unsigned_shift, - into_input, first_shift_count_rtx, - NULL_RTX, unsignedp, next_methods); - outof_temp2 = expand_binop (word_mode, reverse_unsigned_shift, - outof_input, second_shift_count_rtx, - NULL_RTX, unsignedp, next_methods); - - if (inter != 0 && outof_temp1 != 0 && outof_temp2 != 0) - inter = expand_binop (word_mode, ior_optab, - outof_temp1, outof_temp2, - outof_target, unsignedp, next_methods); - - if (inter != 0 && inter != outof_target) - emit_move_insn (outof_target, inter); - } - - insns = get_insns (); - end_sequence (); - - if (inter != 0) - { - emit_insn (insns); - return target; - } - } - - /* These can be done a word at a time by propagating carries. */ - if ((binoptab == add_optab || binoptab == sub_optab) - && is_int_mode (mode, &int_mode) - && GET_MODE_SIZE (int_mode) >= 2 * UNITS_PER_WORD - && optab_handler (binoptab, word_mode) != CODE_FOR_nothing) - { - unsigned int i; - optab otheroptab = binoptab == add_optab ? sub_optab : add_optab; - const unsigned int nwords = GET_MODE_BITSIZE (int_mode) / BITS_PER_WORD; - rtx carry_in = NULL_RTX, carry_out = NULL_RTX; - rtx xop0, xop1, xtarget; - - /* We can handle either a 1 or -1 value for the carry. If STORE_FLAG - value is one of those, use it. Otherwise, use 1 since it is the - one easiest to get. */ -#if STORE_FLAG_VALUE == 1 || STORE_FLAG_VALUE == -1 - int normalizep = STORE_FLAG_VALUE; -#else - int normalizep = 1; -#endif - - /* Prepare the operands. */ - xop0 = force_reg (int_mode, op0); - xop1 = force_reg (int_mode, op1); - - xtarget = gen_reg_rtx (int_mode); - - if (target == 0 || !REG_P (target) || !valid_multiword_target_p (target)) - target = xtarget; - - /* Indicate for flow that the entire target reg is being set. */ - if (REG_P (target)) - emit_clobber (xtarget); - - /* Do the actual arithmetic. */ - for (i = 0; i < nwords; i++) - { - int index = (WORDS_BIG_ENDIAN ? nwords - i - 1 : i); - rtx target_piece = operand_subword (xtarget, index, 1, int_mode); - rtx op0_piece = operand_subword_force (xop0, index, int_mode); - rtx op1_piece = operand_subword_force (xop1, index, int_mode); - rtx x; - - /* Main add/subtract of the input operands. */ - x = expand_binop (word_mode, binoptab, - op0_piece, op1_piece, - target_piece, unsignedp, next_methods); - if (x == 0) - break; - - if (i + 1 < nwords) - { - /* Store carry from main add/subtract. */ - carry_out = gen_reg_rtx (word_mode); - carry_out = emit_store_flag_force (carry_out, - (binoptab == add_optab - ? LT : GT), - x, op0_piece, - word_mode, 1, normalizep); - } - - if (i > 0) - { - rtx newx; - - /* Add/subtract previous carry to main result. */ - newx = expand_binop (word_mode, - normalizep == 1 ? binoptab : otheroptab, - x, carry_in, - NULL_RTX, 1, next_methods); - - if (i + 1 < nwords) - { - /* Get out carry from adding/subtracting carry in. */ - rtx carry_tmp = gen_reg_rtx (word_mode); - carry_tmp = emit_store_flag_force (carry_tmp, - (binoptab == add_optab - ? LT : GT), - newx, x, - word_mode, 1, normalizep); - - /* Logical-ior the two poss. carry together. */ - carry_out = expand_binop (word_mode, ior_optab, - carry_out, carry_tmp, - carry_out, 0, next_methods); - if (carry_out == 0) - break; - } - emit_move_insn (target_piece, newx); - } - else - { - if (x != target_piece) - emit_move_insn (target_piece, x); - } - - carry_in = carry_out; - } - - if (i == GET_MODE_BITSIZE (int_mode) / (unsigned) BITS_PER_WORD) - { - if (optab_handler (mov_optab, int_mode) != CODE_FOR_nothing - || ! rtx_equal_p (target, xtarget)) - { - rtx_insn *temp = emit_move_insn (target, xtarget); - - set_dst_reg_note (temp, REG_EQUAL, - gen_rtx_fmt_ee (optab_to_code (binoptab), - int_mode, copy_rtx (xop0), - copy_rtx (xop1)), - target); - } - else - target = xtarget; - - return target; - } - - else - delete_insns_since (last); - } - - /* Attempt to synthesize double word multiplies using a sequence of word - mode multiplications. We first attempt to generate a sequence using a - more efficient unsigned widening multiply, and if that fails we then - try using a signed widening multiply. */ - - if (binoptab == smul_optab - && is_int_mode (mode, &int_mode) - && GET_MODE_SIZE (int_mode) == 2 * UNITS_PER_WORD - && optab_handler (smul_optab, word_mode) != CODE_FOR_nothing - && optab_handler (add_optab, word_mode) != CODE_FOR_nothing) - { - rtx product = NULL_RTX; - if (convert_optab_handler (umul_widen_optab, int_mode, word_mode) - != CODE_FOR_nothing) - { - product = expand_doubleword_mult (int_mode, op0, op1, target, - true, methods); - if (!product) - delete_insns_since (last); - } - - if (product == NULL_RTX - && (convert_optab_handler (smul_widen_optab, int_mode, word_mode) - != CODE_FOR_nothing)) - { - product = expand_doubleword_mult (int_mode, op0, op1, target, - false, methods); - if (!product) - delete_insns_since (last); - } - - if (product != NULL_RTX) - { - if (optab_handler (mov_optab, int_mode) != CODE_FOR_nothing) - { - rtx_insn *move = emit_move_insn (target ? target : product, - product); - set_dst_reg_note (move, - REG_EQUAL, - gen_rtx_fmt_ee (MULT, int_mode, - copy_rtx (op0), - copy_rtx (op1)), - target ? target : product); - } - return product; - } - } - - /* Attempt to synthetize double word modulo by constant divisor. */ - if ((binoptab == umod_optab - || binoptab == smod_optab - || binoptab == udiv_optab - || binoptab == sdiv_optab) - && optimize - && CONST_INT_P (op1) - && is_int_mode (mode, &int_mode) - && GET_MODE_SIZE (int_mode) == 2 * UNITS_PER_WORD - && optab_handler ((binoptab == umod_optab || binoptab == udiv_optab) - ? udivmod_optab : sdivmod_optab, - int_mode) == CODE_FOR_nothing - && optab_handler (and_optab, word_mode) != CODE_FOR_nothing - && optab_handler (add_optab, word_mode) != CODE_FOR_nothing - && optimize_insn_for_speed_p ()) - { - rtx res = NULL_RTX; - if ((binoptab == umod_optab || binoptab == smod_optab) - && (INTVAL (op1) & 1) == 0) - res = expand_doubleword_mod (int_mode, op0, op1, - binoptab == umod_optab); - else - { - rtx quot = expand_doubleword_divmod (int_mode, op0, op1, &res, - binoptab == umod_optab - || binoptab == udiv_optab); - if (quot == NULL_RTX) - res = NULL_RTX; - else if (binoptab == udiv_optab || binoptab == sdiv_optab) - res = quot; - } - if (res != NULL_RTX) - { - if (optab_handler (mov_optab, int_mode) != CODE_FOR_nothing) - { - rtx_insn *move = emit_move_insn (target ? target : res, - res); - set_dst_reg_note (move, REG_EQUAL, - gen_rtx_fmt_ee (optab_to_code (binoptab), - int_mode, copy_rtx (op0), op1), - target ? target : res); - } - return res; - } - else - delete_insns_since (last); - } - - /* It can't be open-coded in this mode. - Use a library call if one is available and caller says that's ok. */ - - libfunc = optab_libfunc (binoptab, mode); - if (libfunc - && (methods == OPTAB_LIB || methods == OPTAB_LIB_WIDEN)) - { - rtx_insn *insns; - rtx op1x = op1; - machine_mode op1_mode = mode; - rtx value; - - start_sequence (); - - if (shift_optab_p (binoptab)) - { - op1_mode = targetm.libgcc_shift_count_mode (); - /* Specify unsigned here, - since negative shift counts are meaningless. */ - op1x = convert_to_mode (op1_mode, op1, 1); - } - - if (GET_MODE (op0) != VOIDmode - && GET_MODE (op0) != mode) - op0 = convert_to_mode (mode, op0, unsignedp); - - /* Pass 1 for NO_QUEUE so we don't lose any increments - if the libcall is cse'd or moved. */ - value = emit_library_call_value (libfunc, - NULL_RTX, LCT_CONST, mode, - op0, mode, op1x, op1_mode); - - insns = get_insns (); - end_sequence (); - - bool trapv = trapv_binoptab_p (binoptab); - target = gen_reg_rtx (mode); - emit_libcall_block_1 (insns, target, value, - trapv ? NULL_RTX - : gen_rtx_fmt_ee (optab_to_code (binoptab), - mode, op0, op1), trapv); - - return target; - } - - delete_insns_since (last); - - /* It can't be done in this mode. Can we do it in a wider mode? */ - - if (! (methods == OPTAB_WIDEN || methods == OPTAB_LIB_WIDEN - || methods == OPTAB_MUST_WIDEN)) - { - /* Caller says, don't even try. */ - delete_insns_since (entry_last); - return 0; - } - - /* Compute the value of METHODS to pass to recursive calls. - Don't allow widening to be tried recursively. */ - - methods = (methods == OPTAB_LIB_WIDEN ? OPTAB_LIB : OPTAB_DIRECT); - - /* Look for a wider mode of the same class for which it appears we can do - the operation. */ - - if (CLASS_HAS_WIDER_MODES_P (mclass)) - { - /* This code doesn't make sense for conversion optabs, since we - wouldn't then want to extend the operands to be the same size - as the result. */ - gcc_assert (!convert_optab_p (binoptab)); - FOR_EACH_WIDER_MODE (wider_mode, mode) - { - if (optab_handler (binoptab, wider_mode) - || (methods == OPTAB_LIB - && optab_libfunc (binoptab, wider_mode))) - { - rtx xop0 = op0, xop1 = op1; - int no_extend = 0; - - /* For certain integer operations, we need not actually extend - the narrow operands, as long as we will truncate - the results to the same narrowness. */ - - if ((binoptab == ior_optab || binoptab == and_optab - || binoptab == xor_optab - || binoptab == add_optab || binoptab == sub_optab - || binoptab == smul_optab || binoptab == ashl_optab) - && mclass == MODE_INT) - no_extend = 1; - - xop0 = widen_operand (xop0, wider_mode, mode, - unsignedp, no_extend); - - /* The second operand of a shift must always be extended. */ - xop1 = widen_operand (xop1, wider_mode, mode, unsignedp, - no_extend && binoptab != ashl_optab); - - temp = expand_binop (wider_mode, binoptab, xop0, xop1, NULL_RTX, - unsignedp, methods); - if (temp) - { - if (mclass != MODE_INT - || !TRULY_NOOP_TRUNCATION_MODES_P (mode, wider_mode)) - { - if (target == 0) - target = gen_reg_rtx (mode); - convert_move (target, temp, 0); - return target; - } - else - return gen_lowpart (mode, temp); - } - else - delete_insns_since (last); - } - } - } - - delete_insns_since (entry_last); - return 0; -} - -/* Expand a binary operator which has both signed and unsigned forms. - UOPTAB is the optab for unsigned operations, and SOPTAB is for - signed operations. - - If we widen unsigned operands, we may use a signed wider operation instead - of an unsigned wider operation, since the result would be the same. */ - -rtx -sign_expand_binop (machine_mode mode, optab uoptab, optab soptab, - rtx op0, rtx op1, rtx target, int unsignedp, - enum optab_methods methods) -{ - rtx temp; - optab direct_optab = unsignedp ? uoptab : soptab; - bool save_enable; - - /* Do it without widening, if possible. */ - temp = expand_binop (mode, direct_optab, op0, op1, target, - unsignedp, OPTAB_DIRECT); - if (temp || methods == OPTAB_DIRECT) - return temp; - - /* Try widening to a signed int. Disable any direct use of any - signed insn in the current mode. */ - save_enable = swap_optab_enable (soptab, mode, false); - - temp = expand_binop (mode, soptab, op0, op1, target, - unsignedp, OPTAB_WIDEN); - - /* For unsigned operands, try widening to an unsigned int. */ - if (!temp && unsignedp) - temp = expand_binop (mode, uoptab, op0, op1, target, - unsignedp, OPTAB_WIDEN); - if (temp || methods == OPTAB_WIDEN) - goto egress; - - /* Use the right width libcall if that exists. */ - temp = expand_binop (mode, direct_optab, op0, op1, target, - unsignedp, OPTAB_LIB); - if (temp || methods == OPTAB_LIB) - goto egress; - - /* Must widen and use a libcall, use either signed or unsigned. */ - temp = expand_binop (mode, soptab, op0, op1, target, - unsignedp, methods); - if (!temp && unsignedp) - temp = expand_binop (mode, uoptab, op0, op1, target, - unsignedp, methods); - - egress: - /* Undo the fiddling above. */ - if (save_enable) - swap_optab_enable (soptab, mode, true); - return temp; -} - -/* Generate code to perform an operation specified by UNOPPTAB - on operand OP0, with two results to TARG0 and TARG1. - We assume that the order of the operands for the instruction - is TARG0, TARG1, OP0. - - Either TARG0 or TARG1 may be zero, but what that means is that - the result is not actually wanted. We will generate it into - a dummy pseudo-reg and discard it. They may not both be zero. - - Returns 1 if this operation can be performed; 0 if not. */ - -int -expand_twoval_unop (optab unoptab, rtx op0, rtx targ0, rtx targ1, - int unsignedp) -{ - machine_mode mode = GET_MODE (targ0 ? targ0 : targ1); - enum mode_class mclass; - machine_mode wider_mode; - rtx_insn *entry_last = get_last_insn (); - rtx_insn *last; - - mclass = GET_MODE_CLASS (mode); - - if (!targ0) - targ0 = gen_reg_rtx (mode); - if (!targ1) - targ1 = gen_reg_rtx (mode); - - /* Record where to go back to if we fail. */ - last = get_last_insn (); - - if (optab_handler (unoptab, mode) != CODE_FOR_nothing) - { - class expand_operand ops[3]; - enum insn_code icode = optab_handler (unoptab, mode); - - create_fixed_operand (&ops[0], targ0); - create_fixed_operand (&ops[1], targ1); - create_convert_operand_from (&ops[2], op0, mode, unsignedp); - if (maybe_expand_insn (icode, 3, ops)) - return 1; - } - - /* It can't be done in this mode. Can we do it in a wider mode? */ - - if (CLASS_HAS_WIDER_MODES_P (mclass)) - { - FOR_EACH_WIDER_MODE (wider_mode, mode) - { - if (optab_handler (unoptab, wider_mode) != CODE_FOR_nothing) - { - rtx t0 = gen_reg_rtx (wider_mode); - rtx t1 = gen_reg_rtx (wider_mode); - rtx cop0 = convert_modes (wider_mode, mode, op0, unsignedp); - - if (expand_twoval_unop (unoptab, cop0, t0, t1, unsignedp)) - { - convert_move (targ0, t0, unsignedp); - convert_move (targ1, t1, unsignedp); - return 1; - } - else - delete_insns_since (last); - } - } - } - - delete_insns_since (entry_last); - return 0; -} - -/* Generate code to perform an operation specified by BINOPTAB - on operands OP0 and OP1, with two results to TARG1 and TARG2. - We assume that the order of the operands for the instruction - is TARG0, OP0, OP1, TARG1, which would fit a pattern like - [(set TARG0 (operate OP0 OP1)) (set TARG1 (operate ...))]. - - Either TARG0 or TARG1 may be zero, but what that means is that - the result is not actually wanted. We will generate it into - a dummy pseudo-reg and discard it. They may not both be zero. - - Returns 1 if this operation can be performed; 0 if not. */ - -int -expand_twoval_binop (optab binoptab, rtx op0, rtx op1, rtx targ0, rtx targ1, - int unsignedp) -{ - machine_mode mode = GET_MODE (targ0 ? targ0 : targ1); - enum mode_class mclass; - machine_mode wider_mode; - rtx_insn *entry_last = get_last_insn (); - rtx_insn *last; - - mclass = GET_MODE_CLASS (mode); - - if (!targ0) - targ0 = gen_reg_rtx (mode); - if (!targ1) - targ1 = gen_reg_rtx (mode); - - /* Record where to go back to if we fail. */ - last = get_last_insn (); - - if (optab_handler (binoptab, mode) != CODE_FOR_nothing) - { - class expand_operand ops[4]; - enum insn_code icode = optab_handler (binoptab, mode); - machine_mode mode0 = insn_data[icode].operand[1].mode; - machine_mode mode1 = insn_data[icode].operand[2].mode; - rtx xop0 = op0, xop1 = op1; - - /* If we are optimizing, force expensive constants into a register. */ - xop0 = avoid_expensive_constant (mode0, binoptab, 0, xop0, unsignedp); - xop1 = avoid_expensive_constant (mode1, binoptab, 1, xop1, unsignedp); - - create_fixed_operand (&ops[0], targ0); - create_convert_operand_from (&ops[1], xop0, mode, unsignedp); - create_convert_operand_from (&ops[2], xop1, mode, unsignedp); - create_fixed_operand (&ops[3], targ1); - if (maybe_expand_insn (icode, 4, ops)) - return 1; - delete_insns_since (last); - } - - /* It can't be done in this mode. Can we do it in a wider mode? */ - - if (CLASS_HAS_WIDER_MODES_P (mclass)) - { - FOR_EACH_WIDER_MODE (wider_mode, mode) - { - if (optab_handler (binoptab, wider_mode) != CODE_FOR_nothing) - { - rtx t0 = gen_reg_rtx (wider_mode); - rtx t1 = gen_reg_rtx (wider_mode); - rtx cop0 = convert_modes (wider_mode, mode, op0, unsignedp); - rtx cop1 = convert_modes (wider_mode, mode, op1, unsignedp); - - if (expand_twoval_binop (binoptab, cop0, cop1, - t0, t1, unsignedp)) - { - convert_move (targ0, t0, unsignedp); - convert_move (targ1, t1, unsignedp); - return 1; - } - else - delete_insns_since (last); - } - } - } - - delete_insns_since (entry_last); - return 0; -} - -/* Expand the two-valued library call indicated by BINOPTAB, but - preserve only one of the values. If TARG0 is non-NULL, the first - value is placed into TARG0; otherwise the second value is placed - into TARG1. Exactly one of TARG0 and TARG1 must be non-NULL. The - value stored into TARG0 or TARG1 is equivalent to (CODE OP0 OP1). - This routine assumes that the value returned by the library call is - as if the return value was of an integral mode twice as wide as the - mode of OP0. Returns 1 if the call was successful. */ - -bool -expand_twoval_binop_libfunc (optab binoptab, rtx op0, rtx op1, - rtx targ0, rtx targ1, enum rtx_code code) -{ - machine_mode mode; - machine_mode libval_mode; - rtx libval; - rtx_insn *insns; - rtx libfunc; - - /* Exactly one of TARG0 or TARG1 should be non-NULL. */ - gcc_assert (!targ0 != !targ1); - - mode = GET_MODE (op0); - libfunc = optab_libfunc (binoptab, mode); - if (!libfunc) - return false; - - /* The value returned by the library function will have twice as - many bits as the nominal MODE. */ - libval_mode = smallest_int_mode_for_size (2 * GET_MODE_BITSIZE (mode)); - start_sequence (); - libval = emit_library_call_value (libfunc, NULL_RTX, LCT_CONST, - libval_mode, - op0, mode, - op1, mode); - /* Get the part of VAL containing the value that we want. */ - libval = simplify_gen_subreg (mode, libval, libval_mode, - targ0 ? 0 : GET_MODE_SIZE (mode)); - insns = get_insns (); - end_sequence (); - /* Move the into the desired location. */ - emit_libcall_block (insns, targ0 ? targ0 : targ1, libval, - gen_rtx_fmt_ee (code, mode, op0, op1)); - - return true; -} - - -/* Wrapper around expand_unop which takes an rtx code to specify - the operation to perform, not an optab pointer. All other - arguments are the same. */ -rtx -expand_simple_unop (machine_mode mode, enum rtx_code code, rtx op0, - rtx target, int unsignedp) -{ - optab unop = code_to_optab (code); - gcc_assert (unop); - - return expand_unop (mode, unop, op0, target, unsignedp); -} - -/* Try calculating - (clz:narrow x) - as - (clz:wide (zero_extend:wide x)) - ((width wide) - (width narrow)). - - A similar operation can be used for clrsb. UNOPTAB says which operation - we are trying to expand. */ -static rtx -widen_leading (scalar_int_mode mode, rtx op0, rtx target, optab unoptab) -{ - opt_scalar_int_mode wider_mode_iter; - FOR_EACH_WIDER_MODE (wider_mode_iter, mode) - { - scalar_int_mode wider_mode = wider_mode_iter.require (); - if (optab_handler (unoptab, wider_mode) != CODE_FOR_nothing) - { - rtx xop0, temp; - rtx_insn *last; - - last = get_last_insn (); - - if (target == 0) - target = gen_reg_rtx (mode); - xop0 = widen_operand (op0, wider_mode, mode, - unoptab != clrsb_optab, false); - temp = expand_unop (wider_mode, unoptab, xop0, NULL_RTX, - unoptab != clrsb_optab); - if (temp != 0) - temp = expand_binop - (wider_mode, sub_optab, temp, - gen_int_mode (GET_MODE_PRECISION (wider_mode) - - GET_MODE_PRECISION (mode), - wider_mode), - target, true, OPTAB_DIRECT); - if (temp == 0) - delete_insns_since (last); - - return temp; - } - } - return 0; -} - -/* Attempt to emit (clrsb:mode op0) as - (plus:mode (clz:mode (xor:mode op0 (ashr:mode op0 (const_int prec-1)))) - (const_int -1)) - if CLZ_DEFINED_VALUE_AT_ZERO (mode, val) is 2 and val is prec, - or as - (clz:mode (ior:mode (xor:mode (ashl:mode op0 (const_int 1)) - (ashr:mode op0 (const_int prec-1))) - (const_int 1))) - otherwise. */ - -static rtx -expand_clrsb_using_clz (scalar_int_mode mode, rtx op0, rtx target) -{ - if (optimize_insn_for_size_p () - || optab_handler (clz_optab, mode) == CODE_FOR_nothing) - return NULL_RTX; - - start_sequence (); - HOST_WIDE_INT val = 0; - if (CLZ_DEFINED_VALUE_AT_ZERO (mode, val) != 2 - || val != GET_MODE_PRECISION (mode)) - val = 0; - else - val = 1; - - rtx temp2 = op0; - if (!val) - { - temp2 = expand_binop (mode, ashl_optab, op0, const1_rtx, - NULL_RTX, 0, OPTAB_DIRECT); - if (!temp2) - { - fail: - end_sequence (); - return NULL_RTX; - } - } - - rtx temp = expand_binop (mode, ashr_optab, op0, - GEN_INT (GET_MODE_PRECISION (mode) - 1), - NULL_RTX, 0, OPTAB_DIRECT); - if (!temp) - goto fail; - - temp = expand_binop (mode, xor_optab, temp2, temp, NULL_RTX, 0, - OPTAB_DIRECT); - if (!temp) - goto fail; - - if (!val) - { - temp = expand_binop (mode, ior_optab, temp, const1_rtx, - NULL_RTX, 0, OPTAB_DIRECT); - if (!temp) - goto fail; - } - temp = expand_unop_direct (mode, clz_optab, temp, val ? NULL_RTX : target, - true); - if (!temp) - goto fail; - if (val) - { - temp = expand_binop (mode, add_optab, temp, constm1_rtx, - target, 0, OPTAB_DIRECT); - if (!temp) - goto fail; - } - - rtx_insn *seq = get_insns (); - end_sequence (); - - add_equal_note (seq, temp, CLRSB, op0, NULL_RTX, mode); - emit_insn (seq); - return temp; -} - -/* Try calculating clz of a double-word quantity as two clz's of word-sized - quantities, choosing which based on whether the high word is nonzero. */ -static rtx -expand_doubleword_clz (scalar_int_mode mode, rtx op0, rtx target) -{ - rtx xop0 = force_reg (mode, op0); - rtx subhi = gen_highpart (word_mode, xop0); - rtx sublo = gen_lowpart (word_mode, xop0); - rtx_code_label *hi0_label = gen_label_rtx (); - rtx_code_label *after_label = gen_label_rtx (); - rtx_insn *seq; - rtx temp, result; - - /* If we were not given a target, use a word_mode register, not a - 'mode' register. The result will fit, and nobody is expecting - anything bigger (the return type of __builtin_clz* is int). */ - if (!target) - target = gen_reg_rtx (word_mode); - - /* In any case, write to a word_mode scratch in both branches of the - conditional, so we can ensure there is a single move insn setting - 'target' to tag a REG_EQUAL note on. */ - result = gen_reg_rtx (word_mode); - - start_sequence (); - - /* If the high word is not equal to zero, - then clz of the full value is clz of the high word. */ - emit_cmp_and_jump_insns (subhi, CONST0_RTX (word_mode), EQ, 0, - word_mode, true, hi0_label); - - temp = expand_unop_direct (word_mode, clz_optab, subhi, result, true); - if (!temp) - goto fail; - - if (temp != result) - convert_move (result, temp, true); - - emit_jump_insn (targetm.gen_jump (after_label)); - emit_barrier (); - - /* Else clz of the full value is clz of the low word plus the number - of bits in the high word. */ - emit_label (hi0_label); - - temp = expand_unop_direct (word_mode, clz_optab, sublo, 0, true); - if (!temp) - goto fail; - temp = expand_binop (word_mode, add_optab, temp, - gen_int_mode (GET_MODE_BITSIZE (word_mode), word_mode), - result, true, OPTAB_DIRECT); - if (!temp) - goto fail; - if (temp != result) - convert_move (result, temp, true); - - emit_label (after_label); - convert_move (target, result, true); - - seq = get_insns (); - end_sequence (); - - add_equal_note (seq, target, CLZ, xop0, NULL_RTX, mode); - emit_insn (seq); - return target; - - fail: - end_sequence (); - return 0; -} - -/* Try calculating popcount of a double-word quantity as two popcount's of - word-sized quantities and summing up the results. */ -static rtx -expand_doubleword_popcount (scalar_int_mode mode, rtx op0, rtx target) -{ - rtx t0, t1, t; - rtx_insn *seq; - - start_sequence (); - - t0 = expand_unop_direct (word_mode, popcount_optab, - operand_subword_force (op0, 0, mode), NULL_RTX, - true); - t1 = expand_unop_direct (word_mode, popcount_optab, - operand_subword_force (op0, 1, mode), NULL_RTX, - true); - if (!t0 || !t1) - { - end_sequence (); - return NULL_RTX; - } - - /* If we were not given a target, use a word_mode register, not a - 'mode' register. The result will fit, and nobody is expecting - anything bigger (the return type of __builtin_popcount* is int). */ - if (!target) - target = gen_reg_rtx (word_mode); - - t = expand_binop (word_mode, add_optab, t0, t1, target, 0, OPTAB_DIRECT); - - seq = get_insns (); - end_sequence (); - - add_equal_note (seq, t, POPCOUNT, op0, NULL_RTX, mode); - emit_insn (seq); - return t; -} - -/* Try calculating - (parity:wide x) - as - (parity:narrow (low (x) ^ high (x))) */ -static rtx -expand_doubleword_parity (scalar_int_mode mode, rtx op0, rtx target) -{ - rtx t = expand_binop (word_mode, xor_optab, - operand_subword_force (op0, 0, mode), - operand_subword_force (op0, 1, mode), - NULL_RTX, 0, OPTAB_DIRECT); - return expand_unop (word_mode, parity_optab, t, target, true); -} - -/* Try calculating - (bswap:narrow x) - as - (lshiftrt:wide (bswap:wide x) ((width wide) - (width narrow))). */ -static rtx -widen_bswap (scalar_int_mode mode, rtx op0, rtx target) -{ - rtx x; - rtx_insn *last; - opt_scalar_int_mode wider_mode_iter; - - FOR_EACH_WIDER_MODE (wider_mode_iter, mode) - if (optab_handler (bswap_optab, wider_mode_iter.require ()) - != CODE_FOR_nothing) - break; - - if (!wider_mode_iter.exists ()) - return NULL_RTX; - - scalar_int_mode wider_mode = wider_mode_iter.require (); - last = get_last_insn (); - - x = widen_operand (op0, wider_mode, mode, true, true); - x = expand_unop (wider_mode, bswap_optab, x, NULL_RTX, true); - - gcc_assert (GET_MODE_PRECISION (wider_mode) == GET_MODE_BITSIZE (wider_mode) - && GET_MODE_PRECISION (mode) == GET_MODE_BITSIZE (mode)); - if (x != 0) - x = expand_shift (RSHIFT_EXPR, wider_mode, x, - GET_MODE_BITSIZE (wider_mode) - - GET_MODE_BITSIZE (mode), - NULL_RTX, true); - - if (x != 0) - { - if (target == 0) - target = gen_reg_rtx (mode); - emit_move_insn (target, gen_lowpart (mode, x)); - } - else - delete_insns_since (last); - - return target; -} - -/* Try calculating bswap as two bswaps of two word-sized operands. */ - -static rtx -expand_doubleword_bswap (machine_mode mode, rtx op, rtx target) -{ - rtx t0, t1; - - t1 = expand_unop (word_mode, bswap_optab, - operand_subword_force (op, 0, mode), NULL_RTX, true); - t0 = expand_unop (word_mode, bswap_optab, - operand_subword_force (op, 1, mode), NULL_RTX, true); - - if (target == 0 || !valid_multiword_target_p (target)) - target = gen_reg_rtx (mode); - if (REG_P (target)) - emit_clobber (target); - emit_move_insn (operand_subword (target, 0, 1, mode), t0); - emit_move_insn (operand_subword (target, 1, 1, mode), t1); - - return target; -} - -/* Try calculating (parity x) as (and (popcount x) 1), where - popcount can also be done in a wider mode. */ -static rtx -expand_parity (scalar_int_mode mode, rtx op0, rtx target) -{ - enum mode_class mclass = GET_MODE_CLASS (mode); - opt_scalar_int_mode wider_mode_iter; - FOR_EACH_MODE_FROM (wider_mode_iter, mode) - { - scalar_int_mode wider_mode = wider_mode_iter.require (); - if (optab_handler (popcount_optab, wider_mode) != CODE_FOR_nothing) - { - rtx xop0, temp; - rtx_insn *last; - - last = get_last_insn (); - - if (target == 0 || GET_MODE (target) != wider_mode) - target = gen_reg_rtx (wider_mode); - - xop0 = widen_operand (op0, wider_mode, mode, true, false); - temp = expand_unop (wider_mode, popcount_optab, xop0, NULL_RTX, - true); - if (temp != 0) - temp = expand_binop (wider_mode, and_optab, temp, const1_rtx, - target, true, OPTAB_DIRECT); - - if (temp) - { - if (mclass != MODE_INT - || !TRULY_NOOP_TRUNCATION_MODES_P (mode, wider_mode)) - return convert_to_mode (mode, temp, 0); - else - return gen_lowpart (mode, temp); - } - else - delete_insns_since (last); - } - } - return 0; -} - -/* Try calculating ctz(x) as K - clz(x & -x) , - where K is GET_MODE_PRECISION(mode) - 1. - - Both __builtin_ctz and __builtin_clz are undefined at zero, so we - don't have to worry about what the hardware does in that case. (If - the clz instruction produces the usual value at 0, which is K, the - result of this code sequence will be -1; expand_ffs, below, relies - on this. It might be nice to have it be K instead, for consistency - with the (very few) processors that provide a ctz with a defined - value, but that would take one more instruction, and it would be - less convenient for expand_ffs anyway. */ - -static rtx -expand_ctz (scalar_int_mode mode, rtx op0, rtx target) -{ - rtx_insn *seq; - rtx temp; - - if (optab_handler (clz_optab, mode) == CODE_FOR_nothing) - return 0; - - start_sequence (); - - temp = expand_unop_direct (mode, neg_optab, op0, NULL_RTX, true); - if (temp) - temp = expand_binop (mode, and_optab, op0, temp, NULL_RTX, - true, OPTAB_DIRECT); - if (temp) - temp = expand_unop_direct (mode, clz_optab, temp, NULL_RTX, true); - if (temp) - temp = expand_binop (mode, sub_optab, - gen_int_mode (GET_MODE_PRECISION (mode) - 1, mode), - temp, target, - true, OPTAB_DIRECT); - if (temp == 0) - { - end_sequence (); - return 0; - } - - seq = get_insns (); - end_sequence (); - - add_equal_note (seq, temp, CTZ, op0, NULL_RTX, mode); - emit_insn (seq); - return temp; -} - - -/* Try calculating ffs(x) using ctz(x) if we have that instruction, or - else with the sequence used by expand_clz. - - The ffs builtin promises to return zero for a zero value and ctz/clz - may have an undefined value in that case. If they do not give us a - convenient value, we have to generate a test and branch. */ -static rtx -expand_ffs (scalar_int_mode mode, rtx op0, rtx target) -{ - HOST_WIDE_INT val = 0; - bool defined_at_zero = false; - rtx temp; - rtx_insn *seq; - - if (optab_handler (ctz_optab, mode) != CODE_FOR_nothing) - { - start_sequence (); - - temp = expand_unop_direct (mode, ctz_optab, op0, 0, true); - if (!temp) - goto fail; - - defined_at_zero = (CTZ_DEFINED_VALUE_AT_ZERO (mode, val) == 2); - } - else if (optab_handler (clz_optab, mode) != CODE_FOR_nothing) - { - start_sequence (); - temp = expand_ctz (mode, op0, 0); - if (!temp) - goto fail; - - if (CLZ_DEFINED_VALUE_AT_ZERO (mode, val) == 2) - { - defined_at_zero = true; - val = (GET_MODE_PRECISION (mode) - 1) - val; - } - } - else - return 0; - - if (defined_at_zero && val == -1) - /* No correction needed at zero. */; - else - { - /* We don't try to do anything clever with the situation found - on some processors (eg Alpha) where ctz(0:mode) == - bitsize(mode). If someone can think of a way to send N to -1 - and leave alone all values in the range 0..N-1 (where N is a - power of two), cheaper than this test-and-branch, please add it. - - The test-and-branch is done after the operation itself, in case - the operation sets condition codes that can be recycled for this. - (This is true on i386, for instance.) */ - - rtx_code_label *nonzero_label = gen_label_rtx (); - emit_cmp_and_jump_insns (op0, CONST0_RTX (mode), NE, 0, - mode, true, nonzero_label); - - convert_move (temp, GEN_INT (-1), false); - emit_label (nonzero_label); - } - - /* temp now has a value in the range -1..bitsize-1. ffs is supposed - to produce a value in the range 0..bitsize. */ - temp = expand_binop (mode, add_optab, temp, gen_int_mode (1, mode), - target, false, OPTAB_DIRECT); - if (!temp) - goto fail; - - seq = get_insns (); - end_sequence (); - - add_equal_note (seq, temp, FFS, op0, NULL_RTX, mode); - emit_insn (seq); - return temp; - - fail: - end_sequence (); - return 0; -} - -/* Extract the OMODE lowpart from VAL, which has IMODE. Under certain - conditions, VAL may already be a SUBREG against which we cannot generate - a further SUBREG. In this case, we expect forcing the value into a - register will work around the situation. */ - -static rtx -lowpart_subreg_maybe_copy (machine_mode omode, rtx val, - machine_mode imode) -{ - rtx ret; - ret = lowpart_subreg (omode, val, imode); - if (ret == NULL) - { - val = force_reg (imode, val); - ret = lowpart_subreg (omode, val, imode); - gcc_assert (ret != NULL); - } - return ret; -} - -/* Expand a floating point absolute value or negation operation via a - logical operation on the sign bit. */ - -static rtx -expand_absneg_bit (enum rtx_code code, scalar_float_mode mode, - rtx op0, rtx target) -{ - const struct real_format *fmt; - int bitpos, word, nwords, i; - scalar_int_mode imode; - rtx temp; - rtx_insn *insns; - - /* The format has to have a simple sign bit. */ - fmt = REAL_MODE_FORMAT (mode); - if (fmt == NULL) - return NULL_RTX; - - bitpos = fmt->signbit_rw; - if (bitpos < 0) - return NULL_RTX; - - /* Don't create negative zeros if the format doesn't support them. */ - if (code == NEG && !fmt->has_signed_zero) - return NULL_RTX; - - if (GET_MODE_SIZE (mode) <= UNITS_PER_WORD) - { - if (!int_mode_for_mode (mode).exists (&imode)) - return NULL_RTX; - word = 0; - nwords = 1; - } - else - { - imode = word_mode; - - if (FLOAT_WORDS_BIG_ENDIAN) - word = (GET_MODE_BITSIZE (mode) - bitpos) / BITS_PER_WORD; - else - word = bitpos / BITS_PER_WORD; - bitpos = bitpos % BITS_PER_WORD; - nwords = (GET_MODE_BITSIZE (mode) + BITS_PER_WORD - 1) / BITS_PER_WORD; - } - - wide_int mask = wi::set_bit_in_zero (bitpos, GET_MODE_PRECISION (imode)); - if (code == ABS) - mask = ~mask; - - if (target == 0 - || target == op0 - || reg_overlap_mentioned_p (target, op0) - || (nwords > 1 && !valid_multiword_target_p (target))) - target = gen_reg_rtx (mode); - - if (nwords > 1) - { - start_sequence (); - - for (i = 0; i < nwords; ++i) - { - rtx targ_piece = operand_subword (target, i, 1, mode); - rtx op0_piece = operand_subword_force (op0, i, mode); - - if (i == word) - { - temp = expand_binop (imode, code == ABS ? and_optab : xor_optab, - op0_piece, - immed_wide_int_const (mask, imode), - targ_piece, 1, OPTAB_LIB_WIDEN); - if (temp != targ_piece) - emit_move_insn (targ_piece, temp); - } - else - emit_move_insn (targ_piece, op0_piece); - } - - insns = get_insns (); - end_sequence (); - - emit_insn (insns); - } - else - { - temp = expand_binop (imode, code == ABS ? and_optab : xor_optab, - gen_lowpart (imode, op0), - immed_wide_int_const (mask, imode), - gen_lowpart (imode, target), 1, OPTAB_LIB_WIDEN); - target = lowpart_subreg_maybe_copy (mode, temp, imode); - - set_dst_reg_note (get_last_insn (), REG_EQUAL, - gen_rtx_fmt_e (code, mode, copy_rtx (op0)), - target); - } - - return target; -} - -/* As expand_unop, but will fail rather than attempt the operation in a - different mode or with a libcall. */ -static rtx -expand_unop_direct (machine_mode mode, optab unoptab, rtx op0, rtx target, - int unsignedp) -{ - if (optab_handler (unoptab, mode) != CODE_FOR_nothing) - { - class expand_operand ops[2]; - enum insn_code icode = optab_handler (unoptab, mode); - rtx_insn *last = get_last_insn (); - rtx_insn *pat; - - create_output_operand (&ops[0], target, mode); - create_convert_operand_from (&ops[1], op0, mode, unsignedp); - pat = maybe_gen_insn (icode, 2, ops); - if (pat) - { - if (INSN_P (pat) && NEXT_INSN (pat) != NULL_RTX - && ! add_equal_note (pat, ops[0].value, - optab_to_code (unoptab), - ops[1].value, NULL_RTX, mode)) - { - delete_insns_since (last); - return expand_unop (mode, unoptab, op0, NULL_RTX, unsignedp); - } - - emit_insn (pat); - - return ops[0].value; - } - } - return 0; -} - -/* Generate code to perform an operation specified by UNOPTAB - on operand OP0, with result having machine-mode MODE. - - UNSIGNEDP is for the case where we have to widen the operands - to perform the operation. It says to use zero-extension. - - If TARGET is nonzero, the value - is generated there, if it is convenient to do so. - In all cases an rtx is returned for the locus of the value; - this may or may not be TARGET. */ - -rtx -expand_unop (machine_mode mode, optab unoptab, rtx op0, rtx target, - int unsignedp) -{ - enum mode_class mclass = GET_MODE_CLASS (mode); - machine_mode wider_mode; - scalar_int_mode int_mode; - scalar_float_mode float_mode; - rtx temp; - rtx libfunc; - - temp = expand_unop_direct (mode, unoptab, op0, target, unsignedp); - if (temp) - return temp; - - /* It can't be done in this mode. Can we open-code it in a wider mode? */ - - /* Widening (or narrowing) clz needs special treatment. */ - if (unoptab == clz_optab) - { - if (is_a <scalar_int_mode> (mode, &int_mode)) - { - temp = widen_leading (int_mode, op0, target, unoptab); - if (temp) - return temp; - - if (GET_MODE_SIZE (int_mode) == 2 * UNITS_PER_WORD - && optab_handler (unoptab, word_mode) != CODE_FOR_nothing) - { - temp = expand_doubleword_clz (int_mode, op0, target); - if (temp) - return temp; - } - } - - goto try_libcall; - } - - if (unoptab == clrsb_optab) - { - if (is_a <scalar_int_mode> (mode, &int_mode)) - { - temp = widen_leading (int_mode, op0, target, unoptab); - if (temp) - return temp; - temp = expand_clrsb_using_clz (int_mode, op0, target); - if (temp) - return temp; - } - goto try_libcall; - } - - if (unoptab == popcount_optab - && is_a <scalar_int_mode> (mode, &int_mode) - && GET_MODE_SIZE (int_mode) == 2 * UNITS_PER_WORD - && optab_handler (unoptab, word_mode) != CODE_FOR_nothing - && optimize_insn_for_speed_p ()) - { - temp = expand_doubleword_popcount (int_mode, op0, target); - if (temp) - return temp; - } - - if (unoptab == parity_optab - && is_a <scalar_int_mode> (mode, &int_mode) - && GET_MODE_SIZE (int_mode) == 2 * UNITS_PER_WORD - && (optab_handler (unoptab, word_mode) != CODE_FOR_nothing - || optab_handler (popcount_optab, word_mode) != CODE_FOR_nothing) - && optimize_insn_for_speed_p ()) - { - temp = expand_doubleword_parity (int_mode, op0, target); - if (temp) - return temp; - } - - /* Widening (or narrowing) bswap needs special treatment. */ - if (unoptab == bswap_optab) - { - /* HImode is special because in this mode BSWAP is equivalent to ROTATE - or ROTATERT. First try these directly; if this fails, then try the - obvious pair of shifts with allowed widening, as this will probably - be always more efficient than the other fallback methods. */ - if (mode == HImode) - { - rtx_insn *last; - rtx temp1, temp2; - - if (optab_handler (rotl_optab, mode) != CODE_FOR_nothing) - { - temp = expand_binop (mode, rotl_optab, op0, - gen_int_shift_amount (mode, 8), - target, unsignedp, OPTAB_DIRECT); - if (temp) - return temp; - } - - if (optab_handler (rotr_optab, mode) != CODE_FOR_nothing) - { - temp = expand_binop (mode, rotr_optab, op0, - gen_int_shift_amount (mode, 8), - target, unsignedp, OPTAB_DIRECT); - if (temp) - return temp; - } - - last = get_last_insn (); - - temp1 = expand_binop (mode, ashl_optab, op0, - gen_int_shift_amount (mode, 8), NULL_RTX, - unsignedp, OPTAB_WIDEN); - temp2 = expand_binop (mode, lshr_optab, op0, - gen_int_shift_amount (mode, 8), NULL_RTX, - unsignedp, OPTAB_WIDEN); - if (temp1 && temp2) - { - temp = expand_binop (mode, ior_optab, temp1, temp2, target, - unsignedp, OPTAB_WIDEN); - if (temp) - return temp; - } - - delete_insns_since (last); - } - - if (is_a <scalar_int_mode> (mode, &int_mode)) - { - temp = widen_bswap (int_mode, op0, target); - if (temp) - return temp; - - /* We do not provide a 128-bit bswap in libgcc so force the use of - a double bswap for 64-bit targets. */ - if (GET_MODE_SIZE (int_mode) == 2 * UNITS_PER_WORD - && (UNITS_PER_WORD == 8 - || optab_handler (unoptab, word_mode) != CODE_FOR_nothing)) - { - temp = expand_doubleword_bswap (mode, op0, target); - if (temp) - return temp; - } - } - - goto try_libcall; - } - - if (CLASS_HAS_WIDER_MODES_P (mclass)) - FOR_EACH_WIDER_MODE (wider_mode, mode) - { - if (optab_handler (unoptab, wider_mode) != CODE_FOR_nothing) - { - rtx xop0 = op0; - rtx_insn *last = get_last_insn (); - - /* For certain operations, we need not actually extend - the narrow operand, as long as we will truncate the - results to the same narrowness. */ - - xop0 = widen_operand (xop0, wider_mode, mode, unsignedp, - (unoptab == neg_optab - || unoptab == one_cmpl_optab) - && mclass == MODE_INT); - - temp = expand_unop (wider_mode, unoptab, xop0, NULL_RTX, - unsignedp); - - if (temp) - { - if (mclass != MODE_INT - || !TRULY_NOOP_TRUNCATION_MODES_P (mode, wider_mode)) - { - if (target == 0) - target = gen_reg_rtx (mode); - convert_move (target, temp, 0); - return target; - } - else - return gen_lowpart (mode, temp); - } - else - delete_insns_since (last); - } - } - - /* These can be done a word at a time. */ - if (unoptab == one_cmpl_optab - && is_int_mode (mode, &int_mode) - && GET_MODE_SIZE (int_mode) > UNITS_PER_WORD - && optab_handler (unoptab, word_mode) != CODE_FOR_nothing) - { - int i; - rtx_insn *insns; - - if (target == 0 - || target == op0 - || reg_overlap_mentioned_p (target, op0) - || !valid_multiword_target_p (target)) - target = gen_reg_rtx (int_mode); - - start_sequence (); - - /* Do the actual arithmetic. */ - for (i = 0; i < GET_MODE_BITSIZE (int_mode) / BITS_PER_WORD; i++) - { - rtx target_piece = operand_subword (target, i, 1, int_mode); - rtx x = expand_unop (word_mode, unoptab, - operand_subword_force (op0, i, int_mode), - target_piece, unsignedp); - - if (target_piece != x) - emit_move_insn (target_piece, x); - } - - insns = get_insns (); - end_sequence (); - - emit_insn (insns); - return target; - } - - /* Emit ~op0 as op0 ^ -1. */ - if (unoptab == one_cmpl_optab - && (SCALAR_INT_MODE_P (mode) || GET_MODE_CLASS (mode) == MODE_VECTOR_INT) - && optab_handler (xor_optab, mode) != CODE_FOR_nothing) - { - temp = expand_binop (mode, xor_optab, op0, CONSTM1_RTX (mode), - target, unsignedp, OPTAB_DIRECT); - if (temp) - return temp; - } - - if (optab_to_code (unoptab) == NEG) - { - /* Try negating floating point values by flipping the sign bit. */ - if (is_a <scalar_float_mode> (mode, &float_mode)) - { - temp = expand_absneg_bit (NEG, float_mode, op0, target); - if (temp) - return temp; - } - - /* If there is no negation pattern, and we have no negative zero, - try subtracting from zero. */ - if (!HONOR_SIGNED_ZEROS (mode)) - { - temp = expand_binop (mode, (unoptab == negv_optab - ? subv_optab : sub_optab), - CONST0_RTX (mode), op0, target, - unsignedp, OPTAB_DIRECT); - if (temp) - return temp; - } - } - - /* Try calculating parity (x) as popcount (x) % 2. */ - if (unoptab == parity_optab && is_a <scalar_int_mode> (mode, &int_mode)) - { - temp = expand_parity (int_mode, op0, target); - if (temp) - return temp; - } - - /* Try implementing ffs (x) in terms of clz (x). */ - if (unoptab == ffs_optab && is_a <scalar_int_mode> (mode, &int_mode)) - { - temp = expand_ffs (int_mode, op0, target); - if (temp) - return temp; - } - - /* Try implementing ctz (x) in terms of clz (x). */ - if (unoptab == ctz_optab && is_a <scalar_int_mode> (mode, &int_mode)) - { - temp = expand_ctz (int_mode, op0, target); - if (temp) - return temp; - } - - try_libcall: - /* Now try a library call in this mode. */ - libfunc = optab_libfunc (unoptab, mode); - if (libfunc) - { - rtx_insn *insns; - rtx value; - rtx eq_value; - machine_mode outmode = mode; - - /* All of these functions return small values. Thus we choose to - have them return something that isn't a double-word. */ - if (unoptab == ffs_optab || unoptab == clz_optab || unoptab == ctz_optab - || unoptab == clrsb_optab || unoptab == popcount_optab - || unoptab == parity_optab) - outmode - = GET_MODE (hard_libcall_value (TYPE_MODE (integer_type_node), - optab_libfunc (unoptab, mode))); - - start_sequence (); - - /* Pass 1 for NO_QUEUE so we don't lose any increments - if the libcall is cse'd or moved. */ - value = emit_library_call_value (libfunc, NULL_RTX, LCT_CONST, outmode, - op0, mode); - insns = get_insns (); - end_sequence (); - - target = gen_reg_rtx (outmode); - bool trapv = trapv_unoptab_p (unoptab); - if (trapv) - eq_value = NULL_RTX; - else - { - eq_value = gen_rtx_fmt_e (optab_to_code (unoptab), mode, op0); - if (GET_MODE_UNIT_SIZE (outmode) < GET_MODE_UNIT_SIZE (mode)) - eq_value = simplify_gen_unary (TRUNCATE, outmode, eq_value, mode); - else if (GET_MODE_UNIT_SIZE (outmode) > GET_MODE_UNIT_SIZE (mode)) - eq_value = simplify_gen_unary (ZERO_EXTEND, - outmode, eq_value, mode); - } - emit_libcall_block_1 (insns, target, value, eq_value, trapv); - - return target; - } - - /* It can't be done in this mode. Can we do it in a wider mode? */ - - if (CLASS_HAS_WIDER_MODES_P (mclass)) - { - FOR_EACH_WIDER_MODE (wider_mode, mode) - { - if (optab_handler (unoptab, wider_mode) != CODE_FOR_nothing - || optab_libfunc (unoptab, wider_mode)) - { - rtx xop0 = op0; - rtx_insn *last = get_last_insn (); - - /* For certain operations, we need not actually extend - the narrow operand, as long as we will truncate the - results to the same narrowness. */ - xop0 = widen_operand (xop0, wider_mode, mode, unsignedp, - (unoptab == neg_optab - || unoptab == one_cmpl_optab - || unoptab == bswap_optab) - && mclass == MODE_INT); - - temp = expand_unop (wider_mode, unoptab, xop0, NULL_RTX, - unsignedp); - - /* If we are generating clz using wider mode, adjust the - result. Similarly for clrsb. */ - if ((unoptab == clz_optab || unoptab == clrsb_optab) - && temp != 0) - { - scalar_int_mode wider_int_mode - = as_a <scalar_int_mode> (wider_mode); - int_mode = as_a <scalar_int_mode> (mode); - temp = expand_binop - (wider_mode, sub_optab, temp, - gen_int_mode (GET_MODE_PRECISION (wider_int_mode) - - GET_MODE_PRECISION (int_mode), - wider_int_mode), - target, true, OPTAB_DIRECT); - } - - /* Likewise for bswap. */ - if (unoptab == bswap_optab && temp != 0) - { - scalar_int_mode wider_int_mode - = as_a <scalar_int_mode> (wider_mode); - int_mode = as_a <scalar_int_mode> (mode); - gcc_assert (GET_MODE_PRECISION (wider_int_mode) - == GET_MODE_BITSIZE (wider_int_mode) - && GET_MODE_PRECISION (int_mode) - == GET_MODE_BITSIZE (int_mode)); - - temp = expand_shift (RSHIFT_EXPR, wider_int_mode, temp, - GET_MODE_BITSIZE (wider_int_mode) - - GET_MODE_BITSIZE (int_mode), - NULL_RTX, true); - } - - if (temp) - { - if (mclass != MODE_INT) - { - if (target == 0) - target = gen_reg_rtx (mode); - convert_move (target, temp, 0); - return target; - } - else - return gen_lowpart (mode, temp); - } - else - delete_insns_since (last); - } - } - } - - /* One final attempt at implementing negation via subtraction, - this time allowing widening of the operand. */ - if (optab_to_code (unoptab) == NEG && !HONOR_SIGNED_ZEROS (mode)) - { - rtx temp; - temp = expand_binop (mode, - unoptab == negv_optab ? subv_optab : sub_optab, - CONST0_RTX (mode), op0, - target, unsignedp, OPTAB_LIB_WIDEN); - if (temp) - return temp; - } - - return 0; -} - -/* Emit code to compute the absolute value of OP0, with result to - TARGET if convenient. (TARGET may be 0.) The return value says - where the result actually is to be found. - - MODE is the mode of the operand; the mode of the result is - different but can be deduced from MODE. - - */ - -rtx -expand_abs_nojump (machine_mode mode, rtx op0, rtx target, - int result_unsignedp) -{ - rtx temp; - - if (GET_MODE_CLASS (mode) != MODE_INT - || ! flag_trapv) - result_unsignedp = 1; - - /* First try to do it with a special abs instruction. */ - temp = expand_unop (mode, result_unsignedp ? abs_optab : absv_optab, - op0, target, 0); - if (temp != 0) - return temp; - - /* For floating point modes, try clearing the sign bit. */ - scalar_float_mode float_mode; - if (is_a <scalar_float_mode> (mode, &float_mode)) - { - temp = expand_absneg_bit (ABS, float_mode, op0, target); - if (temp) - return temp; - } - - /* If we have a MAX insn, we can do this as MAX (x, -x). */ - if (optab_handler (smax_optab, mode) != CODE_FOR_nothing - && !HONOR_SIGNED_ZEROS (mode)) - { - rtx_insn *last = get_last_insn (); - - temp = expand_unop (mode, result_unsignedp ? neg_optab : negv_optab, - op0, NULL_RTX, 0); - if (temp != 0) - temp = expand_binop (mode, smax_optab, op0, temp, target, 0, - OPTAB_WIDEN); - - if (temp != 0) - return temp; - - delete_insns_since (last); - } - - /* If this machine has expensive jumps, we can do integer absolute - value of X as (((signed) x >> (W-1)) ^ x) - ((signed) x >> (W-1)), - where W is the width of MODE. */ - - scalar_int_mode int_mode; - if (is_int_mode (mode, &int_mode) - && BRANCH_COST (optimize_insn_for_speed_p (), - false) >= 2) - { - rtx extended = expand_shift (RSHIFT_EXPR, int_mode, op0, - GET_MODE_PRECISION (int_mode) - 1, - NULL_RTX, 0); - - temp = expand_binop (int_mode, xor_optab, extended, op0, target, 0, - OPTAB_LIB_WIDEN); - if (temp != 0) - temp = expand_binop (int_mode, - result_unsignedp ? sub_optab : subv_optab, - temp, extended, target, 0, OPTAB_LIB_WIDEN); - - if (temp != 0) - return temp; - } - - return NULL_RTX; -} - -rtx -expand_abs (machine_mode mode, rtx op0, rtx target, - int result_unsignedp, int safe) -{ - rtx temp; - rtx_code_label *op1; - - if (GET_MODE_CLASS (mode) != MODE_INT - || ! flag_trapv) - result_unsignedp = 1; - - temp = expand_abs_nojump (mode, op0, target, result_unsignedp); - if (temp != 0) - return temp; - - /* If that does not win, use conditional jump and negate. */ - - /* It is safe to use the target if it is the same - as the source if this is also a pseudo register */ - if (op0 == target && REG_P (op0) - && REGNO (op0) >= FIRST_PSEUDO_REGISTER) - safe = 1; - - op1 = gen_label_rtx (); - if (target == 0 || ! safe - || GET_MODE (target) != mode - || (MEM_P (target) && MEM_VOLATILE_P (target)) - || (REG_P (target) - && REGNO (target) < FIRST_PSEUDO_REGISTER)) - target = gen_reg_rtx (mode); - - emit_move_insn (target, op0); - NO_DEFER_POP; - - do_compare_rtx_and_jump (target, CONST0_RTX (mode), GE, 0, mode, - NULL_RTX, NULL, op1, - profile_probability::uninitialized ()); - - op0 = expand_unop (mode, result_unsignedp ? neg_optab : negv_optab, - target, target, 0); - if (op0 != target) - emit_move_insn (target, op0); - emit_label (op1); - OK_DEFER_POP; - return target; -} - -/* Emit code to compute the one's complement absolute value of OP0 - (if (OP0 < 0) OP0 = ~OP0), with result to TARGET if convenient. - (TARGET may be NULL_RTX.) The return value says where the result - actually is to be found. - - MODE is the mode of the operand; the mode of the result is - different but can be deduced from MODE. */ - -rtx -expand_one_cmpl_abs_nojump (machine_mode mode, rtx op0, rtx target) -{ - rtx temp; - - /* Not applicable for floating point modes. */ - if (FLOAT_MODE_P (mode)) - return NULL_RTX; - - /* If we have a MAX insn, we can do this as MAX (x, ~x). */ - if (optab_handler (smax_optab, mode) != CODE_FOR_nothing) - { - rtx_insn *last = get_last_insn (); - - temp = expand_unop (mode, one_cmpl_optab, op0, NULL_RTX, 0); - if (temp != 0) - temp = expand_binop (mode, smax_optab, op0, temp, target, 0, - OPTAB_WIDEN); - - if (temp != 0) - return temp; - - delete_insns_since (last); - } - - /* If this machine has expensive jumps, we can do one's complement - absolute value of X as (((signed) x >> (W-1)) ^ x). */ - - scalar_int_mode int_mode; - if (is_int_mode (mode, &int_mode) - && BRANCH_COST (optimize_insn_for_speed_p (), - false) >= 2) - { - rtx extended = expand_shift (RSHIFT_EXPR, int_mode, op0, - GET_MODE_PRECISION (int_mode) - 1, - NULL_RTX, 0); - - temp = expand_binop (int_mode, xor_optab, extended, op0, target, 0, - OPTAB_LIB_WIDEN); - - if (temp != 0) - return temp; - } - - return NULL_RTX; -} - -/* A subroutine of expand_copysign, perform the copysign operation using the - abs and neg primitives advertised to exist on the target. The assumption - is that we have a split register file, and leaving op0 in fp registers, - and not playing with subregs so much, will help the register allocator. */ - -static rtx -expand_copysign_absneg (scalar_float_mode mode, rtx op0, rtx op1, rtx target, - int bitpos, bool op0_is_abs) -{ - scalar_int_mode imode; - enum insn_code icode; - rtx sign; - rtx_code_label *label; - - if (target == op1) - target = NULL_RTX; - - /* Check if the back end provides an insn that handles signbit for the - argument's mode. */ - icode = optab_handler (signbit_optab, mode); - if (icode != CODE_FOR_nothing) - { - imode = as_a <scalar_int_mode> (insn_data[(int) icode].operand[0].mode); - sign = gen_reg_rtx (imode); - emit_unop_insn (icode, sign, op1, UNKNOWN); - } - else - { - if (GET_MODE_SIZE (mode) <= UNITS_PER_WORD) - { - if (!int_mode_for_mode (mode).exists (&imode)) - return NULL_RTX; - op1 = gen_lowpart (imode, op1); - } - else - { - int word; - - imode = word_mode; - if (FLOAT_WORDS_BIG_ENDIAN) - word = (GET_MODE_BITSIZE (mode) - bitpos) / BITS_PER_WORD; - else - word = bitpos / BITS_PER_WORD; - bitpos = bitpos % BITS_PER_WORD; - op1 = operand_subword_force (op1, word, mode); - } - - wide_int mask = wi::set_bit_in_zero (bitpos, GET_MODE_PRECISION (imode)); - sign = expand_binop (imode, and_optab, op1, - immed_wide_int_const (mask, imode), - NULL_RTX, 1, OPTAB_LIB_WIDEN); - } - - if (!op0_is_abs) - { - op0 = expand_unop (mode, abs_optab, op0, target, 0); - if (op0 == NULL) - return NULL_RTX; - target = op0; - } - else - { - if (target == NULL_RTX) - target = copy_to_reg (op0); - else - emit_move_insn (target, op0); - } - - label = gen_label_rtx (); - emit_cmp_and_jump_insns (sign, const0_rtx, EQ, NULL_RTX, imode, 1, label); - - if (CONST_DOUBLE_AS_FLOAT_P (op0)) - op0 = simplify_unary_operation (NEG, mode, op0, mode); - else - op0 = expand_unop (mode, neg_optab, op0, target, 0); - if (op0 != target) - emit_move_insn (target, op0); - - emit_label (label); - - return target; -} - - -/* A subroutine of expand_copysign, perform the entire copysign operation - with integer bitmasks. BITPOS is the position of the sign bit; OP0_IS_ABS - is true if op0 is known to have its sign bit clear. */ - -static rtx -expand_copysign_bit (scalar_float_mode mode, rtx op0, rtx op1, rtx target, - int bitpos, bool op0_is_abs) -{ - scalar_int_mode imode; - int word, nwords, i; - rtx temp; - rtx_insn *insns; - - if (GET_MODE_SIZE (mode) <= UNITS_PER_WORD) - { - if (!int_mode_for_mode (mode).exists (&imode)) - return NULL_RTX; - word = 0; - nwords = 1; - } - else - { - imode = word_mode; - - if (FLOAT_WORDS_BIG_ENDIAN) - word = (GET_MODE_BITSIZE (mode) - bitpos) / BITS_PER_WORD; - else - word = bitpos / BITS_PER_WORD; - bitpos = bitpos % BITS_PER_WORD; - nwords = (GET_MODE_BITSIZE (mode) + BITS_PER_WORD - 1) / BITS_PER_WORD; - } - - wide_int mask = wi::set_bit_in_zero (bitpos, GET_MODE_PRECISION (imode)); - - if (target == 0 - || target == op0 - || target == op1 - || reg_overlap_mentioned_p (target, op0) - || reg_overlap_mentioned_p (target, op1) - || (nwords > 1 && !valid_multiword_target_p (target))) - target = gen_reg_rtx (mode); - - if (nwords > 1) - { - start_sequence (); - - for (i = 0; i < nwords; ++i) - { - rtx targ_piece = operand_subword (target, i, 1, mode); - rtx op0_piece = operand_subword_force (op0, i, mode); - - if (i == word) - { - if (!op0_is_abs) - op0_piece - = expand_binop (imode, and_optab, op0_piece, - immed_wide_int_const (~mask, imode), - NULL_RTX, 1, OPTAB_LIB_WIDEN); - op1 = expand_binop (imode, and_optab, - operand_subword_force (op1, i, mode), - immed_wide_int_const (mask, imode), - NULL_RTX, 1, OPTAB_LIB_WIDEN); - - temp = expand_binop (imode, ior_optab, op0_piece, op1, - targ_piece, 1, OPTAB_LIB_WIDEN); - if (temp != targ_piece) - emit_move_insn (targ_piece, temp); - } - else - emit_move_insn (targ_piece, op0_piece); - } - - insns = get_insns (); - end_sequence (); - - emit_insn (insns); - } - else - { - op1 = expand_binop (imode, and_optab, gen_lowpart (imode, op1), - immed_wide_int_const (mask, imode), - NULL_RTX, 1, OPTAB_LIB_WIDEN); - - op0 = gen_lowpart (imode, op0); - if (!op0_is_abs) - op0 = expand_binop (imode, and_optab, op0, - immed_wide_int_const (~mask, imode), - NULL_RTX, 1, OPTAB_LIB_WIDEN); - - temp = expand_binop (imode, ior_optab, op0, op1, - gen_lowpart (imode, target), 1, OPTAB_LIB_WIDEN); - target = lowpart_subreg_maybe_copy (mode, temp, imode); - } - - return target; -} - -/* Expand the C99 copysign operation. OP0 and OP1 must be the same - scalar floating point mode. Return NULL if we do not know how to - expand the operation inline. */ - -rtx -expand_copysign (rtx op0, rtx op1, rtx target) -{ - scalar_float_mode mode; - const struct real_format *fmt; - bool op0_is_abs; - rtx temp; - - mode = as_a <scalar_float_mode> (GET_MODE (op0)); - gcc_assert (GET_MODE (op1) == mode); - - /* First try to do it with a special instruction. */ - temp = expand_binop (mode, copysign_optab, op0, op1, - target, 0, OPTAB_DIRECT); - if (temp) - return temp; - - fmt = REAL_MODE_FORMAT (mode); - if (fmt == NULL || !fmt->has_signed_zero) - return NULL_RTX; - - op0_is_abs = false; - if (CONST_DOUBLE_AS_FLOAT_P (op0)) - { - if (real_isneg (CONST_DOUBLE_REAL_VALUE (op0))) - op0 = simplify_unary_operation (ABS, mode, op0, mode); - op0_is_abs = true; - } - - if (fmt->signbit_ro >= 0 - && (CONST_DOUBLE_AS_FLOAT_P (op0) - || (optab_handler (neg_optab, mode) != CODE_FOR_nothing - && optab_handler (abs_optab, mode) != CODE_FOR_nothing))) - { - temp = expand_copysign_absneg (mode, op0, op1, target, - fmt->signbit_ro, op0_is_abs); - if (temp) - return temp; - } - - if (fmt->signbit_rw < 0) - return NULL_RTX; - return expand_copysign_bit (mode, op0, op1, target, - fmt->signbit_rw, op0_is_abs); -} - -/* Generate an instruction whose insn-code is INSN_CODE, - with two operands: an output TARGET and an input OP0. - TARGET *must* be nonzero, and the output is always stored there. - CODE is an rtx code such that (CODE OP0) is an rtx that describes - the value that is stored into TARGET. - - Return false if expansion failed. */ - -bool -maybe_emit_unop_insn (enum insn_code icode, rtx target, rtx op0, - enum rtx_code code) -{ - class expand_operand ops[2]; - rtx_insn *pat; - - create_output_operand (&ops[0], target, GET_MODE (target)); - create_input_operand (&ops[1], op0, GET_MODE (op0)); - pat = maybe_gen_insn (icode, 2, ops); - if (!pat) - return false; - - if (INSN_P (pat) && NEXT_INSN (pat) != NULL_RTX - && code != UNKNOWN) - add_equal_note (pat, ops[0].value, code, ops[1].value, NULL_RTX, - GET_MODE (op0)); - - emit_insn (pat); - - if (ops[0].value != target) - emit_move_insn (target, ops[0].value); - return true; -} -/* Generate an instruction whose insn-code is INSN_CODE, - with two operands: an output TARGET and an input OP0. - TARGET *must* be nonzero, and the output is always stored there. - CODE is an rtx code such that (CODE OP0) is an rtx that describes - the value that is stored into TARGET. */ - -void -emit_unop_insn (enum insn_code icode, rtx target, rtx op0, enum rtx_code code) -{ - bool ok = maybe_emit_unop_insn (icode, target, op0, code); - gcc_assert (ok); -} - -struct no_conflict_data -{ - rtx target; - rtx_insn *first, *insn; - bool must_stay; -}; - -/* Called via note_stores by emit_libcall_block. Set P->must_stay if - the currently examined clobber / store has to stay in the list of - insns that constitute the actual libcall block. */ -static void -no_conflict_move_test (rtx dest, const_rtx set, void *p0) -{ - struct no_conflict_data *p= (struct no_conflict_data *) p0; - - /* If this inns directly contributes to setting the target, it must stay. */ - if (reg_overlap_mentioned_p (p->target, dest)) - p->must_stay = true; - /* If we haven't committed to keeping any other insns in the list yet, - there is nothing more to check. */ - else if (p->insn == p->first) - return; - /* If this insn sets / clobbers a register that feeds one of the insns - already in the list, this insn has to stay too. */ - else if (reg_overlap_mentioned_p (dest, PATTERN (p->first)) - || (CALL_P (p->first) && (find_reg_fusage (p->first, USE, dest))) - || reg_used_between_p (dest, p->first, p->insn) - /* Likewise if this insn depends on a register set by a previous - insn in the list, or if it sets a result (presumably a hard - register) that is set or clobbered by a previous insn. - N.B. the modified_*_p (SET_DEST...) tests applied to a MEM - SET_DEST perform the former check on the address, and the latter - check on the MEM. */ - || (GET_CODE (set) == SET - && (modified_in_p (SET_SRC (set), p->first) - || modified_in_p (SET_DEST (set), p->first) - || modified_between_p (SET_SRC (set), p->first, p->insn) - || modified_between_p (SET_DEST (set), p->first, p->insn)))) - p->must_stay = true; -} - - -/* Emit code to make a call to a constant function or a library call. - - INSNS is a list containing all insns emitted in the call. - These insns leave the result in RESULT. Our block is to copy RESULT - to TARGET, which is logically equivalent to EQUIV. - - We first emit any insns that set a pseudo on the assumption that these are - loading constants into registers; doing so allows them to be safely cse'ed - between blocks. Then we emit all the other insns in the block, followed by - an insn to move RESULT to TARGET. This last insn will have a REQ_EQUAL - note with an operand of EQUIV. */ - -static void -emit_libcall_block_1 (rtx_insn *insns, rtx target, rtx result, rtx equiv, - bool equiv_may_trap) -{ - rtx final_dest = target; - rtx_insn *next, *last, *insn; - - /* If this is a reg with REG_USERVAR_P set, then it could possibly turn - into a MEM later. Protect the libcall block from this change. */ - if (! REG_P (target) || REG_USERVAR_P (target)) - target = gen_reg_rtx (GET_MODE (target)); - - /* If we're using non-call exceptions, a libcall corresponding to an - operation that may trap may also trap. */ - /* ??? See the comment in front of make_reg_eh_region_note. */ - if (cfun->can_throw_non_call_exceptions - && (equiv_may_trap || may_trap_p (equiv))) - { - for (insn = insns; insn; insn = NEXT_INSN (insn)) - if (CALL_P (insn)) - { - rtx note = find_reg_note (insn, REG_EH_REGION, NULL_RTX); - if (note) - { - int lp_nr = INTVAL (XEXP (note, 0)); - if (lp_nr == 0 || lp_nr == INT_MIN) - remove_note (insn, note); - } - } - } - else - { - /* Look for any CALL_INSNs in this sequence, and attach a REG_EH_REGION - reg note to indicate that this call cannot throw or execute a nonlocal - goto (unless there is already a REG_EH_REGION note, in which case - we update it). */ - for (insn = insns; insn; insn = NEXT_INSN (insn)) - if (CALL_P (insn)) - make_reg_eh_region_note_nothrow_nononlocal (insn); - } - - /* First emit all insns that set pseudos. Remove them from the list as - we go. Avoid insns that set pseudos which were referenced in previous - insns. These can be generated by move_by_pieces, for example, - to update an address. Similarly, avoid insns that reference things - set in previous insns. */ - - for (insn = insns; insn; insn = next) - { - rtx set = single_set (insn); - - next = NEXT_INSN (insn); - - if (set != 0 && REG_P (SET_DEST (set)) - && REGNO (SET_DEST (set)) >= FIRST_PSEUDO_REGISTER) - { - struct no_conflict_data data; - - data.target = const0_rtx; - data.first = insns; - data.insn = insn; - data.must_stay = 0; - note_stores (insn, no_conflict_move_test, &data); - if (! data.must_stay) - { - if (PREV_INSN (insn)) - SET_NEXT_INSN (PREV_INSN (insn)) = next; - else - insns = next; - - if (next) - SET_PREV_INSN (next) = PREV_INSN (insn); - - add_insn (insn); - } - } - - /* Some ports use a loop to copy large arguments onto the stack. - Don't move anything outside such a loop. */ - if (LABEL_P (insn)) - break; - } - - /* Write the remaining insns followed by the final copy. */ - for (insn = insns; insn; insn = next) - { - next = NEXT_INSN (insn); - - add_insn (insn); - } - - last = emit_move_insn (target, result); - if (equiv) - set_dst_reg_note (last, REG_EQUAL, copy_rtx (equiv), target); - - if (final_dest != target) - emit_move_insn (final_dest, target); -} - -void -emit_libcall_block (rtx_insn *insns, rtx target, rtx result, rtx equiv) -{ - emit_libcall_block_1 (insns, target, result, equiv, false); -} - -/* Nonzero if we can perform a comparison of mode MODE straightforwardly. - PURPOSE describes how this comparison will be used. CODE is the rtx - comparison code we will be using. - - ??? Actually, CODE is slightly weaker than that. A target is still - required to implement all of the normal bcc operations, but not - required to implement all (or any) of the unordered bcc operations. */ - -int -can_compare_p (enum rtx_code code, machine_mode mode, - enum can_compare_purpose purpose) -{ - rtx test; - test = gen_rtx_fmt_ee (code, mode, const0_rtx, const0_rtx); - do - { - enum insn_code icode; - - if (purpose == ccp_jump - && (icode = optab_handler (cbranch_optab, mode)) != CODE_FOR_nothing - && insn_operand_matches (icode, 0, test)) - return 1; - if (purpose == ccp_store_flag - && (icode = optab_handler (cstore_optab, mode)) != CODE_FOR_nothing - && insn_operand_matches (icode, 1, test)) - return 1; - if (purpose == ccp_cmov - && optab_handler (cmov_optab, mode) != CODE_FOR_nothing) - return 1; - - mode = GET_MODE_WIDER_MODE (mode).else_void (); - PUT_MODE (test, mode); - } - while (mode != VOIDmode); - - return 0; -} - -/* Return whether RTL code CODE corresponds to an unsigned optab. */ - -static bool -unsigned_optab_p (enum rtx_code code) -{ - return code == LTU || code == LEU || code == GTU || code == GEU; -} - -/* Return whether the backend-emitted comparison for code CODE, comparing - operands of mode VALUE_MODE and producing a result with MASK_MODE, matches - operand OPNO of pattern ICODE. */ - -static bool -insn_predicate_matches_p (enum insn_code icode, unsigned int opno, - enum rtx_code code, machine_mode mask_mode, - machine_mode value_mode) -{ - rtx reg1 = alloca_raw_REG (value_mode, LAST_VIRTUAL_REGISTER + 1); - rtx reg2 = alloca_raw_REG (value_mode, LAST_VIRTUAL_REGISTER + 2); - rtx test = alloca_rtx_fmt_ee (code, mask_mode, reg1, reg2); - return insn_operand_matches (icode, opno, test); -} - -/* Return whether the backend can emit a vector comparison (vec_cmp/vec_cmpu) - for code CODE, comparing operands of mode VALUE_MODE and producing a result - with MASK_MODE. */ - -bool -can_vec_cmp_compare_p (enum rtx_code code, machine_mode value_mode, - machine_mode mask_mode) -{ - enum insn_code icode - = get_vec_cmp_icode (value_mode, mask_mode, unsigned_optab_p (code)); - if (icode == CODE_FOR_nothing) - return false; - - return insn_predicate_matches_p (icode, 1, code, mask_mode, value_mode); -} - -/* Return whether the backend can emit a vector comparison (vcond/vcondu) for - code CODE, comparing operands of mode CMP_OP_MODE and producing a result - with VALUE_MODE. */ - -bool -can_vcond_compare_p (enum rtx_code code, machine_mode value_mode, - machine_mode cmp_op_mode) -{ - enum insn_code icode - = get_vcond_icode (value_mode, cmp_op_mode, unsigned_optab_p (code)); - if (icode == CODE_FOR_nothing) - return false; - - return insn_predicate_matches_p (icode, 3, code, value_mode, cmp_op_mode); -} - -/* Return whether the backend can emit vector set instructions for inserting - element into vector at variable index position. */ - -bool -can_vec_set_var_idx_p (machine_mode vec_mode) -{ - if (!VECTOR_MODE_P (vec_mode)) - return false; - - machine_mode inner_mode = GET_MODE_INNER (vec_mode); - rtx reg1 = alloca_raw_REG (vec_mode, LAST_VIRTUAL_REGISTER + 1); - rtx reg2 = alloca_raw_REG (inner_mode, LAST_VIRTUAL_REGISTER + 2); - rtx reg3 = alloca_raw_REG (VOIDmode, LAST_VIRTUAL_REGISTER + 3); - - enum insn_code icode = optab_handler (vec_set_optab, vec_mode); - - return icode != CODE_FOR_nothing && insn_operand_matches (icode, 0, reg1) - && insn_operand_matches (icode, 1, reg2) - && insn_operand_matches (icode, 2, reg3); -} - -/* This function is called when we are going to emit a compare instruction that - compares the values found in X and Y, using the rtl operator COMPARISON. - - If they have mode BLKmode, then SIZE specifies the size of both operands. - - UNSIGNEDP nonzero says that the operands are unsigned; - this matters if they need to be widened (as given by METHODS). - - *PTEST is where the resulting comparison RTX is returned or NULL_RTX - if we failed to produce one. - - *PMODE is the mode of the inputs (in case they are const_int). - - This function performs all the setup necessary so that the caller only has - to emit a single comparison insn. This setup can involve doing a BLKmode - comparison or emitting a library call to perform the comparison if no insn - is available to handle it. - The values which are passed in through pointers can be modified; the caller - should perform the comparison on the modified values. Constant - comparisons must have already been folded. */ - -static void -prepare_cmp_insn (rtx x, rtx y, enum rtx_code comparison, rtx size, - int unsignedp, enum optab_methods methods, - rtx *ptest, machine_mode *pmode) -{ - machine_mode mode = *pmode; - rtx libfunc, test; - machine_mode cmp_mode; - enum mode_class mclass; - - /* The other methods are not needed. */ - gcc_assert (methods == OPTAB_DIRECT || methods == OPTAB_WIDEN - || methods == OPTAB_LIB_WIDEN); - - if (CONST_SCALAR_INT_P (y)) - canonicalize_comparison (mode, &comparison, &y); - - /* If we are optimizing, force expensive constants into a register. */ - if (CONSTANT_P (x) && optimize - && (rtx_cost (x, mode, COMPARE, 0, optimize_insn_for_speed_p ()) - > COSTS_N_INSNS (1))) - x = force_reg (mode, x); - - if (CONSTANT_P (y) && optimize - && (rtx_cost (y, mode, COMPARE, 1, optimize_insn_for_speed_p ()) - > COSTS_N_INSNS (1))) - y = force_reg (mode, y); - - /* Don't let both operands fail to indicate the mode. */ - if (GET_MODE (x) == VOIDmode && GET_MODE (y) == VOIDmode) - x = force_reg (mode, x); - if (mode == VOIDmode) - mode = GET_MODE (x) != VOIDmode ? GET_MODE (x) : GET_MODE (y); - - /* Handle all BLKmode compares. */ - - if (mode == BLKmode) - { - machine_mode result_mode; - enum insn_code cmp_code; - rtx result; - rtx opalign - = GEN_INT (MIN (MEM_ALIGN (x), MEM_ALIGN (y)) / BITS_PER_UNIT); - - gcc_assert (size); - - /* Try to use a memory block compare insn - either cmpstr - or cmpmem will do. */ - opt_scalar_int_mode cmp_mode_iter; - FOR_EACH_MODE_IN_CLASS (cmp_mode_iter, MODE_INT) - { - scalar_int_mode cmp_mode = cmp_mode_iter.require (); - cmp_code = direct_optab_handler (cmpmem_optab, cmp_mode); - if (cmp_code == CODE_FOR_nothing) - cmp_code = direct_optab_handler (cmpstr_optab, cmp_mode); - if (cmp_code == CODE_FOR_nothing) - cmp_code = direct_optab_handler (cmpstrn_optab, cmp_mode); - if (cmp_code == CODE_FOR_nothing) - continue; - - /* Must make sure the size fits the insn's mode. */ - if (CONST_INT_P (size) - ? UINTVAL (size) > GET_MODE_MASK (cmp_mode) - : (GET_MODE_BITSIZE (as_a <scalar_int_mode> (GET_MODE (size))) - > GET_MODE_BITSIZE (cmp_mode))) - continue; - - result_mode = insn_data[cmp_code].operand[0].mode; - result = gen_reg_rtx (result_mode); - size = convert_to_mode (cmp_mode, size, 1); - emit_insn (GEN_FCN (cmp_code) (result, x, y, size, opalign)); - - *ptest = gen_rtx_fmt_ee (comparison, VOIDmode, result, const0_rtx); - *pmode = result_mode; - return; - } - - if (methods != OPTAB_LIB && methods != OPTAB_LIB_WIDEN) - goto fail; - - /* Otherwise call a library function. */ - result = emit_block_comp_via_libcall (x, y, size); - - x = result; - y = const0_rtx; - mode = TYPE_MODE (integer_type_node); - methods = OPTAB_LIB_WIDEN; - unsignedp = false; - } - - /* Don't allow operands to the compare to trap, as that can put the - compare and branch in different basic blocks. */ - if (cfun->can_throw_non_call_exceptions) - { - if (may_trap_p (x)) - x = copy_to_reg (x); - if (may_trap_p (y)) - y = copy_to_reg (y); - } - - if (GET_MODE_CLASS (mode) == MODE_CC) - { - enum insn_code icode = optab_handler (cbranch_optab, CCmode); - test = gen_rtx_fmt_ee (comparison, VOIDmode, x, y); - gcc_assert (icode != CODE_FOR_nothing - && insn_operand_matches (icode, 0, test)); - *ptest = test; - return; - } - - mclass = GET_MODE_CLASS (mode); - test = gen_rtx_fmt_ee (comparison, VOIDmode, x, y); - FOR_EACH_MODE_FROM (cmp_mode, mode) - { - enum insn_code icode; - icode = optab_handler (cbranch_optab, cmp_mode); - if (icode != CODE_FOR_nothing - && insn_operand_matches (icode, 0, test)) - { - rtx_insn *last = get_last_insn (); - rtx op0 = prepare_operand (icode, x, 1, mode, cmp_mode, unsignedp); - rtx op1 = prepare_operand (icode, y, 2, mode, cmp_mode, unsignedp); - if (op0 && op1 - && insn_operand_matches (icode, 1, op0) - && insn_operand_matches (icode, 2, op1)) - { - XEXP (test, 0) = op0; - XEXP (test, 1) = op1; - *ptest = test; - *pmode = cmp_mode; - return; - } - delete_insns_since (last); - } - - if (methods == OPTAB_DIRECT || !CLASS_HAS_WIDER_MODES_P (mclass)) - break; - } - - if (methods != OPTAB_LIB_WIDEN) - goto fail; - - if (SCALAR_FLOAT_MODE_P (mode)) - { - /* Small trick if UNORDERED isn't implemented by the hardware. */ - if (comparison == UNORDERED && rtx_equal_p (x, y)) - { - prepare_cmp_insn (x, y, UNLT, NULL_RTX, unsignedp, OPTAB_WIDEN, - ptest, pmode); - if (*ptest) - return; - } - - prepare_float_lib_cmp (x, y, comparison, ptest, pmode); - } - else - { - rtx result; - machine_mode ret_mode; - - /* Handle a libcall just for the mode we are using. */ - libfunc = optab_libfunc (cmp_optab, mode); - gcc_assert (libfunc); - - /* If we want unsigned, and this mode has a distinct unsigned - comparison routine, use that. */ - if (unsignedp) - { - rtx ulibfunc = optab_libfunc (ucmp_optab, mode); - if (ulibfunc) - libfunc = ulibfunc; - } - - ret_mode = targetm.libgcc_cmp_return_mode (); - result = emit_library_call_value (libfunc, NULL_RTX, LCT_CONST, - ret_mode, x, mode, y, mode); - - /* There are two kinds of comparison routines. Biased routines - return 0/1/2, and unbiased routines return -1/0/1. Other parts - of gcc expect that the comparison operation is equivalent - to the modified comparison. For signed comparisons compare the - result against 1 in the biased case, and zero in the unbiased - case. For unsigned comparisons always compare against 1 after - biasing the unbiased result by adding 1. This gives us a way to - represent LTU. - The comparisons in the fixed-point helper library are always - biased. */ - x = result; - y = const1_rtx; - - if (!TARGET_LIB_INT_CMP_BIASED && !ALL_FIXED_POINT_MODE_P (mode)) - { - if (unsignedp) - x = plus_constant (ret_mode, result, 1); - else - y = const0_rtx; - } - - *pmode = ret_mode; - prepare_cmp_insn (x, y, comparison, NULL_RTX, unsignedp, methods, - ptest, pmode); - } - - return; - - fail: - *ptest = NULL_RTX; -} - -/* Before emitting an insn with code ICODE, make sure that X, which is going - to be used for operand OPNUM of the insn, is converted from mode MODE to - WIDER_MODE (UNSIGNEDP determines whether it is an unsigned conversion), and - that it is accepted by the operand predicate. Return the new value. */ - -rtx -prepare_operand (enum insn_code icode, rtx x, int opnum, machine_mode mode, - machine_mode wider_mode, int unsignedp) -{ - if (mode != wider_mode) - x = convert_modes (wider_mode, mode, x, unsignedp); - - if (!insn_operand_matches (icode, opnum, x)) - { - machine_mode op_mode = insn_data[(int) icode].operand[opnum].mode; - if (reload_completed) - return NULL_RTX; - if (GET_MODE (x) != op_mode && GET_MODE (x) != VOIDmode) - return NULL_RTX; - x = copy_to_mode_reg (op_mode, x); - } - - return x; -} - -/* Subroutine of emit_cmp_and_jump_insns; this function is called when we know - we can do the branch. */ - -static void -emit_cmp_and_jump_insn_1 (rtx test, machine_mode mode, rtx label, - profile_probability prob) -{ - machine_mode optab_mode; - enum mode_class mclass; - enum insn_code icode; - rtx_insn *insn; - - mclass = GET_MODE_CLASS (mode); - optab_mode = (mclass == MODE_CC) ? CCmode : mode; - icode = optab_handler (cbranch_optab, optab_mode); - - gcc_assert (icode != CODE_FOR_nothing); - gcc_assert (insn_operand_matches (icode, 0, test)); - insn = emit_jump_insn (GEN_FCN (icode) (test, XEXP (test, 0), - XEXP (test, 1), label)); - if (prob.initialized_p () - && profile_status_for_fn (cfun) != PROFILE_ABSENT - && insn - && JUMP_P (insn) - && any_condjump_p (insn) - && !find_reg_note (insn, REG_BR_PROB, 0)) - add_reg_br_prob_note (insn, prob); -} - -/* Generate code to compare X with Y so that the condition codes are - set and to jump to LABEL if the condition is true. If X is a - constant and Y is not a constant, then the comparison is swapped to - ensure that the comparison RTL has the canonical form. - - UNSIGNEDP nonzero says that X and Y are unsigned; this matters if they - need to be widened. UNSIGNEDP is also used to select the proper - branch condition code. - - If X and Y have mode BLKmode, then SIZE specifies the size of both X and Y. - - MODE is the mode of the inputs (in case they are const_int). - - COMPARISON is the rtl operator to compare with (EQ, NE, GT, etc.). - It will be potentially converted into an unsigned variant based on - UNSIGNEDP to select a proper jump instruction. - - PROB is the probability of jumping to LABEL. */ - -void -emit_cmp_and_jump_insns (rtx x, rtx y, enum rtx_code comparison, rtx size, - machine_mode mode, int unsignedp, rtx label, - profile_probability prob) -{ - rtx op0 = x, op1 = y; - rtx test; - - /* Swap operands and condition to ensure canonical RTL. */ - if (swap_commutative_operands_p (x, y) - && can_compare_p (swap_condition (comparison), mode, ccp_jump)) - { - op0 = y, op1 = x; - comparison = swap_condition (comparison); - } - - /* If OP0 is still a constant, then both X and Y must be constants - or the opposite comparison is not supported. Force X into a register - to create canonical RTL. */ - if (CONSTANT_P (op0)) - op0 = force_reg (mode, op0); - - if (unsignedp) - comparison = unsigned_condition (comparison); - - prepare_cmp_insn (op0, op1, comparison, size, unsignedp, OPTAB_LIB_WIDEN, - &test, &mode); - emit_cmp_and_jump_insn_1 (test, mode, label, prob); -} - - -/* Emit a library call comparison between floating point X and Y. - COMPARISON is the rtl operator to compare with (EQ, NE, GT, etc.). */ - -static void -prepare_float_lib_cmp (rtx x, rtx y, enum rtx_code comparison, - rtx *ptest, machine_mode *pmode) -{ - enum rtx_code swapped = swap_condition (comparison); - enum rtx_code reversed = reverse_condition_maybe_unordered (comparison); - machine_mode orig_mode = GET_MODE (x); - machine_mode mode; - rtx true_rtx, false_rtx; - rtx value, target, equiv; - rtx_insn *insns; - rtx libfunc = 0; - bool reversed_p = false; - scalar_int_mode cmp_mode = targetm.libgcc_cmp_return_mode (); - - FOR_EACH_MODE_FROM (mode, orig_mode) - { - if (code_to_optab (comparison) - && (libfunc = optab_libfunc (code_to_optab (comparison), mode))) - break; - - if (code_to_optab (swapped) - && (libfunc = optab_libfunc (code_to_optab (swapped), mode))) - { - std::swap (x, y); - comparison = swapped; - break; - } - - if (code_to_optab (reversed) - && (libfunc = optab_libfunc (code_to_optab (reversed), mode))) - { - comparison = reversed; - reversed_p = true; - break; - } - } - - gcc_assert (mode != VOIDmode); - - if (mode != orig_mode) - { - x = convert_to_mode (mode, x, 0); - y = convert_to_mode (mode, y, 0); - } - - /* Attach a REG_EQUAL note describing the semantics of the libcall to - the RTL. The allows the RTL optimizers to delete the libcall if the - condition can be determined at compile-time. */ - if (comparison == UNORDERED - || FLOAT_LIB_COMPARE_RETURNS_BOOL (mode, comparison)) - { - true_rtx = const_true_rtx; - false_rtx = const0_rtx; - } - else - { - switch (comparison) - { - case EQ: - true_rtx = const0_rtx; - false_rtx = const_true_rtx; - break; - - case NE: - true_rtx = const_true_rtx; - false_rtx = const0_rtx; - break; - - case GT: - true_rtx = const1_rtx; - false_rtx = const0_rtx; - break; - - case GE: - true_rtx = const0_rtx; - false_rtx = constm1_rtx; - break; - - case LT: - true_rtx = constm1_rtx; - false_rtx = const0_rtx; - break; - - case LE: - true_rtx = const0_rtx; - false_rtx = const1_rtx; - break; - - default: - gcc_unreachable (); - } - } - - if (comparison == UNORDERED) - { - rtx temp = simplify_gen_relational (NE, cmp_mode, mode, x, x); - equiv = simplify_gen_relational (NE, cmp_mode, mode, y, y); - equiv = simplify_gen_ternary (IF_THEN_ELSE, cmp_mode, cmp_mode, - temp, const_true_rtx, equiv); - } - else - { - equiv = simplify_gen_relational (comparison, cmp_mode, mode, x, y); - if (! FLOAT_LIB_COMPARE_RETURNS_BOOL (mode, comparison)) - equiv = simplify_gen_ternary (IF_THEN_ELSE, cmp_mode, cmp_mode, - equiv, true_rtx, false_rtx); - } - - start_sequence (); - value = emit_library_call_value (libfunc, NULL_RTX, LCT_CONST, - cmp_mode, x, mode, y, mode); - insns = get_insns (); - end_sequence (); - - target = gen_reg_rtx (cmp_mode); - emit_libcall_block (insns, target, value, equiv); - - if (comparison == UNORDERED - || FLOAT_LIB_COMPARE_RETURNS_BOOL (mode, comparison) - || reversed_p) - *ptest = gen_rtx_fmt_ee (reversed_p ? EQ : NE, VOIDmode, target, false_rtx); - else - *ptest = gen_rtx_fmt_ee (comparison, VOIDmode, target, const0_rtx); - - *pmode = cmp_mode; -} - -/* Generate code to indirectly jump to a location given in the rtx LOC. */ - -void -emit_indirect_jump (rtx loc) -{ - if (!targetm.have_indirect_jump ()) - sorry ("indirect jumps are not available on this target"); - else - { - class expand_operand ops[1]; - create_address_operand (&ops[0], loc); - expand_jump_insn (targetm.code_for_indirect_jump, 1, ops); - emit_barrier (); - } -} - - -/* Emit a conditional move instruction if the machine supports one for that - condition and machine mode. - - OP0 and OP1 are the operands that should be compared using CODE. CMODE is - the mode to use should they be constants. If it is VOIDmode, they cannot - both be constants. - - OP2 should be stored in TARGET if the comparison is true, otherwise OP3 - should be stored there. MODE is the mode to use should they be constants. - If it is VOIDmode, they cannot both be constants. - - The result is either TARGET (perhaps modified) or NULL_RTX if the operation - is not supported. */ - -rtx -emit_conditional_move (rtx target, enum rtx_code code, rtx op0, rtx op1, - machine_mode cmode, rtx op2, rtx op3, - machine_mode mode, int unsignedp) -{ - rtx comparison; - rtx_insn *last; - enum insn_code icode; - enum rtx_code reversed; - - /* If the two source operands are identical, that's just a move. */ - - if (rtx_equal_p (op2, op3)) - { - if (!target) - target = gen_reg_rtx (mode); - - emit_move_insn (target, op3); - return target; - } - - /* If one operand is constant, make it the second one. Only do this - if the other operand is not constant as well. */ - - if (swap_commutative_operands_p (op0, op1)) - { - std::swap (op0, op1); - code = swap_condition (code); - } - - /* get_condition will prefer to generate LT and GT even if the old - comparison was against zero, so undo that canonicalization here since - comparisons against zero are cheaper. */ - if (code == LT && op1 == const1_rtx) - code = LE, op1 = const0_rtx; - else if (code == GT && op1 == constm1_rtx) - code = GE, op1 = const0_rtx; - - if (cmode == VOIDmode) - cmode = GET_MODE (op0); - - enum rtx_code orig_code = code; - bool swapped = false; - if (swap_commutative_operands_p (op2, op3) - && ((reversed = reversed_comparison_code_parts (code, op0, op1, NULL)) - != UNKNOWN)) - { - std::swap (op2, op3); - code = reversed; - swapped = true; - } - - if (mode == VOIDmode) - mode = GET_MODE (op2); - - icode = direct_optab_handler (movcc_optab, mode); - - if (icode == CODE_FOR_nothing) - return NULL_RTX; - - if (!target) - target = gen_reg_rtx (mode); - - for (int pass = 0; ; pass++) - { - code = unsignedp ? unsigned_condition (code) : code; - comparison = simplify_gen_relational (code, VOIDmode, cmode, op0, op1); - - /* We can get const0_rtx or const_true_rtx in some circumstances. Just - punt and let the caller figure out how best to deal with this - situation. */ - if (COMPARISON_P (comparison)) - { - saved_pending_stack_adjust save; - save_pending_stack_adjust (&save); - last = get_last_insn (); - do_pending_stack_adjust (); - machine_mode cmpmode = cmode; - prepare_cmp_insn (XEXP (comparison, 0), XEXP (comparison, 1), - GET_CODE (comparison), NULL_RTX, unsignedp, - OPTAB_WIDEN, &comparison, &cmpmode); - if (comparison) - { - class expand_operand ops[4]; - - create_output_operand (&ops[0], target, mode); - create_fixed_operand (&ops[1], comparison); - create_input_operand (&ops[2], op2, mode); - create_input_operand (&ops[3], op3, mode); - if (maybe_expand_insn (icode, 4, ops)) - { - if (ops[0].value != target) - convert_move (target, ops[0].value, false); - return target; - } - } - delete_insns_since (last); - restore_pending_stack_adjust (&save); - } - - if (pass == 1) - return NULL_RTX; - - /* If the preferred op2/op3 order is not usable, retry with other - operand order, perhaps it will expand successfully. */ - if (swapped) - code = orig_code; - else if ((reversed = reversed_comparison_code_parts (orig_code, op0, op1, - NULL)) - != UNKNOWN) - code = reversed; - else - return NULL_RTX; - std::swap (op2, op3); - } -} - - -/* Emit a conditional negate or bitwise complement using the - negcc or notcc optabs if available. Return NULL_RTX if such operations - are not available. Otherwise return the RTX holding the result. - TARGET is the desired destination of the result. COMP is the comparison - on which to negate. If COND is true move into TARGET the negation - or bitwise complement of OP1. Otherwise move OP2 into TARGET. - CODE is either NEG or NOT. MODE is the machine mode in which the - operation is performed. */ - -rtx -emit_conditional_neg_or_complement (rtx target, rtx_code code, - machine_mode mode, rtx cond, rtx op1, - rtx op2) -{ - optab op = unknown_optab; - if (code == NEG) - op = negcc_optab; - else if (code == NOT) - op = notcc_optab; - else - gcc_unreachable (); - - insn_code icode = direct_optab_handler (op, mode); - - if (icode == CODE_FOR_nothing) - return NULL_RTX; - - if (!target) - target = gen_reg_rtx (mode); - - rtx_insn *last = get_last_insn (); - class expand_operand ops[4]; - - create_output_operand (&ops[0], target, mode); - create_fixed_operand (&ops[1], cond); - create_input_operand (&ops[2], op1, mode); - create_input_operand (&ops[3], op2, mode); - - if (maybe_expand_insn (icode, 4, ops)) - { - if (ops[0].value != target) - convert_move (target, ops[0].value, false); - - return target; - } - delete_insns_since (last); - return NULL_RTX; -} - -/* Emit a conditional addition instruction if the machine supports one for that - condition and machine mode. - - OP0 and OP1 are the operands that should be compared using CODE. CMODE is - the mode to use should they be constants. If it is VOIDmode, they cannot - both be constants. - - OP2 should be stored in TARGET if the comparison is false, otherwise OP2+OP3 - should be stored there. MODE is the mode to use should they be constants. - If it is VOIDmode, they cannot both be constants. - - The result is either TARGET (perhaps modified) or NULL_RTX if the operation - is not supported. */ - -rtx -emit_conditional_add (rtx target, enum rtx_code code, rtx op0, rtx op1, - machine_mode cmode, rtx op2, rtx op3, - machine_mode mode, int unsignedp) -{ - rtx comparison; - rtx_insn *last; - enum insn_code icode; - - /* If one operand is constant, make it the second one. Only do this - if the other operand is not constant as well. */ - - if (swap_commutative_operands_p (op0, op1)) - { - std::swap (op0, op1); - code = swap_condition (code); - } - - /* get_condition will prefer to generate LT and GT even if the old - comparison was against zero, so undo that canonicalization here since - comparisons against zero are cheaper. */ - if (code == LT && op1 == const1_rtx) - code = LE, op1 = const0_rtx; - else if (code == GT && op1 == constm1_rtx) - code = GE, op1 = const0_rtx; - - if (cmode == VOIDmode) - cmode = GET_MODE (op0); - - if (mode == VOIDmode) - mode = GET_MODE (op2); - - icode = optab_handler (addcc_optab, mode); - - if (icode == CODE_FOR_nothing) - return 0; - - if (!target) - target = gen_reg_rtx (mode); - - code = unsignedp ? unsigned_condition (code) : code; - comparison = simplify_gen_relational (code, VOIDmode, cmode, op0, op1); - - /* We can get const0_rtx or const_true_rtx in some circumstances. Just - return NULL and let the caller figure out how best to deal with this - situation. */ - if (!COMPARISON_P (comparison)) - return NULL_RTX; - - do_pending_stack_adjust (); - last = get_last_insn (); - prepare_cmp_insn (XEXP (comparison, 0), XEXP (comparison, 1), - GET_CODE (comparison), NULL_RTX, unsignedp, OPTAB_WIDEN, - &comparison, &cmode); - if (comparison) - { - class expand_operand ops[4]; - - create_output_operand (&ops[0], target, mode); - create_fixed_operand (&ops[1], comparison); - create_input_operand (&ops[2], op2, mode); - create_input_operand (&ops[3], op3, mode); - if (maybe_expand_insn (icode, 4, ops)) - { - if (ops[0].value != target) - convert_move (target, ops[0].value, false); - return target; - } - } - delete_insns_since (last); - return NULL_RTX; -} - -/* These functions attempt to generate an insn body, rather than - emitting the insn, but if the gen function already emits them, we - make no attempt to turn them back into naked patterns. */ - -/* Generate and return an insn body to add Y to X. */ - -rtx_insn * -gen_add2_insn (rtx x, rtx y) -{ - enum insn_code icode = optab_handler (add_optab, GET_MODE (x)); - - gcc_assert (insn_operand_matches (icode, 0, x)); - gcc_assert (insn_operand_matches (icode, 1, x)); - gcc_assert (insn_operand_matches (icode, 2, y)); - - return GEN_FCN (icode) (x, x, y); -} - -/* Generate and return an insn body to add r1 and c, - storing the result in r0. */ - -rtx_insn * -gen_add3_insn (rtx r0, rtx r1, rtx c) -{ - enum insn_code icode = optab_handler (add_optab, GET_MODE (r0)); - - if (icode == CODE_FOR_nothing - || !insn_operand_matches (icode, 0, r0) - || !insn_operand_matches (icode, 1, r1) - || !insn_operand_matches (icode, 2, c)) - return NULL; - - return GEN_FCN (icode) (r0, r1, c); -} - -int -have_add2_insn (rtx x, rtx y) -{ - enum insn_code icode; - - gcc_assert (GET_MODE (x) != VOIDmode); - - icode = optab_handler (add_optab, GET_MODE (x)); - - if (icode == CODE_FOR_nothing) - return 0; - - if (!insn_operand_matches (icode, 0, x) - || !insn_operand_matches (icode, 1, x) - || !insn_operand_matches (icode, 2, y)) - return 0; - - return 1; -} - -/* Generate and return an insn body to add Y to X. */ - -rtx_insn * -gen_addptr3_insn (rtx x, rtx y, rtx z) -{ - enum insn_code icode = optab_handler (addptr3_optab, GET_MODE (x)); - - gcc_assert (insn_operand_matches (icode, 0, x)); - gcc_assert (insn_operand_matches (icode, 1, y)); - gcc_assert (insn_operand_matches (icode, 2, z)); - - return GEN_FCN (icode) (x, y, z); -} - -/* Return true if the target implements an addptr pattern and X, Y, - and Z are valid for the pattern predicates. */ - -int -have_addptr3_insn (rtx x, rtx y, rtx z) -{ - enum insn_code icode; - - gcc_assert (GET_MODE (x) != VOIDmode); - - icode = optab_handler (addptr3_optab, GET_MODE (x)); - - if (icode == CODE_FOR_nothing) - return 0; - - if (!insn_operand_matches (icode, 0, x) - || !insn_operand_matches (icode, 1, y) - || !insn_operand_matches (icode, 2, z)) - return 0; - - return 1; -} - -/* Generate and return an insn body to subtract Y from X. */ - -rtx_insn * -gen_sub2_insn (rtx x, rtx y) -{ - enum insn_code icode = optab_handler (sub_optab, GET_MODE (x)); - - gcc_assert (insn_operand_matches (icode, 0, x)); - gcc_assert (insn_operand_matches (icode, 1, x)); - gcc_assert (insn_operand_matches (icode, 2, y)); - - return GEN_FCN (icode) (x, x, y); -} - -/* Generate and return an insn body to subtract r1 and c, - storing the result in r0. */ - -rtx_insn * -gen_sub3_insn (rtx r0, rtx r1, rtx c) -{ - enum insn_code icode = optab_handler (sub_optab, GET_MODE (r0)); - - if (icode == CODE_FOR_nothing - || !insn_operand_matches (icode, 0, r0) - || !insn_operand_matches (icode, 1, r1) - || !insn_operand_matches (icode, 2, c)) - return NULL; - - return GEN_FCN (icode) (r0, r1, c); -} - -int -have_sub2_insn (rtx x, rtx y) -{ - enum insn_code icode; - - gcc_assert (GET_MODE (x) != VOIDmode); - - icode = optab_handler (sub_optab, GET_MODE (x)); - - if (icode == CODE_FOR_nothing) - return 0; - - if (!insn_operand_matches (icode, 0, x) - || !insn_operand_matches (icode, 1, x) - || !insn_operand_matches (icode, 2, y)) - return 0; - - return 1; -} - -/* Generate the body of an insn to extend Y (with mode MFROM) - into X (with mode MTO). Do zero-extension if UNSIGNEDP is nonzero. */ - -rtx_insn * -gen_extend_insn (rtx x, rtx y, machine_mode mto, - machine_mode mfrom, int unsignedp) -{ - enum insn_code icode = can_extend_p (mto, mfrom, unsignedp); - return GEN_FCN (icode) (x, y); -} - -/* Generate code to convert FROM to floating point - and store in TO. FROM must be fixed point and not VOIDmode. - UNSIGNEDP nonzero means regard FROM as unsigned. - Normally this is done by correcting the final value - if it is negative. */ - -void -expand_float (rtx to, rtx from, int unsignedp) -{ - enum insn_code icode; - rtx target = to; - scalar_mode from_mode, to_mode; - machine_mode fmode, imode; - bool can_do_signed = false; - - /* Crash now, because we won't be able to decide which mode to use. */ - gcc_assert (GET_MODE (from) != VOIDmode); - - /* Look for an insn to do the conversion. Do it in the specified - modes if possible; otherwise convert either input, output or both to - wider mode. If the integer mode is wider than the mode of FROM, - we can do the conversion signed even if the input is unsigned. */ - - FOR_EACH_MODE_FROM (fmode, GET_MODE (to)) - FOR_EACH_MODE_FROM (imode, GET_MODE (from)) - { - int doing_unsigned = unsignedp; - - if (fmode != GET_MODE (to) - && (significand_size (fmode) - < GET_MODE_UNIT_PRECISION (GET_MODE (from)))) - continue; - - icode = can_float_p (fmode, imode, unsignedp); - if (icode == CODE_FOR_nothing && unsignedp) - { - enum insn_code scode = can_float_p (fmode, imode, 0); - if (scode != CODE_FOR_nothing) - can_do_signed = true; - if (imode != GET_MODE (from)) - icode = scode, doing_unsigned = 0; - } - - if (icode != CODE_FOR_nothing) - { - if (imode != GET_MODE (from)) - from = convert_to_mode (imode, from, unsignedp); - - if (fmode != GET_MODE (to)) - target = gen_reg_rtx (fmode); - - emit_unop_insn (icode, target, from, - doing_unsigned ? UNSIGNED_FLOAT : FLOAT); - - if (target != to) - convert_move (to, target, 0); - return; - } - } - - /* Unsigned integer, and no way to convert directly. Convert as signed, - then unconditionally adjust the result. */ - if (unsignedp - && can_do_signed - && is_a <scalar_mode> (GET_MODE (to), &to_mode) - && is_a <scalar_mode> (GET_MODE (from), &from_mode)) - { - opt_scalar_mode fmode_iter; - rtx_code_label *label = gen_label_rtx (); - rtx temp; - REAL_VALUE_TYPE offset; - - /* Look for a usable floating mode FMODE wider than the source and at - least as wide as the target. Using FMODE will avoid rounding woes - with unsigned values greater than the signed maximum value. */ - - FOR_EACH_MODE_FROM (fmode_iter, to_mode) - { - scalar_mode fmode = fmode_iter.require (); - if (GET_MODE_PRECISION (from_mode) < GET_MODE_BITSIZE (fmode) - && can_float_p (fmode, from_mode, 0) != CODE_FOR_nothing) - break; - } - - if (!fmode_iter.exists (&fmode)) - { - /* There is no such mode. Pretend the target is wide enough. */ - fmode = to_mode; - - /* Avoid double-rounding when TO is narrower than FROM. */ - if ((significand_size (fmode) + 1) - < GET_MODE_PRECISION (from_mode)) - { - rtx temp1; - rtx_code_label *neglabel = gen_label_rtx (); - - /* Don't use TARGET if it isn't a register, is a hard register, - or is the wrong mode. */ - if (!REG_P (target) - || REGNO (target) < FIRST_PSEUDO_REGISTER - || GET_MODE (target) != fmode) - target = gen_reg_rtx (fmode); - - imode = from_mode; - do_pending_stack_adjust (); - - /* Test whether the sign bit is set. */ - emit_cmp_and_jump_insns (from, const0_rtx, LT, NULL_RTX, imode, - 0, neglabel); - - /* The sign bit is not set. Convert as signed. */ - expand_float (target, from, 0); - emit_jump_insn (targetm.gen_jump (label)); - emit_barrier (); - - /* The sign bit is set. - Convert to a usable (positive signed) value by shifting right - one bit, while remembering if a nonzero bit was shifted - out; i.e., compute (from & 1) | (from >> 1). */ - - emit_label (neglabel); - temp = expand_binop (imode, and_optab, from, const1_rtx, - NULL_RTX, 1, OPTAB_LIB_WIDEN); - temp1 = expand_shift (RSHIFT_EXPR, imode, from, 1, NULL_RTX, 1); - temp = expand_binop (imode, ior_optab, temp, temp1, temp, 1, - OPTAB_LIB_WIDEN); - expand_float (target, temp, 0); - - /* Multiply by 2 to undo the shift above. */ - temp = expand_binop (fmode, add_optab, target, target, - target, 0, OPTAB_LIB_WIDEN); - if (temp != target) - emit_move_insn (target, temp); - - do_pending_stack_adjust (); - emit_label (label); - goto done; - } - } - - /* If we are about to do some arithmetic to correct for an - unsigned operand, do it in a pseudo-register. */ - - if (to_mode != fmode - || !REG_P (to) || REGNO (to) < FIRST_PSEUDO_REGISTER) - target = gen_reg_rtx (fmode); - - /* Convert as signed integer to floating. */ - expand_float (target, from, 0); - - /* If FROM is negative (and therefore TO is negative), - correct its value by 2**bitwidth. */ - - do_pending_stack_adjust (); - emit_cmp_and_jump_insns (from, const0_rtx, GE, NULL_RTX, from_mode, - 0, label); - - - real_2expN (&offset, GET_MODE_PRECISION (from_mode), fmode); - temp = expand_binop (fmode, add_optab, target, - const_double_from_real_value (offset, fmode), - target, 0, OPTAB_LIB_WIDEN); - if (temp != target) - emit_move_insn (target, temp); - - do_pending_stack_adjust (); - emit_label (label); - goto done; - } - - /* No hardware instruction available; call a library routine. */ - { - rtx libfunc; - rtx_insn *insns; - rtx value; - convert_optab tab = unsignedp ? ufloat_optab : sfloat_optab; - - if (is_narrower_int_mode (GET_MODE (from), SImode)) - from = convert_to_mode (SImode, from, unsignedp); - - libfunc = convert_optab_libfunc (tab, GET_MODE (to), GET_MODE (from)); - gcc_assert (libfunc); - - start_sequence (); - - value = emit_library_call_value (libfunc, NULL_RTX, LCT_CONST, - GET_MODE (to), from, GET_MODE (from)); - insns = get_insns (); - end_sequence (); - - emit_libcall_block (insns, target, value, - gen_rtx_fmt_e (unsignedp ? UNSIGNED_FLOAT : FLOAT, - GET_MODE (to), from)); - } - - done: - - /* Copy result to requested destination - if we have been computing in a temp location. */ - - if (target != to) - { - if (GET_MODE (target) == GET_MODE (to)) - emit_move_insn (to, target); - else - convert_move (to, target, 0); - } -} - -/* Generate code to convert FROM to fixed point and store in TO. FROM - must be floating point. */ - -void -expand_fix (rtx to, rtx from, int unsignedp) -{ - enum insn_code icode; - rtx target = to; - machine_mode fmode, imode; - opt_scalar_mode fmode_iter; - bool must_trunc = false; - - /* We first try to find a pair of modes, one real and one integer, at - least as wide as FROM and TO, respectively, in which we can open-code - this conversion. If the integer mode is wider than the mode of TO, - we can do the conversion either signed or unsigned. */ - - FOR_EACH_MODE_FROM (fmode, GET_MODE (from)) - FOR_EACH_MODE_FROM (imode, GET_MODE (to)) - { - int doing_unsigned = unsignedp; - - icode = can_fix_p (imode, fmode, unsignedp, &must_trunc); - if (icode == CODE_FOR_nothing && imode != GET_MODE (to) && unsignedp) - icode = can_fix_p (imode, fmode, 0, &must_trunc), doing_unsigned = 0; - - if (icode != CODE_FOR_nothing) - { - rtx_insn *last = get_last_insn (); - rtx from1 = from; - if (fmode != GET_MODE (from)) - from1 = convert_to_mode (fmode, from, 0); - - if (must_trunc) - { - rtx temp = gen_reg_rtx (GET_MODE (from1)); - from1 = expand_unop (GET_MODE (from1), ftrunc_optab, from1, - temp, 0); - } - - if (imode != GET_MODE (to)) - target = gen_reg_rtx (imode); - - if (maybe_emit_unop_insn (icode, target, from1, - doing_unsigned ? UNSIGNED_FIX : FIX)) - { - if (target != to) - convert_move (to, target, unsignedp); - return; - } - delete_insns_since (last); - } - } - - /* For an unsigned conversion, there is one more way to do it. - If we have a signed conversion, we generate code that compares - the real value to the largest representable positive number. If if - is smaller, the conversion is done normally. Otherwise, subtract - one plus the highest signed number, convert, and add it back. - - We only need to check all real modes, since we know we didn't find - anything with a wider integer mode. - - This code used to extend FP value into mode wider than the destination. - This is needed for decimal float modes which cannot accurately - represent one plus the highest signed number of the same size, but - not for binary modes. Consider, for instance conversion from SFmode - into DImode. - - The hot path through the code is dealing with inputs smaller than 2^63 - and doing just the conversion, so there is no bits to lose. - - In the other path we know the value is positive in the range 2^63..2^64-1 - inclusive. (as for other input overflow happens and result is undefined) - So we know that the most important bit set in mantissa corresponds to - 2^63. The subtraction of 2^63 should not generate any rounding as it - simply clears out that bit. The rest is trivial. */ - - scalar_int_mode to_mode; - if (unsignedp - && is_a <scalar_int_mode> (GET_MODE (to), &to_mode) - && HWI_COMPUTABLE_MODE_P (to_mode)) - FOR_EACH_MODE_FROM (fmode_iter, as_a <scalar_mode> (GET_MODE (from))) - { - scalar_mode fmode = fmode_iter.require (); - if (CODE_FOR_nothing != can_fix_p (to_mode, fmode, - 0, &must_trunc) - && (!DECIMAL_FLOAT_MODE_P (fmode) - || (GET_MODE_BITSIZE (fmode) > GET_MODE_PRECISION (to_mode)))) - { - int bitsize; - REAL_VALUE_TYPE offset; - rtx limit; - rtx_code_label *lab1, *lab2; - rtx_insn *insn; - - bitsize = GET_MODE_PRECISION (to_mode); - real_2expN (&offset, bitsize - 1, fmode); - limit = const_double_from_real_value (offset, fmode); - lab1 = gen_label_rtx (); - lab2 = gen_label_rtx (); - - if (fmode != GET_MODE (from)) - from = convert_to_mode (fmode, from, 0); - - /* See if we need to do the subtraction. */ - do_pending_stack_adjust (); - emit_cmp_and_jump_insns (from, limit, GE, NULL_RTX, - GET_MODE (from), 0, lab1); - - /* If not, do the signed "fix" and branch around fixup code. */ - expand_fix (to, from, 0); - emit_jump_insn (targetm.gen_jump (lab2)); - emit_barrier (); - - /* Otherwise, subtract 2**(N-1), convert to signed number, - then add 2**(N-1). Do the addition using XOR since this - will often generate better code. */ - emit_label (lab1); - target = expand_binop (GET_MODE (from), sub_optab, from, limit, - NULL_RTX, 0, OPTAB_LIB_WIDEN); - expand_fix (to, target, 0); - target = expand_binop (to_mode, xor_optab, to, - gen_int_mode - (HOST_WIDE_INT_1 << (bitsize - 1), - to_mode), - to, 1, OPTAB_LIB_WIDEN); - - if (target != to) - emit_move_insn (to, target); - - emit_label (lab2); - - if (optab_handler (mov_optab, to_mode) != CODE_FOR_nothing) - { - /* Make a place for a REG_NOTE and add it. */ - insn = emit_move_insn (to, to); - set_dst_reg_note (insn, REG_EQUAL, - gen_rtx_fmt_e (UNSIGNED_FIX, to_mode, - copy_rtx (from)), - to); - } - - return; - } - } - - /* We can't do it with an insn, so use a library call. But first ensure - that the mode of TO is at least as wide as SImode, since those are the - only library calls we know about. */ - - if (is_narrower_int_mode (GET_MODE (to), SImode)) - { - target = gen_reg_rtx (SImode); - - expand_fix (target, from, unsignedp); - } - else - { - rtx_insn *insns; - rtx value; - rtx libfunc; - - convert_optab tab = unsignedp ? ufix_optab : sfix_optab; - libfunc = convert_optab_libfunc (tab, GET_MODE (to), GET_MODE (from)); - gcc_assert (libfunc); - - start_sequence (); - - value = emit_library_call_value (libfunc, NULL_RTX, LCT_CONST, - GET_MODE (to), from, GET_MODE (from)); - insns = get_insns (); - end_sequence (); - - emit_libcall_block (insns, target, value, - gen_rtx_fmt_e (unsignedp ? UNSIGNED_FIX : FIX, - GET_MODE (to), from)); - } - - if (target != to) - { - if (GET_MODE (to) == GET_MODE (target)) - emit_move_insn (to, target); - else - convert_move (to, target, 0); - } -} - - -/* Promote integer arguments for a libcall if necessary. - emit_library_call_value cannot do the promotion because it does not - know if it should do a signed or unsigned promotion. This is because - there are no tree types defined for libcalls. */ - -static rtx -prepare_libcall_arg (rtx arg, int uintp) -{ - scalar_int_mode mode; - machine_mode arg_mode; - if (is_a <scalar_int_mode> (GET_MODE (arg), &mode)) - { - /* If we need to promote the integer function argument we need to do - it here instead of inside emit_library_call_value because in - emit_library_call_value we don't know if we should do a signed or - unsigned promotion. */ - - int unsigned_p = 0; - arg_mode = promote_function_mode (NULL_TREE, mode, - &unsigned_p, NULL_TREE, 0); - if (arg_mode != mode) - return convert_to_mode (arg_mode, arg, uintp); - } - return arg; -} - -/* Generate code to convert FROM or TO a fixed-point. - If UINTP is true, either TO or FROM is an unsigned integer. - If SATP is true, we need to saturate the result. */ - -void -expand_fixed_convert (rtx to, rtx from, int uintp, int satp) -{ - machine_mode to_mode = GET_MODE (to); - machine_mode from_mode = GET_MODE (from); - convert_optab tab; - enum rtx_code this_code; - enum insn_code code; - rtx_insn *insns; - rtx value; - rtx libfunc; - - if (to_mode == from_mode) - { - emit_move_insn (to, from); - return; - } - - if (uintp) - { - tab = satp ? satfractuns_optab : fractuns_optab; - this_code = satp ? UNSIGNED_SAT_FRACT : UNSIGNED_FRACT_CONVERT; - } - else - { - tab = satp ? satfract_optab : fract_optab; - this_code = satp ? SAT_FRACT : FRACT_CONVERT; - } - code = convert_optab_handler (tab, to_mode, from_mode); - if (code != CODE_FOR_nothing) - { - emit_unop_insn (code, to, from, this_code); - return; - } - - libfunc = convert_optab_libfunc (tab, to_mode, from_mode); - gcc_assert (libfunc); - - from = prepare_libcall_arg (from, uintp); - from_mode = GET_MODE (from); - - start_sequence (); - value = emit_library_call_value (libfunc, NULL_RTX, LCT_CONST, to_mode, - from, from_mode); - insns = get_insns (); - end_sequence (); - - emit_libcall_block (insns, to, value, - gen_rtx_fmt_e (optab_to_code (tab), to_mode, from)); -} - -/* Generate code to convert FROM to fixed point and store in TO. FROM - must be floating point, TO must be signed. Use the conversion optab - TAB to do the conversion. */ - -bool -expand_sfix_optab (rtx to, rtx from, convert_optab tab) -{ - enum insn_code icode; - rtx target = to; - machine_mode fmode, imode; - - /* We first try to find a pair of modes, one real and one integer, at - least as wide as FROM and TO, respectively, in which we can open-code - this conversion. If the integer mode is wider than the mode of TO, - we can do the conversion either signed or unsigned. */ - - FOR_EACH_MODE_FROM (fmode, GET_MODE (from)) - FOR_EACH_MODE_FROM (imode, GET_MODE (to)) - { - icode = convert_optab_handler (tab, imode, fmode); - if (icode != CODE_FOR_nothing) - { - rtx_insn *last = get_last_insn (); - if (fmode != GET_MODE (from)) - from = convert_to_mode (fmode, from, 0); - - if (imode != GET_MODE (to)) - target = gen_reg_rtx (imode); - - if (!maybe_emit_unop_insn (icode, target, from, UNKNOWN)) - { - delete_insns_since (last); - continue; - } - if (target != to) - convert_move (to, target, 0); - return true; - } - } - - return false; -} - -/* Report whether we have an instruction to perform the operation - specified by CODE on operands of mode MODE. */ -int -have_insn_for (enum rtx_code code, machine_mode mode) -{ - return (code_to_optab (code) - && (optab_handler (code_to_optab (code), mode) - != CODE_FOR_nothing)); -} - -/* Print information about the current contents of the optabs on - STDERR. */ - -DEBUG_FUNCTION void -debug_optab_libfuncs (void) -{ - int i, j, k; - - /* Dump the arithmetic optabs. */ - for (i = FIRST_NORM_OPTAB; i <= LAST_NORMLIB_OPTAB; ++i) - for (j = 0; j < NUM_MACHINE_MODES; ++j) - { - rtx l = optab_libfunc ((optab) i, (machine_mode) j); - if (l) - { - gcc_assert (GET_CODE (l) == SYMBOL_REF); - fprintf (stderr, "%s\t%s:\t%s\n", - GET_RTX_NAME (optab_to_code ((optab) i)), - GET_MODE_NAME (j), - XSTR (l, 0)); - } - } - - /* Dump the conversion optabs. */ - for (i = FIRST_CONV_OPTAB; i <= LAST_CONVLIB_OPTAB; ++i) - for (j = 0; j < NUM_MACHINE_MODES; ++j) - for (k = 0; k < NUM_MACHINE_MODES; ++k) - { - rtx l = convert_optab_libfunc ((optab) i, (machine_mode) j, - (machine_mode) k); - if (l) - { - gcc_assert (GET_CODE (l) == SYMBOL_REF); - fprintf (stderr, "%s\t%s\t%s:\t%s\n", - GET_RTX_NAME (optab_to_code ((optab) i)), - GET_MODE_NAME (j), - GET_MODE_NAME (k), - XSTR (l, 0)); - } - } -} - -/* Generate insns to trap with code TCODE if OP1 and OP2 satisfy condition - CODE. Return 0 on failure. */ - -rtx_insn * -gen_cond_trap (enum rtx_code code, rtx op1, rtx op2, rtx tcode) -{ - machine_mode mode = GET_MODE (op1); - enum insn_code icode; - rtx_insn *insn; - rtx trap_rtx; - - if (mode == VOIDmode) - return 0; - - icode = optab_handler (ctrap_optab, mode); - if (icode == CODE_FOR_nothing) - return 0; - - /* Some targets only accept a zero trap code. */ - if (!insn_operand_matches (icode, 3, tcode)) - return 0; - - do_pending_stack_adjust (); - start_sequence (); - prepare_cmp_insn (op1, op2, code, NULL_RTX, false, OPTAB_DIRECT, - &trap_rtx, &mode); - if (!trap_rtx) - insn = NULL; - else - insn = GEN_FCN (icode) (trap_rtx, XEXP (trap_rtx, 0), XEXP (trap_rtx, 1), - tcode); - - /* If that failed, then give up. */ - if (insn == 0) - { - end_sequence (); - return 0; - } - - emit_insn (insn); - insn = get_insns (); - end_sequence (); - return insn; -} - -/* Return rtx code for TCODE or UNKNOWN. Use UNSIGNEDP to select signed - or unsigned operation code. */ - -enum rtx_code -get_rtx_code_1 (enum tree_code tcode, bool unsignedp) -{ - enum rtx_code code; - switch (tcode) - { - case EQ_EXPR: - code = EQ; - break; - case NE_EXPR: - code = NE; - break; - case LT_EXPR: - code = unsignedp ? LTU : LT; - break; - case LE_EXPR: - code = unsignedp ? LEU : LE; - break; - case GT_EXPR: - code = unsignedp ? GTU : GT; - break; - case GE_EXPR: - code = unsignedp ? GEU : GE; - break; - - case UNORDERED_EXPR: - code = UNORDERED; - break; - case ORDERED_EXPR: - code = ORDERED; - break; - case UNLT_EXPR: - code = UNLT; - break; - case UNLE_EXPR: - code = UNLE; - break; - case UNGT_EXPR: - code = UNGT; - break; - case UNGE_EXPR: - code = UNGE; - break; - case UNEQ_EXPR: - code = UNEQ; - break; - case LTGT_EXPR: - code = LTGT; - break; - - case BIT_AND_EXPR: - code = AND; - break; - - case BIT_IOR_EXPR: - code = IOR; - break; - - default: - code = UNKNOWN; - break; - } - return code; -} - -/* Return rtx code for TCODE. Use UNSIGNEDP to select signed - or unsigned operation code. */ - -enum rtx_code -get_rtx_code (enum tree_code tcode, bool unsignedp) -{ - enum rtx_code code = get_rtx_code_1 (tcode, unsignedp); - gcc_assert (code != UNKNOWN); - return code; -} - -/* Return a comparison rtx of mode CMP_MODE for COND. Use UNSIGNEDP to - select signed or unsigned operators. OPNO holds the index of the - first comparison operand for insn ICODE. Do not generate the - compare instruction itself. */ - -rtx -vector_compare_rtx (machine_mode cmp_mode, enum tree_code tcode, - tree t_op0, tree t_op1, bool unsignedp, - enum insn_code icode, unsigned int opno) -{ - class expand_operand ops[2]; - rtx rtx_op0, rtx_op1; - machine_mode m0, m1; - enum rtx_code rcode = get_rtx_code (tcode, unsignedp); - - gcc_assert (TREE_CODE_CLASS (tcode) == tcc_comparison); - - /* Expand operands. For vector types with scalar modes, e.g. where int64x1_t - has mode DImode, this can produce a constant RTX of mode VOIDmode; in such - cases, use the original mode. */ - rtx_op0 = expand_expr (t_op0, NULL_RTX, TYPE_MODE (TREE_TYPE (t_op0)), - EXPAND_STACK_PARM); - m0 = GET_MODE (rtx_op0); - if (m0 == VOIDmode) - m0 = TYPE_MODE (TREE_TYPE (t_op0)); - - rtx_op1 = expand_expr (t_op1, NULL_RTX, TYPE_MODE (TREE_TYPE (t_op1)), - EXPAND_STACK_PARM); - m1 = GET_MODE (rtx_op1); - if (m1 == VOIDmode) - m1 = TYPE_MODE (TREE_TYPE (t_op1)); - - create_input_operand (&ops[0], rtx_op0, m0); - create_input_operand (&ops[1], rtx_op1, m1); - if (!maybe_legitimize_operands (icode, opno, 2, ops)) - gcc_unreachable (); - return gen_rtx_fmt_ee (rcode, cmp_mode, ops[0].value, ops[1].value); -} - -/* Check if vec_perm mask SEL is a constant equivalent to a shift of - the first vec_perm operand, assuming the second operand (for left shift - first operand) is a constant vector of zeros. Return the shift distance - in bits if so, or NULL_RTX if the vec_perm is not a shift. MODE is the - mode of the value being shifted. SHIFT_OPTAB is vec_shr_optab for right - shift or vec_shl_optab for left shift. */ -static rtx -shift_amt_for_vec_perm_mask (machine_mode mode, const vec_perm_indices &sel, - optab shift_optab) -{ - unsigned int bitsize = GET_MODE_UNIT_BITSIZE (mode); - poly_int64 first = sel[0]; - if (maybe_ge (sel[0], GET_MODE_NUNITS (mode))) - return NULL_RTX; - - if (shift_optab == vec_shl_optab) - { - unsigned int nelt; - if (!GET_MODE_NUNITS (mode).is_constant (&nelt)) - return NULL_RTX; - unsigned firstidx = 0; - for (unsigned int i = 0; i < nelt; i++) - { - if (known_eq (sel[i], nelt)) - { - if (i == 0 || firstidx) - return NULL_RTX; - firstidx = i; - } - else if (firstidx - ? maybe_ne (sel[i], nelt + i - firstidx) - : maybe_ge (sel[i], nelt)) - return NULL_RTX; - } - - if (firstidx == 0) - return NULL_RTX; - first = firstidx; - } - else if (!sel.series_p (0, 1, first, 1)) - { - unsigned int nelt; - if (!GET_MODE_NUNITS (mode).is_constant (&nelt)) - return NULL_RTX; - for (unsigned int i = 1; i < nelt; i++) - { - poly_int64 expected = i + first; - /* Indices into the second vector are all equivalent. */ - if (maybe_lt (sel[i], nelt) - ? maybe_ne (sel[i], expected) - : maybe_lt (expected, nelt)) - return NULL_RTX; - } - } - - return gen_int_shift_amount (mode, first * bitsize); -} - -/* A subroutine of expand_vec_perm_var for expanding one vec_perm insn. */ - -static rtx -expand_vec_perm_1 (enum insn_code icode, rtx target, - rtx v0, rtx v1, rtx sel) -{ - machine_mode tmode = GET_MODE (target); - machine_mode smode = GET_MODE (sel); - class expand_operand ops[4]; - - gcc_assert (GET_MODE_CLASS (smode) == MODE_VECTOR_INT - || related_int_vector_mode (tmode).require () == smode); - create_output_operand (&ops[0], target, tmode); - create_input_operand (&ops[3], sel, smode); - - /* Make an effort to preserve v0 == v1. The target expander is able to - rely on this to determine if we're permuting a single input operand. */ - if (rtx_equal_p (v0, v1)) - { - if (!insn_operand_matches (icode, 1, v0)) - v0 = force_reg (tmode, v0); - gcc_checking_assert (insn_operand_matches (icode, 1, v0)); - gcc_checking_assert (insn_operand_matches (icode, 2, v0)); - - create_fixed_operand (&ops[1], v0); - create_fixed_operand (&ops[2], v0); - } - else - { - create_input_operand (&ops[1], v0, tmode); - create_input_operand (&ops[2], v1, tmode); - } - - if (maybe_expand_insn (icode, 4, ops)) - return ops[0].value; - return NULL_RTX; -} - -/* Implement a permutation of vectors v0 and v1 using the permutation - vector in SEL and return the result. Use TARGET to hold the result - if nonnull and convenient. - - MODE is the mode of the vectors being permuted (V0 and V1). SEL_MODE - is the TYPE_MODE associated with SEL, or BLKmode if SEL isn't known - to have a particular mode. */ - -rtx -expand_vec_perm_const (machine_mode mode, rtx v0, rtx v1, - const vec_perm_builder &sel, machine_mode sel_mode, - rtx target) -{ - if (!target || !register_operand (target, mode)) - target = gen_reg_rtx (mode); - - /* Set QIMODE to a different vector mode with byte elements. - If no such mode, or if MODE already has byte elements, use VOIDmode. */ - machine_mode qimode; - if (!qimode_for_vec_perm (mode).exists (&qimode)) - qimode = VOIDmode; - - rtx_insn *last = get_last_insn (); - - bool single_arg_p = rtx_equal_p (v0, v1); - /* Always specify two input vectors here and leave the target to handle - cases in which the inputs are equal. Not all backends can cope with - the single-input representation when testing for a double-input - target instruction. */ - vec_perm_indices indices (sel, 2, GET_MODE_NUNITS (mode)); - - /* See if this can be handled with a vec_shr or vec_shl. We only do this - if the second (for vec_shr) or first (for vec_shl) vector is all - zeroes. */ - insn_code shift_code = CODE_FOR_nothing; - insn_code shift_code_qi = CODE_FOR_nothing; - optab shift_optab = unknown_optab; - rtx v2 = v0; - if (v1 == CONST0_RTX (GET_MODE (v1))) - shift_optab = vec_shr_optab; - else if (v0 == CONST0_RTX (GET_MODE (v0))) - { - shift_optab = vec_shl_optab; - v2 = v1; - } - if (shift_optab != unknown_optab) - { - shift_code = optab_handler (shift_optab, mode); - shift_code_qi = ((qimode != VOIDmode && qimode != mode) - ? optab_handler (shift_optab, qimode) - : CODE_FOR_nothing); - } - if (shift_code != CODE_FOR_nothing || shift_code_qi != CODE_FOR_nothing) - { - rtx shift_amt = shift_amt_for_vec_perm_mask (mode, indices, shift_optab); - if (shift_amt) - { - class expand_operand ops[3]; - if (shift_amt == const0_rtx) - return v2; - if (shift_code != CODE_FOR_nothing) - { - create_output_operand (&ops[0], target, mode); - create_input_operand (&ops[1], v2, mode); - create_convert_operand_from_type (&ops[2], shift_amt, sizetype); - if (maybe_expand_insn (shift_code, 3, ops)) - return ops[0].value; - } - if (shift_code_qi != CODE_FOR_nothing) - { - rtx tmp = gen_reg_rtx (qimode); - create_output_operand (&ops[0], tmp, qimode); - create_input_operand (&ops[1], gen_lowpart (qimode, v2), qimode); - create_convert_operand_from_type (&ops[2], shift_amt, sizetype); - if (maybe_expand_insn (shift_code_qi, 3, ops)) - return gen_lowpart (mode, ops[0].value); - } - } - } - - if (targetm.vectorize.vec_perm_const != NULL) - { - if (single_arg_p) - v1 = v0; - - if (targetm.vectorize.vec_perm_const (mode, target, v0, v1, indices)) - return target; - } - - /* Fall back to a constant byte-based permutation. */ - vec_perm_indices qimode_indices; - rtx target_qi = NULL_RTX, v0_qi = NULL_RTX, v1_qi = NULL_RTX; - if (qimode != VOIDmode) - { - qimode_indices.new_expanded_vector (indices, GET_MODE_UNIT_SIZE (mode)); - target_qi = gen_reg_rtx (qimode); - v0_qi = gen_lowpart (qimode, v0); - v1_qi = gen_lowpart (qimode, v1); - if (targetm.vectorize.vec_perm_const != NULL - && targetm.vectorize.vec_perm_const (qimode, target_qi, v0_qi, - v1_qi, qimode_indices)) - return gen_lowpart (mode, target_qi); - } - - v0 = force_reg (mode, v0); - if (single_arg_p) - v1 = v0; - v1 = force_reg (mode, v1); - - /* Otherwise expand as a fully variable permuation. */ - - /* The optabs are only defined for selectors with the same width - as the values being permuted. */ - machine_mode required_sel_mode; - if (!related_int_vector_mode (mode).exists (&required_sel_mode)) - { - delete_insns_since (last); - return NULL_RTX; - } - - /* We know that it is semantically valid to treat SEL as having SEL_MODE. - If that isn't the mode we want then we need to prove that using - REQUIRED_SEL_MODE is OK. */ - if (sel_mode != required_sel_mode) - { - if (!selector_fits_mode_p (required_sel_mode, indices)) - { - delete_insns_since (last); - return NULL_RTX; - } - sel_mode = required_sel_mode; - } - - insn_code icode = direct_optab_handler (vec_perm_optab, mode); - if (icode != CODE_FOR_nothing) - { - rtx sel_rtx = vec_perm_indices_to_rtx (sel_mode, indices); - rtx tmp = expand_vec_perm_1 (icode, target, v0, v1, sel_rtx); - if (tmp) - return tmp; - } - - if (qimode != VOIDmode - && selector_fits_mode_p (qimode, qimode_indices)) - { - icode = direct_optab_handler (vec_perm_optab, qimode); - if (icode != CODE_FOR_nothing) - { - rtx sel_qi = vec_perm_indices_to_rtx (qimode, qimode_indices); - rtx tmp = expand_vec_perm_1 (icode, target_qi, v0_qi, v1_qi, sel_qi); - if (tmp) - return gen_lowpart (mode, tmp); - } - } - - delete_insns_since (last); - return NULL_RTX; -} - -/* Implement a permutation of vectors v0 and v1 using the permutation - vector in SEL and return the result. Use TARGET to hold the result - if nonnull and convenient. - - MODE is the mode of the vectors being permuted (V0 and V1). - SEL must have the integer equivalent of MODE and is known to be - unsuitable for permutes with a constant permutation vector. */ - -rtx -expand_vec_perm_var (machine_mode mode, rtx v0, rtx v1, rtx sel, rtx target) -{ - enum insn_code icode; - unsigned int i, u; - rtx tmp, sel_qi; - - u = GET_MODE_UNIT_SIZE (mode); - - if (!target || GET_MODE (target) != mode) - target = gen_reg_rtx (mode); - - icode = direct_optab_handler (vec_perm_optab, mode); - if (icode != CODE_FOR_nothing) - { - tmp = expand_vec_perm_1 (icode, target, v0, v1, sel); - if (tmp) - return tmp; - } - - /* As a special case to aid several targets, lower the element-based - permutation to a byte-based permutation and try again. */ - machine_mode qimode; - if (!qimode_for_vec_perm (mode).exists (&qimode) - || maybe_gt (GET_MODE_NUNITS (qimode), GET_MODE_MASK (QImode) + 1)) - return NULL_RTX; - icode = direct_optab_handler (vec_perm_optab, qimode); - if (icode == CODE_FOR_nothing) - return NULL_RTX; - - /* Multiply each element by its byte size. */ - machine_mode selmode = GET_MODE (sel); - if (u == 2) - sel = expand_simple_binop (selmode, PLUS, sel, sel, - NULL, 0, OPTAB_DIRECT); - else - sel = expand_simple_binop (selmode, ASHIFT, sel, - gen_int_shift_amount (selmode, exact_log2 (u)), - NULL, 0, OPTAB_DIRECT); - gcc_assert (sel != NULL); - - /* Broadcast the low byte each element into each of its bytes. - The encoding has U interleaved stepped patterns, one for each - byte of an element. */ - vec_perm_builder const_sel (GET_MODE_SIZE (mode), u, 3); - unsigned int low_byte_in_u = BYTES_BIG_ENDIAN ? u - 1 : 0; - for (i = 0; i < 3; ++i) - for (unsigned int j = 0; j < u; ++j) - const_sel.quick_push (i * u + low_byte_in_u); - sel = gen_lowpart (qimode, sel); - sel = expand_vec_perm_const (qimode, sel, sel, const_sel, qimode, NULL); - gcc_assert (sel != NULL); - - /* Add the byte offset to each byte element. */ - /* Note that the definition of the indicies here is memory ordering, - so there should be no difference between big and little endian. */ - rtx_vector_builder byte_indices (qimode, u, 1); - for (i = 0; i < u; ++i) - byte_indices.quick_push (GEN_INT (i)); - tmp = byte_indices.build (); - sel_qi = expand_simple_binop (qimode, PLUS, sel, tmp, - sel, 0, OPTAB_DIRECT); - gcc_assert (sel_qi != NULL); - - tmp = mode != qimode ? gen_reg_rtx (qimode) : target; - tmp = expand_vec_perm_1 (icode, tmp, gen_lowpart (qimode, v0), - gen_lowpart (qimode, v1), sel_qi); - if (tmp) - tmp = gen_lowpart (mode, tmp); - return tmp; -} - -/* Generate VEC_SERIES_EXPR <OP0, OP1>, returning a value of mode VMODE. - Use TARGET for the result if nonnull and convenient. */ - -rtx -expand_vec_series_expr (machine_mode vmode, rtx op0, rtx op1, rtx target) -{ - class expand_operand ops[3]; - enum insn_code icode; - machine_mode emode = GET_MODE_INNER (vmode); - - icode = direct_optab_handler (vec_series_optab, vmode); - gcc_assert (icode != CODE_FOR_nothing); - - create_output_operand (&ops[0], target, vmode); - create_input_operand (&ops[1], op0, emode); - create_input_operand (&ops[2], op1, emode); - - expand_insn (icode, 3, ops); - return ops[0].value; -} - -/* Generate insns for a vector comparison into a mask. */ - -rtx -expand_vec_cmp_expr (tree type, tree exp, rtx target) -{ - class expand_operand ops[4]; - enum insn_code icode; - rtx comparison; - machine_mode mask_mode = TYPE_MODE (type); - machine_mode vmode; - bool unsignedp; - tree op0a, op0b; - enum tree_code tcode; - - op0a = TREE_OPERAND (exp, 0); - op0b = TREE_OPERAND (exp, 1); - tcode = TREE_CODE (exp); - - unsignedp = TYPE_UNSIGNED (TREE_TYPE (op0a)); - vmode = TYPE_MODE (TREE_TYPE (op0a)); - - icode = get_vec_cmp_icode (vmode, mask_mode, unsignedp); - if (icode == CODE_FOR_nothing) - { - if (tcode == EQ_EXPR || tcode == NE_EXPR) - icode = get_vec_cmp_eq_icode (vmode, mask_mode); - if (icode == CODE_FOR_nothing) - return 0; - } - - comparison = vector_compare_rtx (mask_mode, tcode, op0a, op0b, - unsignedp, icode, 2); - create_output_operand (&ops[0], target, mask_mode); - create_fixed_operand (&ops[1], comparison); - create_fixed_operand (&ops[2], XEXP (comparison, 0)); - create_fixed_operand (&ops[3], XEXP (comparison, 1)); - expand_insn (icode, 4, ops); - return ops[0].value; -} - -/* Expand a highpart multiply. */ - -rtx -expand_mult_highpart (machine_mode mode, rtx op0, rtx op1, - rtx target, bool uns_p) -{ - class expand_operand eops[3]; - enum insn_code icode; - int method, i; - machine_mode wmode; - rtx m1, m2; - optab tab1, tab2; - - method = can_mult_highpart_p (mode, uns_p); - switch (method) - { - case 0: - return NULL_RTX; - case 1: - tab1 = uns_p ? umul_highpart_optab : smul_highpart_optab; - return expand_binop (mode, tab1, op0, op1, target, uns_p, - OPTAB_LIB_WIDEN); - case 2: - tab1 = uns_p ? vec_widen_umult_even_optab : vec_widen_smult_even_optab; - tab2 = uns_p ? vec_widen_umult_odd_optab : vec_widen_smult_odd_optab; - break; - case 3: - tab1 = uns_p ? vec_widen_umult_lo_optab : vec_widen_smult_lo_optab; - tab2 = uns_p ? vec_widen_umult_hi_optab : vec_widen_smult_hi_optab; - if (BYTES_BIG_ENDIAN) - std::swap (tab1, tab2); - break; - default: - gcc_unreachable (); - } - - icode = optab_handler (tab1, mode); - wmode = insn_data[icode].operand[0].mode; - gcc_checking_assert (known_eq (2 * GET_MODE_NUNITS (wmode), - GET_MODE_NUNITS (mode))); - gcc_checking_assert (known_eq (GET_MODE_SIZE (wmode), GET_MODE_SIZE (mode))); - - create_output_operand (&eops[0], gen_reg_rtx (wmode), wmode); - create_input_operand (&eops[1], op0, mode); - create_input_operand (&eops[2], op1, mode); - expand_insn (icode, 3, eops); - m1 = gen_lowpart (mode, eops[0].value); - - create_output_operand (&eops[0], gen_reg_rtx (wmode), wmode); - create_input_operand (&eops[1], op0, mode); - create_input_operand (&eops[2], op1, mode); - expand_insn (optab_handler (tab2, mode), 3, eops); - m2 = gen_lowpart (mode, eops[0].value); - - vec_perm_builder sel; - if (method == 2) - { - /* The encoding has 2 interleaved stepped patterns. */ - sel.new_vector (GET_MODE_NUNITS (mode), 2, 3); - for (i = 0; i < 6; ++i) - sel.quick_push (!BYTES_BIG_ENDIAN + (i & ~1) - + ((i & 1) ? GET_MODE_NUNITS (mode) : 0)); - } - else - { - /* The encoding has a single interleaved stepped pattern. */ - sel.new_vector (GET_MODE_NUNITS (mode), 1, 3); - for (i = 0; i < 3; ++i) - sel.quick_push (2 * i + (BYTES_BIG_ENDIAN ? 0 : 1)); - } - - return expand_vec_perm_const (mode, m1, m2, sel, BLKmode, target); -} - -/* Helper function to find the MODE_CC set in a sync_compare_and_swap - pattern. */ - -static void -find_cc_set (rtx x, const_rtx pat, void *data) -{ - if (REG_P (x) && GET_MODE_CLASS (GET_MODE (x)) == MODE_CC - && GET_CODE (pat) == SET) - { - rtx *p_cc_reg = (rtx *) data; - gcc_assert (!*p_cc_reg); - *p_cc_reg = x; - } -} - -/* This is a helper function for the other atomic operations. This function - emits a loop that contains SEQ that iterates until a compare-and-swap - operation at the end succeeds. MEM is the memory to be modified. SEQ is - a set of instructions that takes a value from OLD_REG as an input and - produces a value in NEW_REG as an output. Before SEQ, OLD_REG will be - set to the current contents of MEM. After SEQ, a compare-and-swap will - attempt to update MEM with NEW_REG. The function returns true when the - loop was generated successfully. */ - -static bool -expand_compare_and_swap_loop (rtx mem, rtx old_reg, rtx new_reg, rtx seq) -{ - machine_mode mode = GET_MODE (mem); - rtx_code_label *label; - rtx cmp_reg, success, oldval; - - /* The loop we want to generate looks like - - cmp_reg = mem; - label: - old_reg = cmp_reg; - seq; - (success, cmp_reg) = compare-and-swap(mem, old_reg, new_reg) - if (success) - goto label; - - Note that we only do the plain load from memory once. Subsequent - iterations use the value loaded by the compare-and-swap pattern. */ - - label = gen_label_rtx (); - cmp_reg = gen_reg_rtx (mode); - - emit_move_insn (cmp_reg, mem); - emit_label (label); - emit_move_insn (old_reg, cmp_reg); - if (seq) - emit_insn (seq); - - success = NULL_RTX; - oldval = cmp_reg; - if (!expand_atomic_compare_and_swap (&success, &oldval, mem, old_reg, - new_reg, false, MEMMODEL_SYNC_SEQ_CST, - MEMMODEL_RELAXED)) - return false; - - if (oldval != cmp_reg) - emit_move_insn (cmp_reg, oldval); - - /* Mark this jump predicted not taken. */ - emit_cmp_and_jump_insns (success, const0_rtx, EQ, const0_rtx, - GET_MODE (success), 1, label, - profile_probability::guessed_never ()); - return true; -} - - -/* This function tries to emit an atomic_exchange intruction. VAL is written - to *MEM using memory model MODEL. The previous contents of *MEM are returned, - using TARGET if possible. */ - -static rtx -maybe_emit_atomic_exchange (rtx target, rtx mem, rtx val, enum memmodel model) -{ - machine_mode mode = GET_MODE (mem); - enum insn_code icode; - - /* If the target supports the exchange directly, great. */ - icode = direct_optab_handler (atomic_exchange_optab, mode); - if (icode != CODE_FOR_nothing) - { - class expand_operand ops[4]; - - create_output_operand (&ops[0], target, mode); - create_fixed_operand (&ops[1], mem); - create_input_operand (&ops[2], val, mode); - create_integer_operand (&ops[3], model); - if (maybe_expand_insn (icode, 4, ops)) - return ops[0].value; - } - - return NULL_RTX; -} - -/* This function tries to implement an atomic exchange operation using - __sync_lock_test_and_set. VAL is written to *MEM using memory model MODEL. - The previous contents of *MEM are returned, using TARGET if possible. - Since this instructionn is an acquire barrier only, stronger memory - models may require additional barriers to be emitted. */ - -static rtx -maybe_emit_sync_lock_test_and_set (rtx target, rtx mem, rtx val, - enum memmodel model) -{ - machine_mode mode = GET_MODE (mem); - enum insn_code icode; - rtx_insn *last_insn = get_last_insn (); - - icode = optab_handler (sync_lock_test_and_set_optab, mode); - - /* Legacy sync_lock_test_and_set is an acquire barrier. If the pattern - exists, and the memory model is stronger than acquire, add a release - barrier before the instruction. */ - - if (is_mm_seq_cst (model) || is_mm_release (model) || is_mm_acq_rel (model)) - expand_mem_thread_fence (model); - - if (icode != CODE_FOR_nothing) - { - class expand_operand ops[3]; - create_output_operand (&ops[0], target, mode); - create_fixed_operand (&ops[1], mem); - create_input_operand (&ops[2], val, mode); - if (maybe_expand_insn (icode, 3, ops)) - return ops[0].value; - } - - /* If an external test-and-set libcall is provided, use that instead of - any external compare-and-swap that we might get from the compare-and- - swap-loop expansion later. */ - if (!can_compare_and_swap_p (mode, false)) - { - rtx libfunc = optab_libfunc (sync_lock_test_and_set_optab, mode); - if (libfunc != NULL) - { - rtx addr; - - addr = convert_memory_address (ptr_mode, XEXP (mem, 0)); - return emit_library_call_value (libfunc, NULL_RTX, LCT_NORMAL, - mode, addr, ptr_mode, - val, mode); - } - } - - /* If the test_and_set can't be emitted, eliminate any barrier that might - have been emitted. */ - delete_insns_since (last_insn); - return NULL_RTX; -} - -/* This function tries to implement an atomic exchange operation using a - compare_and_swap loop. VAL is written to *MEM. The previous contents of - *MEM are returned, using TARGET if possible. No memory model is required - since a compare_and_swap loop is seq-cst. */ - -static rtx -maybe_emit_compare_and_swap_exchange_loop (rtx target, rtx mem, rtx val) -{ - machine_mode mode = GET_MODE (mem); - - if (can_compare_and_swap_p (mode, true)) - { - if (!target || !register_operand (target, mode)) - target = gen_reg_rtx (mode); - if (expand_compare_and_swap_loop (mem, target, val, NULL_RTX)) - return target; - } - - return NULL_RTX; -} - -/* This function tries to implement an atomic test-and-set operation - using the atomic_test_and_set instruction pattern. A boolean value - is returned from the operation, using TARGET if possible. */ - -static rtx -maybe_emit_atomic_test_and_set (rtx target, rtx mem, enum memmodel model) -{ - machine_mode pat_bool_mode; - class expand_operand ops[3]; - - if (!targetm.have_atomic_test_and_set ()) - return NULL_RTX; - - /* While we always get QImode from __atomic_test_and_set, we get - other memory modes from __sync_lock_test_and_set. Note that we - use no endian adjustment here. This matches the 4.6 behavior - in the Sparc backend. */ - enum insn_code icode = targetm.code_for_atomic_test_and_set; - gcc_checking_assert (insn_data[icode].operand[1].mode == QImode); - if (GET_MODE (mem) != QImode) - mem = adjust_address_nv (mem, QImode, 0); - - pat_bool_mode = insn_data[icode].operand[0].mode; - create_output_operand (&ops[0], target, pat_bool_mode); - create_fixed_operand (&ops[1], mem); - create_integer_operand (&ops[2], model); - - if (maybe_expand_insn (icode, 3, ops)) - return ops[0].value; - return NULL_RTX; -} - -/* This function expands the legacy _sync_lock test_and_set operation which is - generally an atomic exchange. Some limited targets only allow the - constant 1 to be stored. This is an ACQUIRE operation. - - TARGET is an optional place to stick the return value. - MEM is where VAL is stored. */ - -rtx -expand_sync_lock_test_and_set (rtx target, rtx mem, rtx val) -{ - rtx ret; - - /* Try an atomic_exchange first. */ - ret = maybe_emit_atomic_exchange (target, mem, val, MEMMODEL_SYNC_ACQUIRE); - if (ret) - return ret; - - ret = maybe_emit_sync_lock_test_and_set (target, mem, val, - MEMMODEL_SYNC_ACQUIRE); - if (ret) - return ret; - - ret = maybe_emit_compare_and_swap_exchange_loop (target, mem, val); - if (ret) - return ret; - - /* If there are no other options, try atomic_test_and_set if the value - being stored is 1. */ - if (val == const1_rtx) - ret = maybe_emit_atomic_test_and_set (target, mem, MEMMODEL_SYNC_ACQUIRE); - - return ret; -} - -/* This function expands the atomic test_and_set operation: - atomically store a boolean TRUE into MEM and return the previous value. - - MEMMODEL is the memory model variant to use. - TARGET is an optional place to stick the return value. */ - -rtx -expand_atomic_test_and_set (rtx target, rtx mem, enum memmodel model) -{ - machine_mode mode = GET_MODE (mem); - rtx ret, trueval, subtarget; - - ret = maybe_emit_atomic_test_and_set (target, mem, model); - if (ret) - return ret; - - /* Be binary compatible with non-default settings of trueval, and different - cpu revisions. E.g. one revision may have atomic-test-and-set, but - another only has atomic-exchange. */ - if (targetm.atomic_test_and_set_trueval == 1) - { - trueval = const1_rtx; - subtarget = target ? target : gen_reg_rtx (mode); - } - else - { - trueval = gen_int_mode (targetm.atomic_test_and_set_trueval, mode); - subtarget = gen_reg_rtx (mode); - } - - /* Try the atomic-exchange optab... */ - ret = maybe_emit_atomic_exchange (subtarget, mem, trueval, model); - - /* ... then an atomic-compare-and-swap loop ... */ - if (!ret) - ret = maybe_emit_compare_and_swap_exchange_loop (subtarget, mem, trueval); - - /* ... before trying the vaguely defined legacy lock_test_and_set. */ - if (!ret) - ret = maybe_emit_sync_lock_test_and_set (subtarget, mem, trueval, model); - - /* Recall that the legacy lock_test_and_set optab was allowed to do magic - things with the value 1. Thus we try again without trueval. */ - if (!ret && targetm.atomic_test_and_set_trueval != 1) - ret = maybe_emit_sync_lock_test_and_set (subtarget, mem, const1_rtx, model); - - /* Failing all else, assume a single threaded environment and simply - perform the operation. */ - if (!ret) - { - /* If the result is ignored skip the move to target. */ - if (subtarget != const0_rtx) - emit_move_insn (subtarget, mem); - - emit_move_insn (mem, trueval); - ret = subtarget; - } - - /* Recall that have to return a boolean value; rectify if trueval - is not exactly one. */ - if (targetm.atomic_test_and_set_trueval != 1) - ret = emit_store_flag_force (target, NE, ret, const0_rtx, mode, 0, 1); - - return ret; -} - -/* This function expands the atomic exchange operation: - atomically store VAL in MEM and return the previous value in MEM. - - MEMMODEL is the memory model variant to use. - TARGET is an optional place to stick the return value. */ - -rtx -expand_atomic_exchange (rtx target, rtx mem, rtx val, enum memmodel model) -{ - machine_mode mode = GET_MODE (mem); - rtx ret; - - /* If loads are not atomic for the required size and we are not called to - provide a __sync builtin, do not do anything so that we stay consistent - with atomic loads of the same size. */ - if (!can_atomic_load_p (mode) && !is_mm_sync (model)) - return NULL_RTX; - - ret = maybe_emit_atomic_exchange (target, mem, val, model); - - /* Next try a compare-and-swap loop for the exchange. */ - if (!ret) - ret = maybe_emit_compare_and_swap_exchange_loop (target, mem, val); - - return ret; -} - -/* This function expands the atomic compare exchange operation: - - *PTARGET_BOOL is an optional place to store the boolean success/failure. - *PTARGET_OVAL is an optional place to store the old value from memory. - Both target parameters may be NULL or const0_rtx to indicate that we do - not care about that return value. Both target parameters are updated on - success to the actual location of the corresponding result. - - MEMMODEL is the memory model variant to use. - - The return value of the function is true for success. */ - -bool -expand_atomic_compare_and_swap (rtx *ptarget_bool, rtx *ptarget_oval, - rtx mem, rtx expected, rtx desired, - bool is_weak, enum memmodel succ_model, - enum memmodel fail_model) -{ - machine_mode mode = GET_MODE (mem); - class expand_operand ops[8]; - enum insn_code icode; - rtx target_oval, target_bool = NULL_RTX; - rtx libfunc; - - /* If loads are not atomic for the required size and we are not called to - provide a __sync builtin, do not do anything so that we stay consistent - with atomic loads of the same size. */ - if (!can_atomic_load_p (mode) && !is_mm_sync (succ_model)) - return false; - - /* Load expected into a register for the compare and swap. */ - if (MEM_P (expected)) - expected = copy_to_reg (expected); - - /* Make sure we always have some place to put the return oldval. - Further, make sure that place is distinct from the input expected, - just in case we need that path down below. */ - if (ptarget_oval && *ptarget_oval == const0_rtx) - ptarget_oval = NULL; - - if (ptarget_oval == NULL - || (target_oval = *ptarget_oval) == NULL - || reg_overlap_mentioned_p (expected, target_oval)) - target_oval = gen_reg_rtx (mode); - - icode = direct_optab_handler (atomic_compare_and_swap_optab, mode); - if (icode != CODE_FOR_nothing) - { - machine_mode bool_mode = insn_data[icode].operand[0].mode; - - if (ptarget_bool && *ptarget_bool == const0_rtx) - ptarget_bool = NULL; - - /* Make sure we always have a place for the bool operand. */ - if (ptarget_bool == NULL - || (target_bool = *ptarget_bool) == NULL - || GET_MODE (target_bool) != bool_mode) - target_bool = gen_reg_rtx (bool_mode); - - /* Emit the compare_and_swap. */ - create_output_operand (&ops[0], target_bool, bool_mode); - create_output_operand (&ops[1], target_oval, mode); - create_fixed_operand (&ops[2], mem); - create_input_operand (&ops[3], expected, mode); - create_input_operand (&ops[4], desired, mode); - create_integer_operand (&ops[5], is_weak); - create_integer_operand (&ops[6], succ_model); - create_integer_operand (&ops[7], fail_model); - if (maybe_expand_insn (icode, 8, ops)) - { - /* Return success/failure. */ - target_bool = ops[0].value; - target_oval = ops[1].value; - goto success; - } - } - - /* Otherwise fall back to the original __sync_val_compare_and_swap - which is always seq-cst. */ - icode = optab_handler (sync_compare_and_swap_optab, mode); - if (icode != CODE_FOR_nothing) - { - rtx cc_reg; - - create_output_operand (&ops[0], target_oval, mode); - create_fixed_operand (&ops[1], mem); - create_input_operand (&ops[2], expected, mode); - create_input_operand (&ops[3], desired, mode); - if (!maybe_expand_insn (icode, 4, ops)) - return false; - - target_oval = ops[0].value; - - /* If the caller isn't interested in the boolean return value, - skip the computation of it. */ - if (ptarget_bool == NULL) - goto success; - - /* Otherwise, work out if the compare-and-swap succeeded. */ - cc_reg = NULL_RTX; - if (have_insn_for (COMPARE, CCmode)) - note_stores (get_last_insn (), find_cc_set, &cc_reg); - if (cc_reg) - { - target_bool = emit_store_flag_force (target_bool, EQ, cc_reg, - const0_rtx, VOIDmode, 0, 1); - goto success; - } - goto success_bool_from_val; - } - - /* Also check for library support for __sync_val_compare_and_swap. */ - libfunc = optab_libfunc (sync_compare_and_swap_optab, mode); - if (libfunc != NULL) - { - rtx addr = convert_memory_address (ptr_mode, XEXP (mem, 0)); - rtx target = emit_library_call_value (libfunc, NULL_RTX, LCT_NORMAL, - mode, addr, ptr_mode, - expected, mode, desired, mode); - emit_move_insn (target_oval, target); - - /* Compute the boolean return value only if requested. */ - if (ptarget_bool) - goto success_bool_from_val; - else - goto success; - } - - /* Failure. */ - return false; - - success_bool_from_val: - target_bool = emit_store_flag_force (target_bool, EQ, target_oval, - expected, VOIDmode, 1, 1); - success: - /* Make sure that the oval output winds up where the caller asked. */ - if (ptarget_oval) - *ptarget_oval = target_oval; - if (ptarget_bool) - *ptarget_bool = target_bool; - return true; -} - -/* Generate asm volatile("" : : : "memory") as the memory blockage. */ - -static void -expand_asm_memory_blockage (void) -{ - rtx asm_op, clob; - - asm_op = gen_rtx_ASM_OPERANDS (VOIDmode, "", "", 0, - rtvec_alloc (0), rtvec_alloc (0), - rtvec_alloc (0), UNKNOWN_LOCATION); - MEM_VOLATILE_P (asm_op) = 1; - - clob = gen_rtx_SCRATCH (VOIDmode); - clob = gen_rtx_MEM (BLKmode, clob); - clob = gen_rtx_CLOBBER (VOIDmode, clob); - - emit_insn (gen_rtx_PARALLEL (VOIDmode, gen_rtvec (2, asm_op, clob))); -} - -/* Do not propagate memory accesses across this point. */ - -static void -expand_memory_blockage (void) -{ - if (targetm.have_memory_blockage ()) - emit_insn (targetm.gen_memory_blockage ()); - else - expand_asm_memory_blockage (); -} - -/* Generate asm volatile("" : : : "memory") as a memory blockage, at the - same time clobbering the register set specified by REGS. */ - -void -expand_asm_reg_clobber_mem_blockage (HARD_REG_SET regs) -{ - rtx asm_op, clob_mem; - - unsigned int num_of_regs = 0; - for (unsigned int i = 0; i < FIRST_PSEUDO_REGISTER; i++) - if (TEST_HARD_REG_BIT (regs, i)) - num_of_regs++; - - asm_op = gen_rtx_ASM_OPERANDS (VOIDmode, "", "", 0, - rtvec_alloc (0), rtvec_alloc (0), - rtvec_alloc (0), UNKNOWN_LOCATION); - MEM_VOLATILE_P (asm_op) = 1; - - rtvec v = rtvec_alloc (num_of_regs + 2); - - clob_mem = gen_rtx_SCRATCH (VOIDmode); - clob_mem = gen_rtx_MEM (BLKmode, clob_mem); - clob_mem = gen_rtx_CLOBBER (VOIDmode, clob_mem); - - RTVEC_ELT (v, 0) = asm_op; - RTVEC_ELT (v, 1) = clob_mem; - - if (num_of_regs > 0) - { - unsigned int j = 2; - for (unsigned int i = 0; i < FIRST_PSEUDO_REGISTER; i++) - if (TEST_HARD_REG_BIT (regs, i)) - { - RTVEC_ELT (v, j) = gen_rtx_CLOBBER (VOIDmode, regno_reg_rtx[i]); - j++; - } - gcc_assert (j == (num_of_regs + 2)); - } - - emit_insn (gen_rtx_PARALLEL (VOIDmode, v)); -} - -/* This routine will either emit the mem_thread_fence pattern or issue a - sync_synchronize to generate a fence for memory model MEMMODEL. */ - -void -expand_mem_thread_fence (enum memmodel model) -{ - if (is_mm_relaxed (model)) - return; - if (targetm.have_mem_thread_fence ()) - { - emit_insn (targetm.gen_mem_thread_fence (GEN_INT (model))); - expand_memory_blockage (); - } - else if (targetm.have_memory_barrier ()) - emit_insn (targetm.gen_memory_barrier ()); - else if (synchronize_libfunc != NULL_RTX) - emit_library_call (synchronize_libfunc, LCT_NORMAL, VOIDmode); - else - expand_memory_blockage (); -} - -/* Emit a signal fence with given memory model. */ - -void -expand_mem_signal_fence (enum memmodel model) -{ - /* No machine barrier is required to implement a signal fence, but - a compiler memory barrier must be issued, except for relaxed MM. */ - if (!is_mm_relaxed (model)) - expand_memory_blockage (); -} - -/* This function expands the atomic load operation: - return the atomically loaded value in MEM. - - MEMMODEL is the memory model variant to use. - TARGET is an option place to stick the return value. */ - -rtx -expand_atomic_load (rtx target, rtx mem, enum memmodel model) -{ - machine_mode mode = GET_MODE (mem); - enum insn_code icode; - - /* If the target supports the load directly, great. */ - icode = direct_optab_handler (atomic_load_optab, mode); - if (icode != CODE_FOR_nothing) - { - class expand_operand ops[3]; - rtx_insn *last = get_last_insn (); - if (is_mm_seq_cst (model)) - expand_memory_blockage (); - - create_output_operand (&ops[0], target, mode); - create_fixed_operand (&ops[1], mem); - create_integer_operand (&ops[2], model); - if (maybe_expand_insn (icode, 3, ops)) - { - if (!is_mm_relaxed (model)) - expand_memory_blockage (); - return ops[0].value; - } - delete_insns_since (last); - } - - /* If the size of the object is greater than word size on this target, - then we assume that a load will not be atomic. We could try to - emulate a load with a compare-and-swap operation, but the store that - doing this could result in would be incorrect if this is a volatile - atomic load or targetting read-only-mapped memory. */ - if (maybe_gt (GET_MODE_PRECISION (mode), BITS_PER_WORD)) - /* If there is no atomic load, leave the library call. */ - return NULL_RTX; - - /* Otherwise assume loads are atomic, and emit the proper barriers. */ - if (!target || target == const0_rtx) - target = gen_reg_rtx (mode); - - /* For SEQ_CST, emit a barrier before the load. */ - if (is_mm_seq_cst (model)) - expand_mem_thread_fence (model); - - emit_move_insn (target, mem); - - /* Emit the appropriate barrier after the load. */ - expand_mem_thread_fence (model); - - return target; -} - -/* This function expands the atomic store operation: - Atomically store VAL in MEM. - MEMMODEL is the memory model variant to use. - USE_RELEASE is true if __sync_lock_release can be used as a fall back. - function returns const0_rtx if a pattern was emitted. */ - -rtx -expand_atomic_store (rtx mem, rtx val, enum memmodel model, bool use_release) -{ - machine_mode mode = GET_MODE (mem); - enum insn_code icode; - class expand_operand ops[3]; - - /* If the target supports the store directly, great. */ - icode = direct_optab_handler (atomic_store_optab, mode); - if (icode != CODE_FOR_nothing) - { - rtx_insn *last = get_last_insn (); - if (!is_mm_relaxed (model)) - expand_memory_blockage (); - create_fixed_operand (&ops[0], mem); - create_input_operand (&ops[1], val, mode); - create_integer_operand (&ops[2], model); - if (maybe_expand_insn (icode, 3, ops)) - { - if (is_mm_seq_cst (model)) - expand_memory_blockage (); - return const0_rtx; - } - delete_insns_since (last); - } - - /* If using __sync_lock_release is a viable alternative, try it. - Note that this will not be set to true if we are expanding a generic - __atomic_store_n. */ - if (use_release) - { - icode = direct_optab_handler (sync_lock_release_optab, mode); - if (icode != CODE_FOR_nothing) - { - create_fixed_operand (&ops[0], mem); - create_input_operand (&ops[1], const0_rtx, mode); - if (maybe_expand_insn (icode, 2, ops)) - { - /* lock_release is only a release barrier. */ - if (is_mm_seq_cst (model)) - expand_mem_thread_fence (model); - return const0_rtx; - } - } - } - - /* If the size of the object is greater than word size on this target, - a default store will not be atomic. */ - if (maybe_gt (GET_MODE_PRECISION (mode), BITS_PER_WORD)) - { - /* If loads are atomic or we are called to provide a __sync builtin, - we can try a atomic_exchange and throw away the result. Otherwise, - don't do anything so that we do not create an inconsistency between - loads and stores. */ - if (can_atomic_load_p (mode) || is_mm_sync (model)) - { - rtx target = maybe_emit_atomic_exchange (NULL_RTX, mem, val, model); - if (!target) - target = maybe_emit_compare_and_swap_exchange_loop (NULL_RTX, mem, - val); - if (target) - return const0_rtx; - } - return NULL_RTX; - } - - /* Otherwise assume stores are atomic, and emit the proper barriers. */ - expand_mem_thread_fence (model); - - emit_move_insn (mem, val); - - /* For SEQ_CST, also emit a barrier after the store. */ - if (is_mm_seq_cst (model)) - expand_mem_thread_fence (model); - - return const0_rtx; -} - - -/* Structure containing the pointers and values required to process the - various forms of the atomic_fetch_op and atomic_op_fetch builtins. */ - -struct atomic_op_functions -{ - direct_optab mem_fetch_before; - direct_optab mem_fetch_after; - direct_optab mem_no_result; - optab fetch_before; - optab fetch_after; - direct_optab no_result; - enum rtx_code reverse_code; -}; - - -/* Fill in structure pointed to by OP with the various optab entries for an - operation of type CODE. */ - -static void -get_atomic_op_for_code (struct atomic_op_functions *op, enum rtx_code code) -{ - gcc_assert (op!= NULL); - - /* If SWITCHABLE_TARGET is defined, then subtargets can be switched - in the source code during compilation, and the optab entries are not - computable until runtime. Fill in the values at runtime. */ - switch (code) - { - case PLUS: - op->mem_fetch_before = atomic_fetch_add_optab; - op->mem_fetch_after = atomic_add_fetch_optab; - op->mem_no_result = atomic_add_optab; - op->fetch_before = sync_old_add_optab; - op->fetch_after = sync_new_add_optab; - op->no_result = sync_add_optab; - op->reverse_code = MINUS; - break; - case MINUS: - op->mem_fetch_before = atomic_fetch_sub_optab; - op->mem_fetch_after = atomic_sub_fetch_optab; - op->mem_no_result = atomic_sub_optab; - op->fetch_before = sync_old_sub_optab; - op->fetch_after = sync_new_sub_optab; - op->no_result = sync_sub_optab; - op->reverse_code = PLUS; - break; - case XOR: - op->mem_fetch_before = atomic_fetch_xor_optab; - op->mem_fetch_after = atomic_xor_fetch_optab; - op->mem_no_result = atomic_xor_optab; - op->fetch_before = sync_old_xor_optab; - op->fetch_after = sync_new_xor_optab; - op->no_result = sync_xor_optab; - op->reverse_code = XOR; - break; - case AND: - op->mem_fetch_before = atomic_fetch_and_optab; - op->mem_fetch_after = atomic_and_fetch_optab; - op->mem_no_result = atomic_and_optab; - op->fetch_before = sync_old_and_optab; - op->fetch_after = sync_new_and_optab; - op->no_result = sync_and_optab; - op->reverse_code = UNKNOWN; - break; - case IOR: - op->mem_fetch_before = atomic_fetch_or_optab; - op->mem_fetch_after = atomic_or_fetch_optab; - op->mem_no_result = atomic_or_optab; - op->fetch_before = sync_old_ior_optab; - op->fetch_after = sync_new_ior_optab; - op->no_result = sync_ior_optab; - op->reverse_code = UNKNOWN; - break; - case NOT: - op->mem_fetch_before = atomic_fetch_nand_optab; - op->mem_fetch_after = atomic_nand_fetch_optab; - op->mem_no_result = atomic_nand_optab; - op->fetch_before = sync_old_nand_optab; - op->fetch_after = sync_new_nand_optab; - op->no_result = sync_nand_optab; - op->reverse_code = UNKNOWN; - break; - default: - gcc_unreachable (); - } -} - -/* See if there is a more optimal way to implement the operation "*MEM CODE VAL" - using memory order MODEL. If AFTER is true the operation needs to return - the value of *MEM after the operation, otherwise the previous value. - TARGET is an optional place to place the result. The result is unused if - it is const0_rtx. - Return the result if there is a better sequence, otherwise NULL_RTX. */ - -static rtx -maybe_optimize_fetch_op (rtx target, rtx mem, rtx val, enum rtx_code code, - enum memmodel model, bool after) -{ - /* If the value is prefetched, or not used, it may be possible to replace - the sequence with a native exchange operation. */ - if (!after || target == const0_rtx) - { - /* fetch_and (&x, 0, m) can be replaced with exchange (&x, 0, m). */ - if (code == AND && val == const0_rtx) - { - if (target == const0_rtx) - target = gen_reg_rtx (GET_MODE (mem)); - return maybe_emit_atomic_exchange (target, mem, val, model); - } - - /* fetch_or (&x, -1, m) can be replaced with exchange (&x, -1, m). */ - if (code == IOR && val == constm1_rtx) - { - if (target == const0_rtx) - target = gen_reg_rtx (GET_MODE (mem)); - return maybe_emit_atomic_exchange (target, mem, val, model); - } - } - - return NULL_RTX; -} - -/* Try to emit an instruction for a specific operation varaition. - OPTAB contains the OP functions. - TARGET is an optional place to return the result. const0_rtx means unused. - MEM is the memory location to operate on. - VAL is the value to use in the operation. - USE_MEMMODEL is TRUE if the variation with a memory model should be tried. - MODEL is the memory model, if used. - AFTER is true if the returned result is the value after the operation. */ - -static rtx -maybe_emit_op (const struct atomic_op_functions *optab, rtx target, rtx mem, - rtx val, bool use_memmodel, enum memmodel model, bool after) -{ - machine_mode mode = GET_MODE (mem); - class expand_operand ops[4]; - enum insn_code icode; - int op_counter = 0; - int num_ops; - - /* Check to see if there is a result returned. */ - if (target == const0_rtx) - { - if (use_memmodel) - { - icode = direct_optab_handler (optab->mem_no_result, mode); - create_integer_operand (&ops[2], model); - num_ops = 3; - } - else - { - icode = direct_optab_handler (optab->no_result, mode); - num_ops = 2; - } - } - /* Otherwise, we need to generate a result. */ - else - { - if (use_memmodel) - { - icode = direct_optab_handler (after ? optab->mem_fetch_after - : optab->mem_fetch_before, mode); - create_integer_operand (&ops[3], model); - num_ops = 4; - } - else - { - icode = optab_handler (after ? optab->fetch_after - : optab->fetch_before, mode); - num_ops = 3; - } - create_output_operand (&ops[op_counter++], target, mode); - } - if (icode == CODE_FOR_nothing) - return NULL_RTX; - - create_fixed_operand (&ops[op_counter++], mem); - /* VAL may have been promoted to a wider mode. Shrink it if so. */ - create_convert_operand_to (&ops[op_counter++], val, mode, true); - - if (maybe_expand_insn (icode, num_ops, ops)) - return (target == const0_rtx ? const0_rtx : ops[0].value); - - return NULL_RTX; -} - - -/* This function expands an atomic fetch_OP or OP_fetch operation: - TARGET is an option place to stick the return value. const0_rtx indicates - the result is unused. - atomically fetch MEM, perform the operation with VAL and return it to MEM. - CODE is the operation being performed (OP) - MEMMODEL is the memory model variant to use. - AFTER is true to return the result of the operation (OP_fetch). - AFTER is false to return the value before the operation (fetch_OP). - - This function will *only* generate instructions if there is a direct - optab. No compare and swap loops or libcalls will be generated. */ - -static rtx -expand_atomic_fetch_op_no_fallback (rtx target, rtx mem, rtx val, - enum rtx_code code, enum memmodel model, - bool after) -{ - machine_mode mode = GET_MODE (mem); - struct atomic_op_functions optab; - rtx result; - bool unused_result = (target == const0_rtx); - - get_atomic_op_for_code (&optab, code); - - /* Check to see if there are any better instructions. */ - result = maybe_optimize_fetch_op (target, mem, val, code, model, after); - if (result) - return result; - - /* Check for the case where the result isn't used and try those patterns. */ - if (unused_result) - { - /* Try the memory model variant first. */ - result = maybe_emit_op (&optab, target, mem, val, true, model, true); - if (result) - return result; - - /* Next try the old style withuot a memory model. */ - result = maybe_emit_op (&optab, target, mem, val, false, model, true); - if (result) - return result; - - /* There is no no-result pattern, so try patterns with a result. */ - target = NULL_RTX; - } - - /* Try the __atomic version. */ - result = maybe_emit_op (&optab, target, mem, val, true, model, after); - if (result) - return result; - - /* Try the older __sync version. */ - result = maybe_emit_op (&optab, target, mem, val, false, model, after); - if (result) - return result; - - /* If the fetch value can be calculated from the other variation of fetch, - try that operation. */ - if (after || unused_result || optab.reverse_code != UNKNOWN) - { - /* Try the __atomic version, then the older __sync version. */ - result = maybe_emit_op (&optab, target, mem, val, true, model, !after); - if (!result) - result = maybe_emit_op (&optab, target, mem, val, false, model, !after); - - if (result) - { - /* If the result isn't used, no need to do compensation code. */ - if (unused_result) - return result; - - /* Issue compensation code. Fetch_after == fetch_before OP val. - Fetch_before == after REVERSE_OP val. */ - if (!after) - code = optab.reverse_code; - if (code == NOT) - { - result = expand_simple_binop (mode, AND, result, val, NULL_RTX, - true, OPTAB_LIB_WIDEN); - result = expand_simple_unop (mode, NOT, result, target, true); - } - else - result = expand_simple_binop (mode, code, result, val, target, - true, OPTAB_LIB_WIDEN); - return result; - } - } - - /* No direct opcode can be generated. */ - return NULL_RTX; -} - - - -/* This function expands an atomic fetch_OP or OP_fetch operation: - TARGET is an option place to stick the return value. const0_rtx indicates - the result is unused. - atomically fetch MEM, perform the operation with VAL and return it to MEM. - CODE is the operation being performed (OP) - MEMMODEL is the memory model variant to use. - AFTER is true to return the result of the operation (OP_fetch). - AFTER is false to return the value before the operation (fetch_OP). */ -rtx -expand_atomic_fetch_op (rtx target, rtx mem, rtx val, enum rtx_code code, - enum memmodel model, bool after) -{ - machine_mode mode = GET_MODE (mem); - rtx result; - bool unused_result = (target == const0_rtx); - - /* If loads are not atomic for the required size and we are not called to - provide a __sync builtin, do not do anything so that we stay consistent - with atomic loads of the same size. */ - if (!can_atomic_load_p (mode) && !is_mm_sync (model)) - return NULL_RTX; - - result = expand_atomic_fetch_op_no_fallback (target, mem, val, code, model, - after); - - if (result) - return result; - - /* Add/sub can be implemented by doing the reverse operation with -(val). */ - if (code == PLUS || code == MINUS) - { - rtx tmp; - enum rtx_code reverse = (code == PLUS ? MINUS : PLUS); - - start_sequence (); - tmp = expand_simple_unop (mode, NEG, val, NULL_RTX, true); - result = expand_atomic_fetch_op_no_fallback (target, mem, tmp, reverse, - model, after); - if (result) - { - /* PLUS worked so emit the insns and return. */ - tmp = get_insns (); - end_sequence (); - emit_insn (tmp); - return result; - } - - /* PLUS did not work, so throw away the negation code and continue. */ - end_sequence (); - } - - /* Try the __sync libcalls only if we can't do compare-and-swap inline. */ - if (!can_compare_and_swap_p (mode, false)) - { - rtx libfunc; - bool fixup = false; - enum rtx_code orig_code = code; - struct atomic_op_functions optab; - - get_atomic_op_for_code (&optab, code); - libfunc = optab_libfunc (after ? optab.fetch_after - : optab.fetch_before, mode); - if (libfunc == NULL - && (after || unused_result || optab.reverse_code != UNKNOWN)) - { - fixup = true; - if (!after) - code = optab.reverse_code; - libfunc = optab_libfunc (after ? optab.fetch_before - : optab.fetch_after, mode); - } - if (libfunc != NULL) - { - rtx addr = convert_memory_address (ptr_mode, XEXP (mem, 0)); - result = emit_library_call_value (libfunc, NULL, LCT_NORMAL, mode, - addr, ptr_mode, val, mode); - - if (!unused_result && fixup) - result = expand_simple_binop (mode, code, result, val, target, - true, OPTAB_LIB_WIDEN); - return result; - } - - /* We need the original code for any further attempts. */ - code = orig_code; - } - - /* If nothing else has succeeded, default to a compare and swap loop. */ - if (can_compare_and_swap_p (mode, true)) - { - rtx_insn *insn; - rtx t0 = gen_reg_rtx (mode), t1; - - start_sequence (); - - /* If the result is used, get a register for it. */ - if (!unused_result) - { - if (!target || !register_operand (target, mode)) - target = gen_reg_rtx (mode); - /* If fetch_before, copy the value now. */ - if (!after) - emit_move_insn (target, t0); - } - else - target = const0_rtx; - - t1 = t0; - if (code == NOT) - { - t1 = expand_simple_binop (mode, AND, t1, val, NULL_RTX, - true, OPTAB_LIB_WIDEN); - t1 = expand_simple_unop (mode, code, t1, NULL_RTX, true); - } - else - t1 = expand_simple_binop (mode, code, t1, val, NULL_RTX, true, - OPTAB_LIB_WIDEN); - - /* For after, copy the value now. */ - if (!unused_result && after) - emit_move_insn (target, t1); - insn = get_insns (); - end_sequence (); - - if (t1 != NULL && expand_compare_and_swap_loop (mem, t0, t1, insn)) - return target; - } - - return NULL_RTX; -} - -/* Return true if OPERAND is suitable for operand number OPNO of - instruction ICODE. */ - -bool -insn_operand_matches (enum insn_code icode, unsigned int opno, rtx operand) -{ - return (!insn_data[(int) icode].operand[opno].predicate - || (insn_data[(int) icode].operand[opno].predicate - (operand, insn_data[(int) icode].operand[opno].mode))); -} - -/* TARGET is a target of a multiword operation that we are going to - implement as a series of word-mode operations. Return true if - TARGET is suitable for this purpose. */ - -bool -valid_multiword_target_p (rtx target) -{ - machine_mode mode; - int i, size; - - mode = GET_MODE (target); - if (!GET_MODE_SIZE (mode).is_constant (&size)) - return false; - for (i = 0; i < size; i += UNITS_PER_WORD) - if (!validate_subreg (word_mode, mode, target, i)) - return false; - return true; -} - -/* Make OP describe an input operand that has value INTVAL and that has - no inherent mode. This function should only be used for operands that - are always expand-time constants. The backend may request that INTVAL - be copied into a different kind of rtx, but it must specify the mode - of that rtx if so. */ - -void -create_integer_operand (class expand_operand *op, poly_int64 intval) -{ - create_expand_operand (op, EXPAND_INTEGER, - gen_int_mode (intval, MAX_MODE_INT), - VOIDmode, false, intval); -} - -/* Like maybe_legitimize_operand, but do not change the code of the - current rtx value. */ - -static bool -maybe_legitimize_operand_same_code (enum insn_code icode, unsigned int opno, - class expand_operand *op) -{ - /* See if the operand matches in its current form. */ - if (insn_operand_matches (icode, opno, op->value)) - return true; - - /* If the operand is a memory whose address has no side effects, - try forcing the address into a non-virtual pseudo register. - The check for side effects is important because copy_to_mode_reg - cannot handle things like auto-modified addresses. */ - if (insn_data[(int) icode].operand[opno].allows_mem && MEM_P (op->value)) - { - rtx addr, mem; - - mem = op->value; - addr = XEXP (mem, 0); - if (!(REG_P (addr) && REGNO (addr) > LAST_VIRTUAL_REGISTER) - && !side_effects_p (addr)) - { - rtx_insn *last; - machine_mode mode; - - last = get_last_insn (); - mode = get_address_mode (mem); - mem = replace_equiv_address (mem, copy_to_mode_reg (mode, addr)); - if (insn_operand_matches (icode, opno, mem)) - { - op->value = mem; - return true; - } - delete_insns_since (last); - } - } - - return false; -} - -/* Try to make OP match operand OPNO of instruction ICODE. Return true - on success, storing the new operand value back in OP. */ - -static bool -maybe_legitimize_operand (enum insn_code icode, unsigned int opno, - class expand_operand *op) -{ - machine_mode mode, imode, tmode; - - mode = op->mode; - switch (op->type) - { - case EXPAND_FIXED: - { - temporary_volatile_ok v (true); - return maybe_legitimize_operand_same_code (icode, opno, op); - } - - case EXPAND_OUTPUT: - gcc_assert (mode != VOIDmode); - if (op->value - && op->value != const0_rtx - && GET_MODE (op->value) == mode - && maybe_legitimize_operand_same_code (icode, opno, op)) - return true; - - op->value = gen_reg_rtx (mode); - op->target = 0; - break; - - case EXPAND_INPUT: - input: - gcc_assert (mode != VOIDmode); - gcc_assert (GET_MODE (op->value) == VOIDmode - || GET_MODE (op->value) == mode); - if (maybe_legitimize_operand_same_code (icode, opno, op)) - return true; - - op->value = copy_to_mode_reg (mode, op->value); - break; - - case EXPAND_CONVERT_TO: - gcc_assert (mode != VOIDmode); - op->value = convert_to_mode (mode, op->value, op->unsigned_p); - goto input; - - case EXPAND_CONVERT_FROM: - if (GET_MODE (op->value) != VOIDmode) - mode = GET_MODE (op->value); - else - /* The caller must tell us what mode this value has. */ - gcc_assert (mode != VOIDmode); - - imode = insn_data[(int) icode].operand[opno].mode; - tmode = (VECTOR_MODE_P (imode) && !VECTOR_MODE_P (mode) - ? GET_MODE_INNER (imode) : imode); - if (tmode != VOIDmode && tmode != mode) - { - op->value = convert_modes (tmode, mode, op->value, op->unsigned_p); - mode = tmode; - } - if (imode != VOIDmode && imode != mode) - { - gcc_assert (VECTOR_MODE_P (imode) && !VECTOR_MODE_P (mode)); - op->value = expand_vector_broadcast (imode, op->value); - mode = imode; - } - goto input; - - case EXPAND_ADDRESS: - op->value = convert_memory_address (as_a <scalar_int_mode> (mode), - op->value); - goto input; - - case EXPAND_INTEGER: - mode = insn_data[(int) icode].operand[opno].mode; - if (mode != VOIDmode - && known_eq (trunc_int_for_mode (op->int_value, mode), - op->int_value)) - { - op->value = gen_int_mode (op->int_value, mode); - goto input; - } - break; - } - return insn_operand_matches (icode, opno, op->value); -} - -/* Make OP describe an input operand that should have the same value - as VALUE, after any mode conversion that the target might request. - TYPE is the type of VALUE. */ - -void -create_convert_operand_from_type (class expand_operand *op, - rtx value, tree type) -{ - create_convert_operand_from (op, value, TYPE_MODE (type), - TYPE_UNSIGNED (type)); -} - -/* Return true if the requirements on operands OP1 and OP2 of instruction - ICODE are similar enough for the result of legitimizing OP1 to be - reusable for OP2. OPNO1 and OPNO2 are the operand numbers associated - with OP1 and OP2 respectively. */ - -static inline bool -can_reuse_operands_p (enum insn_code icode, - unsigned int opno1, unsigned int opno2, - const class expand_operand *op1, - const class expand_operand *op2) -{ - /* Check requirements that are common to all types. */ - if (op1->type != op2->type - || op1->mode != op2->mode - || (insn_data[(int) icode].operand[opno1].mode - != insn_data[(int) icode].operand[opno2].mode)) - return false; - - /* Check the requirements for specific types. */ - switch (op1->type) - { - case EXPAND_OUTPUT: - /* Outputs must remain distinct. */ - return false; - - case EXPAND_FIXED: - case EXPAND_INPUT: - case EXPAND_ADDRESS: - case EXPAND_INTEGER: - return true; - - case EXPAND_CONVERT_TO: - case EXPAND_CONVERT_FROM: - return op1->unsigned_p == op2->unsigned_p; - } - gcc_unreachable (); -} - -/* Try to make operands [OPS, OPS + NOPS) match operands [OPNO, OPNO + NOPS) - of instruction ICODE. Return true on success, leaving the new operand - values in the OPS themselves. Emit no code on failure. */ - -bool -maybe_legitimize_operands (enum insn_code icode, unsigned int opno, - unsigned int nops, class expand_operand *ops) -{ - rtx_insn *last = get_last_insn (); - rtx *orig_values = XALLOCAVEC (rtx, nops); - for (unsigned int i = 0; i < nops; i++) - { - orig_values[i] = ops[i].value; - - /* First try reusing the result of an earlier legitimization. - This avoids duplicate rtl and ensures that tied operands - remain tied. - - This search is linear, but NOPS is bounded at compile time - to a small number (current a single digit). */ - unsigned int j = 0; - for (; j < i; ++j) - if (can_reuse_operands_p (icode, opno + j, opno + i, &ops[j], &ops[i]) - && rtx_equal_p (orig_values[j], orig_values[i]) - && ops[j].value - && insn_operand_matches (icode, opno + i, ops[j].value)) - { - ops[i].value = copy_rtx (ops[j].value); - break; - } - - /* Otherwise try legitimizing the operand on its own. */ - if (j == i && !maybe_legitimize_operand (icode, opno + i, &ops[i])) - { - delete_insns_since (last); - return false; - } - } - return true; -} - -/* Try to generate instruction ICODE, using operands [OPS, OPS + NOPS) - as its operands. Return the instruction pattern on success, - and emit any necessary set-up code. Return null and emit no - code on failure. */ - -rtx_insn * -maybe_gen_insn (enum insn_code icode, unsigned int nops, - class expand_operand *ops) -{ - gcc_assert (nops == (unsigned int) insn_data[(int) icode].n_generator_args); - if (!maybe_legitimize_operands (icode, 0, nops, ops)) - return NULL; - - switch (nops) - { - case 1: - return GEN_FCN (icode) (ops[0].value); - case 2: - return GEN_FCN (icode) (ops[0].value, ops[1].value); - case 3: - return GEN_FCN (icode) (ops[0].value, ops[1].value, ops[2].value); - case 4: - return GEN_FCN (icode) (ops[0].value, ops[1].value, ops[2].value, - ops[3].value); - case 5: - return GEN_FCN (icode) (ops[0].value, ops[1].value, ops[2].value, - ops[3].value, ops[4].value); - case 6: - return GEN_FCN (icode) (ops[0].value, ops[1].value, ops[2].value, - ops[3].value, ops[4].value, ops[5].value); - case 7: - return GEN_FCN (icode) (ops[0].value, ops[1].value, ops[2].value, - ops[3].value, ops[4].value, ops[5].value, - ops[6].value); - case 8: - return GEN_FCN (icode) (ops[0].value, ops[1].value, ops[2].value, - ops[3].value, ops[4].value, ops[5].value, - ops[6].value, ops[7].value); - case 9: - return GEN_FCN (icode) (ops[0].value, ops[1].value, ops[2].value, - ops[3].value, ops[4].value, ops[5].value, - ops[6].value, ops[7].value, ops[8].value); - } - gcc_unreachable (); -} - -/* Try to emit instruction ICODE, using operands [OPS, OPS + NOPS) - as its operands. Return true on success and emit no code on failure. */ - -bool -maybe_expand_insn (enum insn_code icode, unsigned int nops, - class expand_operand *ops) -{ - rtx_insn *pat = maybe_gen_insn (icode, nops, ops); - if (pat) - { - emit_insn (pat); - return true; - } - return false; -} - -/* Like maybe_expand_insn, but for jumps. */ - -bool -maybe_expand_jump_insn (enum insn_code icode, unsigned int nops, - class expand_operand *ops) -{ - rtx_insn *pat = maybe_gen_insn (icode, nops, ops); - if (pat) - { - emit_jump_insn (pat); - return true; - } - return false; -} - -/* Emit instruction ICODE, using operands [OPS, OPS + NOPS) - as its operands. */ - -void -expand_insn (enum insn_code icode, unsigned int nops, - class expand_operand *ops) -{ - if (!maybe_expand_insn (icode, nops, ops)) - gcc_unreachable (); -} - -/* Like expand_insn, but for jumps. */ - -void -expand_jump_insn (enum insn_code icode, unsigned int nops, - class expand_operand *ops) -{ - if (!maybe_expand_jump_insn (icode, nops, ops)) - gcc_unreachable (); -} |