; NOTE: Assertions have been autogenerated by utils/update_llc_test_checks.py UTC_ARGS: --version 5 ; RUN: llc < %s -mtriple=x86_64-unknown-unknown -verify-machineinstrs | FileCheck %s --check-prefixes=SSE2 ; RUN: llc < %s -mtriple=x86_64-unknown-unknown -mattr=+avx -verify-machineinstrs | FileCheck %s --check-prefixes=AVX ; RUN: llc < %s -mtriple=x86_64-unknown-unknown -mattr=+avx512f -verify-machineinstrs | FileCheck %s --check-prefixes=AVX512 define half @test(float %f, ptr %p) nounwind { ; SSE2-LABEL: test: ; SSE2: # %bb.0: ; SSE2-NEXT: pushq %rbx ; SSE2-NEXT: subq $16, %rsp ; SSE2-NEXT: movq %rdi, %rbx ; SSE2-NEXT: callq __truncsfhf2@PLT ; SSE2-NEXT: movss %xmm0, {{[-0-9]+}}(%r{{[sb]}}p) # 4-byte Spill ; SSE2-NEXT: callq __extendhfsf2@PLT ; SSE2-NEXT: movss %xmm0, {{[-0-9]+}}(%r{{[sb]}}p) # 4-byte Spill ; SSE2-NEXT: #APP ; SSE2-NEXT: #NO_APP ; SSE2-NEXT: movss {{[-0-9]+}}(%r{{[sb]}}p), %xmm0 # 4-byte Reload ; SSE2-NEXT: # xmm0 = mem[0],zero,zero,zero ; SSE2-NEXT: movss %xmm0, (%rbx) ; SSE2-NEXT: movss {{[-0-9]+}}(%r{{[sb]}}p), %xmm0 # 4-byte Reload ; SSE2-NEXT: # xmm0 = mem[0],zero,zero,zero ; SSE2-NEXT: addq $16, %rsp ; SSE2-NEXT: popq %rbx ; SSE2-NEXT: retq ; ; AVX-LABEL: test: ; AVX: # %bb.0: ; AVX-NEXT: pushq %rbx ; AVX-NEXT: subq $16, %rsp ; AVX-NEXT: movq %rdi, %rbx ; AVX-NEXT: callq __truncsfhf2@PLT ; AVX-NEXT: vmovss %xmm0, {{[-0-9]+}}(%r{{[sb]}}p) # 4-byte Spill ; AVX-NEXT: callq __extendhfsf2@PLT ; AVX-NEXT: vmovss %xmm0, {{[-0-9]+}}(%r{{[sb]}}p) # 4-byte Spill ; AVX-NEXT: #APP ; AVX-NEXT: #NO_APP ; AVX-NEXT: vmovss {{[-0-9]+}}(%r{{[sb]}}p), %xmm0 # 4-byte Reload ; AVX-NEXT: # xmm0 = mem[0],zero,zero,zero ; AVX-NEXT: vmovss %xmm0, (%rbx) ; AVX-NEXT: vmovss {{[-0-9]+}}(%r{{[sb]}}p), %xmm0 # 4-byte Reload ; AVX-NEXT: # xmm0 = mem[0],zero,zero,zero ; AVX-NEXT: addq $16, %rsp ; AVX-NEXT: popq %rbx ; AVX-NEXT: retq ; ; AVX512-LABEL: test: ; AVX512: # %bb.0: ; AVX512-NEXT: vcvtps2ph $4, %xmm0, %xmm0 ; AVX512-NEXT: vmovaps %xmm0, {{[-0-9]+}}(%r{{[sb]}}p) # 16-byte Spill ; AVX512-NEXT: vcvtph2ps %xmm0, %xmm0 ; AVX512-NEXT: vmovaps %xmm0, {{[-0-9]+}}(%r{{[sb]}}p) # 16-byte Spill ; AVX512-NEXT: #APP ; AVX512-NEXT: #NO_APP ; AVX512-NEXT: vmovaps {{[-0-9]+}}(%r{{[sb]}}p), %xmm0 # 16-byte Reload ; AVX512-NEXT: vmovss %xmm0, (%rdi) ; AVX512-NEXT: vmovaps {{[-0-9]+}}(%r{{[sb]}}p), %xmm0 # 16-byte Reload ; AVX512-NEXT: retq %t = fptrunc float %f to half %t2 = fpext half %t to float tail call void asm sideeffect "", "~{xmm0},~{xmm1},~{xmm2},~{xmm3},~{xmm4},~{xmm5},~{xmm6},~{xmm7},~{xmm8},~{xmm9},~{xmm10},~{xmm11},~{xmm12},~{xmm13},~{xmm14},~{xmm15},~{dirflag},~{fpsr},~{flags}"() store float %t2, ptr %p ret half %t }