| /* Wrapper implementations of vector math functions. |
| Copyright (C) 2014-2018 Free Software Foundation, Inc. |
| This file is part of the GNU C Library. |
| |
| The GNU C Library is free software; you can redistribute it and/or |
| modify it under the terms of the GNU Lesser General Public |
| License as published by the Free Software Foundation; either |
| version 2.1 of the License, or (at your option) any later version. |
| |
| The GNU C Library is distributed in the hope that it will be useful, |
| but WITHOUT ANY WARRANTY; without even the implied warranty of |
| MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU |
| Lesser General Public License for more details. |
| |
| You should have received a copy of the GNU Lesser General Public |
| License along with the GNU C Library; if not, see |
| <http://www.gnu.org/licenses/>. */ |
| |
| /* SSE2 ISA version as wrapper to scalar. */ |
| .macro WRAPPER_IMPL_SSE2 callee |
| subq $40, %rsp |
| cfi_adjust_cfa_offset(40) |
| movaps %xmm0, (%rsp) |
| call JUMPTARGET(\callee) |
| movsd %xmm0, 16(%rsp) |
| movsd 8(%rsp), %xmm0 |
| call JUMPTARGET(\callee) |
| movsd 16(%rsp), %xmm1 |
| movsd %xmm0, 24(%rsp) |
| unpcklpd %xmm0, %xmm1 |
| movaps %xmm1, %xmm0 |
| addq $40, %rsp |
| cfi_adjust_cfa_offset(-40) |
| ret |
| .endm |
| |
| /* 2 argument SSE2 ISA version as wrapper to scalar. */ |
| .macro WRAPPER_IMPL_SSE2_ff callee |
| subq $56, %rsp |
| cfi_adjust_cfa_offset(56) |
| movaps %xmm0, (%rsp) |
| movaps %xmm1, 16(%rsp) |
| call JUMPTARGET(\callee) |
| movsd %xmm0, 32(%rsp) |
| movsd 8(%rsp), %xmm0 |
| movsd 24(%rsp), %xmm1 |
| call JUMPTARGET(\callee) |
| movsd 32(%rsp), %xmm1 |
| movsd %xmm0, 40(%rsp) |
| unpcklpd %xmm0, %xmm1 |
| movaps %xmm1, %xmm0 |
| addq $56, %rsp |
| cfi_adjust_cfa_offset(-56) |
| ret |
| .endm |
| |
| /* 3 argument SSE2 ISA version as wrapper to scalar. */ |
| .macro WRAPPER_IMPL_SSE2_fFF callee |
| pushq %rbp |
| cfi_adjust_cfa_offset (8) |
| cfi_rel_offset (%rbp, 0) |
| pushq %rbx |
| cfi_adjust_cfa_offset (8) |
| cfi_rel_offset (%rbx, 0) |
| movq %rdi, %rbp |
| movq %rsi, %rbx |
| subq $40, %rsp |
| cfi_adjust_cfa_offset(40) |
| leaq 16(%rsp), %rsi |
| leaq 24(%rsp), %rdi |
| movaps %xmm0, (%rsp) |
| call JUMPTARGET(\callee) |
| leaq 16(%rsp), %rsi |
| leaq 24(%rsp), %rdi |
| movsd 24(%rsp), %xmm0 |
| movapd (%rsp), %xmm1 |
| movsd %xmm0, 0(%rbp) |
| unpckhpd %xmm1, %xmm1 |
| movsd 16(%rsp), %xmm0 |
| movsd %xmm0, (%rbx) |
| movapd %xmm1, %xmm0 |
| call JUMPTARGET(\callee) |
| movsd 24(%rsp), %xmm0 |
| movsd %xmm0, 8(%rbp) |
| movsd 16(%rsp), %xmm0 |
| movsd %xmm0, 8(%rbx) |
| addq $40, %rsp |
| cfi_adjust_cfa_offset(-40) |
| popq %rbx |
| cfi_adjust_cfa_offset (-8) |
| cfi_restore (%rbx) |
| popq %rbp |
| cfi_adjust_cfa_offset (-8) |
| cfi_restore (%rbp) |
| ret |
| .endm |
| |
| /* AVX/AVX2 ISA version as wrapper to SSE ISA version. */ |
| .macro WRAPPER_IMPL_AVX callee |
| pushq %rbp |
| cfi_adjust_cfa_offset (8) |
| cfi_rel_offset (%rbp, 0) |
| movq %rsp, %rbp |
| cfi_def_cfa_register (%rbp) |
| andq $-32, %rsp |
| subq $32, %rsp |
| vextractf128 $1, %ymm0, (%rsp) |
| vzeroupper |
| call HIDDEN_JUMPTARGET(\callee) |
| vmovapd %xmm0, 16(%rsp) |
| vmovaps (%rsp), %xmm0 |
| call HIDDEN_JUMPTARGET(\callee) |
| vmovapd %xmm0, %xmm1 |
| vmovapd 16(%rsp), %xmm0 |
| vinsertf128 $1, %xmm1, %ymm0, %ymm0 |
| movq %rbp, %rsp |
| cfi_def_cfa_register (%rsp) |
| popq %rbp |
| cfi_adjust_cfa_offset (-8) |
| cfi_restore (%rbp) |
| ret |
| .endm |
| |
| /* 2 argument AVX/AVX2 ISA version as wrapper to SSE ISA version. */ |
| .macro WRAPPER_IMPL_AVX_ff callee |
| pushq %rbp |
| cfi_adjust_cfa_offset (8) |
| cfi_rel_offset (%rbp, 0) |
| movq %rsp, %rbp |
| cfi_def_cfa_register (%rbp) |
| andq $-32, %rsp |
| subq $64, %rsp |
| vextractf128 $1, %ymm0, 16(%rsp) |
| vextractf128 $1, %ymm1, (%rsp) |
| vzeroupper |
| call HIDDEN_JUMPTARGET(\callee) |
| vmovaps %xmm0, 32(%rsp) |
| vmovaps 16(%rsp), %xmm0 |
| vmovaps (%rsp), %xmm1 |
| call HIDDEN_JUMPTARGET(\callee) |
| vmovaps %xmm0, %xmm1 |
| vmovaps 32(%rsp), %xmm0 |
| vinsertf128 $1, %xmm1, %ymm0, %ymm0 |
| movq %rbp, %rsp |
| cfi_def_cfa_register (%rsp) |
| popq %rbp |
| cfi_adjust_cfa_offset (-8) |
| cfi_restore (%rbp) |
| ret |
| .endm |
| |
| /* 3 argument AVX/AVX2 ISA version as wrapper to SSE ISA version. */ |
| .macro WRAPPER_IMPL_AVX_fFF callee |
| pushq %rbp |
| cfi_adjust_cfa_offset (8) |
| cfi_rel_offset (%rbp, 0) |
| movq %rsp, %rbp |
| cfi_def_cfa_register (%rbp) |
| andq $-32, %rsp |
| pushq %r13 |
| cfi_adjust_cfa_offset (8) |
| cfi_rel_offset (%r13, 0) |
| pushq %r14 |
| cfi_adjust_cfa_offset (8) |
| cfi_rel_offset (%r14, 0) |
| subq $48, %rsp |
| movq %rsi, %r14 |
| movq %rdi, %r13 |
| vextractf128 $1, %ymm0, 32(%rsp) |
| vzeroupper |
| call HIDDEN_JUMPTARGET(\callee) |
| vmovaps 32(%rsp), %xmm0 |
| lea (%rsp), %rdi |
| lea 16(%rsp), %rsi |
| call HIDDEN_JUMPTARGET(\callee) |
| vmovapd (%rsp), %xmm0 |
| vmovapd 16(%rsp), %xmm1 |
| vmovapd %xmm0, 16(%r13) |
| vmovapd %xmm1, 16(%r14) |
| addq $48, %rsp |
| popq %r14 |
| cfi_adjust_cfa_offset (-8) |
| cfi_restore (%r14) |
| popq %r13 |
| cfi_adjust_cfa_offset (-8) |
| cfi_restore (%r13) |
| movq %rbp, %rsp |
| cfi_def_cfa_register (%rsp) |
| popq %rbp |
| cfi_adjust_cfa_offset (-8) |
| cfi_restore (%rbp) |
| ret |
| .endm |
| |
| /* AVX512 ISA version as wrapper to AVX2 ISA version. */ |
| .macro WRAPPER_IMPL_AVX512 callee |
| pushq %rbp |
| cfi_adjust_cfa_offset (8) |
| cfi_rel_offset (%rbp, 0) |
| movq %rsp, %rbp |
| cfi_def_cfa_register (%rbp) |
| andq $-64, %rsp |
| subq $128, %rsp |
| vmovups %zmm0, (%rsp) |
| vmovupd (%rsp), %ymm0 |
| call HIDDEN_JUMPTARGET(\callee) |
| vmovupd %ymm0, 64(%rsp) |
| vmovupd 32(%rsp), %ymm0 |
| call HIDDEN_JUMPTARGET(\callee) |
| vmovupd %ymm0, 96(%rsp) |
| vmovups 64(%rsp), %zmm0 |
| movq %rbp, %rsp |
| cfi_def_cfa_register (%rsp) |
| popq %rbp |
| cfi_adjust_cfa_offset (-8) |
| cfi_restore (%rbp) |
| ret |
| .endm |
| |
| /* 2 argument AVX512 ISA version as wrapper to AVX2 ISA version. */ |
| .macro WRAPPER_IMPL_AVX512_ff callee |
| pushq %rbp |
| cfi_adjust_cfa_offset (8) |
| cfi_rel_offset (%rbp, 0) |
| movq %rsp, %rbp |
| cfi_def_cfa_register (%rbp) |
| andq $-64, %rsp |
| subq $192, %rsp |
| vmovups %zmm0, (%rsp) |
| vmovups %zmm1, 64(%rsp) |
| vmovupd (%rsp), %ymm0 |
| vmovupd 64(%rsp), %ymm1 |
| call HIDDEN_JUMPTARGET(\callee) |
| vmovupd %ymm0, 128(%rsp) |
| vmovupd 32(%rsp), %ymm0 |
| vmovupd 96(%rsp), %ymm1 |
| call HIDDEN_JUMPTARGET(\callee) |
| vmovupd %ymm0, 160(%rsp) |
| vmovups 128(%rsp), %zmm0 |
| movq %rbp, %rsp |
| cfi_def_cfa_register (%rsp) |
| popq %rbp |
| cfi_adjust_cfa_offset (-8) |
| cfi_restore (%rbp) |
| ret |
| .endm |
| |
| /* 3 argument AVX512 ISA version as wrapper to AVX2 ISA version. */ |
| .macro WRAPPER_IMPL_AVX512_fFF callee |
| pushq %rbp |
| cfi_adjust_cfa_offset (8) |
| cfi_rel_offset (%rbp, 0) |
| movq %rsp, %rbp |
| cfi_def_cfa_register (%rbp) |
| andq $-64, %rsp |
| pushq %r12 |
| cfi_adjust_cfa_offset (8) |
| cfi_rel_offset (%r12, 0) |
| pushq %r13 |
| cfi_adjust_cfa_offset (8) |
| cfi_rel_offset (%r13, 0) |
| subq $176, %rsp |
| movq %rsi, %r13 |
| vmovups %zmm0, (%rsp) |
| movq %rdi, %r12 |
| vmovupd (%rsp), %ymm0 |
| call HIDDEN_JUMPTARGET(\callee) |
| vmovupd 32(%rsp), %ymm0 |
| lea 64(%rsp), %rdi |
| lea 96(%rsp), %rsi |
| call HIDDEN_JUMPTARGET(\callee) |
| vmovupd 64(%rsp), %ymm0 |
| vmovupd 96(%rsp), %ymm1 |
| vmovupd %ymm0, 32(%r12) |
| vmovupd %ymm1, 32(%r13) |
| vzeroupper |
| addq $176, %rsp |
| popq %r13 |
| cfi_adjust_cfa_offset (-8) |
| cfi_restore (%r13) |
| popq %r12 |
| cfi_adjust_cfa_offset (-8) |
| cfi_restore (%r12) |
| movq %rbp, %rsp |
| cfi_def_cfa_register (%rsp) |
| popq %rbp |
| cfi_adjust_cfa_offset (-8) |
| cfi_restore (%rbp) |
| ret |
| .endm |