addmul_1.asm
上传用户:qaz666999
上传日期:2022-08-06
资源大小:2570k
文件大小:4k
- dnl mpn_addmul_1 for Pentium 4 and P6 models with SSE2 (i.e., 9,D,E,F).
- dnl Copyright 2005, 2007 Free Software Foundation, Inc.
- dnl
- dnl This file is part of the GNU MP Library.
- dnl
- dnl The GNU MP Library is free software; you can redistribute it and/or modify
- dnl it under the terms of the GNU Lesser General Public License as published
- dnl by the Free Software Foundation; either version 3 of the License, or (at
- dnl your option) any later version.
- dnl
- dnl The GNU MP Library is distributed in the hope that it will be useful, but
- dnl WITHOUT ANY WARRANTY; without even the implied warranty of MERCHANTABILITY
- dnl or FITNESS FOR A PARTICULAR PURPOSE. See the GNU Lesser General Public
- dnl License for more details.
- dnl
- dnl You should have received a copy of the GNU Lesser General Public License
- dnl along with the GNU MP Library. If not, see http://www.gnu.org/licenses/.
- include(`../config.m4')
- C TODO:
- C * Tweak eax/edx offsets in loop as to save some lea's
- C * Perhaps software pipeline small-case code
- C cycles/limb
- C P6 model 0-8,10-12) -
- C P6 model 9 (Banias) ?
- C P6 model 13 (Dothan) 5.24
- C P4 model 0-1 (Willamette): 5
- C P4 model 2 (Northwood): 5
- C P4 model 3-4 (Prescott): 5
- C INPUT PARAMETERS
- C rp sp + 4
- C up sp + 8
- C n sp + 12
- C v0 sp + 16
- TEXT
- ALIGN(16)
- PROLOGUE(mpn_addmul_1c)
- mov 4(%esp), %edx
- mov 8(%esp), %eax
- mov 12(%esp), %ecx
- movd 16(%esp), %mm7
- movd 20(%esp), %mm6
- jmp L(ent)
- EPILOGUE()
- ALIGN(16)
- PROLOGUE(mpn_addmul_1)
- mov 4(%esp), %edx
- mov 8(%esp), %eax
- mov 12(%esp), %ecx
- movd 16(%esp), %mm7
- pxor %mm6, %mm6
- L(ent): cmp $4, %ecx
- jnc L(big)
- L(lp0): movd (%eax), %mm0
- lea 4(%eax), %eax
- movd (%edx), %mm4
- lea 4(%edx), %edx
- pmuludq %mm7, %mm0
- paddq %mm0, %mm4
- paddq %mm4, %mm6
- movd %mm6, -4(%edx)
- psrlq $32, %mm6
- dec %ecx
- jnz L(lp0)
- movd %mm6, %eax
- emms
- ret
- L(big): and $3, %ecx
- je L(0)
- cmp $2, %ecx
- jc L(1)
- je L(2)
- jmp L(3) C FIXME: one case should fall through
- L(0): movd (%eax), %mm3
- sub 12(%esp), %ecx C loop count
- lea -16(%eax), %eax
- lea -12(%edx), %edx
- pmuludq %mm7, %mm3
- movd 20(%eax), %mm0
- movd 12(%edx), %mm5
- pmuludq %mm7, %mm0
- movd 24(%eax), %mm1
- paddq %mm3, %mm5
- movd 16(%edx), %mm4
- jmp L(00)
- L(1): movd (%eax), %mm2
- sub 12(%esp), %ecx
- lea -12(%eax), %eax
- lea -8(%edx), %edx
- movd 8(%edx), %mm4
- pmuludq %mm7, %mm2
- movd 16(%eax), %mm3
- pmuludq %mm7, %mm3
- movd 20(%eax), %mm0
- paddq %mm2, %mm4
- movd 12(%edx), %mm5
- jmp L(01)
- L(2): movd (%eax), %mm1
- sub 12(%esp), %ecx
- lea -8(%eax), %eax
- lea -4(%edx), %edx
- pmuludq %mm7, %mm1
- movd 12(%eax), %mm2
- movd 4(%edx), %mm5
- pmuludq %mm7, %mm2
- movd 16(%eax), %mm3
- paddq %mm1, %mm5
- movd 8(%edx), %mm4
- jmp L(10)
- L(3): movd (%eax), %mm0
- sub 12(%esp), %ecx
- lea -4(%eax), %eax
- pmuludq %mm7, %mm0
- movd 8(%eax), %mm1
- movd (%edx), %mm4
- pmuludq %mm7, %mm1
- movd 12(%eax), %mm2
- paddq %mm0, %mm4
- movd 4(%edx), %mm5
- ALIGN(16)
- L(top): pmuludq %mm7, %mm2
- paddq %mm4, %mm6
- movd 16(%eax), %mm3
- paddq %mm1, %mm5
- movd 8(%edx), %mm4
- movd %mm6, 0(%edx)
- psrlq $32, %mm6
- L(10): pmuludq %mm7, %mm3
- paddq %mm5, %mm6
- movd 20(%eax), %mm0
- paddq %mm2, %mm4
- movd 12(%edx), %mm5
- movd %mm6, 4(%edx)
- psrlq $32, %mm6
- L(01): pmuludq %mm7, %mm0
- paddq %mm4, %mm6
- movd 24(%eax), %mm1
- paddq %mm3, %mm5
- movd 16(%edx), %mm4
- movd %mm6, 8(%edx)
- psrlq $32, %mm6
- L(00): pmuludq %mm7, %mm1
- paddq %mm5, %mm6
- movd 28(%eax), %mm2
- paddq %mm0, %mm4
- movd 20(%edx), %mm5
- movd %mm6, 12(%edx)
- psrlq $32, %mm6
- lea 16(%eax), %eax
- lea 16(%edx), %edx
- add $4, %ecx
- jnz L(top)
- L(end): pmuludq %mm7, %mm2
- paddq %mm4, %mm6
- paddq %mm1, %mm5
- movd 8(%edx), %mm4
- movd %mm6, 0(%edx)
- psrlq $32, %mm6
- paddq %mm5, %mm6
- paddq %mm2, %mm4
- movd %mm6, 4(%edx)
- psrlq $32, %mm6
- paddq %mm4, %mm6
- movd %mm6, 8(%edx)
- psrlq $32, %mm6
- movd %mm6, %eax
- emms
- ret
- EPILOGUE()