mirror of
https://github.com/klzgrad/naiveproxy.git
synced 2024-11-28 16:26:10 +03:00
1067 lines
19 KiB
NASM
1067 lines
19 KiB
NASM
%ifidn __OUTPUT_FORMAT__,obj
|
|
section code use32 class=code align=64
|
|
%elifidn __OUTPUT_FORMAT__,win32
|
|
%ifdef __YASM_VERSION_ID__
|
|
%if __YASM_VERSION_ID__ < 01010000h
|
|
%error yasm version 1.1.0 or later needed.
|
|
%endif
|
|
; Yasm automatically includes .00 and complains about redefining it.
|
|
; https://www.tortall.net/projects/yasm/manual/html/objfmt-win32-safeseh.html
|
|
%else
|
|
$@feat.00 equ 1
|
|
%endif
|
|
section .text code align=64
|
|
%else
|
|
section .text code
|
|
%endif
|
|
global _gcm_gmult_4bit_mmx
|
|
align 16
|
|
_gcm_gmult_4bit_mmx:
|
|
L$_gcm_gmult_4bit_mmx_begin:
|
|
push ebp
|
|
push ebx
|
|
push esi
|
|
push edi
|
|
mov edi,DWORD [20+esp]
|
|
mov esi,DWORD [24+esp]
|
|
call L$000pic_point
|
|
L$000pic_point:
|
|
pop eax
|
|
lea eax,[(L$rem_4bit-L$000pic_point)+eax]
|
|
movzx ebx,BYTE [15+edi]
|
|
xor ecx,ecx
|
|
mov edx,ebx
|
|
mov cl,dl
|
|
mov ebp,14
|
|
shl cl,4
|
|
and edx,240
|
|
movq mm0,[8+ecx*1+esi]
|
|
movq mm1,[ecx*1+esi]
|
|
movd ebx,mm0
|
|
jmp NEAR L$001mmx_loop
|
|
align 16
|
|
L$001mmx_loop:
|
|
psrlq mm0,4
|
|
and ebx,15
|
|
movq mm2,mm1
|
|
psrlq mm1,4
|
|
pxor mm0,[8+edx*1+esi]
|
|
mov cl,BYTE [ebp*1+edi]
|
|
psllq mm2,60
|
|
pxor mm1,[ebx*8+eax]
|
|
dec ebp
|
|
movd ebx,mm0
|
|
pxor mm1,[edx*1+esi]
|
|
mov edx,ecx
|
|
pxor mm0,mm2
|
|
js NEAR L$002mmx_break
|
|
shl cl,4
|
|
and ebx,15
|
|
psrlq mm0,4
|
|
and edx,240
|
|
movq mm2,mm1
|
|
psrlq mm1,4
|
|
pxor mm0,[8+ecx*1+esi]
|
|
psllq mm2,60
|
|
pxor mm1,[ebx*8+eax]
|
|
movd ebx,mm0
|
|
pxor mm1,[ecx*1+esi]
|
|
pxor mm0,mm2
|
|
jmp NEAR L$001mmx_loop
|
|
align 16
|
|
L$002mmx_break:
|
|
shl cl,4
|
|
and ebx,15
|
|
psrlq mm0,4
|
|
and edx,240
|
|
movq mm2,mm1
|
|
psrlq mm1,4
|
|
pxor mm0,[8+ecx*1+esi]
|
|
psllq mm2,60
|
|
pxor mm1,[ebx*8+eax]
|
|
movd ebx,mm0
|
|
pxor mm1,[ecx*1+esi]
|
|
pxor mm0,mm2
|
|
psrlq mm0,4
|
|
and ebx,15
|
|
movq mm2,mm1
|
|
psrlq mm1,4
|
|
pxor mm0,[8+edx*1+esi]
|
|
psllq mm2,60
|
|
pxor mm1,[ebx*8+eax]
|
|
movd ebx,mm0
|
|
pxor mm1,[edx*1+esi]
|
|
pxor mm0,mm2
|
|
psrlq mm0,32
|
|
movd edx,mm1
|
|
psrlq mm1,32
|
|
movd ecx,mm0
|
|
movd ebp,mm1
|
|
bswap ebx
|
|
bswap edx
|
|
bswap ecx
|
|
bswap ebp
|
|
emms
|
|
mov DWORD [12+edi],ebx
|
|
mov DWORD [4+edi],edx
|
|
mov DWORD [8+edi],ecx
|
|
mov DWORD [edi],ebp
|
|
pop edi
|
|
pop esi
|
|
pop ebx
|
|
pop ebp
|
|
ret
|
|
global _gcm_ghash_4bit_mmx
|
|
align 16
|
|
_gcm_ghash_4bit_mmx:
|
|
L$_gcm_ghash_4bit_mmx_begin:
|
|
push ebp
|
|
push ebx
|
|
push esi
|
|
push edi
|
|
mov eax,DWORD [20+esp]
|
|
mov ebx,DWORD [24+esp]
|
|
mov ecx,DWORD [28+esp]
|
|
mov edx,DWORD [32+esp]
|
|
mov ebp,esp
|
|
call L$003pic_point
|
|
L$003pic_point:
|
|
pop esi
|
|
lea esi,[(L$rem_8bit-L$003pic_point)+esi]
|
|
sub esp,544
|
|
and esp,-64
|
|
sub esp,16
|
|
add edx,ecx
|
|
mov DWORD [544+esp],eax
|
|
mov DWORD [552+esp],edx
|
|
mov DWORD [556+esp],ebp
|
|
add ebx,128
|
|
lea edi,[144+esp]
|
|
lea ebp,[400+esp]
|
|
mov edx,DWORD [ebx-120]
|
|
movq mm0,[ebx-120]
|
|
movq mm3,[ebx-128]
|
|
shl edx,4
|
|
mov BYTE [esp],dl
|
|
mov edx,DWORD [ebx-104]
|
|
movq mm2,[ebx-104]
|
|
movq mm5,[ebx-112]
|
|
movq [edi-128],mm0
|
|
psrlq mm0,4
|
|
movq [edi],mm3
|
|
movq mm7,mm3
|
|
psrlq mm3,4
|
|
shl edx,4
|
|
mov BYTE [1+esp],dl
|
|
mov edx,DWORD [ebx-88]
|
|
movq mm1,[ebx-88]
|
|
psllq mm7,60
|
|
movq mm4,[ebx-96]
|
|
por mm0,mm7
|
|
movq [edi-120],mm2
|
|
psrlq mm2,4
|
|
movq [8+edi],mm5
|
|
movq mm6,mm5
|
|
movq [ebp-128],mm0
|
|
psrlq mm5,4
|
|
movq [ebp],mm3
|
|
shl edx,4
|
|
mov BYTE [2+esp],dl
|
|
mov edx,DWORD [ebx-72]
|
|
movq mm0,[ebx-72]
|
|
psllq mm6,60
|
|
movq mm3,[ebx-80]
|
|
por mm2,mm6
|
|
movq [edi-112],mm1
|
|
psrlq mm1,4
|
|
movq [16+edi],mm4
|
|
movq mm7,mm4
|
|
movq [ebp-120],mm2
|
|
psrlq mm4,4
|
|
movq [8+ebp],mm5
|
|
shl edx,4
|
|
mov BYTE [3+esp],dl
|
|
mov edx,DWORD [ebx-56]
|
|
movq mm2,[ebx-56]
|
|
psllq mm7,60
|
|
movq mm5,[ebx-64]
|
|
por mm1,mm7
|
|
movq [edi-104],mm0
|
|
psrlq mm0,4
|
|
movq [24+edi],mm3
|
|
movq mm6,mm3
|
|
movq [ebp-112],mm1
|
|
psrlq mm3,4
|
|
movq [16+ebp],mm4
|
|
shl edx,4
|
|
mov BYTE [4+esp],dl
|
|
mov edx,DWORD [ebx-40]
|
|
movq mm1,[ebx-40]
|
|
psllq mm6,60
|
|
movq mm4,[ebx-48]
|
|
por mm0,mm6
|
|
movq [edi-96],mm2
|
|
psrlq mm2,4
|
|
movq [32+edi],mm5
|
|
movq mm7,mm5
|
|
movq [ebp-104],mm0
|
|
psrlq mm5,4
|
|
movq [24+ebp],mm3
|
|
shl edx,4
|
|
mov BYTE [5+esp],dl
|
|
mov edx,DWORD [ebx-24]
|
|
movq mm0,[ebx-24]
|
|
psllq mm7,60
|
|
movq mm3,[ebx-32]
|
|
por mm2,mm7
|
|
movq [edi-88],mm1
|
|
psrlq mm1,4
|
|
movq [40+edi],mm4
|
|
movq mm6,mm4
|
|
movq [ebp-96],mm2
|
|
psrlq mm4,4
|
|
movq [32+ebp],mm5
|
|
shl edx,4
|
|
mov BYTE [6+esp],dl
|
|
mov edx,DWORD [ebx-8]
|
|
movq mm2,[ebx-8]
|
|
psllq mm6,60
|
|
movq mm5,[ebx-16]
|
|
por mm1,mm6
|
|
movq [edi-80],mm0
|
|
psrlq mm0,4
|
|
movq [48+edi],mm3
|
|
movq mm7,mm3
|
|
movq [ebp-88],mm1
|
|
psrlq mm3,4
|
|
movq [40+ebp],mm4
|
|
shl edx,4
|
|
mov BYTE [7+esp],dl
|
|
mov edx,DWORD [8+ebx]
|
|
movq mm1,[8+ebx]
|
|
psllq mm7,60
|
|
movq mm4,[ebx]
|
|
por mm0,mm7
|
|
movq [edi-72],mm2
|
|
psrlq mm2,4
|
|
movq [56+edi],mm5
|
|
movq mm6,mm5
|
|
movq [ebp-80],mm0
|
|
psrlq mm5,4
|
|
movq [48+ebp],mm3
|
|
shl edx,4
|
|
mov BYTE [8+esp],dl
|
|
mov edx,DWORD [24+ebx]
|
|
movq mm0,[24+ebx]
|
|
psllq mm6,60
|
|
movq mm3,[16+ebx]
|
|
por mm2,mm6
|
|
movq [edi-64],mm1
|
|
psrlq mm1,4
|
|
movq [64+edi],mm4
|
|
movq mm7,mm4
|
|
movq [ebp-72],mm2
|
|
psrlq mm4,4
|
|
movq [56+ebp],mm5
|
|
shl edx,4
|
|
mov BYTE [9+esp],dl
|
|
mov edx,DWORD [40+ebx]
|
|
movq mm2,[40+ebx]
|
|
psllq mm7,60
|
|
movq mm5,[32+ebx]
|
|
por mm1,mm7
|
|
movq [edi-56],mm0
|
|
psrlq mm0,4
|
|
movq [72+edi],mm3
|
|
movq mm6,mm3
|
|
movq [ebp-64],mm1
|
|
psrlq mm3,4
|
|
movq [64+ebp],mm4
|
|
shl edx,4
|
|
mov BYTE [10+esp],dl
|
|
mov edx,DWORD [56+ebx]
|
|
movq mm1,[56+ebx]
|
|
psllq mm6,60
|
|
movq mm4,[48+ebx]
|
|
por mm0,mm6
|
|
movq [edi-48],mm2
|
|
psrlq mm2,4
|
|
movq [80+edi],mm5
|
|
movq mm7,mm5
|
|
movq [ebp-56],mm0
|
|
psrlq mm5,4
|
|
movq [72+ebp],mm3
|
|
shl edx,4
|
|
mov BYTE [11+esp],dl
|
|
mov edx,DWORD [72+ebx]
|
|
movq mm0,[72+ebx]
|
|
psllq mm7,60
|
|
movq mm3,[64+ebx]
|
|
por mm2,mm7
|
|
movq [edi-40],mm1
|
|
psrlq mm1,4
|
|
movq [88+edi],mm4
|
|
movq mm6,mm4
|
|
movq [ebp-48],mm2
|
|
psrlq mm4,4
|
|
movq [80+ebp],mm5
|
|
shl edx,4
|
|
mov BYTE [12+esp],dl
|
|
mov edx,DWORD [88+ebx]
|
|
movq mm2,[88+ebx]
|
|
psllq mm6,60
|
|
movq mm5,[80+ebx]
|
|
por mm1,mm6
|
|
movq [edi-32],mm0
|
|
psrlq mm0,4
|
|
movq [96+edi],mm3
|
|
movq mm7,mm3
|
|
movq [ebp-40],mm1
|
|
psrlq mm3,4
|
|
movq [88+ebp],mm4
|
|
shl edx,4
|
|
mov BYTE [13+esp],dl
|
|
mov edx,DWORD [104+ebx]
|
|
movq mm1,[104+ebx]
|
|
psllq mm7,60
|
|
movq mm4,[96+ebx]
|
|
por mm0,mm7
|
|
movq [edi-24],mm2
|
|
psrlq mm2,4
|
|
movq [104+edi],mm5
|
|
movq mm6,mm5
|
|
movq [ebp-32],mm0
|
|
psrlq mm5,4
|
|
movq [96+ebp],mm3
|
|
shl edx,4
|
|
mov BYTE [14+esp],dl
|
|
mov edx,DWORD [120+ebx]
|
|
movq mm0,[120+ebx]
|
|
psllq mm6,60
|
|
movq mm3,[112+ebx]
|
|
por mm2,mm6
|
|
movq [edi-16],mm1
|
|
psrlq mm1,4
|
|
movq [112+edi],mm4
|
|
movq mm7,mm4
|
|
movq [ebp-24],mm2
|
|
psrlq mm4,4
|
|
movq [104+ebp],mm5
|
|
shl edx,4
|
|
mov BYTE [15+esp],dl
|
|
psllq mm7,60
|
|
por mm1,mm7
|
|
movq [edi-8],mm0
|
|
psrlq mm0,4
|
|
movq [120+edi],mm3
|
|
movq mm6,mm3
|
|
movq [ebp-16],mm1
|
|
psrlq mm3,4
|
|
movq [112+ebp],mm4
|
|
psllq mm6,60
|
|
por mm0,mm6
|
|
movq [ebp-8],mm0
|
|
movq [120+ebp],mm3
|
|
movq mm6,[eax]
|
|
mov ebx,DWORD [8+eax]
|
|
mov edx,DWORD [12+eax]
|
|
align 16
|
|
L$004outer:
|
|
xor edx,DWORD [12+ecx]
|
|
xor ebx,DWORD [8+ecx]
|
|
pxor mm6,[ecx]
|
|
lea ecx,[16+ecx]
|
|
mov DWORD [536+esp],ebx
|
|
movq [528+esp],mm6
|
|
mov DWORD [548+esp],ecx
|
|
xor eax,eax
|
|
rol edx,8
|
|
mov al,dl
|
|
mov ebp,eax
|
|
and al,15
|
|
shr ebp,4
|
|
pxor mm0,mm0
|
|
rol edx,8
|
|
pxor mm1,mm1
|
|
pxor mm2,mm2
|
|
movq mm7,[16+eax*8+esp]
|
|
movq mm6,[144+eax*8+esp]
|
|
mov al,dl
|
|
movd ebx,mm7
|
|
psrlq mm7,8
|
|
movq mm3,mm6
|
|
mov edi,eax
|
|
psrlq mm6,8
|
|
pxor mm7,[272+ebp*8+esp]
|
|
and al,15
|
|
psllq mm3,56
|
|
shr edi,4
|
|
pxor mm7,[16+eax*8+esp]
|
|
rol edx,8
|
|
pxor mm6,[144+eax*8+esp]
|
|
pxor mm7,mm3
|
|
pxor mm6,[400+ebp*8+esp]
|
|
xor bl,BYTE [ebp*1+esp]
|
|
mov al,dl
|
|
movd ecx,mm7
|
|
movzx ebx,bl
|
|
psrlq mm7,8
|
|
movq mm3,mm6
|
|
mov ebp,eax
|
|
psrlq mm6,8
|
|
pxor mm7,[272+edi*8+esp]
|
|
and al,15
|
|
psllq mm3,56
|
|
shr ebp,4
|
|
pinsrw mm2,WORD [ebx*2+esi],2
|
|
pxor mm7,[16+eax*8+esp]
|
|
rol edx,8
|
|
pxor mm6,[144+eax*8+esp]
|
|
pxor mm7,mm3
|
|
pxor mm6,[400+edi*8+esp]
|
|
xor cl,BYTE [edi*1+esp]
|
|
mov al,dl
|
|
mov edx,DWORD [536+esp]
|
|
movd ebx,mm7
|
|
movzx ecx,cl
|
|
psrlq mm7,8
|
|
movq mm3,mm6
|
|
mov edi,eax
|
|
psrlq mm6,8
|
|
pxor mm7,[272+ebp*8+esp]
|
|
and al,15
|
|
psllq mm3,56
|
|
pxor mm6,mm2
|
|
shr edi,4
|
|
pinsrw mm1,WORD [ecx*2+esi],2
|
|
pxor mm7,[16+eax*8+esp]
|
|
rol edx,8
|
|
pxor mm6,[144+eax*8+esp]
|
|
pxor mm7,mm3
|
|
pxor mm6,[400+ebp*8+esp]
|
|
xor bl,BYTE [ebp*1+esp]
|
|
mov al,dl
|
|
movd ecx,mm7
|
|
movzx ebx,bl
|
|
psrlq mm7,8
|
|
movq mm3,mm6
|
|
mov ebp,eax
|
|
psrlq mm6,8
|
|
pxor mm7,[272+edi*8+esp]
|
|
and al,15
|
|
psllq mm3,56
|
|
pxor mm6,mm1
|
|
shr ebp,4
|
|
pinsrw mm0,WORD [ebx*2+esi],2
|
|
pxor mm7,[16+eax*8+esp]
|
|
rol edx,8
|
|
pxor mm6,[144+eax*8+esp]
|
|
pxor mm7,mm3
|
|
pxor mm6,[400+edi*8+esp]
|
|
xor cl,BYTE [edi*1+esp]
|
|
mov al,dl
|
|
movd ebx,mm7
|
|
movzx ecx,cl
|
|
psrlq mm7,8
|
|
movq mm3,mm6
|
|
mov edi,eax
|
|
psrlq mm6,8
|
|
pxor mm7,[272+ebp*8+esp]
|
|
and al,15
|
|
psllq mm3,56
|
|
pxor mm6,mm0
|
|
shr edi,4
|
|
pinsrw mm2,WORD [ecx*2+esi],2
|
|
pxor mm7,[16+eax*8+esp]
|
|
rol edx,8
|
|
pxor mm6,[144+eax*8+esp]
|
|
pxor mm7,mm3
|
|
pxor mm6,[400+ebp*8+esp]
|
|
xor bl,BYTE [ebp*1+esp]
|
|
mov al,dl
|
|
movd ecx,mm7
|
|
movzx ebx,bl
|
|
psrlq mm7,8
|
|
movq mm3,mm6
|
|
mov ebp,eax
|
|
psrlq mm6,8
|
|
pxor mm7,[272+edi*8+esp]
|
|
and al,15
|
|
psllq mm3,56
|
|
pxor mm6,mm2
|
|
shr ebp,4
|
|
pinsrw mm1,WORD [ebx*2+esi],2
|
|
pxor mm7,[16+eax*8+esp]
|
|
rol edx,8
|
|
pxor mm6,[144+eax*8+esp]
|
|
pxor mm7,mm3
|
|
pxor mm6,[400+edi*8+esp]
|
|
xor cl,BYTE [edi*1+esp]
|
|
mov al,dl
|
|
mov edx,DWORD [532+esp]
|
|
movd ebx,mm7
|
|
movzx ecx,cl
|
|
psrlq mm7,8
|
|
movq mm3,mm6
|
|
mov edi,eax
|
|
psrlq mm6,8
|
|
pxor mm7,[272+ebp*8+esp]
|
|
and al,15
|
|
psllq mm3,56
|
|
pxor mm6,mm1
|
|
shr edi,4
|
|
pinsrw mm0,WORD [ecx*2+esi],2
|
|
pxor mm7,[16+eax*8+esp]
|
|
rol edx,8
|
|
pxor mm6,[144+eax*8+esp]
|
|
pxor mm7,mm3
|
|
pxor mm6,[400+ebp*8+esp]
|
|
xor bl,BYTE [ebp*1+esp]
|
|
mov al,dl
|
|
movd ecx,mm7
|
|
movzx ebx,bl
|
|
psrlq mm7,8
|
|
movq mm3,mm6
|
|
mov ebp,eax
|
|
psrlq mm6,8
|
|
pxor mm7,[272+edi*8+esp]
|
|
and al,15
|
|
psllq mm3,56
|
|
pxor mm6,mm0
|
|
shr ebp,4
|
|
pinsrw mm2,WORD [ebx*2+esi],2
|
|
pxor mm7,[16+eax*8+esp]
|
|
rol edx,8
|
|
pxor mm6,[144+eax*8+esp]
|
|
pxor mm7,mm3
|
|
pxor mm6,[400+edi*8+esp]
|
|
xor cl,BYTE [edi*1+esp]
|
|
mov al,dl
|
|
movd ebx,mm7
|
|
movzx ecx,cl
|
|
psrlq mm7,8
|
|
movq mm3,mm6
|
|
mov edi,eax
|
|
psrlq mm6,8
|
|
pxor mm7,[272+ebp*8+esp]
|
|
and al,15
|
|
psllq mm3,56
|
|
pxor mm6,mm2
|
|
shr edi,4
|
|
pinsrw mm1,WORD [ecx*2+esi],2
|
|
pxor mm7,[16+eax*8+esp]
|
|
rol edx,8
|
|
pxor mm6,[144+eax*8+esp]
|
|
pxor mm7,mm3
|
|
pxor mm6,[400+ebp*8+esp]
|
|
xor bl,BYTE [ebp*1+esp]
|
|
mov al,dl
|
|
movd ecx,mm7
|
|
movzx ebx,bl
|
|
psrlq mm7,8
|
|
movq mm3,mm6
|
|
mov ebp,eax
|
|
psrlq mm6,8
|
|
pxor mm7,[272+edi*8+esp]
|
|
and al,15
|
|
psllq mm3,56
|
|
pxor mm6,mm1
|
|
shr ebp,4
|
|
pinsrw mm0,WORD [ebx*2+esi],2
|
|
pxor mm7,[16+eax*8+esp]
|
|
rol edx,8
|
|
pxor mm6,[144+eax*8+esp]
|
|
pxor mm7,mm3
|
|
pxor mm6,[400+edi*8+esp]
|
|
xor cl,BYTE [edi*1+esp]
|
|
mov al,dl
|
|
mov edx,DWORD [528+esp]
|
|
movd ebx,mm7
|
|
movzx ecx,cl
|
|
psrlq mm7,8
|
|
movq mm3,mm6
|
|
mov edi,eax
|
|
psrlq mm6,8
|
|
pxor mm7,[272+ebp*8+esp]
|
|
and al,15
|
|
psllq mm3,56
|
|
pxor mm6,mm0
|
|
shr edi,4
|
|
pinsrw mm2,WORD [ecx*2+esi],2
|
|
pxor mm7,[16+eax*8+esp]
|
|
rol edx,8
|
|
pxor mm6,[144+eax*8+esp]
|
|
pxor mm7,mm3
|
|
pxor mm6,[400+ebp*8+esp]
|
|
xor bl,BYTE [ebp*1+esp]
|
|
mov al,dl
|
|
movd ecx,mm7
|
|
movzx ebx,bl
|
|
psrlq mm7,8
|
|
movq mm3,mm6
|
|
mov ebp,eax
|
|
psrlq mm6,8
|
|
pxor mm7,[272+edi*8+esp]
|
|
and al,15
|
|
psllq mm3,56
|
|
pxor mm6,mm2
|
|
shr ebp,4
|
|
pinsrw mm1,WORD [ebx*2+esi],2
|
|
pxor mm7,[16+eax*8+esp]
|
|
rol edx,8
|
|
pxor mm6,[144+eax*8+esp]
|
|
pxor mm7,mm3
|
|
pxor mm6,[400+edi*8+esp]
|
|
xor cl,BYTE [edi*1+esp]
|
|
mov al,dl
|
|
movd ebx,mm7
|
|
movzx ecx,cl
|
|
psrlq mm7,8
|
|
movq mm3,mm6
|
|
mov edi,eax
|
|
psrlq mm6,8
|
|
pxor mm7,[272+ebp*8+esp]
|
|
and al,15
|
|
psllq mm3,56
|
|
pxor mm6,mm1
|
|
shr edi,4
|
|
pinsrw mm0,WORD [ecx*2+esi],2
|
|
pxor mm7,[16+eax*8+esp]
|
|
rol edx,8
|
|
pxor mm6,[144+eax*8+esp]
|
|
pxor mm7,mm3
|
|
pxor mm6,[400+ebp*8+esp]
|
|
xor bl,BYTE [ebp*1+esp]
|
|
mov al,dl
|
|
movd ecx,mm7
|
|
movzx ebx,bl
|
|
psrlq mm7,8
|
|
movq mm3,mm6
|
|
mov ebp,eax
|
|
psrlq mm6,8
|
|
pxor mm7,[272+edi*8+esp]
|
|
and al,15
|
|
psllq mm3,56
|
|
pxor mm6,mm0
|
|
shr ebp,4
|
|
pinsrw mm2,WORD [ebx*2+esi],2
|
|
pxor mm7,[16+eax*8+esp]
|
|
rol edx,8
|
|
pxor mm6,[144+eax*8+esp]
|
|
pxor mm7,mm3
|
|
pxor mm6,[400+edi*8+esp]
|
|
xor cl,BYTE [edi*1+esp]
|
|
mov al,dl
|
|
mov edx,DWORD [524+esp]
|
|
movd ebx,mm7
|
|
movzx ecx,cl
|
|
psrlq mm7,8
|
|
movq mm3,mm6
|
|
mov edi,eax
|
|
psrlq mm6,8
|
|
pxor mm7,[272+ebp*8+esp]
|
|
and al,15
|
|
psllq mm3,56
|
|
pxor mm6,mm2
|
|
shr edi,4
|
|
pinsrw mm1,WORD [ecx*2+esi],2
|
|
pxor mm7,[16+eax*8+esp]
|
|
pxor mm6,[144+eax*8+esp]
|
|
xor bl,BYTE [ebp*1+esp]
|
|
pxor mm7,mm3
|
|
pxor mm6,[400+ebp*8+esp]
|
|
movzx ebx,bl
|
|
pxor mm2,mm2
|
|
psllq mm1,4
|
|
movd ecx,mm7
|
|
psrlq mm7,4
|
|
movq mm3,mm6
|
|
psrlq mm6,4
|
|
shl ecx,4
|
|
pxor mm7,[16+edi*8+esp]
|
|
psllq mm3,60
|
|
movzx ecx,cl
|
|
pxor mm7,mm3
|
|
pxor mm6,[144+edi*8+esp]
|
|
pinsrw mm0,WORD [ebx*2+esi],2
|
|
pxor mm6,mm1
|
|
movd edx,mm7
|
|
pinsrw mm2,WORD [ecx*2+esi],3
|
|
psllq mm0,12
|
|
pxor mm6,mm0
|
|
psrlq mm7,32
|
|
pxor mm6,mm2
|
|
mov ecx,DWORD [548+esp]
|
|
movd ebx,mm7
|
|
movq mm3,mm6
|
|
psllw mm6,8
|
|
psrlw mm3,8
|
|
por mm6,mm3
|
|
bswap edx
|
|
pshufw mm6,mm6,27
|
|
bswap ebx
|
|
cmp ecx,DWORD [552+esp]
|
|
jne NEAR L$004outer
|
|
mov eax,DWORD [544+esp]
|
|
mov DWORD [12+eax],edx
|
|
mov DWORD [8+eax],ebx
|
|
movq [eax],mm6
|
|
mov esp,DWORD [556+esp]
|
|
emms
|
|
pop edi
|
|
pop esi
|
|
pop ebx
|
|
pop ebp
|
|
ret
|
|
global _gcm_init_clmul
|
|
align 16
|
|
_gcm_init_clmul:
|
|
L$_gcm_init_clmul_begin:
|
|
mov edx,DWORD [4+esp]
|
|
mov eax,DWORD [8+esp]
|
|
call L$005pic
|
|
L$005pic:
|
|
pop ecx
|
|
lea ecx,[(L$bswap-L$005pic)+ecx]
|
|
movdqu xmm2,[eax]
|
|
pshufd xmm2,xmm2,78
|
|
pshufd xmm4,xmm2,255
|
|
movdqa xmm3,xmm2
|
|
psllq xmm2,1
|
|
pxor xmm5,xmm5
|
|
psrlq xmm3,63
|
|
pcmpgtd xmm5,xmm4
|
|
pslldq xmm3,8
|
|
por xmm2,xmm3
|
|
pand xmm5,[16+ecx]
|
|
pxor xmm2,xmm5
|
|
movdqa xmm0,xmm2
|
|
movdqa xmm1,xmm0
|
|
pshufd xmm3,xmm0,78
|
|
pshufd xmm4,xmm2,78
|
|
pxor xmm3,xmm0
|
|
pxor xmm4,xmm2
|
|
db 102,15,58,68,194,0
|
|
db 102,15,58,68,202,17
|
|
db 102,15,58,68,220,0
|
|
xorps xmm3,xmm0
|
|
xorps xmm3,xmm1
|
|
movdqa xmm4,xmm3
|
|
psrldq xmm3,8
|
|
pslldq xmm4,8
|
|
pxor xmm1,xmm3
|
|
pxor xmm0,xmm4
|
|
movdqa xmm4,xmm0
|
|
movdqa xmm3,xmm0
|
|
psllq xmm0,5
|
|
pxor xmm3,xmm0
|
|
psllq xmm0,1
|
|
pxor xmm0,xmm3
|
|
psllq xmm0,57
|
|
movdqa xmm3,xmm0
|
|
pslldq xmm0,8
|
|
psrldq xmm3,8
|
|
pxor xmm0,xmm4
|
|
pxor xmm1,xmm3
|
|
movdqa xmm4,xmm0
|
|
psrlq xmm0,1
|
|
pxor xmm1,xmm4
|
|
pxor xmm4,xmm0
|
|
psrlq xmm0,5
|
|
pxor xmm0,xmm4
|
|
psrlq xmm0,1
|
|
pxor xmm0,xmm1
|
|
pshufd xmm3,xmm2,78
|
|
pshufd xmm4,xmm0,78
|
|
pxor xmm3,xmm2
|
|
movdqu [edx],xmm2
|
|
pxor xmm4,xmm0
|
|
movdqu [16+edx],xmm0
|
|
db 102,15,58,15,227,8
|
|
movdqu [32+edx],xmm4
|
|
ret
|
|
global _gcm_gmult_clmul
|
|
align 16
|
|
_gcm_gmult_clmul:
|
|
L$_gcm_gmult_clmul_begin:
|
|
mov eax,DWORD [4+esp]
|
|
mov edx,DWORD [8+esp]
|
|
call L$006pic
|
|
L$006pic:
|
|
pop ecx
|
|
lea ecx,[(L$bswap-L$006pic)+ecx]
|
|
movdqu xmm0,[eax]
|
|
movdqa xmm5,[ecx]
|
|
movups xmm2,[edx]
|
|
db 102,15,56,0,197
|
|
movups xmm4,[32+edx]
|
|
movdqa xmm1,xmm0
|
|
pshufd xmm3,xmm0,78
|
|
pxor xmm3,xmm0
|
|
db 102,15,58,68,194,0
|
|
db 102,15,58,68,202,17
|
|
db 102,15,58,68,220,0
|
|
xorps xmm3,xmm0
|
|
xorps xmm3,xmm1
|
|
movdqa xmm4,xmm3
|
|
psrldq xmm3,8
|
|
pslldq xmm4,8
|
|
pxor xmm1,xmm3
|
|
pxor xmm0,xmm4
|
|
movdqa xmm4,xmm0
|
|
movdqa xmm3,xmm0
|
|
psllq xmm0,5
|
|
pxor xmm3,xmm0
|
|
psllq xmm0,1
|
|
pxor xmm0,xmm3
|
|
psllq xmm0,57
|
|
movdqa xmm3,xmm0
|
|
pslldq xmm0,8
|
|
psrldq xmm3,8
|
|
pxor xmm0,xmm4
|
|
pxor xmm1,xmm3
|
|
movdqa xmm4,xmm0
|
|
psrlq xmm0,1
|
|
pxor xmm1,xmm4
|
|
pxor xmm4,xmm0
|
|
psrlq xmm0,5
|
|
pxor xmm0,xmm4
|
|
psrlq xmm0,1
|
|
pxor xmm0,xmm1
|
|
db 102,15,56,0,197
|
|
movdqu [eax],xmm0
|
|
ret
|
|
global _gcm_ghash_clmul
|
|
align 16
|
|
_gcm_ghash_clmul:
|
|
L$_gcm_ghash_clmul_begin:
|
|
push ebp
|
|
push ebx
|
|
push esi
|
|
push edi
|
|
mov eax,DWORD [20+esp]
|
|
mov edx,DWORD [24+esp]
|
|
mov esi,DWORD [28+esp]
|
|
mov ebx,DWORD [32+esp]
|
|
call L$007pic
|
|
L$007pic:
|
|
pop ecx
|
|
lea ecx,[(L$bswap-L$007pic)+ecx]
|
|
movdqu xmm0,[eax]
|
|
movdqa xmm5,[ecx]
|
|
movdqu xmm2,[edx]
|
|
db 102,15,56,0,197
|
|
sub ebx,16
|
|
jz NEAR L$008odd_tail
|
|
movdqu xmm3,[esi]
|
|
movdqu xmm6,[16+esi]
|
|
db 102,15,56,0,221
|
|
db 102,15,56,0,245
|
|
movdqu xmm5,[32+edx]
|
|
pxor xmm0,xmm3
|
|
pshufd xmm3,xmm6,78
|
|
movdqa xmm7,xmm6
|
|
pxor xmm3,xmm6
|
|
lea esi,[32+esi]
|
|
db 102,15,58,68,242,0
|
|
db 102,15,58,68,250,17
|
|
db 102,15,58,68,221,0
|
|
movups xmm2,[16+edx]
|
|
nop
|
|
sub ebx,32
|
|
jbe NEAR L$009even_tail
|
|
jmp NEAR L$010mod_loop
|
|
align 32
|
|
L$010mod_loop:
|
|
pshufd xmm4,xmm0,78
|
|
movdqa xmm1,xmm0
|
|
pxor xmm4,xmm0
|
|
nop
|
|
db 102,15,58,68,194,0
|
|
db 102,15,58,68,202,17
|
|
db 102,15,58,68,229,16
|
|
movups xmm2,[edx]
|
|
xorps xmm0,xmm6
|
|
movdqa xmm5,[ecx]
|
|
xorps xmm1,xmm7
|
|
movdqu xmm7,[esi]
|
|
pxor xmm3,xmm0
|
|
movdqu xmm6,[16+esi]
|
|
pxor xmm3,xmm1
|
|
db 102,15,56,0,253
|
|
pxor xmm4,xmm3
|
|
movdqa xmm3,xmm4
|
|
psrldq xmm4,8
|
|
pslldq xmm3,8
|
|
pxor xmm1,xmm4
|
|
pxor xmm0,xmm3
|
|
db 102,15,56,0,245
|
|
pxor xmm1,xmm7
|
|
movdqa xmm7,xmm6
|
|
movdqa xmm4,xmm0
|
|
movdqa xmm3,xmm0
|
|
psllq xmm0,5
|
|
pxor xmm3,xmm0
|
|
psllq xmm0,1
|
|
pxor xmm0,xmm3
|
|
db 102,15,58,68,242,0
|
|
movups xmm5,[32+edx]
|
|
psllq xmm0,57
|
|
movdqa xmm3,xmm0
|
|
pslldq xmm0,8
|
|
psrldq xmm3,8
|
|
pxor xmm0,xmm4
|
|
pxor xmm1,xmm3
|
|
pshufd xmm3,xmm7,78
|
|
movdqa xmm4,xmm0
|
|
psrlq xmm0,1
|
|
pxor xmm3,xmm7
|
|
pxor xmm1,xmm4
|
|
db 102,15,58,68,250,17
|
|
movups xmm2,[16+edx]
|
|
pxor xmm4,xmm0
|
|
psrlq xmm0,5
|
|
pxor xmm0,xmm4
|
|
psrlq xmm0,1
|
|
pxor xmm0,xmm1
|
|
db 102,15,58,68,221,0
|
|
lea esi,[32+esi]
|
|
sub ebx,32
|
|
ja NEAR L$010mod_loop
|
|
L$009even_tail:
|
|
pshufd xmm4,xmm0,78
|
|
movdqa xmm1,xmm0
|
|
pxor xmm4,xmm0
|
|
db 102,15,58,68,194,0
|
|
db 102,15,58,68,202,17
|
|
db 102,15,58,68,229,16
|
|
movdqa xmm5,[ecx]
|
|
xorps xmm0,xmm6
|
|
xorps xmm1,xmm7
|
|
pxor xmm3,xmm0
|
|
pxor xmm3,xmm1
|
|
pxor xmm4,xmm3
|
|
movdqa xmm3,xmm4
|
|
psrldq xmm4,8
|
|
pslldq xmm3,8
|
|
pxor xmm1,xmm4
|
|
pxor xmm0,xmm3
|
|
movdqa xmm4,xmm0
|
|
movdqa xmm3,xmm0
|
|
psllq xmm0,5
|
|
pxor xmm3,xmm0
|
|
psllq xmm0,1
|
|
pxor xmm0,xmm3
|
|
psllq xmm0,57
|
|
movdqa xmm3,xmm0
|
|
pslldq xmm0,8
|
|
psrldq xmm3,8
|
|
pxor xmm0,xmm4
|
|
pxor xmm1,xmm3
|
|
movdqa xmm4,xmm0
|
|
psrlq xmm0,1
|
|
pxor xmm1,xmm4
|
|
pxor xmm4,xmm0
|
|
psrlq xmm0,5
|
|
pxor xmm0,xmm4
|
|
psrlq xmm0,1
|
|
pxor xmm0,xmm1
|
|
test ebx,ebx
|
|
jnz NEAR L$011done
|
|
movups xmm2,[edx]
|
|
L$008odd_tail:
|
|
movdqu xmm3,[esi]
|
|
db 102,15,56,0,221
|
|
pxor xmm0,xmm3
|
|
movdqa xmm1,xmm0
|
|
pshufd xmm3,xmm0,78
|
|
pshufd xmm4,xmm2,78
|
|
pxor xmm3,xmm0
|
|
pxor xmm4,xmm2
|
|
db 102,15,58,68,194,0
|
|
db 102,15,58,68,202,17
|
|
db 102,15,58,68,220,0
|
|
xorps xmm3,xmm0
|
|
xorps xmm3,xmm1
|
|
movdqa xmm4,xmm3
|
|
psrldq xmm3,8
|
|
pslldq xmm4,8
|
|
pxor xmm1,xmm3
|
|
pxor xmm0,xmm4
|
|
movdqa xmm4,xmm0
|
|
movdqa xmm3,xmm0
|
|
psllq xmm0,5
|
|
pxor xmm3,xmm0
|
|
psllq xmm0,1
|
|
pxor xmm0,xmm3
|
|
psllq xmm0,57
|
|
movdqa xmm3,xmm0
|
|
pslldq xmm0,8
|
|
psrldq xmm3,8
|
|
pxor xmm0,xmm4
|
|
pxor xmm1,xmm3
|
|
movdqa xmm4,xmm0
|
|
psrlq xmm0,1
|
|
pxor xmm1,xmm4
|
|
pxor xmm4,xmm0
|
|
psrlq xmm0,5
|
|
pxor xmm0,xmm4
|
|
psrlq xmm0,1
|
|
pxor xmm0,xmm1
|
|
L$011done:
|
|
db 102,15,56,0,197
|
|
movdqu [eax],xmm0
|
|
pop edi
|
|
pop esi
|
|
pop ebx
|
|
pop ebp
|
|
ret
|
|
align 64
|
|
L$bswap:
|
|
db 15,14,13,12,11,10,9,8,7,6,5,4,3,2,1,0
|
|
db 1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,194
|
|
align 64
|
|
L$rem_8bit:
|
|
dw 0,450,900,582,1800,1738,1164,1358
|
|
dw 3600,4050,3476,3158,2328,2266,2716,2910
|
|
dw 7200,7650,8100,7782,6952,6890,6316,6510
|
|
dw 4656,5106,4532,4214,5432,5370,5820,6014
|
|
dw 14400,14722,15300,14854,16200,16010,15564,15630
|
|
dw 13904,14226,13780,13334,12632,12442,13020,13086
|
|
dw 9312,9634,10212,9766,9064,8874,8428,8494
|
|
dw 10864,11186,10740,10294,11640,11450,12028,12094
|
|
dw 28800,28994,29444,29382,30600,30282,29708,30158
|
|
dw 32400,32594,32020,31958,31128,30810,31260,31710
|
|
dw 27808,28002,28452,28390,27560,27242,26668,27118
|
|
dw 25264,25458,24884,24822,26040,25722,26172,26622
|
|
dw 18624,18690,19268,19078,20424,19978,19532,19854
|
|
dw 18128,18194,17748,17558,16856,16410,16988,17310
|
|
dw 21728,21794,22372,22182,21480,21034,20588,20910
|
|
dw 23280,23346,22900,22710,24056,23610,24188,24510
|
|
dw 57600,57538,57988,58182,58888,59338,58764,58446
|
|
dw 61200,61138,60564,60758,59416,59866,60316,59998
|
|
dw 64800,64738,65188,65382,64040,64490,63916,63598
|
|
dw 62256,62194,61620,61814,62520,62970,63420,63102
|
|
dw 55616,55426,56004,56070,56904,57226,56780,56334
|
|
dw 55120,54930,54484,54550,53336,53658,54236,53790
|
|
dw 50528,50338,50916,50982,49768,50090,49644,49198
|
|
dw 52080,51890,51444,51510,52344,52666,53244,52798
|
|
dw 37248,36930,37380,37830,38536,38730,38156,38094
|
|
dw 40848,40530,39956,40406,39064,39258,39708,39646
|
|
dw 36256,35938,36388,36838,35496,35690,35116,35054
|
|
dw 33712,33394,32820,33270,33976,34170,34620,34558
|
|
dw 43456,43010,43588,43910,44744,44810,44364,44174
|
|
dw 42960,42514,42068,42390,41176,41242,41820,41630
|
|
dw 46560,46114,46692,47014,45800,45866,45420,45230
|
|
dw 48112,47666,47220,47542,48376,48442,49020,48830
|
|
align 64
|
|
L$rem_4bit:
|
|
dd 0,0,0,471859200,0,943718400,0,610271232
|
|
dd 0,1887436800,0,1822425088,0,1220542464,0,1423966208
|
|
dd 0,3774873600,0,4246732800,0,3644850176,0,3311403008
|
|
dd 0,2441084928,0,2376073216,0,2847932416,0,3051356160
|
|
db 71,72,65,83,72,32,102,111,114,32,120,56,54,44,32,67
|
|
db 82,89,80,84,79,71,65,77,83,32,98,121,32,60,97,112
|
|
db 112,114,111,64,111,112,101,110,115,115,108,46,111,114,103,62
|
|
db 0
|