Loading crypto/sha/asm/sha1-x86_64.pl +2 −2 Original line number Diff line number Diff line Loading @@ -806,7 +806,7 @@ $code.=<<___; mov %rdi,$ctx # reassigned argument mov %rsi,$inp # reassigned argument mov %rdx,$num # reassigned argument vzeroall vzeroupper shl \$6,$num add $inp,$num Loading Loading @@ -1096,7 +1096,7 @@ ___ &Xtail_avx(\&body_20_39); $code.=<<___; vzeroall vzeroupper add 0($ctx),$A # update context add 4($ctx),@T[0] Loading crypto/sha/asm/sha512-x86_64.pl +6 −6 Original line number Diff line number Diff line Loading @@ -948,7 +948,7 @@ ___ $code.=<<___; .Lprologue_xop: vzeroall vzeroupper mov $SZ*0($ctx),$A mov $SZ*1($ctx),$B mov $SZ*2($ctx),$C Loading Loading @@ -1260,7 +1260,7 @@ $code.=<<___; jb .Lloop_xop mov $_rsp,%rsi vzeroall vzeroupper ___ $code.=<<___ if ($win64); movaps 16*$SZ+32(%rsp),%xmm6 Loading Loading @@ -1324,7 +1324,7 @@ ___ $code.=<<___; .Lprologue_avx: vzeroall vzeroupper mov $SZ*0($ctx),$A mov $SZ*1($ctx),$B mov $SZ*2($ctx),$C Loading Loading @@ -1568,7 +1568,7 @@ $code.=<<___; jb .Lloop_avx mov $_rsp,%rsi vzeroall vzeroupper ___ $code.=<<___ if ($win64); movaps 16*$SZ+32(%rsp),%xmm6 Loading Loading @@ -1676,7 +1676,7 @@ ___ $code.=<<___; .Lprologue_avx2: vzeroall vzeroupper sub \$-16*$SZ,$inp # inp++, size optimization mov $SZ*0($ctx),$A mov $inp,%r12 # borrow $T1 Loading Loading @@ -1942,7 +1942,7 @@ $code.=<<___; .Ldone_avx2: lea ($Tbl),%rsp mov $_rsp,%rsi vzeroall vzeroupper ___ $code.=<<___ if ($win64); movaps 16*$SZ+32(%rsp),%xmm6 Loading Loading
crypto/sha/asm/sha1-x86_64.pl +2 −2 Original line number Diff line number Diff line Loading @@ -806,7 +806,7 @@ $code.=<<___; mov %rdi,$ctx # reassigned argument mov %rsi,$inp # reassigned argument mov %rdx,$num # reassigned argument vzeroall vzeroupper shl \$6,$num add $inp,$num Loading Loading @@ -1096,7 +1096,7 @@ ___ &Xtail_avx(\&body_20_39); $code.=<<___; vzeroall vzeroupper add 0($ctx),$A # update context add 4($ctx),@T[0] Loading
crypto/sha/asm/sha512-x86_64.pl +6 −6 Original line number Diff line number Diff line Loading @@ -948,7 +948,7 @@ ___ $code.=<<___; .Lprologue_xop: vzeroall vzeroupper mov $SZ*0($ctx),$A mov $SZ*1($ctx),$B mov $SZ*2($ctx),$C Loading Loading @@ -1260,7 +1260,7 @@ $code.=<<___; jb .Lloop_xop mov $_rsp,%rsi vzeroall vzeroupper ___ $code.=<<___ if ($win64); movaps 16*$SZ+32(%rsp),%xmm6 Loading Loading @@ -1324,7 +1324,7 @@ ___ $code.=<<___; .Lprologue_avx: vzeroall vzeroupper mov $SZ*0($ctx),$A mov $SZ*1($ctx),$B mov $SZ*2($ctx),$C Loading Loading @@ -1568,7 +1568,7 @@ $code.=<<___; jb .Lloop_avx mov $_rsp,%rsi vzeroall vzeroupper ___ $code.=<<___ if ($win64); movaps 16*$SZ+32(%rsp),%xmm6 Loading Loading @@ -1676,7 +1676,7 @@ ___ $code.=<<___; .Lprologue_avx2: vzeroall vzeroupper sub \$-16*$SZ,$inp # inp++, size optimization mov $SZ*0($ctx),$A mov $inp,%r12 # borrow $T1 Loading Loading @@ -1942,7 +1942,7 @@ $code.=<<___; .Ldone_avx2: lea ($Tbl),%rsp mov $_rsp,%rsi vzeroall vzeroupper ___ $code.=<<___ if ($win64); movaps 16*$SZ+32(%rsp),%xmm6 Loading