Commit 24798c5e authored by Andy Polyakov's avatar Andy Polyakov
Browse files

ghash-sparcv9.pl: 22% improvement on T4.

parent 5b0e3daf
Loading
Loading
Loading
Loading
+152 −118
Original line number Diff line number Diff line
@@ -41,8 +41,10 @@
#
# Add VIS3 lookup-table-free implementation using polynomial
# multiplication xmulx[hi] and extended addition addxc[cc]
# instructions. 3.96/6.26x improvement on T3/T4 or in absolute
# terms 9.02/2.61 cycles per byte.
# instructions. 4.22/7.63x improvement on T3/T4 or in absolute
# terms 8.45/2.14 cycles per byte. On T4 multi-process benchmark
# saturates at ~15x single-process result on 8-core processor, or
# ~19.7GBps per 2.85GHz socket.

$bits=32;
for (@ARGV)     { $bits=64 if (/\-m64/ || /\-xarch\=v9/); }
@@ -335,74 +337,103 @@ gcm_gmult_4bit:
___

{{{
# Straightforward 64-bits-at-a-time approach with pair of 128x64-bit
# multiplications followed by 64-bit reductions. While it might be
# suboptimal with regard to sheer amount of multiplications, other
# methods would require larger amount of 64-bit registers, which we
# don't have in 32-bit application. Also, they [alternative methods
# such as aggregated reduction] kind of thrive on fast 128-bit SIMD
# instructions and these are not option on SPARC...
# Straightforward 128x128-bit multiplication using Karatsuba algorithm
# followed by pair of 64-bit reductions [with a shortcut in first one,
# which allowed to break dependency between reductions and remove one
# mulitplication from critical path]. While it might be suboptimal
# with regard to sheer number of multiplications, other methods [such
# as aggregate reduction] would require more 64-bit registers, which
# we don't have in 32-bit application context.

($Xip,$Htable,$inp,$len)=map("%i$_",(0..3));

($xE1,$Hhi,$Hlo,$Rhi,$Rlo,$M0hi,$M0lo,$M1hi,$M1lo,$Zhi,$Zlo,$X)=
	(map("%g$_",(1..5)),map("%o$_",(0..5,7)));
($shl,$shr)=map("%l$_",(0..7));
($Hhl,$Hlo,$Hhi,$Xlo,$Xhi,$xE1,$x384, $C0,$C1,$C2,$C3,$V)=
	(map("%o$_",(0..5,7)),map("%g$_",(1..5)));

($shl,$shr,$sqr)=map("%l$_",(0..7));

# For details regarding "twisted H" see ghash-x86.pl.
$code.=<<___;
.globl	gcm_gmult_vis3
.globl	gcm_init_vis3
.align	32
gcm_gmult_vis3:
gcm_init_vis3:
	save	%sp,-$frame,%sp

	ldx	[$Xip+8],$X		! load X.lo
	ldx	[$Htable-8], $Hlo	! load H
	ldx	[$Htable-16],$Hhi
	mov	0xE1,$xE1
	sllx	$xE1,57,$xE1

	xmulx	$X,$Hlo,$M0lo		! HX.lo
	xmulxhi	$X,$Hlo,$M0hi
	xmulx	$X,$Hhi,$M1lo
	xmulxhi	$X,$Hhi,$M1hi
	ldx	[$Xip+0],$X		! load X.hi

	addcc	$M0lo,$M0lo,$M0lo	! (HX.lo)<<1
	xor	$M0hi,$M1lo,$M1lo

	xmulx	$xE1,$M0lo,$Rlo		! res=Z.lo(0xE1<<57)
	xmulxhi	$xE1,$M0lo,$Rhi

	addxccc	$M1lo,$M1lo,$Zlo	! Z=((HX.lo)<<1)>>64
	addxc	$M1hi,$M1hi,$Zhi
	xor	$M0lo,$Zhi,$Zhi		! overflow bit from 0xE1<<57

	xmulx	$X,$Hlo,$M0lo		! HX.hi
	xmulxhi	$X,$Hlo,$M0hi
	xmulx	$X,$Hhi,$M1lo
	xmulxhi	$X,$Hhi,$M1hi
	ldx	[%i1+0],$Hhi
	ldx	[%i1+8],$Hlo
	mov	0xE1,$Xhi
	mov	1,$Xlo
	sllx	$Xhi,57,$Xhi
	srax	$Hhi,63,$C0		! carry
	addcc	$Hlo,$Hlo,$Hlo		! H<<=1
	addxc	$Hhi,$Hhi,$Hhi
	and	$Xlo,$C0,$Xlo
	and	$Xhi,$C0,$Xhi
	xor	$Xlo,$Hlo,$Hlo
	xor	$Xhi,$Hhi,$Hhi
	stx	$Hlo,[%i0+8]		! save twisted H
	stx	$Hhi,[%i0+0]

	xor	$Rlo,$Zlo,$Zlo		! Z^=res
	xor	$Rhi,$Zhi,$Zhi

	addcc	$M0lo,$M0lo,$M0lo	! (HX.lo)<<1
	xor	$Zlo, $M0lo,$M0lo
	xor	$M0hi,$M1lo,$M1lo

	xmulx	$xE1,$M0lo,$Rlo		! res=Z.lo(0xE1<<57)
	xmulxhi	$xE1,$M0lo,$Rhi

	addxccc	$M1lo,$M1lo,$M1lo
	addxc	$M1hi,$M1hi,$M1hi

	xor	$M1lo,$Zhi,$Zlo		! Z=(Z^(HX.hi)<<1)>>64
	xor	$M0lo,$M1hi,$Zhi	! overflow bit from 0xE1<<57
	ret
	restore
.type	gcm_init_vis3,#function
.size	gcm_init_vis3,.-gcm_init_vis3

	xor	$Rlo,$Zlo,$Zlo		! Z^=res
	xor	$Rhi,$Zhi,$Zhi
.globl	gcm_gmult_vis3
.align	32
gcm_gmult_vis3:
	save	%sp,-$frame,%sp

	stx	$Zlo,[$Xip+8]		! save Xi
	stx	$Zhi,[$Xip+0]
	ldx	[$Xip+8],$Xlo		! load Xi
	ldx	[$Xip+0],$Xhi
	ldx	[$Htable+8],$Hlo	! load twisted H
	ldx	[$Htable+0],$Hhi

	sethi	%hi(0xA0406080),$V
	sethi	%hi(0x20C0E000),%l0
	or	$V,%lo(0xA0406080),$V
	or	%l0,%lo(0x20C0E000),%l0
	sllx	$V,32,$V
	mov	0xE1,%l1
	or	%l0,$V,$V		! (0xE0i)&0xff=0xA040608020C0E000
	sllx	%l1,57,$xE1		! 57 is not a typo
	sllx	%l1,50,$x384	
	xor	$Hhi,$Hlo,$Hhl		! Karatsuba pre-processing

	xmulx	$Xlo,$Hlo,$C0
	xor	$Xlo,$Xhi,$C2		! Karatsuba pre-processing
	xmulx	$C2,$Hhl,$C1
	xmulxhi	$Xlo,$Hlo,$Xlo
	xmulxhi	$C2,$Hhl,$C2
	xmulxhi	$Xhi,$Hhi,$C3
	xmulx	$Xhi,$Hhi,$Xhi

	sll	$C0,3,$sqr
	srlx	$V,$sqr,$sqr		! 0xE0 [implicit &(7<<3)]
	xor	$C0,$sqr,$sqr
	and	$sqr,0x7f,$sqr

	xor	$C0,$C1,$C1		! Karatsuba post-processing
	xor	$Xlo,$C2,$C2
	xor	$Xhi,$C1,$C1
	xor	$C3,$C2,$C2
	xor	$Xlo,$C1,$C1

	xmulxhi	$C0,$xE1,$Xlo		! 0xE1<<1<<56
	 xor	$Xhi,$C2,$C2
	xmulx	$sqr,$x384,$Xhi		! 0xE1<<2<<48
	 xor	$C0,$C2,$C2
	xmulx	$C1,$xE1,$C0
	 xor	$C1,$C3,$C3
	xmulxhi	$C1,$xE1,$C1

	xor	$Xlo,$C2,$C2
	xor	$Xhi,$C3,$C3
	xor	$C0,$C2,$C2
	xor	$C1,$C3,$C3

	stx	$C2,[$Xip+8]		! save Xi
	stx	$C3,[$Xip+0]

	ret
	restore
@@ -414,80 +445,83 @@ gcm_gmult_vis3:
gcm_ghash_vis3:
	save	%sp,-$frame,%sp

	ldx	[$Xip+0],$Zhi		! load X.hi
	ldx	[$Xip+8],$Zlo		! load X.lo
	ldx	[$Xip+8],$C2		! load Xi
	ldx	[$Xip+0],$C3
	ldx	[$Htable+8],$Hlo	! load twisted H
	ldx	[$Htable+0],$Hhi

	sethi	%hi(0xA0406080),$V
	sethi	%hi(0x20C0E000),%l6
	or	$V,%lo(0xA0406080),$V
	or	%l6,%lo(0x20C0E000),%l6
	sllx	$V,32,$V
	mov	0xE1,%l7
	or	%l6,$V,$V		! (0xE0i)&0xff=0xA040608020C0E000
	sllx	%l7,57,$xE1		! 57 is not a typo
	sllx	%l7,50,$x384	

	and	$inp,7,$shl
	andn	$inp,7,$inp
	ldx	[$Htable-8], $Hlo	! load H
	ldx	[$Htable-16],$Hhi
	sll	$shl,3,$shl
	prefetch [$inp+63], 20
	mov	0xE1,$xE1
	sub	%g0,$shl,$shr
	sllx	$xE1,57,$xE1

	xor	$Hhi,$Hlo,$Hhl		! Karatsuba pre-processing

.Loop:
	ldx	[$inp+8],$Rlo		! load *inp
	ldx	[$inp+8],$Xlo
	brz,pt	$shl,1f
	ldx	[$inp+0],$Rhi

	ldx	[$inp+16],$X		! align data
	srlx	$Rlo,$shr,$M0lo
	sllx	$Rlo,$shl,$Rlo
	sllx	$Rhi,$shl,$Rhi
	srlx	$X,$shr,$X
	or	$M0lo,$Rhi,$Rhi
	or	$X,$Rlo,$Rlo

	ldx	[$inp+0],$Xhi

	ldx	[$inp+16],$C1		! align data
	srlx	$Xlo,$shr,$C0
	sllx	$Xlo,$shl,$Xlo
	sllx	$Xhi,$shl,$Xhi
	srlx	$C1,$shr,$C1
	or	$C0,$Xhi,$Xhi
	or	$C1,$Xlo,$Xlo
1:
	add	$inp,16,$inp
	sub	$len,16,$len
	xor	$Rlo,$Zlo,$X
	xor	$C2,$Xlo,$Xlo
	xor	$C3,$Xhi,$Xhi
	prefetch [$inp+63], 20

	xmulx	$X,$Hlo,$M0lo		! HX.lo
	xmulxhi	$X,$Hlo,$M0hi
	xmulx	$X,$Hhi,$M1lo
	xmulxhi	$X,$Hhi,$M1hi
	xor	$Rhi,$Zhi,$X

	addcc	$M0lo,$M0lo,$M0lo	! (HX.lo)<<1
	xor	$M0hi,$M1lo,$M1lo

	xmulx	$xE1,$M0lo,$Rlo		! res=Z.lo(0xE1<<57)
	xmulxhi	$xE1,$M0lo,$Rhi

	addxccc	$M1lo,$M1lo,$Zlo	! Z=((HX.lo)<<1)>>64
	addxc	$M1hi,$M1hi,$Zhi
	xor	$M0lo,$Zhi,$Zhi		! overflow bit from 0xE1<<57

	xmulx	$X,$Hlo,$M0lo		! HX.hi
	xmulxhi	$X,$Hlo,$M0hi
	xmulx	$X,$Hhi,$M1lo
	xmulxhi	$X,$Hhi,$M1hi

	xor	$Rlo,$Zlo,$Zlo		! Z^=res
	xor	$Rhi,$Zhi,$Zhi

	addcc	$M0lo,$M0lo,$M0lo	! (HX.lo)<<1
	xor	$Zlo, $M0lo,$M0lo
	xor	$M0hi,$M1lo,$M1lo

	xmulx	$xE1,$M0lo,$Rlo		! res=Z.lo(0xE1<<57)
	xmulxhi	$xE1,$M0lo,$Rhi

	addxccc	$M1lo,$M1lo,$M1lo
	addxc	$M1hi,$M1hi,$M1hi

	xor	$M1lo,$Zhi,$Zlo		! Z=(Z^(HX.hi)<<1)>>64
	xor	$M0lo,$M1hi,$Zhi	! overflow bit from 0xE1<<57

	xor	$Rlo,$Zlo,$Zlo		! Z^=res
	xmulx	$Xlo,$Hlo,$C0
	xor	$Xlo,$Xhi,$C2		! Karatsuba pre-processing
	xmulx	$C2,$Hhl,$C1
	xmulxhi	$Xlo,$Hlo,$Xlo
	xmulxhi	$C2,$Hhl,$C2
	xmulxhi	$Xhi,$Hhi,$C3
	xmulx	$Xhi,$Hhi,$Xhi

	sll	$C0,3,$sqr
	srlx	$V,$sqr,$sqr		! 0xE0 [implicit &(7<<3)]
	xor	$C0,$sqr,$sqr
	and	$sqr,0x7f,$sqr

	xor	$C0,$C1,$C1		! Karatsuba post-processing
	xor	$Xlo,$C2,$C2
	xor	$Xhi,$C1,$C1
	xor	$C3,$C2,$C2
	xor	$Xlo,$C1,$C1

	xmulxhi	$C0,$xE1,$Xlo		! 0xE1<<1<<56
	 xor	$Xhi,$C2,$C2
	xmulx	$sqr,$x384,$Xhi		! 0xE1<<2<<48
	 xor	$C0,$C2,$C2
	xmulx	$C1,$xE1,$C0
	 xor	$C1,$C3,$C3
	xmulxhi	$C1,$xE1,$C1

	xor	$Xlo,$C2,$C2
	xor	$Xhi,$C3,$C3
	xor	$C0,$C2,$C2
	brnz,pt	$len,.Loop
	xor	$Rhi,$Zhi,$Zhi
	xor	$C1,$C3,$C3

	stx	$Zlo,[$Xip+8]		! save Xi
	stx	$Zhi,[$Xip+0]
	stx	$C2,[$Xip+8]		! save Xi
	stx	$C3,[$Xip+0]

	ret
	restore
+2 −0
Original line number Diff line number Diff line
@@ -679,6 +679,7 @@ void gcm_ghash_neon(u64 Xi[2],const u128 Htable[16],const u8 *inp,size_t len);
#  define GHASH_ASM_SPARC
#  define GCM_FUNCREF_4BIT
extern unsigned int OPENSSL_sparcv9cap_P[];
void gcm_init_vis3(u128 Htable[16],const u64 Xi[2]);
void gcm_gmult_vis3(u64 Xi[2],const u128 Htable[16]);
void gcm_ghash_vis3(u64 Xi[2],const u128 Htable[16],const u8 *inp,size_t len);
# endif
@@ -759,6 +760,7 @@ void CRYPTO_gcm128_init(GCM128_CONTEXT *ctx,void *key,block128_f block)
	}
# elif	defined(GHASH_ASM_SPARC)
	if (OPENSSL_sparcv9cap_P[0] & SPARCV9_VIS3) {
		gcm_init_vis3(ctx->Htable,ctx->H.u);
		ctx->gmult = gcm_gmult_vis3;
		ctx->ghash = gcm_ghash_vis3;
	} else {