#!/usr/bin/env perl # ==================================================================== # Written by Andy Polyakov for the OpenSSL # project. The module is, however, dual licensed under OpenSSL and # CRYPTOGAMS licenses depending on where you obtain it. For further # details see http://www.openssl.org/~appro/cryptogams/. # ==================================================================== # October 2012. # # SPARCv9 VIS3 Montgomery multiplicaion procedure suitable for T3 and # onward. There are three new instructions used here: umulxhi, # addxc[cc] and initializing store. On T3 RSA private key operations # are 1.54/1.87/2.11/2.26 times faster for 512/1024/2048/4096-bit key # lengths. This is without dedicated squaring procedure. On T4 # corresponding coefficients are 1.47/2.10/2.80/2.90x, which is mostly # for reference purposes, because T4 has dedicated Montgomery # multiplication and squaring *instructions* that deliver even more. $bits=32; for (@ARGV) { $bits=64 if (/\-m64/ || /\-xarch\=v9/); } if ($bits==64) { $bias=2047; $frame=192; } else { $bias=0; $frame=112; } $code.=<<___ if ($bits==64); .register %g2,#scratch .register %g3,#scratch ___ $code.=<<___; .section ".text",#alloc,#execinstr ___ ($n0,$m0,$m1,$lo0,$hi0, $lo1,$hi1,$aj,$alo,$nj,$nlo,$tj)= (map("%g$_",(1..5)),map("%o$_",(0..5,7))); # int bn_mul_mont( $rp="%o0"; # BN_ULONG *rp, $ap="%o1"; # const BN_ULONG *ap, $bp="%o2"; # const BN_ULONG *bp, $np="%o3"; # const BN_ULONG *np, $n0p="%o4"; # const BN_ULONG *n0, $num="%o5"; # int num); # caller ensures that num is even # and >=6 $code.=<<___; .globl bn_mul_mont_vis3 .align 32 bn_mul_mont_vis3: add %sp, $bias, %g4 ! real top of stack sll $num, 2, $num ! size in bytes add $num, 63, %g5 andn %g5, 63, %g5 ! buffer size rounded up to 64 bytes add %g5, %g5, %g1 add %g5, %g1, %g1 ! 3*buffer size sub %g4, %g1, %g1 andn %g1, 63, %g1 ! align at 64 byte sub %g1, $frame, %g1 ! new top of stack sub %g1, %g4, %g1 save %sp, %g1, %sp ___ # +-------------------------------+<----- %sp # . . # +-------------------------------+<----- aligned at 64 bytes # | __int64 tmp[0] | # +-------------------------------+ # . . # . . # +-------------------------------+<----- aligned at 64 bytes # | __int64 ap[1..0] | converted ap[] # +-------------------------------+ # | __int64 np[1..0] | converted np[] # +-------------------------------+ # | __int64 ap[3..2] | # . . # . . # +-------------------------------+ ($rp,$ap,$bp,$np,$n0p,$num)=map("%i$_",(0..5)); ($t0,$t1,$t2,$t3,$cnt,$tp,$bufsz,$anp)=map("%l$_",(0..7)); ($ovf,$i)=($t0,$t1); $code.=<<___; ld [$n0p+0], $t0 ! pull n0[0..1] value add %sp, $bias+$frame, $tp ld [$n0p+4], $t1 add $tp, %g5, $anp ld [$bp+0], $t2 ! m0=bp[0] sllx $t1, 32, $n0 ld [$bp+4], $t3 or $t0, $n0, $n0 add $bp, 8, $bp ld [$ap+0], $t0 ! ap[0] sllx $t3, 32, $m0 ld [$ap+4], $t1 or $t2, $m0, $m0 ld [$ap+8], $t2 ! ap[1] sllx $t1, 32, $aj ld [$ap+12], $t3 or $t0, $aj, $aj add $ap, 16, $ap stx $aj, [$anp] ! converted ap[0] mulx $aj, $m0, $lo0 ! ap[0]*bp[0] umulxhi $aj, $m0, $hi0 ld [$np+0], $t0 ! np[0] sllx $t3, 32, $aj ld [$np+4], $t1 or $t2, $aj, $aj ld [$np+8], $t2 ! np[1] sllx $t1, 32, $nj ld [$np+12], $t3 or $t0, $nj, $nj add $np, 16, $np stx $nj, [$anp+8] ! converted np[0] mulx $lo0, $n0, $m1 ! "tp[0]"*n0 stx $aj, [$anp+16] ! converted ap[1] mulx $aj, $m0, $alo ! ap[1]*bp[0] umulxhi $aj, $m0, $aj ! ahi=aj mulx $nj, $m1, $lo1 ! np[0]*m1 umulxhi $nj, $m1, $hi1 sllx $t3, 32, $nj or $t2, $nj, $nj stx $nj, [$anp+24] ! converted np[1] add $anp, 32, $anp addcc $lo0, $lo1, $lo1 addxc %g0, $hi1, $hi1 mulx $nj, $m1, $nlo ! np[1]*m1 umulxhi $nj, $m1, $nj ! nhi=nj ba .L1st sub $num, 24, $cnt ! cnt=num-3 .align 16 .L1st: ld [$ap+0], $t0 ! ap[j] addcc $alo, $hi0, $lo0 ld [$ap+4], $t1 addxc $aj, %g0, $hi0 sllx $t1, 32, $aj add $ap, 8, $ap or $t0, $aj, $aj stx $aj, [$anp] ! converted ap[j] ld [$np+0], $t2 ! np[j] addcc $nlo, $hi1, $lo1 ld [$np+4], $t3 addxc $nj, %g0, $hi1 ! nhi=nj sllx $t3, 32, $nj add $np, 8, $np mulx $aj, $m0, $alo ! ap[j]*bp[0] or $t2, $nj, $nj umulxhi $aj, $m0, $aj ! ahi=aj stx $nj, [$anp+8] ! converted np[j] add $anp, 16, $anp ! anp++ mulx $nj, $m1, $nlo ! np[j]*m1 addcc $lo0, $lo1, $lo1 ! np[j]*m1+ap[j]*bp[0] umulxhi $nj, $m1, $nj ! nhi=nj addxc %g0, $hi1, $hi1 stx $lo1, [$tp] ! tp[j-1] add $tp, 8, $tp ! tp++ brnz,pt $cnt, .L1st sub $cnt, 8, $cnt ! j-- !.L1st addcc $alo, $hi0, $lo0 addxc $aj, %g0, $hi0 ! ahi=aj addcc $nlo, $hi1, $lo1 addxc $nj, %g0, $hi1 addcc $lo0, $lo1, $lo1 ! np[j]*m1+ap[j]*bp[0] addxc %g0, $hi1, $hi1 stx $lo1, [$tp] ! tp[j-1] add $tp, 8, $tp addcc $hi0, $hi1, $hi1 addxc %g0, %g0, $ovf ! upmost overflow bit stx $hi1, [$tp] add $tp, 8, $tp ba .Louter sub $num, 16, $i ! i=num-2 .align 16 .Louter: ld [$bp+0], $t2 ! m0=bp[i] ld [$bp+4], $t3 sub $anp, $num, $anp ! rewind sub $tp, $num, $tp sub $anp, $num, $anp add $bp, 8, $bp sllx $t3, 32, $m0 ldx [$anp+0], $aj ! ap[0] or $t2, $m0, $m0 ldx [$anp+8], $nj ! np[0] mulx $aj, $m0, $lo0 ! ap[0]*bp[i] ldx [$tp], $tj ! tp[0] umulxhi $aj, $m0, $hi0 ldx [$anp+16], $aj ! ap[1] addcc $lo0, $tj, $lo0 ! ap[0]*bp[i]+tp[0] mulx $aj, $m0, $alo ! ap[1]*bp[i] addxc %g0, $hi0, $hi0 mulx $lo0, $n0, $m1 ! tp[0]*n0 umulxhi $aj, $m0, $aj ! ahi=aj mulx $nj, $m1, $lo1 ! np[0]*m1 umulxhi $nj, $m1, $hi1 ldx [$anp+24], $nj ! np[1] add $anp, 32, $anp addcc $lo1, $lo0, $lo1 mulx $nj, $m1, $nlo ! np[1]*m1 addxc %g0, $hi1, $hi1 umulxhi $nj, $m1, $nj ! nhi=nj ba .Linner sub $num, 24, $cnt ! cnt=num-3 .align 16 .Linner: addcc $alo, $hi0, $lo0 ldx [$tp+8], $tj ! tp[j] addxc $aj, %g0, $hi0 ! ahi=aj ldx [$anp+0], $aj ! ap[j] addcc $nlo, $hi1, $lo1 mulx $aj, $m0, $alo ! ap[j]*bp[i] addxc $nj, %g0, $hi1 ! nhi=nj ldx [$anp+8], $nj ! np[j] add $anp, 16, $anp umulxhi $aj, $m0, $aj ! ahi=aj addcc $lo0, $tj, $lo0 ! ap[j]*bp[i]+tp[j] mulx $nj, $m1, $nlo ! np[j]*m1 addxc %g0, $hi0, $hi0 umulxhi $nj, $m1, $nj ! nhi=nj addcc $lo1, $lo0, $lo1 ! np[j]*m1+ap[j]*bp[i]+tp[j] addxc %g0, $hi1, $hi1 stx $lo1, [$tp] ! tp[j-1] add $tp, 8, $tp brnz,pt $cnt, .Linner sub $cnt, 8, $cnt !.Linner ldx [$tp+8], $tj ! tp[j] addcc $alo, $hi0, $lo0 addxc $aj, %g0, $hi0 ! ahi=aj addcc $lo0, $tj, $lo0 ! ap[j]*bp[i]+tp[j] addxc %g0, $hi0, $hi0 addcc $nlo, $hi1, $lo1 addxc $nj, %g0, $hi1 ! nhi=nj addcc $lo1, $lo0, $lo1 ! np[j]*m1+ap[j]*bp[i]+tp[j] addxc %g0, $hi1, $hi1 stx $lo1, [$tp] ! tp[j-1] subcc %g0, $ovf, %g0 ! move upmost overflow to CCR.xcc addxccc $hi1, $hi0, $hi1 addxc %g0, %g0, $ovf stx $hi1, [$tp+8] add $tp, 16, $tp brnz,pt $i, .Louter sub $i, 8, $i sub $anp, $num, $anp ! rewind sub $tp, $num, $tp sub $anp, $num, $anp ba .Lsub subcc $num, 8, $cnt ! cnt=num-1 and clear CCR.xcc .align 16 .Lsub: ldx [$tp], $tj add $tp, 8, $tp ldx [$anp+8], $nj add $anp, 16, $anp subccc $tj, $nj, $t2 ! tp[j]-np[j] srlx $tj, 32, $tj srlx $nj, 32, $nj subccc $tj, $nj, $t3 add $rp, 8, $rp st $t2, [$rp-4] ! reverse order st $t3, [$rp-8] brnz,pt $cnt, .Lsub sub $cnt, 8, $cnt sub $anp, $num, $anp ! rewind sub $tp, $num, $tp sub $anp, $num, $anp sub $rp, $num, $rp subc $ovf, %g0, $ovf ! handle upmost overflow bit and $tp, $ovf, $ap andn $rp, $ovf, $np or $np, $ap, $ap ! ap=borrow?tp:rp ba .Lcopy sub $num, 8, $cnt .align 16 .Lcopy: ! copy or in-place refresh ld [$ap+0], $t2 ld [$ap+4], $t3 add $ap, 8, $ap stx %g0, [$tp] ! zap add $tp, 8, $tp stx %g0, [$anp] ! zap stx %g0, [$anp+8] add $anp, 16, $anp st $t3, [$rp+0] ! flip order st $t2, [$rp+4] add $rp, 8, $rp brnz $cnt, .Lcopy sub $cnt, 8, $cnt mov 1, %o0 ret restore .type bn_mul_mont_vis3, #function .size bn_mul_mont_vis3, .-bn_mul_mont_vis3 .asciz "Montgomery Multiplication for SPARCv9 VIS3, CRYPTOGAMS by " .align 4 ___ # Purpose of these subroutines is to explicitly encode VIS instructions, # so that one can compile the module without having to specify VIS # extensions on compiler command line, e.g. -xarch=v9 vs. -xarch=v9a. # Idea is to reserve for option to produce "universal" binary and let # programmer detect if current CPU is VIS capable at run-time. sub unvis3 { my ($mnemonic,$rs1,$rs2,$rd)=@_; my %bias = ( "g" => 0, "o" => 8, "l" => 16, "i" => 24 ); my ($ref,$opf); my %visopf = ( "addxc" => 0x011, "addxccc" => 0x013, "umulxhi" => 0x016 ); $ref = "$mnemonic\t$rs1,$rs2,$rd"; if ($opf=$visopf{$mnemonic}) { foreach ($rs1,$rs2,$rd) { return $ref if (!/%([goli])([0-9])/); $_=$bias{$1}+$2; } return sprintf ".word\t0x%08x !%s", 0x81b00000|$rd<<25|$rs1<<14|$opf<<5|$rs2, $ref; } else { return $ref; } } foreach (split("\n",$code)) { s/\`([^\`]*)\`/eval $1/ge; s/\b(umulxhi|addxc[c]{0,2})\s+(%[goli][0-7]),\s*(%[goli][0-7]),\s*(%[goli][0-7])/ &unvis3($1,$2,$3,$4) /ge; print $_,"\n"; } close STDOUT;