#!/usr/bin/env perl # Copyright 2017-2018 The OpenSSL Project Authors. All Rights Reserved. # # Licensed under the OpenSSL license (the "License"). You may not use # this file except in compliance with the License. You can obtain a copy # in the file LICENSE in the source distribution or at # https://www.openssl.org/source/license.html # # ==================================================================== # Written by Andy Polyakov for the OpenSSL # project. The module is, however, dual licensed under OpenSSL and # CRYPTOGAMS licenses depending on where you obtain it. For further # details see http://www.openssl.org/~appro/cryptogams/. # ==================================================================== # # Keccak-1600 for x86_64. # # June 2017. # # Below code is [lane complementing] KECCAK_2X implementation (see # sha/keccak1600.c) with C[5] and D[5] held in register bank. Though # instead of actually unrolling the loop pair-wise I simply flip # pointers to T[][] and A[][] at the end of round. Since number of # rounds is even, last round writes to A[][] and everything works out. # How does it compare to x86_64 assembly module in Keccak Code Package? # Depending on processor it's either as fast or faster by up to 15%... # ######################################################################## # Numbers are cycles per processed byte out of large message. # # r=1088(*) # # P4 25.8 # Core 2 12.9 # Westmere 13.7 # Sandy Bridge 12.9(**) # Haswell 9.6 # Skylake 9.4 # Silvermont 22.8 # Goldmont 15.8 # VIA Nano 17.3 # Sledgehammer 13.3 # Bulldozer 16.5 # Ryzen 8.8 # # (*) Corresponds to SHA3-256. Improvement over compiler-generate # varies a lot, most commont coefficient is 15% in comparison to # gcc-5.x, 50% for gcc-4.x, 90% for gcc-3.x. # (**) Sandy Bridge has broken rotate instruction. Performance can be # improved by 14% by replacing rotates with double-precision # shift with same register as source and destination. $flavour = shift; $output = shift; if ($flavour =~ /\./) { $output = $flavour; undef $flavour; } $win64=0; $win64=1 if ($flavour =~ /[nm]asm|mingw64/ || $output =~ /\.asm$/); $0 =~ m/(.*[\/\\])[^\/\\]+$/; $dir=$1; ( $xlate="${dir}x86_64-xlate.pl" and -f $xlate ) or ( $xlate="${dir}../../perlasm/x86_64-xlate.pl" and -f $xlate) or die "can't locate x86_64-xlate.pl"; open OUT,"| \"$^X\" \"$xlate\" $flavour \"$output\""; *STDOUT=*OUT; my @A = map([ 8*$_-100, 8*($_+1)-100, 8*($_+2)-100, 8*($_+3)-100, 8*($_+4)-100 ], (0,5,10,15,20)); my @C = ("%rax","%rbx","%rcx","%rdx","%rbp"); my @D = map("%r$_",(8..12)); my @T = map("%r$_",(13..14)); my $iotas = "%r15"; my @rhotates = ([ 0, 1, 62, 28, 27 ], [ 36, 44, 6, 55, 20 ], [ 3, 10, 43, 25, 39 ], [ 41, 45, 15, 21, 8 ], [ 18, 2, 61, 56, 14 ]); $code.=<<___; .text .type __KeccakF1600,\@abi-omnipotent .align 32 __KeccakF1600: .cfi_startproc mov $A[4][0](%rdi),@C[0] mov $A[4][1](%rdi),@C[1] mov $A[4][2](%rdi),@C[2] mov $A[4][3](%rdi),@C[3] mov $A[4][4](%rdi),@C[4] jmp .Loop .align 32 .Loop: mov $A[0][0](%rdi),@D[0] mov $A[1][1](%rdi),@D[1] mov $A[2][2](%rdi),@D[2] mov $A[3][3](%rdi),@D[3] xor $A[0][2](%rdi),@C[2] xor $A[0][3](%rdi),@C[3] xor @D[0], @C[0] xor $A[0][1](%rdi),@C[1] xor $A[1][2](%rdi),@C[2] xor $A[1][0](%rdi),@C[0] mov @C[4],@D[4] xor $A[0][4](%rdi),@C[4] xor @D[2], @C[2] xor $A[2][0](%rdi),@C[0] xor $A[1][3](%rdi),@C[3] xor @D[1], @C[1] xor $A[1][4](%rdi),@C[4] xor $A[3][2](%rdi),@C[2] xor $A[3][0](%rdi),@C[0] xor $A[2][3](%rdi),@C[3] xor $A[2][1](%rdi),@C[1] xor $A[2][4](%rdi),@C[4] mov @C[2],@T[0] rol \$1,@C[2] xor @C[0],@C[2] # D[1] = ROL64(C[2], 1) ^ C[0] xor @D[3], @C[3] rol \$1,@C[0] xor @C[3],@C[0] # D[4] = ROL64(C[0], 1) ^ C[3] xor $A[3][1](%rdi),@C[1] rol \$1,@C[3] xor @C[1],@C[3] # D[2] = ROL64(C[3], 1) ^ C[1] xor $A[3][4](%rdi),@C[4] rol \$1,@C[1] xor @C[4],@C[1] # D[0] = ROL64(C[1], 1) ^ C[4] rol \$1,@C[4] xor @T[0],@C[4] # D[3] = ROL64(C[4], 1) ^ C[2] ___ (@D[0..4], @C) = (@C[1..4,0], @D); $code.=<<___; xor @D[1],@C[1] xor @D[2],@C[2] rol \$$rhotates[1][1],@C[1] xor @D[3],@C[3] xor @D[4],@C[4] rol \$$rhotates[2][2],@C[2] xor @D[0],@C[0] mov @C[1],@T[0] rol \$$rhotates[3][3],@C[3] or @C[2],@C[1] xor @C[0],@C[1] # C[0] ^ ( C[1] | C[2]) rol \$$rhotates[4][4],@C[4] xor ($iotas),@C[1] lea 8($iotas),$iotas mov @C[4],@T[1] and @C[3],@C[4] mov @C[1],$A[0][0](%rsi) # R[0][0] = C[0] ^ ( C[1] | C[2]) ^ iotas[i] xor @C[2],@C[4] # C[2] ^ ( C[4] & C[3]) not @C[2] mov @C[4],$A[0][2](%rsi) # R[0][2] = C[2] ^ ( C[4] & C[3]) or @C[3],@C[2] mov $A[4][2](%rdi),@C[4] xor @T[0],@C[2] # C[1] ^ (~C[2] | C[3]) mov @C[2],$A[0][1](%rsi) # R[0][1] = C[1] ^ (~C[2] | C[3]) and @C[0],@T[0] mov $A[1][4](%rdi),@C[1] xor @T[1],@T[0] # C[4] ^ ( C[1] & C[0]) mov $A[2][0](%rdi),@C[2] mov @T[0],$A[0][4](%rsi) # R[0][4] = C[4] ^ ( C[1] & C[0]) or @C[0],@T[1] mov $A[0][3](%rdi),@C[0] xor @C[3],@T[1] # C[3] ^ ( C[4] | C[0]) mov $A[3][1](%rdi),@C[3] mov @T[1],$A[0][3](%rsi) # R[0][3] = C[3] ^ ( C[4] | C[0]) xor @D[3],@C[0] xor @D[2],@C[4] rol \$$rhotates[0][3],@C[0] xor @D[1],@C[3] xor @D[4],@C[1] rol \$$rhotates[4][2],@C[4] rol \$$rhotates[3][1],@C[3] xor @D[0],@C[2] rol \$$rhotates[1][4],@C[1] mov @C[0],@T[0] or @C[4],@C[0] rol \$$rhotates[2][0],@C[2] xor @C[3],@C[0] # C[3] ^ (C[0] | C[4]) mov @C[0],$A[1][3](%rsi) # R[1][3] = C[3] ^ (C[0] | C[4]) mov @C[1],@T[1] and @T[0],@C[1] mov $A[0][1](%rdi),@C[0] xor @C[4],@C[1] # C[4] ^ (C[1] & C[0]) not @C[4] mov @C[1],$A[1][4](%rsi) # R[1][4] = C[4] ^ (C[1] & C[0]) or @C[3],@C[4] mov $A[1][2](%rdi),@C[1] xor @C[2],@C[4] # C[2] ^ (~C[4] | C[3]) mov @C[4],$A[1][2](%rsi) # R[1][2] = C[2] ^ (~C[4] | C[3]) and @C[2],@C[3] mov $A[4][0](%rdi),@C[4] xor @T[1],@C[3] # C[1] ^ (C[3] & C[2]) mov @C[3],$A[1][1](%rsi) # R[1][1] = C[1] ^ (C[3] & C[2]) or @C[2],@T[1] mov $A[2][3](%rdi),@C[2] xor @T[0],@T[1] # C[0] ^ (C[1] | C[2]) mov $A[3][4](%rdi),@C[3] mov @T[1],$A[1][0](%rsi) # R[1][0] = C[0] ^ (C[1] | C[2]) xor @D[3],@C[2] xor @D[4],@C[3] rol \$$rhotates[2][3],@C[2] xor @D[2],@C[1] rol \$$rhotates[3][4],@C[3] xor @D[0],@C[4] rol \$$rhotates[1][2],@C[1] xor @D[1],@C[0] rol \$$rhotates[4][0],@C[4] mov @C[2],@T[0] and @C[3],@C[2] rol \$$rhotates[0][1],@C[0] not @C[3] xor @C[1],@C[2] # C[1] ^ ( C[2] & C[3]) mov @C[2],$A[2][1](%rsi) # R[2][1] = C[1] ^ ( C[2] & C[3]) mov @C[4],@T[1] and @C[3],@C[4] mov $A[2][1](%rdi),@C[2] xor @T[0],@C[4] # C[2] ^ ( C[4] & ~C[3]) mov @C[4],$A[2][2](%rsi) # R[2][2] = C[2] ^ ( C[4] & ~C[3]) or @C[1],@T[0] mov $A[4][3](%rdi),@C[4] xor @C[0],@T[0] # C[0] ^ ( C[2] | C[1]) mov @T[0],$A[2][0](%rsi) # R[2][0] = C[0] ^ ( C[2] | C[1]) and @C[0],@C[1] xor @T[1],@C[1] # C[4] ^ ( C[1] & C[0]) mov @C[1],$A[2][4](%rsi) # R[2][4] = C[4] ^ ( C[1] & C[0]) or @C[0],@T[1] mov $A[1][0](%rdi),@C[1] xor @C[3],@T[1] # ~C[3] ^ ( C[0] | C[4]) mov $A[3][2](%rdi),@C[3] mov @T[1],$A[2][3](%rsi) # R[2][3] = ~C[3] ^ ( C[0] | C[4]) mov $A[0][4](%rdi),@C[0] xor @D[1],@C[2] xor @D[2],@C[3] rol \$$rhotates[2][1],@C[2] xor @D[0],@C[1] rol \$$rhotates[3][2],@C[3] xor @D[3],@C[4] rol \$$rhotates[1][0],@C[1] xor @D[4],@C[0] rol \$$rhotates[4][3],@C[4] mov @C[2],@T[0] or @C[3],@C[2] rol \$$rhotates[0][4],@C[0] not @C[3] xor @C[1],@C[2] # C[1] ^ ( C[2] | C[3]) mov @C[2],$A[3][1](%rsi) # R[3][1] = C[1] ^ ( C[2] | C[3]) mov @C[4],@T[1] or @C[3],@C[4] xor @T[0],@C[4] # C[2] ^ ( C[4] | ~C[3]) mov @C[4],$A[3][2](%rsi) # R[3][2] = C[2] ^ ( C[4] | ~C[3]) and @C[1],@T[0] xor @C[0],@T[0] # C[0] ^ ( C[2] & C[1]) mov @T[0],$A[3][0](%rsi) # R[3][0] = C[0] ^ ( C[2] & C[1]) or @C[0],@C[1] xor @T[1],@C[1] # C[4] ^ ( C[1] | C[0]) mov @C[1],$A[3][4](%rsi) # R[3][4] = C[4] ^ ( C[1] | C[0]) and @T[1],@C[0] xor @C[3],@C[0] # ~C[3] ^ ( C[0] & C[4]) mov @C[0],$A[3][3](%rsi) # R[3][3] = ~C[3] ^ ( C[0] & C[4]) xor $A[0][2](%rdi),@D[2] xor $A[1][3](%rdi),@D[3] rol \$$rhotates[0][2],@D[2] xor $A[4][1](%rdi),@D[1] rol \$$rhotates[1][3],@D[3] xor $A[2][4](%rdi),@D[4] rol \$$rhotates[4][1],@D[1] xor $A[3][0](%rdi),@D[0] xchg %rsi,%rdi rol \$$rhotates[2][4],@D[4] rol \$$rhotates[3][0],@D[0] ___ @C = @D[2..4,0,1]; $code.=<<___; mov @C[0],@T[0] and @C[1],@C[0] not @C[1] xor @C[4],@C[0] # C[4] ^ ( C[0] & C[1]) mov @C[0],$A[4][4](%rdi) # R[4][4] = C[4] ^ ( C[0] & C[1]) mov @C[2],@T[1] and @C[1],@C[2] xor @T[0],@C[2] # C[0] ^ ( C[2] & ~C[1]) mov @C[2],$A[4][0](%rdi) # R[4][0] = C[0] ^ ( C[2] & ~C[1]) or @C[4],@T[0] xor @C[3],@T[0] # C[3] ^ ( C[0] | C[4]) mov @T[0],$A[4][3](%rdi) # R[4][3] = C[3] ^ ( C[0] | C[4]) and @C[3],@C[4] xor @T[1],@C[4] # C[2] ^ ( C[4] & C[3]) mov @C[4],$A[4][2](%rdi) # R[4][2] = C[2] ^ ( C[4] & C[3]) or @T[1],@C[3] xor @C[1],@C[3] # ~C[1] ^ ( C[2] | C[3]) mov @C[3],$A[4][1](%rdi) # R[4][1] = ~C[1] ^ ( C[2] | C[3]) mov @C[0],@C[1] # harmonize with the loop top mov @T[0],@C[0] test \$255,$iotas jnz .Loop lea -192($iotas),$iotas # rewind iotas ret .cfi_endproc .size __KeccakF1600,.-__KeccakF1600 .type KeccakF1600,\@abi-omnipotent .align 32 KeccakF1600: .cfi_startproc push %rbx .cfi_push %rbx push %rbp .cfi_push %rbp push %r12 .cfi_push %r12 push %r13 .cfi_push %r13 push %r14 .cfi_push %r14 push %r15 .cfi_push %r15 lea 100(%rdi),%rdi # size optimization sub \$200,%rsp .cfi_adjust_cfa_offset 200 notq $A[0][1](%rdi) notq $A[0][2](%rdi) notq $A[1][3](%rdi) notq $A[2][2](%rdi) notq $A[3][2](%rdi) notq $A[4][0](%rdi) lea iotas(%rip),$iotas lea 100(%rsp),%rsi # size optimization call __KeccakF1600 notq $A[0][1](%rdi) notq $A[0][2](%rdi) notq $A[1][3](%rdi) notq $A[2][2](%rdi) notq $A[3][2](%rdi) notq $A[4][0](%rdi) lea -100(%rdi),%rdi # preserve A[][] add \$200,%rsp .cfi_adjust_cfa_offset -200 pop %r15 .cfi_pop %r15 pop %r14 .cfi_pop %r14 pop %r13 .cfi_pop %r13 pop %r12 .cfi_pop %r12 pop %rbp .cfi_pop %rbp pop %rbx .cfi_pop %rbx ret .cfi_endproc .size KeccakF1600,.-KeccakF1600 ___ { my ($A_flat,$inp,$len,$bsz) = ("%rdi","%rsi","%rdx","%rcx"); ($A_flat,$inp) = ("%r8","%r9"); $code.=<<___; .globl SHA3_absorb .type SHA3_absorb,\@function,4 .align 32 SHA3_absorb: .cfi_startproc push %rbx .cfi_push %rbx push %rbp .cfi_push %rbp push %r12 .cfi_push %r12 push %r13 .cfi_push %r13 push %r14 .cfi_push %r14 push %r15 .cfi_push %r15 lea 100(%rdi),%rdi # size optimization sub \$232,%rsp .cfi_adjust_cfa_offset 232 mov %rsi,$inp lea 100(%rsp),%rsi # size optimization notq $A[0][1](%rdi) notq $A[0][2](%rdi) notq $A[1][3](%rdi) notq $A[2][2](%rdi) notq $A[3][2](%rdi) notq $A[4][0](%rdi) lea iotas(%rip),$iotas mov $bsz,216-100(%rsi) # save bsz .Loop_absorb: cmp $bsz,$len jc .Ldone_absorb shr \$3,$bsz lea -100(%rdi),$A_flat .Lblock_absorb: mov ($inp),%rax lea 8($inp),$inp xor ($A_flat),%rax lea 8($A_flat),$A_flat sub \$8,$len mov %rax,-8($A_flat) sub \$1,$bsz jnz .Lblock_absorb mov $inp,200-100(%rsi) # save inp mov $len,208-100(%rsi) # save len call __KeccakF1600 mov 200-100(%rsi),$inp # pull inp mov 208-100(%rsi),$len # pull len mov 216-100(%rsi),$bsz # pull bsz jmp .Loop_absorb .align 32 .Ldone_absorb: mov $len,%rax # return value notq $A[0][1](%rdi) notq $A[0][2](%rdi) notq $A[1][3](%rdi) notq $A[2][2](%rdi) notq $A[3][2](%rdi) notq $A[4][0](%rdi) add \$232,%rsp .cfi_adjust_cfa_offset -232 pop %r15 .cfi_pop %r15 pop %r14 .cfi_pop %r14 pop %r13 .cfi_pop %r13 pop %r12 .cfi_pop %r12 pop %rbp .cfi_pop %rbp pop %rbx .cfi_pop %rbx ret .cfi_endproc .size SHA3_absorb,.-SHA3_absorb ___ } { my ($A_flat,$out,$len,$bsz) = ("%rdi","%rsi","%rdx","%rcx"); ($out,$len,$bsz) = ("%r12","%r13","%r14"); $code.=<<___; .globl SHA3_squeeze .type SHA3_squeeze,\@function,4 .align 32 SHA3_squeeze: .cfi_startproc push %r12 .cfi_push %r12 push %r13 .cfi_push %r13 push %r14 .cfi_push %r14 shr \$3,%rcx mov $A_flat,%r8 mov %rsi,$out mov %rdx,$len mov %rcx,$bsz jmp .Loop_squeeze .align 32 .Loop_squeeze: cmp \$8,$len jb .Ltail_squeeze mov (%r8),%rax lea 8(%r8),%r8 mov %rax,($out) lea 8($out),$out sub \$8,$len # len -= 8 jz .Ldone_squeeze sub \$1,%rcx # bsz-- jnz .Loop_squeeze call KeccakF1600 mov $A_flat,%r8 mov $bsz,%rcx jmp .Loop_squeeze .Ltail_squeeze: mov %r8, %rsi mov $out,%rdi mov $len,%rcx .byte 0xf3,0xa4 # rep movsb .Ldone_squeeze: pop %r14 .cfi_pop %r14 pop %r13 .cfi_pop %r13 pop %r12 .cfi_pop %r13 ret .cfi_endproc .size SHA3_squeeze,.-SHA3_squeeze ___ } $code.=<<___; .align 256 .quad 0,0,0,0,0,0,0,0 .type iotas,\@object iotas: .quad 0x0000000000000001 .quad 0x0000000000008082 .quad 0x800000000000808a .quad 0x8000000080008000 .quad 0x000000000000808b .quad 0x0000000080000001 .quad 0x8000000080008081 .quad 0x8000000000008009 .quad 0x000000000000008a .quad 0x0000000000000088 .quad 0x0000000080008009 .quad 0x000000008000000a .quad 0x000000008000808b .quad 0x800000000000008b .quad 0x8000000000008089 .quad 0x8000000000008003 .quad 0x8000000000008002 .quad 0x8000000000000080 .quad 0x000000000000800a .quad 0x800000008000000a .quad 0x8000000080008081 .quad 0x8000000000008080 .quad 0x0000000080000001 .quad 0x8000000080008008 .size iotas,.-iotas .asciz "Keccak-1600 absorb and squeeze for x86_64, CRYPTOGAMS by " ___ foreach (split("\n",$code)) { # Below replacement results in 11.2 on Sandy Bridge, 9.4 on # Haswell, but it hurts other processors by up to 2-3-4x... #s/rol\s+(\$[0-9]+),(%[a-z][a-z0-9]+)/shld\t$1,$2,$2/; # Below replacement results in 9.3 on Haswell [as well as # on Ryzen, i.e. it *hurts* Ryzen]... #s/rol\s+\$([0-9]+),(%[a-z][a-z0-9]+)/rorx\t\$64-$1,$2,$2/; print $_, "\n"; } close STDOUT;