forked from luck/tmp_suning_uos_patched
6dcc5627f6
These are all functions which are invoked from elsewhere, so annotate them as global using the new SYM_FUNC_START and their ENDPROC's by SYM_FUNC_END. Make sure ENTRY/ENDPROC is not defined on X86_64, given these were the last users. Signed-off-by: Jiri Slaby <jslaby@suse.cz> Signed-off-by: Borislav Petkov <bp@suse.de> Reviewed-by: Rafael J. Wysocki <rafael.j.wysocki@intel.com> [hibernate] Reviewed-by: Boris Ostrovsky <boris.ostrovsky@oracle.com> [xen bits] Acked-by: Herbert Xu <herbert@gondor.apana.org.au> [crypto] Cc: Allison Randal <allison@lohutok.net> Cc: Andrey Ryabinin <aryabinin@virtuozzo.com> Cc: Andy Lutomirski <luto@kernel.org> Cc: Andy Shevchenko <andy@infradead.org> Cc: Ard Biesheuvel <ard.biesheuvel@linaro.org> Cc: Armijn Hemel <armijn@tjaldur.nl> Cc: Cao jin <caoj.fnst@cn.fujitsu.com> Cc: Darren Hart <dvhart@infradead.org> Cc: Dave Hansen <dave.hansen@linux.intel.com> Cc: "David S. Miller" <davem@davemloft.net> Cc: Enrico Weigelt <info@metux.net> Cc: Greg Kroah-Hartman <gregkh@linuxfoundation.org> Cc: Herbert Xu <herbert@gondor.apana.org.au> Cc: "H. Peter Anvin" <hpa@zytor.com> Cc: Ingo Molnar <mingo@redhat.com> Cc: Jim Mattson <jmattson@google.com> Cc: Joerg Roedel <joro@8bytes.org> Cc: Josh Poimboeuf <jpoimboe@redhat.com> Cc: Juergen Gross <jgross@suse.com> Cc: Kate Stewart <kstewart@linuxfoundation.org> Cc: "Kirill A. Shutemov" <kirill.shutemov@linux.intel.com> Cc: kvm ML <kvm@vger.kernel.org> Cc: Len Brown <len.brown@intel.com> Cc: linux-arch@vger.kernel.org Cc: linux-crypto@vger.kernel.org Cc: linux-efi <linux-efi@vger.kernel.org> Cc: linux-efi@vger.kernel.org Cc: linux-pm@vger.kernel.org Cc: Mark Rutland <mark.rutland@arm.com> Cc: Matt Fleming <matt@codeblueprint.co.uk> Cc: Paolo Bonzini <pbonzini@redhat.com> Cc: Pavel Machek <pavel@ucw.cz> Cc: Peter Zijlstra <peterz@infradead.org> Cc: platform-driver-x86@vger.kernel.org Cc: "Radim Krčmář" <rkrcmar@redhat.com> Cc: Sean Christopherson <sean.j.christopherson@intel.com> Cc: Stefano Stabellini <sstabellini@kernel.org> Cc: "Steven Rostedt (VMware)" <rostedt@goodmis.org> Cc: Thomas Gleixner <tglx@linutronix.de> Cc: Vitaly Kuznetsov <vkuznets@redhat.com> Cc: Wanpeng Li <wanpengli@tencent.com> Cc: Wei Huang <wei@redhat.com> Cc: x86-ml <x86@kernel.org> Cc: xen-devel@lists.xenproject.org Cc: Xiaoyao Li <xiaoyao.li@linux.intel.com> Link: https://lkml.kernel.org/r/20191011115108.12392-25-jslaby@suse.cz
134 lines
2.7 KiB
ArmAsm
134 lines
2.7 KiB
ArmAsm
/* SPDX-License-Identifier: GPL-2.0-only */
|
|
/*
|
|
* Accelerated GHASH implementation with Intel PCLMULQDQ-NI
|
|
* instructions. This file contains accelerated part of ghash
|
|
* implementation. More information about PCLMULQDQ can be found at:
|
|
*
|
|
* http://software.intel.com/en-us/articles/carry-less-multiplication-and-its-usage-for-computing-the-gcm-mode/
|
|
*
|
|
* Copyright (c) 2009 Intel Corp.
|
|
* Author: Huang Ying <ying.huang@intel.com>
|
|
* Vinodh Gopal
|
|
* Erdinc Ozturk
|
|
* Deniz Karakoyunlu
|
|
*/
|
|
|
|
#include <linux/linkage.h>
|
|
#include <asm/inst.h>
|
|
#include <asm/frame.h>
|
|
|
|
.section .rodata.cst16.bswap_mask, "aM", @progbits, 16
|
|
.align 16
|
|
.Lbswap_mask:
|
|
.octa 0x000102030405060708090a0b0c0d0e0f
|
|
|
|
#define DATA %xmm0
|
|
#define SHASH %xmm1
|
|
#define T1 %xmm2
|
|
#define T2 %xmm3
|
|
#define T3 %xmm4
|
|
#define BSWAP %xmm5
|
|
#define IN1 %xmm6
|
|
|
|
.text
|
|
|
|
/*
|
|
* __clmul_gf128mul_ble: internal ABI
|
|
* input:
|
|
* DATA: operand1
|
|
* SHASH: operand2, hash_key << 1 mod poly
|
|
* output:
|
|
* DATA: operand1 * operand2 mod poly
|
|
* changed:
|
|
* T1
|
|
* T2
|
|
* T3
|
|
*/
|
|
SYM_FUNC_START_LOCAL(__clmul_gf128mul_ble)
|
|
movaps DATA, T1
|
|
pshufd $0b01001110, DATA, T2
|
|
pshufd $0b01001110, SHASH, T3
|
|
pxor DATA, T2
|
|
pxor SHASH, T3
|
|
|
|
PCLMULQDQ 0x00 SHASH DATA # DATA = a0 * b0
|
|
PCLMULQDQ 0x11 SHASH T1 # T1 = a1 * b1
|
|
PCLMULQDQ 0x00 T3 T2 # T2 = (a1 + a0) * (b1 + b0)
|
|
pxor DATA, T2
|
|
pxor T1, T2 # T2 = a0 * b1 + a1 * b0
|
|
|
|
movaps T2, T3
|
|
pslldq $8, T3
|
|
psrldq $8, T2
|
|
pxor T3, DATA
|
|
pxor T2, T1 # <T1:DATA> is result of
|
|
# carry-less multiplication
|
|
|
|
# first phase of the reduction
|
|
movaps DATA, T3
|
|
psllq $1, T3
|
|
pxor DATA, T3
|
|
psllq $5, T3
|
|
pxor DATA, T3
|
|
psllq $57, T3
|
|
movaps T3, T2
|
|
pslldq $8, T2
|
|
psrldq $8, T3
|
|
pxor T2, DATA
|
|
pxor T3, T1
|
|
|
|
# second phase of the reduction
|
|
movaps DATA, T2
|
|
psrlq $5, T2
|
|
pxor DATA, T2
|
|
psrlq $1, T2
|
|
pxor DATA, T2
|
|
psrlq $1, T2
|
|
pxor T2, T1
|
|
pxor T1, DATA
|
|
ret
|
|
SYM_FUNC_END(__clmul_gf128mul_ble)
|
|
|
|
/* void clmul_ghash_mul(char *dst, const u128 *shash) */
|
|
SYM_FUNC_START(clmul_ghash_mul)
|
|
FRAME_BEGIN
|
|
movups (%rdi), DATA
|
|
movups (%rsi), SHASH
|
|
movaps .Lbswap_mask, BSWAP
|
|
PSHUFB_XMM BSWAP DATA
|
|
call __clmul_gf128mul_ble
|
|
PSHUFB_XMM BSWAP DATA
|
|
movups DATA, (%rdi)
|
|
FRAME_END
|
|
ret
|
|
SYM_FUNC_END(clmul_ghash_mul)
|
|
|
|
/*
|
|
* void clmul_ghash_update(char *dst, const char *src, unsigned int srclen,
|
|
* const u128 *shash);
|
|
*/
|
|
SYM_FUNC_START(clmul_ghash_update)
|
|
FRAME_BEGIN
|
|
cmp $16, %rdx
|
|
jb .Lupdate_just_ret # check length
|
|
movaps .Lbswap_mask, BSWAP
|
|
movups (%rdi), DATA
|
|
movups (%rcx), SHASH
|
|
PSHUFB_XMM BSWAP DATA
|
|
.align 4
|
|
.Lupdate_loop:
|
|
movups (%rsi), IN1
|
|
PSHUFB_XMM BSWAP IN1
|
|
pxor IN1, DATA
|
|
call __clmul_gf128mul_ble
|
|
sub $16, %rdx
|
|
add $16, %rsi
|
|
cmp $16, %rdx
|
|
jge .Lupdate_loop
|
|
PSHUFB_XMM BSWAP DATA
|
|
movups DATA, (%rdi)
|
|
.Lupdate_just_ret:
|
|
FRAME_END
|
|
ret
|
|
SYM_FUNC_END(clmul_ghash_update)
|