Reviewed-by: Ard Biesheuvel <ardb@kernel.org> Signed-off-by: Richard Henderson <richard.henderson@linaro.org>
		
			
				
	
	
		
			84 lines
		
	
	
		
			1.5 KiB
		
	
	
	
		
			C
		
	
	
	
	
	
			
		
		
	
	
			84 lines
		
	
	
		
			1.5 KiB
		
	
	
	
		
			C
		
	
	
	
	
	
/*
 | 
						|
 * Carry-less multiply operations.
 | 
						|
 * SPDX-License-Identifier: GPL-2.0-or-later
 | 
						|
 *
 | 
						|
 * Copyright (C) 2023 Linaro, Ltd.
 | 
						|
 */
 | 
						|
 | 
						|
#ifndef CRYPTO_CLMUL_H
 | 
						|
#define CRYPTO_CLMUL_H
 | 
						|
 | 
						|
#include "qemu/int128.h"
 | 
						|
#include "host/crypto/clmul.h"
 | 
						|
 | 
						|
/**
 | 
						|
 * clmul_8x8_low:
 | 
						|
 *
 | 
						|
 * Perform eight 8x8->8 carry-less multiplies.
 | 
						|
 */
 | 
						|
uint64_t clmul_8x8_low(uint64_t, uint64_t);
 | 
						|
 | 
						|
/**
 | 
						|
 * clmul_8x4_even:
 | 
						|
 *
 | 
						|
 * Perform four 8x8->16 carry-less multiplies.
 | 
						|
 * The odd bytes of the inputs are ignored.
 | 
						|
 */
 | 
						|
uint64_t clmul_8x4_even(uint64_t, uint64_t);
 | 
						|
 | 
						|
/**
 | 
						|
 * clmul_8x4_odd:
 | 
						|
 *
 | 
						|
 * Perform four 8x8->16 carry-less multiplies.
 | 
						|
 * The even bytes of the inputs are ignored.
 | 
						|
 */
 | 
						|
uint64_t clmul_8x4_odd(uint64_t, uint64_t);
 | 
						|
 | 
						|
/**
 | 
						|
 * clmul_8x4_packed:
 | 
						|
 *
 | 
						|
 * Perform four 8x8->16 carry-less multiplies.
 | 
						|
 */
 | 
						|
uint64_t clmul_8x4_packed(uint32_t, uint32_t);
 | 
						|
 | 
						|
/**
 | 
						|
 * clmul_16x2_even:
 | 
						|
 *
 | 
						|
 * Perform two 16x16->32 carry-less multiplies.
 | 
						|
 * The odd words of the inputs are ignored.
 | 
						|
 */
 | 
						|
uint64_t clmul_16x2_even(uint64_t, uint64_t);
 | 
						|
 | 
						|
/**
 | 
						|
 * clmul_16x2_odd:
 | 
						|
 *
 | 
						|
 * Perform two 16x16->32 carry-less multiplies.
 | 
						|
 * The even words of the inputs are ignored.
 | 
						|
 */
 | 
						|
uint64_t clmul_16x2_odd(uint64_t, uint64_t);
 | 
						|
 | 
						|
/**
 | 
						|
 * clmul_32:
 | 
						|
 *
 | 
						|
 * Perform a 32x32->64 carry-less multiply.
 | 
						|
 */
 | 
						|
uint64_t clmul_32(uint32_t, uint32_t);
 | 
						|
 | 
						|
/**
 | 
						|
 * clmul_64:
 | 
						|
 *
 | 
						|
 * Perform a 64x64->128 carry-less multiply.
 | 
						|
 */
 | 
						|
Int128 clmul_64_gen(uint64_t, uint64_t);
 | 
						|
 | 
						|
static inline Int128 clmul_64(uint64_t a, uint64_t b)
 | 
						|
{
 | 
						|
    if (HAVE_CLMUL_ACCEL) {
 | 
						|
        return clmul_64_accel(a, b);
 | 
						|
    } else {
 | 
						|
        return clmul_64_gen(a, b);
 | 
						|
    }
 | 
						|
}
 | 
						|
 | 
						|
#endif /* CRYPTO_CLMUL_H */
 |