 00f463b38a
			
		
	
	
		00f463b38a
		
	
	
	
	
		
			
			Reviewed-by: Ard Biesheuvel <ardb@kernel.org> Signed-off-by: Richard Henderson <richard.henderson@linaro.org>
		
			
				
	
	
		
			84 lines
		
	
	
		
			1.5 KiB
		
	
	
	
		
			C
		
	
	
	
	
	
			
		
		
	
	
			84 lines
		
	
	
		
			1.5 KiB
		
	
	
	
		
			C
		
	
	
	
	
	
| /*
 | |
|  * Carry-less multiply operations.
 | |
|  * SPDX-License-Identifier: GPL-2.0-or-later
 | |
|  *
 | |
|  * Copyright (C) 2023 Linaro, Ltd.
 | |
|  */
 | |
| 
 | |
| #ifndef CRYPTO_CLMUL_H
 | |
| #define CRYPTO_CLMUL_H
 | |
| 
 | |
| #include "qemu/int128.h"
 | |
| #include "host/crypto/clmul.h"
 | |
| 
 | |
| /**
 | |
|  * clmul_8x8_low:
 | |
|  *
 | |
|  * Perform eight 8x8->8 carry-less multiplies.
 | |
|  */
 | |
| uint64_t clmul_8x8_low(uint64_t, uint64_t);
 | |
| 
 | |
| /**
 | |
|  * clmul_8x4_even:
 | |
|  *
 | |
|  * Perform four 8x8->16 carry-less multiplies.
 | |
|  * The odd bytes of the inputs are ignored.
 | |
|  */
 | |
| uint64_t clmul_8x4_even(uint64_t, uint64_t);
 | |
| 
 | |
| /**
 | |
|  * clmul_8x4_odd:
 | |
|  *
 | |
|  * Perform four 8x8->16 carry-less multiplies.
 | |
|  * The even bytes of the inputs are ignored.
 | |
|  */
 | |
| uint64_t clmul_8x4_odd(uint64_t, uint64_t);
 | |
| 
 | |
| /**
 | |
|  * clmul_8x4_packed:
 | |
|  *
 | |
|  * Perform four 8x8->16 carry-less multiplies.
 | |
|  */
 | |
| uint64_t clmul_8x4_packed(uint32_t, uint32_t);
 | |
| 
 | |
| /**
 | |
|  * clmul_16x2_even:
 | |
|  *
 | |
|  * Perform two 16x16->32 carry-less multiplies.
 | |
|  * The odd words of the inputs are ignored.
 | |
|  */
 | |
| uint64_t clmul_16x2_even(uint64_t, uint64_t);
 | |
| 
 | |
| /**
 | |
|  * clmul_16x2_odd:
 | |
|  *
 | |
|  * Perform two 16x16->32 carry-less multiplies.
 | |
|  * The even words of the inputs are ignored.
 | |
|  */
 | |
| uint64_t clmul_16x2_odd(uint64_t, uint64_t);
 | |
| 
 | |
| /**
 | |
|  * clmul_32:
 | |
|  *
 | |
|  * Perform a 32x32->64 carry-less multiply.
 | |
|  */
 | |
| uint64_t clmul_32(uint32_t, uint32_t);
 | |
| 
 | |
| /**
 | |
|  * clmul_64:
 | |
|  *
 | |
|  * Perform a 64x64->128 carry-less multiply.
 | |
|  */
 | |
| Int128 clmul_64_gen(uint64_t, uint64_t);
 | |
| 
 | |
| static inline Int128 clmul_64(uint64_t a, uint64_t b)
 | |
| {
 | |
|     if (HAVE_CLMUL_ACCEL) {
 | |
|         return clmul_64_accel(a, b);
 | |
|     } else {
 | |
|         return clmul_64_gen(a, b);
 | |
|     }
 | |
| }
 | |
| 
 | |
| #endif /* CRYPTO_CLMUL_H */
 |