 393c908afb
			
		
	
	
		393c908afb
		
	
	
	
	
		
			
			Introduce a new ast2700 class to support AST2700. Signed-off-by: Jamin Lin <jamin_lin@aspeedtech.com> Reviewed-by: Andrew Jeffery <andrew@codeconstruct.com.au> Link: https://lore.kernel.org/qemu-devel/20250225075622.305515-3-jamin_lin@aspeedtech.com Signed-off-by: Cédric Le Goater <clg@redhat.com>
		
			
				
	
	
		
			585 lines
		
	
	
		
			18 KiB
		
	
	
	
		
			C
		
	
	
	
	
	
			
		
		
	
	
			585 lines
		
	
	
		
			18 KiB
		
	
	
	
		
			C
		
	
	
	
	
	
| /*
 | |
|  * ASPEED Hash and Crypto Engine
 | |
|  *
 | |
|  * Copyright (c) 2024 Seagate Technology LLC and/or its Affiliates
 | |
|  * Copyright (C) 2021 IBM Corp.
 | |
|  *
 | |
|  * Joel Stanley <joel@jms.id.au>
 | |
|  *
 | |
|  * SPDX-License-Identifier: GPL-2.0-or-later
 | |
|  */
 | |
| 
 | |
| #include "qemu/osdep.h"
 | |
| #include "qemu/log.h"
 | |
| #include "qemu/error-report.h"
 | |
| #include "hw/misc/aspeed_hace.h"
 | |
| #include "qapi/error.h"
 | |
| #include "migration/vmstate.h"
 | |
| #include "crypto/hash.h"
 | |
| #include "hw/qdev-properties.h"
 | |
| #include "hw/irq.h"
 | |
| 
 | |
| #define R_CRYPT_CMD     (0x10 / 4)
 | |
| 
 | |
| #define R_STATUS        (0x1c / 4)
 | |
| #define HASH_IRQ        BIT(9)
 | |
| #define CRYPT_IRQ       BIT(12)
 | |
| #define TAG_IRQ         BIT(15)
 | |
| 
 | |
| #define R_HASH_SRC      (0x20 / 4)
 | |
| #define R_HASH_DEST     (0x24 / 4)
 | |
| #define R_HASH_KEY_BUFF (0x28 / 4)
 | |
| #define R_HASH_SRC_LEN  (0x2c / 4)
 | |
| 
 | |
| #define R_HASH_CMD      (0x30 / 4)
 | |
| /* Hash algorithm selection */
 | |
| #define  HASH_ALGO_MASK                 (BIT(4) | BIT(5) | BIT(6))
 | |
| #define  HASH_ALGO_MD5                  0
 | |
| #define  HASH_ALGO_SHA1                 BIT(5)
 | |
| #define  HASH_ALGO_SHA224               BIT(6)
 | |
| #define  HASH_ALGO_SHA256               (BIT(4) | BIT(6))
 | |
| #define  HASH_ALGO_SHA512_SERIES        (BIT(5) | BIT(6))
 | |
| /* SHA512 algorithm selection */
 | |
| #define  SHA512_HASH_ALGO_MASK          (BIT(10) | BIT(11) | BIT(12))
 | |
| #define  HASH_ALGO_SHA512_SHA512        0
 | |
| #define  HASH_ALGO_SHA512_SHA384        BIT(10)
 | |
| #define  HASH_ALGO_SHA512_SHA256        BIT(11)
 | |
| #define  HASH_ALGO_SHA512_SHA224        (BIT(10) | BIT(11))
 | |
| /* HMAC modes */
 | |
| #define  HASH_HMAC_MASK                 (BIT(7) | BIT(8))
 | |
| #define  HASH_DIGEST                    0
 | |
| #define  HASH_DIGEST_HMAC               BIT(7)
 | |
| #define  HASH_DIGEST_ACCUM              BIT(8)
 | |
| #define  HASH_HMAC_KEY                  (BIT(7) | BIT(8))
 | |
| /* Cascaded operation modes */
 | |
| #define  HASH_ONLY                      0
 | |
| #define  HASH_ONLY2                     BIT(0)
 | |
| #define  HASH_CRYPT_THEN_HASH           BIT(1)
 | |
| #define  HASH_HASH_THEN_CRYPT           (BIT(0) | BIT(1))
 | |
| /* Other cmd bits */
 | |
| #define  HASH_IRQ_EN                    BIT(9)
 | |
| #define  HASH_SG_EN                     BIT(18)
 | |
| /* Scatter-gather data list */
 | |
| #define SG_LIST_LEN_SIZE                4
 | |
| #define SG_LIST_LEN_MASK                0x0FFFFFFF
 | |
| #define SG_LIST_LEN_LAST                BIT(31)
 | |
| #define SG_LIST_ADDR_SIZE               4
 | |
| #define SG_LIST_ADDR_MASK               0x7FFFFFFF
 | |
| #define SG_LIST_ENTRY_SIZE              (SG_LIST_LEN_SIZE + SG_LIST_ADDR_SIZE)
 | |
| 
 | |
| static const struct {
 | |
|     uint32_t mask;
 | |
|     QCryptoHashAlgo algo;
 | |
| } hash_algo_map[] = {
 | |
|     { HASH_ALGO_MD5, QCRYPTO_HASH_ALGO_MD5 },
 | |
|     { HASH_ALGO_SHA1, QCRYPTO_HASH_ALGO_SHA1 },
 | |
|     { HASH_ALGO_SHA224, QCRYPTO_HASH_ALGO_SHA224 },
 | |
|     { HASH_ALGO_SHA256, QCRYPTO_HASH_ALGO_SHA256 },
 | |
|     { HASH_ALGO_SHA512_SERIES | HASH_ALGO_SHA512_SHA512,
 | |
|       QCRYPTO_HASH_ALGO_SHA512 },
 | |
|     { HASH_ALGO_SHA512_SERIES | HASH_ALGO_SHA512_SHA384,
 | |
|       QCRYPTO_HASH_ALGO_SHA384 },
 | |
|     { HASH_ALGO_SHA512_SERIES | HASH_ALGO_SHA512_SHA256,
 | |
|       QCRYPTO_HASH_ALGO_SHA256 },
 | |
| };
 | |
| 
 | |
| static int hash_algo_lookup(uint32_t reg)
 | |
| {
 | |
|     int i;
 | |
| 
 | |
|     reg &= HASH_ALGO_MASK | SHA512_HASH_ALGO_MASK;
 | |
| 
 | |
|     for (i = 0; i < ARRAY_SIZE(hash_algo_map); i++) {
 | |
|         if (reg == hash_algo_map[i].mask) {
 | |
|             return hash_algo_map[i].algo;
 | |
|         }
 | |
|     }
 | |
| 
 | |
|     return -1;
 | |
| }
 | |
| 
 | |
| /**
 | |
|  * Check whether the request contains padding message.
 | |
|  *
 | |
|  * @param s             aspeed hace state object
 | |
|  * @param iov           iov of current request
 | |
|  * @param req_len       length of the current request
 | |
|  * @param total_msg_len length of all acc_mode requests(excluding padding msg)
 | |
|  * @param pad_offset    start offset of padding message
 | |
|  */
 | |
| static bool has_padding(AspeedHACEState *s, struct iovec *iov,
 | |
|                         hwaddr req_len, uint32_t *total_msg_len,
 | |
|                         uint32_t *pad_offset)
 | |
| {
 | |
|     *total_msg_len = (uint32_t)(ldq_be_p(iov->iov_base + req_len - 8) / 8);
 | |
|     /*
 | |
|      * SG_LIST_LEN_LAST asserted in the request length doesn't mean it is the
 | |
|      * last request. The last request should contain padding message.
 | |
|      * We check whether message contains padding by
 | |
|      *   1. Get total message length. If the current message contains
 | |
|      *      padding, the last 8 bytes are total message length.
 | |
|      *   2. Check whether the total message length is valid.
 | |
|      *      If it is valid, the value should less than or equal to
 | |
|      *      total_req_len.
 | |
|      *   3. Current request len - padding_size to get padding offset.
 | |
|      *      The padding message's first byte should be 0x80
 | |
|      */
 | |
|     if (*total_msg_len <= s->total_req_len) {
 | |
|         uint32_t padding_size = s->total_req_len - *total_msg_len;
 | |
|         uint8_t *padding = iov->iov_base;
 | |
|         *pad_offset = req_len - padding_size;
 | |
|         if (padding[*pad_offset] == 0x80) {
 | |
|             return true;
 | |
|         }
 | |
|     }
 | |
| 
 | |
|     return false;
 | |
| }
 | |
| 
 | |
| static int reconstruct_iov(AspeedHACEState *s, struct iovec *iov, int id,
 | |
|                            uint32_t *pad_offset)
 | |
| {
 | |
|     int i, iov_count;
 | |
|     if (*pad_offset != 0) {
 | |
|         s->iov_cache[s->iov_count].iov_base = iov[id].iov_base;
 | |
|         s->iov_cache[s->iov_count].iov_len = *pad_offset;
 | |
|         ++s->iov_count;
 | |
|     }
 | |
|     for (i = 0; i < s->iov_count; i++) {
 | |
|         iov[i].iov_base = s->iov_cache[i].iov_base;
 | |
|         iov[i].iov_len = s->iov_cache[i].iov_len;
 | |
|     }
 | |
|     iov_count = s->iov_count;
 | |
|     s->iov_count = 0;
 | |
|     s->total_req_len = 0;
 | |
|     return iov_count;
 | |
| }
 | |
| 
 | |
| static void do_hash_operation(AspeedHACEState *s, int algo, bool sg_mode,
 | |
|                               bool acc_mode)
 | |
| {
 | |
|     struct iovec iov[ASPEED_HACE_MAX_SG];
 | |
|     uint32_t total_msg_len;
 | |
|     uint32_t pad_offset;
 | |
|     g_autofree uint8_t *digest_buf = NULL;
 | |
|     size_t digest_len = 0;
 | |
|     bool sg_acc_mode_final_request = false;
 | |
|     int i;
 | |
|     void *haddr;
 | |
|     Error *local_err = NULL;
 | |
| 
 | |
|     if (acc_mode && s->hash_ctx == NULL) {
 | |
|         s->hash_ctx = qcrypto_hash_new(algo, &local_err);
 | |
|         if (s->hash_ctx == NULL) {
 | |
|             qemu_log_mask(LOG_GUEST_ERROR, "qcrypto hash failed : %s",
 | |
|                           error_get_pretty(local_err));
 | |
|             error_free(local_err);
 | |
|             return;
 | |
|         }
 | |
|     }
 | |
| 
 | |
|     if (sg_mode) {
 | |
|         uint32_t len = 0;
 | |
| 
 | |
|         for (i = 0; !(len & SG_LIST_LEN_LAST); i++) {
 | |
|             uint32_t addr, src;
 | |
|             hwaddr plen;
 | |
| 
 | |
|             if (i == ASPEED_HACE_MAX_SG) {
 | |
|                 qemu_log_mask(LOG_GUEST_ERROR,
 | |
|                         "aspeed_hace: guest failed to set end of sg list marker\n");
 | |
|                 break;
 | |
|             }
 | |
| 
 | |
|             src = s->regs[R_HASH_SRC] + (i * SG_LIST_ENTRY_SIZE);
 | |
| 
 | |
|             len = address_space_ldl_le(&s->dram_as, src,
 | |
|                                        MEMTXATTRS_UNSPECIFIED, NULL);
 | |
| 
 | |
|             addr = address_space_ldl_le(&s->dram_as, src + SG_LIST_LEN_SIZE,
 | |
|                                         MEMTXATTRS_UNSPECIFIED, NULL);
 | |
|             addr &= SG_LIST_ADDR_MASK;
 | |
| 
 | |
|             plen = len & SG_LIST_LEN_MASK;
 | |
|             haddr = address_space_map(&s->dram_as, addr, &plen, false,
 | |
|                                       MEMTXATTRS_UNSPECIFIED);
 | |
|             if (haddr == NULL) {
 | |
|                 qemu_log_mask(LOG_GUEST_ERROR,
 | |
|                               "%s: qcrypto failed\n", __func__);
 | |
|                 return;
 | |
|             }
 | |
|             iov[i].iov_base = haddr;
 | |
|             if (acc_mode) {
 | |
|                 s->total_req_len += plen;
 | |
| 
 | |
|                 if (has_padding(s, &iov[i], plen, &total_msg_len,
 | |
|                                 &pad_offset)) {
 | |
|                     /* Padding being present indicates the final request */
 | |
|                     sg_acc_mode_final_request = true;
 | |
|                     iov[i].iov_len = pad_offset;
 | |
|                 } else {
 | |
|                     iov[i].iov_len = plen;
 | |
|                 }
 | |
|             } else {
 | |
|                 iov[i].iov_len = plen;
 | |
|             }
 | |
|         }
 | |
|     } else {
 | |
|         hwaddr len = s->regs[R_HASH_SRC_LEN];
 | |
| 
 | |
|         haddr = address_space_map(&s->dram_as, s->regs[R_HASH_SRC],
 | |
|                                   &len, false, MEMTXATTRS_UNSPECIFIED);
 | |
|         if (haddr == NULL) {
 | |
|             qemu_log_mask(LOG_GUEST_ERROR, "%s: qcrypto failed\n", __func__);
 | |
|             return;
 | |
|         }
 | |
|         iov[0].iov_base = haddr;
 | |
|         iov[0].iov_len = len;
 | |
|         i = 1;
 | |
| 
 | |
|         if (s->iov_count) {
 | |
|             /*
 | |
|              * In aspeed sdk kernel driver, sg_mode is disabled in hash_final().
 | |
|              * Thus if we received a request with sg_mode disabled, it is
 | |
|              * required to check whether cache is empty. If no, we should
 | |
|              * combine cached iov and the current iov.
 | |
|              */
 | |
|             s->total_req_len += len;
 | |
|             if (has_padding(s, iov, len, &total_msg_len, &pad_offset)) {
 | |
|                 i = reconstruct_iov(s, iov, 0, &pad_offset);
 | |
|             }
 | |
|         }
 | |
|     }
 | |
| 
 | |
|     if (acc_mode) {
 | |
|         if (qcrypto_hash_updatev(s->hash_ctx, iov, i, &local_err) < 0) {
 | |
|             qemu_log_mask(LOG_GUEST_ERROR, "qcrypto hash update failed : %s",
 | |
|                           error_get_pretty(local_err));
 | |
|             error_free(local_err);
 | |
|             return;
 | |
|         }
 | |
| 
 | |
|         if (sg_acc_mode_final_request) {
 | |
|             if (qcrypto_hash_finalize_bytes(s->hash_ctx, &digest_buf,
 | |
|                                             &digest_len, &local_err)) {
 | |
|                 qemu_log_mask(LOG_GUEST_ERROR,
 | |
|                               "qcrypto hash finalize failed : %s",
 | |
|                               error_get_pretty(local_err));
 | |
|                 error_free(local_err);
 | |
|                 local_err = NULL;
 | |
|             }
 | |
| 
 | |
|             qcrypto_hash_free(s->hash_ctx);
 | |
| 
 | |
|             s->hash_ctx = NULL;
 | |
|             s->iov_count = 0;
 | |
|             s->total_req_len = 0;
 | |
|         }
 | |
|     } else if (qcrypto_hash_bytesv(algo, iov, i, &digest_buf,
 | |
|                                    &digest_len, &local_err) < 0) {
 | |
|         qemu_log_mask(LOG_GUEST_ERROR, "qcrypto hash bytesv failed : %s",
 | |
|                       error_get_pretty(local_err));
 | |
|         error_free(local_err);
 | |
|         return;
 | |
|     }
 | |
| 
 | |
|     if (address_space_write(&s->dram_as, s->regs[R_HASH_DEST],
 | |
|                             MEMTXATTRS_UNSPECIFIED,
 | |
|                             digest_buf, digest_len)) {
 | |
|         qemu_log_mask(LOG_GUEST_ERROR,
 | |
|                       "aspeed_hace: address space write failed\n");
 | |
|     }
 | |
| 
 | |
|     for (; i > 0; i--) {
 | |
|         address_space_unmap(&s->dram_as, iov[i - 1].iov_base,
 | |
|                             iov[i - 1].iov_len, false,
 | |
|                             iov[i - 1].iov_len);
 | |
|     }
 | |
| 
 | |
|     /*
 | |
|      * Set status bits to indicate completion. Testing shows hardware sets
 | |
|      * these irrespective of HASH_IRQ_EN.
 | |
|      */
 | |
|     s->regs[R_STATUS] |= HASH_IRQ;
 | |
| }
 | |
| 
 | |
| static uint64_t aspeed_hace_read(void *opaque, hwaddr addr, unsigned int size)
 | |
| {
 | |
|     AspeedHACEState *s = ASPEED_HACE(opaque);
 | |
| 
 | |
|     addr >>= 2;
 | |
| 
 | |
|     if (addr >= ASPEED_HACE_NR_REGS) {
 | |
|         qemu_log_mask(LOG_GUEST_ERROR,
 | |
|                       "%s: Out-of-bounds read at offset 0x%" HWADDR_PRIx "\n",
 | |
|                       __func__, addr << 2);
 | |
|         return 0;
 | |
|     }
 | |
| 
 | |
|     return s->regs[addr];
 | |
| }
 | |
| 
 | |
| static void aspeed_hace_write(void *opaque, hwaddr addr, uint64_t data,
 | |
|                               unsigned int size)
 | |
| {
 | |
|     AspeedHACEState *s = ASPEED_HACE(opaque);
 | |
|     AspeedHACEClass *ahc = ASPEED_HACE_GET_CLASS(s);
 | |
| 
 | |
|     addr >>= 2;
 | |
| 
 | |
|     if (addr >= ASPEED_HACE_NR_REGS) {
 | |
|         qemu_log_mask(LOG_GUEST_ERROR,
 | |
|                       "%s: Out-of-bounds write at offset 0x%" HWADDR_PRIx "\n",
 | |
|                       __func__, addr << 2);
 | |
|         return;
 | |
|     }
 | |
| 
 | |
|     switch (addr) {
 | |
|     case R_STATUS:
 | |
|         if (data & HASH_IRQ) {
 | |
|             data &= ~HASH_IRQ;
 | |
| 
 | |
|             if (s->regs[addr] & HASH_IRQ) {
 | |
|                 qemu_irq_lower(s->irq);
 | |
|             }
 | |
|         }
 | |
|         break;
 | |
|     case R_HASH_SRC:
 | |
|         data &= ahc->src_mask;
 | |
|         break;
 | |
|     case R_HASH_DEST:
 | |
|         data &= ahc->dest_mask;
 | |
|         break;
 | |
|     case R_HASH_KEY_BUFF:
 | |
|         data &= ahc->key_mask;
 | |
|         break;
 | |
|     case R_HASH_SRC_LEN:
 | |
|         data &= 0x0FFFFFFF;
 | |
|         break;
 | |
|     case R_HASH_CMD: {
 | |
|         int algo;
 | |
|         data &= ahc->hash_mask;
 | |
| 
 | |
|         if ((data & HASH_DIGEST_HMAC)) {
 | |
|             qemu_log_mask(LOG_UNIMP,
 | |
|                           "%s: HMAC mode not implemented\n",
 | |
|                           __func__);
 | |
|         }
 | |
|         if (data & BIT(1)) {
 | |
|             qemu_log_mask(LOG_UNIMP,
 | |
|                           "%s: Cascaded mode not implemented\n",
 | |
|                           __func__);
 | |
|         }
 | |
|         algo = hash_algo_lookup(data);
 | |
|         if (algo < 0) {
 | |
|                 qemu_log_mask(LOG_GUEST_ERROR,
 | |
|                         "%s: Invalid hash algorithm selection 0x%"PRIx64"\n",
 | |
|                         __func__, data & ahc->hash_mask);
 | |
|                 break;
 | |
|         }
 | |
|         do_hash_operation(s, algo, data & HASH_SG_EN,
 | |
|                 ((data & HASH_HMAC_MASK) == HASH_DIGEST_ACCUM));
 | |
| 
 | |
|         if (data & HASH_IRQ_EN) {
 | |
|             qemu_irq_raise(s->irq);
 | |
|         }
 | |
|         break;
 | |
|     }
 | |
|     case R_CRYPT_CMD:
 | |
|         qemu_log_mask(LOG_UNIMP, "%s: Crypt commands not implemented\n",
 | |
|                        __func__);
 | |
|         break;
 | |
|     default:
 | |
|         break;
 | |
|     }
 | |
| 
 | |
|     s->regs[addr] = data;
 | |
| }
 | |
| 
 | |
| static const MemoryRegionOps aspeed_hace_ops = {
 | |
|     .read = aspeed_hace_read,
 | |
|     .write = aspeed_hace_write,
 | |
|     .endianness = DEVICE_LITTLE_ENDIAN,
 | |
|     .valid = {
 | |
|         .min_access_size = 1,
 | |
|         .max_access_size = 4,
 | |
|     },
 | |
| };
 | |
| 
 | |
| static void aspeed_hace_reset(DeviceState *dev)
 | |
| {
 | |
|     struct AspeedHACEState *s = ASPEED_HACE(dev);
 | |
| 
 | |
|     if (s->hash_ctx != NULL) {
 | |
|         qcrypto_hash_free(s->hash_ctx);
 | |
|         s->hash_ctx = NULL;
 | |
|     }
 | |
| 
 | |
|     memset(s->regs, 0, sizeof(s->regs));
 | |
|     s->iov_count = 0;
 | |
|     s->total_req_len = 0;
 | |
| }
 | |
| 
 | |
| static void aspeed_hace_realize(DeviceState *dev, Error **errp)
 | |
| {
 | |
|     AspeedHACEState *s = ASPEED_HACE(dev);
 | |
|     SysBusDevice *sbd = SYS_BUS_DEVICE(dev);
 | |
| 
 | |
|     sysbus_init_irq(sbd, &s->irq);
 | |
| 
 | |
|     memory_region_init_io(&s->iomem, OBJECT(s), &aspeed_hace_ops, s,
 | |
|             TYPE_ASPEED_HACE, 0x1000);
 | |
| 
 | |
|     if (!s->dram_mr) {
 | |
|         error_setg(errp, TYPE_ASPEED_HACE ": 'dram' link not set");
 | |
|         return;
 | |
|     }
 | |
| 
 | |
|     address_space_init(&s->dram_as, s->dram_mr, "dram");
 | |
| 
 | |
|     sysbus_init_mmio(sbd, &s->iomem);
 | |
| }
 | |
| 
 | |
| static const Property aspeed_hace_properties[] = {
 | |
|     DEFINE_PROP_LINK("dram", AspeedHACEState, dram_mr,
 | |
|                      TYPE_MEMORY_REGION, MemoryRegion *),
 | |
| };
 | |
| 
 | |
| 
 | |
| static const VMStateDescription vmstate_aspeed_hace = {
 | |
|     .name = TYPE_ASPEED_HACE,
 | |
|     .version_id = 1,
 | |
|     .minimum_version_id = 1,
 | |
|     .fields = (const VMStateField[]) {
 | |
|         VMSTATE_UINT32_ARRAY(regs, AspeedHACEState, ASPEED_HACE_NR_REGS),
 | |
|         VMSTATE_UINT32(total_req_len, AspeedHACEState),
 | |
|         VMSTATE_UINT32(iov_count, AspeedHACEState),
 | |
|         VMSTATE_END_OF_LIST(),
 | |
|     }
 | |
| };
 | |
| 
 | |
| static void aspeed_hace_class_init(ObjectClass *klass, void *data)
 | |
| {
 | |
|     DeviceClass *dc = DEVICE_CLASS(klass);
 | |
| 
 | |
|     dc->realize = aspeed_hace_realize;
 | |
|     device_class_set_legacy_reset(dc, aspeed_hace_reset);
 | |
|     device_class_set_props(dc, aspeed_hace_properties);
 | |
|     dc->vmsd = &vmstate_aspeed_hace;
 | |
| }
 | |
| 
 | |
| static const TypeInfo aspeed_hace_info = {
 | |
|     .name = TYPE_ASPEED_HACE,
 | |
|     .parent = TYPE_SYS_BUS_DEVICE,
 | |
|     .instance_size = sizeof(AspeedHACEState),
 | |
|     .class_init = aspeed_hace_class_init,
 | |
|     .class_size = sizeof(AspeedHACEClass)
 | |
| };
 | |
| 
 | |
| static void aspeed_ast2400_hace_class_init(ObjectClass *klass, void *data)
 | |
| {
 | |
|     DeviceClass *dc = DEVICE_CLASS(klass);
 | |
|     AspeedHACEClass *ahc = ASPEED_HACE_CLASS(klass);
 | |
| 
 | |
|     dc->desc = "AST2400 Hash and Crypto Engine";
 | |
| 
 | |
|     ahc->src_mask = 0x0FFFFFFF;
 | |
|     ahc->dest_mask = 0x0FFFFFF8;
 | |
|     ahc->key_mask = 0x0FFFFFC0;
 | |
|     ahc->hash_mask = 0x000003ff; /* No SG or SHA512 modes */
 | |
| }
 | |
| 
 | |
| static const TypeInfo aspeed_ast2400_hace_info = {
 | |
|     .name = TYPE_ASPEED_AST2400_HACE,
 | |
|     .parent = TYPE_ASPEED_HACE,
 | |
|     .class_init = aspeed_ast2400_hace_class_init,
 | |
| };
 | |
| 
 | |
| static void aspeed_ast2500_hace_class_init(ObjectClass *klass, void *data)
 | |
| {
 | |
|     DeviceClass *dc = DEVICE_CLASS(klass);
 | |
|     AspeedHACEClass *ahc = ASPEED_HACE_CLASS(klass);
 | |
| 
 | |
|     dc->desc = "AST2500 Hash and Crypto Engine";
 | |
| 
 | |
|     ahc->src_mask = 0x3fffffff;
 | |
|     ahc->dest_mask = 0x3ffffff8;
 | |
|     ahc->key_mask = 0x3FFFFFC0;
 | |
|     ahc->hash_mask = 0x000003ff; /* No SG or SHA512 modes */
 | |
| }
 | |
| 
 | |
| static const TypeInfo aspeed_ast2500_hace_info = {
 | |
|     .name = TYPE_ASPEED_AST2500_HACE,
 | |
|     .parent = TYPE_ASPEED_HACE,
 | |
|     .class_init = aspeed_ast2500_hace_class_init,
 | |
| };
 | |
| 
 | |
| static void aspeed_ast2600_hace_class_init(ObjectClass *klass, void *data)
 | |
| {
 | |
|     DeviceClass *dc = DEVICE_CLASS(klass);
 | |
|     AspeedHACEClass *ahc = ASPEED_HACE_CLASS(klass);
 | |
| 
 | |
|     dc->desc = "AST2600 Hash and Crypto Engine";
 | |
| 
 | |
|     ahc->src_mask = 0x7FFFFFFF;
 | |
|     ahc->dest_mask = 0x7FFFFFF8;
 | |
|     ahc->key_mask = 0x7FFFFFF8;
 | |
|     ahc->hash_mask = 0x00147FFF;
 | |
| }
 | |
| 
 | |
| static const TypeInfo aspeed_ast2600_hace_info = {
 | |
|     .name = TYPE_ASPEED_AST2600_HACE,
 | |
|     .parent = TYPE_ASPEED_HACE,
 | |
|     .class_init = aspeed_ast2600_hace_class_init,
 | |
| };
 | |
| 
 | |
| static void aspeed_ast1030_hace_class_init(ObjectClass *klass, void *data)
 | |
| {
 | |
|     DeviceClass *dc = DEVICE_CLASS(klass);
 | |
|     AspeedHACEClass *ahc = ASPEED_HACE_CLASS(klass);
 | |
| 
 | |
|     dc->desc = "AST1030 Hash and Crypto Engine";
 | |
| 
 | |
|     ahc->src_mask = 0x7FFFFFFF;
 | |
|     ahc->dest_mask = 0x7FFFFFF8;
 | |
|     ahc->key_mask = 0x7FFFFFF8;
 | |
|     ahc->hash_mask = 0x00147FFF;
 | |
| }
 | |
| 
 | |
| static const TypeInfo aspeed_ast1030_hace_info = {
 | |
|     .name = TYPE_ASPEED_AST1030_HACE,
 | |
|     .parent = TYPE_ASPEED_HACE,
 | |
|     .class_init = aspeed_ast1030_hace_class_init,
 | |
| };
 | |
| 
 | |
| static void aspeed_ast2700_hace_class_init(ObjectClass *klass, void *data)
 | |
| {
 | |
|     DeviceClass *dc = DEVICE_CLASS(klass);
 | |
|     AspeedHACEClass *ahc = ASPEED_HACE_CLASS(klass);
 | |
| 
 | |
|     dc->desc = "AST2700 Hash and Crypto Engine";
 | |
| 
 | |
|     ahc->src_mask = 0x7FFFFFFF;
 | |
|     ahc->dest_mask = 0x7FFFFFF8;
 | |
|     ahc->key_mask = 0x7FFFFFF8;
 | |
|     ahc->hash_mask = 0x00147FFF;
 | |
| }
 | |
| 
 | |
| static const TypeInfo aspeed_ast2700_hace_info = {
 | |
|     .name = TYPE_ASPEED_AST2700_HACE,
 | |
|     .parent = TYPE_ASPEED_HACE,
 | |
|     .class_init = aspeed_ast2700_hace_class_init,
 | |
| };
 | |
| 
 | |
| static void aspeed_hace_register_types(void)
 | |
| {
 | |
|     type_register_static(&aspeed_ast2400_hace_info);
 | |
|     type_register_static(&aspeed_ast2500_hace_info);
 | |
|     type_register_static(&aspeed_ast2600_hace_info);
 | |
|     type_register_static(&aspeed_ast1030_hace_info);
 | |
|     type_register_static(&aspeed_ast2700_hace_info);
 | |
|     type_register_static(&aspeed_hace_info);
 | |
| }
 | |
| 
 | |
| type_init(aspeed_hace_register_types);
 |