301 lines
6.3 KiB
C
301 lines
6.3 KiB
C
// SPDX-License-Identifier: GPL-2.0
|
|
#include <linux/cpumask.h>
|
|
#include <linux/dma-mapping.h>
|
|
#include <linux/dmapool.h>
|
|
#include <linux/delay.h>
|
|
#include <linux/gfp.h>
|
|
#include <linux/kernel.h>
|
|
#include <linux/module.h>
|
|
#include <linux/pci_regs.h>
|
|
#include <linux/vmalloc.h>
|
|
#include <linux/pci.h>
|
|
|
|
#include "nitrox_dev.h"
|
|
#include "nitrox_common.h"
|
|
#include "nitrox_req.h"
|
|
#include "nitrox_csr.h"
|
|
|
|
#define CRYPTO_CTX_SIZE 256
|
|
|
|
/* packet inuput ring alignments */
|
|
#define PKTIN_Q_ALIGN_BYTES 16
|
|
/* AQM Queue input alignments */
|
|
#define AQM_Q_ALIGN_BYTES 32
|
|
|
|
static int nitrox_cmdq_init(struct nitrox_cmdq *cmdq, int align_bytes)
|
|
{
|
|
struct nitrox_device *ndev = cmdq->ndev;
|
|
|
|
cmdq->qsize = (ndev->qlen * cmdq->instr_size) + align_bytes;
|
|
cmdq->unalign_base = dma_alloc_coherent(DEV(ndev), cmdq->qsize,
|
|
&cmdq->unalign_dma,
|
|
GFP_KERNEL);
|
|
if (!cmdq->unalign_base)
|
|
return -ENOMEM;
|
|
|
|
cmdq->dma = PTR_ALIGN(cmdq->unalign_dma, align_bytes);
|
|
cmdq->base = cmdq->unalign_base + (cmdq->dma - cmdq->unalign_dma);
|
|
cmdq->write_idx = 0;
|
|
|
|
spin_lock_init(&cmdq->cmd_qlock);
|
|
spin_lock_init(&cmdq->resp_qlock);
|
|
spin_lock_init(&cmdq->backlog_qlock);
|
|
|
|
INIT_LIST_HEAD(&cmdq->response_head);
|
|
INIT_LIST_HEAD(&cmdq->backlog_head);
|
|
INIT_WORK(&cmdq->backlog_qflush, backlog_qflush_work);
|
|
|
|
atomic_set(&cmdq->pending_count, 0);
|
|
atomic_set(&cmdq->backlog_count, 0);
|
|
return 0;
|
|
}
|
|
|
|
static void nitrox_cmdq_reset(struct nitrox_cmdq *cmdq)
|
|
{
|
|
cmdq->write_idx = 0;
|
|
atomic_set(&cmdq->pending_count, 0);
|
|
atomic_set(&cmdq->backlog_count, 0);
|
|
}
|
|
|
|
static void nitrox_cmdq_cleanup(struct nitrox_cmdq *cmdq)
|
|
{
|
|
struct nitrox_device *ndev;
|
|
|
|
if (!cmdq)
|
|
return;
|
|
|
|
if (!cmdq->unalign_base)
|
|
return;
|
|
|
|
ndev = cmdq->ndev;
|
|
cancel_work_sync(&cmdq->backlog_qflush);
|
|
|
|
dma_free_coherent(DEV(ndev), cmdq->qsize,
|
|
cmdq->unalign_base, cmdq->unalign_dma);
|
|
nitrox_cmdq_reset(cmdq);
|
|
|
|
cmdq->dbell_csr_addr = NULL;
|
|
cmdq->compl_cnt_csr_addr = NULL;
|
|
cmdq->unalign_base = NULL;
|
|
cmdq->base = NULL;
|
|
cmdq->unalign_dma = 0;
|
|
cmdq->dma = 0;
|
|
cmdq->qsize = 0;
|
|
cmdq->instr_size = 0;
|
|
}
|
|
|
|
static void nitrox_free_aqm_queues(struct nitrox_device *ndev)
|
|
{
|
|
int i;
|
|
|
|
for (i = 0; i < ndev->nr_queues; i++) {
|
|
nitrox_cmdq_cleanup(ndev->aqmq[i]);
|
|
kfree_sensitive(ndev->aqmq[i]);
|
|
ndev->aqmq[i] = NULL;
|
|
}
|
|
}
|
|
|
|
static int nitrox_alloc_aqm_queues(struct nitrox_device *ndev)
|
|
{
|
|
int i, err;
|
|
|
|
for (i = 0; i < ndev->nr_queues; i++) {
|
|
struct nitrox_cmdq *cmdq;
|
|
u64 offset;
|
|
|
|
cmdq = kzalloc_node(sizeof(*cmdq), GFP_KERNEL, ndev->node);
|
|
if (!cmdq) {
|
|
err = -ENOMEM;
|
|
goto aqmq_fail;
|
|
}
|
|
|
|
cmdq->ndev = ndev;
|
|
cmdq->qno = i;
|
|
cmdq->instr_size = sizeof(struct aqmq_command_s);
|
|
|
|
/* AQM Queue Doorbell Counter Register Address */
|
|
offset = AQMQ_DRBLX(i);
|
|
cmdq->dbell_csr_addr = NITROX_CSR_ADDR(ndev, offset);
|
|
/* AQM Queue Commands Completed Count Register Address */
|
|
offset = AQMQ_CMD_CNTX(i);
|
|
cmdq->compl_cnt_csr_addr = NITROX_CSR_ADDR(ndev, offset);
|
|
|
|
err = nitrox_cmdq_init(cmdq, AQM_Q_ALIGN_BYTES);
|
|
if (err) {
|
|
kfree_sensitive(cmdq);
|
|
goto aqmq_fail;
|
|
}
|
|
ndev->aqmq[i] = cmdq;
|
|
}
|
|
|
|
return 0;
|
|
|
|
aqmq_fail:
|
|
nitrox_free_aqm_queues(ndev);
|
|
return err;
|
|
}
|
|
|
|
static void nitrox_free_pktin_queues(struct nitrox_device *ndev)
|
|
{
|
|
int i;
|
|
|
|
for (i = 0; i < ndev->nr_queues; i++) {
|
|
struct nitrox_cmdq *cmdq = &ndev->pkt_inq[i];
|
|
|
|
nitrox_cmdq_cleanup(cmdq);
|
|
}
|
|
kfree(ndev->pkt_inq);
|
|
ndev->pkt_inq = NULL;
|
|
}
|
|
|
|
static int nitrox_alloc_pktin_queues(struct nitrox_device *ndev)
|
|
{
|
|
int i, err;
|
|
|
|
ndev->pkt_inq = kcalloc_node(ndev->nr_queues,
|
|
sizeof(struct nitrox_cmdq),
|
|
GFP_KERNEL, ndev->node);
|
|
if (!ndev->pkt_inq)
|
|
return -ENOMEM;
|
|
|
|
for (i = 0; i < ndev->nr_queues; i++) {
|
|
struct nitrox_cmdq *cmdq;
|
|
u64 offset;
|
|
|
|
cmdq = &ndev->pkt_inq[i];
|
|
cmdq->ndev = ndev;
|
|
cmdq->qno = i;
|
|
cmdq->instr_size = sizeof(struct nps_pkt_instr);
|
|
|
|
/* packet input ring doorbell address */
|
|
offset = NPS_PKT_IN_INSTR_BAOFF_DBELLX(i);
|
|
cmdq->dbell_csr_addr = NITROX_CSR_ADDR(ndev, offset);
|
|
/* packet solicit port completion count address */
|
|
offset = NPS_PKT_SLC_CNTSX(i);
|
|
cmdq->compl_cnt_csr_addr = NITROX_CSR_ADDR(ndev, offset);
|
|
|
|
err = nitrox_cmdq_init(cmdq, PKTIN_Q_ALIGN_BYTES);
|
|
if (err)
|
|
goto pktq_fail;
|
|
}
|
|
return 0;
|
|
|
|
pktq_fail:
|
|
nitrox_free_pktin_queues(ndev);
|
|
return err;
|
|
}
|
|
|
|
static int create_crypto_dma_pool(struct nitrox_device *ndev)
|
|
{
|
|
size_t size;
|
|
|
|
/* Crypto context pool, 16 byte aligned */
|
|
size = CRYPTO_CTX_SIZE + sizeof(struct ctx_hdr);
|
|
ndev->ctx_pool = dma_pool_create("nitrox-context",
|
|
DEV(ndev), size, 16, 0);
|
|
if (!ndev->ctx_pool)
|
|
return -ENOMEM;
|
|
|
|
return 0;
|
|
}
|
|
|
|
static void destroy_crypto_dma_pool(struct nitrox_device *ndev)
|
|
{
|
|
if (!ndev->ctx_pool)
|
|
return;
|
|
|
|
dma_pool_destroy(ndev->ctx_pool);
|
|
ndev->ctx_pool = NULL;
|
|
}
|
|
|
|
/*
|
|
* crypto_alloc_context - Allocate crypto context from pool
|
|
* @ndev: NITROX Device
|
|
*/
|
|
void *crypto_alloc_context(struct nitrox_device *ndev)
|
|
{
|
|
struct ctx_hdr *ctx;
|
|
struct crypto_ctx_hdr *chdr;
|
|
void *vaddr;
|
|
dma_addr_t dma;
|
|
|
|
chdr = kmalloc(sizeof(*chdr), GFP_KERNEL);
|
|
if (!chdr)
|
|
return NULL;
|
|
|
|
vaddr = dma_pool_zalloc(ndev->ctx_pool, GFP_KERNEL, &dma);
|
|
if (!vaddr) {
|
|
kfree(chdr);
|
|
return NULL;
|
|
}
|
|
|
|
/* fill meta data */
|
|
ctx = vaddr;
|
|
ctx->pool = ndev->ctx_pool;
|
|
ctx->dma = dma;
|
|
ctx->ctx_dma = dma + sizeof(struct ctx_hdr);
|
|
|
|
chdr->pool = ndev->ctx_pool;
|
|
chdr->dma = dma;
|
|
chdr->vaddr = vaddr;
|
|
|
|
return chdr;
|
|
}
|
|
|
|
/**
|
|
* crypto_free_context - Free crypto context to pool
|
|
* @ctx: context to free
|
|
*/
|
|
void crypto_free_context(void *ctx)
|
|
{
|
|
struct crypto_ctx_hdr *ctxp;
|
|
|
|
if (!ctx)
|
|
return;
|
|
|
|
ctxp = ctx;
|
|
dma_pool_free(ctxp->pool, ctxp->vaddr, ctxp->dma);
|
|
kfree(ctxp);
|
|
}
|
|
|
|
/**
|
|
* nitrox_common_sw_init - allocate software resources.
|
|
* @ndev: NITROX device
|
|
*
|
|
* Allocates crypto context pools and command queues etc.
|
|
*
|
|
* Return: 0 on success, or a negative error code on error.
|
|
*/
|
|
int nitrox_common_sw_init(struct nitrox_device *ndev)
|
|
{
|
|
int err = 0;
|
|
|
|
/* per device crypto context pool */
|
|
err = create_crypto_dma_pool(ndev);
|
|
if (err)
|
|
return err;
|
|
|
|
err = nitrox_alloc_pktin_queues(ndev);
|
|
if (err)
|
|
destroy_crypto_dma_pool(ndev);
|
|
|
|
err = nitrox_alloc_aqm_queues(ndev);
|
|
if (err) {
|
|
nitrox_free_pktin_queues(ndev);
|
|
destroy_crypto_dma_pool(ndev);
|
|
}
|
|
|
|
return err;
|
|
}
|
|
|
|
/**
|
|
* nitrox_common_sw_cleanup - free software resources.
|
|
* @ndev: NITROX device
|
|
*/
|
|
void nitrox_common_sw_cleanup(struct nitrox_device *ndev)
|
|
{
|
|
nitrox_free_aqm_queues(ndev);
|
|
nitrox_free_pktin_queues(ndev);
|
|
destroy_crypto_dma_pool(ndev);
|
|
}
|