609 lines
16 KiB
C
609 lines
16 KiB
C
// SPDX-License-Identifier: GPL-2.0-only
|
|
/* Copyright (C) 2020 Marvell. */
|
|
|
|
#include "otx2_cpt_common.h"
|
|
#include "otx2_cptpf.h"
|
|
#include "rvu_reg.h"
|
|
|
|
/* Fastpath ipsec opcode with inplace processing */
|
|
#define CPT_INLINE_RX_OPCODE (0x26 | (1 << 6))
|
|
#define CN10K_CPT_INLINE_RX_OPCODE (0x29 | (1 << 6))
|
|
|
|
#define cpt_inline_rx_opcode(pdev) \
|
|
({ \
|
|
u8 opcode; \
|
|
if (is_dev_otx2(pdev)) \
|
|
opcode = CPT_INLINE_RX_OPCODE; \
|
|
else \
|
|
opcode = CN10K_CPT_INLINE_RX_OPCODE; \
|
|
(opcode); \
|
|
})
|
|
|
|
/*
|
|
* CPT PF driver version, It will be incremented by 1 for every feature
|
|
* addition in CPT mailbox messages.
|
|
*/
|
|
#define OTX2_CPT_PF_DRV_VERSION 0x1
|
|
|
|
static int forward_to_af(struct otx2_cptpf_dev *cptpf,
|
|
struct otx2_cptvf_info *vf,
|
|
struct mbox_msghdr *req, int size)
|
|
{
|
|
struct mbox_msghdr *msg;
|
|
int ret;
|
|
|
|
mutex_lock(&cptpf->lock);
|
|
msg = otx2_mbox_alloc_msg(&cptpf->afpf_mbox, 0, size);
|
|
if (msg == NULL) {
|
|
mutex_unlock(&cptpf->lock);
|
|
return -ENOMEM;
|
|
}
|
|
|
|
memcpy((uint8_t *)msg + sizeof(struct mbox_msghdr),
|
|
(uint8_t *)req + sizeof(struct mbox_msghdr), size);
|
|
msg->id = req->id;
|
|
msg->pcifunc = req->pcifunc;
|
|
msg->sig = req->sig;
|
|
msg->ver = req->ver;
|
|
|
|
ret = otx2_cpt_sync_mbox_msg(&cptpf->afpf_mbox);
|
|
/* Error code -EIO indicate there is a communication failure
|
|
* to the AF. Rest of the error codes indicate that AF processed
|
|
* VF messages and set the error codes in response messages
|
|
* (if any) so simply forward responses to VF.
|
|
*/
|
|
if (ret == -EIO) {
|
|
dev_warn(&cptpf->pdev->dev,
|
|
"AF not responding to VF%d messages\n", vf->vf_id);
|
|
mutex_unlock(&cptpf->lock);
|
|
return ret;
|
|
}
|
|
mutex_unlock(&cptpf->lock);
|
|
return 0;
|
|
}
|
|
|
|
static int handle_msg_get_caps(struct otx2_cptpf_dev *cptpf,
|
|
struct otx2_cptvf_info *vf,
|
|
struct mbox_msghdr *req)
|
|
{
|
|
struct otx2_cpt_caps_rsp *rsp;
|
|
|
|
rsp = (struct otx2_cpt_caps_rsp *)
|
|
otx2_mbox_alloc_msg(&cptpf->vfpf_mbox, vf->vf_id,
|
|
sizeof(*rsp));
|
|
if (!rsp)
|
|
return -ENOMEM;
|
|
|
|
rsp->hdr.id = MBOX_MSG_GET_CAPS;
|
|
rsp->hdr.sig = OTX2_MBOX_RSP_SIG;
|
|
rsp->hdr.pcifunc = req->pcifunc;
|
|
rsp->cpt_pf_drv_version = OTX2_CPT_PF_DRV_VERSION;
|
|
rsp->cpt_revision = cptpf->pdev->revision;
|
|
memcpy(&rsp->eng_caps, &cptpf->eng_caps, sizeof(rsp->eng_caps));
|
|
|
|
return 0;
|
|
}
|
|
|
|
static int handle_msg_get_eng_grp_num(struct otx2_cptpf_dev *cptpf,
|
|
struct otx2_cptvf_info *vf,
|
|
struct mbox_msghdr *req)
|
|
{
|
|
struct otx2_cpt_egrp_num_msg *grp_req;
|
|
struct otx2_cpt_egrp_num_rsp *rsp;
|
|
|
|
grp_req = (struct otx2_cpt_egrp_num_msg *)req;
|
|
rsp = (struct otx2_cpt_egrp_num_rsp *)
|
|
otx2_mbox_alloc_msg(&cptpf->vfpf_mbox, vf->vf_id, sizeof(*rsp));
|
|
if (!rsp)
|
|
return -ENOMEM;
|
|
|
|
rsp->hdr.id = MBOX_MSG_GET_ENG_GRP_NUM;
|
|
rsp->hdr.sig = OTX2_MBOX_RSP_SIG;
|
|
rsp->hdr.pcifunc = req->pcifunc;
|
|
rsp->eng_type = grp_req->eng_type;
|
|
rsp->eng_grp_num = otx2_cpt_get_eng_grp(&cptpf->eng_grps,
|
|
grp_req->eng_type);
|
|
|
|
return 0;
|
|
}
|
|
|
|
static int handle_msg_kvf_limits(struct otx2_cptpf_dev *cptpf,
|
|
struct otx2_cptvf_info *vf,
|
|
struct mbox_msghdr *req)
|
|
{
|
|
struct otx2_cpt_kvf_limits_rsp *rsp;
|
|
|
|
rsp = (struct otx2_cpt_kvf_limits_rsp *)
|
|
otx2_mbox_alloc_msg(&cptpf->vfpf_mbox, vf->vf_id, sizeof(*rsp));
|
|
if (!rsp)
|
|
return -ENOMEM;
|
|
|
|
rsp->hdr.id = MBOX_MSG_GET_KVF_LIMITS;
|
|
rsp->hdr.sig = OTX2_MBOX_RSP_SIG;
|
|
rsp->hdr.pcifunc = req->pcifunc;
|
|
rsp->kvf_limits = cptpf->kvf_limits;
|
|
|
|
return 0;
|
|
}
|
|
|
|
static int send_inline_ipsec_inbound_msg(struct otx2_cptpf_dev *cptpf,
|
|
int sso_pf_func, u8 slot)
|
|
{
|
|
struct cpt_inline_ipsec_cfg_msg *req;
|
|
struct pci_dev *pdev = cptpf->pdev;
|
|
|
|
req = (struct cpt_inline_ipsec_cfg_msg *)
|
|
otx2_mbox_alloc_msg_rsp(&cptpf->afpf_mbox, 0,
|
|
sizeof(*req), sizeof(struct msg_rsp));
|
|
if (req == NULL) {
|
|
dev_err(&pdev->dev, "RVU MBOX failed to get message.\n");
|
|
return -EFAULT;
|
|
}
|
|
memset(req, 0, sizeof(*req));
|
|
req->hdr.id = MBOX_MSG_CPT_INLINE_IPSEC_CFG;
|
|
req->hdr.sig = OTX2_MBOX_REQ_SIG;
|
|
req->hdr.pcifunc = OTX2_CPT_RVU_PFFUNC(cptpf->pf_id, 0);
|
|
req->dir = CPT_INLINE_INBOUND;
|
|
req->slot = slot;
|
|
req->sso_pf_func_ovrd = cptpf->sso_pf_func_ovrd;
|
|
req->sso_pf_func = sso_pf_func;
|
|
req->enable = 1;
|
|
|
|
return otx2_cpt_send_mbox_msg(&cptpf->afpf_mbox, pdev);
|
|
}
|
|
|
|
static int rx_inline_ipsec_lf_cfg(struct otx2_cptpf_dev *cptpf, u8 egrp,
|
|
struct otx2_cpt_rx_inline_lf_cfg *req)
|
|
{
|
|
struct nix_inline_ipsec_cfg *nix_req;
|
|
struct pci_dev *pdev = cptpf->pdev;
|
|
int ret;
|
|
|
|
nix_req = (struct nix_inline_ipsec_cfg *)
|
|
otx2_mbox_alloc_msg_rsp(&cptpf->afpf_mbox, 0,
|
|
sizeof(*nix_req),
|
|
sizeof(struct msg_rsp));
|
|
if (nix_req == NULL) {
|
|
dev_err(&pdev->dev, "RVU MBOX failed to get message.\n");
|
|
return -EFAULT;
|
|
}
|
|
memset(nix_req, 0, sizeof(*nix_req));
|
|
nix_req->hdr.id = MBOX_MSG_NIX_INLINE_IPSEC_CFG;
|
|
nix_req->hdr.sig = OTX2_MBOX_REQ_SIG;
|
|
nix_req->enable = 1;
|
|
if (!req->credit || req->credit > OTX2_CPT_INST_QLEN_MSGS)
|
|
nix_req->cpt_credit = OTX2_CPT_INST_QLEN_MSGS - 1;
|
|
else
|
|
nix_req->cpt_credit = req->credit - 1;
|
|
nix_req->gen_cfg.egrp = egrp;
|
|
if (req->opcode)
|
|
nix_req->gen_cfg.opcode = req->opcode;
|
|
else
|
|
nix_req->gen_cfg.opcode = cpt_inline_rx_opcode(pdev);
|
|
nix_req->gen_cfg.param1 = req->param1;
|
|
nix_req->gen_cfg.param2 = req->param2;
|
|
nix_req->inst_qsel.cpt_pf_func = OTX2_CPT_RVU_PFFUNC(cptpf->pf_id, 0);
|
|
nix_req->inst_qsel.cpt_slot = 0;
|
|
ret = otx2_cpt_send_mbox_msg(&cptpf->afpf_mbox, pdev);
|
|
if (ret)
|
|
return ret;
|
|
|
|
if (cptpf->has_cpt1) {
|
|
ret = send_inline_ipsec_inbound_msg(cptpf, req->sso_pf_func, 1);
|
|
if (ret)
|
|
return ret;
|
|
}
|
|
|
|
return send_inline_ipsec_inbound_msg(cptpf, req->sso_pf_func, 0);
|
|
}
|
|
|
|
static int handle_msg_rx_inline_ipsec_lf_cfg(struct otx2_cptpf_dev *cptpf,
|
|
struct mbox_msghdr *req)
|
|
{
|
|
struct otx2_cpt_rx_inline_lf_cfg *cfg_req;
|
|
u8 egrp;
|
|
int ret;
|
|
|
|
cfg_req = (struct otx2_cpt_rx_inline_lf_cfg *)req;
|
|
if (cptpf->lfs.lfs_num) {
|
|
dev_err(&cptpf->pdev->dev,
|
|
"LF is already configured for RX inline ipsec.\n");
|
|
return -EEXIST;
|
|
}
|
|
/*
|
|
* Allow LFs to execute requests destined to only grp IE_TYPES and
|
|
* set queue priority of each LF to high
|
|
*/
|
|
egrp = otx2_cpt_get_eng_grp(&cptpf->eng_grps, OTX2_CPT_IE_TYPES);
|
|
if (egrp == OTX2_CPT_INVALID_CRYPTO_ENG_GRP) {
|
|
dev_err(&cptpf->pdev->dev,
|
|
"Engine group for inline ipsec is not available\n");
|
|
return -ENOENT;
|
|
}
|
|
|
|
otx2_cptlf_set_dev_info(&cptpf->lfs, cptpf->pdev, cptpf->reg_base,
|
|
&cptpf->afpf_mbox, BLKADDR_CPT0);
|
|
ret = otx2_cptlf_init(&cptpf->lfs, 1 << egrp, OTX2_CPT_QUEUE_HI_PRIO,
|
|
1);
|
|
if (ret) {
|
|
dev_err(&cptpf->pdev->dev,
|
|
"LF configuration failed for RX inline ipsec.\n");
|
|
return ret;
|
|
}
|
|
|
|
if (cptpf->has_cpt1) {
|
|
cptpf->rsrc_req_blkaddr = BLKADDR_CPT1;
|
|
otx2_cptlf_set_dev_info(&cptpf->cpt1_lfs, cptpf->pdev,
|
|
cptpf->reg_base, &cptpf->afpf_mbox,
|
|
BLKADDR_CPT1);
|
|
ret = otx2_cptlf_init(&cptpf->cpt1_lfs, 1 << egrp,
|
|
OTX2_CPT_QUEUE_HI_PRIO, 1);
|
|
if (ret) {
|
|
dev_err(&cptpf->pdev->dev,
|
|
"LF configuration failed for RX inline ipsec.\n");
|
|
goto lf_cleanup;
|
|
}
|
|
cptpf->rsrc_req_blkaddr = 0;
|
|
}
|
|
|
|
ret = rx_inline_ipsec_lf_cfg(cptpf, egrp, cfg_req);
|
|
if (ret)
|
|
goto lf1_cleanup;
|
|
|
|
return 0;
|
|
|
|
lf1_cleanup:
|
|
otx2_cptlf_shutdown(&cptpf->cpt1_lfs);
|
|
lf_cleanup:
|
|
otx2_cptlf_shutdown(&cptpf->lfs);
|
|
return ret;
|
|
}
|
|
|
|
static int cptpf_handle_vf_req(struct otx2_cptpf_dev *cptpf,
|
|
struct otx2_cptvf_info *vf,
|
|
struct mbox_msghdr *req, int size)
|
|
{
|
|
int err = 0;
|
|
|
|
/* Check if msg is valid, if not reply with an invalid msg */
|
|
if (req->sig != OTX2_MBOX_REQ_SIG)
|
|
goto inval_msg;
|
|
|
|
switch (req->id) {
|
|
case MBOX_MSG_GET_ENG_GRP_NUM:
|
|
err = handle_msg_get_eng_grp_num(cptpf, vf, req);
|
|
break;
|
|
case MBOX_MSG_GET_CAPS:
|
|
err = handle_msg_get_caps(cptpf, vf, req);
|
|
break;
|
|
case MBOX_MSG_GET_KVF_LIMITS:
|
|
err = handle_msg_kvf_limits(cptpf, vf, req);
|
|
break;
|
|
case MBOX_MSG_RX_INLINE_IPSEC_LF_CFG:
|
|
err = handle_msg_rx_inline_ipsec_lf_cfg(cptpf, req);
|
|
break;
|
|
|
|
default:
|
|
err = forward_to_af(cptpf, vf, req, size);
|
|
break;
|
|
}
|
|
return err;
|
|
|
|
inval_msg:
|
|
otx2_reply_invalid_msg(&cptpf->vfpf_mbox, vf->vf_id, 0, req->id);
|
|
otx2_mbox_msg_send(&cptpf->vfpf_mbox, vf->vf_id);
|
|
return err;
|
|
}
|
|
|
|
irqreturn_t otx2_cptpf_vfpf_mbox_intr(int __always_unused irq, void *arg)
|
|
{
|
|
struct otx2_cptpf_dev *cptpf = arg;
|
|
struct otx2_cptvf_info *vf;
|
|
int i, vf_idx;
|
|
u64 intr;
|
|
|
|
/*
|
|
* Check which VF has raised an interrupt and schedule
|
|
* corresponding work queue to process the messages
|
|
*/
|
|
for (i = 0; i < 2; i++) {
|
|
/* Read the interrupt bits */
|
|
intr = otx2_cpt_read64(cptpf->reg_base, BLKADDR_RVUM, 0,
|
|
RVU_PF_VFPF_MBOX_INTX(i));
|
|
|
|
for (vf_idx = i * 64; vf_idx < cptpf->enabled_vfs; vf_idx++) {
|
|
vf = &cptpf->vf[vf_idx];
|
|
if (intr & (1ULL << vf->intr_idx)) {
|
|
queue_work(cptpf->vfpf_mbox_wq,
|
|
&vf->vfpf_mbox_work);
|
|
/* Clear the interrupt */
|
|
otx2_cpt_write64(cptpf->reg_base, BLKADDR_RVUM,
|
|
0, RVU_PF_VFPF_MBOX_INTX(i),
|
|
BIT_ULL(vf->intr_idx));
|
|
}
|
|
}
|
|
}
|
|
return IRQ_HANDLED;
|
|
}
|
|
|
|
void otx2_cptpf_vfpf_mbox_handler(struct work_struct *work)
|
|
{
|
|
struct otx2_cptpf_dev *cptpf;
|
|
struct otx2_cptvf_info *vf;
|
|
struct otx2_mbox_dev *mdev;
|
|
struct mbox_hdr *req_hdr;
|
|
struct mbox_msghdr *msg;
|
|
struct otx2_mbox *mbox;
|
|
int offset, i, err;
|
|
|
|
vf = container_of(work, struct otx2_cptvf_info, vfpf_mbox_work);
|
|
cptpf = vf->cptpf;
|
|
mbox = &cptpf->vfpf_mbox;
|
|
/* sync with mbox memory region */
|
|
smp_rmb();
|
|
mdev = &mbox->dev[vf->vf_id];
|
|
/* Process received mbox messages */
|
|
req_hdr = (struct mbox_hdr *)(mdev->mbase + mbox->rx_start);
|
|
offset = mbox->rx_start + ALIGN(sizeof(*req_hdr), MBOX_MSG_ALIGN);
|
|
|
|
for (i = 0; i < req_hdr->num_msgs; i++) {
|
|
msg = (struct mbox_msghdr *)(mdev->mbase + offset);
|
|
|
|
/* Set which VF sent this message based on mbox IRQ */
|
|
msg->pcifunc = ((u16)cptpf->pf_id << RVU_PFVF_PF_SHIFT) |
|
|
((vf->vf_id + 1) & RVU_PFVF_FUNC_MASK);
|
|
|
|
err = cptpf_handle_vf_req(cptpf, vf, msg,
|
|
msg->next_msgoff - offset);
|
|
/*
|
|
* Behave as the AF, drop the msg if there is
|
|
* no memory, timeout handling also goes here
|
|
*/
|
|
if (err == -ENOMEM || err == -EIO)
|
|
break;
|
|
offset = msg->next_msgoff;
|
|
/* Write barrier required for VF responses which are handled by
|
|
* PF driver and not forwarded to AF.
|
|
*/
|
|
smp_wmb();
|
|
}
|
|
/* Send mbox responses to VF */
|
|
if (mdev->num_msgs)
|
|
otx2_mbox_msg_send(mbox, vf->vf_id);
|
|
}
|
|
|
|
irqreturn_t otx2_cptpf_afpf_mbox_intr(int __always_unused irq, void *arg)
|
|
{
|
|
struct otx2_cptpf_dev *cptpf = arg;
|
|
struct otx2_mbox_dev *mdev;
|
|
struct otx2_mbox *mbox;
|
|
struct mbox_hdr *hdr;
|
|
u64 intr;
|
|
|
|
/* Read the interrupt bits */
|
|
intr = otx2_cpt_read64(cptpf->reg_base, BLKADDR_RVUM, 0, RVU_PF_INT);
|
|
|
|
if (intr & 0x1ULL) {
|
|
mbox = &cptpf->afpf_mbox;
|
|
mdev = &mbox->dev[0];
|
|
hdr = mdev->mbase + mbox->rx_start;
|
|
if (hdr->num_msgs)
|
|
/* Schedule work queue function to process the MBOX request */
|
|
queue_work(cptpf->afpf_mbox_wq, &cptpf->afpf_mbox_work);
|
|
|
|
mbox = &cptpf->afpf_mbox_up;
|
|
mdev = &mbox->dev[0];
|
|
hdr = mdev->mbase + mbox->rx_start;
|
|
if (hdr->num_msgs)
|
|
/* Schedule work queue function to process the MBOX request */
|
|
queue_work(cptpf->afpf_mbox_wq, &cptpf->afpf_mbox_up_work);
|
|
/* Clear and ack the interrupt */
|
|
otx2_cpt_write64(cptpf->reg_base, BLKADDR_RVUM, 0, RVU_PF_INT,
|
|
0x1ULL);
|
|
}
|
|
return IRQ_HANDLED;
|
|
}
|
|
|
|
static void process_afpf_mbox_msg(struct otx2_cptpf_dev *cptpf,
|
|
struct mbox_msghdr *msg)
|
|
{
|
|
struct otx2_cptlfs_info *lfs = &cptpf->lfs;
|
|
struct device *dev = &cptpf->pdev->dev;
|
|
struct cpt_rd_wr_reg_msg *rsp_rd_wr;
|
|
|
|
if (msg->id >= MBOX_MSG_MAX) {
|
|
dev_err(dev, "MBOX msg with unknown ID %d\n", msg->id);
|
|
return;
|
|
}
|
|
if (msg->sig != OTX2_MBOX_RSP_SIG) {
|
|
dev_err(dev, "MBOX msg with wrong signature %x, ID %d\n",
|
|
msg->sig, msg->id);
|
|
return;
|
|
}
|
|
if (cptpf->rsrc_req_blkaddr == BLKADDR_CPT1)
|
|
lfs = &cptpf->cpt1_lfs;
|
|
|
|
switch (msg->id) {
|
|
case MBOX_MSG_READY:
|
|
cptpf->pf_id = (msg->pcifunc >> RVU_PFVF_PF_SHIFT) &
|
|
RVU_PFVF_PF_MASK;
|
|
break;
|
|
case MBOX_MSG_CPT_RD_WR_REGISTER:
|
|
rsp_rd_wr = (struct cpt_rd_wr_reg_msg *)msg;
|
|
if (msg->rc) {
|
|
dev_err(dev, "Reg %llx rd/wr(%d) failed %d\n",
|
|
rsp_rd_wr->reg_offset, rsp_rd_wr->is_write,
|
|
msg->rc);
|
|
return;
|
|
}
|
|
if (!rsp_rd_wr->is_write)
|
|
*rsp_rd_wr->ret_val = rsp_rd_wr->val;
|
|
break;
|
|
case MBOX_MSG_ATTACH_RESOURCES:
|
|
if (!msg->rc)
|
|
lfs->are_lfs_attached = 1;
|
|
break;
|
|
case MBOX_MSG_DETACH_RESOURCES:
|
|
if (!msg->rc)
|
|
lfs->are_lfs_attached = 0;
|
|
break;
|
|
case MBOX_MSG_CPT_INLINE_IPSEC_CFG:
|
|
case MBOX_MSG_NIX_INLINE_IPSEC_CFG:
|
|
break;
|
|
|
|
default:
|
|
dev_err(dev,
|
|
"Unsupported msg %d received.\n", msg->id);
|
|
break;
|
|
}
|
|
}
|
|
|
|
static void forward_to_vf(struct otx2_cptpf_dev *cptpf, struct mbox_msghdr *msg,
|
|
int vf_id, int size)
|
|
{
|
|
struct otx2_mbox *vfpf_mbox;
|
|
struct mbox_msghdr *fwd;
|
|
|
|
if (msg->id >= MBOX_MSG_MAX) {
|
|
dev_err(&cptpf->pdev->dev,
|
|
"MBOX msg with unknown ID %d\n", msg->id);
|
|
return;
|
|
}
|
|
if (msg->sig != OTX2_MBOX_RSP_SIG) {
|
|
dev_err(&cptpf->pdev->dev,
|
|
"MBOX msg with wrong signature %x, ID %d\n",
|
|
msg->sig, msg->id);
|
|
return;
|
|
}
|
|
vfpf_mbox = &cptpf->vfpf_mbox;
|
|
vf_id--;
|
|
if (vf_id >= cptpf->enabled_vfs) {
|
|
dev_err(&cptpf->pdev->dev,
|
|
"MBOX msg to unknown VF: %d >= %d\n",
|
|
vf_id, cptpf->enabled_vfs);
|
|
return;
|
|
}
|
|
if (msg->id == MBOX_MSG_VF_FLR)
|
|
return;
|
|
|
|
fwd = otx2_mbox_alloc_msg(vfpf_mbox, vf_id, size);
|
|
if (!fwd) {
|
|
dev_err(&cptpf->pdev->dev,
|
|
"Forwarding to VF%d failed.\n", vf_id);
|
|
return;
|
|
}
|
|
memcpy((uint8_t *)fwd + sizeof(struct mbox_msghdr),
|
|
(uint8_t *)msg + sizeof(struct mbox_msghdr), size);
|
|
fwd->id = msg->id;
|
|
fwd->pcifunc = msg->pcifunc;
|
|
fwd->sig = msg->sig;
|
|
fwd->ver = msg->ver;
|
|
fwd->rc = msg->rc;
|
|
}
|
|
|
|
/* Handle mailbox messages received from AF */
|
|
void otx2_cptpf_afpf_mbox_handler(struct work_struct *work)
|
|
{
|
|
struct otx2_cptpf_dev *cptpf;
|
|
struct otx2_mbox *afpf_mbox;
|
|
struct otx2_mbox_dev *mdev;
|
|
struct mbox_hdr *rsp_hdr;
|
|
struct mbox_msghdr *msg;
|
|
int offset, vf_id, i;
|
|
|
|
cptpf = container_of(work, struct otx2_cptpf_dev, afpf_mbox_work);
|
|
afpf_mbox = &cptpf->afpf_mbox;
|
|
mdev = &afpf_mbox->dev[0];
|
|
/* Sync mbox data into memory */
|
|
smp_wmb();
|
|
|
|
rsp_hdr = (struct mbox_hdr *)(mdev->mbase + afpf_mbox->rx_start);
|
|
offset = ALIGN(sizeof(*rsp_hdr), MBOX_MSG_ALIGN);
|
|
|
|
for (i = 0; i < rsp_hdr->num_msgs; i++) {
|
|
msg = (struct mbox_msghdr *)(mdev->mbase + afpf_mbox->rx_start +
|
|
offset);
|
|
vf_id = (msg->pcifunc >> RVU_PFVF_FUNC_SHIFT) &
|
|
RVU_PFVF_FUNC_MASK;
|
|
if (vf_id > 0)
|
|
forward_to_vf(cptpf, msg, vf_id,
|
|
msg->next_msgoff - offset);
|
|
else
|
|
process_afpf_mbox_msg(cptpf, msg);
|
|
|
|
offset = msg->next_msgoff;
|
|
/* Sync VF response ready to be sent */
|
|
smp_wmb();
|
|
mdev->msgs_acked++;
|
|
}
|
|
otx2_mbox_reset(afpf_mbox, 0);
|
|
}
|
|
|
|
static void handle_msg_cpt_inst_lmtst(struct otx2_cptpf_dev *cptpf,
|
|
struct mbox_msghdr *msg)
|
|
{
|
|
struct cpt_inst_lmtst_req *req = (struct cpt_inst_lmtst_req *)msg;
|
|
struct otx2_cptlfs_info *lfs = &cptpf->lfs;
|
|
struct msg_rsp *rsp;
|
|
|
|
if (cptpf->lfs.lfs_num)
|
|
lfs->ops->send_cmd((union otx2_cpt_inst_s *)req->inst, 1,
|
|
&lfs->lf[0]);
|
|
|
|
rsp = (struct msg_rsp *)otx2_mbox_alloc_msg(&cptpf->afpf_mbox_up, 0,
|
|
sizeof(*rsp));
|
|
if (!rsp)
|
|
return;
|
|
|
|
rsp->hdr.id = msg->id;
|
|
rsp->hdr.sig = OTX2_MBOX_RSP_SIG;
|
|
rsp->hdr.pcifunc = 0;
|
|
rsp->hdr.rc = 0;
|
|
}
|
|
|
|
static void process_afpf_mbox_up_msg(struct otx2_cptpf_dev *cptpf,
|
|
struct mbox_msghdr *msg)
|
|
{
|
|
if (msg->id >= MBOX_MSG_MAX) {
|
|
dev_err(&cptpf->pdev->dev,
|
|
"MBOX msg with unknown ID %d\n", msg->id);
|
|
return;
|
|
}
|
|
|
|
switch (msg->id) {
|
|
case MBOX_MSG_CPT_INST_LMTST:
|
|
handle_msg_cpt_inst_lmtst(cptpf, msg);
|
|
break;
|
|
default:
|
|
otx2_reply_invalid_msg(&cptpf->afpf_mbox_up, 0, 0, msg->id);
|
|
}
|
|
}
|
|
|
|
void otx2_cptpf_afpf_mbox_up_handler(struct work_struct *work)
|
|
{
|
|
struct otx2_cptpf_dev *cptpf;
|
|
struct otx2_mbox_dev *mdev;
|
|
struct mbox_hdr *rsp_hdr;
|
|
struct mbox_msghdr *msg;
|
|
struct otx2_mbox *mbox;
|
|
int offset, i;
|
|
|
|
cptpf = container_of(work, struct otx2_cptpf_dev, afpf_mbox_up_work);
|
|
mbox = &cptpf->afpf_mbox_up;
|
|
mdev = &mbox->dev[0];
|
|
/* Sync mbox data into memory */
|
|
smp_wmb();
|
|
|
|
rsp_hdr = (struct mbox_hdr *)(mdev->mbase + mbox->rx_start);
|
|
offset = mbox->rx_start + ALIGN(sizeof(*rsp_hdr), MBOX_MSG_ALIGN);
|
|
|
|
for (i = 0; i < rsp_hdr->num_msgs; i++) {
|
|
msg = (struct mbox_msghdr *)(mdev->mbase + offset);
|
|
|
|
process_afpf_mbox_up_msg(cptpf, msg);
|
|
|
|
offset = mbox->rx_start + msg->next_msgoff;
|
|
}
|
|
otx2_mbox_msg_send(mbox, 0);
|
|
}
|