1869 lines
45 KiB
C
1869 lines
45 KiB
C
// SPDX-License-Identifier: GPL-2.0-only
|
|
/* Copyright (C) 2020 Marvell. */
|
|
|
|
#include <linux/ctype.h>
|
|
#include <linux/firmware.h>
|
|
#include "otx2_cptpf_ucode.h"
|
|
#include "otx2_cpt_common.h"
|
|
#include "otx2_cptpf.h"
|
|
#include "otx2_cptlf.h"
|
|
#include "otx2_cpt_reqmgr.h"
|
|
#include "rvu_reg.h"
|
|
|
|
#define CSR_DELAY 30
|
|
|
|
#define LOADFVC_RLEN 8
|
|
#define LOADFVC_MAJOR_OP 0x01
|
|
#define LOADFVC_MINOR_OP 0x08
|
|
|
|
#define CTX_FLUSH_TIMER_CNT 0xFFFFFF
|
|
|
|
struct fw_info_t {
|
|
struct list_head ucodes;
|
|
};
|
|
|
|
static struct otx2_cpt_bitmap get_cores_bmap(struct device *dev,
|
|
struct otx2_cpt_eng_grp_info *eng_grp)
|
|
{
|
|
struct otx2_cpt_bitmap bmap = { {0} };
|
|
bool found = false;
|
|
int i;
|
|
|
|
if (eng_grp->g->engs_num < 0 ||
|
|
eng_grp->g->engs_num > OTX2_CPT_MAX_ENGINES) {
|
|
dev_err(dev, "unsupported number of engines %d on octeontx2\n",
|
|
eng_grp->g->engs_num);
|
|
return bmap;
|
|
}
|
|
|
|
for (i = 0; i < OTX2_CPT_MAX_ETYPES_PER_GRP; i++) {
|
|
if (eng_grp->engs[i].type) {
|
|
bitmap_or(bmap.bits, bmap.bits,
|
|
eng_grp->engs[i].bmap,
|
|
eng_grp->g->engs_num);
|
|
bmap.size = eng_grp->g->engs_num;
|
|
found = true;
|
|
}
|
|
}
|
|
|
|
if (!found)
|
|
dev_err(dev, "No engines reserved for engine group %d\n",
|
|
eng_grp->idx);
|
|
return bmap;
|
|
}
|
|
|
|
static int is_eng_type(int val, int eng_type)
|
|
{
|
|
return val & (1 << eng_type);
|
|
}
|
|
|
|
static int is_2nd_ucode_used(struct otx2_cpt_eng_grp_info *eng_grp)
|
|
{
|
|
if (eng_grp->ucode[1].type)
|
|
return true;
|
|
else
|
|
return false;
|
|
}
|
|
|
|
static void set_ucode_filename(struct otx2_cpt_ucode *ucode,
|
|
const char *filename)
|
|
{
|
|
strscpy(ucode->filename, filename, OTX2_CPT_NAME_LENGTH);
|
|
}
|
|
|
|
static char *get_eng_type_str(int eng_type)
|
|
{
|
|
char *str = "unknown";
|
|
|
|
switch (eng_type) {
|
|
case OTX2_CPT_SE_TYPES:
|
|
str = "SE";
|
|
break;
|
|
|
|
case OTX2_CPT_IE_TYPES:
|
|
str = "IE";
|
|
break;
|
|
|
|
case OTX2_CPT_AE_TYPES:
|
|
str = "AE";
|
|
break;
|
|
}
|
|
return str;
|
|
}
|
|
|
|
static char *get_ucode_type_str(int ucode_type)
|
|
{
|
|
char *str = "unknown";
|
|
|
|
switch (ucode_type) {
|
|
case (1 << OTX2_CPT_SE_TYPES):
|
|
str = "SE";
|
|
break;
|
|
|
|
case (1 << OTX2_CPT_IE_TYPES):
|
|
str = "IE";
|
|
break;
|
|
|
|
case (1 << OTX2_CPT_AE_TYPES):
|
|
str = "AE";
|
|
break;
|
|
|
|
case (1 << OTX2_CPT_SE_TYPES | 1 << OTX2_CPT_IE_TYPES):
|
|
str = "SE+IPSEC";
|
|
break;
|
|
}
|
|
return str;
|
|
}
|
|
|
|
static int get_ucode_type(struct device *dev,
|
|
struct otx2_cpt_ucode_hdr *ucode_hdr,
|
|
int *ucode_type)
|
|
{
|
|
struct otx2_cptpf_dev *cptpf = dev_get_drvdata(dev);
|
|
char ver_str_prefix[OTX2_CPT_UCODE_VER_STR_SZ];
|
|
char tmp_ver_str[OTX2_CPT_UCODE_VER_STR_SZ];
|
|
struct pci_dev *pdev = cptpf->pdev;
|
|
int i, val = 0;
|
|
u8 nn;
|
|
|
|
strscpy(tmp_ver_str, ucode_hdr->ver_str, OTX2_CPT_UCODE_VER_STR_SZ);
|
|
for (i = 0; i < strlen(tmp_ver_str); i++)
|
|
tmp_ver_str[i] = tolower(tmp_ver_str[i]);
|
|
|
|
sprintf(ver_str_prefix, "ocpt-%02d", pdev->revision);
|
|
if (!strnstr(tmp_ver_str, ver_str_prefix, OTX2_CPT_UCODE_VER_STR_SZ))
|
|
return -EINVAL;
|
|
|
|
nn = ucode_hdr->ver_num.nn;
|
|
if (strnstr(tmp_ver_str, "se-", OTX2_CPT_UCODE_VER_STR_SZ) &&
|
|
(nn == OTX2_CPT_SE_UC_TYPE1 || nn == OTX2_CPT_SE_UC_TYPE2 ||
|
|
nn == OTX2_CPT_SE_UC_TYPE3))
|
|
val |= 1 << OTX2_CPT_SE_TYPES;
|
|
if (strnstr(tmp_ver_str, "ie-", OTX2_CPT_UCODE_VER_STR_SZ) &&
|
|
(nn == OTX2_CPT_IE_UC_TYPE1 || nn == OTX2_CPT_IE_UC_TYPE2 ||
|
|
nn == OTX2_CPT_IE_UC_TYPE3))
|
|
val |= 1 << OTX2_CPT_IE_TYPES;
|
|
if (strnstr(tmp_ver_str, "ae", OTX2_CPT_UCODE_VER_STR_SZ) &&
|
|
nn == OTX2_CPT_AE_UC_TYPE)
|
|
val |= 1 << OTX2_CPT_AE_TYPES;
|
|
|
|
*ucode_type = val;
|
|
|
|
if (!val)
|
|
return -EINVAL;
|
|
|
|
return 0;
|
|
}
|
|
|
|
static int __write_ucode_base(struct otx2_cptpf_dev *cptpf, int eng,
|
|
dma_addr_t dma_addr, int blkaddr)
|
|
{
|
|
return otx2_cpt_write_af_reg(&cptpf->afpf_mbox, cptpf->pdev,
|
|
CPT_AF_EXEX_UCODE_BASE(eng),
|
|
(u64)dma_addr, blkaddr);
|
|
}
|
|
|
|
static int cptx_set_ucode_base(struct otx2_cpt_eng_grp_info *eng_grp,
|
|
struct otx2_cptpf_dev *cptpf, int blkaddr)
|
|
{
|
|
struct otx2_cpt_engs_rsvd *engs;
|
|
dma_addr_t dma_addr;
|
|
int i, bit, ret;
|
|
|
|
/* Set PF number for microcode fetches */
|
|
ret = otx2_cpt_write_af_reg(&cptpf->afpf_mbox, cptpf->pdev,
|
|
CPT_AF_PF_FUNC,
|
|
cptpf->pf_id << RVU_PFVF_PF_SHIFT, blkaddr);
|
|
if (ret)
|
|
return ret;
|
|
|
|
for (i = 0; i < OTX2_CPT_MAX_ETYPES_PER_GRP; i++) {
|
|
engs = &eng_grp->engs[i];
|
|
if (!engs->type)
|
|
continue;
|
|
|
|
dma_addr = engs->ucode->dma;
|
|
|
|
/*
|
|
* Set UCODE_BASE only for the cores which are not used,
|
|
* other cores should have already valid UCODE_BASE set
|
|
*/
|
|
for_each_set_bit(bit, engs->bmap, eng_grp->g->engs_num)
|
|
if (!eng_grp->g->eng_ref_cnt[bit]) {
|
|
ret = __write_ucode_base(cptpf, bit, dma_addr,
|
|
blkaddr);
|
|
if (ret)
|
|
return ret;
|
|
}
|
|
}
|
|
return 0;
|
|
}
|
|
|
|
static int cpt_set_ucode_base(struct otx2_cpt_eng_grp_info *eng_grp, void *obj)
|
|
{
|
|
struct otx2_cptpf_dev *cptpf = obj;
|
|
int ret;
|
|
|
|
if (cptpf->has_cpt1) {
|
|
ret = cptx_set_ucode_base(eng_grp, cptpf, BLKADDR_CPT1);
|
|
if (ret)
|
|
return ret;
|
|
}
|
|
return cptx_set_ucode_base(eng_grp, cptpf, BLKADDR_CPT0);
|
|
}
|
|
|
|
static int cptx_detach_and_disable_cores(struct otx2_cpt_eng_grp_info *eng_grp,
|
|
struct otx2_cptpf_dev *cptpf,
|
|
struct otx2_cpt_bitmap bmap,
|
|
int blkaddr)
|
|
{
|
|
int i, timeout = 10;
|
|
int busy, ret;
|
|
u64 reg = 0;
|
|
|
|
/* Detach the cores from group */
|
|
for_each_set_bit(i, bmap.bits, bmap.size) {
|
|
ret = otx2_cpt_read_af_reg(&cptpf->afpf_mbox, cptpf->pdev,
|
|
CPT_AF_EXEX_CTL2(i), ®, blkaddr);
|
|
if (ret)
|
|
return ret;
|
|
|
|
if (reg & (1ull << eng_grp->idx)) {
|
|
eng_grp->g->eng_ref_cnt[i]--;
|
|
reg &= ~(1ull << eng_grp->idx);
|
|
|
|
ret = otx2_cpt_write_af_reg(&cptpf->afpf_mbox,
|
|
cptpf->pdev,
|
|
CPT_AF_EXEX_CTL2(i), reg,
|
|
blkaddr);
|
|
if (ret)
|
|
return ret;
|
|
}
|
|
}
|
|
|
|
/* Wait for cores to become idle */
|
|
do {
|
|
busy = 0;
|
|
usleep_range(10000, 20000);
|
|
if (timeout-- < 0)
|
|
return -EBUSY;
|
|
|
|
for_each_set_bit(i, bmap.bits, bmap.size) {
|
|
ret = otx2_cpt_read_af_reg(&cptpf->afpf_mbox,
|
|
cptpf->pdev,
|
|
CPT_AF_EXEX_STS(i), ®,
|
|
blkaddr);
|
|
if (ret)
|
|
return ret;
|
|
|
|
if (reg & 0x1) {
|
|
busy = 1;
|
|
break;
|
|
}
|
|
}
|
|
} while (busy);
|
|
|
|
/* Disable the cores only if they are not used anymore */
|
|
for_each_set_bit(i, bmap.bits, bmap.size) {
|
|
if (!eng_grp->g->eng_ref_cnt[i]) {
|
|
ret = otx2_cpt_write_af_reg(&cptpf->afpf_mbox,
|
|
cptpf->pdev,
|
|
CPT_AF_EXEX_CTL(i), 0x0,
|
|
blkaddr);
|
|
if (ret)
|
|
return ret;
|
|
}
|
|
}
|
|
|
|
return 0;
|
|
}
|
|
|
|
static int cpt_detach_and_disable_cores(struct otx2_cpt_eng_grp_info *eng_grp,
|
|
void *obj)
|
|
{
|
|
struct otx2_cptpf_dev *cptpf = obj;
|
|
struct otx2_cpt_bitmap bmap;
|
|
int ret;
|
|
|
|
bmap = get_cores_bmap(&cptpf->pdev->dev, eng_grp);
|
|
if (!bmap.size)
|
|
return -EINVAL;
|
|
|
|
if (cptpf->has_cpt1) {
|
|
ret = cptx_detach_and_disable_cores(eng_grp, cptpf, bmap,
|
|
BLKADDR_CPT1);
|
|
if (ret)
|
|
return ret;
|
|
}
|
|
return cptx_detach_and_disable_cores(eng_grp, cptpf, bmap,
|
|
BLKADDR_CPT0);
|
|
}
|
|
|
|
static int cptx_attach_and_enable_cores(struct otx2_cpt_eng_grp_info *eng_grp,
|
|
struct otx2_cptpf_dev *cptpf,
|
|
struct otx2_cpt_bitmap bmap,
|
|
int blkaddr)
|
|
{
|
|
u64 reg = 0;
|
|
int i, ret;
|
|
|
|
/* Attach the cores to the group */
|
|
for_each_set_bit(i, bmap.bits, bmap.size) {
|
|
ret = otx2_cpt_read_af_reg(&cptpf->afpf_mbox, cptpf->pdev,
|
|
CPT_AF_EXEX_CTL2(i), ®, blkaddr);
|
|
if (ret)
|
|
return ret;
|
|
|
|
if (!(reg & (1ull << eng_grp->idx))) {
|
|
eng_grp->g->eng_ref_cnt[i]++;
|
|
reg |= 1ull << eng_grp->idx;
|
|
|
|
ret = otx2_cpt_write_af_reg(&cptpf->afpf_mbox,
|
|
cptpf->pdev,
|
|
CPT_AF_EXEX_CTL2(i), reg,
|
|
blkaddr);
|
|
if (ret)
|
|
return ret;
|
|
}
|
|
}
|
|
|
|
/* Enable the cores */
|
|
for_each_set_bit(i, bmap.bits, bmap.size) {
|
|
ret = otx2_cpt_add_write_af_reg(&cptpf->afpf_mbox, cptpf->pdev,
|
|
CPT_AF_EXEX_CTL(i), 0x1,
|
|
blkaddr);
|
|
if (ret)
|
|
return ret;
|
|
}
|
|
return otx2_cpt_send_af_reg_requests(&cptpf->afpf_mbox, cptpf->pdev);
|
|
}
|
|
|
|
static int cpt_attach_and_enable_cores(struct otx2_cpt_eng_grp_info *eng_grp,
|
|
void *obj)
|
|
{
|
|
struct otx2_cptpf_dev *cptpf = obj;
|
|
struct otx2_cpt_bitmap bmap;
|
|
int ret;
|
|
|
|
bmap = get_cores_bmap(&cptpf->pdev->dev, eng_grp);
|
|
if (!bmap.size)
|
|
return -EINVAL;
|
|
|
|
if (cptpf->has_cpt1) {
|
|
ret = cptx_attach_and_enable_cores(eng_grp, cptpf, bmap,
|
|
BLKADDR_CPT1);
|
|
if (ret)
|
|
return ret;
|
|
}
|
|
return cptx_attach_and_enable_cores(eng_grp, cptpf, bmap, BLKADDR_CPT0);
|
|
}
|
|
|
|
static int load_fw(struct device *dev, struct fw_info_t *fw_info,
|
|
char *filename)
|
|
{
|
|
struct otx2_cpt_ucode_hdr *ucode_hdr;
|
|
struct otx2_cpt_uc_info_t *uc_info;
|
|
int ucode_type, ucode_size;
|
|
int ret;
|
|
|
|
uc_info = kzalloc(sizeof(*uc_info), GFP_KERNEL);
|
|
if (!uc_info)
|
|
return -ENOMEM;
|
|
|
|
ret = request_firmware(&uc_info->fw, filename, dev);
|
|
if (ret)
|
|
goto free_uc_info;
|
|
|
|
ucode_hdr = (struct otx2_cpt_ucode_hdr *)uc_info->fw->data;
|
|
ret = get_ucode_type(dev, ucode_hdr, &ucode_type);
|
|
if (ret)
|
|
goto release_fw;
|
|
|
|
ucode_size = ntohl(ucode_hdr->code_length) * 2;
|
|
if (!ucode_size) {
|
|
dev_err(dev, "Ucode %s invalid size\n", filename);
|
|
ret = -EINVAL;
|
|
goto release_fw;
|
|
}
|
|
|
|
set_ucode_filename(&uc_info->ucode, filename);
|
|
memcpy(uc_info->ucode.ver_str, ucode_hdr->ver_str,
|
|
OTX2_CPT_UCODE_VER_STR_SZ);
|
|
uc_info->ucode.ver_num = ucode_hdr->ver_num;
|
|
uc_info->ucode.type = ucode_type;
|
|
uc_info->ucode.size = ucode_size;
|
|
list_add_tail(&uc_info->list, &fw_info->ucodes);
|
|
|
|
return 0;
|
|
|
|
release_fw:
|
|
release_firmware(uc_info->fw);
|
|
free_uc_info:
|
|
kfree(uc_info);
|
|
return ret;
|
|
}
|
|
|
|
static void cpt_ucode_release_fw(struct fw_info_t *fw_info)
|
|
{
|
|
struct otx2_cpt_uc_info_t *curr, *temp;
|
|
|
|
if (!fw_info)
|
|
return;
|
|
|
|
list_for_each_entry_safe(curr, temp, &fw_info->ucodes, list) {
|
|
list_del(&curr->list);
|
|
release_firmware(curr->fw);
|
|
kfree(curr);
|
|
}
|
|
}
|
|
|
|
static struct otx2_cpt_uc_info_t *get_ucode(struct fw_info_t *fw_info,
|
|
int ucode_type)
|
|
{
|
|
struct otx2_cpt_uc_info_t *curr;
|
|
|
|
list_for_each_entry(curr, &fw_info->ucodes, list) {
|
|
if (!is_eng_type(curr->ucode.type, ucode_type))
|
|
continue;
|
|
|
|
return curr;
|
|
}
|
|
return NULL;
|
|
}
|
|
|
|
static void print_uc_info(struct fw_info_t *fw_info)
|
|
{
|
|
struct otx2_cpt_uc_info_t *curr;
|
|
|
|
list_for_each_entry(curr, &fw_info->ucodes, list) {
|
|
pr_debug("Ucode filename %s\n", curr->ucode.filename);
|
|
pr_debug("Ucode version string %s\n", curr->ucode.ver_str);
|
|
pr_debug("Ucode version %d.%d.%d.%d\n",
|
|
curr->ucode.ver_num.nn, curr->ucode.ver_num.xx,
|
|
curr->ucode.ver_num.yy, curr->ucode.ver_num.zz);
|
|
pr_debug("Ucode type (%d) %s\n", curr->ucode.type,
|
|
get_ucode_type_str(curr->ucode.type));
|
|
pr_debug("Ucode size %d\n", curr->ucode.size);
|
|
pr_debug("Ucode ptr %p\n", curr->fw->data);
|
|
}
|
|
}
|
|
|
|
static int cpt_ucode_load_fw(struct pci_dev *pdev, struct fw_info_t *fw_info)
|
|
{
|
|
char filename[OTX2_CPT_NAME_LENGTH];
|
|
char eng_type[8] = {0};
|
|
int ret, e, i;
|
|
|
|
INIT_LIST_HEAD(&fw_info->ucodes);
|
|
|
|
for (e = 1; e < OTX2_CPT_MAX_ENG_TYPES; e++) {
|
|
strcpy(eng_type, get_eng_type_str(e));
|
|
for (i = 0; i < strlen(eng_type); i++)
|
|
eng_type[i] = tolower(eng_type[i]);
|
|
|
|
snprintf(filename, sizeof(filename), "mrvl/cpt%02d/%s.out",
|
|
pdev->revision, eng_type);
|
|
/* Request firmware for each engine type */
|
|
ret = load_fw(&pdev->dev, fw_info, filename);
|
|
if (ret)
|
|
goto release_fw;
|
|
}
|
|
print_uc_info(fw_info);
|
|
return 0;
|
|
|
|
release_fw:
|
|
cpt_ucode_release_fw(fw_info);
|
|
return ret;
|
|
}
|
|
|
|
struct otx2_cpt_engs_rsvd *find_engines_by_type(
|
|
struct otx2_cpt_eng_grp_info *eng_grp,
|
|
int eng_type)
|
|
{
|
|
int i;
|
|
|
|
for (i = 0; i < OTX2_CPT_MAX_ETYPES_PER_GRP; i++) {
|
|
if (!eng_grp->engs[i].type)
|
|
continue;
|
|
|
|
if (eng_grp->engs[i].type == eng_type)
|
|
return &eng_grp->engs[i];
|
|
}
|
|
return NULL;
|
|
}
|
|
|
|
static int eng_grp_has_eng_type(struct otx2_cpt_eng_grp_info *eng_grp,
|
|
int eng_type)
|
|
{
|
|
struct otx2_cpt_engs_rsvd *engs;
|
|
|
|
engs = find_engines_by_type(eng_grp, eng_type);
|
|
|
|
return (engs != NULL ? 1 : 0);
|
|
}
|
|
|
|
static int update_engines_avail_count(struct device *dev,
|
|
struct otx2_cpt_engs_available *avail,
|
|
struct otx2_cpt_engs_rsvd *engs, int val)
|
|
{
|
|
switch (engs->type) {
|
|
case OTX2_CPT_SE_TYPES:
|
|
avail->se_cnt += val;
|
|
break;
|
|
|
|
case OTX2_CPT_IE_TYPES:
|
|
avail->ie_cnt += val;
|
|
break;
|
|
|
|
case OTX2_CPT_AE_TYPES:
|
|
avail->ae_cnt += val;
|
|
break;
|
|
|
|
default:
|
|
dev_err(dev, "Invalid engine type %d\n", engs->type);
|
|
return -EINVAL;
|
|
}
|
|
return 0;
|
|
}
|
|
|
|
static int update_engines_offset(struct device *dev,
|
|
struct otx2_cpt_engs_available *avail,
|
|
struct otx2_cpt_engs_rsvd *engs)
|
|
{
|
|
switch (engs->type) {
|
|
case OTX2_CPT_SE_TYPES:
|
|
engs->offset = 0;
|
|
break;
|
|
|
|
case OTX2_CPT_IE_TYPES:
|
|
engs->offset = avail->max_se_cnt;
|
|
break;
|
|
|
|
case OTX2_CPT_AE_TYPES:
|
|
engs->offset = avail->max_se_cnt + avail->max_ie_cnt;
|
|
break;
|
|
|
|
default:
|
|
dev_err(dev, "Invalid engine type %d\n", engs->type);
|
|
return -EINVAL;
|
|
}
|
|
return 0;
|
|
}
|
|
|
|
static int release_engines(struct device *dev,
|
|
struct otx2_cpt_eng_grp_info *grp)
|
|
{
|
|
int i, ret = 0;
|
|
|
|
for (i = 0; i < OTX2_CPT_MAX_ETYPES_PER_GRP; i++) {
|
|
if (!grp->engs[i].type)
|
|
continue;
|
|
|
|
if (grp->engs[i].count > 0) {
|
|
ret = update_engines_avail_count(dev, &grp->g->avail,
|
|
&grp->engs[i],
|
|
grp->engs[i].count);
|
|
if (ret)
|
|
return ret;
|
|
}
|
|
|
|
grp->engs[i].type = 0;
|
|
grp->engs[i].count = 0;
|
|
grp->engs[i].offset = 0;
|
|
grp->engs[i].ucode = NULL;
|
|
bitmap_zero(grp->engs[i].bmap, grp->g->engs_num);
|
|
}
|
|
return 0;
|
|
}
|
|
|
|
static int do_reserve_engines(struct device *dev,
|
|
struct otx2_cpt_eng_grp_info *grp,
|
|
struct otx2_cpt_engines *req_engs)
|
|
{
|
|
struct otx2_cpt_engs_rsvd *engs = NULL;
|
|
int i, ret;
|
|
|
|
for (i = 0; i < OTX2_CPT_MAX_ETYPES_PER_GRP; i++) {
|
|
if (!grp->engs[i].type) {
|
|
engs = &grp->engs[i];
|
|
break;
|
|
}
|
|
}
|
|
|
|
if (!engs)
|
|
return -ENOMEM;
|
|
|
|
engs->type = req_engs->type;
|
|
engs->count = req_engs->count;
|
|
|
|
ret = update_engines_offset(dev, &grp->g->avail, engs);
|
|
if (ret)
|
|
return ret;
|
|
|
|
if (engs->count > 0) {
|
|
ret = update_engines_avail_count(dev, &grp->g->avail, engs,
|
|
-engs->count);
|
|
if (ret)
|
|
return ret;
|
|
}
|
|
|
|
return 0;
|
|
}
|
|
|
|
static int check_engines_availability(struct device *dev,
|
|
struct otx2_cpt_eng_grp_info *grp,
|
|
struct otx2_cpt_engines *req_eng)
|
|
{
|
|
int avail_cnt = 0;
|
|
|
|
switch (req_eng->type) {
|
|
case OTX2_CPT_SE_TYPES:
|
|
avail_cnt = grp->g->avail.se_cnt;
|
|
break;
|
|
|
|
case OTX2_CPT_IE_TYPES:
|
|
avail_cnt = grp->g->avail.ie_cnt;
|
|
break;
|
|
|
|
case OTX2_CPT_AE_TYPES:
|
|
avail_cnt = grp->g->avail.ae_cnt;
|
|
break;
|
|
|
|
default:
|
|
dev_err(dev, "Invalid engine type %d\n", req_eng->type);
|
|
return -EINVAL;
|
|
}
|
|
|
|
if (avail_cnt < req_eng->count) {
|
|
dev_err(dev,
|
|
"Error available %s engines %d < than requested %d\n",
|
|
get_eng_type_str(req_eng->type),
|
|
avail_cnt, req_eng->count);
|
|
return -EBUSY;
|
|
}
|
|
return 0;
|
|
}
|
|
|
|
static int reserve_engines(struct device *dev,
|
|
struct otx2_cpt_eng_grp_info *grp,
|
|
struct otx2_cpt_engines *req_engs, int ucodes_cnt)
|
|
{
|
|
int i, ret = 0;
|
|
|
|
/* Validate if a number of requested engines are available */
|
|
for (i = 0; i < ucodes_cnt; i++) {
|
|
ret = check_engines_availability(dev, grp, &req_engs[i]);
|
|
if (ret)
|
|
return ret;
|
|
}
|
|
|
|
/* Reserve requested engines for this engine group */
|
|
for (i = 0; i < ucodes_cnt; i++) {
|
|
ret = do_reserve_engines(dev, grp, &req_engs[i]);
|
|
if (ret)
|
|
return ret;
|
|
}
|
|
return 0;
|
|
}
|
|
|
|
static void ucode_unload(struct device *dev, struct otx2_cpt_ucode *ucode)
|
|
{
|
|
if (ucode->va) {
|
|
dma_free_coherent(dev, OTX2_CPT_UCODE_SZ, ucode->va,
|
|
ucode->dma);
|
|
ucode->va = NULL;
|
|
ucode->dma = 0;
|
|
ucode->size = 0;
|
|
}
|
|
|
|
memset(&ucode->ver_str, 0, OTX2_CPT_UCODE_VER_STR_SZ);
|
|
memset(&ucode->ver_num, 0, sizeof(struct otx2_cpt_ucode_ver_num));
|
|
set_ucode_filename(ucode, "");
|
|
ucode->type = 0;
|
|
}
|
|
|
|
static int copy_ucode_to_dma_mem(struct device *dev,
|
|
struct otx2_cpt_ucode *ucode,
|
|
const u8 *ucode_data)
|
|
{
|
|
u32 i;
|
|
|
|
/* Allocate DMAable space */
|
|
ucode->va = dma_alloc_coherent(dev, OTX2_CPT_UCODE_SZ, &ucode->dma,
|
|
GFP_KERNEL);
|
|
if (!ucode->va)
|
|
return -ENOMEM;
|
|
|
|
memcpy(ucode->va, ucode_data + sizeof(struct otx2_cpt_ucode_hdr),
|
|
ucode->size);
|
|
|
|
/* Byte swap 64-bit */
|
|
for (i = 0; i < (ucode->size / 8); i++)
|
|
cpu_to_be64s(&((u64 *)ucode->va)[i]);
|
|
/* Ucode needs 16-bit swap */
|
|
for (i = 0; i < (ucode->size / 2); i++)
|
|
cpu_to_be16s(&((u16 *)ucode->va)[i]);
|
|
return 0;
|
|
}
|
|
|
|
static int enable_eng_grp(struct otx2_cpt_eng_grp_info *eng_grp,
|
|
void *obj)
|
|
{
|
|
int ret;
|
|
|
|
/* Point microcode to each core of the group */
|
|
ret = cpt_set_ucode_base(eng_grp, obj);
|
|
if (ret)
|
|
return ret;
|
|
|
|
/* Attach the cores to the group and enable them */
|
|
ret = cpt_attach_and_enable_cores(eng_grp, obj);
|
|
|
|
return ret;
|
|
}
|
|
|
|
static int disable_eng_grp(struct device *dev,
|
|
struct otx2_cpt_eng_grp_info *eng_grp,
|
|
void *obj)
|
|
{
|
|
int i, ret;
|
|
|
|
/* Disable all engines used by this group */
|
|
ret = cpt_detach_and_disable_cores(eng_grp, obj);
|
|
if (ret)
|
|
return ret;
|
|
|
|
/* Unload ucode used by this engine group */
|
|
ucode_unload(dev, &eng_grp->ucode[0]);
|
|
ucode_unload(dev, &eng_grp->ucode[1]);
|
|
|
|
for (i = 0; i < OTX2_CPT_MAX_ETYPES_PER_GRP; i++) {
|
|
if (!eng_grp->engs[i].type)
|
|
continue;
|
|
|
|
eng_grp->engs[i].ucode = &eng_grp->ucode[0];
|
|
}
|
|
|
|
/* Clear UCODE_BASE register for each engine used by this group */
|
|
ret = cpt_set_ucode_base(eng_grp, obj);
|
|
|
|
return ret;
|
|
}
|
|
|
|
static void setup_eng_grp_mirroring(struct otx2_cpt_eng_grp_info *dst_grp,
|
|
struct otx2_cpt_eng_grp_info *src_grp)
|
|
{
|
|
/* Setup fields for engine group which is mirrored */
|
|
src_grp->mirror.is_ena = false;
|
|
src_grp->mirror.idx = 0;
|
|
src_grp->mirror.ref_count++;
|
|
|
|
/* Setup fields for mirroring engine group */
|
|
dst_grp->mirror.is_ena = true;
|
|
dst_grp->mirror.idx = src_grp->idx;
|
|
dst_grp->mirror.ref_count = 0;
|
|
}
|
|
|
|
static void remove_eng_grp_mirroring(struct otx2_cpt_eng_grp_info *dst_grp)
|
|
{
|
|
struct otx2_cpt_eng_grp_info *src_grp;
|
|
|
|
if (!dst_grp->mirror.is_ena)
|
|
return;
|
|
|
|
src_grp = &dst_grp->g->grp[dst_grp->mirror.idx];
|
|
|
|
src_grp->mirror.ref_count--;
|
|
dst_grp->mirror.is_ena = false;
|
|
dst_grp->mirror.idx = 0;
|
|
dst_grp->mirror.ref_count = 0;
|
|
}
|
|
|
|
static void update_requested_engs(struct otx2_cpt_eng_grp_info *mirror_eng_grp,
|
|
struct otx2_cpt_engines *engs, int engs_cnt)
|
|
{
|
|
struct otx2_cpt_engs_rsvd *mirrored_engs;
|
|
int i;
|
|
|
|
for (i = 0; i < engs_cnt; i++) {
|
|
mirrored_engs = find_engines_by_type(mirror_eng_grp,
|
|
engs[i].type);
|
|
if (!mirrored_engs)
|
|
continue;
|
|
|
|
/*
|
|
* If mirrored group has this type of engines attached then
|
|
* there are 3 scenarios possible:
|
|
* 1) mirrored_engs.count == engs[i].count then all engines
|
|
* from mirrored engine group will be shared with this engine
|
|
* group
|
|
* 2) mirrored_engs.count > engs[i].count then only a subset of
|
|
* engines from mirrored engine group will be shared with this
|
|
* engine group
|
|
* 3) mirrored_engs.count < engs[i].count then all engines
|
|
* from mirrored engine group will be shared with this group
|
|
* and additional engines will be reserved for exclusively use
|
|
* by this engine group
|
|
*/
|
|
engs[i].count -= mirrored_engs->count;
|
|
}
|
|
}
|
|
|
|
static struct otx2_cpt_eng_grp_info *find_mirrored_eng_grp(
|
|
struct otx2_cpt_eng_grp_info *grp)
|
|
{
|
|
struct otx2_cpt_eng_grps *eng_grps = grp->g;
|
|
int i;
|
|
|
|
for (i = 0; i < OTX2_CPT_MAX_ENGINE_GROUPS; i++) {
|
|
if (!eng_grps->grp[i].is_enabled)
|
|
continue;
|
|
if (eng_grps->grp[i].ucode[0].type &&
|
|
eng_grps->grp[i].ucode[1].type)
|
|
continue;
|
|
if (grp->idx == i)
|
|
continue;
|
|
if (!strncasecmp(eng_grps->grp[i].ucode[0].ver_str,
|
|
grp->ucode[0].ver_str,
|
|
OTX2_CPT_UCODE_VER_STR_SZ))
|
|
return &eng_grps->grp[i];
|
|
}
|
|
|
|
return NULL;
|
|
}
|
|
|
|
static struct otx2_cpt_eng_grp_info *find_unused_eng_grp(
|
|
struct otx2_cpt_eng_grps *eng_grps)
|
|
{
|
|
int i;
|
|
|
|
for (i = 0; i < OTX2_CPT_MAX_ENGINE_GROUPS; i++) {
|
|
if (!eng_grps->grp[i].is_enabled)
|
|
return &eng_grps->grp[i];
|
|
}
|
|
return NULL;
|
|
}
|
|
|
|
static int eng_grp_update_masks(struct device *dev,
|
|
struct otx2_cpt_eng_grp_info *eng_grp)
|
|
{
|
|
struct otx2_cpt_engs_rsvd *engs, *mirrored_engs;
|
|
struct otx2_cpt_bitmap tmp_bmap = { {0} };
|
|
int i, j, cnt, max_cnt;
|
|
int bit;
|
|
|
|
for (i = 0; i < OTX2_CPT_MAX_ETYPES_PER_GRP; i++) {
|
|
engs = &eng_grp->engs[i];
|
|
if (!engs->type)
|
|
continue;
|
|
if (engs->count <= 0)
|
|
continue;
|
|
|
|
switch (engs->type) {
|
|
case OTX2_CPT_SE_TYPES:
|
|
max_cnt = eng_grp->g->avail.max_se_cnt;
|
|
break;
|
|
|
|
case OTX2_CPT_IE_TYPES:
|
|
max_cnt = eng_grp->g->avail.max_ie_cnt;
|
|
break;
|
|
|
|
case OTX2_CPT_AE_TYPES:
|
|
max_cnt = eng_grp->g->avail.max_ae_cnt;
|
|
break;
|
|
|
|
default:
|
|
dev_err(dev, "Invalid engine type %d\n", engs->type);
|
|
return -EINVAL;
|
|
}
|
|
|
|
cnt = engs->count;
|
|
WARN_ON(engs->offset + max_cnt > OTX2_CPT_MAX_ENGINES);
|
|
bitmap_zero(tmp_bmap.bits, eng_grp->g->engs_num);
|
|
for (j = engs->offset; j < engs->offset + max_cnt; j++) {
|
|
if (!eng_grp->g->eng_ref_cnt[j]) {
|
|
bitmap_set(tmp_bmap.bits, j, 1);
|
|
cnt--;
|
|
if (!cnt)
|
|
break;
|
|
}
|
|
}
|
|
|
|
if (cnt)
|
|
return -ENOSPC;
|
|
|
|
bitmap_copy(engs->bmap, tmp_bmap.bits, eng_grp->g->engs_num);
|
|
}
|
|
|
|
if (!eng_grp->mirror.is_ena)
|
|
return 0;
|
|
|
|
for (i = 0; i < OTX2_CPT_MAX_ETYPES_PER_GRP; i++) {
|
|
engs = &eng_grp->engs[i];
|
|
if (!engs->type)
|
|
continue;
|
|
|
|
mirrored_engs = find_engines_by_type(
|
|
&eng_grp->g->grp[eng_grp->mirror.idx],
|
|
engs->type);
|
|
WARN_ON(!mirrored_engs && engs->count <= 0);
|
|
if (!mirrored_engs)
|
|
continue;
|
|
|
|
bitmap_copy(tmp_bmap.bits, mirrored_engs->bmap,
|
|
eng_grp->g->engs_num);
|
|
if (engs->count < 0) {
|
|
bit = find_first_bit(mirrored_engs->bmap,
|
|
eng_grp->g->engs_num);
|
|
bitmap_clear(tmp_bmap.bits, bit, -engs->count);
|
|
}
|
|
bitmap_or(engs->bmap, engs->bmap, tmp_bmap.bits,
|
|
eng_grp->g->engs_num);
|
|
}
|
|
return 0;
|
|
}
|
|
|
|
static int delete_engine_group(struct device *dev,
|
|
struct otx2_cpt_eng_grp_info *eng_grp)
|
|
{
|
|
int ret;
|
|
|
|
if (!eng_grp->is_enabled)
|
|
return 0;
|
|
|
|
if (eng_grp->mirror.ref_count)
|
|
return -EINVAL;
|
|
|
|
/* Removing engine group mirroring if enabled */
|
|
remove_eng_grp_mirroring(eng_grp);
|
|
|
|
/* Disable engine group */
|
|
ret = disable_eng_grp(dev, eng_grp, eng_grp->g->obj);
|
|
if (ret)
|
|
return ret;
|
|
|
|
/* Release all engines held by this engine group */
|
|
ret = release_engines(dev, eng_grp);
|
|
if (ret)
|
|
return ret;
|
|
|
|
eng_grp->is_enabled = false;
|
|
|
|
return 0;
|
|
}
|
|
|
|
static void update_ucode_ptrs(struct otx2_cpt_eng_grp_info *eng_grp)
|
|
{
|
|
struct otx2_cpt_ucode *ucode;
|
|
|
|
if (eng_grp->mirror.is_ena)
|
|
ucode = &eng_grp->g->grp[eng_grp->mirror.idx].ucode[0];
|
|
else
|
|
ucode = &eng_grp->ucode[0];
|
|
WARN_ON(!eng_grp->engs[0].type);
|
|
eng_grp->engs[0].ucode = ucode;
|
|
|
|
if (eng_grp->engs[1].type) {
|
|
if (is_2nd_ucode_used(eng_grp))
|
|
eng_grp->engs[1].ucode = &eng_grp->ucode[1];
|
|
else
|
|
eng_grp->engs[1].ucode = ucode;
|
|
}
|
|
}
|
|
|
|
static int create_engine_group(struct device *dev,
|
|
struct otx2_cpt_eng_grps *eng_grps,
|
|
struct otx2_cpt_engines *engs, int ucodes_cnt,
|
|
void *ucode_data[], int is_print)
|
|
{
|
|
struct otx2_cpt_eng_grp_info *mirrored_eng_grp;
|
|
struct otx2_cpt_eng_grp_info *eng_grp;
|
|
struct otx2_cpt_uc_info_t *uc_info;
|
|
int i, ret = 0;
|
|
|
|
/* Find engine group which is not used */
|
|
eng_grp = find_unused_eng_grp(eng_grps);
|
|
if (!eng_grp) {
|
|
dev_err(dev, "Error all engine groups are being used\n");
|
|
return -ENOSPC;
|
|
}
|
|
/* Load ucode */
|
|
for (i = 0; i < ucodes_cnt; i++) {
|
|
uc_info = (struct otx2_cpt_uc_info_t *) ucode_data[i];
|
|
eng_grp->ucode[i] = uc_info->ucode;
|
|
ret = copy_ucode_to_dma_mem(dev, &eng_grp->ucode[i],
|
|
uc_info->fw->data);
|
|
if (ret)
|
|
goto unload_ucode;
|
|
}
|
|
|
|
/* Check if this group mirrors another existing engine group */
|
|
mirrored_eng_grp = find_mirrored_eng_grp(eng_grp);
|
|
if (mirrored_eng_grp) {
|
|
/* Setup mirroring */
|
|
setup_eng_grp_mirroring(eng_grp, mirrored_eng_grp);
|
|
|
|
/*
|
|
* Update count of requested engines because some
|
|
* of them might be shared with mirrored group
|
|
*/
|
|
update_requested_engs(mirrored_eng_grp, engs, ucodes_cnt);
|
|
}
|
|
ret = reserve_engines(dev, eng_grp, engs, ucodes_cnt);
|
|
if (ret)
|
|
goto unload_ucode;
|
|
|
|
/* Update ucode pointers used by engines */
|
|
update_ucode_ptrs(eng_grp);
|
|
|
|
/* Update engine masks used by this group */
|
|
ret = eng_grp_update_masks(dev, eng_grp);
|
|
if (ret)
|
|
goto release_engs;
|
|
|
|
/* Enable engine group */
|
|
ret = enable_eng_grp(eng_grp, eng_grps->obj);
|
|
if (ret)
|
|
goto release_engs;
|
|
|
|
/*
|
|
* If this engine group mirrors another engine group
|
|
* then we need to unload ucode as we will use ucode
|
|
* from mirrored engine group
|
|
*/
|
|
if (eng_grp->mirror.is_ena)
|
|
ucode_unload(dev, &eng_grp->ucode[0]);
|
|
|
|
eng_grp->is_enabled = true;
|
|
|
|
if (!is_print)
|
|
return 0;
|
|
|
|
if (mirrored_eng_grp)
|
|
dev_info(dev,
|
|
"Engine_group%d: reuse microcode %s from group %d\n",
|
|
eng_grp->idx, mirrored_eng_grp->ucode[0].ver_str,
|
|
mirrored_eng_grp->idx);
|
|
else
|
|
dev_info(dev, "Engine_group%d: microcode loaded %s\n",
|
|
eng_grp->idx, eng_grp->ucode[0].ver_str);
|
|
if (is_2nd_ucode_used(eng_grp))
|
|
dev_info(dev, "Engine_group%d: microcode loaded %s\n",
|
|
eng_grp->idx, eng_grp->ucode[1].ver_str);
|
|
|
|
return 0;
|
|
|
|
release_engs:
|
|
release_engines(dev, eng_grp);
|
|
unload_ucode:
|
|
ucode_unload(dev, &eng_grp->ucode[0]);
|
|
ucode_unload(dev, &eng_grp->ucode[1]);
|
|
return ret;
|
|
}
|
|
|
|
static void delete_engine_grps(struct pci_dev *pdev,
|
|
struct otx2_cpt_eng_grps *eng_grps)
|
|
{
|
|
int i;
|
|
|
|
/* First delete all mirroring engine groups */
|
|
for (i = 0; i < OTX2_CPT_MAX_ENGINE_GROUPS; i++)
|
|
if (eng_grps->grp[i].mirror.is_ena)
|
|
delete_engine_group(&pdev->dev, &eng_grps->grp[i]);
|
|
|
|
/* Delete remaining engine groups */
|
|
for (i = 0; i < OTX2_CPT_MAX_ENGINE_GROUPS; i++)
|
|
delete_engine_group(&pdev->dev, &eng_grps->grp[i]);
|
|
}
|
|
|
|
#define PCI_DEVID_CN10K_RNM 0xA098
|
|
#define RNM_ENTROPY_STATUS 0x8
|
|
|
|
static void rnm_to_cpt_errata_fixup(struct device *dev)
|
|
{
|
|
struct pci_dev *pdev;
|
|
void __iomem *base;
|
|
int timeout = 5000;
|
|
|
|
pdev = pci_get_device(PCI_VENDOR_ID_CAVIUM, PCI_DEVID_CN10K_RNM, NULL);
|
|
if (!pdev)
|
|
return;
|
|
|
|
base = pci_ioremap_bar(pdev, 0);
|
|
if (!base)
|
|
goto put_pdev;
|
|
|
|
while ((readq(base + RNM_ENTROPY_STATUS) & 0x7F) != 0x40) {
|
|
cpu_relax();
|
|
udelay(1);
|
|
timeout--;
|
|
if (!timeout) {
|
|
dev_warn(dev, "RNM is not producing entropy\n");
|
|
break;
|
|
}
|
|
}
|
|
|
|
iounmap(base);
|
|
|
|
put_pdev:
|
|
pci_dev_put(pdev);
|
|
}
|
|
|
|
int otx2_cpt_get_eng_grp(struct otx2_cpt_eng_grps *eng_grps, int eng_type)
|
|
{
|
|
|
|
int eng_grp_num = OTX2_CPT_INVALID_CRYPTO_ENG_GRP;
|
|
struct otx2_cpt_eng_grp_info *grp;
|
|
int i;
|
|
|
|
for (i = 0; i < OTX2_CPT_MAX_ENGINE_GROUPS; i++) {
|
|
grp = &eng_grps->grp[i];
|
|
if (!grp->is_enabled)
|
|
continue;
|
|
|
|
if (eng_type == OTX2_CPT_SE_TYPES) {
|
|
if (eng_grp_has_eng_type(grp, eng_type) &&
|
|
!eng_grp_has_eng_type(grp, OTX2_CPT_IE_TYPES)) {
|
|
eng_grp_num = i;
|
|
break;
|
|
}
|
|
} else {
|
|
if (eng_grp_has_eng_type(grp, eng_type)) {
|
|
eng_grp_num = i;
|
|
break;
|
|
}
|
|
}
|
|
}
|
|
return eng_grp_num;
|
|
}
|
|
|
|
int otx2_cpt_create_eng_grps(struct otx2_cptpf_dev *cptpf,
|
|
struct otx2_cpt_eng_grps *eng_grps)
|
|
{
|
|
struct otx2_cpt_uc_info_t *uc_info[OTX2_CPT_MAX_ETYPES_PER_GRP] = { };
|
|
struct otx2_cpt_engines engs[OTX2_CPT_MAX_ETYPES_PER_GRP] = { {0} };
|
|
struct pci_dev *pdev = cptpf->pdev;
|
|
struct fw_info_t fw_info;
|
|
u64 reg_val;
|
|
int ret = 0;
|
|
|
|
mutex_lock(&eng_grps->lock);
|
|
/*
|
|
* We don't create engine groups if it was already
|
|
* made (when user enabled VFs for the first time)
|
|
*/
|
|
if (eng_grps->is_grps_created)
|
|
goto unlock;
|
|
|
|
ret = cpt_ucode_load_fw(pdev, &fw_info);
|
|
if (ret)
|
|
goto unlock;
|
|
|
|
/*
|
|
* Create engine group with SE engines for kernel
|
|
* crypto functionality (symmetric crypto)
|
|
*/
|
|
uc_info[0] = get_ucode(&fw_info, OTX2_CPT_SE_TYPES);
|
|
if (uc_info[0] == NULL) {
|
|
dev_err(&pdev->dev, "Unable to find firmware for SE\n");
|
|
ret = -EINVAL;
|
|
goto release_fw;
|
|
}
|
|
engs[0].type = OTX2_CPT_SE_TYPES;
|
|
engs[0].count = eng_grps->avail.max_se_cnt;
|
|
|
|
ret = create_engine_group(&pdev->dev, eng_grps, engs, 1,
|
|
(void **) uc_info, 1);
|
|
if (ret)
|
|
goto release_fw;
|
|
|
|
/*
|
|
* Create engine group with SE+IE engines for IPSec.
|
|
* All SE engines will be shared with engine group 0.
|
|
*/
|
|
uc_info[0] = get_ucode(&fw_info, OTX2_CPT_SE_TYPES);
|
|
uc_info[1] = get_ucode(&fw_info, OTX2_CPT_IE_TYPES);
|
|
|
|
if (uc_info[1] == NULL) {
|
|
dev_err(&pdev->dev, "Unable to find firmware for IE");
|
|
ret = -EINVAL;
|
|
goto delete_eng_grp;
|
|
}
|
|
engs[0].type = OTX2_CPT_SE_TYPES;
|
|
engs[0].count = eng_grps->avail.max_se_cnt;
|
|
engs[1].type = OTX2_CPT_IE_TYPES;
|
|
engs[1].count = eng_grps->avail.max_ie_cnt;
|
|
|
|
ret = create_engine_group(&pdev->dev, eng_grps, engs, 2,
|
|
(void **) uc_info, 1);
|
|
if (ret)
|
|
goto delete_eng_grp;
|
|
|
|
/*
|
|
* Create engine group with AE engines for asymmetric
|
|
* crypto functionality.
|
|
*/
|
|
uc_info[0] = get_ucode(&fw_info, OTX2_CPT_AE_TYPES);
|
|
if (uc_info[0] == NULL) {
|
|
dev_err(&pdev->dev, "Unable to find firmware for AE");
|
|
ret = -EINVAL;
|
|
goto delete_eng_grp;
|
|
}
|
|
engs[0].type = OTX2_CPT_AE_TYPES;
|
|
engs[0].count = eng_grps->avail.max_ae_cnt;
|
|
|
|
ret = create_engine_group(&pdev->dev, eng_grps, engs, 1,
|
|
(void **) uc_info, 1);
|
|
if (ret)
|
|
goto delete_eng_grp;
|
|
|
|
eng_grps->is_grps_created = true;
|
|
|
|
cpt_ucode_release_fw(&fw_info);
|
|
|
|
if (is_dev_otx2(pdev))
|
|
goto unlock;
|
|
|
|
/*
|
|
* Ensure RNM_ENTROPY_STATUS[NORMAL_CNT] = 0x40 before writing
|
|
* CPT_AF_CTL[RNM_REQ_EN] = 1 as a workaround for HW errata.
|
|
*/
|
|
rnm_to_cpt_errata_fixup(&pdev->dev);
|
|
|
|
/*
|
|
* Configure engine group mask to allow context prefetching
|
|
* for the groups and enable random number request, to enable
|
|
* CPT to request random numbers from RNM.
|
|
*/
|
|
otx2_cpt_write_af_reg(&cptpf->afpf_mbox, pdev, CPT_AF_CTL,
|
|
OTX2_CPT_ALL_ENG_GRPS_MASK << 3 | BIT_ULL(16),
|
|
BLKADDR_CPT0);
|
|
/*
|
|
* Set interval to periodically flush dirty data for the next
|
|
* CTX cache entry. Set the interval count to maximum supported
|
|
* value.
|
|
*/
|
|
otx2_cpt_write_af_reg(&cptpf->afpf_mbox, pdev, CPT_AF_CTX_FLUSH_TIMER,
|
|
CTX_FLUSH_TIMER_CNT, BLKADDR_CPT0);
|
|
|
|
/*
|
|
* Set CPT_AF_DIAG[FLT_DIS], as a workaround for HW errata, when
|
|
* CPT_AF_DIAG[FLT_DIS] = 0 and a CPT engine access to LLC/DRAM
|
|
* encounters a fault/poison, a rare case may result in
|
|
* unpredictable data being delivered to a CPT engine.
|
|
*/
|
|
otx2_cpt_read_af_reg(&cptpf->afpf_mbox, pdev, CPT_AF_DIAG, ®_val,
|
|
BLKADDR_CPT0);
|
|
otx2_cpt_write_af_reg(&cptpf->afpf_mbox, pdev, CPT_AF_DIAG,
|
|
reg_val | BIT_ULL(24), BLKADDR_CPT0);
|
|
|
|
mutex_unlock(&eng_grps->lock);
|
|
return 0;
|
|
|
|
delete_eng_grp:
|
|
delete_engine_grps(pdev, eng_grps);
|
|
release_fw:
|
|
cpt_ucode_release_fw(&fw_info);
|
|
unlock:
|
|
mutex_unlock(&eng_grps->lock);
|
|
return ret;
|
|
}
|
|
|
|
static int cptx_disable_all_cores(struct otx2_cptpf_dev *cptpf, int total_cores,
|
|
int blkaddr)
|
|
{
|
|
int timeout = 10, ret;
|
|
int i, busy;
|
|
u64 reg;
|
|
|
|
/* Disengage the cores from groups */
|
|
for (i = 0; i < total_cores; i++) {
|
|
ret = otx2_cpt_add_write_af_reg(&cptpf->afpf_mbox, cptpf->pdev,
|
|
CPT_AF_EXEX_CTL2(i), 0x0,
|
|
blkaddr);
|
|
if (ret)
|
|
return ret;
|
|
|
|
cptpf->eng_grps.eng_ref_cnt[i] = 0;
|
|
}
|
|
ret = otx2_cpt_send_af_reg_requests(&cptpf->afpf_mbox, cptpf->pdev);
|
|
if (ret)
|
|
return ret;
|
|
|
|
/* Wait for cores to become idle */
|
|
do {
|
|
busy = 0;
|
|
usleep_range(10000, 20000);
|
|
if (timeout-- < 0)
|
|
return -EBUSY;
|
|
|
|
for (i = 0; i < total_cores; i++) {
|
|
ret = otx2_cpt_read_af_reg(&cptpf->afpf_mbox,
|
|
cptpf->pdev,
|
|
CPT_AF_EXEX_STS(i), ®,
|
|
blkaddr);
|
|
if (ret)
|
|
return ret;
|
|
|
|
if (reg & 0x1) {
|
|
busy = 1;
|
|
break;
|
|
}
|
|
}
|
|
} while (busy);
|
|
|
|
/* Disable the cores */
|
|
for (i = 0; i < total_cores; i++) {
|
|
ret = otx2_cpt_add_write_af_reg(&cptpf->afpf_mbox, cptpf->pdev,
|
|
CPT_AF_EXEX_CTL(i), 0x0,
|
|
blkaddr);
|
|
if (ret)
|
|
return ret;
|
|
}
|
|
return otx2_cpt_send_af_reg_requests(&cptpf->afpf_mbox, cptpf->pdev);
|
|
}
|
|
|
|
int otx2_cpt_disable_all_cores(struct otx2_cptpf_dev *cptpf)
|
|
{
|
|
int total_cores, ret;
|
|
|
|
total_cores = cptpf->eng_grps.avail.max_se_cnt +
|
|
cptpf->eng_grps.avail.max_ie_cnt +
|
|
cptpf->eng_grps.avail.max_ae_cnt;
|
|
|
|
if (cptpf->has_cpt1) {
|
|
ret = cptx_disable_all_cores(cptpf, total_cores, BLKADDR_CPT1);
|
|
if (ret)
|
|
return ret;
|
|
}
|
|
return cptx_disable_all_cores(cptpf, total_cores, BLKADDR_CPT0);
|
|
}
|
|
|
|
void otx2_cpt_cleanup_eng_grps(struct pci_dev *pdev,
|
|
struct otx2_cpt_eng_grps *eng_grps)
|
|
{
|
|
struct otx2_cpt_eng_grp_info *grp;
|
|
int i, j;
|
|
|
|
mutex_lock(&eng_grps->lock);
|
|
delete_engine_grps(pdev, eng_grps);
|
|
/* Release memory */
|
|
for (i = 0; i < OTX2_CPT_MAX_ENGINE_GROUPS; i++) {
|
|
grp = &eng_grps->grp[i];
|
|
for (j = 0; j < OTX2_CPT_MAX_ETYPES_PER_GRP; j++) {
|
|
kfree(grp->engs[j].bmap);
|
|
grp->engs[j].bmap = NULL;
|
|
}
|
|
}
|
|
mutex_unlock(&eng_grps->lock);
|
|
}
|
|
|
|
int otx2_cpt_init_eng_grps(struct pci_dev *pdev,
|
|
struct otx2_cpt_eng_grps *eng_grps)
|
|
{
|
|
struct otx2_cpt_eng_grp_info *grp;
|
|
int i, j, ret;
|
|
|
|
mutex_init(&eng_grps->lock);
|
|
eng_grps->obj = pci_get_drvdata(pdev);
|
|
eng_grps->avail.se_cnt = eng_grps->avail.max_se_cnt;
|
|
eng_grps->avail.ie_cnt = eng_grps->avail.max_ie_cnt;
|
|
eng_grps->avail.ae_cnt = eng_grps->avail.max_ae_cnt;
|
|
|
|
eng_grps->engs_num = eng_grps->avail.max_se_cnt +
|
|
eng_grps->avail.max_ie_cnt +
|
|
eng_grps->avail.max_ae_cnt;
|
|
if (eng_grps->engs_num > OTX2_CPT_MAX_ENGINES) {
|
|
dev_err(&pdev->dev,
|
|
"Number of engines %d > than max supported %d\n",
|
|
eng_grps->engs_num, OTX2_CPT_MAX_ENGINES);
|
|
ret = -EINVAL;
|
|
goto cleanup_eng_grps;
|
|
}
|
|
|
|
for (i = 0; i < OTX2_CPT_MAX_ENGINE_GROUPS; i++) {
|
|
grp = &eng_grps->grp[i];
|
|
grp->g = eng_grps;
|
|
grp->idx = i;
|
|
|
|
for (j = 0; j < OTX2_CPT_MAX_ETYPES_PER_GRP; j++) {
|
|
grp->engs[j].bmap =
|
|
kcalloc(BITS_TO_LONGS(eng_grps->engs_num),
|
|
sizeof(long), GFP_KERNEL);
|
|
if (!grp->engs[j].bmap) {
|
|
ret = -ENOMEM;
|
|
goto cleanup_eng_grps;
|
|
}
|
|
}
|
|
}
|
|
return 0;
|
|
|
|
cleanup_eng_grps:
|
|
otx2_cpt_cleanup_eng_grps(pdev, eng_grps);
|
|
return ret;
|
|
}
|
|
|
|
static int create_eng_caps_discovery_grps(struct pci_dev *pdev,
|
|
struct otx2_cpt_eng_grps *eng_grps)
|
|
{
|
|
struct otx2_cpt_uc_info_t *uc_info[OTX2_CPT_MAX_ETYPES_PER_GRP] = { };
|
|
struct otx2_cpt_engines engs[OTX2_CPT_MAX_ETYPES_PER_GRP] = { {0} };
|
|
struct fw_info_t fw_info;
|
|
int ret;
|
|
|
|
mutex_lock(&eng_grps->lock);
|
|
ret = cpt_ucode_load_fw(pdev, &fw_info);
|
|
if (ret) {
|
|
mutex_unlock(&eng_grps->lock);
|
|
return ret;
|
|
}
|
|
|
|
uc_info[0] = get_ucode(&fw_info, OTX2_CPT_AE_TYPES);
|
|
if (uc_info[0] == NULL) {
|
|
dev_err(&pdev->dev, "Unable to find firmware for AE\n");
|
|
ret = -EINVAL;
|
|
goto release_fw;
|
|
}
|
|
engs[0].type = OTX2_CPT_AE_TYPES;
|
|
engs[0].count = 2;
|
|
|
|
ret = create_engine_group(&pdev->dev, eng_grps, engs, 1,
|
|
(void **) uc_info, 0);
|
|
if (ret)
|
|
goto release_fw;
|
|
|
|
uc_info[0] = get_ucode(&fw_info, OTX2_CPT_SE_TYPES);
|
|
if (uc_info[0] == NULL) {
|
|
dev_err(&pdev->dev, "Unable to find firmware for SE\n");
|
|
ret = -EINVAL;
|
|
goto delete_eng_grp;
|
|
}
|
|
engs[0].type = OTX2_CPT_SE_TYPES;
|
|
engs[0].count = 2;
|
|
|
|
ret = create_engine_group(&pdev->dev, eng_grps, engs, 1,
|
|
(void **) uc_info, 0);
|
|
if (ret)
|
|
goto delete_eng_grp;
|
|
|
|
uc_info[0] = get_ucode(&fw_info, OTX2_CPT_IE_TYPES);
|
|
if (uc_info[0] == NULL) {
|
|
dev_err(&pdev->dev, "Unable to find firmware for IE\n");
|
|
ret = -EINVAL;
|
|
goto delete_eng_grp;
|
|
}
|
|
engs[0].type = OTX2_CPT_IE_TYPES;
|
|
engs[0].count = 2;
|
|
|
|
ret = create_engine_group(&pdev->dev, eng_grps, engs, 1,
|
|
(void **) uc_info, 0);
|
|
if (ret)
|
|
goto delete_eng_grp;
|
|
|
|
cpt_ucode_release_fw(&fw_info);
|
|
mutex_unlock(&eng_grps->lock);
|
|
return 0;
|
|
|
|
delete_eng_grp:
|
|
delete_engine_grps(pdev, eng_grps);
|
|
release_fw:
|
|
cpt_ucode_release_fw(&fw_info);
|
|
mutex_unlock(&eng_grps->lock);
|
|
return ret;
|
|
}
|
|
|
|
/*
|
|
* Get CPT HW capabilities using LOAD_FVC operation.
|
|
*/
|
|
int otx2_cpt_discover_eng_capabilities(struct otx2_cptpf_dev *cptpf)
|
|
{
|
|
struct otx2_cptlfs_info *lfs = &cptpf->lfs;
|
|
struct otx2_cpt_iq_command iq_cmd;
|
|
union otx2_cpt_opcode opcode;
|
|
union otx2_cpt_res_s *result;
|
|
union otx2_cpt_inst_s inst;
|
|
dma_addr_t rptr_baddr;
|
|
struct pci_dev *pdev;
|
|
u32 len, compl_rlen;
|
|
int ret, etype;
|
|
void *rptr;
|
|
|
|
/*
|
|
* We don't get capabilities if it was already done
|
|
* (when user enabled VFs for the first time)
|
|
*/
|
|
if (cptpf->is_eng_caps_discovered)
|
|
return 0;
|
|
|
|
pdev = cptpf->pdev;
|
|
/*
|
|
* Create engine groups for each type to submit LOAD_FVC op and
|
|
* get engine's capabilities.
|
|
*/
|
|
ret = create_eng_caps_discovery_grps(pdev, &cptpf->eng_grps);
|
|
if (ret)
|
|
goto delete_grps;
|
|
|
|
lfs->pdev = pdev;
|
|
lfs->reg_base = cptpf->reg_base;
|
|
lfs->mbox = &cptpf->afpf_mbox;
|
|
lfs->blkaddr = BLKADDR_CPT0;
|
|
ret = otx2_cptlf_init(&cptpf->lfs, OTX2_CPT_ALL_ENG_GRPS_MASK,
|
|
OTX2_CPT_QUEUE_HI_PRIO, 1);
|
|
if (ret)
|
|
goto delete_grps;
|
|
|
|
compl_rlen = ALIGN(sizeof(union otx2_cpt_res_s), OTX2_CPT_DMA_MINALIGN);
|
|
len = compl_rlen + LOADFVC_RLEN;
|
|
|
|
result = kzalloc(len, GFP_KERNEL);
|
|
if (!result) {
|
|
ret = -ENOMEM;
|
|
goto lf_cleanup;
|
|
}
|
|
rptr_baddr = dma_map_single(&pdev->dev, (void *)result, len,
|
|
DMA_BIDIRECTIONAL);
|
|
if (dma_mapping_error(&pdev->dev, rptr_baddr)) {
|
|
dev_err(&pdev->dev, "DMA mapping failed\n");
|
|
ret = -EFAULT;
|
|
goto free_result;
|
|
}
|
|
rptr = (u8 *)result + compl_rlen;
|
|
|
|
/* Fill in the command */
|
|
opcode.s.major = LOADFVC_MAJOR_OP;
|
|
opcode.s.minor = LOADFVC_MINOR_OP;
|
|
|
|
iq_cmd.cmd.u = 0;
|
|
iq_cmd.cmd.s.opcode = cpu_to_be16(opcode.flags);
|
|
|
|
/* 64-bit swap for microcode data reads, not needed for addresses */
|
|
cpu_to_be64s(&iq_cmd.cmd.u);
|
|
iq_cmd.dptr = 0;
|
|
iq_cmd.rptr = rptr_baddr + compl_rlen;
|
|
iq_cmd.cptr.u = 0;
|
|
|
|
for (etype = 1; etype < OTX2_CPT_MAX_ENG_TYPES; etype++) {
|
|
result->s.compcode = OTX2_CPT_COMPLETION_CODE_INIT;
|
|
iq_cmd.cptr.s.grp = otx2_cpt_get_eng_grp(&cptpf->eng_grps,
|
|
etype);
|
|
otx2_cpt_fill_inst(&inst, &iq_cmd, rptr_baddr);
|
|
lfs->ops->send_cmd(&inst, 1, &cptpf->lfs.lf[0]);
|
|
|
|
while (lfs->ops->cpt_get_compcode(result) ==
|
|
OTX2_CPT_COMPLETION_CODE_INIT)
|
|
cpu_relax();
|
|
|
|
cptpf->eng_caps[etype].u = be64_to_cpup(rptr);
|
|
}
|
|
dma_unmap_single(&pdev->dev, rptr_baddr, len, DMA_BIDIRECTIONAL);
|
|
cptpf->is_eng_caps_discovered = true;
|
|
|
|
free_result:
|
|
kfree(result);
|
|
lf_cleanup:
|
|
otx2_cptlf_shutdown(&cptpf->lfs);
|
|
delete_grps:
|
|
delete_engine_grps(pdev, &cptpf->eng_grps);
|
|
|
|
return ret;
|
|
}
|
|
|
|
int otx2_cpt_dl_custom_egrp_create(struct otx2_cptpf_dev *cptpf,
|
|
struct devlink_param_gset_ctx *ctx)
|
|
{
|
|
struct otx2_cpt_engines engs[OTX2_CPT_MAX_ETYPES_PER_GRP] = { { 0 } };
|
|
struct otx2_cpt_uc_info_t *uc_info[OTX2_CPT_MAX_ETYPES_PER_GRP] = {};
|
|
struct otx2_cpt_eng_grps *eng_grps = &cptpf->eng_grps;
|
|
char *ucode_filename[OTX2_CPT_MAX_ETYPES_PER_GRP];
|
|
char tmp_buf[OTX2_CPT_NAME_LENGTH] = { 0 };
|
|
struct device *dev = &cptpf->pdev->dev;
|
|
char *start, *val, *err_msg, *tmp;
|
|
int grp_idx = 0, ret = -EINVAL;
|
|
bool has_se, has_ie, has_ae;
|
|
struct fw_info_t fw_info;
|
|
int ucode_idx = 0;
|
|
|
|
if (!eng_grps->is_grps_created) {
|
|
dev_err(dev, "Not allowed before creating the default groups\n");
|
|
return -EINVAL;
|
|
}
|
|
err_msg = "Invalid engine group format";
|
|
strscpy(tmp_buf, ctx->val.vstr, strlen(ctx->val.vstr) + 1);
|
|
start = tmp_buf;
|
|
|
|
has_se = has_ie = has_ae = false;
|
|
|
|
for (;;) {
|
|
val = strsep(&start, ";");
|
|
if (!val)
|
|
break;
|
|
val = strim(val);
|
|
if (!*val)
|
|
continue;
|
|
|
|
if (!strncasecmp(val, "se", 2) && strchr(val, ':')) {
|
|
if (has_se || ucode_idx)
|
|
goto err_print;
|
|
tmp = strsep(&val, ":");
|
|
if (!tmp)
|
|
goto err_print;
|
|
tmp = strim(tmp);
|
|
if (!val)
|
|
goto err_print;
|
|
if (strlen(tmp) != 2)
|
|
goto err_print;
|
|
if (kstrtoint(strim(val), 10, &engs[grp_idx].count))
|
|
goto err_print;
|
|
engs[grp_idx++].type = OTX2_CPT_SE_TYPES;
|
|
has_se = true;
|
|
} else if (!strncasecmp(val, "ae", 2) && strchr(val, ':')) {
|
|
if (has_ae || ucode_idx)
|
|
goto err_print;
|
|
tmp = strsep(&val, ":");
|
|
if (!tmp)
|
|
goto err_print;
|
|
tmp = strim(tmp);
|
|
if (!val)
|
|
goto err_print;
|
|
if (strlen(tmp) != 2)
|
|
goto err_print;
|
|
if (kstrtoint(strim(val), 10, &engs[grp_idx].count))
|
|
goto err_print;
|
|
engs[grp_idx++].type = OTX2_CPT_AE_TYPES;
|
|
has_ae = true;
|
|
} else if (!strncasecmp(val, "ie", 2) && strchr(val, ':')) {
|
|
if (has_ie || ucode_idx)
|
|
goto err_print;
|
|
tmp = strsep(&val, ":");
|
|
if (!tmp)
|
|
goto err_print;
|
|
tmp = strim(tmp);
|
|
if (!val)
|
|
goto err_print;
|
|
if (strlen(tmp) != 2)
|
|
goto err_print;
|
|
if (kstrtoint(strim(val), 10, &engs[grp_idx].count))
|
|
goto err_print;
|
|
engs[grp_idx++].type = OTX2_CPT_IE_TYPES;
|
|
has_ie = true;
|
|
} else {
|
|
if (ucode_idx > 1)
|
|
goto err_print;
|
|
if (!strlen(val))
|
|
goto err_print;
|
|
if (strnstr(val, " ", strlen(val)))
|
|
goto err_print;
|
|
ucode_filename[ucode_idx++] = val;
|
|
}
|
|
}
|
|
|
|
/* Validate input parameters */
|
|
if (!(grp_idx && ucode_idx))
|
|
goto err_print;
|
|
|
|
if (ucode_idx > 1 && grp_idx < 2)
|
|
goto err_print;
|
|
|
|
if (grp_idx > OTX2_CPT_MAX_ETYPES_PER_GRP) {
|
|
err_msg = "Error max 2 engine types can be attached";
|
|
goto err_print;
|
|
}
|
|
|
|
if (grp_idx > 1) {
|
|
if ((engs[0].type + engs[1].type) !=
|
|
(OTX2_CPT_SE_TYPES + OTX2_CPT_IE_TYPES)) {
|
|
err_msg = "Only combination of SE+IE engines is allowed";
|
|
goto err_print;
|
|
}
|
|
/* Keep SE engines at zero index */
|
|
if (engs[1].type == OTX2_CPT_SE_TYPES)
|
|
swap(engs[0], engs[1]);
|
|
}
|
|
mutex_lock(&eng_grps->lock);
|
|
|
|
if (cptpf->enabled_vfs) {
|
|
dev_err(dev, "Disable VFs before modifying engine groups\n");
|
|
ret = -EACCES;
|
|
goto err_unlock;
|
|
}
|
|
INIT_LIST_HEAD(&fw_info.ucodes);
|
|
ret = load_fw(dev, &fw_info, ucode_filename[0]);
|
|
if (ret) {
|
|
dev_err(dev, "Unable to load firmware %s\n", ucode_filename[0]);
|
|
goto err_unlock;
|
|
}
|
|
if (ucode_idx > 1) {
|
|
ret = load_fw(dev, &fw_info, ucode_filename[1]);
|
|
if (ret) {
|
|
dev_err(dev, "Unable to load firmware %s\n",
|
|
ucode_filename[1]);
|
|
goto release_fw;
|
|
}
|
|
}
|
|
uc_info[0] = get_ucode(&fw_info, engs[0].type);
|
|
if (uc_info[0] == NULL) {
|
|
dev_err(dev, "Unable to find firmware for %s\n",
|
|
get_eng_type_str(engs[0].type));
|
|
ret = -EINVAL;
|
|
goto release_fw;
|
|
}
|
|
if (ucode_idx > 1) {
|
|
uc_info[1] = get_ucode(&fw_info, engs[1].type);
|
|
if (uc_info[1] == NULL) {
|
|
dev_err(dev, "Unable to find firmware for %s\n",
|
|
get_eng_type_str(engs[1].type));
|
|
ret = -EINVAL;
|
|
goto release_fw;
|
|
}
|
|
}
|
|
ret = create_engine_group(dev, eng_grps, engs, grp_idx,
|
|
(void **)uc_info, 1);
|
|
|
|
release_fw:
|
|
cpt_ucode_release_fw(&fw_info);
|
|
err_unlock:
|
|
mutex_unlock(&eng_grps->lock);
|
|
return ret;
|
|
err_print:
|
|
dev_err(dev, "%s\n", err_msg);
|
|
return ret;
|
|
}
|
|
|
|
int otx2_cpt_dl_custom_egrp_delete(struct otx2_cptpf_dev *cptpf,
|
|
struct devlink_param_gset_ctx *ctx)
|
|
{
|
|
struct otx2_cpt_eng_grps *eng_grps = &cptpf->eng_grps;
|
|
struct device *dev = &cptpf->pdev->dev;
|
|
char *tmp, *err_msg;
|
|
int egrp;
|
|
int ret;
|
|
|
|
err_msg = "Invalid input string format(ex: egrp:0)";
|
|
if (strncasecmp(ctx->val.vstr, "egrp", 4))
|
|
goto err_print;
|
|
tmp = ctx->val.vstr;
|
|
strsep(&tmp, ":");
|
|
if (!tmp)
|
|
goto err_print;
|
|
if (kstrtoint(tmp, 10, &egrp))
|
|
goto err_print;
|
|
|
|
if (egrp < 0 || egrp >= OTX2_CPT_MAX_ENGINE_GROUPS) {
|
|
dev_err(dev, "Invalid engine group %d", egrp);
|
|
return -EINVAL;
|
|
}
|
|
if (!eng_grps->grp[egrp].is_enabled) {
|
|
dev_err(dev, "Error engine_group%d is not configured", egrp);
|
|
return -EINVAL;
|
|
}
|
|
mutex_lock(&eng_grps->lock);
|
|
ret = delete_engine_group(dev, &eng_grps->grp[egrp]);
|
|
mutex_unlock(&eng_grps->lock);
|
|
|
|
return ret;
|
|
|
|
err_print:
|
|
dev_err(dev, "%s\n", err_msg);
|
|
return -EINVAL;
|
|
}
|
|
|
|
static void get_engs_info(struct otx2_cpt_eng_grp_info *eng_grp, char *buf,
|
|
int size, int idx)
|
|
{
|
|
struct otx2_cpt_engs_rsvd *mirrored_engs = NULL;
|
|
struct otx2_cpt_engs_rsvd *engs;
|
|
int len, i;
|
|
|
|
buf[0] = '\0';
|
|
for (i = 0; i < OTX2_CPT_MAX_ETYPES_PER_GRP; i++) {
|
|
engs = &eng_grp->engs[i];
|
|
if (!engs->type)
|
|
continue;
|
|
if (idx != -1 && idx != i)
|
|
continue;
|
|
|
|
if (eng_grp->mirror.is_ena)
|
|
mirrored_engs = find_engines_by_type(
|
|
&eng_grp->g->grp[eng_grp->mirror.idx],
|
|
engs->type);
|
|
if (i > 0 && idx == -1) {
|
|
len = strlen(buf);
|
|
scnprintf(buf + len, size - len, ", ");
|
|
}
|
|
|
|
len = strlen(buf);
|
|
scnprintf(buf + len, size - len, "%d %s ",
|
|
mirrored_engs ? engs->count + mirrored_engs->count :
|
|
engs->count,
|
|
get_eng_type_str(engs->type));
|
|
if (mirrored_engs) {
|
|
len = strlen(buf);
|
|
scnprintf(buf + len, size - len,
|
|
"(%d shared with engine_group%d) ",
|
|
engs->count <= 0 ?
|
|
engs->count + mirrored_engs->count :
|
|
mirrored_engs->count,
|
|
eng_grp->mirror.idx);
|
|
}
|
|
}
|
|
}
|
|
|
|
void otx2_cpt_print_uc_dbg_info(struct otx2_cptpf_dev *cptpf)
|
|
{
|
|
struct otx2_cpt_eng_grps *eng_grps = &cptpf->eng_grps;
|
|
struct otx2_cpt_eng_grp_info *mirrored_grp;
|
|
char engs_info[2 * OTX2_CPT_NAME_LENGTH];
|
|
struct otx2_cpt_eng_grp_info *grp;
|
|
struct otx2_cpt_engs_rsvd *engs;
|
|
int i, j;
|
|
|
|
pr_debug("Engine groups global info");
|
|
pr_debug("max SE %d, max IE %d, max AE %d", eng_grps->avail.max_se_cnt,
|
|
eng_grps->avail.max_ie_cnt, eng_grps->avail.max_ae_cnt);
|
|
pr_debug("free SE %d", eng_grps->avail.se_cnt);
|
|
pr_debug("free IE %d", eng_grps->avail.ie_cnt);
|
|
pr_debug("free AE %d", eng_grps->avail.ae_cnt);
|
|
|
|
for (i = 0; i < OTX2_CPT_MAX_ENGINE_GROUPS; i++) {
|
|
grp = &eng_grps->grp[i];
|
|
pr_debug("engine_group%d, state %s", i,
|
|
grp->is_enabled ? "enabled" : "disabled");
|
|
if (grp->is_enabled) {
|
|
mirrored_grp = &eng_grps->grp[grp->mirror.idx];
|
|
pr_debug("Ucode0 filename %s, version %s",
|
|
grp->mirror.is_ena ?
|
|
mirrored_grp->ucode[0].filename :
|
|
grp->ucode[0].filename,
|
|
grp->mirror.is_ena ?
|
|
mirrored_grp->ucode[0].ver_str :
|
|
grp->ucode[0].ver_str);
|
|
if (is_2nd_ucode_used(grp))
|
|
pr_debug("Ucode1 filename %s, version %s",
|
|
grp->ucode[1].filename,
|
|
grp->ucode[1].ver_str);
|
|
}
|
|
|
|
for (j = 0; j < OTX2_CPT_MAX_ETYPES_PER_GRP; j++) {
|
|
engs = &grp->engs[j];
|
|
if (engs->type) {
|
|
u32 mask[5] = { };
|
|
|
|
get_engs_info(grp, engs_info,
|
|
2 * OTX2_CPT_NAME_LENGTH, j);
|
|
pr_debug("Slot%d: %s", j, engs_info);
|
|
bitmap_to_arr32(mask, engs->bmap,
|
|
eng_grps->engs_num);
|
|
if (is_dev_otx2(cptpf->pdev))
|
|
pr_debug("Mask: %8.8x %8.8x %8.8x %8.8x",
|
|
mask[3], mask[2], mask[1],
|
|
mask[0]);
|
|
else
|
|
pr_debug("Mask: %8.8x %8.8x %8.8x %8.8x %8.8x",
|
|
mask[4], mask[3], mask[2], mask[1],
|
|
mask[0]);
|
|
}
|
|
}
|
|
}
|
|
}
|