1 // SPDX-License-Identifier: GPL-2.0-only
2 /* Marvell RVU Admin Function driver
4 * Copyright (C) 2020 Marvell.
8 #include <linux/bitfield.h>
10 #include "rvu_struct.h"
15 /* CPT PF device id */
16 #define PCI_DEVID_OTX2_CPT_PF 0xA0FD
17 #define PCI_DEVID_OTX2_CPT10K_PF 0xA0F2
19 /* Length of initial context fetch in 128 byte words */
20 #define CPT_CTX_ILEN 1ULL
22 /* Interrupt vector count of CPT RVU and RAS interrupts */
23 #define CPT_10K_AF_RVU_RAS_INT_VEC_CNT 2
25 /* Default CPT_AF_RXC_CFG1:max_rxc_icb_cnt */
26 #define CPT_DFLT_MAX_RXC_ICB_CNT 0xC0ULL
28 #define cpt_get_eng_sts(e_min, e_max, rsp, etype) \
30 u64 free_sts = 0, busy_sts = 0; \
31 typeof(rsp) _rsp = rsp; \
34 for (e = (e_min), i = 0; e < (e_max); e++, i++) { \
35 reg = rvu_read64(rvu, blkaddr, CPT_AF_EXEX_STS(e)); \
37 busy_sts |= 1ULL << i; \
40 free_sts |= 1ULL << i; \
42 (_rsp)->busy_sts_##etype = busy_sts; \
43 (_rsp)->free_sts_##etype = free_sts; \
46 #define MAX_AE GENMASK_ULL(47, 32)
47 #define MAX_IE GENMASK_ULL(31, 16)
48 #define MAX_SE GENMASK_ULL(15, 0)
50 static u16 cpt_max_engines_get(struct rvu *rvu)
52 u16 max_ses, max_ies, max_aes;
55 reg = rvu_read64(rvu, BLKADDR_CPT0, CPT_AF_CONSTANTS1);
56 max_ses = FIELD_GET(MAX_SE, reg);
57 max_ies = FIELD_GET(MAX_IE, reg);
58 max_aes = FIELD_GET(MAX_AE, reg);
60 return max_ses + max_ies + max_aes;
63 /* Number of flt interrupt vectors are depends on number of engines that the
64 * chip has. Each flt vector represents 64 engines.
66 static int cpt_10k_flt_nvecs_get(struct rvu *rvu, u16 max_engs)
70 flt_vecs = DIV_ROUND_UP(max_engs, 64);
72 if (flt_vecs > CPT_10K_AF_INT_VEC_FLT_MAX) {
73 dev_warn_once(rvu->dev, "flt_vecs:%d exceeds the max vectors:%d\n",
74 flt_vecs, CPT_10K_AF_INT_VEC_FLT_MAX);
75 flt_vecs = CPT_10K_AF_INT_VEC_FLT_MAX;
81 static irqreturn_t cpt_af_flt_intr_handler(int vec, void *ptr)
83 struct rvu_block *block = ptr;
84 struct rvu *rvu = block->rvu;
85 int blkaddr = block->addr;
90 reg = rvu_read64(rvu, blkaddr, CPT_AF_FLTX_INT(vec));
91 dev_err_ratelimited(rvu->dev, "Received CPTAF FLT%d irq : 0x%llx", vec, reg);
94 while ((i = find_next_bit((unsigned long *)®, 64, i + 1)) < 64) {
106 grp = rvu_read64(rvu, blkaddr, CPT_AF_EXEX_CTL2(eng)) & 0xFF;
107 /* Disable and enable the engine which triggers fault */
108 rvu_write64(rvu, blkaddr, CPT_AF_EXEX_CTL2(eng), 0x0);
109 val = rvu_read64(rvu, blkaddr, CPT_AF_EXEX_CTL(eng));
110 rvu_write64(rvu, blkaddr, CPT_AF_EXEX_CTL(eng), val & ~1ULL);
112 rvu_write64(rvu, blkaddr, CPT_AF_EXEX_CTL2(eng), grp);
113 rvu_write64(rvu, blkaddr, CPT_AF_EXEX_CTL(eng), val | 1ULL);
115 spin_lock(&rvu->cpt_intr_lock);
116 block->cpt_flt_eng_map[vec] |= BIT_ULL(i);
117 val = rvu_read64(rvu, blkaddr, CPT_AF_EXEX_STS(eng));
119 if (val == 0x1 || val == 0x2)
120 block->cpt_rcvrd_eng_map[vec] |= BIT_ULL(i);
121 spin_unlock(&rvu->cpt_intr_lock);
123 rvu_write64(rvu, blkaddr, CPT_AF_FLTX_INT(vec), reg);
128 static irqreturn_t rvu_cpt_af_flt0_intr_handler(int irq, void *ptr)
130 return cpt_af_flt_intr_handler(CPT_AF_INT_VEC_FLT0, ptr);
133 static irqreturn_t rvu_cpt_af_flt1_intr_handler(int irq, void *ptr)
135 return cpt_af_flt_intr_handler(CPT_AF_INT_VEC_FLT1, ptr);
138 static irqreturn_t rvu_cpt_af_flt2_intr_handler(int irq, void *ptr)
140 return cpt_af_flt_intr_handler(CPT_10K_AF_INT_VEC_FLT2, ptr);
143 static irqreturn_t rvu_cpt_af_rvu_intr_handler(int irq, void *ptr)
145 struct rvu_block *block = ptr;
146 struct rvu *rvu = block->rvu;
147 int blkaddr = block->addr;
150 reg = rvu_read64(rvu, blkaddr, CPT_AF_RVU_INT);
151 dev_err_ratelimited(rvu->dev, "Received CPTAF RVU irq : 0x%llx", reg);
153 rvu_write64(rvu, blkaddr, CPT_AF_RVU_INT, reg);
157 static irqreturn_t rvu_cpt_af_ras_intr_handler(int irq, void *ptr)
159 struct rvu_block *block = ptr;
160 struct rvu *rvu = block->rvu;
161 int blkaddr = block->addr;
164 reg = rvu_read64(rvu, blkaddr, CPT_AF_RAS_INT);
165 dev_err_ratelimited(rvu->dev, "Received CPTAF RAS irq : 0x%llx", reg);
167 rvu_write64(rvu, blkaddr, CPT_AF_RAS_INT, reg);
171 static int rvu_cpt_do_register_interrupt(struct rvu_block *block, int irq_offs,
172 irq_handler_t handler,
175 struct rvu *rvu = block->rvu;
178 ret = request_irq(pci_irq_vector(rvu->pdev, irq_offs), handler, 0,
181 dev_err(rvu->dev, "RVUAF: %s irq registration failed", name);
185 WARN_ON(rvu->irq_allocated[irq_offs]);
186 rvu->irq_allocated[irq_offs] = true;
190 static void cpt_10k_unregister_interrupts(struct rvu_block *block, int off)
192 struct rvu *rvu = block->rvu;
193 int blkaddr = block->addr;
198 max_engs = cpt_max_engines_get(rvu);
199 flt_vecs = cpt_10k_flt_nvecs_get(rvu, max_engs);
201 /* Disable all CPT AF interrupts */
202 for (i = CPT_10K_AF_INT_VEC_FLT0; i < flt_vecs; i++) {
203 nr = (max_engs > 64) ? 64 : max_engs;
205 rvu_write64(rvu, blkaddr, CPT_AF_FLTX_INT_ENA_W1C(i),
209 rvu_write64(rvu, blkaddr, CPT_AF_RVU_INT_ENA_W1C, 0x1);
210 rvu_write64(rvu, blkaddr, CPT_AF_RAS_INT_ENA_W1C, 0x1);
212 /* CPT AF interrupt vectors are flt_int, rvu_int and ras_int. */
213 for (i = 0; i < flt_vecs + CPT_10K_AF_RVU_RAS_INT_VEC_CNT; i++)
214 if (rvu->irq_allocated[off + i]) {
215 free_irq(pci_irq_vector(rvu->pdev, off + i), block);
216 rvu->irq_allocated[off + i] = false;
220 static void cpt_unregister_interrupts(struct rvu *rvu, int blkaddr)
222 struct rvu_hwinfo *hw = rvu->hw;
223 struct rvu_block *block;
226 if (!is_block_implemented(rvu->hw, blkaddr))
228 offs = rvu_read64(rvu, blkaddr, CPT_PRIV_AF_INT_CFG) & 0x7FF;
231 "Failed to get CPT_AF_INT vector offsets\n");
234 block = &hw->block[blkaddr];
235 if (!is_rvu_otx2(rvu))
236 return cpt_10k_unregister_interrupts(block, offs);
238 /* Disable all CPT AF interrupts */
239 for (i = 0; i < CPT_AF_INT_VEC_RVU; i++)
240 rvu_write64(rvu, blkaddr, CPT_AF_FLTX_INT_ENA_W1C(i), ~0ULL);
241 rvu_write64(rvu, blkaddr, CPT_AF_RVU_INT_ENA_W1C, 0x1);
242 rvu_write64(rvu, blkaddr, CPT_AF_RAS_INT_ENA_W1C, 0x1);
244 for (i = 0; i < CPT_AF_INT_VEC_CNT; i++)
245 if (rvu->irq_allocated[offs + i]) {
246 free_irq(pci_irq_vector(rvu->pdev, offs + i), block);
247 rvu->irq_allocated[offs + i] = false;
251 void rvu_cpt_unregister_interrupts(struct rvu *rvu)
253 cpt_unregister_interrupts(rvu, BLKADDR_CPT0);
254 cpt_unregister_interrupts(rvu, BLKADDR_CPT1);
257 static int cpt_10k_register_interrupts(struct rvu_block *block, int off)
259 int rvu_intr_vec, ras_intr_vec;
260 struct rvu *rvu = block->rvu;
261 int blkaddr = block->addr;
262 irq_handler_t flt_fn;
263 int i, ret, flt_vecs;
267 max_engs = cpt_max_engines_get(rvu);
268 flt_vecs = cpt_10k_flt_nvecs_get(rvu, max_engs);
270 for (i = CPT_10K_AF_INT_VEC_FLT0; i < flt_vecs; i++) {
271 sprintf(&rvu->irq_name[(off + i) * NAME_SIZE], "CPTAF FLT%d", i);
274 case CPT_10K_AF_INT_VEC_FLT0:
275 flt_fn = rvu_cpt_af_flt0_intr_handler;
277 case CPT_10K_AF_INT_VEC_FLT1:
278 flt_fn = rvu_cpt_af_flt1_intr_handler;
280 case CPT_10K_AF_INT_VEC_FLT2:
281 flt_fn = rvu_cpt_af_flt2_intr_handler;
284 ret = rvu_cpt_do_register_interrupt(block, off + i,
285 flt_fn, &rvu->irq_name[(off + i) * NAME_SIZE]);
289 nr = (max_engs > 64) ? 64 : max_engs;
291 rvu_write64(rvu, blkaddr, CPT_AF_FLTX_INT_ENA_W1S(i),
295 rvu_intr_vec = flt_vecs;
296 ras_intr_vec = rvu_intr_vec + 1;
298 ret = rvu_cpt_do_register_interrupt(block, off + rvu_intr_vec,
299 rvu_cpt_af_rvu_intr_handler,
303 rvu_write64(rvu, blkaddr, CPT_AF_RVU_INT_ENA_W1S, 0x1);
305 ret = rvu_cpt_do_register_interrupt(block, off + ras_intr_vec,
306 rvu_cpt_af_ras_intr_handler,
310 rvu_write64(rvu, blkaddr, CPT_AF_RAS_INT_ENA_W1S, 0x1);
314 rvu_cpt_unregister_interrupts(rvu);
318 static int cpt_register_interrupts(struct rvu *rvu, int blkaddr)
320 struct rvu_hwinfo *hw = rvu->hw;
321 struct rvu_block *block;
322 irq_handler_t flt_fn;
323 int i, offs, ret = 0;
325 if (!is_block_implemented(rvu->hw, blkaddr))
328 block = &hw->block[blkaddr];
329 offs = rvu_read64(rvu, blkaddr, CPT_PRIV_AF_INT_CFG) & 0x7FF;
332 "Failed to get CPT_AF_INT vector offsets\n");
336 if (!is_rvu_otx2(rvu))
337 return cpt_10k_register_interrupts(block, offs);
339 for (i = CPT_AF_INT_VEC_FLT0; i < CPT_AF_INT_VEC_RVU; i++) {
340 sprintf(&rvu->irq_name[(offs + i) * NAME_SIZE], "CPTAF FLT%d", i);
342 case CPT_AF_INT_VEC_FLT0:
343 flt_fn = rvu_cpt_af_flt0_intr_handler;
345 case CPT_AF_INT_VEC_FLT1:
346 flt_fn = rvu_cpt_af_flt1_intr_handler;
349 ret = rvu_cpt_do_register_interrupt(block, offs + i,
350 flt_fn, &rvu->irq_name[(offs + i) * NAME_SIZE]);
353 rvu_write64(rvu, blkaddr, CPT_AF_FLTX_INT_ENA_W1S(i), ~0ULL);
356 ret = rvu_cpt_do_register_interrupt(block, offs + CPT_AF_INT_VEC_RVU,
357 rvu_cpt_af_rvu_intr_handler,
361 rvu_write64(rvu, blkaddr, CPT_AF_RVU_INT_ENA_W1S, 0x1);
363 ret = rvu_cpt_do_register_interrupt(block, offs + CPT_AF_INT_VEC_RAS,
364 rvu_cpt_af_ras_intr_handler,
368 rvu_write64(rvu, blkaddr, CPT_AF_RAS_INT_ENA_W1S, 0x1);
372 rvu_cpt_unregister_interrupts(rvu);
376 int rvu_cpt_register_interrupts(struct rvu *rvu)
380 ret = cpt_register_interrupts(rvu, BLKADDR_CPT0);
384 return cpt_register_interrupts(rvu, BLKADDR_CPT1);
387 static int get_cpt_pf_num(struct rvu *rvu)
389 int i, domain_nr, cpt_pf_num = -1;
390 struct pci_dev *pdev;
392 domain_nr = pci_domain_nr(rvu->pdev->bus);
393 for (i = 0; i < rvu->hw->total_pfs; i++) {
394 pdev = pci_get_domain_bus_and_slot(domain_nr, i + 1, 0);
398 if (pdev->device == PCI_DEVID_OTX2_CPT_PF ||
399 pdev->device == PCI_DEVID_OTX2_CPT10K_PF) {
401 put_device(&pdev->dev);
404 put_device(&pdev->dev);
409 static bool is_cpt_pf(struct rvu *rvu, u16 pcifunc)
411 int cpt_pf_num = rvu->cpt_pf_num;
413 if (rvu_get_pf(pcifunc) != cpt_pf_num)
415 if (pcifunc & RVU_PFVF_FUNC_MASK)
421 static bool is_cpt_vf(struct rvu *rvu, u16 pcifunc)
423 int cpt_pf_num = rvu->cpt_pf_num;
425 if (rvu_get_pf(pcifunc) != cpt_pf_num)
427 if (!(pcifunc & RVU_PFVF_FUNC_MASK))
433 static int validate_and_get_cpt_blkaddr(int req_blkaddr)
437 blkaddr = req_blkaddr ? req_blkaddr : BLKADDR_CPT0;
438 if (blkaddr != BLKADDR_CPT0 && blkaddr != BLKADDR_CPT1)
444 int rvu_mbox_handler_cpt_lf_alloc(struct rvu *rvu,
445 struct cpt_lf_alloc_req_msg *req,
448 u16 pcifunc = req->hdr.pcifunc;
449 struct rvu_block *block;
454 blkaddr = validate_and_get_cpt_blkaddr(req->blkaddr);
458 if (req->eng_grpmsk == 0x0)
459 return CPT_AF_ERR_GRP_INVALID;
461 block = &rvu->hw->block[blkaddr];
462 num_lfs = rvu_get_rsrc_mapcount(rvu_get_pfvf(rvu, pcifunc),
465 return CPT_AF_ERR_LF_INVALID;
467 /* Check if requested 'CPTLF <=> NIXLF' mapping is valid */
468 if (req->nix_pf_func) {
469 /* If default, use 'this' CPTLF's PFFUNC */
470 if (req->nix_pf_func == RVU_DEFAULT_PF_FUNC)
471 req->nix_pf_func = pcifunc;
472 if (!is_pffunc_map_valid(rvu, req->nix_pf_func, BLKTYPE_NIX))
473 return CPT_AF_ERR_NIX_PF_FUNC_INVALID;
476 /* Check if requested 'CPTLF <=> SSOLF' mapping is valid */
477 if (req->sso_pf_func) {
478 /* If default, use 'this' CPTLF's PFFUNC */
479 if (req->sso_pf_func == RVU_DEFAULT_PF_FUNC)
480 req->sso_pf_func = pcifunc;
481 if (!is_pffunc_map_valid(rvu, req->sso_pf_func, BLKTYPE_SSO))
482 return CPT_AF_ERR_SSO_PF_FUNC_INVALID;
485 for (slot = 0; slot < num_lfs; slot++) {
486 cptlf = rvu_get_lf(rvu, block, pcifunc, slot);
488 return CPT_AF_ERR_LF_INVALID;
490 /* Set CPT LF group and priority */
491 val = (u64)req->eng_grpmsk << 48 | 1;
492 if (!is_rvu_otx2(rvu)) {
493 if (req->ctx_ilen_valid)
494 val |= (req->ctx_ilen << 17);
496 val |= (CPT_CTX_ILEN << 17);
499 rvu_write64(rvu, blkaddr, CPT_AF_LFX_CTL(cptlf), val);
501 /* Set CPT LF NIX_PF_FUNC and SSO_PF_FUNC. EXE_LDWB is set
504 val = rvu_read64(rvu, blkaddr, CPT_AF_LFX_CTL2(cptlf));
505 val &= ~(GENMASK_ULL(63, 48) | GENMASK_ULL(47, 32));
506 val |= ((u64)req->nix_pf_func << 48 |
507 (u64)req->sso_pf_func << 32);
508 rvu_write64(rvu, blkaddr, CPT_AF_LFX_CTL2(cptlf), val);
514 static int cpt_lf_free(struct rvu *rvu, struct msg_req *req, int blkaddr)
516 u16 pcifunc = req->hdr.pcifunc;
517 int num_lfs, cptlf, slot, err;
518 struct rvu_block *block;
520 block = &rvu->hw->block[blkaddr];
521 num_lfs = rvu_get_rsrc_mapcount(rvu_get_pfvf(rvu, pcifunc),
526 for (slot = 0; slot < num_lfs; slot++) {
527 cptlf = rvu_get_lf(rvu, block, pcifunc, slot);
529 return CPT_AF_ERR_LF_INVALID;
531 /* Perform teardown */
532 rvu_cpt_lf_teardown(rvu, pcifunc, blkaddr, cptlf, slot);
535 err = rvu_lf_reset(rvu, block, cptlf);
537 dev_err(rvu->dev, "Failed to reset blkaddr %d LF%d\n",
545 int rvu_mbox_handler_cpt_lf_free(struct rvu *rvu, struct msg_req *req,
550 ret = cpt_lf_free(rvu, req, BLKADDR_CPT0);
554 if (is_block_implemented(rvu->hw, BLKADDR_CPT1))
555 ret = cpt_lf_free(rvu, req, BLKADDR_CPT1);
560 static int cpt_inline_ipsec_cfg_inbound(struct rvu *rvu, int blkaddr, u8 cptlf,
561 struct cpt_inline_ipsec_cfg_msg *req)
563 u16 sso_pf_func = req->sso_pf_func;
567 val = rvu_read64(rvu, blkaddr, CPT_AF_LFX_CTL(cptlf));
568 if (req->enable && (val & BIT_ULL(16))) {
569 /* IPSec inline outbound path is already enabled for a given
570 * CPT LF, HRM states that inline inbound & outbound paths
571 * must not be enabled at the same time for a given CPT LF
573 return CPT_AF_ERR_INLINE_IPSEC_INB_ENA;
575 /* Check if requested 'CPTLF <=> SSOLF' mapping is valid */
576 if (sso_pf_func && !is_pffunc_map_valid(rvu, sso_pf_func, BLKTYPE_SSO))
577 return CPT_AF_ERR_SSO_PF_FUNC_INVALID;
579 nix_sel = (blkaddr == BLKADDR_CPT1) ? 1 : 0;
580 /* Enable CPT LF for IPsec inline inbound operations */
586 val |= (u64)nix_sel << 8;
587 rvu_write64(rvu, blkaddr, CPT_AF_LFX_CTL(cptlf), val);
590 /* Set SSO_PF_FUNC */
591 val = rvu_read64(rvu, blkaddr, CPT_AF_LFX_CTL2(cptlf));
592 val |= (u64)sso_pf_func << 32;
593 val |= (u64)req->nix_pf_func << 48;
594 rvu_write64(rvu, blkaddr, CPT_AF_LFX_CTL2(cptlf), val);
596 if (req->sso_pf_func_ovrd)
597 /* Set SSO_PF_FUNC_OVRD for inline IPSec */
598 rvu_write64(rvu, blkaddr, CPT_AF_ECO, 0x1);
600 /* Configure the X2P Link register with the cpt base channel number and
601 * range of channels it should propagate to X2P
603 if (!is_rvu_otx2(rvu)) {
604 val = (ilog2(NIX_CHAN_CPT_X2P_MASK + 1) << 16);
605 val |= (u64)rvu->hw->cpt_chan_base;
607 rvu_write64(rvu, blkaddr, CPT_AF_X2PX_LINK_CFG(0), val);
608 rvu_write64(rvu, blkaddr, CPT_AF_X2PX_LINK_CFG(1), val);
614 static int cpt_inline_ipsec_cfg_outbound(struct rvu *rvu, int blkaddr, u8 cptlf,
615 struct cpt_inline_ipsec_cfg_msg *req)
617 u16 nix_pf_func = req->nix_pf_func;
622 val = rvu_read64(rvu, blkaddr, CPT_AF_LFX_CTL(cptlf));
623 if (req->enable && (val & BIT_ULL(9))) {
624 /* IPSec inline inbound path is already enabled for a given
625 * CPT LF, HRM states that inline inbound & outbound paths
626 * must not be enabled at the same time for a given CPT LF
628 return CPT_AF_ERR_INLINE_IPSEC_OUT_ENA;
631 /* Check if requested 'CPTLF <=> NIXLF' mapping is valid */
632 if (nix_pf_func && !is_pffunc_map_valid(rvu, nix_pf_func, BLKTYPE_NIX))
633 return CPT_AF_ERR_NIX_PF_FUNC_INVALID;
635 /* Enable CPT LF for IPsec inline outbound operations */
640 rvu_write64(rvu, blkaddr, CPT_AF_LFX_CTL(cptlf), val);
643 /* Set NIX_PF_FUNC */
644 val = rvu_read64(rvu, blkaddr, CPT_AF_LFX_CTL2(cptlf));
645 val |= (u64)nix_pf_func << 48;
646 rvu_write64(rvu, blkaddr, CPT_AF_LFX_CTL2(cptlf), val);
648 nix_blkaddr = rvu_get_blkaddr(rvu, BLKTYPE_NIX, nix_pf_func);
649 nix_sel = (nix_blkaddr == BLKADDR_NIX0) ? 0 : 1;
651 val = rvu_read64(rvu, blkaddr, CPT_AF_LFX_CTL(cptlf));
652 val |= (u64)nix_sel << 8;
653 rvu_write64(rvu, blkaddr, CPT_AF_LFX_CTL(cptlf), val);
659 int rvu_mbox_handler_cpt_inline_ipsec_cfg(struct rvu *rvu,
660 struct cpt_inline_ipsec_cfg_msg *req,
663 u16 pcifunc = req->hdr.pcifunc;
664 struct rvu_block *block;
665 int cptlf, blkaddr, ret;
668 blkaddr = rvu_get_blkaddr_from_slot(rvu, BLKTYPE_CPT, pcifunc,
669 req->slot, &actual_slot);
671 return CPT_AF_ERR_LF_INVALID;
673 block = &rvu->hw->block[blkaddr];
675 cptlf = rvu_get_lf(rvu, block, pcifunc, actual_slot);
677 return CPT_AF_ERR_LF_INVALID;
680 case CPT_INLINE_INBOUND:
681 ret = cpt_inline_ipsec_cfg_inbound(rvu, blkaddr, cptlf, req);
684 case CPT_INLINE_OUTBOUND:
685 ret = cpt_inline_ipsec_cfg_outbound(rvu, blkaddr, cptlf, req);
689 return CPT_AF_ERR_PARAM;
695 static bool validate_and_update_reg_offset(struct rvu *rvu,
696 struct cpt_rd_wr_reg_msg *req,
699 u64 offset = req->reg_offset;
700 int blkaddr, num_lfs, lf;
701 struct rvu_block *block;
702 struct rvu_pfvf *pfvf;
704 blkaddr = validate_and_get_cpt_blkaddr(req->blkaddr);
708 /* Registers that can be accessed from PF/VF */
709 if ((offset & 0xFF000) == CPT_AF_LFX_CTL(0) ||
710 (offset & 0xFF000) == CPT_AF_LFX_CTL2(0)) {
714 lf = (offset & 0xFFF) >> 3;
715 block = &rvu->hw->block[blkaddr];
716 pfvf = rvu_get_pfvf(rvu, req->hdr.pcifunc);
717 num_lfs = rvu_get_rsrc_mapcount(pfvf, block->addr);
719 /* Slot is not valid for that PF/VF */
722 /* Translate local LF used by VFs to global CPT LF */
723 lf = rvu_get_lf(rvu, &rvu->hw->block[blkaddr],
724 req->hdr.pcifunc, lf);
728 /* Translate local LF's offset to global CPT LF's offset to
729 * access LFX register.
731 *reg_offset = (req->reg_offset & 0xFF000) + (lf << 3);
734 } else if (!(req->hdr.pcifunc & RVU_PFVF_FUNC_MASK)) {
735 /* Registers that can be accessed from PF */
741 case CPT_AF_CONSTANTS1:
742 case CPT_AF_CTX_FLUSH_TIMER:
743 case CPT_AF_RXC_CFG1:
747 switch (offset & 0xFF000) {
748 case CPT_AF_EXEX_STS(0):
749 case CPT_AF_EXEX_CTL(0):
750 case CPT_AF_EXEX_CTL2(0):
751 case CPT_AF_EXEX_UCODE_BASE(0):
763 int rvu_mbox_handler_cpt_rd_wr_register(struct rvu *rvu,
764 struct cpt_rd_wr_reg_msg *req,
765 struct cpt_rd_wr_reg_msg *rsp)
767 u64 offset = req->reg_offset;
770 blkaddr = validate_and_get_cpt_blkaddr(req->blkaddr);
774 /* This message is accepted only if sent from CPT PF/VF */
775 if (!is_cpt_pf(rvu, req->hdr.pcifunc) &&
776 !is_cpt_vf(rvu, req->hdr.pcifunc))
777 return CPT_AF_ERR_ACCESS_DENIED;
779 if (!validate_and_update_reg_offset(rvu, req, &offset))
780 return CPT_AF_ERR_ACCESS_DENIED;
782 rsp->reg_offset = req->reg_offset;
783 rsp->ret_val = req->ret_val;
784 rsp->is_write = req->is_write;
787 rvu_write64(rvu, blkaddr, offset, req->val);
789 rsp->val = rvu_read64(rvu, blkaddr, offset);
794 static void get_ctx_pc(struct rvu *rvu, struct cpt_sts_rsp *rsp, int blkaddr)
796 struct rvu_hwinfo *hw = rvu->hw;
798 if (is_rvu_otx2(rvu))
801 rsp->ctx_mis_pc = rvu_read64(rvu, blkaddr, CPT_AF_CTX_MIS_PC);
802 rsp->ctx_hit_pc = rvu_read64(rvu, blkaddr, CPT_AF_CTX_HIT_PC);
803 rsp->ctx_aop_pc = rvu_read64(rvu, blkaddr, CPT_AF_CTX_AOP_PC);
804 rsp->ctx_aop_lat_pc = rvu_read64(rvu, blkaddr,
805 CPT_AF_CTX_AOP_LATENCY_PC);
806 rsp->ctx_ifetch_pc = rvu_read64(rvu, blkaddr, CPT_AF_CTX_IFETCH_PC);
807 rsp->ctx_ifetch_lat_pc = rvu_read64(rvu, blkaddr,
808 CPT_AF_CTX_IFETCH_LATENCY_PC);
809 rsp->ctx_ffetch_pc = rvu_read64(rvu, blkaddr, CPT_AF_CTX_FFETCH_PC);
810 rsp->ctx_ffetch_lat_pc = rvu_read64(rvu, blkaddr,
811 CPT_AF_CTX_FFETCH_LATENCY_PC);
812 rsp->ctx_wback_pc = rvu_read64(rvu, blkaddr, CPT_AF_CTX_FFETCH_PC);
813 rsp->ctx_wback_lat_pc = rvu_read64(rvu, blkaddr,
814 CPT_AF_CTX_FFETCH_LATENCY_PC);
815 rsp->ctx_psh_pc = rvu_read64(rvu, blkaddr, CPT_AF_CTX_FFETCH_PC);
816 rsp->ctx_psh_lat_pc = rvu_read64(rvu, blkaddr,
817 CPT_AF_CTX_FFETCH_LATENCY_PC);
818 rsp->ctx_err = rvu_read64(rvu, blkaddr, CPT_AF_CTX_ERR);
819 rsp->ctx_enc_id = rvu_read64(rvu, blkaddr, CPT_AF_CTX_ENC_ID);
820 rsp->ctx_flush_timer = rvu_read64(rvu, blkaddr, CPT_AF_CTX_FLUSH_TIMER);
821 rsp->x2p_link_cfg0 = rvu_read64(rvu, blkaddr, CPT_AF_X2PX_LINK_CFG(0));
822 rsp->x2p_link_cfg1 = rvu_read64(rvu, blkaddr, CPT_AF_X2PX_LINK_CFG(1));
824 if (!hw->cap.cpt_rxc)
826 rsp->rxc_time = rvu_read64(rvu, blkaddr, CPT_AF_RXC_TIME);
827 rsp->rxc_time_cfg = rvu_read64(rvu, blkaddr, CPT_AF_RXC_TIME_CFG);
828 rsp->rxc_active_sts = rvu_read64(rvu, blkaddr, CPT_AF_RXC_ACTIVE_STS);
829 rsp->rxc_zombie_sts = rvu_read64(rvu, blkaddr, CPT_AF_RXC_ZOMBIE_STS);
830 rsp->rxc_dfrg = rvu_read64(rvu, blkaddr, CPT_AF_RXC_DFRG);
833 static void get_eng_sts(struct rvu *rvu, struct cpt_sts_rsp *rsp, int blkaddr)
835 u16 max_ses, max_ies, max_aes;
836 u32 e_min = 0, e_max = 0;
839 reg = rvu_read64(rvu, blkaddr, CPT_AF_CONSTANTS1);
840 max_ses = reg & 0xffff;
841 max_ies = (reg >> 16) & 0xffff;
842 max_aes = (reg >> 32) & 0xffff;
845 e_min = max_ses + max_ies;
846 e_max = max_ses + max_ies + max_aes;
847 cpt_get_eng_sts(e_min, e_max, rsp, ae);
851 cpt_get_eng_sts(e_min, e_max, rsp, se);
854 e_max = max_ses + max_ies;
855 cpt_get_eng_sts(e_min, e_max, rsp, ie);
858 int rvu_mbox_handler_cpt_sts(struct rvu *rvu, struct cpt_sts_req *req,
859 struct cpt_sts_rsp *rsp)
863 blkaddr = validate_and_get_cpt_blkaddr(req->blkaddr);
867 /* This message is accepted only if sent from CPT PF/VF */
868 if (!is_cpt_pf(rvu, req->hdr.pcifunc) &&
869 !is_cpt_vf(rvu, req->hdr.pcifunc))
870 return CPT_AF_ERR_ACCESS_DENIED;
872 get_ctx_pc(rvu, rsp, blkaddr);
874 /* Get CPT engines status */
875 get_eng_sts(rvu, rsp, blkaddr);
877 /* Read CPT instruction PC registers */
878 rsp->inst_req_pc = rvu_read64(rvu, blkaddr, CPT_AF_INST_REQ_PC);
879 rsp->inst_lat_pc = rvu_read64(rvu, blkaddr, CPT_AF_INST_LATENCY_PC);
880 rsp->rd_req_pc = rvu_read64(rvu, blkaddr, CPT_AF_RD_REQ_PC);
881 rsp->rd_lat_pc = rvu_read64(rvu, blkaddr, CPT_AF_RD_LATENCY_PC);
882 rsp->rd_uc_pc = rvu_read64(rvu, blkaddr, CPT_AF_RD_UC_PC);
883 rsp->active_cycles_pc = rvu_read64(rvu, blkaddr,
884 CPT_AF_ACTIVE_CYCLES_PC);
885 rsp->exe_err_info = rvu_read64(rvu, blkaddr, CPT_AF_EXE_ERR_INFO);
886 rsp->cptclk_cnt = rvu_read64(rvu, blkaddr, CPT_AF_CPTCLK_CNT);
887 rsp->diag = rvu_read64(rvu, blkaddr, CPT_AF_DIAG);
892 #define RXC_ZOMBIE_THRES GENMASK_ULL(59, 48)
893 #define RXC_ZOMBIE_LIMIT GENMASK_ULL(43, 32)
894 #define RXC_ACTIVE_THRES GENMASK_ULL(27, 16)
895 #define RXC_ACTIVE_LIMIT GENMASK_ULL(11, 0)
896 #define RXC_ACTIVE_COUNT GENMASK_ULL(60, 48)
897 #define RXC_ZOMBIE_COUNT GENMASK_ULL(60, 48)
899 static void cpt_rxc_time_cfg(struct rvu *rvu, struct cpt_rxc_time_cfg_req *req,
900 int blkaddr, struct cpt_rxc_time_cfg_req *save)
905 /* Save older config */
906 dfrg_reg = rvu_read64(rvu, blkaddr, CPT_AF_RXC_DFRG);
907 save->zombie_thres = FIELD_GET(RXC_ZOMBIE_THRES, dfrg_reg);
908 save->zombie_limit = FIELD_GET(RXC_ZOMBIE_LIMIT, dfrg_reg);
909 save->active_thres = FIELD_GET(RXC_ACTIVE_THRES, dfrg_reg);
910 save->active_limit = FIELD_GET(RXC_ACTIVE_LIMIT, dfrg_reg);
912 save->step = rvu_read64(rvu, blkaddr, CPT_AF_RXC_TIME_CFG);
915 dfrg_reg = FIELD_PREP(RXC_ZOMBIE_THRES, req->zombie_thres);
916 dfrg_reg |= FIELD_PREP(RXC_ZOMBIE_LIMIT, req->zombie_limit);
917 dfrg_reg |= FIELD_PREP(RXC_ACTIVE_THRES, req->active_thres);
918 dfrg_reg |= FIELD_PREP(RXC_ACTIVE_LIMIT, req->active_limit);
920 rvu_write64(rvu, blkaddr, CPT_AF_RXC_TIME_CFG, req->step);
921 rvu_write64(rvu, blkaddr, CPT_AF_RXC_DFRG, dfrg_reg);
924 int rvu_mbox_handler_cpt_rxc_time_cfg(struct rvu *rvu,
925 struct cpt_rxc_time_cfg_req *req,
930 blkaddr = validate_and_get_cpt_blkaddr(req->blkaddr);
934 /* This message is accepted only if sent from CPT PF/VF */
935 if (!is_cpt_pf(rvu, req->hdr.pcifunc) &&
936 !is_cpt_vf(rvu, req->hdr.pcifunc))
937 return CPT_AF_ERR_ACCESS_DENIED;
939 cpt_rxc_time_cfg(rvu, req, blkaddr, NULL);
944 int rvu_mbox_handler_cpt_ctx_cache_sync(struct rvu *rvu, struct msg_req *req,
947 return rvu_cpt_ctx_flush(rvu, req->hdr.pcifunc);
950 int rvu_mbox_handler_cpt_lf_reset(struct rvu *rvu, struct cpt_lf_rst_req *req,
953 u16 pcifunc = req->hdr.pcifunc;
954 struct rvu_block *block;
955 int cptlf, blkaddr, ret;
959 blkaddr = rvu_get_blkaddr_from_slot(rvu, BLKTYPE_CPT, pcifunc,
960 req->slot, &actual_slot);
962 return CPT_AF_ERR_LF_INVALID;
964 block = &rvu->hw->block[blkaddr];
966 cptlf = rvu_get_lf(rvu, block, pcifunc, actual_slot);
968 return CPT_AF_ERR_LF_INVALID;
969 ctl = rvu_read64(rvu, blkaddr, CPT_AF_LFX_CTL(cptlf));
970 ctl2 = rvu_read64(rvu, blkaddr, CPT_AF_LFX_CTL2(cptlf));
972 ret = rvu_lf_reset(rvu, block, cptlf);
974 dev_err(rvu->dev, "Failed to reset blkaddr %d LF%d\n",
977 rvu_write64(rvu, blkaddr, CPT_AF_LFX_CTL(cptlf), ctl);
978 rvu_write64(rvu, blkaddr, CPT_AF_LFX_CTL2(cptlf), ctl2);
983 int rvu_mbox_handler_cpt_flt_eng_info(struct rvu *rvu, struct cpt_flt_eng_info_req *req,
984 struct cpt_flt_eng_info_rsp *rsp)
986 struct rvu_block *block;
992 blkaddr = validate_and_get_cpt_blkaddr(req->blkaddr);
996 block = &rvu->hw->block[blkaddr];
997 max_engs = cpt_max_engines_get(rvu);
998 flt_vecs = cpt_10k_flt_nvecs_get(rvu, max_engs);
999 for (vec = 0; vec < flt_vecs; vec++) {
1000 spin_lock_irqsave(&rvu->cpt_intr_lock, flags);
1001 rsp->flt_eng_map[vec] = block->cpt_flt_eng_map[vec];
1002 rsp->rcvrd_eng_map[vec] = block->cpt_rcvrd_eng_map[vec];
1004 block->cpt_flt_eng_map[vec] = 0x0;
1005 block->cpt_rcvrd_eng_map[vec] = 0x0;
1007 spin_unlock_irqrestore(&rvu->cpt_intr_lock, flags);
1012 static void cpt_rxc_teardown(struct rvu *rvu, int blkaddr)
1014 struct cpt_rxc_time_cfg_req req, prev;
1015 struct rvu_hwinfo *hw = rvu->hw;
1019 if (!hw->cap.cpt_rxc)
1022 /* Set time limit to minimum values, so that rxc entries will be
1023 * flushed out quickly.
1026 req.zombie_thres = 1;
1027 req.zombie_limit = 1;
1028 req.active_thres = 1;
1029 req.active_limit = 1;
1031 cpt_rxc_time_cfg(rvu, &req, blkaddr, &prev);
1034 reg = rvu_read64(rvu, blkaddr, CPT_AF_RXC_ACTIVE_STS);
1036 if (FIELD_GET(RXC_ACTIVE_COUNT, reg))
1043 dev_warn(rvu->dev, "Poll for RXC active count hits hard loop counter\n");
1047 reg = rvu_read64(rvu, blkaddr, CPT_AF_RXC_ZOMBIE_STS);
1049 if (FIELD_GET(RXC_ZOMBIE_COUNT, reg))
1056 dev_warn(rvu->dev, "Poll for RXC zombie count hits hard loop counter\n");
1058 /* Restore config */
1059 cpt_rxc_time_cfg(rvu, &prev, blkaddr, NULL);
1062 #define INFLIGHT GENMASK_ULL(8, 0)
1063 #define GRB_CNT GENMASK_ULL(39, 32)
1064 #define GWB_CNT GENMASK_ULL(47, 40)
1065 #define XQ_XOR GENMASK_ULL(63, 63)
1066 #define DQPTR GENMASK_ULL(19, 0)
1067 #define NQPTR GENMASK_ULL(51, 32)
1069 static void cpt_lf_disable_iqueue(struct rvu *rvu, int blkaddr, int slot)
1071 int timeout = 1000000;
1072 u64 inprog, inst_ptr;
1076 /* Disable instructions enqueuing */
1077 rvu_write64(rvu, blkaddr, CPT_AF_BAR2_ALIASX(slot, CPT_LF_CTL), 0x0);
1079 inprog = rvu_read64(rvu, blkaddr,
1080 CPT_AF_BAR2_ALIASX(slot, CPT_LF_INPROG));
1081 inprog |= BIT_ULL(16);
1082 rvu_write64(rvu, blkaddr,
1083 CPT_AF_BAR2_ALIASX(slot, CPT_LF_INPROG), inprog);
1085 qsize = rvu_read64(rvu, blkaddr,
1086 CPT_AF_BAR2_ALIASX(slot, CPT_LF_Q_SIZE)) & 0x7FFF;
1088 inst_ptr = rvu_read64(rvu, blkaddr,
1089 CPT_AF_BAR2_ALIASX(slot, CPT_LF_Q_INST_PTR));
1090 pending = (FIELD_GET(XQ_XOR, inst_ptr) * qsize * 40) +
1091 FIELD_GET(NQPTR, inst_ptr) -
1092 FIELD_GET(DQPTR, inst_ptr);
1095 } while ((pending != 0) && (timeout != 0));
1098 dev_warn(rvu->dev, "TIMEOUT: CPT poll on pending instructions\n");
1101 /* Wait for CPT queue to become execution-quiescent */
1103 inprog = rvu_read64(rvu, blkaddr,
1104 CPT_AF_BAR2_ALIASX(slot, CPT_LF_INPROG));
1106 if ((FIELD_GET(INFLIGHT, inprog) == 0) &&
1107 (FIELD_GET(GRB_CNT, inprog) == 0)) {
1113 } while ((timeout != 0) && (i < 10));
1116 dev_warn(rvu->dev, "TIMEOUT: CPT poll on inflight count\n");
1117 /* Wait for 2 us to flush all queue writes to memory */
1121 int rvu_cpt_lf_teardown(struct rvu *rvu, u16 pcifunc, int blkaddr, int lf, int slot)
1125 if (is_cpt_pf(rvu, pcifunc) || is_cpt_vf(rvu, pcifunc))
1126 cpt_rxc_teardown(rvu, blkaddr);
1128 mutex_lock(&rvu->alias_lock);
1129 /* Enable BAR2 ALIAS for this pcifunc. */
1130 reg = BIT_ULL(16) | pcifunc;
1131 rvu_bar2_sel_write64(rvu, blkaddr, CPT_AF_BAR2_SEL, reg);
1133 cpt_lf_disable_iqueue(rvu, blkaddr, slot);
1135 rvu_bar2_sel_write64(rvu, blkaddr, CPT_AF_BAR2_SEL, 0);
1136 mutex_unlock(&rvu->alias_lock);
1141 #define CPT_RES_LEN 16
1142 #define CPT_SE_IE_EGRP 1ULL
1144 static int cpt_inline_inb_lf_cmd_send(struct rvu *rvu, int blkaddr,
1147 int cpt_pf_num = rvu->cpt_pf_num;
1148 struct cpt_inst_lmtst_req *req;
1149 dma_addr_t res_daddr;
1156 res = kzalloc(CPT_RES_LEN, GFP_KERNEL);
1160 res_daddr = dma_map_single(rvu->dev, res, CPT_RES_LEN,
1162 if (dma_mapping_error(rvu->dev, res_daddr)) {
1163 dev_err(rvu->dev, "DMA mapping failed for CPT result\n");
1169 /* Send mbox message to CPT PF */
1170 req = (struct cpt_inst_lmtst_req *)
1171 otx2_mbox_alloc_msg_rsp(&rvu->afpf_wq_info.mbox_up,
1172 cpt_pf_num, sizeof(*req),
1173 sizeof(struct msg_rsp));
1176 goto res_daddr_unmap;
1178 req->hdr.sig = OTX2_MBOX_REQ_SIG;
1179 req->hdr.id = MBOX_MSG_CPT_INST_LMTST;
1182 /* Prepare CPT_INST_S */
1184 inst[1] = res_daddr;
1193 inst[7] = CPT_SE_IE_EGRP << 61;
1195 /* Subtract 1 from the NIX-CPT credit count to preserve
1198 cpt_idx = (blkaddr == BLKADDR_CPT0) ? 0 : 1;
1199 rvu_write64(rvu, nix_blkaddr, NIX_AF_RX_CPTX_CREDIT(cpt_idx),
1202 otx2_mbox_msg_send(&rvu->afpf_wq_info.mbox_up, cpt_pf_num);
1203 rc = otx2_mbox_wait_for_rsp(&rvu->afpf_wq_info.mbox_up, cpt_pf_num);
1205 dev_warn(rvu->dev, "notification to pf %d failed\n",
1207 /* Wait for CPT instruction to be completed */
1217 dev_warn(rvu->dev, "Poll for result hits hard loop counter\n");
1220 dma_unmap_single(rvu->dev, res_daddr, CPT_RES_LEN, DMA_BIDIRECTIONAL);
1227 #define CTX_CAM_PF_FUNC GENMASK_ULL(61, 46)
1228 #define CTX_CAM_CPTR GENMASK_ULL(45, 0)
1230 int rvu_cpt_ctx_flush(struct rvu *rvu, u16 pcifunc)
1232 int nix_blkaddr, blkaddr;
1233 u16 max_ctx_entries, i;
1234 int slot = 0, num_lfs;
1238 nix_blkaddr = rvu_get_blkaddr(rvu, BLKTYPE_NIX, pcifunc);
1239 if (nix_blkaddr < 0)
1242 if (is_rvu_otx2(rvu))
1245 blkaddr = (nix_blkaddr == BLKADDR_NIX1) ? BLKADDR_CPT1 : BLKADDR_CPT0;
1247 /* Submit CPT_INST_S to track when all packets have been
1248 * flushed through for the NIX PF FUNC in inline inbound case.
1250 rc = cpt_inline_inb_lf_cmd_send(rvu, blkaddr, nix_blkaddr);
1254 /* Wait for rxc entries to be flushed out */
1255 cpt_rxc_teardown(rvu, blkaddr);
1257 reg = rvu_read64(rvu, blkaddr, CPT_AF_CONSTANTS0);
1258 max_ctx_entries = (reg >> 48) & 0xFFF;
1260 mutex_lock(&rvu->rsrc_lock);
1262 num_lfs = rvu_get_rsrc_mapcount(rvu_get_pfvf(rvu, pcifunc),
1265 dev_warn(rvu->dev, "CPT LF is not configured\n");
1269 /* Enable BAR2 ALIAS for this pcifunc. */
1270 reg = BIT_ULL(16) | pcifunc;
1271 rvu_bar2_sel_write64(rvu, blkaddr, CPT_AF_BAR2_SEL, reg);
1273 for (i = 0; i < max_ctx_entries; i++) {
1274 cam_data = rvu_read64(rvu, blkaddr, CPT_AF_CTX_CAM_DATA(i));
1276 if ((FIELD_GET(CTX_CAM_PF_FUNC, cam_data) == pcifunc) &&
1277 FIELD_GET(CTX_CAM_CPTR, cam_data)) {
1278 reg = BIT_ULL(46) | FIELD_GET(CTX_CAM_CPTR, cam_data);
1279 rvu_write64(rvu, blkaddr,
1280 CPT_AF_BAR2_ALIASX(slot, CPT_LF_CTX_FLUSH),
1284 rvu_bar2_sel_write64(rvu, blkaddr, CPT_AF_BAR2_SEL, 0);
1287 mutex_unlock(&rvu->rsrc_lock);
1292 #define MAX_RXC_ICB_CNT GENMASK_ULL(40, 32)
1294 int rvu_cpt_init(struct rvu *rvu)
1296 struct rvu_hwinfo *hw = rvu->hw;
1299 /* Retrieve CPT PF number */
1300 rvu->cpt_pf_num = get_cpt_pf_num(rvu);
1301 if (is_block_implemented(rvu->hw, BLKADDR_CPT0) && !is_rvu_otx2(rvu) &&
1303 hw->cap.cpt_rxc = true;
1305 if (hw->cap.cpt_rxc && !is_cn10ka_a0(rvu) && !is_cn10ka_a1(rvu)) {
1306 /* Set CPT_AF_RXC_CFG1:max_rxc_icb_cnt to 0xc0 to not effect
1307 * inline inbound peak performance
1309 reg_val = rvu_read64(rvu, BLKADDR_CPT0, CPT_AF_RXC_CFG1);
1310 reg_val &= ~MAX_RXC_ICB_CNT;
1311 reg_val |= FIELD_PREP(MAX_RXC_ICB_CNT,
1312 CPT_DFLT_MAX_RXC_ICB_CNT);
1313 rvu_write64(rvu, BLKADDR_CPT0, CPT_AF_RXC_CFG1, reg_val);
1316 spin_lock_init(&rvu->cpt_intr_lock);