1 // SPDX-License-Identifier: GPL-2.0 OR BSD-3-Clause
5 /* Copyright (c) 2020-2022, Alibaba Group. */
7 #include "erdma_verbs.h"
9 #define MAX_POLL_CHUNK_SIZE 16
11 void notify_eq(struct erdma_eq *eq)
13 u64 db_data = FIELD_PREP(ERDMA_EQDB_CI_MASK, eq->ci) |
14 FIELD_PREP(ERDMA_EQDB_ARM_MASK, 1);
16 *eq->db_record = db_data;
17 writeq(db_data, eq->db);
19 atomic64_inc(&eq->notify_num);
22 void *get_next_valid_eqe(struct erdma_eq *eq)
24 u64 *eqe = get_queue_entry(eq->qbuf, eq->ci, eq->depth, EQE_SHIFT);
25 u32 owner = FIELD_GET(ERDMA_CEQE_HDR_O_MASK, READ_ONCE(*eqe));
27 return owner ^ !!(eq->ci & eq->depth) ? eqe : NULL;
30 void erdma_aeq_event_handler(struct erdma_dev *dev)
32 struct erdma_aeqe *aeqe;
36 struct ib_event event;
39 memset(&event, 0, sizeof(event));
41 while (poll_cnt < MAX_POLL_CHUNK_SIZE) {
42 aeqe = get_next_valid_eqe(&dev->aeq);
49 atomic64_inc(&dev->aeq.event_num);
52 if (FIELD_GET(ERDMA_AEQE_HDR_TYPE_MASK,
53 le32_to_cpu(aeqe->hdr)) == ERDMA_AE_TYPE_CQ_ERR) {
54 cqn = le32_to_cpu(aeqe->event_data0);
55 cq = find_cq_by_cqn(dev, cqn);
59 event.device = cq->ibcq.device;
60 event.element.cq = &cq->ibcq;
61 event.event = IB_EVENT_CQ_ERR;
62 if (cq->ibcq.event_handler)
63 cq->ibcq.event_handler(&event,
66 qpn = le32_to_cpu(aeqe->event_data0);
67 qp = find_qp_by_qpn(dev, qpn);
71 event.device = qp->ibqp.device;
72 event.element.qp = &qp->ibqp;
73 event.event = IB_EVENT_QP_FATAL;
74 if (qp->ibqp.event_handler)
75 qp->ibqp.event_handler(&event,
83 int erdma_aeq_init(struct erdma_dev *dev)
85 struct erdma_eq *eq = &dev->aeq;
88 eq->depth = ERDMA_DEFAULT_EQ_DEPTH;
89 buf_size = eq->depth << EQE_SHIFT;
92 dma_alloc_coherent(&dev->pdev->dev, WARPPED_BUFSIZE(buf_size),
93 &eq->qbuf_dma_addr, GFP_KERNEL | __GFP_ZERO);
97 spin_lock_init(&eq->lock);
98 atomic64_set(&eq->event_num, 0);
99 atomic64_set(&eq->notify_num, 0);
101 eq->db = dev->func_bar + ERDMA_REGS_AEQ_DB_REG;
102 eq->db_record = (u64 *)(eq->qbuf + buf_size);
104 erdma_reg_write32(dev, ERDMA_REGS_AEQ_ADDR_H_REG,
105 upper_32_bits(eq->qbuf_dma_addr));
106 erdma_reg_write32(dev, ERDMA_REGS_AEQ_ADDR_L_REG,
107 lower_32_bits(eq->qbuf_dma_addr));
108 erdma_reg_write32(dev, ERDMA_REGS_AEQ_DEPTH_REG, eq->depth);
109 erdma_reg_write64(dev, ERDMA_AEQ_DB_HOST_ADDR_REG,
110 eq->qbuf_dma_addr + buf_size);
115 void erdma_aeq_destroy(struct erdma_dev *dev)
117 struct erdma_eq *eq = &dev->aeq;
119 dma_free_coherent(&dev->pdev->dev,
120 WARPPED_BUFSIZE(eq->depth << EQE_SHIFT), eq->qbuf,
124 void erdma_ceq_completion_handler(struct erdma_eq_cb *ceq_cb)
126 struct erdma_dev *dev = ceq_cb->dev;
135 while (poll_cnt < MAX_POLL_CHUNK_SIZE) {
136 ceqe = get_next_valid_eqe(&ceq_cb->eq);
143 cqn = FIELD_GET(ERDMA_CEQE_HDR_CQN_MASK, READ_ONCE(*ceqe));
145 cq = find_cq_by_cqn(dev, cqn);
149 if (rdma_is_kernel_res(&cq->ibcq.res))
152 if (cq->ibcq.comp_handler)
153 cq->ibcq.comp_handler(&cq->ibcq, cq->ibcq.cq_context);
156 notify_eq(&ceq_cb->eq);
159 static irqreturn_t erdma_intr_ceq_handler(int irq, void *data)
161 struct erdma_eq_cb *ceq_cb = data;
163 tasklet_schedule(&ceq_cb->tasklet);
168 static void erdma_intr_ceq_task(unsigned long data)
170 erdma_ceq_completion_handler((struct erdma_eq_cb *)data);
173 static int erdma_set_ceq_irq(struct erdma_dev *dev, u16 ceqn)
175 struct erdma_eq_cb *eqc = &dev->ceqs[ceqn];
178 snprintf(eqc->irq.name, ERDMA_IRQNAME_SIZE, "erdma-ceq%u@pci:%s", ceqn,
179 pci_name(dev->pdev));
180 eqc->irq.msix_vector = pci_irq_vector(dev->pdev, ceqn + 1);
182 tasklet_init(&dev->ceqs[ceqn].tasklet, erdma_intr_ceq_task,
183 (unsigned long)&dev->ceqs[ceqn]);
185 cpumask_set_cpu(cpumask_local_spread(ceqn + 1, dev->attrs.numa_node),
186 &eqc->irq.affinity_hint_mask);
188 err = request_irq(eqc->irq.msix_vector, erdma_intr_ceq_handler, 0,
191 dev_err(&dev->pdev->dev, "failed to request_irq(%d)\n", err);
195 irq_set_affinity_hint(eqc->irq.msix_vector,
196 &eqc->irq.affinity_hint_mask);
201 static void erdma_free_ceq_irq(struct erdma_dev *dev, u16 ceqn)
203 struct erdma_eq_cb *eqc = &dev->ceqs[ceqn];
205 irq_set_affinity_hint(eqc->irq.msix_vector, NULL);
206 free_irq(eqc->irq.msix_vector, eqc);
209 static int create_eq_cmd(struct erdma_dev *dev, u32 eqn, struct erdma_eq *eq)
211 struct erdma_cmdq_create_eq_req req;
212 dma_addr_t db_info_dma_addr;
214 erdma_cmdq_build_reqhdr(&req.hdr, CMDQ_SUBMOD_COMMON,
215 CMDQ_OPCODE_CREATE_EQ);
217 req.depth = ilog2(eq->depth);
218 req.qbuf_addr = eq->qbuf_dma_addr;
219 req.qtype = ERDMA_EQ_TYPE_CEQ;
220 /* Vector index is the same as EQN. */
221 req.vector_idx = eqn;
222 db_info_dma_addr = eq->qbuf_dma_addr + (eq->depth << EQE_SHIFT);
223 req.db_dma_addr_l = lower_32_bits(db_info_dma_addr);
224 req.db_dma_addr_h = upper_32_bits(db_info_dma_addr);
226 return erdma_post_cmd_wait(&dev->cmdq, &req, sizeof(req), NULL, NULL);
229 static int erdma_ceq_init_one(struct erdma_dev *dev, u16 ceqn)
231 struct erdma_eq *eq = &dev->ceqs[ceqn].eq;
232 u32 buf_size = ERDMA_DEFAULT_EQ_DEPTH << EQE_SHIFT;
236 dma_alloc_coherent(&dev->pdev->dev, WARPPED_BUFSIZE(buf_size),
237 &eq->qbuf_dma_addr, GFP_KERNEL | __GFP_ZERO);
241 spin_lock_init(&eq->lock);
242 atomic64_set(&eq->event_num, 0);
243 atomic64_set(&eq->notify_num, 0);
245 eq->depth = ERDMA_DEFAULT_EQ_DEPTH;
246 eq->db = dev->func_bar + ERDMA_REGS_CEQ_DB_BASE_REG +
247 (ceqn + 1) * ERDMA_DB_SIZE;
248 eq->db_record = (u64 *)(eq->qbuf + buf_size);
250 dev->ceqs[ceqn].dev = dev;
252 /* CEQ indexed from 1, 0 rsvd for CMDQ-EQ. */
253 ret = create_eq_cmd(dev, ceqn + 1, eq);
254 dev->ceqs[ceqn].ready = ret ? false : true;
259 static void erdma_ceq_uninit_one(struct erdma_dev *dev, u16 ceqn)
261 struct erdma_eq *eq = &dev->ceqs[ceqn].eq;
262 u32 buf_size = ERDMA_DEFAULT_EQ_DEPTH << EQE_SHIFT;
263 struct erdma_cmdq_destroy_eq_req req;
266 dev->ceqs[ceqn].ready = 0;
268 erdma_cmdq_build_reqhdr(&req.hdr, CMDQ_SUBMOD_COMMON,
269 CMDQ_OPCODE_DESTROY_EQ);
270 /* CEQ indexed from 1, 0 rsvd for CMDQ-EQ. */
272 req.qtype = ERDMA_EQ_TYPE_CEQ;
273 req.vector_idx = ceqn + 1;
275 err = erdma_post_cmd_wait(&dev->cmdq, &req, sizeof(req), NULL, NULL);
279 dma_free_coherent(&dev->pdev->dev, WARPPED_BUFSIZE(buf_size), eq->qbuf,
283 int erdma_ceqs_init(struct erdma_dev *dev)
288 for (i = 0; i < dev->attrs.irq_num - 1; i++) {
289 err = erdma_ceq_init_one(dev, i);
293 err = erdma_set_ceq_irq(dev, i);
295 erdma_ceq_uninit_one(dev, i);
303 for (j = 0; j < i; j++) {
304 erdma_free_ceq_irq(dev, j);
305 erdma_ceq_uninit_one(dev, j);
311 void erdma_ceqs_uninit(struct erdma_dev *dev)
315 for (i = 0; i < dev->attrs.irq_num - 1; i++) {
316 erdma_free_ceq_irq(dev, i);
317 erdma_ceq_uninit_one(dev, i);