]> Git Repo - linux.git/blob - drivers/infiniband/hw/hns/hns_roce_restrack.c
Linux 6.14-rc3
[linux.git] / drivers / infiniband / hw / hns / hns_roce_restrack.c
1 // SPDX-License-Identifier: (GPL-2.0 OR BSD-2-Clause)
2 // Copyright (c) 2019 Hisilicon Limited.
3
4 #include <rdma/rdma_cm.h>
5 #include <rdma/restrack.h>
6 #include <uapi/rdma/rdma_netlink.h>
7 #include "hnae3.h"
8 #include "hns_roce_common.h"
9 #include "hns_roce_device.h"
10 #include "hns_roce_hw_v2.h"
11
12 int hns_roce_fill_res_cq_entry(struct sk_buff *msg, struct ib_cq *ib_cq)
13 {
14         struct hns_roce_cq *hr_cq = to_hr_cq(ib_cq);
15         struct nlattr *table_attr;
16
17         table_attr = nla_nest_start(msg, RDMA_NLDEV_ATTR_DRIVER);
18         if (!table_attr)
19                 return -EMSGSIZE;
20
21         if (rdma_nl_put_driver_u32(msg, "cq_depth", hr_cq->cq_depth))
22                 goto err;
23
24         if (rdma_nl_put_driver_u32(msg, "cons_index", hr_cq->cons_index))
25                 goto err;
26
27         if (rdma_nl_put_driver_u32(msg, "cqe_size", hr_cq->cqe_size))
28                 goto err;
29
30         if (rdma_nl_put_driver_u32(msg, "arm_sn", hr_cq->arm_sn))
31                 goto err;
32
33         nla_nest_end(msg, table_attr);
34
35         return 0;
36
37 err:
38         nla_nest_cancel(msg, table_attr);
39
40         return -EMSGSIZE;
41 }
42
43 int hns_roce_fill_res_cq_entry_raw(struct sk_buff *msg, struct ib_cq *ib_cq)
44 {
45         struct hns_roce_dev *hr_dev = to_hr_dev(ib_cq->device);
46         struct hns_roce_cq *hr_cq = to_hr_cq(ib_cq);
47         struct hns_roce_v2_cq_context context;
48         int ret;
49
50         if (!hr_dev->hw->query_cqc)
51                 return -EINVAL;
52
53         ret = hr_dev->hw->query_cqc(hr_dev, hr_cq->cqn, &context);
54         if (ret)
55                 return -EINVAL;
56
57         ret = nla_put(msg, RDMA_NLDEV_ATTR_RES_RAW, sizeof(context), &context);
58
59         return ret;
60 }
61
62 int hns_roce_fill_res_qp_entry(struct sk_buff *msg, struct ib_qp *ib_qp)
63 {
64         struct hns_roce_qp *hr_qp = to_hr_qp(ib_qp);
65         struct nlattr *table_attr;
66
67         table_attr = nla_nest_start(msg, RDMA_NLDEV_ATTR_DRIVER);
68         if (!table_attr)
69                 return -EMSGSIZE;
70
71         if (rdma_nl_put_driver_u32_hex(msg, "sq_wqe_cnt", hr_qp->sq.wqe_cnt))
72                 goto err;
73
74         if (rdma_nl_put_driver_u32_hex(msg, "sq_max_gs", hr_qp->sq.max_gs))
75                 goto err;
76
77         if (rdma_nl_put_driver_u32_hex(msg, "rq_wqe_cnt", hr_qp->rq.wqe_cnt))
78                 goto err;
79
80         if (rdma_nl_put_driver_u32_hex(msg, "rq_max_gs", hr_qp->rq.max_gs))
81                 goto err;
82
83         if (rdma_nl_put_driver_u32_hex(msg, "ext_sge_sge_cnt", hr_qp->sge.sge_cnt))
84                 goto err;
85
86         nla_nest_end(msg, table_attr);
87
88         return 0;
89
90 err:
91         nla_nest_cancel(msg, table_attr);
92
93         return -EMSGSIZE;
94 }
95
96 int hns_roce_fill_res_qp_entry_raw(struct sk_buff *msg, struct ib_qp *ib_qp)
97 {
98         struct hns_roce_dev *hr_dev = to_hr_dev(ib_qp->device);
99         struct hns_roce_qp *hr_qp = to_hr_qp(ib_qp);
100         struct hns_roce_full_qp_ctx {
101                 struct hns_roce_v2_qp_context qpc;
102                 struct hns_roce_v2_scc_context sccc;
103         } context = {};
104         int ret;
105
106         if (!hr_dev->hw->query_qpc)
107                 return -EINVAL;
108
109         ret = hr_dev->hw->query_qpc(hr_dev, hr_qp->qpn, &context.qpc);
110         if (ret)
111                 return ret;
112
113         /* If SCC is disabled or the query fails, the queried SCCC will
114          * be all 0.
115          */
116         if (!(hr_dev->caps.flags & HNS_ROCE_CAP_FLAG_QP_FLOW_CTRL) ||
117             !hr_dev->hw->query_sccc)
118                 goto out;
119
120         ret = hr_dev->hw->query_sccc(hr_dev, hr_qp->qpn, &context.sccc);
121         if (ret)
122                 ibdev_warn_ratelimited(&hr_dev->ib_dev,
123                                        "failed to query SCCC, ret = %d.\n",
124                                        ret);
125
126 out:
127         ret = nla_put(msg, RDMA_NLDEV_ATTR_RES_RAW, sizeof(context), &context);
128
129         return ret;
130 }
131
132 int hns_roce_fill_res_mr_entry(struct sk_buff *msg, struct ib_mr *ib_mr)
133 {
134         struct hns_roce_mr *hr_mr = to_hr_mr(ib_mr);
135         struct nlattr *table_attr;
136
137         table_attr = nla_nest_start(msg, RDMA_NLDEV_ATTR_DRIVER);
138         if (!table_attr)
139                 return -EMSGSIZE;
140
141         if (rdma_nl_put_driver_u32_hex(msg, "pbl_hop_num", hr_mr->pbl_hop_num))
142                 goto err;
143
144         if (rdma_nl_put_driver_u32_hex(msg, "ba_pg_shift",
145                                        hr_mr->pbl_mtr.hem_cfg.ba_pg_shift))
146                 goto err;
147
148         if (rdma_nl_put_driver_u32_hex(msg, "buf_pg_shift",
149                                        hr_mr->pbl_mtr.hem_cfg.buf_pg_shift))
150                 goto err;
151
152         nla_nest_end(msg, table_attr);
153
154         return 0;
155
156 err:
157         nla_nest_cancel(msg, table_attr);
158
159         return -EMSGSIZE;
160 }
161
162 int hns_roce_fill_res_mr_entry_raw(struct sk_buff *msg, struct ib_mr *ib_mr)
163 {
164         struct hns_roce_dev *hr_dev = to_hr_dev(ib_mr->device);
165         struct hns_roce_mr *hr_mr = to_hr_mr(ib_mr);
166         struct hns_roce_v2_mpt_entry context;
167         int ret;
168
169         if (!hr_dev->hw->query_mpt)
170                 return -EINVAL;
171
172         ret = hr_dev->hw->query_mpt(hr_dev, hr_mr->key, &context);
173         if (ret)
174                 return -EINVAL;
175
176         ret = nla_put(msg, RDMA_NLDEV_ATTR_RES_RAW, sizeof(context), &context);
177
178         return ret;
179 }
180
181 int hns_roce_fill_res_srq_entry(struct sk_buff *msg, struct ib_srq *ib_srq)
182 {
183         struct hns_roce_srq *hr_srq = to_hr_srq(ib_srq);
184         struct nlattr *table_attr;
185
186         table_attr = nla_nest_start(msg, RDMA_NLDEV_ATTR_DRIVER);
187         if (!table_attr)
188                 return -EMSGSIZE;
189
190         if (rdma_nl_put_driver_u32_hex(msg, "srqn", hr_srq->srqn))
191                 goto err;
192
193         if (rdma_nl_put_driver_u32_hex(msg, "wqe_cnt", hr_srq->wqe_cnt))
194                 goto err;
195
196         if (rdma_nl_put_driver_u32_hex(msg, "max_gs", hr_srq->max_gs))
197                 goto err;
198
199         if (rdma_nl_put_driver_u32_hex(msg, "xrcdn", hr_srq->xrcdn))
200                 goto err;
201
202         nla_nest_end(msg, table_attr);
203
204         return 0;
205
206 err:
207         nla_nest_cancel(msg, table_attr);
208         return -EMSGSIZE;
209 }
210
211 int hns_roce_fill_res_srq_entry_raw(struct sk_buff *msg, struct ib_srq *ib_srq)
212 {
213         struct hns_roce_dev *hr_dev = to_hr_dev(ib_srq->device);
214         struct hns_roce_srq *hr_srq = to_hr_srq(ib_srq);
215         struct hns_roce_srq_context context;
216         int ret;
217
218         if (!hr_dev->hw->query_srqc)
219                 return -EINVAL;
220
221         ret = hr_dev->hw->query_srqc(hr_dev, hr_srq->srqn, &context);
222         if (ret)
223                 return ret;
224
225         ret = nla_put(msg, RDMA_NLDEV_ATTR_RES_RAW, sizeof(context), &context);
226
227         return ret;
228 }
This page took 0.044352 seconds and 4 git commands to generate.