1 // SPDX-License-Identifier: GPL-2.0-or-later
3 * linux/drivers/net/ethernet/ibm/ehea/ehea_phyp.c
5 * eHEA ethernet device driver for IBM eServer System p
7 * (C) Copyright IBM Corp. 2006
15 #define pr_fmt(fmt) KBUILD_MODNAME ": " fmt
17 #include "ehea_phyp.h"
20 static inline u16 get_order_of_qentries(u16 queue_entries)
22 u8 ld = 1; /* logarithmus dualis */
23 while (((1U << ld) - 1) < queue_entries)
28 /* Defines for H_CALL H_ALLOC_RESOURCE */
29 #define H_ALL_RES_TYPE_QP 1
30 #define H_ALL_RES_TYPE_CQ 2
31 #define H_ALL_RES_TYPE_EQ 3
32 #define H_ALL_RES_TYPE_MR 5
33 #define H_ALL_RES_TYPE_MW 6
35 static long ehea_plpar_hcall_norets(unsigned long opcode,
47 for (i = 0; i < 5; i++) {
48 ret = plpar_hcall_norets(opcode, arg1, arg2, arg3, arg4,
51 if (H_IS_LONG_BUSY(ret)) {
52 sleep_msecs = get_longbusy_msecs(ret);
53 msleep_interruptible(sleep_msecs);
58 pr_err("opcode=%lx ret=%lx"
59 " arg1=%lx arg2=%lx arg3=%lx arg4=%lx"
60 " arg5=%lx arg6=%lx arg7=%lx\n",
62 arg1, arg2, arg3, arg4, arg5, arg6, arg7);
70 static long ehea_plpar_hcall9(unsigned long opcode,
71 unsigned long *outs, /* array of 9 outputs */
86 for (i = 0; i < 5; i++) {
87 ret = plpar_hcall9(opcode, outs,
88 arg1, arg2, arg3, arg4, arg5,
89 arg6, arg7, arg8, arg9);
91 if (H_IS_LONG_BUSY(ret)) {
92 sleep_msecs = get_longbusy_msecs(ret);
93 msleep_interruptible(sleep_msecs);
97 cb_cat = EHEA_BMASK_GET(H_MEHEAPORT_CAT, arg2);
99 if ((ret < H_SUCCESS) && !(((ret == H_AUTHORITY)
100 && (opcode == H_MODIFY_HEA_PORT))
101 && (((cb_cat == H_PORT_CB4) && ((arg3 == H_PORT_CB4_JUMBO)
102 || (arg3 == H_PORT_CB4_SPEED))) || ((cb_cat == H_PORT_CB7)
103 && (arg3 == H_PORT_CB7_DUCQPN)))))
104 pr_err("opcode=%lx ret=%lx"
105 " arg1=%lx arg2=%lx arg3=%lx arg4=%lx"
106 " arg5=%lx arg6=%lx arg7=%lx arg8=%lx"
108 " out1=%lx out2=%lx out3=%lx out4=%lx"
109 " out5=%lx out6=%lx out7=%lx out8=%lx"
112 arg1, arg2, arg3, arg4, arg5,
113 arg6, arg7, arg8, arg9,
114 outs[0], outs[1], outs[2], outs[3], outs[4],
115 outs[5], outs[6], outs[7], outs[8]);
122 u64 ehea_h_query_ehea_qp(const u64 adapter_handle, const u8 qp_category,
123 const u64 qp_handle, const u64 sel_mask, void *cb_addr)
125 return ehea_plpar_hcall_norets(H_QUERY_HEA_QP,
126 adapter_handle, /* R4 */
127 qp_category, /* R5 */
130 __pa(cb_addr), /* R8 */
135 #define H_ALL_RES_QP_EQPO EHEA_BMASK_IBM(9, 11)
136 #define H_ALL_RES_QP_QPP EHEA_BMASK_IBM(12, 12)
137 #define H_ALL_RES_QP_RQR EHEA_BMASK_IBM(13, 15)
138 #define H_ALL_RES_QP_EQEG EHEA_BMASK_IBM(16, 16)
139 #define H_ALL_RES_QP_LL_QP EHEA_BMASK_IBM(17, 17)
140 #define H_ALL_RES_QP_DMA128 EHEA_BMASK_IBM(19, 19)
141 #define H_ALL_RES_QP_HSM EHEA_BMASK_IBM(20, 21)
142 #define H_ALL_RES_QP_SIGT EHEA_BMASK_IBM(22, 23)
143 #define H_ALL_RES_QP_TENURE EHEA_BMASK_IBM(48, 55)
144 #define H_ALL_RES_QP_RES_TYP EHEA_BMASK_IBM(56, 63)
147 #define H_ALL_RES_QP_TOKEN EHEA_BMASK_IBM(0, 31)
148 #define H_ALL_RES_QP_PD EHEA_BMASK_IBM(32, 63)
150 /* input param R10 */
151 #define H_ALL_RES_QP_MAX_SWQE EHEA_BMASK_IBM(4, 7)
152 #define H_ALL_RES_QP_MAX_R1WQE EHEA_BMASK_IBM(12, 15)
153 #define H_ALL_RES_QP_MAX_R2WQE EHEA_BMASK_IBM(20, 23)
154 #define H_ALL_RES_QP_MAX_R3WQE EHEA_BMASK_IBM(28, 31)
155 /* Max Send Scatter Gather Elements */
156 #define H_ALL_RES_QP_MAX_SSGE EHEA_BMASK_IBM(37, 39)
157 #define H_ALL_RES_QP_MAX_R1SGE EHEA_BMASK_IBM(45, 47)
158 /* Max Receive SG Elements RQ1 */
159 #define H_ALL_RES_QP_MAX_R2SGE EHEA_BMASK_IBM(53, 55)
160 #define H_ALL_RES_QP_MAX_R3SGE EHEA_BMASK_IBM(61, 63)
162 /* input param R11 */
163 #define H_ALL_RES_QP_SWQE_IDL EHEA_BMASK_IBM(0, 7)
164 /* max swqe immediate data length */
165 #define H_ALL_RES_QP_PORT_NUM EHEA_BMASK_IBM(48, 63)
167 /* input param R12 */
168 #define H_ALL_RES_QP_TH_RQ2 EHEA_BMASK_IBM(0, 15)
170 #define H_ALL_RES_QP_TH_RQ3 EHEA_BMASK_IBM(16, 31)
173 /* output param R6 */
174 #define H_ALL_RES_QP_ACT_SWQE EHEA_BMASK_IBM(0, 15)
175 #define H_ALL_RES_QP_ACT_R1WQE EHEA_BMASK_IBM(16, 31)
176 #define H_ALL_RES_QP_ACT_R2WQE EHEA_BMASK_IBM(32, 47)
177 #define H_ALL_RES_QP_ACT_R3WQE EHEA_BMASK_IBM(48, 63)
179 /* output param, R7 */
180 #define H_ALL_RES_QP_ACT_SSGE EHEA_BMASK_IBM(0, 7)
181 #define H_ALL_RES_QP_ACT_R1SGE EHEA_BMASK_IBM(8, 15)
182 #define H_ALL_RES_QP_ACT_R2SGE EHEA_BMASK_IBM(16, 23)
183 #define H_ALL_RES_QP_ACT_R3SGE EHEA_BMASK_IBM(24, 31)
184 #define H_ALL_RES_QP_ACT_SWQE_IDL EHEA_BMASK_IBM(32, 39)
186 /* output param R8,R9 */
187 #define H_ALL_RES_QP_SIZE_SQ EHEA_BMASK_IBM(0, 31)
188 #define H_ALL_RES_QP_SIZE_RQ1 EHEA_BMASK_IBM(32, 63)
189 #define H_ALL_RES_QP_SIZE_RQ2 EHEA_BMASK_IBM(0, 31)
190 #define H_ALL_RES_QP_SIZE_RQ3 EHEA_BMASK_IBM(32, 63)
192 /* output param R11,R12 */
193 #define H_ALL_RES_QP_LIOBN_SQ EHEA_BMASK_IBM(0, 31)
194 #define H_ALL_RES_QP_LIOBN_RQ1 EHEA_BMASK_IBM(32, 63)
195 #define H_ALL_RES_QP_LIOBN_RQ2 EHEA_BMASK_IBM(0, 31)
196 #define H_ALL_RES_QP_LIOBN_RQ3 EHEA_BMASK_IBM(32, 63)
198 u64 ehea_h_alloc_resource_qp(const u64 adapter_handle,
199 struct ehea_qp_init_attr *init_attr, const u32 pd,
200 u64 *qp_handle, struct h_epas *h_epas)
203 unsigned long outs[PLPAR_HCALL9_BUFSIZE];
205 u64 allocate_controls =
206 EHEA_BMASK_SET(H_ALL_RES_QP_EQPO, init_attr->low_lat_rq1 ? 1 : 0)
207 | EHEA_BMASK_SET(H_ALL_RES_QP_QPP, 0)
208 | EHEA_BMASK_SET(H_ALL_RES_QP_RQR, 6) /* rq1 & rq2 & rq3 */
209 | EHEA_BMASK_SET(H_ALL_RES_QP_EQEG, 0) /* EQE gen. disabled */
210 | EHEA_BMASK_SET(H_ALL_RES_QP_LL_QP, init_attr->low_lat_rq1)
211 | EHEA_BMASK_SET(H_ALL_RES_QP_DMA128, 0)
212 | EHEA_BMASK_SET(H_ALL_RES_QP_HSM, 0)
213 | EHEA_BMASK_SET(H_ALL_RES_QP_SIGT, init_attr->signalingtype)
214 | EHEA_BMASK_SET(H_ALL_RES_QP_RES_TYP, H_ALL_RES_TYPE_QP);
216 u64 r9_reg = EHEA_BMASK_SET(H_ALL_RES_QP_PD, pd)
217 | EHEA_BMASK_SET(H_ALL_RES_QP_TOKEN, init_attr->qp_token);
220 EHEA_BMASK_SET(H_ALL_RES_QP_MAX_SWQE,
221 get_order_of_qentries(init_attr->max_nr_send_wqes))
222 | EHEA_BMASK_SET(H_ALL_RES_QP_MAX_R1WQE,
223 get_order_of_qentries(init_attr->max_nr_rwqes_rq1))
224 | EHEA_BMASK_SET(H_ALL_RES_QP_MAX_R2WQE,
225 get_order_of_qentries(init_attr->max_nr_rwqes_rq2))
226 | EHEA_BMASK_SET(H_ALL_RES_QP_MAX_R3WQE,
227 get_order_of_qentries(init_attr->max_nr_rwqes_rq3))
228 | EHEA_BMASK_SET(H_ALL_RES_QP_MAX_SSGE, init_attr->wqe_size_enc_sq)
229 | EHEA_BMASK_SET(H_ALL_RES_QP_MAX_R1SGE,
230 init_attr->wqe_size_enc_rq1)
231 | EHEA_BMASK_SET(H_ALL_RES_QP_MAX_R2SGE,
232 init_attr->wqe_size_enc_rq2)
233 | EHEA_BMASK_SET(H_ALL_RES_QP_MAX_R3SGE,
234 init_attr->wqe_size_enc_rq3);
237 EHEA_BMASK_SET(H_ALL_RES_QP_SWQE_IDL, init_attr->swqe_imm_data_len)
238 | EHEA_BMASK_SET(H_ALL_RES_QP_PORT_NUM, init_attr->port_nr);
240 EHEA_BMASK_SET(H_ALL_RES_QP_TH_RQ2, init_attr->rq2_threshold)
241 | EHEA_BMASK_SET(H_ALL_RES_QP_TH_RQ3, init_attr->rq3_threshold);
243 hret = ehea_plpar_hcall9(H_ALLOC_HEA_RESOURCE,
245 adapter_handle, /* R4 */
246 allocate_controls, /* R5 */
247 init_attr->send_cq_handle, /* R6 */
248 init_attr->recv_cq_handle, /* R7 */
249 init_attr->aff_eq_handle, /* R8 */
251 max_r10_reg, /* R10 */
253 threshold); /* R12 */
255 *qp_handle = outs[0];
256 init_attr->qp_nr = (u32)outs[1];
258 init_attr->act_nr_send_wqes =
259 (u16)EHEA_BMASK_GET(H_ALL_RES_QP_ACT_SWQE, outs[2]);
260 init_attr->act_nr_rwqes_rq1 =
261 (u16)EHEA_BMASK_GET(H_ALL_RES_QP_ACT_R1WQE, outs[2]);
262 init_attr->act_nr_rwqes_rq2 =
263 (u16)EHEA_BMASK_GET(H_ALL_RES_QP_ACT_R2WQE, outs[2]);
264 init_attr->act_nr_rwqes_rq3 =
265 (u16)EHEA_BMASK_GET(H_ALL_RES_QP_ACT_R3WQE, outs[2]);
267 init_attr->act_wqe_size_enc_sq = init_attr->wqe_size_enc_sq;
268 init_attr->act_wqe_size_enc_rq1 = init_attr->wqe_size_enc_rq1;
269 init_attr->act_wqe_size_enc_rq2 = init_attr->wqe_size_enc_rq2;
270 init_attr->act_wqe_size_enc_rq3 = init_attr->wqe_size_enc_rq3;
272 init_attr->nr_sq_pages =
273 (u32)EHEA_BMASK_GET(H_ALL_RES_QP_SIZE_SQ, outs[4]);
274 init_attr->nr_rq1_pages =
275 (u32)EHEA_BMASK_GET(H_ALL_RES_QP_SIZE_RQ1, outs[4]);
276 init_attr->nr_rq2_pages =
277 (u32)EHEA_BMASK_GET(H_ALL_RES_QP_SIZE_RQ2, outs[5]);
278 init_attr->nr_rq3_pages =
279 (u32)EHEA_BMASK_GET(H_ALL_RES_QP_SIZE_RQ3, outs[5]);
281 init_attr->liobn_sq =
282 (u32)EHEA_BMASK_GET(H_ALL_RES_QP_LIOBN_SQ, outs[7]);
283 init_attr->liobn_rq1 =
284 (u32)EHEA_BMASK_GET(H_ALL_RES_QP_LIOBN_RQ1, outs[7]);
285 init_attr->liobn_rq2 =
286 (u32)EHEA_BMASK_GET(H_ALL_RES_QP_LIOBN_RQ2, outs[8]);
287 init_attr->liobn_rq3 =
288 (u32)EHEA_BMASK_GET(H_ALL_RES_QP_LIOBN_RQ3, outs[8]);
291 hcp_epas_ctor(h_epas, outs[6], outs[6]);
296 u64 ehea_h_alloc_resource_cq(const u64 adapter_handle,
297 struct ehea_cq_attr *cq_attr,
298 u64 *cq_handle, struct h_epas *epas)
301 unsigned long outs[PLPAR_HCALL9_BUFSIZE];
303 hret = ehea_plpar_hcall9(H_ALLOC_HEA_RESOURCE,
305 adapter_handle, /* R4 */
306 H_ALL_RES_TYPE_CQ, /* R5 */
307 cq_attr->eq_handle, /* R6 */
308 cq_attr->cq_token, /* R7 */
309 cq_attr->max_nr_of_cqes, /* R8 */
310 0, 0, 0, 0); /* R9-R12 */
312 *cq_handle = outs[0];
313 cq_attr->act_nr_of_cqes = outs[3];
314 cq_attr->nr_pages = outs[4];
317 hcp_epas_ctor(epas, outs[5], outs[6]);
322 /* Defines for H_CALL H_ALLOC_RESOURCE */
323 #define H_ALL_RES_TYPE_QP 1
324 #define H_ALL_RES_TYPE_CQ 2
325 #define H_ALL_RES_TYPE_EQ 3
326 #define H_ALL_RES_TYPE_MR 5
327 #define H_ALL_RES_TYPE_MW 6
330 #define H_ALL_RES_EQ_NEQ EHEA_BMASK_IBM(0, 0)
331 #define H_ALL_RES_EQ_NON_NEQ_ISN EHEA_BMASK_IBM(6, 7)
332 #define H_ALL_RES_EQ_INH_EQE_GEN EHEA_BMASK_IBM(16, 16)
333 #define H_ALL_RES_EQ_RES_TYPE EHEA_BMASK_IBM(56, 63)
335 #define H_ALL_RES_EQ_MAX_EQE EHEA_BMASK_IBM(32, 63)
337 /* output param R6 */
338 #define H_ALL_RES_EQ_LIOBN EHEA_BMASK_IBM(32, 63)
340 /* output param R7 */
341 #define H_ALL_RES_EQ_ACT_EQE EHEA_BMASK_IBM(32, 63)
343 /* output param R8 */
344 #define H_ALL_RES_EQ_ACT_PS EHEA_BMASK_IBM(32, 63)
346 /* output param R9 */
347 #define H_ALL_RES_EQ_ACT_EQ_IST_C EHEA_BMASK_IBM(30, 31)
348 #define H_ALL_RES_EQ_ACT_EQ_IST_1 EHEA_BMASK_IBM(40, 63)
350 /* output param R10 */
351 #define H_ALL_RES_EQ_ACT_EQ_IST_2 EHEA_BMASK_IBM(40, 63)
353 /* output param R11 */
354 #define H_ALL_RES_EQ_ACT_EQ_IST_3 EHEA_BMASK_IBM(40, 63)
356 /* output param R12 */
357 #define H_ALL_RES_EQ_ACT_EQ_IST_4 EHEA_BMASK_IBM(40, 63)
359 u64 ehea_h_alloc_resource_eq(const u64 adapter_handle,
360 struct ehea_eq_attr *eq_attr, u64 *eq_handle)
362 u64 hret, allocate_controls;
363 unsigned long outs[PLPAR_HCALL9_BUFSIZE];
367 EHEA_BMASK_SET(H_ALL_RES_EQ_RES_TYPE, H_ALL_RES_TYPE_EQ)
368 | EHEA_BMASK_SET(H_ALL_RES_EQ_NEQ, eq_attr->type ? 1 : 0)
369 | EHEA_BMASK_SET(H_ALL_RES_EQ_INH_EQE_GEN, !eq_attr->eqe_gen)
370 | EHEA_BMASK_SET(H_ALL_RES_EQ_NON_NEQ_ISN, 1);
372 hret = ehea_plpar_hcall9(H_ALLOC_HEA_RESOURCE,
374 adapter_handle, /* R4 */
375 allocate_controls, /* R5 */
376 eq_attr->max_nr_of_eqes, /* R6 */
377 0, 0, 0, 0, 0, 0); /* R7-R10 */
379 *eq_handle = outs[0];
380 eq_attr->act_nr_of_eqes = outs[3];
381 eq_attr->nr_pages = outs[4];
382 eq_attr->ist1 = outs[5];
383 eq_attr->ist2 = outs[6];
384 eq_attr->ist3 = outs[7];
385 eq_attr->ist4 = outs[8];
390 u64 ehea_h_modify_ehea_qp(const u64 adapter_handle, const u8 cat,
391 const u64 qp_handle, const u64 sel_mask,
392 void *cb_addr, u64 *inv_attr_id, u64 *proc_mask,
393 u16 *out_swr, u16 *out_rwr)
396 unsigned long outs[PLPAR_HCALL9_BUFSIZE];
398 hret = ehea_plpar_hcall9(H_MODIFY_HEA_QP,
400 adapter_handle, /* R4 */
404 __pa(cb_addr), /* R8 */
405 0, 0, 0, 0); /* R9-R12 */
407 *inv_attr_id = outs[0];
410 *proc_mask = outs[5];
415 u64 ehea_h_register_rpage(const u64 adapter_handle, const u8 pagesize,
416 const u8 queue_type, const u64 resource_handle,
417 const u64 log_pageaddr, u64 count)
421 reg_control = EHEA_BMASK_SET(H_REG_RPAGE_PAGE_SIZE, pagesize)
422 | EHEA_BMASK_SET(H_REG_RPAGE_QT, queue_type);
424 return ehea_plpar_hcall_norets(H_REGISTER_HEA_RPAGES,
425 adapter_handle, /* R4 */
426 reg_control, /* R5 */
427 resource_handle, /* R6 */
428 log_pageaddr, /* R7 */
433 u64 ehea_h_register_smr(const u64 adapter_handle, const u64 orig_mr_handle,
434 const u64 vaddr_in, const u32 access_ctrl, const u32 pd,
438 unsigned long outs[PLPAR_HCALL9_BUFSIZE];
440 hret = ehea_plpar_hcall9(H_REGISTER_SMR,
442 adapter_handle , /* R4 */
443 orig_mr_handle, /* R5 */
445 (((u64)access_ctrl) << 32ULL), /* R7 */
447 0, 0, 0, 0); /* R9-R12 */
449 mr->handle = outs[0];
450 mr->lkey = (u32)outs[2];
455 u64 ehea_h_disable_and_get_hea(const u64 adapter_handle, const u64 qp_handle)
457 unsigned long outs[PLPAR_HCALL9_BUFSIZE];
459 return ehea_plpar_hcall9(H_DISABLE_AND_GET_HEA,
461 adapter_handle, /* R4 */
462 H_DISABLE_GET_EHEA_WQE_P, /* R5 */
464 0, 0, 0, 0, 0, 0); /* R7-R12 */
467 u64 ehea_h_free_resource(const u64 adapter_handle, const u64 res_handle,
470 return ehea_plpar_hcall_norets(H_FREE_RESOURCE,
471 adapter_handle, /* R4 */
474 0, 0, 0, 0); /* R7-R10 */
477 u64 ehea_h_alloc_resource_mr(const u64 adapter_handle, const u64 vaddr,
478 const u64 length, const u32 access_ctrl,
479 const u32 pd, u64 *mr_handle, u32 *lkey)
482 unsigned long outs[PLPAR_HCALL9_BUFSIZE];
484 hret = ehea_plpar_hcall9(H_ALLOC_HEA_RESOURCE,
486 adapter_handle, /* R4 */
490 (((u64) access_ctrl) << 32ULL), /* R8 */
492 0, 0, 0); /* R10-R12 */
494 *mr_handle = outs[0];
495 *lkey = (u32)outs[2];
499 u64 ehea_h_register_rpage_mr(const u64 adapter_handle, const u64 mr_handle,
500 const u8 pagesize, const u8 queue_type,
501 const u64 log_pageaddr, const u64 count)
503 if ((count > 1) && (log_pageaddr & ~PAGE_MASK)) {
504 pr_err("not on pageboundary\n");
508 return ehea_h_register_rpage(adapter_handle, pagesize,
509 queue_type, mr_handle,
510 log_pageaddr, count);
513 u64 ehea_h_query_ehea(const u64 adapter_handle, void *cb_addr)
515 u64 hret, cb_logaddr;
517 cb_logaddr = __pa(cb_addr);
519 hret = ehea_plpar_hcall_norets(H_QUERY_HEA,
520 adapter_handle, /* R4 */
522 0, 0, 0, 0, 0); /* R6-R10 */
524 ehea_dump(cb_addr, sizeof(struct hcp_query_ehea), "hcp_query_ehea");
529 u64 ehea_h_query_ehea_port(const u64 adapter_handle, const u16 port_num,
530 const u8 cb_cat, const u64 select_mask,
534 u64 cb_logaddr = __pa(cb_addr);
537 port_info = EHEA_BMASK_SET(H_MEHEAPORT_CAT, cb_cat)
538 | EHEA_BMASK_SET(H_MEHEAPORT_PN, port_num);
540 return ehea_plpar_hcall_norets(H_QUERY_HEA_PORT,
541 adapter_handle, /* R4 */
543 select_mask, /* R6 */
549 u64 ehea_h_modify_ehea_port(const u64 adapter_handle, const u16 port_num,
550 const u8 cb_cat, const u64 select_mask,
553 unsigned long outs[PLPAR_HCALL9_BUFSIZE];
556 u64 cb_logaddr = __pa(cb_addr);
558 port_info = EHEA_BMASK_SET(H_MEHEAPORT_CAT, cb_cat)
559 | EHEA_BMASK_SET(H_MEHEAPORT_PN, port_num);
561 ehea_dump(cb_addr, sizeof(struct hcp_ehea_port_cb0), "Before HCALL");
563 return ehea_plpar_hcall9(H_MODIFY_HEA_PORT,
565 adapter_handle, /* R4 */
567 select_mask, /* R6 */
570 0, 0, 0, 0); /* R9-R12 */
573 u64 ehea_h_reg_dereg_bcmc(const u64 adapter_handle, const u16 port_num,
574 const u8 reg_type, const u64 mc_mac_addr,
575 const u16 vlan_id, const u32 hcall_id)
577 u64 r5_port_num, r6_reg_type, r7_mc_mac_addr, r8_vlan_id;
578 u64 mac_addr = mc_mac_addr >> 16;
580 r5_port_num = EHEA_BMASK_SET(H_REGBCMC_PN, port_num);
581 r6_reg_type = EHEA_BMASK_SET(H_REGBCMC_REGTYPE, reg_type);
582 r7_mc_mac_addr = EHEA_BMASK_SET(H_REGBCMC_MACADDR, mac_addr);
583 r8_vlan_id = EHEA_BMASK_SET(H_REGBCMC_VLANID, vlan_id);
585 return ehea_plpar_hcall_norets(hcall_id,
586 adapter_handle, /* R4 */
587 r5_port_num, /* R5 */
588 r6_reg_type, /* R6 */
589 r7_mc_mac_addr, /* R7 */
594 u64 ehea_h_reset_events(const u64 adapter_handle, const u64 neq_handle,
595 const u64 event_mask)
597 return ehea_plpar_hcall_norets(H_RESET_EVENTS,
598 adapter_handle, /* R4 */
601 0, 0, 0, 0); /* R7-R12 */
604 u64 ehea_h_error_data(const u64 adapter_handle, const u64 ressource_handle,
607 return ehea_plpar_hcall_norets(H_ERROR_DATA,
608 adapter_handle, /* R4 */
609 ressource_handle, /* R5 */
610 __pa(rblock), /* R6 */
611 0, 0, 0, 0); /* R7-R12 */