bnx2: stop using net_device.{base_addr, irq}.
[linux-2.6.git] / drivers / net / ethernet / ibm / ehea / ehea_phyp.h
blob52c456ec4d6cf811567f82f80477b4e32910ee9a
1 /*
2 * linux/drivers/net/ethernet/ibm/ehea/ehea_phyp.h
4 * eHEA ethernet device driver for IBM eServer System p
6 * (C) Copyright IBM Corp. 2006
8 * Authors:
9 * Christoph Raisch <raisch@de.ibm.com>
10 * Jan-Bernd Themann <themann@de.ibm.com>
11 * Thomas Klein <tklein@de.ibm.com>
14 * This program is free software; you can redistribute it and/or modify
15 * it under the terms of the GNU General Public License as published by
16 * the Free Software Foundation; either version 2, or (at your option)
17 * any later version.
19 * This program is distributed in the hope that it will be useful,
20 * but WITHOUT ANY WARRANTY; without even the implied warranty of
21 * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
22 * GNU General Public License for more details.
24 * You should have received a copy of the GNU General Public License
25 * along with this program; if not, write to the Free Software
26 * Foundation, Inc., 675 Mass Ave, Cambridge, MA 02139, USA.
29 #ifndef __EHEA_PHYP_H__
30 #define __EHEA_PHYP_H__
32 #include <linux/delay.h>
33 #include <asm/hvcall.h>
34 #include "ehea.h"
35 #include "ehea_hw.h"
37 /* Some abbreviations used here:
39 * hcp_* - structures, variables and functions releated to Hypervisor Calls
42 static inline u32 get_longbusy_msecs(int long_busy_ret_code)
44 switch (long_busy_ret_code) {
45 case H_LONG_BUSY_ORDER_1_MSEC:
46 return 1;
47 case H_LONG_BUSY_ORDER_10_MSEC:
48 return 10;
49 case H_LONG_BUSY_ORDER_100_MSEC:
50 return 100;
51 case H_LONG_BUSY_ORDER_1_SEC:
52 return 1000;
53 case H_LONG_BUSY_ORDER_10_SEC:
54 return 10000;
55 case H_LONG_BUSY_ORDER_100_SEC:
56 return 100000;
57 default:
58 return 1;
62 /* Number of pages which can be registered at once by H_REGISTER_HEA_RPAGES */
63 #define EHEA_MAX_RPAGE 512
65 /* Notification Event Queue (NEQ) Entry bit masks */
66 #define NEQE_EVENT_CODE EHEA_BMASK_IBM(2, 7)
67 #define NEQE_PORTNUM EHEA_BMASK_IBM(32, 47)
68 #define NEQE_PORT_UP EHEA_BMASK_IBM(16, 16)
69 #define NEQE_EXTSWITCH_PORT_UP EHEA_BMASK_IBM(17, 17)
70 #define NEQE_EXTSWITCH_PRIMARY EHEA_BMASK_IBM(18, 18)
71 #define NEQE_PLID EHEA_BMASK_IBM(16, 47)
73 /* Notification Event Codes */
74 #define EHEA_EC_PORTSTATE_CHG 0x30
75 #define EHEA_EC_ADAPTER_MALFUNC 0x32
76 #define EHEA_EC_PORT_MALFUNC 0x33
78 /* Notification Event Log Register (NELR) bit masks */
79 #define NELR_PORT_MALFUNC EHEA_BMASK_IBM(61, 61)
80 #define NELR_ADAPTER_MALFUNC EHEA_BMASK_IBM(62, 62)
81 #define NELR_PORTSTATE_CHG EHEA_BMASK_IBM(63, 63)
83 static inline void hcp_epas_ctor(struct h_epas *epas, u64 paddr_kernel,
84 u64 paddr_user)
86 /* To support 64k pages we must round to 64k page boundary */
87 epas->kernel.addr = ioremap((paddr_kernel & PAGE_MASK), PAGE_SIZE) +
88 (paddr_kernel & ~PAGE_MASK);
89 epas->user.addr = paddr_user;
92 static inline void hcp_epas_dtor(struct h_epas *epas)
94 if (epas->kernel.addr)
95 iounmap((void __iomem *)((u64)epas->kernel.addr & PAGE_MASK));
97 epas->user.addr = 0;
98 epas->kernel.addr = 0;
101 struct hcp_modify_qp_cb0 {
102 u64 qp_ctl_reg; /* 00 */
103 u32 max_swqe; /* 02 */
104 u32 max_rwqe; /* 03 */
105 u32 port_nb; /* 04 */
106 u32 reserved0; /* 05 */
107 u64 qp_aer; /* 06 */
108 u64 qp_tenure; /* 08 */
111 /* Hcall Query/Modify Queue Pair Control Block 0 Selection Mask Bits */
112 #define H_QPCB0_ALL EHEA_BMASK_IBM(0, 5)
113 #define H_QPCB0_QP_CTL_REG EHEA_BMASK_IBM(0, 0)
114 #define H_QPCB0_MAX_SWQE EHEA_BMASK_IBM(1, 1)
115 #define H_QPCB0_MAX_RWQE EHEA_BMASK_IBM(2, 2)
116 #define H_QPCB0_PORT_NB EHEA_BMASK_IBM(3, 3)
117 #define H_QPCB0_QP_AER EHEA_BMASK_IBM(4, 4)
118 #define H_QPCB0_QP_TENURE EHEA_BMASK_IBM(5, 5)
120 /* Queue Pair Control Register Status Bits */
121 #define H_QP_CR_ENABLED 0x8000000000000000ULL /* QP enabled */
122 /* QP States: */
123 #define H_QP_CR_STATE_RESET 0x0000010000000000ULL /* Reset */
124 #define H_QP_CR_STATE_INITIALIZED 0x0000020000000000ULL /* Initialized */
125 #define H_QP_CR_STATE_RDY2RCV 0x0000030000000000ULL /* Ready to recv */
126 #define H_QP_CR_STATE_RDY2SND 0x0000050000000000ULL /* Ready to send */
127 #define H_QP_CR_STATE_ERROR 0x0000800000000000ULL /* Error */
128 #define H_QP_CR_RES_STATE 0x0000007F00000000ULL /* Resultant state */
130 struct hcp_modify_qp_cb1 {
131 u32 qpn; /* 00 */
132 u32 qp_asyn_ev_eq_nb; /* 01 */
133 u64 sq_cq_handle; /* 02 */
134 u64 rq_cq_handle; /* 04 */
135 /* sgel = scatter gather element */
136 u32 sgel_nb_sq; /* 06 */
137 u32 sgel_nb_rq1; /* 07 */
138 u32 sgel_nb_rq2; /* 08 */
139 u32 sgel_nb_rq3; /* 09 */
142 /* Hcall Query/Modify Queue Pair Control Block 1 Selection Mask Bits */
143 #define H_QPCB1_ALL EHEA_BMASK_IBM(0, 7)
144 #define H_QPCB1_QPN EHEA_BMASK_IBM(0, 0)
145 #define H_QPCB1_ASYN_EV_EQ_NB EHEA_BMASK_IBM(1, 1)
146 #define H_QPCB1_SQ_CQ_HANDLE EHEA_BMASK_IBM(2, 2)
147 #define H_QPCB1_RQ_CQ_HANDLE EHEA_BMASK_IBM(3, 3)
148 #define H_QPCB1_SGEL_NB_SQ EHEA_BMASK_IBM(4, 4)
149 #define H_QPCB1_SGEL_NB_RQ1 EHEA_BMASK_IBM(5, 5)
150 #define H_QPCB1_SGEL_NB_RQ2 EHEA_BMASK_IBM(6, 6)
151 #define H_QPCB1_SGEL_NB_RQ3 EHEA_BMASK_IBM(7, 7)
153 struct hcp_query_ehea {
154 u32 cur_num_qps; /* 00 */
155 u32 cur_num_cqs; /* 01 */
156 u32 cur_num_eqs; /* 02 */
157 u32 cur_num_mrs; /* 03 */
158 u32 auth_level; /* 04 */
159 u32 max_num_qps; /* 05 */
160 u32 max_num_cqs; /* 06 */
161 u32 max_num_eqs; /* 07 */
162 u32 max_num_mrs; /* 08 */
163 u32 reserved0; /* 09 */
164 u32 int_clock_freq; /* 10 */
165 u32 max_num_pds; /* 11 */
166 u32 max_num_addr_handles; /* 12 */
167 u32 max_num_cqes; /* 13 */
168 u32 max_num_wqes; /* 14 */
169 u32 max_num_sgel_rq1wqe; /* 15 */
170 u32 max_num_sgel_rq2wqe; /* 16 */
171 u32 max_num_sgel_rq3wqe; /* 17 */
172 u32 mr_page_size; /* 18 */
173 u32 reserved1; /* 19 */
174 u64 max_mr_size; /* 20 */
175 u64 reserved2; /* 22 */
176 u32 num_ports; /* 24 */
177 u32 reserved3; /* 25 */
178 u32 reserved4; /* 26 */
179 u32 reserved5; /* 27 */
180 u64 max_mc_mac; /* 28 */
181 u64 ehea_cap; /* 30 */
182 u32 max_isn_per_eq; /* 32 */
183 u32 max_num_neq; /* 33 */
184 u64 max_num_vlan_ids; /* 34 */
185 u32 max_num_port_group; /* 36 */
186 u32 max_num_phys_port; /* 37 */
190 /* Hcall Query/Modify Port Control Block defines */
191 #define H_PORT_CB0 0
192 #define H_PORT_CB1 1
193 #define H_PORT_CB2 2
194 #define H_PORT_CB3 3
195 #define H_PORT_CB4 4
196 #define H_PORT_CB5 5
197 #define H_PORT_CB6 6
198 #define H_PORT_CB7 7
200 struct hcp_ehea_port_cb0 {
201 u64 port_mac_addr;
202 u64 port_rc;
203 u64 reserved0;
204 u32 port_op_state;
205 u32 port_speed;
206 u32 ext_swport_op_state;
207 u32 neg_tpf_prpf;
208 u32 num_default_qps;
209 u32 reserved1;
210 u64 default_qpn_arr[16];
213 /* Hcall Query/Modify Port Control Block 0 Selection Mask Bits */
214 #define H_PORT_CB0_ALL EHEA_BMASK_IBM(0, 7) /* Set all bits */
215 #define H_PORT_CB0_MAC EHEA_BMASK_IBM(0, 0) /* MAC address */
216 #define H_PORT_CB0_PRC EHEA_BMASK_IBM(1, 1) /* Port Recv Control */
217 #define H_PORT_CB0_DEFQPNARRAY EHEA_BMASK_IBM(7, 7) /* Default QPN Array */
219 /* Hcall Query Port: Returned port speed values */
220 #define H_SPEED_10M_H 1 /* 10 Mbps, Half Duplex */
221 #define H_SPEED_10M_F 2 /* 10 Mbps, Full Duplex */
222 #define H_SPEED_100M_H 3 /* 100 Mbps, Half Duplex */
223 #define H_SPEED_100M_F 4 /* 100 Mbps, Full Duplex */
224 #define H_SPEED_1G_F 6 /* 1 Gbps, Full Duplex */
225 #define H_SPEED_10G_F 8 /* 10 Gbps, Full Duplex */
227 /* Port Receive Control Status Bits */
228 #define PXLY_RC_VALID EHEA_BMASK_IBM(49, 49)
229 #define PXLY_RC_VLAN_XTRACT EHEA_BMASK_IBM(50, 50)
230 #define PXLY_RC_TCP_6_TUPLE EHEA_BMASK_IBM(51, 51)
231 #define PXLY_RC_UDP_6_TUPLE EHEA_BMASK_IBM(52, 52)
232 #define PXLY_RC_TCP_3_TUPLE EHEA_BMASK_IBM(53, 53)
233 #define PXLY_RC_TCP_2_TUPLE EHEA_BMASK_IBM(54, 54)
234 #define PXLY_RC_LLC_SNAP EHEA_BMASK_IBM(55, 55)
235 #define PXLY_RC_JUMBO_FRAME EHEA_BMASK_IBM(56, 56)
236 #define PXLY_RC_FRAG_IP_PKT EHEA_BMASK_IBM(57, 57)
237 #define PXLY_RC_TCP_UDP_CHKSUM EHEA_BMASK_IBM(58, 58)
238 #define PXLY_RC_IP_CHKSUM EHEA_BMASK_IBM(59, 59)
239 #define PXLY_RC_MAC_FILTER EHEA_BMASK_IBM(60, 60)
240 #define PXLY_RC_UNTAG_FILTER EHEA_BMASK_IBM(61, 61)
241 #define PXLY_RC_VLAN_TAG_FILTER EHEA_BMASK_IBM(62, 63)
243 #define PXLY_RC_VLAN_FILTER 2
244 #define PXLY_RC_VLAN_PERM 0
247 #define H_PORT_CB1_ALL 0x8000000000000000ULL
249 struct hcp_ehea_port_cb1 {
250 u64 vlan_filter[64];
253 #define H_PORT_CB2_ALL 0xFFE0000000000000ULL
255 struct hcp_ehea_port_cb2 {
256 u64 rxo;
257 u64 rxucp;
258 u64 rxufd;
259 u64 rxuerr;
260 u64 rxftl;
261 u64 rxmcp;
262 u64 rxbcp;
263 u64 txo;
264 u64 txucp;
265 u64 txmcp;
266 u64 txbcp;
269 struct hcp_ehea_port_cb3 {
270 u64 vlan_bc_filter[64];
271 u64 vlan_mc_filter[64];
272 u64 vlan_un_filter[64];
273 u64 port_mac_hash_array[64];
276 #define H_PORT_CB4_ALL 0xF000000000000000ULL
277 #define H_PORT_CB4_JUMBO 0x1000000000000000ULL
278 #define H_PORT_CB4_SPEED 0x8000000000000000ULL
280 struct hcp_ehea_port_cb4 {
281 u32 port_speed;
282 u32 pause_frame;
283 u32 ens_port_op_state;
284 u32 jumbo_frame;
285 u32 ens_port_wrap;
288 /* Hcall Query/Modify Port Control Block 5 Selection Mask Bits */
289 #define H_PORT_CB5_RCU 0x0001000000000000ULL
290 #define PXS_RCU EHEA_BMASK_IBM(61, 63)
292 struct hcp_ehea_port_cb5 {
293 u64 prc; /* 00 */
294 u64 uaa; /* 01 */
295 u64 macvc; /* 02 */
296 u64 xpcsc; /* 03 */
297 u64 xpcsp; /* 04 */
298 u64 pcsid; /* 05 */
299 u64 xpcsst; /* 06 */
300 u64 pthlb; /* 07 */
301 u64 pthrb; /* 08 */
302 u64 pqu; /* 09 */
303 u64 pqd; /* 10 */
304 u64 prt; /* 11 */
305 u64 wsth; /* 12 */
306 u64 rcb; /* 13 */
307 u64 rcm; /* 14 */
308 u64 rcu; /* 15 */
309 u64 macc; /* 16 */
310 u64 pc; /* 17 */
311 u64 pst; /* 18 */
312 u64 ducqpn; /* 19 */
313 u64 mcqpn; /* 20 */
314 u64 mma; /* 21 */
315 u64 pmc0h; /* 22 */
316 u64 pmc0l; /* 23 */
317 u64 lbc; /* 24 */
320 #define H_PORT_CB6_ALL 0xFFFFFE7FFFFF8000ULL
322 struct hcp_ehea_port_cb6 {
323 u64 rxo; /* 00 */
324 u64 rx64; /* 01 */
325 u64 rx65; /* 02 */
326 u64 rx128; /* 03 */
327 u64 rx256; /* 04 */
328 u64 rx512; /* 05 */
329 u64 rx1024; /* 06 */
330 u64 rxbfcs; /* 07 */
331 u64 rxime; /* 08 */
332 u64 rxrle; /* 09 */
333 u64 rxorle; /* 10 */
334 u64 rxftl; /* 11 */
335 u64 rxjab; /* 12 */
336 u64 rxse; /* 13 */
337 u64 rxce; /* 14 */
338 u64 rxrf; /* 15 */
339 u64 rxfrag; /* 16 */
340 u64 rxuoc; /* 17 */
341 u64 rxcpf; /* 18 */
342 u64 rxsb; /* 19 */
343 u64 rxfd; /* 20 */
344 u64 rxoerr; /* 21 */
345 u64 rxaln; /* 22 */
346 u64 ducqpn; /* 23 */
347 u64 reserved0; /* 24 */
348 u64 rxmcp; /* 25 */
349 u64 rxbcp; /* 26 */
350 u64 txmcp; /* 27 */
351 u64 txbcp; /* 28 */
352 u64 txo; /* 29 */
353 u64 tx64; /* 30 */
354 u64 tx65; /* 31 */
355 u64 tx128; /* 32 */
356 u64 tx256; /* 33 */
357 u64 tx512; /* 34 */
358 u64 tx1024; /* 35 */
359 u64 txbfcs; /* 36 */
360 u64 txcpf; /* 37 */
361 u64 txlf; /* 38 */
362 u64 txrf; /* 39 */
363 u64 txime; /* 40 */
364 u64 txsc; /* 41 */
365 u64 txmc; /* 42 */
366 u64 txsqe; /* 43 */
367 u64 txdef; /* 44 */
368 u64 txlcol; /* 45 */
369 u64 txexcol; /* 46 */
370 u64 txcse; /* 47 */
371 u64 txbor; /* 48 */
374 #define H_PORT_CB7_DUCQPN 0x8000000000000000ULL
376 struct hcp_ehea_port_cb7 {
377 u64 def_uc_qpn;
380 u64 ehea_h_query_ehea_qp(const u64 adapter_handle,
381 const u8 qp_category,
382 const u64 qp_handle, const u64 sel_mask,
383 void *cb_addr);
385 u64 ehea_h_modify_ehea_qp(const u64 adapter_handle,
386 const u8 cat,
387 const u64 qp_handle,
388 const u64 sel_mask,
389 void *cb_addr,
390 u64 *inv_attr_id,
391 u64 *proc_mask, u16 *out_swr, u16 *out_rwr);
393 u64 ehea_h_alloc_resource_eq(const u64 adapter_handle,
394 struct ehea_eq_attr *eq_attr, u64 *eq_handle);
396 u64 ehea_h_alloc_resource_cq(const u64 adapter_handle,
397 struct ehea_cq_attr *cq_attr,
398 u64 *cq_handle, struct h_epas *epas);
400 u64 ehea_h_alloc_resource_qp(const u64 adapter_handle,
401 struct ehea_qp_init_attr *init_attr,
402 const u32 pd,
403 u64 *qp_handle, struct h_epas *h_epas);
405 #define H_REG_RPAGE_PAGE_SIZE EHEA_BMASK_IBM(48, 55)
406 #define H_REG_RPAGE_QT EHEA_BMASK_IBM(62, 63)
408 u64 ehea_h_register_rpage(const u64 adapter_handle,
409 const u8 pagesize,
410 const u8 queue_type,
411 const u64 resource_handle,
412 const u64 log_pageaddr, u64 count);
414 #define H_DISABLE_GET_EHEA_WQE_P 1
415 #define H_DISABLE_GET_SQ_WQE_P 2
416 #define H_DISABLE_GET_RQC 3
418 u64 ehea_h_disable_and_get_hea(const u64 adapter_handle, const u64 qp_handle);
420 #define FORCE_FREE 1
421 #define NORMAL_FREE 0
423 u64 ehea_h_free_resource(const u64 adapter_handle, const u64 res_handle,
424 u64 force_bit);
426 u64 ehea_h_alloc_resource_mr(const u64 adapter_handle, const u64 vaddr,
427 const u64 length, const u32 access_ctrl,
428 const u32 pd, u64 *mr_handle, u32 *lkey);
430 u64 ehea_h_register_rpage_mr(const u64 adapter_handle, const u64 mr_handle,
431 const u8 pagesize, const u8 queue_type,
432 const u64 log_pageaddr, const u64 count);
434 u64 ehea_h_register_smr(const u64 adapter_handle, const u64 orig_mr_handle,
435 const u64 vaddr_in, const u32 access_ctrl, const u32 pd,
436 struct ehea_mr *mr);
438 u64 ehea_h_query_ehea(const u64 adapter_handle, void *cb_addr);
440 /* output param R5 */
441 #define H_MEHEAPORT_CAT EHEA_BMASK_IBM(40, 47)
442 #define H_MEHEAPORT_PN EHEA_BMASK_IBM(48, 63)
444 u64 ehea_h_query_ehea_port(const u64 adapter_handle, const u16 port_num,
445 const u8 cb_cat, const u64 select_mask,
446 void *cb_addr);
448 u64 ehea_h_modify_ehea_port(const u64 adapter_handle, const u16 port_num,
449 const u8 cb_cat, const u64 select_mask,
450 void *cb_addr);
452 #define H_REGBCMC_PN EHEA_BMASK_IBM(48, 63)
453 #define H_REGBCMC_REGTYPE EHEA_BMASK_IBM(61, 63)
454 #define H_REGBCMC_MACADDR EHEA_BMASK_IBM(16, 63)
455 #define H_REGBCMC_VLANID EHEA_BMASK_IBM(52, 63)
457 u64 ehea_h_reg_dereg_bcmc(const u64 adapter_handle, const u16 port_num,
458 const u8 reg_type, const u64 mc_mac_addr,
459 const u16 vlan_id, const u32 hcall_id);
461 u64 ehea_h_reset_events(const u64 adapter_handle, const u64 neq_handle,
462 const u64 event_mask);
464 u64 ehea_h_error_data(const u64 adapter_handle, const u64 ressource_handle,
465 void *rblock);
467 #endif /* __EHEA_PHYP_H__ */