Linux Audio

Check our new training course

Loading...
v6.8
  1/* SPDX-License-Identifier: GPL-2.0-or-later */
  2/*
  3 *  linux/drivers/net/ethernet/ibm/ehea/ehea_hw.h
  4 *
  5 *  eHEA ethernet device driver for IBM eServer System p
  6 *
  7 *  (C) Copyright IBM Corp. 2006
  8 *
  9 *  Authors:
 10 *       Christoph Raisch <raisch@de.ibm.com>
 11 *       Jan-Bernd Themann <themann@de.ibm.com>
 12 *       Thomas Klein <tklein@de.ibm.com>
 13 */
 14
 15#ifndef __EHEA_HW_H__
 16#define __EHEA_HW_H__
 17
 18#define QPX_SQA_VALUE   EHEA_BMASK_IBM(48, 63)
 19#define QPX_RQ1A_VALUE  EHEA_BMASK_IBM(48, 63)
 20#define QPX_RQ2A_VALUE  EHEA_BMASK_IBM(48, 63)
 21#define QPX_RQ3A_VALUE  EHEA_BMASK_IBM(48, 63)
 22
 23#define QPTEMM_OFFSET(x) offsetof(struct ehea_qptemm, x)
 24
 25struct ehea_qptemm {
 26	u64 qpx_hcr;
 27	u64 qpx_c;
 28	u64 qpx_herr;
 29	u64 qpx_aer;
 30	u64 qpx_sqa;
 31	u64 qpx_sqc;
 32	u64 qpx_rq1a;
 33	u64 qpx_rq1c;
 34	u64 qpx_st;
 35	u64 qpx_aerr;
 36	u64 qpx_tenure;
 37	u64 qpx_reserved1[(0x098 - 0x058) / 8];
 38	u64 qpx_portp;
 39	u64 qpx_reserved2[(0x100 - 0x0A0) / 8];
 40	u64 qpx_t;
 41	u64 qpx_sqhp;
 42	u64 qpx_sqptp;
 43	u64 qpx_reserved3[(0x140 - 0x118) / 8];
 44	u64 qpx_sqwsize;
 45	u64 qpx_reserved4[(0x170 - 0x148) / 8];
 46	u64 qpx_sqsize;
 47	u64 qpx_reserved5[(0x1B0 - 0x178) / 8];
 48	u64 qpx_sigt;
 49	u64 qpx_wqecnt;
 50	u64 qpx_rq1hp;
 51	u64 qpx_rq1ptp;
 52	u64 qpx_rq1size;
 53	u64 qpx_reserved6[(0x220 - 0x1D8) / 8];
 54	u64 qpx_rq1wsize;
 55	u64 qpx_reserved7[(0x240 - 0x228) / 8];
 56	u64 qpx_pd;
 57	u64 qpx_scqn;
 58	u64 qpx_rcqn;
 59	u64 qpx_aeqn;
 60	u64 reserved49;
 61	u64 qpx_ram;
 62	u64 qpx_reserved8[(0x300 - 0x270) / 8];
 63	u64 qpx_rq2a;
 64	u64 qpx_rq2c;
 65	u64 qpx_rq2hp;
 66	u64 qpx_rq2ptp;
 67	u64 qpx_rq2size;
 68	u64 qpx_rq2wsize;
 69	u64 qpx_rq2th;
 70	u64 qpx_rq3a;
 71	u64 qpx_rq3c;
 72	u64 qpx_rq3hp;
 73	u64 qpx_rq3ptp;
 74	u64 qpx_rq3size;
 75	u64 qpx_rq3wsize;
 76	u64 qpx_rq3th;
 77	u64 qpx_lpn;
 78	u64 qpx_reserved9[(0x400 - 0x378) / 8];
 79	u64 reserved_ext[(0x500 - 0x400) / 8];
 80	u64 reserved2[(0x1000 - 0x500) / 8];
 81};
 82
 83#define MRx_HCR_LPARID_VALID EHEA_BMASK_IBM(0, 0)
 84
 85#define MRMWMM_OFFSET(x) offsetof(struct ehea_mrmwmm, x)
 86
 87struct ehea_mrmwmm {
 88	u64 mrx_hcr;
 89	u64 mrx_c;
 90	u64 mrx_herr;
 91	u64 mrx_aer;
 92	u64 mrx_pp;
 93	u64 reserved1;
 94	u64 reserved2;
 95	u64 reserved3;
 96	u64 reserved4[(0x200 - 0x40) / 8];
 97	u64 mrx_ctl[64];
 98};
 99
100#define QPEDMM_OFFSET(x) offsetof(struct ehea_qpedmm, x)
101
102struct ehea_qpedmm {
103
104	u64 reserved0[(0x400) / 8];
105	u64 qpedx_phh;
106	u64 qpedx_ppsgp;
107	u64 qpedx_ppsgu;
108	u64 qpedx_ppdgp;
109	u64 qpedx_ppdgu;
110	u64 qpedx_aph;
111	u64 qpedx_apsgp;
112	u64 qpedx_apsgu;
113	u64 qpedx_apdgp;
114	u64 qpedx_apdgu;
115	u64 qpedx_apav;
116	u64 qpedx_apsav;
117	u64 qpedx_hcr;
118	u64 reserved1[4];
119	u64 qpedx_rrl0;
120	u64 qpedx_rrrkey0;
121	u64 qpedx_rrva0;
122	u64 reserved2;
123	u64 qpedx_rrl1;
124	u64 qpedx_rrrkey1;
125	u64 qpedx_rrva1;
126	u64 reserved3;
127	u64 qpedx_rrl2;
128	u64 qpedx_rrrkey2;
129	u64 qpedx_rrva2;
130	u64 reserved4;
131	u64 qpedx_rrl3;
132	u64 qpedx_rrrkey3;
133	u64 qpedx_rrva3;
134};
135
136#define CQX_FECADDER EHEA_BMASK_IBM(32, 63)
137#define CQX_FEC_CQE_CNT EHEA_BMASK_IBM(32, 63)
138#define CQX_N1_GENERATE_COMP_EVENT EHEA_BMASK_IBM(0, 0)
139#define CQX_EP_EVENT_PENDING EHEA_BMASK_IBM(0, 0)
140
141#define CQTEMM_OFFSET(x) offsetof(struct ehea_cqtemm, x)
142
143struct ehea_cqtemm {
144	u64 cqx_hcr;
145	u64 cqx_c;
146	u64 cqx_herr;
147	u64 cqx_aer;
148	u64 cqx_ptp;
149	u64 cqx_tp;
150	u64 cqx_fec;
151	u64 cqx_feca;
152	u64 cqx_ep;
153	u64 cqx_eq;
154	u64 reserved1;
155	u64 cqx_n0;
156	u64 cqx_n1;
157	u64 reserved2[(0x1000 - 0x60) / 8];
158};
159
160#define EQTEMM_OFFSET(x) offsetof(struct ehea_eqtemm, x)
161
162struct ehea_eqtemm {
163	u64 eqx_hcr;
164	u64 eqx_c;
165	u64 eqx_herr;
166	u64 eqx_aer;
167	u64 eqx_ptp;
168	u64 eqx_tp;
169	u64 eqx_ssba;
170	u64 eqx_psba;
171	u64 eqx_cec;
172	u64 eqx_meql;
173	u64 eqx_xisbi;
174	u64 eqx_xisc;
175	u64 eqx_it;
176};
177
178/*
179 * These access functions will be changed when the dissuccsion about
180 * the new access methods for POWER has settled.
181 */
182
183static inline u64 epa_load(struct h_epa epa, u32 offset)
184{
185	return __raw_readq((void __iomem *)(epa.addr + offset));
186}
187
188static inline void epa_store(struct h_epa epa, u32 offset, u64 value)
189{
190	__raw_writeq(value, (void __iomem *)(epa.addr + offset));
191	epa_load(epa, offset);	/* synchronize explicitly to eHEA */
192}
193
194static inline void epa_store_acc(struct h_epa epa, u32 offset, u64 value)
195{
196	__raw_writeq(value, (void __iomem *)(epa.addr + offset));
197}
198
199#define epa_store_cq(epa, offset, value)\
200	epa_store(epa, CQTEMM_OFFSET(offset), value)
201#define epa_load_cq(epa, offset)\
202	epa_load(epa, CQTEMM_OFFSET(offset))
203
204static inline void ehea_update_sqa(struct ehea_qp *qp, u16 nr_wqes)
205{
206	struct h_epa epa = qp->epas.kernel;
207	epa_store_acc(epa, QPTEMM_OFFSET(qpx_sqa),
208		      EHEA_BMASK_SET(QPX_SQA_VALUE, nr_wqes));
209}
210
211static inline void ehea_update_rq3a(struct ehea_qp *qp, u16 nr_wqes)
212{
213	struct h_epa epa = qp->epas.kernel;
214	epa_store_acc(epa, QPTEMM_OFFSET(qpx_rq3a),
215		      EHEA_BMASK_SET(QPX_RQ1A_VALUE, nr_wqes));
216}
217
218static inline void ehea_update_rq2a(struct ehea_qp *qp, u16 nr_wqes)
219{
220	struct h_epa epa = qp->epas.kernel;
221	epa_store_acc(epa, QPTEMM_OFFSET(qpx_rq2a),
222		      EHEA_BMASK_SET(QPX_RQ2A_VALUE, nr_wqes));
223}
224
225static inline void ehea_update_rq1a(struct ehea_qp *qp, u16 nr_wqes)
226{
227	struct h_epa epa = qp->epas.kernel;
228	epa_store_acc(epa, QPTEMM_OFFSET(qpx_rq1a),
229		      EHEA_BMASK_SET(QPX_RQ3A_VALUE, nr_wqes));
230}
231
232static inline void ehea_update_feca(struct ehea_cq *cq, u32 nr_cqes)
233{
234	struct h_epa epa = cq->epas.kernel;
235	epa_store_acc(epa, CQTEMM_OFFSET(cqx_feca),
236		      EHEA_BMASK_SET(CQX_FECADDER, nr_cqes));
237}
238
239static inline void ehea_reset_cq_n1(struct ehea_cq *cq)
240{
241	struct h_epa epa = cq->epas.kernel;
242	epa_store_cq(epa, cqx_n1,
243		     EHEA_BMASK_SET(CQX_N1_GENERATE_COMP_EVENT, 1));
244}
245
246static inline void ehea_reset_cq_ep(struct ehea_cq *my_cq)
247{
248	struct h_epa epa = my_cq->epas.kernel;
249	epa_store_acc(epa, CQTEMM_OFFSET(cqx_ep),
250		      EHEA_BMASK_SET(CQX_EP_EVENT_PENDING, 0));
251}
252
253#endif	/* __EHEA_HW_H__ */
v6.13.7
  1/* SPDX-License-Identifier: GPL-2.0-or-later */
  2/*
  3 *  linux/drivers/net/ethernet/ibm/ehea/ehea_hw.h
  4 *
  5 *  eHEA ethernet device driver for IBM eServer System p
  6 *
  7 *  (C) Copyright IBM Corp. 2006
  8 *
  9 *  Authors:
 10 *       Christoph Raisch <raisch@de.ibm.com>
 11 *       Jan-Bernd Themann <themann@de.ibm.com>
 12 *       Thomas Klein <tklein@de.ibm.com>
 13 */
 14
 15#ifndef __EHEA_HW_H__
 16#define __EHEA_HW_H__
 17
 18#define QPX_SQA_VALUE   EHEA_BMASK_IBM(48, 63)
 19#define QPX_RQ1A_VALUE  EHEA_BMASK_IBM(48, 63)
 20#define QPX_RQ2A_VALUE  EHEA_BMASK_IBM(48, 63)
 21#define QPX_RQ3A_VALUE  EHEA_BMASK_IBM(48, 63)
 22
 23#define QPTEMM_OFFSET(x) offsetof(struct ehea_qptemm, x)
 24
 25struct ehea_qptemm {
 26	u64 qpx_hcr;
 27	u64 qpx_c;
 28	u64 qpx_herr;
 29	u64 qpx_aer;
 30	u64 qpx_sqa;
 31	u64 qpx_sqc;
 32	u64 qpx_rq1a;
 33	u64 qpx_rq1c;
 34	u64 qpx_st;
 35	u64 qpx_aerr;
 36	u64 qpx_tenure;
 37	u64 qpx_reserved1[(0x098 - 0x058) / 8];
 38	u64 qpx_portp;
 39	u64 qpx_reserved2[(0x100 - 0x0A0) / 8];
 40	u64 qpx_t;
 41	u64 qpx_sqhp;
 42	u64 qpx_sqptp;
 43	u64 qpx_reserved3[(0x140 - 0x118) / 8];
 44	u64 qpx_sqwsize;
 45	u64 qpx_reserved4[(0x170 - 0x148) / 8];
 46	u64 qpx_sqsize;
 47	u64 qpx_reserved5[(0x1B0 - 0x178) / 8];
 48	u64 qpx_sigt;
 49	u64 qpx_wqecnt;
 50	u64 qpx_rq1hp;
 51	u64 qpx_rq1ptp;
 52	u64 qpx_rq1size;
 53	u64 qpx_reserved6[(0x220 - 0x1D8) / 8];
 54	u64 qpx_rq1wsize;
 55	u64 qpx_reserved7[(0x240 - 0x228) / 8];
 56	u64 qpx_pd;
 57	u64 qpx_scqn;
 58	u64 qpx_rcqn;
 59	u64 qpx_aeqn;
 60	u64 reserved49;
 61	u64 qpx_ram;
 62	u64 qpx_reserved8[(0x300 - 0x270) / 8];
 63	u64 qpx_rq2a;
 64	u64 qpx_rq2c;
 65	u64 qpx_rq2hp;
 66	u64 qpx_rq2ptp;
 67	u64 qpx_rq2size;
 68	u64 qpx_rq2wsize;
 69	u64 qpx_rq2th;
 70	u64 qpx_rq3a;
 71	u64 qpx_rq3c;
 72	u64 qpx_rq3hp;
 73	u64 qpx_rq3ptp;
 74	u64 qpx_rq3size;
 75	u64 qpx_rq3wsize;
 76	u64 qpx_rq3th;
 77	u64 qpx_lpn;
 78	u64 qpx_reserved9[(0x400 - 0x378) / 8];
 79	u64 reserved_ext[(0x500 - 0x400) / 8];
 80	u64 reserved2[(0x1000 - 0x500) / 8];
 81};
 82
 83#define MRx_HCR_LPARID_VALID EHEA_BMASK_IBM(0, 0)
 84
 85#define MRMWMM_OFFSET(x) offsetof(struct ehea_mrmwmm, x)
 86
 87struct ehea_mrmwmm {
 88	u64 mrx_hcr;
 89	u64 mrx_c;
 90	u64 mrx_herr;
 91	u64 mrx_aer;
 92	u64 mrx_pp;
 93	u64 reserved1;
 94	u64 reserved2;
 95	u64 reserved3;
 96	u64 reserved4[(0x200 - 0x40) / 8];
 97	u64 mrx_ctl[64];
 98};
 99
100#define QPEDMM_OFFSET(x) offsetof(struct ehea_qpedmm, x)
101
102struct ehea_qpedmm {
103
104	u64 reserved0[(0x400) / 8];
105	u64 qpedx_phh;
106	u64 qpedx_ppsgp;
107	u64 qpedx_ppsgu;
108	u64 qpedx_ppdgp;
109	u64 qpedx_ppdgu;
110	u64 qpedx_aph;
111	u64 qpedx_apsgp;
112	u64 qpedx_apsgu;
113	u64 qpedx_apdgp;
114	u64 qpedx_apdgu;
115	u64 qpedx_apav;
116	u64 qpedx_apsav;
117	u64 qpedx_hcr;
118	u64 reserved1[4];
119	u64 qpedx_rrl0;
120	u64 qpedx_rrrkey0;
121	u64 qpedx_rrva0;
122	u64 reserved2;
123	u64 qpedx_rrl1;
124	u64 qpedx_rrrkey1;
125	u64 qpedx_rrva1;
126	u64 reserved3;
127	u64 qpedx_rrl2;
128	u64 qpedx_rrrkey2;
129	u64 qpedx_rrva2;
130	u64 reserved4;
131	u64 qpedx_rrl3;
132	u64 qpedx_rrrkey3;
133	u64 qpedx_rrva3;
134};
135
136#define CQX_FECADDER EHEA_BMASK_IBM(32, 63)
137#define CQX_FEC_CQE_CNT EHEA_BMASK_IBM(32, 63)
138#define CQX_N1_GENERATE_COMP_EVENT EHEA_BMASK_IBM(0, 0)
139#define CQX_EP_EVENT_PENDING EHEA_BMASK_IBM(0, 0)
140
141#define CQTEMM_OFFSET(x) offsetof(struct ehea_cqtemm, x)
142
143struct ehea_cqtemm {
144	u64 cqx_hcr;
145	u64 cqx_c;
146	u64 cqx_herr;
147	u64 cqx_aer;
148	u64 cqx_ptp;
149	u64 cqx_tp;
150	u64 cqx_fec;
151	u64 cqx_feca;
152	u64 cqx_ep;
153	u64 cqx_eq;
154	u64 reserved1;
155	u64 cqx_n0;
156	u64 cqx_n1;
157	u64 reserved2[(0x1000 - 0x60) / 8];
158};
159
160#define EQTEMM_OFFSET(x) offsetof(struct ehea_eqtemm, x)
161
162struct ehea_eqtemm {
163	u64 eqx_hcr;
164	u64 eqx_c;
165	u64 eqx_herr;
166	u64 eqx_aer;
167	u64 eqx_ptp;
168	u64 eqx_tp;
169	u64 eqx_ssba;
170	u64 eqx_psba;
171	u64 eqx_cec;
172	u64 eqx_meql;
173	u64 eqx_xisbi;
174	u64 eqx_xisc;
175	u64 eqx_it;
176};
177
178/*
179 * These access functions will be changed when the dissuccsion about
180 * the new access methods for POWER has settled.
181 */
182
183static inline u64 epa_load(struct h_epa epa, u32 offset)
184{
185	return __raw_readq((void __iomem *)(epa.addr + offset));
186}
187
188static inline void epa_store(struct h_epa epa, u32 offset, u64 value)
189{
190	__raw_writeq(value, (void __iomem *)(epa.addr + offset));
191	epa_load(epa, offset);	/* synchronize explicitly to eHEA */
192}
193
194static inline void epa_store_acc(struct h_epa epa, u32 offset, u64 value)
195{
196	__raw_writeq(value, (void __iomem *)(epa.addr + offset));
197}
198
199#define epa_store_cq(epa, offset, value)\
200	epa_store(epa, CQTEMM_OFFSET(offset), value)
201#define epa_load_cq(epa, offset)\
202	epa_load(epa, CQTEMM_OFFSET(offset))
203
204static inline void ehea_update_sqa(struct ehea_qp *qp, u16 nr_wqes)
205{
206	struct h_epa epa = qp->epas.kernel;
207	epa_store_acc(epa, QPTEMM_OFFSET(qpx_sqa),
208		      EHEA_BMASK_SET(QPX_SQA_VALUE, nr_wqes));
209}
210
211static inline void ehea_update_rq3a(struct ehea_qp *qp, u16 nr_wqes)
212{
213	struct h_epa epa = qp->epas.kernel;
214	epa_store_acc(epa, QPTEMM_OFFSET(qpx_rq3a),
215		      EHEA_BMASK_SET(QPX_RQ1A_VALUE, nr_wqes));
216}
217
218static inline void ehea_update_rq2a(struct ehea_qp *qp, u16 nr_wqes)
219{
220	struct h_epa epa = qp->epas.kernel;
221	epa_store_acc(epa, QPTEMM_OFFSET(qpx_rq2a),
222		      EHEA_BMASK_SET(QPX_RQ2A_VALUE, nr_wqes));
223}
224
225static inline void ehea_update_rq1a(struct ehea_qp *qp, u16 nr_wqes)
226{
227	struct h_epa epa = qp->epas.kernel;
228	epa_store_acc(epa, QPTEMM_OFFSET(qpx_rq1a),
229		      EHEA_BMASK_SET(QPX_RQ3A_VALUE, nr_wqes));
230}
231
232static inline void ehea_update_feca(struct ehea_cq *cq, u32 nr_cqes)
233{
234	struct h_epa epa = cq->epas.kernel;
235	epa_store_acc(epa, CQTEMM_OFFSET(cqx_feca),
236		      EHEA_BMASK_SET(CQX_FECADDER, nr_cqes));
237}
238
239static inline void ehea_reset_cq_n1(struct ehea_cq *cq)
240{
241	struct h_epa epa = cq->epas.kernel;
242	epa_store_cq(epa, cqx_n1,
243		     EHEA_BMASK_SET(CQX_N1_GENERATE_COMP_EVENT, 1));
244}
245
246static inline void ehea_reset_cq_ep(struct ehea_cq *my_cq)
247{
248	struct h_epa epa = my_cq->epas.kernel;
249	epa_store_acc(epa, CQTEMM_OFFSET(cqx_ep),
250		      EHEA_BMASK_SET(CQX_EP_EVENT_PENDING, 0));
251}
252
253#endif	/* __EHEA_HW_H__ */