Loading...
1/******************************************************************************
2 *
3 * (C)Copyright 1998,1999 SysKonnect,
4 * a business unit of Schneider & Koch & Co. Datensysteme GmbH.
5 *
6 * See the file "skfddi.c" for further information.
7 *
8 * This program is free software; you can redistribute it and/or modify
9 * it under the terms of the GNU General Public License as published by
10 * the Free Software Foundation; either version 2 of the License, or
11 * (at your option) any later version.
12 *
13 * The information in this file is provided "AS IS" without warranty.
14 *
15 ******************************************************************************/
16
17#ifndef lint
18static char const ID_sccs[] = "@(#)hwmtm.c 1.40 99/05/31 (C) SK" ;
19#endif
20
21#define HWMTM
22
23#ifndef FDDI
24#define FDDI
25#endif
26
27#include "h/types.h"
28#include "h/fddi.h"
29#include "h/smc.h"
30#include "h/supern_2.h"
31#include "h/skfbiinc.h"
32
33/*
34 -------------------------------------------------------------
35 DOCUMENTATION
36 -------------------------------------------------------------
37 BEGIN_MANUAL_ENTRY(DOCUMENTATION)
38
39 T B D
40
41 END_MANUAL_ENTRY
42*/
43/*
44 -------------------------------------------------------------
45 LOCAL VARIABLES:
46 -------------------------------------------------------------
47*/
48#ifdef COMMON_MB_POOL
49static SMbuf *mb_start = 0 ;
50static SMbuf *mb_free = 0 ;
51static int mb_init = FALSE ;
52static int call_count = 0 ;
53#endif
54
55/*
56 -------------------------------------------------------------
57 EXTERNE VARIABLES:
58 -------------------------------------------------------------
59*/
60
61#ifdef DEBUG
62#ifndef DEBUG_BRD
63extern struct smt_debug debug ;
64#endif
65#endif
66
67#ifdef NDIS_OS2
68extern u_char offDepth ;
69extern u_char force_irq_pending ;
70#endif
71
72/*
73 -------------------------------------------------------------
74 LOCAL FUNCTIONS:
75 -------------------------------------------------------------
76*/
77
78static void queue_llc_rx(struct s_smc *smc, SMbuf *mb);
79static void smt_to_llc(struct s_smc *smc, SMbuf *mb);
80static void init_txd_ring(struct s_smc *smc);
81static void init_rxd_ring(struct s_smc *smc);
82static void queue_txd_mb(struct s_smc *smc, SMbuf *mb);
83static u_long init_descr_ring(struct s_smc *smc, union s_fp_descr volatile *start,
84 int count);
85static u_long repair_txd_ring(struct s_smc *smc, struct s_smt_tx_queue *queue);
86static u_long repair_rxd_ring(struct s_smc *smc, struct s_smt_rx_queue *queue);
87static SMbuf* get_llc_rx(struct s_smc *smc);
88static SMbuf* get_txd_mb(struct s_smc *smc);
89static void mac_drv_clear_txd(struct s_smc *smc);
90
91/*
92 -------------------------------------------------------------
93 EXTERNAL FUNCTIONS:
94 -------------------------------------------------------------
95*/
96/* The external SMT functions are listed in cmtdef.h */
97
98extern void* mac_drv_get_space(struct s_smc *smc, unsigned int size);
99extern void* mac_drv_get_desc_mem(struct s_smc *smc, unsigned int size);
100extern void mac_drv_fill_rxd(struct s_smc *smc);
101extern void mac_drv_tx_complete(struct s_smc *smc,
102 volatile struct s_smt_fp_txd *txd);
103extern void mac_drv_rx_complete(struct s_smc *smc,
104 volatile struct s_smt_fp_rxd *rxd,
105 int frag_count, int len);
106extern void mac_drv_requeue_rxd(struct s_smc *smc,
107 volatile struct s_smt_fp_rxd *rxd,
108 int frag_count);
109extern void mac_drv_clear_rxd(struct s_smc *smc,
110 volatile struct s_smt_fp_rxd *rxd, int frag_count);
111
112#ifdef USE_OS_CPY
113extern void hwm_cpy_rxd2mb(void);
114extern void hwm_cpy_txd2mb(void);
115#endif
116
117#ifdef ALL_RX_COMPLETE
118extern void mac_drv_all_receives_complete(void);
119#endif
120
121extern u_long mac_drv_virt2phys(struct s_smc *smc, void *virt);
122extern u_long dma_master(struct s_smc *smc, void *virt, int len, int flag);
123
124#ifdef NDIS_OS2
125extern void post_proc(void);
126#else
127extern void dma_complete(struct s_smc *smc, volatile union s_fp_descr *descr,
128 int flag);
129#endif
130
131extern int mac_drv_rx_init(struct s_smc *smc, int len, int fc, char *look_ahead,
132 int la_len);
133
134/*
135 -------------------------------------------------------------
136 PUBLIC FUNCTIONS:
137 -------------------------------------------------------------
138*/
139void process_receive(struct s_smc *smc);
140void fddi_isr(struct s_smc *smc);
141void smt_free_mbuf(struct s_smc *smc, SMbuf *mb);
142void init_driver_fplus(struct s_smc *smc);
143void mac_drv_rx_mode(struct s_smc *smc, int mode);
144void init_fddi_driver(struct s_smc *smc, u_char *mac_addr);
145void mac_drv_clear_tx_queue(struct s_smc *smc);
146void mac_drv_clear_rx_queue(struct s_smc *smc);
147void hwm_tx_frag(struct s_smc *smc, char far *virt, u_long phys, int len,
148 int frame_status);
149void hwm_rx_frag(struct s_smc *smc, char far *virt, u_long phys, int len,
150 int frame_status);
151
152int mac_drv_init(struct s_smc *smc);
153int hwm_tx_init(struct s_smc *smc, u_char fc, int frag_count, int frame_len,
154 int frame_status);
155
156u_int mac_drv_check_space(void);
157
158SMbuf* smt_get_mbuf(struct s_smc *smc);
159
160#ifdef DEBUG
161 void mac_drv_debug_lev(void);
162#endif
163
164/*
165 -------------------------------------------------------------
166 MACROS:
167 -------------------------------------------------------------
168*/
169#ifndef UNUSED
170#ifdef lint
171#define UNUSED(x) (x) = (x)
172#else
173#define UNUSED(x)
174#endif
175#endif
176
177#ifdef USE_CAN_ADDR
178#define MA smc->hw.fddi_canon_addr.a
179#define GROUP_ADDR_BIT 0x01
180#else
181#define MA smc->hw.fddi_home_addr.a
182#define GROUP_ADDR_BIT 0x80
183#endif
184
185#define RXD_TXD_COUNT (HWM_ASYNC_TXD_COUNT+HWM_SYNC_TXD_COUNT+\
186 SMT_R1_RXD_COUNT+SMT_R2_RXD_COUNT)
187
188#ifdef MB_OUTSIDE_SMC
189#define EXT_VIRT_MEM ((RXD_TXD_COUNT+1)*sizeof(struct s_smt_fp_txd) +\
190 MAX_MBUF*sizeof(SMbuf))
191#define EXT_VIRT_MEM_2 ((RXD_TXD_COUNT+1)*sizeof(struct s_smt_fp_txd))
192#else
193#define EXT_VIRT_MEM ((RXD_TXD_COUNT+1)*sizeof(struct s_smt_fp_txd))
194#endif
195
196 /*
197 * define critical read for 16 Bit drivers
198 */
199#if defined(NDIS_OS2) || defined(ODI2)
200#define CR_READ(var) ((var) & 0xffff0000 | ((var) & 0xffff))
201#else
202#define CR_READ(var) (__le32)(var)
203#endif
204
205#define IMASK_SLOW (IS_PLINT1 | IS_PLINT2 | IS_TIMINT | IS_TOKEN | \
206 IS_MINTR1 | IS_MINTR2 | IS_MINTR3 | IS_R1_P | \
207 IS_R1_C | IS_XA_C | IS_XS_C)
208
209/*
210 -------------------------------------------------------------
211 INIT- AND SMT FUNCTIONS:
212 -------------------------------------------------------------
213*/
214
215
216/*
217 * BEGIN_MANUAL_ENTRY(mac_drv_check_space)
218 * u_int mac_drv_check_space()
219 *
220 * function DOWNCALL (drvsr.c)
221 * This function calculates the needed non virtual
222 * memory for MBufs, RxD and TxD descriptors etc.
223 * needed by the driver.
224 *
225 * return u_int memory in bytes
226 *
227 * END_MANUAL_ENTRY
228 */
229u_int mac_drv_check_space(void)
230{
231#ifdef MB_OUTSIDE_SMC
232#ifdef COMMON_MB_POOL
233 call_count++ ;
234 if (call_count == 1) {
235 return EXT_VIRT_MEM;
236 }
237 else {
238 return EXT_VIRT_MEM_2;
239 }
240#else
241 return EXT_VIRT_MEM;
242#endif
243#else
244 return 0;
245#endif
246}
247
248/*
249 * BEGIN_MANUAL_ENTRY(mac_drv_init)
250 * void mac_drv_init(smc)
251 *
252 * function DOWNCALL (drvsr.c)
253 * In this function the hardware module allocates it's
254 * memory.
255 * The operating system dependent module should call
256 * mac_drv_init once, after the adatper is detected.
257 * END_MANUAL_ENTRY
258 */
259int mac_drv_init(struct s_smc *smc)
260{
261 if (sizeof(struct s_smt_fp_rxd) % 16) {
262 SMT_PANIC(smc,HWM_E0001,HWM_E0001_MSG) ;
263 }
264 if (sizeof(struct s_smt_fp_txd) % 16) {
265 SMT_PANIC(smc,HWM_E0002,HWM_E0002_MSG) ;
266 }
267
268 /*
269 * get the required memory for the RxDs and TxDs
270 */
271 if (!(smc->os.hwm.descr_p = (union s_fp_descr volatile *)
272 mac_drv_get_desc_mem(smc,(u_int)
273 (RXD_TXD_COUNT+1)*sizeof(struct s_smt_fp_txd)))) {
274 return 1; /* no space the hwm modul can't work */
275 }
276
277 /*
278 * get the memory for the SMT MBufs
279 */
280#ifndef MB_OUTSIDE_SMC
281 smc->os.hwm.mbuf_pool.mb_start=(SMbuf *)(&smc->os.hwm.mbuf_pool.mb[0]) ;
282#else
283#ifndef COMMON_MB_POOL
284 if (!(smc->os.hwm.mbuf_pool.mb_start = (SMbuf *) mac_drv_get_space(smc,
285 MAX_MBUF*sizeof(SMbuf)))) {
286 return 1; /* no space the hwm modul can't work */
287 }
288#else
289 if (!mb_start) {
290 if (!(mb_start = (SMbuf *) mac_drv_get_space(smc,
291 MAX_MBUF*sizeof(SMbuf)))) {
292 return 1; /* no space the hwm modul can't work */
293 }
294 }
295#endif
296#endif
297 return 0;
298}
299
300/*
301 * BEGIN_MANUAL_ENTRY(init_driver_fplus)
302 * init_driver_fplus(smc)
303 *
304 * Sets hardware modul specific values for the mode register 2
305 * (e.g. the byte alignment for the received frames, the position of the
306 * least significant byte etc.)
307 * END_MANUAL_ENTRY
308 */
309void init_driver_fplus(struct s_smc *smc)
310{
311 smc->hw.fp.mdr2init = FM_LSB | FM_BMMODE | FM_ENNPRQ | FM_ENHSRQ | 3 ;
312
313#ifdef PCI
314 smc->hw.fp.mdr2init |= FM_CHKPAR | FM_PARITY ;
315#endif
316 smc->hw.fp.mdr3init = FM_MENRQAUNLCK | FM_MENRS ;
317
318#ifdef USE_CAN_ADDR
319 /* enable address bit swapping */
320 smc->hw.fp.frselreg_init = FM_ENXMTADSWAP | FM_ENRCVADSWAP ;
321#endif
322}
323
324static u_long init_descr_ring(struct s_smc *smc,
325 union s_fp_descr volatile *start,
326 int count)
327{
328 int i ;
329 union s_fp_descr volatile *d1 ;
330 union s_fp_descr volatile *d2 ;
331 u_long phys ;
332
333 DB_GEN("descr ring starts at = %x ",(void *)start,0,3) ;
334 for (i=count-1, d1=start; i ; i--) {
335 d2 = d1 ;
336 d1++ ; /* descr is owned by the host */
337 d2->r.rxd_rbctrl = cpu_to_le32(BMU_CHECK) ;
338 d2->r.rxd_next = &d1->r ;
339 phys = mac_drv_virt2phys(smc,(void *)d1) ;
340 d2->r.rxd_nrdadr = cpu_to_le32(phys) ;
341 }
342 DB_GEN("descr ring ends at = %x ",(void *)d1,0,3) ;
343 d1->r.rxd_rbctrl = cpu_to_le32(BMU_CHECK) ;
344 d1->r.rxd_next = &start->r ;
345 phys = mac_drv_virt2phys(smc,(void *)start) ;
346 d1->r.rxd_nrdadr = cpu_to_le32(phys) ;
347
348 for (i=count, d1=start; i ; i--) {
349 DRV_BUF_FLUSH(&d1->r,DDI_DMA_SYNC_FORDEV) ;
350 d1++;
351 }
352 return phys;
353}
354
355static void init_txd_ring(struct s_smc *smc)
356{
357 struct s_smt_fp_txd volatile *ds ;
358 struct s_smt_tx_queue *queue ;
359 u_long phys ;
360
361 /*
362 * initialize the transmit descriptors
363 */
364 ds = (struct s_smt_fp_txd volatile *) ((char *)smc->os.hwm.descr_p +
365 SMT_R1_RXD_COUNT*sizeof(struct s_smt_fp_rxd)) ;
366 queue = smc->hw.fp.tx[QUEUE_A0] ;
367 DB_GEN("Init async TxD ring, %d TxDs ",HWM_ASYNC_TXD_COUNT,0,3) ;
368 (void)init_descr_ring(smc,(union s_fp_descr volatile *)ds,
369 HWM_ASYNC_TXD_COUNT) ;
370 phys = le32_to_cpu(ds->txd_ntdadr) ;
371 ds++ ;
372 queue->tx_curr_put = queue->tx_curr_get = ds ;
373 ds-- ;
374 queue->tx_free = HWM_ASYNC_TXD_COUNT ;
375 queue->tx_used = 0 ;
376 outpd(ADDR(B5_XA_DA),phys) ;
377
378 ds = (struct s_smt_fp_txd volatile *) ((char *)ds +
379 HWM_ASYNC_TXD_COUNT*sizeof(struct s_smt_fp_txd)) ;
380 queue = smc->hw.fp.tx[QUEUE_S] ;
381 DB_GEN("Init sync TxD ring, %d TxDs ",HWM_SYNC_TXD_COUNT,0,3) ;
382 (void)init_descr_ring(smc,(union s_fp_descr volatile *)ds,
383 HWM_SYNC_TXD_COUNT) ;
384 phys = le32_to_cpu(ds->txd_ntdadr) ;
385 ds++ ;
386 queue->tx_curr_put = queue->tx_curr_get = ds ;
387 queue->tx_free = HWM_SYNC_TXD_COUNT ;
388 queue->tx_used = 0 ;
389 outpd(ADDR(B5_XS_DA),phys) ;
390}
391
392static void init_rxd_ring(struct s_smc *smc)
393{
394 struct s_smt_fp_rxd volatile *ds ;
395 struct s_smt_rx_queue *queue ;
396 u_long phys ;
397
398 /*
399 * initialize the receive descriptors
400 */
401 ds = (struct s_smt_fp_rxd volatile *) smc->os.hwm.descr_p ;
402 queue = smc->hw.fp.rx[QUEUE_R1] ;
403 DB_GEN("Init RxD ring, %d RxDs ",SMT_R1_RXD_COUNT,0,3) ;
404 (void)init_descr_ring(smc,(union s_fp_descr volatile *)ds,
405 SMT_R1_RXD_COUNT) ;
406 phys = le32_to_cpu(ds->rxd_nrdadr) ;
407 ds++ ;
408 queue->rx_curr_put = queue->rx_curr_get = ds ;
409 queue->rx_free = SMT_R1_RXD_COUNT ;
410 queue->rx_used = 0 ;
411 outpd(ADDR(B4_R1_DA),phys) ;
412}
413
414/*
415 * BEGIN_MANUAL_ENTRY(init_fddi_driver)
416 * void init_fddi_driver(smc,mac_addr)
417 *
418 * initializes the driver and it's variables
419 *
420 * END_MANUAL_ENTRY
421 */
422void init_fddi_driver(struct s_smc *smc, u_char *mac_addr)
423{
424 SMbuf *mb ;
425 int i ;
426
427 init_board(smc,mac_addr) ;
428 (void)init_fplus(smc) ;
429
430 /*
431 * initialize the SMbufs for the SMT
432 */
433#ifndef COMMON_MB_POOL
434 mb = smc->os.hwm.mbuf_pool.mb_start ;
435 smc->os.hwm.mbuf_pool.mb_free = (SMbuf *)NULL ;
436 for (i = 0; i < MAX_MBUF; i++) {
437 mb->sm_use_count = 1 ;
438 smt_free_mbuf(smc,mb) ;
439 mb++ ;
440 }
441#else
442 mb = mb_start ;
443 if (!mb_init) {
444 mb_free = 0 ;
445 for (i = 0; i < MAX_MBUF; i++) {
446 mb->sm_use_count = 1 ;
447 smt_free_mbuf(smc,mb) ;
448 mb++ ;
449 }
450 mb_init = TRUE ;
451 }
452#endif
453
454 /*
455 * initialize the other variables
456 */
457 smc->os.hwm.llc_rx_pipe = smc->os.hwm.llc_rx_tail = (SMbuf *)NULL ;
458 smc->os.hwm.txd_tx_pipe = smc->os.hwm.txd_tx_tail = NULL ;
459 smc->os.hwm.pass_SMT = smc->os.hwm.pass_NSA = smc->os.hwm.pass_DB = 0 ;
460 smc->os.hwm.pass_llc_promisc = TRUE ;
461 smc->os.hwm.queued_rx_frames = smc->os.hwm.queued_txd_mb = 0 ;
462 smc->os.hwm.detec_count = 0 ;
463 smc->os.hwm.rx_break = 0 ;
464 smc->os.hwm.rx_len_error = 0 ;
465 smc->os.hwm.isr_flag = FALSE ;
466
467 /*
468 * make sure that the start pointer is 16 byte aligned
469 */
470 i = 16 - ((long)smc->os.hwm.descr_p & 0xf) ;
471 if (i != 16) {
472 DB_GEN("i = %d",i,0,3) ;
473 smc->os.hwm.descr_p = (union s_fp_descr volatile *)
474 ((char *)smc->os.hwm.descr_p+i) ;
475 }
476 DB_GEN("pt to descr area = %x",(void *)smc->os.hwm.descr_p,0,3) ;
477
478 init_txd_ring(smc) ;
479 init_rxd_ring(smc) ;
480 mac_drv_fill_rxd(smc) ;
481
482 init_plc(smc) ;
483}
484
485
486SMbuf *smt_get_mbuf(struct s_smc *smc)
487{
488 register SMbuf *mb ;
489
490#ifndef COMMON_MB_POOL
491 mb = smc->os.hwm.mbuf_pool.mb_free ;
492#else
493 mb = mb_free ;
494#endif
495 if (mb) {
496#ifndef COMMON_MB_POOL
497 smc->os.hwm.mbuf_pool.mb_free = mb->sm_next ;
498#else
499 mb_free = mb->sm_next ;
500#endif
501 mb->sm_off = 8 ;
502 mb->sm_use_count = 1 ;
503 }
504 DB_GEN("get SMbuf: mb = %x",(void *)mb,0,3) ;
505 return mb; /* May be NULL */
506}
507
508void smt_free_mbuf(struct s_smc *smc, SMbuf *mb)
509{
510
511 if (mb) {
512 mb->sm_use_count-- ;
513 DB_GEN("free_mbuf: sm_use_count = %d",mb->sm_use_count,0,3) ;
514 /*
515 * If the use_count is != zero the MBuf is queued
516 * more than once and must not queued into the
517 * free MBuf queue
518 */
519 if (!mb->sm_use_count) {
520 DB_GEN("free SMbuf: mb = %x",(void *)mb,0,3) ;
521#ifndef COMMON_MB_POOL
522 mb->sm_next = smc->os.hwm.mbuf_pool.mb_free ;
523 smc->os.hwm.mbuf_pool.mb_free = mb ;
524#else
525 mb->sm_next = mb_free ;
526 mb_free = mb ;
527#endif
528 }
529 }
530 else
531 SMT_PANIC(smc,HWM_E0003,HWM_E0003_MSG) ;
532}
533
534
535/*
536 * BEGIN_MANUAL_ENTRY(mac_drv_repair_descr)
537 * void mac_drv_repair_descr(smc)
538 *
539 * function called from SMT (HWM / hwmtm.c)
540 * The BMU is idle when this function is called.
541 * Mac_drv_repair_descr sets up the physical address
542 * for all receive and transmit queues where the BMU
543 * should continue.
544 * It may be that the BMU was reseted during a fragmented
545 * transfer. In this case there are some fragments which will
546 * never completed by the BMU. The OWN bit of this fragments
547 * must be switched to be owned by the host.
548 *
549 * Give a start command to the receive BMU.
550 * Start the transmit BMUs if transmit frames pending.
551 *
552 * END_MANUAL_ENTRY
553 */
554void mac_drv_repair_descr(struct s_smc *smc)
555{
556 u_long phys ;
557
558 if (smc->hw.hw_state != STOPPED) {
559 SK_BREAK() ;
560 SMT_PANIC(smc,HWM_E0013,HWM_E0013_MSG) ;
561 return ;
562 }
563
564 /*
565 * repair tx queues: don't start
566 */
567 phys = repair_txd_ring(smc,smc->hw.fp.tx[QUEUE_A0]) ;
568 outpd(ADDR(B5_XA_DA),phys) ;
569 if (smc->hw.fp.tx_q[QUEUE_A0].tx_used) {
570 outpd(ADDR(B0_XA_CSR),CSR_START) ;
571 }
572 phys = repair_txd_ring(smc,smc->hw.fp.tx[QUEUE_S]) ;
573 outpd(ADDR(B5_XS_DA),phys) ;
574 if (smc->hw.fp.tx_q[QUEUE_S].tx_used) {
575 outpd(ADDR(B0_XS_CSR),CSR_START) ;
576 }
577
578 /*
579 * repair rx queues
580 */
581 phys = repair_rxd_ring(smc,smc->hw.fp.rx[QUEUE_R1]) ;
582 outpd(ADDR(B4_R1_DA),phys) ;
583 outpd(ADDR(B0_R1_CSR),CSR_START) ;
584}
585
586static u_long repair_txd_ring(struct s_smc *smc, struct s_smt_tx_queue *queue)
587{
588 int i ;
589 int tx_used ;
590 u_long phys ;
591 u_long tbctrl ;
592 struct s_smt_fp_txd volatile *t ;
593
594 SK_UNUSED(smc) ;
595
596 t = queue->tx_curr_get ;
597 tx_used = queue->tx_used ;
598 for (i = tx_used+queue->tx_free-1 ; i ; i-- ) {
599 t = t->txd_next ;
600 }
601 phys = le32_to_cpu(t->txd_ntdadr) ;
602
603 t = queue->tx_curr_get ;
604 while (tx_used) {
605 DRV_BUF_FLUSH(t,DDI_DMA_SYNC_FORCPU) ;
606 tbctrl = le32_to_cpu(t->txd_tbctrl) ;
607
608 if (tbctrl & BMU_OWN) {
609 if (tbctrl & BMU_STF) {
610 break ; /* exit the loop */
611 }
612 else {
613 /*
614 * repair the descriptor
615 */
616 t->txd_tbctrl &= ~cpu_to_le32(BMU_OWN) ;
617 }
618 }
619 phys = le32_to_cpu(t->txd_ntdadr) ;
620 DRV_BUF_FLUSH(t,DDI_DMA_SYNC_FORDEV) ;
621 t = t->txd_next ;
622 tx_used-- ;
623 }
624 return phys;
625}
626
627/*
628 * Repairs the receive descriptor ring and returns the physical address
629 * where the BMU should continue working.
630 *
631 * o The physical address where the BMU was stopped has to be
632 * determined. This is the next RxD after rx_curr_get with an OWN
633 * bit set.
634 * o The BMU should start working at beginning of the next frame.
635 * RxDs with an OWN bit set but with a reset STF bit should be
636 * skipped and owned by the driver (OWN = 0).
637 */
638static u_long repair_rxd_ring(struct s_smc *smc, struct s_smt_rx_queue *queue)
639{
640 int i ;
641 int rx_used ;
642 u_long phys ;
643 u_long rbctrl ;
644 struct s_smt_fp_rxd volatile *r ;
645
646 SK_UNUSED(smc) ;
647
648 r = queue->rx_curr_get ;
649 rx_used = queue->rx_used ;
650 for (i = SMT_R1_RXD_COUNT-1 ; i ; i-- ) {
651 r = r->rxd_next ;
652 }
653 phys = le32_to_cpu(r->rxd_nrdadr) ;
654
655 r = queue->rx_curr_get ;
656 while (rx_used) {
657 DRV_BUF_FLUSH(r,DDI_DMA_SYNC_FORCPU) ;
658 rbctrl = le32_to_cpu(r->rxd_rbctrl) ;
659
660 if (rbctrl & BMU_OWN) {
661 if (rbctrl & BMU_STF) {
662 break ; /* exit the loop */
663 }
664 else {
665 /*
666 * repair the descriptor
667 */
668 r->rxd_rbctrl &= ~cpu_to_le32(BMU_OWN) ;
669 }
670 }
671 phys = le32_to_cpu(r->rxd_nrdadr) ;
672 DRV_BUF_FLUSH(r,DDI_DMA_SYNC_FORDEV) ;
673 r = r->rxd_next ;
674 rx_used-- ;
675 }
676 return phys;
677}
678
679
680/*
681 -------------------------------------------------------------
682 INTERRUPT SERVICE ROUTINE:
683 -------------------------------------------------------------
684*/
685
686/*
687 * BEGIN_MANUAL_ENTRY(fddi_isr)
688 * void fddi_isr(smc)
689 *
690 * function DOWNCALL (drvsr.c)
691 * interrupt service routine, handles the interrupt requests
692 * generated by the FDDI adapter.
693 *
694 * NOTE: The operating system dependent module must guarantee that the
695 * interrupts of the adapter are disabled when it calls fddi_isr.
696 *
697 * About the USE_BREAK_ISR mechanismn:
698 *
699 * The main requirement of this mechanismn is to force an timer IRQ when
700 * leaving process_receive() with leave_isr set. process_receive() may
701 * be called at any time from anywhere!
702 * To be sure we don't miss such event we set 'force_irq' per default.
703 * We have to force and Timer IRQ if 'smc->os.hwm.leave_isr' AND
704 * 'force_irq' are set. 'force_irq' may be reset if a receive complete
705 * IRQ is pending.
706 *
707 * END_MANUAL_ENTRY
708 */
709void fddi_isr(struct s_smc *smc)
710{
711 u_long is ; /* ISR source */
712 u_short stu, stl ;
713 SMbuf *mb ;
714
715#ifdef USE_BREAK_ISR
716 int force_irq ;
717#endif
718
719#ifdef ODI2
720 if (smc->os.hwm.rx_break) {
721 mac_drv_fill_rxd(smc) ;
722 if (smc->hw.fp.rx_q[QUEUE_R1].rx_used > 0) {
723 smc->os.hwm.rx_break = 0 ;
724 process_receive(smc) ;
725 }
726 else {
727 smc->os.hwm.detec_count = 0 ;
728 smt_force_irq(smc) ;
729 }
730 }
731#endif
732 smc->os.hwm.isr_flag = TRUE ;
733
734#ifdef USE_BREAK_ISR
735 force_irq = TRUE ;
736 if (smc->os.hwm.leave_isr) {
737 smc->os.hwm.leave_isr = FALSE ;
738 process_receive(smc) ;
739 }
740#endif
741
742 while ((is = GET_ISR() & ISR_MASK)) {
743 NDD_TRACE("CH0B",is,0,0) ;
744 DB_GEN("ISA = 0x%x",is,0,7) ;
745
746 if (is & IMASK_SLOW) {
747 NDD_TRACE("CH1b",is,0,0) ;
748 if (is & IS_PLINT1) { /* PLC1 */
749 plc1_irq(smc) ;
750 }
751 if (is & IS_PLINT2) { /* PLC2 */
752 plc2_irq(smc) ;
753 }
754 if (is & IS_MINTR1) { /* FORMAC+ STU1(U/L) */
755 stu = inpw(FM_A(FM_ST1U)) ;
756 stl = inpw(FM_A(FM_ST1L)) ;
757 DB_GEN("Slow transmit complete",0,0,6) ;
758 mac1_irq(smc,stu,stl) ;
759 }
760 if (is & IS_MINTR2) { /* FORMAC+ STU2(U/L) */
761 stu= inpw(FM_A(FM_ST2U)) ;
762 stl= inpw(FM_A(FM_ST2L)) ;
763 DB_GEN("Slow receive complete",0,0,6) ;
764 DB_GEN("stl = %x : stu = %x",stl,stu,7) ;
765 mac2_irq(smc,stu,stl) ;
766 }
767 if (is & IS_MINTR3) { /* FORMAC+ STU3(U/L) */
768 stu= inpw(FM_A(FM_ST3U)) ;
769 stl= inpw(FM_A(FM_ST3L)) ;
770 DB_GEN("FORMAC Mode Register 3",0,0,6) ;
771 mac3_irq(smc,stu,stl) ;
772 }
773 if (is & IS_TIMINT) { /* Timer 82C54-2 */
774 timer_irq(smc) ;
775#ifdef NDIS_OS2
776 force_irq_pending = 0 ;
777#endif
778 /*
779 * out of RxD detection
780 */
781 if (++smc->os.hwm.detec_count > 4) {
782 /*
783 * check out of RxD condition
784 */
785 process_receive(smc) ;
786 }
787 }
788 if (is & IS_TOKEN) { /* Restricted Token Monitor */
789 rtm_irq(smc) ;
790 }
791 if (is & IS_R1_P) { /* Parity error rx queue 1 */
792 /* clear IRQ */
793 outpd(ADDR(B4_R1_CSR),CSR_IRQ_CL_P) ;
794 SMT_PANIC(smc,HWM_E0004,HWM_E0004_MSG) ;
795 }
796 if (is & IS_R1_C) { /* Encoding error rx queue 1 */
797 /* clear IRQ */
798 outpd(ADDR(B4_R1_CSR),CSR_IRQ_CL_C) ;
799 SMT_PANIC(smc,HWM_E0005,HWM_E0005_MSG) ;
800 }
801 if (is & IS_XA_C) { /* Encoding error async tx q */
802 /* clear IRQ */
803 outpd(ADDR(B5_XA_CSR),CSR_IRQ_CL_C) ;
804 SMT_PANIC(smc,HWM_E0006,HWM_E0006_MSG) ;
805 }
806 if (is & IS_XS_C) { /* Encoding error sync tx q */
807 /* clear IRQ */
808 outpd(ADDR(B5_XS_CSR),CSR_IRQ_CL_C) ;
809 SMT_PANIC(smc,HWM_E0007,HWM_E0007_MSG) ;
810 }
811 }
812
813 /*
814 * Fast Tx complete Async/Sync Queue (BMU service)
815 */
816 if (is & (IS_XS_F|IS_XA_F)) {
817 DB_GEN("Fast tx complete queue",0,0,6) ;
818 /*
819 * clear IRQ, Note: no IRQ is lost, because
820 * we always service both queues
821 */
822 outpd(ADDR(B5_XS_CSR),CSR_IRQ_CL_F) ;
823 outpd(ADDR(B5_XA_CSR),CSR_IRQ_CL_F) ;
824 mac_drv_clear_txd(smc) ;
825 llc_restart_tx(smc) ;
826 }
827
828 /*
829 * Fast Rx Complete (BMU service)
830 */
831 if (is & IS_R1_F) {
832 DB_GEN("Fast receive complete",0,0,6) ;
833 /* clear IRQ */
834#ifndef USE_BREAK_ISR
835 outpd(ADDR(B4_R1_CSR),CSR_IRQ_CL_F) ;
836 process_receive(smc) ;
837#else
838 process_receive(smc) ;
839 if (smc->os.hwm.leave_isr) {
840 force_irq = FALSE ;
841 } else {
842 outpd(ADDR(B4_R1_CSR),CSR_IRQ_CL_F) ;
843 process_receive(smc) ;
844 }
845#endif
846 }
847
848#ifndef NDIS_OS2
849 while ((mb = get_llc_rx(smc))) {
850 smt_to_llc(smc,mb) ;
851 }
852#else
853 if (offDepth)
854 post_proc() ;
855
856 while (!offDepth && (mb = get_llc_rx(smc))) {
857 smt_to_llc(smc,mb) ;
858 }
859
860 if (!offDepth && smc->os.hwm.rx_break) {
861 process_receive(smc) ;
862 }
863#endif
864 if (smc->q.ev_get != smc->q.ev_put) {
865 NDD_TRACE("CH2a",0,0,0) ;
866 ev_dispatcher(smc) ;
867 }
868#ifdef NDIS_OS2
869 post_proc() ;
870 if (offDepth) { /* leave fddi_isr because */
871 break ; /* indications not allowed */
872 }
873#endif
874#ifdef USE_BREAK_ISR
875 if (smc->os.hwm.leave_isr) {
876 break ; /* leave fddi_isr */
877 }
878#endif
879
880 /* NOTE: when the isr is left, no rx is pending */
881 } /* end of interrupt source polling loop */
882
883#ifdef USE_BREAK_ISR
884 if (smc->os.hwm.leave_isr && force_irq) {
885 smt_force_irq(smc) ;
886 }
887#endif
888 smc->os.hwm.isr_flag = FALSE ;
889 NDD_TRACE("CH0E",0,0,0) ;
890}
891
892
893/*
894 -------------------------------------------------------------
895 RECEIVE FUNCTIONS:
896 -------------------------------------------------------------
897*/
898
899#ifndef NDIS_OS2
900/*
901 * BEGIN_MANUAL_ENTRY(mac_drv_rx_mode)
902 * void mac_drv_rx_mode(smc,mode)
903 *
904 * function DOWNCALL (fplus.c)
905 * Corresponding to the parameter mode, the operating system
906 * dependent module can activate several receive modes.
907 *
908 * para mode = 1: RX_ENABLE_ALLMULTI enable all multicasts
909 * = 2: RX_DISABLE_ALLMULTI disable "enable all multicasts"
910 * = 3: RX_ENABLE_PROMISC enable promiscuous
911 * = 4: RX_DISABLE_PROMISC disable promiscuous
912 * = 5: RX_ENABLE_NSA enable rec. of all NSA frames
913 * (disabled after 'driver reset' & 'set station address')
914 * = 6: RX_DISABLE_NSA disable rec. of all NSA frames
915 *
916 * = 21: RX_ENABLE_PASS_SMT ( see description )
917 * = 22: RX_DISABLE_PASS_SMT ( " " )
918 * = 23: RX_ENABLE_PASS_NSA ( " " )
919 * = 24: RX_DISABLE_PASS_NSA ( " " )
920 * = 25: RX_ENABLE_PASS_DB ( " " )
921 * = 26: RX_DISABLE_PASS_DB ( " " )
922 * = 27: RX_DISABLE_PASS_ALL ( " " )
923 * = 28: RX_DISABLE_LLC_PROMISC ( " " )
924 * = 29: RX_ENABLE_LLC_PROMISC ( " " )
925 *
926 *
927 * RX_ENABLE_PASS_SMT / RX_DISABLE_PASS_SMT
928 *
929 * If the operating system dependent module activates the
930 * mode RX_ENABLE_PASS_SMT, the hardware module
931 * duplicates all SMT frames with the frame control
932 * FC_SMT_INFO and passes them to the LLC receive channel
933 * by calling mac_drv_rx_init.
934 * The SMT Frames which are sent by the local SMT and the NSA
935 * frames whose A- and C-Indicator is not set are also duplicated
936 * and passed.
937 * The receive mode RX_DISABLE_PASS_SMT disables the passing
938 * of SMT frames.
939 *
940 * RX_ENABLE_PASS_NSA / RX_DISABLE_PASS_NSA
941 *
942 * If the operating system dependent module activates the
943 * mode RX_ENABLE_PASS_NSA, the hardware module
944 * duplicates all NSA frames with frame control FC_SMT_NSA
945 * and a set A-Indicator and passed them to the LLC
946 * receive channel by calling mac_drv_rx_init.
947 * All NSA Frames which are sent by the local SMT
948 * are also duplicated and passed.
949 * The receive mode RX_DISABLE_PASS_NSA disables the passing
950 * of NSA frames with the A- or C-Indicator set.
951 *
952 * NOTE: For fear that the hardware module receives NSA frames with
953 * a reset A-Indicator, the operating system dependent module
954 * has to call mac_drv_rx_mode with the mode RX_ENABLE_NSA
955 * before activate the RX_ENABLE_PASS_NSA mode and after every
956 * 'driver reset' and 'set station address'.
957 *
958 * RX_ENABLE_PASS_DB / RX_DISABLE_PASS_DB
959 *
960 * If the operating system dependent module activates the
961 * mode RX_ENABLE_PASS_DB, direct BEACON frames
962 * (FC_BEACON frame control) are passed to the LLC receive
963 * channel by mac_drv_rx_init.
964 * The receive mode RX_DISABLE_PASS_DB disables the passing
965 * of direct BEACON frames.
966 *
967 * RX_DISABLE_PASS_ALL
968 *
969 * Disables all special receives modes. It is equal to
970 * call mac_drv_set_rx_mode successively with the
971 * parameters RX_DISABLE_NSA, RX_DISABLE_PASS_SMT,
972 * RX_DISABLE_PASS_NSA and RX_DISABLE_PASS_DB.
973 *
974 * RX_ENABLE_LLC_PROMISC
975 *
976 * (default) all received LLC frames and all SMT/NSA/DBEACON
977 * frames depending on the attitude of the flags
978 * PASS_SMT/PASS_NSA/PASS_DBEACON will be delivered to the
979 * LLC layer
980 *
981 * RX_DISABLE_LLC_PROMISC
982 *
983 * all received SMT/NSA/DBEACON frames depending on the
984 * attitude of the flags PASS_SMT/PASS_NSA/PASS_DBEACON
985 * will be delivered to the LLC layer.
986 * all received LLC frames with a directed address, Multicast
987 * or Broadcast address will be delivered to the LLC
988 * layer too.
989 *
990 * END_MANUAL_ENTRY
991 */
992void mac_drv_rx_mode(struct s_smc *smc, int mode)
993{
994 switch(mode) {
995 case RX_ENABLE_PASS_SMT:
996 smc->os.hwm.pass_SMT = TRUE ;
997 break ;
998 case RX_DISABLE_PASS_SMT:
999 smc->os.hwm.pass_SMT = FALSE ;
1000 break ;
1001 case RX_ENABLE_PASS_NSA:
1002 smc->os.hwm.pass_NSA = TRUE ;
1003 break ;
1004 case RX_DISABLE_PASS_NSA:
1005 smc->os.hwm.pass_NSA = FALSE ;
1006 break ;
1007 case RX_ENABLE_PASS_DB:
1008 smc->os.hwm.pass_DB = TRUE ;
1009 break ;
1010 case RX_DISABLE_PASS_DB:
1011 smc->os.hwm.pass_DB = FALSE ;
1012 break ;
1013 case RX_DISABLE_PASS_ALL:
1014 smc->os.hwm.pass_SMT = smc->os.hwm.pass_NSA = FALSE ;
1015 smc->os.hwm.pass_DB = FALSE ;
1016 smc->os.hwm.pass_llc_promisc = TRUE ;
1017 mac_set_rx_mode(smc,RX_DISABLE_NSA) ;
1018 break ;
1019 case RX_DISABLE_LLC_PROMISC:
1020 smc->os.hwm.pass_llc_promisc = FALSE ;
1021 break ;
1022 case RX_ENABLE_LLC_PROMISC:
1023 smc->os.hwm.pass_llc_promisc = TRUE ;
1024 break ;
1025 case RX_ENABLE_ALLMULTI:
1026 case RX_DISABLE_ALLMULTI:
1027 case RX_ENABLE_PROMISC:
1028 case RX_DISABLE_PROMISC:
1029 case RX_ENABLE_NSA:
1030 case RX_DISABLE_NSA:
1031 default:
1032 mac_set_rx_mode(smc,mode) ;
1033 break ;
1034 }
1035}
1036#endif /* ifndef NDIS_OS2 */
1037
1038/*
1039 * process receive queue
1040 */
1041void process_receive(struct s_smc *smc)
1042{
1043 int i ;
1044 int n ;
1045 int frag_count ; /* number of RxDs of the curr rx buf */
1046 int used_frags ; /* number of RxDs of the curr frame */
1047 struct s_smt_rx_queue *queue ; /* points to the queue ctl struct */
1048 struct s_smt_fp_rxd volatile *r ; /* rxd pointer */
1049 struct s_smt_fp_rxd volatile *rxd ; /* first rxd of rx frame */
1050 u_long rbctrl ; /* receive buffer control word */
1051 u_long rfsw ; /* receive frame status word */
1052 u_short rx_used ;
1053 u_char far *virt ;
1054 char far *data ;
1055 SMbuf *mb ;
1056 u_char fc ; /* Frame control */
1057 int len ; /* Frame length */
1058
1059 smc->os.hwm.detec_count = 0 ;
1060 queue = smc->hw.fp.rx[QUEUE_R1] ;
1061 NDD_TRACE("RHxB",0,0,0) ;
1062 for ( ; ; ) {
1063 r = queue->rx_curr_get ;
1064 rx_used = queue->rx_used ;
1065 frag_count = 0 ;
1066
1067#ifdef USE_BREAK_ISR
1068 if (smc->os.hwm.leave_isr) {
1069 goto rx_end ;
1070 }
1071#endif
1072#ifdef NDIS_OS2
1073 if (offDepth) {
1074 smc->os.hwm.rx_break = 1 ;
1075 goto rx_end ;
1076 }
1077 smc->os.hwm.rx_break = 0 ;
1078#endif
1079#ifdef ODI2
1080 if (smc->os.hwm.rx_break) {
1081 goto rx_end ;
1082 }
1083#endif
1084 n = 0 ;
1085 do {
1086 DB_RX("Check RxD %x for OWN and EOF",(void *)r,0,5) ;
1087 DRV_BUF_FLUSH(r,DDI_DMA_SYNC_FORCPU) ;
1088 rbctrl = le32_to_cpu(CR_READ(r->rxd_rbctrl));
1089
1090 if (rbctrl & BMU_OWN) {
1091 NDD_TRACE("RHxE",r,rfsw,rbctrl) ;
1092 DB_RX("End of RxDs",0,0,4) ;
1093 goto rx_end ;
1094 }
1095 /*
1096 * out of RxD detection
1097 */
1098 if (!rx_used) {
1099 SK_BREAK() ;
1100 SMT_PANIC(smc,HWM_E0009,HWM_E0009_MSG) ;
1101 /* Either we don't have an RxD or all
1102 * RxDs are filled. Therefore it's allowed
1103 * for to set the STOPPED flag */
1104 smc->hw.hw_state = STOPPED ;
1105 mac_drv_clear_rx_queue(smc) ;
1106 smc->hw.hw_state = STARTED ;
1107 mac_drv_fill_rxd(smc) ;
1108 smc->os.hwm.detec_count = 0 ;
1109 goto rx_end ;
1110 }
1111 rfsw = le32_to_cpu(r->rxd_rfsw) ;
1112 if ((rbctrl & BMU_STF) != ((rbctrl & BMU_ST_BUF) <<5)) {
1113 /*
1114 * The BMU_STF bit is deleted, 1 frame is
1115 * placed into more than 1 rx buffer
1116 *
1117 * skip frame by setting the rx len to 0
1118 *
1119 * if fragment count == 0
1120 * The missing STF bit belongs to the
1121 * current frame, search for the
1122 * EOF bit to complete the frame
1123 * else
1124 * the fragment belongs to the next frame,
1125 * exit the loop and process the frame
1126 */
1127 SK_BREAK() ;
1128 rfsw = 0 ;
1129 if (frag_count) {
1130 break ;
1131 }
1132 }
1133 n += rbctrl & 0xffff ;
1134 r = r->rxd_next ;
1135 frag_count++ ;
1136 rx_used-- ;
1137 } while (!(rbctrl & BMU_EOF)) ;
1138 used_frags = frag_count ;
1139 DB_RX("EOF set in RxD, used_frags = %d ",used_frags,0,5) ;
1140
1141 /* may be next 2 DRV_BUF_FLUSH() can be skipped, because */
1142 /* BMU_ST_BUF will not be changed by the ASIC */
1143 DRV_BUF_FLUSH(r,DDI_DMA_SYNC_FORCPU) ;
1144 while (rx_used && !(r->rxd_rbctrl & cpu_to_le32(BMU_ST_BUF))) {
1145 DB_RX("Check STF bit in %x",(void *)r,0,5) ;
1146 r = r->rxd_next ;
1147 DRV_BUF_FLUSH(r,DDI_DMA_SYNC_FORCPU) ;
1148 frag_count++ ;
1149 rx_used-- ;
1150 }
1151 DB_RX("STF bit found",0,0,5) ;
1152
1153 /*
1154 * The received frame is finished for the process receive
1155 */
1156 rxd = queue->rx_curr_get ;
1157 queue->rx_curr_get = r ;
1158 queue->rx_free += frag_count ;
1159 queue->rx_used = rx_used ;
1160
1161 /*
1162 * ASIC Errata no. 7 (STF - Bit Bug)
1163 */
1164 rxd->rxd_rbctrl &= cpu_to_le32(~BMU_STF) ;
1165
1166 for (r=rxd, i=frag_count ; i ; r=r->rxd_next, i--){
1167 DB_RX("dma_complete for RxD %x",(void *)r,0,5) ;
1168 dma_complete(smc,(union s_fp_descr volatile *)r,DMA_WR);
1169 }
1170 smc->hw.fp.err_stats.err_valid++ ;
1171 smc->mib.m[MAC0].fddiMACCopied_Ct++ ;
1172
1173 /* the length of the data including the FC */
1174 len = (rfsw & RD_LENGTH) - 4 ;
1175
1176 DB_RX("frame length = %d",len,0,4) ;
1177 /*
1178 * check the frame_length and all error flags
1179 */
1180 if (rfsw & (RX_MSRABT|RX_FS_E|RX_FS_CRC|RX_FS_IMPL)){
1181 if (rfsw & RD_S_MSRABT) {
1182 DB_RX("Frame aborted by the FORMAC",0,0,2) ;
1183 smc->hw.fp.err_stats.err_abort++ ;
1184 }
1185 /*
1186 * check frame status
1187 */
1188 if (rfsw & RD_S_SEAC2) {
1189 DB_RX("E-Indicator set",0,0,2) ;
1190 smc->hw.fp.err_stats.err_e_indicator++ ;
1191 }
1192 if (rfsw & RD_S_SFRMERR) {
1193 DB_RX("CRC error",0,0,2) ;
1194 smc->hw.fp.err_stats.err_crc++ ;
1195 }
1196 if (rfsw & RX_FS_IMPL) {
1197 DB_RX("Implementer frame",0,0,2) ;
1198 smc->hw.fp.err_stats.err_imp_frame++ ;
1199 }
1200 goto abort_frame ;
1201 }
1202 if (len > FDDI_RAW_MTU-4) {
1203 DB_RX("Frame too long error",0,0,2) ;
1204 smc->hw.fp.err_stats.err_too_long++ ;
1205 goto abort_frame ;
1206 }
1207 /*
1208 * SUPERNET 3 Bug: FORMAC delivers status words
1209 * of aborded frames to the BMU
1210 */
1211 if (len <= 4) {
1212 DB_RX("Frame length = 0",0,0,2) ;
1213 goto abort_frame ;
1214 }
1215
1216 if (len != (n-4)) {
1217 DB_RX("BMU: rx len differs: [%d:%d]",len,n,4);
1218 smc->os.hwm.rx_len_error++ ;
1219 goto abort_frame ;
1220 }
1221
1222 /*
1223 * Check SA == MA
1224 */
1225 virt = (u_char far *) rxd->rxd_virt ;
1226 DB_RX("FC = %x",*virt,0,2) ;
1227 if (virt[12] == MA[5] &&
1228 virt[11] == MA[4] &&
1229 virt[10] == MA[3] &&
1230 virt[9] == MA[2] &&
1231 virt[8] == MA[1] &&
1232 (virt[7] & ~GROUP_ADDR_BIT) == MA[0]) {
1233 goto abort_frame ;
1234 }
1235
1236 /*
1237 * test if LLC frame
1238 */
1239 if (rfsw & RX_FS_LLC) {
1240 /*
1241 * if pass_llc_promisc is disable
1242 * if DA != Multicast or Broadcast or DA!=MA
1243 * abort the frame
1244 */
1245 if (!smc->os.hwm.pass_llc_promisc) {
1246 if(!(virt[1] & GROUP_ADDR_BIT)) {
1247 if (virt[6] != MA[5] ||
1248 virt[5] != MA[4] ||
1249 virt[4] != MA[3] ||
1250 virt[3] != MA[2] ||
1251 virt[2] != MA[1] ||
1252 virt[1] != MA[0]) {
1253 DB_RX("DA != MA and not multi- or broadcast",0,0,2) ;
1254 goto abort_frame ;
1255 }
1256 }
1257 }
1258
1259 /*
1260 * LLC frame received
1261 */
1262 DB_RX("LLC - receive",0,0,4) ;
1263 mac_drv_rx_complete(smc,rxd,frag_count,len) ;
1264 }
1265 else {
1266 if (!(mb = smt_get_mbuf(smc))) {
1267 smc->hw.fp.err_stats.err_no_buf++ ;
1268 DB_RX("No SMbuf; receive terminated",0,0,4) ;
1269 goto abort_frame ;
1270 }
1271 data = smtod(mb,char *) - 1 ;
1272
1273 /*
1274 * copy the frame into a SMT_MBuf
1275 */
1276#ifdef USE_OS_CPY
1277 hwm_cpy_rxd2mb(rxd,data,len) ;
1278#else
1279 for (r=rxd, i=used_frags ; i ; r=r->rxd_next, i--){
1280 n = le32_to_cpu(r->rxd_rbctrl) & RD_LENGTH ;
1281 DB_RX("cp SMT frame to mb: len = %d",n,0,6) ;
1282 memcpy(data,r->rxd_virt,n) ;
1283 data += n ;
1284 }
1285 data = smtod(mb,char *) - 1 ;
1286#endif
1287 fc = *(char *)mb->sm_data = *data ;
1288 mb->sm_len = len - 1 ; /* len - fc */
1289 data++ ;
1290
1291 /*
1292 * SMT frame received
1293 */
1294 switch(fc) {
1295 case FC_SMT_INFO :
1296 smc->hw.fp.err_stats.err_smt_frame++ ;
1297 DB_RX("SMT frame received ",0,0,5) ;
1298
1299 if (smc->os.hwm.pass_SMT) {
1300 DB_RX("pass SMT frame ",0,0,5) ;
1301 mac_drv_rx_complete(smc, rxd,
1302 frag_count,len) ;
1303 }
1304 else {
1305 DB_RX("requeue RxD",0,0,5) ;
1306 mac_drv_requeue_rxd(smc,rxd,frag_count);
1307 }
1308
1309 smt_received_pack(smc,mb,(int)(rfsw>>25)) ;
1310 break ;
1311 case FC_SMT_NSA :
1312 smc->hw.fp.err_stats.err_smt_frame++ ;
1313 DB_RX("SMT frame received ",0,0,5) ;
1314
1315 /* if pass_NSA set pass the NSA frame or */
1316 /* pass_SMT set and the A-Indicator */
1317 /* is not set, pass the NSA frame */
1318 if (smc->os.hwm.pass_NSA ||
1319 (smc->os.hwm.pass_SMT &&
1320 !(rfsw & A_INDIC))) {
1321 DB_RX("pass SMT frame ",0,0,5) ;
1322 mac_drv_rx_complete(smc, rxd,
1323 frag_count,len) ;
1324 }
1325 else {
1326 DB_RX("requeue RxD",0,0,5) ;
1327 mac_drv_requeue_rxd(smc,rxd,frag_count);
1328 }
1329
1330 smt_received_pack(smc,mb,(int)(rfsw>>25)) ;
1331 break ;
1332 case FC_BEACON :
1333 if (smc->os.hwm.pass_DB) {
1334 DB_RX("pass DB frame ",0,0,5) ;
1335 mac_drv_rx_complete(smc, rxd,
1336 frag_count,len) ;
1337 }
1338 else {
1339 DB_RX("requeue RxD",0,0,5) ;
1340 mac_drv_requeue_rxd(smc,rxd,frag_count);
1341 }
1342 smt_free_mbuf(smc,mb) ;
1343 break ;
1344 default :
1345 /*
1346 * unknown FC abord the frame
1347 */
1348 DB_RX("unknown FC error",0,0,2) ;
1349 smt_free_mbuf(smc,mb) ;
1350 DB_RX("requeue RxD",0,0,5) ;
1351 mac_drv_requeue_rxd(smc,rxd,frag_count) ;
1352 if ((fc & 0xf0) == FC_MAC)
1353 smc->hw.fp.err_stats.err_mac_frame++ ;
1354 else
1355 smc->hw.fp.err_stats.err_imp_frame++ ;
1356
1357 break ;
1358 }
1359 }
1360
1361 DB_RX("next RxD is %x ",queue->rx_curr_get,0,3) ;
1362 NDD_TRACE("RHx1",queue->rx_curr_get,0,0) ;
1363
1364 continue ;
1365 /*--------------------------------------------------------------------*/
1366abort_frame:
1367 DB_RX("requeue RxD",0,0,5) ;
1368 mac_drv_requeue_rxd(smc,rxd,frag_count) ;
1369
1370 DB_RX("next RxD is %x ",queue->rx_curr_get,0,3) ;
1371 NDD_TRACE("RHx2",queue->rx_curr_get,0,0) ;
1372 }
1373rx_end:
1374#ifdef ALL_RX_COMPLETE
1375 mac_drv_all_receives_complete(smc) ;
1376#endif
1377 return ; /* lint bug: needs return detect end of function */
1378}
1379
1380static void smt_to_llc(struct s_smc *smc, SMbuf *mb)
1381{
1382 u_char fc ;
1383
1384 DB_RX("send a queued frame to the llc layer",0,0,4) ;
1385 smc->os.hwm.r.len = mb->sm_len ;
1386 smc->os.hwm.r.mb_pos = smtod(mb,char *) ;
1387 fc = *smc->os.hwm.r.mb_pos ;
1388 (void)mac_drv_rx_init(smc,(int)mb->sm_len,(int)fc,
1389 smc->os.hwm.r.mb_pos,(int)mb->sm_len) ;
1390 smt_free_mbuf(smc,mb) ;
1391}
1392
1393/*
1394 * BEGIN_MANUAL_ENTRY(hwm_rx_frag)
1395 * void hwm_rx_frag(smc,virt,phys,len,frame_status)
1396 *
1397 * function MACRO (hardware module, hwmtm.h)
1398 * This function calls dma_master for preparing the
1399 * system hardware for the DMA transfer and initializes
1400 * the current RxD with the length and the physical and
1401 * virtual address of the fragment. Furthermore, it sets the
1402 * STF and EOF bits depending on the frame status byte,
1403 * switches the OWN flag of the RxD, so that it is owned by the
1404 * adapter and issues an rx_start.
1405 *
1406 * para virt virtual pointer to the fragment
1407 * len the length of the fragment
1408 * frame_status status of the frame, see design description
1409 *
1410 * NOTE: It is possible to call this function with a fragment length
1411 * of zero.
1412 *
1413 * END_MANUAL_ENTRY
1414 */
1415void hwm_rx_frag(struct s_smc *smc, char far *virt, u_long phys, int len,
1416 int frame_status)
1417{
1418 struct s_smt_fp_rxd volatile *r ;
1419 __le32 rbctrl;
1420
1421 NDD_TRACE("RHfB",virt,len,frame_status) ;
1422 DB_RX("hwm_rx_frag: len = %d, frame_status = %x\n",len,frame_status,2) ;
1423 r = smc->hw.fp.rx_q[QUEUE_R1].rx_curr_put ;
1424 r->rxd_virt = virt ;
1425 r->rxd_rbadr = cpu_to_le32(phys) ;
1426 rbctrl = cpu_to_le32( (((__u32)frame_status &
1427 (FIRST_FRAG|LAST_FRAG))<<26) |
1428 (((u_long) frame_status & FIRST_FRAG) << 21) |
1429 BMU_OWN | BMU_CHECK | BMU_EN_IRQ_EOF | len) ;
1430 r->rxd_rbctrl = rbctrl ;
1431
1432 DRV_BUF_FLUSH(r,DDI_DMA_SYNC_FORDEV) ;
1433 outpd(ADDR(B0_R1_CSR),CSR_START) ;
1434 smc->hw.fp.rx_q[QUEUE_R1].rx_free-- ;
1435 smc->hw.fp.rx_q[QUEUE_R1].rx_used++ ;
1436 smc->hw.fp.rx_q[QUEUE_R1].rx_curr_put = r->rxd_next ;
1437 NDD_TRACE("RHfE",r,le32_to_cpu(r->rxd_rbadr),0) ;
1438}
1439
1440/*
1441 * BEGINN_MANUAL_ENTRY(mac_drv_clear_rx_queue)
1442 *
1443 * void mac_drv_clear_rx_queue(smc)
1444 * struct s_smc *smc ;
1445 *
1446 * function DOWNCALL (hardware module, hwmtm.c)
1447 * mac_drv_clear_rx_queue is called by the OS-specific module
1448 * after it has issued a card_stop.
1449 * In this case, the frames in the receive queue are obsolete and
1450 * should be removed. For removing mac_drv_clear_rx_queue
1451 * calls dma_master for each RxD and mac_drv_clear_rxd for each
1452 * receive buffer.
1453 *
1454 * NOTE: calling sequence card_stop:
1455 * CLI_FBI(), card_stop(),
1456 * mac_drv_clear_tx_queue(), mac_drv_clear_rx_queue(),
1457 *
1458 * NOTE: The caller is responsible that the BMUs are idle
1459 * when this function is called.
1460 *
1461 * END_MANUAL_ENTRY
1462 */
1463void mac_drv_clear_rx_queue(struct s_smc *smc)
1464{
1465 struct s_smt_fp_rxd volatile *r ;
1466 struct s_smt_fp_rxd volatile *next_rxd ;
1467 struct s_smt_rx_queue *queue ;
1468 int frag_count ;
1469 int i ;
1470
1471 if (smc->hw.hw_state != STOPPED) {
1472 SK_BREAK() ;
1473 SMT_PANIC(smc,HWM_E0012,HWM_E0012_MSG) ;
1474 return ;
1475 }
1476
1477 queue = smc->hw.fp.rx[QUEUE_R1] ;
1478 DB_RX("clear_rx_queue",0,0,5) ;
1479
1480 /*
1481 * dma_complete and mac_drv_clear_rxd for all RxDs / receive buffers
1482 */
1483 r = queue->rx_curr_get ;
1484 while (queue->rx_used) {
1485 DRV_BUF_FLUSH(r,DDI_DMA_SYNC_FORCPU) ;
1486 DB_RX("switch OWN bit of RxD 0x%x ",r,0,5) ;
1487 r->rxd_rbctrl &= ~cpu_to_le32(BMU_OWN) ;
1488 frag_count = 1 ;
1489 DRV_BUF_FLUSH(r,DDI_DMA_SYNC_FORDEV) ;
1490 r = r->rxd_next ;
1491 DRV_BUF_FLUSH(r,DDI_DMA_SYNC_FORCPU) ;
1492 while (r != queue->rx_curr_put &&
1493 !(r->rxd_rbctrl & cpu_to_le32(BMU_ST_BUF))) {
1494 DB_RX("Check STF bit in %x",(void *)r,0,5) ;
1495 r->rxd_rbctrl &= ~cpu_to_le32(BMU_OWN) ;
1496 DRV_BUF_FLUSH(r,DDI_DMA_SYNC_FORDEV) ;
1497 r = r->rxd_next ;
1498 DRV_BUF_FLUSH(r,DDI_DMA_SYNC_FORCPU) ;
1499 frag_count++ ;
1500 }
1501 DB_RX("STF bit found",0,0,5) ;
1502 next_rxd = r ;
1503
1504 for (r=queue->rx_curr_get,i=frag_count; i ; r=r->rxd_next,i--){
1505 DB_RX("dma_complete for RxD %x",(void *)r,0,5) ;
1506 dma_complete(smc,(union s_fp_descr volatile *)r,DMA_WR);
1507 }
1508
1509 DB_RX("mac_drv_clear_rxd: RxD %x frag_count %d ",
1510 (void *)queue->rx_curr_get,frag_count,5) ;
1511 mac_drv_clear_rxd(smc,queue->rx_curr_get,frag_count) ;
1512
1513 queue->rx_curr_get = next_rxd ;
1514 queue->rx_used -= frag_count ;
1515 queue->rx_free += frag_count ;
1516 }
1517}
1518
1519
1520/*
1521 -------------------------------------------------------------
1522 SEND FUNCTIONS:
1523 -------------------------------------------------------------
1524*/
1525
1526/*
1527 * BEGIN_MANUAL_ENTRY(hwm_tx_init)
1528 * int hwm_tx_init(smc,fc,frag_count,frame_len,frame_status)
1529 *
1530 * function DOWN_CALL (hardware module, hwmtm.c)
1531 * hwm_tx_init checks if the frame can be sent through the
1532 * corresponding send queue.
1533 *
1534 * para fc the frame control. To determine through which
1535 * send queue the frame should be transmitted.
1536 * 0x50 - 0x57: asynchronous LLC frame
1537 * 0xD0 - 0xD7: synchronous LLC frame
1538 * 0x41, 0x4F: SMT frame to the network
1539 * 0x42: SMT frame to the network and to the local SMT
1540 * 0x43: SMT frame to the local SMT
1541 * frag_count count of the fragments for this frame
1542 * frame_len length of the frame
1543 * frame_status status of the frame, the send queue bit is already
1544 * specified
1545 *
1546 * return frame_status
1547 *
1548 * END_MANUAL_ENTRY
1549 */
1550int hwm_tx_init(struct s_smc *smc, u_char fc, int frag_count, int frame_len,
1551 int frame_status)
1552{
1553 NDD_TRACE("THiB",fc,frag_count,frame_len) ;
1554 smc->os.hwm.tx_p = smc->hw.fp.tx[frame_status & QUEUE_A0] ;
1555 smc->os.hwm.tx_descr = TX_DESCRIPTOR | (((u_long)(frame_len-1)&3)<<27) ;
1556 smc->os.hwm.tx_len = frame_len ;
1557 DB_TX("hwm_tx_init: fc = %x, len = %d",fc,frame_len,3) ;
1558 if ((fc & ~(FC_SYNC_BIT|FC_LLC_PRIOR)) == FC_ASYNC_LLC) {
1559 frame_status |= LAN_TX ;
1560 }
1561 else {
1562 switch (fc) {
1563 case FC_SMT_INFO :
1564 case FC_SMT_NSA :
1565 frame_status |= LAN_TX ;
1566 break ;
1567 case FC_SMT_LOC :
1568 frame_status |= LOC_TX ;
1569 break ;
1570 case FC_SMT_LAN_LOC :
1571 frame_status |= LAN_TX | LOC_TX ;
1572 break ;
1573 default :
1574 SMT_PANIC(smc,HWM_E0010,HWM_E0010_MSG) ;
1575 }
1576 }
1577 if (!smc->hw.mac_ring_is_up) {
1578 frame_status &= ~LAN_TX ;
1579 frame_status |= RING_DOWN ;
1580 DB_TX("Ring is down: terminate LAN_TX",0,0,2) ;
1581 }
1582 if (frag_count > smc->os.hwm.tx_p->tx_free) {
1583#ifndef NDIS_OS2
1584 mac_drv_clear_txd(smc) ;
1585 if (frag_count > smc->os.hwm.tx_p->tx_free) {
1586 DB_TX("Out of TxDs, terminate LAN_TX",0,0,2) ;
1587 frame_status &= ~LAN_TX ;
1588 frame_status |= OUT_OF_TXD ;
1589 }
1590#else
1591 DB_TX("Out of TxDs, terminate LAN_TX",0,0,2) ;
1592 frame_status &= ~LAN_TX ;
1593 frame_status |= OUT_OF_TXD ;
1594#endif
1595 }
1596 DB_TX("frame_status = %x",frame_status,0,3) ;
1597 NDD_TRACE("THiE",frame_status,smc->os.hwm.tx_p->tx_free,0) ;
1598 return frame_status;
1599}
1600
1601/*
1602 * BEGIN_MANUAL_ENTRY(hwm_tx_frag)
1603 * void hwm_tx_frag(smc,virt,phys,len,frame_status)
1604 *
1605 * function DOWNCALL (hardware module, hwmtm.c)
1606 * If the frame should be sent to the LAN, this function calls
1607 * dma_master, fills the current TxD with the virtual and the
1608 * physical address, sets the STF and EOF bits dependent on
1609 * the frame status, and requests the BMU to start the
1610 * transmit.
1611 * If the frame should be sent to the local SMT, an SMT_MBuf
1612 * is allocated if the FIRST_FRAG bit is set in the frame_status.
1613 * The fragment of the frame is copied into the SMT MBuf.
1614 * The function smt_received_pack is called if the LAST_FRAG
1615 * bit is set in the frame_status word.
1616 *
1617 * para virt virtual pointer to the fragment
1618 * len the length of the fragment
1619 * frame_status status of the frame, see design description
1620 *
1621 * return nothing returned, no parameter is modified
1622 *
1623 * NOTE: It is possible to invoke this macro with a fragment length
1624 * of zero.
1625 *
1626 * END_MANUAL_ENTRY
1627 */
1628void hwm_tx_frag(struct s_smc *smc, char far *virt, u_long phys, int len,
1629 int frame_status)
1630{
1631 struct s_smt_fp_txd volatile *t ;
1632 struct s_smt_tx_queue *queue ;
1633 __le32 tbctrl ;
1634
1635 queue = smc->os.hwm.tx_p ;
1636
1637 NDD_TRACE("THfB",virt,len,frame_status) ;
1638 /* Bug fix: AF / May 31 1999 (#missing)
1639 * snmpinfo problem reported by IBM is caused by invalid
1640 * t-pointer (txd) if LAN_TX is not set but LOC_TX only.
1641 * Set: t = queue->tx_curr_put here !
1642 */
1643 t = queue->tx_curr_put ;
1644
1645 DB_TX("hwm_tx_frag: len = %d, frame_status = %x ",len,frame_status,2) ;
1646 if (frame_status & LAN_TX) {
1647 /* '*t' is already defined */
1648 DB_TX("LAN_TX: TxD = %x, virt = %x ",t,virt,3) ;
1649 t->txd_virt = virt ;
1650 t->txd_txdscr = cpu_to_le32(smc->os.hwm.tx_descr) ;
1651 t->txd_tbadr = cpu_to_le32(phys) ;
1652 tbctrl = cpu_to_le32((((__u32)frame_status &
1653 (FIRST_FRAG|LAST_FRAG|EN_IRQ_EOF))<< 26) |
1654 BMU_OWN|BMU_CHECK |len) ;
1655 t->txd_tbctrl = tbctrl ;
1656
1657#ifndef AIX
1658 DRV_BUF_FLUSH(t,DDI_DMA_SYNC_FORDEV) ;
1659 outpd(queue->tx_bmu_ctl,CSR_START) ;
1660#else /* ifndef AIX */
1661 DRV_BUF_FLUSH(t,DDI_DMA_SYNC_FORDEV) ;
1662 if (frame_status & QUEUE_A0) {
1663 outpd(ADDR(B0_XA_CSR),CSR_START) ;
1664 }
1665 else {
1666 outpd(ADDR(B0_XS_CSR),CSR_START) ;
1667 }
1668#endif
1669 queue->tx_free-- ;
1670 queue->tx_used++ ;
1671 queue->tx_curr_put = t->txd_next ;
1672 if (frame_status & LAST_FRAG) {
1673 smc->mib.m[MAC0].fddiMACTransmit_Ct++ ;
1674 }
1675 }
1676 if (frame_status & LOC_TX) {
1677 DB_TX("LOC_TX: ",0,0,3) ;
1678 if (frame_status & FIRST_FRAG) {
1679 if(!(smc->os.hwm.tx_mb = smt_get_mbuf(smc))) {
1680 smc->hw.fp.err_stats.err_no_buf++ ;
1681 DB_TX("No SMbuf; transmit terminated",0,0,4) ;
1682 }
1683 else {
1684 smc->os.hwm.tx_data =
1685 smtod(smc->os.hwm.tx_mb,char *) - 1 ;
1686#ifdef USE_OS_CPY
1687#ifdef PASS_1ST_TXD_2_TX_COMP
1688 hwm_cpy_txd2mb(t,smc->os.hwm.tx_data,
1689 smc->os.hwm.tx_len) ;
1690#endif
1691#endif
1692 }
1693 }
1694 if (smc->os.hwm.tx_mb) {
1695#ifndef USE_OS_CPY
1696 DB_TX("copy fragment into MBuf ",0,0,3) ;
1697 memcpy(smc->os.hwm.tx_data,virt,len) ;
1698 smc->os.hwm.tx_data += len ;
1699#endif
1700 if (frame_status & LAST_FRAG) {
1701#ifdef USE_OS_CPY
1702#ifndef PASS_1ST_TXD_2_TX_COMP
1703 /*
1704 * hwm_cpy_txd2mb(txd,data,len) copies 'len'
1705 * bytes from the virtual pointer in 'rxd'
1706 * to 'data'. The virtual pointer of the
1707 * os-specific tx-buffer should be written
1708 * in the LAST txd.
1709 */
1710 hwm_cpy_txd2mb(t,smc->os.hwm.tx_data,
1711 smc->os.hwm.tx_len) ;
1712#endif /* nPASS_1ST_TXD_2_TX_COMP */
1713#endif /* USE_OS_CPY */
1714 smc->os.hwm.tx_data =
1715 smtod(smc->os.hwm.tx_mb,char *) - 1 ;
1716 *(char *)smc->os.hwm.tx_mb->sm_data =
1717 *smc->os.hwm.tx_data ;
1718 smc->os.hwm.tx_data++ ;
1719 smc->os.hwm.tx_mb->sm_len =
1720 smc->os.hwm.tx_len - 1 ;
1721 DB_TX("pass LLC frame to SMT ",0,0,3) ;
1722 smt_received_pack(smc,smc->os.hwm.tx_mb,
1723 RD_FS_LOCAL) ;
1724 }
1725 }
1726 }
1727 NDD_TRACE("THfE",t,queue->tx_free,0) ;
1728}
1729
1730
1731/*
1732 * queues a receive for later send
1733 */
1734static void queue_llc_rx(struct s_smc *smc, SMbuf *mb)
1735{
1736 DB_GEN("queue_llc_rx: mb = %x",(void *)mb,0,4) ;
1737 smc->os.hwm.queued_rx_frames++ ;
1738 mb->sm_next = (SMbuf *)NULL ;
1739 if (smc->os.hwm.llc_rx_pipe == NULL) {
1740 smc->os.hwm.llc_rx_pipe = mb ;
1741 }
1742 else {
1743 smc->os.hwm.llc_rx_tail->sm_next = mb ;
1744 }
1745 smc->os.hwm.llc_rx_tail = mb ;
1746
1747 /*
1748 * force an timer IRQ to receive the data
1749 */
1750 if (!smc->os.hwm.isr_flag) {
1751 smt_force_irq(smc) ;
1752 }
1753}
1754
1755/*
1756 * get a SMbuf from the llc_rx_queue
1757 */
1758static SMbuf *get_llc_rx(struct s_smc *smc)
1759{
1760 SMbuf *mb ;
1761
1762 if ((mb = smc->os.hwm.llc_rx_pipe)) {
1763 smc->os.hwm.queued_rx_frames-- ;
1764 smc->os.hwm.llc_rx_pipe = mb->sm_next ;
1765 }
1766 DB_GEN("get_llc_rx: mb = 0x%x",(void *)mb,0,4) ;
1767 return mb;
1768}
1769
1770/*
1771 * queues a transmit SMT MBuf during the time were the MBuf is
1772 * queued the TxD ring
1773 */
1774static void queue_txd_mb(struct s_smc *smc, SMbuf *mb)
1775{
1776 DB_GEN("_rx: queue_txd_mb = %x",(void *)mb,0,4) ;
1777 smc->os.hwm.queued_txd_mb++ ;
1778 mb->sm_next = (SMbuf *)NULL ;
1779 if (smc->os.hwm.txd_tx_pipe == NULL) {
1780 smc->os.hwm.txd_tx_pipe = mb ;
1781 }
1782 else {
1783 smc->os.hwm.txd_tx_tail->sm_next = mb ;
1784 }
1785 smc->os.hwm.txd_tx_tail = mb ;
1786}
1787
1788/*
1789 * get a SMbuf from the txd_tx_queue
1790 */
1791static SMbuf *get_txd_mb(struct s_smc *smc)
1792{
1793 SMbuf *mb ;
1794
1795 if ((mb = smc->os.hwm.txd_tx_pipe)) {
1796 smc->os.hwm.queued_txd_mb-- ;
1797 smc->os.hwm.txd_tx_pipe = mb->sm_next ;
1798 }
1799 DB_GEN("get_txd_mb: mb = 0x%x",(void *)mb,0,4) ;
1800 return mb;
1801}
1802
1803/*
1804 * SMT Send function
1805 */
1806void smt_send_mbuf(struct s_smc *smc, SMbuf *mb, int fc)
1807{
1808 char far *data ;
1809 int len ;
1810 int n ;
1811 int i ;
1812 int frag_count ;
1813 int frame_status ;
1814 SK_LOC_DECL(char far,*virt[3]) ;
1815 int frag_len[3] ;
1816 struct s_smt_tx_queue *queue ;
1817 struct s_smt_fp_txd volatile *t ;
1818 u_long phys ;
1819 __le32 tbctrl;
1820
1821 NDD_TRACE("THSB",mb,fc,0) ;
1822 DB_TX("smt_send_mbuf: mb = 0x%x, fc = 0x%x",mb,fc,4) ;
1823
1824 mb->sm_off-- ; /* set to fc */
1825 mb->sm_len++ ; /* + fc */
1826 data = smtod(mb,char *) ;
1827 *data = fc ;
1828 if (fc == FC_SMT_LOC)
1829 *data = FC_SMT_INFO ;
1830
1831 /*
1832 * determine the frag count and the virt addresses of the frags
1833 */
1834 frag_count = 0 ;
1835 len = mb->sm_len ;
1836 while (len) {
1837 n = SMT_PAGESIZE - ((long)data & (SMT_PAGESIZE-1)) ;
1838 if (n >= len) {
1839 n = len ;
1840 }
1841 DB_TX("frag: virt/len = 0x%x/%d ",(void *)data,n,5) ;
1842 virt[frag_count] = data ;
1843 frag_len[frag_count] = n ;
1844 frag_count++ ;
1845 len -= n ;
1846 data += n ;
1847 }
1848
1849 /*
1850 * determine the frame status
1851 */
1852 queue = smc->hw.fp.tx[QUEUE_A0] ;
1853 if (fc == FC_BEACON || fc == FC_SMT_LOC) {
1854 frame_status = LOC_TX ;
1855 }
1856 else {
1857 frame_status = LAN_TX ;
1858 if ((smc->os.hwm.pass_NSA &&(fc == FC_SMT_NSA)) ||
1859 (smc->os.hwm.pass_SMT &&(fc == FC_SMT_INFO)))
1860 frame_status |= LOC_TX ;
1861 }
1862
1863 if (!smc->hw.mac_ring_is_up || frag_count > queue->tx_free) {
1864 frame_status &= ~LAN_TX;
1865 if (frame_status) {
1866 DB_TX("Ring is down: terminate LAN_TX",0,0,2) ;
1867 }
1868 else {
1869 DB_TX("Ring is down: terminate transmission",0,0,2) ;
1870 smt_free_mbuf(smc,mb) ;
1871 return ;
1872 }
1873 }
1874 DB_TX("frame_status = 0x%x ",frame_status,0,5) ;
1875
1876 if ((frame_status & LAN_TX) && (frame_status & LOC_TX)) {
1877 mb->sm_use_count = 2 ;
1878 }
1879
1880 if (frame_status & LAN_TX) {
1881 t = queue->tx_curr_put ;
1882 frame_status |= FIRST_FRAG ;
1883 for (i = 0; i < frag_count; i++) {
1884 DB_TX("init TxD = 0x%x",(void *)t,0,5) ;
1885 if (i == frag_count-1) {
1886 frame_status |= LAST_FRAG ;
1887 t->txd_txdscr = cpu_to_le32(TX_DESCRIPTOR |
1888 (((__u32)(mb->sm_len-1)&3) << 27)) ;
1889 }
1890 t->txd_virt = virt[i] ;
1891 phys = dma_master(smc, (void far *)virt[i],
1892 frag_len[i], DMA_RD|SMT_BUF) ;
1893 t->txd_tbadr = cpu_to_le32(phys) ;
1894 tbctrl = cpu_to_le32((((__u32)frame_status &
1895 (FIRST_FRAG|LAST_FRAG)) << 26) |
1896 BMU_OWN | BMU_CHECK | BMU_SMT_TX |frag_len[i]) ;
1897 t->txd_tbctrl = tbctrl ;
1898#ifndef AIX
1899 DRV_BUF_FLUSH(t,DDI_DMA_SYNC_FORDEV) ;
1900 outpd(queue->tx_bmu_ctl,CSR_START) ;
1901#else
1902 DRV_BUF_FLUSH(t,DDI_DMA_SYNC_FORDEV) ;
1903 outpd(ADDR(B0_XA_CSR),CSR_START) ;
1904#endif
1905 frame_status &= ~FIRST_FRAG ;
1906 queue->tx_curr_put = t = t->txd_next ;
1907 queue->tx_free-- ;
1908 queue->tx_used++ ;
1909 }
1910 smc->mib.m[MAC0].fddiMACTransmit_Ct++ ;
1911 queue_txd_mb(smc,mb) ;
1912 }
1913
1914 if (frame_status & LOC_TX) {
1915 DB_TX("pass Mbuf to LLC queue",0,0,5) ;
1916 queue_llc_rx(smc,mb) ;
1917 }
1918
1919 /*
1920 * We need to unqueue the free SMT_MBUFs here, because it may
1921 * be that the SMT want's to send more than 1 frame for one down call
1922 */
1923 mac_drv_clear_txd(smc) ;
1924 NDD_TRACE("THSE",t,queue->tx_free,frag_count) ;
1925}
1926
1927/* BEGIN_MANUAL_ENTRY(mac_drv_clear_txd)
1928 * void mac_drv_clear_txd(smc)
1929 *
1930 * function DOWNCALL (hardware module, hwmtm.c)
1931 * mac_drv_clear_txd searches in both send queues for TxD's
1932 * which were finished by the adapter. It calls dma_complete
1933 * for each TxD. If the last fragment of an LLC frame is
1934 * reached, it calls mac_drv_tx_complete to release the
1935 * send buffer.
1936 *
1937 * return nothing
1938 *
1939 * END_MANUAL_ENTRY
1940 */
1941static void mac_drv_clear_txd(struct s_smc *smc)
1942{
1943 struct s_smt_tx_queue *queue ;
1944 struct s_smt_fp_txd volatile *t1 ;
1945 struct s_smt_fp_txd volatile *t2 = NULL ;
1946 SMbuf *mb ;
1947 u_long tbctrl ;
1948 int i ;
1949 int frag_count ;
1950 int n ;
1951
1952 NDD_TRACE("THcB",0,0,0) ;
1953 for (i = QUEUE_S; i <= QUEUE_A0; i++) {
1954 queue = smc->hw.fp.tx[i] ;
1955 t1 = queue->tx_curr_get ;
1956 DB_TX("clear_txd: QUEUE = %d (0=sync/1=async)",i,0,5) ;
1957
1958 for ( ; ; ) {
1959 frag_count = 0 ;
1960
1961 do {
1962 DRV_BUF_FLUSH(t1,DDI_DMA_SYNC_FORCPU) ;
1963 DB_TX("check OWN/EOF bit of TxD 0x%x",t1,0,5) ;
1964 tbctrl = le32_to_cpu(CR_READ(t1->txd_tbctrl));
1965
1966 if (tbctrl & BMU_OWN || !queue->tx_used){
1967 DB_TX("End of TxDs queue %d",i,0,4) ;
1968 goto free_next_queue ; /* next queue */
1969 }
1970 t1 = t1->txd_next ;
1971 frag_count++ ;
1972 } while (!(tbctrl & BMU_EOF)) ;
1973
1974 t1 = queue->tx_curr_get ;
1975 for (n = frag_count; n; n--) {
1976 tbctrl = le32_to_cpu(t1->txd_tbctrl) ;
1977 dma_complete(smc,
1978 (union s_fp_descr volatile *) t1,
1979 (int) (DMA_RD |
1980 ((tbctrl & BMU_SMT_TX) >> 18))) ;
1981 t2 = t1 ;
1982 t1 = t1->txd_next ;
1983 }
1984
1985 if (tbctrl & BMU_SMT_TX) {
1986 mb = get_txd_mb(smc) ;
1987 smt_free_mbuf(smc,mb) ;
1988 }
1989 else {
1990#ifndef PASS_1ST_TXD_2_TX_COMP
1991 DB_TX("mac_drv_tx_comp for TxD 0x%x",t2,0,4) ;
1992 mac_drv_tx_complete(smc,t2) ;
1993#else
1994 DB_TX("mac_drv_tx_comp for TxD 0x%x",
1995 queue->tx_curr_get,0,4) ;
1996 mac_drv_tx_complete(smc,queue->tx_curr_get) ;
1997#endif
1998 }
1999 queue->tx_curr_get = t1 ;
2000 queue->tx_free += frag_count ;
2001 queue->tx_used -= frag_count ;
2002 }
2003free_next_queue: ;
2004 }
2005 NDD_TRACE("THcE",0,0,0) ;
2006}
2007
2008/*
2009 * BEGINN_MANUAL_ENTRY(mac_drv_clear_tx_queue)
2010 *
2011 * void mac_drv_clear_tx_queue(smc)
2012 * struct s_smc *smc ;
2013 *
2014 * function DOWNCALL (hardware module, hwmtm.c)
2015 * mac_drv_clear_tx_queue is called from the SMT when
2016 * the RMT state machine has entered the ISOLATE state.
2017 * This function is also called by the os-specific module
2018 * after it has called the function card_stop().
2019 * In this case, the frames in the send queues are obsolete and
2020 * should be removed.
2021 *
2022 * note calling sequence:
2023 * CLI_FBI(), card_stop(),
2024 * mac_drv_clear_tx_queue(), mac_drv_clear_rx_queue(),
2025 *
2026 * NOTE: The caller is responsible that the BMUs are idle
2027 * when this function is called.
2028 *
2029 * END_MANUAL_ENTRY
2030 */
2031void mac_drv_clear_tx_queue(struct s_smc *smc)
2032{
2033 struct s_smt_fp_txd volatile *t ;
2034 struct s_smt_tx_queue *queue ;
2035 int tx_used ;
2036 int i ;
2037
2038 if (smc->hw.hw_state != STOPPED) {
2039 SK_BREAK() ;
2040 SMT_PANIC(smc,HWM_E0011,HWM_E0011_MSG) ;
2041 return ;
2042 }
2043
2044 for (i = QUEUE_S; i <= QUEUE_A0; i++) {
2045 queue = smc->hw.fp.tx[i] ;
2046 DB_TX("clear_tx_queue: QUEUE = %d (0=sync/1=async)",i,0,5) ;
2047
2048 /*
2049 * switch the OWN bit of all pending frames to the host
2050 */
2051 t = queue->tx_curr_get ;
2052 tx_used = queue->tx_used ;
2053 while (tx_used) {
2054 DRV_BUF_FLUSH(t,DDI_DMA_SYNC_FORCPU) ;
2055 DB_TX("switch OWN bit of TxD 0x%x ",t,0,5) ;
2056 t->txd_tbctrl &= ~cpu_to_le32(BMU_OWN) ;
2057 DRV_BUF_FLUSH(t,DDI_DMA_SYNC_FORDEV) ;
2058 t = t->txd_next ;
2059 tx_used-- ;
2060 }
2061 }
2062
2063 /*
2064 * release all TxD's for both send queues
2065 */
2066 mac_drv_clear_txd(smc) ;
2067
2068 for (i = QUEUE_S; i <= QUEUE_A0; i++) {
2069 queue = smc->hw.fp.tx[i] ;
2070 t = queue->tx_curr_get ;
2071
2072 /*
2073 * write the phys pointer of the NEXT descriptor into the
2074 * BMU's current address descriptor pointer and set
2075 * tx_curr_get and tx_curr_put to this position
2076 */
2077 if (i == QUEUE_S) {
2078 outpd(ADDR(B5_XS_DA),le32_to_cpu(t->txd_ntdadr)) ;
2079 }
2080 else {
2081 outpd(ADDR(B5_XA_DA),le32_to_cpu(t->txd_ntdadr)) ;
2082 }
2083
2084 queue->tx_curr_put = queue->tx_curr_get->txd_next ;
2085 queue->tx_curr_get = queue->tx_curr_put ;
2086 }
2087}
2088
2089
2090/*
2091 -------------------------------------------------------------
2092 TEST FUNCTIONS:
2093 -------------------------------------------------------------
2094*/
2095
2096#ifdef DEBUG
2097/*
2098 * BEGIN_MANUAL_ENTRY(mac_drv_debug_lev)
2099 * void mac_drv_debug_lev(smc,flag,lev)
2100 *
2101 * function DOWNCALL (drvsr.c)
2102 * To get a special debug info the user can assign a debug level
2103 * to any debug flag.
2104 *
2105 * para flag debug flag, possible values are:
2106 * = 0: reset all debug flags (the defined level is
2107 * ignored)
2108 * = 1: debug.d_smtf
2109 * = 2: debug.d_smt
2110 * = 3: debug.d_ecm
2111 * = 4: debug.d_rmt
2112 * = 5: debug.d_cfm
2113 * = 6: debug.d_pcm
2114 *
2115 * = 10: debug.d_os.hwm_rx (hardware module receive path)
2116 * = 11: debug.d_os.hwm_tx(hardware module transmit path)
2117 * = 12: debug.d_os.hwm_gen(hardware module general flag)
2118 *
2119 * lev debug level
2120 *
2121 * END_MANUAL_ENTRY
2122 */
2123void mac_drv_debug_lev(struct s_smc *smc, int flag, int lev)
2124{
2125 switch(flag) {
2126 case (int)NULL:
2127 DB_P.d_smtf = DB_P.d_smt = DB_P.d_ecm = DB_P.d_rmt = 0 ;
2128 DB_P.d_cfm = 0 ;
2129 DB_P.d_os.hwm_rx = DB_P.d_os.hwm_tx = DB_P.d_os.hwm_gen = 0 ;
2130#ifdef SBA
2131 DB_P.d_sba = 0 ;
2132#endif
2133#ifdef ESS
2134 DB_P.d_ess = 0 ;
2135#endif
2136 break ;
2137 case DEBUG_SMTF:
2138 DB_P.d_smtf = lev ;
2139 break ;
2140 case DEBUG_SMT:
2141 DB_P.d_smt = lev ;
2142 break ;
2143 case DEBUG_ECM:
2144 DB_P.d_ecm = lev ;
2145 break ;
2146 case DEBUG_RMT:
2147 DB_P.d_rmt = lev ;
2148 break ;
2149 case DEBUG_CFM:
2150 DB_P.d_cfm = lev ;
2151 break ;
2152 case DEBUG_PCM:
2153 DB_P.d_pcm = lev ;
2154 break ;
2155 case DEBUG_SBA:
2156#ifdef SBA
2157 DB_P.d_sba = lev ;
2158#endif
2159 break ;
2160 case DEBUG_ESS:
2161#ifdef ESS
2162 DB_P.d_ess = lev ;
2163#endif
2164 break ;
2165 case DB_HWM_RX:
2166 DB_P.d_os.hwm_rx = lev ;
2167 break ;
2168 case DB_HWM_TX:
2169 DB_P.d_os.hwm_tx = lev ;
2170 break ;
2171 case DB_HWM_GEN:
2172 DB_P.d_os.hwm_gen = lev ;
2173 break ;
2174 default:
2175 break ;
2176 }
2177}
2178#endif
1// SPDX-License-Identifier: GPL-2.0-or-later
2/******************************************************************************
3 *
4 * (C)Copyright 1998,1999 SysKonnect,
5 * a business unit of Schneider & Koch & Co. Datensysteme GmbH.
6 *
7 * See the file "skfddi.c" for further information.
8 *
9 * The information in this file is provided "AS IS" without warranty.
10 *
11 ******************************************************************************/
12
13#define HWMTM
14
15#ifndef FDDI
16#define FDDI
17#endif
18
19#include "h/types.h"
20#include "h/fddi.h"
21#include "h/smc.h"
22#include "h/supern_2.h"
23#include "h/skfbiinc.h"
24
25/*
26 -------------------------------------------------------------
27 DOCUMENTATION
28 -------------------------------------------------------------
29 BEGIN_MANUAL_ENTRY(DOCUMENTATION)
30
31 T B D
32
33 END_MANUAL_ENTRY
34*/
35/*
36 -------------------------------------------------------------
37 LOCAL VARIABLES:
38 -------------------------------------------------------------
39*/
40#ifdef COMMON_MB_POOL
41static SMbuf *mb_start;
42static SMbuf *mb_free;
43static int mb_init = FALSE ;
44static int call_count;
45#endif
46
47/*
48 -------------------------------------------------------------
49 EXTERNE VARIABLES:
50 -------------------------------------------------------------
51*/
52
53#ifdef DEBUG
54#ifndef DEBUG_BRD
55extern struct smt_debug debug ;
56#endif
57#endif
58
59#ifdef NDIS_OS2
60extern u_char offDepth ;
61extern u_char force_irq_pending ;
62#endif
63
64/*
65 -------------------------------------------------------------
66 LOCAL FUNCTIONS:
67 -------------------------------------------------------------
68*/
69
70static void queue_llc_rx(struct s_smc *smc, SMbuf *mb);
71static void smt_to_llc(struct s_smc *smc, SMbuf *mb);
72static void init_txd_ring(struct s_smc *smc);
73static void init_rxd_ring(struct s_smc *smc);
74static void queue_txd_mb(struct s_smc *smc, SMbuf *mb);
75static u_long init_descr_ring(struct s_smc *smc, union s_fp_descr volatile *start,
76 int count);
77static u_long repair_txd_ring(struct s_smc *smc, struct s_smt_tx_queue *queue);
78static u_long repair_rxd_ring(struct s_smc *smc, struct s_smt_rx_queue *queue);
79static SMbuf* get_llc_rx(struct s_smc *smc);
80static SMbuf* get_txd_mb(struct s_smc *smc);
81static void mac_drv_clear_txd(struct s_smc *smc);
82
83/*
84 -------------------------------------------------------------
85 EXTERNAL FUNCTIONS:
86 -------------------------------------------------------------
87*/
88/* The external SMT functions are listed in cmtdef.h */
89
90extern void* mac_drv_get_space(struct s_smc *smc, unsigned int size);
91extern void* mac_drv_get_desc_mem(struct s_smc *smc, unsigned int size);
92extern void mac_drv_fill_rxd(struct s_smc *smc);
93extern void mac_drv_tx_complete(struct s_smc *smc,
94 volatile struct s_smt_fp_txd *txd);
95extern void mac_drv_rx_complete(struct s_smc *smc,
96 volatile struct s_smt_fp_rxd *rxd,
97 int frag_count, int len);
98extern void mac_drv_requeue_rxd(struct s_smc *smc,
99 volatile struct s_smt_fp_rxd *rxd,
100 int frag_count);
101extern void mac_drv_clear_rxd(struct s_smc *smc,
102 volatile struct s_smt_fp_rxd *rxd, int frag_count);
103
104#ifdef USE_OS_CPY
105extern void hwm_cpy_rxd2mb(void);
106extern void hwm_cpy_txd2mb(void);
107#endif
108
109#ifdef ALL_RX_COMPLETE
110extern void mac_drv_all_receives_complete(void);
111#endif
112
113extern u_long mac_drv_virt2phys(struct s_smc *smc, void *virt);
114extern u_long dma_master(struct s_smc *smc, void *virt, int len, int flag);
115
116#ifdef NDIS_OS2
117extern void post_proc(void);
118#else
119extern void dma_complete(struct s_smc *smc, volatile union s_fp_descr *descr,
120 int flag);
121#endif
122
123extern int mac_drv_rx_init(struct s_smc *smc, int len, int fc, char *look_ahead,
124 int la_len);
125
126/*
127 -------------------------------------------------------------
128 PUBLIC FUNCTIONS:
129 -------------------------------------------------------------
130*/
131void process_receive(struct s_smc *smc);
132void fddi_isr(struct s_smc *smc);
133void smt_free_mbuf(struct s_smc *smc, SMbuf *mb);
134void init_driver_fplus(struct s_smc *smc);
135void mac_drv_rx_mode(struct s_smc *smc, int mode);
136void init_fddi_driver(struct s_smc *smc, u_char *mac_addr);
137void mac_drv_clear_tx_queue(struct s_smc *smc);
138void mac_drv_clear_rx_queue(struct s_smc *smc);
139void hwm_tx_frag(struct s_smc *smc, char far *virt, u_long phys, int len,
140 int frame_status);
141void hwm_rx_frag(struct s_smc *smc, char far *virt, u_long phys, int len,
142 int frame_status);
143
144int mac_drv_init(struct s_smc *smc);
145int hwm_tx_init(struct s_smc *smc, u_char fc, int frag_count, int frame_len,
146 int frame_status);
147
148u_int mac_drv_check_space(void);
149
150SMbuf* smt_get_mbuf(struct s_smc *smc);
151
152#ifdef DEBUG
153 void mac_drv_debug_lev(struct s_smc *smc, int flag, int lev);
154#endif
155
156/*
157 -------------------------------------------------------------
158 MACROS:
159 -------------------------------------------------------------
160*/
161#ifndef UNUSED
162#ifdef lint
163#define UNUSED(x) (x) = (x)
164#else
165#define UNUSED(x)
166#endif
167#endif
168
169#ifdef USE_CAN_ADDR
170#define MA smc->hw.fddi_canon_addr.a
171#define GROUP_ADDR_BIT 0x01
172#else
173#define MA smc->hw.fddi_home_addr.a
174#define GROUP_ADDR_BIT 0x80
175#endif
176
177#define RXD_TXD_COUNT (HWM_ASYNC_TXD_COUNT+HWM_SYNC_TXD_COUNT+\
178 SMT_R1_RXD_COUNT+SMT_R2_RXD_COUNT)
179
180#ifdef MB_OUTSIDE_SMC
181#define EXT_VIRT_MEM ((RXD_TXD_COUNT+1)*sizeof(struct s_smt_fp_txd) +\
182 MAX_MBUF*sizeof(SMbuf))
183#define EXT_VIRT_MEM_2 ((RXD_TXD_COUNT+1)*sizeof(struct s_smt_fp_txd))
184#else
185#define EXT_VIRT_MEM ((RXD_TXD_COUNT+1)*sizeof(struct s_smt_fp_txd))
186#endif
187
188 /*
189 * define critical read for 16 Bit drivers
190 */
191#if defined(NDIS_OS2) || defined(ODI2)
192#define CR_READ(var) ((var) & 0xffff0000 | ((var) & 0xffff))
193#else
194#define CR_READ(var) (__le32)(var)
195#endif
196
197#define IMASK_SLOW (IS_PLINT1 | IS_PLINT2 | IS_TIMINT | IS_TOKEN | \
198 IS_MINTR1 | IS_MINTR2 | IS_MINTR3 | IS_R1_P | \
199 IS_R1_C | IS_XA_C | IS_XS_C)
200
201/*
202 -------------------------------------------------------------
203 INIT- AND SMT FUNCTIONS:
204 -------------------------------------------------------------
205*/
206
207
208/*
209 * BEGIN_MANUAL_ENTRY(mac_drv_check_space)
210 * u_int mac_drv_check_space()
211 *
212 * function DOWNCALL (drvsr.c)
213 * This function calculates the needed non virtual
214 * memory for MBufs, RxD and TxD descriptors etc.
215 * needed by the driver.
216 *
217 * return u_int memory in bytes
218 *
219 * END_MANUAL_ENTRY
220 */
221u_int mac_drv_check_space(void)
222{
223#ifdef MB_OUTSIDE_SMC
224#ifdef COMMON_MB_POOL
225 call_count++ ;
226 if (call_count == 1) {
227 return EXT_VIRT_MEM;
228 }
229 else {
230 return EXT_VIRT_MEM_2;
231 }
232#else
233 return EXT_VIRT_MEM;
234#endif
235#else
236 return 0;
237#endif
238}
239
240/*
241 * BEGIN_MANUAL_ENTRY(mac_drv_init)
242 * void mac_drv_init(smc)
243 *
244 * function DOWNCALL (drvsr.c)
245 * In this function the hardware module allocates it's
246 * memory.
247 * The operating system dependent module should call
248 * mac_drv_init once, after the adatper is detected.
249 * END_MANUAL_ENTRY
250 */
251int mac_drv_init(struct s_smc *smc)
252{
253 if (sizeof(struct s_smt_fp_rxd) % 16) {
254 SMT_PANIC(smc,HWM_E0001,HWM_E0001_MSG) ;
255 }
256 if (sizeof(struct s_smt_fp_txd) % 16) {
257 SMT_PANIC(smc,HWM_E0002,HWM_E0002_MSG) ;
258 }
259
260 /*
261 * get the required memory for the RxDs and TxDs
262 */
263 if (!(smc->os.hwm.descr_p = (union s_fp_descr volatile *)
264 mac_drv_get_desc_mem(smc,(u_int)
265 (RXD_TXD_COUNT+1)*sizeof(struct s_smt_fp_txd)))) {
266 return 1; /* no space the hwm modul can't work */
267 }
268
269 /*
270 * get the memory for the SMT MBufs
271 */
272#ifndef MB_OUTSIDE_SMC
273 smc->os.hwm.mbuf_pool.mb_start=(SMbuf *)(&smc->os.hwm.mbuf_pool.mb[0]) ;
274#else
275#ifndef COMMON_MB_POOL
276 if (!(smc->os.hwm.mbuf_pool.mb_start = (SMbuf *) mac_drv_get_space(smc,
277 MAX_MBUF*sizeof(SMbuf)))) {
278 return 1; /* no space the hwm modul can't work */
279 }
280#else
281 if (!mb_start) {
282 if (!(mb_start = (SMbuf *) mac_drv_get_space(smc,
283 MAX_MBUF*sizeof(SMbuf)))) {
284 return 1; /* no space the hwm modul can't work */
285 }
286 }
287#endif
288#endif
289 return 0;
290}
291
292/*
293 * BEGIN_MANUAL_ENTRY(init_driver_fplus)
294 * init_driver_fplus(smc)
295 *
296 * Sets hardware modul specific values for the mode register 2
297 * (e.g. the byte alignment for the received frames, the position of the
298 * least significant byte etc.)
299 * END_MANUAL_ENTRY
300 */
301void init_driver_fplus(struct s_smc *smc)
302{
303 smc->hw.fp.mdr2init = FM_LSB | FM_BMMODE | FM_ENNPRQ | FM_ENHSRQ | 3 ;
304
305#ifdef PCI
306 smc->hw.fp.mdr2init |= FM_CHKPAR | FM_PARITY ;
307#endif
308 smc->hw.fp.mdr3init = FM_MENRQAUNLCK | FM_MENRS ;
309
310#ifdef USE_CAN_ADDR
311 /* enable address bit swapping */
312 smc->hw.fp.frselreg_init = FM_ENXMTADSWAP | FM_ENRCVADSWAP ;
313#endif
314}
315
316static u_long init_descr_ring(struct s_smc *smc,
317 union s_fp_descr volatile *start,
318 int count)
319{
320 int i ;
321 union s_fp_descr volatile *d1 ;
322 union s_fp_descr volatile *d2 ;
323 u_long phys ;
324
325 DB_GEN(3, "descr ring starts at = %p", start);
326 for (i=count-1, d1=start; i ; i--) {
327 d2 = d1 ;
328 d1++ ; /* descr is owned by the host */
329 d2->r.rxd_rbctrl = cpu_to_le32(BMU_CHECK) ;
330 d2->r.rxd_next = &d1->r ;
331 phys = mac_drv_virt2phys(smc,(void *)d1) ;
332 d2->r.rxd_nrdadr = cpu_to_le32(phys) ;
333 }
334 DB_GEN(3, "descr ring ends at = %p", d1);
335 d1->r.rxd_rbctrl = cpu_to_le32(BMU_CHECK) ;
336 d1->r.rxd_next = &start->r ;
337 phys = mac_drv_virt2phys(smc,(void *)start) ;
338 d1->r.rxd_nrdadr = cpu_to_le32(phys) ;
339
340 for (i=count, d1=start; i ; i--) {
341 DRV_BUF_FLUSH(&d1->r,DDI_DMA_SYNC_FORDEV) ;
342 d1++;
343 }
344 return phys;
345}
346
347static void init_txd_ring(struct s_smc *smc)
348{
349 struct s_smt_fp_txd volatile *ds ;
350 struct s_smt_tx_queue *queue ;
351 u_long phys ;
352
353 /*
354 * initialize the transmit descriptors
355 */
356 ds = (struct s_smt_fp_txd volatile *) ((char *)smc->os.hwm.descr_p +
357 SMT_R1_RXD_COUNT*sizeof(struct s_smt_fp_rxd)) ;
358 queue = smc->hw.fp.tx[QUEUE_A0] ;
359 DB_GEN(3, "Init async TxD ring, %d TxDs", HWM_ASYNC_TXD_COUNT);
360 (void)init_descr_ring(smc,(union s_fp_descr volatile *)ds,
361 HWM_ASYNC_TXD_COUNT) ;
362 phys = le32_to_cpu(ds->txd_ntdadr) ;
363 ds++ ;
364 queue->tx_curr_put = queue->tx_curr_get = ds ;
365 ds-- ;
366 queue->tx_free = HWM_ASYNC_TXD_COUNT ;
367 queue->tx_used = 0 ;
368 outpd(ADDR(B5_XA_DA),phys) ;
369
370 ds = (struct s_smt_fp_txd volatile *) ((char *)ds +
371 HWM_ASYNC_TXD_COUNT*sizeof(struct s_smt_fp_txd)) ;
372 queue = smc->hw.fp.tx[QUEUE_S] ;
373 DB_GEN(3, "Init sync TxD ring, %d TxDs", HWM_SYNC_TXD_COUNT);
374 (void)init_descr_ring(smc,(union s_fp_descr volatile *)ds,
375 HWM_SYNC_TXD_COUNT) ;
376 phys = le32_to_cpu(ds->txd_ntdadr) ;
377 ds++ ;
378 queue->tx_curr_put = queue->tx_curr_get = ds ;
379 queue->tx_free = HWM_SYNC_TXD_COUNT ;
380 queue->tx_used = 0 ;
381 outpd(ADDR(B5_XS_DA),phys) ;
382}
383
384static void init_rxd_ring(struct s_smc *smc)
385{
386 struct s_smt_fp_rxd volatile *ds ;
387 struct s_smt_rx_queue *queue ;
388 u_long phys ;
389
390 /*
391 * initialize the receive descriptors
392 */
393 ds = (struct s_smt_fp_rxd volatile *) smc->os.hwm.descr_p ;
394 queue = smc->hw.fp.rx[QUEUE_R1] ;
395 DB_GEN(3, "Init RxD ring, %d RxDs", SMT_R1_RXD_COUNT);
396 (void)init_descr_ring(smc,(union s_fp_descr volatile *)ds,
397 SMT_R1_RXD_COUNT) ;
398 phys = le32_to_cpu(ds->rxd_nrdadr) ;
399 ds++ ;
400 queue->rx_curr_put = queue->rx_curr_get = ds ;
401 queue->rx_free = SMT_R1_RXD_COUNT ;
402 queue->rx_used = 0 ;
403 outpd(ADDR(B4_R1_DA),phys) ;
404}
405
406/*
407 * BEGIN_MANUAL_ENTRY(init_fddi_driver)
408 * void init_fddi_driver(smc,mac_addr)
409 *
410 * initializes the driver and it's variables
411 *
412 * END_MANUAL_ENTRY
413 */
414void init_fddi_driver(struct s_smc *smc, u_char *mac_addr)
415{
416 SMbuf *mb ;
417 int i ;
418
419 init_board(smc,mac_addr) ;
420 (void)init_fplus(smc) ;
421
422 /*
423 * initialize the SMbufs for the SMT
424 */
425#ifndef COMMON_MB_POOL
426 mb = smc->os.hwm.mbuf_pool.mb_start ;
427 smc->os.hwm.mbuf_pool.mb_free = (SMbuf *)NULL ;
428 for (i = 0; i < MAX_MBUF; i++) {
429 mb->sm_use_count = 1 ;
430 smt_free_mbuf(smc,mb) ;
431 mb++ ;
432 }
433#else
434 mb = mb_start ;
435 if (!mb_init) {
436 mb_free = 0 ;
437 for (i = 0; i < MAX_MBUF; i++) {
438 mb->sm_use_count = 1 ;
439 smt_free_mbuf(smc,mb) ;
440 mb++ ;
441 }
442 mb_init = TRUE ;
443 }
444#endif
445
446 /*
447 * initialize the other variables
448 */
449 smc->os.hwm.llc_rx_pipe = smc->os.hwm.llc_rx_tail = (SMbuf *)NULL ;
450 smc->os.hwm.txd_tx_pipe = smc->os.hwm.txd_tx_tail = NULL ;
451 smc->os.hwm.pass_SMT = smc->os.hwm.pass_NSA = smc->os.hwm.pass_DB = 0 ;
452 smc->os.hwm.pass_llc_promisc = TRUE ;
453 smc->os.hwm.queued_rx_frames = smc->os.hwm.queued_txd_mb = 0 ;
454 smc->os.hwm.detec_count = 0 ;
455 smc->os.hwm.rx_break = 0 ;
456 smc->os.hwm.rx_len_error = 0 ;
457 smc->os.hwm.isr_flag = FALSE ;
458
459 /*
460 * make sure that the start pointer is 16 byte aligned
461 */
462 i = 16 - ((long)smc->os.hwm.descr_p & 0xf) ;
463 if (i != 16) {
464 DB_GEN(3, "i = %d", i);
465 smc->os.hwm.descr_p = (union s_fp_descr volatile *)
466 ((char *)smc->os.hwm.descr_p+i) ;
467 }
468 DB_GEN(3, "pt to descr area = %p", smc->os.hwm.descr_p);
469
470 init_txd_ring(smc) ;
471 init_rxd_ring(smc) ;
472 mac_drv_fill_rxd(smc) ;
473
474 init_plc(smc) ;
475}
476
477
478SMbuf *smt_get_mbuf(struct s_smc *smc)
479{
480 register SMbuf *mb ;
481
482#ifndef COMMON_MB_POOL
483 mb = smc->os.hwm.mbuf_pool.mb_free ;
484#else
485 mb = mb_free ;
486#endif
487 if (mb) {
488#ifndef COMMON_MB_POOL
489 smc->os.hwm.mbuf_pool.mb_free = mb->sm_next ;
490#else
491 mb_free = mb->sm_next ;
492#endif
493 mb->sm_off = 8 ;
494 mb->sm_use_count = 1 ;
495 }
496 DB_GEN(3, "get SMbuf: mb = %p", mb);
497 return mb; /* May be NULL */
498}
499
500void smt_free_mbuf(struct s_smc *smc, SMbuf *mb)
501{
502
503 if (mb) {
504 mb->sm_use_count-- ;
505 DB_GEN(3, "free_mbuf: sm_use_count = %d", mb->sm_use_count);
506 /*
507 * If the use_count is != zero the MBuf is queued
508 * more than once and must not queued into the
509 * free MBuf queue
510 */
511 if (!mb->sm_use_count) {
512 DB_GEN(3, "free SMbuf: mb = %p", mb);
513#ifndef COMMON_MB_POOL
514 mb->sm_next = smc->os.hwm.mbuf_pool.mb_free ;
515 smc->os.hwm.mbuf_pool.mb_free = mb ;
516#else
517 mb->sm_next = mb_free ;
518 mb_free = mb ;
519#endif
520 }
521 }
522 else
523 SMT_PANIC(smc,HWM_E0003,HWM_E0003_MSG) ;
524}
525
526
527/*
528 * BEGIN_MANUAL_ENTRY(mac_drv_repair_descr)
529 * void mac_drv_repair_descr(smc)
530 *
531 * function called from SMT (HWM / hwmtm.c)
532 * The BMU is idle when this function is called.
533 * Mac_drv_repair_descr sets up the physical address
534 * for all receive and transmit queues where the BMU
535 * should continue.
536 * It may be that the BMU was reseted during a fragmented
537 * transfer. In this case there are some fragments which will
538 * never completed by the BMU. The OWN bit of this fragments
539 * must be switched to be owned by the host.
540 *
541 * Give a start command to the receive BMU.
542 * Start the transmit BMUs if transmit frames pending.
543 *
544 * END_MANUAL_ENTRY
545 */
546void mac_drv_repair_descr(struct s_smc *smc)
547{
548 u_long phys ;
549
550 if (smc->hw.hw_state != STOPPED) {
551 SK_BREAK() ;
552 SMT_PANIC(smc,HWM_E0013,HWM_E0013_MSG) ;
553 return ;
554 }
555
556 /*
557 * repair tx queues: don't start
558 */
559 phys = repair_txd_ring(smc,smc->hw.fp.tx[QUEUE_A0]) ;
560 outpd(ADDR(B5_XA_DA),phys) ;
561 if (smc->hw.fp.tx_q[QUEUE_A0].tx_used) {
562 outpd(ADDR(B0_XA_CSR),CSR_START) ;
563 }
564 phys = repair_txd_ring(smc,smc->hw.fp.tx[QUEUE_S]) ;
565 outpd(ADDR(B5_XS_DA),phys) ;
566 if (smc->hw.fp.tx_q[QUEUE_S].tx_used) {
567 outpd(ADDR(B0_XS_CSR),CSR_START) ;
568 }
569
570 /*
571 * repair rx queues
572 */
573 phys = repair_rxd_ring(smc,smc->hw.fp.rx[QUEUE_R1]) ;
574 outpd(ADDR(B4_R1_DA),phys) ;
575 outpd(ADDR(B0_R1_CSR),CSR_START) ;
576}
577
578static u_long repair_txd_ring(struct s_smc *smc, struct s_smt_tx_queue *queue)
579{
580 int i ;
581 int tx_used ;
582 u_long phys ;
583 u_long tbctrl ;
584 struct s_smt_fp_txd volatile *t ;
585
586 SK_UNUSED(smc) ;
587
588 t = queue->tx_curr_get ;
589 tx_used = queue->tx_used ;
590 for (i = tx_used+queue->tx_free-1 ; i ; i-- ) {
591 t = t->txd_next ;
592 }
593 phys = le32_to_cpu(t->txd_ntdadr) ;
594
595 t = queue->tx_curr_get ;
596 while (tx_used) {
597 DRV_BUF_FLUSH(t,DDI_DMA_SYNC_FORCPU) ;
598 tbctrl = le32_to_cpu(t->txd_tbctrl) ;
599
600 if (tbctrl & BMU_OWN) {
601 if (tbctrl & BMU_STF) {
602 break ; /* exit the loop */
603 }
604 else {
605 /*
606 * repair the descriptor
607 */
608 t->txd_tbctrl &= ~cpu_to_le32(BMU_OWN) ;
609 }
610 }
611 phys = le32_to_cpu(t->txd_ntdadr) ;
612 DRV_BUF_FLUSH(t,DDI_DMA_SYNC_FORDEV) ;
613 t = t->txd_next ;
614 tx_used-- ;
615 }
616 return phys;
617}
618
619/*
620 * Repairs the receive descriptor ring and returns the physical address
621 * where the BMU should continue working.
622 *
623 * o The physical address where the BMU was stopped has to be
624 * determined. This is the next RxD after rx_curr_get with an OWN
625 * bit set.
626 * o The BMU should start working at beginning of the next frame.
627 * RxDs with an OWN bit set but with a reset STF bit should be
628 * skipped and owned by the driver (OWN = 0).
629 */
630static u_long repair_rxd_ring(struct s_smc *smc, struct s_smt_rx_queue *queue)
631{
632 int i ;
633 int rx_used ;
634 u_long phys ;
635 u_long rbctrl ;
636 struct s_smt_fp_rxd volatile *r ;
637
638 SK_UNUSED(smc) ;
639
640 r = queue->rx_curr_get ;
641 rx_used = queue->rx_used ;
642 for (i = SMT_R1_RXD_COUNT-1 ; i ; i-- ) {
643 r = r->rxd_next ;
644 }
645 phys = le32_to_cpu(r->rxd_nrdadr) ;
646
647 r = queue->rx_curr_get ;
648 while (rx_used) {
649 DRV_BUF_FLUSH(r,DDI_DMA_SYNC_FORCPU) ;
650 rbctrl = le32_to_cpu(r->rxd_rbctrl) ;
651
652 if (rbctrl & BMU_OWN) {
653 if (rbctrl & BMU_STF) {
654 break ; /* exit the loop */
655 }
656 else {
657 /*
658 * repair the descriptor
659 */
660 r->rxd_rbctrl &= ~cpu_to_le32(BMU_OWN) ;
661 }
662 }
663 phys = le32_to_cpu(r->rxd_nrdadr) ;
664 DRV_BUF_FLUSH(r,DDI_DMA_SYNC_FORDEV) ;
665 r = r->rxd_next ;
666 rx_used-- ;
667 }
668 return phys;
669}
670
671
672/*
673 -------------------------------------------------------------
674 INTERRUPT SERVICE ROUTINE:
675 -------------------------------------------------------------
676*/
677
678/*
679 * BEGIN_MANUAL_ENTRY(fddi_isr)
680 * void fddi_isr(smc)
681 *
682 * function DOWNCALL (drvsr.c)
683 * interrupt service routine, handles the interrupt requests
684 * generated by the FDDI adapter.
685 *
686 * NOTE: The operating system dependent module must guarantee that the
687 * interrupts of the adapter are disabled when it calls fddi_isr.
688 *
689 * About the USE_BREAK_ISR mechanismn:
690 *
691 * The main requirement of this mechanismn is to force an timer IRQ when
692 * leaving process_receive() with leave_isr set. process_receive() may
693 * be called at any time from anywhere!
694 * To be sure we don't miss such event we set 'force_irq' per default.
695 * We have to force and Timer IRQ if 'smc->os.hwm.leave_isr' AND
696 * 'force_irq' are set. 'force_irq' may be reset if a receive complete
697 * IRQ is pending.
698 *
699 * END_MANUAL_ENTRY
700 */
701void fddi_isr(struct s_smc *smc)
702{
703 u_long is ; /* ISR source */
704 u_short stu, stl ;
705 SMbuf *mb ;
706
707#ifdef USE_BREAK_ISR
708 int force_irq ;
709#endif
710
711#ifdef ODI2
712 if (smc->os.hwm.rx_break) {
713 mac_drv_fill_rxd(smc) ;
714 if (smc->hw.fp.rx_q[QUEUE_R1].rx_used > 0) {
715 smc->os.hwm.rx_break = 0 ;
716 process_receive(smc) ;
717 }
718 else {
719 smc->os.hwm.detec_count = 0 ;
720 smt_force_irq(smc) ;
721 }
722 }
723#endif
724 smc->os.hwm.isr_flag = TRUE ;
725
726#ifdef USE_BREAK_ISR
727 force_irq = TRUE ;
728 if (smc->os.hwm.leave_isr) {
729 smc->os.hwm.leave_isr = FALSE ;
730 process_receive(smc) ;
731 }
732#endif
733
734 while ((is = GET_ISR() & ISR_MASK)) {
735 NDD_TRACE("CH0B",is,0,0) ;
736 DB_GEN(7, "ISA = 0x%lx", is);
737
738 if (is & IMASK_SLOW) {
739 NDD_TRACE("CH1b",is,0,0) ;
740 if (is & IS_PLINT1) { /* PLC1 */
741 plc1_irq(smc) ;
742 }
743 if (is & IS_PLINT2) { /* PLC2 */
744 plc2_irq(smc) ;
745 }
746 if (is & IS_MINTR1) { /* FORMAC+ STU1(U/L) */
747 stu = inpw(FM_A(FM_ST1U)) ;
748 stl = inpw(FM_A(FM_ST1L)) ;
749 DB_GEN(6, "Slow transmit complete");
750 mac1_irq(smc,stu,stl) ;
751 }
752 if (is & IS_MINTR2) { /* FORMAC+ STU2(U/L) */
753 stu= inpw(FM_A(FM_ST2U)) ;
754 stl= inpw(FM_A(FM_ST2L)) ;
755 DB_GEN(6, "Slow receive complete");
756 DB_GEN(7, "stl = %x : stu = %x", stl, stu);
757 mac2_irq(smc,stu,stl) ;
758 }
759 if (is & IS_MINTR3) { /* FORMAC+ STU3(U/L) */
760 stu= inpw(FM_A(FM_ST3U)) ;
761 stl= inpw(FM_A(FM_ST3L)) ;
762 DB_GEN(6, "FORMAC Mode Register 3");
763 mac3_irq(smc,stu,stl) ;
764 }
765 if (is & IS_TIMINT) { /* Timer 82C54-2 */
766 timer_irq(smc) ;
767#ifdef NDIS_OS2
768 force_irq_pending = 0 ;
769#endif
770 /*
771 * out of RxD detection
772 */
773 if (++smc->os.hwm.detec_count > 4) {
774 /*
775 * check out of RxD condition
776 */
777 process_receive(smc) ;
778 }
779 }
780 if (is & IS_TOKEN) { /* Restricted Token Monitor */
781 rtm_irq(smc) ;
782 }
783 if (is & IS_R1_P) { /* Parity error rx queue 1 */
784 /* clear IRQ */
785 outpd(ADDR(B4_R1_CSR),CSR_IRQ_CL_P) ;
786 SMT_PANIC(smc,HWM_E0004,HWM_E0004_MSG) ;
787 }
788 if (is & IS_R1_C) { /* Encoding error rx queue 1 */
789 /* clear IRQ */
790 outpd(ADDR(B4_R1_CSR),CSR_IRQ_CL_C) ;
791 SMT_PANIC(smc,HWM_E0005,HWM_E0005_MSG) ;
792 }
793 if (is & IS_XA_C) { /* Encoding error async tx q */
794 /* clear IRQ */
795 outpd(ADDR(B5_XA_CSR),CSR_IRQ_CL_C) ;
796 SMT_PANIC(smc,HWM_E0006,HWM_E0006_MSG) ;
797 }
798 if (is & IS_XS_C) { /* Encoding error sync tx q */
799 /* clear IRQ */
800 outpd(ADDR(B5_XS_CSR),CSR_IRQ_CL_C) ;
801 SMT_PANIC(smc,HWM_E0007,HWM_E0007_MSG) ;
802 }
803 }
804
805 /*
806 * Fast Tx complete Async/Sync Queue (BMU service)
807 */
808 if (is & (IS_XS_F|IS_XA_F)) {
809 DB_GEN(6, "Fast tx complete queue");
810 /*
811 * clear IRQ, Note: no IRQ is lost, because
812 * we always service both queues
813 */
814 outpd(ADDR(B5_XS_CSR),CSR_IRQ_CL_F) ;
815 outpd(ADDR(B5_XA_CSR),CSR_IRQ_CL_F) ;
816 mac_drv_clear_txd(smc) ;
817 llc_restart_tx(smc) ;
818 }
819
820 /*
821 * Fast Rx Complete (BMU service)
822 */
823 if (is & IS_R1_F) {
824 DB_GEN(6, "Fast receive complete");
825 /* clear IRQ */
826#ifndef USE_BREAK_ISR
827 outpd(ADDR(B4_R1_CSR),CSR_IRQ_CL_F) ;
828 process_receive(smc) ;
829#else
830 process_receive(smc) ;
831 if (smc->os.hwm.leave_isr) {
832 force_irq = FALSE ;
833 } else {
834 outpd(ADDR(B4_R1_CSR),CSR_IRQ_CL_F) ;
835 process_receive(smc) ;
836 }
837#endif
838 }
839
840#ifndef NDIS_OS2
841 while ((mb = get_llc_rx(smc))) {
842 smt_to_llc(smc,mb) ;
843 }
844#else
845 if (offDepth)
846 post_proc() ;
847
848 while (!offDepth && (mb = get_llc_rx(smc))) {
849 smt_to_llc(smc,mb) ;
850 }
851
852 if (!offDepth && smc->os.hwm.rx_break) {
853 process_receive(smc) ;
854 }
855#endif
856 if (smc->q.ev_get != smc->q.ev_put) {
857 NDD_TRACE("CH2a",0,0,0) ;
858 ev_dispatcher(smc) ;
859 }
860#ifdef NDIS_OS2
861 post_proc() ;
862 if (offDepth) { /* leave fddi_isr because */
863 break ; /* indications not allowed */
864 }
865#endif
866#ifdef USE_BREAK_ISR
867 if (smc->os.hwm.leave_isr) {
868 break ; /* leave fddi_isr */
869 }
870#endif
871
872 /* NOTE: when the isr is left, no rx is pending */
873 } /* end of interrupt source polling loop */
874
875#ifdef USE_BREAK_ISR
876 if (smc->os.hwm.leave_isr && force_irq) {
877 smt_force_irq(smc) ;
878 }
879#endif
880 smc->os.hwm.isr_flag = FALSE ;
881 NDD_TRACE("CH0E",0,0,0) ;
882}
883
884
885/*
886 -------------------------------------------------------------
887 RECEIVE FUNCTIONS:
888 -------------------------------------------------------------
889*/
890
891#ifndef NDIS_OS2
892/*
893 * BEGIN_MANUAL_ENTRY(mac_drv_rx_mode)
894 * void mac_drv_rx_mode(smc,mode)
895 *
896 * function DOWNCALL (fplus.c)
897 * Corresponding to the parameter mode, the operating system
898 * dependent module can activate several receive modes.
899 *
900 * para mode = 1: RX_ENABLE_ALLMULTI enable all multicasts
901 * = 2: RX_DISABLE_ALLMULTI disable "enable all multicasts"
902 * = 3: RX_ENABLE_PROMISC enable promiscuous
903 * = 4: RX_DISABLE_PROMISC disable promiscuous
904 * = 5: RX_ENABLE_NSA enable rec. of all NSA frames
905 * (disabled after 'driver reset' & 'set station address')
906 * = 6: RX_DISABLE_NSA disable rec. of all NSA frames
907 *
908 * = 21: RX_ENABLE_PASS_SMT ( see description )
909 * = 22: RX_DISABLE_PASS_SMT ( " " )
910 * = 23: RX_ENABLE_PASS_NSA ( " " )
911 * = 24: RX_DISABLE_PASS_NSA ( " " )
912 * = 25: RX_ENABLE_PASS_DB ( " " )
913 * = 26: RX_DISABLE_PASS_DB ( " " )
914 * = 27: RX_DISABLE_PASS_ALL ( " " )
915 * = 28: RX_DISABLE_LLC_PROMISC ( " " )
916 * = 29: RX_ENABLE_LLC_PROMISC ( " " )
917 *
918 *
919 * RX_ENABLE_PASS_SMT / RX_DISABLE_PASS_SMT
920 *
921 * If the operating system dependent module activates the
922 * mode RX_ENABLE_PASS_SMT, the hardware module
923 * duplicates all SMT frames with the frame control
924 * FC_SMT_INFO and passes them to the LLC receive channel
925 * by calling mac_drv_rx_init.
926 * The SMT Frames which are sent by the local SMT and the NSA
927 * frames whose A- and C-Indicator is not set are also duplicated
928 * and passed.
929 * The receive mode RX_DISABLE_PASS_SMT disables the passing
930 * of SMT frames.
931 *
932 * RX_ENABLE_PASS_NSA / RX_DISABLE_PASS_NSA
933 *
934 * If the operating system dependent module activates the
935 * mode RX_ENABLE_PASS_NSA, the hardware module
936 * duplicates all NSA frames with frame control FC_SMT_NSA
937 * and a set A-Indicator and passed them to the LLC
938 * receive channel by calling mac_drv_rx_init.
939 * All NSA Frames which are sent by the local SMT
940 * are also duplicated and passed.
941 * The receive mode RX_DISABLE_PASS_NSA disables the passing
942 * of NSA frames with the A- or C-Indicator set.
943 *
944 * NOTE: For fear that the hardware module receives NSA frames with
945 * a reset A-Indicator, the operating system dependent module
946 * has to call mac_drv_rx_mode with the mode RX_ENABLE_NSA
947 * before activate the RX_ENABLE_PASS_NSA mode and after every
948 * 'driver reset' and 'set station address'.
949 *
950 * RX_ENABLE_PASS_DB / RX_DISABLE_PASS_DB
951 *
952 * If the operating system dependent module activates the
953 * mode RX_ENABLE_PASS_DB, direct BEACON frames
954 * (FC_BEACON frame control) are passed to the LLC receive
955 * channel by mac_drv_rx_init.
956 * The receive mode RX_DISABLE_PASS_DB disables the passing
957 * of direct BEACON frames.
958 *
959 * RX_DISABLE_PASS_ALL
960 *
961 * Disables all special receives modes. It is equal to
962 * call mac_drv_set_rx_mode successively with the
963 * parameters RX_DISABLE_NSA, RX_DISABLE_PASS_SMT,
964 * RX_DISABLE_PASS_NSA and RX_DISABLE_PASS_DB.
965 *
966 * RX_ENABLE_LLC_PROMISC
967 *
968 * (default) all received LLC frames and all SMT/NSA/DBEACON
969 * frames depending on the attitude of the flags
970 * PASS_SMT/PASS_NSA/PASS_DBEACON will be delivered to the
971 * LLC layer
972 *
973 * RX_DISABLE_LLC_PROMISC
974 *
975 * all received SMT/NSA/DBEACON frames depending on the
976 * attitude of the flags PASS_SMT/PASS_NSA/PASS_DBEACON
977 * will be delivered to the LLC layer.
978 * all received LLC frames with a directed address, Multicast
979 * or Broadcast address will be delivered to the LLC
980 * layer too.
981 *
982 * END_MANUAL_ENTRY
983 */
984void mac_drv_rx_mode(struct s_smc *smc, int mode)
985{
986 switch(mode) {
987 case RX_ENABLE_PASS_SMT:
988 smc->os.hwm.pass_SMT = TRUE ;
989 break ;
990 case RX_DISABLE_PASS_SMT:
991 smc->os.hwm.pass_SMT = FALSE ;
992 break ;
993 case RX_ENABLE_PASS_NSA:
994 smc->os.hwm.pass_NSA = TRUE ;
995 break ;
996 case RX_DISABLE_PASS_NSA:
997 smc->os.hwm.pass_NSA = FALSE ;
998 break ;
999 case RX_ENABLE_PASS_DB:
1000 smc->os.hwm.pass_DB = TRUE ;
1001 break ;
1002 case RX_DISABLE_PASS_DB:
1003 smc->os.hwm.pass_DB = FALSE ;
1004 break ;
1005 case RX_DISABLE_PASS_ALL:
1006 smc->os.hwm.pass_SMT = smc->os.hwm.pass_NSA = FALSE ;
1007 smc->os.hwm.pass_DB = FALSE ;
1008 smc->os.hwm.pass_llc_promisc = TRUE ;
1009 mac_set_rx_mode(smc,RX_DISABLE_NSA) ;
1010 break ;
1011 case RX_DISABLE_LLC_PROMISC:
1012 smc->os.hwm.pass_llc_promisc = FALSE ;
1013 break ;
1014 case RX_ENABLE_LLC_PROMISC:
1015 smc->os.hwm.pass_llc_promisc = TRUE ;
1016 break ;
1017 case RX_ENABLE_ALLMULTI:
1018 case RX_DISABLE_ALLMULTI:
1019 case RX_ENABLE_PROMISC:
1020 case RX_DISABLE_PROMISC:
1021 case RX_ENABLE_NSA:
1022 case RX_DISABLE_NSA:
1023 default:
1024 mac_set_rx_mode(smc,mode) ;
1025 break ;
1026 }
1027}
1028#endif /* ifndef NDIS_OS2 */
1029
1030/*
1031 * process receive queue
1032 */
1033void process_receive(struct s_smc *smc)
1034{
1035 int i ;
1036 int n ;
1037 int frag_count ; /* number of RxDs of the curr rx buf */
1038 int used_frags ; /* number of RxDs of the curr frame */
1039 struct s_smt_rx_queue *queue ; /* points to the queue ctl struct */
1040 struct s_smt_fp_rxd volatile *r ; /* rxd pointer */
1041 struct s_smt_fp_rxd volatile *rxd ; /* first rxd of rx frame */
1042 u_long rbctrl ; /* receive buffer control word */
1043 u_long rfsw ; /* receive frame status word */
1044 u_short rx_used ;
1045 u_char far *virt ;
1046 char far *data ;
1047 SMbuf *mb ;
1048 u_char fc ; /* Frame control */
1049 int len ; /* Frame length */
1050
1051 smc->os.hwm.detec_count = 0 ;
1052 queue = smc->hw.fp.rx[QUEUE_R1] ;
1053 NDD_TRACE("RHxB",0,0,0) ;
1054 for ( ; ; ) {
1055 r = queue->rx_curr_get ;
1056 rx_used = queue->rx_used ;
1057 frag_count = 0 ;
1058
1059#ifdef USE_BREAK_ISR
1060 if (smc->os.hwm.leave_isr) {
1061 goto rx_end ;
1062 }
1063#endif
1064#ifdef NDIS_OS2
1065 if (offDepth) {
1066 smc->os.hwm.rx_break = 1 ;
1067 goto rx_end ;
1068 }
1069 smc->os.hwm.rx_break = 0 ;
1070#endif
1071#ifdef ODI2
1072 if (smc->os.hwm.rx_break) {
1073 goto rx_end ;
1074 }
1075#endif
1076 n = 0 ;
1077 do {
1078 DB_RX(5, "Check RxD %p for OWN and EOF", r);
1079 DRV_BUF_FLUSH(r,DDI_DMA_SYNC_FORCPU) ;
1080 rbctrl = le32_to_cpu(CR_READ(r->rxd_rbctrl));
1081
1082 if (rbctrl & BMU_OWN) {
1083 NDD_TRACE("RHxE",r,rfsw,rbctrl) ;
1084 DB_RX(4, "End of RxDs");
1085 goto rx_end ;
1086 }
1087 /*
1088 * out of RxD detection
1089 */
1090 if (!rx_used) {
1091 SK_BREAK() ;
1092 SMT_PANIC(smc,HWM_E0009,HWM_E0009_MSG) ;
1093 /* Either we don't have an RxD or all
1094 * RxDs are filled. Therefore it's allowed
1095 * for to set the STOPPED flag */
1096 smc->hw.hw_state = STOPPED ;
1097 mac_drv_clear_rx_queue(smc) ;
1098 smc->hw.hw_state = STARTED ;
1099 mac_drv_fill_rxd(smc) ;
1100 smc->os.hwm.detec_count = 0 ;
1101 goto rx_end ;
1102 }
1103 rfsw = le32_to_cpu(r->rxd_rfsw) ;
1104 if ((rbctrl & BMU_STF) != ((rbctrl & BMU_ST_BUF) <<5)) {
1105 /*
1106 * The BMU_STF bit is deleted, 1 frame is
1107 * placed into more than 1 rx buffer
1108 *
1109 * skip frame by setting the rx len to 0
1110 *
1111 * if fragment count == 0
1112 * The missing STF bit belongs to the
1113 * current frame, search for the
1114 * EOF bit to complete the frame
1115 * else
1116 * the fragment belongs to the next frame,
1117 * exit the loop and process the frame
1118 */
1119 SK_BREAK() ;
1120 rfsw = 0 ;
1121 if (frag_count) {
1122 break ;
1123 }
1124 }
1125 n += rbctrl & 0xffff ;
1126 r = r->rxd_next ;
1127 frag_count++ ;
1128 rx_used-- ;
1129 } while (!(rbctrl & BMU_EOF)) ;
1130 used_frags = frag_count ;
1131 DB_RX(5, "EOF set in RxD, used_frags = %d", used_frags);
1132
1133 /* may be next 2 DRV_BUF_FLUSH() can be skipped, because */
1134 /* BMU_ST_BUF will not be changed by the ASIC */
1135 DRV_BUF_FLUSH(r,DDI_DMA_SYNC_FORCPU) ;
1136 while (rx_used && !(r->rxd_rbctrl & cpu_to_le32(BMU_ST_BUF))) {
1137 DB_RX(5, "Check STF bit in %p", r);
1138 r = r->rxd_next ;
1139 DRV_BUF_FLUSH(r,DDI_DMA_SYNC_FORCPU) ;
1140 frag_count++ ;
1141 rx_used-- ;
1142 }
1143 DB_RX(5, "STF bit found");
1144
1145 /*
1146 * The received frame is finished for the process receive
1147 */
1148 rxd = queue->rx_curr_get ;
1149 queue->rx_curr_get = r ;
1150 queue->rx_free += frag_count ;
1151 queue->rx_used = rx_used ;
1152
1153 /*
1154 * ASIC Errata no. 7 (STF - Bit Bug)
1155 */
1156 rxd->rxd_rbctrl &= cpu_to_le32(~BMU_STF) ;
1157
1158 for (r=rxd, i=frag_count ; i ; r=r->rxd_next, i--){
1159 DB_RX(5, "dma_complete for RxD %p", r);
1160 dma_complete(smc,(union s_fp_descr volatile *)r,DMA_WR);
1161 }
1162 smc->hw.fp.err_stats.err_valid++ ;
1163 smc->mib.m[MAC0].fddiMACCopied_Ct++ ;
1164
1165 /* the length of the data including the FC */
1166 len = (rfsw & RD_LENGTH) - 4 ;
1167
1168 DB_RX(4, "frame length = %d", len);
1169 /*
1170 * check the frame_length and all error flags
1171 */
1172 if (rfsw & (RX_MSRABT|RX_FS_E|RX_FS_CRC|RX_FS_IMPL)){
1173 if (rfsw & RD_S_MSRABT) {
1174 DB_RX(2, "Frame aborted by the FORMAC");
1175 smc->hw.fp.err_stats.err_abort++ ;
1176 }
1177 /*
1178 * check frame status
1179 */
1180 if (rfsw & RD_S_SEAC2) {
1181 DB_RX(2, "E-Indicator set");
1182 smc->hw.fp.err_stats.err_e_indicator++ ;
1183 }
1184 if (rfsw & RD_S_SFRMERR) {
1185 DB_RX(2, "CRC error");
1186 smc->hw.fp.err_stats.err_crc++ ;
1187 }
1188 if (rfsw & RX_FS_IMPL) {
1189 DB_RX(2, "Implementer frame");
1190 smc->hw.fp.err_stats.err_imp_frame++ ;
1191 }
1192 goto abort_frame ;
1193 }
1194 if (len > FDDI_RAW_MTU-4) {
1195 DB_RX(2, "Frame too long error");
1196 smc->hw.fp.err_stats.err_too_long++ ;
1197 goto abort_frame ;
1198 }
1199 /*
1200 * SUPERNET 3 Bug: FORMAC delivers status words
1201 * of aborted frames to the BMU
1202 */
1203 if (len <= 4) {
1204 DB_RX(2, "Frame length = 0");
1205 goto abort_frame ;
1206 }
1207
1208 if (len != (n-4)) {
1209 DB_RX(4, "BMU: rx len differs: [%d:%d]", len, n);
1210 smc->os.hwm.rx_len_error++ ;
1211 goto abort_frame ;
1212 }
1213
1214 /*
1215 * Check SA == MA
1216 */
1217 virt = (u_char far *) rxd->rxd_virt ;
1218 DB_RX(2, "FC = %x", *virt);
1219 if (virt[12] == MA[5] &&
1220 virt[11] == MA[4] &&
1221 virt[10] == MA[3] &&
1222 virt[9] == MA[2] &&
1223 virt[8] == MA[1] &&
1224 (virt[7] & ~GROUP_ADDR_BIT) == MA[0]) {
1225 goto abort_frame ;
1226 }
1227
1228 /*
1229 * test if LLC frame
1230 */
1231 if (rfsw & RX_FS_LLC) {
1232 /*
1233 * if pass_llc_promisc is disable
1234 * if DA != Multicast or Broadcast or DA!=MA
1235 * abort the frame
1236 */
1237 if (!smc->os.hwm.pass_llc_promisc) {
1238 if(!(virt[1] & GROUP_ADDR_BIT)) {
1239 if (virt[6] != MA[5] ||
1240 virt[5] != MA[4] ||
1241 virt[4] != MA[3] ||
1242 virt[3] != MA[2] ||
1243 virt[2] != MA[1] ||
1244 virt[1] != MA[0]) {
1245 DB_RX(2, "DA != MA and not multi- or broadcast");
1246 goto abort_frame ;
1247 }
1248 }
1249 }
1250
1251 /*
1252 * LLC frame received
1253 */
1254 DB_RX(4, "LLC - receive");
1255 mac_drv_rx_complete(smc,rxd,frag_count,len) ;
1256 }
1257 else {
1258 if (!(mb = smt_get_mbuf(smc))) {
1259 smc->hw.fp.err_stats.err_no_buf++ ;
1260 DB_RX(4, "No SMbuf; receive terminated");
1261 goto abort_frame ;
1262 }
1263 data = smtod(mb,char *) - 1 ;
1264
1265 /*
1266 * copy the frame into a SMT_MBuf
1267 */
1268#ifdef USE_OS_CPY
1269 hwm_cpy_rxd2mb(rxd,data,len) ;
1270#else
1271 for (r=rxd, i=used_frags ; i ; r=r->rxd_next, i--){
1272 n = le32_to_cpu(r->rxd_rbctrl) & RD_LENGTH ;
1273 DB_RX(6, "cp SMT frame to mb: len = %d", n);
1274 memcpy(data,r->rxd_virt,n) ;
1275 data += n ;
1276 }
1277 data = smtod(mb,char *) - 1 ;
1278#endif
1279 fc = *(char *)mb->sm_data = *data ;
1280 mb->sm_len = len - 1 ; /* len - fc */
1281 data++ ;
1282
1283 /*
1284 * SMT frame received
1285 */
1286 switch(fc) {
1287 case FC_SMT_INFO :
1288 smc->hw.fp.err_stats.err_smt_frame++ ;
1289 DB_RX(5, "SMT frame received");
1290
1291 if (smc->os.hwm.pass_SMT) {
1292 DB_RX(5, "pass SMT frame");
1293 mac_drv_rx_complete(smc, rxd,
1294 frag_count,len) ;
1295 }
1296 else {
1297 DB_RX(5, "requeue RxD");
1298 mac_drv_requeue_rxd(smc,rxd,frag_count);
1299 }
1300
1301 smt_received_pack(smc,mb,(int)(rfsw>>25)) ;
1302 break ;
1303 case FC_SMT_NSA :
1304 smc->hw.fp.err_stats.err_smt_frame++ ;
1305 DB_RX(5, "SMT frame received");
1306
1307 /* if pass_NSA set pass the NSA frame or */
1308 /* pass_SMT set and the A-Indicator */
1309 /* is not set, pass the NSA frame */
1310 if (smc->os.hwm.pass_NSA ||
1311 (smc->os.hwm.pass_SMT &&
1312 !(rfsw & A_INDIC))) {
1313 DB_RX(5, "pass SMT frame");
1314 mac_drv_rx_complete(smc, rxd,
1315 frag_count,len) ;
1316 }
1317 else {
1318 DB_RX(5, "requeue RxD");
1319 mac_drv_requeue_rxd(smc,rxd,frag_count);
1320 }
1321
1322 smt_received_pack(smc,mb,(int)(rfsw>>25)) ;
1323 break ;
1324 case FC_BEACON :
1325 if (smc->os.hwm.pass_DB) {
1326 DB_RX(5, "pass DB frame");
1327 mac_drv_rx_complete(smc, rxd,
1328 frag_count,len) ;
1329 }
1330 else {
1331 DB_RX(5, "requeue RxD");
1332 mac_drv_requeue_rxd(smc,rxd,frag_count);
1333 }
1334 smt_free_mbuf(smc,mb) ;
1335 break ;
1336 default :
1337 /*
1338 * unknown FC abort the frame
1339 */
1340 DB_RX(2, "unknown FC error");
1341 smt_free_mbuf(smc,mb) ;
1342 DB_RX(5, "requeue RxD");
1343 mac_drv_requeue_rxd(smc,rxd,frag_count) ;
1344 if ((fc & 0xf0) == FC_MAC)
1345 smc->hw.fp.err_stats.err_mac_frame++ ;
1346 else
1347 smc->hw.fp.err_stats.err_imp_frame++ ;
1348
1349 break ;
1350 }
1351 }
1352
1353 DB_RX(3, "next RxD is %p", queue->rx_curr_get);
1354 NDD_TRACE("RHx1",queue->rx_curr_get,0,0) ;
1355
1356 continue ;
1357 /*--------------------------------------------------------------------*/
1358abort_frame:
1359 DB_RX(5, "requeue RxD");
1360 mac_drv_requeue_rxd(smc,rxd,frag_count) ;
1361
1362 DB_RX(3, "next RxD is %p", queue->rx_curr_get);
1363 NDD_TRACE("RHx2",queue->rx_curr_get,0,0) ;
1364 }
1365rx_end:
1366#ifdef ALL_RX_COMPLETE
1367 mac_drv_all_receives_complete(smc) ;
1368#endif
1369 return ; /* lint bug: needs return detect end of function */
1370}
1371
1372static void smt_to_llc(struct s_smc *smc, SMbuf *mb)
1373{
1374 u_char fc ;
1375
1376 DB_RX(4, "send a queued frame to the llc layer");
1377 smc->os.hwm.r.len = mb->sm_len ;
1378 smc->os.hwm.r.mb_pos = smtod(mb,char *) ;
1379 fc = *smc->os.hwm.r.mb_pos ;
1380 (void)mac_drv_rx_init(smc,(int)mb->sm_len,(int)fc,
1381 smc->os.hwm.r.mb_pos,(int)mb->sm_len) ;
1382 smt_free_mbuf(smc,mb) ;
1383}
1384
1385/*
1386 * BEGIN_MANUAL_ENTRY(hwm_rx_frag)
1387 * void hwm_rx_frag(smc,virt,phys,len,frame_status)
1388 *
1389 * function MACRO (hardware module, hwmtm.h)
1390 * This function calls dma_master for preparing the
1391 * system hardware for the DMA transfer and initializes
1392 * the current RxD with the length and the physical and
1393 * virtual address of the fragment. Furthermore, it sets the
1394 * STF and EOF bits depending on the frame status byte,
1395 * switches the OWN flag of the RxD, so that it is owned by the
1396 * adapter and issues an rx_start.
1397 *
1398 * para virt virtual pointer to the fragment
1399 * len the length of the fragment
1400 * frame_status status of the frame, see design description
1401 *
1402 * NOTE: It is possible to call this function with a fragment length
1403 * of zero.
1404 *
1405 * END_MANUAL_ENTRY
1406 */
1407void hwm_rx_frag(struct s_smc *smc, char far *virt, u_long phys, int len,
1408 int frame_status)
1409{
1410 struct s_smt_fp_rxd volatile *r ;
1411 __le32 rbctrl;
1412
1413 NDD_TRACE("RHfB",virt,len,frame_status) ;
1414 DB_RX(2, "hwm_rx_frag: len = %d, frame_status = %x", len, frame_status);
1415 r = smc->hw.fp.rx_q[QUEUE_R1].rx_curr_put ;
1416 r->rxd_virt = virt ;
1417 r->rxd_rbadr = cpu_to_le32(phys) ;
1418 rbctrl = cpu_to_le32( (((__u32)frame_status &
1419 (FIRST_FRAG|LAST_FRAG))<<26) |
1420 (((u_long) frame_status & FIRST_FRAG) << 21) |
1421 BMU_OWN | BMU_CHECK | BMU_EN_IRQ_EOF | len) ;
1422 r->rxd_rbctrl = rbctrl ;
1423
1424 DRV_BUF_FLUSH(r,DDI_DMA_SYNC_FORDEV) ;
1425 outpd(ADDR(B0_R1_CSR),CSR_START) ;
1426 smc->hw.fp.rx_q[QUEUE_R1].rx_free-- ;
1427 smc->hw.fp.rx_q[QUEUE_R1].rx_used++ ;
1428 smc->hw.fp.rx_q[QUEUE_R1].rx_curr_put = r->rxd_next ;
1429 NDD_TRACE("RHfE",r,le32_to_cpu(r->rxd_rbadr),0) ;
1430}
1431
1432/*
1433 * BEGINN_MANUAL_ENTRY(mac_drv_clear_rx_queue)
1434 *
1435 * void mac_drv_clear_rx_queue(smc)
1436 * struct s_smc *smc ;
1437 *
1438 * function DOWNCALL (hardware module, hwmtm.c)
1439 * mac_drv_clear_rx_queue is called by the OS-specific module
1440 * after it has issued a card_stop.
1441 * In this case, the frames in the receive queue are obsolete and
1442 * should be removed. For removing mac_drv_clear_rx_queue
1443 * calls dma_master for each RxD and mac_drv_clear_rxd for each
1444 * receive buffer.
1445 *
1446 * NOTE: calling sequence card_stop:
1447 * CLI_FBI(), card_stop(),
1448 * mac_drv_clear_tx_queue(), mac_drv_clear_rx_queue(),
1449 *
1450 * NOTE: The caller is responsible that the BMUs are idle
1451 * when this function is called.
1452 *
1453 * END_MANUAL_ENTRY
1454 */
1455void mac_drv_clear_rx_queue(struct s_smc *smc)
1456{
1457 struct s_smt_fp_rxd volatile *r ;
1458 struct s_smt_fp_rxd volatile *next_rxd ;
1459 struct s_smt_rx_queue *queue ;
1460 int frag_count ;
1461 int i ;
1462
1463 if (smc->hw.hw_state != STOPPED) {
1464 SK_BREAK() ;
1465 SMT_PANIC(smc,HWM_E0012,HWM_E0012_MSG) ;
1466 return ;
1467 }
1468
1469 queue = smc->hw.fp.rx[QUEUE_R1] ;
1470 DB_RX(5, "clear_rx_queue");
1471
1472 /*
1473 * dma_complete and mac_drv_clear_rxd for all RxDs / receive buffers
1474 */
1475 r = queue->rx_curr_get ;
1476 while (queue->rx_used) {
1477 DRV_BUF_FLUSH(r,DDI_DMA_SYNC_FORCPU) ;
1478 DB_RX(5, "switch OWN bit of RxD 0x%p", r);
1479 r->rxd_rbctrl &= ~cpu_to_le32(BMU_OWN) ;
1480 frag_count = 1 ;
1481 DRV_BUF_FLUSH(r,DDI_DMA_SYNC_FORDEV) ;
1482 r = r->rxd_next ;
1483 DRV_BUF_FLUSH(r,DDI_DMA_SYNC_FORCPU) ;
1484 while (r != queue->rx_curr_put &&
1485 !(r->rxd_rbctrl & cpu_to_le32(BMU_ST_BUF))) {
1486 DB_RX(5, "Check STF bit in %p", r);
1487 r->rxd_rbctrl &= ~cpu_to_le32(BMU_OWN) ;
1488 DRV_BUF_FLUSH(r,DDI_DMA_SYNC_FORDEV) ;
1489 r = r->rxd_next ;
1490 DRV_BUF_FLUSH(r,DDI_DMA_SYNC_FORCPU) ;
1491 frag_count++ ;
1492 }
1493 DB_RX(5, "STF bit found");
1494 next_rxd = r ;
1495
1496 for (r=queue->rx_curr_get,i=frag_count; i ; r=r->rxd_next,i--){
1497 DB_RX(5, "dma_complete for RxD %p", r);
1498 dma_complete(smc,(union s_fp_descr volatile *)r,DMA_WR);
1499 }
1500
1501 DB_RX(5, "mac_drv_clear_rxd: RxD %p frag_count %d",
1502 queue->rx_curr_get, frag_count);
1503 mac_drv_clear_rxd(smc,queue->rx_curr_get,frag_count) ;
1504
1505 queue->rx_curr_get = next_rxd ;
1506 queue->rx_used -= frag_count ;
1507 queue->rx_free += frag_count ;
1508 }
1509}
1510
1511
1512/*
1513 -------------------------------------------------------------
1514 SEND FUNCTIONS:
1515 -------------------------------------------------------------
1516*/
1517
1518/*
1519 * BEGIN_MANUAL_ENTRY(hwm_tx_init)
1520 * int hwm_tx_init(smc,fc,frag_count,frame_len,frame_status)
1521 *
1522 * function DOWN_CALL (hardware module, hwmtm.c)
1523 * hwm_tx_init checks if the frame can be sent through the
1524 * corresponding send queue.
1525 *
1526 * para fc the frame control. To determine through which
1527 * send queue the frame should be transmitted.
1528 * 0x50 - 0x57: asynchronous LLC frame
1529 * 0xD0 - 0xD7: synchronous LLC frame
1530 * 0x41, 0x4F: SMT frame to the network
1531 * 0x42: SMT frame to the network and to the local SMT
1532 * 0x43: SMT frame to the local SMT
1533 * frag_count count of the fragments for this frame
1534 * frame_len length of the frame
1535 * frame_status status of the frame, the send queue bit is already
1536 * specified
1537 *
1538 * return frame_status
1539 *
1540 * END_MANUAL_ENTRY
1541 */
1542int hwm_tx_init(struct s_smc *smc, u_char fc, int frag_count, int frame_len,
1543 int frame_status)
1544{
1545 NDD_TRACE("THiB",fc,frag_count,frame_len) ;
1546 smc->os.hwm.tx_p = smc->hw.fp.tx[frame_status & QUEUE_A0] ;
1547 smc->os.hwm.tx_descr = TX_DESCRIPTOR | (((u_long)(frame_len-1)&3)<<27) ;
1548 smc->os.hwm.tx_len = frame_len ;
1549 DB_TX(3, "hwm_tx_init: fc = %x, len = %d", fc, frame_len);
1550 if ((fc & ~(FC_SYNC_BIT|FC_LLC_PRIOR)) == FC_ASYNC_LLC) {
1551 frame_status |= LAN_TX ;
1552 }
1553 else {
1554 switch (fc) {
1555 case FC_SMT_INFO :
1556 case FC_SMT_NSA :
1557 frame_status |= LAN_TX ;
1558 break ;
1559 case FC_SMT_LOC :
1560 frame_status |= LOC_TX ;
1561 break ;
1562 case FC_SMT_LAN_LOC :
1563 frame_status |= LAN_TX | LOC_TX ;
1564 break ;
1565 default :
1566 SMT_PANIC(smc,HWM_E0010,HWM_E0010_MSG) ;
1567 }
1568 }
1569 if (!smc->hw.mac_ring_is_up) {
1570 frame_status &= ~LAN_TX ;
1571 frame_status |= RING_DOWN ;
1572 DB_TX(2, "Ring is down: terminate LAN_TX");
1573 }
1574 if (frag_count > smc->os.hwm.tx_p->tx_free) {
1575#ifndef NDIS_OS2
1576 mac_drv_clear_txd(smc) ;
1577 if (frag_count > smc->os.hwm.tx_p->tx_free) {
1578 DB_TX(2, "Out of TxDs, terminate LAN_TX");
1579 frame_status &= ~LAN_TX ;
1580 frame_status |= OUT_OF_TXD ;
1581 }
1582#else
1583 DB_TX(2, "Out of TxDs, terminate LAN_TX");
1584 frame_status &= ~LAN_TX ;
1585 frame_status |= OUT_OF_TXD ;
1586#endif
1587 }
1588 DB_TX(3, "frame_status = %x", frame_status);
1589 NDD_TRACE("THiE",frame_status,smc->os.hwm.tx_p->tx_free,0) ;
1590 return frame_status;
1591}
1592
1593/*
1594 * BEGIN_MANUAL_ENTRY(hwm_tx_frag)
1595 * void hwm_tx_frag(smc,virt,phys,len,frame_status)
1596 *
1597 * function DOWNCALL (hardware module, hwmtm.c)
1598 * If the frame should be sent to the LAN, this function calls
1599 * dma_master, fills the current TxD with the virtual and the
1600 * physical address, sets the STF and EOF bits dependent on
1601 * the frame status, and requests the BMU to start the
1602 * transmit.
1603 * If the frame should be sent to the local SMT, an SMT_MBuf
1604 * is allocated if the FIRST_FRAG bit is set in the frame_status.
1605 * The fragment of the frame is copied into the SMT MBuf.
1606 * The function smt_received_pack is called if the LAST_FRAG
1607 * bit is set in the frame_status word.
1608 *
1609 * para virt virtual pointer to the fragment
1610 * len the length of the fragment
1611 * frame_status status of the frame, see design description
1612 *
1613 * return nothing returned, no parameter is modified
1614 *
1615 * NOTE: It is possible to invoke this macro with a fragment length
1616 * of zero.
1617 *
1618 * END_MANUAL_ENTRY
1619 */
1620void hwm_tx_frag(struct s_smc *smc, char far *virt, u_long phys, int len,
1621 int frame_status)
1622{
1623 struct s_smt_fp_txd volatile *t ;
1624 struct s_smt_tx_queue *queue ;
1625 __le32 tbctrl ;
1626
1627 queue = smc->os.hwm.tx_p ;
1628
1629 NDD_TRACE("THfB",virt,len,frame_status) ;
1630 /* Bug fix: AF / May 31 1999 (#missing)
1631 * snmpinfo problem reported by IBM is caused by invalid
1632 * t-pointer (txd) if LAN_TX is not set but LOC_TX only.
1633 * Set: t = queue->tx_curr_put here !
1634 */
1635 t = queue->tx_curr_put ;
1636
1637 DB_TX(2, "hwm_tx_frag: len = %d, frame_status = %x", len, frame_status);
1638 if (frame_status & LAN_TX) {
1639 /* '*t' is already defined */
1640 DB_TX(3, "LAN_TX: TxD = %p, virt = %p", t, virt);
1641 t->txd_virt = virt ;
1642 t->txd_txdscr = cpu_to_le32(smc->os.hwm.tx_descr) ;
1643 t->txd_tbadr = cpu_to_le32(phys) ;
1644 tbctrl = cpu_to_le32((((__u32)frame_status &
1645 (FIRST_FRAG|LAST_FRAG|EN_IRQ_EOF))<< 26) |
1646 BMU_OWN|BMU_CHECK |len) ;
1647 t->txd_tbctrl = tbctrl ;
1648
1649#ifndef AIX
1650 DRV_BUF_FLUSH(t,DDI_DMA_SYNC_FORDEV) ;
1651 outpd(queue->tx_bmu_ctl,CSR_START) ;
1652#else /* ifndef AIX */
1653 DRV_BUF_FLUSH(t,DDI_DMA_SYNC_FORDEV) ;
1654 if (frame_status & QUEUE_A0) {
1655 outpd(ADDR(B0_XA_CSR),CSR_START) ;
1656 }
1657 else {
1658 outpd(ADDR(B0_XS_CSR),CSR_START) ;
1659 }
1660#endif
1661 queue->tx_free-- ;
1662 queue->tx_used++ ;
1663 queue->tx_curr_put = t->txd_next ;
1664 if (frame_status & LAST_FRAG) {
1665 smc->mib.m[MAC0].fddiMACTransmit_Ct++ ;
1666 }
1667 }
1668 if (frame_status & LOC_TX) {
1669 DB_TX(3, "LOC_TX:");
1670 if (frame_status & FIRST_FRAG) {
1671 if(!(smc->os.hwm.tx_mb = smt_get_mbuf(smc))) {
1672 smc->hw.fp.err_stats.err_no_buf++ ;
1673 DB_TX(4, "No SMbuf; transmit terminated");
1674 }
1675 else {
1676 smc->os.hwm.tx_data =
1677 smtod(smc->os.hwm.tx_mb,char *) - 1 ;
1678#ifdef USE_OS_CPY
1679#ifdef PASS_1ST_TXD_2_TX_COMP
1680 hwm_cpy_txd2mb(t,smc->os.hwm.tx_data,
1681 smc->os.hwm.tx_len) ;
1682#endif
1683#endif
1684 }
1685 }
1686 if (smc->os.hwm.tx_mb) {
1687#ifndef USE_OS_CPY
1688 DB_TX(3, "copy fragment into MBuf");
1689 memcpy(smc->os.hwm.tx_data,virt,len) ;
1690 smc->os.hwm.tx_data += len ;
1691#endif
1692 if (frame_status & LAST_FRAG) {
1693#ifdef USE_OS_CPY
1694#ifndef PASS_1ST_TXD_2_TX_COMP
1695 /*
1696 * hwm_cpy_txd2mb(txd,data,len) copies 'len'
1697 * bytes from the virtual pointer in 'rxd'
1698 * to 'data'. The virtual pointer of the
1699 * os-specific tx-buffer should be written
1700 * in the LAST txd.
1701 */
1702 hwm_cpy_txd2mb(t,smc->os.hwm.tx_data,
1703 smc->os.hwm.tx_len) ;
1704#endif /* nPASS_1ST_TXD_2_TX_COMP */
1705#endif /* USE_OS_CPY */
1706 smc->os.hwm.tx_data =
1707 smtod(smc->os.hwm.tx_mb,char *) - 1 ;
1708 *(char *)smc->os.hwm.tx_mb->sm_data =
1709 *smc->os.hwm.tx_data ;
1710 smc->os.hwm.tx_data++ ;
1711 smc->os.hwm.tx_mb->sm_len =
1712 smc->os.hwm.tx_len - 1 ;
1713 DB_TX(3, "pass LLC frame to SMT");
1714 smt_received_pack(smc,smc->os.hwm.tx_mb,
1715 RD_FS_LOCAL) ;
1716 }
1717 }
1718 }
1719 NDD_TRACE("THfE",t,queue->tx_free,0) ;
1720}
1721
1722
1723/*
1724 * queues a receive for later send
1725 */
1726static void queue_llc_rx(struct s_smc *smc, SMbuf *mb)
1727{
1728 DB_GEN(4, "queue_llc_rx: mb = %p", mb);
1729 smc->os.hwm.queued_rx_frames++ ;
1730 mb->sm_next = (SMbuf *)NULL ;
1731 if (smc->os.hwm.llc_rx_pipe == NULL) {
1732 smc->os.hwm.llc_rx_pipe = mb ;
1733 }
1734 else {
1735 smc->os.hwm.llc_rx_tail->sm_next = mb ;
1736 }
1737 smc->os.hwm.llc_rx_tail = mb ;
1738
1739 /*
1740 * force an timer IRQ to receive the data
1741 */
1742 if (!smc->os.hwm.isr_flag) {
1743 smt_force_irq(smc) ;
1744 }
1745}
1746
1747/*
1748 * get a SMbuf from the llc_rx_queue
1749 */
1750static SMbuf *get_llc_rx(struct s_smc *smc)
1751{
1752 SMbuf *mb ;
1753
1754 if ((mb = smc->os.hwm.llc_rx_pipe)) {
1755 smc->os.hwm.queued_rx_frames-- ;
1756 smc->os.hwm.llc_rx_pipe = mb->sm_next ;
1757 }
1758 DB_GEN(4, "get_llc_rx: mb = 0x%p", mb);
1759 return mb;
1760}
1761
1762/*
1763 * queues a transmit SMT MBuf during the time were the MBuf is
1764 * queued the TxD ring
1765 */
1766static void queue_txd_mb(struct s_smc *smc, SMbuf *mb)
1767{
1768 DB_GEN(4, "_rx: queue_txd_mb = %p", mb);
1769 smc->os.hwm.queued_txd_mb++ ;
1770 mb->sm_next = (SMbuf *)NULL ;
1771 if (smc->os.hwm.txd_tx_pipe == NULL) {
1772 smc->os.hwm.txd_tx_pipe = mb ;
1773 }
1774 else {
1775 smc->os.hwm.txd_tx_tail->sm_next = mb ;
1776 }
1777 smc->os.hwm.txd_tx_tail = mb ;
1778}
1779
1780/*
1781 * get a SMbuf from the txd_tx_queue
1782 */
1783static SMbuf *get_txd_mb(struct s_smc *smc)
1784{
1785 SMbuf *mb ;
1786
1787 if ((mb = smc->os.hwm.txd_tx_pipe)) {
1788 smc->os.hwm.queued_txd_mb-- ;
1789 smc->os.hwm.txd_tx_pipe = mb->sm_next ;
1790 }
1791 DB_GEN(4, "get_txd_mb: mb = 0x%p", mb);
1792 return mb;
1793}
1794
1795/*
1796 * SMT Send function
1797 */
1798void smt_send_mbuf(struct s_smc *smc, SMbuf *mb, int fc)
1799{
1800 char far *data ;
1801 int len ;
1802 int n ;
1803 int i ;
1804 int frag_count ;
1805 int frame_status ;
1806 SK_LOC_DECL(char far,*virt[3]) ;
1807 int frag_len[3] ;
1808 struct s_smt_tx_queue *queue ;
1809 struct s_smt_fp_txd volatile *t ;
1810 u_long phys ;
1811 __le32 tbctrl;
1812
1813 NDD_TRACE("THSB",mb,fc,0) ;
1814 DB_TX(4, "smt_send_mbuf: mb = 0x%p, fc = 0x%x", mb, fc);
1815
1816 mb->sm_off-- ; /* set to fc */
1817 mb->sm_len++ ; /* + fc */
1818 data = smtod(mb,char *) ;
1819 *data = fc ;
1820 if (fc == FC_SMT_LOC)
1821 *data = FC_SMT_INFO ;
1822
1823 /*
1824 * determine the frag count and the virt addresses of the frags
1825 */
1826 frag_count = 0 ;
1827 len = mb->sm_len ;
1828 while (len) {
1829 n = SMT_PAGESIZE - ((long)data & (SMT_PAGESIZE-1)) ;
1830 if (n >= len) {
1831 n = len ;
1832 }
1833 DB_TX(5, "frag: virt/len = 0x%p/%d", data, n);
1834 virt[frag_count] = data ;
1835 frag_len[frag_count] = n ;
1836 frag_count++ ;
1837 len -= n ;
1838 data += n ;
1839 }
1840
1841 /*
1842 * determine the frame status
1843 */
1844 queue = smc->hw.fp.tx[QUEUE_A0] ;
1845 if (fc == FC_BEACON || fc == FC_SMT_LOC) {
1846 frame_status = LOC_TX ;
1847 }
1848 else {
1849 frame_status = LAN_TX ;
1850 if ((smc->os.hwm.pass_NSA &&(fc == FC_SMT_NSA)) ||
1851 (smc->os.hwm.pass_SMT &&(fc == FC_SMT_INFO)))
1852 frame_status |= LOC_TX ;
1853 }
1854
1855 if (!smc->hw.mac_ring_is_up || frag_count > queue->tx_free) {
1856 frame_status &= ~LAN_TX;
1857 if (frame_status) {
1858 DB_TX(2, "Ring is down: terminate LAN_TX");
1859 }
1860 else {
1861 DB_TX(2, "Ring is down: terminate transmission");
1862 smt_free_mbuf(smc,mb) ;
1863 return ;
1864 }
1865 }
1866 DB_TX(5, "frame_status = 0x%x", frame_status);
1867
1868 if ((frame_status & LAN_TX) && (frame_status & LOC_TX)) {
1869 mb->sm_use_count = 2 ;
1870 }
1871
1872 if (frame_status & LAN_TX) {
1873 t = queue->tx_curr_put ;
1874 frame_status |= FIRST_FRAG ;
1875 for (i = 0; i < frag_count; i++) {
1876 DB_TX(5, "init TxD = 0x%p", t);
1877 if (i == frag_count-1) {
1878 frame_status |= LAST_FRAG ;
1879 t->txd_txdscr = cpu_to_le32(TX_DESCRIPTOR |
1880 (((__u32)(mb->sm_len-1)&3) << 27)) ;
1881 }
1882 t->txd_virt = virt[i] ;
1883 phys = dma_master(smc, (void far *)virt[i],
1884 frag_len[i], DMA_RD|SMT_BUF) ;
1885 t->txd_tbadr = cpu_to_le32(phys) ;
1886 tbctrl = cpu_to_le32((((__u32)frame_status &
1887 (FIRST_FRAG|LAST_FRAG)) << 26) |
1888 BMU_OWN | BMU_CHECK | BMU_SMT_TX |frag_len[i]) ;
1889 t->txd_tbctrl = tbctrl ;
1890#ifndef AIX
1891 DRV_BUF_FLUSH(t,DDI_DMA_SYNC_FORDEV) ;
1892 outpd(queue->tx_bmu_ctl,CSR_START) ;
1893#else
1894 DRV_BUF_FLUSH(t,DDI_DMA_SYNC_FORDEV) ;
1895 outpd(ADDR(B0_XA_CSR),CSR_START) ;
1896#endif
1897 frame_status &= ~FIRST_FRAG ;
1898 queue->tx_curr_put = t = t->txd_next ;
1899 queue->tx_free-- ;
1900 queue->tx_used++ ;
1901 }
1902 smc->mib.m[MAC0].fddiMACTransmit_Ct++ ;
1903 queue_txd_mb(smc,mb) ;
1904 }
1905
1906 if (frame_status & LOC_TX) {
1907 DB_TX(5, "pass Mbuf to LLC queue");
1908 queue_llc_rx(smc,mb) ;
1909 }
1910
1911 /*
1912 * We need to unqueue the free SMT_MBUFs here, because it may
1913 * be that the SMT want's to send more than 1 frame for one down call
1914 */
1915 mac_drv_clear_txd(smc) ;
1916 NDD_TRACE("THSE",t,queue->tx_free,frag_count) ;
1917}
1918
1919/* BEGIN_MANUAL_ENTRY(mac_drv_clear_txd)
1920 * void mac_drv_clear_txd(smc)
1921 *
1922 * function DOWNCALL (hardware module, hwmtm.c)
1923 * mac_drv_clear_txd searches in both send queues for TxD's
1924 * which were finished by the adapter. It calls dma_complete
1925 * for each TxD. If the last fragment of an LLC frame is
1926 * reached, it calls mac_drv_tx_complete to release the
1927 * send buffer.
1928 *
1929 * return nothing
1930 *
1931 * END_MANUAL_ENTRY
1932 */
1933static void mac_drv_clear_txd(struct s_smc *smc)
1934{
1935 struct s_smt_tx_queue *queue ;
1936 struct s_smt_fp_txd volatile *t1 ;
1937 struct s_smt_fp_txd volatile *t2 = NULL ;
1938 SMbuf *mb ;
1939 u_long tbctrl ;
1940 int i ;
1941 int frag_count ;
1942 int n ;
1943
1944 NDD_TRACE("THcB",0,0,0) ;
1945 for (i = QUEUE_S; i <= QUEUE_A0; i++) {
1946 queue = smc->hw.fp.tx[i] ;
1947 t1 = queue->tx_curr_get ;
1948 DB_TX(5, "clear_txd: QUEUE = %d (0=sync/1=async)", i);
1949
1950 for ( ; ; ) {
1951 frag_count = 0 ;
1952
1953 do {
1954 DRV_BUF_FLUSH(t1,DDI_DMA_SYNC_FORCPU) ;
1955 DB_TX(5, "check OWN/EOF bit of TxD 0x%p", t1);
1956 tbctrl = le32_to_cpu(CR_READ(t1->txd_tbctrl));
1957
1958 if (tbctrl & BMU_OWN || !queue->tx_used){
1959 DB_TX(4, "End of TxDs queue %d", i);
1960 goto free_next_queue ; /* next queue */
1961 }
1962 t1 = t1->txd_next ;
1963 frag_count++ ;
1964 } while (!(tbctrl & BMU_EOF)) ;
1965
1966 t1 = queue->tx_curr_get ;
1967 for (n = frag_count; n; n--) {
1968 tbctrl = le32_to_cpu(t1->txd_tbctrl) ;
1969 dma_complete(smc,
1970 (union s_fp_descr volatile *) t1,
1971 (int) (DMA_RD |
1972 ((tbctrl & BMU_SMT_TX) >> 18))) ;
1973 t2 = t1 ;
1974 t1 = t1->txd_next ;
1975 }
1976
1977 if (tbctrl & BMU_SMT_TX) {
1978 mb = get_txd_mb(smc) ;
1979 smt_free_mbuf(smc,mb) ;
1980 }
1981 else {
1982#ifndef PASS_1ST_TXD_2_TX_COMP
1983 DB_TX(4, "mac_drv_tx_comp for TxD 0x%p", t2);
1984 mac_drv_tx_complete(smc,t2) ;
1985#else
1986 DB_TX(4, "mac_drv_tx_comp for TxD 0x%x",
1987 queue->tx_curr_get);
1988 mac_drv_tx_complete(smc,queue->tx_curr_get) ;
1989#endif
1990 }
1991 queue->tx_curr_get = t1 ;
1992 queue->tx_free += frag_count ;
1993 queue->tx_used -= frag_count ;
1994 }
1995free_next_queue: ;
1996 }
1997 NDD_TRACE("THcE",0,0,0) ;
1998}
1999
2000/*
2001 * BEGINN_MANUAL_ENTRY(mac_drv_clear_tx_queue)
2002 *
2003 * void mac_drv_clear_tx_queue(smc)
2004 * struct s_smc *smc ;
2005 *
2006 * function DOWNCALL (hardware module, hwmtm.c)
2007 * mac_drv_clear_tx_queue is called from the SMT when
2008 * the RMT state machine has entered the ISOLATE state.
2009 * This function is also called by the os-specific module
2010 * after it has called the function card_stop().
2011 * In this case, the frames in the send queues are obsolete and
2012 * should be removed.
2013 *
2014 * note calling sequence:
2015 * CLI_FBI(), card_stop(),
2016 * mac_drv_clear_tx_queue(), mac_drv_clear_rx_queue(),
2017 *
2018 * NOTE: The caller is responsible that the BMUs are idle
2019 * when this function is called.
2020 *
2021 * END_MANUAL_ENTRY
2022 */
2023void mac_drv_clear_tx_queue(struct s_smc *smc)
2024{
2025 struct s_smt_fp_txd volatile *t ;
2026 struct s_smt_tx_queue *queue ;
2027 int tx_used ;
2028 int i ;
2029
2030 if (smc->hw.hw_state != STOPPED) {
2031 SK_BREAK() ;
2032 SMT_PANIC(smc,HWM_E0011,HWM_E0011_MSG) ;
2033 return ;
2034 }
2035
2036 for (i = QUEUE_S; i <= QUEUE_A0; i++) {
2037 queue = smc->hw.fp.tx[i] ;
2038 DB_TX(5, "clear_tx_queue: QUEUE = %d (0=sync/1=async)", i);
2039
2040 /*
2041 * switch the OWN bit of all pending frames to the host
2042 */
2043 t = queue->tx_curr_get ;
2044 tx_used = queue->tx_used ;
2045 while (tx_used) {
2046 DRV_BUF_FLUSH(t,DDI_DMA_SYNC_FORCPU) ;
2047 DB_TX(5, "switch OWN bit of TxD 0x%p", t);
2048 t->txd_tbctrl &= ~cpu_to_le32(BMU_OWN) ;
2049 DRV_BUF_FLUSH(t,DDI_DMA_SYNC_FORDEV) ;
2050 t = t->txd_next ;
2051 tx_used-- ;
2052 }
2053 }
2054
2055 /*
2056 * release all TxD's for both send queues
2057 */
2058 mac_drv_clear_txd(smc) ;
2059
2060 for (i = QUEUE_S; i <= QUEUE_A0; i++) {
2061 queue = smc->hw.fp.tx[i] ;
2062 t = queue->tx_curr_get ;
2063
2064 /*
2065 * write the phys pointer of the NEXT descriptor into the
2066 * BMU's current address descriptor pointer and set
2067 * tx_curr_get and tx_curr_put to this position
2068 */
2069 if (i == QUEUE_S) {
2070 outpd(ADDR(B5_XS_DA),le32_to_cpu(t->txd_ntdadr)) ;
2071 }
2072 else {
2073 outpd(ADDR(B5_XA_DA),le32_to_cpu(t->txd_ntdadr)) ;
2074 }
2075
2076 queue->tx_curr_put = queue->tx_curr_get->txd_next ;
2077 queue->tx_curr_get = queue->tx_curr_put ;
2078 }
2079}
2080
2081
2082/*
2083 -------------------------------------------------------------
2084 TEST FUNCTIONS:
2085 -------------------------------------------------------------
2086*/
2087
2088#ifdef DEBUG
2089/*
2090 * BEGIN_MANUAL_ENTRY(mac_drv_debug_lev)
2091 * void mac_drv_debug_lev(smc,flag,lev)
2092 *
2093 * function DOWNCALL (drvsr.c)
2094 * To get a special debug info the user can assign a debug level
2095 * to any debug flag.
2096 *
2097 * para flag debug flag, possible values are:
2098 * = 0: reset all debug flags (the defined level is
2099 * ignored)
2100 * = 1: debug.d_smtf
2101 * = 2: debug.d_smt
2102 * = 3: debug.d_ecm
2103 * = 4: debug.d_rmt
2104 * = 5: debug.d_cfm
2105 * = 6: debug.d_pcm
2106 *
2107 * = 10: debug.d_os.hwm_rx (hardware module receive path)
2108 * = 11: debug.d_os.hwm_tx(hardware module transmit path)
2109 * = 12: debug.d_os.hwm_gen(hardware module general flag)
2110 *
2111 * lev debug level
2112 *
2113 * END_MANUAL_ENTRY
2114 */
2115void mac_drv_debug_lev(struct s_smc *smc, int flag, int lev)
2116{
2117 switch(flag) {
2118 case (int)NULL:
2119 DB_P.d_smtf = DB_P.d_smt = DB_P.d_ecm = DB_P.d_rmt = 0 ;
2120 DB_P.d_cfm = 0 ;
2121 DB_P.d_os.hwm_rx = DB_P.d_os.hwm_tx = DB_P.d_os.hwm_gen = 0 ;
2122#ifdef SBA
2123 DB_P.d_sba = 0 ;
2124#endif
2125#ifdef ESS
2126 DB_P.d_ess = 0 ;
2127#endif
2128 break ;
2129 case DEBUG_SMTF:
2130 DB_P.d_smtf = lev ;
2131 break ;
2132 case DEBUG_SMT:
2133 DB_P.d_smt = lev ;
2134 break ;
2135 case DEBUG_ECM:
2136 DB_P.d_ecm = lev ;
2137 break ;
2138 case DEBUG_RMT:
2139 DB_P.d_rmt = lev ;
2140 break ;
2141 case DEBUG_CFM:
2142 DB_P.d_cfm = lev ;
2143 break ;
2144 case DEBUG_PCM:
2145 DB_P.d_pcm = lev ;
2146 break ;
2147 case DEBUG_SBA:
2148#ifdef SBA
2149 DB_P.d_sba = lev ;
2150#endif
2151 break ;
2152 case DEBUG_ESS:
2153#ifdef ESS
2154 DB_P.d_ess = lev ;
2155#endif
2156 break ;
2157 case DB_HWM_RX:
2158 DB_P.d_os.hwm_rx = lev ;
2159 break ;
2160 case DB_HWM_TX:
2161 DB_P.d_os.hwm_tx = lev ;
2162 break ;
2163 case DB_HWM_GEN:
2164 DB_P.d_os.hwm_gen = lev ;
2165 break ;
2166 default:
2167 break ;
2168 }
2169}
2170#endif