5 ForeRunnerHE ATM Adapter driver for ATM on Linux
6 Copyright (C) 1999-2001 Naval Research Laboratory
8 This library is free software; you can redistribute it and/or
9 modify it under the terms of the GNU Lesser General Public
10 License as published by the Free Software Foundation; either
11 version 2.1 of the License, or (at your option) any later version.
13 This library is distributed in the hope that it will be useful,
14 but WITHOUT ANY WARRANTY; without even the implied warranty of
15 MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU
16 Lesser General Public License for more details.
18 You should have received a copy of the GNU Lesser General Public
19 License along with this library; if not, write to the Free Software
20 Foundation, Inc., 59 Temple Place, Suite 330, Boston, MA 02111-1307 USA
28 ForeRunnerHE ATM Adapter driver for ATM on Linux
29 Copyright (C) 1999-2001 Naval Research Laboratory
31 Permission to use, copy, modify and distribute this software and its
32 documentation is hereby granted, provided that both the copyright
33 notice and this permission notice appear in all copies of the software,
34 derivative works or modified versions, and any portions thereof, and
35 that both notices appear in supporting documentation.
37 NRL ALLOWS FREE USE OF THIS SOFTWARE IN ITS "AS IS" CONDITION AND
38 DISCLAIMS ANY LIABILITY OF ANY KIND FOR ANY DAMAGES WHATSOEVER
39 RESULTING FROM THE USE OF THIS SOFTWARE.
41 This driver was written using the "Programmer's Reference Manual for
42 ForeRunnerHE(tm)", MANU0361-01 - Rev. A, 08/21/98.
45 chas williams <chas@cmf.nrl.navy.mil>
46 eric kinzie <ekinzie@cmf.nrl.navy.mil>
49 4096 supported 'connections'
50 group 0 is used for all traffic
51 interrupt queue 0 is used for all interrupts
52 aal0 support (based on work from ulrich.u.muller@nokia.com)
56 #include <linux/module.h>
57 #include <linux/kernel.h>
58 #include <linux/skbuff.h>
59 #include <linux/pci.h>
60 #include <linux/errno.h>
61 #include <linux/types.h>
62 #include <linux/string.h>
63 #include <linux/delay.h>
64 #include <linux/init.h>
66 #include <linux/sched.h>
67 #include <linux/timer.h>
68 #include <linux/interrupt.h>
69 #include <linux/dma-mapping.h>
70 #include <linux/slab.h>
72 #include <asm/byteorder.h>
73 #include <asm/uaccess.h>
75 #include <linux/atmdev.h>
76 #include <linux/atm.h>
77 #include <linux/sonet.h>
79 #undef USE_SCATTERGATHER
80 #undef USE_CHECKSUM_HW /* still confused about this */
85 #include <linux/atm_he.h>
87 #define hprintk(fmt,args...) printk(KERN_ERR DEV_LABEL "%d: " fmt, he_dev->number , ##args)
90 #define HPRINTK(fmt,args...) printk(KERN_DEBUG DEV_LABEL "%d: " fmt, he_dev->number , ##args)
92 #define HPRINTK(fmt,args...) do { } while (0)
97 static int he_open(struct atm_vcc *vcc);
98 static void he_close(struct atm_vcc *vcc);
99 static int he_send(struct atm_vcc *vcc, struct sk_buff *skb);
100 static int he_ioctl(struct atm_dev *dev, unsigned int cmd, void __user *arg);
101 static irqreturn_t he_irq_handler(int irq, void *dev_id);
102 static void he_tasklet(unsigned long data);
103 static int he_proc_read(struct atm_dev *dev,loff_t *pos,char *page);
104 static int he_start(struct atm_dev *dev);
105 static void he_stop(struct he_dev *dev);
106 static void he_phy_put(struct atm_dev *, unsigned char, unsigned long);
107 static unsigned char he_phy_get(struct atm_dev *, unsigned long);
109 static u8 read_prom_byte(struct he_dev *he_dev, int addr);
113 static struct he_dev *he_devs;
114 static int disable64;
115 static short nvpibits = -1;
116 static short nvcibits = -1;
117 static short rx_skb_reserve = 16;
118 static int irq_coalesce = 1;
121 /* Read from EEPROM = 0000 0011b */
122 static unsigned int readtab[] = {
137 CLK_HIGH | SI_HIGH, /* 1 */
139 CLK_HIGH | SI_HIGH /* 1 */
142 /* Clock to read from/write to the EEPROM */
143 static unsigned int clocktab[] = {
163 static struct atmdev_ops he_ops =
169 .phy_put = he_phy_put,
170 .phy_get = he_phy_get,
171 .proc_read = he_proc_read,
175 #define he_writel(dev, val, reg) do { writel(val, (dev)->membase + (reg)); wmb(); } while (0)
176 #define he_readl(dev, reg) readl((dev)->membase + (reg))
178 /* section 2.12 connection memory access */
180 static __inline__ void
181 he_writel_internal(struct he_dev *he_dev, unsigned val, unsigned addr,
184 he_writel(he_dev, val, CON_DAT);
185 (void) he_readl(he_dev, CON_DAT); /* flush posted writes */
186 he_writel(he_dev, flags | CON_CTL_WRITE | CON_CTL_ADDR(addr), CON_CTL);
187 while (he_readl(he_dev, CON_CTL) & CON_CTL_BUSY);
190 #define he_writel_rcm(dev, val, reg) \
191 he_writel_internal(dev, val, reg, CON_CTL_RCM)
193 #define he_writel_tcm(dev, val, reg) \
194 he_writel_internal(dev, val, reg, CON_CTL_TCM)
196 #define he_writel_mbox(dev, val, reg) \
197 he_writel_internal(dev, val, reg, CON_CTL_MBOX)
200 he_readl_internal(struct he_dev *he_dev, unsigned addr, unsigned flags)
202 he_writel(he_dev, flags | CON_CTL_READ | CON_CTL_ADDR(addr), CON_CTL);
203 while (he_readl(he_dev, CON_CTL) & CON_CTL_BUSY);
204 return he_readl(he_dev, CON_DAT);
207 #define he_readl_rcm(dev, reg) \
208 he_readl_internal(dev, reg, CON_CTL_RCM)
210 #define he_readl_tcm(dev, reg) \
211 he_readl_internal(dev, reg, CON_CTL_TCM)
213 #define he_readl_mbox(dev, reg) \
214 he_readl_internal(dev, reg, CON_CTL_MBOX)
217 /* figure 2.2 connection id */
219 #define he_mkcid(dev, vpi, vci) (((vpi << (dev)->vcibits) | vci) & 0x1fff)
221 /* 2.5.1 per connection transmit state registers */
223 #define he_writel_tsr0(dev, val, cid) \
224 he_writel_tcm(dev, val, CONFIG_TSRA | (cid << 3) | 0)
225 #define he_readl_tsr0(dev, cid) \
226 he_readl_tcm(dev, CONFIG_TSRA | (cid << 3) | 0)
228 #define he_writel_tsr1(dev, val, cid) \
229 he_writel_tcm(dev, val, CONFIG_TSRA | (cid << 3) | 1)
231 #define he_writel_tsr2(dev, val, cid) \
232 he_writel_tcm(dev, val, CONFIG_TSRA | (cid << 3) | 2)
234 #define he_writel_tsr3(dev, val, cid) \
235 he_writel_tcm(dev, val, CONFIG_TSRA | (cid << 3) | 3)
237 #define he_writel_tsr4(dev, val, cid) \
238 he_writel_tcm(dev, val, CONFIG_TSRA | (cid << 3) | 4)
242 * NOTE While the transmit connection is active, bits 23 through 0
243 * of this register must not be written by the host. Byte
244 * enables should be used during normal operation when writing
245 * the most significant byte.
248 #define he_writel_tsr4_upper(dev, val, cid) \
249 he_writel_internal(dev, val, CONFIG_TSRA | (cid << 3) | 4, \
251 | CON_BYTE_DISABLE_2 \
252 | CON_BYTE_DISABLE_1 \
253 | CON_BYTE_DISABLE_0)
255 #define he_readl_tsr4(dev, cid) \
256 he_readl_tcm(dev, CONFIG_TSRA | (cid << 3) | 4)
258 #define he_writel_tsr5(dev, val, cid) \
259 he_writel_tcm(dev, val, CONFIG_TSRA | (cid << 3) | 5)
261 #define he_writel_tsr6(dev, val, cid) \
262 he_writel_tcm(dev, val, CONFIG_TSRA | (cid << 3) | 6)
264 #define he_writel_tsr7(dev, val, cid) \
265 he_writel_tcm(dev, val, CONFIG_TSRA | (cid << 3) | 7)
268 #define he_writel_tsr8(dev, val, cid) \
269 he_writel_tcm(dev, val, CONFIG_TSRB | (cid << 2) | 0)
271 #define he_writel_tsr9(dev, val, cid) \
272 he_writel_tcm(dev, val, CONFIG_TSRB | (cid << 2) | 1)
274 #define he_writel_tsr10(dev, val, cid) \
275 he_writel_tcm(dev, val, CONFIG_TSRB | (cid << 2) | 2)
277 #define he_writel_tsr11(dev, val, cid) \
278 he_writel_tcm(dev, val, CONFIG_TSRB | (cid << 2) | 3)
281 #define he_writel_tsr12(dev, val, cid) \
282 he_writel_tcm(dev, val, CONFIG_TSRC | (cid << 1) | 0)
284 #define he_writel_tsr13(dev, val, cid) \
285 he_writel_tcm(dev, val, CONFIG_TSRC | (cid << 1) | 1)
288 #define he_writel_tsr14(dev, val, cid) \
289 he_writel_tcm(dev, val, CONFIG_TSRD | cid)
291 #define he_writel_tsr14_upper(dev, val, cid) \
292 he_writel_internal(dev, val, CONFIG_TSRD | cid, \
294 | CON_BYTE_DISABLE_2 \
295 | CON_BYTE_DISABLE_1 \
296 | CON_BYTE_DISABLE_0)
298 /* 2.7.1 per connection receive state registers */
300 #define he_writel_rsr0(dev, val, cid) \
301 he_writel_rcm(dev, val, 0x00000 | (cid << 3) | 0)
302 #define he_readl_rsr0(dev, cid) \
303 he_readl_rcm(dev, 0x00000 | (cid << 3) | 0)
305 #define he_writel_rsr1(dev, val, cid) \
306 he_writel_rcm(dev, val, 0x00000 | (cid << 3) | 1)
308 #define he_writel_rsr2(dev, val, cid) \
309 he_writel_rcm(dev, val, 0x00000 | (cid << 3) | 2)
311 #define he_writel_rsr3(dev, val, cid) \
312 he_writel_rcm(dev, val, 0x00000 | (cid << 3) | 3)
314 #define he_writel_rsr4(dev, val, cid) \
315 he_writel_rcm(dev, val, 0x00000 | (cid << 3) | 4)
317 #define he_writel_rsr5(dev, val, cid) \
318 he_writel_rcm(dev, val, 0x00000 | (cid << 3) | 5)
320 #define he_writel_rsr6(dev, val, cid) \
321 he_writel_rcm(dev, val, 0x00000 | (cid << 3) | 6)
323 #define he_writel_rsr7(dev, val, cid) \
324 he_writel_rcm(dev, val, 0x00000 | (cid << 3) | 7)
326 static __inline__ struct atm_vcc*
327 __find_vcc(struct he_dev *he_dev, unsigned cid)
329 struct hlist_head *head;
331 struct hlist_node *node;
336 vpi = cid >> he_dev->vcibits;
337 vci = cid & ((1 << he_dev->vcibits) - 1);
338 head = &vcc_hash[vci & (VCC_HTABLE_SIZE -1)];
340 sk_for_each(s, node, head) {
342 if (vcc->dev == he_dev->atm_dev &&
343 vcc->vci == vci && vcc->vpi == vpi &&
344 vcc->qos.rxtp.traffic_class != ATM_NONE) {
352 he_init_one(struct pci_dev *pci_dev, const struct pci_device_id *pci_ent)
354 struct atm_dev *atm_dev = NULL;
355 struct he_dev *he_dev = NULL;
358 printk(KERN_INFO "ATM he driver\n");
360 if (pci_enable_device(pci_dev))
362 if (pci_set_dma_mask(pci_dev, DMA_BIT_MASK(32)) != 0) {
363 printk(KERN_WARNING "he: no suitable dma available\n");
365 goto init_one_failure;
368 atm_dev = atm_dev_register(DEV_LABEL, &he_ops, -1, NULL);
371 goto init_one_failure;
373 pci_set_drvdata(pci_dev, atm_dev);
375 he_dev = kzalloc(sizeof(struct he_dev),
379 goto init_one_failure;
381 he_dev->pci_dev = pci_dev;
382 he_dev->atm_dev = atm_dev;
383 he_dev->atm_dev->dev_data = he_dev;
384 atm_dev->dev_data = he_dev;
385 he_dev->number = atm_dev->number;
386 tasklet_init(&he_dev->tasklet, he_tasklet, (unsigned long) he_dev);
387 spin_lock_init(&he_dev->global_lock);
389 if (he_start(atm_dev)) {
392 goto init_one_failure;
396 he_dev->next = he_devs;
402 atm_dev_deregister(atm_dev);
404 pci_disable_device(pci_dev);
408 static void __devexit
409 he_remove_one (struct pci_dev *pci_dev)
411 struct atm_dev *atm_dev;
412 struct he_dev *he_dev;
414 atm_dev = pci_get_drvdata(pci_dev);
415 he_dev = HE_DEV(atm_dev);
417 /* need to remove from he_devs */
420 atm_dev_deregister(atm_dev);
423 pci_set_drvdata(pci_dev, NULL);
424 pci_disable_device(pci_dev);
429 rate_to_atmf(unsigned rate) /* cps to atm forum format */
431 #define NONZERO (1 << 14)
439 while (rate > 0x3ff) {
444 return (NONZERO | (exp << 9) | (rate & 0x1ff));
447 static void __devinit
448 he_init_rx_lbfp0(struct he_dev *he_dev)
450 unsigned i, lbm_offset, lbufd_index, lbuf_addr, lbuf_count;
451 unsigned lbufs_per_row = he_dev->cells_per_row / he_dev->cells_per_lbuf;
452 unsigned lbuf_bufsize = he_dev->cells_per_lbuf * ATM_CELL_PAYLOAD;
453 unsigned row_offset = he_dev->r0_startrow * he_dev->bytes_per_row;
456 lbm_offset = he_readl(he_dev, RCMLBM_BA);
458 he_writel(he_dev, lbufd_index, RLBF0_H);
460 for (i = 0, lbuf_count = 0; i < he_dev->r0_numbuffs; ++i) {
462 lbuf_addr = (row_offset + (lbuf_count * lbuf_bufsize)) / 32;
464 he_writel_rcm(he_dev, lbuf_addr, lbm_offset);
465 he_writel_rcm(he_dev, lbufd_index, lbm_offset + 1);
467 if (++lbuf_count == lbufs_per_row) {
469 row_offset += he_dev->bytes_per_row;
474 he_writel(he_dev, lbufd_index - 2, RLBF0_T);
475 he_writel(he_dev, he_dev->r0_numbuffs, RLBF0_C);
478 static void __devinit
479 he_init_rx_lbfp1(struct he_dev *he_dev)
481 unsigned i, lbm_offset, lbufd_index, lbuf_addr, lbuf_count;
482 unsigned lbufs_per_row = he_dev->cells_per_row / he_dev->cells_per_lbuf;
483 unsigned lbuf_bufsize = he_dev->cells_per_lbuf * ATM_CELL_PAYLOAD;
484 unsigned row_offset = he_dev->r1_startrow * he_dev->bytes_per_row;
487 lbm_offset = he_readl(he_dev, RCMLBM_BA) + (2 * lbufd_index);
489 he_writel(he_dev, lbufd_index, RLBF1_H);
491 for (i = 0, lbuf_count = 0; i < he_dev->r1_numbuffs; ++i) {
493 lbuf_addr = (row_offset + (lbuf_count * lbuf_bufsize)) / 32;
495 he_writel_rcm(he_dev, lbuf_addr, lbm_offset);
496 he_writel_rcm(he_dev, lbufd_index, lbm_offset + 1);
498 if (++lbuf_count == lbufs_per_row) {
500 row_offset += he_dev->bytes_per_row;
505 he_writel(he_dev, lbufd_index - 2, RLBF1_T);
506 he_writel(he_dev, he_dev->r1_numbuffs, RLBF1_C);
509 static void __devinit
510 he_init_tx_lbfp(struct he_dev *he_dev)
512 unsigned i, lbm_offset, lbufd_index, lbuf_addr, lbuf_count;
513 unsigned lbufs_per_row = he_dev->cells_per_row / he_dev->cells_per_lbuf;
514 unsigned lbuf_bufsize = he_dev->cells_per_lbuf * ATM_CELL_PAYLOAD;
515 unsigned row_offset = he_dev->tx_startrow * he_dev->bytes_per_row;
517 lbufd_index = he_dev->r0_numbuffs + he_dev->r1_numbuffs;
518 lbm_offset = he_readl(he_dev, RCMLBM_BA) + (2 * lbufd_index);
520 he_writel(he_dev, lbufd_index, TLBF_H);
522 for (i = 0, lbuf_count = 0; i < he_dev->tx_numbuffs; ++i) {
524 lbuf_addr = (row_offset + (lbuf_count * lbuf_bufsize)) / 32;
526 he_writel_rcm(he_dev, lbuf_addr, lbm_offset);
527 he_writel_rcm(he_dev, lbufd_index, lbm_offset + 1);
529 if (++lbuf_count == lbufs_per_row) {
531 row_offset += he_dev->bytes_per_row;
536 he_writel(he_dev, lbufd_index - 1, TLBF_T);
540 he_init_tpdrq(struct he_dev *he_dev)
542 he_dev->tpdrq_base = pci_alloc_consistent(he_dev->pci_dev,
543 CONFIG_TPDRQ_SIZE * sizeof(struct he_tpdrq), &he_dev->tpdrq_phys);
544 if (he_dev->tpdrq_base == NULL) {
545 hprintk("failed to alloc tpdrq\n");
548 memset(he_dev->tpdrq_base, 0,
549 CONFIG_TPDRQ_SIZE * sizeof(struct he_tpdrq));
551 he_dev->tpdrq_tail = he_dev->tpdrq_base;
552 he_dev->tpdrq_head = he_dev->tpdrq_base;
554 he_writel(he_dev, he_dev->tpdrq_phys, TPDRQ_B_H);
555 he_writel(he_dev, 0, TPDRQ_T);
556 he_writel(he_dev, CONFIG_TPDRQ_SIZE - 1, TPDRQ_S);
561 static void __devinit
562 he_init_cs_block(struct he_dev *he_dev)
564 unsigned clock, rate, delta;
567 /* 5.1.7 cs block initialization */
569 for (reg = 0; reg < 0x20; ++reg)
570 he_writel_mbox(he_dev, 0x0, CS_STTIM0 + reg);
572 /* rate grid timer reload values */
574 clock = he_is622(he_dev) ? 66667000 : 50000000;
575 rate = he_dev->atm_dev->link_rate;
576 delta = rate / 16 / 2;
578 for (reg = 0; reg < 0x10; ++reg) {
579 /* 2.4 internal transmit function
581 * we initialize the first row in the rate grid.
582 * values are period (in clock cycles) of timer
584 unsigned period = clock / rate;
586 he_writel_mbox(he_dev, period, CS_TGRLD0 + reg);
590 if (he_is622(he_dev)) {
591 /* table 5.2 (4 cells per lbuf) */
592 he_writel_mbox(he_dev, 0x000800fa, CS_ERTHR0);
593 he_writel_mbox(he_dev, 0x000c33cb, CS_ERTHR1);
594 he_writel_mbox(he_dev, 0x0010101b, CS_ERTHR2);
595 he_writel_mbox(he_dev, 0x00181dac, CS_ERTHR3);
596 he_writel_mbox(he_dev, 0x00280600, CS_ERTHR4);
598 /* table 5.3, 5.4, 5.5, 5.6, 5.7 */
599 he_writel_mbox(he_dev, 0x023de8b3, CS_ERCTL0);
600 he_writel_mbox(he_dev, 0x1801, CS_ERCTL1);
601 he_writel_mbox(he_dev, 0x68b3, CS_ERCTL2);
602 he_writel_mbox(he_dev, 0x1280, CS_ERSTAT0);
603 he_writel_mbox(he_dev, 0x68b3, CS_ERSTAT1);
604 he_writel_mbox(he_dev, 0x14585, CS_RTFWR);
606 he_writel_mbox(he_dev, 0x4680, CS_RTATR);
609 he_writel_mbox(he_dev, 0x00159ece, CS_TFBSET);
610 he_writel_mbox(he_dev, 0x68b3, CS_WCRMAX);
611 he_writel_mbox(he_dev, 0x5eb3, CS_WCRMIN);
612 he_writel_mbox(he_dev, 0xe8b3, CS_WCRINC);
613 he_writel_mbox(he_dev, 0xdeb3, CS_WCRDEC);
614 he_writel_mbox(he_dev, 0x68b3, CS_WCRCEIL);
617 he_writel_mbox(he_dev, 0x5, CS_OTPPER);
618 he_writel_mbox(he_dev, 0x14, CS_OTWPER);
620 /* table 5.1 (4 cells per lbuf) */
621 he_writel_mbox(he_dev, 0x000400ea, CS_ERTHR0);
622 he_writel_mbox(he_dev, 0x00063388, CS_ERTHR1);
623 he_writel_mbox(he_dev, 0x00081018, CS_ERTHR2);
624 he_writel_mbox(he_dev, 0x000c1dac, CS_ERTHR3);
625 he_writel_mbox(he_dev, 0x0014051a, CS_ERTHR4);
627 /* table 5.3, 5.4, 5.5, 5.6, 5.7 */
628 he_writel_mbox(he_dev, 0x0235e4b1, CS_ERCTL0);
629 he_writel_mbox(he_dev, 0x4701, CS_ERCTL1);
630 he_writel_mbox(he_dev, 0x64b1, CS_ERCTL2);
631 he_writel_mbox(he_dev, 0x1280, CS_ERSTAT0);
632 he_writel_mbox(he_dev, 0x64b1, CS_ERSTAT1);
633 he_writel_mbox(he_dev, 0xf424, CS_RTFWR);
635 he_writel_mbox(he_dev, 0x4680, CS_RTATR);
638 he_writel_mbox(he_dev, 0x000563b7, CS_TFBSET);
639 he_writel_mbox(he_dev, 0x64b1, CS_WCRMAX);
640 he_writel_mbox(he_dev, 0x5ab1, CS_WCRMIN);
641 he_writel_mbox(he_dev, 0xe4b1, CS_WCRINC);
642 he_writel_mbox(he_dev, 0xdab1, CS_WCRDEC);
643 he_writel_mbox(he_dev, 0x64b1, CS_WCRCEIL);
646 he_writel_mbox(he_dev, 0x6, CS_OTPPER);
647 he_writel_mbox(he_dev, 0x1e, CS_OTWPER);
650 he_writel_mbox(he_dev, 0x8, CS_OTTLIM);
652 for (reg = 0; reg < 0x8; ++reg)
653 he_writel_mbox(he_dev, 0x0, CS_HGRRT0 + reg);
658 he_init_cs_block_rcm(struct he_dev *he_dev)
660 unsigned (*rategrid)[16][16];
661 unsigned rate, delta;
664 unsigned rate_atmf, exp, man;
665 unsigned long long rate_cps;
666 int mult, buf, buf_limit = 4;
668 rategrid = kmalloc( sizeof(unsigned) * 16 * 16, GFP_KERNEL);
672 /* initialize rate grid group table */
674 for (reg = 0x0; reg < 0xff; ++reg)
675 he_writel_rcm(he_dev, 0x0, CONFIG_RCMABR + reg);
677 /* initialize rate controller groups */
679 for (reg = 0x100; reg < 0x1ff; ++reg)
680 he_writel_rcm(he_dev, 0x0, CONFIG_RCMABR + reg);
682 /* initialize tNrm lookup table */
684 /* the manual makes reference to a routine in a sample driver
685 for proper configuration; fortunately, we only need this
686 in order to support abr connection */
688 /* initialize rate to group table */
690 rate = he_dev->atm_dev->link_rate;
694 * 2.4 transmit internal functions
696 * we construct a copy of the rate grid used by the scheduler
697 * in order to construct the rate to group table below
700 for (j = 0; j < 16; j++) {
701 (*rategrid)[0][j] = rate;
705 for (i = 1; i < 16; i++)
706 for (j = 0; j < 16; j++)
708 (*rategrid)[i][j] = (*rategrid)[i - 1][j] / 4;
710 (*rategrid)[i][j] = (*rategrid)[i - 1][j] / 2;
713 * 2.4 transmit internal function
715 * this table maps the upper 5 bits of exponent and mantissa
716 * of the atm forum representation of the rate into an index
721 while (rate_atmf < 0x400) {
722 man = (rate_atmf & 0x1f) << 4;
723 exp = rate_atmf >> 5;
726 instead of '/ 512', use '>> 9' to prevent a call
727 to divdu3 on x86 platforms
729 rate_cps = (unsigned long long) (1 << exp) * (man + 512) >> 9;
732 rate_cps = 10; /* 2.2.1 minimum payload rate is 10 cps */
734 for (i = 255; i > 0; i--)
735 if ((*rategrid)[i/16][i%16] >= rate_cps)
736 break; /* pick nearest rate instead? */
739 * each table entry is 16 bits: (rate grid index (8 bits)
740 * and a buffer limit (8 bits)
741 * there are two table entries in each 32-bit register
745 buf = rate_cps * he_dev->tx_numbuffs /
746 (he_dev->atm_dev->link_rate * 2);
748 /* this is pretty, but avoids _divdu3 and is mostly correct */
749 mult = he_dev->atm_dev->link_rate / ATM_OC3_PCR;
750 if (rate_cps > (272 * mult))
752 else if (rate_cps > (204 * mult))
754 else if (rate_cps > (136 * mult))
756 else if (rate_cps > (68 * mult))
763 reg = (reg << 16) | ((i << 8) | buf);
765 #define RTGTBL_OFFSET 0x400
768 he_writel_rcm(he_dev, reg,
769 CONFIG_RCMABR + RTGTBL_OFFSET + (rate_atmf >> 1));
779 he_init_group(struct he_dev *he_dev, int group)
783 he_writel(he_dev, 0x0, G0_RBPS_S + (group * 32));
784 he_writel(he_dev, 0x0, G0_RBPS_T + (group * 32));
785 he_writel(he_dev, 0x0, G0_RBPS_QI + (group * 32));
786 he_writel(he_dev, RBP_THRESH(0x1) | RBP_QSIZE(0x0),
787 G0_RBPS_BS + (group * 32));
789 /* large buffer pool */
790 he_dev->rbpl_pool = pci_pool_create("rbpl", he_dev->pci_dev,
791 CONFIG_RBPL_BUFSIZE, 8, 0);
792 if (he_dev->rbpl_pool == NULL) {
793 hprintk("unable to create rbpl pool\n");
797 he_dev->rbpl_base = pci_alloc_consistent(he_dev->pci_dev,
798 CONFIG_RBPL_SIZE * sizeof(struct he_rbp), &he_dev->rbpl_phys);
799 if (he_dev->rbpl_base == NULL) {
800 hprintk("failed to alloc rbpl_base\n");
801 goto out_destroy_rbpl_pool;
803 memset(he_dev->rbpl_base, 0, CONFIG_RBPL_SIZE * sizeof(struct he_rbp));
804 he_dev->rbpl_virt = kmalloc(CONFIG_RBPL_SIZE * sizeof(struct he_virt), GFP_KERNEL);
805 if (he_dev->rbpl_virt == NULL) {
806 hprintk("failed to alloc rbpl_virt\n");
807 goto out_free_rbpl_base;
810 for (i = 0; i < CONFIG_RBPL_SIZE; ++i) {
811 dma_addr_t dma_handle;
814 cpuaddr = pci_pool_alloc(he_dev->rbpl_pool, GFP_KERNEL|GFP_DMA, &dma_handle);
816 goto out_free_rbpl_virt;
818 he_dev->rbpl_virt[i].virt = cpuaddr;
819 he_dev->rbpl_base[i].status = RBP_LOANED | (i << RBP_INDEX_OFF);
820 he_dev->rbpl_base[i].phys = dma_handle;
822 he_dev->rbpl_tail = &he_dev->rbpl_base[CONFIG_RBPL_SIZE - 1];
824 he_writel(he_dev, he_dev->rbpl_phys, G0_RBPL_S + (group * 32));
825 he_writel(he_dev, RBPL_MASK(he_dev->rbpl_tail),
826 G0_RBPL_T + (group * 32));
827 he_writel(he_dev, CONFIG_RBPL_BUFSIZE/4,
828 G0_RBPL_BS + (group * 32));
830 RBP_THRESH(CONFIG_RBPL_THRESH) |
831 RBP_QSIZE(CONFIG_RBPL_SIZE - 1) |
833 G0_RBPL_QI + (group * 32));
835 /* rx buffer ready queue */
837 he_dev->rbrq_base = pci_alloc_consistent(he_dev->pci_dev,
838 CONFIG_RBRQ_SIZE * sizeof(struct he_rbrq), &he_dev->rbrq_phys);
839 if (he_dev->rbrq_base == NULL) {
840 hprintk("failed to allocate rbrq\n");
841 goto out_free_rbpl_virt;
843 memset(he_dev->rbrq_base, 0, CONFIG_RBRQ_SIZE * sizeof(struct he_rbrq));
845 he_dev->rbrq_head = he_dev->rbrq_base;
846 he_writel(he_dev, he_dev->rbrq_phys, G0_RBRQ_ST + (group * 16));
847 he_writel(he_dev, 0, G0_RBRQ_H + (group * 16));
849 RBRQ_THRESH(CONFIG_RBRQ_THRESH) | RBRQ_SIZE(CONFIG_RBRQ_SIZE - 1),
850 G0_RBRQ_Q + (group * 16));
852 hprintk("coalescing interrupts\n");
853 he_writel(he_dev, RBRQ_TIME(768) | RBRQ_COUNT(7),
854 G0_RBRQ_I + (group * 16));
856 he_writel(he_dev, RBRQ_TIME(0) | RBRQ_COUNT(1),
857 G0_RBRQ_I + (group * 16));
859 /* tx buffer ready queue */
861 he_dev->tbrq_base = pci_alloc_consistent(he_dev->pci_dev,
862 CONFIG_TBRQ_SIZE * sizeof(struct he_tbrq), &he_dev->tbrq_phys);
863 if (he_dev->tbrq_base == NULL) {
864 hprintk("failed to allocate tbrq\n");
865 goto out_free_rbpq_base;
867 memset(he_dev->tbrq_base, 0, CONFIG_TBRQ_SIZE * sizeof(struct he_tbrq));
869 he_dev->tbrq_head = he_dev->tbrq_base;
871 he_writel(he_dev, he_dev->tbrq_phys, G0_TBRQ_B_T + (group * 16));
872 he_writel(he_dev, 0, G0_TBRQ_H + (group * 16));
873 he_writel(he_dev, CONFIG_TBRQ_SIZE - 1, G0_TBRQ_S + (group * 16));
874 he_writel(he_dev, CONFIG_TBRQ_THRESH, G0_TBRQ_THRESH + (group * 16));
879 pci_free_consistent(he_dev->pci_dev, CONFIG_RBRQ_SIZE *
880 sizeof(struct he_rbrq), he_dev->rbrq_base,
882 i = CONFIG_RBPL_SIZE;
885 pci_pool_free(he_dev->rbpl_pool, he_dev->rbpl_virt[i].virt,
886 he_dev->rbpl_base[i].phys);
887 kfree(he_dev->rbpl_virt);
890 pci_free_consistent(he_dev->pci_dev, CONFIG_RBPL_SIZE *
891 sizeof(struct he_rbp), he_dev->rbpl_base,
893 out_destroy_rbpl_pool:
894 pci_pool_destroy(he_dev->rbpl_pool);
900 he_init_irq(struct he_dev *he_dev)
904 /* 2.9.3.5 tail offset for each interrupt queue is located after the
905 end of the interrupt queue */
907 he_dev->irq_base = pci_alloc_consistent(he_dev->pci_dev,
908 (CONFIG_IRQ_SIZE+1) * sizeof(struct he_irq), &he_dev->irq_phys);
909 if (he_dev->irq_base == NULL) {
910 hprintk("failed to allocate irq\n");
913 he_dev->irq_tailoffset = (unsigned *)
914 &he_dev->irq_base[CONFIG_IRQ_SIZE];
915 *he_dev->irq_tailoffset = 0;
916 he_dev->irq_head = he_dev->irq_base;
917 he_dev->irq_tail = he_dev->irq_base;
919 for (i = 0; i < CONFIG_IRQ_SIZE; ++i)
920 he_dev->irq_base[i].isw = ITYPE_INVALID;
922 he_writel(he_dev, he_dev->irq_phys, IRQ0_BASE);
924 IRQ_SIZE(CONFIG_IRQ_SIZE) | IRQ_THRESH(CONFIG_IRQ_THRESH),
926 he_writel(he_dev, IRQ_INT_A | IRQ_TYPE_LINE, IRQ0_CNTL);
927 he_writel(he_dev, 0x0, IRQ0_DATA);
929 he_writel(he_dev, 0x0, IRQ1_BASE);
930 he_writel(he_dev, 0x0, IRQ1_HEAD);
931 he_writel(he_dev, 0x0, IRQ1_CNTL);
932 he_writel(he_dev, 0x0, IRQ1_DATA);
934 he_writel(he_dev, 0x0, IRQ2_BASE);
935 he_writel(he_dev, 0x0, IRQ2_HEAD);
936 he_writel(he_dev, 0x0, IRQ2_CNTL);
937 he_writel(he_dev, 0x0, IRQ2_DATA);
939 he_writel(he_dev, 0x0, IRQ3_BASE);
940 he_writel(he_dev, 0x0, IRQ3_HEAD);
941 he_writel(he_dev, 0x0, IRQ3_CNTL);
942 he_writel(he_dev, 0x0, IRQ3_DATA);
944 /* 2.9.3.2 interrupt queue mapping registers */
946 he_writel(he_dev, 0x0, GRP_10_MAP);
947 he_writel(he_dev, 0x0, GRP_32_MAP);
948 he_writel(he_dev, 0x0, GRP_54_MAP);
949 he_writel(he_dev, 0x0, GRP_76_MAP);
951 if (request_irq(he_dev->pci_dev->irq, he_irq_handler, IRQF_DISABLED|IRQF_SHARED, DEV_LABEL, he_dev)) {
952 hprintk("irq %d already in use\n", he_dev->pci_dev->irq);
956 he_dev->irq = he_dev->pci_dev->irq;
962 he_start(struct atm_dev *dev)
964 struct he_dev *he_dev;
965 struct pci_dev *pci_dev;
966 unsigned long membase;
969 u32 gen_cntl_0, host_cntl, lb_swap;
970 u8 cache_size, timer;
973 unsigned int status, reg;
976 he_dev = HE_DEV(dev);
977 pci_dev = he_dev->pci_dev;
979 membase = pci_resource_start(pci_dev, 0);
980 HPRINTK("membase = 0x%lx irq = %d.\n", membase, pci_dev->irq);
983 * pci bus controller initialization
986 /* 4.3 pci bus controller-specific initialization */
987 if (pci_read_config_dword(pci_dev, GEN_CNTL_0, &gen_cntl_0) != 0) {
988 hprintk("can't read GEN_CNTL_0\n");
991 gen_cntl_0 |= (MRL_ENB | MRM_ENB | IGNORE_TIMEOUT);
992 if (pci_write_config_dword(pci_dev, GEN_CNTL_0, gen_cntl_0) != 0) {
993 hprintk("can't write GEN_CNTL_0.\n");
997 if (pci_read_config_word(pci_dev, PCI_COMMAND, &command) != 0) {
998 hprintk("can't read PCI_COMMAND.\n");
1002 command |= (PCI_COMMAND_MEMORY | PCI_COMMAND_MASTER | PCI_COMMAND_INVALIDATE);
1003 if (pci_write_config_word(pci_dev, PCI_COMMAND, command) != 0) {
1004 hprintk("can't enable memory.\n");
1008 if (pci_read_config_byte(pci_dev, PCI_CACHE_LINE_SIZE, &cache_size)) {
1009 hprintk("can't read cache line size?\n");
1013 if (cache_size < 16) {
1015 if (pci_write_config_byte(pci_dev, PCI_CACHE_LINE_SIZE, cache_size))
1016 hprintk("can't set cache line size to %d\n", cache_size);
1019 if (pci_read_config_byte(pci_dev, PCI_LATENCY_TIMER, &timer)) {
1020 hprintk("can't read latency timer?\n");
1026 * LAT_TIMER = 1 + AVG_LAT + BURST_SIZE/BUS_SIZE
1028 * AVG_LAT: The average first data read/write latency [maximum 16 clock cycles]
1029 * BURST_SIZE: 1536 bytes (read) for 622, 768 bytes (read) for 155 [192 clock cycles]
1032 #define LAT_TIMER 209
1033 if (timer < LAT_TIMER) {
1034 HPRINTK("latency timer was %d, setting to %d\n", timer, LAT_TIMER);
1036 if (pci_write_config_byte(pci_dev, PCI_LATENCY_TIMER, timer))
1037 hprintk("can't set latency timer to %d\n", timer);
1040 if (!(he_dev->membase = ioremap(membase, HE_REGMAP_SIZE))) {
1041 hprintk("can't set up page mapping\n");
1045 /* 4.4 card reset */
1046 he_writel(he_dev, 0x0, RESET_CNTL);
1047 he_writel(he_dev, 0xff, RESET_CNTL);
1049 udelay(16*1000); /* 16 ms */
1050 status = he_readl(he_dev, RESET_CNTL);
1051 if ((status & BOARD_RST_STATUS) == 0) {
1052 hprintk("reset failed\n");
1056 /* 4.5 set bus width */
1057 host_cntl = he_readl(he_dev, HOST_CNTL);
1058 if (host_cntl & PCI_BUS_SIZE64)
1059 gen_cntl_0 |= ENBL_64;
1061 gen_cntl_0 &= ~ENBL_64;
1063 if (disable64 == 1) {
1064 hprintk("disabling 64-bit pci bus transfers\n");
1065 gen_cntl_0 &= ~ENBL_64;
1068 if (gen_cntl_0 & ENBL_64)
1069 hprintk("64-bit transfers enabled\n");
1071 pci_write_config_dword(pci_dev, GEN_CNTL_0, gen_cntl_0);
1073 /* 4.7 read prom contents */
1074 for (i = 0; i < PROD_ID_LEN; ++i)
1075 he_dev->prod_id[i] = read_prom_byte(he_dev, PROD_ID + i);
1077 he_dev->media = read_prom_byte(he_dev, MEDIA);
1079 for (i = 0; i < 6; ++i)
1080 dev->esi[i] = read_prom_byte(he_dev, MAC_ADDR + i);
1082 hprintk("%s%s, %x:%x:%x:%x:%x:%x\n",
1084 he_dev->media & 0x40 ? "SM" : "MM",
1091 he_dev->atm_dev->link_rate = he_is622(he_dev) ?
1092 ATM_OC12_PCR : ATM_OC3_PCR;
1094 /* 4.6 set host endianess */
1095 lb_swap = he_readl(he_dev, LB_SWAP);
1096 if (he_is622(he_dev))
1097 lb_swap &= ~XFER_SIZE; /* 4 cells */
1099 lb_swap |= XFER_SIZE; /* 8 cells */
1101 lb_swap |= DESC_WR_SWAP | INTR_SWAP | BIG_ENDIAN_HOST;
1103 lb_swap &= ~(DESC_WR_SWAP | INTR_SWAP | BIG_ENDIAN_HOST |
1104 DATA_WR_SWAP | DATA_RD_SWAP | DESC_RD_SWAP);
1105 #endif /* __BIG_ENDIAN */
1106 he_writel(he_dev, lb_swap, LB_SWAP);
1108 /* 4.8 sdram controller initialization */
1109 he_writel(he_dev, he_is622(he_dev) ? LB_64_ENB : 0x0, SDRAM_CTL);
1111 /* 4.9 initialize rnum value */
1112 lb_swap |= SWAP_RNUM_MAX(0xf);
1113 he_writel(he_dev, lb_swap, LB_SWAP);
1115 /* 4.10 initialize the interrupt queues */
1116 if ((err = he_init_irq(he_dev)) != 0)
1119 /* 4.11 enable pci bus controller state machines */
1120 host_cntl |= (OUTFF_ENB | CMDFF_ENB |
1121 QUICK_RD_RETRY | QUICK_WR_RETRY | PERR_INT_ENB);
1122 he_writel(he_dev, host_cntl, HOST_CNTL);
1124 gen_cntl_0 |= INT_PROC_ENBL|INIT_ENB;
1125 pci_write_config_dword(pci_dev, GEN_CNTL_0, gen_cntl_0);
1128 * atm network controller initialization
1131 /* 5.1.1 generic configuration state */
1134 * local (cell) buffer memory map
1138 * 0 ____________1023 bytes 0 _______________________2047 bytes
1140 * | utility | | rx0 | |
1141 * 5|____________| 255|___________________| u |
1144 * | rx0 | row | tx | l |
1146 * | | 767|___________________| t |
1147 * 517|____________| 768| | y |
1148 * row 518| | | rx1 | |
1149 * | | 1023|___________________|___|
1154 * 1535|____________|
1157 * 2047|____________|
1161 /* total 4096 connections */
1162 he_dev->vcibits = CONFIG_DEFAULT_VCIBITS;
1163 he_dev->vpibits = CONFIG_DEFAULT_VPIBITS;
1165 if (nvpibits != -1 && nvcibits != -1 && nvpibits+nvcibits != HE_MAXCIDBITS) {
1166 hprintk("nvpibits + nvcibits != %d\n", HE_MAXCIDBITS);
1170 if (nvpibits != -1) {
1171 he_dev->vpibits = nvpibits;
1172 he_dev->vcibits = HE_MAXCIDBITS - nvpibits;
1175 if (nvcibits != -1) {
1176 he_dev->vcibits = nvcibits;
1177 he_dev->vpibits = HE_MAXCIDBITS - nvcibits;
1181 if (he_is622(he_dev)) {
1182 he_dev->cells_per_row = 40;
1183 he_dev->bytes_per_row = 2048;
1184 he_dev->r0_numrows = 256;
1185 he_dev->tx_numrows = 512;
1186 he_dev->r1_numrows = 256;
1187 he_dev->r0_startrow = 0;
1188 he_dev->tx_startrow = 256;
1189 he_dev->r1_startrow = 768;
1191 he_dev->cells_per_row = 20;
1192 he_dev->bytes_per_row = 1024;
1193 he_dev->r0_numrows = 512;
1194 he_dev->tx_numrows = 1018;
1195 he_dev->r1_numrows = 512;
1196 he_dev->r0_startrow = 6;
1197 he_dev->tx_startrow = 518;
1198 he_dev->r1_startrow = 1536;
1201 he_dev->cells_per_lbuf = 4;
1202 he_dev->buffer_limit = 4;
1203 he_dev->r0_numbuffs = he_dev->r0_numrows *
1204 he_dev->cells_per_row / he_dev->cells_per_lbuf;
1205 if (he_dev->r0_numbuffs > 2560)
1206 he_dev->r0_numbuffs = 2560;
1208 he_dev->r1_numbuffs = he_dev->r1_numrows *
1209 he_dev->cells_per_row / he_dev->cells_per_lbuf;
1210 if (he_dev->r1_numbuffs > 2560)
1211 he_dev->r1_numbuffs = 2560;
1213 he_dev->tx_numbuffs = he_dev->tx_numrows *
1214 he_dev->cells_per_row / he_dev->cells_per_lbuf;
1215 if (he_dev->tx_numbuffs > 5120)
1216 he_dev->tx_numbuffs = 5120;
1218 /* 5.1.2 configure hardware dependent registers */
1221 SLICE_X(0x2) | ARB_RNUM_MAX(0xf) | TH_PRTY(0x3) |
1222 RH_PRTY(0x3) | TL_PRTY(0x2) | RL_PRTY(0x1) |
1223 (he_is622(he_dev) ? BUS_MULTI(0x28) : BUS_MULTI(0x46)) |
1224 (he_is622(he_dev) ? NET_PREF(0x50) : NET_PREF(0x8c)),
1227 he_writel(he_dev, BANK_ON |
1228 (he_is622(he_dev) ? (REF_RATE(0x384) | WIDE_DATA) : REF_RATE(0x150)),
1232 (he_is622(he_dev) ? RM_BANK_WAIT(1) : RM_BANK_WAIT(0)) |
1233 RM_RW_WAIT(1), RCMCONFIG);
1235 (he_is622(he_dev) ? TM_BANK_WAIT(2) : TM_BANK_WAIT(1)) |
1236 TM_RW_WAIT(1), TCMCONFIG);
1238 he_writel(he_dev, he_dev->cells_per_lbuf * ATM_CELL_PAYLOAD, LB_CONFIG);
1241 (he_is622(he_dev) ? UT_RD_DELAY(8) : UT_RD_DELAY(0)) |
1242 (he_is622(he_dev) ? RC_UT_MODE(0) : RC_UT_MODE(1)) |
1243 RX_VALVP(he_dev->vpibits) |
1244 RX_VALVC(he_dev->vcibits), RC_CONFIG);
1246 he_writel(he_dev, DRF_THRESH(0x20) |
1247 (he_is622(he_dev) ? TX_UT_MODE(0) : TX_UT_MODE(1)) |
1248 TX_VCI_MASK(he_dev->vcibits) |
1249 LBFREE_CNT(he_dev->tx_numbuffs), TX_CONFIG);
1251 he_writel(he_dev, 0x0, TXAAL5_PROTO);
1253 he_writel(he_dev, PHY_INT_ENB |
1254 (he_is622(he_dev) ? PTMR_PRE(67 - 1) : PTMR_PRE(50 - 1)),
1257 /* 5.1.3 initialize connection memory */
1259 for (i = 0; i < TCM_MEM_SIZE; ++i)
1260 he_writel_tcm(he_dev, 0, i);
1262 for (i = 0; i < RCM_MEM_SIZE; ++i)
1263 he_writel_rcm(he_dev, 0, i);
1266 * transmit connection memory map
1269 * 0x0 ___________________
1275 * 0x8000|___________________|
1278 * 0xc000|___________________|
1281 * 0xe000|___________________|
1283 * 0xf000|___________________|
1285 * 0x10000|___________________|
1288 * |___________________|
1291 * 0x1ffff|___________________|
1296 he_writel(he_dev, CONFIG_TSRB, TSRB_BA);
1297 he_writel(he_dev, CONFIG_TSRC, TSRC_BA);
1298 he_writel(he_dev, CONFIG_TSRD, TSRD_BA);
1299 he_writel(he_dev, CONFIG_TMABR, TMABR_BA);
1300 he_writel(he_dev, CONFIG_TPDBA, TPD_BA);
1304 * receive connection memory map
1306 * 0x0 ___________________
1312 * 0x8000|___________________|
1315 * | LBM | link lists of local
1316 * | tx | buffer memory
1318 * 0xd000|___________________|
1321 * 0xe000|___________________|
1324 * |___________________|
1327 * 0xffff|___________________|
1330 he_writel(he_dev, 0x08000, RCMLBM_BA);
1331 he_writel(he_dev, 0x0e000, RCMRSRB_BA);
1332 he_writel(he_dev, 0x0d800, RCMABR_BA);
1334 /* 5.1.4 initialize local buffer free pools linked lists */
1336 he_init_rx_lbfp0(he_dev);
1337 he_init_rx_lbfp1(he_dev);
1339 he_writel(he_dev, 0x0, RLBC_H);
1340 he_writel(he_dev, 0x0, RLBC_T);
1341 he_writel(he_dev, 0x0, RLBC_H2);
1343 he_writel(he_dev, 512, RXTHRSH); /* 10% of r0+r1 buffers */
1344 he_writel(he_dev, 256, LITHRSH); /* 5% of r0+r1 buffers */
1346 he_init_tx_lbfp(he_dev);
1348 he_writel(he_dev, he_is622(he_dev) ? 0x104780 : 0x800, UBUFF_BA);
1350 /* 5.1.5 initialize intermediate receive queues */
1352 if (he_is622(he_dev)) {
1353 he_writel(he_dev, 0x000f, G0_INMQ_S);
1354 he_writel(he_dev, 0x200f, G0_INMQ_L);
1356 he_writel(he_dev, 0x001f, G1_INMQ_S);
1357 he_writel(he_dev, 0x201f, G1_INMQ_L);
1359 he_writel(he_dev, 0x002f, G2_INMQ_S);
1360 he_writel(he_dev, 0x202f, G2_INMQ_L);
1362 he_writel(he_dev, 0x003f, G3_INMQ_S);
1363 he_writel(he_dev, 0x203f, G3_INMQ_L);
1365 he_writel(he_dev, 0x004f, G4_INMQ_S);
1366 he_writel(he_dev, 0x204f, G4_INMQ_L);
1368 he_writel(he_dev, 0x005f, G5_INMQ_S);
1369 he_writel(he_dev, 0x205f, G5_INMQ_L);
1371 he_writel(he_dev, 0x006f, G6_INMQ_S);
1372 he_writel(he_dev, 0x206f, G6_INMQ_L);
1374 he_writel(he_dev, 0x007f, G7_INMQ_S);
1375 he_writel(he_dev, 0x207f, G7_INMQ_L);
1377 he_writel(he_dev, 0x0000, G0_INMQ_S);
1378 he_writel(he_dev, 0x0008, G0_INMQ_L);
1380 he_writel(he_dev, 0x0001, G1_INMQ_S);
1381 he_writel(he_dev, 0x0009, G1_INMQ_L);
1383 he_writel(he_dev, 0x0002, G2_INMQ_S);
1384 he_writel(he_dev, 0x000a, G2_INMQ_L);
1386 he_writel(he_dev, 0x0003, G3_INMQ_S);
1387 he_writel(he_dev, 0x000b, G3_INMQ_L);
1389 he_writel(he_dev, 0x0004, G4_INMQ_S);
1390 he_writel(he_dev, 0x000c, G4_INMQ_L);
1392 he_writel(he_dev, 0x0005, G5_INMQ_S);
1393 he_writel(he_dev, 0x000d, G5_INMQ_L);
1395 he_writel(he_dev, 0x0006, G6_INMQ_S);
1396 he_writel(he_dev, 0x000e, G6_INMQ_L);
1398 he_writel(he_dev, 0x0007, G7_INMQ_S);
1399 he_writel(he_dev, 0x000f, G7_INMQ_L);
1402 /* 5.1.6 application tunable parameters */
1404 he_writel(he_dev, 0x0, MCC);
1405 he_writel(he_dev, 0x0, OEC);
1406 he_writel(he_dev, 0x0, DCC);
1407 he_writel(he_dev, 0x0, CEC);
1409 /* 5.1.7 cs block initialization */
1411 he_init_cs_block(he_dev);
1413 /* 5.1.8 cs block connection memory initialization */
1415 if (he_init_cs_block_rcm(he_dev) < 0)
1418 /* 5.1.10 initialize host structures */
1420 he_init_tpdrq(he_dev);
1422 he_dev->tpd_pool = pci_pool_create("tpd", he_dev->pci_dev,
1423 sizeof(struct he_tpd), TPD_ALIGNMENT, 0);
1424 if (he_dev->tpd_pool == NULL) {
1425 hprintk("unable to create tpd pci_pool\n");
1429 INIT_LIST_HEAD(&he_dev->outstanding_tpds);
1431 if (he_init_group(he_dev, 0) != 0)
1434 for (group = 1; group < HE_NUM_GROUPS; ++group) {
1435 he_writel(he_dev, 0x0, G0_RBPS_S + (group * 32));
1436 he_writel(he_dev, 0x0, G0_RBPS_T + (group * 32));
1437 he_writel(he_dev, 0x0, G0_RBPS_QI + (group * 32));
1438 he_writel(he_dev, RBP_THRESH(0x1) | RBP_QSIZE(0x0),
1439 G0_RBPS_BS + (group * 32));
1441 he_writel(he_dev, 0x0, G0_RBPL_S + (group * 32));
1442 he_writel(he_dev, 0x0, G0_RBPL_T + (group * 32));
1443 he_writel(he_dev, RBP_THRESH(0x1) | RBP_QSIZE(0x0),
1444 G0_RBPL_QI + (group * 32));
1445 he_writel(he_dev, 0x0, G0_RBPL_BS + (group * 32));
1447 he_writel(he_dev, 0x0, G0_RBRQ_ST + (group * 16));
1448 he_writel(he_dev, 0x0, G0_RBRQ_H + (group * 16));
1449 he_writel(he_dev, RBRQ_THRESH(0x1) | RBRQ_SIZE(0x0),
1450 G0_RBRQ_Q + (group * 16));
1451 he_writel(he_dev, 0x0, G0_RBRQ_I + (group * 16));
1453 he_writel(he_dev, 0x0, G0_TBRQ_B_T + (group * 16));
1454 he_writel(he_dev, 0x0, G0_TBRQ_H + (group * 16));
1455 he_writel(he_dev, TBRQ_THRESH(0x1),
1456 G0_TBRQ_THRESH + (group * 16));
1457 he_writel(he_dev, 0x0, G0_TBRQ_S + (group * 16));
1460 /* host status page */
1462 he_dev->hsp = pci_alloc_consistent(he_dev->pci_dev,
1463 sizeof(struct he_hsp), &he_dev->hsp_phys);
1464 if (he_dev->hsp == NULL) {
1465 hprintk("failed to allocate host status page\n");
1468 memset(he_dev->hsp, 0, sizeof(struct he_hsp));
1469 he_writel(he_dev, he_dev->hsp_phys, HSP_BA);
1471 /* initialize framer */
1473 #ifdef CONFIG_ATM_HE_USE_SUNI
1474 if (he_isMM(he_dev))
1475 suni_init(he_dev->atm_dev);
1476 if (he_dev->atm_dev->phy && he_dev->atm_dev->phy->start)
1477 he_dev->atm_dev->phy->start(he_dev->atm_dev);
1478 #endif /* CONFIG_ATM_HE_USE_SUNI */
1481 /* this really should be in suni.c but for now... */
1484 val = he_phy_get(he_dev->atm_dev, SUNI_TPOP_APM);
1485 val = (val & ~SUNI_TPOP_APM_S) | (SUNI_TPOP_S_SDH << SUNI_TPOP_APM_S_SHIFT);
1486 he_phy_put(he_dev->atm_dev, val, SUNI_TPOP_APM);
1487 he_phy_put(he_dev->atm_dev, SUNI_TACP_IUCHP_CLP, SUNI_TACP_IUCHP);
1490 /* 5.1.12 enable transmit and receive */
1492 reg = he_readl_mbox(he_dev, CS_ERCTL0);
1493 reg |= TX_ENABLE|ER_ENABLE;
1494 he_writel_mbox(he_dev, reg, CS_ERCTL0);
1496 reg = he_readl(he_dev, RC_CONFIG);
1498 he_writel(he_dev, reg, RC_CONFIG);
1500 for (i = 0; i < HE_NUM_CS_STPER; ++i) {
1501 he_dev->cs_stper[i].inuse = 0;
1502 he_dev->cs_stper[i].pcr = -1;
1504 he_dev->total_bw = 0;
1507 /* atm linux initialization */
1509 he_dev->atm_dev->ci_range.vpi_bits = he_dev->vpibits;
1510 he_dev->atm_dev->ci_range.vci_bits = he_dev->vcibits;
1512 he_dev->irq_peak = 0;
1513 he_dev->rbrq_peak = 0;
1514 he_dev->rbpl_peak = 0;
1515 he_dev->tbrq_peak = 0;
1517 HPRINTK("hell bent for leather!\n");
1523 he_stop(struct he_dev *he_dev)
1526 u32 gen_cntl_0, reg;
1527 struct pci_dev *pci_dev;
1529 pci_dev = he_dev->pci_dev;
1531 /* disable interrupts */
1533 if (he_dev->membase) {
1534 pci_read_config_dword(pci_dev, GEN_CNTL_0, &gen_cntl_0);
1535 gen_cntl_0 &= ~(INT_PROC_ENBL | INIT_ENB);
1536 pci_write_config_dword(pci_dev, GEN_CNTL_0, gen_cntl_0);
1538 tasklet_disable(&he_dev->tasklet);
1540 /* disable recv and transmit */
1542 reg = he_readl_mbox(he_dev, CS_ERCTL0);
1543 reg &= ~(TX_ENABLE|ER_ENABLE);
1544 he_writel_mbox(he_dev, reg, CS_ERCTL0);
1546 reg = he_readl(he_dev, RC_CONFIG);
1547 reg &= ~(RX_ENABLE);
1548 he_writel(he_dev, reg, RC_CONFIG);
1551 #ifdef CONFIG_ATM_HE_USE_SUNI
1552 if (he_dev->atm_dev->phy && he_dev->atm_dev->phy->stop)
1553 he_dev->atm_dev->phy->stop(he_dev->atm_dev);
1554 #endif /* CONFIG_ATM_HE_USE_SUNI */
1557 free_irq(he_dev->irq, he_dev);
1559 if (he_dev->irq_base)
1560 pci_free_consistent(he_dev->pci_dev, (CONFIG_IRQ_SIZE+1)
1561 * sizeof(struct he_irq), he_dev->irq_base, he_dev->irq_phys);
1564 pci_free_consistent(he_dev->pci_dev, sizeof(struct he_hsp),
1565 he_dev->hsp, he_dev->hsp_phys);
1567 if (he_dev->rbpl_base) {
1570 for (i = 0; i < CONFIG_RBPL_SIZE; ++i) {
1571 void *cpuaddr = he_dev->rbpl_virt[i].virt;
1572 dma_addr_t dma_handle = he_dev->rbpl_base[i].phys;
1574 pci_pool_free(he_dev->rbpl_pool, cpuaddr, dma_handle);
1576 pci_free_consistent(he_dev->pci_dev, CONFIG_RBPL_SIZE
1577 * sizeof(struct he_rbp), he_dev->rbpl_base, he_dev->rbpl_phys);
1580 if (he_dev->rbpl_pool)
1581 pci_pool_destroy(he_dev->rbpl_pool);
1583 if (he_dev->rbrq_base)
1584 pci_free_consistent(he_dev->pci_dev, CONFIG_RBRQ_SIZE * sizeof(struct he_rbrq),
1585 he_dev->rbrq_base, he_dev->rbrq_phys);
1587 if (he_dev->tbrq_base)
1588 pci_free_consistent(he_dev->pci_dev, CONFIG_TBRQ_SIZE * sizeof(struct he_tbrq),
1589 he_dev->tbrq_base, he_dev->tbrq_phys);
1591 if (he_dev->tpdrq_base)
1592 pci_free_consistent(he_dev->pci_dev, CONFIG_TBRQ_SIZE * sizeof(struct he_tbrq),
1593 he_dev->tpdrq_base, he_dev->tpdrq_phys);
1595 if (he_dev->tpd_pool)
1596 pci_pool_destroy(he_dev->tpd_pool);
1598 if (he_dev->pci_dev) {
1599 pci_read_config_word(he_dev->pci_dev, PCI_COMMAND, &command);
1600 command &= ~(PCI_COMMAND_MEMORY | PCI_COMMAND_MASTER);
1601 pci_write_config_word(he_dev->pci_dev, PCI_COMMAND, command);
1604 if (he_dev->membase)
1605 iounmap(he_dev->membase);
1608 static struct he_tpd *
1609 __alloc_tpd(struct he_dev *he_dev)
1612 dma_addr_t dma_handle;
1614 tpd = pci_pool_alloc(he_dev->tpd_pool, GFP_ATOMIC|GFP_DMA, &dma_handle);
1618 tpd->status = TPD_ADDR(dma_handle);
1620 tpd->iovec[0].addr = 0; tpd->iovec[0].len = 0;
1621 tpd->iovec[1].addr = 0; tpd->iovec[1].len = 0;
1622 tpd->iovec[2].addr = 0; tpd->iovec[2].len = 0;
1627 #define AAL5_LEN(buf,len) \
1628 ((((unsigned char *)(buf))[(len)-6] << 8) | \
1629 (((unsigned char *)(buf))[(len)-5]))
1633 * aal5 packets can optionally return the tcp checksum in the lower
1634 * 16 bits of the crc (RSR0_TCP_CKSUM)
1637 #define TCP_CKSUM(buf,len) \
1638 ((((unsigned char *)(buf))[(len)-2] << 8) | \
1639 (((unsigned char *)(buf))[(len-1)]))
1642 he_service_rbrq(struct he_dev *he_dev, int group)
1644 struct he_rbrq *rbrq_tail = (struct he_rbrq *)
1645 ((unsigned long)he_dev->rbrq_base |
1646 he_dev->hsp->group[group].rbrq_tail);
1647 struct he_rbp *rbp = NULL;
1648 unsigned cid, lastcid = -1;
1649 unsigned buf_len = 0;
1650 struct sk_buff *skb;
1651 struct atm_vcc *vcc = NULL;
1652 struct he_vcc *he_vcc;
1653 struct he_iovec *iov;
1654 int pdus_assembled = 0;
1657 read_lock(&vcc_sklist_lock);
1658 while (he_dev->rbrq_head != rbrq_tail) {
1661 HPRINTK("%p rbrq%d 0x%x len=%d cid=0x%x %s%s%s%s%s%s\n",
1662 he_dev->rbrq_head, group,
1663 RBRQ_ADDR(he_dev->rbrq_head),
1664 RBRQ_BUFLEN(he_dev->rbrq_head),
1665 RBRQ_CID(he_dev->rbrq_head),
1666 RBRQ_CRC_ERR(he_dev->rbrq_head) ? " CRC_ERR" : "",
1667 RBRQ_LEN_ERR(he_dev->rbrq_head) ? " LEN_ERR" : "",
1668 RBRQ_END_PDU(he_dev->rbrq_head) ? " END_PDU" : "",
1669 RBRQ_AAL5_PROT(he_dev->rbrq_head) ? " AAL5_PROT" : "",
1670 RBRQ_CON_CLOSED(he_dev->rbrq_head) ? " CON_CLOSED" : "",
1671 RBRQ_HBUF_ERR(he_dev->rbrq_head) ? " HBUF_ERR" : "");
1673 rbp = &he_dev->rbpl_base[RBP_INDEX(RBRQ_ADDR(he_dev->rbrq_head))];
1675 buf_len = RBRQ_BUFLEN(he_dev->rbrq_head) * 4;
1676 cid = RBRQ_CID(he_dev->rbrq_head);
1679 vcc = __find_vcc(he_dev, cid);
1683 hprintk("vcc == NULL (cid 0x%x)\n", cid);
1684 if (!RBRQ_HBUF_ERR(he_dev->rbrq_head))
1685 rbp->status &= ~RBP_LOANED;
1687 goto next_rbrq_entry;
1690 he_vcc = HE_VCC(vcc);
1691 if (he_vcc == NULL) {
1692 hprintk("he_vcc == NULL (cid 0x%x)\n", cid);
1693 if (!RBRQ_HBUF_ERR(he_dev->rbrq_head))
1694 rbp->status &= ~RBP_LOANED;
1695 goto next_rbrq_entry;
1698 if (RBRQ_HBUF_ERR(he_dev->rbrq_head)) {
1699 hprintk("HBUF_ERR! (cid 0x%x)\n", cid);
1700 atomic_inc(&vcc->stats->rx_drop);
1701 goto return_host_buffers;
1704 he_vcc->iov_tail->iov_base = RBRQ_ADDR(he_dev->rbrq_head);
1705 he_vcc->iov_tail->iov_len = buf_len;
1706 he_vcc->pdu_len += buf_len;
1709 if (RBRQ_CON_CLOSED(he_dev->rbrq_head)) {
1711 HPRINTK("wake_up rx_waitq (cid 0x%x)\n", cid);
1712 wake_up(&he_vcc->rx_waitq);
1713 goto return_host_buffers;
1717 if ((he_vcc->iov_tail - he_vcc->iov_head) > HE_MAXIOV) {
1718 hprintk("iovec full! cid 0x%x\n", cid);
1719 goto return_host_buffers;
1722 if (!RBRQ_END_PDU(he_dev->rbrq_head))
1723 goto next_rbrq_entry;
1725 if (RBRQ_LEN_ERR(he_dev->rbrq_head)
1726 || RBRQ_CRC_ERR(he_dev->rbrq_head)) {
1727 HPRINTK("%s%s (%d.%d)\n",
1728 RBRQ_CRC_ERR(he_dev->rbrq_head)
1730 RBRQ_LEN_ERR(he_dev->rbrq_head)
1732 vcc->vpi, vcc->vci);
1733 atomic_inc(&vcc->stats->rx_err);
1734 goto return_host_buffers;
1737 skb = atm_alloc_charge(vcc, he_vcc->pdu_len + rx_skb_reserve,
1740 HPRINTK("charge failed (%d.%d)\n", vcc->vpi, vcc->vci);
1741 goto return_host_buffers;
1744 if (rx_skb_reserve > 0)
1745 skb_reserve(skb, rx_skb_reserve);
1747 __net_timestamp(skb);
1749 for (iov = he_vcc->iov_head; iov < he_vcc->iov_tail; ++iov)
1750 memcpy(skb_put(skb, iov->iov_len),
1751 he_dev->rbpl_virt[RBP_INDEX(iov->iov_base)].virt, iov->iov_len);
1753 switch (vcc->qos.aal) {
1755 /* 2.10.1.5 raw cell receive */
1756 skb->len = ATM_AAL0_SDU;
1757 skb_set_tail_pointer(skb, skb->len);
1760 /* 2.10.1.2 aal5 receive */
1762 skb->len = AAL5_LEN(skb->data, he_vcc->pdu_len);
1763 skb_set_tail_pointer(skb, skb->len);
1764 #ifdef USE_CHECKSUM_HW
1765 if (vcc->vpi == 0 && vcc->vci >= ATM_NOT_RSV_VCI) {
1766 skb->ip_summed = CHECKSUM_COMPLETE;
1767 skb->csum = TCP_CKSUM(skb->data,
1774 #ifdef should_never_happen
1775 if (skb->len > vcc->qos.rxtp.max_sdu)
1776 hprintk("pdu_len (%d) > vcc->qos.rxtp.max_sdu (%d)! cid 0x%x\n", skb->len, vcc->qos.rxtp.max_sdu, cid);
1780 ATM_SKB(skb)->vcc = vcc;
1782 spin_unlock(&he_dev->global_lock);
1783 vcc->push(vcc, skb);
1784 spin_lock(&he_dev->global_lock);
1786 atomic_inc(&vcc->stats->rx);
1788 return_host_buffers:
1791 for (iov = he_vcc->iov_head; iov < he_vcc->iov_tail; ++iov) {
1792 rbp = &he_dev->rbpl_base[RBP_INDEX(iov->iov_base)];
1793 rbp->status &= ~RBP_LOANED;
1796 he_vcc->iov_tail = he_vcc->iov_head;
1797 he_vcc->pdu_len = 0;
1800 he_dev->rbrq_head = (struct he_rbrq *)
1801 ((unsigned long) he_dev->rbrq_base |
1802 RBRQ_MASK(++he_dev->rbrq_head));
1805 read_unlock(&vcc_sklist_lock);
1808 if (updated > he_dev->rbrq_peak)
1809 he_dev->rbrq_peak = updated;
1811 he_writel(he_dev, RBRQ_MASK(he_dev->rbrq_head),
1812 G0_RBRQ_H + (group * 16));
1815 return pdus_assembled;
1819 he_service_tbrq(struct he_dev *he_dev, int group)
1821 struct he_tbrq *tbrq_tail = (struct he_tbrq *)
1822 ((unsigned long)he_dev->tbrq_base |
1823 he_dev->hsp->group[group].tbrq_tail);
1825 int slot, updated = 0;
1826 struct he_tpd *__tpd;
1828 /* 2.1.6 transmit buffer return queue */
1830 while (he_dev->tbrq_head != tbrq_tail) {
1833 HPRINTK("tbrq%d 0x%x%s%s\n",
1835 TBRQ_TPD(he_dev->tbrq_head),
1836 TBRQ_EOS(he_dev->tbrq_head) ? " EOS" : "",
1837 TBRQ_MULTIPLE(he_dev->tbrq_head) ? " MULTIPLE" : "");
1839 list_for_each_entry(__tpd, &he_dev->outstanding_tpds, entry) {
1840 if (TPD_ADDR(__tpd->status) == TBRQ_TPD(he_dev->tbrq_head)) {
1842 list_del(&__tpd->entry);
1848 hprintk("unable to locate tpd for dma buffer %x\n",
1849 TBRQ_TPD(he_dev->tbrq_head));
1850 goto next_tbrq_entry;
1853 if (TBRQ_EOS(he_dev->tbrq_head)) {
1854 HPRINTK("wake_up(tx_waitq) cid 0x%x\n",
1855 he_mkcid(he_dev, tpd->vcc->vpi, tpd->vcc->vci));
1857 wake_up(&HE_VCC(tpd->vcc)->tx_waitq);
1859 goto next_tbrq_entry;
1862 for (slot = 0; slot < TPD_MAXIOV; ++slot) {
1863 if (tpd->iovec[slot].addr)
1864 pci_unmap_single(he_dev->pci_dev,
1865 tpd->iovec[slot].addr,
1866 tpd->iovec[slot].len & TPD_LEN_MASK,
1868 if (tpd->iovec[slot].len & TPD_LST)
1873 if (tpd->skb) { /* && !TBRQ_MULTIPLE(he_dev->tbrq_head) */
1874 if (tpd->vcc && tpd->vcc->pop)
1875 tpd->vcc->pop(tpd->vcc, tpd->skb);
1877 dev_kfree_skb_any(tpd->skb);
1882 pci_pool_free(he_dev->tpd_pool, tpd, TPD_ADDR(tpd->status));
1883 he_dev->tbrq_head = (struct he_tbrq *)
1884 ((unsigned long) he_dev->tbrq_base |
1885 TBRQ_MASK(++he_dev->tbrq_head));
1889 if (updated > he_dev->tbrq_peak)
1890 he_dev->tbrq_peak = updated;
1892 he_writel(he_dev, TBRQ_MASK(he_dev->tbrq_head),
1893 G0_TBRQ_H + (group * 16));
1898 he_service_rbpl(struct he_dev *he_dev, int group)
1900 struct he_rbp *newtail;
1901 struct he_rbp *rbpl_head;
1904 rbpl_head = (struct he_rbp *) ((unsigned long)he_dev->rbpl_base |
1905 RBPL_MASK(he_readl(he_dev, G0_RBPL_S)));
1908 newtail = (struct he_rbp *) ((unsigned long)he_dev->rbpl_base |
1909 RBPL_MASK(he_dev->rbpl_tail+1));
1911 /* table 3.42 -- rbpl_tail should never be set to rbpl_head */
1912 if ((newtail == rbpl_head) || (newtail->status & RBP_LOANED))
1915 newtail->status |= RBP_LOANED;
1916 he_dev->rbpl_tail = newtail;
1921 he_writel(he_dev, RBPL_MASK(he_dev->rbpl_tail), G0_RBPL_T);
1925 he_tasklet(unsigned long data)
1927 unsigned long flags;
1928 struct he_dev *he_dev = (struct he_dev *) data;
1932 HPRINTK("tasklet (0x%lx)\n", data);
1933 spin_lock_irqsave(&he_dev->global_lock, flags);
1935 while (he_dev->irq_head != he_dev->irq_tail) {
1938 type = ITYPE_TYPE(he_dev->irq_head->isw);
1939 group = ITYPE_GROUP(he_dev->irq_head->isw);
1942 case ITYPE_RBRQ_THRESH:
1943 HPRINTK("rbrq%d threshold\n", group);
1945 case ITYPE_RBRQ_TIMER:
1946 if (he_service_rbrq(he_dev, group))
1947 he_service_rbpl(he_dev, group);
1949 case ITYPE_TBRQ_THRESH:
1950 HPRINTK("tbrq%d threshold\n", group);
1952 case ITYPE_TPD_COMPLETE:
1953 he_service_tbrq(he_dev, group);
1955 case ITYPE_RBPL_THRESH:
1956 he_service_rbpl(he_dev, group);
1958 case ITYPE_RBPS_THRESH:
1959 /* shouldn't happen unless small buffers enabled */
1962 HPRINTK("phy interrupt\n");
1963 #ifdef CONFIG_ATM_HE_USE_SUNI
1964 spin_unlock_irqrestore(&he_dev->global_lock, flags);
1965 if (he_dev->atm_dev->phy && he_dev->atm_dev->phy->interrupt)
1966 he_dev->atm_dev->phy->interrupt(he_dev->atm_dev);
1967 spin_lock_irqsave(&he_dev->global_lock, flags);
1971 switch (type|group) {
1973 hprintk("parity error\n");
1976 hprintk("abort 0x%x\n", he_readl(he_dev, ABORT_ADDR));
1980 case ITYPE_TYPE(ITYPE_INVALID):
1981 /* see 8.1.1 -- check all queues */
1983 HPRINTK("isw not updated 0x%x\n", he_dev->irq_head->isw);
1985 he_service_rbrq(he_dev, 0);
1986 he_service_rbpl(he_dev, 0);
1987 he_service_tbrq(he_dev, 0);
1990 hprintk("bad isw 0x%x?\n", he_dev->irq_head->isw);
1993 he_dev->irq_head->isw = ITYPE_INVALID;
1995 he_dev->irq_head = (struct he_irq *) NEXT_ENTRY(he_dev->irq_base, he_dev->irq_head, IRQ_MASK);
1999 if (updated > he_dev->irq_peak)
2000 he_dev->irq_peak = updated;
2003 IRQ_SIZE(CONFIG_IRQ_SIZE) |
2004 IRQ_THRESH(CONFIG_IRQ_THRESH) |
2005 IRQ_TAIL(he_dev->irq_tail), IRQ0_HEAD);
2006 (void) he_readl(he_dev, INT_FIFO); /* 8.1.2 controller errata; flush posted writes */
2008 spin_unlock_irqrestore(&he_dev->global_lock, flags);
2012 he_irq_handler(int irq, void *dev_id)
2014 unsigned long flags;
2015 struct he_dev *he_dev = (struct he_dev * )dev_id;
2021 spin_lock_irqsave(&he_dev->global_lock, flags);
2023 he_dev->irq_tail = (struct he_irq *) (((unsigned long)he_dev->irq_base) |
2024 (*he_dev->irq_tailoffset << 2));
2026 if (he_dev->irq_tail == he_dev->irq_head) {
2027 HPRINTK("tailoffset not updated?\n");
2028 he_dev->irq_tail = (struct he_irq *) ((unsigned long)he_dev->irq_base |
2029 ((he_readl(he_dev, IRQ0_BASE) & IRQ_MASK) << 2));
2030 (void) he_readl(he_dev, INT_FIFO); /* 8.1.2 controller errata */
2034 if (he_dev->irq_head == he_dev->irq_tail /* && !IRQ_PENDING */)
2035 hprintk("spurious (or shared) interrupt?\n");
2038 if (he_dev->irq_head != he_dev->irq_tail) {
2040 tasklet_schedule(&he_dev->tasklet);
2041 he_writel(he_dev, INT_CLEAR_A, INT_FIFO); /* clear interrupt */
2042 (void) he_readl(he_dev, INT_FIFO); /* flush posted writes */
2044 spin_unlock_irqrestore(&he_dev->global_lock, flags);
2045 return IRQ_RETVAL(handled);
2049 static __inline__ void
2050 __enqueue_tpd(struct he_dev *he_dev, struct he_tpd *tpd, unsigned cid)
2052 struct he_tpdrq *new_tail;
2054 HPRINTK("tpdrq %p cid 0x%x -> tpdrq_tail %p\n",
2055 tpd, cid, he_dev->tpdrq_tail);
2057 /* new_tail = he_dev->tpdrq_tail; */
2058 new_tail = (struct he_tpdrq *) ((unsigned long) he_dev->tpdrq_base |
2059 TPDRQ_MASK(he_dev->tpdrq_tail+1));
2062 * check to see if we are about to set the tail == head
2063 * if true, update the head pointer from the adapter
2064 * to see if this is really the case (reading the queue
2065 * head for every enqueue would be unnecessarily slow)
2068 if (new_tail == he_dev->tpdrq_head) {
2069 he_dev->tpdrq_head = (struct he_tpdrq *)
2070 (((unsigned long)he_dev->tpdrq_base) |
2071 TPDRQ_MASK(he_readl(he_dev, TPDRQ_B_H)));
2073 if (new_tail == he_dev->tpdrq_head) {
2076 hprintk("tpdrq full (cid 0x%x)\n", cid);
2079 * push tpd onto a transmit backlog queue
2080 * after service_tbrq, service the backlog
2081 * for now, we just drop the pdu
2083 for (slot = 0; slot < TPD_MAXIOV; ++slot) {
2084 if (tpd->iovec[slot].addr)
2085 pci_unmap_single(he_dev->pci_dev,
2086 tpd->iovec[slot].addr,
2087 tpd->iovec[slot].len & TPD_LEN_MASK,
2092 tpd->vcc->pop(tpd->vcc, tpd->skb);
2094 dev_kfree_skb_any(tpd->skb);
2095 atomic_inc(&tpd->vcc->stats->tx_err);
2097 pci_pool_free(he_dev->tpd_pool, tpd, TPD_ADDR(tpd->status));
2102 /* 2.1.5 transmit packet descriptor ready queue */
2103 list_add_tail(&tpd->entry, &he_dev->outstanding_tpds);
2104 he_dev->tpdrq_tail->tpd = TPD_ADDR(tpd->status);
2105 he_dev->tpdrq_tail->cid = cid;
2108 he_dev->tpdrq_tail = new_tail;
2110 he_writel(he_dev, TPDRQ_MASK(he_dev->tpdrq_tail), TPDRQ_T);
2111 (void) he_readl(he_dev, TPDRQ_T); /* flush posted writes */
2115 he_open(struct atm_vcc *vcc)
2117 unsigned long flags;
2118 struct he_dev *he_dev = HE_DEV(vcc->dev);
2119 struct he_vcc *he_vcc;
2121 unsigned cid, rsr0, rsr1, rsr4, tsr0, tsr0_aal, tsr4, period, reg, clock;
2122 short vpi = vcc->vpi;
2125 if (vci == ATM_VCI_UNSPEC || vpi == ATM_VPI_UNSPEC)
2128 HPRINTK("open vcc %p %d.%d\n", vcc, vpi, vci);
2130 set_bit(ATM_VF_ADDR, &vcc->flags);
2132 cid = he_mkcid(he_dev, vpi, vci);
2134 he_vcc = kmalloc(sizeof(struct he_vcc), GFP_ATOMIC);
2135 if (he_vcc == NULL) {
2136 hprintk("unable to allocate he_vcc during open\n");
2140 he_vcc->iov_tail = he_vcc->iov_head;
2141 he_vcc->pdu_len = 0;
2142 he_vcc->rc_index = -1;
2144 init_waitqueue_head(&he_vcc->rx_waitq);
2145 init_waitqueue_head(&he_vcc->tx_waitq);
2147 vcc->dev_data = he_vcc;
2149 if (vcc->qos.txtp.traffic_class != ATM_NONE) {
2152 pcr_goal = atm_pcr_goal(&vcc->qos.txtp);
2154 pcr_goal = he_dev->atm_dev->link_rate;
2155 if (pcr_goal < 0) /* means round down, technically */
2156 pcr_goal = -pcr_goal;
2158 HPRINTK("open tx cid 0x%x pcr_goal %d\n", cid, pcr_goal);
2160 switch (vcc->qos.aal) {
2162 tsr0_aal = TSR0_AAL5;
2166 tsr0_aal = TSR0_AAL0_SDU;
2167 tsr4 = TSR4_AAL0_SDU;
2174 spin_lock_irqsave(&he_dev->global_lock, flags);
2175 tsr0 = he_readl_tsr0(he_dev, cid);
2176 spin_unlock_irqrestore(&he_dev->global_lock, flags);
2178 if (TSR0_CONN_STATE(tsr0) != 0) {
2179 hprintk("cid 0x%x not idle (tsr0 = 0x%x)\n", cid, tsr0);
2184 switch (vcc->qos.txtp.traffic_class) {
2186 /* 2.3.3.1 open connection ubr */
2188 tsr0 = TSR0_UBR | TSR0_GROUP(0) | tsr0_aal |
2189 TSR0_USE_WMIN | TSR0_UPDATE_GER;
2193 /* 2.3.3.2 open connection cbr */
2195 /* 8.2.3 cbr scheduler wrap problem -- limit to 90% total link rate */
2196 if ((he_dev->total_bw + pcr_goal)
2197 > (he_dev->atm_dev->link_rate * 9 / 10))
2203 spin_lock_irqsave(&he_dev->global_lock, flags); /* also protects he_dev->cs_stper[] */
2205 /* find an unused cs_stper register */
2206 for (reg = 0; reg < HE_NUM_CS_STPER; ++reg)
2207 if (he_dev->cs_stper[reg].inuse == 0 ||
2208 he_dev->cs_stper[reg].pcr == pcr_goal)
2211 if (reg == HE_NUM_CS_STPER) {
2213 spin_unlock_irqrestore(&he_dev->global_lock, flags);
2217 he_dev->total_bw += pcr_goal;
2219 he_vcc->rc_index = reg;
2220 ++he_dev->cs_stper[reg].inuse;
2221 he_dev->cs_stper[reg].pcr = pcr_goal;
2223 clock = he_is622(he_dev) ? 66667000 : 50000000;
2224 period = clock / pcr_goal;
2226 HPRINTK("rc_index = %d period = %d\n",
2229 he_writel_mbox(he_dev, rate_to_atmf(period/2),
2231 spin_unlock_irqrestore(&he_dev->global_lock, flags);
2233 tsr0 = TSR0_CBR | TSR0_GROUP(0) | tsr0_aal |
2242 spin_lock_irqsave(&he_dev->global_lock, flags);
2244 he_writel_tsr0(he_dev, tsr0, cid);
2245 he_writel_tsr4(he_dev, tsr4 | 1, cid);
2246 he_writel_tsr1(he_dev, TSR1_MCR(rate_to_atmf(0)) |
2247 TSR1_PCR(rate_to_atmf(pcr_goal)), cid);
2248 he_writel_tsr2(he_dev, TSR2_ACR(rate_to_atmf(pcr_goal)), cid);
2249 he_writel_tsr9(he_dev, TSR9_OPEN_CONN, cid);
2251 he_writel_tsr3(he_dev, 0x0, cid);
2252 he_writel_tsr5(he_dev, 0x0, cid);
2253 he_writel_tsr6(he_dev, 0x0, cid);
2254 he_writel_tsr7(he_dev, 0x0, cid);
2255 he_writel_tsr8(he_dev, 0x0, cid);
2256 he_writel_tsr10(he_dev, 0x0, cid);
2257 he_writel_tsr11(he_dev, 0x0, cid);
2258 he_writel_tsr12(he_dev, 0x0, cid);
2259 he_writel_tsr13(he_dev, 0x0, cid);
2260 he_writel_tsr14(he_dev, 0x0, cid);
2261 (void) he_readl_tsr0(he_dev, cid); /* flush posted writes */
2262 spin_unlock_irqrestore(&he_dev->global_lock, flags);
2265 if (vcc->qos.rxtp.traffic_class != ATM_NONE) {
2268 HPRINTK("open rx cid 0x%x (rx_waitq %p)\n", cid,
2269 &HE_VCC(vcc)->rx_waitq);
2271 switch (vcc->qos.aal) {
2283 spin_lock_irqsave(&he_dev->global_lock, flags);
2285 rsr0 = he_readl_rsr0(he_dev, cid);
2286 if (rsr0 & RSR0_OPEN_CONN) {
2287 spin_unlock_irqrestore(&he_dev->global_lock, flags);
2289 hprintk("cid 0x%x not idle (rsr0 = 0x%x)\n", cid, rsr0);
2294 rsr1 = RSR1_GROUP(0) | RSR1_RBPL_ONLY;
2295 rsr4 = RSR4_GROUP(0) | RSR4_RBPL_ONLY;
2296 rsr0 = vcc->qos.rxtp.traffic_class == ATM_UBR ?
2297 (RSR0_EPD_ENABLE|RSR0_PPD_ENABLE) : 0;
2299 #ifdef USE_CHECKSUM_HW
2300 if (vpi == 0 && vci >= ATM_NOT_RSV_VCI)
2301 rsr0 |= RSR0_TCP_CKSUM;
2304 he_writel_rsr4(he_dev, rsr4, cid);
2305 he_writel_rsr1(he_dev, rsr1, cid);
2306 /* 5.1.11 last parameter initialized should be
2307 the open/closed indication in rsr0 */
2308 he_writel_rsr0(he_dev,
2309 rsr0 | RSR0_START_PDU | RSR0_OPEN_CONN | aal, cid);
2310 (void) he_readl_rsr0(he_dev, cid); /* flush posted writes */
2312 spin_unlock_irqrestore(&he_dev->global_lock, flags);
2319 clear_bit(ATM_VF_ADDR, &vcc->flags);
2322 set_bit(ATM_VF_READY, &vcc->flags);
2328 he_close(struct atm_vcc *vcc)
2330 unsigned long flags;
2331 DECLARE_WAITQUEUE(wait, current);
2332 struct he_dev *he_dev = HE_DEV(vcc->dev);
2335 struct he_vcc *he_vcc = HE_VCC(vcc);
2336 #define MAX_RETRY 30
2337 int retry = 0, sleep = 1, tx_inuse;
2339 HPRINTK("close vcc %p %d.%d\n", vcc, vcc->vpi, vcc->vci);
2341 clear_bit(ATM_VF_READY, &vcc->flags);
2342 cid = he_mkcid(he_dev, vcc->vpi, vcc->vci);
2344 if (vcc->qos.rxtp.traffic_class != ATM_NONE) {
2347 HPRINTK("close rx cid 0x%x\n", cid);
2349 /* 2.7.2.2 close receive operation */
2351 /* wait for previous close (if any) to finish */
2353 spin_lock_irqsave(&he_dev->global_lock, flags);
2354 while (he_readl(he_dev, RCC_STAT) & RCC_BUSY) {
2355 HPRINTK("close cid 0x%x RCC_BUSY\n", cid);
2359 set_current_state(TASK_UNINTERRUPTIBLE);
2360 add_wait_queue(&he_vcc->rx_waitq, &wait);
2362 he_writel_rsr0(he_dev, RSR0_CLOSE_CONN, cid);
2363 (void) he_readl_rsr0(he_dev, cid); /* flush posted writes */
2364 he_writel_mbox(he_dev, cid, RXCON_CLOSE);
2365 spin_unlock_irqrestore(&he_dev->global_lock, flags);
2367 timeout = schedule_timeout(30*HZ);
2369 remove_wait_queue(&he_vcc->rx_waitq, &wait);
2370 set_current_state(TASK_RUNNING);
2373 hprintk("close rx timeout cid 0x%x\n", cid);
2375 HPRINTK("close rx cid 0x%x complete\n", cid);
2379 if (vcc->qos.txtp.traffic_class != ATM_NONE) {
2380 volatile unsigned tsr4, tsr0;
2383 HPRINTK("close tx cid 0x%x\n", cid);
2387 * ... the host must first stop queueing packets to the TPDRQ
2388 * on the connection to be closed, then wait for all outstanding
2389 * packets to be transmitted and their buffers returned to the
2390 * TBRQ. When the last packet on the connection arrives in the
2391 * TBRQ, the host issues the close command to the adapter.
2394 while (((tx_inuse = atomic_read(&sk_atm(vcc)->sk_wmem_alloc)) > 1) &&
2395 (retry < MAX_RETRY)) {
2404 hprintk("close tx cid 0x%x tx_inuse = %d\n", cid, tx_inuse);
2406 /* 2.3.1.1 generic close operations with flush */
2408 spin_lock_irqsave(&he_dev->global_lock, flags);
2409 he_writel_tsr4_upper(he_dev, TSR4_FLUSH_CONN, cid);
2410 /* also clears TSR4_SESSION_ENDED */
2412 switch (vcc->qos.txtp.traffic_class) {
2414 he_writel_tsr1(he_dev,
2415 TSR1_MCR(rate_to_atmf(200000))
2416 | TSR1_PCR(0), cid);
2419 he_writel_tsr14_upper(he_dev, TSR14_DELETE, cid);
2422 (void) he_readl_tsr4(he_dev, cid); /* flush posted writes */
2424 tpd = __alloc_tpd(he_dev);
2426 hprintk("close tx he_alloc_tpd failed cid 0x%x\n", cid);
2427 goto close_tx_incomplete;
2429 tpd->status |= TPD_EOS | TPD_INT;
2434 set_current_state(TASK_UNINTERRUPTIBLE);
2435 add_wait_queue(&he_vcc->tx_waitq, &wait);
2436 __enqueue_tpd(he_dev, tpd, cid);
2437 spin_unlock_irqrestore(&he_dev->global_lock, flags);
2439 timeout = schedule_timeout(30*HZ);
2441 remove_wait_queue(&he_vcc->tx_waitq, &wait);
2442 set_current_state(TASK_RUNNING);
2444 spin_lock_irqsave(&he_dev->global_lock, flags);
2447 hprintk("close tx timeout cid 0x%x\n", cid);
2448 goto close_tx_incomplete;
2451 while (!((tsr4 = he_readl_tsr4(he_dev, cid)) & TSR4_SESSION_ENDED)) {
2452 HPRINTK("close tx cid 0x%x !TSR4_SESSION_ENDED (tsr4 = 0x%x)\n", cid, tsr4);
2456 while (TSR0_CONN_STATE(tsr0 = he_readl_tsr0(he_dev, cid)) != 0) {
2457 HPRINTK("close tx cid 0x%x TSR0_CONN_STATE != 0 (tsr0 = 0x%x)\n", cid, tsr0);
2461 close_tx_incomplete:
2463 if (vcc->qos.txtp.traffic_class == ATM_CBR) {
2464 int reg = he_vcc->rc_index;
2466 HPRINTK("cs_stper reg = %d\n", reg);
2468 if (he_dev->cs_stper[reg].inuse == 0)
2469 hprintk("cs_stper[%d].inuse = 0!\n", reg);
2471 --he_dev->cs_stper[reg].inuse;
2473 he_dev->total_bw -= he_dev->cs_stper[reg].pcr;
2475 spin_unlock_irqrestore(&he_dev->global_lock, flags);
2477 HPRINTK("close tx cid 0x%x complete\n", cid);
2482 clear_bit(ATM_VF_ADDR, &vcc->flags);
2486 he_send(struct atm_vcc *vcc, struct sk_buff *skb)
2488 unsigned long flags;
2489 struct he_dev *he_dev = HE_DEV(vcc->dev);
2490 unsigned cid = he_mkcid(he_dev, vcc->vpi, vcc->vci);
2492 #ifdef USE_SCATTERGATHER
2496 #define HE_TPD_BUFSIZE 0xffff
2498 HPRINTK("send %d.%d\n", vcc->vpi, vcc->vci);
2500 if ((skb->len > HE_TPD_BUFSIZE) ||
2501 ((vcc->qos.aal == ATM_AAL0) && (skb->len != ATM_AAL0_SDU))) {
2502 hprintk("buffer too large (or small) -- %d bytes\n", skb->len );
2506 dev_kfree_skb_any(skb);
2507 atomic_inc(&vcc->stats->tx_err);
2511 #ifndef USE_SCATTERGATHER
2512 if (skb_shinfo(skb)->nr_frags) {
2513 hprintk("no scatter/gather support\n");
2517 dev_kfree_skb_any(skb);
2518 atomic_inc(&vcc->stats->tx_err);
2522 spin_lock_irqsave(&he_dev->global_lock, flags);
2524 tpd = __alloc_tpd(he_dev);
2529 dev_kfree_skb_any(skb);
2530 atomic_inc(&vcc->stats->tx_err);
2531 spin_unlock_irqrestore(&he_dev->global_lock, flags);
2535 if (vcc->qos.aal == ATM_AAL5)
2536 tpd->status |= TPD_CELLTYPE(TPD_USERCELL);
2538 char *pti_clp = (void *) (skb->data + 3);
2541 pti = (*pti_clp & ATM_HDR_PTI_MASK) >> ATM_HDR_PTI_SHIFT;
2542 clp = (*pti_clp & ATM_HDR_CLP);
2543 tpd->status |= TPD_CELLTYPE(pti);
2545 tpd->status |= TPD_CLP;
2547 skb_pull(skb, ATM_AAL0_SDU - ATM_CELL_PAYLOAD);
2550 #ifdef USE_SCATTERGATHER
2551 tpd->iovec[slot].addr = pci_map_single(he_dev->pci_dev, skb->data,
2552 skb_headlen(skb), PCI_DMA_TODEVICE);
2553 tpd->iovec[slot].len = skb_headlen(skb);
2556 for (i = 0; i < skb_shinfo(skb)->nr_frags; i++) {
2557 skb_frag_t *frag = &skb_shinfo(skb)->frags[i];
2559 if (slot == TPD_MAXIOV) { /* queue tpd; start new tpd */
2561 tpd->skb = NULL; /* not the last fragment
2562 so dont ->push() yet */
2565 __enqueue_tpd(he_dev, tpd, cid);
2566 tpd = __alloc_tpd(he_dev);
2571 dev_kfree_skb_any(skb);
2572 atomic_inc(&vcc->stats->tx_err);
2573 spin_unlock_irqrestore(&he_dev->global_lock, flags);
2576 tpd->status |= TPD_USERCELL;
2580 tpd->iovec[slot].addr = pci_map_single(he_dev->pci_dev,
2581 (void *) page_address(frag->page) + frag->page_offset,
2582 frag->size, PCI_DMA_TODEVICE);
2583 tpd->iovec[slot].len = frag->size;
2588 tpd->iovec[slot - 1].len |= TPD_LST;
2590 tpd->address0 = pci_map_single(he_dev->pci_dev, skb->data, skb->len, PCI_DMA_TODEVICE);
2591 tpd->length0 = skb->len | TPD_LST;
2593 tpd->status |= TPD_INT;
2598 ATM_SKB(skb)->vcc = vcc;
2600 __enqueue_tpd(he_dev, tpd, cid);
2601 spin_unlock_irqrestore(&he_dev->global_lock, flags);
2603 atomic_inc(&vcc->stats->tx);
2609 he_ioctl(struct atm_dev *atm_dev, unsigned int cmd, void __user *arg)
2611 unsigned long flags;
2612 struct he_dev *he_dev = HE_DEV(atm_dev);
2613 struct he_ioctl_reg reg;
2618 if (!capable(CAP_NET_ADMIN))
2621 if (copy_from_user(®, arg,
2622 sizeof(struct he_ioctl_reg)))
2625 spin_lock_irqsave(&he_dev->global_lock, flags);
2627 case HE_REGTYPE_PCI:
2628 if (reg.addr >= HE_REGMAP_SIZE) {
2633 reg.val = he_readl(he_dev, reg.addr);
2635 case HE_REGTYPE_RCM:
2637 he_readl_rcm(he_dev, reg.addr);
2639 case HE_REGTYPE_TCM:
2641 he_readl_tcm(he_dev, reg.addr);
2643 case HE_REGTYPE_MBOX:
2645 he_readl_mbox(he_dev, reg.addr);
2651 spin_unlock_irqrestore(&he_dev->global_lock, flags);
2653 if (copy_to_user(arg, ®,
2654 sizeof(struct he_ioctl_reg)))
2658 #ifdef CONFIG_ATM_HE_USE_SUNI
2659 if (atm_dev->phy && atm_dev->phy->ioctl)
2660 err = atm_dev->phy->ioctl(atm_dev, cmd, arg);
2661 #else /* CONFIG_ATM_HE_USE_SUNI */
2663 #endif /* CONFIG_ATM_HE_USE_SUNI */
2671 he_phy_put(struct atm_dev *atm_dev, unsigned char val, unsigned long addr)
2673 unsigned long flags;
2674 struct he_dev *he_dev = HE_DEV(atm_dev);
2676 HPRINTK("phy_put(val 0x%x, addr 0x%lx)\n", val, addr);
2678 spin_lock_irqsave(&he_dev->global_lock, flags);
2679 he_writel(he_dev, val, FRAMER + (addr*4));
2680 (void) he_readl(he_dev, FRAMER + (addr*4)); /* flush posted writes */
2681 spin_unlock_irqrestore(&he_dev->global_lock, flags);
2685 static unsigned char
2686 he_phy_get(struct atm_dev *atm_dev, unsigned long addr)
2688 unsigned long flags;
2689 struct he_dev *he_dev = HE_DEV(atm_dev);
2692 spin_lock_irqsave(&he_dev->global_lock, flags);
2693 reg = he_readl(he_dev, FRAMER + (addr*4));
2694 spin_unlock_irqrestore(&he_dev->global_lock, flags);
2696 HPRINTK("phy_get(addr 0x%lx) =0x%x\n", addr, reg);
2701 he_proc_read(struct atm_dev *dev, loff_t *pos, char *page)
2703 unsigned long flags;
2704 struct he_dev *he_dev = HE_DEV(dev);
2707 struct he_rbrq *rbrq_tail;
2708 struct he_tpdrq *tpdrq_head;
2709 int rbpl_head, rbpl_tail;
2711 static long mcc = 0, oec = 0, dcc = 0, cec = 0;
2716 return sprintf(page, "ATM he driver\n");
2719 return sprintf(page, "%s%s\n\n",
2720 he_dev->prod_id, he_dev->media & 0x40 ? "SM" : "MM");
2723 return sprintf(page, "Mismatched Cells VPI/VCI Not Open Dropped Cells RCM Dropped Cells\n");
2725 spin_lock_irqsave(&he_dev->global_lock, flags);
2726 mcc += he_readl(he_dev, MCC);
2727 oec += he_readl(he_dev, OEC);
2728 dcc += he_readl(he_dev, DCC);
2729 cec += he_readl(he_dev, CEC);
2730 spin_unlock_irqrestore(&he_dev->global_lock, flags);
2733 return sprintf(page, "%16ld %16ld %13ld %17ld\n\n",
2734 mcc, oec, dcc, cec);
2737 return sprintf(page, "irq_size = %d inuse = ? peak = %d\n",
2738 CONFIG_IRQ_SIZE, he_dev->irq_peak);
2741 return sprintf(page, "tpdrq_size = %d inuse = ?\n",
2745 return sprintf(page, "rbrq_size = %d inuse = ? peak = %d\n",
2746 CONFIG_RBRQ_SIZE, he_dev->rbrq_peak);
2749 return sprintf(page, "tbrq_size = %d peak = %d\n",
2750 CONFIG_TBRQ_SIZE, he_dev->tbrq_peak);
2754 rbpl_head = RBPL_MASK(he_readl(he_dev, G0_RBPL_S));
2755 rbpl_tail = RBPL_MASK(he_readl(he_dev, G0_RBPL_T));
2757 inuse = rbpl_head - rbpl_tail;
2759 inuse += CONFIG_RBPL_SIZE * sizeof(struct he_rbp);
2760 inuse /= sizeof(struct he_rbp);
2763 return sprintf(page, "rbpl_size = %d inuse = %d\n\n",
2764 CONFIG_RBPL_SIZE, inuse);
2768 return sprintf(page, "rate controller periods (cbr)\n pcr #vc\n");
2770 for (i = 0; i < HE_NUM_CS_STPER; ++i)
2772 return sprintf(page, "cs_stper%-2d %8ld %3d\n", i,
2773 he_dev->cs_stper[i].pcr,
2774 he_dev->cs_stper[i].inuse);
2777 return sprintf(page, "total bw (cbr): %d (limit %d)\n",
2778 he_dev->total_bw, he_dev->atm_dev->link_rate * 10 / 9);
2783 /* eeprom routines -- see 4.7 */
2785 static u8 read_prom_byte(struct he_dev *he_dev, int addr)
2787 u32 val = 0, tmp_read = 0;
2791 val = readl(he_dev->membase + HOST_CNTL);
2794 /* Turn on write enable */
2796 he_writel(he_dev, val, HOST_CNTL);
2798 /* Send READ instruction */
2799 for (i = 0; i < ARRAY_SIZE(readtab); i++) {
2800 he_writel(he_dev, val | readtab[i], HOST_CNTL);
2801 udelay(EEPROM_DELAY);
2804 /* Next, we need to send the byte address to read from */
2805 for (i = 7; i >= 0; i--) {
2806 he_writel(he_dev, val | clocktab[j++] | (((addr >> i) & 1) << 9), HOST_CNTL);
2807 udelay(EEPROM_DELAY);
2808 he_writel(he_dev, val | clocktab[j++] | (((addr >> i) & 1) << 9), HOST_CNTL);
2809 udelay(EEPROM_DELAY);
2814 val &= 0xFFFFF7FF; /* Turn off write enable */
2815 he_writel(he_dev, val, HOST_CNTL);
2817 /* Now, we can read data from the EEPROM by clocking it in */
2818 for (i = 7; i >= 0; i--) {
2819 he_writel(he_dev, val | clocktab[j++], HOST_CNTL);
2820 udelay(EEPROM_DELAY);
2821 tmp_read = he_readl(he_dev, HOST_CNTL);
2822 byte_read |= (unsigned char)
2823 ((tmp_read & ID_DOUT) >> ID_DOFFSET << i);
2824 he_writel(he_dev, val | clocktab[j++], HOST_CNTL);
2825 udelay(EEPROM_DELAY);
2828 he_writel(he_dev, val | ID_CS, HOST_CNTL);
2829 udelay(EEPROM_DELAY);
2834 MODULE_LICENSE("GPL");
2835 MODULE_AUTHOR("chas williams <chas@cmf.nrl.navy.mil>");
2836 MODULE_DESCRIPTION("ForeRunnerHE ATM Adapter driver");
2837 module_param(disable64, bool, 0);
2838 MODULE_PARM_DESC(disable64, "disable 64-bit pci bus transfers");
2839 module_param(nvpibits, short, 0);
2840 MODULE_PARM_DESC(nvpibits, "numbers of bits for vpi (default 0)");
2841 module_param(nvcibits, short, 0);
2842 MODULE_PARM_DESC(nvcibits, "numbers of bits for vci (default 12)");
2843 module_param(rx_skb_reserve, short, 0);
2844 MODULE_PARM_DESC(rx_skb_reserve, "padding for receive skb (default 16)");
2845 module_param(irq_coalesce, bool, 0);
2846 MODULE_PARM_DESC(irq_coalesce, "use interrupt coalescing (default 1)");
2847 module_param(sdh, bool, 0);
2848 MODULE_PARM_DESC(sdh, "use SDH framing (default 0)");
2850 static struct pci_device_id he_pci_tbl[] = {
2851 { PCI_VENDOR_ID_FORE, PCI_DEVICE_ID_FORE_HE, PCI_ANY_ID, PCI_ANY_ID,
2856 MODULE_DEVICE_TABLE(pci, he_pci_tbl);
2858 static struct pci_driver he_driver = {
2860 .probe = he_init_one,
2861 .remove = __devexit_p(he_remove_one),
2862 .id_table = he_pci_tbl,
2865 static int __init he_init(void)
2867 return pci_register_driver(&he_driver);
2870 static void __exit he_cleanup(void)
2872 pci_unregister_driver(&he_driver);
2875 module_init(he_init);
2876 module_exit(he_cleanup);