Merge tag 'nfs-for-3.8-2' of git://git.linux-nfs.org/projects/trondmy/linux-nfs
[platform/adaptation/renesas_rcar/renesas_kernel.git] / drivers / atm / he.c
1 /*
2
3   he.c
4
5   ForeRunnerHE ATM Adapter driver for ATM on Linux
6   Copyright (C) 1999-2001  Naval Research Laboratory
7
8   This library is free software; you can redistribute it and/or
9   modify it under the terms of the GNU Lesser General Public
10   License as published by the Free Software Foundation; either
11   version 2.1 of the License, or (at your option) any later version.
12
13   This library is distributed in the hope that it will be useful,
14   but WITHOUT ANY WARRANTY; without even the implied warranty of
15   MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE.  See the GNU
16   Lesser General Public License for more details.
17
18   You should have received a copy of the GNU Lesser General Public
19   License along with this library; if not, write to the Free Software
20   Foundation, Inc., 59 Temple Place, Suite 330, Boston, MA  02111-1307  USA
21
22 */
23
24 /*
25
26   he.c
27
28   ForeRunnerHE ATM Adapter driver for ATM on Linux
29   Copyright (C) 1999-2001  Naval Research Laboratory
30
31   Permission to use, copy, modify and distribute this software and its
32   documentation is hereby granted, provided that both the copyright
33   notice and this permission notice appear in all copies of the software,
34   derivative works or modified versions, and any portions thereof, and
35   that both notices appear in supporting documentation.
36
37   NRL ALLOWS FREE USE OF THIS SOFTWARE IN ITS "AS IS" CONDITION AND
38   DISCLAIMS ANY LIABILITY OF ANY KIND FOR ANY DAMAGES WHATSOEVER
39   RESULTING FROM THE USE OF THIS SOFTWARE.
40
41   This driver was written using the "Programmer's Reference Manual for
42   ForeRunnerHE(tm)", MANU0361-01 - Rev. A, 08/21/98.
43
44   AUTHORS:
45         chas williams <chas@cmf.nrl.navy.mil>
46         eric kinzie <ekinzie@cmf.nrl.navy.mil>
47
48   NOTES:
49         4096 supported 'connections'
50         group 0 is used for all traffic
51         interrupt queue 0 is used for all interrupts
52         aal0 support (based on work from ulrich.u.muller@nokia.com)
53
54  */
55
56 #include <linux/module.h>
57 #include <linux/kernel.h>
58 #include <linux/skbuff.h>
59 #include <linux/pci.h>
60 #include <linux/errno.h>
61 #include <linux/types.h>
62 #include <linux/string.h>
63 #include <linux/delay.h>
64 #include <linux/init.h>
65 #include <linux/mm.h>
66 #include <linux/sched.h>
67 #include <linux/timer.h>
68 #include <linux/interrupt.h>
69 #include <linux/dma-mapping.h>
70 #include <linux/bitmap.h>
71 #include <linux/slab.h>
72 #include <asm/io.h>
73 #include <asm/byteorder.h>
74 #include <asm/uaccess.h>
75
76 #include <linux/atmdev.h>
77 #include <linux/atm.h>
78 #include <linux/sonet.h>
79
80 #undef USE_SCATTERGATHER
81 #undef USE_CHECKSUM_HW                  /* still confused about this */
82 /* #undef HE_DEBUG */
83
84 #include "he.h"
85 #include "suni.h"
86 #include <linux/atm_he.h>
87
88 #define hprintk(fmt,args...)    printk(KERN_ERR DEV_LABEL "%d: " fmt, he_dev->number , ##args)
89
90 #ifdef HE_DEBUG
91 #define HPRINTK(fmt,args...)    printk(KERN_DEBUG DEV_LABEL "%d: " fmt, he_dev->number , ##args)
92 #else /* !HE_DEBUG */
93 #define HPRINTK(fmt,args...)    do { } while (0)
94 #endif /* HE_DEBUG */
95
96 /* declarations */
97
98 static int he_open(struct atm_vcc *vcc);
99 static void he_close(struct atm_vcc *vcc);
100 static int he_send(struct atm_vcc *vcc, struct sk_buff *skb);
101 static int he_ioctl(struct atm_dev *dev, unsigned int cmd, void __user *arg);
102 static irqreturn_t he_irq_handler(int irq, void *dev_id);
103 static void he_tasklet(unsigned long data);
104 static int he_proc_read(struct atm_dev *dev,loff_t *pos,char *page);
105 static int he_start(struct atm_dev *dev);
106 static void he_stop(struct he_dev *dev);
107 static void he_phy_put(struct atm_dev *, unsigned char, unsigned long);
108 static unsigned char he_phy_get(struct atm_dev *, unsigned long);
109
110 static u8 read_prom_byte(struct he_dev *he_dev, int addr);
111
112 /* globals */
113
114 static struct he_dev *he_devs;
115 static bool disable64;
116 static short nvpibits = -1;
117 static short nvcibits = -1;
118 static short rx_skb_reserve = 16;
119 static bool irq_coalesce = 1;
120 static bool sdh = 0;
121
122 /* Read from EEPROM = 0000 0011b */
123 static unsigned int readtab[] = {
124         CS_HIGH | CLK_HIGH,
125         CS_LOW | CLK_LOW,
126         CLK_HIGH,               /* 0 */
127         CLK_LOW,
128         CLK_HIGH,               /* 0 */
129         CLK_LOW,
130         CLK_HIGH,               /* 0 */
131         CLK_LOW,
132         CLK_HIGH,               /* 0 */
133         CLK_LOW,
134         CLK_HIGH,               /* 0 */
135         CLK_LOW,
136         CLK_HIGH,               /* 0 */
137         CLK_LOW | SI_HIGH,
138         CLK_HIGH | SI_HIGH,     /* 1 */
139         CLK_LOW | SI_HIGH,
140         CLK_HIGH | SI_HIGH      /* 1 */
141 };     
142  
143 /* Clock to read from/write to the EEPROM */
144 static unsigned int clocktab[] = {
145         CLK_LOW,
146         CLK_HIGH,
147         CLK_LOW,
148         CLK_HIGH,
149         CLK_LOW,
150         CLK_HIGH,
151         CLK_LOW,
152         CLK_HIGH,
153         CLK_LOW,
154         CLK_HIGH,
155         CLK_LOW,
156         CLK_HIGH,
157         CLK_LOW,
158         CLK_HIGH,
159         CLK_LOW,
160         CLK_HIGH,
161         CLK_LOW
162 };     
163
164 static struct atmdev_ops he_ops =
165 {
166         .open =         he_open,
167         .close =        he_close,       
168         .ioctl =        he_ioctl,       
169         .send =         he_send,
170         .phy_put =      he_phy_put,
171         .phy_get =      he_phy_get,
172         .proc_read =    he_proc_read,
173         .owner =        THIS_MODULE
174 };
175
176 #define he_writel(dev, val, reg)        do { writel(val, (dev)->membase + (reg)); wmb(); } while (0)
177 #define he_readl(dev, reg)              readl((dev)->membase + (reg))
178
179 /* section 2.12 connection memory access */
180
181 static __inline__ void
182 he_writel_internal(struct he_dev *he_dev, unsigned val, unsigned addr,
183                                                                 unsigned flags)
184 {
185         he_writel(he_dev, val, CON_DAT);
186         (void) he_readl(he_dev, CON_DAT);               /* flush posted writes */
187         he_writel(he_dev, flags | CON_CTL_WRITE | CON_CTL_ADDR(addr), CON_CTL);
188         while (he_readl(he_dev, CON_CTL) & CON_CTL_BUSY);
189 }
190
191 #define he_writel_rcm(dev, val, reg)                            \
192                         he_writel_internal(dev, val, reg, CON_CTL_RCM)
193
194 #define he_writel_tcm(dev, val, reg)                            \
195                         he_writel_internal(dev, val, reg, CON_CTL_TCM)
196
197 #define he_writel_mbox(dev, val, reg)                           \
198                         he_writel_internal(dev, val, reg, CON_CTL_MBOX)
199
200 static unsigned
201 he_readl_internal(struct he_dev *he_dev, unsigned addr, unsigned flags)
202 {
203         he_writel(he_dev, flags | CON_CTL_READ | CON_CTL_ADDR(addr), CON_CTL);
204         while (he_readl(he_dev, CON_CTL) & CON_CTL_BUSY);
205         return he_readl(he_dev, CON_DAT);
206 }
207
208 #define he_readl_rcm(dev, reg) \
209                         he_readl_internal(dev, reg, CON_CTL_RCM)
210
211 #define he_readl_tcm(dev, reg) \
212                         he_readl_internal(dev, reg, CON_CTL_TCM)
213
214 #define he_readl_mbox(dev, reg) \
215                         he_readl_internal(dev, reg, CON_CTL_MBOX)
216
217
218 /* figure 2.2 connection id */
219
220 #define he_mkcid(dev, vpi, vci)         (((vpi << (dev)->vcibits) | vci) & 0x1fff)
221
222 /* 2.5.1 per connection transmit state registers */
223
224 #define he_writel_tsr0(dev, val, cid) \
225                 he_writel_tcm(dev, val, CONFIG_TSRA | (cid << 3) | 0)
226 #define he_readl_tsr0(dev, cid) \
227                 he_readl_tcm(dev, CONFIG_TSRA | (cid << 3) | 0)
228
229 #define he_writel_tsr1(dev, val, cid) \
230                 he_writel_tcm(dev, val, CONFIG_TSRA | (cid << 3) | 1)
231
232 #define he_writel_tsr2(dev, val, cid) \
233                 he_writel_tcm(dev, val, CONFIG_TSRA | (cid << 3) | 2)
234
235 #define he_writel_tsr3(dev, val, cid) \
236                 he_writel_tcm(dev, val, CONFIG_TSRA | (cid << 3) | 3)
237
238 #define he_writel_tsr4(dev, val, cid) \
239                 he_writel_tcm(dev, val, CONFIG_TSRA | (cid << 3) | 4)
240
241         /* from page 2-20
242          *
243          * NOTE While the transmit connection is active, bits 23 through 0
244          *      of this register must not be written by the host.  Byte
245          *      enables should be used during normal operation when writing
246          *      the most significant byte.
247          */
248
249 #define he_writel_tsr4_upper(dev, val, cid) \
250                 he_writel_internal(dev, val, CONFIG_TSRA | (cid << 3) | 4, \
251                                                         CON_CTL_TCM \
252                                                         | CON_BYTE_DISABLE_2 \
253                                                         | CON_BYTE_DISABLE_1 \
254                                                         | CON_BYTE_DISABLE_0)
255
256 #define he_readl_tsr4(dev, cid) \
257                 he_readl_tcm(dev, CONFIG_TSRA | (cid << 3) | 4)
258
259 #define he_writel_tsr5(dev, val, cid) \
260                 he_writel_tcm(dev, val, CONFIG_TSRA | (cid << 3) | 5)
261
262 #define he_writel_tsr6(dev, val, cid) \
263                 he_writel_tcm(dev, val, CONFIG_TSRA | (cid << 3) | 6)
264
265 #define he_writel_tsr7(dev, val, cid) \
266                 he_writel_tcm(dev, val, CONFIG_TSRA | (cid << 3) | 7)
267
268
269 #define he_writel_tsr8(dev, val, cid) \
270                 he_writel_tcm(dev, val, CONFIG_TSRB | (cid << 2) | 0)
271
272 #define he_writel_tsr9(dev, val, cid) \
273                 he_writel_tcm(dev, val, CONFIG_TSRB | (cid << 2) | 1)
274
275 #define he_writel_tsr10(dev, val, cid) \
276                 he_writel_tcm(dev, val, CONFIG_TSRB | (cid << 2) | 2)
277
278 #define he_writel_tsr11(dev, val, cid) \
279                 he_writel_tcm(dev, val, CONFIG_TSRB | (cid << 2) | 3)
280
281
282 #define he_writel_tsr12(dev, val, cid) \
283                 he_writel_tcm(dev, val, CONFIG_TSRC | (cid << 1) | 0)
284
285 #define he_writel_tsr13(dev, val, cid) \
286                 he_writel_tcm(dev, val, CONFIG_TSRC | (cid << 1) | 1)
287
288
289 #define he_writel_tsr14(dev, val, cid) \
290                 he_writel_tcm(dev, val, CONFIG_TSRD | cid)
291
292 #define he_writel_tsr14_upper(dev, val, cid) \
293                 he_writel_internal(dev, val, CONFIG_TSRD | cid, \
294                                                         CON_CTL_TCM \
295                                                         | CON_BYTE_DISABLE_2 \
296                                                         | CON_BYTE_DISABLE_1 \
297                                                         | CON_BYTE_DISABLE_0)
298
299 /* 2.7.1 per connection receive state registers */
300
301 #define he_writel_rsr0(dev, val, cid) \
302                 he_writel_rcm(dev, val, 0x00000 | (cid << 3) | 0)
303 #define he_readl_rsr0(dev, cid) \
304                 he_readl_rcm(dev, 0x00000 | (cid << 3) | 0)
305
306 #define he_writel_rsr1(dev, val, cid) \
307                 he_writel_rcm(dev, val, 0x00000 | (cid << 3) | 1)
308
309 #define he_writel_rsr2(dev, val, cid) \
310                 he_writel_rcm(dev, val, 0x00000 | (cid << 3) | 2)
311
312 #define he_writel_rsr3(dev, val, cid) \
313                 he_writel_rcm(dev, val, 0x00000 | (cid << 3) | 3)
314
315 #define he_writel_rsr4(dev, val, cid) \
316                 he_writel_rcm(dev, val, 0x00000 | (cid << 3) | 4)
317
318 #define he_writel_rsr5(dev, val, cid) \
319                 he_writel_rcm(dev, val, 0x00000 | (cid << 3) | 5)
320
321 #define he_writel_rsr6(dev, val, cid) \
322                 he_writel_rcm(dev, val, 0x00000 | (cid << 3) | 6)
323
324 #define he_writel_rsr7(dev, val, cid) \
325                 he_writel_rcm(dev, val, 0x00000 | (cid << 3) | 7)
326
327 static __inline__ struct atm_vcc*
328 __find_vcc(struct he_dev *he_dev, unsigned cid)
329 {
330         struct hlist_head *head;
331         struct atm_vcc *vcc;
332         struct hlist_node *node;
333         struct sock *s;
334         short vpi;
335         int vci;
336
337         vpi = cid >> he_dev->vcibits;
338         vci = cid & ((1 << he_dev->vcibits) - 1);
339         head = &vcc_hash[vci & (VCC_HTABLE_SIZE -1)];
340
341         sk_for_each(s, node, head) {
342                 vcc = atm_sk(s);
343                 if (vcc->dev == he_dev->atm_dev &&
344                     vcc->vci == vci && vcc->vpi == vpi &&
345                     vcc->qos.rxtp.traffic_class != ATM_NONE) {
346                                 return vcc;
347                 }
348         }
349         return NULL;
350 }
351
352 static int he_init_one(struct pci_dev *pci_dev,
353                        const struct pci_device_id *pci_ent)
354 {
355         struct atm_dev *atm_dev = NULL;
356         struct he_dev *he_dev = NULL;
357         int err = 0;
358
359         printk(KERN_INFO "ATM he driver\n");
360
361         if (pci_enable_device(pci_dev))
362                 return -EIO;
363         if (pci_set_dma_mask(pci_dev, DMA_BIT_MASK(32)) != 0) {
364                 printk(KERN_WARNING "he: no suitable dma available\n");
365                 err = -EIO;
366                 goto init_one_failure;
367         }
368
369         atm_dev = atm_dev_register(DEV_LABEL, &pci_dev->dev, &he_ops, -1, NULL);
370         if (!atm_dev) {
371                 err = -ENODEV;
372                 goto init_one_failure;
373         }
374         pci_set_drvdata(pci_dev, atm_dev);
375
376         he_dev = kzalloc(sizeof(struct he_dev),
377                                                         GFP_KERNEL);
378         if (!he_dev) {
379                 err = -ENOMEM;
380                 goto init_one_failure;
381         }
382         he_dev->pci_dev = pci_dev;
383         he_dev->atm_dev = atm_dev;
384         he_dev->atm_dev->dev_data = he_dev;
385         atm_dev->dev_data = he_dev;
386         he_dev->number = atm_dev->number;
387         tasklet_init(&he_dev->tasklet, he_tasklet, (unsigned long) he_dev);
388         spin_lock_init(&he_dev->global_lock);
389
390         if (he_start(atm_dev)) {
391                 he_stop(he_dev);
392                 err = -ENODEV;
393                 goto init_one_failure;
394         }
395         he_dev->next = NULL;
396         if (he_devs)
397                 he_dev->next = he_devs;
398         he_devs = he_dev;
399         return 0;
400
401 init_one_failure:
402         if (atm_dev)
403                 atm_dev_deregister(atm_dev);
404         kfree(he_dev);
405         pci_disable_device(pci_dev);
406         return err;
407 }
408
409 static void he_remove_one(struct pci_dev *pci_dev)
410 {
411         struct atm_dev *atm_dev;
412         struct he_dev *he_dev;
413
414         atm_dev = pci_get_drvdata(pci_dev);
415         he_dev = HE_DEV(atm_dev);
416
417         /* need to remove from he_devs */
418
419         he_stop(he_dev);
420         atm_dev_deregister(atm_dev);
421         kfree(he_dev);
422
423         pci_set_drvdata(pci_dev, NULL);
424         pci_disable_device(pci_dev);
425 }
426
427
428 static unsigned
429 rate_to_atmf(unsigned rate)             /* cps to atm forum format */
430 {
431 #define NONZERO (1 << 14)
432
433         unsigned exp = 0;
434
435         if (rate == 0)
436                 return 0;
437
438         rate <<= 9;
439         while (rate > 0x3ff) {
440                 ++exp;
441                 rate >>= 1;
442         }
443
444         return (NONZERO | (exp << 9) | (rate & 0x1ff));
445 }
446
447 static void he_init_rx_lbfp0(struct he_dev *he_dev)
448 {
449         unsigned i, lbm_offset, lbufd_index, lbuf_addr, lbuf_count;
450         unsigned lbufs_per_row = he_dev->cells_per_row / he_dev->cells_per_lbuf;
451         unsigned lbuf_bufsize = he_dev->cells_per_lbuf * ATM_CELL_PAYLOAD;
452         unsigned row_offset = he_dev->r0_startrow * he_dev->bytes_per_row;
453         
454         lbufd_index = 0;
455         lbm_offset = he_readl(he_dev, RCMLBM_BA);
456
457         he_writel(he_dev, lbufd_index, RLBF0_H);
458
459         for (i = 0, lbuf_count = 0; i < he_dev->r0_numbuffs; ++i) {
460                 lbufd_index += 2;
461                 lbuf_addr = (row_offset + (lbuf_count * lbuf_bufsize)) / 32;
462
463                 he_writel_rcm(he_dev, lbuf_addr, lbm_offset);
464                 he_writel_rcm(he_dev, lbufd_index, lbm_offset + 1);
465
466                 if (++lbuf_count == lbufs_per_row) {
467                         lbuf_count = 0;
468                         row_offset += he_dev->bytes_per_row;
469                 }
470                 lbm_offset += 4;
471         }
472                 
473         he_writel(he_dev, lbufd_index - 2, RLBF0_T);
474         he_writel(he_dev, he_dev->r0_numbuffs, RLBF0_C);
475 }
476
477 static void he_init_rx_lbfp1(struct he_dev *he_dev)
478 {
479         unsigned i, lbm_offset, lbufd_index, lbuf_addr, lbuf_count;
480         unsigned lbufs_per_row = he_dev->cells_per_row / he_dev->cells_per_lbuf;
481         unsigned lbuf_bufsize = he_dev->cells_per_lbuf * ATM_CELL_PAYLOAD;
482         unsigned row_offset = he_dev->r1_startrow * he_dev->bytes_per_row;
483         
484         lbufd_index = 1;
485         lbm_offset = he_readl(he_dev, RCMLBM_BA) + (2 * lbufd_index);
486
487         he_writel(he_dev, lbufd_index, RLBF1_H);
488
489         for (i = 0, lbuf_count = 0; i < he_dev->r1_numbuffs; ++i) {
490                 lbufd_index += 2;
491                 lbuf_addr = (row_offset + (lbuf_count * lbuf_bufsize)) / 32;
492
493                 he_writel_rcm(he_dev, lbuf_addr, lbm_offset);
494                 he_writel_rcm(he_dev, lbufd_index, lbm_offset + 1);
495
496                 if (++lbuf_count == lbufs_per_row) {
497                         lbuf_count = 0;
498                         row_offset += he_dev->bytes_per_row;
499                 }
500                 lbm_offset += 4;
501         }
502                 
503         he_writel(he_dev, lbufd_index - 2, RLBF1_T);
504         he_writel(he_dev, he_dev->r1_numbuffs, RLBF1_C);
505 }
506
507 static void he_init_tx_lbfp(struct he_dev *he_dev)
508 {
509         unsigned i, lbm_offset, lbufd_index, lbuf_addr, lbuf_count;
510         unsigned lbufs_per_row = he_dev->cells_per_row / he_dev->cells_per_lbuf;
511         unsigned lbuf_bufsize = he_dev->cells_per_lbuf * ATM_CELL_PAYLOAD;
512         unsigned row_offset = he_dev->tx_startrow * he_dev->bytes_per_row;
513         
514         lbufd_index = he_dev->r0_numbuffs + he_dev->r1_numbuffs;
515         lbm_offset = he_readl(he_dev, RCMLBM_BA) + (2 * lbufd_index);
516
517         he_writel(he_dev, lbufd_index, TLBF_H);
518
519         for (i = 0, lbuf_count = 0; i < he_dev->tx_numbuffs; ++i) {
520                 lbufd_index += 1;
521                 lbuf_addr = (row_offset + (lbuf_count * lbuf_bufsize)) / 32;
522
523                 he_writel_rcm(he_dev, lbuf_addr, lbm_offset);
524                 he_writel_rcm(he_dev, lbufd_index, lbm_offset + 1);
525
526                 if (++lbuf_count == lbufs_per_row) {
527                         lbuf_count = 0;
528                         row_offset += he_dev->bytes_per_row;
529                 }
530                 lbm_offset += 2;
531         }
532                 
533         he_writel(he_dev, lbufd_index - 1, TLBF_T);
534 }
535
536 static int he_init_tpdrq(struct he_dev *he_dev)
537 {
538         he_dev->tpdrq_base = pci_alloc_consistent(he_dev->pci_dev,
539                 CONFIG_TPDRQ_SIZE * sizeof(struct he_tpdrq), &he_dev->tpdrq_phys);
540         if (he_dev->tpdrq_base == NULL) {
541                 hprintk("failed to alloc tpdrq\n");
542                 return -ENOMEM;
543         }
544         memset(he_dev->tpdrq_base, 0,
545                                 CONFIG_TPDRQ_SIZE * sizeof(struct he_tpdrq));
546
547         he_dev->tpdrq_tail = he_dev->tpdrq_base;
548         he_dev->tpdrq_head = he_dev->tpdrq_base;
549
550         he_writel(he_dev, he_dev->tpdrq_phys, TPDRQ_B_H);
551         he_writel(he_dev, 0, TPDRQ_T);  
552         he_writel(he_dev, CONFIG_TPDRQ_SIZE - 1, TPDRQ_S);
553
554         return 0;
555 }
556
557 static void he_init_cs_block(struct he_dev *he_dev)
558 {
559         unsigned clock, rate, delta;
560         int reg;
561
562         /* 5.1.7 cs block initialization */
563
564         for (reg = 0; reg < 0x20; ++reg)
565                 he_writel_mbox(he_dev, 0x0, CS_STTIM0 + reg);
566
567         /* rate grid timer reload values */
568
569         clock = he_is622(he_dev) ? 66667000 : 50000000;
570         rate = he_dev->atm_dev->link_rate;
571         delta = rate / 16 / 2;
572
573         for (reg = 0; reg < 0x10; ++reg) {
574                 /* 2.4 internal transmit function
575                  *
576                  * we initialize the first row in the rate grid.
577                  * values are period (in clock cycles) of timer
578                  */
579                 unsigned period = clock / rate;
580
581                 he_writel_mbox(he_dev, period, CS_TGRLD0 + reg);
582                 rate -= delta;
583         }
584
585         if (he_is622(he_dev)) {
586                 /* table 5.2 (4 cells per lbuf) */
587                 he_writel_mbox(he_dev, 0x000800fa, CS_ERTHR0);
588                 he_writel_mbox(he_dev, 0x000c33cb, CS_ERTHR1);
589                 he_writel_mbox(he_dev, 0x0010101b, CS_ERTHR2);
590                 he_writel_mbox(he_dev, 0x00181dac, CS_ERTHR3);
591                 he_writel_mbox(he_dev, 0x00280600, CS_ERTHR4);
592
593                 /* table 5.3, 5.4, 5.5, 5.6, 5.7 */
594                 he_writel_mbox(he_dev, 0x023de8b3, CS_ERCTL0);
595                 he_writel_mbox(he_dev, 0x1801, CS_ERCTL1);
596                 he_writel_mbox(he_dev, 0x68b3, CS_ERCTL2);
597                 he_writel_mbox(he_dev, 0x1280, CS_ERSTAT0);
598                 he_writel_mbox(he_dev, 0x68b3, CS_ERSTAT1);
599                 he_writel_mbox(he_dev, 0x14585, CS_RTFWR);
600
601                 he_writel_mbox(he_dev, 0x4680, CS_RTATR);
602
603                 /* table 5.8 */
604                 he_writel_mbox(he_dev, 0x00159ece, CS_TFBSET);
605                 he_writel_mbox(he_dev, 0x68b3, CS_WCRMAX);
606                 he_writel_mbox(he_dev, 0x5eb3, CS_WCRMIN);
607                 he_writel_mbox(he_dev, 0xe8b3, CS_WCRINC);
608                 he_writel_mbox(he_dev, 0xdeb3, CS_WCRDEC);
609                 he_writel_mbox(he_dev, 0x68b3, CS_WCRCEIL);
610
611                 /* table 5.9 */
612                 he_writel_mbox(he_dev, 0x5, CS_OTPPER);
613                 he_writel_mbox(he_dev, 0x14, CS_OTWPER);
614         } else {
615                 /* table 5.1 (4 cells per lbuf) */
616                 he_writel_mbox(he_dev, 0x000400ea, CS_ERTHR0);
617                 he_writel_mbox(he_dev, 0x00063388, CS_ERTHR1);
618                 he_writel_mbox(he_dev, 0x00081018, CS_ERTHR2);
619                 he_writel_mbox(he_dev, 0x000c1dac, CS_ERTHR3);
620                 he_writel_mbox(he_dev, 0x0014051a, CS_ERTHR4);
621
622                 /* table 5.3, 5.4, 5.5, 5.6, 5.7 */
623                 he_writel_mbox(he_dev, 0x0235e4b1, CS_ERCTL0);
624                 he_writel_mbox(he_dev, 0x4701, CS_ERCTL1);
625                 he_writel_mbox(he_dev, 0x64b1, CS_ERCTL2);
626                 he_writel_mbox(he_dev, 0x1280, CS_ERSTAT0);
627                 he_writel_mbox(he_dev, 0x64b1, CS_ERSTAT1);
628                 he_writel_mbox(he_dev, 0xf424, CS_RTFWR);
629
630                 he_writel_mbox(he_dev, 0x4680, CS_RTATR);
631
632                 /* table 5.8 */
633                 he_writel_mbox(he_dev, 0x000563b7, CS_TFBSET);
634                 he_writel_mbox(he_dev, 0x64b1, CS_WCRMAX);
635                 he_writel_mbox(he_dev, 0x5ab1, CS_WCRMIN);
636                 he_writel_mbox(he_dev, 0xe4b1, CS_WCRINC);
637                 he_writel_mbox(he_dev, 0xdab1, CS_WCRDEC);
638                 he_writel_mbox(he_dev, 0x64b1, CS_WCRCEIL);
639
640                 /* table 5.9 */
641                 he_writel_mbox(he_dev, 0x6, CS_OTPPER);
642                 he_writel_mbox(he_dev, 0x1e, CS_OTWPER);
643         }
644
645         he_writel_mbox(he_dev, 0x8, CS_OTTLIM);
646
647         for (reg = 0; reg < 0x8; ++reg)
648                 he_writel_mbox(he_dev, 0x0, CS_HGRRT0 + reg);
649
650 }
651
652 static int he_init_cs_block_rcm(struct he_dev *he_dev)
653 {
654         unsigned (*rategrid)[16][16];
655         unsigned rate, delta;
656         int i, j, reg;
657
658         unsigned rate_atmf, exp, man;
659         unsigned long long rate_cps;
660         int mult, buf, buf_limit = 4;
661
662         rategrid = kmalloc( sizeof(unsigned) * 16 * 16, GFP_KERNEL);
663         if (!rategrid)
664                 return -ENOMEM;
665
666         /* initialize rate grid group table */
667
668         for (reg = 0x0; reg < 0xff; ++reg)
669                 he_writel_rcm(he_dev, 0x0, CONFIG_RCMABR + reg);
670
671         /* initialize rate controller groups */
672
673         for (reg = 0x100; reg < 0x1ff; ++reg)
674                 he_writel_rcm(he_dev, 0x0, CONFIG_RCMABR + reg);
675         
676         /* initialize tNrm lookup table */
677
678         /* the manual makes reference to a routine in a sample driver
679            for proper configuration; fortunately, we only need this
680            in order to support abr connection */
681         
682         /* initialize rate to group table */
683
684         rate = he_dev->atm_dev->link_rate;
685         delta = rate / 32;
686
687         /*
688          * 2.4 transmit internal functions
689          * 
690          * we construct a copy of the rate grid used by the scheduler
691          * in order to construct the rate to group table below
692          */
693
694         for (j = 0; j < 16; j++) {
695                 (*rategrid)[0][j] = rate;
696                 rate -= delta;
697         }
698
699         for (i = 1; i < 16; i++)
700                 for (j = 0; j < 16; j++)
701                         if (i > 14)
702                                 (*rategrid)[i][j] = (*rategrid)[i - 1][j] / 4;
703                         else
704                                 (*rategrid)[i][j] = (*rategrid)[i - 1][j] / 2;
705
706         /*
707          * 2.4 transmit internal function
708          *
709          * this table maps the upper 5 bits of exponent and mantissa
710          * of the atm forum representation of the rate into an index
711          * on rate grid  
712          */
713
714         rate_atmf = 0;
715         while (rate_atmf < 0x400) {
716                 man = (rate_atmf & 0x1f) << 4;
717                 exp = rate_atmf >> 5;
718
719                 /* 
720                         instead of '/ 512', use '>> 9' to prevent a call
721                         to divdu3 on x86 platforms
722                 */
723                 rate_cps = (unsigned long long) (1 << exp) * (man + 512) >> 9;
724
725                 if (rate_cps < 10)
726                         rate_cps = 10;  /* 2.2.1 minimum payload rate is 10 cps */
727
728                 for (i = 255; i > 0; i--)
729                         if ((*rategrid)[i/16][i%16] >= rate_cps)
730                                 break;   /* pick nearest rate instead? */
731
732                 /*
733                  * each table entry is 16 bits: (rate grid index (8 bits)
734                  * and a buffer limit (8 bits)
735                  * there are two table entries in each 32-bit register
736                  */
737
738 #ifdef notdef
739                 buf = rate_cps * he_dev->tx_numbuffs /
740                                 (he_dev->atm_dev->link_rate * 2);
741 #else
742                 /* this is pretty, but avoids _divdu3 and is mostly correct */
743                 mult = he_dev->atm_dev->link_rate / ATM_OC3_PCR;
744                 if (rate_cps > (272 * mult))
745                         buf = 4;
746                 else if (rate_cps > (204 * mult))
747                         buf = 3;
748                 else if (rate_cps > (136 * mult))
749                         buf = 2;
750                 else if (rate_cps > (68 * mult))
751                         buf = 1;
752                 else
753                         buf = 0;
754 #endif
755                 if (buf > buf_limit)
756                         buf = buf_limit;
757                 reg = (reg << 16) | ((i << 8) | buf);
758
759 #define RTGTBL_OFFSET 0x400
760           
761                 if (rate_atmf & 0x1)
762                         he_writel_rcm(he_dev, reg,
763                                 CONFIG_RCMABR + RTGTBL_OFFSET + (rate_atmf >> 1));
764
765                 ++rate_atmf;
766         }
767
768         kfree(rategrid);
769         return 0;
770 }
771
772 static int he_init_group(struct he_dev *he_dev, int group)
773 {
774         struct he_buff *heb, *next;
775         dma_addr_t mapping;
776         int i;
777
778         he_writel(he_dev, 0x0, G0_RBPS_S + (group * 32));
779         he_writel(he_dev, 0x0, G0_RBPS_T + (group * 32));
780         he_writel(he_dev, 0x0, G0_RBPS_QI + (group * 32));
781         he_writel(he_dev, RBP_THRESH(0x1) | RBP_QSIZE(0x0),
782                   G0_RBPS_BS + (group * 32));
783
784         /* bitmap table */
785         he_dev->rbpl_table = kmalloc(BITS_TO_LONGS(RBPL_TABLE_SIZE)
786                                      * sizeof(unsigned long), GFP_KERNEL);
787         if (!he_dev->rbpl_table) {
788                 hprintk("unable to allocate rbpl bitmap table\n");
789                 return -ENOMEM;
790         }
791         bitmap_zero(he_dev->rbpl_table, RBPL_TABLE_SIZE);
792
793         /* rbpl_virt 64-bit pointers */
794         he_dev->rbpl_virt = kmalloc(RBPL_TABLE_SIZE
795                                     * sizeof(struct he_buff *), GFP_KERNEL);
796         if (!he_dev->rbpl_virt) {
797                 hprintk("unable to allocate rbpl virt table\n");
798                 goto out_free_rbpl_table;
799         }
800
801         /* large buffer pool */
802         he_dev->rbpl_pool = pci_pool_create("rbpl", he_dev->pci_dev,
803                                             CONFIG_RBPL_BUFSIZE, 64, 0);
804         if (he_dev->rbpl_pool == NULL) {
805                 hprintk("unable to create rbpl pool\n");
806                 goto out_free_rbpl_virt;
807         }
808
809         he_dev->rbpl_base = pci_alloc_consistent(he_dev->pci_dev,
810                 CONFIG_RBPL_SIZE * sizeof(struct he_rbp), &he_dev->rbpl_phys);
811         if (he_dev->rbpl_base == NULL) {
812                 hprintk("failed to alloc rbpl_base\n");
813                 goto out_destroy_rbpl_pool;
814         }
815         memset(he_dev->rbpl_base, 0, CONFIG_RBPL_SIZE * sizeof(struct he_rbp));
816
817         INIT_LIST_HEAD(&he_dev->rbpl_outstanding);
818
819         for (i = 0; i < CONFIG_RBPL_SIZE; ++i) {
820
821                 heb = pci_pool_alloc(he_dev->rbpl_pool, GFP_KERNEL|GFP_DMA, &mapping);
822                 if (!heb)
823                         goto out_free_rbpl;
824                 heb->mapping = mapping;
825                 list_add(&heb->entry, &he_dev->rbpl_outstanding);
826
827                 set_bit(i, he_dev->rbpl_table);
828                 he_dev->rbpl_virt[i] = heb;
829                 he_dev->rbpl_hint = i + 1;
830                 he_dev->rbpl_base[i].idx =  i << RBP_IDX_OFFSET;
831                 he_dev->rbpl_base[i].phys = mapping + offsetof(struct he_buff, data);
832         }
833         he_dev->rbpl_tail = &he_dev->rbpl_base[CONFIG_RBPL_SIZE - 1];
834
835         he_writel(he_dev, he_dev->rbpl_phys, G0_RBPL_S + (group * 32));
836         he_writel(he_dev, RBPL_MASK(he_dev->rbpl_tail),
837                                                 G0_RBPL_T + (group * 32));
838         he_writel(he_dev, (CONFIG_RBPL_BUFSIZE - sizeof(struct he_buff))/4,
839                                                 G0_RBPL_BS + (group * 32));
840         he_writel(he_dev,
841                         RBP_THRESH(CONFIG_RBPL_THRESH) |
842                         RBP_QSIZE(CONFIG_RBPL_SIZE - 1) |
843                         RBP_INT_ENB,
844                                                 G0_RBPL_QI + (group * 32));
845
846         /* rx buffer ready queue */
847
848         he_dev->rbrq_base = pci_alloc_consistent(he_dev->pci_dev,
849                 CONFIG_RBRQ_SIZE * sizeof(struct he_rbrq), &he_dev->rbrq_phys);
850         if (he_dev->rbrq_base == NULL) {
851                 hprintk("failed to allocate rbrq\n");
852                 goto out_free_rbpl;
853         }
854         memset(he_dev->rbrq_base, 0, CONFIG_RBRQ_SIZE * sizeof(struct he_rbrq));
855
856         he_dev->rbrq_head = he_dev->rbrq_base;
857         he_writel(he_dev, he_dev->rbrq_phys, G0_RBRQ_ST + (group * 16));
858         he_writel(he_dev, 0, G0_RBRQ_H + (group * 16));
859         he_writel(he_dev,
860                 RBRQ_THRESH(CONFIG_RBRQ_THRESH) | RBRQ_SIZE(CONFIG_RBRQ_SIZE - 1),
861                                                 G0_RBRQ_Q + (group * 16));
862         if (irq_coalesce) {
863                 hprintk("coalescing interrupts\n");
864                 he_writel(he_dev, RBRQ_TIME(768) | RBRQ_COUNT(7),
865                                                 G0_RBRQ_I + (group * 16));
866         } else
867                 he_writel(he_dev, RBRQ_TIME(0) | RBRQ_COUNT(1),
868                                                 G0_RBRQ_I + (group * 16));
869
870         /* tx buffer ready queue */
871
872         he_dev->tbrq_base = pci_alloc_consistent(he_dev->pci_dev,
873                 CONFIG_TBRQ_SIZE * sizeof(struct he_tbrq), &he_dev->tbrq_phys);
874         if (he_dev->tbrq_base == NULL) {
875                 hprintk("failed to allocate tbrq\n");
876                 goto out_free_rbpq_base;
877         }
878         memset(he_dev->tbrq_base, 0, CONFIG_TBRQ_SIZE * sizeof(struct he_tbrq));
879
880         he_dev->tbrq_head = he_dev->tbrq_base;
881
882         he_writel(he_dev, he_dev->tbrq_phys, G0_TBRQ_B_T + (group * 16));
883         he_writel(he_dev, 0, G0_TBRQ_H + (group * 16));
884         he_writel(he_dev, CONFIG_TBRQ_SIZE - 1, G0_TBRQ_S + (group * 16));
885         he_writel(he_dev, CONFIG_TBRQ_THRESH, G0_TBRQ_THRESH + (group * 16));
886
887         return 0;
888
889 out_free_rbpq_base:
890         pci_free_consistent(he_dev->pci_dev, CONFIG_RBRQ_SIZE *
891                         sizeof(struct he_rbrq), he_dev->rbrq_base,
892                         he_dev->rbrq_phys);
893 out_free_rbpl:
894         list_for_each_entry_safe(heb, next, &he_dev->rbpl_outstanding, entry)
895                 pci_pool_free(he_dev->rbpl_pool, heb, heb->mapping);
896
897         pci_free_consistent(he_dev->pci_dev, CONFIG_RBPL_SIZE *
898                         sizeof(struct he_rbp), he_dev->rbpl_base,
899                         he_dev->rbpl_phys);
900 out_destroy_rbpl_pool:
901         pci_pool_destroy(he_dev->rbpl_pool);
902 out_free_rbpl_virt:
903         kfree(he_dev->rbpl_virt);
904 out_free_rbpl_table:
905         kfree(he_dev->rbpl_table);
906
907         return -ENOMEM;
908 }
909
910 static int he_init_irq(struct he_dev *he_dev)
911 {
912         int i;
913
914         /* 2.9.3.5  tail offset for each interrupt queue is located after the
915                     end of the interrupt queue */
916
917         he_dev->irq_base = pci_alloc_consistent(he_dev->pci_dev,
918                         (CONFIG_IRQ_SIZE+1) * sizeof(struct he_irq), &he_dev->irq_phys);
919         if (he_dev->irq_base == NULL) {
920                 hprintk("failed to allocate irq\n");
921                 return -ENOMEM;
922         }
923         he_dev->irq_tailoffset = (unsigned *)
924                                         &he_dev->irq_base[CONFIG_IRQ_SIZE];
925         *he_dev->irq_tailoffset = 0;
926         he_dev->irq_head = he_dev->irq_base;
927         he_dev->irq_tail = he_dev->irq_base;
928
929         for (i = 0; i < CONFIG_IRQ_SIZE; ++i)
930                 he_dev->irq_base[i].isw = ITYPE_INVALID;
931
932         he_writel(he_dev, he_dev->irq_phys, IRQ0_BASE);
933         he_writel(he_dev,
934                 IRQ_SIZE(CONFIG_IRQ_SIZE) | IRQ_THRESH(CONFIG_IRQ_THRESH),
935                                                                 IRQ0_HEAD);
936         he_writel(he_dev, IRQ_INT_A | IRQ_TYPE_LINE, IRQ0_CNTL);
937         he_writel(he_dev, 0x0, IRQ0_DATA);
938
939         he_writel(he_dev, 0x0, IRQ1_BASE);
940         he_writel(he_dev, 0x0, IRQ1_HEAD);
941         he_writel(he_dev, 0x0, IRQ1_CNTL);
942         he_writel(he_dev, 0x0, IRQ1_DATA);
943
944         he_writel(he_dev, 0x0, IRQ2_BASE);
945         he_writel(he_dev, 0x0, IRQ2_HEAD);
946         he_writel(he_dev, 0x0, IRQ2_CNTL);
947         he_writel(he_dev, 0x0, IRQ2_DATA);
948
949         he_writel(he_dev, 0x0, IRQ3_BASE);
950         he_writel(he_dev, 0x0, IRQ3_HEAD);
951         he_writel(he_dev, 0x0, IRQ3_CNTL);
952         he_writel(he_dev, 0x0, IRQ3_DATA);
953
954         /* 2.9.3.2 interrupt queue mapping registers */
955
956         he_writel(he_dev, 0x0, GRP_10_MAP);
957         he_writel(he_dev, 0x0, GRP_32_MAP);
958         he_writel(he_dev, 0x0, GRP_54_MAP);
959         he_writel(he_dev, 0x0, GRP_76_MAP);
960
961         if (request_irq(he_dev->pci_dev->irq,
962                         he_irq_handler, IRQF_SHARED, DEV_LABEL, he_dev)) {
963                 hprintk("irq %d already in use\n", he_dev->pci_dev->irq);
964                 return -EINVAL;
965         }   
966
967         he_dev->irq = he_dev->pci_dev->irq;
968
969         return 0;
970 }
971
972 static int he_start(struct atm_dev *dev)
973 {
974         struct he_dev *he_dev;
975         struct pci_dev *pci_dev;
976         unsigned long membase;
977
978         u16 command;
979         u32 gen_cntl_0, host_cntl, lb_swap;
980         u8 cache_size, timer;
981         
982         unsigned err;
983         unsigned int status, reg;
984         int i, group;
985
986         he_dev = HE_DEV(dev);
987         pci_dev = he_dev->pci_dev;
988
989         membase = pci_resource_start(pci_dev, 0);
990         HPRINTK("membase = 0x%lx  irq = %d.\n", membase, pci_dev->irq);
991
992         /*
993          * pci bus controller initialization 
994          */
995
996         /* 4.3 pci bus controller-specific initialization */
997         if (pci_read_config_dword(pci_dev, GEN_CNTL_0, &gen_cntl_0) != 0) {
998                 hprintk("can't read GEN_CNTL_0\n");
999                 return -EINVAL;
1000         }
1001         gen_cntl_0 |= (MRL_ENB | MRM_ENB | IGNORE_TIMEOUT);
1002         if (pci_write_config_dword(pci_dev, GEN_CNTL_0, gen_cntl_0) != 0) {
1003                 hprintk("can't write GEN_CNTL_0.\n");
1004                 return -EINVAL;
1005         }
1006
1007         if (pci_read_config_word(pci_dev, PCI_COMMAND, &command) != 0) {
1008                 hprintk("can't read PCI_COMMAND.\n");
1009                 return -EINVAL;
1010         }
1011
1012         command |= (PCI_COMMAND_MEMORY | PCI_COMMAND_MASTER | PCI_COMMAND_INVALIDATE);
1013         if (pci_write_config_word(pci_dev, PCI_COMMAND, command) != 0) {
1014                 hprintk("can't enable memory.\n");
1015                 return -EINVAL;
1016         }
1017
1018         if (pci_read_config_byte(pci_dev, PCI_CACHE_LINE_SIZE, &cache_size)) {
1019                 hprintk("can't read cache line size?\n");
1020                 return -EINVAL;
1021         }
1022
1023         if (cache_size < 16) {
1024                 cache_size = 16;
1025                 if (pci_write_config_byte(pci_dev, PCI_CACHE_LINE_SIZE, cache_size))
1026                         hprintk("can't set cache line size to %d\n", cache_size);
1027         }
1028
1029         if (pci_read_config_byte(pci_dev, PCI_LATENCY_TIMER, &timer)) {
1030                 hprintk("can't read latency timer?\n");
1031                 return -EINVAL;
1032         }
1033
1034         /* from table 3.9
1035          *
1036          * LAT_TIMER = 1 + AVG_LAT + BURST_SIZE/BUS_SIZE
1037          * 
1038          * AVG_LAT: The average first data read/write latency [maximum 16 clock cycles]
1039          * BURST_SIZE: 1536 bytes (read) for 622, 768 bytes (read) for 155 [192 clock cycles]
1040          *
1041          */ 
1042 #define LAT_TIMER 209
1043         if (timer < LAT_TIMER) {
1044                 HPRINTK("latency timer was %d, setting to %d\n", timer, LAT_TIMER);
1045                 timer = LAT_TIMER;
1046                 if (pci_write_config_byte(pci_dev, PCI_LATENCY_TIMER, timer))
1047                         hprintk("can't set latency timer to %d\n", timer);
1048         }
1049
1050         if (!(he_dev->membase = ioremap(membase, HE_REGMAP_SIZE))) {
1051                 hprintk("can't set up page mapping\n");
1052                 return -EINVAL;
1053         }
1054
1055         /* 4.4 card reset */
1056         he_writel(he_dev, 0x0, RESET_CNTL);
1057         he_writel(he_dev, 0xff, RESET_CNTL);
1058
1059         udelay(16*1000);        /* 16 ms */
1060         status = he_readl(he_dev, RESET_CNTL);
1061         if ((status & BOARD_RST_STATUS) == 0) {
1062                 hprintk("reset failed\n");
1063                 return -EINVAL;
1064         }
1065
1066         /* 4.5 set bus width */
1067         host_cntl = he_readl(he_dev, HOST_CNTL);
1068         if (host_cntl & PCI_BUS_SIZE64)
1069                 gen_cntl_0 |= ENBL_64;
1070         else
1071                 gen_cntl_0 &= ~ENBL_64;
1072
1073         if (disable64 == 1) {
1074                 hprintk("disabling 64-bit pci bus transfers\n");
1075                 gen_cntl_0 &= ~ENBL_64;
1076         }
1077
1078         if (gen_cntl_0 & ENBL_64)
1079                 hprintk("64-bit transfers enabled\n");
1080
1081         pci_write_config_dword(pci_dev, GEN_CNTL_0, gen_cntl_0);
1082
1083         /* 4.7 read prom contents */
1084         for (i = 0; i < PROD_ID_LEN; ++i)
1085                 he_dev->prod_id[i] = read_prom_byte(he_dev, PROD_ID + i);
1086
1087         he_dev->media = read_prom_byte(he_dev, MEDIA);
1088
1089         for (i = 0; i < 6; ++i)
1090                 dev->esi[i] = read_prom_byte(he_dev, MAC_ADDR + i);
1091
1092         hprintk("%s%s, %x:%x:%x:%x:%x:%x\n",
1093                                 he_dev->prod_id,
1094                                         he_dev->media & 0x40 ? "SM" : "MM",
1095                                                 dev->esi[0],
1096                                                 dev->esi[1],
1097                                                 dev->esi[2],
1098                                                 dev->esi[3],
1099                                                 dev->esi[4],
1100                                                 dev->esi[5]);
1101         he_dev->atm_dev->link_rate = he_is622(he_dev) ?
1102                                                 ATM_OC12_PCR : ATM_OC3_PCR;
1103
1104         /* 4.6 set host endianess */
1105         lb_swap = he_readl(he_dev, LB_SWAP);
1106         if (he_is622(he_dev))
1107                 lb_swap &= ~XFER_SIZE;          /* 4 cells */
1108         else
1109                 lb_swap |= XFER_SIZE;           /* 8 cells */
1110 #ifdef __BIG_ENDIAN
1111         lb_swap |= DESC_WR_SWAP | INTR_SWAP | BIG_ENDIAN_HOST;
1112 #else
1113         lb_swap &= ~(DESC_WR_SWAP | INTR_SWAP | BIG_ENDIAN_HOST |
1114                         DATA_WR_SWAP | DATA_RD_SWAP | DESC_RD_SWAP);
1115 #endif /* __BIG_ENDIAN */
1116         he_writel(he_dev, lb_swap, LB_SWAP);
1117
1118         /* 4.8 sdram controller initialization */
1119         he_writel(he_dev, he_is622(he_dev) ? LB_64_ENB : 0x0, SDRAM_CTL);
1120
1121         /* 4.9 initialize rnum value */
1122         lb_swap |= SWAP_RNUM_MAX(0xf);
1123         he_writel(he_dev, lb_swap, LB_SWAP);
1124
1125         /* 4.10 initialize the interrupt queues */
1126         if ((err = he_init_irq(he_dev)) != 0)
1127                 return err;
1128
1129         /* 4.11 enable pci bus controller state machines */
1130         host_cntl |= (OUTFF_ENB | CMDFF_ENB |
1131                                 QUICK_RD_RETRY | QUICK_WR_RETRY | PERR_INT_ENB);
1132         he_writel(he_dev, host_cntl, HOST_CNTL);
1133
1134         gen_cntl_0 |= INT_PROC_ENBL|INIT_ENB;
1135         pci_write_config_dword(pci_dev, GEN_CNTL_0, gen_cntl_0);
1136
1137         /*
1138          * atm network controller initialization
1139          */
1140
1141         /* 5.1.1 generic configuration state */
1142
1143         /*
1144          *              local (cell) buffer memory map
1145          *                    
1146          *             HE155                          HE622
1147          *                                                      
1148          *        0 ____________1023 bytes  0 _______________________2047 bytes
1149          *         |            |            |                   |   |
1150          *         |  utility   |            |        rx0        |   |
1151          *        5|____________|         255|___________________| u |
1152          *        6|            |         256|                   | t |
1153          *         |            |            |                   | i |
1154          *         |    rx0     |     row    |        tx         | l |
1155          *         |            |            |                   | i |
1156          *         |            |         767|___________________| t |
1157          *      517|____________|         768|                   | y |
1158          * row  518|            |            |        rx1        |   |
1159          *         |            |        1023|___________________|___|
1160          *         |            |
1161          *         |    tx      |
1162          *         |            |
1163          *         |            |
1164          *     1535|____________|
1165          *     1536|            |
1166          *         |    rx1     |
1167          *     2047|____________|
1168          *
1169          */
1170
1171         /* total 4096 connections */
1172         he_dev->vcibits = CONFIG_DEFAULT_VCIBITS;
1173         he_dev->vpibits = CONFIG_DEFAULT_VPIBITS;
1174
1175         if (nvpibits != -1 && nvcibits != -1 && nvpibits+nvcibits != HE_MAXCIDBITS) {
1176                 hprintk("nvpibits + nvcibits != %d\n", HE_MAXCIDBITS);
1177                 return -ENODEV;
1178         }
1179
1180         if (nvpibits != -1) {
1181                 he_dev->vpibits = nvpibits;
1182                 he_dev->vcibits = HE_MAXCIDBITS - nvpibits;
1183         }
1184
1185         if (nvcibits != -1) {
1186                 he_dev->vcibits = nvcibits;
1187                 he_dev->vpibits = HE_MAXCIDBITS - nvcibits;
1188         }
1189
1190
1191         if (he_is622(he_dev)) {
1192                 he_dev->cells_per_row = 40;
1193                 he_dev->bytes_per_row = 2048;
1194                 he_dev->r0_numrows = 256;
1195                 he_dev->tx_numrows = 512;
1196                 he_dev->r1_numrows = 256;
1197                 he_dev->r0_startrow = 0;
1198                 he_dev->tx_startrow = 256;
1199                 he_dev->r1_startrow = 768;
1200         } else {
1201                 he_dev->cells_per_row = 20;
1202                 he_dev->bytes_per_row = 1024;
1203                 he_dev->r0_numrows = 512;
1204                 he_dev->tx_numrows = 1018;
1205                 he_dev->r1_numrows = 512;
1206                 he_dev->r0_startrow = 6;
1207                 he_dev->tx_startrow = 518;
1208                 he_dev->r1_startrow = 1536;
1209         }
1210
1211         he_dev->cells_per_lbuf = 4;
1212         he_dev->buffer_limit = 4;
1213         he_dev->r0_numbuffs = he_dev->r0_numrows *
1214                                 he_dev->cells_per_row / he_dev->cells_per_lbuf;
1215         if (he_dev->r0_numbuffs > 2560)
1216                 he_dev->r0_numbuffs = 2560;
1217
1218         he_dev->r1_numbuffs = he_dev->r1_numrows *
1219                                 he_dev->cells_per_row / he_dev->cells_per_lbuf;
1220         if (he_dev->r1_numbuffs > 2560)
1221                 he_dev->r1_numbuffs = 2560;
1222
1223         he_dev->tx_numbuffs = he_dev->tx_numrows *
1224                                 he_dev->cells_per_row / he_dev->cells_per_lbuf;
1225         if (he_dev->tx_numbuffs > 5120)
1226                 he_dev->tx_numbuffs = 5120;
1227
1228         /* 5.1.2 configure hardware dependent registers */
1229
1230         he_writel(he_dev, 
1231                 SLICE_X(0x2) | ARB_RNUM_MAX(0xf) | TH_PRTY(0x3) |
1232                 RH_PRTY(0x3) | TL_PRTY(0x2) | RL_PRTY(0x1) |
1233                 (he_is622(he_dev) ? BUS_MULTI(0x28) : BUS_MULTI(0x46)) |
1234                 (he_is622(he_dev) ? NET_PREF(0x50) : NET_PREF(0x8c)),
1235                                                                 LBARB);
1236
1237         he_writel(he_dev, BANK_ON |
1238                 (he_is622(he_dev) ? (REF_RATE(0x384) | WIDE_DATA) : REF_RATE(0x150)),
1239                                                                 SDRAMCON);
1240
1241         he_writel(he_dev,
1242                 (he_is622(he_dev) ? RM_BANK_WAIT(1) : RM_BANK_WAIT(0)) |
1243                                                 RM_RW_WAIT(1), RCMCONFIG);
1244         he_writel(he_dev,
1245                 (he_is622(he_dev) ? TM_BANK_WAIT(2) : TM_BANK_WAIT(1)) |
1246                                                 TM_RW_WAIT(1), TCMCONFIG);
1247
1248         he_writel(he_dev, he_dev->cells_per_lbuf * ATM_CELL_PAYLOAD, LB_CONFIG);
1249
1250         he_writel(he_dev, 
1251                 (he_is622(he_dev) ? UT_RD_DELAY(8) : UT_RD_DELAY(0)) |
1252                 (he_is622(he_dev) ? RC_UT_MODE(0) : RC_UT_MODE(1)) |
1253                 RX_VALVP(he_dev->vpibits) |
1254                 RX_VALVC(he_dev->vcibits),                       RC_CONFIG);
1255
1256         he_writel(he_dev, DRF_THRESH(0x20) |
1257                 (he_is622(he_dev) ? TX_UT_MODE(0) : TX_UT_MODE(1)) |
1258                 TX_VCI_MASK(he_dev->vcibits) |
1259                 LBFREE_CNT(he_dev->tx_numbuffs),                TX_CONFIG);
1260
1261         he_writel(he_dev, 0x0, TXAAL5_PROTO);
1262
1263         he_writel(he_dev, PHY_INT_ENB |
1264                 (he_is622(he_dev) ? PTMR_PRE(67 - 1) : PTMR_PRE(50 - 1)),
1265                                                                 RH_CONFIG);
1266
1267         /* 5.1.3 initialize connection memory */
1268
1269         for (i = 0; i < TCM_MEM_SIZE; ++i)
1270                 he_writel_tcm(he_dev, 0, i);
1271
1272         for (i = 0; i < RCM_MEM_SIZE; ++i)
1273                 he_writel_rcm(he_dev, 0, i);
1274
1275         /*
1276          *      transmit connection memory map
1277          *
1278          *                  tx memory
1279          *          0x0 ___________________
1280          *             |                   |
1281          *             |                   |
1282          *             |       TSRa        |
1283          *             |                   |
1284          *             |                   |
1285          *       0x8000|___________________|
1286          *             |                   |
1287          *             |       TSRb        |
1288          *       0xc000|___________________|
1289          *             |                   |
1290          *             |       TSRc        |
1291          *       0xe000|___________________|
1292          *             |       TSRd        |
1293          *       0xf000|___________________|
1294          *             |       tmABR       |
1295          *      0x10000|___________________|
1296          *             |                   |
1297          *             |       tmTPD       |
1298          *             |___________________|
1299          *             |                   |
1300          *                      ....
1301          *      0x1ffff|___________________|
1302          *
1303          *
1304          */
1305
1306         he_writel(he_dev, CONFIG_TSRB, TSRB_BA);
1307         he_writel(he_dev, CONFIG_TSRC, TSRC_BA);
1308         he_writel(he_dev, CONFIG_TSRD, TSRD_BA);
1309         he_writel(he_dev, CONFIG_TMABR, TMABR_BA);
1310         he_writel(he_dev, CONFIG_TPDBA, TPD_BA);
1311
1312
1313         /*
1314          *      receive connection memory map
1315          *
1316          *          0x0 ___________________
1317          *             |                   |
1318          *             |                   |
1319          *             |       RSRa        |
1320          *             |                   |
1321          *             |                   |
1322          *       0x8000|___________________|
1323          *             |                   |
1324          *             |             rx0/1 |
1325          *             |       LBM         |   link lists of local
1326          *             |             tx    |   buffer memory 
1327          *             |                   |
1328          *       0xd000|___________________|
1329          *             |                   |
1330          *             |      rmABR        |
1331          *       0xe000|___________________|
1332          *             |                   |
1333          *             |       RSRb        |
1334          *             |___________________|
1335          *             |                   |
1336          *                      ....
1337          *       0xffff|___________________|
1338          */
1339
1340         he_writel(he_dev, 0x08000, RCMLBM_BA);
1341         he_writel(he_dev, 0x0e000, RCMRSRB_BA);
1342         he_writel(he_dev, 0x0d800, RCMABR_BA);
1343
1344         /* 5.1.4 initialize local buffer free pools linked lists */
1345
1346         he_init_rx_lbfp0(he_dev);
1347         he_init_rx_lbfp1(he_dev);
1348
1349         he_writel(he_dev, 0x0, RLBC_H);
1350         he_writel(he_dev, 0x0, RLBC_T);
1351         he_writel(he_dev, 0x0, RLBC_H2);
1352
1353         he_writel(he_dev, 512, RXTHRSH);        /* 10% of r0+r1 buffers */
1354         he_writel(he_dev, 256, LITHRSH);        /* 5% of r0+r1 buffers */
1355
1356         he_init_tx_lbfp(he_dev);
1357
1358         he_writel(he_dev, he_is622(he_dev) ? 0x104780 : 0x800, UBUFF_BA);
1359
1360         /* 5.1.5 initialize intermediate receive queues */
1361
1362         if (he_is622(he_dev)) {
1363                 he_writel(he_dev, 0x000f, G0_INMQ_S);
1364                 he_writel(he_dev, 0x200f, G0_INMQ_L);
1365
1366                 he_writel(he_dev, 0x001f, G1_INMQ_S);
1367                 he_writel(he_dev, 0x201f, G1_INMQ_L);
1368
1369                 he_writel(he_dev, 0x002f, G2_INMQ_S);
1370                 he_writel(he_dev, 0x202f, G2_INMQ_L);
1371
1372                 he_writel(he_dev, 0x003f, G3_INMQ_S);
1373                 he_writel(he_dev, 0x203f, G3_INMQ_L);
1374
1375                 he_writel(he_dev, 0x004f, G4_INMQ_S);
1376                 he_writel(he_dev, 0x204f, G4_INMQ_L);
1377
1378                 he_writel(he_dev, 0x005f, G5_INMQ_S);
1379                 he_writel(he_dev, 0x205f, G5_INMQ_L);
1380
1381                 he_writel(he_dev, 0x006f, G6_INMQ_S);
1382                 he_writel(he_dev, 0x206f, G6_INMQ_L);
1383
1384                 he_writel(he_dev, 0x007f, G7_INMQ_S);
1385                 he_writel(he_dev, 0x207f, G7_INMQ_L);
1386         } else {
1387                 he_writel(he_dev, 0x0000, G0_INMQ_S);
1388                 he_writel(he_dev, 0x0008, G0_INMQ_L);
1389
1390                 he_writel(he_dev, 0x0001, G1_INMQ_S);
1391                 he_writel(he_dev, 0x0009, G1_INMQ_L);
1392
1393                 he_writel(he_dev, 0x0002, G2_INMQ_S);
1394                 he_writel(he_dev, 0x000a, G2_INMQ_L);
1395
1396                 he_writel(he_dev, 0x0003, G3_INMQ_S);
1397                 he_writel(he_dev, 0x000b, G3_INMQ_L);
1398
1399                 he_writel(he_dev, 0x0004, G4_INMQ_S);
1400                 he_writel(he_dev, 0x000c, G4_INMQ_L);
1401
1402                 he_writel(he_dev, 0x0005, G5_INMQ_S);
1403                 he_writel(he_dev, 0x000d, G5_INMQ_L);
1404
1405                 he_writel(he_dev, 0x0006, G6_INMQ_S);
1406                 he_writel(he_dev, 0x000e, G6_INMQ_L);
1407
1408                 he_writel(he_dev, 0x0007, G7_INMQ_S);
1409                 he_writel(he_dev, 0x000f, G7_INMQ_L);
1410         }
1411
1412         /* 5.1.6 application tunable parameters */
1413
1414         he_writel(he_dev, 0x0, MCC);
1415         he_writel(he_dev, 0x0, OEC);
1416         he_writel(he_dev, 0x0, DCC);
1417         he_writel(he_dev, 0x0, CEC);
1418         
1419         /* 5.1.7 cs block initialization */
1420
1421         he_init_cs_block(he_dev);
1422
1423         /* 5.1.8 cs block connection memory initialization */
1424         
1425         if (he_init_cs_block_rcm(he_dev) < 0)
1426                 return -ENOMEM;
1427
1428         /* 5.1.10 initialize host structures */
1429
1430         he_init_tpdrq(he_dev);
1431
1432         he_dev->tpd_pool = pci_pool_create("tpd", he_dev->pci_dev,
1433                 sizeof(struct he_tpd), TPD_ALIGNMENT, 0);
1434         if (he_dev->tpd_pool == NULL) {
1435                 hprintk("unable to create tpd pci_pool\n");
1436                 return -ENOMEM;         
1437         }
1438
1439         INIT_LIST_HEAD(&he_dev->outstanding_tpds);
1440
1441         if (he_init_group(he_dev, 0) != 0)
1442                 return -ENOMEM;
1443
1444         for (group = 1; group < HE_NUM_GROUPS; ++group) {
1445                 he_writel(he_dev, 0x0, G0_RBPS_S + (group * 32));
1446                 he_writel(he_dev, 0x0, G0_RBPS_T + (group * 32));
1447                 he_writel(he_dev, 0x0, G0_RBPS_QI + (group * 32));
1448                 he_writel(he_dev, RBP_THRESH(0x1) | RBP_QSIZE(0x0),
1449                                                 G0_RBPS_BS + (group * 32));
1450
1451                 he_writel(he_dev, 0x0, G0_RBPL_S + (group * 32));
1452                 he_writel(he_dev, 0x0, G0_RBPL_T + (group * 32));
1453                 he_writel(he_dev, RBP_THRESH(0x1) | RBP_QSIZE(0x0),
1454                                                 G0_RBPL_QI + (group * 32));
1455                 he_writel(he_dev, 0x0, G0_RBPL_BS + (group * 32));
1456
1457                 he_writel(he_dev, 0x0, G0_RBRQ_ST + (group * 16));
1458                 he_writel(he_dev, 0x0, G0_RBRQ_H + (group * 16));
1459                 he_writel(he_dev, RBRQ_THRESH(0x1) | RBRQ_SIZE(0x0),
1460                                                 G0_RBRQ_Q + (group * 16));
1461                 he_writel(he_dev, 0x0, G0_RBRQ_I + (group * 16));
1462
1463                 he_writel(he_dev, 0x0, G0_TBRQ_B_T + (group * 16));
1464                 he_writel(he_dev, 0x0, G0_TBRQ_H + (group * 16));
1465                 he_writel(he_dev, TBRQ_THRESH(0x1),
1466                                                 G0_TBRQ_THRESH + (group * 16));
1467                 he_writel(he_dev, 0x0, G0_TBRQ_S + (group * 16));
1468         }
1469
1470         /* host status page */
1471
1472         he_dev->hsp = pci_alloc_consistent(he_dev->pci_dev,
1473                                 sizeof(struct he_hsp), &he_dev->hsp_phys);
1474         if (he_dev->hsp == NULL) {
1475                 hprintk("failed to allocate host status page\n");
1476                 return -ENOMEM;
1477         }
1478         memset(he_dev->hsp, 0, sizeof(struct he_hsp));
1479         he_writel(he_dev, he_dev->hsp_phys, HSP_BA);
1480
1481         /* initialize framer */
1482
1483 #ifdef CONFIG_ATM_HE_USE_SUNI
1484         if (he_isMM(he_dev))
1485                 suni_init(he_dev->atm_dev);
1486         if (he_dev->atm_dev->phy && he_dev->atm_dev->phy->start)
1487                 he_dev->atm_dev->phy->start(he_dev->atm_dev);
1488 #endif /* CONFIG_ATM_HE_USE_SUNI */
1489
1490         if (sdh) {
1491                 /* this really should be in suni.c but for now... */
1492                 int val;
1493
1494                 val = he_phy_get(he_dev->atm_dev, SUNI_TPOP_APM);
1495                 val = (val & ~SUNI_TPOP_APM_S) | (SUNI_TPOP_S_SDH << SUNI_TPOP_APM_S_SHIFT);
1496                 he_phy_put(he_dev->atm_dev, val, SUNI_TPOP_APM);
1497                 he_phy_put(he_dev->atm_dev, SUNI_TACP_IUCHP_CLP, SUNI_TACP_IUCHP);
1498         }
1499
1500         /* 5.1.12 enable transmit and receive */
1501
1502         reg = he_readl_mbox(he_dev, CS_ERCTL0);
1503         reg |= TX_ENABLE|ER_ENABLE;
1504         he_writel_mbox(he_dev, reg, CS_ERCTL0);
1505
1506         reg = he_readl(he_dev, RC_CONFIG);
1507         reg |= RX_ENABLE;
1508         he_writel(he_dev, reg, RC_CONFIG);
1509
1510         for (i = 0; i < HE_NUM_CS_STPER; ++i) {
1511                 he_dev->cs_stper[i].inuse = 0;
1512                 he_dev->cs_stper[i].pcr = -1;
1513         }
1514         he_dev->total_bw = 0;
1515
1516
1517         /* atm linux initialization */
1518
1519         he_dev->atm_dev->ci_range.vpi_bits = he_dev->vpibits;
1520         he_dev->atm_dev->ci_range.vci_bits = he_dev->vcibits;
1521
1522         he_dev->irq_peak = 0;
1523         he_dev->rbrq_peak = 0;
1524         he_dev->rbpl_peak = 0;
1525         he_dev->tbrq_peak = 0;
1526
1527         HPRINTK("hell bent for leather!\n");
1528
1529         return 0;
1530 }
1531
1532 static void
1533 he_stop(struct he_dev *he_dev)
1534 {
1535         struct he_buff *heb, *next;
1536         struct pci_dev *pci_dev;
1537         u32 gen_cntl_0, reg;
1538         u16 command;
1539
1540         pci_dev = he_dev->pci_dev;
1541
1542         /* disable interrupts */
1543
1544         if (he_dev->membase) {
1545                 pci_read_config_dword(pci_dev, GEN_CNTL_0, &gen_cntl_0);
1546                 gen_cntl_0 &= ~(INT_PROC_ENBL | INIT_ENB);
1547                 pci_write_config_dword(pci_dev, GEN_CNTL_0, gen_cntl_0);
1548
1549                 tasklet_disable(&he_dev->tasklet);
1550
1551                 /* disable recv and transmit */
1552
1553                 reg = he_readl_mbox(he_dev, CS_ERCTL0);
1554                 reg &= ~(TX_ENABLE|ER_ENABLE);
1555                 he_writel_mbox(he_dev, reg, CS_ERCTL0);
1556
1557                 reg = he_readl(he_dev, RC_CONFIG);
1558                 reg &= ~(RX_ENABLE);
1559                 he_writel(he_dev, reg, RC_CONFIG);
1560         }
1561
1562 #ifdef CONFIG_ATM_HE_USE_SUNI
1563         if (he_dev->atm_dev->phy && he_dev->atm_dev->phy->stop)
1564                 he_dev->atm_dev->phy->stop(he_dev->atm_dev);
1565 #endif /* CONFIG_ATM_HE_USE_SUNI */
1566
1567         if (he_dev->irq)
1568                 free_irq(he_dev->irq, he_dev);
1569
1570         if (he_dev->irq_base)
1571                 pci_free_consistent(he_dev->pci_dev, (CONFIG_IRQ_SIZE+1)
1572                         * sizeof(struct he_irq), he_dev->irq_base, he_dev->irq_phys);
1573
1574         if (he_dev->hsp)
1575                 pci_free_consistent(he_dev->pci_dev, sizeof(struct he_hsp),
1576                                                 he_dev->hsp, he_dev->hsp_phys);
1577
1578         if (he_dev->rbpl_base) {
1579                 list_for_each_entry_safe(heb, next, &he_dev->rbpl_outstanding, entry)
1580                         pci_pool_free(he_dev->rbpl_pool, heb, heb->mapping);
1581
1582                 pci_free_consistent(he_dev->pci_dev, CONFIG_RBPL_SIZE
1583                         * sizeof(struct he_rbp), he_dev->rbpl_base, he_dev->rbpl_phys);
1584         }
1585
1586         kfree(he_dev->rbpl_virt);
1587         kfree(he_dev->rbpl_table);
1588
1589         if (he_dev->rbpl_pool)
1590                 pci_pool_destroy(he_dev->rbpl_pool);
1591
1592         if (he_dev->rbrq_base)
1593                 pci_free_consistent(he_dev->pci_dev, CONFIG_RBRQ_SIZE * sizeof(struct he_rbrq),
1594                                                         he_dev->rbrq_base, he_dev->rbrq_phys);
1595
1596         if (he_dev->tbrq_base)
1597                 pci_free_consistent(he_dev->pci_dev, CONFIG_TBRQ_SIZE * sizeof(struct he_tbrq),
1598                                                         he_dev->tbrq_base, he_dev->tbrq_phys);
1599
1600         if (he_dev->tpdrq_base)
1601                 pci_free_consistent(he_dev->pci_dev, CONFIG_TBRQ_SIZE * sizeof(struct he_tbrq),
1602                                                         he_dev->tpdrq_base, he_dev->tpdrq_phys);
1603
1604         if (he_dev->tpd_pool)
1605                 pci_pool_destroy(he_dev->tpd_pool);
1606
1607         if (he_dev->pci_dev) {
1608                 pci_read_config_word(he_dev->pci_dev, PCI_COMMAND, &command);
1609                 command &= ~(PCI_COMMAND_MEMORY | PCI_COMMAND_MASTER);
1610                 pci_write_config_word(he_dev->pci_dev, PCI_COMMAND, command);
1611         }
1612         
1613         if (he_dev->membase)
1614                 iounmap(he_dev->membase);
1615 }
1616
1617 static struct he_tpd *
1618 __alloc_tpd(struct he_dev *he_dev)
1619 {
1620         struct he_tpd *tpd;
1621         dma_addr_t mapping;
1622
1623         tpd = pci_pool_alloc(he_dev->tpd_pool, GFP_ATOMIC|GFP_DMA, &mapping);
1624         if (tpd == NULL)
1625                 return NULL;
1626                         
1627         tpd->status = TPD_ADDR(mapping);
1628         tpd->reserved = 0; 
1629         tpd->iovec[0].addr = 0; tpd->iovec[0].len = 0;
1630         tpd->iovec[1].addr = 0; tpd->iovec[1].len = 0;
1631         tpd->iovec[2].addr = 0; tpd->iovec[2].len = 0;
1632
1633         return tpd;
1634 }
1635
1636 #define AAL5_LEN(buf,len)                                               \
1637                         ((((unsigned char *)(buf))[(len)-6] << 8) |     \
1638                                 (((unsigned char *)(buf))[(len)-5]))
1639
1640 /* 2.10.1.2 receive
1641  *
1642  * aal5 packets can optionally return the tcp checksum in the lower
1643  * 16 bits of the crc (RSR0_TCP_CKSUM)
1644  */
1645
1646 #define TCP_CKSUM(buf,len)                                              \
1647                         ((((unsigned char *)(buf))[(len)-2] << 8) |     \
1648                                 (((unsigned char *)(buf))[(len-1)]))
1649
1650 static int
1651 he_service_rbrq(struct he_dev *he_dev, int group)
1652 {
1653         struct he_rbrq *rbrq_tail = (struct he_rbrq *)
1654                                 ((unsigned long)he_dev->rbrq_base |
1655                                         he_dev->hsp->group[group].rbrq_tail);
1656         unsigned cid, lastcid = -1;
1657         struct sk_buff *skb;
1658         struct atm_vcc *vcc = NULL;
1659         struct he_vcc *he_vcc;
1660         struct he_buff *heb, *next;
1661         int i;
1662         int pdus_assembled = 0;
1663         int updated = 0;
1664
1665         read_lock(&vcc_sklist_lock);
1666         while (he_dev->rbrq_head != rbrq_tail) {
1667                 ++updated;
1668
1669                 HPRINTK("%p rbrq%d 0x%x len=%d cid=0x%x %s%s%s%s%s%s\n",
1670                         he_dev->rbrq_head, group,
1671                         RBRQ_ADDR(he_dev->rbrq_head),
1672                         RBRQ_BUFLEN(he_dev->rbrq_head),
1673                         RBRQ_CID(he_dev->rbrq_head),
1674                         RBRQ_CRC_ERR(he_dev->rbrq_head) ? " CRC_ERR" : "",
1675                         RBRQ_LEN_ERR(he_dev->rbrq_head) ? " LEN_ERR" : "",
1676                         RBRQ_END_PDU(he_dev->rbrq_head) ? " END_PDU" : "",
1677                         RBRQ_AAL5_PROT(he_dev->rbrq_head) ? " AAL5_PROT" : "",
1678                         RBRQ_CON_CLOSED(he_dev->rbrq_head) ? " CON_CLOSED" : "",
1679                         RBRQ_HBUF_ERR(he_dev->rbrq_head) ? " HBUF_ERR" : "");
1680
1681                 i = RBRQ_ADDR(he_dev->rbrq_head) >> RBP_IDX_OFFSET;
1682                 heb = he_dev->rbpl_virt[i];
1683
1684                 cid = RBRQ_CID(he_dev->rbrq_head);
1685                 if (cid != lastcid)
1686                         vcc = __find_vcc(he_dev, cid);
1687                 lastcid = cid;
1688
1689                 if (vcc == NULL || (he_vcc = HE_VCC(vcc)) == NULL) {
1690                         hprintk("vcc/he_vcc == NULL  (cid 0x%x)\n", cid);
1691                         if (!RBRQ_HBUF_ERR(he_dev->rbrq_head)) {
1692                                 clear_bit(i, he_dev->rbpl_table);
1693                                 list_del(&heb->entry);
1694                                 pci_pool_free(he_dev->rbpl_pool, heb, heb->mapping);
1695                         }
1696                                         
1697                         goto next_rbrq_entry;
1698                 }
1699
1700                 if (RBRQ_HBUF_ERR(he_dev->rbrq_head)) {
1701                         hprintk("HBUF_ERR!  (cid 0x%x)\n", cid);
1702                                 atomic_inc(&vcc->stats->rx_drop);
1703                         goto return_host_buffers;
1704                 }
1705
1706                 heb->len = RBRQ_BUFLEN(he_dev->rbrq_head) * 4;
1707                 clear_bit(i, he_dev->rbpl_table);
1708                 list_move_tail(&heb->entry, &he_vcc->buffers);
1709                 he_vcc->pdu_len += heb->len;
1710
1711                 if (RBRQ_CON_CLOSED(he_dev->rbrq_head)) {
1712                         lastcid = -1;
1713                         HPRINTK("wake_up rx_waitq  (cid 0x%x)\n", cid);
1714                         wake_up(&he_vcc->rx_waitq);
1715                         goto return_host_buffers;
1716                 }
1717
1718                 if (!RBRQ_END_PDU(he_dev->rbrq_head))
1719                         goto next_rbrq_entry;
1720
1721                 if (RBRQ_LEN_ERR(he_dev->rbrq_head)
1722                                 || RBRQ_CRC_ERR(he_dev->rbrq_head)) {
1723                         HPRINTK("%s%s (%d.%d)\n",
1724                                 RBRQ_CRC_ERR(he_dev->rbrq_head)
1725                                                         ? "CRC_ERR " : "",
1726                                 RBRQ_LEN_ERR(he_dev->rbrq_head)
1727                                                         ? "LEN_ERR" : "",
1728                                                         vcc->vpi, vcc->vci);
1729                         atomic_inc(&vcc->stats->rx_err);
1730                         goto return_host_buffers;
1731                 }
1732
1733                 skb = atm_alloc_charge(vcc, he_vcc->pdu_len + rx_skb_reserve,
1734                                                         GFP_ATOMIC);
1735                 if (!skb) {
1736                         HPRINTK("charge failed (%d.%d)\n", vcc->vpi, vcc->vci);
1737                         goto return_host_buffers;
1738                 }
1739
1740                 if (rx_skb_reserve > 0)
1741                         skb_reserve(skb, rx_skb_reserve);
1742
1743                 __net_timestamp(skb);
1744
1745                 list_for_each_entry(heb, &he_vcc->buffers, entry)
1746                         memcpy(skb_put(skb, heb->len), &heb->data, heb->len);
1747
1748                 switch (vcc->qos.aal) {
1749                         case ATM_AAL0:
1750                                 /* 2.10.1.5 raw cell receive */
1751                                 skb->len = ATM_AAL0_SDU;
1752                                 skb_set_tail_pointer(skb, skb->len);
1753                                 break;
1754                         case ATM_AAL5:
1755                                 /* 2.10.1.2 aal5 receive */
1756
1757                                 skb->len = AAL5_LEN(skb->data, he_vcc->pdu_len);
1758                                 skb_set_tail_pointer(skb, skb->len);
1759 #ifdef USE_CHECKSUM_HW
1760                                 if (vcc->vpi == 0 && vcc->vci >= ATM_NOT_RSV_VCI) {
1761                                         skb->ip_summed = CHECKSUM_COMPLETE;
1762                                         skb->csum = TCP_CKSUM(skb->data,
1763                                                         he_vcc->pdu_len);
1764                                 }
1765 #endif
1766                                 break;
1767                 }
1768
1769 #ifdef should_never_happen
1770                 if (skb->len > vcc->qos.rxtp.max_sdu)
1771                         hprintk("pdu_len (%d) > vcc->qos.rxtp.max_sdu (%d)!  cid 0x%x\n", skb->len, vcc->qos.rxtp.max_sdu, cid);
1772 #endif
1773
1774 #ifdef notdef
1775                 ATM_SKB(skb)->vcc = vcc;
1776 #endif
1777                 spin_unlock(&he_dev->global_lock);
1778                 vcc->push(vcc, skb);
1779                 spin_lock(&he_dev->global_lock);
1780
1781                 atomic_inc(&vcc->stats->rx);
1782
1783 return_host_buffers:
1784                 ++pdus_assembled;
1785
1786                 list_for_each_entry_safe(heb, next, &he_vcc->buffers, entry)
1787                         pci_pool_free(he_dev->rbpl_pool, heb, heb->mapping);
1788                 INIT_LIST_HEAD(&he_vcc->buffers);
1789                 he_vcc->pdu_len = 0;
1790
1791 next_rbrq_entry:
1792                 he_dev->rbrq_head = (struct he_rbrq *)
1793                                 ((unsigned long) he_dev->rbrq_base |
1794                                         RBRQ_MASK(he_dev->rbrq_head + 1));
1795
1796         }
1797         read_unlock(&vcc_sklist_lock);
1798
1799         if (updated) {
1800                 if (updated > he_dev->rbrq_peak)
1801                         he_dev->rbrq_peak = updated;
1802
1803                 he_writel(he_dev, RBRQ_MASK(he_dev->rbrq_head),
1804                                                 G0_RBRQ_H + (group * 16));
1805         }
1806
1807         return pdus_assembled;
1808 }
1809
1810 static void
1811 he_service_tbrq(struct he_dev *he_dev, int group)
1812 {
1813         struct he_tbrq *tbrq_tail = (struct he_tbrq *)
1814                                 ((unsigned long)he_dev->tbrq_base |
1815                                         he_dev->hsp->group[group].tbrq_tail);
1816         struct he_tpd *tpd;
1817         int slot, updated = 0;
1818         struct he_tpd *__tpd;
1819
1820         /* 2.1.6 transmit buffer return queue */
1821
1822         while (he_dev->tbrq_head != tbrq_tail) {
1823                 ++updated;
1824
1825                 HPRINTK("tbrq%d 0x%x%s%s\n",
1826                         group,
1827                         TBRQ_TPD(he_dev->tbrq_head), 
1828                         TBRQ_EOS(he_dev->tbrq_head) ? " EOS" : "",
1829                         TBRQ_MULTIPLE(he_dev->tbrq_head) ? " MULTIPLE" : "");
1830                 tpd = NULL;
1831                 list_for_each_entry(__tpd, &he_dev->outstanding_tpds, entry) {
1832                         if (TPD_ADDR(__tpd->status) == TBRQ_TPD(he_dev->tbrq_head)) {
1833                                 tpd = __tpd;
1834                                 list_del(&__tpd->entry);
1835                                 break;
1836                         }
1837                 }
1838
1839                 if (tpd == NULL) {
1840                         hprintk("unable to locate tpd for dma buffer %x\n",
1841                                                 TBRQ_TPD(he_dev->tbrq_head));
1842                         goto next_tbrq_entry;
1843                 }
1844
1845                 if (TBRQ_EOS(he_dev->tbrq_head)) {
1846                         HPRINTK("wake_up(tx_waitq) cid 0x%x\n",
1847                                 he_mkcid(he_dev, tpd->vcc->vpi, tpd->vcc->vci));
1848                         if (tpd->vcc)
1849                                 wake_up(&HE_VCC(tpd->vcc)->tx_waitq);
1850
1851                         goto next_tbrq_entry;
1852                 }
1853
1854                 for (slot = 0; slot < TPD_MAXIOV; ++slot) {
1855                         if (tpd->iovec[slot].addr)
1856                                 pci_unmap_single(he_dev->pci_dev,
1857                                         tpd->iovec[slot].addr,
1858                                         tpd->iovec[slot].len & TPD_LEN_MASK,
1859                                                         PCI_DMA_TODEVICE);
1860                         if (tpd->iovec[slot].len & TPD_LST)
1861                                 break;
1862                                 
1863                 }
1864
1865                 if (tpd->skb) { /* && !TBRQ_MULTIPLE(he_dev->tbrq_head) */
1866                         if (tpd->vcc && tpd->vcc->pop)
1867                                 tpd->vcc->pop(tpd->vcc, tpd->skb);
1868                         else
1869                                 dev_kfree_skb_any(tpd->skb);
1870                 }
1871
1872 next_tbrq_entry:
1873                 if (tpd)
1874                         pci_pool_free(he_dev->tpd_pool, tpd, TPD_ADDR(tpd->status));
1875                 he_dev->tbrq_head = (struct he_tbrq *)
1876                                 ((unsigned long) he_dev->tbrq_base |
1877                                         TBRQ_MASK(he_dev->tbrq_head + 1));
1878         }
1879
1880         if (updated) {
1881                 if (updated > he_dev->tbrq_peak)
1882                         he_dev->tbrq_peak = updated;
1883
1884                 he_writel(he_dev, TBRQ_MASK(he_dev->tbrq_head),
1885                                                 G0_TBRQ_H + (group * 16));
1886         }
1887 }
1888
1889 static void
1890 he_service_rbpl(struct he_dev *he_dev, int group)
1891 {
1892         struct he_rbp *new_tail;
1893         struct he_rbp *rbpl_head;
1894         struct he_buff *heb;
1895         dma_addr_t mapping;
1896         int i;
1897         int moved = 0;
1898
1899         rbpl_head = (struct he_rbp *) ((unsigned long)he_dev->rbpl_base |
1900                                         RBPL_MASK(he_readl(he_dev, G0_RBPL_S)));
1901
1902         for (;;) {
1903                 new_tail = (struct he_rbp *) ((unsigned long)he_dev->rbpl_base |
1904                                                 RBPL_MASK(he_dev->rbpl_tail+1));
1905
1906                 /* table 3.42 -- rbpl_tail should never be set to rbpl_head */
1907                 if (new_tail == rbpl_head)
1908                         break;
1909
1910                 i = find_next_zero_bit(he_dev->rbpl_table, RBPL_TABLE_SIZE, he_dev->rbpl_hint);
1911                 if (i > (RBPL_TABLE_SIZE - 1)) {
1912                         i = find_first_zero_bit(he_dev->rbpl_table, RBPL_TABLE_SIZE);
1913                         if (i > (RBPL_TABLE_SIZE - 1))
1914                                 break;
1915                 }
1916                 he_dev->rbpl_hint = i + 1;
1917
1918                 heb = pci_pool_alloc(he_dev->rbpl_pool, GFP_ATOMIC|GFP_DMA, &mapping);
1919                 if (!heb)
1920                         break;
1921                 heb->mapping = mapping;
1922                 list_add(&heb->entry, &he_dev->rbpl_outstanding);
1923                 he_dev->rbpl_virt[i] = heb;
1924                 set_bit(i, he_dev->rbpl_table);
1925                 new_tail->idx = i << RBP_IDX_OFFSET;
1926                 new_tail->phys = mapping + offsetof(struct he_buff, data);
1927
1928                 he_dev->rbpl_tail = new_tail;
1929                 ++moved;
1930         } 
1931
1932         if (moved)
1933                 he_writel(he_dev, RBPL_MASK(he_dev->rbpl_tail), G0_RBPL_T);
1934 }
1935
1936 static void
1937 he_tasklet(unsigned long data)
1938 {
1939         unsigned long flags;
1940         struct he_dev *he_dev = (struct he_dev *) data;
1941         int group, type;
1942         int updated = 0;
1943
1944         HPRINTK("tasklet (0x%lx)\n", data);
1945         spin_lock_irqsave(&he_dev->global_lock, flags);
1946
1947         while (he_dev->irq_head != he_dev->irq_tail) {
1948                 ++updated;
1949
1950                 type = ITYPE_TYPE(he_dev->irq_head->isw);
1951                 group = ITYPE_GROUP(he_dev->irq_head->isw);
1952
1953                 switch (type) {
1954                         case ITYPE_RBRQ_THRESH:
1955                                 HPRINTK("rbrq%d threshold\n", group);
1956                                 /* fall through */
1957                         case ITYPE_RBRQ_TIMER:
1958                                 if (he_service_rbrq(he_dev, group))
1959                                         he_service_rbpl(he_dev, group);
1960                                 break;
1961                         case ITYPE_TBRQ_THRESH:
1962                                 HPRINTK("tbrq%d threshold\n", group);
1963                                 /* fall through */
1964                         case ITYPE_TPD_COMPLETE:
1965                                 he_service_tbrq(he_dev, group);
1966                                 break;
1967                         case ITYPE_RBPL_THRESH:
1968                                 he_service_rbpl(he_dev, group);
1969                                 break;
1970                         case ITYPE_RBPS_THRESH:
1971                                 /* shouldn't happen unless small buffers enabled */
1972                                 break;
1973                         case ITYPE_PHY:
1974                                 HPRINTK("phy interrupt\n");
1975 #ifdef CONFIG_ATM_HE_USE_SUNI
1976                                 spin_unlock_irqrestore(&he_dev->global_lock, flags);
1977                                 if (he_dev->atm_dev->phy && he_dev->atm_dev->phy->interrupt)
1978                                         he_dev->atm_dev->phy->interrupt(he_dev->atm_dev);
1979                                 spin_lock_irqsave(&he_dev->global_lock, flags);
1980 #endif
1981                                 break;
1982                         case ITYPE_OTHER:
1983                                 switch (type|group) {
1984                                         case ITYPE_PARITY:
1985                                                 hprintk("parity error\n");
1986                                                 break;
1987                                         case ITYPE_ABORT:
1988                                                 hprintk("abort 0x%x\n", he_readl(he_dev, ABORT_ADDR));
1989                                                 break;
1990                                 }
1991                                 break;
1992                         case ITYPE_TYPE(ITYPE_INVALID):
1993                                 /* see 8.1.1 -- check all queues */
1994
1995                                 HPRINTK("isw not updated 0x%x\n", he_dev->irq_head->isw);
1996
1997                                 he_service_rbrq(he_dev, 0);
1998                                 he_service_rbpl(he_dev, 0);
1999                                 he_service_tbrq(he_dev, 0);
2000                                 break;
2001                         default:
2002                                 hprintk("bad isw 0x%x?\n", he_dev->irq_head->isw);
2003                 }
2004
2005                 he_dev->irq_head->isw = ITYPE_INVALID;
2006
2007                 he_dev->irq_head = (struct he_irq *) NEXT_ENTRY(he_dev->irq_base, he_dev->irq_head, IRQ_MASK);
2008         }
2009
2010         if (updated) {
2011                 if (updated > he_dev->irq_peak)
2012                         he_dev->irq_peak = updated;
2013
2014                 he_writel(he_dev,
2015                         IRQ_SIZE(CONFIG_IRQ_SIZE) |
2016                         IRQ_THRESH(CONFIG_IRQ_THRESH) |
2017                         IRQ_TAIL(he_dev->irq_tail), IRQ0_HEAD);
2018                 (void) he_readl(he_dev, INT_FIFO); /* 8.1.2 controller errata; flush posted writes */
2019         }
2020         spin_unlock_irqrestore(&he_dev->global_lock, flags);
2021 }
2022
2023 static irqreturn_t
2024 he_irq_handler(int irq, void *dev_id)
2025 {
2026         unsigned long flags;
2027         struct he_dev *he_dev = (struct he_dev * )dev_id;
2028         int handled = 0;
2029
2030         if (he_dev == NULL)
2031                 return IRQ_NONE;
2032
2033         spin_lock_irqsave(&he_dev->global_lock, flags);
2034
2035         he_dev->irq_tail = (struct he_irq *) (((unsigned long)he_dev->irq_base) |
2036                                                 (*he_dev->irq_tailoffset << 2));
2037
2038         if (he_dev->irq_tail == he_dev->irq_head) {
2039                 HPRINTK("tailoffset not updated?\n");
2040                 he_dev->irq_tail = (struct he_irq *) ((unsigned long)he_dev->irq_base |
2041                         ((he_readl(he_dev, IRQ0_BASE) & IRQ_MASK) << 2));
2042                 (void) he_readl(he_dev, INT_FIFO);      /* 8.1.2 controller errata */
2043         }
2044
2045 #ifdef DEBUG
2046         if (he_dev->irq_head == he_dev->irq_tail /* && !IRQ_PENDING */)
2047                 hprintk("spurious (or shared) interrupt?\n");
2048 #endif
2049
2050         if (he_dev->irq_head != he_dev->irq_tail) {
2051                 handled = 1;
2052                 tasklet_schedule(&he_dev->tasklet);
2053                 he_writel(he_dev, INT_CLEAR_A, INT_FIFO);       /* clear interrupt */
2054                 (void) he_readl(he_dev, INT_FIFO);              /* flush posted writes */
2055         }
2056         spin_unlock_irqrestore(&he_dev->global_lock, flags);
2057         return IRQ_RETVAL(handled);
2058
2059 }
2060
2061 static __inline__ void
2062 __enqueue_tpd(struct he_dev *he_dev, struct he_tpd *tpd, unsigned cid)
2063 {
2064         struct he_tpdrq *new_tail;
2065
2066         HPRINTK("tpdrq %p cid 0x%x -> tpdrq_tail %p\n",
2067                                         tpd, cid, he_dev->tpdrq_tail);
2068
2069         /* new_tail = he_dev->tpdrq_tail; */
2070         new_tail = (struct he_tpdrq *) ((unsigned long) he_dev->tpdrq_base |
2071                                         TPDRQ_MASK(he_dev->tpdrq_tail+1));
2072
2073         /*
2074          * check to see if we are about to set the tail == head
2075          * if true, update the head pointer from the adapter
2076          * to see if this is really the case (reading the queue
2077          * head for every enqueue would be unnecessarily slow)
2078          */
2079
2080         if (new_tail == he_dev->tpdrq_head) {
2081                 he_dev->tpdrq_head = (struct he_tpdrq *)
2082                         (((unsigned long)he_dev->tpdrq_base) |
2083                                 TPDRQ_MASK(he_readl(he_dev, TPDRQ_B_H)));
2084
2085                 if (new_tail == he_dev->tpdrq_head) {
2086                         int slot;
2087
2088                         hprintk("tpdrq full (cid 0x%x)\n", cid);
2089                         /*
2090                          * FIXME
2091                          * push tpd onto a transmit backlog queue
2092                          * after service_tbrq, service the backlog
2093                          * for now, we just drop the pdu
2094                          */
2095                         for (slot = 0; slot < TPD_MAXIOV; ++slot) {
2096                                 if (tpd->iovec[slot].addr)
2097                                         pci_unmap_single(he_dev->pci_dev,
2098                                                 tpd->iovec[slot].addr,
2099                                                 tpd->iovec[slot].len & TPD_LEN_MASK,
2100                                                                 PCI_DMA_TODEVICE);
2101                         }
2102                         if (tpd->skb) {
2103                                 if (tpd->vcc->pop)
2104                                         tpd->vcc->pop(tpd->vcc, tpd->skb);
2105                                 else
2106                                         dev_kfree_skb_any(tpd->skb);
2107                                 atomic_inc(&tpd->vcc->stats->tx_err);
2108                         }
2109                         pci_pool_free(he_dev->tpd_pool, tpd, TPD_ADDR(tpd->status));
2110                         return;
2111                 }
2112         }
2113
2114         /* 2.1.5 transmit packet descriptor ready queue */
2115         list_add_tail(&tpd->entry, &he_dev->outstanding_tpds);
2116         he_dev->tpdrq_tail->tpd = TPD_ADDR(tpd->status);
2117         he_dev->tpdrq_tail->cid = cid;
2118         wmb();
2119
2120         he_dev->tpdrq_tail = new_tail;
2121
2122         he_writel(he_dev, TPDRQ_MASK(he_dev->tpdrq_tail), TPDRQ_T);
2123         (void) he_readl(he_dev, TPDRQ_T);               /* flush posted writes */
2124 }
2125
2126 static int
2127 he_open(struct atm_vcc *vcc)
2128 {
2129         unsigned long flags;
2130         struct he_dev *he_dev = HE_DEV(vcc->dev);
2131         struct he_vcc *he_vcc;
2132         int err = 0;
2133         unsigned cid, rsr0, rsr1, rsr4, tsr0, tsr0_aal, tsr4, period, reg, clock;
2134         short vpi = vcc->vpi;
2135         int vci = vcc->vci;
2136
2137         if (vci == ATM_VCI_UNSPEC || vpi == ATM_VPI_UNSPEC)
2138                 return 0;
2139
2140         HPRINTK("open vcc %p %d.%d\n", vcc, vpi, vci);
2141
2142         set_bit(ATM_VF_ADDR, &vcc->flags);
2143
2144         cid = he_mkcid(he_dev, vpi, vci);
2145
2146         he_vcc = kmalloc(sizeof(struct he_vcc), GFP_ATOMIC);
2147         if (he_vcc == NULL) {
2148                 hprintk("unable to allocate he_vcc during open\n");
2149                 return -ENOMEM;
2150         }
2151
2152         INIT_LIST_HEAD(&he_vcc->buffers);
2153         he_vcc->pdu_len = 0;
2154         he_vcc->rc_index = -1;
2155
2156         init_waitqueue_head(&he_vcc->rx_waitq);
2157         init_waitqueue_head(&he_vcc->tx_waitq);
2158
2159         vcc->dev_data = he_vcc;
2160
2161         if (vcc->qos.txtp.traffic_class != ATM_NONE) {
2162                 int pcr_goal;
2163
2164                 pcr_goal = atm_pcr_goal(&vcc->qos.txtp);
2165                 if (pcr_goal == 0)
2166                         pcr_goal = he_dev->atm_dev->link_rate;
2167                 if (pcr_goal < 0)       /* means round down, technically */
2168                         pcr_goal = -pcr_goal;
2169
2170                 HPRINTK("open tx cid 0x%x pcr_goal %d\n", cid, pcr_goal);
2171
2172                 switch (vcc->qos.aal) {
2173                         case ATM_AAL5:
2174                                 tsr0_aal = TSR0_AAL5;
2175                                 tsr4 = TSR4_AAL5;
2176                                 break;
2177                         case ATM_AAL0:
2178                                 tsr0_aal = TSR0_AAL0_SDU;
2179                                 tsr4 = TSR4_AAL0_SDU;
2180                                 break;
2181                         default:
2182                                 err = -EINVAL;
2183                                 goto open_failed;
2184                 }
2185
2186                 spin_lock_irqsave(&he_dev->global_lock, flags);
2187                 tsr0 = he_readl_tsr0(he_dev, cid);
2188                 spin_unlock_irqrestore(&he_dev->global_lock, flags);
2189
2190                 if (TSR0_CONN_STATE(tsr0) != 0) {
2191                         hprintk("cid 0x%x not idle (tsr0 = 0x%x)\n", cid, tsr0);
2192                         err = -EBUSY;
2193                         goto open_failed;
2194                 }
2195
2196                 switch (vcc->qos.txtp.traffic_class) {
2197                         case ATM_UBR:
2198                                 /* 2.3.3.1 open connection ubr */
2199
2200                                 tsr0 = TSR0_UBR | TSR0_GROUP(0) | tsr0_aal |
2201                                         TSR0_USE_WMIN | TSR0_UPDATE_GER;
2202                                 break;
2203
2204                         case ATM_CBR:
2205                                 /* 2.3.3.2 open connection cbr */
2206
2207                                 /* 8.2.3 cbr scheduler wrap problem -- limit to 90% total link rate */
2208                                 if ((he_dev->total_bw + pcr_goal)
2209                                         > (he_dev->atm_dev->link_rate * 9 / 10))
2210                                 {
2211                                         err = -EBUSY;
2212                                         goto open_failed;
2213                                 }
2214
2215                                 spin_lock_irqsave(&he_dev->global_lock, flags);                 /* also protects he_dev->cs_stper[] */
2216
2217                                 /* find an unused cs_stper register */
2218                                 for (reg = 0; reg < HE_NUM_CS_STPER; ++reg)
2219                                         if (he_dev->cs_stper[reg].inuse == 0 || 
2220                                             he_dev->cs_stper[reg].pcr == pcr_goal)
2221                                                         break;
2222
2223                                 if (reg == HE_NUM_CS_STPER) {
2224                                         err = -EBUSY;
2225                                         spin_unlock_irqrestore(&he_dev->global_lock, flags);
2226                                         goto open_failed;
2227                                 }
2228
2229                                 he_dev->total_bw += pcr_goal;
2230
2231                                 he_vcc->rc_index = reg;
2232                                 ++he_dev->cs_stper[reg].inuse;
2233                                 he_dev->cs_stper[reg].pcr = pcr_goal;
2234
2235                                 clock = he_is622(he_dev) ? 66667000 : 50000000;
2236                                 period = clock / pcr_goal;
2237                                 
2238                                 HPRINTK("rc_index = %d period = %d\n",
2239                                                                 reg, period);
2240
2241                                 he_writel_mbox(he_dev, rate_to_atmf(period/2),
2242                                                         CS_STPER0 + reg);
2243                                 spin_unlock_irqrestore(&he_dev->global_lock, flags);
2244
2245                                 tsr0 = TSR0_CBR | TSR0_GROUP(0) | tsr0_aal |
2246                                                         TSR0_RC_INDEX(reg);
2247
2248                                 break;
2249                         default:
2250                                 err = -EINVAL;
2251                                 goto open_failed;
2252                 }
2253
2254                 spin_lock_irqsave(&he_dev->global_lock, flags);
2255
2256                 he_writel_tsr0(he_dev, tsr0, cid);
2257                 he_writel_tsr4(he_dev, tsr4 | 1, cid);
2258                 he_writel_tsr1(he_dev, TSR1_MCR(rate_to_atmf(0)) |
2259                                         TSR1_PCR(rate_to_atmf(pcr_goal)), cid);
2260                 he_writel_tsr2(he_dev, TSR2_ACR(rate_to_atmf(pcr_goal)), cid);
2261                 he_writel_tsr9(he_dev, TSR9_OPEN_CONN, cid);
2262
2263                 he_writel_tsr3(he_dev, 0x0, cid);
2264                 he_writel_tsr5(he_dev, 0x0, cid);
2265                 he_writel_tsr6(he_dev, 0x0, cid);
2266                 he_writel_tsr7(he_dev, 0x0, cid);
2267                 he_writel_tsr8(he_dev, 0x0, cid);
2268                 he_writel_tsr10(he_dev, 0x0, cid);
2269                 he_writel_tsr11(he_dev, 0x0, cid);
2270                 he_writel_tsr12(he_dev, 0x0, cid);
2271                 he_writel_tsr13(he_dev, 0x0, cid);
2272                 he_writel_tsr14(he_dev, 0x0, cid);
2273                 (void) he_readl_tsr0(he_dev, cid);              /* flush posted writes */
2274                 spin_unlock_irqrestore(&he_dev->global_lock, flags);
2275         }
2276
2277         if (vcc->qos.rxtp.traffic_class != ATM_NONE) {
2278                 unsigned aal;
2279
2280                 HPRINTK("open rx cid 0x%x (rx_waitq %p)\n", cid,
2281                                                 &HE_VCC(vcc)->rx_waitq);
2282
2283                 switch (vcc->qos.aal) {
2284                         case ATM_AAL5:
2285                                 aal = RSR0_AAL5;
2286                                 break;
2287                         case ATM_AAL0:
2288                                 aal = RSR0_RAWCELL;
2289                                 break;
2290                         default:
2291                                 err = -EINVAL;
2292                                 goto open_failed;
2293                 }
2294
2295                 spin_lock_irqsave(&he_dev->global_lock, flags);
2296
2297                 rsr0 = he_readl_rsr0(he_dev, cid);
2298                 if (rsr0 & RSR0_OPEN_CONN) {
2299                         spin_unlock_irqrestore(&he_dev->global_lock, flags);
2300
2301                         hprintk("cid 0x%x not idle (rsr0 = 0x%x)\n", cid, rsr0);
2302                         err = -EBUSY;
2303                         goto open_failed;
2304                 }
2305
2306                 rsr1 = RSR1_GROUP(0) | RSR1_RBPL_ONLY;
2307                 rsr4 = RSR4_GROUP(0) | RSR4_RBPL_ONLY;
2308                 rsr0 = vcc->qos.rxtp.traffic_class == ATM_UBR ? 
2309                                 (RSR0_EPD_ENABLE|RSR0_PPD_ENABLE) : 0;
2310
2311 #ifdef USE_CHECKSUM_HW
2312                 if (vpi == 0 && vci >= ATM_NOT_RSV_VCI)
2313                         rsr0 |= RSR0_TCP_CKSUM;
2314 #endif
2315
2316                 he_writel_rsr4(he_dev, rsr4, cid);
2317                 he_writel_rsr1(he_dev, rsr1, cid);
2318                 /* 5.1.11 last parameter initialized should be
2319                           the open/closed indication in rsr0 */
2320                 he_writel_rsr0(he_dev,
2321                         rsr0 | RSR0_START_PDU | RSR0_OPEN_CONN | aal, cid);
2322                 (void) he_readl_rsr0(he_dev, cid);              /* flush posted writes */
2323
2324                 spin_unlock_irqrestore(&he_dev->global_lock, flags);
2325         }
2326
2327 open_failed:
2328
2329         if (err) {
2330                 kfree(he_vcc);
2331                 clear_bit(ATM_VF_ADDR, &vcc->flags);
2332         }
2333         else
2334                 set_bit(ATM_VF_READY, &vcc->flags);
2335
2336         return err;
2337 }
2338
2339 static void
2340 he_close(struct atm_vcc *vcc)
2341 {
2342         unsigned long flags;
2343         DECLARE_WAITQUEUE(wait, current);
2344         struct he_dev *he_dev = HE_DEV(vcc->dev);
2345         struct he_tpd *tpd;
2346         unsigned cid;
2347         struct he_vcc *he_vcc = HE_VCC(vcc);
2348 #define MAX_RETRY 30
2349         int retry = 0, sleep = 1, tx_inuse;
2350
2351         HPRINTK("close vcc %p %d.%d\n", vcc, vcc->vpi, vcc->vci);
2352
2353         clear_bit(ATM_VF_READY, &vcc->flags);
2354         cid = he_mkcid(he_dev, vcc->vpi, vcc->vci);
2355
2356         if (vcc->qos.rxtp.traffic_class != ATM_NONE) {
2357                 int timeout;
2358
2359                 HPRINTK("close rx cid 0x%x\n", cid);
2360
2361                 /* 2.7.2.2 close receive operation */
2362
2363                 /* wait for previous close (if any) to finish */
2364
2365                 spin_lock_irqsave(&he_dev->global_lock, flags);
2366                 while (he_readl(he_dev, RCC_STAT) & RCC_BUSY) {
2367                         HPRINTK("close cid 0x%x RCC_BUSY\n", cid);
2368                         udelay(250);
2369                 }
2370
2371                 set_current_state(TASK_UNINTERRUPTIBLE);
2372                 add_wait_queue(&he_vcc->rx_waitq, &wait);
2373
2374                 he_writel_rsr0(he_dev, RSR0_CLOSE_CONN, cid);
2375                 (void) he_readl_rsr0(he_dev, cid);              /* flush posted writes */
2376                 he_writel_mbox(he_dev, cid, RXCON_CLOSE);
2377                 spin_unlock_irqrestore(&he_dev->global_lock, flags);
2378
2379                 timeout = schedule_timeout(30*HZ);
2380
2381                 remove_wait_queue(&he_vcc->rx_waitq, &wait);
2382                 set_current_state(TASK_RUNNING);
2383
2384                 if (timeout == 0)
2385                         hprintk("close rx timeout cid 0x%x\n", cid);
2386
2387                 HPRINTK("close rx cid 0x%x complete\n", cid);
2388
2389         }
2390
2391         if (vcc->qos.txtp.traffic_class != ATM_NONE) {
2392                 volatile unsigned tsr4, tsr0;
2393                 int timeout;
2394
2395                 HPRINTK("close tx cid 0x%x\n", cid);
2396                 
2397                 /* 2.1.2
2398                  *
2399                  * ... the host must first stop queueing packets to the TPDRQ
2400                  * on the connection to be closed, then wait for all outstanding
2401                  * packets to be transmitted and their buffers returned to the
2402                  * TBRQ. When the last packet on the connection arrives in the
2403                  * TBRQ, the host issues the close command to the adapter.
2404                  */
2405
2406                 while (((tx_inuse = atomic_read(&sk_atm(vcc)->sk_wmem_alloc)) > 1) &&
2407                        (retry < MAX_RETRY)) {
2408                         msleep(sleep);
2409                         if (sleep < 250)
2410                                 sleep = sleep * 2;
2411
2412                         ++retry;
2413                 }
2414
2415                 if (tx_inuse > 1)
2416                         hprintk("close tx cid 0x%x tx_inuse = %d\n", cid, tx_inuse);
2417
2418                 /* 2.3.1.1 generic close operations with flush */
2419
2420                 spin_lock_irqsave(&he_dev->global_lock, flags);
2421                 he_writel_tsr4_upper(he_dev, TSR4_FLUSH_CONN, cid);
2422                                         /* also clears TSR4_SESSION_ENDED */
2423
2424                 switch (vcc->qos.txtp.traffic_class) {
2425                         case ATM_UBR:
2426                                 he_writel_tsr1(he_dev, 
2427                                         TSR1_MCR(rate_to_atmf(200000))
2428                                         | TSR1_PCR(0), cid);
2429                                 break;
2430                         case ATM_CBR:
2431                                 he_writel_tsr14_upper(he_dev, TSR14_DELETE, cid);
2432                                 break;
2433                 }
2434                 (void) he_readl_tsr4(he_dev, cid);              /* flush posted writes */
2435
2436                 tpd = __alloc_tpd(he_dev);
2437                 if (tpd == NULL) {
2438                         hprintk("close tx he_alloc_tpd failed cid 0x%x\n", cid);
2439                         goto close_tx_incomplete;
2440                 }
2441                 tpd->status |= TPD_EOS | TPD_INT;
2442                 tpd->skb = NULL;
2443                 tpd->vcc = vcc;
2444                 wmb();
2445
2446                 set_current_state(TASK_UNINTERRUPTIBLE);
2447                 add_wait_queue(&he_vcc->tx_waitq, &wait);
2448                 __enqueue_tpd(he_dev, tpd, cid);
2449                 spin_unlock_irqrestore(&he_dev->global_lock, flags);
2450
2451                 timeout = schedule_timeout(30*HZ);
2452
2453                 remove_wait_queue(&he_vcc->tx_waitq, &wait);
2454                 set_current_state(TASK_RUNNING);
2455
2456                 spin_lock_irqsave(&he_dev->global_lock, flags);
2457
2458                 if (timeout == 0) {
2459                         hprintk("close tx timeout cid 0x%x\n", cid);
2460                         goto close_tx_incomplete;
2461                 }
2462
2463                 while (!((tsr4 = he_readl_tsr4(he_dev, cid)) & TSR4_SESSION_ENDED)) {
2464                         HPRINTK("close tx cid 0x%x !TSR4_SESSION_ENDED (tsr4 = 0x%x)\n", cid, tsr4);
2465                         udelay(250);
2466                 }
2467
2468                 while (TSR0_CONN_STATE(tsr0 = he_readl_tsr0(he_dev, cid)) != 0) {
2469                         HPRINTK("close tx cid 0x%x TSR0_CONN_STATE != 0 (tsr0 = 0x%x)\n", cid, tsr0);
2470                         udelay(250);
2471                 }
2472
2473 close_tx_incomplete:
2474
2475                 if (vcc->qos.txtp.traffic_class == ATM_CBR) {
2476                         int reg = he_vcc->rc_index;
2477
2478                         HPRINTK("cs_stper reg = %d\n", reg);
2479
2480                         if (he_dev->cs_stper[reg].inuse == 0)
2481                                 hprintk("cs_stper[%d].inuse = 0!\n", reg);
2482                         else
2483                                 --he_dev->cs_stper[reg].inuse;
2484
2485                         he_dev->total_bw -= he_dev->cs_stper[reg].pcr;
2486                 }
2487                 spin_unlock_irqrestore(&he_dev->global_lock, flags);
2488
2489                 HPRINTK("close tx cid 0x%x complete\n", cid);
2490         }
2491
2492         kfree(he_vcc);
2493
2494         clear_bit(ATM_VF_ADDR, &vcc->flags);
2495 }
2496
2497 static int
2498 he_send(struct atm_vcc *vcc, struct sk_buff *skb)
2499 {
2500         unsigned long flags;
2501         struct he_dev *he_dev = HE_DEV(vcc->dev);
2502         unsigned cid = he_mkcid(he_dev, vcc->vpi, vcc->vci);
2503         struct he_tpd *tpd;
2504 #ifdef USE_SCATTERGATHER
2505         int i, slot = 0;
2506 #endif
2507
2508 #define HE_TPD_BUFSIZE 0xffff
2509
2510         HPRINTK("send %d.%d\n", vcc->vpi, vcc->vci);
2511
2512         if ((skb->len > HE_TPD_BUFSIZE) ||
2513             ((vcc->qos.aal == ATM_AAL0) && (skb->len != ATM_AAL0_SDU))) {
2514                 hprintk("buffer too large (or small) -- %d bytes\n", skb->len );
2515                 if (vcc->pop)
2516                         vcc->pop(vcc, skb);
2517                 else
2518                         dev_kfree_skb_any(skb);
2519                 atomic_inc(&vcc->stats->tx_err);
2520                 return -EINVAL;
2521         }
2522
2523 #ifndef USE_SCATTERGATHER
2524         if (skb_shinfo(skb)->nr_frags) {
2525                 hprintk("no scatter/gather support\n");
2526                 if (vcc->pop)
2527                         vcc->pop(vcc, skb);
2528                 else
2529                         dev_kfree_skb_any(skb);
2530                 atomic_inc(&vcc->stats->tx_err);
2531                 return -EINVAL;
2532         }
2533 #endif
2534         spin_lock_irqsave(&he_dev->global_lock, flags);
2535
2536         tpd = __alloc_tpd(he_dev);
2537         if (tpd == NULL) {
2538                 if (vcc->pop)
2539                         vcc->pop(vcc, skb);
2540                 else
2541                         dev_kfree_skb_any(skb);
2542                 atomic_inc(&vcc->stats->tx_err);
2543                 spin_unlock_irqrestore(&he_dev->global_lock, flags);
2544                 return -ENOMEM;
2545         }
2546
2547         if (vcc->qos.aal == ATM_AAL5)
2548                 tpd->status |= TPD_CELLTYPE(TPD_USERCELL);
2549         else {
2550                 char *pti_clp = (void *) (skb->data + 3);
2551                 int clp, pti;
2552
2553                 pti = (*pti_clp & ATM_HDR_PTI_MASK) >> ATM_HDR_PTI_SHIFT; 
2554                 clp = (*pti_clp & ATM_HDR_CLP);
2555                 tpd->status |= TPD_CELLTYPE(pti);
2556                 if (clp)
2557                         tpd->status |= TPD_CLP;
2558
2559                 skb_pull(skb, ATM_AAL0_SDU - ATM_CELL_PAYLOAD);
2560         }
2561
2562 #ifdef USE_SCATTERGATHER
2563         tpd->iovec[slot].addr = pci_map_single(he_dev->pci_dev, skb->data,
2564                                 skb_headlen(skb), PCI_DMA_TODEVICE);
2565         tpd->iovec[slot].len = skb_headlen(skb);
2566         ++slot;
2567
2568         for (i = 0; i < skb_shinfo(skb)->nr_frags; i++) {
2569                 skb_frag_t *frag = &skb_shinfo(skb)->frags[i];
2570
2571                 if (slot == TPD_MAXIOV) {       /* queue tpd; start new tpd */
2572                         tpd->vcc = vcc;
2573                         tpd->skb = NULL;        /* not the last fragment
2574                                                    so dont ->push() yet */
2575                         wmb();
2576
2577                         __enqueue_tpd(he_dev, tpd, cid);
2578                         tpd = __alloc_tpd(he_dev);
2579                         if (tpd == NULL) {
2580                                 if (vcc->pop)
2581                                         vcc->pop(vcc, skb);
2582                                 else
2583                                         dev_kfree_skb_any(skb);
2584                                 atomic_inc(&vcc->stats->tx_err);
2585                                 spin_unlock_irqrestore(&he_dev->global_lock, flags);
2586                                 return -ENOMEM;
2587                         }
2588                         tpd->status |= TPD_USERCELL;
2589                         slot = 0;
2590                 }
2591
2592                 tpd->iovec[slot].addr = pci_map_single(he_dev->pci_dev,
2593                         (void *) page_address(frag->page) + frag->page_offset,
2594                                 frag->size, PCI_DMA_TODEVICE);
2595                 tpd->iovec[slot].len = frag->size;
2596                 ++slot;
2597
2598         }
2599
2600         tpd->iovec[slot - 1].len |= TPD_LST;
2601 #else
2602         tpd->address0 = pci_map_single(he_dev->pci_dev, skb->data, skb->len, PCI_DMA_TODEVICE);
2603         tpd->length0 = skb->len | TPD_LST;
2604 #endif
2605         tpd->status |= TPD_INT;
2606
2607         tpd->vcc = vcc;
2608         tpd->skb = skb;
2609         wmb();
2610         ATM_SKB(skb)->vcc = vcc;
2611
2612         __enqueue_tpd(he_dev, tpd, cid);
2613         spin_unlock_irqrestore(&he_dev->global_lock, flags);
2614
2615         atomic_inc(&vcc->stats->tx);
2616
2617         return 0;
2618 }
2619
2620 static int
2621 he_ioctl(struct atm_dev *atm_dev, unsigned int cmd, void __user *arg)
2622 {
2623         unsigned long flags;
2624         struct he_dev *he_dev = HE_DEV(atm_dev);
2625         struct he_ioctl_reg reg;
2626         int err = 0;
2627
2628         switch (cmd) {
2629                 case HE_GET_REG:
2630                         if (!capable(CAP_NET_ADMIN))
2631                                 return -EPERM;
2632
2633                         if (copy_from_user(&reg, arg,
2634                                            sizeof(struct he_ioctl_reg)))
2635                                 return -EFAULT;
2636
2637                         spin_lock_irqsave(&he_dev->global_lock, flags);
2638                         switch (reg.type) {
2639                                 case HE_REGTYPE_PCI:
2640                                         if (reg.addr >= HE_REGMAP_SIZE) {
2641                                                 err = -EINVAL;
2642                                                 break;
2643                                         }
2644
2645                                         reg.val = he_readl(he_dev, reg.addr);
2646                                         break;
2647                                 case HE_REGTYPE_RCM:
2648                                         reg.val =
2649                                                 he_readl_rcm(he_dev, reg.addr);
2650                                         break;
2651                                 case HE_REGTYPE_TCM:
2652                                         reg.val =
2653                                                 he_readl_tcm(he_dev, reg.addr);
2654                                         break;
2655                                 case HE_REGTYPE_MBOX:
2656                                         reg.val =
2657                                                 he_readl_mbox(he_dev, reg.addr);
2658                                         break;
2659                                 default:
2660                                         err = -EINVAL;
2661                                         break;
2662                         }
2663                         spin_unlock_irqrestore(&he_dev->global_lock, flags);
2664                         if (err == 0)
2665                                 if (copy_to_user(arg, &reg,
2666                                                         sizeof(struct he_ioctl_reg)))
2667                                         return -EFAULT;
2668                         break;
2669                 default:
2670 #ifdef CONFIG_ATM_HE_USE_SUNI
2671                         if (atm_dev->phy && atm_dev->phy->ioctl)
2672                                 err = atm_dev->phy->ioctl(atm_dev, cmd, arg);
2673 #else /* CONFIG_ATM_HE_USE_SUNI */
2674                         err = -EINVAL;
2675 #endif /* CONFIG_ATM_HE_USE_SUNI */
2676                         break;
2677         }
2678
2679         return err;
2680 }
2681
2682 static void
2683 he_phy_put(struct atm_dev *atm_dev, unsigned char val, unsigned long addr)
2684 {
2685         unsigned long flags;
2686         struct he_dev *he_dev = HE_DEV(atm_dev);
2687
2688         HPRINTK("phy_put(val 0x%x, addr 0x%lx)\n", val, addr);
2689
2690         spin_lock_irqsave(&he_dev->global_lock, flags);
2691         he_writel(he_dev, val, FRAMER + (addr*4));
2692         (void) he_readl(he_dev, FRAMER + (addr*4));             /* flush posted writes */
2693         spin_unlock_irqrestore(&he_dev->global_lock, flags);
2694 }
2695  
2696         
2697 static unsigned char
2698 he_phy_get(struct atm_dev *atm_dev, unsigned long addr)
2699
2700         unsigned long flags;
2701         struct he_dev *he_dev = HE_DEV(atm_dev);
2702         unsigned reg;
2703
2704         spin_lock_irqsave(&he_dev->global_lock, flags);
2705         reg = he_readl(he_dev, FRAMER + (addr*4));
2706         spin_unlock_irqrestore(&he_dev->global_lock, flags);
2707
2708         HPRINTK("phy_get(addr 0x%lx) =0x%x\n", addr, reg);
2709         return reg;
2710 }
2711
2712 static int
2713 he_proc_read(struct atm_dev *dev, loff_t *pos, char *page)
2714 {
2715         unsigned long flags;
2716         struct he_dev *he_dev = HE_DEV(dev);
2717         int left, i;
2718 #ifdef notdef
2719         struct he_rbrq *rbrq_tail;
2720         struct he_tpdrq *tpdrq_head;
2721         int rbpl_head, rbpl_tail;
2722 #endif
2723         static long mcc = 0, oec = 0, dcc = 0, cec = 0;
2724
2725
2726         left = *pos;
2727         if (!left--)
2728                 return sprintf(page, "ATM he driver\n");
2729
2730         if (!left--)
2731                 return sprintf(page, "%s%s\n\n",
2732                         he_dev->prod_id, he_dev->media & 0x40 ? "SM" : "MM");
2733
2734         if (!left--)
2735                 return sprintf(page, "Mismatched Cells  VPI/VCI Not Open  Dropped Cells  RCM Dropped Cells\n");
2736
2737         spin_lock_irqsave(&he_dev->global_lock, flags);
2738         mcc += he_readl(he_dev, MCC);
2739         oec += he_readl(he_dev, OEC);
2740         dcc += he_readl(he_dev, DCC);
2741         cec += he_readl(he_dev, CEC);
2742         spin_unlock_irqrestore(&he_dev->global_lock, flags);
2743
2744         if (!left--)
2745                 return sprintf(page, "%16ld  %16ld  %13ld  %17ld\n\n", 
2746                                                         mcc, oec, dcc, cec);
2747
2748         if (!left--)
2749                 return sprintf(page, "irq_size = %d  inuse = ?  peak = %d\n",
2750                                 CONFIG_IRQ_SIZE, he_dev->irq_peak);
2751
2752         if (!left--)
2753                 return sprintf(page, "tpdrq_size = %d  inuse = ?\n",
2754                                                 CONFIG_TPDRQ_SIZE);
2755
2756         if (!left--)
2757                 return sprintf(page, "rbrq_size = %d  inuse = ?  peak = %d\n",
2758                                 CONFIG_RBRQ_SIZE, he_dev->rbrq_peak);
2759
2760         if (!left--)
2761                 return sprintf(page, "tbrq_size = %d  peak = %d\n",
2762                                         CONFIG_TBRQ_SIZE, he_dev->tbrq_peak);
2763
2764
2765 #ifdef notdef
2766         rbpl_head = RBPL_MASK(he_readl(he_dev, G0_RBPL_S));
2767         rbpl_tail = RBPL_MASK(he_readl(he_dev, G0_RBPL_T));
2768
2769         inuse = rbpl_head - rbpl_tail;
2770         if (inuse < 0)
2771                 inuse += CONFIG_RBPL_SIZE * sizeof(struct he_rbp);
2772         inuse /= sizeof(struct he_rbp);
2773
2774         if (!left--)
2775                 return sprintf(page, "rbpl_size = %d  inuse = %d\n\n",
2776                                                 CONFIG_RBPL_SIZE, inuse);
2777 #endif
2778
2779         if (!left--)
2780                 return sprintf(page, "rate controller periods (cbr)\n                 pcr  #vc\n");
2781
2782         for (i = 0; i < HE_NUM_CS_STPER; ++i)
2783                 if (!left--)
2784                         return sprintf(page, "cs_stper%-2d  %8ld  %3d\n", i,
2785                                                 he_dev->cs_stper[i].pcr,
2786                                                 he_dev->cs_stper[i].inuse);
2787
2788         if (!left--)
2789                 return sprintf(page, "total bw (cbr): %d  (limit %d)\n",
2790                         he_dev->total_bw, he_dev->atm_dev->link_rate * 10 / 9);
2791
2792         return 0;
2793 }
2794
2795 /* eeprom routines  -- see 4.7 */
2796
2797 static u8 read_prom_byte(struct he_dev *he_dev, int addr)
2798 {
2799         u32 val = 0, tmp_read = 0;
2800         int i, j = 0;
2801         u8 byte_read = 0;
2802
2803         val = readl(he_dev->membase + HOST_CNTL);
2804         val &= 0xFFFFE0FF;
2805        
2806         /* Turn on write enable */
2807         val |= 0x800;
2808         he_writel(he_dev, val, HOST_CNTL);
2809        
2810         /* Send READ instruction */
2811         for (i = 0; i < ARRAY_SIZE(readtab); i++) {
2812                 he_writel(he_dev, val | readtab[i], HOST_CNTL);
2813                 udelay(EEPROM_DELAY);
2814         }
2815        
2816         /* Next, we need to send the byte address to read from */
2817         for (i = 7; i >= 0; i--) {
2818                 he_writel(he_dev, val | clocktab[j++] | (((addr >> i) & 1) << 9), HOST_CNTL);
2819                 udelay(EEPROM_DELAY);
2820                 he_writel(he_dev, val | clocktab[j++] | (((addr >> i) & 1) << 9), HOST_CNTL);
2821                 udelay(EEPROM_DELAY);
2822         }
2823        
2824         j = 0;
2825
2826         val &= 0xFFFFF7FF;      /* Turn off write enable */
2827         he_writel(he_dev, val, HOST_CNTL);
2828        
2829         /* Now, we can read data from the EEPROM by clocking it in */
2830         for (i = 7; i >= 0; i--) {
2831                 he_writel(he_dev, val | clocktab[j++], HOST_CNTL);
2832                 udelay(EEPROM_DELAY);
2833                 tmp_read = he_readl(he_dev, HOST_CNTL);
2834                 byte_read |= (unsigned char)
2835                            ((tmp_read & ID_DOUT) >> ID_DOFFSET << i);
2836                 he_writel(he_dev, val | clocktab[j++], HOST_CNTL);
2837                 udelay(EEPROM_DELAY);
2838         }
2839        
2840         he_writel(he_dev, val | ID_CS, HOST_CNTL);
2841         udelay(EEPROM_DELAY);
2842
2843         return byte_read;
2844 }
2845
2846 MODULE_LICENSE("GPL");
2847 MODULE_AUTHOR("chas williams <chas@cmf.nrl.navy.mil>");
2848 MODULE_DESCRIPTION("ForeRunnerHE ATM Adapter driver");
2849 module_param(disable64, bool, 0);
2850 MODULE_PARM_DESC(disable64, "disable 64-bit pci bus transfers");
2851 module_param(nvpibits, short, 0);
2852 MODULE_PARM_DESC(nvpibits, "numbers of bits for vpi (default 0)");
2853 module_param(nvcibits, short, 0);
2854 MODULE_PARM_DESC(nvcibits, "numbers of bits for vci (default 12)");
2855 module_param(rx_skb_reserve, short, 0);
2856 MODULE_PARM_DESC(rx_skb_reserve, "padding for receive skb (default 16)");
2857 module_param(irq_coalesce, bool, 0);
2858 MODULE_PARM_DESC(irq_coalesce, "use interrupt coalescing (default 1)");
2859 module_param(sdh, bool, 0);
2860 MODULE_PARM_DESC(sdh, "use SDH framing (default 0)");
2861
2862 static struct pci_device_id he_pci_tbl[] = {
2863         { PCI_VDEVICE(FORE, PCI_DEVICE_ID_FORE_HE), 0 },
2864         { 0, }
2865 };
2866
2867 MODULE_DEVICE_TABLE(pci, he_pci_tbl);
2868
2869 static struct pci_driver he_driver = {
2870         .name =         "he",
2871         .probe =        he_init_one,
2872         .remove =       he_remove_one,
2873         .id_table =     he_pci_tbl,
2874 };
2875
2876 static int __init he_init(void)
2877 {
2878         return pci_register_driver(&he_driver);
2879 }
2880
2881 static void __exit he_cleanup(void)
2882 {
2883         pci_unregister_driver(&he_driver);
2884 }
2885
2886 module_init(he_init);
2887 module_exit(he_cleanup);