1c35e1d637a0e30aecd6f3369a0f89cdbf34b972
[linux-2.6.git] / drivers / net / pcnet32.c
1 /* pcnet32.c: An AMD PCnet32 ethernet driver for linux. */
2 /*
3  *      Copyright 1996-1999 Thomas Bogendoerfer
4  *
5  *      Derived from the lance driver written 1993,1994,1995 by Donald Becker.
6  *
7  *      Copyright 1993 United States Government as represented by the
8  *      Director, National Security Agency.
9  *
10  *      This software may be used and distributed according to the terms
11  *      of the GNU General Public License, incorporated herein by reference.
12  *
13  *      This driver is for PCnet32 and PCnetPCI based ethercards
14  */
15 /**************************************************************************
16  *  23 Oct, 2000.
17  *  Fixed a few bugs, related to running the controller in 32bit mode.
18  *
19  *  Carsten Langgaard, carstenl@mips.com
20  *  Copyright (C) 2000 MIPS Technologies, Inc.  All rights reserved.
21  *
22  *************************************************************************/
23
24 #define DRV_NAME        "pcnet32"
25 #define DRV_VERSION     "1.35"
26 #define DRV_RELDATE     "21.Apr.2008"
27 #define PFX             DRV_NAME ": "
28
29 static const char *const version =
30     DRV_NAME ".c:v" DRV_VERSION " " DRV_RELDATE " tsbogend@alpha.franken.de\n";
31
32 #include <linux/module.h>
33 #include <linux/kernel.h>
34 #include <linux/string.h>
35 #include <linux/errno.h>
36 #include <linux/ioport.h>
37 #include <linux/slab.h>
38 #include <linux/interrupt.h>
39 #include <linux/pci.h>
40 #include <linux/delay.h>
41 #include <linux/init.h>
42 #include <linux/ethtool.h>
43 #include <linux/mii.h>
44 #include <linux/crc32.h>
45 #include <linux/netdevice.h>
46 #include <linux/etherdevice.h>
47 #include <linux/skbuff.h>
48 #include <linux/spinlock.h>
49 #include <linux/moduleparam.h>
50 #include <linux/bitops.h>
51
52 #include <asm/dma.h>
53 #include <asm/io.h>
54 #include <asm/uaccess.h>
55 #include <asm/irq.h>
56
57 /*
58  * PCI device identifiers for "new style" Linux PCI Device Drivers
59  */
60 static struct pci_device_id pcnet32_pci_tbl[] = {
61         { PCI_DEVICE(PCI_VENDOR_ID_AMD, PCI_DEVICE_ID_AMD_LANCE_HOME), },
62         { PCI_DEVICE(PCI_VENDOR_ID_AMD, PCI_DEVICE_ID_AMD_LANCE), },
63
64         /*
65          * Adapters that were sold with IBM's RS/6000 or pSeries hardware have
66          * the incorrect vendor id.
67          */
68         { PCI_DEVICE(PCI_VENDOR_ID_TRIDENT, PCI_DEVICE_ID_AMD_LANCE),
69           .class = (PCI_CLASS_NETWORK_ETHERNET << 8), .class_mask = 0xffff00, },
70
71         { }     /* terminate list */
72 };
73
74 MODULE_DEVICE_TABLE(pci, pcnet32_pci_tbl);
75
76 static int cards_found;
77
78 /*
79  * VLB I/O addresses
80  */
81 static unsigned int pcnet32_portlist[] __initdata =
82     { 0x300, 0x320, 0x340, 0x360, 0 };
83
84 static int pcnet32_debug = 0;
85 static int tx_start = 1;        /* Mapping -- 0:20, 1:64, 2:128, 3:~220 (depends on chip vers) */
86 static int pcnet32vlb;          /* check for VLB cards ? */
87
88 static struct net_device *pcnet32_dev;
89
90 static int max_interrupt_work = 2;
91 static int rx_copybreak = 200;
92
93 #define PCNET32_PORT_AUI      0x00
94 #define PCNET32_PORT_10BT     0x01
95 #define PCNET32_PORT_GPSI     0x02
96 #define PCNET32_PORT_MII      0x03
97
98 #define PCNET32_PORT_PORTSEL  0x03
99 #define PCNET32_PORT_ASEL     0x04
100 #define PCNET32_PORT_100      0x40
101 #define PCNET32_PORT_FD       0x80
102
103 #define PCNET32_DMA_MASK 0xffffffff
104
105 #define PCNET32_WATCHDOG_TIMEOUT (jiffies + (2 * HZ))
106 #define PCNET32_BLINK_TIMEOUT   (jiffies + (HZ/4))
107
108 /*
109  * table to translate option values from tulip
110  * to internal options
111  */
112 static const unsigned char options_mapping[] = {
113         PCNET32_PORT_ASEL,                      /*  0 Auto-select      */
114         PCNET32_PORT_AUI,                       /*  1 BNC/AUI          */
115         PCNET32_PORT_AUI,                       /*  2 AUI/BNC          */
116         PCNET32_PORT_ASEL,                      /*  3 not supported    */
117         PCNET32_PORT_10BT | PCNET32_PORT_FD,    /*  4 10baseT-FD       */
118         PCNET32_PORT_ASEL,                      /*  5 not supported    */
119         PCNET32_PORT_ASEL,                      /*  6 not supported    */
120         PCNET32_PORT_ASEL,                      /*  7 not supported    */
121         PCNET32_PORT_ASEL,                      /*  8 not supported    */
122         PCNET32_PORT_MII,                       /*  9 MII 10baseT      */
123         PCNET32_PORT_MII | PCNET32_PORT_FD,     /* 10 MII 10baseT-FD   */
124         PCNET32_PORT_MII,                       /* 11 MII (autosel)    */
125         PCNET32_PORT_10BT,                      /* 12 10BaseT          */
126         PCNET32_PORT_MII | PCNET32_PORT_100,    /* 13 MII 100BaseTx    */
127                                                 /* 14 MII 100BaseTx-FD */
128         PCNET32_PORT_MII | PCNET32_PORT_100 | PCNET32_PORT_FD,
129         PCNET32_PORT_ASEL                       /* 15 not supported    */
130 };
131
132 static const char pcnet32_gstrings_test[][ETH_GSTRING_LEN] = {
133         "Loopback test  (offline)"
134 };
135
136 #define PCNET32_TEST_LEN        ARRAY_SIZE(pcnet32_gstrings_test)
137
138 #define PCNET32_NUM_REGS 136
139
140 #define MAX_UNITS 8             /* More are supported, limit only on options */
141 static int options[MAX_UNITS];
142 static int full_duplex[MAX_UNITS];
143 static int homepna[MAX_UNITS];
144
145 /*
146  *                              Theory of Operation
147  *
148  * This driver uses the same software structure as the normal lance
149  * driver. So look for a verbose description in lance.c. The differences
150  * to the normal lance driver is the use of the 32bit mode of PCnet32
151  * and PCnetPCI chips. Because these chips are 32bit chips, there is no
152  * 16MB limitation and we don't need bounce buffers.
153  */
154
155 /*
156  * Set the number of Tx and Rx buffers, using Log_2(# buffers).
157  * Reasonable default values are 4 Tx buffers, and 16 Rx buffers.
158  * That translates to 2 (4 == 2^^2) and 4 (16 == 2^^4).
159  */
160 #ifndef PCNET32_LOG_TX_BUFFERS
161 #define PCNET32_LOG_TX_BUFFERS          4
162 #define PCNET32_LOG_RX_BUFFERS          5
163 #define PCNET32_LOG_MAX_TX_BUFFERS      9       /* 2^9 == 512 */
164 #define PCNET32_LOG_MAX_RX_BUFFERS      9
165 #endif
166
167 #define TX_RING_SIZE            (1 << (PCNET32_LOG_TX_BUFFERS))
168 #define TX_MAX_RING_SIZE        (1 << (PCNET32_LOG_MAX_TX_BUFFERS))
169
170 #define RX_RING_SIZE            (1 << (PCNET32_LOG_RX_BUFFERS))
171 #define RX_MAX_RING_SIZE        (1 << (PCNET32_LOG_MAX_RX_BUFFERS))
172
173 #define PKT_BUF_SKB             1544
174 /* actual buffer length after being aligned */
175 #define PKT_BUF_SIZE            (PKT_BUF_SKB - NET_IP_ALIGN)
176 /* chip wants twos complement of the (aligned) buffer length */
177 #define NEG_BUF_SIZE            (NET_IP_ALIGN - PKT_BUF_SKB)
178
179 /* Offsets from base I/O address. */
180 #define PCNET32_WIO_RDP         0x10
181 #define PCNET32_WIO_RAP         0x12
182 #define PCNET32_WIO_RESET       0x14
183 #define PCNET32_WIO_BDP         0x16
184
185 #define PCNET32_DWIO_RDP        0x10
186 #define PCNET32_DWIO_RAP        0x14
187 #define PCNET32_DWIO_RESET      0x18
188 #define PCNET32_DWIO_BDP        0x1C
189
190 #define PCNET32_TOTAL_SIZE      0x20
191
192 #define CSR0            0
193 #define CSR0_INIT       0x1
194 #define CSR0_START      0x2
195 #define CSR0_STOP       0x4
196 #define CSR0_TXPOLL     0x8
197 #define CSR0_INTEN      0x40
198 #define CSR0_IDON       0x0100
199 #define CSR0_NORMAL     (CSR0_START | CSR0_INTEN)
200 #define PCNET32_INIT_LOW        1
201 #define PCNET32_INIT_HIGH       2
202 #define CSR3            3
203 #define CSR4            4
204 #define CSR5            5
205 #define CSR5_SUSPEND    0x0001
206 #define CSR15           15
207 #define PCNET32_MC_FILTER       8
208
209 #define PCNET32_79C970A 0x2621
210
211 /* The PCNET32 Rx and Tx ring descriptors. */
212 struct pcnet32_rx_head {
213         __le32  base;
214         __le16  buf_length;     /* two`s complement of length */
215         __le16  status;
216         __le32  msg_length;
217         __le32  reserved;
218 };
219
220 struct pcnet32_tx_head {
221         __le32  base;
222         __le16  length;         /* two`s complement of length */
223         __le16  status;
224         __le32  misc;
225         __le32  reserved;
226 };
227
228 /* The PCNET32 32-Bit initialization block, described in databook. */
229 struct pcnet32_init_block {
230         __le16  mode;
231         __le16  tlen_rlen;
232         u8      phys_addr[6];
233         __le16  reserved;
234         __le32  filter[2];
235         /* Receive and transmit ring base, along with extra bits. */
236         __le32  rx_ring;
237         __le32  tx_ring;
238 };
239
240 /* PCnet32 access functions */
241 struct pcnet32_access {
242         u16     (*read_csr) (unsigned long, int);
243         void    (*write_csr) (unsigned long, int, u16);
244         u16     (*read_bcr) (unsigned long, int);
245         void    (*write_bcr) (unsigned long, int, u16);
246         u16     (*read_rap) (unsigned long);
247         void    (*write_rap) (unsigned long, u16);
248         void    (*reset) (unsigned long);
249 };
250
251 /*
252  * The first field of pcnet32_private is read by the ethernet device
253  * so the structure should be allocated using pci_alloc_consistent().
254  */
255 struct pcnet32_private {
256         struct pcnet32_init_block *init_block;
257         /* The Tx and Rx ring entries must be aligned on 16-byte boundaries in 32bit mode. */
258         struct pcnet32_rx_head  *rx_ring;
259         struct pcnet32_tx_head  *tx_ring;
260         dma_addr_t              init_dma_addr;/* DMA address of beginning of the init block,
261                                    returned by pci_alloc_consistent */
262         struct pci_dev          *pci_dev;
263         const char              *name;
264         /* The saved address of a sent-in-place packet/buffer, for skfree(). */
265         struct sk_buff          **tx_skbuff;
266         struct sk_buff          **rx_skbuff;
267         dma_addr_t              *tx_dma_addr;
268         dma_addr_t              *rx_dma_addr;
269         struct pcnet32_access   a;
270         spinlock_t              lock;           /* Guard lock */
271         unsigned int            cur_rx, cur_tx; /* The next free ring entry */
272         unsigned int            rx_ring_size;   /* current rx ring size */
273         unsigned int            tx_ring_size;   /* current tx ring size */
274         unsigned int            rx_mod_mask;    /* rx ring modular mask */
275         unsigned int            tx_mod_mask;    /* tx ring modular mask */
276         unsigned short          rx_len_bits;
277         unsigned short          tx_len_bits;
278         dma_addr_t              rx_ring_dma_addr;
279         dma_addr_t              tx_ring_dma_addr;
280         unsigned int            dirty_rx,       /* ring entries to be freed. */
281                                 dirty_tx;
282
283         struct net_device       *dev;
284         struct napi_struct      napi;
285         char                    tx_full;
286         char                    phycount;       /* number of phys found */
287         int                     options;
288         unsigned int            shared_irq:1,   /* shared irq possible */
289                                 dxsuflo:1,   /* disable transmit stop on uflo */
290                                 mii:1;          /* mii port available */
291         struct net_device       *next;
292         struct mii_if_info      mii_if;
293         struct timer_list       watchdog_timer;
294         struct timer_list       blink_timer;
295         u32                     msg_enable;     /* debug message level */
296
297         /* each bit indicates an available PHY */
298         u32                     phymask;
299         unsigned short          chip_version;   /* which variant this is */
300 };
301
302 static int pcnet32_probe_pci(struct pci_dev *, const struct pci_device_id *);
303 static int pcnet32_probe1(unsigned long, int, struct pci_dev *);
304 static int pcnet32_open(struct net_device *);
305 static int pcnet32_init_ring(struct net_device *);
306 static int pcnet32_start_xmit(struct sk_buff *, struct net_device *);
307 static void pcnet32_tx_timeout(struct net_device *dev);
308 static irqreturn_t pcnet32_interrupt(int, void *);
309 static int pcnet32_close(struct net_device *);
310 static struct net_device_stats *pcnet32_get_stats(struct net_device *);
311 static void pcnet32_load_multicast(struct net_device *dev);
312 static void pcnet32_set_multicast_list(struct net_device *);
313 static int pcnet32_ioctl(struct net_device *, struct ifreq *, int);
314 static void pcnet32_watchdog(struct net_device *);
315 static int mdio_read(struct net_device *dev, int phy_id, int reg_num);
316 static void mdio_write(struct net_device *dev, int phy_id, int reg_num,
317                        int val);
318 static void pcnet32_restart(struct net_device *dev, unsigned int csr0_bits);
319 static void pcnet32_ethtool_test(struct net_device *dev,
320                                  struct ethtool_test *eth_test, u64 * data);
321 static int pcnet32_loopback_test(struct net_device *dev, uint64_t * data1);
322 static int pcnet32_phys_id(struct net_device *dev, u32 data);
323 static void pcnet32_led_blink_callback(struct net_device *dev);
324 static int pcnet32_get_regs_len(struct net_device *dev);
325 static void pcnet32_get_regs(struct net_device *dev, struct ethtool_regs *regs,
326                              void *ptr);
327 static void pcnet32_purge_tx_ring(struct net_device *dev);
328 static int pcnet32_alloc_ring(struct net_device *dev, const char *name);
329 static void pcnet32_free_ring(struct net_device *dev);
330 static void pcnet32_check_media(struct net_device *dev, int verbose);
331
332 static u16 pcnet32_wio_read_csr(unsigned long addr, int index)
333 {
334         outw(index, addr + PCNET32_WIO_RAP);
335         return inw(addr + PCNET32_WIO_RDP);
336 }
337
338 static void pcnet32_wio_write_csr(unsigned long addr, int index, u16 val)
339 {
340         outw(index, addr + PCNET32_WIO_RAP);
341         outw(val, addr + PCNET32_WIO_RDP);
342 }
343
344 static u16 pcnet32_wio_read_bcr(unsigned long addr, int index)
345 {
346         outw(index, addr + PCNET32_WIO_RAP);
347         return inw(addr + PCNET32_WIO_BDP);
348 }
349
350 static void pcnet32_wio_write_bcr(unsigned long addr, int index, u16 val)
351 {
352         outw(index, addr + PCNET32_WIO_RAP);
353         outw(val, addr + PCNET32_WIO_BDP);
354 }
355
356 static u16 pcnet32_wio_read_rap(unsigned long addr)
357 {
358         return inw(addr + PCNET32_WIO_RAP);
359 }
360
361 static void pcnet32_wio_write_rap(unsigned long addr, u16 val)
362 {
363         outw(val, addr + PCNET32_WIO_RAP);
364 }
365
366 static void pcnet32_wio_reset(unsigned long addr)
367 {
368         inw(addr + PCNET32_WIO_RESET);
369 }
370
371 static int pcnet32_wio_check(unsigned long addr)
372 {
373         outw(88, addr + PCNET32_WIO_RAP);
374         return (inw(addr + PCNET32_WIO_RAP) == 88);
375 }
376
377 static struct pcnet32_access pcnet32_wio = {
378         .read_csr = pcnet32_wio_read_csr,
379         .write_csr = pcnet32_wio_write_csr,
380         .read_bcr = pcnet32_wio_read_bcr,
381         .write_bcr = pcnet32_wio_write_bcr,
382         .read_rap = pcnet32_wio_read_rap,
383         .write_rap = pcnet32_wio_write_rap,
384         .reset = pcnet32_wio_reset
385 };
386
387 static u16 pcnet32_dwio_read_csr(unsigned long addr, int index)
388 {
389         outl(index, addr + PCNET32_DWIO_RAP);
390         return (inl(addr + PCNET32_DWIO_RDP) & 0xffff);
391 }
392
393 static void pcnet32_dwio_write_csr(unsigned long addr, int index, u16 val)
394 {
395         outl(index, addr + PCNET32_DWIO_RAP);
396         outl(val, addr + PCNET32_DWIO_RDP);
397 }
398
399 static u16 pcnet32_dwio_read_bcr(unsigned long addr, int index)
400 {
401         outl(index, addr + PCNET32_DWIO_RAP);
402         return (inl(addr + PCNET32_DWIO_BDP) & 0xffff);
403 }
404
405 static void pcnet32_dwio_write_bcr(unsigned long addr, int index, u16 val)
406 {
407         outl(index, addr + PCNET32_DWIO_RAP);
408         outl(val, addr + PCNET32_DWIO_BDP);
409 }
410
411 static u16 pcnet32_dwio_read_rap(unsigned long addr)
412 {
413         return (inl(addr + PCNET32_DWIO_RAP) & 0xffff);
414 }
415
416 static void pcnet32_dwio_write_rap(unsigned long addr, u16 val)
417 {
418         outl(val, addr + PCNET32_DWIO_RAP);
419 }
420
421 static void pcnet32_dwio_reset(unsigned long addr)
422 {
423         inl(addr + PCNET32_DWIO_RESET);
424 }
425
426 static int pcnet32_dwio_check(unsigned long addr)
427 {
428         outl(88, addr + PCNET32_DWIO_RAP);
429         return ((inl(addr + PCNET32_DWIO_RAP) & 0xffff) == 88);
430 }
431
432 static struct pcnet32_access pcnet32_dwio = {
433         .read_csr = pcnet32_dwio_read_csr,
434         .write_csr = pcnet32_dwio_write_csr,
435         .read_bcr = pcnet32_dwio_read_bcr,
436         .write_bcr = pcnet32_dwio_write_bcr,
437         .read_rap = pcnet32_dwio_read_rap,
438         .write_rap = pcnet32_dwio_write_rap,
439         .reset = pcnet32_dwio_reset
440 };
441
442 static void pcnet32_netif_stop(struct net_device *dev)
443 {
444         struct pcnet32_private *lp = netdev_priv(dev);
445
446         dev->trans_start = jiffies;
447         napi_disable(&lp->napi);
448         netif_tx_disable(dev);
449 }
450
451 static void pcnet32_netif_start(struct net_device *dev)
452 {
453         struct pcnet32_private *lp = netdev_priv(dev);
454         ulong ioaddr = dev->base_addr;
455         u16 val;
456
457         netif_wake_queue(dev);
458         val = lp->a.read_csr(ioaddr, CSR3);
459         val &= 0x00ff;
460         lp->a.write_csr(ioaddr, CSR3, val);
461         napi_enable(&lp->napi);
462 }
463
464 /*
465  * Allocate space for the new sized tx ring.
466  * Free old resources
467  * Save new resources.
468  * Any failure keeps old resources.
469  * Must be called with lp->lock held.
470  */
471 static void pcnet32_realloc_tx_ring(struct net_device *dev,
472                                     struct pcnet32_private *lp,
473                                     unsigned int size)
474 {
475         dma_addr_t new_ring_dma_addr;
476         dma_addr_t *new_dma_addr_list;
477         struct pcnet32_tx_head *new_tx_ring;
478         struct sk_buff **new_skb_list;
479
480         pcnet32_purge_tx_ring(dev);
481
482         new_tx_ring = pci_alloc_consistent(lp->pci_dev,
483                                            sizeof(struct pcnet32_tx_head) *
484                                            (1 << size),
485                                            &new_ring_dma_addr);
486         if (new_tx_ring == NULL) {
487                 if (netif_msg_drv(lp))
488                         printk("\n" KERN_ERR
489                                "%s: Consistent memory allocation failed.\n",
490                                dev->name);
491                 return;
492         }
493         memset(new_tx_ring, 0, sizeof(struct pcnet32_tx_head) * (1 << size));
494
495         new_dma_addr_list = kcalloc((1 << size), sizeof(dma_addr_t),
496                                 GFP_ATOMIC);
497         if (!new_dma_addr_list) {
498                 if (netif_msg_drv(lp))
499                         printk("\n" KERN_ERR
500                                "%s: Memory allocation failed.\n", dev->name);
501                 goto free_new_tx_ring;
502         }
503
504         new_skb_list = kcalloc((1 << size), sizeof(struct sk_buff *),
505                                 GFP_ATOMIC);
506         if (!new_skb_list) {
507                 if (netif_msg_drv(lp))
508                         printk("\n" KERN_ERR
509                                "%s: Memory allocation failed.\n", dev->name);
510                 goto free_new_lists;
511         }
512
513         kfree(lp->tx_skbuff);
514         kfree(lp->tx_dma_addr);
515         pci_free_consistent(lp->pci_dev,
516                             sizeof(struct pcnet32_tx_head) *
517                             lp->tx_ring_size, lp->tx_ring,
518                             lp->tx_ring_dma_addr);
519
520         lp->tx_ring_size = (1 << size);
521         lp->tx_mod_mask = lp->tx_ring_size - 1;
522         lp->tx_len_bits = (size << 12);
523         lp->tx_ring = new_tx_ring;
524         lp->tx_ring_dma_addr = new_ring_dma_addr;
525         lp->tx_dma_addr = new_dma_addr_list;
526         lp->tx_skbuff = new_skb_list;
527         return;
528
529     free_new_lists:
530         kfree(new_dma_addr_list);
531     free_new_tx_ring:
532         pci_free_consistent(lp->pci_dev,
533                             sizeof(struct pcnet32_tx_head) *
534                             (1 << size),
535                             new_tx_ring,
536                             new_ring_dma_addr);
537         return;
538 }
539
540 /*
541  * Allocate space for the new sized rx ring.
542  * Re-use old receive buffers.
543  *   alloc extra buffers
544  *   free unneeded buffers
545  *   free unneeded buffers
546  * Save new resources.
547  * Any failure keeps old resources.
548  * Must be called with lp->lock held.
549  */
550 static void pcnet32_realloc_rx_ring(struct net_device *dev,
551                                     struct pcnet32_private *lp,
552                                     unsigned int size)
553 {
554         dma_addr_t new_ring_dma_addr;
555         dma_addr_t *new_dma_addr_list;
556         struct pcnet32_rx_head *new_rx_ring;
557         struct sk_buff **new_skb_list;
558         int new, overlap;
559
560         new_rx_ring = pci_alloc_consistent(lp->pci_dev,
561                                            sizeof(struct pcnet32_rx_head) *
562                                            (1 << size),
563                                            &new_ring_dma_addr);
564         if (new_rx_ring == NULL) {
565                 if (netif_msg_drv(lp))
566                         printk("\n" KERN_ERR
567                                "%s: Consistent memory allocation failed.\n",
568                                dev->name);
569                 return;
570         }
571         memset(new_rx_ring, 0, sizeof(struct pcnet32_rx_head) * (1 << size));
572
573         new_dma_addr_list = kcalloc((1 << size), sizeof(dma_addr_t),
574                                 GFP_ATOMIC);
575         if (!new_dma_addr_list) {
576                 if (netif_msg_drv(lp))
577                         printk("\n" KERN_ERR
578                                "%s: Memory allocation failed.\n", dev->name);
579                 goto free_new_rx_ring;
580         }
581
582         new_skb_list = kcalloc((1 << size), sizeof(struct sk_buff *),
583                                 GFP_ATOMIC);
584         if (!new_skb_list) {
585                 if (netif_msg_drv(lp))
586                         printk("\n" KERN_ERR
587                                "%s: Memory allocation failed.\n", dev->name);
588                 goto free_new_lists;
589         }
590
591         /* first copy the current receive buffers */
592         overlap = min(size, lp->rx_ring_size);
593         for (new = 0; new < overlap; new++) {
594                 new_rx_ring[new] = lp->rx_ring[new];
595                 new_dma_addr_list[new] = lp->rx_dma_addr[new];
596                 new_skb_list[new] = lp->rx_skbuff[new];
597         }
598         /* now allocate any new buffers needed */
599         for (; new < size; new++ ) {
600                 struct sk_buff *rx_skbuff;
601                 new_skb_list[new] = dev_alloc_skb(PKT_BUF_SKB);
602                 if (!(rx_skbuff = new_skb_list[new])) {
603                         /* keep the original lists and buffers */
604                         if (netif_msg_drv(lp))
605                                 printk(KERN_ERR
606                                        "%s: pcnet32_realloc_rx_ring dev_alloc_skb failed.\n",
607                                        dev->name);
608                         goto free_all_new;
609                 }
610                 skb_reserve(rx_skbuff, NET_IP_ALIGN);
611
612                 new_dma_addr_list[new] =
613                             pci_map_single(lp->pci_dev, rx_skbuff->data,
614                                            PKT_BUF_SIZE, PCI_DMA_FROMDEVICE);
615                 new_rx_ring[new].base = cpu_to_le32(new_dma_addr_list[new]);
616                 new_rx_ring[new].buf_length = cpu_to_le16(NEG_BUF_SIZE);
617                 new_rx_ring[new].status = cpu_to_le16(0x8000);
618         }
619         /* and free any unneeded buffers */
620         for (; new < lp->rx_ring_size; new++) {
621                 if (lp->rx_skbuff[new]) {
622                         pci_unmap_single(lp->pci_dev, lp->rx_dma_addr[new],
623                                          PKT_BUF_SIZE, PCI_DMA_FROMDEVICE);
624                         dev_kfree_skb(lp->rx_skbuff[new]);
625                 }
626         }
627
628         kfree(lp->rx_skbuff);
629         kfree(lp->rx_dma_addr);
630         pci_free_consistent(lp->pci_dev,
631                             sizeof(struct pcnet32_rx_head) *
632                             lp->rx_ring_size, lp->rx_ring,
633                             lp->rx_ring_dma_addr);
634
635         lp->rx_ring_size = (1 << size);
636         lp->rx_mod_mask = lp->rx_ring_size - 1;
637         lp->rx_len_bits = (size << 4);
638         lp->rx_ring = new_rx_ring;
639         lp->rx_ring_dma_addr = new_ring_dma_addr;
640         lp->rx_dma_addr = new_dma_addr_list;
641         lp->rx_skbuff = new_skb_list;
642         return;
643
644     free_all_new:
645         for (; --new >= lp->rx_ring_size; ) {
646                 if (new_skb_list[new]) {
647                         pci_unmap_single(lp->pci_dev, new_dma_addr_list[new],
648                                          PKT_BUF_SIZE, PCI_DMA_FROMDEVICE);
649                         dev_kfree_skb(new_skb_list[new]);
650                 }
651         }
652         kfree(new_skb_list);
653     free_new_lists:
654         kfree(new_dma_addr_list);
655     free_new_rx_ring:
656         pci_free_consistent(lp->pci_dev,
657                             sizeof(struct pcnet32_rx_head) *
658                             (1 << size),
659                             new_rx_ring,
660                             new_ring_dma_addr);
661         return;
662 }
663
664 static void pcnet32_purge_rx_ring(struct net_device *dev)
665 {
666         struct pcnet32_private *lp = netdev_priv(dev);
667         int i;
668
669         /* free all allocated skbuffs */
670         for (i = 0; i < lp->rx_ring_size; i++) {
671                 lp->rx_ring[i].status = 0;      /* CPU owns buffer */
672                 wmb();          /* Make sure adapter sees owner change */
673                 if (lp->rx_skbuff[i]) {
674                         pci_unmap_single(lp->pci_dev, lp->rx_dma_addr[i],
675                                          PKT_BUF_SIZE, PCI_DMA_FROMDEVICE);
676                         dev_kfree_skb_any(lp->rx_skbuff[i]);
677                 }
678                 lp->rx_skbuff[i] = NULL;
679                 lp->rx_dma_addr[i] = 0;
680         }
681 }
682
683 #ifdef CONFIG_NET_POLL_CONTROLLER
684 static void pcnet32_poll_controller(struct net_device *dev)
685 {
686         disable_irq(dev->irq);
687         pcnet32_interrupt(0, dev);
688         enable_irq(dev->irq);
689 }
690 #endif
691
692 static int pcnet32_get_settings(struct net_device *dev, struct ethtool_cmd *cmd)
693 {
694         struct pcnet32_private *lp = netdev_priv(dev);
695         unsigned long flags;
696         int r = -EOPNOTSUPP;
697
698         if (lp->mii) {
699                 spin_lock_irqsave(&lp->lock, flags);
700                 mii_ethtool_gset(&lp->mii_if, cmd);
701                 spin_unlock_irqrestore(&lp->lock, flags);
702                 r = 0;
703         }
704         return r;
705 }
706
707 static int pcnet32_set_settings(struct net_device *dev, struct ethtool_cmd *cmd)
708 {
709         struct pcnet32_private *lp = netdev_priv(dev);
710         unsigned long flags;
711         int r = -EOPNOTSUPP;
712
713         if (lp->mii) {
714                 spin_lock_irqsave(&lp->lock, flags);
715                 r = mii_ethtool_sset(&lp->mii_if, cmd);
716                 spin_unlock_irqrestore(&lp->lock, flags);
717         }
718         return r;
719 }
720
721 static void pcnet32_get_drvinfo(struct net_device *dev,
722                                 struct ethtool_drvinfo *info)
723 {
724         struct pcnet32_private *lp = netdev_priv(dev);
725
726         strcpy(info->driver, DRV_NAME);
727         strcpy(info->version, DRV_VERSION);
728         if (lp->pci_dev)
729                 strcpy(info->bus_info, pci_name(lp->pci_dev));
730         else
731                 sprintf(info->bus_info, "VLB 0x%lx", dev->base_addr);
732 }
733
734 static u32 pcnet32_get_link(struct net_device *dev)
735 {
736         struct pcnet32_private *lp = netdev_priv(dev);
737         unsigned long flags;
738         int r;
739
740         spin_lock_irqsave(&lp->lock, flags);
741         if (lp->mii) {
742                 r = mii_link_ok(&lp->mii_if);
743         } else if (lp->chip_version >= PCNET32_79C970A) {
744                 ulong ioaddr = dev->base_addr;  /* card base I/O address */
745                 r = (lp->a.read_bcr(ioaddr, 4) != 0xc0);
746         } else {        /* can not detect link on really old chips */
747                 r = 1;
748         }
749         spin_unlock_irqrestore(&lp->lock, flags);
750
751         return r;
752 }
753
754 static u32 pcnet32_get_msglevel(struct net_device *dev)
755 {
756         struct pcnet32_private *lp = netdev_priv(dev);
757         return lp->msg_enable;
758 }
759
760 static void pcnet32_set_msglevel(struct net_device *dev, u32 value)
761 {
762         struct pcnet32_private *lp = netdev_priv(dev);
763         lp->msg_enable = value;
764 }
765
766 static int pcnet32_nway_reset(struct net_device *dev)
767 {
768         struct pcnet32_private *lp = netdev_priv(dev);
769         unsigned long flags;
770         int r = -EOPNOTSUPP;
771
772         if (lp->mii) {
773                 spin_lock_irqsave(&lp->lock, flags);
774                 r = mii_nway_restart(&lp->mii_if);
775                 spin_unlock_irqrestore(&lp->lock, flags);
776         }
777         return r;
778 }
779
780 static void pcnet32_get_ringparam(struct net_device *dev,
781                                   struct ethtool_ringparam *ering)
782 {
783         struct pcnet32_private *lp = netdev_priv(dev);
784
785         ering->tx_max_pending = TX_MAX_RING_SIZE;
786         ering->tx_pending = lp->tx_ring_size;
787         ering->rx_max_pending = RX_MAX_RING_SIZE;
788         ering->rx_pending = lp->rx_ring_size;
789 }
790
791 static int pcnet32_set_ringparam(struct net_device *dev,
792                                  struct ethtool_ringparam *ering)
793 {
794         struct pcnet32_private *lp = netdev_priv(dev);
795         unsigned long flags;
796         unsigned int size;
797         ulong ioaddr = dev->base_addr;
798         int i;
799
800         if (ering->rx_mini_pending || ering->rx_jumbo_pending)
801                 return -EINVAL;
802
803         if (netif_running(dev))
804                 pcnet32_netif_stop(dev);
805
806         spin_lock_irqsave(&lp->lock, flags);
807         lp->a.write_csr(ioaddr, CSR0, CSR0_STOP);       /* stop the chip */
808
809         size = min(ering->tx_pending, (unsigned int)TX_MAX_RING_SIZE);
810
811         /* set the minimum ring size to 4, to allow the loopback test to work
812          * unchanged.
813          */
814         for (i = 2; i <= PCNET32_LOG_MAX_TX_BUFFERS; i++) {
815                 if (size <= (1 << i))
816                         break;
817         }
818         if ((1 << i) != lp->tx_ring_size)
819                 pcnet32_realloc_tx_ring(dev, lp, i);
820
821         size = min(ering->rx_pending, (unsigned int)RX_MAX_RING_SIZE);
822         for (i = 2; i <= PCNET32_LOG_MAX_RX_BUFFERS; i++) {
823                 if (size <= (1 << i))
824                         break;
825         }
826         if ((1 << i) != lp->rx_ring_size)
827                 pcnet32_realloc_rx_ring(dev, lp, i);
828
829         lp->napi.weight = lp->rx_ring_size / 2;
830
831         if (netif_running(dev)) {
832                 pcnet32_netif_start(dev);
833                 pcnet32_restart(dev, CSR0_NORMAL);
834         }
835
836         spin_unlock_irqrestore(&lp->lock, flags);
837
838         if (netif_msg_drv(lp))
839                 printk(KERN_INFO
840                        "%s: Ring Param Settings: RX: %d, TX: %d\n", dev->name,
841                        lp->rx_ring_size, lp->tx_ring_size);
842
843         return 0;
844 }
845
846 static void pcnet32_get_strings(struct net_device *dev, u32 stringset,
847                                 u8 * data)
848 {
849         memcpy(data, pcnet32_gstrings_test, sizeof(pcnet32_gstrings_test));
850 }
851
852 static int pcnet32_get_sset_count(struct net_device *dev, int sset)
853 {
854         switch (sset) {
855         case ETH_SS_TEST:
856                 return PCNET32_TEST_LEN;
857         default:
858                 return -EOPNOTSUPP;
859         }
860 }
861
862 static void pcnet32_ethtool_test(struct net_device *dev,
863                                  struct ethtool_test *test, u64 * data)
864 {
865         struct pcnet32_private *lp = netdev_priv(dev);
866         int rc;
867
868         if (test->flags == ETH_TEST_FL_OFFLINE) {
869                 rc = pcnet32_loopback_test(dev, data);
870                 if (rc) {
871                         if (netif_msg_hw(lp))
872                                 printk(KERN_DEBUG "%s: Loopback test failed.\n",
873                                        dev->name);
874                         test->flags |= ETH_TEST_FL_FAILED;
875                 } else if (netif_msg_hw(lp))
876                         printk(KERN_DEBUG "%s: Loopback test passed.\n",
877                                dev->name);
878         } else if (netif_msg_hw(lp))
879                 printk(KERN_DEBUG
880                        "%s: No tests to run (specify 'Offline' on ethtool).",
881                        dev->name);
882 }                               /* end pcnet32_ethtool_test */
883
884 static int pcnet32_loopback_test(struct net_device *dev, uint64_t * data1)
885 {
886         struct pcnet32_private *lp = netdev_priv(dev);
887         struct pcnet32_access *a = &lp->a;      /* access to registers */
888         ulong ioaddr = dev->base_addr;  /* card base I/O address */
889         struct sk_buff *skb;    /* sk buff */
890         int x, i;               /* counters */
891         int numbuffs = 4;       /* number of TX/RX buffers and descs */
892         u16 status = 0x8300;    /* TX ring status */
893         __le16 teststatus;      /* test of ring status */
894         int rc;                 /* return code */
895         int size;               /* size of packets */
896         unsigned char *packet;  /* source packet data */
897         static const int data_len = 60; /* length of source packets */
898         unsigned long flags;
899         unsigned long ticks;
900
901         rc = 1;                 /* default to fail */
902
903         if (netif_running(dev))
904                 pcnet32_netif_stop(dev);
905
906         spin_lock_irqsave(&lp->lock, flags);
907         lp->a.write_csr(ioaddr, CSR0, CSR0_STOP);       /* stop the chip */
908
909         numbuffs = min(numbuffs, (int)min(lp->rx_ring_size, lp->tx_ring_size));
910
911         /* Reset the PCNET32 */
912         lp->a.reset(ioaddr);
913         lp->a.write_csr(ioaddr, CSR4, 0x0915);  /* auto tx pad */
914
915         /* switch pcnet32 to 32bit mode */
916         lp->a.write_bcr(ioaddr, 20, 2);
917
918         /* purge & init rings but don't actually restart */
919         pcnet32_restart(dev, 0x0000);
920
921         lp->a.write_csr(ioaddr, CSR0, CSR0_STOP);       /* Set STOP bit */
922
923         /* Initialize Transmit buffers. */
924         size = data_len + 15;
925         for (x = 0; x < numbuffs; x++) {
926                 if (!(skb = dev_alloc_skb(size))) {
927                         if (netif_msg_hw(lp))
928                                 printk(KERN_DEBUG
929                                        "%s: Cannot allocate skb at line: %d!\n",
930                                        dev->name, __LINE__);
931                         goto clean_up;
932                 } else {
933                         packet = skb->data;
934                         skb_put(skb, size);     /* create space for data */
935                         lp->tx_skbuff[x] = skb;
936                         lp->tx_ring[x].length = cpu_to_le16(-skb->len);
937                         lp->tx_ring[x].misc = 0;
938
939                         /* put DA and SA into the skb */
940                         for (i = 0; i < 6; i++)
941                                 *packet++ = dev->dev_addr[i];
942                         for (i = 0; i < 6; i++)
943                                 *packet++ = dev->dev_addr[i];
944                         /* type */
945                         *packet++ = 0x08;
946                         *packet++ = 0x06;
947                         /* packet number */
948                         *packet++ = x;
949                         /* fill packet with data */
950                         for (i = 0; i < data_len; i++)
951                                 *packet++ = i;
952
953                         lp->tx_dma_addr[x] =
954                             pci_map_single(lp->pci_dev, skb->data, skb->len,
955                                            PCI_DMA_TODEVICE);
956                         lp->tx_ring[x].base = cpu_to_le32(lp->tx_dma_addr[x]);
957                         wmb();  /* Make sure owner changes after all others are visible */
958                         lp->tx_ring[x].status = cpu_to_le16(status);
959                 }
960         }
961
962         x = a->read_bcr(ioaddr, 32);    /* set internal loopback in BCR32 */
963         a->write_bcr(ioaddr, 32, x | 0x0002);
964
965         /* set int loopback in CSR15 */
966         x = a->read_csr(ioaddr, CSR15) & 0xfffc;
967         lp->a.write_csr(ioaddr, CSR15, x | 0x0044);
968
969         teststatus = cpu_to_le16(0x8000);
970         lp->a.write_csr(ioaddr, CSR0, CSR0_START);      /* Set STRT bit */
971
972         /* Check status of descriptors */
973         for (x = 0; x < numbuffs; x++) {
974                 ticks = 0;
975                 rmb();
976                 while ((lp->rx_ring[x].status & teststatus) && (ticks < 200)) {
977                         spin_unlock_irqrestore(&lp->lock, flags);
978                         msleep(1);
979                         spin_lock_irqsave(&lp->lock, flags);
980                         rmb();
981                         ticks++;
982                 }
983                 if (ticks == 200) {
984                         if (netif_msg_hw(lp))
985                                 printk("%s: Desc %d failed to reset!\n",
986                                        dev->name, x);
987                         break;
988                 }
989         }
990
991         lp->a.write_csr(ioaddr, CSR0, CSR0_STOP);       /* Set STOP bit */
992         wmb();
993         if (netif_msg_hw(lp) && netif_msg_pktdata(lp)) {
994                 printk(KERN_DEBUG "%s: RX loopback packets:\n", dev->name);
995
996                 for (x = 0; x < numbuffs; x++) {
997                         printk(KERN_DEBUG "%s: Packet %d:\n", dev->name, x);
998                         skb = lp->rx_skbuff[x];
999                         for (i = 0; i < size; i++) {
1000                                 printk("%02x ", *(skb->data + i));
1001                         }
1002                         printk("\n");
1003                 }
1004         }
1005
1006         x = 0;
1007         rc = 0;
1008         while (x < numbuffs && !rc) {
1009                 skb = lp->rx_skbuff[x];
1010                 packet = lp->tx_skbuff[x]->data;
1011                 for (i = 0; i < size; i++) {
1012                         if (*(skb->data + i) != packet[i]) {
1013                                 if (netif_msg_hw(lp))
1014                                         printk(KERN_DEBUG
1015                                                "%s: Error in compare! %2x - %02x %02x\n",
1016                                                dev->name, i, *(skb->data + i),
1017                                                packet[i]);
1018                                 rc = 1;
1019                                 break;
1020                         }
1021                 }
1022                 x++;
1023         }
1024
1025       clean_up:
1026         *data1 = rc;
1027         pcnet32_purge_tx_ring(dev);
1028
1029         x = a->read_csr(ioaddr, CSR15);
1030         a->write_csr(ioaddr, CSR15, (x & ~0x0044));     /* reset bits 6 and 2 */
1031
1032         x = a->read_bcr(ioaddr, 32);    /* reset internal loopback */
1033         a->write_bcr(ioaddr, 32, (x & ~0x0002));
1034
1035         if (netif_running(dev)) {
1036                 pcnet32_netif_start(dev);
1037                 pcnet32_restart(dev, CSR0_NORMAL);
1038         } else {
1039                 pcnet32_purge_rx_ring(dev);
1040                 lp->a.write_bcr(ioaddr, 20, 4); /* return to 16bit mode */
1041         }
1042         spin_unlock_irqrestore(&lp->lock, flags);
1043
1044         return (rc);
1045 }                               /* end pcnet32_loopback_test  */
1046
1047 static void pcnet32_led_blink_callback(struct net_device *dev)
1048 {
1049         struct pcnet32_private *lp = netdev_priv(dev);
1050         struct pcnet32_access *a = &lp->a;
1051         ulong ioaddr = dev->base_addr;
1052         unsigned long flags;
1053         int i;
1054
1055         spin_lock_irqsave(&lp->lock, flags);
1056         for (i = 4; i < 8; i++) {
1057                 a->write_bcr(ioaddr, i, a->read_bcr(ioaddr, i) ^ 0x4000);
1058         }
1059         spin_unlock_irqrestore(&lp->lock, flags);
1060
1061         mod_timer(&lp->blink_timer, PCNET32_BLINK_TIMEOUT);
1062 }
1063
1064 static int pcnet32_phys_id(struct net_device *dev, u32 data)
1065 {
1066         struct pcnet32_private *lp = netdev_priv(dev);
1067         struct pcnet32_access *a = &lp->a;
1068         ulong ioaddr = dev->base_addr;
1069         unsigned long flags;
1070         int i, regs[4];
1071
1072         if (!lp->blink_timer.function) {
1073                 init_timer(&lp->blink_timer);
1074                 lp->blink_timer.function = (void *)pcnet32_led_blink_callback;
1075                 lp->blink_timer.data = (unsigned long)dev;
1076         }
1077
1078         /* Save the current value of the bcrs */
1079         spin_lock_irqsave(&lp->lock, flags);
1080         for (i = 4; i < 8; i++) {
1081                 regs[i - 4] = a->read_bcr(ioaddr, i);
1082         }
1083         spin_unlock_irqrestore(&lp->lock, flags);
1084
1085         mod_timer(&lp->blink_timer, jiffies);
1086         set_current_state(TASK_INTERRUPTIBLE);
1087
1088         /* AV: the limit here makes no sense whatsoever */
1089         if ((!data) || (data > (u32) (MAX_SCHEDULE_TIMEOUT / HZ)))
1090                 data = (u32) (MAX_SCHEDULE_TIMEOUT / HZ);
1091
1092         msleep_interruptible(data * 1000);
1093         del_timer_sync(&lp->blink_timer);
1094
1095         /* Restore the original value of the bcrs */
1096         spin_lock_irqsave(&lp->lock, flags);
1097         for (i = 4; i < 8; i++) {
1098                 a->write_bcr(ioaddr, i, regs[i - 4]);
1099         }
1100         spin_unlock_irqrestore(&lp->lock, flags);
1101
1102         return 0;
1103 }
1104
1105 /*
1106  * lp->lock must be held.
1107  */
1108 static int pcnet32_suspend(struct net_device *dev, unsigned long *flags,
1109                 int can_sleep)
1110 {
1111         int csr5;
1112         struct pcnet32_private *lp = netdev_priv(dev);
1113         struct pcnet32_access *a = &lp->a;
1114         ulong ioaddr = dev->base_addr;
1115         int ticks;
1116
1117         /* really old chips have to be stopped. */
1118         if (lp->chip_version < PCNET32_79C970A)
1119                 return 0;
1120
1121         /* set SUSPEND (SPND) - CSR5 bit 0 */
1122         csr5 = a->read_csr(ioaddr, CSR5);
1123         a->write_csr(ioaddr, CSR5, csr5 | CSR5_SUSPEND);
1124
1125         /* poll waiting for bit to be set */
1126         ticks = 0;
1127         while (!(a->read_csr(ioaddr, CSR5) & CSR5_SUSPEND)) {
1128                 spin_unlock_irqrestore(&lp->lock, *flags);
1129                 if (can_sleep)
1130                         msleep(1);
1131                 else
1132                         mdelay(1);
1133                 spin_lock_irqsave(&lp->lock, *flags);
1134                 ticks++;
1135                 if (ticks > 200) {
1136                         if (netif_msg_hw(lp))
1137                                 printk(KERN_DEBUG
1138                                        "%s: Error getting into suspend!\n",
1139                                        dev->name);
1140                         return 0;
1141                 }
1142         }
1143         return 1;
1144 }
1145
1146 /*
1147  * process one receive descriptor entry
1148  */
1149
1150 static void pcnet32_rx_entry(struct net_device *dev,
1151                              struct pcnet32_private *lp,
1152                              struct pcnet32_rx_head *rxp,
1153                              int entry)
1154 {
1155         int status = (short)le16_to_cpu(rxp->status) >> 8;
1156         int rx_in_place = 0;
1157         struct sk_buff *skb;
1158         short pkt_len;
1159
1160         if (status != 0x03) {   /* There was an error. */
1161                 /*
1162                  * There is a tricky error noted by John Murphy,
1163                  * <murf@perftech.com> to Russ Nelson: Even with full-sized
1164                  * buffers it's possible for a jabber packet to use two
1165                  * buffers, with only the last correctly noting the error.
1166                  */
1167                 if (status & 0x01)      /* Only count a general error at the */
1168                         dev->stats.rx_errors++; /* end of a packet. */
1169                 if (status & 0x20)
1170                         dev->stats.rx_frame_errors++;
1171                 if (status & 0x10)
1172                         dev->stats.rx_over_errors++;
1173                 if (status & 0x08)
1174                         dev->stats.rx_crc_errors++;
1175                 if (status & 0x04)
1176                         dev->stats.rx_fifo_errors++;
1177                 return;
1178         }
1179
1180         pkt_len = (le32_to_cpu(rxp->msg_length) & 0xfff) - 4;
1181
1182         /* Discard oversize frames. */
1183         if (unlikely(pkt_len > PKT_BUF_SIZE)) {
1184                 if (netif_msg_drv(lp))
1185                         printk(KERN_ERR "%s: Impossible packet size %d!\n",
1186                                dev->name, pkt_len);
1187                 dev->stats.rx_errors++;
1188                 return;
1189         }
1190         if (pkt_len < 60) {
1191                 if (netif_msg_rx_err(lp))
1192                         printk(KERN_ERR "%s: Runt packet!\n", dev->name);
1193                 dev->stats.rx_errors++;
1194                 return;
1195         }
1196
1197         if (pkt_len > rx_copybreak) {
1198                 struct sk_buff *newskb;
1199
1200                 if ((newskb = dev_alloc_skb(PKT_BUF_SKB))) {
1201                         skb_reserve(newskb, NET_IP_ALIGN);
1202                         skb = lp->rx_skbuff[entry];
1203                         pci_unmap_single(lp->pci_dev,
1204                                          lp->rx_dma_addr[entry],
1205                                          PKT_BUF_SIZE,
1206                                          PCI_DMA_FROMDEVICE);
1207                         skb_put(skb, pkt_len);
1208                         lp->rx_skbuff[entry] = newskb;
1209                         lp->rx_dma_addr[entry] =
1210                                             pci_map_single(lp->pci_dev,
1211                                                            newskb->data,
1212                                                            PKT_BUF_SIZE,
1213                                                            PCI_DMA_FROMDEVICE);
1214                         rxp->base = cpu_to_le32(lp->rx_dma_addr[entry]);
1215                         rx_in_place = 1;
1216                 } else
1217                         skb = NULL;
1218         } else {
1219                 skb = dev_alloc_skb(pkt_len + NET_IP_ALIGN);
1220         }
1221
1222         if (skb == NULL) {
1223                 if (netif_msg_drv(lp))
1224                         printk(KERN_ERR
1225                                "%s: Memory squeeze, dropping packet.\n",
1226                                dev->name);
1227                 dev->stats.rx_dropped++;
1228                 return;
1229         }
1230         if (!rx_in_place) {
1231                 skb_reserve(skb, NET_IP_ALIGN);
1232                 skb_put(skb, pkt_len);  /* Make room */
1233                 pci_dma_sync_single_for_cpu(lp->pci_dev,
1234                                             lp->rx_dma_addr[entry],
1235                                             pkt_len,
1236                                             PCI_DMA_FROMDEVICE);
1237                 skb_copy_to_linear_data(skb,
1238                                  (unsigned char *)(lp->rx_skbuff[entry]->data),
1239                                  pkt_len);
1240                 pci_dma_sync_single_for_device(lp->pci_dev,
1241                                                lp->rx_dma_addr[entry],
1242                                                pkt_len,
1243                                                PCI_DMA_FROMDEVICE);
1244         }
1245         dev->stats.rx_bytes += skb->len;
1246         skb->protocol = eth_type_trans(skb, dev);
1247         netif_receive_skb(skb);
1248         dev->stats.rx_packets++;
1249         return;
1250 }
1251
1252 static int pcnet32_rx(struct net_device *dev, int budget)
1253 {
1254         struct pcnet32_private *lp = netdev_priv(dev);
1255         int entry = lp->cur_rx & lp->rx_mod_mask;
1256         struct pcnet32_rx_head *rxp = &lp->rx_ring[entry];
1257         int npackets = 0;
1258
1259         /* If we own the next entry, it's a new packet. Send it up. */
1260         while (npackets < budget && (short)le16_to_cpu(rxp->status) >= 0) {
1261                 pcnet32_rx_entry(dev, lp, rxp, entry);
1262                 npackets += 1;
1263                 /*
1264                  * The docs say that the buffer length isn't touched, but Andrew
1265                  * Boyd of QNX reports that some revs of the 79C965 clear it.
1266                  */
1267                 rxp->buf_length = cpu_to_le16(NEG_BUF_SIZE);
1268                 wmb();  /* Make sure owner changes after others are visible */
1269                 rxp->status = cpu_to_le16(0x8000);
1270                 entry = (++lp->cur_rx) & lp->rx_mod_mask;
1271                 rxp = &lp->rx_ring[entry];
1272         }
1273
1274         return npackets;
1275 }
1276
1277 static int pcnet32_tx(struct net_device *dev)
1278 {
1279         struct pcnet32_private *lp = netdev_priv(dev);
1280         unsigned int dirty_tx = lp->dirty_tx;
1281         int delta;
1282         int must_restart = 0;
1283
1284         while (dirty_tx != lp->cur_tx) {
1285                 int entry = dirty_tx & lp->tx_mod_mask;
1286                 int status = (short)le16_to_cpu(lp->tx_ring[entry].status);
1287
1288                 if (status < 0)
1289                         break;  /* It still hasn't been Txed */
1290
1291                 lp->tx_ring[entry].base = 0;
1292
1293                 if (status & 0x4000) {
1294                         /* There was a major error, log it. */
1295                         int err_status = le32_to_cpu(lp->tx_ring[entry].misc);
1296                         dev->stats.tx_errors++;
1297                         if (netif_msg_tx_err(lp))
1298                                 printk(KERN_ERR
1299                                        "%s: Tx error status=%04x err_status=%08x\n",
1300                                        dev->name, status,
1301                                        err_status);
1302                         if (err_status & 0x04000000)
1303                                 dev->stats.tx_aborted_errors++;
1304                         if (err_status & 0x08000000)
1305                                 dev->stats.tx_carrier_errors++;
1306                         if (err_status & 0x10000000)
1307                                 dev->stats.tx_window_errors++;
1308 #ifndef DO_DXSUFLO
1309                         if (err_status & 0x40000000) {
1310                                 dev->stats.tx_fifo_errors++;
1311                                 /* Ackk!  On FIFO errors the Tx unit is turned off! */
1312                                 /* Remove this verbosity later! */
1313                                 if (netif_msg_tx_err(lp))
1314                                         printk(KERN_ERR
1315                                                "%s: Tx FIFO error!\n",
1316                                                dev->name);
1317                                 must_restart = 1;
1318                         }
1319 #else
1320                         if (err_status & 0x40000000) {
1321                                 dev->stats.tx_fifo_errors++;
1322                                 if (!lp->dxsuflo) {     /* If controller doesn't recover ... */
1323                                         /* Ackk!  On FIFO errors the Tx unit is turned off! */
1324                                         /* Remove this verbosity later! */
1325                                         if (netif_msg_tx_err(lp))
1326                                                 printk(KERN_ERR
1327                                                        "%s: Tx FIFO error!\n",
1328                                                        dev->name);
1329                                         must_restart = 1;
1330                                 }
1331                         }
1332 #endif
1333                 } else {
1334                         if (status & 0x1800)
1335                                 dev->stats.collisions++;
1336                         dev->stats.tx_packets++;
1337                 }
1338
1339                 /* We must free the original skb */
1340                 if (lp->tx_skbuff[entry]) {
1341                         pci_unmap_single(lp->pci_dev,
1342                                          lp->tx_dma_addr[entry],
1343                                          lp->tx_skbuff[entry]->
1344                                          len, PCI_DMA_TODEVICE);
1345                         dev_kfree_skb_any(lp->tx_skbuff[entry]);
1346                         lp->tx_skbuff[entry] = NULL;
1347                         lp->tx_dma_addr[entry] = 0;
1348                 }
1349                 dirty_tx++;
1350         }
1351
1352         delta = (lp->cur_tx - dirty_tx) & (lp->tx_mod_mask + lp->tx_ring_size);
1353         if (delta > lp->tx_ring_size) {
1354                 if (netif_msg_drv(lp))
1355                         printk(KERN_ERR
1356                                "%s: out-of-sync dirty pointer, %d vs. %d, full=%d.\n",
1357                                dev->name, dirty_tx, lp->cur_tx,
1358                                lp->tx_full);
1359                 dirty_tx += lp->tx_ring_size;
1360                 delta -= lp->tx_ring_size;
1361         }
1362
1363         if (lp->tx_full &&
1364             netif_queue_stopped(dev) &&
1365             delta < lp->tx_ring_size - 2) {
1366                 /* The ring is no longer full, clear tbusy. */
1367                 lp->tx_full = 0;
1368                 netif_wake_queue(dev);
1369         }
1370         lp->dirty_tx = dirty_tx;
1371
1372         return must_restart;
1373 }
1374
1375 static int pcnet32_poll(struct napi_struct *napi, int budget)
1376 {
1377         struct pcnet32_private *lp = container_of(napi, struct pcnet32_private, napi);
1378         struct net_device *dev = lp->dev;
1379         unsigned long ioaddr = dev->base_addr;
1380         unsigned long flags;
1381         int work_done;
1382         u16 val;
1383
1384         work_done = pcnet32_rx(dev, budget);
1385
1386         spin_lock_irqsave(&lp->lock, flags);
1387         if (pcnet32_tx(dev)) {
1388                 /* reset the chip to clear the error condition, then restart */
1389                 lp->a.reset(ioaddr);
1390                 lp->a.write_csr(ioaddr, CSR4, 0x0915);  /* auto tx pad */
1391                 pcnet32_restart(dev, CSR0_START);
1392                 netif_wake_queue(dev);
1393         }
1394         spin_unlock_irqrestore(&lp->lock, flags);
1395
1396         if (work_done < budget) {
1397                 spin_lock_irqsave(&lp->lock, flags);
1398
1399                 __napi_complete(napi);
1400
1401                 /* clear interrupt masks */
1402                 val = lp->a.read_csr(ioaddr, CSR3);
1403                 val &= 0x00ff;
1404                 lp->a.write_csr(ioaddr, CSR3, val);
1405
1406                 /* Set interrupt enable. */
1407                 lp->a.write_csr(ioaddr, CSR0, CSR0_INTEN);
1408
1409                 spin_unlock_irqrestore(&lp->lock, flags);
1410         }
1411         return work_done;
1412 }
1413
1414 #define PCNET32_REGS_PER_PHY    32
1415 #define PCNET32_MAX_PHYS        32
1416 static int pcnet32_get_regs_len(struct net_device *dev)
1417 {
1418         struct pcnet32_private *lp = netdev_priv(dev);
1419         int j = lp->phycount * PCNET32_REGS_PER_PHY;
1420
1421         return ((PCNET32_NUM_REGS + j) * sizeof(u16));
1422 }
1423
1424 static void pcnet32_get_regs(struct net_device *dev, struct ethtool_regs *regs,
1425                              void *ptr)
1426 {
1427         int i, csr0;
1428         u16 *buff = ptr;
1429         struct pcnet32_private *lp = netdev_priv(dev);
1430         struct pcnet32_access *a = &lp->a;
1431         ulong ioaddr = dev->base_addr;
1432         unsigned long flags;
1433
1434         spin_lock_irqsave(&lp->lock, flags);
1435
1436         csr0 = a->read_csr(ioaddr, CSR0);
1437         if (!(csr0 & CSR0_STOP))        /* If not stopped */
1438                 pcnet32_suspend(dev, &flags, 1);
1439
1440         /* read address PROM */
1441         for (i = 0; i < 16; i += 2)
1442                 *buff++ = inw(ioaddr + i);
1443
1444         /* read control and status registers */
1445         for (i = 0; i < 90; i++) {
1446                 *buff++ = a->read_csr(ioaddr, i);
1447         }
1448
1449         *buff++ = a->read_csr(ioaddr, 112);
1450         *buff++ = a->read_csr(ioaddr, 114);
1451
1452         /* read bus configuration registers */
1453         for (i = 0; i < 30; i++) {
1454                 *buff++ = a->read_bcr(ioaddr, i);
1455         }
1456         *buff++ = 0;            /* skip bcr30 so as not to hang 79C976 */
1457         for (i = 31; i < 36; i++) {
1458                 *buff++ = a->read_bcr(ioaddr, i);
1459         }
1460
1461         /* read mii phy registers */
1462         if (lp->mii) {
1463                 int j;
1464                 for (j = 0; j < PCNET32_MAX_PHYS; j++) {
1465                         if (lp->phymask & (1 << j)) {
1466                                 for (i = 0; i < PCNET32_REGS_PER_PHY; i++) {
1467                                         lp->a.write_bcr(ioaddr, 33,
1468                                                         (j << 5) | i);
1469                                         *buff++ = lp->a.read_bcr(ioaddr, 34);
1470                                 }
1471                         }
1472                 }
1473         }
1474
1475         if (!(csr0 & CSR0_STOP)) {      /* If not stopped */
1476                 int csr5;
1477
1478                 /* clear SUSPEND (SPND) - CSR5 bit 0 */
1479                 csr5 = a->read_csr(ioaddr, CSR5);
1480                 a->write_csr(ioaddr, CSR5, csr5 & (~CSR5_SUSPEND));
1481         }
1482
1483         spin_unlock_irqrestore(&lp->lock, flags);
1484 }
1485
1486 static const struct ethtool_ops pcnet32_ethtool_ops = {
1487         .get_settings           = pcnet32_get_settings,
1488         .set_settings           = pcnet32_set_settings,
1489         .get_drvinfo            = pcnet32_get_drvinfo,
1490         .get_msglevel           = pcnet32_get_msglevel,
1491         .set_msglevel           = pcnet32_set_msglevel,
1492         .nway_reset             = pcnet32_nway_reset,
1493         .get_link               = pcnet32_get_link,
1494         .get_ringparam          = pcnet32_get_ringparam,
1495         .set_ringparam          = pcnet32_set_ringparam,
1496         .get_strings            = pcnet32_get_strings,
1497         .self_test              = pcnet32_ethtool_test,
1498         .phys_id                = pcnet32_phys_id,
1499         .get_regs_len           = pcnet32_get_regs_len,
1500         .get_regs               = pcnet32_get_regs,
1501         .get_sset_count         = pcnet32_get_sset_count,
1502 };
1503
1504 /* only probes for non-PCI devices, the rest are handled by
1505  * pci_register_driver via pcnet32_probe_pci */
1506
1507 static void __devinit pcnet32_probe_vlbus(unsigned int *pcnet32_portlist)
1508 {
1509         unsigned int *port, ioaddr;
1510
1511         /* search for PCnet32 VLB cards at known addresses */
1512         for (port = pcnet32_portlist; (ioaddr = *port); port++) {
1513                 if (request_region
1514                     (ioaddr, PCNET32_TOTAL_SIZE, "pcnet32_probe_vlbus")) {
1515                         /* check if there is really a pcnet chip on that ioaddr */
1516                         if ((inb(ioaddr + 14) == 0x57)
1517                             && (inb(ioaddr + 15) == 0x57)) {
1518                                 pcnet32_probe1(ioaddr, 0, NULL);
1519                         } else {
1520                                 release_region(ioaddr, PCNET32_TOTAL_SIZE);
1521                         }
1522                 }
1523         }
1524 }
1525
1526 static int __devinit
1527 pcnet32_probe_pci(struct pci_dev *pdev, const struct pci_device_id *ent)
1528 {
1529         unsigned long ioaddr;
1530         int err;
1531
1532         err = pci_enable_device(pdev);
1533         if (err < 0) {
1534                 if (pcnet32_debug & NETIF_MSG_PROBE)
1535                         printk(KERN_ERR PFX
1536                                "failed to enable device -- err=%d\n", err);
1537                 return err;
1538         }
1539         pci_set_master(pdev);
1540
1541         ioaddr = pci_resource_start(pdev, 0);
1542         if (!ioaddr) {
1543                 if (pcnet32_debug & NETIF_MSG_PROBE)
1544                         printk(KERN_ERR PFX
1545                                "card has no PCI IO resources, aborting\n");
1546                 return -ENODEV;
1547         }
1548
1549         if (!pci_dma_supported(pdev, PCNET32_DMA_MASK)) {
1550                 if (pcnet32_debug & NETIF_MSG_PROBE)
1551                         printk(KERN_ERR PFX
1552                                "architecture does not support 32bit PCI busmaster DMA\n");
1553                 return -ENODEV;
1554         }
1555         if (request_region(ioaddr, PCNET32_TOTAL_SIZE, "pcnet32_probe_pci") ==
1556             NULL) {
1557                 if (pcnet32_debug & NETIF_MSG_PROBE)
1558                         printk(KERN_ERR PFX
1559                                "io address range already allocated\n");
1560                 return -EBUSY;
1561         }
1562
1563         err = pcnet32_probe1(ioaddr, 1, pdev);
1564         if (err < 0) {
1565                 pci_disable_device(pdev);
1566         }
1567         return err;
1568 }
1569
1570 static const struct net_device_ops pcnet32_netdev_ops = {
1571         .ndo_open               = pcnet32_open,
1572         .ndo_stop               = pcnet32_close,
1573         .ndo_start_xmit         = pcnet32_start_xmit,
1574         .ndo_tx_timeout         = pcnet32_tx_timeout,
1575         .ndo_get_stats          = pcnet32_get_stats,
1576         .ndo_set_multicast_list = pcnet32_set_multicast_list,
1577         .ndo_do_ioctl           = pcnet32_ioctl,
1578         .ndo_change_mtu         = eth_change_mtu,
1579         .ndo_set_mac_address    = eth_mac_addr,
1580         .ndo_validate_addr      = eth_validate_addr,
1581 #ifdef CONFIG_NET_POLL_CONTROLLER
1582         .ndo_poll_controller    = pcnet32_poll_controller,
1583 #endif
1584 };
1585
1586 /* pcnet32_probe1
1587  *  Called from both pcnet32_probe_vlbus and pcnet_probe_pci.
1588  *  pdev will be NULL when called from pcnet32_probe_vlbus.
1589  */
1590 static int __devinit
1591 pcnet32_probe1(unsigned long ioaddr, int shared, struct pci_dev *pdev)
1592 {
1593         struct pcnet32_private *lp;
1594         int i, media;
1595         int fdx, mii, fset, dxsuflo;
1596         int chip_version;
1597         char *chipname;
1598         struct net_device *dev;
1599         struct pcnet32_access *a = NULL;
1600         u8 promaddr[6];
1601         int ret = -ENODEV;
1602
1603         /* reset the chip */
1604         pcnet32_wio_reset(ioaddr);
1605
1606         /* NOTE: 16-bit check is first, otherwise some older PCnet chips fail */
1607         if (pcnet32_wio_read_csr(ioaddr, 0) == 4 && pcnet32_wio_check(ioaddr)) {
1608                 a = &pcnet32_wio;
1609         } else {
1610                 pcnet32_dwio_reset(ioaddr);
1611                 if (pcnet32_dwio_read_csr(ioaddr, 0) == 4
1612                     && pcnet32_dwio_check(ioaddr)) {
1613                         a = &pcnet32_dwio;
1614                 } else
1615                         goto err_release_region;
1616         }
1617
1618         chip_version =
1619             a->read_csr(ioaddr, 88) | (a->read_csr(ioaddr, 89) << 16);
1620         if ((pcnet32_debug & NETIF_MSG_PROBE) && (pcnet32_debug & NETIF_MSG_HW))
1621                 printk(KERN_INFO "  PCnet chip version is %#x.\n",
1622                        chip_version);
1623         if ((chip_version & 0xfff) != 0x003) {
1624                 if (pcnet32_debug & NETIF_MSG_PROBE)
1625                         printk(KERN_INFO PFX "Unsupported chip version.\n");
1626                 goto err_release_region;
1627         }
1628
1629         /* initialize variables */
1630         fdx = mii = fset = dxsuflo = 0;
1631         chip_version = (chip_version >> 12) & 0xffff;
1632
1633         switch (chip_version) {
1634         case 0x2420:
1635                 chipname = "PCnet/PCI 79C970";  /* PCI */
1636                 break;
1637         case 0x2430:
1638                 if (shared)
1639                         chipname = "PCnet/PCI 79C970";  /* 970 gives the wrong chip id back */
1640                 else
1641                         chipname = "PCnet/32 79C965";   /* 486/VL bus */
1642                 break;
1643         case 0x2621:
1644                 chipname = "PCnet/PCI II 79C970A";      /* PCI */
1645                 fdx = 1;
1646                 break;
1647         case 0x2623:
1648                 chipname = "PCnet/FAST 79C971"; /* PCI */
1649                 fdx = 1;
1650                 mii = 1;
1651                 fset = 1;
1652                 break;
1653         case 0x2624:
1654                 chipname = "PCnet/FAST+ 79C972";        /* PCI */
1655                 fdx = 1;
1656                 mii = 1;
1657                 fset = 1;
1658                 break;
1659         case 0x2625:
1660                 chipname = "PCnet/FAST III 79C973";     /* PCI */
1661                 fdx = 1;
1662                 mii = 1;
1663                 break;
1664         case 0x2626:
1665                 chipname = "PCnet/Home 79C978"; /* PCI */
1666                 fdx = 1;
1667                 /*
1668                  * This is based on specs published at www.amd.com.  This section
1669                  * assumes that a card with a 79C978 wants to go into standard
1670                  * ethernet mode.  The 79C978 can also go into 1Mb HomePNA mode,
1671                  * and the module option homepna=1 can select this instead.
1672                  */
1673                 media = a->read_bcr(ioaddr, 49);
1674                 media &= ~3;    /* default to 10Mb ethernet */
1675                 if (cards_found < MAX_UNITS && homepna[cards_found])
1676                         media |= 1;     /* switch to home wiring mode */
1677                 if (pcnet32_debug & NETIF_MSG_PROBE)
1678                         printk(KERN_DEBUG PFX "media set to %sMbit mode.\n",
1679                                (media & 1) ? "1" : "10");
1680                 a->write_bcr(ioaddr, 49, media);
1681                 break;
1682         case 0x2627:
1683                 chipname = "PCnet/FAST III 79C975";     /* PCI */
1684                 fdx = 1;
1685                 mii = 1;
1686                 break;
1687         case 0x2628:
1688                 chipname = "PCnet/PRO 79C976";
1689                 fdx = 1;
1690                 mii = 1;
1691                 break;
1692         default:
1693                 if (pcnet32_debug & NETIF_MSG_PROBE)
1694                         printk(KERN_INFO PFX
1695                                "PCnet version %#x, no PCnet32 chip.\n",
1696                                chip_version);
1697                 goto err_release_region;
1698         }
1699
1700         /*
1701          *  On selected chips turn on the BCR18:NOUFLO bit. This stops transmit
1702          *  starting until the packet is loaded. Strike one for reliability, lose
1703          *  one for latency - although on PCI this isnt a big loss. Older chips
1704          *  have FIFO's smaller than a packet, so you can't do this.
1705          *  Turn on BCR18:BurstRdEn and BCR18:BurstWrEn.
1706          */
1707
1708         if (fset) {
1709                 a->write_bcr(ioaddr, 18, (a->read_bcr(ioaddr, 18) | 0x0860));
1710                 a->write_csr(ioaddr, 80,
1711                              (a->read_csr(ioaddr, 80) & 0x0C00) | 0x0c00);
1712                 dxsuflo = 1;
1713         }
1714
1715         dev = alloc_etherdev(sizeof(*lp));
1716         if (!dev) {
1717                 if (pcnet32_debug & NETIF_MSG_PROBE)
1718                         printk(KERN_ERR PFX "Memory allocation failed.\n");
1719                 ret = -ENOMEM;
1720                 goto err_release_region;
1721         }
1722         SET_NETDEV_DEV(dev, &pdev->dev);
1723
1724         if (pcnet32_debug & NETIF_MSG_PROBE)
1725                 printk(KERN_INFO PFX "%s at %#3lx,", chipname, ioaddr);
1726
1727         /* In most chips, after a chip reset, the ethernet address is read from the
1728          * station address PROM at the base address and programmed into the
1729          * "Physical Address Registers" CSR12-14.
1730          * As a precautionary measure, we read the PROM values and complain if
1731          * they disagree with the CSRs.  If they miscompare, and the PROM addr
1732          * is valid, then the PROM addr is used.
1733          */
1734         for (i = 0; i < 3; i++) {
1735                 unsigned int val;
1736                 val = a->read_csr(ioaddr, i + 12) & 0x0ffff;
1737                 /* There may be endianness issues here. */
1738                 dev->dev_addr[2 * i] = val & 0x0ff;
1739                 dev->dev_addr[2 * i + 1] = (val >> 8) & 0x0ff;
1740         }
1741
1742         /* read PROM address and compare with CSR address */
1743         for (i = 0; i < 6; i++)
1744                 promaddr[i] = inb(ioaddr + i);
1745
1746         if (memcmp(promaddr, dev->dev_addr, 6)
1747             || !is_valid_ether_addr(dev->dev_addr)) {
1748                 if (is_valid_ether_addr(promaddr)) {
1749                         if (pcnet32_debug & NETIF_MSG_PROBE) {
1750                                 printk(" warning: CSR address invalid,\n");
1751                                 printk(KERN_INFO
1752                                        "    using instead PROM address of");
1753                         }
1754                         memcpy(dev->dev_addr, promaddr, 6);
1755                 }
1756         }
1757         memcpy(dev->perm_addr, dev->dev_addr, dev->addr_len);
1758
1759         /* if the ethernet address is not valid, force to 00:00:00:00:00:00 */
1760         if (!is_valid_ether_addr(dev->perm_addr))
1761                 memset(dev->dev_addr, 0, sizeof(dev->dev_addr));
1762
1763         if (pcnet32_debug & NETIF_MSG_PROBE) {
1764                 printk(" %pM", dev->dev_addr);
1765
1766                 /* Version 0x2623 and 0x2624 */
1767                 if (((chip_version + 1) & 0xfffe) == 0x2624) {
1768                         i = a->read_csr(ioaddr, 80) & 0x0C00;   /* Check tx_start_pt */
1769                         printk("\n" KERN_INFO "    tx_start_pt(0x%04x):", i);
1770                         switch (i >> 10) {
1771                         case 0:
1772                                 printk("  20 bytes,");
1773                                 break;
1774                         case 1:
1775                                 printk("  64 bytes,");
1776                                 break;
1777                         case 2:
1778                                 printk(" 128 bytes,");
1779                                 break;
1780                         case 3:
1781                                 printk("~220 bytes,");
1782                                 break;
1783                         }
1784                         i = a->read_bcr(ioaddr, 18);    /* Check Burst/Bus control */
1785                         printk(" BCR18(%x):", i & 0xffff);
1786                         if (i & (1 << 5))
1787                                 printk("BurstWrEn ");
1788                         if (i & (1 << 6))
1789                                 printk("BurstRdEn ");
1790                         if (i & (1 << 7))
1791                                 printk("DWordIO ");
1792                         if (i & (1 << 11))
1793                                 printk("NoUFlow ");
1794                         i = a->read_bcr(ioaddr, 25);
1795                         printk("\n" KERN_INFO "    SRAMSIZE=0x%04x,", i << 8);
1796                         i = a->read_bcr(ioaddr, 26);
1797                         printk(" SRAM_BND=0x%04x,", i << 8);
1798                         i = a->read_bcr(ioaddr, 27);
1799                         if (i & (1 << 14))
1800                                 printk("LowLatRx");
1801                 }
1802         }
1803
1804         dev->base_addr = ioaddr;
1805         lp = netdev_priv(dev);
1806         /* pci_alloc_consistent returns page-aligned memory, so we do not have to check the alignment */
1807         if ((lp->init_block =
1808              pci_alloc_consistent(pdev, sizeof(*lp->init_block), &lp->init_dma_addr)) == NULL) {
1809                 if (pcnet32_debug & NETIF_MSG_PROBE)
1810                         printk(KERN_ERR PFX
1811                                "Consistent memory allocation failed.\n");
1812                 ret = -ENOMEM;
1813                 goto err_free_netdev;
1814         }
1815         lp->pci_dev = pdev;
1816
1817         lp->dev = dev;
1818
1819         spin_lock_init(&lp->lock);
1820
1821         SET_NETDEV_DEV(dev, &pdev->dev);
1822         lp->name = chipname;
1823         lp->shared_irq = shared;
1824         lp->tx_ring_size = TX_RING_SIZE;        /* default tx ring size */
1825         lp->rx_ring_size = RX_RING_SIZE;        /* default rx ring size */
1826         lp->tx_mod_mask = lp->tx_ring_size - 1;
1827         lp->rx_mod_mask = lp->rx_ring_size - 1;
1828         lp->tx_len_bits = (PCNET32_LOG_TX_BUFFERS << 12);
1829         lp->rx_len_bits = (PCNET32_LOG_RX_BUFFERS << 4);
1830         lp->mii_if.full_duplex = fdx;
1831         lp->mii_if.phy_id_mask = 0x1f;
1832         lp->mii_if.reg_num_mask = 0x1f;
1833         lp->dxsuflo = dxsuflo;
1834         lp->mii = mii;
1835         lp->chip_version = chip_version;
1836         lp->msg_enable = pcnet32_debug;
1837         if ((cards_found >= MAX_UNITS)
1838             || (options[cards_found] > sizeof(options_mapping)))
1839                 lp->options = PCNET32_PORT_ASEL;
1840         else
1841                 lp->options = options_mapping[options[cards_found]];
1842         lp->mii_if.dev = dev;
1843         lp->mii_if.mdio_read = mdio_read;
1844         lp->mii_if.mdio_write = mdio_write;
1845
1846         /* napi.weight is used in both the napi and non-napi cases */
1847         lp->napi.weight = lp->rx_ring_size / 2;
1848
1849         netif_napi_add(dev, &lp->napi, pcnet32_poll, lp->rx_ring_size / 2);
1850
1851         if (fdx && !(lp->options & PCNET32_PORT_ASEL) &&
1852             ((cards_found >= MAX_UNITS) || full_duplex[cards_found]))
1853                 lp->options |= PCNET32_PORT_FD;
1854
1855         if (!a) {
1856                 if (pcnet32_debug & NETIF_MSG_PROBE)
1857                         printk(KERN_ERR PFX "No access methods\n");
1858                 ret = -ENODEV;
1859                 goto err_free_consistent;
1860         }
1861         lp->a = *a;
1862
1863         /* prior to register_netdev, dev->name is not yet correct */
1864         if (pcnet32_alloc_ring(dev, pci_name(lp->pci_dev))) {
1865                 ret = -ENOMEM;
1866                 goto err_free_ring;
1867         }
1868         /* detect special T1/E1 WAN card by checking for MAC address */
1869         if (dev->dev_addr[0] == 0x00 && dev->dev_addr[1] == 0xe0
1870             && dev->dev_addr[2] == 0x75)
1871                 lp->options = PCNET32_PORT_FD | PCNET32_PORT_GPSI;
1872
1873         lp->init_block->mode = cpu_to_le16(0x0003);     /* Disable Rx and Tx. */
1874         lp->init_block->tlen_rlen =
1875             cpu_to_le16(lp->tx_len_bits | lp->rx_len_bits);
1876         for (i = 0; i < 6; i++)
1877                 lp->init_block->phys_addr[i] = dev->dev_addr[i];
1878         lp->init_block->filter[0] = 0x00000000;
1879         lp->init_block->filter[1] = 0x00000000;
1880         lp->init_block->rx_ring = cpu_to_le32(lp->rx_ring_dma_addr);
1881         lp->init_block->tx_ring = cpu_to_le32(lp->tx_ring_dma_addr);
1882
1883         /* switch pcnet32 to 32bit mode */
1884         a->write_bcr(ioaddr, 20, 2);
1885
1886         a->write_csr(ioaddr, 1, (lp->init_dma_addr & 0xffff));
1887         a->write_csr(ioaddr, 2, (lp->init_dma_addr >> 16));
1888
1889         if (pdev) {             /* use the IRQ provided by PCI */
1890                 dev->irq = pdev->irq;
1891                 if (pcnet32_debug & NETIF_MSG_PROBE)
1892                         printk(" assigned IRQ %d.\n", dev->irq);
1893         } else {
1894                 unsigned long irq_mask = probe_irq_on();
1895
1896                 /*
1897                  * To auto-IRQ we enable the initialization-done and DMA error
1898                  * interrupts. For ISA boards we get a DMA error, but VLB and PCI
1899                  * boards will work.
1900                  */
1901                 /* Trigger an initialization just for the interrupt. */
1902                 a->write_csr(ioaddr, CSR0, CSR0_INTEN | CSR0_INIT);
1903                 mdelay(1);
1904
1905                 dev->irq = probe_irq_off(irq_mask);
1906                 if (!dev->irq) {
1907                         if (pcnet32_debug & NETIF_MSG_PROBE)
1908                                 printk(", failed to detect IRQ line.\n");
1909                         ret = -ENODEV;
1910                         goto err_free_ring;
1911                 }
1912                 if (pcnet32_debug & NETIF_MSG_PROBE)
1913                         printk(", probed IRQ %d.\n", dev->irq);
1914         }
1915
1916         /* Set the mii phy_id so that we can query the link state */
1917         if (lp->mii) {
1918                 /* lp->phycount and lp->phymask are set to 0 by memset above */
1919
1920                 lp->mii_if.phy_id = ((lp->a.read_bcr(ioaddr, 33)) >> 5) & 0x1f;
1921                 /* scan for PHYs */
1922                 for (i = 0; i < PCNET32_MAX_PHYS; i++) {
1923                         unsigned short id1, id2;
1924
1925                         id1 = mdio_read(dev, i, MII_PHYSID1);
1926                         if (id1 == 0xffff)
1927                                 continue;
1928                         id2 = mdio_read(dev, i, MII_PHYSID2);
1929                         if (id2 == 0xffff)
1930                                 continue;
1931                         if (i == 31 && ((chip_version + 1) & 0xfffe) == 0x2624)
1932                                 continue;       /* 79C971 & 79C972 have phantom phy at id 31 */
1933                         lp->phycount++;
1934                         lp->phymask |= (1 << i);
1935                         lp->mii_if.phy_id = i;
1936                         if (pcnet32_debug & NETIF_MSG_PROBE)
1937                                 printk(KERN_INFO PFX
1938                                        "Found PHY %04x:%04x at address %d.\n",
1939                                        id1, id2, i);
1940                 }
1941                 lp->a.write_bcr(ioaddr, 33, (lp->mii_if.phy_id) << 5);
1942                 if (lp->phycount > 1) {
1943                         lp->options |= PCNET32_PORT_MII;
1944                 }
1945         }
1946
1947         init_timer(&lp->watchdog_timer);
1948         lp->watchdog_timer.data = (unsigned long)dev;
1949         lp->watchdog_timer.function = (void *)&pcnet32_watchdog;
1950
1951         /* The PCNET32-specific entries in the device structure. */
1952         dev->netdev_ops = &pcnet32_netdev_ops;
1953         dev->ethtool_ops = &pcnet32_ethtool_ops;
1954         dev->watchdog_timeo = (5 * HZ);
1955
1956         /* Fill in the generic fields of the device structure. */
1957         if (register_netdev(dev))
1958                 goto err_free_ring;
1959
1960         if (pdev) {
1961                 pci_set_drvdata(pdev, dev);
1962         } else {
1963                 lp->next = pcnet32_dev;
1964                 pcnet32_dev = dev;
1965         }
1966
1967         if (pcnet32_debug & NETIF_MSG_PROBE)
1968                 printk(KERN_INFO "%s: registered as %s\n", dev->name, lp->name);
1969         cards_found++;
1970
1971         /* enable LED writes */
1972         a->write_bcr(ioaddr, 2, a->read_bcr(ioaddr, 2) | 0x1000);
1973
1974         return 0;
1975
1976       err_free_ring:
1977         pcnet32_free_ring(dev);
1978       err_free_consistent:
1979         pci_free_consistent(lp->pci_dev, sizeof(*lp->init_block),
1980                             lp->init_block, lp->init_dma_addr);
1981       err_free_netdev:
1982         free_netdev(dev);
1983       err_release_region:
1984         release_region(ioaddr, PCNET32_TOTAL_SIZE);
1985         return ret;
1986 }
1987
1988 /* if any allocation fails, caller must also call pcnet32_free_ring */
1989 static int pcnet32_alloc_ring(struct net_device *dev, const char *name)
1990 {
1991         struct pcnet32_private *lp = netdev_priv(dev);
1992
1993         lp->tx_ring = pci_alloc_consistent(lp->pci_dev,
1994                                            sizeof(struct pcnet32_tx_head) *
1995                                            lp->tx_ring_size,
1996                                            &lp->tx_ring_dma_addr);
1997         if (lp->tx_ring == NULL) {
1998                 if (netif_msg_drv(lp))
1999                         printk("\n" KERN_ERR PFX
2000                                "%s: Consistent memory allocation failed.\n",
2001                                name);
2002                 return -ENOMEM;
2003         }
2004
2005         lp->rx_ring = pci_alloc_consistent(lp->pci_dev,
2006                                            sizeof(struct pcnet32_rx_head) *
2007                                            lp->rx_ring_size,
2008                                            &lp->rx_ring_dma_addr);
2009         if (lp->rx_ring == NULL) {
2010                 if (netif_msg_drv(lp))
2011                         printk("\n" KERN_ERR PFX
2012                                "%s: Consistent memory allocation failed.\n",
2013                                name);
2014                 return -ENOMEM;
2015         }
2016
2017         lp->tx_dma_addr = kcalloc(lp->tx_ring_size, sizeof(dma_addr_t),
2018                                   GFP_ATOMIC);
2019         if (!lp->tx_dma_addr) {
2020                 if (netif_msg_drv(lp))
2021                         printk("\n" KERN_ERR PFX
2022                                "%s: Memory allocation failed.\n", name);
2023                 return -ENOMEM;
2024         }
2025
2026         lp->rx_dma_addr = kcalloc(lp->rx_ring_size, sizeof(dma_addr_t),
2027                                   GFP_ATOMIC);
2028         if (!lp->rx_dma_addr) {
2029                 if (netif_msg_drv(lp))
2030                         printk("\n" KERN_ERR PFX
2031                                "%s: Memory allocation failed.\n", name);
2032                 return -ENOMEM;
2033         }
2034
2035         lp->tx_skbuff = kcalloc(lp->tx_ring_size, sizeof(struct sk_buff *),
2036                                 GFP_ATOMIC);
2037         if (!lp->tx_skbuff) {
2038                 if (netif_msg_drv(lp))
2039                         printk("\n" KERN_ERR PFX
2040                                "%s: Memory allocation failed.\n", name);
2041                 return -ENOMEM;
2042         }
2043
2044         lp->rx_skbuff = kcalloc(lp->rx_ring_size, sizeof(struct sk_buff *),
2045                                 GFP_ATOMIC);
2046         if (!lp->rx_skbuff) {
2047                 if (netif_msg_drv(lp))
2048                         printk("\n" KERN_ERR PFX
2049                                "%s: Memory allocation failed.\n", name);
2050                 return -ENOMEM;
2051         }
2052
2053         return 0;
2054 }
2055
2056 static void pcnet32_free_ring(struct net_device *dev)
2057 {
2058         struct pcnet32_private *lp = netdev_priv(dev);
2059
2060         kfree(lp->tx_skbuff);
2061         lp->tx_skbuff = NULL;
2062
2063         kfree(lp->rx_skbuff);
2064         lp->rx_skbuff = NULL;
2065
2066         kfree(lp->tx_dma_addr);
2067         lp->tx_dma_addr = NULL;
2068
2069         kfree(lp->rx_dma_addr);
2070         lp->rx_dma_addr = NULL;
2071
2072         if (lp->tx_ring) {
2073                 pci_free_consistent(lp->pci_dev,
2074                                     sizeof(struct pcnet32_tx_head) *
2075                                     lp->tx_ring_size, lp->tx_ring,
2076                                     lp->tx_ring_dma_addr);
2077                 lp->tx_ring = NULL;
2078         }
2079
2080         if (lp->rx_ring) {
2081                 pci_free_consistent(lp->pci_dev,
2082                                     sizeof(struct pcnet32_rx_head) *
2083                                     lp->rx_ring_size, lp->rx_ring,
2084                                     lp->rx_ring_dma_addr);
2085                 lp->rx_ring = NULL;
2086         }
2087 }
2088
2089 static int pcnet32_open(struct net_device *dev)
2090 {
2091         struct pcnet32_private *lp = netdev_priv(dev);
2092         unsigned long ioaddr = dev->base_addr;
2093         u16 val;
2094         int i;
2095         int rc;
2096         unsigned long flags;
2097
2098         if (request_irq(dev->irq, &pcnet32_interrupt,
2099                         lp->shared_irq ? IRQF_SHARED : 0, dev->name,
2100                         (void *)dev)) {
2101                 return -EAGAIN;
2102         }
2103
2104         spin_lock_irqsave(&lp->lock, flags);
2105         /* Check for a valid station address */
2106         if (!is_valid_ether_addr(dev->dev_addr)) {
2107                 rc = -EINVAL;
2108                 goto err_free_irq;
2109         }
2110
2111         /* Reset the PCNET32 */
2112         lp->a.reset(ioaddr);
2113
2114         /* switch pcnet32 to 32bit mode */
2115         lp->a.write_bcr(ioaddr, 20, 2);
2116
2117         if (netif_msg_ifup(lp))
2118                 printk(KERN_DEBUG
2119                        "%s: pcnet32_open() irq %d tx/rx rings %#x/%#x init %#x.\n",
2120                        dev->name, dev->irq, (u32) (lp->tx_ring_dma_addr),
2121                        (u32) (lp->rx_ring_dma_addr),
2122                        (u32) (lp->init_dma_addr));
2123
2124         /* set/reset autoselect bit */
2125         val = lp->a.read_bcr(ioaddr, 2) & ~2;
2126         if (lp->options & PCNET32_PORT_ASEL)
2127                 val |= 2;
2128         lp->a.write_bcr(ioaddr, 2, val);
2129
2130         /* handle full duplex setting */
2131         if (lp->mii_if.full_duplex) {
2132                 val = lp->a.read_bcr(ioaddr, 9) & ~3;
2133                 if (lp->options & PCNET32_PORT_FD) {
2134                         val |= 1;
2135                         if (lp->options == (PCNET32_PORT_FD | PCNET32_PORT_AUI))
2136                                 val |= 2;
2137                 } else if (lp->options & PCNET32_PORT_ASEL) {
2138                         /* workaround of xSeries250, turn on for 79C975 only */
2139                         if (lp->chip_version == 0x2627)
2140                                 val |= 3;
2141                 }
2142                 lp->a.write_bcr(ioaddr, 9, val);
2143         }
2144
2145         /* set/reset GPSI bit in test register */
2146         val = lp->a.read_csr(ioaddr, 124) & ~0x10;
2147         if ((lp->options & PCNET32_PORT_PORTSEL) == PCNET32_PORT_GPSI)
2148                 val |= 0x10;
2149         lp->a.write_csr(ioaddr, 124, val);
2150
2151         /* Allied Telesyn AT 2700/2701 FX are 100Mbit only and do not negotiate */
2152         if (lp->pci_dev->subsystem_vendor == PCI_VENDOR_ID_AT &&
2153             (lp->pci_dev->subsystem_device == PCI_SUBDEVICE_ID_AT_2700FX ||
2154              lp->pci_dev->subsystem_device == PCI_SUBDEVICE_ID_AT_2701FX)) {
2155                 if (lp->options & PCNET32_PORT_ASEL) {
2156                         lp->options = PCNET32_PORT_FD | PCNET32_PORT_100;
2157                         if (netif_msg_link(lp))
2158                                 printk(KERN_DEBUG
2159                                        "%s: Setting 100Mb-Full Duplex.\n",
2160                                        dev->name);
2161                 }
2162         }
2163         if (lp->phycount < 2) {
2164                 /*
2165                  * 24 Jun 2004 according AMD, in order to change the PHY,
2166                  * DANAS (or DISPM for 79C976) must be set; then select the speed,
2167                  * duplex, and/or enable auto negotiation, and clear DANAS
2168                  */
2169                 if (lp->mii && !(lp->options & PCNET32_PORT_ASEL)) {
2170                         lp->a.write_bcr(ioaddr, 32,
2171                                         lp->a.read_bcr(ioaddr, 32) | 0x0080);
2172                         /* disable Auto Negotiation, set 10Mpbs, HD */
2173                         val = lp->a.read_bcr(ioaddr, 32) & ~0xb8;
2174                         if (lp->options & PCNET32_PORT_FD)
2175                                 val |= 0x10;
2176                         if (lp->options & PCNET32_PORT_100)
2177                                 val |= 0x08;
2178                         lp->a.write_bcr(ioaddr, 32, val);
2179                 } else {
2180                         if (lp->options & PCNET32_PORT_ASEL) {
2181                                 lp->a.write_bcr(ioaddr, 32,
2182                                                 lp->a.read_bcr(ioaddr,
2183                                                                32) | 0x0080);
2184                                 /* enable auto negotiate, setup, disable fd */
2185                                 val = lp->a.read_bcr(ioaddr, 32) & ~0x98;
2186                                 val |= 0x20;
2187                                 lp->a.write_bcr(ioaddr, 32, val);
2188                         }
2189                 }
2190         } else {
2191                 int first_phy = -1;
2192                 u16 bmcr;
2193                 u32 bcr9;
2194                 struct ethtool_cmd ecmd;
2195
2196                 /*
2197                  * There is really no good other way to handle multiple PHYs
2198                  * other than turning off all automatics
2199                  */
2200                 val = lp->a.read_bcr(ioaddr, 2);
2201                 lp->a.write_bcr(ioaddr, 2, val & ~2);
2202                 val = lp->a.read_bcr(ioaddr, 32);
2203                 lp->a.write_bcr(ioaddr, 32, val & ~(1 << 7));   /* stop MII manager */
2204
2205                 if (!(lp->options & PCNET32_PORT_ASEL)) {
2206                         /* setup ecmd */
2207                         ecmd.port = PORT_MII;
2208                         ecmd.transceiver = XCVR_INTERNAL;
2209                         ecmd.autoneg = AUTONEG_DISABLE;
2210                         ecmd.speed =
2211                             lp->
2212                             options & PCNET32_PORT_100 ? SPEED_100 : SPEED_10;
2213                         bcr9 = lp->a.read_bcr(ioaddr, 9);
2214
2215                         if (lp->options & PCNET32_PORT_FD) {
2216                                 ecmd.duplex = DUPLEX_FULL;
2217                                 bcr9 |= (1 << 0);
2218                         } else {
2219                                 ecmd.duplex = DUPLEX_HALF;
2220                                 bcr9 |= ~(1 << 0);
2221                         }
2222                         lp->a.write_bcr(ioaddr, 9, bcr9);
2223                 }
2224
2225                 for (i = 0; i < PCNET32_MAX_PHYS; i++) {
2226                         if (lp->phymask & (1 << i)) {
2227                                 /* isolate all but the first PHY */
2228                                 bmcr = mdio_read(dev, i, MII_BMCR);
2229                                 if (first_phy == -1) {
2230                                         first_phy = i;
2231                                         mdio_write(dev, i, MII_BMCR,
2232                                                    bmcr & ~BMCR_ISOLATE);
2233                                 } else {
2234                                         mdio_write(dev, i, MII_BMCR,
2235                                                    bmcr | BMCR_ISOLATE);
2236                                 }
2237                                 /* use mii_ethtool_sset to setup PHY */
2238                                 lp->mii_if.phy_id = i;
2239                                 ecmd.phy_address = i;
2240                                 if (lp->options & PCNET32_PORT_ASEL) {
2241                                         mii_ethtool_gset(&lp->mii_if, &ecmd);
2242                                         ecmd.autoneg = AUTONEG_ENABLE;
2243                                 }
2244                                 mii_ethtool_sset(&lp->mii_if, &ecmd);
2245                         }
2246                 }
2247                 lp->mii_if.phy_id = first_phy;
2248                 if (netif_msg_link(lp))
2249                         printk(KERN_INFO "%s: Using PHY number %d.\n",
2250                                dev->name, first_phy);
2251         }
2252
2253 #ifdef DO_DXSUFLO
2254         if (lp->dxsuflo) {      /* Disable transmit stop on underflow */
2255                 val = lp->a.read_csr(ioaddr, CSR3);
2256                 val |= 0x40;
2257                 lp->a.write_csr(ioaddr, CSR3, val);
2258         }
2259 #endif
2260
2261         lp->init_block->mode =
2262             cpu_to_le16((lp->options & PCNET32_PORT_PORTSEL) << 7);
2263         pcnet32_load_multicast(dev);
2264
2265         if (pcnet32_init_ring(dev)) {
2266                 rc = -ENOMEM;
2267                 goto err_free_ring;
2268         }
2269
2270         napi_enable(&lp->napi);
2271
2272         /* Re-initialize the PCNET32, and start it when done. */
2273         lp->a.write_csr(ioaddr, 1, (lp->init_dma_addr & 0xffff));
2274         lp->a.write_csr(ioaddr, 2, (lp->init_dma_addr >> 16));
2275
2276         lp->a.write_csr(ioaddr, CSR4, 0x0915);  /* auto tx pad */
2277         lp->a.write_csr(ioaddr, CSR0, CSR0_INIT);
2278
2279         netif_start_queue(dev);
2280
2281         if (lp->chip_version >= PCNET32_79C970A) {
2282                 /* Print the link status and start the watchdog */
2283                 pcnet32_check_media(dev, 1);
2284                 mod_timer(&lp->watchdog_timer, PCNET32_WATCHDOG_TIMEOUT);
2285         }
2286
2287         i = 0;
2288         while (i++ < 100)
2289                 if (lp->a.read_csr(ioaddr, CSR0) & CSR0_IDON)
2290                         break;
2291         /*
2292          * We used to clear the InitDone bit, 0x0100, here but Mark Stockton
2293          * reports that doing so triggers a bug in the '974.
2294          */
2295         lp->a.write_csr(ioaddr, CSR0, CSR0_NORMAL);
2296
2297         if (netif_msg_ifup(lp))
2298                 printk(KERN_DEBUG
2299                        "%s: pcnet32 open after %d ticks, init block %#x csr0 %4.4x.\n",
2300                        dev->name, i,
2301                        (u32) (lp->init_dma_addr),
2302                        lp->a.read_csr(ioaddr, CSR0));
2303
2304         spin_unlock_irqrestore(&lp->lock, flags);
2305
2306         return 0;               /* Always succeed */
2307
2308       err_free_ring:
2309         /* free any allocated skbuffs */
2310         pcnet32_purge_rx_ring(dev);
2311
2312         /*
2313          * Switch back to 16bit mode to avoid problems with dumb
2314          * DOS packet driver after a warm reboot
2315          */
2316         lp->a.write_bcr(ioaddr, 20, 4);
2317
2318       err_free_irq:
2319         spin_unlock_irqrestore(&lp->lock, flags);
2320         free_irq(dev->irq, dev);
2321         return rc;
2322 }
2323
2324 /*
2325  * The LANCE has been halted for one reason or another (busmaster memory
2326  * arbitration error, Tx FIFO underflow, driver stopped it to reconfigure,
2327  * etc.).  Modern LANCE variants always reload their ring-buffer
2328  * configuration when restarted, so we must reinitialize our ring
2329  * context before restarting.  As part of this reinitialization,
2330  * find all packets still on the Tx ring and pretend that they had been
2331  * sent (in effect, drop the packets on the floor) - the higher-level
2332  * protocols will time out and retransmit.  It'd be better to shuffle
2333  * these skbs to a temp list and then actually re-Tx them after
2334  * restarting the chip, but I'm too lazy to do so right now.  dplatt@3do.com
2335  */
2336
2337 static void pcnet32_purge_tx_ring(struct net_device *dev)
2338 {
2339         struct pcnet32_private *lp = netdev_priv(dev);
2340         int i;
2341
2342         for (i = 0; i < lp->tx_ring_size; i++) {
2343                 lp->tx_ring[i].status = 0;      /* CPU owns buffer */
2344                 wmb();          /* Make sure adapter sees owner change */
2345                 if (lp->tx_skbuff[i]) {
2346                         pci_unmap_single(lp->pci_dev, lp->tx_dma_addr[i],
2347                                          lp->tx_skbuff[i]->len,
2348                                          PCI_DMA_TODEVICE);
2349                         dev_kfree_skb_any(lp->tx_skbuff[i]);
2350                 }
2351                 lp->tx_skbuff[i] = NULL;
2352                 lp->tx_dma_addr[i] = 0;
2353         }
2354 }
2355
2356 /* Initialize the PCNET32 Rx and Tx rings. */
2357 static int pcnet32_init_ring(struct net_device *dev)
2358 {
2359         struct pcnet32_private *lp = netdev_priv(dev);
2360         int i;
2361
2362         lp->tx_full = 0;
2363         lp->cur_rx = lp->cur_tx = 0;
2364         lp->dirty_rx = lp->dirty_tx = 0;
2365
2366         for (i = 0; i < lp->rx_ring_size; i++) {
2367                 struct sk_buff *rx_skbuff = lp->rx_skbuff[i];
2368                 if (rx_skbuff == NULL) {
2369                         if (!
2370                             (rx_skbuff = lp->rx_skbuff[i] =
2371                              dev_alloc_skb(PKT_BUF_SKB))) {
2372                                 /* there is not much, we can do at this point */
2373                                 if (netif_msg_drv(lp))
2374                                         printk(KERN_ERR
2375                                                "%s: pcnet32_init_ring dev_alloc_skb failed.\n",
2376                                                dev->name);
2377                                 return -1;
2378                         }
2379                         skb_reserve(rx_skbuff, NET_IP_ALIGN);
2380                 }
2381
2382                 rmb();
2383                 if (lp->rx_dma_addr[i] == 0)
2384                         lp->rx_dma_addr[i] =
2385                             pci_map_single(lp->pci_dev, rx_skbuff->data,
2386                                            PKT_BUF_SIZE, PCI_DMA_FROMDEVICE);
2387                 lp->rx_ring[i].base = cpu_to_le32(lp->rx_dma_addr[i]);
2388                 lp->rx_ring[i].buf_length = cpu_to_le16(NEG_BUF_SIZE);
2389                 wmb();          /* Make sure owner changes after all others are visible */
2390                 lp->rx_ring[i].status = cpu_to_le16(0x8000);
2391         }
2392         /* The Tx buffer address is filled in as needed, but we do need to clear
2393          * the upper ownership bit. */
2394         for (i = 0; i < lp->tx_ring_size; i++) {
2395                 lp->tx_ring[i].status = 0;      /* CPU owns buffer */
2396                 wmb();          /* Make sure adapter sees owner change */
2397                 lp->tx_ring[i].base = 0;
2398                 lp->tx_dma_addr[i] = 0;
2399         }
2400
2401         lp->init_block->tlen_rlen =
2402             cpu_to_le16(lp->tx_len_bits | lp->rx_len_bits);
2403         for (i = 0; i < 6; i++)
2404                 lp->init_block->phys_addr[i] = dev->dev_addr[i];
2405         lp->init_block->rx_ring = cpu_to_le32(lp->rx_ring_dma_addr);
2406         lp->init_block->tx_ring = cpu_to_le32(lp->tx_ring_dma_addr);
2407         wmb();                  /* Make sure all changes are visible */
2408         return 0;
2409 }
2410
2411 /* the pcnet32 has been issued a stop or reset.  Wait for the stop bit
2412  * then flush the pending transmit operations, re-initialize the ring,
2413  * and tell the chip to initialize.
2414  */
2415 static void pcnet32_restart(struct net_device *dev, unsigned int csr0_bits)
2416 {
2417         struct pcnet32_private *lp = netdev_priv(dev);
2418         unsigned long ioaddr = dev->base_addr;
2419         int i;
2420
2421         /* wait for stop */
2422         for (i = 0; i < 100; i++)
2423                 if (lp->a.read_csr(ioaddr, CSR0) & CSR0_STOP)
2424                         break;
2425
2426         if (i >= 100 && netif_msg_drv(lp))
2427                 printk(KERN_ERR
2428                        "%s: pcnet32_restart timed out waiting for stop.\n",
2429                        dev->name);
2430
2431         pcnet32_purge_tx_ring(dev);
2432         if (pcnet32_init_ring(dev))
2433                 return;
2434
2435         /* ReInit Ring */
2436         lp->a.write_csr(ioaddr, CSR0, CSR0_INIT);
2437         i = 0;
2438         while (i++ < 1000)
2439                 if (lp->a.read_csr(ioaddr, CSR0) & CSR0_IDON)
2440                         break;
2441
2442         lp->a.write_csr(ioaddr, CSR0, csr0_bits);
2443 }
2444
2445 static void pcnet32_tx_timeout(struct net_device *dev)
2446 {
2447         struct pcnet32_private *lp = netdev_priv(dev);
2448         unsigned long ioaddr = dev->base_addr, flags;
2449
2450         spin_lock_irqsave(&lp->lock, flags);
2451         /* Transmitter timeout, serious problems. */
2452         if (pcnet32_debug & NETIF_MSG_DRV)
2453                 printk(KERN_ERR
2454                        "%s: transmit timed out, status %4.4x, resetting.\n",
2455                        dev->name, lp->a.read_csr(ioaddr, CSR0));
2456         lp->a.write_csr(ioaddr, CSR0, CSR0_STOP);
2457         dev->stats.tx_errors++;
2458         if (netif_msg_tx_err(lp)) {
2459                 int i;
2460                 printk(KERN_DEBUG
2461                        " Ring data dump: dirty_tx %d cur_tx %d%s cur_rx %d.",
2462                        lp->dirty_tx, lp->cur_tx, lp->tx_full ? " (full)" : "",
2463                        lp->cur_rx);
2464                 for (i = 0; i < lp->rx_ring_size; i++)
2465                         printk("%s %08x %04x %08x %04x", i & 1 ? "" : "\n ",
2466                                le32_to_cpu(lp->rx_ring[i].base),
2467                                (-le16_to_cpu(lp->rx_ring[i].buf_length)) &
2468                                0xffff, le32_to_cpu(lp->rx_ring[i].msg_length),
2469                                le16_to_cpu(lp->rx_ring[i].status));
2470                 for (i = 0; i < lp->tx_ring_size; i++)
2471                         printk("%s %08x %04x %08x %04x", i & 1 ? "" : "\n ",
2472                                le32_to_cpu(lp->tx_ring[i].base),
2473                                (-le16_to_cpu(lp->tx_ring[i].length)) & 0xffff,
2474                                le32_to_cpu(lp->tx_ring[i].misc),
2475                                le16_to_cpu(lp->tx_ring[i].status));
2476                 printk("\n");
2477         }
2478         pcnet32_restart(dev, CSR0_NORMAL);
2479
2480         dev->trans_start = jiffies;
2481         netif_wake_queue(dev);
2482
2483         spin_unlock_irqrestore(&lp->lock, flags);
2484 }
2485
2486 static int pcnet32_start_xmit(struct sk_buff *skb, struct net_device *dev)
2487 {
2488         struct pcnet32_private *lp = netdev_priv(dev);
2489         unsigned long ioaddr = dev->base_addr;
2490         u16 status;
2491         int entry;
2492         unsigned long flags;
2493
2494         spin_lock_irqsave(&lp->lock, flags);
2495
2496         if (netif_msg_tx_queued(lp)) {
2497                 printk(KERN_DEBUG
2498                        "%s: pcnet32_start_xmit() called, csr0 %4.4x.\n",
2499                        dev->name, lp->a.read_csr(ioaddr, CSR0));
2500         }
2501
2502         /* Default status -- will not enable Successful-TxDone
2503          * interrupt when that option is available to us.
2504          */
2505         status = 0x8300;
2506
2507         /* Fill in a Tx ring entry */
2508
2509         /* Mask to ring buffer boundary. */
2510         entry = lp->cur_tx & lp->tx_mod_mask;
2511
2512         /* Caution: the write order is important here, set the status
2513          * with the "ownership" bits last. */
2514
2515         lp->tx_ring[entry].length = cpu_to_le16(-skb->len);
2516
2517         lp->tx_ring[entry].misc = 0x00000000;
2518
2519         lp->tx_skbuff[entry] = skb;
2520         lp->tx_dma_addr[entry] =
2521             pci_map_single(lp->pci_dev, skb->data, skb->len, PCI_DMA_TODEVICE);
2522         lp->tx_ring[entry].base = cpu_to_le32(lp->tx_dma_addr[entry]);
2523         wmb();                  /* Make sure owner changes after all others are visible */
2524         lp->tx_ring[entry].status = cpu_to_le16(status);
2525
2526         lp->cur_tx++;
2527         dev->stats.tx_bytes += skb->len;
2528
2529         /* Trigger an immediate send poll. */
2530         lp->a.write_csr(ioaddr, CSR0, CSR0_INTEN | CSR0_TXPOLL);
2531
2532         dev->trans_start = jiffies;
2533
2534         if (lp->tx_ring[(entry + 1) & lp->tx_mod_mask].base != 0) {
2535                 lp->tx_full = 1;
2536                 netif_stop_queue(dev);
2537         }
2538         spin_unlock_irqrestore(&lp->lock, flags);
2539         return 0;
2540 }
2541
2542 /* The PCNET32 interrupt handler. */
2543 static irqreturn_t
2544 pcnet32_interrupt(int irq, void *dev_id)
2545 {
2546         struct net_device *dev = dev_id;
2547         struct pcnet32_private *lp;
2548         unsigned long ioaddr;
2549         u16 csr0;
2550         int boguscnt = max_interrupt_work;
2551
2552         ioaddr = dev->base_addr;
2553         lp = netdev_priv(dev);
2554
2555         spin_lock(&lp->lock);
2556
2557         csr0 = lp->a.read_csr(ioaddr, CSR0);
2558         while ((csr0 & 0x8f00) && --boguscnt >= 0) {
2559                 if (csr0 == 0xffff) {
2560                         break;  /* PCMCIA remove happened */
2561                 }
2562                 /* Acknowledge all of the current interrupt sources ASAP. */
2563                 lp->a.write_csr(ioaddr, CSR0, csr0 & ~0x004f);
2564
2565                 if (netif_msg_intr(lp))
2566                         printk(KERN_DEBUG
2567                                "%s: interrupt  csr0=%#2.2x new csr=%#2.2x.\n",
2568                                dev->name, csr0, lp->a.read_csr(ioaddr, CSR0));
2569
2570                 /* Log misc errors. */
2571                 if (csr0 & 0x4000)
2572                         dev->stats.tx_errors++; /* Tx babble. */
2573                 if (csr0 & 0x1000) {
2574                         /*
2575                          * This happens when our receive ring is full. This
2576                          * shouldn't be a problem as we will see normal rx
2577                          * interrupts for the frames in the receive ring.  But
2578                          * there are some PCI chipsets (I can reproduce this
2579                          * on SP3G with Intel saturn chipset) which have
2580                          * sometimes problems and will fill up the receive
2581                          * ring with error descriptors.  In this situation we
2582                          * don't get a rx interrupt, but a missed frame
2583                          * interrupt sooner or later.
2584                          */
2585                         dev->stats.rx_errors++; /* Missed a Rx frame. */
2586                 }
2587                 if (csr0 & 0x0800) {
2588                         if (netif_msg_drv(lp))
2589                                 printk(KERN_ERR
2590                                        "%s: Bus master arbitration failure, status %4.4x.\n",
2591                                        dev->name, csr0);
2592                         /* unlike for the lance, there is no restart needed */
2593                 }
2594                 if (napi_schedule_prep(&lp->napi)) {
2595                         u16 val;
2596                         /* set interrupt masks */
2597                         val = lp->a.read_csr(ioaddr, CSR3);
2598                         val |= 0x5f00;
2599                         lp->a.write_csr(ioaddr, CSR3, val);
2600
2601                         __napi_schedule(&lp->napi);
2602                         break;
2603                 }
2604                 csr0 = lp->a.read_csr(ioaddr, CSR0);
2605         }
2606
2607         if (netif_msg_intr(lp))
2608                 printk(KERN_DEBUG "%s: exiting interrupt, csr0=%#4.4x.\n",
2609                        dev->name, lp->a.read_csr(ioaddr, CSR0));
2610
2611         spin_unlock(&lp->lock);
2612
2613         return IRQ_HANDLED;
2614 }
2615
2616 static int pcnet32_close(struct net_device *dev)
2617 {
2618         unsigned long ioaddr = dev->base_addr;
2619         struct pcnet32_private *lp = netdev_priv(dev);
2620         unsigned long flags;
2621
2622         del_timer_sync(&lp->watchdog_timer);
2623
2624         netif_stop_queue(dev);
2625         napi_disable(&lp->napi);
2626
2627         spin_lock_irqsave(&lp->lock, flags);
2628
2629         dev->stats.rx_missed_errors = lp->a.read_csr(ioaddr, 112);
2630
2631         if (netif_msg_ifdown(lp))
2632                 printk(KERN_DEBUG
2633                        "%s: Shutting down ethercard, status was %2.2x.\n",
2634                        dev->name, lp->a.read_csr(ioaddr, CSR0));
2635
2636         /* We stop the PCNET32 here -- it occasionally polls memory if we don't. */
2637         lp->a.write_csr(ioaddr, CSR0, CSR0_STOP);
2638
2639         /*
2640          * Switch back to 16bit mode to avoid problems with dumb
2641          * DOS packet driver after a warm reboot
2642          */
2643         lp->a.write_bcr(ioaddr, 20, 4);
2644
2645         spin_unlock_irqrestore(&lp->lock, flags);
2646
2647         free_irq(dev->irq, dev);
2648
2649         spin_lock_irqsave(&lp->lock, flags);
2650
2651         pcnet32_purge_rx_ring(dev);
2652         pcnet32_purge_tx_ring(dev);
2653
2654         spin_unlock_irqrestore(&lp->lock, flags);
2655
2656         return 0;
2657 }
2658
2659 static struct net_device_stats *pcnet32_get_stats(struct net_device *dev)
2660 {
2661         struct pcnet32_private *lp = netdev_priv(dev);
2662         unsigned long ioaddr = dev->base_addr;
2663         unsigned long flags;
2664
2665         spin_lock_irqsave(&lp->lock, flags);
2666         dev->stats.rx_missed_errors = lp->a.read_csr(ioaddr, 112);
2667         spin_unlock_irqrestore(&lp->lock, flags);
2668
2669         return &dev->stats;
2670 }
2671
2672 /* taken from the sunlance driver, which it took from the depca driver */
2673 static void pcnet32_load_multicast(struct net_device *dev)
2674 {
2675         struct pcnet32_private *lp = netdev_priv(dev);
2676         volatile struct pcnet32_init_block *ib = lp->init_block;
2677         volatile __le16 *mcast_table = (__le16 *)ib->filter;
2678         struct dev_mc_list *dmi = dev->mc_list;
2679         unsigned long ioaddr = dev->base_addr;
2680         char *addrs;
2681         int i;
2682         u32 crc;
2683
2684         /* set all multicast bits */
2685         if (dev->flags & IFF_ALLMULTI) {
2686                 ib->filter[0] = cpu_to_le32(~0U);
2687                 ib->filter[1] = cpu_to_le32(~0U);
2688                 lp->a.write_csr(ioaddr, PCNET32_MC_FILTER, 0xffff);
2689                 lp->a.write_csr(ioaddr, PCNET32_MC_FILTER+1, 0xffff);
2690                 lp->a.write_csr(ioaddr, PCNET32_MC_FILTER+2, 0xffff);
2691                 lp->a.write_csr(ioaddr, PCNET32_MC_FILTER+3, 0xffff);
2692                 return;
2693         }
2694         /* clear the multicast filter */
2695         ib->filter[0] = 0;
2696         ib->filter[1] = 0;
2697
2698         /* Add addresses */
2699         for (i = 0; i < dev->mc_count; i++) {
2700                 addrs = dmi->dmi_addr;
2701                 dmi = dmi->next;
2702
2703                 /* multicast address? */
2704                 if (!(*addrs & 1))
2705                         continue;
2706
2707                 crc = ether_crc_le(6, addrs);
2708                 crc = crc >> 26;
2709                 mcast_table[crc >> 4] |= cpu_to_le16(1 << (crc & 0xf));
2710         }
2711         for (i = 0; i < 4; i++)
2712                 lp->a.write_csr(ioaddr, PCNET32_MC_FILTER + i,
2713                                 le16_to_cpu(mcast_table[i]));
2714         return;
2715 }
2716
2717 /*
2718  * Set or clear the multicast filter for this adaptor.
2719  */
2720 static void pcnet32_set_multicast_list(struct net_device *dev)
2721 {
2722         unsigned long ioaddr = dev->base_addr, flags;
2723         struct pcnet32_private *lp = netdev_priv(dev);
2724         int csr15, suspended;
2725
2726         spin_lock_irqsave(&lp->lock, flags);
2727         suspended = pcnet32_suspend(dev, &flags, 0);
2728         csr15 = lp->a.read_csr(ioaddr, CSR15);
2729         if (dev->flags & IFF_PROMISC) {
2730                 /* Log any net taps. */
2731                 if (netif_msg_hw(lp))
2732                         printk(KERN_INFO "%s: Promiscuous mode enabled.\n",
2733                                dev->name);
2734                 lp->init_block->mode =
2735                     cpu_to_le16(0x8000 | (lp->options & PCNET32_PORT_PORTSEL) <<
2736                                 7);
2737                 lp->a.write_csr(ioaddr, CSR15, csr15 | 0x8000);
2738         } else {
2739                 lp->init_block->mode =
2740                     cpu_to_le16((lp->options & PCNET32_PORT_PORTSEL) << 7);
2741                 lp->a.write_csr(ioaddr, CSR15, csr15 & 0x7fff);
2742                 pcnet32_load_multicast(dev);
2743         }
2744
2745         if (suspended) {
2746                 int csr5;
2747                 /* clear SUSPEND (SPND) - CSR5 bit 0 */
2748                 csr5 = lp->a.read_csr(ioaddr, CSR5);
2749                 lp->a.write_csr(ioaddr, CSR5, csr5 & (~CSR5_SUSPEND));
2750         } else {
2751                 lp->a.write_csr(ioaddr, CSR0, CSR0_STOP);
2752                 pcnet32_restart(dev, CSR0_NORMAL);
2753                 netif_wake_queue(dev);
2754         }
2755
2756         spin_unlock_irqrestore(&lp->lock, flags);
2757 }
2758
2759 /* This routine assumes that the lp->lock is held */
2760 static int mdio_read(struct net_device *dev, int phy_id, int reg_num)
2761 {
2762         struct pcnet32_private *lp = netdev_priv(dev);
2763         unsigned long ioaddr = dev->base_addr;
2764         u16 val_out;
2765
2766         if (!lp->mii)
2767                 return 0;
2768
2769         lp->a.write_bcr(ioaddr, 33, ((phy_id & 0x1f) << 5) | (reg_num & 0x1f));
2770         val_out = lp->a.read_bcr(ioaddr, 34);
2771
2772         return val_out;
2773 }
2774
2775 /* This routine assumes that the lp->lock is held */
2776 static void mdio_write(struct net_device *dev, int phy_id, int reg_num, int val)
2777 {
2778         struct pcnet32_private *lp = netdev_priv(dev);
2779         unsigned long ioaddr = dev->base_addr;
2780
2781         if (!lp->mii)
2782                 return;
2783
2784         lp->a.write_bcr(ioaddr, 33, ((phy_id & 0x1f) << 5) | (reg_num & 0x1f));
2785         lp->a.write_bcr(ioaddr, 34, val);
2786 }
2787
2788 static int pcnet32_ioctl(struct net_device *dev, struct ifreq *rq, int cmd)
2789 {
2790         struct pcnet32_private *lp = netdev_priv(dev);
2791         int rc;
2792         unsigned long flags;
2793
2794         /* SIOC[GS]MIIxxx ioctls */
2795         if (lp->mii) {
2796                 spin_lock_irqsave(&lp->lock, flags);
2797                 rc = generic_mii_ioctl(&lp->mii_if, if_mii(rq), cmd, NULL);
2798                 spin_unlock_irqrestore(&lp->lock, flags);
2799         } else {
2800                 rc = -EOPNOTSUPP;
2801         }
2802
2803         return rc;
2804 }
2805
2806 static int pcnet32_check_otherphy(struct net_device *dev)
2807 {
2808         struct pcnet32_private *lp = netdev_priv(dev);
2809         struct mii_if_info mii = lp->mii_if;
2810         u16 bmcr;
2811         int i;
2812
2813         for (i = 0; i < PCNET32_MAX_PHYS; i++) {
2814                 if (i == lp->mii_if.phy_id)
2815                         continue;       /* skip active phy */
2816                 if (lp->phymask & (1 << i)) {
2817                         mii.phy_id = i;
2818                         if (mii_link_ok(&mii)) {
2819                                 /* found PHY with active link */
2820                                 if (netif_msg_link(lp))
2821                                         printk(KERN_INFO
2822                                                "%s: Using PHY number %d.\n",
2823                                                dev->name, i);
2824
2825                                 /* isolate inactive phy */
2826                                 bmcr =
2827                                     mdio_read(dev, lp->mii_if.phy_id, MII_BMCR);
2828                                 mdio_write(dev, lp->mii_if.phy_id, MII_BMCR,
2829                                            bmcr | BMCR_ISOLATE);
2830
2831                                 /* de-isolate new phy */
2832                                 bmcr = mdio_read(dev, i, MII_BMCR);
2833                                 mdio_write(dev, i, MII_BMCR,
2834                                            bmcr & ~BMCR_ISOLATE);
2835
2836                                 /* set new phy address */
2837                                 lp->mii_if.phy_id = i;
2838                                 return 1;
2839                         }
2840                 }
2841         }
2842         return 0;
2843 }
2844
2845 /*
2846  * Show the status of the media.  Similar to mii_check_media however it
2847  * correctly shows the link speed for all (tested) pcnet32 variants.
2848  * Devices with no mii just report link state without speed.
2849  *
2850  * Caller is assumed to hold and release the lp->lock.
2851  */
2852
2853 static void pcnet32_check_media(struct net_device *dev, int verbose)
2854 {
2855         struct pcnet32_private *lp = netdev_priv(dev);
2856         int curr_link;
2857         int prev_link = netif_carrier_ok(dev) ? 1 : 0;
2858         u32 bcr9;
2859
2860         if (lp->mii) {
2861                 curr_link = mii_link_ok(&lp->mii_if);
2862         } else {
2863                 ulong ioaddr = dev->base_addr;  /* card base I/O address */
2864                 curr_link = (lp->a.read_bcr(ioaddr, 4) != 0xc0);
2865         }
2866         if (!curr_link) {
2867                 if (prev_link || verbose) {
2868                         netif_carrier_off(dev);
2869                         if (netif_msg_link(lp))
2870                                 printk(KERN_INFO "%s: link down\n", dev->name);
2871                 }
2872                 if (lp->phycount > 1) {
2873                         curr_link = pcnet32_check_otherphy(dev);
2874                         prev_link = 0;
2875                 }
2876         } else if (verbose || !prev_link) {
2877                 netif_carrier_on(dev);
2878                 if (lp->mii) {
2879                         if (netif_msg_link(lp)) {
2880                                 struct ethtool_cmd ecmd;
2881                                 mii_ethtool_gset(&lp->mii_if, &ecmd);
2882                                 printk(KERN_INFO
2883                                        "%s: link up, %sMbps, %s-duplex\n",
2884                                        dev->name,
2885                                        (ecmd.speed == SPEED_100) ? "100" : "10",
2886                                        (ecmd.duplex ==
2887                                         DUPLEX_FULL) ? "full" : "half");
2888                         }
2889                         bcr9 = lp->a.read_bcr(dev->base_addr, 9);
2890                         if ((bcr9 & (1 << 0)) != lp->mii_if.full_duplex) {
2891                                 if (lp->mii_if.full_duplex)
2892                                         bcr9 |= (1 << 0);
2893                                 else
2894                                         bcr9 &= ~(1 << 0);
2895                                 lp->a.write_bcr(dev->base_addr, 9, bcr9);
2896                         }
2897                 } else {
2898                         if (netif_msg_link(lp))
2899                                 printk(KERN_INFO "%s: link up\n", dev->name);
2900                 }
2901         }
2902 }
2903
2904 /*
2905  * Check for loss of link and link establishment.
2906  * Can not use mii_check_media because it does nothing if mode is forced.
2907  */
2908
2909 static void pcnet32_watchdog(struct net_device *dev)
2910 {
2911         struct pcnet32_private *lp = netdev_priv(dev);
2912         unsigned long flags;
2913
2914         /* Print the link status if it has changed */
2915         spin_lock_irqsave(&lp->lock, flags);
2916         pcnet32_check_media(dev, 0);
2917         spin_unlock_irqrestore(&lp->lock, flags);
2918
2919         mod_timer(&lp->watchdog_timer, round_jiffies(PCNET32_WATCHDOG_TIMEOUT));
2920 }
2921
2922 static int pcnet32_pm_suspend(struct pci_dev *pdev, pm_message_t state)
2923 {
2924         struct net_device *dev = pci_get_drvdata(pdev);
2925
2926         if (netif_running(dev)) {
2927                 netif_device_detach(dev);
2928                 pcnet32_close(dev);
2929         }
2930         pci_save_state(pdev);
2931         pci_set_power_state(pdev, pci_choose_state(pdev, state));
2932         return 0;
2933 }
2934
2935 static int pcnet32_pm_resume(struct pci_dev *pdev)
2936 {
2937         struct net_device *dev = pci_get_drvdata(pdev);
2938
2939         pci_set_power_state(pdev, PCI_D0);
2940         pci_restore_state(pdev);
2941
2942         if (netif_running(dev)) {
2943                 pcnet32_open(dev);
2944                 netif_device_attach(dev);
2945         }
2946         return 0;
2947 }
2948
2949 static void __devexit pcnet32_remove_one(struct pci_dev *pdev)
2950 {
2951         struct net_device *dev = pci_get_drvdata(pdev);
2952
2953         if (dev) {
2954                 struct pcnet32_private *lp = netdev_priv(dev);
2955
2956                 unregister_netdev(dev);
2957                 pcnet32_free_ring(dev);
2958                 release_region(dev->base_addr, PCNET32_TOTAL_SIZE);
2959                 pci_free_consistent(lp->pci_dev, sizeof(*lp->init_block),
2960                                     lp->init_block, lp->init_dma_addr);
2961                 free_netdev(dev);
2962                 pci_disable_device(pdev);
2963                 pci_set_drvdata(pdev, NULL);
2964         }
2965 }
2966
2967 static struct pci_driver pcnet32_driver = {
2968         .name = DRV_NAME,
2969         .probe = pcnet32_probe_pci,
2970         .remove = __devexit_p(pcnet32_remove_one),
2971         .id_table = pcnet32_pci_tbl,
2972         .suspend = pcnet32_pm_suspend,
2973         .resume = pcnet32_pm_resume,
2974 };
2975
2976 /* An additional parameter that may be passed in... */
2977 static int debug = -1;
2978 static int tx_start_pt = -1;
2979 static int pcnet32_have_pci;
2980
2981 module_param(debug, int, 0);
2982 MODULE_PARM_DESC(debug, DRV_NAME " debug level");
2983 module_param(max_interrupt_work, int, 0);
2984 MODULE_PARM_DESC(max_interrupt_work,
2985                  DRV_NAME " maximum events handled per interrupt");
2986 module_param(rx_copybreak, int, 0);
2987 MODULE_PARM_DESC(rx_copybreak,
2988                  DRV_NAME " copy breakpoint for copy-only-tiny-frames");
2989 module_param(tx_start_pt, int, 0);
2990 MODULE_PARM_DESC(tx_start_pt, DRV_NAME " transmit start point (0-3)");
2991 module_param(pcnet32vlb, int, 0);
2992 MODULE_PARM_DESC(pcnet32vlb, DRV_NAME " Vesa local bus (VLB) support (0/1)");
2993 module_param_array(options, int, NULL, 0);
2994 MODULE_PARM_DESC(options, DRV_NAME " initial option setting(s) (0-15)");
2995 module_param_array(full_duplex, int, NULL, 0);
2996 MODULE_PARM_DESC(full_duplex, DRV_NAME " full duplex setting(s) (1)");
2997 /* Module Parameter for HomePNA cards added by Patrick Simmons, 2004 */
2998 module_param_array(homepna, int, NULL, 0);
2999 MODULE_PARM_DESC(homepna,
3000                  DRV_NAME
3001                  " mode for 79C978 cards (1 for HomePNA, 0 for Ethernet, default Ethernet");
3002
3003 MODULE_AUTHOR("Thomas Bogendoerfer");
3004 MODULE_DESCRIPTION("Driver for PCnet32 and PCnetPCI based ethercards");
3005 MODULE_LICENSE("GPL");
3006
3007 #define PCNET32_MSG_DEFAULT (NETIF_MSG_DRV | NETIF_MSG_PROBE | NETIF_MSG_LINK)
3008
3009 static int __init pcnet32_init_module(void)
3010 {
3011         printk(KERN_INFO "%s", version);
3012
3013         pcnet32_debug = netif_msg_init(debug, PCNET32_MSG_DEFAULT);
3014
3015         if ((tx_start_pt >= 0) && (tx_start_pt <= 3))
3016                 tx_start = tx_start_pt;
3017
3018         /* find the PCI devices */
3019         if (!pci_register_driver(&pcnet32_driver))
3020                 pcnet32_have_pci = 1;
3021
3022         /* should we find any remaining VLbus devices ? */
3023         if (pcnet32vlb)
3024                 pcnet32_probe_vlbus(pcnet32_portlist);
3025
3026         if (cards_found && (pcnet32_debug & NETIF_MSG_PROBE))
3027                 printk(KERN_INFO PFX "%d cards_found.\n", cards_found);
3028
3029         return (pcnet32_have_pci + cards_found) ? 0 : -ENODEV;
3030 }
3031
3032 static void __exit pcnet32_cleanup_module(void)
3033 {
3034         struct net_device *next_dev;
3035
3036         while (pcnet32_dev) {
3037                 struct pcnet32_private *lp = netdev_priv(pcnet32_dev);
3038                 next_dev = lp->next;
3039                 unregister_netdev(pcnet32_dev);
3040                 pcnet32_free_ring(pcnet32_dev);
3041                 release_region(pcnet32_dev->base_addr, PCNET32_TOTAL_SIZE);
3042                 pci_free_consistent(lp->pci_dev, sizeof(*lp->init_block),
3043                                     lp->init_block, lp->init_dma_addr);
3044                 free_netdev(pcnet32_dev);
3045                 pcnet32_dev = next_dev;
3046         }
3047
3048         if (pcnet32_have_pci)
3049                 pci_unregister_driver(&pcnet32_driver);
3050 }
3051
3052 module_init(pcnet32_init_module);
3053 module_exit(pcnet32_cleanup_module);
3054
3055 /*
3056  * Local variables:
3057  *  c-indent-level: 4
3058  *  tab-width: 8
3059  * End:
3060  */