2dfc60308648153b9cdea7fba12f7742687a0d77
[cascardo/linux.git] / drivers / net / ethernet / ibm / emac / core.c
1 /*
2  * drivers/net/ethernet/ibm/emac/core.c
3  *
4  * Driver for PowerPC 4xx on-chip ethernet controller.
5  *
6  * Copyright 2007 Benjamin Herrenschmidt, IBM Corp.
7  *                <benh@kernel.crashing.org>
8  *
9  * Based on the arch/ppc version of the driver:
10  *
11  * Copyright (c) 2004, 2005 Zultys Technologies.
12  * Eugene Surovegin <eugene.surovegin@zultys.com> or <ebs@ebshome.net>
13  *
14  * Based on original work by
15  *      Matt Porter <mporter@kernel.crashing.org>
16  *      (c) 2003 Benjamin Herrenschmidt <benh@kernel.crashing.org>
17  *      Armin Kuster <akuster@mvista.com>
18  *      Johnnie Peters <jpeters@mvista.com>
19  *
20  * This program is free software; you can redistribute  it and/or modify it
21  * under  the terms of  the GNU General  Public License as published by the
22  * Free Software Foundation;  either version 2 of the  License, or (at your
23  * option) any later version.
24  *
25  */
26
27 #include <linux/module.h>
28 #include <linux/sched.h>
29 #include <linux/string.h>
30 #include <linux/errno.h>
31 #include <linux/delay.h>
32 #include <linux/types.h>
33 #include <linux/pci.h>
34 #include <linux/etherdevice.h>
35 #include <linux/skbuff.h>
36 #include <linux/crc32.h>
37 #include <linux/ethtool.h>
38 #include <linux/mii.h>
39 #include <linux/bitops.h>
40 #include <linux/workqueue.h>
41 #include <linux/of.h>
42 #include <linux/of_address.h>
43 #include <linux/of_irq.h>
44 #include <linux/of_net.h>
45 #include <linux/slab.h>
46
47 #include <asm/processor.h>
48 #include <asm/io.h>
49 #include <asm/dma.h>
50 #include <asm/uaccess.h>
51 #include <asm/dcr.h>
52 #include <asm/dcr-regs.h>
53
54 #include "core.h"
55
56 /*
57  * Lack of dma_unmap_???? calls is intentional.
58  *
59  * API-correct usage requires additional support state information to be
60  * maintained for every RX and TX buffer descriptor (BD). Unfortunately, due to
61  * EMAC design (e.g. TX buffer passed from network stack can be split into
62  * several BDs, dma_map_single/dma_map_page can be used to map particular BD),
63  * maintaining such information will add additional overhead.
64  * Current DMA API implementation for 4xx processors only ensures cache coherency
65  * and dma_unmap_???? routines are empty and are likely to stay this way.
66  * I decided to omit dma_unmap_??? calls because I don't want to add additional
67  * complexity just for the sake of following some abstract API, when it doesn't
68  * add any real benefit to the driver. I understand that this decision maybe
69  * controversial, but I really tried to make code API-correct and efficient
70  * at the same time and didn't come up with code I liked :(.                --ebs
71  */
72
73 #define DRV_NAME        "emac"
74 #define DRV_VERSION     "3.54"
75 #define DRV_DESC        "PPC 4xx OCP EMAC driver"
76
77 MODULE_DESCRIPTION(DRV_DESC);
78 MODULE_AUTHOR
79     ("Eugene Surovegin <eugene.surovegin@zultys.com> or <ebs@ebshome.net>");
80 MODULE_LICENSE("GPL");
81
82 /* minimum number of free TX descriptors required to wake up TX process */
83 #define EMAC_TX_WAKEUP_THRESH           (NUM_TX_BUFF / 4)
84
85 /* If packet size is less than this number, we allocate small skb and copy packet
86  * contents into it instead of just sending original big skb up
87  */
88 #define EMAC_RX_COPY_THRESH             CONFIG_IBM_EMAC_RX_COPY_THRESHOLD
89
90 /* Since multiple EMACs share MDIO lines in various ways, we need
91  * to avoid re-using the same PHY ID in cases where the arch didn't
92  * setup precise phy_map entries
93  *
94  * XXX This is something that needs to be reworked as we can have multiple
95  * EMAC "sets" (multiple ASICs containing several EMACs) though we can
96  * probably require in that case to have explicit PHY IDs in the device-tree
97  */
98 static u32 busy_phy_map;
99 static DEFINE_MUTEX(emac_phy_map_lock);
100
101 /* This is the wait queue used to wait on any event related to probe, that
102  * is discovery of MALs, other EMACs, ZMII/RGMIIs, etc...
103  */
104 static DECLARE_WAIT_QUEUE_HEAD(emac_probe_wait);
105
106 /* Having stable interface names is a doomed idea. However, it would be nice
107  * if we didn't have completely random interface names at boot too :-) It's
108  * just a matter of making everybody's life easier. Since we are doing
109  * threaded probing, it's a bit harder though. The base idea here is that
110  * we make up a list of all emacs in the device-tree before we register the
111  * driver. Every emac will then wait for the previous one in the list to
112  * initialize before itself. We should also keep that list ordered by
113  * cell_index.
114  * That list is only 4 entries long, meaning that additional EMACs don't
115  * get ordering guarantees unless EMAC_BOOT_LIST_SIZE is increased.
116  */
117
118 #define EMAC_BOOT_LIST_SIZE     4
119 static struct device_node *emac_boot_list[EMAC_BOOT_LIST_SIZE];
120
121 /* How long should I wait for dependent devices ? */
122 #define EMAC_PROBE_DEP_TIMEOUT  (HZ * 5)
123
124 /* I don't want to litter system log with timeout errors
125  * when we have brain-damaged PHY.
126  */
127 static inline void emac_report_timeout_error(struct emac_instance *dev,
128                                              const char *error)
129 {
130         if (emac_has_feature(dev, EMAC_FTR_440GX_PHY_CLK_FIX |
131                                   EMAC_FTR_460EX_PHY_CLK_FIX |
132                                   EMAC_FTR_440EP_PHY_CLK_FIX))
133                 DBG(dev, "%s" NL, error);
134         else if (net_ratelimit())
135                 printk(KERN_ERR "%s: %s\n", dev->ofdev->dev.of_node->full_name,
136                         error);
137 }
138
139 /* EMAC PHY clock workaround:
140  * 440EP/440GR has more sane SDR0_MFR register implementation than 440GX,
141  * which allows controlling each EMAC clock
142  */
143 static inline void emac_rx_clk_tx(struct emac_instance *dev)
144 {
145 #ifdef CONFIG_PPC_DCR_NATIVE
146         if (emac_has_feature(dev, EMAC_FTR_440EP_PHY_CLK_FIX))
147                 dcri_clrset(SDR0, SDR0_MFR,
148                             0, SDR0_MFR_ECS >> dev->cell_index);
149 #endif
150 }
151
152 static inline void emac_rx_clk_default(struct emac_instance *dev)
153 {
154 #ifdef CONFIG_PPC_DCR_NATIVE
155         if (emac_has_feature(dev, EMAC_FTR_440EP_PHY_CLK_FIX))
156                 dcri_clrset(SDR0, SDR0_MFR,
157                             SDR0_MFR_ECS >> dev->cell_index, 0);
158 #endif
159 }
160
161 /* PHY polling intervals */
162 #define PHY_POLL_LINK_ON        HZ
163 #define PHY_POLL_LINK_OFF       (HZ / 5)
164
165 /* Graceful stop timeouts in us.
166  * We should allow up to 1 frame time (full-duplex, ignoring collisions)
167  */
168 #define STOP_TIMEOUT_10         1230
169 #define STOP_TIMEOUT_100        124
170 #define STOP_TIMEOUT_1000       13
171 #define STOP_TIMEOUT_1000_JUMBO 73
172
173 static unsigned char default_mcast_addr[] = {
174         0x01, 0x80, 0xC2, 0x00, 0x00, 0x01
175 };
176
177 /* Please, keep in sync with struct ibm_emac_stats/ibm_emac_error_stats */
178 static const char emac_stats_keys[EMAC_ETHTOOL_STATS_COUNT][ETH_GSTRING_LEN] = {
179         "rx_packets", "rx_bytes", "tx_packets", "tx_bytes", "rx_packets_csum",
180         "tx_packets_csum", "tx_undo", "rx_dropped_stack", "rx_dropped_oom",
181         "rx_dropped_error", "rx_dropped_resize", "rx_dropped_mtu",
182         "rx_stopped", "rx_bd_errors", "rx_bd_overrun", "rx_bd_bad_packet",
183         "rx_bd_runt_packet", "rx_bd_short_event", "rx_bd_alignment_error",
184         "rx_bd_bad_fcs", "rx_bd_packet_too_long", "rx_bd_out_of_range",
185         "rx_bd_in_range", "rx_parity", "rx_fifo_overrun", "rx_overrun",
186         "rx_bad_packet", "rx_runt_packet", "rx_short_event",
187         "rx_alignment_error", "rx_bad_fcs", "rx_packet_too_long",
188         "rx_out_of_range", "rx_in_range", "tx_dropped", "tx_bd_errors",
189         "tx_bd_bad_fcs", "tx_bd_carrier_loss", "tx_bd_excessive_deferral",
190         "tx_bd_excessive_collisions", "tx_bd_late_collision",
191         "tx_bd_multple_collisions", "tx_bd_single_collision",
192         "tx_bd_underrun", "tx_bd_sqe", "tx_parity", "tx_underrun", "tx_sqe",
193         "tx_errors"
194 };
195
196 static irqreturn_t emac_irq(int irq, void *dev_instance);
197 static void emac_clean_tx_ring(struct emac_instance *dev);
198 static void __emac_set_multicast_list(struct emac_instance *dev);
199
200 static inline int emac_phy_supports_gige(int phy_mode)
201 {
202         return  phy_mode == PHY_MODE_GMII ||
203                 phy_mode == PHY_MODE_RGMII ||
204                 phy_mode == PHY_MODE_SGMII ||
205                 phy_mode == PHY_MODE_TBI ||
206                 phy_mode == PHY_MODE_RTBI;
207 }
208
209 static inline int emac_phy_gpcs(int phy_mode)
210 {
211         return  phy_mode == PHY_MODE_SGMII ||
212                 phy_mode == PHY_MODE_TBI ||
213                 phy_mode == PHY_MODE_RTBI;
214 }
215
216 static inline void emac_tx_enable(struct emac_instance *dev)
217 {
218         struct emac_regs __iomem *p = dev->emacp;
219         u32 r;
220
221         DBG(dev, "tx_enable" NL);
222
223         r = in_be32(&p->mr0);
224         if (!(r & EMAC_MR0_TXE))
225                 out_be32(&p->mr0, r | EMAC_MR0_TXE);
226 }
227
228 static void emac_tx_disable(struct emac_instance *dev)
229 {
230         struct emac_regs __iomem *p = dev->emacp;
231         u32 r;
232
233         DBG(dev, "tx_disable" NL);
234
235         r = in_be32(&p->mr0);
236         if (r & EMAC_MR0_TXE) {
237                 int n = dev->stop_timeout;
238                 out_be32(&p->mr0, r & ~EMAC_MR0_TXE);
239                 while (!(in_be32(&p->mr0) & EMAC_MR0_TXI) && n) {
240                         udelay(1);
241                         --n;
242                 }
243                 if (unlikely(!n))
244                         emac_report_timeout_error(dev, "TX disable timeout");
245         }
246 }
247
248 static void emac_rx_enable(struct emac_instance *dev)
249 {
250         struct emac_regs __iomem *p = dev->emacp;
251         u32 r;
252
253         if (unlikely(test_bit(MAL_COMMAC_RX_STOPPED, &dev->commac.flags)))
254                 goto out;
255
256         DBG(dev, "rx_enable" NL);
257
258         r = in_be32(&p->mr0);
259         if (!(r & EMAC_MR0_RXE)) {
260                 if (unlikely(!(r & EMAC_MR0_RXI))) {
261                         /* Wait if previous async disable is still in progress */
262                         int n = dev->stop_timeout;
263                         while (!(r = in_be32(&p->mr0) & EMAC_MR0_RXI) && n) {
264                                 udelay(1);
265                                 --n;
266                         }
267                         if (unlikely(!n))
268                                 emac_report_timeout_error(dev,
269                                                           "RX disable timeout");
270                 }
271                 out_be32(&p->mr0, r | EMAC_MR0_RXE);
272         }
273  out:
274         ;
275 }
276
277 static void emac_rx_disable(struct emac_instance *dev)
278 {
279         struct emac_regs __iomem *p = dev->emacp;
280         u32 r;
281
282         DBG(dev, "rx_disable" NL);
283
284         r = in_be32(&p->mr0);
285         if (r & EMAC_MR0_RXE) {
286                 int n = dev->stop_timeout;
287                 out_be32(&p->mr0, r & ~EMAC_MR0_RXE);
288                 while (!(in_be32(&p->mr0) & EMAC_MR0_RXI) && n) {
289                         udelay(1);
290                         --n;
291                 }
292                 if (unlikely(!n))
293                         emac_report_timeout_error(dev, "RX disable timeout");
294         }
295 }
296
297 static inline void emac_netif_stop(struct emac_instance *dev)
298 {
299         netif_tx_lock_bh(dev->ndev);
300         netif_addr_lock(dev->ndev);
301         dev->no_mcast = 1;
302         netif_addr_unlock(dev->ndev);
303         netif_tx_unlock_bh(dev->ndev);
304         netif_trans_update(dev->ndev);  /* prevent tx timeout */
305         mal_poll_disable(dev->mal, &dev->commac);
306         netif_tx_disable(dev->ndev);
307 }
308
309 static inline void emac_netif_start(struct emac_instance *dev)
310 {
311         netif_tx_lock_bh(dev->ndev);
312         netif_addr_lock(dev->ndev);
313         dev->no_mcast = 0;
314         if (dev->mcast_pending && netif_running(dev->ndev))
315                 __emac_set_multicast_list(dev);
316         netif_addr_unlock(dev->ndev);
317         netif_tx_unlock_bh(dev->ndev);
318
319         netif_wake_queue(dev->ndev);
320
321         /* NOTE: unconditional netif_wake_queue is only appropriate
322          * so long as all callers are assured to have free tx slots
323          * (taken from tg3... though the case where that is wrong is
324          *  not terribly harmful)
325          */
326         mal_poll_enable(dev->mal, &dev->commac);
327 }
328
329 static inline void emac_rx_disable_async(struct emac_instance *dev)
330 {
331         struct emac_regs __iomem *p = dev->emacp;
332         u32 r;
333
334         DBG(dev, "rx_disable_async" NL);
335
336         r = in_be32(&p->mr0);
337         if (r & EMAC_MR0_RXE)
338                 out_be32(&p->mr0, r & ~EMAC_MR0_RXE);
339 }
340
341 static int emac_reset(struct emac_instance *dev)
342 {
343         struct emac_regs __iomem *p = dev->emacp;
344         int n = 20;
345
346         DBG(dev, "reset" NL);
347
348         if (!dev->reset_failed) {
349                 /* 40x erratum suggests stopping RX channel before reset,
350                  * we stop TX as well
351                  */
352                 emac_rx_disable(dev);
353                 emac_tx_disable(dev);
354         }
355
356 #ifdef CONFIG_PPC_DCR_NATIVE
357         /*
358          * PPC460EX/GT Embedded Processor Advanced User's Manual
359          * section 28.10.1 Mode Register 0 (EMACx_MR0) states:
360          * Note: The PHY must provide a TX Clk in order to perform a soft reset
361          * of the EMAC. If none is present, select the internal clock
362          * (SDR0_ETH_CFG[EMACx_PHY_CLK] = 1).
363          * After a soft reset, select the external clock.
364          */
365         if (emac_has_feature(dev, EMAC_FTR_460EX_PHY_CLK_FIX)) {
366                 if (dev->phy_address == 0xffffffff &&
367                     dev->phy_map == 0xffffffff) {
368                         /* No PHY: select internal loop clock before reset */
369                         dcri_clrset(SDR0, SDR0_ETH_CFG,
370                                     0, SDR0_ETH_CFG_ECS << dev->cell_index);
371                 } else {
372                         /* PHY present: select external clock before reset */
373                         dcri_clrset(SDR0, SDR0_ETH_CFG,
374                                     SDR0_ETH_CFG_ECS << dev->cell_index, 0);
375                 }
376         }
377 #endif
378
379         out_be32(&p->mr0, EMAC_MR0_SRST);
380         while ((in_be32(&p->mr0) & EMAC_MR0_SRST) && n)
381                 --n;
382
383 #ifdef CONFIG_PPC_DCR_NATIVE
384         if (emac_has_feature(dev, EMAC_FTR_460EX_PHY_CLK_FIX)) {
385                 if (dev->phy_address == 0xffffffff &&
386                     dev->phy_map == 0xffffffff) {
387                         /* No PHY: restore external clock source after reset */
388                         dcri_clrset(SDR0, SDR0_ETH_CFG,
389                                     SDR0_ETH_CFG_ECS << dev->cell_index, 0);
390                 }
391         }
392 #endif
393
394         if (n) {
395                 dev->reset_failed = 0;
396                 return 0;
397         } else {
398                 emac_report_timeout_error(dev, "reset timeout");
399                 dev->reset_failed = 1;
400                 return -ETIMEDOUT;
401         }
402 }
403
404 static void emac_hash_mc(struct emac_instance *dev)
405 {
406         const int regs = EMAC_XAHT_REGS(dev);
407         u32 *gaht_base = emac_gaht_base(dev);
408         u32 gaht_temp[regs];
409         struct netdev_hw_addr *ha;
410         int i;
411
412         DBG(dev, "hash_mc %d" NL, netdev_mc_count(dev->ndev));
413
414         memset(gaht_temp, 0, sizeof (gaht_temp));
415
416         netdev_for_each_mc_addr(ha, dev->ndev) {
417                 int slot, reg, mask;
418                 DBG2(dev, "mc %pM" NL, ha->addr);
419
420                 slot = EMAC_XAHT_CRC_TO_SLOT(dev,
421                                              ether_crc(ETH_ALEN, ha->addr));
422                 reg = EMAC_XAHT_SLOT_TO_REG(dev, slot);
423                 mask = EMAC_XAHT_SLOT_TO_MASK(dev, slot);
424
425                 gaht_temp[reg] |= mask;
426         }
427
428         for (i = 0; i < regs; i++)
429                 out_be32(gaht_base + i, gaht_temp[i]);
430 }
431
432 static inline u32 emac_iff2rmr(struct net_device *ndev)
433 {
434         struct emac_instance *dev = netdev_priv(ndev);
435         u32 r;
436
437         r = EMAC_RMR_SP | EMAC_RMR_SFCS | EMAC_RMR_IAE | EMAC_RMR_BAE;
438
439         if (emac_has_feature(dev, EMAC_FTR_EMAC4))
440             r |= EMAC4_RMR_BASE;
441         else
442             r |= EMAC_RMR_BASE;
443
444         if (ndev->flags & IFF_PROMISC)
445                 r |= EMAC_RMR_PME;
446         else if (ndev->flags & IFF_ALLMULTI ||
447                          (netdev_mc_count(ndev) > EMAC_XAHT_SLOTS(dev)))
448                 r |= EMAC_RMR_PMME;
449         else if (!netdev_mc_empty(ndev))
450                 r |= EMAC_RMR_MAE;
451
452         if (emac_has_feature(dev, EMAC_APM821XX_REQ_JUMBO_FRAME_SIZE)) {
453                 r &= ~EMAC4_RMR_MJS_MASK;
454                 r |= EMAC4_RMR_MJS(ndev->mtu);
455         }
456
457         return r;
458 }
459
460 static u32 __emac_calc_base_mr1(struct emac_instance *dev, int tx_size, int rx_size)
461 {
462         u32 ret = EMAC_MR1_VLE | EMAC_MR1_IST | EMAC_MR1_TR0_MULT;
463
464         DBG2(dev, "__emac_calc_base_mr1" NL);
465
466         switch(tx_size) {
467         case 2048:
468                 ret |= EMAC_MR1_TFS_2K;
469                 break;
470         default:
471                 printk(KERN_WARNING "%s: Unknown Tx FIFO size %d\n",
472                        dev->ndev->name, tx_size);
473         }
474
475         switch(rx_size) {
476         case 16384:
477                 ret |= EMAC_MR1_RFS_16K;
478                 break;
479         case 4096:
480                 ret |= EMAC_MR1_RFS_4K;
481                 break;
482         default:
483                 printk(KERN_WARNING "%s: Unknown Rx FIFO size %d\n",
484                        dev->ndev->name, rx_size);
485         }
486
487         return ret;
488 }
489
490 static u32 __emac4_calc_base_mr1(struct emac_instance *dev, int tx_size, int rx_size)
491 {
492         u32 ret = EMAC_MR1_VLE | EMAC_MR1_IST | EMAC4_MR1_TR |
493                 EMAC4_MR1_OBCI(dev->opb_bus_freq / 1000000);
494
495         DBG2(dev, "__emac4_calc_base_mr1" NL);
496
497         switch(tx_size) {
498         case 16384:
499                 ret |= EMAC4_MR1_TFS_16K;
500                 break;
501         case 4096:
502                 ret |= EMAC4_MR1_TFS_4K;
503                 break;
504         case 2048:
505                 ret |= EMAC4_MR1_TFS_2K;
506                 break;
507         default:
508                 printk(KERN_WARNING "%s: Unknown Tx FIFO size %d\n",
509                        dev->ndev->name, tx_size);
510         }
511
512         switch(rx_size) {
513         case 16384:
514                 ret |= EMAC4_MR1_RFS_16K;
515                 break;
516         case 4096:
517                 ret |= EMAC4_MR1_RFS_4K;
518                 break;
519         case 2048:
520                 ret |= EMAC4_MR1_RFS_2K;
521                 break;
522         default:
523                 printk(KERN_WARNING "%s: Unknown Rx FIFO size %d\n",
524                        dev->ndev->name, rx_size);
525         }
526
527         return ret;
528 }
529
530 static u32 emac_calc_base_mr1(struct emac_instance *dev, int tx_size, int rx_size)
531 {
532         return emac_has_feature(dev, EMAC_FTR_EMAC4) ?
533                 __emac4_calc_base_mr1(dev, tx_size, rx_size) :
534                 __emac_calc_base_mr1(dev, tx_size, rx_size);
535 }
536
537 static inline u32 emac_calc_trtr(struct emac_instance *dev, unsigned int size)
538 {
539         if (emac_has_feature(dev, EMAC_FTR_EMAC4))
540                 return ((size >> 6) - 1) << EMAC_TRTR_SHIFT_EMAC4;
541         else
542                 return ((size >> 6) - 1) << EMAC_TRTR_SHIFT;
543 }
544
545 static inline u32 emac_calc_rwmr(struct emac_instance *dev,
546                                  unsigned int low, unsigned int high)
547 {
548         if (emac_has_feature(dev, EMAC_FTR_EMAC4))
549                 return (low << 22) | ( (high & 0x3ff) << 6);
550         else
551                 return (low << 23) | ( (high & 0x1ff) << 7);
552 }
553
554 static int emac_configure(struct emac_instance *dev)
555 {
556         struct emac_regs __iomem *p = dev->emacp;
557         struct net_device *ndev = dev->ndev;
558         int tx_size, rx_size, link = netif_carrier_ok(dev->ndev);
559         u32 r, mr1 = 0;
560
561         DBG(dev, "configure" NL);
562
563         if (!link) {
564                 out_be32(&p->mr1, in_be32(&p->mr1)
565                          | EMAC_MR1_FDE | EMAC_MR1_ILE);
566                 udelay(100);
567         } else if (emac_reset(dev) < 0)
568                 return -ETIMEDOUT;
569
570         if (emac_has_feature(dev, EMAC_FTR_HAS_TAH))
571                 tah_reset(dev->tah_dev);
572
573         DBG(dev, " link = %d duplex = %d, pause = %d, asym_pause = %d\n",
574             link, dev->phy.duplex, dev->phy.pause, dev->phy.asym_pause);
575
576         /* Default fifo sizes */
577         tx_size = dev->tx_fifo_size;
578         rx_size = dev->rx_fifo_size;
579
580         /* No link, force loopback */
581         if (!link)
582                 mr1 = EMAC_MR1_FDE | EMAC_MR1_ILE;
583
584         /* Check for full duplex */
585         else if (dev->phy.duplex == DUPLEX_FULL)
586                 mr1 |= EMAC_MR1_FDE | EMAC_MR1_MWSW_001;
587
588         /* Adjust fifo sizes, mr1 and timeouts based on link speed */
589         dev->stop_timeout = STOP_TIMEOUT_10;
590         switch (dev->phy.speed) {
591         case SPEED_1000:
592                 if (emac_phy_gpcs(dev->phy.mode)) {
593                         mr1 |= EMAC_MR1_MF_1000GPCS | EMAC_MR1_MF_IPPA(
594                                 (dev->phy.gpcs_address != 0xffffffff) ?
595                                  dev->phy.gpcs_address : dev->phy.address);
596
597                         /* Put some arbitrary OUI, Manuf & Rev IDs so we can
598                          * identify this GPCS PHY later.
599                          */
600                         out_be32(&p->u1.emac4.ipcr, 0xdeadbeef);
601                 } else
602                         mr1 |= EMAC_MR1_MF_1000;
603
604                 /* Extended fifo sizes */
605                 tx_size = dev->tx_fifo_size_gige;
606                 rx_size = dev->rx_fifo_size_gige;
607
608                 if (dev->ndev->mtu > ETH_DATA_LEN) {
609                         if (emac_has_feature(dev, EMAC_FTR_EMAC4))
610                                 mr1 |= EMAC4_MR1_JPSM;
611                         else
612                                 mr1 |= EMAC_MR1_JPSM;
613                         dev->stop_timeout = STOP_TIMEOUT_1000_JUMBO;
614                 } else
615                         dev->stop_timeout = STOP_TIMEOUT_1000;
616                 break;
617         case SPEED_100:
618                 mr1 |= EMAC_MR1_MF_100;
619                 dev->stop_timeout = STOP_TIMEOUT_100;
620                 break;
621         default: /* make gcc happy */
622                 break;
623         }
624
625         if (emac_has_feature(dev, EMAC_FTR_HAS_RGMII))
626                 rgmii_set_speed(dev->rgmii_dev, dev->rgmii_port,
627                                 dev->phy.speed);
628         if (emac_has_feature(dev, EMAC_FTR_HAS_ZMII))
629                 zmii_set_speed(dev->zmii_dev, dev->zmii_port, dev->phy.speed);
630
631         /* on 40x erratum forces us to NOT use integrated flow control,
632          * let's hope it works on 44x ;)
633          */
634         if (!emac_has_feature(dev, EMAC_FTR_NO_FLOW_CONTROL_40x) &&
635             dev->phy.duplex == DUPLEX_FULL) {
636                 if (dev->phy.pause)
637                         mr1 |= EMAC_MR1_EIFC | EMAC_MR1_APP;
638                 else if (dev->phy.asym_pause)
639                         mr1 |= EMAC_MR1_APP;
640         }
641
642         /* Add base settings & fifo sizes & program MR1 */
643         mr1 |= emac_calc_base_mr1(dev, tx_size, rx_size);
644         out_be32(&p->mr1, mr1);
645
646         /* Set individual MAC address */
647         out_be32(&p->iahr, (ndev->dev_addr[0] << 8) | ndev->dev_addr[1]);
648         out_be32(&p->ialr, (ndev->dev_addr[2] << 24) |
649                  (ndev->dev_addr[3] << 16) | (ndev->dev_addr[4] << 8) |
650                  ndev->dev_addr[5]);
651
652         /* VLAN Tag Protocol ID */
653         out_be32(&p->vtpid, 0x8100);
654
655         /* Receive mode register */
656         r = emac_iff2rmr(ndev);
657         if (r & EMAC_RMR_MAE)
658                 emac_hash_mc(dev);
659         out_be32(&p->rmr, r);
660
661         /* FIFOs thresholds */
662         if (emac_has_feature(dev, EMAC_FTR_EMAC4))
663                 r = EMAC4_TMR1((dev->mal_burst_size / dev->fifo_entry_size) + 1,
664                                tx_size / 2 / dev->fifo_entry_size);
665         else
666                 r = EMAC_TMR1((dev->mal_burst_size / dev->fifo_entry_size) + 1,
667                               tx_size / 2 / dev->fifo_entry_size);
668         out_be32(&p->tmr1, r);
669         out_be32(&p->trtr, emac_calc_trtr(dev, tx_size / 2));
670
671         /* PAUSE frame is sent when RX FIFO reaches its high-water mark,
672            there should be still enough space in FIFO to allow the our link
673            partner time to process this frame and also time to send PAUSE
674            frame itself.
675
676            Here is the worst case scenario for the RX FIFO "headroom"
677            (from "The Switch Book") (100Mbps, without preamble, inter-frame gap):
678
679            1) One maximum-length frame on TX                    1522 bytes
680            2) One PAUSE frame time                                64 bytes
681            3) PAUSE frame decode time allowance                   64 bytes
682            4) One maximum-length frame on RX                    1522 bytes
683            5) Round-trip propagation delay of the link (100Mb)    15 bytes
684            ----------
685            3187 bytes
686
687            I chose to set high-water mark to RX_FIFO_SIZE / 4 (1024 bytes)
688            low-water mark  to RX_FIFO_SIZE / 8 (512 bytes)
689          */
690         r = emac_calc_rwmr(dev, rx_size / 8 / dev->fifo_entry_size,
691                            rx_size / 4 / dev->fifo_entry_size);
692         out_be32(&p->rwmr, r);
693
694         /* Set PAUSE timer to the maximum */
695         out_be32(&p->ptr, 0xffff);
696
697         /* IRQ sources */
698         r = EMAC_ISR_OVR | EMAC_ISR_BP | EMAC_ISR_SE |
699                 EMAC_ISR_ALE | EMAC_ISR_BFCS | EMAC_ISR_PTLE | EMAC_ISR_ORE |
700                 EMAC_ISR_IRE | EMAC_ISR_TE;
701         if (emac_has_feature(dev, EMAC_FTR_EMAC4))
702             r |= EMAC4_ISR_TXPE | EMAC4_ISR_RXPE /* | EMAC4_ISR_TXUE |
703                                                   EMAC4_ISR_RXOE | */;
704         out_be32(&p->iser,  r);
705
706         /* We need to take GPCS PHY out of isolate mode after EMAC reset */
707         if (emac_phy_gpcs(dev->phy.mode)) {
708                 if (dev->phy.gpcs_address != 0xffffffff)
709                         emac_mii_reset_gpcs(&dev->phy);
710                 else
711                         emac_mii_reset_phy(&dev->phy);
712         }
713
714         return 0;
715 }
716
717 static void emac_reinitialize(struct emac_instance *dev)
718 {
719         DBG(dev, "reinitialize" NL);
720
721         emac_netif_stop(dev);
722         if (!emac_configure(dev)) {
723                 emac_tx_enable(dev);
724                 emac_rx_enable(dev);
725         }
726         emac_netif_start(dev);
727 }
728
729 static void emac_full_tx_reset(struct emac_instance *dev)
730 {
731         DBG(dev, "full_tx_reset" NL);
732
733         emac_tx_disable(dev);
734         mal_disable_tx_channel(dev->mal, dev->mal_tx_chan);
735         emac_clean_tx_ring(dev);
736         dev->tx_cnt = dev->tx_slot = dev->ack_slot = 0;
737
738         emac_configure(dev);
739
740         mal_enable_tx_channel(dev->mal, dev->mal_tx_chan);
741         emac_tx_enable(dev);
742         emac_rx_enable(dev);
743 }
744
745 static void emac_reset_work(struct work_struct *work)
746 {
747         struct emac_instance *dev = container_of(work, struct emac_instance, reset_work);
748
749         DBG(dev, "reset_work" NL);
750
751         mutex_lock(&dev->link_lock);
752         if (dev->opened) {
753                 emac_netif_stop(dev);
754                 emac_full_tx_reset(dev);
755                 emac_netif_start(dev);
756         }
757         mutex_unlock(&dev->link_lock);
758 }
759
760 static void emac_tx_timeout(struct net_device *ndev)
761 {
762         struct emac_instance *dev = netdev_priv(ndev);
763
764         DBG(dev, "tx_timeout" NL);
765
766         schedule_work(&dev->reset_work);
767 }
768
769
770 static inline int emac_phy_done(struct emac_instance *dev, u32 stacr)
771 {
772         int done = !!(stacr & EMAC_STACR_OC);
773
774         if (emac_has_feature(dev, EMAC_FTR_STACR_OC_INVERT))
775                 done = !done;
776
777         return done;
778 };
779
780 static int __emac_mdio_read(struct emac_instance *dev, u8 id, u8 reg)
781 {
782         struct emac_regs __iomem *p = dev->emacp;
783         u32 r = 0;
784         int n, err = -ETIMEDOUT;
785
786         mutex_lock(&dev->mdio_lock);
787
788         DBG2(dev, "mdio_read(%02x,%02x)" NL, id, reg);
789
790         /* Enable proper MDIO port */
791         if (emac_has_feature(dev, EMAC_FTR_HAS_ZMII))
792                 zmii_get_mdio(dev->zmii_dev, dev->zmii_port);
793         if (emac_has_feature(dev, EMAC_FTR_HAS_RGMII))
794                 rgmii_get_mdio(dev->rgmii_dev, dev->rgmii_port);
795
796         /* Wait for management interface to become idle */
797         n = 20;
798         while (!emac_phy_done(dev, in_be32(&p->stacr))) {
799                 udelay(1);
800                 if (!--n) {
801                         DBG2(dev, " -> timeout wait idle\n");
802                         goto bail;
803                 }
804         }
805
806         /* Issue read command */
807         if (emac_has_feature(dev, EMAC_FTR_EMAC4))
808                 r = EMAC4_STACR_BASE(dev->opb_bus_freq);
809         else
810                 r = EMAC_STACR_BASE(dev->opb_bus_freq);
811         if (emac_has_feature(dev, EMAC_FTR_STACR_OC_INVERT))
812                 r |= EMAC_STACR_OC;
813         if (emac_has_feature(dev, EMAC_FTR_HAS_NEW_STACR))
814                 r |= EMACX_STACR_STAC_READ;
815         else
816                 r |= EMAC_STACR_STAC_READ;
817         r |= (reg & EMAC_STACR_PRA_MASK)
818                 | ((id & EMAC_STACR_PCDA_MASK) << EMAC_STACR_PCDA_SHIFT);
819         out_be32(&p->stacr, r);
820
821         /* Wait for read to complete */
822         n = 200;
823         while (!emac_phy_done(dev, (r = in_be32(&p->stacr)))) {
824                 udelay(1);
825                 if (!--n) {
826                         DBG2(dev, " -> timeout wait complete\n");
827                         goto bail;
828                 }
829         }
830
831         if (unlikely(r & EMAC_STACR_PHYE)) {
832                 DBG(dev, "mdio_read(%02x, %02x) failed" NL, id, reg);
833                 err = -EREMOTEIO;
834                 goto bail;
835         }
836
837         r = ((r >> EMAC_STACR_PHYD_SHIFT) & EMAC_STACR_PHYD_MASK);
838
839         DBG2(dev, "mdio_read -> %04x" NL, r);
840         err = 0;
841  bail:
842         if (emac_has_feature(dev, EMAC_FTR_HAS_RGMII))
843                 rgmii_put_mdio(dev->rgmii_dev, dev->rgmii_port);
844         if (emac_has_feature(dev, EMAC_FTR_HAS_ZMII))
845                 zmii_put_mdio(dev->zmii_dev, dev->zmii_port);
846         mutex_unlock(&dev->mdio_lock);
847
848         return err == 0 ? r : err;
849 }
850
851 static void __emac_mdio_write(struct emac_instance *dev, u8 id, u8 reg,
852                               u16 val)
853 {
854         struct emac_regs __iomem *p = dev->emacp;
855         u32 r = 0;
856         int n, err = -ETIMEDOUT;
857
858         mutex_lock(&dev->mdio_lock);
859
860         DBG2(dev, "mdio_write(%02x,%02x,%04x)" NL, id, reg, val);
861
862         /* Enable proper MDIO port */
863         if (emac_has_feature(dev, EMAC_FTR_HAS_ZMII))
864                 zmii_get_mdio(dev->zmii_dev, dev->zmii_port);
865         if (emac_has_feature(dev, EMAC_FTR_HAS_RGMII))
866                 rgmii_get_mdio(dev->rgmii_dev, dev->rgmii_port);
867
868         /* Wait for management interface to be idle */
869         n = 20;
870         while (!emac_phy_done(dev, in_be32(&p->stacr))) {
871                 udelay(1);
872                 if (!--n) {
873                         DBG2(dev, " -> timeout wait idle\n");
874                         goto bail;
875                 }
876         }
877
878         /* Issue write command */
879         if (emac_has_feature(dev, EMAC_FTR_EMAC4))
880                 r = EMAC4_STACR_BASE(dev->opb_bus_freq);
881         else
882                 r = EMAC_STACR_BASE(dev->opb_bus_freq);
883         if (emac_has_feature(dev, EMAC_FTR_STACR_OC_INVERT))
884                 r |= EMAC_STACR_OC;
885         if (emac_has_feature(dev, EMAC_FTR_HAS_NEW_STACR))
886                 r |= EMACX_STACR_STAC_WRITE;
887         else
888                 r |= EMAC_STACR_STAC_WRITE;
889         r |= (reg & EMAC_STACR_PRA_MASK) |
890                 ((id & EMAC_STACR_PCDA_MASK) << EMAC_STACR_PCDA_SHIFT) |
891                 (val << EMAC_STACR_PHYD_SHIFT);
892         out_be32(&p->stacr, r);
893
894         /* Wait for write to complete */
895         n = 200;
896         while (!emac_phy_done(dev, in_be32(&p->stacr))) {
897                 udelay(1);
898                 if (!--n) {
899                         DBG2(dev, " -> timeout wait complete\n");
900                         goto bail;
901                 }
902         }
903         err = 0;
904  bail:
905         if (emac_has_feature(dev, EMAC_FTR_HAS_RGMII))
906                 rgmii_put_mdio(dev->rgmii_dev, dev->rgmii_port);
907         if (emac_has_feature(dev, EMAC_FTR_HAS_ZMII))
908                 zmii_put_mdio(dev->zmii_dev, dev->zmii_port);
909         mutex_unlock(&dev->mdio_lock);
910 }
911
912 static int emac_mdio_read(struct net_device *ndev, int id, int reg)
913 {
914         struct emac_instance *dev = netdev_priv(ndev);
915         int res;
916
917         res = __emac_mdio_read((dev->mdio_instance &&
918                                 dev->phy.gpcs_address != id) ?
919                                 dev->mdio_instance : dev,
920                                (u8) id, (u8) reg);
921         return res;
922 }
923
924 static void emac_mdio_write(struct net_device *ndev, int id, int reg, int val)
925 {
926         struct emac_instance *dev = netdev_priv(ndev);
927
928         __emac_mdio_write((dev->mdio_instance &&
929                            dev->phy.gpcs_address != id) ?
930                            dev->mdio_instance : dev,
931                           (u8) id, (u8) reg, (u16) val);
932 }
933
934 /* Tx lock BH */
935 static void __emac_set_multicast_list(struct emac_instance *dev)
936 {
937         struct emac_regs __iomem *p = dev->emacp;
938         u32 rmr = emac_iff2rmr(dev->ndev);
939
940         DBG(dev, "__multicast %08x" NL, rmr);
941
942         /* I decided to relax register access rules here to avoid
943          * full EMAC reset.
944          *
945          * There is a real problem with EMAC4 core if we use MWSW_001 bit
946          * in MR1 register and do a full EMAC reset.
947          * One TX BD status update is delayed and, after EMAC reset, it
948          * never happens, resulting in TX hung (it'll be recovered by TX
949          * timeout handler eventually, but this is just gross).
950          * So we either have to do full TX reset or try to cheat here :)
951          *
952          * The only required change is to RX mode register, so I *think* all
953          * we need is just to stop RX channel. This seems to work on all
954          * tested SoCs.                                                --ebs
955          *
956          * If we need the full reset, we might just trigger the workqueue
957          * and do it async... a bit nasty but should work --BenH
958          */
959         dev->mcast_pending = 0;
960         emac_rx_disable(dev);
961         if (rmr & EMAC_RMR_MAE)
962                 emac_hash_mc(dev);
963         out_be32(&p->rmr, rmr);
964         emac_rx_enable(dev);
965 }
966
967 /* Tx lock BH */
968 static void emac_set_multicast_list(struct net_device *ndev)
969 {
970         struct emac_instance *dev = netdev_priv(ndev);
971
972         DBG(dev, "multicast" NL);
973
974         BUG_ON(!netif_running(dev->ndev));
975
976         if (dev->no_mcast) {
977                 dev->mcast_pending = 1;
978                 return;
979         }
980         __emac_set_multicast_list(dev);
981 }
982
983 static int emac_set_mac_address(struct net_device *ndev, void *sa)
984 {
985         struct emac_instance *dev = netdev_priv(ndev);
986         struct sockaddr *addr = sa;
987         struct emac_regs __iomem *p = dev->emacp;
988
989         if (!is_valid_ether_addr(addr->sa_data))
990                return -EADDRNOTAVAIL;
991
992         mutex_lock(&dev->link_lock);
993
994         memcpy(ndev->dev_addr, addr->sa_data, ndev->addr_len);
995
996         emac_rx_disable(dev);
997         emac_tx_disable(dev);
998         out_be32(&p->iahr, (ndev->dev_addr[0] << 8) | ndev->dev_addr[1]);
999         out_be32(&p->ialr, (ndev->dev_addr[2] << 24) |
1000                 (ndev->dev_addr[3] << 16) | (ndev->dev_addr[4] << 8) |
1001                 ndev->dev_addr[5]);
1002         emac_tx_enable(dev);
1003         emac_rx_enable(dev);
1004
1005         mutex_unlock(&dev->link_lock);
1006
1007         return 0;
1008 }
1009
1010 static int emac_resize_rx_ring(struct emac_instance *dev, int new_mtu)
1011 {
1012         int rx_sync_size = emac_rx_sync_size(new_mtu);
1013         int rx_skb_size = emac_rx_skb_size(new_mtu);
1014         int i, ret = 0;
1015         int mr1_jumbo_bit_change = 0;
1016
1017         mutex_lock(&dev->link_lock);
1018         emac_netif_stop(dev);
1019         emac_rx_disable(dev);
1020         mal_disable_rx_channel(dev->mal, dev->mal_rx_chan);
1021
1022         if (dev->rx_sg_skb) {
1023                 ++dev->estats.rx_dropped_resize;
1024                 dev_kfree_skb(dev->rx_sg_skb);
1025                 dev->rx_sg_skb = NULL;
1026         }
1027
1028         /* Make a first pass over RX ring and mark BDs ready, dropping
1029          * non-processed packets on the way. We need this as a separate pass
1030          * to simplify error recovery in the case of allocation failure later.
1031          */
1032         for (i = 0; i < NUM_RX_BUFF; ++i) {
1033                 if (dev->rx_desc[i].ctrl & MAL_RX_CTRL_FIRST)
1034                         ++dev->estats.rx_dropped_resize;
1035
1036                 dev->rx_desc[i].data_len = 0;
1037                 dev->rx_desc[i].ctrl = MAL_RX_CTRL_EMPTY |
1038                     (i == (NUM_RX_BUFF - 1) ? MAL_RX_CTRL_WRAP : 0);
1039         }
1040
1041         /* Reallocate RX ring only if bigger skb buffers are required */
1042         if (rx_skb_size <= dev->rx_skb_size)
1043                 goto skip;
1044
1045         /* Second pass, allocate new skbs */
1046         for (i = 0; i < NUM_RX_BUFF; ++i) {
1047                 struct sk_buff *skb = alloc_skb(rx_skb_size, GFP_ATOMIC);
1048                 if (!skb) {
1049                         ret = -ENOMEM;
1050                         goto oom;
1051                 }
1052
1053                 BUG_ON(!dev->rx_skb[i]);
1054                 dev_kfree_skb(dev->rx_skb[i]);
1055
1056                 skb_reserve(skb, EMAC_RX_SKB_HEADROOM + 2);
1057                 dev->rx_desc[i].data_ptr =
1058                     dma_map_single(&dev->ofdev->dev, skb->data - 2, rx_sync_size,
1059                                    DMA_FROM_DEVICE) + 2;
1060                 dev->rx_skb[i] = skb;
1061         }
1062  skip:
1063         /* Check if we need to change "Jumbo" bit in MR1 */
1064         if (emac_has_feature(dev, EMAC_APM821XX_REQ_JUMBO_FRAME_SIZE)) {
1065                 mr1_jumbo_bit_change = (new_mtu > ETH_DATA_LEN) ||
1066                                 (dev->ndev->mtu > ETH_DATA_LEN);
1067         } else {
1068                 mr1_jumbo_bit_change = (new_mtu > ETH_DATA_LEN) ^
1069                                 (dev->ndev->mtu > ETH_DATA_LEN);
1070         }
1071
1072         if (mr1_jumbo_bit_change) {
1073                 /* This is to prevent starting RX channel in emac_rx_enable() */
1074                 set_bit(MAL_COMMAC_RX_STOPPED, &dev->commac.flags);
1075
1076                 dev->ndev->mtu = new_mtu;
1077                 emac_full_tx_reset(dev);
1078         }
1079
1080         mal_set_rcbs(dev->mal, dev->mal_rx_chan, emac_rx_size(new_mtu));
1081  oom:
1082         /* Restart RX */
1083         clear_bit(MAL_COMMAC_RX_STOPPED, &dev->commac.flags);
1084         dev->rx_slot = 0;
1085         mal_enable_rx_channel(dev->mal, dev->mal_rx_chan);
1086         emac_rx_enable(dev);
1087         emac_netif_start(dev);
1088         mutex_unlock(&dev->link_lock);
1089
1090         return ret;
1091 }
1092
1093 /* Process ctx, rtnl_lock semaphore */
1094 static int emac_change_mtu(struct net_device *ndev, int new_mtu)
1095 {
1096         struct emac_instance *dev = netdev_priv(ndev);
1097         int ret = 0;
1098
1099         if (new_mtu < EMAC_MIN_MTU || new_mtu > dev->max_mtu)
1100                 return -EINVAL;
1101
1102         DBG(dev, "change_mtu(%d)" NL, new_mtu);
1103
1104         if (netif_running(ndev)) {
1105                 /* Check if we really need to reinitialize RX ring */
1106                 if (emac_rx_skb_size(ndev->mtu) != emac_rx_skb_size(new_mtu))
1107                         ret = emac_resize_rx_ring(dev, new_mtu);
1108         }
1109
1110         if (!ret) {
1111                 ndev->mtu = new_mtu;
1112                 dev->rx_skb_size = emac_rx_skb_size(new_mtu);
1113                 dev->rx_sync_size = emac_rx_sync_size(new_mtu);
1114         }
1115
1116         return ret;
1117 }
1118
1119 static void emac_clean_tx_ring(struct emac_instance *dev)
1120 {
1121         int i;
1122
1123         for (i = 0; i < NUM_TX_BUFF; ++i) {
1124                 if (dev->tx_skb[i]) {
1125                         dev_kfree_skb(dev->tx_skb[i]);
1126                         dev->tx_skb[i] = NULL;
1127                         if (dev->tx_desc[i].ctrl & MAL_TX_CTRL_READY)
1128                                 ++dev->estats.tx_dropped;
1129                 }
1130                 dev->tx_desc[i].ctrl = 0;
1131                 dev->tx_desc[i].data_ptr = 0;
1132         }
1133 }
1134
1135 static void emac_clean_rx_ring(struct emac_instance *dev)
1136 {
1137         int i;
1138
1139         for (i = 0; i < NUM_RX_BUFF; ++i)
1140                 if (dev->rx_skb[i]) {
1141                         dev->rx_desc[i].ctrl = 0;
1142                         dev_kfree_skb(dev->rx_skb[i]);
1143                         dev->rx_skb[i] = NULL;
1144                         dev->rx_desc[i].data_ptr = 0;
1145                 }
1146
1147         if (dev->rx_sg_skb) {
1148                 dev_kfree_skb(dev->rx_sg_skb);
1149                 dev->rx_sg_skb = NULL;
1150         }
1151 }
1152
1153 static inline int emac_alloc_rx_skb(struct emac_instance *dev, int slot,
1154                                     gfp_t flags)
1155 {
1156         struct sk_buff *skb = alloc_skb(dev->rx_skb_size, flags);
1157         if (unlikely(!skb))
1158                 return -ENOMEM;
1159
1160         dev->rx_skb[slot] = skb;
1161         dev->rx_desc[slot].data_len = 0;
1162
1163         skb_reserve(skb, EMAC_RX_SKB_HEADROOM + 2);
1164         dev->rx_desc[slot].data_ptr =
1165             dma_map_single(&dev->ofdev->dev, skb->data - 2, dev->rx_sync_size,
1166                            DMA_FROM_DEVICE) + 2;
1167         wmb();
1168         dev->rx_desc[slot].ctrl = MAL_RX_CTRL_EMPTY |
1169             (slot == (NUM_RX_BUFF - 1) ? MAL_RX_CTRL_WRAP : 0);
1170
1171         return 0;
1172 }
1173
1174 static void emac_print_link_status(struct emac_instance *dev)
1175 {
1176         if (netif_carrier_ok(dev->ndev))
1177                 printk(KERN_INFO "%s: link is up, %d %s%s\n",
1178                        dev->ndev->name, dev->phy.speed,
1179                        dev->phy.duplex == DUPLEX_FULL ? "FDX" : "HDX",
1180                        dev->phy.pause ? ", pause enabled" :
1181                        dev->phy.asym_pause ? ", asymmetric pause enabled" : "");
1182         else
1183                 printk(KERN_INFO "%s: link is down\n", dev->ndev->name);
1184 }
1185
1186 /* Process ctx, rtnl_lock semaphore */
1187 static int emac_open(struct net_device *ndev)
1188 {
1189         struct emac_instance *dev = netdev_priv(ndev);
1190         int err, i;
1191
1192         DBG(dev, "open" NL);
1193
1194         /* Setup error IRQ handler */
1195         err = request_irq(dev->emac_irq, emac_irq, 0, "EMAC", dev);
1196         if (err) {
1197                 printk(KERN_ERR "%s: failed to request IRQ %d\n",
1198                        ndev->name, dev->emac_irq);
1199                 return err;
1200         }
1201
1202         /* Allocate RX ring */
1203         for (i = 0; i < NUM_RX_BUFF; ++i)
1204                 if (emac_alloc_rx_skb(dev, i, GFP_KERNEL)) {
1205                         printk(KERN_ERR "%s: failed to allocate RX ring\n",
1206                                ndev->name);
1207                         goto oom;
1208                 }
1209
1210         dev->tx_cnt = dev->tx_slot = dev->ack_slot = dev->rx_slot = 0;
1211         clear_bit(MAL_COMMAC_RX_STOPPED, &dev->commac.flags);
1212         dev->rx_sg_skb = NULL;
1213
1214         mutex_lock(&dev->link_lock);
1215         dev->opened = 1;
1216
1217         /* Start PHY polling now.
1218          */
1219         if (dev->phy.address >= 0) {
1220                 int link_poll_interval;
1221                 if (dev->phy.def->ops->poll_link(&dev->phy)) {
1222                         dev->phy.def->ops->read_link(&dev->phy);
1223                         emac_rx_clk_default(dev);
1224                         netif_carrier_on(dev->ndev);
1225                         link_poll_interval = PHY_POLL_LINK_ON;
1226                 } else {
1227                         emac_rx_clk_tx(dev);
1228                         netif_carrier_off(dev->ndev);
1229                         link_poll_interval = PHY_POLL_LINK_OFF;
1230                 }
1231                 dev->link_polling = 1;
1232                 wmb();
1233                 schedule_delayed_work(&dev->link_work, link_poll_interval);
1234                 emac_print_link_status(dev);
1235         } else
1236                 netif_carrier_on(dev->ndev);
1237
1238         /* Required for Pause packet support in EMAC */
1239         dev_mc_add_global(ndev, default_mcast_addr);
1240
1241         emac_configure(dev);
1242         mal_poll_add(dev->mal, &dev->commac);
1243         mal_enable_tx_channel(dev->mal, dev->mal_tx_chan);
1244         mal_set_rcbs(dev->mal, dev->mal_rx_chan, emac_rx_size(ndev->mtu));
1245         mal_enable_rx_channel(dev->mal, dev->mal_rx_chan);
1246         emac_tx_enable(dev);
1247         emac_rx_enable(dev);
1248         emac_netif_start(dev);
1249
1250         mutex_unlock(&dev->link_lock);
1251
1252         return 0;
1253  oom:
1254         emac_clean_rx_ring(dev);
1255         free_irq(dev->emac_irq, dev);
1256
1257         return -ENOMEM;
1258 }
1259
1260 /* BHs disabled */
1261 #if 0
1262 static int emac_link_differs(struct emac_instance *dev)
1263 {
1264         u32 r = in_be32(&dev->emacp->mr1);
1265
1266         int duplex = r & EMAC_MR1_FDE ? DUPLEX_FULL : DUPLEX_HALF;
1267         int speed, pause, asym_pause;
1268
1269         if (r & EMAC_MR1_MF_1000)
1270                 speed = SPEED_1000;
1271         else if (r & EMAC_MR1_MF_100)
1272                 speed = SPEED_100;
1273         else
1274                 speed = SPEED_10;
1275
1276         switch (r & (EMAC_MR1_EIFC | EMAC_MR1_APP)) {
1277         case (EMAC_MR1_EIFC | EMAC_MR1_APP):
1278                 pause = 1;
1279                 asym_pause = 0;
1280                 break;
1281         case EMAC_MR1_APP:
1282                 pause = 0;
1283                 asym_pause = 1;
1284                 break;
1285         default:
1286                 pause = asym_pause = 0;
1287         }
1288         return speed != dev->phy.speed || duplex != dev->phy.duplex ||
1289             pause != dev->phy.pause || asym_pause != dev->phy.asym_pause;
1290 }
1291 #endif
1292
1293 static void emac_link_timer(struct work_struct *work)
1294 {
1295         struct emac_instance *dev =
1296                 container_of(to_delayed_work(work),
1297                              struct emac_instance, link_work);
1298         int link_poll_interval;
1299
1300         mutex_lock(&dev->link_lock);
1301         DBG2(dev, "link timer" NL);
1302
1303         if (!dev->opened)
1304                 goto bail;
1305
1306         if (dev->phy.def->ops->poll_link(&dev->phy)) {
1307                 if (!netif_carrier_ok(dev->ndev)) {
1308                         emac_rx_clk_default(dev);
1309                         /* Get new link parameters */
1310                         dev->phy.def->ops->read_link(&dev->phy);
1311
1312                         netif_carrier_on(dev->ndev);
1313                         emac_netif_stop(dev);
1314                         emac_full_tx_reset(dev);
1315                         emac_netif_start(dev);
1316                         emac_print_link_status(dev);
1317                 }
1318                 link_poll_interval = PHY_POLL_LINK_ON;
1319         } else {
1320                 if (netif_carrier_ok(dev->ndev)) {
1321                         emac_rx_clk_tx(dev);
1322                         netif_carrier_off(dev->ndev);
1323                         netif_tx_disable(dev->ndev);
1324                         emac_reinitialize(dev);
1325                         emac_print_link_status(dev);
1326                 }
1327                 link_poll_interval = PHY_POLL_LINK_OFF;
1328         }
1329         schedule_delayed_work(&dev->link_work, link_poll_interval);
1330  bail:
1331         mutex_unlock(&dev->link_lock);
1332 }
1333
1334 static void emac_force_link_update(struct emac_instance *dev)
1335 {
1336         netif_carrier_off(dev->ndev);
1337         smp_rmb();
1338         if (dev->link_polling) {
1339                 cancel_delayed_work_sync(&dev->link_work);
1340                 if (dev->link_polling)
1341                         schedule_delayed_work(&dev->link_work,  PHY_POLL_LINK_OFF);
1342         }
1343 }
1344
1345 /* Process ctx, rtnl_lock semaphore */
1346 static int emac_close(struct net_device *ndev)
1347 {
1348         struct emac_instance *dev = netdev_priv(ndev);
1349
1350         DBG(dev, "close" NL);
1351
1352         if (dev->phy.address >= 0) {
1353                 dev->link_polling = 0;
1354                 cancel_delayed_work_sync(&dev->link_work);
1355         }
1356         mutex_lock(&dev->link_lock);
1357         emac_netif_stop(dev);
1358         dev->opened = 0;
1359         mutex_unlock(&dev->link_lock);
1360
1361         emac_rx_disable(dev);
1362         emac_tx_disable(dev);
1363         mal_disable_rx_channel(dev->mal, dev->mal_rx_chan);
1364         mal_disable_tx_channel(dev->mal, dev->mal_tx_chan);
1365         mal_poll_del(dev->mal, &dev->commac);
1366
1367         emac_clean_tx_ring(dev);
1368         emac_clean_rx_ring(dev);
1369
1370         free_irq(dev->emac_irq, dev);
1371
1372         netif_carrier_off(ndev);
1373
1374         return 0;
1375 }
1376
1377 static inline u16 emac_tx_csum(struct emac_instance *dev,
1378                                struct sk_buff *skb)
1379 {
1380         if (emac_has_feature(dev, EMAC_FTR_HAS_TAH) &&
1381                 (skb->ip_summed == CHECKSUM_PARTIAL)) {
1382                 ++dev->stats.tx_packets_csum;
1383                 return EMAC_TX_CTRL_TAH_CSUM;
1384         }
1385         return 0;
1386 }
1387
1388 static inline int emac_xmit_finish(struct emac_instance *dev, int len)
1389 {
1390         struct emac_regs __iomem *p = dev->emacp;
1391         struct net_device *ndev = dev->ndev;
1392
1393         /* Send the packet out. If the if makes a significant perf
1394          * difference, then we can store the TMR0 value in "dev"
1395          * instead
1396          */
1397         if (emac_has_feature(dev, EMAC_FTR_EMAC4))
1398                 out_be32(&p->tmr0, EMAC4_TMR0_XMIT);
1399         else
1400                 out_be32(&p->tmr0, EMAC_TMR0_XMIT);
1401
1402         if (unlikely(++dev->tx_cnt == NUM_TX_BUFF)) {
1403                 netif_stop_queue(ndev);
1404                 DBG2(dev, "stopped TX queue" NL);
1405         }
1406
1407         netif_trans_update(ndev);
1408         ++dev->stats.tx_packets;
1409         dev->stats.tx_bytes += len;
1410
1411         return NETDEV_TX_OK;
1412 }
1413
1414 /* Tx lock BH */
1415 static int emac_start_xmit(struct sk_buff *skb, struct net_device *ndev)
1416 {
1417         struct emac_instance *dev = netdev_priv(ndev);
1418         unsigned int len = skb->len;
1419         int slot;
1420
1421         u16 ctrl = EMAC_TX_CTRL_GFCS | EMAC_TX_CTRL_GP | MAL_TX_CTRL_READY |
1422             MAL_TX_CTRL_LAST | emac_tx_csum(dev, skb);
1423
1424         slot = dev->tx_slot++;
1425         if (dev->tx_slot == NUM_TX_BUFF) {
1426                 dev->tx_slot = 0;
1427                 ctrl |= MAL_TX_CTRL_WRAP;
1428         }
1429
1430         DBG2(dev, "xmit(%u) %d" NL, len, slot);
1431
1432         dev->tx_skb[slot] = skb;
1433         dev->tx_desc[slot].data_ptr = dma_map_single(&dev->ofdev->dev,
1434                                                      skb->data, len,
1435                                                      DMA_TO_DEVICE);
1436         dev->tx_desc[slot].data_len = (u16) len;
1437         wmb();
1438         dev->tx_desc[slot].ctrl = ctrl;
1439
1440         return emac_xmit_finish(dev, len);
1441 }
1442
1443 static inline int emac_xmit_split(struct emac_instance *dev, int slot,
1444                                   u32 pd, int len, int last, u16 base_ctrl)
1445 {
1446         while (1) {
1447                 u16 ctrl = base_ctrl;
1448                 int chunk = min(len, MAL_MAX_TX_SIZE);
1449                 len -= chunk;
1450
1451                 slot = (slot + 1) % NUM_TX_BUFF;
1452
1453                 if (last && !len)
1454                         ctrl |= MAL_TX_CTRL_LAST;
1455                 if (slot == NUM_TX_BUFF - 1)
1456                         ctrl |= MAL_TX_CTRL_WRAP;
1457
1458                 dev->tx_skb[slot] = NULL;
1459                 dev->tx_desc[slot].data_ptr = pd;
1460                 dev->tx_desc[slot].data_len = (u16) chunk;
1461                 dev->tx_desc[slot].ctrl = ctrl;
1462                 ++dev->tx_cnt;
1463
1464                 if (!len)
1465                         break;
1466
1467                 pd += chunk;
1468         }
1469         return slot;
1470 }
1471
1472 /* Tx lock BH disabled (SG version for TAH equipped EMACs) */
1473 static int emac_start_xmit_sg(struct sk_buff *skb, struct net_device *ndev)
1474 {
1475         struct emac_instance *dev = netdev_priv(ndev);
1476         int nr_frags = skb_shinfo(skb)->nr_frags;
1477         int len = skb->len, chunk;
1478         int slot, i;
1479         u16 ctrl;
1480         u32 pd;
1481
1482         /* This is common "fast" path */
1483         if (likely(!nr_frags && len <= MAL_MAX_TX_SIZE))
1484                 return emac_start_xmit(skb, ndev);
1485
1486         len -= skb->data_len;
1487
1488         /* Note, this is only an *estimation*, we can still run out of empty
1489          * slots because of the additional fragmentation into
1490          * MAL_MAX_TX_SIZE-sized chunks
1491          */
1492         if (unlikely(dev->tx_cnt + nr_frags + mal_tx_chunks(len) > NUM_TX_BUFF))
1493                 goto stop_queue;
1494
1495         ctrl = EMAC_TX_CTRL_GFCS | EMAC_TX_CTRL_GP | MAL_TX_CTRL_READY |
1496             emac_tx_csum(dev, skb);
1497         slot = dev->tx_slot;
1498
1499         /* skb data */
1500         dev->tx_skb[slot] = NULL;
1501         chunk = min(len, MAL_MAX_TX_SIZE);
1502         dev->tx_desc[slot].data_ptr = pd =
1503             dma_map_single(&dev->ofdev->dev, skb->data, len, DMA_TO_DEVICE);
1504         dev->tx_desc[slot].data_len = (u16) chunk;
1505         len -= chunk;
1506         if (unlikely(len))
1507                 slot = emac_xmit_split(dev, slot, pd + chunk, len, !nr_frags,
1508                                        ctrl);
1509         /* skb fragments */
1510         for (i = 0; i < nr_frags; ++i) {
1511                 struct skb_frag_struct *frag = &skb_shinfo(skb)->frags[i];
1512                 len = skb_frag_size(frag);
1513
1514                 if (unlikely(dev->tx_cnt + mal_tx_chunks(len) >= NUM_TX_BUFF))
1515                         goto undo_frame;
1516
1517                 pd = skb_frag_dma_map(&dev->ofdev->dev, frag, 0, len,
1518                                       DMA_TO_DEVICE);
1519
1520                 slot = emac_xmit_split(dev, slot, pd, len, i == nr_frags - 1,
1521                                        ctrl);
1522         }
1523
1524         DBG2(dev, "xmit_sg(%u) %d - %d" NL, skb->len, dev->tx_slot, slot);
1525
1526         /* Attach skb to the last slot so we don't release it too early */
1527         dev->tx_skb[slot] = skb;
1528
1529         /* Send the packet out */
1530         if (dev->tx_slot == NUM_TX_BUFF - 1)
1531                 ctrl |= MAL_TX_CTRL_WRAP;
1532         wmb();
1533         dev->tx_desc[dev->tx_slot].ctrl = ctrl;
1534         dev->tx_slot = (slot + 1) % NUM_TX_BUFF;
1535
1536         return emac_xmit_finish(dev, skb->len);
1537
1538  undo_frame:
1539         /* Well, too bad. Our previous estimation was overly optimistic.
1540          * Undo everything.
1541          */
1542         while (slot != dev->tx_slot) {
1543                 dev->tx_desc[slot].ctrl = 0;
1544                 --dev->tx_cnt;
1545                 if (--slot < 0)
1546                         slot = NUM_TX_BUFF - 1;
1547         }
1548         ++dev->estats.tx_undo;
1549
1550  stop_queue:
1551         netif_stop_queue(ndev);
1552         DBG2(dev, "stopped TX queue" NL);
1553         return NETDEV_TX_BUSY;
1554 }
1555
1556 /* Tx lock BHs */
1557 static void emac_parse_tx_error(struct emac_instance *dev, u16 ctrl)
1558 {
1559         struct emac_error_stats *st = &dev->estats;
1560
1561         DBG(dev, "BD TX error %04x" NL, ctrl);
1562
1563         ++st->tx_bd_errors;
1564         if (ctrl & EMAC_TX_ST_BFCS)
1565                 ++st->tx_bd_bad_fcs;
1566         if (ctrl & EMAC_TX_ST_LCS)
1567                 ++st->tx_bd_carrier_loss;
1568         if (ctrl & EMAC_TX_ST_ED)
1569                 ++st->tx_bd_excessive_deferral;
1570         if (ctrl & EMAC_TX_ST_EC)
1571                 ++st->tx_bd_excessive_collisions;
1572         if (ctrl & EMAC_TX_ST_LC)
1573                 ++st->tx_bd_late_collision;
1574         if (ctrl & EMAC_TX_ST_MC)
1575                 ++st->tx_bd_multple_collisions;
1576         if (ctrl & EMAC_TX_ST_SC)
1577                 ++st->tx_bd_single_collision;
1578         if (ctrl & EMAC_TX_ST_UR)
1579                 ++st->tx_bd_underrun;
1580         if (ctrl & EMAC_TX_ST_SQE)
1581                 ++st->tx_bd_sqe;
1582 }
1583
1584 static void emac_poll_tx(void *param)
1585 {
1586         struct emac_instance *dev = param;
1587         u32 bad_mask;
1588
1589         DBG2(dev, "poll_tx, %d %d" NL, dev->tx_cnt, dev->ack_slot);
1590
1591         if (emac_has_feature(dev, EMAC_FTR_HAS_TAH))
1592                 bad_mask = EMAC_IS_BAD_TX_TAH;
1593         else
1594                 bad_mask = EMAC_IS_BAD_TX;
1595
1596         netif_tx_lock_bh(dev->ndev);
1597         if (dev->tx_cnt) {
1598                 u16 ctrl;
1599                 int slot = dev->ack_slot, n = 0;
1600         again:
1601                 ctrl = dev->tx_desc[slot].ctrl;
1602                 if (!(ctrl & MAL_TX_CTRL_READY)) {
1603                         struct sk_buff *skb = dev->tx_skb[slot];
1604                         ++n;
1605
1606                         if (skb) {
1607                                 dev_kfree_skb(skb);
1608                                 dev->tx_skb[slot] = NULL;
1609                         }
1610                         slot = (slot + 1) % NUM_TX_BUFF;
1611
1612                         if (unlikely(ctrl & bad_mask))
1613                                 emac_parse_tx_error(dev, ctrl);
1614
1615                         if (--dev->tx_cnt)
1616                                 goto again;
1617                 }
1618                 if (n) {
1619                         dev->ack_slot = slot;
1620                         if (netif_queue_stopped(dev->ndev) &&
1621                             dev->tx_cnt < EMAC_TX_WAKEUP_THRESH)
1622                                 netif_wake_queue(dev->ndev);
1623
1624                         DBG2(dev, "tx %d pkts" NL, n);
1625                 }
1626         }
1627         netif_tx_unlock_bh(dev->ndev);
1628 }
1629
1630 static inline void emac_recycle_rx_skb(struct emac_instance *dev, int slot,
1631                                        int len)
1632 {
1633         struct sk_buff *skb = dev->rx_skb[slot];
1634
1635         DBG2(dev, "recycle %d %d" NL, slot, len);
1636
1637         if (len)
1638                 dma_map_single(&dev->ofdev->dev, skb->data - 2,
1639                                EMAC_DMA_ALIGN(len + 2), DMA_FROM_DEVICE);
1640
1641         dev->rx_desc[slot].data_len = 0;
1642         wmb();
1643         dev->rx_desc[slot].ctrl = MAL_RX_CTRL_EMPTY |
1644             (slot == (NUM_RX_BUFF - 1) ? MAL_RX_CTRL_WRAP : 0);
1645 }
1646
1647 static void emac_parse_rx_error(struct emac_instance *dev, u16 ctrl)
1648 {
1649         struct emac_error_stats *st = &dev->estats;
1650
1651         DBG(dev, "BD RX error %04x" NL, ctrl);
1652
1653         ++st->rx_bd_errors;
1654         if (ctrl & EMAC_RX_ST_OE)
1655                 ++st->rx_bd_overrun;
1656         if (ctrl & EMAC_RX_ST_BP)
1657                 ++st->rx_bd_bad_packet;
1658         if (ctrl & EMAC_RX_ST_RP)
1659                 ++st->rx_bd_runt_packet;
1660         if (ctrl & EMAC_RX_ST_SE)
1661                 ++st->rx_bd_short_event;
1662         if (ctrl & EMAC_RX_ST_AE)
1663                 ++st->rx_bd_alignment_error;
1664         if (ctrl & EMAC_RX_ST_BFCS)
1665                 ++st->rx_bd_bad_fcs;
1666         if (ctrl & EMAC_RX_ST_PTL)
1667                 ++st->rx_bd_packet_too_long;
1668         if (ctrl & EMAC_RX_ST_ORE)
1669                 ++st->rx_bd_out_of_range;
1670         if (ctrl & EMAC_RX_ST_IRE)
1671                 ++st->rx_bd_in_range;
1672 }
1673
1674 static inline void emac_rx_csum(struct emac_instance *dev,
1675                                 struct sk_buff *skb, u16 ctrl)
1676 {
1677 #ifdef CONFIG_IBM_EMAC_TAH
1678         if (!ctrl && dev->tah_dev) {
1679                 skb->ip_summed = CHECKSUM_UNNECESSARY;
1680                 ++dev->stats.rx_packets_csum;
1681         }
1682 #endif
1683 }
1684
1685 static inline int emac_rx_sg_append(struct emac_instance *dev, int slot)
1686 {
1687         if (likely(dev->rx_sg_skb != NULL)) {
1688                 int len = dev->rx_desc[slot].data_len;
1689                 int tot_len = dev->rx_sg_skb->len + len;
1690
1691                 if (unlikely(tot_len + 2 > dev->rx_skb_size)) {
1692                         ++dev->estats.rx_dropped_mtu;
1693                         dev_kfree_skb(dev->rx_sg_skb);
1694                         dev->rx_sg_skb = NULL;
1695                 } else {
1696                         memcpy(skb_tail_pointer(dev->rx_sg_skb),
1697                                          dev->rx_skb[slot]->data, len);
1698                         skb_put(dev->rx_sg_skb, len);
1699                         emac_recycle_rx_skb(dev, slot, len);
1700                         return 0;
1701                 }
1702         }
1703         emac_recycle_rx_skb(dev, slot, 0);
1704         return -1;
1705 }
1706
1707 /* NAPI poll context */
1708 static int emac_poll_rx(void *param, int budget)
1709 {
1710         struct emac_instance *dev = param;
1711         int slot = dev->rx_slot, received = 0;
1712
1713         DBG2(dev, "poll_rx(%d)" NL, budget);
1714
1715  again:
1716         while (budget > 0) {
1717                 int len;
1718                 struct sk_buff *skb;
1719                 u16 ctrl = dev->rx_desc[slot].ctrl;
1720
1721                 if (ctrl & MAL_RX_CTRL_EMPTY)
1722                         break;
1723
1724                 skb = dev->rx_skb[slot];
1725                 mb();
1726                 len = dev->rx_desc[slot].data_len;
1727
1728                 if (unlikely(!MAL_IS_SINGLE_RX(ctrl)))
1729                         goto sg;
1730
1731                 ctrl &= EMAC_BAD_RX_MASK;
1732                 if (unlikely(ctrl && ctrl != EMAC_RX_TAH_BAD_CSUM)) {
1733                         emac_parse_rx_error(dev, ctrl);
1734                         ++dev->estats.rx_dropped_error;
1735                         emac_recycle_rx_skb(dev, slot, 0);
1736                         len = 0;
1737                         goto next;
1738                 }
1739
1740                 if (len < ETH_HLEN) {
1741                         ++dev->estats.rx_dropped_stack;
1742                         emac_recycle_rx_skb(dev, slot, len);
1743                         goto next;
1744                 }
1745
1746                 if (len && len < EMAC_RX_COPY_THRESH) {
1747                         struct sk_buff *copy_skb =
1748                             alloc_skb(len + EMAC_RX_SKB_HEADROOM + 2, GFP_ATOMIC);
1749                         if (unlikely(!copy_skb))
1750                                 goto oom;
1751
1752                         skb_reserve(copy_skb, EMAC_RX_SKB_HEADROOM + 2);
1753                         memcpy(copy_skb->data - 2, skb->data - 2, len + 2);
1754                         emac_recycle_rx_skb(dev, slot, len);
1755                         skb = copy_skb;
1756                 } else if (unlikely(emac_alloc_rx_skb(dev, slot, GFP_ATOMIC)))
1757                         goto oom;
1758
1759                 skb_put(skb, len);
1760         push_packet:
1761                 skb->protocol = eth_type_trans(skb, dev->ndev);
1762                 emac_rx_csum(dev, skb, ctrl);
1763
1764                 if (unlikely(netif_receive_skb(skb) == NET_RX_DROP))
1765                         ++dev->estats.rx_dropped_stack;
1766         next:
1767                 ++dev->stats.rx_packets;
1768         skip:
1769                 dev->stats.rx_bytes += len;
1770                 slot = (slot + 1) % NUM_RX_BUFF;
1771                 --budget;
1772                 ++received;
1773                 continue;
1774         sg:
1775                 if (ctrl & MAL_RX_CTRL_FIRST) {
1776                         BUG_ON(dev->rx_sg_skb);
1777                         if (unlikely(emac_alloc_rx_skb(dev, slot, GFP_ATOMIC))) {
1778                                 DBG(dev, "rx OOM %d" NL, slot);
1779                                 ++dev->estats.rx_dropped_oom;
1780                                 emac_recycle_rx_skb(dev, slot, 0);
1781                         } else {
1782                                 dev->rx_sg_skb = skb;
1783                                 skb_put(skb, len);
1784                         }
1785                 } else if (!emac_rx_sg_append(dev, slot) &&
1786                            (ctrl & MAL_RX_CTRL_LAST)) {
1787
1788                         skb = dev->rx_sg_skb;
1789                         dev->rx_sg_skb = NULL;
1790
1791                         ctrl &= EMAC_BAD_RX_MASK;
1792                         if (unlikely(ctrl && ctrl != EMAC_RX_TAH_BAD_CSUM)) {
1793                                 emac_parse_rx_error(dev, ctrl);
1794                                 ++dev->estats.rx_dropped_error;
1795                                 dev_kfree_skb(skb);
1796                                 len = 0;
1797                         } else
1798                                 goto push_packet;
1799                 }
1800                 goto skip;
1801         oom:
1802                 DBG(dev, "rx OOM %d" NL, slot);
1803                 /* Drop the packet and recycle skb */
1804                 ++dev->estats.rx_dropped_oom;
1805                 emac_recycle_rx_skb(dev, slot, 0);
1806                 goto next;
1807         }
1808
1809         if (received) {
1810                 DBG2(dev, "rx %d BDs" NL, received);
1811                 dev->rx_slot = slot;
1812         }
1813
1814         if (unlikely(budget && test_bit(MAL_COMMAC_RX_STOPPED, &dev->commac.flags))) {
1815                 mb();
1816                 if (!(dev->rx_desc[slot].ctrl & MAL_RX_CTRL_EMPTY)) {
1817                         DBG2(dev, "rx restart" NL);
1818                         received = 0;
1819                         goto again;
1820                 }
1821
1822                 if (dev->rx_sg_skb) {
1823                         DBG2(dev, "dropping partial rx packet" NL);
1824                         ++dev->estats.rx_dropped_error;
1825                         dev_kfree_skb(dev->rx_sg_skb);
1826                         dev->rx_sg_skb = NULL;
1827                 }
1828
1829                 clear_bit(MAL_COMMAC_RX_STOPPED, &dev->commac.flags);
1830                 mal_enable_rx_channel(dev->mal, dev->mal_rx_chan);
1831                 emac_rx_enable(dev);
1832                 dev->rx_slot = 0;
1833         }
1834         return received;
1835 }
1836
1837 /* NAPI poll context */
1838 static int emac_peek_rx(void *param)
1839 {
1840         struct emac_instance *dev = param;
1841
1842         return !(dev->rx_desc[dev->rx_slot].ctrl & MAL_RX_CTRL_EMPTY);
1843 }
1844
1845 /* NAPI poll context */
1846 static int emac_peek_rx_sg(void *param)
1847 {
1848         struct emac_instance *dev = param;
1849
1850         int slot = dev->rx_slot;
1851         while (1) {
1852                 u16 ctrl = dev->rx_desc[slot].ctrl;
1853                 if (ctrl & MAL_RX_CTRL_EMPTY)
1854                         return 0;
1855                 else if (ctrl & MAL_RX_CTRL_LAST)
1856                         return 1;
1857
1858                 slot = (slot + 1) % NUM_RX_BUFF;
1859
1860                 /* I'm just being paranoid here :) */
1861                 if (unlikely(slot == dev->rx_slot))
1862                         return 0;
1863         }
1864 }
1865
1866 /* Hard IRQ */
1867 static void emac_rxde(void *param)
1868 {
1869         struct emac_instance *dev = param;
1870
1871         ++dev->estats.rx_stopped;
1872         emac_rx_disable_async(dev);
1873 }
1874
1875 /* Hard IRQ */
1876 static irqreturn_t emac_irq(int irq, void *dev_instance)
1877 {
1878         struct emac_instance *dev = dev_instance;
1879         struct emac_regs __iomem *p = dev->emacp;
1880         struct emac_error_stats *st = &dev->estats;
1881         u32 isr;
1882
1883         spin_lock(&dev->lock);
1884
1885         isr = in_be32(&p->isr);
1886         out_be32(&p->isr, isr);
1887
1888         DBG(dev, "isr = %08x" NL, isr);
1889
1890         if (isr & EMAC4_ISR_TXPE)
1891                 ++st->tx_parity;
1892         if (isr & EMAC4_ISR_RXPE)
1893                 ++st->rx_parity;
1894         if (isr & EMAC4_ISR_TXUE)
1895                 ++st->tx_underrun;
1896         if (isr & EMAC4_ISR_RXOE)
1897                 ++st->rx_fifo_overrun;
1898         if (isr & EMAC_ISR_OVR)
1899                 ++st->rx_overrun;
1900         if (isr & EMAC_ISR_BP)
1901                 ++st->rx_bad_packet;
1902         if (isr & EMAC_ISR_RP)
1903                 ++st->rx_runt_packet;
1904         if (isr & EMAC_ISR_SE)
1905                 ++st->rx_short_event;
1906         if (isr & EMAC_ISR_ALE)
1907                 ++st->rx_alignment_error;
1908         if (isr & EMAC_ISR_BFCS)
1909                 ++st->rx_bad_fcs;
1910         if (isr & EMAC_ISR_PTLE)
1911                 ++st->rx_packet_too_long;
1912         if (isr & EMAC_ISR_ORE)
1913                 ++st->rx_out_of_range;
1914         if (isr & EMAC_ISR_IRE)
1915                 ++st->rx_in_range;
1916         if (isr & EMAC_ISR_SQE)
1917                 ++st->tx_sqe;
1918         if (isr & EMAC_ISR_TE)
1919                 ++st->tx_errors;
1920
1921         spin_unlock(&dev->lock);
1922
1923         return IRQ_HANDLED;
1924 }
1925
1926 static struct net_device_stats *emac_stats(struct net_device *ndev)
1927 {
1928         struct emac_instance *dev = netdev_priv(ndev);
1929         struct emac_stats *st = &dev->stats;
1930         struct emac_error_stats *est = &dev->estats;
1931         struct net_device_stats *nst = &dev->nstats;
1932         unsigned long flags;
1933
1934         DBG2(dev, "stats" NL);
1935
1936         /* Compute "legacy" statistics */
1937         spin_lock_irqsave(&dev->lock, flags);
1938         nst->rx_packets = (unsigned long)st->rx_packets;
1939         nst->rx_bytes = (unsigned long)st->rx_bytes;
1940         nst->tx_packets = (unsigned long)st->tx_packets;
1941         nst->tx_bytes = (unsigned long)st->tx_bytes;
1942         nst->rx_dropped = (unsigned long)(est->rx_dropped_oom +
1943                                           est->rx_dropped_error +
1944                                           est->rx_dropped_resize +
1945                                           est->rx_dropped_mtu);
1946         nst->tx_dropped = (unsigned long)est->tx_dropped;
1947
1948         nst->rx_errors = (unsigned long)est->rx_bd_errors;
1949         nst->rx_fifo_errors = (unsigned long)(est->rx_bd_overrun +
1950                                               est->rx_fifo_overrun +
1951                                               est->rx_overrun);
1952         nst->rx_frame_errors = (unsigned long)(est->rx_bd_alignment_error +
1953                                                est->rx_alignment_error);
1954         nst->rx_crc_errors = (unsigned long)(est->rx_bd_bad_fcs +
1955                                              est->rx_bad_fcs);
1956         nst->rx_length_errors = (unsigned long)(est->rx_bd_runt_packet +
1957                                                 est->rx_bd_short_event +
1958                                                 est->rx_bd_packet_too_long +
1959                                                 est->rx_bd_out_of_range +
1960                                                 est->rx_bd_in_range +
1961                                                 est->rx_runt_packet +
1962                                                 est->rx_short_event +
1963                                                 est->rx_packet_too_long +
1964                                                 est->rx_out_of_range +
1965                                                 est->rx_in_range);
1966
1967         nst->tx_errors = (unsigned long)(est->tx_bd_errors + est->tx_errors);
1968         nst->tx_fifo_errors = (unsigned long)(est->tx_bd_underrun +
1969                                               est->tx_underrun);
1970         nst->tx_carrier_errors = (unsigned long)est->tx_bd_carrier_loss;
1971         nst->collisions = (unsigned long)(est->tx_bd_excessive_deferral +
1972                                           est->tx_bd_excessive_collisions +
1973                                           est->tx_bd_late_collision +
1974                                           est->tx_bd_multple_collisions);
1975         spin_unlock_irqrestore(&dev->lock, flags);
1976         return nst;
1977 }
1978
1979 static struct mal_commac_ops emac_commac_ops = {
1980         .poll_tx = &emac_poll_tx,
1981         .poll_rx = &emac_poll_rx,
1982         .peek_rx = &emac_peek_rx,
1983         .rxde = &emac_rxde,
1984 };
1985
1986 static struct mal_commac_ops emac_commac_sg_ops = {
1987         .poll_tx = &emac_poll_tx,
1988         .poll_rx = &emac_poll_rx,
1989         .peek_rx = &emac_peek_rx_sg,
1990         .rxde = &emac_rxde,
1991 };
1992
1993 /* Ethtool support */
1994 static int emac_ethtool_get_settings(struct net_device *ndev,
1995                                      struct ethtool_cmd *cmd)
1996 {
1997         struct emac_instance *dev = netdev_priv(ndev);
1998
1999         cmd->supported = dev->phy.features;
2000         cmd->port = PORT_MII;
2001         cmd->phy_address = dev->phy.address;
2002         cmd->transceiver =
2003             dev->phy.address >= 0 ? XCVR_EXTERNAL : XCVR_INTERNAL;
2004
2005         mutex_lock(&dev->link_lock);
2006         cmd->advertising = dev->phy.advertising;
2007         cmd->autoneg = dev->phy.autoneg;
2008         cmd->speed = dev->phy.speed;
2009         cmd->duplex = dev->phy.duplex;
2010         mutex_unlock(&dev->link_lock);
2011
2012         return 0;
2013 }
2014
2015 static int emac_ethtool_set_settings(struct net_device *ndev,
2016                                      struct ethtool_cmd *cmd)
2017 {
2018         struct emac_instance *dev = netdev_priv(ndev);
2019         u32 f = dev->phy.features;
2020
2021         DBG(dev, "set_settings(%d, %d, %d, 0x%08x)" NL,
2022             cmd->autoneg, cmd->speed, cmd->duplex, cmd->advertising);
2023
2024         /* Basic sanity checks */
2025         if (dev->phy.address < 0)
2026                 return -EOPNOTSUPP;
2027         if (cmd->autoneg != AUTONEG_ENABLE && cmd->autoneg != AUTONEG_DISABLE)
2028                 return -EINVAL;
2029         if (cmd->autoneg == AUTONEG_ENABLE && cmd->advertising == 0)
2030                 return -EINVAL;
2031         if (cmd->duplex != DUPLEX_HALF && cmd->duplex != DUPLEX_FULL)
2032                 return -EINVAL;
2033
2034         if (cmd->autoneg == AUTONEG_DISABLE) {
2035                 switch (cmd->speed) {
2036                 case SPEED_10:
2037                         if (cmd->duplex == DUPLEX_HALF &&
2038                             !(f & SUPPORTED_10baseT_Half))
2039                                 return -EINVAL;
2040                         if (cmd->duplex == DUPLEX_FULL &&
2041                             !(f & SUPPORTED_10baseT_Full))
2042                                 return -EINVAL;
2043                         break;
2044                 case SPEED_100:
2045                         if (cmd->duplex == DUPLEX_HALF &&
2046                             !(f & SUPPORTED_100baseT_Half))
2047                                 return -EINVAL;
2048                         if (cmd->duplex == DUPLEX_FULL &&
2049                             !(f & SUPPORTED_100baseT_Full))
2050                                 return -EINVAL;
2051                         break;
2052                 case SPEED_1000:
2053                         if (cmd->duplex == DUPLEX_HALF &&
2054                             !(f & SUPPORTED_1000baseT_Half))
2055                                 return -EINVAL;
2056                         if (cmd->duplex == DUPLEX_FULL &&
2057                             !(f & SUPPORTED_1000baseT_Full))
2058                                 return -EINVAL;
2059                         break;
2060                 default:
2061                         return -EINVAL;
2062                 }
2063
2064                 mutex_lock(&dev->link_lock);
2065                 dev->phy.def->ops->setup_forced(&dev->phy, cmd->speed,
2066                                                 cmd->duplex);
2067                 mutex_unlock(&dev->link_lock);
2068
2069         } else {
2070                 if (!(f & SUPPORTED_Autoneg))
2071                         return -EINVAL;
2072
2073                 mutex_lock(&dev->link_lock);
2074                 dev->phy.def->ops->setup_aneg(&dev->phy,
2075                                               (cmd->advertising & f) |
2076                                               (dev->phy.advertising &
2077                                                (ADVERTISED_Pause |
2078                                                 ADVERTISED_Asym_Pause)));
2079                 mutex_unlock(&dev->link_lock);
2080         }
2081         emac_force_link_update(dev);
2082
2083         return 0;
2084 }
2085
2086 static void emac_ethtool_get_ringparam(struct net_device *ndev,
2087                                        struct ethtool_ringparam *rp)
2088 {
2089         rp->rx_max_pending = rp->rx_pending = NUM_RX_BUFF;
2090         rp->tx_max_pending = rp->tx_pending = NUM_TX_BUFF;
2091 }
2092
2093 static void emac_ethtool_get_pauseparam(struct net_device *ndev,
2094                                         struct ethtool_pauseparam *pp)
2095 {
2096         struct emac_instance *dev = netdev_priv(ndev);
2097
2098         mutex_lock(&dev->link_lock);
2099         if ((dev->phy.features & SUPPORTED_Autoneg) &&
2100             (dev->phy.advertising & (ADVERTISED_Pause | ADVERTISED_Asym_Pause)))
2101                 pp->autoneg = 1;
2102
2103         if (dev->phy.duplex == DUPLEX_FULL) {
2104                 if (dev->phy.pause)
2105                         pp->rx_pause = pp->tx_pause = 1;
2106                 else if (dev->phy.asym_pause)
2107                         pp->tx_pause = 1;
2108         }
2109         mutex_unlock(&dev->link_lock);
2110 }
2111
2112 static int emac_get_regs_len(struct emac_instance *dev)
2113 {
2114                 return sizeof(struct emac_ethtool_regs_subhdr) +
2115                         sizeof(struct emac_regs);
2116 }
2117
2118 static int emac_ethtool_get_regs_len(struct net_device *ndev)
2119 {
2120         struct emac_instance *dev = netdev_priv(ndev);
2121         int size;
2122
2123         size = sizeof(struct emac_ethtool_regs_hdr) +
2124                 emac_get_regs_len(dev) + mal_get_regs_len(dev->mal);
2125         if (emac_has_feature(dev, EMAC_FTR_HAS_ZMII))
2126                 size += zmii_get_regs_len(dev->zmii_dev);
2127         if (emac_has_feature(dev, EMAC_FTR_HAS_RGMII))
2128                 size += rgmii_get_regs_len(dev->rgmii_dev);
2129         if (emac_has_feature(dev, EMAC_FTR_HAS_TAH))
2130                 size += tah_get_regs_len(dev->tah_dev);
2131
2132         return size;
2133 }
2134
2135 static void *emac_dump_regs(struct emac_instance *dev, void *buf)
2136 {
2137         struct emac_ethtool_regs_subhdr *hdr = buf;
2138
2139         hdr->index = dev->cell_index;
2140         if (emac_has_feature(dev, EMAC_FTR_EMAC4SYNC)) {
2141                 hdr->version = EMAC4SYNC_ETHTOOL_REGS_VER;
2142         } else if (emac_has_feature(dev, EMAC_FTR_EMAC4)) {
2143                 hdr->version = EMAC4_ETHTOOL_REGS_VER;
2144         } else {
2145                 hdr->version = EMAC_ETHTOOL_REGS_VER;
2146         }
2147         memcpy_fromio(hdr + 1, dev->emacp, sizeof(struct emac_regs));
2148         return (void *)(hdr + 1) + sizeof(struct emac_regs);
2149 }
2150
2151 static void emac_ethtool_get_regs(struct net_device *ndev,
2152                                   struct ethtool_regs *regs, void *buf)
2153 {
2154         struct emac_instance *dev = netdev_priv(ndev);
2155         struct emac_ethtool_regs_hdr *hdr = buf;
2156
2157         hdr->components = 0;
2158         buf = hdr + 1;
2159
2160         buf = mal_dump_regs(dev->mal, buf);
2161         buf = emac_dump_regs(dev, buf);
2162         if (emac_has_feature(dev, EMAC_FTR_HAS_ZMII)) {
2163                 hdr->components |= EMAC_ETHTOOL_REGS_ZMII;
2164                 buf = zmii_dump_regs(dev->zmii_dev, buf);
2165         }
2166         if (emac_has_feature(dev, EMAC_FTR_HAS_RGMII)) {
2167                 hdr->components |= EMAC_ETHTOOL_REGS_RGMII;
2168                 buf = rgmii_dump_regs(dev->rgmii_dev, buf);
2169         }
2170         if (emac_has_feature(dev, EMAC_FTR_HAS_TAH)) {
2171                 hdr->components |= EMAC_ETHTOOL_REGS_TAH;
2172                 buf = tah_dump_regs(dev->tah_dev, buf);
2173         }
2174 }
2175
2176 static int emac_ethtool_nway_reset(struct net_device *ndev)
2177 {
2178         struct emac_instance *dev = netdev_priv(ndev);
2179         int res = 0;
2180
2181         DBG(dev, "nway_reset" NL);
2182
2183         if (dev->phy.address < 0)
2184                 return -EOPNOTSUPP;
2185
2186         mutex_lock(&dev->link_lock);
2187         if (!dev->phy.autoneg) {
2188                 res = -EINVAL;
2189                 goto out;
2190         }
2191
2192         dev->phy.def->ops->setup_aneg(&dev->phy, dev->phy.advertising);
2193  out:
2194         mutex_unlock(&dev->link_lock);
2195         emac_force_link_update(dev);
2196         return res;
2197 }
2198
2199 static int emac_ethtool_get_sset_count(struct net_device *ndev, int stringset)
2200 {
2201         if (stringset == ETH_SS_STATS)
2202                 return EMAC_ETHTOOL_STATS_COUNT;
2203         else
2204                 return -EINVAL;
2205 }
2206
2207 static void emac_ethtool_get_strings(struct net_device *ndev, u32 stringset,
2208                                      u8 * buf)
2209 {
2210         if (stringset == ETH_SS_STATS)
2211                 memcpy(buf, &emac_stats_keys, sizeof(emac_stats_keys));
2212 }
2213
2214 static void emac_ethtool_get_ethtool_stats(struct net_device *ndev,
2215                                            struct ethtool_stats *estats,
2216                                            u64 * tmp_stats)
2217 {
2218         struct emac_instance *dev = netdev_priv(ndev);
2219
2220         memcpy(tmp_stats, &dev->stats, sizeof(dev->stats));
2221         tmp_stats += sizeof(dev->stats) / sizeof(u64);
2222         memcpy(tmp_stats, &dev->estats, sizeof(dev->estats));
2223 }
2224
2225 static void emac_ethtool_get_drvinfo(struct net_device *ndev,
2226                                      struct ethtool_drvinfo *info)
2227 {
2228         struct emac_instance *dev = netdev_priv(ndev);
2229
2230         strlcpy(info->driver, "ibm_emac", sizeof(info->driver));
2231         strlcpy(info->version, DRV_VERSION, sizeof(info->version));
2232         snprintf(info->bus_info, sizeof(info->bus_info), "PPC 4xx EMAC-%d %s",
2233                  dev->cell_index, dev->ofdev->dev.of_node->full_name);
2234 }
2235
2236 static const struct ethtool_ops emac_ethtool_ops = {
2237         .get_settings = emac_ethtool_get_settings,
2238         .set_settings = emac_ethtool_set_settings,
2239         .get_drvinfo = emac_ethtool_get_drvinfo,
2240
2241         .get_regs_len = emac_ethtool_get_regs_len,
2242         .get_regs = emac_ethtool_get_regs,
2243
2244         .nway_reset = emac_ethtool_nway_reset,
2245
2246         .get_ringparam = emac_ethtool_get_ringparam,
2247         .get_pauseparam = emac_ethtool_get_pauseparam,
2248
2249         .get_strings = emac_ethtool_get_strings,
2250         .get_sset_count = emac_ethtool_get_sset_count,
2251         .get_ethtool_stats = emac_ethtool_get_ethtool_stats,
2252
2253         .get_link = ethtool_op_get_link,
2254 };
2255
2256 static int emac_ioctl(struct net_device *ndev, struct ifreq *rq, int cmd)
2257 {
2258         struct emac_instance *dev = netdev_priv(ndev);
2259         struct mii_ioctl_data *data = if_mii(rq);
2260
2261         DBG(dev, "ioctl %08x" NL, cmd);
2262
2263         if (dev->phy.address < 0)
2264                 return -EOPNOTSUPP;
2265
2266         switch (cmd) {
2267         case SIOCGMIIPHY:
2268                 data->phy_id = dev->phy.address;
2269                 /* Fall through */
2270         case SIOCGMIIREG:
2271                 data->val_out = emac_mdio_read(ndev, dev->phy.address,
2272                                                data->reg_num);
2273                 return 0;
2274
2275         case SIOCSMIIREG:
2276                 emac_mdio_write(ndev, dev->phy.address, data->reg_num,
2277                                 data->val_in);
2278                 return 0;
2279         default:
2280                 return -EOPNOTSUPP;
2281         }
2282 }
2283
2284 struct emac_depentry {
2285         u32                     phandle;
2286         struct device_node      *node;
2287         struct platform_device  *ofdev;
2288         void                    *drvdata;
2289 };
2290
2291 #define EMAC_DEP_MAL_IDX        0
2292 #define EMAC_DEP_ZMII_IDX       1
2293 #define EMAC_DEP_RGMII_IDX      2
2294 #define EMAC_DEP_TAH_IDX        3
2295 #define EMAC_DEP_MDIO_IDX       4
2296 #define EMAC_DEP_PREV_IDX       5
2297 #define EMAC_DEP_COUNT          6
2298
2299 static int emac_check_deps(struct emac_instance *dev,
2300                            struct emac_depentry *deps)
2301 {
2302         int i, there = 0;
2303         struct device_node *np;
2304
2305         for (i = 0; i < EMAC_DEP_COUNT; i++) {
2306                 /* no dependency on that item, allright */
2307                 if (deps[i].phandle == 0) {
2308                         there++;
2309                         continue;
2310                 }
2311                 /* special case for blist as the dependency might go away */
2312                 if (i == EMAC_DEP_PREV_IDX) {
2313                         np = *(dev->blist - 1);
2314                         if (np == NULL) {
2315                                 deps[i].phandle = 0;
2316                                 there++;
2317                                 continue;
2318                         }
2319                         if (deps[i].node == NULL)
2320                                 deps[i].node = of_node_get(np);
2321                 }
2322                 if (deps[i].node == NULL)
2323                         deps[i].node = of_find_node_by_phandle(deps[i].phandle);
2324                 if (deps[i].node == NULL)
2325                         continue;
2326                 if (deps[i].ofdev == NULL)
2327                         deps[i].ofdev = of_find_device_by_node(deps[i].node);
2328                 if (deps[i].ofdev == NULL)
2329                         continue;
2330                 if (deps[i].drvdata == NULL)
2331                         deps[i].drvdata = platform_get_drvdata(deps[i].ofdev);
2332                 if (deps[i].drvdata != NULL)
2333                         there++;
2334         }
2335         return there == EMAC_DEP_COUNT;
2336 }
2337
2338 static void emac_put_deps(struct emac_instance *dev)
2339 {
2340         of_dev_put(dev->mal_dev);
2341         of_dev_put(dev->zmii_dev);
2342         of_dev_put(dev->rgmii_dev);
2343         of_dev_put(dev->mdio_dev);
2344         of_dev_put(dev->tah_dev);
2345 }
2346
2347 static int emac_of_bus_notify(struct notifier_block *nb, unsigned long action,
2348                               void *data)
2349 {
2350         /* We are only intereted in device addition */
2351         if (action == BUS_NOTIFY_BOUND_DRIVER)
2352                 wake_up_all(&emac_probe_wait);
2353         return 0;
2354 }
2355
2356 static struct notifier_block emac_of_bus_notifier = {
2357         .notifier_call = emac_of_bus_notify
2358 };
2359
2360 static int emac_wait_deps(struct emac_instance *dev)
2361 {
2362         struct emac_depentry deps[EMAC_DEP_COUNT];
2363         int i, err;
2364
2365         memset(&deps, 0, sizeof(deps));
2366
2367         deps[EMAC_DEP_MAL_IDX].phandle = dev->mal_ph;
2368         deps[EMAC_DEP_ZMII_IDX].phandle = dev->zmii_ph;
2369         deps[EMAC_DEP_RGMII_IDX].phandle = dev->rgmii_ph;
2370         if (dev->tah_ph)
2371                 deps[EMAC_DEP_TAH_IDX].phandle = dev->tah_ph;
2372         if (dev->mdio_ph)
2373                 deps[EMAC_DEP_MDIO_IDX].phandle = dev->mdio_ph;
2374         if (dev->blist && dev->blist > emac_boot_list)
2375                 deps[EMAC_DEP_PREV_IDX].phandle = 0xffffffffu;
2376         bus_register_notifier(&platform_bus_type, &emac_of_bus_notifier);
2377         wait_event_timeout(emac_probe_wait,
2378                            emac_check_deps(dev, deps),
2379                            EMAC_PROBE_DEP_TIMEOUT);
2380         bus_unregister_notifier(&platform_bus_type, &emac_of_bus_notifier);
2381         err = emac_check_deps(dev, deps) ? 0 : -ENODEV;
2382         for (i = 0; i < EMAC_DEP_COUNT; i++) {
2383                 of_node_put(deps[i].node);
2384                 if (err)
2385                         of_dev_put(deps[i].ofdev);
2386         }
2387         if (err == 0) {
2388                 dev->mal_dev = deps[EMAC_DEP_MAL_IDX].ofdev;
2389                 dev->zmii_dev = deps[EMAC_DEP_ZMII_IDX].ofdev;
2390                 dev->rgmii_dev = deps[EMAC_DEP_RGMII_IDX].ofdev;
2391                 dev->tah_dev = deps[EMAC_DEP_TAH_IDX].ofdev;
2392                 dev->mdio_dev = deps[EMAC_DEP_MDIO_IDX].ofdev;
2393         }
2394         of_dev_put(deps[EMAC_DEP_PREV_IDX].ofdev);
2395         return err;
2396 }
2397
2398 static int emac_read_uint_prop(struct device_node *np, const char *name,
2399                                u32 *val, int fatal)
2400 {
2401         int len;
2402         const u32 *prop = of_get_property(np, name, &len);
2403         if (prop == NULL || len < sizeof(u32)) {
2404                 if (fatal)
2405                         printk(KERN_ERR "%s: missing %s property\n",
2406                                np->full_name, name);
2407                 return -ENODEV;
2408         }
2409         *val = *prop;
2410         return 0;
2411 }
2412
2413 static int emac_init_phy(struct emac_instance *dev)
2414 {
2415         struct device_node *np = dev->ofdev->dev.of_node;
2416         struct net_device *ndev = dev->ndev;
2417         u32 phy_map, adv;
2418         int i;
2419
2420         dev->phy.dev = ndev;
2421         dev->phy.mode = dev->phy_mode;
2422
2423         /* PHY-less configuration.
2424          * XXX I probably should move these settings to the dev tree
2425          */
2426         if (dev->phy_address == 0xffffffff && dev->phy_map == 0xffffffff) {
2427                 emac_reset(dev);
2428
2429                 /* PHY-less configuration.
2430                  * XXX I probably should move these settings to the dev tree
2431                  */
2432                 dev->phy.address = -1;
2433                 dev->phy.features = SUPPORTED_MII;
2434                 if (emac_phy_supports_gige(dev->phy_mode))
2435                         dev->phy.features |= SUPPORTED_1000baseT_Full;
2436                 else
2437                         dev->phy.features |= SUPPORTED_100baseT_Full;
2438                 dev->phy.pause = 1;
2439
2440                 return 0;
2441         }
2442
2443         mutex_lock(&emac_phy_map_lock);
2444         phy_map = dev->phy_map | busy_phy_map;
2445
2446         DBG(dev, "PHY maps %08x %08x" NL, dev->phy_map, busy_phy_map);
2447
2448         dev->phy.mdio_read = emac_mdio_read;
2449         dev->phy.mdio_write = emac_mdio_write;
2450
2451         /* Enable internal clock source */
2452 #ifdef CONFIG_PPC_DCR_NATIVE
2453         if (emac_has_feature(dev, EMAC_FTR_440GX_PHY_CLK_FIX))
2454                 dcri_clrset(SDR0, SDR0_MFR, 0, SDR0_MFR_ECS);
2455 #endif
2456         /* PHY clock workaround */
2457         emac_rx_clk_tx(dev);
2458
2459         /* Enable internal clock source on 440GX*/
2460 #ifdef CONFIG_PPC_DCR_NATIVE
2461         if (emac_has_feature(dev, EMAC_FTR_440GX_PHY_CLK_FIX))
2462                 dcri_clrset(SDR0, SDR0_MFR, 0, SDR0_MFR_ECS);
2463 #endif
2464         /* Configure EMAC with defaults so we can at least use MDIO
2465          * This is needed mostly for 440GX
2466          */
2467         if (emac_phy_gpcs(dev->phy.mode)) {
2468                 /* XXX
2469                  * Make GPCS PHY address equal to EMAC index.
2470                  * We probably should take into account busy_phy_map
2471                  * and/or phy_map here.
2472                  *
2473                  * Note that the busy_phy_map is currently global
2474                  * while it should probably be per-ASIC...
2475                  */
2476                 dev->phy.gpcs_address = dev->gpcs_address;
2477                 if (dev->phy.gpcs_address == 0xffffffff)
2478                         dev->phy.address = dev->cell_index;
2479         }
2480
2481         emac_configure(dev);
2482
2483         if (dev->phy_address != 0xffffffff)
2484                 phy_map = ~(1 << dev->phy_address);
2485
2486         for (i = 0; i < 0x20; phy_map >>= 1, ++i)
2487                 if (!(phy_map & 1)) {
2488                         int r;
2489                         busy_phy_map |= 1 << i;
2490
2491                         /* Quick check if there is a PHY at the address */
2492                         r = emac_mdio_read(dev->ndev, i, MII_BMCR);
2493                         if (r == 0xffff || r < 0)
2494                                 continue;
2495                         if (!emac_mii_phy_probe(&dev->phy, i))
2496                                 break;
2497                 }
2498
2499         /* Enable external clock source */
2500 #ifdef CONFIG_PPC_DCR_NATIVE
2501         if (emac_has_feature(dev, EMAC_FTR_440GX_PHY_CLK_FIX))
2502                 dcri_clrset(SDR0, SDR0_MFR, SDR0_MFR_ECS, 0);
2503 #endif
2504         mutex_unlock(&emac_phy_map_lock);
2505         if (i == 0x20) {
2506                 printk(KERN_WARNING "%s: can't find PHY!\n", np->full_name);
2507                 return -ENXIO;
2508         }
2509
2510         /* Init PHY */
2511         if (dev->phy.def->ops->init)
2512                 dev->phy.def->ops->init(&dev->phy);
2513
2514         /* Disable any PHY features not supported by the platform */
2515         dev->phy.def->features &= ~dev->phy_feat_exc;
2516         dev->phy.features &= ~dev->phy_feat_exc;
2517
2518         /* Setup initial link parameters */
2519         if (dev->phy.features & SUPPORTED_Autoneg) {
2520                 adv = dev->phy.features;
2521                 if (!emac_has_feature(dev, EMAC_FTR_NO_FLOW_CONTROL_40x))
2522                         adv |= ADVERTISED_Pause | ADVERTISED_Asym_Pause;
2523                 /* Restart autonegotiation */
2524                 dev->phy.def->ops->setup_aneg(&dev->phy, adv);
2525         } else {
2526                 u32 f = dev->phy.def->features;
2527                 int speed = SPEED_10, fd = DUPLEX_HALF;
2528
2529                 /* Select highest supported speed/duplex */
2530                 if (f & SUPPORTED_1000baseT_Full) {
2531                         speed = SPEED_1000;
2532                         fd = DUPLEX_FULL;
2533                 } else if (f & SUPPORTED_1000baseT_Half)
2534                         speed = SPEED_1000;
2535                 else if (f & SUPPORTED_100baseT_Full) {
2536                         speed = SPEED_100;
2537                         fd = DUPLEX_FULL;
2538                 } else if (f & SUPPORTED_100baseT_Half)
2539                         speed = SPEED_100;
2540                 else if (f & SUPPORTED_10baseT_Full)
2541                         fd = DUPLEX_FULL;
2542
2543                 /* Force link parameters */
2544                 dev->phy.def->ops->setup_forced(&dev->phy, speed, fd);
2545         }
2546         return 0;
2547 }
2548
2549 static int emac_init_config(struct emac_instance *dev)
2550 {
2551         struct device_node *np = dev->ofdev->dev.of_node;
2552         const void *p;
2553
2554         /* Read config from device-tree */
2555         if (emac_read_uint_prop(np, "mal-device", &dev->mal_ph, 1))
2556                 return -ENXIO;
2557         if (emac_read_uint_prop(np, "mal-tx-channel", &dev->mal_tx_chan, 1))
2558                 return -ENXIO;
2559         if (emac_read_uint_prop(np, "mal-rx-channel", &dev->mal_rx_chan, 1))
2560                 return -ENXIO;
2561         if (emac_read_uint_prop(np, "cell-index", &dev->cell_index, 1))
2562                 return -ENXIO;
2563         if (emac_read_uint_prop(np, "max-frame-size", &dev->max_mtu, 0))
2564                 dev->max_mtu = 1500;
2565         if (emac_read_uint_prop(np, "rx-fifo-size", &dev->rx_fifo_size, 0))
2566                 dev->rx_fifo_size = 2048;
2567         if (emac_read_uint_prop(np, "tx-fifo-size", &dev->tx_fifo_size, 0))
2568                 dev->tx_fifo_size = 2048;
2569         if (emac_read_uint_prop(np, "rx-fifo-size-gige", &dev->rx_fifo_size_gige, 0))
2570                 dev->rx_fifo_size_gige = dev->rx_fifo_size;
2571         if (emac_read_uint_prop(np, "tx-fifo-size-gige", &dev->tx_fifo_size_gige, 0))
2572                 dev->tx_fifo_size_gige = dev->tx_fifo_size;
2573         if (emac_read_uint_prop(np, "phy-address", &dev->phy_address, 0))
2574                 dev->phy_address = 0xffffffff;
2575         if (emac_read_uint_prop(np, "phy-map", &dev->phy_map, 0))
2576                 dev->phy_map = 0xffffffff;
2577         if (emac_read_uint_prop(np, "gpcs-address", &dev->gpcs_address, 0))
2578                 dev->gpcs_address = 0xffffffff;
2579         if (emac_read_uint_prop(np->parent, "clock-frequency", &dev->opb_bus_freq, 1))
2580                 return -ENXIO;
2581         if (emac_read_uint_prop(np, "tah-device", &dev->tah_ph, 0))
2582                 dev->tah_ph = 0;
2583         if (emac_read_uint_prop(np, "tah-channel", &dev->tah_port, 0))
2584                 dev->tah_port = 0;
2585         if (emac_read_uint_prop(np, "mdio-device", &dev->mdio_ph, 0))
2586                 dev->mdio_ph = 0;
2587         if (emac_read_uint_prop(np, "zmii-device", &dev->zmii_ph, 0))
2588                 dev->zmii_ph = 0;
2589         if (emac_read_uint_prop(np, "zmii-channel", &dev->zmii_port, 0))
2590                 dev->zmii_port = 0xffffffff;
2591         if (emac_read_uint_prop(np, "rgmii-device", &dev->rgmii_ph, 0))
2592                 dev->rgmii_ph = 0;
2593         if (emac_read_uint_prop(np, "rgmii-channel", &dev->rgmii_port, 0))
2594                 dev->rgmii_port = 0xffffffff;
2595         if (emac_read_uint_prop(np, "fifo-entry-size", &dev->fifo_entry_size, 0))
2596                 dev->fifo_entry_size = 16;
2597         if (emac_read_uint_prop(np, "mal-burst-size", &dev->mal_burst_size, 0))
2598                 dev->mal_burst_size = 256;
2599
2600         /* PHY mode needs some decoding */
2601         dev->phy_mode = of_get_phy_mode(np);
2602         if (dev->phy_mode < 0)
2603                 dev->phy_mode = PHY_MODE_NA;
2604
2605         /* Check EMAC version */
2606         if (of_device_is_compatible(np, "ibm,emac4sync")) {
2607                 dev->features |= (EMAC_FTR_EMAC4 | EMAC_FTR_EMAC4SYNC);
2608                 if (of_device_is_compatible(np, "ibm,emac-460ex") ||
2609                     of_device_is_compatible(np, "ibm,emac-460gt"))
2610                         dev->features |= EMAC_FTR_460EX_PHY_CLK_FIX;
2611                 if (of_device_is_compatible(np, "ibm,emac-405ex") ||
2612                     of_device_is_compatible(np, "ibm,emac-405exr"))
2613                         dev->features |= EMAC_FTR_440EP_PHY_CLK_FIX;
2614                 if (of_device_is_compatible(np, "ibm,emac-apm821xx")) {
2615                         dev->features |= (EMAC_APM821XX_REQ_JUMBO_FRAME_SIZE |
2616                                           EMAC_FTR_APM821XX_NO_HALF_DUPLEX |
2617                                           EMAC_FTR_460EX_PHY_CLK_FIX);
2618                 }
2619         } else if (of_device_is_compatible(np, "ibm,emac4")) {
2620                 dev->features |= EMAC_FTR_EMAC4;
2621                 if (of_device_is_compatible(np, "ibm,emac-440gx"))
2622                         dev->features |= EMAC_FTR_440GX_PHY_CLK_FIX;
2623         } else {
2624                 if (of_device_is_compatible(np, "ibm,emac-440ep") ||
2625                     of_device_is_compatible(np, "ibm,emac-440gr"))
2626                         dev->features |= EMAC_FTR_440EP_PHY_CLK_FIX;
2627                 if (of_device_is_compatible(np, "ibm,emac-405ez")) {
2628 #ifdef CONFIG_IBM_EMAC_NO_FLOW_CTRL
2629                         dev->features |= EMAC_FTR_NO_FLOW_CONTROL_40x;
2630 #else
2631                         printk(KERN_ERR "%s: Flow control not disabled!\n",
2632                                         np->full_name);
2633                         return -ENXIO;
2634 #endif
2635                 }
2636
2637         }
2638
2639         /* Fixup some feature bits based on the device tree */
2640         if (of_get_property(np, "has-inverted-stacr-oc", NULL))
2641                 dev->features |= EMAC_FTR_STACR_OC_INVERT;
2642         if (of_get_property(np, "has-new-stacr-staopc", NULL))
2643                 dev->features |= EMAC_FTR_HAS_NEW_STACR;
2644
2645         /* CAB lacks the appropriate properties */
2646         if (of_device_is_compatible(np, "ibm,emac-axon"))
2647                 dev->features |= EMAC_FTR_HAS_NEW_STACR |
2648                         EMAC_FTR_STACR_OC_INVERT;
2649
2650         /* Enable TAH/ZMII/RGMII features as found */
2651         if (dev->tah_ph != 0) {
2652 #ifdef CONFIG_IBM_EMAC_TAH
2653                 dev->features |= EMAC_FTR_HAS_TAH;
2654 #else
2655                 printk(KERN_ERR "%s: TAH support not enabled !\n",
2656                        np->full_name);
2657                 return -ENXIO;
2658 #endif
2659         }
2660
2661         if (dev->zmii_ph != 0) {
2662 #ifdef CONFIG_IBM_EMAC_ZMII
2663                 dev->features |= EMAC_FTR_HAS_ZMII;
2664 #else
2665                 printk(KERN_ERR "%s: ZMII support not enabled !\n",
2666                        np->full_name);
2667                 return -ENXIO;
2668 #endif
2669         }
2670
2671         if (dev->rgmii_ph != 0) {
2672 #ifdef CONFIG_IBM_EMAC_RGMII
2673                 dev->features |= EMAC_FTR_HAS_RGMII;
2674 #else
2675                 printk(KERN_ERR "%s: RGMII support not enabled !\n",
2676                        np->full_name);
2677                 return -ENXIO;
2678 #endif
2679         }
2680
2681         /* Read MAC-address */
2682         p = of_get_property(np, "local-mac-address", NULL);
2683         if (p == NULL) {
2684                 printk(KERN_ERR "%s: Can't find local-mac-address property\n",
2685                        np->full_name);
2686                 return -ENXIO;
2687         }
2688         memcpy(dev->ndev->dev_addr, p, ETH_ALEN);
2689
2690         /* IAHT and GAHT filter parameterization */
2691         if (emac_has_feature(dev, EMAC_FTR_EMAC4SYNC)) {
2692                 dev->xaht_slots_shift = EMAC4SYNC_XAHT_SLOTS_SHIFT;
2693                 dev->xaht_width_shift = EMAC4SYNC_XAHT_WIDTH_SHIFT;
2694         } else {
2695                 dev->xaht_slots_shift = EMAC4_XAHT_SLOTS_SHIFT;
2696                 dev->xaht_width_shift = EMAC4_XAHT_WIDTH_SHIFT;
2697         }
2698
2699         DBG(dev, "features     : 0x%08x / 0x%08x\n", dev->features, EMAC_FTRS_POSSIBLE);
2700         DBG(dev, "tx_fifo_size : %d (%d gige)\n", dev->tx_fifo_size, dev->tx_fifo_size_gige);
2701         DBG(dev, "rx_fifo_size : %d (%d gige)\n", dev->rx_fifo_size, dev->rx_fifo_size_gige);
2702         DBG(dev, "max_mtu      : %d\n", dev->max_mtu);
2703         DBG(dev, "OPB freq     : %d\n", dev->opb_bus_freq);
2704
2705         return 0;
2706 }
2707
2708 static const struct net_device_ops emac_netdev_ops = {
2709         .ndo_open               = emac_open,
2710         .ndo_stop               = emac_close,
2711         .ndo_get_stats          = emac_stats,
2712         .ndo_set_rx_mode        = emac_set_multicast_list,
2713         .ndo_do_ioctl           = emac_ioctl,
2714         .ndo_tx_timeout         = emac_tx_timeout,
2715         .ndo_validate_addr      = eth_validate_addr,
2716         .ndo_set_mac_address    = emac_set_mac_address,
2717         .ndo_start_xmit         = emac_start_xmit,
2718         .ndo_change_mtu         = eth_change_mtu,
2719 };
2720
2721 static const struct net_device_ops emac_gige_netdev_ops = {
2722         .ndo_open               = emac_open,
2723         .ndo_stop               = emac_close,
2724         .ndo_get_stats          = emac_stats,
2725         .ndo_set_rx_mode        = emac_set_multicast_list,
2726         .ndo_do_ioctl           = emac_ioctl,
2727         .ndo_tx_timeout         = emac_tx_timeout,
2728         .ndo_validate_addr      = eth_validate_addr,
2729         .ndo_set_mac_address    = emac_set_mac_address,
2730         .ndo_start_xmit         = emac_start_xmit_sg,
2731         .ndo_change_mtu         = emac_change_mtu,
2732 };
2733
2734 static int emac_probe(struct platform_device *ofdev)
2735 {
2736         struct net_device *ndev;
2737         struct emac_instance *dev;
2738         struct device_node *np = ofdev->dev.of_node;
2739         struct device_node **blist = NULL;
2740         int err, i;
2741
2742         /* Skip unused/unwired EMACS.  We leave the check for an unused
2743          * property here for now, but new flat device trees should set a
2744          * status property to "disabled" instead.
2745          */
2746         if (of_get_property(np, "unused", NULL) || !of_device_is_available(np))
2747                 return -ENODEV;
2748
2749         /* Find ourselves in the bootlist if we are there */
2750         for (i = 0; i < EMAC_BOOT_LIST_SIZE; i++)
2751                 if (emac_boot_list[i] == np)
2752                         blist = &emac_boot_list[i];
2753
2754         /* Allocate our net_device structure */
2755         err = -ENOMEM;
2756         ndev = alloc_etherdev(sizeof(struct emac_instance));
2757         if (!ndev)
2758                 goto err_gone;
2759
2760         dev = netdev_priv(ndev);
2761         dev->ndev = ndev;
2762         dev->ofdev = ofdev;
2763         dev->blist = blist;
2764         SET_NETDEV_DEV(ndev, &ofdev->dev);
2765
2766         /* Initialize some embedded data structures */
2767         mutex_init(&dev->mdio_lock);
2768         mutex_init(&dev->link_lock);
2769         spin_lock_init(&dev->lock);
2770         INIT_WORK(&dev->reset_work, emac_reset_work);
2771
2772         /* Init various config data based on device-tree */
2773         err = emac_init_config(dev);
2774         if (err != 0)
2775                 goto err_free;
2776
2777         /* Get interrupts. EMAC irq is mandatory, WOL irq is optional */
2778         dev->emac_irq = irq_of_parse_and_map(np, 0);
2779         dev->wol_irq = irq_of_parse_and_map(np, 1);
2780         if (dev->emac_irq == NO_IRQ) {
2781                 printk(KERN_ERR "%s: Can't map main interrupt\n", np->full_name);
2782                 goto err_free;
2783         }
2784         ndev->irq = dev->emac_irq;
2785
2786         /* Map EMAC regs */
2787         if (of_address_to_resource(np, 0, &dev->rsrc_regs)) {
2788                 printk(KERN_ERR "%s: Can't get registers address\n",
2789                        np->full_name);
2790                 goto err_irq_unmap;
2791         }
2792         // TODO : request_mem_region
2793         dev->emacp = ioremap(dev->rsrc_regs.start,
2794                              resource_size(&dev->rsrc_regs));
2795         if (dev->emacp == NULL) {
2796                 printk(KERN_ERR "%s: Can't map device registers!\n",
2797                        np->full_name);
2798                 err = -ENOMEM;
2799                 goto err_irq_unmap;
2800         }
2801
2802         /* Wait for dependent devices */
2803         err = emac_wait_deps(dev);
2804         if (err) {
2805                 printk(KERN_ERR
2806                        "%s: Timeout waiting for dependent devices\n",
2807                        np->full_name);
2808                 /*  display more info about what's missing ? */
2809                 goto err_reg_unmap;
2810         }
2811         dev->mal = platform_get_drvdata(dev->mal_dev);
2812         if (dev->mdio_dev != NULL)
2813                 dev->mdio_instance = platform_get_drvdata(dev->mdio_dev);
2814
2815         /* Register with MAL */
2816         dev->commac.ops = &emac_commac_ops;
2817         dev->commac.dev = dev;
2818         dev->commac.tx_chan_mask = MAL_CHAN_MASK(dev->mal_tx_chan);
2819         dev->commac.rx_chan_mask = MAL_CHAN_MASK(dev->mal_rx_chan);
2820         err = mal_register_commac(dev->mal, &dev->commac);
2821         if (err) {
2822                 printk(KERN_ERR "%s: failed to register with mal %s!\n",
2823                        np->full_name, dev->mal_dev->dev.of_node->full_name);
2824                 goto err_rel_deps;
2825         }
2826         dev->rx_skb_size = emac_rx_skb_size(ndev->mtu);
2827         dev->rx_sync_size = emac_rx_sync_size(ndev->mtu);
2828
2829         /* Get pointers to BD rings */
2830         dev->tx_desc =
2831             dev->mal->bd_virt + mal_tx_bd_offset(dev->mal, dev->mal_tx_chan);
2832         dev->rx_desc =
2833             dev->mal->bd_virt + mal_rx_bd_offset(dev->mal, dev->mal_rx_chan);
2834
2835         DBG(dev, "tx_desc %p" NL, dev->tx_desc);
2836         DBG(dev, "rx_desc %p" NL, dev->rx_desc);
2837
2838         /* Clean rings */
2839         memset(dev->tx_desc, 0, NUM_TX_BUFF * sizeof(struct mal_descriptor));
2840         memset(dev->rx_desc, 0, NUM_RX_BUFF * sizeof(struct mal_descriptor));
2841         memset(dev->tx_skb, 0, NUM_TX_BUFF * sizeof(struct sk_buff *));
2842         memset(dev->rx_skb, 0, NUM_RX_BUFF * sizeof(struct sk_buff *));
2843
2844         /* Attach to ZMII, if needed */
2845         if (emac_has_feature(dev, EMAC_FTR_HAS_ZMII) &&
2846             (err = zmii_attach(dev->zmii_dev, dev->zmii_port, &dev->phy_mode)) != 0)
2847                 goto err_unreg_commac;
2848
2849         /* Attach to RGMII, if needed */
2850         if (emac_has_feature(dev, EMAC_FTR_HAS_RGMII) &&
2851             (err = rgmii_attach(dev->rgmii_dev, dev->rgmii_port, dev->phy_mode)) != 0)
2852                 goto err_detach_zmii;
2853
2854         /* Attach to TAH, if needed */
2855         if (emac_has_feature(dev, EMAC_FTR_HAS_TAH) &&
2856             (err = tah_attach(dev->tah_dev, dev->tah_port)) != 0)
2857                 goto err_detach_rgmii;
2858
2859         /* Set some link defaults before we can find out real parameters */
2860         dev->phy.speed = SPEED_100;
2861         dev->phy.duplex = DUPLEX_FULL;
2862         dev->phy.autoneg = AUTONEG_DISABLE;
2863         dev->phy.pause = dev->phy.asym_pause = 0;
2864         dev->stop_timeout = STOP_TIMEOUT_100;
2865         INIT_DELAYED_WORK(&dev->link_work, emac_link_timer);
2866
2867         /* Some SoCs like APM821xx does not support Half Duplex mode. */
2868         if (emac_has_feature(dev, EMAC_FTR_APM821XX_NO_HALF_DUPLEX)) {
2869                 dev->phy_feat_exc = (SUPPORTED_1000baseT_Half |
2870                                      SUPPORTED_100baseT_Half |
2871                                      SUPPORTED_10baseT_Half);
2872         }
2873
2874         /* Find PHY if any */
2875         err = emac_init_phy(dev);
2876         if (err != 0)
2877                 goto err_detach_tah;
2878
2879         if (dev->tah_dev) {
2880                 ndev->hw_features = NETIF_F_IP_CSUM | NETIF_F_SG;
2881                 ndev->features |= ndev->hw_features | NETIF_F_RXCSUM;
2882         }
2883         ndev->watchdog_timeo = 5 * HZ;
2884         if (emac_phy_supports_gige(dev->phy_mode)) {
2885                 ndev->netdev_ops = &emac_gige_netdev_ops;
2886                 dev->commac.ops = &emac_commac_sg_ops;
2887         } else
2888                 ndev->netdev_ops = &emac_netdev_ops;
2889         ndev->ethtool_ops = &emac_ethtool_ops;
2890
2891         netif_carrier_off(ndev);
2892
2893         err = register_netdev(ndev);
2894         if (err) {
2895                 printk(KERN_ERR "%s: failed to register net device (%d)!\n",
2896                        np->full_name, err);
2897                 goto err_detach_tah;
2898         }
2899
2900         /* Set our drvdata last as we don't want them visible until we are
2901          * fully initialized
2902          */
2903         wmb();
2904         platform_set_drvdata(ofdev, dev);
2905
2906         /* There's a new kid in town ! Let's tell everybody */
2907         wake_up_all(&emac_probe_wait);
2908
2909
2910         printk(KERN_INFO "%s: EMAC-%d %s, MAC %pM\n",
2911                ndev->name, dev->cell_index, np->full_name, ndev->dev_addr);
2912
2913         if (dev->phy_mode == PHY_MODE_SGMII)
2914                 printk(KERN_NOTICE "%s: in SGMII mode\n", ndev->name);
2915
2916         if (dev->phy.address >= 0)
2917                 printk("%s: found %s PHY (0x%02x)\n", ndev->name,
2918                        dev->phy.def->name, dev->phy.address);
2919
2920         emac_dbg_register(dev);
2921
2922         /* Life is good */
2923         return 0;
2924
2925         /* I have a bad feeling about this ... */
2926
2927  err_detach_tah:
2928         if (emac_has_feature(dev, EMAC_FTR_HAS_TAH))
2929                 tah_detach(dev->tah_dev, dev->tah_port);
2930  err_detach_rgmii:
2931         if (emac_has_feature(dev, EMAC_FTR_HAS_RGMII))
2932                 rgmii_detach(dev->rgmii_dev, dev->rgmii_port);
2933  err_detach_zmii:
2934         if (emac_has_feature(dev, EMAC_FTR_HAS_ZMII))
2935                 zmii_detach(dev->zmii_dev, dev->zmii_port);
2936  err_unreg_commac:
2937         mal_unregister_commac(dev->mal, &dev->commac);
2938  err_rel_deps:
2939         emac_put_deps(dev);
2940  err_reg_unmap:
2941         iounmap(dev->emacp);
2942  err_irq_unmap:
2943         if (dev->wol_irq != NO_IRQ)
2944                 irq_dispose_mapping(dev->wol_irq);
2945         if (dev->emac_irq != NO_IRQ)
2946                 irq_dispose_mapping(dev->emac_irq);
2947  err_free:
2948         free_netdev(ndev);
2949  err_gone:
2950         /* if we were on the bootlist, remove us as we won't show up and
2951          * wake up all waiters to notify them in case they were waiting
2952          * on us
2953          */
2954         if (blist) {
2955                 *blist = NULL;
2956                 wake_up_all(&emac_probe_wait);
2957         }
2958         return err;
2959 }
2960
2961 static int emac_remove(struct platform_device *ofdev)
2962 {
2963         struct emac_instance *dev = platform_get_drvdata(ofdev);
2964
2965         DBG(dev, "remove" NL);
2966
2967         unregister_netdev(dev->ndev);
2968
2969         cancel_work_sync(&dev->reset_work);
2970
2971         if (emac_has_feature(dev, EMAC_FTR_HAS_TAH))
2972                 tah_detach(dev->tah_dev, dev->tah_port);
2973         if (emac_has_feature(dev, EMAC_FTR_HAS_RGMII))
2974                 rgmii_detach(dev->rgmii_dev, dev->rgmii_port);
2975         if (emac_has_feature(dev, EMAC_FTR_HAS_ZMII))
2976                 zmii_detach(dev->zmii_dev, dev->zmii_port);
2977
2978         busy_phy_map &= ~(1 << dev->phy.address);
2979         DBG(dev, "busy_phy_map now %#x" NL, busy_phy_map);
2980
2981         mal_unregister_commac(dev->mal, &dev->commac);
2982         emac_put_deps(dev);
2983
2984         emac_dbg_unregister(dev);
2985         iounmap(dev->emacp);
2986
2987         if (dev->wol_irq != NO_IRQ)
2988                 irq_dispose_mapping(dev->wol_irq);
2989         if (dev->emac_irq != NO_IRQ)
2990                 irq_dispose_mapping(dev->emac_irq);
2991
2992         free_netdev(dev->ndev);
2993
2994         return 0;
2995 }
2996
2997 /* XXX Features in here should be replaced by properties... */
2998 static const struct of_device_id emac_match[] =
2999 {
3000         {
3001                 .type           = "network",
3002                 .compatible     = "ibm,emac",
3003         },
3004         {
3005                 .type           = "network",
3006                 .compatible     = "ibm,emac4",
3007         },
3008         {
3009                 .type           = "network",
3010                 .compatible     = "ibm,emac4sync",
3011         },
3012         {},
3013 };
3014 MODULE_DEVICE_TABLE(of, emac_match);
3015
3016 static struct platform_driver emac_driver = {
3017         .driver = {
3018                 .name = "emac",
3019                 .of_match_table = emac_match,
3020         },
3021         .probe = emac_probe,
3022         .remove = emac_remove,
3023 };
3024
3025 static void __init emac_make_bootlist(void)
3026 {
3027         struct device_node *np = NULL;
3028         int j, max, i = 0;
3029         int cell_indices[EMAC_BOOT_LIST_SIZE];
3030
3031         /* Collect EMACs */
3032         while((np = of_find_all_nodes(np)) != NULL) {
3033                 const u32 *idx;
3034
3035                 if (of_match_node(emac_match, np) == NULL)
3036                         continue;
3037                 if (of_get_property(np, "unused", NULL))
3038                         continue;
3039                 idx = of_get_property(np, "cell-index", NULL);
3040                 if (idx == NULL)
3041                         continue;
3042                 cell_indices[i] = *idx;
3043                 emac_boot_list[i++] = of_node_get(np);
3044                 if (i >= EMAC_BOOT_LIST_SIZE) {
3045                         of_node_put(np);
3046                         break;
3047                 }
3048         }
3049         max = i;
3050
3051         /* Bubble sort them (doh, what a creative algorithm :-) */
3052         for (i = 0; max > 1 && (i < (max - 1)); i++)
3053                 for (j = i; j < max; j++) {
3054                         if (cell_indices[i] > cell_indices[j]) {
3055                                 swap(emac_boot_list[i], emac_boot_list[j]);
3056                                 swap(cell_indices[i], cell_indices[j]);
3057                         }
3058                 }
3059 }
3060
3061 static int __init emac_init(void)
3062 {
3063         int rc;
3064
3065         printk(KERN_INFO DRV_DESC ", version " DRV_VERSION "\n");
3066
3067         /* Init debug stuff */
3068         emac_init_debug();
3069
3070         /* Build EMAC boot list */
3071         emac_make_bootlist();
3072
3073         /* Init submodules */
3074         rc = mal_init();
3075         if (rc)
3076                 goto err;
3077         rc = zmii_init();
3078         if (rc)
3079                 goto err_mal;
3080         rc = rgmii_init();
3081         if (rc)
3082                 goto err_zmii;
3083         rc = tah_init();
3084         if (rc)
3085                 goto err_rgmii;
3086         rc = platform_driver_register(&emac_driver);
3087         if (rc)
3088                 goto err_tah;
3089
3090         return 0;
3091
3092  err_tah:
3093         tah_exit();
3094  err_rgmii:
3095         rgmii_exit();
3096  err_zmii:
3097         zmii_exit();
3098  err_mal:
3099         mal_exit();
3100  err:
3101         return rc;
3102 }
3103
3104 static void __exit emac_exit(void)
3105 {
3106         int i;
3107
3108         platform_driver_unregister(&emac_driver);
3109
3110         tah_exit();
3111         rgmii_exit();
3112         zmii_exit();
3113         mal_exit();
3114         emac_fini_debug();
3115
3116         /* Destroy EMAC boot list */
3117         for (i = 0; i < EMAC_BOOT_LIST_SIZE; i++)
3118                 of_node_put(emac_boot_list[i]);
3119 }
3120
3121 module_init(emac_init);
3122 module_exit(emac_exit);