1 /*-----------------------------------------------------------------------------+
3 * This source code has been made available to you by IBM on an AS-IS
4 * basis. Anyone receiving this source is licensed under IBM
5 * copyrights to use it in any way he or she deems fit, including
6 * copying it, modifying it, compiling it, and redistributing it either
7 * with or without modifications. No license under IBM patents or
8 * patent applications is to be implied by the copyright license.
10 * Any user of this software should understand that IBM cannot provide
11 * technical support for this software and will not be responsible for
12 * any consequences resulting from the use of this software.
14 * Any person who transfers this source code or any derivative work
15 * must include the IBM copyright notice, this paragraph, and the
16 * preceding two paragraphs in the transferred software.
18 * COPYRIGHT I B M CORPORATION 1995
19 * LICENSED MATERIAL - PROGRAM PROPERTY OF I B M
20 *-----------------------------------------------------------------------------*/
21 /*-----------------------------------------------------------------------------+
23 * File Name: enetemac.c
25 * Function: Device driver for the ethernet EMAC3 macro on the 405GP.
31 * Date Description of Change BY
32 * --------- --------------------- ---
33 * 05-May-99 Created MKW
34 * 27-Jun-99 Clean up JWB
35 * 16-Jul-99 Added MAL error recovery and better IP packet handling MKW
36 * 29-Jul-99 Added Full duplex support MKW
37 * 06-Aug-99 Changed names for Mal CR reg MKW
38 * 23-Aug-99 Turned off SYE when running at 10Mbs MKW
39 * 24-Aug-99 Marked descriptor empty after call_xlc MKW
40 * 07-Sep-99 Set MAL RX buffer size reg to ENET_MAX_MTU_ALIGNED / 16 MCG
41 * to avoid chaining maximum sized packets. Push starting
42 * RX descriptor address up to the next cache line boundary.
43 * 16-Jan-00 Added support for booting with IP of 0x0 MKW
44 * 15-Mar-00 Updated enetInit() to enable broadcast addresses in the
45 * EMAC_RXM register. JWB
46 * 12-Mar-01 anne-sophie.harnois@nextream.fr
47 * - Variables are compatible with those already defined in
49 * - Receive buffer descriptor ring is used to send buffers
51 * - Info print about send/received/handled packet number if
52 * INFO_405_ENET is set
53 * 17-Apr-01 stefan.roese@esd-electronics.com
54 * - MAL reset in "eth_halt" included
55 * - Enet speed and duplex output now in one line
56 * 08-May-01 stefan.roese@esd-electronics.com
57 * - MAL error handling added (eth_init called again)
58 * 13-Nov-01 stefan.roese@esd-electronics.com
59 * - Set IST bit in EMAC_M1 reg upon 100MBit or full duplex
60 * 04-Jan-02 stefan.roese@esd-electronics.com
61 * - Wait for PHY auto negotiation to complete added
62 * 06-Feb-02 stefan.roese@esd-electronics.com
63 * - Bug fixed in waiting for auto negotiation to complete
64 * 26-Feb-02 stefan.roese@esd-electronics.com
65 * - rx and tx buffer descriptors now allocated (no fixed address
67 * 17-Jun-02 stefan.roese@esd-electronics.com
68 * - MAL error debug printf 'M' removed (rx de interrupt may
69 * occur upon many incoming packets with only 4 rx buffers).
70 *-----------------------------------------------------------------------------*
71 * 17-Nov-03 travis.sawyer@sandburst.com
72 * - ported from 405gp_enet.c to utilized upto 4 EMAC ports
73 * in the 440GX. This port should work with the 440GP
75 * 15-Aug-05 sr@denx.de
76 * - merged 405gp_enet.c and 440gx_enet.c to generic 4xx_enet.c
77 now handling all 4xx cpu's.
78 *-----------------------------------------------------------------------------*/
83 #include <asm/processor.h>
86 #include <ppc4xx_enet.h>
93 * Only compile for platform with AMCC EMAC ethernet controller and
94 * network support enabled.
95 * Remark: CONFIG_405 describes Xilinx PPC405 FPGA without EMAC controller!
97 #if (CONFIG_COMMANDS & CFG_CMD_NET) && !defined(CONFIG_405) && !defined(CONFIG_IOP480)
99 #if !(defined(CONFIG_MII) || (CONFIG_COMMANDS & CFG_CMD_MII))
100 #error "CONFIG_MII has to be defined!"
103 #define EMAC_RESET_TIMEOUT 1000 /* 1000 ms reset timeout */
104 #define PHY_AUTONEGOTIATE_TIMEOUT 4000 /* 4000 ms autonegotiate timeout */
106 /* Ethernet Transmit and Receive Buffers */
108 * In the same way ENET_MAX_MTU and ENET_MAX_MTU_ALIGNED are set from
109 * PKTSIZE and PKTSIZE_ALIGN (include/net.h)
111 #define ENET_MAX_MTU PKTSIZE
112 #define ENET_MAX_MTU_ALIGNED PKTSIZE_ALIGN
114 /* define the number of channels implemented */
115 #define EMAC_RXCHL EMAC_NUM_DEV
116 #define EMAC_TXCHL EMAC_NUM_DEV
118 /*-----------------------------------------------------------------------------+
119 * Defines for MAL/EMAC interrupt conditions as reported in the UIC (Universal
120 * Interrupt Controller).
121 *-----------------------------------------------------------------------------*/
122 #define MAL_UIC_ERR ( UIC_MAL_SERR | UIC_MAL_TXDE | UIC_MAL_RXDE)
123 #define MAL_UIC_DEF (UIC_MAL_RXEOB | MAL_UIC_ERR)
124 #define EMAC_UIC_DEF UIC_ENET
125 #define EMAC_UIC_DEF1 UIC_ENET1
126 #define SEL_UIC_DEF(p) (p ? UIC_ENET1 : UIC_ENET )
130 #define BI_PHYMODE_NONE 0
131 #define BI_PHYMODE_ZMII 1
132 #define BI_PHYMODE_RGMII 2
135 /*-----------------------------------------------------------------------------+
136 * Global variables. TX and RX descriptors and buffers.
137 *-----------------------------------------------------------------------------*/
139 static uint32_t mal_ier;
141 #if !defined(CONFIG_NET_MULTI)
142 struct eth_device *emac0_dev = NULL;
146 /*-----------------------------------------------------------------------------+
147 * Prototypes and externals.
148 *-----------------------------------------------------------------------------*/
149 static void enet_rcv (struct eth_device *dev, unsigned long malisr);
151 int enetInt (struct eth_device *dev);
152 static void mal_err (struct eth_device *dev, unsigned long isr,
153 unsigned long uic, unsigned long maldef,
154 unsigned long mal_errr);
155 static void emac_err (struct eth_device *dev, unsigned long isr);
158 /*-----------------------------------------------------------------------------+
160 | Disable MAL channel, and EMACn
161 +-----------------------------------------------------------------------------*/
162 static void ppc_4xx_eth_halt (struct eth_device *dev)
164 EMAC_4XX_HW_PST hw_p = dev->priv;
165 uint32_t failsafe = 10000;
167 out32 (EMAC_IER + hw_p->hw_addr, 0x00000000); /* disable emac interrupts */
169 /* 1st reset MAL channel */
170 /* Note: writing a 0 to a channel has no effect */
171 #if defined(CONFIG_405EP) || defined(CONFIG_440EP) || defined(CONFIG_440GR)
172 mtdcr (maltxcarr, (MAL_CR_MMSR >> (hw_p->devnum * 2)));
174 mtdcr (maltxcarr, (MAL_CR_MMSR >> hw_p->devnum));
176 mtdcr (malrxcarr, (MAL_CR_MMSR >> hw_p->devnum));
179 while (mfdcr (malrxcasr) & (MAL_CR_MMSR >> hw_p->devnum)) {
180 udelay (1000); /* Delay 1 MS so as not to hammer the register */
187 out32 (EMAC_M0 + hw_p->hw_addr, EMAC_M0_SRST);
189 #ifndef CONFIG_NETCONSOLE
190 hw_p->print_speed = 1; /* print speed message again next time */
196 extern int phy_setup_aneg (unsigned char addr);
197 extern int miiphy_reset (unsigned char addr);
199 #if defined (CONFIG_440GX)
200 int ppc_4xx_eth_setup_bridge(int devnum, bd_t * bis)
203 unsigned long zmiifer;
204 unsigned long rmiifer;
206 mfsdr(sdr_pfc1, pfc1);
207 pfc1 = SDR0_PFC1_EPS_DECODE(pfc1);
214 zmiifer |= ZMII_FER_RMII << ZMII_FER_V(0);
215 zmiifer |= ZMII_FER_RMII << ZMII_FER_V(1);
216 zmiifer |= ZMII_FER_RMII << ZMII_FER_V(2);
217 zmiifer |= ZMII_FER_RMII << ZMII_FER_V(3);
218 bis->bi_phymode[0] = BI_PHYMODE_ZMII;
219 bis->bi_phymode[1] = BI_PHYMODE_ZMII;
220 bis->bi_phymode[2] = BI_PHYMODE_ZMII;
221 bis->bi_phymode[3] = BI_PHYMODE_ZMII;
224 zmiifer = ZMII_FER_SMII << ZMII_FER_V(0);
225 zmiifer = ZMII_FER_SMII << ZMII_FER_V(1);
226 zmiifer = ZMII_FER_SMII << ZMII_FER_V(2);
227 zmiifer = ZMII_FER_SMII << ZMII_FER_V(3);
228 bis->bi_phymode[0] = BI_PHYMODE_ZMII;
229 bis->bi_phymode[1] = BI_PHYMODE_ZMII;
230 bis->bi_phymode[2] = BI_PHYMODE_ZMII;
231 bis->bi_phymode[3] = BI_PHYMODE_ZMII;
234 zmiifer |= ZMII_FER_RMII << ZMII_FER_V(0);
235 rmiifer |= RGMII_FER_RGMII << RGMII_FER_V(2);
236 bis->bi_phymode[0] = BI_PHYMODE_ZMII;
237 bis->bi_phymode[1] = BI_PHYMODE_NONE;
238 bis->bi_phymode[2] = BI_PHYMODE_RGMII;
239 bis->bi_phymode[3] = BI_PHYMODE_NONE;
242 zmiifer |= ZMII_FER_SMII << ZMII_FER_V(0);
243 zmiifer |= ZMII_FER_SMII << ZMII_FER_V(1);
244 rmiifer |= RGMII_FER_RGMII << RGMII_FER_V (2);
245 rmiifer |= RGMII_FER_RGMII << RGMII_FER_V (3);
246 bis->bi_phymode[0] = BI_PHYMODE_ZMII;
247 bis->bi_phymode[1] = BI_PHYMODE_ZMII;
248 bis->bi_phymode[2] = BI_PHYMODE_RGMII;
249 bis->bi_phymode[3] = BI_PHYMODE_RGMII;
252 zmiifer |= ZMII_FER_SMII << ZMII_FER_V (0);
253 zmiifer |= ZMII_FER_SMII << ZMII_FER_V (1);
254 zmiifer |= ZMII_FER_SMII << ZMII_FER_V (2);
255 rmiifer |= RGMII_FER_RGMII << RGMII_FER_V(3);
256 bis->bi_phymode[0] = BI_PHYMODE_ZMII;
257 bis->bi_phymode[1] = BI_PHYMODE_ZMII;
258 bis->bi_phymode[2] = BI_PHYMODE_ZMII;
259 bis->bi_phymode[3] = BI_PHYMODE_RGMII;
262 zmiifer |= ZMII_FER_SMII << ZMII_FER_V (0);
263 zmiifer |= ZMII_FER_SMII << ZMII_FER_V (1);
264 rmiifer |= RGMII_FER_RGMII << RGMII_FER_V(2);
265 bis->bi_phymode[0] = BI_PHYMODE_ZMII;
266 bis->bi_phymode[1] = BI_PHYMODE_ZMII;
267 bis->bi_phymode[2] = BI_PHYMODE_RGMII;
271 zmiifer = ZMII_FER_MII << ZMII_FER_V(devnum);
273 bis->bi_phymode[0] = BI_PHYMODE_ZMII;
274 bis->bi_phymode[1] = BI_PHYMODE_ZMII;
275 bis->bi_phymode[2] = BI_PHYMODE_ZMII;
276 bis->bi_phymode[3] = BI_PHYMODE_ZMII;
280 /* Ensure we setup mdio for this devnum and ONLY this devnum */
281 zmiifer |= (ZMII_FER_MDI) << ZMII_FER_V(devnum);
283 out32 (ZMII_FER, zmiifer);
284 out32 (RGMII_FER, rmiifer);
291 static int ppc_4xx_eth_init (struct eth_device *dev, bd_t * bis)
294 unsigned long reg = 0;
297 unsigned long duplex;
298 unsigned long failsafe;
300 unsigned short devnum;
301 unsigned short reg_short;
302 #if defined(CONFIG_440GX)
307 EMAC_4XX_HW_PST hw_p = dev->priv;
309 /* before doing anything, figure out if we have a MAC address */
311 if (memcmp (dev->enetaddr, "\0\0\0\0\0\0", 6) == 0) {
312 printf("ERROR: ethaddr not set!\n");
316 #if defined(CONFIG_440GX)
317 /* Need to get the OPB frequency so we can access the PHY */
318 get_sys_info (&sysinfo);
322 mtmsr (msr & ~(MSR_EE)); /* disable interrupts */
324 devnum = hw_p->devnum;
329 * hw_p->stats.pkts_handled <= hw_p->stats.pkts_rx <= hw_p->stats.pkts_handled+PKTBUFSRX
330 * In the most cases hw_p->stats.pkts_handled = hw_p->stats.pkts_rx, but it
331 * is possible that new packets (without relationship with
332 * current transfer) have got the time to arrived before
333 * netloop calls eth_halt
335 printf ("About preceeding transfer (eth%d):\n"
336 "- Sent packet number %d\n"
337 "- Received packet number %d\n"
338 "- Handled packet number %d\n",
341 hw_p->stats.pkts_rx, hw_p->stats.pkts_handled);
343 hw_p->stats.pkts_tx = 0;
344 hw_p->stats.pkts_rx = 0;
345 hw_p->stats.pkts_handled = 0;
348 hw_p->tx_err_index = 0; /* Transmit Error Index for tx_err_log */
349 hw_p->rx_err_index = 0; /* Receive Error Index for rx_err_log */
351 hw_p->rx_slot = 0; /* MAL Receive Slot */
352 hw_p->rx_i_index = 0; /* Receive Interrupt Queue Index */
353 hw_p->rx_u_index = 0; /* Receive User Queue Index */
355 hw_p->tx_slot = 0; /* MAL Transmit Slot */
356 hw_p->tx_i_index = 0; /* Transmit Interrupt Queue Index */
357 hw_p->tx_u_index = 0; /* Transmit User Queue Index */
359 #if defined(CONFIG_440)
361 /* NOTE: 440GX spec states that mode is mutually exclusive */
362 /* NOTE: Therefore, disable all other EMACS, since we handle */
363 /* NOTE: only one emac at a time */
368 #if defined(CONFIG_440EP) || defined(CONFIG_440GR)
369 out32 (ZMII_FER, (ZMII_FER_RMII | ZMII_FER_MDI) << ZMII_FER_V (devnum));
370 #elif defined(CONFIG_440GX)
371 ethgroup = ppc_4xx_eth_setup_bridge(devnum, bis);
372 #elif defined(CONFIG_440GP)
374 out32 (ZMII_FER, ZMII_RMII | ZMII_MDI0);
376 if ((devnum == 0) || (devnum == 1)) {
377 out32 (ZMII_FER, (ZMII_FER_SMII | ZMII_FER_MDI) << ZMII_FER_V (devnum));
379 else { /* ((devnum == 2) || (devnum == 3)) */
380 out32 (ZMII_FER, ZMII_FER_MDI << ZMII_FER_V (devnum));
381 out32 (RGMII_FER, ((RGMII_FER_RGMII << RGMII_FER_V (2)) |
382 (RGMII_FER_RGMII << RGMII_FER_V (3))));
386 out32 (ZMII_SSR, ZMII_SSR_SP << ZMII_SSR_V(devnum));
387 #endif /* defined(CONFIG_440) */
389 __asm__ volatile ("eieio");
391 /* reset emac so we have access to the phy */
393 out32 (EMAC_M0 + hw_p->hw_addr, EMAC_M0_SRST);
394 __asm__ volatile ("eieio");
397 while ((in32 (EMAC_M0 + hw_p->hw_addr) & (EMAC_M0_SRST)) && failsafe) {
402 #if defined(CONFIG_440GX)
403 /* Whack the M1 register */
405 mode_reg &= ~0x00000038;
406 if (sysinfo.freqOPB <= 50000000);
407 else if (sysinfo.freqOPB <= 66666667)
408 mode_reg |= EMAC_M1_OBCI_66;
409 else if (sysinfo.freqOPB <= 83333333)
410 mode_reg |= EMAC_M1_OBCI_83;
411 else if (sysinfo.freqOPB <= 100000000)
412 mode_reg |= EMAC_M1_OBCI_100;
414 mode_reg |= EMAC_M1_OBCI_GT100;
416 out32 (EMAC_M1 + hw_p->hw_addr, mode_reg);
417 #endif /* defined(CONFIG_440GX) */
419 /* wait for PHY to complete auto negotiation */
421 #ifndef CONFIG_CS8952_PHY
424 reg = CONFIG_PHY_ADDR;
426 #if defined (CONFIG_PHY1_ADDR)
428 reg = CONFIG_PHY1_ADDR;
431 #if defined (CONFIG_440GX)
433 reg = CONFIG_PHY2_ADDR;
436 reg = CONFIG_PHY3_ADDR;
440 reg = CONFIG_PHY_ADDR;
444 bis->bi_phynum[devnum] = reg;
446 #if defined(CONFIG_PHY_RESET)
448 * Reset the phy, only if its the first time through
449 * otherwise, just check the speeds & feeds
451 if (hw_p->first_init == 0) {
454 #if defined(CONFIG_440GX)
455 #if defined(CONFIG_CIS8201_PHY)
457 * Cicada 8201 PHY needs to have an extended register whacked
460 if ( ((devnum == 2) || (devnum ==3)) && (4 == ethgroup) ) {
461 #if defined(CONFIG_CIS8201_SHORT_ETCH)
462 miiphy_write (reg, 23, 0x1300);
464 miiphy_write (reg, 23, 0x1000);
467 * Vitesse VSC8201/Cicada CIS8201 errata:
468 * Interoperability problem with Intel 82547EI phys
469 * This work around (provided by Vitesse) changes
470 * the default timer convergence from 8ms to 12ms
472 miiphy_write (reg, 0x1f, 0x2a30);
473 miiphy_write (reg, 0x08, 0x0200);
474 miiphy_write (reg, 0x1f, 0x52b5);
475 miiphy_write (reg, 0x02, 0x0004);
476 miiphy_write (reg, 0x01, 0x0671);
477 miiphy_write (reg, 0x00, 0x8fae);
478 miiphy_write (reg, 0x1f, 0x2a30);
479 miiphy_write (reg, 0x08, 0x0000);
480 miiphy_write (reg, 0x1f, 0x0000);
481 /* end Vitesse/Cicada errata */
485 /* Start/Restart autonegotiation */
486 phy_setup_aneg (reg);
489 #endif /* defined(CONFIG_PHY_RESET) */
491 miiphy_read (reg, PHY_BMSR, ®_short);
494 * Wait if PHY is capable of autonegotiation and autonegotiation is not complete
496 if ((reg_short & PHY_BMSR_AUTN_ABLE)
497 && !(reg_short & PHY_BMSR_AUTN_COMP)) {
498 puts ("Waiting for PHY auto negotiation to complete");
500 while (!(reg_short & PHY_BMSR_AUTN_COMP)) {
504 if (i > PHY_AUTONEGOTIATE_TIMEOUT) {
505 puts (" TIMEOUT !\n");
509 if ((i++ % 1000) == 0) {
512 udelay (1000); /* 1 ms */
513 miiphy_read (reg, PHY_BMSR, ®_short);
517 udelay (500000); /* another 500 ms (results in faster booting) */
519 #endif /* #ifndef CONFIG_CS8952_PHY */
521 speed = miiphy_speed (reg);
522 duplex = miiphy_duplex (reg);
524 if (hw_p->print_speed) {
525 hw_p->print_speed = 0;
526 printf ("ENET Speed is %d Mbps - %s duplex connection\n",
527 (int) speed, (duplex == HALF) ? "HALF" : "FULL");
530 #if defined(CONFIG_440)
531 #if defined(CONFIG_440EP) || defined(CONFIG_440GR)
534 reg = (reg & ~SDR0_MFR_ZMII_MODE_MASK) | SDR0_MFR_ZMII_MODE_RMII_100M;
536 reg = (reg & ~SDR0_MFR_ZMII_MODE_MASK) | SDR0_MFR_ZMII_MODE_RMII_10M;
541 /* Set ZMII/RGMII speed according to the phy link speed */
542 reg = in32 (ZMII_SSR);
543 if ( (speed == 100) || (speed == 1000) )
544 out32 (ZMII_SSR, reg | (ZMII_SSR_SP << ZMII_SSR_V (devnum)));
546 out32 (ZMII_SSR, reg & (~(ZMII_SSR_SP << ZMII_SSR_V (devnum))));
548 if ((devnum == 2) || (devnum == 3)) {
550 reg = (RGMII_SSR_SP_1000MBPS << RGMII_SSR_V (devnum));
551 else if (speed == 100)
552 reg = (RGMII_SSR_SP_100MBPS << RGMII_SSR_V (devnum));
554 reg = (RGMII_SSR_SP_10MBPS << RGMII_SSR_V (devnum));
556 out32 (RGMII_SSR, reg);
558 #endif /* defined(CONFIG_440) */
560 /* set the Mal configuration reg */
561 #if defined(CONFIG_440GX)
562 mtdcr (malmcr, MAL_CR_PLBB | MAL_CR_OPBBL | MAL_CR_LEA |
563 MAL_CR_PLBLT_DEFAULT | MAL_CR_EOPIE | 0x00330000);
565 mtdcr (malmcr, MAL_CR_PLBB | MAL_CR_OPBBL | MAL_CR_LEA | MAL_CR_PLBLT_DEFAULT);
566 /* Errata 1.12: MAL_1 -- Disable MAL bursting */
567 if (get_pvr() == PVR_440GP_RB) {
568 mtdcr (malmcr, mfdcr(malmcr) & ~MAL_CR_PLBB);
572 /* Free "old" buffers */
573 if (hw_p->alloc_tx_buf)
574 free (hw_p->alloc_tx_buf);
575 if (hw_p->alloc_rx_buf)
576 free (hw_p->alloc_rx_buf);
579 * Malloc MAL buffer desciptors, make sure they are
580 * aligned on cache line boundary size
581 * (401/403/IOP480 = 16, 405 = 32)
582 * and doesn't cross cache block boundaries.
585 (mal_desc_t *) malloc ((sizeof (mal_desc_t) * NUM_TX_BUFF) +
586 ((2 * CFG_CACHELINE_SIZE) - 2));
587 if (NULL == hw_p->alloc_tx_buf)
589 if (((int) hw_p->alloc_tx_buf & CACHELINE_MASK) != 0) {
591 (mal_desc_t *) ((int) hw_p->alloc_tx_buf +
594 alloc_tx_buf & CACHELINE_MASK));
596 hw_p->tx = hw_p->alloc_tx_buf;
600 (mal_desc_t *) malloc ((sizeof (mal_desc_t) * NUM_RX_BUFF) +
601 ((2 * CFG_CACHELINE_SIZE) - 2));
602 if (NULL == hw_p->alloc_rx_buf) {
603 free(hw_p->alloc_tx_buf);
604 hw_p->alloc_tx_buf = NULL;
608 if (((int) hw_p->alloc_rx_buf & CACHELINE_MASK) != 0) {
610 (mal_desc_t *) ((int) hw_p->alloc_rx_buf +
613 alloc_rx_buf & CACHELINE_MASK));
615 hw_p->rx = hw_p->alloc_rx_buf;
618 for (i = 0; i < NUM_TX_BUFF; i++) {
619 hw_p->tx[i].ctrl = 0;
620 hw_p->tx[i].data_len = 0;
621 if (hw_p->first_init == 0) {
623 (char *) malloc (ENET_MAX_MTU_ALIGNED);
624 if (NULL == hw_p->txbuf_ptr) {
625 free(hw_p->alloc_rx_buf);
626 free(hw_p->alloc_tx_buf);
627 hw_p->alloc_rx_buf = NULL;
628 hw_p->alloc_tx_buf = NULL;
629 for(j = 0; j < i; j++) {
630 free(hw_p->tx[i].data_ptr);
631 hw_p->tx[i].data_ptr = NULL;
635 hw_p->tx[i].data_ptr = hw_p->txbuf_ptr;
636 if ((NUM_TX_BUFF - 1) == i)
637 hw_p->tx[i].ctrl |= MAL_TX_CTRL_WRAP;
638 hw_p->tx_run[i] = -1;
640 printf ("TX_BUFF %d @ 0x%08lx\n", i,
641 (ulong) hw_p->tx[i].data_ptr);
645 for (i = 0; i < NUM_RX_BUFF; i++) {
646 hw_p->rx[i].ctrl = 0;
647 hw_p->rx[i].data_len = 0;
648 /* rx[i].data_ptr = (char *) &rx_buff[i]; */
649 hw_p->rx[i].data_ptr = (char *) NetRxPackets[i];
650 if ((NUM_RX_BUFF - 1) == i)
651 hw_p->rx[i].ctrl |= MAL_RX_CTRL_WRAP;
652 hw_p->rx[i].ctrl |= MAL_RX_CTRL_EMPTY | MAL_RX_CTRL_INTR;
653 hw_p->rx_ready[i] = -1;
655 printf ("RX_BUFF %d @ 0x%08lx\n", i, (ulong) rx[i].data_ptr);
661 reg |= dev->enetaddr[0]; /* set high address */
663 reg |= dev->enetaddr[1];
665 out32 (EMAC_IAH + hw_p->hw_addr, reg);
668 reg |= dev->enetaddr[2]; /* set low address */
670 reg |= dev->enetaddr[3];
672 reg |= dev->enetaddr[4];
674 reg |= dev->enetaddr[5];
676 out32 (EMAC_IAL + hw_p->hw_addr, reg);
680 /* setup MAL tx & rx channel pointers */
681 #if defined (CONFIG_405EP) || defined (CONFIG_440EP) || defined (CONFIG_440GR)
682 mtdcr (maltxctp2r, hw_p->tx);
684 mtdcr (maltxctp1r, hw_p->tx);
686 #if defined(CONFIG_440)
687 mtdcr (maltxbattr, 0x0);
688 mtdcr (malrxbattr, 0x0);
690 mtdcr (malrxctp1r, hw_p->rx);
691 /* set RX buffer size */
692 mtdcr (malrcbs1, ENET_MAX_MTU_ALIGNED / 16);
694 #if defined (CONFIG_440GX)
696 /* setup MAL tx & rx channel pointers */
697 mtdcr (maltxbattr, 0x0);
698 mtdcr (malrxbattr, 0x0);
699 mtdcr (maltxctp2r, hw_p->tx);
700 mtdcr (malrxctp2r, hw_p->rx);
701 /* set RX buffer size */
702 mtdcr (malrcbs2, ENET_MAX_MTU_ALIGNED / 16);
705 /* setup MAL tx & rx channel pointers */
706 mtdcr (maltxbattr, 0x0);
707 mtdcr (maltxctp3r, hw_p->tx);
708 mtdcr (malrxbattr, 0x0);
709 mtdcr (malrxctp3r, hw_p->rx);
710 /* set RX buffer size */
711 mtdcr (malrcbs3, ENET_MAX_MTU_ALIGNED / 16);
713 #endif /* CONFIG_440GX */
716 /* setup MAL tx & rx channel pointers */
717 #if defined(CONFIG_440)
718 mtdcr (maltxbattr, 0x0);
719 mtdcr (malrxbattr, 0x0);
721 mtdcr (maltxctp0r, hw_p->tx);
722 mtdcr (malrxctp0r, hw_p->rx);
723 /* set RX buffer size */
724 mtdcr (malrcbs0, ENET_MAX_MTU_ALIGNED / 16);
728 /* Enable MAL transmit and receive channels */
729 #if defined(CONFIG_405EP) || defined(CONFIG_440EP) || defined(CONFIG_440GR)
730 mtdcr (maltxcasr, (MAL_TXRX_CASR >> (hw_p->devnum*2)));
732 mtdcr (maltxcasr, (MAL_TXRX_CASR >> hw_p->devnum));
734 mtdcr (malrxcasr, (MAL_TXRX_CASR >> hw_p->devnum));
736 /* set transmit enable & receive enable */
737 out32 (EMAC_M0 + hw_p->hw_addr, EMAC_M0_TXE | EMAC_M0_RXE);
739 /* set receive fifo to 4k and tx fifo to 2k */
740 mode_reg = in32 (EMAC_M1 + hw_p->hw_addr);
741 mode_reg |= EMAC_M1_RFS_4K | EMAC_M1_TX_FIFO_2K;
744 if (speed == _1000BASET)
745 mode_reg = mode_reg | EMAC_M1_MF_1000MBPS | EMAC_M1_IST;
746 else if (speed == _100BASET)
747 mode_reg = mode_reg | EMAC_M1_MF_100MBPS | EMAC_M1_IST;
749 mode_reg = mode_reg & ~0x00C00000; /* 10 MBPS */
751 mode_reg = mode_reg | 0x80000000 | EMAC_M1_IST;
753 out32 (EMAC_M1 + hw_p->hw_addr, mode_reg);
755 /* Enable broadcast and indvidual address */
756 /* TBS: enabling runts as some misbehaved nics will send runts */
757 out32 (EMAC_RXM + hw_p->hw_addr, EMAC_RMR_BAE | EMAC_RMR_IAE);
759 /* we probably need to set the tx mode1 reg? maybe at tx time */
761 /* set transmit request threshold register */
762 out32 (EMAC_TRTR + hw_p->hw_addr, 0x18000000); /* 256 byte threshold */
764 /* set receive low/high water mark register */
765 #if defined(CONFIG_440)
766 /* 440GP has a 64 byte burst length */
767 out32 (EMAC_RX_HI_LO_WMARK + hw_p->hw_addr, 0x80009000);
769 /* 405s have a 16 byte burst length */
770 out32 (EMAC_RX_HI_LO_WMARK + hw_p->hw_addr, 0x0f002000);
771 #endif /* defined(CONFIG_440) */
772 out32 (EMAC_TXM1 + hw_p->hw_addr, 0xf8640000);
774 /* Set fifo limit entry in tx mode 0 */
775 out32 (EMAC_TXM0 + hw_p->hw_addr, 0x00000003);
777 out32 (EMAC_I_FRAME_GAP_REG + hw_p->hw_addr, 0x00000008);
780 hw_p->emac_ier = EMAC_ISR_PTLE | EMAC_ISR_BFCS | EMAC_ISR_ORE | EMAC_ISR_IRE;
781 if (speed == _100BASET)
782 hw_p->emac_ier = hw_p->emac_ier | EMAC_ISR_SYE;
784 out32 (EMAC_ISR + hw_p->hw_addr, 0xffffffff); /* clear pending interrupts */
785 out32 (EMAC_IER + hw_p->hw_addr, hw_p->emac_ier);
787 if (hw_p->first_init == 0) {
789 * Connect interrupt service routines
791 irq_install_handler (VECNUM_ETH0 + (hw_p->devnum * 2),
792 (interrupt_handler_t *) enetInt, dev);
795 mtmsr (msr); /* enable interrupts again */
798 hw_p->first_init = 1;
804 static int ppc_4xx_eth_send (struct eth_device *dev, volatile void *ptr,
807 struct enet_frame *ef_ptr;
808 ulong time_start, time_now;
809 unsigned long temp_txm0;
810 EMAC_4XX_HW_PST hw_p = dev->priv;
812 ef_ptr = (struct enet_frame *) ptr;
814 /*-----------------------------------------------------------------------+
815 * Copy in our address into the frame.
816 *-----------------------------------------------------------------------*/
817 (void) memcpy (ef_ptr->source_addr, dev->enetaddr, ENET_ADDR_LENGTH);
819 /*-----------------------------------------------------------------------+
820 * If frame is too long or too short, modify length.
821 *-----------------------------------------------------------------------*/
822 /* TBS: where does the fragment go???? */
823 if (len > ENET_MAX_MTU)
826 /* memcpy ((void *) &tx_buff[tx_slot], (const void *) ptr, len); */
827 memcpy ((void *) hw_p->txbuf_ptr, (const void *) ptr, len);
829 /*-----------------------------------------------------------------------+
830 * set TX Buffer busy, and send it
831 *-----------------------------------------------------------------------*/
832 hw_p->tx[hw_p->tx_slot].ctrl = (MAL_TX_CTRL_LAST |
833 EMAC_TX_CTRL_GFCS | EMAC_TX_CTRL_GP) &
834 ~(EMAC_TX_CTRL_ISA | EMAC_TX_CTRL_RSA);
835 if ((NUM_TX_BUFF - 1) == hw_p->tx_slot)
836 hw_p->tx[hw_p->tx_slot].ctrl |= MAL_TX_CTRL_WRAP;
838 hw_p->tx[hw_p->tx_slot].data_len = (short) len;
839 hw_p->tx[hw_p->tx_slot].ctrl |= MAL_TX_CTRL_READY;
841 __asm__ volatile ("eieio");
843 out32 (EMAC_TXM0 + hw_p->hw_addr,
844 in32 (EMAC_TXM0 + hw_p->hw_addr) | EMAC_TXM0_GNP0);
846 hw_p->stats.pkts_tx++;
849 /*-----------------------------------------------------------------------+
850 * poll unitl the packet is sent and then make sure it is OK
851 *-----------------------------------------------------------------------*/
852 time_start = get_timer (0);
854 temp_txm0 = in32 (EMAC_TXM0 + hw_p->hw_addr);
855 /* loop until either TINT turns on or 3 seconds elapse */
856 if ((temp_txm0 & EMAC_TXM0_GNP0) != 0) {
857 /* transmit is done, so now check for errors
858 * If there is an error, an interrupt should
859 * happen when we return
861 time_now = get_timer (0);
862 if ((time_now - time_start) > 3000) {
871 #if defined (CONFIG_440)
873 int enetInt (struct eth_device *dev)
876 int rc = -1; /* default to not us */
877 unsigned long mal_isr;
878 unsigned long emac_isr = 0;
879 unsigned long mal_rx_eob;
880 unsigned long my_uic0msr, my_uic1msr;
882 #if defined(CONFIG_440GX)
883 unsigned long my_uic2msr;
885 EMAC_4XX_HW_PST hw_p;
888 * Because the mal is generic, we need to get the current
891 #if defined(CONFIG_NET_MULTI)
900 /* enter loop that stays in interrupt code until nothing to service */
904 my_uic0msr = mfdcr (uic0msr);
905 my_uic1msr = mfdcr (uic1msr);
906 #if defined(CONFIG_440GX)
907 my_uic2msr = mfdcr (uic2msr);
909 if (!(my_uic0msr & (UIC_MRE | UIC_MTE))
911 (UIC_ETH0 | UIC_ETH1 | UIC_MS | UIC_MTDE |
916 #if defined (CONFIG_440GX)
917 if (!(my_uic0msr & (UIC_MRE | UIC_MTE))
918 && !(my_uic2msr & (UIC_ETH2 | UIC_ETH3))) {
923 /* get and clear controller status interrupts */
924 /* look at Mal and EMAC interrupts */
925 if ((my_uic0msr & (UIC_MRE | UIC_MTE))
926 || (my_uic1msr & (UIC_MS | UIC_MTDE | UIC_MRDE))) {
927 /* we have a MAL interrupt */
928 mal_isr = mfdcr (malesr);
929 /* look for mal error */
930 if (my_uic1msr & (UIC_MS | UIC_MTDE | UIC_MRDE)) {
931 mal_err (dev, mal_isr, my_uic0msr,
932 MAL_UIC_DEF, MAL_UIC_ERR);
938 /* port by port dispatch of emac interrupts */
939 if (hw_p->devnum == 0) {
940 if (UIC_ETH0 & my_uic1msr) { /* look for EMAC errors */
941 emac_isr = in32 (EMAC_ISR + hw_p->hw_addr);
942 if ((hw_p->emac_ier & emac_isr) != 0) {
943 emac_err (dev, emac_isr);
948 if ((hw_p->emac_ier & emac_isr)
949 || (my_uic1msr & (UIC_MS | UIC_MTDE | UIC_MRDE))) {
950 mtdcr (uic0sr, UIC_MRE | UIC_MTE); /* Clear */
951 mtdcr (uic1sr, UIC_ETH0 | UIC_MS | UIC_MTDE | UIC_MRDE); /* Clear */
952 return (rc); /* we had errors so get out */
956 if (hw_p->devnum == 1) {
957 if (UIC_ETH1 & my_uic1msr) { /* look for EMAC errors */
958 emac_isr = in32 (EMAC_ISR + hw_p->hw_addr);
959 if ((hw_p->emac_ier & emac_isr) != 0) {
960 emac_err (dev, emac_isr);
965 if ((hw_p->emac_ier & emac_isr)
966 || (my_uic1msr & (UIC_MS | UIC_MTDE | UIC_MRDE))) {
967 mtdcr (uic0sr, UIC_MRE | UIC_MTE); /* Clear */
968 mtdcr (uic1sr, UIC_ETH1 | UIC_MS | UIC_MTDE | UIC_MRDE); /* Clear */
969 return (rc); /* we had errors so get out */
972 #if defined (CONFIG_440GX)
973 if (hw_p->devnum == 2) {
974 if (UIC_ETH2 & my_uic2msr) { /* look for EMAC errors */
975 emac_isr = in32 (EMAC_ISR + hw_p->hw_addr);
976 if ((hw_p->emac_ier & emac_isr) != 0) {
977 emac_err (dev, emac_isr);
982 if ((hw_p->emac_ier & emac_isr)
983 || (my_uic1msr & (UIC_MS | UIC_MTDE | UIC_MRDE))) {
984 mtdcr (uic0sr, UIC_MRE | UIC_MTE); /* Clear */
985 mtdcr (uic1sr, UIC_MS | UIC_MTDE | UIC_MRDE); /* Clear */
986 mtdcr (uic2sr, UIC_ETH2);
987 return (rc); /* we had errors so get out */
991 if (hw_p->devnum == 3) {
992 if (UIC_ETH3 & my_uic2msr) { /* look for EMAC errors */
993 emac_isr = in32 (EMAC_ISR + hw_p->hw_addr);
994 if ((hw_p->emac_ier & emac_isr) != 0) {
995 emac_err (dev, emac_isr);
1000 if ((hw_p->emac_ier & emac_isr)
1001 || (my_uic1msr & (UIC_MS | UIC_MTDE | UIC_MRDE))) {
1002 mtdcr (uic0sr, UIC_MRE | UIC_MTE); /* Clear */
1003 mtdcr (uic1sr, UIC_MS | UIC_MTDE | UIC_MRDE); /* Clear */
1004 mtdcr (uic2sr, UIC_ETH3);
1005 return (rc); /* we had errors so get out */
1008 #endif /* CONFIG_440GX */
1009 /* handle MAX TX EOB interrupt from a tx */
1010 if (my_uic0msr & UIC_MTE) {
1011 mal_rx_eob = mfdcr (maltxeobisr);
1012 mtdcr (maltxeobisr, mal_rx_eob);
1013 mtdcr (uic0sr, UIC_MTE);
1015 /* handle MAL RX EOB interupt from a receive */
1016 /* check for EOB on valid channels */
1017 if (my_uic0msr & UIC_MRE) {
1018 mal_rx_eob = mfdcr (malrxeobisr);
1019 if ((mal_rx_eob & (0x80000000 >> hw_p->devnum)) != 0) { /* call emac routine for channel x */
1021 mtdcr(malrxeobisr, mal_rx_eob); */
1022 enet_rcv (dev, emac_isr);
1023 /* indicate that we serviced an interrupt */
1028 mtdcr (uic0sr, UIC_MRE); /* Clear */
1029 mtdcr (uic1sr, UIC_MS | UIC_MTDE | UIC_MRDE); /* Clear */
1030 switch (hw_p->devnum) {
1032 mtdcr (uic1sr, UIC_ETH0);
1035 mtdcr (uic1sr, UIC_ETH1);
1037 #if defined (CONFIG_440GX)
1039 mtdcr (uic2sr, UIC_ETH2);
1042 mtdcr (uic2sr, UIC_ETH3);
1044 #endif /* CONFIG_440GX */
1053 #else /* CONFIG_440 */
1055 int enetInt (struct eth_device *dev)
1058 int rc = -1; /* default to not us */
1059 unsigned long mal_isr;
1060 unsigned long emac_isr = 0;
1061 unsigned long mal_rx_eob;
1062 unsigned long my_uicmsr;
1064 EMAC_4XX_HW_PST hw_p;
1067 * Because the mal is generic, we need to get the current
1070 #if defined(CONFIG_NET_MULTI)
1071 dev = eth_get_dev();
1078 /* enter loop that stays in interrupt code until nothing to service */
1082 my_uicmsr = mfdcr (uicmsr);
1084 if ((my_uicmsr & (MAL_UIC_DEF | EMAC_UIC_DEF)) == 0) { /* not for us */
1087 /* get and clear controller status interrupts */
1088 /* look at Mal and EMAC interrupts */
1089 if ((MAL_UIC_DEF & my_uicmsr) != 0) { /* we have a MAL interrupt */
1090 mal_isr = mfdcr (malesr);
1091 /* look for mal error */
1092 if ((my_uicmsr & MAL_UIC_ERR) != 0) {
1093 mal_err (dev, mal_isr, my_uicmsr, MAL_UIC_DEF, MAL_UIC_ERR);
1099 /* port by port dispatch of emac interrupts */
1101 if ((SEL_UIC_DEF(hw_p->devnum) & my_uicmsr) != 0) { /* look for EMAC errors */
1102 emac_isr = in32 (EMAC_ISR + hw_p->hw_addr);
1103 if ((hw_p->emac_ier & emac_isr) != 0) {
1104 emac_err (dev, emac_isr);
1109 if (((hw_p->emac_ier & emac_isr) != 0) || ((MAL_UIC_ERR & my_uicmsr) != 0)) {
1110 mtdcr (uicsr, MAL_UIC_DEF | SEL_UIC_DEF(hw_p->devnum)); /* Clear */
1111 return (rc); /* we had errors so get out */
1114 /* handle MAX TX EOB interrupt from a tx */
1115 if (my_uicmsr & UIC_MAL_TXEOB) {
1116 mal_rx_eob = mfdcr (maltxeobisr);
1117 mtdcr (maltxeobisr, mal_rx_eob);
1118 mtdcr (uicsr, UIC_MAL_TXEOB);
1120 /* handle MAL RX EOB interupt from a receive */
1121 /* check for EOB on valid channels */
1122 if (my_uicmsr & UIC_MAL_RXEOB)
1124 mal_rx_eob = mfdcr (malrxeobisr);
1125 if ((mal_rx_eob & (0x80000000 >> hw_p->devnum)) != 0) { /* call emac routine for channel x */
1127 mtdcr(malrxeobisr, mal_rx_eob); */
1128 enet_rcv (dev, emac_isr);
1129 /* indicate that we serviced an interrupt */
1134 mtdcr (uicsr, MAL_UIC_DEF|EMAC_UIC_DEF|EMAC_UIC_DEF1); /* Clear */
1141 #endif /* CONFIG_440 */
1143 /*-----------------------------------------------------------------------------+
1145 *-----------------------------------------------------------------------------*/
1146 static void mal_err (struct eth_device *dev, unsigned long isr,
1147 unsigned long uic, unsigned long maldef,
1148 unsigned long mal_errr)
1150 EMAC_4XX_HW_PST hw_p = dev->priv;
1152 mtdcr (malesr, isr); /* clear interrupt */
1154 /* clear DE interrupt */
1155 mtdcr (maltxdeir, 0xC0000000);
1156 mtdcr (malrxdeir, 0x80000000);
1158 #ifdef INFO_4XX_ENET
1159 printf ("\nMAL error occured.... ISR = %lx UIC = = %lx MAL_DEF = %lx MAL_ERR= %lx \n", isr, uic, maldef, mal_errr);
1162 eth_init (hw_p->bis); /* start again... */
1165 /*-----------------------------------------------------------------------------+
1166 * EMAC Error Routine
1167 *-----------------------------------------------------------------------------*/
1168 static void emac_err (struct eth_device *dev, unsigned long isr)
1170 EMAC_4XX_HW_PST hw_p = dev->priv;
1172 printf ("EMAC%d error occured.... ISR = %lx\n", hw_p->devnum, isr);
1173 out32 (EMAC_ISR + hw_p->hw_addr, isr);
1176 /*-----------------------------------------------------------------------------+
1177 * enet_rcv() handles the ethernet receive data
1178 *-----------------------------------------------------------------------------*/
1179 static void enet_rcv (struct eth_device *dev, unsigned long malisr)
1181 struct enet_frame *ef_ptr;
1182 unsigned long data_len;
1183 unsigned long rx_eob_isr;
1184 EMAC_4XX_HW_PST hw_p = dev->priv;
1190 rx_eob_isr = mfdcr (malrxeobisr);
1191 if ((0x80000000 >> hw_p->devnum) & rx_eob_isr) {
1193 mtdcr (malrxeobisr, rx_eob_isr);
1196 while (1) { /* do all */
1199 if ((MAL_RX_CTRL_EMPTY & hw_p->rx[i].ctrl)
1200 || (loop_count >= NUM_RX_BUFF))
1204 if (NUM_RX_BUFF == hw_p->rx_slot)
1207 data_len = (unsigned long) hw_p->rx[i].data_len; /* Get len */
1209 if (data_len > ENET_MAX_MTU) /* Check len */
1212 if (EMAC_RX_ERRORS & hw_p->rx[i].ctrl) { /* Check Errors */
1214 hw_p->stats.rx_err_log[hw_p->
1217 hw_p->rx_err_index++;
1218 if (hw_p->rx_err_index ==
1220 hw_p->rx_err_index =
1223 } /* data_len < max mtu */
1225 if (!data_len) { /* no data */
1226 hw_p->rx[i].ctrl |= MAL_RX_CTRL_EMPTY; /* Free Recv Buffer */
1228 hw_p->stats.data_len_err++; /* Error at Rx */
1233 /* Check if user has already eaten buffer */
1234 /* if not => ERROR */
1235 else if (hw_p->rx_ready[hw_p->rx_i_index] != -1) {
1236 if (hw_p->is_receiving)
1237 printf ("ERROR : Receive buffers are full!\n");
1240 hw_p->stats.rx_frames++;
1241 hw_p->stats.rx += data_len;
1242 ef_ptr = (struct enet_frame *) hw_p->rx[i].
1244 #ifdef INFO_4XX_ENET
1245 hw_p->stats.pkts_rx++;
1250 hw_p->rx_ready[hw_p->rx_i_index] = i;
1252 if (NUM_RX_BUFF == hw_p->rx_i_index)
1253 hw_p->rx_i_index = 0;
1256 * free receive buffer only when
1257 * buffer has been handled (eth_rx)
1258 rx[i].ctrl |= MAL_RX_CTRL_EMPTY;
1262 } /* if EMACK_RXCHL */
1266 static int ppc_4xx_eth_rx (struct eth_device *dev)
1271 EMAC_4XX_HW_PST hw_p = dev->priv;
1273 hw_p->is_receiving = 1; /* tell driver */
1277 * use ring buffer and
1278 * get index from rx buffer desciptor queue
1280 user_index = hw_p->rx_ready[hw_p->rx_u_index];
1281 if (user_index == -1) {
1283 break; /* nothing received - leave for() loop */
1287 mtmsr (msr & ~(MSR_EE));
1289 length = hw_p->rx[user_index].data_len;
1291 /* Pass the packet up to the protocol layers. */
1292 /* NetReceive(NetRxPackets[rxIdx], length - 4); */
1293 /* NetReceive(NetRxPackets[i], length); */
1294 NetReceive (NetRxPackets[user_index], length - 4);
1295 /* Free Recv Buffer */
1296 hw_p->rx[user_index].ctrl |= MAL_RX_CTRL_EMPTY;
1297 /* Free rx buffer descriptor queue */
1298 hw_p->rx_ready[hw_p->rx_u_index] = -1;
1300 if (NUM_RX_BUFF == hw_p->rx_u_index)
1301 hw_p->rx_u_index = 0;
1303 #ifdef INFO_4XX_ENET
1304 hw_p->stats.pkts_handled++;
1307 mtmsr (msr); /* Enable IRQ's */
1310 hw_p->is_receiving = 0; /* tell driver */
1315 int ppc_4xx_eth_initialize (bd_t * bis)
1317 static int virgin = 0;
1318 struct eth_device *dev;
1320 EMAC_4XX_HW_PST hw = NULL;
1322 #if defined(CONFIG_440GX)
1325 mfsdr (sdr_pfc1, pfc1);
1326 pfc1 &= ~(0x01e00000);
1328 mtsdr (sdr_pfc1, pfc1);
1330 /* set phy num and mode */
1331 bis->bi_phynum[0] = CONFIG_PHY_ADDR;
1332 #if defined(CONFIG_PHY1_ADDR)
1333 bis->bi_phynum[1] = CONFIG_PHY1_ADDR;
1335 #if defined(CONFIG_440GX)
1336 bis->bi_phynum[2] = CONFIG_PHY2_ADDR;
1337 bis->bi_phynum[3] = CONFIG_PHY3_ADDR;
1338 bis->bi_phymode[0] = 0;
1339 bis->bi_phymode[1] = 0;
1340 bis->bi_phymode[2] = 2;
1341 bis->bi_phymode[3] = 2;
1343 #if defined (CONFIG_440GX)
1344 ppc_4xx_eth_setup_bridge(0, bis);
1348 for (eth_num = 0; eth_num < EMAC_NUM_DEV; eth_num++) {
1350 /* See if we can actually bring up the interface, otherwise, skip it */
1352 default: /* fall through */
1354 if (memcmp (bis->bi_enetaddr, "\0\0\0\0\0\0", 6) == 0) {
1355 bis->bi_phymode[eth_num] = BI_PHYMODE_NONE;
1359 #ifdef CONFIG_HAS_ETH1
1361 if (memcmp (bis->bi_enet1addr, "\0\0\0\0\0\0", 6) == 0) {
1362 bis->bi_phymode[eth_num] = BI_PHYMODE_NONE;
1367 #ifdef CONFIG_HAS_ETH2
1369 if (memcmp (bis->bi_enet2addr, "\0\0\0\0\0\0", 6) == 0) {
1370 bis->bi_phymode[eth_num] = BI_PHYMODE_NONE;
1375 #ifdef CONFIG_HAS_ETH3
1377 if (memcmp (bis->bi_enet3addr, "\0\0\0\0\0\0", 6) == 0) {
1378 bis->bi_phymode[eth_num] = BI_PHYMODE_NONE;
1385 /* Allocate device structure */
1386 dev = (struct eth_device *) malloc (sizeof (*dev));
1388 printf ("ppc_4xx_eth_initialize: "
1389 "Cannot allocate eth_device %d\n", eth_num);
1392 memset(dev, 0, sizeof(*dev));
1394 /* Allocate our private use data */
1395 hw = (EMAC_4XX_HW_PST) malloc (sizeof (*hw));
1397 printf ("ppc_4xx_eth_initialize: "
1398 "Cannot allocate private hw data for eth_device %d",
1403 memset(hw, 0, sizeof(*hw));
1406 default: /* fall through */
1409 memcpy (dev->enetaddr, bis->bi_enetaddr, 6);
1411 #ifdef CONFIG_HAS_ETH1
1413 hw->hw_addr = 0x100;
1414 memcpy (dev->enetaddr, bis->bi_enet1addr, 6);
1417 #ifdef CONFIG_HAS_ETH2
1419 hw->hw_addr = 0x400;
1420 memcpy (dev->enetaddr, bis->bi_enet2addr, 6);
1423 #ifdef CONFIG_HAS_ETH3
1425 hw->hw_addr = 0x600;
1426 memcpy (dev->enetaddr, bis->bi_enet3addr, 6);
1431 hw->devnum = eth_num;
1432 hw->print_speed = 1;
1434 sprintf (dev->name, "ppc_4xx_eth%d", eth_num);
1435 dev->priv = (void *) hw;
1436 dev->init = ppc_4xx_eth_init;
1437 dev->halt = ppc_4xx_eth_halt;
1438 dev->send = ppc_4xx_eth_send;
1439 dev->recv = ppc_4xx_eth_rx;
1442 /* set the MAL IER ??? names may change with new spec ??? */
1444 MAL_IER_DE | MAL_IER_NE | MAL_IER_TE |
1445 MAL_IER_OPBE | MAL_IER_PLBE;
1446 mtdcr (malesr, 0xffffffff); /* clear pending interrupts */
1447 mtdcr (maltxdeir, 0xffffffff); /* clear pending interrupts */
1448 mtdcr (malrxdeir, 0xffffffff); /* clear pending interrupts */
1449 mtdcr (malier, mal_ier);
1451 /* install MAL interrupt handler */
1452 irq_install_handler (VECNUM_MS,
1453 (interrupt_handler_t *) enetInt,
1455 irq_install_handler (VECNUM_MTE,
1456 (interrupt_handler_t *) enetInt,
1458 irq_install_handler (VECNUM_MRE,
1459 (interrupt_handler_t *) enetInt,
1461 irq_install_handler (VECNUM_TXDE,
1462 (interrupt_handler_t *) enetInt,
1464 irq_install_handler (VECNUM_RXDE,
1465 (interrupt_handler_t *) enetInt,
1470 #if defined(CONFIG_NET_MULTI)
1476 } /* end for each supported device */
1481 #if !defined(CONFIG_NET_MULTI)
1482 void eth_halt (void) {
1484 ppc_4xx_eth_halt(emac0_dev);
1490 int eth_init (bd_t *bis)
1492 ppc_4xx_eth_initialize(bis);
1494 return ppc_4xx_eth_init(emac0_dev, bis);
1496 printf("ERROR: ethaddr not set!\n");
1501 int eth_send(volatile void *packet, int length)
1503 return (ppc_4xx_eth_send(emac0_dev, packet, length));
1508 return (ppc_4xx_eth_rx(emac0_dev));
1510 #endif /* !defined(CONFIG_NET_MULTI) */
1512 #endif /* #if (CONFIG_COMMANDS & CFG_CMD_NET) */