| Linus Torvalds | 1da177e | 2005-04-16 15:20:36 -0700 | [diff] [blame] | 1 | /* | 
|  | 2 | * Network device driver for the BMAC ethernet controller on | 
|  | 3 | * Apple Powermacs.  Assumes it's under a DBDMA controller. | 
|  | 4 | * | 
|  | 5 | * Copyright (C) 1998 Randy Gobbel. | 
|  | 6 | * | 
|  | 7 | * May 1999, Al Viro: proper release of /proc/net/bmac entry, switched to | 
|  | 8 | * dynamic procfs inode. | 
|  | 9 | */ | 
|  | 10 | #include <linux/config.h> | 
|  | 11 | #include <linux/module.h> | 
|  | 12 | #include <linux/kernel.h> | 
|  | 13 | #include <linux/netdevice.h> | 
|  | 14 | #include <linux/etherdevice.h> | 
|  | 15 | #include <linux/delay.h> | 
|  | 16 | #include <linux/string.h> | 
|  | 17 | #include <linux/timer.h> | 
|  | 18 | #include <linux/proc_fs.h> | 
|  | 19 | #include <linux/init.h> | 
|  | 20 | #include <linux/spinlock.h> | 
|  | 21 | #include <linux/crc32.h> | 
|  | 22 | #include <asm/prom.h> | 
|  | 23 | #include <asm/dbdma.h> | 
|  | 24 | #include <asm/io.h> | 
|  | 25 | #include <asm/page.h> | 
|  | 26 | #include <asm/pgtable.h> | 
|  | 27 | #include <asm/machdep.h> | 
|  | 28 | #include <asm/pmac_feature.h> | 
|  | 29 | #include <asm/macio.h> | 
|  | 30 | #include <asm/irq.h> | 
|  | 31 |  | 
|  | 32 | #include "bmac.h" | 
|  | 33 |  | 
|  | 34 | #define trunc_page(x)	((void *)(((unsigned long)(x)) & ~((unsigned long)(PAGE_SIZE - 1)))) | 
|  | 35 | #define round_page(x)	trunc_page(((unsigned long)(x)) + ((unsigned long)(PAGE_SIZE - 1))) | 
|  | 36 |  | 
|  | 37 | /* | 
|  | 38 | * CRC polynomial - used in working out multicast filter bits. | 
|  | 39 | */ | 
|  | 40 | #define ENET_CRCPOLY 0x04c11db7 | 
|  | 41 |  | 
|  | 42 | /* switch to use multicast code lifted from sunhme driver */ | 
|  | 43 | #define SUNHME_MULTICAST | 
|  | 44 |  | 
|  | 45 | #define N_RX_RING	64 | 
|  | 46 | #define N_TX_RING	32 | 
|  | 47 | #define MAX_TX_ACTIVE	1 | 
|  | 48 | #define ETHERCRC	4 | 
|  | 49 | #define ETHERMINPACKET	64 | 
|  | 50 | #define ETHERMTU	1500 | 
|  | 51 | #define RX_BUFLEN	(ETHERMTU + 14 + ETHERCRC + 2) | 
|  | 52 | #define TX_TIMEOUT	HZ	/* 1 second */ | 
|  | 53 |  | 
|  | 54 | /* Bits in transmit DMA status */ | 
|  | 55 | #define TX_DMA_ERR	0x80 | 
|  | 56 |  | 
|  | 57 | #define XXDEBUG(args) | 
|  | 58 |  | 
|  | 59 | struct bmac_data { | 
|  | 60 | /* volatile struct bmac *bmac; */ | 
|  | 61 | struct sk_buff_head *queue; | 
|  | 62 | volatile struct dbdma_regs __iomem *tx_dma; | 
|  | 63 | int tx_dma_intr; | 
|  | 64 | volatile struct dbdma_regs __iomem *rx_dma; | 
|  | 65 | int rx_dma_intr; | 
|  | 66 | volatile struct dbdma_cmd *tx_cmds;	/* xmit dma command list */ | 
|  | 67 | volatile struct dbdma_cmd *rx_cmds;	/* recv dma command list */ | 
|  | 68 | struct macio_dev *mdev; | 
|  | 69 | int is_bmac_plus; | 
|  | 70 | struct sk_buff *rx_bufs[N_RX_RING]; | 
|  | 71 | int rx_fill; | 
|  | 72 | int rx_empty; | 
|  | 73 | struct sk_buff *tx_bufs[N_TX_RING]; | 
|  | 74 | int tx_fill; | 
|  | 75 | int tx_empty; | 
|  | 76 | unsigned char tx_fullup; | 
|  | 77 | struct net_device_stats stats; | 
|  | 78 | struct timer_list tx_timeout; | 
|  | 79 | int timeout_active; | 
|  | 80 | int sleeping; | 
|  | 81 | int opened; | 
|  | 82 | unsigned short hash_use_count[64]; | 
|  | 83 | unsigned short hash_table_mask[4]; | 
|  | 84 | spinlock_t lock; | 
|  | 85 | }; | 
|  | 86 |  | 
|  | 87 | #if 0 /* Move that to ethtool */ | 
|  | 88 |  | 
|  | 89 | typedef struct bmac_reg_entry { | 
|  | 90 | char *name; | 
|  | 91 | unsigned short reg_offset; | 
|  | 92 | } bmac_reg_entry_t; | 
|  | 93 |  | 
|  | 94 | #define N_REG_ENTRIES 31 | 
|  | 95 |  | 
|  | 96 | static bmac_reg_entry_t reg_entries[N_REG_ENTRIES] = { | 
|  | 97 | {"MEMADD", MEMADD}, | 
|  | 98 | {"MEMDATAHI", MEMDATAHI}, | 
|  | 99 | {"MEMDATALO", MEMDATALO}, | 
|  | 100 | {"TXPNTR", TXPNTR}, | 
|  | 101 | {"RXPNTR", RXPNTR}, | 
|  | 102 | {"IPG1", IPG1}, | 
|  | 103 | {"IPG2", IPG2}, | 
|  | 104 | {"ALIMIT", ALIMIT}, | 
|  | 105 | {"SLOT", SLOT}, | 
|  | 106 | {"PALEN", PALEN}, | 
|  | 107 | {"PAPAT", PAPAT}, | 
|  | 108 | {"TXSFD", TXSFD}, | 
|  | 109 | {"JAM", JAM}, | 
|  | 110 | {"TXCFG", TXCFG}, | 
|  | 111 | {"TXMAX", TXMAX}, | 
|  | 112 | {"TXMIN", TXMIN}, | 
|  | 113 | {"PAREG", PAREG}, | 
|  | 114 | {"DCNT", DCNT}, | 
|  | 115 | {"NCCNT", NCCNT}, | 
|  | 116 | {"NTCNT", NTCNT}, | 
|  | 117 | {"EXCNT", EXCNT}, | 
|  | 118 | {"LTCNT", LTCNT}, | 
|  | 119 | {"TXSM", TXSM}, | 
|  | 120 | {"RXCFG", RXCFG}, | 
|  | 121 | {"RXMAX", RXMAX}, | 
|  | 122 | {"RXMIN", RXMIN}, | 
|  | 123 | {"FRCNT", FRCNT}, | 
|  | 124 | {"AECNT", AECNT}, | 
|  | 125 | {"FECNT", FECNT}, | 
|  | 126 | {"RXSM", RXSM}, | 
|  | 127 | {"RXCV", RXCV} | 
|  | 128 | }; | 
|  | 129 |  | 
|  | 130 | #endif | 
|  | 131 |  | 
|  | 132 | static unsigned char *bmac_emergency_rxbuf; | 
|  | 133 |  | 
|  | 134 | /* | 
|  | 135 | * Number of bytes of private data per BMAC: allow enough for | 
|  | 136 | * the rx and tx dma commands plus a branch dma command each, | 
|  | 137 | * and another 16 bytes to allow us to align the dma command | 
|  | 138 | * buffers on a 16 byte boundary. | 
|  | 139 | */ | 
|  | 140 | #define PRIV_BYTES	(sizeof(struct bmac_data) \ | 
|  | 141 | + (N_RX_RING + N_TX_RING + 4) * sizeof(struct dbdma_cmd) \ | 
|  | 142 | + sizeof(struct sk_buff_head)) | 
|  | 143 |  | 
|  | 144 | static unsigned char bitrev(unsigned char b); | 
|  | 145 | static int bmac_open(struct net_device *dev); | 
|  | 146 | static int bmac_close(struct net_device *dev); | 
|  | 147 | static int bmac_transmit_packet(struct sk_buff *skb, struct net_device *dev); | 
|  | 148 | static struct net_device_stats *bmac_stats(struct net_device *dev); | 
|  | 149 | static void bmac_set_multicast(struct net_device *dev); | 
|  | 150 | static void bmac_reset_and_enable(struct net_device *dev); | 
|  | 151 | static void bmac_start_chip(struct net_device *dev); | 
|  | 152 | static void bmac_init_chip(struct net_device *dev); | 
|  | 153 | static void bmac_init_registers(struct net_device *dev); | 
|  | 154 | static void bmac_enable_and_reset_chip(struct net_device *dev); | 
|  | 155 | static int bmac_set_address(struct net_device *dev, void *addr); | 
|  | 156 | static irqreturn_t bmac_misc_intr(int irq, void *dev_id, struct pt_regs *regs); | 
|  | 157 | static irqreturn_t bmac_txdma_intr(int irq, void *dev_id, struct pt_regs *regs); | 
|  | 158 | static irqreturn_t bmac_rxdma_intr(int irq, void *dev_id, struct pt_regs *regs); | 
|  | 159 | static void bmac_set_timeout(struct net_device *dev); | 
|  | 160 | static void bmac_tx_timeout(unsigned long data); | 
|  | 161 | static int bmac_output(struct sk_buff *skb, struct net_device *dev); | 
|  | 162 | static void bmac_start(struct net_device *dev); | 
|  | 163 |  | 
|  | 164 | #define	DBDMA_SET(x)	( ((x) | (x) << 16) ) | 
|  | 165 | #define	DBDMA_CLEAR(x)	( (x) << 16) | 
|  | 166 |  | 
|  | 167 | static inline void | 
|  | 168 | dbdma_st32(volatile __u32 __iomem *a, unsigned long x) | 
|  | 169 | { | 
|  | 170 | __asm__ volatile( "stwbrx %0,0,%1" : : "r" (x), "r" (a) : "memory"); | 
|  | 171 | return; | 
|  | 172 | } | 
|  | 173 |  | 
|  | 174 | static inline unsigned long | 
|  | 175 | dbdma_ld32(volatile __u32 __iomem *a) | 
|  | 176 | { | 
|  | 177 | __u32 swap; | 
|  | 178 | __asm__ volatile ("lwbrx %0,0,%1" :  "=r" (swap) : "r" (a)); | 
|  | 179 | return swap; | 
|  | 180 | } | 
|  | 181 |  | 
|  | 182 | static void | 
|  | 183 | dbdma_continue(volatile struct dbdma_regs __iomem *dmap) | 
|  | 184 | { | 
|  | 185 | dbdma_st32(&dmap->control, | 
|  | 186 | DBDMA_SET(RUN|WAKE) | DBDMA_CLEAR(PAUSE|DEAD)); | 
|  | 187 | eieio(); | 
|  | 188 | } | 
|  | 189 |  | 
|  | 190 | static void | 
|  | 191 | dbdma_reset(volatile struct dbdma_regs __iomem *dmap) | 
|  | 192 | { | 
|  | 193 | dbdma_st32(&dmap->control, | 
|  | 194 | DBDMA_CLEAR(ACTIVE|DEAD|WAKE|FLUSH|PAUSE|RUN)); | 
|  | 195 | eieio(); | 
|  | 196 | while (dbdma_ld32(&dmap->status) & RUN) | 
|  | 197 | eieio(); | 
|  | 198 | } | 
|  | 199 |  | 
|  | 200 | static void | 
|  | 201 | dbdma_setcmd(volatile struct dbdma_cmd *cp, | 
|  | 202 | unsigned short cmd, unsigned count, unsigned long addr, | 
|  | 203 | unsigned long cmd_dep) | 
|  | 204 | { | 
|  | 205 | out_le16(&cp->command, cmd); | 
|  | 206 | out_le16(&cp->req_count, count); | 
|  | 207 | out_le32(&cp->phy_addr, addr); | 
|  | 208 | out_le32(&cp->cmd_dep, cmd_dep); | 
|  | 209 | out_le16(&cp->xfer_status, 0); | 
|  | 210 | out_le16(&cp->res_count, 0); | 
|  | 211 | } | 
|  | 212 |  | 
|  | 213 | static inline | 
|  | 214 | void bmwrite(struct net_device *dev, unsigned long reg_offset, unsigned data ) | 
|  | 215 | { | 
|  | 216 | out_le16((void __iomem *)dev->base_addr + reg_offset, data); | 
|  | 217 | } | 
|  | 218 |  | 
|  | 219 |  | 
|  | 220 | static inline | 
|  | 221 | volatile unsigned short bmread(struct net_device *dev, unsigned long reg_offset ) | 
|  | 222 | { | 
|  | 223 | return in_le16((void __iomem *)dev->base_addr + reg_offset); | 
|  | 224 | } | 
|  | 225 |  | 
|  | 226 | static void | 
|  | 227 | bmac_enable_and_reset_chip(struct net_device *dev) | 
|  | 228 | { | 
|  | 229 | struct bmac_data *bp = netdev_priv(dev); | 
|  | 230 | volatile struct dbdma_regs __iomem *rd = bp->rx_dma; | 
|  | 231 | volatile struct dbdma_regs __iomem *td = bp->tx_dma; | 
|  | 232 |  | 
|  | 233 | if (rd) | 
|  | 234 | dbdma_reset(rd); | 
|  | 235 | if (td) | 
|  | 236 | dbdma_reset(td); | 
|  | 237 |  | 
|  | 238 | pmac_call_feature(PMAC_FTR_BMAC_ENABLE, macio_get_of_node(bp->mdev), 0, 1); | 
|  | 239 | } | 
|  | 240 |  | 
|  | 241 | #define MIFDELAY	udelay(10) | 
|  | 242 |  | 
|  | 243 | static unsigned int | 
|  | 244 | bmac_mif_readbits(struct net_device *dev, int nb) | 
|  | 245 | { | 
|  | 246 | unsigned int val = 0; | 
|  | 247 |  | 
|  | 248 | while (--nb >= 0) { | 
|  | 249 | bmwrite(dev, MIFCSR, 0); | 
|  | 250 | MIFDELAY; | 
|  | 251 | if (bmread(dev, MIFCSR) & 8) | 
|  | 252 | val |= 1 << nb; | 
|  | 253 | bmwrite(dev, MIFCSR, 1); | 
|  | 254 | MIFDELAY; | 
|  | 255 | } | 
|  | 256 | bmwrite(dev, MIFCSR, 0); | 
|  | 257 | MIFDELAY; | 
|  | 258 | bmwrite(dev, MIFCSR, 1); | 
|  | 259 | MIFDELAY; | 
|  | 260 | return val; | 
|  | 261 | } | 
|  | 262 |  | 
|  | 263 | static void | 
|  | 264 | bmac_mif_writebits(struct net_device *dev, unsigned int val, int nb) | 
|  | 265 | { | 
|  | 266 | int b; | 
|  | 267 |  | 
|  | 268 | while (--nb >= 0) { | 
|  | 269 | b = (val & (1 << nb))? 6: 4; | 
|  | 270 | bmwrite(dev, MIFCSR, b); | 
|  | 271 | MIFDELAY; | 
|  | 272 | bmwrite(dev, MIFCSR, b|1); | 
|  | 273 | MIFDELAY; | 
|  | 274 | } | 
|  | 275 | } | 
|  | 276 |  | 
|  | 277 | static unsigned int | 
|  | 278 | bmac_mif_read(struct net_device *dev, unsigned int addr) | 
|  | 279 | { | 
|  | 280 | unsigned int val; | 
|  | 281 |  | 
|  | 282 | bmwrite(dev, MIFCSR, 4); | 
|  | 283 | MIFDELAY; | 
|  | 284 | bmac_mif_writebits(dev, ~0U, 32); | 
|  | 285 | bmac_mif_writebits(dev, 6, 4); | 
|  | 286 | bmac_mif_writebits(dev, addr, 10); | 
|  | 287 | bmwrite(dev, MIFCSR, 2); | 
|  | 288 | MIFDELAY; | 
|  | 289 | bmwrite(dev, MIFCSR, 1); | 
|  | 290 | MIFDELAY; | 
|  | 291 | val = bmac_mif_readbits(dev, 17); | 
|  | 292 | bmwrite(dev, MIFCSR, 4); | 
|  | 293 | MIFDELAY; | 
|  | 294 | return val; | 
|  | 295 | } | 
|  | 296 |  | 
|  | 297 | static void | 
|  | 298 | bmac_mif_write(struct net_device *dev, unsigned int addr, unsigned int val) | 
|  | 299 | { | 
|  | 300 | bmwrite(dev, MIFCSR, 4); | 
|  | 301 | MIFDELAY; | 
|  | 302 | bmac_mif_writebits(dev, ~0U, 32); | 
|  | 303 | bmac_mif_writebits(dev, 5, 4); | 
|  | 304 | bmac_mif_writebits(dev, addr, 10); | 
|  | 305 | bmac_mif_writebits(dev, 2, 2); | 
|  | 306 | bmac_mif_writebits(dev, val, 16); | 
|  | 307 | bmac_mif_writebits(dev, 3, 2); | 
|  | 308 | } | 
|  | 309 |  | 
|  | 310 | static void | 
|  | 311 | bmac_init_registers(struct net_device *dev) | 
|  | 312 | { | 
|  | 313 | struct bmac_data *bp = netdev_priv(dev); | 
|  | 314 | volatile unsigned short regValue; | 
|  | 315 | unsigned short *pWord16; | 
|  | 316 | int i; | 
|  | 317 |  | 
|  | 318 | /* XXDEBUG(("bmac: enter init_registers\n")); */ | 
|  | 319 |  | 
|  | 320 | bmwrite(dev, RXRST, RxResetValue); | 
|  | 321 | bmwrite(dev, TXRST, TxResetBit); | 
|  | 322 |  | 
|  | 323 | i = 100; | 
|  | 324 | do { | 
|  | 325 | --i; | 
|  | 326 | udelay(10000); | 
|  | 327 | regValue = bmread(dev, TXRST); /* wait for reset to clear..acknowledge */ | 
|  | 328 | } while ((regValue & TxResetBit) && i > 0); | 
|  | 329 |  | 
|  | 330 | if (!bp->is_bmac_plus) { | 
|  | 331 | regValue = bmread(dev, XCVRIF); | 
|  | 332 | regValue |= ClkBit | SerialMode | COLActiveLow; | 
|  | 333 | bmwrite(dev, XCVRIF, regValue); | 
|  | 334 | udelay(10000); | 
|  | 335 | } | 
|  | 336 |  | 
|  | 337 | bmwrite(dev, RSEED, (unsigned short)0x1968); | 
|  | 338 |  | 
|  | 339 | regValue = bmread(dev, XIFC); | 
|  | 340 | regValue |= TxOutputEnable; | 
|  | 341 | bmwrite(dev, XIFC, regValue); | 
|  | 342 |  | 
|  | 343 | bmread(dev, PAREG); | 
|  | 344 |  | 
|  | 345 | /* set collision counters to 0 */ | 
|  | 346 | bmwrite(dev, NCCNT, 0); | 
|  | 347 | bmwrite(dev, NTCNT, 0); | 
|  | 348 | bmwrite(dev, EXCNT, 0); | 
|  | 349 | bmwrite(dev, LTCNT, 0); | 
|  | 350 |  | 
|  | 351 | /* set rx counters to 0 */ | 
|  | 352 | bmwrite(dev, FRCNT, 0); | 
|  | 353 | bmwrite(dev, LECNT, 0); | 
|  | 354 | bmwrite(dev, AECNT, 0); | 
|  | 355 | bmwrite(dev, FECNT, 0); | 
|  | 356 | bmwrite(dev, RXCV, 0); | 
|  | 357 |  | 
|  | 358 | /* set tx fifo information */ | 
|  | 359 | bmwrite(dev, TXTH, 4);	/* 4 octets before tx starts */ | 
|  | 360 |  | 
|  | 361 | bmwrite(dev, TXFIFOCSR, 0);	/* first disable txFIFO */ | 
|  | 362 | bmwrite(dev, TXFIFOCSR, TxFIFOEnable ); | 
|  | 363 |  | 
|  | 364 | /* set rx fifo information */ | 
|  | 365 | bmwrite(dev, RXFIFOCSR, 0);	/* first disable rxFIFO */ | 
|  | 366 | bmwrite(dev, RXFIFOCSR, RxFIFOEnable ); | 
|  | 367 |  | 
|  | 368 | //bmwrite(dev, TXCFG, TxMACEnable);	       	/* TxNeverGiveUp maybe later */ | 
|  | 369 | bmread(dev, STATUS);		/* read it just to clear it */ | 
|  | 370 |  | 
|  | 371 | /* zero out the chip Hash Filter registers */ | 
|  | 372 | for (i=0; i<4; i++) bp->hash_table_mask[i] = 0; | 
|  | 373 | bmwrite(dev, BHASH3, bp->hash_table_mask[0]); 	/* bits 15 - 0 */ | 
|  | 374 | bmwrite(dev, BHASH2, bp->hash_table_mask[1]); 	/* bits 31 - 16 */ | 
|  | 375 | bmwrite(dev, BHASH1, bp->hash_table_mask[2]); 	/* bits 47 - 32 */ | 
|  | 376 | bmwrite(dev, BHASH0, bp->hash_table_mask[3]); 	/* bits 63 - 48 */ | 
|  | 377 |  | 
|  | 378 | pWord16 = (unsigned short *)dev->dev_addr; | 
|  | 379 | bmwrite(dev, MADD0, *pWord16++); | 
|  | 380 | bmwrite(dev, MADD1, *pWord16++); | 
|  | 381 | bmwrite(dev, MADD2, *pWord16); | 
|  | 382 |  | 
|  | 383 | bmwrite(dev, RXCFG, RxCRCNoStrip | RxHashFilterEnable | RxRejectOwnPackets); | 
|  | 384 |  | 
|  | 385 | bmwrite(dev, INTDISABLE, EnableNormal); | 
|  | 386 |  | 
|  | 387 | return; | 
|  | 388 | } | 
|  | 389 |  | 
|  | 390 | #if 0 | 
|  | 391 | static void | 
|  | 392 | bmac_disable_interrupts(struct net_device *dev) | 
|  | 393 | { | 
|  | 394 | bmwrite(dev, INTDISABLE, DisableAll); | 
|  | 395 | } | 
|  | 396 |  | 
|  | 397 | static void | 
|  | 398 | bmac_enable_interrupts(struct net_device *dev) | 
|  | 399 | { | 
|  | 400 | bmwrite(dev, INTDISABLE, EnableNormal); | 
|  | 401 | } | 
|  | 402 | #endif | 
|  | 403 |  | 
|  | 404 |  | 
|  | 405 | static void | 
|  | 406 | bmac_start_chip(struct net_device *dev) | 
|  | 407 | { | 
|  | 408 | struct bmac_data *bp = netdev_priv(dev); | 
|  | 409 | volatile struct dbdma_regs __iomem *rd = bp->rx_dma; | 
|  | 410 | unsigned short	oldConfig; | 
|  | 411 |  | 
|  | 412 | /* enable rx dma channel */ | 
|  | 413 | dbdma_continue(rd); | 
|  | 414 |  | 
|  | 415 | oldConfig = bmread(dev, TXCFG); | 
|  | 416 | bmwrite(dev, TXCFG, oldConfig | TxMACEnable ); | 
|  | 417 |  | 
|  | 418 | /* turn on rx plus any other bits already on (promiscuous possibly) */ | 
|  | 419 | oldConfig = bmread(dev, RXCFG); | 
|  | 420 | bmwrite(dev, RXCFG, oldConfig | RxMACEnable ); | 
|  | 421 | udelay(20000); | 
|  | 422 | } | 
|  | 423 |  | 
|  | 424 | static void | 
|  | 425 | bmac_init_phy(struct net_device *dev) | 
|  | 426 | { | 
|  | 427 | unsigned int addr; | 
|  | 428 | struct bmac_data *bp = netdev_priv(dev); | 
|  | 429 |  | 
|  | 430 | printk(KERN_DEBUG "phy registers:"); | 
|  | 431 | for (addr = 0; addr < 32; ++addr) { | 
|  | 432 | if ((addr & 7) == 0) | 
|  | 433 | printk("\n" KERN_DEBUG); | 
|  | 434 | printk(" %.4x", bmac_mif_read(dev, addr)); | 
|  | 435 | } | 
|  | 436 | printk("\n"); | 
|  | 437 | if (bp->is_bmac_plus) { | 
|  | 438 | unsigned int capable, ctrl; | 
|  | 439 |  | 
|  | 440 | ctrl = bmac_mif_read(dev, 0); | 
|  | 441 | capable = ((bmac_mif_read(dev, 1) & 0xf800) >> 6) | 1; | 
|  | 442 | if (bmac_mif_read(dev, 4) != capable | 
|  | 443 | || (ctrl & 0x1000) == 0) { | 
|  | 444 | bmac_mif_write(dev, 4, capable); | 
|  | 445 | bmac_mif_write(dev, 0, 0x1200); | 
|  | 446 | } else | 
|  | 447 | bmac_mif_write(dev, 0, 0x1000); | 
|  | 448 | } | 
|  | 449 | } | 
|  | 450 |  | 
|  | 451 | static void bmac_init_chip(struct net_device *dev) | 
|  | 452 | { | 
|  | 453 | bmac_init_phy(dev); | 
|  | 454 | bmac_init_registers(dev); | 
|  | 455 | } | 
|  | 456 |  | 
|  | 457 | #ifdef CONFIG_PM | 
| Pavel Machek | 05adc3b | 2005-04-16 15:25:25 -0700 | [diff] [blame] | 458 | static int bmac_suspend(struct macio_dev *mdev, pm_message_t state) | 
| Linus Torvalds | 1da177e | 2005-04-16 15:20:36 -0700 | [diff] [blame] | 459 | { | 
|  | 460 | struct net_device* dev = macio_get_drvdata(mdev); | 
|  | 461 | struct bmac_data *bp = netdev_priv(dev); | 
|  | 462 | unsigned long flags; | 
|  | 463 | unsigned short config; | 
|  | 464 | int i; | 
|  | 465 |  | 
|  | 466 | netif_device_detach(dev); | 
|  | 467 | /* prolly should wait for dma to finish & turn off the chip */ | 
|  | 468 | spin_lock_irqsave(&bp->lock, flags); | 
|  | 469 | if (bp->timeout_active) { | 
|  | 470 | del_timer(&bp->tx_timeout); | 
|  | 471 | bp->timeout_active = 0; | 
|  | 472 | } | 
|  | 473 | disable_irq(dev->irq); | 
|  | 474 | disable_irq(bp->tx_dma_intr); | 
|  | 475 | disable_irq(bp->rx_dma_intr); | 
|  | 476 | bp->sleeping = 1; | 
|  | 477 | spin_unlock_irqrestore(&bp->lock, flags); | 
|  | 478 | if (bp->opened) { | 
|  | 479 | volatile struct dbdma_regs __iomem *rd = bp->rx_dma; | 
|  | 480 | volatile struct dbdma_regs __iomem *td = bp->tx_dma; | 
|  | 481 |  | 
|  | 482 | config = bmread(dev, RXCFG); | 
|  | 483 | bmwrite(dev, RXCFG, (config & ~RxMACEnable)); | 
|  | 484 | config = bmread(dev, TXCFG); | 
|  | 485 | bmwrite(dev, TXCFG, (config & ~TxMACEnable)); | 
|  | 486 | bmwrite(dev, INTDISABLE, DisableAll); /* disable all intrs */ | 
|  | 487 | /* disable rx and tx dma */ | 
|  | 488 | st_le32(&rd->control, DBDMA_CLEAR(RUN|PAUSE|FLUSH|WAKE));	/* clear run bit */ | 
|  | 489 | st_le32(&td->control, DBDMA_CLEAR(RUN|PAUSE|FLUSH|WAKE));	/* clear run bit */ | 
|  | 490 | /* free some skb's */ | 
|  | 491 | for (i=0; i<N_RX_RING; i++) { | 
|  | 492 | if (bp->rx_bufs[i] != NULL) { | 
|  | 493 | dev_kfree_skb(bp->rx_bufs[i]); | 
|  | 494 | bp->rx_bufs[i] = NULL; | 
|  | 495 | } | 
|  | 496 | } | 
|  | 497 | for (i = 0; i<N_TX_RING; i++) { | 
|  | 498 | if (bp->tx_bufs[i] != NULL) { | 
|  | 499 | dev_kfree_skb(bp->tx_bufs[i]); | 
|  | 500 | bp->tx_bufs[i] = NULL; | 
|  | 501 | } | 
|  | 502 | } | 
|  | 503 | } | 
|  | 504 | pmac_call_feature(PMAC_FTR_BMAC_ENABLE, macio_get_of_node(bp->mdev), 0, 0); | 
|  | 505 | return 0; | 
|  | 506 | } | 
|  | 507 |  | 
|  | 508 | static int bmac_resume(struct macio_dev *mdev) | 
|  | 509 | { | 
|  | 510 | struct net_device* dev = macio_get_drvdata(mdev); | 
|  | 511 | struct bmac_data *bp = netdev_priv(dev); | 
|  | 512 |  | 
|  | 513 | /* see if this is enough */ | 
|  | 514 | if (bp->opened) | 
|  | 515 | bmac_reset_and_enable(dev); | 
|  | 516 |  | 
|  | 517 | enable_irq(dev->irq); | 
|  | 518 | enable_irq(bp->tx_dma_intr); | 
|  | 519 | enable_irq(bp->rx_dma_intr); | 
|  | 520 | netif_device_attach(dev); | 
|  | 521 |  | 
|  | 522 | return 0; | 
|  | 523 | } | 
|  | 524 | #endif /* CONFIG_PM */ | 
|  | 525 |  | 
|  | 526 | static int bmac_set_address(struct net_device *dev, void *addr) | 
|  | 527 | { | 
|  | 528 | struct bmac_data *bp = netdev_priv(dev); | 
|  | 529 | unsigned char *p = addr; | 
|  | 530 | unsigned short *pWord16; | 
|  | 531 | unsigned long flags; | 
|  | 532 | int i; | 
|  | 533 |  | 
|  | 534 | XXDEBUG(("bmac: enter set_address\n")); | 
|  | 535 | spin_lock_irqsave(&bp->lock, flags); | 
|  | 536 |  | 
|  | 537 | for (i = 0; i < 6; ++i) { | 
|  | 538 | dev->dev_addr[i] = p[i]; | 
|  | 539 | } | 
|  | 540 | /* load up the hardware address */ | 
|  | 541 | pWord16  = (unsigned short *)dev->dev_addr; | 
|  | 542 | bmwrite(dev, MADD0, *pWord16++); | 
|  | 543 | bmwrite(dev, MADD1, *pWord16++); | 
|  | 544 | bmwrite(dev, MADD2, *pWord16); | 
|  | 545 |  | 
|  | 546 | spin_unlock_irqrestore(&bp->lock, flags); | 
|  | 547 | XXDEBUG(("bmac: exit set_address\n")); | 
|  | 548 | return 0; | 
|  | 549 | } | 
|  | 550 |  | 
|  | 551 | static inline void bmac_set_timeout(struct net_device *dev) | 
|  | 552 | { | 
|  | 553 | struct bmac_data *bp = netdev_priv(dev); | 
|  | 554 | unsigned long flags; | 
|  | 555 |  | 
|  | 556 | spin_lock_irqsave(&bp->lock, flags); | 
|  | 557 | if (bp->timeout_active) | 
|  | 558 | del_timer(&bp->tx_timeout); | 
|  | 559 | bp->tx_timeout.expires = jiffies + TX_TIMEOUT; | 
|  | 560 | bp->tx_timeout.function = bmac_tx_timeout; | 
|  | 561 | bp->tx_timeout.data = (unsigned long) dev; | 
|  | 562 | add_timer(&bp->tx_timeout); | 
|  | 563 | bp->timeout_active = 1; | 
|  | 564 | spin_unlock_irqrestore(&bp->lock, flags); | 
|  | 565 | } | 
|  | 566 |  | 
|  | 567 | static void | 
|  | 568 | bmac_construct_xmt(struct sk_buff *skb, volatile struct dbdma_cmd *cp) | 
|  | 569 | { | 
|  | 570 | void *vaddr; | 
|  | 571 | unsigned long baddr; | 
|  | 572 | unsigned long len; | 
|  | 573 |  | 
|  | 574 | len = skb->len; | 
|  | 575 | vaddr = skb->data; | 
|  | 576 | baddr = virt_to_bus(vaddr); | 
|  | 577 |  | 
|  | 578 | dbdma_setcmd(cp, (OUTPUT_LAST | INTR_ALWAYS | WAIT_IFCLR), len, baddr, 0); | 
|  | 579 | } | 
|  | 580 |  | 
|  | 581 | static void | 
|  | 582 | bmac_construct_rxbuff(struct sk_buff *skb, volatile struct dbdma_cmd *cp) | 
|  | 583 | { | 
|  | 584 | unsigned char *addr = skb? skb->data: bmac_emergency_rxbuf; | 
|  | 585 |  | 
|  | 586 | dbdma_setcmd(cp, (INPUT_LAST | INTR_ALWAYS), RX_BUFLEN, | 
|  | 587 | virt_to_bus(addr), 0); | 
|  | 588 | } | 
|  | 589 |  | 
|  | 590 | /* Bit-reverse one byte of an ethernet hardware address. */ | 
|  | 591 | static unsigned char | 
|  | 592 | bitrev(unsigned char b) | 
|  | 593 | { | 
|  | 594 | int d = 0, i; | 
|  | 595 |  | 
|  | 596 | for (i = 0; i < 8; ++i, b >>= 1) | 
|  | 597 | d = (d << 1) | (b & 1); | 
|  | 598 | return d; | 
|  | 599 | } | 
|  | 600 |  | 
|  | 601 |  | 
|  | 602 | static void | 
|  | 603 | bmac_init_tx_ring(struct bmac_data *bp) | 
|  | 604 | { | 
|  | 605 | volatile struct dbdma_regs __iomem *td = bp->tx_dma; | 
|  | 606 |  | 
|  | 607 | memset((char *)bp->tx_cmds, 0, (N_TX_RING+1) * sizeof(struct dbdma_cmd)); | 
|  | 608 |  | 
|  | 609 | bp->tx_empty = 0; | 
|  | 610 | bp->tx_fill = 0; | 
|  | 611 | bp->tx_fullup = 0; | 
|  | 612 |  | 
|  | 613 | /* put a branch at the end of the tx command list */ | 
|  | 614 | dbdma_setcmd(&bp->tx_cmds[N_TX_RING], | 
|  | 615 | (DBDMA_NOP | BR_ALWAYS), 0, 0, virt_to_bus(bp->tx_cmds)); | 
|  | 616 |  | 
|  | 617 | /* reset tx dma */ | 
|  | 618 | dbdma_reset(td); | 
|  | 619 | out_le32(&td->wait_sel, 0x00200020); | 
|  | 620 | out_le32(&td->cmdptr, virt_to_bus(bp->tx_cmds)); | 
|  | 621 | } | 
|  | 622 |  | 
|  | 623 | static int | 
|  | 624 | bmac_init_rx_ring(struct bmac_data *bp) | 
|  | 625 | { | 
|  | 626 | volatile struct dbdma_regs __iomem *rd = bp->rx_dma; | 
|  | 627 | int i; | 
|  | 628 | struct sk_buff *skb; | 
|  | 629 |  | 
|  | 630 | /* initialize list of sk_buffs for receiving and set up recv dma */ | 
|  | 631 | memset((char *)bp->rx_cmds, 0, | 
|  | 632 | (N_RX_RING + 1) * sizeof(struct dbdma_cmd)); | 
|  | 633 | for (i = 0; i < N_RX_RING; i++) { | 
|  | 634 | if ((skb = bp->rx_bufs[i]) == NULL) { | 
|  | 635 | bp->rx_bufs[i] = skb = dev_alloc_skb(RX_BUFLEN+2); | 
|  | 636 | if (skb != NULL) | 
|  | 637 | skb_reserve(skb, 2); | 
|  | 638 | } | 
|  | 639 | bmac_construct_rxbuff(skb, &bp->rx_cmds[i]); | 
|  | 640 | } | 
|  | 641 |  | 
|  | 642 | bp->rx_empty = 0; | 
|  | 643 | bp->rx_fill = i; | 
|  | 644 |  | 
|  | 645 | /* Put a branch back to the beginning of the receive command list */ | 
|  | 646 | dbdma_setcmd(&bp->rx_cmds[N_RX_RING], | 
|  | 647 | (DBDMA_NOP | BR_ALWAYS), 0, 0, virt_to_bus(bp->rx_cmds)); | 
|  | 648 |  | 
|  | 649 | /* start rx dma */ | 
|  | 650 | dbdma_reset(rd); | 
|  | 651 | out_le32(&rd->cmdptr, virt_to_bus(bp->rx_cmds)); | 
|  | 652 |  | 
|  | 653 | return 1; | 
|  | 654 | } | 
|  | 655 |  | 
|  | 656 |  | 
|  | 657 | static int bmac_transmit_packet(struct sk_buff *skb, struct net_device *dev) | 
|  | 658 | { | 
|  | 659 | struct bmac_data *bp = netdev_priv(dev); | 
|  | 660 | volatile struct dbdma_regs __iomem *td = bp->tx_dma; | 
|  | 661 | int i; | 
|  | 662 |  | 
|  | 663 | /* see if there's a free slot in the tx ring */ | 
|  | 664 | /* XXDEBUG(("bmac_xmit_start: empty=%d fill=%d\n", */ | 
|  | 665 | /* 	     bp->tx_empty, bp->tx_fill)); */ | 
|  | 666 | i = bp->tx_fill + 1; | 
|  | 667 | if (i >= N_TX_RING) | 
|  | 668 | i = 0; | 
|  | 669 | if (i == bp->tx_empty) { | 
|  | 670 | netif_stop_queue(dev); | 
|  | 671 | bp->tx_fullup = 1; | 
|  | 672 | XXDEBUG(("bmac_transmit_packet: tx ring full\n")); | 
|  | 673 | return -1;		/* can't take it at the moment */ | 
|  | 674 | } | 
|  | 675 |  | 
|  | 676 | dbdma_setcmd(&bp->tx_cmds[i], DBDMA_STOP, 0, 0, 0); | 
|  | 677 |  | 
|  | 678 | bmac_construct_xmt(skb, &bp->tx_cmds[bp->tx_fill]); | 
|  | 679 |  | 
|  | 680 | bp->tx_bufs[bp->tx_fill] = skb; | 
|  | 681 | bp->tx_fill = i; | 
|  | 682 |  | 
|  | 683 | bp->stats.tx_bytes += skb->len; | 
|  | 684 |  | 
|  | 685 | dbdma_continue(td); | 
|  | 686 |  | 
|  | 687 | return 0; | 
|  | 688 | } | 
|  | 689 |  | 
|  | 690 | static int rxintcount; | 
|  | 691 |  | 
|  | 692 | static irqreturn_t bmac_rxdma_intr(int irq, void *dev_id, struct pt_regs *regs) | 
|  | 693 | { | 
|  | 694 | struct net_device *dev = (struct net_device *) dev_id; | 
|  | 695 | struct bmac_data *bp = netdev_priv(dev); | 
|  | 696 | volatile struct dbdma_regs __iomem *rd = bp->rx_dma; | 
|  | 697 | volatile struct dbdma_cmd *cp; | 
|  | 698 | int i, nb, stat; | 
|  | 699 | struct sk_buff *skb; | 
|  | 700 | unsigned int residual; | 
|  | 701 | int last; | 
|  | 702 | unsigned long flags; | 
|  | 703 |  | 
|  | 704 | spin_lock_irqsave(&bp->lock, flags); | 
|  | 705 |  | 
|  | 706 | if (++rxintcount < 10) { | 
|  | 707 | XXDEBUG(("bmac_rxdma_intr\n")); | 
|  | 708 | } | 
|  | 709 |  | 
|  | 710 | last = -1; | 
|  | 711 | i = bp->rx_empty; | 
|  | 712 |  | 
|  | 713 | while (1) { | 
|  | 714 | cp = &bp->rx_cmds[i]; | 
|  | 715 | stat = ld_le16(&cp->xfer_status); | 
|  | 716 | residual = ld_le16(&cp->res_count); | 
|  | 717 | if ((stat & ACTIVE) == 0) | 
|  | 718 | break; | 
|  | 719 | nb = RX_BUFLEN - residual - 2; | 
|  | 720 | if (nb < (ETHERMINPACKET - ETHERCRC)) { | 
|  | 721 | skb = NULL; | 
|  | 722 | bp->stats.rx_length_errors++; | 
|  | 723 | bp->stats.rx_errors++; | 
|  | 724 | } else { | 
|  | 725 | skb = bp->rx_bufs[i]; | 
|  | 726 | bp->rx_bufs[i] = NULL; | 
|  | 727 | } | 
|  | 728 | if (skb != NULL) { | 
|  | 729 | nb -= ETHERCRC; | 
|  | 730 | skb_put(skb, nb); | 
|  | 731 | skb->dev = dev; | 
|  | 732 | skb->protocol = eth_type_trans(skb, dev); | 
|  | 733 | netif_rx(skb); | 
|  | 734 | dev->last_rx = jiffies; | 
|  | 735 | ++bp->stats.rx_packets; | 
|  | 736 | bp->stats.rx_bytes += nb; | 
|  | 737 | } else { | 
|  | 738 | ++bp->stats.rx_dropped; | 
|  | 739 | } | 
|  | 740 | dev->last_rx = jiffies; | 
|  | 741 | if ((skb = bp->rx_bufs[i]) == NULL) { | 
|  | 742 | bp->rx_bufs[i] = skb = dev_alloc_skb(RX_BUFLEN+2); | 
|  | 743 | if (skb != NULL) | 
|  | 744 | skb_reserve(bp->rx_bufs[i], 2); | 
|  | 745 | } | 
|  | 746 | bmac_construct_rxbuff(skb, &bp->rx_cmds[i]); | 
|  | 747 | st_le16(&cp->res_count, 0); | 
|  | 748 | st_le16(&cp->xfer_status, 0); | 
|  | 749 | last = i; | 
|  | 750 | if (++i >= N_RX_RING) i = 0; | 
|  | 751 | } | 
|  | 752 |  | 
|  | 753 | if (last != -1) { | 
|  | 754 | bp->rx_fill = last; | 
|  | 755 | bp->rx_empty = i; | 
|  | 756 | } | 
|  | 757 |  | 
|  | 758 | dbdma_continue(rd); | 
|  | 759 | spin_unlock_irqrestore(&bp->lock, flags); | 
|  | 760 |  | 
|  | 761 | if (rxintcount < 10) { | 
|  | 762 | XXDEBUG(("bmac_rxdma_intr done\n")); | 
|  | 763 | } | 
|  | 764 | return IRQ_HANDLED; | 
|  | 765 | } | 
|  | 766 |  | 
|  | 767 | static int txintcount; | 
|  | 768 |  | 
|  | 769 | static irqreturn_t bmac_txdma_intr(int irq, void *dev_id, struct pt_regs *regs) | 
|  | 770 | { | 
|  | 771 | struct net_device *dev = (struct net_device *) dev_id; | 
|  | 772 | struct bmac_data *bp = netdev_priv(dev); | 
|  | 773 | volatile struct dbdma_cmd *cp; | 
|  | 774 | int stat; | 
|  | 775 | unsigned long flags; | 
|  | 776 |  | 
|  | 777 | spin_lock_irqsave(&bp->lock, flags); | 
|  | 778 |  | 
|  | 779 | if (txintcount++ < 10) { | 
|  | 780 | XXDEBUG(("bmac_txdma_intr\n")); | 
|  | 781 | } | 
|  | 782 |  | 
|  | 783 | /*     del_timer(&bp->tx_timeout); */ | 
|  | 784 | /*     bp->timeout_active = 0; */ | 
|  | 785 |  | 
|  | 786 | while (1) { | 
|  | 787 | cp = &bp->tx_cmds[bp->tx_empty]; | 
|  | 788 | stat = ld_le16(&cp->xfer_status); | 
|  | 789 | if (txintcount < 10) { | 
|  | 790 | XXDEBUG(("bmac_txdma_xfer_stat=%#0x\n", stat)); | 
|  | 791 | } | 
|  | 792 | if (!(stat & ACTIVE)) { | 
|  | 793 | /* | 
|  | 794 | * status field might not have been filled by DBDMA | 
|  | 795 | */ | 
|  | 796 | if (cp == bus_to_virt(in_le32(&bp->tx_dma->cmdptr))) | 
|  | 797 | break; | 
|  | 798 | } | 
|  | 799 |  | 
|  | 800 | if (bp->tx_bufs[bp->tx_empty]) { | 
|  | 801 | ++bp->stats.tx_packets; | 
|  | 802 | dev_kfree_skb_irq(bp->tx_bufs[bp->tx_empty]); | 
|  | 803 | } | 
|  | 804 | bp->tx_bufs[bp->tx_empty] = NULL; | 
|  | 805 | bp->tx_fullup = 0; | 
|  | 806 | netif_wake_queue(dev); | 
|  | 807 | if (++bp->tx_empty >= N_TX_RING) | 
|  | 808 | bp->tx_empty = 0; | 
|  | 809 | if (bp->tx_empty == bp->tx_fill) | 
|  | 810 | break; | 
|  | 811 | } | 
|  | 812 |  | 
|  | 813 | spin_unlock_irqrestore(&bp->lock, flags); | 
|  | 814 |  | 
|  | 815 | if (txintcount < 10) { | 
|  | 816 | XXDEBUG(("bmac_txdma_intr done->bmac_start\n")); | 
|  | 817 | } | 
|  | 818 |  | 
|  | 819 | bmac_start(dev); | 
|  | 820 | return IRQ_HANDLED; | 
|  | 821 | } | 
|  | 822 |  | 
|  | 823 | static struct net_device_stats *bmac_stats(struct net_device *dev) | 
|  | 824 | { | 
|  | 825 | struct bmac_data *p = netdev_priv(dev); | 
|  | 826 |  | 
|  | 827 | return &p->stats; | 
|  | 828 | } | 
|  | 829 |  | 
|  | 830 | #ifndef SUNHME_MULTICAST | 
|  | 831 | /* Real fast bit-reversal algorithm, 6-bit values */ | 
|  | 832 | static int reverse6[64] = { | 
|  | 833 | 0x0,0x20,0x10,0x30,0x8,0x28,0x18,0x38, | 
|  | 834 | 0x4,0x24,0x14,0x34,0xc,0x2c,0x1c,0x3c, | 
|  | 835 | 0x2,0x22,0x12,0x32,0xa,0x2a,0x1a,0x3a, | 
|  | 836 | 0x6,0x26,0x16,0x36,0xe,0x2e,0x1e,0x3e, | 
|  | 837 | 0x1,0x21,0x11,0x31,0x9,0x29,0x19,0x39, | 
|  | 838 | 0x5,0x25,0x15,0x35,0xd,0x2d,0x1d,0x3d, | 
|  | 839 | 0x3,0x23,0x13,0x33,0xb,0x2b,0x1b,0x3b, | 
|  | 840 | 0x7,0x27,0x17,0x37,0xf,0x2f,0x1f,0x3f | 
|  | 841 | }; | 
|  | 842 |  | 
|  | 843 | static unsigned int | 
|  | 844 | crc416(unsigned int curval, unsigned short nxtval) | 
|  | 845 | { | 
|  | 846 | register unsigned int counter, cur = curval, next = nxtval; | 
|  | 847 | register int high_crc_set, low_data_set; | 
|  | 848 |  | 
|  | 849 | /* Swap bytes */ | 
|  | 850 | next = ((next & 0x00FF) << 8) | (next >> 8); | 
|  | 851 |  | 
|  | 852 | /* Compute bit-by-bit */ | 
|  | 853 | for (counter = 0; counter < 16; ++counter) { | 
|  | 854 | /* is high CRC bit set? */ | 
|  | 855 | if ((cur & 0x80000000) == 0) high_crc_set = 0; | 
|  | 856 | else high_crc_set = 1; | 
|  | 857 |  | 
|  | 858 | cur = cur << 1; | 
|  | 859 |  | 
|  | 860 | if ((next & 0x0001) == 0) low_data_set = 0; | 
|  | 861 | else low_data_set = 1; | 
|  | 862 |  | 
|  | 863 | next = next >> 1; | 
|  | 864 |  | 
|  | 865 | /* do the XOR */ | 
|  | 866 | if (high_crc_set ^ low_data_set) cur = cur ^ ENET_CRCPOLY; | 
|  | 867 | } | 
|  | 868 | return cur; | 
|  | 869 | } | 
|  | 870 |  | 
|  | 871 | static unsigned int | 
|  | 872 | bmac_crc(unsigned short *address) | 
|  | 873 | { | 
|  | 874 | unsigned int newcrc; | 
|  | 875 |  | 
|  | 876 | XXDEBUG(("bmac_crc: addr=%#04x, %#04x, %#04x\n", *address, address[1], address[2])); | 
|  | 877 | newcrc = crc416(0xffffffff, *address);	/* address bits 47 - 32 */ | 
|  | 878 | newcrc = crc416(newcrc, address[1]);	/* address bits 31 - 16 */ | 
|  | 879 | newcrc = crc416(newcrc, address[2]);	/* address bits 15 - 0  */ | 
|  | 880 |  | 
|  | 881 | return(newcrc); | 
|  | 882 | } | 
|  | 883 |  | 
|  | 884 | /* | 
|  | 885 | * Add requested mcast addr to BMac's hash table filter. | 
|  | 886 | * | 
|  | 887 | */ | 
|  | 888 |  | 
|  | 889 | static void | 
|  | 890 | bmac_addhash(struct bmac_data *bp, unsigned char *addr) | 
|  | 891 | { | 
|  | 892 | unsigned int	 crc; | 
|  | 893 | unsigned short	 mask; | 
|  | 894 |  | 
|  | 895 | if (!(*addr)) return; | 
|  | 896 | crc = bmac_crc((unsigned short *)addr) & 0x3f; /* Big-endian alert! */ | 
|  | 897 | crc = reverse6[crc];	/* Hyperfast bit-reversing algorithm */ | 
|  | 898 | if (bp->hash_use_count[crc]++) return; /* This bit is already set */ | 
|  | 899 | mask = crc % 16; | 
|  | 900 | mask = (unsigned char)1 << mask; | 
|  | 901 | bp->hash_use_count[crc/16] |= mask; | 
|  | 902 | } | 
|  | 903 |  | 
|  | 904 | static void | 
|  | 905 | bmac_removehash(struct bmac_data *bp, unsigned char *addr) | 
|  | 906 | { | 
|  | 907 | unsigned int crc; | 
|  | 908 | unsigned char mask; | 
|  | 909 |  | 
|  | 910 | /* Now, delete the address from the filter copy, as indicated */ | 
|  | 911 | crc = bmac_crc((unsigned short *)addr) & 0x3f; /* Big-endian alert! */ | 
|  | 912 | crc = reverse6[crc];	/* Hyperfast bit-reversing algorithm */ | 
|  | 913 | if (bp->hash_use_count[crc] == 0) return; /* That bit wasn't in use! */ | 
|  | 914 | if (--bp->hash_use_count[crc]) return; /* That bit is still in use */ | 
|  | 915 | mask = crc % 16; | 
|  | 916 | mask = ((unsigned char)1 << mask) ^ 0xffff; /* To turn off bit */ | 
|  | 917 | bp->hash_table_mask[crc/16] &= mask; | 
|  | 918 | } | 
|  | 919 |  | 
|  | 920 | /* | 
|  | 921 | * Sync the adapter with the software copy of the multicast mask | 
|  | 922 | *  (logical address filter). | 
|  | 923 | */ | 
|  | 924 |  | 
|  | 925 | static void | 
|  | 926 | bmac_rx_off(struct net_device *dev) | 
|  | 927 | { | 
|  | 928 | unsigned short rx_cfg; | 
|  | 929 |  | 
|  | 930 | rx_cfg = bmread(dev, RXCFG); | 
|  | 931 | rx_cfg &= ~RxMACEnable; | 
|  | 932 | bmwrite(dev, RXCFG, rx_cfg); | 
|  | 933 | do { | 
|  | 934 | rx_cfg = bmread(dev, RXCFG); | 
|  | 935 | }  while (rx_cfg & RxMACEnable); | 
|  | 936 | } | 
|  | 937 |  | 
|  | 938 | unsigned short | 
|  | 939 | bmac_rx_on(struct net_device *dev, int hash_enable, int promisc_enable) | 
|  | 940 | { | 
|  | 941 | unsigned short rx_cfg; | 
|  | 942 |  | 
|  | 943 | rx_cfg = bmread(dev, RXCFG); | 
|  | 944 | rx_cfg |= RxMACEnable; | 
|  | 945 | if (hash_enable) rx_cfg |= RxHashFilterEnable; | 
|  | 946 | else rx_cfg &= ~RxHashFilterEnable; | 
|  | 947 | if (promisc_enable) rx_cfg |= RxPromiscEnable; | 
|  | 948 | else rx_cfg &= ~RxPromiscEnable; | 
|  | 949 | bmwrite(dev, RXRST, RxResetValue); | 
|  | 950 | bmwrite(dev, RXFIFOCSR, 0);	/* first disable rxFIFO */ | 
|  | 951 | bmwrite(dev, RXFIFOCSR, RxFIFOEnable ); | 
|  | 952 | bmwrite(dev, RXCFG, rx_cfg ); | 
|  | 953 | return rx_cfg; | 
|  | 954 | } | 
|  | 955 |  | 
|  | 956 | static void | 
|  | 957 | bmac_update_hash_table_mask(struct net_device *dev, struct bmac_data *bp) | 
|  | 958 | { | 
|  | 959 | bmwrite(dev, BHASH3, bp->hash_table_mask[0]); /* bits 15 - 0 */ | 
|  | 960 | bmwrite(dev, BHASH2, bp->hash_table_mask[1]); /* bits 31 - 16 */ | 
|  | 961 | bmwrite(dev, BHASH1, bp->hash_table_mask[2]); /* bits 47 - 32 */ | 
|  | 962 | bmwrite(dev, BHASH0, bp->hash_table_mask[3]); /* bits 63 - 48 */ | 
|  | 963 | } | 
|  | 964 |  | 
|  | 965 | #if 0 | 
|  | 966 | static void | 
|  | 967 | bmac_add_multi(struct net_device *dev, | 
|  | 968 | struct bmac_data *bp, unsigned char *addr) | 
|  | 969 | { | 
|  | 970 | /* XXDEBUG(("bmac: enter bmac_add_multi\n")); */ | 
|  | 971 | bmac_addhash(bp, addr); | 
|  | 972 | bmac_rx_off(dev); | 
|  | 973 | bmac_update_hash_table_mask(dev, bp); | 
|  | 974 | bmac_rx_on(dev, 1, (dev->flags & IFF_PROMISC)? 1 : 0); | 
|  | 975 | /* XXDEBUG(("bmac: exit bmac_add_multi\n")); */ | 
|  | 976 | } | 
|  | 977 |  | 
|  | 978 | static void | 
|  | 979 | bmac_remove_multi(struct net_device *dev, | 
|  | 980 | struct bmac_data *bp, unsigned char *addr) | 
|  | 981 | { | 
|  | 982 | bmac_removehash(bp, addr); | 
|  | 983 | bmac_rx_off(dev); | 
|  | 984 | bmac_update_hash_table_mask(dev, bp); | 
|  | 985 | bmac_rx_on(dev, 1, (dev->flags & IFF_PROMISC)? 1 : 0); | 
|  | 986 | } | 
|  | 987 | #endif | 
|  | 988 |  | 
|  | 989 | /* Set or clear the multicast filter for this adaptor. | 
|  | 990 | num_addrs == -1	Promiscuous mode, receive all packets | 
|  | 991 | num_addrs == 0	Normal mode, clear multicast list | 
|  | 992 | num_addrs > 0	Multicast mode, receive normal and MC packets, and do | 
|  | 993 | best-effort filtering. | 
|  | 994 | */ | 
|  | 995 | static void bmac_set_multicast(struct net_device *dev) | 
|  | 996 | { | 
|  | 997 | struct dev_mc_list *dmi; | 
|  | 998 | struct bmac_data *bp = netdev_priv(dev); | 
|  | 999 | int num_addrs = dev->mc_count; | 
|  | 1000 | unsigned short rx_cfg; | 
|  | 1001 | int i; | 
|  | 1002 |  | 
|  | 1003 | if (bp->sleeping) | 
|  | 1004 | return; | 
|  | 1005 |  | 
|  | 1006 | XXDEBUG(("bmac: enter bmac_set_multicast, n_addrs=%d\n", num_addrs)); | 
|  | 1007 |  | 
|  | 1008 | if((dev->flags & IFF_ALLMULTI) || (dev->mc_count > 64)) { | 
|  | 1009 | for (i=0; i<4; i++) bp->hash_table_mask[i] = 0xffff; | 
|  | 1010 | bmac_update_hash_table_mask(dev, bp); | 
|  | 1011 | rx_cfg = bmac_rx_on(dev, 1, 0); | 
|  | 1012 | XXDEBUG(("bmac: all multi, rx_cfg=%#08x\n")); | 
|  | 1013 | } else if ((dev->flags & IFF_PROMISC) || (num_addrs < 0)) { | 
|  | 1014 | rx_cfg = bmread(dev, RXCFG); | 
|  | 1015 | rx_cfg |= RxPromiscEnable; | 
|  | 1016 | bmwrite(dev, RXCFG, rx_cfg); | 
|  | 1017 | rx_cfg = bmac_rx_on(dev, 0, 1); | 
|  | 1018 | XXDEBUG(("bmac: promisc mode enabled, rx_cfg=%#08x\n", rx_cfg)); | 
|  | 1019 | } else { | 
|  | 1020 | for (i=0; i<4; i++) bp->hash_table_mask[i] = 0; | 
|  | 1021 | for (i=0; i<64; i++) bp->hash_use_count[i] = 0; | 
|  | 1022 | if (num_addrs == 0) { | 
|  | 1023 | rx_cfg = bmac_rx_on(dev, 0, 0); | 
|  | 1024 | XXDEBUG(("bmac: multi disabled, rx_cfg=%#08x\n", rx_cfg)); | 
|  | 1025 | } else { | 
|  | 1026 | for (dmi=dev->mc_list; dmi!=NULL; dmi=dmi->next) | 
|  | 1027 | bmac_addhash(bp, dmi->dmi_addr); | 
|  | 1028 | bmac_update_hash_table_mask(dev, bp); | 
|  | 1029 | rx_cfg = bmac_rx_on(dev, 1, 0); | 
|  | 1030 | XXDEBUG(("bmac: multi enabled, rx_cfg=%#08x\n", rx_cfg)); | 
|  | 1031 | } | 
|  | 1032 | } | 
|  | 1033 | /* XXDEBUG(("bmac: exit bmac_set_multicast\n")); */ | 
|  | 1034 | } | 
|  | 1035 | #else /* ifdef SUNHME_MULTICAST */ | 
|  | 1036 |  | 
|  | 1037 | /* The version of set_multicast below was lifted from sunhme.c */ | 
|  | 1038 |  | 
|  | 1039 | static void bmac_set_multicast(struct net_device *dev) | 
|  | 1040 | { | 
|  | 1041 | struct dev_mc_list *dmi = dev->mc_list; | 
|  | 1042 | char *addrs; | 
|  | 1043 | int i; | 
|  | 1044 | unsigned short rx_cfg; | 
|  | 1045 | u32 crc; | 
|  | 1046 |  | 
|  | 1047 | if((dev->flags & IFF_ALLMULTI) || (dev->mc_count > 64)) { | 
|  | 1048 | bmwrite(dev, BHASH0, 0xffff); | 
|  | 1049 | bmwrite(dev, BHASH1, 0xffff); | 
|  | 1050 | bmwrite(dev, BHASH2, 0xffff); | 
|  | 1051 | bmwrite(dev, BHASH3, 0xffff); | 
|  | 1052 | } else if(dev->flags & IFF_PROMISC) { | 
|  | 1053 | rx_cfg = bmread(dev, RXCFG); | 
|  | 1054 | rx_cfg |= RxPromiscEnable; | 
|  | 1055 | bmwrite(dev, RXCFG, rx_cfg); | 
|  | 1056 | } else { | 
|  | 1057 | u16 hash_table[4]; | 
|  | 1058 |  | 
|  | 1059 | rx_cfg = bmread(dev, RXCFG); | 
|  | 1060 | rx_cfg &= ~RxPromiscEnable; | 
|  | 1061 | bmwrite(dev, RXCFG, rx_cfg); | 
|  | 1062 |  | 
|  | 1063 | for(i = 0; i < 4; i++) hash_table[i] = 0; | 
|  | 1064 |  | 
|  | 1065 | for(i = 0; i < dev->mc_count; i++) { | 
|  | 1066 | addrs = dmi->dmi_addr; | 
|  | 1067 | dmi = dmi->next; | 
|  | 1068 |  | 
|  | 1069 | if(!(*addrs & 1)) | 
|  | 1070 | continue; | 
|  | 1071 |  | 
|  | 1072 | crc = ether_crc_le(6, addrs); | 
|  | 1073 | crc >>= 26; | 
|  | 1074 | hash_table[crc >> 4] |= 1 << (crc & 0xf); | 
|  | 1075 | } | 
|  | 1076 | bmwrite(dev, BHASH0, hash_table[0]); | 
|  | 1077 | bmwrite(dev, BHASH1, hash_table[1]); | 
|  | 1078 | bmwrite(dev, BHASH2, hash_table[2]); | 
|  | 1079 | bmwrite(dev, BHASH3, hash_table[3]); | 
|  | 1080 | } | 
|  | 1081 | } | 
|  | 1082 | #endif /* SUNHME_MULTICAST */ | 
|  | 1083 |  | 
|  | 1084 | static int miscintcount; | 
|  | 1085 |  | 
|  | 1086 | static irqreturn_t bmac_misc_intr(int irq, void *dev_id, struct pt_regs *regs) | 
|  | 1087 | { | 
|  | 1088 | struct net_device *dev = (struct net_device *) dev_id; | 
|  | 1089 | struct bmac_data *bp = netdev_priv(dev); | 
|  | 1090 | unsigned int status = bmread(dev, STATUS); | 
|  | 1091 | if (miscintcount++ < 10) { | 
|  | 1092 | XXDEBUG(("bmac_misc_intr\n")); | 
|  | 1093 | } | 
|  | 1094 | /* XXDEBUG(("bmac_misc_intr, status=%#08x\n", status)); */ | 
|  | 1095 | /*     bmac_txdma_intr_inner(irq, dev_id, regs); */ | 
|  | 1096 | /*   if (status & FrameReceived) bp->stats.rx_dropped++; */ | 
|  | 1097 | if (status & RxErrorMask) bp->stats.rx_errors++; | 
|  | 1098 | if (status & RxCRCCntExp) bp->stats.rx_crc_errors++; | 
|  | 1099 | if (status & RxLenCntExp) bp->stats.rx_length_errors++; | 
|  | 1100 | if (status & RxOverFlow) bp->stats.rx_over_errors++; | 
|  | 1101 | if (status & RxAlignCntExp) bp->stats.rx_frame_errors++; | 
|  | 1102 |  | 
|  | 1103 | /*   if (status & FrameSent) bp->stats.tx_dropped++; */ | 
|  | 1104 | if (status & TxErrorMask) bp->stats.tx_errors++; | 
|  | 1105 | if (status & TxUnderrun) bp->stats.tx_fifo_errors++; | 
|  | 1106 | if (status & TxNormalCollExp) bp->stats.collisions++; | 
|  | 1107 | return IRQ_HANDLED; | 
|  | 1108 | } | 
|  | 1109 |  | 
|  | 1110 | /* | 
|  | 1111 | * Procedure for reading EEPROM | 
|  | 1112 | */ | 
|  | 1113 | #define SROMAddressLength	5 | 
|  | 1114 | #define DataInOn		0x0008 | 
|  | 1115 | #define DataInOff		0x0000 | 
|  | 1116 | #define Clk			0x0002 | 
|  | 1117 | #define ChipSelect		0x0001 | 
|  | 1118 | #define SDIShiftCount		3 | 
|  | 1119 | #define SD0ShiftCount		2 | 
|  | 1120 | #define	DelayValue		1000	/* number of microseconds */ | 
|  | 1121 | #define SROMStartOffset		10	/* this is in words */ | 
|  | 1122 | #define SROMReadCount		3	/* number of words to read from SROM */ | 
|  | 1123 | #define SROMAddressBits		6 | 
|  | 1124 | #define EnetAddressOffset	20 | 
|  | 1125 |  | 
|  | 1126 | static unsigned char | 
|  | 1127 | bmac_clock_out_bit(struct net_device *dev) | 
|  | 1128 | { | 
|  | 1129 | unsigned short         data; | 
|  | 1130 | unsigned short         val; | 
|  | 1131 |  | 
|  | 1132 | bmwrite(dev, SROMCSR, ChipSelect | Clk); | 
|  | 1133 | udelay(DelayValue); | 
|  | 1134 |  | 
|  | 1135 | data = bmread(dev, SROMCSR); | 
|  | 1136 | udelay(DelayValue); | 
|  | 1137 | val = (data >> SD0ShiftCount) & 1; | 
|  | 1138 |  | 
|  | 1139 | bmwrite(dev, SROMCSR, ChipSelect); | 
|  | 1140 | udelay(DelayValue); | 
|  | 1141 |  | 
|  | 1142 | return val; | 
|  | 1143 | } | 
|  | 1144 |  | 
|  | 1145 | static void | 
|  | 1146 | bmac_clock_in_bit(struct net_device *dev, unsigned int val) | 
|  | 1147 | { | 
|  | 1148 | unsigned short data; | 
|  | 1149 |  | 
|  | 1150 | if (val != 0 && val != 1) return; | 
|  | 1151 |  | 
|  | 1152 | data = (val << SDIShiftCount); | 
|  | 1153 | bmwrite(dev, SROMCSR, data | ChipSelect  ); | 
|  | 1154 | udelay(DelayValue); | 
|  | 1155 |  | 
|  | 1156 | bmwrite(dev, SROMCSR, data | ChipSelect | Clk ); | 
|  | 1157 | udelay(DelayValue); | 
|  | 1158 |  | 
|  | 1159 | bmwrite(dev, SROMCSR, data | ChipSelect); | 
|  | 1160 | udelay(DelayValue); | 
|  | 1161 | } | 
|  | 1162 |  | 
|  | 1163 | static void | 
|  | 1164 | reset_and_select_srom(struct net_device *dev) | 
|  | 1165 | { | 
|  | 1166 | /* first reset */ | 
|  | 1167 | bmwrite(dev, SROMCSR, 0); | 
|  | 1168 | udelay(DelayValue); | 
|  | 1169 |  | 
|  | 1170 | /* send it the read command (110) */ | 
|  | 1171 | bmac_clock_in_bit(dev, 1); | 
|  | 1172 | bmac_clock_in_bit(dev, 1); | 
|  | 1173 | bmac_clock_in_bit(dev, 0); | 
|  | 1174 | } | 
|  | 1175 |  | 
|  | 1176 | static unsigned short | 
|  | 1177 | read_srom(struct net_device *dev, unsigned int addr, unsigned int addr_len) | 
|  | 1178 | { | 
|  | 1179 | unsigned short data, val; | 
|  | 1180 | int i; | 
|  | 1181 |  | 
|  | 1182 | /* send out the address we want to read from */ | 
|  | 1183 | for (i = 0; i < addr_len; i++)	{ | 
|  | 1184 | val = addr >> (addr_len-i-1); | 
|  | 1185 | bmac_clock_in_bit(dev, val & 1); | 
|  | 1186 | } | 
|  | 1187 |  | 
|  | 1188 | /* Now read in the 16-bit data */ | 
|  | 1189 | data = 0; | 
|  | 1190 | for (i = 0; i < 16; i++)	{ | 
|  | 1191 | val = bmac_clock_out_bit(dev); | 
|  | 1192 | data <<= 1; | 
|  | 1193 | data |= val; | 
|  | 1194 | } | 
|  | 1195 | bmwrite(dev, SROMCSR, 0); | 
|  | 1196 |  | 
|  | 1197 | return data; | 
|  | 1198 | } | 
|  | 1199 |  | 
|  | 1200 | /* | 
|  | 1201 | * It looks like Cogent and SMC use different methods for calculating | 
|  | 1202 | * checksums. What a pain.. | 
|  | 1203 | */ | 
|  | 1204 |  | 
|  | 1205 | static int | 
|  | 1206 | bmac_verify_checksum(struct net_device *dev) | 
|  | 1207 | { | 
|  | 1208 | unsigned short data, storedCS; | 
|  | 1209 |  | 
|  | 1210 | reset_and_select_srom(dev); | 
|  | 1211 | data = read_srom(dev, 3, SROMAddressBits); | 
|  | 1212 | storedCS = ((data >> 8) & 0x0ff) | ((data << 8) & 0xff00); | 
|  | 1213 |  | 
|  | 1214 | return 0; | 
|  | 1215 | } | 
|  | 1216 |  | 
|  | 1217 |  | 
|  | 1218 | static void | 
|  | 1219 | bmac_get_station_address(struct net_device *dev, unsigned char *ea) | 
|  | 1220 | { | 
|  | 1221 | int i; | 
|  | 1222 | unsigned short data; | 
|  | 1223 |  | 
|  | 1224 | for (i = 0; i < 6; i++) | 
|  | 1225 | { | 
|  | 1226 | reset_and_select_srom(dev); | 
|  | 1227 | data = read_srom(dev, i + EnetAddressOffset/2, SROMAddressBits); | 
|  | 1228 | ea[2*i]   = bitrev(data & 0x0ff); | 
|  | 1229 | ea[2*i+1] = bitrev((data >> 8) & 0x0ff); | 
|  | 1230 | } | 
|  | 1231 | } | 
|  | 1232 |  | 
|  | 1233 | static void bmac_reset_and_enable(struct net_device *dev) | 
|  | 1234 | { | 
|  | 1235 | struct bmac_data *bp = netdev_priv(dev); | 
|  | 1236 | unsigned long flags; | 
|  | 1237 | struct sk_buff *skb; | 
|  | 1238 | unsigned char *data; | 
|  | 1239 |  | 
|  | 1240 | spin_lock_irqsave(&bp->lock, flags); | 
|  | 1241 | bmac_enable_and_reset_chip(dev); | 
|  | 1242 | bmac_init_tx_ring(bp); | 
|  | 1243 | bmac_init_rx_ring(bp); | 
|  | 1244 | bmac_init_chip(dev); | 
|  | 1245 | bmac_start_chip(dev); | 
|  | 1246 | bmwrite(dev, INTDISABLE, EnableNormal); | 
|  | 1247 | bp->sleeping = 0; | 
|  | 1248 |  | 
|  | 1249 | /* | 
|  | 1250 | * It seems that the bmac can't receive until it's transmitted | 
|  | 1251 | * a packet.  So we give it a dummy packet to transmit. | 
|  | 1252 | */ | 
|  | 1253 | skb = dev_alloc_skb(ETHERMINPACKET); | 
|  | 1254 | if (skb != NULL) { | 
|  | 1255 | data = skb_put(skb, ETHERMINPACKET); | 
|  | 1256 | memset(data, 0, ETHERMINPACKET); | 
|  | 1257 | memcpy(data, dev->dev_addr, 6); | 
|  | 1258 | memcpy(data+6, dev->dev_addr, 6); | 
|  | 1259 | bmac_transmit_packet(skb, dev); | 
|  | 1260 | } | 
|  | 1261 | spin_unlock_irqrestore(&bp->lock, flags); | 
|  | 1262 | } | 
|  | 1263 |  | 
| Jeff Mahoney | 5e65577 | 2005-07-06 15:44:41 -0400 | [diff] [blame] | 1264 | static int __devinit bmac_probe(struct macio_dev *mdev, const struct of_device_id *match) | 
| Linus Torvalds | 1da177e | 2005-04-16 15:20:36 -0700 | [diff] [blame] | 1265 | { | 
|  | 1266 | int j, rev, ret; | 
|  | 1267 | struct bmac_data *bp; | 
|  | 1268 | unsigned char *addr; | 
|  | 1269 | struct net_device *dev; | 
|  | 1270 | int is_bmac_plus = ((int)match->data) != 0; | 
|  | 1271 |  | 
|  | 1272 | if (macio_resource_count(mdev) != 3 || macio_irq_count(mdev) != 3) { | 
|  | 1273 | printk(KERN_ERR "BMAC: can't use, need 3 addrs and 3 intrs\n"); | 
|  | 1274 | return -ENODEV; | 
|  | 1275 | } | 
|  | 1276 | addr = get_property(macio_get_of_node(mdev), "mac-address", NULL); | 
|  | 1277 | if (addr == NULL) { | 
|  | 1278 | addr = get_property(macio_get_of_node(mdev), "local-mac-address", NULL); | 
|  | 1279 | if (addr == NULL) { | 
|  | 1280 | printk(KERN_ERR "BMAC: Can't get mac-address\n"); | 
|  | 1281 | return -ENODEV; | 
|  | 1282 | } | 
|  | 1283 | } | 
|  | 1284 |  | 
|  | 1285 | dev = alloc_etherdev(PRIV_BYTES); | 
|  | 1286 | if (!dev) { | 
|  | 1287 | printk(KERN_ERR "BMAC: alloc_etherdev failed, out of memory\n"); | 
|  | 1288 | return -ENOMEM; | 
|  | 1289 | } | 
|  | 1290 |  | 
|  | 1291 | bp = netdev_priv(dev); | 
|  | 1292 | SET_MODULE_OWNER(dev); | 
|  | 1293 | SET_NETDEV_DEV(dev, &mdev->ofdev.dev); | 
|  | 1294 | macio_set_drvdata(mdev, dev); | 
|  | 1295 |  | 
|  | 1296 | bp->mdev = mdev; | 
|  | 1297 | spin_lock_init(&bp->lock); | 
|  | 1298 |  | 
|  | 1299 | if (macio_request_resources(mdev, "bmac")) { | 
|  | 1300 | printk(KERN_ERR "BMAC: can't request IO resource !\n"); | 
|  | 1301 | goto out_free; | 
|  | 1302 | } | 
|  | 1303 |  | 
|  | 1304 | dev->base_addr = (unsigned long) | 
|  | 1305 | ioremap(macio_resource_start(mdev, 0), macio_resource_len(mdev, 0)); | 
|  | 1306 | if (dev->base_addr == 0) | 
|  | 1307 | goto out_release; | 
|  | 1308 |  | 
|  | 1309 | dev->irq = macio_irq(mdev, 0); | 
|  | 1310 |  | 
|  | 1311 | bmac_enable_and_reset_chip(dev); | 
|  | 1312 | bmwrite(dev, INTDISABLE, DisableAll); | 
|  | 1313 |  | 
|  | 1314 | rev = addr[0] == 0 && addr[1] == 0xA0; | 
|  | 1315 | for (j = 0; j < 6; ++j) | 
|  | 1316 | dev->dev_addr[j] = rev? bitrev(addr[j]): addr[j]; | 
|  | 1317 |  | 
|  | 1318 | /* Enable chip without interrupts for now */ | 
|  | 1319 | bmac_enable_and_reset_chip(dev); | 
|  | 1320 | bmwrite(dev, INTDISABLE, DisableAll); | 
|  | 1321 |  | 
|  | 1322 | dev->open = bmac_open; | 
|  | 1323 | dev->stop = bmac_close; | 
|  | 1324 | dev->hard_start_xmit = bmac_output; | 
|  | 1325 | dev->get_stats = bmac_stats; | 
|  | 1326 | dev->set_multicast_list = bmac_set_multicast; | 
|  | 1327 | dev->set_mac_address = bmac_set_address; | 
|  | 1328 |  | 
|  | 1329 | bmac_get_station_address(dev, addr); | 
|  | 1330 | if (bmac_verify_checksum(dev) != 0) | 
|  | 1331 | goto err_out_iounmap; | 
|  | 1332 |  | 
|  | 1333 | bp->is_bmac_plus = is_bmac_plus; | 
|  | 1334 | bp->tx_dma = ioremap(macio_resource_start(mdev, 1), macio_resource_len(mdev, 1)); | 
|  | 1335 | if (!bp->tx_dma) | 
|  | 1336 | goto err_out_iounmap; | 
|  | 1337 | bp->tx_dma_intr = macio_irq(mdev, 1); | 
|  | 1338 | bp->rx_dma = ioremap(macio_resource_start(mdev, 2), macio_resource_len(mdev, 2)); | 
|  | 1339 | if (!bp->rx_dma) | 
|  | 1340 | goto err_out_iounmap_tx; | 
|  | 1341 | bp->rx_dma_intr = macio_irq(mdev, 2); | 
|  | 1342 |  | 
|  | 1343 | bp->tx_cmds = (volatile struct dbdma_cmd *) DBDMA_ALIGN(bp + 1); | 
|  | 1344 | bp->rx_cmds = bp->tx_cmds + N_TX_RING + 1; | 
|  | 1345 |  | 
|  | 1346 | bp->queue = (struct sk_buff_head *)(bp->rx_cmds + N_RX_RING + 1); | 
|  | 1347 | skb_queue_head_init(bp->queue); | 
|  | 1348 |  | 
|  | 1349 | init_timer(&bp->tx_timeout); | 
|  | 1350 |  | 
|  | 1351 | ret = request_irq(dev->irq, bmac_misc_intr, 0, "BMAC-misc", dev); | 
|  | 1352 | if (ret) { | 
|  | 1353 | printk(KERN_ERR "BMAC: can't get irq %d\n", dev->irq); | 
|  | 1354 | goto err_out_iounmap_rx; | 
|  | 1355 | } | 
|  | 1356 | ret = request_irq(bp->tx_dma_intr, bmac_txdma_intr, 0, "BMAC-txdma", dev); | 
|  | 1357 | if (ret) { | 
|  | 1358 | printk(KERN_ERR "BMAC: can't get irq %d\n", bp->tx_dma_intr); | 
|  | 1359 | goto err_out_irq0; | 
|  | 1360 | } | 
|  | 1361 | ret = request_irq(bp->rx_dma_intr, bmac_rxdma_intr, 0, "BMAC-rxdma", dev); | 
|  | 1362 | if (ret) { | 
|  | 1363 | printk(KERN_ERR "BMAC: can't get irq %d\n", bp->rx_dma_intr); | 
|  | 1364 | goto err_out_irq1; | 
|  | 1365 | } | 
|  | 1366 |  | 
|  | 1367 | /* Mask chip interrupts and disable chip, will be | 
|  | 1368 | * re-enabled on open() | 
|  | 1369 | */ | 
|  | 1370 | disable_irq(dev->irq); | 
|  | 1371 | pmac_call_feature(PMAC_FTR_BMAC_ENABLE, macio_get_of_node(bp->mdev), 0, 0); | 
|  | 1372 |  | 
|  | 1373 | if (register_netdev(dev) != 0) { | 
|  | 1374 | printk(KERN_ERR "BMAC: Ethernet registration failed\n"); | 
|  | 1375 | goto err_out_irq2; | 
|  | 1376 | } | 
|  | 1377 |  | 
|  | 1378 | printk(KERN_INFO "%s: BMAC%s at", dev->name, (is_bmac_plus? "+": "")); | 
|  | 1379 | for (j = 0; j < 6; ++j) | 
|  | 1380 | printk("%c%.2x", (j? ':': ' '), dev->dev_addr[j]); | 
|  | 1381 | XXDEBUG((", base_addr=%#0lx", dev->base_addr)); | 
|  | 1382 | printk("\n"); | 
|  | 1383 |  | 
|  | 1384 | return 0; | 
|  | 1385 |  | 
|  | 1386 | err_out_irq2: | 
|  | 1387 | free_irq(bp->rx_dma_intr, dev); | 
|  | 1388 | err_out_irq1: | 
|  | 1389 | free_irq(bp->tx_dma_intr, dev); | 
|  | 1390 | err_out_irq0: | 
|  | 1391 | free_irq(dev->irq, dev); | 
|  | 1392 | err_out_iounmap_rx: | 
|  | 1393 | iounmap(bp->rx_dma); | 
|  | 1394 | err_out_iounmap_tx: | 
|  | 1395 | iounmap(bp->tx_dma); | 
|  | 1396 | err_out_iounmap: | 
|  | 1397 | iounmap((void __iomem *)dev->base_addr); | 
|  | 1398 | out_release: | 
|  | 1399 | macio_release_resources(mdev); | 
|  | 1400 | out_free: | 
|  | 1401 | pmac_call_feature(PMAC_FTR_BMAC_ENABLE, macio_get_of_node(bp->mdev), 0, 0); | 
|  | 1402 | free_netdev(dev); | 
|  | 1403 |  | 
|  | 1404 | return -ENODEV; | 
|  | 1405 | } | 
|  | 1406 |  | 
|  | 1407 | static int bmac_open(struct net_device *dev) | 
|  | 1408 | { | 
|  | 1409 | struct bmac_data *bp = netdev_priv(dev); | 
|  | 1410 | /* XXDEBUG(("bmac: enter open\n")); */ | 
|  | 1411 | /* reset the chip */ | 
|  | 1412 | bp->opened = 1; | 
|  | 1413 | bmac_reset_and_enable(dev); | 
|  | 1414 | enable_irq(dev->irq); | 
| Linus Torvalds | 1da177e | 2005-04-16 15:20:36 -0700 | [diff] [blame] | 1415 | return 0; | 
|  | 1416 | } | 
|  | 1417 |  | 
|  | 1418 | static int bmac_close(struct net_device *dev) | 
|  | 1419 | { | 
|  | 1420 | struct bmac_data *bp = netdev_priv(dev); | 
|  | 1421 | volatile struct dbdma_regs __iomem *rd = bp->rx_dma; | 
|  | 1422 | volatile struct dbdma_regs __iomem *td = bp->tx_dma; | 
|  | 1423 | unsigned short config; | 
|  | 1424 | int i; | 
|  | 1425 |  | 
|  | 1426 | bp->sleeping = 1; | 
| Linus Torvalds | 1da177e | 2005-04-16 15:20:36 -0700 | [diff] [blame] | 1427 |  | 
|  | 1428 | /* disable rx and tx */ | 
|  | 1429 | config = bmread(dev, RXCFG); | 
|  | 1430 | bmwrite(dev, RXCFG, (config & ~RxMACEnable)); | 
|  | 1431 |  | 
|  | 1432 | config = bmread(dev, TXCFG); | 
|  | 1433 | bmwrite(dev, TXCFG, (config & ~TxMACEnable)); | 
|  | 1434 |  | 
|  | 1435 | bmwrite(dev, INTDISABLE, DisableAll); /* disable all intrs */ | 
|  | 1436 |  | 
|  | 1437 | /* disable rx and tx dma */ | 
|  | 1438 | st_le32(&rd->control, DBDMA_CLEAR(RUN|PAUSE|FLUSH|WAKE));	/* clear run bit */ | 
|  | 1439 | st_le32(&td->control, DBDMA_CLEAR(RUN|PAUSE|FLUSH|WAKE));	/* clear run bit */ | 
|  | 1440 |  | 
|  | 1441 | /* free some skb's */ | 
|  | 1442 | XXDEBUG(("bmac: free rx bufs\n")); | 
|  | 1443 | for (i=0; i<N_RX_RING; i++) { | 
|  | 1444 | if (bp->rx_bufs[i] != NULL) { | 
|  | 1445 | dev_kfree_skb(bp->rx_bufs[i]); | 
|  | 1446 | bp->rx_bufs[i] = NULL; | 
|  | 1447 | } | 
|  | 1448 | } | 
|  | 1449 | XXDEBUG(("bmac: free tx bufs\n")); | 
|  | 1450 | for (i = 0; i<N_TX_RING; i++) { | 
|  | 1451 | if (bp->tx_bufs[i] != NULL) { | 
|  | 1452 | dev_kfree_skb(bp->tx_bufs[i]); | 
|  | 1453 | bp->tx_bufs[i] = NULL; | 
|  | 1454 | } | 
|  | 1455 | } | 
|  | 1456 | XXDEBUG(("bmac: all bufs freed\n")); | 
|  | 1457 |  | 
|  | 1458 | bp->opened = 0; | 
|  | 1459 | disable_irq(dev->irq); | 
|  | 1460 | pmac_call_feature(PMAC_FTR_BMAC_ENABLE, macio_get_of_node(bp->mdev), 0, 0); | 
|  | 1461 |  | 
|  | 1462 | return 0; | 
|  | 1463 | } | 
|  | 1464 |  | 
|  | 1465 | static void | 
|  | 1466 | bmac_start(struct net_device *dev) | 
|  | 1467 | { | 
|  | 1468 | struct bmac_data *bp = netdev_priv(dev); | 
|  | 1469 | int i; | 
|  | 1470 | struct sk_buff *skb; | 
|  | 1471 | unsigned long flags; | 
|  | 1472 |  | 
|  | 1473 | if (bp->sleeping) | 
|  | 1474 | return; | 
|  | 1475 |  | 
|  | 1476 | spin_lock_irqsave(&bp->lock, flags); | 
|  | 1477 | while (1) { | 
|  | 1478 | i = bp->tx_fill + 1; | 
|  | 1479 | if (i >= N_TX_RING) | 
|  | 1480 | i = 0; | 
|  | 1481 | if (i == bp->tx_empty) | 
|  | 1482 | break; | 
|  | 1483 | skb = skb_dequeue(bp->queue); | 
|  | 1484 | if (skb == NULL) | 
|  | 1485 | break; | 
|  | 1486 | bmac_transmit_packet(skb, dev); | 
|  | 1487 | } | 
|  | 1488 | spin_unlock_irqrestore(&bp->lock, flags); | 
|  | 1489 | } | 
|  | 1490 |  | 
|  | 1491 | static int | 
|  | 1492 | bmac_output(struct sk_buff *skb, struct net_device *dev) | 
|  | 1493 | { | 
|  | 1494 | struct bmac_data *bp = netdev_priv(dev); | 
|  | 1495 | skb_queue_tail(bp->queue, skb); | 
|  | 1496 | bmac_start(dev); | 
|  | 1497 | return 0; | 
|  | 1498 | } | 
|  | 1499 |  | 
|  | 1500 | static void bmac_tx_timeout(unsigned long data) | 
|  | 1501 | { | 
|  | 1502 | struct net_device *dev = (struct net_device *) data; | 
|  | 1503 | struct bmac_data *bp = netdev_priv(dev); | 
|  | 1504 | volatile struct dbdma_regs __iomem *td = bp->tx_dma; | 
|  | 1505 | volatile struct dbdma_regs __iomem *rd = bp->rx_dma; | 
|  | 1506 | volatile struct dbdma_cmd *cp; | 
|  | 1507 | unsigned long flags; | 
|  | 1508 | unsigned short config, oldConfig; | 
|  | 1509 | int i; | 
|  | 1510 |  | 
|  | 1511 | XXDEBUG(("bmac: tx_timeout called\n")); | 
|  | 1512 | spin_lock_irqsave(&bp->lock, flags); | 
|  | 1513 | bp->timeout_active = 0; | 
|  | 1514 |  | 
|  | 1515 | /* update various counters */ | 
|  | 1516 | /*     	bmac_handle_misc_intrs(bp, 0); */ | 
|  | 1517 |  | 
|  | 1518 | cp = &bp->tx_cmds[bp->tx_empty]; | 
|  | 1519 | /*	XXDEBUG((KERN_DEBUG "bmac: tx dmastat=%x %x runt=%d pr=%x fs=%x fc=%x\n", */ | 
|  | 1520 | /* 	   ld_le32(&td->status), ld_le16(&cp->xfer_status), bp->tx_bad_runt, */ | 
|  | 1521 | /* 	   mb->pr, mb->xmtfs, mb->fifofc)); */ | 
|  | 1522 |  | 
|  | 1523 | /* turn off both tx and rx and reset the chip */ | 
|  | 1524 | config = bmread(dev, RXCFG); | 
|  | 1525 | bmwrite(dev, RXCFG, (config & ~RxMACEnable)); | 
|  | 1526 | config = bmread(dev, TXCFG); | 
|  | 1527 | bmwrite(dev, TXCFG, (config & ~TxMACEnable)); | 
|  | 1528 | out_le32(&td->control, DBDMA_CLEAR(RUN|PAUSE|FLUSH|WAKE|ACTIVE|DEAD)); | 
|  | 1529 | printk(KERN_ERR "bmac: transmit timeout - resetting\n"); | 
|  | 1530 | bmac_enable_and_reset_chip(dev); | 
|  | 1531 |  | 
|  | 1532 | /* restart rx dma */ | 
|  | 1533 | cp = bus_to_virt(ld_le32(&rd->cmdptr)); | 
|  | 1534 | out_le32(&rd->control, DBDMA_CLEAR(RUN|PAUSE|FLUSH|WAKE|ACTIVE|DEAD)); | 
|  | 1535 | out_le16(&cp->xfer_status, 0); | 
|  | 1536 | out_le32(&rd->cmdptr, virt_to_bus(cp)); | 
|  | 1537 | out_le32(&rd->control, DBDMA_SET(RUN|WAKE)); | 
|  | 1538 |  | 
|  | 1539 | /* fix up the transmit side */ | 
|  | 1540 | XXDEBUG((KERN_DEBUG "bmac: tx empty=%d fill=%d fullup=%d\n", | 
|  | 1541 | bp->tx_empty, bp->tx_fill, bp->tx_fullup)); | 
|  | 1542 | i = bp->tx_empty; | 
|  | 1543 | ++bp->stats.tx_errors; | 
|  | 1544 | if (i != bp->tx_fill) { | 
|  | 1545 | dev_kfree_skb(bp->tx_bufs[i]); | 
|  | 1546 | bp->tx_bufs[i] = NULL; | 
|  | 1547 | if (++i >= N_TX_RING) i = 0; | 
|  | 1548 | bp->tx_empty = i; | 
|  | 1549 | } | 
|  | 1550 | bp->tx_fullup = 0; | 
|  | 1551 | netif_wake_queue(dev); | 
|  | 1552 | if (i != bp->tx_fill) { | 
|  | 1553 | cp = &bp->tx_cmds[i]; | 
|  | 1554 | out_le16(&cp->xfer_status, 0); | 
|  | 1555 | out_le16(&cp->command, OUTPUT_LAST); | 
|  | 1556 | out_le32(&td->cmdptr, virt_to_bus(cp)); | 
|  | 1557 | out_le32(&td->control, DBDMA_SET(RUN)); | 
|  | 1558 | /* 	bmac_set_timeout(dev); */ | 
|  | 1559 | XXDEBUG((KERN_DEBUG "bmac: starting %d\n", i)); | 
|  | 1560 | } | 
|  | 1561 |  | 
|  | 1562 | /* turn it back on */ | 
|  | 1563 | oldConfig = bmread(dev, RXCFG); | 
|  | 1564 | bmwrite(dev, RXCFG, oldConfig | RxMACEnable ); | 
|  | 1565 | oldConfig = bmread(dev, TXCFG); | 
|  | 1566 | bmwrite(dev, TXCFG, oldConfig | TxMACEnable ); | 
|  | 1567 |  | 
|  | 1568 | spin_unlock_irqrestore(&bp->lock, flags); | 
|  | 1569 | } | 
|  | 1570 |  | 
|  | 1571 | #if 0 | 
|  | 1572 | static void dump_dbdma(volatile struct dbdma_cmd *cp,int count) | 
|  | 1573 | { | 
|  | 1574 | int i,*ip; | 
|  | 1575 |  | 
|  | 1576 | for (i=0;i< count;i++) { | 
|  | 1577 | ip = (int*)(cp+i); | 
|  | 1578 |  | 
|  | 1579 | printk("dbdma req 0x%x addr 0x%x baddr 0x%x xfer/res 0x%x\n", | 
|  | 1580 | ld_le32(ip+0), | 
|  | 1581 | ld_le32(ip+1), | 
|  | 1582 | ld_le32(ip+2), | 
|  | 1583 | ld_le32(ip+3)); | 
|  | 1584 | } | 
|  | 1585 |  | 
|  | 1586 | } | 
|  | 1587 | #endif | 
|  | 1588 |  | 
|  | 1589 | #if 0 | 
|  | 1590 | static int | 
|  | 1591 | bmac_proc_info(char *buffer, char **start, off_t offset, int length) | 
|  | 1592 | { | 
|  | 1593 | int len = 0; | 
|  | 1594 | off_t pos   = 0; | 
|  | 1595 | off_t begin = 0; | 
|  | 1596 | int i; | 
|  | 1597 |  | 
|  | 1598 | if (bmac_devs == NULL) | 
|  | 1599 | return (-ENOSYS); | 
|  | 1600 |  | 
|  | 1601 | len += sprintf(buffer, "BMAC counters & registers\n"); | 
|  | 1602 |  | 
|  | 1603 | for (i = 0; i<N_REG_ENTRIES; i++) { | 
|  | 1604 | len += sprintf(buffer + len, "%s: %#08x\n", | 
|  | 1605 | reg_entries[i].name, | 
|  | 1606 | bmread(bmac_devs, reg_entries[i].reg_offset)); | 
|  | 1607 | pos = begin + len; | 
|  | 1608 |  | 
|  | 1609 | if (pos < offset) { | 
|  | 1610 | len = 0; | 
|  | 1611 | begin = pos; | 
|  | 1612 | } | 
|  | 1613 |  | 
|  | 1614 | if (pos > offset+length) break; | 
|  | 1615 | } | 
|  | 1616 |  | 
|  | 1617 | *start = buffer + (offset - begin); | 
|  | 1618 | len -= (offset - begin); | 
|  | 1619 |  | 
|  | 1620 | if (len > length) len = length; | 
|  | 1621 |  | 
|  | 1622 | return len; | 
|  | 1623 | } | 
|  | 1624 | #endif | 
|  | 1625 |  | 
|  | 1626 | static int __devexit bmac_remove(struct macio_dev *mdev) | 
|  | 1627 | { | 
|  | 1628 | struct net_device *dev = macio_get_drvdata(mdev); | 
|  | 1629 | struct bmac_data *bp = netdev_priv(dev); | 
|  | 1630 |  | 
|  | 1631 | unregister_netdev(dev); | 
|  | 1632 |  | 
|  | 1633 | free_irq(dev->irq, dev); | 
|  | 1634 | free_irq(bp->tx_dma_intr, dev); | 
|  | 1635 | free_irq(bp->rx_dma_intr, dev); | 
|  | 1636 |  | 
|  | 1637 | iounmap((void __iomem *)dev->base_addr); | 
|  | 1638 | iounmap(bp->tx_dma); | 
|  | 1639 | iounmap(bp->rx_dma); | 
|  | 1640 |  | 
|  | 1641 | macio_release_resources(mdev); | 
|  | 1642 |  | 
|  | 1643 | free_netdev(dev); | 
|  | 1644 |  | 
|  | 1645 | return 0; | 
|  | 1646 | } | 
|  | 1647 |  | 
| Jeff Mahoney | 5e65577 | 2005-07-06 15:44:41 -0400 | [diff] [blame] | 1648 | static struct of_device_id bmac_match[] = | 
| Linus Torvalds | 1da177e | 2005-04-16 15:20:36 -0700 | [diff] [blame] | 1649 | { | 
|  | 1650 | { | 
|  | 1651 | .name 		= "bmac", | 
| Linus Torvalds | 1da177e | 2005-04-16 15:20:36 -0700 | [diff] [blame] | 1652 | .data		= (void *)0, | 
|  | 1653 | }, | 
|  | 1654 | { | 
| Linus Torvalds | 1da177e | 2005-04-16 15:20:36 -0700 | [diff] [blame] | 1655 | .type		= "network", | 
|  | 1656 | .compatible	= "bmac+", | 
|  | 1657 | .data		= (void *)1, | 
|  | 1658 | }, | 
|  | 1659 | {}, | 
|  | 1660 | }; | 
|  | 1661 |  | 
|  | 1662 | static struct macio_driver bmac_driver = | 
|  | 1663 | { | 
|  | 1664 | .name 		= "bmac", | 
|  | 1665 | .match_table	= bmac_match, | 
|  | 1666 | .probe		= bmac_probe, | 
|  | 1667 | .remove		= bmac_remove, | 
|  | 1668 | #ifdef CONFIG_PM | 
|  | 1669 | .suspend	= bmac_suspend, | 
|  | 1670 | .resume		= bmac_resume, | 
|  | 1671 | #endif | 
|  | 1672 | }; | 
|  | 1673 |  | 
|  | 1674 |  | 
|  | 1675 | static int __init bmac_init(void) | 
|  | 1676 | { | 
|  | 1677 | if (bmac_emergency_rxbuf == NULL) { | 
|  | 1678 | bmac_emergency_rxbuf = kmalloc(RX_BUFLEN, GFP_KERNEL); | 
|  | 1679 | if (bmac_emergency_rxbuf == NULL) { | 
|  | 1680 | printk(KERN_ERR "BMAC: can't allocate emergency RX buffer\n"); | 
|  | 1681 | return -ENOMEM; | 
|  | 1682 | } | 
|  | 1683 | } | 
|  | 1684 |  | 
|  | 1685 | return macio_register_driver(&bmac_driver); | 
|  | 1686 | } | 
|  | 1687 |  | 
|  | 1688 | static void __exit bmac_exit(void) | 
|  | 1689 | { | 
|  | 1690 | macio_unregister_driver(&bmac_driver); | 
|  | 1691 |  | 
|  | 1692 | if (bmac_emergency_rxbuf != NULL) { | 
|  | 1693 | kfree(bmac_emergency_rxbuf); | 
|  | 1694 | bmac_emergency_rxbuf = NULL; | 
|  | 1695 | } | 
|  | 1696 | } | 
|  | 1697 |  | 
|  | 1698 | MODULE_AUTHOR("Randy Gobbel/Paul Mackerras"); | 
|  | 1699 | MODULE_DESCRIPTION("PowerMac BMAC ethernet driver."); | 
|  | 1700 | MODULE_LICENSE("GPL"); | 
|  | 1701 |  | 
|  | 1702 | module_init(bmac_init); | 
|  | 1703 | module_exit(bmac_exit); |