| Matt Porter | f89efd5 | 2005-09-09 12:10:10 -0700 | [diff] [blame] | 1 | /* | 
|  | 2 | * rionet - Ethernet driver over RapidIO messaging services | 
|  | 3 | * | 
|  | 4 | * Copyright 2005 MontaVista Software, Inc. | 
|  | 5 | * Matt Porter <mporter@kernel.crashing.org> | 
|  | 6 | * | 
|  | 7 | * This program is free software; you can redistribute  it and/or modify it | 
|  | 8 | * under  the terms of  the GNU General  Public License as published by the | 
|  | 9 | * Free Software Foundation;  either version 2 of the  License, or (at your | 
|  | 10 | * option) any later version. | 
|  | 11 | */ | 
|  | 12 |  | 
|  | 13 | #include <linux/module.h> | 
|  | 14 | #include <linux/kernel.h> | 
|  | 15 | #include <linux/dma-mapping.h> | 
|  | 16 | #include <linux/delay.h> | 
|  | 17 | #include <linux/rio.h> | 
|  | 18 | #include <linux/rio_drv.h> | 
| Tejun Heo | 5a0e3ad | 2010-03-24 17:04:11 +0900 | [diff] [blame] | 19 | #include <linux/slab.h> | 
| Matt Porter | f89efd5 | 2005-09-09 12:10:10 -0700 | [diff] [blame] | 20 | #include <linux/rio_ids.h> | 
|  | 21 |  | 
|  | 22 | #include <linux/netdevice.h> | 
|  | 23 | #include <linux/etherdevice.h> | 
|  | 24 | #include <linux/skbuff.h> | 
|  | 25 | #include <linux/crc32.h> | 
|  | 26 | #include <linux/ethtool.h> | 
|  | 27 |  | 
|  | 28 | #define DRV_NAME        "rionet" | 
|  | 29 | #define DRV_VERSION     "0.2" | 
|  | 30 | #define DRV_AUTHOR      "Matt Porter <mporter@kernel.crashing.org>" | 
|  | 31 | #define DRV_DESC        "Ethernet over RapidIO" | 
|  | 32 |  | 
|  | 33 | MODULE_AUTHOR(DRV_AUTHOR); | 
|  | 34 | MODULE_DESCRIPTION(DRV_DESC); | 
|  | 35 | MODULE_LICENSE("GPL"); | 
|  | 36 |  | 
|  | 37 | #define RIONET_DEFAULT_MSGLEVEL \ | 
|  | 38 | (NETIF_MSG_DRV          | \ | 
|  | 39 | NETIF_MSG_LINK         | \ | 
|  | 40 | NETIF_MSG_RX_ERR       | \ | 
|  | 41 | NETIF_MSG_TX_ERR) | 
|  | 42 |  | 
|  | 43 | #define RIONET_DOORBELL_JOIN	0x1000 | 
|  | 44 | #define RIONET_DOORBELL_LEAVE	0x1001 | 
|  | 45 |  | 
|  | 46 | #define RIONET_MAILBOX		0 | 
|  | 47 |  | 
|  | 48 | #define RIONET_TX_RING_SIZE	CONFIG_RIONET_TX_SIZE | 
|  | 49 | #define RIONET_RX_RING_SIZE	CONFIG_RIONET_RX_SIZE | 
|  | 50 |  | 
|  | 51 | static LIST_HEAD(rionet_peers); | 
|  | 52 |  | 
|  | 53 | struct rionet_private { | 
|  | 54 | struct rio_mport *mport; | 
|  | 55 | struct sk_buff *rx_skb[RIONET_RX_RING_SIZE]; | 
|  | 56 | struct sk_buff *tx_skb[RIONET_TX_RING_SIZE]; | 
| Matt Porter | f89efd5 | 2005-09-09 12:10:10 -0700 | [diff] [blame] | 57 | int rx_slot; | 
|  | 58 | int tx_slot; | 
|  | 59 | int tx_cnt; | 
|  | 60 | int ack_slot; | 
|  | 61 | spinlock_t lock; | 
|  | 62 | spinlock_t tx_lock; | 
|  | 63 | u32 msg_enable; | 
|  | 64 | }; | 
|  | 65 |  | 
|  | 66 | struct rionet_peer { | 
|  | 67 | struct list_head node; | 
|  | 68 | struct rio_dev *rdev; | 
|  | 69 | struct resource *res; | 
|  | 70 | }; | 
|  | 71 |  | 
|  | 72 | static int rionet_check = 0; | 
|  | 73 | static int rionet_capable = 1; | 
|  | 74 |  | 
|  | 75 | /* | 
| Anand Gadiyar | 411c940 | 2009-07-07 15:24:23 +0530 | [diff] [blame] | 76 | * This is a fast lookup table for translating TX | 
| Matt Porter | f89efd5 | 2005-09-09 12:10:10 -0700 | [diff] [blame] | 77 | * Ethernet packets into a destination RIO device. It | 
|  | 78 | * could be made into a hash table to save memory depending | 
|  | 79 | * on system trade-offs. | 
|  | 80 | */ | 
| Zhang Wei | e042323 | 2008-04-18 13:33:42 -0700 | [diff] [blame] | 81 | static struct rio_dev **rionet_active; | 
| Matt Porter | f89efd5 | 2005-09-09 12:10:10 -0700 | [diff] [blame] | 82 |  | 
|  | 83 | #define is_rionet_capable(pef, src_ops, dst_ops)		\ | 
|  | 84 | ((pef & RIO_PEF_INB_MBOX) &&		\ | 
|  | 85 | (pef & RIO_PEF_INB_DOORBELL) &&	\ | 
|  | 86 | (src_ops & RIO_SRC_OPS_DOORBELL) &&	\ | 
|  | 87 | (dst_ops & RIO_DST_OPS_DOORBELL)) | 
|  | 88 | #define dev_rionet_capable(dev) \ | 
|  | 89 | is_rionet_capable(dev->pef, dev->src_ops, dev->dst_ops) | 
|  | 90 |  | 
|  | 91 | #define RIONET_MAC_MATCH(x)	(*(u32 *)x == 0x00010001) | 
|  | 92 | #define RIONET_GET_DESTID(x)	(*(u16 *)(x + 4)) | 
|  | 93 |  | 
| Matt Porter | f89efd5 | 2005-09-09 12:10:10 -0700 | [diff] [blame] | 94 | static int rionet_rx_clean(struct net_device *ndev) | 
|  | 95 | { | 
|  | 96 | int i; | 
|  | 97 | int error = 0; | 
| Wang Chen | 4cf1653 | 2008-11-12 23:38:14 -0800 | [diff] [blame] | 98 | struct rionet_private *rnet = netdev_priv(ndev); | 
| Matt Porter | f89efd5 | 2005-09-09 12:10:10 -0700 | [diff] [blame] | 99 | void *data; | 
|  | 100 |  | 
|  | 101 | i = rnet->rx_slot; | 
|  | 102 |  | 
|  | 103 | do { | 
|  | 104 | if (!rnet->rx_skb[i]) | 
|  | 105 | continue; | 
|  | 106 |  | 
|  | 107 | if (!(data = rio_get_inb_message(rnet->mport, RIONET_MAILBOX))) | 
|  | 108 | break; | 
|  | 109 |  | 
|  | 110 | rnet->rx_skb[i]->data = data; | 
|  | 111 | skb_put(rnet->rx_skb[i], RIO_MAX_MSG_SIZE); | 
| Matt Porter | f89efd5 | 2005-09-09 12:10:10 -0700 | [diff] [blame] | 112 | rnet->rx_skb[i]->protocol = | 
|  | 113 | eth_type_trans(rnet->rx_skb[i], ndev); | 
|  | 114 | error = netif_rx(rnet->rx_skb[i]); | 
|  | 115 |  | 
|  | 116 | if (error == NET_RX_DROP) { | 
| Jeff Garzik | 09f75cd | 2007-10-03 17:41:50 -0700 | [diff] [blame] | 117 | ndev->stats.rx_dropped++; | 
| Matt Porter | f89efd5 | 2005-09-09 12:10:10 -0700 | [diff] [blame] | 118 | } else { | 
| Jeff Garzik | 09f75cd | 2007-10-03 17:41:50 -0700 | [diff] [blame] | 119 | ndev->stats.rx_packets++; | 
|  | 120 | ndev->stats.rx_bytes += RIO_MAX_MSG_SIZE; | 
| Matt Porter | f89efd5 | 2005-09-09 12:10:10 -0700 | [diff] [blame] | 121 | } | 
|  | 122 |  | 
|  | 123 | } while ((i = (i + 1) % RIONET_RX_RING_SIZE) != rnet->rx_slot); | 
|  | 124 |  | 
|  | 125 | return i; | 
|  | 126 | } | 
|  | 127 |  | 
|  | 128 | static void rionet_rx_fill(struct net_device *ndev, int end) | 
|  | 129 | { | 
|  | 130 | int i; | 
| Wang Chen | 4cf1653 | 2008-11-12 23:38:14 -0800 | [diff] [blame] | 131 | struct rionet_private *rnet = netdev_priv(ndev); | 
| Matt Porter | f89efd5 | 2005-09-09 12:10:10 -0700 | [diff] [blame] | 132 |  | 
|  | 133 | i = rnet->rx_slot; | 
|  | 134 | do { | 
|  | 135 | rnet->rx_skb[i] = dev_alloc_skb(RIO_MAX_MSG_SIZE); | 
|  | 136 |  | 
|  | 137 | if (!rnet->rx_skb[i]) | 
|  | 138 | break; | 
|  | 139 |  | 
|  | 140 | rio_add_inb_buffer(rnet->mport, RIONET_MAILBOX, | 
|  | 141 | rnet->rx_skb[i]->data); | 
|  | 142 | } while ((i = (i + 1) % RIONET_RX_RING_SIZE) != end); | 
|  | 143 |  | 
|  | 144 | rnet->rx_slot = i; | 
|  | 145 | } | 
|  | 146 |  | 
|  | 147 | static int rionet_queue_tx_msg(struct sk_buff *skb, struct net_device *ndev, | 
|  | 148 | struct rio_dev *rdev) | 
|  | 149 | { | 
| Wang Chen | 4cf1653 | 2008-11-12 23:38:14 -0800 | [diff] [blame] | 150 | struct rionet_private *rnet = netdev_priv(ndev); | 
| Matt Porter | f89efd5 | 2005-09-09 12:10:10 -0700 | [diff] [blame] | 151 |  | 
|  | 152 | rio_add_outb_message(rnet->mport, rdev, 0, skb->data, skb->len); | 
|  | 153 | rnet->tx_skb[rnet->tx_slot] = skb; | 
|  | 154 |  | 
| Jeff Garzik | 09f75cd | 2007-10-03 17:41:50 -0700 | [diff] [blame] | 155 | ndev->stats.tx_packets++; | 
|  | 156 | ndev->stats.tx_bytes += skb->len; | 
| Matt Porter | f89efd5 | 2005-09-09 12:10:10 -0700 | [diff] [blame] | 157 |  | 
|  | 158 | if (++rnet->tx_cnt == RIONET_TX_RING_SIZE) | 
|  | 159 | netif_stop_queue(ndev); | 
|  | 160 |  | 
|  | 161 | ++rnet->tx_slot; | 
|  | 162 | rnet->tx_slot &= (RIONET_TX_RING_SIZE - 1); | 
|  | 163 |  | 
|  | 164 | if (netif_msg_tx_queued(rnet)) | 
|  | 165 | printk(KERN_INFO "%s: queued skb %8.8x len %8.8x\n", DRV_NAME, | 
|  | 166 | (u32) skb, skb->len); | 
|  | 167 |  | 
|  | 168 | return 0; | 
|  | 169 | } | 
|  | 170 |  | 
|  | 171 | static int rionet_start_xmit(struct sk_buff *skb, struct net_device *ndev) | 
|  | 172 | { | 
|  | 173 | int i; | 
| Wang Chen | 4cf1653 | 2008-11-12 23:38:14 -0800 | [diff] [blame] | 174 | struct rionet_private *rnet = netdev_priv(ndev); | 
| Matt Porter | f89efd5 | 2005-09-09 12:10:10 -0700 | [diff] [blame] | 175 | struct ethhdr *eth = (struct ethhdr *)skb->data; | 
|  | 176 | u16 destid; | 
|  | 177 | unsigned long flags; | 
|  | 178 |  | 
|  | 179 | local_irq_save(flags); | 
|  | 180 | if (!spin_trylock(&rnet->tx_lock)) { | 
|  | 181 | local_irq_restore(flags); | 
|  | 182 | return NETDEV_TX_LOCKED; | 
|  | 183 | } | 
|  | 184 |  | 
|  | 185 | if ((rnet->tx_cnt + 1) > RIONET_TX_RING_SIZE) { | 
|  | 186 | netif_stop_queue(ndev); | 
|  | 187 | spin_unlock_irqrestore(&rnet->tx_lock, flags); | 
|  | 188 | printk(KERN_ERR "%s: BUG! Tx Ring full when queue awake!\n", | 
|  | 189 | ndev->name); | 
|  | 190 | return NETDEV_TX_BUSY; | 
|  | 191 | } | 
|  | 192 |  | 
|  | 193 | if (eth->h_dest[0] & 0x01) { | 
| Zhang Wei | e042323 | 2008-04-18 13:33:42 -0700 | [diff] [blame] | 194 | for (i = 0; i < RIO_MAX_ROUTE_ENTRIES(rnet->mport->sys_size); | 
|  | 195 | i++) | 
| Matt Porter | f89efd5 | 2005-09-09 12:10:10 -0700 | [diff] [blame] | 196 | if (rionet_active[i]) | 
|  | 197 | rionet_queue_tx_msg(skb, ndev, | 
|  | 198 | rionet_active[i]); | 
|  | 199 | } else if (RIONET_MAC_MATCH(eth->h_dest)) { | 
|  | 200 | destid = RIONET_GET_DESTID(eth->h_dest); | 
|  | 201 | if (rionet_active[destid]) | 
|  | 202 | rionet_queue_tx_msg(skb, ndev, rionet_active[destid]); | 
|  | 203 | } | 
|  | 204 |  | 
|  | 205 | spin_unlock_irqrestore(&rnet->tx_lock, flags); | 
|  | 206 |  | 
| Patrick McHardy | 6ed1065 | 2009-06-23 06:03:08 +0000 | [diff] [blame] | 207 | return NETDEV_TX_OK; | 
| Matt Porter | f89efd5 | 2005-09-09 12:10:10 -0700 | [diff] [blame] | 208 | } | 
|  | 209 |  | 
|  | 210 | static void rionet_dbell_event(struct rio_mport *mport, void *dev_id, u16 sid, u16 tid, | 
|  | 211 | u16 info) | 
|  | 212 | { | 
|  | 213 | struct net_device *ndev = dev_id; | 
| Wang Chen | 4cf1653 | 2008-11-12 23:38:14 -0800 | [diff] [blame] | 214 | struct rionet_private *rnet = netdev_priv(ndev); | 
| Matt Porter | f89efd5 | 2005-09-09 12:10:10 -0700 | [diff] [blame] | 215 | struct rionet_peer *peer; | 
|  | 216 |  | 
|  | 217 | if (netif_msg_intr(rnet)) | 
|  | 218 | printk(KERN_INFO "%s: doorbell sid %4.4x tid %4.4x info %4.4x", | 
|  | 219 | DRV_NAME, sid, tid, info); | 
|  | 220 | if (info == RIONET_DOORBELL_JOIN) { | 
|  | 221 | if (!rionet_active[sid]) { | 
|  | 222 | list_for_each_entry(peer, &rionet_peers, node) { | 
|  | 223 | if (peer->rdev->destid == sid) | 
|  | 224 | rionet_active[sid] = peer->rdev; | 
|  | 225 | } | 
|  | 226 | rio_mport_send_doorbell(mport, sid, | 
|  | 227 | RIONET_DOORBELL_JOIN); | 
|  | 228 | } | 
|  | 229 | } else if (info == RIONET_DOORBELL_LEAVE) { | 
|  | 230 | rionet_active[sid] = NULL; | 
|  | 231 | } else { | 
|  | 232 | if (netif_msg_intr(rnet)) | 
|  | 233 | printk(KERN_WARNING "%s: unhandled doorbell\n", | 
|  | 234 | DRV_NAME); | 
|  | 235 | } | 
|  | 236 | } | 
|  | 237 |  | 
|  | 238 | static void rionet_inb_msg_event(struct rio_mport *mport, void *dev_id, int mbox, int slot) | 
|  | 239 | { | 
|  | 240 | int n; | 
|  | 241 | struct net_device *ndev = dev_id; | 
| Wang Chen | 4cf1653 | 2008-11-12 23:38:14 -0800 | [diff] [blame] | 242 | struct rionet_private *rnet = netdev_priv(ndev); | 
| Matt Porter | f89efd5 | 2005-09-09 12:10:10 -0700 | [diff] [blame] | 243 |  | 
|  | 244 | if (netif_msg_intr(rnet)) | 
|  | 245 | printk(KERN_INFO "%s: inbound message event, mbox %d slot %d\n", | 
|  | 246 | DRV_NAME, mbox, slot); | 
|  | 247 |  | 
|  | 248 | spin_lock(&rnet->lock); | 
|  | 249 | if ((n = rionet_rx_clean(ndev)) != rnet->rx_slot) | 
|  | 250 | rionet_rx_fill(ndev, n); | 
|  | 251 | spin_unlock(&rnet->lock); | 
|  | 252 | } | 
|  | 253 |  | 
|  | 254 | static void rionet_outb_msg_event(struct rio_mport *mport, void *dev_id, int mbox, int slot) | 
|  | 255 | { | 
|  | 256 | struct net_device *ndev = dev_id; | 
| Wang Chen | 4cf1653 | 2008-11-12 23:38:14 -0800 | [diff] [blame] | 257 | struct rionet_private *rnet = netdev_priv(ndev); | 
| Matt Porter | f89efd5 | 2005-09-09 12:10:10 -0700 | [diff] [blame] | 258 |  | 
|  | 259 | spin_lock(&rnet->lock); | 
|  | 260 |  | 
|  | 261 | if (netif_msg_intr(rnet)) | 
|  | 262 | printk(KERN_INFO | 
|  | 263 | "%s: outbound message event, mbox %d slot %d\n", | 
|  | 264 | DRV_NAME, mbox, slot); | 
|  | 265 |  | 
|  | 266 | while (rnet->tx_cnt && (rnet->ack_slot != slot)) { | 
|  | 267 | /* dma unmap single */ | 
|  | 268 | dev_kfree_skb_irq(rnet->tx_skb[rnet->ack_slot]); | 
|  | 269 | rnet->tx_skb[rnet->ack_slot] = NULL; | 
|  | 270 | ++rnet->ack_slot; | 
|  | 271 | rnet->ack_slot &= (RIONET_TX_RING_SIZE - 1); | 
|  | 272 | rnet->tx_cnt--; | 
|  | 273 | } | 
|  | 274 |  | 
|  | 275 | if (rnet->tx_cnt < RIONET_TX_RING_SIZE) | 
|  | 276 | netif_wake_queue(ndev); | 
|  | 277 |  | 
|  | 278 | spin_unlock(&rnet->lock); | 
|  | 279 | } | 
|  | 280 |  | 
|  | 281 | static int rionet_open(struct net_device *ndev) | 
|  | 282 | { | 
|  | 283 | int i, rc = 0; | 
|  | 284 | struct rionet_peer *peer, *tmp; | 
|  | 285 | u32 pwdcsr; | 
| Wang Chen | 4cf1653 | 2008-11-12 23:38:14 -0800 | [diff] [blame] | 286 | struct rionet_private *rnet = netdev_priv(ndev); | 
| Matt Porter | f89efd5 | 2005-09-09 12:10:10 -0700 | [diff] [blame] | 287 |  | 
|  | 288 | if (netif_msg_ifup(rnet)) | 
|  | 289 | printk(KERN_INFO "%s: open\n", DRV_NAME); | 
|  | 290 |  | 
|  | 291 | if ((rc = rio_request_inb_dbell(rnet->mport, | 
|  | 292 | (void *)ndev, | 
|  | 293 | RIONET_DOORBELL_JOIN, | 
|  | 294 | RIONET_DOORBELL_LEAVE, | 
|  | 295 | rionet_dbell_event)) < 0) | 
|  | 296 | goto out; | 
|  | 297 |  | 
|  | 298 | if ((rc = rio_request_inb_mbox(rnet->mport, | 
|  | 299 | (void *)ndev, | 
|  | 300 | RIONET_MAILBOX, | 
|  | 301 | RIONET_RX_RING_SIZE, | 
|  | 302 | rionet_inb_msg_event)) < 0) | 
|  | 303 | goto out; | 
|  | 304 |  | 
|  | 305 | if ((rc = rio_request_outb_mbox(rnet->mport, | 
|  | 306 | (void *)ndev, | 
|  | 307 | RIONET_MAILBOX, | 
|  | 308 | RIONET_TX_RING_SIZE, | 
|  | 309 | rionet_outb_msg_event)) < 0) | 
|  | 310 | goto out; | 
|  | 311 |  | 
|  | 312 | /* Initialize inbound message ring */ | 
|  | 313 | for (i = 0; i < RIONET_RX_RING_SIZE; i++) | 
|  | 314 | rnet->rx_skb[i] = NULL; | 
|  | 315 | rnet->rx_slot = 0; | 
|  | 316 | rionet_rx_fill(ndev, 0); | 
|  | 317 |  | 
|  | 318 | rnet->tx_slot = 0; | 
|  | 319 | rnet->tx_cnt = 0; | 
|  | 320 | rnet->ack_slot = 0; | 
|  | 321 |  | 
|  | 322 | netif_carrier_on(ndev); | 
|  | 323 | netif_start_queue(ndev); | 
|  | 324 |  | 
|  | 325 | list_for_each_entry_safe(peer, tmp, &rionet_peers, node) { | 
|  | 326 | if (!(peer->res = rio_request_outb_dbell(peer->rdev, | 
|  | 327 | RIONET_DOORBELL_JOIN, | 
|  | 328 | RIONET_DOORBELL_LEAVE))) | 
|  | 329 | { | 
|  | 330 | printk(KERN_ERR "%s: error requesting doorbells\n", | 
|  | 331 | DRV_NAME); | 
|  | 332 | continue; | 
|  | 333 | } | 
|  | 334 |  | 
|  | 335 | /* | 
|  | 336 | * If device has initialized inbound doorbells, | 
|  | 337 | * send a join message | 
|  | 338 | */ | 
|  | 339 | rio_read_config_32(peer->rdev, RIO_WRITE_PORT_CSR, &pwdcsr); | 
|  | 340 | if (pwdcsr & RIO_DOORBELL_AVAIL) | 
|  | 341 | rio_send_doorbell(peer->rdev, RIONET_DOORBELL_JOIN); | 
|  | 342 | } | 
|  | 343 |  | 
|  | 344 | out: | 
|  | 345 | return rc; | 
|  | 346 | } | 
|  | 347 |  | 
|  | 348 | static int rionet_close(struct net_device *ndev) | 
|  | 349 | { | 
| Wang Chen | 4cf1653 | 2008-11-12 23:38:14 -0800 | [diff] [blame] | 350 | struct rionet_private *rnet = netdev_priv(ndev); | 
| Matt Porter | f89efd5 | 2005-09-09 12:10:10 -0700 | [diff] [blame] | 351 | struct rionet_peer *peer, *tmp; | 
|  | 352 | int i; | 
|  | 353 |  | 
|  | 354 | if (netif_msg_ifup(rnet)) | 
|  | 355 | printk(KERN_INFO "%s: close\n", DRV_NAME); | 
|  | 356 |  | 
|  | 357 | netif_stop_queue(ndev); | 
|  | 358 | netif_carrier_off(ndev); | 
|  | 359 |  | 
|  | 360 | for (i = 0; i < RIONET_RX_RING_SIZE; i++) | 
| Wei Yongjun | aaff1e1 | 2009-02-25 00:18:12 +0000 | [diff] [blame] | 361 | kfree_skb(rnet->rx_skb[i]); | 
| Matt Porter | f89efd5 | 2005-09-09 12:10:10 -0700 | [diff] [blame] | 362 |  | 
|  | 363 | list_for_each_entry_safe(peer, tmp, &rionet_peers, node) { | 
|  | 364 | if (rionet_active[peer->rdev->destid]) { | 
|  | 365 | rio_send_doorbell(peer->rdev, RIONET_DOORBELL_LEAVE); | 
|  | 366 | rionet_active[peer->rdev->destid] = NULL; | 
|  | 367 | } | 
|  | 368 | rio_release_outb_dbell(peer->rdev, peer->res); | 
|  | 369 | } | 
|  | 370 |  | 
|  | 371 | rio_release_inb_dbell(rnet->mport, RIONET_DOORBELL_JOIN, | 
|  | 372 | RIONET_DOORBELL_LEAVE); | 
|  | 373 | rio_release_inb_mbox(rnet->mport, RIONET_MAILBOX); | 
|  | 374 | rio_release_outb_mbox(rnet->mport, RIONET_MAILBOX); | 
|  | 375 |  | 
|  | 376 | return 0; | 
|  | 377 | } | 
|  | 378 |  | 
|  | 379 | static void rionet_remove(struct rio_dev *rdev) | 
|  | 380 | { | 
|  | 381 | struct net_device *ndev = NULL; | 
|  | 382 | struct rionet_peer *peer, *tmp; | 
|  | 383 |  | 
| Zhang Wei | e042323 | 2008-04-18 13:33:42 -0700 | [diff] [blame] | 384 | free_pages((unsigned long)rionet_active, rdev->net->hport->sys_size ? | 
| Alexandre Bounine | 388b78a | 2011-03-23 16:43:03 -0700 | [diff] [blame] | 385 | __fls(sizeof(void *)) + 4 : 0); | 
| Matt Porter | f89efd5 | 2005-09-09 12:10:10 -0700 | [diff] [blame] | 386 | unregister_netdev(ndev); | 
| Kulikov Vasiliy | 22138d3 | 2010-09-25 23:58:03 +0000 | [diff] [blame] | 387 | free_netdev(ndev); | 
| Matt Porter | f89efd5 | 2005-09-09 12:10:10 -0700 | [diff] [blame] | 388 |  | 
|  | 389 | list_for_each_entry_safe(peer, tmp, &rionet_peers, node) { | 
|  | 390 | list_del(&peer->node); | 
|  | 391 | kfree(peer); | 
|  | 392 | } | 
|  | 393 | } | 
|  | 394 |  | 
|  | 395 | static void rionet_get_drvinfo(struct net_device *ndev, | 
|  | 396 | struct ethtool_drvinfo *info) | 
|  | 397 | { | 
| Wang Chen | 4cf1653 | 2008-11-12 23:38:14 -0800 | [diff] [blame] | 398 | struct rionet_private *rnet = netdev_priv(ndev); | 
| Matt Porter | f89efd5 | 2005-09-09 12:10:10 -0700 | [diff] [blame] | 399 |  | 
|  | 400 | strcpy(info->driver, DRV_NAME); | 
|  | 401 | strcpy(info->version, DRV_VERSION); | 
|  | 402 | strcpy(info->fw_version, "n/a"); | 
|  | 403 | strcpy(info->bus_info, rnet->mport->name); | 
|  | 404 | } | 
|  | 405 |  | 
|  | 406 | static u32 rionet_get_msglevel(struct net_device *ndev) | 
|  | 407 | { | 
| Wang Chen | 4cf1653 | 2008-11-12 23:38:14 -0800 | [diff] [blame] | 408 | struct rionet_private *rnet = netdev_priv(ndev); | 
| Matt Porter | f89efd5 | 2005-09-09 12:10:10 -0700 | [diff] [blame] | 409 |  | 
|  | 410 | return rnet->msg_enable; | 
|  | 411 | } | 
|  | 412 |  | 
|  | 413 | static void rionet_set_msglevel(struct net_device *ndev, u32 value) | 
|  | 414 | { | 
| Wang Chen | 4cf1653 | 2008-11-12 23:38:14 -0800 | [diff] [blame] | 415 | struct rionet_private *rnet = netdev_priv(ndev); | 
| Matt Porter | f89efd5 | 2005-09-09 12:10:10 -0700 | [diff] [blame] | 416 |  | 
|  | 417 | rnet->msg_enable = value; | 
|  | 418 | } | 
|  | 419 |  | 
| Jeff Garzik | 7282d49 | 2006-09-13 14:30:00 -0400 | [diff] [blame] | 420 | static const struct ethtool_ops rionet_ethtool_ops = { | 
| Matt Porter | f89efd5 | 2005-09-09 12:10:10 -0700 | [diff] [blame] | 421 | .get_drvinfo = rionet_get_drvinfo, | 
|  | 422 | .get_msglevel = rionet_get_msglevel, | 
|  | 423 | .set_msglevel = rionet_set_msglevel, | 
|  | 424 | .get_link = ethtool_op_get_link, | 
|  | 425 | }; | 
|  | 426 |  | 
| Alexander Beregalov | a33a2bb | 2009-04-15 12:52:56 +0000 | [diff] [blame] | 427 | static const struct net_device_ops rionet_netdev_ops = { | 
|  | 428 | .ndo_open		= rionet_open, | 
|  | 429 | .ndo_stop		= rionet_close, | 
|  | 430 | .ndo_start_xmit		= rionet_start_xmit, | 
|  | 431 | .ndo_change_mtu		= eth_change_mtu, | 
|  | 432 | .ndo_validate_addr	= eth_validate_addr, | 
|  | 433 | .ndo_set_mac_address	= eth_mac_addr, | 
|  | 434 | }; | 
|  | 435 |  | 
| Matt Porter | f89efd5 | 2005-09-09 12:10:10 -0700 | [diff] [blame] | 436 | static int rionet_setup_netdev(struct rio_mport *mport) | 
|  | 437 | { | 
|  | 438 | int rc = 0; | 
|  | 439 | struct net_device *ndev = NULL; | 
|  | 440 | struct rionet_private *rnet; | 
|  | 441 | u16 device_id; | 
|  | 442 |  | 
|  | 443 | /* Allocate our net_device structure */ | 
|  | 444 | ndev = alloc_etherdev(sizeof(struct rionet_private)); | 
|  | 445 | if (ndev == NULL) { | 
|  | 446 | printk(KERN_INFO "%s: could not allocate ethernet device.\n", | 
|  | 447 | DRV_NAME); | 
|  | 448 | rc = -ENOMEM; | 
|  | 449 | goto out; | 
|  | 450 | } | 
|  | 451 |  | 
| Zhang Wei | e042323 | 2008-04-18 13:33:42 -0700 | [diff] [blame] | 452 | rionet_active = (struct rio_dev **)__get_free_pages(GFP_KERNEL, | 
| Alexandre Bounine | 388b78a | 2011-03-23 16:43:03 -0700 | [diff] [blame] | 453 | mport->sys_size ? __fls(sizeof(void *)) + 4 : 0); | 
| Zhang Wei | e042323 | 2008-04-18 13:33:42 -0700 | [diff] [blame] | 454 | if (!rionet_active) { | 
|  | 455 | rc = -ENOMEM; | 
|  | 456 | goto out; | 
|  | 457 | } | 
|  | 458 | memset((void *)rionet_active, 0, sizeof(void *) * | 
|  | 459 | RIO_MAX_ROUTE_ENTRIES(mport->sys_size)); | 
|  | 460 |  | 
| Matt Porter | f89efd5 | 2005-09-09 12:10:10 -0700 | [diff] [blame] | 461 | /* Set up private area */ | 
| Wang Chen | 4cf1653 | 2008-11-12 23:38:14 -0800 | [diff] [blame] | 462 | rnet = netdev_priv(ndev); | 
| Matt Porter | f89efd5 | 2005-09-09 12:10:10 -0700 | [diff] [blame] | 463 | rnet->mport = mport; | 
|  | 464 |  | 
|  | 465 | /* Set the default MAC address */ | 
|  | 466 | device_id = rio_local_get_device_id(mport); | 
|  | 467 | ndev->dev_addr[0] = 0x00; | 
|  | 468 | ndev->dev_addr[1] = 0x01; | 
|  | 469 | ndev->dev_addr[2] = 0x00; | 
|  | 470 | ndev->dev_addr[3] = 0x01; | 
|  | 471 | ndev->dev_addr[4] = device_id >> 8; | 
|  | 472 | ndev->dev_addr[5] = device_id & 0xff; | 
|  | 473 |  | 
| Alexander Beregalov | a33a2bb | 2009-04-15 12:52:56 +0000 | [diff] [blame] | 474 | ndev->netdev_ops = &rionet_netdev_ops; | 
| Matt Porter | f89efd5 | 2005-09-09 12:10:10 -0700 | [diff] [blame] | 475 | ndev->mtu = RIO_MAX_MSG_SIZE - 14; | 
|  | 476 | ndev->features = NETIF_F_LLTX; | 
|  | 477 | SET_ETHTOOL_OPS(ndev, &rionet_ethtool_ops); | 
|  | 478 |  | 
| Matt Porter | f89efd5 | 2005-09-09 12:10:10 -0700 | [diff] [blame] | 479 | spin_lock_init(&rnet->lock); | 
|  | 480 | spin_lock_init(&rnet->tx_lock); | 
|  | 481 |  | 
|  | 482 | rnet->msg_enable = RIONET_DEFAULT_MSGLEVEL; | 
|  | 483 |  | 
|  | 484 | rc = register_netdev(ndev); | 
|  | 485 | if (rc != 0) | 
|  | 486 | goto out; | 
|  | 487 |  | 
| Johannes Berg | e174961 | 2008-10-27 15:59:26 -0700 | [diff] [blame] | 488 | printk("%s: %s %s Version %s, MAC %pM\n", | 
| Matt Porter | f89efd5 | 2005-09-09 12:10:10 -0700 | [diff] [blame] | 489 | ndev->name, | 
|  | 490 | DRV_NAME, | 
|  | 491 | DRV_DESC, | 
|  | 492 | DRV_VERSION, | 
| Johannes Berg | e174961 | 2008-10-27 15:59:26 -0700 | [diff] [blame] | 493 | ndev->dev_addr); | 
| Matt Porter | f89efd5 | 2005-09-09 12:10:10 -0700 | [diff] [blame] | 494 |  | 
|  | 495 | out: | 
|  | 496 | return rc; | 
|  | 497 | } | 
|  | 498 |  | 
|  | 499 | /* | 
|  | 500 | * XXX Make multi-net safe | 
|  | 501 | */ | 
|  | 502 | static int rionet_probe(struct rio_dev *rdev, const struct rio_device_id *id) | 
|  | 503 | { | 
|  | 504 | int rc = -ENODEV; | 
|  | 505 | u32 lpef, lsrc_ops, ldst_ops; | 
|  | 506 | struct rionet_peer *peer; | 
|  | 507 |  | 
|  | 508 | /* If local device is not rionet capable, give up quickly */ | 
|  | 509 | if (!rionet_capable) | 
|  | 510 | goto out; | 
|  | 511 |  | 
|  | 512 | /* | 
|  | 513 | * First time through, make sure local device is rionet | 
|  | 514 | * capable, setup netdev,  and set flags so this is skipped | 
|  | 515 | * on later probes | 
|  | 516 | */ | 
|  | 517 | if (!rionet_check) { | 
|  | 518 | rio_local_read_config_32(rdev->net->hport, RIO_PEF_CAR, &lpef); | 
|  | 519 | rio_local_read_config_32(rdev->net->hport, RIO_SRC_OPS_CAR, | 
|  | 520 | &lsrc_ops); | 
|  | 521 | rio_local_read_config_32(rdev->net->hport, RIO_DST_OPS_CAR, | 
|  | 522 | &ldst_ops); | 
|  | 523 | if (!is_rionet_capable(lpef, lsrc_ops, ldst_ops)) { | 
|  | 524 | printk(KERN_ERR | 
|  | 525 | "%s: local device is not network capable\n", | 
|  | 526 | DRV_NAME); | 
|  | 527 | rionet_check = 1; | 
|  | 528 | rionet_capable = 0; | 
|  | 529 | goto out; | 
|  | 530 | } | 
|  | 531 |  | 
|  | 532 | rc = rionet_setup_netdev(rdev->net->hport); | 
|  | 533 | rionet_check = 1; | 
|  | 534 | } | 
|  | 535 |  | 
|  | 536 | /* | 
|  | 537 | * If the remote device has mailbox/doorbell capabilities, | 
|  | 538 | * add it to the peer list. | 
|  | 539 | */ | 
|  | 540 | if (dev_rionet_capable(rdev)) { | 
|  | 541 | if (!(peer = kmalloc(sizeof(struct rionet_peer), GFP_KERNEL))) { | 
|  | 542 | rc = -ENOMEM; | 
|  | 543 | goto out; | 
|  | 544 | } | 
|  | 545 | peer->rdev = rdev; | 
|  | 546 | list_add_tail(&peer->node, &rionet_peers); | 
|  | 547 | } | 
|  | 548 |  | 
|  | 549 | out: | 
|  | 550 | return rc; | 
|  | 551 | } | 
|  | 552 |  | 
|  | 553 | static struct rio_device_id rionet_id_table[] = { | 
|  | 554 | {RIO_DEVICE(RIO_ANY_ID, RIO_ANY_ID)} | 
|  | 555 | }; | 
|  | 556 |  | 
|  | 557 | static struct rio_driver rionet_driver = { | 
|  | 558 | .name = "rionet", | 
|  | 559 | .id_table = rionet_id_table, | 
|  | 560 | .probe = rionet_probe, | 
|  | 561 | .remove = rionet_remove, | 
|  | 562 | }; | 
|  | 563 |  | 
|  | 564 | static int __init rionet_init(void) | 
|  | 565 | { | 
|  | 566 | return rio_register_driver(&rionet_driver); | 
|  | 567 | } | 
|  | 568 |  | 
|  | 569 | static void __exit rionet_exit(void) | 
|  | 570 | { | 
|  | 571 | rio_unregister_driver(&rionet_driver); | 
|  | 572 | } | 
|  | 573 |  | 
| Alexandre Bounine | 2f80998 | 2011-03-23 16:43:04 -0700 | [diff] [blame] | 574 | late_initcall(rionet_init); | 
| Matt Porter | f89efd5 | 2005-09-09 12:10:10 -0700 | [diff] [blame] | 575 | module_exit(rionet_exit); |