blob: f410050d0fbb4d97930d966e11edf32d9bf123ba [file] [log] [blame]
Chiranjeevi Velempati9179d542013-02-18 15:22:54 +05301/* Copyright (c) 2011-2013, Linux Foundation. All rights reserved.
Bryan Huntsman3f2bc4d2011-08-16 17:27:22 -07002 *
3 * This program is free software; you can redistribute it and/or modify
4 * it under the terms of the GNU General Public License version 2 and
5 * only version 2 as published by the Free Software Foundation.
6 *
7 * This program is distributed in the hope that it will be useful,
8 * but WITHOUT ANY WARRANTY; without even the implied warranty of
9 * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
10 * GNU General Public License for more details.
11 */
12
13#include <linux/kernel.h>
14#include <linux/interrupt.h>
15#include <linux/device.h>
16#include <linux/delay.h>
17#include <linux/slab.h>
18#include <linux/termios.h>
19#include <mach/msm_smd.h>
20#include <linux/netdevice.h>
21#include <mach/bam_dmux.h>
22#include <linux/debugfs.h>
23#include <linux/bitops.h>
24#include <linux/termios.h>
25
Ofir Cohena1c2a872011-12-14 10:26:34 +020026#include <mach/usb_gadget_xport.h>
Shimrit Malichi194fe122012-07-25 13:50:41 +030027#include <linux/usb/msm_hsusb.h>
Ofir Cohena1c2a872011-12-14 10:26:34 +020028#include <mach/usb_bam.h>
29
Bryan Huntsman3f2bc4d2011-08-16 17:27:22 -070030#include "u_rmnet.h"
31
32#define BAM_N_PORTS 1
Anna Perel21515162012-02-02 20:50:02 +020033#define BAM2BAM_N_PORTS 3
Bryan Huntsman3f2bc4d2011-08-16 17:27:22 -070034
35static struct workqueue_struct *gbam_wq;
36static int n_bam_ports;
Ofir Cohena1c2a872011-12-14 10:26:34 +020037static int n_bam2bam_ports;
Vijayavardhan Vennapusa08c31252011-12-21 13:02:49 +053038static unsigned n_tx_req_queued;
Bryan Huntsman3f2bc4d2011-08-16 17:27:22 -070039static unsigned bam_ch_ids[] = { 8 };
40
Jack Phameffd4ae2011-08-03 16:49:36 -070041static const char *bam_ch_names[] = { "bam_dmux_ch_8" };
42
Vamsi Krishna84579552011-11-09 15:33:22 -080043#define BAM_PENDING_LIMIT 220
Vamsi Krishna8f24f252011-11-02 11:46:08 -070044#define BAM_MUX_TX_PKT_DROP_THRESHOLD 1000
Vamsi Krishna84579552011-11-09 15:33:22 -080045#define BAM_MUX_RX_PKT_FCTRL_EN_TSHOLD 500
46#define BAM_MUX_RX_PKT_FCTRL_DIS_TSHOLD 300
Vamsi Krishna8f24f252011-11-02 11:46:08 -070047#define BAM_MUX_RX_PKT_FLOW_CTRL_SUPPORT 1
Bryan Huntsman3f2bc4d2011-08-16 17:27:22 -070048
49#define BAM_MUX_HDR 8
50
Vamsi Krishna8f24f252011-11-02 11:46:08 -070051#define BAM_MUX_RX_Q_SIZE 16
52#define BAM_MUX_TX_Q_SIZE 200
53#define BAM_MUX_RX_REQ_SIZE (2048 - BAM_MUX_HDR)
Bryan Huntsman3f2bc4d2011-08-16 17:27:22 -070054
Vijayavardhan Vennapusa08c31252011-12-21 13:02:49 +053055#define DL_INTR_THRESHOLD 20
56
Vamsi Krishna8f24f252011-11-02 11:46:08 -070057unsigned int bam_mux_tx_pkt_drop_thld = BAM_MUX_TX_PKT_DROP_THRESHOLD;
58module_param(bam_mux_tx_pkt_drop_thld, uint, S_IRUGO | S_IWUSR);
Bryan Huntsman3f2bc4d2011-08-16 17:27:22 -070059
Vamsi Krishna8f24f252011-11-02 11:46:08 -070060unsigned int bam_mux_rx_fctrl_en_thld = BAM_MUX_RX_PKT_FCTRL_EN_TSHOLD;
61module_param(bam_mux_rx_fctrl_en_thld, uint, S_IRUGO | S_IWUSR);
Bryan Huntsman3f2bc4d2011-08-16 17:27:22 -070062
Vamsi Krishna8f24f252011-11-02 11:46:08 -070063unsigned int bam_mux_rx_fctrl_support = BAM_MUX_RX_PKT_FLOW_CTRL_SUPPORT;
64module_param(bam_mux_rx_fctrl_support, uint, S_IRUGO | S_IWUSR);
Bryan Huntsman3f2bc4d2011-08-16 17:27:22 -070065
Vamsi Krishna8f24f252011-11-02 11:46:08 -070066unsigned int bam_mux_rx_fctrl_dis_thld = BAM_MUX_RX_PKT_FCTRL_DIS_TSHOLD;
67module_param(bam_mux_rx_fctrl_dis_thld, uint, S_IRUGO | S_IWUSR);
Bryan Huntsman3f2bc4d2011-08-16 17:27:22 -070068
Vamsi Krishna8f24f252011-11-02 11:46:08 -070069unsigned int bam_mux_tx_q_size = BAM_MUX_TX_Q_SIZE;
70module_param(bam_mux_tx_q_size, uint, S_IRUGO | S_IWUSR);
Bryan Huntsman3f2bc4d2011-08-16 17:27:22 -070071
Vamsi Krishna8f24f252011-11-02 11:46:08 -070072unsigned int bam_mux_rx_q_size = BAM_MUX_RX_Q_SIZE;
73module_param(bam_mux_rx_q_size, uint, S_IRUGO | S_IWUSR);
Bryan Huntsman3f2bc4d2011-08-16 17:27:22 -070074
Vamsi Krishna8f24f252011-11-02 11:46:08 -070075unsigned int bam_mux_rx_req_size = BAM_MUX_RX_REQ_SIZE;
76module_param(bam_mux_rx_req_size, uint, S_IRUGO | S_IWUSR);
Bryan Huntsman3f2bc4d2011-08-16 17:27:22 -070077
Vijayavardhan Vennapusa08c31252011-12-21 13:02:49 +053078unsigned int dl_intr_threshold = DL_INTR_THRESHOLD;
79module_param(dl_intr_threshold, uint, S_IRUGO | S_IWUSR);
80
Jack Phameffd4ae2011-08-03 16:49:36 -070081#define BAM_CH_OPENED BIT(0)
82#define BAM_CH_READY BIT(1)
Ofir Cohena1c2a872011-12-14 10:26:34 +020083
Bryan Huntsman3f2bc4d2011-08-16 17:27:22 -070084struct bam_ch_info {
Jack Phameffd4ae2011-08-03 16:49:36 -070085 unsigned long flags;
Bryan Huntsman3f2bc4d2011-08-16 17:27:22 -070086 unsigned id;
87
88 struct list_head tx_idle;
89 struct sk_buff_head tx_skb_q;
90
91 struct list_head rx_idle;
92 struct sk_buff_head rx_skb_q;
93
94 struct gbam_port *port;
95 struct work_struct write_tobam_w;
Vijayavardhan Vennapusa929e5792011-12-12 17:34:53 +053096 struct work_struct write_tohost_w;
Bryan Huntsman3f2bc4d2011-08-16 17:27:22 -070097
Ofir Cohena1c2a872011-12-14 10:26:34 +020098 struct usb_request *rx_req;
99 struct usb_request *tx_req;
100
Shimrit Malichi255b5342012-08-02 21:01:43 +0300101 u32 src_pipe_idx;
102 u32 dst_pipe_idx;
Ofir Cohena1c2a872011-12-14 10:26:34 +0200103 u8 connection_idx;
104
Bryan Huntsman3f2bc4d2011-08-16 17:27:22 -0700105 /* stats */
106 unsigned int pending_with_bam;
107 unsigned int tohost_drp_cnt;
108 unsigned int tomodem_drp_cnt;
109 unsigned int tx_len;
110 unsigned int rx_len;
111 unsigned long to_modem;
112 unsigned long to_host;
113};
114
115struct gbam_port {
116 unsigned port_num;
Vijayavardhan Vennapusaa8808532012-01-09 15:24:02 +0530117 spinlock_t port_lock_ul;
118 spinlock_t port_lock_dl;
Bryan Huntsman3f2bc4d2011-08-16 17:27:22 -0700119
120 struct grmnet *port_usb;
Ofir Cohena1c2a872011-12-14 10:26:34 +0200121 struct grmnet *gr;
Bryan Huntsman3f2bc4d2011-08-16 17:27:22 -0700122
123 struct bam_ch_info data_ch;
124
125 struct work_struct connect_w;
Vamsi Krishna1ad076d2011-11-10 15:03:30 -0800126 struct work_struct disconnect_w;
Bryan Huntsman3f2bc4d2011-08-16 17:27:22 -0700127};
128
129static struct bam_portmaster {
130 struct gbam_port *port;
Jack Phameffd4ae2011-08-03 16:49:36 -0700131 struct platform_driver pdrv;
132} bam_ports[BAM_N_PORTS];
Bryan Huntsman3f2bc4d2011-08-16 17:27:22 -0700133
Ofir Cohena1c2a872011-12-14 10:26:34 +0200134struct gbam_port *bam2bam_ports[BAM2BAM_N_PORTS];
Bryan Huntsman3f2bc4d2011-08-16 17:27:22 -0700135static void gbam_start_rx(struct gbam_port *port);
Ofir Cohena1c2a872011-12-14 10:26:34 +0200136static void gbam_start_endless_rx(struct gbam_port *port);
137static void gbam_start_endless_tx(struct gbam_port *port);
Bryan Huntsman3f2bc4d2011-08-16 17:27:22 -0700138
139/*---------------misc functions---------------- */
140static void gbam_free_requests(struct usb_ep *ep, struct list_head *head)
141{
142 struct usb_request *req;
143
144 while (!list_empty(head)) {
145 req = list_entry(head->next, struct usb_request, list);
146 list_del(&req->list);
147 usb_ep_free_request(ep, req);
148 }
149}
150
151static int gbam_alloc_requests(struct usb_ep *ep, struct list_head *head,
152 int num,
153 void (*cb)(struct usb_ep *ep, struct usb_request *),
154 gfp_t flags)
155{
156 int i;
157 struct usb_request *req;
158
159 pr_debug("%s: ep:%p head:%p num:%d cb:%p", __func__,
160 ep, head, num, cb);
161
162 for (i = 0; i < num; i++) {
163 req = usb_ep_alloc_request(ep, flags);
164 if (!req) {
165 pr_debug("%s: req allocated:%d\n", __func__, i);
166 return list_empty(head) ? -ENOMEM : 0;
167 }
168 req->complete = cb;
169 list_add(&req->list, head);
170 }
171
172 return 0;
173}
174/*--------------------------------------------- */
175
176/*------------data_path----------------------------*/
Chiranjeevi Velempatie5105922012-01-19 12:25:26 +0530177static void gbam_write_data_tohost(struct gbam_port *port)
Bryan Huntsman3f2bc4d2011-08-16 17:27:22 -0700178{
179 unsigned long flags;
Chiranjeevi Velempatie5105922012-01-19 12:25:26 +0530180 struct bam_ch_info *d = &port->data_ch;
Bryan Huntsman3f2bc4d2011-08-16 17:27:22 -0700181 struct sk_buff *skb;
182 int ret;
183 struct usb_request *req;
184 struct usb_ep *ep;
185
Vijayavardhan Vennapusaa8808532012-01-09 15:24:02 +0530186 spin_lock_irqsave(&port->port_lock_dl, flags);
Bryan Huntsman3f2bc4d2011-08-16 17:27:22 -0700187 if (!port->port_usb) {
Vijayavardhan Vennapusaa8808532012-01-09 15:24:02 +0530188 spin_unlock_irqrestore(&port->port_lock_dl, flags);
Bryan Huntsman3f2bc4d2011-08-16 17:27:22 -0700189 return;
190 }
191
192 ep = port->port_usb->in;
193
194 while (!list_empty(&d->tx_idle)) {
195 skb = __skb_dequeue(&d->tx_skb_q);
196 if (!skb) {
Vijayavardhan Vennapusaa8808532012-01-09 15:24:02 +0530197 spin_unlock_irqrestore(&port->port_lock_dl, flags);
Bryan Huntsman3f2bc4d2011-08-16 17:27:22 -0700198 return;
199 }
200 req = list_first_entry(&d->tx_idle,
201 struct usb_request,
202 list);
203 req->context = skb;
204 req->buf = skb->data;
205 req->length = skb->len;
Vijayavardhan Vennapusa08c31252011-12-21 13:02:49 +0530206 n_tx_req_queued++;
207 if (n_tx_req_queued == dl_intr_threshold) {
208 req->no_interrupt = 0;
209 n_tx_req_queued = 0;
210 } else {
211 req->no_interrupt = 1;
212 }
Bryan Huntsman3f2bc4d2011-08-16 17:27:22 -0700213
214 list_del(&req->list);
215
Vijayavardhan Vennapusaa8808532012-01-09 15:24:02 +0530216 spin_unlock(&port->port_lock_dl);
Bryan Huntsman3f2bc4d2011-08-16 17:27:22 -0700217 ret = usb_ep_queue(ep, req, GFP_ATOMIC);
Vijayavardhan Vennapusaa8808532012-01-09 15:24:02 +0530218 spin_lock(&port->port_lock_dl);
Bryan Huntsman3f2bc4d2011-08-16 17:27:22 -0700219 if (ret) {
220 pr_err("%s: usb epIn failed\n", __func__);
221 list_add(&req->list, &d->tx_idle);
222 dev_kfree_skb_any(skb);
223 break;
224 }
225 d->to_host++;
226 }
Vijayavardhan Vennapusaa8808532012-01-09 15:24:02 +0530227 spin_unlock_irqrestore(&port->port_lock_dl, flags);
Bryan Huntsman3f2bc4d2011-08-16 17:27:22 -0700228}
229
Chiranjeevi Velempatie5105922012-01-19 12:25:26 +0530230static void gbam_write_data_tohost_w(struct work_struct *w)
231{
232 struct bam_ch_info *d;
233 struct gbam_port *port;
234
235 d = container_of(w, struct bam_ch_info, write_tohost_w);
236 port = d->port;
237
238 gbam_write_data_tohost(port);
239}
240
Bryan Huntsman3f2bc4d2011-08-16 17:27:22 -0700241void gbam_data_recv_cb(void *p, struct sk_buff *skb)
242{
243 struct gbam_port *port = p;
244 struct bam_ch_info *d = &port->data_ch;
245 unsigned long flags;
246
247 if (!skb)
248 return;
249
250 pr_debug("%s: p:%p#%d d:%p skb_len:%d\n", __func__,
251 port, port->port_num, d, skb->len);
252
Vijayavardhan Vennapusaa8808532012-01-09 15:24:02 +0530253 spin_lock_irqsave(&port->port_lock_dl, flags);
Bryan Huntsman3f2bc4d2011-08-16 17:27:22 -0700254 if (!port->port_usb) {
Vijayavardhan Vennapusaa8808532012-01-09 15:24:02 +0530255 spin_unlock_irqrestore(&port->port_lock_dl, flags);
Bryan Huntsman3f2bc4d2011-08-16 17:27:22 -0700256 dev_kfree_skb_any(skb);
257 return;
258 }
259
Vamsi Krishna8f24f252011-11-02 11:46:08 -0700260 if (d->tx_skb_q.qlen > bam_mux_tx_pkt_drop_thld) {
Bryan Huntsman3f2bc4d2011-08-16 17:27:22 -0700261 d->tohost_drp_cnt++;
262 if (printk_ratelimit())
263 pr_err("%s: tx pkt dropped: tx_drop_cnt:%u\n",
264 __func__, d->tohost_drp_cnt);
Vijayavardhan Vennapusaa8808532012-01-09 15:24:02 +0530265 spin_unlock_irqrestore(&port->port_lock_dl, flags);
Bryan Huntsman3f2bc4d2011-08-16 17:27:22 -0700266 dev_kfree_skb_any(skb);
267 return;
268 }
269
270 __skb_queue_tail(&d->tx_skb_q, skb);
Vijayavardhan Vennapusaa8808532012-01-09 15:24:02 +0530271 spin_unlock_irqrestore(&port->port_lock_dl, flags);
Bryan Huntsman3f2bc4d2011-08-16 17:27:22 -0700272
Chiranjeevi Velempatie5105922012-01-19 12:25:26 +0530273 gbam_write_data_tohost(port);
Bryan Huntsman3f2bc4d2011-08-16 17:27:22 -0700274}
275
276void gbam_data_write_done(void *p, struct sk_buff *skb)
277{
278 struct gbam_port *port = p;
279 struct bam_ch_info *d = &port->data_ch;
280 unsigned long flags;
281
282 if (!skb)
283 return;
284
285 dev_kfree_skb_any(skb);
286
Vijayavardhan Vennapusaa8808532012-01-09 15:24:02 +0530287 spin_lock_irqsave(&port->port_lock_ul, flags);
Bryan Huntsman3f2bc4d2011-08-16 17:27:22 -0700288
289 d->pending_with_bam--;
290
291 pr_debug("%s: port:%p d:%p tom:%lu pbam:%u, pno:%d\n", __func__,
292 port, d, d->to_modem,
293 d->pending_with_bam, port->port_num);
294
Vijayavardhan Vennapusaa8808532012-01-09 15:24:02 +0530295 spin_unlock_irqrestore(&port->port_lock_ul, flags);
Bryan Huntsman3f2bc4d2011-08-16 17:27:22 -0700296
Vamsi Krishna84579552011-11-09 15:33:22 -0800297 queue_work(gbam_wq, &d->write_tobam_w);
Bryan Huntsman3f2bc4d2011-08-16 17:27:22 -0700298}
299
300static void gbam_data_write_tobam(struct work_struct *w)
301{
302 struct gbam_port *port;
303 struct bam_ch_info *d;
304 struct sk_buff *skb;
305 unsigned long flags;
306 int ret;
Vamsi Krishna84579552011-11-09 15:33:22 -0800307 int qlen;
Bryan Huntsman3f2bc4d2011-08-16 17:27:22 -0700308
309 d = container_of(w, struct bam_ch_info, write_tobam_w);
310 port = d->port;
311
Vijayavardhan Vennapusaa8808532012-01-09 15:24:02 +0530312 spin_lock_irqsave(&port->port_lock_ul, flags);
Bryan Huntsman3f2bc4d2011-08-16 17:27:22 -0700313 if (!port->port_usb) {
Vijayavardhan Vennapusaa8808532012-01-09 15:24:02 +0530314 spin_unlock_irqrestore(&port->port_lock_ul, flags);
Bryan Huntsman3f2bc4d2011-08-16 17:27:22 -0700315 return;
316 }
317
Vamsi Krishna84579552011-11-09 15:33:22 -0800318 while (d->pending_with_bam < BAM_PENDING_LIMIT) {
Vamsi Krishna2327c79152011-11-08 16:12:42 -0800319 skb = __skb_dequeue(&d->rx_skb_q);
Vamsi Krishna625c28e2011-12-16 22:34:49 -0800320 if (!skb)
321 break;
322
Bryan Huntsman3f2bc4d2011-08-16 17:27:22 -0700323 d->pending_with_bam++;
324 d->to_modem++;
325
326 pr_debug("%s: port:%p d:%p tom:%lu pbam:%u pno:%d\n", __func__,
327 port, d, d->to_modem, d->pending_with_bam,
328 port->port_num);
329
Vijayavardhan Vennapusaa8808532012-01-09 15:24:02 +0530330 spin_unlock_irqrestore(&port->port_lock_ul, flags);
Bryan Huntsman3f2bc4d2011-08-16 17:27:22 -0700331 ret = msm_bam_dmux_write(d->id, skb);
Vijayavardhan Vennapusaa8808532012-01-09 15:24:02 +0530332 spin_lock_irqsave(&port->port_lock_ul, flags);
Bryan Huntsman3f2bc4d2011-08-16 17:27:22 -0700333 if (ret) {
334 pr_debug("%s: write error:%d\n", __func__, ret);
335 d->pending_with_bam--;
336 d->to_modem--;
337 d->tomodem_drp_cnt++;
338 dev_kfree_skb_any(skb);
339 break;
340 }
341 }
Vamsi Krishna84579552011-11-09 15:33:22 -0800342
343 qlen = d->rx_skb_q.qlen;
344
Vijayavardhan Vennapusaa8808532012-01-09 15:24:02 +0530345 spin_unlock_irqrestore(&port->port_lock_ul, flags);
Vamsi Krishna84579552011-11-09 15:33:22 -0800346
347 if (qlen < BAM_MUX_RX_PKT_FCTRL_DIS_TSHOLD)
348 gbam_start_rx(port);
Bryan Huntsman3f2bc4d2011-08-16 17:27:22 -0700349}
350/*-------------------------------------------------------------*/
351
352static void gbam_epin_complete(struct usb_ep *ep, struct usb_request *req)
353{
354 struct gbam_port *port = ep->driver_data;
355 struct bam_ch_info *d;
356 struct sk_buff *skb = req->context;
357 int status = req->status;
358
359 switch (status) {
360 case 0:
361 /* successful completion */
362 case -ECONNRESET:
363 case -ESHUTDOWN:
364 /* connection gone */
365 break;
366 default:
367 pr_err("%s: data tx ep error %d\n",
368 __func__, status);
369 break;
370 }
371
372 dev_kfree_skb_any(skb);
373
374 if (!port)
375 return;
376
Vijayavardhan Vennapusaa8808532012-01-09 15:24:02 +0530377 spin_lock(&port->port_lock_dl);
Bryan Huntsman3f2bc4d2011-08-16 17:27:22 -0700378 d = &port->data_ch;
379 list_add_tail(&req->list, &d->tx_idle);
Vijayavardhan Vennapusaa8808532012-01-09 15:24:02 +0530380 spin_unlock(&port->port_lock_dl);
Bryan Huntsman3f2bc4d2011-08-16 17:27:22 -0700381
Vijayavardhan Vennapusa929e5792011-12-12 17:34:53 +0530382 queue_work(gbam_wq, &d->write_tohost_w);
Bryan Huntsman3f2bc4d2011-08-16 17:27:22 -0700383}
384
385static void
386gbam_epout_complete(struct usb_ep *ep, struct usb_request *req)
387{
388 struct gbam_port *port = ep->driver_data;
389 struct bam_ch_info *d = &port->data_ch;
390 struct sk_buff *skb = req->context;
391 int status = req->status;
392 int queue = 0;
393
394 switch (status) {
395 case 0:
396 skb_put(skb, req->actual);
397 queue = 1;
398 break;
399 case -ECONNRESET:
400 case -ESHUTDOWN:
401 /* cable disconnection */
402 dev_kfree_skb_any(skb);
403 req->buf = 0;
404 usb_ep_free_request(ep, req);
405 return;
406 default:
407 if (printk_ratelimit())
408 pr_err("%s: %s response error %d, %d/%d\n",
409 __func__, ep->name, status,
410 req->actual, req->length);
411 dev_kfree_skb_any(skb);
412 break;
413 }
414
Vijayavardhan Vennapusaa8808532012-01-09 15:24:02 +0530415 spin_lock(&port->port_lock_ul);
Bryan Huntsman3f2bc4d2011-08-16 17:27:22 -0700416 if (queue) {
417 __skb_queue_tail(&d->rx_skb_q, skb);
418 queue_work(gbam_wq, &d->write_tobam_w);
419 }
420
421 /* TODO: Handle flow control gracefully by having
422 * having call back mechanism from bam driver
423 */
Vamsi Krishna8f24f252011-11-02 11:46:08 -0700424 if (bam_mux_rx_fctrl_support &&
Vamsi Krishna84579552011-11-09 15:33:22 -0800425 d->rx_skb_q.qlen >= bam_mux_rx_fctrl_en_thld) {
Bryan Huntsman3f2bc4d2011-08-16 17:27:22 -0700426
427 list_add_tail(&req->list, &d->rx_idle);
Vijayavardhan Vennapusaa8808532012-01-09 15:24:02 +0530428 spin_unlock(&port->port_lock_ul);
Bryan Huntsman3f2bc4d2011-08-16 17:27:22 -0700429 return;
430 }
Vijayavardhan Vennapusaa8808532012-01-09 15:24:02 +0530431 spin_unlock(&port->port_lock_ul);
Bryan Huntsman3f2bc4d2011-08-16 17:27:22 -0700432
Vamsi Krishna8f24f252011-11-02 11:46:08 -0700433 skb = alloc_skb(bam_mux_rx_req_size + BAM_MUX_HDR, GFP_ATOMIC);
Bryan Huntsman3f2bc4d2011-08-16 17:27:22 -0700434 if (!skb) {
Vijayavardhan Vennapusaa8808532012-01-09 15:24:02 +0530435 spin_lock(&port->port_lock_ul);
Bryan Huntsman3f2bc4d2011-08-16 17:27:22 -0700436 list_add_tail(&req->list, &d->rx_idle);
Vijayavardhan Vennapusaa8808532012-01-09 15:24:02 +0530437 spin_unlock(&port->port_lock_ul);
Bryan Huntsman3f2bc4d2011-08-16 17:27:22 -0700438 return;
439 }
440 skb_reserve(skb, BAM_MUX_HDR);
441
442 req->buf = skb->data;
Vamsi Krishna8f24f252011-11-02 11:46:08 -0700443 req->length = bam_mux_rx_req_size;
Bryan Huntsman3f2bc4d2011-08-16 17:27:22 -0700444 req->context = skb;
445
446 status = usb_ep_queue(ep, req, GFP_ATOMIC);
447 if (status) {
448 dev_kfree_skb_any(skb);
449
450 if (printk_ratelimit())
451 pr_err("%s: data rx enqueue err %d\n",
452 __func__, status);
453
Vijayavardhan Vennapusaa8808532012-01-09 15:24:02 +0530454 spin_lock(&port->port_lock_ul);
Bryan Huntsman3f2bc4d2011-08-16 17:27:22 -0700455 list_add_tail(&req->list, &d->rx_idle);
Vijayavardhan Vennapusaa8808532012-01-09 15:24:02 +0530456 spin_unlock(&port->port_lock_ul);
Bryan Huntsman3f2bc4d2011-08-16 17:27:22 -0700457 }
458}
459
Ofir Cohena1c2a872011-12-14 10:26:34 +0200460static void gbam_endless_rx_complete(struct usb_ep *ep, struct usb_request *req)
461{
462 int status = req->status;
463
464 pr_debug("%s status: %d\n", __func__, status);
465}
466
467static void gbam_endless_tx_complete(struct usb_ep *ep, struct usb_request *req)
468{
469 int status = req->status;
470
471 pr_debug("%s status: %d\n", __func__, status);
472}
473
Bryan Huntsman3f2bc4d2011-08-16 17:27:22 -0700474static void gbam_start_rx(struct gbam_port *port)
475{
476 struct usb_request *req;
477 struct bam_ch_info *d;
478 struct usb_ep *ep;
479 unsigned long flags;
480 int ret;
481 struct sk_buff *skb;
482
Vijayavardhan Vennapusaa8808532012-01-09 15:24:02 +0530483 spin_lock_irqsave(&port->port_lock_ul, flags);
Bryan Huntsman3f2bc4d2011-08-16 17:27:22 -0700484 if (!port->port_usb) {
Vijayavardhan Vennapusaa8808532012-01-09 15:24:02 +0530485 spin_unlock_irqrestore(&port->port_lock_ul, flags);
Bryan Huntsman3f2bc4d2011-08-16 17:27:22 -0700486 return;
487 }
488
489 d = &port->data_ch;
490 ep = port->port_usb->out;
491
492 while (port->port_usb && !list_empty(&d->rx_idle)) {
Vamsi Krishna84579552011-11-09 15:33:22 -0800493
494 if (bam_mux_rx_fctrl_support &&
495 d->rx_skb_q.qlen >= bam_mux_rx_fctrl_en_thld)
496 break;
497
Bryan Huntsman3f2bc4d2011-08-16 17:27:22 -0700498 req = list_first_entry(&d->rx_idle, struct usb_request, list);
499
Vamsi Krishna8f24f252011-11-02 11:46:08 -0700500 skb = alloc_skb(bam_mux_rx_req_size + BAM_MUX_HDR, GFP_ATOMIC);
Bryan Huntsman3f2bc4d2011-08-16 17:27:22 -0700501 if (!skb)
502 break;
503 skb_reserve(skb, BAM_MUX_HDR);
504
505 list_del(&req->list);
506 req->buf = skb->data;
Vamsi Krishna8f24f252011-11-02 11:46:08 -0700507 req->length = bam_mux_rx_req_size;
Bryan Huntsman3f2bc4d2011-08-16 17:27:22 -0700508 req->context = skb;
509
Vijayavardhan Vennapusaa8808532012-01-09 15:24:02 +0530510 spin_unlock_irqrestore(&port->port_lock_ul, flags);
Bryan Huntsman3f2bc4d2011-08-16 17:27:22 -0700511 ret = usb_ep_queue(ep, req, GFP_ATOMIC);
Vijayavardhan Vennapusaa8808532012-01-09 15:24:02 +0530512 spin_lock_irqsave(&port->port_lock_ul, flags);
Bryan Huntsman3f2bc4d2011-08-16 17:27:22 -0700513 if (ret) {
514 dev_kfree_skb_any(skb);
515
516 if (printk_ratelimit())
517 pr_err("%s: rx queue failed\n", __func__);
518
519 if (port->port_usb)
520 list_add(&req->list, &d->rx_idle);
521 else
522 usb_ep_free_request(ep, req);
523 break;
524 }
525 }
Vijayavardhan Vennapusaa8808532012-01-09 15:24:02 +0530526 spin_unlock_irqrestore(&port->port_lock_ul, flags);
Bryan Huntsman3f2bc4d2011-08-16 17:27:22 -0700527}
528
Ofir Cohena1c2a872011-12-14 10:26:34 +0200529static void gbam_start_endless_rx(struct gbam_port *port)
530{
531 struct bam_ch_info *d = &port->data_ch;
532 int status;
533
534 status = usb_ep_queue(port->port_usb->out, d->rx_req, GFP_ATOMIC);
535 if (status)
536 pr_err("%s: error enqueuing transfer, %d\n", __func__, status);
537}
538
539static void gbam_start_endless_tx(struct gbam_port *port)
540{
541 struct bam_ch_info *d = &port->data_ch;
542 int status;
543
544 status = usb_ep_queue(port->port_usb->in, d->tx_req, GFP_ATOMIC);
545 if (status)
546 pr_err("%s: error enqueuing transfer, %d\n", __func__, status);
547}
548
Bryan Huntsman3f2bc4d2011-08-16 17:27:22 -0700549static void gbam_start_io(struct gbam_port *port)
550{
551 unsigned long flags;
552 struct usb_ep *ep;
553 int ret;
554 struct bam_ch_info *d;
555
556 pr_debug("%s: port:%p\n", __func__, port);
557
Vijayavardhan Vennapusaa8808532012-01-09 15:24:02 +0530558 spin_lock_irqsave(&port->port_lock_ul, flags);
Bryan Huntsman3f2bc4d2011-08-16 17:27:22 -0700559 if (!port->port_usb) {
Vijayavardhan Vennapusaa8808532012-01-09 15:24:02 +0530560 spin_unlock_irqrestore(&port->port_lock_ul, flags);
Bryan Huntsman3f2bc4d2011-08-16 17:27:22 -0700561 return;
562 }
563
564 d = &port->data_ch;
565 ep = port->port_usb->out;
Vamsi Krishna8f24f252011-11-02 11:46:08 -0700566 ret = gbam_alloc_requests(ep, &d->rx_idle, bam_mux_rx_q_size,
Bryan Huntsman3f2bc4d2011-08-16 17:27:22 -0700567 gbam_epout_complete, GFP_ATOMIC);
568 if (ret) {
569 pr_err("%s: rx req allocation failed\n", __func__);
570 return;
571 }
572
Vijayavardhan Vennapusaa8808532012-01-09 15:24:02 +0530573 spin_unlock_irqrestore(&port->port_lock_ul, flags);
574 spin_lock_irqsave(&port->port_lock_dl, flags);
Chiranjeevi Velempati9179d542013-02-18 15:22:54 +0530575 if (!port->port_usb) {
576 gbam_free_requests(ep, &d->rx_idle);
577 spin_unlock_irqrestore(&port->port_lock_dl, flags);
578 return;
579 }
Bryan Huntsman3f2bc4d2011-08-16 17:27:22 -0700580 ep = port->port_usb->in;
Vamsi Krishna8f24f252011-11-02 11:46:08 -0700581 ret = gbam_alloc_requests(ep, &d->tx_idle, bam_mux_tx_q_size,
Bryan Huntsman3f2bc4d2011-08-16 17:27:22 -0700582 gbam_epin_complete, GFP_ATOMIC);
583 if (ret) {
584 pr_err("%s: tx req allocation failed\n", __func__);
585 gbam_free_requests(ep, &d->rx_idle);
586 return;
587 }
588
Vijayavardhan Vennapusaa8808532012-01-09 15:24:02 +0530589 spin_unlock_irqrestore(&port->port_lock_dl, flags);
Bryan Huntsman3f2bc4d2011-08-16 17:27:22 -0700590
591 /* queue out requests */
592 gbam_start_rx(port);
593}
594
Jeff Hugo1c4531c2011-08-02 14:55:37 -0600595static void gbam_notify(void *p, int event, unsigned long data)
596{
597 switch (event) {
598 case BAM_DMUX_RECEIVE:
599 gbam_data_recv_cb(p, (struct sk_buff *)(data));
600 break;
601 case BAM_DMUX_WRITE_DONE:
602 gbam_data_write_done(p, (struct sk_buff *)(data));
603 break;
604 }
605}
606
Ofir Cohena1c2a872011-12-14 10:26:34 +0200607static void gbam_free_buffers(struct gbam_port *port)
608{
609 struct sk_buff *skb;
610 unsigned long flags;
611 struct bam_ch_info *d;
612
Vijayavardhan Vennapusaa8808532012-01-09 15:24:02 +0530613 spin_lock_irqsave(&port->port_lock_ul, flags);
Jeff Ohlstein6f8c5fc2012-01-17 13:08:39 -0800614 spin_lock(&port->port_lock_dl);
Ofir Cohena1c2a872011-12-14 10:26:34 +0200615
616 if (!port || !port->port_usb)
617 goto free_buf_out;
618
619 d = &port->data_ch;
620
621 gbam_free_requests(port->port_usb->in, &d->tx_idle);
622 gbam_free_requests(port->port_usb->out, &d->rx_idle);
623
624 while ((skb = __skb_dequeue(&d->tx_skb_q)))
625 dev_kfree_skb_any(skb);
626
627 while ((skb = __skb_dequeue(&d->rx_skb_q)))
628 dev_kfree_skb_any(skb);
629
630free_buf_out:
Jeff Ohlstein6f8c5fc2012-01-17 13:08:39 -0800631 spin_unlock(&port->port_lock_dl);
Vijayavardhan Vennapusaa8808532012-01-09 15:24:02 +0530632 spin_unlock_irqrestore(&port->port_lock_ul, flags);
Ofir Cohena1c2a872011-12-14 10:26:34 +0200633}
634
Vamsi Krishna1ad076d2011-11-10 15:03:30 -0800635static void gbam_disconnect_work(struct work_struct *w)
636{
637 struct gbam_port *port =
638 container_of(w, struct gbam_port, disconnect_w);
639 struct bam_ch_info *d = &port->data_ch;
640
641 if (!test_bit(BAM_CH_OPENED, &d->flags))
642 return;
643
644 msm_bam_dmux_close(d->id);
645 clear_bit(BAM_CH_OPENED, &d->flags);
646}
647
Ofir Cohena1c2a872011-12-14 10:26:34 +0200648static void gbam2bam_disconnect_work(struct work_struct *w)
649{
650 struct gbam_port *port =
651 container_of(w, struct gbam_port, disconnect_w);
652 unsigned long flags;
653
Vijayavardhan Vennapusaa8808532012-01-09 15:24:02 +0530654 spin_lock_irqsave(&port->port_lock_ul, flags);
Jeff Ohlstein6f8c5fc2012-01-17 13:08:39 -0800655 spin_lock(&port->port_lock_dl);
Ofir Cohena1c2a872011-12-14 10:26:34 +0200656 port->port_usb = 0;
Jeff Ohlstein6f8c5fc2012-01-17 13:08:39 -0800657 spin_unlock(&port->port_lock_dl);
Vijayavardhan Vennapusaa8808532012-01-09 15:24:02 +0530658 spin_unlock_irqrestore(&port->port_lock_ul, flags);
Ofir Cohena1c2a872011-12-14 10:26:34 +0200659
660 /* disable endpoints */
661 usb_ep_disable(port->gr->out);
662 usb_ep_disable(port->gr->in);
663
Anna Perel97b8c222012-01-18 10:08:14 +0200664 port->gr->in->driver_data = NULL;
665 port->gr->out->driver_data = NULL;
Ofir Cohena1c2a872011-12-14 10:26:34 +0200666}
667
Bryan Huntsman3f2bc4d2011-08-16 17:27:22 -0700668static void gbam_connect_work(struct work_struct *w)
669{
670 struct gbam_port *port = container_of(w, struct gbam_port, connect_w);
671 struct bam_ch_info *d = &port->data_ch;
672 int ret;
Vamsi Krishna1ad076d2011-11-10 15:03:30 -0800673 unsigned long flags;
674
Vijayavardhan Vennapusaa8808532012-01-09 15:24:02 +0530675 spin_lock_irqsave(&port->port_lock_ul, flags);
Jeff Ohlstein6f8c5fc2012-01-17 13:08:39 -0800676 spin_lock(&port->port_lock_dl);
Vamsi Krishna1ad076d2011-11-10 15:03:30 -0800677 if (!port->port_usb) {
Jeff Ohlstein6f8c5fc2012-01-17 13:08:39 -0800678 spin_unlock(&port->port_lock_dl);
Vijayavardhan Vennapusaa8808532012-01-09 15:24:02 +0530679 spin_unlock_irqrestore(&port->port_lock_ul, flags);
Vamsi Krishna1ad076d2011-11-10 15:03:30 -0800680 return;
681 }
Jeff Ohlstein6f8c5fc2012-01-17 13:08:39 -0800682 spin_unlock(&port->port_lock_dl);
Vijayavardhan Vennapusaa8808532012-01-09 15:24:02 +0530683 spin_unlock_irqrestore(&port->port_lock_ul, flags);
Bryan Huntsman3f2bc4d2011-08-16 17:27:22 -0700684
Jack Phameffd4ae2011-08-03 16:49:36 -0700685 if (!test_bit(BAM_CH_READY, &d->flags))
686 return;
687
Jeff Hugo1c4531c2011-08-02 14:55:37 -0600688 ret = msm_bam_dmux_open(d->id, port, gbam_notify);
Bryan Huntsman3f2bc4d2011-08-16 17:27:22 -0700689 if (ret) {
690 pr_err("%s: unable open bam ch:%d err:%d\n",
691 __func__, d->id, ret);
692 return;
693 }
Jack Phameffd4ae2011-08-03 16:49:36 -0700694 set_bit(BAM_CH_OPENED, &d->flags);
Bryan Huntsman3f2bc4d2011-08-16 17:27:22 -0700695
696 gbam_start_io(port);
697
698 pr_debug("%s: done\n", __func__);
699}
700
Ofir Cohena1c2a872011-12-14 10:26:34 +0200701static void gbam2bam_connect_work(struct work_struct *w)
Jack Phameffd4ae2011-08-03 16:49:36 -0700702{
Ofir Cohena1c2a872011-12-14 10:26:34 +0200703 struct gbam_port *port = container_of(w, struct gbam_port, connect_w);
704 struct bam_ch_info *d = &port->data_ch;
705 u32 sps_params;
706 int ret;
Ofir Cohen4da266f2012-01-03 10:19:29 +0200707 unsigned long flags;
708
Tatyana Brokhmancf709c12011-06-28 16:33:48 +0300709 ret = usb_ep_enable(port->gr->in);
Ofir Cohen4da266f2012-01-03 10:19:29 +0200710 if (ret) {
711 pr_err("%s: usb_ep_enable failed eptype:IN ep:%p",
712 __func__, port->gr->in);
713 return;
714 }
715 port->gr->in->driver_data = port;
716
Tatyana Brokhmancf709c12011-06-28 16:33:48 +0300717 ret = usb_ep_enable(port->gr->out);
Ofir Cohen4da266f2012-01-03 10:19:29 +0200718 if (ret) {
719 pr_err("%s: usb_ep_enable failed eptype:OUT ep:%p",
720 __func__, port->gr->out);
721 port->gr->in->driver_data = 0;
722 return;
723 }
724 port->gr->out->driver_data = port;
Vijayavardhan Vennapusaa8808532012-01-09 15:24:02 +0530725 spin_lock_irqsave(&port->port_lock_ul, flags);
Jeff Ohlstein6f8c5fc2012-01-17 13:08:39 -0800726 spin_lock(&port->port_lock_dl);
Ofir Cohen4da266f2012-01-03 10:19:29 +0200727 port->port_usb = port->gr;
Jeff Ohlstein6f8c5fc2012-01-17 13:08:39 -0800728 spin_unlock(&port->port_lock_dl);
Vijayavardhan Vennapusaa8808532012-01-09 15:24:02 +0530729 spin_unlock_irqrestore(&port->port_lock_ul, flags);
Jack Phameffd4ae2011-08-03 16:49:36 -0700730
Ofir Cohena1c2a872011-12-14 10:26:34 +0200731 ret = usb_bam_connect(d->connection_idx, &d->src_pipe_idx,
732 &d->dst_pipe_idx);
733 if (ret) {
734 pr_err("%s: usb_bam_connect failed: err:%d\n",
735 __func__, ret);
736 return;
737 }
Jack Phameffd4ae2011-08-03 16:49:36 -0700738
Ofir Cohena1c2a872011-12-14 10:26:34 +0200739 d->rx_req = usb_ep_alloc_request(port->port_usb->out, GFP_KERNEL);
740 if (!d->rx_req)
741 return;
Jack Phameffd4ae2011-08-03 16:49:36 -0700742
Ofir Cohena1c2a872011-12-14 10:26:34 +0200743 d->rx_req->context = port;
744 d->rx_req->complete = gbam_endless_rx_complete;
745 d->rx_req->length = 0;
Ido Shayevitzd1cb16c2012-03-28 18:57:47 +0200746 sps_params = (MSM_SPS_MODE | d->src_pipe_idx |
747 MSM_VENDOR_ID) & ~MSM_IS_FINITE_TRANSFER;
Ofir Cohena1c2a872011-12-14 10:26:34 +0200748 d->rx_req->udc_priv = sps_params;
749 d->tx_req = usb_ep_alloc_request(port->port_usb->in, GFP_KERNEL);
750 if (!d->tx_req)
751 return;
Jack Phameffd4ae2011-08-03 16:49:36 -0700752
Ofir Cohena1c2a872011-12-14 10:26:34 +0200753 d->tx_req->context = port;
754 d->tx_req->complete = gbam_endless_tx_complete;
755 d->tx_req->length = 0;
Ido Shayevitzd1cb16c2012-03-28 18:57:47 +0200756 sps_params = (MSM_SPS_MODE | d->dst_pipe_idx |
757 MSM_VENDOR_ID) & ~MSM_IS_FINITE_TRANSFER;
Ofir Cohena1c2a872011-12-14 10:26:34 +0200758 d->tx_req->udc_priv = sps_params;
Jack Phameffd4ae2011-08-03 16:49:36 -0700759
Ofir Cohena1c2a872011-12-14 10:26:34 +0200760 /* queue in & out requests */
761 gbam_start_endless_rx(port);
762 gbam_start_endless_tx(port);
Jack Phameffd4ae2011-08-03 16:49:36 -0700763
Ofir Cohena1c2a872011-12-14 10:26:34 +0200764 pr_debug("%s: done\n", __func__);
Jack Phameffd4ae2011-08-03 16:49:36 -0700765}
766
767/* BAM data channel ready, allow attempt to open */
768static int gbam_data_ch_probe(struct platform_device *pdev)
769{
770 struct gbam_port *port;
771 struct bam_ch_info *d;
772 int i;
773 unsigned long flags;
774
775 pr_debug("%s: name:%s\n", __func__, pdev->name);
776
777 for (i = 0; i < n_bam_ports; i++) {
778 port = bam_ports[i].port;
779 d = &port->data_ch;
780
781 if (!strncmp(bam_ch_names[i], pdev->name,
782 BAM_DMUX_CH_NAME_MAX_LEN)) {
783 set_bit(BAM_CH_READY, &d->flags);
784
785 /* if usb is online, try opening bam_ch */
Vijayavardhan Vennapusaa8808532012-01-09 15:24:02 +0530786 spin_lock_irqsave(&port->port_lock_ul, flags);
Jeff Ohlstein6f8c5fc2012-01-17 13:08:39 -0800787 spin_lock(&port->port_lock_dl);
Jack Phameffd4ae2011-08-03 16:49:36 -0700788 if (port->port_usb)
789 queue_work(gbam_wq, &port->connect_w);
Jeff Ohlstein6f8c5fc2012-01-17 13:08:39 -0800790 spin_unlock(&port->port_lock_dl);
Vijayavardhan Vennapusaa8808532012-01-09 15:24:02 +0530791 spin_unlock_irqrestore(&port->port_lock_ul, flags);
Jack Phameffd4ae2011-08-03 16:49:36 -0700792
793 break;
794 }
795 }
796
797 return 0;
798}
799
800/* BAM data channel went inactive, so close it */
801static int gbam_data_ch_remove(struct platform_device *pdev)
802{
803 struct gbam_port *port;
804 struct bam_ch_info *d;
805 struct usb_ep *ep_in = NULL;
806 struct usb_ep *ep_out = NULL;
807 unsigned long flags;
808 int i;
809
810 pr_debug("%s: name:%s\n", __func__, pdev->name);
811
812 for (i = 0; i < n_bam_ports; i++) {
813 if (!strncmp(bam_ch_names[i], pdev->name,
814 BAM_DMUX_CH_NAME_MAX_LEN)) {
815 port = bam_ports[i].port;
816 d = &port->data_ch;
817
Vijayavardhan Vennapusaa8808532012-01-09 15:24:02 +0530818 spin_lock_irqsave(&port->port_lock_ul, flags);
Jeff Ohlstein6f8c5fc2012-01-17 13:08:39 -0800819 spin_lock(&port->port_lock_dl);
Jack Phameffd4ae2011-08-03 16:49:36 -0700820 if (port->port_usb) {
821 ep_in = port->port_usb->in;
822 ep_out = port->port_usb->out;
823 }
Jeff Ohlstein6f8c5fc2012-01-17 13:08:39 -0800824 spin_unlock(&port->port_lock_dl);
Vijayavardhan Vennapusaa8808532012-01-09 15:24:02 +0530825 spin_unlock_irqrestore(&port->port_lock_ul, flags);
Jack Phameffd4ae2011-08-03 16:49:36 -0700826
827 if (ep_in)
828 usb_ep_fifo_flush(ep_in);
829 if (ep_out)
830 usb_ep_fifo_flush(ep_out);
831
832 gbam_free_buffers(port);
833
834 msm_bam_dmux_close(d->id);
835
Vamsi Krishna7658bd12012-01-13 10:32:00 -0800836 /* bam dmux will free all pending skbs */
837 d->pending_with_bam = 0;
838
Jack Phameffd4ae2011-08-03 16:49:36 -0700839 clear_bit(BAM_CH_READY, &d->flags);
840 clear_bit(BAM_CH_OPENED, &d->flags);
841 }
842 }
843
844 return 0;
845}
846
Bryan Huntsman3f2bc4d2011-08-16 17:27:22 -0700847static void gbam_port_free(int portno)
848{
849 struct gbam_port *port = bam_ports[portno].port;
Jack Phameffd4ae2011-08-03 16:49:36 -0700850 struct platform_driver *pdrv = &bam_ports[portno].pdrv;
Bryan Huntsman3f2bc4d2011-08-16 17:27:22 -0700851
Jack Phameffd4ae2011-08-03 16:49:36 -0700852 if (port) {
Bryan Huntsman3f2bc4d2011-08-16 17:27:22 -0700853 kfree(port);
Jack Phameffd4ae2011-08-03 16:49:36 -0700854 platform_driver_unregister(pdrv);
855 }
Bryan Huntsman3f2bc4d2011-08-16 17:27:22 -0700856}
857
Ofir Cohena1c2a872011-12-14 10:26:34 +0200858static void gbam2bam_port_free(int portno)
859{
860 struct gbam_port *port = bam2bam_ports[portno];
861
862 kfree(port);
863}
864
Bryan Huntsman3f2bc4d2011-08-16 17:27:22 -0700865static int gbam_port_alloc(int portno)
866{
867 struct gbam_port *port;
868 struct bam_ch_info *d;
Jack Phameffd4ae2011-08-03 16:49:36 -0700869 struct platform_driver *pdrv;
Bryan Huntsman3f2bc4d2011-08-16 17:27:22 -0700870
871 port = kzalloc(sizeof(struct gbam_port), GFP_KERNEL);
872 if (!port)
873 return -ENOMEM;
874
875 port->port_num = portno;
876
877 /* port initialization */
Vijayavardhan Vennapusaa8808532012-01-09 15:24:02 +0530878 spin_lock_init(&port->port_lock_ul);
879 spin_lock_init(&port->port_lock_dl);
Bryan Huntsman3f2bc4d2011-08-16 17:27:22 -0700880 INIT_WORK(&port->connect_w, gbam_connect_work);
Vamsi Krishna1ad076d2011-11-10 15:03:30 -0800881 INIT_WORK(&port->disconnect_w, gbam_disconnect_work);
Bryan Huntsman3f2bc4d2011-08-16 17:27:22 -0700882
883 /* data ch */
884 d = &port->data_ch;
885 d->port = port;
886 INIT_LIST_HEAD(&d->tx_idle);
887 INIT_LIST_HEAD(&d->rx_idle);
888 INIT_WORK(&d->write_tobam_w, gbam_data_write_tobam);
Chiranjeevi Velempatie5105922012-01-19 12:25:26 +0530889 INIT_WORK(&d->write_tohost_w, gbam_write_data_tohost_w);
Bryan Huntsman3f2bc4d2011-08-16 17:27:22 -0700890 skb_queue_head_init(&d->tx_skb_q);
891 skb_queue_head_init(&d->rx_skb_q);
892 d->id = bam_ch_ids[portno];
893
894 bam_ports[portno].port = port;
895
Jack Phameffd4ae2011-08-03 16:49:36 -0700896 pdrv = &bam_ports[portno].pdrv;
897 pdrv->probe = gbam_data_ch_probe;
898 pdrv->remove = gbam_data_ch_remove;
899 pdrv->driver.name = bam_ch_names[portno];
900 pdrv->driver.owner = THIS_MODULE;
901
902 platform_driver_register(pdrv);
Ofir Cohena1c2a872011-12-14 10:26:34 +0200903 pr_debug("%s: port:%p portno:%d\n", __func__, port, portno);
904
905 return 0;
906}
907
908static int gbam2bam_port_alloc(int portno)
909{
910 struct gbam_port *port;
911 struct bam_ch_info *d;
912
913 port = kzalloc(sizeof(struct gbam_port), GFP_KERNEL);
914 if (!port)
915 return -ENOMEM;
916
917 port->port_num = portno;
918
919 /* port initialization */
Vijayavardhan Vennapusaa8808532012-01-09 15:24:02 +0530920 spin_lock_init(&port->port_lock_ul);
921 spin_lock_init(&port->port_lock_dl);
Ofir Cohena1c2a872011-12-14 10:26:34 +0200922
923 INIT_WORK(&port->connect_w, gbam2bam_connect_work);
924 INIT_WORK(&port->disconnect_w, gbam2bam_disconnect_work);
925
926 /* data ch */
927 d = &port->data_ch;
928 d->port = port;
929 bam2bam_ports[portno] = port;
Jack Phameffd4ae2011-08-03 16:49:36 -0700930
Bryan Huntsman3f2bc4d2011-08-16 17:27:22 -0700931 pr_debug("%s: port:%p portno:%d\n", __func__, port, portno);
932
933 return 0;
934}
935
936#if defined(CONFIG_DEBUG_FS)
937#define DEBUG_BUF_SIZE 1024
938static ssize_t gbam_read_stats(struct file *file, char __user *ubuf,
939 size_t count, loff_t *ppos)
940{
941 struct gbam_port *port;
942 struct bam_ch_info *d;
943 char *buf;
944 unsigned long flags;
945 int ret;
946 int i;
947 int temp = 0;
948
949 buf = kzalloc(sizeof(char) * DEBUG_BUF_SIZE, GFP_KERNEL);
950 if (!buf)
951 return -ENOMEM;
952
953 for (i = 0; i < n_bam_ports; i++) {
954 port = bam_ports[i].port;
955 if (!port)
956 continue;
Vijayavardhan Vennapusaa8808532012-01-09 15:24:02 +0530957 spin_lock_irqsave(&port->port_lock_ul, flags);
Jeff Ohlstein6f8c5fc2012-01-17 13:08:39 -0800958 spin_lock(&port->port_lock_dl);
Bryan Huntsman3f2bc4d2011-08-16 17:27:22 -0700959
960 d = &port->data_ch;
961
962 temp += scnprintf(buf + temp, DEBUG_BUF_SIZE - temp,
963 "#PORT:%d port:%p data_ch:%p#\n"
964 "dpkts_to_usbhost: %lu\n"
965 "dpkts_to_modem: %lu\n"
966 "dpkts_pwith_bam: %u\n"
967 "to_usbhost_dcnt: %u\n"
968 "tomodem__dcnt: %u\n"
969 "tx_buf_len: %u\n"
Vamsi Krishna84579552011-11-09 15:33:22 -0800970 "rx_buf_len: %u\n"
Jack Phameffd4ae2011-08-03 16:49:36 -0700971 "data_ch_open: %d\n"
972 "data_ch_ready: %d\n",
Bryan Huntsman3f2bc4d2011-08-16 17:27:22 -0700973 i, port, &port->data_ch,
974 d->to_host, d->to_modem,
975 d->pending_with_bam,
976 d->tohost_drp_cnt, d->tomodem_drp_cnt,
Vamsi Krishna84579552011-11-09 15:33:22 -0800977 d->tx_skb_q.qlen, d->rx_skb_q.qlen,
Jack Phameffd4ae2011-08-03 16:49:36 -0700978 test_bit(BAM_CH_OPENED, &d->flags),
979 test_bit(BAM_CH_READY, &d->flags));
Bryan Huntsman3f2bc4d2011-08-16 17:27:22 -0700980
Jeff Ohlstein6f8c5fc2012-01-17 13:08:39 -0800981 spin_unlock(&port->port_lock_dl);
Vijayavardhan Vennapusaa8808532012-01-09 15:24:02 +0530982 spin_unlock_irqrestore(&port->port_lock_ul, flags);
Bryan Huntsman3f2bc4d2011-08-16 17:27:22 -0700983 }
984
985 ret = simple_read_from_buffer(ubuf, count, ppos, buf, temp);
986
987 kfree(buf);
988
989 return ret;
990}
991
992static ssize_t gbam_reset_stats(struct file *file, const char __user *buf,
993 size_t count, loff_t *ppos)
994{
995 struct gbam_port *port;
996 struct bam_ch_info *d;
997 int i;
998 unsigned long flags;
999
1000 for (i = 0; i < n_bam_ports; i++) {
1001 port = bam_ports[i].port;
1002 if (!port)
1003 continue;
1004
Vijayavardhan Vennapusaa8808532012-01-09 15:24:02 +05301005 spin_lock_irqsave(&port->port_lock_ul, flags);
Jeff Ohlstein6f8c5fc2012-01-17 13:08:39 -08001006 spin_lock(&port->port_lock_dl);
Bryan Huntsman3f2bc4d2011-08-16 17:27:22 -07001007
1008 d = &port->data_ch;
1009
1010 d->to_host = 0;
1011 d->to_modem = 0;
1012 d->pending_with_bam = 0;
1013 d->tohost_drp_cnt = 0;
1014 d->tomodem_drp_cnt = 0;
1015
Jeff Ohlstein6f8c5fc2012-01-17 13:08:39 -08001016 spin_unlock(&port->port_lock_dl);
Vijayavardhan Vennapusaa8808532012-01-09 15:24:02 +05301017 spin_unlock_irqrestore(&port->port_lock_ul, flags);
Bryan Huntsman3f2bc4d2011-08-16 17:27:22 -07001018 }
1019 return count;
1020}
1021
1022const struct file_operations gbam_stats_ops = {
1023 .read = gbam_read_stats,
1024 .write = gbam_reset_stats,
1025};
1026
1027static void gbam_debugfs_init(void)
1028{
1029 struct dentry *dent;
1030 struct dentry *dfile;
1031
1032 dent = debugfs_create_dir("usb_rmnet", 0);
1033 if (IS_ERR(dent))
1034 return;
1035
1036 /* TODO: Implement cleanup function to remove created file */
1037 dfile = debugfs_create_file("status", 0444, dent, 0, &gbam_stats_ops);
1038 if (!dfile || IS_ERR(dfile))
1039 debugfs_remove(dent);
1040}
1041#else
1042static void gam_debugfs_init(void) { }
1043#endif
1044
Ofir Cohena1c2a872011-12-14 10:26:34 +02001045void gbam_disconnect(struct grmnet *gr, u8 port_num, enum transport_type trans)
Bryan Huntsman3f2bc4d2011-08-16 17:27:22 -07001046{
1047 struct gbam_port *port;
1048 unsigned long flags;
1049 struct bam_ch_info *d;
1050
1051 pr_debug("%s: grmnet:%p port#%d\n", __func__, gr, port_num);
1052
Ofir Cohena1c2a872011-12-14 10:26:34 +02001053 if (trans == USB_GADGET_XPORT_BAM &&
1054 port_num >= n_bam_ports) {
1055 pr_err("%s: invalid bam portno#%d\n",
1056 __func__, port_num);
1057 return;
1058 }
1059
1060 if (trans == USB_GADGET_XPORT_BAM2BAM &&
1061 port_num >= n_bam2bam_ports) {
1062 pr_err("%s: invalid bam2bam portno#%d\n",
1063 __func__, port_num);
Bryan Huntsman3f2bc4d2011-08-16 17:27:22 -07001064 return;
1065 }
1066
1067 if (!gr) {
1068 pr_err("%s: grmnet port is null\n", __func__);
1069 return;
1070 }
Ofir Cohena1c2a872011-12-14 10:26:34 +02001071 if (trans == USB_GADGET_XPORT_BAM)
1072 port = bam_ports[port_num].port;
1073 else
1074 port = bam2bam_ports[port_num];
Bryan Huntsman3f2bc4d2011-08-16 17:27:22 -07001075
Bryan Huntsman3f2bc4d2011-08-16 17:27:22 -07001076 d = &port->data_ch;
Ofir Cohena1c2a872011-12-14 10:26:34 +02001077 port->gr = gr;
Bryan Huntsman3f2bc4d2011-08-16 17:27:22 -07001078
Ofir Cohena1c2a872011-12-14 10:26:34 +02001079 if (trans == USB_GADGET_XPORT_BAM) {
1080 gbam_free_buffers(port);
Bryan Huntsman3f2bc4d2011-08-16 17:27:22 -07001081
Vijayavardhan Vennapusaa8808532012-01-09 15:24:02 +05301082 spin_lock_irqsave(&port->port_lock_ul, flags);
Jeff Ohlstein6f8c5fc2012-01-17 13:08:39 -08001083 spin_lock(&port->port_lock_dl);
Vijayavardhan Vennapusaa8808532012-01-09 15:24:02 +05301084 port->port_usb = 0;
Vijayavardhan Vennapusa08c31252011-12-21 13:02:49 +05301085 n_tx_req_queued = 0;
Jeff Ohlstein6f8c5fc2012-01-17 13:08:39 -08001086 spin_unlock(&port->port_lock_dl);
Vijayavardhan Vennapusaa8808532012-01-09 15:24:02 +05301087 spin_unlock_irqrestore(&port->port_lock_ul, flags);
Bryan Huntsman3f2bc4d2011-08-16 17:27:22 -07001088
Ofir Cohena1c2a872011-12-14 10:26:34 +02001089 /* disable endpoints */
1090 usb_ep_disable(gr->out);
1091 usb_ep_disable(gr->in);
1092 }
Bryan Huntsman3f2bc4d2011-08-16 17:27:22 -07001093
Vamsi Krishna1ad076d2011-11-10 15:03:30 -08001094 queue_work(gbam_wq, &port->disconnect_w);
Bryan Huntsman3f2bc4d2011-08-16 17:27:22 -07001095}
1096
Ofir Cohena1c2a872011-12-14 10:26:34 +02001097int gbam_connect(struct grmnet *gr, u8 port_num,
1098 enum transport_type trans, u8 connection_idx)
Bryan Huntsman3f2bc4d2011-08-16 17:27:22 -07001099{
1100 struct gbam_port *port;
1101 struct bam_ch_info *d;
1102 int ret;
1103 unsigned long flags;
1104
1105 pr_debug("%s: grmnet:%p port#%d\n", __func__, gr, port_num);
1106
Ofir Cohena1c2a872011-12-14 10:26:34 +02001107 if (trans == USB_GADGET_XPORT_BAM && port_num >= n_bam_ports) {
1108 pr_err("%s: invalid portno#%d\n", __func__, port_num);
1109 return -ENODEV;
1110 }
1111
1112 if (trans == USB_GADGET_XPORT_BAM2BAM && port_num >= n_bam2bam_ports) {
Bryan Huntsman3f2bc4d2011-08-16 17:27:22 -07001113 pr_err("%s: invalid portno#%d\n", __func__, port_num);
1114 return -ENODEV;
1115 }
1116
1117 if (!gr) {
1118 pr_err("%s: grmnet port is null\n", __func__);
1119 return -ENODEV;
1120 }
1121
Ofir Cohena1c2a872011-12-14 10:26:34 +02001122 if (trans == USB_GADGET_XPORT_BAM)
1123 port = bam_ports[port_num].port;
1124 else
1125 port = bam2bam_ports[port_num];
1126
Bryan Huntsman3f2bc4d2011-08-16 17:27:22 -07001127 d = &port->data_ch;
1128
Ofir Cohena1c2a872011-12-14 10:26:34 +02001129 if (trans == USB_GADGET_XPORT_BAM) {
Tatyana Brokhmancf709c12011-06-28 16:33:48 +03001130 ret = usb_ep_enable(gr->in);
Ofir Cohen4da266f2012-01-03 10:19:29 +02001131 if (ret) {
1132 pr_err("%s: usb_ep_enable failed eptype:IN ep:%p",
1133 __func__, gr->in);
1134 return ret;
1135 }
1136 gr->in->driver_data = port;
1137
Tatyana Brokhmancf709c12011-06-28 16:33:48 +03001138 ret = usb_ep_enable(gr->out);
Ofir Cohen4da266f2012-01-03 10:19:29 +02001139 if (ret) {
1140 pr_err("%s: usb_ep_enable failed eptype:OUT ep:%p",
1141 __func__, gr->out);
1142 gr->in->driver_data = 0;
1143 return ret;
1144 }
1145 gr->out->driver_data = port;
1146
Vijayavardhan Vennapusaa8808532012-01-09 15:24:02 +05301147 spin_lock_irqsave(&port->port_lock_ul, flags);
Jeff Ohlstein6f8c5fc2012-01-17 13:08:39 -08001148 spin_lock(&port->port_lock_dl);
Vijayavardhan Vennapusaa8808532012-01-09 15:24:02 +05301149 port->port_usb = gr;
Ofir Cohen4da266f2012-01-03 10:19:29 +02001150
Ofir Cohena1c2a872011-12-14 10:26:34 +02001151 d->to_host = 0;
1152 d->to_modem = 0;
1153 d->pending_with_bam = 0;
1154 d->tohost_drp_cnt = 0;
1155 d->tomodem_drp_cnt = 0;
Jeff Ohlstein6f8c5fc2012-01-17 13:08:39 -08001156 spin_unlock(&port->port_lock_dl);
Vijayavardhan Vennapusaa8808532012-01-09 15:24:02 +05301157 spin_unlock_irqrestore(&port->port_lock_ul, flags);
Ofir Cohena1c2a872011-12-14 10:26:34 +02001158 }
Bryan Huntsman3f2bc4d2011-08-16 17:27:22 -07001159
Ofir Cohen4da266f2012-01-03 10:19:29 +02001160 if (trans == USB_GADGET_XPORT_BAM2BAM) {
1161 port->gr = gr;
Ofir Cohena1c2a872011-12-14 10:26:34 +02001162 d->connection_idx = connection_idx;
Ofir Cohen4da266f2012-01-03 10:19:29 +02001163 }
Bryan Huntsman3f2bc4d2011-08-16 17:27:22 -07001164
1165 queue_work(gbam_wq, &port->connect_w);
1166
1167 return 0;
1168}
1169
Ofir Cohena1c2a872011-12-14 10:26:34 +02001170int gbam_setup(unsigned int no_bam_port, unsigned int no_bam2bam_port)
Bryan Huntsman3f2bc4d2011-08-16 17:27:22 -07001171{
1172 int i;
1173 int ret;
1174
Ofir Cohena1c2a872011-12-14 10:26:34 +02001175 pr_debug("%s: requested BAM ports:%d and BAM2BAM ports:%d\n",
1176 __func__, no_bam_port, no_bam2bam_port);
Bryan Huntsman3f2bc4d2011-08-16 17:27:22 -07001177
Ofir Cohena1c2a872011-12-14 10:26:34 +02001178 if ((!no_bam_port && !no_bam2bam_port) || no_bam_port > BAM_N_PORTS
1179 || no_bam2bam_port > BAM2BAM_N_PORTS) {
1180 pr_err("%s: Invalid num of ports count:%d,%d\n",
1181 __func__, no_bam_port, no_bam2bam_port);
Bryan Huntsman3f2bc4d2011-08-16 17:27:22 -07001182 return -EINVAL;
1183 }
1184
1185 gbam_wq = alloc_workqueue("k_gbam", WQ_UNBOUND | WQ_MEM_RECLAIM, 1);
1186 if (!gbam_wq) {
1187 pr_err("%s: Unable to create workqueue gbam_wq\n",
1188 __func__);
1189 return -ENOMEM;
1190 }
1191
Ofir Cohena1c2a872011-12-14 10:26:34 +02001192 for (i = 0; i < no_bam_port; i++) {
Manu Gautamd59b5d32011-09-09 14:47:08 +05301193 n_bam_ports++;
Bryan Huntsman3f2bc4d2011-08-16 17:27:22 -07001194 ret = gbam_port_alloc(i);
1195 if (ret) {
Manu Gautamd59b5d32011-09-09 14:47:08 +05301196 n_bam_ports--;
Bryan Huntsman3f2bc4d2011-08-16 17:27:22 -07001197 pr_err("%s: Unable to alloc port:%d\n", __func__, i);
1198 goto free_bam_ports;
1199 }
Bryan Huntsman3f2bc4d2011-08-16 17:27:22 -07001200 }
1201
Ofir Cohena1c2a872011-12-14 10:26:34 +02001202 for (i = 0; i < no_bam2bam_port; i++) {
1203 n_bam2bam_ports++;
1204 ret = gbam2bam_port_alloc(i);
1205 if (ret) {
1206 n_bam2bam_ports--;
1207 pr_err("%s: Unable to alloc port:%d\n", __func__, i);
1208 goto free_bam_ports;
1209 }
1210 }
Bryan Huntsman3f2bc4d2011-08-16 17:27:22 -07001211 gbam_debugfs_init();
Bryan Huntsman3f2bc4d2011-08-16 17:27:22 -07001212 return 0;
Ofir Cohena1c2a872011-12-14 10:26:34 +02001213
Bryan Huntsman3f2bc4d2011-08-16 17:27:22 -07001214free_bam_ports:
1215 for (i = 0; i < n_bam_ports; i++)
1216 gbam_port_free(i);
Ofir Cohena1c2a872011-12-14 10:26:34 +02001217 for (i = 0; i < n_bam2bam_ports; i++)
1218 gbam2bam_port_free(i);
Bryan Huntsman3f2bc4d2011-08-16 17:27:22 -07001219 destroy_workqueue(gbam_wq);
1220
1221 return ret;
1222}
Amit Blaye5bb35e2012-05-08 20:38:20 +03001223
1224static int gbam_wake_cb(void *param)
1225{
1226 struct gbam_port *port = (struct gbam_port *)param;
1227 struct bam_ch_info *d;
1228 struct f_rmnet *dev;
1229
1230 dev = port_to_rmnet(port->gr);
1231 d = &port->data_ch;
1232
1233 pr_debug("%s: woken up by peer\n", __func__);
1234
1235 return usb_gadget_wakeup(dev->cdev->gadget);
1236}
1237
1238void gbam_suspend(struct grmnet *gr, u8 port_num, enum transport_type trans)
1239{
1240 struct gbam_port *port;
1241 struct bam_ch_info *d;
1242
1243 if (trans != USB_GADGET_XPORT_BAM2BAM)
1244 return;
1245
1246 port = bam2bam_ports[port_num];
1247 d = &port->data_ch;
1248
1249 pr_debug("%s: suspended port %d\n", __func__, port_num);
1250
1251 usb_bam_register_wake_cb(d->connection_idx, gbam_wake_cb, port);
1252}
1253
1254void gbam_resume(struct grmnet *gr, u8 port_num, enum transport_type trans)
1255{
1256 struct gbam_port *port;
1257 struct bam_ch_info *d;
1258
1259 if (trans != USB_GADGET_XPORT_BAM2BAM)
1260 return;
1261
1262 port = bam2bam_ports[port_num];
1263 d = &port->data_ch;
1264
1265 pr_debug("%s: resumed port %d\n", __func__, port_num);
1266
1267 usb_bam_register_wake_cb(d->connection_idx, NULL, NULL);
1268}