blob: 382291b91f7b6fe75368a1344a45182dee8a626e [file] [log] [blame]
Mian Yousaf Kaukab8dcc8f72011-03-22 15:55:58 +01001/*
2 * drivers/usb/musb/ux500_dma.c
3 *
Fabio Baltieri3ee1f2e2013-04-03 10:45:02 +02004 * U8500 DMA support code
Mian Yousaf Kaukab8dcc8f72011-03-22 15:55:58 +01005 *
6 * Copyright (C) 2009 STMicroelectronics
7 * Copyright (C) 2011 ST-Ericsson SA
8 * Authors:
9 * Mian Yousaf Kaukab <mian.yousaf.kaukab@stericsson.com>
10 * Praveena Nadahally <praveen.nadahally@stericsson.com>
11 * Rajaram Regupathy <ragupathy.rajaram@stericsson.com>
12 *
13 * This program is free software: you can redistribute it and/or modify
14 * it under the terms of the GNU General Public License as published by
15 * the Free Software Foundation, either version 2 of the License, or
16 * (at your option) any later version.
17 *
18 * This program is distributed in the hope that it will be useful,
19 * but WITHOUT ANY WARRANTY; without even the implied warranty of
20 * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
21 * GNU General Public License for more details.
22 *
23 * You should have received a copy of the GNU General Public License
24 * along with this program. If not, see <http://www.gnu.org/licenses/>.
25 */
26
27#include <linux/device.h>
28#include <linux/interrupt.h>
29#include <linux/platform_device.h>
30#include <linux/dma-mapping.h>
31#include <linux/dmaengine.h>
32#include <linux/pfn.h>
Felipe Balbi0f53e482013-02-06 09:47:58 +020033#include <linux/sizes.h>
Arnd Bergmanndb298da2012-08-24 15:19:33 +020034#include <linux/platform_data/usb-musb-ux500.h>
Mian Yousaf Kaukab8dcc8f72011-03-22 15:55:58 +010035#include "musb_core.h"
36
37struct ux500_dma_channel {
38 struct dma_channel channel;
39 struct ux500_dma_controller *controller;
40 struct musb_hw_ep *hw_ep;
Mian Yousaf Kaukab8dcc8f72011-03-22 15:55:58 +010041 struct dma_chan *dma_chan;
42 unsigned int cur_len;
43 dma_cookie_t cookie;
44 u8 ch_num;
45 u8 is_tx;
46 u8 is_allocated;
47};
48
49struct ux500_dma_controller {
50 struct dma_controller controller;
Lee Jonesbe2dbb02013-05-15 10:51:43 +010051 struct ux500_dma_channel rx_channel[UX500_MUSB_DMA_NUM_RX_TX_CHANNELS];
52 struct ux500_dma_channel tx_channel[UX500_MUSB_DMA_NUM_RX_TX_CHANNELS];
Mian Yousaf Kaukab8dcc8f72011-03-22 15:55:58 +010053 void *private_data;
54 dma_addr_t phy_base;
55};
56
Mian Yousaf Kaukab8dcc8f72011-03-22 15:55:58 +010057/* Work function invoked from DMA callback to handle rx transfers. */
Felipe Balbi6b0cfc62013-03-22 17:03:32 +020058static void ux500_dma_callback(void *private_data)
Mian Yousaf Kaukab8dcc8f72011-03-22 15:55:58 +010059{
Per Forlinbe18a252011-08-17 11:03:40 +020060 struct dma_channel *channel = private_data;
61 struct ux500_dma_channel *ux500_channel = channel->private_data;
Mian Yousaf Kaukab8dcc8f72011-03-22 15:55:58 +010062 struct musb_hw_ep *hw_ep = ux500_channel->hw_ep;
63 struct musb *musb = hw_ep->musb;
64 unsigned long flags;
65
Per Forlinafbd0742011-08-03 14:22:17 +020066 dev_dbg(musb->controller, "DMA rx transfer done on hw_ep=%d\n",
67 hw_ep->epnum);
Mian Yousaf Kaukab8dcc8f72011-03-22 15:55:58 +010068
69 spin_lock_irqsave(&musb->lock, flags);
70 ux500_channel->channel.actual_len = ux500_channel->cur_len;
71 ux500_channel->channel.status = MUSB_DMA_STATUS_FREE;
72 musb_dma_completion(musb, hw_ep->epnum,
73 ux500_channel->is_tx);
74 spin_unlock_irqrestore(&musb->lock, flags);
Mian Yousaf Kaukab8dcc8f72011-03-22 15:55:58 +010075
Mian Yousaf Kaukab8dcc8f72011-03-22 15:55:58 +010076}
77
78static bool ux500_configure_channel(struct dma_channel *channel,
79 u16 packet_sz, u8 mode,
80 dma_addr_t dma_addr, u32 len)
81{
82 struct ux500_dma_channel *ux500_channel = channel->private_data;
83 struct musb_hw_ep *hw_ep = ux500_channel->hw_ep;
84 struct dma_chan *dma_chan = ux500_channel->dma_chan;
85 struct dma_async_tx_descriptor *dma_desc;
Vinod Koul83415442011-10-14 21:50:31 +053086 enum dma_transfer_direction direction;
Mian Yousaf Kaukab8dcc8f72011-03-22 15:55:58 +010087 struct scatterlist sg;
88 struct dma_slave_config slave_conf;
89 enum dma_slave_buswidth addr_width;
90 dma_addr_t usb_fifo_addr = (MUSB_FIFO_OFFSET(hw_ep->epnum) +
91 ux500_channel->controller->phy_base);
Per Forlinafbd0742011-08-03 14:22:17 +020092 struct musb *musb = ux500_channel->controller->private_data;
Mian Yousaf Kaukab8dcc8f72011-03-22 15:55:58 +010093
Per Forlinafbd0742011-08-03 14:22:17 +020094 dev_dbg(musb->controller,
Felipe Balbi6a3b0032013-02-06 09:53:01 +020095 "packet_sz=%d, mode=%d, dma_addr=0x%llu, len=%d is_tx=%d\n",
96 packet_sz, mode, (unsigned long long) dma_addr,
97 len, ux500_channel->is_tx);
Mian Yousaf Kaukab8dcc8f72011-03-22 15:55:58 +010098
99 ux500_channel->cur_len = len;
100
101 sg_init_table(&sg, 1);
102 sg_set_page(&sg, pfn_to_page(PFN_DOWN(dma_addr)), len,
103 offset_in_page(dma_addr));
104 sg_dma_address(&sg) = dma_addr;
105 sg_dma_len(&sg) = len;
106
Vinod Koul83415442011-10-14 21:50:31 +0530107 direction = ux500_channel->is_tx ? DMA_MEM_TO_DEV : DMA_DEV_TO_MEM;
Mian Yousaf Kaukab8dcc8f72011-03-22 15:55:58 +0100108 addr_width = (len & 0x3) ? DMA_SLAVE_BUSWIDTH_1_BYTE :
109 DMA_SLAVE_BUSWIDTH_4_BYTES;
110
111 slave_conf.direction = direction;
Per Forlind366d39b2011-08-02 17:33:39 +0200112 slave_conf.src_addr = usb_fifo_addr;
113 slave_conf.src_addr_width = addr_width;
114 slave_conf.src_maxburst = 16;
115 slave_conf.dst_addr = usb_fifo_addr;
116 slave_conf.dst_addr_width = addr_width;
117 slave_conf.dst_maxburst = 16;
Viresh Kumar258aea72012-02-01 16:12:19 +0530118 slave_conf.device_fc = false;
Per Forlind366d39b2011-08-02 17:33:39 +0200119
Mian Yousaf Kaukab8dcc8f72011-03-22 15:55:58 +0100120 dma_chan->device->device_control(dma_chan, DMA_SLAVE_CONFIG,
121 (unsigned long) &slave_conf);
122
Alexandre Bounine16052822012-03-08 16:11:18 -0500123 dma_desc = dmaengine_prep_slave_sg(dma_chan, &sg, 1, direction,
Mian Yousaf Kaukab8dcc8f72011-03-22 15:55:58 +0100124 DMA_PREP_INTERRUPT | DMA_CTRL_ACK);
125 if (!dma_desc)
126 return false;
127
128 dma_desc->callback = ux500_dma_callback;
129 dma_desc->callback_param = channel;
130 ux500_channel->cookie = dma_desc->tx_submit(dma_desc);
131
132 dma_async_issue_pending(dma_chan);
133
134 return true;
135}
136
137static struct dma_channel *ux500_dma_channel_allocate(struct dma_controller *c,
138 struct musb_hw_ep *hw_ep, u8 is_tx)
139{
140 struct ux500_dma_controller *controller = container_of(c,
141 struct ux500_dma_controller, controller);
142 struct ux500_dma_channel *ux500_channel = NULL;
Per Forlinafbd0742011-08-03 14:22:17 +0200143 struct musb *musb = controller->private_data;
Mian Yousaf Kaukab8dcc8f72011-03-22 15:55:58 +0100144 u8 ch_num = hw_ep->epnum - 1;
Mian Yousaf Kaukab8dcc8f72011-03-22 15:55:58 +0100145
Lee Jonesbe2dbb02013-05-15 10:51:43 +0100146 /* 8 DMA channels (0 - 7). Each DMA channel can only be allocated
Mian Yousaf Kaukab8dcc8f72011-03-22 15:55:58 +0100147 * to specified hw_ep. For example DMA channel 0 can only be allocated
148 * to hw_ep 1 and 9.
149 */
150 if (ch_num > 7)
151 ch_num -= 8;
152
Lee Jonesbe2dbb02013-05-15 10:51:43 +0100153 if (ch_num >= UX500_MUSB_DMA_NUM_RX_TX_CHANNELS)
Mian Yousaf Kaukab8dcc8f72011-03-22 15:55:58 +0100154 return NULL;
155
156 ux500_channel = is_tx ? &(controller->tx_channel[ch_num]) :
157 &(controller->rx_channel[ch_num]) ;
158
159 /* Check if channel is already used. */
160 if (ux500_channel->is_allocated)
161 return NULL;
162
163 ux500_channel->hw_ep = hw_ep;
164 ux500_channel->is_allocated = 1;
165
Per Forlinafbd0742011-08-03 14:22:17 +0200166 dev_dbg(musb->controller, "hw_ep=%d, is_tx=0x%x, channel=%d\n",
Mian Yousaf Kaukab8dcc8f72011-03-22 15:55:58 +0100167 hw_ep->epnum, is_tx, ch_num);
168
169 return &(ux500_channel->channel);
170}
171
172static void ux500_dma_channel_release(struct dma_channel *channel)
173{
174 struct ux500_dma_channel *ux500_channel = channel->private_data;
Per Forlinafbd0742011-08-03 14:22:17 +0200175 struct musb *musb = ux500_channel->controller->private_data;
Mian Yousaf Kaukab8dcc8f72011-03-22 15:55:58 +0100176
Per Forlinafbd0742011-08-03 14:22:17 +0200177 dev_dbg(musb->controller, "channel=%d\n", ux500_channel->ch_num);
Mian Yousaf Kaukab8dcc8f72011-03-22 15:55:58 +0100178
179 if (ux500_channel->is_allocated) {
180 ux500_channel->is_allocated = 0;
181 channel->status = MUSB_DMA_STATUS_FREE;
182 channel->actual_len = 0;
183 }
184}
185
186static int ux500_dma_is_compatible(struct dma_channel *channel,
187 u16 maxpacket, void *buf, u32 length)
188{
189 if ((maxpacket & 0x3) ||
Felipe Balbi6a3b0032013-02-06 09:53:01 +0200190 ((unsigned long int) buf & 0x3) ||
Mian Yousaf Kaukab8dcc8f72011-03-22 15:55:58 +0100191 (length < 512) ||
192 (length & 0x3))
193 return false;
194 else
195 return true;
196}
197
198static int ux500_dma_channel_program(struct dma_channel *channel,
199 u16 packet_sz, u8 mode,
200 dma_addr_t dma_addr, u32 len)
201{
202 int ret;
203
204 BUG_ON(channel->status == MUSB_DMA_STATUS_UNKNOWN ||
205 channel->status == MUSB_DMA_STATUS_BUSY);
206
207 if (!ux500_dma_is_compatible(channel, packet_sz, (void *)dma_addr, len))
208 return false;
209
210 channel->status = MUSB_DMA_STATUS_BUSY;
211 channel->actual_len = 0;
212 ret = ux500_configure_channel(channel, packet_sz, mode, dma_addr, len);
213 if (!ret)
214 channel->status = MUSB_DMA_STATUS_FREE;
215
216 return ret;
217}
218
219static int ux500_dma_channel_abort(struct dma_channel *channel)
220{
221 struct ux500_dma_channel *ux500_channel = channel->private_data;
222 struct ux500_dma_controller *controller = ux500_channel->controller;
223 struct musb *musb = controller->private_data;
224 void __iomem *epio = musb->endpoints[ux500_channel->hw_ep->epnum].regs;
225 u16 csr;
226
Per Forlinafbd0742011-08-03 14:22:17 +0200227 dev_dbg(musb->controller, "channel=%d, is_tx=%d\n",
228 ux500_channel->ch_num, ux500_channel->is_tx);
Mian Yousaf Kaukab8dcc8f72011-03-22 15:55:58 +0100229
230 if (channel->status == MUSB_DMA_STATUS_BUSY) {
231 if (ux500_channel->is_tx) {
232 csr = musb_readw(epio, MUSB_TXCSR);
233 csr &= ~(MUSB_TXCSR_AUTOSET |
234 MUSB_TXCSR_DMAENAB |
235 MUSB_TXCSR_DMAMODE);
236 musb_writew(epio, MUSB_TXCSR, csr);
237 } else {
238 csr = musb_readw(epio, MUSB_RXCSR);
239 csr &= ~(MUSB_RXCSR_AUTOCLEAR |
240 MUSB_RXCSR_DMAENAB |
241 MUSB_RXCSR_DMAMODE);
242 musb_writew(epio, MUSB_RXCSR, csr);
243 }
244
245 ux500_channel->dma_chan->device->
246 device_control(ux500_channel->dma_chan,
247 DMA_TERMINATE_ALL, 0);
248 channel->status = MUSB_DMA_STATUS_FREE;
249 }
250 return 0;
251}
252
253static int ux500_dma_controller_stop(struct dma_controller *c)
254{
255 struct ux500_dma_controller *controller = container_of(c,
256 struct ux500_dma_controller, controller);
257 struct ux500_dma_channel *ux500_channel;
258 struct dma_channel *channel;
259 u8 ch_num;
260
Lee Jonesbe2dbb02013-05-15 10:51:43 +0100261 for (ch_num = 0; ch_num < UX500_MUSB_DMA_NUM_RX_TX_CHANNELS; ch_num++) {
Mian Yousaf Kaukab8dcc8f72011-03-22 15:55:58 +0100262 channel = &controller->rx_channel[ch_num].channel;
263 ux500_channel = channel->private_data;
264
265 ux500_dma_channel_release(channel);
266
267 if (ux500_channel->dma_chan)
268 dma_release_channel(ux500_channel->dma_chan);
269 }
270
Lee Jonesbe2dbb02013-05-15 10:51:43 +0100271 for (ch_num = 0; ch_num < UX500_MUSB_DMA_NUM_RX_TX_CHANNELS; ch_num++) {
Mian Yousaf Kaukab8dcc8f72011-03-22 15:55:58 +0100272 channel = &controller->tx_channel[ch_num].channel;
273 ux500_channel = channel->private_data;
274
275 ux500_dma_channel_release(channel);
276
277 if (ux500_channel->dma_chan)
278 dma_release_channel(ux500_channel->dma_chan);
279 }
280
281 return 0;
282}
283
284static int ux500_dma_controller_start(struct dma_controller *c)
285{
286 struct ux500_dma_controller *controller = container_of(c,
287 struct ux500_dma_controller, controller);
288 struct ux500_dma_channel *ux500_channel = NULL;
289 struct musb *musb = controller->private_data;
290 struct device *dev = musb->controller;
291 struct musb_hdrc_platform_data *plat = dev->platform_data;
292 struct ux500_musb_board_data *data = plat->board_data;
293 struct dma_channel *dma_channel = NULL;
294 u32 ch_num;
295 u8 dir;
296 u8 is_tx = 0;
297
298 void **param_array;
299 struct ux500_dma_channel *channel_array;
Mian Yousaf Kaukab8dcc8f72011-03-22 15:55:58 +0100300 dma_cap_mask_t mask;
301
Mian Yousaf Kaukab8dcc8f72011-03-22 15:55:58 +0100302
Mian Yousaf Kaukab8dcc8f72011-03-22 15:55:58 +0100303
304 dma_cap_zero(mask);
305 dma_cap_set(DMA_SLAVE, mask);
306
307 /* Prepare the loop for RX channels */
308 channel_array = controller->rx_channel;
Mian Yousaf Kaukab8dcc8f72011-03-22 15:55:58 +0100309 param_array = data->dma_rx_param_array;
Mian Yousaf Kaukab8dcc8f72011-03-22 15:55:58 +0100310
311 for (dir = 0; dir < 2; dir++) {
Lee Jonesbe2dbb02013-05-15 10:51:43 +0100312 for (ch_num = 0;
313 ch_num < UX500_MUSB_DMA_NUM_RX_TX_CHANNELS;
314 ch_num++) {
Mian Yousaf Kaukab8dcc8f72011-03-22 15:55:58 +0100315 ux500_channel = &channel_array[ch_num];
316 ux500_channel->controller = controller;
317 ux500_channel->ch_num = ch_num;
318 ux500_channel->is_tx = is_tx;
319
320 dma_channel = &(ux500_channel->channel);
321 dma_channel->private_data = ux500_channel;
322 dma_channel->status = MUSB_DMA_STATUS_FREE;
323 dma_channel->max_len = SZ_16M;
324
325 ux500_channel->dma_chan = dma_request_channel(mask,
326 data->dma_filter,
327 param_array[ch_num]);
328 if (!ux500_channel->dma_chan) {
329 ERR("Dma pipe allocation error dir=%d ch=%d\n",
330 dir, ch_num);
331
332 /* Release already allocated channels */
333 ux500_dma_controller_stop(c);
334
335 return -EBUSY;
336 }
337
Mian Yousaf Kaukab8dcc8f72011-03-22 15:55:58 +0100338 }
339
340 /* Prepare the loop for TX channels */
341 channel_array = controller->tx_channel;
Mian Yousaf Kaukab8dcc8f72011-03-22 15:55:58 +0100342 param_array = data->dma_tx_param_array;
Mian Yousaf Kaukab8dcc8f72011-03-22 15:55:58 +0100343 is_tx = 1;
344 }
345
346 return 0;
347}
348
349void dma_controller_destroy(struct dma_controller *c)
350{
351 struct ux500_dma_controller *controller = container_of(c,
352 struct ux500_dma_controller, controller);
353
354 kfree(controller);
355}
356
Bill Pemberton41ac7b32012-11-19 13:21:48 -0500357struct dma_controller *dma_controller_create(struct musb *musb, void __iomem *base)
Mian Yousaf Kaukab8dcc8f72011-03-22 15:55:58 +0100358{
359 struct ux500_dma_controller *controller;
360 struct platform_device *pdev = to_platform_device(musb->controller);
361 struct resource *iomem;
362
363 controller = kzalloc(sizeof(*controller), GFP_KERNEL);
364 if (!controller)
Virupax Sadashivpetimath399e0f42013-03-08 10:27:05 +0800365 goto kzalloc_fail;
Mian Yousaf Kaukab8dcc8f72011-03-22 15:55:58 +0100366
367 controller->private_data = musb;
368
369 /* Save physical address for DMA controller. */
370 iomem = platform_get_resource(pdev, IORESOURCE_MEM, 0);
Virupax Sadashivpetimath399e0f42013-03-08 10:27:05 +0800371 if (!iomem) {
372 dev_err(musb->controller, "no memory resource defined\n");
373 goto plat_get_fail;
374 }
375
Mian Yousaf Kaukab8dcc8f72011-03-22 15:55:58 +0100376 controller->phy_base = (dma_addr_t) iomem->start;
377
378 controller->controller.start = ux500_dma_controller_start;
379 controller->controller.stop = ux500_dma_controller_stop;
380 controller->controller.channel_alloc = ux500_dma_channel_allocate;
381 controller->controller.channel_release = ux500_dma_channel_release;
382 controller->controller.channel_program = ux500_dma_channel_program;
383 controller->controller.channel_abort = ux500_dma_channel_abort;
384 controller->controller.is_compatible = ux500_dma_is_compatible;
385
386 return &controller->controller;
Virupax Sadashivpetimath399e0f42013-03-08 10:27:05 +0800387
388plat_get_fail:
389 kfree(controller);
390kzalloc_fail:
391 return NULL;
Mian Yousaf Kaukab8dcc8f72011-03-22 15:55:58 +0100392}