blob: 3036b6113ffdf1fbdb34df5c171148d7f5f5c2d7 [file] [log] [blame]
Madhusudhan Chikkature9f2bc792008-11-12 13:27:09 -08001/*
2 * drivers/w1/masters/omap_hdq.c
3 *
4 * Copyright (C) 2007 Texas Instruments, Inc.
5 *
6 * This file is licensed under the terms of the GNU General Public License
7 * version 2. This program is licensed "as is" without any warranty of any
8 * kind, whether express or implied.
9 *
10 */
11#include <linux/kernel.h>
12#include <linux/module.h>
13#include <linux/platform_device.h>
14#include <linux/interrupt.h>
Tejun Heo5a0e3ad2010-03-24 17:04:11 +090015#include <linux/slab.h>
Madhusudhan Chikkature9f2bc792008-11-12 13:27:09 -080016#include <linux/err.h>
17#include <linux/clk.h>
18#include <linux/io.h>
Amit Kucheria81fa08f2010-04-23 13:18:03 -040019#include <linux/sched.h>
Madhusudhan Chikkature9f2bc792008-11-12 13:27:09 -080020
21#include <asm/irq.h>
22#include <mach/hardware.h>
23
24#include "../w1.h"
25#include "../w1_int.h"
26
27#define MOD_NAME "OMAP_HDQ:"
28
29#define OMAP_HDQ_REVISION 0x00
30#define OMAP_HDQ_TX_DATA 0x04
31#define OMAP_HDQ_RX_DATA 0x08
32#define OMAP_HDQ_CTRL_STATUS 0x0c
33#define OMAP_HDQ_CTRL_STATUS_INTERRUPTMASK (1<<6)
34#define OMAP_HDQ_CTRL_STATUS_CLOCKENABLE (1<<5)
35#define OMAP_HDQ_CTRL_STATUS_GO (1<<4)
36#define OMAP_HDQ_CTRL_STATUS_INITIALIZATION (1<<2)
37#define OMAP_HDQ_CTRL_STATUS_DIR (1<<1)
38#define OMAP_HDQ_CTRL_STATUS_MODE (1<<0)
39#define OMAP_HDQ_INT_STATUS 0x10
40#define OMAP_HDQ_INT_STATUS_TXCOMPLETE (1<<2)
41#define OMAP_HDQ_INT_STATUS_RXCOMPLETE (1<<1)
42#define OMAP_HDQ_INT_STATUS_TIMEOUT (1<<0)
43#define OMAP_HDQ_SYSCONFIG 0x14
44#define OMAP_HDQ_SYSCONFIG_SOFTRESET (1<<1)
45#define OMAP_HDQ_SYSCONFIG_AUTOIDLE (1<<0)
46#define OMAP_HDQ_SYSSTATUS 0x18
47#define OMAP_HDQ_SYSSTATUS_RESETDONE (1<<0)
48
49#define OMAP_HDQ_FLAG_CLEAR 0
50#define OMAP_HDQ_FLAG_SET 1
51#define OMAP_HDQ_TIMEOUT (HZ/5)
52
53#define OMAP_HDQ_MAX_USER 4
54
55static DECLARE_WAIT_QUEUE_HEAD(hdq_wait_queue);
56static int w1_id;
57
58struct hdq_data {
59 struct device *dev;
60 void __iomem *hdq_base;
61 /* lock status update */
62 struct mutex hdq_mutex;
63 int hdq_usecount;
64 struct clk *hdq_ick;
65 struct clk *hdq_fck;
66 u8 hdq_irqstatus;
67 /* device lock */
68 spinlock_t hdq_spinlock;
69 /*
70 * Used to control the call to omap_hdq_get and omap_hdq_put.
71 * HDQ Protocol: Write the CMD|REG_address first, followed by
72 * the data wrire or read.
73 */
74 int init_trans;
75};
76
Uwe Kleine-Königa96b9122010-02-04 20:56:54 +010077static int __devinit omap_hdq_probe(struct platform_device *pdev);
Madhusudhan Chikkature9f2bc792008-11-12 13:27:09 -080078static int omap_hdq_remove(struct platform_device *pdev);
79
80static struct platform_driver omap_hdq_driver = {
81 .probe = omap_hdq_probe,
82 .remove = omap_hdq_remove,
83 .driver = {
84 .name = "omap_hdq",
85 },
86};
87
88static u8 omap_w1_read_byte(void *_hdq);
89static void omap_w1_write_byte(void *_hdq, u8 byte);
90static u8 omap_w1_reset_bus(void *_hdq);
Stanley.Miao06b0d4d2008-11-19 15:36:50 -080091static void omap_w1_search_bus(void *_hdq, struct w1_master *master_dev,
92 u8 search_type, w1_slave_found_callback slave_found);
Madhusudhan Chikkature9f2bc792008-11-12 13:27:09 -080093
94
95static struct w1_bus_master omap_w1_master = {
96 .read_byte = omap_w1_read_byte,
97 .write_byte = omap_w1_write_byte,
98 .reset_bus = omap_w1_reset_bus,
99 .search = omap_w1_search_bus,
100};
101
102/* HDQ register I/O routines */
103static inline u8 hdq_reg_in(struct hdq_data *hdq_data, u32 offset)
104{
105 return __raw_readb(hdq_data->hdq_base + offset);
106}
107
108static inline void hdq_reg_out(struct hdq_data *hdq_data, u32 offset, u8 val)
109{
110 __raw_writeb(val, hdq_data->hdq_base + offset);
111}
112
113static inline u8 hdq_reg_merge(struct hdq_data *hdq_data, u32 offset,
114 u8 val, u8 mask)
115{
116 u8 new_val = (__raw_readb(hdq_data->hdq_base + offset) & ~mask)
117 | (val & mask);
118 __raw_writeb(new_val, hdq_data->hdq_base + offset);
119
120 return new_val;
121}
122
123/*
124 * Wait for one or more bits in flag change.
125 * HDQ_FLAG_SET: wait until any bit in the flag is set.
126 * HDQ_FLAG_CLEAR: wait until all bits in the flag are cleared.
127 * return 0 on success and -ETIMEDOUT in the case of timeout.
128 */
129static int hdq_wait_for_flag(struct hdq_data *hdq_data, u32 offset,
130 u8 flag, u8 flag_set, u8 *status)
131{
132 int ret = 0;
133 unsigned long timeout = jiffies + OMAP_HDQ_TIMEOUT;
134
135 if (flag_set == OMAP_HDQ_FLAG_CLEAR) {
136 /* wait for the flag clear */
137 while (((*status = hdq_reg_in(hdq_data, offset)) & flag)
138 && time_before(jiffies, timeout)) {
139 schedule_timeout_uninterruptible(1);
140 }
141 if (*status & flag)
142 ret = -ETIMEDOUT;
143 } else if (flag_set == OMAP_HDQ_FLAG_SET) {
144 /* wait for the flag set */
145 while (!((*status = hdq_reg_in(hdq_data, offset)) & flag)
146 && time_before(jiffies, timeout)) {
147 schedule_timeout_uninterruptible(1);
148 }
149 if (!(*status & flag))
150 ret = -ETIMEDOUT;
151 } else
152 return -EINVAL;
153
154 return ret;
155}
156
157/* write out a byte and fill *status with HDQ_INT_STATUS */
158static int hdq_write_byte(struct hdq_data *hdq_data, u8 val, u8 *status)
159{
160 int ret;
161 u8 tmp_status;
162 unsigned long irqflags;
163
164 *status = 0;
165
166 spin_lock_irqsave(&hdq_data->hdq_spinlock, irqflags);
167 /* clear interrupt flags via a dummy read */
168 hdq_reg_in(hdq_data, OMAP_HDQ_INT_STATUS);
169 /* ISR loads it with new INT_STATUS */
170 hdq_data->hdq_irqstatus = 0;
171 spin_unlock_irqrestore(&hdq_data->hdq_spinlock, irqflags);
172
173 hdq_reg_out(hdq_data, OMAP_HDQ_TX_DATA, val);
174
175 /* set the GO bit */
176 hdq_reg_merge(hdq_data, OMAP_HDQ_CTRL_STATUS, OMAP_HDQ_CTRL_STATUS_GO,
177 OMAP_HDQ_CTRL_STATUS_DIR | OMAP_HDQ_CTRL_STATUS_GO);
178 /* wait for the TXCOMPLETE bit */
179 ret = wait_event_timeout(hdq_wait_queue,
180 hdq_data->hdq_irqstatus, OMAP_HDQ_TIMEOUT);
181 if (ret == 0) {
182 dev_dbg(hdq_data->dev, "TX wait elapsed\n");
NeilBrown7b5362a2012-05-22 09:43:02 +1000183 ret = -ETIMEDOUT;
Madhusudhan Chikkature9f2bc792008-11-12 13:27:09 -0800184 goto out;
185 }
186
187 *status = hdq_data->hdq_irqstatus;
188 /* check irqstatus */
189 if (!(*status & OMAP_HDQ_INT_STATUS_TXCOMPLETE)) {
190 dev_dbg(hdq_data->dev, "timeout waiting for"
NeilBrown7b5362a2012-05-22 09:43:02 +1000191 " TXCOMPLETE/RXCOMPLETE, %x", *status);
Madhusudhan Chikkature9f2bc792008-11-12 13:27:09 -0800192 ret = -ETIMEDOUT;
193 goto out;
194 }
195
196 /* wait for the GO bit return to zero */
197 ret = hdq_wait_for_flag(hdq_data, OMAP_HDQ_CTRL_STATUS,
198 OMAP_HDQ_CTRL_STATUS_GO,
199 OMAP_HDQ_FLAG_CLEAR, &tmp_status);
200 if (ret) {
201 dev_dbg(hdq_data->dev, "timeout waiting GO bit"
NeilBrown7b5362a2012-05-22 09:43:02 +1000202 " return to zero, %x", tmp_status);
Madhusudhan Chikkature9f2bc792008-11-12 13:27:09 -0800203 }
204
205out:
206 return ret;
207}
208
209/* HDQ Interrupt service routine */
210static irqreturn_t hdq_isr(int irq, void *_hdq)
211{
212 struct hdq_data *hdq_data = _hdq;
213 unsigned long irqflags;
214
215 spin_lock_irqsave(&hdq_data->hdq_spinlock, irqflags);
216 hdq_data->hdq_irqstatus = hdq_reg_in(hdq_data, OMAP_HDQ_INT_STATUS);
217 spin_unlock_irqrestore(&hdq_data->hdq_spinlock, irqflags);
218 dev_dbg(hdq_data->dev, "hdq_isr: %x", hdq_data->hdq_irqstatus);
219
220 if (hdq_data->hdq_irqstatus &
221 (OMAP_HDQ_INT_STATUS_TXCOMPLETE | OMAP_HDQ_INT_STATUS_RXCOMPLETE
222 | OMAP_HDQ_INT_STATUS_TIMEOUT)) {
223 /* wake up sleeping process */
224 wake_up(&hdq_wait_queue);
225 }
226
227 return IRQ_HANDLED;
228}
229
230/* HDQ Mode: always return success */
231static u8 omap_w1_reset_bus(void *_hdq)
232{
233 return 0;
234}
235
236/* W1 search callback function */
Stanley.Miao06b0d4d2008-11-19 15:36:50 -0800237static void omap_w1_search_bus(void *_hdq, struct w1_master *master_dev,
238 u8 search_type, w1_slave_found_callback slave_found)
Madhusudhan Chikkature9f2bc792008-11-12 13:27:09 -0800239{
240 u64 module_id, rn_le, cs, id;
241
242 if (w1_id)
243 module_id = w1_id;
244 else
245 module_id = 0x1;
246
247 rn_le = cpu_to_le64(module_id);
248 /*
249 * HDQ might not obey truly the 1-wire spec.
250 * So calculate CRC based on module parameter.
251 */
252 cs = w1_calc_crc8((u8 *)&rn_le, 7);
253 id = (cs << 56) | module_id;
254
Stanley.Miao06b0d4d2008-11-19 15:36:50 -0800255 slave_found(master_dev, id);
Madhusudhan Chikkature9f2bc792008-11-12 13:27:09 -0800256}
257
258static int _omap_hdq_reset(struct hdq_data *hdq_data)
259{
260 int ret;
261 u8 tmp_status;
262
263 hdq_reg_out(hdq_data, OMAP_HDQ_SYSCONFIG, OMAP_HDQ_SYSCONFIG_SOFTRESET);
264 /*
265 * Select HDQ mode & enable clocks.
266 * It is observed that INT flags can't be cleared via a read and GO/INIT
267 * won't return to zero if interrupt is disabled. So we always enable
268 * interrupt.
269 */
270 hdq_reg_out(hdq_data, OMAP_HDQ_CTRL_STATUS,
271 OMAP_HDQ_CTRL_STATUS_CLOCKENABLE |
272 OMAP_HDQ_CTRL_STATUS_INTERRUPTMASK);
273
274 /* wait for reset to complete */
275 ret = hdq_wait_for_flag(hdq_data, OMAP_HDQ_SYSSTATUS,
276 OMAP_HDQ_SYSSTATUS_RESETDONE, OMAP_HDQ_FLAG_SET, &tmp_status);
277 if (ret)
278 dev_dbg(hdq_data->dev, "timeout waiting HDQ reset, %x",
279 tmp_status);
280 else {
281 hdq_reg_out(hdq_data, OMAP_HDQ_CTRL_STATUS,
282 OMAP_HDQ_CTRL_STATUS_CLOCKENABLE |
283 OMAP_HDQ_CTRL_STATUS_INTERRUPTMASK);
284 hdq_reg_out(hdq_data, OMAP_HDQ_SYSCONFIG,
285 OMAP_HDQ_SYSCONFIG_AUTOIDLE);
286 }
287
288 return ret;
289}
290
291/* Issue break pulse to the device */
292static int omap_hdq_break(struct hdq_data *hdq_data)
293{
294 int ret = 0;
295 u8 tmp_status;
296 unsigned long irqflags;
297
298 ret = mutex_lock_interruptible(&hdq_data->hdq_mutex);
299 if (ret < 0) {
300 dev_dbg(hdq_data->dev, "Could not acquire mutex\n");
301 ret = -EINTR;
302 goto rtn;
303 }
304
305 spin_lock_irqsave(&hdq_data->hdq_spinlock, irqflags);
306 /* clear interrupt flags via a dummy read */
307 hdq_reg_in(hdq_data, OMAP_HDQ_INT_STATUS);
308 /* ISR loads it with new INT_STATUS */
309 hdq_data->hdq_irqstatus = 0;
310 spin_unlock_irqrestore(&hdq_data->hdq_spinlock, irqflags);
311
312 /* set the INIT and GO bit */
313 hdq_reg_merge(hdq_data, OMAP_HDQ_CTRL_STATUS,
314 OMAP_HDQ_CTRL_STATUS_INITIALIZATION | OMAP_HDQ_CTRL_STATUS_GO,
315 OMAP_HDQ_CTRL_STATUS_DIR | OMAP_HDQ_CTRL_STATUS_INITIALIZATION |
316 OMAP_HDQ_CTRL_STATUS_GO);
317
318 /* wait for the TIMEOUT bit */
319 ret = wait_event_timeout(hdq_wait_queue,
320 hdq_data->hdq_irqstatus, OMAP_HDQ_TIMEOUT);
321 if (ret == 0) {
322 dev_dbg(hdq_data->dev, "break wait elapsed\n");
323 ret = -EINTR;
324 goto out;
325 }
326
327 tmp_status = hdq_data->hdq_irqstatus;
328 /* check irqstatus */
329 if (!(tmp_status & OMAP_HDQ_INT_STATUS_TIMEOUT)) {
330 dev_dbg(hdq_data->dev, "timeout waiting for TIMEOUT, %x",
331 tmp_status);
332 ret = -ETIMEDOUT;
333 goto out;
334 }
335 /*
336 * wait for both INIT and GO bits rerurn to zero.
337 * zero wait time expected for interrupt mode.
338 */
339 ret = hdq_wait_for_flag(hdq_data, OMAP_HDQ_CTRL_STATUS,
340 OMAP_HDQ_CTRL_STATUS_INITIALIZATION |
341 OMAP_HDQ_CTRL_STATUS_GO, OMAP_HDQ_FLAG_CLEAR,
342 &tmp_status);
343 if (ret)
344 dev_dbg(hdq_data->dev, "timeout waiting INIT&GO bits"
NeilBrown7b5362a2012-05-22 09:43:02 +1000345 " return to zero, %x", tmp_status);
Madhusudhan Chikkature9f2bc792008-11-12 13:27:09 -0800346
347out:
348 mutex_unlock(&hdq_data->hdq_mutex);
349rtn:
350 return ret;
351}
352
353static int hdq_read_byte(struct hdq_data *hdq_data, u8 *val)
354{
355 int ret = 0;
356 u8 status;
357 unsigned long timeout = jiffies + OMAP_HDQ_TIMEOUT;
358
359 ret = mutex_lock_interruptible(&hdq_data->hdq_mutex);
360 if (ret < 0) {
361 ret = -EINTR;
362 goto rtn;
363 }
364
365 if (!hdq_data->hdq_usecount) {
366 ret = -EINVAL;
367 goto out;
368 }
369
370 if (!(hdq_data->hdq_irqstatus & OMAP_HDQ_INT_STATUS_RXCOMPLETE)) {
371 hdq_reg_merge(hdq_data, OMAP_HDQ_CTRL_STATUS,
372 OMAP_HDQ_CTRL_STATUS_DIR | OMAP_HDQ_CTRL_STATUS_GO,
373 OMAP_HDQ_CTRL_STATUS_DIR | OMAP_HDQ_CTRL_STATUS_GO);
374 /*
375 * The RX comes immediately after TX. It
376 * triggers another interrupt before we
377 * sleep. So we have to wait for RXCOMPLETE bit.
378 */
379 while (!(hdq_data->hdq_irqstatus
380 & OMAP_HDQ_INT_STATUS_RXCOMPLETE)
381 && time_before(jiffies, timeout)) {
382 schedule_timeout_uninterruptible(1);
383 }
384 hdq_reg_merge(hdq_data, OMAP_HDQ_CTRL_STATUS, 0,
385 OMAP_HDQ_CTRL_STATUS_DIR);
386 status = hdq_data->hdq_irqstatus;
387 /* check irqstatus */
388 if (!(status & OMAP_HDQ_INT_STATUS_RXCOMPLETE)) {
389 dev_dbg(hdq_data->dev, "timeout waiting for"
NeilBrown7b5362a2012-05-22 09:43:02 +1000390 " RXCOMPLETE, %x", status);
Madhusudhan Chikkature9f2bc792008-11-12 13:27:09 -0800391 ret = -ETIMEDOUT;
392 goto out;
393 }
394 }
395 /* the data is ready. Read it in! */
396 *val = hdq_reg_in(hdq_data, OMAP_HDQ_RX_DATA);
397out:
398 mutex_unlock(&hdq_data->hdq_mutex);
399rtn:
NeilBrown7b5362a2012-05-22 09:43:02 +1000400 return ret;
Madhusudhan Chikkature9f2bc792008-11-12 13:27:09 -0800401
402}
403
404/* Enable clocks and set the controller to HDQ mode */
405static int omap_hdq_get(struct hdq_data *hdq_data)
406{
407 int ret = 0;
408
409 ret = mutex_lock_interruptible(&hdq_data->hdq_mutex);
410 if (ret < 0) {
411 ret = -EINTR;
412 goto rtn;
413 }
414
415 if (OMAP_HDQ_MAX_USER == hdq_data->hdq_usecount) {
416 dev_dbg(hdq_data->dev, "attempt to exceed the max use count");
417 ret = -EINVAL;
418 goto out;
419 } else {
420 hdq_data->hdq_usecount++;
421 try_module_get(THIS_MODULE);
422 if (1 == hdq_data->hdq_usecount) {
423 if (clk_enable(hdq_data->hdq_ick)) {
424 dev_dbg(hdq_data->dev, "Can not enable ick\n");
425 ret = -ENODEV;
426 goto clk_err;
427 }
428 if (clk_enable(hdq_data->hdq_fck)) {
429 dev_dbg(hdq_data->dev, "Can not enable fck\n");
430 clk_disable(hdq_data->hdq_ick);
431 ret = -ENODEV;
432 goto clk_err;
433 }
434
435 /* make sure HDQ is out of reset */
436 if (!(hdq_reg_in(hdq_data, OMAP_HDQ_SYSSTATUS) &
437 OMAP_HDQ_SYSSTATUS_RESETDONE)) {
438 ret = _omap_hdq_reset(hdq_data);
439 if (ret)
440 /* back up the count */
441 hdq_data->hdq_usecount--;
442 } else {
443 /* select HDQ mode & enable clocks */
444 hdq_reg_out(hdq_data, OMAP_HDQ_CTRL_STATUS,
445 OMAP_HDQ_CTRL_STATUS_CLOCKENABLE |
446 OMAP_HDQ_CTRL_STATUS_INTERRUPTMASK);
447 hdq_reg_out(hdq_data, OMAP_HDQ_SYSCONFIG,
448 OMAP_HDQ_SYSCONFIG_AUTOIDLE);
449 hdq_reg_in(hdq_data, OMAP_HDQ_INT_STATUS);
450 }
451 }
452 }
453
454clk_err:
455 clk_put(hdq_data->hdq_ick);
456 clk_put(hdq_data->hdq_fck);
457out:
458 mutex_unlock(&hdq_data->hdq_mutex);
459rtn:
460 return ret;
461}
462
463/* Disable clocks to the module */
464static int omap_hdq_put(struct hdq_data *hdq_data)
465{
466 int ret = 0;
467
468 ret = mutex_lock_interruptible(&hdq_data->hdq_mutex);
469 if (ret < 0)
470 return -EINTR;
471
472 if (0 == hdq_data->hdq_usecount) {
473 dev_dbg(hdq_data->dev, "attempt to decrement use count"
NeilBrown7b5362a2012-05-22 09:43:02 +1000474 " when it is zero");
Madhusudhan Chikkature9f2bc792008-11-12 13:27:09 -0800475 ret = -EINVAL;
476 } else {
477 hdq_data->hdq_usecount--;
478 module_put(THIS_MODULE);
479 if (0 == hdq_data->hdq_usecount) {
480 clk_disable(hdq_data->hdq_ick);
481 clk_disable(hdq_data->hdq_fck);
482 }
483 }
484 mutex_unlock(&hdq_data->hdq_mutex);
485
486 return ret;
487}
488
489/* Read a byte of data from the device */
490static u8 omap_w1_read_byte(void *_hdq)
491{
492 struct hdq_data *hdq_data = _hdq;
493 u8 val = 0;
494 int ret;
495
496 ret = hdq_read_byte(hdq_data, &val);
497 if (ret) {
498 ret = mutex_lock_interruptible(&hdq_data->hdq_mutex);
499 if (ret < 0) {
500 dev_dbg(hdq_data->dev, "Could not acquire mutex\n");
501 return -EINTR;
502 }
503 hdq_data->init_trans = 0;
504 mutex_unlock(&hdq_data->hdq_mutex);
505 omap_hdq_put(hdq_data);
506 return -1;
507 }
508
509 /* Write followed by a read, release the module */
510 if (hdq_data->init_trans) {
511 ret = mutex_lock_interruptible(&hdq_data->hdq_mutex);
512 if (ret < 0) {
513 dev_dbg(hdq_data->dev, "Could not acquire mutex\n");
514 return -EINTR;
515 }
516 hdq_data->init_trans = 0;
517 mutex_unlock(&hdq_data->hdq_mutex);
518 omap_hdq_put(hdq_data);
519 }
520
521 return val;
522}
523
524/* Write a byte of data to the device */
525static void omap_w1_write_byte(void *_hdq, u8 byte)
526{
527 struct hdq_data *hdq_data = _hdq;
528 int ret;
529 u8 status;
530
531 /* First write to initialize the transfer */
532 if (hdq_data->init_trans == 0)
533 omap_hdq_get(hdq_data);
534
535 ret = mutex_lock_interruptible(&hdq_data->hdq_mutex);
536 if (ret < 0) {
537 dev_dbg(hdq_data->dev, "Could not acquire mutex\n");
538 return;
539 }
540 hdq_data->init_trans++;
541 mutex_unlock(&hdq_data->hdq_mutex);
542
543 ret = hdq_write_byte(hdq_data, byte, &status);
NeilBrown7b5362a2012-05-22 09:43:02 +1000544 if (ret < 0) {
Madhusudhan Chikkature9f2bc792008-11-12 13:27:09 -0800545 dev_dbg(hdq_data->dev, "TX failure:Ctrl status %x\n", status);
546 return;
547 }
548
Lucas De Marchi25985ed2011-03-30 22:57:33 -0300549 /* Second write, data transferred. Release the module */
Madhusudhan Chikkature9f2bc792008-11-12 13:27:09 -0800550 if (hdq_data->init_trans > 1) {
551 omap_hdq_put(hdq_data);
552 ret = mutex_lock_interruptible(&hdq_data->hdq_mutex);
553 if (ret < 0) {
554 dev_dbg(hdq_data->dev, "Could not acquire mutex\n");
555 return;
556 }
557 hdq_data->init_trans = 0;
558 mutex_unlock(&hdq_data->hdq_mutex);
559 }
560
561 return;
562}
563
Uwe Kleine-Königa96b9122010-02-04 20:56:54 +0100564static int __devinit omap_hdq_probe(struct platform_device *pdev)
Madhusudhan Chikkature9f2bc792008-11-12 13:27:09 -0800565{
566 struct hdq_data *hdq_data;
567 struct resource *res;
568 int ret, irq;
569 u8 rev;
570
571 hdq_data = kmalloc(sizeof(*hdq_data), GFP_KERNEL);
572 if (!hdq_data) {
573 dev_dbg(&pdev->dev, "unable to allocate memory\n");
574 ret = -ENOMEM;
575 goto err_kmalloc;
576 }
577
578 hdq_data->dev = &pdev->dev;
579 platform_set_drvdata(pdev, hdq_data);
580
581 res = platform_get_resource(pdev, IORESOURCE_MEM, 0);
582 if (!res) {
583 dev_dbg(&pdev->dev, "unable to get resource\n");
584 ret = -ENXIO;
585 goto err_resource;
586 }
587
588 hdq_data->hdq_base = ioremap(res->start, SZ_4K);
589 if (!hdq_data->hdq_base) {
590 dev_dbg(&pdev->dev, "ioremap failed\n");
591 ret = -EINVAL;
592 goto err_ioremap;
593 }
594
595 /* get interface & functional clock objects */
Russell Kingcc51c9d2009-01-22 10:12:04 +0000596 hdq_data->hdq_ick = clk_get(&pdev->dev, "ick");
Julia Lawall80d02d22011-02-10 15:01:37 -0800597 if (IS_ERR(hdq_data->hdq_ick)) {
598 dev_dbg(&pdev->dev, "Can't get HDQ ick clock object\n");
599 ret = PTR_ERR(hdq_data->hdq_ick);
600 goto err_ick;
601 }
Madhusudhan Chikkature9f2bc792008-11-12 13:27:09 -0800602
Julia Lawall80d02d22011-02-10 15:01:37 -0800603 hdq_data->hdq_fck = clk_get(&pdev->dev, "fck");
604 if (IS_ERR(hdq_data->hdq_fck)) {
605 dev_dbg(&pdev->dev, "Can't get HDQ fck clock object\n");
606 ret = PTR_ERR(hdq_data->hdq_fck);
607 goto err_fck;
Madhusudhan Chikkature9f2bc792008-11-12 13:27:09 -0800608 }
609
610 hdq_data->hdq_usecount = 0;
611 mutex_init(&hdq_data->hdq_mutex);
612
613 if (clk_enable(hdq_data->hdq_ick)) {
614 dev_dbg(&pdev->dev, "Can not enable ick\n");
615 ret = -ENODEV;
616 goto err_intfclk;
617 }
618
619 if (clk_enable(hdq_data->hdq_fck)) {
620 dev_dbg(&pdev->dev, "Can not enable fck\n");
621 ret = -ENODEV;
622 goto err_fnclk;
623 }
624
625 rev = hdq_reg_in(hdq_data, OMAP_HDQ_REVISION);
626 dev_info(&pdev->dev, "OMAP HDQ Hardware Rev %c.%c. Driver in %s mode\n",
627 (rev >> 4) + '0', (rev & 0x0f) + '0', "Interrupt");
628
629 spin_lock_init(&hdq_data->hdq_spinlock);
630
631 irq = platform_get_irq(pdev, 0);
632 if (irq < 0) {
633 ret = -ENXIO;
634 goto err_irq;
635 }
636
637 ret = request_irq(irq, hdq_isr, IRQF_DISABLED, "omap_hdq", hdq_data);
638 if (ret < 0) {
639 dev_dbg(&pdev->dev, "could not request irq\n");
640 goto err_irq;
641 }
642
643 omap_hdq_break(hdq_data);
644
645 /* don't clock the HDQ until it is needed */
646 clk_disable(hdq_data->hdq_ick);
647 clk_disable(hdq_data->hdq_fck);
648
649 omap_w1_master.data = hdq_data;
650
651 ret = w1_add_master_device(&omap_w1_master);
652 if (ret) {
653 dev_dbg(&pdev->dev, "Failure in registering w1 master\n");
654 goto err_w1;
655 }
656
657 return 0;
658
659err_w1:
660err_irq:
661 clk_disable(hdq_data->hdq_fck);
662
663err_fnclk:
664 clk_disable(hdq_data->hdq_ick);
665
666err_intfclk:
Madhusudhan Chikkature9f2bc792008-11-12 13:27:09 -0800667 clk_put(hdq_data->hdq_fck);
668
Julia Lawall80d02d22011-02-10 15:01:37 -0800669err_fck:
670 clk_put(hdq_data->hdq_ick);
671
672err_ick:
Madhusudhan Chikkature9f2bc792008-11-12 13:27:09 -0800673 iounmap(hdq_data->hdq_base);
674
675err_ioremap:
676err_resource:
677 platform_set_drvdata(pdev, NULL);
678 kfree(hdq_data);
679
680err_kmalloc:
681 return ret;
682
683}
684
685static int omap_hdq_remove(struct platform_device *pdev)
686{
687 struct hdq_data *hdq_data = platform_get_drvdata(pdev);
688
689 mutex_lock(&hdq_data->hdq_mutex);
690
691 if (hdq_data->hdq_usecount) {
692 dev_dbg(&pdev->dev, "removed when use count is not zero\n");
Stoyan Gaydarov20200022009-08-06 15:07:28 -0700693 mutex_unlock(&hdq_data->hdq_mutex);
Madhusudhan Chikkature9f2bc792008-11-12 13:27:09 -0800694 return -EBUSY;
695 }
696
697 mutex_unlock(&hdq_data->hdq_mutex);
698
699 /* remove module dependency */
700 clk_put(hdq_data->hdq_ick);
701 clk_put(hdq_data->hdq_fck);
702 free_irq(INT_24XX_HDQ_IRQ, hdq_data);
703 platform_set_drvdata(pdev, NULL);
704 iounmap(hdq_data->hdq_base);
705 kfree(hdq_data);
706
707 return 0;
708}
709
710static int __init
711omap_hdq_init(void)
712{
713 return platform_driver_register(&omap_hdq_driver);
714}
715module_init(omap_hdq_init);
716
717static void __exit
718omap_hdq_exit(void)
719{
720 platform_driver_unregister(&omap_hdq_driver);
721}
722module_exit(omap_hdq_exit);
723
724module_param(w1_id, int, S_IRUSR);
725MODULE_PARM_DESC(w1_id, "1-wire id for the slave detection");
726
727MODULE_AUTHOR("Texas Instruments");
728MODULE_DESCRIPTION("HDQ driver Library");
729MODULE_LICENSE("GPL");