blob: 76781cb69ec64b05fa17f319c82567cd368375cc [file] [log] [blame]
Brian Swetland2eb44eb2008-09-29 16:00:48 -07001/* arch/arm/mach-msm/smd.c
2 *
3 * Copyright (C) 2007 Google, Inc.
Bryan Huntsman3f2bc4d2011-08-16 17:27:22 -07004 * Copyright (c) 2008-2011, Code Aurora Forum. All rights reserved.
Brian Swetland2eb44eb2008-09-29 16:00:48 -07005 * Author: Brian Swetland <swetland@google.com>
6 *
7 * This software is licensed under the terms of the GNU General Public
8 * License version 2, as published by the Free Software Foundation, and
9 * may be copied, distributed, and modified under those terms.
10 *
11 * This program is distributed in the hope that it will be useful,
12 * but WITHOUT ANY WARRANTY; without even the implied warranty of
13 * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
14 * GNU General Public License for more details.
15 *
16 */
17
18#include <linux/platform_device.h>
19#include <linux/module.h>
20#include <linux/fs.h>
21#include <linux/cdev.h>
22#include <linux/device.h>
23#include <linux/wait.h>
24#include <linux/interrupt.h>
25#include <linux/irq.h>
26#include <linux/list.h>
27#include <linux/slab.h>
Brian Swetland2eb44eb2008-09-29 16:00:48 -070028#include <linux/delay.h>
Bryan Huntsman3f2bc4d2011-08-16 17:27:22 -070029#include <linux/io.h>
30#include <linux/termios.h>
31#include <linux/ctype.h>
32#include <linux/remote_spinlock.h>
33#include <linux/uaccess.h>
Brian Swetland2eb44eb2008-09-29 16:00:48 -070034#include <mach/msm_smd.h>
Bryan Huntsman3f2bc4d2011-08-16 17:27:22 -070035#include <mach/msm_iomap.h>
Brian Swetland2eb44eb2008-09-29 16:00:48 -070036#include <mach/system.h>
Bryan Huntsman3f2bc4d2011-08-16 17:27:22 -070037#include <mach/subsystem_notif.h>
Angshuman Sarkaread67bd2011-09-21 20:13:12 +053038#include <mach/socinfo.h>
Brian Swetland2eb44eb2008-09-29 16:00:48 -070039
40#include "smd_private.h"
41#include "proc_comm.h"
Bryan Huntsman3f2bc4d2011-08-16 17:27:22 -070042#include "modem_notifier.h"
Brian Swetland2eb44eb2008-09-29 16:00:48 -070043
Bryan Huntsman3f2bc4d2011-08-16 17:27:22 -070044#if defined(CONFIG_ARCH_QSD8X50) || defined(CONFIG_ARCH_MSM8X60) \
Jeff Hugo56b933a2011-09-28 14:42:05 -060045 || defined(CONFIG_ARCH_MSM8960) || defined(CONFIG_ARCH_FSM9XXX) \
Jeff Hugo0c0f5e92011-09-28 13:55:45 -060046 || defined(CONFIG_ARCH_MSM9615) || defined(CONFIG_ARCH_APQ8064)
Brian Swetland37521a32009-07-01 18:30:47 -070047#define CONFIG_QDSP6 1
48#endif
49
Jeff Hugo0c0f5e92011-09-28 13:55:45 -060050#if defined(CONFIG_ARCH_MSM8X60) || defined(CONFIG_ARCH_MSM8960) \
51 || defined(CONFIG_ARCH_APQ8064)
Bryan Huntsman3f2bc4d2011-08-16 17:27:22 -070052#define CONFIG_DSPS 1
53#endif
54
Jeff Hugo0c0f5e92011-09-28 13:55:45 -060055#if defined(CONFIG_ARCH_MSM8960) \
56 || defined(CONFIG_ARCH_APQ8064)
Bryan Huntsman3f2bc4d2011-08-16 17:27:22 -070057#define CONFIG_WCNSS 1
Jeff Hugo6a8057c2011-08-16 13:47:12 -060058#define CONFIG_DSPS_SMSM 1
Bryan Huntsman3f2bc4d2011-08-16 17:27:22 -070059#endif
Brian Swetland2eb44eb2008-09-29 16:00:48 -070060
61#define MODULE_NAME "msm_smd"
Bryan Huntsman3f2bc4d2011-08-16 17:27:22 -070062#define SMEM_VERSION 0x000B
63#define SMD_VERSION 0x00020000
64
65uint32_t SMSM_NUM_ENTRIES = 8;
66uint32_t SMSM_NUM_HOSTS = 3;
Brian Swetland2eb44eb2008-09-29 16:00:48 -070067
68enum {
69 MSM_SMD_DEBUG = 1U << 0,
Bryan Huntsman3f2bc4d2011-08-16 17:27:22 -070070 MSM_SMSM_DEBUG = 1U << 1,
71 MSM_SMD_INFO = 1U << 2,
72 MSM_SMSM_INFO = 1U << 3,
73};
74
75struct smsm_shared_info {
76 uint32_t *state;
77 uint32_t *intr_mask;
78 uint32_t *intr_mux;
79};
80
81static struct smsm_shared_info smsm_info;
82
83struct smsm_size_info_type {
84 uint32_t num_hosts;
85 uint32_t num_entries;
86 uint32_t reserved0;
87 uint32_t reserved1;
88};
89
90struct smsm_state_cb_info {
91 struct list_head cb_list;
92 uint32_t mask;
93 void *data;
94 void (*notify)(void *data, uint32_t old_state, uint32_t new_state);
95};
96
97struct smsm_state_info {
98 struct list_head callbacks;
99 uint32_t last_value;
100};
101
102#define SMSM_STATE_ADDR(entry) (smsm_info.state + entry)
103#define SMSM_INTR_MASK_ADDR(entry, host) (smsm_info.intr_mask + \
104 entry * SMSM_NUM_HOSTS + host)
105#define SMSM_INTR_MUX_ADDR(entry) (smsm_info.intr_mux + entry)
106
107/* Internal definitions which are not exported in some targets */
108enum {
109 SMSM_APPS_DEM_I = 3,
Brian Swetland2eb44eb2008-09-29 16:00:48 -0700110};
111
112static int msm_smd_debug_mask;
Brian Swetland2eb44eb2008-09-29 16:00:48 -0700113module_param_named(debug_mask, msm_smd_debug_mask,
114 int, S_IRUGO | S_IWUSR | S_IWGRP);
115
Bryan Huntsman3f2bc4d2011-08-16 17:27:22 -0700116#if defined(CONFIG_MSM_SMD_DEBUG)
117#define SMD_DBG(x...) do { \
118 if (msm_smd_debug_mask & MSM_SMD_DEBUG) \
119 printk(KERN_DEBUG x); \
120 } while (0)
121
122#define SMSM_DBG(x...) do { \
123 if (msm_smd_debug_mask & MSM_SMSM_DEBUG) \
124 printk(KERN_DEBUG x); \
125 } while (0)
126
127#define SMD_INFO(x...) do { \
128 if (msm_smd_debug_mask & MSM_SMD_INFO) \
129 printk(KERN_INFO x); \
130 } while (0)
131
132#define SMSM_INFO(x...) do { \
133 if (msm_smd_debug_mask & MSM_SMSM_INFO) \
134 printk(KERN_INFO x); \
135 } while (0)
136#else
137#define SMD_DBG(x...) do { } while (0)
138#define SMSM_DBG(x...) do { } while (0)
139#define SMD_INFO(x...) do { } while (0)
140#define SMSM_INFO(x...) do { } while (0)
141#endif
142
Brian Swetland2eb44eb2008-09-29 16:00:48 -0700143static unsigned last_heap_free = 0xffffffff;
144
Bryan Huntsman3f2bc4d2011-08-16 17:27:22 -0700145static inline void smd_write_intr(unsigned int val,
146 const void __iomem *addr);
147
148#if defined(CONFIG_ARCH_MSM7X30)
149#define MSM_TRIG_A2M_SMD_INT \
150 (smd_write_intr(1 << 0, MSM_GCC_BASE + 0x8))
151#define MSM_TRIG_A2Q6_SMD_INT \
152 (smd_write_intr(1 << 8, MSM_GCC_BASE + 0x8))
153#define MSM_TRIG_A2M_SMSM_INT \
154 (smd_write_intr(1 << 5, MSM_GCC_BASE + 0x8))
155#define MSM_TRIG_A2Q6_SMSM_INT \
156 (smd_write_intr(1 << 8, MSM_GCC_BASE + 0x8))
157#define MSM_TRIG_A2DSPS_SMD_INT
Jeff Hugo6a8057c2011-08-16 13:47:12 -0600158#define MSM_TRIG_A2DSPS_SMSM_INT
Bryan Huntsman3f2bc4d2011-08-16 17:27:22 -0700159#define MSM_TRIG_A2WCNSS_SMD_INT
160#define MSM_TRIG_A2WCNSS_SMSM_INT
161#elif defined(CONFIG_ARCH_MSM8X60)
162#define MSM_TRIG_A2M_SMD_INT \
163 (smd_write_intr(1 << 3, MSM_GCC_BASE + 0x8))
164#define MSM_TRIG_A2Q6_SMD_INT \
165 (smd_write_intr(1 << 15, MSM_GCC_BASE + 0x8))
166#define MSM_TRIG_A2M_SMSM_INT \
167 (smd_write_intr(1 << 4, MSM_GCC_BASE + 0x8))
168#define MSM_TRIG_A2Q6_SMSM_INT \
169 (smd_write_intr(1 << 14, MSM_GCC_BASE + 0x8))
170#define MSM_TRIG_A2DSPS_SMD_INT \
171 (smd_write_intr(1, MSM_SIC_NON_SECURE_BASE + 0x4080))
Jeff Hugo6a8057c2011-08-16 13:47:12 -0600172#define MSM_TRIG_A2DSPS_SMSM_INT
Bryan Huntsman3f2bc4d2011-08-16 17:27:22 -0700173#define MSM_TRIG_A2WCNSS_SMD_INT
174#define MSM_TRIG_A2WCNSS_SMSM_INT
Stepan Moskovchenko0dd0eba2011-10-20 18:20:59 -0700175#elif defined(CONFIG_ARCH_MSM8960) || defined(CONFIG_ARCH_MSM8930) || \
176 defined(CONFIG_ARCH_APQ8064)
Bryan Huntsman3f2bc4d2011-08-16 17:27:22 -0700177#define MSM_TRIG_A2M_SMD_INT \
178 (smd_write_intr(1 << 3, MSM_APCS_GCC_BASE + 0x8))
179#define MSM_TRIG_A2Q6_SMD_INT \
180 (smd_write_intr(1 << 15, MSM_APCS_GCC_BASE + 0x8))
181#define MSM_TRIG_A2M_SMSM_INT \
182 (smd_write_intr(1 << 4, MSM_APCS_GCC_BASE + 0x8))
183#define MSM_TRIG_A2Q6_SMSM_INT \
184 (smd_write_intr(1 << 14, MSM_APCS_GCC_BASE + 0x8))
185#define MSM_TRIG_A2DSPS_SMD_INT \
186 (smd_write_intr(1, MSM_SIC_NON_SECURE_BASE + 0x4080))
Jeff Hugo6a8057c2011-08-16 13:47:12 -0600187#define MSM_TRIG_A2DSPS_SMSM_INT \
188 (smd_write_intr(1, MSM_SIC_NON_SECURE_BASE + 0x4094))
Bryan Huntsman3f2bc4d2011-08-16 17:27:22 -0700189#define MSM_TRIG_A2WCNSS_SMD_INT \
190 (smd_write_intr(1 << 25, MSM_APCS_GCC_BASE + 0x8))
191#define MSM_TRIG_A2WCNSS_SMSM_INT \
192 (smd_write_intr(1 << 23, MSM_APCS_GCC_BASE + 0x8))
Jeff Hugo56b933a2011-09-28 14:42:05 -0600193#elif defined(CONFIG_ARCH_MSM9615)
194#define MSM_TRIG_A2M_SMD_INT \
195 (smd_write_intr(1 << 3, MSM_APCS_GCC_BASE + 0x8))
196#define MSM_TRIG_A2Q6_SMD_INT \
197 (smd_write_intr(1 << 15, MSM_APCS_GCC_BASE + 0x8))
198#define MSM_TRIG_A2M_SMSM_INT \
199 (smd_write_intr(1 << 4, MSM_APCS_GCC_BASE + 0x8))
200#define MSM_TRIG_A2Q6_SMSM_INT \
201 (smd_write_intr(1 << 14, MSM_APCS_GCC_BASE + 0x8))
202#define MSM_TRIG_A2DSPS_SMD_INT
203#define MSM_TRIG_A2DSPS_SMSM_INT
204#define MSM_TRIG_A2WCNSS_SMD_INT
205#define MSM_TRIG_A2WCNSS_SMSM_INT
Bryan Huntsman3f2bc4d2011-08-16 17:27:22 -0700206#elif defined(CONFIG_ARCH_FSM9XXX)
207#define MSM_TRIG_A2Q6_SMD_INT \
208 (smd_write_intr(1 << 10, MSM_GCC_BASE + 0x8))
209#define MSM_TRIG_A2Q6_SMSM_INT \
210 (smd_write_intr(1 << 10, MSM_GCC_BASE + 0x8))
211#define MSM_TRIG_A2M_SMD_INT \
212 (smd_write_intr(1 << 0, MSM_GCC_BASE + 0x8))
213#define MSM_TRIG_A2M_SMSM_INT \
214 (smd_write_intr(1 << 5, MSM_GCC_BASE + 0x8))
215#define MSM_TRIG_A2DSPS_SMD_INT
Jeff Hugo6a8057c2011-08-16 13:47:12 -0600216#define MSM_TRIG_A2DSPS_SMSM_INT
Bryan Huntsman3f2bc4d2011-08-16 17:27:22 -0700217#define MSM_TRIG_A2WCNSS_SMD_INT
218#define MSM_TRIG_A2WCNSS_SMSM_INT
219#else
220#define MSM_TRIG_A2M_SMD_INT \
221 (smd_write_intr(1, MSM_CSR_BASE + 0x400 + (0) * 4))
222#define MSM_TRIG_A2Q6_SMD_INT \
223 (smd_write_intr(1, MSM_CSR_BASE + 0x400 + (8) * 4))
224#define MSM_TRIG_A2M_SMSM_INT \
225 (smd_write_intr(1, MSM_CSR_BASE + 0x400 + (5) * 4))
226#define MSM_TRIG_A2Q6_SMSM_INT \
227 (smd_write_intr(1, MSM_CSR_BASE + 0x400 + (8) * 4))
228#define MSM_TRIG_A2DSPS_SMD_INT
Jeff Hugo6a8057c2011-08-16 13:47:12 -0600229#define MSM_TRIG_A2DSPS_SMSM_INT
Bryan Huntsman3f2bc4d2011-08-16 17:27:22 -0700230#define MSM_TRIG_A2WCNSS_SMD_INT
231#define MSM_TRIG_A2WCNSS_SMSM_INT
Brian Swetland37521a32009-07-01 18:30:47 -0700232#endif
Bryan Huntsman3f2bc4d2011-08-16 17:27:22 -0700233
234#define SMD_LOOPBACK_CID 100
235
Eric Holmbergf6d7d1a2011-09-23 18:31:04 -0600236#define SMEM_SPINLOCK_SMEM_ALLOC "S:3"
237static remote_spinlock_t remote_spinlock;
238
Bryan Huntsman3f2bc4d2011-08-16 17:27:22 -0700239static LIST_HEAD(smd_ch_list_loopback);
240static irqreturn_t smsm_irq_handler(int irq, void *data);
241static void smd_fake_irq_handler(unsigned long arg);
242
243static void notify_smsm_cb_clients_worker(struct work_struct *work);
244static DECLARE_WORK(smsm_cb_work, notify_smsm_cb_clients_worker);
Eric Holmbergc8002902011-09-16 13:55:57 -0600245static DEFINE_MUTEX(smsm_lock);
Bryan Huntsman3f2bc4d2011-08-16 17:27:22 -0700246static struct smsm_state_info *smsm_states;
Angshuman Sarkar7ee0dca2011-08-22 21:37:34 +0530247static int spinlocks_initialized;
Bryan Huntsman3f2bc4d2011-08-16 17:27:22 -0700248
249static inline void smd_write_intr(unsigned int val,
250 const void __iomem *addr)
251{
252 wmb();
253 __raw_writel(val, addr);
254}
255
256#ifdef CONFIG_WCNSS
257static inline void wakeup_v1_riva(void)
258{
259 /*
260 * workaround hack for RIVA v1 hardware bug
261 * trigger GPIO 40 to wake up RIVA from power collaspe
262 * not to be sent to customers
263 */
Jeff Hugoafb8c4a2011-10-27 15:57:27 -0600264 if (SOCINFO_VERSION_MAJOR(socinfo_get_version()) == 1) {
265 __raw_writel(0x0, MSM_TLMM_BASE + 0x1284);
266 __raw_writel(0x2, MSM_TLMM_BASE + 0x1284);
267 }
Bryan Huntsman3f2bc4d2011-08-16 17:27:22 -0700268 /* end workaround */
269}
270#else
271static inline void wakeup_v1_riva(void) {}
272#endif
273
274static void notify_other_smsm(uint32_t smsm_entry, uint32_t notify_mask)
275{
276 /* older protocol don't use smsm_intr_mask,
277 but still communicates with modem */
278 if (!smsm_info.intr_mask ||
279 (__raw_readl(SMSM_INTR_MASK_ADDR(smsm_entry, SMSM_MODEM))
280 & notify_mask))
281 MSM_TRIG_A2M_SMSM_INT;
282
283 if (smsm_info.intr_mask &&
284 (__raw_readl(SMSM_INTR_MASK_ADDR(smsm_entry, SMSM_Q6))
285 & notify_mask)) {
Bryan Huntsman3f2bc4d2011-08-16 17:27:22 -0700286 uint32_t mux_val;
287
Eric Holmberg6282c5d2011-10-27 17:30:57 -0600288 if (cpu_is_qsd8x50() && smsm_info.intr_mux) {
Bryan Huntsman3f2bc4d2011-08-16 17:27:22 -0700289 mux_val = __raw_readl(
290 SMSM_INTR_MUX_ADDR(SMEM_APPS_Q6_SMSM));
291 mux_val++;
292 __raw_writel(mux_val,
293 SMSM_INTR_MUX_ADDR(SMEM_APPS_Q6_SMSM));
294 }
Bryan Huntsman3f2bc4d2011-08-16 17:27:22 -0700295 MSM_TRIG_A2Q6_SMSM_INT;
296 }
297
298 if (smsm_info.intr_mask &&
299 (__raw_readl(SMSM_INTR_MASK_ADDR(smsm_entry, SMSM_WCNSS))
300 & notify_mask)) {
301 wakeup_v1_riva();
302 MSM_TRIG_A2WCNSS_SMSM_INT;
303 }
304
Jeff Hugo6a8057c2011-08-16 13:47:12 -0600305 if (smsm_info.intr_mask &&
306 (__raw_readl(SMSM_INTR_MASK_ADDR(smsm_entry, SMSM_DSPS))
307 & notify_mask)) {
308 MSM_TRIG_A2DSPS_SMSM_INT;
309 }
310
Bryan Huntsman3f2bc4d2011-08-16 17:27:22 -0700311 schedule_work(&smsm_cb_work);
Brian Swetland2eb44eb2008-09-29 16:00:48 -0700312}
313
Brian Swetland5b0f5a32009-04-26 18:38:49 -0700314static inline void notify_modem_smd(void)
Brian Swetland2eb44eb2008-09-29 16:00:48 -0700315{
Bryan Huntsman3f2bc4d2011-08-16 17:27:22 -0700316 MSM_TRIG_A2M_SMD_INT;
Brian Swetland2eb44eb2008-09-29 16:00:48 -0700317}
318
Brian Swetland5b0f5a32009-04-26 18:38:49 -0700319static inline void notify_dsp_smd(void)
320{
Bryan Huntsman3f2bc4d2011-08-16 17:27:22 -0700321 MSM_TRIG_A2Q6_SMD_INT;
Brian Swetland5b0f5a32009-04-26 18:38:49 -0700322}
323
Bryan Huntsman3f2bc4d2011-08-16 17:27:22 -0700324static inline void notify_dsps_smd(void)
325{
326 MSM_TRIG_A2DSPS_SMD_INT;
327}
328
329static inline void notify_wcnss_smd(void)
330{
331 wakeup_v1_riva();
332 MSM_TRIG_A2WCNSS_SMD_INT;
333}
334
335void smd_diag(void)
Brian Swetland2eb44eb2008-09-29 16:00:48 -0700336{
337 char *x;
Bryan Huntsman3f2bc4d2011-08-16 17:27:22 -0700338 int size;
Brian Swetland2eb44eb2008-09-29 16:00:48 -0700339
340 x = smem_find(ID_DIAG_ERR_MSG, SZ_DIAG_ERR_MSG);
341 if (x != 0) {
342 x[SZ_DIAG_ERR_MSG - 1] = 0;
Bryan Huntsman3f2bc4d2011-08-16 17:27:22 -0700343 SMD_INFO("smem: DIAG '%s'\n", x);
344 }
345
346 x = smem_get_entry(SMEM_ERR_CRASH_LOG, &size);
347 if (x != 0) {
348 x[size - 1] = 0;
349 pr_err("smem: CRASH LOG\n'%s'\n", x);
Brian Swetland2eb44eb2008-09-29 16:00:48 -0700350 }
351}
352
Bryan Huntsman3f2bc4d2011-08-16 17:27:22 -0700353
Brian Swetland2eb44eb2008-09-29 16:00:48 -0700354static void handle_modem_crash(void)
355{
Bryan Huntsman3f2bc4d2011-08-16 17:27:22 -0700356 pr_err("MODEM/AMSS has CRASHED\n");
Brian Swetland2eb44eb2008-09-29 16:00:48 -0700357 smd_diag();
358
Bryan Huntsman3f2bc4d2011-08-16 17:27:22 -0700359 /* hard reboot if possible FIXME
360 if (msm_reset_hook)
361 msm_reset_hook();
362 */
Brian Swetland2eb44eb2008-09-29 16:00:48 -0700363
364 /* in this case the modem or watchdog should reboot us */
365 for (;;)
366 ;
367}
368
Bryan Huntsman3f2bc4d2011-08-16 17:27:22 -0700369int smsm_check_for_modem_crash(void)
Arve Hjønnevåg28379412009-05-20 16:52:36 -0700370{
Bryan Huntsman3f2bc4d2011-08-16 17:27:22 -0700371 /* if the modem's not ready yet, we have to hope for the best */
372 if (!smsm_info.state)
373 return 0;
Arve Hjønnevåg28379412009-05-20 16:52:36 -0700374
Bryan Huntsman3f2bc4d2011-08-16 17:27:22 -0700375 if (__raw_readl(SMSM_STATE_ADDR(SMSM_MODEM_STATE)) & SMSM_RESET) {
Brian Swetland2eb44eb2008-09-29 16:00:48 -0700376 handle_modem_crash();
377 return -1;
Brian Swetland2eb44eb2008-09-29 16:00:48 -0700378 }
Brian Swetland5b0f5a32009-04-26 18:38:49 -0700379 return 0;
Brian Swetland2eb44eb2008-09-29 16:00:48 -0700380}
Bryan Huntsman3f2bc4d2011-08-16 17:27:22 -0700381EXPORT_SYMBOL(smsm_check_for_modem_crash);
Brian Swetland2eb44eb2008-09-29 16:00:48 -0700382
Brian Swetland2eb44eb2008-09-29 16:00:48 -0700383/* the spinlock is used to synchronize between the
Brian Swetland03e00cd2009-07-01 17:58:37 -0700384 * irq handler and code that mutates the channel
385 * list or fiddles with channel state
386 */
Bryan Huntsman3f2bc4d2011-08-16 17:27:22 -0700387static DEFINE_SPINLOCK(smd_lock);
Brian Swetland03e00cd2009-07-01 17:58:37 -0700388DEFINE_SPINLOCK(smem_lock);
Brian Swetland2eb44eb2008-09-29 16:00:48 -0700389
390/* the mutex is used during open() and close()
Brian Swetland03e00cd2009-07-01 17:58:37 -0700391 * operations to avoid races while creating or
392 * destroying smd_channel structures
393 */
Brian Swetland2eb44eb2008-09-29 16:00:48 -0700394static DEFINE_MUTEX(smd_creation_mutex);
395
396static int smd_initialized;
397
Bryan Huntsman3f2bc4d2011-08-16 17:27:22 -0700398struct smd_shared_v1 {
399 struct smd_half_channel ch0;
400 unsigned char data0[SMD_BUF_SIZE];
401 struct smd_half_channel ch1;
402 unsigned char data1[SMD_BUF_SIZE];
403};
404
405struct smd_shared_v2 {
406 struct smd_half_channel ch0;
407 struct smd_half_channel ch1;
408};
409
410struct smd_channel {
411 volatile struct smd_half_channel *send;
412 volatile struct smd_half_channel *recv;
413 unsigned char *send_data;
414 unsigned char *recv_data;
415 unsigned fifo_size;
416 unsigned fifo_mask;
417 struct list_head ch_list;
418
419 unsigned current_packet;
420 unsigned n;
421 void *priv;
422 void (*notify)(void *priv, unsigned flags);
423
424 int (*read)(smd_channel_t *ch, void *data, int len, int user_buf);
425 int (*write)(smd_channel_t *ch, const void *data, int len,
426 int user_buf);
427 int (*read_avail)(smd_channel_t *ch);
428 int (*write_avail)(smd_channel_t *ch);
429 int (*read_from_cb)(smd_channel_t *ch, void *data, int len,
430 int user_buf);
431
432 void (*update_state)(smd_channel_t *ch);
433 unsigned last_state;
434 void (*notify_other_cpu)(void);
435
436 char name[20];
437 struct platform_device pdev;
438 unsigned type;
439
440 int pending_pkt_sz;
441
442 char is_pkt_ch;
443};
444
445struct edge_to_pid {
446 uint32_t local_pid;
447 uint32_t remote_pid;
448};
449
Eric Holmberg7dfd1972011-09-09 16:07:57 -0600450/*
451 * SMD Processor ID's.
452 *
453 * For all processors that have both SMSM and SMD clients,
454 * the SMSM Processor ID and the SMD Processor ID will
455 * be the same. In cases where a processor only supports
456 * SMD, the entry will only exist in this enum.
457 */
458enum {
459 SMD_APPS = SMSM_APPS,
460 SMD_MODEM = SMSM_MODEM,
461 SMD_Q6 = SMSM_Q6,
462 SMD_WCNSS = SMSM_WCNSS,
463 SMD_DSPS = SMSM_DSPS,
464 SMD_MODEM_Q6_FW,
465};
466
Bryan Huntsman3f2bc4d2011-08-16 17:27:22 -0700467/**
468 * Maps edge type to local and remote processor ID's.
469 */
470static struct edge_to_pid edge_to_pids[] = {
471 [SMD_APPS_MODEM] = {SMSM_APPS, SMSM_MODEM},
472 [SMD_APPS_QDSP] = {SMSM_APPS, SMSM_Q6},
473 [SMD_MODEM_QDSP] = {SMSM_MODEM, SMSM_Q6},
474 [SMD_APPS_DSPS] = {SMSM_APPS, SMSM_DSPS},
475 [SMD_MODEM_DSPS] = {SMSM_MODEM, SMSM_DSPS},
476 [SMD_QDSP_DSPS] = {SMSM_Q6, SMSM_DSPS},
477 [SMD_APPS_WCNSS] = {SMSM_APPS, SMSM_WCNSS},
478 [SMD_MODEM_WCNSS] = {SMSM_MODEM, SMSM_WCNSS},
479 [SMD_QDSP_WCNSS] = {SMSM_Q6, SMSM_WCNSS},
480 [SMD_DSPS_WCNSS] = {SMSM_DSPS, SMSM_WCNSS},
Eric Holmberg7dfd1972011-09-09 16:07:57 -0600481 [SMD_APPS_Q6FW] = {SMSM_APPS, SMD_MODEM_Q6_FW},
482 [SMD_MODEM_Q6FW] = {SMSM_MODEM, SMD_MODEM_Q6_FW},
483 [SMD_QDSP_Q6FW] = {SMSM_Q6, SMD_MODEM_Q6_FW},
484 [SMD_DSPS_Q6FW] = {SMSM_DSPS, SMD_MODEM_Q6_FW},
485 [SMD_WCNSS_Q6FW] = {SMSM_WCNSS, SMD_MODEM_Q6_FW},
Bryan Huntsman3f2bc4d2011-08-16 17:27:22 -0700486};
487
488struct restart_notifier_block {
489 unsigned processor;
490 char *name;
491 struct notifier_block nb;
492};
493
494static struct platform_device loopback_tty_pdev = {.name = "LOOPBACK_TTY"};
495
496static LIST_HEAD(smd_ch_closed_list);
497static LIST_HEAD(smd_ch_closing_list);
498static LIST_HEAD(smd_ch_to_close_list);
499static LIST_HEAD(smd_ch_list_modem);
500static LIST_HEAD(smd_ch_list_dsp);
501static LIST_HEAD(smd_ch_list_dsps);
502static LIST_HEAD(smd_ch_list_wcnss);
Brian Swetland2eb44eb2008-09-29 16:00:48 -0700503
504static unsigned char smd_ch_allocated[64];
505static struct work_struct probe_work;
506
Bryan Huntsman3f2bc4d2011-08-16 17:27:22 -0700507static void finalize_channel_close_fn(struct work_struct *work);
508static DECLARE_WORK(finalize_channel_close_work, finalize_channel_close_fn);
509static struct workqueue_struct *channel_close_wq;
510
511static int smd_alloc_channel(struct smd_alloc_elm *alloc_elm);
512
513/* on smp systems, the probe might get called from multiple cores,
514 hence use a lock */
515static DEFINE_MUTEX(smd_probe_lock);
516
517static void smd_channel_probe_worker(struct work_struct *work)
518{
519 struct smd_alloc_elm *shared;
520 unsigned n;
521 uint32_t type;
522
523 shared = smem_find(ID_CH_ALLOC_TBL, sizeof(*shared) * 64);
524
525 if (!shared) {
526 pr_err("%s: allocation table not initialized\n", __func__);
527 return;
528 }
529
530 mutex_lock(&smd_probe_lock);
531 for (n = 0; n < 64; n++) {
532 if (smd_ch_allocated[n])
533 continue;
534
535 /* channel should be allocated only if APPS
536 processor is involved */
537 type = SMD_CHANNEL_TYPE(shared[n].type);
538 if ((type != SMD_APPS_MODEM) && (type != SMD_APPS_QDSP) &&
539 (type != SMD_APPS_DSPS) && (type != SMD_APPS_WCNSS))
540 continue;
541 if (!shared[n].ref_count)
542 continue;
543 if (!shared[n].name[0])
544 continue;
545
546 if (!smd_alloc_channel(&shared[n]))
547 smd_ch_allocated[n] = 1;
548 else
549 SMD_INFO("Probe skipping ch %d, not allocated\n", n);
550 }
551 mutex_unlock(&smd_probe_lock);
552}
553
554/**
555 * Lookup processor ID and determine if it belongs to the proved edge
556 * type.
557 *
558 * @shared2: Pointer to v2 shared channel structure
559 * @type: Edge type
560 * @pid: Processor ID of processor on edge
561 * @local_ch: Channel that belongs to processor @pid
562 * @remote_ch: Other side of edge contained @pid
563 *
564 * Returns 0 for not on edge, 1 for found on edge
565 */
566static int pid_is_on_edge(struct smd_shared_v2 *shared2,
567 uint32_t type, uint32_t pid,
568 struct smd_half_channel **local_ch,
569 struct smd_half_channel **remote_ch
570 )
571{
572 int ret = 0;
573 struct edge_to_pid *edge;
574
575 *local_ch = 0;
576 *remote_ch = 0;
577
578 if (!shared2 || (type >= ARRAY_SIZE(edge_to_pids)))
579 return 0;
580
581 edge = &edge_to_pids[type];
582 if (edge->local_pid != edge->remote_pid) {
583 if (pid == edge->local_pid) {
584 *local_ch = &shared2->ch0;
585 *remote_ch = &shared2->ch1;
586 ret = 1;
587 } else if (pid == edge->remote_pid) {
588 *local_ch = &shared2->ch1;
589 *remote_ch = &shared2->ch0;
590 ret = 1;
591 }
592 }
593
594 return ret;
595}
596
Eric Holmberg2a563c32011-10-05 14:51:43 -0600597static void smd_reset_edge(struct smd_half_channel *ch, unsigned new_state)
598{
599 if (ch->state != SMD_SS_CLOSED) {
600 ch->state = new_state;
601 ch->fDSR = 0;
602 ch->fCTS = 0;
603 ch->fCD = 0;
604 ch->fSTATE = 1;
605 }
606}
Bryan Huntsman3f2bc4d2011-08-16 17:27:22 -0700607
608static void smd_channel_reset_state(struct smd_alloc_elm *shared,
609 unsigned new_state, unsigned pid)
610{
611 unsigned n;
612 struct smd_shared_v2 *shared2;
613 uint32_t type;
614 struct smd_half_channel *local_ch;
615 struct smd_half_channel *remote_ch;
616
617 for (n = 0; n < SMD_CHANNELS; n++) {
618 if (!shared[n].ref_count)
619 continue;
620 if (!shared[n].name[0])
621 continue;
622
623 type = SMD_CHANNEL_TYPE(shared[n].type);
624 shared2 = smem_alloc(SMEM_SMD_BASE_ID + n, sizeof(*shared2));
625 if (!shared2)
626 continue;
627
Eric Holmberg2a563c32011-10-05 14:51:43 -0600628 if (pid_is_on_edge(shared2, type, pid, &local_ch, &remote_ch))
629 smd_reset_edge(local_ch, new_state);
Bryan Huntsman3f2bc4d2011-08-16 17:27:22 -0700630
Eric Holmberg2a563c32011-10-05 14:51:43 -0600631 /*
632 * ModemFW is in the same subsystem as ModemSW, but has
633 * separate SMD edges that need to be reset.
634 */
635 if (pid == SMSM_MODEM &&
636 pid_is_on_edge(shared2, type, SMD_MODEM_Q6_FW,
637 &local_ch, &remote_ch))
638 smd_reset_edge(local_ch, new_state);
Bryan Huntsman3f2bc4d2011-08-16 17:27:22 -0700639 }
640}
641
642
643void smd_channel_reset(uint32_t restart_pid)
644{
645 struct smd_alloc_elm *shared;
Bryan Huntsman3f2bc4d2011-08-16 17:27:22 -0700646 unsigned long flags;
647
648 SMD_DBG("%s: starting reset\n", __func__);
649 shared = smem_find(ID_CH_ALLOC_TBL, sizeof(*shared) * 64);
650 if (!shared) {
651 pr_err("%s: allocation table not initialized\n", __func__);
652 return;
653 }
654
Eric Holmbergf6d7d1a2011-09-23 18:31:04 -0600655 /* release any held spinlocks */
656 remote_spin_release(&remote_spinlock, restart_pid);
657 remote_spin_release_all(restart_pid);
Bryan Huntsman3f2bc4d2011-08-16 17:27:22 -0700658
659 /* reset SMSM entry */
660 if (smsm_info.state) {
661 writel_relaxed(0, SMSM_STATE_ADDR(restart_pid));
662
663 /* clear apps SMSM to restart SMSM init handshake */
664 if (restart_pid == SMSM_MODEM)
665 writel_relaxed(0, SMSM_STATE_ADDR(SMSM_APPS));
666
667 /* notify SMSM processors */
668 smsm_irq_handler(0, 0);
669 MSM_TRIG_A2M_SMSM_INT;
670 MSM_TRIG_A2Q6_SMSM_INT;
Jeff Hugo6a8057c2011-08-16 13:47:12 -0600671 MSM_TRIG_A2DSPS_SMSM_INT;
Bryan Huntsman3f2bc4d2011-08-16 17:27:22 -0700672 }
673
674 /* change all remote states to CLOSING */
675 mutex_lock(&smd_probe_lock);
676 spin_lock_irqsave(&smd_lock, flags);
677 smd_channel_reset_state(shared, SMD_SS_CLOSING, restart_pid);
678 spin_unlock_irqrestore(&smd_lock, flags);
679 mutex_unlock(&smd_probe_lock);
680
681 /* notify SMD processors */
682 mb();
683 smd_fake_irq_handler(0);
684 notify_modem_smd();
685 notify_dsp_smd();
686 notify_dsps_smd();
687 notify_wcnss_smd();
688
689 /* change all remote states to CLOSED */
690 mutex_lock(&smd_probe_lock);
691 spin_lock_irqsave(&smd_lock, flags);
692 smd_channel_reset_state(shared, SMD_SS_CLOSED, restart_pid);
693 spin_unlock_irqrestore(&smd_lock, flags);
694 mutex_unlock(&smd_probe_lock);
695
696 /* notify SMD processors */
697 mb();
698 smd_fake_irq_handler(0);
699 notify_modem_smd();
700 notify_dsp_smd();
701 notify_dsps_smd();
702 notify_wcnss_smd();
703
704 SMD_DBG("%s: finished reset\n", __func__);
705}
706
Brian Swetland2eb44eb2008-09-29 16:00:48 -0700707/* how many bytes are available for reading */
708static int smd_stream_read_avail(struct smd_channel *ch)
709{
Brian Swetland5b0f5a32009-04-26 18:38:49 -0700710 return (ch->recv->head - ch->recv->tail) & ch->fifo_mask;
Brian Swetland2eb44eb2008-09-29 16:00:48 -0700711}
712
713/* how many bytes we are free to write */
714static int smd_stream_write_avail(struct smd_channel *ch)
715{
Brian Swetland5b0f5a32009-04-26 18:38:49 -0700716 return ch->fifo_mask -
717 ((ch->send->head - ch->send->tail) & ch->fifo_mask);
Brian Swetland2eb44eb2008-09-29 16:00:48 -0700718}
719
720static int smd_packet_read_avail(struct smd_channel *ch)
721{
722 if (ch->current_packet) {
723 int n = smd_stream_read_avail(ch);
724 if (n > ch->current_packet)
725 n = ch->current_packet;
726 return n;
727 } else {
728 return 0;
729 }
730}
731
732static int smd_packet_write_avail(struct smd_channel *ch)
733{
734 int n = smd_stream_write_avail(ch);
735 return n > SMD_HEADER_SIZE ? n - SMD_HEADER_SIZE : 0;
736}
737
738static int ch_is_open(struct smd_channel *ch)
739{
Bryan Huntsman3f2bc4d2011-08-16 17:27:22 -0700740 return (ch->recv->state == SMD_SS_OPENED ||
741 ch->recv->state == SMD_SS_FLUSHING)
742 && (ch->send->state == SMD_SS_OPENED);
Brian Swetland2eb44eb2008-09-29 16:00:48 -0700743}
744
745/* provide a pointer and length to readable data in the fifo */
746static unsigned ch_read_buffer(struct smd_channel *ch, void **ptr)
747{
748 unsigned head = ch->recv->head;
749 unsigned tail = ch->recv->tail;
Brian Swetland5b0f5a32009-04-26 18:38:49 -0700750 *ptr = (void *) (ch->recv_data + tail);
Brian Swetland2eb44eb2008-09-29 16:00:48 -0700751
752 if (tail <= head)
753 return head - tail;
754 else
Brian Swetland5b0f5a32009-04-26 18:38:49 -0700755 return ch->fifo_size - tail;
Brian Swetland2eb44eb2008-09-29 16:00:48 -0700756}
757
Bryan Huntsman3f2bc4d2011-08-16 17:27:22 -0700758static int read_intr_blocked(struct smd_channel *ch)
759{
760 return ch->recv->fBLOCKREADINTR;
761}
762
Brian Swetland2eb44eb2008-09-29 16:00:48 -0700763/* advance the fifo read pointer after data from ch_read_buffer is consumed */
764static void ch_read_done(struct smd_channel *ch, unsigned count)
765{
766 BUG_ON(count > smd_stream_read_avail(ch));
Brian Swetland5b0f5a32009-04-26 18:38:49 -0700767 ch->recv->tail = (ch->recv->tail + count) & ch->fifo_mask;
Bryan Huntsman3f2bc4d2011-08-16 17:27:22 -0700768 wmb();
Haley Teng7632fba2009-10-12 10:38:10 -0700769 ch->send->fTAIL = 1;
Brian Swetland2eb44eb2008-09-29 16:00:48 -0700770}
771
772/* basic read interface to ch_read_{buffer,done} used
Brian Swetland03e00cd2009-07-01 17:58:37 -0700773 * by smd_*_read() and update_packet_state()
774 * will read-and-discard if the _data pointer is null
775 */
Bryan Huntsman3f2bc4d2011-08-16 17:27:22 -0700776static int ch_read(struct smd_channel *ch, void *_data, int len, int user_buf)
Brian Swetland2eb44eb2008-09-29 16:00:48 -0700777{
778 void *ptr;
779 unsigned n;
780 unsigned char *data = _data;
781 int orig_len = len;
Bryan Huntsman3f2bc4d2011-08-16 17:27:22 -0700782 int r = 0;
Brian Swetland2eb44eb2008-09-29 16:00:48 -0700783
784 while (len > 0) {
785 n = ch_read_buffer(ch, &ptr);
786 if (n == 0)
787 break;
788
789 if (n > len)
790 n = len;
Bryan Huntsman3f2bc4d2011-08-16 17:27:22 -0700791 if (_data) {
792 if (user_buf) {
793 r = copy_to_user(data, ptr, n);
794 if (r > 0) {
795 pr_err("%s: "
796 "copy_to_user could not copy "
797 "%i bytes.\n",
798 __func__,
799 r);
800 }
801 } else
802 memcpy(data, ptr, n);
803 }
Brian Swetland2eb44eb2008-09-29 16:00:48 -0700804
805 data += n;
806 len -= n;
807 ch_read_done(ch, n);
808 }
809
810 return orig_len - len;
811}
812
813static void update_stream_state(struct smd_channel *ch)
814{
815 /* streams have no special state requiring updating */
816}
817
818static void update_packet_state(struct smd_channel *ch)
819{
820 unsigned hdr[5];
821 int r;
822
823 /* can't do anything if we're in the middle of a packet */
Bryan Huntsman3f2bc4d2011-08-16 17:27:22 -0700824 while (ch->current_packet == 0) {
825 /* discard 0 length packets if any */
Brian Swetland2eb44eb2008-09-29 16:00:48 -0700826
Bryan Huntsman3f2bc4d2011-08-16 17:27:22 -0700827 /* don't bother unless we can get the full header */
828 if (smd_stream_read_avail(ch) < SMD_HEADER_SIZE)
829 return;
Brian Swetland2eb44eb2008-09-29 16:00:48 -0700830
Bryan Huntsman3f2bc4d2011-08-16 17:27:22 -0700831 r = ch_read(ch, hdr, SMD_HEADER_SIZE, 0);
832 BUG_ON(r != SMD_HEADER_SIZE);
Brian Swetland2eb44eb2008-09-29 16:00:48 -0700833
Bryan Huntsman3f2bc4d2011-08-16 17:27:22 -0700834 ch->current_packet = hdr[0];
835 }
Brian Swetland2eb44eb2008-09-29 16:00:48 -0700836}
837
838/* provide a pointer and length to next free space in the fifo */
839static unsigned ch_write_buffer(struct smd_channel *ch, void **ptr)
840{
841 unsigned head = ch->send->head;
842 unsigned tail = ch->send->tail;
Brian Swetland5b0f5a32009-04-26 18:38:49 -0700843 *ptr = (void *) (ch->send_data + head);
Brian Swetland2eb44eb2008-09-29 16:00:48 -0700844
845 if (head < tail) {
846 return tail - head - 1;
847 } else {
848 if (tail == 0)
Brian Swetland5b0f5a32009-04-26 18:38:49 -0700849 return ch->fifo_size - head - 1;
Brian Swetland2eb44eb2008-09-29 16:00:48 -0700850 else
Brian Swetland5b0f5a32009-04-26 18:38:49 -0700851 return ch->fifo_size - head;
Brian Swetland2eb44eb2008-09-29 16:00:48 -0700852 }
853}
854
855/* advace the fifo write pointer after freespace
856 * from ch_write_buffer is filled
857 */
858static void ch_write_done(struct smd_channel *ch, unsigned count)
859{
860 BUG_ON(count > smd_stream_write_avail(ch));
Brian Swetland5b0f5a32009-04-26 18:38:49 -0700861 ch->send->head = (ch->send->head + count) & ch->fifo_mask;
Bryan Huntsman3f2bc4d2011-08-16 17:27:22 -0700862 wmb();
Brian Swetland2eb44eb2008-09-29 16:00:48 -0700863 ch->send->fHEAD = 1;
864}
865
Brian Swetland5b0f5a32009-04-26 18:38:49 -0700866static void ch_set_state(struct smd_channel *ch, unsigned n)
Brian Swetland2eb44eb2008-09-29 16:00:48 -0700867{
868 if (n == SMD_SS_OPENED) {
Brian Swetland5b0f5a32009-04-26 18:38:49 -0700869 ch->send->fDSR = 1;
870 ch->send->fCTS = 1;
871 ch->send->fCD = 1;
Brian Swetland2eb44eb2008-09-29 16:00:48 -0700872 } else {
Brian Swetland5b0f5a32009-04-26 18:38:49 -0700873 ch->send->fDSR = 0;
874 ch->send->fCTS = 0;
875 ch->send->fCD = 0;
Brian Swetland2eb44eb2008-09-29 16:00:48 -0700876 }
Brian Swetland5b0f5a32009-04-26 18:38:49 -0700877 ch->send->state = n;
878 ch->send->fSTATE = 1;
879 ch->notify_other_cpu();
Brian Swetland2eb44eb2008-09-29 16:00:48 -0700880}
881
882static void do_smd_probe(void)
883{
884 struct smem_shared *shared = (void *) MSM_SHARED_RAM_BASE;
885 if (shared->heap_info.free_offset != last_heap_free) {
886 last_heap_free = shared->heap_info.free_offset;
887 schedule_work(&probe_work);
888 }
889}
890
891static void smd_state_change(struct smd_channel *ch,
892 unsigned last, unsigned next)
893{
894 ch->last_state = next;
895
Bryan Huntsman3f2bc4d2011-08-16 17:27:22 -0700896 SMD_INFO("SMD: ch %d %d -> %d\n", ch->n, last, next);
Brian Swetland2eb44eb2008-09-29 16:00:48 -0700897
898 switch (next) {
899 case SMD_SS_OPENING:
Bryan Huntsman3f2bc4d2011-08-16 17:27:22 -0700900 if (ch->send->state == SMD_SS_CLOSING ||
901 ch->send->state == SMD_SS_CLOSED) {
902 ch->recv->tail = 0;
903 ch->send->head = 0;
904 ch->send->fBLOCKREADINTR = 0;
905 ch_set_state(ch, SMD_SS_OPENING);
906 }
907 break;
Brian Swetland2eb44eb2008-09-29 16:00:48 -0700908 case SMD_SS_OPENED:
Bryan Huntsman3f2bc4d2011-08-16 17:27:22 -0700909 if (ch->send->state == SMD_SS_OPENING) {
Brian Swetland5b0f5a32009-04-26 18:38:49 -0700910 ch_set_state(ch, SMD_SS_OPENED);
Bryan Huntsman3f2bc4d2011-08-16 17:27:22 -0700911 ch->notify(ch->priv, SMD_EVENT_OPEN);
912 }
Brian Swetland2eb44eb2008-09-29 16:00:48 -0700913 break;
914 case SMD_SS_FLUSHING:
915 case SMD_SS_RESET:
916 /* we should force them to close? */
Bryan Huntsman3f2bc4d2011-08-16 17:27:22 -0700917 break;
918 case SMD_SS_CLOSED:
919 if (ch->send->state == SMD_SS_OPENED) {
920 ch_set_state(ch, SMD_SS_CLOSING);
921 ch->current_packet = 0;
Eric Holmbergad4fa8d2011-11-11 16:55:13 -0700922 ch->pending_pkt_sz = 0;
Bryan Huntsman3f2bc4d2011-08-16 17:27:22 -0700923 ch->notify(ch->priv, SMD_EVENT_CLOSE);
924 }
925 break;
926 case SMD_SS_CLOSING:
927 if (ch->send->state == SMD_SS_CLOSED) {
928 list_move(&ch->ch_list,
929 &smd_ch_to_close_list);
930 queue_work(channel_close_wq,
931 &finalize_channel_close_work);
932 }
933 break;
Brian Swetland2eb44eb2008-09-29 16:00:48 -0700934 }
935}
936
Bryan Huntsman3f2bc4d2011-08-16 17:27:22 -0700937static void handle_smd_irq_closing_list(void)
938{
939 unsigned long flags;
940 struct smd_channel *ch;
941 struct smd_channel *index;
942 unsigned tmp;
943
944 spin_lock_irqsave(&smd_lock, flags);
945 list_for_each_entry_safe(ch, index, &smd_ch_closing_list, ch_list) {
946 if (ch->recv->fSTATE)
947 ch->recv->fSTATE = 0;
948 tmp = ch->recv->state;
949 if (tmp != ch->last_state)
950 smd_state_change(ch, ch->last_state, tmp);
951 }
952 spin_unlock_irqrestore(&smd_lock, flags);
953}
954
Brian Swetland5b0f5a32009-04-26 18:38:49 -0700955static void handle_smd_irq(struct list_head *list, void (*notify)(void))
Brian Swetland2eb44eb2008-09-29 16:00:48 -0700956{
957 unsigned long flags;
958 struct smd_channel *ch;
Brian Swetland2eb44eb2008-09-29 16:00:48 -0700959 unsigned ch_flags;
960 unsigned tmp;
Bryan Huntsman3f2bc4d2011-08-16 17:27:22 -0700961 unsigned char state_change;
Brian Swetland2eb44eb2008-09-29 16:00:48 -0700962
963 spin_lock_irqsave(&smd_lock, flags);
Brian Swetland5b0f5a32009-04-26 18:38:49 -0700964 list_for_each_entry(ch, list, ch_list) {
Bryan Huntsman3f2bc4d2011-08-16 17:27:22 -0700965 state_change = 0;
Brian Swetland2eb44eb2008-09-29 16:00:48 -0700966 ch_flags = 0;
967 if (ch_is_open(ch)) {
968 if (ch->recv->fHEAD) {
969 ch->recv->fHEAD = 0;
970 ch_flags |= 1;
Brian Swetland2eb44eb2008-09-29 16:00:48 -0700971 }
972 if (ch->recv->fTAIL) {
973 ch->recv->fTAIL = 0;
974 ch_flags |= 2;
Brian Swetland2eb44eb2008-09-29 16:00:48 -0700975 }
976 if (ch->recv->fSTATE) {
977 ch->recv->fSTATE = 0;
978 ch_flags |= 4;
Brian Swetland2eb44eb2008-09-29 16:00:48 -0700979 }
980 }
981 tmp = ch->recv->state;
Bryan Huntsman3f2bc4d2011-08-16 17:27:22 -0700982 if (tmp != ch->last_state) {
Brian Swetland2eb44eb2008-09-29 16:00:48 -0700983 smd_state_change(ch, ch->last_state, tmp);
Bryan Huntsman3f2bc4d2011-08-16 17:27:22 -0700984 state_change = 1;
985 }
Brian Swetland2eb44eb2008-09-29 16:00:48 -0700986 if (ch_flags) {
987 ch->update_state(ch);
988 ch->notify(ch->priv, SMD_EVENT_DATA);
989 }
Bryan Huntsman3f2bc4d2011-08-16 17:27:22 -0700990 if (ch_flags & 0x4 && !state_change)
991 ch->notify(ch->priv, SMD_EVENT_STATUS);
Brian Swetland2eb44eb2008-09-29 16:00:48 -0700992 }
Brian Swetland2eb44eb2008-09-29 16:00:48 -0700993 spin_unlock_irqrestore(&smd_lock, flags);
994 do_smd_probe();
Brian Swetland5b0f5a32009-04-26 18:38:49 -0700995}
996
Brian Swetland37521a32009-07-01 18:30:47 -0700997static irqreturn_t smd_modem_irq_handler(int irq, void *data)
Brian Swetland5b0f5a32009-04-26 18:38:49 -0700998{
Brian Swetland37521a32009-07-01 18:30:47 -0700999 handle_smd_irq(&smd_ch_list_modem, notify_modem_smd);
Bryan Huntsman3f2bc4d2011-08-16 17:27:22 -07001000 handle_smd_irq_closing_list();
Brian Swetland37521a32009-07-01 18:30:47 -07001001 return IRQ_HANDLED;
1002}
1003
Daniel Walkerb13525c2010-03-18 10:10:30 -07001004#if defined(CONFIG_QDSP6)
Brian Swetland37521a32009-07-01 18:30:47 -07001005static irqreturn_t smd_dsp_irq_handler(int irq, void *data)
1006{
1007 handle_smd_irq(&smd_ch_list_dsp, notify_dsp_smd);
Bryan Huntsman3f2bc4d2011-08-16 17:27:22 -07001008 handle_smd_irq_closing_list();
1009 return IRQ_HANDLED;
1010}
1011#endif
1012
1013#if defined(CONFIG_DSPS)
1014static irqreturn_t smd_dsps_irq_handler(int irq, void *data)
1015{
1016 handle_smd_irq(&smd_ch_list_dsps, notify_dsps_smd);
1017 handle_smd_irq_closing_list();
1018 return IRQ_HANDLED;
1019}
1020#endif
1021
1022#if defined(CONFIG_WCNSS)
1023static irqreturn_t smd_wcnss_irq_handler(int irq, void *data)
1024{
1025 handle_smd_irq(&smd_ch_list_wcnss, notify_wcnss_smd);
1026 handle_smd_irq_closing_list();
Brian Swetland2eb44eb2008-09-29 16:00:48 -07001027 return IRQ_HANDLED;
1028}
Daniel Walkerb13525c2010-03-18 10:10:30 -07001029#endif
Brian Swetland2eb44eb2008-09-29 16:00:48 -07001030
1031static void smd_fake_irq_handler(unsigned long arg)
1032{
Brian Swetland37521a32009-07-01 18:30:47 -07001033 handle_smd_irq(&smd_ch_list_modem, notify_modem_smd);
1034 handle_smd_irq(&smd_ch_list_dsp, notify_dsp_smd);
Bryan Huntsman3f2bc4d2011-08-16 17:27:22 -07001035 handle_smd_irq(&smd_ch_list_dsps, notify_dsps_smd);
1036 handle_smd_irq(&smd_ch_list_wcnss, notify_wcnss_smd);
1037 handle_smd_irq_closing_list();
Brian Swetland2eb44eb2008-09-29 16:00:48 -07001038}
1039
1040static DECLARE_TASKLET(smd_fake_irq_tasklet, smd_fake_irq_handler, 0);
1041
Brian Swetland37521a32009-07-01 18:30:47 -07001042static inline int smd_need_int(struct smd_channel *ch)
1043{
1044 if (ch_is_open(ch)) {
1045 if (ch->recv->fHEAD || ch->recv->fTAIL || ch->recv->fSTATE)
1046 return 1;
1047 if (ch->recv->state != ch->last_state)
1048 return 1;
1049 }
1050 return 0;
1051}
1052
Brian Swetland2eb44eb2008-09-29 16:00:48 -07001053void smd_sleep_exit(void)
1054{
1055 unsigned long flags;
1056 struct smd_channel *ch;
Brian Swetland2eb44eb2008-09-29 16:00:48 -07001057 int need_int = 0;
1058
1059 spin_lock_irqsave(&smd_lock, flags);
Brian Swetland37521a32009-07-01 18:30:47 -07001060 list_for_each_entry(ch, &smd_ch_list_modem, ch_list) {
1061 if (smd_need_int(ch)) {
1062 need_int = 1;
1063 break;
1064 }
1065 }
1066 list_for_each_entry(ch, &smd_ch_list_dsp, ch_list) {
1067 if (smd_need_int(ch)) {
1068 need_int = 1;
1069 break;
Brian Swetland2eb44eb2008-09-29 16:00:48 -07001070 }
1071 }
Bryan Huntsman3f2bc4d2011-08-16 17:27:22 -07001072 list_for_each_entry(ch, &smd_ch_list_dsps, ch_list) {
1073 if (smd_need_int(ch)) {
1074 need_int = 1;
1075 break;
1076 }
1077 }
1078 list_for_each_entry(ch, &smd_ch_list_wcnss, ch_list) {
1079 if (smd_need_int(ch)) {
1080 need_int = 1;
1081 break;
1082 }
1083 }
Brian Swetland2eb44eb2008-09-29 16:00:48 -07001084 spin_unlock_irqrestore(&smd_lock, flags);
1085 do_smd_probe();
Brian Swetland37521a32009-07-01 18:30:47 -07001086
Brian Swetland2eb44eb2008-09-29 16:00:48 -07001087 if (need_int) {
Bryan Huntsman3f2bc4d2011-08-16 17:27:22 -07001088 SMD_DBG("smd_sleep_exit need interrupt\n");
Brian Swetland2eb44eb2008-09-29 16:00:48 -07001089 tasklet_schedule(&smd_fake_irq_tasklet);
1090 }
1091}
Bryan Huntsman3f2bc4d2011-08-16 17:27:22 -07001092EXPORT_SYMBOL(smd_sleep_exit);
Brian Swetland2eb44eb2008-09-29 16:00:48 -07001093
Bryan Huntsman3f2bc4d2011-08-16 17:27:22 -07001094static int smd_is_packet(struct smd_alloc_elm *alloc_elm)
Brian Swetland2eb44eb2008-09-29 16:00:48 -07001095{
Bryan Huntsman3f2bc4d2011-08-16 17:27:22 -07001096 if (SMD_XFER_TYPE(alloc_elm->type) == 1)
1097 return 0;
1098 else if (SMD_XFER_TYPE(alloc_elm->type) == 2)
Brian Swetland5b0f5a32009-04-26 18:38:49 -07001099 return 1;
Bryan Huntsman3f2bc4d2011-08-16 17:27:22 -07001100
1101 /* for cases where xfer type is 0 */
1102 if (!strncmp(alloc_elm->name, "DAL", 3))
Brian Swetland5b0f5a32009-04-26 18:38:49 -07001103 return 0;
1104
Bryan Huntsman3f2bc4d2011-08-16 17:27:22 -07001105 /* for cases where xfer type is 0 */
1106 if (!strncmp(alloc_elm->name, "RPCCALL_QDSP", 12))
1107 return 0;
1108
1109 if (alloc_elm->cid > 4 || alloc_elm->cid == 1)
Brian Swetland2eb44eb2008-09-29 16:00:48 -07001110 return 1;
1111 else
1112 return 0;
1113}
1114
Bryan Huntsman3f2bc4d2011-08-16 17:27:22 -07001115static int smd_stream_write(smd_channel_t *ch, const void *_data, int len,
1116 int user_buf)
Brian Swetland2eb44eb2008-09-29 16:00:48 -07001117{
1118 void *ptr;
1119 const unsigned char *buf = _data;
1120 unsigned xfer;
1121 int orig_len = len;
Bryan Huntsman3f2bc4d2011-08-16 17:27:22 -07001122 int r = 0;
Brian Swetland2eb44eb2008-09-29 16:00:48 -07001123
Bryan Huntsman3f2bc4d2011-08-16 17:27:22 -07001124 SMD_DBG("smd_stream_write() %d -> ch%d\n", len, ch->n);
Brian Swetland2eb44eb2008-09-29 16:00:48 -07001125 if (len < 0)
1126 return -EINVAL;
Bryan Huntsman3f2bc4d2011-08-16 17:27:22 -07001127 else if (len == 0)
1128 return 0;
Brian Swetland2eb44eb2008-09-29 16:00:48 -07001129
1130 while ((xfer = ch_write_buffer(ch, &ptr)) != 0) {
1131 if (!ch_is_open(ch))
1132 break;
1133 if (xfer > len)
1134 xfer = len;
Bryan Huntsman3f2bc4d2011-08-16 17:27:22 -07001135 if (user_buf) {
1136 r = copy_from_user(ptr, buf, xfer);
1137 if (r > 0) {
1138 pr_err("%s: "
1139 "copy_from_user could not copy %i "
1140 "bytes.\n",
1141 __func__,
1142 r);
1143 }
1144 } else
1145 memcpy(ptr, buf, xfer);
Brian Swetland2eb44eb2008-09-29 16:00:48 -07001146 ch_write_done(ch, xfer);
1147 len -= xfer;
1148 buf += xfer;
1149 if (len == 0)
1150 break;
1151 }
1152
Bryan Huntsman3f2bc4d2011-08-16 17:27:22 -07001153 if (orig_len - len)
1154 ch->notify_other_cpu();
Brian Swetland2eb44eb2008-09-29 16:00:48 -07001155
1156 return orig_len - len;
1157}
1158
Bryan Huntsman3f2bc4d2011-08-16 17:27:22 -07001159static int smd_packet_write(smd_channel_t *ch, const void *_data, int len,
1160 int user_buf)
Brian Swetland2eb44eb2008-09-29 16:00:48 -07001161{
Bryan Huntsman3f2bc4d2011-08-16 17:27:22 -07001162 int ret;
Brian Swetland2eb44eb2008-09-29 16:00:48 -07001163 unsigned hdr[5];
1164
Bryan Huntsman3f2bc4d2011-08-16 17:27:22 -07001165 SMD_DBG("smd_packet_write() %d -> ch%d\n", len, ch->n);
Brian Swetland2eb44eb2008-09-29 16:00:48 -07001166 if (len < 0)
1167 return -EINVAL;
Bryan Huntsman3f2bc4d2011-08-16 17:27:22 -07001168 else if (len == 0)
1169 return 0;
Brian Swetland2eb44eb2008-09-29 16:00:48 -07001170
1171 if (smd_stream_write_avail(ch) < (len + SMD_HEADER_SIZE))
1172 return -ENOMEM;
1173
1174 hdr[0] = len;
1175 hdr[1] = hdr[2] = hdr[3] = hdr[4] = 0;
1176
Bryan Huntsman3f2bc4d2011-08-16 17:27:22 -07001177
1178 ret = smd_stream_write(ch, hdr, sizeof(hdr), 0);
1179 if (ret < 0 || ret != sizeof(hdr)) {
1180 SMD_DBG("%s failed to write pkt header: "
1181 "%d returned\n", __func__, ret);
1182 return -1;
1183 }
1184
1185
1186 ret = smd_stream_write(ch, _data, len, user_buf);
1187 if (ret < 0 || ret != len) {
1188 SMD_DBG("%s failed to write pkt data: "
1189 "%d returned\n", __func__, ret);
1190 return ret;
1191 }
Brian Swetland2eb44eb2008-09-29 16:00:48 -07001192
1193 return len;
1194}
1195
Bryan Huntsman3f2bc4d2011-08-16 17:27:22 -07001196static int smd_stream_read(smd_channel_t *ch, void *data, int len, int user_buf)
Brian Swetland2eb44eb2008-09-29 16:00:48 -07001197{
1198 int r;
1199
1200 if (len < 0)
1201 return -EINVAL;
1202
Bryan Huntsman3f2bc4d2011-08-16 17:27:22 -07001203 r = ch_read(ch, data, len, user_buf);
Brian Swetland2eb44eb2008-09-29 16:00:48 -07001204 if (r > 0)
Bryan Huntsman3f2bc4d2011-08-16 17:27:22 -07001205 if (!read_intr_blocked(ch))
1206 ch->notify_other_cpu();
Brian Swetland2eb44eb2008-09-29 16:00:48 -07001207
1208 return r;
1209}
1210
Bryan Huntsman3f2bc4d2011-08-16 17:27:22 -07001211static int smd_packet_read(smd_channel_t *ch, void *data, int len, int user_buf)
Brian Swetland2eb44eb2008-09-29 16:00:48 -07001212{
1213 unsigned long flags;
1214 int r;
1215
1216 if (len < 0)
1217 return -EINVAL;
1218
1219 if (len > ch->current_packet)
1220 len = ch->current_packet;
1221
Bryan Huntsman3f2bc4d2011-08-16 17:27:22 -07001222 r = ch_read(ch, data, len, user_buf);
Brian Swetland2eb44eb2008-09-29 16:00:48 -07001223 if (r > 0)
Bryan Huntsman3f2bc4d2011-08-16 17:27:22 -07001224 if (!read_intr_blocked(ch))
1225 ch->notify_other_cpu();
Brian Swetland2eb44eb2008-09-29 16:00:48 -07001226
1227 spin_lock_irqsave(&smd_lock, flags);
1228 ch->current_packet -= r;
1229 update_packet_state(ch);
1230 spin_unlock_irqrestore(&smd_lock, flags);
1231
1232 return r;
1233}
1234
Bryan Huntsman3f2bc4d2011-08-16 17:27:22 -07001235static int smd_packet_read_from_cb(smd_channel_t *ch, void *data, int len,
1236 int user_buf)
1237{
1238 int r;
1239
1240 if (len < 0)
1241 return -EINVAL;
1242
1243 if (len > ch->current_packet)
1244 len = ch->current_packet;
1245
1246 r = ch_read(ch, data, len, user_buf);
1247 if (r > 0)
1248 if (!read_intr_blocked(ch))
1249 ch->notify_other_cpu();
1250
1251 ch->current_packet -= r;
1252 update_packet_state(ch);
1253
1254 return r;
1255}
1256
Angshuman Sarkarac7d6252011-09-30 18:20:59 +05301257#if (defined(CONFIG_MSM_SMD_PKG4) || defined(CONFIG_MSM_SMD_PKG3))
Bryan Huntsman3f2bc4d2011-08-16 17:27:22 -07001258static int smd_alloc_v2(struct smd_channel *ch)
1259{
1260 struct smd_shared_v2 *shared2;
1261 void *buffer;
1262 unsigned buffer_sz;
1263
1264 shared2 = smem_alloc(SMEM_SMD_BASE_ID + ch->n, sizeof(*shared2));
1265 if (!shared2) {
1266 SMD_INFO("smem_alloc failed ch=%d\n", ch->n);
Angshuman Sarkarac7d6252011-09-30 18:20:59 +05301267 return -EINVAL;
Bryan Huntsman3f2bc4d2011-08-16 17:27:22 -07001268 }
1269 buffer = smem_get_entry(SMEM_SMD_FIFO_BASE_ID + ch->n, &buffer_sz);
1270 if (!buffer) {
Angshuman Sarkarac7d6252011-09-30 18:20:59 +05301271 SMD_INFO("smem_get_entry failed\n");
1272 return -EINVAL;
Bryan Huntsman3f2bc4d2011-08-16 17:27:22 -07001273 }
1274
1275 /* buffer must be a power-of-two size */
Angshuman Sarkarac7d6252011-09-30 18:20:59 +05301276 if (buffer_sz & (buffer_sz - 1)) {
1277 SMD_INFO("Buffer size: %u not power of two\n", buffer_sz);
1278 return -EINVAL;
1279 }
Bryan Huntsman3f2bc4d2011-08-16 17:27:22 -07001280 buffer_sz /= 2;
1281 ch->send = &shared2->ch0;
1282 ch->recv = &shared2->ch1;
1283 ch->send_data = buffer;
1284 ch->recv_data = buffer + buffer_sz;
1285 ch->fifo_size = buffer_sz;
1286 return 0;
1287}
1288
1289static int smd_alloc_v1(struct smd_channel *ch)
1290{
Angshuman Sarkarac7d6252011-09-30 18:20:59 +05301291 return -EINVAL;
1292}
1293
1294#else /* define v1 for older targets */
1295static int smd_alloc_v2(struct smd_channel *ch)
1296{
1297 return -EINVAL;
1298}
1299
1300static int smd_alloc_v1(struct smd_channel *ch)
1301{
Bryan Huntsman3f2bc4d2011-08-16 17:27:22 -07001302 struct smd_shared_v1 *shared1;
1303 shared1 = smem_alloc(ID_SMD_CHANNELS + ch->n, sizeof(*shared1));
1304 if (!shared1) {
1305 pr_err("smd_alloc_channel() cid %d does not exist\n", ch->n);
Angshuman Sarkarac7d6252011-09-30 18:20:59 +05301306 return -EINVAL;
Bryan Huntsman3f2bc4d2011-08-16 17:27:22 -07001307 }
1308 ch->send = &shared1->ch0;
1309 ch->recv = &shared1->ch1;
1310 ch->send_data = shared1->data0;
1311 ch->recv_data = shared1->data1;
1312 ch->fifo_size = SMD_BUF_SIZE;
1313 return 0;
1314}
1315
Angshuman Sarkarac7d6252011-09-30 18:20:59 +05301316#endif
1317
Bryan Huntsman3f2bc4d2011-08-16 17:27:22 -07001318static int smd_alloc_channel(struct smd_alloc_elm *alloc_elm)
Brian Swetland2eb44eb2008-09-29 16:00:48 -07001319{
1320 struct smd_channel *ch;
Brian Swetland2eb44eb2008-09-29 16:00:48 -07001321
1322 ch = kzalloc(sizeof(struct smd_channel), GFP_KERNEL);
1323 if (ch == 0) {
1324 pr_err("smd_alloc_channel() out of memory\n");
Brian Swetland34f719b2009-10-30 16:22:05 -07001325 return -1;
Brian Swetland2eb44eb2008-09-29 16:00:48 -07001326 }
Bryan Huntsman3f2bc4d2011-08-16 17:27:22 -07001327 ch->n = alloc_elm->cid;
Brian Swetland2eb44eb2008-09-29 16:00:48 -07001328
Bryan Huntsman3f2bc4d2011-08-16 17:27:22 -07001329 if (smd_alloc_v2(ch) && smd_alloc_v1(ch)) {
Brian Swetland5b0f5a32009-04-26 18:38:49 -07001330 kfree(ch);
Brian Swetland34f719b2009-10-30 16:22:05 -07001331 return -1;
Brian Swetland5b0f5a32009-04-26 18:38:49 -07001332 }
1333
1334 ch->fifo_mask = ch->fifo_size - 1;
Bryan Huntsman3f2bc4d2011-08-16 17:27:22 -07001335 ch->type = SMD_CHANNEL_TYPE(alloc_elm->type);
Brian Swetland5b0f5a32009-04-26 18:38:49 -07001336
Bryan Huntsman3f2bc4d2011-08-16 17:27:22 -07001337 if (ch->type == SMD_APPS_MODEM)
Brian Swetland5b0f5a32009-04-26 18:38:49 -07001338 ch->notify_other_cpu = notify_modem_smd;
Bryan Huntsman3f2bc4d2011-08-16 17:27:22 -07001339 else if (ch->type == SMD_APPS_QDSP)
Brian Swetland5b0f5a32009-04-26 18:38:49 -07001340 ch->notify_other_cpu = notify_dsp_smd;
Bryan Huntsman3f2bc4d2011-08-16 17:27:22 -07001341 else if (ch->type == SMD_APPS_DSPS)
1342 ch->notify_other_cpu = notify_dsps_smd;
1343 else
1344 ch->notify_other_cpu = notify_wcnss_smd;
Brian Swetland5b0f5a32009-04-26 18:38:49 -07001345
Bryan Huntsman3f2bc4d2011-08-16 17:27:22 -07001346 if (smd_is_packet(alloc_elm)) {
Brian Swetland2eb44eb2008-09-29 16:00:48 -07001347 ch->read = smd_packet_read;
1348 ch->write = smd_packet_write;
1349 ch->read_avail = smd_packet_read_avail;
1350 ch->write_avail = smd_packet_write_avail;
1351 ch->update_state = update_packet_state;
Bryan Huntsman3f2bc4d2011-08-16 17:27:22 -07001352 ch->read_from_cb = smd_packet_read_from_cb;
1353 ch->is_pkt_ch = 1;
Brian Swetland2eb44eb2008-09-29 16:00:48 -07001354 } else {
1355 ch->read = smd_stream_read;
1356 ch->write = smd_stream_write;
1357 ch->read_avail = smd_stream_read_avail;
1358 ch->write_avail = smd_stream_write_avail;
1359 ch->update_state = update_stream_state;
Bryan Huntsman3f2bc4d2011-08-16 17:27:22 -07001360 ch->read_from_cb = smd_stream_read;
Brian Swetland2eb44eb2008-09-29 16:00:48 -07001361 }
1362
Bryan Huntsman3f2bc4d2011-08-16 17:27:22 -07001363 memcpy(ch->name, alloc_elm->name, SMD_MAX_CH_NAME_LEN);
1364 ch->name[SMD_MAX_CH_NAME_LEN-1] = 0;
Brian Swetland2eb44eb2008-09-29 16:00:48 -07001365
Bryan Huntsman3f2bc4d2011-08-16 17:27:22 -07001366 ch->pdev.name = ch->name;
1367 ch->pdev.id = ch->type;
1368
1369 SMD_INFO("smd_alloc_channel() '%s' cid=%d\n",
1370 ch->name, ch->n);
1371
1372 mutex_lock(&smd_creation_mutex);
1373 list_add(&ch->ch_list, &smd_ch_closed_list);
1374 mutex_unlock(&smd_creation_mutex);
1375
1376 platform_device_register(&ch->pdev);
1377 if (!strncmp(ch->name, "LOOPBACK", 8) && ch->type == SMD_APPS_MODEM) {
1378 /* create a platform driver to be used by smd_tty driver
1379 * so that it can access the loopback port
1380 */
1381 loopback_tty_pdev.id = ch->type;
1382 platform_device_register(&loopback_tty_pdev);
1383 }
1384 return 0;
1385}
1386
1387static inline void notify_loopback_smd(void)
1388{
1389 unsigned long flags;
1390 struct smd_channel *ch;
1391
1392 spin_lock_irqsave(&smd_lock, flags);
1393 list_for_each_entry(ch, &smd_ch_list_loopback, ch_list) {
1394 ch->notify(ch->priv, SMD_EVENT_DATA);
1395 }
1396 spin_unlock_irqrestore(&smd_lock, flags);
1397}
1398
1399static int smd_alloc_loopback_channel(void)
1400{
1401 static struct smd_half_channel smd_loopback_ctl;
1402 static char smd_loopback_data[SMD_BUF_SIZE];
1403 struct smd_channel *ch;
1404
1405 ch = kzalloc(sizeof(struct smd_channel), GFP_KERNEL);
1406 if (ch == 0) {
1407 pr_err("%s: out of memory\n", __func__);
1408 return -1;
1409 }
1410 ch->n = SMD_LOOPBACK_CID;
1411
1412 ch->send = &smd_loopback_ctl;
1413 ch->recv = &smd_loopback_ctl;
1414 ch->send_data = smd_loopback_data;
1415 ch->recv_data = smd_loopback_data;
1416 ch->fifo_size = SMD_BUF_SIZE;
1417
1418 ch->fifo_mask = ch->fifo_size - 1;
1419 ch->type = SMD_LOOPBACK_TYPE;
1420 ch->notify_other_cpu = notify_loopback_smd;
1421
1422 ch->read = smd_stream_read;
1423 ch->write = smd_stream_write;
1424 ch->read_avail = smd_stream_read_avail;
1425 ch->write_avail = smd_stream_write_avail;
1426 ch->update_state = update_stream_state;
1427 ch->read_from_cb = smd_stream_read;
1428
1429 memset(ch->name, 0, 20);
1430 memcpy(ch->name, "local_loopback", 14);
1431
1432 ch->pdev.name = ch->name;
1433 ch->pdev.id = ch->type;
1434
1435 SMD_INFO("%s: '%s' cid=%d\n", __func__, ch->name, ch->n);
Brian Swetland2eb44eb2008-09-29 16:00:48 -07001436
1437 mutex_lock(&smd_creation_mutex);
1438 list_add(&ch->ch_list, &smd_ch_closed_list);
1439 mutex_unlock(&smd_creation_mutex);
1440
1441 platform_device_register(&ch->pdev);
Brian Swetland34f719b2009-10-30 16:22:05 -07001442 return 0;
Brian Swetland2eb44eb2008-09-29 16:00:48 -07001443}
1444
1445static void do_nothing_notify(void *priv, unsigned flags)
1446{
1447}
1448
Bryan Huntsman3f2bc4d2011-08-16 17:27:22 -07001449static void finalize_channel_close_fn(struct work_struct *work)
1450{
1451 unsigned long flags;
1452 struct smd_channel *ch;
1453 struct smd_channel *index;
1454
1455 spin_lock_irqsave(&smd_lock, flags);
1456 list_for_each_entry_safe(ch, index, &smd_ch_to_close_list, ch_list) {
1457 list_del(&ch->ch_list);
1458 spin_unlock_irqrestore(&smd_lock, flags);
1459 mutex_lock(&smd_creation_mutex);
1460 list_add(&ch->ch_list, &smd_ch_closed_list);
1461 mutex_unlock(&smd_creation_mutex);
1462 ch->notify(ch->priv, SMD_EVENT_REOPEN_READY);
1463 ch->notify = do_nothing_notify;
1464 spin_lock_irqsave(&smd_lock, flags);
1465 }
1466 spin_unlock_irqrestore(&smd_lock, flags);
1467}
1468
1469struct smd_channel *smd_get_channel(const char *name, uint32_t type)
Brian Swetland2eb44eb2008-09-29 16:00:48 -07001470{
1471 struct smd_channel *ch;
1472
1473 mutex_lock(&smd_creation_mutex);
1474 list_for_each_entry(ch, &smd_ch_closed_list, ch_list) {
Bryan Huntsman3f2bc4d2011-08-16 17:27:22 -07001475 if (!strcmp(name, ch->name) &&
1476 (type == ch->type)) {
Brian Swetland2eb44eb2008-09-29 16:00:48 -07001477 list_del(&ch->ch_list);
1478 mutex_unlock(&smd_creation_mutex);
1479 return ch;
1480 }
1481 }
1482 mutex_unlock(&smd_creation_mutex);
1483
1484 return NULL;
1485}
1486
Bryan Huntsman3f2bc4d2011-08-16 17:27:22 -07001487int smd_named_open_on_edge(const char *name, uint32_t edge,
1488 smd_channel_t **_ch,
1489 void *priv, void (*notify)(void *, unsigned))
Brian Swetland2eb44eb2008-09-29 16:00:48 -07001490{
1491 struct smd_channel *ch;
1492 unsigned long flags;
1493
1494 if (smd_initialized == 0) {
Bryan Huntsman3f2bc4d2011-08-16 17:27:22 -07001495 SMD_INFO("smd_open() before smd_init()\n");
Brian Swetland2eb44eb2008-09-29 16:00:48 -07001496 return -ENODEV;
1497 }
1498
Bryan Huntsman3f2bc4d2011-08-16 17:27:22 -07001499 SMD_DBG("smd_open('%s', %p, %p)\n", name, priv, notify);
1500
1501 ch = smd_get_channel(name, edge);
Brian Swetland2eb44eb2008-09-29 16:00:48 -07001502 if (!ch)
1503 return -ENODEV;
1504
1505 if (notify == 0)
1506 notify = do_nothing_notify;
1507
1508 ch->notify = notify;
1509 ch->current_packet = 0;
1510 ch->last_state = SMD_SS_CLOSED;
1511 ch->priv = priv;
1512
Bryan Huntsman3f2bc4d2011-08-16 17:27:22 -07001513 if (edge == SMD_LOOPBACK_TYPE) {
1514 ch->last_state = SMD_SS_OPENED;
1515 ch->send->state = SMD_SS_OPENED;
1516 ch->send->fDSR = 1;
1517 ch->send->fCTS = 1;
1518 ch->send->fCD = 1;
1519 }
1520
Brian Swetland2eb44eb2008-09-29 16:00:48 -07001521 *_ch = ch;
1522
Bryan Huntsman3f2bc4d2011-08-16 17:27:22 -07001523 SMD_DBG("smd_open: opening '%s'\n", ch->name);
1524
Brian Swetland2eb44eb2008-09-29 16:00:48 -07001525 spin_lock_irqsave(&smd_lock, flags);
Bryan Huntsman3f2bc4d2011-08-16 17:27:22 -07001526 if (SMD_CHANNEL_TYPE(ch->type) == SMD_APPS_MODEM)
Brian Swetland37521a32009-07-01 18:30:47 -07001527 list_add(&ch->ch_list, &smd_ch_list_modem);
Bryan Huntsman3f2bc4d2011-08-16 17:27:22 -07001528 else if (SMD_CHANNEL_TYPE(ch->type) == SMD_APPS_QDSP)
Brian Swetland37521a32009-07-01 18:30:47 -07001529 list_add(&ch->ch_list, &smd_ch_list_dsp);
Bryan Huntsman3f2bc4d2011-08-16 17:27:22 -07001530 else if (SMD_CHANNEL_TYPE(ch->type) == SMD_APPS_DSPS)
1531 list_add(&ch->ch_list, &smd_ch_list_dsps);
1532 else if (SMD_CHANNEL_TYPE(ch->type) == SMD_APPS_WCNSS)
1533 list_add(&ch->ch_list, &smd_ch_list_wcnss);
1534 else
1535 list_add(&ch->ch_list, &smd_ch_list_loopback);
Brian Swetland2eb44eb2008-09-29 16:00:48 -07001536
Bryan Huntsman3f2bc4d2011-08-16 17:27:22 -07001537 SMD_DBG("%s: opening ch %d\n", __func__, ch->n);
1538
1539 if (edge != SMD_LOOPBACK_TYPE)
1540 smd_state_change(ch, ch->last_state, SMD_SS_OPENING);
1541
Brian Swetland2eb44eb2008-09-29 16:00:48 -07001542 spin_unlock_irqrestore(&smd_lock, flags);
Brian Swetland2eb44eb2008-09-29 16:00:48 -07001543
1544 return 0;
1545}
Bryan Huntsman3f2bc4d2011-08-16 17:27:22 -07001546EXPORT_SYMBOL(smd_named_open_on_edge);
1547
1548
1549int smd_open(const char *name, smd_channel_t **_ch,
1550 void *priv, void (*notify)(void *, unsigned))
1551{
1552 return smd_named_open_on_edge(name, SMD_APPS_MODEM, _ch, priv,
1553 notify);
1554}
1555EXPORT_SYMBOL(smd_open);
Brian Swetland2eb44eb2008-09-29 16:00:48 -07001556
1557int smd_close(smd_channel_t *ch)
1558{
1559 unsigned long flags;
1560
Brian Swetland2eb44eb2008-09-29 16:00:48 -07001561 if (ch == 0)
1562 return -1;
1563
Bryan Huntsman3f2bc4d2011-08-16 17:27:22 -07001564 SMD_INFO("smd_close(%s)\n", ch->name);
Brian Swetland2eb44eb2008-09-29 16:00:48 -07001565
Bryan Huntsman3f2bc4d2011-08-16 17:27:22 -07001566 spin_lock_irqsave(&smd_lock, flags);
1567 list_del(&ch->ch_list);
1568 if (ch->n == SMD_LOOPBACK_CID) {
1569 ch->send->fDSR = 0;
1570 ch->send->fCTS = 0;
1571 ch->send->fCD = 0;
1572 ch->send->state = SMD_SS_CLOSED;
1573 } else
1574 ch_set_state(ch, SMD_SS_CLOSED);
1575
1576 if (ch->recv->state == SMD_SS_OPENED) {
1577 list_add(&ch->ch_list, &smd_ch_closing_list);
1578 spin_unlock_irqrestore(&smd_lock, flags);
1579 } else {
1580 spin_unlock_irqrestore(&smd_lock, flags);
1581 ch->notify = do_nothing_notify;
1582 mutex_lock(&smd_creation_mutex);
1583 list_add(&ch->ch_list, &smd_ch_closed_list);
1584 mutex_unlock(&smd_creation_mutex);
1585 }
Brian Swetland2eb44eb2008-09-29 16:00:48 -07001586
1587 return 0;
1588}
Bryan Huntsman3f2bc4d2011-08-16 17:27:22 -07001589EXPORT_SYMBOL(smd_close);
1590
1591int smd_write_start(smd_channel_t *ch, int len)
1592{
1593 int ret;
1594 unsigned hdr[5];
1595
1596 if (!ch) {
1597 pr_err("%s: Invalid channel specified\n", __func__);
1598 return -ENODEV;
1599 }
1600 if (!ch->is_pkt_ch) {
1601 pr_err("%s: non-packet channel specified\n", __func__);
1602 return -EACCES;
1603 }
1604 if (len < 1) {
1605 pr_err("%s: invalid length: %d\n", __func__, len);
1606 return -EINVAL;
1607 }
1608
1609 if (ch->pending_pkt_sz) {
1610 pr_err("%s: packet of size: %d in progress\n", __func__,
1611 ch->pending_pkt_sz);
1612 return -EBUSY;
1613 }
1614 ch->pending_pkt_sz = len;
1615
1616 if (smd_stream_write_avail(ch) < (SMD_HEADER_SIZE)) {
1617 ch->pending_pkt_sz = 0;
1618 SMD_DBG("%s: no space to write packet header\n", __func__);
1619 return -EAGAIN;
1620 }
1621
1622 hdr[0] = len;
1623 hdr[1] = hdr[2] = hdr[3] = hdr[4] = 0;
1624
1625
1626 ret = smd_stream_write(ch, hdr, sizeof(hdr), 0);
1627 if (ret < 0 || ret != sizeof(hdr)) {
1628 ch->pending_pkt_sz = 0;
1629 pr_err("%s: packet header failed to write\n", __func__);
1630 return -EPERM;
1631 }
1632 return 0;
1633}
1634EXPORT_SYMBOL(smd_write_start);
1635
1636int smd_write_segment(smd_channel_t *ch, void *data, int len, int user_buf)
1637{
1638 int bytes_written;
1639
1640 if (!ch) {
1641 pr_err("%s: Invalid channel specified\n", __func__);
1642 return -ENODEV;
1643 }
1644 if (len < 1) {
1645 pr_err("%s: invalid length: %d\n", __func__, len);
1646 return -EINVAL;
1647 }
1648
1649 if (!ch->pending_pkt_sz) {
1650 pr_err("%s: no transaction in progress\n", __func__);
1651 return -ENOEXEC;
1652 }
1653 if (ch->pending_pkt_sz - len < 0) {
1654 pr_err("%s: segment of size: %d will make packet go over "
1655 "length\n", __func__, len);
1656 return -EINVAL;
1657 }
1658
1659 bytes_written = smd_stream_write(ch, data, len, user_buf);
1660
1661 ch->pending_pkt_sz -= bytes_written;
1662
1663 return bytes_written;
1664}
1665EXPORT_SYMBOL(smd_write_segment);
1666
1667int smd_write_end(smd_channel_t *ch)
1668{
1669
1670 if (!ch) {
1671 pr_err("%s: Invalid channel specified\n", __func__);
1672 return -ENODEV;
1673 }
1674 if (ch->pending_pkt_sz) {
1675 pr_err("%s: current packet not completely written\n", __func__);
1676 return -E2BIG;
1677 }
1678
1679 return 0;
1680}
1681EXPORT_SYMBOL(smd_write_end);
Brian Swetland2eb44eb2008-09-29 16:00:48 -07001682
1683int smd_read(smd_channel_t *ch, void *data, int len)
1684{
Bryan Huntsman3f2bc4d2011-08-16 17:27:22 -07001685 return ch->read(ch, data, len, 0);
Brian Swetland2eb44eb2008-09-29 16:00:48 -07001686}
Bryan Huntsman3f2bc4d2011-08-16 17:27:22 -07001687EXPORT_SYMBOL(smd_read);
1688
1689int smd_read_user_buffer(smd_channel_t *ch, void *data, int len)
1690{
1691 return ch->read(ch, data, len, 1);
1692}
1693EXPORT_SYMBOL(smd_read_user_buffer);
1694
1695int smd_read_from_cb(smd_channel_t *ch, void *data, int len)
1696{
1697 return ch->read_from_cb(ch, data, len, 0);
1698}
1699EXPORT_SYMBOL(smd_read_from_cb);
Brian Swetland2eb44eb2008-09-29 16:00:48 -07001700
1701int smd_write(smd_channel_t *ch, const void *data, int len)
1702{
Bryan Huntsman3f2bc4d2011-08-16 17:27:22 -07001703 return ch->pending_pkt_sz ? -EBUSY : ch->write(ch, data, len, 0);
Brian Swetland2eb44eb2008-09-29 16:00:48 -07001704}
Bryan Huntsman3f2bc4d2011-08-16 17:27:22 -07001705EXPORT_SYMBOL(smd_write);
Brian Swetland2eb44eb2008-09-29 16:00:48 -07001706
Bryan Huntsman3f2bc4d2011-08-16 17:27:22 -07001707int smd_write_user_buffer(smd_channel_t *ch, const void *data, int len)
Brian Swetland636eb9c2009-12-07 15:28:08 -08001708{
Bryan Huntsman3f2bc4d2011-08-16 17:27:22 -07001709 return ch->pending_pkt_sz ? -EBUSY : ch->write(ch, data, len, 1);
Brian Swetland636eb9c2009-12-07 15:28:08 -08001710}
Bryan Huntsman3f2bc4d2011-08-16 17:27:22 -07001711EXPORT_SYMBOL(smd_write_user_buffer);
Brian Swetland636eb9c2009-12-07 15:28:08 -08001712
Brian Swetland2eb44eb2008-09-29 16:00:48 -07001713int smd_read_avail(smd_channel_t *ch)
1714{
1715 return ch->read_avail(ch);
1716}
Bryan Huntsman3f2bc4d2011-08-16 17:27:22 -07001717EXPORT_SYMBOL(smd_read_avail);
Brian Swetland2eb44eb2008-09-29 16:00:48 -07001718
1719int smd_write_avail(smd_channel_t *ch)
1720{
1721 return ch->write_avail(ch);
1722}
Bryan Huntsman3f2bc4d2011-08-16 17:27:22 -07001723EXPORT_SYMBOL(smd_write_avail);
1724
1725void smd_enable_read_intr(smd_channel_t *ch)
1726{
1727 if (ch)
1728 ch->send->fBLOCKREADINTR = 0;
1729}
1730EXPORT_SYMBOL(smd_enable_read_intr);
1731
1732void smd_disable_read_intr(smd_channel_t *ch)
1733{
1734 if (ch)
1735 ch->send->fBLOCKREADINTR = 1;
1736}
1737EXPORT_SYMBOL(smd_disable_read_intr);
Brian Swetland2eb44eb2008-09-29 16:00:48 -07001738
1739int smd_wait_until_readable(smd_channel_t *ch, int bytes)
1740{
1741 return -1;
1742}
1743
1744int smd_wait_until_writable(smd_channel_t *ch, int bytes)
1745{
1746 return -1;
1747}
1748
1749int smd_cur_packet_size(smd_channel_t *ch)
1750{
1751 return ch->current_packet;
1752}
Bryan Huntsman3f2bc4d2011-08-16 17:27:22 -07001753EXPORT_SYMBOL(smd_cur_packet_size);
1754
1755int smd_tiocmget(smd_channel_t *ch)
1756{
1757 return (ch->recv->fDSR ? TIOCM_DSR : 0) |
1758 (ch->recv->fCTS ? TIOCM_CTS : 0) |
1759 (ch->recv->fCD ? TIOCM_CD : 0) |
1760 (ch->recv->fRI ? TIOCM_RI : 0) |
1761 (ch->send->fCTS ? TIOCM_RTS : 0) |
1762 (ch->send->fDSR ? TIOCM_DTR : 0);
1763}
1764EXPORT_SYMBOL(smd_tiocmget);
1765
Vamsi Krishnacb12a102011-08-17 15:18:26 -07001766/* this api will be called while holding smd_lock */
1767int
1768smd_tiocmset_from_cb(smd_channel_t *ch, unsigned int set, unsigned int clear)
Bryan Huntsman3f2bc4d2011-08-16 17:27:22 -07001769{
Bryan Huntsman3f2bc4d2011-08-16 17:27:22 -07001770 if (set & TIOCM_DTR)
1771 ch->send->fDSR = 1;
1772
1773 if (set & TIOCM_RTS)
1774 ch->send->fCTS = 1;
1775
1776 if (clear & TIOCM_DTR)
1777 ch->send->fDSR = 0;
1778
1779 if (clear & TIOCM_RTS)
1780 ch->send->fCTS = 0;
1781
1782 ch->send->fSTATE = 1;
1783 barrier();
1784 ch->notify_other_cpu();
Vamsi Krishnacb12a102011-08-17 15:18:26 -07001785
1786 return 0;
1787}
1788EXPORT_SYMBOL(smd_tiocmset_from_cb);
1789
1790int smd_tiocmset(smd_channel_t *ch, unsigned int set, unsigned int clear)
1791{
1792 unsigned long flags;
1793
1794 spin_lock_irqsave(&smd_lock, flags);
1795 smd_tiocmset_from_cb(ch, set, clear);
Bryan Huntsman3f2bc4d2011-08-16 17:27:22 -07001796 spin_unlock_irqrestore(&smd_lock, flags);
1797
1798 return 0;
1799}
1800EXPORT_SYMBOL(smd_tiocmset);
Brian Swetland2eb44eb2008-09-29 16:00:48 -07001801
1802
Bryan Huntsman3f2bc4d2011-08-16 17:27:22 -07001803/* -------------------------------------------------------------------------- */
Brian Swetland2eb44eb2008-09-29 16:00:48 -07001804
Bryan Huntsman3f2bc4d2011-08-16 17:27:22 -07001805/* smem_alloc returns the pointer to smem item if it is already allocated.
1806 * Otherwise, it returns NULL.
1807 */
Brian Swetland2eb44eb2008-09-29 16:00:48 -07001808void *smem_alloc(unsigned id, unsigned size)
1809{
1810 return smem_find(id, size);
1811}
Bryan Huntsman3f2bc4d2011-08-16 17:27:22 -07001812EXPORT_SYMBOL(smem_alloc);
Brian Swetland2eb44eb2008-09-29 16:00:48 -07001813
Bryan Huntsman3f2bc4d2011-08-16 17:27:22 -07001814/* smem_alloc2 returns the pointer to smem item. If it is not allocated,
1815 * it allocates it and then returns the pointer to it.
1816 */
Angshuman Sarkar4eade0d2011-08-17 14:06:23 +05301817void *smem_alloc2(unsigned id, unsigned size_in)
Bryan Huntsman3f2bc4d2011-08-16 17:27:22 -07001818{
1819 struct smem_shared *shared = (void *) MSM_SHARED_RAM_BASE;
1820 struct smem_heap_entry *toc = shared->heap_toc;
1821 unsigned long flags;
1822 void *ret = NULL;
1823
1824 if (!shared->heap_info.initialized) {
1825 pr_err("%s: smem heap info not initialized\n", __func__);
1826 return NULL;
1827 }
1828
1829 if (id >= SMEM_NUM_ITEMS)
1830 return NULL;
1831
1832 size_in = ALIGN(size_in, 8);
1833 remote_spin_lock_irqsave(&remote_spinlock, flags);
1834 if (toc[id].allocated) {
1835 SMD_DBG("%s: %u already allocated\n", __func__, id);
1836 if (size_in != toc[id].size)
1837 pr_err("%s: wrong size %u (expected %u)\n",
1838 __func__, toc[id].size, size_in);
1839 else
1840 ret = (void *)(MSM_SHARED_RAM_BASE + toc[id].offset);
1841 } else if (id > SMEM_FIXED_ITEM_LAST) {
1842 SMD_DBG("%s: allocating %u\n", __func__, id);
1843 if (shared->heap_info.heap_remaining >= size_in) {
1844 toc[id].offset = shared->heap_info.free_offset;
1845 toc[id].size = size_in;
1846 wmb();
1847 toc[id].allocated = 1;
1848
1849 shared->heap_info.free_offset += size_in;
1850 shared->heap_info.heap_remaining -= size_in;
1851 ret = (void *)(MSM_SHARED_RAM_BASE + toc[id].offset);
1852 } else
1853 pr_err("%s: not enough memory %u (required %u)\n",
1854 __func__, shared->heap_info.heap_remaining,
1855 size_in);
1856 }
1857 wmb();
1858 remote_spin_unlock_irqrestore(&remote_spinlock, flags);
1859 return ret;
1860}
Angshuman Sarkar4eade0d2011-08-17 14:06:23 +05301861EXPORT_SYMBOL(smem_alloc2);
Bryan Huntsman3f2bc4d2011-08-16 17:27:22 -07001862
1863void *smem_get_entry(unsigned id, unsigned *size)
Brian Swetland2eb44eb2008-09-29 16:00:48 -07001864{
1865 struct smem_shared *shared = (void *) MSM_SHARED_RAM_BASE;
1866 struct smem_heap_entry *toc = shared->heap_toc;
Angshuman Sarkar7ee0dca2011-08-22 21:37:34 +05301867 int use_spinlocks = spinlocks_initialized;
1868 void *ret = 0;
1869 unsigned long flags = 0;
Brian Swetland2eb44eb2008-09-29 16:00:48 -07001870
1871 if (id >= SMEM_NUM_ITEMS)
Angshuman Sarkar7ee0dca2011-08-22 21:37:34 +05301872 return ret;
Brian Swetland2eb44eb2008-09-29 16:00:48 -07001873
Angshuman Sarkar7ee0dca2011-08-22 21:37:34 +05301874 if (use_spinlocks)
1875 remote_spin_lock_irqsave(&remote_spinlock, flags);
Bryan Huntsman3f2bc4d2011-08-16 17:27:22 -07001876 /* toc is in device memory and cannot be speculatively accessed */
Brian Swetland2eb44eb2008-09-29 16:00:48 -07001877 if (toc[id].allocated) {
1878 *size = toc[id].size;
Bryan Huntsman3f2bc4d2011-08-16 17:27:22 -07001879 barrier();
Angshuman Sarkar7ee0dca2011-08-22 21:37:34 +05301880 ret = (void *) (MSM_SHARED_RAM_BASE + toc[id].offset);
Brian Swetland5b0f5a32009-04-26 18:38:49 -07001881 } else {
1882 *size = 0;
Brian Swetland2eb44eb2008-09-29 16:00:48 -07001883 }
Angshuman Sarkar7ee0dca2011-08-22 21:37:34 +05301884 if (use_spinlocks)
1885 remote_spin_unlock_irqrestore(&remote_spinlock, flags);
Brian Swetland2eb44eb2008-09-29 16:00:48 -07001886
Angshuman Sarkar7ee0dca2011-08-22 21:37:34 +05301887 return ret;
Brian Swetland2eb44eb2008-09-29 16:00:48 -07001888}
Bryan Huntsman3f2bc4d2011-08-16 17:27:22 -07001889EXPORT_SYMBOL(smem_get_entry);
Brian Swetland2eb44eb2008-09-29 16:00:48 -07001890
1891void *smem_find(unsigned id, unsigned size_in)
1892{
1893 unsigned size;
1894 void *ptr;
1895
Bryan Huntsman3f2bc4d2011-08-16 17:27:22 -07001896 ptr = smem_get_entry(id, &size);
Brian Swetland2eb44eb2008-09-29 16:00:48 -07001897 if (!ptr)
1898 return 0;
1899
1900 size_in = ALIGN(size_in, 8);
1901 if (size_in != size) {
1902 pr_err("smem_find(%d, %d): wrong size %d\n",
1903 id, size_in, size);
1904 return 0;
1905 }
1906
1907 return ptr;
1908}
Bryan Huntsman3f2bc4d2011-08-16 17:27:22 -07001909EXPORT_SYMBOL(smem_find);
1910
1911static int smsm_cb_init(void)
1912{
Bryan Huntsman3f2bc4d2011-08-16 17:27:22 -07001913 struct smsm_state_info *state_info;
1914 int n;
1915 int ret = 0;
1916
1917 smsm_states = kmalloc(sizeof(struct smsm_state_info)*SMSM_NUM_ENTRIES,
1918 GFP_KERNEL);
1919
1920 if (!smsm_states) {
1921 pr_err("%s: SMSM init failed\n", __func__);
1922 return -ENOMEM;
1923 }
1924
Eric Holmbergc8002902011-09-16 13:55:57 -06001925 mutex_lock(&smsm_lock);
Bryan Huntsman3f2bc4d2011-08-16 17:27:22 -07001926 for (n = 0; n < SMSM_NUM_ENTRIES; n++) {
1927 state_info = &smsm_states[n];
1928 state_info->last_value = __raw_readl(SMSM_STATE_ADDR(n));
1929 INIT_LIST_HEAD(&state_info->callbacks);
1930 }
Eric Holmbergc8002902011-09-16 13:55:57 -06001931 mutex_unlock(&smsm_lock);
Bryan Huntsman3f2bc4d2011-08-16 17:27:22 -07001932
1933 return ret;
1934}
1935
1936static int smsm_init(void)
1937{
1938 struct smem_shared *shared = (void *) MSM_SHARED_RAM_BASE;
1939 int i;
1940 struct smsm_size_info_type *smsm_size_info;
1941
1942 i = remote_spin_lock_init(&remote_spinlock, SMEM_SPINLOCK_SMEM_ALLOC);
1943 if (i) {
1944 pr_err("%s: remote spinlock init failed %d\n", __func__, i);
1945 return i;
1946 }
Angshuman Sarkar7ee0dca2011-08-22 21:37:34 +05301947 spinlocks_initialized = 1;
Bryan Huntsman3f2bc4d2011-08-16 17:27:22 -07001948
1949 smsm_size_info = smem_alloc(SMEM_SMSM_SIZE_INFO,
1950 sizeof(struct smsm_size_info_type));
1951 if (smsm_size_info) {
1952 SMSM_NUM_ENTRIES = smsm_size_info->num_entries;
1953 SMSM_NUM_HOSTS = smsm_size_info->num_hosts;
1954 }
1955
1956 if (!smsm_info.state) {
1957 smsm_info.state = smem_alloc2(ID_SHARED_STATE,
1958 SMSM_NUM_ENTRIES *
1959 sizeof(uint32_t));
1960
1961 if (smsm_info.state) {
1962 __raw_writel(0, SMSM_STATE_ADDR(SMSM_APPS_STATE));
1963 if ((shared->version[VERSION_MODEM] >> 16) >= 0xB)
1964 __raw_writel(0, \
1965 SMSM_STATE_ADDR(SMSM_APPS_DEM_I));
1966 }
1967 }
1968
1969 if (!smsm_info.intr_mask) {
1970 smsm_info.intr_mask = smem_alloc2(SMEM_SMSM_CPU_INTR_MASK,
1971 SMSM_NUM_ENTRIES *
1972 SMSM_NUM_HOSTS *
1973 sizeof(uint32_t));
1974
1975 if (smsm_info.intr_mask)
1976 for (i = 0; i < SMSM_NUM_ENTRIES; i++)
1977 __raw_writel(0xffffffff,
1978 SMSM_INTR_MASK_ADDR(i, SMSM_APPS));
1979 }
1980
1981 if (!smsm_info.intr_mux)
1982 smsm_info.intr_mux = smem_alloc2(SMEM_SMD_SMSM_INTR_MUX,
1983 SMSM_NUM_INTR_MUX *
1984 sizeof(uint32_t));
1985
1986 i = smsm_cb_init();
1987 if (i)
1988 return i;
1989
1990 wmb();
1991 return 0;
1992}
1993
1994void smsm_reset_modem(unsigned mode)
1995{
1996 if (mode == SMSM_SYSTEM_DOWNLOAD) {
1997 mode = SMSM_RESET | SMSM_SYSTEM_DOWNLOAD;
1998 } else if (mode == SMSM_MODEM_WAIT) {
1999 mode = SMSM_RESET | SMSM_MODEM_WAIT;
2000 } else { /* reset_mode is SMSM_RESET or default */
2001 mode = SMSM_RESET;
2002 }
2003
2004 smsm_change_state(SMSM_APPS_STATE, mode, mode);
2005}
2006EXPORT_SYMBOL(smsm_reset_modem);
2007
2008void smsm_reset_modem_cont(void)
2009{
2010 unsigned long flags;
2011 uint32_t state;
2012
2013 if (!smsm_info.state)
2014 return;
2015
2016 spin_lock_irqsave(&smem_lock, flags);
2017 state = __raw_readl(SMSM_STATE_ADDR(SMSM_APPS_STATE)) \
2018 & ~SMSM_MODEM_WAIT;
2019 __raw_writel(state, SMSM_STATE_ADDR(SMSM_APPS_STATE));
2020 wmb();
2021 spin_unlock_irqrestore(&smem_lock, flags);
2022}
2023EXPORT_SYMBOL(smsm_reset_modem_cont);
Brian Swetland2eb44eb2008-09-29 16:00:48 -07002024
2025static irqreturn_t smsm_irq_handler(int irq, void *data)
2026{
2027 unsigned long flags;
Bryan Huntsman3f2bc4d2011-08-16 17:27:22 -07002028
Bryan Huntsman3f2bc4d2011-08-16 17:27:22 -07002029 if (irq == INT_ADSP_A11_SMSM) {
Eric Holmberg6282c5d2011-10-27 17:30:57 -06002030 uint32_t mux_val;
2031 static uint32_t prev_smem_q6_apps_smsm;
2032
2033 if (smsm_info.intr_mux && cpu_is_qsd8x50()) {
2034 mux_val = __raw_readl(
2035 SMSM_INTR_MUX_ADDR(SMEM_Q6_APPS_SMSM));
2036 if (mux_val != prev_smem_q6_apps_smsm)
2037 prev_smem_q6_apps_smsm = mux_val;
2038 }
2039
2040 schedule_work(&smsm_cb_work);
Bryan Huntsman3f2bc4d2011-08-16 17:27:22 -07002041 return IRQ_HANDLED;
2042 }
Brian Swetland2eb44eb2008-09-29 16:00:48 -07002043
2044 spin_lock_irqsave(&smem_lock, flags);
Bryan Huntsman3f2bc4d2011-08-16 17:27:22 -07002045 if (!smsm_info.state) {
2046 SMSM_INFO("<SM NO STATE>\n");
2047 } else {
2048 unsigned old_apps, apps;
2049 unsigned modm = __raw_readl(SMSM_STATE_ADDR(SMSM_MODEM_STATE));
Brian Swetland2eb44eb2008-09-29 16:00:48 -07002050
Bryan Huntsman3f2bc4d2011-08-16 17:27:22 -07002051 old_apps = apps = __raw_readl(SMSM_STATE_ADDR(SMSM_APPS_STATE));
Brian Swetland2eb44eb2008-09-29 16:00:48 -07002052
Bryan Huntsman3f2bc4d2011-08-16 17:27:22 -07002053 SMSM_DBG("<SM %08x %08x>\n", apps, modm);
2054 if (apps & SMSM_RESET) {
2055 /* If we get an interrupt and the apps SMSM_RESET
2056 bit is already set, the modem is acking the
2057 app's reset ack. */
Stepan Moskovchenkoa1ca7582011-10-25 14:45:09 -07002058 if (!cpu_is_msm8960() && !cpu_is_msm8930())
Angshuman Sarkaread67bd2011-09-21 20:13:12 +05302059 apps &= ~SMSM_RESET;
Bryan Huntsman3f2bc4d2011-08-16 17:27:22 -07002060 /* Issue a fake irq to handle any
2061 * smd state changes during reset
2062 */
2063 smd_fake_irq_handler(0);
Brian Swetland5b0f5a32009-04-26 18:38:49 -07002064
Bryan Huntsman3f2bc4d2011-08-16 17:27:22 -07002065 /* queue modem restart notify chain */
2066 modem_queue_start_reset_notify();
2067
2068 } else if (modm & SMSM_RESET) {
Stepan Moskovchenkoa1ca7582011-10-25 14:45:09 -07002069 if (!cpu_is_msm8960() && !cpu_is_msm8930())
Angshuman Sarkaread67bd2011-09-21 20:13:12 +05302070 apps |= SMSM_RESET;
Bryan Huntsman3f2bc4d2011-08-16 17:27:22 -07002071
2072 pr_err("\nSMSM: Modem SMSM state changed to SMSM_RESET.");
2073 modem_queue_start_reset_notify();
2074
2075 } else if (modm & SMSM_INIT) {
2076 if (!(apps & SMSM_INIT)) {
2077 apps |= SMSM_INIT;
2078 modem_queue_smsm_init_notify();
2079 }
2080
2081 if (modm & SMSM_SMDINIT)
2082 apps |= SMSM_SMDINIT;
2083 if ((apps & (SMSM_INIT | SMSM_SMDINIT | SMSM_RPCINIT)) ==
2084 (SMSM_INIT | SMSM_SMDINIT | SMSM_RPCINIT))
2085 apps |= SMSM_RUN;
2086 } else if (modm & SMSM_SYSTEM_DOWNLOAD) {
2087 pr_err("\nSMSM: Modem SMSM state changed to SMSM_SYSTEM_DOWNLOAD.");
2088 modem_queue_start_reset_notify();
2089 }
2090
2091 if (old_apps != apps) {
2092 SMSM_DBG("<SM %08x NOTIFY>\n", apps);
2093 __raw_writel(apps, SMSM_STATE_ADDR(SMSM_APPS_STATE));
2094 do_smd_probe();
2095 notify_other_smsm(SMSM_APPS_STATE, (old_apps ^ apps));
2096 }
2097
2098 schedule_work(&smsm_cb_work);
2099 }
Brian Swetland2eb44eb2008-09-29 16:00:48 -07002100 spin_unlock_irqrestore(&smem_lock, flags);
2101 return IRQ_HANDLED;
2102}
2103
Bryan Huntsman3f2bc4d2011-08-16 17:27:22 -07002104int smsm_change_intr_mask(uint32_t smsm_entry,
2105 uint32_t clear_mask, uint32_t set_mask)
Brian Swetland2eb44eb2008-09-29 16:00:48 -07002106{
Bryan Huntsman3f2bc4d2011-08-16 17:27:22 -07002107 uint32_t old_mask, new_mask;
Brian Swetland2eb44eb2008-09-29 16:00:48 -07002108 unsigned long flags;
Brian Swetland5b0f5a32009-04-26 18:38:49 -07002109
Bryan Huntsman3f2bc4d2011-08-16 17:27:22 -07002110 if (smsm_entry >= SMSM_NUM_ENTRIES) {
2111 pr_err("smsm_change_state: Invalid entry %d\n",
2112 smsm_entry);
2113 return -EINVAL;
2114 }
2115
2116 if (!smsm_info.intr_mask) {
2117 pr_err("smsm_change_intr_mask <SM NO STATE>\n");
Brian Swetland5b0f5a32009-04-26 18:38:49 -07002118 return -EIO;
Bryan Huntsman3f2bc4d2011-08-16 17:27:22 -07002119 }
Brian Swetland2eb44eb2008-09-29 16:00:48 -07002120
2121 spin_lock_irqsave(&smem_lock, flags);
2122
Bryan Huntsman3f2bc4d2011-08-16 17:27:22 -07002123 old_mask = __raw_readl(SMSM_INTR_MASK_ADDR(smsm_entry, SMSM_APPS));
2124 new_mask = (old_mask & ~clear_mask) | set_mask;
2125 __raw_writel(new_mask, SMSM_INTR_MASK_ADDR(smsm_entry, SMSM_APPS));
Brian Swetland2eb44eb2008-09-29 16:00:48 -07002126
Bryan Huntsman3f2bc4d2011-08-16 17:27:22 -07002127 wmb();
2128 spin_unlock_irqrestore(&smem_lock, flags);
Brian Swetland5b0f5a32009-04-26 18:38:49 -07002129
Bryan Huntsman3f2bc4d2011-08-16 17:27:22 -07002130 return 0;
2131}
2132EXPORT_SYMBOL(smsm_change_intr_mask);
2133
2134int smsm_get_intr_mask(uint32_t smsm_entry, uint32_t *intr_mask)
2135{
2136 if (smsm_entry >= SMSM_NUM_ENTRIES) {
2137 pr_err("smsm_change_state: Invalid entry %d\n",
2138 smsm_entry);
2139 return -EINVAL;
2140 }
2141
2142 if (!smsm_info.intr_mask) {
2143 pr_err("smsm_change_intr_mask <SM NO STATE>\n");
2144 return -EIO;
2145 }
2146
2147 *intr_mask = __raw_readl(SMSM_INTR_MASK_ADDR(smsm_entry, SMSM_APPS));
2148 return 0;
2149}
2150EXPORT_SYMBOL(smsm_get_intr_mask);
2151
2152int smsm_change_state(uint32_t smsm_entry,
2153 uint32_t clear_mask, uint32_t set_mask)
2154{
2155 unsigned long flags;
2156 uint32_t old_state, new_state;
2157
2158 if (smsm_entry >= SMSM_NUM_ENTRIES) {
2159 pr_err("smsm_change_state: Invalid entry %d",
2160 smsm_entry);
2161 return -EINVAL;
2162 }
2163
2164 if (!smsm_info.state) {
2165 pr_err("smsm_change_state <SM NO STATE>\n");
2166 return -EIO;
2167 }
2168 spin_lock_irqsave(&smem_lock, flags);
2169
2170 old_state = __raw_readl(SMSM_STATE_ADDR(smsm_entry));
2171 new_state = (old_state & ~clear_mask) | set_mask;
2172 __raw_writel(new_state, SMSM_STATE_ADDR(smsm_entry));
2173 SMSM_DBG("smsm_change_state %x\n", new_state);
2174 notify_other_smsm(SMSM_APPS_STATE, (old_state ^ new_state));
Brian Swetland2eb44eb2008-09-29 16:00:48 -07002175
2176 spin_unlock_irqrestore(&smem_lock, flags);
2177
Brian Swetland2eb44eb2008-09-29 16:00:48 -07002178 return 0;
2179}
Bryan Huntsman3f2bc4d2011-08-16 17:27:22 -07002180EXPORT_SYMBOL(smsm_change_state);
Brian Swetland2eb44eb2008-09-29 16:00:48 -07002181
Bryan Huntsman3f2bc4d2011-08-16 17:27:22 -07002182uint32_t smsm_get_state(uint32_t smsm_entry)
Brian Swetland2eb44eb2008-09-29 16:00:48 -07002183{
Bryan Huntsman3f2bc4d2011-08-16 17:27:22 -07002184 uint32_t rv = 0;
Brian Swetland2eb44eb2008-09-29 16:00:48 -07002185
Bryan Huntsman3f2bc4d2011-08-16 17:27:22 -07002186 /* needs interface change to return error code */
2187 if (smsm_entry >= SMSM_NUM_ENTRIES) {
2188 pr_err("smsm_change_state: Invalid entry %d",
2189 smsm_entry);
2190 return 0;
2191 }
Brian Swetland2eb44eb2008-09-29 16:00:48 -07002192
Bryan Huntsman3f2bc4d2011-08-16 17:27:22 -07002193 if (!smsm_info.state) {
2194 pr_err("smsm_get_state <SM NO STATE>\n");
2195 } else {
2196 rv = __raw_readl(SMSM_STATE_ADDR(smsm_entry));
2197 }
Brian Swetland2eb44eb2008-09-29 16:00:48 -07002198
Brian Swetland2eb44eb2008-09-29 16:00:48 -07002199 return rv;
2200}
Bryan Huntsman3f2bc4d2011-08-16 17:27:22 -07002201EXPORT_SYMBOL(smsm_get_state);
Brian Swetland2eb44eb2008-09-29 16:00:48 -07002202
Bryan Huntsman3f2bc4d2011-08-16 17:27:22 -07002203/**
2204 * Performs SMSM callback client notifiction.
2205 */
2206void notify_smsm_cb_clients_worker(struct work_struct *work)
Arve Hjønnevågec9d3d12009-06-16 14:48:21 -07002207{
Bryan Huntsman3f2bc4d2011-08-16 17:27:22 -07002208 struct smsm_state_cb_info *cb_info;
2209 struct smsm_state_info *state_info;
2210 int n;
2211 uint32_t new_state;
2212 uint32_t state_changes;
Brian Swetland03e00cd2009-07-01 17:58:37 -07002213
Eric Holmbergc8002902011-09-16 13:55:57 -06002214 mutex_lock(&smsm_lock);
Bryan Huntsman3f2bc4d2011-08-16 17:27:22 -07002215
2216 if (!smsm_states) {
2217 /* smsm not yet initialized */
Eric Holmbergc8002902011-09-16 13:55:57 -06002218 mutex_unlock(&smsm_lock);
Bryan Huntsman3f2bc4d2011-08-16 17:27:22 -07002219 return;
Arve Hjønnevågec9d3d12009-06-16 14:48:21 -07002220 }
Bryan Huntsman3f2bc4d2011-08-16 17:27:22 -07002221
2222 for (n = 0; n < SMSM_NUM_ENTRIES; n++) {
2223 state_info = &smsm_states[n];
2224 new_state = __raw_readl(SMSM_STATE_ADDR(n));
2225
2226 if (new_state != state_info->last_value) {
2227 state_changes = state_info->last_value ^ new_state;
2228
2229 list_for_each_entry(cb_info,
2230 &state_info->callbacks, cb_list) {
2231
2232 if (cb_info->mask & state_changes)
2233 cb_info->notify(cb_info->data,
2234 state_info->last_value,
2235 new_state);
2236 }
2237 state_info->last_value = new_state;
2238 }
2239 }
2240
Eric Holmbergc8002902011-09-16 13:55:57 -06002241 mutex_unlock(&smsm_lock);
Arve Hjønnevågec9d3d12009-06-16 14:48:21 -07002242}
2243
Arve Hjønnevågec9d3d12009-06-16 14:48:21 -07002244
Bryan Huntsman3f2bc4d2011-08-16 17:27:22 -07002245/**
2246 * Registers callback for SMSM state notifications when the specified
2247 * bits change.
2248 *
2249 * @smsm_entry Processor entry to deregister
2250 * @mask Bits to deregister (if result is 0, callback is removed)
2251 * @notify Notification function to deregister
2252 * @data Opaque data passed in to callback
2253 *
2254 * @returns Status code
2255 * <0 error code
2256 * 0 inserted new entry
2257 * 1 updated mask of existing entry
2258 */
2259int smsm_state_cb_register(uint32_t smsm_entry, uint32_t mask,
2260 void (*notify)(void *, uint32_t, uint32_t), void *data)
Brian Swetland2eb44eb2008-09-29 16:00:48 -07002261{
Bryan Huntsman3f2bc4d2011-08-16 17:27:22 -07002262 struct smsm_state_cb_info *cb_info;
2263 struct smsm_state_cb_info *cb_found = 0;
2264 int ret = 0;
Brian Swetland2eb44eb2008-09-29 16:00:48 -07002265
Bryan Huntsman3f2bc4d2011-08-16 17:27:22 -07002266 if (smsm_entry >= SMSM_NUM_ENTRIES)
2267 return -EINVAL;
2268
Eric Holmbergc8002902011-09-16 13:55:57 -06002269 mutex_lock(&smsm_lock);
Bryan Huntsman3f2bc4d2011-08-16 17:27:22 -07002270
2271 if (!smsm_states) {
2272 /* smsm not yet initialized */
2273 ret = -ENODEV;
2274 goto cleanup;
Brian Swetland2eb44eb2008-09-29 16:00:48 -07002275 }
Brian Swetland2eb44eb2008-09-29 16:00:48 -07002276
Bryan Huntsman3f2bc4d2011-08-16 17:27:22 -07002277 list_for_each_entry(cb_info,
2278 &smsm_states[smsm_entry].callbacks, cb_list) {
2279 if ((cb_info->notify == notify) &&
2280 (cb_info->data == data)) {
2281 cb_info->mask |= mask;
2282 cb_found = cb_info;
2283 ret = 1;
Brian Swetland5b0f5a32009-04-26 18:38:49 -07002284 break;
2285 }
2286 }
2287
Bryan Huntsman3f2bc4d2011-08-16 17:27:22 -07002288 if (!cb_found) {
2289 cb_info = kmalloc(sizeof(struct smsm_state_cb_info),
2290 GFP_ATOMIC);
2291 if (!cb_info) {
2292 ret = -ENOMEM;
2293 goto cleanup;
2294 }
2295
2296 cb_info->mask = mask;
2297 cb_info->notify = notify;
2298 cb_info->data = data;
2299 INIT_LIST_HEAD(&cb_info->cb_list);
2300 list_add_tail(&cb_info->cb_list,
2301 &smsm_states[smsm_entry].callbacks);
2302 }
2303
2304cleanup:
Eric Holmbergc8002902011-09-16 13:55:57 -06002305 mutex_unlock(&smsm_lock);
Bryan Huntsman3f2bc4d2011-08-16 17:27:22 -07002306 return ret;
2307}
2308EXPORT_SYMBOL(smsm_state_cb_register);
2309
2310
2311/**
2312 * Deregisters for SMSM state notifications for the specified bits.
2313 *
2314 * @smsm_entry Processor entry to deregister
2315 * @mask Bits to deregister (if result is 0, callback is removed)
2316 * @notify Notification function to deregister
2317 * @data Opaque data passed in to callback
2318 *
2319 * @returns Status code
2320 * <0 error code
2321 * 0 not found
2322 * 1 updated mask
2323 * 2 removed callback
2324 */
2325int smsm_state_cb_deregister(uint32_t smsm_entry, uint32_t mask,
2326 void (*notify)(void *, uint32_t, uint32_t), void *data)
2327{
Bryan Huntsman3f2bc4d2011-08-16 17:27:22 -07002328 struct smsm_state_cb_info *cb_info;
2329 int ret = 0;
2330
2331 if (smsm_entry >= SMSM_NUM_ENTRIES)
2332 return -EINVAL;
2333
Eric Holmbergc8002902011-09-16 13:55:57 -06002334 mutex_lock(&smsm_lock);
Bryan Huntsman3f2bc4d2011-08-16 17:27:22 -07002335
2336 if (!smsm_states) {
2337 /* smsm not yet initialized */
Eric Holmbergc8002902011-09-16 13:55:57 -06002338 mutex_unlock(&smsm_lock);
Bryan Huntsman3f2bc4d2011-08-16 17:27:22 -07002339 return -ENODEV;
2340 }
2341
2342 list_for_each_entry(cb_info,
2343 &smsm_states[smsm_entry].callbacks, cb_list) {
2344 if ((cb_info->notify == notify) &&
2345 (cb_info->data == data)) {
2346 cb_info->mask &= ~mask;
2347 ret = 1;
2348 if (!cb_info->mask) {
2349 /* no mask bits set, remove callback */
2350 list_del(&cb_info->cb_list);
2351 kfree(cb_info);
2352 ret = 2;
2353 }
2354 break;
2355 }
2356 }
2357
Eric Holmbergc8002902011-09-16 13:55:57 -06002358 mutex_unlock(&smsm_lock);
Bryan Huntsman3f2bc4d2011-08-16 17:27:22 -07002359 return ret;
2360}
2361EXPORT_SYMBOL(smsm_state_cb_deregister);
2362
2363
2364int smd_core_init(void)
2365{
2366 int r;
2367 unsigned long flags = IRQF_TRIGGER_RISING;
2368 SMD_INFO("smd_core_init()\n");
Brian Swetland5b0f5a32009-04-26 18:38:49 -07002369
Brian Swetland37521a32009-07-01 18:30:47 -07002370 r = request_irq(INT_A9_M2A_0, smd_modem_irq_handler,
Bryan Huntsman3f2bc4d2011-08-16 17:27:22 -07002371 flags, "smd_dev", 0);
Brian Swetland2eb44eb2008-09-29 16:00:48 -07002372 if (r < 0)
2373 return r;
2374 r = enable_irq_wake(INT_A9_M2A_0);
2375 if (r < 0)
Bryan Huntsman3f2bc4d2011-08-16 17:27:22 -07002376 pr_err("smd_core_init: "
2377 "enable_irq_wake failed for INT_A9_M2A_0\n");
Brian Swetland2eb44eb2008-09-29 16:00:48 -07002378
2379 r = request_irq(INT_A9_M2A_5, smsm_irq_handler,
Bryan Huntsman3f2bc4d2011-08-16 17:27:22 -07002380 flags, "smsm_dev", 0);
Brian Swetland2eb44eb2008-09-29 16:00:48 -07002381 if (r < 0) {
2382 free_irq(INT_A9_M2A_0, 0);
2383 return r;
2384 }
2385 r = enable_irq_wake(INT_A9_M2A_5);
2386 if (r < 0)
Bryan Huntsman3f2bc4d2011-08-16 17:27:22 -07002387 pr_err("smd_core_init: "
2388 "enable_irq_wake failed for INT_A9_M2A_5\n");
Brian Swetland2eb44eb2008-09-29 16:00:48 -07002389
Brian Swetland37521a32009-07-01 18:30:47 -07002390#if defined(CONFIG_QDSP6)
Bryan Huntsman3f2bc4d2011-08-16 17:27:22 -07002391#if (INT_ADSP_A11 == INT_ADSP_A11_SMSM)
2392 flags |= IRQF_SHARED;
2393#endif
Brian Swetland37521a32009-07-01 18:30:47 -07002394 r = request_irq(INT_ADSP_A11, smd_dsp_irq_handler,
Bryan Huntsman3f2bc4d2011-08-16 17:27:22 -07002395 flags, "smd_dev", smd_dsp_irq_handler);
Brian Swetland37521a32009-07-01 18:30:47 -07002396 if (r < 0) {
2397 free_irq(INT_A9_M2A_0, 0);
2398 free_irq(INT_A9_M2A_5, 0);
2399 return r;
2400 }
Bryan Huntsman3f2bc4d2011-08-16 17:27:22 -07002401
2402 r = request_irq(INT_ADSP_A11_SMSM, smsm_irq_handler,
2403 flags, "smsm_dev", smsm_irq_handler);
2404 if (r < 0) {
2405 free_irq(INT_A9_M2A_0, 0);
2406 free_irq(INT_A9_M2A_5, 0);
2407 free_irq(INT_ADSP_A11, smd_dsp_irq_handler);
2408 return r;
2409 }
2410
2411 r = enable_irq_wake(INT_ADSP_A11);
2412 if (r < 0)
2413 pr_err("smd_core_init: "
2414 "enable_irq_wake failed for INT_ADSP_A11\n");
2415
2416#if (INT_ADSP_A11 != INT_ADSP_A11_SMSM)
2417 r = enable_irq_wake(INT_ADSP_A11_SMSM);
2418 if (r < 0)
2419 pr_err("smd_core_init: enable_irq_wake "
2420 "failed for INT_ADSP_A11_SMSM\n");
2421#endif
2422 flags &= ~IRQF_SHARED;
Brian Swetland37521a32009-07-01 18:30:47 -07002423#endif
2424
Bryan Huntsman3f2bc4d2011-08-16 17:27:22 -07002425#if defined(CONFIG_DSPS)
2426 r = request_irq(INT_DSPS_A11, smd_dsps_irq_handler,
2427 flags, "smd_dev", smd_dsps_irq_handler);
2428 if (r < 0) {
2429 free_irq(INT_A9_M2A_0, 0);
2430 free_irq(INT_A9_M2A_5, 0);
2431 free_irq(INT_ADSP_A11, smd_dsp_irq_handler);
2432 free_irq(INT_ADSP_A11_SMSM, smsm_irq_handler);
2433 return r;
2434 }
Brian Swetland5b0f5a32009-04-26 18:38:49 -07002435
Bryan Huntsman3f2bc4d2011-08-16 17:27:22 -07002436 r = enable_irq_wake(INT_DSPS_A11);
2437 if (r < 0)
2438 pr_err("smd_core_init: "
2439 "enable_irq_wake failed for INT_ADSP_A11\n");
Arve Hjønnevågec9d3d12009-06-16 14:48:21 -07002440#endif
Brian Swetland2eb44eb2008-09-29 16:00:48 -07002441
Bryan Huntsman3f2bc4d2011-08-16 17:27:22 -07002442#if defined(CONFIG_WCNSS)
2443 r = request_irq(INT_WCNSS_A11, smd_wcnss_irq_handler,
2444 flags, "smd_dev", smd_wcnss_irq_handler);
2445 if (r < 0) {
2446 free_irq(INT_A9_M2A_0, 0);
2447 free_irq(INT_A9_M2A_5, 0);
2448 free_irq(INT_ADSP_A11, smd_dsp_irq_handler);
2449 free_irq(INT_ADSP_A11_SMSM, smsm_irq_handler);
2450 free_irq(INT_DSPS_A11, smd_dsps_irq_handler);
2451 return r;
2452 }
2453
2454 r = enable_irq_wake(INT_WCNSS_A11);
2455 if (r < 0)
2456 pr_err("smd_core_init: "
2457 "enable_irq_wake failed for INT_WCNSS_A11\n");
2458
2459 r = request_irq(INT_WCNSS_A11_SMSM, smsm_irq_handler,
2460 flags, "smsm_dev", smsm_irq_handler);
2461 if (r < 0) {
2462 free_irq(INT_A9_M2A_0, 0);
2463 free_irq(INT_A9_M2A_5, 0);
2464 free_irq(INT_ADSP_A11, smd_dsp_irq_handler);
2465 free_irq(INT_ADSP_A11_SMSM, smsm_irq_handler);
2466 free_irq(INT_DSPS_A11, smd_dsps_irq_handler);
2467 free_irq(INT_WCNSS_A11, smd_wcnss_irq_handler);
2468 return r;
2469 }
2470
2471 r = enable_irq_wake(INT_WCNSS_A11_SMSM);
2472 if (r < 0)
2473 pr_err("smd_core_init: "
2474 "enable_irq_wake failed for INT_WCNSS_A11_SMSM\n");
2475#endif
2476
Jeff Hugo6a8057c2011-08-16 13:47:12 -06002477#if defined(CONFIG_DSPS_SMSM)
2478 r = request_irq(INT_DSPS_A11_SMSM, smsm_irq_handler,
2479 flags, "smsm_dev", smsm_irq_handler);
2480 if (r < 0) {
2481 free_irq(INT_A9_M2A_0, 0);
2482 free_irq(INT_A9_M2A_5, 0);
2483 free_irq(INT_ADSP_A11, smd_dsp_irq_handler);
2484 free_irq(INT_ADSP_A11_SMSM, smsm_irq_handler);
2485 free_irq(INT_DSPS_A11, smd_dsps_irq_handler);
2486 free_irq(INT_WCNSS_A11, smd_wcnss_irq_handler);
2487 free_irq(INT_WCNSS_A11_SMSM, smsm_irq_handler);
2488 return r;
2489 }
2490
2491 r = enable_irq_wake(INT_DSPS_A11_SMSM);
2492 if (r < 0)
2493 pr_err("smd_core_init: "
2494 "enable_irq_wake failed for INT_DSPS_A11_SMSM\n");
2495#endif
2496
Bryan Huntsman3f2bc4d2011-08-16 17:27:22 -07002497 /* we may have missed a signal while booting -- fake
2498 * an interrupt to make sure we process any existing
2499 * state
2500 */
2501 smsm_irq_handler(0, 0);
2502
2503 SMD_INFO("smd_core_init() done\n");
2504
Brian Swetland2eb44eb2008-09-29 16:00:48 -07002505 return 0;
2506}
2507
Gregory Bean4416e9e2010-07-28 10:22:12 -07002508static int __devinit msm_smd_probe(struct platform_device *pdev)
Brian Swetland2eb44eb2008-09-29 16:00:48 -07002509{
Bryan Huntsman3f2bc4d2011-08-16 17:27:22 -07002510 SMD_INFO("smd probe\n");
Daniel Walker0aec66d2010-03-18 12:31:08 -07002511
Brian Swetland2eb44eb2008-09-29 16:00:48 -07002512 INIT_WORK(&probe_work, smd_channel_probe_worker);
2513
Bryan Huntsman3f2bc4d2011-08-16 17:27:22 -07002514 channel_close_wq = create_singlethread_workqueue("smd_channel_close");
2515 if (IS_ERR(channel_close_wq)) {
2516 pr_err("%s: create_singlethread_workqueue ENOMEM\n", __func__);
2517 return -ENOMEM;
2518 }
2519
2520 if (smsm_init()) {
2521 pr_err("smsm_init() failed\n");
2522 return -1;
2523 }
2524
Brian Swetland2eb44eb2008-09-29 16:00:48 -07002525 if (smd_core_init()) {
2526 pr_err("smd_core_init() failed\n");
2527 return -1;
2528 }
2529
Brian Swetland2eb44eb2008-09-29 16:00:48 -07002530 smd_initialized = 1;
2531
Bryan Huntsman3f2bc4d2011-08-16 17:27:22 -07002532 smd_alloc_loopback_channel();
Eric Holmbergc33d4ab2011-10-24 10:28:25 -06002533 smsm_irq_handler(0, 0);
2534 tasklet_schedule(&smd_fake_irq_tasklet);
Bryan Huntsman3f2bc4d2011-08-16 17:27:22 -07002535
Brian Swetland2eb44eb2008-09-29 16:00:48 -07002536 return 0;
2537}
2538
Bryan Huntsman3f2bc4d2011-08-16 17:27:22 -07002539static int restart_notifier_cb(struct notifier_block *this,
2540 unsigned long code,
2541 void *data);
2542
2543static struct restart_notifier_block restart_notifiers[] = {
Eric Holmbergca7ead22011-12-01 17:21:15 -07002544 {SMD_MODEM, "modem", .nb.notifier_call = restart_notifier_cb},
2545 {SMD_Q6, "lpass", .nb.notifier_call = restart_notifier_cb},
2546 {SMD_WCNSS, "riva", .nb.notifier_call = restart_notifier_cb},
2547 {SMD_DSPS, "dsps", .nb.notifier_call = restart_notifier_cb},
Bryan Huntsman3f2bc4d2011-08-16 17:27:22 -07002548};
2549
2550static int restart_notifier_cb(struct notifier_block *this,
2551 unsigned long code,
2552 void *data)
2553{
2554 if (code == SUBSYS_AFTER_SHUTDOWN) {
2555 struct restart_notifier_block *notifier;
2556
2557 notifier = container_of(this,
2558 struct restart_notifier_block, nb);
2559 SMD_INFO("%s: ssrestart for processor %d ('%s')\n",
2560 __func__, notifier->processor,
2561 notifier->name);
2562
2563 smd_channel_reset(notifier->processor);
2564 }
2565
2566 return NOTIFY_DONE;
2567}
2568
2569static __init int modem_restart_late_init(void)
2570{
2571 int i;
2572 void *handle;
2573 struct restart_notifier_block *nb;
2574
2575 for (i = 0; i < ARRAY_SIZE(restart_notifiers); i++) {
2576 nb = &restart_notifiers[i];
2577 handle = subsys_notif_register_notifier(nb->name, &nb->nb);
2578 SMD_DBG("%s: registering notif for '%s', handle=%p\n",
2579 __func__, nb->name, handle);
2580 }
2581 return 0;
2582}
2583late_initcall(modem_restart_late_init);
2584
Brian Swetland2eb44eb2008-09-29 16:00:48 -07002585static struct platform_driver msm_smd_driver = {
2586 .probe = msm_smd_probe,
2587 .driver = {
2588 .name = MODULE_NAME,
2589 .owner = THIS_MODULE,
2590 },
2591};
2592
2593static int __init msm_smd_init(void)
2594{
2595 return platform_driver_register(&msm_smd_driver);
2596}
2597
2598module_init(msm_smd_init);
2599
2600MODULE_DESCRIPTION("MSM Shared Memory Core");
2601MODULE_AUTHOR("Brian Swetland <swetland@google.com>");
2602MODULE_LICENSE("GPL");