/* Copyright (c) 2010-2012, The Linux Foundation. All rights reserved. * * This program is free software; you can redistribute it and/or modify * it under the terms of the GNU General Public License version 2 and * only version 2 as published by the Free Software Foundation. * * This program is distributed in the hope that it will be useful, * but WITHOUT ANY WARRANTY; without even the implied warranty of * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the * GNU General Public License for more details. * */ #include #include #include #include #include #include #include #include #include #include #include #include #include #include #include #include #include #include #include #include /****************************************************************************** * Data type and structure definitions *****************************************************************************/ struct msm_rpm_request { struct msm_rpm_iv_pair *req; int count; uint32_t *ctx_mask_ack; uint32_t *sel_masks_ack; struct completion *done; }; struct msm_rpm_notif_config { struct msm_rpm_iv_pair iv[SEL_MASK_SIZE * 2]; }; #define configured_iv(notif_cfg) ((notif_cfg)->iv) #define registered_iv(notif_cfg) ((notif_cfg)->iv + msm_rpm_sel_mask_size) static uint32_t msm_rpm_sel_mask_size; static struct msm_rpm_platform_data msm_rpm_data; static DEFINE_MUTEX(msm_rpm_mutex); static DEFINE_SPINLOCK(msm_rpm_lock); static DEFINE_SPINLOCK(msm_rpm_irq_lock); static struct msm_rpm_request *msm_rpm_request; static struct msm_rpm_request msm_rpm_request_irq_mode; static struct msm_rpm_request msm_rpm_request_poll_mode; static LIST_HEAD(msm_rpm_notifications); static struct msm_rpm_notif_config msm_rpm_notif_cfgs[MSM_RPM_CTX_SET_COUNT]; static bool msm_rpm_init_notif_done; /****************************************************************************** * Internal functions *****************************************************************************/ static inline unsigned int target_enum(unsigned int id) { BUG_ON(id >= MSM_RPM_ID_LAST); return msm_rpm_data.target_id[id].id; } static inline unsigned int target_status(unsigned int id) { BUG_ON(id >= MSM_RPM_STATUS_ID_LAST); return msm_rpm_data.target_status[id]; } static inline unsigned int target_ctrl(unsigned int id) { BUG_ON(id >= MSM_RPM_CTRL_LAST); return msm_rpm_data.target_ctrl_id[id]; } static inline uint32_t msm_rpm_read(unsigned int page, unsigned int reg) { return __raw_readl(msm_rpm_data.reg_base_addrs[page] + reg * 4); } static inline void msm_rpm_write( unsigned int page, unsigned int reg, uint32_t value) { __raw_writel(value, msm_rpm_data.reg_base_addrs[page] + reg * 4); } static inline void msm_rpm_read_contiguous( unsigned int page, unsigned int reg, uint32_t *values, int count) { int i; for (i = 0; i < count; i++) values[i] = msm_rpm_read(page, reg + i); } static inline void msm_rpm_write_contiguous( unsigned int page, unsigned int reg, uint32_t *values, int count) { int i; for (i = 0; i < count; i++) msm_rpm_write(page, reg + i, values[i]); } static inline void msm_rpm_write_contiguous_zeros( unsigned int page, unsigned int reg, int count) { int i; for (i = 0; i < count; i++) msm_rpm_write(page, reg + i, 0); } static inline uint32_t msm_rpm_map_id_to_sel(uint32_t id) { return (id >= MSM_RPM_ID_LAST) ? msm_rpm_data.sel_last + 1 : msm_rpm_data.target_id[id].sel; } /* * Note: the function does not clear the masks before filling them. * * Return value: * 0: success * -EINVAL: invalid id in array */ static int msm_rpm_fill_sel_masks( uint32_t *sel_masks, struct msm_rpm_iv_pair *req, int count) { uint32_t sel; int i; for (i = 0; i < count; i++) { sel = msm_rpm_map_id_to_sel(req[i].id); if (sel > msm_rpm_data.sel_last) { pr_err("%s(): RPM ID %d not defined for target\n", __func__, req[i].id); return -EINVAL; } sel_masks[msm_rpm_get_sel_mask_reg(sel)] |= msm_rpm_get_sel_mask(sel); } return 0; } static inline void msm_rpm_send_req_interrupt(void) { __raw_writel(msm_rpm_data.ipc_rpm_val, msm_rpm_data.ipc_rpm_reg); } /* * Note: assumes caller has acquired . * * Return value: * 0: request acknowledgement * 1: notification * 2: spurious interrupt */ static int msm_rpm_process_ack_interrupt(void) { uint32_t ctx_mask_ack; uint32_t sel_masks_ack[SEL_MASK_SIZE] = {0}; ctx_mask_ack = msm_rpm_read(MSM_RPM_PAGE_CTRL, target_ctrl(MSM_RPM_CTRL_ACK_CTX_0)); msm_rpm_read_contiguous(MSM_RPM_PAGE_CTRL, target_ctrl(MSM_RPM_CTRL_ACK_SEL_0), sel_masks_ack, msm_rpm_sel_mask_size); if (ctx_mask_ack & msm_rpm_get_ctx_mask(MSM_RPM_CTX_NOTIFICATION)) { struct msm_rpm_notification *n; int i; list_for_each_entry(n, &msm_rpm_notifications, list) for (i = 0; i < msm_rpm_sel_mask_size; i++) if (sel_masks_ack[i] & n->sel_masks[i]) { up(&n->sem); break; } msm_rpm_write_contiguous_zeros(MSM_RPM_PAGE_CTRL, target_ctrl(MSM_RPM_CTRL_ACK_SEL_0), msm_rpm_sel_mask_size); msm_rpm_write(MSM_RPM_PAGE_CTRL, target_ctrl(MSM_RPM_CTRL_ACK_CTX_0), 0); /* Ensure the write is complete before return */ mb(); return 1; } if (msm_rpm_request) { int i; *(msm_rpm_request->ctx_mask_ack) = ctx_mask_ack; memcpy(msm_rpm_request->sel_masks_ack, sel_masks_ack, sizeof(sel_masks_ack)); for (i = 0; i < msm_rpm_request->count; i++) msm_rpm_request->req[i].value = msm_rpm_read(MSM_RPM_PAGE_ACK, target_enum(msm_rpm_request->req[i].id)); msm_rpm_write_contiguous_zeros(MSM_RPM_PAGE_CTRL, target_ctrl(MSM_RPM_CTRL_ACK_SEL_0), msm_rpm_sel_mask_size); msm_rpm_write(MSM_RPM_PAGE_CTRL, target_ctrl(MSM_RPM_CTRL_ACK_CTX_0), 0); /* Ensure the write is complete before return */ mb(); if (msm_rpm_request->done) complete_all(msm_rpm_request->done); msm_rpm_request = NULL; return 0; } return 2; } static void msm_rpm_err_fatal(void) { /* Tell RPM that we're handling the interrupt */ __raw_writel(0x1, msm_rpm_data.ipc_rpm_reg); panic("RPM error fataled"); } static irqreturn_t msm_rpm_err_interrupt(int irq, void *dev_id) { msm_rpm_err_fatal(); return IRQ_HANDLED; } static irqreturn_t msm_rpm_ack_interrupt(int irq, void *dev_id) { unsigned long flags; int rc; if (dev_id != &msm_rpm_ack_interrupt) return IRQ_NONE; spin_lock_irqsave(&msm_rpm_irq_lock, flags); rc = msm_rpm_process_ack_interrupt(); spin_unlock_irqrestore(&msm_rpm_irq_lock, flags); return IRQ_HANDLED; } /* * Note: assumes caller has acquired . */ static void msm_rpm_busy_wait_for_request_completion( bool allow_async_completion) { int rc; do { while (!gic_is_irq_pending(msm_rpm_data.irq_ack) && msm_rpm_request) { if (allow_async_completion) spin_unlock(&msm_rpm_irq_lock); if (gic_is_irq_pending(msm_rpm_data.irq_err)) msm_rpm_err_fatal(); gic_clear_irq_pending(msm_rpm_data.irq_err); udelay(1); if (allow_async_completion) spin_lock(&msm_rpm_irq_lock); } if (!msm_rpm_request) break; rc = msm_rpm_process_ack_interrupt(); gic_clear_irq_pending(msm_rpm_data.irq_ack); } while (rc); } /* Upon return, the array will contain values from the ack page. * * Note: assumes caller has acquired . * * Return value: * 0: success * -ENOSPC: request rejected */ static int msm_rpm_set_exclusive(int ctx, uint32_t *sel_masks, struct msm_rpm_iv_pair *req, int count) { DECLARE_COMPLETION_ONSTACK(ack); unsigned long flags; uint32_t ctx_mask = msm_rpm_get_ctx_mask(ctx); uint32_t ctx_mask_ack = 0; uint32_t sel_masks_ack[SEL_MASK_SIZE] = {0}; int i; msm_rpm_request_irq_mode.req = req; msm_rpm_request_irq_mode.count = count; msm_rpm_request_irq_mode.ctx_mask_ack = &ctx_mask_ack; msm_rpm_request_irq_mode.sel_masks_ack = sel_masks_ack; msm_rpm_request_irq_mode.done = &ack; spin_lock_irqsave(&msm_rpm_lock, flags); spin_lock(&msm_rpm_irq_lock); BUG_ON(msm_rpm_request); msm_rpm_request = &msm_rpm_request_irq_mode; for (i = 0; i < count; i++) { BUG_ON(target_enum(req[i].id) >= MSM_RPM_ID_LAST); msm_rpm_write(MSM_RPM_PAGE_REQ, target_enum(req[i].id), req[i].value); } msm_rpm_write_contiguous(MSM_RPM_PAGE_CTRL, target_ctrl(MSM_RPM_CTRL_REQ_SEL_0), sel_masks, msm_rpm_sel_mask_size); msm_rpm_write(MSM_RPM_PAGE_CTRL, target_ctrl(MSM_RPM_CTRL_REQ_CTX_0), ctx_mask); /* Ensure RPM data is written before sending the interrupt */ mb(); msm_rpm_send_req_interrupt(); spin_unlock(&msm_rpm_irq_lock); spin_unlock_irqrestore(&msm_rpm_lock, flags); wait_for_completion(&ack); BUG_ON((ctx_mask_ack & ~(msm_rpm_get_ctx_mask(MSM_RPM_CTX_REJECTED))) != ctx_mask); BUG_ON(memcmp(sel_masks, sel_masks_ack, sizeof(sel_masks_ack))); return (ctx_mask_ack & msm_rpm_get_ctx_mask(MSM_RPM_CTX_REJECTED)) ? -ENOSPC : 0; } /* Upon return, the array will contain values from the ack page. * * Note: assumes caller has acquired . * * Return value: * 0: success * -ENOSPC: request rejected */ static int msm_rpm_set_exclusive_noirq(int ctx, uint32_t *sel_masks, struct msm_rpm_iv_pair *req, int count) { unsigned int irq = msm_rpm_data.irq_ack; unsigned long flags; uint32_t ctx_mask = msm_rpm_get_ctx_mask(ctx); uint32_t ctx_mask_ack = 0; uint32_t sel_masks_ack[SEL_MASK_SIZE] = {0}; struct irq_chip *irq_chip, *err_chip; int i; msm_rpm_request_poll_mode.req = req; msm_rpm_request_poll_mode.count = count; msm_rpm_request_poll_mode.ctx_mask_ack = &ctx_mask_ack; msm_rpm_request_poll_mode.sel_masks_ack = sel_masks_ack; msm_rpm_request_poll_mode.done = NULL; spin_lock_irqsave(&msm_rpm_irq_lock, flags); irq_chip = irq_get_chip(irq); if (!irq_chip) { spin_unlock_irqrestore(&msm_rpm_irq_lock, flags); return -ENOSPC; } irq_chip->irq_mask(irq_get_irq_data(irq)); err_chip = irq_get_chip(msm_rpm_data.irq_err); if (!err_chip) { irq_chip->irq_unmask(irq_get_irq_data(irq)); spin_unlock_irqrestore(&msm_rpm_irq_lock, flags); return -ENOSPC; } err_chip->irq_mask(irq_get_irq_data(msm_rpm_data.irq_err)); if (msm_rpm_request) { msm_rpm_busy_wait_for_request_completion(true); BUG_ON(msm_rpm_request); } msm_rpm_request = &msm_rpm_request_poll_mode; for (i = 0; i < count; i++) { BUG_ON(target_enum(req[i].id) >= MSM_RPM_ID_LAST); msm_rpm_write(MSM_RPM_PAGE_REQ, target_enum(req[i].id), req[i].value); } msm_rpm_write_contiguous(MSM_RPM_PAGE_CTRL, target_ctrl(MSM_RPM_CTRL_REQ_SEL_0), sel_masks, msm_rpm_sel_mask_size); msm_rpm_write(MSM_RPM_PAGE_CTRL, target_ctrl(MSM_RPM_CTRL_REQ_CTX_0), ctx_mask); /* Ensure RPM data is written before sending the interrupt */ mb(); msm_rpm_send_req_interrupt(); msm_rpm_busy_wait_for_request_completion(false); BUG_ON(msm_rpm_request); err_chip->irq_unmask(irq_get_irq_data(msm_rpm_data.irq_err)); irq_chip->irq_unmask(irq_get_irq_data(irq)); spin_unlock_irqrestore(&msm_rpm_irq_lock, flags); BUG_ON((ctx_mask_ack & ~(msm_rpm_get_ctx_mask(MSM_RPM_CTX_REJECTED))) != ctx_mask); BUG_ON(memcmp(sel_masks, sel_masks_ack, sizeof(sel_masks_ack))); return (ctx_mask_ack & msm_rpm_get_ctx_mask(MSM_RPM_CTX_REJECTED)) ? -ENOSPC : 0; } /* Upon return, the array will contain values from the ack page. * * Return value: * 0: success * -EINVAL: invalid or invalid id in array * -ENOSPC: request rejected * -ENODEV: RPM driver not initialized */ static int msm_rpm_set_common( int ctx, struct msm_rpm_iv_pair *req, int count, bool noirq) { uint32_t sel_masks[SEL_MASK_SIZE] = {}; int rc; if (ctx >= MSM_RPM_CTX_SET_COUNT) { rc = -EINVAL; goto set_common_exit; } rc = msm_rpm_fill_sel_masks(sel_masks, req, count); if (rc) goto set_common_exit; if (noirq) { unsigned long flags; spin_lock_irqsave(&msm_rpm_lock, flags); rc = msm_rpm_set_exclusive_noirq(ctx, sel_masks, req, count); spin_unlock_irqrestore(&msm_rpm_lock, flags); } else { mutex_lock(&msm_rpm_mutex); rc = msm_rpm_set_exclusive(ctx, sel_masks, req, count); mutex_unlock(&msm_rpm_mutex); } set_common_exit: return rc; } /* * Return value: * 0: success * -EINVAL: invalid or invalid id in array * -ENODEV: RPM driver not initialized. */ static int msm_rpm_clear_common( int ctx, struct msm_rpm_iv_pair *req, int count, bool noirq) { uint32_t sel_masks[SEL_MASK_SIZE] = {}; struct msm_rpm_iv_pair r[SEL_MASK_SIZE]; int rc; int i; if (ctx >= MSM_RPM_CTX_SET_COUNT) { rc = -EINVAL; goto clear_common_exit; } rc = msm_rpm_fill_sel_masks(sel_masks, req, count); if (rc) goto clear_common_exit; for (i = 0; i < ARRAY_SIZE(r); i++) { r[i].id = MSM_RPM_ID_INVALIDATE_0 + i; r[i].value = sel_masks[i]; } memset(sel_masks, 0, sizeof(sel_masks)); sel_masks[msm_rpm_get_sel_mask_reg(msm_rpm_data.sel_invalidate)] |= msm_rpm_get_sel_mask(msm_rpm_data.sel_invalidate); if (noirq) { unsigned long flags; spin_lock_irqsave(&msm_rpm_lock, flags); rc = msm_rpm_set_exclusive_noirq(ctx, sel_masks, r, ARRAY_SIZE(r)); spin_unlock_irqrestore(&msm_rpm_lock, flags); BUG_ON(rc); } else { mutex_lock(&msm_rpm_mutex); rc = msm_rpm_set_exclusive(ctx, sel_masks, r, ARRAY_SIZE(r)); mutex_unlock(&msm_rpm_mutex); BUG_ON(rc); } clear_common_exit: return rc; } /* * Note: assumes caller has acquired . */ static void msm_rpm_update_notification(uint32_t ctx, struct msm_rpm_notif_config *curr_cfg, struct msm_rpm_notif_config *new_cfg) { unsigned int sel_notif = msm_rpm_data.sel_notification; if (memcmp(curr_cfg, new_cfg, sizeof(*new_cfg))) { uint32_t sel_masks[SEL_MASK_SIZE] = {}; int rc; sel_masks[msm_rpm_get_sel_mask_reg(sel_notif)] |= msm_rpm_get_sel_mask(sel_notif); rc = msm_rpm_set_exclusive(ctx, sel_masks, new_cfg->iv, ARRAY_SIZE(new_cfg->iv)); BUG_ON(rc); memcpy(curr_cfg, new_cfg, sizeof(*new_cfg)); } } /* * Note: assumes caller has acquired . */ static void msm_rpm_initialize_notification(void) { struct msm_rpm_notif_config cfg; unsigned int ctx; int i; for (ctx = MSM_RPM_CTX_SET_0; ctx <= MSM_RPM_CTX_SET_SLEEP; ctx++) { cfg = msm_rpm_notif_cfgs[ctx]; for (i = 0; i < msm_rpm_sel_mask_size; i++) { configured_iv(&cfg)[i].id = MSM_RPM_ID_NOTIFICATION_CONFIGURED_0 + i; configured_iv(&cfg)[i].value = ~0UL; registered_iv(&cfg)[i].id = MSM_RPM_ID_NOTIFICATION_REGISTERED_0 + i; registered_iv(&cfg)[i].value = 0; } msm_rpm_update_notification(ctx, &msm_rpm_notif_cfgs[ctx], &cfg); } } /****************************************************************************** * Public functions *****************************************************************************/ int msm_rpm_local_request_is_outstanding(void) { unsigned long flags; int outstanding = 0; if (!spin_trylock_irqsave(&msm_rpm_lock, flags)) goto local_request_is_outstanding_exit; if (!spin_trylock(&msm_rpm_irq_lock)) goto local_request_is_outstanding_unlock; outstanding = (msm_rpm_request != NULL); spin_unlock(&msm_rpm_irq_lock); local_request_is_outstanding_unlock: spin_unlock_irqrestore(&msm_rpm_lock, flags); local_request_is_outstanding_exit: return outstanding; } /* * Read the specified status registers and return their values. * * status: array of id-value pairs. Each specifies a status register, * i.e, one of MSM_RPM_STATUS_ID_xxxx. Upon return, each will * contain the value of the status register. * count: number of id-value pairs in the array * * Return value: * 0: success * -EBUSY: RPM is updating the status page; values across different registers * may not be consistent * -EINVAL: invalid id in array * -ENODEV: RPM driver not initialized */ int msm_rpm_get_status(struct msm_rpm_iv_pair *status, int count) { uint32_t seq_begin; uint32_t seq_end; int rc; int i; seq_begin = msm_rpm_read(MSM_RPM_PAGE_STATUS, target_status(MSM_RPM_STATUS_ID_SEQUENCE)); for (i = 0; i < count; i++) { int target_status_id; if (status[i].id >= MSM_RPM_STATUS_ID_LAST) { pr_err("%s(): Status ID beyond limits\n", __func__); rc = -EINVAL; goto get_status_exit; } target_status_id = target_status(status[i].id); if (target_status_id >= MSM_RPM_STATUS_ID_LAST) { pr_err("%s(): Status id %d not defined for target\n", __func__, target_status_id); rc = -EINVAL; goto get_status_exit; } status[i].value = msm_rpm_read(MSM_RPM_PAGE_STATUS, target_status_id); } seq_end = msm_rpm_read(MSM_RPM_PAGE_STATUS, target_status(MSM_RPM_STATUS_ID_SEQUENCE)); rc = (seq_begin != seq_end || (seq_begin & 0x01)) ? -EBUSY : 0; get_status_exit: return rc; } EXPORT_SYMBOL(msm_rpm_get_status); /* * Issue a resource request to RPM to set resource values. * * Note: the function may sleep and must be called in a task context. * * ctx: the request's context. * There two contexts that a RPM driver client can use: * MSM_RPM_CTX_SET_0 and MSM_RPM_CTX_SET_SLEEP. For resource values * that are intended to take effect when the CPU is active, * MSM_RPM_CTX_SET_0 should be used. For resource values that are * intended to take effect when the CPU is not active, * MSM_RPM_CTX_SET_SLEEP should be used. * req: array of id-value pairs. Each specifies a RPM resource, * i.e, one of MSM_RPM_ID_xxxx. Each specifies the requested * resource value. * count: number of id-value pairs in the array * * Return value: * 0: success * -EINVAL: invalid or invalid id in array * -ENOSPC: request rejected * -ENODEV: RPM driver not initialized */ int msm_rpm_set(int ctx, struct msm_rpm_iv_pair *req, int count) { return msm_rpm_set_common(ctx, req, count, false); } EXPORT_SYMBOL(msm_rpm_set); /* * Issue a resource request to RPM to set resource values. * * Note: the function is similar to msm_rpm_set() except that it must be * called with interrupts masked. If possible, use msm_rpm_set() * instead, to maximize CPU throughput. */ int msm_rpm_set_noirq(int ctx, struct msm_rpm_iv_pair *req, int count) { WARN(!irqs_disabled(), "msm_rpm_set_noirq can only be called " "safely when local irqs are disabled. Consider using " "msm_rpm_set or msm_rpm_set_nosleep instead."); return msm_rpm_set_common(ctx, req, count, true); } EXPORT_SYMBOL(msm_rpm_set_noirq); /* * Issue a resource request to RPM to clear resource values. Once the * values are cleared, the resources revert back to their default values * for this RPM master. * * Note: the function may sleep and must be called in a task context. * * ctx: the request's context. * req: array of id-value pairs. Each specifies a RPM resource, * i.e, one of MSM_RPM_ID_xxxx. 's are ignored. * count: number of id-value pairs in the array * * Return value: * 0: success * -EINVAL: invalid or invalid id in array */ int msm_rpm_clear(int ctx, struct msm_rpm_iv_pair *req, int count) { return msm_rpm_clear_common(ctx, req, count, false); } EXPORT_SYMBOL(msm_rpm_clear); /* * Issue a resource request to RPM to clear resource values. * * Note: the function is similar to msm_rpm_clear() except that it must be * called with interrupts masked. If possible, use msm_rpm_clear() * instead, to maximize CPU throughput. */ int msm_rpm_clear_noirq(int ctx, struct msm_rpm_iv_pair *req, int count) { WARN(!irqs_disabled(), "msm_rpm_clear_noirq can only be called " "safely when local irqs are disabled. Consider using " "msm_rpm_clear or msm_rpm_clear_nosleep instead."); return msm_rpm_clear_common(ctx, req, count, true); } EXPORT_SYMBOL(msm_rpm_clear_noirq); /* * Register for RPM notification. When the specified resources * change their status on RPM, RPM sends out notifications and the * driver will "up" the semaphore in struct msm_rpm_notification. * * Note: the function may sleep and must be called in a task context. * * Memory for must not be freed until the notification is * unregistered. Memory for can be freed after this * function returns. * * n: the notifcation object. Caller should initialize only the * semaphore field. When a notification arrives later, the * semaphore will be "up"ed. * req: array of id-value pairs. Each specifies a status register, * i.e, one of MSM_RPM_STATUS_ID_xxxx. 's are ignored. * count: number of id-value pairs in the array * * Return value: * 0: success * -EINVAL: invalid id in array * -ENODEV: RPM driver not initialized */ int msm_rpm_register_notification(struct msm_rpm_notification *n, struct msm_rpm_iv_pair *req, int count) { unsigned long flags; unsigned int ctx; struct msm_rpm_notif_config cfg; int rc; int i; INIT_LIST_HEAD(&n->list); rc = msm_rpm_fill_sel_masks(n->sel_masks, req, count); if (rc) goto register_notification_exit; mutex_lock(&msm_rpm_mutex); if (!msm_rpm_init_notif_done) { msm_rpm_initialize_notification(); msm_rpm_init_notif_done = true; } spin_lock_irqsave(&msm_rpm_irq_lock, flags); list_add(&n->list, &msm_rpm_notifications); spin_unlock_irqrestore(&msm_rpm_irq_lock, flags); ctx = MSM_RPM_CTX_SET_0; cfg = msm_rpm_notif_cfgs[ctx]; for (i = 0; i < msm_rpm_sel_mask_size; i++) registered_iv(&cfg)[i].value |= n->sel_masks[i]; msm_rpm_update_notification(ctx, &msm_rpm_notif_cfgs[ctx], &cfg); mutex_unlock(&msm_rpm_mutex); register_notification_exit: return rc; } EXPORT_SYMBOL(msm_rpm_register_notification); /* * Unregister a notification. * * Note: the function may sleep and must be called in a task context. * * n: the notifcation object that was registered previously. * * Return value: * 0: success * -ENODEV: RPM driver not initialized */ int msm_rpm_unregister_notification(struct msm_rpm_notification *n) { unsigned long flags; unsigned int ctx; struct msm_rpm_notif_config cfg; int rc = 0; int i; mutex_lock(&msm_rpm_mutex); ctx = MSM_RPM_CTX_SET_0; cfg = msm_rpm_notif_cfgs[ctx]; for (i = 0; i < msm_rpm_sel_mask_size; i++) registered_iv(&cfg)[i].value = 0; spin_lock_irqsave(&msm_rpm_irq_lock, flags); list_del(&n->list); list_for_each_entry(n, &msm_rpm_notifications, list) for (i = 0; i < msm_rpm_sel_mask_size; i++) registered_iv(&cfg)[i].value |= n->sel_masks[i]; spin_unlock_irqrestore(&msm_rpm_irq_lock, flags); msm_rpm_update_notification(ctx, &msm_rpm_notif_cfgs[ctx], &cfg); mutex_unlock(&msm_rpm_mutex); return rc; } EXPORT_SYMBOL(msm_rpm_unregister_notification); static uint32_t fw_major, fw_minor, fw_build; static ssize_t driver_version_show(struct kobject *kobj, struct kobj_attribute *attr, char *buf) { return snprintf(buf, PAGE_SIZE, "%u.%u.%u\n", msm_rpm_data.ver[0], msm_rpm_data.ver[1], msm_rpm_data.ver[2]); } static ssize_t fw_version_show(struct kobject *kobj, struct kobj_attribute *attr, char *buf) { return snprintf(buf, PAGE_SIZE, "%u.%u.%u\n", fw_major, fw_minor, fw_build); } static struct kobj_attribute driver_version_attr = __ATTR_RO(driver_version); static struct kobj_attribute fw_version_attr = __ATTR_RO(fw_version); static struct attribute *driver_attributes[] = { &driver_version_attr.attr, &fw_version_attr.attr, NULL }; static struct attribute_group driver_attr_group = { .attrs = driver_attributes, }; static int __devinit msm_rpm_probe(struct platform_device *pdev) { return sysfs_create_group(&pdev->dev.kobj, &driver_attr_group); } static int __devexit msm_rpm_remove(struct platform_device *pdev) { sysfs_remove_group(&pdev->dev.kobj, &driver_attr_group); return 0; } static struct platform_driver msm_rpm_platform_driver = { .probe = msm_rpm_probe, .remove = __devexit_p(msm_rpm_remove), .driver = { .name = "msm_rpm", .owner = THIS_MODULE, }, }; static void __init msm_rpm_populate_map(struct msm_rpm_platform_data *data) { int i, j; struct msm_rpm_map_data *src = NULL; struct msm_rpm_map_data *dst = NULL; for (i = 0; i < MSM_RPM_ID_LAST;) { src = &data->target_id[i]; dst = &msm_rpm_data.target_id[i]; dst->id = MSM_RPM_ID_LAST; dst->sel = msm_rpm_data.sel_last + 1; /* * copy the target specific id of the current and also of * all the #count id's that follow the current. * [MSM_RPM_ID_PM8921_S1_0] = { MSM_RPM_8960_ID_PM8921_S1_0, * MSM_RPM_8960_SEL_PM8921_S1, * 2}, * [MSM_RPM_ID_PM8921_S1_1] = { 0, 0, 0 }, * should translate to * [MSM_RPM_ID_PM8921_S1_0] = { MSM_RPM_8960_ID_PM8921_S1_0, * MSM_RPM_8960_SEL_PM8921, * 2 }, * [MSM_RPM_ID_PM8921_S1_1] = { MSM_RPM_8960_ID_PM8921_S1_0 + 1, * MSM_RPM_8960_SEL_PM8921, * 0 }, */ for (j = 0; j < src->count; j++) { dst = &msm_rpm_data.target_id[i + j]; dst->id = src->id + j; dst->sel = src->sel; } i += (src->count) ? src->count : 1; } for (i = 0; i < MSM_RPM_STATUS_ID_LAST; i++) { if (data->target_status[i] & MSM_RPM_STATUS_ID_VALID) msm_rpm_data.target_status[i] &= ~MSM_RPM_STATUS_ID_VALID; else msm_rpm_data.target_status[i] = MSM_RPM_STATUS_ID_LAST; } } static irqreturn_t msm_pm_rpm_wakeup_interrupt(int irq, void *dev_id) { if (dev_id != &msm_pm_rpm_wakeup_interrupt) return IRQ_NONE; return IRQ_HANDLED; } int __init msm_rpm_init(struct msm_rpm_platform_data *data) { int rc; memcpy(&msm_rpm_data, data, sizeof(struct msm_rpm_platform_data)); msm_rpm_sel_mask_size = msm_rpm_data.sel_last / 32 + 1; BUG_ON(SEL_MASK_SIZE < msm_rpm_sel_mask_size); fw_major = msm_rpm_read(MSM_RPM_PAGE_STATUS, target_status(MSM_RPM_STATUS_ID_VERSION_MAJOR)); fw_minor = msm_rpm_read(MSM_RPM_PAGE_STATUS, target_status(MSM_RPM_STATUS_ID_VERSION_MINOR)); fw_build = msm_rpm_read(MSM_RPM_PAGE_STATUS, target_status(MSM_RPM_STATUS_ID_VERSION_BUILD)); pr_info("%s: RPM firmware %u.%u.%u\n", __func__, fw_major, fw_minor, fw_build); if (fw_major != msm_rpm_data.ver[0]) { pr_err("%s: RPM version %u.%u.%u incompatible with " "this driver version %u.%u.%u\n", __func__, fw_major, fw_minor, fw_build, msm_rpm_data.ver[0], msm_rpm_data.ver[1], msm_rpm_data.ver[2]); return -EFAULT; } msm_rpm_write(MSM_RPM_PAGE_CTRL, target_ctrl(MSM_RPM_CTRL_VERSION_MAJOR), msm_rpm_data.ver[0]); msm_rpm_write(MSM_RPM_PAGE_CTRL, target_ctrl(MSM_RPM_CTRL_VERSION_MINOR), msm_rpm_data.ver[1]); msm_rpm_write(MSM_RPM_PAGE_CTRL, target_ctrl(MSM_RPM_CTRL_VERSION_BUILD), msm_rpm_data.ver[2]); rc = request_irq(data->irq_ack, msm_rpm_ack_interrupt, IRQF_TRIGGER_RISING | IRQF_NO_SUSPEND, "rpm_drv", msm_rpm_ack_interrupt); if (rc) { pr_err("%s: failed to request irq %d: %d\n", __func__, data->irq_ack, rc); return rc; } rc = irq_set_irq_wake(data->irq_ack, 1); if (rc) { pr_err("%s: failed to set wakeup irq %u: %d\n", __func__, data->irq_ack, rc); return rc; } rc = request_irq(data->irq_err, msm_rpm_err_interrupt, IRQF_TRIGGER_RISING, "rpm_err", NULL); if (rc) { pr_err("%s: failed to request error interrupt: %d\n", __func__, rc); return rc; } rc = request_irq(data->irq_wakeup, msm_pm_rpm_wakeup_interrupt, IRQF_TRIGGER_RISING, "pm_drv", msm_pm_rpm_wakeup_interrupt); if (rc) { pr_err("%s: failed to request irq %u: %d\n", __func__, data->irq_wakeup, rc); return rc; } rc = irq_set_irq_wake(data->irq_wakeup, 1); if (rc) { pr_err("%s: failed to set wakeup irq %u: %d\n", __func__, data->irq_wakeup, rc); return rc; } msm_rpm_populate_map(data); return platform_driver_register(&msm_rpm_platform_driver); }