M7350/kernel/drivers/iommu/msm_iommu_perfmon.c
2024-09-09 08:52:07 +00:00

840 lines
21 KiB
C

/* Copyright (c) 2012-2013, The Linux Foundation. All rights reserved.
*
* This program is free software; you can redistribute it and/or modify
* it under the terms of the GNU General Public License version 2 and
* only version 2 as published by the Free Software Foundation.
*
* This program is distributed in the hope that it will be useful,
* but WITHOUT ANY WARRANTY; without even the implied warranty of
* MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
* GNU General Public License for more details.
*/
#define pr_fmt(fmt) KBUILD_MODNAME ": " fmt
#include <linux/kernel.h>
#include <linux/module.h>
#include <linux/string.h>
#include <linux/iommu.h>
#include <linux/slab.h>
#include <linux/device.h>
#include <linux/interrupt.h>
#include <linux/bitops.h>
#include <linux/debugfs.h>
#include <mach/iommu.h>
#include <mach/iommu_perfmon.h>
static LIST_HEAD(iommu_list);
static struct dentry *msm_iommu_root_debugfs_dir;
static const char *NO_EVENT_CLASS_NAME = "none";
static const unsigned int MAX_EVEN_CLASS_NAME_LEN = 36;
struct event_class {
unsigned int event_number;
const char *desc;
};
static struct event_class pmu_event_classes[] = {
{ 0x00, "cycle_count" },
{ 0x01, "cycle_count64" },
{ 0x08, "tlb_refill" },
{ 0x09, "tlb_refill_read" },
{ 0x0A, "tlb_refill_write" },
{ 0x10, "access" },
{ 0x11, "access_read" },
{ 0x12, "access_write" },
{ 0x80, "full_misses" },
{ 0x81, "partial_miss_1lbfb_hit" },
{ 0x82, "partial_miss_2lbfb_hit" },
{ 0x83, "full_hit" },
{ 0x90, "pred_req_full_miss" },
{ 0x91, "pred_req_partial_miss_1lbfb_hit" },
{ 0x92, "pred_req_partial_miss_2lbfb_hit" },
{ 0xb0, "tot_num_miss_axi_htw_read_req" },
{ 0xb1, "tot_num_pred_axi_htw_read_req" },
};
static unsigned int iommu_pm_create_sup_cls_str(char **buf,
struct iommu_pmon *pmon)
{
unsigned long buf_size = ARRAY_SIZE(pmu_event_classes) *
MAX_EVEN_CLASS_NAME_LEN;
unsigned int pos = 0;
unsigned int nevent_cls = pmon->nevent_cls_supported;
*buf = kzalloc(buf_size, GFP_KERNEL);
if (*buf) {
unsigned int j;
int i;
struct event_class *ptr;
size_t array_len = ARRAY_SIZE(pmu_event_classes);
ptr = pmu_event_classes;
for (j = 0; j < nevent_cls; ++j) {
for (i = 0; i < array_len; ++i) {
if (ptr[i].event_number !=
pmon->event_cls_supported[j])
continue;
if (pos < buf_size) {
pos += snprintf(&(*buf)[pos],
buf_size-pos,
"[%u] %s\n",
ptr[i].event_number,
ptr[i].desc);
}
break;
}
}
}
return pos;
}
static int iommu_pm_event_class_supported(struct iommu_pmon *pmon,
int event_class)
{
unsigned int nevent_cls = pmon->nevent_cls_supported;
unsigned int i;
for (i = 0; i < nevent_cls; ++i) {
if (event_class == pmon->event_cls_supported[i])
return event_class;
}
return MSM_IOMMU_PMU_NO_EVENT_CLASS;
}
static const char *iommu_pm_find_event_class_name(int event_class)
{
size_t array_len;
struct event_class *ptr;
int i;
const char *event_class_name = NO_EVENT_CLASS_NAME;
if (event_class < 0)
goto out;
array_len = ARRAY_SIZE(pmu_event_classes);
ptr = pmu_event_classes;
for (i = 0; i < array_len; ++i) {
if (ptr[i].event_number == event_class) {
event_class_name = ptr[i].desc;
break;
}
}
out:
return event_class_name;
}
static int iommu_pm_find_event_class(struct iommu_pmon *pmon,
const char *event_class_name)
{
size_t array_len;
struct event_class *ptr;
int i;
int event_class = MSM_IOMMU_PMU_NO_EVENT_CLASS;
if (strcmp(event_class_name, NO_EVENT_CLASS_NAME) == 0)
goto out;
array_len = ARRAY_SIZE(pmu_event_classes);
ptr = pmu_event_classes;
for (i = 0; i < array_len; ++i) {
if (strcmp(ptr[i].desc, event_class_name) == 0) {
event_class = ptr[i].event_number;
goto out;
}
}
out:
event_class = iommu_pm_event_class_supported(pmon, event_class);
return event_class;
}
static inline void iommu_pm_add_to_iommu_list(struct iommu_pmon *iommu_pmon)
{
list_add(&iommu_pmon->iommu_list, &iommu_list);
}
static inline void iommu_pm_del_from_iommu_list(struct iommu_pmon *iommu_pmon)
{
list_del(&iommu_pmon->iommu_list);
}
static struct iommu_pmon *iommu_pm_get_pm_by_dev(struct device *dev)
{
struct iommu_pmon *pmon;
struct iommu_info *info;
struct list_head *ent;
list_for_each(ent, &iommu_list) {
pmon = list_entry(ent, struct iommu_pmon, iommu_list);
info = &pmon->iommu;
if (dev == info->iommu_dev)
return pmon;
}
return NULL;
}
static void iommu_pm_set_event_type(struct iommu_pmon *pmon,
struct iommu_pmon_counter *counter)
{
int event_class;
unsigned int count_no;
struct iommu_info *iommu = &pmon->iommu;
event_class = counter->current_event_class;
count_no = counter->absolute_counter_no;
if (event_class == MSM_IOMMU_PMU_NO_EVENT_CLASS) {
if (iommu->hw_ops->is_hw_access_OK(pmon)) {
iommu->ops->iommu_lock_acquire();
iommu->hw_ops->counter_disable(iommu, counter);
iommu->hw_ops->ovfl_int_disable(iommu, counter);
iommu->hw_ops->set_event_class(pmon, count_no, 0);
iommu->ops->iommu_lock_release();
}
counter->overflow_count = 0;
counter->value = 0;
} else {
counter->overflow_count = 0;
counter->value = 0;
if (iommu->hw_ops->is_hw_access_OK(pmon)) {
iommu->ops->iommu_lock_acquire();
iommu->hw_ops->set_event_class(pmon, count_no,
event_class);
iommu->hw_ops->ovfl_int_enable(iommu, counter);
iommu->hw_ops->counter_enable(iommu, counter);
iommu->ops->iommu_lock_release();
}
}
}
static void iommu_pm_reset_counts(struct iommu_pmon *pmon)
{
unsigned int i;
unsigned int j;
for (i = 0; i < pmon->num_groups; ++i) {
struct iommu_pmon_cnt_group *cnt_grp = &pmon->cnt_grp[i];
for (j = 0; j < cnt_grp->num_counters; ++j) {
cnt_grp->counters[j].value = 0;
cnt_grp->counters[j].overflow_count = 0;
}
}
}
static void iommu_pm_set_all_counters(struct iommu_pmon *pmon)
{
unsigned int i;
unsigned int j;
for (i = 0; i < pmon->num_groups; ++i) {
struct iommu_pmon_cnt_group *cnt_grp = &pmon->cnt_grp[i];
for (j = 0; j < cnt_grp->num_counters; ++j)
iommu_pm_set_event_type(pmon, &cnt_grp->counters[j]);
}
}
static void iommu_pm_read_all_counters(struct iommu_pmon *pmon)
{
unsigned int i;
unsigned int j;
struct iommu_info *iommu = &pmon->iommu;
for (i = 0; i < pmon->num_groups; ++i) {
struct iommu_pmon_cnt_group *cnt_grp = &pmon->cnt_grp[i];
for (j = 0; j < cnt_grp->num_counters; ++j) {
struct iommu_pmon_counter *counter;
counter = &cnt_grp->counters[j];
counter->value = iommu->hw_ops->read_counter(counter);
}
}
}
static void iommu_pm_on(struct iommu_pmon *pmon)
{
unsigned int i;
struct iommu_info *iommu = &pmon->iommu;
struct msm_iommu_drvdata *iommu_drvdata =
dev_get_drvdata(iommu->iommu_dev);
iommu->ops->iommu_power_on(iommu_drvdata);
iommu->ops->iommu_bus_vote(iommu_drvdata, 1);
iommu->ops->iommu_clk_on(iommu_drvdata);
/* Reset counters in HW */
iommu->ops->iommu_lock_acquire();
iommu->hw_ops->reset_counters(&pmon->iommu);
iommu->ops->iommu_lock_release();
/* Reset SW counters */
iommu_pm_reset_counts(pmon);
pmon->enabled = 1;
iommu_pm_set_all_counters(pmon);
iommu->ops->iommu_lock_acquire();
/* enable all counter group */
for (i = 0; i < pmon->num_groups; ++i)
iommu->hw_ops->grp_enable(iommu, i);
/* enable global counters */
iommu->hw_ops->enable_pm(iommu);
iommu->ops->iommu_lock_release();
pr_info("%s: TLB performance monitoring turned ON\n",
pmon->iommu.iommu_name);
}
static void iommu_pm_off(struct iommu_pmon *pmon)
{
unsigned int i;
struct iommu_info *iommu = &pmon->iommu;
struct msm_iommu_drvdata *iommu_drvdata =
dev_get_drvdata(iommu->iommu_dev);
pmon->enabled = 0;
iommu->ops->iommu_lock_acquire();
/* disable global counters */
iommu->hw_ops->disable_pm(iommu);
/* Check if we overflowed just before turning off pmon */
iommu->hw_ops->check_for_overflow(pmon);
/* disable all counter group */
for (i = 0; i < pmon->num_groups; ++i)
iommu->hw_ops->grp_disable(iommu, i);
/* Update cached copy of counters before turning off power */
iommu_pm_read_all_counters(pmon);
iommu->ops->iommu_lock_release();
iommu->ops->iommu_clk_off(iommu_drvdata);
iommu->ops->iommu_bus_vote(iommu_drvdata, 0);
iommu->ops->iommu_power_off(iommu_drvdata);
pr_info("%s: TLB performance monitoring turned OFF\n",
pmon->iommu.iommu_name);
}
static int iommu_pm_debug_open(struct inode *inode, struct file *file)
{
file->private_data = inode->i_private;
return 0;
}
static ssize_t iommu_pm_count_value_read(struct file *fp,
char __user *user_buff,
size_t count, loff_t *pos)
{
size_t rd_cnt;
unsigned long long full_count;
struct iommu_pmon_counter *counter = fp->private_data;
struct iommu_pmon *pmon = counter->cnt_group->pmon;
struct iommu_info *iommu = &pmon->iommu;
char buf[50];
size_t len;
mutex_lock(&pmon->lock);
if (iommu->hw_ops->is_hw_access_OK(pmon)) {
iommu->ops->iommu_lock_acquire();
counter->value = iommu->hw_ops->read_counter(counter);
iommu->ops->iommu_lock_release();
}
full_count = (unsigned long long) counter->value +
((unsigned long long)counter->overflow_count *
0x100000000ULL);
len = snprintf(buf, 50, "%llu\n", full_count);
rd_cnt = simple_read_from_buffer(user_buff, count, pos, buf, len);
mutex_unlock(&pmon->lock);
return rd_cnt;
}
static const struct file_operations cnt_value_file_ops = {
.open = iommu_pm_debug_open,
.read = iommu_pm_count_value_read,
};
static ssize_t iommu_pm_event_class_read(struct file *fp,
char __user *user_buff,
size_t count, loff_t *pos)
{
size_t rd_cnt;
struct iommu_pmon_counter *counter = fp->private_data;
struct iommu_pmon *pmon = counter->cnt_group->pmon;
char buf[50];
const char *event_class_name;
size_t len;
mutex_lock(&pmon->lock);
event_class_name = iommu_pm_find_event_class_name(
counter->current_event_class);
len = snprintf(buf, 50, "%s\n", event_class_name);
rd_cnt = simple_read_from_buffer(user_buff, count, pos, buf, len);
mutex_unlock(&pmon->lock);
return rd_cnt;
}
static ssize_t iommu_pm_event_class_write(struct file *fp,
const char __user *user_buff,
size_t count, loff_t *pos)
{
size_t wr_cnt;
char buf[50];
size_t buf_size = sizeof(buf);
struct iommu_pmon_counter *counter = fp->private_data;
struct iommu_pmon *pmon = counter->cnt_group->pmon;
int current_event_class;
if ((count + *pos) >= buf_size)
return -EINVAL;
mutex_lock(&pmon->lock);
current_event_class = counter->current_event_class;
wr_cnt = simple_write_to_buffer(buf, buf_size, pos, user_buff, count);
if (wr_cnt >= 1) {
int rv;
long value;
buf[wr_cnt-1] = '\0';
rv = kstrtol(buf, 10, &value);
if (!rv) {
counter->current_event_class =
iommu_pm_find_event_class(pmon,
iommu_pm_find_event_class_name(value));
} else {
counter->current_event_class =
iommu_pm_find_event_class(pmon, buf);
} }
if (current_event_class != counter->current_event_class)
iommu_pm_set_event_type(pmon, counter);
mutex_unlock(&pmon->lock);
return wr_cnt;
}
static const struct file_operations event_class_file_ops = {
.open = iommu_pm_debug_open,
.read = iommu_pm_event_class_read,
.write = iommu_pm_event_class_write,
};
static ssize_t iommu_reset_counters_write(struct file *fp,
const char __user *user_buff,
size_t count, loff_t *pos)
{
size_t wr_cnt;
char buf[10];
size_t buf_size = sizeof(buf);
struct iommu_pmon *pmon = fp->private_data;
struct iommu_info *iommu = &pmon->iommu;
if ((count + *pos) >= buf_size)
return -EINVAL;
mutex_lock(&pmon->lock);
wr_cnt = simple_write_to_buffer(buf, buf_size, pos, user_buff, count);
if (wr_cnt >= 1) {
unsigned long cmd = 0;
int rv;
buf[wr_cnt-1] = '\0';
rv = kstrtoul(buf, 10, &cmd);
if (!rv && (cmd == 1)) {
if (iommu->hw_ops->is_hw_access_OK(pmon)) {
iommu->ops->iommu_lock_acquire();
iommu->hw_ops->reset_counters(&pmon->iommu);
iommu->ops->iommu_lock_release();
}
iommu_pm_reset_counts(pmon);
pr_info("TLB performance counters reset\n");
} else {
pr_err("Unknown performance monitor command: %lu\n",
cmd);
}
}
mutex_unlock(&pmon->lock);
return wr_cnt;
}
static const struct file_operations reset_file_ops = {
.open = iommu_pm_debug_open,
.write = iommu_reset_counters_write,
};
static ssize_t iommu_pm_enable_counters_read(struct file *fp,
char __user *user_buff,
size_t count, loff_t *pos)
{
size_t rd_cnt;
char buf[5];
size_t len;
struct iommu_pmon *pmon = fp->private_data;
mutex_lock(&pmon->lock);
len = snprintf(buf, 5, "%u\n", pmon->enabled);
rd_cnt = simple_read_from_buffer(user_buff, count, pos, buf, len);
mutex_unlock(&pmon->lock);
return rd_cnt;
}
static ssize_t iommu_pm_enable_counters_write(struct file *fp,
const char __user *user_buff,
size_t count, loff_t *pos)
{
size_t wr_cnt;
char buf[10];
size_t buf_size = sizeof(buf);
struct iommu_pmon *pmon = fp->private_data;
if ((count + *pos) >= buf_size)
return -EINVAL;
mutex_lock(&pmon->lock);
wr_cnt = simple_write_to_buffer(buf, buf_size, pos, user_buff, count);
if (wr_cnt >= 1) {
unsigned long cmd;
int rv;
buf[wr_cnt-1] = '\0';
rv = kstrtoul(buf, 10, &cmd);
if (!rv && (cmd < 2)) {
if (pmon->enabled == 1 && cmd == 0) {
if (pmon->iommu.always_on ||
pmon->iommu_attach_count > 0)
iommu_pm_off(pmon);
} else if (pmon->enabled == 0 && cmd == 1) {
/* We can only turn on perf. monitoring if
* iommu is attached (if not always on).
* Delay turning on perf. monitoring until
* we are attached.
*/
if (pmon->iommu.always_on ||
pmon->iommu_attach_count > 0)
iommu_pm_on(pmon);
else
pmon->enabled = 1;
}
} else {
pr_err("Unknown performance monitor command: %lu\n",
cmd);
}
}
mutex_unlock(&pmon->lock);
return wr_cnt;
}
static const struct file_operations event_enable_file_ops = {
.open = iommu_pm_debug_open,
.read = iommu_pm_enable_counters_read,
.write = iommu_pm_enable_counters_write,
};
static ssize_t iommu_pm_avail_event_cls_read(struct file *fp,
char __user *user_buff,
size_t count, loff_t *pos)
{
size_t rd_cnt = 0;
struct iommu_pmon *pmon = fp->private_data;
char *buf;
size_t len;
mutex_lock(&pmon->lock);
len = iommu_pm_create_sup_cls_str(&buf, pmon);
if (buf) {
rd_cnt = simple_read_from_buffer(user_buff, count, pos,
buf, len);
kfree(buf);
}
mutex_unlock(&pmon->lock);
return rd_cnt;
}
static const struct file_operations available_event_cls_file_ops = {
.open = iommu_pm_debug_open,
.read = iommu_pm_avail_event_cls_read,
};
static int iommu_pm_create_grp_debugfs_counters_hierarchy(
struct iommu_pmon_cnt_group *cnt_grp,
unsigned int *abs_counter_no)
{
int ret = 0;
int j;
char name[20];
for (j = 0; j < cnt_grp->num_counters; ++j) {
struct dentry *grp_dir = cnt_grp->group_dir;
struct dentry *counter_dir;
cnt_grp->counters[j].cnt_group = cnt_grp;
cnt_grp->counters[j].counter_no = j;
cnt_grp->counters[j].absolute_counter_no = *abs_counter_no;
(*abs_counter_no)++;
cnt_grp->counters[j].value = 0;
cnt_grp->counters[j].overflow_count = 0;
cnt_grp->counters[j].current_event_class =
MSM_IOMMU_PMU_NO_EVENT_CLASS;
snprintf(name, 20, "counter%u", j);
counter_dir = debugfs_create_dir(name, grp_dir);
if (IS_ERR_OR_NULL(counter_dir)) {
pr_err("unable to create counter debugfs dir %s\n",
name);
ret = -ENOMEM;
goto out;
}
cnt_grp->counters[j].counter_dir = counter_dir;
if (!debugfs_create_file("value", 0644, counter_dir,
&cnt_grp->counters[j],
&cnt_value_file_ops)) {
ret = -EIO;
goto out;
}
if (!debugfs_create_file("current_event_class", 0644,
counter_dir, &cnt_grp->counters[j],
&event_class_file_ops)) {
ret = -EIO;
goto out;
}
}
out:
return ret;
}
static int iommu_pm_create_group_debugfs_hierarchy(struct iommu_info *iommu,
struct iommu_pmon *pmon_entry)
{
int i;
int ret = 0;
char name[20];
unsigned int abs_counter_no = 0;
for (i = 0; i < pmon_entry->num_groups; ++i) {
pmon_entry->cnt_grp[i].pmon = pmon_entry;
pmon_entry->cnt_grp[i].grp_no = i;
pmon_entry->cnt_grp[i].num_counters = pmon_entry->num_counters;
pmon_entry->cnt_grp[i].counters =
kzalloc(sizeof(*pmon_entry->cnt_grp[i].counters)
* pmon_entry->cnt_grp[i].num_counters, GFP_KERNEL);
if (!pmon_entry->cnt_grp[i].counters) {
pr_err("Unable to allocate memory for counters\n");
ret = -ENOMEM;
goto out;
}
snprintf(name, 20, "group%u", i);
pmon_entry->cnt_grp[i].group_dir = debugfs_create_dir(name,
pmon_entry->iommu_dir);
if (IS_ERR_OR_NULL(pmon_entry->cnt_grp[i].group_dir)) {
pr_err("unable to create group debugfs dir %s\n", name);
ret = -ENOMEM;
goto out;
}
ret = iommu_pm_create_grp_debugfs_counters_hierarchy(
&pmon_entry->cnt_grp[i],
&abs_counter_no);
if (ret)
goto out;
}
out:
return ret;
}
int msm_iommu_pm_iommu_register(struct iommu_pmon *pmon_entry)
{
int ret = 0;
struct iommu_info *iommu = &pmon_entry->iommu;
int i;
if (!iommu->ops || !iommu->iommu_name || !iommu->base
|| !iommu->iommu_dev) {
ret = -EINVAL;
goto out;
}
if (!msm_iommu_root_debugfs_dir) {
msm_iommu_root_debugfs_dir = debugfs_create_dir("iommu", NULL);
if (IS_ERR_OR_NULL(msm_iommu_root_debugfs_dir)) {
pr_err("Failed creating iommu debugfs dir \"iommu\"\n");
ret = -EIO;
goto out;
}
}
pmon_entry->cnt_grp = kzalloc(sizeof(*pmon_entry->cnt_grp)
* pmon_entry->num_groups, GFP_KERNEL);
if (!pmon_entry->cnt_grp) {
pr_err("Unable to allocate memory for counter groups\n");
ret = -ENOMEM;
goto file_err;
}
pmon_entry->iommu_dir = debugfs_create_dir(iommu->iommu_name,
msm_iommu_root_debugfs_dir);
if (IS_ERR_OR_NULL(pmon_entry->iommu_dir)) {
pr_err("unable to create iommu debugfs dir %s\n",
iommu->iommu_name);
ret = -ENOMEM;
goto free_mem;
}
if (!debugfs_create_file("reset_counters", 0644,
pmon_entry->iommu_dir, pmon_entry, &reset_file_ops)) {
ret = -EIO;
goto free_mem;
}
if (!debugfs_create_file("enable_counters", 0644,
pmon_entry->iommu_dir, pmon_entry, &event_enable_file_ops)) {
ret = -EIO;
goto free_mem;
}
if (!debugfs_create_file("available_event_classes", 0644,
pmon_entry->iommu_dir, pmon_entry,
&available_event_cls_file_ops)) {
ret = -EIO;
goto free_mem;
}
ret = iommu_pm_create_group_debugfs_hierarchy(iommu, pmon_entry);
if (ret)
goto free_mem;
iommu->hw_ops->initialize_hw(pmon_entry);
if (iommu->evt_irq > 0) {
ret = request_threaded_irq(iommu->evt_irq, NULL,
iommu->hw_ops->evt_ovfl_int_handler,
IRQF_ONESHOT | IRQF_SHARED,
"msm_iommu_pmon_nonsecure_irq", pmon_entry);
if (ret) {
pr_err("Request IRQ %d failed with ret=%d\n",
iommu->evt_irq,
ret);
goto free_mem;
}
} else {
pr_info("%s: Overflow interrupt not available\n", __func__);
}
dev_dbg(iommu->iommu_dev, "%s iommu registered\n", iommu->iommu_name);
goto out;
free_mem:
if (pmon_entry->cnt_grp) {
for (i = 0; i < pmon_entry->num_groups; ++i) {
kfree(pmon_entry->cnt_grp[i].counters);
pmon_entry->cnt_grp[i].counters = 0;
}
}
kfree(pmon_entry->cnt_grp);
pmon_entry->cnt_grp = 0;
file_err:
debugfs_remove_recursive(msm_iommu_root_debugfs_dir);
out:
return ret;
}
EXPORT_SYMBOL(msm_iommu_pm_iommu_register);
void msm_iommu_pm_iommu_unregister(struct device *dev)
{
int i;
struct iommu_pmon *pmon_entry = iommu_pm_get_pm_by_dev(dev);
if (!pmon_entry)
return;
free_irq(pmon_entry->iommu.evt_irq, pmon_entry->iommu.iommu_dev);
if (!pmon_entry)
goto remove_debugfs;
if (pmon_entry->cnt_grp) {
for (i = 0; i < pmon_entry->num_groups; ++i)
kfree(pmon_entry->cnt_grp[i].counters);
}
kfree(pmon_entry->cnt_grp);
remove_debugfs:
debugfs_remove_recursive(msm_iommu_root_debugfs_dir);
return;
}
EXPORT_SYMBOL(msm_iommu_pm_iommu_unregister);
struct iommu_pmon *msm_iommu_pm_alloc(struct device *dev)
{
struct iommu_pmon *pmon_entry;
struct iommu_info *info;
pmon_entry = devm_kzalloc(dev, sizeof(*pmon_entry), GFP_KERNEL);
if (!pmon_entry)
return NULL;
info = &pmon_entry->iommu;
info->iommu_dev = dev;
mutex_init(&pmon_entry->lock);
iommu_pm_add_to_iommu_list(pmon_entry);
return pmon_entry;
}
EXPORT_SYMBOL(msm_iommu_pm_alloc);
void msm_iommu_pm_free(struct device *dev)
{
struct iommu_pmon *pmon = iommu_pm_get_pm_by_dev(dev);
if (pmon)
iommu_pm_del_from_iommu_list(pmon);
}
EXPORT_SYMBOL(msm_iommu_pm_free);
void msm_iommu_attached(struct device *dev)
{
struct iommu_pmon *pmon = iommu_pm_get_pm_by_dev(dev);
if (pmon) {
mutex_lock(&pmon->lock);
++pmon->iommu_attach_count;
if (pmon->iommu_attach_count == 1) {
/* If perf. mon was enabled before we attached we do
* the actual enabling after we attach.
*/
if (pmon->enabled && !pmon->iommu.always_on)
iommu_pm_on(pmon);
}
mutex_unlock(&pmon->lock);
}
}
EXPORT_SYMBOL(msm_iommu_attached);
void msm_iommu_detached(struct device *dev)
{
struct iommu_pmon *pmon = iommu_pm_get_pm_by_dev(dev);
if (pmon) {
mutex_lock(&pmon->lock);
if (pmon->iommu_attach_count == 1) {
/* If perf. mon is still enabled we have to disable
* before we do the detach if iommu is not always on.
*/
if (pmon->enabled && !pmon->iommu.always_on)
iommu_pm_off(pmon);
}
BUG_ON(pmon->iommu_attach_count == 0);
--pmon->iommu_attach_count;
mutex_unlock(&pmon->lock);
}
}
EXPORT_SYMBOL(msm_iommu_detached);