M7350/kernel/drivers/slimbus/slimbus.c
2024-09-09 08:57:42 +00:00

3338 lines
88 KiB
C

/* Copyright (c) 2011-2015, The Linux Foundation. All rights reserved.
*
* This program is free software; you can redistribute it and/or modify
* it under the terms of the GNU General Public License version 2 and
* only version 2 as published by the Free Software Foundation.
*
* This program is distributed in the hope that it will be useful,
* but WITHOUT ANY WARRANTY; without even the implied warranty of
* MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
* GNU General Public License for more details.
*/
#include <linux/kernel.h>
#include <linux/errno.h>
#include <linux/slab.h>
#include <linux/init.h>
#include <linux/completion.h>
#include <linux/idr.h>
#include <linux/pm_runtime.h>
#include <linux/slimbus/slimbus.h>
#define SLIM_PORT_HDL(la, f, p) ((la)<<24 | (f) << 16 | (p))
#define SLIM_HDL_TO_LA(hdl) ((u32)((hdl) & 0xFF000000) >> 24)
#define SLIM_HDL_TO_FLOW(hdl) (((u32)(hdl) & 0xFF0000) >> 16)
#define SLIM_HDL_TO_PORT(hdl) ((u32)(hdl) & 0xFF)
#define SLIM_HDL_TO_CHIDX(hdl) ((u16)(hdl) & 0xFF)
#define SLIM_GRP_TO_NCHAN(hdl) ((u16)(hdl >> 8) & 0xFF)
#define SLIM_SLAVE_PORT(p, la) (((la)<<16) | (p))
#define SLIM_MGR_PORT(p) ((0xFF << 16) | (p))
#define SLIM_LA_MANAGER 0xFF
#define SLIM_START_GRP (1 << 8)
#define SLIM_END_GRP (1 << 9)
#define SLIM_MAX_INTR_COEFF_3 (SLIM_SL_PER_SUPERFRAME/3)
#define SLIM_MAX_INTR_COEFF_1 SLIM_SL_PER_SUPERFRAME
static DEFINE_MUTEX(slim_lock);
static DEFINE_IDR(ctrl_idr);
static struct device_type slim_dev_type;
static struct device_type slim_ctrl_type;
#define DEFINE_SLIM_LDEST_TXN(name, mc, len, rl, rbuf, wbuf, la) \
struct slim_msg_txn name = { rl, 0, mc, SLIM_MSG_DEST_LOGICALADDR, 0,\
len, 0, la, false, rbuf, wbuf, NULL, }
#define DEFINE_SLIM_BCAST_TXN(name, mc, len, rl, rbuf, wbuf, la) \
struct slim_msg_txn name = { rl, 0, mc, SLIM_MSG_DEST_BROADCAST, 0,\
len, 0, la, false, rbuf, wbuf, NULL, }
static const struct slim_device_id *slim_match(const struct slim_device_id *id,
const struct slim_device *slim_dev)
{
while (id->name[0]) {
if (strncmp(slim_dev->name, id->name, SLIMBUS_NAME_SIZE) == 0)
return id;
id++;
}
return NULL;
}
static int slim_device_match(struct device *dev, struct device_driver *driver)
{
struct slim_device *slim_dev;
struct slim_driver *drv = to_slim_driver(driver);
if (dev->type == &slim_dev_type)
slim_dev = to_slim_device(dev);
else
return 0;
if (drv->id_table)
return slim_match(drv->id_table, slim_dev) != NULL;
if (driver->name)
return strncmp(slim_dev->name, driver->name, SLIMBUS_NAME_SIZE)
== 0;
return 0;
}
#ifdef CONFIG_PM_SLEEP
static int slim_legacy_suspend(struct device *dev, pm_message_t mesg)
{
struct slim_device *slim_dev = NULL;
struct slim_driver *driver;
if (dev->type == &slim_dev_type)
slim_dev = to_slim_device(dev);
if (!slim_dev || !dev->driver)
return 0;
driver = to_slim_driver(dev->driver);
if (!driver->suspend)
return 0;
return driver->suspend(slim_dev, mesg);
}
static int slim_legacy_resume(struct device *dev)
{
struct slim_device *slim_dev = NULL;
struct slim_driver *driver;
if (dev->type == &slim_dev_type)
slim_dev = to_slim_device(dev);
if (!slim_dev || !dev->driver)
return 0;
driver = to_slim_driver(dev->driver);
if (!driver->resume)
return 0;
return driver->resume(slim_dev);
}
static int slim_pm_suspend(struct device *dev)
{
const struct dev_pm_ops *pm = dev->driver ? dev->driver->pm : NULL;
if (pm)
return pm_generic_suspend(dev);
else
return slim_legacy_suspend(dev, PMSG_SUSPEND);
}
static int slim_pm_resume(struct device *dev)
{
const struct dev_pm_ops *pm = dev->driver ? dev->driver->pm : NULL;
if (pm)
return pm_generic_resume(dev);
else
return slim_legacy_resume(dev);
}
#else
#define slim_pm_suspend NULL
#define slim_pm_resume NULL
#endif
static const struct dev_pm_ops slimbus_pm = {
.suspend = slim_pm_suspend,
.resume = slim_pm_resume,
SET_RUNTIME_PM_OPS(
pm_generic_suspend,
pm_generic_resume,
NULL
)
};
struct bus_type slimbus_type = {
.name = "slimbus",
.match = slim_device_match,
.pm = &slimbus_pm,
};
EXPORT_SYMBOL_GPL(slimbus_type);
struct device slimbus_dev = {
.init_name = "slimbus",
};
static void __exit slimbus_exit(void)
{
device_unregister(&slimbus_dev);
bus_unregister(&slimbus_type);
}
static int __init slimbus_init(void)
{
int retval;
retval = bus_register(&slimbus_type);
if (!retval)
retval = device_register(&slimbus_dev);
if (retval)
bus_unregister(&slimbus_type);
return retval;
}
postcore_initcall(slimbus_init);
module_exit(slimbus_exit);
static int slim_drv_probe(struct device *dev)
{
const struct slim_driver *sdrv = to_slim_driver(dev->driver);
struct slim_device *sbdev = to_slim_device(dev);
struct slim_controller *ctrl = sbdev->ctrl;
if (sdrv->probe) {
int ret;
ret = sdrv->probe(sbdev);
if (ret)
return ret;
if (sdrv->device_up)
queue_work(ctrl->wq, &sbdev->wd);
return 0;
}
return -ENODEV;
}
static int slim_drv_remove(struct device *dev)
{
const struct slim_driver *sdrv = to_slim_driver(dev->driver);
struct slim_device *sbdev = to_slim_device(dev);
sbdev->notified = false;
if (sdrv->remove)
return sdrv->remove(to_slim_device(dev));
return -ENODEV;
}
static void slim_drv_shutdown(struct device *dev)
{
const struct slim_driver *sdrv = to_slim_driver(dev->driver);
if (sdrv->shutdown)
sdrv->shutdown(to_slim_device(dev));
}
/*
* slim_driver_register: Client driver registration with slimbus
* @drv:Client driver to be associated with client-device.
* This API will register the client driver with the slimbus
* It is called from the driver's module-init function.
*/
int slim_driver_register(struct slim_driver *drv)
{
drv->driver.bus = &slimbus_type;
if (drv->probe)
drv->driver.probe = slim_drv_probe;
if (drv->remove)
drv->driver.remove = slim_drv_remove;
if (drv->shutdown)
drv->driver.shutdown = slim_drv_shutdown;
return driver_register(&drv->driver);
}
EXPORT_SYMBOL_GPL(slim_driver_register);
/*
* slim_driver_unregister: Undo effects of slim_driver_register
* @drv: Client driver to be unregistered
*/
void slim_driver_unregister(struct slim_driver *drv)
{
if (drv)
driver_unregister(&drv->driver);
}
EXPORT_SYMBOL_GPL(slim_driver_unregister);
#define slim_ctrl_attr_gr NULL
static void slim_ctrl_release(struct device *dev)
{
struct slim_controller *ctrl = to_slim_controller(dev);
complete(&ctrl->dev_released);
}
static struct device_type slim_ctrl_type = {
.groups = slim_ctrl_attr_gr,
.release = slim_ctrl_release,
};
static struct slim_controller *slim_ctrl_get(struct slim_controller *ctrl)
{
if (!ctrl || !get_device(&ctrl->dev))
return NULL;
return ctrl;
}
static void slim_ctrl_put(struct slim_controller *ctrl)
{
if (ctrl)
put_device(&ctrl->dev);
}
#define slim_device_attr_gr NULL
#define slim_device_uevent NULL
static void slim_dev_release(struct device *dev)
{
struct slim_device *sbdev = to_slim_device(dev);
slim_ctrl_put(sbdev->ctrl);
}
static struct device_type slim_dev_type = {
.groups = slim_device_attr_gr,
.uevent = slim_device_uevent,
.release = slim_dev_release,
};
static void slim_report(struct work_struct *work)
{
struct slim_driver *sbdrv;
struct slim_device *sbdev =
container_of(work, struct slim_device, wd);
if (!sbdev->dev.driver)
return;
/* check if device-up or down needs to be called */
if ((!sbdev->reported && !sbdev->notified) ||
(sbdev->reported && sbdev->notified))
return;
sbdrv = to_slim_driver(sbdev->dev.driver);
/*
* address no longer valid, means device reported absent, whereas
* address valid, means device reported present
*/
if (sbdev->notified && !sbdev->reported) {
sbdev->notified = false;
if (sbdrv->device_down)
sbdrv->device_down(sbdev);
} else if (!sbdev->notified && sbdev->reported) {
sbdev->notified = true;
if (sbdrv->device_up)
sbdrv->device_up(sbdev);
}
}
/*
* slim_add_device: Add a new device without register board info.
* @ctrl: Controller to which this device is to be added to.
* Called when device doesn't have an explicit client-driver to be probed, or
* the client-driver is a module installed dynamically.
*/
int slim_add_device(struct slim_controller *ctrl, struct slim_device *sbdev)
{
sbdev->dev.bus = &slimbus_type;
sbdev->dev.parent = ctrl->dev.parent;
sbdev->dev.type = &slim_dev_type;
sbdev->dev.driver = NULL;
sbdev->ctrl = ctrl;
slim_ctrl_get(ctrl);
dev_set_name(&sbdev->dev, "%s", sbdev->name);
mutex_init(&sbdev->sldev_reconf);
INIT_LIST_HEAD(&sbdev->mark_define);
INIT_LIST_HEAD(&sbdev->mark_suspend);
INIT_LIST_HEAD(&sbdev->mark_removal);
INIT_WORK(&sbdev->wd, slim_report);
mutex_lock(&ctrl->m_ctrl);
list_add_tail(&sbdev->dev_list, &ctrl->devs);
mutex_unlock(&ctrl->m_ctrl);
/* probe slave on this controller */
return device_register(&sbdev->dev);
}
EXPORT_SYMBOL_GPL(slim_add_device);
struct sbi_boardinfo {
struct list_head list;
struct slim_boardinfo board_info;
};
static LIST_HEAD(board_list);
static LIST_HEAD(slim_ctrl_list);
static DEFINE_MUTEX(board_lock);
/* If controller is not present, only add to boards list */
static void slim_match_ctrl_to_boardinfo(struct slim_controller *ctrl,
struct slim_boardinfo *bi)
{
int ret;
if (ctrl->nr != bi->bus_num)
return;
ret = slim_add_device(ctrl, bi->slim_slave);
if (ret != 0)
dev_err(ctrl->dev.parent, "can't create new device for %s\n",
bi->slim_slave->name);
}
/*
* slim_register_board_info: Board-initialization routine.
* @info: List of all devices on all controllers present on the board.
* @n: number of entries.
* API enumerates respective devices on corresponding controller.
* Called from board-init function.
*/
int slim_register_board_info(struct slim_boardinfo const *info, unsigned n)
{
struct sbi_boardinfo *bi;
int i;
bi = kzalloc(n * sizeof(*bi), GFP_KERNEL);
if (!bi)
return -ENOMEM;
for (i = 0; i < n; i++, bi++, info++) {
struct slim_controller *ctrl;
memcpy(&bi->board_info, info, sizeof(*info));
mutex_lock(&board_lock);
list_add_tail(&bi->list, &board_list);
list_for_each_entry(ctrl, &slim_ctrl_list, list)
slim_match_ctrl_to_boardinfo(ctrl, &bi->board_info);
mutex_unlock(&board_lock);
}
return 0;
}
EXPORT_SYMBOL_GPL(slim_register_board_info);
/*
* slim_ctrl_add_boarddevs: Add devices registered by board-info
* @ctrl: Controller to which these devices are to be added to.
* This API is called by controller when it is up and running.
* If devices on a controller were registered before controller,
* this will make sure that they get probed when controller is up.
*/
void slim_ctrl_add_boarddevs(struct slim_controller *ctrl)
{
struct sbi_boardinfo *bi;
mutex_lock(&board_lock);
list_add_tail(&ctrl->list, &slim_ctrl_list);
list_for_each_entry(bi, &board_list, list)
slim_match_ctrl_to_boardinfo(ctrl, &bi->board_info);
mutex_unlock(&board_lock);
}
EXPORT_SYMBOL_GPL(slim_ctrl_add_boarddevs);
/*
* slim_busnum_to_ctrl: Map bus number to controller
* @busnum: Bus number
* Returns controller representing this bus number
*/
struct slim_controller *slim_busnum_to_ctrl(u32 bus_num)
{
struct slim_controller *ctrl;
mutex_lock(&board_lock);
list_for_each_entry(ctrl, &slim_ctrl_list, list)
if (bus_num == ctrl->nr) {
mutex_unlock(&board_lock);
return ctrl;
}
mutex_unlock(&board_lock);
return NULL;
}
EXPORT_SYMBOL_GPL(slim_busnum_to_ctrl);
static int slim_register_controller(struct slim_controller *ctrl)
{
int ret = 0;
/* Can't register until after driver model init */
if (WARN_ON(!slimbus_type.p)) {
ret = -EPROBE_DEFER;
goto out_list;
}
dev_set_name(&ctrl->dev, "sb-%d", ctrl->nr);
ctrl->dev.bus = &slimbus_type;
ctrl->dev.type = &slim_ctrl_type;
ctrl->num_dev = 0;
if (!ctrl->min_cg)
ctrl->min_cg = SLIM_MIN_CLK_GEAR;
if (!ctrl->max_cg)
ctrl->max_cg = SLIM_MAX_CLK_GEAR;
spin_lock_init(&ctrl->txn_lock);
mutex_init(&ctrl->m_ctrl);
mutex_init(&ctrl->sched.m_reconf);
ret = device_register(&ctrl->dev);
if (ret)
goto out_list;
dev_dbg(&ctrl->dev, "Bus [%s] registered:dev:%p\n", ctrl->name,
&ctrl->dev);
if (ctrl->nports) {
ctrl->ports = kzalloc(ctrl->nports * sizeof(struct slim_port),
GFP_KERNEL);
if (!ctrl->ports) {
ret = -ENOMEM;
goto err_port_failed;
}
}
if (ctrl->nchans) {
ctrl->chans = kzalloc(ctrl->nchans * sizeof(struct slim_ich),
GFP_KERNEL);
if (!ctrl->chans) {
ret = -ENOMEM;
goto err_chan_failed;
}
ctrl->sched.chc1 =
kzalloc(ctrl->nchans * sizeof(struct slim_ich *),
GFP_KERNEL);
if (!ctrl->sched.chc1) {
kfree(ctrl->chans);
ret = -ENOMEM;
goto err_chan_failed;
}
ctrl->sched.chc3 =
kzalloc(ctrl->nchans * sizeof(struct slim_ich *),
GFP_KERNEL);
if (!ctrl->sched.chc3) {
kfree(ctrl->sched.chc1);
kfree(ctrl->chans);
ret = -ENOMEM;
goto err_chan_failed;
}
}
#ifdef DEBUG
ctrl->sched.slots = kzalloc(SLIM_SL_PER_SUPERFRAME, GFP_KERNEL);
#endif
init_completion(&ctrl->pause_comp);
INIT_LIST_HEAD(&ctrl->devs);
ctrl->wq = create_singlethread_workqueue(dev_name(&ctrl->dev));
if (!ctrl->wq)
goto err_workq_failed;
return 0;
err_workq_failed:
kfree(ctrl->sched.chc3);
kfree(ctrl->sched.chc1);
kfree(ctrl->chans);
err_chan_failed:
kfree(ctrl->ports);
err_port_failed:
device_unregister(&ctrl->dev);
out_list:
mutex_lock(&slim_lock);
idr_remove(&ctrl_idr, ctrl->nr);
mutex_unlock(&slim_lock);
return ret;
}
/* slim_remove_device: Remove the effect of slim_add_device() */
void slim_remove_device(struct slim_device *sbdev)
{
struct slim_controller *ctrl = sbdev->ctrl;
mutex_lock(&ctrl->m_ctrl);
list_del_init(&sbdev->dev_list);
mutex_unlock(&ctrl->m_ctrl);
device_unregister(&sbdev->dev);
}
EXPORT_SYMBOL_GPL(slim_remove_device);
static void slim_ctrl_remove_device(struct slim_controller *ctrl,
struct slim_boardinfo *bi)
{
if (ctrl->nr == bi->bus_num)
slim_remove_device(bi->slim_slave);
}
/*
* slim_del_controller: Controller tear-down.
* Controller added with the above API is teared down using this API.
*/
int slim_del_controller(struct slim_controller *ctrl)
{
struct slim_controller *found;
struct sbi_boardinfo *bi;
/* First make sure that this bus was added */
mutex_lock(&slim_lock);
found = idr_find(&ctrl_idr, ctrl->nr);
mutex_unlock(&slim_lock);
if (found != ctrl)
return -EINVAL;
/* Remove all clients */
mutex_lock(&board_lock);
list_for_each_entry(bi, &board_list, list)
slim_ctrl_remove_device(ctrl, &bi->board_info);
mutex_unlock(&board_lock);
init_completion(&ctrl->dev_released);
device_unregister(&ctrl->dev);
wait_for_completion(&ctrl->dev_released);
list_del(&ctrl->list);
destroy_workqueue(ctrl->wq);
/* free bus id */
mutex_lock(&slim_lock);
idr_remove(&ctrl_idr, ctrl->nr);
mutex_unlock(&slim_lock);
kfree(ctrl->sched.chc1);
kfree(ctrl->sched.chc3);
#ifdef DEBUG
kfree(ctrl->sched.slots);
#endif
kfree(ctrl->chans);
kfree(ctrl->ports);
return 0;
}
EXPORT_SYMBOL_GPL(slim_del_controller);
/*
* slim_add_numbered_controller: Controller bring-up.
* @ctrl: Controller to be registered.
* A controller is registered with the framework using this API. ctrl->nr is the
* desired number with which slimbus framework registers the controller.
* Function will return -EBUSY if the number is in use.
*/
int slim_add_numbered_controller(struct slim_controller *ctrl)
{
int id;
mutex_lock(&slim_lock);
id = idr_alloc(&ctrl_idr, ctrl, ctrl->nr, ctrl->nr + 1, GFP_KERNEL);
mutex_unlock(&slim_lock);
if (id < 0)
return id;
ctrl->nr = id;
return slim_register_controller(ctrl);
}
EXPORT_SYMBOL_GPL(slim_add_numbered_controller);
/*
* slim_report_absent: Controller calls this function when a device
* reports absent, OR when the device cannot be communicated with
* @sbdev: Device that cannot be reached, or sent report absent
*/
void slim_report_absent(struct slim_device *sbdev)
{
struct slim_controller *ctrl;
int i;
if (!sbdev)
return;
ctrl = sbdev->ctrl;
if (!ctrl)
return;
/* invalidate logical addresses */
mutex_lock(&ctrl->m_ctrl);
for (i = 0; i < ctrl->num_dev; i++) {
if (sbdev->laddr == ctrl->addrt[i].laddr)
ctrl->addrt[i].valid = false;
}
mutex_unlock(&ctrl->m_ctrl);
sbdev->reported = false;
queue_work(ctrl->wq, &sbdev->wd);
}
EXPORT_SYMBOL(slim_report_absent);
/*
* slim_framer_booted: This function is called by controller after the active
* framer has booted (using Bus Reset sequence, or after it has shutdown and has
* come back up). Components, devices on the bus may be in undefined state,
* and this function triggers their drivers to do the needful
* to bring them back in Reset state so that they can acquire sync, report
* present and be operational again.
*/
void slim_framer_booted(struct slim_controller *ctrl)
{
struct slim_device *sbdev;
struct list_head *pos, *next;
if (!ctrl)
return;
mutex_lock(&ctrl->m_ctrl);
list_for_each_safe(pos, next, &ctrl->devs) {
struct slim_driver *sbdrv;
sbdev = list_entry(pos, struct slim_device, dev_list);
mutex_unlock(&ctrl->m_ctrl);
if (sbdev && sbdev->dev.driver) {
sbdrv = to_slim_driver(sbdev->dev.driver);
if (sbdrv->reset_device)
sbdrv->reset_device(sbdev);
}
mutex_lock(&ctrl->m_ctrl);
}
mutex_unlock(&ctrl->m_ctrl);
}
EXPORT_SYMBOL(slim_framer_booted);
/*
* slim_msg_response: Deliver Message response received from a device to the
* framework.
* @ctrl: Controller handle
* @reply: Reply received from the device
* @len: Length of the reply
* @tid: Transaction ID received with which framework can associate reply.
* Called by controller to inform framework about the response received.
* This helps in making the API asynchronous, and controller-driver doesn't need
* to manage 1 more table other than the one managed by framework mapping TID
* with buffers
*/
void slim_msg_response(struct slim_controller *ctrl, u8 *reply, u8 tid, u8 len)
{
int i;
unsigned long flags;
bool async;
struct slim_msg_txn *txn;
spin_lock_irqsave(&ctrl->txn_lock, flags);
txn = ctrl->txnt[tid];
if (txn == NULL || txn->rbuf == NULL) {
spin_unlock_irqrestore(&ctrl->txn_lock, flags);
if (txn == NULL)
dev_err(&ctrl->dev, "Got response to invalid TID:%d, len:%d",
tid, len);
else
dev_err(&ctrl->dev, "Invalid client buffer passed\n");
return;
}
async = txn->async;
for (i = 0; i < len; i++)
txn->rbuf[i] = reply[i];
if (txn->comp)
complete(txn->comp);
ctrl->txnt[tid] = NULL;
spin_unlock_irqrestore(&ctrl->txn_lock, flags);
if (async)
kfree(txn);
}
EXPORT_SYMBOL_GPL(slim_msg_response);
static int slim_processtxn(struct slim_controller *ctrl,
struct slim_msg_txn *txn, bool need_tid)
{
u8 i = 0;
int ret = 0;
unsigned long flags;
if (need_tid) {
spin_lock_irqsave(&ctrl->txn_lock, flags);
for (i = 0; i < ctrl->last_tid; i++) {
if (ctrl->txnt[i] == NULL)
break;
}
if (i >= ctrl->last_tid) {
if (ctrl->last_tid == 255) {
spin_unlock_irqrestore(&ctrl->txn_lock, flags);
return -ENOMEM;
}
ctrl->last_tid++;
}
ctrl->txnt[i] = txn;
txn->tid = i;
spin_unlock_irqrestore(&ctrl->txn_lock, flags);
}
ret = ctrl->xfer_msg(ctrl, txn);
return ret;
}
static int ctrl_getlogical_addr(struct slim_controller *ctrl, const u8 *eaddr,
u8 e_len, u8 *entry)
{
u8 i;
for (i = 0; i < ctrl->num_dev; i++) {
if (ctrl->addrt[i].valid &&
memcmp(ctrl->addrt[i].eaddr, eaddr, e_len) == 0) {
*entry = i;
return 0;
}
}
return -ENXIO;
}
/*
* slim_assign_laddr: Assign logical address to a device enumerated.
* @ctrl: Controller with which device is enumerated.
* @e_addr: 6-byte elemental address of the device.
* @e_len: buffer length for e_addr
* @laddr: Return logical address (if valid flag is false)
* @valid: true if laddr holds a valid address that controller wants to
* set for this enumeration address. Otherwise framework sets index into
* address table as logical address.
* Called by controller in response to REPORT_PRESENT. Framework will assign
* a logical address to this enumeration address.
* Function returns -EXFULL to indicate that all logical addresses are already
* taken.
*/
int slim_assign_laddr(struct slim_controller *ctrl, const u8 *e_addr,
u8 e_len, u8 *laddr, bool valid)
{
int ret;
u8 i = 0;
bool exists = false;
struct slim_device *sbdev;
struct list_head *pos, *next;
mutex_lock(&ctrl->m_ctrl);
/* already assigned */
if (ctrl_getlogical_addr(ctrl, e_addr, e_len, &i) == 0) {
*laddr = ctrl->addrt[i].laddr;
exists = true;
} else {
if (ctrl->num_dev >= 254) {
ret = -EXFULL;
goto ret_assigned_laddr;
}
for (i = 0; i < ctrl->num_dev; i++) {
if (ctrl->addrt[i].valid == false)
break;
}
if (i == ctrl->num_dev) {
ctrl->addrt = krealloc(ctrl->addrt,
(ctrl->num_dev + 1) *
sizeof(struct slim_addrt),
GFP_KERNEL);
if (!ctrl->addrt) {
ret = -ENOMEM;
goto ret_assigned_laddr;
}
ctrl->num_dev++;
}
memcpy(ctrl->addrt[i].eaddr, e_addr, e_len);
ctrl->addrt[i].valid = true;
/* Preferred address is index into table */
if (!valid)
*laddr = i;
}
ret = ctrl->set_laddr(ctrl, (const u8 *)&ctrl->addrt[i].eaddr, 6,
*laddr);
if (ret) {
ctrl->addrt[i].valid = false;
goto ret_assigned_laddr;
}
ctrl->addrt[i].laddr = *laddr;
dev_dbg(&ctrl->dev, "setting slimbus l-addr:%x\n", *laddr);
ret_assigned_laddr:
mutex_unlock(&ctrl->m_ctrl);
if (exists || ret)
return ret;
pr_info("slimbus:%d laddr:0x%x, EAPC:0x%x:0x%x", ctrl->nr, *laddr,
e_addr[1], e_addr[2]);
mutex_lock(&ctrl->m_ctrl);
list_for_each_safe(pos, next, &ctrl->devs) {
sbdev = list_entry(pos, struct slim_device, dev_list);
if (memcmp(sbdev->e_addr, e_addr, 6) == 0) {
struct slim_driver *sbdrv;
sbdev->laddr = *laddr;
sbdev->reported = true;
if (sbdev->dev.driver) {
sbdrv = to_slim_driver(sbdev->dev.driver);
if (sbdrv->device_up)
queue_work(ctrl->wq, &sbdev->wd);
}
break;
}
}
mutex_unlock(&ctrl->m_ctrl);
return 0;
}
EXPORT_SYMBOL_GPL(slim_assign_laddr);
/*
* slim_get_logical_addr: Return the logical address of a slimbus device.
* @sb: client handle requesting the adddress.
* @e_addr: Elemental address of the device.
* @e_len: Length of e_addr
* @laddr: output buffer to store the address
* context: can sleep
* -EINVAL is returned in case of invalid parameters, and -ENXIO is returned if
* the device with this elemental address is not found.
*/
int slim_get_logical_addr(struct slim_device *sb, const u8 *e_addr,
u8 e_len, u8 *laddr)
{
int ret = 0;
u8 entry;
struct slim_controller *ctrl = sb->ctrl;
if (!ctrl || !laddr || !e_addr || e_len != 6)
return -EINVAL;
mutex_lock(&ctrl->m_ctrl);
ret = ctrl_getlogical_addr(ctrl, e_addr, e_len, &entry);
if (!ret)
*laddr = ctrl->addrt[entry].laddr;
mutex_unlock(&ctrl->m_ctrl);
if (ret == -ENXIO && ctrl->get_laddr) {
ret = ctrl->get_laddr(ctrl, e_addr, e_len, laddr);
if (!ret)
ret = slim_assign_laddr(ctrl, e_addr, e_len, laddr,
true);
}
return ret;
}
EXPORT_SYMBOL_GPL(slim_get_logical_addr);
static int slim_ele_access_sanity(struct slim_ele_access *msg, int oper,
u8 *rbuf, const u8 *wbuf, u8 len)
{
if (!msg || msg->num_bytes > 16 || msg->start_offset + len > 0xC00)
return -EINVAL;
switch (oper) {
case SLIM_MSG_MC_REQUEST_VALUE:
case SLIM_MSG_MC_REQUEST_INFORMATION:
if (rbuf == NULL)
return -EINVAL;
return 0;
case SLIM_MSG_MC_CHANGE_VALUE:
case SLIM_MSG_MC_CLEAR_INFORMATION:
if (wbuf == NULL)
return -EINVAL;
return 0;
case SLIM_MSG_MC_REQUEST_CHANGE_VALUE:
case SLIM_MSG_MC_REQUEST_CLEAR_INFORMATION:
if (rbuf == NULL || wbuf == NULL)
return -EINVAL;
return 0;
default:
return -EINVAL;
}
}
static u16 slim_slicecodefromsize(u32 req)
{
u8 codetosize[8] = {1, 2, 3, 4, 6, 8, 12, 16};
if (req >= 8)
return 0;
else
return codetosize[req];
}
static u16 slim_slicesize(u32 code)
{
u8 sizetocode[16] = {0, 1, 2, 3, 3, 4, 4, 5, 5, 5, 5, 6, 6, 6, 6, 7};
if (code == 0)
code = 1;
if (code > 16)
code = 16;
return sizetocode[code - 1];
}
/* Message APIs Unicast message APIs used by slimbus slave drivers */
/*
* Message API access routines.
* @sb: client handle requesting elemental message reads, writes.
* @msg: Input structure for start-offset, number of bytes to read.
* @rbuf: data buffer to be filled with values read.
* @len: data buffer size
* @wbuf: data buffer containing value/information to be written
* context: can sleep
* Returns:
* -EINVAL: Invalid parameters
* -ETIMEDOUT: If controller could not complete the request. This may happen if
* the bus lines are not clocked, controller is not powered-on, slave with
* given address is not enumerated/responding.
*/
int slim_request_val_element(struct slim_device *sb,
struct slim_ele_access *msg, u8 *buf, u8 len)
{
struct slim_controller *ctrl = sb->ctrl;
if (!ctrl)
return -EINVAL;
return slim_xfer_msg(ctrl, sb, msg, SLIM_MSG_MC_REQUEST_VALUE, buf,
NULL, len);
}
EXPORT_SYMBOL_GPL(slim_request_val_element);
int slim_request_inf_element(struct slim_device *sb,
struct slim_ele_access *msg, u8 *buf, u8 len)
{
struct slim_controller *ctrl = sb->ctrl;
if (!ctrl)
return -EINVAL;
return slim_xfer_msg(ctrl, sb, msg, SLIM_MSG_MC_REQUEST_INFORMATION,
buf, NULL, len);
}
EXPORT_SYMBOL_GPL(slim_request_inf_element);
int slim_change_val_element(struct slim_device *sb, struct slim_ele_access *msg,
const u8 *buf, u8 len)
{
struct slim_controller *ctrl = sb->ctrl;
if (!ctrl)
return -EINVAL;
return slim_xfer_msg(ctrl, sb, msg, SLIM_MSG_MC_CHANGE_VALUE, NULL, buf,
len);
}
EXPORT_SYMBOL_GPL(slim_change_val_element);
int slim_clear_inf_element(struct slim_device *sb, struct slim_ele_access *msg,
u8 *buf, u8 len)
{
struct slim_controller *ctrl = sb->ctrl;
if (!ctrl)
return -EINVAL;
return slim_xfer_msg(ctrl, sb, msg, SLIM_MSG_MC_CLEAR_INFORMATION, NULL,
buf, len);
}
EXPORT_SYMBOL_GPL(slim_clear_inf_element);
int slim_request_change_val_element(struct slim_device *sb,
struct slim_ele_access *msg, u8 *rbuf,
const u8 *wbuf, u8 len)
{
struct slim_controller *ctrl = sb->ctrl;
if (!ctrl)
return -EINVAL;
return slim_xfer_msg(ctrl, sb, msg, SLIM_MSG_MC_REQUEST_CHANGE_VALUE,
rbuf, wbuf, len);
}
EXPORT_SYMBOL_GPL(slim_request_change_val_element);
int slim_request_clear_inf_element(struct slim_device *sb,
struct slim_ele_access *msg, u8 *rbuf,
const u8 *wbuf, u8 len)
{
struct slim_controller *ctrl = sb->ctrl;
if (!ctrl)
return -EINVAL;
return slim_xfer_msg(ctrl, sb, msg,
SLIM_MSG_MC_REQUEST_CLEAR_INFORMATION,
rbuf, wbuf, len);
}
EXPORT_SYMBOL_GPL(slim_request_clear_inf_element);
/*
* Broadcast message API:
* call this API directly with sbdev = NULL.
* For broadcast reads, make sure that buffers are big-enough to incorporate
* replies from all logical addresses.
* All controllers may not support broadcast
*/
int slim_xfer_msg(struct slim_controller *ctrl, struct slim_device *sbdev,
struct slim_ele_access *msg, u16 mc, u8 *rbuf,
const u8 *wbuf, u8 len)
{
DECLARE_COMPLETION_ONSTACK(complete);
DEFINE_SLIM_LDEST_TXN(txn_stack, mc, len, 6, rbuf, wbuf, sbdev->laddr);
struct slim_msg_txn *txn;
int ret;
u16 sl, cur;
if (msg->comp && rbuf) {
txn = kmalloc(sizeof(struct slim_msg_txn),
GFP_KERNEL);
if (IS_ERR_OR_NULL(txn))
return PTR_ERR(txn);
*txn = txn_stack;
txn->async = true;
txn->comp = msg->comp;
} else {
txn = &txn_stack;
if (rbuf)
txn->comp = &complete;
}
ret = slim_ele_access_sanity(msg, mc, rbuf, wbuf, len);
if (ret)
goto xfer_err;
sl = slim_slicesize(len);
dev_dbg(&ctrl->dev, "SB xfer msg:os:%x, len:%d, MC:%x, sl:%x\n",
msg->start_offset, len, mc, sl);
cur = slim_slicecodefromsize(sl);
txn->ec = ((sl | (1 << 3)) | ((msg->start_offset & 0xFFF) << 4));
if (wbuf)
txn->rl += len;
if (rbuf) {
unsigned long flags;
txn->rl++;
ret = slim_processtxn(ctrl, txn, true);
/* sync read */
if (!ret && !msg->comp) {
ret = wait_for_completion_timeout(&complete, HZ);
if (!ret) {
dev_err(&ctrl->dev, "slimbus Read timed out");
spin_lock_irqsave(&ctrl->txn_lock, flags);
/* Invalidate the transaction */
ctrl->txnt[txn->tid] = NULL;
spin_unlock_irqrestore(&ctrl->txn_lock, flags);
ret = -ETIMEDOUT;
} else
ret = 0;
} else if (ret < 0 && !msg->comp) {
dev_err(&ctrl->dev, "slimbus Read error");
spin_lock_irqsave(&ctrl->txn_lock, flags);
/* Invalidate the transaction */
ctrl->txnt[txn->tid] = NULL;
spin_unlock_irqrestore(&ctrl->txn_lock, flags);
}
} else
ret = slim_processtxn(ctrl, txn, false);
xfer_err:
return ret;
}
EXPORT_SYMBOL_GPL(slim_xfer_msg);
/*
* User message:
* slim_user_msg: Send user message that is interpreted by destination device
* @sb: Client handle sending the message
* @la: Destination device for this user message
* @mt: Message Type (Soruce-referred, or Destination-referred)
* @mc: Message Code
* @msg: Message structure (start offset, number of bytes) to be sent
* @buf: data buffer to be sent
* @len: data buffer size in bytes
*/
int slim_user_msg(struct slim_device *sb, u8 la, u8 mt, u8 mc,
struct slim_ele_access *msg, u8 *buf, u8 len)
{
if (!sb || !sb->ctrl || !msg || mt == SLIM_MSG_MT_CORE)
return -EINVAL;
if (!sb->ctrl->xfer_user_msg)
return -EPROTONOSUPPORT;
return sb->ctrl->xfer_user_msg(sb->ctrl, la, mt, mc, msg, buf, len);
}
EXPORT_SYMBOL(slim_user_msg);
/*
* Queue bulk of message writes:
* slim_bulk_msg_write: Write bulk of messages (e.g. downloading FW)
* @sb: Client handle sending these messages
* @la: Destination device for these messages
* @mt: Message Type
* @mc: Message Code
* @msgs: List of messages to be written in bulk
* @n: Number of messages in the list
* @cb: Callback if client needs this to be non-blocking
* @ctx: Context for this callback
* If supported by controller, this message list will be sent in bulk to the HW
* If the client specifies this to be non-blocking, the callback will be
* called from atomic context.
*/
int slim_bulk_msg_write(struct slim_device *sb, u8 mt, u8 mc,
struct slim_val_inf msgs[], int n,
int (*comp_cb)(void *ctx, int err), void *ctx)
{
int i, ret;
if (!sb || !sb->ctrl || !msgs)
return -EINVAL;
if (!sb->ctrl->xfer_bulk_wr) {
pr_warn("controller does not support bulk WR, serializing");
for (i = 0; i < n; i++) {
struct slim_ele_access ele;
ele.comp = NULL;
ele.start_offset = msgs[i].start_offset;
ele.num_bytes = msgs[i].num_bytes;
ret = slim_xfer_msg(sb->ctrl, sb, &ele, mc,
msgs[i].rbuf, msgs[i].wbuf,
ele.num_bytes);
if (ret)
return ret;
}
return ret;
}
return sb->ctrl->xfer_bulk_wr(sb->ctrl, sb->laddr, mt, mc, msgs, n,
comp_cb, ctx);
}
EXPORT_SYMBOL(slim_bulk_msg_write);
/*
* slim_alloc_mgrports: Allocate port on manager side.
* @sb: device/client handle.
* @req: Port request type.
* @nports: Number of ports requested
* @rh: output buffer to store the port handles
* @hsz: size of buffer storing handles
* context: can sleep
* This port will be typically used by SW. e.g. client driver wants to receive
* some data from audio codec HW using a data channel.
* Port allocated using this API will be used to receive the data.
* If half-duplex ports are requested, two adjacent ports are allocated for
* 1 half-duplex port. So the handle-buffer size should be twice the number
* of half-duplex ports to be allocated.
* -EDQUOT is returned if all ports are in use.
*/
int slim_alloc_mgrports(struct slim_device *sb, enum slim_port_req req,
int nports, u32 *rh, int hsz)
{
int i, j;
int ret = -EINVAL;
int nphysp = nports;
struct slim_controller *ctrl = sb->ctrl;
if (!rh || !ctrl)
return -EINVAL;
if (req == SLIM_REQ_HALF_DUP)
nphysp *= 2;
if (hsz/sizeof(u32) < nphysp)
return -EINVAL;
mutex_lock(&ctrl->m_ctrl);
for (i = 0; i < ctrl->nports; i++) {
bool multiok = true;
if (ctrl->ports[i].state != SLIM_P_FREE)
continue;
/* Start half duplex channel at even port */
if (req == SLIM_REQ_HALF_DUP && (i % 2))
continue;
/* Allocate ports contiguously for multi-ch */
if (ctrl->nports < (i + nphysp)) {
i = ctrl->nports;
break;
}
if (req == SLIM_REQ_MULTI_CH) {
multiok = true;
for (j = i; j < i + nphysp; j++) {
if (ctrl->ports[j].state != SLIM_P_FREE) {
multiok = false;
break;
}
}
if (!multiok)
continue;
}
break;
}
if (i >= ctrl->nports) {
ret = -EDQUOT;
goto alloc_err;
}
ret = 0;
for (j = i; j < i + nphysp; j++) {
ctrl->ports[j].state = SLIM_P_UNCFG;
ctrl->ports[j].req = req;
if (req == SLIM_REQ_HALF_DUP && (j % 2))
ctrl->ports[j].flow = SLIM_SINK;
else
ctrl->ports[j].flow = SLIM_SRC;
if (ctrl->alloc_port)
ret = ctrl->alloc_port(ctrl, j);
if (ret) {
for (; j >= i; j--)
ctrl->ports[j].state = SLIM_P_FREE;
goto alloc_err;
}
*rh++ = SLIM_PORT_HDL(SLIM_LA_MANAGER, 0, j);
}
alloc_err:
mutex_unlock(&ctrl->m_ctrl);
return ret;
}
EXPORT_SYMBOL_GPL(slim_alloc_mgrports);
/* Deallocate the port(s) allocated using the API above */
int slim_dealloc_mgrports(struct slim_device *sb, u32 *hdl, int nports)
{
int i;
struct slim_controller *ctrl = sb->ctrl;
if (!ctrl || !hdl)
return -EINVAL;
mutex_lock(&ctrl->m_ctrl);
for (i = 0; i < nports; i++) {
u8 pn;
pn = SLIM_HDL_TO_PORT(hdl[i]);
if (pn >= ctrl->nports || ctrl->ports[pn].state == SLIM_P_CFG) {
int j, ret;
if (pn >= ctrl->nports) {
dev_err(&ctrl->dev, "invalid port number");
ret = -EINVAL;
} else {
dev_err(&ctrl->dev,
"Can't dealloc connected port:%d", i);
ret = -EISCONN;
}
for (j = i - 1; j >= 0; j--) {
pn = SLIM_HDL_TO_PORT(hdl[j]);
ctrl->ports[pn].state = SLIM_P_UNCFG;
}
mutex_unlock(&ctrl->m_ctrl);
return ret;
}
if (ctrl->dealloc_port)
ctrl->dealloc_port(ctrl, pn);
ctrl->ports[pn].state = SLIM_P_FREE;
}
mutex_unlock(&ctrl->m_ctrl);
return 0;
}
EXPORT_SYMBOL_GPL(slim_dealloc_mgrports);
/*
* slim_config_mgrports: Configure manager side ports
* @sb: device/client handle.
* @ph: array of port handles for which this configuration is valid
* @nports: Number of ports in ph
* @cfg: configuration requested for port(s)
* Configure port settings if they are different than the default ones.
* Returns success if the config could be applied. Returns -EISCONN if the
* port is in use
*/
int slim_config_mgrports(struct slim_device *sb, u32 *ph, int nports,
struct slim_port_cfg *cfg)
{
int i;
struct slim_controller *ctrl;
if (!sb || !ph || !nports || !sb->ctrl || !cfg)
return -EINVAL;
ctrl = sb->ctrl;
mutex_lock(&ctrl->sched.m_reconf);
for (i = 0; i < nports; i++) {
u8 pn = SLIM_HDL_TO_PORT(ph[i]);
if (ctrl->ports[pn].state == SLIM_P_CFG)
return -EISCONN;
ctrl->ports[pn].cfg = *cfg;
}
mutex_unlock(&ctrl->sched.m_reconf);
return 0;
}
EXPORT_SYMBOL(slim_config_mgrports);
/*
* slim_get_slaveport: Get slave port handle
* @la: slave device logical address.
* @idx: port index at slave
* @rh: return handle
* @flw: Flow type (source or destination)
* This API only returns a slave port's representation as expected by slimbus
* driver. This port is not managed by the slimbus driver. Caller is expected
* to have visibility of this port since it's a device-port.
*/
int slim_get_slaveport(u8 la, int idx, u32 *rh, enum slim_port_flow flw)
{
if (rh == NULL)
return -EINVAL;
*rh = SLIM_PORT_HDL(la, flw, idx);
return 0;
}
EXPORT_SYMBOL_GPL(slim_get_slaveport);
static int connect_port_ch(struct slim_controller *ctrl, u8 ch, u32 ph,
enum slim_port_flow flow)
{
int ret;
u8 buf[2];
u32 la = SLIM_HDL_TO_LA(ph);
u8 pn = (u8)SLIM_HDL_TO_PORT(ph);
DEFINE_SLIM_LDEST_TXN(txn, 0, 2, 6, NULL, buf, la);
if (flow == SLIM_SRC)
txn.mc = SLIM_MSG_MC_CONNECT_SOURCE;
else
txn.mc = SLIM_MSG_MC_CONNECT_SINK;
buf[0] = pn;
buf[1] = ctrl->chans[ch].chan;
if (la == SLIM_LA_MANAGER)
ctrl->ports[pn].flow = flow;
ret = slim_processtxn(ctrl, &txn, false);
if (!ret && la == SLIM_LA_MANAGER)
ctrl->ports[pn].state = SLIM_P_CFG;
return ret;
}
static int disconnect_port_ch(struct slim_controller *ctrl, u32 ph)
{
int ret;
u32 la = SLIM_HDL_TO_LA(ph);
u8 pn = (u8)SLIM_HDL_TO_PORT(ph);
DEFINE_SLIM_LDEST_TXN(txn, 0, 1, 5, NULL, &pn, la);
txn.mc = SLIM_MSG_MC_DISCONNECT_PORT;
ret = slim_processtxn(ctrl, &txn, false);
if (ret)
return ret;
if (la == SLIM_LA_MANAGER) {
ctrl->ports[pn].state = SLIM_P_UNCFG;
ctrl->ports[pn].cfg.watermark = 0;
ctrl->ports[pn].cfg.port_opts = 0;
ctrl->ports[pn].ch = NULL;
}
return 0;
}
/*
* slim_connect_src: Connect source port to channel.
* @sb: client handle
* @srch: source handle to be connected to this channel
* @chanh: Channel with which the ports need to be associated with.
* Per slimbus specification, a channel may have 1 source port.
* Channel specified in chanh needs to be allocated first.
* Returns -EALREADY if source is already configured for this channel.
* Returns -ENOTCONN if channel is not allocated
* Returns -EINVAL if invalid direction is specified for non-manager port,
* or if the manager side port number is out of bounds, or in incorrect state
*/
int slim_connect_src(struct slim_device *sb, u32 srch, u16 chanh)
{
struct slim_controller *ctrl = sb->ctrl;
int ret;
u8 chan = SLIM_HDL_TO_CHIDX(chanh);
struct slim_ich *slc = &ctrl->chans[chan];
enum slim_port_flow flow = SLIM_HDL_TO_FLOW(srch);
u8 la = SLIM_HDL_TO_LA(srch);
u8 pn = SLIM_HDL_TO_PORT(srch);
/* manager ports don't have direction when they are allocated */
if (la != SLIM_LA_MANAGER && flow != SLIM_SRC)
return -EINVAL;
mutex_lock(&ctrl->sched.m_reconf);
if (la == SLIM_LA_MANAGER) {
if (pn >= ctrl->nports ||
ctrl->ports[pn].state != SLIM_P_UNCFG) {
ret = -EINVAL;
goto connect_src_err;
}
}
if (slc->state == SLIM_CH_FREE) {
ret = -ENOTCONN;
goto connect_src_err;
}
/*
* Once channel is removed, its ports can be considered disconnected
* So its ports can be reassigned. Source port is zeroed
* when channel is deallocated.
*/
if (slc->srch) {
ret = -EALREADY;
goto connect_src_err;
}
ctrl->ports[pn].ch = &slc->prop;
ret = connect_port_ch(ctrl, chan, srch, SLIM_SRC);
if (!ret)
slc->srch = srch;
connect_src_err:
mutex_unlock(&ctrl->sched.m_reconf);
return ret;
}
EXPORT_SYMBOL_GPL(slim_connect_src);
/*
* slim_connect_sink: Connect sink port(s) to channel.
* @sb: client handle
* @sinkh: sink handle(s) to be connected to this channel
* @nsink: number of sinks
* @chanh: Channel with which the ports need to be associated with.
* Per slimbus specification, a channel may have multiple sink-ports.
* Channel specified in chanh needs to be allocated first.
* Returns -EALREADY if sink is already configured for this channel.
* Returns -ENOTCONN if channel is not allocated
* Returns -EINVAL if invalid parameters are passed, or invalid direction is
* specified for non-manager port, or if the manager side port number is out of
* bounds, or in incorrect state
*/
int slim_connect_sink(struct slim_device *sb, u32 *sinkh, int nsink, u16 chanh)
{
struct slim_controller *ctrl = sb->ctrl;
int j;
int ret = 0;
u8 chan = SLIM_HDL_TO_CHIDX(chanh);
struct slim_ich *slc = &ctrl->chans[chan];
if (!sinkh || !nsink)
return -EINVAL;
mutex_lock(&ctrl->sched.m_reconf);
/*
* Once channel is removed, its ports can be considered disconnected
* So its ports can be reassigned. Sink ports are freed when channel
* is deallocated.
*/
if (slc->state == SLIM_CH_FREE) {
ret = -ENOTCONN;
goto connect_sink_err;
}
for (j = 0; j < nsink; j++) {
enum slim_port_flow flow = SLIM_HDL_TO_FLOW(sinkh[j]);
u8 la = SLIM_HDL_TO_LA(sinkh[j]);
u8 pn = SLIM_HDL_TO_PORT(sinkh[j]);
if (la != SLIM_LA_MANAGER && flow != SLIM_SINK)
ret = -EINVAL;
else if (la == SLIM_LA_MANAGER &&
(pn >= ctrl->nports ||
ctrl->ports[pn].state != SLIM_P_UNCFG))
ret = -EINVAL;
else {
ctrl->ports[pn].ch = &slc->prop;
ret = connect_port_ch(ctrl, chan, sinkh[j], SLIM_SINK);
}
if (ret) {
for (j = j - 1; j >= 0; j--)
disconnect_port_ch(ctrl, sinkh[j]);
goto connect_sink_err;
}
}
slc->sinkh = krealloc(slc->sinkh, (sizeof(u32) * (slc->nsink + nsink)),
GFP_KERNEL);
if (!slc->sinkh) {
ret = -ENOMEM;
for (j = 0; j < nsink; j++)
disconnect_port_ch(ctrl, sinkh[j]);
goto connect_sink_err;
}
memcpy(slc->sinkh + slc->nsink, sinkh, (sizeof(u32) * nsink));
slc->nsink += nsink;
connect_sink_err:
mutex_unlock(&ctrl->sched.m_reconf);
return ret;
}
EXPORT_SYMBOL_GPL(slim_connect_sink);
/*
* slim_disconnect_ports: Disconnect port(s) from channel
* @sb: client handle
* @ph: ports to be disconnected
* @nph: number of ports.
* Disconnects ports from a channel.
*/
int slim_disconnect_ports(struct slim_device *sb, u32 *ph, int nph)
{
struct slim_controller *ctrl = sb->ctrl;
int i;
mutex_lock(&ctrl->sched.m_reconf);
for (i = 0; i < nph; i++)
disconnect_port_ch(ctrl, ph[i]);
mutex_unlock(&ctrl->sched.m_reconf);
return 0;
}
EXPORT_SYMBOL_GPL(slim_disconnect_ports);
/*
* slim_port_xfer: Schedule buffer to be transferred/received using port-handle.
* @sb: client handle
* @ph: port-handle
* @iobuf: buffer to be transferred or populated
* @len: buffer size.
* @comp: completion signal to indicate transfer done or error.
* context: can sleep
* Returns number of bytes transferred/received if used synchronously.
* Will return 0 if used asynchronously.
* Client will call slim_port_get_xfer_status to get error and/or number of
* bytes transferred if used asynchronously.
*/
int slim_port_xfer(struct slim_device *sb, u32 ph, phys_addr_t iobuf, u32 len,
struct completion *comp)
{
struct slim_controller *ctrl = sb->ctrl;
u8 pn = SLIM_HDL_TO_PORT(ph);
dev_dbg(&ctrl->dev, "port xfer: num:%d", pn);
return ctrl->port_xfer(ctrl, pn, iobuf, len, comp);
}
EXPORT_SYMBOL_GPL(slim_port_xfer);
/*
* slim_port_get_xfer_status: Poll for port transfers, or get transfer status
* after completion is done.
* @sb: client handle
* @ph: port-handle
* @done_buf: return pointer (iobuf from slim_port_xfer) which is processed.
* @done_len: Number of bytes transferred.
* This can be called when port_xfer complition is signalled.
* The API will return port transfer error (underflow/overflow/disconnect)
* and/or done_len will reflect number of bytes transferred. Note that
* done_len may be valid even if port error (overflow/underflow) has happened.
* e.g. If the transfer was scheduled with a few bytes to be transferred and
* client has not supplied more data to be transferred, done_len will indicate
* number of bytes transferred with underflow error. To avoid frequent underflow
* errors, multiple transfers can be queued (e.g. ping-pong buffers) so that
* channel has data to be transferred even if client is not ready to transfer
* data all the time. done_buf will indicate address of the last buffer
* processed from the multiple transfers.
*/
enum slim_port_err slim_port_get_xfer_status(struct slim_device *sb, u32 ph,
phys_addr_t *done_buf, u32 *done_len)
{
struct slim_controller *ctrl = sb->ctrl;
u8 pn = SLIM_HDL_TO_PORT(ph);
u32 la = SLIM_HDL_TO_LA(ph);
enum slim_port_err err;
dev_dbg(&ctrl->dev, "get status port num:%d", pn);
/*
* Framework only has insight into ports managed by ported device
* used by the manager and not slave
*/
if (la != SLIM_LA_MANAGER) {
if (done_buf)
*done_buf = 0;
if (done_len)
*done_len = 0;
return SLIM_P_NOT_OWNED;
}
err = ctrl->port_xfer_status(ctrl, pn, done_buf, done_len);
if (err == SLIM_P_INPROGRESS)
err = ctrl->ports[pn].err;
return err;
}
EXPORT_SYMBOL_GPL(slim_port_get_xfer_status);
static void slim_add_ch(struct slim_controller *ctrl, struct slim_ich *slc)
{
struct slim_ich **arr;
int i, j;
int *len;
int sl = slc->seglen << slc->rootexp;
/* Channel is already active and other end is transmitting data */
if (slc->state >= SLIM_CH_ACTIVE)
return;
if (slc->coeff == SLIM_COEFF_1) {
arr = ctrl->sched.chc1;
len = &ctrl->sched.num_cc1;
} else {
arr = ctrl->sched.chc3;
len = &ctrl->sched.num_cc3;
sl *= 3;
}
*len += 1;
/* Insert the channel based on rootexp and seglen */
for (i = 0; i < *len - 1; i++) {
/*
* Primary key: exp low to high.
* Secondary key: seglen: high to low
*/
if ((slc->rootexp > arr[i]->rootexp) ||
((slc->rootexp == arr[i]->rootexp) &&
(slc->seglen < arr[i]->seglen)))
continue;
else
break;
}
for (j = *len - 1; j > i; j--)
arr[j] = arr[j - 1];
arr[i] = slc;
if (!ctrl->allocbw)
ctrl->sched.usedslots += sl;
return;
}
static int slim_remove_ch(struct slim_controller *ctrl, struct slim_ich *slc)
{
struct slim_ich **arr;
int i;
u32 la, ph;
int *len;
if (slc->coeff == SLIM_COEFF_1) {
arr = ctrl->sched.chc1;
len = &ctrl->sched.num_cc1;
} else {
arr = ctrl->sched.chc3;
len = &ctrl->sched.num_cc3;
}
for (i = 0; i < *len; i++) {
if (arr[i] == slc)
break;
}
if (i >= *len)
return -EXFULL;
for (; i < *len - 1; i++)
arr[i] = arr[i + 1];
*len -= 1;
arr[*len] = NULL;
slc->state = SLIM_CH_ALLOCATED;
slc->newintr = 0;
slc->newoff = 0;
for (i = 0; i < slc->nsink; i++) {
ph = slc->sinkh[i];
la = SLIM_HDL_TO_LA(ph);
/*
* For ports managed by manager's ported device, no need to send
* disconnect. It is client's responsibility to call disconnect
* on ports owned by the slave device
*/
if (la == SLIM_LA_MANAGER) {
ctrl->ports[SLIM_HDL_TO_PORT(ph)].state = SLIM_P_UNCFG;
ctrl->ports[SLIM_HDL_TO_PORT(ph)].ch = NULL;
}
}
ph = slc->srch;
la = SLIM_HDL_TO_LA(ph);
if (la == SLIM_LA_MANAGER) {
u8 pn = SLIM_HDL_TO_PORT(ph);
ctrl->ports[pn].state = SLIM_P_UNCFG;
ctrl->ports[pn].cfg.watermark = 0;
ctrl->ports[pn].cfg.port_opts = 0;
}
kfree(slc->sinkh);
slc->sinkh = NULL;
slc->srch = 0;
slc->nsink = 0;
return 0;
}
static u32 slim_calc_prrate(struct slim_controller *ctrl, struct slim_ch *prop)
{
u32 rate = 0, rate4k = 0, rate11k = 0;
u32 exp = 0;
u32 pr = 0;
bool exact = true;
bool done = false;
enum slim_ch_rate ratefam;
if (prop->prot >= SLIM_ASYNC_SMPLX)
return 0;
if (prop->baser == SLIM_RATE_1HZ) {
rate = prop->ratem / 4000;
rate4k = rate;
if (rate * 4000 == prop->ratem)
ratefam = SLIM_RATE_4000HZ;
else {
rate = prop->ratem / 11025;
rate11k = rate;
if (rate * 11025 == prop->ratem)
ratefam = SLIM_RATE_11025HZ;
else
ratefam = SLIM_RATE_1HZ;
}
} else {
ratefam = prop->baser;
rate = prop->ratem;
}
if (ratefam == SLIM_RATE_1HZ) {
exact = false;
if ((rate4k + 1) * 4000 < (rate11k + 1) * 11025) {
rate = rate4k + 1;
ratefam = SLIM_RATE_4000HZ;
} else {
rate = rate11k + 1;
ratefam = SLIM_RATE_11025HZ;
}
}
/* covert rate to coeff-exp */
while (!done) {
while ((rate & 0x1) != 0x1) {
rate >>= 1;
exp++;
}
if (rate > 3) {
/* roundup if not exact */
rate++;
exact = false;
} else
done = true;
}
if (ratefam == SLIM_RATE_4000HZ) {
if (rate == 1)
pr = 0x10;
else {
pr = 0;
exp++;
}
} else {
pr = 8;
exp++;
}
if (exp <= 7) {
pr |= exp;
if (exact)
pr |= 0x80;
} else
pr = 0;
return pr;
}
static int slim_nextdefine_ch(struct slim_device *sb, u8 chan)
{
struct slim_controller *ctrl = sb->ctrl;
u32 chrate = 0;
u32 exp = 0;
u32 coeff = 0;
bool exact = true;
bool done = false;
int ret = 0;
struct slim_ich *slc = &ctrl->chans[chan];
struct slim_ch *prop = &slc->prop;
slc->prrate = slim_calc_prrate(ctrl, prop);
dev_dbg(&ctrl->dev, "ch:%d, chan PR rate:%x\n", chan, slc->prrate);
if (prop->baser == SLIM_RATE_4000HZ)
chrate = 4000 * prop->ratem;
else if (prop->baser == SLIM_RATE_11025HZ)
chrate = 11025 * prop->ratem;
else
chrate = prop->ratem;
/* max allowed sample freq = 768 seg/frame */
if (chrate > 3600000)
return -EDQUOT;
if (prop->baser == SLIM_RATE_4000HZ &&
ctrl->a_framer->superfreq == 4000)
coeff = prop->ratem;
else if (prop->baser == SLIM_RATE_11025HZ &&
ctrl->a_framer->superfreq == 3675)
coeff = 3 * prop->ratem;
else {
u32 tempr = 0;
tempr = chrate * SLIM_CL_PER_SUPERFRAME_DIV8;
coeff = tempr / ctrl->a_framer->rootfreq;
if (coeff * ctrl->a_framer->rootfreq != tempr) {
coeff++;
exact = false;
}
}
/* convert coeff to coeff-exponent */
exp = 0;
while (!done) {
while ((coeff & 0x1) != 0x1) {
coeff >>= 1;
exp++;
}
if (coeff > 3) {
coeff++;
exact = false;
} else
done = true;
}
if (prop->prot == SLIM_HARD_ISO && !exact)
return -EPROTONOSUPPORT;
else if (prop->prot == SLIM_AUTO_ISO) {
if (exact)
prop->prot = SLIM_HARD_ISO;
else
prop->prot = SLIM_PUSH;
}
slc->rootexp = exp;
slc->seglen = prop->sampleszbits/SLIM_CL_PER_SL;
if (prop->prot != SLIM_HARD_ISO)
slc->seglen++;
if (prop->prot >= SLIM_EXT_SMPLX)
slc->seglen++;
/* convert coeff to enum */
if (coeff == 1) {
if (exp > 9)
ret = -EIO;
coeff = SLIM_COEFF_1;
} else {
if (exp > 8)
ret = -EIO;
coeff = SLIM_COEFF_3;
}
slc->coeff = coeff;
return ret;
}
/*
* slim_alloc_ch: Allocate a slimbus channel and return its handle.
* @sb: client handle.
* @chanh: return channel handle
* Slimbus channels are limited to 256 per specification.
* -EXFULL is returned if all channels are in use.
* Although slimbus specification supports 256 channels, a controller may not
* support that many channels.
*/
int slim_alloc_ch(struct slim_device *sb, u16 *chanh)
{
struct slim_controller *ctrl = sb->ctrl;
u16 i;
if (!ctrl)
return -EINVAL;
mutex_lock(&ctrl->sched.m_reconf);
for (i = 0; i < ctrl->nchans; i++) {
if (ctrl->chans[i].state == SLIM_CH_FREE)
break;
}
if (i >= ctrl->nchans) {
mutex_unlock(&ctrl->sched.m_reconf);
return -EXFULL;
}
*chanh = i;
ctrl->chans[i].nextgrp = 0;
ctrl->chans[i].state = SLIM_CH_ALLOCATED;
ctrl->chans[i].chan = (u8)(ctrl->reserved + i);
mutex_unlock(&ctrl->sched.m_reconf);
return 0;
}
EXPORT_SYMBOL_GPL(slim_alloc_ch);
/*
* slim_query_ch: Get reference-counted handle for a channel number. Every
* channel is reference counted by upto one as producer and the others as
* consumer)
* @sb: client handle
* @chan: slimbus channel number
* @chanh: return channel handle
* If request channel number is not in use, it is allocated, and reference
* count is set to one. If the channel was was already allocated, this API
* will return handle to that channel and reference count is incremented.
* -EXFULL is returned if all channels are in use
*/
int slim_query_ch(struct slim_device *sb, u8 ch, u16 *chanh)
{
struct slim_controller *ctrl = sb->ctrl;
u16 i, j;
int ret = 0;
if (!ctrl || !chanh)
return -EINVAL;
mutex_lock(&ctrl->sched.m_reconf);
/* start with modulo number */
i = ch % ctrl->nchans;
for (j = 0; j < ctrl->nchans; j++) {
if (ctrl->chans[i].chan == ch) {
*chanh = i;
ctrl->chans[i].ref++;
if (ctrl->chans[i].state == SLIM_CH_FREE)
ctrl->chans[i].state = SLIM_CH_ALLOCATED;
goto query_out;
}
i = (i + 1) % ctrl->nchans;
}
/* Channel not in table yet */
ret = -EXFULL;
for (j = 0; j < ctrl->nchans; j++) {
if (ctrl->chans[i].state == SLIM_CH_FREE) {
ctrl->chans[i].state =
SLIM_CH_ALLOCATED;
*chanh = i;
ctrl->chans[i].ref++;
ctrl->chans[i].chan = ch;
ctrl->chans[i].nextgrp = 0;
ret = 0;
break;
}
i = (i + 1) % ctrl->nchans;
}
query_out:
mutex_unlock(&ctrl->sched.m_reconf);
dev_dbg(&ctrl->dev, "query ch:%d,hdl:%d,ref:%d,ret:%d",
ch, i, ctrl->chans[i].ref, ret);
return ret;
}
EXPORT_SYMBOL_GPL(slim_query_ch);
/*
* slim_dealloc_ch: Deallocate channel allocated using the API above
* -EISCONN is returned if the channel is tried to be deallocated without
* being removed first.
* -ENOTCONN is returned if deallocation is tried on a channel that's not
* allocated.
*/
int slim_dealloc_ch(struct slim_device *sb, u16 chanh)
{
struct slim_controller *ctrl = sb->ctrl;
u8 chan = SLIM_HDL_TO_CHIDX(chanh);
struct slim_ich *slc = &ctrl->chans[chan];
if (!ctrl)
return -EINVAL;
mutex_lock(&ctrl->sched.m_reconf);
if (slc->state == SLIM_CH_FREE) {
mutex_unlock(&ctrl->sched.m_reconf);
return -ENOTCONN;
}
if (slc->ref > 1) {
slc->ref--;
mutex_unlock(&ctrl->sched.m_reconf);
dev_dbg(&ctrl->dev, "remove chan:%d,hdl:%d,ref:%d",
slc->chan, chanh, slc->ref);
return 0;
}
if (slc->state >= SLIM_CH_PENDING_ACTIVE) {
dev_err(&ctrl->dev, "Channel:%d should be removed first", chan);
mutex_unlock(&ctrl->sched.m_reconf);
return -EISCONN;
}
slc->ref--;
slc->state = SLIM_CH_FREE;
mutex_unlock(&ctrl->sched.m_reconf);
dev_dbg(&ctrl->dev, "remove chan:%d,hdl:%d,ref:%d",
slc->chan, chanh, slc->ref);
return 0;
}
EXPORT_SYMBOL_GPL(slim_dealloc_ch);
/*
* slim_get_ch_state: Channel state.
* This API returns the channel's state (active, suspended, inactive etc)
*/
enum slim_ch_state slim_get_ch_state(struct slim_device *sb, u16 chanh)
{
u8 chan = SLIM_HDL_TO_CHIDX(chanh);
struct slim_ich *slc = &sb->ctrl->chans[chan];
return slc->state;
}
EXPORT_SYMBOL_GPL(slim_get_ch_state);
/*
* slim_define_ch: Define a channel.This API defines channel parameters for a
* given channel.
* @sb: client handle.
* @prop: slim_ch structure with channel parameters desired to be used.
* @chanh: list of channels to be defined.
* @nchan: number of channels in a group (1 if grp is false)
* @grp: Are the channels grouped
* @grph: return group handle if grouping of channels is desired.
* Channels can be grouped if multiple channels use same parameters
* (e.g. 5.1 audio has 6 channels with same parameters. They will all be grouped
* and given 1 handle for simplicity and avoid repeatedly calling the API)
* -EISCONN is returned if channel is already used with different parameters.
* -ENXIO is returned if the channel is not yet allocated.
*/
int slim_define_ch(struct slim_device *sb, struct slim_ch *prop, u16 *chanh,
u8 nchan, bool grp, u16 *grph)
{
struct slim_controller *ctrl = sb->ctrl;
int i, ret = 0;
if (!ctrl || !chanh || !prop || !nchan)
return -EINVAL;
mutex_lock(&ctrl->sched.m_reconf);
for (i = 0; i < nchan; i++) {
u8 chan = SLIM_HDL_TO_CHIDX(chanh[i]);
struct slim_ich *slc = &ctrl->chans[chan];
dev_dbg(&ctrl->dev, "define_ch: ch:%d, state:%d", chan,
(int)ctrl->chans[chan].state);
if (slc->state < SLIM_CH_ALLOCATED) {
ret = -ENXIO;
goto err_define_ch;
}
if (slc->state >= SLIM_CH_DEFINED && slc->ref >= 2) {
if (prop->ratem != slc->prop.ratem ||
prop->sampleszbits != slc->prop.sampleszbits ||
prop->baser != slc->prop.baser) {
ret = -EISCONN;
goto err_define_ch;
}
} else if (slc->state > SLIM_CH_DEFINED) {
ret = -EISCONN;
goto err_define_ch;
} else {
ctrl->chans[chan].prop = *prop;
ret = slim_nextdefine_ch(sb, chan);
if (ret)
goto err_define_ch;
}
if (i < (nchan - 1))
ctrl->chans[chan].nextgrp = chanh[i + 1];
if (i == 0)
ctrl->chans[chan].nextgrp |= SLIM_START_GRP;
if (i == (nchan - 1))
ctrl->chans[chan].nextgrp |= SLIM_END_GRP;
}
if (grp)
*grph = ((nchan << 8) | SLIM_HDL_TO_CHIDX(chanh[0]));
for (i = 0; i < nchan; i++) {
u8 chan = SLIM_HDL_TO_CHIDX(chanh[i]);
struct slim_ich *slc = &ctrl->chans[chan];
if (slc->state == SLIM_CH_ALLOCATED)
slc->state = SLIM_CH_DEFINED;
}
err_define_ch:
dev_dbg(&ctrl->dev, "define_ch: ch:%d, ret:%d", *chanh, ret);
mutex_unlock(&ctrl->sched.m_reconf);
return ret;
}
EXPORT_SYMBOL_GPL(slim_define_ch);
static u32 getsubfrmcoding(u32 *ctrlw, u32 *subfrml, u32 *msgsl)
{
u32 code = 0;
if (*ctrlw == *subfrml) {
*ctrlw = 8;
*subfrml = 8;
*msgsl = SLIM_SL_PER_SUPERFRAME - SLIM_FRM_SLOTS_PER_SUPERFRAME
- SLIM_GDE_SLOTS_PER_SUPERFRAME;
return 0;
}
if (*subfrml == 6) {
code = 0;
*msgsl = 256;
} else if (*subfrml == 8) {
code = 1;
*msgsl = 192;
} else if (*subfrml == 24) {
code = 2;
*msgsl = 64;
} else { /* 32 */
code = 3;
*msgsl = 48;
}
if (*ctrlw < 8) {
if (*ctrlw >= 6) {
*ctrlw = 6;
code |= 0x14;
} else {
if (*ctrlw == 5)
*ctrlw = 4;
code |= (*ctrlw << 2);
}
} else {
code -= 2;
if (*ctrlw >= 24) {
*ctrlw = 24;
code |= 0x1e;
} else if (*ctrlw >= 16) {
*ctrlw = 16;
code |= 0x1c;
} else if (*ctrlw >= 12) {
*ctrlw = 12;
code |= 0x1a;
} else {
*ctrlw = 8;
code |= 0x18;
}
}
*msgsl = (*msgsl * *ctrlw) - SLIM_FRM_SLOTS_PER_SUPERFRAME -
SLIM_GDE_SLOTS_PER_SUPERFRAME;
return code;
}
static void shiftsegoffsets(struct slim_controller *ctrl, struct slim_ich **ach,
int sz, u32 shft)
{
int i;
u32 oldoff;
for (i = 0; i < sz; i++) {
struct slim_ich *slc;
if (ach[i] == NULL)
continue;
slc = ach[i];
if (slc->state == SLIM_CH_PENDING_REMOVAL)
continue;
oldoff = slc->newoff;
slc->newoff += shft;
/* seg. offset must be <= interval */
if (slc->newoff >= slc->newintr)
slc->newoff -= slc->newintr;
}
}
static int slim_sched_chans(struct slim_device *sb, u32 clkgear,
u32 *ctrlw, u32 *subfrml)
{
int coeff1, coeff3;
enum slim_ch_coeff bias;
struct slim_controller *ctrl = sb->ctrl;
int last1 = ctrl->sched.num_cc1 - 1;
int last3 = ctrl->sched.num_cc3 - 1;
/*
* Find first channels with coeff 1 & 3 as starting points for
* scheduling
*/
for (coeff3 = 0; coeff3 < ctrl->sched.num_cc3; coeff3++) {
struct slim_ich *slc = ctrl->sched.chc3[coeff3];
if (slc->state == SLIM_CH_PENDING_REMOVAL)
continue;
else
break;
}
for (coeff1 = 0; coeff1 < ctrl->sched.num_cc1; coeff1++) {
struct slim_ich *slc = ctrl->sched.chc1[coeff1];
if (slc->state == SLIM_CH_PENDING_REMOVAL)
continue;
else
break;
}
if (coeff3 == ctrl->sched.num_cc3 && coeff1 == ctrl->sched.num_cc1) {
*ctrlw = 8;
*subfrml = 8;
return 0;
} else if (coeff3 == ctrl->sched.num_cc3)
bias = SLIM_COEFF_1;
else
bias = SLIM_COEFF_3;
/*
* Find last chan in coeff1, 3 list, we will use to know when we
* have done scheduling all coeff1 channels
*/
while (last1 >= 0) {
if (ctrl->sched.chc1[last1] != NULL &&
(ctrl->sched.chc1[last1])->state !=
SLIM_CH_PENDING_REMOVAL)
break;
last1--;
}
while (last3 >= 0) {
if (ctrl->sched.chc3[last3] != NULL &&
(ctrl->sched.chc3[last3])->state !=
SLIM_CH_PENDING_REMOVAL)
break;
last3--;
}
if (bias == SLIM_COEFF_1) {
struct slim_ich *slc1 = ctrl->sched.chc1[coeff1];
u32 expshft = SLIM_MAX_CLK_GEAR - clkgear;
int curexp, finalexp;
u32 curintr, curmaxsl;
int opensl1[2];
int maxctrlw1;
finalexp = (ctrl->sched.chc1[last1])->rootexp;
curexp = (int)expshft - 1;
curintr = (SLIM_MAX_INTR_COEFF_1 * 2) >> (curexp + 1);
curmaxsl = curintr >> 1;
opensl1[0] = opensl1[1] = curmaxsl;
while ((coeff1 < ctrl->sched.num_cc1) || (curintr > 24)) {
curintr >>= 1;
curmaxsl >>= 1;
/* update 4K family open slot records */
if (opensl1[1] < opensl1[0])
opensl1[1] -= curmaxsl;
else
opensl1[1] = opensl1[0] - curmaxsl;
opensl1[0] = curmaxsl;
if (opensl1[1] < 0) {
opensl1[0] += opensl1[1];
opensl1[1] = 0;
}
if (opensl1[0] <= 0) {
dev_dbg(&ctrl->dev, "reconfig failed:%d\n",
__LINE__);
return -EXFULL;
}
curexp++;
/* schedule 4k family channels */
while ((coeff1 < ctrl->sched.num_cc1) && (curexp ==
(int)(slc1->rootexp + expshft))) {
if (slc1->state == SLIM_CH_PENDING_REMOVAL) {
coeff1++;
slc1 = ctrl->sched.chc1[coeff1];
continue;
}
if (opensl1[1] >= opensl1[0] ||
(finalexp == (int)slc1->rootexp &&
curintr <= 24 &&
opensl1[0] == curmaxsl)) {
opensl1[1] -= slc1->seglen;
slc1->newoff = curmaxsl + opensl1[1];
if (opensl1[1] < 0 &&
opensl1[0] == curmaxsl) {
opensl1[0] += opensl1[1];
opensl1[1] = 0;
if (opensl1[0] < 0) {
dev_dbg(&ctrl->dev,
"reconfig failed:%d\n",
__LINE__);
return -EXFULL;
}
}
} else {
if (slc1->seglen > opensl1[0]) {
dev_dbg(&ctrl->dev,
"reconfig failed:%d\n",
__LINE__);
return -EXFULL;
}
slc1->newoff = opensl1[0] -
slc1->seglen;
opensl1[0] = slc1->newoff;
}
slc1->newintr = curintr;
coeff1++;
slc1 = ctrl->sched.chc1[coeff1];
}
}
/* Leave some slots for messaging space */
if (opensl1[1] <= 0 && opensl1[0] <= 0)
return -EXFULL;
if (opensl1[1] > opensl1[0]) {
int temp = opensl1[0];
opensl1[0] = opensl1[1];
opensl1[1] = temp;
shiftsegoffsets(ctrl, ctrl->sched.chc1,
ctrl->sched.num_cc1, curmaxsl);
}
/* choose subframe mode to maximize bw */
maxctrlw1 = opensl1[0];
if (opensl1[0] == curmaxsl)
maxctrlw1 += opensl1[1];
if (curintr >= 24) {
*subfrml = 24;
*ctrlw = maxctrlw1;
} else if (curintr == 12) {
if (maxctrlw1 > opensl1[1] * 4) {
*subfrml = 24;
*ctrlw = maxctrlw1;
} else {
*subfrml = 6;
*ctrlw = opensl1[1];
}
} else {
*subfrml = 6;
*ctrlw = maxctrlw1;
}
} else {
struct slim_ich *slc1 = NULL;
struct slim_ich *slc3 = ctrl->sched.chc3[coeff3];
u32 expshft = SLIM_MAX_CLK_GEAR - clkgear;
int curexp, finalexp, exp1;
u32 curintr, curmaxsl;
int opensl3[2];
int opensl1[6];
bool opensl1valid = false;
int maxctrlw1, maxctrlw3, i;
finalexp = (ctrl->sched.chc3[last3])->rootexp;
if (last1 >= 0) {
slc1 = ctrl->sched.chc1[coeff1];
exp1 = (ctrl->sched.chc1[last1])->rootexp;
if (exp1 > finalexp)
finalexp = exp1;
}
curexp = (int)expshft - 1;
curintr = (SLIM_MAX_INTR_COEFF_3 * 2) >> (curexp + 1);
curmaxsl = curintr >> 1;
opensl3[0] = opensl3[1] = curmaxsl;
while (coeff1 < ctrl->sched.num_cc1 ||
coeff3 < ctrl->sched.num_cc3 ||
curintr > 32) {
curintr >>= 1;
curmaxsl >>= 1;
/* update 12k family open slot records */
if (opensl3[1] < opensl3[0])
opensl3[1] -= curmaxsl;
else
opensl3[1] = opensl3[0] - curmaxsl;
opensl3[0] = curmaxsl;
if (opensl3[1] < 0) {
opensl3[0] += opensl3[1];
opensl3[1] = 0;
}
if (opensl3[0] <= 0) {
dev_dbg(&ctrl->dev, "reconfig failed:%d\n",
__LINE__);
return -EXFULL;
}
curexp++;
/* schedule 12k family channels */
while (coeff3 < ctrl->sched.num_cc3 &&
curexp == (int)slc3->rootexp + expshft) {
if (slc3->state == SLIM_CH_PENDING_REMOVAL) {
coeff3++;
slc3 = ctrl->sched.chc3[coeff3];
continue;
}
opensl1valid = false;
if (opensl3[1] >= opensl3[0] ||
(finalexp == (int)slc3->rootexp &&
curintr <= 32 &&
opensl3[0] == curmaxsl &&
last1 < 0)) {
opensl3[1] -= slc3->seglen;
slc3->newoff = curmaxsl + opensl3[1];
if (opensl3[1] < 0 &&
opensl3[0] == curmaxsl) {
opensl3[0] += opensl3[1];
opensl3[1] = 0;
}
if (opensl3[0] < 0) {
dev_dbg(&ctrl->dev,
"reconfig failed:%d\n",
__LINE__);
return -EXFULL;
}
} else {
if (slc3->seglen > opensl3[0]) {
dev_dbg(&ctrl->dev,
"reconfig failed:%d\n",
__LINE__);
return -EXFULL;
}
slc3->newoff = opensl3[0] -
slc3->seglen;
opensl3[0] = slc3->newoff;
}
slc3->newintr = curintr;
coeff3++;
slc3 = ctrl->sched.chc3[coeff3];
}
/* update 4k openslot records */
if (opensl1valid == false) {
for (i = 0; i < 3; i++) {
opensl1[i * 2] = opensl3[0];
opensl1[(i * 2) + 1] = opensl3[1];
}
} else {
int opensl1p[6];
memcpy(opensl1p, opensl1, sizeof(opensl1));
for (i = 0; i < 3; i++) {
if (opensl1p[i] < opensl1p[i + 3])
opensl1[(i * 2) + 1] =
opensl1p[i];
else
opensl1[(i * 2) + 1] =
opensl1p[i + 3];
}
for (i = 0; i < 3; i++) {
opensl1[(i * 2) + 1] -= curmaxsl;
opensl1[i * 2] = curmaxsl;
if (opensl1[(i * 2) + 1] < 0) {
opensl1[i * 2] +=
opensl1[(i * 2) + 1];
opensl1[(i * 2) + 1] = 0;
}
if (opensl1[i * 2] < 0) {
dev_dbg(&ctrl->dev,
"reconfig failed:%d\n",
__LINE__);
return -EXFULL;
}
}
}
/* schedule 4k family channels */
while (coeff1 < ctrl->sched.num_cc1 &&
curexp == (int)slc1->rootexp + expshft) {
/* searchorder effective when opensl valid */
static const int srcho[] = { 5, 2, 4, 1, 3, 0 };
int maxopensl = 0;
int maxi = 0;
if (slc1->state == SLIM_CH_PENDING_REMOVAL) {
coeff1++;
slc1 = ctrl->sched.chc1[coeff1];
continue;
}
opensl1valid = true;
for (i = 0; i < 6; i++) {
if (opensl1[srcho[i]] > maxopensl) {
maxopensl = opensl1[srcho[i]];
maxi = srcho[i];
}
}
opensl1[maxi] -= slc1->seglen;
slc1->newoff = (curmaxsl * maxi) +
opensl1[maxi];
if (opensl1[maxi] < 0) {
if (((maxi & 1) == 1) &&
(opensl1[maxi - 1] == curmaxsl)) {
opensl1[maxi - 1] +=
opensl1[maxi];
if (opensl3[0] >
opensl1[maxi - 1])
opensl3[0] =
opensl1[maxi - 1];
opensl3[1] = 0;
opensl1[maxi] = 0;
if (opensl1[maxi - 1] < 0) {
dev_dbg(&ctrl->dev,
"reconfig failed:%d\n",
__LINE__);
return -EXFULL;
}
} else {
dev_dbg(&ctrl->dev,
"reconfig failed:%d\n",
__LINE__);
return -EXFULL;
}
} else {
if (opensl3[maxi & 1] > opensl1[maxi])
opensl3[maxi & 1] =
opensl1[maxi];
}
slc1->newintr = curintr * 3;
coeff1++;
slc1 = ctrl->sched.chc1[coeff1];
}
}
/* Leave some slots for messaging space */
if (opensl3[1] <= 0 && opensl3[0] <= 0)
return -EXFULL;
/* swap 1st and 2nd bucket if 2nd bucket has more open slots */
if (opensl3[1] > opensl3[0]) {
int temp = opensl3[0];
opensl3[0] = opensl3[1];
opensl3[1] = temp;
temp = opensl1[5];
opensl1[5] = opensl1[4];
opensl1[4] = opensl1[3];
opensl1[3] = opensl1[2];
opensl1[2] = opensl1[1];
opensl1[1] = opensl1[0];
opensl1[0] = temp;
shiftsegoffsets(ctrl, ctrl->sched.chc1,
ctrl->sched.num_cc1, curmaxsl);
shiftsegoffsets(ctrl, ctrl->sched.chc3,
ctrl->sched.num_cc3, curmaxsl);
}
/* subframe mode to maximize BW */
maxctrlw3 = opensl3[0];
maxctrlw1 = opensl1[0];
if (opensl3[0] == curmaxsl)
maxctrlw3 += opensl3[1];
for (i = 0; i < 5 && opensl1[i] == curmaxsl; i++)
maxctrlw1 += opensl1[i + 1];
if (curintr >= 32) {
*subfrml = 32;
*ctrlw = maxctrlw3;
} else if (curintr == 16) {
if (maxctrlw3 > (opensl3[1] * 4)) {
*subfrml = 32;
*ctrlw = maxctrlw3;
} else {
*subfrml = 8;
*ctrlw = opensl3[1];
}
} else {
if ((maxctrlw1 * 8) >= (maxctrlw3 * 24)) {
*subfrml = 24;
*ctrlw = maxctrlw1;
} else {
*subfrml = 8;
*ctrlw = maxctrlw3;
}
}
}
return 0;
}
#ifdef DEBUG
static int slim_verifychansched(struct slim_controller *ctrl, u32 ctrlw,
u32 subfrml, u32 clkgear)
{
int sl, i;
int cc1 = 0;
int cc3 = 0;
struct slim_ich *slc = NULL;
if (!ctrl->sched.slots)
return 0;
memset(ctrl->sched.slots, 0, SLIM_SL_PER_SUPERFRAME);
dev_dbg(&ctrl->dev, "Clock gear is:%d\n", clkgear);
for (sl = 0; sl < SLIM_SL_PER_SUPERFRAME; sl += subfrml) {
for (i = 0; i < ctrlw; i++)
ctrl->sched.slots[sl + i] = 33;
}
while (cc1 < ctrl->sched.num_cc1) {
slc = ctrl->sched.chc1[cc1];
if (slc == NULL) {
dev_err(&ctrl->dev, "SLC1 null in verify: chan%d\n",
cc1);
return -EIO;
}
dev_dbg(&ctrl->dev, "chan:%d, offset:%d, intr:%d, seglen:%d\n",
(slc - ctrl->chans), slc->newoff,
slc->newintr, slc->seglen);
if (slc->state != SLIM_CH_PENDING_REMOVAL) {
for (sl = slc->newoff;
sl < SLIM_SL_PER_SUPERFRAME;
sl += slc->newintr) {
for (i = 0; i < slc->seglen; i++) {
if (ctrl->sched.slots[sl + i])
return -EXFULL;
ctrl->sched.slots[sl + i] = cc1 + 1;
}
}
}
cc1++;
}
while (cc3 < ctrl->sched.num_cc3) {
slc = ctrl->sched.chc3[cc3];
if (slc == NULL) {
dev_err(&ctrl->dev, "SLC3 null in verify: chan%d\n",
cc3);
return -EIO;
}
dev_dbg(&ctrl->dev, "chan:%d, offset:%d, intr:%d, seglen:%d\n",
(slc - ctrl->chans), slc->newoff,
slc->newintr, slc->seglen);
if (slc->state != SLIM_CH_PENDING_REMOVAL) {
for (sl = slc->newoff;
sl < SLIM_SL_PER_SUPERFRAME;
sl += slc->newintr) {
for (i = 0; i < slc->seglen; i++) {
if (ctrl->sched.slots[sl + i])
return -EXFULL;
ctrl->sched.slots[sl + i] = cc3 + 1;
}
}
}
cc3++;
}
return 0;
}
#else
static int slim_verifychansched(struct slim_controller *ctrl, u32 ctrlw,
u32 subfrml, u32 clkgear)
{
return 0;
}
#endif
static void slim_sort_chan_grp(struct slim_controller *ctrl,
struct slim_ich *slc)
{
u8 last = (u8)-1;
u8 second = 0;
for (; last > 0; last--) {
struct slim_ich *slc1 = slc;
struct slim_ich *slc2;
u8 next = SLIM_HDL_TO_CHIDX(slc1->nextgrp);
slc2 = &ctrl->chans[next];
for (second = 1; second <= last && slc2 &&
(slc2->state == SLIM_CH_ACTIVE ||
slc2->state == SLIM_CH_PENDING_ACTIVE); second++) {
if (slc1->newoff > slc2->newoff) {
u32 temp = slc2->newoff;
slc2->newoff = slc1->newoff;
slc1->newoff = temp;
}
if (slc2->nextgrp & SLIM_END_GRP) {
last = second;
break;
}
slc1 = slc2;
next = SLIM_HDL_TO_CHIDX(slc1->nextgrp);
slc2 = &ctrl->chans[next];
}
if (slc2 == NULL)
last = second - 1;
}
}
static int slim_allocbw(struct slim_device *sb, int *subfrmc, int *clkgear)
{
u32 msgsl = 0;
u32 ctrlw = 0;
u32 subfrml = 0;
int ret = -EIO;
struct slim_controller *ctrl = sb->ctrl;
u32 usedsl = ctrl->sched.usedslots + ctrl->sched.pending_msgsl;
u32 availsl = SLIM_SL_PER_SUPERFRAME - SLIM_FRM_SLOTS_PER_SUPERFRAME -
SLIM_GDE_SLOTS_PER_SUPERFRAME;
*clkgear = SLIM_MAX_CLK_GEAR;
dev_dbg(&ctrl->dev, "used sl:%u, availlable sl:%u\n", usedsl, availsl);
dev_dbg(&ctrl->dev, "pending:chan sl:%u, :msg sl:%u, clkgear:%u\n",
ctrl->sched.usedslots,
ctrl->sched.pending_msgsl, *clkgear);
/*
* If number of slots are 0, that means channels are inactive.
* It is very likely that the manager will call clock pause very soon.
* By making sure that bus is in MAX_GEAR, clk pause sequence will take
* minimum amount of time.
*/
if (ctrl->sched.usedslots != 0) {
while ((usedsl * 2 <= availsl) && (*clkgear > ctrl->min_cg)) {
*clkgear -= 1;
usedsl *= 2;
}
}
/*
* Try scheduling data channels at current clock gear, if all channels
* can be scheduled, or reserved BW can't be satisfied, increase clock
* gear and try again
*/
for (; *clkgear <= ctrl->max_cg; (*clkgear)++) {
ret = slim_sched_chans(sb, *clkgear, &ctrlw, &subfrml);
if (ret == 0) {
*subfrmc = getsubfrmcoding(&ctrlw, &subfrml, &msgsl);
if ((msgsl >> (ctrl->max_cg - *clkgear) <
ctrl->sched.pending_msgsl) &&
(*clkgear < ctrl->max_cg))
continue;
else
break;
}
}
if (ret == 0) {
int i;
/* Sort channel-groups */
for (i = 0; i < ctrl->sched.num_cc1; i++) {
struct slim_ich *slc = ctrl->sched.chc1[i];
if (slc->state == SLIM_CH_PENDING_REMOVAL)
continue;
if ((slc->nextgrp & SLIM_START_GRP) &&
!(slc->nextgrp & SLIM_END_GRP)) {
slim_sort_chan_grp(ctrl, slc);
}
}
for (i = 0; i < ctrl->sched.num_cc3; i++) {
struct slim_ich *slc = ctrl->sched.chc3[i];
if (slc->state == SLIM_CH_PENDING_REMOVAL)
continue;
if ((slc->nextgrp & SLIM_START_GRP) &&
!(slc->nextgrp & SLIM_END_GRP)) {
slim_sort_chan_grp(ctrl, slc);
}
}
ret = slim_verifychansched(ctrl, ctrlw, subfrml, *clkgear);
}
return ret;
}
static void slim_change_existing_chans(struct slim_controller *ctrl, int coeff)
{
struct slim_ich **arr;
int len, i;
if (coeff == SLIM_COEFF_1) {
arr = ctrl->sched.chc1;
len = ctrl->sched.num_cc1;
} else {
arr = ctrl->sched.chc3;
len = ctrl->sched.num_cc3;
}
for (i = 0; i < len; i++) {
struct slim_ich *slc = arr[i];
if (slc->state == SLIM_CH_ACTIVE ||
slc->state == SLIM_CH_SUSPENDED)
slc->offset = slc->newoff;
slc->interval = slc->newintr;
}
}
static void slim_chan_changes(struct slim_device *sb, bool revert)
{
struct slim_controller *ctrl = sb->ctrl;
while (!list_empty(&sb->mark_define)) {
struct slim_ich *slc;
struct slim_pending_ch *pch =
list_entry(sb->mark_define.next,
struct slim_pending_ch, pending);
slc = &ctrl->chans[pch->chan];
if (revert) {
if (slc->state == SLIM_CH_PENDING_ACTIVE) {
u32 sl = slc->seglen << slc->rootexp;
if (slc->coeff == SLIM_COEFF_3)
sl *= 3;
if (!ctrl->allocbw)
ctrl->sched.usedslots -= sl;
slim_remove_ch(ctrl, slc);
slc->state = SLIM_CH_DEFINED;
}
} else {
slc->state = SLIM_CH_ACTIVE;
slc->def++;
}
list_del_init(&pch->pending);
kfree(pch);
}
while (!list_empty(&sb->mark_removal)) {
struct slim_pending_ch *pch =
list_entry(sb->mark_removal.next,
struct slim_pending_ch, pending);
struct slim_ich *slc = &ctrl->chans[pch->chan];
u32 sl = slc->seglen << slc->rootexp;
if (revert || slc->def > 0) {
if (slc->coeff == SLIM_COEFF_3)
sl *= 3;
if (!ctrl->allocbw)
ctrl->sched.usedslots += sl;
if (revert)
slc->def++;
slc->state = SLIM_CH_ACTIVE;
} else
slim_remove_ch(ctrl, slc);
list_del_init(&pch->pending);
kfree(pch);
}
while (!list_empty(&sb->mark_suspend)) {
struct slim_pending_ch *pch =
list_entry(sb->mark_suspend.next,
struct slim_pending_ch, pending);
struct slim_ich *slc = &ctrl->chans[pch->chan];
if (revert)
slc->state = SLIM_CH_ACTIVE;
list_del_init(&pch->pending);
kfree(pch);
}
/* Change already active channel if reconfig succeeded */
if (!revert) {
slim_change_existing_chans(ctrl, SLIM_COEFF_1);
slim_change_existing_chans(ctrl, SLIM_COEFF_3);
}
}
/*
* slim_reconfigure_now: Request reconfiguration now.
* @sb: client handle
* This API does what commit flag in other scheduling APIs do.
* -EXFULL is returned if there is no space in TDM to reserve the
* bandwidth. -EBUSY is returned if reconfiguration request is already in
* progress.
*/
int slim_reconfigure_now(struct slim_device *sb)
{
u8 i;
u8 wbuf[4];
u32 clkgear, subframe;
u32 curexp;
int ret;
struct slim_controller *ctrl = sb->ctrl;
u32 expshft;
u32 segdist;
struct slim_pending_ch *pch;
DEFINE_SLIM_BCAST_TXN(txn, SLIM_MSG_MC_BEGIN_RECONFIGURATION, 0, 3,
NULL, NULL, sb->laddr);
mutex_lock(&ctrl->sched.m_reconf);
/*
* If there are no pending changes from this client, avoid sending
* the reconfiguration sequence
*/
if (sb->pending_msgsl == sb->cur_msgsl &&
list_empty(&sb->mark_define) &&
list_empty(&sb->mark_suspend)) {
struct list_head *pos, *next;
list_for_each_safe(pos, next, &sb->mark_removal) {
struct slim_ich *slc;
pch = list_entry(pos, struct slim_pending_ch, pending);
slc = &ctrl->chans[pch->chan];
if (slc->def > 0)
slc->def--;
/* Disconnect source port to free it up */
if (SLIM_HDL_TO_LA(slc->srch) == sb->laddr)
slc->srch = 0;
/*
* If controller overrides BW allocation,
* delete this in remove channel itself
*/
if (slc->def != 0 && !ctrl->allocbw) {
list_del(&pch->pending);
kfree(pch);
}
}
if (list_empty(&sb->mark_removal)) {
mutex_unlock(&ctrl->sched.m_reconf);
pr_info("SLIM_CL: skip reconfig sequence");
return 0;
}
}
ctrl->sched.pending_msgsl += sb->pending_msgsl - sb->cur_msgsl;
list_for_each_entry(pch, &sb->mark_define, pending) {
struct slim_ich *slc = &ctrl->chans[pch->chan];
slim_add_ch(ctrl, slc);
if (slc->state < SLIM_CH_ACTIVE)
slc->state = SLIM_CH_PENDING_ACTIVE;
}
list_for_each_entry(pch, &sb->mark_removal, pending) {
struct slim_ich *slc = &ctrl->chans[pch->chan];
u32 sl = slc->seglen << slc->rootexp;
if (slc->coeff == SLIM_COEFF_3)
sl *= 3;
if (!ctrl->allocbw)
ctrl->sched.usedslots -= sl;
slc->state = SLIM_CH_PENDING_REMOVAL;
}
list_for_each_entry(pch, &sb->mark_suspend, pending) {
struct slim_ich *slc = &ctrl->chans[pch->chan];
slc->state = SLIM_CH_SUSPENDED;
}
/*
* Controller can override default channel scheduling algorithm.
* (e.g. if controller needs to use fixed channel scheduling based
* on number of channels)
*/
if (ctrl->allocbw)
ret = ctrl->allocbw(sb, &subframe, &clkgear);
else
ret = slim_allocbw(sb, &subframe, &clkgear);
if (!ret) {
ret = slim_processtxn(ctrl, &txn, false);
dev_dbg(&ctrl->dev, "sending begin_reconfig:ret:%d\n", ret);
}
if (!ret && subframe != ctrl->sched.subfrmcode) {
wbuf[0] = (u8)(subframe & 0xFF);
txn.mc = SLIM_MSG_MC_NEXT_SUBFRAME_MODE;
txn.len = 1;
txn.rl = 4;
txn.wbuf = wbuf;
ret = slim_processtxn(ctrl, &txn, false);
dev_dbg(&ctrl->dev, "sending subframe:%d,ret:%d\n",
(int)wbuf[0], ret);
}
if (!ret && clkgear != ctrl->clkgear) {
wbuf[0] = (u8)(clkgear & 0xFF);
txn.mc = SLIM_MSG_MC_NEXT_CLOCK_GEAR;
txn.len = 1;
txn.rl = 4;
txn.wbuf = wbuf;
ret = slim_processtxn(ctrl, &txn, false);
dev_dbg(&ctrl->dev, "sending clkgear:%d,ret:%d\n",
(int)wbuf[0], ret);
}
if (ret)
goto revert_reconfig;
expshft = SLIM_MAX_CLK_GEAR - clkgear;
/* activate/remove channel */
list_for_each_entry(pch, &sb->mark_define, pending) {
struct slim_ich *slc = &ctrl->chans[pch->chan];
/* Define content */
wbuf[0] = slc->chan;
wbuf[1] = slc->prrate;
wbuf[2] = slc->prop.dataf | (slc->prop.auxf << 4);
wbuf[3] = slc->prop.sampleszbits / SLIM_CL_PER_SL;
txn.mc = SLIM_MSG_MC_NEXT_DEFINE_CONTENT;
txn.len = 4;
txn.rl = 7;
txn.wbuf = wbuf;
dev_dbg(&ctrl->dev, "define content, activate:%x, %x, %x, %x\n",
wbuf[0], wbuf[1], wbuf[2], wbuf[3]);
/* Right now, channel link bit is not supported */
ret = slim_processtxn(ctrl, &txn, false);
if (ret)
goto revert_reconfig;
txn.mc = SLIM_MSG_MC_NEXT_ACTIVATE_CHANNEL;
txn.len = 1;
txn.rl = 4;
ret = slim_processtxn(ctrl, &txn, false);
if (ret)
goto revert_reconfig;
}
list_for_each_entry(pch, &sb->mark_removal, pending) {
struct slim_ich *slc = &ctrl->chans[pch->chan];
dev_dbg(&ctrl->dev, "remove chan:%x\n", pch->chan);
wbuf[0] = slc->chan;
txn.mc = SLIM_MSG_MC_NEXT_REMOVE_CHANNEL;
txn.len = 1;
txn.rl = 4;
txn.wbuf = wbuf;
ret = slim_processtxn(ctrl, &txn, false);
if (ret)
goto revert_reconfig;
}
list_for_each_entry(pch, &sb->mark_suspend, pending) {
struct slim_ich *slc = &ctrl->chans[pch->chan];
dev_dbg(&ctrl->dev, "suspend chan:%x\n", pch->chan);
wbuf[0] = slc->chan;
txn.mc = SLIM_MSG_MC_NEXT_DEACTIVATE_CHANNEL;
txn.len = 1;
txn.rl = 4;
txn.wbuf = wbuf;
ret = slim_processtxn(ctrl, &txn, false);
if (ret)
goto revert_reconfig;
}
/* Define CC1 channel */
for (i = 0; i < ctrl->sched.num_cc1; i++) {
struct slim_ich *slc = ctrl->sched.chc1[i];
if (slc->state == SLIM_CH_PENDING_REMOVAL)
continue;
curexp = slc->rootexp + expshft;
segdist = (slc->newoff << curexp) & 0x1FF;
expshft = SLIM_MAX_CLK_GEAR - clkgear;
dev_dbg(&ctrl->dev, "new-intr:%d, old-intr:%d, dist:%d\n",
slc->newintr, slc->interval, segdist);
dev_dbg(&ctrl->dev, "new-off:%d, old-off:%d\n",
slc->newoff, slc->offset);
if (slc->state < SLIM_CH_ACTIVE || slc->def < slc->ref ||
slc->newintr != slc->interval ||
slc->newoff != slc->offset) {
segdist |= 0x200;
segdist >>= curexp;
segdist |= (slc->newoff << (curexp + 1)) & 0xC00;
wbuf[0] = slc->chan;
wbuf[1] = (u8)(segdist & 0xFF);
wbuf[2] = (u8)((segdist & 0xF00) >> 8) |
(slc->prop.prot << 4);
wbuf[3] = slc->seglen;
txn.mc = SLIM_MSG_MC_NEXT_DEFINE_CHANNEL;
txn.len = 4;
txn.rl = 7;
txn.wbuf = wbuf;
ret = slim_processtxn(ctrl, &txn, false);
if (ret)
goto revert_reconfig;
}
}
/* Define CC3 channels */
for (i = 0; i < ctrl->sched.num_cc3; i++) {
struct slim_ich *slc = ctrl->sched.chc3[i];
if (slc->state == SLIM_CH_PENDING_REMOVAL)
continue;
curexp = slc->rootexp + expshft;
segdist = (slc->newoff << curexp) & 0x1FF;
expshft = SLIM_MAX_CLK_GEAR - clkgear;
dev_dbg(&ctrl->dev, "new-intr:%d, old-intr:%d, dist:%d\n",
slc->newintr, slc->interval, segdist);
dev_dbg(&ctrl->dev, "new-off:%d, old-off:%d\n",
slc->newoff, slc->offset);
if (slc->state < SLIM_CH_ACTIVE || slc->def < slc->ref ||
slc->newintr != slc->interval ||
slc->newoff != slc->offset) {
segdist |= 0x200;
segdist >>= curexp;
segdist |= 0xC00;
wbuf[0] = slc->chan;
wbuf[1] = (u8)(segdist & 0xFF);
wbuf[2] = (u8)((segdist & 0xF00) >> 8) |
(slc->prop.prot << 4);
wbuf[3] = (u8)(slc->seglen);
txn.mc = SLIM_MSG_MC_NEXT_DEFINE_CHANNEL;
txn.len = 4;
txn.rl = 7;
txn.wbuf = wbuf;
ret = slim_processtxn(ctrl, &txn, false);
if (ret)
goto revert_reconfig;
}
}
txn.mc = SLIM_MSG_MC_RECONFIGURE_NOW;
txn.len = 0;
txn.rl = 3;
txn.wbuf = NULL;
ret = slim_processtxn(ctrl, &txn, false);
dev_dbg(&ctrl->dev, "reconfig now:ret:%d\n", ret);
if (!ret) {
ctrl->sched.subfrmcode = subframe;
ctrl->clkgear = clkgear;
ctrl->sched.msgsl = ctrl->sched.pending_msgsl;
sb->cur_msgsl = sb->pending_msgsl;
slim_chan_changes(sb, false);
mutex_unlock(&ctrl->sched.m_reconf);
return 0;
}
revert_reconfig:
/* Revert channel changes */
slim_chan_changes(sb, true);
mutex_unlock(&ctrl->sched.m_reconf);
return ret;
}
EXPORT_SYMBOL_GPL(slim_reconfigure_now);
static int add_pending_ch(struct list_head *listh, u8 chan)
{
struct slim_pending_ch *pch;
pch = kmalloc(sizeof(struct slim_pending_ch), GFP_KERNEL);
if (!pch)
return -ENOMEM;
pch->chan = chan;
list_add_tail(&pch->pending, listh);
return 0;
}
/*
* slim_control_ch: Channel control API.
* @sb: client handle
* @chanh: group or channel handle to be controlled
* @chctrl: Control command (activate/suspend/remove)
* @commit: flag to indicate whether the control should take effect right-away.
* This API activates, removes or suspends a channel (or group of channels)
* chanh indicates the channel or group handle (returned by the define_ch API).
* Reconfiguration may be time-consuming since it can change all other active
* channel allocations on the bus, change in clock gear used by the slimbus,
* and change in the control space width used for messaging.
* commit makes sure that multiple channels can be activated/deactivated before
* reconfiguration is started.
* -EXFULL is returned if there is no space in TDM to reserve the bandwidth.
* -EISCONN/-ENOTCONN is returned if the channel is already connected or not
* yet defined.
* -EINVAL is returned if individual control of a grouped-channel is attempted.
*/
int slim_control_ch(struct slim_device *sb, u16 chanh,
enum slim_ch_control chctrl, bool commit)
{
struct slim_controller *ctrl = sb->ctrl;
int ret = 0;
/* Get rid of the group flag in MSB if any */
u8 chan = SLIM_HDL_TO_CHIDX(chanh);
u8 nchan = 0;
struct slim_ich *slc = &ctrl->chans[chan];
if (!(slc->nextgrp & SLIM_START_GRP))
return -EINVAL;
mutex_lock(&sb->sldev_reconf);
do {
struct slim_pending_ch *pch;
u8 add_mark_removal = true;
slc = &ctrl->chans[chan];
dev_dbg(&ctrl->dev, "chan:%d,ctrl:%d,def:%d", chan, chctrl,
slc->def);
if (slc->state < SLIM_CH_DEFINED) {
ret = -ENOTCONN;
break;
}
if (chctrl == SLIM_CH_SUSPEND) {
ret = add_pending_ch(&sb->mark_suspend, chan);
if (ret)
break;
} else if (chctrl == SLIM_CH_ACTIVATE) {
if (slc->state > SLIM_CH_ACTIVE) {
ret = -EISCONN;
break;
}
ret = add_pending_ch(&sb->mark_define, chan);
if (ret)
break;
} else {
if (slc->state < SLIM_CH_ACTIVE) {
ret = -ENOTCONN;
break;
}
/* If channel removal request comes when pending
* in the mark_define, remove it from the define
* list instead of adding it to removal list
*/
if (!list_empty(&sb->mark_define)) {
struct list_head *pos, *next;
list_for_each_safe(pos, next,
&sb->mark_define) {
pch = list_entry(pos,
struct slim_pending_ch,
pending);
if (pch->chan == chan) {
list_del(&pch->pending);
kfree(pch);
add_mark_removal = false;
break;
}
}
}
if (add_mark_removal == true) {
ret = add_pending_ch(&sb->mark_removal, chan);
if (ret)
break;
}
}
nchan++;
if (nchan < SLIM_GRP_TO_NCHAN(chanh))
chan = SLIM_HDL_TO_CHIDX(slc->nextgrp);
} while (nchan < SLIM_GRP_TO_NCHAN(chanh));
if (!ret && commit == true)
ret = slim_reconfigure_now(sb);
mutex_unlock(&sb->sldev_reconf);
return ret;
}
EXPORT_SYMBOL_GPL(slim_control_ch);
/*
* slim_reservemsg_bw: Request to reserve bandwidth for messages.
* @sb: client handle
* @bw_bps: message bandwidth in bits per second to be requested
* @commit: indicates whether the reconfiguration needs to be acted upon.
* This API call can be grouped with slim_control_ch API call with only one of
* the APIs specifying the commit flag to avoid reconfiguration being called too
* frequently. -EXFULL is returned if there is no space in TDM to reserve the
* bandwidth. -EBUSY is returned if reconfiguration is requested, but a request
* is already in progress.
*/
int slim_reservemsg_bw(struct slim_device *sb, u32 bw_bps, bool commit)
{
struct slim_controller *ctrl = sb->ctrl;
int ret = 0;
int sl;
mutex_lock(&sb->sldev_reconf);
if ((bw_bps >> 3) >= ctrl->a_framer->rootfreq)
sl = SLIM_SL_PER_SUPERFRAME;
else {
sl = (bw_bps * (SLIM_CL_PER_SUPERFRAME_DIV8/SLIM_CL_PER_SL/2) +
(ctrl->a_framer->rootfreq/2 - 1)) /
(ctrl->a_framer->rootfreq/2);
}
dev_dbg(&ctrl->dev, "request:bw:%d, slots:%d, current:%d\n", bw_bps, sl,
sb->cur_msgsl);
sb->pending_msgsl = sl;
if (commit == true)
ret = slim_reconfigure_now(sb);
mutex_unlock(&sb->sldev_reconf);
return ret;
}
EXPORT_SYMBOL_GPL(slim_reservemsg_bw);
/*
* slim_ctrl_clk_pause: Called by slimbus controller to request clock to be
* paused or woken up out of clock pause
* or woken up from clock pause
* @ctrl: controller requesting bus to be paused or woken up
* @wakeup: Wakeup this controller from clock pause.
* @restart: Restart time value per spec used for clock pause. This value
* isn't used when controller is to be woken up.
* This API executes clock pause reconfiguration sequence if wakeup is false.
* If wakeup is true, controller's wakeup is called
* Slimbus clock is idle and can be disabled by the controller later.
*/
int slim_ctrl_clk_pause(struct slim_controller *ctrl, bool wakeup, u8 restart)
{
int ret = 0;
int i;
DEFINE_SLIM_BCAST_TXN(txn, SLIM_MSG_CLK_PAUSE_SEQ_FLG |
SLIM_MSG_MC_BEGIN_RECONFIGURATION, 0, 3,
NULL, NULL, 0);
if (wakeup == false && restart > SLIM_CLK_UNSPECIFIED)
return -EINVAL;
mutex_lock(&ctrl->m_ctrl);
if (wakeup) {
if (ctrl->clk_state == SLIM_CLK_ACTIVE) {
mutex_unlock(&ctrl->m_ctrl);
return 0;
}
wait_for_completion(&ctrl->pause_comp);
/*
* Slimbus framework will call controller wakeup
* Controller should make sure that it sets active framer
* out of clock pause by doing appropriate setting
*/
if (ctrl->clk_state == SLIM_CLK_PAUSED && ctrl->wakeup)
ret = ctrl->wakeup(ctrl);
/*
* If wakeup fails, make sure that next attempt can succeed.
* Since we already consumed pause_comp, complete it so
* that next wakeup isn't blocked forever
*/
if (!ret)
ctrl->clk_state = SLIM_CLK_ACTIVE;
else
complete(&ctrl->pause_comp);
mutex_unlock(&ctrl->m_ctrl);
return ret;
} else {
switch (ctrl->clk_state) {
case SLIM_CLK_ENTERING_PAUSE:
case SLIM_CLK_PAUSE_FAILED:
/*
* If controller is already trying to enter clock pause,
* let it finish.
* In case of error, retry
* In both cases, previous clock pause has signalled
* completion.
*/
wait_for_completion(&ctrl->pause_comp);
/* retry upon failure */
if (ctrl->clk_state == SLIM_CLK_PAUSE_FAILED) {
ctrl->clk_state = SLIM_CLK_ACTIVE;
break;
} else {
mutex_unlock(&ctrl->m_ctrl);
/*
* Signal completion so that wakeup can wait on
* it.
*/
complete(&ctrl->pause_comp);
return 0;
}
break;
case SLIM_CLK_PAUSED:
/* already paused */
mutex_unlock(&ctrl->m_ctrl);
return 0;
case SLIM_CLK_ACTIVE:
default:
break;
}
}
/* Pending response for a message */
for (i = 0; i < ctrl->last_tid; i++) {
if (ctrl->txnt[i]) {
ret = -EBUSY;
pr_info("slim_clk_pause: txn-rsp for %d pending", i);
mutex_unlock(&ctrl->m_ctrl);
return -EBUSY;
}
}
ctrl->clk_state = SLIM_CLK_ENTERING_PAUSE;
mutex_unlock(&ctrl->m_ctrl);
mutex_lock(&ctrl->sched.m_reconf);
/* Data channels active */
if (ctrl->sched.usedslots) {
pr_info("slim_clk_pause: data channel active");
ret = -EBUSY;
goto clk_pause_ret;
}
ret = slim_processtxn(ctrl, &txn, false);
if (ret)
goto clk_pause_ret;
txn.mc = SLIM_MSG_CLK_PAUSE_SEQ_FLG | SLIM_MSG_MC_NEXT_PAUSE_CLOCK;
txn.len = 1;
txn.rl = 4;
txn.wbuf = &restart;
ret = slim_processtxn(ctrl, &txn, false);
if (ret)
goto clk_pause_ret;
txn.mc = SLIM_MSG_CLK_PAUSE_SEQ_FLG | SLIM_MSG_MC_RECONFIGURE_NOW;
txn.len = 0;
txn.rl = 3;
txn.wbuf = NULL;
ret = slim_processtxn(ctrl, &txn, false);
if (ret)
goto clk_pause_ret;
clk_pause_ret:
if (ret)
ctrl->clk_state = SLIM_CLK_PAUSE_FAILED;
else
ctrl->clk_state = SLIM_CLK_PAUSED;
complete(&ctrl->pause_comp);
mutex_unlock(&ctrl->sched.m_reconf);
return ret;
}
EXPORT_SYMBOL_GPL(slim_ctrl_clk_pause);
MODULE_LICENSE("GPL v2");
MODULE_VERSION("0.1");
MODULE_DESCRIPTION("Slimbus module");
MODULE_ALIAS("platform:slimbus");