M7350/kernel/drivers/media/platform/msm/vcap/vcap_vp.c

898 lines
25 KiB
C
Raw Normal View History

2024-09-09 08:52:07 +00:00
/* Copyright (c) 2012-2013, The Linux Foundation. All rights reserved.
*
* This program is free software; you can redistribute it and/or modify
* it under the terms of the GNU General Public License version 2 and
* only version 2 as published by the Free Software Foundation.
*
* This program is distributed in the hope that it will be useful,
* but WITHOUT ANY WARRANTY; without even the implied warranty of
* MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
* GNU General Public License for more details.
*
*/
#include <linux/kernel.h>
#include <linux/init.h>
#include <linux/sched.h>
#include <linux/kthread.h>
#include <linux/freezer.h>
#include <linux/clk.h>
#include <linux/io.h>
#include <mach/camera.h>
#include <mach/clk.h>
#include <media/v4l2-event.h>
#include <media/vcap_v4l2.h>
#include <media/vcap_fmt.h>
#include "vcap_vp.h"
void config_nr_buffer(struct vcap_client_data *c_data,
struct vcap_buffer *buf)
{
struct vcap_dev *dev = c_data->dev;
int size = c_data->vp_in_fmt.height * c_data->vp_in_fmt.width;
writel_relaxed(buf->paddr, VCAP_VP_NR_T2_Y_BASE_ADDR);
writel_relaxed(buf->paddr + size, VCAP_VP_NR_T2_C_BASE_ADDR);
}
void config_in_buffer(struct vcap_client_data *c_data,
struct vcap_buffer *buf)
{
struct vcap_dev *dev = c_data->dev;
int size = c_data->vp_in_fmt.height * c_data->vp_in_fmt.width;
writel_relaxed(buf->paddr, VCAP_VP_T2_Y_BASE_ADDR);
writel_relaxed(buf->paddr + size, VCAP_VP_T2_C_BASE_ADDR);
}
void config_out_buffer(struct vcap_client_data *c_data,
struct vcap_buffer *buf)
{
struct vcap_dev *dev = c_data->dev;
int size;
size = c_data->vp_out_fmt.height * c_data->vp_out_fmt.width;
writel_relaxed(buf->paddr, VCAP_VP_OUT_Y_BASE_ADDR);
writel_relaxed(buf->paddr + size, VCAP_VP_OUT_C_BASE_ADDR);
}
int vp_setup_buffers(struct vcap_client_data *c_data)
{
struct vp_action *vp_act;
struct vcap_dev *dev;
unsigned long flags = 0;
if (!c_data->streaming)
return -ENOEXEC;
dev = c_data->dev;
pr_debug("VP: Start setup buffers\n");
if (dev->vp_shutdown) {
pr_debug("%s: VP shutting down, no buf setup\n",
__func__);
return -EPERM;
}
/* No need to verify vp_client is not NULL caller does so */
vp_act = &dev->vp_client->vp_action;
spin_lock_irqsave(&dev->vp_client->cap_slock, flags);
if (list_empty(&vp_act->in_active)) {
spin_unlock_irqrestore(&dev->vp_client->cap_slock, flags);
pr_debug("%s: VP We have no more input buffers\n",
__func__);
return -EAGAIN;
}
if (list_empty(&vp_act->out_active)) {
spin_unlock_irqrestore(&dev->vp_client->cap_slock,
flags);
pr_debug("%s: VP We have no more output buffers\n",
__func__);
return -EAGAIN;
}
vp_act->bufT2 = list_entry(vp_act->in_active.next,
struct vcap_buffer, list);
list_del(&vp_act->bufT2->list);
vp_act->bufOut = list_entry(vp_act->out_active.next,
struct vcap_buffer, list);
list_del(&vp_act->bufOut->list);
spin_unlock_irqrestore(&dev->vp_client->cap_slock, flags);
config_in_buffer(c_data, vp_act->bufT2);
config_out_buffer(c_data, vp_act->bufOut);
return 0;
}
static void mov_buf_to_vc(struct work_struct *work)
{
struct vp_work_t *vp_work = container_of(work, struct vp_work_t, work);
struct v4l2_buffer p;
struct vb2_buffer *vb_vc;
struct vcap_buffer *buf_vc;
struct vb2_buffer *vb_vp;
struct vcap_buffer *buf_vp;
int rc;
p.memory = V4L2_MEMORY_USERPTR;
/* This loop exits when there is no more buffers left */
while (1) {
p.type = V4L2_BUF_TYPE_INTERLACED_IN_DECODER;
if (!vp_work->cd->streaming)
return;
rc = vcvp_dqbuf(&vp_work->cd->vp_in_vidq, &p);
if (rc < 0)
return;
vb_vc = vp_work->cd->vc_vidq.bufs[p.index];
if (NULL == vb_vc) {
pr_debug("%s: buffer is NULL\n", __func__);
vcvp_qbuf(&vp_work->cd->vp_in_vidq, &p);
return;
}
buf_vc = container_of(vb_vc, struct vcap_buffer, vb);
vb_vp = vp_work->cd->vp_in_vidq.bufs[p.index];
if (NULL == vb_vp) {
pr_debug("%s: buffer is NULL\n", __func__);
vcvp_qbuf(&vp_work->cd->vp_in_vidq, &p);
return;
}
buf_vp = container_of(vb_vp, struct vcap_buffer, vb);
buf_vc->ion_handle = buf_vp->ion_handle;
buf_vc->paddr = buf_vp->paddr;
buf_vp->ion_handle = NULL;
buf_vp->paddr = 0;
p.type = V4L2_BUF_TYPE_VIDEO_CAPTURE;
/* This call should not fail */
rc = vcvp_qbuf(&vp_work->cd->vc_vidq, &p);
if (rc < 0) {
pr_err("%s: qbuf to vc failed\n", __func__);
buf_vp->ion_handle = buf_vc->ion_handle;
buf_vp->paddr = buf_vc->paddr;
buf_vc->ion_handle = NULL;
buf_vc->paddr = 0;
p.type = V4L2_BUF_TYPE_INTERLACED_IN_DECODER;
vcvp_qbuf(&vp_work->cd->vp_in_vidq, &p);
}
}
}
void update_nr_value(struct vcap_dev *dev)
{
struct nr_param *par;
uint32_t val = 0;
par = &dev->nr_param;
if (par->mode == NR_MANUAL) {
writel_relaxed(par->window << 24 | par->decay_ratio << 20,
VCAP_VP_NR_CONFIG);
if (par->threshold)
val = VP_NR_DYNAMIC_THRESHOLD;
writel_relaxed(val |
par->luma.max_blend_ratio << 24 |
par->luma.scale_diff_ratio << 12 |
par->luma.diff_limit_ratio << 8 |
par->luma.scale_motion_ratio << 4 |
par->luma.blend_limit_ratio << 0,
VCAP_VP_NR_LUMA_CONFIG);
writel_relaxed(val |
par->chroma.max_blend_ratio << 24 |
par->chroma.scale_diff_ratio << 12 |
par->chroma.diff_limit_ratio << 8 |
par->chroma.scale_motion_ratio << 4 |
par->chroma.blend_limit_ratio << 0,
VCAP_VP_NR_CHROMA_CONFIG);
}
dev->nr_update = false;
}
static void vp_wq_fnc(struct work_struct *work)
{
struct vp_work_t *vp_work = container_of(work, struct vp_work_t, work);
struct vcap_dev *dev;
struct vp_action *vp_act;
struct timeval tv;
unsigned long flags = 0;
uint32_t irq;
int rc;
bool top_field = 0;
if (vp_work && vp_work->cd && vp_work->cd->dev)
dev = vp_work->cd->dev;
else
return;
vp_act = &dev->vp_client->vp_action;
rc = readl_relaxed(VCAP_OFFSET(0x048));
while (!(rc & 0x00000100))
rc = readl_relaxed(VCAP_OFFSET(0x048));
irq = readl_relaxed(VCAP_VP_INT_STATUS);
writel_relaxed(0x00000000, VCAP_VP_BAL_VMOTION_STATE);
writel_relaxed(0x40000000, VCAP_VP_REDUCT_AVG_MOTION2);
spin_lock_irqsave(&dev->vp_client->cap_slock, flags);
if (dev->nr_update == true)
update_nr_value(dev);
spin_unlock_irqrestore(&dev->vp_client->cap_slock, flags);
/* Queue the done buffers */
if (vp_act->vp_state == VP_NORMAL &&
vp_act->bufNR.nr_pos != TM1_BUF) {
vb2_buffer_done(&vp_act->bufTm1->vb, VB2_BUF_STATE_DONE);
if (vp_work->cd->op_mode == VC_AND_VP_VCAP_OP)
queue_work(dev->vcap_wq, &dev->vp_to_vc_work.work);
}
if (vp_act->bufT0 != NULL && vp_act->vp_state == VP_NORMAL) {
vp_act->bufOut->vb.v4l2_buf.timestamp =
vp_act->bufT0->vb.v4l2_buf.timestamp;
}
vb2_buffer_done(&vp_act->bufOut->vb, VB2_BUF_STATE_DONE);
/* Cycle to next state */
if (vp_act->vp_state != VP_NORMAL)
vp_act->vp_state++;
/* Cycle Buffers*/
if (dev->nr_param.mode) {
if (vp_act->bufNR.nr_pos == TM1_BUF)
vp_act->bufNR.nr_pos = BUF_NOT_IN_USE;
if (vp_act->bufNR.nr_pos != BUF_NOT_IN_USE)
vp_act->bufNR.nr_pos++;
vp_act->bufTm1 = vp_act->bufT0;
vp_act->bufT0 = vp_act->bufT1;
vp_act->bufT1 = vp_act->bufNRT2;
vp_act->bufNRT2 = vp_act->bufT2;
config_nr_buffer(vp_work->cd, vp_act->bufNRT2);
} else {
vp_act->bufTm1 = vp_act->bufT0;
vp_act->bufT0 = vp_act->bufT1;
vp_act->bufT1 = vp_act->bufT2;
}
rc = vp_setup_buffers(vp_work->cd);
if (rc < 0) {
/* setup_buf failed because we are waiting for buffers */
writel_relaxed(0x00000000, VCAP_VP_INTERRUPT_ENABLE);
writel_iowmb(irq, VCAP_VP_INT_CLEAR);
atomic_set(&dev->vp_enabled, 0);
if (dev->vp_shutdown)
wake_up(&dev->vp_dummy_waitq);
return;
}
/* Config VP */
if (vp_act->bufT2->vb.v4l2_buf.field == V4L2_FIELD_BOTTOM)
top_field = 1;
writel_iowmb(0x00000000 | top_field, VCAP_VP_CTRL);
writel_iowmb(0x00010000 | top_field, VCAP_VP_CTRL);
enable_irq(dev->vpirq->start);
do_gettimeofday(&tv);
dev->dbg_p.vp_timestamp = (uint32_t) (tv.tv_sec * VCAP_USEC +
tv.tv_usec);
writel_iowmb(irq, VCAP_VP_INT_CLEAR);
}
irqreturn_t vp_handler(struct vcap_dev *dev)
{
struct vcap_client_data *c_data;
struct vp_action *vp_act;
struct v4l2_event v4l2_evt;
uint32_t irq;
int rc;
struct timeval tv;
uint32_t new_ts;
irq = readl_relaxed(VCAP_VP_INT_STATUS);
if (dev->vp_dummy_event == true) {
writel_relaxed(irq, VCAP_VP_INT_CLEAR);
dev->vp_dummy_complete = true;
wake_up(&dev->vp_dummy_waitq);
return IRQ_HANDLED;
}
if (irq & 0x02000000) {
v4l2_evt.type = V4L2_EVENT_PRIVATE_START +
VCAP_VP_REG_R_ERR_EVENT;
v4l2_event_queue(dev->vfd, &v4l2_evt);
}
if (irq & 0x01000000) {
v4l2_evt.type = V4L2_EVENT_PRIVATE_START +
VCAP_VP_REG_W_ERR_EVENT;
v4l2_event_queue(dev->vfd, &v4l2_evt);
}
if (irq & 0x00020000) {
v4l2_evt.type = V4L2_EVENT_PRIVATE_START +
VCAP_VP_IN_HEIGHT_ERR_EVENT;
v4l2_event_queue(dev->vfd, &v4l2_evt);
}
if (irq & 0x00010000) {
v4l2_evt.type = V4L2_EVENT_PRIVATE_START +
VCAP_VP_IN_WIDTH_ERR_EVENT;
v4l2_event_queue(dev->vfd, &v4l2_evt);
}
pr_debug("%s: irq=0x%08x\n", __func__, irq);
if (!(irq & (VP_PIC_DONE | VP_MODE_CHANGE))) {
writel_relaxed(irq, VCAP_VP_INT_CLEAR);
pr_err("VP IRQ shows some error\n");
return IRQ_HANDLED;
}
if (dev->vp_client == NULL) {
writel_relaxed(irq, VCAP_VP_INT_CLEAR);
pr_err("VC: There is no active vp client\n");
return IRQ_HANDLED;
}
vp_act = &dev->vp_client->vp_action;
c_data = dev->vp_client;
if (vp_act->vp_state == VP_UNKNOWN) {
writel_relaxed(irq, VCAP_VP_INT_CLEAR);
pr_err("%s: VP is in an unknown state\n",
__func__);
return -EAGAIN;
}
do_gettimeofday(&tv);
new_ts = (uint32_t) (tv.tv_sec * VCAP_USEC +
tv.tv_usec);
if (new_ts > dev->dbg_p.vp_timestamp) {
dev->dbg_p.vp_ewma = ((new_ts - dev->dbg_p.vp_timestamp) /
10 + (dev->dbg_p.vp_ewma / 10 * 9));
}
dev->dbg_p.vp_timestamp = (uint32_t) (tv.tv_sec * VCAP_USEC +
tv.tv_usec);
INIT_WORK(&dev->vp_work.work, vp_wq_fnc);
dev->vp_work.cd = c_data;
rc = queue_work(dev->vcap_wq, &dev->vp_work.work);
disable_irq_nosync(dev->vpirq->start);
return IRQ_HANDLED;
}
int vp_sw_reset(struct vcap_dev *dev)
{
int timeout;
writel_iowmb(0x00000010, VCAP_SW_RESET_REQ);
timeout = 10000;
while (1) {
if (!(readl_relaxed(VCAP_SW_RESET_STATUS) & 0x10))
break;
timeout--;
if (timeout == 0) {
/* This should not happen */
pr_err("VP is not resetting properly\n");
writel_iowmb(0x00000000, VCAP_SW_RESET_REQ);
return -EINVAL;
}
}
return 0;
}
void vp_stop_capture(struct vcap_client_data *c_data)
{
struct vcap_dev *dev = c_data->dev;
int rc;
dev->vp_shutdown = true;
flush_workqueue(dev->vcap_wq);
if (atomic_read(&dev->vp_enabled) == 1) {
rc = wait_event_interruptible_timeout(dev->vp_dummy_waitq,
!atomic_read(&dev->vp_enabled),
msecs_to_jiffies(50));
if (rc == 0 && atomic_read(&dev->vp_enabled) == 1) {
/* This should not happen, if it does hw is stuck */
disable_irq_nosync(dev->vpirq->start);
atomic_set(&dev->vp_enabled, 0);
pr_err("%s: VP Timeout and VP still running\n",
__func__);
}
}
vp_sw_reset(dev);
dev->vp_shutdown = false;
}
int config_vp_format(struct vcap_client_data *c_data)
{
struct vcap_dev *dev = c_data->dev;
int rc;
INIT_WORK(&dev->vp_to_vc_work.work, mov_buf_to_vc);
dev->vp_to_vc_work.cd = c_data;
/* SW restart VP */
rc = vp_sw_reset(dev);
if (rc < 0)
return rc;
/* Film Mode related settings */
writel_iowmb(0x00000000, VCAP_VP_FILM_PROJECTION_T0);
writel_relaxed(0x00000000, VCAP_VP_FILM_PROJECTION_T2);
writel_relaxed(0x00000000, VCAP_VP_FILM_PAST_MAX_PROJ);
writel_relaxed(0x00000000, VCAP_VP_FILM_PAST_MIN_PROJ);
writel_relaxed(0x00000000, VCAP_VP_FILM_SEQUENCE_HIST);
writel_relaxed(0x00000000, VCAP_VP_FILM_MODE_STATE);
writel_relaxed(0x00000000, VCAP_VP_BAL_VMOTION_STATE);
writel_relaxed(0x00000010, VCAP_VP_REDUCT_AVG_MOTION);
writel_relaxed(0x40000000, VCAP_VP_REDUCT_AVG_MOTION2);
writel_relaxed(0x40000000, VCAP_VP_NR_AVG_LUMA);
writel_relaxed(0x40000000, VCAP_VP_NR_AVG_CHROMA);
writel_relaxed(0x40000000, VCAP_VP_NR_CTRL_LUMA);
writel_relaxed(0x40000000, VCAP_VP_NR_CTRL_CHROMA);
writel_relaxed(0x00000000, VCAP_VP_BAL_AVG_BLEND);
writel_relaxed(0x00000000, VCAP_VP_VMOTION_HIST);
writel_relaxed(0x05047D19, VCAP_VP_FILM_ANALYSIS_CONFIG);
writel_relaxed(0x20260200, VCAP_VP_FILM_STATE_CONFIG);
writel_relaxed(0x23A60114, VCAP_VP_FVM_CONFIG);
writel_relaxed(0x03043210, VCAP_VP_FILM_ANALYSIS_CONFIG2);
writel_relaxed(0x04DB7A51, VCAP_VP_MIXED_ANALYSIS_CONFIG);
writel_relaxed(0x14224916, VCAP_VP_SPATIAL_CONFIG);
writel_relaxed(0x83270400, VCAP_VP_SPATIAL_CONFIG2);
writel_relaxed(0x0F000F92, VCAP_VP_SPATIAL_CONFIG3);
writel_relaxed(0x00000000, VCAP_VP_TEMPORAL_CONFIG);
writel_relaxed(0x00000000, VCAP_VP_PIXEL_DIFF_CONFIG);
writel_relaxed(0x0C090511, VCAP_VP_H_FREQ_CONFIG);
writel_relaxed(0x0A000000, VCAP_VP_NR_CONFIG);
writel_relaxed(0x008F4149, VCAP_VP_NR_LUMA_CONFIG);
writel_relaxed(0x008F4149, VCAP_VP_NR_CHROMA_CONFIG);
writel_relaxed(0x43C0FD0C, VCAP_VP_BAL_CONFIG);
writel_relaxed(0x00000255, VCAP_VP_BAL_MOTION_CONFIG);
writel_relaxed(0x24154252, VCAP_VP_BAL_LIGHT_COMB);
writel_relaxed(0x10024414, VCAP_VP_BAL_VMOTION_CONFIG);
writel_relaxed(0x00000002, VCAP_VP_NR_CONFIG2);
writel_relaxed((c_data->vp_out_fmt.height-1)<<16 |
(c_data->vp_out_fmt.width - 1), VCAP_VP_FRAME_SIZE);
writel_relaxed(0x00000000, VCAP_VP_SPLIT_SCRN_CTRL);
return 0;
}
int init_motion_buf(struct vcap_client_data *c_data)
{
int rc;
struct vcap_dev *dev = c_data->dev;
struct ion_handle *handle = NULL;
unsigned long paddr, len;
void *vaddr;
size_t size = ((c_data->vp_out_fmt.width + 63) >> 6) *
((c_data->vp_out_fmt.height + 7) >> 3) * 16;
if (c_data->vp_action.motionHandle) {
pr_err("Motion buffer has already been created");
return -ENOEXEC;
}
handle = ion_alloc(dev->ion_client, size, SZ_4K,
ION_HEAP(ION_CP_MM_HEAP_ID), 0);
if (IS_ERR_OR_NULL(handle)) {
pr_err("%s: ion_alloc failed\n", __func__);
return -ENOMEM;
}
vaddr = ion_map_kernel(dev->ion_client, handle);
if (IS_ERR(vaddr)) {
pr_err("%s: Map motion buffer failed\n", __func__);
ion_free(dev->ion_client, handle);
rc = -ENOMEM;
return rc;
}
memset(vaddr, 0, size);
ion_unmap_kernel(dev->ion_client, handle);
rc = ion_map_iommu(dev->ion_client, handle,
dev->domain_num, 0, SZ_4K, 0, &paddr, &len,
0, 0);
if (rc < 0) {
pr_err("%s: map_iommu failed\n", __func__);
ion_free(dev->ion_client, handle);
return rc;
}
c_data->vp_action.motionHandle = handle;
vaddr = NULL;
writel_iowmb(paddr, VCAP_VP_MOTION_EST_ADDR);
return 0;
}
void deinit_motion_buf(struct vcap_client_data *c_data)
{
struct vcap_dev *dev = c_data->dev;
if (!c_data->vp_action.motionHandle) {
pr_err("Motion buffer has not been created");
return;
}
writel_iowmb(0x00000000, VCAP_VP_MOTION_EST_ADDR);
ion_unmap_iommu(dev->ion_client, c_data->vp_action.motionHandle,
dev->domain_num, 0);
ion_free(dev->ion_client, c_data->vp_action.motionHandle);
c_data->vp_action.motionHandle = NULL;
return;
}
int init_nr_buf(struct vcap_client_data *c_data)
{
struct vcap_dev *dev = c_data->dev;
struct ion_handle *handle = NULL;
size_t frame_size, tot_size;
unsigned long paddr, len;
int rc;
if (c_data->vp_action.bufNR.nr_handle) {
pr_err("NR buffer has already been created");
return -ENOEXEC;
}
frame_size = c_data->vp_in_fmt.width * c_data->vp_in_fmt.height;
if (c_data->vp_in_fmt.pixfmt == V4L2_PIX_FMT_NV16)
tot_size = frame_size * 2;
else
tot_size = frame_size / 2 * 3;
handle = ion_alloc(dev->ion_client, tot_size, SZ_4K,
ION_HEAP(ION_CP_MM_HEAP_ID), 0);
if (IS_ERR_OR_NULL(handle)) {
pr_err("%s: ion_alloc failed\n", __func__);
return -ENOMEM;
}
rc = ion_map_iommu(dev->ion_client, handle,
dev->domain_num, 0, SZ_4K, 0, &paddr, &len,
0, 0);
if (rc < 0) {
pr_err("%s: map_iommu failed\n", __func__);
ion_free(dev->ion_client, handle);
return rc;
}
c_data->vp_action.bufNR.nr_handle = handle;
update_nr_value(dev);
c_data->vp_action.bufNR.paddr = paddr;
rc = readl_relaxed(VCAP_VP_NR_CONFIG2);
rc |= (((c_data->vp_out_fmt.width / 16) << 20) | 0x1);
writel_relaxed(rc, VCAP_VP_NR_CONFIG2);
writel_relaxed(paddr, VCAP_VP_NR_T2_Y_BASE_ADDR);
writel_relaxed(paddr + frame_size, VCAP_VP_NR_T2_C_BASE_ADDR);
c_data->vp_action.bufNR.nr_pos = NRT2_BUF;
return 0;
}
void deinit_nr_buf(struct vcap_client_data *c_data)
{
struct vcap_dev *dev = c_data->dev;
struct nr_buffer *buf;
uint32_t rc;
if (!c_data->vp_action.bufNR.nr_handle) {
pr_err("NR buffer has not been created");
return;
}
buf = &c_data->vp_action.bufNR;
rc = readl_relaxed(VCAP_VP_NR_CONFIG2);
rc &= !(0x0FF00001);
writel_relaxed(rc, VCAP_VP_NR_CONFIG2);
ion_unmap_iommu(dev->ion_client, buf->nr_handle, dev->domain_num, 0);
ion_free(dev->ion_client, buf->nr_handle);
buf->nr_handle = NULL;
buf->paddr = 0;
return;
}
int nr_s_param(struct vcap_client_data *c_data, struct nr_param *param)
{
if (param->mode != NR_MANUAL)
return 0;
/* Verify values in range */
if (param->window > VP_NR_MAX_WINDOW)
return -EINVAL;
if (param->luma.max_blend_ratio > VP_NR_MAX_RATIO)
return -EINVAL;
if (param->luma.scale_diff_ratio > VP_NR_MAX_RATIO)
return -EINVAL;
if (param->luma.diff_limit_ratio > VP_NR_MAX_RATIO)
return -EINVAL;
if (param->luma.scale_motion_ratio > VP_NR_MAX_RATIO)
return -EINVAL;
if (param->luma.blend_limit_ratio > VP_NR_MAX_RATIO)
return -EINVAL;
if (param->chroma.max_blend_ratio > VP_NR_MAX_RATIO)
return -EINVAL;
if (param->chroma.scale_diff_ratio > VP_NR_MAX_RATIO)
return -EINVAL;
if (param->chroma.diff_limit_ratio > VP_NR_MAX_RATIO)
return -EINVAL;
if (param->chroma.scale_motion_ratio > VP_NR_MAX_RATIO)
return -EINVAL;
if (param->chroma.blend_limit_ratio > VP_NR_MAX_RATIO)
return -EINVAL;
return 0;
}
void nr_g_param(struct vcap_client_data *c_data, struct nr_param *param)
{
struct vcap_dev *dev = c_data->dev;
uint32_t rc;
rc = readl_relaxed(VCAP_VP_NR_CONFIG);
param->window = BITS_VALUE(rc, 24, 4);
param->decay_ratio = BITS_VALUE(rc, 20, 3);
rc = readl_relaxed(VCAP_VP_NR_LUMA_CONFIG);
param->threshold = NR_THRESHOLD_STATIC;
if (BITS_VALUE(rc, 16, 1))
param->threshold = NR_THRESHOLD_DYNAMIC;
param->luma.max_blend_ratio = BITS_VALUE(rc, 24, 4);
param->luma.scale_diff_ratio = BITS_VALUE(rc, 12, 4);
param->luma.diff_limit_ratio = BITS_VALUE(rc, 8, 4);
param->luma.scale_motion_ratio = BITS_VALUE(rc, 4, 4);
param->luma.blend_limit_ratio = BITS_VALUE(rc, 0, 4);
rc = readl_relaxed(VCAP_VP_NR_CHROMA_CONFIG);
param->chroma.max_blend_ratio = BITS_VALUE(rc, 24, 4);
param->chroma.scale_diff_ratio = BITS_VALUE(rc, 12, 4);
param->chroma.diff_limit_ratio = BITS_VALUE(rc, 8, 4);
param->chroma.scale_motion_ratio = BITS_VALUE(rc, 4, 4);
param->chroma.blend_limit_ratio = BITS_VALUE(rc, 0, 4);
}
void s_default_nr_val(struct nr_param *param)
{
param->threshold = NR_THRESHOLD_STATIC;
param->window = 10;
param->decay_ratio = 0;
param->luma.max_blend_ratio = 0;
param->luma.scale_diff_ratio = 4;
param->luma.diff_limit_ratio = 1;
param->luma.scale_motion_ratio = 4;
param->luma.blend_limit_ratio = 9;
param->chroma.max_blend_ratio = 0;
param->chroma.scale_diff_ratio = 4;
param->chroma.diff_limit_ratio = 1;
param->chroma.scale_motion_ratio = 4;
param->chroma.blend_limit_ratio = 9;
}
int vp_dummy_event(struct vcap_client_data *c_data)
{
struct vcap_dev *dev = c_data->dev;
unsigned int width, height;
struct ion_handle *handle = NULL;
unsigned long paddr, len;
uint32_t reg;
int rc = 0;
pr_debug("%s: Start VP dummy event\n", __func__);
handle = ion_alloc(dev->ion_client, 0x1200, SZ_4K,
ION_HEAP(ION_CP_MM_HEAP_ID), 0);
if (IS_ERR_OR_NULL(handle)) {
pr_err("%s: ion_alloc failed\n", __func__);
return -ENOMEM;
}
rc = ion_map_iommu(dev->ion_client, handle,
dev->domain_num, 0, SZ_4K, 0, &paddr, &len,
0, 0);
if (rc < 0) {
pr_err("%s: map_iommu failed\n", __func__);
ion_free(dev->ion_client, handle);
return rc;
}
width = c_data->vp_out_fmt.width;
height = c_data->vp_out_fmt.height;
c_data->vp_out_fmt.width = 0x3F;
c_data->vp_out_fmt.height = 0x16;
config_vp_format(c_data);
writel_relaxed(paddr, VCAP_VP_T1_Y_BASE_ADDR);
writel_relaxed(paddr + 0x2C0, VCAP_VP_T1_C_BASE_ADDR);
writel_relaxed(paddr + 0x440, VCAP_VP_T2_Y_BASE_ADDR);
writel_relaxed(paddr + 0x700, VCAP_VP_T2_C_BASE_ADDR);
writel_relaxed(paddr + 0x880, VCAP_VP_OUT_Y_BASE_ADDR);
writel_relaxed(paddr + 0xB40, VCAP_VP_OUT_C_BASE_ADDR);
writel_iowmb(paddr + 0x1100, VCAP_VP_MOTION_EST_ADDR);
writel_relaxed(4 << 20 | 0x2 << 4, VCAP_VP_IN_CONFIG);
writel_relaxed(4 << 20 | 0x1 << 4, VCAP_VP_OUT_CONFIG);
dev->vp_dummy_event = true;
enable_irq(dev->vpirq->start);
writel_relaxed(0x01100101, VCAP_VP_INTERRUPT_ENABLE);
writel_iowmb(0x00000000, VCAP_VP_CTRL);
writel_iowmb(0x00010000, VCAP_VP_CTRL);
rc = wait_event_interruptible_timeout(dev->vp_dummy_waitq,
dev->vp_dummy_complete, msecs_to_jiffies(50));
if (!rc && !dev->vp_dummy_complete) {
pr_err("%s: VP dummy event timeout", __func__);
rc = -ETIME;
vp_sw_reset(dev);
dev->vp_dummy_complete = false;
}
writel_relaxed(0x00000000, VCAP_VP_INTERRUPT_ENABLE);
disable_irq(dev->vpirq->start);
dev->vp_dummy_event = false;
reg = readl_relaxed(VCAP_OFFSET(0x0D94));
writel_relaxed(reg, VCAP_OFFSET(0x0D9C));
c_data->vp_out_fmt.width = width;
c_data->vp_out_fmt.height = height;
ion_unmap_iommu(dev->ion_client, handle, dev->domain_num, 0);
ion_free(dev->ion_client, handle);
pr_debug("%s: Exit VP dummy event\n", __func__);
return rc;
}
int kickoff_vp(struct vcap_client_data *c_data)
{
struct vcap_dev *dev;
struct vp_action *vp_act;
struct timeval tv;
unsigned long flags = 0;
unsigned int chroma_fmt = 0;
int size;
bool top_field = 0;
if (!c_data->streaming)
return -ENOEXEC;
dev = c_data->dev;
pr_debug("Start VP Kickoff\n");
if (dev->vp_client == NULL) {
pr_err("No active vp client\n");
return -ENODEV;
}
vp_act = &dev->vp_client->vp_action;
spin_lock_irqsave(&dev->vp_client->cap_slock, flags);
if (list_empty(&vp_act->in_active)) {
spin_unlock_irqrestore(&dev->vp_client->cap_slock, flags);
pr_err("%s: VP We have no more input buffers\n",
__func__);
return -EAGAIN;
}
vp_act->bufT1 = list_entry(vp_act->in_active.next,
struct vcap_buffer, list);
list_del(&vp_act->bufT1->list);
if (list_empty(&vp_act->in_active)) {
spin_unlock_irqrestore(&dev->vp_client->cap_slock, flags);
list_add(&vp_act->bufT1->list, &vp_act->in_active);
pr_err("%s: VP We have no more input buffers\n",
__func__);
return -EAGAIN;
}
vp_act->bufT2 = list_entry(vp_act->in_active.next,
struct vcap_buffer, list);
list_del(&vp_act->bufT2->list);
if (list_empty(&vp_act->out_active)) {
spin_unlock_irqrestore(&dev->vp_client->cap_slock, flags);
list_add(&vp_act->bufT2->list, &vp_act->in_active);
list_add(&vp_act->bufT1->list, &vp_act->in_active);
pr_err("%s: VP We have no more output buffers\n",
__func__);
return -EAGAIN;
}
vp_act->bufOut = list_entry(vp_act->out_active.next,
struct vcap_buffer, list);
list_del(&vp_act->bufOut->list);
spin_unlock_irqrestore(&dev->vp_client->cap_slock, flags);
size = c_data->vp_in_fmt.height * c_data->vp_in_fmt.width;
writel_relaxed(vp_act->bufT1->paddr, VCAP_VP_T1_Y_BASE_ADDR);
writel_relaxed(vp_act->bufT1->paddr + size, VCAP_VP_T1_C_BASE_ADDR);
config_in_buffer(c_data, vp_act->bufT2);
config_out_buffer(c_data, vp_act->bufOut);
/* Config VP */
if (c_data->vp_in_fmt.pixfmt == V4L2_PIX_FMT_NV16)
chroma_fmt = 1;
writel_relaxed((c_data->vp_in_fmt.width / 16) << 20 |
chroma_fmt << 11 | 0x2 << 4, VCAP_VP_IN_CONFIG);
chroma_fmt = 0;
if (c_data->vp_out_fmt.pixfmt == V4L2_PIX_FMT_NV16)
chroma_fmt = 1;
writel_relaxed((c_data->vp_out_fmt.width / 16) << 20 |
chroma_fmt << 11 | 0x1 << 4, VCAP_VP_OUT_CONFIG);
/* Enable Interrupt */
if (vp_act->bufT2->vb.v4l2_buf.field == V4L2_FIELD_BOTTOM)
top_field = 1;
vp_act->vp_state = VP_FRAME2;
writel_relaxed(0x01100001, VCAP_VP_INTERRUPT_ENABLE);
writel_iowmb(0x00000000 | top_field, VCAP_VP_CTRL);
writel_iowmb(0x00010000 | top_field, VCAP_VP_CTRL);
atomic_set(&c_data->dev->vp_enabled, 1);
enable_irq(dev->vpirq->start);
do_gettimeofday(&tv);
dev->dbg_p.vp_timestamp = (uint32_t) (tv.tv_sec * VCAP_USEC +
tv.tv_usec);
return 0;
}
int continue_vp(struct vcap_client_data *c_data)
{
struct vcap_dev *dev;
struct vp_action *vp_act;
struct timeval tv;
int rc;
bool top_field = 0;
pr_debug("Start VP Continue\n");
dev = c_data->dev;
if (dev->vp_client == NULL) {
pr_err("No active vp client\n");
return -ENODEV;
}
vp_act = &dev->vp_client->vp_action;
if (vp_act->vp_state == VP_UNKNOWN) {
pr_err("%s: VP is in an unknown state\n",
__func__);
return -EAGAIN;
}
rc = vp_setup_buffers(c_data);
if (rc < 0)
return rc;
if (vp_act->bufT2->vb.v4l2_buf.field == V4L2_FIELD_BOTTOM)
top_field = 1;
/* Config VP & Enable Interrupt */
writel_relaxed(0x01100001, VCAP_VP_INTERRUPT_ENABLE);
writel_iowmb(0x00000000 | top_field, VCAP_VP_CTRL);
writel_iowmb(0x00010000 | top_field, VCAP_VP_CTRL);
atomic_set(&c_data->dev->vp_enabled, 1);
enable_irq(dev->vpirq->start);
do_gettimeofday(&tv);
dev->dbg_p.vp_timestamp = (uint32_t) (tv.tv_sec * VCAP_USEC +
tv.tv_usec);
return 0;
}