Hi Jungo,

On Fri, May 10, 2019 at 09:58:04AM +0800, Jungo Lin wrote:
> This patch adds the Mediatek ISP P1 HW control device driver.
> It handles the ISP HW configuration, provides interrupt handling and
> initializes the V4L2 device nodes and other functions.

A few comments inline.

> 
> Signed-off-by: Jungo Lin <jungo....@mediatek.com>
> ---
>  .../mtk-isp/isp_50/cam/mtk_cam-regs.h         |  149 ++
>  .../platform/mtk-isp/isp_50/cam/mtk_cam.c     | 1206 +++++++++++++++++
>  .../platform/mtk-isp/isp_50/cam/mtk_cam.h     |  300 ++++
>  3 files changed, 1655 insertions(+)
>  create mode 100644 drivers/media/platform/mtk-isp/isp_50/cam/mtk_cam-regs.h
>  create mode 100644 drivers/media/platform/mtk-isp/isp_50/cam/mtk_cam.c
>  create mode 100644 drivers/media/platform/mtk-isp/isp_50/cam/mtk_cam.h
> 
> diff --git a/drivers/media/platform/mtk-isp/isp_50/cam/mtk_cam-regs.h 
> b/drivers/media/platform/mtk-isp/isp_50/cam/mtk_cam-regs.h
> new file mode 100644
> index 000000000000..342f0e0e9837
> --- /dev/null
> +++ b/drivers/media/platform/mtk-isp/isp_50/cam/mtk_cam-regs.h
> @@ -0,0 +1,149 @@
> +/* SPDX-License-Identifier: GPL-2.0 */
> +/*
> + * Copyright (c) 2018 MediaTek Inc.
> + * Author: Ryan Yu <ryan...@mediatek.com>
> + *
> + * This program is free software; you can redistribute it and/or modify
> + * it under the terms of the GNU General Public License version 2 as
> + * published by the Free Software Foundation.
> + *
> + * This program is distributed in the hope that it will be useful,
> + * but WITHOUT ANY WARRANTY; without even the implied warranty of
> + * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
> + * GNU General Public License for more details.
> + */
> +
> +#ifndef _CAM_REGS_H
> +#define _CAM_REGS_H
> +
> +/* TG Bit Mask */
> +#define VFDATA_EN_BIT        BIT(0)
> +#define CMOS_EN_BIT  BIT(0)
> +
> +/* normal signal bit */
> +#define VS_INT_ST    BIT(0)
> +#define HW_PASS1_DON_ST      BIT(11)
> +#define SOF_INT_ST   BIT(12)
> +#define SW_PASS1_DON_ST      BIT(30)
> +
> +/* err status bit */
> +#define TG_ERR_ST    BIT(4)
> +#define TG_GBERR_ST  BIT(5)
> +#define CQ_CODE_ERR_ST       BIT(6)
> +#define CQ_APB_ERR_ST        BIT(7)
> +#define CQ_VS_ERR_ST BIT(8)
> +#define AMX_ERR_ST   BIT(15)
> +#define RMX_ERR_ST   BIT(16)
> +#define BMX_ERR_ST   BIT(17)
> +#define RRZO_ERR_ST  BIT(18)
> +#define AFO_ERR_ST   BIT(19)
> +#define IMGO_ERR_ST  BIT(20)
> +#define AAO_ERR_ST   BIT(21)
> +#define PSO_ERR_ST   BIT(22)
> +#define LCSO_ERR_ST  BIT(23)
> +#define BNR_ERR_ST   BIT(24)
> +#define LSCI_ERR_ST  BIT(25)
> +#define DMA_ERR_ST   BIT(29)
> +
> +/* CAM DMA done status */
> +#define FLKO_DONE_ST BIT(4)
> +#define AFO_DONE_ST  BIT(5)
> +#define AAO_DONE_ST  BIT(7)
> +#define PSO_DONE_ST  BIT(14)

Please align the values using tabs here and elsewhere.

> +
> +/* IRQ signal mask */
> +#define INT_ST_MASK_CAM      ( \
> +                     VS_INT_ST |\
> +                     SOF_INT_ST |\
> +                     HW_PASS1_DON_ST |\
> +                     SW_PASS1_DON_ST)
> +
> +/* IRQ Warning Mask */
> +#define INT_ST_MASK_CAM_WARN (\
> +                             RRZO_ERR_ST |\
> +                             AFO_ERR_ST |\
> +                             IMGO_ERR_ST |\
> +                             AAO_ERR_ST |\
> +                             PSO_ERR_ST | \
> +                             LCSO_ERR_ST |\
> +                             BNR_ERR_ST |\
> +                             LSCI_ERR_ST)
> +
> +/* IRQ Error Mask */
> +#define INT_ST_MASK_CAM_ERR  (\
> +                             TG_ERR_ST |\
> +                             TG_GBERR_ST |\
> +                             CQ_CODE_ERR_ST |\
> +                             CQ_APB_ERR_ST |\
> +                             CQ_VS_ERR_ST |\
> +                             BNR_ERR_ST |\
> +                             RMX_ERR_ST |\
> +                             BMX_ERR_ST |\
> +                             BNR_ERR_ST |\
> +                             LSCI_ERR_ST |\
> +                             DMA_ERR_ST)
> +
> +/* IRQ Signal Log Mask */
> +#define INT_ST_LOG_MASK_CAM  (\
> +                             SOF_INT_ST |\
> +                             SW_PASS1_DON_ST |\
> +                             VS_INT_ST |\
> +                             TG_ERR_ST |\
> +                             TG_GBERR_ST |\
> +                             RRZO_ERR_ST |\
> +                             AFO_ERR_ST |\
> +                             IMGO_ERR_ST |\
> +                             AAO_ERR_ST |\
> +                             DMA_ERR_ST)
> +
> +/* DMA Event Notification Mask */
> +#define DMA_ST_MASK_CAM      (\
> +                     AFO_DONE_ST |\
> +                     AAO_DONE_ST |\
> +                     PSO_DONE_ST |\
> +                     FLKO_DONE_ST)
> +
> +/* Status check */
> +#define REG_CTL_EN           0x0004
> +#define REG_CTL_DMA_EN               0x0008
> +#define REG_CTL_FMT_SEL              0x0010
> +#define REG_CTL_EN2          0x0018
> +#define REG_CTL_RAW_INT_EN   0x0020
> +#define REG_CTL_RAW_INT_STAT 0x0024
> +#define REG_CTL_RAW_INT2_STAT        0x0034
> +#define REG_CTL_RAW_INT3_STAT        0x00c4
> +#define REG_CTL_TWIN_STAT    0x0050
> +
> +#define REG_TG_SEN_MODE              0x0230
> +#define REG_TG_SEN_GRAB_PIX  0x0238
> +#define REG_TG_SEN_GRAB_LIN  0x023c
> +#define REG_TG_VF_CON                0x0234
> +#define REG_TG_SUB_PERIOD    0x02a4
> +
> +#define REG_IMGO_BASE_ADDR   0x1020
> +#define REG_RRZO_BASE_ADDR   0x1050
> +
> +/* Error status log */
> +#define REG_IMGO_ERR_STAT    0x1360
> +#define REG_RRZO_ERR_STAT    0x1364
> +#define REG_AAO_ERR_STAT     0x1368
> +#define REG_AFO_ERR_STAT     0x136c
> +#define REG_LCSO_ERR_STAT    0x1370
> +#define REG_UFEO_ERR_STAT    0x1374
> +#define REG_PDO_ERR_STAT     0x1378
> +#define REG_BPCI_ERR_STAT    0x137c
> +#define REG_LSCI_ERR_STAT    0x1384
> +#define REG_PDI_ERR_STAT     0x138c
> +#define REG_LMVO_ERR_STAT    0x1390
> +#define REG_FLKO_ERR_STAT    0x1394
> +#define REG_PSO_ERR_STAT     0x13a0
> +
> +/* ISP command */
> +#define REG_CQ_THR0_BASEADDR 0x0198
> +#define REG_HW_FRAME_NUM     0x13b8
> +
> +/* META */
> +#define REG_META0_VB2_INDEX  0x14dc
> +#define REG_META1_VB2_INDEX  0x151c
> +
> +#endif       /* _CAM_REGS_H */
> diff --git a/drivers/media/platform/mtk-isp/isp_50/cam/mtk_cam.c 
> b/drivers/media/platform/mtk-isp/isp_50/cam/mtk_cam.c
> new file mode 100644
> index 000000000000..fc874ec8f7f0
> --- /dev/null
> +++ b/drivers/media/platform/mtk-isp/isp_50/cam/mtk_cam.c
> @@ -0,0 +1,1206 @@
> +// SPDX-License-Identifier: GPL-2.0
> +/*
> + * Copyright (c) 2018 MediaTek Inc.
> + * Author: Ryan Yu <ryan...@mediatek.com>
> + *
> + * This program is free software; you can redistribute it and/or modify
> + * it under the terms of the GNU General Public License version 2 as
> + * published by the Free Software Foundation.
> + *
> + * This program is distributed in the hope that it will be useful,
> + * but WITHOUT ANY WARRANTY; without even the implied warranty of
> + * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
> + * GNU General Public License for more details.
> + */
> +
> +#include <linux/atomic.h>
> +#include <linux/cdev.h>
> +#include <linux/compat.h>
> +#include <linux/fs.h>
> +#include <linux/interrupt.h>
> +#include <linux/jiffies.h>
> +#include <linux/kernel.h>
> +#include <linux/ktime.h>
> +#include <linux/module.h>
> +#include <linux/of_platform.h>
> +#include <linux/of_irq.h>
> +#include <linux/of_address.h>
> +#include <linux/platform_device.h>
> +#include <linux/platform_data/mtk_scp.h>
> +#include <linux/pm_runtime.h>
> +#include <linux/remoteproc.h>
> +#include <linux/sched/clock.h>
> +#include <linux/spinlock.h>
> +#include <linux/types.h>
> +#include <linux/videodev2.h>
> +#include <linux/vmalloc.h>
> +
> +#include "mtk_cam.h"
> +#include "mtk_cam-regs.h"
> +#include "mtk_cam-smem.h"
> +
> +static const struct of_device_id mtk_isp_of_ids[] = {
> +     {.compatible = "mediatek,mt8183-camisp",},
> +     {}
> +};
> +MODULE_DEVICE_TABLE(of, mtk_isp_of_ids);
> +
> +/* list of clocks required by isp cam */
> +static const char * const mtk_isp_clks[] = {
> +     "CAMSYS_CAM_CGPDN", "CAMSYS_CAMTG_CGPDN"
> +};
> +
> +static void isp_dump_dma_status(struct isp_device *isp_dev)
> +{
> +     dev_err(isp_dev->dev,
> +             "IMGO:0x%x, RRZO:0x%x, AAO=0x%x, AFO=0x%x, LMVO=0x%x\n",
> +             readl(isp_dev->regs + REG_IMGO_ERR_STAT),
> +             readl(isp_dev->regs + REG_RRZO_ERR_STAT),
> +             readl(isp_dev->regs + REG_AAO_ERR_STAT),
> +             readl(isp_dev->regs + REG_AFO_ERR_STAT),
> +             readl(isp_dev->regs + REG_LMVO_ERR_STAT));
> +     dev_err(isp_dev->dev,
> +             "LCSO=0x%x, PSO=0x%x, FLKO=0x%x, BPCI:0x%x, LSCI=0x%x\n",
> +             readl(isp_dev->regs + REG_LCSO_ERR_STAT),
> +             readl(isp_dev->regs + REG_PSO_ERR_STAT),
> +             readl(isp_dev->regs + REG_FLKO_ERR_STAT),
> +             readl(isp_dev->regs + REG_BPCI_ERR_STAT),
> +             readl(isp_dev->regs + REG_LSCI_ERR_STAT));
> +}
> +
> +static void mtk_isp_notify(struct mtk_isp_p1_ctx *isp_ctx,
> +                        unsigned int request_fd,
> +                        unsigned int frame_seq_no,
> +                        struct list_head *list_buf,
> +                        enum vb2_buffer_state state)
> +{
> +     struct isp_p1_device *p1_dev = p1_ctx_to_dev(isp_ctx);
> +     struct device *dev = &p1_dev->pdev->dev;
> +     struct mtk_cam_dev_finish_param fram_param;
> +
> +     fram_param.list_buf = list_buf;
> +     fram_param.request_fd = request_fd;
> +     fram_param.frame_seq_no = frame_seq_no;
> +     fram_param.state = state;
> +     dev_dbg(dev, "request fd:%d frame_seq_no:%d\n",
> +             fram_param.request_fd,
> +             fram_param.frame_seq_no);
> +     mtk_cam_dev_job_finish(p1_dev->cam_dev, &fram_param);
> +}
> +
> +static void isp_deque_frame(struct isp_p1_device *p1_dev,
> +                         unsigned int node_id, int vb2_index,
> +                         int frame_seq_no)
> +{
> +     struct mtk_cam_dev *cam_dev = p1_dev->cam_dev;
> +     struct device *dev = &p1_dev->pdev->dev;
> +     struct vb2_queue *vb2_queue = &cam_dev->mem2mem2_nodes[node_id].vbq;
> +     struct vb2_buffer *vb;
> +     struct vb2_v4l2_buffer *vbb;
> +
> +     if (!cam_dev->mem2mem2_nodes[node_id].enabled)
> +             return;
> +
> +     mutex_lock(vb2_queue->lock);
> +     list_for_each_entry(vb, &vb2_queue->queued_list, queued_entry) {
> +             vbb = to_vb2_v4l2_buffer(vb);
> +             if (vbb->request_fd < 0 &&
> +                 vb->index == vb2_index &&
> +                 vb->state == VB2_BUF_STATE_ACTIVE) {
> +                     dev_dbg(dev, "%s:%d:%d", __func__, node_id, vb2_index);
> +                     vbb->vb2_buf.timestamp = ktime_get_ns();
> +                     vbb->sequence = frame_seq_no;
> +                     vb2_buffer_done(vb, VB2_BUF_STATE_DONE);
> +             }
> +     }
> +     mutex_unlock(vb2_queue->lock);
> +}
> +
> +static void isp_deque_request_frame(struct isp_p1_device *p1_dev,
> +                                 int frame_seq_no)
> +{
> +     struct mtk_isp_p1_ctx *isp_ctx = &p1_dev->isp_ctx;
> +     struct device *dev = &p1_dev->pdev->dev;
> +     struct mtk_isp_queue_job *framejob, *tmp;
> +     struct isp_queue *p1_enqueue_list = &isp_ctx->p1_enqueue_list;
> +
> +     /* Match dequeue work and enqueue frame */
> +     spin_lock(&p1_enqueue_list->lock);
> +     list_for_each_entry_safe(framejob, tmp, &p1_enqueue_list->queue,
> +                              list_entry) {
> +             dev_dbg(dev,
> +                     "%s frame_seq_no:%d, target frame_seq_no:%d\n",
> +                     __func__,
> +                     framejob->frame_seq_no, frame_seq_no);
> +             /* Match by the en-queued request number */
> +             if (framejob->frame_seq_no == frame_seq_no) {
> +                     /* Pass to user space */
> +                     mtk_isp_notify(isp_ctx,
> +                                    framejob->request_fd,
> +                                    framejob->frame_seq_no,
> +                                    &framejob->list_buf,
> +                                    VB2_BUF_STATE_DONE);
> +                     atomic_dec(&p1_enqueue_list->queue_cnt);
> +                     dev_dbg(dev,
> +                             "frame_seq_no:%d is done, queue_cnt:%d\n",
> +                             framejob->frame_seq_no,
> +                             atomic_read(&p1_enqueue_list->queue_cnt));
> +
> +                     /* remove only when frame ready */
> +                     list_del(&framejob->list_entry);
> +                     kfree(framejob);
> +                     break;
> +             } else if (framejob->frame_seq_no < frame_seq_no) {
> +                     /* Pass to user space for frame drop */
> +                     mtk_isp_notify(isp_ctx,
> +                                    framejob->request_fd,
> +                                    framejob->frame_seq_no,
> +                                    &framejob->list_buf,
> +                                    VB2_BUF_STATE_ERROR);
> +                     atomic_dec(&p1_enqueue_list->queue_cnt);
> +                     dev_dbg(dev,
> +                             "frame_seq_no:%d drop, queue_cnt:%d\n",
> +                             framejob->frame_seq_no,
> +                             atomic_read(&p1_enqueue_list->queue_cnt));
> +
> +                     /* remove only drop frame */
> +                     list_del(&framejob->list_entry);
> +                     kfree(framejob);
> +             }
> +     }
> +     spin_unlock(&p1_enqueue_list->lock);
> +}
> +
> +static int isp_deque_work(void *data)
> +{
> +     struct isp_p1_device *p1_dev = (struct isp_p1_device *)data;
> +     struct mtk_isp_p1_ctx *isp_ctx = &p1_dev->isp_ctx;
> +     struct device *dev = &p1_dev->pdev->dev;
> +     struct mtk_cam_dev *cam_dev = p1_dev->cam_dev;
> +     struct mtk_cam_dev_stat_event_data event_data;
> +     atomic_t *irq_data_end = &isp_ctx->irq_data_end;
> +     atomic_t *irq_data_start = &isp_ctx->irq_data_start;
> +     unsigned long flags;
> +     int ret, i;
> +
> +     while (1) {
> +             ret = wait_event_interruptible(isp_ctx->isp_deque_thread.wq,
> +                                            (atomic_read(irq_data_end) !=
> +                                            atomic_read(irq_data_start)) ||
> +                                            kthread_should_stop());
> +
> +             if (kthread_should_stop())
> +                     break;
> +
> +             if (ret == ERESTARTSYS) {
> +                     dev_err(dev, "interrupted by a signal!\n");
> +                     continue;
> +             }
> +
> +             spin_lock_irqsave(&isp_ctx->irq_dequeue_lock, flags);
> +             i = atomic_read(&isp_ctx->irq_data_start);
> +             memcpy(&event_data, &isp_ctx->irq_event_datas[i],
> +                    sizeof(event_data));
> +             memset(&isp_ctx->irq_event_datas[i], 0x00, sizeof(event_data));
> +             atomic_set(&isp_ctx->irq_data_start, ++i & 0x3);
> +             spin_unlock_irqrestore(&isp_ctx->irq_dequeue_lock, flags);
> +
> +             if (event_data.irq_status_mask & VS_INT_ST) {
> +                     /* Notify specific HW events to user space */
> +                     mtk_cam_dev_event_frame_sync(cam_dev,
> +                                                  event_data.frame_seq_no);
> +                     dev_dbg(dev,
> +                             "event IRQ:0x%x DMA:0x%x is sent\n",
> +                             event_data.irq_status_mask,
> +                             event_data.dma_status_mask);
> +             }
> +
> +             if (event_data.dma_status_mask & AAO_DONE_ST) {
> +                     isp_deque_frame(p1_dev,
> +                                     MTK_CAM_P1_META_OUT_0,
> +                                     event_data.meta0_vb2_index,
> +                                     event_data.frame_seq_no);
> +             }
> +
> +             if (event_data.irq_status_mask & SW_PASS1_DON_ST) {
> +                     isp_deque_frame(p1_dev,
> +                                     MTK_CAM_P1_META_OUT_0,
> +                                     event_data.meta0_vb2_index,
> +                                     event_data.frame_seq_no);
> +
> +                     isp_deque_request_frame(p1_dev,
> +                                             event_data.frame_seq_no);
> +             }
> +     }
> +     return 0;
> +}
> +
> +static int irq_handle_sof(struct isp_device *isp_dev,
> +                       dma_addr_t base_addr,
> +                       unsigned int frame_num)
> +{
> +     unsigned int cq_addr_index;
> +     struct isp_p1_device *p1_dev = get_p1_device(isp_dev->dev);
> +     int cq_num = atomic_read(&p1_dev->isp_ctx.composed_frame_id);
> +
> +     if (cq_num > frame_num) {
> +             cq_addr_index = frame_num % CQ_BUFFER_COUNT;
> +
> +             writel(base_addr +
> +                     (dma_addr_t)(CQ_ADDRESS_OFFSET * cq_addr_index),
> +                     isp_dev->regs + REG_CQ_THR0_BASEADDR);
> +             dev_dbg(isp_dev->dev,
> +                     "SOF_INT_ST, update next, cq_num:%d, frame_num:%d 
> cq_addr:%d",
> +                     cq_num, frame_num, cq_addr_index);
> +     } else {
> +             dev_dbg(isp_dev->dev,
> +                     "SOF_INT_ST, wait next, cq_num:%d, frame_num:%d",
> +                     cq_num, frame_num);
> +     }
> +
> +     isp_dev->sof_count += 1;
> +
> +     return cq_num;
> +}
> +
> +static int irq_handle_notify_event(struct isp_device *isp_dev,
> +                                unsigned int irqstatus,
> +                                unsigned int dmastatus)
> +{
> +     struct isp_p1_device *p1_dev = get_p1_device(isp_dev->dev);
> +     struct mtk_isp_p1_ctx *isp_ctx = &p1_dev->isp_ctx;
> +     unsigned long flags;
> +     int i;
> +
> +     spin_lock_irqsave(&isp_ctx->irq_dequeue_lock, flags);
> +     i = atomic_read(&isp_ctx->irq_data_end);
> +     isp_ctx->irq_event_datas[i].frame_seq_no = isp_dev->current_frame;
> +     isp_ctx->irq_event_datas[i].meta0_vb2_index = isp_dev->meta0_vb2_index;
> +     isp_ctx->irq_event_datas[i].meta1_vb2_index = isp_dev->meta1_vb2_index;
> +     isp_ctx->irq_event_datas[i].irq_status_mask |=
> +             (irqstatus & INT_ST_MASK_CAM);
> +     isp_ctx->irq_event_datas[i].dma_status_mask |=
> +             (dmastatus & DMA_ST_MASK_CAM);
> +     atomic_set(&isp_ctx->irq_data_end, ++i & 0x3);
> +     spin_unlock_irqrestore(&isp_ctx->irq_dequeue_lock, flags);
> +
> +     wake_up_interruptible(&isp_ctx->isp_deque_thread.wq);
> +
> +     dev_dbg(isp_dev->dev,
> +             "%s IRQ:0x%x DMA:0x%x seq:%d idx0:%d idx1:%d\n",
> +             __func__,
> +             (irqstatus & INT_ST_MASK_CAM),
> +             (dmastatus & DMA_ST_MASK_CAM),
> +             isp_dev->current_frame,
> +             isp_dev->meta0_vb2_index,
> +             isp_dev->meta1_vb2_index);
> +
> +     return 0;
> +}
> +
> +irqreturn_t isp_irq_cam(int irq, void *data)
> +{
> +     struct isp_device *isp_dev = (struct isp_device *)data;
> +     struct isp_p1_device *p1_dev = get_p1_device(isp_dev->dev);
> +     struct mtk_isp_p1_ctx *isp_ctx = &p1_dev->isp_ctx;
> +     struct device *dev = isp_dev->dev;
> +     unsigned int cardinalnum, cq_num, hw_frame_num;
> +     unsigned int meta0_vb2_index, meta1_vb2_index;
> +     unsigned int irqstatus, errstatus, warnstatus, dmastatus;
> +     unsigned long flags;
> +
> +     /* Check the streaming is off or not */
> +     if (!p1_dev->cam_dev->streaming)
> +             return IRQ_HANDLED;
> +
> +     cardinalnum = isp_dev->isp_hw_module - ISP_CAM_A_IDX;
> +     cq_num = 0;
> +
> +     spin_lock_irqsave(&isp_dev->spinlock_irq, flags);
> +     irqstatus = readl(isp_dev->regs + REG_CTL_RAW_INT_STAT);
> +     dmastatus = readl(isp_dev->regs + REG_CTL_RAW_INT2_STAT);
> +     hw_frame_num = readl(isp_dev->regs + REG_HW_FRAME_NUM);
> +     meta0_vb2_index = readl(isp_dev->regs + REG_META0_VB2_INDEX);
> +     meta1_vb2_index = readl(isp_dev->regs + REG_META1_VB2_INDEX);
> +     spin_unlock_irqrestore(&isp_dev->spinlock_irq, flags);
> +
> +     /* Ignore unnecessary IRQ */
> +     if (irqstatus == 0)
> +             return IRQ_HANDLED;
> +
> +     errstatus = irqstatus & INT_ST_MASK_CAM_ERR;
> +     warnstatus = irqstatus & INT_ST_MASK_CAM_WARN;
> +     irqstatus = irqstatus & INT_ST_MASK_CAM;
> +
> +     /* sof , done order check . */
> +     spin_lock_irqsave(&isp_dev->spinlock_irq, flags);
> +     if ((irqstatus & HW_PASS1_DON_ST) && (irqstatus & SOF_INT_ST)) {
> +             dev_warn(dev,
> +                      "isp sof_don block, sof_cnt:%d\n",
> +                      isp_dev->sof_count);
> +
> +             /* Notify IRQ event and enqueue ready frame */
> +             irq_handle_notify_event(isp_dev, irqstatus, dmastatus);
> +             isp_dev->current_frame = hw_frame_num;
> +             isp_dev->meta0_vb2_index = meta0_vb2_index;
> +             isp_dev->meta1_vb2_index = meta1_vb2_index;
> +     } else {
> +             if (irqstatus & SOF_INT_ST) {
> +                     isp_dev->current_frame = hw_frame_num;
> +                     isp_dev->meta0_vb2_index = meta0_vb2_index;
> +                     isp_dev->meta1_vb2_index = meta1_vb2_index;
> +             }
> +
> +             if ((irqstatus & INT_ST_MASK_CAM) ||
> +                 (dmastatus & DMA_ST_MASK_CAM))
> +                     irq_handle_notify_event(isp_dev, irqstatus, dmastatus);
> +     }
> +     spin_unlock_irqrestore(&isp_dev->spinlock_irq, flags);
> +
> +     if (irqstatus & SOF_INT_ST)
> +             cq_num = irq_handle_sof(isp_dev, isp_ctx->scp_mem_iova,
> +                                     hw_frame_num);
> +
> +     if (irqstatus & SW_PASS1_DON_ST) {
> +             int num = atomic_dec_return(&isp_ctx->composing_frame);
> +
> +             dev_dbg(dev, "SW_PASS1_DON_ST queued frame:%d\n", num);
> +             /* Notify TX thread to send if TX frame is blocked */
> +             wake_up_interruptible
> +                             (&isp_ctx->composer_tx_thread.wq);
> +     }
> +
> +     /* check ISP error status */
> +     if (errstatus) {
> +             dev_err(dev,
> +                     "raw_int_err:0x%x/0x%x/0x%x\n",
> +                     irqstatus, warnstatus, errstatus);
> +
> +             /* show DMA errors in detail */
> +             if (errstatus & DMA_ERR_ST)
> +                     isp_dump_dma_status(isp_dev);
> +     }
> +
> +     if (irqstatus & INT_ST_LOG_MASK_CAM)
> +             dev_dbg(dev, IRQ_STAT_STR,
> +                     'A' + cardinalnum,
> +                     isp_dev->sof_count,
> +                     irqstatus,
> +                     dmastatus,
> +                     hw_frame_num,
> +                     cq_num);
> +     return IRQ_HANDLED;
> +}
> +
> +static int enable_sys_clock(struct isp_p1_device *p1_dev)
> +{
> +     struct device *dev = &p1_dev->pdev->dev;
> +     int ret;
> +
> +     dev_info(dev, "- %s dev id:%d\n", __func__, dev->id);
> +
> +     ret = clk_bulk_prepare_enable(p1_dev->isp_clk.num_clks,
> +                                   p1_dev->isp_clk.clk_list);
> +     if (ret < 0)
> +             goto clk_err;
> +     return 0;
> +clk_err:
> +     dev_err(dev, "cannot pre-en isp_cam clock:%d\n", ret);
> +     clk_bulk_disable_unprepare(p1_dev->isp_clk.num_clks,
> +                                p1_dev->isp_clk.clk_list);
> +     return ret;
> +}
> +
> +static void disable_sys_clock(struct isp_p1_device *p1_dev)
> +{
> +     struct device *dev = &p1_dev->pdev->dev;
> +
> +     dev_info(dev, "- %s dev id:%d\n", __func__, dev->id);
> +     clk_bulk_disable_unprepare(p1_dev->isp_clk.num_clks,
> +                                p1_dev->isp_clk.clk_list);
> +}
> +
> +static int mtk_isp_probe(struct platform_device *pdev)
> +{
> +     struct isp_p1_device *p1_dev;
> +     struct mtk_isp_p1_ctx *isp_ctx;
> +     struct isp_device *isp_dev;
> +     struct device *dev = &pdev->dev;
> +     struct resource *res;
> +     int ret;
> +     unsigned int i;
> +
> +     /* Allocate context */
> +     p1_dev = devm_kzalloc(dev, sizeof(*p1_dev), GFP_KERNEL);
> +     if (!p1_dev)
> +             return -ENOMEM;
> +
> +     dev_set_drvdata(dev, p1_dev);
> +     isp_ctx = &p1_dev->isp_ctx;
> +     p1_dev->pdev = pdev;
> +
> +     p1_dev->isp_devs =
> +             devm_kzalloc(dev,
> +                          sizeof(struct isp_device) * ISP_DEV_NODE_NUM,
> +                          GFP_KERNEL);
> +     if (!p1_dev->isp_devs)
> +             return -ENOMEM;
> +
> +     p1_dev->cam_dev =
> +             devm_kzalloc(dev, sizeof(struct mtk_cam_dev), GFP_KERNEL);
> +     if (!p1_dev->isp_devs)
> +             return -ENOMEM;
> +
> +     /* iomap registers */
> +     for (i = ISP_CAMSYS_CONFIG_IDX; i < ISP_DEV_NODE_NUM; i++) {
> +             isp_dev = &p1_dev->isp_devs[i];
> +             isp_dev->isp_hw_module = i;
> +             isp_dev->dev = dev;
> +             res = platform_get_resource(pdev, IORESOURCE_MEM, i);
> +             isp_dev->regs = devm_ioremap_resource(dev, res);
> +
> +             dev_info(dev, "cam%u, map_addr=0x%lx\n",
> +                      i, (unsigned long)isp_dev->regs);
> +
> +             if (!isp_dev->regs)
> +                     return PTR_ERR(isp_dev->regs);
> +
> +             /* support IRQ from ISP_CAM_A_IDX */
> +             if (i >= ISP_CAM_A_IDX) {
> +                     /* reg & interrupts index is shifted with 1  */
> +                     isp_dev->irq = platform_get_irq(pdev, i - 1);
> +                     if (isp_dev->irq > 0) {
> +                             ret = devm_request_irq(dev, isp_dev->irq,
> +                                                    isp_irq_cam,
> +                                                    IRQF_SHARED,
> +                                                    dev_driver_string(dev),
> +                                                    (void *)isp_dev);
> +                             if (ret) {
> +                                     dev_err(dev,
> +                                             "req_irq fail, dev:%s irq=%d\n",
> +                                             dev->of_node->name,
> +                                             isp_dev->irq);
> +                                     return ret;
> +                             }
> +                             dev_info(dev, "Registered irq=%d, ISR:%s\n",
> +                                      isp_dev->irq, dev_driver_string(dev));
> +                     }
> +             }
> +             spin_lock_init(&isp_dev->spinlock_irq);
> +     }
> +
> +     p1_dev->isp_clk.num_clks = ARRAY_SIZE(mtk_isp_clks);
> +     p1_dev->isp_clk.clk_list =
> +             devm_kcalloc(dev,
> +                          p1_dev->isp_clk.num_clks,
> +                          sizeof(*p1_dev->isp_clk.clk_list),
> +                          GFP_KERNEL);
> +     if (!p1_dev->isp_clk.clk_list)
> +             return -ENOMEM;
> +
> +     for (i = 0; i < p1_dev->isp_clk.num_clks; ++i)
> +             p1_dev->isp_clk.clk_list->id = mtk_isp_clks[i];
> +
> +     ret = devm_clk_bulk_get(dev,
> +                             p1_dev->isp_clk.num_clks,
> +                             p1_dev->isp_clk.clk_list);
> +     if (ret) {
> +             dev_err(dev, "cannot get isp cam clock:%d\n", ret);
> +             return ret;
> +     }
> +
> +     /* Initialize reserved DMA memory */
> +     ret = mtk_cam_reserved_memory_init(p1_dev);
> +     if (ret) {
> +             dev_err(dev, "failed to configure DMA memory\n");
> +             return ret;
> +     }
> +
> +     /* Initialize the v4l2 common part */
> +     ret = mtk_cam_dev_init(pdev, p1_dev->cam_dev);
> +     if (ret)
> +             return ret;
> +
> +     spin_lock_init(&isp_ctx->p1_enqueue_list.lock);
> +     atomic_set(&p1_dev->isp_ctx.isp_user_cnt, 0);
> +     pm_runtime_enable(dev);
> +
> +     return 0;
> +}
> +
> +static int mtk_isp_remove(struct platform_device *pdev)
> +{
> +     struct device *dev = &pdev->dev;
> +     struct isp_p1_device *p1_dev = dev_get_drvdata(dev);
> +
> +     pm_runtime_disable(dev);
> +     mtk_cam_dev_release(pdev, p1_dev->cam_dev);
> +
> +     return 0;
> +}
> +
> +static int mtk_isp_suspend(struct device *dev)
> +{
> +     struct isp_p1_device *p1_dev = get_p1_device(dev);
> +     struct isp_device *isp_dev;
> +     unsigned int reg_val;
> +     int usercount, module;
> +
> +     module = p1_dev->isp_ctx.isp_hw_module;
> +     usercount = atomic_read(&p1_dev->isp_ctx.isp_user_cnt);
> +
> +     dev_dbg(dev, "- %s:%d\n", __func__, usercount);
> +
> +     /* If no user count, no further action */
> +     if (!usercount)
> +             return 0;
> +
> +     isp_dev = &p1_dev->isp_devs[module];
> +     reg_val = readl(isp_dev->regs + REG_TG_VF_CON);
> +     if (reg_val & VFDATA_EN_BIT) {
> +             dev_dbg(dev, "Cam:%d suspend, disable VF\n", module);
> +             /* disable VF */
> +             writel((reg_val & (~VFDATA_EN_BIT)),
> +                    isp_dev->regs + REG_TG_VF_CON);
> +             /*
> +              * After VF enable, The TG frame count will be reset to 0;
> +              */
> +             reg_val = readl(isp_dev->regs + REG_TG_SEN_MODE);
> +             writel((reg_val & (~CMOS_EN_BIT)),
> +                    isp_dev->regs +  + REG_TG_SEN_MODE);
> +     }
> +
> +     disable_sys_clock(p1_dev);
> +
> +     return 0;
> +}
> +
> +static int mtk_isp_resume(struct device *dev)
> +{
> +     struct isp_p1_device *p1_dev = get_p1_device(dev);
> +     struct isp_device *isp_dev;
> +     unsigned int reg_val;
> +     int module, usercount;
> +
> +     module = p1_dev->isp_ctx.isp_hw_module;
> +     usercount = atomic_read(&p1_dev->isp_ctx.isp_user_cnt);
> +
> +     dev_dbg(dev, "- %s:%d\n", __func__, usercount);
> +
> +     /* If no user count, no further action */
> +     if (!usercount)
> +             return 0;
> +
> +     enable_sys_clock(p1_dev);
> +
> +     /* V4L2 stream-on phase & restore HW stream-on status */
> +     if (p1_dev->cam_dev->streaming) {
> +             isp_dev = &p1_dev->isp_devs[module];
> +             dev_dbg(dev, "Cam:%d resume,enable VF\n", module);
> +             /* Enable CMOS */
> +             reg_val = readl(isp_dev->regs + REG_TG_SEN_MODE);
> +             writel((reg_val | CMOS_EN_BIT),
> +                    isp_dev->regs + REG_TG_SEN_MODE);
> +             /* Enable VF */
> +             reg_val = readl(isp_dev->regs + REG_TG_VF_CON);
> +             writel((reg_val | VFDATA_EN_BIT),
> +                    isp_dev->regs + REG_TG_VF_CON);
> +     }
> +     return 0;
> +}
> +
> +static int isp_setup_scp_rproc(struct isp_p1_device *p1_dev)
> +{
> +     phandle rproc_phandle;
> +     struct device *dev = &p1_dev->pdev->dev;
> +     int ret;
> +
> +     p1_dev->scp_pdev = scp_get_pdev(p1_dev->pdev);
> +     if (!p1_dev->scp_pdev) {
> +             dev_err(dev, "Failed to get scp device\n");
> +             return -ENODEV;
> +     }
> +     ret = of_property_read_u32(dev->of_node, "mediatek,scp",
> +                                &rproc_phandle);
> +     if (ret) {
> +             dev_err(dev, "fail to get rproc_phandle:%d\n", ret);
> +             return -EINVAL;
> +     }
> +
> +     p1_dev->rproc_handle = rproc_get_by_phandle(rproc_phandle);
> +     dev_dbg(dev, "p1 rproc_phandle: 0x%pK\n\n",
> +             p1_dev->rproc_handle);
> +     if (!p1_dev->rproc_handle) {
> +             dev_err(dev, "fail to get rproc_handle\n");
> +             return -EINVAL;
> +     }
> +
> +     ret = rproc_boot(p1_dev->rproc_handle);
> +     if (ret < 0) {
> +             /*
> +              * Return 0 if downloading firmware successfully,
> +              * otherwise it is failed
> +              */
> +             return -ENODEV;
> +     }
> +
> +     return 0;
> +}
> +
> +static int isp_init_context(struct isp_p1_device *p1_dev)
> +{
> +     struct mtk_isp_p1_ctx *isp_ctx = &p1_dev->isp_ctx;
> +     struct device *dev = &p1_dev->pdev->dev;
> +     unsigned int i;
> +
> +     dev_dbg(dev, "init irq work thread\n");
> +     if (!isp_ctx->isp_deque_thread.thread) {
> +             mutex_init(&isp_ctx->composer_tx_lock);
> +             init_waitqueue_head(&isp_ctx->isp_deque_thread.wq);
> +             isp_ctx->isp_deque_thread.thread =
> +                     kthread_run(isp_deque_work, (void *)p1_dev,
> +                                 "isp_deque_work");
> +             if (IS_ERR(isp_ctx->isp_deque_thread.thread)) {
> +                     dev_err(dev, "unable to alloc kthread\n");
> +                     isp_ctx->isp_deque_thread.thread = NULL;
> +                     return -ENOMEM;
> +             }
> +     }
> +     spin_lock_init(&isp_ctx->irq_dequeue_lock);
> +
> +     INIT_LIST_HEAD(&isp_ctx->p1_enqueue_list.queue);
> +     atomic_set(&isp_ctx->p1_enqueue_list.queue_cnt, 0);
> +
> +     for (i = 0; i < ISP_DEV_NODE_NUM; i++)
> +             spin_lock_init(&p1_dev->isp_devs[i].spinlock_irq);
> +
> +     spin_lock_init(&isp_ctx->p1_enqueue_list.lock);
> +     spin_lock_init(&isp_ctx->composer_txlist.lock);
> +
> +     atomic_set(&isp_ctx->irq_data_end, 0);
> +     atomic_set(&isp_ctx->irq_data_start, 0);
> +     return 0;
> +}
> +
> +static int isp_uninit_context(struct isp_p1_device *p1_dev)
> +{
> +     struct mtk_isp_p1_ctx *isp_ctx = &p1_dev->isp_ctx;
> +     struct mtk_isp_queue_job *framejob, *tmp_framejob;
> +
> +     spin_lock_irq(&isp_ctx->p1_enqueue_list.lock);
> +     list_for_each_entry_safe(framejob, tmp_framejob,
> +                              &isp_ctx->p1_enqueue_list.queue, list_entry) {
> +             list_del(&framejob->list_entry);
> +             kfree(framejob);
> +     }
> +     spin_unlock_irq(&isp_ctx->p1_enqueue_list.lock);
> +
> +     atomic_set(&isp_ctx->isp_user_cnt, 0);
> +
> +     if (!IS_ERR(isp_ctx->isp_deque_thread.thread)) {
> +             kthread_stop(isp_ctx->isp_deque_thread.thread);
> +             wake_up_interruptible(&isp_ctx->isp_deque_thread.wq);
> +             isp_ctx->isp_deque_thread.thread = NULL;
> +     }
> +
> +     return 0;
> +}
> +
> +static unsigned int get_enable_dma_ports(struct mtk_cam_dev *cam_dev)

I think s/enable/enabled would be a clearer name for both the function
and local variable.

> +{
> +     unsigned int enable_dma_ports, i;
> +
> +     /* Get the enabled meta DMA ports */
> +     enable_dma_ports = 0;
> +     for (i = 0; i < cam_dev->dev_node_num; i++) {
> +             if (cam_dev->mem2mem2_nodes[i].enabled)
> +                     enable_dma_ports |=
> +                             cam_dev->mem2mem2_nodes[i].desc.dma_port;
> +     }
> +     dev_dbg(&cam_dev->pdev->dev, "%s enable_dma_ports:0x%x",
> +             __func__, enable_dma_ports);
> +
> +     return enable_dma_ports;
> +}
> +
> +/* Utility functions */
> +static unsigned int get_sensor_pixel_id(unsigned int fmt)
> +{
> +     switch (fmt) {
> +     case MEDIA_BUS_FMT_SBGGR8_1X8:
> +     case MEDIA_BUS_FMT_SBGGR10_1X10:
> +     case MEDIA_BUS_FMT_SBGGR12_1X12:
> +     case MEDIA_BUS_FMT_SBGGR14_1X14:
> +             return raw_pxl_id_b;
> +     case MEDIA_BUS_FMT_SGBRG8_1X8:
> +     case MEDIA_BUS_FMT_SGBRG10_1X10:
> +     case MEDIA_BUS_FMT_SGBRG12_1X12:
> +     case MEDIA_BUS_FMT_SGBRG14_1X14:
> +             return raw_pxl_id_gb;
> +     case MEDIA_BUS_FMT_SGRBG8_1X8:
> +     case MEDIA_BUS_FMT_SGRBG10_1X10:
> +     case MEDIA_BUS_FMT_SGRBG12_1X12:
> +     case MEDIA_BUS_FMT_SGRBG14_1X14:
> +             return raw_pxl_id_gr;
> +     case MEDIA_BUS_FMT_SRGGB8_1X8:
> +     case MEDIA_BUS_FMT_SRGGB10_1X10:
> +     case MEDIA_BUS_FMT_SRGGB12_1X12:
> +     case MEDIA_BUS_FMT_SRGGB14_1X14:
> +             return raw_pxl_id_r;
> +     default:
> +             return raw_pxl_id_b;
> +     }
> +}
> +
> +static unsigned int get_sensor_fmt(unsigned int fmt)
> +{
> +     switch (fmt) {
> +     case MEDIA_BUS_FMT_SBGGR8_1X8:
> +     case MEDIA_BUS_FMT_SGBRG8_1X8:
> +     case MEDIA_BUS_FMT_SGRBG8_1X8:
> +     case MEDIA_BUS_FMT_SRGGB8_1X8:
> +             return img_fmt_bayer8;
> +     case MEDIA_BUS_FMT_SBGGR10_1X10:
> +     case MEDIA_BUS_FMT_SGBRG10_1X10:
> +     case MEDIA_BUS_FMT_SGRBG10_1X10:
> +     case MEDIA_BUS_FMT_SRGGB10_1X10:
> +             return img_fmt_bayer10;
> +     case MEDIA_BUS_FMT_SBGGR12_1X12:
> +     case MEDIA_BUS_FMT_SGBRG12_1X12:
> +     case MEDIA_BUS_FMT_SGRBG12_1X12:
> +     case MEDIA_BUS_FMT_SRGGB12_1X12:
> +             return img_fmt_bayer12;
> +     case MEDIA_BUS_FMT_SBGGR14_1X14:
> +     case MEDIA_BUS_FMT_SGBRG14_1X14:
> +     case MEDIA_BUS_FMT_SGRBG14_1X14:
> +     case MEDIA_BUS_FMT_SRGGB14_1X14:
> +             return img_fmt_bayer14;
> +     default:
> +             return img_fmt_unknown;
> +     }
> +}
> +
> +static unsigned int get_img_fmt(unsigned int fourcc)
> +{
> +     switch (fourcc) {
> +     case V4L2_PIX_FMT_MTISP_B8:
> +             return img_fmt_bayer8;
> +     case V4L2_PIX_FMT_MTISP_F8:
> +             return img_fmt_fg_bayer8;
> +     case V4L2_PIX_FMT_MTISP_B10:
> +             return img_fmt_bayer10;
> +     case V4L2_PIX_FMT_MTISP_F10:
> +             return img_fmt_fg_bayer10;
> +     case V4L2_PIX_FMT_MTISP_B12:
> +             return img_fmt_bayer12;
> +     case V4L2_PIX_FMT_MTISP_F12:
> +             return img_fmt_fg_bayer12;
> +     case V4L2_PIX_FMT_MTISP_B14:
> +             return img_fmt_bayer14;
> +     case V4L2_PIX_FMT_MTISP_F14:
> +             return img_fmt_fg_bayer14;
> +     default:
> +             return img_fmt_unknown;
> +     }
> +}
> +
> +static unsigned int get_pixel_byte(unsigned int fourcc)
> +{
> +     switch (fourcc) {
> +     case V4L2_PIX_FMT_MTISP_B8:
> +     case V4L2_PIX_FMT_MTISP_F8:
> +             return 8;
> +     case V4L2_PIX_FMT_MTISP_B10:
> +     case V4L2_PIX_FMT_MTISP_F10:
> +             return 10;
> +     case V4L2_PIX_FMT_MTISP_B12:
> +     case V4L2_PIX_FMT_MTISP_F12:
> +             return 12;
> +     case V4L2_PIX_FMT_MTISP_B14:
> +     case V4L2_PIX_FMT_MTISP_F14:
> +             return 14;
> +     case V4L2_PIX_FMT_MTISP_U8:
> +     case V4L2_PIX_FMT_MTISP_U10:
> +     case V4L2_PIX_FMT_MTISP_U12:
> +     case V4L2_PIX_FMT_MTISP_U14:
> +             return 16;
> +     default:
> +             return 10;
> +     }
> +}
> +
> +static void composer_deinit_done_cb(void *data)
> +{
> +     struct isp_p1_device *p1_dev = p1_ctx_to_dev(data);
> +
> +     disable_sys_clock(p1_dev);
> +     /* Notify PM */
> +     pm_runtime_put_sync(&p1_dev->pdev->dev);
> +}
> +
> +/* ISP P1 interface functions */
> +int mtk_isp_open(struct device *dev)
> +{
> +     struct isp_p1_device *p1_dev = get_p1_device(dev);
> +     struct mtk_isp_p1_ctx *isp_ctx = &p1_dev->isp_ctx;
> +     s32 usercount = atomic_inc_return(&isp_ctx->isp_user_cnt);
> +     int ret;
> +
> +     dev_dbg(dev, "%s usercount=%d\n", __func__, usercount);
> +
> +     if (usercount == 1) {
> +             ret = isp_setup_scp_rproc(p1_dev);
> +             if (ret)
> +                     goto scp_err;
> +
> +             /* ISP HW INIT */
> +             isp_ctx->isp_hw_module = ISP_CAM_B_IDX;
> +             /* Use pure RAW as default HW path */
> +             isp_ctx->isp_raw_path = ISP_PURE_RAW_PATH;
> +             /* Check enabled DMAs which is configured by media setup */
> +             isp_ctx->enable_dma_ports =
> +                     get_enable_dma_ports(p1_dev->cam_dev);
> +
> +             if (!isp_ctx->enable_dma_ports) {
> +                     dev_dbg(dev, "No DMAs are enabled\n");
> +                     ret = -EINVAL;
> +                     goto scp_err;
> +             }
> +
> +             pm_runtime_get_sync(dev);
> +
> +             ret = isp_init_context(p1_dev);
> +             if (ret)
> +                     goto ctx_err;
> +             ret = isp_composer_init(isp_ctx);
> +             if (ret)
> +                     goto composer_err;
> +             ret = isp_composer_hw_init(isp_ctx);
> +             if (ret)
> +                     goto composer_err;
> +
> +             isp_composer_meta_config(&p1_dev->isp_ctx,
> +                                      isp_ctx->enable_dma_ports);
> +     }
> +
> +     return 0;
> +composer_err:
> +     isp_uninit_context(p1_dev);
> +ctx_err:
> +     pm_runtime_put_sync(dev);
> +scp_err:
> +     atomic_dec_return(&isp_ctx->isp_user_cnt);
> +     return ret;
> +}
> +
> +int mtk_isp_release(struct device *dev)
> +{
> +     struct isp_p1_device *p1_dev = get_p1_device(dev);
> +     struct mtk_isp_p1_ctx *isp_ctx = &p1_dev->isp_ctx;
> +
> +     if (atomic_dec_and_test(&p1_dev->isp_ctx.isp_user_cnt)) {
> +             isp_composer_hw_deinit(isp_ctx, composer_deinit_done_cb);
> +             isp_uninit_context(p1_dev);
> +     }
> +
> +     dev_dbg(dev, "%s usercount=%d\n", __func__,
> +             atomic_read(&p1_dev->isp_ctx.isp_user_cnt));
> +
> +     return 0;
> +}
> +
> +int mtk_isp_config(struct device *dev)
> +{
> +     struct isp_p1_device *p1_dev = get_p1_device(dev);
> +     struct mtk_isp_p1_ctx *isp_ctx = &p1_dev->isp_ctx;
> +     struct p1_config_param config_param;
> +     struct mtk_cam_dev *cam_dev = p1_dev->cam_dev;
> +     struct v4l2_subdev_format sd_format;
> +     unsigned int sd_width, sd_height;
> +     unsigned int enable_dma_ports, idx;
> +     int ret;
> +
> +     p1_dev->isp_devs[isp_ctx->isp_hw_module].current_frame = 0;
> +     p1_dev->isp_devs[isp_ctx->isp_hw_module].sof_count = 0;
> +
> +     isp_ctx->frame_seq_no = 1;
> +     atomic_set(&isp_ctx->composed_frame_id, 0);
> +
> +     /* Get the enabled DMA ports */
> +     enable_dma_ports = isp_ctx->enable_dma_ports;
> +     dev_dbg(dev, "%s enable_dma_ports:0x%x", __func__, enable_dma_ports);
> +
> +     /* sensor config */
> +     sd_format.which = V4L2_SUBDEV_FORMAT_ACTIVE;
> +     ret = v4l2_subdev_call(cam_dev->sensor,
> +                            pad, get_fmt, NULL, &sd_format);
> +
> +     if (ret) {
> +             dev_dbg(dev, "sensor:%s g_fmt on failed:%d\n",
> +                     cam_dev->sensor->entity.name, ret);
> +             return -EPERM;
> +     }
> +
> +     dev_dbg(dev,
> +             "sensor get_fmt ret=%d, w=%d, h=%d, code=0x%x, field=%d, 
> color=%d\n",
> +             ret, sd_format.format.width, sd_format.format.height,
> +             sd_format.format.code, sd_format.format.field,
> +             sd_format.format.colorspace);
> +
> +     config_param.cfg_in_param.continuous = 0x1;
> +     config_param.cfg_in_param.subsample = 0x0;
> +     /* fix to one pixel mode in default */
> +     config_param.cfg_in_param.pixel_mode = one_pixel_mode;
> +     /* support normal pattern in default */
> +     config_param.cfg_in_param.data_pattern = 0x0;
> +
> +     config_param.cfg_in_param.crop.left = 0x0;
> +     config_param.cfg_in_param.crop.top = 0x0;
> +
> +     config_param.cfg_in_param.raw_pixel_id =
> +             get_sensor_pixel_id(sd_format.format.code);
> +     config_param.cfg_in_param.img_fmt =
> +             get_sensor_fmt(sd_format.format.code);
> +     config_param.cfg_in_param.crop.width = sd_format.format.width;
> +     config_param.cfg_in_param.crop.height = sd_format.format.height;
> +     sd_width = sd_format.format.width;
> +     sd_height = sd_format.format.height;
> +
> +     idx = MTK_CAM_P1_MAIN_STREAM_OUT;

The idx variable is unnecessary. Just use MTK_CAM_P1_... to index into
mem2mem2_nodes directly here and below.

> +     if ((enable_dma_ports & R_IMGO) == R_IMGO) {
> +             struct v4l2_format *imgo_fmt =
> +                     &p1_dev->cam_dev->mem2mem2_nodes[idx].vdev_fmt;
> +
> +             config_param.cfg_main_param.pure_raw = isp_ctx->isp_raw_path;
> +             config_param.cfg_main_param.pure_raw_pack = 1;
> +             config_param.cfg_main_param.bypass = 0;
> +
> +             config_param.cfg_main_param.output.img_fmt =
> +                     get_img_fmt(imgo_fmt->fmt.pix_mp.pixelformat);
> +             config_param.cfg_main_param.output.pixel_byte =
> +                     get_pixel_byte(imgo_fmt->fmt.pix_mp.pixelformat);
> +             config_param.cfg_main_param.output.size.w =
> +                     imgo_fmt->fmt.pix_mp.width;
> +             config_param.cfg_main_param.output.size.h =
> +                     imgo_fmt->fmt.pix_mp.height;
> +
> +             config_param.cfg_main_param.output.size.stride =
> +                     imgo_fmt->fmt.pix_mp.plane_fmt[0].bytesperline;
> +             config_param.cfg_main_param.output.size.xsize =
> +                     imgo_fmt->fmt.pix_mp.plane_fmt[0].bytesperline;
> +
> +             config_param.cfg_main_param.output.crop.left = 0x0;
> +             config_param.cfg_main_param.output.crop.top = 0x0;
> +
> +             config_param.cfg_main_param.output.crop.width = sd_width;
> +             config_param.cfg_main_param.output.crop.height = sd_height;
> +
> +             WARN_ONCE(imgo_fmt->fmt.pix_mp.width > sd_width ||
> +                       imgo_fmt->fmt.pix_mp.height > sd_height,
> +                       "img out:%d:%d in:%d:%d",
> +                       imgo_fmt->fmt.pix_mp.width,
> +                       imgo_fmt->fmt.pix_mp.height,
> +                       sd_width,
> +                       sd_height);
> +
> +             dev_dbg(dev,
> +                     "imgo pixel_byte:%d img_fmt:0x%x raw:%d\n",
> +                     config_param.cfg_main_param.output.pixel_byte,
> +                     config_param.cfg_main_param.output.img_fmt,
> +                     config_param.cfg_main_param.pure_raw);
> +             dev_dbg(dev,
> +                     "imgo param:size=%0dx%0d, 
> stride:%d,xsize:%d,crop=%0dx%0d\n",
> +                     config_param.cfg_main_param.output.size.w,
> +                     config_param.cfg_main_param.output.size.h,
> +                     config_param.cfg_main_param.output.size.stride,
> +                     config_param.cfg_main_param.output.size.xsize,
> +                     config_param.cfg_main_param.output.crop.width,
> +                     config_param.cfg_main_param.output.crop.height);
> +     } else {
> +             config_param.cfg_main_param.bypass = 1;
> +     }
> +
> +     idx = MTK_CAM_P1_PACKED_BIN_OUT;
> +     if ((enable_dma_ports & R_RRZO) == R_RRZO) {
> +             struct v4l2_format *rrzo_fmt =
> +                     &p1_dev->cam_dev->mem2mem2_nodes[idx].vdev_fmt;
> +
> +             config_param.cfg_resize_param.bypass = 0;
> +             config_param.cfg_resize_param.output.img_fmt =
> +                     get_img_fmt(rrzo_fmt->fmt.pix_mp.pixelformat);
> +             config_param.cfg_resize_param.output.pixel_byte =
> +                     get_pixel_byte(rrzo_fmt->fmt.pix_mp.pixelformat);
> +             config_param.cfg_resize_param.output.size.w =
> +                     rrzo_fmt->fmt.pix_mp.width;
> +             config_param.cfg_resize_param.output.size.h =
> +                     rrzo_fmt->fmt.pix_mp.height;
> +             config_param.cfg_resize_param.output.size.stride =
> +                     rrzo_fmt->fmt.pix_mp.plane_fmt[0].bytesperline;
> +             config_param.cfg_resize_param.output.size.xsize =
> +                     rrzo_fmt->fmt.pix_mp.plane_fmt[0].bytesperline;
> +
> +             config_param.cfg_resize_param.output.crop.left = 0x0;
> +             config_param.cfg_resize_param.output.crop.top = 0x0;
> +             config_param.cfg_resize_param.output.crop.width = sd_width;
> +             config_param.cfg_resize_param.output.crop.height = sd_height;
> +
> +             WARN_ONCE(rrzo_fmt->fmt.pix_mp.width > sd_width ||
> +                       rrzo_fmt->fmt.pix_mp.height > sd_height,
> +                       "rrz out:%d:%d in:%d:%d",
> +                       rrzo_fmt->fmt.pix_mp.width,
> +                       rrzo_fmt->fmt.pix_mp.height,
> +                       sd_width,
> +                       sd_height);
> +
> +             dev_dbg(dev, "rrzo pixel_byte:%d img_fmt:0x%x\n",
> +                     config_param.cfg_resize_param.output.pixel_byte,
> +                     config_param.cfg_resize_param.output.img_fmt);
> +             dev_dbg(dev,
> +                     "rrzo 
> param:size=%0dx%0d,stride:%d,xsize:%d,crop=%0dx%0d\n",
> +                     config_param.cfg_resize_param.output.size.w,
> +                     config_param.cfg_resize_param.output.size.h,
> +                     config_param.cfg_resize_param.output.size.stride,
> +                     config_param.cfg_resize_param.output.size.xsize,
> +                     config_param.cfg_resize_param.output.crop.width,
> +                     config_param.cfg_resize_param.output.crop.height);
> +     } else {
> +             config_param.cfg_resize_param.bypass = 1;
> +     }
> +
> +     /* Configure meta DMAs info. */
> +     config_param.cfg_meta_param.enabled_meta_dmas = enable_dma_ports;
> +
> +     isp_composer_hw_config(isp_ctx, &config_param);
> +
> +     dev_dbg(dev, "%s done\n", __func__);
> +     return 0;
> +}
> +
> +int mtk_isp_enqueue(struct device *dev,
> +                 unsigned int dma_port,
> +                 struct mtk_cam_dev_buffer *buffer)
> +{
> +     struct isp_p1_device *p1_dev = get_p1_device(dev);
> +     struct mtk_isp_p1_ctx *isp_ctx = &p1_dev->isp_ctx;
> +     struct mtk_isp_scp_p1_cmd frameparams;
> +
> +     memset(&frameparams, 0, sizeof(frameparams));
> +
> +     frameparams.cmd_id = ISP_CMD_ENQUEUE_META;
> +     frameparams.meta_frame.enabled_dma = dma_port;
> +     frameparams.meta_frame.vb_index = buffer->vbb.vb2_buf.index;
> +     frameparams.meta_frame.meta_addr.iova = buffer->daddr;
> +     frameparams.meta_frame.meta_addr.scp_addr = buffer->scp_addr;
> +
> +     isp_composer_enqueue(isp_ctx, &frameparams, SCP_ISP_CMD);
> +
> +     return 0;
> +}
> +
> +int mtk_isp_req_enqueue(struct device *dev,
> +                     struct mtk_cam_dev_start_param *frameparamsbase)
> +{
> +     struct isp_p1_device *p1_dev = get_p1_device(dev);
> +     struct mtk_isp_p1_ctx *isp_ctx = &p1_dev->isp_ctx;
> +     struct p1_frame_param frameparams;
> +     struct mtk_isp_queue_job *framejob;
> +     struct mtk_cam_dev_buffer **bundle_buffers;
> +     unsigned int i, idx;
> +
> +     framejob = kzalloc(sizeof(*framejob), GFP_ATOMIC);
> +     memset(framejob, 0, sizeof(*framejob));
> +     memset(&frameparams, 0, sizeof(frameparams));
> +     INIT_LIST_HEAD(&framejob->list_buf);
> +
> +     bundle_buffers = &frameparamsbase->buffers[0];
> +     frameparams.frame_seq_no = isp_ctx->frame_seq_no++;
> +     frameparams.sof_idx =
> +             p1_dev->isp_devs[isp_ctx->isp_hw_module].sof_count;
> +     framejob->request_fd = frameparamsbase->request_fd;
> +     framejob->frame_seq_no = frameparams.frame_seq_no;
> +
> +     idx = MTK_CAM_P1_META_IN_0;
> +     if (bundle_buffers[idx]) {
> +             frameparams.tuning_addr.iova =
> +                     bundle_buffers[idx]->daddr;
> +             frameparams.tuning_addr.scp_addr =
> +                     bundle_buffers[idx]->scp_addr;
> +             list_add_tail(&bundle_buffers[idx]->list,
> +                           &framejob->list_buf);
> +     }
> +
> +     /* Image output */
> +     idx = MTK_CAM_P1_MAIN_STREAM_OUT;
> +     if (bundle_buffers[idx]) {
> +             frameparams.img_dma_buffers[0].buffer.iova =
> +                     bundle_buffers[idx]->daddr;
> +             frameparams.img_dma_buffers[0].buffer.scp_addr =
> +                     bundle_buffers[idx]->scp_addr;
> +             dev_dbg(dev, "main stream address iova:0x%x\n",
> +                     frameparams.img_dma_buffers[0].buffer.iova);
> +             list_add_tail(&bundle_buffers[idx]->list,
> +                           &framejob->list_buf);
> +     }
> +
> +     /* Resize output */
> +     idx = MTK_CAM_P1_PACKED_BIN_OUT;
> +     if (bundle_buffers[idx]) {
> +             frameparams.img_dma_buffers[1].buffer.iova =
> +                     bundle_buffers[idx]->daddr;
> +             frameparams.img_dma_buffers[1].buffer.scp_addr =
> +                     bundle_buffers[idx]->scp_addr;
> +             dev_dbg(dev, "packed out address iova:0x%x\n",
> +                     frameparams.img_dma_buffers[1].buffer.iova);
> +             list_add_tail(&bundle_buffers[idx]->list,
> +                           &framejob->list_buf);
> +     }
> +
> +     /* Meta output DMAs */
> +     for (i = 0; i < MAX_META_DMA_NODES; i++) {
> +             idx = MTK_CAM_P1_META_OUT_0 + i;
> +             if (bundle_buffers[idx]) {
> +                     frameparams.meta_addrs[i].iova =
> +                       bundle_buffers[idx]->daddr;
> +                     frameparams.meta_addrs[i].scp_addr =
> +                       bundle_buffers[idx]->scp_addr;
> +                     list_add_tail(&bundle_buffers[idx]->list,
> +                                   &framejob->list_buf);
> +             } else {
> +                     frameparams.meta_addrs[i].iova = 0;
> +                     frameparams.meta_addrs[i].scp_addr = 0;
> +             }
> +     }
> +
> +     spin_lock(&isp_ctx->p1_enqueue_list.lock);
> +     list_add_tail(&framejob->list_entry, &isp_ctx->p1_enqueue_list.queue);
> +     atomic_inc(&isp_ctx->p1_enqueue_list.queue_cnt);
> +     spin_unlock(&isp_ctx->p1_enqueue_list.lock);
> +
> +     isp_composer_enqueue(isp_ctx, &frameparams, SCP_ISP_FRAME);
> +     dev_dbg(dev, "request fd:%d frame_seq_no:%d is queued cnt:%d\n",
> +             frameparamsbase->request_fd,
> +             frameparams.frame_seq_no,
> +             atomic_read(&isp_ctx->p1_enqueue_list.queue_cnt));
> +
> +     return 0;
> +}
> +
> +static const struct dev_pm_ops mtk_isp_pm_ops = {
> +     SET_SYSTEM_SLEEP_PM_OPS(mtk_isp_suspend, mtk_isp_resume)
> +     SET_RUNTIME_PM_OPS(mtk_isp_suspend, mtk_isp_resume, NULL)
> +};
> +
> +static struct platform_driver mtk_isp_driver = {
> +     .probe   = mtk_isp_probe,
> +     .remove  = mtk_isp_remove,
> +     .driver  = {
> +             .name  = "mtk-cam",
> +             .of_match_table = of_match_ptr(mtk_isp_of_ids),
> +             .pm     = &mtk_isp_pm_ops,
> +     }
> +};
> +
> +module_platform_driver(mtk_isp_driver);
> +
> +MODULE_DESCRIPTION("Camera ISP driver");
> +MODULE_LICENSE("GPL");
> diff --git a/drivers/media/platform/mtk-isp/isp_50/cam/mtk_cam.h 
> b/drivers/media/platform/mtk-isp/isp_50/cam/mtk_cam.h
> new file mode 100644
> index 000000000000..6cf8bb4ba93a
> --- /dev/null
> +++ b/drivers/media/platform/mtk-isp/isp_50/cam/mtk_cam.h
> @@ -0,0 +1,300 @@
> +/* SPDX-License-Identifier: GPL-2.0 */
> +/*
> + * Copyright (c) 2018 MediaTek Inc.
> + * Author: Ryan Yu <ryan...@mediatek.com>
> + *
> + * This program is free software; you can redistribute it and/or modify
> + * it under the terms of the GNU General Public License version 2 as
> + * published by the Free Software Foundation.
> + *
> + * This program is distributed in the hope that it will be useful,
> + * but WITHOUT ANY WARRANTY; without even the implied warranty of
> + * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
> + * GNU General Public License for more details.
> + */
> +
> +#ifndef __CAMERA_ISP_H
> +#define __CAMERA_ISP_H
> +
> +#include <linux/cdev.h>
> +#include <linux/clk.h>
> +#include <linux/interrupt.h>
> +#include <linux/ioctl.h>
> +#include <linux/irqreturn.h>
> +#include <linux/miscdevice.h>
> +#include <linux/pm_qos.h>
> +#include <linux/scatterlist.h>
> +
> +#include "mtk_cam-dev.h"
> +#include "mtk_cam-scp.h"
> +
> +#define CAM_A_MAX_WIDTH              3328U
> +#define CAM_A_MAX_HEIGHT     2496U
> +#define CAM_B_MAX_WIDTH              5376U
> +#define CAM_B_MAX_HEIGHT     4032U
> +
> +#define CAM_MIN_WIDTH                80U
> +#define CAM_MIN_HEIGHT               60U
> +
> +#define IMG_MAX_WIDTH                CAM_B_MAX_WIDTH
> +#define IMG_MAX_HEIGHT               CAM_B_MAX_HEIGHT
> +#define IMG_MIN_WIDTH                CAM_MIN_WIDTH
> +#define IMG_MIN_HEIGHT               CAM_MIN_HEIGHT
> +
> +#define RRZ_MAX_WIDTH                CAM_B_MAX_WIDTH
> +#define RRZ_MAX_HEIGHT               CAM_B_MAX_HEIGHT
> +#define RRZ_MIN_WIDTH                CAM_MIN_WIDTH
> +#define RRZ_MIN_HEIGHT               CAM_MIN_HEIGHT
> +
> +#define R_IMGO               BIT(0)
> +#define R_RRZO               BIT(1)
> +#define R_AAO                BIT(3)
> +#define R_AFO                BIT(4)
> +#define R_LCSO               BIT(5)
> +#define R_PDO                BIT(6)
> +#define R_LMVO               BIT(7)
> +#define R_FLKO               BIT(8)
> +#define R_RSSO               BIT(9)
> +#define R_PSO                BIT(10)
> +
> +#define ISP_COMPOSING_MAX_NUM                4
> +#define ISP_FRAME_COMPOSING_MAX_NUM  3
> +
> +#define IRQ_DATA_BUF_SIZE            4
> +#define COMPOSRE_EVENT_BUF_SIZE              4
> +
> +#define CQ_ADDRESS_OFFSET            0x640
> +#define CQ_BUFFER_COUNT                      3

Please align macro values using tabs.

> +
> +#define IRQ_STAT_STR "cam%c, SOF_%d irq(0x%x), " \
> +                     "dma(0x%x), frame_num(%d)/cq_num(%d)\n"
> +
> +/*
> + * In order with the sequence of device nodes defined in dtsi rule,
> + * one hardware module should be mapping to one node.
> + */
> +enum isp_dev_node_enum {
> +     ISP_CAMSYS_CONFIG_IDX = 0,
> +     ISP_CAM_UNI_IDX,
> +     ISP_CAM_A_IDX,
> +     ISP_CAM_B_IDX,
> +     ISP_DEV_NODE_NUM
> +};
> +
> +/* Image RAW path for ISP P1 module. */
> +enum isp_raw_path_enum {
> +     ISP_PROCESS_RAW_PATH = 0,
> +     ISP_PURE_RAW_PATH
> +};
> +
> +enum {
> +     img_fmt_unknown         = 0x0000,
> +     img_fmt_raw_start       = 0x2200,
> +     img_fmt_bayer8          = img_fmt_raw_start,
> +     img_fmt_bayer10,
> +     img_fmt_bayer12,
> +     img_fmt_bayer14,
> +     img_fmt_fg_bayer8,
> +     img_fmt_fg_bayer10,
> +     img_fmt_fg_bayer12,
> +     img_fmt_fg_bayer14,
> +};
> +
> +enum {
> +     raw_pxl_id_b   = 0,
> +     raw_pxl_id_gb,
> +     raw_pxl_id_gr,
> +     raw_pxl_id_r
> +};
> +
> +enum {
> +     default_pixel_mode = 0,
> +     one_pixel_mode,
> +     two_pixel_mode,
> +     four_pixel_mode,
> +     pixel_mode_num,
> +};
> +
> +enum mtk_isp_scp_ipi_type {
> +     SCP_ISP_CMD = 0,
> +     SCP_ISP_FRAME,
> +};
> +
> +struct isp_queue {
> +     struct list_head queue;
> +     atomic_t queue_cnt;
> +     spinlock_t lock; /* queue attributes protection */
> +};
> +
> +struct isp_thread {
> +     struct task_struct *thread;
> +     wait_queue_head_t wq;
> +};
> +
> +struct mtk_isp_queue_work {
> +     union {
> +             struct mtk_isp_scp_p1_cmd cmd;
> +             struct p1_frame_param frameparams;
> +     };
> +     struct list_head list_entry;
> +     enum mtk_isp_scp_ipi_type type;
> +};
> +
> +struct mtk_cam_dev_stat_event_data {
> +     __u32 frame_seq_no;
> +     __u32 meta0_vb2_index;
> +     __u32 meta1_vb2_index;
> +     __u32 irq_status_mask;
> +     __u32 dma_status_mask;
> +};
> +
> +struct mtk_isp_queue_job {
> +     struct list_head list_entry;
> +     struct list_head list_buf;
> +     unsigned int request_fd;
> +     unsigned int frame_seq_no;
> +};
> +
> +struct isp_clk_struct {
> +     int num_clks;
> +     struct clk_bulk_data *clk_list;
> +};
> +
> +struct isp_device {
> +     struct device *dev;
> +     void __iomem *regs;
> +     int irq;
> +     spinlock_t spinlock_irq; /* ISP reg setting integrity */
> +     unsigned int current_frame;
> +     unsigned int meta0_vb2_index;
> +     unsigned int meta1_vb2_index;
> +     u8 sof_count;
> +     u8 isp_hw_module;
> +};
> +
> +struct mtk_isp_p1_ctx {
> +     struct isp_queue composer_txlist;
> +     struct isp_thread composer_tx_thread;
> +     atomic_t cmd_queued;
> +     struct mutex composer_tx_lock; /* isp composer work protection */
> +
> +     struct isp_thread composer_rx_thread;
> +     struct mtk_isp_scp_p1_cmd composer_evts[COMPOSRE_EVENT_BUF_SIZE];
> +     atomic_t composer_evts_start;
> +     atomic_t composer_evts_end;
> +     spinlock_t composer_evts_lock; /* SCP events protection */
> +     /* increase after ipi */
> +     atomic_t ipi_occupied;
> +     /* increase after frame enqueue */
> +     atomic_t composing_frame;
> +     /* current composed frame id */
> +     atomic_t composed_frame_id;
> +
> +     struct isp_queue p1_enqueue_list;
> +
> +     struct isp_thread isp_deque_thread;
> +     struct mtk_cam_dev_stat_event_data irq_event_datas[IRQ_DATA_BUF_SIZE];
> +     atomic_t irq_data_start;
> +     atomic_t irq_data_end;
> +     spinlock_t irq_dequeue_lock; /* ISP frame dequeuq protection */
> +
> +     dma_addr_t scp_mem_pa;
> +     dma_addr_t scp_mem_iova;
> +     struct sg_table sgtable;
> +
> +     /* increase after open, decrease when close */
> +     atomic_t isp_user_cnt;
> +     /* frame sequence number, increase per en-queue*/
> +     int frame_seq_no;
> +     unsigned int isp_hw_module;
> +     unsigned int isp_raw_path;
> +     unsigned int enable_dma_ports;
> +
> +     void (*composer_deinit_donecb)(void *isp_ctx);
> +
> +     struct list_head list;
> +};
> +
> +struct isp_p1_device {
> +     struct platform_device *pdev;
> +
> +     /* for SCP driver  */
> +     struct platform_device *scp_pdev;
> +     struct rproc *rproc_handle;
> +
> +     struct mtk_isp_p1_ctx isp_ctx;
> +     struct isp_clk_struct isp_clk;

What's the benefit of having mtk_isp_p1_ctx and isp_clk_struct in
separate structs? They are only ever used in isp_p1_device.

> +     struct mtk_cam_dev *cam_dev;
> +     struct isp_device *isp_devs;

Similarly, why are these allocated at runtime rather then just members
of the struct?

In mtk_isp_probe the struct isp_p1_device is allocated, and immediately
afterwards the struct mtk_cam_dev and struct isp_devices are. There will
only ever be ISP_DEV_NODE_NUM isp_devices.

Could this be changed to:
        struct mtk_cam_dev cam_dev;
        struct isp_device isp_devs[ISP_DEV_NODE_NUM];
?

> +};
> +
> +static inline struct isp_p1_device *
> +p1_ctx_to_dev(const struct mtk_isp_p1_ctx *__p1_ctx)
> +{
> +     return container_of(__p1_ctx, struct isp_p1_device, isp_ctx);
> +}
> +
> +static inline struct isp_p1_device *get_p1_device(struct device *dev)
> +{
> +     return ((struct isp_p1_device *)dev_get_drvdata(dev));
> +}
> +
> +int isp_composer_init(struct mtk_isp_p1_ctx *isp_ctx);
> +int isp_composer_hw_init(struct mtk_isp_p1_ctx *isp_ctx);
> +void isp_composer_meta_config(struct mtk_isp_p1_ctx *isp_ctx,
> +                           unsigned int dma);
> +void isp_composer_hw_config(struct mtk_isp_p1_ctx *isp_ctx,
> +                         struct p1_config_param *config_param);
> +void isp_composer_stream(struct mtk_isp_p1_ctx *isp_ctx, int on);
> +void isp_composer_hw_deinit(struct mtk_isp_p1_ctx *isp_ctx,
> +                         void (*donecb)(void *data));
> +void isp_composer_enqueue(struct mtk_isp_p1_ctx *isp_ctx,
> +                       void *data,
> +                       enum mtk_isp_scp_ipi_type type);

These functions are declared here, but implemented in mtk_cam-scp.c.
Can the funtion declarations be moved to mtk_cam-scp.h?

> +
> +/**
> + * mtk_isp_open - open isp driver and initialize related resources.
> + *
> + * @dev:     isp device.
> + *
> + */
> +int mtk_isp_open(struct device *dev);
> +
> +/**
> + * mtk_isp_release - release isp driver and related resources.
> + *
> + * @dev:     isp device.
> + *
> + */
> +int mtk_isp_release(struct device *dev);
> +
> +/**
> + * mtk_isp_config - output image & meta data configuration.
> + *
> + * @dev:     isp device.
> + *
> + */
> +int mtk_isp_config(struct device *dev);
> +
> +/**
> + * mtk_isp_req_enqueue - enqueue a frame bundle (per-frame basis) to ISP 
> driver.
> + *
> + * @dev:     isp device.
> + * @frameparamsbase: pointer to &struct mtk_cam_dev_start_param.
> + *
> + */
> +int mtk_isp_req_enqueue(struct device *dev,
> +                     struct mtk_cam_dev_start_param *frameparamsbase);
> +
> +/**
> + * mtk_isp_enqueue - enqueue a single frame to ISP driver
> + * for non-per-frame DMA.
> + *
> + * @dev:     isp device.
> + * @buffer: pointer to &struct mtk_cam_dev_buffer.
> + *
> + */
> +int mtk_isp_enqueue(struct device *dev,
> +                 unsigned int dma_idx,
> +                 struct mtk_cam_dev_buffer *buffer);
> +#endif /*__CAMERA_ISP_H*/
> -- 
> 2.18.0
> 

Reply via email to