Dear Brian,

I appreciate your comments.


On Wed, 2019-02-27 at 19:24 -0800, Brian Norris wrote:
> Hi Frederic,
> 
> On Sat, Feb 23, 2019 at 02:18:54PM +0800, Frederic Chen wrote:
> > Dear Brian,
> > 
> > I appreciate your comments. I'm really sorry for the delay in responding
> > to the comments due to some mail subscribing failed issue inside my company.
> 
> No problem.
> 
> > On Thu, 2019-02-21 at 21:36 +0800, Jungo Lin wrote:
> > > On Thu, 2019-02-07 at 11:08 -0800, Brian Norris wrote:
> > > > On Fri, Feb 01, 2019 at 07:21:31PM +0800, Frederic Chen wrote:
> 
> > > > > +static void dip_submit_worker(struct work_struct *work)
> > > > > +{
> > > > > +     struct mtk_dip_submit_work *dip_submit_work =
> > > > > +             container_of(work, struct mtk_dip_submit_work, 
> > > > > frame_work);
> > > > > +
> > > > > +     struct mtk_dip_hw_ctx  *dip_ctx = dip_submit_work->dip_ctx;
> > > > > +     struct mtk_dip_work *dip_work;
> > > > > +     struct dip_device *dip_dev;
> > > > > +     struct dip_subframe *buf;
> > > > > +     u32 len, num;
> > > > > +     int ret;
> > > > > +
> > > > > +     dip_dev = container_of(dip_ctx, struct dip_device, dip_ctx);
> > > > > +     num  = atomic_read(&dip_ctx->num_composing);
> > > > > +
> > > > > +     mutex_lock(&dip_ctx->dip_worklist.queuelock);
> > > > > +     dip_work = list_first_entry(&dip_ctx->dip_worklist.queue,
> > > > > +                                 struct mtk_dip_work, list_entry);
> > > > > +     mutex_unlock(&dip_ctx->dip_worklist.queuelock);
> > > > 
> > > > I see you grab the head of the list here, but then you release the lock.
> > > > Then you later assume that reference is still valid, throughout this
> > > > function.
> > > > 
> > > > That's usually true, because you only remove/delete entries from this
> > > > list within this same workqueue (at the end of this function). But it's
> > > > not true in dip_release_context() (which doesn't even grab the lock,
> > > > BTW).
> > > > 
> > > > I think there could be several ways to solve this, but judging by how
> > > > this list entry is used...couldn't you just remove it from the list
> > > > here, while holding the lock? Then you only have to kfree() it when
> > > > you're done under the free_work_list label.
> > > > 
> > 
> > I see. I would like to modify the codes as following:
> > 
> > mutex_lock(&dip_ctx->dip_useridlist.queuelock);
> 
> You missed the part where you get the head of the list:
> 
>       dip_work = list_first_entry(...);
> 
> But otherwise mostly looks OK.
> 
> > dip_work->user_id->num--;
> 
> Why do you need to do that with the queuelock held? Once you remove this
> work item from the list (safely under the lock), shouldn't you be the
> only one accessing it?
> 
> (Note, I don't actually know what that 'num' really means. I'm just
> looking at basic driver mechanics.)
> 

Yes, there is only one user of the dip work at that time.

I made a mistake on the usage of dip_useridlist.queuelock and 
dip_worklist.queuelock here. What I would like to do is to decrease the
total number of the frames of the user, which is protected by
dip_useridlist.queuelock. (user_id->num saves the total number of the
dip frames belongs to a user; the user may be the preview or capture
context.)

On the other hand, the list of dip work is protected by another lock, 
dip_worklist.queuelock.

In regarding to that point, I would like change the codes as following:

mutex_lock(&dip_ctx->dip_worklist.queuelock);
dip_work = list_first_entry(&dip_ctx->dip_worklist.queue,
                            struct mtk_dip_work, list_entry);
list_del(&dip_work->list_entry);
dip_ctx->dip_worklist.queue_cnt--;
len = dip_ctx->dip_worklist.queue_cnt;
mutex_unlock(&dip_ctx->dip_worklist.queuelock);

/* If the frame's user (preview or capture device) */
/* is in stream off state, */
/* return and release the buffers of the frame */
mutex_lock(&dip_ctx->dip_useridlist.queuelock);
if (dip_work->user_id->state == DIP_STATE_STREAMOFF) {
        dip_work->user_id->num--;
        mutex_unlock(&dip_ctx->dip_useridlist.queuelock);

        dip_work->frameparams.state = FRAME_STATE_STREAMOFF;
        call_mtk_dip_ctx_finish(dip_dev, &dip_work->frameparams);

        goto free_work_list;
mutex_unlock(&dip_ctx->dip_useridlist.queuelock);

> > list_del(&dip_work->list_entry);
> > dip_ctx->dip_worklist.queue_cnt--;
> > len = dip_ctx->dip_worklist.queue_cnt;
> > mutex_unlock(&dip_ctx->dip_useridlist.queuelock);
> > 
> > goto free_work_list;
> > 
> > /* ...... */
> > 
> > free_work_list:
> >     kfree(dip_work);
> > 
> > > > > +
> > > > > +     mutex_lock(&dip_ctx->dip_useridlist.queuelock);
> > > > > +     if (dip_work->user_id->state == DIP_STATE_STREAMOFF) {
> > > > > +             mutex_unlock(&dip_ctx->dip_useridlist.queuelock);
> > > > > +
> > > > > +             dip_work->frameparams.state = FRAME_STATE_STREAMOFF;
> > > > > +             call_mtk_dip_ctx_finish(dip_dev, 
> > > > > &dip_work->frameparams);
> > > > > +
> > > > > +             mutex_lock(&dip_ctx->dip_useridlist.queuelock);
> > > > > +             dip_work->user_id->num--;
> > > > > +             dev_dbg(&dip_dev->pdev->dev,
> > > > > +                     "user_id(%x) is streamoff and num: %d, 
> > > > > frame_no(%d) index: %x\n",
> > > > > +                     dip_work->user_id->id, dip_work->user_id->num,
> > > > > +                     dip_work->frameparams.frame_no,
> > > > > +                     dip_work->frameparams.index);
> > > > > +             mutex_unlock(&dip_ctx->dip_useridlist.queuelock);
> > > > > +
> > > > > +             goto free_work_list;
> > > > > +     }
> > > > > +     mutex_unlock(&dip_ctx->dip_useridlist.queuelock);
> > > > > +
> > > > > +     while (num >= DIP_COMPOSING_MAX_NUM) {
> > > > > +             ret = wait_event_interruptible_timeout
> > > > > +                     (dip_ctx->composing_wq,
> > > > > +                      (num < DIP_COMPOSING_MAX_NUM),
> > > > > +                      msecs_to_jiffies(DIP_COMPOSING_WQ_TIMEOUT));
> > > > > +
> > > > > +             if (ret == -ERESTARTSYS)
> > > > > +                     dev_err(&dip_dev->pdev->dev,
> > > > > +                             "interrupted by a signal!\n");
> > > > > +             else if (ret == 0)
> > > > > +                     dev_dbg(&dip_dev->pdev->dev,
> > > > > +                             "timeout frame_no(%d), num: %d\n",
> > > > > +                             dip_work->frameparams.frame_no, num);
> > > > > +             else
> > > > > +                     dev_dbg(&dip_dev->pdev->dev,
> > > > > +                             "wakeup frame_no(%d), num: %d\n",
> > > > > +                             dip_work->frameparams.frame_no, num);
> > > > > +
> > > > > +             num = atomic_read(&dip_ctx->num_composing);
> > > > > +     };
> > > > > +
> > > > > +     mutex_lock(&dip_ctx->dip_freebufferlist.queuelock);
> > > > > +     if (list_empty(&dip_ctx->dip_freebufferlist.queue)) {
> > > > > +             mutex_unlock(&dip_ctx->dip_freebufferlist.queuelock);
> > > > > +
> > > > > +             dev_err(&dip_dev->pdev->dev,
> > > > > +                     "frame_no(%d) index: %x no free buffer: %d\n",
> > > > > +                     dip_work->frameparams.frame_no,
> > > > > +                     dip_work->frameparams.index,
> > > > > +                     dip_ctx->dip_freebufferlist.queue_cnt);
> > > > > +
> > > > > +             /* Call callback to notify V4L2 common framework
> > > > > +              * for failure of enqueue
> > > > > +              */
> > > > > +             dip_work->frameparams.state = FRAME_STATE_ERROR;
> > > > > +             call_mtk_dip_ctx_finish(dip_dev, 
> > > > > &dip_work->frameparams);
> > > > > +
> > > > > +             mutex_lock(&dip_ctx->dip_useridlist.queuelock);
> > > > > +             dip_work->user_id->num--;
> > > > > +             mutex_unlock(&dip_ctx->dip_useridlist.queuelock);
> > > > > +
> > > > > +             goto free_work_list;
> > > > > +     }
> > > > > +
> > > > > +     buf = list_first_entry(&dip_ctx->dip_freebufferlist.queue,
> > > > > +                            struct dip_subframe,
> > > > > +                            list_entry);
> > > > > +     list_del(&buf->list_entry);
> > > > > +     dip_ctx->dip_freebufferlist.queue_cnt--;
> > > > > +     mutex_unlock(&dip_ctx->dip_freebufferlist.queuelock);
> > > > > +
> > > > > +     mutex_lock(&dip_ctx->dip_usedbufferlist.queuelock);
> > > > > +     list_add_tail(&buf->list_entry, 
> > > > > &dip_ctx->dip_usedbufferlist.queue);
> > > > > +     dip_ctx->dip_usedbufferlist.queue_cnt++;
> > > > > +     mutex_unlock(&dip_ctx->dip_usedbufferlist.queuelock);
> > > > > +
> > > > > +     memcpy(&dip_work->frameparams.subfrm_data,
> > > > > +            &buf->buffer, sizeof(buf->buffer));
> > > > > +
> > > > > +     memset((char *)buf->buffer.va, 0, DIP_SUB_FRM_SZ);
> > > > > +
> > > > > +     memcpy(&dip_work->frameparams.config_data,
> > > > > +            &buf->config_data, sizeof(buf->config_data));
> > > > > +
> > > > > +     memset((char *)buf->config_data.va, 0, DIP_COMP_SZ);
> > > > > +
> > > > > +     if (dip_work->frameparams.tuning_data.pa == 0) {
> > > > > +             dev_dbg(&dip_dev->pdev->dev,
> > > > > +                     "frame_no(%d) has no tuning_data\n",
> > > > > +                     dip_work->frameparams.frame_no);
> > > > > +
> > > > > +             memcpy(&dip_work->frameparams.tuning_data,
> > > > > +                    &buf->tuning_buf, sizeof(buf->tuning_buf));
> > > > > +
> > > > > +             memset((char *)buf->tuning_buf.va, 0, DIP_TUNING_SZ);
> > > > > +             /* When user enqueued without tuning buffer,
> > > > > +              * it would use driver internal buffer.
> > > > > +              * So, tuning_data.va should be 0
> > > > > +              */
> > > > > +             dip_work->frameparams.tuning_data.va = 0;
> > > > > +     }
> > > > > +
> > > > > +     dip_work->frameparams.drv_data = (u64)dip_ctx;
> > > > > +     dip_work->frameparams.state = FRAME_STATE_COMPOSING;
> > > > > +
> > > > > +     memcpy((void *)buf->frameparam.va, &dip_work->frameparams,
> > > > > +            sizeof(dip_work->frameparams));
> > > > > +
> > > > > +     dip_send(dip_ctx->vpu_pdev, IPI_DIP_FRAME,
> > > > > +              (void *)&dip_work->frameparams,
> > > > > +              sizeof(dip_work->frameparams), 0);
> > > > > +     num = atomic_inc_return(&dip_ctx->num_composing);
> > > > > +
> > > > > +free_work_list:
> > > > > +
> > > > > +     mutex_lock(&dip_ctx->dip_worklist.queuelock);
> > > > > +     list_del(&dip_work->list_entry);
> > > > > +     dip_ctx->dip_worklist.queue_cnt--;
> > > > > +     len = dip_ctx->dip_worklist.queue_cnt;
> > > > > +     mutex_unlock(&dip_ctx->dip_worklist.queuelock);
> > > > > +
> > > > > +     dev_dbg(&dip_dev->pdev->dev,
> > > > > +             "frame_no(%d) index: %x, worklist count: %d, composing 
> > > > > num: %d\n",
> > > > > +             dip_work->frameparams.frame_no, 
> > > > > dip_work->frameparams.index,
> > > > > +             len, num);
> > > > > +
> > > > > +     kfree(dip_work);
> > > > > +}
> 
> 
> > > > > +int dip_release_context(struct dip_device *dip_dev)
> > > > 
> > > > Should be static.
> > > > 
> > 
> > I will change it to static.
> > 
> > > > > +{
> > > > > +     u32 i = 0;
> > > > > +     struct dip_subframe *buf, *tmpbuf;
> > > > > +     struct mtk_dip_work *dip_work, *tmp_work;
> > > > > +     struct dip_user_id  *dip_userid, *tmp_id;
> > > > > +     struct mtk_dip_hw_ctx *dip_ctx;
> > > > > +
> > > > > +     dip_ctx = &dip_dev->dip_ctx;
> > > > > +     dev_dbg(&dip_dev->pdev->dev, "composer work queue = %d\n",
> > > > > +             dip_ctx->dip_worklist.queue_cnt);
> > > > > +
> > > > > +     list_for_each_entry_safe(dip_work, tmp_work,
> > > > > +                              &dip_ctx->dip_worklist.queue,
> > > > > +                              list_entry) {
> > > > 
> > > > Shouldn't you be holding the mutex for this? Or alternatively, cancel
> > > > any outstanding work and move the flush_workqueue()/destroy_workqueue()
> > > > up.
> > > > 
> > > > Similar questions for the other lists we're going through here.
> > > > 
> > 
> > We missed the mutex holding here. I would like to change the codes as 
> > following:
> > 
> > mutex_lock(&dip_ctx->dip_worklist.queuelock);
> > list_for_each_entry_safe(dip_work, tmp_work,
> >                      &dip_ctx->dip_worklist.queue,
> >                      list_entry) {
> >     list_del(&dip_work->list_entry);
> >     dip_ctx->dip_worklist.queue_cnt--;
> >     kfree(dip_work);
> > }
> > mutex_unlock(&dip_ctx->dip_worklist.queuelock);
> > 
> > I will also modify dip_useridlist and dip_ctx->dip_freebufferlist 
> > parts like dip_ctx->dip_worklist.
> 
> Seems about right.
> 
> Brian
> 
> > > > > +             list_del(&dip_work->list_entry);
> > > > > +             dev_dbg(&dip_dev->pdev->dev, "dip work frame no: %d\n",
> > > > > +                     dip_work->frameparams.frame_no);
> > > > > +             kfree(dip_work);
> > > > > +             dip_ctx->dip_worklist.queue_cnt--;
> > > > > +     }
> > > > > +
> > > > > +     if (dip_ctx->dip_worklist.queue_cnt != 0)
> > > > > +             dev_dbg(&dip_dev->pdev->dev,
> > > > > +                     "dip_worklist is not empty (%d)\n",
> > > > > +                     dip_ctx->dip_worklist.queue_cnt);
> > > > > +
> > > > > +     list_for_each_entry_safe(dip_userid, tmp_id,
> > > > > +                              &dip_ctx->dip_useridlist.queue,
> > > > > +                              list_entry) {
> > > > > +             list_del(&dip_userid->list_entry);
> > > > > +             dev_dbg(&dip_dev->pdev->dev, "dip user id: %x\n",
> > > > > +                     dip_userid->id);
> > > > > +             kfree(dip_userid);
> > > > > +             dip_ctx->dip_useridlist.queue_cnt--;
> > > > > +     }
> > > > > +
> > > > > +     if (dip_ctx->dip_useridlist.queue_cnt != 0)
> > > > > +             dev_dbg(&dip_dev->pdev->dev,
> > > > > +                     "dip_useridlist is not empty (%d)\n",
> > > > > +                     dip_ctx->dip_useridlist.queue_cnt);
> > > > > +
> > > > > +     flush_workqueue(dip_ctx->mdpcb_workqueue);
> > > > > +     destroy_workqueue(dip_ctx->mdpcb_workqueue);
> > > > > +     dip_ctx->mdpcb_workqueue = NULL;
> > > > > +
> > > > > +     flush_workqueue(dip_ctx->composer_wq);
> > > > > +     destroy_workqueue(dip_ctx->composer_wq);
> > > > > +     dip_ctx->composer_wq = NULL;
> > > > > +
> > > > > +     atomic_set(&dip_ctx->num_composing, 0);
> > > > > +     atomic_set(&dip_ctx->num_running, 0);
> > > > > +
> > > > > +     kthread_stop(dip_ctx->dip_runner_thread.thread);
> > > > > +     dip_ctx->dip_runner_thread.thread = NULL;
> > > > > +
> > > > > +     atomic_set(&dip_ctx->dip_user_cnt, 0);
> > > > > +     atomic_set(&dip_ctx->dip_stream_cnt, 0);
> > > > > +     atomic_set(&dip_ctx->dip_enque_cnt, 0);
> > > > > +
> > > > > +     /* All the buffer should be in the freebufferlist when release 
> > > > > */
> > > > > +     list_for_each_entry_safe(buf, tmpbuf,
> > > > > +                              &dip_ctx->dip_freebufferlist.queue,
> > > > > +                              list_entry) {
> > > > > +             struct sg_table *sgt = &buf->table;
> > > > > +
> > > > > +             dev_dbg(&dip_dev->pdev->dev,
> > > > > +                     "buf pa (%d): %x\n", i, buf->buffer.pa);
> > > > > +             dip_ctx->dip_freebufferlist.queue_cnt--;
> > > > > +             dma_unmap_sg_attrs(&dip_dev->pdev->dev, sgt->sgl,
> > > > > +                                sgt->orig_nents,
> > > > > +                                DMA_BIDIRECTIONAL, 
> > > > > DMA_ATTR_SKIP_CPU_SYNC);
> > > > > +             sg_free_table(sgt);
> > > > > +             list_del(&buf->list_entry);
> > > > > +             kfree(buf);
> > > > > +             buf = NULL;
> > > > > +             i++;
> > > > > +     }
> > > > > +
> > > > > +     if (dip_ctx->dip_freebufferlist.queue_cnt != 0 &&
> > > > > +         i != DIP_SUB_FRM_DATA_NUM)
> > > > > +             dev_err(&dip_dev->pdev->dev,
> > > > > +                     "dip_freebufferlist is not empty (%d/%d)\n",
> > > > > +                     dip_ctx->dip_freebufferlist.queue_cnt, i);
> > > > > +
> > > > > +     mutex_destroy(&dip_ctx->dip_useridlist.queuelock);
> > > > > +     mutex_destroy(&dip_ctx->dip_worklist.queuelock);
> > > > > +     mutex_destroy(&dip_ctx->dip_usedbufferlist.queuelock);
> > > > > +     mutex_destroy(&dip_ctx->dip_freebufferlist.queuelock);
> > > > > +
> > > > > +     return 0;
> > > > > +}
> > > > > +

Sincerely,

Frederic Chen

Reply via email to