Merge tag 'arm-dt-5.7' of git://git.kernel.org/pub/scm/linux/kernel/git/soc/soc
[linux-2.6-microblaze.git] / drivers / mailbox / mtk-cmdq-mailbox.c
index 9a6ce9f..b24822a 100644 (file)
@@ -56,7 +56,6 @@ struct cmdq_thread {
        void __iomem            *base;
        struct list_head        task_busy_list;
        u32                     priority;
-       bool                    atomic_exec;
 };
 
 struct cmdq_task {
@@ -162,48 +161,11 @@ static void cmdq_task_insert_into_thread(struct cmdq_task *task)
        cmdq_thread_invalidate_fetched_data(thread);
 }
 
-static bool cmdq_command_is_wfe(u64 cmd)
-{
-       u64 wfe_option = CMDQ_WFE_UPDATE | CMDQ_WFE_WAIT | CMDQ_WFE_WAIT_VALUE;
-       u64 wfe_op = (u64)(CMDQ_CODE_WFE << CMDQ_OP_CODE_SHIFT) << 32;
-       u64 wfe_mask = (u64)CMDQ_OP_CODE_MASK << 32 | 0xffffffff;
-
-       return ((cmd & wfe_mask) == (wfe_op | wfe_option));
-}
-
-/* we assume tasks in the same display GCE thread are waiting the same event. */
-static void cmdq_task_remove_wfe(struct cmdq_task *task)
-{
-       struct device *dev = task->cmdq->mbox.dev;
-       u64 *base = task->pkt->va_base;
-       int i;
-
-       dma_sync_single_for_cpu(dev, task->pa_base, task->pkt->cmd_buf_size,
-                               DMA_TO_DEVICE);
-       for (i = 0; i < CMDQ_NUM_CMD(task->pkt); i++)
-               if (cmdq_command_is_wfe(base[i]))
-                       base[i] = (u64)CMDQ_JUMP_BY_OFFSET << 32 |
-                                 CMDQ_JUMP_PASS;
-       dma_sync_single_for_device(dev, task->pa_base, task->pkt->cmd_buf_size,
-                                  DMA_TO_DEVICE);
-}
-
 static bool cmdq_thread_is_in_wfe(struct cmdq_thread *thread)
 {
        return readl(thread->base + CMDQ_THR_WAIT_TOKEN) & CMDQ_THR_IS_WAITING;
 }
 
-static void cmdq_thread_wait_end(struct cmdq_thread *thread,
-                                unsigned long end_pa)
-{
-       struct device *dev = thread->chan->mbox->dev;
-       unsigned long curr_pa;
-
-       if (readl_poll_timeout_atomic(thread->base + CMDQ_THR_CURR_ADDR,
-                       curr_pa, curr_pa == end_pa, 1, 20))
-               dev_err(dev, "GCE thread cannot run to end.\n");
-}
-
 static void cmdq_task_exec_done(struct cmdq_task *task, enum cmdq_cb_status sta)
 {
        struct cmdq_task_cb *cb = &task->pkt->async_cb;
@@ -383,36 +345,15 @@ static int cmdq_mbox_send_data(struct mbox_chan *chan, void *data)
                WARN_ON(cmdq_thread_suspend(cmdq, thread) < 0);
                curr_pa = readl(thread->base + CMDQ_THR_CURR_ADDR);
                end_pa = readl(thread->base + CMDQ_THR_END_ADDR);
-
-               /*
-                * Atomic execution should remove the following wfe, i.e. only
-                * wait event at first task, and prevent to pause when running.
-                */
-               if (thread->atomic_exec) {
-                       /* GCE is executing if command is not WFE */
-                       if (!cmdq_thread_is_in_wfe(thread)) {
-                               cmdq_thread_resume(thread);
-                               cmdq_thread_wait_end(thread, end_pa);
-                               WARN_ON(cmdq_thread_suspend(cmdq, thread) < 0);
-                               /* set to this task directly */
-                               writel(task->pa_base,
-                                      thread->base + CMDQ_THR_CURR_ADDR);
-                       } else {
-                               cmdq_task_insert_into_thread(task);
-                               cmdq_task_remove_wfe(task);
-                               smp_mb(); /* modify jump before enable thread */
-                       }
+               /* check boundary */
+               if (curr_pa == end_pa - CMDQ_INST_SIZE ||
+                   curr_pa == end_pa) {
+                       /* set to this task directly */
+                       writel(task->pa_base,
+                              thread->base + CMDQ_THR_CURR_ADDR);
                } else {
-                       /* check boundary */
-                       if (curr_pa == end_pa - CMDQ_INST_SIZE ||
-                           curr_pa == end_pa) {
-                               /* set to this task directly */
-                               writel(task->pa_base,
-                                      thread->base + CMDQ_THR_CURR_ADDR);
-                       } else {
-                               cmdq_task_insert_into_thread(task);
-                               smp_mb(); /* modify jump before enable thread */
-                       }
+                       cmdq_task_insert_into_thread(task);
+                       smp_mb(); /* modify jump before enable thread */
                }
                writel(task->pa_base + pkt->cmd_buf_size,
                       thread->base + CMDQ_THR_END_ADDR);
@@ -432,10 +373,62 @@ static void cmdq_mbox_shutdown(struct mbox_chan *chan)
 {
 }
 
+static int cmdq_mbox_flush(struct mbox_chan *chan, unsigned long timeout)
+{
+       struct cmdq_thread *thread = (struct cmdq_thread *)chan->con_priv;
+       struct cmdq_task_cb *cb;
+       struct cmdq_cb_data data;
+       struct cmdq *cmdq = dev_get_drvdata(chan->mbox->dev);
+       struct cmdq_task *task, *tmp;
+       unsigned long flags;
+       u32 enable;
+
+       spin_lock_irqsave(&thread->chan->lock, flags);
+       if (list_empty(&thread->task_busy_list))
+               goto out;
+
+       WARN_ON(cmdq_thread_suspend(cmdq, thread) < 0);
+       if (!cmdq_thread_is_in_wfe(thread))
+               goto wait;
+
+       list_for_each_entry_safe(task, tmp, &thread->task_busy_list,
+                                list_entry) {
+               cb = &task->pkt->async_cb;
+               if (cb->cb) {
+                       data.sta = CMDQ_CB_ERROR;
+                       data.data = cb->data;
+                       cb->cb(data);
+               }
+               list_del(&task->list_entry);
+               kfree(task);
+       }
+
+       cmdq_thread_resume(thread);
+       cmdq_thread_disable(cmdq, thread);
+       clk_disable(cmdq->clock);
+
+out:
+       spin_unlock_irqrestore(&thread->chan->lock, flags);
+       return 0;
+
+wait:
+       cmdq_thread_resume(thread);
+       spin_unlock_irqrestore(&thread->chan->lock, flags);
+       if (readl_poll_timeout_atomic(thread->base + CMDQ_THR_ENABLE_TASK,
+                                     enable, enable == 0, 1, timeout)) {
+               dev_err(cmdq->mbox.dev, "Fail to wait GCE thread 0x%x done\n",
+                       (u32)(thread->base - cmdq->base));
+
+               return -EFAULT;
+       }
+       return 0;
+}
+
 static const struct mbox_chan_ops cmdq_mbox_chan_ops = {
        .send_data = cmdq_mbox_send_data,
        .startup = cmdq_mbox_startup,
        .shutdown = cmdq_mbox_shutdown,
+       .flush = cmdq_mbox_flush,
 };
 
 static struct mbox_chan *cmdq_xlate(struct mbox_controller *mbox,
@@ -449,7 +442,6 @@ static struct mbox_chan *cmdq_xlate(struct mbox_controller *mbox,
 
        thread = (struct cmdq_thread *)mbox->chans[ind].con_priv;
        thread->priority = sp->args[1];
-       thread->atomic_exec = (sp->args[2] != 0);
        thread->chan = &mbox->chans[ind];
 
        return &mbox->chans[ind];