habanalabs: Rename hw_queues_mirror to cs_mirror
authorTomer Tayar <ttayar@habana.ai>
Fri, 30 Oct 2020 09:16:23 +0000 (11:16 +0200)
committerOded Gabbay <ogabbay@kernel.org>
Mon, 30 Nov 2020 08:47:34 +0000 (10:47 +0200)
Future command submission types might be submitted to HW not via the
QMAN queues path. However, it would be still required to have the TDR
mechanism for these CS, and thus the patch renames the TDR fields and
replaces the hw_queues_ prefix with cs_.

Signed-off-by: Tomer Tayar <ttayar@habana.ai>
Reviewed-by: Oded Gabbay <ogabbay@kernel.org>
Signed-off-by: Oded Gabbay <ogabbay@kernel.org>
drivers/misc/habanalabs/common/command_submission.c
drivers/misc/habanalabs/common/device.c
drivers/misc/habanalabs/common/habanalabs.h
drivers/misc/habanalabs/common/hw_queue.c

index 20b34fb..749ec6c 100644 (file)
@@ -343,25 +343,24 @@ static void cs_do_release(struct kref *ref)
        /* Need to update CI for internal queues */
        hl_int_hw_queue_update_ci(cs);
 
-       spin_lock(&hdev->hw_queues_mirror_lock);
-       /* remove CS from hw_queues mirror list */
+       /* remove CS from CS mirror list */
+       spin_lock(&hdev->cs_mirror_lock);
        list_del_init(&cs->mirror_node);
-       spin_unlock(&hdev->hw_queues_mirror_lock);
+       spin_unlock(&hdev->cs_mirror_lock);
 
        /* Don't cancel TDR in case this CS was timedout because we might be
         * running from the TDR context
         */
-       if (!cs->timedout &&
-                       hdev->timeout_jiffies != MAX_SCHEDULE_TIMEOUT) {
+       if (!cs->timedout && hdev->timeout_jiffies != MAX_SCHEDULE_TIMEOUT) {
                struct hl_cs *next;
 
                if (cs->tdr_active)
                        cancel_delayed_work_sync(&cs->work_tdr);
 
-               spin_lock(&hdev->hw_queues_mirror_lock);
+               spin_lock(&hdev->cs_mirror_lock);
 
                /* queue TDR for next CS */
-               next = list_first_entry_or_null(&hdev->hw_queues_mirror_list,
+               next = list_first_entry_or_null(&hdev->cs_mirror_list,
                                                struct hl_cs, mirror_node);
 
                if (next && !next->tdr_active) {
@@ -370,7 +369,7 @@ static void cs_do_release(struct kref *ref)
                                                hdev->timeout_jiffies);
                }
 
-               spin_unlock(&hdev->hw_queues_mirror_lock);
+               spin_unlock(&hdev->cs_mirror_lock);
        }
 
 out:
@@ -534,8 +533,7 @@ void hl_cs_rollback_all(struct hl_device *hdev)
                flush_workqueue(hdev->cq_wq[i]);
 
        /* Make sure we don't have leftovers in the H/W queues mirror list */
-       list_for_each_entry_safe(cs, tmp, &hdev->hw_queues_mirror_list,
-                               mirror_node) {
+       list_for_each_entry_safe(cs, tmp, &hdev->cs_mirror_list, mirror_node) {
                cs_get(cs);
                cs->aborted = true;
                dev_warn_ratelimited(hdev->dev, "Killing CS %d.%llu\n",
index 348faf3..ce0a127 100644 (file)
@@ -343,8 +343,8 @@ static int device_early_init(struct hl_device *hdev)
        mutex_init(&hdev->send_cpu_message_lock);
        mutex_init(&hdev->debug_lock);
        mutex_init(&hdev->mmu_cache_lock);
-       INIT_LIST_HEAD(&hdev->hw_queues_mirror_list);
-       spin_lock_init(&hdev->hw_queues_mirror_lock);
+       INIT_LIST_HEAD(&hdev->cs_mirror_list);
+       spin_lock_init(&hdev->cs_mirror_lock);
        INIT_LIST_HEAD(&hdev->fpriv_list);
        mutex_init(&hdev->fpriv_list_lock);
        atomic_set(&hdev->in_reset, 0);
index 40b566c..eb968c3 100644 (file)
@@ -1699,8 +1699,8 @@ struct hl_mmu_funcs {
  * @eq_wq: work queue of event queue for executing work in process context.
  * @kernel_ctx: Kernel driver context structure.
  * @kernel_queues: array of hl_hw_queue.
- * @hw_queues_mirror_list: CS mirror list for TDR.
- * @hw_queues_mirror_lock: protects hw_queues_mirror_list.
+ * @cs_mirror_list: CS mirror list for TDR.
+ * @cs_mirror_lock: protects cs_mirror_list.
  * @kernel_cb_mgr: command buffer manager for creating/destroying/handling CGs.
  * @event_queue: event queue for IRQ from CPU-CP.
  * @dma_pool: DMA pool for small allocations.
@@ -1810,8 +1810,8 @@ struct hl_device {
        struct workqueue_struct         *eq_wq;
        struct hl_ctx                   *kernel_ctx;
        struct hl_hw_queue              *kernel_queues;
-       struct list_head                hw_queues_mirror_list;
-       spinlock_t                      hw_queues_mirror_lock;
+       struct list_head                cs_mirror_list;
+       spinlock_t                      cs_mirror_lock;
        struct hl_cb_mgr                kernel_cb_mgr;
        struct hl_eq                    event_queue;
        struct dma_pool                 *dma_pool;
index f9550fc..d1d30fb 100644 (file)
@@ -578,20 +578,20 @@ int hl_hw_queue_schedule_cs(struct hl_cs *cs)
        else if (cs->type == CS_TYPE_COLLECTIVE_WAIT)
                hdev->asic_funcs->collective_wait_init_cs(cs);
 
-       spin_lock(&hdev->hw_queues_mirror_lock);
-       list_add_tail(&cs->mirror_node, &hdev->hw_queues_mirror_list);
+       spin_lock(&hdev->cs_mirror_lock);
+       list_add_tail(&cs->mirror_node, &hdev->cs_mirror_list);
 
        /* Queue TDR if the CS is the first entry and if timeout is wanted */
        if ((hdev->timeout_jiffies != MAX_SCHEDULE_TIMEOUT) &&
-                       (list_first_entry(&hdev->hw_queues_mirror_list,
+                       (list_first_entry(&hdev->cs_mirror_list,
                                        struct hl_cs, mirror_node) == cs)) {
                cs->tdr_active = true;
                schedule_delayed_work(&cs->work_tdr, hdev->timeout_jiffies);
-               spin_unlock(&hdev->hw_queues_mirror_lock);
-       } else {
-               spin_unlock(&hdev->hw_queues_mirror_lock);
+
        }
 
+       spin_unlock(&hdev->cs_mirror_lock);
+
        if (!hdev->cs_active_cnt++) {
                struct hl_device_idle_busy_ts *ts;