drm/amdgpu: Prevent race between late signaled fences and GPU reset.
authorAndrey Grodzovsky <andrey.grodzovsky@amd.com>
Sat, 18 Jun 2022 04:28:50 +0000 (00:28 -0400)
committerAlex Deucher <alexander.deucher@amd.com>
Tue, 28 Jun 2022 15:24:24 +0000 (11:24 -0400)
Problem:
After we start handling timed out jobs we assume there fences won't be
signaled but we cannot be sure and sometimes they fire late. We need
to prevent concurrent accesses to fence array from
amdgpu_fence_driver_clear_job_fences during GPU reset and amdgpu_fence_process
from a late EOP interrupt.

Fix:
Before accessing fence array in GPU disable EOP interrupt and flush
all pending interrupt handlers for amdgpu device's interrupt line.

v2: Switch from irq_get/put to full enable/disable_irq for amdgpu

Signed-off-by: Andrey Grodzovsky <andrey.grodzovsky@amd.com>
Acked-by: Christian König <christian.koenig@amd.com>
Signed-off-by: Alex Deucher <alexander.deucher@amd.com>
drivers/gpu/drm/amd/amdgpu/amdgpu_device.c
drivers/gpu/drm/amd/amdgpu/amdgpu_fence.c
drivers/gpu/drm/amd/amdgpu/amdgpu_ring.h

index f2a4c26..9d2395a 100644 (file)
@@ -4606,6 +4606,8 @@ int amdgpu_device_pre_asic_reset(struct amdgpu_device *adev,
                amdgpu_virt_fini_data_exchange(adev);
        }
 
+       amdgpu_fence_driver_isr_toggle(adev, true);
+
        /* block all schedulers and reset given job's ring */
        for (i = 0; i < AMDGPU_MAX_RINGS; ++i) {
                struct amdgpu_ring *ring = adev->rings[i];
@@ -4621,6 +4623,8 @@ int amdgpu_device_pre_asic_reset(struct amdgpu_device *adev,
                amdgpu_fence_driver_force_completion(ring);
        }
 
+       amdgpu_fence_driver_isr_toggle(adev, false);
+
        if (job && job->vm)
                drm_sched_increase_karma(&job->base);
 
index a9ae3be..c1d04ea 100644 (file)
@@ -532,6 +532,24 @@ void amdgpu_fence_driver_hw_fini(struct amdgpu_device *adev)
        }
 }
 
+/* Will either stop and flush handlers for amdgpu interrupt or reanble it */
+void amdgpu_fence_driver_isr_toggle(struct amdgpu_device *adev, bool stop)
+{
+       int i;
+
+       for (i = 0; i < AMDGPU_MAX_RINGS; i++) {
+               struct amdgpu_ring *ring = adev->rings[i];
+
+               if (!ring || !ring->fence_drv.initialized || !ring->fence_drv.irq_src)
+                       continue;
+
+               if (stop)
+                       disable_irq(adev->irq.irq);
+               else
+                       enable_irq(adev->irq.irq);
+       }
+}
+
 void amdgpu_fence_driver_sw_fini(struct amdgpu_device *adev)
 {
        unsigned int i, j;
index 7d89a52..82c178a 100644 (file)
@@ -143,6 +143,7 @@ signed long amdgpu_fence_wait_polling(struct amdgpu_ring *ring,
                                      uint32_t wait_seq,
                                      signed long timeout);
 unsigned amdgpu_fence_count_emitted(struct amdgpu_ring *ring);
+void amdgpu_fence_driver_isr_toggle(struct amdgpu_device *adev, bool stop);
 
 /*
  * Rings.