drm/amdgpu: move subfunctions to the front of vce_v2_0.c.
authorRex Zhu <Rex.Zhu@amd.com>
Thu, 26 Jan 2017 08:46:22 +0000 (16:46 +0800)
committerAlex Deucher <alexander.deucher@amd.com>
Wed, 8 Feb 2017 22:23:03 +0000 (17:23 -0500)
Signed-off-by: Rex Zhu <Rex.Zhu@amd.com>
Acked-by: Christian König <christian.koenig@amd.com>
Reviewed-by: Alex Deucher <alexander.deucher@amd.com>
Signed-off-by: Alex Deucher <alexander.deucher@amd.com>
drivers/gpu/drm/amd/amdgpu/vce_v2_0.c

index 13c7949..9ea9934 100644 (file)
 #define VCE_V2_0_DATA_SIZE     (23552 * AMDGPU_MAX_VCE_HANDLES)
 #define VCE_STATUS_VCPU_REPORT_FW_LOADED_MASK  0x02
 
-static void vce_v2_0_mc_resume(struct amdgpu_device *adev);
 static void vce_v2_0_set_ring_funcs(struct amdgpu_device *adev);
 static void vce_v2_0_set_irq_funcs(struct amdgpu_device *adev);
-static int vce_v2_0_wait_for_idle(void *handle);
-static void vce_v2_0_init_cg(struct amdgpu_device *adev);
-static void vce_v2_0_disable_cg(struct amdgpu_device *adev);
-static void vce_v2_0_enable_mgcg(struct amdgpu_device *adev, bool enable,
-                                                               bool sw_cg);
 
 /**
  * vce_v2_0_ring_get_rptr - get read pointer
@@ -145,6 +139,86 @@ static int vce_v2_0_firmware_loaded(struct amdgpu_device *adev)
        return -ETIMEDOUT;
 }
 
+static void vce_v2_0_disable_cg(struct amdgpu_device *adev)
+{
+       WREG32(mmVCE_CGTT_CLK_OVERRIDE, 7);
+}
+
+static void vce_v2_0_init_cg(struct amdgpu_device *adev)
+{
+       u32 tmp;
+
+       tmp = RREG32(mmVCE_CLOCK_GATING_A);
+       tmp &= ~0xfff;
+       tmp |= ((0 << 0) | (4 << 4));
+       tmp |= 0x40000;
+       WREG32(mmVCE_CLOCK_GATING_A, tmp);
+
+       tmp = RREG32(mmVCE_UENC_CLOCK_GATING);
+       tmp &= ~0xfff;
+       tmp |= ((0 << 0) | (4 << 4));
+       WREG32(mmVCE_UENC_CLOCK_GATING, tmp);
+
+       tmp = RREG32(mmVCE_CLOCK_GATING_B);
+       tmp |= 0x10;
+       tmp &= ~0x100000;
+       WREG32(mmVCE_CLOCK_GATING_B, tmp);
+}
+
+static void vce_v2_0_mc_resume(struct amdgpu_device *adev)
+{
+       uint64_t addr = adev->vce.gpu_addr;
+       uint32_t size;
+
+       WREG32_P(mmVCE_CLOCK_GATING_A, 0, ~(1 << 16));
+       WREG32_P(mmVCE_UENC_CLOCK_GATING, 0x1FF000, ~0xFF9FF000);
+       WREG32_P(mmVCE_UENC_REG_CLOCK_GATING, 0x3F, ~0x3F);
+       WREG32(mmVCE_CLOCK_GATING_B, 0xf7);
+
+       WREG32(mmVCE_LMI_CTRL, 0x00398000);
+       WREG32_P(mmVCE_LMI_CACHE_CTRL, 0x0, ~0x1);
+       WREG32(mmVCE_LMI_SWAP_CNTL, 0);
+       WREG32(mmVCE_LMI_SWAP_CNTL1, 0);
+       WREG32(mmVCE_LMI_VM_CTRL, 0);
+
+       addr += AMDGPU_VCE_FIRMWARE_OFFSET;
+       size = VCE_V2_0_FW_SIZE;
+       WREG32(mmVCE_VCPU_CACHE_OFFSET0, addr & 0x7fffffff);
+       WREG32(mmVCE_VCPU_CACHE_SIZE0, size);
+
+       addr += size;
+       size = VCE_V2_0_STACK_SIZE;
+       WREG32(mmVCE_VCPU_CACHE_OFFSET1, addr & 0x7fffffff);
+       WREG32(mmVCE_VCPU_CACHE_SIZE1, size);
+
+       addr += size;
+       size = VCE_V2_0_DATA_SIZE;
+       WREG32(mmVCE_VCPU_CACHE_OFFSET2, addr & 0x7fffffff);
+       WREG32(mmVCE_VCPU_CACHE_SIZE2, size);
+
+       WREG32_P(mmVCE_LMI_CTRL2, 0x0, ~0x100);
+       WREG32_FIELD(VCE_SYS_INT_EN, VCE_SYS_INT_TRAP_INTERRUPT_EN, 1);
+}
+
+static bool vce_v2_0_is_idle(void *handle)
+{
+       struct amdgpu_device *adev = (struct amdgpu_device *)handle;
+
+       return !(RREG32(mmSRBM_STATUS2) & SRBM_STATUS2__VCE_BUSY_MASK);
+}
+
+static int vce_v2_0_wait_for_idle(void *handle)
+{
+       struct amdgpu_device *adev = (struct amdgpu_device *)handle;
+       unsigned i;
+
+       for (i = 0; i < adev->usec_timeout; i++) {
+               if (vce_v2_0_is_idle(handle))
+                       return 0;
+       }
+       return -ETIMEDOUT;
+}
+
 /**
  * vce_v2_0_start - start VCE block
  *
@@ -245,6 +319,97 @@ static int vce_v2_0_stop(struct amdgpu_device *adev)
        return 0;
 }
 
+static void vce_v2_0_set_sw_cg(struct amdgpu_device *adev, bool gated)
+{
+       u32 tmp;
+
+       if (gated) {
+               tmp = RREG32(mmVCE_CLOCK_GATING_B);
+               tmp |= 0xe70000;
+               WREG32(mmVCE_CLOCK_GATING_B, tmp);
+
+               tmp = RREG32(mmVCE_UENC_CLOCK_GATING);
+               tmp |= 0xff000000;
+               WREG32(mmVCE_UENC_CLOCK_GATING, tmp);
+
+               tmp = RREG32(mmVCE_UENC_REG_CLOCK_GATING);
+               tmp &= ~0x3fc;
+               WREG32(mmVCE_UENC_REG_CLOCK_GATING, tmp);
+
+               WREG32(mmVCE_CGTT_CLK_OVERRIDE, 0);
+       } else {
+               tmp = RREG32(mmVCE_CLOCK_GATING_B);
+               tmp |= 0xe7;
+               tmp &= ~0xe70000;
+               WREG32(mmVCE_CLOCK_GATING_B, tmp);
+
+               tmp = RREG32(mmVCE_UENC_CLOCK_GATING);
+               tmp |= 0x1fe000;
+               tmp &= ~0xff000000;
+               WREG32(mmVCE_UENC_CLOCK_GATING, tmp);
+
+               tmp = RREG32(mmVCE_UENC_REG_CLOCK_GATING);
+               tmp |= 0x3fc;
+               WREG32(mmVCE_UENC_REG_CLOCK_GATING, tmp);
+       }
+}
+
+static void vce_v2_0_set_dyn_cg(struct amdgpu_device *adev, bool gated)
+{
+       u32 orig, tmp;
+
+/* LMI_MC/LMI_UMC always set in dynamic,
+ * set {CGC_*_GATE_MODE, CGC_*_SW_GATE} = {0, 0}
+ */
+       tmp = RREG32(mmVCE_CLOCK_GATING_B);
+       tmp &= ~0x00060006;
+
+/* Exception for ECPU, IH, SEM, SYS blocks needs to be turned on/off by SW */
+       if (gated) {
+               tmp |= 0xe10000;
+               WREG32(mmVCE_CLOCK_GATING_B, tmp);
+       } else {
+               tmp |= 0xe1;
+               tmp &= ~0xe10000;
+               WREG32(mmVCE_CLOCK_GATING_B, tmp);
+       }
+
+       orig = tmp = RREG32(mmVCE_UENC_CLOCK_GATING);
+       tmp &= ~0x1fe000;
+       tmp &= ~0xff000000;
+       if (tmp != orig)
+               WREG32(mmVCE_UENC_CLOCK_GATING, tmp);
+
+       orig = tmp = RREG32(mmVCE_UENC_REG_CLOCK_GATING);
+       tmp &= ~0x3fc;
+       if (tmp != orig)
+               WREG32(mmVCE_UENC_REG_CLOCK_GATING, tmp);
+
+       /* set VCE_UENC_REG_CLOCK_GATING always in dynamic mode */
+       WREG32(mmVCE_UENC_REG_CLOCK_GATING, 0x00);
+
+       if(gated)
+               WREG32(mmVCE_CGTT_CLK_OVERRIDE, 0);
+}
+
+static void vce_v2_0_enable_mgcg(struct amdgpu_device *adev, bool enable,
+                                                               bool sw_cg)
+{
+       if (enable && (adev->cg_flags & AMD_CG_SUPPORT_VCE_MGCG)) {
+               if (sw_cg)
+                       vce_v2_0_set_sw_cg(adev, true);
+               else
+                       vce_v2_0_set_dyn_cg(adev, true);
+       } else {
+               vce_v2_0_disable_cg(adev);
+
+               if (sw_cg)
+                       vce_v2_0_set_sw_cg(adev, false);
+               else
+                       vce_v2_0_set_dyn_cg(adev, false);
+       }
+}
+
 static int vce_v2_0_early_init(void *handle)
 {
        struct amdgpu_device *adev = (struct amdgpu_device *)handle;
@@ -365,177 +530,6 @@ static int vce_v2_0_resume(void *handle)
        return r;
 }
 
-static void vce_v2_0_set_sw_cg(struct amdgpu_device *adev, bool gated)
-{
-       u32 tmp;
-
-       if (gated) {
-               tmp = RREG32(mmVCE_CLOCK_GATING_B);
-               tmp |= 0xe70000;
-               WREG32(mmVCE_CLOCK_GATING_B, tmp);
-
-               tmp = RREG32(mmVCE_UENC_CLOCK_GATING);
-               tmp |= 0xff000000;
-               WREG32(mmVCE_UENC_CLOCK_GATING, tmp);
-
-               tmp = RREG32(mmVCE_UENC_REG_CLOCK_GATING);
-               tmp &= ~0x3fc;
-               WREG32(mmVCE_UENC_REG_CLOCK_GATING, tmp);
-
-               WREG32(mmVCE_CGTT_CLK_OVERRIDE, 0);
-       } else {
-               tmp = RREG32(mmVCE_CLOCK_GATING_B);
-               tmp |= 0xe7;
-               tmp &= ~0xe70000;
-               WREG32(mmVCE_CLOCK_GATING_B, tmp);
-
-               tmp = RREG32(mmVCE_UENC_CLOCK_GATING);
-               tmp |= 0x1fe000;
-               tmp &= ~0xff000000;
-               WREG32(mmVCE_UENC_CLOCK_GATING, tmp);
-
-               tmp = RREG32(mmVCE_UENC_REG_CLOCK_GATING);
-               tmp |= 0x3fc;
-               WREG32(mmVCE_UENC_REG_CLOCK_GATING, tmp);
-       }
-}
-
-static void vce_v2_0_set_dyn_cg(struct amdgpu_device *adev, bool gated)
-{
-       u32 orig, tmp;
-
-/* LMI_MC/LMI_UMC always set in dynamic,
- * set {CGC_*_GATE_MODE, CGC_*_SW_GATE} = {0, 0}
- */
-       tmp = RREG32(mmVCE_CLOCK_GATING_B);
-       tmp &= ~0x00060006;
-
-/* Exception for ECPU, IH, SEM, SYS blocks needs to be turned on/off by SW */
-       if (gated) {
-               tmp |= 0xe10000;
-               WREG32(mmVCE_CLOCK_GATING_B, tmp);
-       } else {
-               tmp |= 0xe1;
-               tmp &= ~0xe10000;
-               WREG32(mmVCE_CLOCK_GATING_B, tmp);
-       }
-
-       orig = tmp = RREG32(mmVCE_UENC_CLOCK_GATING);
-       tmp &= ~0x1fe000;
-       tmp &= ~0xff000000;
-       if (tmp != orig)
-               WREG32(mmVCE_UENC_CLOCK_GATING, tmp);
-
-       orig = tmp = RREG32(mmVCE_UENC_REG_CLOCK_GATING);
-       tmp &= ~0x3fc;
-       if (tmp != orig)
-               WREG32(mmVCE_UENC_REG_CLOCK_GATING, tmp);
-
-       /* set VCE_UENC_REG_CLOCK_GATING always in dynamic mode */
-       WREG32(mmVCE_UENC_REG_CLOCK_GATING, 0x00);
-
-       if(gated)
-               WREG32(mmVCE_CGTT_CLK_OVERRIDE, 0);
-}
-
-static void vce_v2_0_disable_cg(struct amdgpu_device *adev)
-{
-       WREG32(mmVCE_CGTT_CLK_OVERRIDE, 7);
-}
-
-static void vce_v2_0_enable_mgcg(struct amdgpu_device *adev, bool enable,
-                                                               bool sw_cg)
-{
-       if (enable && (adev->cg_flags & AMD_CG_SUPPORT_VCE_MGCG)) {
-               if (sw_cg)
-                       vce_v2_0_set_sw_cg(adev, true);
-               else
-                       vce_v2_0_set_dyn_cg(adev, true);
-       } else {
-               vce_v2_0_disable_cg(adev);
-
-               if (sw_cg)
-                       vce_v2_0_set_sw_cg(adev, false);
-               else
-                       vce_v2_0_set_dyn_cg(adev, false);
-       }
-}
-
-static void vce_v2_0_init_cg(struct amdgpu_device *adev)
-{
-       u32 tmp;
-
-       tmp = RREG32(mmVCE_CLOCK_GATING_A);
-       tmp &= ~0xfff;
-       tmp |= ((0 << 0) | (4 << 4));
-       tmp |= 0x40000;
-       WREG32(mmVCE_CLOCK_GATING_A, tmp);
-
-       tmp = RREG32(mmVCE_UENC_CLOCK_GATING);
-       tmp &= ~0xfff;
-       tmp |= ((0 << 0) | (4 << 4));
-       WREG32(mmVCE_UENC_CLOCK_GATING, tmp);
-
-       tmp = RREG32(mmVCE_CLOCK_GATING_B);
-       tmp |= 0x10;
-       tmp &= ~0x100000;
-       WREG32(mmVCE_CLOCK_GATING_B, tmp);
-}
-
-static void vce_v2_0_mc_resume(struct amdgpu_device *adev)
-{
-       uint64_t addr = adev->vce.gpu_addr;
-       uint32_t size;
-
-       WREG32_P(mmVCE_CLOCK_GATING_A, 0, ~(1 << 16));
-       WREG32_P(mmVCE_UENC_CLOCK_GATING, 0x1FF000, ~0xFF9FF000);
-       WREG32_P(mmVCE_UENC_REG_CLOCK_GATING, 0x3F, ~0x3F);
-       WREG32(mmVCE_CLOCK_GATING_B, 0xf7);
-
-       WREG32(mmVCE_LMI_CTRL, 0x00398000);
-       WREG32_P(mmVCE_LMI_CACHE_CTRL, 0x0, ~0x1);
-       WREG32(mmVCE_LMI_SWAP_CNTL, 0);
-       WREG32(mmVCE_LMI_SWAP_CNTL1, 0);
-       WREG32(mmVCE_LMI_VM_CTRL, 0);
-
-       addr += AMDGPU_VCE_FIRMWARE_OFFSET;
-       size = VCE_V2_0_FW_SIZE;
-       WREG32(mmVCE_VCPU_CACHE_OFFSET0, addr & 0x7fffffff);
-       WREG32(mmVCE_VCPU_CACHE_SIZE0, size);
-
-       addr += size;
-       size = VCE_V2_0_STACK_SIZE;
-       WREG32(mmVCE_VCPU_CACHE_OFFSET1, addr & 0x7fffffff);
-       WREG32(mmVCE_VCPU_CACHE_SIZE1, size);
-
-       addr += size;
-       size = VCE_V2_0_DATA_SIZE;
-       WREG32(mmVCE_VCPU_CACHE_OFFSET2, addr & 0x7fffffff);
-       WREG32(mmVCE_VCPU_CACHE_SIZE2, size);
-
-       WREG32_P(mmVCE_LMI_CTRL2, 0x0, ~0x100);
-       WREG32_FIELD(VCE_SYS_INT_EN, VCE_SYS_INT_TRAP_INTERRUPT_EN, 1);
-}
-
-static bool vce_v2_0_is_idle(void *handle)
-{
-       struct amdgpu_device *adev = (struct amdgpu_device *)handle;
-
-       return !(RREG32(mmSRBM_STATUS2) & SRBM_STATUS2__VCE_BUSY_MASK);
-}
-
-static int vce_v2_0_wait_for_idle(void *handle)
-{
-       struct amdgpu_device *adev = (struct amdgpu_device *)handle;
-       unsigned i;
-
-       for (i = 0; i < adev->usec_timeout; i++) {
-               if (vce_v2_0_is_idle(handle))
-                       return 0;
-       }
-       return -ETIMEDOUT;
-}
-
 static int vce_v2_0_soft_reset(void *handle)
 {
        struct amdgpu_device *adev = (struct amdgpu_device *)handle;