Commit dbff57bc authored by Alex Deucher's avatar Alex Deucher

drm/amdgpu/gfx: rework fiji cg functions so they can be shared

They can be shared with other asics with minor modifications.
Acked-by: default avatarTom St Denis <tom.stdenis@amd.com>
Signed-off-by: default avatarAlex Deucher <alexander.deucher@amd.com>
parent 79e5412c
...@@ -558,6 +558,7 @@ static const u32 stoney_mgcg_cgcg_init[] = ...@@ -558,6 +558,7 @@ static const u32 stoney_mgcg_cgcg_init[] =
static void gfx_v8_0_set_ring_funcs(struct amdgpu_device *adev); static void gfx_v8_0_set_ring_funcs(struct amdgpu_device *adev);
static void gfx_v8_0_set_irq_funcs(struct amdgpu_device *adev); static void gfx_v8_0_set_irq_funcs(struct amdgpu_device *adev);
static void gfx_v8_0_set_gds_init(struct amdgpu_device *adev); static void gfx_v8_0_set_gds_init(struct amdgpu_device *adev);
static void gfx_v8_0_set_rlc_funcs(struct amdgpu_device *adev);
static void gfx_v8_0_init_golden_registers(struct amdgpu_device *adev) static void gfx_v8_0_init_golden_registers(struct amdgpu_device *adev)
{ {
...@@ -4250,6 +4251,7 @@ static int gfx_v8_0_early_init(void *handle) ...@@ -4250,6 +4251,7 @@ static int gfx_v8_0_early_init(void *handle)
gfx_v8_0_set_ring_funcs(adev); gfx_v8_0_set_ring_funcs(adev);
gfx_v8_0_set_irq_funcs(adev); gfx_v8_0_set_irq_funcs(adev);
gfx_v8_0_set_gds_init(adev); gfx_v8_0_set_gds_init(adev);
gfx_v8_0_set_rlc_funcs(adev);
return 0; return 0;
} }
...@@ -4322,11 +4324,166 @@ static void gfx_v8_0_send_serdes_cmd(struct amdgpu_device *adev, ...@@ -4322,11 +4324,166 @@ static void gfx_v8_0_send_serdes_cmd(struct amdgpu_device *adev,
WREG32(mmRLC_SERDES_WR_CTRL, data); WREG32(mmRLC_SERDES_WR_CTRL, data);
} }
static void fiji_update_medium_grain_clock_gating(struct amdgpu_device *adev, #define MSG_ENTER_RLC_SAFE_MODE 1
bool enable) #define MSG_EXIT_RLC_SAFE_MODE 0
#define RLC_GPR_REG2__REQ_MASK 0x00000001
#define RLC_GPR_REG2__MESSAGE__SHIFT 0x00000001
#define RLC_GPR_REG2__MESSAGE_MASK 0x0000001e
static void cz_enter_rlc_safe_mode(struct amdgpu_device *adev)
{
u32 data = 0;
unsigned i;
data = RREG32(mmRLC_CNTL);
if ((data & RLC_CNTL__RLC_ENABLE_F32_MASK) == 0)
return;
if ((adev->cg_flags & (AMD_CG_SUPPORT_GFX_CGCG | AMD_CG_SUPPORT_GFX_MGCG)) ||
(adev->pg_flags & (AMD_PG_SUPPORT_GFX_PG | AMD_PG_SUPPORT_GFX_SMG |
AMD_PG_SUPPORT_GFX_DMG))) {
data |= RLC_GPR_REG2__REQ_MASK;
data &= ~RLC_GPR_REG2__MESSAGE_MASK;
data |= (MSG_ENTER_RLC_SAFE_MODE << RLC_GPR_REG2__MESSAGE__SHIFT);
WREG32(mmRLC_GPR_REG2, data);
for (i = 0; i < adev->usec_timeout; i++) {
if ((RREG32(mmRLC_GPM_STAT) &
(RLC_GPM_STAT__GFX_CLOCK_STATUS_MASK |
RLC_GPM_STAT__GFX_POWER_STATUS_MASK)) ==
(RLC_GPM_STAT__GFX_CLOCK_STATUS_MASK |
RLC_GPM_STAT__GFX_POWER_STATUS_MASK))
break;
udelay(1);
}
for (i = 0; i < adev->usec_timeout; i++) {
if ((RREG32(mmRLC_GPR_REG2) & RLC_GPR_REG2__REQ_MASK) == 0)
break;
udelay(1);
}
adev->gfx.rlc.in_safe_mode = true;
}
}
static void cz_exit_rlc_safe_mode(struct amdgpu_device *adev)
{
u32 data;
unsigned i;
data = RREG32(mmRLC_CNTL);
if ((data & RLC_CNTL__RLC_ENABLE_F32_MASK) == 0)
return;
if ((adev->cg_flags & (AMD_CG_SUPPORT_GFX_CGCG | AMD_CG_SUPPORT_GFX_MGCG)) ||
(adev->pg_flags & (AMD_PG_SUPPORT_GFX_PG | AMD_PG_SUPPORT_GFX_SMG |
AMD_PG_SUPPORT_GFX_DMG))) {
data |= RLC_GPR_REG2__REQ_MASK;
data &= ~RLC_GPR_REG2__MESSAGE_MASK;
data |= (MSG_EXIT_RLC_SAFE_MODE << RLC_GPR_REG2__MESSAGE__SHIFT);
WREG32(mmRLC_GPR_REG2, data);
adev->gfx.rlc.in_safe_mode = false;
}
for (i = 0; i < adev->usec_timeout; i++) {
if ((RREG32(mmRLC_GPR_REG2) & RLC_GPR_REG2__REQ_MASK) == 0)
break;
udelay(1);
}
}
static void iceland_enter_rlc_safe_mode(struct amdgpu_device *adev)
{
u32 data;
unsigned i;
data = RREG32(mmRLC_CNTL);
if (!(data & RLC_CNTL__RLC_ENABLE_F32_MASK))
return;
if (adev->cg_flags & (AMD_CG_SUPPORT_GFX_CGCG | AMD_CG_SUPPORT_GFX_MGCG)) {
data |= RLC_SAFE_MODE__CMD_MASK;
data &= ~RLC_SAFE_MODE__MESSAGE_MASK;
data |= (1 << RLC_SAFE_MODE__MESSAGE__SHIFT);
WREG32(mmRLC_SAFE_MODE, data);
for (i = 0; i < adev->usec_timeout; i++) {
if ((RREG32(mmRLC_GPM_STAT) &
(RLC_GPM_STAT__GFX_CLOCK_STATUS_MASK |
RLC_GPM_STAT__GFX_POWER_STATUS_MASK)) ==
(RLC_GPM_STAT__GFX_CLOCK_STATUS_MASK |
RLC_GPM_STAT__GFX_POWER_STATUS_MASK))
break;
udelay(1);
}
for (i = 0; i < adev->usec_timeout; i++) {
if ((RREG32(mmRLC_SAFE_MODE) & RLC_SAFE_MODE__CMD_MASK) == 0)
break;
udelay(1);
}
adev->gfx.rlc.in_safe_mode = true;
}
}
static void iceland_exit_rlc_safe_mode(struct amdgpu_device *adev)
{
u32 data = 0;
unsigned i;
data = RREG32(mmRLC_CNTL);
if (!(data & RLC_CNTL__RLC_ENABLE_F32_MASK))
return;
if (adev->cg_flags & (AMD_CG_SUPPORT_GFX_CGCG | AMD_CG_SUPPORT_GFX_MGCG)) {
if (adev->gfx.rlc.in_safe_mode) {
data |= RLC_SAFE_MODE__CMD_MASK;
data &= ~RLC_SAFE_MODE__MESSAGE_MASK;
WREG32(mmRLC_SAFE_MODE, data);
adev->gfx.rlc.in_safe_mode = false;
}
}
for (i = 0; i < adev->usec_timeout; i++) {
if ((RREG32(mmRLC_SAFE_MODE) & RLC_SAFE_MODE__CMD_MASK) == 0)
break;
udelay(1);
}
}
static void gfx_v8_0_nop_enter_rlc_safe_mode(struct amdgpu_device *adev)
{
adev->gfx.rlc.in_safe_mode = true;
}
static void gfx_v8_0_nop_exit_rlc_safe_mode(struct amdgpu_device *adev)
{
adev->gfx.rlc.in_safe_mode = false;
}
static const struct amdgpu_rlc_funcs cz_rlc_funcs = {
.enter_safe_mode = cz_enter_rlc_safe_mode,
.exit_safe_mode = cz_exit_rlc_safe_mode
};
static const struct amdgpu_rlc_funcs iceland_rlc_funcs = {
.enter_safe_mode = iceland_enter_rlc_safe_mode,
.exit_safe_mode = iceland_exit_rlc_safe_mode
};
static const struct amdgpu_rlc_funcs gfx_v8_0_nop_rlc_funcs = {
.enter_safe_mode = gfx_v8_0_nop_enter_rlc_safe_mode,
.exit_safe_mode = gfx_v8_0_nop_exit_rlc_safe_mode
};
static void gfx_v8_0_update_medium_grain_clock_gating(struct amdgpu_device *adev,
bool enable)
{ {
uint32_t temp, data; uint32_t temp, data;
adev->gfx.rlc.funcs->enter_safe_mode(adev);
/* It is disabled by HW by default */ /* It is disabled by HW by default */
if (enable && (adev->cg_flags & AMD_CG_SUPPORT_GFX_MGCG)) { if (enable && (adev->cg_flags & AMD_CG_SUPPORT_GFX_MGCG)) {
if (adev->cg_flags & AMD_CG_SUPPORT_GFX_MGLS) { if (adev->cg_flags & AMD_CG_SUPPORT_GFX_MGLS) {
...@@ -4349,10 +4506,15 @@ static void fiji_update_medium_grain_clock_gating(struct amdgpu_device *adev, ...@@ -4349,10 +4506,15 @@ static void fiji_update_medium_grain_clock_gating(struct amdgpu_device *adev,
/* 3 - RLC_CGTT_MGCG_OVERRIDE */ /* 3 - RLC_CGTT_MGCG_OVERRIDE */
temp = data = RREG32(mmRLC_CGTT_MGCG_OVERRIDE); temp = data = RREG32(mmRLC_CGTT_MGCG_OVERRIDE);
data &= ~(RLC_CGTT_MGCG_OVERRIDE__CPF_MASK | if (adev->flags & AMD_IS_APU)
RLC_CGTT_MGCG_OVERRIDE__RLC_MASK | data &= ~(RLC_CGTT_MGCG_OVERRIDE__CPF_MASK |
RLC_CGTT_MGCG_OVERRIDE__MGCG_MASK | RLC_CGTT_MGCG_OVERRIDE__RLC_MASK |
RLC_CGTT_MGCG_OVERRIDE__GRBM_MASK); RLC_CGTT_MGCG_OVERRIDE__MGCG_MASK);
else
data &= ~(RLC_CGTT_MGCG_OVERRIDE__CPF_MASK |
RLC_CGTT_MGCG_OVERRIDE__RLC_MASK |
RLC_CGTT_MGCG_OVERRIDE__MGCG_MASK |
RLC_CGTT_MGCG_OVERRIDE__GRBM_MASK);
if (temp != data) if (temp != data)
WREG32(mmRLC_CGTT_MGCG_OVERRIDE, data); WREG32(mmRLC_CGTT_MGCG_OVERRIDE, data);
...@@ -4424,15 +4586,19 @@ static void fiji_update_medium_grain_clock_gating(struct amdgpu_device *adev, ...@@ -4424,15 +4586,19 @@ static void fiji_update_medium_grain_clock_gating(struct amdgpu_device *adev,
/* 7- wait for RLC_SERDES_CU_MASTER & RLC_SERDES_NONCU_MASTER idle */ /* 7- wait for RLC_SERDES_CU_MASTER & RLC_SERDES_NONCU_MASTER idle */
gfx_v8_0_wait_for_rlc_serdes(adev); gfx_v8_0_wait_for_rlc_serdes(adev);
} }
adev->gfx.rlc.funcs->exit_safe_mode(adev);
} }
static void fiji_update_coarse_grain_clock_gating(struct amdgpu_device *adev, static void gfx_v8_0_update_coarse_grain_clock_gating(struct amdgpu_device *adev,
bool enable) bool enable)
{ {
uint32_t temp, temp1, data, data1; uint32_t temp, temp1, data, data1;
temp = data = RREG32(mmRLC_CGCG_CGLS_CTRL); temp = data = RREG32(mmRLC_CGCG_CGLS_CTRL);
adev->gfx.rlc.funcs->enter_safe_mode(adev);
if (enable && (adev->cg_flags & AMD_CG_SUPPORT_GFX_CGCG)) { if (enable && (adev->cg_flags & AMD_CG_SUPPORT_GFX_CGCG)) {
/* 1 enable cntx_empty_int_enable/cntx_busy_int_enable/ /* 1 enable cntx_empty_int_enable/cntx_busy_int_enable/
* Cmp_busy/GFX_Idle interrupts * Cmp_busy/GFX_Idle interrupts
...@@ -4509,22 +4675,24 @@ static void fiji_update_coarse_grain_clock_gating(struct amdgpu_device *adev, ...@@ -4509,22 +4675,24 @@ static void fiji_update_coarse_grain_clock_gating(struct amdgpu_device *adev,
if (temp != data) if (temp != data)
WREG32(mmRLC_CGCG_CGLS_CTRL, data); WREG32(mmRLC_CGCG_CGLS_CTRL, data);
} }
adev->gfx.rlc.funcs->exit_safe_mode(adev);
} }
static int fiji_update_gfx_clock_gating(struct amdgpu_device *adev, static int gfx_v8_0_update_gfx_clock_gating(struct amdgpu_device *adev,
bool enable) bool enable)
{ {
if (enable) { if (enable) {
/* CGCG/CGLS should be enabled after MGCG/MGLS/TS(CG/LS) /* CGCG/CGLS should be enabled after MGCG/MGLS/TS(CG/LS)
* === MGCG + MGLS + TS(CG/LS) === * === MGCG + MGLS + TS(CG/LS) ===
*/ */
fiji_update_medium_grain_clock_gating(adev, enable); gfx_v8_0_update_medium_grain_clock_gating(adev, enable);
fiji_update_coarse_grain_clock_gating(adev, enable); gfx_v8_0_update_coarse_grain_clock_gating(adev, enable);
} else { } else {
/* CGCG/CGLS should be disabled before MGCG/MGLS/TS(CG/LS) /* CGCG/CGLS should be disabled before MGCG/MGLS/TS(CG/LS)
* === CGCG + CGLS === * === CGCG + CGLS ===
*/ */
fiji_update_coarse_grain_clock_gating(adev, enable); gfx_v8_0_update_coarse_grain_clock_gating(adev, enable);
fiji_update_medium_grain_clock_gating(adev, enable); gfx_v8_0_update_medium_grain_clock_gating(adev, enable);
} }
return 0; return 0;
} }
...@@ -4536,8 +4704,10 @@ static int gfx_v8_0_set_clockgating_state(void *handle, ...@@ -4536,8 +4704,10 @@ static int gfx_v8_0_set_clockgating_state(void *handle,
switch (adev->asic_type) { switch (adev->asic_type) {
case CHIP_FIJI: case CHIP_FIJI:
fiji_update_gfx_clock_gating(adev, case CHIP_CARRIZO:
state == AMD_CG_STATE_GATE ? true : false); case CHIP_STONEY:
gfx_v8_0_update_gfx_clock_gating(adev,
state == AMD_CG_STATE_GATE ? true : false);
break; break;
default: default:
break; break;
...@@ -5137,6 +5307,22 @@ static void gfx_v8_0_set_irq_funcs(struct amdgpu_device *adev) ...@@ -5137,6 +5307,22 @@ static void gfx_v8_0_set_irq_funcs(struct amdgpu_device *adev)
adev->gfx.priv_inst_irq.funcs = &gfx_v8_0_priv_inst_irq_funcs; adev->gfx.priv_inst_irq.funcs = &gfx_v8_0_priv_inst_irq_funcs;
} }
static void gfx_v8_0_set_rlc_funcs(struct amdgpu_device *adev)
{
switch (adev->asic_type) {
case CHIP_TOPAZ:
case CHIP_STONEY:
adev->gfx.rlc.funcs = &iceland_rlc_funcs;
break;
case CHIP_CARRIZO:
adev->gfx.rlc.funcs = &cz_rlc_funcs;
break;
default:
adev->gfx.rlc.funcs = &gfx_v8_0_nop_rlc_funcs;
break;
}
}
static void gfx_v8_0_set_gds_init(struct amdgpu_device *adev) static void gfx_v8_0_set_gds_init(struct amdgpu_device *adev)
{ {
/* init asci gds info */ /* init asci gds info */
......
Markdown is supported
0%
or
You are about to add 0 people to the discussion. Proceed with caution.
Finish editing this message first!
Please register or to comment