Commit e88bbc91 authored by Rob Clark's avatar Rob Clark

Revert "drm/msm/mdp5: provide dynamic bandwidth management"

This reverts commit c1d12c19.

Breaks armv7 build

  ERROR: modpost: "__aeabi_ldivmod" [drivers/gpu/drm/msm/msm.ko] undefined!
Signed-off-by: default avatarRob Clark <robdclark@chromium.org>
parent c1d12c19
......@@ -43,9 +43,6 @@ struct mdp5_crtc {
/* for unref'ing cursor bo's after scanout completes: */
struct drm_flip_work unref_cursor_work;
/* for lowering down the bandwidth after previous frame is complete */
struct drm_flip_work lower_bw_work;
struct mdp_irq vblank;
struct mdp_irq err;
struct mdp_irq pp_done;
......@@ -174,28 +171,12 @@ static void unref_cursor_worker(struct drm_flip_work *work, void *val)
drm_gem_object_put(val);
}
static void lower_bw_worker(struct drm_flip_work *work, void *val)
{
struct mdp5_crtc *mdp5_crtc =
container_of(work, struct mdp5_crtc, lower_bw_work);
struct drm_crtc *crtc = &mdp5_crtc->base;
struct mdp5_crtc_state *mdp5_cstate = to_mdp5_crtc_state(crtc->state);
struct mdp5_kms *mdp5_kms = get_kms(&mdp5_crtc->base);
if (mdp5_cstate->old_crtc_bw > mdp5_cstate->new_crtc_bw) {
DBG("DOWN BW to %lld\n", mdp5_cstate->new_crtc_bw);
mdp5_kms_set_bandwidth(mdp5_kms);
mdp5_cstate->old_crtc_bw = mdp5_cstate->new_crtc_bw;
}
}
static void mdp5_crtc_destroy(struct drm_crtc *crtc)
{
struct mdp5_crtc *mdp5_crtc = to_mdp5_crtc(crtc);
drm_crtc_cleanup(crtc);
drm_flip_work_cleanup(&mdp5_crtc->unref_cursor_work);
drm_flip_work_cleanup(&mdp5_crtc->lower_bw_work);
kfree(mdp5_crtc);
}
......@@ -710,7 +691,6 @@ static int mdp5_crtc_atomic_check(struct drm_crtc *crtc,
struct drm_crtc_state *crtc_state = drm_atomic_get_new_crtc_state(state,
crtc);
struct mdp5_kms *mdp5_kms = get_kms(crtc);
struct mdp5_crtc_state *mdp5_cstate = to_mdp5_crtc_state(crtc_state);
struct drm_plane *plane;
struct drm_device *dev = crtc->dev;
struct plane_state pstates[STAGE_MAX + 1];
......@@ -721,7 +701,6 @@ static int mdp5_crtc_atomic_check(struct drm_crtc *crtc,
bool need_right_mixer = false;
int cnt = 0, i;
int ret;
u64 crtc_bw = 0;
enum mdp_mixer_stage_id start;
DBG("%s: check", crtc->name);
......@@ -739,9 +718,6 @@ static int mdp5_crtc_atomic_check(struct drm_crtc *crtc,
*/
if (pstates[cnt].state->r_hwpipe)
need_right_mixer = true;
crtc_bw += pstates[cnt].state->plane_bw;
cnt++;
if (plane->type == DRM_PLANE_TYPE_CURSOR)
......@@ -754,10 +730,6 @@ static int mdp5_crtc_atomic_check(struct drm_crtc *crtc,
hw_cfg = mdp5_cfg_get_hw_config(mdp5_kms->cfg);
if (hw_cfg->perf.ab_inefficiency)
crtc_bw = mult_frac(crtc_bw, hw_cfg->perf.ab_inefficiency, 100);
mdp5_cstate->new_crtc_bw = crtc_bw;
/*
* we need a right hwmixer if the mode's width is greater than a single
* LM's max width
......@@ -813,7 +785,6 @@ static void mdp5_crtc_atomic_flush(struct drm_crtc *crtc,
{
struct mdp5_crtc *mdp5_crtc = to_mdp5_crtc(crtc);
struct mdp5_crtc_state *mdp5_cstate = to_mdp5_crtc_state(crtc->state);
struct mdp5_kms *mdp5_kms = get_kms(crtc);
struct drm_device *dev = crtc->dev;
unsigned long flags;
......@@ -837,12 +808,6 @@ static void mdp5_crtc_atomic_flush(struct drm_crtc *crtc,
blend_setup(crtc);
if (mdp5_cstate->old_crtc_bw < mdp5_cstate->new_crtc_bw) {
DBG("UP BW to %lld\n", mdp5_cstate->new_crtc_bw);
mdp5_kms_set_bandwidth(mdp5_kms);
mdp5_cstate->old_crtc_bw = mdp5_cstate->new_crtc_bw;
}
/* PP_DONE irq is only used by command mode for now.
* It is better to request pending before FLUSH and START trigger
* to make sure no pp_done irq missed.
......@@ -1190,7 +1155,6 @@ static void mdp5_crtc_vblank_irq(struct mdp_irq *irq, uint32_t irqstatus)
{
struct mdp5_crtc *mdp5_crtc = container_of(irq, struct mdp5_crtc, vblank);
struct drm_crtc *crtc = &mdp5_crtc->base;
struct mdp5_crtc_state *mdp5_cstate = to_mdp5_crtc_state(crtc->state);
struct msm_drm_private *priv = crtc->dev->dev_private;
unsigned pending;
......@@ -1198,11 +1162,6 @@ static void mdp5_crtc_vblank_irq(struct mdp_irq *irq, uint32_t irqstatus)
pending = atomic_xchg(&mdp5_crtc->pending, 0);
if (mdp5_cstate->old_crtc_bw > mdp5_cstate->new_crtc_bw) {
drm_flip_work_queue(&mdp5_crtc->lower_bw_work, NULL);
drm_flip_work_commit(&mdp5_crtc->lower_bw_work, priv->wq);
}
if (pending & PENDING_FLIP) {
complete_flip(crtc, NULL);
}
......@@ -1359,9 +1318,6 @@ struct drm_crtc *mdp5_crtc_init(struct drm_device *dev,
drm_flip_work_init(&mdp5_crtc->unref_cursor_work,
"unref cursor", unref_cursor_worker);
drm_flip_work_init(&mdp5_crtc->lower_bw_work,
"lower bw", lower_bw_worker);
drm_crtc_helper_add(crtc, &mdp5_crtc_helper_funcs);
return crtc;
......
......@@ -19,8 +19,6 @@
#include "msm_mmu.h"
#include "mdp5_kms.h"
#define MDP5_DEFAULT_BW MBps_to_icc(6400)
static int mdp5_hw_init(struct msm_kms *kms)
{
struct mdp5_kms *mdp5_kms = to_mdp5_kms(to_mdp_kms(kms));
......@@ -298,28 +296,6 @@ static const struct mdp_kms_funcs kms_funcs = {
.set_irqmask = mdp5_set_irqmask,
};
void mdp5_kms_set_bandwidth(struct mdp5_kms *mdp5_kms)
{
int i;
u32 full_bw = 0;
struct drm_crtc *tmp_crtc;
if (!mdp5_kms->num_paths)
return;
drm_for_each_crtc(tmp_crtc, mdp5_kms->dev) {
if (!tmp_crtc->enabled)
continue;
full_bw += Bps_to_icc(to_mdp5_crtc_state(tmp_crtc->state)->new_crtc_bw / mdp5_kms->num_paths);
}
DBG("SET BW to %d\n", full_bw);
for (i = 0; i < mdp5_kms->num_paths; i++)
icc_set_bw(mdp5_kms->paths[i], full_bw, full_bw);
}
static int mdp5_disable(struct mdp5_kms *mdp5_kms)
{
DBG("");
......@@ -337,14 +313,6 @@ static int mdp5_disable(struct mdp5_kms *mdp5_kms)
if (mdp5_kms->lut_clk)
clk_disable_unprepare(mdp5_kms->lut_clk);
if (!mdp5_kms->enable_count) {
int i;
for (i = 0; i < mdp5_kms->num_paths; i++)
icc_set_bw(mdp5_kms->paths[i], 0, 0);
icc_set_bw(mdp5_kms->path_rot, 0, 0);
}
return 0;
}
......@@ -354,14 +322,6 @@ static int mdp5_enable(struct mdp5_kms *mdp5_kms)
mdp5_kms->enable_count++;
if (mdp5_kms->enable_count == 1) {
int i;
for (i = 0; i < mdp5_kms->num_paths; i++)
icc_set_bw(mdp5_kms->paths[i], 0, MDP5_DEFAULT_BW);
icc_set_bw(mdp5_kms->path_rot, 0, MDP5_DEFAULT_BW);
}
clk_prepare_enable(mdp5_kms->ahb_clk);
clk_prepare_enable(mdp5_kms->axi_clk);
clk_prepare_enable(mdp5_kms->core_clk);
......@@ -868,40 +828,6 @@ static int interface_init(struct mdp5_kms *mdp5_kms)
return 0;
}
static int mdp5_setup_interconnect(struct mdp5_kms *mdp5_kms)
{
struct icc_path *path0 = of_icc_get(&mdp5_kms->pdev->dev, "mdp0-mem");
struct icc_path *path1 = of_icc_get(&mdp5_kms->pdev->dev, "mdp1-mem");
struct icc_path *path_rot = of_icc_get(&mdp5_kms->pdev->dev, "rotator-mem");
if (IS_ERR(path0))
return PTR_ERR(path0);
if (!path0) {
/* no interconnect support is not necessarily a fatal
* condition, the platform may simply not have an
* interconnect driver yet. But warn about it in case
* bootloader didn't setup bus clocks high enough for
* scanout.
*/
dev_warn(&mdp5_kms->pdev->dev, "No interconnect support may cause display underflows!\n");
return 0;
}
mdp5_kms->paths[0] = path0;
mdp5_kms->num_paths = 1;
if (!IS_ERR_OR_NULL(path1)) {
mdp5_kms->paths[1] = path1;
mdp5_kms->num_paths++;
}
if (!IS_ERR_OR_NULL(path_rot))
mdp5_kms->path_rot = path_rot;
return 0;
}
static int mdp5_init(struct platform_device *pdev, struct drm_device *dev)
{
struct msm_drm_private *priv = dev->dev_private;
......@@ -909,7 +835,6 @@ static int mdp5_init(struct platform_device *pdev, struct drm_device *dev)
struct mdp5_cfg *config;
u32 major, minor;
int ret;
int i;
mdp5_kms = devm_kzalloc(&pdev->dev, sizeof(*mdp5_kms), GFP_KERNEL);
if (!mdp5_kms) {
......@@ -934,14 +859,6 @@ static int mdp5_init(struct platform_device *pdev, struct drm_device *dev)
goto fail;
}
ret = mdp5_setup_interconnect(mdp5_kms);
if (ret)
goto fail;
for (i = 0; i < mdp5_kms->num_paths; i++)
icc_set_bw(mdp5_kms->paths[i], 0, MDP5_DEFAULT_BW);
icc_set_bw(mdp5_kms->path_rot, 0, MDP5_DEFAULT_BW);
/* mandatory clocks: */
ret = get_clk(pdev, &mdp5_kms->axi_clk, "bus", true);
if (ret)
......@@ -1051,10 +968,46 @@ static const struct component_ops mdp5_ops = {
.unbind = mdp5_unbind,
};
static int mdp5_setup_interconnect(struct platform_device *pdev)
{
struct icc_path *path0 = of_icc_get(&pdev->dev, "mdp0-mem");
struct icc_path *path1 = of_icc_get(&pdev->dev, "mdp1-mem");
struct icc_path *path_rot = of_icc_get(&pdev->dev, "rotator-mem");
if (IS_ERR(path0))
return PTR_ERR(path0);
if (!path0) {
/* no interconnect support is not necessarily a fatal
* condition, the platform may simply not have an
* interconnect driver yet. But warn about it in case
* bootloader didn't setup bus clocks high enough for
* scanout.
*/
dev_warn(&pdev->dev, "No interconnect support may cause display underflows!\n");
return 0;
}
icc_set_bw(path0, 0, MBps_to_icc(6400));
if (!IS_ERR_OR_NULL(path1))
icc_set_bw(path1, 0, MBps_to_icc(6400));
if (!IS_ERR_OR_NULL(path_rot))
icc_set_bw(path_rot, 0, MBps_to_icc(6400));
return 0;
}
static int mdp5_dev_probe(struct platform_device *pdev)
{
int ret;
DBG("");
ret = mdp5_setup_interconnect(pdev);
if (ret)
return ret;
return component_add(&pdev->dev, &mdp5_ops);
}
......
......@@ -17,7 +17,6 @@
#include "mdp5_ctl.h"
#include "mdp5_smp.h"
struct icc_path;
struct mdp5_kms {
struct mdp_kms base;
......@@ -69,10 +68,6 @@ struct mdp5_kms {
struct mdp_irq error_handler;
int enable_count;
int num_paths;
struct icc_path *paths[2];
struct icc_path *path_rot;
};
#define to_mdp5_kms(x) container_of(x, struct mdp5_kms, base)
......@@ -105,8 +100,6 @@ struct mdp5_plane_state {
/* assigned by crtc blender */
enum mdp_mixer_stage_id stage;
u64 plane_bw;
};
#define to_mdp5_plane_state(x) \
container_of(x, struct mdp5_plane_state, base)
......@@ -137,9 +130,6 @@ struct mdp5_crtc_state {
* writing CTL[n].START until encoder->enable()
*/
bool defer_start;
u64 new_crtc_bw;
u64 old_crtc_bw;
};
#define to_mdp5_crtc_state(x) \
container_of(x, struct mdp5_crtc_state, base)
......@@ -302,8 +292,6 @@ void mdp5_encoder_set_intf_mode(struct drm_encoder *encoder, bool cmd_mode);
int mdp5_encoder_get_linecount(struct drm_encoder *encoder);
u32 mdp5_encoder_get_framecount(struct drm_encoder *encoder);
void mdp5_kms_set_bandwidth(struct mdp5_kms *mdp5_kms);
#ifdef CONFIG_DRM_MSM_DSI
void mdp5_cmd_encoder_mode_set(struct drm_encoder *encoder,
struct drm_display_mode *mode,
......
......@@ -154,46 +154,6 @@ static void mdp5_plane_cleanup_fb(struct drm_plane *plane,
msm_framebuffer_cleanup(fb, kms->aspace);
}
/* based on _dpu_plane_calc_bw */
static void mdp5_plane_calc_bw(struct drm_plane_state *state, struct drm_crtc_state *crtc_state)
{
struct drm_framebuffer *fb = state->fb;
struct mdp5_plane_state *pstate = to_mdp5_plane_state(state);
struct drm_display_mode *mode = &crtc_state->mode;
int bpp;
int src_width, src_height, dst_height, fps;
u64 plane_bw;
u32 hw_latency_lines;
u32 prefill_div;
u64 scale_factor;
int vbp, vpw, vfp;
src_width = drm_rect_width(&state->src) >> 16;
src_height = drm_rect_height(&state->src) >> 16;
dst_height = drm_rect_height(&state->dst);
fps = drm_mode_vrefresh(mode);
vbp = mode->vtotal - mode->vsync_end;
vpw = mode->vsync_end - mode->vsync_start;
vfp = mode->vsync_start - mode->vdisplay;
scale_factor = src_height > dst_height ?
mult_frac(src_height, 1, dst_height) : 1;
bpp = to_mdp_format(msm_framebuffer_format(fb))->cpp;
plane_bw = src_width * mode->vtotal * fps * bpp * scale_factor;
hw_latency_lines = 21; /* or 24? */
prefill_div = hw_latency_lines;
if (vbp + vpw > hw_latency_lines)
prefill_div = vbp + vpw;
#if 0
else if (vbp + vpw + vfp < hw_latency_lines)
prefill_div = vbp + vpw + vfp;
#endif
pstate->plane_bw = max(plane_bw, mult_frac(plane_bw, hw_latency_lines, prefill_div));
}
static int mdp5_plane_atomic_check_with_state(struct drm_crtc_state *crtc_state,
struct drm_plane_state *state)
{
......@@ -337,8 +297,6 @@ static int mdp5_plane_atomic_check_with_state(struct drm_crtc_state *crtc_state,
mdp5_pipe_release(state->state, old_hwpipe);
mdp5_pipe_release(state->state, old_right_hwpipe);
}
mdp5_plane_calc_bw(state, crtc_state);
} else {
mdp5_pipe_release(state->state, mdp5_state->hwpipe);
mdp5_pipe_release(state->state, mdp5_state->r_hwpipe);
......
Markdown is supported
0%
or
You are about to add 0 people to the discussion. Proceed with caution.
Finish editing this message first!
Please register or to comment