6db4831e98
Android 14
2893 lines
77 KiB
C
2893 lines
77 KiB
C
// SPDX-License-Identifier: GPL-2.0
|
|
/*
|
|
* Copyright (c) 2020 MediaTek Inc.
|
|
*/
|
|
|
|
#include <linux/module.h>
|
|
#include <linux/clk.h>
|
|
#include <linux/kthread.h>
|
|
#include <linux/platform_device.h>
|
|
#include <linux/of.h>
|
|
#include <linux/string.h>
|
|
#include <linux/math64.h>
|
|
#include <linux/mutex.h>
|
|
#include <linux/wait.h>
|
|
#include <linux/workqueue.h>
|
|
#include <linux/slab.h>
|
|
#include <linux/delay.h>
|
|
#include <linux/sched/clock.h>
|
|
|
|
#ifdef PLL_HOPPING_READY
|
|
#include <mtk_freqhopping_drv.h>
|
|
#endif
|
|
|
|
#if defined(USE_MEDIATEK_EMI)
|
|
#include <memory/mediatek/emi.h>
|
|
#include <memory/mediatek/dramc.h>
|
|
#elif defined(USE_MTK_DRAMC)
|
|
#include <mtk_dramc.h>
|
|
#endif
|
|
|
|
#include "mmdvfs_pmqos.h"
|
|
#include "mmdvfs_plat.h"
|
|
#include <mt-plat/aee.h>
|
|
|
|
#ifdef APPLY_CLK_LOG
|
|
#ifdef SMI_LAF
|
|
#include "mt6779_clkmgr.h"
|
|
#endif
|
|
#endif
|
|
#include "smi_pmqos.h"
|
|
#include "smi_public.h"
|
|
|
|
#define CREATE_TRACE_POINTS
|
|
#include "mmdvfs_events.h"
|
|
|
|
#include <linux/soc/mediatek/mtk_dvfsrc.h>
|
|
#include <dvfsrc-exp.h>
|
|
|
|
#ifdef MMDVFS_MMP
|
|
#include "mmprofile.h"
|
|
#include "mmprofile_function.h"
|
|
#endif
|
|
|
|
#if !(defined(CONFIG_MACH_MT6761) || defined(CONFIG_MACH_MT6765) || defined(CONFIG_MACH_MT6779))
|
|
#include <helio-dvfsrc-opp.h>
|
|
#endif
|
|
|
|
#ifdef QOS_BOUND_DETECT
|
|
#include "mtk_qos_bound.h"
|
|
#endif
|
|
|
|
#include "swpm_me.h"
|
|
|
|
#include <linux/regulator/consumer.h>
|
|
static struct regulator *vcore_reg_id;
|
|
|
|
|
|
#undef pr_fmt
|
|
#define pr_fmt(fmt) "[mmdvfs]" fmt
|
|
|
|
#define CLK_TYPE_NONE 0
|
|
#define CLK_TYPE_MUX 1
|
|
#define CLK_TYPE_PLL 2
|
|
|
|
#ifdef MMDVFS_MMP
|
|
struct mmdvfs_mmp_events_t {
|
|
mmp_event mmdvfs;
|
|
mmp_event freq_change;
|
|
mmp_event ext_freq_change;
|
|
mmp_event limit_change;
|
|
mmp_event hrt_change;
|
|
mmp_event cam_bw_mismatch;
|
|
mmp_event larb_soft_mode;
|
|
mmp_event larb_bwl;
|
|
mmp_event larb_port;
|
|
mmp_event smi_freq;
|
|
};
|
|
static struct mmdvfs_mmp_events_t mmdvfs_mmp_events;
|
|
#endif
|
|
|
|
enum {
|
|
VIRTUAL_DISP_LARB_ID = SMI_LARB_NUM,
|
|
VIRTUAL_MD_LARB_ID,
|
|
VIRTUAL_CCU_COMMON_ID,
|
|
VIRTUAL_CCU_COMMON2_ID,
|
|
MAX_LARB_COUNT
|
|
};
|
|
|
|
#define PORT_VIRTUAL_DISP SMI_PMQOS_ENC(VIRTUAL_DISP_LARB_ID, 0)
|
|
#define PORT_VIRTUAL_MD SMI_PMQOS_ENC(VIRTUAL_MD_LARB_ID, 0)
|
|
#define PORT_VIRTUAL_CCU_COMMON SMI_PMQOS_ENC(VIRTUAL_CCU_COMMON_ID, 0)
|
|
#define PORT_VIRTUAL_CCU_COMMON2 SMI_PMQOS_ENC(VIRTUAL_CCU_COMMON2_ID, 0)
|
|
|
|
static u32 log_level;
|
|
enum mmdvfs_log_level {
|
|
log_freq = 0,
|
|
log_bw,
|
|
log_limit,
|
|
log_smi_freq,
|
|
log_qos_validation,
|
|
};
|
|
|
|
#define STEP_UNREQUEST -1
|
|
|
|
struct mm_freq_step_config {
|
|
u32 clk_type; /* 0: don't set, 1: clk_mux, 2: pll hopping */
|
|
struct clk *clk_mux;
|
|
struct clk *clk_source;
|
|
u32 clk_mux_id;
|
|
u32 clk_source_id;
|
|
u32 pll_id;
|
|
u32 pll_value;
|
|
u64 freq_step;
|
|
};
|
|
|
|
struct mm_freq_limit_config {
|
|
u32 limit_size;
|
|
u32 limit_level;
|
|
u32 limit_value;
|
|
struct mm_freq_step_config **limit_steps;
|
|
};
|
|
|
|
struct mm_freq_config {
|
|
struct notifier_block nb;
|
|
const char *prop_name;
|
|
u32 pm_qos_class;
|
|
s32 current_step;
|
|
struct mm_freq_step_config step_config[MAX_FREQ_STEP];
|
|
struct mm_freq_limit_config limit_config;
|
|
};
|
|
|
|
enum mm_dprop { /*dprop: dts property */
|
|
mm_dp_freq = 0,
|
|
mm_dp_clk_type, /* 1 */
|
|
mm_dp_clk_param1, /* 2 */
|
|
mm_dp_clk_mux = mm_dp_clk_param1,
|
|
mm_dp_pll_id = mm_dp_clk_param1,
|
|
mm_dp_clk_param2, /* 3 */
|
|
mm_dp_clk_source = mm_dp_clk_param2,
|
|
mm_dp_pll_value = mm_dp_clk_param2,
|
|
mm_dp_max /* put max in the end */
|
|
};
|
|
|
|
#define FMETER_MUX_NODE_NAME "fmeter_mux_ids"
|
|
#define MAX_MUX_SIZE 9
|
|
static u32 mux_size;
|
|
static u32 fmeter_mux_ids[MAX_MUX_SIZE];
|
|
#ifdef APPLY_CLK_LOG
|
|
static u32 mux_real_freqs[MAX_MUX_SIZE];
|
|
#endif
|
|
|
|
#define UNINITIALIZED_VALUE (-1)
|
|
#define MAX_OSTD_NODE_NAME "max_ostd"
|
|
static s32 max_ostd = UNINITIALIZED_VALUE;
|
|
#define MAX_OSTD_LARB_NODE_NAME "max_ostd_larb"
|
|
#define CAM_LARB_NODE_NAME "cam_larb"
|
|
static u32 cam_larb_size;
|
|
static u32 cam_larb_ids[MAX_LARB_COUNT];
|
|
|
|
static u32 max_bw_bound;
|
|
#define MAX_COMM_NUM (2)
|
|
|
|
#define VCORE_NODE_NAME "vopp_steps"
|
|
#define MAX_USER_SIZE (12) /* Must be multiple of 4 */
|
|
static u32 step_size;
|
|
static s32 vopp_steps[MAX_FREQ_STEP];
|
|
static s32 current_max_step = STEP_UNREQUEST;
|
|
static s32 force_step = STEP_UNREQUEST;
|
|
static bool mmdvfs_enable;
|
|
static bool mmdvfs_autok_enable;
|
|
static struct mtk_pm_qos_request vcore_request;
|
|
static struct mtk_pm_qos_request mm_bw_request;
|
|
static struct mtk_pm_qos_request smi_freq_request[MAX_COMM_NUM];
|
|
static DEFINE_MUTEX(step_mutex);
|
|
static DEFINE_MUTEX(bw_mutex);
|
|
static s32 total_hrt_bw = UNINITIALIZED_VALUE;
|
|
static s32 total_ui_only_hrt_bw = UNINITIALIZED_VALUE;
|
|
static BLOCKING_NOTIFIER_HEAD(hrt_bw_throttle_notifier);
|
|
static BLOCKING_NOTIFIER_HEAD(cam_max_bw_notifier);
|
|
|
|
|
|
static int mm_freq_notify(struct notifier_block *nb,
|
|
unsigned long freq_value, void *v);
|
|
|
|
static struct mm_freq_config disp_freq = {
|
|
.nb.notifier_call = mm_freq_notify,
|
|
.prop_name = "disp_freq",
|
|
.pm_qos_class = PM_QOS_DISP_FREQ,
|
|
.current_step = STEP_UNREQUEST,
|
|
};
|
|
|
|
static struct mm_freq_config mdp_freq = {
|
|
.nb.notifier_call = mm_freq_notify,
|
|
.prop_name = "mdp_freq",
|
|
.pm_qos_class = PM_QOS_MDP_FREQ,
|
|
.current_step = STEP_UNREQUEST,
|
|
};
|
|
|
|
static struct mm_freq_config vdec_freq = {
|
|
.nb.notifier_call = mm_freq_notify,
|
|
.prop_name = "vdec_freq",
|
|
.pm_qos_class = PM_QOS_VDEC_FREQ,
|
|
.current_step = STEP_UNREQUEST,
|
|
};
|
|
|
|
static struct mm_freq_config venc_freq = {
|
|
.nb.notifier_call = mm_freq_notify,
|
|
.prop_name = "venc_freq",
|
|
.pm_qos_class = PM_QOS_VENC_FREQ,
|
|
.current_step = STEP_UNREQUEST,
|
|
};
|
|
|
|
static struct mm_freq_config cam_freq = {
|
|
.nb.notifier_call = mm_freq_notify,
|
|
.prop_name = "cam_freq",
|
|
.pm_qos_class = PM_QOS_CAM_FREQ,
|
|
.current_step = STEP_UNREQUEST,
|
|
};
|
|
|
|
static struct mm_freq_config img_freq = {
|
|
.nb.notifier_call = mm_freq_notify,
|
|
.prop_name = "img_freq",
|
|
.pm_qos_class = PM_QOS_IMG_FREQ,
|
|
.current_step = STEP_UNREQUEST,
|
|
};
|
|
|
|
static struct mm_freq_config dpe_freq = {
|
|
.nb.notifier_call = mm_freq_notify,
|
|
.prop_name = "dpe_freq",
|
|
.pm_qos_class = PM_QOS_DPE_FREQ,
|
|
.current_step = STEP_UNREQUEST,
|
|
};
|
|
|
|
static struct mm_freq_config ipe_freq = {
|
|
.prop_name = "ipe_freq",
|
|
.pm_qos_class = PM_QOS_RESERVED,
|
|
.current_step = STEP_UNREQUEST,
|
|
};
|
|
|
|
static struct mm_freq_config ccu_freq = {
|
|
.prop_name = "ccu_freq",
|
|
.pm_qos_class = PM_QOS_RESERVED,
|
|
.current_step = STEP_UNREQUEST,
|
|
};
|
|
|
|
static struct mm_freq_config img2_freq = {
|
|
.prop_name = "img2_freq",
|
|
.pm_qos_class = PM_QOS_RESERVED,
|
|
.current_step = STEP_UNREQUEST,
|
|
};
|
|
|
|
/* order should be same as pm_qos_class order for mmdvfs_qos_get_freq() */
|
|
struct mm_freq_config *all_freqs[] = {
|
|
&disp_freq, &mdp_freq,
|
|
&vdec_freq, &venc_freq,
|
|
&img_freq, &cam_freq, &dpe_freq, &ipe_freq, &ccu_freq, &img2_freq};
|
|
|
|
int __attribute__ ((weak)) is_dvfsrc_opp_fixed(void) { return 1; }
|
|
|
|
|
|
static void mm_apply_vcore(s32 vopp)
|
|
{
|
|
mtk_pm_qos_update_request(&vcore_request, vopp);
|
|
|
|
if (vcore_reg_id) {
|
|
#ifdef CHECK_VOLTAGE
|
|
u32 v_real, v_target, max_opp;
|
|
|
|
max_opp = mtk_dvfsrc_vcore_opp_count();
|
|
if (vopp >= 0 && vopp < max_opp) {
|
|
v_real = regulator_get_voltage(vcore_reg_id);
|
|
v_target = mtk_dvfsrc_vcore_uv_table(vopp);
|
|
if (v_real < v_target)
|
|
pr_info("err vcore %d < %d\n",
|
|
v_real, v_target);
|
|
|
|
}
|
|
#endif
|
|
}
|
|
}
|
|
|
|
static s32 mm_set_mux_clk(s32 src_mux_id, const char *name,
|
|
struct mm_freq_step_config *step_config, u32 step)
|
|
{
|
|
s32 ret = 0;
|
|
|
|
if (step_config->clk_mux == NULL ||
|
|
step_config->clk_source == NULL) {
|
|
pr_notice("CCF handle can't be NULL during MMDVFS\n");
|
|
return -EINVAL;
|
|
}
|
|
|
|
ret = clk_prepare_enable(step_config->clk_mux);
|
|
|
|
if (ret) {
|
|
pr_notice("prepare clk(%d): %s-%u\n",
|
|
ret, name, step);
|
|
return -EFAULT;
|
|
}
|
|
|
|
ret = clk_set_parent(
|
|
step_config->clk_mux, step_config->clk_source);
|
|
|
|
if (ret)
|
|
pr_notice(
|
|
"set parent(%d): %s-%u\n",
|
|
ret, name, step);
|
|
#ifdef APPLY_CLK_LOG
|
|
if (step_config->clk_mux_id == src_mux_id)
|
|
mux_real_freqs[src_mux_id] =
|
|
mt_get_ckgen_freq(fmeter_mux_ids[src_mux_id])/1000;
|
|
#endif
|
|
|
|
clk_disable_unprepare(step_config->clk_mux);
|
|
if (ret)
|
|
pr_notice(
|
|
"unprepare clk(%d): %s-%u\n",
|
|
ret, name, step);
|
|
return ret;
|
|
}
|
|
|
|
static s32 mm_set_freq_hopping_clk(const char *name,
|
|
struct mm_freq_step_config *step_config, u32 step)
|
|
{
|
|
s32 ret = 0;
|
|
|
|
#ifdef PLL_HOPPING_READY
|
|
ret = mt_dfs_general_pll(
|
|
step_config->pll_id, step_config->pll_value);
|
|
#endif
|
|
|
|
if (ret)
|
|
pr_notice("hopping rate(%d):(%u)-0x%08x, %s-%u\n",
|
|
ret, step_config->pll_id, step_config->pll_value,
|
|
name, step);
|
|
return ret;
|
|
}
|
|
|
|
static s32 apply_clk_by_type(u32 clk_type, s32 src_mux_id,
|
|
const char *name, struct mm_freq_step_config *config, s32 step)
|
|
{
|
|
s32 ret = 0;
|
|
|
|
if (clk_type == CLK_TYPE_MUX)
|
|
ret = mm_set_mux_clk(src_mux_id, name, config, step);
|
|
else if (clk_type == CLK_TYPE_PLL)
|
|
ret = mm_set_freq_hopping_clk(name, config, step);
|
|
return ret;
|
|
}
|
|
|
|
static void mm_check_limit(struct mm_freq_config *config,
|
|
struct mm_freq_step_config **step_config, u32 step)
|
|
{
|
|
struct mm_freq_step_config *normal_step = &config->step_config[step];
|
|
struct mm_freq_step_config *limit_step;
|
|
u32 level = config->limit_config.limit_level;
|
|
|
|
if (unlikely(level)) {
|
|
limit_step = &config->limit_config.limit_steps[level-1][step];
|
|
*step_config = limit_step;
|
|
if (log_level & 1 << log_limit)
|
|
pr_notice(
|
|
"limit %s: freq %llu -> %llu in step %u\n",
|
|
config->prop_name, normal_step->freq_step,
|
|
limit_step->freq_step, step);
|
|
#ifdef MMDVFS_MMP
|
|
mmprofile_log_ex(
|
|
mmdvfs_mmp_events.freq_change,
|
|
MMPROFILE_FLAG_PULSE, limit_step->freq_step,
|
|
config->pm_qos_class);
|
|
#endif
|
|
} else {
|
|
*step_config = normal_step;
|
|
}
|
|
}
|
|
|
|
static s32 mm_apply_clk(s32 src_mux_id,
|
|
struct mm_freq_config *config, u32 step, s32 old_step)
|
|
{
|
|
struct mm_freq_step_config *step_config;
|
|
s32 ret = 0;
|
|
s32 operations[2];
|
|
u32 i;
|
|
|
|
if (step >= MAX_FREQ_STEP) {
|
|
pr_notice(
|
|
"Invalid clk apply step %d in %s\n",
|
|
step, config->prop_name);
|
|
return -EINVAL;
|
|
}
|
|
|
|
mm_check_limit(config, &step_config, step);
|
|
|
|
if (step_config->clk_type == CLK_TYPE_NONE) {
|
|
pr_notice("No need to change clk of %s\n", config->prop_name);
|
|
return 0;
|
|
}
|
|
|
|
operations[0] = (step < old_step) ? CLK_TYPE_PLL : CLK_TYPE_MUX;
|
|
operations[1] = (step < old_step) ? CLK_TYPE_MUX : CLK_TYPE_PLL;
|
|
|
|
for (i = 0; i < ARRAY_SIZE(operations); i++) {
|
|
if (step_config->clk_type & operations[i])
|
|
ret = apply_clk_by_type(operations[i], src_mux_id,
|
|
config->prop_name, step_config, step);
|
|
}
|
|
|
|
return ret;
|
|
|
|
}
|
|
|
|
/*
|
|
* Each freq occupies 8 bits => 0~3:current_step 4~7:id
|
|
* (id is mapping to index of all_freqs)
|
|
*/
|
|
static inline u32 set_freq_for_log(u32 freq, s32 cur_step, u32 id)
|
|
{
|
|
cur_step &= 0xF;
|
|
id <<= 4;
|
|
return (freq | cur_step | id);
|
|
}
|
|
|
|
static void mm_apply_clk_for_all(u32 pm_qos_class, s32 src_mux_id,
|
|
u32 step, s32 old_step)
|
|
{
|
|
u32 i;
|
|
u32 clk_mux_id;
|
|
u32 real_freq = 0;
|
|
u8 freq[MAX_USER_SIZE] = {0};
|
|
bool set[ARRAY_SIZE(all_freqs)] = {false};
|
|
u32 first_log;
|
|
|
|
for (i = 0; i < ARRAY_SIZE(all_freqs); i++) {
|
|
clk_mux_id = all_freqs[i]->step_config[step].clk_mux_id;
|
|
if (!set[clk_mux_id]) {
|
|
if (!mm_apply_clk(src_mux_id,
|
|
all_freqs[i], step, old_step))
|
|
set[clk_mux_id] = true;
|
|
}
|
|
#ifdef APPLY_CLK_LOG
|
|
if (all_freqs[i]->pm_qos_class == pm_qos_class)
|
|
real_freq = mux_real_freqs[clk_mux_id];
|
|
#endif
|
|
freq[i] = set_freq_for_log(
|
|
freq[i], all_freqs[i]->current_step, i);
|
|
}
|
|
set_swpm_me_freq(all_freqs[3]->step_config[step].freq_step,
|
|
all_freqs[2]->step_config[step].freq_step,
|
|
all_freqs[1]->step_config[step].freq_step);
|
|
first_log = (pm_qos_class << 16) | step;
|
|
|
|
#ifdef MMDVFS_MMP
|
|
mmprofile_log_ex(
|
|
mmdvfs_mmp_events.freq_change,
|
|
MMPROFILE_FLAG_PULSE, first_log, real_freq);
|
|
|
|
mmprofile_log_ex(
|
|
mmdvfs_mmp_events.ext_freq_change,
|
|
MMPROFILE_FLAG_PULSE, *((u32 *)&freq[0]), *((u32 *)&freq[4]));
|
|
#endif
|
|
if (log_level & 1 << log_freq)
|
|
pr_notice(
|
|
"freq change:%u class:%u step:%u f0:%x f1:%x\n",
|
|
real_freq, pm_qos_class, step,
|
|
*((u32 *)&freq[0]), *((u32 *)&freq[4]));
|
|
}
|
|
|
|
/* id is from SMI_LARB_L1ARB */
|
|
static void get_comm_port_by_id(u32 id, u32 *comm, u32 *comm_port)
|
|
{
|
|
*comm = id >> 16;
|
|
*comm_port = id & 0xffff;
|
|
}
|
|
|
|
static inline u32 get_id_by_comm_port(u32 comm, u32 comm_port)
|
|
{
|
|
return ((comm << 16) | (comm_port & 0xffff));
|
|
}
|
|
|
|
static bool larb_soft = true;
|
|
static u32 default_bwl = 0x200;
|
|
static s32 force_larb_mode = -1;
|
|
static s32 comm_port_limit[MAX_COMM_NUM][SMI_COMM_MASTER_NUM] = {};
|
|
static s32 comm_port_hrt[MAX_COMM_NUM][SMI_COMM_MASTER_NUM] = {};
|
|
static s32 force_comm_bwl[MAX_COMM_NUM][SMI_COMM_MASTER_NUM] = {};
|
|
static u32 comm_freq_class[MAX_COMM_NUM] = {};
|
|
#ifdef MMDVFS_SKIP_SMI_CONFIG
|
|
static bool skip_smi_config = true;
|
|
#else
|
|
static bool skip_smi_config;
|
|
#endif
|
|
void mm_qos_update_larb_bwl(u32 larb_update, bool bw_change)
|
|
{
|
|
u32 i, larb_bw, comm, comm_port;
|
|
bool larb_soft_mode = larb_soft;
|
|
s32 freq[MAX_COMM_NUM];
|
|
const u32 length = MAX_COMM_NUM * SMI_COMM_MASTER_NUM;
|
|
|
|
if (unlikely(force_larb_mode >= 0))
|
|
larb_soft_mode = force_larb_mode;
|
|
|
|
for (i = 0; i < MAX_COMM_NUM; i++) {
|
|
if (comm_freq_class[i] == 0)
|
|
freq[i] = 0;
|
|
else
|
|
freq[i] = mmdvfs_qos_get_freq(comm_freq_class[i]);
|
|
}
|
|
|
|
for (i = 0; i < length; i++) {
|
|
if (!(larb_update & (1 << i)))
|
|
continue;
|
|
comm = i / SMI_COMM_MASTER_NUM;
|
|
if (freq[comm] <= 0)
|
|
continue;
|
|
comm_port = i % SMI_COMM_MASTER_NUM;
|
|
larb_bw = 0;
|
|
if (force_comm_bwl[comm][comm_port] != 0) {
|
|
larb_bw = force_comm_bwl[comm][comm_port];
|
|
if (log_level & 1 << log_bw)
|
|
pr_notice("force comm:%d port:%d bwl:%#x\n",
|
|
comm, comm_port, larb_bw);
|
|
} else if (comm_port_limit[comm][comm_port]) {
|
|
larb_bw = (comm_port_limit[comm][comm_port] << 8)
|
|
/ freq[comm];
|
|
if (log_level & 1 << log_bw)
|
|
pr_notice("comm:%d port:%d bwl:%#x bw:%u\n",
|
|
comm, comm_port, larb_bw,
|
|
comm_port_limit[comm][comm_port]);
|
|
}
|
|
if (larb_bw) {
|
|
smi_bwl_update(get_id_by_comm_port(comm, comm_port),
|
|
larb_bw, (comm_port_hrt[comm][comm_port] > 0) ?
|
|
true : larb_soft_mode, "MMDVFS");
|
|
trace_mmqos__update_larb(comm, comm_port,
|
|
comm_port_limit[comm][comm_port], larb_bw,
|
|
(comm_port_hrt[comm][comm_port] > 0) ?
|
|
true : larb_soft_mode);
|
|
#ifdef MMDVFS_MMP
|
|
if (mmdvfs_log_larb_mmp(comm_port, -1))
|
|
mmprofile_log_ex(
|
|
mmdvfs_mmp_events.larb_bwl,
|
|
MMPROFILE_FLAG_PULSE,
|
|
(comm_port << 28) | larb_bw,
|
|
larb_soft_mode);
|
|
#endif
|
|
} else if (bw_change) {
|
|
/* if no bwl_bw, set default bwl with soft-mode */
|
|
smi_bwl_update(get_id_by_comm_port(comm, comm_port),
|
|
default_bwl, true, "MMDVFS");
|
|
trace_mmqos__update_larb(comm, comm_port,
|
|
comm_port_limit[comm][comm_port],
|
|
default_bwl, true);
|
|
#ifdef MMDVFS_MMP
|
|
if (mmdvfs_log_larb_mmp(comm_port, -1))
|
|
mmprofile_log_ex(
|
|
mmdvfs_mmp_events.larb_bwl,
|
|
MMPROFILE_FLAG_PULSE,
|
|
(comm_port << 28) | default_bwl, 2);
|
|
#endif
|
|
}
|
|
}
|
|
}
|
|
|
|
static u32 mmdvfs_get_limit_status(u32 pm_qos_class)
|
|
{
|
|
u32 i = pm_qos_class - PM_QOS_DISP_FREQ;
|
|
|
|
if (i >= ARRAY_SIZE(all_freqs)) {
|
|
pr_notice("[GET]Invalid class: %u\n", pm_qos_class);
|
|
return false;
|
|
}
|
|
|
|
return all_freqs[i]->limit_config.limit_level;
|
|
}
|
|
|
|
static void update_step(u32 pm_qos_class, s32 src_mux_id)
|
|
{
|
|
u32 i;
|
|
s32 old_max_step;
|
|
|
|
if (!mmdvfs_enable || !mmdvfs_autok_enable) {
|
|
pr_notice("mmdvfs qos is disabled(%d)\n", pm_qos_class);
|
|
return;
|
|
}
|
|
|
|
if (!step_size) {
|
|
pr_notice("no step available skip\n");
|
|
return;
|
|
}
|
|
|
|
mutex_lock(&step_mutex);
|
|
old_max_step = current_max_step;
|
|
current_max_step = step_size;
|
|
if (force_step != STEP_UNREQUEST) {
|
|
current_max_step = force_step;
|
|
} else {
|
|
for (i = 0; i < ARRAY_SIZE(all_freqs); i++) {
|
|
if (all_freqs[i]->current_step != -1 &&
|
|
all_freqs[i]->current_step < current_max_step)
|
|
current_max_step = all_freqs[i]->current_step;
|
|
}
|
|
if (current_max_step == step_size)
|
|
current_max_step = STEP_UNREQUEST;
|
|
}
|
|
|
|
if (current_max_step == old_max_step) {
|
|
mutex_unlock(&step_mutex);
|
|
return;
|
|
}
|
|
|
|
if (current_max_step != STEP_UNREQUEST
|
|
&& (current_max_step < old_max_step
|
|
|| old_max_step == STEP_UNREQUEST)) {
|
|
/* configuration for higher freq */
|
|
mm_apply_vcore(vopp_steps[current_max_step]);
|
|
mm_apply_clk_for_all(pm_qos_class, src_mux_id,
|
|
current_max_step, old_max_step);
|
|
} else {
|
|
/* configuration for lower freq */
|
|
s32 vopp_step = STEP_UNREQUEST;
|
|
u32 freq_step = step_size - 1;
|
|
|
|
if (current_max_step != STEP_UNREQUEST) {
|
|
vopp_step = vopp_steps[current_max_step];
|
|
freq_step = current_max_step;
|
|
}
|
|
mm_apply_clk_for_all(
|
|
pm_qos_class, src_mux_id, freq_step, old_max_step);
|
|
mm_apply_vcore(vopp_step);
|
|
}
|
|
mutex_unlock(&step_mutex);
|
|
|
|
if (!skip_smi_config) {
|
|
/* update bwl due to freq change */
|
|
mutex_lock(&bw_mutex);
|
|
mm_qos_update_larb_bwl(0xFFFF, false);
|
|
mutex_unlock(&bw_mutex);
|
|
}
|
|
}
|
|
|
|
static int mm_freq_notify(struct notifier_block *nb,
|
|
unsigned long freq_value, void *v)
|
|
{
|
|
struct mm_freq_config *mm_freq;
|
|
s32 step;
|
|
|
|
mm_freq = container_of(nb, struct mm_freq_config, nb);
|
|
if (!step_size) {
|
|
pr_notice(
|
|
"no step available in %s, skip\n", mm_freq->prop_name);
|
|
return NOTIFY_OK;
|
|
}
|
|
|
|
step = step_size - 1;
|
|
if (freq_value == PM_QOS_MM_FREQ_DEFAULT_VALUE) {
|
|
mm_freq->current_step = STEP_UNREQUEST;
|
|
} else {
|
|
for (; step >= 1; step--) {
|
|
if (freq_value <= mm_freq->step_config[step].freq_step)
|
|
break;
|
|
}
|
|
mm_freq->current_step = step;
|
|
}
|
|
update_step(mm_freq->pm_qos_class,
|
|
mm_freq->step_config[step].clk_mux_id);
|
|
|
|
return NOTIFY_OK;
|
|
}
|
|
|
|
int mmdvfs_qos_get_freq_steps(u32 pm_qos_class,
|
|
u64 *out_freq_steps, u32 *out_step_size)
|
|
{
|
|
struct mm_freq_config *mm_freq = NULL;
|
|
u32 i;
|
|
|
|
if (!out_freq_steps || !out_step_size)
|
|
return -EINVAL;
|
|
|
|
for (i = 0; i < ARRAY_SIZE(all_freqs); i++) {
|
|
if (pm_qos_class == all_freqs[i]->pm_qos_class) {
|
|
mm_freq = all_freqs[i];
|
|
break;
|
|
}
|
|
}
|
|
|
|
if (!mm_freq)
|
|
return -ENXIO;
|
|
|
|
*out_step_size = step_size;
|
|
for (i = 0; i < step_size; i++)
|
|
out_freq_steps[i] = mm_freq->step_config[i].freq_step;
|
|
|
|
return 0;
|
|
}
|
|
EXPORT_SYMBOL_GPL(mmdvfs_qos_get_freq_steps);
|
|
|
|
#define MAX_LARB_NAME 16
|
|
|
|
static struct mm_larb_request larb_req[MAX_LARB_COUNT] = {};
|
|
#define LARB_NODE_NAME "larb_groups"
|
|
|
|
#define MAX_CH_COUNT 2
|
|
static s32 channel_srt_bw[MAX_COMM_NUM][MAX_CH_COUNT] = {};
|
|
static s32 channel_hrt_bw[MAX_COMM_NUM][MAX_CH_COUNT] = {};
|
|
static s32 channel_disp_hrt_cnt[MAX_COMM_NUM][MAX_CH_COUNT] = {};
|
|
|
|
#define MULTIPLY_BW_THRESH_HIGH(value) ((value)*1/2)
|
|
#define MULTIPLY_BW_THRESHOLD_LOW(value) ((value)*2/5)
|
|
#define MULTIPLY_RATIO(value) ((value)*1000)
|
|
#define DIVIDE_RATIO(value) ((value)/1000)
|
|
static s32 current_hrt_bw;
|
|
static u32 camera_max_bw;
|
|
static s32 get_cam_hrt_bw(void)
|
|
{
|
|
u32 i;
|
|
s32 result = 0;
|
|
|
|
for (i = 0; i < cam_larb_size; i++)
|
|
result += larb_req[cam_larb_ids[i]].total_hrt_data;
|
|
|
|
return result;
|
|
}
|
|
|
|
static bool is_camera_larb(u32 master_id)
|
|
{
|
|
u32 i;
|
|
bool result = false;
|
|
|
|
for (i = 0; i < cam_larb_size; i++) {
|
|
if (SMI_PMQOS_LARB_DEC(master_id) == cam_larb_ids[i]) {
|
|
result = true;
|
|
break;
|
|
}
|
|
}
|
|
|
|
return result;
|
|
}
|
|
|
|
static s32 get_total_used_hrt_bw(void)
|
|
{
|
|
s32 cam_hrt_bw;
|
|
s32 disp_hrt_bw;
|
|
s32 md_hrt_bw;
|
|
|
|
/* HRT Write BW should multiply a weight */
|
|
cam_hrt_bw = dram_write_weight(MULTIPLY_RATIO(get_cam_hrt_bw())/cam_occ_ratio());
|
|
disp_hrt_bw =
|
|
MULTIPLY_RATIO(larb_req[SMI_PMQOS_LARB_DEC(PORT_VIRTUAL_DISP)].total_hrt_data)
|
|
/disp_occ_ratio();
|
|
md_hrt_bw =
|
|
larb_req[SMI_PMQOS_LARB_DEC(PORT_VIRTUAL_MD)].total_hrt_data;
|
|
return (cam_hrt_bw + disp_hrt_bw + md_hrt_bw);
|
|
}
|
|
|
|
#if defined(USE_MEDIATEK_EMI)
|
|
static s32 get_io_width(void)
|
|
{
|
|
s32 io_width;
|
|
s32 ddr_type = mtk_dramc_get_ddr_type();
|
|
|
|
if (ddr_type == TYPE_LPDDR4 || ddr_type == TYPE_LPDDR4X
|
|
|| ddr_type == TYPE_LPDDR4P || ddr_type == TYPE_LPDDR5)
|
|
io_width = 2;
|
|
else
|
|
io_width = 4;
|
|
|
|
return io_width;
|
|
}
|
|
#elif defined(USE_MTK_DRAMC)
|
|
static s32 get_io_width(void)
|
|
{
|
|
s32 io_width;
|
|
s32 ddr_type = get_ddr_type();
|
|
|
|
if (ddr_type == TYPE_LPDDR3)
|
|
io_width = 4;
|
|
else if (ddr_type == TYPE_LPDDR4 || ddr_type == TYPE_LPDDR4X)
|
|
io_width = 2;
|
|
else
|
|
io_width = 4;
|
|
|
|
return io_width;
|
|
}
|
|
#endif
|
|
|
|
#ifdef HRT_MECHANISM
|
|
#ifdef SIMULATE_DVFSRC
|
|
static s32 bw_threshold_high[DDR_OPP_NUM] = {0};
|
|
static s32 bw_threshold_low[DDR_OPP_NUM] = {0};
|
|
static struct mtk_pm_qos_request ddr_request;
|
|
|
|
|
|
static void init_simulation(void)
|
|
{
|
|
u32 i = 0;
|
|
|
|
for (i = 0; i < DDR_OPP_NUM; i++) {
|
|
s32 freq = 0;
|
|
#ifdef USE_MTK_DRAMC
|
|
s32 ch_num = get_emi_ch_num();
|
|
/* Todo: Use API from DRAM owner */
|
|
s32 io_width = get_io_width();
|
|
|
|
/* Todo: It should be modified in P80 */
|
|
if (i == 0)
|
|
freq = dram_steps_freq(i) * ch_num * io_width;
|
|
else
|
|
freq = dram_steps_freq(i+1) * ch_num * io_width;
|
|
#endif
|
|
bw_threshold_high[i] =
|
|
(s32)MULTIPLY_BW_THRESH_HIGH(freq);
|
|
bw_threshold_low[i] =
|
|
(s32)MULTIPLY_BW_THRESHOLD_LOW(freq);
|
|
}
|
|
|
|
mtk_pm_qos_add_request(
|
|
&ddr_request, MTK_PM_QOS_DDR_OPP,
|
|
MTK_PM_QOS_DDR_OPP_DEFAULT_VALUE);
|
|
}
|
|
|
|
static u32 get_ddr_opp_by_threshold(s32 bw, s32 *threshold_array)
|
|
{
|
|
s32 i = 0;
|
|
u32 opp = 0;
|
|
|
|
/**
|
|
* From small value to large value.
|
|
* Find the first threshold which is larger than input bw.
|
|
* If no threshold is found, it must be highest level of DDR.
|
|
*/
|
|
for (i = DDR_OPP_NUM-1; i >= 0; i--) {
|
|
if (bw < threshold_array[i]) {
|
|
opp = i;
|
|
break;
|
|
}
|
|
}
|
|
return opp;
|
|
}
|
|
|
|
static void simulate_dvfsrc(s32 next_hrt_bw)
|
|
{
|
|
u32 current_opp, next_opp;
|
|
s32 *threshold_array;
|
|
bool is_up = false;
|
|
|
|
if (next_hrt_bw > current_hrt_bw) {
|
|
threshold_array = &bw_threshold_high[0];
|
|
is_up = true;
|
|
} else
|
|
threshold_array = &bw_threshold_low[0];
|
|
|
|
current_opp = get_ddr_opp_by_threshold(current_hrt_bw, threshold_array);
|
|
next_opp = get_ddr_opp_by_threshold(next_hrt_bw, threshold_array);
|
|
|
|
if ((is_up && next_opp < current_opp) ||
|
|
(!is_up && next_opp > current_opp)) {
|
|
mtk_pm_qos_update_request(&ddr_request, next_opp);
|
|
if (log_level & 1 << log_bw)
|
|
pr_notice("up=%d copp=%d nopp=%d cbw=%d nbw=%d\n",
|
|
is_up, current_opp, next_opp,
|
|
current_hrt_bw, next_hrt_bw);
|
|
}
|
|
}
|
|
#else
|
|
static struct mtk_pm_qos_request dvfsrc_isp_hrt_req;
|
|
static void init_dvfsrc(void)
|
|
{
|
|
mtk_pm_qos_add_request(
|
|
&dvfsrc_isp_hrt_req, MTK_PM_QOS_HRT_BANDWIDTH,
|
|
MTK_PM_QOS_HRT_BANDWIDTH_DEFAULT_VALUE);
|
|
}
|
|
#endif
|
|
|
|
static void log_hrt_bw_info(u32 master_id)
|
|
{
|
|
s32 ccu_hrt_bw = get_ccu_hrt_bw(larb_req);
|
|
s32 p1_hrt_bw = get_cam_hrt_bw() - ccu_hrt_bw;
|
|
s32 disp_hrt_bw =
|
|
larb_req[SMI_PMQOS_LARB_DEC(PORT_VIRTUAL_DISP)].total_hrt_data;
|
|
|
|
/* Note: Should be enabled when mmdvfs gki complete
|
|
* u32 ddr_opp = mtk_dvfsrc_query_opp_info(MTK_DVFSRC_CURR_DRAM_OPP);
|
|
*/
|
|
u32 ddr_opp = 0;
|
|
#ifdef MMDVFS_MMP
|
|
u32 param1 = (SMI_PMQOS_LARB_DEC(master_id) << 24) |
|
|
(ddr_opp << 16) | disp_hrt_bw;
|
|
u32 param2 = (ccu_hrt_bw << 16) | p1_hrt_bw;
|
|
|
|
mmprofile_log_ex(
|
|
mmdvfs_mmp_events.hrt_change,
|
|
MMPROFILE_FLAG_PULSE, param1, param2);
|
|
#endif
|
|
|
|
if (log_level & 1 << log_bw)
|
|
pr_notice("%s larb=%d p1=%d ccu=%d disp=%d ddr_opp=%d\n",
|
|
__func__, SMI_PMQOS_LARB_DEC(master_id), p1_hrt_bw,
|
|
ccu_hrt_bw, disp_hrt_bw, ddr_opp);
|
|
}
|
|
|
|
static void update_hrt_bw_to_dvfsrc(s32 next_hrt_bw)
|
|
{
|
|
#ifdef SIMULATE_DVFSRC
|
|
simulate_dvfsrc(next_hrt_bw);
|
|
#else
|
|
u32 md_larb_id = SMI_PMQOS_LARB_DEC(PORT_VIRTUAL_MD);
|
|
s32 mm_used_hrt_bw =
|
|
next_hrt_bw - larb_req[md_larb_id].total_hrt_data;
|
|
|
|
mtk_pm_qos_update_request(&dvfsrc_isp_hrt_req, mm_used_hrt_bw);
|
|
if (log_level & 1 << log_bw)
|
|
pr_notice("%s report dvfsrc mm_hrt_bw=%d\n",
|
|
__func__, mm_used_hrt_bw);
|
|
#endif
|
|
}
|
|
|
|
#endif /* HRT_MECHANISM */
|
|
|
|
|
|
#ifdef BLOCKING_MECHANISM
|
|
static atomic_t lock_cam_count = ATOMIC_INIT(0);
|
|
static wait_queue_head_t hrt_wait;
|
|
#define WAIT_TIMEOUT_MS 200
|
|
|
|
static void blocking_camera(void)
|
|
{
|
|
u32 wait_result;
|
|
|
|
pr_notice("begin to blocking for camera_max_bw=%d\n", camera_max_bw);
|
|
wait_result = wait_event_timeout(
|
|
hrt_wait, atomic_read(&lock_cam_count) == 0,
|
|
msecs_to_jiffies(WAIT_TIMEOUT_MS));
|
|
pr_notice("blocking wait_result=%d\n", wait_result);
|
|
}
|
|
#endif
|
|
|
|
static void trace_qos_validation(void)
|
|
{
|
|
struct mm_qos_request *req = NULL;
|
|
u16 port_index_list[MAX_PORT_COUNT];
|
|
u32 i, j, port_id;
|
|
s32 bw;
|
|
|
|
for (i = 0; i < ARRAY_SIZE(larb_req); i++) {
|
|
if (!larb_req[i].port_count)
|
|
continue;
|
|
for (j = 0; j < MAX_PORT_COUNT; j++)
|
|
port_index_list[j] = 0;
|
|
list_for_each_entry(req, &larb_req[i].larb_list, larb_node) {
|
|
/* Make one trace for each request instead of for each
|
|
* port because it's hard to calculate data size when
|
|
* one port with many requests (BW and fps are mixed)
|
|
*/
|
|
port_id = SMI_PMQOS_PORT_MASK(req->master_id);
|
|
port_index_list[port_id]++;
|
|
bw = get_comp_value(req->bw_value,
|
|
req->comp_type, true);
|
|
if (req->updated || bw > 0)
|
|
trace_mmqos__update_qosbw(i, port_id,
|
|
port_index_list[port_id], bw);
|
|
}
|
|
}
|
|
}
|
|
|
|
static inline void init_larb_list(u32 larb_id)
|
|
{
|
|
if (!larb_req[larb_id].larb_list_init) {
|
|
INIT_LIST_HEAD(&(larb_req[larb_id].larb_list));
|
|
larb_req[larb_id].larb_list_init = true;
|
|
}
|
|
}
|
|
|
|
s32 mm_qos_add_request(struct plist_head *owner_list,
|
|
struct mm_qos_request *req, u32 smi_master_id)
|
|
{
|
|
u32 larb_id, port_id;
|
|
struct mm_qos_request *enum_req = NULL;
|
|
|
|
larb_id = SMI_PMQOS_LARB_DEC(smi_master_id);
|
|
port_id = SMI_PMQOS_PORT_MASK(smi_master_id);
|
|
if (!req) {
|
|
pr_notice("mm_add: Invalid req pointer\n");
|
|
return -EINVAL;
|
|
}
|
|
if (larb_id >= MAX_LARB_COUNT || port_id >= MAX_PORT_COUNT) {
|
|
pr_notice("mm_add(0x%08x) Invalid master_id\n", smi_master_id);
|
|
return -EINVAL;
|
|
}
|
|
if (req->init) {
|
|
pr_notice("mm_add(0x%08x) req is init\n", req->master_id);
|
|
return -EINVAL;
|
|
}
|
|
|
|
req->master_id = smi_master_id;
|
|
req->bw_value = 0;
|
|
req->hrt_value = 0;
|
|
plist_node_init(&(req->owner_node), smi_master_id);
|
|
plist_add(&(req->owner_node), owner_list);
|
|
INIT_LIST_HEAD(&(req->larb_node));
|
|
INIT_LIST_HEAD(&(req->port_node));
|
|
init_larb_list(larb_id);
|
|
|
|
mutex_lock(&bw_mutex);
|
|
list_add_tail(&(req->larb_node), &(larb_req[larb_id].larb_list));
|
|
req->init = true;
|
|
|
|
list_for_each_entry(enum_req, &larb_req[larb_id].larb_list, larb_node) {
|
|
if (enum_req != req && req->master_id == enum_req->master_id) {
|
|
list_add_tail(&(req->port_node),
|
|
&(enum_req->port_node));
|
|
break;
|
|
}
|
|
}
|
|
mutex_unlock(&bw_mutex);
|
|
|
|
if (log_level & 1 << log_bw) {
|
|
pr_notice("mm_add larb=%u port=%d\n", larb_id, port_id);
|
|
pr_notice("req=%p\n", req);
|
|
}
|
|
return 0;
|
|
}
|
|
EXPORT_SYMBOL_GPL(mm_qos_add_request);
|
|
|
|
#define SHIFT_ROUND(a, b) ((((a) - 1) >> (b)) + 1)
|
|
s32 mm_qos_set_request(struct mm_qos_request *req, u32 bw_value,
|
|
u32 hrt_value, u32 comp_type)
|
|
{
|
|
u32 larb, port, bw, old_larb_mix_value;
|
|
u32 old_comp_bw, old_comp_limit, new_comp_bw, new_comp_limit;
|
|
u32 comm, comm_port;
|
|
struct mm_qos_request *enum_req = NULL;
|
|
bool hrt_port = false;
|
|
#if defined(USE_MEDIATEK_EMI)
|
|
#if IS_ENABLED(CONFIG_MACH_MT6877)
|
|
s32 ddr_type = mtk_dramc_get_ddr_type();
|
|
#endif
|
|
#endif
|
|
|
|
if (!req)
|
|
return -EINVAL;
|
|
|
|
larb = SMI_PMQOS_LARB_DEC(req->master_id);
|
|
port = SMI_PMQOS_PORT_MASK(req->master_id);
|
|
if (!req->init || larb >= MAX_LARB_COUNT ||
|
|
port >= MAX_PORT_COUNT || comp_type >= BW_COMP_END) {
|
|
pr_notice("mm_set(0x%08x) init=%d larb=%d port=%d comp=%d\n",
|
|
req->master_id, req->init, larb, port, comp_type);
|
|
dump_stack();
|
|
return -EINVAL;
|
|
}
|
|
if (!larb_req[larb].port_count || !larb_req[larb].ratio[port]) {
|
|
pr_notice("mm_set(0x%08x) invalid port_cnt=%d ratio=%d\n",
|
|
req->master_id, larb_req[larb].port_count,
|
|
larb_req[larb].ratio[port]);
|
|
return -EINVAL;
|
|
}
|
|
|
|
if (bw_value > max_bw_bound || hrt_value > max_bw_bound) {
|
|
pr_notice("mm_set(0x%08x) invalid bw=%d hrt=%d bw_bound=%d\n",
|
|
req->master_id, bw_value,
|
|
hrt_value, max_bw_bound);
|
|
return -EINVAL;
|
|
}
|
|
|
|
if (req->hrt_value == hrt_value &&
|
|
req->bw_value == bw_value &&
|
|
req->comp_type == comp_type) {
|
|
if (log_level & 1 << log_bw)
|
|
pr_notice("mm_set(0x%08x) no change\n", req->master_id);
|
|
return 0;
|
|
}
|
|
|
|
mutex_lock(&bw_mutex);
|
|
|
|
req->updated = true;
|
|
old_comp_bw = get_comp_value(req->bw_value, req->comp_type, true);
|
|
old_comp_limit = get_comp_value(req->bw_value, req->comp_type, false);
|
|
new_comp_bw = get_comp_value(bw_value, comp_type, true);
|
|
new_comp_limit = get_comp_value(bw_value, comp_type, false);
|
|
/* Update Total QoS BW */
|
|
larb_req[larb].total_bw_data -= old_comp_bw;
|
|
larb_req[larb].total_bw_data += new_comp_bw;
|
|
|
|
old_larb_mix_value = larb_req[larb].total_mix_limit;
|
|
get_comm_port_by_id(larb_req[larb].comm_port, &comm, &comm_port);
|
|
if (req->hrt_value) {
|
|
larb_req[larb].total_hrt_data -= req->hrt_value;
|
|
larb_req[larb].total_mix_limit -= req->hrt_value;
|
|
if (larb < MAX_LARB_COUNT &&
|
|
comm_port < SMI_COMM_MASTER_NUM)
|
|
comm_port_hrt[comm][comm_port] -=
|
|
req->hrt_value;
|
|
if (larb < MAX_LARB_COUNT &&
|
|
larb_req[larb].channel < MAX_CH_COUNT) {
|
|
if (larb_req[larb].is_max_ostd)
|
|
channel_disp_hrt_cnt[comm][larb_req[
|
|
larb].channel]--;
|
|
else
|
|
channel_hrt_bw[comm][larb_req[
|
|
larb].channel] -= req->hrt_value;
|
|
}
|
|
} else
|
|
larb_req[larb].total_mix_limit -= old_comp_limit;
|
|
|
|
if (hrt_value) {
|
|
larb_req[larb].total_hrt_data += hrt_value;
|
|
larb_req[larb].total_mix_limit += hrt_value;
|
|
if (larb < MAX_LARB_COUNT &&
|
|
comm_port < SMI_COMM_MASTER_NUM)
|
|
comm_port_hrt[comm][comm_port] += hrt_value;
|
|
if (larb < MAX_LARB_COUNT &&
|
|
larb_req[larb].channel < MAX_CH_COUNT) {
|
|
if (larb_req[larb].is_max_ostd)
|
|
channel_disp_hrt_cnt[comm][larb_req[
|
|
larb].channel]++;
|
|
else
|
|
channel_hrt_bw[comm][larb_req[
|
|
larb].channel] += hrt_value;
|
|
}
|
|
} else
|
|
larb_req[larb].total_mix_limit += new_comp_limit;
|
|
|
|
if (larb < MAX_LARB_COUNT && larb_req[larb].channel < MAX_CH_COUNT) {
|
|
channel_srt_bw[comm][larb_req[larb].channel] -= old_comp_bw;
|
|
channel_srt_bw[comm][larb_req[larb].channel] += new_comp_bw;
|
|
}
|
|
|
|
if (larb < MAX_LARB_COUNT &&
|
|
comm_port < SMI_COMM_MASTER_NUM) {
|
|
comm_port_limit[comm][comm_port] -= old_larb_mix_value;
|
|
comm_port_limit[comm][comm_port] +=
|
|
larb_req[larb].total_mix_limit;
|
|
}
|
|
|
|
if (log_level & 1 << log_bw) {
|
|
pr_notice("set=0x%08x comp=%u,%u\n", req->master_id,
|
|
comp_type, req->comp_type);
|
|
pr_notice("set=0x%08x bw=%u,%u total_bw=%d\n", req->master_id,
|
|
bw_value, req->bw_value, larb_req[larb].total_bw_data);
|
|
pr_notice("set=0x%08x hrt=%u,%u total_hrt=%d\n", req->master_id,
|
|
hrt_value, req->hrt_value, larb_req[larb].total_hrt_data);
|
|
pr_notice("set=0x%08x o_mix=%u total_mix=%d\n", req->master_id,
|
|
old_larb_mix_value, larb_req[larb].total_mix_limit);
|
|
}
|
|
|
|
req->hrt_value = hrt_value;
|
|
req->bw_value = bw_value;
|
|
req->comp_type = comp_type;
|
|
|
|
bw = hrt_value ? SHIFT_ROUND(hrt_value * 3, 1) : new_comp_limit;
|
|
hrt_port = hrt_value;
|
|
list_for_each_entry(enum_req, &(req->port_node), port_node) {
|
|
if (enum_req->hrt_value) {
|
|
bw += enum_req->hrt_value;
|
|
hrt_port = true;
|
|
} else
|
|
bw += get_comp_value(enum_req->bw_value,
|
|
enum_req->comp_type, false);
|
|
}
|
|
|
|
req->ostd = bw ? SHIFT_ROUND(bw, larb_req[larb].ratio[port]) : 1;
|
|
if (hrt_port) {
|
|
req->ostd = SHIFT_ROUND(req->ostd * 3, 1);
|
|
if (larb_req[larb].is_max_ostd)
|
|
req->ostd = max_ostd;
|
|
}
|
|
#if defined(USE_MEDIATEK_EMI)
|
|
#if IS_ENABLED(CONFIG_MACH_MT6877)
|
|
if (ddr_type == TYPE_LPDDR5) {
|
|
if (larb == 16 && port == 15)
|
|
req->ostd = 4;
|
|
if (larb == 17 && port == 15)
|
|
req->ostd = 4;
|
|
}
|
|
#endif
|
|
#endif
|
|
list_for_each_entry(enum_req, &(req->port_node), port_node)
|
|
enum_req->ostd = req->ostd;
|
|
|
|
if (log_level & 1 << log_bw)
|
|
pr_notice("mm_set=0x%08x bw=%u ostd=%u hrt=%u comp=%u\n",
|
|
req->master_id, req->bw_value, req->ostd,
|
|
req->hrt_value, req->comp_type);
|
|
|
|
mutex_unlock(&bw_mutex);
|
|
|
|
return 0;
|
|
}
|
|
EXPORT_SYMBOL_GPL(mm_qos_set_request);
|
|
|
|
s32 mm_qos_set_bw_request(struct mm_qos_request *req,
|
|
u32 bw_value, s32 comp_type)
|
|
{
|
|
return mm_qos_set_request(req, bw_value, req->hrt_value, comp_type);
|
|
}
|
|
EXPORT_SYMBOL_GPL(mm_qos_set_bw_request);
|
|
|
|
s32 mm_qos_set_hrt_request(struct mm_qos_request *req,
|
|
u32 hrt_value)
|
|
{
|
|
return mm_qos_set_request(req, req->bw_value, hrt_value, 0);
|
|
}
|
|
EXPORT_SYMBOL_GPL(mm_qos_set_hrt_request);
|
|
|
|
static u64 cam_scen_start_time;
|
|
static bool cam_scen_change;
|
|
void mm_qos_update_all_request(struct plist_head *owner_list)
|
|
{
|
|
struct mm_qos_request *req = NULL;
|
|
u64 profile;
|
|
u32 i = 0, larb_update = 0, mm_bw = 0;
|
|
s32 next_hrt_bw;
|
|
s32 cam_bw;
|
|
u32 larb_count = 0, larb_id = 0, larb_port_id = 0, larb_port_bw = 0;
|
|
u32 port_id = 0;
|
|
u32 comm, comm_port;
|
|
s32 smi_srt_clk = 0, smi_hrt_clk = 0;
|
|
s32 max_ch_srt_bw = 0, max_ch_hrt_bw = 0;
|
|
s32 final_chn_hrt_bw[MAX_COMM_NUM][MAX_CH_COUNT];
|
|
#ifdef CHECK_OSTD_UPDATE
|
|
bool update_ostd;
|
|
struct mm_qos_request *enum_req = NULL;
|
|
#endif
|
|
|
|
if (!owner_list || plist_head_empty(owner_list)) {
|
|
pr_notice("%s: owner_list is invalid\n", __func__);
|
|
return;
|
|
}
|
|
|
|
req = plist_first_entry(owner_list, struct mm_qos_request, owner_node);
|
|
|
|
if (is_camera_larb(req->master_id)) {
|
|
cam_bw = dram_write_weight(get_cam_hrt_bw());
|
|
if (cam_bw > camera_max_bw) {
|
|
pr_notice("cam_bw(%d) > camera_max_bw(%d)\n",
|
|
cam_bw, camera_max_bw);
|
|
#ifdef MMDVFS_MMP
|
|
mmprofile_log_ex(
|
|
mmdvfs_mmp_events.cam_bw_mismatch,
|
|
MMPROFILE_FLAG_PULSE,
|
|
cam_bw, camera_max_bw);
|
|
#endif
|
|
#ifdef AEE_CAM_BW_MISMATCH
|
|
aee_kernel_warning("mmdvfs",
|
|
"cam_bw(%d) > camera_max_bw(%d)\n",
|
|
cam_bw, camera_max_bw);
|
|
#endif
|
|
}
|
|
if (cam_scen_change) {
|
|
pr_notice("scenario change time=%u cam_bw=%d\n",
|
|
jiffies_to_msecs(jiffies-cam_scen_start_time),
|
|
cam_bw);
|
|
cam_scen_change = false;
|
|
}
|
|
#ifdef BLOCKING_MECHANISM
|
|
if (atomic_read(&lock_cam_count) > 0)
|
|
blocking_camera();
|
|
#endif
|
|
if (total_hrt_bw != UNINITIALIZED_VALUE &&
|
|
get_total_used_hrt_bw() > total_hrt_bw)
|
|
pr_notice("hrt bw overflow used=%d avail=%d\n",
|
|
get_total_used_hrt_bw(), total_hrt_bw);
|
|
}
|
|
|
|
mutex_lock(&bw_mutex);
|
|
next_hrt_bw = get_total_used_hrt_bw();
|
|
if (next_hrt_bw != current_hrt_bw) {
|
|
#ifdef HRT_MECHANISM
|
|
update_hrt_bw_to_dvfsrc(next_hrt_bw);
|
|
log_hrt_bw_info(req->master_id);
|
|
#endif
|
|
current_hrt_bw = next_hrt_bw;
|
|
}
|
|
mutex_unlock(&bw_mutex);
|
|
|
|
if (log_level & 1 << log_qos_validation)
|
|
trace_qos_validation();
|
|
|
|
plist_for_each_entry(req, owner_list, owner_node) {
|
|
if (!req->updated)
|
|
continue;
|
|
i++;
|
|
larb_id = SMI_PMQOS_LARB_DEC(req->master_id);
|
|
port_id = SMI_PMQOS_PORT_MASK(req->master_id);
|
|
get_comm_port_by_id(larb_req[larb_id].comm_port,
|
|
&comm, &comm_port);
|
|
larb_update |= 1 << (comm * SMI_COMM_MASTER_NUM + comm_port);
|
|
if (log_level & 1 << log_bw)
|
|
pr_notice("update(0x%08x) ostd=%d value=%d hrt=%d\n",
|
|
req->master_id, req->ostd,
|
|
req->bw_value, req->hrt_value);
|
|
trace_mmqos__update_port(larb_id, port_id,
|
|
req->bw_value, req->ostd);
|
|
if (larb_port_id && larb_count == 4) {
|
|
#ifdef MMDVFS_MMP
|
|
mmprofile_log_ex(mmdvfs_mmp_events.larb_port,
|
|
MMPROFILE_FLAG_PULSE,
|
|
larb_port_id, larb_port_bw);
|
|
#endif
|
|
larb_count = larb_port_bw = larb_port_id = 0;
|
|
}
|
|
if (mmdvfs_log_larb_mmp(-1, larb_id)) {
|
|
larb_port_bw |= req->ostd << (8 * larb_count);
|
|
larb_port_id |= port_id << (8 * larb_count);
|
|
larb_count++;
|
|
}
|
|
#ifdef CHECK_OSTD_UPDATE
|
|
mutex_lock(&bw_mutex);
|
|
if (!req->bw_value && !req->hrt_value) {
|
|
update_ostd = false;
|
|
list_for_each_entry(enum_req,
|
|
&(req->port_node), port_node) {
|
|
if (enum_req->bw_value ||
|
|
enum_req->hrt_value) {
|
|
update_ostd = true;
|
|
break;
|
|
}
|
|
}
|
|
req->updated = update_ostd;
|
|
}
|
|
mutex_unlock(&bw_mutex);
|
|
#endif
|
|
}
|
|
#ifdef MMDVFS_MMP
|
|
if (larb_count)
|
|
mmprofile_log_ex(
|
|
mmdvfs_mmp_events.larb_port,
|
|
MMPROFILE_FLAG_PULSE, larb_port_id, larb_port_bw);
|
|
#endif
|
|
if (!skip_smi_config) {
|
|
profile = sched_clock();
|
|
smi_ostd_update(owner_list, "MMDVFS");
|
|
if (log_level & 1 << log_bw)
|
|
pr_notice("config SMI (%d) cost: %llu us\n",
|
|
i, div_u64(sched_clock() - profile, 1000));
|
|
}
|
|
|
|
/* update SMI clock */
|
|
for (comm = 0; comm < MAX_COMM_NUM; comm++) {
|
|
if (comm_freq_class[comm] == 0)
|
|
continue;
|
|
max_ch_srt_bw = 0;
|
|
max_ch_hrt_bw = 0;
|
|
for (i = 0; i < MAX_CH_COUNT; i++) {
|
|
/* channel_hrt_bw[] doesn't contain disp HRT BW, so
|
|
* add one HRT BW to it if disp HRT count > 0
|
|
*/
|
|
final_chn_hrt_bw[comm][i] =
|
|
channel_disp_hrt_cnt[comm][i] > 0 ?
|
|
channel_hrt_bw[comm][i] +
|
|
larb_req[SMI_PMQOS_LARB_DEC(
|
|
PORT_VIRTUAL_DISP)].total_hrt_data :
|
|
channel_hrt_bw[comm][i];
|
|
max_ch_srt_bw = max_t(s32,
|
|
channel_srt_bw[comm][i], max_ch_srt_bw);
|
|
max_ch_hrt_bw = max_t(s32,
|
|
final_chn_hrt_bw[comm][i], max_ch_hrt_bw);
|
|
if (log_level & 1 << log_smi_freq)
|
|
pr_notice("comm:%d chn:%d s_bw:%d h_bw:%d\n",
|
|
comm, i, channel_srt_bw[comm][i],
|
|
final_chn_hrt_bw[comm][i]);
|
|
#ifdef MMDVFS_MMP
|
|
mmprofile_log_ex(
|
|
mmdvfs_mmp_events.smi_freq,
|
|
MMPROFILE_FLAG_PULSE,
|
|
((comm+1) << 28) | (i << 24) | min_t(s32,
|
|
channel_srt_bw[comm][i], 0xffff),
|
|
((comm+1) << 28) | (i << 24) | min_t(s32,
|
|
final_chn_hrt_bw[comm][i], 0xffff));
|
|
#endif
|
|
}
|
|
smi_srt_clk = max_ch_srt_bw ?
|
|
SHIFT_ROUND(max_ch_srt_bw, 4) : 0;
|
|
smi_hrt_clk = max_ch_hrt_bw ?
|
|
SHIFT_ROUND(max_ch_hrt_bw, 4) : 0;
|
|
mtk_pm_qos_update_request(&smi_freq_request[comm],
|
|
max_t(s32, smi_srt_clk, smi_hrt_clk));
|
|
if (log_level & 1 << log_smi_freq)
|
|
pr_notice("comm:%d smi_srt_clk:%d smi_hrt_clk:%d\n",
|
|
comm, smi_srt_clk, smi_hrt_clk);
|
|
#ifdef MMDVFS_MMP
|
|
mmprofile_log_ex(
|
|
mmdvfs_mmp_events.smi_freq,
|
|
MMPROFILE_FLAG_PULSE,
|
|
comm, (min_t(s32, smi_srt_clk, 0xffff) << 16) |
|
|
min_t(s32, smi_hrt_clk, 0xffff));
|
|
#endif
|
|
}
|
|
|
|
mutex_lock(&bw_mutex);
|
|
/* update larb-level BW */
|
|
if (!skip_smi_config)
|
|
mm_qos_update_larb_bwl(larb_update, true);
|
|
|
|
#ifdef QOS_BOUND_DETECT
|
|
mmdvfs_update_qos_sram(larb_req, larb_update);
|
|
#endif
|
|
|
|
/* update mm total bw */
|
|
for (i = 0; i < MAX_LARB_COUNT; i++)
|
|
mm_bw += (larb_req[i].comm_port != SMI_COMM_MASTER_NUM) ?
|
|
larb_req[i].total_bw_data : 0;
|
|
mtk_pm_qos_update_request(&mm_bw_request, mm_bw);
|
|
if (log_level & 1 << log_bw)
|
|
pr_notice("config mm_bw=%d\n", mm_bw);
|
|
mutex_unlock(&bw_mutex);
|
|
}
|
|
EXPORT_SYMBOL_GPL(mm_qos_update_all_request);
|
|
|
|
void mm_qos_update_all_request_zero(struct plist_head *owner_list)
|
|
{
|
|
struct mm_qos_request *req = NULL;
|
|
|
|
plist_for_each_entry(req, owner_list, owner_node) {
|
|
mm_qos_set_request(req, 0, 0, 0);
|
|
}
|
|
mm_qos_update_all_request(owner_list);
|
|
}
|
|
EXPORT_SYMBOL_GPL(mm_qos_update_all_request_zero);
|
|
|
|
void mm_qos_remove_all_request(struct plist_head *owner_list)
|
|
{
|
|
struct mm_qos_request *temp, *req = NULL;
|
|
|
|
mutex_lock(&bw_mutex);
|
|
plist_for_each_entry_safe(req, temp, owner_list, owner_node) {
|
|
pr_notice("mm_del(0x%08x)\n", req->master_id);
|
|
plist_del(&(req->owner_node), owner_list);
|
|
list_del(&(req->larb_node));
|
|
list_del(&(req->port_node));
|
|
req->init = false;
|
|
}
|
|
mutex_unlock(&bw_mutex);
|
|
}
|
|
EXPORT_SYMBOL_GPL(mm_qos_remove_all_request);
|
|
|
|
static s32 disp_bw_ceiling;
|
|
static bool wait_next_max_cam_bw_set;
|
|
s32 mm_hrt_get_available_hrt_bw(u32 master_id)
|
|
{
|
|
s32 total_used_hrt_bw = get_total_used_hrt_bw();
|
|
s32 src_hrt_bw = larb_req[SMI_PMQOS_LARB_DEC(master_id)].total_hrt_data;
|
|
s32 cam_occ_bw;
|
|
s32 cam_occ_max_bw;
|
|
s32 result;
|
|
|
|
if (total_hrt_bw == UNINITIALIZED_VALUE)
|
|
return UNINITIALIZED_VALUE;
|
|
if (total_ui_only_hrt_bw == UNINITIALIZED_VALUE)
|
|
return UNINITIALIZED_VALUE;
|
|
|
|
cam_occ_bw = dram_write_weight(MULTIPLY_RATIO(get_cam_hrt_bw())/cam_occ_ratio());
|
|
if (is_camera_larb(master_id))
|
|
src_hrt_bw = cam_occ_bw;
|
|
else
|
|
src_hrt_bw = MULTIPLY_RATIO(src_hrt_bw)/disp_occ_ratio();
|
|
|
|
if (camera_max_bw > 0)
|
|
result = total_hrt_bw - total_used_hrt_bw + src_hrt_bw;
|
|
else
|
|
result = total_ui_only_hrt_bw - total_used_hrt_bw + src_hrt_bw;
|
|
|
|
if (SMI_PMQOS_LARB_DEC(master_id) ==
|
|
SMI_PMQOS_LARB_DEC(PORT_VIRTUAL_DISP)) {
|
|
/* Consider worst camera bw if camera is on */
|
|
cam_occ_max_bw = MULTIPLY_RATIO(camera_max_bw)/cam_occ_ratio();
|
|
if (cam_occ_max_bw > 0)
|
|
result = result + cam_occ_bw - cam_occ_max_bw;
|
|
|
|
if (disp_bw_ceiling > 0 && !wait_next_max_cam_bw_set
|
|
&& disp_bw_ceiling < result)
|
|
result = disp_bw_ceiling;
|
|
}
|
|
|
|
if (is_camera_larb(master_id))
|
|
result = DIVIDE_RATIO(result * cam_occ_ratio());
|
|
else
|
|
result = DIVIDE_RATIO(result * disp_occ_ratio());
|
|
return ((result < 0)?0:result);
|
|
}
|
|
EXPORT_SYMBOL_GPL(mm_hrt_get_available_hrt_bw);
|
|
|
|
s32 mm_hrt_add_bw_throttle_notifier(struct notifier_block *nb)
|
|
{
|
|
return blocking_notifier_chain_register(
|
|
&hrt_bw_throttle_notifier,
|
|
nb);
|
|
}
|
|
EXPORT_SYMBOL_GPL(mm_hrt_add_bw_throttle_notifier);
|
|
|
|
s32 mm_hrt_remove_bw_throttle_notifier(struct notifier_block *nb)
|
|
{
|
|
return blocking_notifier_chain_unregister(
|
|
&hrt_bw_throttle_notifier,
|
|
nb);
|
|
}
|
|
EXPORT_SYMBOL_GPL(mm_hrt_remove_bw_throttle_notifier);
|
|
|
|
s32 add_cam_max_bw_notifier(struct notifier_block *nb)
|
|
{
|
|
return blocking_notifier_chain_register(
|
|
&cam_max_bw_notifier,
|
|
nb);
|
|
}
|
|
EXPORT_SYMBOL_GPL(add_cam_max_bw_notifier);
|
|
|
|
s32 remove_cam_max_bw_notifier(struct notifier_block *nb)
|
|
{
|
|
return blocking_notifier_chain_unregister(
|
|
&cam_max_bw_notifier,
|
|
nb);
|
|
}
|
|
EXPORT_SYMBOL_GPL(remove_cam_max_bw_notifier);
|
|
|
|
#ifdef HRT_MECHANISM
|
|
static int notify_bw_throttle(void *data)
|
|
{
|
|
u64 start_jiffies = jiffies;
|
|
|
|
blocking_notifier_call_chain(&hrt_bw_throttle_notifier,
|
|
(camera_max_bw > 0)?BW_THROTTLE_START:BW_THROTTLE_END, NULL);
|
|
|
|
pr_notice("notify_time=%u\n",
|
|
jiffies_to_msecs(jiffies-start_jiffies));
|
|
return 0;
|
|
}
|
|
|
|
#ifdef BLOCKING_MECHANISM
|
|
static int notify_bw_throttle_blocking(void *data)
|
|
{
|
|
notify_bw_throttle(data);
|
|
|
|
atomic_dec(&lock_cam_count);
|
|
wake_up(&hrt_wait);
|
|
pr_notice("decrease lock_cam_count=%d\n",
|
|
atomic_read(&lock_cam_count));
|
|
return 0;
|
|
}
|
|
#endif
|
|
|
|
static u32 camera_overlap_bw;
|
|
static void set_camera_max_bw(u32 occ_bw)
|
|
{
|
|
struct task_struct *pKThread;
|
|
|
|
camera_max_bw = occ_bw;
|
|
wait_next_max_cam_bw_set = false;
|
|
pr_notice("set cam max occupy_bw=%d\n", occ_bw);
|
|
#ifdef BLOCKING_MECHANISM
|
|
/* No need to blocking if cam bw is decreasing */
|
|
if (camera_overlap_bw == 0) {
|
|
atomic_inc(&lock_cam_count);
|
|
pr_notice("increase lock_cam_count=%d\n",
|
|
atomic_read(&lock_cam_count));
|
|
pKThread = kthread_run(notify_bw_throttle_blocking,
|
|
NULL, "notify bw throttle blocking");
|
|
return;
|
|
}
|
|
#endif
|
|
pKThread = kthread_run(notify_bw_throttle,
|
|
NULL, "notify bw throttle");
|
|
}
|
|
|
|
static void delay_work_handler(struct work_struct *work)
|
|
{
|
|
set_camera_max_bw(camera_overlap_bw);
|
|
}
|
|
static DECLARE_DELAYED_WORK(g_delay_work, delay_work_handler);
|
|
#endif /* HRT_MECHANISM */
|
|
|
|
#ifdef SMI_TAL
|
|
static int notify_bw_throttle(void *data)
|
|
{
|
|
u64 start_jiffies = jiffies;
|
|
|
|
blocking_notifier_call_chain(&cam_max_bw_notifier,
|
|
camera_max_bw, NULL);
|
|
|
|
pr_notice("notify_time=%u\n",
|
|
jiffies_to_msecs(jiffies-start_jiffies));
|
|
return 0;
|
|
}
|
|
|
|
static u32 camera_overlap_bw;
|
|
static void set_camera_max_bw(u32 occ_bw)
|
|
{
|
|
struct task_struct *pKThread;
|
|
|
|
camera_max_bw = occ_bw;
|
|
wait_next_max_cam_bw_set = false;
|
|
pr_notice("set cam max occupy_bw=%d\n", occ_bw);
|
|
pKThread = kthread_run(notify_bw_throttle,
|
|
NULL, "notify bw throttle");
|
|
}
|
|
static void delay_work_handler(struct work_struct *work)
|
|
{
|
|
set_camera_max_bw(camera_overlap_bw);
|
|
}
|
|
static DECLARE_DELAYED_WORK(g_delay_work, delay_work_handler);
|
|
#endif /* SMI_TAL */
|
|
|
|
|
|
#define MULTIPLY_W_DRAM_WEIGHT(value) ((value)*6/5) /* Write DRAM Weight*/
|
|
|
|
void mmdvfs_set_max_camera_hrt_bw(u32 bw)
|
|
{
|
|
#ifdef SMI_TAL
|
|
u32 mw_hrt_bw;
|
|
|
|
cancel_delayed_work_sync(&g_delay_work);
|
|
|
|
mw_hrt_bw = MULTIPLY_W_DRAM_WEIGHT(bw);
|
|
if (mw_hrt_bw < camera_max_bw) {
|
|
camera_overlap_bw = mw_hrt_bw;
|
|
schedule_delayed_work(&g_delay_work, 2 * HZ);
|
|
} else {
|
|
camera_overlap_bw = 0;
|
|
set_camera_max_bw(mw_hrt_bw);
|
|
}
|
|
|
|
pr_notice("middleware set max camera hrt bw:%d\n", bw);
|
|
|
|
#endif
|
|
|
|
#ifdef HRT_MECHANISM
|
|
u32 mw_hrt_bw;
|
|
|
|
cam_scen_change = true;
|
|
cam_scen_start_time = jiffies;
|
|
|
|
cancel_delayed_work_sync(&g_delay_work);
|
|
|
|
mw_hrt_bw = dram_write_weight(bw);
|
|
if (mw_hrt_bw < camera_max_bw) {
|
|
camera_overlap_bw = mw_hrt_bw;
|
|
schedule_delayed_work(&g_delay_work, 2 * HZ);
|
|
} else {
|
|
camera_overlap_bw = 0;
|
|
set_camera_max_bw(mw_hrt_bw);
|
|
}
|
|
|
|
pr_notice("middleware set max camera hrt bw:%d\n", bw);
|
|
#endif
|
|
}
|
|
EXPORT_SYMBOL_GPL(mmdvfs_set_max_camera_hrt_bw);
|
|
|
|
static s32 get_total_hrt_bw(bool ui_only)
|
|
{
|
|
s32 result = 0;
|
|
#if defined(USE_MEDIATEK_EMI)
|
|
s32 max_freq = get_opp_ddr_freq(0)/1000;
|
|
s32 ch_num = mtk_emicen_get_ch_cnt();
|
|
s32 io_width = get_io_width();
|
|
|
|
if (ui_only)
|
|
result = DIVIDE_RATIO(max_freq * ch_num * io_width * emi_occ_ui_only());
|
|
else
|
|
result = DIVIDE_RATIO(max_freq * ch_num * io_width * emi_occ_ratio());
|
|
#elif defined(USE_MTK_DRAMC)
|
|
s32 max_freq = dram_steps_freq(0);
|
|
s32 ch_num = get_emi_ch_num();
|
|
s32 io_width = get_io_width();
|
|
|
|
result = MULTIPLY_BW_THRESH_HIGH(max_freq * ch_num * io_width);
|
|
#else
|
|
result = UNINITIALIZED_VALUE;
|
|
#endif
|
|
return result;
|
|
}
|
|
|
|
static void get_module_clock_by_index(struct device *dev,
|
|
u32 index, struct clk **clk_module)
|
|
{
|
|
const char *clk_name;
|
|
s32 result;
|
|
|
|
result = of_property_read_string_index(dev->of_node, "clock-names",
|
|
index, &clk_name);
|
|
if (unlikely(result)) {
|
|
pr_notice("Cannot get module name of index (%u), result (%d)\n",
|
|
index, result);
|
|
return;
|
|
}
|
|
|
|
*clk_module = devm_clk_get(dev, clk_name);
|
|
if (IS_ERR(*clk_module)) {
|
|
/* error status print */
|
|
pr_notice("Cannot get module clock: %s\n", clk_name);
|
|
*clk_module = NULL;
|
|
} else {
|
|
/* message print */
|
|
pr_notice("Get module clock: %s\n", clk_name);
|
|
}
|
|
}
|
|
|
|
static void mmdvfs_get_step_node(struct device *dev,
|
|
const char *name, struct mm_freq_step_config *step_config)
|
|
{
|
|
s32 result;
|
|
u32 step[mm_dp_max] = {0};
|
|
|
|
result = of_property_read_u32_array(dev->of_node,
|
|
name, step, mm_dp_max);
|
|
if (likely(!result)) {
|
|
step_config->freq_step = step[mm_dp_freq];
|
|
step_config->clk_type |= step[mm_dp_clk_type];
|
|
if (step[mm_dp_clk_type] == CLK_TYPE_MUX) {
|
|
step_config->clk_mux_id =
|
|
step[mm_dp_clk_mux];
|
|
step_config->clk_source_id =
|
|
step[mm_dp_clk_source];
|
|
get_module_clock_by_index(dev,
|
|
step[mm_dp_clk_mux],
|
|
&step_config->clk_mux);
|
|
get_module_clock_by_index(dev,
|
|
step[mm_dp_clk_source],
|
|
&step_config->clk_source);
|
|
} else if (step[mm_dp_clk_type] == CLK_TYPE_PLL) {
|
|
step_config->pll_id =
|
|
step[mm_dp_pll_id];
|
|
step_config->pll_value =
|
|
step[mm_dp_pll_value];
|
|
}
|
|
pr_notice("%s: %lluMHz, clk:%u/%u/%u\n",
|
|
name, step_config->freq_step,
|
|
step_config->clk_type,
|
|
step[mm_dp_clk_param1], step[mm_dp_clk_param2]);
|
|
} else {
|
|
pr_notice("read freq steps %s failed (%d)\n", name, result);
|
|
}
|
|
}
|
|
|
|
static void mmdvfs_get_larb_node(struct device *dev, u32 larb_id)
|
|
{
|
|
u32 value, count = 0;
|
|
const __be32 *p;
|
|
struct property *prop;
|
|
char larb_name[MAX_LARB_NAME];
|
|
s32 ret = 0;
|
|
|
|
if (larb_id >= MAX_LARB_COUNT) {
|
|
pr_notice("larb_id:%d is over MAX_LARB_COUNT:%d\n",
|
|
larb_id, MAX_LARB_COUNT);
|
|
return;
|
|
}
|
|
|
|
|
|
ret = snprintf(larb_name, MAX_LARB_NAME, "larb%d", larb_id);
|
|
if (ret < 0)
|
|
pr_notice("snprintf return error, ret:%d, larb_id:%d\n", ret, larb_id);
|
|
of_property_for_each_u32(dev->of_node, larb_name, prop, p, value) {
|
|
if (count >= MAX_PORT_COUNT) {
|
|
pr_notice("port size is over (%d)\n", MAX_PORT_COUNT);
|
|
break;
|
|
}
|
|
|
|
larb_req[larb_id].ratio[count] = value;
|
|
count++;
|
|
}
|
|
|
|
larb_req[larb_id].port_count = count;
|
|
if (!count)
|
|
pr_notice("no data in larb (%s)\n", larb_name);
|
|
else
|
|
init_larb_list(larb_id);
|
|
}
|
|
|
|
static void init_virtual_larbs(void)
|
|
{
|
|
larb_req[SMI_PMQOS_LARB_DEC(PORT_VIRTUAL_DISP)].port_count = 1;
|
|
larb_req[SMI_PMQOS_LARB_DEC(PORT_VIRTUAL_DISP)].ratio[0] = 1;
|
|
larb_req[SMI_PMQOS_LARB_DEC(PORT_VIRTUAL_DISP)].channel = MAX_CH_COUNT;
|
|
larb_req[SMI_PMQOS_LARB_DEC(
|
|
PORT_VIRTUAL_DISP)].comm_port = SMI_COMM_MASTER_NUM;
|
|
init_larb_list(SMI_PMQOS_LARB_DEC(PORT_VIRTUAL_DISP));
|
|
|
|
larb_req[SMI_PMQOS_LARB_DEC(PORT_VIRTUAL_CCU_COMMON)].port_count = 1;
|
|
larb_req[SMI_PMQOS_LARB_DEC(PORT_VIRTUAL_CCU_COMMON)].ratio[0] = 1;
|
|
larb_req[SMI_PMQOS_LARB_DEC(PORT_VIRTUAL_CCU_COMMON)].channel =
|
|
SMI_COMM_BUS_SEL[mmdvfs_get_ccu_smi_common_port(
|
|
PORT_VIRTUAL_CCU_COMMON) & 0xffff];
|
|
larb_req[SMI_PMQOS_LARB_DEC(
|
|
PORT_VIRTUAL_CCU_COMMON)].comm_port =
|
|
mmdvfs_get_ccu_smi_common_port(PORT_VIRTUAL_CCU_COMMON);
|
|
init_larb_list(SMI_PMQOS_LARB_DEC(PORT_VIRTUAL_CCU_COMMON));
|
|
|
|
larb_req[SMI_PMQOS_LARB_DEC(PORT_VIRTUAL_CCU_COMMON2)].port_count = 1;
|
|
larb_req[SMI_PMQOS_LARB_DEC(PORT_VIRTUAL_CCU_COMMON2)].ratio[0] = 1;
|
|
larb_req[SMI_PMQOS_LARB_DEC(PORT_VIRTUAL_CCU_COMMON2)].channel =
|
|
SMI_COMM_BUS_SEL[mmdvfs_get_ccu_smi_common_port(
|
|
PORT_VIRTUAL_CCU_COMMON2) & 0xffff];
|
|
larb_req[SMI_PMQOS_LARB_DEC(
|
|
PORT_VIRTUAL_CCU_COMMON2)].comm_port =
|
|
mmdvfs_get_ccu_smi_common_port(PORT_VIRTUAL_CCU_COMMON2);
|
|
init_larb_list(SMI_PMQOS_LARB_DEC(PORT_VIRTUAL_CCU_COMMON2));
|
|
|
|
larb_req[SMI_PMQOS_LARB_DEC(PORT_VIRTUAL_MD)].port_count = 1;
|
|
larb_req[SMI_PMQOS_LARB_DEC(PORT_VIRTUAL_MD)].ratio[0] = 1;
|
|
larb_req[SMI_PMQOS_LARB_DEC(PORT_VIRTUAL_MD)].channel = MAX_CH_COUNT;
|
|
larb_req[SMI_PMQOS_LARB_DEC(
|
|
PORT_VIRTUAL_MD)].comm_port = SMI_COMM_MASTER_NUM;
|
|
init_larb_list(SMI_PMQOS_LARB_DEC(PORT_VIRTUAL_MD));
|
|
larb_req[SMI_PMQOS_LARB_DEC(PORT_VIRTUAL_MD)].total_hrt_data =
|
|
get_md_hrt_bw();
|
|
}
|
|
|
|
static void mmdvfs_get_step_array_node(struct device *dev,
|
|
const char *freq_name, struct mm_freq_step_config step_configs[])
|
|
{
|
|
struct property *prop;
|
|
u32 count = 0;
|
|
const char *name;
|
|
char ext_name[32] = {0};
|
|
|
|
pr_notice("start get step node of %s\n", freq_name);
|
|
of_property_for_each_string(dev->of_node, freq_name, prop, name) {
|
|
if (count >= MAX_FREQ_STEP) {
|
|
pr_notice("freq setting %s is over the MAX_STEP (%d)\n",
|
|
freq_name, MAX_FREQ_STEP);
|
|
break;
|
|
}
|
|
pr_notice(" node name %s\n", name);
|
|
mmdvfs_get_step_node(dev, name, &step_configs[count]);
|
|
strncpy(ext_name, name, sizeof(ext_name)-1);
|
|
strncat(ext_name, "_ext",
|
|
sizeof(ext_name)-strlen(name)-1);
|
|
mmdvfs_get_step_node(dev,
|
|
ext_name, &step_configs[count]);
|
|
count++;
|
|
}
|
|
if (count != step_size)
|
|
pr_notice("freq setting %s is not same as vcore_steps (%d)\n",
|
|
freq_name, step_size);
|
|
pr_notice("%s: step size:%u\n", freq_name, step_size);
|
|
}
|
|
|
|
static void mmdvfs_get_limit_step_node(struct device *dev,
|
|
const char *freq_name,
|
|
struct mm_freq_limit_config *limit_config)
|
|
{
|
|
#ifdef MMDVFS_LIMIT
|
|
s32 result, i;
|
|
char ext_name[32] = {0};
|
|
u32 limit_size = 0;
|
|
|
|
strncpy(ext_name, freq_name, sizeof(ext_name)-1);
|
|
strncat(ext_name, "_limit_size",
|
|
sizeof(ext_name)-strlen(freq_name)-1);
|
|
result = of_property_read_u32(dev->of_node, ext_name, &limit_size);
|
|
if (result < 0 || !limit_size)
|
|
return;
|
|
|
|
pr_notice("[limit]%s size: %u\n", freq_name, limit_size);
|
|
limit_config->limit_size = limit_size;
|
|
limit_config->limit_steps = kcalloc(limit_size,
|
|
sizeof(*limit_config->limit_steps), GFP_KERNEL);
|
|
for (i = 0; i < limit_size; i++) {
|
|
limit_config->limit_steps[i] = kcalloc(MAX_FREQ_STEP,
|
|
sizeof(*limit_config->limit_steps[i]), GFP_KERNEL);
|
|
result = snprintf(ext_name, sizeof(ext_name) - 1,
|
|
"%s_limit_%d", freq_name, i);
|
|
if (result < 0) {
|
|
pr_notice("snprintf error(%d) limit name:%s id:%d\n",
|
|
result, freq_name, i);
|
|
continue;
|
|
}
|
|
pr_notice("[limit]%s-%d: %s\n", freq_name, i, ext_name);
|
|
mmdvfs_get_step_array_node(dev, ext_name,
|
|
limit_config->limit_steps[i]);
|
|
}
|
|
#else
|
|
pr_notice("MMDVFS limit is off\n");
|
|
#endif
|
|
}
|
|
|
|
static int mmdvfs_probe(struct platform_device *pdev)
|
|
{
|
|
u32 i, value, comm_count = 0;
|
|
struct device_node *node = pdev->dev.of_node;
|
|
struct property *prop;
|
|
struct mm_freq_config *mm_freq;
|
|
const __be32 *p;
|
|
u64 freq_steps[MAX_FREQ_STEP] = {0};
|
|
const char *mux_name;
|
|
|
|
#ifdef MMDVFS_MMP
|
|
mmprofile_enable(1);
|
|
if (mmdvfs_mmp_events.mmdvfs == 0) {
|
|
mmdvfs_mmp_events.mmdvfs =
|
|
mmprofile_register_event(MMP_ROOT_EVENT, "MMDVFS");
|
|
mmdvfs_mmp_events.freq_change = mmprofile_register_event(
|
|
mmdvfs_mmp_events.mmdvfs, "freq_change");
|
|
mmdvfs_mmp_events.ext_freq_change = mmprofile_register_event(
|
|
mmdvfs_mmp_events.mmdvfs, "ext_freq_change");
|
|
mmdvfs_mmp_events.limit_change = mmprofile_register_event(
|
|
mmdvfs_mmp_events.mmdvfs, "limit_change");
|
|
mmdvfs_mmp_events.hrt_change = mmprofile_register_event(
|
|
mmdvfs_mmp_events.mmdvfs, "hrt_change");
|
|
mmdvfs_mmp_events.cam_bw_mismatch = mmprofile_register_event(
|
|
mmdvfs_mmp_events.mmdvfs, "cam_bw_mismatch");
|
|
mmdvfs_mmp_events.larb_soft_mode = mmprofile_register_event(
|
|
mmdvfs_mmp_events.mmdvfs, "larb_soft_mode");
|
|
mmdvfs_mmp_events.larb_bwl = mmprofile_register_event(
|
|
mmdvfs_mmp_events.mmdvfs, "larb_bwl");
|
|
mmdvfs_mmp_events.larb_port = mmprofile_register_event(
|
|
mmdvfs_mmp_events.mmdvfs, "larb_port");
|
|
mmdvfs_mmp_events.smi_freq = mmprofile_register_event(
|
|
mmdvfs_mmp_events.mmdvfs, "smi_freq");
|
|
mmprofile_enable_event_recursive(mmdvfs_mmp_events.mmdvfs, 1);
|
|
}
|
|
mmprofile_start(1);
|
|
#endif
|
|
|
|
mmdvfs_enable = true;
|
|
mmdvfs_autok_enable = true;
|
|
mtk_pm_qos_add_request(&vcore_request, MTK_PM_QOS_VCORE_OPP,
|
|
MTK_PM_QOS_VCORE_OPP_DEFAULT_VALUE);
|
|
mtk_pm_qos_add_request(&mm_bw_request, MTK_PM_QOS_MEMORY_BANDWIDTH,
|
|
MTK_PM_QOS_MEMORY_BANDWIDTH_DEFAULT_VALUE);
|
|
step_size = 0;
|
|
of_property_for_each_u32(node, VCORE_NODE_NAME, prop, p, value) {
|
|
if (step_size >= MAX_FREQ_STEP) {
|
|
pr_notice(
|
|
"vcore_steps is over the MAX_STEP (%d)\n",
|
|
MAX_FREQ_STEP);
|
|
break;
|
|
}
|
|
vopp_steps[step_size] = value;
|
|
step_size++;
|
|
}
|
|
|
|
mux_size = 0;
|
|
of_property_for_each_u32(node, FMETER_MUX_NODE_NAME, prop, p, value) {
|
|
if (mux_size >= MAX_MUX_SIZE) {
|
|
pr_notice(
|
|
"fmeter_mux_ids is over the MAX_MUX_SIZE (%d)\n",
|
|
MAX_MUX_SIZE);
|
|
break;
|
|
}
|
|
fmeter_mux_ids[mux_size] = value;
|
|
mux_size++;
|
|
}
|
|
|
|
pr_notice("vcore_steps: [%u, %u, %u, %u, %u, %u], count:%u\n",
|
|
vopp_steps[0], vopp_steps[1], vopp_steps[2],
|
|
vopp_steps[3], vopp_steps[4], vopp_steps[5], step_size);
|
|
|
|
for (i = 0; i < ARRAY_SIZE(all_freqs); i++) {
|
|
mm_freq = all_freqs[i];
|
|
mmdvfs_get_step_array_node(&pdev->dev, mm_freq->prop_name,
|
|
mm_freq->step_config);
|
|
|
|
if (likely(mm_freq->pm_qos_class >= PM_QOS_DISP_FREQ)) {
|
|
mtk_pm_qos_add_notifier(mm_freq->pm_qos_class,
|
|
&mm_freq->nb);
|
|
pr_notice("%s: add notifier\n", mm_freq->prop_name);
|
|
}
|
|
|
|
mmdvfs_get_limit_step_node(&pdev->dev, mm_freq->prop_name,
|
|
&mm_freq->limit_config);
|
|
}
|
|
|
|
of_property_for_each_string(node, "comm_freq", prop, mux_name) {
|
|
if (comm_count >= MAX_COMM_NUM) {
|
|
pr_notice("comm_count > MAX_COMM_NUM (%d)\n",
|
|
MAX_COMM_NUM);
|
|
break;
|
|
}
|
|
for (i = 0; i < ARRAY_SIZE(all_freqs); i++) {
|
|
if (!strcmp(mux_name, all_freqs[i]->prop_name)) {
|
|
comm_freq_class[comm_count] =
|
|
all_freqs[i]->pm_qos_class;
|
|
break;
|
|
}
|
|
}
|
|
if (i == ARRAY_SIZE(all_freqs)) {
|
|
pr_notice("wrong comm_freq name:%s\n", mux_name);
|
|
break;
|
|
}
|
|
mtk_pm_qos_add_request(&smi_freq_request[comm_count],
|
|
comm_freq_class[comm_count],
|
|
PM_QOS_MM_FREQ_DEFAULT_VALUE);
|
|
comm_count++;
|
|
}
|
|
|
|
cam_larb_size = 0;
|
|
of_property_for_each_u32(node, CAM_LARB_NODE_NAME, prop, p, value) {
|
|
if (cam_larb_size >= MAX_LARB_COUNT) {
|
|
pr_notice(
|
|
"cam_larb is over the MAX_LARB_COUNT (%d)\n",
|
|
MAX_LARB_COUNT);
|
|
break;
|
|
}
|
|
cam_larb_ids[cam_larb_size] = value;
|
|
cam_larb_size++;
|
|
}
|
|
|
|
of_property_for_each_u32(
|
|
node, MAX_OSTD_LARB_NODE_NAME, prop, p, value) {
|
|
if (value >= MAX_LARB_COUNT) {
|
|
pr_notice(
|
|
"max_ostd_larb (%d) is over the MAX_LARB_COUNT (%d)\n",
|
|
value, MAX_LARB_COUNT);
|
|
continue;
|
|
}
|
|
larb_req[value].is_max_ostd = true;
|
|
}
|
|
of_property_read_s32(node, MAX_OSTD_NODE_NAME, &max_ostd);
|
|
if (max_ostd != UNINITIALIZED_VALUE)
|
|
max_bw_bound = max_ostd * 256 * 2; /* 256:Write BW, 2: HRT */
|
|
|
|
of_property_for_each_u32(node, LARB_NODE_NAME, prop, p, value) {
|
|
mmdvfs_get_larb_node(&pdev->dev, value);
|
|
}
|
|
|
|
#ifdef HRT_MECHANISM
|
|
#ifdef SIMULATE_DVFSRC
|
|
init_simulation();
|
|
#else
|
|
init_dvfsrc();
|
|
#endif
|
|
#endif
|
|
|
|
if (SMI_LARB_NUM != 0)
|
|
init_virtual_larbs();
|
|
|
|
for (i = 0; i < SMI_LARB_NUM; i++) {
|
|
value = SMI_LARB_L1ARB[i];
|
|
larb_req[i].comm_port = value;
|
|
if (value != SMI_COMM_MASTER_NUM)
|
|
larb_req[i].channel =
|
|
SMI_COMM_BUS_SEL[value & 0xffff];
|
|
pr_notice("larb[%d].comm_port=%d channel=%d\n",
|
|
i, value, larb_req[i].channel);
|
|
}
|
|
|
|
mmdvfs_qos_get_freq_steps(PM_QOS_DISP_FREQ, freq_steps, &value);
|
|
pr_notice("disp step size:%u\n", value);
|
|
for (i = 0; i < value && i < MAX_FREQ_STEP; i++)
|
|
pr_notice(" - step[%d]: %llu\n", i, freq_steps[i]);
|
|
|
|
#ifdef BLOCKING_MECHANISM
|
|
init_waitqueue_head(&hrt_wait);
|
|
#endif
|
|
|
|
vcore_reg_id = regulator_get(&pdev->dev, "vcore");
|
|
if (!vcore_reg_id)
|
|
pr_info("regulator_get vcore_reg_id failed\n");
|
|
return 0;
|
|
|
|
}
|
|
|
|
static int mmdvfs_remove(struct platform_device *pdev)
|
|
{
|
|
u32 i;
|
|
|
|
mtk_pm_qos_remove_request(&vcore_request);
|
|
mtk_pm_qos_remove_request(&mm_bw_request);
|
|
for (i = 0; i < MAX_COMM_NUM; i++) {
|
|
if (comm_freq_class[i] == 0)
|
|
continue;
|
|
mtk_pm_qos_remove_request(&smi_freq_request[i]);
|
|
}
|
|
for (i = 0; i < ARRAY_SIZE(all_freqs); i++)
|
|
pm_qos_remove_notifier(
|
|
all_freqs[i]->pm_qos_class, &all_freqs[i]->nb);
|
|
|
|
#ifdef HRT_MECHANISM
|
|
#ifdef SIMULATE_DVFSRC
|
|
mtk_pm_qos_remove_request(&ddr_request);
|
|
#else
|
|
mtk_pm_qos_remove_request(&dvfsrc_isp_hrt_req);
|
|
#endif
|
|
#endif
|
|
|
|
return 0;
|
|
}
|
|
|
|
static const struct of_device_id mmdvfs_of_ids[] = {
|
|
{.compatible = "mediatek,mmdvfs_pmqos",},
|
|
{}
|
|
};
|
|
|
|
static struct platform_driver mmdvfs_pmqos_driver = {
|
|
.probe = mmdvfs_probe,
|
|
.remove = mmdvfs_remove,
|
|
.driver = {
|
|
.name = "mtk_mmdvfs_pmqos",
|
|
.owner = THIS_MODULE,
|
|
.of_match_table = mmdvfs_of_ids,
|
|
}
|
|
};
|
|
|
|
static int __init mmdvfs_pmqos_init(void)
|
|
{
|
|
#ifdef CONFIG_FPGA_EARLY_PORTING
|
|
return 0;
|
|
#else
|
|
s32 status;
|
|
|
|
status = platform_driver_register(&mmdvfs_pmqos_driver);
|
|
if (status != 0) {
|
|
pr_notice(
|
|
"Failed to register MMDVFS-PMQOS driver(%d)\n", status);
|
|
return -ENODEV;
|
|
}
|
|
|
|
pr_notice("%s\n", __func__);
|
|
return 0;
|
|
#endif /* CONFIG_FPGA_EARLY_PORTING */
|
|
}
|
|
|
|
#ifdef QOS_BOUND_DETECT
|
|
static int system_qos_update(struct notifier_block *nb,
|
|
unsigned long qos_status, void *v)
|
|
{
|
|
larb_soft = !(qos_status > QOS_BOUND_BW_FREE);
|
|
#ifdef MMDVFS_MMP
|
|
mmprofile_log_ex(
|
|
mmdvfs_mmp_events.larb_soft_mode,
|
|
MMPROFILE_FLAG_PULSE, larb_soft, qos_status);
|
|
#endif
|
|
if (likely(force_larb_mode < 0) && !skip_smi_config) {
|
|
mutex_lock(&bw_mutex);
|
|
mm_qos_update_larb_bwl(0xFFFF, false);
|
|
mutex_unlock(&bw_mutex);
|
|
}
|
|
|
|
return NOTIFY_OK;
|
|
}
|
|
|
|
struct system_qos_status {
|
|
struct notifier_block nb;
|
|
};
|
|
|
|
static struct system_qos_status system_qos = {
|
|
.nb.notifier_call = system_qos_update,
|
|
};
|
|
#endif
|
|
|
|
static void __exit mmdvfs_pmqos_exit(void)
|
|
{
|
|
platform_driver_unregister(&mmdvfs_pmqos_driver);
|
|
#ifdef QOS_BOUND_DETECT
|
|
unregister_qos_notifier(&system_qos.nb);
|
|
#endif
|
|
}
|
|
|
|
static int __init mmdvfs_pmqos_late_init(void)
|
|
{
|
|
#ifdef QOS_BOUND_DETECT
|
|
register_qos_notifier(&system_qos.nb);
|
|
#endif
|
|
#ifdef MMDVFS_FORCE_STEP0
|
|
mmdvfs_qos_force_step(0);
|
|
mmdvfs_enable = false;
|
|
pr_notice("force set step0 when late_init\n");
|
|
#else
|
|
mmdvfs_qos_force_step(0);
|
|
mmdvfs_qos_force_step(-1);
|
|
pr_notice("force flip step0 when late_init\n");
|
|
#endif
|
|
total_hrt_bw = get_total_hrt_bw(false);
|
|
total_ui_only_hrt_bw = get_total_hrt_bw(true);
|
|
init_me_swpm();
|
|
return 0;
|
|
}
|
|
|
|
u64 mmdvfs_qos_get_freq(u32 pm_qos_class)
|
|
{
|
|
u32 i = pm_qos_class - PM_QOS_DISP_FREQ;
|
|
u32 l, s;
|
|
|
|
if (!step_size)
|
|
return 0;
|
|
if (i >= ARRAY_SIZE(all_freqs))
|
|
i = 0;
|
|
if (current_max_step < 0 || current_max_step >= step_size)
|
|
s = step_size - 1;
|
|
else
|
|
s = current_max_step;
|
|
l = all_freqs[i]->limit_config.limit_level;
|
|
if (l)
|
|
return all_freqs[i]->limit_config.limit_steps[l-1][s].freq_step;
|
|
return all_freqs[i]->step_config[s].freq_step;
|
|
}
|
|
EXPORT_SYMBOL_GPL(mmdvfs_qos_get_freq);
|
|
|
|
void mmdvfs_qos_limit_config(u32 pm_qos_class, u32 limit_value,
|
|
enum mmdvfs_limit_source source)
|
|
{
|
|
u32 i = pm_qos_class - PM_QOS_DISP_FREQ;
|
|
s32 old_level = 0;
|
|
|
|
if (unlikely(i >= ARRAY_SIZE(all_freqs))) {
|
|
pr_notice("[%d]Invalid class=%u %d\n",
|
|
source, pm_qos_class, old_level);
|
|
return;
|
|
}
|
|
|
|
if (!all_freqs[i]->limit_config.limit_size) {
|
|
pr_notice("[%d]Not support limit: %u\n", source, pm_qos_class);
|
|
return;
|
|
}
|
|
|
|
if (log_level & log_limit)
|
|
pr_notice("[%d][%d]limit score update=(%d, %u, %u)\n",
|
|
source, pm_qos_class, limit_value,
|
|
all_freqs[i]->limit_config.limit_value,
|
|
all_freqs[i]->limit_config.limit_level);
|
|
|
|
#ifdef MMDVFS_LIMIT
|
|
mutex_lock(&step_mutex);
|
|
old_level = all_freqs[i]->limit_config.limit_level;
|
|
mmdvfs_update_limit_config(source, limit_value,
|
|
&all_freqs[i]->limit_config.limit_value,
|
|
&all_freqs[i]->limit_config.limit_level);
|
|
|
|
if (old_level != all_freqs[i]->limit_config.limit_level) {
|
|
pr_notice("MMDVFS limit level changed for %s %d->%d\n",
|
|
all_freqs[i]->prop_name, old_level,
|
|
all_freqs[i]->limit_config.limit_level);
|
|
mm_apply_clk(-1, all_freqs[i], current_max_step,
|
|
current_max_step);
|
|
}
|
|
mutex_unlock(&step_mutex);
|
|
#endif
|
|
#ifdef MMDVFS_MMP
|
|
mmprofile_log_ex(
|
|
mmdvfs_mmp_events.limit_change, MMPROFILE_FLAG_PULSE,
|
|
all_freqs[i]->limit_config.limit_value, pm_qos_class);
|
|
#endif
|
|
}
|
|
EXPORT_SYMBOL_GPL(mmdvfs_qos_limit_config);
|
|
|
|
static int print_freq(char *buf, int length,
|
|
struct mm_freq_step_config step_configs[], s32 current_step)
|
|
{
|
|
u32 i;
|
|
|
|
for (i = 0; i < step_size; i++) {
|
|
length += snprintf(buf + length, PAGE_SIZE - length,
|
|
(i == current_step) ? " v" : " ");
|
|
length += snprintf(buf + length, PAGE_SIZE - length,
|
|
"[%u]vopp=%d freq=%llu clk=%u/%u/%u/%u/0x%08x\n",
|
|
i, vopp_steps[i],
|
|
step_configs[i].freq_step,
|
|
step_configs[i].clk_type,
|
|
step_configs[i].clk_mux_id,
|
|
step_configs[i].clk_source_id,
|
|
step_configs[i].pll_id,
|
|
step_configs[i].pll_value);
|
|
if (length >= PAGE_SIZE)
|
|
break;
|
|
}
|
|
return length;
|
|
}
|
|
|
|
#define MAX_DUMP (PAGE_SIZE - 1)
|
|
int dump_setting(char *buf, const struct kernel_param *kp)
|
|
{
|
|
u32 i, l;
|
|
int length = 0;
|
|
struct mm_freq_config *mm_freq;
|
|
|
|
length += snprintf(buf + length, MAX_DUMP - length,
|
|
"force_step: %d\n", force_step);
|
|
for (i = 0; i < ARRAY_SIZE(all_freqs); i++) {
|
|
mm_freq = all_freqs[i];
|
|
length += snprintf(buf + length, MAX_DUMP - length,
|
|
"[%s] step_size: %u current_step:%d (%lluMhz)\n",
|
|
mm_freq->prop_name, step_size, mm_freq->current_step,
|
|
mmdvfs_qos_get_freq(PM_QOS_DISP_FREQ + i));
|
|
length = print_freq(buf, length,
|
|
mm_freq->step_config, mm_freq->current_step);
|
|
l = mm_freq->limit_config.limit_level;
|
|
if (l) {
|
|
length += snprintf(buf + length, MAX_DUMP - length,
|
|
"-[limit] level=%u value=0x%x\n",
|
|
mm_freq->limit_config.limit_level,
|
|
mm_freq->limit_config.limit_value);
|
|
length = print_freq(buf, length,
|
|
mm_freq->limit_config.limit_steps[l-1],
|
|
mm_freq->current_step);
|
|
}
|
|
if (length >= MAX_DUMP)
|
|
break;
|
|
}
|
|
if (length >= MAX_DUMP)
|
|
length = MAX_DUMP - 1;
|
|
|
|
return length;
|
|
}
|
|
|
|
static struct kernel_param_ops dump_param_ops = {.get = dump_setting};
|
|
module_param_cb(dump_setting, &dump_param_ops, NULL, 0444);
|
|
MODULE_PARM_DESC(dump_setting, "dump mmdvfs current setting");
|
|
|
|
int mmdvfs_qos_force_step(int step)
|
|
{
|
|
if (step >= (s32)step_size || step < STEP_UNREQUEST) {
|
|
pr_notice("force set step invalid: %d\n", step);
|
|
return -EINVAL;
|
|
}
|
|
force_step = step;
|
|
#if defined(CONFIG_MACH_MT6785) || defined(CONFIG_MACH_MT6768)
|
|
#if (defined(CONFIG_MTK_MT6382_BDG) && defined(CONFIG_MTK_MT6382_VDO_MODE))
|
|
force_step = 0;
|
|
#endif
|
|
#endif
|
|
update_step(PM_QOS_NUM_CLASSES, -1);
|
|
return 0;
|
|
}
|
|
EXPORT_SYMBOL_GPL(mmdvfs_qos_force_step);
|
|
|
|
int set_force_step(const char *val, const struct kernel_param *kp)
|
|
{
|
|
int result;
|
|
int new_force_step;
|
|
|
|
result = kstrtoint(val, 0, &new_force_step);
|
|
if (result) {
|
|
pr_notice("force set step failed: %d\n", result);
|
|
return result;
|
|
}
|
|
return mmdvfs_qos_force_step(new_force_step);
|
|
}
|
|
|
|
static struct kernel_param_ops force_step_ops = {
|
|
.set = set_force_step,
|
|
.get = param_get_int,
|
|
};
|
|
module_param_cb(force_step, &force_step_ops, &force_step, 0644);
|
|
MODULE_PARM_DESC(force_step, "force mmdvfs to specified step, -1 for unset");
|
|
|
|
void mmdvfs_autok_qos_enable(bool enable)
|
|
{
|
|
pr_notice("%s: step_size=%d current_max_step=%d\n",
|
|
__func__, step_size, current_max_step);
|
|
if (!enable && step_size > 0 && current_max_step == STEP_UNREQUEST)
|
|
mmdvfs_qos_force_step(step_size - 1);
|
|
|
|
mmdvfs_autok_enable = enable;
|
|
if (enable && step_size > 0)
|
|
mmdvfs_qos_force_step(-1);
|
|
pr_notice("mmdvfs_autok enabled? %d\n", enable);
|
|
}
|
|
EXPORT_SYMBOL_GPL(mmdvfs_autok_qos_enable);
|
|
|
|
void mmdvfs_qos_enable(bool enable)
|
|
{
|
|
mmdvfs_enable = enable;
|
|
pr_notice("mmdvfs enabled? %d\n", enable);
|
|
}
|
|
EXPORT_SYMBOL_GPL(mmdvfs_qos_enable);
|
|
|
|
int set_enable(const char *val, const struct kernel_param *kp)
|
|
{
|
|
int result;
|
|
bool enable;
|
|
|
|
result = kstrtobool(val, &enable);
|
|
if (result) {
|
|
pr_notice("force set enable: %d\n", result);
|
|
return result;
|
|
}
|
|
mmdvfs_qos_enable(enable);
|
|
return 0;
|
|
}
|
|
|
|
static struct kernel_param_ops mmdvfs_enable_ops = {
|
|
.set = set_enable,
|
|
.get = param_get_bool,
|
|
};
|
|
module_param_cb(
|
|
mmdvfs_enable, &mmdvfs_enable_ops, &mmdvfs_enable, 0644);
|
|
MODULE_PARM_DESC(mmdvfs_enable, "enable or disable mmdvfs");
|
|
|
|
void mmdvfs_prepare_action(enum mmdvfs_prepare_event event)
|
|
{
|
|
if (event == MMDVFS_PREPARE_CALIBRATION_START) {
|
|
mmdvfs_autok_qos_enable(false);
|
|
pr_notice("mmdvfs service is disabled for calibration\n");
|
|
} else if (event == MMDVFS_PREPARE_CALIBRATION_END) {
|
|
mmdvfs_autok_qos_enable(true);
|
|
pr_notice("mmdvfs service is enabled after calibration\n");
|
|
} else {
|
|
pr_notice("%s: unknown event code:%d\n", __func__, event);
|
|
}
|
|
}
|
|
|
|
s32 get_virtual_port(enum virtual_source_id id)
|
|
{
|
|
switch (id) {
|
|
case VIRTUAL_DISP:
|
|
return PORT_VIRTUAL_DISP;
|
|
case VIRTUAL_MD:
|
|
return PORT_VIRTUAL_MD;
|
|
case VIRTUAL_CCU_COMMON:
|
|
return PORT_VIRTUAL_CCU_COMMON;
|
|
case VIRTUAL_CCU_COMMON2:
|
|
return PORT_VIRTUAL_CCU_COMMON2;
|
|
default:
|
|
pr_notice("invalid source id:%u\n", id);
|
|
return -1;
|
|
}
|
|
}
|
|
|
|
module_param(log_level, uint, 0644);
|
|
MODULE_PARM_DESC(log_level, "mmdvfs log level");
|
|
|
|
module_param(skip_smi_config, bool, 0644);
|
|
MODULE_PARM_DESC(skip_smi_config, "mmdvfs smi config");
|
|
|
|
static u32 dump_larbs = 0xFFFFFFFF;
|
|
#define MAX_DUMP (PAGE_SIZE - 1)
|
|
int get_larbs_info(char *buf)
|
|
{
|
|
u32 i, j;
|
|
int length = 0;
|
|
struct mm_qos_request *req = NULL;
|
|
|
|
for (i = 0; i < ARRAY_SIZE(larb_req); i++) {
|
|
if (!larb_req[i].port_count || !(dump_larbs & 1 << i))
|
|
continue;
|
|
length += snprintf(buf + length, MAX_DUMP - length,
|
|
"[%u] port count: %u\n", i, larb_req[i].port_count);
|
|
for (j = 0; j < ARRAY_SIZE(larb_req[i].ratio); j++) {
|
|
if (!larb_req[i].ratio[j])
|
|
break;
|
|
length += snprintf(buf + length, MAX_DUMP - length,
|
|
" %u", larb_req[i].ratio[j]);
|
|
if (length >= MAX_DUMP)
|
|
break;
|
|
}
|
|
length += snprintf(buf + length, MAX_DUMP - length, "\n");
|
|
|
|
mutex_lock(&bw_mutex);
|
|
list_for_each_entry(req, &larb_req[i].larb_list, larb_node) {
|
|
if (!req->bw_value && !req->hrt_value)
|
|
continue;
|
|
length += snprintf(buf + length, MAX_DUMP - length,
|
|
" [port-%u]: bw=%u ostd=%u hrt=%u comp=%d\n",
|
|
req->master_id & 0x1F, req->bw_value, req->ostd,
|
|
req->hrt_value, req->comp_type);
|
|
if (length >= MAX_DUMP)
|
|
break;
|
|
}
|
|
mutex_unlock(&bw_mutex);
|
|
|
|
if (length >= MAX_DUMP)
|
|
break;
|
|
}
|
|
if (length >= MAX_DUMP)
|
|
length = MAX_DUMP - 1;
|
|
|
|
return length;
|
|
}
|
|
|
|
void mmdvfs_print_larbs_info(void)
|
|
{
|
|
int len;
|
|
char *ptr, *tmp_str;
|
|
char *log_str = kmalloc(PAGE_SIZE, GFP_KERNEL);
|
|
|
|
if (log_str) {
|
|
len = get_larbs_info(log_str);
|
|
tmp_str = log_str;
|
|
if (len > 0) {
|
|
while ((ptr = strsep(&tmp_str, "\n")) != NULL)
|
|
pr_notice("%s\n", ptr);
|
|
} else
|
|
pr_notice("no larbs info to print\n");
|
|
kfree(log_str);
|
|
} else
|
|
pr_notice("kmalloc fails!\n");
|
|
}
|
|
|
|
int get_dump_larbs(char *buf, const struct kernel_param *kp)
|
|
{
|
|
int len;
|
|
|
|
smi_debug_bus_hang_detect(false, "MMDVFS");
|
|
len = get_larbs_info(buf);
|
|
return len;
|
|
}
|
|
|
|
static struct kernel_param_ops dump_larb_param_ops = {
|
|
.get = get_dump_larbs,
|
|
.set = param_set_uint,
|
|
};
|
|
module_param_cb(dump_larbs, &dump_larb_param_ops, &dump_larbs, 0644);
|
|
MODULE_PARM_DESC(dump_larbs, "dump mmdvfs current larb setting");
|
|
|
|
int get_larb_mode(char *buf, const struct kernel_param *kp)
|
|
{
|
|
int length = 0;
|
|
|
|
length += snprintf(buf + length, PAGE_SIZE - length,
|
|
"current mode: %d\n", larb_soft);
|
|
length += snprintf(buf + length, PAGE_SIZE - length,
|
|
"force mode: %d\n", force_larb_mode);
|
|
buf[length] = '\0';
|
|
|
|
return length;
|
|
}
|
|
|
|
static struct kernel_param_ops larb_mode_ops = {
|
|
.get = get_larb_mode,
|
|
.set = param_set_int,
|
|
};
|
|
module_param_cb(larb_mode, &larb_mode_ops, &force_larb_mode, 0644);
|
|
MODULE_PARM_DESC(larb_mode, "set or get current larb mode");
|
|
static s32 vote_freq;
|
|
static bool vote_req_init;
|
|
struct mtk_pm_qos_request vote_req;
|
|
int set_vote_freq(const char *val, const struct kernel_param *kp)
|
|
{
|
|
int result;
|
|
int new_vote_freq;
|
|
|
|
result = kstrtoint(val, 0, &new_vote_freq);
|
|
if (result) {
|
|
pr_notice("force set step failed: %d\n", result);
|
|
return result;
|
|
}
|
|
|
|
if (!vote_req_init) {
|
|
mtk_pm_qos_add_request(
|
|
&vote_req, PM_QOS_DISP_FREQ,
|
|
PM_QOS_MM_FREQ_DEFAULT_VALUE);
|
|
vote_req_init = true;
|
|
}
|
|
vote_freq = new_vote_freq;
|
|
mtk_pm_qos_update_request(&vote_req, vote_freq);
|
|
return 0;
|
|
}
|
|
static struct kernel_param_ops vote_freq_ops = {
|
|
.set = set_vote_freq,
|
|
.get = param_get_int,
|
|
};
|
|
|
|
module_param_cb(vote_freq, &vote_freq_ops, &vote_freq, 0644);
|
|
MODULE_PARM_DESC(vote_freq, "vote mmdvfs to specified freq, 0 for unset");
|
|
|
|
#define UT_MAX_REQUEST 10
|
|
static s32 qos_ut_case;
|
|
static struct plist_head ut_req_list;
|
|
static bool ut_req_init;
|
|
struct mm_qos_request ut_req[UT_MAX_REQUEST] = {};
|
|
static DECLARE_COMPLETION(comp);
|
|
|
|
static int test_event(struct notifier_block *nb,
|
|
unsigned long value, void *v)
|
|
{
|
|
pr_notice("ut test notifier: value=%lu\n", value);
|
|
/*msleep(50);*/ /* Use it when disp's notifier callback not ready*/
|
|
complete(&comp);
|
|
return 0;
|
|
}
|
|
static struct notifier_block test_notifier = {
|
|
.notifier_call = test_event,
|
|
};
|
|
|
|
static int make_cam_hrt_bw(void *data)
|
|
{
|
|
struct plist_head cam_req_list;
|
|
struct mm_qos_request cam_req = {};
|
|
|
|
plist_head_init(&cam_req_list);
|
|
mm_qos_add_request(&cam_req_list,
|
|
&cam_req, SMI_PMQOS_ENC(cam_larb_ids[0], 0));
|
|
mm_qos_set_request(&cam_req, 100, 100, 0);
|
|
mm_qos_update_all_request(&cam_req_list);
|
|
mm_qos_update_all_request_zero(&cam_req_list);
|
|
mm_qos_remove_all_request(&cam_req_list);
|
|
return 0;
|
|
}
|
|
|
|
int mmdvfs_qos_ut_set(const char *val, const struct kernel_param *kp)
|
|
{
|
|
int result, value;
|
|
u32 old_log_level = log_level;
|
|
u32 req_id, master;
|
|
struct task_struct *pKThread;
|
|
u64 start_jiffies;
|
|
|
|
result = sscanf(val, "%d %d %i %d", &qos_ut_case,
|
|
&req_id, &master, &value);
|
|
if (result != 4) {
|
|
pr_notice("invalid input: %s, result(%d)\n", val, result);
|
|
return -EINVAL;
|
|
}
|
|
if (req_id >= UT_MAX_REQUEST) {
|
|
pr_notice("invalid req_id: %u\n", req_id);
|
|
return -EINVAL;
|
|
}
|
|
|
|
pr_notice("ut with (case_id,req_id,master,value)=(%d,%u,%#x,%d)\n",
|
|
qos_ut_case, req_id, master, value);
|
|
log_level = 1 << log_bw | 1 << log_freq | 1 << log_smi_freq;
|
|
if (!ut_req_init) {
|
|
plist_head_init(&ut_req_list);
|
|
ut_req_init = true;
|
|
}
|
|
switch (qos_ut_case) {
|
|
case 0:
|
|
mm_qos_add_request(&ut_req_list, &ut_req[req_id], master);
|
|
mm_qos_set_request(&ut_req[req_id], value, 0, BW_COMP_NONE);
|
|
mm_qos_update_all_request(&ut_req_list);
|
|
break;
|
|
case 1:
|
|
mm_qos_add_request(&ut_req_list, &ut_req[req_id], master);
|
|
mm_qos_set_request(&ut_req[req_id], value, value, BW_COMP_NONE);
|
|
mm_qos_update_all_request(&ut_req_list);
|
|
break;
|
|
case 2:
|
|
mm_qos_add_request(&ut_req_list, &ut_req[req_id], master);
|
|
mm_qos_set_bw_request(&ut_req[req_id], value, BW_COMP_NONE);
|
|
mm_qos_update_all_request(&ut_req_list);
|
|
break;
|
|
case 3:
|
|
mm_qos_add_request(&ut_req_list, &ut_req[req_id], master);
|
|
mm_qos_set_hrt_request(&ut_req[req_id], value);
|
|
mm_qos_update_all_request(&ut_req_list);
|
|
break;
|
|
case 4:
|
|
mm_qos_add_request(&ut_req_list, &ut_req[req_id], master);
|
|
mm_qos_set_request(&ut_req[req_id], value, 0, BW_COMP_DEFAULT);
|
|
mm_qos_update_all_request(&ut_req_list);
|
|
break;
|
|
case 5:
|
|
mm_qos_add_request(&ut_req_list, &ut_req[req_id], master);
|
|
mm_qos_set_request(&ut_req[req_id], value,
|
|
value, BW_COMP_DEFAULT);
|
|
mm_qos_update_all_request(&ut_req_list);
|
|
break;
|
|
case 6:
|
|
/* Test blocking mechanism */
|
|
reinit_completion(&comp);
|
|
mm_hrt_add_bw_throttle_notifier(&test_notifier);
|
|
/* Make camera block and trigger an event sent to notifier */
|
|
mmdvfs_set_max_camera_hrt_bw(2000);
|
|
pKThread = kthread_run(make_cam_hrt_bw,
|
|
NULL, "make_cam_hrt_bw");
|
|
if (IS_ERR(pKThread))
|
|
pr_notice("create cam hrt bw thread failed\n");
|
|
/* Notifier will call complete */
|
|
wait_for_completion(&comp);
|
|
reinit_completion(&comp);
|
|
start_jiffies = jiffies;
|
|
mmdvfs_set_max_camera_hrt_bw(0);
|
|
wait_for_completion(&comp);
|
|
pr_notice("wait time should > 2000 msecs:%u\n",
|
|
jiffies_to_msecs(jiffies-start_jiffies));
|
|
mm_hrt_remove_bw_throttle_notifier(&test_notifier);
|
|
break;
|
|
case 7:
|
|
mmdvfs_set_max_camera_hrt_bw(5400);
|
|
make_cam_hrt_bw(NULL);
|
|
mmdvfs_set_max_camera_hrt_bw(0);
|
|
break;
|
|
case -1:
|
|
mm_qos_remove_all_request(&ut_req_list);
|
|
break;
|
|
case -2:
|
|
mm_qos_update_all_request_zero(&ut_req_list);
|
|
break;
|
|
default:
|
|
pr_notice("invalid case_id: %d\n", qos_ut_case);
|
|
break;
|
|
}
|
|
|
|
pr_notice("Call SMI Dump API Begin\n");
|
|
/* smi_debug_bus_hang_detect(false, "MMDVFS"); */
|
|
pr_notice("Call SMI Dump API END\n");
|
|
log_level = old_log_level;
|
|
return 0;
|
|
}
|
|
|
|
static struct kernel_param_ops qos_ut_case_ops = {
|
|
.set = mmdvfs_qos_ut_set,
|
|
.get = param_get_int,
|
|
};
|
|
module_param_cb(qos_ut_case, &qos_ut_case_ops, &qos_ut_case, 0644);
|
|
MODULE_PARM_DESC(qos_ut_case, "force mmdvfs UT test case");
|
|
|
|
static s32 mmdvfs_ut_case;
|
|
int mmdvfs_ut_set(const char *val, const struct kernel_param *kp)
|
|
{
|
|
int result;
|
|
int value1, value2;
|
|
u32 old_log_level = log_level;
|
|
struct mtk_pm_qos_request disp_req = {};
|
|
|
|
result = sscanf(val, "%d %d", &mmdvfs_ut_case, &value1);
|
|
if (result != 2) {
|
|
pr_notice("invalid input: %s, result(%d)\n", val, result);
|
|
return -EINVAL;
|
|
}
|
|
pr_notice("%s (case_id, value): (%d,%d)\n",
|
|
__func__, mmdvfs_ut_case, value1);
|
|
|
|
log_level = 1 << log_freq |
|
|
1 << log_limit;
|
|
mtk_pm_qos_add_request(&disp_req, PM_QOS_DISP_FREQ,
|
|
PM_QOS_MM_FREQ_DEFAULT_VALUE);
|
|
|
|
switch (mmdvfs_ut_case) {
|
|
case 0:
|
|
result = sscanf(val, "%d %d %d", &mmdvfs_ut_case,
|
|
&value1, &value2);
|
|
if (result != 3) {
|
|
pr_notice("invalid arguments: %s\n", val);
|
|
break;
|
|
}
|
|
pr_notice("limit test score: %d\n", value2);
|
|
pr_notice("limit initial: %d\n",
|
|
mmdvfs_get_limit_status(value1));
|
|
/* limit enable then opp1 -> opp0 */
|
|
mmdvfs_qos_limit_config(value1, 1, MMDVFS_LIMIT_THERMAL);
|
|
mmdvfs_qos_limit_config(value1, value2, MMDVFS_LIMIT_CAM);
|
|
mtk_pm_qos_update_request(&disp_req, 1000);
|
|
pr_notice("limit enable then opp up: %d freq=%llu MHz\n",
|
|
mmdvfs_get_limit_status(value1),
|
|
mmdvfs_qos_get_freq(value1));
|
|
/* limit disable when opp0 */
|
|
mmdvfs_qos_limit_config(value1, 0, MMDVFS_LIMIT_THERMAL);
|
|
pr_notice("limit disable when opp up: %d freq=%llu MHz\n",
|
|
mmdvfs_get_limit_status(value1),
|
|
mmdvfs_qos_get_freq(value1));
|
|
/* limit enable when opp0 */
|
|
mmdvfs_qos_limit_config(value1, 1, MMDVFS_LIMIT_THERMAL);
|
|
pr_notice("limit enable when opp up: %d freq=%llu MHz\n",
|
|
mmdvfs_get_limit_status(value1),
|
|
mmdvfs_qos_get_freq(value1));
|
|
/* limit disable then opp0 -> opp1 */
|
|
mmdvfs_qos_limit_config(value1, 0, MMDVFS_LIMIT_THERMAL);
|
|
mtk_pm_qos_update_request(&disp_req, 0);
|
|
pr_notice("limit disable then opp down: %d freq=%llu MHz\n",
|
|
mmdvfs_get_limit_status(value1),
|
|
mmdvfs_qos_get_freq(value1));
|
|
/* limit enable when opp1 */
|
|
mmdvfs_qos_limit_config(value1, 1, MMDVFS_LIMIT_THERMAL);
|
|
mtk_pm_qos_update_request(&disp_req, 0);
|
|
pr_notice("limit enable when opp down: %d freq=%llu MHz\n",
|
|
mmdvfs_get_limit_status(value1),
|
|
mmdvfs_qos_get_freq(value1));
|
|
/* limit disable when opp1 */
|
|
mmdvfs_qos_limit_config(value1, 0, MMDVFS_LIMIT_THERMAL);
|
|
mtk_pm_qos_update_request(&disp_req, 0);
|
|
pr_notice("limit disable when opp down: %d freq=%llu MHz\n",
|
|
mmdvfs_get_limit_status(value1),
|
|
mmdvfs_qos_get_freq(value1));
|
|
|
|
break;
|
|
case 1:
|
|
result = sscanf(val, "%d %d %d", &mmdvfs_ut_case,
|
|
&value1, &value2);
|
|
if (result != 3) {
|
|
pr_notice("invalid arguments: %s\n", val);
|
|
mmdvfs_qos_limit_config(value1, 0,
|
|
MMDVFS_LIMIT_THERMAL);
|
|
break;
|
|
}
|
|
pr_notice("limit test score: %d\n", value2);
|
|
pr_notice("limit initial: %d freq=%llu MHz\n",
|
|
mmdvfs_get_limit_status(value1),
|
|
mmdvfs_qos_get_freq(value1));
|
|
mmdvfs_qos_limit_config(value1, 1, MMDVFS_LIMIT_THERMAL);
|
|
mmdvfs_qos_limit_config(value1, value2, MMDVFS_LIMIT_CAM);
|
|
pr_notice("limit now: %d freq=%llu MHz\n",
|
|
mmdvfs_get_limit_status(value1),
|
|
mmdvfs_qos_get_freq(value1));
|
|
|
|
break;
|
|
default:
|
|
pr_notice("invalid case_id: %d\n", mmdvfs_ut_case);
|
|
break;
|
|
}
|
|
|
|
mtk_pm_qos_remove_request(&disp_req);
|
|
|
|
pr_notice("%s END\n", __func__);
|
|
log_level = old_log_level;
|
|
return 0;
|
|
}
|
|
|
|
static struct kernel_param_ops mmdvfs_ut_ops = {
|
|
.set = mmdvfs_ut_set,
|
|
.get = param_get_int,
|
|
};
|
|
module_param_cb(mmdvfs_ut_case, &mmdvfs_ut_ops, &mmdvfs_ut_case, 0644);
|
|
MODULE_PARM_DESC(mmdvfs_ut_case, "force mmdvfs UT test case");
|
|
|
|
int set_disp_bw_ceiling(const char *val, const struct kernel_param *kp)
|
|
{
|
|
int result;
|
|
s32 disp_bw, wait;
|
|
s32 disp_avail_hrt_bw;
|
|
|
|
|
|
result = sscanf(val, "%d %d", &disp_bw, &wait);
|
|
if (result != 2) {
|
|
pr_notice("invalid input: %s, result(%d)\n", val, result);
|
|
return -EINVAL;
|
|
}
|
|
pr_notice("%s (disp_bw, wait): (%d,%d)\n",
|
|
__func__, disp_bw, wait);
|
|
|
|
disp_bw_ceiling = (disp_bw < 0)?0:disp_bw;
|
|
wait_next_max_cam_bw_set = wait;
|
|
|
|
disp_avail_hrt_bw = mm_hrt_get_available_hrt_bw(PORT_VIRTUAL_DISP);
|
|
pr_notice("disp_bw_ceiling=%d total_hrt_bw=%d disp_avail_hrt_bw=%d\n",
|
|
disp_bw_ceiling, total_hrt_bw, disp_avail_hrt_bw);
|
|
|
|
if (!wait_next_max_cam_bw_set)
|
|
blocking_notifier_call_chain(
|
|
&hrt_bw_throttle_notifier,
|
|
BW_THROTTLE_START, NULL);
|
|
|
|
return 0;
|
|
}
|
|
|
|
static struct kernel_param_ops disp_bw_ceiling_ops = {
|
|
.set = set_disp_bw_ceiling,
|
|
.get = param_get_int,
|
|
};
|
|
module_param_cb(disp_bw_ceiling, &disp_bw_ceiling_ops,
|
|
&disp_bw_ceiling, 0644);
|
|
MODULE_PARM_DESC(disp_bw_ceiling,
|
|
"set display bw to test repaint and decouple");
|
|
|
|
int set_force_bwl(const char *val, const struct kernel_param *kp)
|
|
{
|
|
int result;
|
|
int comm, port, bwl;
|
|
|
|
result = sscanf(val, "%d %d %d", &comm, &port, &bwl);
|
|
if (result != 3) {
|
|
pr_notice("invalid input: %s, result(%d)\n", val, result);
|
|
return -EINVAL;
|
|
}
|
|
|
|
if (comm >= 0 && comm < MAX_COMM_NUM
|
|
&& port >= 0 && port < SMI_COMM_MASTER_NUM)
|
|
force_comm_bwl[comm][port] = bwl;
|
|
|
|
return 0;
|
|
}
|
|
|
|
int get_force_bwl(char *buf, const struct kernel_param *kp)
|
|
{
|
|
int i, j, length = 0;
|
|
|
|
for (i = 0; i < MAX_COMM_NUM; i++)
|
|
for (j = 0; j < SMI_COMM_MASTER_NUM; j++)
|
|
length += snprintf(buf + length, PAGE_SIZE - length,
|
|
"%d ", force_comm_bwl[i][j]);
|
|
|
|
length += snprintf(buf + length, PAGE_SIZE - length, "\n");
|
|
|
|
return length;
|
|
}
|
|
|
|
static struct kernel_param_ops force_bwl_ops = {
|
|
.set = set_force_bwl,
|
|
.get = get_force_bwl,
|
|
};
|
|
module_param_cb(force_bwl, &force_bwl_ops,
|
|
NULL, 0644);
|
|
MODULE_PARM_DESC(force_bwl,
|
|
"force bwl for each larb");
|
|
|
|
late_initcall(mmdvfs_pmqos_late_init);
|
|
module_init(mmdvfs_pmqos_init);
|
|
module_exit(mmdvfs_pmqos_exit);
|
|
|
|
MODULE_DESCRIPTION("MTK MMDVFS driver");
|
|
MODULE_AUTHOR("Damon Chu<damon.chu@mediatek.com>");
|
|
MODULE_LICENSE("GPL");
|