kernel_samsung_a34x-permissive/drivers/misc/mediatek/trusted_mem/proc.c
2024-04-28 15:51:13 +02:00

372 lines
9.5 KiB
C

/* SPDX-License-Identifier: GPL-2.0 */
/*
* Copyright (c) 2019 MediaTek Inc.
*/
#define PR_FMT_HEADER_MUST_BE_INCLUDED_BEFORE_ALL_HDRS
#include "private/tmem_pr_fmt.h" PR_FMT_HEADER_MUST_BE_INCLUDED_BEFORE_ALL_HDRS
#include <linux/errno.h>
#include <linux/fs.h>
#include <linux/kernel.h>
#include <linux/proc_fs.h>
#include <linux/string.h>
#include <linux/uaccess.h>
#include <linux/unistd.h>
#include <linux/version.h>
#include <linux/module.h>
#include <linux/moduleparam.h>
#include <linux/sizes.h>
#include <linux/mod_devicetable.h>
#include <linux/platform_device.h>
#include <linux/dma-mapping.h>
#include <linux/dma-direct.h>
#include <linux/kallsyms.h>
#include "private/tmem_error.h"
#include "private/tmem_utils.h"
#include "private/tmem_priv.h"
#include "private/tmem_entry.h"
#include "private/ut_cmd.h"
#include "tee_impl/tee_invoke.h"
#include "memory_ssmr.h"
static int tmem_open(struct inode *inode, struct file *file)
{
UNUSED(inode);
UNUSED(file);
pr_info("%s:%d\n", __func__, __LINE__);
return TMEM_OK;
}
static int tmem_release(struct inode *inode, struct file *file)
{
UNUSED(inode);
UNUSED(file);
pr_info("%s:%d\n", __func__, __LINE__);
return TMEM_OK;
}
static u32 g_common_mem_handle[TRUSTED_MEM_MAX];
static void trusted_mem_device_chunk_alloc(enum TRUSTED_MEM_TYPE mem_type)
{
int ret = TMEM_OK;
u32 alignment = 0, ref_count;
u32 min_chunk_sz = tmem_core_get_min_chunk_size(mem_type);
if (IS_ZERO(g_common_mem_handle[mem_type]))
ret = tmem_core_alloc_chunk(
mem_type, alignment, min_chunk_sz, &ref_count,
&g_common_mem_handle[mem_type], NULL, 0, 0);
else
pr_info("%d chunk is already allocated, handle:0x%x\n",
mem_type, g_common_mem_handle[mem_type]);
if (ret)
pr_err("%d alloc chunk failed:%d\n", mem_type, ret);
}
static void trusted_mem_device_chunk_free(enum TRUSTED_MEM_TYPE mem_type)
{
int ret = TMEM_OK;
if (!IS_ZERO(g_common_mem_handle[mem_type]))
ret = tmem_core_unref_chunk(
mem_type, g_common_mem_handle[mem_type], NULL, 0);
if (ret)
pr_err("%d free chunk failed:%d\n", mem_type, ret);
else
g_common_mem_handle[mem_type] = 0;
}
static void trusted_mem_device_ion_alloc_free(enum TRUSTED_MEM_TYPE mem_type)
{
pr_info("%d ion interface is not supported!\n", mem_type);
}
static void trusted_mem_device_common_operations(u64 cmd, u64 param1,
u64 param2, u64 param3)
{
int device_mem_type = ((u32)cmd) / 10;
int device_cmd = ((u32)cmd) % 10;
if (device_mem_type >= TRUSTED_MEM_MAX) {
pr_err("unsupported mem type: %d (user cmd:%lld)\n",
device_mem_type, cmd);
return;
}
switch (device_cmd) {
case TMEM_DEVICE_COMMON_OPERATION_SSMR_ALLOC:
pr_info("TMEM_DEVICE_COMMON_OPERATION_SSMR_ALLOC\n");
tmem_core_ssmr_allocate(device_mem_type);
break;
case TMEM_DEVICE_COMMON_OPERATION_SSMR_RELEASE:
pr_info("TMEM_DEVICE_COMMON_OPERATION_SSMR_RELEASE\n");
tmem_core_ssmr_release(device_mem_type);
break;
case TMEM_DEVICE_COMMON_OPERATION_SESSION_OPEN:
pr_info("TMEM_DEVICE_COMMON_OPERATION_SESSION_OPEN\n");
tmem_core_session_open(device_mem_type);
break;
case TMEM_DEVICE_COMMON_OPERATION_SESSION_CLOSE:
pr_info("TMEM_DEVICE_COMMON_OPERATION_SESSION_CLOSE\n");
tmem_core_session_close(device_mem_type);
break;
case TMEM_DEVICE_COMMON_OPERATION_REGION_ON:
pr_info("TMEM_DEVICE_COMMON_OPERATION_REGION_ON\n");
tmem_core_regmgr_online(device_mem_type);
break;
case TMEM_DEVICE_COMMON_OPERATION_REGION_OFF:
pr_info("TMEM_DEVICE_COMMON_OPERATION_REGION_OFF\n");
tmem_core_regmgr_offline(device_mem_type);
break;
case TMEM_DEVICE_COMMON_OPERATION_CHUNK_ALLOC:
pr_info("TMEM_DEVICE_COMMON_OPERATION_CHUNK_ALLOC\n");
trusted_mem_device_chunk_alloc(device_mem_type);
break;
case TMEM_DEVICE_COMMON_OPERATION_CHUNK_FREE:
pr_info("TMEM_DEVICE_COMMON_OPERATION_CHUNK_FREE\n");
trusted_mem_device_chunk_free(device_mem_type);
break;
case TMEM_DEVICE_COMMON_OPERATION_ION_ALLOC_FREE:
pr_info("TMEM_DEVICE_COMMON_OPERATION_ION_ALLOC_FREE\n");
trusted_mem_device_ion_alloc_free(device_mem_type);
break;
default:
pr_err("unsupported device cmd: %d, mem type: %d (user cmd:%lld)\n",
device_cmd, device_mem_type, cmd);
break;
}
}
static void trusted_mem_region_status_dump(void)
{
int mem_idx;
bool is_region_on;
bool is_dev_registered;
for (mem_idx = 0; mem_idx < TRUSTED_MEM_MAX; mem_idx++) {
is_region_on = tmem_core_is_regmgr_region_on(mem_idx);
is_dev_registered = tmem_core_is_device_registered(mem_idx);
pr_info("mem%d reg_state:%s registered:%s\n", mem_idx,
is_region_on ? "BUSY" : "IDLE",
is_dev_registered ? "YES" : "NO");
}
}
static void trusted_mem_manual_cmd_invoke(u64 cmd, u64 param1, u64 param2,
u64 param3)
{
if (cmd >= TMEM_DEVICE_COMMON_OPERATION_START
&& cmd <= TMEM_DEVICE_COMMON_OPERATION_END) {
trusted_mem_device_common_operations(cmd, param1, param2,
param3);
return;
}
switch (cmd) {
case TMEM_REGION_STATUS_DUMP:
pr_info("TMEM_REGION_STATUS_DUMP\n");
trusted_mem_region_status_dump();
break;
case TMEM_SECMEM_SVP_DUMP_INFO:
pr_info("TMEM_SECMEM_SVP_DUMP_INFO\n");
#if defined(CONFIG_MTK_SECURE_MEM_SUPPORT)
secmem_svp_dump_info();
#endif
break;
case TMEM_SECMEM_FR_DUMP_INFO:
pr_info("TMEM_SECMEM_FR_DUMP_INFO\n");
#if defined(CONFIG_MTK_SECURE_MEM_SUPPORT) \
&& defined(CONFIG_MTK_CAM_SECURITY_SUPPORT)
secmem_fr_dump_info();
#endif
break;
case TMEM_SECMEM_WFD_DUMP_INFO:
pr_info("TMEM_SECMEM_WFD_DUMP_INFO\n");
#if defined(CONFIG_MTK_WFD_SMEM_SUPPORT)
wfd_smem_dump_info();
#endif
break;
case TMEM_SECMEM_DYNAMIC_DEBUG_ENABLE:
pr_info("TMEM_SECMEM_DYNAMIC_DEBUG_ENABLE\n");
#if defined(CONFIG_MTK_SECURE_MEM_SUPPORT)
secmem_dynamic_debug_control(true);
#endif
break;
case TMEM_SECMEM_DYNAMIC_DEBUG_DISABLE:
pr_info("TMEM_SECMEM_DYNAMIC_DEBUG_DISABLE\n");
#if defined(CONFIG_MTK_SECURE_MEM_SUPPORT)
secmem_dynamic_debug_control(false);
#endif
break;
case TMEM_SECMEM_FORCE_HW_PROTECTION:
pr_info("TMEM_SECMEM_FORCE_HW_PROTECTION\n");
#if defined(CONFIG_MTK_SECURE_MEM_SUPPORT)
secmem_force_hw_protection();
#endif
break;
default:
break;
}
}
static ssize_t tmem_write(struct file *file, const char __user *buffer,
size_t count, loff_t *data)
{
char desc[32];
unsigned int len = 0;
long cmd;
len = (count < (sizeof(desc) - 1)) ? count : (sizeof(desc) - 1);
if (copy_from_user(desc, buffer, len))
return 0;
desc[len] = '\0';
if (kstrtol(desc, 10, &cmd) != 0)
return count;
pr_debug("receives user space cmd '%ld'\n", cmd);
if ((cmd >= TMEM_MANUAL_CMD_RESERVE_START)
&& (cmd <= TMEM_MANUAL_CMD_RESERVE_END)) {
trusted_mem_manual_cmd_invoke((u64)cmd, (u64)0, (u64)0, (u64)0);
} else {
trusted_mem_ut_cmd_invoke((u64)cmd, (u64)0, (u64)0, (u64)0);
}
return count;
}
static const struct file_operations tmem_fops = {
.owner = THIS_MODULE,
.open = tmem_open,
.release = tmem_release,
.unlocked_ioctl = NULL,
#ifdef CONFIG_COMPAT
.compat_ioctl = NULL,
#endif
.write = tmem_write,
};
static void trusted_mem_create_proc_entry(void)
{
proc_create("tmem0", 0664, NULL, &tmem_fops);
}
#ifdef TCORE_UT_TESTS_SUPPORT
#ifdef CONFIG_MTK_ENG_BUILD
#define UT_MULTITHREAD_TEST_DEFAULT_WAIT_COMPLETION_TIMEOUT_MS (900000)
#define UT_SATURATION_STRESS_PMEM_MIN_CHUNK_SIZE (SIZE_8M)
#else
#define UT_MULTITHREAD_TEST_DEFAULT_WAIT_COMPLETION_TIMEOUT_MS (5000)
#define UT_SATURATION_STRESS_PMEM_MIN_CHUNK_SIZE (SIZE_2M)
#endif
static unsigned int ut_multithread_wait_completion_timeout_ms =
UT_MULTITHREAD_TEST_DEFAULT_WAIT_COMPLETION_TIMEOUT_MS;
int get_multithread_test_wait_completion_time(void)
{
return ut_multithread_wait_completion_timeout_ms;
}
module_param_named(wait_comp_ms, ut_multithread_wait_completion_timeout_ms,
uint, 0644);
MODULE_PARM_DESC(ut_multithread_wait_completion_timeout_ms,
"set wait completion timeout in ms for multithread UT tests");
static unsigned int ut_saturation_stress_pmem_min_chunk_size =
UT_SATURATION_STRESS_PMEM_MIN_CHUNK_SIZE;
int get_saturation_stress_pmem_min_chunk_size(void)
{
return ut_saturation_stress_pmem_min_chunk_size;
}
module_param_named(pmem_min_chunk_size,
ut_saturation_stress_pmem_min_chunk_size, uint, 0644);
MODULE_PARM_DESC(ut_saturation_stress_pmem_min_chunk_size,
"set pmem minimal chunk size for saturation stress tests");
#endif
static int trusted_mem_init(struct platform_device *pdev)
{
pr_info("%s:%d\n", __func__, __LINE__);
ssmr_probe(pdev);
trusted_mem_subsys_init();
#ifdef TCORE_UT_TESTS_SUPPORT
tmem_ut_server_init();
tmem_ut_cases_init();
#endif
#ifdef TEE_DEVICES_SUPPORT
tee_smem_devs_init();
#endif
#ifdef MTEE_DEVICES_SUPPORT
mtee_mchunks_init();
#endif
#if IS_ENABLED(CONFIG_MTK_GZ_KREE)
tmem_mpu_vio_init();
#endif
trusted_mem_create_proc_entry();
pr_info("%s:%d (end)\n", __func__, __LINE__);
return TMEM_OK;
}
static int trusted_mem_exit(struct platform_device *pdev)
{
#if IS_ENABLED(CONFIG_MTK_GZ_KREE)
tmem_mpu_vio_exit();
#endif
#ifdef MTEE_DEVICES_SUPPORT
mtee_mchunks_exit();
#endif
#ifdef TEE_DEVICES_SUPPORT
tee_smem_devs_exit();
#endif
#ifdef TCORE_UT_TESTS_SUPPORT
tmem_ut_cases_exit();
tmem_ut_server_exit();
#endif
trusted_mem_subsys_exit();
return 0;
}
static const struct of_device_id tm_of_match_table[] = {
{ .compatible = "mediatek,trusted_mem"},
{},
};
static struct platform_driver trusted_mem_driver = {
.probe = trusted_mem_init,
.remove = trusted_mem_exit,
.driver = {
.name = "trusted_mem",
.of_match_table = tm_of_match_table,
},
};
module_platform_driver(trusted_mem_driver);
MODULE_AUTHOR("MediaTek Inc.");
MODULE_LICENSE("GPL");
MODULE_DESCRIPTION("MediaTek Trusted Memory Driver");