提交 853bec28 编写于 作者: C Cheng Jian 提交者: Zheng Zengkai

livepatch/arm64: check active func in consistency stack checking

hulk inclusion
category: bugfix
bugzilla: 51921
CVE: NA

---------------------------

When doing consistency stack checking, if we try to patch a
function which has been patched already. We should check the
new function(not the origin func) that is activeness currently,
it's always the first entry in list func_node->func_stack.

Example :
	module : origin			livepatch v1		livepatch v2
	func   : old func A -[enable]=> new func A' -[enable]=> new func A''
	check  :		A			A'

when we try to patch function A to new function A'' by livepatch
v2, but the func A has already patched to function A' by livepatch
v1, so function A' which provided in livepatch v1 is active in the
stack instead of origin function A. Even if the long jump method is
used, we jump to the new function A' using a call without LR, the
origin function A will not appear in the stack. We must check the
active function A' in consistency stack checking.
Reviewed-By: NXie XiuQi <xiexiuqi@huawei.com>
Signed-off-by: NCheng Jian <cj.chengjian@huawei.com>
Signed-off-by: NYang Yingliang <yangyingliang@huawei.com>
Signed-off-by: NWang ShaoBo <bobo.shaobowang@huawei.com>
Signed-off-by: NYe Weihua <yeweihua4@huawei.com>
Reviewed-by: NYang Jihong <yangjihong1@huawei.com>
Signed-off-by: NZheng Zengkai <zhengzengkai@huawei.com>
上级 2085b555
......@@ -44,6 +44,33 @@ static inline bool offset_in_range(unsigned long pc, unsigned long addr,
}
#endif
#define LJMP_INSN_SIZE 4
struct klp_func_node {
struct list_head node;
struct list_head func_stack;
unsigned long old_addr;
#ifdef CONFIG_ARM64_MODULE_PLTS
u32 old_insns[LJMP_INSN_SIZE];
#else
u32 old_insn;
#endif
};
static LIST_HEAD(klp_func_list);
static struct klp_func_node *klp_find_func_node(unsigned long old_addr)
{
struct klp_func_node *func_node;
list_for_each_entry(func_node, &klp_func_list, node) {
if (func_node->old_addr == old_addr)
return func_node;
}
return NULL;
}
#ifdef CONFIG_LIVEPATCH_STOP_MACHINE_CONSISTENCY
struct walk_stackframe_args {
struct klp_patch *patch;
......@@ -69,6 +96,7 @@ static bool klp_check_activeness_func(void *data, unsigned long pc)
struct klp_func *func;
unsigned long func_addr, func_size;
const char *func_name;
struct klp_func_node *func_node;
if (args->ret)
return false;
......@@ -76,9 +104,33 @@ static bool klp_check_activeness_func(void *data, unsigned long pc)
for (obj = patch->objs; obj->funcs; obj++) {
for (func = obj->funcs; func->old_name; func++) {
if (args->enable) {
func_addr = (unsigned long)func->old_func;
func_size = func->old_size;
/*
* When enable, checking the currently
* active functions.
*/
func_node = klp_find_func_node((unsigned long)func->old_func);
if (!func_node ||
list_empty(&func_node->func_stack)) {
func_addr = (unsigned long)func->old_func;
func_size = func->old_size;
} else {
/*
* Previously patched function
* [the active one]
*/
struct klp_func *prev;
prev = list_first_or_null_rcu(
&func_node->func_stack,
struct klp_func, stack_node);
func_addr = (unsigned long)prev->new_func;
func_size = prev->new_size;
}
} else {
/*
* When disable, check for the function
* itself which to be unpatched.
*/
func_addr = (unsigned long)func->new_func;
func_size = func->new_size;
}
......@@ -147,33 +199,6 @@ int klp_check_calltrace(struct klp_patch *patch, int enable)
}
#endif
#define LJMP_INSN_SIZE 4
struct klp_func_node {
struct list_head node;
struct list_head func_stack;
void *old_func;
#ifdef CONFIG_ARM64_MODULE_PLTS
u32 old_insns[LJMP_INSN_SIZE];
#else
u32 old_insn;
#endif
};
static LIST_HEAD(klp_func_list);
static struct klp_func_node *klp_find_func_node(void *old_func)
{
struct klp_func_node *func_node;
list_for_each_entry(func_node, &klp_func_list, node) {
if (func_node->old_func == old_func)
return func_node;
}
return NULL;
}
int arch_klp_patch_func(struct klp_func *func)
{
struct klp_func_node *func_node;
......@@ -186,7 +211,7 @@ int arch_klp_patch_func(struct klp_func *func)
#endif
int ret = 0;
func_node = klp_find_func_node(func->old_func);
func_node = klp_find_func_node((unsigned long)func->old_func);
if (!func_node) {
func_node = kzalloc(sizeof(*func_node), GFP_ATOMIC);
if (!func_node)
......@@ -194,7 +219,7 @@ int arch_klp_patch_func(struct klp_func *func)
memory_flag = 1;
INIT_LIST_HEAD(&func_node->func_stack);
func_node->old_func = func->old_func;
func_node->old_addr = (unsigned long)func->old_func;
#ifdef CONFIG_ARM64_MODULE_PLTS
for (i = 0; i < LJMP_INSN_SIZE; i++) {
......@@ -265,11 +290,11 @@ void arch_klp_unpatch_func(struct klp_func *func)
int i;
u32 insns[LJMP_INSN_SIZE];
#endif
func_node = klp_find_func_node(func->old_func);
func_node = klp_find_func_node((unsigned long)func->old_func);
if (WARN_ON(!func_node))
return;
pc = (unsigned long)func_node->old_func;
pc = func_node->old_addr;
if (list_is_singular(&func_node->func_stack)) {
#ifdef CONFIG_ARM64_MODULE_PLTS
for (i = 0; i < LJMP_INSN_SIZE; i++)
......
Markdown is supported
0% .
You are about to add 0 people to the discussion. Proceed with caution.
先完成此消息的编辑!
想要评论请 注册