Skip to content
体验新版
项目
组织
正在加载...
登录
切换导航
打开侧边栏
openeuler
Kernel
提交
94f2be50
K
Kernel
项目概览
openeuler
/
Kernel
1 年多 前同步成功
通知
8
Star
0
Fork
0
代码
文件
提交
分支
Tags
贡献者
分支图
Diff
Issue
0
列表
看板
标记
里程碑
合并请求
0
DevOps
流水线
流水线任务
计划
Wiki
0
Wiki
分析
仓库
DevOps
项目成员
Pages
K
Kernel
项目概览
项目概览
详情
发布
仓库
仓库
文件
提交
分支
标签
贡献者
分支图
比较
Issue
0
Issue
0
列表
看板
标记
里程碑
合并请求
0
合并请求
0
Pages
DevOps
DevOps
流水线
流水线任务
计划
分析
分析
仓库分析
DevOps
Wiki
0
Wiki
成员
成员
收起侧边栏
关闭侧边栏
动态
分支图
创建新Issue
流水线任务
提交
Issue看板
提交
94f2be50
编写于
6月 29, 2021
作者:
P
Petr Mladek
浏览文件
操作
浏览文件
下载
差异文件
Merge branch 'printk-rework' into for-linus
上级
d8c03214
3342aa8e
变更
3
隐藏空白更改
内联
并排
Showing
3 changed file
with
159 addition
and
36 deletion
+159
-36
include/linux/printk.h
include/linux/printk.h
+41
-0
kernel/printk/printk.c
kernel/printk/printk.c
+116
-0
lib/dump_stack.c
lib/dump_stack.c
+2
-36
未找到文件。
include/linux/printk.h
浏览文件 @
94f2be50
...
...
@@ -282,6 +282,47 @@ static inline void printk_safe_flush_on_panic(void)
}
#endif
#ifdef CONFIG_SMP
extern
int
__printk_cpu_trylock
(
void
);
extern
void
__printk_wait_on_cpu_lock
(
void
);
extern
void
__printk_cpu_unlock
(
void
);
/**
* printk_cpu_lock_irqsave() - Acquire the printk cpu-reentrant spinning
* lock and disable interrupts.
* @flags: Stack-allocated storage for saving local interrupt state,
* to be passed to printk_cpu_unlock_irqrestore().
*
* If the lock is owned by another CPU, spin until it becomes available.
* Interrupts are restored while spinning.
*/
#define printk_cpu_lock_irqsave(flags) \
for (;;) { \
local_irq_save(flags); \
if (__printk_cpu_trylock()) \
break; \
local_irq_restore(flags); \
__printk_wait_on_cpu_lock(); \
}
/**
* printk_cpu_unlock_irqrestore() - Release the printk cpu-reentrant spinning
* lock and restore interrupts.
* @flags: Caller's saved interrupt state, from printk_cpu_lock_irqsave().
*/
#define printk_cpu_unlock_irqrestore(flags) \
do { \
__printk_cpu_unlock(); \
local_irq_restore(flags); \
} while (0) \
#else
#define printk_cpu_lock_irqsave(flags) ((void)flags)
#define printk_cpu_unlock_irqrestore(flags) ((void)flags)
#endif
/* CONFIG_SMP */
extern
int
kptr_restrict
;
/**
...
...
kernel/printk/printk.c
浏览文件 @
94f2be50
...
...
@@ -3531,3 +3531,119 @@ void kmsg_dump_rewind(struct kmsg_dump_iter *iter)
EXPORT_SYMBOL_GPL
(
kmsg_dump_rewind
);
#endif
#ifdef CONFIG_SMP
static
atomic_t
printk_cpulock_owner
=
ATOMIC_INIT
(
-
1
);
static
atomic_t
printk_cpulock_nested
=
ATOMIC_INIT
(
0
);
/**
* __printk_wait_on_cpu_lock() - Busy wait until the printk cpu-reentrant
* spinning lock is not owned by any CPU.
*
* Context: Any context.
*/
void
__printk_wait_on_cpu_lock
(
void
)
{
do
{
cpu_relax
();
}
while
(
atomic_read
(
&
printk_cpulock_owner
)
!=
-
1
);
}
EXPORT_SYMBOL
(
__printk_wait_on_cpu_lock
);
/**
* __printk_cpu_trylock() - Try to acquire the printk cpu-reentrant
* spinning lock.
*
* If no processor has the lock, the calling processor takes the lock and
* becomes the owner. If the calling processor is already the owner of the
* lock, this function succeeds immediately.
*
* Context: Any context. Expects interrupts to be disabled.
* Return: 1 on success, otherwise 0.
*/
int
__printk_cpu_trylock
(
void
)
{
int
cpu
;
int
old
;
cpu
=
smp_processor_id
();
/*
* Guarantee loads and stores from this CPU when it is the lock owner
* are _not_ visible to the previous lock owner. This pairs with
* __printk_cpu_unlock:B.
*
* Memory barrier involvement:
*
* If __printk_cpu_trylock:A reads from __printk_cpu_unlock:B, then
* __printk_cpu_unlock:A can never read from __printk_cpu_trylock:B.
*
* Relies on:
*
* RELEASE from __printk_cpu_unlock:A to __printk_cpu_unlock:B
* of the previous CPU
* matching
* ACQUIRE from __printk_cpu_trylock:A to __printk_cpu_trylock:B
* of this CPU
*/
old
=
atomic_cmpxchg_acquire
(
&
printk_cpulock_owner
,
-
1
,
cpu
);
/* LMM(__printk_cpu_trylock:A) */
if
(
old
==
-
1
)
{
/*
* This CPU is now the owner and begins loading/storing
* data: LMM(__printk_cpu_trylock:B)
*/
return
1
;
}
else
if
(
old
==
cpu
)
{
/* This CPU is already the owner. */
atomic_inc
(
&
printk_cpulock_nested
);
return
1
;
}
return
0
;
}
EXPORT_SYMBOL
(
__printk_cpu_trylock
);
/**
* __printk_cpu_unlock() - Release the printk cpu-reentrant spinning lock.
*
* The calling processor must be the owner of the lock.
*
* Context: Any context. Expects interrupts to be disabled.
*/
void
__printk_cpu_unlock
(
void
)
{
if
(
atomic_read
(
&
printk_cpulock_nested
))
{
atomic_dec
(
&
printk_cpulock_nested
);
return
;
}
/*
* This CPU is finished loading/storing data:
* LMM(__printk_cpu_unlock:A)
*/
/*
* Guarantee loads and stores from this CPU when it was the
* lock owner are visible to the next lock owner. This pairs
* with __printk_cpu_trylock:A.
*
* Memory barrier involvement:
*
* If __printk_cpu_trylock:A reads from __printk_cpu_unlock:B,
* then __printk_cpu_trylock:B reads from __printk_cpu_unlock:A.
*
* Relies on:
*
* RELEASE from __printk_cpu_unlock:A to __printk_cpu_unlock:B
* of this CPU
* matching
* ACQUIRE from __printk_cpu_trylock:A to __printk_cpu_trylock:B
* of the next CPU
*/
atomic_set_release
(
&
printk_cpulock_owner
,
-
1
);
/* LMM(__printk_cpu_unlock:B) */
}
EXPORT_SYMBOL
(
__printk_cpu_unlock
);
#endif
/* CONFIG_SMP */
lib/dump_stack.c
浏览文件 @
94f2be50
...
...
@@ -84,50 +84,16 @@ static void __dump_stack(void)
*
* Architectures can override this implementation by implementing its own.
*/
#ifdef CONFIG_SMP
static
atomic_t
dump_lock
=
ATOMIC_INIT
(
-
1
);
asmlinkage
__visible
void
dump_stack
(
void
)
{
unsigned
long
flags
;
int
was_locked
;
int
old
;
int
cpu
;
/*
* Permit this cpu to perform nested stack dumps while serialising
* against other CPUs
*/
retry:
local_irq_save
(
flags
);
cpu
=
smp_processor_id
();
old
=
atomic_cmpxchg
(
&
dump_lock
,
-
1
,
cpu
);
if
(
old
==
-
1
)
{
was_locked
=
0
;
}
else
if
(
old
==
cpu
)
{
was_locked
=
1
;
}
else
{
local_irq_restore
(
flags
);
/*
* Wait for the lock to release before jumping to
* atomic_cmpxchg() in order to mitigate the thundering herd
* problem.
*/
do
{
cpu_relax
();
}
while
(
atomic_read
(
&
dump_lock
)
!=
-
1
);
goto
retry
;
}
__dump_stack
();
if
(
!
was_locked
)
atomic_set
(
&
dump_lock
,
-
1
);
local_irq_restore
(
flags
);
}
#else
asmlinkage
__visible
void
dump_stack
(
void
)
{
printk_cpu_lock_irqsave
(
flags
);
__dump_stack
();
printk_cpu_unlock_irqrestore
(
flags
);
}
#endif
EXPORT_SYMBOL
(
dump_stack
);
编辑
预览
Markdown
is supported
0%
请重试
或
添加新附件
.
添加附件
取消
You are about to add
0
people
to the discussion. Proceed with caution.
先完成此消息的编辑!
取消
想要评论请
注册
或
登录