|
@@ -31,8 +31,8 @@ static char x86_stack_ids[][8] = {
|
|
|
#endif
|
|
|
};
|
|
|
|
|
|
-static unsigned long *in_exception_stack(unsigned cpu, unsigned long stack,
|
|
|
- unsigned *usedp, char **idp)
|
|
|
+static unsigned long *in_exception_stack(unsigned long stack, unsigned *usedp,
|
|
|
+ char **idp)
|
|
|
{
|
|
|
unsigned k;
|
|
|
|
|
@@ -41,7 +41,7 @@ static unsigned long *in_exception_stack(unsigned cpu, unsigned long stack,
|
|
|
* 'stack' is in one of them:
|
|
|
*/
|
|
|
for (k = 0; k < N_EXCEPTION_STACKS; k++) {
|
|
|
- unsigned long end = per_cpu(orig_ist, cpu).ist[k];
|
|
|
+ unsigned long end = raw_cpu_ptr(&orig_ist)->ist[k];
|
|
|
/*
|
|
|
* Is 'stack' above this exception frame's end?
|
|
|
* If yes then skip to the next frame.
|
|
@@ -111,7 +111,7 @@ enum stack_type {
|
|
|
};
|
|
|
|
|
|
static enum stack_type
|
|
|
-analyze_stack(int cpu, struct task_struct *task, unsigned long *stack,
|
|
|
+analyze_stack(struct task_struct *task, unsigned long *stack,
|
|
|
unsigned long **stack_end, unsigned long *irq_stack,
|
|
|
unsigned *used, char **id)
|
|
|
{
|
|
@@ -121,8 +121,7 @@ analyze_stack(int cpu, struct task_struct *task, unsigned long *stack,
|
|
|
if ((unsigned long)task_stack_page(task) == addr)
|
|
|
return STACK_IS_NORMAL;
|
|
|
|
|
|
- *stack_end = in_exception_stack(cpu, (unsigned long)stack,
|
|
|
- used, id);
|
|
|
+ *stack_end = in_exception_stack((unsigned long)stack, used, id);
|
|
|
if (*stack_end)
|
|
|
return STACK_IS_EXCEPTION;
|
|
|
|
|
@@ -149,8 +148,7 @@ void dump_trace(struct task_struct *task, struct pt_regs *regs,
|
|
|
unsigned long *stack, unsigned long bp,
|
|
|
const struct stacktrace_ops *ops, void *data)
|
|
|
{
|
|
|
- const unsigned cpu = get_cpu();
|
|
|
- unsigned long *irq_stack = (unsigned long *)per_cpu(irq_stack_ptr, cpu);
|
|
|
+ unsigned long *irq_stack = (unsigned long *)this_cpu_read(irq_stack_ptr);
|
|
|
unsigned used = 0;
|
|
|
int graph = 0;
|
|
|
int done = 0;
|
|
@@ -169,8 +167,8 @@ void dump_trace(struct task_struct *task, struct pt_regs *regs,
|
|
|
enum stack_type stype;
|
|
|
char *id;
|
|
|
|
|
|
- stype = analyze_stack(cpu, task, stack, &stack_end,
|
|
|
- irq_stack, &used, &id);
|
|
|
+ stype = analyze_stack(task, stack, &stack_end, irq_stack, &used,
|
|
|
+ &id);
|
|
|
|
|
|
/* Default finish unless specified to continue */
|
|
|
done = 1;
|
|
@@ -225,7 +223,6 @@ void dump_trace(struct task_struct *task, struct pt_regs *regs,
|
|
|
* This handles the process stack:
|
|
|
*/
|
|
|
bp = ops->walk_stack(task, stack, bp, ops, data, NULL, &graph);
|
|
|
- put_cpu();
|
|
|
}
|
|
|
EXPORT_SYMBOL(dump_trace);
|
|
|
|
|
@@ -236,13 +233,9 @@ show_stack_log_lvl(struct task_struct *task, struct pt_regs *regs,
|
|
|
unsigned long *irq_stack_end;
|
|
|
unsigned long *irq_stack;
|
|
|
unsigned long *stack;
|
|
|
- int cpu;
|
|
|
int i;
|
|
|
|
|
|
- preempt_disable();
|
|
|
- cpu = smp_processor_id();
|
|
|
-
|
|
|
- irq_stack_end = (unsigned long *)(per_cpu(irq_stack_ptr, cpu));
|
|
|
+ irq_stack_end = (unsigned long *)this_cpu_read(irq_stack_ptr);
|
|
|
irq_stack = irq_stack_end - (IRQ_STACK_SIZE / sizeof(long));
|
|
|
|
|
|
sp = sp ? : get_stack_pointer(task, regs);
|
|
@@ -274,7 +267,6 @@ show_stack_log_lvl(struct task_struct *task, struct pt_regs *regs,
|
|
|
stack++;
|
|
|
touch_nmi_watchdog();
|
|
|
}
|
|
|
- preempt_enable();
|
|
|
|
|
|
pr_cont("\n");
|
|
|
show_trace_log_lvl(task, regs, sp, bp, log_lvl);
|