|
@@ -322,7 +322,6 @@ void prepare_ftrace_return(unsigned long *parent_ra_addr, unsigned long self_ra,
|
|
|
unsigned long fp)
|
|
|
{
|
|
|
unsigned long old_parent_ra;
|
|
|
- struct ftrace_graph_ent trace;
|
|
|
unsigned long return_hooker = (unsigned long)
|
|
|
&return_to_handler;
|
|
|
int faulted, insns;
|
|
@@ -369,12 +368,6 @@ void prepare_ftrace_return(unsigned long *parent_ra_addr, unsigned long self_ra,
|
|
|
if (unlikely(faulted))
|
|
|
goto out;
|
|
|
|
|
|
- if (ftrace_push_return_trace(old_parent_ra, self_ra, &trace.depth, fp,
|
|
|
- NULL) == -EBUSY) {
|
|
|
- *parent_ra_addr = old_parent_ra;
|
|
|
- return;
|
|
|
- }
|
|
|
-
|
|
|
/*
|
|
|
* Get the recorded ip of the current mcount calling site in the
|
|
|
* __mcount_loc section, which will be used to filter the function
|
|
@@ -382,13 +375,10 @@ void prepare_ftrace_return(unsigned long *parent_ra_addr, unsigned long self_ra,
|
|
|
*/
|
|
|
|
|
|
insns = core_kernel_text(self_ra) ? 2 : MCOUNT_OFFSET_INSNS + 1;
|
|
|
- trace.func = self_ra - (MCOUNT_INSN_SIZE * insns);
|
|
|
+ self_ra -= (MCOUNT_INSN_SIZE * insns);
|
|
|
|
|
|
- /* Only trace if the calling function expects to */
|
|
|
- if (!ftrace_graph_entry(&trace)) {
|
|
|
- current->curr_ret_stack--;
|
|
|
+ if (function_graph_enter(old_parent_ra, self_ra, fp, NULL))
|
|
|
*parent_ra_addr = old_parent_ra;
|
|
|
- }
|
|
|
return;
|
|
|
out:
|
|
|
ftrace_graph_stop();
|