|
@@ -652,21 +652,117 @@ static const struct file_operations uprobe_profile_ops = {
|
|
.release = seq_release,
|
|
.release = seq_release,
|
|
};
|
|
};
|
|
|
|
|
|
|
|
+struct uprobe_cpu_buffer {
|
|
|
|
+ struct mutex mutex;
|
|
|
|
+ void *buf;
|
|
|
|
+};
|
|
|
|
+static struct uprobe_cpu_buffer __percpu *uprobe_cpu_buffer;
|
|
|
|
+static int uprobe_buffer_refcnt;
|
|
|
|
+
|
|
|
|
+static int uprobe_buffer_init(void)
|
|
|
|
+{
|
|
|
|
+ int cpu, err_cpu;
|
|
|
|
+
|
|
|
|
+ uprobe_cpu_buffer = alloc_percpu(struct uprobe_cpu_buffer);
|
|
|
|
+ if (uprobe_cpu_buffer == NULL)
|
|
|
|
+ return -ENOMEM;
|
|
|
|
+
|
|
|
|
+ for_each_possible_cpu(cpu) {
|
|
|
|
+ struct page *p = alloc_pages_node(cpu_to_node(cpu),
|
|
|
|
+ GFP_KERNEL, 0);
|
|
|
|
+ if (p == NULL) {
|
|
|
|
+ err_cpu = cpu;
|
|
|
|
+ goto err;
|
|
|
|
+ }
|
|
|
|
+ per_cpu_ptr(uprobe_cpu_buffer, cpu)->buf = page_address(p);
|
|
|
|
+ mutex_init(&per_cpu_ptr(uprobe_cpu_buffer, cpu)->mutex);
|
|
|
|
+ }
|
|
|
|
+
|
|
|
|
+ return 0;
|
|
|
|
+
|
|
|
|
+err:
|
|
|
|
+ for_each_possible_cpu(cpu) {
|
|
|
|
+ if (cpu == err_cpu)
|
|
|
|
+ break;
|
|
|
|
+ free_page((unsigned long)per_cpu_ptr(uprobe_cpu_buffer, cpu)->buf);
|
|
|
|
+ }
|
|
|
|
+
|
|
|
|
+ free_percpu(uprobe_cpu_buffer);
|
|
|
|
+ return -ENOMEM;
|
|
|
|
+}
|
|
|
|
+
|
|
|
|
+static int uprobe_buffer_enable(void)
|
|
|
|
+{
|
|
|
|
+ int ret = 0;
|
|
|
|
+
|
|
|
|
+ BUG_ON(!mutex_is_locked(&event_mutex));
|
|
|
|
+
|
|
|
|
+ if (uprobe_buffer_refcnt++ == 0) {
|
|
|
|
+ ret = uprobe_buffer_init();
|
|
|
|
+ if (ret < 0)
|
|
|
|
+ uprobe_buffer_refcnt--;
|
|
|
|
+ }
|
|
|
|
+
|
|
|
|
+ return ret;
|
|
|
|
+}
|
|
|
|
+
|
|
|
|
+static void uprobe_buffer_disable(void)
|
|
|
|
+{
|
|
|
|
+ BUG_ON(!mutex_is_locked(&event_mutex));
|
|
|
|
+
|
|
|
|
+ if (--uprobe_buffer_refcnt == 0) {
|
|
|
|
+ free_percpu(uprobe_cpu_buffer);
|
|
|
|
+ uprobe_cpu_buffer = NULL;
|
|
|
|
+ }
|
|
|
|
+}
|
|
|
|
+
|
|
|
|
+static struct uprobe_cpu_buffer *uprobe_buffer_get(void)
|
|
|
|
+{
|
|
|
|
+ struct uprobe_cpu_buffer *ucb;
|
|
|
|
+ int cpu;
|
|
|
|
+
|
|
|
|
+ cpu = raw_smp_processor_id();
|
|
|
|
+ ucb = per_cpu_ptr(uprobe_cpu_buffer, cpu);
|
|
|
|
+
|
|
|
|
+ /*
|
|
|
|
+ * Use per-cpu buffers for fastest access, but we might migrate
|
|
|
|
+ * so the mutex makes sure we have sole access to it.
|
|
|
|
+ */
|
|
|
|
+ mutex_lock(&ucb->mutex);
|
|
|
|
+
|
|
|
|
+ return ucb;
|
|
|
|
+}
|
|
|
|
+
|
|
|
|
+static void uprobe_buffer_put(struct uprobe_cpu_buffer *ucb)
|
|
|
|
+{
|
|
|
|
+ mutex_unlock(&ucb->mutex);
|
|
|
|
+}
|
|
|
|
+
|
|
static void uprobe_trace_print(struct trace_uprobe *tu,
|
|
static void uprobe_trace_print(struct trace_uprobe *tu,
|
|
unsigned long func, struct pt_regs *regs)
|
|
unsigned long func, struct pt_regs *regs)
|
|
{
|
|
{
|
|
struct uprobe_trace_entry_head *entry;
|
|
struct uprobe_trace_entry_head *entry;
|
|
struct ring_buffer_event *event;
|
|
struct ring_buffer_event *event;
|
|
struct ring_buffer *buffer;
|
|
struct ring_buffer *buffer;
|
|
|
|
+ struct uprobe_cpu_buffer *ucb;
|
|
void *data;
|
|
void *data;
|
|
- int size, i;
|
|
|
|
|
|
+ int size, dsize, esize;
|
|
struct ftrace_event_call *call = &tu->tp.call;
|
|
struct ftrace_event_call *call = &tu->tp.call;
|
|
|
|
|
|
- size = SIZEOF_TRACE_ENTRY(is_ret_probe(tu));
|
|
|
|
|
|
+ dsize = __get_data_size(&tu->tp, regs);
|
|
|
|
+ esize = SIZEOF_TRACE_ENTRY(is_ret_probe(tu));
|
|
|
|
+
|
|
|
|
+ if (WARN_ON_ONCE(!uprobe_cpu_buffer || tu->tp.size + dsize > PAGE_SIZE))
|
|
|
|
+ return;
|
|
|
|
+
|
|
|
|
+ ucb = uprobe_buffer_get();
|
|
|
|
+ store_trace_args(esize, &tu->tp, regs, ucb->buf, dsize);
|
|
|
|
+
|
|
|
|
+ size = esize + tu->tp.size + dsize;
|
|
event = trace_current_buffer_lock_reserve(&buffer, call->event.type,
|
|
event = trace_current_buffer_lock_reserve(&buffer, call->event.type,
|
|
- size + tu->tp.size, 0, 0);
|
|
|
|
|
|
+ size, 0, 0);
|
|
if (!event)
|
|
if (!event)
|
|
- return;
|
|
|
|
|
|
+ goto out;
|
|
|
|
|
|
entry = ring_buffer_event_data(event);
|
|
entry = ring_buffer_event_data(event);
|
|
if (is_ret_probe(tu)) {
|
|
if (is_ret_probe(tu)) {
|
|
@@ -678,13 +774,13 @@ static void uprobe_trace_print(struct trace_uprobe *tu,
|
|
data = DATAOF_TRACE_ENTRY(entry, false);
|
|
data = DATAOF_TRACE_ENTRY(entry, false);
|
|
}
|
|
}
|
|
|
|
|
|
- for (i = 0; i < tu->tp.nr_args; i++) {
|
|
|
|
- call_fetch(&tu->tp.args[i].fetch, regs,
|
|
|
|
- data + tu->tp.args[i].offset);
|
|
|
|
- }
|
|
|
|
|
|
+ memcpy(data, ucb->buf, tu->tp.size + dsize);
|
|
|
|
|
|
if (!call_filter_check_discard(call, entry, buffer, event))
|
|
if (!call_filter_check_discard(call, entry, buffer, event))
|
|
trace_buffer_unlock_commit(buffer, event, 0, 0);
|
|
trace_buffer_unlock_commit(buffer, event, 0, 0);
|
|
|
|
+
|
|
|
|
+out:
|
|
|
|
+ uprobe_buffer_put(ucb);
|
|
}
|
|
}
|
|
|
|
|
|
/* uprobe handler */
|
|
/* uprobe handler */
|
|
@@ -752,6 +848,10 @@ probe_event_enable(struct trace_uprobe *tu, int flag, filter_func_t filter)
|
|
if (trace_probe_is_enabled(&tu->tp))
|
|
if (trace_probe_is_enabled(&tu->tp))
|
|
return -EINTR;
|
|
return -EINTR;
|
|
|
|
|
|
|
|
+ ret = uprobe_buffer_enable();
|
|
|
|
+ if (ret < 0)
|
|
|
|
+ return ret;
|
|
|
|
+
|
|
WARN_ON(!uprobe_filter_is_empty(&tu->filter));
|
|
WARN_ON(!uprobe_filter_is_empty(&tu->filter));
|
|
|
|
|
|
tu->tp.flags |= flag;
|
|
tu->tp.flags |= flag;
|
|
@@ -772,6 +872,8 @@ static void probe_event_disable(struct trace_uprobe *tu, int flag)
|
|
|
|
|
|
uprobe_unregister(tu->inode, tu->offset, &tu->consumer);
|
|
uprobe_unregister(tu->inode, tu->offset, &tu->consumer);
|
|
tu->tp.flags &= ~flag;
|
|
tu->tp.flags &= ~flag;
|
|
|
|
+
|
|
|
|
+ uprobe_buffer_disable();
|
|
}
|
|
}
|
|
|
|
|
|
static int uprobe_event_define_fields(struct ftrace_event_call *event_call)
|
|
static int uprobe_event_define_fields(struct ftrace_event_call *event_call)
|
|
@@ -898,11 +1000,24 @@ static void uprobe_perf_print(struct trace_uprobe *tu,
|
|
struct ftrace_event_call *call = &tu->tp.call;
|
|
struct ftrace_event_call *call = &tu->tp.call;
|
|
struct uprobe_trace_entry_head *entry;
|
|
struct uprobe_trace_entry_head *entry;
|
|
struct hlist_head *head;
|
|
struct hlist_head *head;
|
|
|
|
+ struct uprobe_cpu_buffer *ucb;
|
|
void *data;
|
|
void *data;
|
|
- int size, rctx, i;
|
|
|
|
|
|
+ int size, dsize, esize;
|
|
|
|
+ int rctx;
|
|
|
|
+
|
|
|
|
+ dsize = __get_data_size(&tu->tp, regs);
|
|
|
|
+ esize = SIZEOF_TRACE_ENTRY(is_ret_probe(tu));
|
|
|
|
|
|
- size = SIZEOF_TRACE_ENTRY(is_ret_probe(tu));
|
|
|
|
- size = ALIGN(size + tu->tp.size + sizeof(u32), sizeof(u64)) - sizeof(u32);
|
|
|
|
|
|
+ if (WARN_ON_ONCE(!uprobe_cpu_buffer))
|
|
|
|
+ return;
|
|
|
|
+
|
|
|
|
+ size = esize + tu->tp.size + dsize;
|
|
|
|
+ size = ALIGN(size + sizeof(u32), sizeof(u64)) - sizeof(u32);
|
|
|
|
+ if (WARN_ONCE(size > PERF_MAX_TRACE_SIZE, "profile buffer not large enough"))
|
|
|
|
+ return;
|
|
|
|
+
|
|
|
|
+ ucb = uprobe_buffer_get();
|
|
|
|
+ store_trace_args(esize, &tu->tp, regs, ucb->buf, dsize);
|
|
|
|
|
|
preempt_disable();
|
|
preempt_disable();
|
|
head = this_cpu_ptr(call->perf_events);
|
|
head = this_cpu_ptr(call->perf_events);
|
|
@@ -922,15 +1037,18 @@ static void uprobe_perf_print(struct trace_uprobe *tu,
|
|
data = DATAOF_TRACE_ENTRY(entry, false);
|
|
data = DATAOF_TRACE_ENTRY(entry, false);
|
|
}
|
|
}
|
|
|
|
|
|
- for (i = 0; i < tu->tp.nr_args; i++) {
|
|
|
|
- struct probe_arg *parg = &tu->tp.args[i];
|
|
|
|
|
|
+ memcpy(data, ucb->buf, tu->tp.size + dsize);
|
|
|
|
+
|
|
|
|
+ if (size - esize > tu->tp.size + dsize) {
|
|
|
|
+ int len = tu->tp.size + dsize;
|
|
|
|
|
|
- call_fetch(&parg->fetch, regs, data + parg->offset);
|
|
|
|
|
|
+ memset(data + len, 0, size - esize - len);
|
|
}
|
|
}
|
|
|
|
|
|
perf_trace_buf_submit(entry, size, rctx, 0, 1, regs, head, NULL);
|
|
perf_trace_buf_submit(entry, size, rctx, 0, 1, regs, head, NULL);
|
|
out:
|
|
out:
|
|
preempt_enable();
|
|
preempt_enable();
|
|
|
|
+ uprobe_buffer_put(ucb);
|
|
}
|
|
}
|
|
|
|
|
|
/* uprobe profile handler */
|
|
/* uprobe profile handler */
|