|
@@ -39,6 +39,7 @@ static enum {
|
|
|
#define LBR_IND_JMP_BIT 6 /* do not capture indirect jumps */
|
|
|
#define LBR_REL_JMP_BIT 7 /* do not capture relative jumps */
|
|
|
#define LBR_FAR_BIT 8 /* do not capture far branches */
|
|
|
+#define LBR_CALL_STACK_BIT 9 /* enable call stack */
|
|
|
|
|
|
#define LBR_KERNEL (1 << LBR_KERNEL_BIT)
|
|
|
#define LBR_USER (1 << LBR_USER_BIT)
|
|
@@ -49,6 +50,7 @@ static enum {
|
|
|
#define LBR_REL_JMP (1 << LBR_REL_JMP_BIT)
|
|
|
#define LBR_IND_JMP (1 << LBR_IND_JMP_BIT)
|
|
|
#define LBR_FAR (1 << LBR_FAR_BIT)
|
|
|
+#define LBR_CALL_STACK (1 << LBR_CALL_STACK_BIT)
|
|
|
|
|
|
#define LBR_PLM (LBR_KERNEL | LBR_USER)
|
|
|
|
|
@@ -74,24 +76,25 @@ static enum {
|
|
|
* x86control flow changes include branches, interrupts, traps, faults
|
|
|
*/
|
|
|
enum {
|
|
|
- X86_BR_NONE = 0, /* unknown */
|
|
|
-
|
|
|
- X86_BR_USER = 1 << 0, /* branch target is user */
|
|
|
- X86_BR_KERNEL = 1 << 1, /* branch target is kernel */
|
|
|
-
|
|
|
- X86_BR_CALL = 1 << 2, /* call */
|
|
|
- X86_BR_RET = 1 << 3, /* return */
|
|
|
- X86_BR_SYSCALL = 1 << 4, /* syscall */
|
|
|
- X86_BR_SYSRET = 1 << 5, /* syscall return */
|
|
|
- X86_BR_INT = 1 << 6, /* sw interrupt */
|
|
|
- X86_BR_IRET = 1 << 7, /* return from interrupt */
|
|
|
- X86_BR_JCC = 1 << 8, /* conditional */
|
|
|
- X86_BR_JMP = 1 << 9, /* jump */
|
|
|
- X86_BR_IRQ = 1 << 10,/* hw interrupt or trap or fault */
|
|
|
- X86_BR_IND_CALL = 1 << 11,/* indirect calls */
|
|
|
- X86_BR_ABORT = 1 << 12,/* transaction abort */
|
|
|
- X86_BR_IN_TX = 1 << 13,/* in transaction */
|
|
|
- X86_BR_NO_TX = 1 << 14,/* not in transaction */
|
|
|
+ X86_BR_NONE = 0, /* unknown */
|
|
|
+
|
|
|
+ X86_BR_USER = 1 << 0, /* branch target is user */
|
|
|
+ X86_BR_KERNEL = 1 << 1, /* branch target is kernel */
|
|
|
+
|
|
|
+ X86_BR_CALL = 1 << 2, /* call */
|
|
|
+ X86_BR_RET = 1 << 3, /* return */
|
|
|
+ X86_BR_SYSCALL = 1 << 4, /* syscall */
|
|
|
+ X86_BR_SYSRET = 1 << 5, /* syscall return */
|
|
|
+ X86_BR_INT = 1 << 6, /* sw interrupt */
|
|
|
+ X86_BR_IRET = 1 << 7, /* return from interrupt */
|
|
|
+ X86_BR_JCC = 1 << 8, /* conditional */
|
|
|
+ X86_BR_JMP = 1 << 9, /* jump */
|
|
|
+ X86_BR_IRQ = 1 << 10,/* hw interrupt or trap or fault */
|
|
|
+ X86_BR_IND_CALL = 1 << 11,/* indirect calls */
|
|
|
+ X86_BR_ABORT = 1 << 12,/* transaction abort */
|
|
|
+ X86_BR_IN_TX = 1 << 13,/* in transaction */
|
|
|
+ X86_BR_NO_TX = 1 << 14,/* not in transaction */
|
|
|
+ X86_BR_CALL_STACK = 1 << 15,/* call stack */
|
|
|
};
|
|
|
|
|
|
#define X86_BR_PLM (X86_BR_USER | X86_BR_KERNEL)
|
|
@@ -373,7 +376,7 @@ void intel_pmu_lbr_read(void)
|
|
|
* - in case there is no HW filter
|
|
|
* - in case the HW filter has errata or limitations
|
|
|
*/
|
|
|
-static void intel_pmu_setup_sw_lbr_filter(struct perf_event *event)
|
|
|
+static int intel_pmu_setup_sw_lbr_filter(struct perf_event *event)
|
|
|
{
|
|
|
u64 br_type = event->attr.branch_sample_type;
|
|
|
int mask = 0;
|
|
@@ -410,11 +413,21 @@ static void intel_pmu_setup_sw_lbr_filter(struct perf_event *event)
|
|
|
if (br_type & PERF_SAMPLE_BRANCH_COND)
|
|
|
mask |= X86_BR_JCC;
|
|
|
|
|
|
+ if (br_type & PERF_SAMPLE_BRANCH_CALL_STACK) {
|
|
|
+ if (!x86_pmu_has_lbr_callstack())
|
|
|
+ return -EOPNOTSUPP;
|
|
|
+ if (mask & ~(X86_BR_USER | X86_BR_KERNEL))
|
|
|
+ return -EINVAL;
|
|
|
+ mask |= X86_BR_CALL | X86_BR_IND_CALL | X86_BR_RET |
|
|
|
+ X86_BR_CALL_STACK;
|
|
|
+ }
|
|
|
+
|
|
|
/*
|
|
|
* stash actual user request into reg, it may
|
|
|
* be used by fixup code for some CPU
|
|
|
*/
|
|
|
event->hw.branch_reg.reg = mask;
|
|
|
+ return 0;
|
|
|
}
|
|
|
|
|
|
/*
|
|
@@ -443,8 +456,12 @@ static int intel_pmu_setup_hw_lbr_filter(struct perf_event *event)
|
|
|
reg = &event->hw.branch_reg;
|
|
|
reg->idx = EXTRA_REG_LBR;
|
|
|
|
|
|
- /* LBR_SELECT operates in suppress mode so invert mask */
|
|
|
- reg->config = ~mask & x86_pmu.lbr_sel_mask;
|
|
|
+ /*
|
|
|
+ * The first 9 bits (LBR_SEL_MASK) in LBR_SELECT operate
|
|
|
+ * in suppress mode. So LBR_SELECT should be set to
|
|
|
+ * (~mask & LBR_SEL_MASK) | (mask & ~LBR_SEL_MASK)
|
|
|
+ */
|
|
|
+ reg->config = mask ^ x86_pmu.lbr_sel_mask;
|
|
|
|
|
|
return 0;
|
|
|
}
|
|
@@ -462,7 +479,9 @@ int intel_pmu_setup_lbr_filter(struct perf_event *event)
|
|
|
/*
|
|
|
* setup SW LBR filter
|
|
|
*/
|
|
|
- intel_pmu_setup_sw_lbr_filter(event);
|
|
|
+ ret = intel_pmu_setup_sw_lbr_filter(event);
|
|
|
+ if (ret)
|
|
|
+ return ret;
|
|
|
|
|
|
/*
|
|
|
* setup HW LBR filter, if any
|
|
@@ -732,6 +751,20 @@ static const int snb_lbr_sel_map[PERF_SAMPLE_BRANCH_SELECT_MAP_SIZE] = {
|
|
|
[PERF_SAMPLE_BRANCH_COND_SHIFT] = LBR_JCC,
|
|
|
};
|
|
|
|
|
|
+static const int hsw_lbr_sel_map[PERF_SAMPLE_BRANCH_SELECT_MAP_SIZE] = {
|
|
|
+ [PERF_SAMPLE_BRANCH_ANY_SHIFT] = LBR_ANY,
|
|
|
+ [PERF_SAMPLE_BRANCH_USER_SHIFT] = LBR_USER,
|
|
|
+ [PERF_SAMPLE_BRANCH_KERNEL_SHIFT] = LBR_KERNEL,
|
|
|
+ [PERF_SAMPLE_BRANCH_HV_SHIFT] = LBR_IGN,
|
|
|
+ [PERF_SAMPLE_BRANCH_ANY_RETURN_SHIFT] = LBR_RETURN | LBR_FAR,
|
|
|
+ [PERF_SAMPLE_BRANCH_ANY_CALL_SHIFT] = LBR_REL_CALL | LBR_IND_CALL
|
|
|
+ | LBR_FAR,
|
|
|
+ [PERF_SAMPLE_BRANCH_IND_CALL_SHIFT] = LBR_IND_CALL,
|
|
|
+ [PERF_SAMPLE_BRANCH_COND_SHIFT] = LBR_JCC,
|
|
|
+ [PERF_SAMPLE_BRANCH_CALL_STACK_SHIFT] = LBR_REL_CALL | LBR_IND_CALL
|
|
|
+ | LBR_RETURN | LBR_CALL_STACK,
|
|
|
+};
|
|
|
+
|
|
|
/* core */
|
|
|
void __init intel_pmu_lbr_init_core(void)
|
|
|
{
|
|
@@ -788,6 +821,20 @@ void __init intel_pmu_lbr_init_snb(void)
|
|
|
pr_cont("16-deep LBR, ");
|
|
|
}
|
|
|
|
|
|
+/* haswell */
|
|
|
+void intel_pmu_lbr_init_hsw(void)
|
|
|
+{
|
|
|
+ x86_pmu.lbr_nr = 16;
|
|
|
+ x86_pmu.lbr_tos = MSR_LBR_TOS;
|
|
|
+ x86_pmu.lbr_from = MSR_LBR_NHM_FROM;
|
|
|
+ x86_pmu.lbr_to = MSR_LBR_NHM_TO;
|
|
|
+
|
|
|
+ x86_pmu.lbr_sel_mask = LBR_SEL_MASK;
|
|
|
+ x86_pmu.lbr_sel_map = hsw_lbr_sel_map;
|
|
|
+
|
|
|
+ pr_cont("16-deep LBR, ");
|
|
|
+}
|
|
|
+
|
|
|
/* atom */
|
|
|
void __init intel_pmu_lbr_init_atom(void)
|
|
|
{
|