record.c 5.4 KB

123456789101112131415161718192021222324252627282930313233343536373839404142434445464748495051525354555657585960616263646566676869707172737475767778798081828384858687888990919293949596979899100101102103104105106107108109110111112113114115116117118119120121122123124125126127128129130131132133134135136137138139140141142143144145146147148149150151152153154155156157158159160161162163164165166167168169170171172173174175176177178179180181182183184185186187188189190191192193194195196197198199200201202203204205206207208209210211212213214215216217218219220221222223224225226227228229230231232233234235236237238239240241242243244245246247248249250251252
  1. #include "evlist.h"
  2. #include "evsel.h"
  3. #include "cpumap.h"
  4. #include "parse-events.h"
  5. #include <api/fs/fs.h>
  6. #include "util.h"
  7. #include "cloexec.h"
  8. typedef void (*setup_probe_fn_t)(struct perf_evsel *evsel);
  9. static int perf_do_probe_api(setup_probe_fn_t fn, int cpu, const char *str)
  10. {
  11. struct perf_evlist *evlist;
  12. struct perf_evsel *evsel;
  13. unsigned long flags = perf_event_open_cloexec_flag();
  14. int err = -EAGAIN, fd;
  15. static pid_t pid = -1;
  16. evlist = perf_evlist__new();
  17. if (!evlist)
  18. return -ENOMEM;
  19. if (parse_events(evlist, str, NULL))
  20. goto out_delete;
  21. evsel = perf_evlist__first(evlist);
  22. while (1) {
  23. fd = sys_perf_event_open(&evsel->attr, pid, cpu, -1, flags);
  24. if (fd < 0) {
  25. if (pid == -1 && errno == EACCES) {
  26. pid = 0;
  27. continue;
  28. }
  29. goto out_delete;
  30. }
  31. break;
  32. }
  33. close(fd);
  34. fn(evsel);
  35. fd = sys_perf_event_open(&evsel->attr, pid, cpu, -1, flags);
  36. if (fd < 0) {
  37. if (errno == EINVAL)
  38. err = -EINVAL;
  39. goto out_delete;
  40. }
  41. close(fd);
  42. err = 0;
  43. out_delete:
  44. perf_evlist__delete(evlist);
  45. return err;
  46. }
  47. static bool perf_probe_api(setup_probe_fn_t fn)
  48. {
  49. const char *try[] = {"cycles:u", "instructions:u", "cpu-clock:u", NULL};
  50. struct cpu_map *cpus;
  51. int cpu, ret, i = 0;
  52. cpus = cpu_map__new(NULL);
  53. if (!cpus)
  54. return false;
  55. cpu = cpus->map[0];
  56. cpu_map__put(cpus);
  57. do {
  58. ret = perf_do_probe_api(fn, cpu, try[i++]);
  59. if (!ret)
  60. return true;
  61. } while (ret == -EAGAIN && try[i]);
  62. return false;
  63. }
  64. static void perf_probe_sample_identifier(struct perf_evsel *evsel)
  65. {
  66. evsel->attr.sample_type |= PERF_SAMPLE_IDENTIFIER;
  67. }
  68. static void perf_probe_comm_exec(struct perf_evsel *evsel)
  69. {
  70. evsel->attr.comm_exec = 1;
  71. }
  72. bool perf_can_sample_identifier(void)
  73. {
  74. return perf_probe_api(perf_probe_sample_identifier);
  75. }
  76. static bool perf_can_comm_exec(void)
  77. {
  78. return perf_probe_api(perf_probe_comm_exec);
  79. }
  80. void perf_evlist__config(struct perf_evlist *evlist, struct record_opts *opts)
  81. {
  82. struct perf_evsel *evsel;
  83. bool use_sample_identifier = false;
  84. bool use_comm_exec;
  85. /*
  86. * Set the evsel leader links before we configure attributes,
  87. * since some might depend on this info.
  88. */
  89. if (opts->group)
  90. perf_evlist__set_leader(evlist);
  91. if (evlist->cpus->map[0] < 0)
  92. opts->no_inherit = true;
  93. use_comm_exec = perf_can_comm_exec();
  94. evlist__for_each(evlist, evsel) {
  95. perf_evsel__config(evsel, opts);
  96. if (evsel->tracking && use_comm_exec)
  97. evsel->attr.comm_exec = 1;
  98. }
  99. if (opts->full_auxtrace) {
  100. /*
  101. * Need to be able to synthesize and parse selected events with
  102. * arbitrary sample types, which requires always being able to
  103. * match the id.
  104. */
  105. use_sample_identifier = perf_can_sample_identifier();
  106. evlist__for_each(evlist, evsel)
  107. perf_evsel__set_sample_id(evsel, use_sample_identifier);
  108. } else if (evlist->nr_entries > 1) {
  109. struct perf_evsel *first = perf_evlist__first(evlist);
  110. evlist__for_each(evlist, evsel) {
  111. if (evsel->attr.sample_type == first->attr.sample_type)
  112. continue;
  113. use_sample_identifier = perf_can_sample_identifier();
  114. break;
  115. }
  116. evlist__for_each(evlist, evsel)
  117. perf_evsel__set_sample_id(evsel, use_sample_identifier);
  118. }
  119. perf_evlist__set_id_pos(evlist);
  120. }
  121. static int get_max_rate(unsigned int *rate)
  122. {
  123. return sysctl__read_int("kernel/perf_event_max_sample_rate", (int *)rate);
  124. }
  125. static int record_opts__config_freq(struct record_opts *opts)
  126. {
  127. bool user_freq = opts->user_freq != UINT_MAX;
  128. unsigned int max_rate;
  129. if (opts->user_interval != ULLONG_MAX)
  130. opts->default_interval = opts->user_interval;
  131. if (user_freq)
  132. opts->freq = opts->user_freq;
  133. /*
  134. * User specified count overrides default frequency.
  135. */
  136. if (opts->default_interval)
  137. opts->freq = 0;
  138. else if (opts->freq) {
  139. opts->default_interval = opts->freq;
  140. } else {
  141. pr_err("frequency and count are zero, aborting\n");
  142. return -1;
  143. }
  144. if (get_max_rate(&max_rate))
  145. return 0;
  146. /*
  147. * User specified frequency is over current maximum.
  148. */
  149. if (user_freq && (max_rate < opts->freq)) {
  150. pr_err("Maximum frequency rate (%u) reached.\n"
  151. "Please use -F freq option with lower value or consider\n"
  152. "tweaking /proc/sys/kernel/perf_event_max_sample_rate.\n",
  153. max_rate);
  154. return -1;
  155. }
  156. /*
  157. * Default frequency is over current maximum.
  158. */
  159. if (max_rate < opts->freq) {
  160. pr_warning("Lowering default frequency rate to %u.\n"
  161. "Please consider tweaking "
  162. "/proc/sys/kernel/perf_event_max_sample_rate.\n",
  163. max_rate);
  164. opts->freq = max_rate;
  165. }
  166. return 0;
  167. }
  168. int record_opts__config(struct record_opts *opts)
  169. {
  170. return record_opts__config_freq(opts);
  171. }
  172. bool perf_evlist__can_select_event(struct perf_evlist *evlist, const char *str)
  173. {
  174. struct perf_evlist *temp_evlist;
  175. struct perf_evsel *evsel;
  176. int err, fd, cpu;
  177. bool ret = false;
  178. pid_t pid = -1;
  179. temp_evlist = perf_evlist__new();
  180. if (!temp_evlist)
  181. return false;
  182. err = parse_events(temp_evlist, str, NULL);
  183. if (err)
  184. goto out_delete;
  185. evsel = perf_evlist__last(temp_evlist);
  186. if (!evlist || cpu_map__empty(evlist->cpus)) {
  187. struct cpu_map *cpus = cpu_map__new(NULL);
  188. cpu = cpus ? cpus->map[0] : 0;
  189. cpu_map__put(cpus);
  190. } else {
  191. cpu = evlist->cpus->map[0];
  192. }
  193. while (1) {
  194. fd = sys_perf_event_open(&evsel->attr, pid, cpu, -1,
  195. perf_event_open_cloexec_flag());
  196. if (fd < 0) {
  197. if (pid == -1 && errno == EACCES) {
  198. pid = 0;
  199. continue;
  200. }
  201. goto out_delete;
  202. }
  203. break;
  204. }
  205. close(fd);
  206. ret = true;
  207. out_delete:
  208. perf_evlist__delete(temp_evlist);
  209. return ret;
  210. }