map_perf_ring.c 6.6 KB

123456789101112131415161718192021222324252627282930313233343536373839404142434445464748495051525354555657585960616263646566676869707172737475767778798081828384858687888990919293949596979899100101102103104105106107108109110111112113114115116117118119120121122123124125126127128129130131132133134135136137138139140141142143144145146147148149150151152153154155156157158159160161162163164165166167168169170171172173174175176177178179180181182183184185186187188189190191192193194195196197198199200201202203204205206207208209210211212213214215216217218219220221222223224225226227228229230231232233234235236237238239240241242243244245246247248249250251252253254255256257258259260261262263264265266267268269270271272273274275276277278279280281282283284285286287288289290291292293294295296297298299300301302303304305306307308309310311
  1. // SPDX-License-Identifier: GPL-2.0-only
  2. /* Copyright (C) 2018 Netronome Systems, Inc. */
  3. /* This program is free software; you can redistribute it and/or
  4. * modify it under the terms of version 2 of the GNU General Public
  5. * License as published by the Free Software Foundation.
  6. */
  7. #include <errno.h>
  8. #include <fcntl.h>
  9. #include <libbpf.h>
  10. #include <poll.h>
  11. #include <signal.h>
  12. #include <stdbool.h>
  13. #include <stdio.h>
  14. #include <stdlib.h>
  15. #include <string.h>
  16. #include <time.h>
  17. #include <unistd.h>
  18. #include <linux/bpf.h>
  19. #include <linux/perf_event.h>
  20. #include <sys/ioctl.h>
  21. #include <sys/mman.h>
  22. #include <sys/syscall.h>
  23. #include <bpf.h>
  24. #include <perf-sys.h>
  25. #include "main.h"
  26. #define MMAP_PAGE_CNT 16
  27. static bool stop;
  28. struct event_ring_info {
  29. int fd;
  30. int key;
  31. unsigned int cpu;
  32. void *mem;
  33. };
  34. struct perf_event_sample {
  35. struct perf_event_header header;
  36. u64 time;
  37. __u32 size;
  38. unsigned char data[];
  39. };
  40. static void int_exit(int signo)
  41. {
  42. fprintf(stderr, "Stopping...\n");
  43. stop = true;
  44. }
  45. static enum bpf_perf_event_ret
  46. print_bpf_output(struct perf_event_header *event, void *private_data)
  47. {
  48. struct perf_event_sample *e = container_of(event, struct perf_event_sample,
  49. header);
  50. struct event_ring_info *ring = private_data;
  51. struct {
  52. struct perf_event_header header;
  53. __u64 id;
  54. __u64 lost;
  55. } *lost = (typeof(lost))event;
  56. if (json_output) {
  57. jsonw_start_object(json_wtr);
  58. jsonw_name(json_wtr, "type");
  59. jsonw_uint(json_wtr, e->header.type);
  60. jsonw_name(json_wtr, "cpu");
  61. jsonw_uint(json_wtr, ring->cpu);
  62. jsonw_name(json_wtr, "index");
  63. jsonw_uint(json_wtr, ring->key);
  64. if (e->header.type == PERF_RECORD_SAMPLE) {
  65. jsonw_name(json_wtr, "timestamp");
  66. jsonw_uint(json_wtr, e->time);
  67. jsonw_name(json_wtr, "data");
  68. print_data_json(e->data, e->size);
  69. } else if (e->header.type == PERF_RECORD_LOST) {
  70. jsonw_name(json_wtr, "lost");
  71. jsonw_start_object(json_wtr);
  72. jsonw_name(json_wtr, "id");
  73. jsonw_uint(json_wtr, lost->id);
  74. jsonw_name(json_wtr, "count");
  75. jsonw_uint(json_wtr, lost->lost);
  76. jsonw_end_object(json_wtr);
  77. }
  78. jsonw_end_object(json_wtr);
  79. } else {
  80. if (e->header.type == PERF_RECORD_SAMPLE) {
  81. printf("== @%lld.%09lld CPU: %d index: %d =====\n",
  82. e->time / 1000000000ULL, e->time % 1000000000ULL,
  83. ring->cpu, ring->key);
  84. fprint_hex(stdout, e->data, e->size, " ");
  85. printf("\n");
  86. } else if (e->header.type == PERF_RECORD_LOST) {
  87. printf("lost %lld events\n", lost->lost);
  88. } else {
  89. printf("unknown event type=%d size=%d\n",
  90. e->header.type, e->header.size);
  91. }
  92. }
  93. return LIBBPF_PERF_EVENT_CONT;
  94. }
  95. static void
  96. perf_event_read(struct event_ring_info *ring, void **buf, size_t *buf_len)
  97. {
  98. enum bpf_perf_event_ret ret;
  99. ret = bpf_perf_event_read_simple(ring->mem,
  100. MMAP_PAGE_CNT * get_page_size(),
  101. get_page_size(), buf, buf_len,
  102. print_bpf_output, ring);
  103. if (ret != LIBBPF_PERF_EVENT_CONT) {
  104. fprintf(stderr, "perf read loop failed with %d\n", ret);
  105. stop = true;
  106. }
  107. }
  108. static int perf_mmap_size(void)
  109. {
  110. return get_page_size() * (MMAP_PAGE_CNT + 1);
  111. }
  112. static void *perf_event_mmap(int fd)
  113. {
  114. int mmap_size = perf_mmap_size();
  115. void *base;
  116. base = mmap(NULL, mmap_size, PROT_READ | PROT_WRITE, MAP_SHARED, fd, 0);
  117. if (base == MAP_FAILED) {
  118. p_err("event mmap failed: %s\n", strerror(errno));
  119. return NULL;
  120. }
  121. return base;
  122. }
  123. static void perf_event_unmap(void *mem)
  124. {
  125. if (munmap(mem, perf_mmap_size()))
  126. fprintf(stderr, "Can't unmap ring memory!\n");
  127. }
  128. static int bpf_perf_event_open(int map_fd, int key, int cpu)
  129. {
  130. struct perf_event_attr attr = {
  131. .sample_type = PERF_SAMPLE_RAW | PERF_SAMPLE_TIME,
  132. .type = PERF_TYPE_SOFTWARE,
  133. .config = PERF_COUNT_SW_BPF_OUTPUT,
  134. };
  135. int pmu_fd;
  136. pmu_fd = sys_perf_event_open(&attr, -1, cpu, -1, 0);
  137. if (pmu_fd < 0) {
  138. p_err("failed to open perf event %d for CPU %d", key, cpu);
  139. return -1;
  140. }
  141. if (bpf_map_update_elem(map_fd, &key, &pmu_fd, BPF_ANY)) {
  142. p_err("failed to update map for event %d for CPU %d", key, cpu);
  143. goto err_close;
  144. }
  145. if (ioctl(pmu_fd, PERF_EVENT_IOC_ENABLE, 0)) {
  146. p_err("failed to enable event %d for CPU %d", key, cpu);
  147. goto err_close;
  148. }
  149. return pmu_fd;
  150. err_close:
  151. close(pmu_fd);
  152. return -1;
  153. }
  154. int do_event_pipe(int argc, char **argv)
  155. {
  156. int i, nfds, map_fd, index = -1, cpu = -1;
  157. struct bpf_map_info map_info = {};
  158. struct event_ring_info *rings;
  159. size_t tmp_buf_sz = 0;
  160. void *tmp_buf = NULL;
  161. struct pollfd *pfds;
  162. __u32 map_info_len;
  163. bool do_all = true;
  164. map_info_len = sizeof(map_info);
  165. map_fd = map_parse_fd_and_info(&argc, &argv, &map_info, &map_info_len);
  166. if (map_fd < 0)
  167. return -1;
  168. if (map_info.type != BPF_MAP_TYPE_PERF_EVENT_ARRAY) {
  169. p_err("map is not a perf event array");
  170. goto err_close_map;
  171. }
  172. while (argc) {
  173. if (argc < 2) {
  174. BAD_ARG();
  175. goto err_close_map;
  176. }
  177. if (is_prefix(*argv, "cpu")) {
  178. char *endptr;
  179. NEXT_ARG();
  180. cpu = strtoul(*argv, &endptr, 0);
  181. if (*endptr) {
  182. p_err("can't parse %s as CPU ID", **argv);
  183. goto err_close_map;
  184. }
  185. NEXT_ARG();
  186. } else if (is_prefix(*argv, "index")) {
  187. char *endptr;
  188. NEXT_ARG();
  189. index = strtoul(*argv, &endptr, 0);
  190. if (*endptr) {
  191. p_err("can't parse %s as index", **argv);
  192. goto err_close_map;
  193. }
  194. NEXT_ARG();
  195. } else {
  196. BAD_ARG();
  197. goto err_close_map;
  198. }
  199. do_all = false;
  200. }
  201. if (!do_all) {
  202. if (index == -1 || cpu == -1) {
  203. p_err("cpu and index must be specified together");
  204. goto err_close_map;
  205. }
  206. nfds = 1;
  207. } else {
  208. nfds = min(get_possible_cpus(), map_info.max_entries);
  209. cpu = 0;
  210. index = 0;
  211. }
  212. rings = calloc(nfds, sizeof(rings[0]));
  213. if (!rings)
  214. goto err_close_map;
  215. pfds = calloc(nfds, sizeof(pfds[0]));
  216. if (!pfds)
  217. goto err_free_rings;
  218. for (i = 0; i < nfds; i++) {
  219. rings[i].cpu = cpu + i;
  220. rings[i].key = index + i;
  221. rings[i].fd = bpf_perf_event_open(map_fd, rings[i].key,
  222. rings[i].cpu);
  223. if (rings[i].fd < 0)
  224. goto err_close_fds_prev;
  225. rings[i].mem = perf_event_mmap(rings[i].fd);
  226. if (!rings[i].mem)
  227. goto err_close_fds_current;
  228. pfds[i].fd = rings[i].fd;
  229. pfds[i].events = POLLIN;
  230. }
  231. signal(SIGINT, int_exit);
  232. signal(SIGHUP, int_exit);
  233. signal(SIGTERM, int_exit);
  234. if (json_output)
  235. jsonw_start_array(json_wtr);
  236. while (!stop) {
  237. poll(pfds, nfds, 200);
  238. for (i = 0; i < nfds; i++)
  239. perf_event_read(&rings[i], &tmp_buf, &tmp_buf_sz);
  240. }
  241. free(tmp_buf);
  242. if (json_output)
  243. jsonw_end_array(json_wtr);
  244. for (i = 0; i < nfds; i++) {
  245. perf_event_unmap(rings[i].mem);
  246. close(rings[i].fd);
  247. }
  248. free(pfds);
  249. free(rings);
  250. close(map_fd);
  251. return 0;
  252. err_close_fds_prev:
  253. while (i--) {
  254. perf_event_unmap(rings[i].mem);
  255. err_close_fds_current:
  256. close(rings[i].fd);
  257. }
  258. free(pfds);
  259. err_free_rings:
  260. free(rings);
  261. err_close_map:
  262. close(map_fd);
  263. return -1;
  264. }