hv-24x7.c 32 KB

12345678910111213141516171819202122232425262728293031323334353637383940414243444546474849505152535455565758596061626364656667686970717273747576777879808182838485868788899091929394959697989910010110210310410510610710810911011111211311411511611711811912012112212312412512612712812913013113213313413513613713813914014114214314414514614714814915015115215315415515615715815916016116216316416516616716816917017117217317417517617717817918018118218318418518618718818919019119219319419519619719819920020120220320420520620720820921021121221321421521621721821922022122222322422522622722822923023123223323423523623723823924024124224324424524624724824925025125225325425525625725825926026126226326426526626726826927027127227327427527627727827928028128228328428528628728828929029129229329429529629729829930030130230330430530630730830931031131231331431531631731831932032132232332432532632732832933033133233333433533633733833934034134234334434534634734834935035135235335435535635735835936036136236336436536636736836937037137237337437537637737837938038138238338438538638738838939039139239339439539639739839940040140240340440540640740840941041141241341441541641741841942042142242342442542642742842943043143243343443543643743843944044144244344444544644744844945045145245345445545645745845946046146246346446546646746846947047147247347447547647747847948048148248348448548648748848949049149249349449549649749849950050150250350450550650750850951051151251351451551651751851952052152252352452552652752852953053153253353453553653753853954054154254354454554654754854955055155255355455555655755855956056156256356456556656756856957057157257357457557657757857958058158258358458558658758858959059159259359459559659759859960060160260360460560660760860961061161261361461561661761861962062162262362462562662762862963063163263363463563663763863964064164264364464564664764864965065165265365465565665765865966066166266366466566666766866967067167267367467567667767867968068168268368468568668768868969069169269369469569669769869970070170270370470570670770870971071171271371471571671771871972072172272372472572672772872973073173273373473573673773873974074174274374474574674774874975075175275375475575675775875976076176276376476576676776876977077177277377477577677777877978078178278378478578678778878979079179279379479579679779879980080180280380480580680780880981081181281381481581681781881982082182282382482582682782882983083183283383483583683783883984084184284384484584684784884985085185285385485585685785885986086186286386486586686786886987087187287387487587687787887988088188288388488588688788888989089189289389489589689789889990090190290390490590690790890991091191291391491591691791891992092192292392492592692792892993093193293393493593693793893994094194294394494594694794894995095195295395495595695795895996096196296396496596696796896997097197297397497597697797897998098198298398498598698798898999099199299399499599699799899910001001100210031004100510061007100810091010101110121013101410151016101710181019102010211022102310241025102610271028102910301031103210331034103510361037103810391040104110421043104410451046104710481049105010511052105310541055105610571058105910601061106210631064106510661067106810691070107110721073107410751076107710781079108010811082108310841085108610871088108910901091109210931094109510961097109810991100110111021103110411051106110711081109111011111112111311141115111611171118111911201121112211231124112511261127112811291130113111321133113411351136113711381139114011411142114311441145114611471148114911501151115211531154115511561157115811591160116111621163116411651166116711681169117011711172117311741175117611771178117911801181118211831184118511861187118811891190119111921193119411951196119711981199120012011202120312041205120612071208120912101211121212131214121512161217121812191220122112221223122412251226122712281229123012311232123312341235123612371238123912401241124212431244124512461247124812491250125112521253125412551256125712581259126012611262126312641265126612671268126912701271127212731274127512761277127812791280128112821283128412851286128712881289129012911292129312941295129612971298129913001301130213031304130513061307130813091310
  1. /*
  2. * Hypervisor supplied "24x7" performance counter support
  3. *
  4. * Author: Cody P Schafer <cody@linux.vnet.ibm.com>
  5. * Copyright 2014 IBM Corporation.
  6. *
  7. * This program is free software; you can redistribute it and/or
  8. * modify it under the terms of the GNU General Public License
  9. * as published by the Free Software Foundation; either version
  10. * 2 of the License, or (at your option) any later version.
  11. */
  12. #define pr_fmt(fmt) "hv-24x7: " fmt
  13. #include <linux/perf_event.h>
  14. #include <linux/rbtree.h>
  15. #include <linux/module.h>
  16. #include <linux/slab.h>
  17. #include <linux/vmalloc.h>
  18. #include <asm/firmware.h>
  19. #include <asm/hvcall.h>
  20. #include <asm/io.h>
  21. #include <linux/byteorder/generic.h>
  22. #include "hv-24x7.h"
  23. #include "hv-24x7-catalog.h"
  24. #include "hv-common.h"
  25. static const char *event_domain_suffix(unsigned domain)
  26. {
  27. switch (domain) {
  28. #define DOMAIN(n, v, x, c) \
  29. case HV_PERF_DOMAIN_##n: \
  30. return "__" #n;
  31. #include "hv-24x7-domains.h"
  32. #undef DOMAIN
  33. default:
  34. WARN(1, "unknown domain %d\n", domain);
  35. return "__UNKNOWN_DOMAIN_SUFFIX";
  36. }
  37. }
  38. static bool domain_is_valid(unsigned domain)
  39. {
  40. switch (domain) {
  41. #define DOMAIN(n, v, x, c) \
  42. case HV_PERF_DOMAIN_##n: \
  43. /* fall through */
  44. #include "hv-24x7-domains.h"
  45. #undef DOMAIN
  46. return true;
  47. default:
  48. return false;
  49. }
  50. }
  51. static bool is_physical_domain(unsigned domain)
  52. {
  53. switch (domain) {
  54. #define DOMAIN(n, v, x, c) \
  55. case HV_PERF_DOMAIN_##n: \
  56. return c;
  57. #include "hv-24x7-domains.h"
  58. #undef DOMAIN
  59. default:
  60. return false;
  61. }
  62. }
  63. static bool catalog_entry_domain_is_valid(unsigned domain)
  64. {
  65. return is_physical_domain(domain);
  66. }
  67. /*
  68. * TODO: Merging events:
  69. * - Think of the hcall as an interface to a 4d array of counters:
  70. * - x = domains
  71. * - y = indexes in the domain (core, chip, vcpu, node, etc)
  72. * - z = offset into the counter space
  73. * - w = lpars (guest vms, "logical partitions")
  74. * - A single request is: x,y,y_last,z,z_last,w,w_last
  75. * - this means we can retrieve a rectangle of counters in y,z for a single x.
  76. *
  77. * - Things to consider (ignoring w):
  78. * - input cost_per_request = 16
  79. * - output cost_per_result(ys,zs) = 8 + 8 * ys + ys * zs
  80. * - limited number of requests per hcall (must fit into 4K bytes)
  81. * - 4k = 16 [buffer header] - 16 [request size] * request_count
  82. * - 255 requests per hcall
  83. * - sometimes it will be more efficient to read extra data and discard
  84. */
  85. /*
  86. * Example usage:
  87. * perf stat -e 'hv_24x7/domain=2,offset=8,vcpu=0,lpar=0xffffffff/'
  88. */
  89. /* u3 0-6, one of HV_24X7_PERF_DOMAIN */
  90. EVENT_DEFINE_RANGE_FORMAT(domain, config, 0, 3);
  91. /* u16 */
  92. EVENT_DEFINE_RANGE_FORMAT(core, config, 16, 31);
  93. EVENT_DEFINE_RANGE_FORMAT(vcpu, config, 16, 31);
  94. /* u32, see "data_offset" */
  95. EVENT_DEFINE_RANGE_FORMAT(offset, config, 32, 63);
  96. /* u16 */
  97. EVENT_DEFINE_RANGE_FORMAT(lpar, config1, 0, 15);
  98. EVENT_DEFINE_RANGE(reserved1, config, 4, 15);
  99. EVENT_DEFINE_RANGE(reserved2, config1, 16, 63);
  100. EVENT_DEFINE_RANGE(reserved3, config2, 0, 63);
  101. static struct attribute *format_attrs[] = {
  102. &format_attr_domain.attr,
  103. &format_attr_offset.attr,
  104. &format_attr_core.attr,
  105. &format_attr_vcpu.attr,
  106. &format_attr_lpar.attr,
  107. NULL,
  108. };
  109. static struct attribute_group format_group = {
  110. .name = "format",
  111. .attrs = format_attrs,
  112. };
  113. static struct attribute_group event_group = {
  114. .name = "events",
  115. /* .attrs is set in init */
  116. };
  117. static struct attribute_group event_desc_group = {
  118. .name = "event_descs",
  119. /* .attrs is set in init */
  120. };
  121. static struct attribute_group event_long_desc_group = {
  122. .name = "event_long_descs",
  123. /* .attrs is set in init */
  124. };
  125. static struct kmem_cache *hv_page_cache;
  126. /*
  127. * request_buffer and result_buffer are not required to be 4k aligned,
  128. * but are not allowed to cross any 4k boundary. Aligning them to 4k is
  129. * the simplest way to ensure that.
  130. */
  131. #define H24x7_DATA_BUFFER_SIZE 4096
  132. DEFINE_PER_CPU(char, hv_24x7_reqb[H24x7_DATA_BUFFER_SIZE]) __aligned(4096);
  133. DEFINE_PER_CPU(char, hv_24x7_resb[H24x7_DATA_BUFFER_SIZE]) __aligned(4096);
  134. static char *event_name(struct hv_24x7_event_data *ev, int *len)
  135. {
  136. *len = be16_to_cpu(ev->event_name_len) - 2;
  137. return (char *)ev->remainder;
  138. }
  139. static char *event_desc(struct hv_24x7_event_data *ev, int *len)
  140. {
  141. unsigned nl = be16_to_cpu(ev->event_name_len);
  142. __be16 *desc_len = (__be16 *)(ev->remainder + nl - 2);
  143. *len = be16_to_cpu(*desc_len) - 2;
  144. return (char *)ev->remainder + nl;
  145. }
  146. static char *event_long_desc(struct hv_24x7_event_data *ev, int *len)
  147. {
  148. unsigned nl = be16_to_cpu(ev->event_name_len);
  149. __be16 *desc_len_ = (__be16 *)(ev->remainder + nl - 2);
  150. unsigned desc_len = be16_to_cpu(*desc_len_);
  151. __be16 *long_desc_len = (__be16 *)(ev->remainder + nl + desc_len - 2);
  152. *len = be16_to_cpu(*long_desc_len) - 2;
  153. return (char *)ev->remainder + nl + desc_len;
  154. }
  155. static bool event_fixed_portion_is_within(struct hv_24x7_event_data *ev,
  156. void *end)
  157. {
  158. void *start = ev;
  159. return (start + offsetof(struct hv_24x7_event_data, remainder)) < end;
  160. }
  161. /*
  162. * Things we don't check:
  163. * - padding for desc, name, and long/detailed desc is required to be '\0'
  164. * bytes.
  165. *
  166. * Return NULL if we pass end,
  167. * Otherwise return the address of the byte just following the event.
  168. */
  169. static void *event_end(struct hv_24x7_event_data *ev, void *end)
  170. {
  171. void *start = ev;
  172. __be16 *dl_, *ldl_;
  173. unsigned dl, ldl;
  174. unsigned nl = be16_to_cpu(ev->event_name_len);
  175. if (nl < 2) {
  176. pr_debug("%s: name length too short: %d", __func__, nl);
  177. return NULL;
  178. }
  179. if (start + nl > end) {
  180. pr_debug("%s: start=%p + nl=%u > end=%p",
  181. __func__, start, nl, end);
  182. return NULL;
  183. }
  184. dl_ = (__be16 *)(ev->remainder + nl - 2);
  185. if (!IS_ALIGNED((uintptr_t)dl_, 2))
  186. pr_warn("desc len not aligned %p", dl_);
  187. dl = be16_to_cpu(*dl_);
  188. if (dl < 2) {
  189. pr_debug("%s: desc len too short: %d", __func__, dl);
  190. return NULL;
  191. }
  192. if (start + nl + dl > end) {
  193. pr_debug("%s: (start=%p + nl=%u + dl=%u)=%p > end=%p",
  194. __func__, start, nl, dl, start + nl + dl, end);
  195. return NULL;
  196. }
  197. ldl_ = (__be16 *)(ev->remainder + nl + dl - 2);
  198. if (!IS_ALIGNED((uintptr_t)ldl_, 2))
  199. pr_warn("long desc len not aligned %p", ldl_);
  200. ldl = be16_to_cpu(*ldl_);
  201. if (ldl < 2) {
  202. pr_debug("%s: long desc len too short (ldl=%u)",
  203. __func__, ldl);
  204. return NULL;
  205. }
  206. if (start + nl + dl + ldl > end) {
  207. pr_debug("%s: start=%p + nl=%u + dl=%u + ldl=%u > end=%p",
  208. __func__, start, nl, dl, ldl, end);
  209. return NULL;
  210. }
  211. return start + nl + dl + ldl;
  212. }
  213. static unsigned long h_get_24x7_catalog_page_(unsigned long phys_4096,
  214. unsigned long version,
  215. unsigned long index)
  216. {
  217. pr_devel("h_get_24x7_catalog_page(0x%lx, %lu, %lu)",
  218. phys_4096, version, index);
  219. WARN_ON(!IS_ALIGNED(phys_4096, 4096));
  220. return plpar_hcall_norets(H_GET_24X7_CATALOG_PAGE,
  221. phys_4096, version, index);
  222. }
  223. static unsigned long h_get_24x7_catalog_page(char page[],
  224. u64 version, u32 index)
  225. {
  226. return h_get_24x7_catalog_page_(virt_to_phys(page),
  227. version, index);
  228. }
  229. static unsigned core_domains[] = {
  230. HV_PERF_DOMAIN_PHYS_CORE,
  231. HV_PERF_DOMAIN_VCPU_HOME_CORE,
  232. HV_PERF_DOMAIN_VCPU_HOME_CHIP,
  233. HV_PERF_DOMAIN_VCPU_HOME_NODE,
  234. HV_PERF_DOMAIN_VCPU_REMOTE_NODE,
  235. };
  236. /* chip event data always yeilds a single event, core yeilds multiple */
  237. #define MAX_EVENTS_PER_EVENT_DATA ARRAY_SIZE(core_domains)
  238. static char *event_fmt(struct hv_24x7_event_data *event, unsigned domain)
  239. {
  240. const char *sindex;
  241. const char *lpar;
  242. if (is_physical_domain(domain)) {
  243. lpar = "0x0";
  244. sindex = "core";
  245. } else {
  246. lpar = "?";
  247. sindex = "vcpu";
  248. }
  249. return kasprintf(GFP_KERNEL,
  250. "domain=0x%x,offset=0x%x,%s=?,lpar=%s",
  251. domain,
  252. be16_to_cpu(event->event_counter_offs) +
  253. be16_to_cpu(event->event_group_record_offs),
  254. sindex,
  255. lpar);
  256. }
  257. /* Avoid trusting fw to NUL terminate strings */
  258. static char *memdup_to_str(char *maybe_str, int max_len, gfp_t gfp)
  259. {
  260. return kasprintf(gfp, "%.*s", max_len, maybe_str);
  261. }
  262. static ssize_t device_show_string(struct device *dev,
  263. struct device_attribute *attr, char *buf)
  264. {
  265. struct dev_ext_attribute *d;
  266. d = container_of(attr, struct dev_ext_attribute, attr);
  267. return sprintf(buf, "%s\n", (char *)d->var);
  268. }
  269. static struct attribute *device_str_attr_create_(char *name, char *str)
  270. {
  271. struct dev_ext_attribute *attr = kzalloc(sizeof(*attr), GFP_KERNEL);
  272. if (!attr)
  273. return NULL;
  274. attr->var = str;
  275. attr->attr.attr.name = name;
  276. attr->attr.attr.mode = 0444;
  277. attr->attr.show = device_show_string;
  278. return &attr->attr.attr;
  279. }
  280. static struct attribute *device_str_attr_create(char *name, int name_max,
  281. int name_nonce,
  282. char *str, size_t str_max)
  283. {
  284. char *n;
  285. char *s = memdup_to_str(str, str_max, GFP_KERNEL);
  286. struct attribute *a;
  287. if (!s)
  288. return NULL;
  289. if (!name_nonce)
  290. n = kasprintf(GFP_KERNEL, "%.*s", name_max, name);
  291. else
  292. n = kasprintf(GFP_KERNEL, "%.*s__%d", name_max, name,
  293. name_nonce);
  294. if (!n)
  295. goto out_s;
  296. a = device_str_attr_create_(n, s);
  297. if (!a)
  298. goto out_n;
  299. return a;
  300. out_n:
  301. kfree(n);
  302. out_s:
  303. kfree(s);
  304. return NULL;
  305. }
  306. static void device_str_attr_destroy(struct attribute *attr)
  307. {
  308. struct dev_ext_attribute *d;
  309. d = container_of(attr, struct dev_ext_attribute, attr.attr);
  310. kfree(d->var);
  311. kfree(d->attr.attr.name);
  312. kfree(d);
  313. }
  314. static struct attribute *event_to_attr(unsigned ix,
  315. struct hv_24x7_event_data *event,
  316. unsigned domain,
  317. int nonce)
  318. {
  319. int event_name_len;
  320. char *ev_name, *a_ev_name, *val;
  321. const char *ev_suffix;
  322. struct attribute *attr;
  323. if (!domain_is_valid(domain)) {
  324. pr_warn("catalog event %u has invalid domain %u\n",
  325. ix, domain);
  326. return NULL;
  327. }
  328. val = event_fmt(event, domain);
  329. if (!val)
  330. return NULL;
  331. ev_suffix = event_domain_suffix(domain);
  332. ev_name = event_name(event, &event_name_len);
  333. if (!nonce)
  334. a_ev_name = kasprintf(GFP_KERNEL, "%.*s%s",
  335. (int)event_name_len, ev_name, ev_suffix);
  336. else
  337. a_ev_name = kasprintf(GFP_KERNEL, "%.*s%s__%d",
  338. (int)event_name_len, ev_name, ev_suffix, nonce);
  339. if (!a_ev_name)
  340. goto out_val;
  341. attr = device_str_attr_create_(a_ev_name, val);
  342. if (!attr)
  343. goto out_name;
  344. return attr;
  345. out_name:
  346. kfree(a_ev_name);
  347. out_val:
  348. kfree(val);
  349. return NULL;
  350. }
  351. static struct attribute *event_to_desc_attr(struct hv_24x7_event_data *event,
  352. int nonce)
  353. {
  354. int nl, dl;
  355. char *name = event_name(event, &nl);
  356. char *desc = event_desc(event, &dl);
  357. /* If there isn't a description, don't create the sysfs file */
  358. if (!dl)
  359. return NULL;
  360. return device_str_attr_create(name, nl, nonce, desc, dl);
  361. }
  362. static struct attribute *
  363. event_to_long_desc_attr(struct hv_24x7_event_data *event, int nonce)
  364. {
  365. int nl, dl;
  366. char *name = event_name(event, &nl);
  367. char *desc = event_long_desc(event, &dl);
  368. /* If there isn't a description, don't create the sysfs file */
  369. if (!dl)
  370. return NULL;
  371. return device_str_attr_create(name, nl, nonce, desc, dl);
  372. }
  373. static ssize_t event_data_to_attrs(unsigned ix, struct attribute **attrs,
  374. struct hv_24x7_event_data *event, int nonce)
  375. {
  376. unsigned i;
  377. switch (event->domain) {
  378. case HV_PERF_DOMAIN_PHYS_CHIP:
  379. *attrs = event_to_attr(ix, event, event->domain, nonce);
  380. return 1;
  381. case HV_PERF_DOMAIN_PHYS_CORE:
  382. for (i = 0; i < ARRAY_SIZE(core_domains); i++) {
  383. attrs[i] = event_to_attr(ix, event, core_domains[i],
  384. nonce);
  385. if (!attrs[i]) {
  386. pr_warn("catalog event %u: individual attr %u "
  387. "creation failure\n", ix, i);
  388. for (; i; i--)
  389. device_str_attr_destroy(attrs[i - 1]);
  390. return -1;
  391. }
  392. }
  393. return i;
  394. default:
  395. pr_warn("catalog event %u: domain %u is not allowed in the "
  396. "catalog\n", ix, event->domain);
  397. return -1;
  398. }
  399. }
  400. static size_t event_to_attr_ct(struct hv_24x7_event_data *event)
  401. {
  402. switch (event->domain) {
  403. case HV_PERF_DOMAIN_PHYS_CHIP:
  404. return 1;
  405. case HV_PERF_DOMAIN_PHYS_CORE:
  406. return ARRAY_SIZE(core_domains);
  407. default:
  408. return 0;
  409. }
  410. }
  411. static unsigned long vmalloc_to_phys(void *v)
  412. {
  413. struct page *p = vmalloc_to_page(v);
  414. BUG_ON(!p);
  415. return page_to_phys(p) + offset_in_page(v);
  416. }
  417. /* */
  418. struct event_uniq {
  419. struct rb_node node;
  420. const char *name;
  421. int nl;
  422. unsigned ct;
  423. unsigned domain;
  424. };
  425. static int memord(const void *d1, size_t s1, const void *d2, size_t s2)
  426. {
  427. if (s1 < s2)
  428. return 1;
  429. if (s2 > s1)
  430. return -1;
  431. return memcmp(d1, d2, s1);
  432. }
  433. static int ev_uniq_ord(const void *v1, size_t s1, unsigned d1, const void *v2,
  434. size_t s2, unsigned d2)
  435. {
  436. int r = memord(v1, s1, v2, s2);
  437. if (r)
  438. return r;
  439. if (d1 > d2)
  440. return 1;
  441. if (d2 > d1)
  442. return -1;
  443. return 0;
  444. }
  445. static int event_uniq_add(struct rb_root *root, const char *name, int nl,
  446. unsigned domain)
  447. {
  448. struct rb_node **new = &(root->rb_node), *parent = NULL;
  449. struct event_uniq *data;
  450. /* Figure out where to put new node */
  451. while (*new) {
  452. struct event_uniq *it;
  453. int result;
  454. it = container_of(*new, struct event_uniq, node);
  455. result = ev_uniq_ord(name, nl, domain, it->name, it->nl,
  456. it->domain);
  457. parent = *new;
  458. if (result < 0)
  459. new = &((*new)->rb_left);
  460. else if (result > 0)
  461. new = &((*new)->rb_right);
  462. else {
  463. it->ct++;
  464. pr_info("found a duplicate event %.*s, ct=%u\n", nl,
  465. name, it->ct);
  466. return it->ct;
  467. }
  468. }
  469. data = kmalloc(sizeof(*data), GFP_KERNEL);
  470. if (!data)
  471. return -ENOMEM;
  472. *data = (struct event_uniq) {
  473. .name = name,
  474. .nl = nl,
  475. .ct = 0,
  476. .domain = domain,
  477. };
  478. /* Add new node and rebalance tree. */
  479. rb_link_node(&data->node, parent, new);
  480. rb_insert_color(&data->node, root);
  481. /* data->ct */
  482. return 0;
  483. }
  484. static void event_uniq_destroy(struct rb_root *root)
  485. {
  486. /*
  487. * the strings we point to are in the giant block of memory filled by
  488. * the catalog, and are freed separately.
  489. */
  490. struct event_uniq *pos, *n;
  491. rbtree_postorder_for_each_entry_safe(pos, n, root, node)
  492. kfree(pos);
  493. }
  494. /*
  495. * ensure the event structure's sizes are self consistent and don't cause us to
  496. * read outside of the event
  497. *
  498. * On success, return the event length in bytes.
  499. * Otherwise, return -1 (and print as appropriate).
  500. */
  501. static ssize_t catalog_event_len_validate(struct hv_24x7_event_data *event,
  502. size_t event_idx,
  503. size_t event_data_bytes,
  504. size_t event_entry_count,
  505. size_t offset, void *end)
  506. {
  507. ssize_t ev_len;
  508. void *ev_end, *calc_ev_end;
  509. if (offset >= event_data_bytes)
  510. return -1;
  511. if (event_idx >= event_entry_count) {
  512. pr_devel("catalog event data has %zu bytes of padding after last event\n",
  513. event_data_bytes - offset);
  514. return -1;
  515. }
  516. if (!event_fixed_portion_is_within(event, end)) {
  517. pr_warn("event %zu fixed portion is not within range\n",
  518. event_idx);
  519. return -1;
  520. }
  521. ev_len = be16_to_cpu(event->length);
  522. if (ev_len % 16)
  523. pr_info("event %zu has length %zu not divisible by 16: event=%pK\n",
  524. event_idx, ev_len, event);
  525. ev_end = (__u8 *)event + ev_len;
  526. if (ev_end > end) {
  527. pr_warn("event %zu has .length=%zu, ends after buffer end: ev_end=%pK > end=%pK, offset=%zu\n",
  528. event_idx, ev_len, ev_end, end,
  529. offset);
  530. return -1;
  531. }
  532. calc_ev_end = event_end(event, end);
  533. if (!calc_ev_end) {
  534. pr_warn("event %zu has a calculated length which exceeds buffer length %zu: event=%pK end=%pK, offset=%zu\n",
  535. event_idx, event_data_bytes, event, end,
  536. offset);
  537. return -1;
  538. }
  539. if (calc_ev_end > ev_end) {
  540. pr_warn("event %zu exceeds it's own length: event=%pK, end=%pK, offset=%zu, calc_ev_end=%pK\n",
  541. event_idx, event, ev_end, offset, calc_ev_end);
  542. return -1;
  543. }
  544. return ev_len;
  545. }
  546. #define MAX_4K (SIZE_MAX / 4096)
  547. static int create_events_from_catalog(struct attribute ***events_,
  548. struct attribute ***event_descs_,
  549. struct attribute ***event_long_descs_)
  550. {
  551. unsigned long hret;
  552. size_t catalog_len, catalog_page_len, event_entry_count,
  553. event_data_len, event_data_offs,
  554. event_data_bytes, junk_events, event_idx, event_attr_ct, i,
  555. attr_max, event_idx_last, desc_ct, long_desc_ct;
  556. ssize_t ct, ev_len;
  557. uint32_t catalog_version_num;
  558. struct attribute **events, **event_descs, **event_long_descs;
  559. struct hv_24x7_catalog_page_0 *page_0 =
  560. kmem_cache_alloc(hv_page_cache, GFP_KERNEL);
  561. void *page = page_0;
  562. void *event_data, *end;
  563. struct hv_24x7_event_data *event;
  564. struct rb_root ev_uniq = RB_ROOT;
  565. int ret = 0;
  566. if (!page) {
  567. ret = -ENOMEM;
  568. goto e_out;
  569. }
  570. hret = h_get_24x7_catalog_page(page, 0, 0);
  571. if (hret) {
  572. ret = -EIO;
  573. goto e_free;
  574. }
  575. catalog_version_num = be64_to_cpu(page_0->version);
  576. catalog_page_len = be32_to_cpu(page_0->length);
  577. if (MAX_4K < catalog_page_len) {
  578. pr_err("invalid page count: %zu\n", catalog_page_len);
  579. ret = -EIO;
  580. goto e_free;
  581. }
  582. catalog_len = catalog_page_len * 4096;
  583. event_entry_count = be16_to_cpu(page_0->event_entry_count);
  584. event_data_offs = be16_to_cpu(page_0->event_data_offs);
  585. event_data_len = be16_to_cpu(page_0->event_data_len);
  586. pr_devel("cv %zu cl %zu eec %zu edo %zu edl %zu\n",
  587. (size_t)catalog_version_num, catalog_len,
  588. event_entry_count, event_data_offs, event_data_len);
  589. if ((MAX_4K < event_data_len)
  590. || (MAX_4K < event_data_offs)
  591. || (MAX_4K - event_data_offs < event_data_len)) {
  592. pr_err("invalid event data offs %zu and/or len %zu\n",
  593. event_data_offs, event_data_len);
  594. ret = -EIO;
  595. goto e_free;
  596. }
  597. if ((event_data_offs + event_data_len) > catalog_page_len) {
  598. pr_err("event data %zu-%zu does not fit inside catalog 0-%zu\n",
  599. event_data_offs,
  600. event_data_offs + event_data_len,
  601. catalog_page_len);
  602. ret = -EIO;
  603. goto e_free;
  604. }
  605. if (SIZE_MAX / MAX_EVENTS_PER_EVENT_DATA - 1 < event_entry_count) {
  606. pr_err("event_entry_count %zu is invalid\n",
  607. event_entry_count);
  608. ret = -EIO;
  609. goto e_free;
  610. }
  611. event_data_bytes = event_data_len * 4096;
  612. /*
  613. * event data can span several pages, events can cross between these
  614. * pages. Use vmalloc to make this easier.
  615. */
  616. event_data = vmalloc(event_data_bytes);
  617. if (!event_data) {
  618. pr_err("could not allocate event data\n");
  619. ret = -ENOMEM;
  620. goto e_free;
  621. }
  622. end = event_data + event_data_bytes;
  623. /*
  624. * using vmalloc_to_phys() like this only works if PAGE_SIZE is
  625. * divisible by 4096
  626. */
  627. BUILD_BUG_ON(PAGE_SIZE % 4096);
  628. for (i = 0; i < event_data_len; i++) {
  629. hret = h_get_24x7_catalog_page_(
  630. vmalloc_to_phys(event_data + i * 4096),
  631. catalog_version_num,
  632. i + event_data_offs);
  633. if (hret) {
  634. pr_err("failed to get event data in page %zu\n",
  635. i + event_data_offs);
  636. ret = -EIO;
  637. goto e_event_data;
  638. }
  639. }
  640. /*
  641. * scan the catalog to determine the number of attributes we need, and
  642. * verify it at the same time.
  643. */
  644. for (junk_events = 0, event = event_data, event_idx = 0, attr_max = 0;
  645. ;
  646. event_idx++, event = (void *)event + ev_len) {
  647. size_t offset = (void *)event - (void *)event_data;
  648. char *name;
  649. int nl;
  650. ev_len = catalog_event_len_validate(event, event_idx,
  651. event_data_bytes,
  652. event_entry_count,
  653. offset, end);
  654. if (ev_len < 0)
  655. break;
  656. name = event_name(event, &nl);
  657. if (event->event_group_record_len == 0) {
  658. pr_devel("invalid event %zu (%.*s): group_record_len == 0, skipping\n",
  659. event_idx, nl, name);
  660. junk_events++;
  661. continue;
  662. }
  663. if (!catalog_entry_domain_is_valid(event->domain)) {
  664. pr_info("event %zu (%.*s) has invalid domain %d\n",
  665. event_idx, nl, name, event->domain);
  666. junk_events++;
  667. continue;
  668. }
  669. attr_max += event_to_attr_ct(event);
  670. }
  671. event_idx_last = event_idx;
  672. if (event_idx_last != event_entry_count)
  673. pr_warn("event buffer ended before listed # of events were parsed (got %zu, wanted %zu, junk %zu)\n",
  674. event_idx_last, event_entry_count, junk_events);
  675. events = kmalloc_array(attr_max + 1, sizeof(*events), GFP_KERNEL);
  676. if (!events) {
  677. ret = -ENOMEM;
  678. goto e_event_data;
  679. }
  680. event_descs = kmalloc_array(event_idx + 1, sizeof(*event_descs),
  681. GFP_KERNEL);
  682. if (!event_descs) {
  683. ret = -ENOMEM;
  684. goto e_event_attrs;
  685. }
  686. event_long_descs = kmalloc_array(event_idx + 1,
  687. sizeof(*event_long_descs), GFP_KERNEL);
  688. if (!event_long_descs) {
  689. ret = -ENOMEM;
  690. goto e_event_descs;
  691. }
  692. /* Iterate over the catalog filling in the attribute vector */
  693. for (junk_events = 0, event_attr_ct = 0, desc_ct = 0, long_desc_ct = 0,
  694. event = event_data, event_idx = 0;
  695. event_idx < event_idx_last;
  696. event_idx++, ev_len = be16_to_cpu(event->length),
  697. event = (void *)event + ev_len) {
  698. char *name;
  699. int nl;
  700. int nonce;
  701. /*
  702. * these are the only "bad" events that are intermixed and that
  703. * we can ignore without issue. make sure to skip them here
  704. */
  705. if (event->event_group_record_len == 0)
  706. continue;
  707. if (!catalog_entry_domain_is_valid(event->domain))
  708. continue;
  709. name = event_name(event, &nl);
  710. nonce = event_uniq_add(&ev_uniq, name, nl, event->domain);
  711. ct = event_data_to_attrs(event_idx, events + event_attr_ct,
  712. event, nonce);
  713. if (ct <= 0) {
  714. pr_warn("event %zu (%.*s) creation failure, skipping\n",
  715. event_idx, nl, name);
  716. junk_events++;
  717. } else {
  718. event_attr_ct += ct;
  719. event_descs[desc_ct] = event_to_desc_attr(event, nonce);
  720. if (event_descs[desc_ct])
  721. desc_ct++;
  722. event_long_descs[long_desc_ct] =
  723. event_to_long_desc_attr(event, nonce);
  724. if (event_long_descs[long_desc_ct])
  725. long_desc_ct++;
  726. }
  727. }
  728. pr_info("read %zu catalog entries, created %zu event attrs (%zu failures), %zu descs\n",
  729. event_idx, event_attr_ct, junk_events, desc_ct);
  730. events[event_attr_ct] = NULL;
  731. event_descs[desc_ct] = NULL;
  732. event_long_descs[long_desc_ct] = NULL;
  733. event_uniq_destroy(&ev_uniq);
  734. vfree(event_data);
  735. kmem_cache_free(hv_page_cache, page);
  736. *events_ = events;
  737. *event_descs_ = event_descs;
  738. *event_long_descs_ = event_long_descs;
  739. return 0;
  740. e_event_descs:
  741. kfree(event_descs);
  742. e_event_attrs:
  743. kfree(events);
  744. e_event_data:
  745. vfree(event_data);
  746. e_free:
  747. kmem_cache_free(hv_page_cache, page);
  748. e_out:
  749. *events_ = NULL;
  750. *event_descs_ = NULL;
  751. *event_long_descs_ = NULL;
  752. return ret;
  753. }
  754. static ssize_t catalog_read(struct file *filp, struct kobject *kobj,
  755. struct bin_attribute *bin_attr, char *buf,
  756. loff_t offset, size_t count)
  757. {
  758. unsigned long hret;
  759. ssize_t ret = 0;
  760. size_t catalog_len = 0, catalog_page_len = 0;
  761. loff_t page_offset = 0;
  762. loff_t offset_in_page;
  763. size_t copy_len;
  764. uint64_t catalog_version_num = 0;
  765. void *page = kmem_cache_alloc(hv_page_cache, GFP_USER);
  766. struct hv_24x7_catalog_page_0 *page_0 = page;
  767. if (!page)
  768. return -ENOMEM;
  769. hret = h_get_24x7_catalog_page(page, 0, 0);
  770. if (hret) {
  771. ret = -EIO;
  772. goto e_free;
  773. }
  774. catalog_version_num = be64_to_cpu(page_0->version);
  775. catalog_page_len = be32_to_cpu(page_0->length);
  776. catalog_len = catalog_page_len * 4096;
  777. page_offset = offset / 4096;
  778. offset_in_page = offset % 4096;
  779. if (page_offset >= catalog_page_len)
  780. goto e_free;
  781. if (page_offset != 0) {
  782. hret = h_get_24x7_catalog_page(page, catalog_version_num,
  783. page_offset);
  784. if (hret) {
  785. ret = -EIO;
  786. goto e_free;
  787. }
  788. }
  789. copy_len = 4096 - offset_in_page;
  790. if (copy_len > count)
  791. copy_len = count;
  792. memcpy(buf, page+offset_in_page, copy_len);
  793. ret = copy_len;
  794. e_free:
  795. if (hret)
  796. pr_err("h_get_24x7_catalog_page(ver=%lld, page=%lld) failed:"
  797. " rc=%ld\n",
  798. catalog_version_num, page_offset, hret);
  799. kmem_cache_free(hv_page_cache, page);
  800. pr_devel("catalog_read: offset=%lld(%lld) count=%zu "
  801. "catalog_len=%zu(%zu) => %zd\n", offset, page_offset,
  802. count, catalog_len, catalog_page_len, ret);
  803. return ret;
  804. }
  805. #define PAGE_0_ATTR(_name, _fmt, _expr) \
  806. static ssize_t _name##_show(struct device *dev, \
  807. struct device_attribute *dev_attr, \
  808. char *buf) \
  809. { \
  810. unsigned long hret; \
  811. ssize_t ret = 0; \
  812. void *page = kmem_cache_alloc(hv_page_cache, GFP_USER); \
  813. struct hv_24x7_catalog_page_0 *page_0 = page; \
  814. if (!page) \
  815. return -ENOMEM; \
  816. hret = h_get_24x7_catalog_page(page, 0, 0); \
  817. if (hret) { \
  818. ret = -EIO; \
  819. goto e_free; \
  820. } \
  821. ret = sprintf(buf, _fmt, _expr); \
  822. e_free: \
  823. kmem_cache_free(hv_page_cache, page); \
  824. return ret; \
  825. } \
  826. static DEVICE_ATTR_RO(_name)
  827. PAGE_0_ATTR(catalog_version, "%lld\n",
  828. (unsigned long long)be64_to_cpu(page_0->version));
  829. PAGE_0_ATTR(catalog_len, "%lld\n",
  830. (unsigned long long)be32_to_cpu(page_0->length) * 4096);
  831. static BIN_ATTR_RO(catalog, 0/* real length varies */);
  832. static struct bin_attribute *if_bin_attrs[] = {
  833. &bin_attr_catalog,
  834. NULL,
  835. };
  836. static struct attribute *if_attrs[] = {
  837. &dev_attr_catalog_len.attr,
  838. &dev_attr_catalog_version.attr,
  839. NULL,
  840. };
  841. static struct attribute_group if_group = {
  842. .name = "interface",
  843. .bin_attrs = if_bin_attrs,
  844. .attrs = if_attrs,
  845. };
  846. static const struct attribute_group *attr_groups[] = {
  847. &format_group,
  848. &event_group,
  849. &event_desc_group,
  850. &event_long_desc_group,
  851. &if_group,
  852. NULL,
  853. };
  854. static void log_24x7_hcall(struct hv_24x7_request_buffer *request_buffer,
  855. struct hv_24x7_data_result_buffer *result_buffer,
  856. unsigned long ret)
  857. {
  858. struct hv_24x7_request *req;
  859. req = &request_buffer->requests[0];
  860. pr_notice_ratelimited("hcall failed: [%d %#x %#x %d] => "
  861. "ret 0x%lx (%ld) detail=0x%x failing ix=%x\n",
  862. req->performance_domain, req->data_offset,
  863. req->starting_ix, req->starting_lpar_ix, ret, ret,
  864. result_buffer->detailed_rc,
  865. result_buffer->failing_request_ix);
  866. }
  867. /*
  868. * Start the process for a new H_GET_24x7_DATA hcall.
  869. */
  870. static void init_24x7_request(struct hv_24x7_request_buffer *request_buffer,
  871. struct hv_24x7_data_result_buffer *result_buffer)
  872. {
  873. memset(request_buffer, 0, 4096);
  874. memset(result_buffer, 0, 4096);
  875. request_buffer->interface_version = HV_24X7_IF_VERSION_CURRENT;
  876. /* memset above set request_buffer->num_requests to 0 */
  877. }
  878. /*
  879. * Commit (i.e perform) the H_GET_24x7_DATA hcall using the data collected
  880. * by 'init_24x7_request()' and 'add_event_to_24x7_request()'.
  881. */
  882. static int make_24x7_request(struct hv_24x7_request_buffer *request_buffer,
  883. struct hv_24x7_data_result_buffer *result_buffer)
  884. {
  885. unsigned long ret;
  886. /*
  887. * NOTE: Due to variable number of array elements in request and
  888. * result buffer(s), sizeof() is not reliable. Use the actual
  889. * allocated buffer size, H24x7_DATA_BUFFER_SIZE.
  890. */
  891. ret = plpar_hcall_norets(H_GET_24X7_DATA,
  892. virt_to_phys(request_buffer), H24x7_DATA_BUFFER_SIZE,
  893. virt_to_phys(result_buffer), H24x7_DATA_BUFFER_SIZE);
  894. if (ret)
  895. log_24x7_hcall(request_buffer, result_buffer, ret);
  896. return ret;
  897. }
  898. /*
  899. * Add the given @event to the next slot in the 24x7 request_buffer.
  900. *
  901. * Note that H_GET_24X7_DATA hcall allows reading several counters'
  902. * values in a single HCALL. We expect the caller to add events to the
  903. * request buffer one by one, make the HCALL and process the results.
  904. */
  905. static int add_event_to_24x7_request(struct perf_event *event,
  906. struct hv_24x7_request_buffer *request_buffer)
  907. {
  908. u16 idx;
  909. int i;
  910. struct hv_24x7_request *req;
  911. if (request_buffer->num_requests > 254) {
  912. pr_devel("Too many requests for 24x7 HCALL %d\n",
  913. request_buffer->num_requests);
  914. return -EINVAL;
  915. }
  916. if (is_physical_domain(event_get_domain(event)))
  917. idx = event_get_core(event);
  918. else
  919. idx = event_get_vcpu(event);
  920. i = request_buffer->num_requests++;
  921. req = &request_buffer->requests[i];
  922. req->performance_domain = event_get_domain(event);
  923. req->data_size = cpu_to_be16(8);
  924. req->data_offset = cpu_to_be32(event_get_offset(event));
  925. req->starting_lpar_ix = cpu_to_be16(event_get_lpar(event)),
  926. req->max_num_lpars = cpu_to_be16(1);
  927. req->starting_ix = cpu_to_be16(idx);
  928. req->max_ix = cpu_to_be16(1);
  929. return 0;
  930. }
  931. static unsigned long single_24x7_request(struct perf_event *event, u64 *count)
  932. {
  933. unsigned long ret;
  934. struct hv_24x7_request_buffer *request_buffer;
  935. struct hv_24x7_data_result_buffer *result_buffer;
  936. struct hv_24x7_result *resb;
  937. BUILD_BUG_ON(sizeof(*request_buffer) > 4096);
  938. BUILD_BUG_ON(sizeof(*result_buffer) > 4096);
  939. request_buffer = (void *)get_cpu_var(hv_24x7_reqb);
  940. result_buffer = (void *)get_cpu_var(hv_24x7_resb);
  941. init_24x7_request(request_buffer, result_buffer);
  942. ret = add_event_to_24x7_request(event, request_buffer);
  943. if (ret)
  944. goto out;
  945. ret = make_24x7_request(request_buffer, result_buffer);
  946. if (ret) {
  947. log_24x7_hcall(request_buffer, result_buffer, ret);
  948. goto out;
  949. }
  950. /* process result from hcall */
  951. resb = &result_buffer->results[0];
  952. *count = be64_to_cpu(resb->elements[0].element_data[0]);
  953. out:
  954. put_cpu_var(hv_24x7_reqb);
  955. put_cpu_var(hv_24x7_resb);
  956. return ret;
  957. }
  958. static int h_24x7_event_init(struct perf_event *event)
  959. {
  960. struct hv_perf_caps caps;
  961. unsigned domain;
  962. unsigned long hret;
  963. u64 ct;
  964. /* Not our event */
  965. if (event->attr.type != event->pmu->type)
  966. return -ENOENT;
  967. /* Unused areas must be 0 */
  968. if (event_get_reserved1(event) ||
  969. event_get_reserved2(event) ||
  970. event_get_reserved3(event)) {
  971. pr_devel("reserved set when forbidden 0x%llx(0x%llx) 0x%llx(0x%llx) 0x%llx(0x%llx)\n",
  972. event->attr.config,
  973. event_get_reserved1(event),
  974. event->attr.config1,
  975. event_get_reserved2(event),
  976. event->attr.config2,
  977. event_get_reserved3(event));
  978. return -EINVAL;
  979. }
  980. /* unsupported modes and filters */
  981. if (event->attr.exclude_user ||
  982. event->attr.exclude_kernel ||
  983. event->attr.exclude_hv ||
  984. event->attr.exclude_idle ||
  985. event->attr.exclude_host ||
  986. event->attr.exclude_guest)
  987. return -EINVAL;
  988. /* no branch sampling */
  989. if (has_branch_stack(event))
  990. return -EOPNOTSUPP;
  991. /* offset must be 8 byte aligned */
  992. if (event_get_offset(event) % 8) {
  993. pr_devel("bad alignment\n");
  994. return -EINVAL;
  995. }
  996. /* Domains above 6 are invalid */
  997. domain = event_get_domain(event);
  998. if (domain > 6) {
  999. pr_devel("invalid domain %d\n", domain);
  1000. return -EINVAL;
  1001. }
  1002. hret = hv_perf_caps_get(&caps);
  1003. if (hret) {
  1004. pr_devel("could not get capabilities: rc=%ld\n", hret);
  1005. return -EIO;
  1006. }
  1007. /* Physical domains & other lpars require extra capabilities */
  1008. if (!caps.collect_privileged && (is_physical_domain(domain) ||
  1009. (event_get_lpar(event) != event_get_lpar_max()))) {
  1010. pr_devel("hv permissions disallow: is_physical_domain:%d, lpar=0x%llx\n",
  1011. is_physical_domain(domain),
  1012. event_get_lpar(event));
  1013. return -EACCES;
  1014. }
  1015. /* see if the event complains */
  1016. if (single_24x7_request(event, &ct)) {
  1017. pr_devel("test hcall failed\n");
  1018. return -EIO;
  1019. }
  1020. return 0;
  1021. }
  1022. static u64 h_24x7_get_value(struct perf_event *event)
  1023. {
  1024. unsigned long ret;
  1025. u64 ct;
  1026. ret = single_24x7_request(event, &ct);
  1027. if (ret)
  1028. /* We checked this in event init, shouldn't fail here... */
  1029. return 0;
  1030. return ct;
  1031. }
  1032. static void update_event_count(struct perf_event *event, u64 now)
  1033. {
  1034. s64 prev;
  1035. prev = local64_xchg(&event->hw.prev_count, now);
  1036. local64_add(now - prev, &event->count);
  1037. }
  1038. static void h_24x7_event_read(struct perf_event *event)
  1039. {
  1040. u64 now;
  1041. now = h_24x7_get_value(event);
  1042. update_event_count(event, now);
  1043. }
  1044. static void h_24x7_event_start(struct perf_event *event, int flags)
  1045. {
  1046. if (flags & PERF_EF_RELOAD)
  1047. local64_set(&event->hw.prev_count, h_24x7_get_value(event));
  1048. }
  1049. static void h_24x7_event_stop(struct perf_event *event, int flags)
  1050. {
  1051. h_24x7_event_read(event);
  1052. }
  1053. static int h_24x7_event_add(struct perf_event *event, int flags)
  1054. {
  1055. if (flags & PERF_EF_START)
  1056. h_24x7_event_start(event, flags);
  1057. return 0;
  1058. }
  1059. static struct pmu h_24x7_pmu = {
  1060. .task_ctx_nr = perf_invalid_context,
  1061. .name = "hv_24x7",
  1062. .attr_groups = attr_groups,
  1063. .event_init = h_24x7_event_init,
  1064. .add = h_24x7_event_add,
  1065. .del = h_24x7_event_stop,
  1066. .start = h_24x7_event_start,
  1067. .stop = h_24x7_event_stop,
  1068. .read = h_24x7_event_read,
  1069. };
  1070. static int hv_24x7_init(void)
  1071. {
  1072. int r;
  1073. unsigned long hret;
  1074. struct hv_perf_caps caps;
  1075. if (!firmware_has_feature(FW_FEATURE_LPAR)) {
  1076. pr_debug("not a virtualized system, not enabling\n");
  1077. return -ENODEV;
  1078. }
  1079. hret = hv_perf_caps_get(&caps);
  1080. if (hret) {
  1081. pr_debug("could not obtain capabilities, not enabling, rc=%ld\n",
  1082. hret);
  1083. return -ENODEV;
  1084. }
  1085. hv_page_cache = kmem_cache_create("hv-page-4096", 4096, 4096, 0, NULL);
  1086. if (!hv_page_cache)
  1087. return -ENOMEM;
  1088. /* sampling not supported */
  1089. h_24x7_pmu.capabilities |= PERF_PMU_CAP_NO_INTERRUPT;
  1090. r = create_events_from_catalog(&event_group.attrs,
  1091. &event_desc_group.attrs,
  1092. &event_long_desc_group.attrs);
  1093. if (r)
  1094. return r;
  1095. r = perf_pmu_register(&h_24x7_pmu, h_24x7_pmu.name, -1);
  1096. if (r)
  1097. return r;
  1098. return 0;
  1099. }
  1100. device_initcall(hv_24x7_init);