inet_diag.c 29 KB

1234567891011121314151617181920212223242526272829303132333435363738394041424344454647484950515253545556575859606162636465666768697071727374757677787980818283848586878889909192939495969798991001011021031041051061071081091101111121131141151161171181191201211221231241251261271281291301311321331341351361371381391401411421431441451461471481491501511521531541551561571581591601611621631641651661671681691701711721731741751761771781791801811821831841851861871881891901911921931941951961971981992002012022032042052062072082092102112122132142152162172182192202212222232242252262272282292302312322332342352362372382392402412422432442452462472482492502512522532542552562572582592602612622632642652662672682692702712722732742752762772782792802812822832842852862872882892902912922932942952962972982993003013023033043053063073083093103113123133143153163173183193203213223233243253263273283293303313323333343353363373383393403413423433443453463473483493503513523533543553563573583593603613623633643653663673683693703713723733743753763773783793803813823833843853863873883893903913923933943953963973983994004014024034044054064074084094104114124134144154164174184194204214224234244254264274284294304314324334344354364374384394404414424434444454464474484494504514524534544554564574584594604614624634644654664674684694704714724734744754764774784794804814824834844854864874884894904914924934944954964974984995005015025035045055065075085095105115125135145155165175185195205215225235245255265275285295305315325335345355365375385395405415425435445455465475485495505515525535545555565575585595605615625635645655665675685695705715725735745755765775785795805815825835845855865875885895905915925935945955965975985996006016026036046056066076086096106116126136146156166176186196206216226236246256266276286296306316326336346356366376386396406416426436446456466476486496506516526536546556566576586596606616626636646656666676686696706716726736746756766776786796806816826836846856866876886896906916926936946956966976986997007017027037047057067077087097107117127137147157167177187197207217227237247257267277287297307317327337347357367377387397407417427437447457467477487497507517527537547557567577587597607617627637647657667677687697707717727737747757767777787797807817827837847857867877887897907917927937947957967977987998008018028038048058068078088098108118128138148158168178188198208218228238248258268278288298308318328338348358368378388398408418428438448458468478488498508518528538548558568578588598608618628638648658668678688698708718728738748758768778788798808818828838848858868878888898908918928938948958968978988999009019029039049059069079089099109119129139149159169179189199209219229239249259269279289299309319329339349359369379389399409419429439449459469479489499509519529539549559569579589599609619629639649659669679689699709719729739749759769779789799809819829839849859869879889899909919929939949959969979989991000100110021003100410051006100710081009101010111012101310141015101610171018101910201021102210231024102510261027102810291030103110321033103410351036103710381039104010411042104310441045104610471048104910501051105210531054105510561057105810591060106110621063106410651066106710681069107010711072107310741075107610771078107910801081108210831084108510861087108810891090109110921093109410951096109710981099110011011102110311041105110611071108110911101111111211131114111511161117111811191120112111221123112411251126112711281129113011311132113311341135113611371138113911401141114211431144114511461147114811491150115111521153115411551156115711581159116011611162116311641165116611671168116911701171117211731174117511761177117811791180118111821183118411851186118711881189119011911192119311941195119611971198119912001201120212031204120512061207
  1. /*
  2. * inet_diag.c Module for monitoring INET transport protocols sockets.
  3. *
  4. * Authors: Alexey Kuznetsov, <kuznet@ms2.inr.ac.ru>
  5. *
  6. * This program is free software; you can redistribute it and/or
  7. * modify it under the terms of the GNU General Public License
  8. * as published by the Free Software Foundation; either version
  9. * 2 of the License, or (at your option) any later version.
  10. */
  11. #include <linux/kernel.h>
  12. #include <linux/module.h>
  13. #include <linux/types.h>
  14. #include <linux/fcntl.h>
  15. #include <linux/random.h>
  16. #include <linux/slab.h>
  17. #include <linux/cache.h>
  18. #include <linux/init.h>
  19. #include <linux/time.h>
  20. #include <net/icmp.h>
  21. #include <net/tcp.h>
  22. #include <net/ipv6.h>
  23. #include <net/inet_common.h>
  24. #include <net/inet_connection_sock.h>
  25. #include <net/inet_hashtables.h>
  26. #include <net/inet_timewait_sock.h>
  27. #include <net/inet6_hashtables.h>
  28. #include <net/netlink.h>
  29. #include <linux/inet.h>
  30. #include <linux/stddef.h>
  31. #include <linux/inet_diag.h>
  32. #include <linux/sock_diag.h>
  33. static const struct inet_diag_handler **inet_diag_table;
  34. struct inet_diag_entry {
  35. __be32 *saddr;
  36. __be32 *daddr;
  37. u16 sport;
  38. u16 dport;
  39. u16 family;
  40. u16 userlocks;
  41. #if IS_ENABLED(CONFIG_IPV6)
  42. struct in6_addr saddr_storage; /* for IPv4-mapped-IPv6 addresses */
  43. struct in6_addr daddr_storage; /* for IPv4-mapped-IPv6 addresses */
  44. #endif
  45. };
  46. static DEFINE_MUTEX(inet_diag_table_mutex);
  47. static const struct inet_diag_handler *inet_diag_lock_handler(int proto)
  48. {
  49. if (!inet_diag_table[proto])
  50. request_module("net-pf-%d-proto-%d-type-%d-%d", PF_NETLINK,
  51. NETLINK_SOCK_DIAG, AF_INET, proto);
  52. mutex_lock(&inet_diag_table_mutex);
  53. if (!inet_diag_table[proto])
  54. return ERR_PTR(-ENOENT);
  55. return inet_diag_table[proto];
  56. }
  57. static inline void inet_diag_unlock_handler(
  58. const struct inet_diag_handler *handler)
  59. {
  60. mutex_unlock(&inet_diag_table_mutex);
  61. }
  62. static size_t inet_sk_attr_size(void)
  63. {
  64. return nla_total_size(sizeof(struct tcp_info))
  65. + nla_total_size(1) /* INET_DIAG_SHUTDOWN */
  66. + nla_total_size(1) /* INET_DIAG_TOS */
  67. + nla_total_size(1) /* INET_DIAG_TCLASS */
  68. + nla_total_size(sizeof(struct inet_diag_meminfo))
  69. + nla_total_size(sizeof(struct inet_diag_msg))
  70. + nla_total_size(SK_MEMINFO_VARS * sizeof(u32))
  71. + nla_total_size(TCP_CA_NAME_MAX)
  72. + nla_total_size(sizeof(struct tcpvegas_info))
  73. + 64;
  74. }
  75. int inet_sk_diag_fill(struct sock *sk, struct inet_connection_sock *icsk,
  76. struct sk_buff *skb, struct inet_diag_req_v2 *req,
  77. struct user_namespace *user_ns,
  78. u32 portid, u32 seq, u16 nlmsg_flags,
  79. const struct nlmsghdr *unlh)
  80. {
  81. const struct inet_sock *inet = inet_sk(sk);
  82. struct inet_diag_msg *r;
  83. struct nlmsghdr *nlh;
  84. struct nlattr *attr;
  85. void *info = NULL;
  86. const struct inet_diag_handler *handler;
  87. int ext = req->idiag_ext;
  88. handler = inet_diag_table[req->sdiag_protocol];
  89. BUG_ON(handler == NULL);
  90. nlh = nlmsg_put(skb, portid, seq, unlh->nlmsg_type, sizeof(*r),
  91. nlmsg_flags);
  92. if (!nlh)
  93. return -EMSGSIZE;
  94. r = nlmsg_data(nlh);
  95. BUG_ON(sk->sk_state == TCP_TIME_WAIT);
  96. r->idiag_family = sk->sk_family;
  97. r->idiag_state = sk->sk_state;
  98. r->idiag_timer = 0;
  99. r->idiag_retrans = 0;
  100. r->id.idiag_if = sk->sk_bound_dev_if;
  101. sock_diag_save_cookie(sk, r->id.idiag_cookie);
  102. r->id.idiag_sport = inet->inet_sport;
  103. r->id.idiag_dport = inet->inet_dport;
  104. memset(&r->id.idiag_src, 0, sizeof(r->id.idiag_src));
  105. memset(&r->id.idiag_dst, 0, sizeof(r->id.idiag_dst));
  106. r->id.idiag_src[0] = inet->inet_rcv_saddr;
  107. r->id.idiag_dst[0] = inet->inet_daddr;
  108. if (nla_put_u8(skb, INET_DIAG_SHUTDOWN, sk->sk_shutdown))
  109. goto errout;
  110. /* IPv6 dual-stack sockets use inet->tos for IPv4 connections,
  111. * hence this needs to be included regardless of socket family.
  112. */
  113. if (ext & (1 << (INET_DIAG_TOS - 1)))
  114. if (nla_put_u8(skb, INET_DIAG_TOS, inet->tos) < 0)
  115. goto errout;
  116. #if IS_ENABLED(CONFIG_IPV6)
  117. if (r->idiag_family == AF_INET6) {
  118. *(struct in6_addr *)r->id.idiag_src = sk->sk_v6_rcv_saddr;
  119. *(struct in6_addr *)r->id.idiag_dst = sk->sk_v6_daddr;
  120. if (ext & (1 << (INET_DIAG_TCLASS - 1)))
  121. if (nla_put_u8(skb, INET_DIAG_TCLASS,
  122. inet6_sk(sk)->tclass) < 0)
  123. goto errout;
  124. }
  125. #endif
  126. r->idiag_uid = from_kuid_munged(user_ns, sock_i_uid(sk));
  127. r->idiag_inode = sock_i_ino(sk);
  128. if (ext & (1 << (INET_DIAG_MEMINFO - 1))) {
  129. struct inet_diag_meminfo minfo = {
  130. .idiag_rmem = sk_rmem_alloc_get(sk),
  131. .idiag_wmem = sk->sk_wmem_queued,
  132. .idiag_fmem = sk->sk_forward_alloc,
  133. .idiag_tmem = sk_wmem_alloc_get(sk),
  134. };
  135. if (nla_put(skb, INET_DIAG_MEMINFO, sizeof(minfo), &minfo) < 0)
  136. goto errout;
  137. }
  138. if (ext & (1 << (INET_DIAG_SKMEMINFO - 1)))
  139. if (sock_diag_put_meminfo(sk, skb, INET_DIAG_SKMEMINFO))
  140. goto errout;
  141. if (icsk == NULL) {
  142. handler->idiag_get_info(sk, r, NULL);
  143. goto out;
  144. }
  145. #define EXPIRES_IN_MS(tmo) DIV_ROUND_UP((tmo - jiffies) * 1000, HZ)
  146. if (icsk->icsk_pending == ICSK_TIME_RETRANS ||
  147. icsk->icsk_pending == ICSK_TIME_EARLY_RETRANS ||
  148. icsk->icsk_pending == ICSK_TIME_LOSS_PROBE) {
  149. r->idiag_timer = 1;
  150. r->idiag_retrans = icsk->icsk_retransmits;
  151. r->idiag_expires = EXPIRES_IN_MS(icsk->icsk_timeout);
  152. } else if (icsk->icsk_pending == ICSK_TIME_PROBE0) {
  153. r->idiag_timer = 4;
  154. r->idiag_retrans = icsk->icsk_probes_out;
  155. r->idiag_expires = EXPIRES_IN_MS(icsk->icsk_timeout);
  156. } else if (timer_pending(&sk->sk_timer)) {
  157. r->idiag_timer = 2;
  158. r->idiag_retrans = icsk->icsk_probes_out;
  159. r->idiag_expires = EXPIRES_IN_MS(sk->sk_timer.expires);
  160. } else {
  161. r->idiag_timer = 0;
  162. r->idiag_expires = 0;
  163. }
  164. #undef EXPIRES_IN_MS
  165. if (ext & (1 << (INET_DIAG_INFO - 1))) {
  166. attr = nla_reserve(skb, INET_DIAG_INFO,
  167. sizeof(struct tcp_info));
  168. if (!attr)
  169. goto errout;
  170. info = nla_data(attr);
  171. }
  172. if ((ext & (1 << (INET_DIAG_CONG - 1))) && icsk->icsk_ca_ops)
  173. if (nla_put_string(skb, INET_DIAG_CONG,
  174. icsk->icsk_ca_ops->name) < 0)
  175. goto errout;
  176. handler->idiag_get_info(sk, r, info);
  177. if (sk->sk_state < TCP_TIME_WAIT &&
  178. icsk->icsk_ca_ops && icsk->icsk_ca_ops->get_info)
  179. icsk->icsk_ca_ops->get_info(sk, ext, skb);
  180. out:
  181. nlmsg_end(skb, nlh);
  182. return 0;
  183. errout:
  184. nlmsg_cancel(skb, nlh);
  185. return -EMSGSIZE;
  186. }
  187. EXPORT_SYMBOL_GPL(inet_sk_diag_fill);
  188. static int inet_csk_diag_fill(struct sock *sk,
  189. struct sk_buff *skb, struct inet_diag_req_v2 *req,
  190. struct user_namespace *user_ns,
  191. u32 portid, u32 seq, u16 nlmsg_flags,
  192. const struct nlmsghdr *unlh)
  193. {
  194. return inet_sk_diag_fill(sk, inet_csk(sk),
  195. skb, req, user_ns, portid, seq, nlmsg_flags, unlh);
  196. }
  197. static int inet_twsk_diag_fill(struct inet_timewait_sock *tw,
  198. struct sk_buff *skb, struct inet_diag_req_v2 *req,
  199. u32 portid, u32 seq, u16 nlmsg_flags,
  200. const struct nlmsghdr *unlh)
  201. {
  202. s32 tmo;
  203. struct inet_diag_msg *r;
  204. struct nlmsghdr *nlh;
  205. nlh = nlmsg_put(skb, portid, seq, unlh->nlmsg_type, sizeof(*r),
  206. nlmsg_flags);
  207. if (!nlh)
  208. return -EMSGSIZE;
  209. r = nlmsg_data(nlh);
  210. BUG_ON(tw->tw_state != TCP_TIME_WAIT);
  211. tmo = tw->tw_ttd - inet_tw_time_stamp();
  212. if (tmo < 0)
  213. tmo = 0;
  214. r->idiag_family = tw->tw_family;
  215. r->idiag_retrans = 0;
  216. r->id.idiag_if = tw->tw_bound_dev_if;
  217. sock_diag_save_cookie(tw, r->id.idiag_cookie);
  218. r->id.idiag_sport = tw->tw_sport;
  219. r->id.idiag_dport = tw->tw_dport;
  220. memset(&r->id.idiag_src, 0, sizeof(r->id.idiag_src));
  221. memset(&r->id.idiag_dst, 0, sizeof(r->id.idiag_dst));
  222. r->id.idiag_src[0] = tw->tw_rcv_saddr;
  223. r->id.idiag_dst[0] = tw->tw_daddr;
  224. r->idiag_state = tw->tw_substate;
  225. r->idiag_timer = 3;
  226. r->idiag_expires = jiffies_to_msecs(tmo);
  227. r->idiag_rqueue = 0;
  228. r->idiag_wqueue = 0;
  229. r->idiag_uid = 0;
  230. r->idiag_inode = 0;
  231. #if IS_ENABLED(CONFIG_IPV6)
  232. if (tw->tw_family == AF_INET6) {
  233. *(struct in6_addr *)r->id.idiag_src = tw->tw_v6_rcv_saddr;
  234. *(struct in6_addr *)r->id.idiag_dst = tw->tw_v6_daddr;
  235. }
  236. #endif
  237. nlmsg_end(skb, nlh);
  238. return 0;
  239. }
  240. static int sk_diag_fill(struct sock *sk, struct sk_buff *skb,
  241. struct inet_diag_req_v2 *r,
  242. struct user_namespace *user_ns,
  243. u32 portid, u32 seq, u16 nlmsg_flags,
  244. const struct nlmsghdr *unlh)
  245. {
  246. if (sk->sk_state == TCP_TIME_WAIT)
  247. return inet_twsk_diag_fill(inet_twsk(sk), skb, r, portid, seq,
  248. nlmsg_flags, unlh);
  249. return inet_csk_diag_fill(sk, skb, r, user_ns, portid, seq,
  250. nlmsg_flags, unlh);
  251. }
  252. int inet_diag_dump_one_icsk(struct inet_hashinfo *hashinfo, struct sk_buff *in_skb,
  253. const struct nlmsghdr *nlh, struct inet_diag_req_v2 *req)
  254. {
  255. int err;
  256. struct sock *sk;
  257. struct sk_buff *rep;
  258. struct net *net = sock_net(in_skb->sk);
  259. err = -EINVAL;
  260. if (req->sdiag_family == AF_INET) {
  261. sk = inet_lookup(net, hashinfo, req->id.idiag_dst[0],
  262. req->id.idiag_dport, req->id.idiag_src[0],
  263. req->id.idiag_sport, req->id.idiag_if);
  264. }
  265. #if IS_ENABLED(CONFIG_IPV6)
  266. else if (req->sdiag_family == AF_INET6) {
  267. sk = inet6_lookup(net, hashinfo,
  268. (struct in6_addr *)req->id.idiag_dst,
  269. req->id.idiag_dport,
  270. (struct in6_addr *)req->id.idiag_src,
  271. req->id.idiag_sport,
  272. req->id.idiag_if);
  273. }
  274. #endif
  275. else {
  276. goto out_nosk;
  277. }
  278. err = -ENOENT;
  279. if (sk == NULL)
  280. goto out_nosk;
  281. err = sock_diag_check_cookie(sk, req->id.idiag_cookie);
  282. if (err)
  283. goto out;
  284. rep = nlmsg_new(inet_sk_attr_size(), GFP_KERNEL);
  285. if (!rep) {
  286. err = -ENOMEM;
  287. goto out;
  288. }
  289. err = sk_diag_fill(sk, rep, req,
  290. sk_user_ns(NETLINK_CB(in_skb).sk),
  291. NETLINK_CB(in_skb).portid,
  292. nlh->nlmsg_seq, 0, nlh);
  293. if (err < 0) {
  294. WARN_ON(err == -EMSGSIZE);
  295. nlmsg_free(rep);
  296. goto out;
  297. }
  298. err = netlink_unicast(net->diag_nlsk, rep, NETLINK_CB(in_skb).portid,
  299. MSG_DONTWAIT);
  300. if (err > 0)
  301. err = 0;
  302. out:
  303. if (sk)
  304. sock_gen_put(sk);
  305. out_nosk:
  306. return err;
  307. }
  308. EXPORT_SYMBOL_GPL(inet_diag_dump_one_icsk);
  309. static int inet_diag_get_exact(struct sk_buff *in_skb,
  310. const struct nlmsghdr *nlh,
  311. struct inet_diag_req_v2 *req)
  312. {
  313. const struct inet_diag_handler *handler;
  314. int err;
  315. handler = inet_diag_lock_handler(req->sdiag_protocol);
  316. if (IS_ERR(handler))
  317. err = PTR_ERR(handler);
  318. else
  319. err = handler->dump_one(in_skb, nlh, req);
  320. inet_diag_unlock_handler(handler);
  321. return err;
  322. }
  323. static int bitstring_match(const __be32 *a1, const __be32 *a2, int bits)
  324. {
  325. int words = bits >> 5;
  326. bits &= 0x1f;
  327. if (words) {
  328. if (memcmp(a1, a2, words << 2))
  329. return 0;
  330. }
  331. if (bits) {
  332. __be32 w1, w2;
  333. __be32 mask;
  334. w1 = a1[words];
  335. w2 = a2[words];
  336. mask = htonl((0xffffffff) << (32 - bits));
  337. if ((w1 ^ w2) & mask)
  338. return 0;
  339. }
  340. return 1;
  341. }
  342. static int inet_diag_bc_run(const struct nlattr *_bc,
  343. const struct inet_diag_entry *entry)
  344. {
  345. const void *bc = nla_data(_bc);
  346. int len = nla_len(_bc);
  347. while (len > 0) {
  348. int yes = 1;
  349. const struct inet_diag_bc_op *op = bc;
  350. switch (op->code) {
  351. case INET_DIAG_BC_NOP:
  352. break;
  353. case INET_DIAG_BC_JMP:
  354. yes = 0;
  355. break;
  356. case INET_DIAG_BC_S_GE:
  357. yes = entry->sport >= op[1].no;
  358. break;
  359. case INET_DIAG_BC_S_LE:
  360. yes = entry->sport <= op[1].no;
  361. break;
  362. case INET_DIAG_BC_D_GE:
  363. yes = entry->dport >= op[1].no;
  364. break;
  365. case INET_DIAG_BC_D_LE:
  366. yes = entry->dport <= op[1].no;
  367. break;
  368. case INET_DIAG_BC_AUTO:
  369. yes = !(entry->userlocks & SOCK_BINDPORT_LOCK);
  370. break;
  371. case INET_DIAG_BC_S_COND:
  372. case INET_DIAG_BC_D_COND: {
  373. struct inet_diag_hostcond *cond;
  374. __be32 *addr;
  375. cond = (struct inet_diag_hostcond *)(op + 1);
  376. if (cond->port != -1 &&
  377. cond->port != (op->code == INET_DIAG_BC_S_COND ?
  378. entry->sport : entry->dport)) {
  379. yes = 0;
  380. break;
  381. }
  382. if (op->code == INET_DIAG_BC_S_COND)
  383. addr = entry->saddr;
  384. else
  385. addr = entry->daddr;
  386. if (cond->family != AF_UNSPEC &&
  387. cond->family != entry->family) {
  388. if (entry->family == AF_INET6 &&
  389. cond->family == AF_INET) {
  390. if (addr[0] == 0 && addr[1] == 0 &&
  391. addr[2] == htonl(0xffff) &&
  392. bitstring_match(addr + 3,
  393. cond->addr,
  394. cond->prefix_len))
  395. break;
  396. }
  397. yes = 0;
  398. break;
  399. }
  400. if (cond->prefix_len == 0)
  401. break;
  402. if (bitstring_match(addr, cond->addr,
  403. cond->prefix_len))
  404. break;
  405. yes = 0;
  406. break;
  407. }
  408. }
  409. if (yes) {
  410. len -= op->yes;
  411. bc += op->yes;
  412. } else {
  413. len -= op->no;
  414. bc += op->no;
  415. }
  416. }
  417. return len == 0;
  418. }
  419. int inet_diag_bc_sk(const struct nlattr *bc, struct sock *sk)
  420. {
  421. struct inet_diag_entry entry;
  422. struct inet_sock *inet = inet_sk(sk);
  423. if (bc == NULL)
  424. return 1;
  425. entry.family = sk->sk_family;
  426. #if IS_ENABLED(CONFIG_IPV6)
  427. if (entry.family == AF_INET6) {
  428. entry.saddr = sk->sk_v6_rcv_saddr.s6_addr32;
  429. entry.daddr = sk->sk_v6_daddr.s6_addr32;
  430. } else
  431. #endif
  432. {
  433. entry.saddr = &inet->inet_rcv_saddr;
  434. entry.daddr = &inet->inet_daddr;
  435. }
  436. entry.sport = inet->inet_num;
  437. entry.dport = ntohs(inet->inet_dport);
  438. entry.userlocks = sk->sk_userlocks;
  439. return inet_diag_bc_run(bc, &entry);
  440. }
  441. EXPORT_SYMBOL_GPL(inet_diag_bc_sk);
  442. static int valid_cc(const void *bc, int len, int cc)
  443. {
  444. while (len >= 0) {
  445. const struct inet_diag_bc_op *op = bc;
  446. if (cc > len)
  447. return 0;
  448. if (cc == len)
  449. return 1;
  450. if (op->yes < 4 || op->yes & 3)
  451. return 0;
  452. len -= op->yes;
  453. bc += op->yes;
  454. }
  455. return 0;
  456. }
  457. /* Validate an inet_diag_hostcond. */
  458. static bool valid_hostcond(const struct inet_diag_bc_op *op, int len,
  459. int *min_len)
  460. {
  461. int addr_len;
  462. struct inet_diag_hostcond *cond;
  463. /* Check hostcond space. */
  464. *min_len += sizeof(struct inet_diag_hostcond);
  465. if (len < *min_len)
  466. return false;
  467. cond = (struct inet_diag_hostcond *)(op + 1);
  468. /* Check address family and address length. */
  469. switch (cond->family) {
  470. case AF_UNSPEC:
  471. addr_len = 0;
  472. break;
  473. case AF_INET:
  474. addr_len = sizeof(struct in_addr);
  475. break;
  476. case AF_INET6:
  477. addr_len = sizeof(struct in6_addr);
  478. break;
  479. default:
  480. return false;
  481. }
  482. *min_len += addr_len;
  483. if (len < *min_len)
  484. return false;
  485. /* Check prefix length (in bits) vs address length (in bytes). */
  486. if (cond->prefix_len > 8 * addr_len)
  487. return false;
  488. return true;
  489. }
  490. /* Validate a port comparison operator. */
  491. static inline bool valid_port_comparison(const struct inet_diag_bc_op *op,
  492. int len, int *min_len)
  493. {
  494. /* Port comparisons put the port in a follow-on inet_diag_bc_op. */
  495. *min_len += sizeof(struct inet_diag_bc_op);
  496. if (len < *min_len)
  497. return false;
  498. return true;
  499. }
  500. static int inet_diag_bc_audit(const void *bytecode, int bytecode_len)
  501. {
  502. const void *bc = bytecode;
  503. int len = bytecode_len;
  504. while (len > 0) {
  505. const struct inet_diag_bc_op *op = bc;
  506. int min_len = sizeof(struct inet_diag_bc_op);
  507. //printk("BC: %d %d %d {%d} / %d\n", op->code, op->yes, op->no, op[1].no, len);
  508. switch (op->code) {
  509. case INET_DIAG_BC_S_COND:
  510. case INET_DIAG_BC_D_COND:
  511. if (!valid_hostcond(bc, len, &min_len))
  512. return -EINVAL;
  513. break;
  514. case INET_DIAG_BC_S_GE:
  515. case INET_DIAG_BC_S_LE:
  516. case INET_DIAG_BC_D_GE:
  517. case INET_DIAG_BC_D_LE:
  518. if (!valid_port_comparison(bc, len, &min_len))
  519. return -EINVAL;
  520. break;
  521. case INET_DIAG_BC_AUTO:
  522. case INET_DIAG_BC_JMP:
  523. case INET_DIAG_BC_NOP:
  524. break;
  525. default:
  526. return -EINVAL;
  527. }
  528. if (op->code != INET_DIAG_BC_NOP) {
  529. if (op->no < min_len || op->no > len + 4 || op->no & 3)
  530. return -EINVAL;
  531. if (op->no < len &&
  532. !valid_cc(bytecode, bytecode_len, len - op->no))
  533. return -EINVAL;
  534. }
  535. if (op->yes < min_len || op->yes > len + 4 || op->yes & 3)
  536. return -EINVAL;
  537. bc += op->yes;
  538. len -= op->yes;
  539. }
  540. return len == 0 ? 0 : -EINVAL;
  541. }
  542. static int inet_csk_diag_dump(struct sock *sk,
  543. struct sk_buff *skb,
  544. struct netlink_callback *cb,
  545. struct inet_diag_req_v2 *r,
  546. const struct nlattr *bc)
  547. {
  548. if (!inet_diag_bc_sk(bc, sk))
  549. return 0;
  550. return inet_csk_diag_fill(sk, skb, r,
  551. sk_user_ns(NETLINK_CB(cb->skb).sk),
  552. NETLINK_CB(cb->skb).portid,
  553. cb->nlh->nlmsg_seq, NLM_F_MULTI, cb->nlh);
  554. }
  555. static int inet_twsk_diag_dump(struct sock *sk,
  556. struct sk_buff *skb,
  557. struct netlink_callback *cb,
  558. struct inet_diag_req_v2 *r,
  559. const struct nlattr *bc)
  560. {
  561. struct inet_timewait_sock *tw = inet_twsk(sk);
  562. if (bc != NULL) {
  563. struct inet_diag_entry entry;
  564. entry.family = tw->tw_family;
  565. #if IS_ENABLED(CONFIG_IPV6)
  566. if (tw->tw_family == AF_INET6) {
  567. entry.saddr = tw->tw_v6_rcv_saddr.s6_addr32;
  568. entry.daddr = tw->tw_v6_daddr.s6_addr32;
  569. } else
  570. #endif
  571. {
  572. entry.saddr = &tw->tw_rcv_saddr;
  573. entry.daddr = &tw->tw_daddr;
  574. }
  575. entry.sport = tw->tw_num;
  576. entry.dport = ntohs(tw->tw_dport);
  577. entry.userlocks = 0;
  578. if (!inet_diag_bc_run(bc, &entry))
  579. return 0;
  580. }
  581. return inet_twsk_diag_fill(tw, skb, r,
  582. NETLINK_CB(cb->skb).portid,
  583. cb->nlh->nlmsg_seq, NLM_F_MULTI, cb->nlh);
  584. }
  585. /* Get the IPv4, IPv6, or IPv4-mapped-IPv6 local and remote addresses
  586. * from a request_sock. For IPv4-mapped-IPv6 we must map IPv4 to IPv6.
  587. */
  588. static inline void inet_diag_req_addrs(const struct sock *sk,
  589. const struct request_sock *req,
  590. struct inet_diag_entry *entry)
  591. {
  592. struct inet_request_sock *ireq = inet_rsk(req);
  593. #if IS_ENABLED(CONFIG_IPV6)
  594. if (sk->sk_family == AF_INET6) {
  595. if (req->rsk_ops->family == AF_INET6) {
  596. entry->saddr = ireq->ir_v6_loc_addr.s6_addr32;
  597. entry->daddr = ireq->ir_v6_rmt_addr.s6_addr32;
  598. } else if (req->rsk_ops->family == AF_INET) {
  599. ipv6_addr_set_v4mapped(ireq->ir_loc_addr,
  600. &entry->saddr_storage);
  601. ipv6_addr_set_v4mapped(ireq->ir_rmt_addr,
  602. &entry->daddr_storage);
  603. entry->saddr = entry->saddr_storage.s6_addr32;
  604. entry->daddr = entry->daddr_storage.s6_addr32;
  605. }
  606. } else
  607. #endif
  608. {
  609. entry->saddr = &ireq->ir_loc_addr;
  610. entry->daddr = &ireq->ir_rmt_addr;
  611. }
  612. }
  613. static int inet_diag_fill_req(struct sk_buff *skb, struct sock *sk,
  614. struct request_sock *req,
  615. struct user_namespace *user_ns,
  616. u32 portid, u32 seq,
  617. const struct nlmsghdr *unlh)
  618. {
  619. const struct inet_request_sock *ireq = inet_rsk(req);
  620. struct inet_sock *inet = inet_sk(sk);
  621. struct inet_diag_msg *r;
  622. struct nlmsghdr *nlh;
  623. long tmo;
  624. nlh = nlmsg_put(skb, portid, seq, unlh->nlmsg_type, sizeof(*r),
  625. NLM_F_MULTI);
  626. if (!nlh)
  627. return -EMSGSIZE;
  628. r = nlmsg_data(nlh);
  629. r->idiag_family = sk->sk_family;
  630. r->idiag_state = TCP_SYN_RECV;
  631. r->idiag_timer = 1;
  632. r->idiag_retrans = req->num_retrans;
  633. r->id.idiag_if = sk->sk_bound_dev_if;
  634. sock_diag_save_cookie(req, r->id.idiag_cookie);
  635. tmo = req->expires - jiffies;
  636. if (tmo < 0)
  637. tmo = 0;
  638. r->id.idiag_sport = inet->inet_sport;
  639. r->id.idiag_dport = ireq->ir_rmt_port;
  640. memset(&r->id.idiag_src, 0, sizeof(r->id.idiag_src));
  641. memset(&r->id.idiag_dst, 0, sizeof(r->id.idiag_dst));
  642. r->id.idiag_src[0] = ireq->ir_loc_addr;
  643. r->id.idiag_dst[0] = ireq->ir_rmt_addr;
  644. r->idiag_expires = jiffies_to_msecs(tmo);
  645. r->idiag_rqueue = 0;
  646. r->idiag_wqueue = 0;
  647. r->idiag_uid = from_kuid_munged(user_ns, sock_i_uid(sk));
  648. r->idiag_inode = 0;
  649. #if IS_ENABLED(CONFIG_IPV6)
  650. if (r->idiag_family == AF_INET6) {
  651. struct inet_diag_entry entry;
  652. inet_diag_req_addrs(sk, req, &entry);
  653. memcpy(r->id.idiag_src, entry.saddr, sizeof(struct in6_addr));
  654. memcpy(r->id.idiag_dst, entry.daddr, sizeof(struct in6_addr));
  655. }
  656. #endif
  657. nlmsg_end(skb, nlh);
  658. return 0;
  659. }
  660. static int inet_diag_dump_reqs(struct sk_buff *skb, struct sock *sk,
  661. struct netlink_callback *cb,
  662. struct inet_diag_req_v2 *r,
  663. const struct nlattr *bc)
  664. {
  665. struct inet_diag_entry entry;
  666. struct inet_connection_sock *icsk = inet_csk(sk);
  667. struct listen_sock *lopt;
  668. struct inet_sock *inet = inet_sk(sk);
  669. int j, s_j;
  670. int reqnum, s_reqnum;
  671. int err = 0;
  672. s_j = cb->args[3];
  673. s_reqnum = cb->args[4];
  674. if (s_j > 0)
  675. s_j--;
  676. entry.family = sk->sk_family;
  677. read_lock_bh(&icsk->icsk_accept_queue.syn_wait_lock);
  678. lopt = icsk->icsk_accept_queue.listen_opt;
  679. if (!lopt || !lopt->qlen)
  680. goto out;
  681. if (bc != NULL) {
  682. entry.sport = inet->inet_num;
  683. entry.userlocks = sk->sk_userlocks;
  684. }
  685. for (j = s_j; j < lopt->nr_table_entries; j++) {
  686. struct request_sock *req, *head = lopt->syn_table[j];
  687. reqnum = 0;
  688. for (req = head; req; reqnum++, req = req->dl_next) {
  689. struct inet_request_sock *ireq = inet_rsk(req);
  690. if (reqnum < s_reqnum)
  691. continue;
  692. if (r->id.idiag_dport != ireq->ir_rmt_port &&
  693. r->id.idiag_dport)
  694. continue;
  695. if (bc) {
  696. inet_diag_req_addrs(sk, req, &entry);
  697. entry.dport = ntohs(ireq->ir_rmt_port);
  698. if (!inet_diag_bc_run(bc, &entry))
  699. continue;
  700. }
  701. err = inet_diag_fill_req(skb, sk, req,
  702. sk_user_ns(NETLINK_CB(cb->skb).sk),
  703. NETLINK_CB(cb->skb).portid,
  704. cb->nlh->nlmsg_seq, cb->nlh);
  705. if (err < 0) {
  706. cb->args[3] = j + 1;
  707. cb->args[4] = reqnum;
  708. goto out;
  709. }
  710. }
  711. s_reqnum = 0;
  712. }
  713. out:
  714. read_unlock_bh(&icsk->icsk_accept_queue.syn_wait_lock);
  715. return err;
  716. }
  717. void inet_diag_dump_icsk(struct inet_hashinfo *hashinfo, struct sk_buff *skb,
  718. struct netlink_callback *cb, struct inet_diag_req_v2 *r, struct nlattr *bc)
  719. {
  720. int i, num;
  721. int s_i, s_num;
  722. struct net *net = sock_net(skb->sk);
  723. s_i = cb->args[1];
  724. s_num = num = cb->args[2];
  725. if (cb->args[0] == 0) {
  726. if (!(r->idiag_states & (TCPF_LISTEN | TCPF_SYN_RECV)))
  727. goto skip_listen_ht;
  728. for (i = s_i; i < INET_LHTABLE_SIZE; i++) {
  729. struct sock *sk;
  730. struct hlist_nulls_node *node;
  731. struct inet_listen_hashbucket *ilb;
  732. num = 0;
  733. ilb = &hashinfo->listening_hash[i];
  734. spin_lock_bh(&ilb->lock);
  735. sk_nulls_for_each(sk, node, &ilb->head) {
  736. struct inet_sock *inet = inet_sk(sk);
  737. if (!net_eq(sock_net(sk), net))
  738. continue;
  739. if (num < s_num) {
  740. num++;
  741. continue;
  742. }
  743. if (r->sdiag_family != AF_UNSPEC &&
  744. sk->sk_family != r->sdiag_family)
  745. goto next_listen;
  746. if (r->id.idiag_sport != inet->inet_sport &&
  747. r->id.idiag_sport)
  748. goto next_listen;
  749. if (!(r->idiag_states & TCPF_LISTEN) ||
  750. r->id.idiag_dport ||
  751. cb->args[3] > 0)
  752. goto syn_recv;
  753. if (inet_csk_diag_dump(sk, skb, cb, r, bc) < 0) {
  754. spin_unlock_bh(&ilb->lock);
  755. goto done;
  756. }
  757. syn_recv:
  758. if (!(r->idiag_states & TCPF_SYN_RECV))
  759. goto next_listen;
  760. if (inet_diag_dump_reqs(skb, sk, cb, r, bc) < 0) {
  761. spin_unlock_bh(&ilb->lock);
  762. goto done;
  763. }
  764. next_listen:
  765. cb->args[3] = 0;
  766. cb->args[4] = 0;
  767. ++num;
  768. }
  769. spin_unlock_bh(&ilb->lock);
  770. s_num = 0;
  771. cb->args[3] = 0;
  772. cb->args[4] = 0;
  773. }
  774. skip_listen_ht:
  775. cb->args[0] = 1;
  776. s_i = num = s_num = 0;
  777. }
  778. if (!(r->idiag_states & ~(TCPF_LISTEN | TCPF_SYN_RECV)))
  779. goto out;
  780. for (i = s_i; i <= hashinfo->ehash_mask; i++) {
  781. struct inet_ehash_bucket *head = &hashinfo->ehash[i];
  782. spinlock_t *lock = inet_ehash_lockp(hashinfo, i);
  783. struct sock *sk;
  784. struct hlist_nulls_node *node;
  785. num = 0;
  786. if (hlist_nulls_empty(&head->chain))
  787. continue;
  788. if (i > s_i)
  789. s_num = 0;
  790. spin_lock_bh(lock);
  791. sk_nulls_for_each(sk, node, &head->chain) {
  792. int res;
  793. int state;
  794. if (!net_eq(sock_net(sk), net))
  795. continue;
  796. if (num < s_num)
  797. goto next_normal;
  798. state = (sk->sk_state == TCP_TIME_WAIT) ?
  799. inet_twsk(sk)->tw_substate : sk->sk_state;
  800. if (!(r->idiag_states & (1 << state)))
  801. goto next_normal;
  802. if (r->sdiag_family != AF_UNSPEC &&
  803. sk->sk_family != r->sdiag_family)
  804. goto next_normal;
  805. if (r->id.idiag_sport != htons(sk->sk_num) &&
  806. r->id.idiag_sport)
  807. goto next_normal;
  808. if (r->id.idiag_dport != sk->sk_dport &&
  809. r->id.idiag_dport)
  810. goto next_normal;
  811. if (sk->sk_state == TCP_TIME_WAIT)
  812. res = inet_twsk_diag_dump(sk, skb, cb, r, bc);
  813. else
  814. res = inet_csk_diag_dump(sk, skb, cb, r, bc);
  815. if (res < 0) {
  816. spin_unlock_bh(lock);
  817. goto done;
  818. }
  819. next_normal:
  820. ++num;
  821. }
  822. spin_unlock_bh(lock);
  823. }
  824. done:
  825. cb->args[1] = i;
  826. cb->args[2] = num;
  827. out:
  828. ;
  829. }
  830. EXPORT_SYMBOL_GPL(inet_diag_dump_icsk);
  831. static int __inet_diag_dump(struct sk_buff *skb, struct netlink_callback *cb,
  832. struct inet_diag_req_v2 *r, struct nlattr *bc)
  833. {
  834. const struct inet_diag_handler *handler;
  835. int err = 0;
  836. handler = inet_diag_lock_handler(r->sdiag_protocol);
  837. if (!IS_ERR(handler))
  838. handler->dump(skb, cb, r, bc);
  839. else
  840. err = PTR_ERR(handler);
  841. inet_diag_unlock_handler(handler);
  842. return err ? : skb->len;
  843. }
  844. static int inet_diag_dump(struct sk_buff *skb, struct netlink_callback *cb)
  845. {
  846. struct nlattr *bc = NULL;
  847. int hdrlen = sizeof(struct inet_diag_req_v2);
  848. if (nlmsg_attrlen(cb->nlh, hdrlen))
  849. bc = nlmsg_find_attr(cb->nlh, hdrlen, INET_DIAG_REQ_BYTECODE);
  850. return __inet_diag_dump(skb, cb, nlmsg_data(cb->nlh), bc);
  851. }
  852. static inline int inet_diag_type2proto(int type)
  853. {
  854. switch (type) {
  855. case TCPDIAG_GETSOCK:
  856. return IPPROTO_TCP;
  857. case DCCPDIAG_GETSOCK:
  858. return IPPROTO_DCCP;
  859. default:
  860. return 0;
  861. }
  862. }
  863. static int inet_diag_dump_compat(struct sk_buff *skb, struct netlink_callback *cb)
  864. {
  865. struct inet_diag_req *rc = nlmsg_data(cb->nlh);
  866. struct inet_diag_req_v2 req;
  867. struct nlattr *bc = NULL;
  868. int hdrlen = sizeof(struct inet_diag_req);
  869. req.sdiag_family = AF_UNSPEC; /* compatibility */
  870. req.sdiag_protocol = inet_diag_type2proto(cb->nlh->nlmsg_type);
  871. req.idiag_ext = rc->idiag_ext;
  872. req.idiag_states = rc->idiag_states;
  873. req.id = rc->id;
  874. if (nlmsg_attrlen(cb->nlh, hdrlen))
  875. bc = nlmsg_find_attr(cb->nlh, hdrlen, INET_DIAG_REQ_BYTECODE);
  876. return __inet_diag_dump(skb, cb, &req, bc);
  877. }
  878. static int inet_diag_get_exact_compat(struct sk_buff *in_skb,
  879. const struct nlmsghdr *nlh)
  880. {
  881. struct inet_diag_req *rc = nlmsg_data(nlh);
  882. struct inet_diag_req_v2 req;
  883. req.sdiag_family = rc->idiag_family;
  884. req.sdiag_protocol = inet_diag_type2proto(nlh->nlmsg_type);
  885. req.idiag_ext = rc->idiag_ext;
  886. req.idiag_states = rc->idiag_states;
  887. req.id = rc->id;
  888. return inet_diag_get_exact(in_skb, nlh, &req);
  889. }
  890. static int inet_diag_rcv_msg_compat(struct sk_buff *skb, struct nlmsghdr *nlh)
  891. {
  892. int hdrlen = sizeof(struct inet_diag_req);
  893. struct net *net = sock_net(skb->sk);
  894. if (nlh->nlmsg_type >= INET_DIAG_GETSOCK_MAX ||
  895. nlmsg_len(nlh) < hdrlen)
  896. return -EINVAL;
  897. if (nlh->nlmsg_flags & NLM_F_DUMP) {
  898. if (nlmsg_attrlen(nlh, hdrlen)) {
  899. struct nlattr *attr;
  900. attr = nlmsg_find_attr(nlh, hdrlen,
  901. INET_DIAG_REQ_BYTECODE);
  902. if (attr == NULL ||
  903. nla_len(attr) < sizeof(struct inet_diag_bc_op) ||
  904. inet_diag_bc_audit(nla_data(attr), nla_len(attr)))
  905. return -EINVAL;
  906. }
  907. {
  908. struct netlink_dump_control c = {
  909. .dump = inet_diag_dump_compat,
  910. };
  911. return netlink_dump_start(net->diag_nlsk, skb, nlh, &c);
  912. }
  913. }
  914. return inet_diag_get_exact_compat(skb, nlh);
  915. }
  916. static int inet_diag_handler_dump(struct sk_buff *skb, struct nlmsghdr *h)
  917. {
  918. int hdrlen = sizeof(struct inet_diag_req_v2);
  919. struct net *net = sock_net(skb->sk);
  920. if (nlmsg_len(h) < hdrlen)
  921. return -EINVAL;
  922. if (h->nlmsg_flags & NLM_F_DUMP) {
  923. if (nlmsg_attrlen(h, hdrlen)) {
  924. struct nlattr *attr;
  925. attr = nlmsg_find_attr(h, hdrlen,
  926. INET_DIAG_REQ_BYTECODE);
  927. if (attr == NULL ||
  928. nla_len(attr) < sizeof(struct inet_diag_bc_op) ||
  929. inet_diag_bc_audit(nla_data(attr), nla_len(attr)))
  930. return -EINVAL;
  931. }
  932. {
  933. struct netlink_dump_control c = {
  934. .dump = inet_diag_dump,
  935. };
  936. return netlink_dump_start(net->diag_nlsk, skb, h, &c);
  937. }
  938. }
  939. return inet_diag_get_exact(skb, h, nlmsg_data(h));
  940. }
  941. static const struct sock_diag_handler inet_diag_handler = {
  942. .family = AF_INET,
  943. .dump = inet_diag_handler_dump,
  944. };
  945. static const struct sock_diag_handler inet6_diag_handler = {
  946. .family = AF_INET6,
  947. .dump = inet_diag_handler_dump,
  948. };
  949. int inet_diag_register(const struct inet_diag_handler *h)
  950. {
  951. const __u16 type = h->idiag_type;
  952. int err = -EINVAL;
  953. if (type >= IPPROTO_MAX)
  954. goto out;
  955. mutex_lock(&inet_diag_table_mutex);
  956. err = -EEXIST;
  957. if (inet_diag_table[type] == NULL) {
  958. inet_diag_table[type] = h;
  959. err = 0;
  960. }
  961. mutex_unlock(&inet_diag_table_mutex);
  962. out:
  963. return err;
  964. }
  965. EXPORT_SYMBOL_GPL(inet_diag_register);
  966. void inet_diag_unregister(const struct inet_diag_handler *h)
  967. {
  968. const __u16 type = h->idiag_type;
  969. if (type >= IPPROTO_MAX)
  970. return;
  971. mutex_lock(&inet_diag_table_mutex);
  972. inet_diag_table[type] = NULL;
  973. mutex_unlock(&inet_diag_table_mutex);
  974. }
  975. EXPORT_SYMBOL_GPL(inet_diag_unregister);
  976. static int __init inet_diag_init(void)
  977. {
  978. const int inet_diag_table_size = (IPPROTO_MAX *
  979. sizeof(struct inet_diag_handler *));
  980. int err = -ENOMEM;
  981. inet_diag_table = kzalloc(inet_diag_table_size, GFP_KERNEL);
  982. if (!inet_diag_table)
  983. goto out;
  984. err = sock_diag_register(&inet_diag_handler);
  985. if (err)
  986. goto out_free_nl;
  987. err = sock_diag_register(&inet6_diag_handler);
  988. if (err)
  989. goto out_free_inet;
  990. sock_diag_register_inet_compat(inet_diag_rcv_msg_compat);
  991. out:
  992. return err;
  993. out_free_inet:
  994. sock_diag_unregister(&inet_diag_handler);
  995. out_free_nl:
  996. kfree(inet_diag_table);
  997. goto out;
  998. }
  999. static void __exit inet_diag_exit(void)
  1000. {
  1001. sock_diag_unregister(&inet6_diag_handler);
  1002. sock_diag_unregister(&inet_diag_handler);
  1003. sock_diag_unregister_inet_compat(inet_diag_rcv_msg_compat);
  1004. kfree(inet_diag_table);
  1005. }
  1006. module_init(inet_diag_init);
  1007. module_exit(inet_diag_exit);
  1008. MODULE_LICENSE("GPL");
  1009. MODULE_ALIAS_NET_PF_PROTO_TYPE(PF_NETLINK, NETLINK_SOCK_DIAG, 2 /* AF_INET */);
  1010. MODULE_ALIAS_NET_PF_PROTO_TYPE(PF_NETLINK, NETLINK_SOCK_DIAG, 10 /* AF_INET6 */);