br_mdb.c 11 KB

123456789101112131415161718192021222324252627282930313233343536373839404142434445464748495051525354555657585960616263646566676869707172737475767778798081828384858687888990919293949596979899100101102103104105106107108109110111112113114115116117118119120121122123124125126127128129130131132133134135136137138139140141142143144145146147148149150151152153154155156157158159160161162163164165166167168169170171172173174175176177178179180181182183184185186187188189190191192193194195196197198199200201202203204205206207208209210211212213214215216217218219220221222223224225226227228229230231232233234235236237238239240241242243244245246247248249250251252253254255256257258259260261262263264265266267268269270271272273274275276277278279280281282283284285286287288289290291292293294295296297298299300301302303304305306307308309310311312313314315316317318319320321322323324325326327328329330331332333334335336337338339340341342343344345346347348349350351352353354355356357358359360361362363364365366367368369370371372373374375376377378379380381382383384385386387388389390391392393394395396397398399400401402403404405406407408409410411412413414415416417418419420421422423424425426427428429430431432433434435436437438439440441442443444445446447448449450451452453454455456457458459460461462463464465466467468469470471472473474475476477478479480481482483484485486487488489490491492493494495496497498499500
  1. #include <linux/err.h>
  2. #include <linux/igmp.h>
  3. #include <linux/kernel.h>
  4. #include <linux/netdevice.h>
  5. #include <linux/rculist.h>
  6. #include <linux/skbuff.h>
  7. #include <linux/if_ether.h>
  8. #include <net/ip.h>
  9. #include <net/netlink.h>
  10. #if IS_ENABLED(CONFIG_IPV6)
  11. #include <net/ipv6.h>
  12. #include <net/addrconf.h>
  13. #endif
  14. #include "br_private.h"
  15. static int br_rports_fill_info(struct sk_buff *skb, struct netlink_callback *cb,
  16. struct net_device *dev)
  17. {
  18. struct net_bridge *br = netdev_priv(dev);
  19. struct net_bridge_port *p;
  20. struct nlattr *nest;
  21. if (!br->multicast_router || hlist_empty(&br->router_list))
  22. return 0;
  23. nest = nla_nest_start(skb, MDBA_ROUTER);
  24. if (nest == NULL)
  25. return -EMSGSIZE;
  26. hlist_for_each_entry_rcu(p, &br->router_list, rlist) {
  27. if (p && nla_put_u32(skb, MDBA_ROUTER_PORT, p->dev->ifindex))
  28. goto fail;
  29. }
  30. nla_nest_end(skb, nest);
  31. return 0;
  32. fail:
  33. nla_nest_cancel(skb, nest);
  34. return -EMSGSIZE;
  35. }
  36. static int br_mdb_fill_info(struct sk_buff *skb, struct netlink_callback *cb,
  37. struct net_device *dev)
  38. {
  39. struct net_bridge *br = netdev_priv(dev);
  40. struct net_bridge_mdb_htable *mdb;
  41. struct nlattr *nest, *nest2;
  42. int i, err = 0;
  43. int idx = 0, s_idx = cb->args[1];
  44. if (br->multicast_disabled)
  45. return 0;
  46. mdb = rcu_dereference(br->mdb);
  47. if (!mdb)
  48. return 0;
  49. nest = nla_nest_start(skb, MDBA_MDB);
  50. if (nest == NULL)
  51. return -EMSGSIZE;
  52. for (i = 0; i < mdb->max; i++) {
  53. struct net_bridge_mdb_entry *mp;
  54. struct net_bridge_port_group *p;
  55. struct net_bridge_port_group __rcu **pp;
  56. struct net_bridge_port *port;
  57. hlist_for_each_entry_rcu(mp, &mdb->mhash[i], hlist[mdb->ver]) {
  58. if (idx < s_idx)
  59. goto skip;
  60. nest2 = nla_nest_start(skb, MDBA_MDB_ENTRY);
  61. if (nest2 == NULL) {
  62. err = -EMSGSIZE;
  63. goto out;
  64. }
  65. for (pp = &mp->ports;
  66. (p = rcu_dereference(*pp)) != NULL;
  67. pp = &p->next) {
  68. port = p->port;
  69. if (port) {
  70. struct br_mdb_entry e;
  71. memset(&e, 0, sizeof(e));
  72. e.ifindex = port->dev->ifindex;
  73. e.state = p->state;
  74. if (p->addr.proto == htons(ETH_P_IP))
  75. e.addr.u.ip4 = p->addr.u.ip4;
  76. #if IS_ENABLED(CONFIG_IPV6)
  77. if (p->addr.proto == htons(ETH_P_IPV6))
  78. e.addr.u.ip6 = p->addr.u.ip6;
  79. #endif
  80. e.addr.proto = p->addr.proto;
  81. if (nla_put(skb, MDBA_MDB_ENTRY_INFO, sizeof(e), &e)) {
  82. nla_nest_cancel(skb, nest2);
  83. err = -EMSGSIZE;
  84. goto out;
  85. }
  86. }
  87. }
  88. nla_nest_end(skb, nest2);
  89. skip:
  90. idx++;
  91. }
  92. }
  93. out:
  94. cb->args[1] = idx;
  95. nla_nest_end(skb, nest);
  96. return err;
  97. }
  98. static int br_mdb_dump(struct sk_buff *skb, struct netlink_callback *cb)
  99. {
  100. struct net_device *dev;
  101. struct net *net = sock_net(skb->sk);
  102. struct nlmsghdr *nlh = NULL;
  103. int idx = 0, s_idx;
  104. s_idx = cb->args[0];
  105. rcu_read_lock();
  106. /* In theory this could be wrapped to 0... */
  107. cb->seq = net->dev_base_seq + br_mdb_rehash_seq;
  108. for_each_netdev_rcu(net, dev) {
  109. if (dev->priv_flags & IFF_EBRIDGE) {
  110. struct br_port_msg *bpm;
  111. if (idx < s_idx)
  112. goto skip;
  113. nlh = nlmsg_put(skb, NETLINK_CB(cb->skb).portid,
  114. cb->nlh->nlmsg_seq, RTM_GETMDB,
  115. sizeof(*bpm), NLM_F_MULTI);
  116. if (nlh == NULL)
  117. break;
  118. bpm = nlmsg_data(nlh);
  119. memset(bpm, 0, sizeof(*bpm));
  120. bpm->ifindex = dev->ifindex;
  121. if (br_mdb_fill_info(skb, cb, dev) < 0)
  122. goto out;
  123. if (br_rports_fill_info(skb, cb, dev) < 0)
  124. goto out;
  125. cb->args[1] = 0;
  126. nlmsg_end(skb, nlh);
  127. skip:
  128. idx++;
  129. }
  130. }
  131. out:
  132. if (nlh)
  133. nlmsg_end(skb, nlh);
  134. rcu_read_unlock();
  135. cb->args[0] = idx;
  136. return skb->len;
  137. }
  138. static int nlmsg_populate_mdb_fill(struct sk_buff *skb,
  139. struct net_device *dev,
  140. struct br_mdb_entry *entry, u32 pid,
  141. u32 seq, int type, unsigned int flags)
  142. {
  143. struct nlmsghdr *nlh;
  144. struct br_port_msg *bpm;
  145. struct nlattr *nest, *nest2;
  146. nlh = nlmsg_put(skb, pid, seq, type, sizeof(*bpm), 0);
  147. if (!nlh)
  148. return -EMSGSIZE;
  149. bpm = nlmsg_data(nlh);
  150. memset(bpm, 0, sizeof(*bpm));
  151. bpm->family = AF_BRIDGE;
  152. bpm->ifindex = dev->ifindex;
  153. nest = nla_nest_start(skb, MDBA_MDB);
  154. if (nest == NULL)
  155. goto cancel;
  156. nest2 = nla_nest_start(skb, MDBA_MDB_ENTRY);
  157. if (nest2 == NULL)
  158. goto end;
  159. if (nla_put(skb, MDBA_MDB_ENTRY_INFO, sizeof(*entry), entry))
  160. goto end;
  161. nla_nest_end(skb, nest2);
  162. nla_nest_end(skb, nest);
  163. nlmsg_end(skb, nlh);
  164. return 0;
  165. end:
  166. nla_nest_end(skb, nest);
  167. cancel:
  168. nlmsg_cancel(skb, nlh);
  169. return -EMSGSIZE;
  170. }
  171. static inline size_t rtnl_mdb_nlmsg_size(void)
  172. {
  173. return NLMSG_ALIGN(sizeof(struct br_port_msg))
  174. + nla_total_size(sizeof(struct br_mdb_entry));
  175. }
  176. static void __br_mdb_notify(struct net_device *dev, struct br_mdb_entry *entry,
  177. int type)
  178. {
  179. struct net *net = dev_net(dev);
  180. struct sk_buff *skb;
  181. int err = -ENOBUFS;
  182. skb = nlmsg_new(rtnl_mdb_nlmsg_size(), GFP_ATOMIC);
  183. if (!skb)
  184. goto errout;
  185. err = nlmsg_populate_mdb_fill(skb, dev, entry, 0, 0, type, NTF_SELF);
  186. if (err < 0) {
  187. kfree_skb(skb);
  188. goto errout;
  189. }
  190. rtnl_notify(skb, net, 0, RTNLGRP_MDB, NULL, GFP_ATOMIC);
  191. return;
  192. errout:
  193. rtnl_set_sk_err(net, RTNLGRP_MDB, err);
  194. }
  195. void br_mdb_notify(struct net_device *dev, struct net_bridge_port *port,
  196. struct br_ip *group, int type)
  197. {
  198. struct br_mdb_entry entry;
  199. memset(&entry, 0, sizeof(entry));
  200. entry.ifindex = port->dev->ifindex;
  201. entry.addr.proto = group->proto;
  202. entry.addr.u.ip4 = group->u.ip4;
  203. #if IS_ENABLED(CONFIG_IPV6)
  204. entry.addr.u.ip6 = group->u.ip6;
  205. #endif
  206. __br_mdb_notify(dev, &entry, type);
  207. }
  208. static bool is_valid_mdb_entry(struct br_mdb_entry *entry)
  209. {
  210. if (entry->ifindex == 0)
  211. return false;
  212. if (entry->addr.proto == htons(ETH_P_IP)) {
  213. if (!ipv4_is_multicast(entry->addr.u.ip4))
  214. return false;
  215. if (ipv4_is_local_multicast(entry->addr.u.ip4))
  216. return false;
  217. #if IS_ENABLED(CONFIG_IPV6)
  218. } else if (entry->addr.proto == htons(ETH_P_IPV6)) {
  219. if (ipv6_addr_is_ll_all_nodes(&entry->addr.u.ip6))
  220. return false;
  221. #endif
  222. } else
  223. return false;
  224. if (entry->state != MDB_PERMANENT && entry->state != MDB_TEMPORARY)
  225. return false;
  226. return true;
  227. }
  228. static int br_mdb_parse(struct sk_buff *skb, struct nlmsghdr *nlh,
  229. struct net_device **pdev, struct br_mdb_entry **pentry)
  230. {
  231. struct net *net = sock_net(skb->sk);
  232. struct br_mdb_entry *entry;
  233. struct br_port_msg *bpm;
  234. struct nlattr *tb[MDBA_SET_ENTRY_MAX+1];
  235. struct net_device *dev;
  236. int err;
  237. err = nlmsg_parse(nlh, sizeof(*bpm), tb, MDBA_SET_ENTRY_MAX, NULL);
  238. if (err < 0)
  239. return err;
  240. bpm = nlmsg_data(nlh);
  241. if (bpm->ifindex == 0) {
  242. pr_info("PF_BRIDGE: br_mdb_parse() with invalid ifindex\n");
  243. return -EINVAL;
  244. }
  245. dev = __dev_get_by_index(net, bpm->ifindex);
  246. if (dev == NULL) {
  247. pr_info("PF_BRIDGE: br_mdb_parse() with unknown ifindex\n");
  248. return -ENODEV;
  249. }
  250. if (!(dev->priv_flags & IFF_EBRIDGE)) {
  251. pr_info("PF_BRIDGE: br_mdb_parse() with non-bridge\n");
  252. return -EOPNOTSUPP;
  253. }
  254. *pdev = dev;
  255. if (!tb[MDBA_SET_ENTRY] ||
  256. nla_len(tb[MDBA_SET_ENTRY]) != sizeof(struct br_mdb_entry)) {
  257. pr_info("PF_BRIDGE: br_mdb_parse() with invalid attr\n");
  258. return -EINVAL;
  259. }
  260. entry = nla_data(tb[MDBA_SET_ENTRY]);
  261. if (!is_valid_mdb_entry(entry)) {
  262. pr_info("PF_BRIDGE: br_mdb_parse() with invalid entry\n");
  263. return -EINVAL;
  264. }
  265. *pentry = entry;
  266. return 0;
  267. }
  268. static int br_mdb_add_group(struct net_bridge *br, struct net_bridge_port *port,
  269. struct br_ip *group, unsigned char state)
  270. {
  271. struct net_bridge_mdb_entry *mp;
  272. struct net_bridge_port_group *p;
  273. struct net_bridge_port_group __rcu **pp;
  274. struct net_bridge_mdb_htable *mdb;
  275. int err;
  276. mdb = mlock_dereference(br->mdb, br);
  277. mp = br_mdb_ip_get(mdb, group);
  278. if (!mp) {
  279. mp = br_multicast_new_group(br, port, group);
  280. err = PTR_ERR(mp);
  281. if (IS_ERR(mp))
  282. return err;
  283. }
  284. for (pp = &mp->ports;
  285. (p = mlock_dereference(*pp, br)) != NULL;
  286. pp = &p->next) {
  287. if (p->port == port)
  288. return -EEXIST;
  289. if ((unsigned long)p->port < (unsigned long)port)
  290. break;
  291. }
  292. p = br_multicast_new_port_group(port, group, *pp, state);
  293. if (unlikely(!p))
  294. return -ENOMEM;
  295. rcu_assign_pointer(*pp, p);
  296. br_mdb_notify(br->dev, port, group, RTM_NEWMDB);
  297. return 0;
  298. }
  299. static int __br_mdb_add(struct net *net, struct net_bridge *br,
  300. struct br_mdb_entry *entry)
  301. {
  302. struct br_ip ip;
  303. struct net_device *dev;
  304. struct net_bridge_port *p;
  305. int ret;
  306. if (!netif_running(br->dev) || br->multicast_disabled)
  307. return -EINVAL;
  308. dev = __dev_get_by_index(net, entry->ifindex);
  309. if (!dev)
  310. return -ENODEV;
  311. p = br_port_get_rtnl(dev);
  312. if (!p || p->br != br || p->state == BR_STATE_DISABLED)
  313. return -EINVAL;
  314. ip.proto = entry->addr.proto;
  315. if (ip.proto == htons(ETH_P_IP))
  316. ip.u.ip4 = entry->addr.u.ip4;
  317. #if IS_ENABLED(CONFIG_IPV6)
  318. else
  319. ip.u.ip6 = entry->addr.u.ip6;
  320. #endif
  321. spin_lock_bh(&br->multicast_lock);
  322. ret = br_mdb_add_group(br, p, &ip, entry->state);
  323. spin_unlock_bh(&br->multicast_lock);
  324. return ret;
  325. }
  326. static int br_mdb_add(struct sk_buff *skb, struct nlmsghdr *nlh)
  327. {
  328. struct net *net = sock_net(skb->sk);
  329. struct br_mdb_entry *entry;
  330. struct net_device *dev;
  331. struct net_bridge *br;
  332. int err;
  333. err = br_mdb_parse(skb, nlh, &dev, &entry);
  334. if (err < 0)
  335. return err;
  336. br = netdev_priv(dev);
  337. err = __br_mdb_add(net, br, entry);
  338. if (!err)
  339. __br_mdb_notify(dev, entry, RTM_NEWMDB);
  340. return err;
  341. }
  342. static int __br_mdb_del(struct net_bridge *br, struct br_mdb_entry *entry)
  343. {
  344. struct net_bridge_mdb_htable *mdb;
  345. struct net_bridge_mdb_entry *mp;
  346. struct net_bridge_port_group *p;
  347. struct net_bridge_port_group __rcu **pp;
  348. struct br_ip ip;
  349. int err = -EINVAL;
  350. if (!netif_running(br->dev) || br->multicast_disabled)
  351. return -EINVAL;
  352. ip.proto = entry->addr.proto;
  353. if (ip.proto == htons(ETH_P_IP)) {
  354. if (timer_pending(&br->ip4_other_query.timer))
  355. return -EBUSY;
  356. ip.u.ip4 = entry->addr.u.ip4;
  357. #if IS_ENABLED(CONFIG_IPV6)
  358. } else {
  359. if (timer_pending(&br->ip6_other_query.timer))
  360. return -EBUSY;
  361. ip.u.ip6 = entry->addr.u.ip6;
  362. #endif
  363. }
  364. spin_lock_bh(&br->multicast_lock);
  365. mdb = mlock_dereference(br->mdb, br);
  366. mp = br_mdb_ip_get(mdb, &ip);
  367. if (!mp)
  368. goto unlock;
  369. for (pp = &mp->ports;
  370. (p = mlock_dereference(*pp, br)) != NULL;
  371. pp = &p->next) {
  372. if (!p->port || p->port->dev->ifindex != entry->ifindex)
  373. continue;
  374. if (p->port->state == BR_STATE_DISABLED)
  375. goto unlock;
  376. rcu_assign_pointer(*pp, p->next);
  377. hlist_del_init(&p->mglist);
  378. del_timer(&p->timer);
  379. call_rcu_bh(&p->rcu, br_multicast_free_pg);
  380. err = 0;
  381. if (!mp->ports && !mp->mglist &&
  382. netif_running(br->dev))
  383. mod_timer(&mp->timer, jiffies);
  384. break;
  385. }
  386. unlock:
  387. spin_unlock_bh(&br->multicast_lock);
  388. return err;
  389. }
  390. static int br_mdb_del(struct sk_buff *skb, struct nlmsghdr *nlh)
  391. {
  392. struct net_device *dev;
  393. struct br_mdb_entry *entry;
  394. struct net_bridge *br;
  395. int err;
  396. err = br_mdb_parse(skb, nlh, &dev, &entry);
  397. if (err < 0)
  398. return err;
  399. br = netdev_priv(dev);
  400. err = __br_mdb_del(br, entry);
  401. if (!err)
  402. __br_mdb_notify(dev, entry, RTM_DELMDB);
  403. return err;
  404. }
  405. void br_mdb_init(void)
  406. {
  407. rtnl_register(PF_BRIDGE, RTM_GETMDB, NULL, br_mdb_dump, NULL);
  408. rtnl_register(PF_BRIDGE, RTM_NEWMDB, br_mdb_add, NULL, NULL);
  409. rtnl_register(PF_BRIDGE, RTM_DELMDB, br_mdb_del, NULL, NULL);
  410. }
  411. void br_mdb_uninit(void)
  412. {
  413. rtnl_unregister(PF_BRIDGE, RTM_GETMDB);
  414. rtnl_unregister(PF_BRIDGE, RTM_NEWMDB);
  415. rtnl_unregister(PF_BRIDGE, RTM_DELMDB);
  416. }