nfnetlink_log.c 28 KB

12345678910111213141516171819202122232425262728293031323334353637383940414243444546474849505152535455565758596061626364656667686970717273747576777879808182838485868788899091929394959697989910010110210310410510610710810911011111211311411511611711811912012112212312412512612712812913013113213313413513613713813914014114214314414514614714814915015115215315415515615715815916016116216316416516616716816917017117217317417517617717817918018118218318418518618718818919019119219319419519619719819920020120220320420520620720820921021121221321421521621721821922022122222322422522622722822923023123223323423523623723823924024124224324424524624724824925025125225325425525625725825926026126226326426526626726826927027127227327427527627727827928028128228328428528628728828929029129229329429529629729829930030130230330430530630730830931031131231331431531631731831932032132232332432532632732832933033133233333433533633733833934034134234334434534634734834935035135235335435535635735835936036136236336436536636736836937037137237337437537637737837938038138238338438538638738838939039139239339439539639739839940040140240340440540640740840941041141241341441541641741841942042142242342442542642742842943043143243343443543643743843944044144244344444544644744844945045145245345445545645745845946046146246346446546646746846947047147247347447547647747847948048148248348448548648748848949049149249349449549649749849950050150250350450550650750850951051151251351451551651751851952052152252352452552652752852953053153253353453553653753853954054154254354454554654754854955055155255355455555655755855956056156256356456556656756856957057157257357457557657757857958058158258358458558658758858959059159259359459559659759859960060160260360460560660760860961061161261361461561661761861962062162262362462562662762862963063163263363463563663763863964064164264364464564664764864965065165265365465565665765865966066166266366466566666766866967067167267367467567667767867968068168268368468568668768868969069169269369469569669769869970070170270370470570670770870971071171271371471571671771871972072172272372472572672772872973073173273373473573673773873974074174274374474574674774874975075175275375475575675775875976076176276376476576676776876977077177277377477577677777877978078178278378478578678778878979079179279379479579679779879980080180280380480580680780880981081181281381481581681781881982082182282382482582682782882983083183283383483583683783883984084184284384484584684784884985085185285385485585685785885986086186286386486586686786886987087187287387487587687787887988088188288388488588688788888989089189289389489589689789889990090190290390490590690790890991091191291391491591691791891992092192292392492592692792892993093193293393493593693793893994094194294394494594694794894995095195295395495595695795895996096196296396496596696796896997097197297397497597697797897998098198298398498598698798898999099199299399499599699799899910001001100210031004100510061007100810091010101110121013101410151016101710181019102010211022102310241025102610271028102910301031103210331034103510361037103810391040104110421043104410451046104710481049105010511052105310541055105610571058105910601061106210631064106510661067106810691070107110721073107410751076107710781079108010811082108310841085108610871088108910901091109210931094109510961097109810991100110111021103110411051106110711081109111011111112111311141115111611171118111911201121112211231124112511261127112811291130113111321133113411351136113711381139114011411142114311441145114611471148114911501151
  1. /*
  2. * This is a module which is used for logging packets to userspace via
  3. * nfetlink.
  4. *
  5. * (C) 2005 by Harald Welte <laforge@netfilter.org>
  6. * (C) 2006-2012 Patrick McHardy <kaber@trash.net>
  7. *
  8. * Based on the old ipv4-only ipt_ULOG.c:
  9. * (C) 2000-2004 by Harald Welte <laforge@netfilter.org>
  10. *
  11. * This program is free software; you can redistribute it and/or modify
  12. * it under the terms of the GNU General Public License version 2 as
  13. * published by the Free Software Foundation.
  14. */
  15. #define pr_fmt(fmt) KBUILD_MODNAME ": " fmt
  16. #include <linux/module.h>
  17. #include <linux/skbuff.h>
  18. #include <linux/if_arp.h>
  19. #include <linux/init.h>
  20. #include <linux/ip.h>
  21. #include <linux/ipv6.h>
  22. #include <linux/netdevice.h>
  23. #include <linux/netfilter.h>
  24. #include <linux/netfilter_bridge.h>
  25. #include <net/netlink.h>
  26. #include <linux/netfilter/nfnetlink.h>
  27. #include <linux/netfilter/nfnetlink_log.h>
  28. #include <linux/netfilter/nf_conntrack_common.h>
  29. #include <linux/spinlock.h>
  30. #include <linux/sysctl.h>
  31. #include <linux/proc_fs.h>
  32. #include <linux/security.h>
  33. #include <linux/list.h>
  34. #include <linux/slab.h>
  35. #include <net/sock.h>
  36. #include <net/netfilter/nf_log.h>
  37. #include <net/netns/generic.h>
  38. #include <linux/atomic.h>
  39. #include <linux/refcount.h>
  40. #if IS_ENABLED(CONFIG_BRIDGE_NETFILTER)
  41. #include "../bridge/br_private.h"
  42. #endif
  43. #define NFULNL_COPY_DISABLED 0xff
  44. #define NFULNL_NLBUFSIZ_DEFAULT NLMSG_GOODSIZE
  45. #define NFULNL_TIMEOUT_DEFAULT 100 /* every second */
  46. #define NFULNL_QTHRESH_DEFAULT 100 /* 100 packets */
  47. /* max packet size is limited by 16-bit struct nfattr nfa_len field */
  48. #define NFULNL_COPY_RANGE_MAX (0xFFFF - NLA_HDRLEN)
  49. #define PRINTR(x, args...) do { if (net_ratelimit()) \
  50. printk(x, ## args); } while (0);
  51. struct nfulnl_instance {
  52. struct hlist_node hlist; /* global list of instances */
  53. spinlock_t lock;
  54. refcount_t use; /* use count */
  55. unsigned int qlen; /* number of nlmsgs in skb */
  56. struct sk_buff *skb; /* pre-allocatd skb */
  57. struct timer_list timer;
  58. struct net *net;
  59. struct user_namespace *peer_user_ns; /* User namespace of the peer process */
  60. u32 peer_portid; /* PORTID of the peer process */
  61. /* configurable parameters */
  62. unsigned int flushtimeout; /* timeout until queue flush */
  63. unsigned int nlbufsiz; /* netlink buffer allocation size */
  64. unsigned int qthreshold; /* threshold of the queue */
  65. u_int32_t copy_range;
  66. u_int32_t seq; /* instance-local sequential counter */
  67. u_int16_t group_num; /* number of this queue */
  68. u_int16_t flags;
  69. u_int8_t copy_mode;
  70. struct rcu_head rcu;
  71. };
  72. #define INSTANCE_BUCKETS 16
  73. static unsigned int nfnl_log_net_id __read_mostly;
  74. struct nfnl_log_net {
  75. spinlock_t instances_lock;
  76. struct hlist_head instance_table[INSTANCE_BUCKETS];
  77. atomic_t global_seq;
  78. };
  79. static struct nfnl_log_net *nfnl_log_pernet(struct net *net)
  80. {
  81. return net_generic(net, nfnl_log_net_id);
  82. }
  83. static inline u_int8_t instance_hashfn(u_int16_t group_num)
  84. {
  85. return ((group_num & 0xff) % INSTANCE_BUCKETS);
  86. }
  87. static struct nfulnl_instance *
  88. __instance_lookup(struct nfnl_log_net *log, u_int16_t group_num)
  89. {
  90. struct hlist_head *head;
  91. struct nfulnl_instance *inst;
  92. head = &log->instance_table[instance_hashfn(group_num)];
  93. hlist_for_each_entry_rcu(inst, head, hlist) {
  94. if (inst->group_num == group_num)
  95. return inst;
  96. }
  97. return NULL;
  98. }
  99. static inline void
  100. instance_get(struct nfulnl_instance *inst)
  101. {
  102. refcount_inc(&inst->use);
  103. }
  104. static struct nfulnl_instance *
  105. instance_lookup_get(struct nfnl_log_net *log, u_int16_t group_num)
  106. {
  107. struct nfulnl_instance *inst;
  108. rcu_read_lock_bh();
  109. inst = __instance_lookup(log, group_num);
  110. if (inst && !refcount_inc_not_zero(&inst->use))
  111. inst = NULL;
  112. rcu_read_unlock_bh();
  113. return inst;
  114. }
  115. static void nfulnl_instance_free_rcu(struct rcu_head *head)
  116. {
  117. struct nfulnl_instance *inst =
  118. container_of(head, struct nfulnl_instance, rcu);
  119. put_net(inst->net);
  120. kfree(inst);
  121. module_put(THIS_MODULE);
  122. }
  123. static void
  124. instance_put(struct nfulnl_instance *inst)
  125. {
  126. if (inst && refcount_dec_and_test(&inst->use))
  127. call_rcu_bh(&inst->rcu, nfulnl_instance_free_rcu);
  128. }
  129. static void nfulnl_timer(struct timer_list *t);
  130. static struct nfulnl_instance *
  131. instance_create(struct net *net, u_int16_t group_num,
  132. u32 portid, struct user_namespace *user_ns)
  133. {
  134. struct nfulnl_instance *inst;
  135. struct nfnl_log_net *log = nfnl_log_pernet(net);
  136. int err;
  137. spin_lock_bh(&log->instances_lock);
  138. if (__instance_lookup(log, group_num)) {
  139. err = -EEXIST;
  140. goto out_unlock;
  141. }
  142. inst = kzalloc(sizeof(*inst), GFP_ATOMIC);
  143. if (!inst) {
  144. err = -ENOMEM;
  145. goto out_unlock;
  146. }
  147. if (!try_module_get(THIS_MODULE)) {
  148. kfree(inst);
  149. err = -EAGAIN;
  150. goto out_unlock;
  151. }
  152. INIT_HLIST_NODE(&inst->hlist);
  153. spin_lock_init(&inst->lock);
  154. /* needs to be two, since we _put() after creation */
  155. refcount_set(&inst->use, 2);
  156. timer_setup(&inst->timer, nfulnl_timer, 0);
  157. inst->net = get_net(net);
  158. inst->peer_user_ns = user_ns;
  159. inst->peer_portid = portid;
  160. inst->group_num = group_num;
  161. inst->qthreshold = NFULNL_QTHRESH_DEFAULT;
  162. inst->flushtimeout = NFULNL_TIMEOUT_DEFAULT;
  163. inst->nlbufsiz = NFULNL_NLBUFSIZ_DEFAULT;
  164. inst->copy_mode = NFULNL_COPY_PACKET;
  165. inst->copy_range = NFULNL_COPY_RANGE_MAX;
  166. hlist_add_head_rcu(&inst->hlist,
  167. &log->instance_table[instance_hashfn(group_num)]);
  168. spin_unlock_bh(&log->instances_lock);
  169. return inst;
  170. out_unlock:
  171. spin_unlock_bh(&log->instances_lock);
  172. return ERR_PTR(err);
  173. }
  174. static void __nfulnl_flush(struct nfulnl_instance *inst);
  175. /* called with BH disabled */
  176. static void
  177. __instance_destroy(struct nfulnl_instance *inst)
  178. {
  179. /* first pull it out of the global list */
  180. hlist_del_rcu(&inst->hlist);
  181. /* then flush all pending packets from skb */
  182. spin_lock(&inst->lock);
  183. /* lockless readers wont be able to use us */
  184. inst->copy_mode = NFULNL_COPY_DISABLED;
  185. if (inst->skb)
  186. __nfulnl_flush(inst);
  187. spin_unlock(&inst->lock);
  188. /* and finally put the refcount */
  189. instance_put(inst);
  190. }
  191. static inline void
  192. instance_destroy(struct nfnl_log_net *log,
  193. struct nfulnl_instance *inst)
  194. {
  195. spin_lock_bh(&log->instances_lock);
  196. __instance_destroy(inst);
  197. spin_unlock_bh(&log->instances_lock);
  198. }
  199. static int
  200. nfulnl_set_mode(struct nfulnl_instance *inst, u_int8_t mode,
  201. unsigned int range)
  202. {
  203. int status = 0;
  204. spin_lock_bh(&inst->lock);
  205. switch (mode) {
  206. case NFULNL_COPY_NONE:
  207. case NFULNL_COPY_META:
  208. inst->copy_mode = mode;
  209. inst->copy_range = 0;
  210. break;
  211. case NFULNL_COPY_PACKET:
  212. inst->copy_mode = mode;
  213. if (range == 0)
  214. range = NFULNL_COPY_RANGE_MAX;
  215. inst->copy_range = min_t(unsigned int,
  216. range, NFULNL_COPY_RANGE_MAX);
  217. break;
  218. default:
  219. status = -EINVAL;
  220. break;
  221. }
  222. spin_unlock_bh(&inst->lock);
  223. return status;
  224. }
  225. static int
  226. nfulnl_set_nlbufsiz(struct nfulnl_instance *inst, u_int32_t nlbufsiz)
  227. {
  228. int status;
  229. spin_lock_bh(&inst->lock);
  230. if (nlbufsiz < NFULNL_NLBUFSIZ_DEFAULT)
  231. status = -ERANGE;
  232. else if (nlbufsiz > 131072)
  233. status = -ERANGE;
  234. else {
  235. inst->nlbufsiz = nlbufsiz;
  236. status = 0;
  237. }
  238. spin_unlock_bh(&inst->lock);
  239. return status;
  240. }
  241. static void
  242. nfulnl_set_timeout(struct nfulnl_instance *inst, u_int32_t timeout)
  243. {
  244. spin_lock_bh(&inst->lock);
  245. inst->flushtimeout = timeout;
  246. spin_unlock_bh(&inst->lock);
  247. }
  248. static void
  249. nfulnl_set_qthresh(struct nfulnl_instance *inst, u_int32_t qthresh)
  250. {
  251. spin_lock_bh(&inst->lock);
  252. inst->qthreshold = qthresh;
  253. spin_unlock_bh(&inst->lock);
  254. }
  255. static int
  256. nfulnl_set_flags(struct nfulnl_instance *inst, u_int16_t flags)
  257. {
  258. spin_lock_bh(&inst->lock);
  259. inst->flags = flags;
  260. spin_unlock_bh(&inst->lock);
  261. return 0;
  262. }
  263. static struct sk_buff *
  264. nfulnl_alloc_skb(struct net *net, u32 peer_portid, unsigned int inst_size,
  265. unsigned int pkt_size)
  266. {
  267. struct sk_buff *skb;
  268. unsigned int n;
  269. /* alloc skb which should be big enough for a whole multipart
  270. * message. WARNING: has to be <= 128k due to slab restrictions */
  271. n = max(inst_size, pkt_size);
  272. skb = alloc_skb(n, GFP_ATOMIC | __GFP_NOWARN);
  273. if (!skb) {
  274. if (n > pkt_size) {
  275. /* try to allocate only as much as we need for current
  276. * packet */
  277. skb = alloc_skb(pkt_size, GFP_ATOMIC);
  278. }
  279. }
  280. return skb;
  281. }
  282. static void
  283. __nfulnl_send(struct nfulnl_instance *inst)
  284. {
  285. if (inst->qlen > 1) {
  286. struct nlmsghdr *nlh = nlmsg_put(inst->skb, 0, 0,
  287. NLMSG_DONE,
  288. sizeof(struct nfgenmsg),
  289. 0);
  290. if (WARN_ONCE(!nlh, "bad nlskb size: %u, tailroom %d\n",
  291. inst->skb->len, skb_tailroom(inst->skb))) {
  292. kfree_skb(inst->skb);
  293. goto out;
  294. }
  295. }
  296. nfnetlink_unicast(inst->skb, inst->net, inst->peer_portid,
  297. MSG_DONTWAIT);
  298. out:
  299. inst->qlen = 0;
  300. inst->skb = NULL;
  301. }
  302. static void
  303. __nfulnl_flush(struct nfulnl_instance *inst)
  304. {
  305. /* timer holds a reference */
  306. if (del_timer(&inst->timer))
  307. instance_put(inst);
  308. if (inst->skb)
  309. __nfulnl_send(inst);
  310. }
  311. static void
  312. nfulnl_timer(struct timer_list *t)
  313. {
  314. struct nfulnl_instance *inst = from_timer(inst, t, timer);
  315. spin_lock_bh(&inst->lock);
  316. if (inst->skb)
  317. __nfulnl_send(inst);
  318. spin_unlock_bh(&inst->lock);
  319. instance_put(inst);
  320. }
  321. /* This is an inline function, we don't really care about a long
  322. * list of arguments */
  323. static inline int
  324. __build_packet_message(struct nfnl_log_net *log,
  325. struct nfulnl_instance *inst,
  326. const struct sk_buff *skb,
  327. unsigned int data_len,
  328. u_int8_t pf,
  329. unsigned int hooknum,
  330. const struct net_device *indev,
  331. const struct net_device *outdev,
  332. const char *prefix, unsigned int plen,
  333. const struct nfnl_ct_hook *nfnl_ct,
  334. struct nf_conn *ct, enum ip_conntrack_info ctinfo)
  335. {
  336. struct nfulnl_msg_packet_hdr pmsg;
  337. struct nlmsghdr *nlh;
  338. struct nfgenmsg *nfmsg;
  339. sk_buff_data_t old_tail = inst->skb->tail;
  340. struct sock *sk;
  341. const unsigned char *hwhdrp;
  342. nlh = nlmsg_put(inst->skb, 0, 0,
  343. nfnl_msg_type(NFNL_SUBSYS_ULOG, NFULNL_MSG_PACKET),
  344. sizeof(struct nfgenmsg), 0);
  345. if (!nlh)
  346. return -1;
  347. nfmsg = nlmsg_data(nlh);
  348. nfmsg->nfgen_family = pf;
  349. nfmsg->version = NFNETLINK_V0;
  350. nfmsg->res_id = htons(inst->group_num);
  351. memset(&pmsg, 0, sizeof(pmsg));
  352. pmsg.hw_protocol = skb->protocol;
  353. pmsg.hook = hooknum;
  354. if (nla_put(inst->skb, NFULA_PACKET_HDR, sizeof(pmsg), &pmsg))
  355. goto nla_put_failure;
  356. if (prefix &&
  357. nla_put(inst->skb, NFULA_PREFIX, plen, prefix))
  358. goto nla_put_failure;
  359. if (indev) {
  360. #if !IS_ENABLED(CONFIG_BRIDGE_NETFILTER)
  361. if (nla_put_be32(inst->skb, NFULA_IFINDEX_INDEV,
  362. htonl(indev->ifindex)))
  363. goto nla_put_failure;
  364. #else
  365. if (pf == PF_BRIDGE) {
  366. /* Case 1: outdev is physical input device, we need to
  367. * look for bridge group (when called from
  368. * netfilter_bridge) */
  369. if (nla_put_be32(inst->skb, NFULA_IFINDEX_PHYSINDEV,
  370. htonl(indev->ifindex)) ||
  371. /* this is the bridge group "brX" */
  372. /* rcu_read_lock()ed by nf_hook_thresh or
  373. * nf_log_packet.
  374. */
  375. nla_put_be32(inst->skb, NFULA_IFINDEX_INDEV,
  376. htonl(br_port_get_rcu(indev)->br->dev->ifindex)))
  377. goto nla_put_failure;
  378. } else {
  379. struct net_device *physindev;
  380. /* Case 2: indev is bridge group, we need to look for
  381. * physical device (when called from ipv4) */
  382. if (nla_put_be32(inst->skb, NFULA_IFINDEX_INDEV,
  383. htonl(indev->ifindex)))
  384. goto nla_put_failure;
  385. physindev = nf_bridge_get_physindev(skb);
  386. if (physindev &&
  387. nla_put_be32(inst->skb, NFULA_IFINDEX_PHYSINDEV,
  388. htonl(physindev->ifindex)))
  389. goto nla_put_failure;
  390. }
  391. #endif
  392. }
  393. if (outdev) {
  394. #if !IS_ENABLED(CONFIG_BRIDGE_NETFILTER)
  395. if (nla_put_be32(inst->skb, NFULA_IFINDEX_OUTDEV,
  396. htonl(outdev->ifindex)))
  397. goto nla_put_failure;
  398. #else
  399. if (pf == PF_BRIDGE) {
  400. /* Case 1: outdev is physical output device, we need to
  401. * look for bridge group (when called from
  402. * netfilter_bridge) */
  403. if (nla_put_be32(inst->skb, NFULA_IFINDEX_PHYSOUTDEV,
  404. htonl(outdev->ifindex)) ||
  405. /* this is the bridge group "brX" */
  406. /* rcu_read_lock()ed by nf_hook_thresh or
  407. * nf_log_packet.
  408. */
  409. nla_put_be32(inst->skb, NFULA_IFINDEX_OUTDEV,
  410. htonl(br_port_get_rcu(outdev)->br->dev->ifindex)))
  411. goto nla_put_failure;
  412. } else {
  413. struct net_device *physoutdev;
  414. /* Case 2: indev is a bridge group, we need to look
  415. * for physical device (when called from ipv4) */
  416. if (nla_put_be32(inst->skb, NFULA_IFINDEX_OUTDEV,
  417. htonl(outdev->ifindex)))
  418. goto nla_put_failure;
  419. physoutdev = nf_bridge_get_physoutdev(skb);
  420. if (physoutdev &&
  421. nla_put_be32(inst->skb, NFULA_IFINDEX_PHYSOUTDEV,
  422. htonl(physoutdev->ifindex)))
  423. goto nla_put_failure;
  424. }
  425. #endif
  426. }
  427. if (skb->mark &&
  428. nla_put_be32(inst->skb, NFULA_MARK, htonl(skb->mark)))
  429. goto nla_put_failure;
  430. if (indev && skb->dev &&
  431. skb->mac_header != skb->network_header) {
  432. struct nfulnl_msg_packet_hw phw;
  433. int len;
  434. memset(&phw, 0, sizeof(phw));
  435. len = dev_parse_header(skb, phw.hw_addr);
  436. if (len > 0) {
  437. phw.hw_addrlen = htons(len);
  438. if (nla_put(inst->skb, NFULA_HWADDR, sizeof(phw), &phw))
  439. goto nla_put_failure;
  440. }
  441. }
  442. if (indev && skb_mac_header_was_set(skb)) {
  443. if (nla_put_be16(inst->skb, NFULA_HWTYPE, htons(skb->dev->type)) ||
  444. nla_put_be16(inst->skb, NFULA_HWLEN,
  445. htons(skb->dev->hard_header_len)))
  446. goto nla_put_failure;
  447. hwhdrp = skb_mac_header(skb);
  448. if (skb->dev->type == ARPHRD_SIT)
  449. hwhdrp -= ETH_HLEN;
  450. if (hwhdrp >= skb->head &&
  451. nla_put(inst->skb, NFULA_HWHEADER,
  452. skb->dev->hard_header_len, hwhdrp))
  453. goto nla_put_failure;
  454. }
  455. if (skb->tstamp) {
  456. struct nfulnl_msg_packet_timestamp ts;
  457. struct timespec64 kts = ktime_to_timespec64(skb->tstamp);
  458. ts.sec = cpu_to_be64(kts.tv_sec);
  459. ts.usec = cpu_to_be64(kts.tv_nsec / NSEC_PER_USEC);
  460. if (nla_put(inst->skb, NFULA_TIMESTAMP, sizeof(ts), &ts))
  461. goto nla_put_failure;
  462. }
  463. /* UID */
  464. sk = skb->sk;
  465. if (sk && sk_fullsock(sk)) {
  466. read_lock_bh(&sk->sk_callback_lock);
  467. if (sk->sk_socket && sk->sk_socket->file) {
  468. struct file *file = sk->sk_socket->file;
  469. const struct cred *cred = file->f_cred;
  470. struct user_namespace *user_ns = inst->peer_user_ns;
  471. __be32 uid = htonl(from_kuid_munged(user_ns, cred->fsuid));
  472. __be32 gid = htonl(from_kgid_munged(user_ns, cred->fsgid));
  473. read_unlock_bh(&sk->sk_callback_lock);
  474. if (nla_put_be32(inst->skb, NFULA_UID, uid) ||
  475. nla_put_be32(inst->skb, NFULA_GID, gid))
  476. goto nla_put_failure;
  477. } else
  478. read_unlock_bh(&sk->sk_callback_lock);
  479. }
  480. /* local sequence number */
  481. if ((inst->flags & NFULNL_CFG_F_SEQ) &&
  482. nla_put_be32(inst->skb, NFULA_SEQ, htonl(inst->seq++)))
  483. goto nla_put_failure;
  484. /* global sequence number */
  485. if ((inst->flags & NFULNL_CFG_F_SEQ_GLOBAL) &&
  486. nla_put_be32(inst->skb, NFULA_SEQ_GLOBAL,
  487. htonl(atomic_inc_return(&log->global_seq))))
  488. goto nla_put_failure;
  489. if (ct && nfnl_ct->build(inst->skb, ct, ctinfo,
  490. NFULA_CT, NFULA_CT_INFO) < 0)
  491. goto nla_put_failure;
  492. if (data_len) {
  493. struct nlattr *nla;
  494. int size = nla_attr_size(data_len);
  495. if (skb_tailroom(inst->skb) < nla_total_size(data_len))
  496. goto nla_put_failure;
  497. nla = skb_put(inst->skb, nla_total_size(data_len));
  498. nla->nla_type = NFULA_PAYLOAD;
  499. nla->nla_len = size;
  500. if (skb_copy_bits(skb, 0, nla_data(nla), data_len))
  501. BUG();
  502. }
  503. nlh->nlmsg_len = inst->skb->tail - old_tail;
  504. return 0;
  505. nla_put_failure:
  506. PRINTR(KERN_ERR "nfnetlink_log: error creating log nlmsg\n");
  507. return -1;
  508. }
  509. static const struct nf_loginfo default_loginfo = {
  510. .type = NF_LOG_TYPE_ULOG,
  511. .u = {
  512. .ulog = {
  513. .copy_len = 0xffff,
  514. .group = 0,
  515. .qthreshold = 1,
  516. },
  517. },
  518. };
  519. /* log handler for internal netfilter logging api */
  520. static void
  521. nfulnl_log_packet(struct net *net,
  522. u_int8_t pf,
  523. unsigned int hooknum,
  524. const struct sk_buff *skb,
  525. const struct net_device *in,
  526. const struct net_device *out,
  527. const struct nf_loginfo *li_user,
  528. const char *prefix)
  529. {
  530. size_t size;
  531. unsigned int data_len;
  532. struct nfulnl_instance *inst;
  533. const struct nf_loginfo *li;
  534. unsigned int qthreshold;
  535. unsigned int plen = 0;
  536. struct nfnl_log_net *log = nfnl_log_pernet(net);
  537. const struct nfnl_ct_hook *nfnl_ct = NULL;
  538. struct nf_conn *ct = NULL;
  539. enum ip_conntrack_info uninitialized_var(ctinfo);
  540. if (li_user && li_user->type == NF_LOG_TYPE_ULOG)
  541. li = li_user;
  542. else
  543. li = &default_loginfo;
  544. inst = instance_lookup_get(log, li->u.ulog.group);
  545. if (!inst)
  546. return;
  547. if (prefix)
  548. plen = strlen(prefix) + 1;
  549. /* FIXME: do we want to make the size calculation conditional based on
  550. * what is actually present? way more branches and checks, but more
  551. * memory efficient... */
  552. size = nlmsg_total_size(sizeof(struct nfgenmsg))
  553. + nla_total_size(sizeof(struct nfulnl_msg_packet_hdr))
  554. + nla_total_size(sizeof(u_int32_t)) /* ifindex */
  555. + nla_total_size(sizeof(u_int32_t)) /* ifindex */
  556. #if IS_ENABLED(CONFIG_BRIDGE_NETFILTER)
  557. + nla_total_size(sizeof(u_int32_t)) /* ifindex */
  558. + nla_total_size(sizeof(u_int32_t)) /* ifindex */
  559. #endif
  560. + nla_total_size(sizeof(u_int32_t)) /* mark */
  561. + nla_total_size(sizeof(u_int32_t)) /* uid */
  562. + nla_total_size(sizeof(u_int32_t)) /* gid */
  563. + nla_total_size(plen) /* prefix */
  564. + nla_total_size(sizeof(struct nfulnl_msg_packet_hw))
  565. + nla_total_size(sizeof(struct nfulnl_msg_packet_timestamp))
  566. + nla_total_size(sizeof(struct nfgenmsg)); /* NLMSG_DONE */
  567. if (in && skb_mac_header_was_set(skb)) {
  568. size += nla_total_size(skb->dev->hard_header_len)
  569. + nla_total_size(sizeof(u_int16_t)) /* hwtype */
  570. + nla_total_size(sizeof(u_int16_t)); /* hwlen */
  571. }
  572. spin_lock_bh(&inst->lock);
  573. if (inst->flags & NFULNL_CFG_F_SEQ)
  574. size += nla_total_size(sizeof(u_int32_t));
  575. if (inst->flags & NFULNL_CFG_F_SEQ_GLOBAL)
  576. size += nla_total_size(sizeof(u_int32_t));
  577. if (inst->flags & NFULNL_CFG_F_CONNTRACK) {
  578. nfnl_ct = rcu_dereference(nfnl_ct_hook);
  579. if (nfnl_ct != NULL) {
  580. ct = nfnl_ct->get_ct(skb, &ctinfo);
  581. if (ct != NULL)
  582. size += nfnl_ct->build_size(ct);
  583. }
  584. }
  585. qthreshold = inst->qthreshold;
  586. /* per-rule qthreshold overrides per-instance */
  587. if (li->u.ulog.qthreshold)
  588. if (qthreshold > li->u.ulog.qthreshold)
  589. qthreshold = li->u.ulog.qthreshold;
  590. switch (inst->copy_mode) {
  591. case NFULNL_COPY_META:
  592. case NFULNL_COPY_NONE:
  593. data_len = 0;
  594. break;
  595. case NFULNL_COPY_PACKET:
  596. data_len = inst->copy_range;
  597. if ((li->u.ulog.flags & NF_LOG_F_COPY_LEN) &&
  598. (li->u.ulog.copy_len < data_len))
  599. data_len = li->u.ulog.copy_len;
  600. if (data_len > skb->len)
  601. data_len = skb->len;
  602. size += nla_total_size(data_len);
  603. break;
  604. case NFULNL_COPY_DISABLED:
  605. default:
  606. goto unlock_and_release;
  607. }
  608. if (inst->skb && size > skb_tailroom(inst->skb)) {
  609. /* either the queue len is too high or we don't have
  610. * enough room in the skb left. flush to userspace. */
  611. __nfulnl_flush(inst);
  612. }
  613. if (!inst->skb) {
  614. inst->skb = nfulnl_alloc_skb(net, inst->peer_portid,
  615. inst->nlbufsiz, size);
  616. if (!inst->skb)
  617. goto alloc_failure;
  618. }
  619. inst->qlen++;
  620. __build_packet_message(log, inst, skb, data_len, pf,
  621. hooknum, in, out, prefix, plen,
  622. nfnl_ct, ct, ctinfo);
  623. if (inst->qlen >= qthreshold)
  624. __nfulnl_flush(inst);
  625. /* timer_pending always called within inst->lock, so there
  626. * is no chance of a race here */
  627. else if (!timer_pending(&inst->timer)) {
  628. instance_get(inst);
  629. inst->timer.expires = jiffies + (inst->flushtimeout*HZ/100);
  630. add_timer(&inst->timer);
  631. }
  632. unlock_and_release:
  633. spin_unlock_bh(&inst->lock);
  634. instance_put(inst);
  635. return;
  636. alloc_failure:
  637. /* FIXME: statistics */
  638. goto unlock_and_release;
  639. }
  640. static int
  641. nfulnl_rcv_nl_event(struct notifier_block *this,
  642. unsigned long event, void *ptr)
  643. {
  644. struct netlink_notify *n = ptr;
  645. struct nfnl_log_net *log = nfnl_log_pernet(n->net);
  646. if (event == NETLINK_URELEASE && n->protocol == NETLINK_NETFILTER) {
  647. int i;
  648. /* destroy all instances for this portid */
  649. spin_lock_bh(&log->instances_lock);
  650. for (i = 0; i < INSTANCE_BUCKETS; i++) {
  651. struct hlist_node *t2;
  652. struct nfulnl_instance *inst;
  653. struct hlist_head *head = &log->instance_table[i];
  654. hlist_for_each_entry_safe(inst, t2, head, hlist) {
  655. if (n->portid == inst->peer_portid)
  656. __instance_destroy(inst);
  657. }
  658. }
  659. spin_unlock_bh(&log->instances_lock);
  660. }
  661. return NOTIFY_DONE;
  662. }
  663. static struct notifier_block nfulnl_rtnl_notifier = {
  664. .notifier_call = nfulnl_rcv_nl_event,
  665. };
  666. static int nfulnl_recv_unsupp(struct net *net, struct sock *ctnl,
  667. struct sk_buff *skb, const struct nlmsghdr *nlh,
  668. const struct nlattr * const nfqa[],
  669. struct netlink_ext_ack *extack)
  670. {
  671. return -ENOTSUPP;
  672. }
  673. static struct nf_logger nfulnl_logger __read_mostly = {
  674. .name = "nfnetlink_log",
  675. .type = NF_LOG_TYPE_ULOG,
  676. .logfn = nfulnl_log_packet,
  677. .me = THIS_MODULE,
  678. };
  679. static const struct nla_policy nfula_cfg_policy[NFULA_CFG_MAX+1] = {
  680. [NFULA_CFG_CMD] = { .len = sizeof(struct nfulnl_msg_config_cmd) },
  681. [NFULA_CFG_MODE] = { .len = sizeof(struct nfulnl_msg_config_mode) },
  682. [NFULA_CFG_TIMEOUT] = { .type = NLA_U32 },
  683. [NFULA_CFG_QTHRESH] = { .type = NLA_U32 },
  684. [NFULA_CFG_NLBUFSIZ] = { .type = NLA_U32 },
  685. [NFULA_CFG_FLAGS] = { .type = NLA_U16 },
  686. };
  687. static int nfulnl_recv_config(struct net *net, struct sock *ctnl,
  688. struct sk_buff *skb, const struct nlmsghdr *nlh,
  689. const struct nlattr * const nfula[],
  690. struct netlink_ext_ack *extack)
  691. {
  692. struct nfgenmsg *nfmsg = nlmsg_data(nlh);
  693. u_int16_t group_num = ntohs(nfmsg->res_id);
  694. struct nfulnl_instance *inst;
  695. struct nfulnl_msg_config_cmd *cmd = NULL;
  696. struct nfnl_log_net *log = nfnl_log_pernet(net);
  697. int ret = 0;
  698. u16 flags = 0;
  699. if (nfula[NFULA_CFG_CMD]) {
  700. u_int8_t pf = nfmsg->nfgen_family;
  701. cmd = nla_data(nfula[NFULA_CFG_CMD]);
  702. /* Commands without queue context */
  703. switch (cmd->command) {
  704. case NFULNL_CFG_CMD_PF_BIND:
  705. return nf_log_bind_pf(net, pf, &nfulnl_logger);
  706. case NFULNL_CFG_CMD_PF_UNBIND:
  707. nf_log_unbind_pf(net, pf);
  708. return 0;
  709. }
  710. }
  711. inst = instance_lookup_get(log, group_num);
  712. if (inst && inst->peer_portid != NETLINK_CB(skb).portid) {
  713. ret = -EPERM;
  714. goto out_put;
  715. }
  716. /* Check if we support these flags in first place, dependencies should
  717. * be there too not to break atomicity.
  718. */
  719. if (nfula[NFULA_CFG_FLAGS]) {
  720. flags = ntohs(nla_get_be16(nfula[NFULA_CFG_FLAGS]));
  721. if ((flags & NFULNL_CFG_F_CONNTRACK) &&
  722. !rcu_access_pointer(nfnl_ct_hook)) {
  723. #ifdef CONFIG_MODULES
  724. nfnl_unlock(NFNL_SUBSYS_ULOG);
  725. request_module("ip_conntrack_netlink");
  726. nfnl_lock(NFNL_SUBSYS_ULOG);
  727. if (rcu_access_pointer(nfnl_ct_hook)) {
  728. ret = -EAGAIN;
  729. goto out_put;
  730. }
  731. #endif
  732. ret = -EOPNOTSUPP;
  733. goto out_put;
  734. }
  735. }
  736. if (cmd != NULL) {
  737. switch (cmd->command) {
  738. case NFULNL_CFG_CMD_BIND:
  739. if (inst) {
  740. ret = -EBUSY;
  741. goto out_put;
  742. }
  743. inst = instance_create(net, group_num,
  744. NETLINK_CB(skb).portid,
  745. sk_user_ns(NETLINK_CB(skb).sk));
  746. if (IS_ERR(inst)) {
  747. ret = PTR_ERR(inst);
  748. goto out;
  749. }
  750. break;
  751. case NFULNL_CFG_CMD_UNBIND:
  752. if (!inst) {
  753. ret = -ENODEV;
  754. goto out;
  755. }
  756. instance_destroy(log, inst);
  757. goto out_put;
  758. default:
  759. ret = -ENOTSUPP;
  760. goto out_put;
  761. }
  762. } else if (!inst) {
  763. ret = -ENODEV;
  764. goto out;
  765. }
  766. if (nfula[NFULA_CFG_MODE]) {
  767. struct nfulnl_msg_config_mode *params =
  768. nla_data(nfula[NFULA_CFG_MODE]);
  769. nfulnl_set_mode(inst, params->copy_mode,
  770. ntohl(params->copy_range));
  771. }
  772. if (nfula[NFULA_CFG_TIMEOUT]) {
  773. __be32 timeout = nla_get_be32(nfula[NFULA_CFG_TIMEOUT]);
  774. nfulnl_set_timeout(inst, ntohl(timeout));
  775. }
  776. if (nfula[NFULA_CFG_NLBUFSIZ]) {
  777. __be32 nlbufsiz = nla_get_be32(nfula[NFULA_CFG_NLBUFSIZ]);
  778. nfulnl_set_nlbufsiz(inst, ntohl(nlbufsiz));
  779. }
  780. if (nfula[NFULA_CFG_QTHRESH]) {
  781. __be32 qthresh = nla_get_be32(nfula[NFULA_CFG_QTHRESH]);
  782. nfulnl_set_qthresh(inst, ntohl(qthresh));
  783. }
  784. if (nfula[NFULA_CFG_FLAGS])
  785. nfulnl_set_flags(inst, flags);
  786. out_put:
  787. instance_put(inst);
  788. out:
  789. return ret;
  790. }
  791. static const struct nfnl_callback nfulnl_cb[NFULNL_MSG_MAX] = {
  792. [NFULNL_MSG_PACKET] = { .call = nfulnl_recv_unsupp,
  793. .attr_count = NFULA_MAX, },
  794. [NFULNL_MSG_CONFIG] = { .call = nfulnl_recv_config,
  795. .attr_count = NFULA_CFG_MAX,
  796. .policy = nfula_cfg_policy },
  797. };
  798. static const struct nfnetlink_subsystem nfulnl_subsys = {
  799. .name = "log",
  800. .subsys_id = NFNL_SUBSYS_ULOG,
  801. .cb_count = NFULNL_MSG_MAX,
  802. .cb = nfulnl_cb,
  803. };
  804. #ifdef CONFIG_PROC_FS
  805. struct iter_state {
  806. struct seq_net_private p;
  807. unsigned int bucket;
  808. };
  809. static struct hlist_node *get_first(struct net *net, struct iter_state *st)
  810. {
  811. struct nfnl_log_net *log;
  812. if (!st)
  813. return NULL;
  814. log = nfnl_log_pernet(net);
  815. for (st->bucket = 0; st->bucket < INSTANCE_BUCKETS; st->bucket++) {
  816. struct hlist_head *head = &log->instance_table[st->bucket];
  817. if (!hlist_empty(head))
  818. return rcu_dereference_bh(hlist_first_rcu(head));
  819. }
  820. return NULL;
  821. }
  822. static struct hlist_node *get_next(struct net *net, struct iter_state *st,
  823. struct hlist_node *h)
  824. {
  825. h = rcu_dereference_bh(hlist_next_rcu(h));
  826. while (!h) {
  827. struct nfnl_log_net *log;
  828. struct hlist_head *head;
  829. if (++st->bucket >= INSTANCE_BUCKETS)
  830. return NULL;
  831. log = nfnl_log_pernet(net);
  832. head = &log->instance_table[st->bucket];
  833. h = rcu_dereference_bh(hlist_first_rcu(head));
  834. }
  835. return h;
  836. }
  837. static struct hlist_node *get_idx(struct net *net, struct iter_state *st,
  838. loff_t pos)
  839. {
  840. struct hlist_node *head;
  841. head = get_first(net, st);
  842. if (head)
  843. while (pos && (head = get_next(net, st, head)))
  844. pos--;
  845. return pos ? NULL : head;
  846. }
  847. static void *seq_start(struct seq_file *s, loff_t *pos)
  848. __acquires(rcu_bh)
  849. {
  850. rcu_read_lock_bh();
  851. return get_idx(seq_file_net(s), s->private, *pos);
  852. }
  853. static void *seq_next(struct seq_file *s, void *v, loff_t *pos)
  854. {
  855. (*pos)++;
  856. return get_next(seq_file_net(s), s->private, v);
  857. }
  858. static void seq_stop(struct seq_file *s, void *v)
  859. __releases(rcu_bh)
  860. {
  861. rcu_read_unlock_bh();
  862. }
  863. static int seq_show(struct seq_file *s, void *v)
  864. {
  865. const struct nfulnl_instance *inst = v;
  866. seq_printf(s, "%5u %6u %5u %1u %5u %6u %2u\n",
  867. inst->group_num,
  868. inst->peer_portid, inst->qlen,
  869. inst->copy_mode, inst->copy_range,
  870. inst->flushtimeout, refcount_read(&inst->use));
  871. return 0;
  872. }
  873. static const struct seq_operations nful_seq_ops = {
  874. .start = seq_start,
  875. .next = seq_next,
  876. .stop = seq_stop,
  877. .show = seq_show,
  878. };
  879. #endif /* PROC_FS */
  880. static int __net_init nfnl_log_net_init(struct net *net)
  881. {
  882. unsigned int i;
  883. struct nfnl_log_net *log = nfnl_log_pernet(net);
  884. #ifdef CONFIG_PROC_FS
  885. struct proc_dir_entry *proc;
  886. kuid_t root_uid;
  887. kgid_t root_gid;
  888. #endif
  889. for (i = 0; i < INSTANCE_BUCKETS; i++)
  890. INIT_HLIST_HEAD(&log->instance_table[i]);
  891. spin_lock_init(&log->instances_lock);
  892. #ifdef CONFIG_PROC_FS
  893. proc = proc_create_net("nfnetlink_log", 0440, net->nf.proc_netfilter,
  894. &nful_seq_ops, sizeof(struct iter_state));
  895. if (!proc)
  896. return -ENOMEM;
  897. root_uid = make_kuid(net->user_ns, 0);
  898. root_gid = make_kgid(net->user_ns, 0);
  899. if (uid_valid(root_uid) && gid_valid(root_gid))
  900. proc_set_user(proc, root_uid, root_gid);
  901. #endif
  902. return 0;
  903. }
  904. static void __net_exit nfnl_log_net_exit(struct net *net)
  905. {
  906. struct nfnl_log_net *log = nfnl_log_pernet(net);
  907. unsigned int i;
  908. #ifdef CONFIG_PROC_FS
  909. remove_proc_entry("nfnetlink_log", net->nf.proc_netfilter);
  910. #endif
  911. nf_log_unset(net, &nfulnl_logger);
  912. for (i = 0; i < INSTANCE_BUCKETS; i++)
  913. WARN_ON_ONCE(!hlist_empty(&log->instance_table[i]));
  914. }
  915. static struct pernet_operations nfnl_log_net_ops = {
  916. .init = nfnl_log_net_init,
  917. .exit = nfnl_log_net_exit,
  918. .id = &nfnl_log_net_id,
  919. .size = sizeof(struct nfnl_log_net),
  920. };
  921. static int __init nfnetlink_log_init(void)
  922. {
  923. int status;
  924. status = register_pernet_subsys(&nfnl_log_net_ops);
  925. if (status < 0) {
  926. pr_err("failed to register pernet ops\n");
  927. goto out;
  928. }
  929. netlink_register_notifier(&nfulnl_rtnl_notifier);
  930. status = nfnetlink_subsys_register(&nfulnl_subsys);
  931. if (status < 0) {
  932. pr_err("failed to create netlink socket\n");
  933. goto cleanup_netlink_notifier;
  934. }
  935. status = nf_log_register(NFPROTO_UNSPEC, &nfulnl_logger);
  936. if (status < 0) {
  937. pr_err("failed to register logger\n");
  938. goto cleanup_subsys;
  939. }
  940. return status;
  941. cleanup_subsys:
  942. nfnetlink_subsys_unregister(&nfulnl_subsys);
  943. cleanup_netlink_notifier:
  944. netlink_unregister_notifier(&nfulnl_rtnl_notifier);
  945. unregister_pernet_subsys(&nfnl_log_net_ops);
  946. out:
  947. return status;
  948. }
  949. static void __exit nfnetlink_log_fini(void)
  950. {
  951. nfnetlink_subsys_unregister(&nfulnl_subsys);
  952. netlink_unregister_notifier(&nfulnl_rtnl_notifier);
  953. unregister_pernet_subsys(&nfnl_log_net_ops);
  954. nf_log_unregister(&nfulnl_logger);
  955. }
  956. MODULE_DESCRIPTION("netfilter userspace logging");
  957. MODULE_AUTHOR("Harald Welte <laforge@netfilter.org>");
  958. MODULE_LICENSE("GPL");
  959. MODULE_ALIAS_NFNL_SUBSYS(NFNL_SUBSYS_ULOG);
  960. MODULE_ALIAS_NF_LOGGER(AF_INET, 1);
  961. MODULE_ALIAS_NF_LOGGER(AF_INET6, 1);
  962. MODULE_ALIAS_NF_LOGGER(AF_BRIDGE, 1);
  963. MODULE_ALIAS_NF_LOGGER(3, 1); /* NFPROTO_ARP */
  964. MODULE_ALIAS_NF_LOGGER(5, 1); /* NFPROTO_NETDEV */
  965. module_init(nfnetlink_log_init);
  966. module_exit(nfnetlink_log_fini);