sch_skbprio.c 7.5 KB

123456789101112131415161718192021222324252627282930313233343536373839404142434445464748495051525354555657585960616263646566676869707172737475767778798081828384858687888990919293949596979899100101102103104105106107108109110111112113114115116117118119120121122123124125126127128129130131132133134135136137138139140141142143144145146147148149150151152153154155156157158159160161162163164165166167168169170171172173174175176177178179180181182183184185186187188189190191192193194195196197198199200201202203204205206207208209210211212213214215216217218219220221222223224225226227228229230231232233234235236237238239240241242243244245246247248249250251252253254255256257258259260261262263264265266267268269270271272273274275276277278279280281282283284285286287288289290291292293294295296297298299300301302303304305306307308309310
  1. // SPDX-License-Identifier: GPL-2.0-or-later
  2. /*
  3. * net/sched/sch_skbprio.c SKB Priority Queue.
  4. *
  5. * Authors: Nishanth Devarajan, <ndev2021@gmail.com>
  6. * Cody Doucette, <doucette@bu.edu>
  7. * original idea by Michel Machado, Cody Doucette, and Qiaobin Fu
  8. */
  9. #include <linux/string.h>
  10. #include <linux/module.h>
  11. #include <linux/slab.h>
  12. #include <linux/types.h>
  13. #include <linux/kernel.h>
  14. #include <linux/errno.h>
  15. #include <linux/skbuff.h>
  16. #include <net/pkt_sched.h>
  17. #include <net/sch_generic.h>
  18. #include <net/inet_ecn.h>
  19. /* SKB Priority Queue
  20. * =================================
  21. *
  22. * Skbprio (SKB Priority Queue) is a queueing discipline that prioritizes
  23. * packets according to their skb->priority field. Under congestion,
  24. * Skbprio drops already-enqueued lower priority packets to make space
  25. * available for higher priority packets; it was conceived as a solution
  26. * for denial-of-service defenses that need to route packets with different
  27. * priorities as a mean to overcome DoS attacks.
  28. */
  29. struct skbprio_sched_data {
  30. /* Queue state. */
  31. struct sk_buff_head qdiscs[SKBPRIO_MAX_PRIORITY];
  32. struct gnet_stats_queue qstats[SKBPRIO_MAX_PRIORITY];
  33. u16 highest_prio;
  34. u16 lowest_prio;
  35. };
  36. static u16 calc_new_high_prio(const struct skbprio_sched_data *q)
  37. {
  38. int prio;
  39. for (prio = q->highest_prio - 1; prio >= q->lowest_prio; prio--) {
  40. if (!skb_queue_empty(&q->qdiscs[prio]))
  41. return prio;
  42. }
  43. /* SKB queue is empty, return 0 (default highest priority setting). */
  44. return 0;
  45. }
  46. static u16 calc_new_low_prio(const struct skbprio_sched_data *q)
  47. {
  48. int prio;
  49. for (prio = q->lowest_prio + 1; prio <= q->highest_prio; prio++) {
  50. if (!skb_queue_empty(&q->qdiscs[prio]))
  51. return prio;
  52. }
  53. /* SKB queue is empty, return SKBPRIO_MAX_PRIORITY - 1
  54. * (default lowest priority setting).
  55. */
  56. return SKBPRIO_MAX_PRIORITY - 1;
  57. }
  58. static int skbprio_enqueue(struct sk_buff *skb, struct Qdisc *sch,
  59. struct sk_buff **to_free)
  60. {
  61. const unsigned int max_priority = SKBPRIO_MAX_PRIORITY - 1;
  62. struct skbprio_sched_data *q = qdisc_priv(sch);
  63. struct sk_buff_head *qdisc;
  64. struct sk_buff_head *lp_qdisc;
  65. struct sk_buff *to_drop;
  66. u16 prio, lp;
  67. /* Obtain the priority of @skb. */
  68. prio = min(skb->priority, max_priority);
  69. qdisc = &q->qdiscs[prio];
  70. /* sch->limit can change under us from skbprio_change() */
  71. if (sch->q.qlen < READ_ONCE(sch->limit)) {
  72. __skb_queue_tail(qdisc, skb);
  73. qdisc_qstats_backlog_inc(sch, skb);
  74. q->qstats[prio].backlog += qdisc_pkt_len(skb);
  75. /* Check to update highest and lowest priorities. */
  76. if (prio > q->highest_prio)
  77. q->highest_prio = prio;
  78. if (prio < q->lowest_prio)
  79. q->lowest_prio = prio;
  80. sch->q.qlen++;
  81. return NET_XMIT_SUCCESS;
  82. }
  83. /* If this packet has the lowest priority, drop it. */
  84. lp = q->lowest_prio;
  85. if (prio <= lp) {
  86. q->qstats[prio].drops++;
  87. q->qstats[prio].overlimits++;
  88. return qdisc_drop(skb, sch, to_free);
  89. }
  90. __skb_queue_tail(qdisc, skb);
  91. qdisc_qstats_backlog_inc(sch, skb);
  92. q->qstats[prio].backlog += qdisc_pkt_len(skb);
  93. /* Drop the packet at the tail of the lowest priority qdisc. */
  94. lp_qdisc = &q->qdiscs[lp];
  95. to_drop = __skb_dequeue_tail(lp_qdisc);
  96. BUG_ON(!to_drop);
  97. qdisc_qstats_backlog_dec(sch, to_drop);
  98. qdisc_drop(to_drop, sch, to_free);
  99. q->qstats[lp].backlog -= qdisc_pkt_len(to_drop);
  100. q->qstats[lp].drops++;
  101. q->qstats[lp].overlimits++;
  102. /* Check to update highest and lowest priorities. */
  103. if (skb_queue_empty(lp_qdisc)) {
  104. if (q->lowest_prio == q->highest_prio) {
  105. q->lowest_prio = prio;
  106. q->highest_prio = prio;
  107. } else {
  108. q->lowest_prio = calc_new_low_prio(q);
  109. }
  110. }
  111. if (prio > q->highest_prio)
  112. q->highest_prio = prio;
  113. return NET_XMIT_CN;
  114. }
  115. static struct sk_buff *skbprio_dequeue(struct Qdisc *sch)
  116. {
  117. struct skbprio_sched_data *q = qdisc_priv(sch);
  118. struct sk_buff_head *hpq = &q->qdiscs[q->highest_prio];
  119. struct sk_buff *skb = __skb_dequeue(hpq);
  120. if (unlikely(!skb))
  121. return NULL;
  122. sch->q.qlen--;
  123. qdisc_qstats_backlog_dec(sch, skb);
  124. qdisc_bstats_update(sch, skb);
  125. q->qstats[q->highest_prio].backlog -= qdisc_pkt_len(skb);
  126. /* Update highest priority field. */
  127. if (skb_queue_empty(hpq)) {
  128. if (q->lowest_prio == q->highest_prio) {
  129. q->highest_prio = 0;
  130. q->lowest_prio = SKBPRIO_MAX_PRIORITY - 1;
  131. } else {
  132. q->highest_prio = calc_new_high_prio(q);
  133. }
  134. }
  135. return skb;
  136. }
  137. static int skbprio_change(struct Qdisc *sch, struct nlattr *opt,
  138. struct netlink_ext_ack *extack)
  139. {
  140. struct tc_skbprio_qopt *ctl = nla_data(opt);
  141. if (opt->nla_len != nla_attr_size(sizeof(*ctl)))
  142. return -EINVAL;
  143. WRITE_ONCE(sch->limit, ctl->limit);
  144. return 0;
  145. }
  146. static int skbprio_init(struct Qdisc *sch, struct nlattr *opt,
  147. struct netlink_ext_ack *extack)
  148. {
  149. struct skbprio_sched_data *q = qdisc_priv(sch);
  150. int prio;
  151. /* Initialise all queues, one for each possible priority. */
  152. for (prio = 0; prio < SKBPRIO_MAX_PRIORITY; prio++)
  153. __skb_queue_head_init(&q->qdiscs[prio]);
  154. memset(&q->qstats, 0, sizeof(q->qstats));
  155. q->highest_prio = 0;
  156. q->lowest_prio = SKBPRIO_MAX_PRIORITY - 1;
  157. sch->limit = 64;
  158. if (!opt)
  159. return 0;
  160. return skbprio_change(sch, opt, extack);
  161. }
  162. static int skbprio_dump(struct Qdisc *sch, struct sk_buff *skb)
  163. {
  164. struct tc_skbprio_qopt opt;
  165. opt.limit = READ_ONCE(sch->limit);
  166. if (nla_put(skb, TCA_OPTIONS, sizeof(opt), &opt))
  167. return -1;
  168. return skb->len;
  169. }
  170. static void skbprio_reset(struct Qdisc *sch)
  171. {
  172. struct skbprio_sched_data *q = qdisc_priv(sch);
  173. int prio;
  174. for (prio = 0; prio < SKBPRIO_MAX_PRIORITY; prio++)
  175. __skb_queue_purge(&q->qdiscs[prio]);
  176. memset(&q->qstats, 0, sizeof(q->qstats));
  177. q->highest_prio = 0;
  178. q->lowest_prio = SKBPRIO_MAX_PRIORITY - 1;
  179. }
  180. static void skbprio_destroy(struct Qdisc *sch)
  181. {
  182. struct skbprio_sched_data *q = qdisc_priv(sch);
  183. int prio;
  184. for (prio = 0; prio < SKBPRIO_MAX_PRIORITY; prio++)
  185. __skb_queue_purge(&q->qdiscs[prio]);
  186. }
  187. static struct Qdisc *skbprio_leaf(struct Qdisc *sch, unsigned long arg)
  188. {
  189. return NULL;
  190. }
  191. static unsigned long skbprio_find(struct Qdisc *sch, u32 classid)
  192. {
  193. return 0;
  194. }
  195. static int skbprio_dump_class(struct Qdisc *sch, unsigned long cl,
  196. struct sk_buff *skb, struct tcmsg *tcm)
  197. {
  198. tcm->tcm_handle |= TC_H_MIN(cl);
  199. return 0;
  200. }
  201. static int skbprio_dump_class_stats(struct Qdisc *sch, unsigned long cl,
  202. struct gnet_dump *d)
  203. {
  204. struct skbprio_sched_data *q = qdisc_priv(sch);
  205. if (gnet_stats_copy_queue(d, NULL, &q->qstats[cl - 1],
  206. q->qstats[cl - 1].qlen) < 0)
  207. return -1;
  208. return 0;
  209. }
  210. static void skbprio_walk(struct Qdisc *sch, struct qdisc_walker *arg)
  211. {
  212. unsigned int i;
  213. if (arg->stop)
  214. return;
  215. for (i = 0; i < SKBPRIO_MAX_PRIORITY; i++) {
  216. if (!tc_qdisc_stats_dump(sch, i + 1, arg))
  217. break;
  218. }
  219. }
  220. static const struct Qdisc_class_ops skbprio_class_ops = {
  221. .leaf = skbprio_leaf,
  222. .find = skbprio_find,
  223. .dump = skbprio_dump_class,
  224. .dump_stats = skbprio_dump_class_stats,
  225. .walk = skbprio_walk,
  226. };
  227. static struct Qdisc_ops skbprio_qdisc_ops __read_mostly = {
  228. .cl_ops = &skbprio_class_ops,
  229. .id = "skbprio",
  230. .priv_size = sizeof(struct skbprio_sched_data),
  231. .enqueue = skbprio_enqueue,
  232. .dequeue = skbprio_dequeue,
  233. .peek = qdisc_peek_dequeued,
  234. .init = skbprio_init,
  235. .reset = skbprio_reset,
  236. .change = skbprio_change,
  237. .dump = skbprio_dump,
  238. .destroy = skbprio_destroy,
  239. .owner = THIS_MODULE,
  240. };
  241. MODULE_ALIAS_NET_SCH("skbprio");
  242. static int __init skbprio_module_init(void)
  243. {
  244. return register_qdisc(&skbprio_qdisc_ops);
  245. }
  246. static void __exit skbprio_module_exit(void)
  247. {
  248. unregister_qdisc(&skbprio_qdisc_ops);
  249. }
  250. module_init(skbprio_module_init)
  251. module_exit(skbprio_module_exit)
  252. MODULE_LICENSE("GPL");
  253. MODULE_DESCRIPTION("SKB priority based scheduling qdisc");