sch_skbprio.c 7.8 KB

123456789101112131415161718192021222324252627282930313233343536373839404142434445464748495051525354555657585960616263646566676869707172737475767778798081828384858687888990919293949596979899100101102103104105106107108109110111112113114115116117118119120121122123124125126127128129130131132133134135136137138139140141142143144145146147148149150151152153154155156157158159160161162163164165166167168169170171172173174175176177178179180181182183184185186187188189190191192193194195196197198199200201202203204205206207208209210211212213214215216217218219220221222223224225226227228229230231232233234235236237238239240241242243244245246247248249250251252253254255256257258259260261262263264265266267268269270271272273274275276277278279280281282283284285286287288289290291292293294295296297298299300301302303304305306307308309310311312313314315316317318319320321322323
  1. /*
  2. * net/sched/sch_skbprio.c SKB Priority Queue.
  3. *
  4. * This program is free software; you can redistribute it and/or
  5. * modify it under the terms of the GNU General Public License
  6. * as published by the Free Software Foundation; either version
  7. * 2 of the License, or (at your option) any later version.
  8. *
  9. * Authors: Nishanth Devarajan, <ndev2021@gmail.com>
  10. * Cody Doucette, <doucette@bu.edu>
  11. * original idea by Michel Machado, Cody Doucette, and Qiaobin Fu
  12. */
  13. #include <linux/string.h>
  14. #include <linux/module.h>
  15. #include <linux/slab.h>
  16. #include <linux/types.h>
  17. #include <linux/kernel.h>
  18. #include <linux/errno.h>
  19. #include <linux/skbuff.h>
  20. #include <net/pkt_sched.h>
  21. #include <net/sch_generic.h>
  22. #include <net/inet_ecn.h>
  23. /* SKB Priority Queue
  24. * =================================
  25. *
  26. * Skbprio (SKB Priority Queue) is a queueing discipline that prioritizes
  27. * packets according to their skb->priority field. Under congestion,
  28. * Skbprio drops already-enqueued lower priority packets to make space
  29. * available for higher priority packets; it was conceived as a solution
  30. * for denial-of-service defenses that need to route packets with different
  31. * priorities as a mean to overcome DoS attacks.
  32. */
  33. struct skbprio_sched_data {
  34. /* Queue state. */
  35. struct sk_buff_head qdiscs[SKBPRIO_MAX_PRIORITY];
  36. struct gnet_stats_queue qstats[SKBPRIO_MAX_PRIORITY];
  37. u16 highest_prio;
  38. u16 lowest_prio;
  39. };
  40. static u16 calc_new_high_prio(const struct skbprio_sched_data *q)
  41. {
  42. int prio;
  43. for (prio = q->highest_prio - 1; prio >= q->lowest_prio; prio--) {
  44. if (!skb_queue_empty(&q->qdiscs[prio]))
  45. return prio;
  46. }
  47. /* SKB queue is empty, return 0 (default highest priority setting). */
  48. return 0;
  49. }
  50. static u16 calc_new_low_prio(const struct skbprio_sched_data *q)
  51. {
  52. int prio;
  53. for (prio = q->lowest_prio + 1; prio <= q->highest_prio; prio++) {
  54. if (!skb_queue_empty(&q->qdiscs[prio]))
  55. return prio;
  56. }
  57. /* SKB queue is empty, return SKBPRIO_MAX_PRIORITY - 1
  58. * (default lowest priority setting).
  59. */
  60. return SKBPRIO_MAX_PRIORITY - 1;
  61. }
  62. static int skbprio_enqueue(struct sk_buff *skb, struct Qdisc *sch,
  63. struct sk_buff **to_free)
  64. {
  65. const unsigned int max_priority = SKBPRIO_MAX_PRIORITY - 1;
  66. struct skbprio_sched_data *q = qdisc_priv(sch);
  67. struct sk_buff_head *qdisc;
  68. struct sk_buff_head *lp_qdisc;
  69. struct sk_buff *to_drop;
  70. u16 prio, lp;
  71. /* Obtain the priority of @skb. */
  72. prio = min(skb->priority, max_priority);
  73. qdisc = &q->qdiscs[prio];
  74. if (sch->q.qlen < sch->limit) {
  75. __skb_queue_tail(qdisc, skb);
  76. qdisc_qstats_backlog_inc(sch, skb);
  77. q->qstats[prio].backlog += qdisc_pkt_len(skb);
  78. /* Check to update highest and lowest priorities. */
  79. if (prio > q->highest_prio)
  80. q->highest_prio = prio;
  81. if (prio < q->lowest_prio)
  82. q->lowest_prio = prio;
  83. sch->q.qlen++;
  84. return NET_XMIT_SUCCESS;
  85. }
  86. /* If this packet has the lowest priority, drop it. */
  87. lp = q->lowest_prio;
  88. if (prio <= lp) {
  89. q->qstats[prio].drops++;
  90. q->qstats[prio].overlimits++;
  91. return qdisc_drop(skb, sch, to_free);
  92. }
  93. __skb_queue_tail(qdisc, skb);
  94. qdisc_qstats_backlog_inc(sch, skb);
  95. q->qstats[prio].backlog += qdisc_pkt_len(skb);
  96. /* Drop the packet at the tail of the lowest priority qdisc. */
  97. lp_qdisc = &q->qdiscs[lp];
  98. to_drop = __skb_dequeue_tail(lp_qdisc);
  99. BUG_ON(!to_drop);
  100. qdisc_qstats_backlog_dec(sch, to_drop);
  101. qdisc_drop(to_drop, sch, to_free);
  102. q->qstats[lp].backlog -= qdisc_pkt_len(to_drop);
  103. q->qstats[lp].drops++;
  104. q->qstats[lp].overlimits++;
  105. /* Check to update highest and lowest priorities. */
  106. if (skb_queue_empty(lp_qdisc)) {
  107. if (q->lowest_prio == q->highest_prio) {
  108. /* The incoming packet is the only packet in queue. */
  109. BUG_ON(sch->q.qlen != 1);
  110. q->lowest_prio = prio;
  111. q->highest_prio = prio;
  112. } else {
  113. q->lowest_prio = calc_new_low_prio(q);
  114. }
  115. }
  116. if (prio > q->highest_prio)
  117. q->highest_prio = prio;
  118. return NET_XMIT_CN;
  119. }
  120. static struct sk_buff *skbprio_dequeue(struct Qdisc *sch)
  121. {
  122. struct skbprio_sched_data *q = qdisc_priv(sch);
  123. struct sk_buff_head *hpq = &q->qdiscs[q->highest_prio];
  124. struct sk_buff *skb = __skb_dequeue(hpq);
  125. if (unlikely(!skb))
  126. return NULL;
  127. sch->q.qlen--;
  128. qdisc_qstats_backlog_dec(sch, skb);
  129. qdisc_bstats_update(sch, skb);
  130. q->qstats[q->highest_prio].backlog -= qdisc_pkt_len(skb);
  131. /* Update highest priority field. */
  132. if (skb_queue_empty(hpq)) {
  133. if (q->lowest_prio == q->highest_prio) {
  134. BUG_ON(sch->q.qlen);
  135. q->highest_prio = 0;
  136. q->lowest_prio = SKBPRIO_MAX_PRIORITY - 1;
  137. } else {
  138. q->highest_prio = calc_new_high_prio(q);
  139. }
  140. }
  141. return skb;
  142. }
  143. static int skbprio_change(struct Qdisc *sch, struct nlattr *opt,
  144. struct netlink_ext_ack *extack)
  145. {
  146. struct tc_skbprio_qopt *ctl = nla_data(opt);
  147. if (opt->nla_len != nla_attr_size(sizeof(*ctl)))
  148. return -EINVAL;
  149. sch->limit = ctl->limit;
  150. return 0;
  151. }
  152. static int skbprio_init(struct Qdisc *sch, struct nlattr *opt,
  153. struct netlink_ext_ack *extack)
  154. {
  155. struct skbprio_sched_data *q = qdisc_priv(sch);
  156. int prio;
  157. /* Initialise all queues, one for each possible priority. */
  158. for (prio = 0; prio < SKBPRIO_MAX_PRIORITY; prio++)
  159. __skb_queue_head_init(&q->qdiscs[prio]);
  160. memset(&q->qstats, 0, sizeof(q->qstats));
  161. q->highest_prio = 0;
  162. q->lowest_prio = SKBPRIO_MAX_PRIORITY - 1;
  163. sch->limit = 64;
  164. if (!opt)
  165. return 0;
  166. return skbprio_change(sch, opt, extack);
  167. }
  168. static int skbprio_dump(struct Qdisc *sch, struct sk_buff *skb)
  169. {
  170. struct tc_skbprio_qopt opt;
  171. opt.limit = sch->limit;
  172. if (nla_put(skb, TCA_OPTIONS, sizeof(opt), &opt))
  173. return -1;
  174. return skb->len;
  175. }
  176. static void skbprio_reset(struct Qdisc *sch)
  177. {
  178. struct skbprio_sched_data *q = qdisc_priv(sch);
  179. int prio;
  180. sch->qstats.backlog = 0;
  181. sch->q.qlen = 0;
  182. for (prio = 0; prio < SKBPRIO_MAX_PRIORITY; prio++)
  183. __skb_queue_purge(&q->qdiscs[prio]);
  184. memset(&q->qstats, 0, sizeof(q->qstats));
  185. q->highest_prio = 0;
  186. q->lowest_prio = SKBPRIO_MAX_PRIORITY - 1;
  187. }
  188. static void skbprio_destroy(struct Qdisc *sch)
  189. {
  190. struct skbprio_sched_data *q = qdisc_priv(sch);
  191. int prio;
  192. for (prio = 0; prio < SKBPRIO_MAX_PRIORITY; prio++)
  193. __skb_queue_purge(&q->qdiscs[prio]);
  194. }
  195. static struct Qdisc *skbprio_leaf(struct Qdisc *sch, unsigned long arg)
  196. {
  197. return NULL;
  198. }
  199. static unsigned long skbprio_find(struct Qdisc *sch, u32 classid)
  200. {
  201. return 0;
  202. }
  203. static int skbprio_dump_class(struct Qdisc *sch, unsigned long cl,
  204. struct sk_buff *skb, struct tcmsg *tcm)
  205. {
  206. tcm->tcm_handle |= TC_H_MIN(cl);
  207. return 0;
  208. }
  209. static int skbprio_dump_class_stats(struct Qdisc *sch, unsigned long cl,
  210. struct gnet_dump *d)
  211. {
  212. struct skbprio_sched_data *q = qdisc_priv(sch);
  213. if (gnet_stats_copy_queue(d, NULL, &q->qstats[cl - 1],
  214. q->qstats[cl - 1].qlen) < 0)
  215. return -1;
  216. return 0;
  217. }
  218. static void skbprio_walk(struct Qdisc *sch, struct qdisc_walker *arg)
  219. {
  220. unsigned int i;
  221. if (arg->stop)
  222. return;
  223. for (i = 0; i < SKBPRIO_MAX_PRIORITY; i++) {
  224. if (arg->count < arg->skip) {
  225. arg->count++;
  226. continue;
  227. }
  228. if (arg->fn(sch, i + 1, arg) < 0) {
  229. arg->stop = 1;
  230. break;
  231. }
  232. arg->count++;
  233. }
  234. }
  235. static const struct Qdisc_class_ops skbprio_class_ops = {
  236. .leaf = skbprio_leaf,
  237. .find = skbprio_find,
  238. .dump = skbprio_dump_class,
  239. .dump_stats = skbprio_dump_class_stats,
  240. .walk = skbprio_walk,
  241. };
  242. static struct Qdisc_ops skbprio_qdisc_ops __read_mostly = {
  243. .cl_ops = &skbprio_class_ops,
  244. .id = "skbprio",
  245. .priv_size = sizeof(struct skbprio_sched_data),
  246. .enqueue = skbprio_enqueue,
  247. .dequeue = skbprio_dequeue,
  248. .peek = qdisc_peek_dequeued,
  249. .init = skbprio_init,
  250. .reset = skbprio_reset,
  251. .change = skbprio_change,
  252. .dump = skbprio_dump,
  253. .destroy = skbprio_destroy,
  254. .owner = THIS_MODULE,
  255. };
  256. static int __init skbprio_module_init(void)
  257. {
  258. return register_qdisc(&skbprio_qdisc_ops);
  259. }
  260. static void __exit skbprio_module_exit(void)
  261. {
  262. unregister_qdisc(&skbprio_qdisc_ops);
  263. }
  264. module_init(skbprio_module_init)
  265. module_exit(skbprio_module_exit)
  266. MODULE_LICENSE("GPL");