2 * Copyright (c) 2008-2011, Intel Corporation.
4 * This program is free software; you can redistribute it and/or modify it
5 * under the terms and conditions of the GNU General Public License,
6 * version 2, as published by the Free Software Foundation.
8 * This program is distributed in the hope it will be useful, but WITHOUT
9 * ANY WARRANTY; without even the implied warranty of MERCHANTABILITY or
10 * FITNESS FOR A PARTICULAR PURPOSE. See the GNU General Public License for
13 * You should have received a copy of the GNU General Public License along with
14 * this program; if not, write to the Free Software Foundation, Inc., 59 Temple
15 * Place - Suite 330, Boston, MA 02111-1307 USA.
17 * Author: Lucy Liu <lucy.liu@intel.com>
20 #include <linux/netdevice.h>
21 #include <linux/netlink.h>
22 #include <linux/slab.h>
23 #include <net/netlink.h>
24 #include <net/rtnetlink.h>
25 #include <linux/dcbnl.h>
26 #include <net/dcbevent.h>
27 #include <linux/rtnetlink.h>
28 #include <linux/module.h>
32 * Data Center Bridging (DCB) is a collection of Ethernet enhancements
33 * intended to allow network traffic with differing requirements
34 * (highly reliable, no drops vs. best effort vs. low latency) to operate
35 * and co-exist on Ethernet. Current DCB features are:
37 * Enhanced Transmission Selection (aka Priority Grouping [PG]) - provides a
38 * framework for assigning bandwidth guarantees to traffic classes.
40 * Priority-based Flow Control (PFC) - provides a flow control mechanism which
41 * can work independently for each 802.1p priority.
43 * Congestion Notification - provides a mechanism for end-to-end congestion
44 * control for protocols which do not have built-in congestion management.
46 * More information about the emerging standards for these Ethernet features
47 * can be found at: http://www.ieee802.org/1/pages/dcbridges.html
49 * This file implements an rtnetlink interface to allow configuration of DCB
50 * features for capable devices.
53 MODULE_AUTHOR("Lucy Liu, <lucy.liu@intel.com>");
54 MODULE_DESCRIPTION("Data Center Bridging netlink interface");
55 MODULE_LICENSE("GPL");
57 /**************** DCB attribute policies *************************************/
59 /* DCB netlink attributes policy */
60 static const struct nla_policy dcbnl_rtnl_policy[DCB_ATTR_MAX + 1] = {
61 [DCB_ATTR_IFNAME] = {.type = NLA_NUL_STRING, .len = IFNAMSIZ - 1},
62 [DCB_ATTR_STATE] = {.type = NLA_U8},
63 [DCB_ATTR_PFC_CFG] = {.type = NLA_NESTED},
64 [DCB_ATTR_PG_CFG] = {.type = NLA_NESTED},
65 [DCB_ATTR_SET_ALL] = {.type = NLA_U8},
66 [DCB_ATTR_PERM_HWADDR] = {.type = NLA_FLAG},
67 [DCB_ATTR_CAP] = {.type = NLA_NESTED},
68 [DCB_ATTR_PFC_STATE] = {.type = NLA_U8},
69 [DCB_ATTR_BCN] = {.type = NLA_NESTED},
70 [DCB_ATTR_APP] = {.type = NLA_NESTED},
71 [DCB_ATTR_IEEE] = {.type = NLA_NESTED},
72 [DCB_ATTR_DCBX] = {.type = NLA_U8},
73 [DCB_ATTR_FEATCFG] = {.type = NLA_NESTED},
76 /* DCB priority flow control to User Priority nested attributes */
77 static const struct nla_policy dcbnl_pfc_up_nest[DCB_PFC_UP_ATTR_MAX + 1] = {
78 [DCB_PFC_UP_ATTR_0] = {.type = NLA_U8},
79 [DCB_PFC_UP_ATTR_1] = {.type = NLA_U8},
80 [DCB_PFC_UP_ATTR_2] = {.type = NLA_U8},
81 [DCB_PFC_UP_ATTR_3] = {.type = NLA_U8},
82 [DCB_PFC_UP_ATTR_4] = {.type = NLA_U8},
83 [DCB_PFC_UP_ATTR_5] = {.type = NLA_U8},
84 [DCB_PFC_UP_ATTR_6] = {.type = NLA_U8},
85 [DCB_PFC_UP_ATTR_7] = {.type = NLA_U8},
86 [DCB_PFC_UP_ATTR_ALL] = {.type = NLA_FLAG},
89 /* DCB priority grouping nested attributes */
90 static const struct nla_policy dcbnl_pg_nest[DCB_PG_ATTR_MAX + 1] = {
91 [DCB_PG_ATTR_TC_0] = {.type = NLA_NESTED},
92 [DCB_PG_ATTR_TC_1] = {.type = NLA_NESTED},
93 [DCB_PG_ATTR_TC_2] = {.type = NLA_NESTED},
94 [DCB_PG_ATTR_TC_3] = {.type = NLA_NESTED},
95 [DCB_PG_ATTR_TC_4] = {.type = NLA_NESTED},
96 [DCB_PG_ATTR_TC_5] = {.type = NLA_NESTED},
97 [DCB_PG_ATTR_TC_6] = {.type = NLA_NESTED},
98 [DCB_PG_ATTR_TC_7] = {.type = NLA_NESTED},
99 [DCB_PG_ATTR_TC_ALL] = {.type = NLA_NESTED},
100 [DCB_PG_ATTR_BW_ID_0] = {.type = NLA_U8},
101 [DCB_PG_ATTR_BW_ID_1] = {.type = NLA_U8},
102 [DCB_PG_ATTR_BW_ID_2] = {.type = NLA_U8},
103 [DCB_PG_ATTR_BW_ID_3] = {.type = NLA_U8},
104 [DCB_PG_ATTR_BW_ID_4] = {.type = NLA_U8},
105 [DCB_PG_ATTR_BW_ID_5] = {.type = NLA_U8},
106 [DCB_PG_ATTR_BW_ID_6] = {.type = NLA_U8},
107 [DCB_PG_ATTR_BW_ID_7] = {.type = NLA_U8},
108 [DCB_PG_ATTR_BW_ID_ALL] = {.type = NLA_FLAG},
111 /* DCB traffic class nested attributes. */
112 static const struct nla_policy dcbnl_tc_param_nest[DCB_TC_ATTR_PARAM_MAX + 1] = {
113 [DCB_TC_ATTR_PARAM_PGID] = {.type = NLA_U8},
114 [DCB_TC_ATTR_PARAM_UP_MAPPING] = {.type = NLA_U8},
115 [DCB_TC_ATTR_PARAM_STRICT_PRIO] = {.type = NLA_U8},
116 [DCB_TC_ATTR_PARAM_BW_PCT] = {.type = NLA_U8},
117 [DCB_TC_ATTR_PARAM_ALL] = {.type = NLA_FLAG},
120 /* DCB capabilities nested attributes. */
121 static const struct nla_policy dcbnl_cap_nest[DCB_CAP_ATTR_MAX + 1] = {
122 [DCB_CAP_ATTR_ALL] = {.type = NLA_FLAG},
123 [DCB_CAP_ATTR_PG] = {.type = NLA_U8},
124 [DCB_CAP_ATTR_PFC] = {.type = NLA_U8},
125 [DCB_CAP_ATTR_UP2TC] = {.type = NLA_U8},
126 [DCB_CAP_ATTR_PG_TCS] = {.type = NLA_U8},
127 [DCB_CAP_ATTR_PFC_TCS] = {.type = NLA_U8},
128 [DCB_CAP_ATTR_GSP] = {.type = NLA_U8},
129 [DCB_CAP_ATTR_BCN] = {.type = NLA_U8},
130 [DCB_CAP_ATTR_DCBX] = {.type = NLA_U8},
133 /* DCB capabilities nested attributes. */
134 static const struct nla_policy dcbnl_numtcs_nest[DCB_NUMTCS_ATTR_MAX + 1] = {
135 [DCB_NUMTCS_ATTR_ALL] = {.type = NLA_FLAG},
136 [DCB_NUMTCS_ATTR_PG] = {.type = NLA_U8},
137 [DCB_NUMTCS_ATTR_PFC] = {.type = NLA_U8},
140 /* DCB BCN nested attributes. */
141 static const struct nla_policy dcbnl_bcn_nest[DCB_BCN_ATTR_MAX + 1] = {
142 [DCB_BCN_ATTR_RP_0] = {.type = NLA_U8},
143 [DCB_BCN_ATTR_RP_1] = {.type = NLA_U8},
144 [DCB_BCN_ATTR_RP_2] = {.type = NLA_U8},
145 [DCB_BCN_ATTR_RP_3] = {.type = NLA_U8},
146 [DCB_BCN_ATTR_RP_4] = {.type = NLA_U8},
147 [DCB_BCN_ATTR_RP_5] = {.type = NLA_U8},
148 [DCB_BCN_ATTR_RP_6] = {.type = NLA_U8},
149 [DCB_BCN_ATTR_RP_7] = {.type = NLA_U8},
150 [DCB_BCN_ATTR_RP_ALL] = {.type = NLA_FLAG},
151 [DCB_BCN_ATTR_BCNA_0] = {.type = NLA_U32},
152 [DCB_BCN_ATTR_BCNA_1] = {.type = NLA_U32},
153 [DCB_BCN_ATTR_ALPHA] = {.type = NLA_U32},
154 [DCB_BCN_ATTR_BETA] = {.type = NLA_U32},
155 [DCB_BCN_ATTR_GD] = {.type = NLA_U32},
156 [DCB_BCN_ATTR_GI] = {.type = NLA_U32},
157 [DCB_BCN_ATTR_TMAX] = {.type = NLA_U32},
158 [DCB_BCN_ATTR_TD] = {.type = NLA_U32},
159 [DCB_BCN_ATTR_RMIN] = {.type = NLA_U32},
160 [DCB_BCN_ATTR_W] = {.type = NLA_U32},
161 [DCB_BCN_ATTR_RD] = {.type = NLA_U32},
162 [DCB_BCN_ATTR_RU] = {.type = NLA_U32},
163 [DCB_BCN_ATTR_WRTT] = {.type = NLA_U32},
164 [DCB_BCN_ATTR_RI] = {.type = NLA_U32},
165 [DCB_BCN_ATTR_C] = {.type = NLA_U32},
166 [DCB_BCN_ATTR_ALL] = {.type = NLA_FLAG},
169 /* DCB APP nested attributes. */
170 static const struct nla_policy dcbnl_app_nest[DCB_APP_ATTR_MAX + 1] = {
171 [DCB_APP_ATTR_IDTYPE] = {.type = NLA_U8},
172 [DCB_APP_ATTR_ID] = {.type = NLA_U16},
173 [DCB_APP_ATTR_PRIORITY] = {.type = NLA_U8},
176 /* IEEE 802.1Qaz nested attributes. */
177 static const struct nla_policy dcbnl_ieee_policy[DCB_ATTR_IEEE_MAX + 1] = {
178 [DCB_ATTR_IEEE_ETS] = {.len = sizeof(struct ieee_ets)},
179 [DCB_ATTR_IEEE_PFC] = {.len = sizeof(struct ieee_pfc)},
180 [DCB_ATTR_IEEE_APP_TABLE] = {.type = NLA_NESTED},
181 [DCB_ATTR_IEEE_MAXRATE] = {.len = sizeof(struct ieee_maxrate)},
184 static const struct nla_policy dcbnl_ieee_app[DCB_ATTR_IEEE_APP_MAX + 1] = {
185 [DCB_ATTR_IEEE_APP] = {.len = sizeof(struct dcb_app)},
188 /* DCB number of traffic classes nested attributes. */
189 static const struct nla_policy dcbnl_featcfg_nest[DCB_FEATCFG_ATTR_MAX + 1] = {
190 [DCB_FEATCFG_ATTR_ALL] = {.type = NLA_FLAG},
191 [DCB_FEATCFG_ATTR_PG] = {.type = NLA_U8},
192 [DCB_FEATCFG_ATTR_PFC] = {.type = NLA_U8},
193 [DCB_FEATCFG_ATTR_APP] = {.type = NLA_U8},
196 static LIST_HEAD(dcb_app_list);
197 static DEFINE_SPINLOCK(dcb_lock);
199 static struct sk_buff *dcbnl_newmsg(int type, u8 cmd, u32 port, u32 seq,
200 u32 flags, struct nlmsghdr **nlhp)
204 struct nlmsghdr *nlh;
206 skb = nlmsg_new(NLMSG_DEFAULT_SIZE, GFP_KERNEL);
210 nlh = nlmsg_put(skb, port, seq, type, sizeof(*dcb), flags);
213 dcb = nlmsg_data(nlh);
214 dcb->dcb_family = AF_UNSPEC;
224 static int dcbnl_getstate(struct net_device *netdev, struct nlmsghdr *nlh,
225 u32 seq, struct nlattr **tb, struct sk_buff *skb)
227 /* if (!tb[DCB_ATTR_STATE] || !netdev->dcbnl_ops->getstate) */
228 if (!netdev->dcbnl_ops->getstate)
231 return nla_put_u8(skb, DCB_ATTR_STATE,
232 netdev->dcbnl_ops->getstate(netdev));
235 static int dcbnl_getpfccfg(struct net_device *netdev, struct nlmsghdr *nlh,
236 u32 seq, struct nlattr **tb, struct sk_buff *skb)
238 struct nlattr *data[DCB_PFC_UP_ATTR_MAX + 1], *nest;
244 if (!tb[DCB_ATTR_PFC_CFG])
247 if (!netdev->dcbnl_ops->getpfccfg)
250 ret = nla_parse_nested(data, DCB_PFC_UP_ATTR_MAX,
251 tb[DCB_ATTR_PFC_CFG],
256 nest = nla_nest_start(skb, DCB_ATTR_PFC_CFG);
260 if (data[DCB_PFC_UP_ATTR_ALL])
263 for (i = DCB_PFC_UP_ATTR_0; i <= DCB_PFC_UP_ATTR_7; i++) {
264 if (!getall && !data[i])
267 netdev->dcbnl_ops->getpfccfg(netdev, i - DCB_PFC_UP_ATTR_0,
269 ret = nla_put_u8(skb, i, value);
271 nla_nest_cancel(skb, nest);
275 nla_nest_end(skb, nest);
280 static int dcbnl_getperm_hwaddr(struct net_device *netdev, struct nlmsghdr *nlh,
281 u32 seq, struct nlattr **tb, struct sk_buff *skb)
283 u8 perm_addr[MAX_ADDR_LEN];
285 if (!netdev->dcbnl_ops->getpermhwaddr)
288 netdev->dcbnl_ops->getpermhwaddr(netdev, perm_addr);
290 return nla_put(skb, DCB_ATTR_PERM_HWADDR, sizeof(perm_addr), perm_addr);
293 static int dcbnl_getcap(struct net_device *netdev, struct nlmsghdr *nlh,
294 u32 seq, struct nlattr **tb, struct sk_buff *skb)
296 struct nlattr *data[DCB_CAP_ATTR_MAX + 1], *nest;
302 if (!tb[DCB_ATTR_CAP])
305 if (!netdev->dcbnl_ops->getcap)
308 ret = nla_parse_nested(data, DCB_CAP_ATTR_MAX, tb[DCB_ATTR_CAP],
313 nest = nla_nest_start(skb, DCB_ATTR_CAP);
317 if (data[DCB_CAP_ATTR_ALL])
320 for (i = DCB_CAP_ATTR_ALL+1; i <= DCB_CAP_ATTR_MAX; i++) {
321 if (!getall && !data[i])
324 if (!netdev->dcbnl_ops->getcap(netdev, i, &value)) {
325 ret = nla_put_u8(skb, i, value);
327 nla_nest_cancel(skb, nest);
332 nla_nest_end(skb, nest);
337 static int dcbnl_getnumtcs(struct net_device *netdev, struct nlmsghdr *nlh,
338 u32 seq, struct nlattr **tb, struct sk_buff *skb)
340 struct nlattr *data[DCB_NUMTCS_ATTR_MAX + 1], *nest;
346 if (!tb[DCB_ATTR_NUMTCS])
349 if (!netdev->dcbnl_ops->getnumtcs)
352 ret = nla_parse_nested(data, DCB_NUMTCS_ATTR_MAX, tb[DCB_ATTR_NUMTCS],
357 nest = nla_nest_start(skb, DCB_ATTR_NUMTCS);
361 if (data[DCB_NUMTCS_ATTR_ALL])
364 for (i = DCB_NUMTCS_ATTR_ALL+1; i <= DCB_NUMTCS_ATTR_MAX; i++) {
365 if (!getall && !data[i])
368 ret = netdev->dcbnl_ops->getnumtcs(netdev, i, &value);
370 ret = nla_put_u8(skb, i, value);
372 nla_nest_cancel(skb, nest);
378 nla_nest_end(skb, nest);
383 static int dcbnl_setnumtcs(struct net_device *netdev, struct nlmsghdr *nlh,
384 u32 seq, struct nlattr **tb, struct sk_buff *skb)
386 struct nlattr *data[DCB_NUMTCS_ATTR_MAX + 1];
391 if (!tb[DCB_ATTR_NUMTCS])
394 if (!netdev->dcbnl_ops->setnumtcs)
397 ret = nla_parse_nested(data, DCB_NUMTCS_ATTR_MAX, tb[DCB_ATTR_NUMTCS],
402 for (i = DCB_NUMTCS_ATTR_ALL+1; i <= DCB_NUMTCS_ATTR_MAX; i++) {
406 value = nla_get_u8(data[i]);
408 ret = netdev->dcbnl_ops->setnumtcs(netdev, i, value);
413 return nla_put_u8(skb, DCB_ATTR_NUMTCS, !!ret);
416 static int dcbnl_getpfcstate(struct net_device *netdev, struct nlmsghdr *nlh,
417 u32 seq, struct nlattr **tb, struct sk_buff *skb)
419 if (!netdev->dcbnl_ops->getpfcstate)
422 return nla_put_u8(skb, DCB_ATTR_PFC_STATE,
423 netdev->dcbnl_ops->getpfcstate(netdev));
426 static int dcbnl_setpfcstate(struct net_device *netdev, struct nlmsghdr *nlh,
427 u32 seq, struct nlattr **tb, struct sk_buff *skb)
431 if (!tb[DCB_ATTR_PFC_STATE])
434 if (!netdev->dcbnl_ops->setpfcstate)
437 value = nla_get_u8(tb[DCB_ATTR_PFC_STATE]);
439 netdev->dcbnl_ops->setpfcstate(netdev, value);
441 return nla_put_u8(skb, DCB_ATTR_PFC_STATE, 0);
444 static int dcbnl_getapp(struct net_device *netdev, struct nlmsghdr *nlh,
445 u32 seq, struct nlattr **tb, struct sk_buff *skb)
447 struct nlattr *app_nest;
448 struct nlattr *app_tb[DCB_APP_ATTR_MAX + 1];
453 if (!tb[DCB_ATTR_APP])
456 ret = nla_parse_nested(app_tb, DCB_APP_ATTR_MAX, tb[DCB_ATTR_APP],
461 /* all must be non-null */
462 if ((!app_tb[DCB_APP_ATTR_IDTYPE]) ||
463 (!app_tb[DCB_APP_ATTR_ID]))
466 /* either by eth type or by socket number */
467 idtype = nla_get_u8(app_tb[DCB_APP_ATTR_IDTYPE]);
468 if ((idtype != DCB_APP_IDTYPE_ETHTYPE) &&
469 (idtype != DCB_APP_IDTYPE_PORTNUM))
472 id = nla_get_u16(app_tb[DCB_APP_ATTR_ID]);
474 if (netdev->dcbnl_ops->getapp) {
475 up = netdev->dcbnl_ops->getapp(netdev, idtype, id);
477 struct dcb_app app = {
481 up = dcb_getapp(netdev, &app);
484 app_nest = nla_nest_start(skb, DCB_ATTR_APP);
488 ret = nla_put_u8(skb, DCB_APP_ATTR_IDTYPE, idtype);
492 ret = nla_put_u16(skb, DCB_APP_ATTR_ID, id);
496 ret = nla_put_u8(skb, DCB_APP_ATTR_PRIORITY, up);
500 nla_nest_end(skb, app_nest);
505 nla_nest_cancel(skb, app_nest);
509 static int dcbnl_setapp(struct net_device *netdev, struct nlmsghdr *nlh,
510 u32 seq, struct nlattr **tb, struct sk_buff *skb)
515 struct nlattr *app_tb[DCB_APP_ATTR_MAX + 1];
517 if (!tb[DCB_ATTR_APP])
520 ret = nla_parse_nested(app_tb, DCB_APP_ATTR_MAX, tb[DCB_ATTR_APP],
525 /* all must be non-null */
526 if ((!app_tb[DCB_APP_ATTR_IDTYPE]) ||
527 (!app_tb[DCB_APP_ATTR_ID]) ||
528 (!app_tb[DCB_APP_ATTR_PRIORITY]))
531 /* either by eth type or by socket number */
532 idtype = nla_get_u8(app_tb[DCB_APP_ATTR_IDTYPE]);
533 if ((idtype != DCB_APP_IDTYPE_ETHTYPE) &&
534 (idtype != DCB_APP_IDTYPE_PORTNUM))
537 id = nla_get_u16(app_tb[DCB_APP_ATTR_ID]);
538 up = nla_get_u8(app_tb[DCB_APP_ATTR_PRIORITY]);
540 if (netdev->dcbnl_ops->setapp) {
541 ret = netdev->dcbnl_ops->setapp(netdev, idtype, id, up);
544 app.selector = idtype;
547 ret = dcb_setapp(netdev, &app);
550 ret = nla_put_u8(skb, DCB_ATTR_APP, ret);
551 dcbnl_cee_notify(netdev, RTM_SETDCB, DCB_CMD_SAPP, seq, 0);
556 static int __dcbnl_pg_getcfg(struct net_device *netdev, struct nlmsghdr *nlh,
557 struct nlattr **tb, struct sk_buff *skb, int dir)
559 struct nlattr *pg_nest, *param_nest, *data;
560 struct nlattr *pg_tb[DCB_PG_ATTR_MAX + 1];
561 struct nlattr *param_tb[DCB_TC_ATTR_PARAM_MAX + 1];
562 u8 prio, pgid, tc_pct, up_map;
567 if (!tb[DCB_ATTR_PG_CFG])
570 if (!netdev->dcbnl_ops->getpgtccfgtx ||
571 !netdev->dcbnl_ops->getpgtccfgrx ||
572 !netdev->dcbnl_ops->getpgbwgcfgtx ||
573 !netdev->dcbnl_ops->getpgbwgcfgrx)
576 ret = nla_parse_nested(pg_tb, DCB_PG_ATTR_MAX,
577 tb[DCB_ATTR_PG_CFG], dcbnl_pg_nest);
581 pg_nest = nla_nest_start(skb, DCB_ATTR_PG_CFG);
585 if (pg_tb[DCB_PG_ATTR_TC_ALL])
588 for (i = DCB_PG_ATTR_TC_0; i <= DCB_PG_ATTR_TC_7; i++) {
589 if (!getall && !pg_tb[i])
592 if (pg_tb[DCB_PG_ATTR_TC_ALL])
593 data = pg_tb[DCB_PG_ATTR_TC_ALL];
596 ret = nla_parse_nested(param_tb, DCB_TC_ATTR_PARAM_MAX,
597 data, dcbnl_tc_param_nest);
601 param_nest = nla_nest_start(skb, i);
605 pgid = DCB_ATTR_VALUE_UNDEFINED;
606 prio = DCB_ATTR_VALUE_UNDEFINED;
607 tc_pct = DCB_ATTR_VALUE_UNDEFINED;
608 up_map = DCB_ATTR_VALUE_UNDEFINED;
612 netdev->dcbnl_ops->getpgtccfgrx(netdev,
613 i - DCB_PG_ATTR_TC_0, &prio,
614 &pgid, &tc_pct, &up_map);
617 netdev->dcbnl_ops->getpgtccfgtx(netdev,
618 i - DCB_PG_ATTR_TC_0, &prio,
619 &pgid, &tc_pct, &up_map);
622 if (param_tb[DCB_TC_ATTR_PARAM_PGID] ||
623 param_tb[DCB_TC_ATTR_PARAM_ALL]) {
624 ret = nla_put_u8(skb,
625 DCB_TC_ATTR_PARAM_PGID, pgid);
629 if (param_tb[DCB_TC_ATTR_PARAM_UP_MAPPING] ||
630 param_tb[DCB_TC_ATTR_PARAM_ALL]) {
631 ret = nla_put_u8(skb,
632 DCB_TC_ATTR_PARAM_UP_MAPPING, up_map);
636 if (param_tb[DCB_TC_ATTR_PARAM_STRICT_PRIO] ||
637 param_tb[DCB_TC_ATTR_PARAM_ALL]) {
638 ret = nla_put_u8(skb,
639 DCB_TC_ATTR_PARAM_STRICT_PRIO, prio);
643 if (param_tb[DCB_TC_ATTR_PARAM_BW_PCT] ||
644 param_tb[DCB_TC_ATTR_PARAM_ALL]) {
645 ret = nla_put_u8(skb, DCB_TC_ATTR_PARAM_BW_PCT,
650 nla_nest_end(skb, param_nest);
653 if (pg_tb[DCB_PG_ATTR_BW_ID_ALL])
658 for (i = DCB_PG_ATTR_BW_ID_0; i <= DCB_PG_ATTR_BW_ID_7; i++) {
659 if (!getall && !pg_tb[i])
662 tc_pct = DCB_ATTR_VALUE_UNDEFINED;
666 netdev->dcbnl_ops->getpgbwgcfgrx(netdev,
667 i - DCB_PG_ATTR_BW_ID_0, &tc_pct);
670 netdev->dcbnl_ops->getpgbwgcfgtx(netdev,
671 i - DCB_PG_ATTR_BW_ID_0, &tc_pct);
673 ret = nla_put_u8(skb, i, tc_pct);
678 nla_nest_end(skb, pg_nest);
683 nla_nest_cancel(skb, param_nest);
685 nla_nest_cancel(skb, pg_nest);
690 static int dcbnl_pgtx_getcfg(struct net_device *netdev, struct nlmsghdr *nlh,
691 u32 seq, struct nlattr **tb, struct sk_buff *skb)
693 return __dcbnl_pg_getcfg(netdev, nlh, tb, skb, 0);
696 static int dcbnl_pgrx_getcfg(struct net_device *netdev, struct nlmsghdr *nlh,
697 u32 seq, struct nlattr **tb, struct sk_buff *skb)
699 return __dcbnl_pg_getcfg(netdev, nlh, tb, skb, 1);
702 static int dcbnl_setstate(struct net_device *netdev, struct nlmsghdr *nlh,
703 u32 seq, struct nlattr **tb, struct sk_buff *skb)
707 if (!tb[DCB_ATTR_STATE])
710 if (!netdev->dcbnl_ops->setstate)
713 value = nla_get_u8(tb[DCB_ATTR_STATE]);
715 return nla_put_u8(skb, DCB_ATTR_STATE,
716 netdev->dcbnl_ops->setstate(netdev, value));
719 static int dcbnl_setpfccfg(struct net_device *netdev, struct nlmsghdr *nlh,
720 u32 seq, struct nlattr **tb, struct sk_buff *skb)
722 struct nlattr *data[DCB_PFC_UP_ATTR_MAX + 1];
727 if (!tb[DCB_ATTR_PFC_CFG])
730 if (!netdev->dcbnl_ops->setpfccfg)
733 ret = nla_parse_nested(data, DCB_PFC_UP_ATTR_MAX,
734 tb[DCB_ATTR_PFC_CFG],
739 for (i = DCB_PFC_UP_ATTR_0; i <= DCB_PFC_UP_ATTR_7; i++) {
742 value = nla_get_u8(data[i]);
743 netdev->dcbnl_ops->setpfccfg(netdev,
744 data[i]->nla_type - DCB_PFC_UP_ATTR_0, value);
747 return nla_put_u8(skb, DCB_ATTR_PFC_CFG, 0);
750 static int dcbnl_setall(struct net_device *netdev, struct nlmsghdr *nlh,
751 u32 seq, struct nlattr **tb, struct sk_buff *skb)
755 if (!tb[DCB_ATTR_SET_ALL])
758 if (!netdev->dcbnl_ops->setall)
761 ret = nla_put_u8(skb, DCB_ATTR_SET_ALL,
762 netdev->dcbnl_ops->setall(netdev));
763 dcbnl_cee_notify(netdev, RTM_SETDCB, DCB_CMD_SET_ALL, seq, 0);
768 static int __dcbnl_pg_setcfg(struct net_device *netdev, struct nlmsghdr *nlh,
769 u32 seq, struct nlattr **tb, struct sk_buff *skb,
772 struct nlattr *pg_tb[DCB_PG_ATTR_MAX + 1];
773 struct nlattr *param_tb[DCB_TC_ATTR_PARAM_MAX + 1];
781 if (!tb[DCB_ATTR_PG_CFG])
784 if (!netdev->dcbnl_ops->setpgtccfgtx ||
785 !netdev->dcbnl_ops->setpgtccfgrx ||
786 !netdev->dcbnl_ops->setpgbwgcfgtx ||
787 !netdev->dcbnl_ops->setpgbwgcfgrx)
790 ret = nla_parse_nested(pg_tb, DCB_PG_ATTR_MAX,
791 tb[DCB_ATTR_PG_CFG], dcbnl_pg_nest);
795 for (i = DCB_PG_ATTR_TC_0; i <= DCB_PG_ATTR_TC_7; i++) {
799 ret = nla_parse_nested(param_tb, DCB_TC_ATTR_PARAM_MAX,
800 pg_tb[i], dcbnl_tc_param_nest);
804 pgid = DCB_ATTR_VALUE_UNDEFINED;
805 prio = DCB_ATTR_VALUE_UNDEFINED;
806 tc_pct = DCB_ATTR_VALUE_UNDEFINED;
807 up_map = DCB_ATTR_VALUE_UNDEFINED;
809 if (param_tb[DCB_TC_ATTR_PARAM_STRICT_PRIO])
811 nla_get_u8(param_tb[DCB_TC_ATTR_PARAM_STRICT_PRIO]);
813 if (param_tb[DCB_TC_ATTR_PARAM_PGID])
814 pgid = nla_get_u8(param_tb[DCB_TC_ATTR_PARAM_PGID]);
816 if (param_tb[DCB_TC_ATTR_PARAM_BW_PCT])
817 tc_pct = nla_get_u8(param_tb[DCB_TC_ATTR_PARAM_BW_PCT]);
819 if (param_tb[DCB_TC_ATTR_PARAM_UP_MAPPING])
821 nla_get_u8(param_tb[DCB_TC_ATTR_PARAM_UP_MAPPING]);
823 /* dir: Tx = 0, Rx = 1 */
826 netdev->dcbnl_ops->setpgtccfgrx(netdev,
827 i - DCB_PG_ATTR_TC_0,
828 prio, pgid, tc_pct, up_map);
831 netdev->dcbnl_ops->setpgtccfgtx(netdev,
832 i - DCB_PG_ATTR_TC_0,
833 prio, pgid, tc_pct, up_map);
837 for (i = DCB_PG_ATTR_BW_ID_0; i <= DCB_PG_ATTR_BW_ID_7; i++) {
841 tc_pct = nla_get_u8(pg_tb[i]);
843 /* dir: Tx = 0, Rx = 1 */
846 netdev->dcbnl_ops->setpgbwgcfgrx(netdev,
847 i - DCB_PG_ATTR_BW_ID_0, tc_pct);
850 netdev->dcbnl_ops->setpgbwgcfgtx(netdev,
851 i - DCB_PG_ATTR_BW_ID_0, tc_pct);
855 return nla_put_u8(skb, DCB_ATTR_PG_CFG, 0);
858 static int dcbnl_pgtx_setcfg(struct net_device *netdev, struct nlmsghdr *nlh,
859 u32 seq, struct nlattr **tb, struct sk_buff *skb)
861 return __dcbnl_pg_setcfg(netdev, nlh, seq, tb, skb, 0);
864 static int dcbnl_pgrx_setcfg(struct net_device *netdev, struct nlmsghdr *nlh,
865 u32 seq, struct nlattr **tb, struct sk_buff *skb)
867 return __dcbnl_pg_setcfg(netdev, nlh, seq, tb, skb, 1);
870 static int dcbnl_bcn_getcfg(struct net_device *netdev, struct nlmsghdr *nlh,
871 u32 seq, struct nlattr **tb, struct sk_buff *skb)
873 struct nlattr *bcn_nest;
874 struct nlattr *bcn_tb[DCB_BCN_ATTR_MAX + 1];
881 if (!tb[DCB_ATTR_BCN])
884 if (!netdev->dcbnl_ops->getbcnrp ||
885 !netdev->dcbnl_ops->getbcncfg)
888 ret = nla_parse_nested(bcn_tb, DCB_BCN_ATTR_MAX,
889 tb[DCB_ATTR_BCN], dcbnl_bcn_nest);
893 bcn_nest = nla_nest_start(skb, DCB_ATTR_BCN);
897 if (bcn_tb[DCB_BCN_ATTR_ALL])
900 for (i = DCB_BCN_ATTR_RP_0; i <= DCB_BCN_ATTR_RP_7; i++) {
901 if (!getall && !bcn_tb[i])
904 netdev->dcbnl_ops->getbcnrp(netdev, i - DCB_BCN_ATTR_RP_0,
906 ret = nla_put_u8(skb, i, value_byte);
911 for (i = DCB_BCN_ATTR_BCNA_0; i <= DCB_BCN_ATTR_RI; i++) {
912 if (!getall && !bcn_tb[i])
915 netdev->dcbnl_ops->getbcncfg(netdev, i,
917 ret = nla_put_u32(skb, i, value_integer);
922 nla_nest_end(skb, bcn_nest);
927 nla_nest_cancel(skb, bcn_nest);
931 static int dcbnl_bcn_setcfg(struct net_device *netdev, struct nlmsghdr *nlh,
932 u32 seq, struct nlattr **tb, struct sk_buff *skb)
934 struct nlattr *data[DCB_BCN_ATTR_MAX + 1];
940 if (!tb[DCB_ATTR_BCN])
943 if (!netdev->dcbnl_ops->setbcncfg ||
944 !netdev->dcbnl_ops->setbcnrp)
947 ret = nla_parse_nested(data, DCB_BCN_ATTR_MAX,
953 for (i = DCB_BCN_ATTR_RP_0; i <= DCB_BCN_ATTR_RP_7; i++) {
956 value_byte = nla_get_u8(data[i]);
957 netdev->dcbnl_ops->setbcnrp(netdev,
958 data[i]->nla_type - DCB_BCN_ATTR_RP_0, value_byte);
961 for (i = DCB_BCN_ATTR_BCNA_0; i <= DCB_BCN_ATTR_RI; i++) {
964 value_int = nla_get_u32(data[i]);
965 netdev->dcbnl_ops->setbcncfg(netdev,
969 return nla_put_u8(skb, DCB_ATTR_BCN, 0);
972 static int dcbnl_build_peer_app(struct net_device *netdev, struct sk_buff* skb,
973 int app_nested_type, int app_info_type,
976 struct dcb_peer_app_info info;
977 struct dcb_app *table = NULL;
978 const struct dcbnl_rtnl_ops *ops = netdev->dcbnl_ops;
984 * retrieve the peer app configuration form the driver. If the driver
985 * handlers fail exit without doing anything
987 err = ops->peer_getappinfo(netdev, &info, &app_count);
988 if (!err && app_count) {
989 table = kmalloc(sizeof(struct dcb_app) * app_count, GFP_KERNEL);
993 err = ops->peer_getapptable(netdev, table);
1001 * build the message, from here on the only possible failure
1002 * is due to the skb size
1006 app = nla_nest_start(skb, app_nested_type);
1008 goto nla_put_failure;
1010 if (app_info_type &&
1011 nla_put(skb, app_info_type, sizeof(info), &info))
1012 goto nla_put_failure;
1014 for (i = 0; i < app_count; i++) {
1015 if (nla_put(skb, app_entry_type, sizeof(struct dcb_app),
1017 goto nla_put_failure;
1019 nla_nest_end(skb, app);
1028 /* Handle IEEE 802.1Qaz GET commands. */
1029 static int dcbnl_ieee_fill(struct sk_buff *skb, struct net_device *netdev)
1031 struct nlattr *ieee, *app;
1032 struct dcb_app_type *itr;
1033 const struct dcbnl_rtnl_ops *ops = netdev->dcbnl_ops;
1037 if (nla_put_string(skb, DCB_ATTR_IFNAME, netdev->name))
1040 ieee = nla_nest_start(skb, DCB_ATTR_IEEE);
1044 if (ops->ieee_getets) {
1045 struct ieee_ets ets;
1046 err = ops->ieee_getets(netdev, &ets);
1048 nla_put(skb, DCB_ATTR_IEEE_ETS, sizeof(ets), &ets))
1052 if (ops->ieee_getmaxrate) {
1053 struct ieee_maxrate maxrate;
1054 err = ops->ieee_getmaxrate(netdev, &maxrate);
1056 err = nla_put(skb, DCB_ATTR_IEEE_MAXRATE,
1057 sizeof(maxrate), &maxrate);
1063 if (ops->ieee_getpfc) {
1064 struct ieee_pfc pfc;
1065 err = ops->ieee_getpfc(netdev, &pfc);
1067 nla_put(skb, DCB_ATTR_IEEE_PFC, sizeof(pfc), &pfc))
1071 app = nla_nest_start(skb, DCB_ATTR_IEEE_APP_TABLE);
1075 spin_lock(&dcb_lock);
1076 list_for_each_entry(itr, &dcb_app_list, list) {
1077 if (itr->ifindex == netdev->ifindex) {
1078 err = nla_put(skb, DCB_ATTR_IEEE_APP, sizeof(itr->app),
1081 spin_unlock(&dcb_lock);
1087 if (netdev->dcbnl_ops->getdcbx)
1088 dcbx = netdev->dcbnl_ops->getdcbx(netdev);
1092 spin_unlock(&dcb_lock);
1093 nla_nest_end(skb, app);
1095 /* get peer info if available */
1096 if (ops->ieee_peer_getets) {
1097 struct ieee_ets ets;
1098 err = ops->ieee_peer_getets(netdev, &ets);
1100 nla_put(skb, DCB_ATTR_IEEE_PEER_ETS, sizeof(ets), &ets))
1104 if (ops->ieee_peer_getpfc) {
1105 struct ieee_pfc pfc;
1106 err = ops->ieee_peer_getpfc(netdev, &pfc);
1108 nla_put(skb, DCB_ATTR_IEEE_PEER_PFC, sizeof(pfc), &pfc))
1112 if (ops->peer_getappinfo && ops->peer_getapptable) {
1113 err = dcbnl_build_peer_app(netdev, skb,
1114 DCB_ATTR_IEEE_PEER_APP,
1115 DCB_ATTR_IEEE_APP_UNSPEC,
1121 nla_nest_end(skb, ieee);
1123 err = nla_put_u8(skb, DCB_ATTR_DCBX, dcbx);
1131 static int dcbnl_cee_pg_fill(struct sk_buff *skb, struct net_device *dev,
1134 u8 pgid, up_map, prio, tc_pct;
1135 const struct dcbnl_rtnl_ops *ops = dev->dcbnl_ops;
1136 int i = dir ? DCB_ATTR_CEE_TX_PG : DCB_ATTR_CEE_RX_PG;
1137 struct nlattr *pg = nla_nest_start(skb, i);
1142 for (i = DCB_PG_ATTR_TC_0; i <= DCB_PG_ATTR_TC_7; i++) {
1143 struct nlattr *tc_nest = nla_nest_start(skb, i);
1148 pgid = DCB_ATTR_VALUE_UNDEFINED;
1149 prio = DCB_ATTR_VALUE_UNDEFINED;
1150 tc_pct = DCB_ATTR_VALUE_UNDEFINED;
1151 up_map = DCB_ATTR_VALUE_UNDEFINED;
1154 ops->getpgtccfgrx(dev, i - DCB_PG_ATTR_TC_0,
1155 &prio, &pgid, &tc_pct, &up_map);
1157 ops->getpgtccfgtx(dev, i - DCB_PG_ATTR_TC_0,
1158 &prio, &pgid, &tc_pct, &up_map);
1160 if (nla_put_u8(skb, DCB_TC_ATTR_PARAM_PGID, pgid) ||
1161 nla_put_u8(skb, DCB_TC_ATTR_PARAM_UP_MAPPING, up_map) ||
1162 nla_put_u8(skb, DCB_TC_ATTR_PARAM_STRICT_PRIO, prio) ||
1163 nla_put_u8(skb, DCB_TC_ATTR_PARAM_BW_PCT, tc_pct))
1165 nla_nest_end(skb, tc_nest);
1168 for (i = DCB_PG_ATTR_BW_ID_0; i <= DCB_PG_ATTR_BW_ID_7; i++) {
1169 tc_pct = DCB_ATTR_VALUE_UNDEFINED;
1172 ops->getpgbwgcfgrx(dev, i - DCB_PG_ATTR_BW_ID_0,
1175 ops->getpgbwgcfgtx(dev, i - DCB_PG_ATTR_BW_ID_0,
1177 if (nla_put_u8(skb, i, tc_pct))
1180 nla_nest_end(skb, pg);
1184 static int dcbnl_cee_fill(struct sk_buff *skb, struct net_device *netdev)
1186 struct nlattr *cee, *app;
1187 struct dcb_app_type *itr;
1188 const struct dcbnl_rtnl_ops *ops = netdev->dcbnl_ops;
1189 int dcbx, i, err = -EMSGSIZE;
1192 if (nla_put_string(skb, DCB_ATTR_IFNAME, netdev->name))
1193 goto nla_put_failure;
1194 cee = nla_nest_start(skb, DCB_ATTR_CEE);
1196 goto nla_put_failure;
1199 if (ops->getpgtccfgtx && ops->getpgbwgcfgtx) {
1200 err = dcbnl_cee_pg_fill(skb, netdev, 1);
1202 goto nla_put_failure;
1205 if (ops->getpgtccfgrx && ops->getpgbwgcfgrx) {
1206 err = dcbnl_cee_pg_fill(skb, netdev, 0);
1208 goto nla_put_failure;
1212 if (ops->getpfccfg) {
1213 struct nlattr *pfc_nest = nla_nest_start(skb, DCB_ATTR_CEE_PFC);
1216 goto nla_put_failure;
1218 for (i = DCB_PFC_UP_ATTR_0; i <= DCB_PFC_UP_ATTR_7; i++) {
1219 ops->getpfccfg(netdev, i - DCB_PFC_UP_ATTR_0, &value);
1220 if (nla_put_u8(skb, i, value))
1221 goto nla_put_failure;
1223 nla_nest_end(skb, pfc_nest);
1227 spin_lock(&dcb_lock);
1228 app = nla_nest_start(skb, DCB_ATTR_CEE_APP_TABLE);
1232 list_for_each_entry(itr, &dcb_app_list, list) {
1233 if (itr->ifindex == netdev->ifindex) {
1234 struct nlattr *app_nest = nla_nest_start(skb,
1239 err = nla_put_u8(skb, DCB_APP_ATTR_IDTYPE,
1244 err = nla_put_u16(skb, DCB_APP_ATTR_ID,
1249 err = nla_put_u8(skb, DCB_APP_ATTR_PRIORITY,
1254 nla_nest_end(skb, app_nest);
1257 nla_nest_end(skb, app);
1259 if (netdev->dcbnl_ops->getdcbx)
1260 dcbx = netdev->dcbnl_ops->getdcbx(netdev);
1264 spin_unlock(&dcb_lock);
1266 /* features flags */
1267 if (ops->getfeatcfg) {
1268 struct nlattr *feat = nla_nest_start(skb, DCB_ATTR_CEE_FEAT);
1270 goto nla_put_failure;
1272 for (i = DCB_FEATCFG_ATTR_ALL + 1; i <= DCB_FEATCFG_ATTR_MAX;
1274 if (!ops->getfeatcfg(netdev, i, &value) &&
1275 nla_put_u8(skb, i, value))
1276 goto nla_put_failure;
1278 nla_nest_end(skb, feat);
1281 /* peer info if available */
1282 if (ops->cee_peer_getpg) {
1284 err = ops->cee_peer_getpg(netdev, &pg);
1286 nla_put(skb, DCB_ATTR_CEE_PEER_PG, sizeof(pg), &pg))
1287 goto nla_put_failure;
1290 if (ops->cee_peer_getpfc) {
1292 err = ops->cee_peer_getpfc(netdev, &pfc);
1294 nla_put(skb, DCB_ATTR_CEE_PEER_PFC, sizeof(pfc), &pfc))
1295 goto nla_put_failure;
1298 if (ops->peer_getappinfo && ops->peer_getapptable) {
1299 err = dcbnl_build_peer_app(netdev, skb,
1300 DCB_ATTR_CEE_PEER_APP_TABLE,
1301 DCB_ATTR_CEE_PEER_APP_INFO,
1302 DCB_ATTR_CEE_PEER_APP);
1304 goto nla_put_failure;
1306 nla_nest_end(skb, cee);
1310 err = nla_put_u8(skb, DCB_ATTR_DCBX, dcbx);
1312 goto nla_put_failure;
1317 spin_unlock(&dcb_lock);
1322 static int dcbnl_notify(struct net_device *dev, int event, int cmd,
1323 u32 seq, u32 pid, int dcbx_ver)
1325 struct net *net = dev_net(dev);
1326 struct sk_buff *skb;
1327 struct nlmsghdr *nlh;
1328 const struct dcbnl_rtnl_ops *ops = dev->dcbnl_ops;
1334 skb = dcbnl_newmsg(event, cmd, pid, seq, 0, &nlh);
1338 if (dcbx_ver == DCB_CAP_DCBX_VER_IEEE)
1339 err = dcbnl_ieee_fill(skb, dev);
1341 err = dcbnl_cee_fill(skb, dev);
1344 /* Report error to broadcast listeners */
1346 rtnl_set_sk_err(net, RTNLGRP_DCB, err);
1348 /* End nlmsg and notify broadcast listeners */
1349 nlmsg_end(skb, nlh);
1350 rtnl_notify(skb, net, 0, RTNLGRP_DCB, NULL, GFP_KERNEL);
1356 int dcbnl_ieee_notify(struct net_device *dev, int event, int cmd,
1359 return dcbnl_notify(dev, event, cmd, seq, pid, DCB_CAP_DCBX_VER_IEEE);
1361 EXPORT_SYMBOL(dcbnl_ieee_notify);
1363 int dcbnl_cee_notify(struct net_device *dev, int event, int cmd,
1366 return dcbnl_notify(dev, event, cmd, seq, pid, DCB_CAP_DCBX_VER_CEE);
1368 EXPORT_SYMBOL(dcbnl_cee_notify);
1370 /* Handle IEEE 802.1Qaz SET commands. If any requested operation can not
1371 * be completed the entire msg is aborted and error value is returned.
1372 * No attempt is made to reconcile the case where only part of the
1373 * cmd can be completed.
1375 static int dcbnl_ieee_set(struct net_device *netdev, struct nlmsghdr *nlh,
1376 u32 seq, struct nlattr **tb, struct sk_buff *skb)
1378 const struct dcbnl_rtnl_ops *ops = netdev->dcbnl_ops;
1379 struct nlattr *ieee[DCB_ATTR_IEEE_MAX + 1];
1385 if (!tb[DCB_ATTR_IEEE])
1388 err = nla_parse_nested(ieee, DCB_ATTR_IEEE_MAX,
1389 tb[DCB_ATTR_IEEE], dcbnl_ieee_policy);
1393 if (ieee[DCB_ATTR_IEEE_ETS] && ops->ieee_setets) {
1394 struct ieee_ets *ets = nla_data(ieee[DCB_ATTR_IEEE_ETS]);
1395 err = ops->ieee_setets(netdev, ets);
1400 if (ieee[DCB_ATTR_IEEE_MAXRATE] && ops->ieee_setmaxrate) {
1401 struct ieee_maxrate *maxrate =
1402 nla_data(ieee[DCB_ATTR_IEEE_MAXRATE]);
1403 err = ops->ieee_setmaxrate(netdev, maxrate);
1408 if (ieee[DCB_ATTR_IEEE_PFC] && ops->ieee_setpfc) {
1409 struct ieee_pfc *pfc = nla_data(ieee[DCB_ATTR_IEEE_PFC]);
1410 err = ops->ieee_setpfc(netdev, pfc);
1415 if (ieee[DCB_ATTR_IEEE_APP_TABLE]) {
1416 struct nlattr *attr;
1419 nla_for_each_nested(attr, ieee[DCB_ATTR_IEEE_APP_TABLE], rem) {
1420 struct dcb_app *app_data;
1421 if (nla_type(attr) != DCB_ATTR_IEEE_APP)
1423 app_data = nla_data(attr);
1424 if (ops->ieee_setapp)
1425 err = ops->ieee_setapp(netdev, app_data);
1427 err = dcb_ieee_setapp(netdev, app_data);
1434 err = nla_put_u8(skb, DCB_ATTR_IEEE, err);
1435 dcbnl_ieee_notify(netdev, RTM_SETDCB, DCB_CMD_IEEE_SET, seq, 0);
1439 static int dcbnl_ieee_get(struct net_device *netdev, struct nlmsghdr *nlh,
1440 u32 seq, struct nlattr **tb, struct sk_buff *skb)
1442 const struct dcbnl_rtnl_ops *ops = netdev->dcbnl_ops;
1447 return dcbnl_ieee_fill(skb, netdev);
1450 static int dcbnl_ieee_del(struct net_device *netdev, struct nlmsghdr *nlh,
1451 u32 seq, struct nlattr **tb, struct sk_buff *skb)
1453 const struct dcbnl_rtnl_ops *ops = netdev->dcbnl_ops;
1454 struct nlattr *ieee[DCB_ATTR_IEEE_MAX + 1];
1460 if (!tb[DCB_ATTR_IEEE])
1463 err = nla_parse_nested(ieee, DCB_ATTR_IEEE_MAX,
1464 tb[DCB_ATTR_IEEE], dcbnl_ieee_policy);
1468 if (ieee[DCB_ATTR_IEEE_APP_TABLE]) {
1469 struct nlattr *attr;
1472 nla_for_each_nested(attr, ieee[DCB_ATTR_IEEE_APP_TABLE], rem) {
1473 struct dcb_app *app_data;
1475 if (nla_type(attr) != DCB_ATTR_IEEE_APP)
1477 app_data = nla_data(attr);
1478 if (ops->ieee_delapp)
1479 err = ops->ieee_delapp(netdev, app_data);
1481 err = dcb_ieee_delapp(netdev, app_data);
1488 err = nla_put_u8(skb, DCB_ATTR_IEEE, err);
1489 dcbnl_ieee_notify(netdev, RTM_SETDCB, DCB_CMD_IEEE_DEL, seq, 0);
1494 /* DCBX configuration */
1495 static int dcbnl_getdcbx(struct net_device *netdev, struct nlmsghdr *nlh,
1496 u32 seq, struct nlattr **tb, struct sk_buff *skb)
1498 if (!netdev->dcbnl_ops->getdcbx)
1501 return nla_put_u8(skb, DCB_ATTR_DCBX,
1502 netdev->dcbnl_ops->getdcbx(netdev));
1505 static int dcbnl_setdcbx(struct net_device *netdev, struct nlmsghdr *nlh,
1506 u32 seq, struct nlattr **tb, struct sk_buff *skb)
1510 if (!netdev->dcbnl_ops->setdcbx)
1513 if (!tb[DCB_ATTR_DCBX])
1516 value = nla_get_u8(tb[DCB_ATTR_DCBX]);
1518 return nla_put_u8(skb, DCB_ATTR_DCBX,
1519 netdev->dcbnl_ops->setdcbx(netdev, value));
1522 static int dcbnl_getfeatcfg(struct net_device *netdev, struct nlmsghdr *nlh,
1523 u32 seq, struct nlattr **tb, struct sk_buff *skb)
1525 struct nlattr *data[DCB_FEATCFG_ATTR_MAX + 1], *nest;
1530 if (!netdev->dcbnl_ops->getfeatcfg)
1533 if (!tb[DCB_ATTR_FEATCFG])
1536 ret = nla_parse_nested(data, DCB_FEATCFG_ATTR_MAX, tb[DCB_ATTR_FEATCFG],
1537 dcbnl_featcfg_nest);
1541 nest = nla_nest_start(skb, DCB_ATTR_FEATCFG);
1545 if (data[DCB_FEATCFG_ATTR_ALL])
1548 for (i = DCB_FEATCFG_ATTR_ALL+1; i <= DCB_FEATCFG_ATTR_MAX; i++) {
1549 if (!getall && !data[i])
1552 ret = netdev->dcbnl_ops->getfeatcfg(netdev, i, &value);
1554 ret = nla_put_u8(skb, i, value);
1557 nla_nest_cancel(skb, nest);
1558 goto nla_put_failure;
1561 nla_nest_end(skb, nest);
1567 static int dcbnl_setfeatcfg(struct net_device *netdev, struct nlmsghdr *nlh,
1568 u32 seq, struct nlattr **tb, struct sk_buff *skb)
1570 struct nlattr *data[DCB_FEATCFG_ATTR_MAX + 1];
1574 if (!netdev->dcbnl_ops->setfeatcfg)
1577 if (!tb[DCB_ATTR_FEATCFG])
1580 ret = nla_parse_nested(data, DCB_FEATCFG_ATTR_MAX, tb[DCB_ATTR_FEATCFG],
1581 dcbnl_featcfg_nest);
1586 for (i = DCB_FEATCFG_ATTR_ALL+1; i <= DCB_FEATCFG_ATTR_MAX; i++) {
1587 if (data[i] == NULL)
1590 value = nla_get_u8(data[i]);
1592 ret = netdev->dcbnl_ops->setfeatcfg(netdev, i, value);
1598 ret = nla_put_u8(skb, DCB_ATTR_FEATCFG, ret);
1603 /* Handle CEE DCBX GET commands. */
1604 static int dcbnl_cee_get(struct net_device *netdev, struct nlmsghdr *nlh,
1605 u32 seq, struct nlattr **tb, struct sk_buff *skb)
1607 const struct dcbnl_rtnl_ops *ops = netdev->dcbnl_ops;
1612 return dcbnl_cee_fill(skb, netdev);
1616 /* reply netlink message type */
1619 /* function to fill message contents */
1620 int (*cb)(struct net_device *, struct nlmsghdr *, u32,
1621 struct nlattr **, struct sk_buff *);
1624 static const struct reply_func reply_funcs[DCB_CMD_MAX+1] = {
1625 [DCB_CMD_GSTATE] = { RTM_GETDCB, dcbnl_getstate },
1626 [DCB_CMD_SSTATE] = { RTM_SETDCB, dcbnl_setstate },
1627 [DCB_CMD_PFC_GCFG] = { RTM_GETDCB, dcbnl_getpfccfg },
1628 [DCB_CMD_PFC_SCFG] = { RTM_SETDCB, dcbnl_setpfccfg },
1629 [DCB_CMD_GPERM_HWADDR] = { RTM_GETDCB, dcbnl_getperm_hwaddr },
1630 [DCB_CMD_GCAP] = { RTM_GETDCB, dcbnl_getcap },
1631 [DCB_CMD_GNUMTCS] = { RTM_GETDCB, dcbnl_getnumtcs },
1632 [DCB_CMD_SNUMTCS] = { RTM_SETDCB, dcbnl_setnumtcs },
1633 [DCB_CMD_PFC_GSTATE] = { RTM_GETDCB, dcbnl_getpfcstate },
1634 [DCB_CMD_PFC_SSTATE] = { RTM_SETDCB, dcbnl_setpfcstate },
1635 [DCB_CMD_GAPP] = { RTM_GETDCB, dcbnl_getapp },
1636 [DCB_CMD_SAPP] = { RTM_SETDCB, dcbnl_setapp },
1637 [DCB_CMD_PGTX_GCFG] = { RTM_GETDCB, dcbnl_pgtx_getcfg },
1638 [DCB_CMD_PGTX_SCFG] = { RTM_SETDCB, dcbnl_pgtx_setcfg },
1639 [DCB_CMD_PGRX_GCFG] = { RTM_GETDCB, dcbnl_pgrx_getcfg },
1640 [DCB_CMD_PGRX_SCFG] = { RTM_SETDCB, dcbnl_pgrx_setcfg },
1641 [DCB_CMD_SET_ALL] = { RTM_SETDCB, dcbnl_setall },
1642 [DCB_CMD_BCN_GCFG] = { RTM_GETDCB, dcbnl_bcn_getcfg },
1643 [DCB_CMD_BCN_SCFG] = { RTM_SETDCB, dcbnl_bcn_setcfg },
1644 [DCB_CMD_IEEE_GET] = { RTM_GETDCB, dcbnl_ieee_get },
1645 [DCB_CMD_IEEE_SET] = { RTM_SETDCB, dcbnl_ieee_set },
1646 [DCB_CMD_IEEE_DEL] = { RTM_SETDCB, dcbnl_ieee_del },
1647 [DCB_CMD_GDCBX] = { RTM_GETDCB, dcbnl_getdcbx },
1648 [DCB_CMD_SDCBX] = { RTM_SETDCB, dcbnl_setdcbx },
1649 [DCB_CMD_GFEATCFG] = { RTM_GETDCB, dcbnl_getfeatcfg },
1650 [DCB_CMD_SFEATCFG] = { RTM_SETDCB, dcbnl_setfeatcfg },
1651 [DCB_CMD_CEE_GET] = { RTM_GETDCB, dcbnl_cee_get },
1654 static int dcb_doit(struct sk_buff *skb, struct nlmsghdr *nlh, void *arg)
1656 struct net *net = sock_net(skb->sk);
1657 struct net_device *netdev;
1658 struct dcbmsg *dcb = nlmsg_data(nlh);
1659 struct nlattr *tb[DCB_ATTR_MAX + 1];
1660 u32 pid = skb ? NETLINK_CB(skb).pid : 0;
1662 struct sk_buff *reply_skb;
1663 struct nlmsghdr *reply_nlh = NULL;
1664 const struct reply_func *fn;
1666 if (!net_eq(net, &init_net))
1669 ret = nlmsg_parse(nlh, sizeof(*dcb), tb, DCB_ATTR_MAX,
1674 if (dcb->cmd > DCB_CMD_MAX)
1677 /* check if a reply function has been defined for the command */
1678 fn = &reply_funcs[dcb->cmd];
1682 if (!tb[DCB_ATTR_IFNAME])
1685 netdev = dev_get_by_name(&init_net, nla_data(tb[DCB_ATTR_IFNAME]));
1689 if (!netdev->dcbnl_ops) {
1694 reply_skb = dcbnl_newmsg(fn->type, dcb->cmd, pid, nlh->nlmsg_seq,
1695 nlh->nlmsg_flags, &reply_nlh);
1701 ret = fn->cb(netdev, nlh, nlh->nlmsg_seq, tb, reply_skb);
1703 nlmsg_free(reply_skb);
1707 nlmsg_end(reply_skb, reply_nlh);
1709 ret = rtnl_unicast(reply_skb, &init_net, pid);
1715 static struct dcb_app_type *dcb_app_lookup(const struct dcb_app *app,
1716 int ifindex, int prio)
1718 struct dcb_app_type *itr;
1720 list_for_each_entry(itr, &dcb_app_list, list) {
1721 if (itr->app.selector == app->selector &&
1722 itr->app.protocol == app->protocol &&
1723 itr->ifindex == ifindex &&
1724 (!prio || itr->app.priority == prio))
1731 static int dcb_app_add(const struct dcb_app *app, int ifindex)
1733 struct dcb_app_type *entry;
1735 entry = kmalloc(sizeof(*entry), GFP_ATOMIC);
1739 memcpy(&entry->app, app, sizeof(*app));
1740 entry->ifindex = ifindex;
1741 list_add(&entry->list, &dcb_app_list);
1747 * dcb_getapp - retrieve the DCBX application user priority
1749 * On success returns a non-zero 802.1p user priority bitmap
1750 * otherwise returns 0 as the invalid user priority bitmap to
1751 * indicate an error.
1753 u8 dcb_getapp(struct net_device *dev, struct dcb_app *app)
1755 struct dcb_app_type *itr;
1758 spin_lock(&dcb_lock);
1759 if ((itr = dcb_app_lookup(app, dev->ifindex, 0)))
1760 prio = itr->app.priority;
1761 spin_unlock(&dcb_lock);
1765 EXPORT_SYMBOL(dcb_getapp);
1768 * dcb_setapp - add CEE dcb application data to app list
1770 * Priority 0 is an invalid priority in CEE spec. This routine
1771 * removes applications from the app list if the priority is
1774 int dcb_setapp(struct net_device *dev, struct dcb_app *new)
1776 struct dcb_app_type *itr;
1777 struct dcb_app_type event;
1780 event.ifindex = dev->ifindex;
1781 memcpy(&event.app, new, sizeof(event.app));
1782 if (dev->dcbnl_ops->getdcbx)
1783 event.dcbx = dev->dcbnl_ops->getdcbx(dev);
1785 spin_lock(&dcb_lock);
1786 /* Search for existing match and replace */
1787 if ((itr = dcb_app_lookup(new, dev->ifindex, 0))) {
1789 itr->app.priority = new->priority;
1791 list_del(&itr->list);
1796 /* App type does not exist add new application type */
1798 err = dcb_app_add(new, dev->ifindex);
1800 spin_unlock(&dcb_lock);
1802 call_dcbevent_notifiers(DCB_APP_EVENT, &event);
1805 EXPORT_SYMBOL(dcb_setapp);
1808 * dcb_ieee_getapp_mask - retrieve the IEEE DCB application priority
1810 * Helper routine which on success returns a non-zero 802.1Qaz user
1811 * priority bitmap otherwise returns 0 to indicate the dcb_app was
1812 * not found in APP list.
1814 u8 dcb_ieee_getapp_mask(struct net_device *dev, struct dcb_app *app)
1816 struct dcb_app_type *itr;
1819 spin_lock(&dcb_lock);
1820 if ((itr = dcb_app_lookup(app, dev->ifindex, 0)))
1821 prio |= 1 << itr->app.priority;
1822 spin_unlock(&dcb_lock);
1826 EXPORT_SYMBOL(dcb_ieee_getapp_mask);
1829 * dcb_ieee_setapp - add IEEE dcb application data to app list
1831 * This adds Application data to the list. Multiple application
1832 * entries may exists for the same selector and protocol as long
1833 * as the priorities are different.
1835 int dcb_ieee_setapp(struct net_device *dev, struct dcb_app *new)
1837 struct dcb_app_type event;
1840 event.ifindex = dev->ifindex;
1841 memcpy(&event.app, new, sizeof(event.app));
1842 if (dev->dcbnl_ops->getdcbx)
1843 event.dcbx = dev->dcbnl_ops->getdcbx(dev);
1845 spin_lock(&dcb_lock);
1846 /* Search for existing match and abort if found */
1847 if (dcb_app_lookup(new, dev->ifindex, new->priority)) {
1852 err = dcb_app_add(new, dev->ifindex);
1854 spin_unlock(&dcb_lock);
1856 call_dcbevent_notifiers(DCB_APP_EVENT, &event);
1859 EXPORT_SYMBOL(dcb_ieee_setapp);
1862 * dcb_ieee_delapp - delete IEEE dcb application data from list
1864 * This removes a matching APP data from the APP list
1866 int dcb_ieee_delapp(struct net_device *dev, struct dcb_app *del)
1868 struct dcb_app_type *itr;
1869 struct dcb_app_type event;
1872 event.ifindex = dev->ifindex;
1873 memcpy(&event.app, del, sizeof(event.app));
1874 if (dev->dcbnl_ops->getdcbx)
1875 event.dcbx = dev->dcbnl_ops->getdcbx(dev);
1877 spin_lock(&dcb_lock);
1878 /* Search for existing match and remove it. */
1879 if ((itr = dcb_app_lookup(del, dev->ifindex, del->priority))) {
1880 list_del(&itr->list);
1885 spin_unlock(&dcb_lock);
1887 call_dcbevent_notifiers(DCB_APP_EVENT, &event);
1890 EXPORT_SYMBOL(dcb_ieee_delapp);
1892 static void dcb_flushapp(void)
1894 struct dcb_app_type *app;
1895 struct dcb_app_type *tmp;
1897 spin_lock(&dcb_lock);
1898 list_for_each_entry_safe(app, tmp, &dcb_app_list, list) {
1899 list_del(&app->list);
1902 spin_unlock(&dcb_lock);
1905 static int __init dcbnl_init(void)
1907 INIT_LIST_HEAD(&dcb_app_list);
1909 rtnl_register(PF_UNSPEC, RTM_GETDCB, dcb_doit, NULL, NULL);
1910 rtnl_register(PF_UNSPEC, RTM_SETDCB, dcb_doit, NULL, NULL);
1914 module_init(dcbnl_init);
1916 static void __exit dcbnl_exit(void)
1918 rtnl_unregister(PF_UNSPEC, RTM_GETDCB);
1919 rtnl_unregister(PF_UNSPEC, RTM_SETDCB);
1922 module_exit(dcbnl_exit);