4 * Copyright (C) 2007 OpenVZ http://openvz.org, SWsoft Inc
6 * Author: Pavel Emelianov <xemul@openvz.org>
7 * Ethtool interface from: Eric W. Biederman <ebiederm@xmission.com>
11 #include <linux/netdevice.h>
12 #include <linux/slab.h>
13 #include <linux/ethtool.h>
14 #include <linux/etherdevice.h>
15 #include <linux/u64_stats_sync.h>
17 #include <net/rtnetlink.h>
20 #include <linux/veth.h>
21 #include <linux/module.h>
23 #define DRV_NAME "veth"
24 #define DRV_VERSION "1.0"
26 #define MIN_MTU 68 /* Min L3 MTU */
27 #define MAX_MTU 65535 /* Max L3 MTU (arbitrary) */
32 struct u64_stats_sync syncp;
36 struct net_device __rcu *peer;
45 const char string[ETH_GSTRING_LEN];
46 } ethtool_stats_keys[] = {
50 static int veth_get_settings(struct net_device *dev, struct ethtool_cmd *cmd)
54 ethtool_cmd_speed_set(cmd, SPEED_10000);
55 cmd->duplex = DUPLEX_FULL;
58 cmd->transceiver = XCVR_INTERNAL;
59 cmd->autoneg = AUTONEG_DISABLE;
65 static void veth_get_drvinfo(struct net_device *dev, struct ethtool_drvinfo *info)
67 strlcpy(info->driver, DRV_NAME, sizeof(info->driver));
68 strlcpy(info->version, DRV_VERSION, sizeof(info->version));
71 static void veth_get_strings(struct net_device *dev, u32 stringset, u8 *buf)
75 memcpy(buf, ðtool_stats_keys, sizeof(ethtool_stats_keys));
80 static int veth_get_sset_count(struct net_device *dev, int sset)
84 return ARRAY_SIZE(ethtool_stats_keys);
90 static void veth_get_ethtool_stats(struct net_device *dev,
91 struct ethtool_stats *stats, u64 *data)
93 struct veth_priv *priv = netdev_priv(dev);
94 struct net_device *peer = rtnl_dereference(priv->peer);
96 data[0] = peer ? peer->ifindex : 0;
99 static const struct ethtool_ops veth_ethtool_ops = {
100 .get_settings = veth_get_settings,
101 .get_drvinfo = veth_get_drvinfo,
102 .get_link = ethtool_op_get_link,
103 .get_strings = veth_get_strings,
104 .get_sset_count = veth_get_sset_count,
105 .get_ethtool_stats = veth_get_ethtool_stats,
108 static netdev_tx_t veth_xmit(struct sk_buff *skb, struct net_device *dev)
110 struct veth_priv *priv = netdev_priv(dev);
111 struct net_device *rcv;
112 int length = skb->len;
115 rcv = rcu_dereference(priv->peer);
116 if (unlikely(!rcv)) {
120 /* don't change ip_summed == CHECKSUM_PARTIAL, as that
121 * will cause bad checksum on forwarded packets
123 if (skb->ip_summed == CHECKSUM_NONE &&
124 rcv->features & NETIF_F_RXCSUM)
125 skb->ip_summed = CHECKSUM_UNNECESSARY;
127 if (likely(dev_forward_skb(rcv, skb) == NET_RX_SUCCESS)) {
128 struct pcpu_vstats *stats = this_cpu_ptr(dev->vstats);
130 u64_stats_update_begin(&stats->syncp);
131 stats->bytes += length;
133 u64_stats_update_end(&stats->syncp);
136 atomic64_inc(&priv->dropped);
146 static u64 veth_stats_one(struct pcpu_vstats *result, struct net_device *dev)
148 struct veth_priv *priv = netdev_priv(dev);
153 for_each_possible_cpu(cpu) {
154 struct pcpu_vstats *stats = per_cpu_ptr(dev->vstats, cpu);
159 start = u64_stats_fetch_begin_irq(&stats->syncp);
160 packets = stats->packets;
161 bytes = stats->bytes;
162 } while (u64_stats_fetch_retry_irq(&stats->syncp, start));
163 result->packets += packets;
164 result->bytes += bytes;
166 return atomic64_read(&priv->dropped);
169 static struct rtnl_link_stats64 *veth_get_stats64(struct net_device *dev,
170 struct rtnl_link_stats64 *tot)
172 struct veth_priv *priv = netdev_priv(dev);
173 struct net_device *peer;
174 struct pcpu_vstats one;
176 tot->tx_dropped = veth_stats_one(&one, dev);
177 tot->tx_bytes = one.bytes;
178 tot->tx_packets = one.packets;
181 peer = rcu_dereference(priv->peer);
183 tot->rx_dropped = veth_stats_one(&one, peer);
184 tot->rx_bytes = one.bytes;
185 tot->rx_packets = one.packets;
192 /* fake multicast ability */
193 static void veth_set_multicast_list(struct net_device *dev)
197 static int veth_open(struct net_device *dev)
199 struct veth_priv *priv = netdev_priv(dev);
200 struct net_device *peer = rtnl_dereference(priv->peer);
205 if (peer->flags & IFF_UP) {
206 netif_carrier_on(dev);
207 netif_carrier_on(peer);
212 static int veth_close(struct net_device *dev)
214 struct veth_priv *priv = netdev_priv(dev);
215 struct net_device *peer = rtnl_dereference(priv->peer);
217 netif_carrier_off(dev);
219 netif_carrier_off(peer);
224 static int is_valid_veth_mtu(int new_mtu)
226 return new_mtu >= MIN_MTU && new_mtu <= MAX_MTU;
229 static int veth_change_mtu(struct net_device *dev, int new_mtu)
231 if (!is_valid_veth_mtu(new_mtu))
237 static int veth_dev_init(struct net_device *dev)
239 dev->vstats = netdev_alloc_pcpu_stats(struct pcpu_vstats);
245 static void veth_dev_free(struct net_device *dev)
247 free_percpu(dev->vstats);
251 #ifdef CONFIG_NET_POLL_CONTROLLER
252 static void veth_poll_controller(struct net_device *dev)
254 /* veth only receives frames when its peer sends one
255 * Since it's a synchronous operation, we are guaranteed
256 * never to have pending data when we poll for it so
257 * there is nothing to do here.
259 * We need this though so netpoll recognizes us as an interface that
260 * supports polling, which enables bridge devices in virt setups to
261 * still use netconsole
264 #endif /* CONFIG_NET_POLL_CONTROLLER */
266 static const struct net_device_ops veth_netdev_ops = {
267 .ndo_init = veth_dev_init,
268 .ndo_open = veth_open,
269 .ndo_stop = veth_close,
270 .ndo_start_xmit = veth_xmit,
271 .ndo_change_mtu = veth_change_mtu,
272 .ndo_get_stats64 = veth_get_stats64,
273 .ndo_set_rx_mode = veth_set_multicast_list,
274 .ndo_set_mac_address = eth_mac_addr,
275 #ifdef CONFIG_NET_POLL_CONTROLLER
276 .ndo_poll_controller = veth_poll_controller,
280 #define VETH_FEATURES (NETIF_F_SG | NETIF_F_FRAGLIST | NETIF_F_ALL_TSO | \
281 NETIF_F_HW_CSUM | NETIF_F_RXCSUM | NETIF_F_HIGHDMA | \
282 NETIF_F_GSO_GRE | NETIF_F_GSO_UDP_TUNNEL | \
283 NETIF_F_GSO_IPIP | NETIF_F_GSO_SIT | NETIF_F_UFO | \
284 NETIF_F_HW_VLAN_CTAG_TX | NETIF_F_HW_VLAN_CTAG_RX | \
285 NETIF_F_HW_VLAN_STAG_TX | NETIF_F_HW_VLAN_STAG_RX )
287 static void veth_setup(struct net_device *dev)
291 dev->priv_flags &= ~IFF_TX_SKB_SHARING;
292 dev->priv_flags |= IFF_LIVE_ADDR_CHANGE;
294 dev->netdev_ops = &veth_netdev_ops;
295 dev->ethtool_ops = &veth_ethtool_ops;
296 dev->features |= NETIF_F_LLTX;
297 dev->features |= VETH_FEATURES;
298 dev->vlan_features = dev->features &
299 ~(NETIF_F_HW_VLAN_CTAG_TX |
300 NETIF_F_HW_VLAN_STAG_TX |
301 NETIF_F_HW_VLAN_CTAG_RX |
302 NETIF_F_HW_VLAN_STAG_RX);
303 dev->destructor = veth_dev_free;
305 dev->hw_features = VETH_FEATURES;
306 dev->hw_enc_features = VETH_FEATURES;
313 static int veth_validate(struct nlattr *tb[], struct nlattr *data[])
315 if (tb[IFLA_ADDRESS]) {
316 if (nla_len(tb[IFLA_ADDRESS]) != ETH_ALEN)
318 if (!is_valid_ether_addr(nla_data(tb[IFLA_ADDRESS])))
319 return -EADDRNOTAVAIL;
322 if (!is_valid_veth_mtu(nla_get_u32(tb[IFLA_MTU])))
328 static struct rtnl_link_ops veth_link_ops;
330 static int veth_newlink(struct net *src_net, struct net_device *dev,
331 struct nlattr *tb[], struct nlattr *data[])
334 struct net_device *peer;
335 struct veth_priv *priv;
336 char ifname[IFNAMSIZ];
337 struct nlattr *peer_tb[IFLA_MAX + 1], **tbp;
338 unsigned char name_assign_type;
339 struct ifinfomsg *ifmp;
343 * create and register peer first
345 if (data != NULL && data[VETH_INFO_PEER] != NULL) {
346 struct nlattr *nla_peer;
348 nla_peer = data[VETH_INFO_PEER];
349 ifmp = nla_data(nla_peer);
350 err = rtnl_nla_parse_ifla(peer_tb,
351 nla_data(nla_peer) + sizeof(struct ifinfomsg),
352 nla_len(nla_peer) - sizeof(struct ifinfomsg));
356 err = veth_validate(peer_tb, NULL);
366 if (tbp[IFLA_IFNAME]) {
367 nla_strlcpy(ifname, tbp[IFLA_IFNAME], IFNAMSIZ);
368 name_assign_type = NET_NAME_USER;
370 snprintf(ifname, IFNAMSIZ, DRV_NAME "%%d");
371 name_assign_type = NET_NAME_ENUM;
374 net = rtnl_link_get_net(src_net, tbp);
378 peer = rtnl_create_link(net, ifname, name_assign_type,
379 &veth_link_ops, tbp);
382 return PTR_ERR(peer);
385 if (tbp[IFLA_ADDRESS] == NULL)
386 eth_hw_addr_random(peer);
388 if (ifmp && (dev->ifindex != 0))
389 peer->ifindex = ifmp->ifi_index;
391 err = register_netdevice(peer);
395 goto err_register_peer;
397 netif_carrier_off(peer);
399 err = rtnl_configure_link(peer, ifmp);
401 goto err_configure_peer;
406 * note, that since we've registered new device the dev's name
407 * should be re-allocated
410 if (tb[IFLA_ADDRESS] == NULL)
411 eth_hw_addr_random(dev);
414 nla_strlcpy(dev->name, tb[IFLA_IFNAME], IFNAMSIZ);
416 snprintf(dev->name, IFNAMSIZ, DRV_NAME "%%d");
418 err = register_netdevice(dev);
420 goto err_register_dev;
422 netif_carrier_off(dev);
425 * tie the deviced together
428 priv = netdev_priv(dev);
429 rcu_assign_pointer(priv->peer, peer);
431 priv = netdev_priv(peer);
432 rcu_assign_pointer(priv->peer, dev);
438 unregister_netdevice(peer);
446 static void veth_dellink(struct net_device *dev, struct list_head *head)
448 struct veth_priv *priv;
449 struct net_device *peer;
451 priv = netdev_priv(dev);
452 peer = rtnl_dereference(priv->peer);
454 /* Note : dellink() is called from default_device_exit_batch(),
455 * before a rcu_synchronize() point. The devices are guaranteed
456 * not being freed before one RCU grace period.
458 RCU_INIT_POINTER(priv->peer, NULL);
459 unregister_netdevice_queue(dev, head);
462 priv = netdev_priv(peer);
463 RCU_INIT_POINTER(priv->peer, NULL);
464 unregister_netdevice_queue(peer, head);
468 static const struct nla_policy veth_policy[VETH_INFO_MAX + 1] = {
469 [VETH_INFO_PEER] = { .len = sizeof(struct ifinfomsg) },
472 static struct rtnl_link_ops veth_link_ops = {
474 .priv_size = sizeof(struct veth_priv),
476 .validate = veth_validate,
477 .newlink = veth_newlink,
478 .dellink = veth_dellink,
479 .policy = veth_policy,
480 .maxtype = VETH_INFO_MAX,
487 static __init int veth_init(void)
489 return rtnl_link_register(&veth_link_ops);
492 static __exit void veth_exit(void)
494 rtnl_link_unregister(&veth_link_ops);
497 module_init(veth_init);
498 module_exit(veth_exit);
500 MODULE_DESCRIPTION("Virtual Ethernet Tunnel");
501 MODULE_LICENSE("GPL v2");
502 MODULE_ALIAS_RTNL_LINK(DRV_NAME);