cregit-Linux how code gets into the kernel

Release 4.14 net/8021q/vlan_dev.c

Directory: net/8021q
/* -*- linux-c -*-
 * INET         802.1Q VLAN
 *              Ethernet-type device handling.
 *
 * Authors:     Ben Greear <greearb@candelatech.com>
 *              Please send support related email to: netdev@vger.kernel.org
 *              VLAN Home Page: http://www.candelatech.com/~greear/vlan.html
 *
 * Fixes:       Mar 22 2001: Martin Bokaemper <mbokaemper@unispherenetworks.com>
 *                - reset skb->pkt_type on incoming packets when MAC was changed
 *                - see that changed MAC is saddr for outgoing packets
 *              Oct 20, 2001:  Ard van Breeman:
 *                - Fix MC-list, finally.
 *                - Flush MC-list on VLAN destroy.
 *
 *
 *              This program is free software; you can redistribute it and/or
 *              modify it under the terms of the GNU General Public License
 *              as published by the Free Software Foundation; either version
 *              2 of the License, or (at your option) any later version.
 */


#define pr_fmt(fmt) KBUILD_MODNAME ": " fmt

#include <linux/module.h>
#include <linux/slab.h>
#include <linux/skbuff.h>
#include <linux/netdevice.h>
#include <linux/net_tstamp.h>
#include <linux/etherdevice.h>
#include <linux/ethtool.h>
#include <net/arp.h>
#include <net/switchdev.h>

#include "vlan.h"
#include "vlanproc.h"
#include <linux/if_vlan.h>
#include <linux/netpoll.h>

/*
 *      Create the VLAN header for an arbitrary protocol layer
 *
 *      saddr=NULL      means use device source address
 *      daddr=NULL      means leave destination address (eg unresolved arp)
 *
 *  This is called when the SKB is moving down the stack towards the
 *  physical devices.
 */

static int vlan_dev_hard_header(struct sk_buff *skb, struct net_device *dev, unsigned short type, const void *daddr, const void *saddr, unsigned int len) { struct vlan_dev_priv *vlan = vlan_dev_priv(dev); struct vlan_hdr *vhdr; unsigned int vhdrlen = 0; u16 vlan_tci = 0; int rc; if (!(vlan->flags & VLAN_FLAG_REORDER_HDR)) { vhdr = skb_push(skb, VLAN_HLEN); vlan_tci = vlan->vlan_id; vlan_tci |= vlan_dev_get_egress_qos_mask(dev, skb->priority); vhdr->h_vlan_TCI = htons(vlan_tci); /* * Set the protocol type. For a packet of type ETH_P_802_3/2 we * put the length in here instead. */ if (type != ETH_P_802_3 && type != ETH_P_802_2) vhdr->h_vlan_encapsulated_proto = htons(type); else vhdr->h_vlan_encapsulated_proto = htons(len); skb->protocol = vlan->vlan_proto; type = ntohs(vlan->vlan_proto); vhdrlen = VLAN_HLEN; } /* Before delegating work to the lower layer, enter our MAC-address */ if (saddr == NULL) saddr = dev->dev_addr; /* Now make the underlying real hard header */ dev = vlan->real_dev; rc = dev_hard_header(skb, dev, type, daddr, saddr, len + vhdrlen); if (rc > 0) rc += vhdrlen; return rc; }

Contributors

PersonTokensPropCommitsCommitProp
Linus Torvalds14969.63%17.14%
Patrick McHardy4621.50%642.86%
Jerome Borsboom52.34%17.14%
Octavian Purdila52.34%17.14%
Stephen Hemminger31.40%214.29%
Wang Sheng-Hui31.40%17.14%
Eyal Perry20.93%17.14%
David S. Miller10.47%17.14%
Total214100.00%14100.00%


static inline netdev_tx_t vlan_netpoll_send_skb(struct vlan_dev_priv *vlan, struct sk_buff *skb) { #ifdef CONFIG_NET_POLL_CONTROLLER if (vlan->netpoll) netpoll_send_skb(vlan->netpoll, skb); #else BUG(); #endif return NETDEV_TX_OK; }

Contributors

PersonTokensPropCommitsCommitProp
Américo Wang45100.00%1100.00%
Total45100.00%1100.00%


static netdev_tx_t vlan_dev_hard_start_xmit(struct sk_buff *skb, struct net_device *dev) { struct vlan_dev_priv *vlan = vlan_dev_priv(dev); struct vlan_ethhdr *veth = (struct vlan_ethhdr *)(skb->data); unsigned int len; int ret; /* Handle non-VLAN frames if they are sent to us, for example by DHCP. * * NOTE: THIS ASSUMES DIX ETHERNET, SPECIFICALLY NOT SUPPORTING * OTHER THINGS LIKE FDDI/TokenRing/802.3 SNAPs... */ if (veth->h_vlan_proto != vlan->vlan_proto || vlan->flags & VLAN_FLAG_REORDER_HDR) { u16 vlan_tci; vlan_tci = vlan->vlan_id; vlan_tci |= vlan_dev_get_egress_qos_mask(dev, skb->priority); __vlan_hwaccel_put_tag(skb, vlan->vlan_proto, vlan_tci); } skb->dev = vlan->real_dev; len = skb->len; if (unlikely(netpoll_tx_running(dev))) return vlan_netpoll_send_skb(vlan, skb); ret = dev_queue_xmit(skb); if (likely(ret == NET_XMIT_SUCCESS || ret == NET_XMIT_CN)) { struct vlan_pcpu_stats *stats; stats = this_cpu_ptr(vlan->vlan_pcpu_stats); u64_stats_update_begin(&stats->syncp); stats->tx_packets++; stats->tx_bytes += len; u64_stats_update_end(&stats->syncp); } else { this_cpu_inc(vlan->vlan_pcpu_stats->tx_dropped); } return ret; }

Contributors

PersonTokensPropCommitsCommitProp
Eric Dumazet8038.28%314.29%
Linus Torvalds5425.84%14.76%
Américo Wang209.57%14.76%
Patrick McHardy146.70%523.81%
Benjamin LaHaise136.22%14.76%
David S. Miller136.22%419.05%
Shmulik Hen52.39%14.76%
Joonwoo Park52.39%14.76%
Eyal Perry20.96%14.76%
John Fastabend10.48%14.76%
Wei Yongjun10.48%14.76%
Stephen Hemminger10.48%14.76%
Total209100.00%21100.00%


static int vlan_dev_change_mtu(struct net_device *dev, int new_mtu) { struct net_device *real_dev = vlan_dev_priv(dev)->real_dev; unsigned int max_mtu = real_dev->mtu; if (netif_reduces_vlan_mtu(real_dev)) max_mtu -= VLAN_HLEN; if (max_mtu < new_mtu) return -ERANGE; dev->mtu = new_mtu; return 0; }

Contributors

PersonTokensPropCommitsCommitProp
Paolo Abeni2640.62%116.67%
Linus Torvalds2031.25%116.67%
David S. Miller1523.44%116.67%
Patrick McHardy11.56%116.67%
Jiri Pirko11.56%116.67%
John W. Linville11.56%116.67%
Total64100.00%6100.00%


void vlan_dev_set_ingress_priority(const struct net_device *dev, u32 skb_prio, u16 vlan_prio) { struct vlan_dev_priv *vlan = vlan_dev_priv(dev); if (vlan->ingress_priority_map[vlan_prio & 0x7] && !skb_prio) vlan->nr_ingress_mappings--; else if (!vlan->ingress_priority_map[vlan_prio & 0x7] && skb_prio) vlan->nr_ingress_mappings++; vlan->ingress_priority_map[vlan_prio & 0x7] = skb_prio; }

Contributors

PersonTokensPropCommitsCommitProp
Patrick McHardy5267.53%360.00%
Linus Torvalds2329.87%120.00%
Jiri Pirko22.60%120.00%
Total77100.00%5100.00%


int vlan_dev_set_egress_priority(const struct net_device *dev, u32 skb_prio, u16 vlan_prio) { struct vlan_dev_priv *vlan = vlan_dev_priv(dev); struct vlan_priority_tci_mapping *mp = NULL; struct vlan_priority_tci_mapping *np; u32 vlan_qos = (vlan_prio << VLAN_PRIO_SHIFT) & VLAN_PRIO_MASK; /* See if a priority mapping exists.. */ mp = vlan->egress_priority_map[skb_prio & 0xF]; while (mp) { if (mp->priority == skb_prio) { if (mp->vlan_qos && !vlan_qos) vlan->nr_egress_mappings--; else if (!mp->vlan_qos && vlan_qos) vlan->nr_egress_mappings++; mp->vlan_qos = vlan_qos; return 0; } mp = mp->next; } /* Create a new mapping then. */ mp = vlan->egress_priority_map[skb_prio & 0xF]; np = kmalloc(sizeof(struct vlan_priority_tci_mapping), GFP_KERNEL); if (!np) return -ENOBUFS; np->next = mp; np->priority = skb_prio; np->vlan_qos = vlan_qos; /* Before inserting this element in hash table, make sure all its fields * are committed to memory. * coupled with smp_rmb() in vlan_dev_get_egress_qos_mask() */ smp_wmb(); vlan->egress_priority_map[skb_prio & 0xF] = np; if (vlan_qos) vlan->nr_egress_mappings++; return 0; }

Contributors

PersonTokensPropCommitsCommitProp
Linus Torvalds11657.71%111.11%
Patrick McHardy7135.32%333.33%
Björn Andersson62.99%111.11%
Eric Dumazet52.49%222.22%
Jiri Pirko21.00%111.11%
David S. Miller10.50%111.11%
Total201100.00%9100.00%

/* Flags are defined in the vlan_flags enum in include/linux/if_vlan.h file. */
int vlan_dev_change_flags(const struct net_device *dev, u32 flags, u32 mask) { struct vlan_dev_priv *vlan = vlan_dev_priv(dev); u32 old_flags = vlan->flags; if (mask & ~(VLAN_FLAG_REORDER_HDR | VLAN_FLAG_GVRP | VLAN_FLAG_LOOSE_BINDING | VLAN_FLAG_MVRP)) return -EINVAL; vlan->flags = (old_flags & ~mask) | (flags & mask); if (netif_running(dev) && (vlan->flags ^ old_flags) & VLAN_FLAG_GVRP) { if (vlan->flags & VLAN_FLAG_GVRP) vlan_gvrp_request_join(dev); else vlan_gvrp_request_leave(dev); } if (netif_running(dev) && (vlan->flags ^ old_flags) & VLAN_FLAG_MVRP) { if (vlan->flags & VLAN_FLAG_MVRP) vlan_mvrp_request_join(dev); else vlan_mvrp_request_leave(dev); } return 0; }

Contributors

PersonTokensPropCommitsCommitProp
Patrick McHardy8758.39%562.50%
David Ward4026.85%112.50%
Linus Torvalds2013.42%112.50%
Jiri Pirko21.34%112.50%
Total149100.00%8100.00%


void vlan_dev_get_realdev_name(const struct net_device *dev, char *result) { strncpy(result, vlan_dev_priv(dev)->real_dev->name, 23); }

Contributors

PersonTokensPropCommitsCommitProp
Ben Greear2683.87%133.33%
Patrick McHardy412.90%133.33%
Jiri Pirko13.23%133.33%
Total31100.00%3100.00%


bool vlan_dev_inherit_address(struct net_device *dev, struct net_device *real_dev) { if (dev->addr_assign_type != NET_ADDR_STOLEN) return false; ether_addr_copy(dev->dev_addr, real_dev->dev_addr); call_netdevice_notifiers(NETDEV_CHANGEADDR, dev); return true; }

Contributors

PersonTokensPropCommitsCommitProp
Mike Manning47100.00%1100.00%
Total47100.00%1100.00%


static int vlan_dev_open(struct net_device *dev) { struct vlan_dev_priv *vlan = vlan_dev_priv(dev); struct net_device *real_dev = vlan->real_dev; int err; if (!(real_dev->flags & IFF_UP) && !(vlan->flags & VLAN_FLAG_LOOSE_BINDING)) return -ENETDOWN; if (!ether_addr_equal(dev->dev_addr, real_dev->dev_addr) && !vlan_dev_inherit_address(dev, real_dev)) { err = dev_uc_add(real_dev, dev->dev_addr); if (err < 0) goto out; } if (dev->flags & IFF_ALLMULTI) { err = dev_set_allmulti(real_dev, 1); if (err < 0) goto del_unicast; } if (dev->flags & IFF_PROMISC) { err = dev_set_promiscuity(real_dev, 1); if (err < 0) goto clear_allmulti; } ether_addr_copy(vlan->real_dev_addr, real_dev->dev_addr); if (vlan->flags & VLAN_FLAG_GVRP) vlan_gvrp_request_join(dev); if (vlan->flags & VLAN_FLAG_MVRP) vlan_mvrp_request_join(dev); if (netif_carrier_ok(real_dev)) netif_carrier_on(dev); return 0; clear_allmulti: if (dev->flags & IFF_ALLMULTI) dev_set_allmulti(real_dev, -1); del_unicast: if (!ether_addr_equal(dev->dev_addr, real_dev->dev_addr)) dev_uc_del(real_dev, dev->dev_addr); out: netif_carrier_off(dev); return err; }

Contributors

PersonTokensPropCommitsCommitProp
Patrick McHardy10339.31%533.33%
Wang Chen8432.06%16.67%
David S. Miller2810.69%16.67%
David Ward134.96%16.67%
Jay Vosburgh103.82%16.67%
Mike Manning83.05%16.67%
Phil Oester72.67%16.67%
Joe Perches51.91%213.33%
Jiri Pirko41.53%213.33%
Total262100.00%15100.00%


static int vlan_dev_stop(struct net_device *dev) { struct vlan_dev_priv *vlan = vlan_dev_priv(dev); struct net_device *real_dev = vlan->real_dev; dev_mc_unsync(real_dev, dev); dev_uc_unsync(real_dev, dev); if (dev->flags & IFF_ALLMULTI) dev_set_allmulti(real_dev, -1); if (dev->flags & IFF_PROMISC) dev_set_promiscuity(real_dev, -1); if (!ether_addr_equal(dev->dev_addr, real_dev->dev_addr)) dev_uc_del(real_dev, dev->dev_addr); netif_carrier_off(dev); return 0; }

Contributors

PersonTokensPropCommitsCommitProp
Patrick McHardy7368.22%545.45%
David S. Miller1715.89%19.09%
Christopher Leech65.61%19.09%
Jay Vosburgh54.67%19.09%
Jiri Pirko43.74%218.18%
Joe Perches21.87%19.09%
Total107100.00%11100.00%


static int vlan_dev_set_mac_address(struct net_device *dev, void *p) { struct net_device *real_dev = vlan_dev_priv(dev)->real_dev; struct sockaddr *addr = p; int err; if (!is_valid_ether_addr(addr->sa_data)) return -EADDRNOTAVAIL; if (!(dev->flags & IFF_UP)) goto out; if (!ether_addr_equal(addr->sa_data, real_dev->dev_addr)) { err = dev_uc_add(real_dev, addr->sa_data); if (err < 0) return err; } if (!ether_addr_equal(dev->dev_addr, real_dev->dev_addr)) dev_uc_del(real_dev, dev->dev_addr); out: ether_addr_copy(dev->dev_addr, addr->sa_data); return 0; }

Contributors

PersonTokensPropCommitsCommitProp
Patrick McHardy13294.29%233.33%
Joe Perches53.57%233.33%
Jiri Pirko32.14%233.33%
Total140100.00%6100.00%


static int vlan_dev_ioctl(struct net_device *dev, struct ifreq *ifr, int cmd) { struct net_device *real_dev = vlan_dev_priv(dev)->real_dev; const struct net_device_ops *ops = real_dev->netdev_ops; struct ifreq ifrr; int err = -EOPNOTSUPP; strncpy(ifrr.ifr_name, real_dev->name, IFNAMSIZ); ifrr.ifr_ifru = ifr->ifr_ifru; switch (cmd) { case SIOCGMIIPHY: case SIOCGMIIREG: case SIOCSMIIREG: case SIOCSHWTSTAMP: case SIOCGHWTSTAMP: if (netif_device_present(real_dev) && ops->ndo_do_ioctl) err = ops->ndo_do_ioctl(real_dev, &ifrr, cmd); break; } if (!err) ifr->ifr_ifru = ifrr.ifr_ifru; return err; }

Contributors

PersonTokensPropCommitsCommitProp
Stephen Hemminger12794.07%350.00%
Stefan Sörensen64.44%116.67%
Jiri Pirko10.74%116.67%
Patrick McHardy10.74%116.67%
Total135100.00%6100.00%


static int vlan_dev_neigh_setup(struct net_device *dev, struct neigh_parms *pa) { struct net_device *real_dev = vlan_dev_priv(dev)->real_dev; const struct net_device_ops *ops = real_dev->netdev_ops; int err = 0; if (netif_device_present(real_dev) && ops->ndo_neigh_setup) err = ops->ndo_neigh_setup(real_dev, pa); return err; }

Contributors

PersonTokensPropCommitsCommitProp
Frank Blaschka6697.06%133.33%
David S. Miller11.47%133.33%
Jiri Pirko11.47%133.33%
Total68100.00%3100.00%

#if IS_ENABLED(CONFIG_FCOE)
static int vlan_dev_fcoe_ddp_setup(struct net_device *dev, u16 xid, struct scatterlist *sgl, unsigned int sgc) { struct net_device *real_dev = vlan_dev_priv(dev)->real_dev; const struct net_device_ops *ops = real_dev->netdev_ops; int rc = 0; if (ops->ndo_fcoe_ddp_setup) rc = ops->ndo_fcoe_ddp_setup(real_dev, xid, sgl, sgc); return rc; }

Contributors

PersonTokensPropCommitsCommitProp
Vasu Dev7398.65%150.00%
Jiri Pirko11.35%150.00%
Total74100.00%2100.00%


static int vlan_dev_fcoe_ddp_done(struct net_device *dev, u16 xid) { struct net_device *real_dev = vlan_dev_priv(dev)->real_dev; const struct net_device_ops *ops = real_dev->netdev_ops; int len = 0; if (ops->ndo_fcoe_ddp_done) len = ops->ndo_fcoe_ddp_done(real_dev, xid); return len; }

Contributors

PersonTokensPropCommitsCommitProp
Vasu Dev6098.36%150.00%
Jiri Pirko11.64%150.00%
Total61100.00%2100.00%


static int vlan_dev_fcoe_enable(struct net_device *dev) { struct net_device *real_dev = vlan_dev_priv(dev)->real_dev; const struct net_device_ops *ops = real_dev->netdev_ops; int rc = -EINVAL; if (ops->ndo_fcoe_enable) rc = ops->ndo_fcoe_enable(real_dev); return rc; }

Contributors

PersonTokensPropCommitsCommitProp
Yi Zou5698.25%150.00%
Jiri Pirko11.75%150.00%
Total57100.00%2100.00%


static int vlan_dev_fcoe_disable(struct net_device *dev) { struct net_device *real_dev = vlan_dev_priv(dev)->real_dev; const struct net_device_ops *ops = real_dev->netdev_ops; int rc = -EINVAL; if (ops->ndo_fcoe_disable) rc = ops->ndo_fcoe_disable(real_dev); return rc; }

Contributors

PersonTokensPropCommitsCommitProp
Yi Zou5698.25%150.00%
Jiri Pirko11.75%150.00%
Total57100.00%2100.00%


static int vlan_dev_fcoe_get_wwn(struct net_device *dev, u64 *wwn, int type) { struct net_device *real_dev = vlan_dev_priv(dev)->real_dev; const struct net_device_ops *ops = real_dev->netdev_ops; int rc = -EINVAL; if (ops->ndo_fcoe_get_wwn) rc = ops->ndo_fcoe_get_wwn(real_dev, wwn, type); return rc; }

Contributors

PersonTokensPropCommitsCommitProp
Yi Zou6798.53%150.00%
Jiri Pirko11.47%150.00%
Total68100.00%2100.00%


static int vlan_dev_fcoe_ddp_target(struct net_device *dev, u16 xid, struct scatterlist *sgl, unsigned int sgc) { struct net_device *real_dev = vlan_dev_priv(dev)->real_dev; const struct net_device_ops *ops = real_dev->netdev_ops; int rc = 0; if (ops->ndo_fcoe_ddp_target) rc = ops->ndo_fcoe_ddp_target(real_dev, xid, sgl, sgc); return rc; }

Contributors

PersonTokensPropCommitsCommitProp
Yi Zou7398.65%150.00%
Jiri Pirko11.35%150.00%
Total74100.00%2100.00%

#endif
static void vlan_dev_change_rx_flags(struct net_device *dev, int change) { struct net_device *real_dev = vlan_dev_priv(dev)->real_dev; if (dev->flags & IFF_UP) { if (change & IFF_ALLMULTI) dev_set_allmulti(real_dev, dev->flags & IFF_ALLMULTI ? 1 : -1); if (change & IFF_PROMISC) dev_set_promiscuity(real_dev, dev->flags & IFF_PROMISC ? 1 : -1); } }

Contributors

PersonTokensPropCommitsCommitProp
Patrick McHardy6986.25%250.00%
Matthijs Kooijman1012.50%125.00%
Jiri Pirko11.25%125.00%
Total80100.00%4100.00%


static void vlan_dev_set_rx_mode(struct net_device *vlan_dev) { dev_mc_sync(vlan_dev_priv(vlan_dev)->real_dev, vlan_dev); dev_uc_sync(vlan_dev_priv(vlan_dev)->real_dev, vlan_dev); }

Contributors

PersonTokensPropCommitsCommitProp
Christopher Leech1131.43%116.67%
David S. Miller1028.57%116.67%
Linus Torvalds925.71%116.67%
Jiri Pirko25.71%116.67%
Vlad Yasevich25.71%116.67%
Patrick McHardy12.86%116.67%
Total35100.00%6100.00%

/* * vlan network devices have devices nesting below it, and are a special * "super class" of normal network devices; split their locks off into a * separate class since they always nest. */ static struct lock_class_key vlan_netdev_xmit_lock_key; static struct lock_class_key vlan_netdev_addr_lock_key;
static void vlan_dev_set_lockdep_one(struct net_device *dev, struct netdev_queue *txq, void *_subclass) { lockdep_set_class_and_subclass(&txq->_xmit_lock, &vlan_netdev_xmit_lock_key, *(int *)_subclass); }

Contributors

PersonTokensPropCommitsCommitProp
David S. Miller38100.00%2100.00%
Total38100.00%2100.00%


static void vlan_dev_set_lockdep_class(struct net_device *dev, int subclass) { lockdep_set_class_and_subclass(&dev->addr_list_lock, &vlan_netdev_addr_lock_key, subclass); netdev_for_each_tx_queue(dev, vlan_dev_set_lockdep_one, &subclass); }

Contributors

PersonTokensPropCommitsCommitProp
David S. Miller37100.00%3100.00%
Total37100.00%3100.00%


static int vlan_dev_get_lock_subclass(struct net_device *dev) { return vlan_dev_priv(dev)->nest_level; }

Contributors

PersonTokensPropCommitsCommitProp
Vlad Yasevich19100.00%1100.00%
Total19100.00%1100.00%

static const struct header_ops vlan_header_ops = { .create = vlan_dev_hard_header, .parse = eth_header_parse, };
static int vlan_passthru_hard_header(struct sk_buff *skb, struct net_device *dev, unsigned short type, const void *daddr, const void *saddr, unsigned int len) { struct vlan_dev_priv *vlan = vlan_dev_priv(dev); struct net_device *real_dev = vlan->real_dev; if (saddr == NULL) saddr = dev->dev_addr; return dev_hard_header(skb, real_dev, type, daddr, saddr, len); }

Contributors

PersonTokensPropCommitsCommitProp
David S. Miller6985.19%150.00%
Peter Boström1214.81%150.00%
Total81100.00%2100.00%

static const struct header_ops vlan_passthru_header_ops = { .create = vlan_passthru_hard_header, .parse = eth_header_parse, }; static struct device_type vlan_type = { .name = "vlan", }; static const struct net_device_ops vlan_netdev_ops;
static int vlan_dev_init(struct net_device *dev) { struct net_device *real_dev = vlan_dev_priv(dev)->real_dev; netif_carrier_off(dev); /* IFF_BROADCAST|IFF_MULTICAST; ??? */ dev->flags = real_dev->flags & ~(IFF_UP | IFF_PROMISC | IFF_ALLMULTI | IFF_MASTER | IFF_SLAVE); dev->state = (real_dev->state & ((1<<__LINK_STATE_NOCARRIER) | (1<<__LINK_STATE_DORMANT))) | (1<<__LINK_STATE_PRESENT); dev->hw_features = NETIF_F_HW_CSUM | NETIF_F_SG | NETIF_F_FRAGLIST | NETIF_F_GSO_SOFTWARE | NETIF_F_HIGHDMA | NETIF_F_SCTP_CRC | NETIF_F_ALL_FCOE; dev->features |= dev->hw_features | NETIF_F_LLTX; dev->gso_max_size = real_dev->gso_max_size; dev->gso_max_segs = real_dev->gso_max_segs; if (dev->features & NETIF_F_VLAN_FEATURES) netdev_warn(real_dev, "VLAN features are set incorrectly. Q-in-Q configurations may not work correctly.\n"); dev->vlan_features = real_dev->vlan_features & ~NETIF_F_ALL_FCOE; /* ipv6 shared card related stuff */ dev->dev_id = real_dev->dev_id; if (is_zero_ether_addr(dev->dev_addr)) { ether_addr_copy(dev->dev_addr, real_dev->dev_addr); dev->addr_assign_type = NET_ADDR_STOLEN; } if (is_zero_ether_addr(dev->broadcast)) memcpy(dev->broadcast, real_dev->broadcast, dev->addr_len); #if IS_ENABLED(CONFIG_FCOE) dev->fcoe_ddp_xid = real_dev->fcoe_ddp_xid; #endif dev->needed_headroom = real_dev->needed_headroom; if (vlan_hw_offload_capable(real_dev->features, vlan_dev_priv(dev)->vlan_proto)) { dev->header_ops = &vlan_passthru_header_ops; dev->hard_header_len = real_dev->hard_header_len; } else { dev->header_ops = &vlan_header_ops; dev->hard_header_len = real_dev->hard_header_len + VLAN_HLEN; } dev->netdev_ops = &vlan_netdev_ops; SET_NETDEV_DEVTYPE(dev, &vlan_type); vlan_dev_set_lockdep_class(dev, vlan_dev_get_lock_subclass(dev)); vlan_dev_priv(dev)->vlan_pcpu_stats = netdev_alloc_pcpu_stats(struct vlan_pcpu_stats); if (!vlan_dev_priv(dev)->vlan_pcpu_stats) return -ENOMEM; return 0; }

Contributors

PersonTokensPropCommitsCommitProp
Patrick McHardy17350.44%310.00%
Eric Dumazet4914.29%516.67%
Vlad Yasevich298.45%310.00%
Vasu Dev154.37%13.33%
Michał Mirosław144.08%26.67%
Mike Manning133.79%13.33%
Toshiaki Makita123.50%26.67%
Doug Goldstein82.33%13.33%
Alexander Duyck82.33%13.33%
Jay Vosburgh51.46%13.33%
John Fastabend51.46%26.67%
Jiri Pirko30.87%13.33%
David S. Miller30.87%26.67%
Américo Wang20.58%26.67%
Andrey Vagin20.58%13.33%
Tom Herbert20.58%26.67%
Total343100.00%30100.00%


static void vlan_dev_uninit(struct net_device *dev) { struct vlan_priority_tci_mapping *pm; struct vlan_dev_priv *vlan = vlan_dev_priv(dev); int i; for (i = 0; i < ARRAY_SIZE(vlan->egress_priority_map); i++) { while ((pm = vlan->egress_priority_map[i]) != NULL) { vlan->egress_priority_map[i] = pm->next; kfree(pm); } } }

Contributors

PersonTokensPropCommitsCommitProp
Pavel Emelyanov8097.56%150.00%
Jiri Pirko22.44%150.00%
Total82100.00%2100.00%


static netdev_features_t vlan_dev_fix_features(struct net_device *dev, netdev_features_t features) { struct net_device *real_dev = vlan_dev_priv(dev)->real_dev; netdev_features_t old_features = features; netdev_features_t lower_features; lower_features = netdev_intersect_features((real_dev->vlan_features | NETIF_F_RXCSUM), real_dev->features); /* Add HW_CSUM setting to preserve user ability to control * checksum offload on the vlan device. */ if (lower_features & (NETIF_F_IP_CSUM|NETIF_F_IPV6_CSUM)) lower_features |= NETIF_F_HW_CSUM; features = netdev_intersect_features(features, lower_features); features |= old_features & (NETIF_F_SOFT_FEATURES | NETIF_F_GSO_SOFTWARE); features |= NETIF_F_LLTX; return features; }

Contributors

PersonTokensPropCommitsCommitProp
Michał Mirosław4043.48%440.00%
Vlad Yasevich2628.26%110.00%
Shan Wei1010.87%110.00%
Michal Kubeček1010.87%110.00%
Toshiaki Makita44.35%110.00%
Björn Mork11.09%110.00%
Jiri Pirko11.09%110.00%
Total92100.00%10100.00%


static int vlan_ethtool_get_link_ksettings(struct net_device *dev, struct ethtool_link_ksettings *cmd) { const struct vlan_dev_priv *vlan = vlan_dev_priv(dev); return __ethtool_get_link_ksettings(vlan->real_dev, cmd); }

Contributors

PersonTokensPropCommitsCommitProp
Stephen Hemminger3081.08%125.00%
David Decotigny38.11%125.00%
Patrick McHardy25.41%125.00%
Jiri Pirko25.41%125.00%
Total37100.00%4100.00%


static void vlan_ethtool_get_drvinfo(struct net_device *dev, struct ethtool_drvinfo *info) { strlcpy(info->driver, vlan_fullname, sizeof(info->driver)); strlcpy(info->version, vlan_version, sizeof(info->version)); strlcpy(info->fw_version, "N/A", sizeof(info->fw_version)); }

Contributors

PersonTokensPropCommitsCommitProp
Stephen Hemminger4062.50%150.00%
Jiri Pirko2437.50%150.00%
Total64100.00%2100.00%


static int vlan_ethtool_get_ts_info(struct net_device *dev, struct ethtool_ts_info *info) { const struct vlan_dev_priv *vlan = vlan_dev_priv(dev); const struct ethtool_ops *ops = vlan->real_dev->ethtool_ops; if (ops->get_ts_info) { return ops->get_ts_info(vlan->real_dev, info); } else { info->so_timestamping = SOF_TIMESTAMPING_RX_SOFTWARE | SOF_TIMESTAMPING_SOFTWARE; info->phc_index = -1; } return 0; }

Contributors

PersonTokensPropCommitsCommitProp
Richard Cochran80100.00%1100.00%
Total80100.00%1100.00%


static void vlan_dev_get_stats64(struct net_device *dev, struct rtnl_link_stats64 *stats) { struct vlan_pcpu_stats *p; u32 rx_errors = 0, tx_dropped = 0; int i; for_each_possible_cpu(i) { u64 rxpackets, rxbytes, rxmulticast, txpackets, txbytes; unsigned int start; p = per_cpu_ptr(vlan_dev_priv(dev)->vlan_pcpu_stats, i); do { start = u64_stats_fetch_begin_irq(&p->syncp); rxpackets = p->rx_packets; rxbytes = p->rx_bytes; rxmulticast = p->rx_multicast; txpackets = p->tx_packets; txbytes = p->tx_bytes; } while (u64_stats_fetch_retry_irq(&p->syncp, start)); stats->rx_packets += rxpackets; stats->rx_bytes += rxbytes; stats->multicast += rxmulticast; stats->tx_packets += txpackets; stats->tx_bytes += txbytes; /* rx_errors & tx_dropped are u32 */ rx_errors += p->rx_errors; tx_dropped += p->tx_dropped; } stats->rx_errors = rx_errors; stats->tx_dropped = tx_dropped; }

Contributors

PersonTokensPropCommitsCommitProp
Eric Dumazet17597.77%457.14%
Eric W. Biedermann21.12%114.29%
Jiri Pirko10.56%114.29%
Stephen Hemminger10.56%114.29%
Total179100.00%7100.00%

#ifdef CONFIG_NET_POLL_CONTROLLER
static void vlan_dev_poll_controller(struct net_device *dev) { return; }

Contributors

PersonTokensPropCommitsCommitProp
Benjamin LaHaise1191.67%150.00%
Eric Dumazet18.33%150.00%
Total12100.00%2100.00%


static int vlan_dev_netpoll_setup(struct net_device *dev, struct netpoll_info *npinfo) { struct vlan_dev_priv *vlan = vlan_dev_priv(dev); struct net_device *real_dev = vlan->real_dev; struct netpoll *netpoll; int err = 0; netpoll = kzalloc(sizeof(*netpoll), GFP_KERNEL); err = -ENOMEM; if (!netpoll) goto out; err = __netpoll_setup(netpoll, real_dev); if (err) { kfree(netpoll); goto out; } vlan->netpoll = netpoll; out: return err; }

Contributors

PersonTokensPropCommitsCommitProp
Benjamin LaHaise9893.33%120.00%
Américo Wang32.86%120.00%
Jiri Pirko21.90%120.00%
Eric Dumazet10.95%120.00%
Eric W. Biedermann10.95%120.00%
Total105100.00%5100.00%


static void vlan_dev_netpoll_cleanup(struct net_device *dev) { struct vlan_dev_priv *vlan= vlan_dev_priv(dev); struct netpoll *netpoll = vlan->netpoll; if (!netpoll) return; vlan->netpoll = NULL; __netpoll_free_async(netpoll); }

Contributors

PersonTokensPropCommitsCommitProp
Benjamin LaHaise4289.36%125.00%
Américo Wang36.38%125.00%
Neil Horman12.13%125.00%
Eric Dumazet12.13%125.00%
Total47100.00%4100.00%

#endif /* CONFIG_NET_POLL_CONTROLLER */
static int vlan_dev_get_iflink(const struct net_device *dev) { struct net_device *real_dev = vlan_dev_priv(dev)->real_dev; return real_dev->ifindex; }

Contributors

PersonTokensPropCommitsCommitProp
Nicolas Dichtel29100.00%1100.00%
Total29100.00%1100.00%

static const struct ethtool_ops vlan_ethtool_ops = { .get_link_ksettings = vlan_ethtool_get_link_ksettings, .get_drvinfo = vlan_ethtool_get_drvinfo, .get_link = ethtool_op_get_link, .get_ts_info = vlan_ethtool_get_ts_info, }; static const struct net_device_ops vlan_netdev_ops = { .ndo_change_mtu = vlan_dev_change_mtu, .ndo_init = vlan_dev_init, .ndo_uninit = vlan_dev_uninit, .ndo_open = vlan_dev_open, .ndo_stop = vlan_dev_stop, .ndo_start_xmit = vlan_dev_hard_start_xmit, .ndo_validate_addr = eth_validate_addr, .ndo_set_mac_address = vlan_dev_set_mac_address, .ndo_set_rx_mode = vlan_dev_set_rx_mode, .ndo_change_rx_flags = vlan_dev_change_rx_flags, .ndo_do_ioctl = vlan_dev_ioctl, .ndo_neigh_setup = vlan_dev_neigh_setup, .ndo_get_stats64 = vlan_dev_get_stats64, #if IS_ENABLED(CONFIG_FCOE) .ndo_fcoe_ddp_setup = vlan_dev_fcoe_ddp_setup, .ndo_fcoe_ddp_done = vlan_dev_fcoe_ddp_done, .ndo_fcoe_enable = vlan_dev_fcoe_enable, .ndo_fcoe_disable = vlan_dev_fcoe_disable, .ndo_fcoe_get_wwn = vlan_dev_fcoe_get_wwn, .ndo_fcoe_ddp_target = vlan_dev_fcoe_ddp_target, #endif #ifdef CONFIG_NET_POLL_CONTROLLER .ndo_poll_controller = vlan_dev_poll_controller, .ndo_netpoll_setup = vlan_dev_netpoll_setup, .ndo_netpoll_cleanup = vlan_dev_netpoll_cleanup, #endif .ndo_fix_features = vlan_dev_fix_features, .ndo_get_lock_subclass = vlan_dev_get_lock_subclass, .ndo_get_iflink = vlan_dev_get_iflink, };
static void vlan_dev_free(struct net_device *dev) { struct vlan_dev_priv *vlan = vlan_dev_priv(dev); free_percpu(vlan->vlan_pcpu_stats); vlan->vlan_pcpu_stats = NULL; }

Contributors

PersonTokensPropCommitsCommitProp
Eric Dumazet34100.00%1100.00%
Total34100.00%1100.00%


void vlan_setup(struct net_device *dev) { ether_setup(dev); dev->priv_flags |= IFF_802_1Q_VLAN | IFF_NO_QUEUE; dev->priv_flags |= IFF_UNICAST_FLT; dev->priv_flags &= ~IFF_TX_SKB_SHARING; netif_keep_dst(dev); dev->netdev_ops = &vlan_netdev_ops; dev->needs_free_netdev = true; dev->priv_destructor = vlan_dev_free; dev->ethtool_ops = &vlan_ethtool_ops; dev->min_mtu = 0; dev->max_mtu = ETH_MAX_MTU; eth_zero_addr(dev->broadcast); }

Contributors

PersonTokensPropCommitsCommitProp
Patrick McHardy4248.84%216.67%
Eric Dumazet1213.95%325.00%
Jarod Wilson1213.95%18.33%
David S. Miller78.14%18.33%
Zhang Shengju66.98%18.33%
Stephen Hemminger33.49%18.33%
Phil Sutter22.33%18.33%
Joe Perches11.16%18.33%
Neil Horman11.16%18.33%
Total86100.00%12100.00%


Overall Contributors

PersonTokensPropCommitsCommitProp
Patrick McHardy91323.28%2217.05%
Linus Torvalds41510.58%10.78%
Eric Dumazet3819.71%1511.63%
Yi Zou2726.94%32.33%
David S. Miller2636.71%107.75%
Stephen Hemminger2636.71%86.20%
Benjamin LaHaise1934.92%10.78%
Vasu Dev1734.41%10.78%
Richard Cochran882.24%10.78%
Wang Chen842.14%10.78%
Vlad Yasevich812.07%43.10%
Pavel Emelyanov802.04%10.78%
Américo Wang751.91%43.10%
Frank Blaschka711.81%10.78%
Mike Manning681.73%10.78%
Jiri Pirko651.66%43.10%
Michał Mirosław591.50%53.88%
David Ward531.35%10.78%
Nicolas Dichtel340.87%10.78%
Paolo Abeni260.66%10.78%
Ben Greear260.66%10.78%
Doug Goldstein200.51%10.78%
Joe Perches200.51%43.10%
Jay Vosburgh200.51%10.78%
Christopher Leech170.43%10.78%
Toshiaki Makita160.41%21.55%
Jarod Wilson120.31%10.78%
Peter Boström120.31%10.78%
Matthijs Kooijman100.25%10.78%
Michal Kubeček100.25%10.78%
Shan Wei100.25%10.78%
Alexander Duyck80.20%10.78%
Phil Oester70.18%10.78%
John Fastabend60.15%21.55%
Björn Andersson60.15%10.78%
Stefan Sörensen60.15%10.78%
Zhang Shengju60.15%10.78%
Joonwoo Park50.13%10.78%
David Decotigny50.13%10.78%
Shmulik Hen50.13%10.78%
Octavian Purdila50.13%10.78%
Jerome Borsboom50.13%10.78%
Eyal Perry40.10%10.78%
Ido Schimmel30.08%10.78%
Wang Sheng-Hui30.08%10.78%
Tejun Heo30.08%10.78%
Eric W. Biedermann30.08%21.55%
Phil Sutter20.05%10.78%
Tom Herbert20.05%21.55%
Neil Horman20.05%21.55%
Andrey Vagin20.05%10.78%
John W. Linville10.03%10.78%
Björn Mork10.03%10.78%
Hideaki Yoshifuji / 吉藤英明10.03%10.78%
Wei Yongjun10.03%10.78%
Total3922100.00%129100.00%
Directory: net/8021q
Information contained on this website is for historical information purposes only and does not indicate or represent copyright ownership.
Created with cregit.