Loading net/core/dev.c +31 −88 Original line number Diff line number Diff line Loading @@ -3096,54 +3096,23 @@ void netdev_rx_handler_unregister(struct net_device *dev) } EXPORT_SYMBOL_GPL(netdev_rx_handler_unregister); static inline void skb_bond_set_mac_by_master(struct sk_buff *skb, struct net_device *master) static void vlan_on_bond_hook(struct sk_buff *skb) { if (skb->pkt_type == PACKET_HOST) { u16 *dest = (u16 *) eth_hdr(skb)->h_dest; memcpy(dest, master->dev_addr, ETH_ALEN); } } /* On bonding slaves other than the currently active slave, suppress * duplicates except for 802.3ad ETH_P_SLOW, alb non-mcast/bcast, and * ARP on active-backup slaves with arp_validate enabled. */ static int __skb_bond_should_drop(struct sk_buff *skb, struct net_device *master) { struct net_device *dev = skb->dev; if (master->priv_flags & IFF_MASTER_ARPMON) dev->last_rx = jiffies; if ((master->priv_flags & IFF_MASTER_ALB) && (master->priv_flags & IFF_BRIDGE_PORT)) { /* Do address unmangle. The local destination address * will be always the one master has. Provides the right * functionality in a bridge. /* * Make sure ARP frames received on VLAN interfaces stacked on * bonding interfaces still make their way to any base bonding * device that may have registered for a specific ptype. */ skb_bond_set_mac_by_master(skb, master); } if (skb->dev->priv_flags & IFF_802_1Q_VLAN && vlan_dev_real_dev(skb->dev)->priv_flags & IFF_BONDING && skb->protocol == htons(ETH_P_ARP)) { struct sk_buff *skb2 = skb_clone(skb, GFP_ATOMIC); if (dev->priv_flags & IFF_SLAVE_INACTIVE) { if ((dev->priv_flags & IFF_SLAVE_NEEDARP) && skb->protocol == __cpu_to_be16(ETH_P_ARP)) return 0; if (master->priv_flags & IFF_MASTER_ALB) { if (skb->pkt_type != PACKET_BROADCAST && skb->pkt_type != PACKET_MULTICAST) return 0; } if (master->priv_flags & IFF_MASTER_8023AD && skb->protocol == __cpu_to_be16(ETH_P_SLOW)) return 0; return 1; if (!skb2) return; skb2->dev = vlan_dev_real_dev(skb->dev); netif_rx(skb2); } return 0; } static int __netif_receive_skb(struct sk_buff *skb) Loading @@ -3151,8 +3120,7 @@ static int __netif_receive_skb(struct sk_buff *skb) struct packet_type *ptype, *pt_prev; rx_handler_func_t *rx_handler; struct net_device *orig_dev; struct net_device *null_or_orig; struct net_device *orig_or_bond; struct net_device *null_or_dev; int ret = NET_RX_DROP; __be16 type; Loading @@ -3167,32 +3135,8 @@ static int __netif_receive_skb(struct sk_buff *skb) if (!skb->skb_iif) skb->skb_iif = skb->dev->ifindex; /* * bonding note: skbs received on inactive slaves should only * be delivered to pkt handlers that are exact matches. Also * the deliver_no_wcard flag will be set. If packet handlers * are sensitive to duplicate packets these skbs will need to * be dropped at the handler. */ null_or_orig = NULL; orig_dev = skb->dev; if (skb->deliver_no_wcard) null_or_orig = orig_dev; else if (netif_is_bond_slave(orig_dev)) { struct net_device *bond_master = ACCESS_ONCE(orig_dev->master); if (likely(bond_master)) { if (__skb_bond_should_drop(skb, bond_master)) { skb->deliver_no_wcard = 1; /* deliver only exact match */ null_or_orig = orig_dev; } else skb->dev = bond_master; } } __this_cpu_inc(softnet_data.processed); skb_reset_network_header(skb); skb_reset_transport_header(skb); skb->mac_len = skb->network_header - skb->mac_header; Loading @@ -3201,6 +3145,10 @@ static int __netif_receive_skb(struct sk_buff *skb) rcu_read_lock(); another_round: __this_cpu_inc(softnet_data.processed); #ifdef CONFIG_NET_CLS_ACT if (skb->tc_verd & TC_NCLS) { skb->tc_verd = CLR_TC_NCLS(skb->tc_verd); Loading @@ -3209,8 +3157,7 @@ static int __netif_receive_skb(struct sk_buff *skb) #endif list_for_each_entry_rcu(ptype, &ptype_all, list) { if (ptype->dev == null_or_orig || ptype->dev == skb->dev || ptype->dev == orig_dev) { if (!ptype->dev || ptype->dev == skb->dev) { if (pt_prev) ret = deliver_skb(skb, pt_prev, orig_dev); pt_prev = ptype; Loading @@ -3224,16 +3171,20 @@ static int __netif_receive_skb(struct sk_buff *skb) ncls: #endif /* Handle special case of bridge or macvlan */ rx_handler = rcu_dereference(skb->dev->rx_handler); if (rx_handler) { struct net_device *prev_dev; if (pt_prev) { ret = deliver_skb(skb, pt_prev, orig_dev); pt_prev = NULL; } prev_dev = skb->dev; skb = rx_handler(skb); if (!skb) goto out; if (skb->dev != prev_dev) goto another_round; } if (vlan_tx_tag_present(skb)) { Loading @@ -3248,24 +3199,16 @@ ncls: goto out; } /* * Make sure frames received on VLAN interfaces stacked on * bonding interfaces still make their way to any base bonding * device that may have registered for a specific ptype. The * handler may have to adjust skb->dev and orig_dev. */ orig_or_bond = orig_dev; if ((skb->dev->priv_flags & IFF_802_1Q_VLAN) && (vlan_dev_real_dev(skb->dev)->priv_flags & IFF_BONDING)) { orig_or_bond = vlan_dev_real_dev(skb->dev); } vlan_on_bond_hook(skb); /* deliver only exact match when indicated */ null_or_dev = skb->deliver_no_wcard ? skb->dev : NULL; type = skb->protocol; list_for_each_entry_rcu(ptype, &ptype_base[ntohs(type) & PTYPE_HASH_MASK], list) { if (ptype->type == type && (ptype->dev == null_or_orig || ptype->dev == skb->dev || ptype->dev == orig_dev || ptype->dev == orig_or_bond)) { if (ptype->type == type && (ptype->dev == null_or_dev || ptype->dev == skb->dev)) { if (pt_prev) ret = deliver_skb(skb, pt_prev, orig_dev); pt_prev = ptype; Loading Loading
net/core/dev.c +31 −88 Original line number Diff line number Diff line Loading @@ -3096,54 +3096,23 @@ void netdev_rx_handler_unregister(struct net_device *dev) } EXPORT_SYMBOL_GPL(netdev_rx_handler_unregister); static inline void skb_bond_set_mac_by_master(struct sk_buff *skb, struct net_device *master) static void vlan_on_bond_hook(struct sk_buff *skb) { if (skb->pkt_type == PACKET_HOST) { u16 *dest = (u16 *) eth_hdr(skb)->h_dest; memcpy(dest, master->dev_addr, ETH_ALEN); } } /* On bonding slaves other than the currently active slave, suppress * duplicates except for 802.3ad ETH_P_SLOW, alb non-mcast/bcast, and * ARP on active-backup slaves with arp_validate enabled. */ static int __skb_bond_should_drop(struct sk_buff *skb, struct net_device *master) { struct net_device *dev = skb->dev; if (master->priv_flags & IFF_MASTER_ARPMON) dev->last_rx = jiffies; if ((master->priv_flags & IFF_MASTER_ALB) && (master->priv_flags & IFF_BRIDGE_PORT)) { /* Do address unmangle. The local destination address * will be always the one master has. Provides the right * functionality in a bridge. /* * Make sure ARP frames received on VLAN interfaces stacked on * bonding interfaces still make their way to any base bonding * device that may have registered for a specific ptype. */ skb_bond_set_mac_by_master(skb, master); } if (skb->dev->priv_flags & IFF_802_1Q_VLAN && vlan_dev_real_dev(skb->dev)->priv_flags & IFF_BONDING && skb->protocol == htons(ETH_P_ARP)) { struct sk_buff *skb2 = skb_clone(skb, GFP_ATOMIC); if (dev->priv_flags & IFF_SLAVE_INACTIVE) { if ((dev->priv_flags & IFF_SLAVE_NEEDARP) && skb->protocol == __cpu_to_be16(ETH_P_ARP)) return 0; if (master->priv_flags & IFF_MASTER_ALB) { if (skb->pkt_type != PACKET_BROADCAST && skb->pkt_type != PACKET_MULTICAST) return 0; } if (master->priv_flags & IFF_MASTER_8023AD && skb->protocol == __cpu_to_be16(ETH_P_SLOW)) return 0; return 1; if (!skb2) return; skb2->dev = vlan_dev_real_dev(skb->dev); netif_rx(skb2); } return 0; } static int __netif_receive_skb(struct sk_buff *skb) Loading @@ -3151,8 +3120,7 @@ static int __netif_receive_skb(struct sk_buff *skb) struct packet_type *ptype, *pt_prev; rx_handler_func_t *rx_handler; struct net_device *orig_dev; struct net_device *null_or_orig; struct net_device *orig_or_bond; struct net_device *null_or_dev; int ret = NET_RX_DROP; __be16 type; Loading @@ -3167,32 +3135,8 @@ static int __netif_receive_skb(struct sk_buff *skb) if (!skb->skb_iif) skb->skb_iif = skb->dev->ifindex; /* * bonding note: skbs received on inactive slaves should only * be delivered to pkt handlers that are exact matches. Also * the deliver_no_wcard flag will be set. If packet handlers * are sensitive to duplicate packets these skbs will need to * be dropped at the handler. */ null_or_orig = NULL; orig_dev = skb->dev; if (skb->deliver_no_wcard) null_or_orig = orig_dev; else if (netif_is_bond_slave(orig_dev)) { struct net_device *bond_master = ACCESS_ONCE(orig_dev->master); if (likely(bond_master)) { if (__skb_bond_should_drop(skb, bond_master)) { skb->deliver_no_wcard = 1; /* deliver only exact match */ null_or_orig = orig_dev; } else skb->dev = bond_master; } } __this_cpu_inc(softnet_data.processed); skb_reset_network_header(skb); skb_reset_transport_header(skb); skb->mac_len = skb->network_header - skb->mac_header; Loading @@ -3201,6 +3145,10 @@ static int __netif_receive_skb(struct sk_buff *skb) rcu_read_lock(); another_round: __this_cpu_inc(softnet_data.processed); #ifdef CONFIG_NET_CLS_ACT if (skb->tc_verd & TC_NCLS) { skb->tc_verd = CLR_TC_NCLS(skb->tc_verd); Loading @@ -3209,8 +3157,7 @@ static int __netif_receive_skb(struct sk_buff *skb) #endif list_for_each_entry_rcu(ptype, &ptype_all, list) { if (ptype->dev == null_or_orig || ptype->dev == skb->dev || ptype->dev == orig_dev) { if (!ptype->dev || ptype->dev == skb->dev) { if (pt_prev) ret = deliver_skb(skb, pt_prev, orig_dev); pt_prev = ptype; Loading @@ -3224,16 +3171,20 @@ static int __netif_receive_skb(struct sk_buff *skb) ncls: #endif /* Handle special case of bridge or macvlan */ rx_handler = rcu_dereference(skb->dev->rx_handler); if (rx_handler) { struct net_device *prev_dev; if (pt_prev) { ret = deliver_skb(skb, pt_prev, orig_dev); pt_prev = NULL; } prev_dev = skb->dev; skb = rx_handler(skb); if (!skb) goto out; if (skb->dev != prev_dev) goto another_round; } if (vlan_tx_tag_present(skb)) { Loading @@ -3248,24 +3199,16 @@ ncls: goto out; } /* * Make sure frames received on VLAN interfaces stacked on * bonding interfaces still make their way to any base bonding * device that may have registered for a specific ptype. The * handler may have to adjust skb->dev and orig_dev. */ orig_or_bond = orig_dev; if ((skb->dev->priv_flags & IFF_802_1Q_VLAN) && (vlan_dev_real_dev(skb->dev)->priv_flags & IFF_BONDING)) { orig_or_bond = vlan_dev_real_dev(skb->dev); } vlan_on_bond_hook(skb); /* deliver only exact match when indicated */ null_or_dev = skb->deliver_no_wcard ? skb->dev : NULL; type = skb->protocol; list_for_each_entry_rcu(ptype, &ptype_base[ntohs(type) & PTYPE_HASH_MASK], list) { if (ptype->type == type && (ptype->dev == null_or_orig || ptype->dev == skb->dev || ptype->dev == orig_dev || ptype->dev == orig_or_bond)) { if (ptype->type == type && (ptype->dev == null_or_dev || ptype->dev == skb->dev)) { if (pt_prev) ret = deliver_skb(skb, pt_prev, orig_dev); pt_prev = ptype; Loading