···16681668 if (napi_schedule_prep(&priv->napi)) {16691669 gfar_write(&priv->regs->imask, IMASK_RTX_DISABLED);16701670 __napi_schedule(&priv->napi);16711671+ } else {16721672+ /*16731673+ * Clear IEVENT, so interrupts aren't called again16741674+ * because of the packets that have already arrived.16751675+ */16761676+ gfar_write(&priv->regs->ievent, IEVENT_RTX_MASK);16711677 }1672167816731679 spin_unlock(&priv->rxlock);
···260260261261 nexact = n;262262263263- /* The rest is hashed */263263+ /* Remaining multicast addresses are hashed,264264+ * unicast will leave the filter disabled. */264265 memset(filter->mask, 0, sizeof(filter->mask));265265- for (; n < uf.count; n++)266266+ for (; n < uf.count; n++) {267267+ if (!is_multicast_ether_addr(addr[n].u)) {268268+ err = 0; /* no filter */269269+ goto done;270270+ }266271 addr_hash_set(filter->mask, addr[n].u);272272+ }267273268274 /* For ALLMULTI just set the mask to all ones.269275 * This overrides the mask populated above. */
+6-1
net/bridge/br_forward.c
···6767{6868 struct net_device *indev;69697070+ if (skb_warn_if_lro(skb)) {7171+ kfree_skb(skb);7272+ return;7373+ }7474+7075 indev = skb->dev;7176 skb->dev = to->dev;7277 skb_forward_csum(skb);···9489/* called with rcu_read_lock */9590void br_forward(const struct net_bridge_port *to, struct sk_buff *skb)9691{9797- if (!skb_warn_if_lro(skb) && should_deliver(to, skb)) {9292+ if (should_deliver(to, skb)) {9893 __br_forward(to, skb);9994 return;10095 }
+1-1
net/ipv6/ip6_tunnel.c
···249249 }250250251251 t = netdev_priv(dev);252252- ip6_tnl_dev_init(dev);253252 t->parms = *p;253253+ ip6_tnl_dev_init(dev);254254255255 if ((err = register_netdevice(dev)) < 0)256256 goto failed_free;