2008-07-08 14:23:36 +04:00
# include <linux/skbuff.h>
# include <linux/netdevice.h>
# include <linux/if_vlan.h>
2009-03-01 11:11:52 +03:00
# include <linux/netpoll.h>
2008-07-08 14:23:36 +04:00
# include "vlan.h"
2010-10-20 17:56:06 +04:00
bool vlan_hwaccel_do_receive ( struct sk_buff * * skbp )
2008-07-08 14:23:36 +04:00
{
2010-10-20 17:56:06 +04:00
struct sk_buff * skb = * skbp ;
u16 vlan_id = skb - > vlan_tci & VLAN_VID_MASK ;
vlan_dev: VLAN 0 should be treated as "no vlan tag" (802.1p packet)
- Without the 8021q module loaded in the kernel, all 802.1p packets
(VLAN 0 but QoS tagging) are silently discarded (as expected, as
the protocol is not loaded).
- Without this patch in 8021q module, these packets are forwarded to
the module, but they are discarded also if VLAN 0 is not configured,
which should not be the default behaviour, as VLAN 0 is not really
a VLANed packet but a 802.1p packet. Defining VLAN 0 makes it almost
impossible to communicate with mixed 802.1p and non 802.1p devices on
the same network due to arp table issues.
- Changed logic to skip vlan specific code in vlan_skb_recv if VLAN
is 0 and we have not defined a VLAN with ID 0, but we accept the
packet with the encapsulated proto and pass it later to netif_rx.
- In the vlan device event handler, added some logic to add VLAN 0
to HW filter in devices that support it (this prevented any traffic
in VLAN 0 to reach the stack in e1000e with HW filter under 2.6.35,
and probably also with other HW filtered cards, so we fix it here).
- In the vlan unregister logic, prevent the elimination of VLAN 0
in devices with HW filter.
- The default behaviour is to ignore the VLAN 0 tagging and accept
the packet as if it was not tagged, but we can still define a
VLAN 0 if desired (so it is backwards compatible).
Signed-off-by: Pedro Garcia <pedro.netdev@dondevamos.com>
Signed-off-by: David S. Miller <davem@davemloft.net>
2010-07-19 02:38:44 +04:00
struct net_device * vlan_dev ;
2010-11-11 02:42:00 +03:00
struct vlan_pcpu_stats * rx_stats ;
2008-07-08 14:23:36 +04:00
2010-10-20 17:56:06 +04:00
vlan_dev = vlan_find_dev ( skb - > dev , vlan_id ) ;
if ( ! vlan_dev ) {
if ( vlan_id )
skb - > pkt_type = PACKET_OTHERHOST ;
return false ;
2010-09-30 06:16:44 +04:00
}
2008-11-05 01:49:57 +03:00
2010-10-20 17:56:06 +04:00
skb = * skbp = skb_share_check ( skb , GFP_ATOMIC ) ;
if ( unlikely ( ! skb ) )
return false ;
2009-01-06 21:50:09 +03:00
2010-10-20 17:56:06 +04:00
skb - > dev = vlan_dev ;
skb - > priority = vlan_get_ingress_priority ( vlan_dev , skb - > vlan_tci ) ;
2008-07-15 09:49:30 +04:00
skb - > vlan_tci = 0 ;
2008-07-08 14:23:36 +04:00
2010-11-11 02:42:00 +03:00
rx_stats = this_cpu_ptr ( vlan_dev_info ( vlan_dev ) - > vlan_pcpu_stats ) ;
2009-11-17 07:53:09 +03:00
2010-06-24 04:55:06 +04:00
u64_stats_update_begin ( & rx_stats - > syncp ) ;
2009-11-17 07:53:09 +03:00
rx_stats - > rx_packets + + ;
rx_stats - > rx_bytes + = skb - > len ;
2008-07-08 14:23:36 +04:00
switch ( skb - > pkt_type ) {
case PACKET_BROADCAST :
break ;
case PACKET_MULTICAST :
2010-06-24 04:55:06 +04:00
rx_stats - > rx_multicast + + ;
2008-07-08 14:23:36 +04:00
break ;
case PACKET_OTHERHOST :
/* Our lower layer thinks this is not local, let's make sure.
* This allows the VLAN to have a different MAC than the
* underlying device , and still route correctly . */
if ( ! compare_ether_addr ( eth_hdr ( skb ) - > h_dest ,
2010-10-20 17:56:06 +04:00
vlan_dev - > dev_addr ) )
2008-07-08 14:23:36 +04:00
skb - > pkt_type = PACKET_HOST ;
break ;
2010-05-14 14:58:26 +04:00
}
2010-06-24 04:55:06 +04:00
u64_stats_update_end ( & rx_stats - > syncp ) ;
2010-10-20 17:56:06 +04:00
return true ;
2008-07-08 14:23:36 +04:00
}
2008-07-08 14:23:57 +04:00
struct net_device * vlan_dev_real_dev ( const struct net_device * dev )
{
return vlan_dev_info ( dev ) - > real_dev ;
}
2009-01-26 23:37:53 +03:00
EXPORT_SYMBOL ( vlan_dev_real_dev ) ;
2008-07-08 14:23:57 +04:00
u16 vlan_dev_vlan_id ( const struct net_device * dev )
{
return vlan_dev_info ( dev ) - > vlan_id ;
}
2009-01-26 23:37:53 +03:00
EXPORT_SYMBOL ( vlan_dev_vlan_id ) ;
2009-01-06 21:50:09 +03:00
2010-10-20 17:56:06 +04:00
/* VLAN rx hw acceleration helper. This acts like netif_{rx,receive_skb}(). */
int __vlan_hwaccel_rx ( struct sk_buff * skb , struct vlan_group * grp ,
u16 vlan_tci , int polling )
2009-01-06 21:50:09 +03:00
{
2009-11-13 09:33:11 +03:00
__vlan_hwaccel_put_tag ( skb , vlan_tci ) ;
2010-10-20 17:56:06 +04:00
return polling ? netif_receive_skb ( skb ) : netif_rx ( skb ) ;
2009-01-06 21:50:09 +03:00
}
2010-10-20 17:56:06 +04:00
EXPORT_SYMBOL ( __vlan_hwaccel_rx ) ;
2009-01-06 21:50:09 +03:00
2009-10-30 07:36:53 +03:00
gro_result_t vlan_gro_receive ( struct napi_struct * napi , struct vlan_group * grp ,
unsigned int vlan_tci , struct sk_buff * skb )
2009-01-06 21:50:09 +03:00
{
2010-10-20 17:56:06 +04:00
__vlan_hwaccel_put_tag ( skb , vlan_tci ) ;
return napi_gro_receive ( napi , skb ) ;
2009-01-06 21:50:09 +03:00
}
EXPORT_SYMBOL ( vlan_gro_receive ) ;
2009-10-30 07:36:53 +03:00
gro_result_t vlan_gro_frags ( struct napi_struct * napi , struct vlan_group * grp ,
unsigned int vlan_tci )
2009-01-06 21:50:09 +03:00
{
2010-10-20 17:56:06 +04:00
__vlan_hwaccel_put_tag ( napi - > skb , vlan_tci ) ;
return napi_gro_frags ( napi ) ;
2009-01-06 21:50:09 +03:00
}
EXPORT_SYMBOL ( vlan_gro_frags ) ;