2019-06-04 11:11:33 +03:00
// SPDX-License-Identifier: GPL-2.0-only
2008-02-23 17:17:15 +03:00
/*
2009-11-10 02:46:58 +03:00
* Copyright ( c ) 2008 , 2009 open80211s Ltd .
2023-06-06 15:49:26 +03:00
* Copyright ( C ) 2019 , 2021 - 2023 Intel Corporation
2008-02-23 17:17:15 +03:00
* Author : Luis Carlos Cobo < luisca @ cozybit . com >
*/
include cleanup: Update gfp.h and slab.h includes to prepare for breaking implicit slab.h inclusion from percpu.h
percpu.h is included by sched.h and module.h and thus ends up being
included when building most .c files. percpu.h includes slab.h which
in turn includes gfp.h making everything defined by the two files
universally available and complicating inclusion dependencies.
percpu.h -> slab.h dependency is about to be removed. Prepare for
this change by updating users of gfp and slab facilities include those
headers directly instead of assuming availability. As this conversion
needs to touch large number of source files, the following script is
used as the basis of conversion.
http://userweb.kernel.org/~tj/misc/slabh-sweep.py
The script does the followings.
* Scan files for gfp and slab usages and update includes such that
only the necessary includes are there. ie. if only gfp is used,
gfp.h, if slab is used, slab.h.
* When the script inserts a new include, it looks at the include
blocks and try to put the new include such that its order conforms
to its surrounding. It's put in the include block which contains
core kernel includes, in the same order that the rest are ordered -
alphabetical, Christmas tree, rev-Xmas-tree or at the end if there
doesn't seem to be any matching order.
* If the script can't find a place to put a new include (mostly
because the file doesn't have fitting include block), it prints out
an error message indicating which .h file needs to be added to the
file.
The conversion was done in the following steps.
1. The initial automatic conversion of all .c files updated slightly
over 4000 files, deleting around 700 includes and adding ~480 gfp.h
and ~3000 slab.h inclusions. The script emitted errors for ~400
files.
2. Each error was manually checked. Some didn't need the inclusion,
some needed manual addition while adding it to implementation .h or
embedding .c file was more appropriate for others. This step added
inclusions to around 150 files.
3. The script was run again and the output was compared to the edits
from #2 to make sure no file was left behind.
4. Several build tests were done and a couple of problems were fixed.
e.g. lib/decompress_*.c used malloc/free() wrappers around slab
APIs requiring slab.h to be added manually.
5. The script was run on all .h files but without automatically
editing them as sprinkling gfp.h and slab.h inclusions around .h
files could easily lead to inclusion dependency hell. Most gfp.h
inclusion directives were ignored as stuff from gfp.h was usually
wildly available and often used in preprocessor macros. Each
slab.h inclusion directive was examined and added manually as
necessary.
6. percpu.h was updated not to include slab.h.
7. Build test were done on the following configurations and failures
were fixed. CONFIG_GCOV_KERNEL was turned off for all tests (as my
distributed build env didn't work with gcov compiles) and a few
more options had to be turned off depending on archs to make things
build (like ipr on powerpc/64 which failed due to missing writeq).
* x86 and x86_64 UP and SMP allmodconfig and a custom test config.
* powerpc and powerpc64 SMP allmodconfig
* sparc and sparc64 SMP allmodconfig
* ia64 SMP allmodconfig
* s390 SMP allmodconfig
* alpha SMP allmodconfig
* um on x86_64 SMP allmodconfig
8. percpu.h modifications were reverted so that it could be applied as
a separate patch and serve as bisection point.
Given the fact that I had only a couple of failures from tests on step
6, I'm fairly confident about the coverage of this conversion patch.
If there is a breakage, it's likely to be something in one of the arch
headers which should be easily discoverable easily on most builds of
the specific arch.
Signed-off-by: Tejun Heo <tj@kernel.org>
Guess-its-ok-by: Christoph Lameter <cl@linux-foundation.org>
Cc: Ingo Molnar <mingo@redhat.com>
Cc: Lee Schermerhorn <Lee.Schermerhorn@hp.com>
2010-03-24 11:04:11 +03:00
# include <linux/slab.h>
2012-03-01 18:22:09 +04:00
# include <linux/etherdevice.h>
2012-02-20 14:38:41 +04:00
# include <asm/unaligned.h>
2011-09-06 23:10:43 +04:00
# include "wme.h"
2008-02-23 17:17:15 +03:00
# include "mesh.h"
# define TEST_FRAME_LEN 8192
# define MAX_METRIC 0xffffffff
# define ARITH_SHIFT 8
2017-02-15 23:46:50 +03:00
# define LINK_FAIL_THRESH 95
2008-02-23 17:17:15 +03:00
# define MAX_PREQ_QUEUE_LEN 64
2009-11-11 03:01:31 +03:00
static void mesh_queue_preq ( struct mesh_path * , u8 ) ;
2013-02-12 19:43:19 +04:00
static inline u32 u32_field_get ( const u8 * preq_elem , int offset , bool ae )
2008-03-01 04:07:54 +03:00
{
if ( ae )
offset + = 6 ;
2008-05-02 09:19:33 +04:00
return get_unaligned_le32 ( preq_elem + offset ) ;
2008-03-01 04:07:54 +03:00
}
2014-04-15 18:43:06 +04:00
static inline u16 u16_field_get ( const u8 * preq_elem , int offset , bool ae )
2009-11-10 02:46:50 +03:00
{
if ( ae )
offset + = 6 ;
return get_unaligned_le16 ( preq_elem + offset ) ;
}
2008-02-23 17:17:15 +03:00
/* HWMP IE processing macros */
2008-03-01 04:07:54 +03:00
# define AE_F (1<<6)
# define AE_F_SET(x) (*x & AE_F)
# define PREQ_IE_FLAGS(x) (*(x))
# define PREQ_IE_HOPCOUNT(x) (*(x + 1))
# define PREQ_IE_TTL(x) (*(x + 2))
# define PREQ_IE_PREQ_ID(x) u32_field_get(x, 3, 0)
# define PREQ_IE_ORIG_ADDR(x) (x + 7)
2011-07-14 16:07:13 +04:00
# define PREQ_IE_ORIG_SN(x) u32_field_get(x, 13, 0)
# define PREQ_IE_LIFETIME(x) u32_field_get(x, 17, AE_F_SET(x))
# define PREQ_IE_METRIC(x) u32_field_get(x, 21, AE_F_SET(x))
2009-11-10 02:46:55 +03:00
# define PREQ_IE_TARGET_F(x) (*(AE_F_SET(x) ? x + 32 : x + 26))
# define PREQ_IE_TARGET_ADDR(x) (AE_F_SET(x) ? x + 33 : x + 27)
2011-07-14 16:07:13 +04:00
# define PREQ_IE_TARGET_SN(x) u32_field_get(x, 33, AE_F_SET(x))
2008-03-01 04:07:54 +03:00
# define PREP_IE_FLAGS(x) PREQ_IE_FLAGS(x)
# define PREP_IE_HOPCOUNT(x) PREQ_IE_HOPCOUNT(x)
# define PREP_IE_TTL(x) PREQ_IE_TTL(x)
2011-08-12 06:35:15 +04:00
# define PREP_IE_ORIG_ADDR(x) (AE_F_SET(x) ? x + 27 : x + 21)
# define PREP_IE_ORIG_SN(x) u32_field_get(x, 27, AE_F_SET(x))
2011-07-14 16:07:13 +04:00
# define PREP_IE_LIFETIME(x) u32_field_get(x, 13, AE_F_SET(x))
# define PREP_IE_METRIC(x) u32_field_get(x, 17, AE_F_SET(x))
2011-08-12 06:35:15 +04:00
# define PREP_IE_TARGET_ADDR(x) (x + 3)
# define PREP_IE_TARGET_SN(x) u32_field_get(x, 9, 0)
2008-03-01 04:07:54 +03:00
2009-11-10 02:46:50 +03:00
# define PERR_IE_TTL(x) (*(x))
2009-11-10 02:46:55 +03:00
# define PERR_IE_TARGET_FLAGS(x) (*(x + 2))
# define PERR_IE_TARGET_ADDR(x) (x + 3)
2011-07-14 16:07:13 +04:00
# define PERR_IE_TARGET_SN(x) u32_field_get(x, 9, 0)
# define PERR_IE_TARGET_RCODE(x) u16_field_get(x, 13, 0)
2008-02-23 17:17:15 +03:00
# define MSEC_TO_TU(x) (x*1000 / 1024)
2012-03-23 14:48:51 +04:00
# define SN_GT(x, y) ((s32)(y - x) < 0)
# define SN_LT(x, y) ((s32)(x - y) < 0)
2015-06-13 01:38:07 +03:00
# define MAX_SANE_SN_DELTA 32
static inline u32 SN_DELTA ( u32 x , u32 y )
{
return x > = y ? x - y : y - x ;
}
2008-02-23 17:17:15 +03:00
# define net_traversal_jiffies(s) \
2008-09-11 02:01:49 +04:00
msecs_to_jiffies ( s - > u . mesh . mshcfg . dot11MeshHWMPnetDiameterTraversalTime )
2008-02-23 17:17:15 +03:00
# define default_lifetime(s) \
2008-09-11 02:01:49 +04:00
MSEC_TO_TU ( s - > u . mesh . mshcfg . dot11MeshHWMPactivePathTimeout )
2008-02-23 17:17:15 +03:00
# define min_preq_int_jiff(s) \
2008-09-11 02:01:49 +04:00
( msecs_to_jiffies ( s - > u . mesh . mshcfg . dot11MeshHWMPpreqMinInterval ) )
# define max_preq_retries(s) (s->u.mesh.mshcfg.dot11MeshHWMPmaxPREQretries)
2008-02-23 17:17:15 +03:00
# define disc_timeout_jiff(s) \
2008-09-11 02:01:49 +04:00
msecs_to_jiffies ( sdata - > u . mesh . mshcfg . min_discovery_timeout )
2012-06-13 22:06:10 +04:00
# define root_path_confirmation_jiffies(s) \
msecs_to_jiffies ( sdata - > u . mesh . mshcfg . dot11MeshHWMPconfirmationInterval )
2008-02-23 17:17:15 +03:00
enum mpath_frame_type {
MPATH_PREQ = 0 ,
MPATH_PREP ,
2009-11-11 03:01:31 +03:00
MPATH_PERR ,
MPATH_RANN
2008-02-23 17:17:15 +03:00
} ;
2009-11-17 15:34:04 +03:00
static const u8 broadcast_addr [ ETH_ALEN ] = { 0xff , 0xff , 0xff , 0xff , 0xff , 0xff } ;
2008-02-23 17:17:15 +03:00
static int mesh_path_sel_frame_tx ( enum mpath_frame_type action , u8 flags ,
2013-11-13 11:39:12 +04:00
const u8 * orig_addr , u32 orig_sn ,
2013-02-12 19:43:19 +04:00
u8 target_flags , const u8 * target ,
2013-11-13 11:39:12 +04:00
u32 target_sn , const u8 * da ,
2013-02-12 19:43:19 +04:00
u8 hop_count , u8 ttl ,
2013-11-13 11:39:12 +04:00
u32 lifetime , u32 metric , u32 preq_id ,
2013-02-12 19:43:19 +04:00
struct ieee80211_sub_if_data * sdata )
2008-02-23 17:17:15 +03:00
{
2008-08-03 04:04:37 +04:00
struct ieee80211_local * local = sdata - > local ;
2011-10-27 01:47:25 +04:00
struct sk_buff * skb ;
2008-02-23 17:17:15 +03:00
struct ieee80211_mgmt * mgmt ;
2011-10-27 01:47:25 +04:00
u8 * pos , ie_len ;
2017-09-08 12:54:46 +03:00
int hdr_len = offsetofend ( struct ieee80211_mgmt ,
u . action . u . mesh_action ) ;
2008-02-23 17:17:15 +03:00
2012-01-18 06:17:46 +04:00
skb = dev_alloc_skb ( local - > tx_headroom +
2011-10-27 01:47:25 +04:00
hdr_len +
2 + 37 ) ; /* max HWMP IE */
2008-02-23 17:17:15 +03:00
if ( ! skb )
return - 1 ;
2012-01-18 06:17:46 +04:00
skb_reserve ( skb , local - > tx_headroom ) ;
networking: convert many more places to skb_put_zero()
There were many places that my previous spatch didn't find,
as pointed out by yuan linyu in various patches.
The following spatch found many more and also removes the
now unnecessary casts:
@@
identifier p, p2;
expression len;
expression skb;
type t, t2;
@@
(
-p = skb_put(skb, len);
+p = skb_put_zero(skb, len);
|
-p = (t)skb_put(skb, len);
+p = skb_put_zero(skb, len);
)
... when != p
(
p2 = (t2)p;
-memset(p2, 0, len);
|
-memset(p, 0, len);
)
@@
type t, t2;
identifier p, p2;
expression skb;
@@
t *p;
...
(
-p = skb_put(skb, sizeof(t));
+p = skb_put_zero(skb, sizeof(t));
|
-p = (t *)skb_put(skb, sizeof(t));
+p = skb_put_zero(skb, sizeof(t));
)
... when != p
(
p2 = (t2)p;
-memset(p2, 0, sizeof(*p));
|
-memset(p, 0, sizeof(*p));
)
@@
expression skb, len;
@@
-memset(skb_put(skb, len), 0, len);
+skb_put_zero(skb, len);
Apply it to the tree (with one manual fixup to keep the
comment in vxlan.c, which spatch removed.)
Signed-off-by: Johannes Berg <johannes.berg@intel.com>
Signed-off-by: David S. Miller <davem@davemloft.net>
2017-06-16 15:29:19 +03:00
mgmt = skb_put_zero ( skb , hdr_len ) ;
2008-07-16 05:44:13 +04:00
mgmt - > frame_control = cpu_to_le16 ( IEEE80211_FTYPE_MGMT |
IEEE80211_STYPE_ACTION ) ;
2008-02-23 17:17:15 +03:00
memcpy ( mgmt - > da , da , ETH_ALEN ) ;
2009-11-25 19:46:19 +03:00
memcpy ( mgmt - > sa , sdata - > vif . addr , ETH_ALEN ) ;
2009-11-11 03:01:31 +03:00
/* BSSID == SA */
2009-11-25 19:46:19 +03:00
memcpy ( mgmt - > bssid , sdata - > vif . addr , ETH_ALEN ) ;
2011-08-12 06:35:15 +04:00
mgmt - > u . action . category = WLAN_CATEGORY_MESH_ACTION ;
mgmt - > u . action . u . mesh_action . action_code =
WLAN_MESH_ACTION_HWMP_PATH_SELECTION ;
2008-02-23 17:17:15 +03:00
switch ( action ) {
case MPATH_PREQ :
2012-06-22 13:29:50 +04:00
mhwmp_dbg ( sdata , " sending PREQ to %pM \n " , target ) ;
2008-02-23 17:17:15 +03:00
ie_len = 37 ;
pos = skb_put ( skb , 2 + ie_len ) ;
* pos + + = WLAN_EID_PREQ ;
break ;
case MPATH_PREP :
2013-04-03 13:49:53 +04:00
mhwmp_dbg ( sdata , " sending PREP to %pM \n " , orig_addr ) ;
2008-02-23 17:17:15 +03:00
ie_len = 31 ;
pos = skb_put ( skb , 2 + ie_len ) ;
* pos + + = WLAN_EID_PREP ;
break ;
2009-11-11 03:01:31 +03:00
case MPATH_RANN :
2012-06-22 13:29:50 +04:00
mhwmp_dbg ( sdata , " sending RANN from %pM \n " , orig_addr ) ;
2009-11-11 03:01:31 +03:00
ie_len = sizeof ( struct ieee80211_rann_ie ) ;
pos = skb_put ( skb , 2 + ie_len ) ;
* pos + + = WLAN_EID_RANN ;
break ;
2008-02-23 17:17:15 +03:00
default :
2008-05-06 14:52:07 +04:00
kfree_skb ( skb ) ;
2023-12-11 10:05:25 +03:00
return - EOPNOTSUPP ;
2008-02-23 17:17:15 +03:00
}
* pos + + = ie_len ;
* pos + + = flags ;
* pos + + = hop_count ;
* pos + + = ttl ;
2011-08-12 06:35:15 +04:00
if ( action = = MPATH_PREP ) {
memcpy ( pos , target , ETH_ALEN ) ;
pos + = ETH_ALEN ;
2013-11-13 11:39:12 +04:00
put_unaligned_le32 ( target_sn , pos ) ;
2008-02-23 17:17:15 +03:00
pos + = 4 ;
2011-08-12 06:35:15 +04:00
} else {
if ( action = = MPATH_PREQ ) {
2013-11-13 11:39:12 +04:00
put_unaligned_le32 ( preq_id , pos ) ;
2011-08-12 06:35:15 +04:00
pos + = 4 ;
}
memcpy ( pos , orig_addr , ETH_ALEN ) ;
pos + = ETH_ALEN ;
2013-11-13 11:39:12 +04:00
put_unaligned_le32 ( orig_sn , pos ) ;
2009-11-11 03:01:31 +03:00
pos + = 4 ;
}
2013-11-13 11:39:12 +04:00
put_unaligned_le32 ( lifetime , pos ) ; /* interval for RANN */
2011-08-12 06:35:15 +04:00
pos + = 4 ;
2013-11-13 11:39:12 +04:00
put_unaligned_le32 ( metric , pos ) ;
2008-02-23 17:17:15 +03:00
pos + = 4 ;
if ( action = = MPATH_PREQ ) {
2011-08-12 06:35:15 +04:00
* pos + + = 1 ; /* destination count */
2009-11-10 02:46:55 +03:00
* pos + + = target_flags ;
memcpy ( pos , target , ETH_ALEN ) ;
2009-11-11 03:01:31 +03:00
pos + = ETH_ALEN ;
2013-11-13 11:39:12 +04:00
put_unaligned_le32 ( target_sn , pos ) ;
2011-08-12 06:35:15 +04:00
pos + = 4 ;
} else if ( action = = MPATH_PREP ) {
memcpy ( pos , orig_addr , ETH_ALEN ) ;
pos + = ETH_ALEN ;
2013-11-13 11:39:12 +04:00
put_unaligned_le32 ( orig_sn , pos ) ;
2011-08-12 06:35:15 +04:00
pos + = 4 ;
2009-11-11 03:01:31 +03:00
}
2008-02-23 17:17:15 +03:00
2009-11-18 20:42:05 +03:00
ieee80211_tx_skb ( sdata , skb ) ;
2008-02-23 17:17:15 +03:00
return 0 ;
}
2011-09-06 23:10:43 +04:00
/* Headroom is not adjusted. Caller should ensure that skb has sufficient
* headroom in case the frame is encrypted . */
static void prepare_frame_for_deferred_tx ( struct ieee80211_sub_if_data * sdata ,
struct sk_buff * skb )
{
struct ieee80211_tx_info * info = IEEE80211_SKB_CB ( skb ) ;
2013-01-30 21:14:08 +04:00
struct ieee80211_hdr * hdr = ( struct ieee80211_hdr * ) skb - > data ;
2011-09-06 23:10:43 +04:00
2016-03-03 04:16:56 +03:00
skb_reset_mac_header ( skb ) ;
skb_reset_network_header ( skb ) ;
skb_reset_transport_header ( skb ) ;
2011-09-06 23:10:43 +04:00
/* Send all internal mgmt frames on VO. Accordingly set TID to 7. */
skb_set_queue_mapping ( skb , IEEE80211_AC_VO ) ;
skb - > priority = 7 ;
info - > control . vif = & sdata - > vif ;
2020-09-08 15:36:57 +03:00
info - > control . flags | = IEEE80211_TX_INTCFL_NEED_TXPROCESSING ;
2011-09-08 04:49:53 +04:00
ieee80211_set_qos_hdr ( sdata , skb ) ;
2013-01-30 21:14:08 +04:00
ieee80211_mps_set_frame_flags ( sdata , NULL , hdr ) ;
2011-09-06 23:10:43 +04:00
}
2008-02-23 17:17:15 +03:00
/**
2013-01-03 09:09:46 +04:00
* mesh_path_error_tx - Sends a PERR mesh management frame
2008-02-23 17:17:15 +03:00
*
2013-01-03 09:09:46 +04:00
* @ ttl : allowed remaining hops
2009-11-10 02:46:55 +03:00
* @ target : broken destination
* @ target_sn : SN of the broken destination
* @ target_rcode : reason code for this PERR
2008-02-23 17:17:15 +03:00
* @ ra : node this frame is addressed to
2013-01-03 09:09:46 +04:00
* @ sdata : local mesh subif
2011-09-06 23:10:43 +04:00
*
* Note : This function may be called with driver locks taken that the driver
* also acquires in the TX path . To avoid a deadlock we don ' t transmit the
* frame directly but add it to the pending queue instead .
2023-09-28 17:35:29 +03:00
*
* Returns : 0 on success
2008-02-23 17:17:15 +03:00
*/
2013-02-15 17:40:31 +04:00
int mesh_path_error_tx ( struct ieee80211_sub_if_data * sdata ,
2013-11-13 11:39:12 +04:00
u8 ttl , const u8 * target , u32 target_sn ,
u16 target_rcode , const u8 * ra )
2008-02-23 17:17:15 +03:00
{
2008-08-03 04:04:37 +04:00
struct ieee80211_local * local = sdata - > local ;
2011-10-27 01:47:25 +04:00
struct sk_buff * skb ;
2011-11-25 05:15:24 +04:00
struct ieee80211_if_mesh * ifmsh = & sdata - > u . mesh ;
2008-02-23 17:17:15 +03:00
struct ieee80211_mgmt * mgmt ;
2011-10-27 01:47:25 +04:00
u8 * pos , ie_len ;
2017-09-08 12:54:46 +03:00
int hdr_len = offsetofend ( struct ieee80211_mgmt ,
u . action . u . mesh_action ) ;
2008-02-23 17:17:15 +03:00
2011-11-25 05:15:24 +04:00
if ( time_before ( jiffies , ifmsh - > next_perr ) )
return - EAGAIN ;
2012-01-18 06:17:46 +04:00
skb = dev_alloc_skb ( local - > tx_headroom +
2022-02-09 15:14:26 +03:00
IEEE80211_ENCRYPT_HEADROOM +
2013-01-09 21:34:56 +04:00
IEEE80211_ENCRYPT_TAILROOM +
2011-10-27 01:47:25 +04:00
hdr_len +
2 + 15 /* PERR IE */ ) ;
2008-02-23 17:17:15 +03:00
if ( ! skb )
return - 1 ;
2022-02-09 15:14:26 +03:00
skb_reserve ( skb , local - > tx_headroom + IEEE80211_ENCRYPT_HEADROOM ) ;
networking: convert many more places to skb_put_zero()
There were many places that my previous spatch didn't find,
as pointed out by yuan linyu in various patches.
The following spatch found many more and also removes the
now unnecessary casts:
@@
identifier p, p2;
expression len;
expression skb;
type t, t2;
@@
(
-p = skb_put(skb, len);
+p = skb_put_zero(skb, len);
|
-p = (t)skb_put(skb, len);
+p = skb_put_zero(skb, len);
)
... when != p
(
p2 = (t2)p;
-memset(p2, 0, len);
|
-memset(p, 0, len);
)
@@
type t, t2;
identifier p, p2;
expression skb;
@@
t *p;
...
(
-p = skb_put(skb, sizeof(t));
+p = skb_put_zero(skb, sizeof(t));
|
-p = (t *)skb_put(skb, sizeof(t));
+p = skb_put_zero(skb, sizeof(t));
)
... when != p
(
p2 = (t2)p;
-memset(p2, 0, sizeof(*p));
|
-memset(p, 0, sizeof(*p));
)
@@
expression skb, len;
@@
-memset(skb_put(skb, len), 0, len);
+skb_put_zero(skb, len);
Apply it to the tree (with one manual fixup to keep the
comment in vxlan.c, which spatch removed.)
Signed-off-by: Johannes Berg <johannes.berg@intel.com>
Signed-off-by: David S. Miller <davem@davemloft.net>
2017-06-16 15:29:19 +03:00
mgmt = skb_put_zero ( skb , hdr_len ) ;
2008-07-16 05:44:13 +04:00
mgmt - > frame_control = cpu_to_le16 ( IEEE80211_FTYPE_MGMT |
IEEE80211_STYPE_ACTION ) ;
2008-02-23 17:17:15 +03:00
memcpy ( mgmt - > da , ra , ETH_ALEN ) ;
2009-11-25 19:46:19 +03:00
memcpy ( mgmt - > sa , sdata - > vif . addr , ETH_ALEN ) ;
2011-08-12 06:35:15 +04:00
/* BSSID == SA */
memcpy ( mgmt - > bssid , sdata - > vif . addr , ETH_ALEN ) ;
mgmt - > u . action . category = WLAN_CATEGORY_MESH_ACTION ;
mgmt - > u . action . u . mesh_action . action_code =
WLAN_MESH_ACTION_HWMP_PATH_SELECTION ;
2009-11-10 02:46:50 +03:00
ie_len = 15 ;
2008-02-23 17:17:15 +03:00
pos = skb_put ( skb , 2 + ie_len ) ;
* pos + + = WLAN_EID_PERR ;
* pos + + = ie_len ;
2009-11-10 02:46:50 +03:00
/* ttl */
2010-12-03 11:20:40 +03:00
* pos + + = ttl ;
2008-02-23 17:17:15 +03:00
/* number of destinations */
* pos + + = 1 ;
2015-06-15 06:58:53 +03:00
/* Flags field has AE bit only as defined in
* sec 8.4 .2 .117 IEEE802 .11 - 2012
2009-11-10 02:46:50 +03:00
*/
* pos = 0 ;
pos + + ;
2009-11-10 02:46:55 +03:00
memcpy ( pos , target , ETH_ALEN ) ;
2008-02-23 17:17:15 +03:00
pos + = ETH_ALEN ;
2013-11-13 11:39:12 +04:00
put_unaligned_le32 ( target_sn , pos ) ;
2009-11-10 02:46:50 +03:00
pos + = 4 ;
2013-11-13 11:39:12 +04:00
put_unaligned_le16 ( target_rcode , pos ) ;
2008-02-23 17:17:15 +03:00
2011-09-06 23:10:43 +04:00
/* see note in function header */
prepare_frame_for_deferred_tx ( sdata , skb ) ;
2011-11-25 05:15:24 +04:00
ifmsh - > next_perr = TU_TO_EXP_TIME (
ifmsh - > mshcfg . dot11MeshHWMPperrMinInterval ) ;
2011-09-06 23:10:43 +04:00
ieee80211_add_pending_skb ( local , skb ) ;
2008-02-23 17:17:15 +03:00
return 0 ;
}
2009-08-18 04:15:55 +04:00
void ieee80211s_update_metric ( struct ieee80211_local * local ,
2018-09-06 11:57:48 +03:00
struct sta_info * sta ,
struct ieee80211_tx_status * st )
2009-08-18 04:15:55 +04:00
{
2018-09-06 11:57:48 +03:00
struct ieee80211_tx_info * txinfo = st - > info ;
2009-08-18 04:15:55 +04:00
int failed ;
2019-01-16 02:31:56 +03:00
struct rate_info rinfo ;
2009-08-18 04:15:55 +04:00
failed = ! ( txinfo - > flags & IEEE80211_TX_STAT_ACK ) ;
2017-02-14 23:27:16 +03:00
/* moving average, scaled to 100.
* feed failure as 100 and success as 0
*/
ewma_mesh_fail_avg_add ( & sta - > mesh - > fail_avg , failed * 100 ) ;
2017-02-15 23:46:50 +03:00
if ( ewma_mesh_fail_avg_read ( & sta - > mesh - > fail_avg ) >
LINK_FAIL_THRESH )
2012-06-09 00:30:25 +04:00
mesh_plink_broken ( sta ) ;
2019-01-16 02:31:56 +03:00
2022-07-01 16:36:11 +03:00
/* use rate info set by the driver directly if present */
if ( st - > n_rates )
rinfo = sta - > deflink . tx_stats . last_rate_info ;
else
sta_set_rate_info_tx ( sta , & sta - > deflink . tx_stats . last_rate , & rinfo ) ;
2019-01-16 02:31:56 +03:00
ewma_mesh_tx_rate_avg_add ( & sta - > mesh - > tx_rate_avg ,
cfg80211_calculate_bitrate ( & rinfo ) ) ;
2009-08-18 04:15:55 +04:00
}
2019-02-07 23:16:05 +03:00
u32 airtime_link_metric_get ( struct ieee80211_local * local ,
struct sta_info * sta )
2008-02-23 17:17:15 +03:00
{
/* This should be adjusted for each device */
int device_constant = 1 < < ARITH_SHIFT ;
int test_frame_len = TEST_FRAME_LEN < < ARITH_SHIFT ;
int s_unit = 1 < < ARITH_SHIFT ;
int rate , err ;
u32 tx_time , estimated_retx ;
u64 result ;
2017-02-14 23:27:16 +03:00
unsigned long fail_avg =
ewma_mesh_fail_avg_read ( & sta - > mesh - > fail_avg ) ;
2008-02-23 17:17:15 +03:00
2019-12-03 21:06:44 +03:00
if ( sta - > mesh - > plink_state ! = NL80211_PLINK_ESTAB )
return MAX_METRIC ;
2016-07-11 17:15:24 +03:00
/* Try to get rate based on HW/SW RC algorithm.
* Rate is returned in units of Kbps , correct this
* to comply with airtime calculation units
* Round up in case we get rate < 100 Kbps
*/
rate = DIV_ROUND_UP ( sta_get_expected_throughput ( sta ) , 100 ) ;
if ( rate ) {
err = 0 ;
} else {
2017-02-15 23:46:50 +03:00
if ( fail_avg > LINK_FAIL_THRESH )
2016-07-11 17:15:24 +03:00
return MAX_METRIC ;
2008-10-21 14:40:02 +04:00
2019-01-16 02:31:56 +03:00
rate = ewma_mesh_tx_rate_avg_read ( & sta - > mesh - > tx_rate_avg ) ;
2016-07-11 17:15:24 +03:00
if ( WARN_ON ( ! rate ) )
return MAX_METRIC ;
2008-10-21 14:40:02 +04:00
2017-02-14 23:27:16 +03:00
err = ( fail_avg < < ARITH_SHIFT ) / 100 ;
2016-07-11 17:15:24 +03:00
}
2008-02-23 17:17:15 +03:00
/* bitrate is in units of 100 Kbps, while we need rate in units of
* 1 Mbps . This will be corrected on tx_time computation .
*/
tx_time = ( device_constant + 10 * test_frame_len / rate ) ;
estimated_retx = ( ( 1 < < ( 2 * ARITH_SHIFT ) ) / ( s_unit - err ) ) ;
2021-02-05 20:53:52 +03:00
result = ( ( u64 ) tx_time * estimated_retx ) > > ( 2 * ARITH_SHIFT ) ;
2008-02-23 17:17:15 +03:00
return ( u32 ) result ;
}
/**
* hwmp_route_info_get - Update routing info to originator and transmitter
*
2008-08-03 04:04:37 +04:00
* @ sdata : local mesh subif
2008-02-23 17:17:15 +03:00
* @ mgmt : mesh management frame
* @ hwmp_ie : hwmp information element ( PREP or PREQ )
2013-01-03 09:09:46 +04:00
* @ action : type of hwmp ie
2008-02-23 17:17:15 +03:00
*
* This function updates the path routing information to the originator and the
2009-10-20 23:17:34 +04:00
* transmitter of a HWMP PREQ or PREP frame .
2008-02-23 17:17:15 +03:00
*
* Returns : metric to frame originator or 0 if the frame should not be further
* processed
*
* Notes : this function is the only place ( besides user - provided info ) where
* path routing information is updated .
*/
2008-08-03 04:04:37 +04:00
static u32 hwmp_route_info_get ( struct ieee80211_sub_if_data * sdata ,
2013-02-12 19:43:19 +04:00
struct ieee80211_mgmt * mgmt ,
const u8 * hwmp_ie , enum mpath_frame_type action )
2008-02-23 17:17:15 +03:00
{
2008-08-03 04:04:37 +04:00
struct ieee80211_local * local = sdata - > local ;
2008-02-23 17:17:15 +03:00
struct mesh_path * mpath ;
struct sta_info * sta ;
bool fresh_info ;
2013-02-12 19:43:19 +04:00
const u8 * orig_addr , * ta ;
2009-11-10 02:46:55 +03:00
u32 orig_sn , orig_metric ;
2008-02-23 17:17:15 +03:00
unsigned long orig_lifetime , exp_time ;
u32 last_hop_metric , new_metric ;
wifi: mac80211: mesh fast xmit support
Previously, fast xmit only worked on interface types where initially a
sta lookup is performed, and a cached header can be attached to the sta,
requiring only some fields to be updated at runtime.
This technique is not directly applicable for a mesh device type due
to the dynamic nature of the topology and protocol. There are more
addresses that need to be filled, and there is an extra header with a
dynamic length based on the addressing mode.
Change the code to cache entries contain a copy of the mesh subframe header +
bridge tunnel header, as well as an embedded struct ieee80211_fast_tx, which
contains the information for building the 802.11 header.
Add a mesh specific early fast xmit call, which looks up a cached entry and
adds only the mesh subframe header, before passing it over to the generic
fast xmit code.
To ensure the changes in network are reflected in these cached headers,
flush affected cached entries on path changes, as well as other conditions
that currently trigger a fast xmit check in other modes (key changes etc.)
This code is loosely based on a previous implementation by:
Sriram R <quic_srirrama@quicinc.com>
Cc: Sriram R <quic_srirrama@quicinc.com>
Signed-off-by: Ryder Lee <ryder.lee@mediatek.com>
Signed-off-by: Felix Fietkau <nbd@nbd.name>
Link: https://lore.kernel.org/r/20230314095956.62085-4-nbd@nbd.name
Signed-off-by: Johannes Berg <johannes.berg@intel.com>
2023-03-14 12:59:53 +03:00
bool flush_mpath = false ;
2008-02-23 17:17:15 +03:00
bool process = true ;
2019-01-16 02:28:42 +03:00
u8 hopcount ;
2008-02-23 17:17:15 +03:00
rcu_read_lock ( ) ;
2009-11-25 19:46:18 +03:00
sta = sta_info_get ( sdata , mgmt - > sa ) ;
2008-02-23 17:17:20 +03:00
if ( ! sta ) {
rcu_read_unlock ( ) ;
2008-02-23 17:17:15 +03:00
return 0 ;
2008-02-23 17:17:20 +03:00
}
2008-02-23 17:17:15 +03:00
last_hop_metric = airtime_link_metric_get ( local , sta ) ;
/* Update and check originator routing info */
fresh_info = true ;
switch ( action ) {
case MPATH_PREQ :
orig_addr = PREQ_IE_ORIG_ADDR ( hwmp_ie ) ;
2009-11-10 02:46:55 +03:00
orig_sn = PREQ_IE_ORIG_SN ( hwmp_ie ) ;
2008-02-23 17:17:15 +03:00
orig_lifetime = PREQ_IE_LIFETIME ( hwmp_ie ) ;
orig_metric = PREQ_IE_METRIC ( hwmp_ie ) ;
2019-01-16 02:28:42 +03:00
hopcount = PREQ_IE_HOPCOUNT ( hwmp_ie ) + 1 ;
2008-02-23 17:17:15 +03:00
break ;
case MPATH_PREP :
2011-11-25 05:15:22 +04:00
/* Originator here refers to the MP that was the target in the
* Path Request . We divert from the nomenclature in the draft
2008-02-23 17:17:15 +03:00
* so that we can easily use a single function to gather path
* information from both PREQ and PREP frames .
*/
2011-11-25 05:15:22 +04:00
orig_addr = PREP_IE_TARGET_ADDR ( hwmp_ie ) ;
orig_sn = PREP_IE_TARGET_SN ( hwmp_ie ) ;
2008-02-23 17:17:15 +03:00
orig_lifetime = PREP_IE_LIFETIME ( hwmp_ie ) ;
orig_metric = PREP_IE_METRIC ( hwmp_ie ) ;
2019-01-16 02:28:42 +03:00
hopcount = PREP_IE_HOPCOUNT ( hwmp_ie ) + 1 ;
2008-02-23 17:17:15 +03:00
break ;
default :
2008-02-23 17:17:20 +03:00
rcu_read_unlock ( ) ;
2008-02-23 17:17:15 +03:00
return 0 ;
}
new_metric = orig_metric + last_hop_metric ;
if ( new_metric < orig_metric )
new_metric = MAX_METRIC ;
exp_time = TU_TO_EXP_TIME ( orig_lifetime ) ;
mac80211: Convert compare_ether_addr to ether_addr_equal
Use the new bool function ether_addr_equal to add
some clarity and reduce the likelihood for misuse
of compare_ether_addr for sorting.
Done via cocci script:
$ cat compare_ether_addr.cocci
@@
expression a,b;
@@
- !compare_ether_addr(a, b)
+ ether_addr_equal(a, b)
@@
expression a,b;
@@
- compare_ether_addr(a, b)
+ !ether_addr_equal(a, b)
@@
expression a,b;
@@
- !ether_addr_equal(a, b) == 0
+ ether_addr_equal(a, b)
@@
expression a,b;
@@
- !ether_addr_equal(a, b) != 0
+ !ether_addr_equal(a, b)
@@
expression a,b;
@@
- ether_addr_equal(a, b) == 0
+ !ether_addr_equal(a, b)
@@
expression a,b;
@@
- ether_addr_equal(a, b) != 0
+ ether_addr_equal(a, b)
@@
expression a,b;
@@
- !!ether_addr_equal(a, b)
+ ether_addr_equal(a, b)
Signed-off-by: Joe Perches <joe@perches.com>
Signed-off-by: David S. Miller <davem@davemloft.net>
2012-05-08 22:56:52 +04:00
if ( ether_addr_equal ( orig_addr , sdata - > vif . addr ) ) {
2008-02-23 17:17:15 +03:00
/* This MP is the originator, we are not interested in this
* frame , except for updating transmitter ' s path info .
*/
process = false ;
fresh_info = false ;
} else {
2013-02-15 17:40:31 +04:00
mpath = mesh_path_lookup ( sdata , orig_addr ) ;
2008-02-23 17:17:15 +03:00
if ( mpath ) {
spin_lock_bh ( & mpath - > state_lock ) ;
if ( mpath - > flags & MESH_PATH_FIXED )
fresh_info = false ;
else if ( ( mpath - > flags & MESH_PATH_ACTIVE ) & &
2009-11-10 02:46:55 +03:00
( mpath - > flags & MESH_PATH_SN_VALID ) ) {
if ( SN_GT ( mpath - > sn , orig_sn ) | |
( mpath - > sn = = orig_sn & &
2019-01-16 02:33:03 +03:00
( rcu_access_pointer ( mpath - > next_hop ) ! =
sta ?
mult_frac ( new_metric , 10 , 9 ) :
new_metric ) > = mpath - > metric ) ) {
2008-02-23 17:17:15 +03:00
process = false ;
fresh_info = false ;
}
2015-06-13 01:38:07 +03:00
} else if ( ! ( mpath - > flags & MESH_PATH_ACTIVE ) ) {
bool have_sn , newer_sn , bounced ;
have_sn = mpath - > flags & MESH_PATH_SN_VALID ;
newer_sn = have_sn & & SN_GT ( orig_sn , mpath - > sn ) ;
bounced = have_sn & &
( SN_DELTA ( orig_sn , mpath - > sn ) >
MAX_SANE_SN_DELTA ) ;
if ( ! have_sn | | newer_sn ) {
/* if SN is newer than what we had
* then we can take it */ ;
} else if ( bounced ) {
/* if SN is way different than what
* we had then assume the other side
* rebooted or restarted */ ;
} else {
process = false ;
fresh_info = false ;
}
2008-02-23 17:17:15 +03:00
}
} else {
2013-03-29 17:38:39 +04:00
mpath = mesh_path_add ( sdata , orig_addr ) ;
if ( IS_ERR ( mpath ) ) {
2008-02-23 17:17:15 +03:00
rcu_read_unlock ( ) ;
return 0 ;
}
spin_lock_bh ( & mpath - > state_lock ) ;
}
if ( fresh_info ) {
wifi: mac80211: mesh fast xmit support
Previously, fast xmit only worked on interface types where initially a
sta lookup is performed, and a cached header can be attached to the sta,
requiring only some fields to be updated at runtime.
This technique is not directly applicable for a mesh device type due
to the dynamic nature of the topology and protocol. There are more
addresses that need to be filled, and there is an extra header with a
dynamic length based on the addressing mode.
Change the code to cache entries contain a copy of the mesh subframe header +
bridge tunnel header, as well as an embedded struct ieee80211_fast_tx, which
contains the information for building the 802.11 header.
Add a mesh specific early fast xmit call, which looks up a cached entry and
adds only the mesh subframe header, before passing it over to the generic
fast xmit code.
To ensure the changes in network are reflected in these cached headers,
flush affected cached entries on path changes, as well as other conditions
that currently trigger a fast xmit check in other modes (key changes etc.)
This code is loosely based on a previous implementation by:
Sriram R <quic_srirrama@quicinc.com>
Cc: Sriram R <quic_srirrama@quicinc.com>
Signed-off-by: Ryder Lee <ryder.lee@mediatek.com>
Signed-off-by: Felix Fietkau <nbd@nbd.name>
Link: https://lore.kernel.org/r/20230314095956.62085-4-nbd@nbd.name
Signed-off-by: Johannes Berg <johannes.berg@intel.com>
2023-03-14 12:59:53 +03:00
if ( rcu_access_pointer ( mpath - > next_hop ) ! = sta ) {
2019-01-16 02:28:43 +03:00
mpath - > path_change_count + + ;
wifi: mac80211: mesh fast xmit support
Previously, fast xmit only worked on interface types where initially a
sta lookup is performed, and a cached header can be attached to the sta,
requiring only some fields to be updated at runtime.
This technique is not directly applicable for a mesh device type due
to the dynamic nature of the topology and protocol. There are more
addresses that need to be filled, and there is an extra header with a
dynamic length based on the addressing mode.
Change the code to cache entries contain a copy of the mesh subframe header +
bridge tunnel header, as well as an embedded struct ieee80211_fast_tx, which
contains the information for building the 802.11 header.
Add a mesh specific early fast xmit call, which looks up a cached entry and
adds only the mesh subframe header, before passing it over to the generic
fast xmit code.
To ensure the changes in network are reflected in these cached headers,
flush affected cached entries on path changes, as well as other conditions
that currently trigger a fast xmit check in other modes (key changes etc.)
This code is loosely based on a previous implementation by:
Sriram R <quic_srirrama@quicinc.com>
Cc: Sriram R <quic_srirrama@quicinc.com>
Signed-off-by: Ryder Lee <ryder.lee@mediatek.com>
Signed-off-by: Felix Fietkau <nbd@nbd.name>
Link: https://lore.kernel.org/r/20230314095956.62085-4-nbd@nbd.name
Signed-off-by: Johannes Berg <johannes.berg@intel.com>
2023-03-14 12:59:53 +03:00
flush_mpath = true ;
}
2008-02-23 17:17:15 +03:00
mesh_path_assign_nexthop ( mpath , sta ) ;
2009-11-10 02:46:55 +03:00
mpath - > flags | = MESH_PATH_SN_VALID ;
2008-02-23 17:17:15 +03:00
mpath - > metric = new_metric ;
2009-11-10 02:46:55 +03:00
mpath - > sn = orig_sn ;
2008-02-23 17:17:15 +03:00
mpath - > exp_time = time_after ( mpath - > exp_time , exp_time )
? mpath - > exp_time : exp_time ;
2019-01-16 02:28:42 +03:00
mpath - > hop_count = hopcount ;
2008-02-23 17:17:15 +03:00
mesh_path_activate ( mpath ) ;
spin_unlock_bh ( & mpath - > state_lock ) ;
wifi: mac80211: mesh fast xmit support
Previously, fast xmit only worked on interface types where initially a
sta lookup is performed, and a cached header can be attached to the sta,
requiring only some fields to be updated at runtime.
This technique is not directly applicable for a mesh device type due
to the dynamic nature of the topology and protocol. There are more
addresses that need to be filled, and there is an extra header with a
dynamic length based on the addressing mode.
Change the code to cache entries contain a copy of the mesh subframe header +
bridge tunnel header, as well as an embedded struct ieee80211_fast_tx, which
contains the information for building the 802.11 header.
Add a mesh specific early fast xmit call, which looks up a cached entry and
adds only the mesh subframe header, before passing it over to the generic
fast xmit code.
To ensure the changes in network are reflected in these cached headers,
flush affected cached entries on path changes, as well as other conditions
that currently trigger a fast xmit check in other modes (key changes etc.)
This code is loosely based on a previous implementation by:
Sriram R <quic_srirrama@quicinc.com>
Cc: Sriram R <quic_srirrama@quicinc.com>
Signed-off-by: Ryder Lee <ryder.lee@mediatek.com>
Signed-off-by: Felix Fietkau <nbd@nbd.name>
Link: https://lore.kernel.org/r/20230314095956.62085-4-nbd@nbd.name
Signed-off-by: Johannes Berg <johannes.berg@intel.com>
2023-03-14 12:59:53 +03:00
if ( flush_mpath )
mesh_fast_tx_flush_mpath ( mpath ) ;
2017-02-14 23:27:16 +03:00
ewma_mesh_fail_avg_init ( & sta - > mesh - > fail_avg ) ;
/* init it at a low value - 0 start is tricky */
ewma_mesh_fail_avg_add ( & sta - > mesh - > fail_avg , 1 ) ;
2008-02-23 17:17:15 +03:00
mesh_path_tx_pending ( mpath ) ;
/* draft says preq_id should be saved to, but there does
* not seem to be any use for it , skipping by now
*/
} else
spin_unlock_bh ( & mpath - > state_lock ) ;
}
/* Update and check transmitter routing info */
ta = mgmt - > sa ;
mac80211: Convert compare_ether_addr to ether_addr_equal
Use the new bool function ether_addr_equal to add
some clarity and reduce the likelihood for misuse
of compare_ether_addr for sorting.
Done via cocci script:
$ cat compare_ether_addr.cocci
@@
expression a,b;
@@
- !compare_ether_addr(a, b)
+ ether_addr_equal(a, b)
@@
expression a,b;
@@
- compare_ether_addr(a, b)
+ !ether_addr_equal(a, b)
@@
expression a,b;
@@
- !ether_addr_equal(a, b) == 0
+ ether_addr_equal(a, b)
@@
expression a,b;
@@
- !ether_addr_equal(a, b) != 0
+ !ether_addr_equal(a, b)
@@
expression a,b;
@@
- ether_addr_equal(a, b) == 0
+ !ether_addr_equal(a, b)
@@
expression a,b;
@@
- ether_addr_equal(a, b) != 0
+ ether_addr_equal(a, b)
@@
expression a,b;
@@
- !!ether_addr_equal(a, b)
+ ether_addr_equal(a, b)
Signed-off-by: Joe Perches <joe@perches.com>
Signed-off-by: David S. Miller <davem@davemloft.net>
2012-05-08 22:56:52 +04:00
if ( ether_addr_equal ( orig_addr , ta ) )
2008-02-23 17:17:15 +03:00
fresh_info = false ;
else {
fresh_info = true ;
2013-02-15 17:40:31 +04:00
mpath = mesh_path_lookup ( sdata , ta ) ;
2008-02-23 17:17:15 +03:00
if ( mpath ) {
spin_lock_bh ( & mpath - > state_lock ) ;
if ( ( mpath - > flags & MESH_PATH_FIXED ) | |
2019-01-16 02:33:03 +03:00
( ( mpath - > flags & MESH_PATH_ACTIVE ) & &
( ( rcu_access_pointer ( mpath - > next_hop ) ! = sta ?
mult_frac ( last_hop_metric , 10 , 9 ) :
last_hop_metric ) > mpath - > metric ) ) )
2008-02-23 17:17:15 +03:00
fresh_info = false ;
} else {
2013-03-29 17:38:39 +04:00
mpath = mesh_path_add ( sdata , ta ) ;
if ( IS_ERR ( mpath ) ) {
2008-02-23 17:17:15 +03:00
rcu_read_unlock ( ) ;
return 0 ;
}
spin_lock_bh ( & mpath - > state_lock ) ;
}
if ( fresh_info ) {
wifi: mac80211: mesh fast xmit support
Previously, fast xmit only worked on interface types where initially a
sta lookup is performed, and a cached header can be attached to the sta,
requiring only some fields to be updated at runtime.
This technique is not directly applicable for a mesh device type due
to the dynamic nature of the topology and protocol. There are more
addresses that need to be filled, and there is an extra header with a
dynamic length based on the addressing mode.
Change the code to cache entries contain a copy of the mesh subframe header +
bridge tunnel header, as well as an embedded struct ieee80211_fast_tx, which
contains the information for building the 802.11 header.
Add a mesh specific early fast xmit call, which looks up a cached entry and
adds only the mesh subframe header, before passing it over to the generic
fast xmit code.
To ensure the changes in network are reflected in these cached headers,
flush affected cached entries on path changes, as well as other conditions
that currently trigger a fast xmit check in other modes (key changes etc.)
This code is loosely based on a previous implementation by:
Sriram R <quic_srirrama@quicinc.com>
Cc: Sriram R <quic_srirrama@quicinc.com>
Signed-off-by: Ryder Lee <ryder.lee@mediatek.com>
Signed-off-by: Felix Fietkau <nbd@nbd.name>
Link: https://lore.kernel.org/r/20230314095956.62085-4-nbd@nbd.name
Signed-off-by: Johannes Berg <johannes.berg@intel.com>
2023-03-14 12:59:53 +03:00
if ( rcu_access_pointer ( mpath - > next_hop ) ! = sta ) {
2019-01-16 02:28:43 +03:00
mpath - > path_change_count + + ;
wifi: mac80211: mesh fast xmit support
Previously, fast xmit only worked on interface types where initially a
sta lookup is performed, and a cached header can be attached to the sta,
requiring only some fields to be updated at runtime.
This technique is not directly applicable for a mesh device type due
to the dynamic nature of the topology and protocol. There are more
addresses that need to be filled, and there is an extra header with a
dynamic length based on the addressing mode.
Change the code to cache entries contain a copy of the mesh subframe header +
bridge tunnel header, as well as an embedded struct ieee80211_fast_tx, which
contains the information for building the 802.11 header.
Add a mesh specific early fast xmit call, which looks up a cached entry and
adds only the mesh subframe header, before passing it over to the generic
fast xmit code.
To ensure the changes in network are reflected in these cached headers,
flush affected cached entries on path changes, as well as other conditions
that currently trigger a fast xmit check in other modes (key changes etc.)
This code is loosely based on a previous implementation by:
Sriram R <quic_srirrama@quicinc.com>
Cc: Sriram R <quic_srirrama@quicinc.com>
Signed-off-by: Ryder Lee <ryder.lee@mediatek.com>
Signed-off-by: Felix Fietkau <nbd@nbd.name>
Link: https://lore.kernel.org/r/20230314095956.62085-4-nbd@nbd.name
Signed-off-by: Johannes Berg <johannes.berg@intel.com>
2023-03-14 12:59:53 +03:00
flush_mpath = true ;
}
2008-02-23 17:17:15 +03:00
mesh_path_assign_nexthop ( mpath , sta ) ;
mpath - > metric = last_hop_metric ;
mpath - > exp_time = time_after ( mpath - > exp_time , exp_time )
? mpath - > exp_time : exp_time ;
2019-01-16 02:28:42 +03:00
mpath - > hop_count = 1 ;
2008-02-23 17:17:15 +03:00
mesh_path_activate ( mpath ) ;
spin_unlock_bh ( & mpath - > state_lock ) ;
wifi: mac80211: mesh fast xmit support
Previously, fast xmit only worked on interface types where initially a
sta lookup is performed, and a cached header can be attached to the sta,
requiring only some fields to be updated at runtime.
This technique is not directly applicable for a mesh device type due
to the dynamic nature of the topology and protocol. There are more
addresses that need to be filled, and there is an extra header with a
dynamic length based on the addressing mode.
Change the code to cache entries contain a copy of the mesh subframe header +
bridge tunnel header, as well as an embedded struct ieee80211_fast_tx, which
contains the information for building the 802.11 header.
Add a mesh specific early fast xmit call, which looks up a cached entry and
adds only the mesh subframe header, before passing it over to the generic
fast xmit code.
To ensure the changes in network are reflected in these cached headers,
flush affected cached entries on path changes, as well as other conditions
that currently trigger a fast xmit check in other modes (key changes etc.)
This code is loosely based on a previous implementation by:
Sriram R <quic_srirrama@quicinc.com>
Cc: Sriram R <quic_srirrama@quicinc.com>
Signed-off-by: Ryder Lee <ryder.lee@mediatek.com>
Signed-off-by: Felix Fietkau <nbd@nbd.name>
Link: https://lore.kernel.org/r/20230314095956.62085-4-nbd@nbd.name
Signed-off-by: Johannes Berg <johannes.berg@intel.com>
2023-03-14 12:59:53 +03:00
if ( flush_mpath )
mesh_fast_tx_flush_mpath ( mpath ) ;
2017-02-14 23:27:16 +03:00
ewma_mesh_fail_avg_init ( & sta - > mesh - > fail_avg ) ;
/* init it at a low value - 0 start is tricky */
ewma_mesh_fail_avg_add ( & sta - > mesh - > fail_avg , 1 ) ;
2008-02-23 17:17:15 +03:00
mesh_path_tx_pending ( mpath ) ;
} else
spin_unlock_bh ( & mpath - > state_lock ) ;
}
rcu_read_unlock ( ) ;
return process ? new_metric : 0 ;
}
2008-08-03 04:04:37 +04:00
static void hwmp_preq_frame_process ( struct ieee80211_sub_if_data * sdata ,
2008-02-23 17:17:15 +03:00
struct ieee80211_mgmt * mgmt ,
2015-06-10 02:20:24 +03:00
const u8 * preq_elem , u32 orig_metric )
2009-08-12 22:03:43 +04:00
{
2008-09-11 02:01:49 +04:00
struct ieee80211_if_mesh * ifmsh = & sdata - > u . mesh ;
2012-02-28 18:00:06 +04:00
struct mesh_path * mpath = NULL ;
2013-02-12 19:43:19 +04:00
const u8 * target_addr , * orig_addr ;
2012-02-28 18:00:06 +04:00
const u8 * da ;
2012-06-16 22:27:40 +04:00
u8 target_flags , ttl , flags ;
2016-04-04 21:15:23 +03:00
u32 orig_sn , target_sn , lifetime , target_metric = 0 ;
2008-02-23 17:17:15 +03:00
bool reply = false ;
bool forward = true ;
2012-06-16 22:27:40 +04:00
bool root_is_gate ;
2008-02-23 17:17:15 +03:00
2009-11-10 02:46:55 +03:00
/* Update target SN, if present */
target_addr = PREQ_IE_TARGET_ADDR ( preq_elem ) ;
2008-02-23 17:17:15 +03:00
orig_addr = PREQ_IE_ORIG_ADDR ( preq_elem ) ;
2009-11-10 02:46:55 +03:00
target_sn = PREQ_IE_TARGET_SN ( preq_elem ) ;
orig_sn = PREQ_IE_ORIG_SN ( preq_elem ) ;
target_flags = PREQ_IE_TARGET_F ( preq_elem ) ;
2012-06-16 22:27:40 +04:00
/* Proactive PREQ gate announcements */
flags = PREQ_IE_FLAGS ( preq_elem ) ;
root_is_gate = ! ! ( flags & RANN_FLAG_IS_GATE ) ;
2008-02-23 17:17:15 +03:00
2012-06-22 13:29:50 +04:00
mhwmp_dbg ( sdata , " received PREQ from %pM \n " , orig_addr ) ;
2009-11-10 02:46:45 +03:00
mac80211: Convert compare_ether_addr to ether_addr_equal
Use the new bool function ether_addr_equal to add
some clarity and reduce the likelihood for misuse
of compare_ether_addr for sorting.
Done via cocci script:
$ cat compare_ether_addr.cocci
@@
expression a,b;
@@
- !compare_ether_addr(a, b)
+ ether_addr_equal(a, b)
@@
expression a,b;
@@
- compare_ether_addr(a, b)
+ !ether_addr_equal(a, b)
@@
expression a,b;
@@
- !ether_addr_equal(a, b) == 0
+ ether_addr_equal(a, b)
@@
expression a,b;
@@
- !ether_addr_equal(a, b) != 0
+ !ether_addr_equal(a, b)
@@
expression a,b;
@@
- ether_addr_equal(a, b) == 0
+ !ether_addr_equal(a, b)
@@
expression a,b;
@@
- ether_addr_equal(a, b) != 0
+ ether_addr_equal(a, b)
@@
expression a,b;
@@
- !!ether_addr_equal(a, b)
+ ether_addr_equal(a, b)
Signed-off-by: Joe Perches <joe@perches.com>
Signed-off-by: David S. Miller <davem@davemloft.net>
2012-05-08 22:56:52 +04:00
if ( ether_addr_equal ( target_addr , sdata - > vif . addr ) ) {
2012-06-22 13:29:50 +04:00
mhwmp_dbg ( sdata , " PREQ is for us \n " ) ;
2008-02-23 17:17:15 +03:00
forward = false ;
reply = true ;
2015-06-10 02:20:24 +03:00
target_metric = 0 ;
2018-08-29 04:30:08 +03:00
if ( SN_GT ( target_sn , ifmsh - > sn ) )
ifmsh - > sn = target_sn ;
2009-11-10 02:46:55 +03:00
if ( time_after ( jiffies , ifmsh - > last_sn_update +
2008-02-23 17:17:15 +03:00
net_traversal_jiffies ( sdata ) ) | |
2009-11-10 02:46:55 +03:00
time_before ( jiffies , ifmsh - > last_sn_update ) ) {
2014-04-15 18:43:08 +04:00
+ + ifmsh - > sn ;
2009-11-10 02:46:55 +03:00
ifmsh - > last_sn_update = jiffies ;
2008-02-23 17:17:15 +03:00
}
2014-04-15 18:43:08 +04:00
target_sn = ifmsh - > sn ;
2012-06-16 22:27:40 +04:00
} else if ( is_broadcast_ether_addr ( target_addr ) & &
( target_flags & IEEE80211_PREQ_TO_FLAG ) ) {
rcu_read_lock ( ) ;
2013-02-15 17:40:31 +04:00
mpath = mesh_path_lookup ( sdata , orig_addr ) ;
2012-06-16 22:27:40 +04:00
if ( mpath ) {
if ( flags & IEEE80211_PREQ_PROACTIVE_PREP_FLAG ) {
reply = true ;
target_addr = sdata - > vif . addr ;
target_sn = + + ifmsh - > sn ;
2015-06-10 02:20:24 +03:00
target_metric = 0 ;
2012-06-16 22:27:40 +04:00
ifmsh - > last_sn_update = jiffies ;
}
if ( root_is_gate )
mesh_path_add_gate ( mpath ) ;
}
rcu_read_unlock ( ) ;
2008-02-23 17:17:15 +03:00
} else {
rcu_read_lock ( ) ;
2013-02-15 17:40:31 +04:00
mpath = mesh_path_lookup ( sdata , target_addr ) ;
2008-02-23 17:17:15 +03:00
if ( mpath ) {
2009-11-10 02:46:55 +03:00
if ( ( ! ( mpath - > flags & MESH_PATH_SN_VALID ) ) | |
SN_LT ( mpath - > sn , target_sn ) ) {
mpath - > sn = target_sn ;
mpath - > flags | = MESH_PATH_SN_VALID ;
2015-06-15 06:58:53 +03:00
} else if ( ( ! ( target_flags & IEEE80211_PREQ_TO_FLAG ) ) & &
2008-02-23 17:17:15 +03:00
( mpath - > flags & MESH_PATH_ACTIVE ) ) {
reply = true ;
2015-06-10 02:20:24 +03:00
target_metric = mpath - > metric ;
2009-11-10 02:46:55 +03:00
target_sn = mpath - > sn ;
2015-06-15 06:58:53 +03:00
/* Case E2 of sec 13.10.9.3 IEEE 802.11-2012*/
target_flags | = IEEE80211_PREQ_TO_FLAG ;
2008-02-23 17:17:15 +03:00
}
}
rcu_read_unlock ( ) ;
}
if ( reply ) {
lifetime = PREQ_IE_LIFETIME ( preq_elem ) ;
2010-12-03 11:20:40 +03:00
ttl = ifmsh - > mshcfg . element_ttl ;
2009-11-10 02:46:45 +03:00
if ( ttl ! = 0 ) {
2012-06-22 13:29:50 +04:00
mhwmp_dbg ( sdata , " replying to the PREQ \n " ) ;
2011-11-25 05:15:22 +04:00
mesh_path_sel_frame_tx ( MPATH_PREP , 0 , orig_addr ,
2013-11-13 11:39:12 +04:00
orig_sn , 0 , target_addr ,
target_sn , mgmt - > sa , 0 , ttl ,
2015-06-10 02:20:24 +03:00
lifetime , target_metric , 0 ,
sdata ) ;
2012-06-16 22:27:40 +04:00
} else {
2008-09-11 02:01:49 +04:00
ifmsh - > mshstats . dropped_frames_ttl + + ;
2012-06-16 22:27:40 +04:00
}
2008-02-23 17:17:15 +03:00
}
2012-01-20 21:02:16 +04:00
if ( forward & & ifmsh - > mshcfg . dot11MeshForwarding ) {
2008-02-23 17:17:15 +03:00
u32 preq_id ;
2012-06-16 22:27:40 +04:00
u8 hopcount ;
2008-02-23 17:17:15 +03:00
ttl = PREQ_IE_TTL ( preq_elem ) ;
lifetime = PREQ_IE_LIFETIME ( preq_elem ) ;
if ( ttl < = 1 ) {
2008-09-11 02:01:49 +04:00
ifmsh - > mshstats . dropped_frames_ttl + + ;
2008-02-23 17:17:15 +03:00
return ;
}
2012-06-22 13:29:50 +04:00
mhwmp_dbg ( sdata , " forwarding the PREQ from %pM \n " , orig_addr ) ;
2008-02-23 17:17:15 +03:00
- - ttl ;
preq_id = PREQ_IE_PREQ_ID ( preq_elem ) ;
hopcount = PREQ_IE_HOPCOUNT ( preq_elem ) + 1 ;
2012-02-28 18:00:06 +04:00
da = ( mpath & & mpath - > is_root ) ?
mpath - > rann_snd_addr : broadcast_addr ;
2012-06-16 22:27:40 +04:00
if ( flags & IEEE80211_PREQ_PROACTIVE_PREP_FLAG ) {
target_addr = PREQ_IE_TARGET_ADDR ( preq_elem ) ;
target_sn = PREQ_IE_TARGET_SN ( preq_elem ) ;
}
2008-02-23 17:17:15 +03:00
mesh_path_sel_frame_tx ( MPATH_PREQ , flags , orig_addr ,
2013-11-13 11:39:12 +04:00
orig_sn , target_flags , target_addr ,
target_sn , da , hopcount , ttl , lifetime ,
2015-06-10 02:20:24 +03:00
orig_metric , preq_id , sdata ) ;
2012-05-04 10:57:50 +04:00
if ( ! is_multicast_ether_addr ( da ) )
ifmsh - > mshstats . fwded_unicast + + ;
else
ifmsh - > mshstats . fwded_mcast + + ;
2008-09-11 02:01:49 +04:00
ifmsh - > mshstats . fwded_frames + + ;
2008-02-23 17:17:15 +03:00
}
}
2011-05-13 16:15:49 +04:00
static inline struct sta_info *
next_hop_deref_protected ( struct mesh_path * mpath )
{
return rcu_dereference_protected ( mpath - > next_hop ,
lockdep_is_held ( & mpath - > state_lock ) ) ;
}
2008-08-03 04:04:37 +04:00
static void hwmp_prep_frame_process ( struct ieee80211_sub_if_data * sdata ,
2008-02-23 17:17:15 +03:00
struct ieee80211_mgmt * mgmt ,
2013-02-12 19:43:19 +04:00
const u8 * prep_elem , u32 metric )
2008-02-23 17:17:15 +03:00
{
2012-03-01 22:03:19 +04:00
struct ieee80211_if_mesh * ifmsh = & sdata - > u . mesh ;
2008-02-23 17:17:15 +03:00
struct mesh_path * mpath ;
2013-02-12 19:43:19 +04:00
const u8 * target_addr , * orig_addr ;
2008-02-23 17:17:15 +03:00
u8 ttl , hopcount , flags ;
u8 next_hop [ ETH_ALEN ] ;
2009-11-10 02:46:55 +03:00
u32 target_sn , orig_sn , lifetime ;
2008-02-23 17:17:15 +03:00
2012-06-22 13:29:50 +04:00
mhwmp_dbg ( sdata , " received PREP from %pM \n " ,
2013-04-03 13:49:53 +04:00
PREP_IE_TARGET_ADDR ( prep_elem ) ) ;
2009-11-10 02:46:46 +03:00
2011-11-25 05:15:22 +04:00
orig_addr = PREP_IE_ORIG_ADDR ( prep_elem ) ;
mac80211: Convert compare_ether_addr to ether_addr_equal
Use the new bool function ether_addr_equal to add
some clarity and reduce the likelihood for misuse
of compare_ether_addr for sorting.
Done via cocci script:
$ cat compare_ether_addr.cocci
@@
expression a,b;
@@
- !compare_ether_addr(a, b)
+ ether_addr_equal(a, b)
@@
expression a,b;
@@
- compare_ether_addr(a, b)
+ !ether_addr_equal(a, b)
@@
expression a,b;
@@
- !ether_addr_equal(a, b) == 0
+ ether_addr_equal(a, b)
@@
expression a,b;
@@
- !ether_addr_equal(a, b) != 0
+ !ether_addr_equal(a, b)
@@
expression a,b;
@@
- ether_addr_equal(a, b) == 0
+ !ether_addr_equal(a, b)
@@
expression a,b;
@@
- ether_addr_equal(a, b) != 0
+ ether_addr_equal(a, b)
@@
expression a,b;
@@
- !!ether_addr_equal(a, b)
+ ether_addr_equal(a, b)
Signed-off-by: Joe Perches <joe@perches.com>
Signed-off-by: David S. Miller <davem@davemloft.net>
2012-05-08 22:56:52 +04:00
if ( ether_addr_equal ( orig_addr , sdata - > vif . addr ) )
2008-02-23 17:17:15 +03:00
/* destination, no forwarding required */
return ;
2012-03-01 22:03:19 +04:00
if ( ! ifmsh - > mshcfg . dot11MeshForwarding )
return ;
2008-02-23 17:17:15 +03:00
ttl = PREP_IE_TTL ( prep_elem ) ;
if ( ttl < = 1 ) {
2008-09-11 02:01:49 +04:00
sdata - > u . mesh . mshstats . dropped_frames_ttl + + ;
2008-02-23 17:17:15 +03:00
return ;
}
rcu_read_lock ( ) ;
2013-02-15 17:40:31 +04:00
mpath = mesh_path_lookup ( sdata , orig_addr ) ;
2008-02-23 17:17:15 +03:00
if ( mpath )
spin_lock_bh ( & mpath - > state_lock ) ;
else
goto fail ;
if ( ! ( mpath - > flags & MESH_PATH_ACTIVE ) ) {
spin_unlock_bh ( & mpath - > state_lock ) ;
goto fail ;
}
2011-05-13 16:15:49 +04:00
memcpy ( next_hop , next_hop_deref_protected ( mpath ) - > sta . addr , ETH_ALEN ) ;
2008-02-23 17:17:15 +03:00
spin_unlock_bh ( & mpath - > state_lock ) ;
- - ttl ;
flags = PREP_IE_FLAGS ( prep_elem ) ;
lifetime = PREP_IE_LIFETIME ( prep_elem ) ;
hopcount = PREP_IE_HOPCOUNT ( prep_elem ) + 1 ;
2011-11-25 05:15:22 +04:00
target_addr = PREP_IE_TARGET_ADDR ( prep_elem ) ;
2009-11-10 02:46:55 +03:00
target_sn = PREP_IE_TARGET_SN ( prep_elem ) ;
orig_sn = PREP_IE_ORIG_SN ( prep_elem ) ;
2008-02-23 17:17:15 +03:00
2013-11-13 11:39:12 +04:00
mesh_path_sel_frame_tx ( MPATH_PREP , flags , orig_addr , orig_sn , 0 ,
target_addr , target_sn , next_hop , hopcount ,
ttl , lifetime , metric , 0 , sdata ) ;
2008-02-23 17:17:15 +03:00
rcu_read_unlock ( ) ;
2009-08-18 21:59:00 +04:00
sdata - > u . mesh . mshstats . fwded_unicast + + ;
2008-09-11 02:01:49 +04:00
sdata - > u . mesh . mshstats . fwded_frames + + ;
2008-02-23 17:17:15 +03:00
return ;
fail :
rcu_read_unlock ( ) ;
2008-09-11 02:01:49 +04:00
sdata - > u . mesh . mshstats . dropped_frames_no_route + + ;
2008-02-23 17:17:15 +03:00
}
2008-08-03 04:04:37 +04:00
static void hwmp_perr_frame_process ( struct ieee80211_sub_if_data * sdata ,
2013-02-12 19:43:19 +04:00
struct ieee80211_mgmt * mgmt ,
const u8 * perr_elem )
2008-02-23 17:17:15 +03:00
{
2009-11-10 02:46:50 +03:00
struct ieee80211_if_mesh * ifmsh = & sdata - > u . mesh ;
2008-02-23 17:17:15 +03:00
struct mesh_path * mpath ;
2009-11-10 02:46:50 +03:00
u8 ttl ;
2013-02-12 19:43:19 +04:00
const u8 * ta , * target_addr ;
2009-11-10 02:46:55 +03:00
u32 target_sn ;
u16 target_rcode ;
2008-02-23 17:17:15 +03:00
ta = mgmt - > sa ;
2009-11-10 02:46:50 +03:00
ttl = PERR_IE_TTL ( perr_elem ) ;
if ( ttl < = 1 ) {
ifmsh - > mshstats . dropped_frames_ttl + + ;
return ;
}
ttl - - ;
2009-11-10 02:46:55 +03:00
target_addr = PERR_IE_TARGET_ADDR ( perr_elem ) ;
target_sn = PERR_IE_TARGET_SN ( perr_elem ) ;
target_rcode = PERR_IE_TARGET_RCODE ( perr_elem ) ;
2009-11-10 02:46:50 +03:00
2008-02-23 17:17:15 +03:00
rcu_read_lock ( ) ;
2013-02-15 17:40:31 +04:00
mpath = mesh_path_lookup ( sdata , target_addr ) ;
2008-02-23 17:17:15 +03:00
if ( mpath ) {
2012-03-01 18:22:09 +04:00
struct sta_info * sta ;
2008-02-23 17:17:15 +03:00
spin_lock_bh ( & mpath - > state_lock ) ;
2012-03-01 18:22:09 +04:00
sta = next_hop_deref_protected ( mpath ) ;
2008-02-23 17:17:15 +03:00
if ( mpath - > flags & MESH_PATH_ACTIVE & &
mac80211: Convert compare_ether_addr to ether_addr_equal
Use the new bool function ether_addr_equal to add
some clarity and reduce the likelihood for misuse
of compare_ether_addr for sorting.
Done via cocci script:
$ cat compare_ether_addr.cocci
@@
expression a,b;
@@
- !compare_ether_addr(a, b)
+ ether_addr_equal(a, b)
@@
expression a,b;
@@
- compare_ether_addr(a, b)
+ !ether_addr_equal(a, b)
@@
expression a,b;
@@
- !ether_addr_equal(a, b) == 0
+ ether_addr_equal(a, b)
@@
expression a,b;
@@
- !ether_addr_equal(a, b) != 0
+ !ether_addr_equal(a, b)
@@
expression a,b;
@@
- ether_addr_equal(a, b) == 0
+ !ether_addr_equal(a, b)
@@
expression a,b;
@@
- ether_addr_equal(a, b) != 0
+ ether_addr_equal(a, b)
@@
expression a,b;
@@
- !!ether_addr_equal(a, b)
+ ether_addr_equal(a, b)
Signed-off-by: Joe Perches <joe@perches.com>
Signed-off-by: David S. Miller <davem@davemloft.net>
2012-05-08 22:56:52 +04:00
ether_addr_equal ( ta , sta - > sta . addr ) & &
2016-09-06 21:59:00 +03:00
! ( mpath - > flags & MESH_PATH_FIXED ) & &
2009-11-10 02:46:55 +03:00
( ! ( mpath - > flags & MESH_PATH_SN_VALID ) | |
2015-06-10 21:02:09 +03:00
SN_GT ( target_sn , mpath - > sn ) | | target_sn = = 0 ) ) {
2008-02-23 17:17:15 +03:00
mpath - > flags & = ~ MESH_PATH_ACTIVE ;
2015-06-10 21:02:09 +03:00
if ( target_sn ! = 0 )
mpath - > sn = target_sn ;
else
mpath - > sn + = 1 ;
2008-02-23 17:17:15 +03:00
spin_unlock_bh ( & mpath - > state_lock ) ;
2012-03-01 22:03:19 +04:00
if ( ! ifmsh - > mshcfg . dot11MeshForwarding )
goto endperr ;
2013-02-15 17:40:31 +04:00
mesh_path_error_tx ( sdata , ttl , target_addr ,
2013-11-13 11:39:12 +04:00
target_sn , target_rcode ,
2013-02-15 17:40:31 +04:00
broadcast_addr ) ;
2008-02-23 17:17:15 +03:00
} else
spin_unlock_bh ( & mpath - > state_lock ) ;
}
2012-03-01 22:03:19 +04:00
endperr :
2008-02-23 17:17:15 +03:00
rcu_read_unlock ( ) ;
}
2009-11-11 03:01:31 +03:00
static void hwmp_rann_frame_process ( struct ieee80211_sub_if_data * sdata ,
2013-02-12 19:43:19 +04:00
struct ieee80211_mgmt * mgmt ,
const struct ieee80211_rann_ie * rann )
2009-11-11 03:01:31 +03:00
{
struct ieee80211_if_mesh * ifmsh = & sdata - > u . mesh ;
2012-03-23 14:48:51 +04:00
struct ieee80211_local * local = sdata - > local ;
struct sta_info * sta ;
2009-11-11 03:01:31 +03:00
struct mesh_path * mpath ;
u8 ttl , flags , hopcount ;
2013-02-12 19:43:19 +04:00
const u8 * orig_addr ;
2017-11-14 18:20:05 +03:00
u32 orig_sn , new_metric , orig_metric , last_hop_metric , interval ;
2011-08-10 03:45:08 +04:00
bool root_is_gate ;
2009-11-11 03:01:31 +03:00
ttl = rann - > rann_ttl ;
flags = rann - > rann_flags ;
2011-08-10 03:45:08 +04:00
root_is_gate = ! ! ( flags & RANN_FLAG_IS_GATE ) ;
2009-11-11 03:01:31 +03:00
orig_addr = rann - > rann_addr ;
2012-03-19 17:38:46 +04:00
orig_sn = le32_to_cpu ( rann - > rann_seq ) ;
2012-03-23 14:48:51 +04:00
interval = le32_to_cpu ( rann - > rann_interval ) ;
2009-11-11 03:01:31 +03:00
hopcount = rann - > rann_hopcount ;
2009-11-10 02:46:51 +03:00
hopcount + + ;
2017-11-14 18:20:05 +03:00
orig_metric = le32_to_cpu ( rann - > rann_metric ) ;
2011-08-10 03:45:08 +04:00
/* Ignore our own RANNs */
mac80211: Convert compare_ether_addr to ether_addr_equal
Use the new bool function ether_addr_equal to add
some clarity and reduce the likelihood for misuse
of compare_ether_addr for sorting.
Done via cocci script:
$ cat compare_ether_addr.cocci
@@
expression a,b;
@@
- !compare_ether_addr(a, b)
+ ether_addr_equal(a, b)
@@
expression a,b;
@@
- compare_ether_addr(a, b)
+ !ether_addr_equal(a, b)
@@
expression a,b;
@@
- !ether_addr_equal(a, b) == 0
+ ether_addr_equal(a, b)
@@
expression a,b;
@@
- !ether_addr_equal(a, b) != 0
+ !ether_addr_equal(a, b)
@@
expression a,b;
@@
- ether_addr_equal(a, b) == 0
+ !ether_addr_equal(a, b)
@@
expression a,b;
@@
- ether_addr_equal(a, b) != 0
+ ether_addr_equal(a, b)
@@
expression a,b;
@@
- !!ether_addr_equal(a, b)
+ ether_addr_equal(a, b)
Signed-off-by: Joe Perches <joe@perches.com>
Signed-off-by: David S. Miller <davem@davemloft.net>
2012-05-08 22:56:52 +04:00
if ( ether_addr_equal ( orig_addr , sdata - > vif . addr ) )
2011-08-10 03:45:08 +04:00
return ;
2012-06-22 13:29:50 +04:00
mhwmp_dbg ( sdata ,
" received RANN from %pM via neighbour %pM (is_gate=%d) \n " ,
orig_addr , mgmt - > sa , root_is_gate ) ;
2009-11-11 03:01:31 +03:00
rcu_read_lock ( ) ;
2012-03-23 14:48:51 +04:00
sta = sta_info_get ( sdata , mgmt - > sa ) ;
if ( ! sta ) {
rcu_read_unlock ( ) ;
return ;
}
2017-11-14 18:20:05 +03:00
last_hop_metric = airtime_link_metric_get ( local , sta ) ;
new_metric = orig_metric + last_hop_metric ;
if ( new_metric < orig_metric )
new_metric = MAX_METRIC ;
2012-03-23 14:48:51 +04:00
2013-02-15 17:40:31 +04:00
mpath = mesh_path_lookup ( sdata , orig_addr ) ;
2009-11-11 03:01:31 +03:00
if ( ! mpath ) {
2013-03-29 17:38:39 +04:00
mpath = mesh_path_add ( sdata , orig_addr ) ;
if ( IS_ERR ( mpath ) ) {
2009-11-11 03:01:31 +03:00
rcu_read_unlock ( ) ;
sdata - > u . mesh . mshstats . dropped_frames_no_route + + ;
return ;
}
}
2011-08-10 03:45:08 +04:00
2012-06-15 06:20:02 +04:00
if ( ! ( SN_LT ( mpath - > sn , orig_sn ) ) & &
2017-11-14 18:20:05 +03:00
! ( mpath - > sn = = orig_sn & & new_metric < mpath - > rann_metric ) ) {
2012-06-15 06:20:02 +04:00
rcu_read_unlock ( ) ;
return ;
}
2011-08-10 03:45:08 +04:00
if ( ( ! ( mpath - > flags & ( MESH_PATH_ACTIVE | MESH_PATH_RESOLVING ) ) | |
2012-06-13 22:06:10 +04:00
( time_after ( jiffies , mpath - > last_preq_to_root +
root_path_confirmation_jiffies ( sdata ) ) | |
time_before ( jiffies , mpath - > last_preq_to_root ) ) ) & &
2012-06-15 06:20:02 +04:00
! ( mpath - > flags & MESH_PATH_FIXED ) & & ( ttl ! = 0 ) ) {
2012-06-22 13:29:50 +04:00
mhwmp_dbg ( sdata ,
" time to refresh root mpath %pM \n " ,
orig_addr ) ;
2011-08-10 03:45:08 +04:00
mesh_queue_preq ( mpath , PREQ_Q_F_START | PREQ_Q_F_REFRESH ) ;
2012-06-13 22:06:10 +04:00
mpath - > last_preq_to_root = jiffies ;
2011-08-10 03:45:08 +04:00
}
2012-06-15 06:20:02 +04:00
mpath - > sn = orig_sn ;
2017-11-14 18:20:05 +03:00
mpath - > rann_metric = new_metric ;
2012-06-15 06:20:02 +04:00
mpath - > is_root = true ;
/* Recording RANNs sender address to send individually
* addressed PREQs destined for root mesh STA */
memcpy ( mpath - > rann_snd_addr , mgmt - > sa , ETH_ALEN ) ;
if ( root_is_gate )
mesh_path_add_gate ( mpath ) ;
if ( ttl < = 1 ) {
ifmsh - > mshstats . dropped_frames_ttl + + ;
rcu_read_unlock ( ) ;
return ;
}
ttl - - ;
if ( ifmsh - > mshcfg . dot11MeshForwarding ) {
2009-11-11 03:01:31 +03:00
mesh_path_sel_frame_tx ( MPATH_RANN , flags , orig_addr ,
2013-11-13 11:39:12 +04:00
orig_sn , 0 , NULL , 0 , broadcast_addr ,
hopcount , ttl , interval ,
2017-11-14 18:20:05 +03:00
new_metric , 0 , sdata ) ;
2009-11-11 03:01:31 +03:00
}
2012-02-28 18:00:06 +04:00
2009-11-11 03:01:31 +03:00
rcu_read_unlock ( ) ;
}
2008-02-23 17:17:15 +03:00
2008-08-03 04:04:37 +04:00
void mesh_rx_path_sel_frame ( struct ieee80211_sub_if_data * sdata ,
2013-02-15 17:40:31 +04:00
struct ieee80211_mgmt * mgmt , size_t len )
2008-02-23 17:17:15 +03:00
{
mac80211: always allocate struct ieee802_11_elems
As the 802.11 spec evolves, we need to parse more and more
elements. This is causing the struct to grow, and we can no
longer get away with putting it on the stack.
Change the API to always dynamically allocate and return an
allocated pointer that must be kfree()d later.
As an alternative, I contemplated a scheme whereby we'd say
in the code which elements we needed, e.g.
DECLARE_ELEMENT_PARSER(elems,
SUPPORTED_CHANNELS,
CHANNEL_SWITCH,
EXT(KEY_DELIVERY));
ieee802_11_parse_elems(..., &elems, ...);
and while I think this is possible and will save us a lot
since most individual places only care about a small subset
of the elements, it ended up being a bit more work since a
lot of places do the parsing and then pass the struct to
other functions, sometimes with multiple levels.
Link: https://lore.kernel.org/r/20210920154009.26caff6b5998.I05ae58768e990e611aee8eca8abefd9d7bc15e05@changeid
Signed-off-by: Johannes Berg <johannes.berg@intel.com>
2021-09-20 16:40:10 +03:00
struct ieee802_11_elems * elems ;
2008-02-23 17:17:15 +03:00
size_t baselen ;
2015-06-05 22:43:54 +03:00
u32 path_metric ;
2011-10-07 01:54:22 +04:00
struct sta_info * sta ;
2008-02-23 17:17:15 +03:00
2008-09-08 17:41:59 +04:00
/* need action_code */
if ( len < IEEE80211_MIN_ACTION_SIZE + 1 )
return ;
2011-10-07 01:54:22 +04:00
rcu_read_lock ( ) ;
sta = sta_info_get ( sdata , mgmt - > sa ) ;
2015-06-17 11:31:00 +03:00
if ( ! sta | | sta - > mesh - > plink_state ! = NL80211_PLINK_ESTAB ) {
2011-10-07 01:54:22 +04:00
rcu_read_unlock ( ) ;
return ;
}
rcu_read_unlock ( ) ;
2008-02-23 17:17:15 +03:00
baselen = ( u8 * ) mgmt - > u . action . u . mesh_action . variable - ( u8 * ) mgmt ;
mac80211: always allocate struct ieee802_11_elems
As the 802.11 spec evolves, we need to parse more and more
elements. This is causing the struct to grow, and we can no
longer get away with putting it on the stack.
Change the API to always dynamically allocate and return an
allocated pointer that must be kfree()d later.
As an alternative, I contemplated a scheme whereby we'd say
in the code which elements we needed, e.g.
DECLARE_ELEMENT_PARSER(elems,
SUPPORTED_CHANNELS,
CHANNEL_SWITCH,
EXT(KEY_DELIVERY));
ieee802_11_parse_elems(..., &elems, ...);
and while I think this is possible and will save us a lot
since most individual places only care about a small subset
of the elements, it ended up being a bit more work since a
lot of places do the parsing and then pass the struct to
other functions, sometimes with multiple levels.
Link: https://lore.kernel.org/r/20210920154009.26caff6b5998.I05ae58768e990e611aee8eca8abefd9d7bc15e05@changeid
Signed-off-by: Johannes Berg <johannes.berg@intel.com>
2021-09-20 16:40:10 +03:00
elems = ieee802_11_parse_elems ( mgmt - > u . action . u . mesh_action . variable ,
2022-06-29 14:29:05 +03:00
len - baselen , false , NULL ) ;
mac80211: always allocate struct ieee802_11_elems
As the 802.11 spec evolves, we need to parse more and more
elements. This is causing the struct to grow, and we can no
longer get away with putting it on the stack.
Change the API to always dynamically allocate and return an
allocated pointer that must be kfree()d later.
As an alternative, I contemplated a scheme whereby we'd say
in the code which elements we needed, e.g.
DECLARE_ELEMENT_PARSER(elems,
SUPPORTED_CHANNELS,
CHANNEL_SWITCH,
EXT(KEY_DELIVERY));
ieee802_11_parse_elems(..., &elems, ...);
and while I think this is possible and will save us a lot
since most individual places only care about a small subset
of the elements, it ended up being a bit more work since a
lot of places do the parsing and then pass the struct to
other functions, sometimes with multiple levels.
Link: https://lore.kernel.org/r/20210920154009.26caff6b5998.I05ae58768e990e611aee8eca8abefd9d7bc15e05@changeid
Signed-off-by: Johannes Berg <johannes.berg@intel.com>
2021-09-20 16:40:10 +03:00
if ( ! elems )
return ;
2008-02-23 17:17:15 +03:00
mac80211: always allocate struct ieee802_11_elems
As the 802.11 spec evolves, we need to parse more and more
elements. This is causing the struct to grow, and we can no
longer get away with putting it on the stack.
Change the API to always dynamically allocate and return an
allocated pointer that must be kfree()d later.
As an alternative, I contemplated a scheme whereby we'd say
in the code which elements we needed, e.g.
DECLARE_ELEMENT_PARSER(elems,
SUPPORTED_CHANNELS,
CHANNEL_SWITCH,
EXT(KEY_DELIVERY));
ieee802_11_parse_elems(..., &elems, ...);
and while I think this is possible and will save us a lot
since most individual places only care about a small subset
of the elements, it ended up being a bit more work since a
lot of places do the parsing and then pass the struct to
other functions, sometimes with multiple levels.
Link: https://lore.kernel.org/r/20210920154009.26caff6b5998.I05ae58768e990e611aee8eca8abefd9d7bc15e05@changeid
Signed-off-by: Johannes Berg <johannes.berg@intel.com>
2021-09-20 16:40:10 +03:00
if ( elems - > preq ) {
if ( elems - > preq_len ! = 37 )
2008-02-23 17:17:15 +03:00
/* Right now we support just 1 destination and no AE */
mac80211: always allocate struct ieee802_11_elems
As the 802.11 spec evolves, we need to parse more and more
elements. This is causing the struct to grow, and we can no
longer get away with putting it on the stack.
Change the API to always dynamically allocate and return an
allocated pointer that must be kfree()d later.
As an alternative, I contemplated a scheme whereby we'd say
in the code which elements we needed, e.g.
DECLARE_ELEMENT_PARSER(elems,
SUPPORTED_CHANNELS,
CHANNEL_SWITCH,
EXT(KEY_DELIVERY));
ieee802_11_parse_elems(..., &elems, ...);
and while I think this is possible and will save us a lot
since most individual places only care about a small subset
of the elements, it ended up being a bit more work since a
lot of places do the parsing and then pass the struct to
other functions, sometimes with multiple levels.
Link: https://lore.kernel.org/r/20210920154009.26caff6b5998.I05ae58768e990e611aee8eca8abefd9d7bc15e05@changeid
Signed-off-by: Johannes Berg <johannes.berg@intel.com>
2021-09-20 16:40:10 +03:00
goto free ;
path_metric = hwmp_route_info_get ( sdata , mgmt , elems - > preq ,
2015-06-05 22:43:54 +03:00
MPATH_PREQ ) ;
if ( path_metric )
mac80211: always allocate struct ieee802_11_elems
As the 802.11 spec evolves, we need to parse more and more
elements. This is causing the struct to grow, and we can no
longer get away with putting it on the stack.
Change the API to always dynamically allocate and return an
allocated pointer that must be kfree()d later.
As an alternative, I contemplated a scheme whereby we'd say
in the code which elements we needed, e.g.
DECLARE_ELEMENT_PARSER(elems,
SUPPORTED_CHANNELS,
CHANNEL_SWITCH,
EXT(KEY_DELIVERY));
ieee802_11_parse_elems(..., &elems, ...);
and while I think this is possible and will save us a lot
since most individual places only care about a small subset
of the elements, it ended up being a bit more work since a
lot of places do the parsing and then pass the struct to
other functions, sometimes with multiple levels.
Link: https://lore.kernel.org/r/20210920154009.26caff6b5998.I05ae58768e990e611aee8eca8abefd9d7bc15e05@changeid
Signed-off-by: Johannes Berg <johannes.berg@intel.com>
2021-09-20 16:40:10 +03:00
hwmp_preq_frame_process ( sdata , mgmt , elems - > preq ,
2015-06-05 22:43:54 +03:00
path_metric ) ;
2009-11-10 02:46:46 +03:00
}
mac80211: always allocate struct ieee802_11_elems
As the 802.11 spec evolves, we need to parse more and more
elements. This is causing the struct to grow, and we can no
longer get away with putting it on the stack.
Change the API to always dynamically allocate and return an
allocated pointer that must be kfree()d later.
As an alternative, I contemplated a scheme whereby we'd say
in the code which elements we needed, e.g.
DECLARE_ELEMENT_PARSER(elems,
SUPPORTED_CHANNELS,
CHANNEL_SWITCH,
EXT(KEY_DELIVERY));
ieee802_11_parse_elems(..., &elems, ...);
and while I think this is possible and will save us a lot
since most individual places only care about a small subset
of the elements, it ended up being a bit more work since a
lot of places do the parsing and then pass the struct to
other functions, sometimes with multiple levels.
Link: https://lore.kernel.org/r/20210920154009.26caff6b5998.I05ae58768e990e611aee8eca8abefd9d7bc15e05@changeid
Signed-off-by: Johannes Berg <johannes.berg@intel.com>
2021-09-20 16:40:10 +03:00
if ( elems - > prep ) {
if ( elems - > prep_len ! = 31 )
2008-02-23 17:17:15 +03:00
/* Right now we support no AE */
mac80211: always allocate struct ieee802_11_elems
As the 802.11 spec evolves, we need to parse more and more
elements. This is causing the struct to grow, and we can no
longer get away with putting it on the stack.
Change the API to always dynamically allocate and return an
allocated pointer that must be kfree()d later.
As an alternative, I contemplated a scheme whereby we'd say
in the code which elements we needed, e.g.
DECLARE_ELEMENT_PARSER(elems,
SUPPORTED_CHANNELS,
CHANNEL_SWITCH,
EXT(KEY_DELIVERY));
ieee802_11_parse_elems(..., &elems, ...);
and while I think this is possible and will save us a lot
since most individual places only care about a small subset
of the elements, it ended up being a bit more work since a
lot of places do the parsing and then pass the struct to
other functions, sometimes with multiple levels.
Link: https://lore.kernel.org/r/20210920154009.26caff6b5998.I05ae58768e990e611aee8eca8abefd9d7bc15e05@changeid
Signed-off-by: Johannes Berg <johannes.berg@intel.com>
2021-09-20 16:40:10 +03:00
goto free ;
path_metric = hwmp_route_info_get ( sdata , mgmt , elems - > prep ,
2015-06-05 22:43:54 +03:00
MPATH_PREP ) ;
if ( path_metric )
mac80211: always allocate struct ieee802_11_elems
As the 802.11 spec evolves, we need to parse more and more
elements. This is causing the struct to grow, and we can no
longer get away with putting it on the stack.
Change the API to always dynamically allocate and return an
allocated pointer that must be kfree()d later.
As an alternative, I contemplated a scheme whereby we'd say
in the code which elements we needed, e.g.
DECLARE_ELEMENT_PARSER(elems,
SUPPORTED_CHANNELS,
CHANNEL_SWITCH,
EXT(KEY_DELIVERY));
ieee802_11_parse_elems(..., &elems, ...);
and while I think this is possible and will save us a lot
since most individual places only care about a small subset
of the elements, it ended up being a bit more work since a
lot of places do the parsing and then pass the struct to
other functions, sometimes with multiple levels.
Link: https://lore.kernel.org/r/20210920154009.26caff6b5998.I05ae58768e990e611aee8eca8abefd9d7bc15e05@changeid
Signed-off-by: Johannes Berg <johannes.berg@intel.com>
2021-09-20 16:40:10 +03:00
hwmp_prep_frame_process ( sdata , mgmt , elems - > prep ,
2015-06-05 22:43:54 +03:00
path_metric ) ;
2009-11-10 02:46:46 +03:00
}
mac80211: always allocate struct ieee802_11_elems
As the 802.11 spec evolves, we need to parse more and more
elements. This is causing the struct to grow, and we can no
longer get away with putting it on the stack.
Change the API to always dynamically allocate and return an
allocated pointer that must be kfree()d later.
As an alternative, I contemplated a scheme whereby we'd say
in the code which elements we needed, e.g.
DECLARE_ELEMENT_PARSER(elems,
SUPPORTED_CHANNELS,
CHANNEL_SWITCH,
EXT(KEY_DELIVERY));
ieee802_11_parse_elems(..., &elems, ...);
and while I think this is possible and will save us a lot
since most individual places only care about a small subset
of the elements, it ended up being a bit more work since a
lot of places do the parsing and then pass the struct to
other functions, sometimes with multiple levels.
Link: https://lore.kernel.org/r/20210920154009.26caff6b5998.I05ae58768e990e611aee8eca8abefd9d7bc15e05@changeid
Signed-off-by: Johannes Berg <johannes.berg@intel.com>
2021-09-20 16:40:10 +03:00
if ( elems - > perr ) {
if ( elems - > perr_len ! = 15 )
2008-02-23 17:17:15 +03:00
/* Right now we support only one destination per PERR */
mac80211: always allocate struct ieee802_11_elems
As the 802.11 spec evolves, we need to parse more and more
elements. This is causing the struct to grow, and we can no
longer get away with putting it on the stack.
Change the API to always dynamically allocate and return an
allocated pointer that must be kfree()d later.
As an alternative, I contemplated a scheme whereby we'd say
in the code which elements we needed, e.g.
DECLARE_ELEMENT_PARSER(elems,
SUPPORTED_CHANNELS,
CHANNEL_SWITCH,
EXT(KEY_DELIVERY));
ieee802_11_parse_elems(..., &elems, ...);
and while I think this is possible and will save us a lot
since most individual places only care about a small subset
of the elements, it ended up being a bit more work since a
lot of places do the parsing and then pass the struct to
other functions, sometimes with multiple levels.
Link: https://lore.kernel.org/r/20210920154009.26caff6b5998.I05ae58768e990e611aee8eca8abefd9d7bc15e05@changeid
Signed-off-by: Johannes Berg <johannes.berg@intel.com>
2021-09-20 16:40:10 +03:00
goto free ;
hwmp_perr_frame_process ( sdata , mgmt , elems - > perr ) ;
2008-02-23 17:17:15 +03:00
}
mac80211: always allocate struct ieee802_11_elems
As the 802.11 spec evolves, we need to parse more and more
elements. This is causing the struct to grow, and we can no
longer get away with putting it on the stack.
Change the API to always dynamically allocate and return an
allocated pointer that must be kfree()d later.
As an alternative, I contemplated a scheme whereby we'd say
in the code which elements we needed, e.g.
DECLARE_ELEMENT_PARSER(elems,
SUPPORTED_CHANNELS,
CHANNEL_SWITCH,
EXT(KEY_DELIVERY));
ieee802_11_parse_elems(..., &elems, ...);
and while I think this is possible and will save us a lot
since most individual places only care about a small subset
of the elements, it ended up being a bit more work since a
lot of places do the parsing and then pass the struct to
other functions, sometimes with multiple levels.
Link: https://lore.kernel.org/r/20210920154009.26caff6b5998.I05ae58768e990e611aee8eca8abefd9d7bc15e05@changeid
Signed-off-by: Johannes Berg <johannes.berg@intel.com>
2021-09-20 16:40:10 +03:00
if ( elems - > rann )
hwmp_rann_frame_process ( sdata , mgmt , elems - > rann ) ;
free :
kfree ( elems ) ;
2008-02-23 17:17:15 +03:00
}
/**
* mesh_queue_preq - queue a PREQ to a given destination
*
* @ mpath : mesh path to discover
* @ flags : special attributes of the PREQ to be sent
*
* Locking : the function must be called from within a rcu read lock block .
*
*/
static void mesh_queue_preq ( struct mesh_path * mpath , u8 flags )
{
2008-08-03 04:04:37 +04:00
struct ieee80211_sub_if_data * sdata = mpath - > sdata ;
2008-09-11 02:01:49 +04:00
struct ieee80211_if_mesh * ifmsh = & sdata - > u . mesh ;
2008-02-23 17:17:15 +03:00
struct mesh_preq_queue * preq_node ;
2009-06-26 03:07:42 +04:00
preq_node = kmalloc ( sizeof ( struct mesh_preq_queue ) , GFP_ATOMIC ) ;
2008-02-23 17:17:15 +03:00
if ( ! preq_node ) {
2012-06-22 13:29:50 +04:00
mhwmp_dbg ( sdata , " could not allocate PREQ node \n " ) ;
2008-02-23 17:17:15 +03:00
return ;
}
2011-07-28 09:51:05 +04:00
spin_lock_bh ( & ifmsh - > mesh_preq_queue_lock ) ;
2008-09-11 02:01:49 +04:00
if ( ifmsh - > preq_queue_len = = MAX_PREQ_QUEUE_LEN ) {
2011-07-28 09:51:05 +04:00
spin_unlock_bh ( & ifmsh - > mesh_preq_queue_lock ) ;
2008-02-23 17:17:15 +03:00
kfree ( preq_node ) ;
if ( printk_ratelimit ( ) )
2012-06-22 13:29:50 +04:00
mhwmp_dbg ( sdata , " PREQ node queue full \n " ) ;
2008-02-23 17:17:15 +03:00
return ;
}
2011-11-18 18:27:31 +04:00
spin_lock ( & mpath - > state_lock ) ;
2011-11-04 08:11:10 +04:00
if ( mpath - > flags & MESH_PATH_REQ_QUEUED ) {
2011-11-18 18:27:31 +04:00
spin_unlock ( & mpath - > state_lock ) ;
2011-11-04 08:11:10 +04:00
spin_unlock_bh ( & ifmsh - > mesh_preq_queue_lock ) ;
2011-11-15 10:33:31 +04:00
kfree ( preq_node ) ;
2011-11-04 08:11:10 +04:00
return ;
}
2008-02-23 17:17:15 +03:00
memcpy ( preq_node - > dst , mpath - > dst , ETH_ALEN ) ;
preq_node - > flags = flags ;
2011-11-04 08:11:10 +04:00
mpath - > flags | = MESH_PATH_REQ_QUEUED ;
2011-11-18 18:27:31 +04:00
spin_unlock ( & mpath - > state_lock ) ;
2011-11-04 08:11:10 +04:00
2008-09-11 02:01:49 +04:00
list_add_tail ( & preq_node - > list , & ifmsh - > preq_queue . list ) ;
+ + ifmsh - > preq_queue_len ;
2011-07-28 09:51:05 +04:00
spin_unlock_bh ( & ifmsh - > mesh_preq_queue_lock ) ;
2008-02-23 17:17:15 +03:00
2008-09-11 02:01:49 +04:00
if ( time_after ( jiffies , ifmsh - > last_preq + min_preq_int_jiff ( sdata ) ) )
2023-06-06 15:49:26 +03:00
wiphy_work_queue ( sdata - > local - > hw . wiphy , & sdata - > work ) ;
2008-02-23 17:17:15 +03:00
2008-09-11 02:01:49 +04:00
else if ( time_before ( jiffies , ifmsh - > last_preq ) ) {
2008-02-23 17:17:15 +03:00
/* avoid long wait if did not send preqs for a long time
* and jiffies wrapped around
*/
2008-09-11 02:01:49 +04:00
ifmsh - > last_preq = jiffies - min_preq_int_jiff ( sdata ) - 1 ;
2023-06-06 15:49:26 +03:00
wiphy_work_queue ( sdata - > local - > hw . wiphy , & sdata - > work ) ;
2008-02-23 17:17:15 +03:00
} else
2008-09-11 02:01:49 +04:00
mod_timer ( & ifmsh - > mesh_path_timer , ifmsh - > last_preq +
2008-02-23 17:17:15 +03:00
min_preq_int_jiff ( sdata ) ) ;
}
/**
* mesh_path_start_discovery - launch a path discovery from the PREQ queue
*
2008-08-03 04:04:37 +04:00
* @ sdata : local mesh subif
2008-02-23 17:17:15 +03:00
*/
2008-08-03 04:04:37 +04:00
void mesh_path_start_discovery ( struct ieee80211_sub_if_data * sdata )
2008-02-23 17:17:15 +03:00
{
2008-09-11 02:01:49 +04:00
struct ieee80211_if_mesh * ifmsh = & sdata - > u . mesh ;
2008-02-23 17:17:15 +03:00
struct mesh_preq_queue * preq_node ;
struct mesh_path * mpath ;
2015-06-15 06:58:53 +03:00
u8 ttl , target_flags = 0 ;
2012-02-28 18:00:06 +04:00
const u8 * da ;
2008-02-23 17:17:15 +03:00
u32 lifetime ;
2009-07-10 13:39:26 +04:00
spin_lock_bh ( & ifmsh - > mesh_preq_queue_lock ) ;
2008-09-11 02:01:49 +04:00
if ( ! ifmsh - > preq_queue_len | |
time_before ( jiffies , ifmsh - > last_preq +
2008-02-23 17:17:15 +03:00
min_preq_int_jiff ( sdata ) ) ) {
2009-07-10 13:39:26 +04:00
spin_unlock_bh ( & ifmsh - > mesh_preq_queue_lock ) ;
2008-02-23 17:17:15 +03:00
return ;
}
2008-09-11 02:01:49 +04:00
preq_node = list_first_entry ( & ifmsh - > preq_queue . list ,
2008-02-23 17:17:15 +03:00
struct mesh_preq_queue , list ) ;
list_del ( & preq_node - > list ) ;
2008-09-11 02:01:49 +04:00
- - ifmsh - > preq_queue_len ;
2009-07-10 13:39:26 +04:00
spin_unlock_bh ( & ifmsh - > mesh_preq_queue_lock ) ;
2008-02-23 17:17:15 +03:00
rcu_read_lock ( ) ;
2013-02-15 17:40:31 +04:00
mpath = mesh_path_lookup ( sdata , preq_node - > dst ) ;
2008-02-23 17:17:15 +03:00
if ( ! mpath )
goto enddiscovery ;
spin_lock_bh ( & mpath - > state_lock ) ;
2016-09-06 21:59:00 +03:00
if ( mpath - > flags & ( MESH_PATH_DELETED | MESH_PATH_FIXED ) ) {
2016-03-19 05:03:24 +03:00
spin_unlock_bh ( & mpath - > state_lock ) ;
goto enddiscovery ;
}
2011-11-04 08:11:10 +04:00
mpath - > flags & = ~ MESH_PATH_REQ_QUEUED ;
2008-02-23 17:17:15 +03:00
if ( preq_node - > flags & PREQ_Q_F_START ) {
if ( mpath - > flags & MESH_PATH_RESOLVING ) {
spin_unlock_bh ( & mpath - > state_lock ) ;
goto enddiscovery ;
} else {
mpath - > flags & = ~ MESH_PATH_RESOLVED ;
mpath - > flags | = MESH_PATH_RESOLVING ;
mpath - > discovery_retries = 0 ;
mpath - > discovery_timeout = disc_timeout_jiff ( sdata ) ;
}
} else if ( ! ( mpath - > flags & MESH_PATH_RESOLVING ) | |
mpath - > flags & MESH_PATH_RESOLVED ) {
mpath - > flags & = ~ MESH_PATH_RESOLVING ;
spin_unlock_bh ( & mpath - > state_lock ) ;
goto enddiscovery ;
}
2008-09-11 02:01:49 +04:00
ifmsh - > last_preq = jiffies ;
2008-02-23 17:17:15 +03:00
2009-11-10 02:46:55 +03:00
if ( time_after ( jiffies , ifmsh - > last_sn_update +
2008-02-23 17:17:15 +03:00
net_traversal_jiffies ( sdata ) ) | |
2009-11-10 02:46:55 +03:00
time_before ( jiffies , ifmsh - > last_sn_update ) ) {
+ + ifmsh - > sn ;
sdata - > u . mesh . last_sn_update = jiffies ;
2008-02-23 17:17:15 +03:00
}
lifetime = default_lifetime ( sdata ) ;
2010-12-03 11:20:40 +03:00
ttl = sdata - > u . mesh . mshcfg . element_ttl ;
2008-02-23 17:17:15 +03:00
if ( ttl = = 0 ) {
2008-09-11 02:01:49 +04:00
sdata - > u . mesh . mshstats . dropped_frames_ttl + + ;
2008-02-23 17:17:15 +03:00
spin_unlock_bh ( & mpath - > state_lock ) ;
goto enddiscovery ;
}
if ( preq_node - > flags & PREQ_Q_F_REFRESH )
2015-06-15 06:58:53 +03:00
target_flags | = IEEE80211_PREQ_TO_FLAG ;
2008-02-23 17:17:15 +03:00
else
2015-06-15 06:58:53 +03:00
target_flags & = ~ IEEE80211_PREQ_TO_FLAG ;
2008-02-23 17:17:15 +03:00
spin_unlock_bh ( & mpath - > state_lock ) ;
2012-02-28 18:00:06 +04:00
da = ( mpath - > is_root ) ? mpath - > rann_snd_addr : broadcast_addr ;
2013-11-13 11:39:12 +04:00
mesh_path_sel_frame_tx ( MPATH_PREQ , 0 , sdata - > vif . addr , ifmsh - > sn ,
target_flags , mpath - > dst , mpath - > sn , da , 0 ,
ttl , lifetime , 0 , ifmsh - > preq_id + + , sdata ) ;
mac80211: mesh: fix discovery timer re-arming issue / crash
On a non-forwarding 802.11s link between two fairly busy
neighboring nodes (iperf with -P 16 at ~850MBit/s TCP;
1733.3 MBit/s VHT-MCS 9 80MHz short GI VHT-NSS 4), so with
frequent PREQ retries, usually after around 30-40 seconds the
following crash would occur:
[ 1110.822428] Unable to handle kernel read from unreadable memory at virtual address 00000000
[ 1110.830786] Mem abort info:
[ 1110.833573] Exception class = IABT (current EL), IL = 32 bits
[ 1110.839494] SET = 0, FnV = 0
[ 1110.842546] EA = 0, S1PTW = 0
[ 1110.845678] user pgtable: 4k pages, 48-bit VAs, pgd = ffff800076386000
[ 1110.852204] [0000000000000000] *pgd=00000000f6322003, *pud=00000000f62de003, *pmd=0000000000000000
[ 1110.861167] Internal error: Oops: 86000004 [#1] PREEMPT SMP
[ 1110.866730] Modules linked in: pppoe ppp_async batman_adv ath10k_pci ath10k_core ath pppox ppp_generic nf_conntrack_ipv6 mac80211 iptable_nat ipt_REJECT ipt_MASQUERADE cfg80211 xt_time xt_tcpudp xt_state xt_nat xt_multiport xt_mark xt_mac xt_limit xt_conntrack xt_comment xt_TCPMSS xt_REDIRECT xt_LOG xt_FLOWOFFLOAD slhc nf_reject_ipv4 nf_nat_redirect nf_nat_masquerade_ipv4 nf_conntrack_ipv4 nf_nat_ipv4 nf_nat nf_log_ipv4 nf_flow_table_hw nf_flow_table nf_defrag_ipv6 nf_defrag_ipv4 nf_conntrack_rtcache nf_conntrack iptable_mangle iptable_filter ip_tables crc_ccitt compat nf_log_ipv6 nf_log_common ip6table_mangle ip6table_filter ip6_tables ip6t_REJECT x_tables nf_reject_ipv6 usb_storage xhci_plat_hcd xhci_pci xhci_hcd dwc3 usbcore usb_common
[ 1110.932190] Process swapper/3 (pid: 0, stack limit = 0xffff0000090c8000)
[ 1110.938884] CPU: 3 PID: 0 Comm: swapper/3 Not tainted 4.14.162 #0
[ 1110.944965] Hardware name: LS1043A RGW Board (DT)
[ 1110.949658] task: ffff8000787a81c0 task.stack: ffff0000090c8000
[ 1110.955568] PC is at 0x0
[ 1110.958097] LR is at call_timer_fn.isra.27+0x24/0x78
[ 1110.963055] pc : [<0000000000000000>] lr : [<ffff0000080ff29c>] pstate: 00400145
[ 1110.970440] sp : ffff00000801be10
[ 1110.973744] x29: ffff00000801be10 x28: ffff000008bf7018
[ 1110.979047] x27: ffff000008bf87c8 x26: ffff000008c160c0
[ 1110.984352] x25: 0000000000000000 x24: 0000000000000000
[ 1110.989657] x23: dead000000000200 x22: 0000000000000000
[ 1110.994959] x21: 0000000000000000 x20: 0000000000000101
[ 1111.000262] x19: ffff8000787a81c0 x18: 0000000000000000
[ 1111.005565] x17: ffff0000089167b0 x16: 0000000000000058
[ 1111.010868] x15: ffff0000089167b0 x14: 0000000000000000
[ 1111.016172] x13: ffff000008916788 x12: 0000000000000040
[ 1111.021475] x11: ffff80007fda9af0 x10: 0000000000000001
[ 1111.026777] x9 : ffff00000801bea0 x8 : 0000000000000004
[ 1111.032080] x7 : 0000000000000000 x6 : ffff80007fda9aa8
[ 1111.037383] x5 : ffff00000801bea0 x4 : 0000000000000010
[ 1111.042685] x3 : ffff00000801be98 x2 : 0000000000000614
[ 1111.047988] x1 : 0000000000000000 x0 : 0000000000000000
[ 1111.053290] Call trace:
[ 1111.055728] Exception stack(0xffff00000801bcd0 to 0xffff00000801be10)
[ 1111.062158] bcc0: 0000000000000000 0000000000000000
[ 1111.069978] bce0: 0000000000000614 ffff00000801be98 0000000000000010 ffff00000801bea0
[ 1111.077798] bd00: ffff80007fda9aa8 0000000000000000 0000000000000004 ffff00000801bea0
[ 1111.085618] bd20: 0000000000000001 ffff80007fda9af0 0000000000000040 ffff000008916788
[ 1111.093437] bd40: 0000000000000000 ffff0000089167b0 0000000000000058 ffff0000089167b0
[ 1111.101256] bd60: 0000000000000000 ffff8000787a81c0 0000000000000101 0000000000000000
[ 1111.109075] bd80: 0000000000000000 dead000000000200 0000000000000000 0000000000000000
[ 1111.116895] bda0: ffff000008c160c0 ffff000008bf87c8 ffff000008bf7018 ffff00000801be10
[ 1111.124715] bdc0: ffff0000080ff29c ffff00000801be10 0000000000000000 0000000000400145
[ 1111.132534] bde0: ffff8000787a81c0 ffff00000801bde8 0000ffffffffffff 000001029eb19be8
[ 1111.140353] be00: ffff00000801be10 0000000000000000
[ 1111.145220] [< (null)>] (null)
[ 1111.149917] [<ffff0000080ff77c>] run_timer_softirq+0x184/0x398
[ 1111.155741] [<ffff000008081938>] __do_softirq+0x100/0x1fc
[ 1111.161130] [<ffff0000080a2e28>] irq_exit+0x80/0xd8
[ 1111.166002] [<ffff0000080ea708>] __handle_domain_irq+0x88/0xb0
[ 1111.171825] [<ffff000008081678>] gic_handle_irq+0x68/0xb0
[ 1111.177213] Exception stack(0xffff0000090cbe30 to 0xffff0000090cbf70)
[ 1111.183642] be20: 0000000000000020 0000000000000000
[ 1111.191461] be40: 0000000000000001 0000000000000000 00008000771af000 0000000000000000
[ 1111.199281] be60: ffff000008c95180 0000000000000000 ffff000008c19360 ffff0000090cbef0
[ 1111.207101] be80: 0000000000000810 0000000000000400 0000000000000098 ffff000000000000
[ 1111.214920] bea0: 0000000000000001 ffff0000089167b0 0000000000000000 ffff0000089167b0
[ 1111.222740] bec0: 0000000000000000 ffff000008c198e8 ffff000008bf7018 ffff000008c19000
[ 1111.230559] bee0: 0000000000000000 0000000000000000 ffff8000787a81c0 ffff000008018000
[ 1111.238380] bf00: ffff00000801c000 ffff00000913ba34 ffff8000787a81c0 ffff0000090cbf70
[ 1111.246199] bf20: ffff0000080857cc ffff0000090cbf70 ffff0000080857d0 0000000000400145
[ 1111.254020] bf40: ffff000008018000 ffff00000801c000 ffffffffffffffff ffff0000080fa574
[ 1111.261838] bf60: ffff0000090cbf70 ffff0000080857d0
[ 1111.266706] [<ffff0000080832e8>] el1_irq+0xe8/0x18c
[ 1111.271576] [<ffff0000080857d0>] arch_cpu_idle+0x10/0x18
[ 1111.276880] [<ffff0000080d7de4>] do_idle+0xec/0x1b8
[ 1111.281748] [<ffff0000080d8020>] cpu_startup_entry+0x20/0x28
[ 1111.287399] [<ffff00000808f81c>] secondary_start_kernel+0x104/0x110
[ 1111.293662] Code: bad PC value
[ 1111.296710] ---[ end trace 555b6ca4363c3edd ]---
[ 1111.301318] Kernel panic - not syncing: Fatal exception in interrupt
[ 1111.307661] SMP: stopping secondary CPUs
[ 1111.311574] Kernel Offset: disabled
[ 1111.315053] CPU features: 0x0002000
[ 1111.318530] Memory Limit: none
[ 1111.321575] Rebooting in 3 seconds..
With some added debug output / delays we were able to push the crash from
the timer callback runner into the callback function and by that shedding
some light on which object holding the timer gets corrupted:
[ 401.720899] Unable to handle kernel read from unreadable memory at virtual address 00000868
[...]
[ 402.335836] [<ffff0000088fafa4>] _raw_spin_lock_bh+0x14/0x48
[ 402.341548] [<ffff000000dbe684>] mesh_path_timer+0x10c/0x248 [mac80211]
[ 402.348154] [<ffff0000080ff29c>] call_timer_fn.isra.27+0x24/0x78
[ 402.354150] [<ffff0000080ff77c>] run_timer_softirq+0x184/0x398
[ 402.359974] [<ffff000008081938>] __do_softirq+0x100/0x1fc
[ 402.365362] [<ffff0000080a2e28>] irq_exit+0x80/0xd8
[ 402.370231] [<ffff0000080ea708>] __handle_domain_irq+0x88/0xb0
[ 402.376053] [<ffff000008081678>] gic_handle_irq+0x68/0xb0
The issue happens due to the following sequence of events:
1) mesh_path_start_discovery():
-> spin_unlock_bh(&mpath->state_lock) before mesh_path_sel_frame_tx()
2) mesh_path_free_rcu()
-> del_timer_sync(&mpath->timer)
[...]
-> kfree_rcu(mpath)
3) mesh_path_start_discovery():
-> mod_timer(&mpath->timer, ...)
[...]
-> rcu_read_unlock()
4) mesh_path_free_rcu()'s kfree_rcu():
-> kfree(mpath)
5) mesh_path_timer() starts after timeout, using freed mpath object
So a use-after-free issue due to a timer re-arming bug caused by an
early spin-unlocking.
This patch fixes this issue by re-checking if mpath is about to be
free'd and if so bails out of re-arming the timer.
Cc: stable@vger.kernel.org
Fixes: 050ac52cbe1f ("mac80211: code for on-demand Hybrid Wireless Mesh Protocol")
Cc: Simon Wunderlich <sw@simonwunderlich.de>
Signed-off-by: Linus Lüssing <ll@simonwunderlich.de>
Link: https://lore.kernel.org/r/20200522170413.14973-1-linus.luessing@c0d3.blue
Signed-off-by: Johannes Berg <johannes.berg@intel.com>
2020-05-22 20:04:13 +03:00
spin_lock_bh ( & mpath - > state_lock ) ;
2020-06-05 00:41:57 +03:00
if ( ! ( mpath - > flags & MESH_PATH_DELETED ) )
mod_timer ( & mpath - > timer , jiffies + mpath - > discovery_timeout ) ;
mac80211: mesh: fix discovery timer re-arming issue / crash
On a non-forwarding 802.11s link between two fairly busy
neighboring nodes (iperf with -P 16 at ~850MBit/s TCP;
1733.3 MBit/s VHT-MCS 9 80MHz short GI VHT-NSS 4), so with
frequent PREQ retries, usually after around 30-40 seconds the
following crash would occur:
[ 1110.822428] Unable to handle kernel read from unreadable memory at virtual address 00000000
[ 1110.830786] Mem abort info:
[ 1110.833573] Exception class = IABT (current EL), IL = 32 bits
[ 1110.839494] SET = 0, FnV = 0
[ 1110.842546] EA = 0, S1PTW = 0
[ 1110.845678] user pgtable: 4k pages, 48-bit VAs, pgd = ffff800076386000
[ 1110.852204] [0000000000000000] *pgd=00000000f6322003, *pud=00000000f62de003, *pmd=0000000000000000
[ 1110.861167] Internal error: Oops: 86000004 [#1] PREEMPT SMP
[ 1110.866730] Modules linked in: pppoe ppp_async batman_adv ath10k_pci ath10k_core ath pppox ppp_generic nf_conntrack_ipv6 mac80211 iptable_nat ipt_REJECT ipt_MASQUERADE cfg80211 xt_time xt_tcpudp xt_state xt_nat xt_multiport xt_mark xt_mac xt_limit xt_conntrack xt_comment xt_TCPMSS xt_REDIRECT xt_LOG xt_FLOWOFFLOAD slhc nf_reject_ipv4 nf_nat_redirect nf_nat_masquerade_ipv4 nf_conntrack_ipv4 nf_nat_ipv4 nf_nat nf_log_ipv4 nf_flow_table_hw nf_flow_table nf_defrag_ipv6 nf_defrag_ipv4 nf_conntrack_rtcache nf_conntrack iptable_mangle iptable_filter ip_tables crc_ccitt compat nf_log_ipv6 nf_log_common ip6table_mangle ip6table_filter ip6_tables ip6t_REJECT x_tables nf_reject_ipv6 usb_storage xhci_plat_hcd xhci_pci xhci_hcd dwc3 usbcore usb_common
[ 1110.932190] Process swapper/3 (pid: 0, stack limit = 0xffff0000090c8000)
[ 1110.938884] CPU: 3 PID: 0 Comm: swapper/3 Not tainted 4.14.162 #0
[ 1110.944965] Hardware name: LS1043A RGW Board (DT)
[ 1110.949658] task: ffff8000787a81c0 task.stack: ffff0000090c8000
[ 1110.955568] PC is at 0x0
[ 1110.958097] LR is at call_timer_fn.isra.27+0x24/0x78
[ 1110.963055] pc : [<0000000000000000>] lr : [<ffff0000080ff29c>] pstate: 00400145
[ 1110.970440] sp : ffff00000801be10
[ 1110.973744] x29: ffff00000801be10 x28: ffff000008bf7018
[ 1110.979047] x27: ffff000008bf87c8 x26: ffff000008c160c0
[ 1110.984352] x25: 0000000000000000 x24: 0000000000000000
[ 1110.989657] x23: dead000000000200 x22: 0000000000000000
[ 1110.994959] x21: 0000000000000000 x20: 0000000000000101
[ 1111.000262] x19: ffff8000787a81c0 x18: 0000000000000000
[ 1111.005565] x17: ffff0000089167b0 x16: 0000000000000058
[ 1111.010868] x15: ffff0000089167b0 x14: 0000000000000000
[ 1111.016172] x13: ffff000008916788 x12: 0000000000000040
[ 1111.021475] x11: ffff80007fda9af0 x10: 0000000000000001
[ 1111.026777] x9 : ffff00000801bea0 x8 : 0000000000000004
[ 1111.032080] x7 : 0000000000000000 x6 : ffff80007fda9aa8
[ 1111.037383] x5 : ffff00000801bea0 x4 : 0000000000000010
[ 1111.042685] x3 : ffff00000801be98 x2 : 0000000000000614
[ 1111.047988] x1 : 0000000000000000 x0 : 0000000000000000
[ 1111.053290] Call trace:
[ 1111.055728] Exception stack(0xffff00000801bcd0 to 0xffff00000801be10)
[ 1111.062158] bcc0: 0000000000000000 0000000000000000
[ 1111.069978] bce0: 0000000000000614 ffff00000801be98 0000000000000010 ffff00000801bea0
[ 1111.077798] bd00: ffff80007fda9aa8 0000000000000000 0000000000000004 ffff00000801bea0
[ 1111.085618] bd20: 0000000000000001 ffff80007fda9af0 0000000000000040 ffff000008916788
[ 1111.093437] bd40: 0000000000000000 ffff0000089167b0 0000000000000058 ffff0000089167b0
[ 1111.101256] bd60: 0000000000000000 ffff8000787a81c0 0000000000000101 0000000000000000
[ 1111.109075] bd80: 0000000000000000 dead000000000200 0000000000000000 0000000000000000
[ 1111.116895] bda0: ffff000008c160c0 ffff000008bf87c8 ffff000008bf7018 ffff00000801be10
[ 1111.124715] bdc0: ffff0000080ff29c ffff00000801be10 0000000000000000 0000000000400145
[ 1111.132534] bde0: ffff8000787a81c0 ffff00000801bde8 0000ffffffffffff 000001029eb19be8
[ 1111.140353] be00: ffff00000801be10 0000000000000000
[ 1111.145220] [< (null)>] (null)
[ 1111.149917] [<ffff0000080ff77c>] run_timer_softirq+0x184/0x398
[ 1111.155741] [<ffff000008081938>] __do_softirq+0x100/0x1fc
[ 1111.161130] [<ffff0000080a2e28>] irq_exit+0x80/0xd8
[ 1111.166002] [<ffff0000080ea708>] __handle_domain_irq+0x88/0xb0
[ 1111.171825] [<ffff000008081678>] gic_handle_irq+0x68/0xb0
[ 1111.177213] Exception stack(0xffff0000090cbe30 to 0xffff0000090cbf70)
[ 1111.183642] be20: 0000000000000020 0000000000000000
[ 1111.191461] be40: 0000000000000001 0000000000000000 00008000771af000 0000000000000000
[ 1111.199281] be60: ffff000008c95180 0000000000000000 ffff000008c19360 ffff0000090cbef0
[ 1111.207101] be80: 0000000000000810 0000000000000400 0000000000000098 ffff000000000000
[ 1111.214920] bea0: 0000000000000001 ffff0000089167b0 0000000000000000 ffff0000089167b0
[ 1111.222740] bec0: 0000000000000000 ffff000008c198e8 ffff000008bf7018 ffff000008c19000
[ 1111.230559] bee0: 0000000000000000 0000000000000000 ffff8000787a81c0 ffff000008018000
[ 1111.238380] bf00: ffff00000801c000 ffff00000913ba34 ffff8000787a81c0 ffff0000090cbf70
[ 1111.246199] bf20: ffff0000080857cc ffff0000090cbf70 ffff0000080857d0 0000000000400145
[ 1111.254020] bf40: ffff000008018000 ffff00000801c000 ffffffffffffffff ffff0000080fa574
[ 1111.261838] bf60: ffff0000090cbf70 ffff0000080857d0
[ 1111.266706] [<ffff0000080832e8>] el1_irq+0xe8/0x18c
[ 1111.271576] [<ffff0000080857d0>] arch_cpu_idle+0x10/0x18
[ 1111.276880] [<ffff0000080d7de4>] do_idle+0xec/0x1b8
[ 1111.281748] [<ffff0000080d8020>] cpu_startup_entry+0x20/0x28
[ 1111.287399] [<ffff00000808f81c>] secondary_start_kernel+0x104/0x110
[ 1111.293662] Code: bad PC value
[ 1111.296710] ---[ end trace 555b6ca4363c3edd ]---
[ 1111.301318] Kernel panic - not syncing: Fatal exception in interrupt
[ 1111.307661] SMP: stopping secondary CPUs
[ 1111.311574] Kernel Offset: disabled
[ 1111.315053] CPU features: 0x0002000
[ 1111.318530] Memory Limit: none
[ 1111.321575] Rebooting in 3 seconds..
With some added debug output / delays we were able to push the crash from
the timer callback runner into the callback function and by that shedding
some light on which object holding the timer gets corrupted:
[ 401.720899] Unable to handle kernel read from unreadable memory at virtual address 00000868
[...]
[ 402.335836] [<ffff0000088fafa4>] _raw_spin_lock_bh+0x14/0x48
[ 402.341548] [<ffff000000dbe684>] mesh_path_timer+0x10c/0x248 [mac80211]
[ 402.348154] [<ffff0000080ff29c>] call_timer_fn.isra.27+0x24/0x78
[ 402.354150] [<ffff0000080ff77c>] run_timer_softirq+0x184/0x398
[ 402.359974] [<ffff000008081938>] __do_softirq+0x100/0x1fc
[ 402.365362] [<ffff0000080a2e28>] irq_exit+0x80/0xd8
[ 402.370231] [<ffff0000080ea708>] __handle_domain_irq+0x88/0xb0
[ 402.376053] [<ffff000008081678>] gic_handle_irq+0x68/0xb0
The issue happens due to the following sequence of events:
1) mesh_path_start_discovery():
-> spin_unlock_bh(&mpath->state_lock) before mesh_path_sel_frame_tx()
2) mesh_path_free_rcu()
-> del_timer_sync(&mpath->timer)
[...]
-> kfree_rcu(mpath)
3) mesh_path_start_discovery():
-> mod_timer(&mpath->timer, ...)
[...]
-> rcu_read_unlock()
4) mesh_path_free_rcu()'s kfree_rcu():
-> kfree(mpath)
5) mesh_path_timer() starts after timeout, using freed mpath object
So a use-after-free issue due to a timer re-arming bug caused by an
early spin-unlocking.
This patch fixes this issue by re-checking if mpath is about to be
free'd and if so bails out of re-arming the timer.
Cc: stable@vger.kernel.org
Fixes: 050ac52cbe1f ("mac80211: code for on-demand Hybrid Wireless Mesh Protocol")
Cc: Simon Wunderlich <sw@simonwunderlich.de>
Signed-off-by: Linus Lüssing <ll@simonwunderlich.de>
Link: https://lore.kernel.org/r/20200522170413.14973-1-linus.luessing@c0d3.blue
Signed-off-by: Johannes Berg <johannes.berg@intel.com>
2020-05-22 20:04:13 +03:00
spin_unlock_bh ( & mpath - > state_lock ) ;
2008-02-23 17:17:15 +03:00
enddiscovery :
rcu_read_unlock ( ) ;
kfree ( preq_node ) ;
}
2012-07-10 14:55:09 +04:00
/**
* mesh_nexthop_resolve - lookup next hop ; conditionally start path discovery
2008-02-23 17:17:15 +03:00
*
2008-08-05 21:34:52 +04:00
* @ skb : 802.11 frame to be sent
2008-08-03 04:04:37 +04:00
* @ sdata : network subif the frame will be sent through
2008-02-23 17:17:15 +03:00
*
2012-07-10 14:55:09 +04:00
* Lookup next hop for given skb and start path discovery if no
* forwarding information is found .
*
2011-11-25 05:15:25 +04:00
* Returns : 0 if the next hop was found and - ENOENT if the frame was queued .
2021-06-07 18:00:47 +03:00
* skb is freed here if no mpath could be allocated .
2008-02-23 17:17:15 +03:00
*/
2013-02-15 17:40:31 +04:00
int mesh_nexthop_resolve ( struct ieee80211_sub_if_data * sdata ,
struct sk_buff * skb )
2008-02-23 17:17:15 +03:00
{
2008-08-05 21:34:52 +04:00
struct ieee80211_hdr * hdr = ( struct ieee80211_hdr * ) skb - > data ;
2011-11-25 05:15:25 +04:00
struct ieee80211_tx_info * info = IEEE80211_SKB_CB ( skb ) ;
struct mesh_path * mpath ;
struct sk_buff * skb_to_free = NULL ;
2009-11-10 02:46:55 +03:00
u8 * target_addr = hdr - > addr3 ;
2008-02-23 17:17:15 +03:00
2013-01-30 21:14:08 +04:00
/* Nulls are only sent to peers for PS and should be pre-addressed */
if ( ieee80211_is_qos_nullfunc ( hdr - > frame_control ) )
return 0 ;
2019-04-11 23:47:26 +03:00
/* Allow injected packets to bypass mesh routing */
if ( info - > control . flags & IEEE80211_TX_CTRL_SKIP_MPATH_LOOKUP )
return 0 ;
2019-03-16 20:06:30 +03:00
if ( ! mesh_nexthop_lookup ( sdata , skb ) )
return 0 ;
2008-02-23 17:17:15 +03:00
2011-11-25 05:15:25 +04:00
/* no nexthop found, start resolving */
2013-02-15 17:40:31 +04:00
mpath = mesh_path_lookup ( sdata , target_addr ) ;
2008-02-23 17:17:15 +03:00
if ( ! mpath ) {
2013-03-29 17:38:39 +04:00
mpath = mesh_path_add ( sdata , target_addr ) ;
if ( IS_ERR ( mpath ) ) {
2013-02-15 17:40:31 +04:00
mesh_path_discard_frame ( sdata , skb ) ;
2019-03-16 20:06:30 +03:00
return PTR_ERR ( mpath ) ;
2008-02-23 17:17:15 +03:00
}
}
2020-03-05 17:04:09 +03:00
if ( ! ( mpath - > flags & MESH_PATH_RESOLVING ) & &
mesh_path_sel_is_hwmp ( sdata ) )
2011-11-25 05:15:25 +04:00
mesh_queue_preq ( mpath , PREQ_Q_F_START ) ;
if ( skb_queue_len ( & mpath - > frame_queue ) > = MESH_FRAME_QUEUE_LEN )
skb_to_free = skb_dequeue ( & mpath - > frame_queue ) ;
2020-09-08 15:36:57 +03:00
info - > control . flags | = IEEE80211_TX_INTCFL_NEED_TXPROCESSING ;
2011-11-25 05:15:25 +04:00
ieee80211_set_qos_hdr ( sdata , skb ) ;
skb_queue_tail ( & mpath - > frame_queue , skb ) ;
if ( skb_to_free )
2013-02-15 17:40:31 +04:00
mesh_path_discard_frame ( sdata , skb_to_free ) ;
2011-11-25 05:15:25 +04:00
2019-03-16 20:06:30 +03:00
return - ENOENT ;
2011-11-25 05:15:25 +04:00
}
2013-02-15 17:40:31 +04:00
2020-06-17 10:30:33 +03:00
/**
* mesh_nexthop_lookup_nolearn - try to set next hop without path discovery
* @ skb : 802.11 frame to be sent
* @ sdata : network subif the frame will be sent through
*
* Check if the meshDA ( addr3 ) of a unicast frame is a direct neighbor .
* And if so , set the RA ( addr1 ) to it to transmit to this node directly ,
* avoiding PREQ / PREP path discovery .
*
* Returns : 0 if the next hop was found and - ENOENT otherwise .
*/
static int mesh_nexthop_lookup_nolearn ( struct ieee80211_sub_if_data * sdata ,
struct sk_buff * skb )
{
struct ieee80211_hdr * hdr = ( struct ieee80211_hdr * ) skb - > data ;
struct sta_info * sta ;
if ( is_multicast_ether_addr ( hdr - > addr1 ) )
return - ENOENT ;
rcu_read_lock ( ) ;
sta = sta_info_get ( sdata , hdr - > addr3 ) ;
if ( ! sta | | sta - > mesh - > plink_state ! = NL80211_PLINK_ESTAB ) {
rcu_read_unlock ( ) ;
return - ENOENT ;
}
rcu_read_unlock ( ) ;
memcpy ( hdr - > addr1 , hdr - > addr3 , ETH_ALEN ) ;
memcpy ( hdr - > addr2 , sdata - > vif . addr , ETH_ALEN ) ;
return 0 ;
}
wifi: mac80211: mesh fast xmit support
Previously, fast xmit only worked on interface types where initially a
sta lookup is performed, and a cached header can be attached to the sta,
requiring only some fields to be updated at runtime.
This technique is not directly applicable for a mesh device type due
to the dynamic nature of the topology and protocol. There are more
addresses that need to be filled, and there is an extra header with a
dynamic length based on the addressing mode.
Change the code to cache entries contain a copy of the mesh subframe header +
bridge tunnel header, as well as an embedded struct ieee80211_fast_tx, which
contains the information for building the 802.11 header.
Add a mesh specific early fast xmit call, which looks up a cached entry and
adds only the mesh subframe header, before passing it over to the generic
fast xmit code.
To ensure the changes in network are reflected in these cached headers,
flush affected cached entries on path changes, as well as other conditions
that currently trigger a fast xmit check in other modes (key changes etc.)
This code is loosely based on a previous implementation by:
Sriram R <quic_srirrama@quicinc.com>
Cc: Sriram R <quic_srirrama@quicinc.com>
Signed-off-by: Ryder Lee <ryder.lee@mediatek.com>
Signed-off-by: Felix Fietkau <nbd@nbd.name>
Link: https://lore.kernel.org/r/20230314095956.62085-4-nbd@nbd.name
Signed-off-by: Johannes Berg <johannes.berg@intel.com>
2023-03-14 12:59:53 +03:00
void mesh_path_refresh ( struct ieee80211_sub_if_data * sdata ,
struct mesh_path * mpath , const u8 * addr )
{
if ( mpath - > flags & ( MESH_PATH_REQ_QUEUED | MESH_PATH_FIXED |
MESH_PATH_RESOLVING ) )
return ;
if ( time_after ( jiffies ,
mpath - > exp_time -
msecs_to_jiffies ( sdata - > u . mesh . mshcfg . path_refresh_time ) ) & &
( ! addr | | ether_addr_equal ( sdata - > vif . addr , addr ) ) )
mesh_queue_preq ( mpath , PREQ_Q_F_START | PREQ_Q_F_REFRESH ) ;
}
2011-11-25 05:15:25 +04:00
/**
* mesh_nexthop_lookup - put the appropriate next hop on a mesh frame . Calling
* this function is considered " using " the associated mpath , so preempt a path
* refresh if this mpath expires soon .
*
* @ skb : 802.11 frame to be sent
* @ sdata : network subif the frame will be sent through
*
* Returns : 0 if the next hop was found . Nonzero otherwise .
*/
2013-02-15 17:40:31 +04:00
int mesh_nexthop_lookup ( struct ieee80211_sub_if_data * sdata ,
struct sk_buff * skb )
2011-11-25 05:15:25 +04:00
{
2020-06-17 10:30:33 +03:00
struct ieee80211_if_mesh * ifmsh = & sdata - > u . mesh ;
2011-11-25 05:15:25 +04:00
struct mesh_path * mpath ;
struct sta_info * next_hop ;
struct ieee80211_hdr * hdr = ( struct ieee80211_hdr * ) skb - > data ;
u8 * target_addr = hdr - > addr3 ;
2008-02-23 17:17:15 +03:00
2020-06-17 10:30:33 +03:00
if ( ifmsh - > mshcfg . dot11MeshNolearn & &
! mesh_nexthop_lookup_nolearn ( sdata , skb ) )
return 0 ;
2013-02-15 17:40:31 +04:00
mpath = mesh_path_lookup ( sdata , target_addr ) ;
2011-11-25 05:15:25 +04:00
if ( ! mpath | | ! ( mpath - > flags & MESH_PATH_ACTIVE ) )
2019-03-16 20:06:30 +03:00
return - ENOENT ;
2011-11-25 05:15:25 +04:00
wifi: mac80211: mesh fast xmit support
Previously, fast xmit only worked on interface types where initially a
sta lookup is performed, and a cached header can be attached to the sta,
requiring only some fields to be updated at runtime.
This technique is not directly applicable for a mesh device type due
to the dynamic nature of the topology and protocol. There are more
addresses that need to be filled, and there is an extra header with a
dynamic length based on the addressing mode.
Change the code to cache entries contain a copy of the mesh subframe header +
bridge tunnel header, as well as an embedded struct ieee80211_fast_tx, which
contains the information for building the 802.11 header.
Add a mesh specific early fast xmit call, which looks up a cached entry and
adds only the mesh subframe header, before passing it over to the generic
fast xmit code.
To ensure the changes in network are reflected in these cached headers,
flush affected cached entries on path changes, as well as other conditions
that currently trigger a fast xmit check in other modes (key changes etc.)
This code is loosely based on a previous implementation by:
Sriram R <quic_srirrama@quicinc.com>
Cc: Sriram R <quic_srirrama@quicinc.com>
Signed-off-by: Ryder Lee <ryder.lee@mediatek.com>
Signed-off-by: Felix Fietkau <nbd@nbd.name>
Link: https://lore.kernel.org/r/20230314095956.62085-4-nbd@nbd.name
Signed-off-by: Johannes Berg <johannes.berg@intel.com>
2023-03-14 12:59:53 +03:00
mesh_path_refresh ( sdata , mpath , hdr - > addr4 ) ;
2008-02-23 17:17:15 +03:00
2011-11-25 05:15:25 +04:00
next_hop = rcu_dereference ( mpath - > next_hop ) ;
if ( next_hop ) {
memcpy ( hdr - > addr1 , next_hop - > sta . addr , ETH_ALEN ) ;
memcpy ( hdr - > addr2 , sdata - > vif . addr , ETH_ALEN ) ;
2013-01-30 21:14:08 +04:00
ieee80211_mps_set_frame_flags ( sdata , next_hop , hdr ) ;
wifi: mac80211: mesh fast xmit support
Previously, fast xmit only worked on interface types where initially a
sta lookup is performed, and a cached header can be attached to the sta,
requiring only some fields to be updated at runtime.
This technique is not directly applicable for a mesh device type due
to the dynamic nature of the topology and protocol. There are more
addresses that need to be filled, and there is an extra header with a
dynamic length based on the addressing mode.
Change the code to cache entries contain a copy of the mesh subframe header +
bridge tunnel header, as well as an embedded struct ieee80211_fast_tx, which
contains the information for building the 802.11 header.
Add a mesh specific early fast xmit call, which looks up a cached entry and
adds only the mesh subframe header, before passing it over to the generic
fast xmit code.
To ensure the changes in network are reflected in these cached headers,
flush affected cached entries on path changes, as well as other conditions
that currently trigger a fast xmit check in other modes (key changes etc.)
This code is loosely based on a previous implementation by:
Sriram R <quic_srirrama@quicinc.com>
Cc: Sriram R <quic_srirrama@quicinc.com>
Signed-off-by: Ryder Lee <ryder.lee@mediatek.com>
Signed-off-by: Felix Fietkau <nbd@nbd.name>
Link: https://lore.kernel.org/r/20230314095956.62085-4-nbd@nbd.name
Signed-off-by: Johannes Berg <johannes.berg@intel.com>
2023-03-14 12:59:53 +03:00
if ( ieee80211_hw_check ( & sdata - > local - > hw , SUPPORT_FAST_XMIT ) )
mesh_fast_tx_cache ( sdata , skb , mpath ) ;
2019-03-16 20:06:30 +03:00
return 0 ;
2008-02-23 17:17:15 +03:00
}
2019-03-16 20:06:30 +03:00
return - ENOENT ;
2008-02-23 17:17:15 +03:00
}
2017-10-17 02:35:49 +03:00
void mesh_path_timer ( struct timer_list * t )
2008-02-23 17:17:15 +03:00
{
2017-10-17 02:35:49 +03:00
struct mesh_path * mpath = from_timer ( mpath , t , timer ) ;
2011-05-12 17:03:32 +04:00
struct ieee80211_sub_if_data * sdata = mpath - > sdata ;
2011-08-10 03:45:08 +04:00
int ret ;
2009-05-17 13:40:42 +04:00
2011-05-12 17:03:32 +04:00
if ( sdata - > local - > quiescing )
2009-05-17 13:40:42 +04:00
return ;
spin_lock_bh ( & mpath - > state_lock ) ;
2008-03-01 02:04:13 +03:00
if ( mpath - > flags & MESH_PATH_RESOLVED | |
2011-08-10 03:45:08 +04:00
( ! ( mpath - > flags & MESH_PATH_RESOLVING ) ) ) {
2008-02-23 17:17:15 +03:00
mpath - > flags & = ~ ( MESH_PATH_RESOLVING | MESH_PATH_RESOLVED ) ;
2011-08-10 03:45:08 +04:00
spin_unlock_bh ( & mpath - > state_lock ) ;
} else if ( mpath - > discovery_retries < max_preq_retries ( sdata ) ) {
2008-02-23 17:17:15 +03:00
+ + mpath - > discovery_retries ;
mpath - > discovery_timeout * = 2 ;
2011-11-04 08:11:10 +04:00
mpath - > flags & = ~ MESH_PATH_REQ_QUEUED ;
2011-08-10 03:45:08 +04:00
spin_unlock_bh ( & mpath - > state_lock ) ;
2008-02-23 17:17:15 +03:00
mesh_queue_preq ( mpath , 0 ) ;
} else {
2015-06-13 00:13:09 +03:00
mpath - > flags & = ~ ( MESH_PATH_RESOLVING |
MESH_PATH_RESOLVED |
MESH_PATH_REQ_QUEUED ) ;
2008-02-23 17:17:15 +03:00
mpath - > exp_time = jiffies ;
2011-08-10 03:45:08 +04:00
spin_unlock_bh ( & mpath - > state_lock ) ;
if ( ! mpath - > is_gate & & mesh_gate_num ( sdata ) > 0 ) {
ret = mesh_path_send_to_gates ( mpath ) ;
if ( ret )
2012-06-22 13:29:50 +04:00
mhwmp_dbg ( sdata , " no gate was reachable \n " ) ;
2011-08-10 03:45:08 +04:00
} else
mesh_path_flush_pending ( mpath ) ;
2008-02-23 17:17:15 +03:00
}
}
2009-11-10 02:46:56 +03:00
2013-02-15 17:40:31 +04:00
void mesh_path_tx_root_frame ( struct ieee80211_sub_if_data * sdata )
2009-11-10 02:46:56 +03:00
{
struct ieee80211_if_mesh * ifmsh = & sdata - > u . mesh ;
2011-08-10 03:45:10 +04:00
u32 interval = ifmsh - > mshcfg . dot11MeshHWMPRannInterval ;
2012-06-13 22:06:07 +04:00
u8 flags , target_flags = 0 ;
2009-11-10 02:46:56 +03:00
2011-08-10 03:45:11 +04:00
flags = ( ifmsh - > mshcfg . dot11MeshGateAnnouncementProtocol )
? RANN_FLAG_IS_GATE : 0 ;
2012-06-13 22:06:07 +04:00
switch ( ifmsh - > mshcfg . dot11MeshHWMPRootMode ) {
case IEEE80211_PROACTIVE_RANN :
mesh_path_sel_frame_tx ( MPATH_RANN , flags , sdata - > vif . addr ,
2013-11-13 11:39:12 +04:00
+ + ifmsh - > sn , 0 , NULL , 0 , broadcast_addr ,
0 , ifmsh - > mshcfg . element_ttl ,
interval , 0 , 0 , sdata ) ;
2012-06-13 22:06:07 +04:00
break ;
case IEEE80211_PROACTIVE_PREQ_WITH_PREP :
flags | = IEEE80211_PREQ_PROACTIVE_PREP_FLAG ;
2020-07-07 23:45:48 +03:00
fallthrough ;
2012-06-13 22:06:07 +04:00
case IEEE80211_PROACTIVE_PREQ_NO_PREP :
interval = ifmsh - > mshcfg . dot11MeshHWMPactivePathToRootTimeout ;
target_flags | = IEEE80211_PREQ_TO_FLAG |
IEEE80211_PREQ_USN_FLAG ;
mesh_path_sel_frame_tx ( MPATH_PREQ , flags , sdata - > vif . addr ,
2013-11-13 11:39:12 +04:00
+ + ifmsh - > sn , target_flags ,
( u8 * ) broadcast_addr , 0 , broadcast_addr ,
0 , ifmsh - > mshcfg . element_ttl , interval ,
0 , ifmsh - > preq_id + + , sdata ) ;
2012-06-13 22:06:07 +04:00
break ;
default :
2012-06-22 13:29:50 +04:00
mhwmp_dbg ( sdata , " Proactive mechanism not supported \n " ) ;
2012-06-13 22:06:07 +04:00
return ;
}
2009-11-10 02:46:56 +03:00
}