License cleanup: add SPDX GPL-2.0 license identifier to files with no license
Many source files in the tree are missing licensing information, which
makes it harder for compliance tools to determine the correct license.
By default all files without license information are under the default
license of the kernel, which is GPL version 2.
Update the files which contain no license information with the 'GPL-2.0'
SPDX license identifier. The SPDX identifier is a legally binding
shorthand, which can be used instead of the full boiler plate text.
This patch is based on work done by Thomas Gleixner and Kate Stewart and
Philippe Ombredanne.
How this work was done:
Patches were generated and checked against linux-4.14-rc6 for a subset of
the use cases:
- file had no licensing information it it.
- file was a */uapi/* one with no licensing information in it,
- file was a */uapi/* one with existing licensing information,
Further patches will be generated in subsequent months to fix up cases
where non-standard license headers were used, and references to license
had to be inferred by heuristics based on keywords.
The analysis to determine which SPDX License Identifier to be applied to
a file was done in a spreadsheet of side by side results from of the
output of two independent scanners (ScanCode & Windriver) producing SPDX
tag:value files created by Philippe Ombredanne. Philippe prepared the
base worksheet, and did an initial spot review of a few 1000 files.
The 4.13 kernel was the starting point of the analysis with 60,537 files
assessed. Kate Stewart did a file by file comparison of the scanner
results in the spreadsheet to determine which SPDX license identifier(s)
to be applied to the file. She confirmed any determination that was not
immediately clear with lawyers working with the Linux Foundation.
Criteria used to select files for SPDX license identifier tagging was:
- Files considered eligible had to be source code files.
- Make and config files were included as candidates if they contained >5
lines of source
- File already had some variant of a license header in it (even if <5
lines).
All documentation files were explicitly excluded.
The following heuristics were used to determine which SPDX license
identifiers to apply.
- when both scanners couldn't find any license traces, file was
considered to have no license information in it, and the top level
COPYING file license applied.
For non */uapi/* files that summary was:
SPDX license identifier # files
---------------------------------------------------|-------
GPL-2.0 11139
and resulted in the first patch in this series.
If that file was a */uapi/* path one, it was "GPL-2.0 WITH
Linux-syscall-note" otherwise it was "GPL-2.0". Results of that was:
SPDX license identifier # files
---------------------------------------------------|-------
GPL-2.0 WITH Linux-syscall-note 930
and resulted in the second patch in this series.
- if a file had some form of licensing information in it, and was one
of the */uapi/* ones, it was denoted with the Linux-syscall-note if
any GPL family license was found in the file or had no licensing in
it (per prior point). Results summary:
SPDX license identifier # files
---------------------------------------------------|------
GPL-2.0 WITH Linux-syscall-note 270
GPL-2.0+ WITH Linux-syscall-note 169
((GPL-2.0 WITH Linux-syscall-note) OR BSD-2-Clause) 21
((GPL-2.0 WITH Linux-syscall-note) OR BSD-3-Clause) 17
LGPL-2.1+ WITH Linux-syscall-note 15
GPL-1.0+ WITH Linux-syscall-note 14
((GPL-2.0+ WITH Linux-syscall-note) OR BSD-3-Clause) 5
LGPL-2.0+ WITH Linux-syscall-note 4
LGPL-2.1 WITH Linux-syscall-note 3
((GPL-2.0 WITH Linux-syscall-note) OR MIT) 3
((GPL-2.0 WITH Linux-syscall-note) AND MIT) 1
and that resulted in the third patch in this series.
- when the two scanners agreed on the detected license(s), that became
the concluded license(s).
- when there was disagreement between the two scanners (one detected a
license but the other didn't, or they both detected different
licenses) a manual inspection of the file occurred.
- In most cases a manual inspection of the information in the file
resulted in a clear resolution of the license that should apply (and
which scanner probably needed to revisit its heuristics).
- When it was not immediately clear, the license identifier was
confirmed with lawyers working with the Linux Foundation.
- If there was any question as to the appropriate license identifier,
the file was flagged for further research and to be revisited later
in time.
In total, over 70 hours of logged manual review was done on the
spreadsheet to determine the SPDX license identifiers to apply to the
source files by Kate, Philippe, Thomas and, in some cases, confirmation
by lawyers working with the Linux Foundation.
Kate also obtained a third independent scan of the 4.13 code base from
FOSSology, and compared selected files where the other two scanners
disagreed against that SPDX file, to see if there was new insights. The
Windriver scanner is based on an older version of FOSSology in part, so
they are related.
Thomas did random spot checks in about 500 files from the spreadsheets
for the uapi headers and agreed with SPDX license identifier in the
files he inspected. For the non-uapi files Thomas did random spot checks
in about 15000 files.
In initial set of patches against 4.14-rc6, 3 files were found to have
copy/paste license identifier errors, and have been fixed to reflect the
correct identifier.
Additionally Philippe spent 10 hours this week doing a detailed manual
inspection and review of the 12,461 patched files from the initial patch
version early this week with:
- a full scancode scan run, collecting the matched texts, detected
license ids and scores
- reviewing anything where there was a license detected (about 500+
files) to ensure that the applied SPDX license was correct
- reviewing anything where there was no detection but the patch license
was not GPL-2.0 WITH Linux-syscall-note to ensure that the applied
SPDX license was correct
This produced a worksheet with 20 files needing minor correction. This
worksheet was then exported into 3 different .csv files for the
different types of files to be modified.
These .csv files were then reviewed by Greg. Thomas wrote a script to
parse the csv files and add the proper SPDX tag to the file, in the
format that the file expected. This script was further refined by Greg
based on the output to detect more types of files automatically and to
distinguish between header and source .c files (which need different
comment types.) Finally Greg ran the script using the .csv files to
generate the patches.
Reviewed-by: Kate Stewart <kstewart@linuxfoundation.org>
Reviewed-by: Philippe Ombredanne <pombredanne@nexb.com>
Reviewed-by: Thomas Gleixner <tglx@linutronix.de>
Signed-off-by: Greg Kroah-Hartman <gregkh@linuxfoundation.org>
2017-11-01 17:07:57 +03:00
// SPDX-License-Identifier: GPL-2.0
2012-07-10 11:49:14 +04:00
# include <linux/rcupdate.h>
# include <linux/spinlock.h>
# include <linux/jiffies.h>
2012-07-10 03:19:30 +04:00
# include <linux/module.h>
2012-07-10 03:07:30 +04:00
# include <linux/cache.h>
2012-07-10 11:49:14 +04:00
# include <linux/slab.h>
# include <linux/init.h>
2012-07-10 03:07:30 +04:00
# include <linux/tcp.h>
2012-07-20 03:02:34 +04:00
# include <linux/hash.h>
2012-09-04 15:03:15 +04:00
# include <linux/tcp_metrics.h>
2012-11-16 09:31:53 +04:00
# include <linux/vmalloc.h>
2012-07-10 03:07:30 +04:00
# include <net/inet_connection_sock.h>
2012-07-10 11:49:14 +04:00
# include <net/net_namespace.h>
2012-07-10 03:19:30 +04:00
# include <net/request_sock.h>
2012-07-10 11:49:14 +04:00
# include <net/inetpeer.h>
2012-07-10 03:07:30 +04:00
# include <net/sock.h>
2012-07-10 11:49:14 +04:00
# include <net/ipv6.h>
2012-07-10 03:07:30 +04:00
# include <net/dst.h>
# include <net/tcp.h>
2012-09-04 15:03:15 +04:00
# include <net/genetlink.h>
2012-07-10 03:07:30 +04:00
2014-01-18 12:55:41 +04:00
static struct tcp_metrics_block * __tcp_get_metrics ( const struct inetpeer_addr * saddr ,
const struct inetpeer_addr * daddr ,
2014-01-16 23:01:21 +04:00
struct net * net , unsigned int hash ) ;
2012-07-19 10:43:06 +04:00
struct tcp_fastopen_metrics {
u16 mss ;
2015-04-07 00:37:27 +03:00
u16 syn_loss : 10 , /* Recurring Fast Open SYN losses */
try_exp : 2 ; /* Request w/ exp. option (once) */
2012-07-19 10:43:10 +04:00
unsigned long last_syn_loss ; /* Last Fast Open SYN loss */
2012-07-19 10:43:06 +04:00
struct tcp_fastopen_cookie cookie ;
} ;
2014-02-27 02:02:48 +04:00
/* TCP_METRIC_MAX includes 2 extra fields for userspace compatibility
* Kernel only stores RTT and RTTVAR in usec resolution
*/
# define TCP_METRIC_MAX_KERNEL (TCP_METRIC_MAX - 2)
2012-07-10 11:49:14 +04:00
struct tcp_metrics_block {
struct tcp_metrics_block __rcu * tcpm_next ;
2015-03-13 08:05:52 +03:00
possible_net_t tcpm_net ;
2014-01-08 19:05:56 +04:00
struct inetpeer_addr tcpm_saddr ;
2014-01-08 19:05:55 +04:00
struct inetpeer_addr tcpm_daddr ;
2012-07-10 11:49:14 +04:00
unsigned long tcpm_stamp ;
u32 tcpm_lock ;
2014-02-27 02:02:48 +04:00
u32 tcpm_vals [ TCP_METRIC_MAX_KERNEL + 1 ] ;
2012-07-19 10:43:06 +04:00
struct tcp_fastopen_metrics tcpm_fastopen ;
2012-09-04 15:03:15 +04:00
struct rcu_head rcu_head ;
2012-07-10 11:49:14 +04:00
} ;
2015-03-13 08:05:52 +03:00
static inline struct net * tm_net ( struct tcp_metrics_block * tm )
{
return read_pnet ( & tm - > tcpm_net ) ;
}
2012-07-10 11:49:14 +04:00
static bool tcp_metric_locked ( struct tcp_metrics_block * tm ,
enum tcp_metric_index idx )
{
return tm - > tcpm_lock & ( 1 < < idx ) ;
}
static u32 tcp_metric_get ( struct tcp_metrics_block * tm ,
enum tcp_metric_index idx )
{
return tm - > tcpm_vals [ idx ] ;
}
static void tcp_metric_set ( struct tcp_metrics_block * tm ,
enum tcp_metric_index idx ,
u32 val )
{
tm - > tcpm_vals [ idx ] = val ;
}
static bool addr_same ( const struct inetpeer_addr * a ,
const struct inetpeer_addr * b )
{
2015-08-28 02:07:01 +03:00
return inetpeer_addr_cmp ( a , b ) = = 0 ;
2012-07-10 11:49:14 +04:00
}
struct tcpm_hash_bucket {
struct tcp_metrics_block __rcu * chain ;
} ;
2015-03-13 08:07:44 +03:00
static struct tcpm_hash_bucket * tcp_metrics_hash __read_mostly ;
static unsigned int tcp_metrics_hash_log __read_mostly ;
2012-07-10 11:49:14 +04:00
static DEFINE_SPINLOCK ( tcp_metrics_lock ) ;
2014-02-27 02:02:48 +04:00
static void tcpm_suck_dst ( struct tcp_metrics_block * tm ,
const struct dst_entry * dst ,
2013-05-03 23:12:45 +04:00
bool fastopen_clear )
2012-07-10 11:49:14 +04:00
{
2014-02-27 02:02:48 +04:00
u32 msval ;
2012-07-10 11:49:14 +04:00
u32 val ;
2012-07-23 11:46:38 +04:00
tm - > tcpm_stamp = jiffies ;
2012-07-10 11:49:14 +04:00
val = 0 ;
if ( dst_metric_locked ( dst , RTAX_RTT ) )
val | = 1 < < TCP_METRIC_RTT ;
if ( dst_metric_locked ( dst , RTAX_RTTVAR ) )
val | = 1 < < TCP_METRIC_RTTVAR ;
if ( dst_metric_locked ( dst , RTAX_SSTHRESH ) )
val | = 1 < < TCP_METRIC_SSTHRESH ;
if ( dst_metric_locked ( dst , RTAX_CWND ) )
val | = 1 < < TCP_METRIC_CWND ;
if ( dst_metric_locked ( dst , RTAX_REORDERING ) )
val | = 1 < < TCP_METRIC_REORDERING ;
tm - > tcpm_lock = val ;
2014-02-27 02:02:48 +04:00
msval = dst_metric_raw ( dst , RTAX_RTT ) ;
tm - > tcpm_vals [ TCP_METRIC_RTT ] = msval * USEC_PER_MSEC ;
msval = dst_metric_raw ( dst , RTAX_RTTVAR ) ;
tm - > tcpm_vals [ TCP_METRIC_RTTVAR ] = msval * USEC_PER_MSEC ;
2012-07-10 11:49:14 +04:00
tm - > tcpm_vals [ TCP_METRIC_SSTHRESH ] = dst_metric_raw ( dst , RTAX_SSTHRESH ) ;
tm - > tcpm_vals [ TCP_METRIC_CWND ] = dst_metric_raw ( dst , RTAX_CWND ) ;
tm - > tcpm_vals [ TCP_METRIC_REORDERING ] = dst_metric_raw ( dst , RTAX_REORDERING ) ;
2013-05-03 23:12:45 +04:00
if ( fastopen_clear ) {
tm - > tcpm_fastopen . mss = 0 ;
tm - > tcpm_fastopen . syn_loss = 0 ;
2015-04-07 00:37:27 +03:00
tm - > tcpm_fastopen . try_exp = 0 ;
tm - > tcpm_fastopen . cookie . exp = false ;
2013-05-03 23:12:45 +04:00
tm - > tcpm_fastopen . cookie . len = 0 ;
}
2012-07-10 11:49:14 +04:00
}
2014-01-16 23:01:21 +04:00
# define TCP_METRICS_TIMEOUT (60 * 60 * HZ)
static void tcpm_check_stamp ( struct tcp_metrics_block * tm , struct dst_entry * dst )
{
if ( tm & & unlikely ( time_after ( jiffies , tm - > tcpm_stamp + TCP_METRICS_TIMEOUT ) ) )
tcpm_suck_dst ( tm , dst , false ) ;
}
# define TCP_METRICS_RECLAIM_DEPTH 5
# define TCP_METRICS_RECLAIM_PTR (struct tcp_metrics_block *) 0x1UL
2015-03-16 17:14:34 +03:00
# define deref_locked(p) \
rcu_dereference_protected ( p , lockdep_is_held ( & tcp_metrics_lock ) )
2012-07-10 11:49:14 +04:00
static struct tcp_metrics_block * tcpm_new ( struct dst_entry * dst ,
2014-01-08 19:05:56 +04:00
struct inetpeer_addr * saddr ,
2014-01-08 19:05:55 +04:00
struct inetpeer_addr * daddr ,
2014-01-16 23:01:21 +04:00
unsigned int hash )
2012-07-10 11:49:14 +04:00
{
struct tcp_metrics_block * tm ;
struct net * net ;
2014-01-16 23:01:21 +04:00
bool reclaim = false ;
2012-07-10 11:49:14 +04:00
spin_lock_bh ( & tcp_metrics_lock ) ;
net = dev_net ( dst - > dev ) ;
2014-01-16 23:01:21 +04:00
/* While waiting for the spin-lock the cache might have been populated
* with this entry and so we have to check again .
*/
2014-01-18 12:55:41 +04:00
tm = __tcp_get_metrics ( saddr , daddr , net , hash ) ;
2014-01-16 23:01:21 +04:00
if ( tm = = TCP_METRICS_RECLAIM_PTR ) {
reclaim = true ;
tm = NULL ;
}
if ( tm ) {
tcpm_check_stamp ( tm , dst ) ;
goto out_unlock ;
}
2012-07-10 11:49:14 +04:00
if ( unlikely ( reclaim ) ) {
struct tcp_metrics_block * oldest ;
2015-03-16 17:14:34 +03:00
oldest = deref_locked ( tcp_metrics_hash [ hash ] . chain ) ;
for ( tm = deref_locked ( oldest - > tcpm_next ) ; tm ;
tm = deref_locked ( tm - > tcpm_next ) ) {
2012-07-10 11:49:14 +04:00
if ( time_before ( tm - > tcpm_stamp , oldest - > tcpm_stamp ) )
oldest = tm ;
}
tm = oldest ;
} else {
tm = kmalloc ( sizeof ( * tm ) , GFP_ATOMIC ) ;
if ( ! tm )
goto out_unlock ;
}
2015-03-13 08:05:52 +03:00
write_pnet ( & tm - > tcpm_net , net ) ;
2014-01-08 19:05:56 +04:00
tm - > tcpm_saddr = * saddr ;
2014-01-08 19:05:55 +04:00
tm - > tcpm_daddr = * daddr ;
2012-07-10 11:49:14 +04:00
2013-05-03 23:12:45 +04:00
tcpm_suck_dst ( tm , dst , true ) ;
2012-07-10 11:49:14 +04:00
if ( likely ( ! reclaim ) ) {
2015-03-13 08:07:44 +03:00
tm - > tcpm_next = tcp_metrics_hash [ hash ] . chain ;
rcu_assign_pointer ( tcp_metrics_hash [ hash ] . chain , tm ) ;
2012-07-10 11:49:14 +04:00
}
out_unlock :
spin_unlock_bh ( & tcp_metrics_lock ) ;
return tm ;
}
static struct tcp_metrics_block * tcp_get_encode ( struct tcp_metrics_block * tm , int depth )
{
if ( tm )
return tm ;
if ( depth > TCP_METRICS_RECLAIM_DEPTH )
return TCP_METRICS_RECLAIM_PTR ;
return NULL ;
}
2014-01-08 19:05:56 +04:00
static struct tcp_metrics_block * __tcp_get_metrics ( const struct inetpeer_addr * saddr ,
const struct inetpeer_addr * daddr ,
2012-07-10 11:49:14 +04:00
struct net * net , unsigned int hash )
{
struct tcp_metrics_block * tm ;
int depth = 0 ;
2015-03-13 08:07:44 +03:00
for ( tm = rcu_dereference ( tcp_metrics_hash [ hash ] . chain ) ; tm ;
2012-07-10 11:49:14 +04:00
tm = rcu_dereference ( tm - > tcpm_next ) ) {
2014-01-08 19:05:56 +04:00
if ( addr_same ( & tm - > tcpm_saddr , saddr ) & &
2015-03-13 08:05:52 +03:00
addr_same ( & tm - > tcpm_daddr , daddr ) & &
net_eq ( tm_net ( tm ) , net ) )
2012-07-10 11:49:14 +04:00
break ;
depth + + ;
}
return tcp_get_encode ( tm , depth ) ;
}
static struct tcp_metrics_block * __tcp_get_metrics_req ( struct request_sock * req ,
struct dst_entry * dst )
{
struct tcp_metrics_block * tm ;
2014-01-08 19:05:56 +04:00
struct inetpeer_addr saddr , daddr ;
2012-07-10 11:49:14 +04:00
unsigned int hash ;
struct net * net ;
2014-01-08 19:05:56 +04:00
saddr . family = req - > rsk_ops - > family ;
2014-01-08 19:05:55 +04:00
daddr . family = req - > rsk_ops - > family ;
switch ( daddr . family ) {
2012-07-10 11:49:14 +04:00
case AF_INET :
2015-08-28 02:07:00 +03:00
inetpeer_set_addr_v4 ( & saddr , inet_rsk ( req ) - > ir_loc_addr ) ;
inetpeer_set_addr_v4 ( & daddr , inet_rsk ( req ) - > ir_rmt_addr ) ;
2015-08-28 02:06:59 +03:00
hash = ipv4_addr_hash ( inet_rsk ( req ) - > ir_rmt_addr ) ;
2012-07-10 11:49:14 +04:00
break ;
2013-10-10 02:21:29 +04:00
# if IS_ENABLED(CONFIG_IPV6)
2012-07-10 11:49:14 +04:00
case AF_INET6 :
2015-08-28 02:07:00 +03:00
inetpeer_set_addr_v6 ( & saddr , & inet_rsk ( req ) - > ir_v6_loc_addr ) ;
inetpeer_set_addr_v6 ( & daddr , & inet_rsk ( req ) - > ir_v6_rmt_addr ) ;
2013-10-10 02:21:29 +04:00
hash = ipv6_addr_hash ( & inet_rsk ( req ) - > ir_v6_rmt_addr ) ;
2012-07-10 11:49:14 +04:00
break ;
2013-10-10 02:21:29 +04:00
# endif
2012-07-10 11:49:14 +04:00
default :
return NULL ;
}
net = dev_net ( dst - > dev ) ;
2015-03-13 08:05:24 +03:00
hash ^ = net_hash_mix ( net ) ;
2015-03-13 08:07:44 +03:00
hash = hash_32 ( hash , tcp_metrics_hash_log ) ;
2012-07-10 11:49:14 +04:00
2015-03-13 08:07:44 +03:00
for ( tm = rcu_dereference ( tcp_metrics_hash [ hash ] . chain ) ; tm ;
2012-07-10 11:49:14 +04:00
tm = rcu_dereference ( tm - > tcpm_next ) ) {
2014-01-08 19:05:56 +04:00
if ( addr_same ( & tm - > tcpm_saddr , & saddr ) & &
2015-03-13 08:05:52 +03:00
addr_same ( & tm - > tcpm_daddr , & daddr ) & &
net_eq ( tm_net ( tm ) , net ) )
2012-07-10 11:49:14 +04:00
break ;
}
tcpm_check_stamp ( tm , dst ) ;
return tm ;
}
static struct tcp_metrics_block * tcp_get_metrics ( struct sock * sk ,
struct dst_entry * dst ,
bool create )
{
struct tcp_metrics_block * tm ;
2014-01-08 19:05:56 +04:00
struct inetpeer_addr saddr , daddr ;
2012-07-10 11:49:14 +04:00
unsigned int hash ;
struct net * net ;
2014-01-22 16:58:44 +04:00
if ( sk - > sk_family = = AF_INET ) {
2015-08-28 02:07:00 +03:00
inetpeer_set_addr_v4 ( & saddr , inet_sk ( sk ) - > inet_saddr ) ;
inetpeer_set_addr_v4 ( & daddr , inet_sk ( sk ) - > inet_daddr ) ;
2015-08-28 02:06:59 +03:00
hash = ipv4_addr_hash ( inet_sk ( sk ) - > inet_daddr ) ;
2014-01-22 16:58:44 +04:00
}
2013-10-09 14:05:48 +04:00
# if IS_ENABLED(CONFIG_IPV6)
2014-01-22 16:58:44 +04:00
else if ( sk - > sk_family = = AF_INET6 ) {
if ( ipv6_addr_v4mapped ( & sk - > sk_v6_daddr ) ) {
2015-08-28 02:07:00 +03:00
inetpeer_set_addr_v4 ( & saddr , inet_sk ( sk ) - > inet_saddr ) ;
inetpeer_set_addr_v4 ( & daddr , inet_sk ( sk ) - > inet_daddr ) ;
2015-08-28 02:06:59 +03:00
hash = ipv4_addr_hash ( inet_sk ( sk ) - > inet_daddr ) ;
2014-01-22 16:58:44 +04:00
} else {
2015-08-28 02:07:00 +03:00
inetpeer_set_addr_v6 ( & saddr , & sk - > sk_v6_rcv_saddr ) ;
inetpeer_set_addr_v6 ( & daddr , & sk - > sk_v6_daddr ) ;
2014-01-22 16:58:44 +04:00
hash = ipv6_addr_hash ( & sk - > sk_v6_daddr ) ;
}
}
2013-10-09 14:05:48 +04:00
# endif
2014-01-22 16:58:44 +04:00
else
2012-07-10 11:49:14 +04:00
return NULL ;
net = dev_net ( dst - > dev ) ;
2015-03-13 08:05:24 +03:00
hash ^ = net_hash_mix ( net ) ;
2015-03-13 08:07:44 +03:00
hash = hash_32 ( hash , tcp_metrics_hash_log ) ;
2012-07-10 11:49:14 +04:00
2014-01-08 19:05:56 +04:00
tm = __tcp_get_metrics ( & saddr , & daddr , net , hash ) ;
2014-01-16 23:01:21 +04:00
if ( tm = = TCP_METRICS_RECLAIM_PTR )
2012-07-10 11:49:14 +04:00
tm = NULL ;
if ( ! tm & & create )
2014-01-18 12:55:41 +04:00
tm = tcpm_new ( dst , & saddr , & daddr , hash ) ;
2012-07-10 11:49:14 +04:00
else
tcpm_check_stamp ( tm , dst ) ;
return tm ;
}
2012-07-10 03:07:30 +04:00
/* Save metrics learned by this TCP session. This function is called
* only , when TCP finishes successfully i . e . when it enters TIME - WAIT
* or goes from LAST - ACK to CLOSE .
*/
void tcp_update_metrics ( struct sock * sk )
{
2012-07-10 11:49:14 +04:00
const struct inet_connection_sock * icsk = inet_csk ( sk ) ;
2012-07-10 03:07:30 +04:00
struct dst_entry * dst = __sk_dst_get ( sk ) ;
2012-07-10 11:49:14 +04:00
struct tcp_sock * tp = tcp_sk ( sk ) ;
2016-02-03 10:46:52 +03:00
struct net * net = sock_net ( sk ) ;
2012-07-10 11:49:14 +04:00
struct tcp_metrics_block * tm ;
unsigned long rtt ;
u32 val ;
int m ;
2012-07-10 03:07:30 +04:00
2017-02-07 00:14:14 +03:00
sk_dst_confirm ( sk ) ;
2017-10-27 17:47:21 +03:00
if ( net - > ipv4 . sysctl_tcp_nometrics_save | | ! dst )
2012-07-10 03:07:30 +04:00
return ;
2012-07-10 11:49:14 +04:00
rcu_read_lock ( ) ;
2014-02-27 02:02:48 +04:00
if ( icsk - > icsk_backoff | | ! tp - > srtt_us ) {
2012-07-10 11:49:14 +04:00
/* This session failed to estimate rtt. Why?
* Probably , no packets returned in time . Reset our
* results .
*/
tm = tcp_get_metrics ( sk , dst , false ) ;
if ( tm & & ! tcp_metric_locked ( tm , TCP_METRIC_RTT ) )
tcp_metric_set ( tm , TCP_METRIC_RTT , 0 ) ;
goto out_unlock ;
} else
tm = tcp_get_metrics ( sk , dst , true ) ;
2012-07-10 03:07:30 +04:00
2012-07-10 11:49:14 +04:00
if ( ! tm )
goto out_unlock ;
2012-07-10 03:07:30 +04:00
2014-02-27 02:02:48 +04:00
rtt = tcp_metric_get ( tm , TCP_METRIC_RTT ) ;
m = rtt - tp - > srtt_us ;
2012-07-10 03:07:30 +04:00
2012-07-10 11:49:14 +04:00
/* If newly calculated rtt larger than stored one, store new
* one . Otherwise , use EWMA . Remember , rtt overestimation is
* always better than underestimation .
*/
if ( ! tcp_metric_locked ( tm , TCP_METRIC_RTT ) ) {
if ( m < = 0 )
2014-02-27 02:02:48 +04:00
rtt = tp - > srtt_us ;
2012-07-10 11:49:14 +04:00
else
rtt - = ( m > > 3 ) ;
2014-02-27 02:02:48 +04:00
tcp_metric_set ( tm , TCP_METRIC_RTT , rtt ) ;
2012-07-10 11:49:14 +04:00
}
2012-07-10 03:07:30 +04:00
2012-07-10 11:49:14 +04:00
if ( ! tcp_metric_locked ( tm , TCP_METRIC_RTTVAR ) ) {
unsigned long var ;
2012-07-10 03:07:30 +04:00
2012-07-10 11:49:14 +04:00
if ( m < 0 )
m = - m ;
2012-07-10 03:07:30 +04:00
2012-07-10 11:49:14 +04:00
/* Scale deviation to rttvar fixed point */
m > > = 1 ;
2014-02-27 02:02:48 +04:00
if ( m < tp - > mdev_us )
m = tp - > mdev_us ;
2012-07-10 03:07:30 +04:00
2014-02-27 02:02:48 +04:00
var = tcp_metric_get ( tm , TCP_METRIC_RTTVAR ) ;
2012-07-10 11:49:14 +04:00
if ( m > = var )
var = m ;
else
var - = ( var - m ) > > 2 ;
2012-07-10 03:07:30 +04:00
2014-02-27 02:02:48 +04:00
tcp_metric_set ( tm , TCP_METRIC_RTTVAR , var ) ;
2012-07-10 11:49:14 +04:00
}
if ( tcp_in_initial_slowstart ( tp ) ) {
/* Slow start still did not finish. */
2019-12-09 22:19:59 +03:00
if ( ! net - > ipv4 . sysctl_tcp_no_ssthresh_metrics_save & &
! tcp_metric_locked ( tm , TCP_METRIC_SSTHRESH ) ) {
2012-07-10 11:49:14 +04:00
val = tcp_metric_get ( tm , TCP_METRIC_SSTHRESH ) ;
if ( val & & ( tp - > snd_cwnd > > 1 ) > val )
tcp_metric_set ( tm , TCP_METRIC_SSTHRESH ,
tp - > snd_cwnd > > 1 ) ;
}
if ( ! tcp_metric_locked ( tm , TCP_METRIC_CWND ) ) {
val = tcp_metric_get ( tm , TCP_METRIC_CWND ) ;
if ( tp - > snd_cwnd > val )
tcp_metric_set ( tm , TCP_METRIC_CWND ,
tp - > snd_cwnd ) ;
}
2015-07-09 23:16:29 +03:00
} else if ( ! tcp_in_slow_start ( tp ) & &
2012-07-10 11:49:14 +04:00
icsk - > icsk_ca_state = = TCP_CA_Open ) {
/* Cong. avoidance phase, cwnd is reliable. */
2019-12-09 22:19:59 +03:00
if ( ! net - > ipv4 . sysctl_tcp_no_ssthresh_metrics_save & &
! tcp_metric_locked ( tm , TCP_METRIC_SSTHRESH ) )
2012-07-10 11:49:14 +04:00
tcp_metric_set ( tm , TCP_METRIC_SSTHRESH ,
max ( tp - > snd_cwnd > > 1 , tp - > snd_ssthresh ) ) ;
if ( ! tcp_metric_locked ( tm , TCP_METRIC_CWND ) ) {
val = tcp_metric_get ( tm , TCP_METRIC_CWND ) ;
2012-07-12 04:18:04 +04:00
tcp_metric_set ( tm , TCP_METRIC_CWND , ( val + tp - > snd_cwnd ) > > 1 ) ;
2012-07-10 11:49:14 +04:00
}
} else {
/* Else slow start did not finish, cwnd is non-sense,
* ssthresh may be also invalid .
*/
if ( ! tcp_metric_locked ( tm , TCP_METRIC_CWND ) ) {
val = tcp_metric_get ( tm , TCP_METRIC_CWND ) ;
tcp_metric_set ( tm , TCP_METRIC_CWND ,
( val + tp - > snd_ssthresh ) > > 1 ) ;
}
2019-12-09 22:19:59 +03:00
if ( ! net - > ipv4 . sysctl_tcp_no_ssthresh_metrics_save & &
! tcp_metric_locked ( tm , TCP_METRIC_SSTHRESH ) ) {
2012-07-10 11:49:14 +04:00
val = tcp_metric_get ( tm , TCP_METRIC_SSTHRESH ) ;
if ( val & & tp - > snd_ssthresh > val )
tcp_metric_set ( tm , TCP_METRIC_SSTHRESH ,
tp - > snd_ssthresh ) ;
}
if ( ! tcp_metric_locked ( tm , TCP_METRIC_REORDERING ) ) {
val = tcp_metric_get ( tm , TCP_METRIC_REORDERING ) ;
if ( val < tp - > reordering & &
2016-02-03 10:46:52 +03:00
tp - > reordering ! = net - > ipv4 . sysctl_tcp_reordering )
2012-07-10 11:49:14 +04:00
tcp_metric_set ( tm , TCP_METRIC_REORDERING ,
tp - > reordering ) ;
2012-07-10 03:07:30 +04:00
}
}
2012-07-10 11:49:14 +04:00
tm - > tcpm_stamp = jiffies ;
out_unlock :
rcu_read_unlock ( ) ;
2012-07-10 03:07:30 +04:00
}
/* Initialize metrics on socket. */
void tcp_init_metrics ( struct sock * sk )
{
struct dst_entry * dst = __sk_dst_get ( sk ) ;
2012-07-10 11:49:14 +04:00
struct tcp_sock * tp = tcp_sk ( sk ) ;
2019-12-09 22:19:59 +03:00
struct net * net = sock_net ( sk ) ;
2012-07-10 11:49:14 +04:00
struct tcp_metrics_block * tm ;
2013-08-30 19:35:53 +04:00
u32 val , crtt = 0 ; /* cached RTT scaled by 8 */
2012-07-10 03:07:30 +04:00
2017-02-07 00:14:14 +03:00
sk_dst_confirm ( sk ) ;
2015-04-03 11:17:26 +03:00
if ( ! dst )
2012-07-10 03:07:30 +04:00
goto reset ;
2012-07-10 11:49:14 +04:00
rcu_read_lock ( ) ;
tm = tcp_get_metrics ( sk , dst , true ) ;
if ( ! tm ) {
rcu_read_unlock ( ) ;
goto reset ;
}
if ( tcp_metric_locked ( tm , TCP_METRIC_CWND ) )
tp - > snd_cwnd_clamp = tcp_metric_get ( tm , TCP_METRIC_CWND ) ;
2019-12-09 22:19:59 +03:00
val = net - > ipv4 . sysctl_tcp_no_ssthresh_metrics_save ?
0 : tcp_metric_get ( tm , TCP_METRIC_SSTHRESH ) ;
2012-07-10 11:49:14 +04:00
if ( val ) {
tp - > snd_ssthresh = val ;
2012-07-10 03:07:30 +04:00
if ( tp - > snd_ssthresh > tp - > snd_cwnd_clamp )
tp - > snd_ssthresh = tp - > snd_cwnd_clamp ;
} else {
/* ssthresh may have been reduced unnecessarily during.
* 3 WHS . Restore it back to its initial default .
*/
tp - > snd_ssthresh = TCP_INFINITE_SSTHRESH ;
}
2012-07-10 11:49:14 +04:00
val = tcp_metric_get ( tm , TCP_METRIC_REORDERING ) ;
2017-11-09 00:01:26 +03:00
if ( val & & tp - > reordering ! = val )
2012-07-10 11:49:14 +04:00
tp - > reordering = val ;
2012-07-10 03:07:30 +04:00
2014-02-27 02:02:48 +04:00
crtt = tcp_metric_get ( tm , TCP_METRIC_RTT ) ;
2012-07-10 11:49:14 +04:00
rcu_read_unlock ( ) ;
2012-07-10 03:07:30 +04:00
reset :
2013-09-04 01:14:35 +04:00
/* The initial RTT measurement from the SYN/SYN-ACK is not ideal
* to seed the RTO for later data packets because SYN packets are
* small . Use the per - dst cached values to seed the RTO but keep
* the RTT estimator variables intact ( e . g . , srtt , mdev , rttvar ) .
* Later the RTO will be updated immediately upon obtaining the first
* data RTT sample ( tcp_rtt_estimator ( ) ) . Hence the cached RTT only
* influences the first RTO but not later RTT estimation .
*
* But if RTT is not available from the SYN ( due to retransmits or
* syn cookies ) or the cache , force a conservative 3 secs timeout .
*
* A bit of theory . RTT is time passed after " normal " sized packet
* is sent until it is ACKed . In normal circumstances sending small
* packets force peer to delay ACKs and calculation is correct too .
* The algorithm is adaptive and , provided we follow specs , it
* NEVER underestimate RTT . BUT ! If peer tries to make some clever
* tricks sort of " quick acks " for time long enough to decrease RTT
* to low value , and then abruptly stops to do it and starts to delay
* ACKs , wait for troubles .
*/
2014-02-27 02:02:48 +04:00
if ( crtt > tp - > srtt_us ) {
2013-09-17 05:44:20 +04:00
/* Set RTO like tcp_rtt_estimator(), but from cached RTT. */
2016-02-21 10:12:39 +03:00
crtt / = 8 * USEC_PER_SEC / HZ ;
2013-09-17 05:44:20 +04:00
inet_csk ( sk ) - > icsk_rto = crtt + max ( 2 * crtt , tcp_rto_min ( sk ) ) ;
2014-02-27 02:02:48 +04:00
} else if ( tp - > srtt_us = = 0 ) {
2012-07-10 03:07:30 +04:00
/* RFC6298: 5.7 We've failed to get a valid RTT sample from
* 3 WHS . This is most likely due to retransmission ,
* including spurious one . Reset the RTO back to 3 secs
* from the more aggressive 1 sec to avoid more spurious
* retransmission .
*/
2014-02-27 02:02:48 +04:00
tp - > rttvar_us = jiffies_to_usecs ( TCP_TIMEOUT_FALLBACK ) ;
tp - > mdev_us = tp - > mdev_max_us = tp - > rttvar_us ;
2012-07-10 03:07:30 +04:00
inet_csk ( sk ) - > icsk_rto = TCP_TIMEOUT_FALLBACK ;
}
}
2012-07-10 03:19:30 +04:00
2017-03-15 23:30:45 +03:00
bool tcp_peer_is_proven ( struct request_sock * req , struct dst_entry * dst )
2012-07-10 03:19:30 +04:00
{
2012-07-10 11:49:14 +04:00
struct tcp_metrics_block * tm ;
bool ret ;
2012-07-10 03:19:30 +04:00
if ( ! dst )
return false ;
2012-07-10 11:49:14 +04:00
rcu_read_lock ( ) ;
tm = __tcp_get_metrics_req ( req , dst ) ;
2017-03-15 23:30:45 +03:00
if ( tm & & tcp_metric_get ( tm , TCP_METRIC_RTT ) )
2012-07-10 14:14:24 +04:00
ret = true ;
2017-03-15 23:30:45 +03:00
else
ret = false ;
2012-07-10 14:14:24 +04:00
rcu_read_unlock ( ) ;
return ret ;
}
2012-07-19 10:43:06 +04:00
static DEFINE_SEQLOCK ( fastopen_seqlock ) ;
void tcp_fastopen_cache_get ( struct sock * sk , u16 * mss ,
2017-12-13 00:10:40 +03:00
struct tcp_fastopen_cookie * cookie )
2012-07-19 10:43:06 +04:00
{
struct tcp_metrics_block * tm ;
rcu_read_lock ( ) ;
tm = tcp_get_metrics ( sk , __sk_dst_get ( sk ) , false ) ;
if ( tm ) {
struct tcp_fastopen_metrics * tfom = & tm - > tcpm_fastopen ;
unsigned int seq ;
do {
seq = read_seqbegin ( & fastopen_seqlock ) ;
if ( tfom - > mss )
* mss = tfom - > mss ;
* cookie = tfom - > cookie ;
2015-04-07 00:37:27 +03:00
if ( cookie - > len < = 0 & & tfom - > try_exp = = 1 )
cookie - > exp = true ;
2012-07-19 10:43:06 +04:00
} while ( read_seqretry ( & fastopen_seqlock , seq ) ) ;
}
rcu_read_unlock ( ) ;
}
void tcp_fastopen_cache_set ( struct sock * sk , u16 mss ,
2015-04-07 00:37:27 +03:00
struct tcp_fastopen_cookie * cookie , bool syn_lost ,
u16 try_exp )
2012-07-19 10:43:06 +04:00
{
2013-11-14 03:00:46 +04:00
struct dst_entry * dst = __sk_dst_get ( sk ) ;
2012-07-19 10:43:06 +04:00
struct tcp_metrics_block * tm ;
2013-11-14 03:00:46 +04:00
if ( ! dst )
return ;
2012-07-19 10:43:06 +04:00
rcu_read_lock ( ) ;
2013-11-14 03:00:46 +04:00
tm = tcp_get_metrics ( sk , dst , true ) ;
2012-07-19 10:43:06 +04:00
if ( tm ) {
struct tcp_fastopen_metrics * tfom = & tm - > tcpm_fastopen ;
write_seqlock_bh ( & fastopen_seqlock ) ;
2013-10-29 21:09:05 +04:00
if ( mss )
tfom - > mss = mss ;
if ( cookie & & cookie - > len > 0 )
2012-07-19 10:43:06 +04:00
tfom - > cookie = * cookie ;
2015-04-07 00:37:27 +03:00
else if ( try_exp > tfom - > try_exp & &
tfom - > cookie . len < = 0 & & ! tfom - > cookie . exp )
tfom - > try_exp = try_exp ;
2012-07-19 10:43:10 +04:00
if ( syn_lost ) {
+ + tfom - > syn_loss ;
tfom - > last_syn_loss = jiffies ;
} else
tfom - > syn_loss = 0 ;
2012-07-19 10:43:06 +04:00
write_sequnlock_bh ( & fastopen_seqlock ) ;
}
rcu_read_unlock ( ) ;
}
2016-10-24 15:40:03 +03:00
static struct genl_family tcp_metrics_nl_family ;
2012-09-04 15:03:15 +04:00
2016-09-01 01:21:37 +03:00
static const struct nla_policy tcp_metrics_nl_policy [ TCP_METRICS_ATTR_MAX + 1 ] = {
2012-09-04 15:03:15 +04:00
[ TCP_METRICS_ATTR_ADDR_IPV4 ] = { . type = NLA_U32 , } ,
[ TCP_METRICS_ATTR_ADDR_IPV6 ] = { . type = NLA_BINARY ,
. len = sizeof ( struct in6_addr ) , } ,
/* Following attributes are not received for GET/DEL,
* we keep them for reference
*/
#if 0
[ TCP_METRICS_ATTR_AGE ] = { . type = NLA_MSECS , } ,
[ TCP_METRICS_ATTR_TW_TSVAL ] = { . type = NLA_U32 , } ,
[ TCP_METRICS_ATTR_TW_TS_STAMP ] = { . type = NLA_S32 , } ,
[ TCP_METRICS_ATTR_VALS ] = { . type = NLA_NESTED , } ,
[ TCP_METRICS_ATTR_FOPEN_MSS ] = { . type = NLA_U16 , } ,
[ TCP_METRICS_ATTR_FOPEN_SYN_DROPS ] = { . type = NLA_U16 , } ,
[ TCP_METRICS_ATTR_FOPEN_SYN_DROP_TS ] = { . type = NLA_MSECS , } ,
[ TCP_METRICS_ATTR_FOPEN_COOKIE ] = { . type = NLA_BINARY ,
. len = TCP_FASTOPEN_COOKIE_MAX , } ,
# endif
} ;
/* Add attributes, caller cancels its header on failure */
static int tcp_metrics_fill_info ( struct sk_buff * msg ,
struct tcp_metrics_block * tm )
{
struct nlattr * nest ;
int i ;
2014-01-08 19:05:55 +04:00
switch ( tm - > tcpm_daddr . family ) {
2012-09-04 15:03:15 +04:00
case AF_INET :
2015-03-29 17:59:25 +03:00
if ( nla_put_in_addr ( msg , TCP_METRICS_ATTR_ADDR_IPV4 ,
2015-08-28 02:07:00 +03:00
inetpeer_get_addr_v4 ( & tm - > tcpm_daddr ) ) < 0 )
2012-09-04 15:03:15 +04:00
goto nla_put_failure ;
2015-03-29 17:59:25 +03:00
if ( nla_put_in_addr ( msg , TCP_METRICS_ATTR_SADDR_IPV4 ,
2015-08-28 02:07:00 +03:00
inetpeer_get_addr_v4 ( & tm - > tcpm_saddr ) ) < 0 )
2014-01-08 19:05:57 +04:00
goto nla_put_failure ;
2012-09-04 15:03:15 +04:00
break ;
case AF_INET6 :
2015-03-29 17:59:25 +03:00
if ( nla_put_in6_addr ( msg , TCP_METRICS_ATTR_ADDR_IPV6 ,
2015-08-28 02:07:00 +03:00
inetpeer_get_addr_v6 ( & tm - > tcpm_daddr ) ) < 0 )
2012-09-04 15:03:15 +04:00
goto nla_put_failure ;
2015-03-29 17:59:25 +03:00
if ( nla_put_in6_addr ( msg , TCP_METRICS_ATTR_SADDR_IPV6 ,
2015-08-28 02:07:00 +03:00
inetpeer_get_addr_v6 ( & tm - > tcpm_saddr ) ) < 0 )
2014-01-08 19:05:57 +04:00
goto nla_put_failure ;
2012-09-04 15:03:15 +04:00
break ;
default :
return - EAFNOSUPPORT ;
}
if ( nla_put_msecs ( msg , TCP_METRICS_ATTR_AGE ,
2016-04-22 18:31:21 +03:00
jiffies - tm - > tcpm_stamp ,
TCP_METRICS_ATTR_PAD ) < 0 )
2012-09-04 15:03:15 +04:00
goto nla_put_failure ;
{
int n = 0 ;
2019-04-26 12:13:06 +03:00
nest = nla_nest_start_noflag ( msg , TCP_METRICS_ATTR_VALS ) ;
2012-09-04 15:03:15 +04:00
if ( ! nest )
goto nla_put_failure ;
2014-02-27 02:02:48 +04:00
for ( i = 0 ; i < TCP_METRIC_MAX_KERNEL + 1 ; i + + ) {
u32 val = tm - > tcpm_vals [ i ] ;
if ( ! val )
2012-09-04 15:03:15 +04:00
continue ;
2014-02-27 02:02:48 +04:00
if ( i = = TCP_METRIC_RTT ) {
if ( nla_put_u32 ( msg , TCP_METRIC_RTT_US + 1 ,
val ) < 0 )
goto nla_put_failure ;
n + + ;
val = max ( val / 1000 , 1U ) ;
}
if ( i = = TCP_METRIC_RTTVAR ) {
if ( nla_put_u32 ( msg , TCP_METRIC_RTTVAR_US + 1 ,
val ) < 0 )
goto nla_put_failure ;
n + + ;
val = max ( val / 1000 , 1U ) ;
}
if ( nla_put_u32 ( msg , i + 1 , val ) < 0 )
2012-09-04 15:03:15 +04:00
goto nla_put_failure ;
n + + ;
}
if ( n )
nla_nest_end ( msg , nest ) ;
else
nla_nest_cancel ( msg , nest ) ;
}
{
struct tcp_fastopen_metrics tfom_copy [ 1 ] , * tfom ;
unsigned int seq ;
do {
seq = read_seqbegin ( & fastopen_seqlock ) ;
tfom_copy [ 0 ] = tm - > tcpm_fastopen ;
} while ( read_seqretry ( & fastopen_seqlock , seq ) ) ;
tfom = tfom_copy ;
if ( tfom - > mss & &
nla_put_u16 ( msg , TCP_METRICS_ATTR_FOPEN_MSS ,
tfom - > mss ) < 0 )
goto nla_put_failure ;
if ( tfom - > syn_loss & &
( nla_put_u16 ( msg , TCP_METRICS_ATTR_FOPEN_SYN_DROPS ,
tfom - > syn_loss ) < 0 | |
nla_put_msecs ( msg , TCP_METRICS_ATTR_FOPEN_SYN_DROP_TS ,
2016-04-22 18:31:21 +03:00
jiffies - tfom - > last_syn_loss ,
TCP_METRICS_ATTR_PAD ) < 0 ) )
2012-09-04 15:03:15 +04:00
goto nla_put_failure ;
if ( tfom - > cookie . len > 0 & &
nla_put ( msg , TCP_METRICS_ATTR_FOPEN_COOKIE ,
tfom - > cookie . len , tfom - > cookie . val ) < 0 )
goto nla_put_failure ;
}
return 0 ;
nla_put_failure :
return - EMSGSIZE ;
}
static int tcp_metrics_dump_info ( struct sk_buff * skb ,
struct netlink_callback * cb ,
struct tcp_metrics_block * tm )
{
void * hdr ;
2012-09-08 00:12:54 +04:00
hdr = genlmsg_put ( skb , NETLINK_CB ( cb - > skb ) . portid , cb - > nlh - > nlmsg_seq ,
2012-09-04 15:03:15 +04:00
& tcp_metrics_nl_family , NLM_F_MULTI ,
TCP_METRICS_CMD_GET ) ;
if ( ! hdr )
return - EMSGSIZE ;
if ( tcp_metrics_fill_info ( skb , tm ) < 0 )
goto nla_put_failure ;
2015-01-17 00:09:00 +03:00
genlmsg_end ( skb , hdr ) ;
return 0 ;
2012-09-04 15:03:15 +04:00
nla_put_failure :
genlmsg_cancel ( skb , hdr ) ;
return - EMSGSIZE ;
}
static int tcp_metrics_nl_dump ( struct sk_buff * skb ,
struct netlink_callback * cb )
{
struct net * net = sock_net ( skb - > sk ) ;
2015-03-13 08:07:44 +03:00
unsigned int max_rows = 1U < < tcp_metrics_hash_log ;
2012-09-04 15:03:15 +04:00
unsigned int row , s_row = cb - > args [ 0 ] ;
int s_col = cb - > args [ 1 ] , col = s_col ;
for ( row = s_row ; row < max_rows ; row + + , s_col = 0 ) {
struct tcp_metrics_block * tm ;
2015-03-13 08:07:44 +03:00
struct tcpm_hash_bucket * hb = tcp_metrics_hash + row ;
2012-09-04 15:03:15 +04:00
rcu_read_lock ( ) ;
for ( col = 0 , tm = rcu_dereference ( hb - > chain ) ; tm ;
tm = rcu_dereference ( tm - > tcpm_next ) , col + + ) {
2015-03-13 08:05:52 +03:00
if ( ! net_eq ( tm_net ( tm ) , net ) )
continue ;
2012-09-04 15:03:15 +04:00
if ( col < s_col )
continue ;
if ( tcp_metrics_dump_info ( skb , cb , tm ) < 0 ) {
rcu_read_unlock ( ) ;
goto done ;
}
}
rcu_read_unlock ( ) ;
}
done :
cb - > args [ 0 ] = row ;
cb - > args [ 1 ] = col ;
return skb - > len ;
}
2014-01-08 19:05:59 +04:00
static int __parse_nl_addr ( struct genl_info * info , struct inetpeer_addr * addr ,
unsigned int * hash , int optional , int v4 , int v6 )
2012-09-04 15:03:15 +04:00
{
struct nlattr * a ;
2014-01-08 19:05:59 +04:00
a = info - > attrs [ v4 ] ;
2012-09-04 15:03:15 +04:00
if ( a ) {
2015-08-28 02:07:00 +03:00
inetpeer_set_addr_v4 ( addr , nla_get_in_addr ( a ) ) ;
2014-01-08 19:05:59 +04:00
if ( hash )
2015-08-28 02:07:00 +03:00
* hash = ipv4_addr_hash ( inetpeer_get_addr_v4 ( addr ) ) ;
2012-09-04 15:03:15 +04:00
return 0 ;
}
2014-01-08 19:05:59 +04:00
a = info - > attrs [ v6 ] ;
2012-09-04 15:03:15 +04:00
if ( a ) {
2015-08-28 02:07:00 +03:00
struct in6_addr in6 ;
2012-10-30 16:03:09 +04:00
if ( nla_len ( a ) ! = sizeof ( struct in6_addr ) )
2012-09-04 15:03:15 +04:00
return - EINVAL ;
2015-08-28 02:07:00 +03:00
in6 = nla_get_in6_addr ( a ) ;
inetpeer_set_addr_v6 ( addr , & in6 ) ;
2014-01-08 19:05:59 +04:00
if ( hash )
2015-08-28 02:07:00 +03:00
* hash = ipv6_addr_hash ( inetpeer_get_addr_v6 ( addr ) ) ;
2012-09-04 15:03:15 +04:00
return 0 ;
}
return optional ? 1 : - EAFNOSUPPORT ;
}
2014-01-08 19:05:59 +04:00
static int parse_nl_addr ( struct genl_info * info , struct inetpeer_addr * addr ,
unsigned int * hash , int optional )
{
return __parse_nl_addr ( info , addr , hash , optional ,
TCP_METRICS_ATTR_ADDR_IPV4 ,
TCP_METRICS_ATTR_ADDR_IPV6 ) ;
}
static int parse_nl_saddr ( struct genl_info * info , struct inetpeer_addr * addr )
{
return __parse_nl_addr ( info , addr , NULL , 0 ,
TCP_METRICS_ATTR_SADDR_IPV4 ,
TCP_METRICS_ATTR_SADDR_IPV6 ) ;
}
2012-09-04 15:03:15 +04:00
static int tcp_metrics_nl_cmd_get ( struct sk_buff * skb , struct genl_info * info )
{
struct tcp_metrics_block * tm ;
2014-01-08 19:05:59 +04:00
struct inetpeer_addr saddr , daddr ;
2012-09-04 15:03:15 +04:00
unsigned int hash ;
struct sk_buff * msg ;
struct net * net = genl_info_net ( info ) ;
void * reply ;
int ret ;
2014-01-08 19:05:59 +04:00
bool src = true ;
2012-09-04 15:03:15 +04:00
2014-01-08 19:05:55 +04:00
ret = parse_nl_addr ( info , & daddr , & hash , 0 ) ;
2012-09-04 15:03:15 +04:00
if ( ret < 0 )
return ret ;
2014-01-08 19:05:59 +04:00
ret = parse_nl_saddr ( info , & saddr ) ;
if ( ret < 0 )
src = false ;
2012-09-04 15:03:15 +04:00
msg = nlmsg_new ( NLMSG_DEFAULT_SIZE , GFP_KERNEL ) ;
if ( ! msg )
return - ENOMEM ;
reply = genlmsg_put_reply ( msg , info , & tcp_metrics_nl_family , 0 ,
info - > genlhdr - > cmd ) ;
if ( ! reply )
goto nla_put_failure ;
2015-03-13 08:05:24 +03:00
hash ^ = net_hash_mix ( net ) ;
2015-03-13 08:07:44 +03:00
hash = hash_32 ( hash , tcp_metrics_hash_log ) ;
2012-09-04 15:03:15 +04:00
ret = - ESRCH ;
rcu_read_lock ( ) ;
2015-03-13 08:07:44 +03:00
for ( tm = rcu_dereference ( tcp_metrics_hash [ hash ] . chain ) ; tm ;
2012-09-04 15:03:15 +04:00
tm = rcu_dereference ( tm - > tcpm_next ) ) {
2014-01-08 19:05:59 +04:00
if ( addr_same ( & tm - > tcpm_daddr , & daddr ) & &
2015-03-13 08:05:52 +03:00
( ! src | | addr_same ( & tm - > tcpm_saddr , & saddr ) ) & &
net_eq ( tm_net ( tm ) , net ) ) {
2012-09-04 15:03:15 +04:00
ret = tcp_metrics_fill_info ( msg , tm ) ;
break ;
}
}
rcu_read_unlock ( ) ;
if ( ret < 0 )
goto out_free ;
genlmsg_end ( msg , reply ) ;
return genlmsg_reply ( msg , info ) ;
nla_put_failure :
ret = - EMSGSIZE ;
out_free :
nlmsg_free ( msg ) ;
return ret ;
}
2015-03-13 08:06:43 +03:00
static void tcp_metrics_flush_all ( struct net * net )
2012-09-04 15:03:15 +04:00
{
2015-03-13 08:07:44 +03:00
unsigned int max_rows = 1U < < tcp_metrics_hash_log ;
struct tcpm_hash_bucket * hb = tcp_metrics_hash ;
2012-09-04 15:03:15 +04:00
struct tcp_metrics_block * tm ;
unsigned int row ;
for ( row = 0 ; row < max_rows ; row + + , hb + + ) {
2015-03-13 08:07:10 +03:00
struct tcp_metrics_block __rcu * * pp ;
2017-09-20 02:27:07 +03:00
bool match ;
2012-09-04 15:03:15 +04:00
spin_lock_bh ( & tcp_metrics_lock ) ;
2015-03-13 08:07:10 +03:00
pp = & hb - > chain ;
2015-03-16 17:14:34 +03:00
for ( tm = deref_locked ( * pp ) ; tm ; tm = deref_locked ( * pp ) ) {
2017-09-20 02:27:07 +03:00
match = net ? net_eq ( tm_net ( tm ) , net ) :
2018-01-12 18:28:31 +03:00
! refcount_read ( & tm_net ( tm ) - > count ) ;
2017-09-20 02:27:07 +03:00
if ( match ) {
2015-03-13 08:07:10 +03:00
* pp = tm - > tcpm_next ;
kfree_rcu ( tm , rcu_head ) ;
} else {
pp = & tm - > tcpm_next ;
}
2012-09-04 15:03:15 +04:00
}
2015-03-13 08:07:10 +03:00
spin_unlock_bh ( & tcp_metrics_lock ) ;
2012-09-04 15:03:15 +04:00
}
}
static int tcp_metrics_nl_cmd_del ( struct sk_buff * skb , struct genl_info * info )
{
struct tcpm_hash_bucket * hb ;
2014-01-21 16:30:26 +04:00
struct tcp_metrics_block * tm ;
2012-09-04 15:03:15 +04:00
struct tcp_metrics_block __rcu * * pp ;
2014-01-08 19:05:59 +04:00
struct inetpeer_addr saddr , daddr ;
2012-09-04 15:03:15 +04:00
unsigned int hash ;
struct net * net = genl_info_net ( info ) ;
int ret ;
2014-01-21 16:30:26 +04:00
bool src = true , found = false ;
2012-09-04 15:03:15 +04:00
2014-01-08 19:05:55 +04:00
ret = parse_nl_addr ( info , & daddr , & hash , 1 ) ;
2012-09-04 15:03:15 +04:00
if ( ret < 0 )
return ret ;
2015-03-13 08:06:43 +03:00
if ( ret > 0 ) {
tcp_metrics_flush_all ( net ) ;
return 0 ;
}
2014-01-08 19:05:59 +04:00
ret = parse_nl_saddr ( info , & saddr ) ;
if ( ret < 0 )
src = false ;
2012-09-04 15:03:15 +04:00
2015-03-13 08:05:24 +03:00
hash ^ = net_hash_mix ( net ) ;
2015-03-13 08:07:44 +03:00
hash = hash_32 ( hash , tcp_metrics_hash_log ) ;
hb = tcp_metrics_hash + hash ;
2012-09-04 15:03:15 +04:00
pp = & hb - > chain ;
spin_lock_bh ( & tcp_metrics_lock ) ;
2015-03-16 17:14:34 +03:00
for ( tm = deref_locked ( * pp ) ; tm ; tm = deref_locked ( * pp ) ) {
2014-01-08 19:05:59 +04:00
if ( addr_same ( & tm - > tcpm_daddr , & daddr ) & &
2015-03-13 08:05:52 +03:00
( ! src | | addr_same ( & tm - > tcpm_saddr , & saddr ) ) & &
net_eq ( tm_net ( tm ) , net ) ) {
2012-09-04 15:03:15 +04:00
* pp = tm - > tcpm_next ;
2014-01-21 16:30:26 +04:00
kfree_rcu ( tm , rcu_head ) ;
found = true ;
2014-01-08 19:05:58 +04:00
} else {
pp = & tm - > tcpm_next ;
2012-09-04 15:03:15 +04:00
}
}
spin_unlock_bh ( & tcp_metrics_lock ) ;
2014-01-21 16:30:26 +04:00
if ( ! found )
2012-09-04 15:03:15 +04:00
return - ESRCH ;
return 0 ;
}
2013-11-14 20:14:46 +04:00
static const struct genl_ops tcp_metrics_nl_ops [ ] = {
2012-09-04 15:03:15 +04:00
{
. cmd = TCP_METRICS_CMD_GET ,
2019-04-26 15:07:31 +03:00
. validate = GENL_DONT_VALIDATE_STRICT | GENL_DONT_VALIDATE_DUMP ,
2012-09-04 15:03:15 +04:00
. doit = tcp_metrics_nl_cmd_get ,
. dumpit = tcp_metrics_nl_dump ,
} ,
{
. cmd = TCP_METRICS_CMD_DEL ,
2019-04-26 15:07:31 +03:00
. validate = GENL_DONT_VALIDATE_STRICT | GENL_DONT_VALIDATE_DUMP ,
2012-09-04 15:03:15 +04:00
. doit = tcp_metrics_nl_cmd_del ,
. flags = GENL_ADMIN_PERM ,
} ,
} ;
2016-10-24 15:40:05 +03:00
static struct genl_family tcp_metrics_nl_family __ro_after_init = {
2016-10-24 15:40:03 +03:00
. hdrsize = 0 ,
. name = TCP_METRICS_GENL_NAME ,
. version = TCP_METRICS_GENL_VERSION ,
. maxattr = TCP_METRICS_ATTR_MAX ,
genetlink: make policy common to family
Since maxattr is common, the policy can't really differ sanely,
so make it common as well.
The only user that did in fact manage to make a non-common policy
is taskstats, which has to be really careful about it (since it's
still using a common maxattr!). This is no longer supported, but
we can fake it using pre_doit.
This reduces the size of e.g. nl80211.o (which has lots of commands):
text data bss dec hex filename
398745 14323 2240 415308 6564c net/wireless/nl80211.o (before)
397913 14331 2240 414484 65314 net/wireless/nl80211.o (after)
--------------------------------
-832 +8 0 -824
Which is obviously just 8 bytes for each command, and an added 8
bytes for the new policy pointer. I'm not sure why the ops list is
counted as .text though.
Most of the code transformations were done using the following spatch:
@ops@
identifier OPS;
expression POLICY;
@@
struct genl_ops OPS[] = {
...,
{
- .policy = POLICY,
},
...
};
@@
identifier ops.OPS;
expression ops.POLICY;
identifier fam;
expression M;
@@
struct genl_family fam = {
.ops = OPS,
.maxattr = M,
+ .policy = POLICY,
...
};
This also gets rid of devlink_nl_cmd_region_read_dumpit() accessing
the cb->data as ops, which we want to change in a later genl patch.
Signed-off-by: Johannes Berg <johannes.berg@intel.com>
Signed-off-by: David S. Miller <davem@davemloft.net>
2019-03-22 00:51:02 +03:00
. policy = tcp_metrics_nl_policy ,
2016-10-24 15:40:03 +03:00
. netnsok = true ,
. module = THIS_MODULE ,
. ops = tcp_metrics_nl_ops ,
. n_ops = ARRAY_SIZE ( tcp_metrics_nl_ops ) ,
} ;
2012-07-20 03:02:34 +04:00
static unsigned int tcpmhash_entries ;
2012-07-10 11:49:14 +04:00
static int __init set_tcpmhash_entries ( char * str )
{
ssize_t ret ;
if ( ! str )
return 0 ;
2012-07-20 03:02:34 +04:00
ret = kstrtouint ( str , 0 , & tcpmhash_entries ) ;
2012-07-10 11:49:14 +04:00
if ( ret )
return 0 ;
return 1 ;
}
__setup ( " tcpmhash_entries= " , set_tcpmhash_entries ) ;
static int __net_init tcp_net_metrics_init ( struct net * net )
{
2012-07-20 03:02:34 +04:00
size_t size ;
unsigned int slots ;
2012-07-10 11:49:14 +04:00
2015-03-13 08:07:44 +03:00
if ( ! net_eq ( net , & init_net ) )
return 0 ;
2012-07-10 11:49:14 +04:00
slots = tcpmhash_entries ;
if ( ! slots ) {
2018-12-28 11:34:29 +03:00
if ( totalram_pages ( ) > = 128 * 1024 )
2012-07-10 11:49:14 +04:00
slots = 16 * 1024 ;
else
slots = 8 * 1024 ;
}
2015-03-13 08:07:44 +03:00
tcp_metrics_hash_log = order_base_2 ( slots ) ;
size = sizeof ( struct tcpm_hash_bucket ) < < tcp_metrics_hash_log ;
2012-07-10 11:49:14 +04:00
2017-05-09 01:57:27 +03:00
tcp_metrics_hash = kvzalloc ( size , GFP_KERNEL ) ;
2015-03-13 08:07:44 +03:00
if ( ! tcp_metrics_hash )
2012-07-10 11:49:14 +04:00
return - ENOMEM ;
return 0 ;
}
2017-09-20 02:27:07 +03:00
static void __net_exit tcp_net_metrics_exit_batch ( struct list_head * net_exit_list )
2012-07-10 11:49:14 +04:00
{
2017-09-20 02:27:07 +03:00
tcp_metrics_flush_all ( NULL ) ;
2012-07-10 11:49:14 +04:00
}
static __net_initdata struct pernet_operations tcp_net_metrics_ops = {
2017-09-20 02:27:07 +03:00
. init = tcp_net_metrics_init ,
. exit_batch = tcp_net_metrics_exit_batch ,
2012-07-10 11:49:14 +04:00
} ;
void __init tcp_metrics_init ( void )
{
2012-09-04 15:03:15 +04:00
int ret ;
ret = register_pernet_subsys ( & tcp_net_metrics_ops ) ;
if ( ret < 0 )
2015-03-13 08:04:51 +03:00
panic ( " Could not allocate the tcp_metrics hash table \n " ) ;
2016-10-24 15:40:03 +03:00
ret = genl_register_family ( & tcp_metrics_nl_family ) ;
2012-09-04 15:03:15 +04:00
if ( ret < 0 )
2015-03-13 08:04:51 +03:00
panic ( " Could not register tcp_metrics generic netlink \n " ) ;
2012-07-10 11:49:14 +04:00
}