7e68dd7d07
Core ---- - Allow live renaming when an interface is up - Add retpoline wrappers for tc, improving considerably the performances of complex queue discipline configurations. - Add inet drop monitor support. - A few GRO performance improvements. - Add infrastructure for atomic dev stats, addressing long standing data races. - De-duplicate common code between OVS and conntrack offloading infrastructure. - A bunch of UBSAN_BOUNDS/FORTIFY_SOURCE improvements. - Netfilter: introduce packet parser for tunneled packets - Replace IPVS timer-based estimators with kthreads to scale up the workload with the number of available CPUs. - Add the helper support for connection-tracking OVS offload. BPF --- - Support for user defined BPF objects: the use case is to allocate own objects, build own object hierarchies and use the building blocks to build own data structures flexibly, for example, linked lists in BPF. - Make cgroup local storage available to non-cgroup attached BPF programs. - Avoid unnecessary deadlock detection and failures wrt BPF task storage helpers. - A relevant bunch of BPF verifier fixes and improvements. - Veristat tool improvements to support custom filtering, sorting, and replay of results. - Add LLVM disassembler as default library for dumping JITed code. - Lots of new BPF documentation for various BPF maps. - Add bpf_rcu_read_{,un}lock() support for sleepable programs. - Add RCU grace period chaining to BPF to wait for the completion of access from both sleepable and non-sleepable BPF programs. - Add support storing struct task_struct objects as kptrs in maps. - Improve helper UAPI by explicitly defining BPF_FUNC_xxx integer values. - Add libbpf *_opts API-variants for bpf_*_get_fd_by_id() functions. Protocols --------- - TCP: implement Protective Load Balancing across switch links. - TCP: allow dynamically disabling TCP-MD5 static key, reverting back to fast[er]-path. - UDP: Introduce optional per-netns hash lookup table. - IPv6: simplify and cleanup sockets disposal. - Netlink: support different type policies for each generic netlink operation. - MPTCP: add MSG_FASTOPEN and FastOpen listener side support. - MPTCP: add netlink notification support for listener sockets events. - SCTP: add VRF support, allowing sctp sockets binding to VRF devices. - Add bridging MAC Authentication Bypass (MAB) support. - Extensions for Ethernet VPN bridging implementation to better support multicast scenarios. - More work for Wi-Fi 7 support, comprising conversion of all the existing drivers to internal TX queue usage. - IPSec: introduce a new offload type (packet offload) allowing complete header processing and crypto offloading. - IPSec: extended ack support for more descriptive XFRM error reporting. - RXRPC: increase SACK table size and move processing into a per-local endpoint kernel thread, reducing considerably the required locking. - IEEE 802154: synchronous send frame and extended filtering support, initial support for scanning available 15.4 networks. - Tun: bump the link speed from 10Mbps to 10Gbps. - Tun/VirtioNet: implement UDP segmentation offload support. Driver API ---------- - PHY/SFP: improve power level switching between standard level 1 and the higher power levels. - New API for netdev <-> devlink_port linkage. - PTP: convert existing drivers to new frequency adjustment implementation. - DSA: add support for rx offloading. - Autoload DSA tagging driver when dynamically changing protocol. - Add new PCP and APPTRUST attributes to Data Center Bridging. - Add configuration support for 800Gbps link speed. - Add devlink port function attribute to enable/disable RoCE and migratable. - Extend devlink-rate to support strict prioriry and weighted fair queuing. - Add devlink support to directly reading from region memory. - New device tree helper to fetch MAC address from nvmem. - New big TCP helper to simplify temporary header stripping. New hardware / drivers ---------------------- - Ethernet: - Marvel Octeon CNF95N and CN10KB Ethernet Switches. - Marvel Prestera AC5X Ethernet Switch. - WangXun 10 Gigabit NIC. - Motorcomm yt8521 Gigabit Ethernet. - Microchip ksz9563 Gigabit Ethernet Switch. - Microsoft Azure Network Adapter. - Linux Automation 10Base-T1L adapter. - PHY: - Aquantia AQR112 and AQR412. - Motorcomm YT8531S. - PTP: - Orolia ART-CARD. - WiFi: - MediaTek Wi-Fi 7 (802.11be) devices. - RealTek rtw8821cu, rtw8822bu, rtw8822cu and rtw8723du USB devices. - Bluetooth: - Broadcom BCM4377/4378/4387 Bluetooth chipsets. - Realtek RTL8852BE and RTL8723DS. - Cypress.CYW4373A0 WiFi + Bluetooth combo device. Drivers ------- - CAN: - gs_usb: bus error reporting support. - kvaser_usb: listen only and bus error reporting support. - Ethernet NICs: - Intel (100G): - extend action skbedit to RX queue mapping. - implement devlink-rate support. - support direct read from memory. - nVidia/Mellanox (mlx5): - SW steering improvements, increasing rules update rate. - Support for enhanced events compression. - extend H/W offload packet manipulation capabilities. - implement IPSec packet offload mode. - nVidia/Mellanox (mlx4): - better big TCP support. - Netronome Ethernet NICs (nfp): - IPsec offload support. - add support for multicast filter. - Broadcom: - RSS and PTP support improvements. - AMD/SolarFlare: - netlink extened ack improvements. - add basic flower matches to offload, and related stats. - Virtual NICs: - ibmvnic: introduce affinity hint support. - small / embedded: - FreeScale fec: add initial XDP support. - Marvel mv643xx_eth: support MII/GMII/RGMII modes for Kirkwood. - TI am65-cpsw: add suspend/resume support. - Mediatek MT7986: add RX wireless wthernet dispatch support. - Realtek 8169: enable GRO software interrupt coalescing per default. - Ethernet high-speed switches: - Microchip (sparx5): - add support for Sparx5 TC/flower H/W offload via VCAP. - Mellanox mlxsw: - add 802.1X and MAC Authentication Bypass offload support. - add ip6gre support. - Embedded Ethernet switches: - Mediatek (mtk_eth_soc): - improve PCS implementation, add DSA untag support. - enable flow offload support. - Renesas: - add rswitch R-Car Gen4 gPTP support. - Microchip (lan966x): - add full XDP support. - add TC H/W offload via VCAP. - enable PTP on bridge interfaces. - Microchip (ksz8): - add MTU support for KSZ8 series. - Qualcomm 802.11ax WiFi (ath11k): - support configuring channel dwell time during scan. - MediaTek WiFi (mt76): - enable Wireless Ethernet Dispatch (WED) offload support. - add ack signal support. - enable coredump support. - remain_on_channel support. - Intel WiFi (iwlwifi): - enable Wi-Fi 7 Extremely High Throughput (EHT) PHY capabilities. - 320 MHz channels support. - RealTek WiFi (rtw89): - new dynamic header firmware format support. - wake-over-WLAN support. Signed-off-by: Paolo Abeni <pabeni@redhat.com> -----BEGIN PGP SIGNATURE----- iQJGBAABCAAwFiEEg1AjqC77wbdLX2LbKSR5jcyPE6QFAmOYXUcSHHBhYmVuaUBy ZWRoYXQuY29tAAoJECkkeY3MjxOk8zQP/R7BZtbJMTPiWkRnSoKHnAyupDVwrz5U ktukLkwPsCyJuEbAjgxrxf4EEEQ9uq2FFlxNSYuKiiQMqIpFxV6KED7LCUygn4Tc kxtkp0Q+5XiqisWlQmtfExf2OjuuPqcjV9tWCDBI6GebKUbfNwY/eI44RcMu4BSv DzIlW5GkX/kZAPqnnuqaLsN3FudDTJHGEAD7NbA++7wJ076RWYSLXlFv0Z+SCSPS H8/PEG0/ZK/65rIWMAFRClJ9BNIDwGVgp0GrsIvs1gqbRUOlA1hl1rDM21TqtNFf 5QPQT7sIfTcCE/nerxKJD5JE3JyP+XRlRn96PaRw3rt4MgI6I/EOj/HOKQ5tMCNc oPiqb7N70+hkLZyr42qX+vN9eDPjp2koEQm7EO2Zs+/534/zWDs24Zfk/Aa1ps0I Fa82oGjAgkBhGe/FZ6i5cYoLcyxqRqZV1Ws9XQMl72qRC7/BwvNbIW6beLpCRyeM yYIU+0e9dEm+wHQEdh2niJuVtR63hy8tvmPx56lyh+6u0+pondkwbfSiC5aD3kAC ikKsN5DyEsdXyiBAlytCEBxnaOjQy4RAz+3YXSiS0eBNacXp03UUrNGx4Pzpu/D0 QLFJhBnMFFCgy5to8/DvKnrTPgZdSURwqbIUcZdvU21f1HLR8tUTpaQnYffc/Whm V8gnt1EL+0cc =CbJC -----END PGP SIGNATURE----- Merge tag 'net-next-6.2' of git://git.kernel.org/pub/scm/linux/kernel/git/netdev/net-next Pull networking updates from Paolo Abeni: "Core: - Allow live renaming when an interface is up - Add retpoline wrappers for tc, improving considerably the performances of complex queue discipline configurations - Add inet drop monitor support - A few GRO performance improvements - Add infrastructure for atomic dev stats, addressing long standing data races - De-duplicate common code between OVS and conntrack offloading infrastructure - A bunch of UBSAN_BOUNDS/FORTIFY_SOURCE improvements - Netfilter: introduce packet parser for tunneled packets - Replace IPVS timer-based estimators with kthreads to scale up the workload with the number of available CPUs - Add the helper support for connection-tracking OVS offload BPF: - Support for user defined BPF objects: the use case is to allocate own objects, build own object hierarchies and use the building blocks to build own data structures flexibly, for example, linked lists in BPF - Make cgroup local storage available to non-cgroup attached BPF programs - Avoid unnecessary deadlock detection and failures wrt BPF task storage helpers - A relevant bunch of BPF verifier fixes and improvements - Veristat tool improvements to support custom filtering, sorting, and replay of results - Add LLVM disassembler as default library for dumping JITed code - Lots of new BPF documentation for various BPF maps - Add bpf_rcu_read_{,un}lock() support for sleepable programs - Add RCU grace period chaining to BPF to wait for the completion of access from both sleepable and non-sleepable BPF programs - Add support storing struct task_struct objects as kptrs in maps - Improve helper UAPI by explicitly defining BPF_FUNC_xxx integer values - Add libbpf *_opts API-variants for bpf_*_get_fd_by_id() functions Protocols: - TCP: implement Protective Load Balancing across switch links - TCP: allow dynamically disabling TCP-MD5 static key, reverting back to fast[er]-path - UDP: Introduce optional per-netns hash lookup table - IPv6: simplify and cleanup sockets disposal - Netlink: support different type policies for each generic netlink operation - MPTCP: add MSG_FASTOPEN and FastOpen listener side support - MPTCP: add netlink notification support for listener sockets events - SCTP: add VRF support, allowing sctp sockets binding to VRF devices - Add bridging MAC Authentication Bypass (MAB) support - Extensions for Ethernet VPN bridging implementation to better support multicast scenarios - More work for Wi-Fi 7 support, comprising conversion of all the existing drivers to internal TX queue usage - IPSec: introduce a new offload type (packet offload) allowing complete header processing and crypto offloading - IPSec: extended ack support for more descriptive XFRM error reporting - RXRPC: increase SACK table size and move processing into a per-local endpoint kernel thread, reducing considerably the required locking - IEEE 802154: synchronous send frame and extended filtering support, initial support for scanning available 15.4 networks - Tun: bump the link speed from 10Mbps to 10Gbps - Tun/VirtioNet: implement UDP segmentation offload support Driver API: - PHY/SFP: improve power level switching between standard level 1 and the higher power levels - New API for netdev <-> devlink_port linkage - PTP: convert existing drivers to new frequency adjustment implementation - DSA: add support for rx offloading - Autoload DSA tagging driver when dynamically changing protocol - Add new PCP and APPTRUST attributes to Data Center Bridging - Add configuration support for 800Gbps link speed - Add devlink port function attribute to enable/disable RoCE and migratable - Extend devlink-rate to support strict prioriry and weighted fair queuing - Add devlink support to directly reading from region memory - New device tree helper to fetch MAC address from nvmem - New big TCP helper to simplify temporary header stripping New hardware / drivers: - Ethernet: - Marvel Octeon CNF95N and CN10KB Ethernet Switches - Marvel Prestera AC5X Ethernet Switch - WangXun 10 Gigabit NIC - Motorcomm yt8521 Gigabit Ethernet - Microchip ksz9563 Gigabit Ethernet Switch - Microsoft Azure Network Adapter - Linux Automation 10Base-T1L adapter - PHY: - Aquantia AQR112 and AQR412 - Motorcomm YT8531S - PTP: - Orolia ART-CARD - WiFi: - MediaTek Wi-Fi 7 (802.11be) devices - RealTek rtw8821cu, rtw8822bu, rtw8822cu and rtw8723du USB devices - Bluetooth: - Broadcom BCM4377/4378/4387 Bluetooth chipsets - Realtek RTL8852BE and RTL8723DS - Cypress.CYW4373A0 WiFi + Bluetooth combo device Drivers: - CAN: - gs_usb: bus error reporting support - kvaser_usb: listen only and bus error reporting support - Ethernet NICs: - Intel (100G): - extend action skbedit to RX queue mapping - implement devlink-rate support - support direct read from memory - nVidia/Mellanox (mlx5): - SW steering improvements, increasing rules update rate - Support for enhanced events compression - extend H/W offload packet manipulation capabilities - implement IPSec packet offload mode - nVidia/Mellanox (mlx4): - better big TCP support - Netronome Ethernet NICs (nfp): - IPsec offload support - add support for multicast filter - Broadcom: - RSS and PTP support improvements - AMD/SolarFlare: - netlink extened ack improvements - add basic flower matches to offload, and related stats - Virtual NICs: - ibmvnic: introduce affinity hint support - small / embedded: - FreeScale fec: add initial XDP support - Marvel mv643xx_eth: support MII/GMII/RGMII modes for Kirkwood - TI am65-cpsw: add suspend/resume support - Mediatek MT7986: add RX wireless wthernet dispatch support - Realtek 8169: enable GRO software interrupt coalescing per default - Ethernet high-speed switches: - Microchip (sparx5): - add support for Sparx5 TC/flower H/W offload via VCAP - Mellanox mlxsw: - add 802.1X and MAC Authentication Bypass offload support - add ip6gre support - Embedded Ethernet switches: - Mediatek (mtk_eth_soc): - improve PCS implementation, add DSA untag support - enable flow offload support - Renesas: - add rswitch R-Car Gen4 gPTP support - Microchip (lan966x): - add full XDP support - add TC H/W offload via VCAP - enable PTP on bridge interfaces - Microchip (ksz8): - add MTU support for KSZ8 series - Qualcomm 802.11ax WiFi (ath11k): - support configuring channel dwell time during scan - MediaTek WiFi (mt76): - enable Wireless Ethernet Dispatch (WED) offload support - add ack signal support - enable coredump support - remain_on_channel support - Intel WiFi (iwlwifi): - enable Wi-Fi 7 Extremely High Throughput (EHT) PHY capabilities - 320 MHz channels support - RealTek WiFi (rtw89): - new dynamic header firmware format support - wake-over-WLAN support" * tag 'net-next-6.2' of git://git.kernel.org/pub/scm/linux/kernel/git/netdev/net-next: (2002 commits) ipvs: fix type warning in do_div() on 32 bit net: lan966x: Remove a useless test in lan966x_ptp_add_trap() net: ipa: add IPA v4.7 support dt-bindings: net: qcom,ipa: Add SM6350 compatible bnxt: Use generic HBH removal helper in tx path IPv6/GRO: generic helper to remove temporary HBH/jumbo header in driver selftests: forwarding: Add bridge MDB test selftests: forwarding: Rename bridge_mdb test bridge: mcast: Support replacement of MDB port group entries bridge: mcast: Allow user space to specify MDB entry routing protocol bridge: mcast: Allow user space to add (*, G) with a source list and filter mode bridge: mcast: Add support for (*, G) with a source list and filter mode bridge: mcast: Avoid arming group timer when (S, G) corresponds to a source bridge: mcast: Add a flag for user installed source entries bridge: mcast: Expose __br_multicast_del_group_src() bridge: mcast: Expose br_multicast_new_group_src() bridge: mcast: Add a centralized error path bridge: mcast: Place netlink policy before validation functions bridge: mcast: Split (*, G) and (S, G) addition into different functions bridge: mcast: Do not derive entry type from its filter mode ...
814 lines
20 KiB
C
814 lines
20 KiB
C
// SPDX-License-Identifier: GPL-2.0-only
|
|
/*
|
|
* Resizable, Scalable, Concurrent Hash Table
|
|
*
|
|
* Copyright (c) 2014-2015 Thomas Graf <tgraf@suug.ch>
|
|
* Copyright (c) 2008-2014 Patrick McHardy <kaber@trash.net>
|
|
*/
|
|
|
|
/**************************************************************************
|
|
* Self Test
|
|
**************************************************************************/
|
|
|
|
#include <linux/init.h>
|
|
#include <linux/jhash.h>
|
|
#include <linux/kernel.h>
|
|
#include <linux/kthread.h>
|
|
#include <linux/module.h>
|
|
#include <linux/rcupdate.h>
|
|
#include <linux/rhashtable.h>
|
|
#include <linux/slab.h>
|
|
#include <linux/sched.h>
|
|
#include <linux/random.h>
|
|
#include <linux/vmalloc.h>
|
|
#include <linux/wait.h>
|
|
|
|
#define MAX_ENTRIES 1000000
|
|
#define TEST_INSERT_FAIL INT_MAX
|
|
|
|
static int parm_entries = 50000;
|
|
module_param(parm_entries, int, 0);
|
|
MODULE_PARM_DESC(parm_entries, "Number of entries to add (default: 50000)");
|
|
|
|
static int runs = 4;
|
|
module_param(runs, int, 0);
|
|
MODULE_PARM_DESC(runs, "Number of test runs per variant (default: 4)");
|
|
|
|
static int max_size = 0;
|
|
module_param(max_size, int, 0);
|
|
MODULE_PARM_DESC(max_size, "Maximum table size (default: calculated)");
|
|
|
|
static bool shrinking = false;
|
|
module_param(shrinking, bool, 0);
|
|
MODULE_PARM_DESC(shrinking, "Enable automatic shrinking (default: off)");
|
|
|
|
static int size = 8;
|
|
module_param(size, int, 0);
|
|
MODULE_PARM_DESC(size, "Initial size hint of table (default: 8)");
|
|
|
|
static int tcount = 10;
|
|
module_param(tcount, int, 0);
|
|
MODULE_PARM_DESC(tcount, "Number of threads to spawn (default: 10)");
|
|
|
|
static bool enomem_retry = false;
|
|
module_param(enomem_retry, bool, 0);
|
|
MODULE_PARM_DESC(enomem_retry, "Retry insert even if -ENOMEM was returned (default: off)");
|
|
|
|
struct test_obj_val {
|
|
int id;
|
|
int tid;
|
|
};
|
|
|
|
struct test_obj {
|
|
struct test_obj_val value;
|
|
struct rhash_head node;
|
|
};
|
|
|
|
struct test_obj_rhl {
|
|
struct test_obj_val value;
|
|
struct rhlist_head list_node;
|
|
};
|
|
|
|
struct thread_data {
|
|
unsigned int entries;
|
|
int id;
|
|
struct task_struct *task;
|
|
struct test_obj *objs;
|
|
};
|
|
|
|
static u32 my_hashfn(const void *data, u32 len, u32 seed)
|
|
{
|
|
const struct test_obj_rhl *obj = data;
|
|
|
|
return (obj->value.id % 10);
|
|
}
|
|
|
|
static int my_cmpfn(struct rhashtable_compare_arg *arg, const void *obj)
|
|
{
|
|
const struct test_obj_rhl *test_obj = obj;
|
|
const struct test_obj_val *val = arg->key;
|
|
|
|
return test_obj->value.id - val->id;
|
|
}
|
|
|
|
static struct rhashtable_params test_rht_params = {
|
|
.head_offset = offsetof(struct test_obj, node),
|
|
.key_offset = offsetof(struct test_obj, value),
|
|
.key_len = sizeof(struct test_obj_val),
|
|
.hashfn = jhash,
|
|
};
|
|
|
|
static struct rhashtable_params test_rht_params_dup = {
|
|
.head_offset = offsetof(struct test_obj_rhl, list_node),
|
|
.key_offset = offsetof(struct test_obj_rhl, value),
|
|
.key_len = sizeof(struct test_obj_val),
|
|
.hashfn = jhash,
|
|
.obj_hashfn = my_hashfn,
|
|
.obj_cmpfn = my_cmpfn,
|
|
.nelem_hint = 128,
|
|
.automatic_shrinking = false,
|
|
};
|
|
|
|
static atomic_t startup_count;
|
|
static DECLARE_WAIT_QUEUE_HEAD(startup_wait);
|
|
|
|
static int insert_retry(struct rhashtable *ht, struct test_obj *obj,
|
|
const struct rhashtable_params params)
|
|
{
|
|
int err, retries = -1, enomem_retries = 0;
|
|
|
|
do {
|
|
retries++;
|
|
cond_resched();
|
|
err = rhashtable_insert_fast(ht, &obj->node, params);
|
|
if (err == -ENOMEM && enomem_retry) {
|
|
enomem_retries++;
|
|
err = -EBUSY;
|
|
}
|
|
} while (err == -EBUSY);
|
|
|
|
if (enomem_retries)
|
|
pr_info(" %u insertions retried after -ENOMEM\n",
|
|
enomem_retries);
|
|
|
|
return err ? : retries;
|
|
}
|
|
|
|
static int __init test_rht_lookup(struct rhashtable *ht, struct test_obj *array,
|
|
unsigned int entries)
|
|
{
|
|
unsigned int i;
|
|
|
|
for (i = 0; i < entries; i++) {
|
|
struct test_obj *obj;
|
|
bool expected = !(i % 2);
|
|
struct test_obj_val key = {
|
|
.id = i,
|
|
};
|
|
|
|
if (array[i / 2].value.id == TEST_INSERT_FAIL)
|
|
expected = false;
|
|
|
|
obj = rhashtable_lookup_fast(ht, &key, test_rht_params);
|
|
|
|
if (expected && !obj) {
|
|
pr_warn("Test failed: Could not find key %u\n", key.id);
|
|
return -ENOENT;
|
|
} else if (!expected && obj) {
|
|
pr_warn("Test failed: Unexpected entry found for key %u\n",
|
|
key.id);
|
|
return -EEXIST;
|
|
} else if (expected && obj) {
|
|
if (obj->value.id != i) {
|
|
pr_warn("Test failed: Lookup value mismatch %u!=%u\n",
|
|
obj->value.id, i);
|
|
return -EINVAL;
|
|
}
|
|
}
|
|
|
|
cond_resched_rcu();
|
|
}
|
|
|
|
return 0;
|
|
}
|
|
|
|
static void test_bucket_stats(struct rhashtable *ht, unsigned int entries)
|
|
{
|
|
unsigned int total = 0, chain_len = 0;
|
|
struct rhashtable_iter hti;
|
|
struct rhash_head *pos;
|
|
|
|
rhashtable_walk_enter(ht, &hti);
|
|
rhashtable_walk_start(&hti);
|
|
|
|
while ((pos = rhashtable_walk_next(&hti))) {
|
|
if (PTR_ERR(pos) == -EAGAIN) {
|
|
pr_info("Info: encountered resize\n");
|
|
chain_len++;
|
|
continue;
|
|
} else if (IS_ERR(pos)) {
|
|
pr_warn("Test failed: rhashtable_walk_next() error: %ld\n",
|
|
PTR_ERR(pos));
|
|
break;
|
|
}
|
|
|
|
total++;
|
|
}
|
|
|
|
rhashtable_walk_stop(&hti);
|
|
rhashtable_walk_exit(&hti);
|
|
|
|
pr_info(" Traversal complete: counted=%u, nelems=%u, entries=%d, table-jumps=%u\n",
|
|
total, atomic_read(&ht->nelems), entries, chain_len);
|
|
|
|
if (total != atomic_read(&ht->nelems) || total != entries)
|
|
pr_warn("Test failed: Total count mismatch ^^^");
|
|
}
|
|
|
|
static s64 __init test_rhashtable(struct rhashtable *ht, struct test_obj *array,
|
|
unsigned int entries)
|
|
{
|
|
struct test_obj *obj;
|
|
int err;
|
|
unsigned int i, insert_retries = 0;
|
|
s64 start, end;
|
|
|
|
/*
|
|
* Insertion Test:
|
|
* Insert entries into table with all keys even numbers
|
|
*/
|
|
pr_info(" Adding %d keys\n", entries);
|
|
start = ktime_get_ns();
|
|
for (i = 0; i < entries; i++) {
|
|
struct test_obj *obj = &array[i];
|
|
|
|
obj->value.id = i * 2;
|
|
err = insert_retry(ht, obj, test_rht_params);
|
|
if (err > 0)
|
|
insert_retries += err;
|
|
else if (err)
|
|
return err;
|
|
}
|
|
|
|
if (insert_retries)
|
|
pr_info(" %u insertions retried due to memory pressure\n",
|
|
insert_retries);
|
|
|
|
test_bucket_stats(ht, entries);
|
|
rcu_read_lock();
|
|
test_rht_lookup(ht, array, entries);
|
|
rcu_read_unlock();
|
|
|
|
test_bucket_stats(ht, entries);
|
|
|
|
pr_info(" Deleting %d keys\n", entries);
|
|
for (i = 0; i < entries; i++) {
|
|
struct test_obj_val key = {
|
|
.id = i * 2,
|
|
};
|
|
|
|
if (array[i].value.id != TEST_INSERT_FAIL) {
|
|
obj = rhashtable_lookup_fast(ht, &key, test_rht_params);
|
|
BUG_ON(!obj);
|
|
|
|
rhashtable_remove_fast(ht, &obj->node, test_rht_params);
|
|
}
|
|
|
|
cond_resched();
|
|
}
|
|
|
|
end = ktime_get_ns();
|
|
pr_info(" Duration of test: %lld ns\n", end - start);
|
|
|
|
return end - start;
|
|
}
|
|
|
|
static struct rhashtable ht;
|
|
static struct rhltable rhlt;
|
|
|
|
static int __init test_rhltable(unsigned int entries)
|
|
{
|
|
struct test_obj_rhl *rhl_test_objects;
|
|
unsigned long *obj_in_table;
|
|
unsigned int i, j, k;
|
|
int ret, err;
|
|
|
|
if (entries == 0)
|
|
entries = 1;
|
|
|
|
rhl_test_objects = vzalloc(array_size(entries,
|
|
sizeof(*rhl_test_objects)));
|
|
if (!rhl_test_objects)
|
|
return -ENOMEM;
|
|
|
|
ret = -ENOMEM;
|
|
obj_in_table = vzalloc(array_size(sizeof(unsigned long),
|
|
BITS_TO_LONGS(entries)));
|
|
if (!obj_in_table)
|
|
goto out_free;
|
|
|
|
err = rhltable_init(&rhlt, &test_rht_params);
|
|
if (WARN_ON(err))
|
|
goto out_free;
|
|
|
|
k = get_random_u32();
|
|
ret = 0;
|
|
for (i = 0; i < entries; i++) {
|
|
rhl_test_objects[i].value.id = k;
|
|
err = rhltable_insert(&rhlt, &rhl_test_objects[i].list_node,
|
|
test_rht_params);
|
|
if (WARN(err, "error %d on element %d\n", err, i))
|
|
break;
|
|
if (err == 0)
|
|
set_bit(i, obj_in_table);
|
|
}
|
|
|
|
if (err)
|
|
ret = err;
|
|
|
|
pr_info("test %d add/delete pairs into rhlist\n", entries);
|
|
for (i = 0; i < entries; i++) {
|
|
struct rhlist_head *h, *pos;
|
|
struct test_obj_rhl *obj;
|
|
struct test_obj_val key = {
|
|
.id = k,
|
|
};
|
|
bool found;
|
|
|
|
rcu_read_lock();
|
|
h = rhltable_lookup(&rhlt, &key, test_rht_params);
|
|
if (WARN(!h, "key not found during iteration %d of %d", i, entries)) {
|
|
rcu_read_unlock();
|
|
break;
|
|
}
|
|
|
|
if (i) {
|
|
j = i - 1;
|
|
rhl_for_each_entry_rcu(obj, pos, h, list_node) {
|
|
if (WARN(pos == &rhl_test_objects[j].list_node, "old element found, should be gone"))
|
|
break;
|
|
}
|
|
}
|
|
|
|
cond_resched_rcu();
|
|
|
|
found = false;
|
|
|
|
rhl_for_each_entry_rcu(obj, pos, h, list_node) {
|
|
if (pos == &rhl_test_objects[i].list_node) {
|
|
found = true;
|
|
break;
|
|
}
|
|
}
|
|
|
|
rcu_read_unlock();
|
|
|
|
if (WARN(!found, "element %d not found", i))
|
|
break;
|
|
|
|
err = rhltable_remove(&rhlt, &rhl_test_objects[i].list_node, test_rht_params);
|
|
WARN(err, "rhltable_remove: err %d for iteration %d\n", err, i);
|
|
if (err == 0)
|
|
clear_bit(i, obj_in_table);
|
|
}
|
|
|
|
if (ret == 0 && err)
|
|
ret = err;
|
|
|
|
for (i = 0; i < entries; i++) {
|
|
WARN(test_bit(i, obj_in_table), "elem %d allegedly still present", i);
|
|
|
|
err = rhltable_insert(&rhlt, &rhl_test_objects[i].list_node,
|
|
test_rht_params);
|
|
if (WARN(err, "error %d on element %d\n", err, i))
|
|
break;
|
|
if (err == 0)
|
|
set_bit(i, obj_in_table);
|
|
}
|
|
|
|
pr_info("test %d random rhlist add/delete operations\n", entries);
|
|
for (j = 0; j < entries; j++) {
|
|
u32 i = get_random_u32_below(entries);
|
|
u32 prand = get_random_u32_below(4);
|
|
|
|
cond_resched();
|
|
|
|
err = rhltable_remove(&rhlt, &rhl_test_objects[i].list_node, test_rht_params);
|
|
if (test_bit(i, obj_in_table)) {
|
|
clear_bit(i, obj_in_table);
|
|
if (WARN(err, "cannot remove element at slot %d", i))
|
|
continue;
|
|
} else {
|
|
if (WARN(err != -ENOENT, "removed non-existent element %d, error %d not %d",
|
|
i, err, -ENOENT))
|
|
continue;
|
|
}
|
|
|
|
if (prand & 1) {
|
|
err = rhltable_insert(&rhlt, &rhl_test_objects[i].list_node, test_rht_params);
|
|
if (err == 0) {
|
|
if (WARN(test_and_set_bit(i, obj_in_table), "succeeded to insert same object %d", i))
|
|
continue;
|
|
} else {
|
|
if (WARN(!test_bit(i, obj_in_table), "failed to insert object %d", i))
|
|
continue;
|
|
}
|
|
}
|
|
|
|
if (prand & 2) {
|
|
i = get_random_u32_below(entries);
|
|
if (test_bit(i, obj_in_table)) {
|
|
err = rhltable_remove(&rhlt, &rhl_test_objects[i].list_node, test_rht_params);
|
|
WARN(err, "cannot remove element at slot %d", i);
|
|
if (err == 0)
|
|
clear_bit(i, obj_in_table);
|
|
} else {
|
|
err = rhltable_insert(&rhlt, &rhl_test_objects[i].list_node, test_rht_params);
|
|
WARN(err, "failed to insert object %d", i);
|
|
if (err == 0)
|
|
set_bit(i, obj_in_table);
|
|
}
|
|
}
|
|
}
|
|
|
|
for (i = 0; i < entries; i++) {
|
|
cond_resched();
|
|
err = rhltable_remove(&rhlt, &rhl_test_objects[i].list_node, test_rht_params);
|
|
if (test_bit(i, obj_in_table)) {
|
|
if (WARN(err, "cannot remove element at slot %d", i))
|
|
continue;
|
|
} else {
|
|
if (WARN(err != -ENOENT, "removed non-existent element, error %d not %d",
|
|
err, -ENOENT))
|
|
continue;
|
|
}
|
|
}
|
|
|
|
rhltable_destroy(&rhlt);
|
|
out_free:
|
|
vfree(rhl_test_objects);
|
|
vfree(obj_in_table);
|
|
return ret;
|
|
}
|
|
|
|
static int __init test_rhashtable_max(struct test_obj *array,
|
|
unsigned int entries)
|
|
{
|
|
unsigned int i;
|
|
int err;
|
|
|
|
test_rht_params.max_size = roundup_pow_of_two(entries / 8);
|
|
err = rhashtable_init(&ht, &test_rht_params);
|
|
if (err)
|
|
return err;
|
|
|
|
for (i = 0; i < ht.max_elems; i++) {
|
|
struct test_obj *obj = &array[i];
|
|
|
|
obj->value.id = i * 2;
|
|
err = insert_retry(&ht, obj, test_rht_params);
|
|
if (err < 0)
|
|
return err;
|
|
}
|
|
|
|
err = insert_retry(&ht, &array[ht.max_elems], test_rht_params);
|
|
if (err == -E2BIG) {
|
|
err = 0;
|
|
} else {
|
|
pr_info("insert element %u should have failed with %d, got %d\n",
|
|
ht.max_elems, -E2BIG, err);
|
|
if (err == 0)
|
|
err = -1;
|
|
}
|
|
|
|
rhashtable_destroy(&ht);
|
|
|
|
return err;
|
|
}
|
|
|
|
static unsigned int __init print_ht(struct rhltable *rhlt)
|
|
{
|
|
struct rhashtable *ht;
|
|
const struct bucket_table *tbl;
|
|
char buff[512] = "";
|
|
int offset = 0;
|
|
unsigned int i, cnt = 0;
|
|
|
|
ht = &rhlt->ht;
|
|
/* Take the mutex to avoid RCU warning */
|
|
mutex_lock(&ht->mutex);
|
|
tbl = rht_dereference(ht->tbl, ht);
|
|
for (i = 0; i < tbl->size; i++) {
|
|
struct rhash_head *pos, *next;
|
|
struct test_obj_rhl *p;
|
|
|
|
pos = rht_ptr_exclusive(tbl->buckets + i);
|
|
next = !rht_is_a_nulls(pos) ? rht_dereference(pos->next, ht) : NULL;
|
|
|
|
if (!rht_is_a_nulls(pos)) {
|
|
offset += sprintf(buff + offset, "\nbucket[%d] -> ", i);
|
|
}
|
|
|
|
while (!rht_is_a_nulls(pos)) {
|
|
struct rhlist_head *list = container_of(pos, struct rhlist_head, rhead);
|
|
offset += sprintf(buff + offset, "[[");
|
|
do {
|
|
pos = &list->rhead;
|
|
list = rht_dereference(list->next, ht);
|
|
p = rht_obj(ht, pos);
|
|
|
|
offset += sprintf(buff + offset, " val %d (tid=%d)%s", p->value.id, p->value.tid,
|
|
list? ", " : " ");
|
|
cnt++;
|
|
} while (list);
|
|
|
|
pos = next,
|
|
next = !rht_is_a_nulls(pos) ?
|
|
rht_dereference(pos->next, ht) : NULL;
|
|
|
|
offset += sprintf(buff + offset, "]]%s", !rht_is_a_nulls(pos) ? " -> " : "");
|
|
}
|
|
}
|
|
printk(KERN_ERR "\n---- ht: ----%s\n-------------\n", buff);
|
|
mutex_unlock(&ht->mutex);
|
|
|
|
return cnt;
|
|
}
|
|
|
|
static int __init test_insert_dup(struct test_obj_rhl *rhl_test_objects,
|
|
int cnt, bool slow)
|
|
{
|
|
struct rhltable *rhlt;
|
|
unsigned int i, ret;
|
|
const char *key;
|
|
int err = 0;
|
|
|
|
rhlt = kmalloc(sizeof(*rhlt), GFP_KERNEL);
|
|
if (WARN_ON(!rhlt))
|
|
return -EINVAL;
|
|
|
|
err = rhltable_init(rhlt, &test_rht_params_dup);
|
|
if (WARN_ON(err)) {
|
|
kfree(rhlt);
|
|
return err;
|
|
}
|
|
|
|
for (i = 0; i < cnt; i++) {
|
|
rhl_test_objects[i].value.tid = i;
|
|
key = rht_obj(&rhlt->ht, &rhl_test_objects[i].list_node.rhead);
|
|
key += test_rht_params_dup.key_offset;
|
|
|
|
if (slow) {
|
|
err = PTR_ERR(rhashtable_insert_slow(&rhlt->ht, key,
|
|
&rhl_test_objects[i].list_node.rhead));
|
|
if (err == -EAGAIN)
|
|
err = 0;
|
|
} else
|
|
err = rhltable_insert(rhlt,
|
|
&rhl_test_objects[i].list_node,
|
|
test_rht_params_dup);
|
|
if (WARN(err, "error %d on element %d/%d (%s)\n", err, i, cnt, slow? "slow" : "fast"))
|
|
goto skip_print;
|
|
}
|
|
|
|
ret = print_ht(rhlt);
|
|
WARN(ret != cnt, "missing rhltable elements (%d != %d, %s)\n", ret, cnt, slow? "slow" : "fast");
|
|
|
|
skip_print:
|
|
rhltable_destroy(rhlt);
|
|
kfree(rhlt);
|
|
|
|
return 0;
|
|
}
|
|
|
|
static int __init test_insert_duplicates_run(void)
|
|
{
|
|
struct test_obj_rhl rhl_test_objects[3] = {};
|
|
|
|
pr_info("test inserting duplicates\n");
|
|
|
|
/* two different values that map to same bucket */
|
|
rhl_test_objects[0].value.id = 1;
|
|
rhl_test_objects[1].value.id = 21;
|
|
|
|
/* and another duplicate with same as [0] value
|
|
* which will be second on the bucket list */
|
|
rhl_test_objects[2].value.id = rhl_test_objects[0].value.id;
|
|
|
|
test_insert_dup(rhl_test_objects, 2, false);
|
|
test_insert_dup(rhl_test_objects, 3, false);
|
|
test_insert_dup(rhl_test_objects, 2, true);
|
|
test_insert_dup(rhl_test_objects, 3, true);
|
|
|
|
return 0;
|
|
}
|
|
|
|
static int thread_lookup_test(struct thread_data *tdata)
|
|
{
|
|
unsigned int entries = tdata->entries;
|
|
int i, err = 0;
|
|
|
|
for (i = 0; i < entries; i++) {
|
|
struct test_obj *obj;
|
|
struct test_obj_val key = {
|
|
.id = i,
|
|
.tid = tdata->id,
|
|
};
|
|
|
|
obj = rhashtable_lookup_fast(&ht, &key, test_rht_params);
|
|
if (obj && (tdata->objs[i].value.id == TEST_INSERT_FAIL)) {
|
|
pr_err(" found unexpected object %d-%d\n", key.tid, key.id);
|
|
err++;
|
|
} else if (!obj && (tdata->objs[i].value.id != TEST_INSERT_FAIL)) {
|
|
pr_err(" object %d-%d not found!\n", key.tid, key.id);
|
|
err++;
|
|
} else if (obj && memcmp(&obj->value, &key, sizeof(key))) {
|
|
pr_err(" wrong object returned (got %d-%d, expected %d-%d)\n",
|
|
obj->value.tid, obj->value.id, key.tid, key.id);
|
|
err++;
|
|
}
|
|
|
|
cond_resched();
|
|
}
|
|
return err;
|
|
}
|
|
|
|
static int threadfunc(void *data)
|
|
{
|
|
int i, step, err = 0, insert_retries = 0;
|
|
struct thread_data *tdata = data;
|
|
|
|
if (atomic_dec_and_test(&startup_count))
|
|
wake_up(&startup_wait);
|
|
if (wait_event_interruptible(startup_wait, atomic_read(&startup_count) == -1)) {
|
|
pr_err(" thread[%d]: interrupted\n", tdata->id);
|
|
goto out;
|
|
}
|
|
|
|
for (i = 0; i < tdata->entries; i++) {
|
|
tdata->objs[i].value.id = i;
|
|
tdata->objs[i].value.tid = tdata->id;
|
|
err = insert_retry(&ht, &tdata->objs[i], test_rht_params);
|
|
if (err > 0) {
|
|
insert_retries += err;
|
|
} else if (err) {
|
|
pr_err(" thread[%d]: rhashtable_insert_fast failed\n",
|
|
tdata->id);
|
|
goto out;
|
|
}
|
|
}
|
|
if (insert_retries)
|
|
pr_info(" thread[%d]: %u insertions retried due to memory pressure\n",
|
|
tdata->id, insert_retries);
|
|
|
|
err = thread_lookup_test(tdata);
|
|
if (err) {
|
|
pr_err(" thread[%d]: rhashtable_lookup_test failed\n",
|
|
tdata->id);
|
|
goto out;
|
|
}
|
|
|
|
for (step = 10; step > 0; step--) {
|
|
for (i = 0; i < tdata->entries; i += step) {
|
|
if (tdata->objs[i].value.id == TEST_INSERT_FAIL)
|
|
continue;
|
|
err = rhashtable_remove_fast(&ht, &tdata->objs[i].node,
|
|
test_rht_params);
|
|
if (err) {
|
|
pr_err(" thread[%d]: rhashtable_remove_fast failed\n",
|
|
tdata->id);
|
|
goto out;
|
|
}
|
|
tdata->objs[i].value.id = TEST_INSERT_FAIL;
|
|
|
|
cond_resched();
|
|
}
|
|
err = thread_lookup_test(tdata);
|
|
if (err) {
|
|
pr_err(" thread[%d]: rhashtable_lookup_test (2) failed\n",
|
|
tdata->id);
|
|
goto out;
|
|
}
|
|
}
|
|
out:
|
|
while (!kthread_should_stop()) {
|
|
set_current_state(TASK_INTERRUPTIBLE);
|
|
schedule();
|
|
}
|
|
return err;
|
|
}
|
|
|
|
static int __init test_rht_init(void)
|
|
{
|
|
unsigned int entries;
|
|
int i, err, started_threads = 0, failed_threads = 0;
|
|
u64 total_time = 0;
|
|
struct thread_data *tdata;
|
|
struct test_obj *objs;
|
|
|
|
if (parm_entries < 0)
|
|
parm_entries = 1;
|
|
|
|
entries = min(parm_entries, MAX_ENTRIES);
|
|
|
|
test_rht_params.automatic_shrinking = shrinking;
|
|
test_rht_params.max_size = max_size ? : roundup_pow_of_two(entries);
|
|
test_rht_params.nelem_hint = size;
|
|
|
|
objs = vzalloc(array_size(sizeof(struct test_obj),
|
|
test_rht_params.max_size + 1));
|
|
if (!objs)
|
|
return -ENOMEM;
|
|
|
|
pr_info("Running rhashtable test nelem=%d, max_size=%d, shrinking=%d\n",
|
|
size, max_size, shrinking);
|
|
|
|
for (i = 0; i < runs; i++) {
|
|
s64 time;
|
|
|
|
pr_info("Test %02d:\n", i);
|
|
memset(objs, 0, test_rht_params.max_size * sizeof(struct test_obj));
|
|
|
|
err = rhashtable_init(&ht, &test_rht_params);
|
|
if (err < 0) {
|
|
pr_warn("Test failed: Unable to initialize hashtable: %d\n",
|
|
err);
|
|
continue;
|
|
}
|
|
|
|
time = test_rhashtable(&ht, objs, entries);
|
|
rhashtable_destroy(&ht);
|
|
if (time < 0) {
|
|
vfree(objs);
|
|
pr_warn("Test failed: return code %lld\n", time);
|
|
return -EINVAL;
|
|
}
|
|
|
|
total_time += time;
|
|
}
|
|
|
|
pr_info("test if its possible to exceed max_size %d: %s\n",
|
|
test_rht_params.max_size, test_rhashtable_max(objs, entries) == 0 ?
|
|
"no, ok" : "YES, failed");
|
|
vfree(objs);
|
|
|
|
do_div(total_time, runs);
|
|
pr_info("Average test time: %llu\n", total_time);
|
|
|
|
test_insert_duplicates_run();
|
|
|
|
if (!tcount)
|
|
return 0;
|
|
|
|
pr_info("Testing concurrent rhashtable access from %d threads\n",
|
|
tcount);
|
|
atomic_set(&startup_count, tcount);
|
|
tdata = vzalloc(array_size(tcount, sizeof(struct thread_data)));
|
|
if (!tdata)
|
|
return -ENOMEM;
|
|
objs = vzalloc(array3_size(sizeof(struct test_obj), tcount, entries));
|
|
if (!objs) {
|
|
vfree(tdata);
|
|
return -ENOMEM;
|
|
}
|
|
|
|
test_rht_params.max_size = max_size ? :
|
|
roundup_pow_of_two(tcount * entries);
|
|
err = rhashtable_init(&ht, &test_rht_params);
|
|
if (err < 0) {
|
|
pr_warn("Test failed: Unable to initialize hashtable: %d\n",
|
|
err);
|
|
vfree(tdata);
|
|
vfree(objs);
|
|
return -EINVAL;
|
|
}
|
|
for (i = 0; i < tcount; i++) {
|
|
tdata[i].id = i;
|
|
tdata[i].entries = entries;
|
|
tdata[i].objs = objs + i * entries;
|
|
tdata[i].task = kthread_run(threadfunc, &tdata[i],
|
|
"rhashtable_thrad[%d]", i);
|
|
if (IS_ERR(tdata[i].task)) {
|
|
pr_err(" kthread_run failed for thread %d\n", i);
|
|
atomic_dec(&startup_count);
|
|
} else {
|
|
started_threads++;
|
|
}
|
|
}
|
|
if (wait_event_interruptible(startup_wait, atomic_read(&startup_count) == 0))
|
|
pr_err(" wait_event interruptible failed\n");
|
|
/* count is 0 now, set it to -1 and wake up all threads together */
|
|
atomic_dec(&startup_count);
|
|
wake_up_all(&startup_wait);
|
|
for (i = 0; i < tcount; i++) {
|
|
if (IS_ERR(tdata[i].task))
|
|
continue;
|
|
if ((err = kthread_stop(tdata[i].task))) {
|
|
pr_warn("Test failed: thread %d returned: %d\n",
|
|
i, err);
|
|
failed_threads++;
|
|
}
|
|
}
|
|
rhashtable_destroy(&ht);
|
|
vfree(tdata);
|
|
vfree(objs);
|
|
|
|
/*
|
|
* rhltable_remove is very expensive, default values can cause test
|
|
* to run for 2 minutes or more, use a smaller number instead.
|
|
*/
|
|
err = test_rhltable(entries / 16);
|
|
pr_info("Started %d threads, %d failed, rhltable test returns %d\n",
|
|
started_threads, failed_threads, err);
|
|
return 0;
|
|
}
|
|
|
|
static void __exit test_rht_exit(void)
|
|
{
|
|
}
|
|
|
|
module_init(test_rht_init);
|
|
module_exit(test_rht_exit);
|
|
|
|
MODULE_LICENSE("GPL v2");
|