Merge tag 'net-next-5.13' of git://git.kernel.org/pub/scm/linux/kernel/git/netdev/net-next
Pull networking updates from Jakub Kicinski: "Core: - bpf: - allow bpf programs calling kernel functions (initially to reuse TCP congestion control implementations) - enable task local storage for tracing programs - remove the need to store per-task state in hash maps, and allow tracing programs access to task local storage previously added for BPF_LSM - add bpf_for_each_map_elem() helper, allowing programs to walk all map elements in a more robust and easier to verify fashion - sockmap: support UDP and cross-protocol BPF_SK_SKB_VERDICT redirection - lpm: add support for batched ops in LPM trie - add BTF_KIND_FLOAT support - mostly to allow use of BTF on s390 which has floats in its headers files - improve BPF syscall documentation and extend the use of kdoc parsing scripts we already employ for bpf-helpers - libbpf, bpftool: support static linking of BPF ELF files - improve support for encapsulation of L2 packets - xdp: restructure redirect actions to avoid a runtime lookup, improving performance by 4-8% in microbenchmarks - xsk: build skb by page (aka generic zerocopy xmit) - improve performance of software AF_XDP path by 33% for devices which don't need headers in the linear skb part (e.g. virtio) - nexthop: resilient next-hop groups - improve path stability on next-hops group changes (incl. offload for mlxsw) - ipv6: segment routing: add support for IPv4 decapsulation - icmp: add support for RFC 8335 extended PROBE messages - inet: use bigger hash table for IP ID generation - tcp: deal better with delayed TX completions - make sure we don't give up on fast TCP retransmissions only because driver is slow in reporting that it completed transmitting the original - tcp: reorder tcp_congestion_ops for better cache locality - mptcp: - add sockopt support for common TCP options - add support for common TCP msg flags - include multiple address ids in RM_ADDR - add reset option support for resetting one subflow - udp: GRO L4 improvements - improve 'forward' / 'frag_list' co-existence with UDP tunnel GRO, allowing the first to take place correctly even for encapsulated UDP traffic - micro-optimize dev_gro_receive() and flow dissection, avoid retpoline overhead on VLAN and TEB GRO - use less memory for sysctls, add a new sysctl type, to allow using u8 instead of "int" and "long" and shrink networking sysctls - veth: allow GRO without XDP - this allows aggregating UDP packets before handing them off to routing, bridge, OvS, etc. - allow specifing ifindex when device is moved to another namespace - netfilter: - nft_socket: add support for cgroupsv2 - nftables: add catch-all set element - special element used to define a default action in case normal lookup missed - use net_generic infra in many modules to avoid allocating per-ns memory unnecessarily - xps: improve the xps handling to avoid potential out-of-bound accesses and use-after-free when XPS change race with other re-configuration under traffic - add a config knob to turn off per-cpu netdev refcnt to catch underflows in testing Device APIs: - add WWAN subsystem to organize the WWAN interfaces better and hopefully start driving towards more unified and vendor- independent APIs - ethtool: - add interface for reading IEEE MIB stats (incl. mlx5 and bnxt support) - allow network drivers to dump arbitrary SFP EEPROM data, current offset+length API was a poor fit for modern SFP which define EEPROM in terms of pages (incl. mlx5 support) - act_police, flow_offload: add support for packet-per-second policing (incl. offload for nfp) - psample: add additional metadata attributes like transit delay for packets sampled from switch HW (and corresponding egress and policy-based sampling in the mlxsw driver) - dsa: improve support for sandwiched LAGs with bridge and DSA - netfilter: - flowtable: use direct xmit in topologies with IP forwarding, bridging, vlans etc. - nftables: counter hardware offload support - Bluetooth: - improvements for firmware download w/ Intel devices - add support for reading AOSP vendor capabilities - add support for virtio transport driver - mac80211: - allow concurrent monitor iface and ethernet rx decap - set priority and queue mapping for injected frames - phy: add support for Clause-45 PHY Loopback - pci/iov: add sysfs MSI-X vector assignment interface to distribute MSI-X resources to VFs (incl. mlx5 support) New hardware/drivers: - dsa: mv88e6xxx: add support for Marvell mv88e6393x - 11-port Ethernet switch with 8x 1-Gigabit Ethernet and 3x 10-Gigabit interfaces. - dsa: support for legacy Broadcom tags used on BCM5325, BCM5365 and BCM63xx switches - Microchip KSZ8863 and KSZ8873; 3x 10/100Mbps Ethernet switches - ath11k: support for QCN9074 a 802.11ax device - Bluetooth: Broadcom BCM4330 and BMC4334 - phy: Marvell 88X2222 transceiver support - mdio: add BCM6368 MDIO mux bus controller - r8152: support RTL8153 and RTL8156 (USB Ethernet) chips - mana: driver for Microsoft Azure Network Adapter (MANA) - Actions Semi Owl Ethernet MAC - can: driver for ETAS ES58X CAN/USB interfaces Pure driver changes: - add XDP support to: enetc, igc, stmmac - add AF_XDP support to: stmmac - virtio: - page_to_skb() use build_skb when there's sufficient tailroom (21% improvement for 1000B UDP frames) - support XDP even without dedicated Tx queues - share the Tx queues with the stack when necessary - mlx5: - flow rules: add support for mirroring with conntrack, matching on ICMP, GTP, flex filters and more - support packet sampling with flow offloads - persist uplink representor netdev across eswitch mode changes - allow coexistence of CQE compression and HW time-stamping - add ethtool extended link error state reporting - ice, iavf: support flow filters, UDP Segmentation Offload - dpaa2-switch: - move the driver out of staging - add spanning tree (STP) support - add rx copybreak support - add tc flower hardware offload on ingress traffic - ionic: - implement Rx page reuse - support HW PTP time-stamping - octeon: support TC hardware offloads - flower matching on ingress and egress ratelimitting. - stmmac: - add RX frame steering based on VLAN priority in tc flower - support frame preemption (FPE) - intel: add cross time-stamping freq difference adjustment - ocelot: - support forwarding of MRP frames in HW - support multiple bridges - support PTP Sync one-step timestamping - dsa: mv88e6xxx, dpaa2-switch: offload bridge port flags like learning, flooding etc. - ipa: add IPA v4.5, v4.9 and v4.11 support (Qualcomm SDX55, SM8350, SC7280 SoCs) - mt7601u: enable TDLS support - mt76: - add support for 802.3 rx frames (mt7915/mt7615) - mt7915 flash pre-calibration support - mt7921/mt7663 runtime power management fixes" * tag 'net-next-5.13' of git://git.kernel.org/pub/scm/linux/kernel/git/netdev/net-next: (2451 commits) net: selftest: fix build issue if INET is disabled net: netrom: nr_in: Remove redundant assignment to ns net: tun: Remove redundant assignment to ret net: phy: marvell: add downshift support for M88E1240 net: dsa: ksz: Make reg_mib_cnt a u8 as it never exceeds 255 net/sched: act_ct: Remove redundant ct get and check icmp: standardize naming of RFC 8335 PROBE constants bpf, selftests: Update array map tests for per-cpu batched ops bpf: Add batched ops support for percpu array bpf: Implement formatted output helpers with bstr_printf seq_file: Add a seq_bprintf function sfc: adjust efx->xdp_tx_queue_count with the real number of initialized queues net:nfc:digital: Fix a double free in digital_tg_recv_dep_req net: fix a concurrency bug in l2tp_tunnel_register() net/smc: Remove redundant assignment to rc mpls: Remove redundant assignment to err llc2: Remove redundant assignment to rc net/tls: Remove redundant initialization of record rds: Remove redundant assignment to nr_sig dt-bindings: net: mdio-gpio: add compatible for microchip,mdio-smi0 ...
This commit is contained in:
@ -3,7 +3,7 @@
|
||||
*
|
||||
* Copyright (c) 2000-2006, 2014-2018, Ericsson AB
|
||||
* Copyright (c) 2004-2008, 2010-2014, Wind River Systems
|
||||
* Copyright (c) 2020, Red Hat Inc
|
||||
* Copyright (c) 2020-2021, Red Hat Inc
|
||||
* All rights reserved.
|
||||
*
|
||||
* Redistribution and use in source and binary forms, with or without
|
||||
@ -222,59 +222,57 @@ static int hash(int x)
|
||||
|
||||
/**
|
||||
* tipc_publ_create - create a publication structure
|
||||
* @type: name sequence type
|
||||
* @lower: name sequence lower bound
|
||||
* @upper: name sequence upper bound
|
||||
* @scope: publication scope
|
||||
* @node: network address of publishing socket
|
||||
* @port: publishing port
|
||||
* @ua: the service range the user is binding to
|
||||
* @sk: the address of the socket that is bound
|
||||
* @key: publication key
|
||||
*/
|
||||
static struct publication *tipc_publ_create(u32 type, u32 lower, u32 upper,
|
||||
u32 scope, u32 node, u32 port,
|
||||
static struct publication *tipc_publ_create(struct tipc_uaddr *ua,
|
||||
struct tipc_socket_addr *sk,
|
||||
u32 key)
|
||||
{
|
||||
struct publication *publ = kzalloc(sizeof(*publ), GFP_ATOMIC);
|
||||
struct publication *p = kzalloc(sizeof(*p), GFP_ATOMIC);
|
||||
|
||||
if (!publ)
|
||||
if (!p)
|
||||
return NULL;
|
||||
|
||||
publ->type = type;
|
||||
publ->lower = lower;
|
||||
publ->upper = upper;
|
||||
publ->scope = scope;
|
||||
publ->node = node;
|
||||
publ->port = port;
|
||||
publ->key = key;
|
||||
INIT_LIST_HEAD(&publ->binding_sock);
|
||||
INIT_LIST_HEAD(&publ->binding_node);
|
||||
INIT_LIST_HEAD(&publ->local_publ);
|
||||
INIT_LIST_HEAD(&publ->all_publ);
|
||||
INIT_LIST_HEAD(&publ->list);
|
||||
return publ;
|
||||
p->sr = ua->sr;
|
||||
p->sk = *sk;
|
||||
p->scope = ua->scope;
|
||||
p->key = key;
|
||||
INIT_LIST_HEAD(&p->binding_sock);
|
||||
INIT_LIST_HEAD(&p->binding_node);
|
||||
INIT_LIST_HEAD(&p->local_publ);
|
||||
INIT_LIST_HEAD(&p->all_publ);
|
||||
INIT_LIST_HEAD(&p->list);
|
||||
return p;
|
||||
}
|
||||
|
||||
/**
|
||||
* tipc_service_create - create a service structure for the specified 'type'
|
||||
* @type: service type
|
||||
* @hd: name_table services list
|
||||
* @net: network namespace
|
||||
* @ua: address representing the service to be bound
|
||||
*
|
||||
* Allocates a single range structure and sets it to all 0's.
|
||||
*/
|
||||
static struct tipc_service *tipc_service_create(u32 type, struct hlist_head *hd)
|
||||
static struct tipc_service *tipc_service_create(struct net *net,
|
||||
struct tipc_uaddr *ua)
|
||||
{
|
||||
struct tipc_service *service = kzalloc(sizeof(*service), GFP_ATOMIC);
|
||||
struct name_table *nt = tipc_name_table(net);
|
||||
struct tipc_service *service;
|
||||
struct hlist_head *hd;
|
||||
|
||||
service = kzalloc(sizeof(*service), GFP_ATOMIC);
|
||||
if (!service) {
|
||||
pr_warn("Service creation failed, no memory\n");
|
||||
return NULL;
|
||||
}
|
||||
|
||||
spin_lock_init(&service->lock);
|
||||
service->type = type;
|
||||
service->type = ua->sr.type;
|
||||
service->ranges = RB_ROOT;
|
||||
INIT_HLIST_NODE(&service->service_list);
|
||||
INIT_LIST_HEAD(&service->subscriptions);
|
||||
hd = &nt->services[hash(ua->sr.type)];
|
||||
hlist_add_head_rcu(&service->service_list, hd);
|
||||
return service;
|
||||
}
|
||||
@ -282,13 +280,13 @@ static struct tipc_service *tipc_service_create(u32 type, struct hlist_head *hd)
|
||||
/* tipc_service_find_range - find service range matching publication parameters
|
||||
*/
|
||||
static struct service_range *tipc_service_find_range(struct tipc_service *sc,
|
||||
u32 lower, u32 upper)
|
||||
struct tipc_uaddr *ua)
|
||||
{
|
||||
struct service_range *sr;
|
||||
|
||||
service_range_foreach_match(sr, sc, lower, upper) {
|
||||
service_range_foreach_match(sr, sc, ua->sr.lower, ua->sr.upper) {
|
||||
/* Look for exact match */
|
||||
if (sr->lower == lower && sr->upper == upper)
|
||||
if (sr->lower == ua->sr.lower && sr->upper == ua->sr.upper)
|
||||
return sr;
|
||||
}
|
||||
|
||||
@ -296,10 +294,12 @@ static struct service_range *tipc_service_find_range(struct tipc_service *sc,
|
||||
}
|
||||
|
||||
static struct service_range *tipc_service_create_range(struct tipc_service *sc,
|
||||
u32 lower, u32 upper)
|
||||
struct publication *p)
|
||||
{
|
||||
struct rb_node **n, *parent = NULL;
|
||||
struct service_range *sr;
|
||||
u32 lower = p->sr.lower;
|
||||
u32 upper = p->sr.upper;
|
||||
|
||||
n = &sc->ranges.rb_node;
|
||||
while (*n) {
|
||||
@ -327,64 +327,68 @@ static struct service_range *tipc_service_create_range(struct tipc_service *sc,
|
||||
return sr;
|
||||
}
|
||||
|
||||
static struct publication *tipc_service_insert_publ(struct net *net,
|
||||
struct tipc_service *sc,
|
||||
u32 type, u32 lower,
|
||||
u32 upper, u32 scope,
|
||||
u32 node, u32 port,
|
||||
u32 key)
|
||||
static bool tipc_service_insert_publ(struct net *net,
|
||||
struct tipc_service *sc,
|
||||
struct publication *p)
|
||||
{
|
||||
struct tipc_subscription *sub, *tmp;
|
||||
struct service_range *sr;
|
||||
struct publication *p;
|
||||
struct publication *_p;
|
||||
u32 node = p->sk.node;
|
||||
bool first = false;
|
||||
bool res = false;
|
||||
u32 key = p->key;
|
||||
|
||||
sr = tipc_service_create_range(sc, lower, upper);
|
||||
spin_lock_bh(&sc->lock);
|
||||
sr = tipc_service_create_range(sc, p);
|
||||
if (!sr)
|
||||
goto err;
|
||||
goto exit;
|
||||
|
||||
first = list_empty(&sr->all_publ);
|
||||
|
||||
/* Return if the publication already exists */
|
||||
list_for_each_entry(p, &sr->all_publ, all_publ) {
|
||||
if (p->key == key && (!p->node || p->node == node))
|
||||
return NULL;
|
||||
list_for_each_entry(_p, &sr->all_publ, all_publ) {
|
||||
if (_p->key == key && (!_p->sk.node || _p->sk.node == node)) {
|
||||
pr_debug("Failed to bind duplicate %u,%u,%u/%u:%u/%u\n",
|
||||
p->sr.type, p->sr.lower, p->sr.upper,
|
||||
node, p->sk.ref, key);
|
||||
goto exit;
|
||||
}
|
||||
}
|
||||
|
||||
/* Create and insert publication */
|
||||
p = tipc_publ_create(type, lower, upper, scope, node, port, key);
|
||||
if (!p)
|
||||
goto err;
|
||||
/* Suppose there shouldn't be a huge gap btw publs i.e. >INT_MAX */
|
||||
p->id = sc->publ_cnt++;
|
||||
if (in_own_node(net, node))
|
||||
if (in_own_node(net, p->sk.node))
|
||||
list_add(&p->local_publ, &sr->local_publ);
|
||||
list_add(&p->all_publ, &sr->all_publ);
|
||||
p->id = sc->publ_cnt++;
|
||||
|
||||
/* Any subscriptions waiting for notification? */
|
||||
list_for_each_entry_safe(sub, tmp, &sc->subscriptions, service_list) {
|
||||
tipc_sub_report_overlap(sub, p->lower, p->upper, TIPC_PUBLISHED,
|
||||
p->port, p->node, p->scope, first);
|
||||
tipc_sub_report_overlap(sub, p, TIPC_PUBLISHED, first);
|
||||
}
|
||||
return p;
|
||||
err:
|
||||
pr_warn("Failed to bind to %u,%u,%u, no memory\n", type, lower, upper);
|
||||
return NULL;
|
||||
res = true;
|
||||
exit:
|
||||
if (!res)
|
||||
pr_warn("Failed to bind to %u,%u,%u\n",
|
||||
p->sr.type, p->sr.lower, p->sr.upper);
|
||||
spin_unlock_bh(&sc->lock);
|
||||
return res;
|
||||
}
|
||||
|
||||
/**
|
||||
* tipc_service_remove_publ - remove a publication from a service
|
||||
* @sr: service_range to remove publication from
|
||||
* @node: target node
|
||||
* @r: service_range to remove publication from
|
||||
* @sk: address publishing socket
|
||||
* @key: target publication key
|
||||
*/
|
||||
static struct publication *tipc_service_remove_publ(struct service_range *sr,
|
||||
u32 node, u32 key)
|
||||
static struct publication *tipc_service_remove_publ(struct service_range *r,
|
||||
struct tipc_socket_addr *sk,
|
||||
u32 key)
|
||||
{
|
||||
struct publication *p;
|
||||
u32 node = sk->node;
|
||||
|
||||
list_for_each_entry(p, &sr->all_publ, all_publ) {
|
||||
if (p->key != key || (node && node != p->node))
|
||||
list_for_each_entry(p, &r->all_publ, all_publ) {
|
||||
if (p->key != key || (node && node != p->sk.node))
|
||||
continue;
|
||||
list_del(&p->all_publ);
|
||||
list_del(&p->local_publ);
|
||||
@ -417,17 +421,14 @@ static int tipc_publ_sort(void *priv, const struct list_head *a,
|
||||
static void tipc_service_subscribe(struct tipc_service *service,
|
||||
struct tipc_subscription *sub)
|
||||
{
|
||||
struct tipc_subscr *sb = &sub->evt.s;
|
||||
struct publication *p, *first, *tmp;
|
||||
struct list_head publ_list;
|
||||
struct service_range *sr;
|
||||
struct tipc_service_range r;
|
||||
u32 filter;
|
||||
u32 filter, lower, upper;
|
||||
|
||||
r.type = tipc_sub_read(sb, seq.type);
|
||||
r.lower = tipc_sub_read(sb, seq.lower);
|
||||
r.upper = tipc_sub_read(sb, seq.upper);
|
||||
filter = tipc_sub_read(sb, filter);
|
||||
filter = sub->s.filter;
|
||||
lower = sub->s.seq.lower;
|
||||
upper = sub->s.seq.upper;
|
||||
|
||||
tipc_sub_get(sub);
|
||||
list_add(&sub->service_list, &service->subscriptions);
|
||||
@ -436,7 +437,7 @@ static void tipc_service_subscribe(struct tipc_service *service,
|
||||
return;
|
||||
|
||||
INIT_LIST_HEAD(&publ_list);
|
||||
service_range_foreach_match(sr, service, r.lower, r.upper) {
|
||||
service_range_foreach_match(sr, service, lower, upper) {
|
||||
first = NULL;
|
||||
list_for_each_entry(p, &sr->all_publ, all_publ) {
|
||||
if (filter & TIPC_SUB_PORTS)
|
||||
@ -452,80 +453,74 @@ static void tipc_service_subscribe(struct tipc_service *service,
|
||||
/* Sort the publications before reporting */
|
||||
list_sort(NULL, &publ_list, tipc_publ_sort);
|
||||
list_for_each_entry_safe(p, tmp, &publ_list, list) {
|
||||
tipc_sub_report_overlap(sub, p->lower, p->upper,
|
||||
TIPC_PUBLISHED, p->port, p->node,
|
||||
p->scope, true);
|
||||
tipc_sub_report_overlap(sub, p, TIPC_PUBLISHED, true);
|
||||
list_del_init(&p->list);
|
||||
}
|
||||
}
|
||||
|
||||
static struct tipc_service *tipc_service_find(struct net *net, u32 type)
|
||||
static struct tipc_service *tipc_service_find(struct net *net,
|
||||
struct tipc_uaddr *ua)
|
||||
{
|
||||
struct name_table *nt = tipc_name_table(net);
|
||||
struct hlist_head *service_head;
|
||||
struct tipc_service *service;
|
||||
|
||||
service_head = &nt->services[hash(type)];
|
||||
service_head = &nt->services[hash(ua->sr.type)];
|
||||
hlist_for_each_entry_rcu(service, service_head, service_list) {
|
||||
if (service->type == type)
|
||||
if (service->type == ua->sr.type)
|
||||
return service;
|
||||
}
|
||||
return NULL;
|
||||
};
|
||||
|
||||
struct publication *tipc_nametbl_insert_publ(struct net *net, u32 type,
|
||||
u32 lower, u32 upper,
|
||||
u32 scope, u32 node,
|
||||
u32 port, u32 key)
|
||||
struct publication *tipc_nametbl_insert_publ(struct net *net,
|
||||
struct tipc_uaddr *ua,
|
||||
struct tipc_socket_addr *sk,
|
||||
u32 key)
|
||||
{
|
||||
struct name_table *nt = tipc_name_table(net);
|
||||
struct tipc_service *sc;
|
||||
struct publication *p;
|
||||
|
||||
if (scope > TIPC_NODE_SCOPE || lower > upper) {
|
||||
pr_debug("Failed to bind illegal {%u,%u,%u} with scope %u\n",
|
||||
type, lower, upper, scope);
|
||||
return NULL;
|
||||
}
|
||||
sc = tipc_service_find(net, type);
|
||||
if (!sc)
|
||||
sc = tipc_service_create(type, &nt->services[hash(type)]);
|
||||
if (!sc)
|
||||
p = tipc_publ_create(ua, sk, key);
|
||||
if (!p)
|
||||
return NULL;
|
||||
|
||||
spin_lock_bh(&sc->lock);
|
||||
p = tipc_service_insert_publ(net, sc, type, lower, upper,
|
||||
scope, node, port, key);
|
||||
spin_unlock_bh(&sc->lock);
|
||||
return p;
|
||||
sc = tipc_service_find(net, ua);
|
||||
if (!sc)
|
||||
sc = tipc_service_create(net, ua);
|
||||
if (sc && tipc_service_insert_publ(net, sc, p))
|
||||
return p;
|
||||
kfree(p);
|
||||
return NULL;
|
||||
}
|
||||
|
||||
struct publication *tipc_nametbl_remove_publ(struct net *net, u32 type,
|
||||
u32 lower, u32 upper,
|
||||
u32 node, u32 key)
|
||||
struct publication *tipc_nametbl_remove_publ(struct net *net,
|
||||
struct tipc_uaddr *ua,
|
||||
struct tipc_socket_addr *sk,
|
||||
u32 key)
|
||||
{
|
||||
struct tipc_service *sc = tipc_service_find(net, type);
|
||||
struct tipc_subscription *sub, *tmp;
|
||||
struct service_range *sr = NULL;
|
||||
struct publication *p = NULL;
|
||||
struct service_range *sr;
|
||||
struct tipc_service *sc;
|
||||
bool last;
|
||||
|
||||
sc = tipc_service_find(net, ua);
|
||||
if (!sc)
|
||||
return NULL;
|
||||
goto exit;
|
||||
|
||||
spin_lock_bh(&sc->lock);
|
||||
sr = tipc_service_find_range(sc, lower, upper);
|
||||
sr = tipc_service_find_range(sc, ua);
|
||||
if (!sr)
|
||||
goto exit;
|
||||
p = tipc_service_remove_publ(sr, node, key);
|
||||
goto unlock;
|
||||
p = tipc_service_remove_publ(sr, sk, key);
|
||||
if (!p)
|
||||
goto exit;
|
||||
goto unlock;
|
||||
|
||||
/* Notify any waiting subscriptions */
|
||||
last = list_empty(&sr->all_publ);
|
||||
list_for_each_entry_safe(sub, tmp, &sc->subscriptions, service_list) {
|
||||
tipc_sub_report_overlap(sub, lower, upper, TIPC_WITHDRAWN,
|
||||
p->port, node, p->scope, last);
|
||||
tipc_sub_report_overlap(sub, p, TIPC_WITHDRAWN, last);
|
||||
}
|
||||
|
||||
/* Remove service range item if this was its last publication */
|
||||
@ -534,77 +529,85 @@ struct publication *tipc_nametbl_remove_publ(struct net *net, u32 type,
|
||||
kfree(sr);
|
||||
}
|
||||
|
||||
/* Delete service item if this no more publications and subscriptions */
|
||||
/* Delete service item if no more publications and subscriptions */
|
||||
if (RB_EMPTY_ROOT(&sc->ranges) && list_empty(&sc->subscriptions)) {
|
||||
hlist_del_init_rcu(&sc->service_list);
|
||||
kfree_rcu(sc, rcu);
|
||||
}
|
||||
exit:
|
||||
unlock:
|
||||
spin_unlock_bh(&sc->lock);
|
||||
exit:
|
||||
if (!p) {
|
||||
pr_err("Failed to remove unknown binding: %u,%u,%u/%u:%u/%u\n",
|
||||
ua->sr.type, ua->sr.lower, ua->sr.upper,
|
||||
sk->node, sk->ref, key);
|
||||
}
|
||||
return p;
|
||||
}
|
||||
|
||||
/**
|
||||
* tipc_nametbl_translate - perform service instance to socket translation
|
||||
* tipc_nametbl_lookup_anycast - perform service instance to socket translation
|
||||
* @net: network namespace
|
||||
* @type: message type
|
||||
* @instance: message instance
|
||||
* @dnode: the search domain used during translation
|
||||
* @ua: service address to look up
|
||||
* @sk: address to socket we want to find
|
||||
*
|
||||
* On entry, a non-zero 'sk->node' indicates the node where we want lookup to be
|
||||
* performed, which may not be this one.
|
||||
*
|
||||
* On exit:
|
||||
* - if translation is deferred to another node, leave 'dnode' unchanged and
|
||||
* return 0
|
||||
* - if translation is attempted and succeeds, set 'dnode' to the publishing
|
||||
* node and return the published (non-zero) port number
|
||||
* - if translation is attempted and fails, set 'dnode' to 0 and return 0
|
||||
*
|
||||
* - If lookup is deferred to another node, leave 'sk->node' unchanged and
|
||||
* return 'true'.
|
||||
* - If lookup is successful, set the 'sk->node' and 'sk->ref' (== portid) which
|
||||
* represent the bound socket and return 'true'.
|
||||
* - If lookup fails, return 'false'
|
||||
*
|
||||
* Note that for legacy users (node configured with Z.C.N address format) the
|
||||
* 'closest-first' lookup algorithm must be maintained, i.e., if dnode is 0
|
||||
* 'closest-first' lookup algorithm must be maintained, i.e., if sk.node is 0
|
||||
* we must look in the local binding list first
|
||||
*/
|
||||
u32 tipc_nametbl_translate(struct net *net, u32 type, u32 instance, u32 *dnode)
|
||||
bool tipc_nametbl_lookup_anycast(struct net *net,
|
||||
struct tipc_uaddr *ua,
|
||||
struct tipc_socket_addr *sk)
|
||||
{
|
||||
struct tipc_net *tn = tipc_net(net);
|
||||
bool legacy = tn->legacy_addr_format;
|
||||
u32 self = tipc_own_addr(net);
|
||||
struct service_range *sr;
|
||||
u32 inst = ua->sa.instance;
|
||||
struct service_range *r;
|
||||
struct tipc_service *sc;
|
||||
struct list_head *list;
|
||||
struct publication *p;
|
||||
u32 port = 0;
|
||||
u32 node = 0;
|
||||
struct list_head *l;
|
||||
bool res = false;
|
||||
|
||||
if (!tipc_in_scope(legacy, *dnode, self))
|
||||
return 0;
|
||||
if (!tipc_in_scope(legacy, sk->node, self))
|
||||
return true;
|
||||
|
||||
rcu_read_lock();
|
||||
sc = tipc_service_find(net, type);
|
||||
sc = tipc_service_find(net, ua);
|
||||
if (unlikely(!sc))
|
||||
goto exit;
|
||||
|
||||
spin_lock_bh(&sc->lock);
|
||||
service_range_foreach_match(sr, sc, instance, instance) {
|
||||
service_range_foreach_match(r, sc, inst, inst) {
|
||||
/* Select lookup algo: local, closest-first or round-robin */
|
||||
if (*dnode == self) {
|
||||
list = &sr->local_publ;
|
||||
if (list_empty(list))
|
||||
if (sk->node == self) {
|
||||
l = &r->local_publ;
|
||||
if (list_empty(l))
|
||||
continue;
|
||||
p = list_first_entry(list, struct publication,
|
||||
local_publ);
|
||||
list_move_tail(&p->local_publ, &sr->local_publ);
|
||||
} else if (legacy && !*dnode && !list_empty(&sr->local_publ)) {
|
||||
list = &sr->local_publ;
|
||||
p = list_first_entry(list, struct publication,
|
||||
local_publ);
|
||||
list_move_tail(&p->local_publ, &sr->local_publ);
|
||||
p = list_first_entry(l, struct publication, local_publ);
|
||||
list_move_tail(&p->local_publ, &r->local_publ);
|
||||
} else if (legacy && !sk->node && !list_empty(&r->local_publ)) {
|
||||
l = &r->local_publ;
|
||||
p = list_first_entry(l, struct publication, local_publ);
|
||||
list_move_tail(&p->local_publ, &r->local_publ);
|
||||
} else {
|
||||
list = &sr->all_publ;
|
||||
p = list_first_entry(list, struct publication,
|
||||
all_publ);
|
||||
list_move_tail(&p->all_publ, &sr->all_publ);
|
||||
l = &r->all_publ;
|
||||
p = list_first_entry(l, struct publication, all_publ);
|
||||
list_move_tail(&p->all_publ, &r->all_publ);
|
||||
}
|
||||
port = p->port;
|
||||
node = p->node;
|
||||
*sk = p->sk;
|
||||
res = true;
|
||||
/* Todo: as for legacy, pick the first matching range only, a
|
||||
* "true" round-robin will be performed as needed.
|
||||
*/
|
||||
@ -614,40 +617,45 @@ u32 tipc_nametbl_translate(struct net *net, u32 type, u32 instance, u32 *dnode)
|
||||
|
||||
exit:
|
||||
rcu_read_unlock();
|
||||
*dnode = node;
|
||||
return port;
|
||||
return res;
|
||||
}
|
||||
|
||||
bool tipc_nametbl_lookup(struct net *net, u32 type, u32 instance, u32 scope,
|
||||
struct list_head *dsts, int *dstcnt, u32 exclude,
|
||||
bool all)
|
||||
/* tipc_nametbl_lookup_group(): lookup destinaton(s) in a communication group
|
||||
* Returns a list of one (== group anycast) or more (== group multicast)
|
||||
* destination socket/node pairs matching the given address.
|
||||
* The requester may or may not want to exclude himself from the list.
|
||||
*/
|
||||
bool tipc_nametbl_lookup_group(struct net *net, struct tipc_uaddr *ua,
|
||||
struct list_head *dsts, int *dstcnt,
|
||||
u32 exclude, bool mcast)
|
||||
{
|
||||
u32 self = tipc_own_addr(net);
|
||||
u32 inst = ua->sa.instance;
|
||||
struct service_range *sr;
|
||||
struct tipc_service *sc;
|
||||
struct publication *p;
|
||||
|
||||
*dstcnt = 0;
|
||||
rcu_read_lock();
|
||||
sc = tipc_service_find(net, type);
|
||||
sc = tipc_service_find(net, ua);
|
||||
if (unlikely(!sc))
|
||||
goto exit;
|
||||
|
||||
spin_lock_bh(&sc->lock);
|
||||
|
||||
/* Todo: a full search i.e. service_range_foreach_match() instead? */
|
||||
sr = service_range_match_first(sc->ranges.rb_node, instance, instance);
|
||||
sr = service_range_match_first(sc->ranges.rb_node, inst, inst);
|
||||
if (!sr)
|
||||
goto no_match;
|
||||
|
||||
list_for_each_entry(p, &sr->all_publ, all_publ) {
|
||||
if (p->scope != scope)
|
||||
if (p->scope != ua->scope)
|
||||
continue;
|
||||
if (p->port == exclude && p->node == self)
|
||||
if (p->sk.ref == exclude && p->sk.node == self)
|
||||
continue;
|
||||
tipc_dest_push(dsts, p->node, p->port);
|
||||
tipc_dest_push(dsts, p->sk.node, p->sk.ref);
|
||||
(*dstcnt)++;
|
||||
if (all)
|
||||
if (mcast)
|
||||
continue;
|
||||
list_move_tail(&p->all_publ, &sr->all_publ);
|
||||
break;
|
||||
@ -659,23 +667,29 @@ exit:
|
||||
return !list_empty(dsts);
|
||||
}
|
||||
|
||||
void tipc_nametbl_mc_lookup(struct net *net, u32 type, u32 lower, u32 upper,
|
||||
u32 scope, bool exact, struct list_head *dports)
|
||||
/* tipc_nametbl_lookup_mcast_sockets(): look up node local destinaton sockets
|
||||
* matching the given address
|
||||
* Used on nodes which have received a multicast/broadcast message
|
||||
* Returns a list of local sockets
|
||||
*/
|
||||
void tipc_nametbl_lookup_mcast_sockets(struct net *net, struct tipc_uaddr *ua,
|
||||
bool exact, struct list_head *dports)
|
||||
{
|
||||
struct service_range *sr;
|
||||
struct tipc_service *sc;
|
||||
struct publication *p;
|
||||
u32 scope = ua->scope;
|
||||
|
||||
rcu_read_lock();
|
||||
sc = tipc_service_find(net, type);
|
||||
sc = tipc_service_find(net, ua);
|
||||
if (!sc)
|
||||
goto exit;
|
||||
|
||||
spin_lock_bh(&sc->lock);
|
||||
service_range_foreach_match(sr, sc, lower, upper) {
|
||||
service_range_foreach_match(sr, sc, ua->sr.lower, ua->sr.upper) {
|
||||
list_for_each_entry(p, &sr->local_publ, local_publ) {
|
||||
if (p->scope == scope || (!exact && p->scope < scope))
|
||||
tipc_dest_push(dports, 0, p->port);
|
||||
tipc_dest_push(dports, 0, p->sk.ref);
|
||||
}
|
||||
}
|
||||
spin_unlock_bh(&sc->lock);
|
||||
@ -683,26 +697,27 @@ exit:
|
||||
rcu_read_unlock();
|
||||
}
|
||||
|
||||
/* tipc_nametbl_lookup_dst_nodes - find broadcast destination nodes
|
||||
* - Creates list of nodes that overlap the given multicast address
|
||||
* - Determines if any node local destinations overlap
|
||||
/* tipc_nametbl_lookup_mcast_nodes(): look up all destination nodes matching
|
||||
* the given address. Used in sending node.
|
||||
* Used on nodes which are sending out a multicast/broadcast message
|
||||
* Returns a list of nodes, including own node if applicable
|
||||
*/
|
||||
void tipc_nametbl_lookup_dst_nodes(struct net *net, u32 type, u32 lower,
|
||||
u32 upper, struct tipc_nlist *nodes)
|
||||
void tipc_nametbl_lookup_mcast_nodes(struct net *net, struct tipc_uaddr *ua,
|
||||
struct tipc_nlist *nodes)
|
||||
{
|
||||
struct service_range *sr;
|
||||
struct tipc_service *sc;
|
||||
struct publication *p;
|
||||
|
||||
rcu_read_lock();
|
||||
sc = tipc_service_find(net, type);
|
||||
sc = tipc_service_find(net, ua);
|
||||
if (!sc)
|
||||
goto exit;
|
||||
|
||||
spin_lock_bh(&sc->lock);
|
||||
service_range_foreach_match(sr, sc, lower, upper) {
|
||||
service_range_foreach_match(sr, sc, ua->sr.lower, ua->sr.upper) {
|
||||
list_for_each_entry(p, &sr->all_publ, all_publ) {
|
||||
tipc_nlist_add(nodes, p->node);
|
||||
tipc_nlist_add(nodes, p->sk.node);
|
||||
}
|
||||
}
|
||||
spin_unlock_bh(&sc->lock);
|
||||
@ -713,7 +728,7 @@ exit:
|
||||
/* tipc_nametbl_build_group - build list of communication group members
|
||||
*/
|
||||
void tipc_nametbl_build_group(struct net *net, struct tipc_group *grp,
|
||||
u32 type, u32 scope)
|
||||
struct tipc_uaddr *ua)
|
||||
{
|
||||
struct service_range *sr;
|
||||
struct tipc_service *sc;
|
||||
@ -721,7 +736,7 @@ void tipc_nametbl_build_group(struct net *net, struct tipc_group *grp,
|
||||
struct rb_node *n;
|
||||
|
||||
rcu_read_lock();
|
||||
sc = tipc_service_find(net, type);
|
||||
sc = tipc_service_find(net, ua);
|
||||
if (!sc)
|
||||
goto exit;
|
||||
|
||||
@ -729,9 +744,10 @@ void tipc_nametbl_build_group(struct net *net, struct tipc_group *grp,
|
||||
for (n = rb_first(&sc->ranges); n; n = rb_next(n)) {
|
||||
sr = container_of(n, struct service_range, tree_node);
|
||||
list_for_each_entry(p, &sr->all_publ, all_publ) {
|
||||
if (p->scope != scope)
|
||||
if (p->scope != ua->scope)
|
||||
continue;
|
||||
tipc_group_add_member(grp, p->node, p->port, p->lower);
|
||||
tipc_group_add_member(grp, p->sk.node, p->sk.ref,
|
||||
p->sr.lower);
|
||||
}
|
||||
}
|
||||
spin_unlock_bh(&sc->lock);
|
||||
@ -741,9 +757,8 @@ exit:
|
||||
|
||||
/* tipc_nametbl_publish - add service binding to name table
|
||||
*/
|
||||
struct publication *tipc_nametbl_publish(struct net *net, u32 type, u32 lower,
|
||||
u32 upper, u32 scope, u32 port,
|
||||
u32 key)
|
||||
struct publication *tipc_nametbl_publish(struct net *net, struct tipc_uaddr *ua,
|
||||
struct tipc_socket_addr *sk, u32 key)
|
||||
{
|
||||
struct name_table *nt = tipc_name_table(net);
|
||||
struct tipc_net *tn = tipc_net(net);
|
||||
@ -758,8 +773,7 @@ struct publication *tipc_nametbl_publish(struct net *net, u32 type, u32 lower,
|
||||
goto exit;
|
||||
}
|
||||
|
||||
p = tipc_nametbl_insert_publ(net, type, lower, upper, scope,
|
||||
tipc_own_addr(net), port, key);
|
||||
p = tipc_nametbl_insert_publ(net, ua, sk, key);
|
||||
if (p) {
|
||||
nt->local_publ_count++;
|
||||
skb = tipc_named_publish(net, p);
|
||||
@ -777,41 +791,33 @@ exit:
|
||||
/**
|
||||
* tipc_nametbl_withdraw - withdraw a service binding
|
||||
* @net: network namespace
|
||||
* @type: service type
|
||||
* @lower: service range lower bound
|
||||
* @upper: service range upper bound
|
||||
* @ua: service address/range being unbound
|
||||
* @sk: address of the socket being unbound from
|
||||
* @key: target publication key
|
||||
*/
|
||||
int tipc_nametbl_withdraw(struct net *net, u32 type, u32 lower,
|
||||
u32 upper, u32 key)
|
||||
void tipc_nametbl_withdraw(struct net *net, struct tipc_uaddr *ua,
|
||||
struct tipc_socket_addr *sk, u32 key)
|
||||
{
|
||||
struct name_table *nt = tipc_name_table(net);
|
||||
struct tipc_net *tn = tipc_net(net);
|
||||
u32 self = tipc_own_addr(net);
|
||||
struct sk_buff *skb = NULL;
|
||||
struct publication *p;
|
||||
u32 rc_dests;
|
||||
|
||||
spin_lock_bh(&tn->nametbl_lock);
|
||||
|
||||
p = tipc_nametbl_remove_publ(net, type, lower, upper, self, key);
|
||||
p = tipc_nametbl_remove_publ(net, ua, sk, key);
|
||||
if (p) {
|
||||
nt->local_publ_count--;
|
||||
skb = tipc_named_withdraw(net, p);
|
||||
list_del_init(&p->binding_sock);
|
||||
kfree_rcu(p, rcu);
|
||||
} else {
|
||||
pr_err("Failed to remove local publication {%u,%u,%u}/%u\n",
|
||||
type, lower, upper, key);
|
||||
}
|
||||
rc_dests = nt->rc_dests;
|
||||
spin_unlock_bh(&tn->nametbl_lock);
|
||||
|
||||
if (skb) {
|
||||
if (skb)
|
||||
tipc_node_broadcast(net, skb, rc_dests);
|
||||
return 1;
|
||||
}
|
||||
return 0;
|
||||
}
|
||||
|
||||
/**
|
||||
@ -820,25 +826,25 @@ int tipc_nametbl_withdraw(struct net *net, u32 type, u32 lower,
|
||||
*/
|
||||
bool tipc_nametbl_subscribe(struct tipc_subscription *sub)
|
||||
{
|
||||
struct name_table *nt = tipc_name_table(sub->net);
|
||||
struct tipc_net *tn = tipc_net(sub->net);
|
||||
struct tipc_subscr *s = &sub->evt.s;
|
||||
u32 type = tipc_sub_read(s, seq.type);
|
||||
u32 type = sub->s.seq.type;
|
||||
struct tipc_service *sc;
|
||||
struct tipc_uaddr ua;
|
||||
bool res = true;
|
||||
|
||||
tipc_uaddr(&ua, TIPC_SERVICE_RANGE, TIPC_NODE_SCOPE, type,
|
||||
sub->s.seq.lower, sub->s.seq.upper);
|
||||
spin_lock_bh(&tn->nametbl_lock);
|
||||
sc = tipc_service_find(sub->net, type);
|
||||
sc = tipc_service_find(sub->net, &ua);
|
||||
if (!sc)
|
||||
sc = tipc_service_create(type, &nt->services[hash(type)]);
|
||||
sc = tipc_service_create(sub->net, &ua);
|
||||
if (sc) {
|
||||
spin_lock_bh(&sc->lock);
|
||||
tipc_service_subscribe(sc, sub);
|
||||
spin_unlock_bh(&sc->lock);
|
||||
} else {
|
||||
pr_warn("Failed to subscribe for {%u,%u,%u}\n", type,
|
||||
tipc_sub_read(s, seq.lower),
|
||||
tipc_sub_read(s, seq.upper));
|
||||
pr_warn("Failed to subscribe for {%u,%u,%u}\n",
|
||||
type, sub->s.seq.lower, sub->s.seq.upper);
|
||||
res = false;
|
||||
}
|
||||
spin_unlock_bh(&tn->nametbl_lock);
|
||||
@ -852,12 +858,13 @@ bool tipc_nametbl_subscribe(struct tipc_subscription *sub)
|
||||
void tipc_nametbl_unsubscribe(struct tipc_subscription *sub)
|
||||
{
|
||||
struct tipc_net *tn = tipc_net(sub->net);
|
||||
struct tipc_subscr *s = &sub->evt.s;
|
||||
u32 type = tipc_sub_read(s, seq.type);
|
||||
struct tipc_service *sc;
|
||||
struct tipc_uaddr ua;
|
||||
|
||||
tipc_uaddr(&ua, TIPC_SERVICE_RANGE, TIPC_NODE_SCOPE,
|
||||
sub->s.seq.type, sub->s.seq.lower, sub->s.seq.upper);
|
||||
spin_lock_bh(&tn->nametbl_lock);
|
||||
sc = tipc_service_find(sub->net, type);
|
||||
sc = tipc_service_find(sub->net, &ua);
|
||||
if (!sc)
|
||||
goto exit;
|
||||
|
||||
@ -909,7 +916,7 @@ static void tipc_service_delete(struct net *net, struct tipc_service *sc)
|
||||
spin_lock_bh(&sc->lock);
|
||||
rbtree_postorder_for_each_entry_safe(sr, tmpr, &sc->ranges, tree_node) {
|
||||
list_for_each_entry_safe(p, tmp, &sr->all_publ, all_publ) {
|
||||
tipc_service_remove_publ(sr, p->node, p->key);
|
||||
tipc_service_remove_publ(sr, &p->sk, p->key);
|
||||
kfree_rcu(p, rcu);
|
||||
}
|
||||
rb_erase_augmented(&sr->tree_node, &sc->ranges, &sr_callbacks);
|
||||
@ -993,9 +1000,9 @@ static int __tipc_nl_add_nametable_publ(struct tipc_nl_msg *msg,
|
||||
goto publ_msg_full;
|
||||
if (nla_put_u32(msg->skb, TIPC_NLA_PUBL_SCOPE, p->scope))
|
||||
goto publ_msg_full;
|
||||
if (nla_put_u32(msg->skb, TIPC_NLA_PUBL_NODE, p->node))
|
||||
if (nla_put_u32(msg->skb, TIPC_NLA_PUBL_NODE, p->sk.node))
|
||||
goto publ_msg_full;
|
||||
if (nla_put_u32(msg->skb, TIPC_NLA_PUBL_REF, p->port))
|
||||
if (nla_put_u32(msg->skb, TIPC_NLA_PUBL_REF, p->sk.ref))
|
||||
goto publ_msg_full;
|
||||
if (nla_put_u32(msg->skb, TIPC_NLA_PUBL_KEY, p->key))
|
||||
goto publ_msg_full;
|
||||
@ -1046,6 +1053,7 @@ static int tipc_nl_service_list(struct net *net, struct tipc_nl_msg *msg,
|
||||
struct tipc_net *tn = tipc_net(net);
|
||||
struct tipc_service *service = NULL;
|
||||
struct hlist_head *head;
|
||||
struct tipc_uaddr ua;
|
||||
int err;
|
||||
int i;
|
||||
|
||||
@ -1059,7 +1067,9 @@ static int tipc_nl_service_list(struct net *net, struct tipc_nl_msg *msg,
|
||||
|
||||
if (*last_type ||
|
||||
(!i && *last_key && (*last_lower == *last_key))) {
|
||||
service = tipc_service_find(net, *last_type);
|
||||
tipc_uaddr(&ua, TIPC_SERVICE_RANGE, TIPC_NODE_SCOPE,
|
||||
*last_type, *last_lower, *last_lower);
|
||||
service = tipc_service_find(net, &ua);
|
||||
if (!service)
|
||||
return -EPIPE;
|
||||
} else {
|
||||
|
Reference in New Issue
Block a user