2010-12-13 14:19:28 +03:00
/*
2011-01-27 12:38:15 +03:00
* Copyright ( C ) 2009 - 2011 B . A . T . M . A . N . contributors :
2010-12-13 14:19:28 +03:00
*
* Marek Lindner , Simon Wunderlich
*
* This program is free software ; you can redistribute it and / or
* modify it under the terms of version 2 of the GNU General Public
* License as published by the Free Software Foundation .
*
* This program is distributed in the hope that it will be useful , but
* WITHOUT ANY WARRANTY ; without even the implied warranty of
* MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE . See the GNU
* General Public License for more details .
*
* You should have received a copy of the GNU General Public License
* along with this program ; if not , write to the Free Software
* Foundation , Inc . , 51 Franklin Street , Fifth Floor , Boston , MA
* 02110 - 1301 , USA
*
*/
/* increase the reference counter for this originator */
# include "main.h"
# include "originator.h"
# include "hash.h"
# include "translation-table.h"
# include "routing.h"
# include "gateway_client.h"
# include "hard-interface.h"
# include "unicast.h"
# include "soft-interface.h"
static void purge_orig ( struct work_struct * work ) ;
static void start_purge_timer ( struct bat_priv * bat_priv )
{
INIT_DELAYED_WORK ( & bat_priv - > orig_work , purge_orig ) ;
queue_delayed_work ( bat_event_workqueue , & bat_priv - > orig_work , 1 * HZ ) ;
}
int originator_init ( struct bat_priv * bat_priv )
{
if ( bat_priv - > orig_hash )
return 1 ;
spin_lock_bh ( & bat_priv - > orig_hash_lock ) ;
bat_priv - > orig_hash = hash_new ( 1024 ) ;
if ( ! bat_priv - > orig_hash )
goto err ;
spin_unlock_bh ( & bat_priv - > orig_hash_lock ) ;
start_purge_timer ( bat_priv ) ;
return 1 ;
err :
spin_unlock_bh ( & bat_priv - > orig_hash_lock ) ;
return 0 ;
}
2010-12-13 00:57:10 +03:00
void neigh_node_free_ref ( struct kref * refcount )
{
struct neigh_node * neigh_node ;
neigh_node = container_of ( refcount , struct neigh_node , refcount ) ;
kfree ( neigh_node ) ;
}
2010-12-13 00:57:12 +03:00
static void neigh_node_free_rcu ( struct rcu_head * rcu )
{
struct neigh_node * neigh_node ;
neigh_node = container_of ( rcu , struct neigh_node , rcu ) ;
kref_put ( & neigh_node - > refcount , neigh_node_free_ref ) ;
}
2010-12-13 00:57:10 +03:00
struct neigh_node * create_neighbor ( struct orig_node * orig_node ,
struct orig_node * orig_neigh_node ,
uint8_t * neigh ,
struct batman_if * if_incoming )
2010-12-13 14:19:28 +03:00
{
struct bat_priv * bat_priv = netdev_priv ( if_incoming - > soft_iface ) ;
struct neigh_node * neigh_node ;
bat_dbg ( DBG_BATMAN , bat_priv ,
" Creating new last-hop neighbor of originator \n " ) ;
neigh_node = kzalloc ( sizeof ( struct neigh_node ) , GFP_ATOMIC ) ;
if ( ! neigh_node )
return NULL ;
2010-12-13 00:57:11 +03:00
INIT_HLIST_NODE ( & neigh_node - > list ) ;
2010-12-13 14:19:28 +03:00
memcpy ( neigh_node - > addr , neigh , ETH_ALEN ) ;
neigh_node - > orig_node = orig_neigh_node ;
neigh_node - > if_incoming = if_incoming ;
2010-12-13 00:57:10 +03:00
kref_init ( & neigh_node - > refcount ) ;
2010-12-13 14:19:28 +03:00
2010-12-13 00:57:12 +03:00
spin_lock_bh ( & orig_node - > neigh_list_lock ) ;
hlist_add_head_rcu ( & neigh_node - > list , & orig_node - > neigh_list ) ;
spin_unlock_bh ( & orig_node - > neigh_list_lock ) ;
2010-12-13 14:19:28 +03:00
return neigh_node ;
}
2011-01-19 23:01:42 +03:00
void orig_node_free_ref ( struct kref * refcount )
2010-12-13 14:19:28 +03:00
{
2010-12-13 00:57:11 +03:00
struct hlist_node * node , * node_tmp ;
2010-12-13 14:19:28 +03:00
struct neigh_node * neigh_node ;
2011-01-19 23:01:42 +03:00
struct orig_node * orig_node ;
orig_node = container_of ( refcount , struct orig_node , refcount ) ;
2010-12-13 14:19:28 +03:00
2010-12-13 00:57:12 +03:00
spin_lock_bh ( & orig_node - > neigh_list_lock ) ;
2010-12-13 14:19:28 +03:00
/* for all neighbors towards this originator ... */
2010-12-13 00:57:11 +03:00
hlist_for_each_entry_safe ( neigh_node , node , node_tmp ,
& orig_node - > neigh_list , list ) {
2010-12-13 00:57:12 +03:00
hlist_del_rcu ( & neigh_node - > list ) ;
call_rcu ( & neigh_node - > rcu , neigh_node_free_rcu ) ;
2010-12-13 14:19:28 +03:00
}
2010-12-13 00:57:12 +03:00
spin_unlock_bh ( & orig_node - > neigh_list_lock ) ;
2010-12-13 14:19:28 +03:00
frag_list_free ( & orig_node - > frag_list ) ;
2011-01-19 23:01:42 +03:00
hna_global_del_orig ( orig_node - > bat_priv , orig_node ,
" originator timed out " ) ;
2010-12-13 14:19:28 +03:00
kfree ( orig_node - > bcast_own ) ;
kfree ( orig_node - > bcast_own_sum ) ;
kfree ( orig_node ) ;
}
void originator_free ( struct bat_priv * bat_priv )
{
2011-01-19 23:01:42 +03:00
struct hashtable_t * hash = bat_priv - > orig_hash ;
struct hlist_node * walk , * safe ;
struct hlist_head * head ;
struct element_t * bucket ;
spinlock_t * list_lock ; /* spinlock to protect write access */
struct orig_node * orig_node ;
int i ;
if ( ! hash )
2010-12-13 14:19:28 +03:00
return ;
cancel_delayed_work_sync ( & bat_priv - > orig_work ) ;
spin_lock_bh ( & bat_priv - > orig_hash_lock ) ;
bat_priv - > orig_hash = NULL ;
2011-01-19 23:01:42 +03:00
for ( i = 0 ; i < hash - > size ; i + + ) {
head = & hash - > table [ i ] ;
list_lock = & hash - > list_locks [ i ] ;
spin_lock_bh ( list_lock ) ;
hlist_for_each_entry_safe ( bucket , walk , safe , head , hlist ) {
orig_node = bucket - > data ;
hlist_del_rcu ( walk ) ;
call_rcu ( & bucket - > rcu , bucket_free_rcu ) ;
kref_put ( & orig_node - > refcount , orig_node_free_ref ) ;
}
spin_unlock_bh ( list_lock ) ;
}
hash_destroy ( hash ) ;
2010-12-13 14:19:28 +03:00
spin_unlock_bh ( & bat_priv - > orig_hash_lock ) ;
}
2011-01-19 23:01:42 +03:00
static void bucket_free_orig_rcu ( struct rcu_head * rcu )
{
struct element_t * bucket ;
struct orig_node * orig_node ;
bucket = container_of ( rcu , struct element_t , rcu ) ;
orig_node = bucket - > data ;
kref_put ( & orig_node - > refcount , orig_node_free_ref ) ;
kfree ( bucket ) ;
}
2010-12-13 14:19:28 +03:00
/* this function finds or creates an originator entry for the given
* address if it does not exits */
struct orig_node * get_orig_node ( struct bat_priv * bat_priv , uint8_t * addr )
{
struct orig_node * orig_node ;
int size ;
int hash_added ;
2011-01-19 23:01:40 +03:00
rcu_read_lock ( ) ;
2010-12-13 14:19:28 +03:00
orig_node = ( ( struct orig_node * ) hash_find ( bat_priv - > orig_hash ,
compare_orig , choose_orig ,
addr ) ) ;
2011-01-19 23:01:40 +03:00
rcu_read_unlock ( ) ;
2010-12-13 14:19:28 +03:00
2011-01-19 23:01:42 +03:00
if ( orig_node ) {
kref_get ( & orig_node - > refcount ) ;
2010-12-13 14:19:28 +03:00
return orig_node ;
2011-01-19 23:01:42 +03:00
}
2010-12-13 14:19:28 +03:00
bat_dbg ( DBG_BATMAN , bat_priv ,
" Creating new originator: %pM \n " , addr ) ;
orig_node = kzalloc ( sizeof ( struct orig_node ) , GFP_ATOMIC ) ;
if ( ! orig_node )
return NULL ;
2010-12-13 00:57:11 +03:00
INIT_HLIST_HEAD ( & orig_node - > neigh_list ) ;
2010-12-13 00:57:12 +03:00
spin_lock_init ( & orig_node - > neigh_list_lock ) ;
2011-01-19 23:01:42 +03:00
kref_init ( & orig_node - > refcount ) ;
2010-12-13 14:19:28 +03:00
2011-01-19 23:01:42 +03:00
orig_node - > bat_priv = bat_priv ;
2010-12-13 14:19:28 +03:00
memcpy ( orig_node - > orig , addr , ETH_ALEN ) ;
orig_node - > router = NULL ;
orig_node - > hna_buff = NULL ;
orig_node - > bcast_seqno_reset = jiffies - 1
- msecs_to_jiffies ( RESET_PROTECTION_MS ) ;
orig_node - > batman_seqno_reset = jiffies - 1
- msecs_to_jiffies ( RESET_PROTECTION_MS ) ;
size = bat_priv - > num_ifaces * sizeof ( unsigned long ) * NUM_WORDS ;
orig_node - > bcast_own = kzalloc ( size , GFP_ATOMIC ) ;
if ( ! orig_node - > bcast_own )
goto free_orig_node ;
size = bat_priv - > num_ifaces * sizeof ( uint8_t ) ;
orig_node - > bcast_own_sum = kzalloc ( size , GFP_ATOMIC ) ;
INIT_LIST_HEAD ( & orig_node - > frag_list ) ;
orig_node - > last_frag_packet = 0 ;
if ( ! orig_node - > bcast_own_sum )
goto free_bcast_own ;
hash_added = hash_add ( bat_priv - > orig_hash , compare_orig , choose_orig ,
orig_node ) ;
if ( hash_added < 0 )
goto free_bcast_own_sum ;
2011-01-19 23:01:42 +03:00
/* extra reference for return */
kref_get ( & orig_node - > refcount ) ;
2010-12-13 14:19:28 +03:00
return orig_node ;
free_bcast_own_sum :
kfree ( orig_node - > bcast_own_sum ) ;
free_bcast_own :
kfree ( orig_node - > bcast_own ) ;
free_orig_node :
kfree ( orig_node ) ;
return NULL ;
}
static bool purge_orig_neighbors ( struct bat_priv * bat_priv ,
struct orig_node * orig_node ,
struct neigh_node * * best_neigh_node )
{
2010-12-13 00:57:11 +03:00
struct hlist_node * node , * node_tmp ;
2010-12-13 14:19:28 +03:00
struct neigh_node * neigh_node ;
bool neigh_purged = false ;
* best_neigh_node = NULL ;
2010-12-13 00:57:12 +03:00
spin_lock_bh ( & orig_node - > neigh_list_lock ) ;
2010-12-13 14:19:28 +03:00
/* for all neighbors towards this originator ... */
2010-12-13 00:57:11 +03:00
hlist_for_each_entry_safe ( neigh_node , node , node_tmp ,
& orig_node - > neigh_list , list ) {
2010-12-13 14:19:28 +03:00
if ( ( time_after ( jiffies ,
neigh_node - > last_valid + PURGE_TIMEOUT * HZ ) ) | |
( neigh_node - > if_incoming - > if_status = = IF_INACTIVE ) | |
2011-01-19 22:16:10 +03:00
( neigh_node - > if_incoming - > if_status = = IF_NOT_IN_USE ) | |
2010-12-13 14:19:28 +03:00
( neigh_node - > if_incoming - > if_status = = IF_TO_BE_REMOVED ) ) {
2011-01-19 22:16:10 +03:00
if ( ( neigh_node - > if_incoming - > if_status = =
IF_INACTIVE ) | |
( neigh_node - > if_incoming - > if_status = =
IF_NOT_IN_USE ) | |
( neigh_node - > if_incoming - > if_status = =
IF_TO_BE_REMOVED ) )
2010-12-13 14:19:28 +03:00
bat_dbg ( DBG_BATMAN , bat_priv ,
" neighbor purge: originator %pM, "
" neighbor: %pM, iface: %s \n " ,
orig_node - > orig , neigh_node - > addr ,
neigh_node - > if_incoming - > net_dev - > name ) ;
else
bat_dbg ( DBG_BATMAN , bat_priv ,
" neighbor timeout: originator %pM, "
" neighbor: %pM, last_valid: %lu \n " ,
orig_node - > orig , neigh_node - > addr ,
( neigh_node - > last_valid / HZ ) ) ;
neigh_purged = true ;
2010-12-13 00:57:11 +03:00
2010-12-13 00:57:12 +03:00
hlist_del_rcu ( & neigh_node - > list ) ;
call_rcu ( & neigh_node - > rcu , neigh_node_free_rcu ) ;
2010-12-13 14:19:28 +03:00
} else {
if ( ( ! * best_neigh_node ) | |
( neigh_node - > tq_avg > ( * best_neigh_node ) - > tq_avg ) )
* best_neigh_node = neigh_node ;
}
}
2010-12-13 00:57:12 +03:00
spin_unlock_bh ( & orig_node - > neigh_list_lock ) ;
2010-12-13 14:19:28 +03:00
return neigh_purged ;
}
static bool purge_orig_node ( struct bat_priv * bat_priv ,
struct orig_node * orig_node )
{
struct neigh_node * best_neigh_node ;
if ( time_after ( jiffies ,
orig_node - > last_valid + 2 * PURGE_TIMEOUT * HZ ) ) {
bat_dbg ( DBG_BATMAN , bat_priv ,
" Originator timeout: originator %pM, last_valid %lu \n " ,
orig_node - > orig , ( orig_node - > last_valid / HZ ) ) ;
return true ;
} else {
if ( purge_orig_neighbors ( bat_priv , orig_node ,
& best_neigh_node ) ) {
update_routes ( bat_priv , orig_node ,
best_neigh_node ,
orig_node - > hna_buff ,
orig_node - > hna_buff_len ) ;
/* update bonding candidates, we could have lost
* some candidates . */
2010-12-29 19:15:19 +03:00
update_bonding_candidates ( orig_node ) ;
2010-12-13 14:19:28 +03:00
}
}
return false ;
}
static void _purge_orig ( struct bat_priv * bat_priv )
{
struct hashtable_t * hash = bat_priv - > orig_hash ;
struct hlist_node * walk , * safe ;
struct hlist_head * head ;
struct element_t * bucket ;
2011-01-19 23:01:40 +03:00
spinlock_t * list_lock ; /* spinlock to protect write access */
2010-12-13 14:19:28 +03:00
struct orig_node * orig_node ;
int i ;
if ( ! hash )
return ;
spin_lock_bh ( & bat_priv - > orig_hash_lock ) ;
/* for all origins... */
for ( i = 0 ; i < hash - > size ; i + + ) {
head = & hash - > table [ i ] ;
2011-01-19 23:01:40 +03:00
list_lock = & hash - > list_locks [ i ] ;
2010-12-13 14:19:28 +03:00
2011-01-19 23:01:40 +03:00
spin_lock_bh ( list_lock ) ;
2010-12-13 14:19:28 +03:00
hlist_for_each_entry_safe ( bucket , walk , safe , head , hlist ) {
orig_node = bucket - > data ;
if ( purge_orig_node ( bat_priv , orig_node ) ) {
if ( orig_node - > gw_flags )
gw_node_delete ( bat_priv , orig_node ) ;
2011-01-19 23:01:40 +03:00
hlist_del_rcu ( walk ) ;
2011-01-19 23:01:42 +03:00
call_rcu ( & bucket - > rcu , bucket_free_orig_rcu ) ;
2011-01-19 23:01:40 +03:00
continue ;
2010-12-13 14:19:28 +03:00
}
if ( time_after ( jiffies , orig_node - > last_frag_packet +
msecs_to_jiffies ( FRAG_TIMEOUT ) ) )
frag_list_free ( & orig_node - > frag_list ) ;
}
2011-01-19 23:01:40 +03:00
spin_unlock_bh ( list_lock ) ;
2010-12-13 14:19:28 +03:00
}
spin_unlock_bh ( & bat_priv - > orig_hash_lock ) ;
gw_node_purge ( bat_priv ) ;
gw_election ( bat_priv ) ;
softif_neigh_purge ( bat_priv ) ;
}
static void purge_orig ( struct work_struct * work )
{
struct delayed_work * delayed_work =
container_of ( work , struct delayed_work , work ) ;
struct bat_priv * bat_priv =
container_of ( delayed_work , struct bat_priv , orig_work ) ;
_purge_orig ( bat_priv ) ;
start_purge_timer ( bat_priv ) ;
}
void purge_orig_ref ( struct bat_priv * bat_priv )
{
_purge_orig ( bat_priv ) ;
}
int orig_seq_print_text ( struct seq_file * seq , void * offset )
{
struct net_device * net_dev = ( struct net_device * ) seq - > private ;
struct bat_priv * bat_priv = netdev_priv ( net_dev ) ;
struct hashtable_t * hash = bat_priv - > orig_hash ;
2010-12-13 00:57:11 +03:00
struct hlist_node * walk , * node ;
2010-12-13 14:19:28 +03:00
struct hlist_head * head ;
struct element_t * bucket ;
struct orig_node * orig_node ;
struct neigh_node * neigh_node ;
int batman_count = 0 ;
int last_seen_secs ;
int last_seen_msecs ;
int i ;
if ( ( ! bat_priv - > primary_if ) | |
( bat_priv - > primary_if - > if_status ! = IF_ACTIVE ) ) {
if ( ! bat_priv - > primary_if )
return seq_printf ( seq , " BATMAN mesh %s disabled - "
" please specify interfaces to enable it \n " ,
net_dev - > name ) ;
return seq_printf ( seq , " BATMAN mesh %s "
" disabled - primary interface not active \n " ,
net_dev - > name ) ;
}
seq_printf ( seq , " [B.A.T.M.A.N. adv %s%s, MainIF/MAC: %s/%pM (%s)] \n " ,
SOURCE_VERSION , REVISION_VERSION_STR ,
bat_priv - > primary_if - > net_dev - > name ,
bat_priv - > primary_if - > net_dev - > dev_addr , net_dev - > name ) ;
seq_printf ( seq , " %-15s %s (%s/%i) %17s [%10s]: %20s ... \n " ,
" Originator " , " last-seen " , " # " , TQ_MAX_VALUE , " Nexthop " ,
" outgoingIF " , " Potential nexthops " ) ;
spin_lock_bh ( & bat_priv - > orig_hash_lock ) ;
for ( i = 0 ; i < hash - > size ; i + + ) {
head = & hash - > table [ i ] ;
2011-01-19 23:01:40 +03:00
rcu_read_lock ( ) ;
hlist_for_each_entry_rcu ( bucket , walk , head , hlist ) {
2010-12-13 14:19:28 +03:00
orig_node = bucket - > data ;
if ( ! orig_node - > router )
continue ;
if ( orig_node - > router - > tq_avg = = 0 )
continue ;
last_seen_secs = jiffies_to_msecs ( jiffies -
orig_node - > last_valid ) / 1000 ;
last_seen_msecs = jiffies_to_msecs ( jiffies -
orig_node - > last_valid ) % 1000 ;
neigh_node = orig_node - > router ;
seq_printf ( seq , " %pM %4i.%03is (%3i) %pM [%10s]: " ,
orig_node - > orig , last_seen_secs ,
last_seen_msecs , neigh_node - > tq_avg ,
neigh_node - > addr ,
neigh_node - > if_incoming - > net_dev - > name ) ;
2010-12-13 00:57:12 +03:00
hlist_for_each_entry_rcu ( neigh_node , node ,
& orig_node - > neigh_list , list ) {
2010-12-13 14:19:28 +03:00
seq_printf ( seq , " %pM (%3i) " , neigh_node - > addr ,
neigh_node - > tq_avg ) ;
}
seq_printf ( seq , " \n " ) ;
batman_count + + ;
}
2011-01-19 23:01:40 +03:00
rcu_read_unlock ( ) ;
2010-12-13 14:19:28 +03:00
}
spin_unlock_bh ( & bat_priv - > orig_hash_lock ) ;
if ( ( batman_count = = 0 ) )
seq_printf ( seq , " No batman nodes in range ... \n " ) ;
return 0 ;
}
static int orig_node_add_if ( struct orig_node * orig_node , int max_if_num )
{
void * data_ptr ;
data_ptr = kmalloc ( max_if_num * sizeof ( unsigned long ) * NUM_WORDS ,
GFP_ATOMIC ) ;
if ( ! data_ptr ) {
pr_err ( " Can't resize orig: out of memory \n " ) ;
return - 1 ;
}
memcpy ( data_ptr , orig_node - > bcast_own ,
( max_if_num - 1 ) * sizeof ( unsigned long ) * NUM_WORDS ) ;
kfree ( orig_node - > bcast_own ) ;
orig_node - > bcast_own = data_ptr ;
data_ptr = kmalloc ( max_if_num * sizeof ( uint8_t ) , GFP_ATOMIC ) ;
if ( ! data_ptr ) {
pr_err ( " Can't resize orig: out of memory \n " ) ;
return - 1 ;
}
memcpy ( data_ptr , orig_node - > bcast_own_sum ,
( max_if_num - 1 ) * sizeof ( uint8_t ) ) ;
kfree ( orig_node - > bcast_own_sum ) ;
orig_node - > bcast_own_sum = data_ptr ;
return 0 ;
}
int orig_hash_add_if ( struct batman_if * batman_if , int max_if_num )
{
struct bat_priv * bat_priv = netdev_priv ( batman_if - > soft_iface ) ;
struct hashtable_t * hash = bat_priv - > orig_hash ;
struct hlist_node * walk ;
struct hlist_head * head ;
struct element_t * bucket ;
struct orig_node * orig_node ;
int i ;
/* resize all orig nodes because orig_node->bcast_own(_sum) depend on
* if_num */
spin_lock_bh ( & bat_priv - > orig_hash_lock ) ;
for ( i = 0 ; i < hash - > size ; i + + ) {
head = & hash - > table [ i ] ;
2011-01-19 23:01:40 +03:00
rcu_read_lock ( ) ;
hlist_for_each_entry_rcu ( bucket , walk , head , hlist ) {
2010-12-13 14:19:28 +03:00
orig_node = bucket - > data ;
if ( orig_node_add_if ( orig_node , max_if_num ) = = - 1 )
goto err ;
}
2011-01-19 23:01:40 +03:00
rcu_read_unlock ( ) ;
2010-12-13 14:19:28 +03:00
}
spin_unlock_bh ( & bat_priv - > orig_hash_lock ) ;
return 0 ;
err :
2011-01-19 23:01:40 +03:00
rcu_read_unlock ( ) ;
2010-12-13 14:19:28 +03:00
spin_unlock_bh ( & bat_priv - > orig_hash_lock ) ;
return - ENOMEM ;
}
static int orig_node_del_if ( struct orig_node * orig_node ,
int max_if_num , int del_if_num )
{
void * data_ptr = NULL ;
int chunk_size ;
/* last interface was removed */
if ( max_if_num = = 0 )
goto free_bcast_own ;
chunk_size = sizeof ( unsigned long ) * NUM_WORDS ;
data_ptr = kmalloc ( max_if_num * chunk_size , GFP_ATOMIC ) ;
if ( ! data_ptr ) {
pr_err ( " Can't resize orig: out of memory \n " ) ;
return - 1 ;
}
/* copy first part */
memcpy ( data_ptr , orig_node - > bcast_own , del_if_num * chunk_size ) ;
/* copy second part */
memcpy ( data_ptr + del_if_num * chunk_size ,
orig_node - > bcast_own + ( ( del_if_num + 1 ) * chunk_size ) ,
( max_if_num - del_if_num ) * chunk_size ) ;
free_bcast_own :
kfree ( orig_node - > bcast_own ) ;
orig_node - > bcast_own = data_ptr ;
if ( max_if_num = = 0 )
goto free_own_sum ;
data_ptr = kmalloc ( max_if_num * sizeof ( uint8_t ) , GFP_ATOMIC ) ;
if ( ! data_ptr ) {
pr_err ( " Can't resize orig: out of memory \n " ) ;
return - 1 ;
}
memcpy ( data_ptr , orig_node - > bcast_own_sum ,
del_if_num * sizeof ( uint8_t ) ) ;
memcpy ( data_ptr + del_if_num * sizeof ( uint8_t ) ,
orig_node - > bcast_own_sum + ( ( del_if_num + 1 ) * sizeof ( uint8_t ) ) ,
( max_if_num - del_if_num ) * sizeof ( uint8_t ) ) ;
free_own_sum :
kfree ( orig_node - > bcast_own_sum ) ;
orig_node - > bcast_own_sum = data_ptr ;
return 0 ;
}
int orig_hash_del_if ( struct batman_if * batman_if , int max_if_num )
{
struct bat_priv * bat_priv = netdev_priv ( batman_if - > soft_iface ) ;
struct hashtable_t * hash = bat_priv - > orig_hash ;
struct hlist_node * walk ;
struct hlist_head * head ;
struct element_t * bucket ;
struct batman_if * batman_if_tmp ;
struct orig_node * orig_node ;
int i , ret ;
/* resize all orig nodes because orig_node->bcast_own(_sum) depend on
* if_num */
spin_lock_bh ( & bat_priv - > orig_hash_lock ) ;
for ( i = 0 ; i < hash - > size ; i + + ) {
head = & hash - > table [ i ] ;
2011-01-19 23:01:40 +03:00
rcu_read_lock ( ) ;
hlist_for_each_entry_rcu ( bucket , walk , head , hlist ) {
2010-12-13 14:19:28 +03:00
orig_node = bucket - > data ;
ret = orig_node_del_if ( orig_node , max_if_num ,
batman_if - > if_num ) ;
if ( ret = = - 1 )
goto err ;
}
2011-01-19 23:01:40 +03:00
rcu_read_unlock ( ) ;
2010-12-13 14:19:28 +03:00
}
/* renumber remaining batman interfaces _inside_ of orig_hash_lock */
rcu_read_lock ( ) ;
list_for_each_entry_rcu ( batman_if_tmp , & if_list , list ) {
if ( batman_if_tmp - > if_status = = IF_NOT_IN_USE )
continue ;
if ( batman_if = = batman_if_tmp )
continue ;
if ( batman_if - > soft_iface ! = batman_if_tmp - > soft_iface )
continue ;
if ( batman_if_tmp - > if_num > batman_if - > if_num )
batman_if_tmp - > if_num - - ;
}
rcu_read_unlock ( ) ;
batman_if - > if_num = - 1 ;
spin_unlock_bh ( & bat_priv - > orig_hash_lock ) ;
return 0 ;
err :
2011-01-19 23:01:40 +03:00
rcu_read_unlock ( ) ;
2010-12-13 14:19:28 +03:00
spin_unlock_bh ( & bat_priv - > orig_hash_lock ) ;
return - ENOMEM ;
}