mirror of
https://github.com/FEX-Emu/linux.git
synced 2024-12-14 21:01:29 +00:00
7b36e8eef9
It might be possible that 2 threads access the same data in the same rcu grace period. The first thread calls call_rcu() to decrement the refcount and free the data while the second thread increases the refcount to use the data. To avoid this race condition all refcount operations have to be atomic. Reported-by: Sven Eckelmann <sven@narfation.org> Signed-off-by: Marek Lindner <lindner_marek@yahoo.de>
609 lines
15 KiB
C
609 lines
15 KiB
C
/*
|
|
* Copyright (C) 2007-2011 B.A.T.M.A.N. contributors:
|
|
*
|
|
* Marek Lindner, Simon Wunderlich
|
|
*
|
|
* This program is free software; you can redistribute it and/or
|
|
* modify it under the terms of version 2 of the GNU General Public
|
|
* License as published by the Free Software Foundation.
|
|
*
|
|
* This program is distributed in the hope that it will be useful, but
|
|
* WITHOUT ANY WARRANTY; without even the implied warranty of
|
|
* MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU
|
|
* General Public License for more details.
|
|
*
|
|
* You should have received a copy of the GNU General Public License
|
|
* along with this program; if not, write to the Free Software
|
|
* Foundation, Inc., 51 Franklin Street, Fifth Floor, Boston, MA
|
|
* 02110-1301, USA
|
|
*
|
|
*/
|
|
|
|
#include "main.h"
|
|
#include "translation-table.h"
|
|
#include "soft-interface.h"
|
|
#include "hash.h"
|
|
#include "originator.h"
|
|
|
|
static void hna_local_purge(struct work_struct *work);
|
|
static void _hna_global_del_orig(struct bat_priv *bat_priv,
|
|
struct hna_global_entry *hna_global_entry,
|
|
char *message);
|
|
|
|
/* returns 1 if they are the same mac addr */
|
|
static int compare_lhna(struct hlist_node *node, void *data2)
|
|
{
|
|
void *data1 = container_of(node, struct hna_local_entry, hash_entry);
|
|
|
|
return (memcmp(data1, data2, ETH_ALEN) == 0 ? 1 : 0);
|
|
}
|
|
|
|
/* returns 1 if they are the same mac addr */
|
|
static int compare_ghna(struct hlist_node *node, void *data2)
|
|
{
|
|
void *data1 = container_of(node, struct hna_global_entry, hash_entry);
|
|
|
|
return (memcmp(data1, data2, ETH_ALEN) == 0 ? 1 : 0);
|
|
}
|
|
|
|
static void hna_local_start_timer(struct bat_priv *bat_priv)
|
|
{
|
|
INIT_DELAYED_WORK(&bat_priv->hna_work, hna_local_purge);
|
|
queue_delayed_work(bat_event_workqueue, &bat_priv->hna_work, 10 * HZ);
|
|
}
|
|
|
|
static struct hna_local_entry *hna_local_hash_find(struct bat_priv *bat_priv,
|
|
void *data)
|
|
{
|
|
struct hashtable_t *hash = bat_priv->hna_local_hash;
|
|
struct hlist_head *head;
|
|
struct hlist_node *node;
|
|
struct hna_local_entry *hna_local_entry, *hna_local_entry_tmp = NULL;
|
|
int index;
|
|
|
|
if (!hash)
|
|
return NULL;
|
|
|
|
index = choose_orig(data, hash->size);
|
|
head = &hash->table[index];
|
|
|
|
rcu_read_lock();
|
|
hlist_for_each_entry_rcu(hna_local_entry, node, head, hash_entry) {
|
|
if (!compare_eth(hna_local_entry, data))
|
|
continue;
|
|
|
|
hna_local_entry_tmp = hna_local_entry;
|
|
break;
|
|
}
|
|
rcu_read_unlock();
|
|
|
|
return hna_local_entry_tmp;
|
|
}
|
|
|
|
static struct hna_global_entry *hna_global_hash_find(struct bat_priv *bat_priv,
|
|
void *data)
|
|
{
|
|
struct hashtable_t *hash = bat_priv->hna_global_hash;
|
|
struct hlist_head *head;
|
|
struct hlist_node *node;
|
|
struct hna_global_entry *hna_global_entry;
|
|
struct hna_global_entry *hna_global_entry_tmp = NULL;
|
|
int index;
|
|
|
|
if (!hash)
|
|
return NULL;
|
|
|
|
index = choose_orig(data, hash->size);
|
|
head = &hash->table[index];
|
|
|
|
rcu_read_lock();
|
|
hlist_for_each_entry_rcu(hna_global_entry, node, head, hash_entry) {
|
|
if (!compare_eth(hna_global_entry, data))
|
|
continue;
|
|
|
|
hna_global_entry_tmp = hna_global_entry;
|
|
break;
|
|
}
|
|
rcu_read_unlock();
|
|
|
|
return hna_global_entry_tmp;
|
|
}
|
|
|
|
int hna_local_init(struct bat_priv *bat_priv)
|
|
{
|
|
if (bat_priv->hna_local_hash)
|
|
return 1;
|
|
|
|
bat_priv->hna_local_hash = hash_new(1024);
|
|
|
|
if (!bat_priv->hna_local_hash)
|
|
return 0;
|
|
|
|
atomic_set(&bat_priv->hna_local_changed, 0);
|
|
hna_local_start_timer(bat_priv);
|
|
|
|
return 1;
|
|
}
|
|
|
|
void hna_local_add(struct net_device *soft_iface, uint8_t *addr)
|
|
{
|
|
struct bat_priv *bat_priv = netdev_priv(soft_iface);
|
|
struct hna_local_entry *hna_local_entry;
|
|
struct hna_global_entry *hna_global_entry;
|
|
int required_bytes;
|
|
|
|
spin_lock_bh(&bat_priv->hna_lhash_lock);
|
|
hna_local_entry = hna_local_hash_find(bat_priv, addr);
|
|
spin_unlock_bh(&bat_priv->hna_lhash_lock);
|
|
|
|
if (hna_local_entry) {
|
|
hna_local_entry->last_seen = jiffies;
|
|
return;
|
|
}
|
|
|
|
/* only announce as many hosts as possible in the batman-packet and
|
|
space in batman_packet->num_hna That also should give a limit to
|
|
MAC-flooding. */
|
|
required_bytes = (bat_priv->num_local_hna + 1) * ETH_ALEN;
|
|
required_bytes += BAT_PACKET_LEN;
|
|
|
|
if ((required_bytes > ETH_DATA_LEN) ||
|
|
(atomic_read(&bat_priv->aggregated_ogms) &&
|
|
required_bytes > MAX_AGGREGATION_BYTES) ||
|
|
(bat_priv->num_local_hna + 1 > 255)) {
|
|
bat_dbg(DBG_ROUTES, bat_priv,
|
|
"Can't add new local hna entry (%pM): "
|
|
"number of local hna entries exceeds packet size\n",
|
|
addr);
|
|
return;
|
|
}
|
|
|
|
bat_dbg(DBG_ROUTES, bat_priv,
|
|
"Creating new local hna entry: %pM\n", addr);
|
|
|
|
hna_local_entry = kmalloc(sizeof(struct hna_local_entry), GFP_ATOMIC);
|
|
if (!hna_local_entry)
|
|
return;
|
|
|
|
memcpy(hna_local_entry->addr, addr, ETH_ALEN);
|
|
hna_local_entry->last_seen = jiffies;
|
|
|
|
/* the batman interface mac address should never be purged */
|
|
if (compare_eth(addr, soft_iface->dev_addr))
|
|
hna_local_entry->never_purge = 1;
|
|
else
|
|
hna_local_entry->never_purge = 0;
|
|
|
|
spin_lock_bh(&bat_priv->hna_lhash_lock);
|
|
|
|
hash_add(bat_priv->hna_local_hash, compare_lhna, choose_orig,
|
|
hna_local_entry, &hna_local_entry->hash_entry);
|
|
bat_priv->num_local_hna++;
|
|
atomic_set(&bat_priv->hna_local_changed, 1);
|
|
|
|
spin_unlock_bh(&bat_priv->hna_lhash_lock);
|
|
|
|
/* remove address from global hash if present */
|
|
spin_lock_bh(&bat_priv->hna_ghash_lock);
|
|
|
|
hna_global_entry = hna_global_hash_find(bat_priv, addr);
|
|
|
|
if (hna_global_entry)
|
|
_hna_global_del_orig(bat_priv, hna_global_entry,
|
|
"local hna received");
|
|
|
|
spin_unlock_bh(&bat_priv->hna_ghash_lock);
|
|
}
|
|
|
|
int hna_local_fill_buffer(struct bat_priv *bat_priv,
|
|
unsigned char *buff, int buff_len)
|
|
{
|
|
struct hashtable_t *hash = bat_priv->hna_local_hash;
|
|
struct hna_local_entry *hna_local_entry;
|
|
struct hlist_node *node;
|
|
struct hlist_head *head;
|
|
int i, count = 0;
|
|
|
|
spin_lock_bh(&bat_priv->hna_lhash_lock);
|
|
|
|
for (i = 0; i < hash->size; i++) {
|
|
head = &hash->table[i];
|
|
|
|
rcu_read_lock();
|
|
hlist_for_each_entry_rcu(hna_local_entry, node,
|
|
head, hash_entry) {
|
|
if (buff_len < (count + 1) * ETH_ALEN)
|
|
break;
|
|
|
|
memcpy(buff + (count * ETH_ALEN), hna_local_entry->addr,
|
|
ETH_ALEN);
|
|
|
|
count++;
|
|
}
|
|
rcu_read_unlock();
|
|
}
|
|
|
|
/* if we did not get all new local hnas see you next time ;-) */
|
|
if (count == bat_priv->num_local_hna)
|
|
atomic_set(&bat_priv->hna_local_changed, 0);
|
|
|
|
spin_unlock_bh(&bat_priv->hna_lhash_lock);
|
|
return count;
|
|
}
|
|
|
|
int hna_local_seq_print_text(struct seq_file *seq, void *offset)
|
|
{
|
|
struct net_device *net_dev = (struct net_device *)seq->private;
|
|
struct bat_priv *bat_priv = netdev_priv(net_dev);
|
|
struct hashtable_t *hash = bat_priv->hna_local_hash;
|
|
struct hna_local_entry *hna_local_entry;
|
|
struct hlist_node *node;
|
|
struct hlist_head *head;
|
|
size_t buf_size, pos;
|
|
char *buff;
|
|
int i;
|
|
|
|
if (!bat_priv->primary_if) {
|
|
return seq_printf(seq, "BATMAN mesh %s disabled - "
|
|
"please specify interfaces to enable it\n",
|
|
net_dev->name);
|
|
}
|
|
|
|
seq_printf(seq, "Locally retrieved addresses (from %s) "
|
|
"announced via HNA:\n",
|
|
net_dev->name);
|
|
|
|
spin_lock_bh(&bat_priv->hna_lhash_lock);
|
|
|
|
buf_size = 1;
|
|
/* Estimate length for: " * xx:xx:xx:xx:xx:xx\n" */
|
|
for (i = 0; i < hash->size; i++) {
|
|
head = &hash->table[i];
|
|
|
|
rcu_read_lock();
|
|
__hlist_for_each_rcu(node, head)
|
|
buf_size += 21;
|
|
rcu_read_unlock();
|
|
}
|
|
|
|
buff = kmalloc(buf_size, GFP_ATOMIC);
|
|
if (!buff) {
|
|
spin_unlock_bh(&bat_priv->hna_lhash_lock);
|
|
return -ENOMEM;
|
|
}
|
|
|
|
buff[0] = '\0';
|
|
pos = 0;
|
|
|
|
for (i = 0; i < hash->size; i++) {
|
|
head = &hash->table[i];
|
|
|
|
rcu_read_lock();
|
|
hlist_for_each_entry_rcu(hna_local_entry, node,
|
|
head, hash_entry) {
|
|
pos += snprintf(buff + pos, 22, " * %pM\n",
|
|
hna_local_entry->addr);
|
|
}
|
|
rcu_read_unlock();
|
|
}
|
|
|
|
spin_unlock_bh(&bat_priv->hna_lhash_lock);
|
|
|
|
seq_printf(seq, "%s", buff);
|
|
kfree(buff);
|
|
return 0;
|
|
}
|
|
|
|
static void _hna_local_del(struct hlist_node *node, void *arg)
|
|
{
|
|
struct bat_priv *bat_priv = (struct bat_priv *)arg;
|
|
void *data = container_of(node, struct hna_local_entry, hash_entry);
|
|
|
|
kfree(data);
|
|
bat_priv->num_local_hna--;
|
|
atomic_set(&bat_priv->hna_local_changed, 1);
|
|
}
|
|
|
|
static void hna_local_del(struct bat_priv *bat_priv,
|
|
struct hna_local_entry *hna_local_entry,
|
|
char *message)
|
|
{
|
|
bat_dbg(DBG_ROUTES, bat_priv, "Deleting local hna entry (%pM): %s\n",
|
|
hna_local_entry->addr, message);
|
|
|
|
hash_remove(bat_priv->hna_local_hash, compare_lhna, choose_orig,
|
|
hna_local_entry->addr);
|
|
_hna_local_del(&hna_local_entry->hash_entry, bat_priv);
|
|
}
|
|
|
|
void hna_local_remove(struct bat_priv *bat_priv,
|
|
uint8_t *addr, char *message)
|
|
{
|
|
struct hna_local_entry *hna_local_entry;
|
|
|
|
spin_lock_bh(&bat_priv->hna_lhash_lock);
|
|
|
|
hna_local_entry = hna_local_hash_find(bat_priv, addr);
|
|
|
|
if (hna_local_entry)
|
|
hna_local_del(bat_priv, hna_local_entry, message);
|
|
|
|
spin_unlock_bh(&bat_priv->hna_lhash_lock);
|
|
}
|
|
|
|
static void hna_local_purge(struct work_struct *work)
|
|
{
|
|
struct delayed_work *delayed_work =
|
|
container_of(work, struct delayed_work, work);
|
|
struct bat_priv *bat_priv =
|
|
container_of(delayed_work, struct bat_priv, hna_work);
|
|
struct hashtable_t *hash = bat_priv->hna_local_hash;
|
|
struct hna_local_entry *hna_local_entry;
|
|
struct hlist_node *node, *node_tmp;
|
|
struct hlist_head *head;
|
|
unsigned long timeout;
|
|
int i;
|
|
|
|
spin_lock_bh(&bat_priv->hna_lhash_lock);
|
|
|
|
for (i = 0; i < hash->size; i++) {
|
|
head = &hash->table[i];
|
|
|
|
hlist_for_each_entry_safe(hna_local_entry, node, node_tmp,
|
|
head, hash_entry) {
|
|
if (hna_local_entry->never_purge)
|
|
continue;
|
|
|
|
timeout = hna_local_entry->last_seen;
|
|
timeout += LOCAL_HNA_TIMEOUT * HZ;
|
|
|
|
if (time_before(jiffies, timeout))
|
|
continue;
|
|
|
|
hna_local_del(bat_priv, hna_local_entry,
|
|
"address timed out");
|
|
}
|
|
}
|
|
|
|
spin_unlock_bh(&bat_priv->hna_lhash_lock);
|
|
hna_local_start_timer(bat_priv);
|
|
}
|
|
|
|
void hna_local_free(struct bat_priv *bat_priv)
|
|
{
|
|
if (!bat_priv->hna_local_hash)
|
|
return;
|
|
|
|
cancel_delayed_work_sync(&bat_priv->hna_work);
|
|
hash_delete(bat_priv->hna_local_hash, _hna_local_del, bat_priv);
|
|
bat_priv->hna_local_hash = NULL;
|
|
}
|
|
|
|
int hna_global_init(struct bat_priv *bat_priv)
|
|
{
|
|
if (bat_priv->hna_global_hash)
|
|
return 1;
|
|
|
|
bat_priv->hna_global_hash = hash_new(1024);
|
|
|
|
if (!bat_priv->hna_global_hash)
|
|
return 0;
|
|
|
|
return 1;
|
|
}
|
|
|
|
void hna_global_add_orig(struct bat_priv *bat_priv,
|
|
struct orig_node *orig_node,
|
|
unsigned char *hna_buff, int hna_buff_len)
|
|
{
|
|
struct hna_global_entry *hna_global_entry;
|
|
struct hna_local_entry *hna_local_entry;
|
|
int hna_buff_count = 0;
|
|
unsigned char *hna_ptr;
|
|
|
|
while ((hna_buff_count + 1) * ETH_ALEN <= hna_buff_len) {
|
|
spin_lock_bh(&bat_priv->hna_ghash_lock);
|
|
|
|
hna_ptr = hna_buff + (hna_buff_count * ETH_ALEN);
|
|
hna_global_entry = hna_global_hash_find(bat_priv, hna_ptr);
|
|
|
|
if (!hna_global_entry) {
|
|
spin_unlock_bh(&bat_priv->hna_ghash_lock);
|
|
|
|
hna_global_entry =
|
|
kmalloc(sizeof(struct hna_global_entry),
|
|
GFP_ATOMIC);
|
|
|
|
if (!hna_global_entry)
|
|
break;
|
|
|
|
memcpy(hna_global_entry->addr, hna_ptr, ETH_ALEN);
|
|
|
|
bat_dbg(DBG_ROUTES, bat_priv,
|
|
"Creating new global hna entry: "
|
|
"%pM (via %pM)\n",
|
|
hna_global_entry->addr, orig_node->orig);
|
|
|
|
spin_lock_bh(&bat_priv->hna_ghash_lock);
|
|
hash_add(bat_priv->hna_global_hash, compare_ghna,
|
|
choose_orig, hna_global_entry,
|
|
&hna_global_entry->hash_entry);
|
|
|
|
}
|
|
|
|
hna_global_entry->orig_node = orig_node;
|
|
spin_unlock_bh(&bat_priv->hna_ghash_lock);
|
|
|
|
/* remove address from local hash if present */
|
|
spin_lock_bh(&bat_priv->hna_lhash_lock);
|
|
|
|
hna_ptr = hna_buff + (hna_buff_count * ETH_ALEN);
|
|
hna_local_entry = hna_local_hash_find(bat_priv, hna_ptr);
|
|
|
|
if (hna_local_entry)
|
|
hna_local_del(bat_priv, hna_local_entry,
|
|
"global hna received");
|
|
|
|
spin_unlock_bh(&bat_priv->hna_lhash_lock);
|
|
|
|
hna_buff_count++;
|
|
}
|
|
|
|
/* initialize, and overwrite if malloc succeeds */
|
|
orig_node->hna_buff = NULL;
|
|
orig_node->hna_buff_len = 0;
|
|
|
|
if (hna_buff_len > 0) {
|
|
orig_node->hna_buff = kmalloc(hna_buff_len, GFP_ATOMIC);
|
|
if (orig_node->hna_buff) {
|
|
memcpy(orig_node->hna_buff, hna_buff, hna_buff_len);
|
|
orig_node->hna_buff_len = hna_buff_len;
|
|
}
|
|
}
|
|
}
|
|
|
|
int hna_global_seq_print_text(struct seq_file *seq, void *offset)
|
|
{
|
|
struct net_device *net_dev = (struct net_device *)seq->private;
|
|
struct bat_priv *bat_priv = netdev_priv(net_dev);
|
|
struct hashtable_t *hash = bat_priv->hna_global_hash;
|
|
struct hna_global_entry *hna_global_entry;
|
|
struct hlist_node *node;
|
|
struct hlist_head *head;
|
|
size_t buf_size, pos;
|
|
char *buff;
|
|
int i;
|
|
|
|
if (!bat_priv->primary_if) {
|
|
return seq_printf(seq, "BATMAN mesh %s disabled - "
|
|
"please specify interfaces to enable it\n",
|
|
net_dev->name);
|
|
}
|
|
|
|
seq_printf(seq, "Globally announced HNAs received via the mesh %s\n",
|
|
net_dev->name);
|
|
|
|
spin_lock_bh(&bat_priv->hna_ghash_lock);
|
|
|
|
buf_size = 1;
|
|
/* Estimate length for: " * xx:xx:xx:xx:xx:xx via xx:xx:xx:xx:xx:xx\n"*/
|
|
for (i = 0; i < hash->size; i++) {
|
|
head = &hash->table[i];
|
|
|
|
rcu_read_lock();
|
|
__hlist_for_each_rcu(node, head)
|
|
buf_size += 43;
|
|
rcu_read_unlock();
|
|
}
|
|
|
|
buff = kmalloc(buf_size, GFP_ATOMIC);
|
|
if (!buff) {
|
|
spin_unlock_bh(&bat_priv->hna_ghash_lock);
|
|
return -ENOMEM;
|
|
}
|
|
buff[0] = '\0';
|
|
pos = 0;
|
|
|
|
for (i = 0; i < hash->size; i++) {
|
|
head = &hash->table[i];
|
|
|
|
rcu_read_lock();
|
|
hlist_for_each_entry_rcu(hna_global_entry, node,
|
|
head, hash_entry) {
|
|
pos += snprintf(buff + pos, 44,
|
|
" * %pM via %pM\n",
|
|
hna_global_entry->addr,
|
|
hna_global_entry->orig_node->orig);
|
|
}
|
|
rcu_read_unlock();
|
|
}
|
|
|
|
spin_unlock_bh(&bat_priv->hna_ghash_lock);
|
|
|
|
seq_printf(seq, "%s", buff);
|
|
kfree(buff);
|
|
return 0;
|
|
}
|
|
|
|
static void _hna_global_del_orig(struct bat_priv *bat_priv,
|
|
struct hna_global_entry *hna_global_entry,
|
|
char *message)
|
|
{
|
|
bat_dbg(DBG_ROUTES, bat_priv,
|
|
"Deleting global hna entry %pM (via %pM): %s\n",
|
|
hna_global_entry->addr, hna_global_entry->orig_node->orig,
|
|
message);
|
|
|
|
hash_remove(bat_priv->hna_global_hash, compare_ghna, choose_orig,
|
|
hna_global_entry->addr);
|
|
kfree(hna_global_entry);
|
|
}
|
|
|
|
void hna_global_del_orig(struct bat_priv *bat_priv,
|
|
struct orig_node *orig_node, char *message)
|
|
{
|
|
struct hna_global_entry *hna_global_entry;
|
|
int hna_buff_count = 0;
|
|
unsigned char *hna_ptr;
|
|
|
|
if (orig_node->hna_buff_len == 0)
|
|
return;
|
|
|
|
spin_lock_bh(&bat_priv->hna_ghash_lock);
|
|
|
|
while ((hna_buff_count + 1) * ETH_ALEN <= orig_node->hna_buff_len) {
|
|
hna_ptr = orig_node->hna_buff + (hna_buff_count * ETH_ALEN);
|
|
hna_global_entry = hna_global_hash_find(bat_priv, hna_ptr);
|
|
|
|
if ((hna_global_entry) &&
|
|
(hna_global_entry->orig_node == orig_node))
|
|
_hna_global_del_orig(bat_priv, hna_global_entry,
|
|
message);
|
|
|
|
hna_buff_count++;
|
|
}
|
|
|
|
spin_unlock_bh(&bat_priv->hna_ghash_lock);
|
|
|
|
orig_node->hna_buff_len = 0;
|
|
kfree(orig_node->hna_buff);
|
|
orig_node->hna_buff = NULL;
|
|
}
|
|
|
|
static void hna_global_del(struct hlist_node *node, void *arg)
|
|
{
|
|
void *data = container_of(node, struct hna_global_entry, hash_entry);
|
|
|
|
kfree(data);
|
|
}
|
|
|
|
void hna_global_free(struct bat_priv *bat_priv)
|
|
{
|
|
if (!bat_priv->hna_global_hash)
|
|
return;
|
|
|
|
hash_delete(bat_priv->hna_global_hash, hna_global_del, NULL);
|
|
bat_priv->hna_global_hash = NULL;
|
|
}
|
|
|
|
struct orig_node *transtable_search(struct bat_priv *bat_priv, uint8_t *addr)
|
|
{
|
|
struct hna_global_entry *hna_global_entry;
|
|
struct orig_node *orig_node = NULL;
|
|
|
|
spin_lock_bh(&bat_priv->hna_ghash_lock);
|
|
hna_global_entry = hna_global_hash_find(bat_priv, addr);
|
|
|
|
if (!hna_global_entry)
|
|
goto out;
|
|
|
|
if (!atomic_inc_not_zero(&hna_global_entry->orig_node->refcount))
|
|
goto out;
|
|
|
|
orig_node = hna_global_entry->orig_node;
|
|
|
|
out:
|
|
spin_unlock_bh(&bat_priv->hna_ghash_lock);
|
|
return orig_node;
|
|
}
|