forked from luck/tmp_suning_uos_patched
593a5f22d8
Updated copyright notice to include the year the file was actually created. Information about file creation dates was extracted from the files in the old CVS repository at tipc.sourceforge.net. Signed-off-by: Per Liden <per.liden@nospam.ericsson.com>
577 lines
14 KiB
C
577 lines
14 KiB
C
/*
|
|
* net/tipc/cluster.c: TIPC cluster management routines
|
|
*
|
|
* Copyright (c) 2000-2006, Ericsson AB
|
|
* Copyright (c) 2005, Wind River Systems
|
|
* All rights reserved.
|
|
*
|
|
* Redistribution and use in source and binary forms, with or without
|
|
* modification, are permitted provided that the following conditions are met:
|
|
*
|
|
* 1. Redistributions of source code must retain the above copyright
|
|
* notice, this list of conditions and the following disclaimer.
|
|
* 2. Redistributions in binary form must reproduce the above copyright
|
|
* notice, this list of conditions and the following disclaimer in the
|
|
* documentation and/or other materials provided with the distribution.
|
|
* 3. Neither the names of the copyright holders nor the names of its
|
|
* contributors may be used to endorse or promote products derived from
|
|
* this software without specific prior written permission.
|
|
*
|
|
* Alternatively, this software may be distributed under the terms of the
|
|
* GNU General Public License ("GPL") version 2 as published by the Free
|
|
* Software Foundation.
|
|
*
|
|
* THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS "AS IS"
|
|
* AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE
|
|
* IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE
|
|
* ARE DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT OWNER OR CONTRIBUTORS BE
|
|
* LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR
|
|
* CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF
|
|
* SUBSTITUTE GOODS OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS
|
|
* INTERRUPTION) HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN
|
|
* CONTRACT, STRICT LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE)
|
|
* ARISING IN ANY WAY OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE
|
|
* POSSIBILITY OF SUCH DAMAGE.
|
|
*/
|
|
|
|
#include "core.h"
|
|
#include "cluster.h"
|
|
#include "addr.h"
|
|
#include "node_subscr.h"
|
|
#include "link.h"
|
|
#include "node.h"
|
|
#include "net.h"
|
|
#include "msg.h"
|
|
#include "bearer.h"
|
|
|
|
void cluster_multicast(struct cluster *c_ptr, struct sk_buff *buf,
|
|
u32 lower, u32 upper);
|
|
struct sk_buff *cluster_prepare_routing_msg(u32 data_size, u32 dest);
|
|
|
|
struct node **local_nodes = 0;
|
|
struct node_map cluster_bcast_nodes = {0,{0,}};
|
|
u32 highest_allowed_slave = 0;
|
|
|
|
struct cluster *cluster_create(u32 addr)
|
|
{
|
|
struct _zone *z_ptr;
|
|
struct cluster *c_ptr;
|
|
int max_nodes;
|
|
int alloc;
|
|
|
|
c_ptr = (struct cluster *)kmalloc(sizeof(*c_ptr), GFP_ATOMIC);
|
|
if (c_ptr == NULL)
|
|
return 0;
|
|
memset(c_ptr, 0, sizeof(*c_ptr));
|
|
|
|
c_ptr->addr = tipc_addr(tipc_zone(addr), tipc_cluster(addr), 0);
|
|
if (in_own_cluster(addr))
|
|
max_nodes = LOWEST_SLAVE + tipc_max_slaves;
|
|
else
|
|
max_nodes = tipc_max_nodes + 1;
|
|
alloc = sizeof(void *) * (max_nodes + 1);
|
|
c_ptr->nodes = (struct node **)kmalloc(alloc, GFP_ATOMIC);
|
|
if (c_ptr->nodes == NULL) {
|
|
kfree(c_ptr);
|
|
return 0;
|
|
}
|
|
memset(c_ptr->nodes, 0, alloc);
|
|
if (in_own_cluster(addr))
|
|
local_nodes = c_ptr->nodes;
|
|
c_ptr->highest_slave = LOWEST_SLAVE - 1;
|
|
c_ptr->highest_node = 0;
|
|
|
|
z_ptr = zone_find(tipc_zone(addr));
|
|
if (z_ptr == NULL) {
|
|
z_ptr = zone_create(addr);
|
|
}
|
|
if (z_ptr != NULL) {
|
|
zone_attach_cluster(z_ptr, c_ptr);
|
|
c_ptr->owner = z_ptr;
|
|
}
|
|
else {
|
|
kfree(c_ptr);
|
|
c_ptr = 0;
|
|
}
|
|
|
|
return c_ptr;
|
|
}
|
|
|
|
void cluster_delete(struct cluster *c_ptr)
|
|
{
|
|
u32 n_num;
|
|
|
|
if (!c_ptr)
|
|
return;
|
|
for (n_num = 1; n_num <= c_ptr->highest_node; n_num++) {
|
|
node_delete(c_ptr->nodes[n_num]);
|
|
}
|
|
for (n_num = LOWEST_SLAVE; n_num <= c_ptr->highest_slave; n_num++) {
|
|
node_delete(c_ptr->nodes[n_num]);
|
|
}
|
|
kfree(c_ptr->nodes);
|
|
kfree(c_ptr);
|
|
}
|
|
|
|
u32 cluster_next_node(struct cluster *c_ptr, u32 addr)
|
|
{
|
|
struct node *n_ptr;
|
|
u32 n_num = tipc_node(addr) + 1;
|
|
|
|
if (!c_ptr)
|
|
return addr;
|
|
for (; n_num <= c_ptr->highest_node; n_num++) {
|
|
n_ptr = c_ptr->nodes[n_num];
|
|
if (n_ptr && node_has_active_links(n_ptr))
|
|
return n_ptr->addr;
|
|
}
|
|
for (n_num = 1; n_num < tipc_node(addr); n_num++) {
|
|
n_ptr = c_ptr->nodes[n_num];
|
|
if (n_ptr && node_has_active_links(n_ptr))
|
|
return n_ptr->addr;
|
|
}
|
|
return 0;
|
|
}
|
|
|
|
void cluster_attach_node(struct cluster *c_ptr, struct node *n_ptr)
|
|
{
|
|
u32 n_num = tipc_node(n_ptr->addr);
|
|
u32 max_n_num = tipc_max_nodes;
|
|
|
|
if (in_own_cluster(n_ptr->addr))
|
|
max_n_num = highest_allowed_slave;
|
|
assert(n_num > 0);
|
|
assert(n_num <= max_n_num);
|
|
assert(c_ptr->nodes[n_num] == 0);
|
|
c_ptr->nodes[n_num] = n_ptr;
|
|
if (n_num > c_ptr->highest_node)
|
|
c_ptr->highest_node = n_num;
|
|
}
|
|
|
|
/**
|
|
* cluster_select_router - select router to a cluster
|
|
*
|
|
* Uses deterministic and fair algorithm.
|
|
*/
|
|
|
|
u32 cluster_select_router(struct cluster *c_ptr, u32 ref)
|
|
{
|
|
u32 n_num;
|
|
u32 ulim = c_ptr->highest_node;
|
|
u32 mask;
|
|
u32 tstart;
|
|
|
|
assert(!in_own_cluster(c_ptr->addr));
|
|
if (!ulim)
|
|
return 0;
|
|
|
|
/* Start entry must be random */
|
|
mask = tipc_max_nodes;
|
|
while (mask > ulim)
|
|
mask >>= 1;
|
|
tstart = ref & mask;
|
|
n_num = tstart;
|
|
|
|
/* Lookup upwards with wrap-around */
|
|
do {
|
|
if (node_is_up(c_ptr->nodes[n_num]))
|
|
break;
|
|
} while (++n_num <= ulim);
|
|
if (n_num > ulim) {
|
|
n_num = 1;
|
|
do {
|
|
if (node_is_up(c_ptr->nodes[n_num]))
|
|
break;
|
|
} while (++n_num < tstart);
|
|
if (n_num == tstart)
|
|
return 0;
|
|
}
|
|
assert(n_num <= ulim);
|
|
return node_select_router(c_ptr->nodes[n_num], ref);
|
|
}
|
|
|
|
/**
|
|
* cluster_select_node - select destination node within a remote cluster
|
|
*
|
|
* Uses deterministic and fair algorithm.
|
|
*/
|
|
|
|
struct node *cluster_select_node(struct cluster *c_ptr, u32 selector)
|
|
{
|
|
u32 n_num;
|
|
u32 mask = tipc_max_nodes;
|
|
u32 start_entry;
|
|
|
|
assert(!in_own_cluster(c_ptr->addr));
|
|
if (!c_ptr->highest_node)
|
|
return 0;
|
|
|
|
/* Start entry must be random */
|
|
while (mask > c_ptr->highest_node) {
|
|
mask >>= 1;
|
|
}
|
|
start_entry = (selector & mask) ? selector & mask : 1u;
|
|
assert(start_entry <= c_ptr->highest_node);
|
|
|
|
/* Lookup upwards with wrap-around */
|
|
for (n_num = start_entry; n_num <= c_ptr->highest_node; n_num++) {
|
|
if (node_has_active_links(c_ptr->nodes[n_num]))
|
|
return c_ptr->nodes[n_num];
|
|
}
|
|
for (n_num = 1; n_num < start_entry; n_num++) {
|
|
if (node_has_active_links(c_ptr->nodes[n_num]))
|
|
return c_ptr->nodes[n_num];
|
|
}
|
|
return 0;
|
|
}
|
|
|
|
/*
|
|
* Routing table management: See description in node.c
|
|
*/
|
|
|
|
struct sk_buff *cluster_prepare_routing_msg(u32 data_size, u32 dest)
|
|
{
|
|
u32 size = INT_H_SIZE + data_size;
|
|
struct sk_buff *buf = buf_acquire(size);
|
|
struct tipc_msg *msg;
|
|
|
|
if (buf) {
|
|
msg = buf_msg(buf);
|
|
memset((char *)msg, 0, size);
|
|
msg_init(msg, ROUTE_DISTRIBUTOR, 0, TIPC_OK, INT_H_SIZE, dest);
|
|
}
|
|
return buf;
|
|
}
|
|
|
|
void cluster_bcast_new_route(struct cluster *c_ptr, u32 dest,
|
|
u32 lower, u32 upper)
|
|
{
|
|
struct sk_buff *buf = cluster_prepare_routing_msg(0, c_ptr->addr);
|
|
struct tipc_msg *msg;
|
|
|
|
if (buf) {
|
|
msg = buf_msg(buf);
|
|
msg_set_remote_node(msg, dest);
|
|
msg_set_type(msg, ROUTE_ADDITION);
|
|
cluster_multicast(c_ptr, buf, lower, upper);
|
|
} else {
|
|
warn("Memory squeeze: broadcast of new route failed\n");
|
|
}
|
|
}
|
|
|
|
void cluster_bcast_lost_route(struct cluster *c_ptr, u32 dest,
|
|
u32 lower, u32 upper)
|
|
{
|
|
struct sk_buff *buf = cluster_prepare_routing_msg(0, c_ptr->addr);
|
|
struct tipc_msg *msg;
|
|
|
|
if (buf) {
|
|
msg = buf_msg(buf);
|
|
msg_set_remote_node(msg, dest);
|
|
msg_set_type(msg, ROUTE_REMOVAL);
|
|
cluster_multicast(c_ptr, buf, lower, upper);
|
|
} else {
|
|
warn("Memory squeeze: broadcast of lost route failed\n");
|
|
}
|
|
}
|
|
|
|
void cluster_send_slave_routes(struct cluster *c_ptr, u32 dest)
|
|
{
|
|
struct sk_buff *buf;
|
|
struct tipc_msg *msg;
|
|
u32 highest = c_ptr->highest_slave;
|
|
u32 n_num;
|
|
int send = 0;
|
|
|
|
assert(!is_slave(dest));
|
|
assert(in_own_cluster(dest));
|
|
assert(in_own_cluster(c_ptr->addr));
|
|
if (highest <= LOWEST_SLAVE)
|
|
return;
|
|
buf = cluster_prepare_routing_msg(highest - LOWEST_SLAVE + 1,
|
|
c_ptr->addr);
|
|
if (buf) {
|
|
msg = buf_msg(buf);
|
|
msg_set_remote_node(msg, c_ptr->addr);
|
|
msg_set_type(msg, SLAVE_ROUTING_TABLE);
|
|
for (n_num = LOWEST_SLAVE; n_num <= highest; n_num++) {
|
|
if (c_ptr->nodes[n_num] &&
|
|
node_has_active_links(c_ptr->nodes[n_num])) {
|
|
send = 1;
|
|
msg_set_dataoctet(msg, n_num);
|
|
}
|
|
}
|
|
if (send)
|
|
link_send(buf, dest, dest);
|
|
else
|
|
buf_discard(buf);
|
|
} else {
|
|
warn("Memory squeeze: broadcast of lost route failed\n");
|
|
}
|
|
}
|
|
|
|
void cluster_send_ext_routes(struct cluster *c_ptr, u32 dest)
|
|
{
|
|
struct sk_buff *buf;
|
|
struct tipc_msg *msg;
|
|
u32 highest = c_ptr->highest_node;
|
|
u32 n_num;
|
|
int send = 0;
|
|
|
|
if (in_own_cluster(c_ptr->addr))
|
|
return;
|
|
assert(!is_slave(dest));
|
|
assert(in_own_cluster(dest));
|
|
highest = c_ptr->highest_node;
|
|
buf = cluster_prepare_routing_msg(highest + 1, c_ptr->addr);
|
|
if (buf) {
|
|
msg = buf_msg(buf);
|
|
msg_set_remote_node(msg, c_ptr->addr);
|
|
msg_set_type(msg, EXT_ROUTING_TABLE);
|
|
for (n_num = 1; n_num <= highest; n_num++) {
|
|
if (c_ptr->nodes[n_num] &&
|
|
node_has_active_links(c_ptr->nodes[n_num])) {
|
|
send = 1;
|
|
msg_set_dataoctet(msg, n_num);
|
|
}
|
|
}
|
|
if (send)
|
|
link_send(buf, dest, dest);
|
|
else
|
|
buf_discard(buf);
|
|
} else {
|
|
warn("Memory squeeze: broadcast of external route failed\n");
|
|
}
|
|
}
|
|
|
|
void cluster_send_local_routes(struct cluster *c_ptr, u32 dest)
|
|
{
|
|
struct sk_buff *buf;
|
|
struct tipc_msg *msg;
|
|
u32 highest = c_ptr->highest_node;
|
|
u32 n_num;
|
|
int send = 0;
|
|
|
|
assert(is_slave(dest));
|
|
assert(in_own_cluster(c_ptr->addr));
|
|
buf = cluster_prepare_routing_msg(highest, c_ptr->addr);
|
|
if (buf) {
|
|
msg = buf_msg(buf);
|
|
msg_set_remote_node(msg, c_ptr->addr);
|
|
msg_set_type(msg, LOCAL_ROUTING_TABLE);
|
|
for (n_num = 1; n_num <= highest; n_num++) {
|
|
if (c_ptr->nodes[n_num] &&
|
|
node_has_active_links(c_ptr->nodes[n_num])) {
|
|
send = 1;
|
|
msg_set_dataoctet(msg, n_num);
|
|
}
|
|
}
|
|
if (send)
|
|
link_send(buf, dest, dest);
|
|
else
|
|
buf_discard(buf);
|
|
} else {
|
|
warn("Memory squeeze: broadcast of local route failed\n");
|
|
}
|
|
}
|
|
|
|
void cluster_recv_routing_table(struct sk_buff *buf)
|
|
{
|
|
struct tipc_msg *msg = buf_msg(buf);
|
|
struct cluster *c_ptr;
|
|
struct node *n_ptr;
|
|
unchar *node_table;
|
|
u32 table_size;
|
|
u32 router;
|
|
u32 rem_node = msg_remote_node(msg);
|
|
u32 z_num;
|
|
u32 c_num;
|
|
u32 n_num;
|
|
|
|
c_ptr = cluster_find(rem_node);
|
|
if (!c_ptr) {
|
|
c_ptr = cluster_create(rem_node);
|
|
if (!c_ptr) {
|
|
buf_discard(buf);
|
|
return;
|
|
}
|
|
}
|
|
|
|
node_table = buf->data + msg_hdr_sz(msg);
|
|
table_size = msg_size(msg) - msg_hdr_sz(msg);
|
|
router = msg_prevnode(msg);
|
|
z_num = tipc_zone(rem_node);
|
|
c_num = tipc_cluster(rem_node);
|
|
|
|
switch (msg_type(msg)) {
|
|
case LOCAL_ROUTING_TABLE:
|
|
assert(is_slave(tipc_own_addr));
|
|
case EXT_ROUTING_TABLE:
|
|
for (n_num = 1; n_num < table_size; n_num++) {
|
|
if (node_table[n_num]) {
|
|
u32 addr = tipc_addr(z_num, c_num, n_num);
|
|
n_ptr = c_ptr->nodes[n_num];
|
|
if (!n_ptr) {
|
|
n_ptr = node_create(addr);
|
|
}
|
|
if (n_ptr)
|
|
node_add_router(n_ptr, router);
|
|
}
|
|
}
|
|
break;
|
|
case SLAVE_ROUTING_TABLE:
|
|
assert(!is_slave(tipc_own_addr));
|
|
assert(in_own_cluster(c_ptr->addr));
|
|
for (n_num = 1; n_num < table_size; n_num++) {
|
|
if (node_table[n_num]) {
|
|
u32 slave_num = n_num + LOWEST_SLAVE;
|
|
u32 addr = tipc_addr(z_num, c_num, slave_num);
|
|
n_ptr = c_ptr->nodes[slave_num];
|
|
if (!n_ptr) {
|
|
n_ptr = node_create(addr);
|
|
}
|
|
if (n_ptr)
|
|
node_add_router(n_ptr, router);
|
|
}
|
|
}
|
|
break;
|
|
case ROUTE_ADDITION:
|
|
if (!is_slave(tipc_own_addr)) {
|
|
assert(!in_own_cluster(c_ptr->addr)
|
|
|| is_slave(rem_node));
|
|
} else {
|
|
assert(in_own_cluster(c_ptr->addr)
|
|
&& !is_slave(rem_node));
|
|
}
|
|
n_ptr = c_ptr->nodes[tipc_node(rem_node)];
|
|
if (!n_ptr)
|
|
n_ptr = node_create(rem_node);
|
|
if (n_ptr)
|
|
node_add_router(n_ptr, router);
|
|
break;
|
|
case ROUTE_REMOVAL:
|
|
if (!is_slave(tipc_own_addr)) {
|
|
assert(!in_own_cluster(c_ptr->addr)
|
|
|| is_slave(rem_node));
|
|
} else {
|
|
assert(in_own_cluster(c_ptr->addr)
|
|
&& !is_slave(rem_node));
|
|
}
|
|
n_ptr = c_ptr->nodes[tipc_node(rem_node)];
|
|
if (n_ptr)
|
|
node_remove_router(n_ptr, router);
|
|
break;
|
|
default:
|
|
assert(!"Illegal routing manager message received\n");
|
|
}
|
|
buf_discard(buf);
|
|
}
|
|
|
|
void cluster_remove_as_router(struct cluster *c_ptr, u32 router)
|
|
{
|
|
u32 start_entry;
|
|
u32 tstop;
|
|
u32 n_num;
|
|
|
|
if (is_slave(router))
|
|
return; /* Slave nodes can not be routers */
|
|
|
|
if (in_own_cluster(c_ptr->addr)) {
|
|
start_entry = LOWEST_SLAVE;
|
|
tstop = c_ptr->highest_slave;
|
|
} else {
|
|
start_entry = 1;
|
|
tstop = c_ptr->highest_node;
|
|
}
|
|
|
|
for (n_num = start_entry; n_num <= tstop; n_num++) {
|
|
if (c_ptr->nodes[n_num]) {
|
|
node_remove_router(c_ptr->nodes[n_num], router);
|
|
}
|
|
}
|
|
}
|
|
|
|
/**
|
|
* cluster_multicast - multicast message to local nodes
|
|
*/
|
|
|
|
void cluster_multicast(struct cluster *c_ptr, struct sk_buff *buf,
|
|
u32 lower, u32 upper)
|
|
{
|
|
struct sk_buff *buf_copy;
|
|
struct node *n_ptr;
|
|
u32 n_num;
|
|
u32 tstop;
|
|
|
|
assert(lower <= upper);
|
|
assert(((lower >= 1) && (lower <= tipc_max_nodes)) ||
|
|
((lower >= LOWEST_SLAVE) && (lower <= highest_allowed_slave)));
|
|
assert(((upper >= 1) && (upper <= tipc_max_nodes)) ||
|
|
((upper >= LOWEST_SLAVE) && (upper <= highest_allowed_slave)));
|
|
assert(in_own_cluster(c_ptr->addr));
|
|
|
|
tstop = is_slave(upper) ? c_ptr->highest_slave : c_ptr->highest_node;
|
|
if (tstop > upper)
|
|
tstop = upper;
|
|
for (n_num = lower; n_num <= tstop; n_num++) {
|
|
n_ptr = c_ptr->nodes[n_num];
|
|
if (n_ptr && node_has_active_links(n_ptr)) {
|
|
buf_copy = skb_copy(buf, GFP_ATOMIC);
|
|
if (buf_copy == NULL)
|
|
break;
|
|
msg_set_destnode(buf_msg(buf_copy), n_ptr->addr);
|
|
link_send(buf_copy, n_ptr->addr, n_ptr->addr);
|
|
}
|
|
}
|
|
buf_discard(buf);
|
|
}
|
|
|
|
/**
|
|
* cluster_broadcast - broadcast message to all nodes within cluster
|
|
*/
|
|
|
|
void cluster_broadcast(struct sk_buff *buf)
|
|
{
|
|
struct sk_buff *buf_copy;
|
|
struct cluster *c_ptr;
|
|
struct node *n_ptr;
|
|
u32 n_num;
|
|
u32 tstart;
|
|
u32 tstop;
|
|
u32 node_type;
|
|
|
|
if (tipc_mode == TIPC_NET_MODE) {
|
|
c_ptr = cluster_find(tipc_own_addr);
|
|
assert(in_own_cluster(c_ptr->addr)); /* For now */
|
|
|
|
/* Send to standard nodes, then repeat loop sending to slaves */
|
|
tstart = 1;
|
|
tstop = c_ptr->highest_node;
|
|
for (node_type = 1; node_type <= 2; node_type++) {
|
|
for (n_num = tstart; n_num <= tstop; n_num++) {
|
|
n_ptr = c_ptr->nodes[n_num];
|
|
if (n_ptr && node_has_active_links(n_ptr)) {
|
|
buf_copy = skb_copy(buf, GFP_ATOMIC);
|
|
if (buf_copy == NULL)
|
|
goto exit;
|
|
msg_set_destnode(buf_msg(buf_copy),
|
|
n_ptr->addr);
|
|
link_send(buf_copy, n_ptr->addr,
|
|
n_ptr->addr);
|
|
}
|
|
}
|
|
tstart = LOWEST_SLAVE;
|
|
tstop = c_ptr->highest_slave;
|
|
}
|
|
}
|
|
exit:
|
|
buf_discard(buf);
|
|
}
|
|
|
|
int cluster_init(void)
|
|
{
|
|
highest_allowed_slave = LOWEST_SLAVE + tipc_max_slaves;
|
|
return cluster_create(tipc_own_addr) ? TIPC_OK : -ENOMEM;
|
|
}
|
|
|