Reliable communication but unreliable networks
17 October 2013
Now that the internet’s basic protocols are more than 30 years old, network scientists are increasingly turning their attention to ad-hoc networks — communications networks set up, on the fly, by wireless devices — where unsolved problems still abound.
Most theoretical analyses of ad-hoc networks have assumed that the communications links within the network are stable. But that often is not the case with real-world wireless devices, as anyone who has used a mobile phone knows.
Researchers from the Theory of Distributed Systems Group at Massachusetts Institute of Technology’s (MIT’s) Computer Science and Artificial Intelligence Laboratory have presented a new framework for analysing ad-hoc networks in which the quality of the communications links fluctuates. Within that framework, they provide mathematical bounds on the efficiency with which messages can propagate through the network, and they describe new algorithms that can achieve maximal efficiency.
“There’s been a discrepancy between the theory, with its idealised models, and the reality of wireless networks,” said Nancy Lynch, the NEC professor of software science and engineering at MIT and head of the Theory of Distributed Systems Group. “When people start designing theoretical algorithms, they tend to rely too heavily on the specific assumptions of the models. So the algorithms tend to be unrealistic and fragile.”
In the past, some researchers have tried to model the unreliability of network links as random fluctuations. “But if you assume real randomness, then you can count on the randomness,” Lynch said. “Somehow you can use that in your algorithm. Maybe randomness itself is giving you an assumption that’s too strong.”
Lynch and her co-authors on the new paper, Mohsen Ghaffari and Cal Newport, instead modeled the fluctuations in the links’ quality as the willful manipulations of an ‘adversary’. Adversaries cannot control all the links in the network: some will remain up throughout the execution of the communication algorithm. But they can change the bandwidth of the others at will and network designers do not know in advance which links are reliable and which are not.
“Your algorithm needs to work for all possible adversaries, some of which are benign and some of which might be doing the worst possible thing for your algorithm,” Newport said. “In other words, it needs to work for all possible strategies for controlling the network.”
In a paper that appeared two years ago, Newport, Lynch and colleagues assumed a very powerful adversary indeed — one that knew in advance of every decision that every node in the network would make while trying to disseminate a message. In that context, they proved, efficient communication is impossible.
In the new paper, they weakened the adversary significantly. He may know exactly how the communications algorithm works, and he may intentionally try to thwart it, but he has to determine his pattern of link manipulation in advance, before the algorithm begins to run. Even this weakened adversary, however, has the potential to be much more disruptive than the types of interference that real-world wireless networks are likely to encounter — such as doors opening and closing, people turning on microwaves, or rain falling.
Lynch, Newport and Ghaffari examined two types of message dissemination. In the first, a single node of the network is trying to broadcast a message to all other nodes. In that case, they found, efficient communication is possible, even in the adversary’s presence.
The second case is that in which a number of nodes are each transmitting messages, and every one of their immediate neighbors has to receive a message from at least one transmitter. As it turns out, many common problems in the analysis of ad hoc networks boil down to this one.
Here, the researchers found that the adversary’s presence can thwart efficient communication, but only if the network has an odd shape, in which a central node is connected to many nearby nodes that are not connected to each other. That type of network layout is improbable in the real world: if two wireless devices are close enough to a third to communicate with it, they’re likely to be able to communicate with each other, too.
Once the researchers added another assumption — that two devices connected to a third will at least sometimes be able to establish links with each other, too — efficient communication again becomes possible. In both cases, the researchers’ communication algorithms were able to thwart the adversary by using randomness.
One of the problems with designing communications protocols for ad-hoc wireless networks is that if two nearby nodes begin transmitting at the same time at the same frequency, they can interfere with each other, preventing either transmission from being received. The best-performing protocols thus assign each node a probability of transmitting during any one round of communication (where a round is defined by the time it takes for a node to send a message to its immediate neighbours).
The researchers’ algorithms adhere to this basic scheme, but rather than cycling through a prescribed sequence of steadily shrinking probabilities, they scramble the sequence up. In the case of the local broadcast, each separate message has to have its own unique sequence of probabilities. So clusters of nodes also temporarily elect local leaders that co-ordinate the probabilities for different transmitters. The researchers were able to show, however, that this extra computation did not slow communication egregiously.
This article is based on materials supplied by Massachusetts Institute of Technology. The original article was written by Larry Hardesty.http://www.engineersjournal.ie/2013/10/17/reliable-communication-but-unreliable-networks%e2%80%a8/http://www.engineersjournal.ie/wp-content/uploads/2013/10/Netework-1024x724.jpghttp://www.engineersjournal.ie/wp-content/uploads/2013/10/Netework-300x300.jpgTechinternet,MIT,research