You may have to Search all our reviewed books and magazines, click the sign up button below to create a free account.
This book aims at being a comprehensive and pedagogical introduction to the concept of self-stabilization, introduced by Edsger Wybe Dijkstra in 1973. Self-stabilization characterizes the ability of a distributed algorithm to converge within finite time to a configuration from which its behavior is correct (i.e., satisfies a given specification), regardless the arbitrary initial configuration of the system. This arbitrary initial configuration may be the result of the occurrence of a finite number of transient faults. Hence, self-stabilization is actually considered as a versatile non-masking fault tolerance approach, since it recovers from the effect of any finite number of such faults in a...
This book constitutes the proceedings of the 18th International Conference on Distributed Computing and Intelligent Technology, ICDCIT 2022, held in Bhubaneswar, India, in January 20212. The 11 full papers presented together with 4 short papers were carefully reviewed and selected from 50 submissions. There are also 4 invited papers included. The papers were organized in topical sections named: invited papers, distributed computing and intelligent technology.
This book constitutes the refereed proceedings of the 18th Annual Symposium on Theoretical Aspects of Computer Science, STACS 2001, held in Dresden, Germany in February 2001. The 46 revised full papers presented together with three invited papers were carefully reviewed and selected from a total of 153 submissions. The papers address foundational aspects from all current areas of theoretical computer science including algorithms, data structures, automata, formal languages, complexity, verification, logic, graph theory, optimization, etc.
This book constitutes the proceedings of the 13th International Symposium on Stabilization, Safety, and Security of Distributed Systems, SSS 2011, held in Grenoble, France, in October 2011. The 29 papers presented were carefully reviewed and selected from 79 submissions. They cover the following areas: ad-hoc, sensor, and peer-to-peer networks; safety and verification; security; self-organizing and autonomic systems; and self-stabilization.
This book constitutes the refereed proceedings of the 13th International Conference on Distributed Computing and Networking, ICDCN 2012, held in Hong Kong, China, during January 3-6, 2012. The 36 revised full papers and 1 short paper presented together with 4 poster papers were carefully reviewed and selected from 100 submissions. The papers address all current issues in the field of distributed computing and networking. Being a leading forum for researchers and practitioners to exchange ideas and share best practices, ICDCN also hosts as a forum for PhD students to discuss their research ideas and get quality feedback from the well-renowned experts in the field of distributed computing and computer networking.
Providing a shared memory abstraction in distributed systems is a powerful tool that can simplify the design and implementation of software systems for networked platforms. This enables the system designers to work with abstract readable and writable objects without the need to deal with the complexity and dynamism of the underlying platform. The key property of shared memory implementations is the consistency guarantee that it provides under concurrent access to the shared objects. The most intuitive memory consistency model is atomicity because of its equivalence with a memory system where accesses occur serially, one at a time. Emulations of shared atomic memory in distributed systems is an active area of research and development. The problem proves to be challenging, and especially so in distributed message passing settings with unreliable components, as is often the case in networked systems. We present several approaches to implementing shared memory services with the help of replication on top of message-passing distributed platforms subject to a variety of perturbations in the computing medium.
Formal methods are coming of age. Mathematical techniques and tools are now regarded as an important part of the development process in a wide range of industrial and governmental organisations. A transfer of technology into the mainstream of systems development is slowly, but surely, taking place. FM’99, the First World Congress on Formal Methods in the Development of Computing Systems, is a result, and a measure, of this new-found maturity. It brings an impressive array of industrial and applications-oriented papers that show how formal methods have been used to tackle real problems. These proceedings are a record of the technical symposium ofFM’99:alo- side the papers describingapplic...
This book constitutes the refereed proceedings of the 14th International Conference on Tests and Proofs, TAP 2020, held as part of the 4th World Congress on Formal Methods 2020, Bergen, Norway, in June 2020. The 7 regular papers, 1 short paper and 2 demonstration papers presented in this volume were carefully reviewed and selected from 209 submissions. The TAP conference promotes research in verification and formal methods that targets the interplay of proofs and testing: the advancement of techniques of each kind and their combination, with the ultimate goal of improving software and system dependability.
This book constitutes the refereed proceedings of the 19th International Colloquium on Structural Information and Communication Complexity, SIROCCO 2012, held in Reykjavik, Iceland for 3 days starting June 30, 2012. The 28 revised full papers presented were carefully reviewed and selected from 54 submissions. SIROCCO is devoted to the study of communication and knowledge in distributed systems. Special emphasis is given to innovative approaches and fundamental understanding, in addition to efforts to optimize current designs. The typical areas include distributed computing, communication networks, game theory, parallel computing, social networks, mobile computing (including autonomous robots), peer to peer systems, communication complexity, fault tolerant graph theories, and randomized/probabilistic issues in networks.