You may have to Search all our reviewed books and magazines, click the sign up button below to create a free account.
This volume contains the papers which have been accepted for presentation atthe Third International Symposium on Programming Language Implementation andLogic Programming (PLILP '91) held in Passau, Germany, August 26-28, 1991. The aim of the symposium was to explore new declarative concepts, methods and techniques relevant for the implementation of all kinds of programming languages, whether algorithmic or declarative ones. The intention was to gather researchers from the fields of algorithmic programming languages as well as logic, functional and object-oriented programming. This volume contains the two invited talks given at the symposium by H. Ait-Kaci and D.B. MacQueen, 32 selected papers, and abstracts of several system demonstrations. The proceedings of PLILP '88 and PLILP '90 are available as Lecture Notes in Computer Science Volumes 348 and 456.
The second half of the 1970s was marked with impressive advances in array/vector architectures and vectorization techniques and compilers. This progress continued with a particular focus on vector machines until the middle of the 1980s. The major ity of supercomputers during this period were register-to-register (Cray 1) or memory-to-memory (CDC Cyber 205) vector (pipelined) machines. However, the increasing demand for higher computational rates lead naturally to parallel comput ers and software. Through the replication of autonomous processors in a coordinated system, one can skip over performance barriers due technology limitations. In princi ple, parallelism offers unlimited performance p...
Supercomputers are the largest and fastest computers available at any point in time. The term was used for the first time in the New York World, March 1920, to describe "new statistical machines with the mental power of 100 skilled mathematicians in solving even highly complex algebraic problems. " Invented by Mendenhall and Warren, these machines were used at Columbia University'S Statistical Bureau. Recently, supercomputers have been used primarily to solve large-scale prob lems in science and engineering. Solutions of systems of partial differential equa tions, such as those found in nuclear physics, meteorology, and computational fluid dynamics, account for the majority of supercomputer ...
This book has been written for practitioners, researchers and stu dents in the fields of parallel and distributed computing. Its objective is to provide detailed coverage of the applications of graph theoretic tech niques to the problems of matching resources and requirements in multi ple computer systems. There has been considerable research in this area over the last decade and intense work continues even as this is being written. For the practitioner, this book serves as a rich source of solution techniques for problems that are routinely encountered in the real world. Algorithms are presented in sufficient detail to permit easy implementa tion; background material and fundamental concepts are covered in full. The researcher will find a clear exposition of graph theoretic tech niques applied to parallel and distributed computing. Research results are covered and many hitherto unpublished spanning the last decade results by the author are included. There are many unsolved problems in this field-it is hoped that this book will stimulate further research.
None
The Complexity of Robot Motion Planning makes original contributions both to roboticsand to the analysis of algorithms. In this groundbreaking monograph John Canny resolveslong-standing problems concerning the complexity of motion planning and, for the central problem offinding a collision free path for a jointed robot in the presence of obstacles, obtains exponentialspeedups over existing algorithms by applying high-powered new mathematical techniques.Canny's newalgorithm for this "generalized movers' problem," the most-studied and basic robot motion planningproblem, has a single exponential running time, and is polynomial for any given robot. The algorithmhas an optimal running time expone...
A reissue of Paul "Bear" Bryant's autobiography, this edition features a completely new introduction and an accompanying audio CD of Bryant himself, in his own voice, talking about his life and football. It's all here, in his own inimitable words and with a candor that is both remarkable and eminently revealing. From his hardscrabble youth as the third youngest of 13 children of a dirt-poor farmer in Moro Bottom, Arkansas, to his playing days at the University of Alabama and fortuitous marriage to the remarkable Mary Harmon Black, to his first stabs at coaching as an assistant coach, to his 38 years as a head coach, coaching marquis names like Namath and Crow and Parilli, to his 323 victories and a record six National Championships.
One suspects that the people who use computers for their livelihood are growing more "sophisticated" as the field of computer science evolves. This view might be defended by the expanding use of languages such as C and Lisp in contrast to the languages such as FORTRAN and COBOL. This hypothesis is false however - computer languages are not like natural languages where successive generations stick with the language of their ancestors. Computer programmers do not grow more sophisticated - programmers simply take the time to muddle through the increasingly complex language semantics in an attempt to write useful programs. Of course, these programmers are "sophisticated" in the same sense as are...
Proceedings of the international conference held in Beijing, April 17-21, 1989. Topics covered include: knowledge representation, knowledge base systems, intelligent environment, knowledge acquisition, on parallel architecture, computation models, Prolog machines, knowledge base machines, functional