You may have to Search all our reviewed books and magazines, click the sign up button below to create a free account.
Will provide a more elementary introduction to these topics than other books available; Gentle is the author of two other Springer books
Twenty-four contributions, intended for a wide audience from various disciplines, cover a variety of applications of heavy-tailed modeling involving telecommunications, the Web, insurance, and finance. Along with discussion of specific applications are several papers devoted to time series analysis, regression, classical signal/noise detection problems, and the general structure of stable processes, viewed from a modeling standpoint. Emphasis is placed on developments in handling the numerical problems associated with stable distribution (a main technical difficulty until recently). No index. Annotation copyrighted by Book News, Inc., Portland, OR
Computational inference is based on an approach to statistical methods that uses modern computational power to simulate distributional properties of estimators and test statistics. This book describes computationally intensive statistical methods in a unified presentation, emphasizing techniques, such as the PDF decomposition, that arise in a wide range of methods.
Statistical science as organized in formal academic departments is relatively new. With a few exceptions, most Statistics and Biostatistics departments have been created within the past 60 years. This book consists of a set of memoirs, one for each department in the U.S. created by the mid-1960s. The memoirs describe key aspects of the department’s history -- its founding, its growth, key people in its development, success stories (such as major research accomplishments) and the occasional failure story, PhD graduates who have had a significant impact, its impact on statistical education, and a summary of where the department stands today and its vision for the future. Read here all about how departments such as at Berkeley, Chicago, Harvard, and Stanford started and how they got to where they are today. The book should also be of interests to scholars in the field of disciplinary history.
This book provides a self-contained presentation on the structure of a large class of stable processes, known as self-similar mixed moving averages. The authors present a way to describe and classify these processes by relating them to so-called deterministic flows. The first sections in the book review random variables, stochastic processes, and integrals, moving on to rigidity and flows, and finally ending with mixed moving averages and self-similarity. In-depth appendices are also included. This book is aimed at graduate students and researchers working in probability theory and statistics.
Describes statistical intervals to quantify sampling uncertainty,focusing on key application needs and recently developed methodology in an easy-to-apply format Statistical intervals provide invaluable tools for quantifying sampling uncertainty. The widely hailed first edition, published in 1991, described the use and construction of the most important statistical intervals. Particular emphasis was given to intervals—such as prediction intervals, tolerance intervals and confidence intervals on distribution quantiles—frequently needed in practice, but often neglected in introductory courses. Vastly improved computer capabilities over the past 25 years have resulted in an explosion of the ...