You may have to Search all our reviewed books and magazines, click the sign up button below to create a free account.
Markov Decision Processes (MDPs) are a mathematical framework for modeling sequential decision problems under uncertainty as well as reinforcement learning problems. Written by experts in the field, this book provides a global view of current research using MDPs in artificial intelligence. It starts with an introductory presentation of the fundamental aspects of MDPs (planning in MDPs, reinforcement learning, partially observable MDPs, Markov games and the use of non-classical criteria). It then presents more advanced research trends in the field and gives some concrete examples using illustrative real life applications.
This volume contains the 137 papers accepted for presentation at the 15th European Conference on Artificial Intelligence (ECAI '02), which is organized by the European Co-ordination Committee on Artificial Intelligence.
Software testing has greatly evolved since the first edition of this book in 2011. Testers are now required to work in "agile" teams and focus on automating test cases. It has thus been necessary to update this work, in order to provide fundamental knowledge that testers should have to be effective and efficient in today's world. This book describes the fundamental aspects of testing in the different lifecycles, and how to implement and benefit from reviews and static analysis. Multiple other techniques are approached, such as equivalence partitioning, boundary value analysis, use case testing, decision tables and state transitions. This second edition also covers test management, test progress monitoring and incident management, in order to ensure that the testing information is correctly provided to the stakeholders. This book provides detailed course-study material for the 2023 version of the ISTQB Foundation level syllabus, including sample questions to help prepare for exams.
The purpose of this book is to provide an overview of AI research, ranging from basic work to interfaces and applications, with as much emphasis on results as on current issues. It is aimed at an audience of master students and Ph.D. students, and can be of interest as well for researchers and engineers who want to know more about AI. The book is split into three volumes: - the first volume brings together twenty-three chapters dealing with the foundations of knowledge representation and the formalization of reasoning and learning (Volume 1. Knowledge representation, reasoning and learning) - the second volume offers a view of AI, in fourteen chapters, from the side of the algorithms (Volume...
This book constitutes the refereed proceedings of the Workshops which complemented the 11th International Conference on Practical Applications of Agents and Multi-Agent Systems, PAAMS 2013, held in Salamanca, Spain, in May 2013. This volume presents the papers that have been accepted for the workshops: Workshop on Agent-based Approaches for the Transportation Modeling and Optimization, Workshop on Agent-Based Solutions for Manufacturing and Supply Chain, Workshop on User-Centric Technologies and Applications, Workshop on Conflict Resolution in Decision Making, Workshop on Multi-Agent System Based Learning Environments, Workshop on Multi-agent based Applications for Sustainable Energy Systems, Workshop on Agents and multi-agent Systems for AAL and e-Health
The role of artificial intelligence (AI) applications in fields as diverse as medicine, economics, linguistics, logical analysis and industry continues to grow in scope and importance. AI has become integral to the effective functioning of much of the technical infrastructure we all now take for granted as part of our daily lives. This book presents the papers from the 21st biennial European Conference on Artificial Intelligence, ECAI 2014, held in Prague, Czech Republic, in August 2014. The ECAI conference remains Europe's principal opportunity for researchers and practitioners of Artificial Intelligence to gather and to discuss the latest trends and challenges in all subfields of AI, as we...
This book is a collection of 15 reviewed technical reports summarizing the presentations at the 2011 Joint Workshop of Fraunhofer IOSB and Institute for Anthropomatics, Vision and Fusion Laboratory. The covered topics include image processing, optical signal processing, visual inspection, pattern recognition and classification, human-machine interaction, world and situation modeling, autonomous system localization and mapping, information fusion, and trust propagation in sensor networks.
This book constitutes the refereed proceedings of the 19th Australian Joint Conference on Artificial Intelligence, AI 2006, held in Hobart, Australia, December 2006. Coverage includes foundations and knowledge based system, machine learning, connectionist AI, data mining, intelligent agents, cognition and user interface, vision and image processing, natural language processing and Web intelligence, neural networks, robotics, and AI applications.
Markov Decision Processes (MDPs) are widely popular in Artificial Intelligence for modeling sequential decision-making scenarios with probabilistic dynamics. They are the framework of choice when designing an intelligent agent that needs to act for long periods of time in an environment where its actions could have uncertain outcomes. MDPs are actively researched in two related subareas of AI, probabilistic planning and reinforcement learning. Probabilistic planning assumes known models for the agent's goals and domain dynamics, and focuses on determining how the agent should behave to achieve its objectives. On the other hand, reinforcement learning additionally learns these models based on...
This book constitutes revised and selected papers of the 9th European Workshop on Reinforcement Learning, EWRL 2011, which took place in Athens, Greece in September 2011. The papers presented were carefully reviewed and selected from 40 submissions. The papers are organized in topical sections online reinforcement learning, learning and exploring MDPs, function approximation methods for reinforcement learning, macro-actions in reinforcement learning, policy search and bounds, multi-task and transfer reinforcement learning, multi-agent reinforcement learning, apprenticeship and inverse reinforcement learning and real-world reinforcement learning.