You may have to Search all our reviewed books and magazines, click the sign up button below to create a free account.
This book constitutes the proceedings of the 7th International Conference on Statistical Language and Speech Processing, SLSP 2019, held in Ljubljana, Slovenia, in October 2019. The 25 full papers presented together with one invited paper in this volume were carefully reviewed and selected from 48 submissions. They were organized in topical sections named: Dialogue and Spoken Language Understanding; Language Analysis and Generation; Speech Analysis and Synthesis; Speech Recognition; Text Analysis and Classification.
Computational creativity is an emerging field of research within AI that focuses on the capacity of machines to both generate and evaluate novel outputs that would, if produced by a human, be considered creative. This book is intended to be a canonical text for this new discipline, through which researchers and students can absorb the philosophy of the field and learn its methods. After a comprehensive introduction to the idea of systematizing creativity the contributions address topics such as autonomous intentionality, conceptual blending, literature mining, computational design, models of novelty, evaluating progress in related research, computer-supported human creativity and human-supported computer creativity, common-sense knowledge, and models of social creativity. Products of this research will have real consequences for the worlds of entertainment, culture, science, education, design, and art, in addition to artificial intelligence, and the book will be of value to practitioners and students in all these domains.
This book constitutes the proceedings of the 23rd International Conference on Discovery Science, DS 2020, which took place during October 19-21, 2020. The conference was planned to take place in Thessaloniki, Greece, but had to change to an online format due to the COVID-19 pandemic. The 26 full and 19 short papers presented in this volume were carefully reviewed and selected from 76 submissions. The contributions were organized in topical sections named: classification; clustering; data and knowledge representation; data streams; distributed processing; ensembles; explainable and interpretable machine learning; graph and network mining; multi-target models; neural networks and deep learning; and spatial, temporal and spatiotemporal data.
The two-volume set LNAI 14115 and 14116 constitutes the refereed proceedings of the 22nd EPIA Conference on Progress in Artificial Intelligence, EPIA 2023, held in Faial Island, Azores, in September 2023. The 85 full papers presented in these proceedings were carefully reviewed and selected from 163 submissions. The papers have been organized in the following topical sections: ambient intelligence and affective environments; ethics and responsibility in artificial intelligence; general artificial intelligence; intelligent robotics; knowledge discovery and business intelligence; multi-agent systems: theory and applications; natural language processing, text mining and applications; planning, scheduling and decision-making in AI; social simulation and modelling; artifical intelligence, generation and creativity; artificial intelligence and law; artificial intelligence in power and energy systems; artificial intelligence in medicine; artificial intelligence and IoT in agriculture; artificial intelligence in transportation systems; artificial intelligence in smart computing; artificial intelligence for industry and societies.
Embeddings have undoubtedly been one of the most influential research areas in Natural Language Processing (NLP). Encoding information into a low-dimensional vector representation, which is easily integrable in modern machine learning models, has played a central role in the development of NLP. Embedding techniques initially focused on words, but the attention soon started to shift to other forms: from graph structures, such as knowledge bases, to other types of textual content, such as sentences and documents. This book provides a high-level synthesis of the main embedding techniques in NLP, in the broad sense. The book starts by explaining conventional word vector space models and word embeddings (e.g., Word2Vec and GloVe) and then moves to other types of embeddings, such as word sense, sentence and document, and graph embeddings. The book also provides an overview of recent developments in contextualized representations (e.g., ELMo and BERT) and explains their potential in NLP. Throughout the book, the reader can find both essential information for understanding a certain topic from scratch and a broad overview of the most successful techniques developed in the literature.
Semantic change — how the meanings of words change over time — has preoccupied scholars since well before modern linguistics emerged in the late 19th and early 20th century, ushering in a new methodological turn in the study of language change. Compared to changes in sound and grammar, semantic change is the least understood. Ever since, the study of semantic change has progressed steadily, accumulating a vast store of knowledge for over a century, encompassing many languages and language families. Historical linguists also early on realized the potential of computers as research tools, with papers at the very first international conferences in computational linguistics in the 1960s. Suc...
This monograph addresses advances in representation learning, a cutting-edge research area of machine learning. Representation learning refers to modern data transformation techniques that convert data of different modalities and complexity, including texts, graphs, and relations, into compact tabular representations, which effectively capture their semantic properties and relations. The monograph focuses on (i) propositionalization approaches, established in relational learning and inductive logic programming, and (ii) embedding approaches, which have gained popularity with recent advances in deep learning. The authors establish a unifying perspective on representation learning techniques developed in these various areas of modern data science, enabling the reader to understand the common underlying principles and to gain insight using selected examples and sample Python code. The monograph should be of interest to a wide audience, ranging from data scientists, machine learning researchers and students to developers, software engineers and industrial researchers interested in hands-on AI solutions.
This book constitutes the refereed proceedings of the 40th European Conference on IR Research, ECIR 2018, held in Grenoble, France, in March 2018. The 39 full papers and 39 short papers presented together with 6 demos, 5 workshops and 3 tutorials, were carefully reviewed and selected from 303 submissions. Accepted papers cover the state of the art in information retrieval including topics such as: topic modeling, deep learning, evaluation, user behavior, document representation, recommendation systems, retrieval methods, learning and classication, and micro-blogs.
None
Exchange between the translation studies and the computational linguistics communities has traditionally not been very intense. Among other things, this is reflected by the different views on parallel corpora. While computational linguistics does not always strictly pay attention to the translation direction (e.g. when translation rules are extracted from (sub)corpora which actually only consist of translations), translation studies are amongst other things concerned with exactly comparing source and target texts (e.g. to draw conclusions on interference and standardization effects). However, there has recently been more exchange between the two fields – especially when it comes to the annotation of parallel corpora. This special issue brings together the different research perspectives. Its contributions show – from both perspectives – how the communities have come to interact in recent years.