Statistical Significance Testing for Natural Language Processing

Statistical Significance Testing for Natural Language Processing
Author :
Publisher : Springer Nature
Total Pages : 98
Release :
ISBN-10 : 9783031021749
ISBN-13 : 3031021746
Rating : 4/5 (49 Downloads)

Synopsis Statistical Significance Testing for Natural Language Processing by : Rotem Dror

Data-driven experimental analysis has become the main evaluation tool of Natural Language Processing (NLP) algorithms. In fact, in the last decade, it has become rare to see an NLP paper, particularly one that proposes a new algorithm, that does not include extensive experimental analysis, and the number of involved tasks, datasets, domains, and languages is constantly growing. This emphasis on empirical results highlights the role of statistical significance testing in NLP research: If we, as a community, rely on empirical evaluation to validate our hypotheses and reveal the correct language processing mechanisms, we better be sure that our results are not coincidental. The goal of this book is to discuss the main aspects of statistical significance testing in NLP. Our guiding assumption throughout the book is that the basic question NLP researchers and engineers deal with is whether or not one algorithm can be considered better than another one. This question drives the field forward as it allows the constant progress of developing better technology for language processing challenges. In practice, researchers and engineers would like to draw the right conclusion from a limited set of experiments, and this conclusion should hold for other experiments with datasets they do not have at their disposal or that they cannot perform due to limited time and resources. The book hence discusses the opportunities and challenges in using statistical significance testing in NLP, from the point of view of experimental comparison between two algorithms. We cover topics such as choosing an appropriate significance test for the major NLP tasks, dealing with the unique aspects of significance testing for non-convex deep neural networks, accounting for a large number of comparisons between two NLP algorithms in a statistically valid manner (multiple hypothesis testing), and, finally, the unique challenges yielded by the nature of the data and practices of the field.

Statistical Significance Testing for Natural Language Processing

Statistical Significance Testing for Natural Language Processing
Author :
Publisher : Morgan & Claypool Publishers
Total Pages : 118
Release :
ISBN-10 : 9781681737966
ISBN-13 : 1681737965
Rating : 4/5 (66 Downloads)

Synopsis Statistical Significance Testing for Natural Language Processing by : Rotem Dror

Data-driven experimental analysis has become the main evaluation tool of Natural Language Processing (NLP) algorithms. In fact, in the last decade, it has become rare to see an NLP paper, particularly one that proposes a new algorithm, that does not include extensive experimental analysis, and the number of involved tasks, datasets, domains, and languages is constantly growing. This emphasis on empirical results highlights the role of statistical significance testing in NLP research: If we, as a community, rely on empirical evaluation to validate our hypotheses and reveal the correct language processing mechanisms, we better be sure that our results are not coincidental. The goal of this book is to discuss the main aspects of statistical significance testing in NLP. Our guiding assumption throughout the book is that the basic question NLP researchers and engineers deal with is whether or not one algorithm can be considered better than another one. This question drives the field forward as it allows the constant progress of developing better technology for language processing challenges. In practice, researchers and engineers would like to draw the right conclusion from a limited set of experiments, and this conclusion should hold for other experiments with datasets they do not have at their disposal or that they cannot perform due to limited time and resources. The book hence discusses the opportunities and challenges in using statistical significance testing in NLP, from the point of view of experimental comparison between two algorithms. We cover topics such as choosing an appropriate significance test for the major NLP tasks, dealing with the unique aspects of significance testing for non-convex deep neural networks, accounting for a large number of comparisons between two NLP algorithms in a statistically valid manner (multiple hypothesis testing), and, finally, the unique challenges yielded by the nature of the data and practices of the field.

Foundations of Statistical Natural Language Processing

Foundations of Statistical Natural Language Processing
Author :
Publisher : MIT Press
Total Pages : 719
Release :
ISBN-10 : 9780262303798
ISBN-13 : 0262303795
Rating : 4/5 (98 Downloads)

Synopsis Foundations of Statistical Natural Language Processing by : Christopher Manning

Statistical approaches to processing natural language text have become dominant in recent years. This foundational text is the first comprehensive introduction to statistical natural language processing (NLP) to appear. The book contains all the theory and algorithms needed for building NLP tools. It provides broad but rigorous coverage of mathematical and linguistic foundations, as well as detailed discussion of statistical methods, allowing students and researchers to construct their own implementations. The book covers collocation finding, word sense disambiguation, probabilistic parsing, information retrieval, and other applications.

Validity, Reliability, and Significance

Validity, Reliability, and Significance
Author :
Publisher : Springer Nature
Total Pages : 179
Release :
ISBN-10 : 9783031570650
ISBN-13 : 3031570650
Rating : 4/5 (50 Downloads)

Synopsis Validity, Reliability, and Significance by : Stefan Riezler

Speech & Language Processing

Speech & Language Processing
Author :
Publisher : Pearson Education India
Total Pages : 912
Release :
ISBN-10 : 8131716724
ISBN-13 : 9788131716724
Rating : 4/5 (24 Downloads)

Synopsis Speech & Language Processing by : Dan Jurafsky

Foundations of Statistical Natural Language Processing

Foundations of Statistical Natural Language Processing
Author :
Publisher : MIT Press
Total Pages : 722
Release :
ISBN-10 : 0262133601
ISBN-13 : 9780262133609
Rating : 4/5 (01 Downloads)

Synopsis Foundations of Statistical Natural Language Processing by : Christopher Manning

Statistical approaches to processing natural language text have become dominant in recent years. This foundational text is the first comprehensive introduction to statistical natural language processing (NLP) to appear. The book contains all the theory and algorithms needed for building NLP tools. It provides broad but rigorous coverage of mathematical and linguistic foundations, as well as detailed discussion of statistical methods, allowing students and researchers to construct their own implementations. The book covers collocation finding, word sense disambiguation, probabilistic parsing, information retrieval, and other applications.

Evaluating Natural Language Processing Systems

Evaluating Natural Language Processing Systems
Author :
Publisher : Springer Science & Business Media
Total Pages : 256
Release :
ISBN-10 : 3540613099
ISBN-13 : 9783540613091
Rating : 4/5 (99 Downloads)

Synopsis Evaluating Natural Language Processing Systems by : Karen Sparck Jones

This book is about the patterns of connections between brain structures. It reviews progress on the analysis of neuroanatomical connection data and presents six different approaches to data analysis. The results of their application to data from cat and monkey cortex are explored. This volume sheds light on the organization of the brain that is specified by its wiring.

Pretrained Transformers for Text Ranking

Pretrained Transformers for Text Ranking
Author :
Publisher : Springer Nature
Total Pages : 307
Release :
ISBN-10 : 9783031021817
ISBN-13 : 3031021819
Rating : 4/5 (17 Downloads)

Synopsis Pretrained Transformers for Text Ranking by : Jimmy Lin

The goal of text ranking is to generate an ordered list of texts retrieved from a corpus in response to a query. Although the most common formulation of text ranking is search, instances of the task can also be found in many natural language processing (NLP) applications.This book provides an overview of text ranking with neural network architectures known as transformers, of which BERT (Bidirectional Encoder Representations from Transformers) is the best-known example. The combination of transformers and self-supervised pretraining has been responsible for a paradigm shift in NLP, information retrieval (IR), and beyond. This book provides a synthesis of existing work as a single point of entry for practitioners who wish to gain a better understanding of how to apply transformers to text ranking problems and researchers who wish to pursue work in this area. It covers a wide range of modern techniques, grouped into two high-level categories: transformer models that perform reranking in multi-stage architectures and dense retrieval techniques that perform ranking directly. Two themes pervade the book: techniques for handling long documents, beyond typical sentence-by-sentence processing in NLP, and techniques for addressing the tradeoff between effectiveness (i.e., result quality) and efficiency (e.g., query latency, model and index size). Although transformer architectures and pretraining techniques are recent innovations, many aspects of how they are applied to text ranking are relatively well understood and represent mature techniques. However, there remain many open research questions, and thus in addition to laying out the foundations of pretrained transformers for text ranking, this book also attempts to prognosticate where the field is heading.

Automated Essay Scoring

Automated Essay Scoring
Author :
Publisher : Springer Nature
Total Pages : 294
Release :
ISBN-10 : 9783031021824
ISBN-13 : 3031021827
Rating : 4/5 (24 Downloads)

Synopsis Automated Essay Scoring by : Beata Beigman Klebanov

This book discusses the state of the art of automated essay scoring, its challenges and its potential. One of the earliest applications of artificial intelligence to language data (along with machine translation and speech recognition), automated essay scoring has evolved to become both a revenue-generating industry and a vast field of research, with many subfields and connections to other NLP tasks. In this book, we review the developments in this field against the backdrop of Elias Page's seminal 1966 paper titled "The Imminence of Grading Essays by Computer." Part 1 establishes what automated essay scoring is about, why it exists, where the technology stands, and what are some of the main issues. In Part 2, the book presents guided exercises to illustrate how one would go about building and evaluating a simple automated scoring system, while Part 3 offers readers a survey of the literature on different types of scoring models, the aspects of essay quality studied in prior research, and the implementation and evaluation of a scoring engine. Part 4 offers a broader view of the field inclusive of some neighboring areas, and Part \ref{part5} closes with summary and discussion. This book grew out of a week-long course on automated evaluation of language production at the North American Summer School for Logic, Language, and Information (NASSLLI), attended by advanced undergraduates and early-stage graduate students from a variety of disciplines. Teachers of natural language processing, in particular, will find that the book offers a useful foundation for a supplemental module on automated scoring. Professionals and students in linguistics, applied linguistics, educational technology, and other related disciplines will also find the material here useful.

Finite-State Text Processing

Finite-State Text Processing
Author :
Publisher : Springer Nature
Total Pages : 140
Release :
ISBN-10 : 9783031021794
ISBN-13 : 3031021797
Rating : 4/5 (94 Downloads)

Synopsis Finite-State Text Processing by : Kyle Gorman

Weighted finite-state transducers (WFSTs) are commonly used by engineers and computational linguists for processing and generating speech and text. This book first provides a detailed introduction to this formalism. It then introduces Pynini, a Python library for compiling finite-state grammars and for combining, optimizing, applying, and searching finite-state transducers. This book illustrates this library's conventions and use with a series of case studies. These include the compilation and application of context-dependent rewrite rules, the construction of morphological analyzers and generators, and text generation and processing applications.