Summary
Natural Language Processing or NLP uses computational techniques to analyze large amounts of natural language data. The goal of NLP is to develop algorithms and systems that can understand, interpret and generate human language. NLP aims at enabling computers to communicate with people in natural language using speech and text. It is a broad interdisciplinary field that combines computer science, artificial intelligence and linguistics. Information Extraction: Extract structured information from unstructured text. This can be used to extract entities, keywords, relationships, events etc. from text. Machine Translation: Translate text from one language to another. This enables communication across languages. Sentiment Analysis: Analyze the sentiment or emotion behind text. This can be used to analyze reviews, social media posts etc. to determine sentiments like positive, negative or neutral. Text Summarization: Summarize large documents into shorter versions preserving the key ideas and concepts. This can help in quickly grasping the main essence from lengthy documents. Chatbots: Automate conversations with users through messaging applications. Chatbots use NLP to understand user messages and respond appropriately. Question Answering: Build systems that can automatically answer questions posed by humans in natural language. (1) Document Acquisition: Obtain the raw text data which needs to be processed. This could come from websites, apps, pdfs etc. (2) Pre-processing: The raw text needs to be cleaned by converting to lowercase, removing stop words, stemming etc. This prepares the text for further analysis. (3) Linguistic Analysis: Analyze the text linguistically by determining things like parts of speech, entities, semantic roles etc. This helps understand the structure and meaning of the text. (4)Generation or Application: Use the linguistic analysis to generate a summary or translation of the text or build an application like a chatbot. In recent years, deep learning techniques have enabled huge progress in NLP. Methods like word embeddings, recurrent neural networks, transformers etc. have led to state-of-the-art results in many NLP tasks. Pre-trained language models like BERT, GPT-3 have further boosted NLP capabilities. NLP is a fast growing field with many exciting applications and a active area of research.
Transformer-based models have pushed state of the art in many areas of NLP, but our understanding of what is behind their success is still limited. This paper is the first survey of over 150 studies of the popular BERT model. We review the current state ofknowledge about how BERT works, what kind of information it learns and how it is represented, common modifications to its training objectives and architecture, the over parameterization issue, and approaches to compression. We then outline directions for future research.
Published By:
Anna Rogers - Transactions of the Association for Computational Linguistics
2020
Cited By:
1060
Large language models present privacy risks through memorization of training data though little attention goes to fine-tuning risks or comparing methods. We study memorization of the methods via attacks to show they differ.
Published By:
Fatemehsadat Mireshghallah - arXiv.org
2022
Cited By:
14
LegalAI focuses on applying AI technology to legal tasks. In recent years, LegalAI has drawn increasing attention from AI researchers and legal professionals, as LegalAI is beneficial to the legal system for liberating legal professionals from paperwork. Legal professionals focus on rule-based and symbolic methods, while NLP researchers focus on data-driven and embedding methods. We introduce LegalAI's history, current state, and future directions. We illustrate tasks from legal and NLP perspectives and show applications. We analyze existing works' pros and cons to explore future directions. implementation: https://github.com/thunlp/CLAIM.
Published By:
Haoxiang Zhong - Annual Meeting of the Association for Computational Linguistics
2020
Cited By:
162
Natural Language Processing(NLP)is widely used to support the automation of different Requirements Engineering(RE)tasks.Most of the proposed approaches start with various NLP stepsthat analyze requirements statements, extract their linguistic information, and convert them to easy-to-process representations,such as lists of features or embedding-based vector representations.These NLP-based representations usually used at a later stage as inputs for machine learning techniques or rule-based methods.Thus, requirements representations play a major role determining the accuracy of different approaches. After compiling an initial pool of 2,227 papers, and applying a set of inclusion/exclusion criteria, we obtained a final pool containing 104 relevant papers. Our survey shows that the research direction has changed from the use of lexical and syntactic features to the use of advanced embedding techniques, especially in the last two years. Using embeddingrepresentations has proved its effectiveness in most RE tasks(such as requirement analysis, extracting requirements from reviews and forums, and semantic-level quality tasks).However,representations that are based on lexical and syntactic featuresare still more appropriate for other RE tasks(such as modeling and syntax-level quality tasks) since they provide the required information for the rules and regular expressions used when handling these tasks.In addition, we identify four gaps in the existing literature, why they matter, and how future research can begin to address them.
Published By:
R. Sonbol - IEEE Access
2022
Cited By:
10
To explain NLP models a popular approach is to use importance measures, such as attention, which inform input tokens are important for making a prediction. However, an open question is how well these explanations accurately reflect a model's logic, a property called faithfulness. To answer this question, we propose Recursive ROAR, a new faithfulness metric.This works by recursively masking allegedly important tokens and then retraining the model. The result is a performance curve given a masking-ratio. Furthermore, we propose a summarizing metric using relative area-between-curves (RACU), which allows for easy comparison across papers, models, and tasks. We evaluate 4 different importance measures on 8 different datasets, using both LSTM-attention models and RoBERTa models. We find that the faithfulness of importance measures is both model-dependent and task-dependent. This conclusion contradicts previous evaluations in both computer vision and faithfulness of attention literature.
Published By:
Andreas Madsen - Conference on Empirical Methods in Natural Language Processing
2021
Cited By:
21
We propose a backdoor scanning technique to detect backdoors in NLP models. It transforms a subject model to an equivalent but differentiable form and uses optimization to invert a distribution of words denoting their likelihood in the trigger. It then uses word analysis to see if the model is discriminative for trigger words. Evaluation on 3839 NLP models shows it is highly effective, achieving over 0.9 detection accuracy and outperforming state-of-the-art scanners.
Published By:
Yingqi Liu - IEEE Symposium on Security and Privacy
2022
Cited By:
34
There are many resources for English NLP but few for Greek. A survey of Greek NLP tools and methods could benefit researchers.
Published By:
Katerina Papantoniou - Hellenic Conference on Artificial Intelligence
2020
Cited By:
12
We propose using page layout features to improve text extraction from legal documents. Splitting pages by type improves accuracy, achieving 0.98 for single-column and 0.96 for double-column layouts.
Published By:
Frieda Josi - undefined
2022
Cited By:
2
NLMs challenges are explained through machine rationales, less utilizing rationales to improve NLM behavior.Explanation regularization aims to improve NLM behavior by aligning machine and human rationales.Prior works evaluate ER models on in-distribution generalization, little on out-of-distribution.Little is understood how ER model performance is affected by the choice of ER criteria or by the number/choice of training instances with human rationales. We propose ER-TEST, evaluating ER models’ out-of-distribution generalization: (1) unseen datasets, (2) contrast set tests, (3) functional tests.Using ER-TEST, we study: (A) Effective ER criteria for given out-of-distribution setting? (B) How is ER affected by number/choice of instances with human rationales?
Published By:
Brihi Joshi - undefined
2022
Cited By:
3