Li, Zhengyan
Learning Implicit Sentiment in Aspect-based Sentiment Analysis with Supervised Contrastive Pre-Training
Li, Zhengyan, Zou, Yicheng, Zhang, Chong, Zhang, Qi, Wei, Zhongyu
Aspect-based sentiment analysis aims to identify the sentiment polarity of a specific aspect in product reviews. We notice that about 30% of reviews do not contain obvious opinion words, but still convey clear human-aware sentiment orientation, which is known as implicit sentiment. However, recent neural network-based approaches paid little attention to implicit sentiment entailed in the reviews. To overcome this issue, we adopt Supervised Contrastive Pre-training on large-scale sentiment-annotated corpora retrieved from in-domain language resources. By aligning the representation of implicit sentiment expressions to those with the same sentiment label, the pre-training process leads to better capture of both implicit and explicit sentiment orientation towards aspects in reviews. Experimental results show that our method achieves state-of-the-art performance on SemEval2014 benchmarks, and comprehensive analysis validates its effectiveness on learning implicit sentiment.
TextFlint: Unified Multilingual Robustness Evaluation Toolkit for Natural Language Processing
Gui, Tao, Wang, Xiao, Zhang, Qi, Liu, Qin, Zou, Yicheng, Zhou, Xin, Zheng, Rui, Zhang, Chong, Wu, Qinzhuo, Ye, Jiacheng, Pang, Zexiong, Zhang, Yongxin, Li, Zhengyan, Ma, Ruotian, Fei, Zichu, Cai, Ruijian, Zhao, Jun, Hu, Xinwu, Yan, Zhiheng, Tan, Yiding, Hu, Yuan, Bian, Qiyuan, Liu, Zhihua, Zhu, Bolin, Qin, Shan, Xing, Xiaoyu, Fu, Jinlan, Zhang, Yue, Peng, Minlong, Zheng, Xiaoqing, Zhou, Yaqian, Wei, Zhongyu, Qiu, Xipeng, Huang, Xuanjing
Various robustness evaluation methodologies from different perspectives have been proposed for different natural language processing (NLP) tasks. These methods have often focused on either universal or task-specific generalization capabilities. In this work, we propose a multilingual robustness evaluation platform for NLP tasks (TextFlint) that incorporates universal text transformation, task-specific transformation, adversarial attack, subpopulation, and their combinations to provide comprehensive robustness analysis. TextFlint enables practitioners to automatically evaluate their models from all aspects or to customize their evaluations as desired with just a few lines of code. To guarantee user acceptability, all the text transformations are linguistically based, and we provide a human evaluation for each one. TextFlint generates complete analytical reports as well as targeted augmented data to address the shortcomings of the model's robustness. To validate TextFlint's utility, we performed large-scale empirical evaluations (over 67,000 evaluations) on state-of-the-art deep learning models, classic supervised methods, and real-world systems. Almost all models showed significant performance degradation, including a decline of more than 50% of BERT's prediction accuracy on tasks such as aspect-level sentiment classification, named entity recognition, and natural language inference. Therefore, we call for the robustness to be included in the model evaluation, so as to promote the healthy development of NLP technology.