Collaborating Authors




If the question'What is sentiment analysis?' popped up in your mind as you clicked on this blog, I think you will find my first blog in this series interesting. Essentially, sentiment analysis is a natural language processing technique used to determine the emotional tone of textual data. It is primarily used to understand customer satisfaction, and gauge brand reputation, call center interactions as well as customer feedback and messages. There are various types of sentiment analysis that are common in the real world. In this part of my blog series, let me walk you through the implementation of sentiment analysis.

Understand your Customer Better with Sentiment Analysis


Originally published on Towards AI the World's Leading AI and Technology News and Media Company. If you are building an AI-related product or service, we invite you to consider becoming an AI sponsor. At Towards AI, we help scale AI and technology startups. Let us help you unleash your technology to the masses. "Your most unhappy customers are your greatest source of learning."

How to leverage AI for social media sentiment analysis - ET CIO


In a world where a single tweet can make or break a brand, it is crucial for companies and brands to invest in social media automation and analysis to derive actionable insights on brand perception. You would not like to wait for 12 hours to reply to that negative comment while #quit prefixed with your brand name trends on Twitter and Instagram, would you? Studies have shown that customers tend to be more vocal and frank with their views on social media. How they perceive a particular brand, its products/services fundamentally influence their behavior. So, for brands, being able to dig deep into the comments, replies, conversations, etc from customers can help uncover an unbiased view of their customers' behavior and persona, helping them understand customer intent and sentiments better.

Logical Reasoning for Task Oriented Dialogue Systems Artificial Intelligence

In recent years, large pretrained models have been used in dialogue systems to improve successful task completion rates. However, lack of reasoning capabilities of dialogue platforms make it difficult to provide relevant and fluent responses, unless the designers of a conversational experience spend a considerable amount of time implementing these capabilities in external rule based modules. In this work, we propose a novel method to fine-tune pretrained transformer models such as Roberta and T5. to reason over a set of facts in a given dialogue context. Our method includes a synthetic data generation mechanism which helps the model learn logical relations, such as comparison between list of numerical values, inverse relations (and negation), inclusion and exclusion for categorical attributes, and application of a combination of attributes over both numerical and categorical values, and spoken form for numerical values, without need for additional training dataset. We show that the transformer based model can perform logical reasoning to answer questions when the dialogue context contains all the required information, otherwise it is able to extract appropriate constraints to pass to downstream components (e.g. a knowledge base) when partial information is available. We observe that transformer based models such as UnifiedQA-T5 can be fine-tuned to perform logical reasoning (such as numerical and categorical attributes' comparison) over attributes that been seen in training time (e.g., accuracy of 90\%+ for comparison of smaller than $k_{\max}$=5 values over heldout test dataset).

Automatic Recognition of the General-Purpose Communicative Functions Defined by the ISO 24617-2 Standard for Dialog Act Annotation

Journal of Artificial Intelligence Research

From the perspective of a dialog system, it is important to identify the intention behind the segments in a dialog, since it provides an important cue regarding the information that is present in the segments and how they should be interpreted. ISO 24617-2, the standard for dialog act annotation, defines a hierarchically organized set of general-purpose communicative functions which correspond to different intentions that are relevant in the context of a dialog. We explore the automatic recognition of these communicative functions in the DialogBank, which is a reference set of dialogs annotated according to this standard. To do so, we propose adaptations of existing approaches to flat dialog act recognition that allow them to deal with the hierarchical classification problem. More specifically, we propose the use of an end-to-end hierarchical network with cascading outputs and maximum a posteriori path estimation to predict the communicative function at each level of the hierarchy, preserve the dependencies between the functions in the path, and decide at which level to stop. Furthermore, since the amount of dialogs in the DialogBank is small, we rely on transfer learning processes to reduce overfitting and improve performance. The results of our experiments show that our approach outperforms both a flat one and hierarchical approaches based on multiple classifiers and that each of its components plays an important role towards the recognition of general-purpose communicative functions.

Natural Language Processing and Sentiment Analysis


You're likely familiar with the saying, "Texting is a brilliant way to miscommunicate how you feel and misinterpret what other people mean." You've probably even experienced it directly! Substitute "texting" with "email" or "online reviews" and you've struck the nerve of businesses worldwide. Gaining a proper understanding of what clients and consumers have to say about your product or service or, more importantly, how they feel about your brand, is a universal struggle for businesses everywhere. What if I told you it doesn't have to be this way?

TourBERT: A pretrained language model for the tourism industry Artificial Intelligence

Tourism is one of the most important economic sectors in the world (Hollenhorst The Bidirectional Encoder Representations et al., 2014), and its services have many from Transformers (BERT) is currently the characteristics that distinguish them from most important and state-of-the-art natural other products. Services are not tangible language model (Tenney et al., 2019) since and cannot be tested in advance, which is its launch in 2018 by Google. BERT Large, why the customer assumes an increased which is based on a Transformer risk before starting the trip. The service is architecture, is considered one of the most co-created together with the customer, so powerful language models with 24 layers, the customer is an active co-creator of the 16 attention heads, and 340 million service. Services are subject to the unoactu parameters (Lan et al. 2019). BERT is a principle, which means they are pretrained model and can be fine-tuned to produced at the same time as they are perform numerous downstream tasks such consumed, and they are considered as text classification, question answering, bilateral, i.e. a reciprocal relationship sentiment analysis, extractive between persons (Chehimi, 2014). In summarization, named entity recognition, addition, tourism services are relatively or sentence similarity (Egger, 2022). The expensive compared to everyday products model was pretrained on a huge English and have an intercultural dimension.

Knowledge Graph Augmented Network Towards Multiview Representation Learning for Aspect-based Sentiment Analysis Artificial Intelligence

Aspect-based sentiment analysis (ABSA) is a fine-grained task of sentiment analysis. To better comprehend long complicated sentences and obtain accurate aspect-specific information, linguistic and commonsense knowledge are generally required in this task. However, most methods employ complicated and inefficient approaches to incorporate external knowledge, e.g., directly searching the graph nodes. Additionally, the complementarity between external knowledge and linguistic information has not been thoroughly studied. To this end, we propose a knowledge graph augmented network (KGAN), which aims to effectively incorporate external knowledge with explicitly syntactic and contextual information. In particular, KGAN captures the sentiment feature representations from multiple different perspectives, i.e., context-, syntax- and knowledge-based. First, KGAN learns the contextual and syntactic representations in parallel to fully extract the semantic features. Then, KGAN integrates the knowledge graphs into the embedding space, based on which the aspect-specific knowledge representations are further obtained via an attention mechanism. Last, we propose a hierarchical fusion module to complement these multiview representations in a local-to-global manner. Extensive experiments on three popular ABSA benchmarks demonstrate the effectiveness and robustness of our KGAN. Notably, with the help of the pretrained model of RoBERTa, KGAN achieves a new record of state-of-the-art performance.

Semantic and sentiment analysis of selected Bhagavad Gita translations using BERT-based language framework Artificial Intelligence

It is well known that translations of songs and poems not only breaks rhythm and rhyming patterns, but also results in loss of semantic information. The Bhagavad Gita is an ancient Hindu philosophical text originally written in Sanskrit that features a conversation between Lord Krishna and Arjuna prior to the Mahabharata war. The Bhagavad Gita is also one of the key sacred texts in Hinduism and known as the forefront of the Vedic corpus of Hinduism. In the last two centuries, there has been a lot of interest in Hindu philosophy by western scholars and hence the Bhagavad Gita has been translated in a number of languages. However, there is not much work that validates the quality of the English translations. Recent progress of language models powered by deep learning has enabled not only translations but better understanding of language and texts with semantic and sentiment analysis. Our work is motivated by the recent progress of language models powered by deep learning methods. In this paper, we compare selected translations (mostly from Sanskrit to English) of the Bhagavad Gita using semantic and sentiment analyses. We use hand-labelled sentiment dataset for tuning state-of-art deep learning-based language model known as \textit{bidirectional encoder representations from transformers} (BERT). We use novel sentence embedding models to provide semantic analysis for selected chapters and verses across translations. Finally, we use the aforementioned models for sentiment and semantic analyses and provide visualisation of results. Our results show that although the style and vocabulary in the respective Bhagavad Gita translations vary widely, the sentiment analysis and semantic similarity shows that the message conveyed are mostly similar across the translations.

Auto-ABSA: Automatic Detection of Aspects in Aspect-Based Sentiment Analysis Artificial Intelligence

After transformer is proposed, lots of pre-trained language models have been come up with and sentiment analysis (SA) task has been improved. In this paper, we proposed a method that uses an auxiliary sentence about aspects that the sentence contains to help sentiment prediction. The first is aspect detection, which uses a multi-aspects detection model to predict all aspects that the sentence has. Combining the predicted aspects and the original sentence as Sentiment Analysis (SA) model's input. The second is to do out-of-domain aspect-based sentiment analysis(ABSA), train sentiment classification model with one kind of dataset and validate it with another kind of dataset. Finally, we created two baselines, they use no aspect and all aspects as sentiment classification model's input, respectively. Compare two baselines performance to our method, found that our method really makes sense.