ID: 2107.02499

Transfer Learning for Improving Results on Russian Sentiment Datasets

July 6, 2021

View on ArXiv
Anton Golubev, Natalia Loukachevitch
Computer Science
Computation and Language

In this study, we test transfer learning approach on Russian sentiment benchmark datasets using additional train sample created with distant supervision technique. We compare several variants of combining additional data with benchmark train samples. The best results were achieved using three-step approach of sequential training on general, thematic and original train samples. For most datasets, the results were improved by more than 3% to the current state-of-the-art methods. The BERT-NLI model treating sentiment classification problem as a natural language inference task reached the human level of sentiment analysis on one of the datasets.

Similar papers 1

Improving Results on Russian Sentiment Datasets

July 28, 2020

94% Match
Anton Golubev, Natalia Loukachevitch
Computation and Language

In this study, we test standard neural network architectures (CNN, LSTM, BiLSTM) and recently appeared BERT architectures on previous Russian sentiment evaluation datasets. We compare two variants of Russian BERT and show that for all sentiment tasks in this study the conversational variant of Russian BERT performs better. The best results were achieved by BERT-NLI model, which treats sentiment classification tasks as a natural language inference task. On one of the datasets,...

Find SimilarView on arXiv

Current Landscape of the Russian Sentiment Corpora

June 28, 2021

91% Match
Evgeny Kotelnikov
Computation and Language

Currently, there are more than a dozen Russian-language corpora for sentiment analysis, differing in the source of the texts, domain, size, number and ratio of sentiment classes, and annotation method. This work examines publicly available Russian-language corpora, presents their qualitative and quantitative characteristics, which make it possible to get an idea of the current landscape of the corpora for sentiment analysis. The ranking of corpora by annotation quality is pro...

Find SimilarView on arXiv

Lexicon-based Methods vs. BERT for Text Sentiment Analysis

November 19, 2021

91% Match
Anastasia Kotelnikova, Danil Paschenko, ... , Kotelnikov Evgeny
Computation and Language

The performance of sentiment analysis methods has greatly increased in recent years. This is due to the use of various models based on the Transformer architecture, in particular BERT. However, deep neural network models are difficult to train and poorly interpretable. An alternative approach is rule-based methods using sentiment lexicons. They are fast, require no training, and are well interpreted. But recently, due to the widespread use of deep learning, lexicon-based meth...

Find SimilarView on arXiv

Large Language Models in Targeted Sentiment Analysis

April 18, 2024

90% Match
Nicolay Rusnachenko, Anton Golubev, Natalia Loukachevitch
Computation and Language

In this paper we investigate the use of decoder-based generative transformers for extracting sentiment towards the named entities in Russian news articles. We study sentiment analysis capabilities of instruction-tuned large language models (LLMs). We consider the dataset of RuSentNE-2023 in our study. The first group of experiments was aimed at the evaluation of zero-shot capabilities of LLMs with closed and open transparencies. The second covers the fine-tuning of Flan-T5 us...

Find SimilarView on arXiv

Natural Language Processing Through Transfer Learning: A Case Study on Sentiment Analysis

November 28, 2023

89% Match
Aman Yadav, Abhishek Vichare
Computation and Language

Artificial intelligence and machine learning have significantly bolstered the technological world. This paper explores the potential of transfer learning in natural language processing focusing mainly on sentiment analysis. The models trained on the big data can also be used where data are scarce. The claim is that, compared to training models from scratch, transfer learning, using pre-trained BERT models, can increase sentiment classification accuracy. The study adopts a sop...

Find SimilarView on arXiv

Adaptation of Deep Bidirectional Multilingual Transformers for Russian Language

May 17, 2019

89% Match
Yuri Kuratov, Mikhail Arkhipov
Computation and Language

The paper introduces methods of adaptation of multilingual masked language models for a specific language. Pre-trained bidirectional language models show state-of-the-art performance on a wide range of tasks including reading comprehension, natural language inference, and sentiment analysis. At the moment there are two alternative approaches to train such models: monolingual and multilingual. While language specific models show superior performance, multilingual models allow ...

Find SimilarView on arXiv

RuSentNE-2023: Evaluating Entity-Oriented Sentiment Analysis on Russian News Texts

May 28, 2023

89% Match
Anton Golubev, Nicolay Rusnachenko, Natalia Loukachevitch
Computation and Language

The paper describes the RuSentNE-2023 evaluation devoted to targeted sentiment analysis in Russian news texts. The task is to predict sentiment towards a named entity in a single sentence. The dataset for RuSentNE-2023 evaluation is based on the Russian news corpus RuSentNE having rich sentiment-related annotation. The corpus is annotated with named entities and sentiments towards these entities, along with related effects and emotional states. The evaluation was organized us...

Find SimilarView on arXiv

Pretraining and Fine-Tuning Strategies for Sentiment Analysis of Latvian Tweets

October 23, 2020

89% Match
Gaurish Thakkar, Marcis Pinnis
Computation and Language

In this paper, we present various pre-training strategies that aid in im-proving the accuracy of the sentiment classification task. We, at first, pre-trainlanguage representation models using these strategies and then fine-tune them onthe downstream task. Experimental results on a time-balanced tweet evaluation setshow the improvement over the previous technique. We achieve 76% accuracy forsentiment analysis on Latvian tweets, which is a substantial improvement over pre-vious...

Find SimilarView on arXiv

Low-Shot Classification: A Comparison of Classical and Deep Transfer Machine Learning Approaches

July 17, 2019

89% Match
Peter Usherwood, Steven Smit
Machine Learning
Machine Learning

Despite the recent success of deep transfer learning approaches in NLP, there is a lack of quantitative studies demonstrating the gains these models offer in low-shot text classification tasks over existing paradigms. Deep transfer learning approaches such as BERT and ULMFiT demonstrate that they can beat state-of-the-art results on larger datasets, however when one has only 100-1000 labelled examples per class, the choice of approach is less clear, with classical machine lea...

Find SimilarView on arXiv

Modelling Sentiment Analysis: LLMs and data augmentation techniques

November 7, 2023

89% Match
Guillem Senabre Prades
Computation and Language

This paper provides different approaches for a binary sentiment classification on a small training dataset. LLMs that provided state-of-the-art results in sentiment analysis and similar domains are being used, such as BERT, RoBERTa and XLNet.

Find SimilarView on arXiv