Journal Pre-proof SACPC: A framework based on probabilistic linguistic terms for short text sentiment analysis Chao Song, Xiao-Kang Wang, Peng-fei Cheng, Jian-qiang Wang, Lin Li
PII: DOI: Reference:
S0950-7051(20)30059-9 https://doi.org/10.1016/j.knosys.2020.105572 KNOSYS 105572
To appear in:
Knowledge-Based Systems
Received date : 4 April 2019 Revised date : 28 October 2019 Accepted date : 23 January 2020 Please cite this article as: C. Song, X.-K. Wang, P.-f. Cheng et al., SACPC: A framework based on probabilistic linguistic terms for short text sentiment analysis, Knowledge-Based Systems (2020), doi: https://doi.org/10.1016/j.knosys.2020.105572. This is a PDF file of an article that has undergone enhancements after acceptance, such as the addition of a cover page and metadata, and formatting for readability, but it is not yet the definitive version of record. This version will undergo additional copyediting, typesetting and review before it is published in its final form, but we are providing this version to give early visibility of the article. Please note that, during the production process, errors may be discovered which could affect the content, and all legal disclaimers that apply to the journal pertain.
© 2020 Published by Elsevier B.V.
*Revised Manuscript (Clean Version) Click here to view linked References
Journal Pre-proof
SACPC: A framework based on probabilistic linguistic terms for short text sentiment analysis Chao Song1, Xiao-Kang Wang1*, Peng-fei Cheng2, Jian-qiang Wang2,1, Lin Li3
pro of
1.School of Business, Central South University, Changsha 410083, PR China 2.Hunan Engineering Research Center for Intelligent Decision Making and Big Data on Industrial Development, Hunan University of Science and Technology, Xiangtan 411201, PR China 3. School of Business, Hunan University, Changsha 410082, PR China Correspondence should be addressed to Xiao-Kang Wang:
[email protected]
re-
Abstract: Short text sentiment analysis is challenging because short texts are limited in length and lack context. Short texts are usually rather ambiguous because of polysemy and the typos these texts contain. Polysemy is the coexistence of multiple word meanings and commonly appears in every language. Various uses of a word may assign the word both positive and negative meanings. In previous studies, the variability of words is often
lP
ignored, which may cause analysis errors. In this study, to resolve this problem, we proposed a novel text representation model named Word2PLTS for short text sentiment analysis by introducing probabilistic linguistic terms sets (PLTSs) and the relevant theory. In this model, every word is represented as a PLTS that
urn a
fully describes the possibilities for the sentiment polarity of the word. Then, by using support vector machines (SVM), a novel sentiment analysis and polarity classification framework named SACPC is obtained. This framework is a technique that combines supervised learning and unsupervised learning. We compare SAPCP to lexicon-based approaches and machine learning approaches on three benchmark datasets. A noticeable improvement in performance is achieved. To further verify the superiority of SAPCP, state of the art
Jo
performance comparisons are conducted, and the results are impressive. Keywords: semantic change; sentiment analysis; probabilistic linguistic terms; polarity classification 1. Introduction
In recent decades, the rapid development of the Internet has promoted a transformation from the "reading Internet" to the "interactive Internet". Social networking has revolutionized the daily life of people. The Internet not only has become an important source of information for people but also provides an important platform for people to voice their views, share their experiences and express their feelings of anger, sadness 1
Journal Pre-proof and joy. The Internet allows for the presence of a large number of sentiment reviews. Such a huge flow of information seems trivial and complex, but the huge potential value hidden in the information should not be ignored. Extensive natural language processing (NLP) is applied to process this vital information. Sentiment analysis is an import branch of NLP that involves various disciplines such as statistics, linguistics, psychology
pro of
and artificial intelligence. NLP usually captures an evaluative factor (positive or negative) and potency or strength (degree to which the word, phrase, sentence, or document in question is positive or negative) towards a subject topic, person, or idea [1]. Sentiment analysis systems have been applied to many different kinds of texts including customer reviews [2-4], news [5, 6] and tweets [7-10].
There are two main categories for the automatic extraction of sentiment information: lexicon-based approaches and machine-learning approaches. An overview of the categorization of classification methods is given in Fig. 1. In most lexicon-based approaches, the sentiment polarity score of a word is expressed by a real
re-
number; when the number greater than zero, the word is labelled positive. Otherwise, the word is classified negative. In machine learning, a vector space model is a frequently used method for representing text documents, and each dimension corresponds to a specific word. The value of the word, known as the weight, is
lP
calculated on the basis of the frequency of the word. However, almost all approaches seem to ignore the
Jo
urn a
problem of the variability of words, and the sentiment polarity of a word is considered to be constant.
Fig. 1. Sentiment analysis methods.
Unlike analysis of sentences and documents, sentiment analysis of short texts is difficult. Short text sentiment analysis is rather ambiguous and fuzzy because of limited contextual information, polysemy and the typos the text contains. Polysemy is the coexistence of multiple word meanings and appears widely in every language. 2
Journal Pre-proof For example, in Fig. 2, the word “funny ” may mean someone or something that makes you laugh or smile, such as “a funny story”. However, "funny" can also imply something suspicious or illegal, such as “I suspect there may be something funny going on.” This double use of language, which gives the same word a positive and negative meaning, is not an accident. The sentiment polarity of a word fluctuates with changing context. Thus, it is unreasonable to express the emotional information of a word with a real number, which not only leads to the loss of information but also the fuzziness and uncertainty in human language is ignored. In
pro of
machine learning, the “bag of words” (BOW) model and word embedding model (like word2vec) ignore the issues of semantic change. Short texts lack contextual information and contain typos that pose challenges to machine learning approaches. Therefore, it is probably better to use a set of possible sentiment polarities to
urn a
lP
re-
describe a word when we do sentiment analysis.
Fig. 2. Different uses of the word “funny”
Similarly, there may be various sentiment polarities in a text review. Simply dividing reviews into negative
Jo
or positive reviews is only approximately 70% accurate [11]. Each individual text review may include active, neutral and passive information [12]. For instance, a consumer may comment on a restaurant where the food tastes only so-so, the price is suitable but service is rude. Neutral, positive, and negative, respectively, three kinds of sentiment information are coming together in a text review. The consumer may express that he is 20% sure that an item is good, 30% sure that the item is neither better nor worse, and 50% sure that the item is awful. Then, the value of the linguistic variable can be denoted:
3
Journal Pre-proof To fully describe and quantify this information, it is necessary to consider both possible sentiment polarity and the associated probabilistic information. This information can be interpreted as the probabilistic distribution [13, 14], importance [15], degree of belief [16] and so on. Ignoring this information leads to inaccurate analysis and may cause erroneous subsequent decisions. Thus, it could be better to calculate the probabilistic distribution of all levels of sentiment polarity when
pro of
we do research on sentiment analysis for short texts. In our research, we introduce PLTS and the relevant theory into short texts sentiment analysis and put forward a novel text representation model named Word2PLTS. Then, a novel framework named sentiment analysis and polarity classification (SAPCP) is proposed. In this framework, the sentiment information of a word can be expressed as a PLTS that is used to depict the fuzziness and uncertainty of the word. By aggregating the PLTSs of words, the sentiment information of a sentence can be obtained. Last, support vector machines (SVMs) are used to classify the text. This process can be roughly divided into three steps: pre-processing, text representation and classification. In
re-
addition, we evaluate the performance of our framework on three different datasets: movie reviews, Stanford twitter sentiment and TripAdvisor reviews. The experimental results show the validity and reliability of our proposed framework.
lP
The rest of this study is organized as follows. The related work is presented in section 2. Section 3 elaborates the sentiment analysis framework. In section 4, we specifically introduce the dataset that is used in this research. The results, evaluation and discussion are presented in section 5. Finally, section 6 concludes
2. Related work
urn a
this study and provides certain suggestions for future research directions.
In this section, several sentiment analysis techniques and certain basic concepts related to PLTS theory are be described.
2.1 Sentiment analysis
Sentiment analysis, also referred to as opinion mining, is a popular research topic in the field of NLP.
Jo
Sentiment analysis is used to analyse the writer's opinions, valuations, attitudes, and emotions towards a particular thing. All the approaches can be divided into two groups: lexicon-based approaches and machine learning approaches.
2.1.1 Lexicon-based approaches There are a large number of sentiment analysis methods that rely heavily on sentiment lexicons and related tools. We briefly provide an overview of five widely used techniques: a manually sentient lexicon, two sentiment analysis tools based on sentiment lexicons and two sentiment lexicons based on supervised 4
Journal Pre-proof learning. The General Inquiry1 (GI) [17] is one of the oldest manually sentient lexicons but is still popular. This lexicon was built in 1966 and has been improved constantly since; GI has been widely used in the sociological, psychological, economic, and anthropological fields [18]. This lexicon is composed of approximately 11789 words, and each word is labelled with one or more labels, such as negative, positive,
pro of
and pain. In this research, we only focused on the words labelled negative or positive. However, GI ignores the sentiment intensity. Thus, this lexicon is inappropriate for fine-grained sentient analysis. We downloaded the GI dictionary from this web page2.
TextBlob3 (TB) [19] is an open resource Python library for processing textual data that stands on the giant shoulders of NLTK4 and pattern5. Sentiment analysis is one of the most basic functions of TextBlob. TB has a lexicon of adjectives (e.g., awful, funny, worst, and badly) that occur frequently in reviews. By using regular expressions and this lexicon, TextBlob calculates the polarity intensity of given text. The polarity
re-
scores of words are improved based on the frequency of the adjectives and successive words. The polarity score is a real number within the range [−1.0, 1.0], where −1.0 is the most negative and 1.0 is the most positive.
lP
Valence Aware Dictionary and sEntiment Reasoner (VADER)6 [20] is rule-based sentiment analysis tool. This tool is specifically attuned to sentiments expressed in social media and works well on texts from other domains. VADER collected intensity ratings on common sentiment words (including adjectives, nouns, and
urn a
adverbs) from 10 independent human ratters. The sentient intensity is given in a range of
, which is
extreme negative to extreme positive. Compared with other well-established lexicons, VADER incorporates numerous lexical features common to sentiment expression in social media. A full list of Western-style emoticons, sentiment-related acronyms and commonly used slang with sentiment values are included in this lexicon. We accessed the VADER polarity score by access the online VADER API and a publicly available Python package.
Jo
SentiWordNet7 (SWN) [21, 22] is a lexical resource for opinion mining based on WordNet8. More than
1. http://www.wjh.harvard.edu/~inquirer 2. http://www.wjh.harvard.edu/~inquirer/spreadsheet_guide.htm 3. https://textblob.readthedocs.io/en/dev/ 4. http://www.nltk.org 5. https://www.clips.uantwerpen.be/pages/pattern-en 6 https://pypi.org/project/vaderSentiment/ 7. http://sentiwordnet.isti.cnr.it 5
Journal Pre-proof 140000 words are included in this lexicon, and each word has one or more synsets. SWN assigns two sentiment scores to each synset: a positive score (PosScore) and negative score (NegScore). The scores range from 0 to 1 and are calculated using a complex mix of semisupervised algorithms. A sample is given in Table 1. The POS tag can be ‘a’, ‘v’, ‘r’ or ‘n’, representing adjectives, verbs, adverbs and nouns, respectively. Although SWN contains many words, the performance level reported is not impressive. The SWN lexicon is
NLTK. Table 1. SentiWordNet 3.0 sample
pro of
very noisy; a large majority of synsets have no positive or negative polarity. We interfaced with SWN via
Synsets
PosScore
NegScore
Happy
happy.a.01
0.875
0
felicitous .s.02
0.75
glad.s.02
0.50
re-
Word
happy.s.04
0.125
0 0 0
SenticNet (SCN)9 is a public resource for emotionally aware intelligence [23]. There are more than 14000
lP
common sense concepts in the SCN lexicon, such as wrath, adoration, woe, and admiration. In the SCN lexicon, sentiment information is expressed in terms of four affective dimensions (pleasantness, attention, sensitivity, and aptitude) and the sentiment polarity of the concept is a numeric value ranging from
.
Instead of simply relying on counting word cooccurrence frequencies, SCN is constructed by the
urn a
bag-of-concepts model, a paradigm that exploits both AI and Semantic Web techniques. SCN 5 is available by calling to the Python library.
Almost all lexicons except SWN ignore the problem of polysemy. Although SWN provides a sentiment score for each synset, this lexicon does not assign a probability of occurrence for each synset. In certain studies [24, 25] based on SWN, the average sentiment score of the synsets is regarded as the sentiment score of a word. There is no doubt that information loss and deviation exist.
Jo
2.1.2 Machine learning approaches
It takes a large amount of time and labour to manually build a comprehensive sentiment lexicon and continually improve the lexicon. Many researchers have explored other methods of automatic identification of sentiment features [26]. Machine learning is technique that can “learn” the sentiment features of text. This
8. https://wordnet.princeton.edu 9 http://www.sentic.net/ 6
Journal Pre-proof method applies a training dataset to learn the model by employing a classifier that is used to the test dataset. There are a variety of classifiers; we cover several classifiers that are used in our experiments. The naive Bayes (NB) classifier [27] is a simple classifier that is extensively used and usually achieves good results. This classifier relies on Bayesian probability and assumes that the feature probabilities are independent of each other. Logistic regression (LR) [28] is a general machine learning technique that can be regard as a case
pro of
of the generalized linear model. By employing logistic functions, LR always gains good classification results. Unlike NB and LR, SVM [29] is a nonprobability classifier. This classifier constructs a hyperplane in a high-dimensional space, separates datapoints, then classifies. Pang[30] compared three machine learning approaches (NB, SVM and maximum entropy classification) on sentiment classification performance. SWM tends to do a little better than the other approaches.
Scikit-learn is a Python library that provides various machine learning algorithms; we access this library for the NB, LR and SVM classification models. However, there are certain problems with machine learning
re-
approaches. First, these approaches require a large amount of training data, and sometimes, the data are not easy to obtain. Machine learning requires considerable amounts of time and resources to manually label the dataset. Second, data sparsity is a major concern when the quantity of labelled data is insufficient to train a
lP
machine learning classifier. Third, when classification errors occur, the errors are difficult to explain and modify. 2.2 PLTS
urn a
A PLTS is a great mathematic tool to depict the fuzziness and uncertainty of language. Pang et al.[31] were the first to provide definitions and certain operations for PLTSs. However, there are still certain problems with the arithmetic operation rules of a PLTS. The operated values could go beyond the bounds of the given linguistic term set, which may lead to the loss of linguistic information in the results. To solve this problem, Gou et al. [32, 33] soon redefined more logical operations for PLTSs that have been widely used in later studies. The basic theory, operations and operators of PLTSs are provided in Appendix A.
Jo
PLTSs have caused widespread concern for scholars, and many studies based on PLTSs have emerged. Zhang et al. [34] applied a PLTS to assess investment risk and proposed a novel concept named probabilistic linguistic preference relation (PLPR). Peng et al. [35] proposed a cloud decision support model named probabilistic linguistic integrated cloud (PLIG). The novel model is used for hotel selection on TripAdvisor.com. Regarding problems with multicriteria decision making, Liao et al. [36] presented the probabilistic linguistic linear programming method to evaluate the hospital levels in China. Tian et al. [37] 7
Journal Pre-proof presented a multicriteria decision-making method based on PLTS and evidential reasoning. Luo et al.[38] used PLTS to evaluate the sustainability of constructed wetlands. Krishankumar [39] proposed a probabilistic linguistic preference relation (PLPR)-based decision framework. There are many studies about the application of PLTS in group decision making (GMD). Wu et al. [40] proposed a comprehensive multiple criteria group decision making method (MCGDMC) with a PLTS based on consensus measures and outranking methods.
pro of
Mao [41] applied an MCGDMC with a PLTS to finance. Nie [42] proposed a GDM support model integrated with prospect theory and PLTS.
An effectual approach to solving uncertain problems is the application of a PLTS to group decision making. A PLTS can comprehensively depict sentimental information, uncertainty and fuzziness in detail. In this study, the sentiment information of each word can be translated into a PLTS, as , where If
, greater values of
is the number of levels and
represents a sentiment polarity level.
indicate more positive sentiment polarity rankings. If
re-
indicate more negative sentiment polarity rankings; this way, the subset
.
, lower values of
is the probability of
. In
can quantify the negative meaning, the subset
can quantify the neutral meaning, and the subset
represented by a PLTS
lP
quantify the positive meaning. For example, supposing
can
, the sentiment information of “funny” may be . This representation means that the word
conveys positive information of level 3 at 70% probability and negative information of level 1 and level 3 at
urn a
20% and 10%, respectively. We can assign sentiment scores and probabilities to each meaning of a word and form a set. Then, the problem of polysemy can be readily solved. Obviously, for larger values of
, the
sentiment information can be depicted more adequately. In this way, PLTS depicts the sentiment information of word to the greatest extent.
3. Application of probabilistic linguistic terms in sentiment analysis As illustrated in section 1, certain deficiencies exist in prior studies; to account for these shortcomings, a
Jo
novel sentiment analysis classification framework is proposed. The framework is divided into three process, Preprocessing, Word2PLTS and Classification. The flowchart of this framework is shown in Fig. 3.
Fig. 3. Flowchart
8
Journal Pre-proof 3.1 Preprocessing To enhance data quality and perform better analysis, preprocessing is necessary. Preprocessing includes spelling correction, negation word checking, part of speech tagging, and stop word removal. 3.1.1 Spelling correction Because spelling errors in online reviews are very common, spelling correction is necessary. The spelling correction is performed by using TextBlob10, which is based on Peter Norvig’s “How to Write a Spelling
pro of
Corrector”11 as implemented in the pattern library. For example, the sentence "I havv goood speling!" can be corrected to “ I have good spelling!” . 3.1.2 Negation words checking
When negation words appear in the text, the sentiment polarity of the text may shift or reverse. When “not” or “no” negation words appear in the text, such as “not happy” and “no success.” “Happy” and “success" are positive words. If these words are modified by "not" and "no", the sentiment polarity of these words is
negative coefficient
is shown below:
re-
reversed. When there are multiple negation words in a paragraph, the polarity is reversed multiple times. The
-
is the number of negation words. If
sentence is reversed, and if
, the sentence is negative, and the sentiment polarity of
, the sentiment polarity remains unchanged.
lP
where
(1)
3.1.3 Part of speech (POS) tagging
POS tagging, also called grammatical tagging, is the process of tagging words as nouns, verbs, adjectives,
urn a
adverbs, etc. The Stanford POS tagger12 was employed in our study. The tagger uses Penn Treebank POS tags, which are converted into tags for verbs, nouns, adjectives and adverbs. The details are shown in Table 2. Only verbs, nouns, adjectives and adverbs are tagged in this research while the rest of the possibilities, such as prepositions and numerals, are ignored. Table 2. POS tags conversion table
Verb
Penn Tag
Jo
POS
VB (verb, base form) VBD (verb, past tense) VBG (verb, gerund or present participle)
10. https://textblob.readthedocs.io/en/dev/ 11. http://norvig.com/spell-correct.htmachine learning
9
Journal Pre-proof VBN (verb, past participle) VBP (verb, non-3rd person singular present) VBZ (verb, 3rd person singular present) Noun
NN (noun, singular or mass) NNP (proper noun, singular)
NNS(noun, plural) Adjective
JJ (adjective)
pro of
NNPS (proper noun, plural)
JJR (adjective, comparative) JJS (adjective, superlative) Adverb
RB (adverb)
RBR (adverb, comparative)
Other
Other
3.1.4 Stop word removing
re-
RBS (adverb, superlative)
lP
In a language, stop words refer to the most common words that have no special meaning, such as ‘it’, ’is’, and ‘the’. These words are not useful for text sentiment analysis. To reduce the calculations, storage space and influence of disturbance, a stop word list13 is used to filter these words. 3.2 Word2PLTS: A novel text representation model based on PLTS
urn a
As illustrated in section 2.2, a PLTS can fully depict the sentiment information of each meaning of a word and effectively solve the problem of uncertainty and fuzziness. In this phase, the novel text representation model based on PLTS is introduced in detail. In this model, every word can be converted to a PLTS, and we called the model “Word2PLTS”. The processes of this model can be divided into three steps: transformation, modification and calculation.
Jo
3.2.1 Transformation
In this section, a simple transformation between a word and the PLTS is provided, and we named this model “Word2PLTS”. We use the difference between the positive and negative scores of each synset from SWN as the sentiment valence of a word to distinguish differences in the sentiment intensities of words. The SynsetScore can be calculated as:
12. https://www.ranks.nl/stopwords 10
Journal Pre-proof (2) There are one or more synsets for each word, so we can obtain a one-dimensional array for the SynsetScore (DAOS) (3) is the number of synsets for a word.
Next, a function
is defined that can map SynsetScore into
:
pro of
where
In this way, we transform SynsetScore into
, the subscript of the linguistic term
(4) , and the DAOS
can be converted to an one-dimensional array of the linguistic term (DAOL).
Now, we count the occurrence frequency of each linguistic term
and roughly calculate the probabilities
. The sentiment polarity of a word can be converted to a PLTS:
(6)
re-
of each term
(5)
3.2.2 Modification
In the previous section, the procedures for extracting the sentiment information of a word from SWN and
lP
converting the information to a PLTS are introduced. However, these procedures are still insufficient. First, the PosScore and NegScore extracted from SWN are not 100% accurate. The scores were calculated using a complex mix of semisupervised algorithms (propagation methods and classifiers). This calculation inevitably contains many errors. For instance, the sentiment polarity scores of “absurd” are
obtained:
in
Table
4,
and
urn a
shown
the
PLTS
of
“absurd”
can
be
, which implies that “absurd” conveys a positive
intensity. However, “absurd” is a derogatory word. Table 4. Sentiment information of “absurd” from SWN Word
PosScore
NegScore
absurd.n.01
0
0
absurd.s.01
0.3750
0
absurd.s.02
0.6250
0
Jo
absurd
Synsets
Second, polysemy is a pervasive phenomenon in human language. However, the probabilities for the appearance of each meaning in the text are not equal. For example, there are ten meanings of "fine" in the SWN; certain meanings may be common, and certain meanings are rare. In SWN, only the sentiment polarity 11
Journal Pre-proof of each synset is provided but does not include the probability information. Last, the sentiment polarities of the words are sometimes determined by context. In different contexts, the sentiment polarity is transferred or strengthened. The same word may have the opposite polarity in different domains. We should fully take context into account rather than only extracting sentiment information from SWN.
pro of
In consideration of these problems that may lead to underfitting, it is necessary modify the PLTS from SWN. We used a training set to build a simple dictionary though word frequency distribution statistics. A novel concept named the context sentiment score (CSS) is introduced, and CSS is calculated by the equation: -
where
is a word,
frequency of
is the frequency of
(7)
in text labelled positive, and
in text labelled negative.
is the
. For example,
re-
means that the word ‘bad’ only appears in text labelled negative. Similarly, by employing Eq. (4), we map
from [-1,1] to a linguistic term subscript
and create a word frequency distribution information dictionary. We named the proposed dictionary CSS-V. Samples from CSS-V are shown in Table 5.
Term
lP
Table 5. Samples from the proposed CSS-V vocabulary Linguistic term subscript -2
nothing
-3
work unfaithful
Linguistic term subscript
usually
1
dull
-6
0
delightful
6
-2
wise
3
urn a
chicken
Term
Finally, by taking advantage of CSS-V, the PLTS extracted from SWN is modified, and the algorithm for
where
Jo
the modification is defined in the following:
is a word,
is queried from CSS-V.
(8)
is the PLTS of the word,
is the linguistic term of word and the subscript
can be considered as a modifying factor, and
modification is not conducted. When causes overfitting. In our study,
. When
, the
is larger, the degree of modification is higher but more easily was used to modify the PLTS from SWN.
3.2.3 Calculation A sentence can be considered as a sequence of words; this sequence implies a semantic sequence that 12
Journal Pre-proof decides the sentiment information of the sentence. Thus, by aggregating the sentiment information of words, the sentiment information of the sentence is available. Supposed, there are
words after preprocessing in a sentence, and
words can be converted into
PLTSs; thus, the operational laws of PLTSs are overly complex. To reduce computational complexity, neutral words are removed. The criterion for determining whether a word is neutral is defined as follows:
where
is the probability of
pro of
(9) . If a PLTS is neutral, that PLTS is removed.
After neutral word filtering, there are
PLTSs left. The sentiment polarity of the sentence can be
calculated by employing a probability linguistic averaging (PLA) operator.
is the PLTS of the sentence and ; otherwise,
.
Usually, there is a large number of
in
is a negative coefficient; if the sentence is negative,
re-
where
(10)
and little difference between adjacent
. It is
necessary to decrease computational complexity, reduce storage consumption and normalize storage schemas. to
and merge
lP
We map example, mapped into
. For
can
be
. that describes the possibilities for the
urn a
In Word2PLTS, a text-like sentence is represented as a
that have identical subscripts
sentiment polarity. This processing makes full use of the sentiment information from the sentiment lexicon and word frequency distribution information. 3.3 Classification
It may be difficult to judge the sentiment polarity of a sentence only though are still many elements in
because there
. To improve the classification performance, we make use of a
Jo
supervised learning algorithm, SVM, that is known to work best according to many researchers, and the radial basis function kernel is employed. In this classification model, linguistic terms corresponding probability
are regarded as features and feature values, respectively.
and the is then
converted into a format that is understandable for the SVM classifier. This format is as given below and in Fig. 4 :
13
Journal Pre-proof where is
and -1 for positive and negative, respectively, is the linguistic term
of the linguistic term.
pro of
is the probability
, and
Fig. 4. Input format for SVM 3.4 Example for SACPC
For easier understanding, in this section, we provide a step-by-step example of how to use SACPC for text sentiment analysis.
re-
Suppose that the SVM classifier is trained using the training set. Then, a model is generated that can be applied to classification. For example, there is a movie review that needs to be classified. “In its chicken heart, crush goes to absurd lengths to duck the very issues it raises.” The detailed procedure is given as follows, and visualization processing is provided in Appendix B:
tagging, are performed.
lP
(1) First, the preprocessing steps, including application of spelling correction, word segment and POS
('in', 'IN'), ('its', 'PRP$'), ('chicken', 'JJ'), ('heart', 'NN'), ('crush', 'NN'), ('goes', 'VBZ'), ('to', 'TO'),
urn a
('absurd', 'VB'), ('lengths', 'NNS'), ('to', 'TO'), ('duck', 'VB'), ('the', 'DT'), ('very', 'RB'), ('issues', 'NNS'), ('it', 'PRP'), ('raises', 'VBZ') (2) Second, stop word removal is applied, and only words with tags of verb, noun, adjective and adverb are kept.
('chicken', 'JJ'), ('heart', 'NN'), ('crush', 'NN'), ('goes', 'VBZ'), ('absurd', 'VB'), ('lengths', 'NNS'),
Jo
('duck', 'VB'), ('issues', 'NNS'), ('raises', 'VBZ') (3) Next, we extract the sentiment polarity score of the synsets of each word and convert the scores to PLTSs. For clarity, we take “funny” as an example to describe the extraction and conversion process. (3.1) First, we obtain the synsets of the word and corresponding sentiment intensity by access to SWN. The synsets for "funny" are shown in Table 3. Table 3. Sentiment information of “funny” from SWN Word
Synsets
PosScore 14
NegScore
Journal Pre-proof funny
funny_story.n.01
0
0
amusing.s.02
0.5
0
curious.s.01
0.125
0.375
fishy.s.02
0
0.5
funny.s.04
0
0.5
(3.3) In this step, using mapping function
pro of
(3.2) Next, by employing Eq. (3), we calculate the SynsetScore, and a DAOS can be obtained.
, the SynsetScore is transformed into
and we obtain the DAOL.
,
re-
(3.4) Finally, the transformation between “funny” and the PLTS is achieved.
Similarly, the PLTSs of word in this sentence can be obtained. chicken:
lP
heart: crush:
absurd: lengths: issues: raises:
urn a
goes:
,
(4) In this step, by referring to CSS-V, the PLTSs of the words are modified. For instance, the PLTS of “funny” extracted from SWN is
Jo
. Then, by
querying CSS-V, we can obtain the linguistic term subscript (8), the modification is conducted:
Last, by employing Eq. .
Similarly, the modification of PLTSs of word in this sentence are conducted. chicken:
S-8
heart:
15
Journal Pre-proof ,
crush: goes: absurd: lengths: issues:
pro of
raises: (5) To reduce the computational complexity, we eliminate the neutral words, and only 4 words are left. chicken: heart: crush: absurd:
to
and transform
into a format
urn a
(7) Now, we map
.
lP
calculated, where the negative coefficient
, is
re-
(6) At this step, by employing the PLA operator, the sentiment polarity of sentence,
.
that can be understood by machine learning. “33:0.0006 …34:0.0053 … 69:0.0015” (8) We use the model selection approach to classify; the review is then classified as positive or negative. 4. Datasets
In this study, we applied SAPCP to different short text corpora from three different domains.
Jo
Movies reviews (MR): This dataset was first used by Bo Pang and Lillian Lee14 [43]. There is a total of 10662 snippets that contains roughly one single sentence. Half the snippets are labelled positive, and half the snippets are labelled negative. The snippets are collected from Rotten Tomatoes webpages, a website in the United States that provides film related comments. These researchers assumed that snippets (from the Rotten Tomatoes webpages) from reviews marked ‘fresh’ are positive, and those from reviews marked ‘rotten’ are
13 http://www.cs.cornell.edu/people/pabo/movie-review-data/ 16
Journal Pre-proof negative. A total of 80% of all the data were randomly selected as training data, and the remaining 20% were testing data. The average length of snippets is 20. Stanford Twitter Sentiment (STS): The Stanford Twitter Sentiment15 corpus [44] is our second labelled corpus. The training set consists of 1.6 million tweets that were automatically labelled positive/negative. To reduce training time, we randomly selected 100k tweets as training data. The test set contains 182 positive
pro of
and 177 negative sentiment tweets with manual annotation. The average length of tweets is 19 words. TripAdvisor reviews (TR): TripAdvisor.com is the largest and most successful social networking and community site for tourism in the world. We crawled more than 10000 text reviews from this web involving 60 restaurants in New York and 1945 customers. We spilt the reviews into sentence-level snippets and used manual annotation. A total of 900 positive and 900 negative sentence-level snippets are available16. We randomly selected 720 positive and 720 negative snippets as training data and the remainder of the snippets were testing data. The average length of the reviews is 22 words.
re-
5. Results and discussion
As the most common evaluation indicators, accuracy (Acc), precision (Prec), recall (Rec), and F-measure
urn a
lP
(F-M) were used to comprehensively evaluate the classification model. (11) (12) (13) (14)
where TP, TN, FP, and FN represent true positives, true negatives, false positives, and false negatives, respectively. Accuracy is an important indicator for classification but only works when the samples are balanced. Precision measures how accurate your model predictions are, and recall measures how well your
Jo
model finds all the positives. However, precision and recall are mutually restrictive and one-sided. The F-measure is the harmonic mean of the precision and recall; this parameter is a comprehensive evaluation index. For the four indicators, the higher the index values are, the better the classification of the method is. First, we compared the performances of five well-established sentiment lexicons on the datasets: General Inquirer (GI), TextBlob (TB), VADER, SenticNet(SCN), and SentiWordNet (SWN). Detailed results for the
14 http://cs.stanford.edu/people/alecmgo/trainingandtestdata.zip 15 https://tianchi.aliyun.com/dataset/dataDetail?dataId=5177, I have uploaded the data to aliyun for everyone to download 17
Journal Pre-proof comparison are shown in Table 6. VADER achieves the best performance compared with the other lexicons, and TB is slightly inferior to VADER. There is little difference between SCN and SWN. GI tends to perform the worst, but the differences are not very large. The average precision, recall and F-measure for the five sentiment analysis tools/techniques on the movie reviews are 58.50%, 69.39% and 63.36%, respectively, on Stanford Twitter Sentiment are 62.36%, 57.92% and 59.77%, respectively, and on TripAdvisor reviews are
pro of
63.99%, 78.38% and 70.41%, respectively. The recall is higher than the precision for the lexicons, which means that the lexicon-based method is more sensitive to false positives than false negative. The highest values for F-M for MR, STS, TR are 68.90%, 66.50% and 77.67%, respectively, which shows that the performances of approaches based on lexicons are far from satisfactory.
Table 6. Detailed results for several sentiment analysis tools/techniques MR
STS
TR
Rec
F-M
Prec
Rec
F-M
Prec
Rec
F-M
GI
0.5542
0.6314
0.5902
0.5674
0.5969
0.5818
0.5892
0.6524
0.6192
TB
0.6052
0.7184
0.6569
0.6683
0.5623
0.6107
0.6896
0.8667
0.7680
VADER
0.6419
0.7551
0.6890
0.7275
0.6124
0.6650
0.6897
0.8888
0.7767
SCN
0.5563
0.6997
0.6198
0.5071
0.5902
0.5455
0.5992
0.7511
0.6666
SWN
0.5674
0.6648
0.6122
0.5854
0.6316
0.7600
0.6899
lP
re-
Prec
0.5341
Jo
urn a
0.6475
Fig. 5. Comparison of results averaged over the evaluation datasets Second, we compared SAPCP to certain machine learning approaches, such as NB, SVM, and LR. We trained the machine learning classifiers based on “bag of words" (BOW), a text representation model [45]. 18
Journal Pre-proof The detailed results are given in Table 7. By comparing the results of Table 6 and Table 7, the highest values of F-M for MR, STS and TR are 76%, 72% and 87%, respectively. The result show that methods based machine learning perform much better than lexicon-based methods. Then, the mean values of the results of the machine learning methods as the baseline and the improvement of SAPCP were calculated, which are given in Table 8. As a technique combines that unsupervised and supervised learning, SAPCP
pro of
achieves the highest performance improvement for the STS. These results show that as the coverage of the training data sets increases, the accuracy and F-measure improve. Smaller improvements are obtained in recall (6.5%) than in precision (9.4%). This observation leads to that our model is more sensitive to false negatives. It is possible that in TripAdvisor reviews people usually focus on the quality of food and hotels and tend to use similar words to convey emotion. However, in movie reviews and tweets, the contents that are expressed are more extensive and diffuse. Many obscure words are used to express feelings. In this case, it is not easy to extract effective features for machine learning approaches, and thus, the BOW model should
re-
be used because many word words only appear in the test set. However, SAPCP can effectively solve these problems by employing a lexicon and word frequency distribution information. Table 7. Detailed results of comparison between machine learning and SAPCP STS
Rec
F-M
Bow+NB
0.75
0.78
0.76
Bow+SVM
0.75
0.76
0.75
Bow+LR
0.76
SAPCP
0.87
TR
Prec
Rec
F-M
Prec
Rec
F-M
0.67
0.79
0.71
0.87
0.86
0.86
0.65
0.82
0.72
0.86
0.86
0.86
urn a
Prec
lP
MR
0.77
0.76
0.68
0.79
0.72
0.87
0.88
0.87
0.81
0.84
0.81
0.92
0.86
0.89
0.90
0.89
Table 8. SAPCP improvement over the baseline
STS TR Average
Rec
F-M
0.117
0.040
0.083
0.143
0.120
0.143
0.023
0.033
0.023
0.094
0.065
0.083
Jo
MR
Prec
Finally, many researchers have used movie reviews and the Stanford Twitter Sentiment as a benchmark to evaluate sentiment classification methods. To verify effectiveness and superiority of SAPCP, we compared this method to other techniques. Detailed comparison results are presented in Table 9. The results are 19
Journal Pre-proof encouraging. For movie reviews, SAPCP outperforms other state of the art techniques. For the Stanford Twitter Sentiment, SAPCP has slightly poorer performance than certain approaches based on deep learning, possibly because movie reviews are usually written in a more standardized and orderly way than tweets. Tweets always contain a variety of incomplete expressions, such as acronyms, slang, ill-formed words and emoji, that contain sentiment information. It is not easy for a lexicon to recognize incomplete expressions and provide sentiment information, and only word frequency distribution information works.
pro of
Table 9. Comparison with state of the art approaches for movie reviews and Stanford Twitter Sentiment Movie reviews Author/Year
Stanford Twitter Sentiment Method
Socher et al. Semi-Supervised (2011) [46]
Acc
Author/Year
77.7%
Speriosu et al. LProp
Recursive
et
Fuzzy Sets al CNN
(2014) [50]
(2014) [52]
81.52%
Socher et al. RNTN
85.4%
77.01%
Dos Santos et CharSCNN
85.7%
al. (2014) [53]
al. KPCNN
(2017) [54] SAPCP
82.9%
(2013) [49]
Word2PLTS+SW M
83.25
Zhao
urn a
et
84.7%
(2013) [51]
Zhang et al. CharCNN
Wang
Socher et al. MV-RNN
lP
Kim
Rules, 76.0%
re-
(RAE)
(2016) [48]
Acc
(2011) [47]
Auto-encoders
Appel et al. Semantic
Method
et
al. GloVe-DCNN
87.62%
(2018) [55]
84.22%
SAPCP
Word2PLTS+SWM
85.22%
In general, SAPCP is an approach than perfectly combines both the unsupervised and supervised
Jo
techniques. First, we extract the sentiment information of words from an existing lexicon and word frequency distribution dictionary. Every word can be represented as a PLTSs. Then, by aggregating the PLTSs of words, the PLTS of a sentence
is available. Last, SWM is used to improve the
classification performance. Though the comparison of results, SAPCP displayed excellent performance, this method has steady classification performances for different datasets. Compared with unsupervised classifiers, SAPCP can solve the problems of domain dependence and contextual information by using word frequency distribution information and introduction of SVM. Compared with supervised classifiers, SAPCP can extract 20
Journal Pre-proof the sentiment information of a word from an existing lexicon, which may not be easily obtained by only relying on data-training. This method can efficiently solve the problems of data unavailability and data sparsity. More importantly, by introducing PLTS and the relevant theory, we effectively solved the problem of polysemy. 6. Conclusions and future study
pro of
In this study, we proposed a novel text representation model named Word2PLTS. Then, by using SVMs, a short text sentiment analysis framework named SAPCP is presented. In the framework, we extract the sentiment information of words from SWN and a word frequency distribution dictionary. The information can be converted into PLTSs. By aggregating the PLTS of the words, the sentiment information of short text is available. Finally, an SVM algorithm is applied to improve classification performance. A noticeable performance improvement is achieved by evaluating SAPCP. In conclusion, several advantages of our work are summarized in the following.
re-
(1) The novel text representation model, Word2PLTS, introduces the idea of fuzzy mathematics into sentiment analysis. This model fully considers the fuzziness and uncertainty of human language. This model is a good way to effectively solve the problem of the variation of lexical meaning and describe
lP
sentiment information for short text.
(2) SAPCP is a technique that combines supervised and unsupervised methods. By comparison with unsupervised methods, this technique takes context into account and achieves great improvement in
urn a
performance. Compared with supervised methods, this method can extract sentiment information from a lexicon, which is effective in solving the problems of data unavailability and data sparsity. There are also some suggestion for future research and the development of the proposed method. (1) The operational laws of PLTSs are overly complex, which is why SAPCP is only applicable to short text sentiment analysis and not long text. Simplifying the arithmetic operation rules is necessary and significant.
Jo
(2) SAPCP achieved a great success in binary (positive/negative) classification. This method may be suitable for fine-grained sentiment classification. The relevant research is meaningful and promising. (3) Recently, some flexible and brilliant linguistic approaches, such as the intuitionistic linguistic approach[56], grey linguistic approach[57] and interval linguistic approach [58], have been developed to deal with ambiguity in human language. Therefore, the linkage of these linguistic approaches to sentiment analysis is interesting. 21
Journal Pre-proof
Acknowledgements We are grateful to the editors and anonymous reviewers for their helpful comments and suggestions. This work was supported by the National Natural Science Foundation of China (Nos. 71571193 and 71871228). Appendix A
pro of
Some basic concepts related to PLTS are introduced, including properties and operators. Definition 1. Let
where
be a LTS, a PLTS can be defined as:
denotes the linguistic term is the number of the elements in
associated with the corresponding probability
.
Definition 2. Let
be an LTS, . ,
. where
are the equivalent transformation functions.
lP
g and
and
be three PLTSs,
is the subscript of linguistic term
re-
be a positive real number, and
(15)
urn a
Then, the sum is given by the following.
(16)
(17)
Jo
(18) (19)
In our study, the multiplication of negative numbers and PLTS is defined. (20)
Definition 3. Let
be
PLTSs, where
linguistic term and the corresponding probability, respectively, in 22
and
are the
. Then, we can obtain the
Journal Pre-proof probability linguistic averaging (PLA) operator.
(21) Definition 4. Let
be
PLTSs, where
is the weight vector of
,
are the
, where
pro of
linguistic term and the corresponding probability, respectively, in
and
and
. We
can obtain the probability linguistic weighted averaging (PLAW) operator.
urn a
lP
re-
, then the PLAW reduces to PLA.
Jo
When
23
(22)
Journal Pre-proof
Jo
urn a
lP
re-
pro of
Appendix B
24
Journal Pre-proof References [1] C.E. Osgood, G.J. Suci, P.H. Tannenbaum, The measurement of meaning, University of Illinois Press, 1957. [2] J.A. Morente-Molinera, G. Kou, C. Pang, F.J. Cabrerizo, E. Herrera-Viedma, An automatic procedure to create fuzzy ontologies from users’ opinions using sentiment analysis procedures and multi-granular fuzzy
pro of
linguistic modelling methods, Information Sciences, 476 (2019) 222-238. [3] P. Ji, H.-Y. Zhang, J.-Q. Wang, A fuzzy decision support model with sentiment analysis for items comparison in e-commerce: The case study of PConline. com, IEEE Transactions on Systems, Man, and Cybernetics: Systems, 49(10) (2019) 1993-2004.
[4] L. Wang, X. Wang, J. Peng, J. Wang, The differences in hotel selection among various types of travellers: A comparative analysis with a useful bounded rationality behavioural decision support model, Tourism Management, 76 (2020) 103961.
re-
[5] N. Godbole, M. Srinivasaiah, S. Skiena, Large-scale sentiment analysis for news and blogs, International Conference on Weblogs and Social Media,(2007).
[6] J. Li, S. Fong, Y. Zhuang, R. Khoury, Hierarchical classification in text mining for sentiment analysis of
lP
online news, Soft Computing - A Fusion of Foundations, Methodologies and Applications, 20 (2016) 3411-3420.
[7] K. Ioannis, N. Azadeh, S. Matthew, S. Abeed, A. Sophia, G.H. Gonzalez, Analysis of the effect of
urn a
sentiment analysis on extracting adverse drug reactions from tweets and forum posts, Journal of Biomedical Informatics, 62 (2016) 148-158.
[8] P. Grandin, J.M. Adan, Piegas: A Systems for Sentiment Analysis of Tweets in Portuguese, IEEE Latin America Transactions, 14 (2016) 3467-3473.
[9] J.A. Morente-Molinera, G. Kou, K. Samuylov, R. Ureña, E. Herrera-Viedma, Carrying out consensual Group Decision Making processes under social networks using sentiment analysis over comparative
Jo
expressions, Knowledge-Based Systems, 165 (2019) 335-345. [10] F. Ali, D. Kwak, P. Khan, S. El-Sappagh, A. Ali, S. Ullah, K.H. Kim, K.-S. Kwak, Transportation sentiment analysis using word embedding and ontology-based topic modeling, Knowledge-Based Systems, 174 (2019) 27-42.
[11] M. Schuckert, X. Liu, R. Law, Hospitality and Tourism Online Reviews: Recent Trends and Future Directions, Journal of Travel & Tourism Marketing, 32 (2015) 608-621. 25
Journal Pre-proof [12] S.O. Proksch, W. Lowe, J. Wäckerle, S. Soroka, Multilingual sentiment analysis: A new approach to measuring conflict in legislative speeches, Legislative Studies Quarterly, 44 (2019) 97-131. [13] Y. Dong, Y. Wu, H. Zhang, G. Zhang, Multi-granular unbalanced linguistic distribution assessments with interval symbolic proportions, Knowledge-Based Systems, 82 (2015) 139-151. [14] Z. Wu, J. Xu, Possibility Distribution-Based Approach for MAGDM With Hesitant Fuzzy Linguistic
pro of
Information, IEEE Transactions on Cybernetics, 46 (2016) 694-705. [15] H. Liu, A fuzzy envelope for hesitant fuzzy linguistic term set and its application to multicriteria decision making, Information Sciences, 258 (2014) 220-238.
[16] J.B. Yang, Rule and utility based evidential reasoning approach for multiattribute decision analysis under uncertainties, European Journal of Operational Research, 131 (2001) 31-61.
[17] P.J. Stone, D.C. Dunphy, M.S. Smith, D.M. Ogilvie, General Inquirer, Philosophy, (1966). [18] P.D. Turney, M.L. Littman, Measuring praise and criticism:Inference of semantic orientation from
re-
association, Acm Transactions on Information Systems, 21 (2003) 315-346.
[19] L. Subirats, N. Reguera, A. Bañón, B. Gómez-Zúñiga, J. Minguillón, M. Armayones, Mining Facebook Data of People with Rare Diseases: A Content-Based and Temporal Analysis, International Journal of
lP
Environmental Research and Public Health, 15 (2018) 1877.
[20] C.J. Hutto, E. Gilbert, VADER: A Parsimonious Rule-Based Model for Sentiment Analysis of Social Media Text, Eighth international AAAI conference on weblogs and social media, (2014).
urn a
[21] A. Esuli, F. Sebastiani, SentiWordNet: A Publicly Available Lexical Resource for Opinion Mining,(2006). 417--422.
[22] S. Baccianella, A. Esuli, F. Sebastiani, SentiWordNet 3.0: An Enhanced Lexical Resource for Sentiment Analysis and Opinion Mining, International Conference on Language Resources and Evaluation, Lrec 2010, 17-23 May 2010, Valletta, Malta,(2010). 83-90. [23] E. Cambria, R. Speer, C. Havasi, A. Hussain, SenticNet: A publicly available semantic resource for
Jo
opinion mining, Aaai Csk, (2010).
[24] F.H. Khan, U. Qamar, S. Bashir, SentiMI: Introducing point-wise mutual information with SentiWordNet to improve sentiment polarity detection, Applied Soft Computing, 39 (2016) 140-153. [25] F.H. Khan, U. Qamar, S. Bashir, eSAP: A Decision Support Framework for Enhanced Sentiment Analysis and Polarity Classification, Information Sciences, 367-368 (2016) 862-873. [26] X. Xie, S. Ge, F. Hu, M. Xie, N. Jiang, An improved algorithm for sentiment analysis based on 26
Journal Pre-proof maximum entropy, Soft Computing, 23 (2019) 599-611. [27] H. Turtle, W.B. Croft, Inference networks for document retrieval,(1989). 1-24. [28] S.H. Walker, D.B. Duncan, Estimation of the probability of an event as a function of several independent variables, Biometrika, 54 (1967) 167-179. [29] V. Vapnik, SVM method of estimating density, conditional probability, and conditional density, IEEE
pro of
International Symposium on Circuits and Systems, 2000. Proceedings. ISCAS,(2000). 749-752 [30] T.B. Pang, B. Pang, L. Lee, Thumbs up? Sentiment Classification using Machine Learning, Empirical Methods in Natural Language Processing, (2002) 79-86.
[31] Q. Pang, H. Wang, Z. Xu, Probabilistic Linguistic Term Sets in Multi-Attribute Group Decision Making, Information Sciences, 369 (2016) 128-143.
[32] X. Gou, Z. Xu, Novel basic operational laws for linguistic terms, hesitant fuzzy linguistic term sets and probabilistic linguistic term sets, Information Sciences, 372 (2016) 407-427.
re-
[33] X. Gou, H. Liao, Z. Xu, F. Herrera, Double Hierarchy Hesitant Fuzzy Linguistic Term Set and MULTIMOORA Method: A Case of Study to Evaluate the Implementation Status of Haze Controlling Measures, Information Fusion, 38 (2017) 22-34.
lP
[34] Y. Zhang, Z. Xu, H. Wang, H. Liao, Consistency-based risk assessment with probabilistic linguistic preference relation, Applied Soft Computing, 49 (2016) 817-833. [35] H.G. Peng, H.Y. Zhang, J.Q. Wang, Cloud decision support model for selecting hotels on
68 (2018) 124-138.
urn a
TripAdvisor.com with probabilistic linguistic information, International Journal of Hospitality Management,
[36] H. Liao, L. Jiang, Z. Xu, J. Xu, F. Herrera, A linear programming method for multiple criteria decision making with probabilistic linguistic information, Information Sciences, 415 (2017). [37] Z.-P. Tian, R.-X. Nie, J.-Q. Wang, Probabilistic linguistic multi-criteria decision-making based on evidential reasoning and combined ranking methods considering decision-makers’ psychological preferences,
Jo
Journal of the Operational Research Society, (2019) 1-18. [38] S.-z. Luo, H.-y. Zhang, J.-q. Wang, L. Li, Group decision-making approach for evaluating the sustainability of constructed wetlands with probabilistic linguistic preference relations, Journal of the Operational Research Society, 70 (2019) 2039-2055. [39] R. Krishankumar, K. Ravichandran, M. Ahmed, S. Kar, S. Tyagi, Probabilistic Linguistic Preference Relation-Based Decision Framework for Multi-Attribute Group Decision Making, Symmetry, 11 (2019) 2. 27
Journal Pre-proof [40] X. Wu, H. Liao, A consensus-based probabilistic linguistic gained and lost dominance score method, European Journal of Operational Research, 272 (2019) 1017-1027. [41] X.-B. Mao, M. Wu, J.-Y. Dong, S.-P. Wan, Z. Jin, A new method for probabilistic linguistic multi-attribute group decision making: Application to the selection of financial technologies, Applied Soft Computing, 77 (2019) 155-175.
pro of
[42] R.-x. Nie, J.-q. Wang, Prospect Theory-Based Consistency Recovery Strategies with Multiplicative Probabilistic Linguistic Preference Relations in Managing Group Decision Making, Arabian Journal for Science and Engineering, (2019) 1-18.
[43] L.L. Bo Pang Seeing stars Exploiting class relationships for sentiment categorization with, Proceedings of the 43rd Annual Meeting of the ACL, ACL (2005) 115-124.
[44] A. Go, R. Bhayani, L. Huang, Twitter sentiment classification using distant supervision, CS224N Project Report, Stanford, 1 (2009).
re-
[45] D. Larlus, J. Verbeek, F. Jurie, Category Level Object Segmentation by Combining Bag-of-Words Models with Dirichlet Processes and Random Fields, International Journal of Computer Vision, 88 (2010) 238-253.
lP
[46] R. Socher, J. Pennington, E.H. Huang, A.Y. Ng, C.D. Manning, Semi-supervised recursive autoencoders for predicting sentiment distributions, Conference on Empirical Methods in Natural Language Processing, EMNLP 2011, 27-31 July 2011, John Mcintyre Conference Centre, Edinburgh, Uk, A Meeting of
urn a
Sigdat, A Special Interest Group of the ACL,(2011). 151-161. [47] M. Speriosu, N. Sudan, S. Upadhyay, J. Baldridge, Twitter polarity classification with label propagation over lexical links and the follower graph, Association for Computational LinguisticsProceedings of the First workshop on Unsupervised Learning in NLP,(2011). 53-63. [48] O. Appel, F. Chiclana, J. Carter, H. Fujita, A hybrid approach to sentiment analysis, Evolutionary Computation,(2016). 242-254.
Jo
[49] R. Socher, A. Perelygin, J. Wu, J. Chuang, C.D. Manning, A. Ng, C. Potts, Recursive deep models for semantic compositionality over a sentiment treebank, Proceedings of the 2013 conference on empirical methods in natural language processing,(2013). 1631-1642. [50] Y. Kim, Convolutional neural networks for sentence classification, arXiv preprint arXiv:1408.5882, (2014). [51] R. Socher, J. Bauer, C.D. Manning, Parsing with compositional vector grammars, Proceedings of the 28
Journal Pre-proof 51st Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers),(2013). 455-465. [52] X. Zhang, J. Zhao, Y. LeCun, Character-level convolutional networks for text classification, Advances in neural information processing systems,(2015). 649-657. [53] C. Dos Santos, M. Gatti, Deep convolutional neural networks for sentiment analysis of short texts,
pro of
Proceedings of COLING 2014, the 25th International Conference on Computational Linguistics: Technical Papers,(2014). 69-78.
[54] J. Wang, Z. Wang, D. Zhang, J. Yan, Combining Knowledge with Deep Convolutional Neural Networks for Short Text Classification, IJCAI,(2017). 2915-2921.
[55] J. Zhao, X. Gui, Deep Convolution Neural Networks for Twitter Sentiment Analysis, IEEE Access, PP (2018) 1-1.
[56] J.Q. Wang, Z.Q. Han, H.Y. Zhang, Multi-criteria Group Decision-Making Method Based on
re-
Intuitionistic Interval Fuzzy Information, Group Decision & Negotiation, 23 (2014) 715-733. [57] Z.P. Tian, J. Wang, J.Q. Wang, X.H. Chen, Multicriteria decision‐ making approach based on gray linguistic weighted Bonferroni mean operator, International Transactions in Operational Research, (2015)
lP
1635-1658.
[58] J.Q. Wang, J.J. Peng, H.Y. Zhang, T. Liu, X.H. Chen, An Uncertain Linguistic Multi-criteria Group
Jo
urn a
Decision-Making Method Based on a Cloud Model, Group Decision & Negotiation, 24 (2015) 171-192.
29
Journal Pre-proof *conflict of Interest Statement
Declaration of interests
pro of
☒ The authors declare that they have no known competing financial interests or personal relationships that could have appeared to influence the work reported in this paper.
☐The authors declare the following financial interests/personal
urn a
lP
re-
relationships which may be considered as potential competing interests:
Jo
Chao Song, Xiao-Kang Wang, Peng-fei Cheng, Jian-qiang Wang, Lin Li