Sentiment Analysis Tutorial: Prof. Ronen Feldman Hebrew University, JERUSALEM Digital Trowel, Empire State Building
Sentiment Analysis Tutorial: Prof. Ronen Feldman Hebrew University, JERUSALEM Digital Trowel, Empire State Building
TUTORIAL
Prof. Ronen Feldman
Hebrew University, JERUSALEM
Digital Trowel, Empire State Building
[email protected]
The Text Mining Handbook
CACM Article
4
INTRODUCTION TO
SENTIMENT ANALYSIS
Based on slides from Bing Liu
and some of our work
5
Introduction
Sentiment analysis
Computational study of opinions, sentiments, evaluations, attitudes,
appraisal, affects, views, emotions, subjectivity, etc., expressed in
text.
7H[W 5HYLHZVEORJVGLVFXVVLRQVQHZVFRPPHQWVIHHGEDFN
LingPipe
9
Individuals
Make decisions to purchase products or to use services
Find public opinions about political candidates and issues
Roadmap
Sentiment Analysis Problem
Document sentiment classification
Sentence subjectivity & sentiment classification
Aspect-based sentiment analysis
Opinion summarization
Sentiment lexicon generation
Mining comparative opinions
Some complications
Generalized sentiment analysis
12
A (regular) opinion
An opinion has the following basic components
(gi, soijkl, hi, tl),
where
gj is a target
soijl is the sentiment value of the opinion from opinion holder hi on
target gj at time tl. soijl is positive, negative or neutral, or a
rating score
hi is an opinion holder.
tl is the time when the opinion is expressed.
17
Opinion target
In some cases, opinion target is a single entity or
topic.
,ORYHiPhoneand ,VXSSRUWtax cut
An opinion is a quintuple
(ej, ajk, soijkl, hi, tl),
where
ej is a target entity.
ajk is an aspect/feature of the entity ej.
soijkl is the sentiment value of the opinion from the opinion holder hi
on aspect ajk of entity ej at time tl. soijkl is +ve, -ve, or neu, or a more
granular rating.
hi is an opinion holder.
tl is the time when the opinion is expressed.
20
&RQIXVLQJWHUPLQRORJLHV
Entity is also called object.
Aspect is also called feature, attribute, facet, etc
Opinion holder is also called opinion source
23
5HDGHUVVWDQGLQJSRLQW
See this sentence
,DPVRKDSS\WKDW*RRJOHSULFHVKRWXSWRGD\
Although the sentence gives an explicit sentiment,
different readers may feel very differently.
If a reader sold his Google shares yesterday, he will not
be that happy.
If a reader bought a lot of Google shares yesterday, he
will be very happy.
Current research mostly ignores the issue.
24
Subjectivity
Sentence subjectivity: An objective sentence presents
some factual information, while a subjective sentence
expresses some personal opinions, beliefs, views,
feelings, or emotions.
Not the same as emotion
26
Subjectivity
Subjective expressions come in many forms, e.g.,
opinions, allegations, desires, beliefs, suspicions,
and speculations (Wiebe, 2000; Riloff et al 2005).
A subjective sentence may contain a positive or negative
opinion
Most opinionated sentences are subjective, but
objective (factual) sentences can imply opinions too
(Liu, 2010)
7KHPDFKLQHVWRSSHGZRUNLQJLQWKHVHFRQGGD\
:HEURXJKWWKHPDWWUHVV\HVWHUGD\DQGDERG\
LPSUHVVLRQKDVIRUPHG
$IWHUWDNLQJWKHGUXJWKHUHLVQRPRUHSDLQ
27
_
Voice Screen Battery Size Weight
Comparison of +
reviews of
Cell Phone 1
Cell Phone 2
_
31
Roadmap
Sentiment Analysis Problem
Document sentiment classification
Sentence subjectivity & sentiment classification
Aspect-based sentiment analysis
Opinion summarization
Sentiment lexicon generation
Mining comparative opinions
Some complications
Generalized sentiment analysis
37
Sentiment classification
Classify a whole opinion document (e.g., a
review) based on the overall sentiment of the
opinion holder (Pang et al 2002; Turney 2002)
Classes: Positive, negative (possibly neutral)
Neutral or no opinion is hard. Most papers ignore it.
An example review:
,ERXJKWDQL3KRQHDIHZGD\VDJR,WLVVXFKDQLFH
phone, although a little large. The touch screen is cool.
7KHYRLFHTXDOLW\LVFOHDUWRR,VLPSO\ORYHLW
Classification: positive or negative?
Perhaps the most widely studied problem.
38
Neutral is ignored.
SVM gives the best classification accuracy based on
balanced training data
Typical result: 80-90% accuracy
Features: unigrams (bag of individual words)
42
Roadmap
Sentiment Analysis Problem
Document sentiment classification
Sentence subjectivity & sentiment classification
Aspect-based sentiment analysis
Opinion summarization
Sentiment lexicon generation
Mining comparative opinions
Some complications
Generalized sentiment analysis
44
Subjectivity classification
Document-level sentiment classification is too coarse
for most applications.
So do sentence level analysis
Assumes a single sentiment per sentence
not always true, so one can classify clauses instead
45
A bootstrapping approach.
A high precision classifier is first used to automatically identify
some subjective and objective sentences.
Two high precision (but low recall) classifiers are used,
A high precision subjective classifier
A high precision objective classifier
Based on manually collected lexical items, single words and n-grams,
which are good subjective clues.
A set of patterns are then learned from these identified
subjective and objective sentences.
Syntactic templates are provided to restrict the kinds of patterns to
be discovered, e.g., <subj> passive-verb.
The learned patterns are then used to extract more subjective
and objective sentences (the process can be repeated).
47
Roadmap
Sentiment Analysis Problem
Document sentiment classification
Sentence subjectivity & sentiment classification
Feature/Aspect-based sentiment analysis
Sentiment lexicon generation
Mining comparative opinions
Some complications
Generalized sentiment analysis
48
Feature/Aspect extraction
Extraction may use:
frequent nouns and noun phrases
Sometimes limited to a set known to be related to the entity of interest or
using part discriminators
e.g., for a scanner entiityRI VFDQQHUVFDQQHUKDV
opinion and target relations
Proximity or syntactic dependency
Standard IE methods
Rule-based or supervised learning
Often HMMs or CRFs (like standard IE)
53
The DP method
DP is a bootstrapping method
Input: a set of seed opinion words,
no aspect seeds needed
Based on dependency grammar (Tesniere 1959).
7KLVphone has good screen
55
Roadmap
Opinion Mining Problem
Document sentiment classification
Sentence subjectivity & sentiment classification
Aspect-based sentiment analysis
Sentiment lexicon generation
Mining comparative opinions
Some complications
Generalized sentiment analysis
56
Sentiment lexicon
Sentiment words or phrases (also called polar words,
opinion bearing words, etc). E.g.,
Positive: beautiful, wonderful, good, amazing,
Negative: bad, poor, terrible, cost an arm and a leg.
Corpus-based approaches
Rely on syntactic patterns in large corpora.
(Hazivassiloglou and McKeown, 1997; Turney, 2002; Yu
and Hazivassiloglou, 2003; Kanayama and Nasukawa,
2006; Ding, Liu and Yu, 2008)
Can find domain dependent orientations (positive, negative,
or neutral).
(Turney, 2002) and (Yu and Hazivassiloglou, 2003)
are similar.
Assign opinion orientations (polarities) to words/phrases.
(Yu and Hazivassiloglou, 2003) is slightly different from
(Turney, 2002)
use more seed words (rather than two) and use log-likelihood ratio
(rather than PMI).
60
The technique
Sentiment analysis to determine whether the context is +ve
or ve.
(J,VDZDvalley LQWZRGD\VZKLFKLVWHUULEOH
This is a negative context.
Statistical test to find +ve and ve candidates.
Pruning
For an aspect with an implied opinion, it has a fixed
opinion, either +ve or ve, but not both.
We find two direct modification relations using a
dependency parser.
Type 1: 22-'HS$
HJThis TV has good picture quality
Type 2: 22-'HS+$-'HS$
HJThe springs of the mattress are bad
Dictionary-based methods
Typically use :RUG1HWV synsets and hierarchies to
acquire opinion words
Start with a small seed set of opinion words.
Bootstrap the set to search for synonyms and antonyms in
WordNet iteratively (Hu and Liu, 2004; Kim and Hovy, 2004;
Valitutti, Strapparava and Stock, 2004; Mohammad, Dunne
and Dorr, 2009).
Kamps et al., (2004) proposed a WordNet distance method to
determine the sentiment orientation of a given adjective.
69
Semi-supervised learning
(Esuti and Sebastiani, 2005)
Use supervised learning
Given two seed sets: positive set P, negative set N
The two seed sets are then expanded using synonym
and antonymy relations in an online dictionary to
JHQHUDWHWKHH[SDQGHGVHWV3DQG1
3DQG1IRUPWKHWUDLQLQJVHWV
Using all the glosses in a dictionary for each term
LQ3 1DQGFRQYHUWLQJWKHPWRDYHFWRU
Build a binary classifier
Tried various learners.
70
Roadmap
Sentiment Analysis Problem
Document sentiment classification
Sentence subjectivity & sentiment classification
Aspect-based sentiment analysis
Sentiment lexicon generation
Mining comparative opinions
Some complications
Generalized sentiment analysis
74
Comparative Opinions
(Jindal and Liu, 2006)
Gradable
Non-Equal Gradable: Relations of the type greater or less than
([RSWLFVRIFDPHUD$LVEHWWHUWKDQWKDWRIFDPHUD
%
Equative: Relations of the type equal to
([camera A and camera B both come in 7MP
Superlative: Relations of the type greater or less than all others
([camera A is the cheapest in market
75
An example
Consider the comparative sentence
&DQRQVRSWLFVLVEHWWHUWKDQWKRVHRI6RQ\DQG1LNRQ
Written by John in 2010.
The extracted comparative opinion/relation:
({Canon}, {Sony, Nikon}, {optics}, preferred:{Canon}, John, 2010)
76
Common comparatives
In English, comparatives are usually formed by
adding -er and superlatives are formed by adding -
est to their base adjectives and adverbs
Adjectives and adverbs with two syllables or more
and not ending in y do not form comparatives or
superlatives by adding -er or -est.
Instead, more, most, less, and least are used before such
words, e.g., more beautiful.
Irregular comparatives and superlatives, i.e., more
most, less, least, better, best, worse, worst, etc
77
Roadmap
Sentiment Analysis Problem
Document sentiment classification
Sentence subjectivity & sentiment classification
Aspect-based sentiment analysis
Sentiment lexicon generation
Mining comparative opinions
Some complications
Generalized sentiment analysis
78
Roadmap
Sentiment Analysis Problem
Document sentiment classification
Sentence subjectivity & sentiment classification
Aspect-based sentiment analysis
Sentiment lexicon generation
Mining comparative opinions
Some complications
Generalized sentiment analysis
84
Sentiment Summary
Sentiment analysis has many sub-problems.
Despite the challenges, applications are flourishing!
Building a reasonably accurate domain specific system is
possible.
Building an accurate generic system is very hard.
But I am optimistic!
92
UNSUPERVISED
INFORMATION
EXTRACTION
99
Why?
takes too long to develop
too expensive
not accurate enough
lacks complete coverage
The Evolution of Information Extraction Technology
Unsupervised
Care 2.0 + Corpus Based
IE
Learning
Care 2.0
Generic Grammar
Augmented IE
HMM, CRF
Supervised Information Extraction
Relation extraction
Relation Extraction (RE) is the task of recognizing
instances of specific relationships between two or more
entities in a natural language text.
In a traditional setting, the target relation types are known
to a RE system in advance, and it can be prepared for its
task either
1. by a knowledge engineer hand-crafting the extraction
rules
2. or by the system itself learning the rules from a set of
hand-labeled training examples.
Both ways require a large expenditure of manual labor.
103
Relation extraction
In recent years, [Banko and Etzioni 2008]
introduced a new setting for the RE task, called
Open Information Extraction (Open IE).
In this setting, the RE system
1. does not know the target relations in advance, and
2. cannot have any relation-specific human input.
The task requires the system itself to identify the
target relations and to train itself for extracting them.
104
To parse!!!
We do perform deep parsing, using a parser
which is built specifically for the task of
Information Extraction.
The SDUVHUVunderlying framework, called CARE-
II, is capable of parsing arbitrary weighted typed-
feature-structure-based context free grammars.
This gives the framework a unique power,
allowing it to use a high-level unification-based
grammar, such as HPSG, while still being able to
flexibly interface with feature-rich sequence
classifiers, such as CRF-based NER and PoS
106
Parsing approach
The parser we built on top of the CARE-II framework
is generic it uses an HPSG-like grammar (derived
primarily from the example grammar in [Sag,
Wasow et al. 2003]).
The parser relies on the CRF-trained NER and PoS
sequence classifiers to provide weights for different
possible typed-feature structure assignments for
different words.
For any input sentence, the parser is able to
generate a single highest- weight parse the parse
which is the most consistent with the NER and POS
classifiers.
107
Sequence classifiers
A sequence classifier is a component that takes as input a
sequence of tokens (the input sentence) and selects for
each token a single label from a small predefined set of
labels.
For example:
1. a named entity recognizer (NER) would select the
ODEHOVIURPDVHWOLNH^3HUVRQ2UJDQL]DWLRQ
/RFDWLRQ1RQH`
2. a part-of-speech tagger (PoS) would select the labels
IURPDVHWOLNH^11113--9%9%*`
109
example
Qualcomm has acquired Elata for 57 million in cash.
Acquisition(Acquirer = QualcommAcquired Elata
It is sufficient to define a single content word WKHYHUEWR
DFTXLUH7KHGHILQLWLRQcan look like this:
113
Example - cont.
The additional positive weight encourages the parses
involving the YHUEDFTXLUHRYHUJHQHULFDOO\GHILQHGZRUGV
If this definition is added to the grammar, the sentence
above will be parsed into the following:
<1: RELN=Org> Qualcomm </1>
<2: RELN=perfect ARG=3> has </2>
<3: RELN=Acquire ACQUIRER=1 ACQUIRED=4
PAST=true> acquired </3>
<4: RELN=Org> Elata </4>
114
Pattern types
The patterns are pieces of parses, and the parses are
word dependency graphs. Thus, all patterns are
connected dependency sub-graphs, and each one
includes at least two entity placeholders. There are three
sets of pattern types:
1. verb-based
2. noun-based
3. BE-based.
116
BE patterns
BE-patterns DUHKHDGHGE\WKHYHUEEHLQits predicative
(non-auxiliary) sense:
;3HUVRQLV mayor m of c Y/Loc
;3HUVRQ mayor m of c Y/Loc
;2UJposs V m headquarters (are) in c
Y/Loc
119
121
'LJLWDO7URZHOV9LVXDO&$5(
122
Visual CARE game changing benefits
development time reduced by 99%
GLVFRYHUWKHUXOHV\RXGLGQWNQRZWRORRNIRU
unprecedented accuracy of over 95%
domain agnostic
A disruptive change in the way industries
are able to utilize content
123
124
Hpsg parser
CARE-II-HPSG is an English grammar written for the
CARE-II framework, based on the principles of HPSG
grammar theory.
The grammar's lexicon is largely underspecified. Only
the most frequent and functional words have full
definitions, while the open classes of words are defined
using generic underspecified lexical entries and tightly-
integrated feature-rich sequence classification models
for part-of-speech tagging (PoS) and named entity
recognition (NER)
For any input sentence, the parser generates a single
highest-weight parse the parse which is the most
consistent with both the grammar rules and the NER and
PoS classifiers.
128
sentences IProblematic
I had severe knee swelling and pain from Levemir
insulin and the dr doesn't think Levemir had
anything to do with the severe pain because knee
swelling wasn't listed as a side effect even though
hand and foot swelling was.
The Charniak's parser missed the important
domain-specific relation between "knee swelling"
and "Levemir insulin" by forming a noun phrase
"pain from Levemir insulin and the dr" and
interpreting it as the subject of "doesn't think".
129
&KDUQLDNV Parse
130
9&V3DUVH
131
sentences IIProblematic
Financial Systems Innovation LLC has entered
into a settlement agreement covering a patent
that applies to credit card fraud protection
technology with Lone Star Steakhouse, Inc.
The wide-coverage parser makes a PP-
attachment mistake, attaching "with Lone Star
Steakhouse, Inc" to the immediately preceding
NP instead of "settlement agreement".
132
&KDUQLDNV Parse
133
9&VSDUVH
134
Entity extraction
There 4 different methods to define additional types of
entity types:
1. Using a separately-trained NER model
2. CARE-II rules
3. CARE-II-HPSG lexicon definitions
4. Lists of allowed values (for entity types for which the
sets of entities are closed).
Arbitrary mixing of these methods is also possible and
effective.
138
Handling negation
Nestle S.A. (NESN.VX), the world's largest food and
beverages producer, Tuesday said it won't bid for U.K.-
based confectionery company Cadbury Plc (CBY).
The syntax and semantics of such forms of negation are
handled in the generic grammar in a way compatible with
the HPSG grammar theory.
In practice, if either the main verb or one of the slots of a
relation is modified by a negating modifier ("not" and its
various forms, and other negating words), then the
extracted relation is marked as negated.
141
example
The Rel_ORG_enter_into_agreement_with_ORG pattern adds
four entries: the prepositions "into" and "with", the noun
"agreement", and the verb "enter".
1. "Into" DQGZLWKDUHGHILQHGas argument prepositions
2. "agreement" is defined as a noun with a special
SYN.HEAD.FORM and without any non-generic semantics
3. "enter" is defined as a verb with three complements and with
the output relation semantics.
The definitions of non-pattern-specific words, such as "into",
"with" and "agreement" can be reused by many patterns.
The pattern names show only the complements the required
pieces of the pattern. The actual instances of the relation may
also contain optional pieces, specified by modifiers.
143
Relation clustering
The automatic pattern clustering uses a variant of the
HAC (Hierarchical Agglomerative Clustering) algorithm
with single linkage, which was shown in (Rosenfeld and
Feldman 2007) to be superior for the relation identification
task.
The direct similarity estimation between patterns takes
into account:
1. the structural similarity between patterns, including
standard syntactic transformations;
2. identity of slots' entity types;
3. identity or synonymy or other association between
specific words in the patterns, discovered using
WordNet (Fellbaum 1998).
144
Clustering example
145
examples
In order for a non-default domain-specific modifier to be useful,
and therefore learnable, its NP part must contain a relation
argument extractable entity either directly, as in:
In January 1997, Hays bought German distributor Daufenbach
for 30 million GBP, ...
or via a PP-NP chain, as in:
Ms. Bennett succeeds William J. Viveen, Jr., as a member of
the Interleukin Board of Directors.
assuming the verb relation patterns Rel_ORG_buy_ORG and
Rel_PERSON_succeed_PERSON are already learned, the two
sentences above would generate the PP modifier patterns
Mod_verb_in_DATE and Mod_verb_as_member_of_ORG.
147
Co-reference resolution
we implemented the co-reference resolution system of (Lee,
Peirsman et al. 2011), which was the best-performing in
CoNLL-2011 shared task, and which is perfectly adaptable to
the CARE-II-HPSG/VC environment.
The method for resolution of co-references is based on locating
all noun phrases, identifying their properties, and then
clustering them in several deterministic iterations (called
sieves), starting with the highest-confidence rules and moving
to lower-confidence higher-recall ones. In each iteration, the
order of candidate checks is deterministic, and any matching
noun phrases with matching properties are immediately
clustered together.
Despite its simplicity, the method showed state-of-the-art
accuracy, outperforming the other systems in the CoNLL
shared task.
149
Co-reference resolution
The method is especially suitable for the our framework,
because all information the method requires is already
extracted: the noun phrases are located from the parses,
together with their properties, which are identified from
HPSG feature structures.
Co-reference resolution module tries to resolve all noun
phrases, although non-entity ones are discarded. This is
necessary for improving the accuracy on the relevant
entity mentions, by removing irrelevant candidates.
150
Examples
The entity to resolve to is the first found one with matching features. The order of checking
is strictly by distance (closest first), except for pronominal/generic-nominal references, for
which
the entities in subject positions are checked before the others.
Microsoft announced it plans to acquire Visio. The company said it will finalize its plans
within a week.
Only pronomial and generic nominal references are here. The first "LWVWUDLJKWIRUZDUGO\
matches "Microsoft". "The company" also matches "Microsoft" first, because it's a
pronominal reference, and "Microsoft" is in the subject position. The second "it" matches
"the company", and so is also resolved to "Microsoft".
Mark said that he used Symlin and it caused him to get a rash. He said that it bothered
him.
First "he" resolves to "Mark". "It" does not match "Mark", and so is resolved to
"Symlin". Other "Him" and "he" are also resolved to "Mark". The second "it" is resolved to
"a rash", because it is closer and neither of the matching candidates - "Symlin" and "a
151
Post processing
The post-processing phase tried to combines a few simple
relation to one complex relation.
The slots of any adjacent relation-pieces are merged, if the
pieces are compatible, and if none of the slots contradict each
other.
For the DrugReplacement relation, StopUsingDrug can merge
with StartUsingDrug and with UseDrug. For the SideEffect
relation, HasSymptom can merge with any of the three:
StartUsingDrug, StopUsingDrug, or UseDrug.
After merging, all StartUsingDrug, UseDrug, and HasSymptom
relations that were not identified as parts of a larger relation
can be removed, since we are interested only in the full
DrugReplacement and SideEffect relations in this project.
152
evaluation
The input corpus for the medical domain consists of about
650,000 messages downloaded from the archives of
several public medical forums.
The contents of the messages were separated into
sentences, filtered and deduplicated, leaving about
300,000 sentences totaling in about 40MB of ASCII text.
153
Evaluation results
Without Coreference With Coreference
Evaluation methodology
Precision
We evaluated on a random set of 200 extractions.
Recall
We checked only the sentences that contained at least two
DRUG entities for the DrugReplacement and at least one
DRUG and at least one SYMPTOM entity for the SideEffect,
together with an indicative word, such as "replace", "switch",
"change", or one of their derivatives for the DrugReplacement.
155
Coverage experiment
we used VC to learn relations from Corpus B
independently from Corpus A, and compared the sets of
relations extracted from the Corpus B by domain-specific
lexicons learned from itself and from Corpus A.
We have found that out of 1742 relation instances found
in Corpus B, we were able to extract 1367 of them
(78.4%) using the patterns learned from Corpus A.
Sentiment Analysis of Stocks from News
Sites
3. Score aggregation:
daily and cumulative
score
Phrase-level SA:
highly successful top selling positive
Or at best neutral
Taking into account voluntary recall negative
1HHGWRUHFRJQL]HWKHZKROHVHQWHQFHDVDSURGXFWUHFDOOHYHQW
CaRE extraction
Engine
Template Based Approach to Content
Filtering
Hybrid Sentiment Analysis
Events
(Predicate)
Patterns Dictionaries
(Phrasal) (Lexical)
All levels are part of the same rulebook, and are therefore
considered simultaneously by CaRE
Dictionary-based sentiment
Started with available sentiment lexicons
Domain-specific and general
Improved by our content experts
Examples
Modifiers: attractive, superior, inefficient, risky
Verbs: invents, advancing, failed, lost
Nouns: opportunity, success, weakness, crisis
Expressions: exceeding expectations, chapter 11
Emphasis and reversal
successful, extremely successful,
far from successful
Event-Based Sentiment
Product release/approval/recall, litigations, acquisitions, workforce
change, analyst recommendations and many more
Semantic role matters:
Google is being VXHGLVVXLQJ
Need to address historical/speculative events
Google acquired YouTube in 2006
What if Google buys Yahoo and the software giant Microsoft
remains a single company fighting for the power of the Internet?
Extract sentiment and events of public companies from online
news articles to enable better trading decisions
TEVA
167
MOS
SNDK
GM
MU
CLF
0DF\V
JC Penny
POT
Ford
Monsanto
Mining Medical User Forums
178
diabetesforums.com
Cleaning
healthboards.com
Chunking
forum.lowcarber.org
diabetes.blog.com**
Information
Extraction
diabetesdaily.com
Comparison
s
** Messages in Diabets.blog.com were focused mainly on Byetta
Drug Analysis
Drug Co-Occurrence - Spring Graph Perceptual Map
180
Several
Pockets of
drugs that were
mentioned
frequently
together in a
message were
identified
Byetta was
mentioned
frequently with:
Glucotrol
Januvia
Amaryl
Actos
Avandia
Prandin
Symlin
Lifts larger than 3
Width of edge reflects how frequently the two drugs appeared together over and beyond what
one would have expected by chance
Drug Usage Analysis
Drug Co-taking 'UXJVPHQWLRQHGDV7DNHQ7RJHWKHU
181
There are two main clusters
of drugs that are mentioned
DVWDNHQWRJHWKHU
Rel_cause_DRUvSYM(DRUG, SYMPTOM)
Negative (1)
Actos hasn't caused any weight gain, I am still losing some.
Positive (25)
I also am on Synthroid, Atenolol, Diovan, Lotrel, Lexapro, Vitorin and Prilosec OTC. I didn't realize that Actos can cause
a weight gain
as I had never read it as a side effect; however, after reading all of the
comments on this site, I now know why my weight has increased over the past few months since taking on it.
I don't take any oral meds, but from what I have read here, Actos causes weight gain because of water retention.
why does the endo think you're a type 1? oral meds are usually given only to type 2's,as type 2's have insulin resistance.
oral meds
treat the insulin resistance. type 1's require insulin.....i take actoplus met- which is actos and metformin.actos is like
avandia and
i've had no heart issues.....tho-avandia and actos can cause weight gain....take care,trish
Actos causes edema and weight gain also.
Actos can cause weight gain (so can Avandia, it's cousin)
Now I have started to see a lot of reports of Actos causing weight gain, among other things.
for the record, Actos can, and does, cause weight gain/water retention.
I'm on both - what did you hate about Metformin? (Actos causes weight gain, metformin weight loss)
Also I hear that the Actos causes weight gain, so now I am afraid the new pill will cause me to gain weight.
I'm type 1 so only on insulin, but I have heard that Actos can cause weight gain.
Avandia & Actos, especially in combination with insulin, causes fluid retention and/or fat weight gain.
My endocrinologist warned me that Actos can cause significant weight gain.
Actos caused weight gain and fluid retention in my chest.
Metformin causes weight loss, Avandia and Actos causes the birth of new fat cells and weight gain.
Side Effects
Side Effects and Remedies
Red lines side effects/symptoms
Blue lines - Remedies
See which
symptoms are most
complained about
Drugs Taken in Combination
192
XML output of VC
USING MULTI-VIEW LEARNING
TO IMPROVE DETECTION
OF INVESTOR SENTIMENTS ON
TWITTER
Motivation
Stocks-related messages on social media, Twitter in
particular, have several interesting properties with regards
to the sentiment analysis task.
The analysis is particularly challenging:
frequent typos
bad grammar
idiosyncratic expressions specific to the domain and the media.
Stocks-related messages almost always refer to the state
of specific entities companies and their stocks at
specific times (times of sending). This state is an objective
property and even has a measurable numeric
characteristic, namely the stock price.
Multi View Learning
Given a large dataset of twitter messages ("twits"), it is
possible to create two separate "views" on the dataset, by
analyzing the text of the twits and their external properties
separately.
Using the two views, we can expand the coverage from
sentiments detectable using generic sentiment analysis
tools to many other twits in the dataset.
We can use the results to learn new sentiment
expressions.
Agenda
1. Complexity and challenges associated with analyzing
finance-related message.
2. Related work.
3. Methodological approach for multi-view learning.
4. The SA system.
5. Experimental design and results.
6. Concludes and Future Research.
Complexity and challenges associated
with analyzing finance-related message
Language Challenges
Sarcasm - ,W
VUXPRUHGWKDW)DFHERRNZLOODQQRXQFHD
new kind of free stock dividend tonight, namely
CrashVille. $FB $ZNGA #SoCalledEarnings
Price Dependency - 6,1$ZDWFK57.50 which now is a
resistance, $59.21 that will open the door to $62.48
before earning, if not we will break $50.
If the price at the time of the analysis is 59.21, the sentiment is very
positive, as the author believes it can reach 62.48,
otherwise, it is negative as the author believes it can drop to $50.
General SA Issues
Sentiment PRGLILHUVHJKLJKO\
Emphasis PRGLILHUVHJPRVWO\
Opposite PRGLILHUVHJIDUIURP
Sentiment VKLIWHUVHJQRW
Anaphora resolution is also a challenge when conducting
sentiment analysis, although, the short messages on
social media often bear little anaphora.
Trading MU (Micron)
Related work Text mining and
Finance
Engelberg (2008)
Found that qualitative earnings information has additional predictability for
asset prices beyond the predictability in quantitative information.
Loughran and McDonald (2011)
Uses textual analysis to examine the sentiment of corporate 10-K reports.
Feldman at al. (2011)
The Stock Sonar uses state-of-the-art information extraction technologies to
analyze news relating to stocks and aggregate the sentiments into a score
and graphically present the information.
Boudoukh et al. (2012)
Found evidence of a strong relationship between stock price changes
and information arrival by using relevant news articles as a proxy.
Information arriving to investors sooner can provide a significant competitive
advantage to those investors. Vis--vis, it is reasonable to assume that
social media response to news is much quicker than professional media.
Related work Social networks
Sprenger & Welpe (2010)
Found that tweet sentiment associated with abnormal stock returns and
message volume to be a good predictor for next-day trading volume.
Bar-haim et al. (2011)
Proposed a general framework for identifying expert investors, and used it
as a basis for several models that predict stock rise from stock microblogging
messages.
Bollen, Mao, and Zeng (2011)
Investigated whether measurements of collective mood states derived from
large-scale Twitter feeds are correlated to the value of the DJIA over time.
Their results suggested that the accuracy of DJIA predictions can be
significantly improved by including specific public mood dimensions.
Xu (2012)
Used NLP to get the public sentiment on individual stocks from social media.
)RXQGWKDWXVHUVactivity overnight significantly correlates positively to the
stock trading volume the next business day.
Found Granger Causality (in 9 out of 15 stocks studied) between collective
sentiments for afterhours and the change of stock price for the next day
Analysis
One particular point in time that should be noted in the
figure is that on April 1st 2011, there were both positive
and negative messages about the company.
However, the positive messages discussed the stock past
performance (which was, indeed, positive for the
referenced period),
while the negative messages reflect the investors current
and future expectations from the stock (which appears to
FRUUHODWHZLWKWKHDFWXDOVWRFNSULFHVWUHQGVLQWKH
presented example).
Basic Learning Model
Assume first, that we have a large corpus T = {t1, t2, ...} of text
messages. Each message t has a true polarity
Pol(t) POLS={POS, NEG, NEUTRAL}, which can always be
identified (by people) by reading and understanding the text.
We make further assumption that the polarity of a message is
determined by occurrence of various sentiment expressions
within the message's text.
The nature of the expressions is not essential at this point.
They can be individual words, multi-word sequences,
sequences with gaps, syntactic patterns, etc. The important
point is that given a message ti, it must be easy to list all
expressions occurring within it. For the purposes of sentiment
expression learning, we represent message texts as bags-of-
expressions: ti = {wi1, wi2, ...}.
Observations
If the true polarity Pol(t) were known for each message tT,
there would be a natural way to search for sentimental
expressions by simply counting their occurrences. However, in
a large unlabeled corpus, the true polarities of messages are
unknown.
On the other hand, the polarity of a message is causally-
independently influenced by external (relative to the message's
text) factors:
if a given stock is doing good or bad on a given day, then the polarity
of the messages about that stock would tend to be correspondingly
positive or negative;
messages from the same author about the same stock would tend to
have the same polarities; etc.
Thus, we have two parallel views on a set of twits, which allows
multi-view learning.
Parallel Views Textual View
Given a large corpus T we first process it with some text-
based SA (sentiment analysis) system, which performs as
a function SA : T o POLS, producing a classification
T = TSA-POS TSA-NEG TSA-NEUTRAL.
The SA system is assumed to have a relatively high
precision for polarized twits, but insufficient recall.
TSA-POS and TSA-NEG generally contain mostly positive and
negative messages, respectively
TSA-NEUTRAL cannot be assumed to contain only neutral
twits. It is also much bigger than the two polarized sets.
Parallel Views : The Second View
We process the corpus T with a feature extractor, which
generates a real-valued high-dimensional vector for each
message, using any possible property of it that is conditionally
independent from its text content given its polarity.
Using this representation, we train a binary SVM classifier with
TSA-POS and TSA-NEG as the training data. This classifier then
produces a score f(t) for each message t in TSA-NEUTRAL.
The significant properties of f(t) are:
(1) because it is grounded on generic SA and external
properties, its sign and magnitude correlates with the true
polarity of t, but
(2) it is independent from the text patterns within t (conditional
on t's true polarity).
Estimating P(Pol(w) = A) - I
Let there be a previously unknown text pattern w, appearing in
TSA-NEUTRAL. We are interested in probabilistically estimating the
polarity of w, that is, in the value of P(Pol(w) = A), where
A{POS, NEG}.
Let Tw = { tTSA-NEUTRAL : wt } be the set of all messages
containing w. Then the probability P(Pol(w) = A) can be
estimated from f(t) scores of messages in Tw:
The constant prior P(Pol(w) = A) can be ignored, assuming the
set Tw is sufficiently large. In the main factor, we can safely
assume that the different twits in Tw are independent from each
other, so:
Estimating P(Pol(w) = A) - II
The marginal P(f(t)) can be estimated directly from the
SVM classifier's scores on T. In the other part of the
formula above, we are only dealing with twits that contain
w whose polarity is non-neutral. According to our
simplifying assumptions, we proceed as if there could be
no conflicts, and the polarities of all twits in Tw are equal to
the polarity of w. And then, the likelihood P(f(t) | Pol(w)=A)
can be reduced to:
The constant marginal P(Pol(t) = A) can be directly
estimated from a manually labeled development test set.
And the rest can be estimated using the SA-polarized
sets.
Estimating P(Pol(w) = A) - III
Due to the conditional independence of SA(t) from f(t),
given the true polarity of t, we have:
P(f(t) | SA(t)=B) and P(SA(t)=B) are estimated directly
from the SA results and SVM results.
P(SA(t) = B | Pol(t) = A) is estimated on a development
test set.
architecture of the learning system
Unlabeled
Corpus
Seed SA SA-Neutrals
SA-POS SA-NEG
Trained Classifier based on
External Features
Content Profiling
Concept Mining
Understand the negative and positive concepts
that consumers associate with top shows in their
tweets, Facebook and Google+ updates
Visualize and track the trending concepts
associated with each show over time
238
Concept Sentiment
A&E S how Related Messages
Concept
Identification
Sentiment
Processing
Negative Concept Positive Concept
Associations Associations
239
Hoarders
The First 48
Longmire
Intervention
Storage Wars
Criminal Minds
Bates Motel
Duck Dynasty
Opinions/Feedback
161
Emotional
What I love about Connection
it
1,990 2,556
Great Show
13,889
22.8%
19.0%
17.0%
12.9%
10.9%
4.9%
4.1%
2.9%
1.2%
0.9%
0.6%
0.5%
DYNASTY
0.5%
0.5%
0.4%
0.3%
WHAT I LOVE ABOUT DUCK
0.2%
0.2%
0.2%
241
0.2%
0.1%
242
Makes
Wednesdays
Better Funniest Show
9.8% 10.4%
Best Show on TV
Opinions/Feedback 19.9%
The Perfect Date Beards are back
Love Reruns
29.3% 2.5%
3.1%
2 new episodes
4.3%
Pay talent
more
9.3%
Great video
43.5%
Scripted
12.4%
Darius Rucker
24.8%
243
RESEARCH INSIGHTS
244
Research Insights
Social Media Popularity Tracking
Track daily by show the total mentions, positive and
negative across Facebook, Google+ and Twitter
8QGHUVWDQGHDFKVKRZVGDLO\Pos/Neg ratio the true measure of a
SURJUDPVUHVRQDQFHZLWK\RXUDXGLHQFH
Crossover Profiling
Understand the off-network shows that consumers
mention most frequently with your shows across social
media
Understand both positive and negative frequencies
246
Popularity Tracking
247
%LJJHVW)DQVDQG%LJJHVW
Identify the authors who make the most comments about
your shows: both positive and negative.
Total PositiveNegative
Most Mentions (by single author) 98 24 94