Generalizing Dependency Features for Opinion Mining

Similar documents
Extracting Opinion Targets in a Single- and Cross-Domain Setting with Conditional Random Fields

Lecture 10: POS Tagging Review. LING 1330/2330: Introduction to Computational Linguistics Na-Rae Han

May All Your Wishes Come True: A Study of Wishes and How to Recognize Them

Motivation. Motivation. Motivation. Finding Deceptive Opinion Spam by Any Stretch of the Imagination

Sentiment Analysis of Reviews: Should we analyze writer intentions or reader perceptions?

Looking for Subjectivity in Medical Discharge Summaries The Obesity NLP i2b2 Challenge (2008)

Detecting Deception in On and Off-line Communications

Rating prediction on Amazon Fine Foods Reviews

Introduction to Sentiment Analysis

Learning from Online Health Communities. Noémie Elhadad

Text Mining of Patient Demographics and Diagnoses from Psychiatric Assessments

Our goal in this section is to explain a few more concepts about experiments. Don t be concerned with the details.

Asthma Surveillance Using Social Media Data

Words: Evaluative, Emotional, Colourful, Musical!!

Joint Inference for Heterogeneous Dependency Parsing

Combining unsupervised and supervised methods for PP attachment disambiguation

Learning the Fine-Grained Information Status of Discourse Entities

CSE 258 Lecture 2. Web Mining and Recommender Systems. Supervised learning Regression

Parsing Discourse Relations. Giuseppe Riccardi Signals and Interactive Systems Lab University of Trento, Italy

Analyzing Emotional Statements Roles of General and Physiological Variables

Discovering and Understanding Self-harm Images in Social Media. Neil O Hare, MFSec Bucharest, Romania, June 6 th, 2017

. Semi-automatic WordNet Linking using Word Embeddings. Kevin Patel, Diptesh Kanojia and Pushpak Bhattacharyya Presented by: Ritesh Panjwani

Outline. Teager Energy and Modulation Features for Speech Applications. Dept. of ECE Technical Univ. of Crete

A Computational Approach to Automatic Prediction of Drunk-Texting

Exploring Normalization Techniques for Human Judgments of Machine Translation Adequacy Collected Using Amazon Mechanical Turk

Committee-based Decision Making in Probabilistic Partial Parsing

How much data is enough? Predicting how accuracy varies with training data size

Foundations of Natural Language Processing Lecture 13 Heads, Dependency parsing

Exploiting Ordinality in Predicting Star Reviews

CSE 258 Lecture 1.5. Web Mining and Recommender Systems. Supervised learning Regression

Improving Binary Classification on Text Problems using Differential Word Features

Visualizing the Affective Structure of a Text Document

An Improved Algorithm To Predict Recurrence Of Breast Cancer

Extraction of Adverse Drug Effects from Clinical Records

Kai-Wei Chang UCLA. What It Takes to Control Societal Bias in Natural Language Processing. References:

Semi-Automatic Construction of Thyroid Cancer Intervention Corpus from Biomedical Abstracts

Textual Emotion Processing From Event Analysis

Processing MWEs: Neurocognitive Bases of Verbal MWEs and Lexical Cohesiveness within MWEs

1 Pattern Recognition 2 1

Gene Selection for Tumor Classification Using Microarray Gene Expression Data

Predicting Breast Cancer Survival Using Treatment and Patient Factors

Shopping online? Don t fall into the ratings trap. solace outliers mediocre exposure banish blackout blind grating clings fond spookily

Negative Deceptive Opinion Spam

Machine learning II. Juhan Ernits ITI8600

Deep Learning based Information Extraction Framework on Chinese Electronic Health Records

Sentiment Classification of Chinese Reviews in Different Domain: A Comparative Study

ASL ONE CURRICULUM UNIT ONE: Lesson 1.1 Fingerspelling I Rule #1 of Deaf Culture Assessment: Assignment 1.1

Reader s Emotion Prediction Based on Partitioned Latent Dirichlet Allocation Model

Cocktail Preference Prediction

Positive and Unlabeled Relational Classification through Label Frequency Estimation

Positive and Unlabeled Relational Classification through Label Frequency Estimation

Protocol Title: A Comparison of Interventions to Teach Melanoma Patients Skin Selfexamination

IBM Research Report. Low-Cost Call Type Classification for Contact Center Calls Using Partial Transcripts

TWITTER SENTIMENT ANALYSIS TO STUDY ASSOCIATION BETWEEN FOOD HABIT AND DIABETES. A Thesis by. Nazila Massoudian

Feature Engineering for Depression Detection in Social Media

Sound Texture Classification Using Statistics from an Auditory Model

Predicting Sleep Using Consumer Wearable Sensing Devices

IT S A WONDER WE UNDERSTAND EACH OTHER AT ALL!

MODEL SELECTION STRATEGIES. Tony Panzarella

Applied Machine Learning in Biomedicine. Enrico Grisan

controlling for confounds in online measures of sentence complexity

GfK Verein. Detecting Emotions from Voice

TKT CLIL LESSON PLAN

Past Progressive Story 1 (Past Continuous)

Word2Vec and Doc2Vec in Unsupervised Sentiment Analysis of Clinical Discharge Summaries

RESOURCE CATALOGUE. I Can Sign A Rainbow

Speech Processing / Speech Translation Case study: Transtac Details

Filippo Chiarello, Andrea Bonaccorsi, Gualtiero Fantoni, Giacomo Ossola, Andrea Cimino and Felice Dell Orletta

Survey Research. We can learn a lot simply by asking people what we want to know... THE PREVALENCE OF SURVEYS IN COMMUNICATION RESEARCH

Classification. Methods Course: Gene Expression Data Analysis -Day Five. Rainer Spang

Machine Learning to Inform Breast Cancer Post-Recovery Surveillance

UniNE at CLEF 2015: Author Profiling

Connecting Distant Entities with Induction through Conditional Random Fields for Named Entity Recognition: Precursor-Induced CRF

Lillian Lee Dept. of Computer Science, Cornell University

Look back at the prediction you made about what would happen in act 3. Discuss with the person next to you what you think will happen.

Lesson 1 Sentence Structure

Modeling Sentiment with Ridge Regression

A Comparison of Collaborative Filtering Methods for Medication Reconciliation

Correlation Analysis between Sentiment of Tweet Messages and Re-tweet Activity on Twitter

Presenting the Read-Aloud

Automatic Extraction of ICD-O-3 Primary Sites from Cancer Pathology Reports

Signing for the Deaf using Virtual Humans. Ian Marshall Mike Lincoln J.A. Bangham S.J.Cox (UEA) M. Tutt M.Wells (TeleVirtual, Norwich)

What makes us special? Ages 3-5

Supporting Information Identification of Amino Acids with Sensitive Nanoporous MoS 2 : Towards Machine Learning-Based Prediction

Rumor Detection on Twitter with Tree-structured Recursive Neural Networks

Stepwise method Modern Model Selection Methods Quantile-Quantile plot and tests for normality

An Avatar-Based Weather Forecast Sign Language System for the Hearing-Impaired

Identifying Adverse Drug Events from Patient Social Media: A Case Study for Diabetes

Microblog Retrieval for Disaster Relief: How To Create Ground Truths? Ribhav Soni and Sukomal Pal

S URVEY ON EMOTION CLASSIFICATION USING SEMANTIC WEB

GSE Teacher Toolkit Download

Discourse Level Opinion Relations: An Annotation Study

Yes, you do have a choice. Understanding and using Choice Theory with clients Presented by: Dana Hampson, MBA, BCC, LPC-S The Balanced Life, LLC

The Rough Guide to AAC! Essential ideas for success. Suzanne Martin Speech & Language Therapist: AAC Consultant ACE Centre

Using a grammar implementation to teach writing skills

Lesson 11.1: The Alpha Value

9-10 Issue 181 VIBE ACTIVITIES. Healthy Vibe - I Quit Because... page 22. Issue 181 Page 1 Y E A R. Name:

college essays help victorylandgroup >>>CLICK HERE<<<

Data mining for Obstructive Sleep Apnea Detection. 18 October 2017 Konstantinos Nikolaidis

Transcription:

Generalizing Dependency Features for Mahesh Joshi 1 and Carolyn Rosé 1,2 1 Language Technologies Institute 2 Human-Computer Interaction Institute Carnegie Mellon University ACL-IJCNLP 2009 Short Papers Track

One Slide Summary Goal: utilize structure in language to improve opinion vs. not-opinion classification in product reviews Key idea: Back off the head word in a dependency relation to it s part-of-speech tag, use these back-off features amod(camera,great) => amod(nn,great) Result: It works! Improvement in accuracy from 0.652 to 0.679 yes, it s significant! 2

One Slide Summary Goal: utilize structure in language to improve opinion vs. not-opinion classification in product reviews Key idea: Back off the head word in a dependency relation to it s part-of-speech tag, use these back-off features amod(camera,great) => amod(nn,great) Result: It works! Improvement in accuracy from 0.652 to 0.679 yes, it s significant! 2

Task Description Opinion mining in product reviews Given a sentence from a product review Predict whether or not it is an opinion sentence Opinion Sentence: If a sentence contains one or more product features and one or more opinion words, then the sentence is called an opinion sentence. [Hu and Liu; SIGKDD 2004] 3

Examples Opinion sentences It is very light weight and has good signal strength. Player works and looks great if you can get the dvd s to play. Non-opinion sentences Had it for a week. If this doesnt bring back the picture, try pressing this button without playing a dvd. 4

Research Question How can we better utilize structure in language for opinion classification? 5

Dependency Relations 6

Dependency Relations it is a fantastic camera and well worth the price. 6

Dependency Relations it is a fantastic camera and well worth the price. (Stanford parser output) ==> nsubj(camera-5,it-1) cop(camera-5,is-2) det(camera-5,a-3) amod(camera-5,fantastic-4) advmod(worth-8,well-7) det(price-10,the-9) amod(price-10,worth-8) conj_and(camera-5,price-10 ) 6

Dependency Relations it is a fantastic camera and well worth the price. (Stanford parser output) ==> Relation nsubj(camera-5,it-1) cop(camera-5,is-2) det(camera-5,a-3) amod(camera-5,fantastic-4) advmod(worth-8,well-7) det(price-10,the-9) amod(price-10,worth-8) conj_and(camera-5,price-10 ) 6

Dependency Relations it is a fantastic camera and well worth the price. (Stanford parser output) ==> nsubj(camera-5,it-1) cop(camera-5,is-2) det(camera-5,a-3) amod(camera-5,fantastic-4) advmod(worth-8,well-7) det(price-10,the-9) amod(price-10,worth-8) conj_and(camera-5,price-10 ) 6

Dependency Relations it is a fantastic camera and well worth the price. (Stanford parser output) ==> Head Word nsubj(camera-5,it-1) cop(camera-5,is-2) det(camera-5,a-3) amod(camera-5,fantastic-4) advmod(worth-8,well-7) det(price-10,the-9) amod(price-10,worth-8) conj_and(camera-5,price-10 ) 6

Dependency Relations it is a fantastic camera and well worth the price. (Stanford parser output) ==> nsubj(camera-5,it-1) cop(camera-5,is-2) det(camera-5,a-3) amod(camera-5,fantastic-4) advmod(worth-8,well-7) det(price-10,the-9) amod(price-10,worth-8) conj_and(camera-5,price-10 ) 6

Dependency Relations it is a fantastic camera and well worth the price. (Stanford parser output) ==> Modifier Word nsubj(camera-5,it-1) cop(camera-5,is-2) det(camera-5,a-3) amod(camera-5,fantastic-4) advmod(worth-8,well-7) det(price-10,the-9) amod(price-10,worth-8) conj_and(camera-5,price-10 ) 6

Dependency Relations 7

Dependency Relations amod(camera-5,fantastic-4) adjectival modifier relationship between camera (head, noun) and fantastic (modifier, adjective) 7

Dependency Relations amod(camera-5,fantastic-4) adjectival modifier relationship between camera (head, noun) and fantastic (modifier, adjective) advmod(worth-8,well-7) adverbial modifier relationship between worth (head, adjective) and well (modifier, adverb) 7

Previous Work 8

Previous Work Success in using dependency relations: [Gamon; COLING 2004], [Matsumoto et al.; PAKDD 2005] Different task: predicting customer satisfaction ratings, and polarity (positive / negative) for movie reviews respectively [Wilson et al.; AAAI 2004] Different task: predicting strength of subjective language Use full set of dependency relations 8

Previous Work Success in using dependency relations: [Gamon; COLING 2004], [Matsumoto et al.; PAKDD 2005] Different task: predicting customer satisfaction ratings, and polarity (positive / negative) for movie reviews respectively [Wilson et al.; AAAI 2004] Different task: predicting strength of subjective language Use full set of dependency relations We propose multiple generalization approaches one of our approaches was used by Gamon as well as Wilson et al. 8

Previous Work 9

Previous Work Dependency relations not found to be useful: [Dave et al.; WWW 2003], [Ng et al.; ACL 2006] Different task: polarity prediction in product reviews and movie reviews respectively Both use a subset of dependency relations manually chosen grammatical relations 9

Previous Work Dependency relations not found to be useful: [Dave et al.; WWW 2003], [Ng et al.; ACL 2006] Different task: polarity prediction in product reviews and movie reviews respectively Both use a subset of dependency relations manually chosen grammatical relations We use the full set of dependency relations 9

Recent Work - not in paper 10

Recent Work - not in paper Transformation of dependency relations [Greene and Resnik; NAACL 2009] Given dependency relations of form: relation(head_word,modifier_word) Create features of form: head_word-relation & relation-modifier_word 10

Recent Work - not in paper Transformation of dependency relations [Greene and Resnik; NAACL 2009] Given dependency relations of form: relation(head_word,modifier_word) Create features of form: head_word-relation & relation-modifier_word 10

Recent Work - not in paper Transformation of dependency relations [Greene and Resnik; NAACL 2009] Given dependency relations of form: relation(head_word,modifier_word) Create features of form: head_word-relation & relation-modifier_word Dependency relations chosen if they contain domain relevant verbs and/or nouns Most closely related to our approach 10

Lesson from Past Work 11

Lesson from Past Work Using full set of dependency relations works better 11

Lesson from Past Work Using full set of dependency relations works better However, there might be overfitting in scarce data scenarios common to NLP where annotated data is expensive! Need features that can generalize well 11

Motivating Our Approach Consider two opinion sentences: This is a great camera! Despite its few negligible flaws, this great mp3 player won my vote. 12

Motivating Our Approach Both have the dependency relation amod with different pair of words participating: amod(camera,great) amod(player,great) We can see the structural similarity of these features A machine learning algorithm can t! 13

Motivating Our Approach Both have the dependency relation amod with different pair of words participating: amod(camera,great) amod(player,great) We can see the structural similarity of these features A machine learning algorithm can t! 13

So Lets Back Off! Back off the head word to its part-of-speech tag amod(camera,great) => amod(nn,great) amod(player,great) => amod(nn,great) Now the algorithm can see that these are similar features (we made them identical) 14

Advantages of Backing Off 15

Advantages of Backing Off Stronger evidence of association of a generalized feature with the opinion category 15

Advantages of Backing Off Stronger evidence of association of a generalized feature with the opinion category New test sentence: this is a great phone amod(phone,great): may not be useful because we might have never seen it! amod(nn,great): still valid 15

Backing Off / Generalizing 16

Backing Off / Generalizing Composite Back-off Features Head word => its part-of-speech tag amod(camera,great) => amod(nn,great) 16

Backing Off / Generalizing Composite Back-off Features Head word => its part-of-speech tag amod(camera,great) => amod(nn,great) 16

Backing Off / Generalizing Composite Back-off Features Head word => its part-of-speech tag amod(camera,great) => amod(nn,great) Modifier word => its part-of-speech tag amod(camera,great) => amod(camera,jj) 16

Backing Off / Generalizing Composite Back-off Features Head word => its part-of-speech tag amod(camera,great) => amod(nn,great) Modifier word => its part-of-speech tag amod(camera,great) => amod(camera,jj) 16

Backing Off / Generalizing Composite Back-off Features Head word => its part-of-speech tag amod(camera,great) => amod(nn,great) Modifier word => its part-of-speech tag amod(camera,great) => amod(camera,jj) Full Back-off Features Head word => its part-of-speech tag; Also modifier word => its part-of-speech tag amod(camera,great) => amod(nn,jj) used by [Gamon; COLING 2004] and [Wilson et al.; AAAI 2004] 16

Backing Off / Generalizing Composite Back-off Features Head word => its part-of-speech tag amod(camera,great) => amod(nn,great) Modifier word => its part-of-speech tag amod(camera,great) => amod(camera,jj) Full Back-off Features Head word => its part-of-speech tag; Also modifier word => its part-of-speech tag amod(camera,great) => amod(nn,jj) used by [Gamon; COLING 2004] and [Wilson et al.; AAAI 2004] 16

Backing Off / Generalizing Composite Back-off Features Head word => its part-of-speech tag amod(camera,great) => amod(nn,great) Modifier word => its part-of-speech tag amod(camera,great) => amod(camera,jj) Full Back-off Features Head word => its part-of-speech tag; Also modifier word => its part-of-speech tag amod(camera,great) => amod(nn,jj) used by [Gamon; COLING 2004] and [Wilson et al.; AAAI 2004] 16

Experiments 17

Experiments 2,200 sentences (Randomly sampled from Amazon.com and Cnet.com reviews for 11 products, 200 per product) Subset of dataset released by [Hu and Liu; SIGKDD 2004] 17

Experiments 2,200 sentences (Randomly sampled from Amazon.com and Cnet.com reviews for 11 products, 200 per product) Subset of dataset released by [Hu and Liu; SIGKDD 2004] Support Vector Machine classifier, linear kernel 17

Experiments 2,200 sentences (Randomly sampled from Amazon.com and Cnet.com reviews for 11 products, 200 per product) Subset of dataset released by [Hu and Liu; SIGKDD 2004] Support Vector Machine classifier, linear kernel Chi-squared feature selection used significant features at α = 0.05 17

Evaluation 11-fold cross-validation, sentences for each product used in test fold once Reporting average accuracy, Cohen s kappa 18

Baselines 19

Baselines Several standard feature sets ngrams (n = 1, 2, 3) Part-of-Speech ngrams (n = 2, 3) Dependency relations (no back-off) 19

Baselines Several standard feature sets ngrams (n = 1, 2, 3) Part-of-Speech ngrams (n = 2, 3) Dependency relations (no back-off) Back-off ngrams (n = 2, 3) similar to [McDonald et al.; ACL 2007] back off words in an ngram to POS tags create features using all possible (2 n -1) back-off combinations 19

Results Feature Set Accuracy Kappa Unigrams 0.652 (±0.048) 0.295 (±0.049) Uni.+Bigrams 0.657 (±0.066) 0.304 (±0.089) Uni.+Trigrams 0.655 (±0.062) 0.306 (±0.077) Uni.+Back-off Bigrams 0.650 (±0.056) 0.299 (±0.079) Uni.+Back-off Trigrams 0.647 (±0.051) 0.287 (±0.075) Uni.+POS Bigrams 0.676 (±0.057) 0.349 (±0.083) Uni.+POS Trigrams 0.661 (±0.050) 0.317 (±0.064) Uni.+Dep. Lex 0.639 (±0.055) 0.268 (±0.079) Uni.+Dep. Head-Back-off 0.679 (±0.063) 0.351 (±0.097) Uni.+Dep. Mod-Back-off 0.657 (±0.056) 0.308 (±0.063) Uni.+Dep. Head-Mod-Back-off 0.670 (±0.046) 0.336 (±0.065) 20

Results Feature Set Accuracy Kappa Unigrams 0.652 (±0.048) 0.295 (±0.049) Uni.+Bigrams 0.657 (±0.066) 0.304 (±0.089) Uni.+Trigrams 0.655 (±0.062) 0.306 (±0.077) Uni.+Back-off Bigrams 0.650 (±0.056) 0.299 (±0.079) Uni.+Back-off Trigrams 0.647 (±0.051) 0.287 (±0.075) Uni.+POS Bigrams 0.676 (±0.057) 0.349 (±0.083) Uni.+POS Trigrams 0.661 (±0.050) 0.317 (±0.064) Uni.+Dep. Lex 0.639 (±0.055) 0.268 (±0.079) Uni.+Dep. Head-Back-off 0.679 (±0.063) 0.351 (±0.097) Uni.+Dep. Mod-Back-off 0.657 (±0.056) 0.308 (±0.063) Uni.+Dep. Head-Mod-Back-off 0.670 (±0.046) 0.336 (±0.065) 20

Results Feature Set Accuracy Kappa Unigrams 0.652 (±0.048) 0.295 (±0.049) Uni.+Bigrams 0.657 (±0.066) 0.304 (±0.089) Uni.+Trigrams 0.655 (±0.062) 0.306 (±0.077) Uni.+Back-off Bigrams 0.650 (±0.056) 0.299 (±0.079) Uni.+Back-off Trigrams 0.647 (±0.051) 0.287 (±0.075) Uni.+POS Bigrams 0.676 (±0.057) 0.349 (±0.083) Uni.+POS Trigrams 0.661 (±0.050) 0.317 (±0.064) Uni.+Dep. Lex 0.639 (±0.055) 0.268 (±0.079) Uni.+Dep. Head-Back-off 0.679 (±0.063) 0.351 (±0.097) Uni.+Dep. Mod-Back-off 0.657 (±0.056) 0.308 (±0.063) Uni.+Dep. Head-Mod-Back-off 0.670 (±0.046) 0.336 (±0.065) 20

Results Feature Set Accuracy Kappa Unigrams 0.652 (±0.048) 0.295 (±0.049) Uni.+Bigrams 0.657 (±0.066) 0.304 (±0.089) Uni.+Trigrams 0.655 (±0.062) 0.306 (±0.077) Uni.+Back-off Bigrams 0.650 (±0.056) 0.299 (±0.079) Uni.+Back-off Trigrams 0.647 (±0.051) 0.287 (±0.075) Uni.+POS Bigrams 0.676 (±0.057) 0.349 (±0.083) Uni.+POS Trigrams 0.661 (±0.050) 0.317 (±0.064) Uni.+Dep. Lex 0.639 (±0.055) 0.268 (±0.079) Uni.+Dep. Head-Back-off 0.679 (±0.063) 0.351 (±0.097) Uni.+Dep. Mod-Back-off 0.657 (±0.056) 0.308 (±0.063) Uni.+Dep. Head-Mod-Back-off 0.670 (±0.046) 0.336 (±0.065) 20

Results Feature Set Accuracy Kappa Unigrams 0.652 (±0.048) 0.295 (±0.049) Uni.+Bigrams 0.657 (±0.066) 0.304 (±0.089) Uni.+Trigrams 0.655 (±0.062) 0.306 (±0.077) Uni.+Back-off Bigrams 0.650 (±0.056) 0.299 (±0.079) Uni.+Back-off Trigrams 0.647 (±0.051) 0.287 (±0.075) Uni.+POS Bigrams 0.676 (±0.057) 0.349 (±0.083) Uni.+POS Trigrams 0.661 (±0.050) 0.317 (±0.064) Uni.+Dep. Lex 0.639 (±0.055) 0.268 (±0.079) Uni.+Dep. Head-Back-off 0.679 (±0.063) 0.351 (±0.097) Uni.+Dep. Mod-Back-off 0.657 (±0.056) 0.308 (±0.063) Uni.+Dep. Head-Mod-Back-off 0.670 (±0.046) 0.336 (±0.065) 20

Results Feature Set Accuracy Kappa Unigrams 0.652 (±0.048) 0.295 (±0.049) Uni.+Bigrams 0.657 (±0.066) 0.304 (±0.089) Uni.+Trigrams 0.655 (±0.062) 0.306 (±0.077) Uni.+Back-off Bigrams 0.650 (±0.056) 0.299 (±0.079) Uni.+Back-off Trigrams 0.647 (±0.051) 0.287 (±0.075) Uni.+POS Bigrams 0.676 (±0.057) 0.349 (±0.083) Uni.+POS Trigrams 0.661 (±0.050) 0.317 (±0.064) Uni.+Dep. Lex 0.639 (±0.055) 0.268 (±0.079) Uni.+Dep. Head-Back-off 0.679 (±0.063) 0.351 (±0.097) Uni.+Dep. Mod-Back-off 0.657 (±0.056) 0.308 (±0.063) Uni.+Dep. Head-Mod-Back-off 0.670 (±0.046) 0.336 (±0.065) 20

Discussion Head-Back-off features significantly better than unigrams-only baseline represent a better way to use dependency relations Generalizing them further (full back-off features) worsens performance 21

Head-Back-off Successes It is perhaps too small. cop(jj,is) The lens retracts and has its own metal cover so you don't need to fuss with a lens cap. det(nn,the), poss(nn,its), neg(vb,n t) The panorama setting is unbelievable! cop(jj,is), det(nn,the) 22

Head-Back-off Successes Even with the waterproof housing it is small. cop(jj,is), nsubj(jj,it), det(nn,the) The auto color balance is often fooled by the clouds. det(nn,the) First, I have to say that I have NEVER had the slightest problem with this camera or the software. det(nn,this) 23

Error Analysis Sentences that are opinions, but not about the product s features Really like them, they work well and the macro function of the 2500 really helps my Ebay biz. Really like them is about two other cameras the reviewer owns, not the camera being reviewed. I have had this little gem for four months now. About the product as a whole (not any particular explicitly mentioned feature) 24

Error Analysis Few misclassifications due to Head-Back-off features misfiring Some people, in their reviews, complain about its small size, and how it doesn t compare with larger cameras. Misclassified as opinion sentence poss(nn,its), neg(vb,n t) The closest competitor is the SONY DSC-P1 (3.3mp). Misclassified as opinion sentence cop(nn,is), det(nn,the) 25

Conclusions Head-Back-off features are a sweet spot not too specific, not too general Future work: Explore relation to supervised domain adaptation 26

Questions? 27