Rumor Detection on Twitter with Tree-structured Recursive Neural Networks

Similar documents
Motivation: Attention: Focusing on specific parts of the input. Inspired by neuroscience.

Learning Convolutional Neural Networks for Graphs

Inferring Clinical Correlations from EEG Reports with Deep Neural Learning

Automatic Classification of Perceived Gender from Facial Images

A Convolutional Approach for Misinformation Identification

Beyond R-CNN detection: Learning to Merge Contextual Attribute

An Artificial Neural Network Architecture Based on Context Transformations in Cortical Minicolumns

Predicting Breast Cancer Survivability Rates

A HMM-based Pre-training Approach for Sequential Data

READ-BIOMED-SS: ADVERSE DRUG REACTION CLASSIFICATION OF MICROBLOGS USING EMOTIONAL AND CONCEPTUAL ENRICHMENT

Formulating Emotion Perception as a Probabilistic Model with Application to Categorical Emotion Classification

On Shape And the Computability of Emotions X. Lu, et al.

DeepASL: Enabling Ubiquitous and Non-Intrusive Word and Sentence-Level Sign Language Translation

Discovering and Understanding Self-harm Images in Social Media. Neil O Hare, MFSec Bucharest, Romania, June 6 th, 2017

Memory-Augmented Active Deep Learning for Identifying Relations Between Distant Medical Concepts in Electroencephalography Reports

基于循环神经网络的序列推荐 吴书智能感知与计算研究中心模式识别国家重点实验室中国科学院自动化研究所. Spe 17, 中国科学院自动化研究所 Institute of Automation Chinese Academy of Sciences

arxiv: v1 [stat.ml] 23 Jan 2017

BLADDERSCAN PRIME PLUS TM DEEP LEARNING

More Examples and Applications on AVL Tree

Predicting Breast Cancer Recurrence Using Machine Learning Techniques

M.Sc. in Cognitive Systems. Model Curriculum

Human Activities: Handling Uncertainties Using Fuzzy Time Intervals

POC Brain Tumor Segmentation. vlife Use Case

3D Deep Learning for Multi-modal Imaging-Guided Survival Time Prediction of Brain Tumor Patients

EECS 433 Statistical Pattern Recognition

Introduction to Sentiment Analysis

A deep learning model integrating FCNNs and CRFs for brain tumor segmentation

DIABETIC RISK PREDICTION FOR WOMEN USING BOOTSTRAP AGGREGATION ON BACK-PROPAGATION NEURAL NETWORKS

Exploiting Ordinality in Predicting Star Reviews

TITLE: A Data-Driven Approach to Patient Risk Stratification for Acute Respiratory Distress Syndrome (ARDS)

Analyzing Spammers Social Networks for Fun and Profit

Deep Learning based Information Extraction Framework on Chinese Electronic Health Records

Variable Features Selection for Classification of Medical Data using SVM

Convolutional Neural Networks for Text Classification

Deep Learning Analytics for Predicting Prognosis of Acute Myeloid Leukemia with Cytogenetics, Age, and Mutations

Brain Tumor segmentation and classification using Fcm and support vector machine

Algorithms in Nature. Pruning in neural networks

Detection of Glaucoma and Diabetic Retinopathy from Fundus Images by Bloodvessel Segmentation

Mining Human-Place Interaction Patterns from Location-Based Social Networks to Enrich Place Categorization Systems

Confluence: Conformity Influence in Large Social Networks

Deep Learning for Computer Vision

Sentiment Analysis of Reviews: Should we analyze writer intentions or reader perceptions?

On the Use of Brainprints as Passwords

Causal Knowledge Modeling for Traditional Chinese Medicine using OWL 2

Enhanced Detection of Lung Cancer using Hybrid Method of Image Segmentation

Lung Cancer Diagnosis from CT Images Using Fuzzy Inference System

Case-based reasoning using electronic health records efficiently identifies eligible patients for clinical trials

Show the truth. Picture warnings save lives. WORLD NO TOBACCO DAY 31 MAY 2009

Enhance micro-blogging recommendations of posts with an homophily-based graph

Cardiac Arrest Prediction to Prevent Code Blue Situation

Data Mining in Bioinformatics Day 9: String & Text Mining in Bioinformatics

Shu Kong. Department of Computer Science, UC Irvine

A Novel Capsule Neural Network Based Model For Drowsiness Detection Using Electroencephalography Signals

IDENTIFYING STRESS BASED ON COMMUNICATIONS IN SOCIAL NETWORKS

Predicting Sleep Using Consumer Wearable Sensing Devices

Audiovisual to Sign Language Translator

Multi Parametric Approach Using Fuzzification On Heart Disease Analysis Upasana Juneja #1, Deepti #2 *

Predicting Potential Domestic Violence Re-offenders Using Machine Learning. Rajhas Balaraman Supervisor : Dr. Timothy Graham

PathGAN: Visual Scanpath Prediction with Generative Adversarial Networks

arxiv: v2 [cs.cv] 19 Dec 2017

CSE 258 Lecture 1.5. Web Mining and Recommender Systems. Supervised learning Regression

A Deep Learning Approach to Identify Diabetes

Data mining for Obstructive Sleep Apnea Detection. 18 October 2017 Konstantinos Nikolaidis

Brain Tumor Detection using Watershed Algorithm

Identifying Adverse Drug Events from Patient Social Media: A Case Study for Diabetes

Classification of benign and malignant masses in breast mammograms

Supplementary Online Content

Medical Knowledge Attention Enhanced Neural Model. for Named Entity Recognition in Chinese EMR

ARTERY/VEIN CLASSIFICATION IN FUNDUS IMAGES USING CNN AND LIKELIHOOD SCORE PROPAGATION

A model to explain the emergence of reward expectancy neurons using reinforcement learning and neural network $

PREDICTION OF BREAST CANCER USING STACKING ENSEMBLE APPROACH

UNIVERSITY of PENNSYLVANIA CIS 520: Machine Learning Final, Fall 2014

Asthma Surveillance Using Social Media Data

Deep Diabetologist: Learning to Prescribe Hypoglycemia Medications with Hierarchical Recurrent Neural Networks

Social Network Data Analysis for User Stress Discovery and Recovery

Rich feature hierarchies for accurate object detection and semantic segmentation

AUTOMATED DEBUGGING. Session Leader: Zongxu Mu

A New Monotony Advanced Decision Tree Using Graft Algorithm to Predict the Diagnosis of Diabetes Mellitus

Brain Tumor Segmentation Based On a Various Classification Algorithm

Consumer Review Analysis with Linear Regression

PSYCHOLOGICAL STRESS DETECTION FROM CROSS-MEDIA MICROBLOG DATA USING DEEP SPARSE NEURAL NETWORK

Agent-Based Models. Maksudul Alam, Wei Wang

EXTRACTION OF RETINAL BLOOD VESSELS USING IMAGE PROCESSING TECHNIQUES

Feature selection methods for early predictive biomarker discovery using untargeted metabolomic data

Deep Learning Models for Time Series Data Analysis with Applications to Health Care

Classıfıcatıon of Dıabetes Dısease Usıng Backpropagatıon and Radıal Basıs Functıon Network

Recognition of sign language gestures using neural networks

Evaluating Classifiers for Disease Gene Discovery

Predicting Breast Cancer Survival Using Treatment and Patient Factors

Patient Subtyping via Time-Aware LSTM Networks

Microblog Retrieval for Disaster Relief: How To Create Ground Truths? Ribhav Soni and Sukomal Pal

Extraction and Identification of Tumor Regions from MRI using Zernike Moments and SVM

Sound Texture Classification Using Statistics from an Auditory Model

The Good News. More storage capacity allows information to be saved Economic and social forces creating more aggregation of data

An Improved Algorithm To Predict Recurrence Of Breast Cancer

Shu Kong. Department of Computer Science, UC Irvine

Neuromorphic convolutional recurrent neural network for road safety or safety near the road

Computational modeling of visual attention and saliency in the Smart Playroom

to Cues Present at Test

Transcription:

1 Rumor Detection on Twitter with Tree-structured Recursive Neural Networks Jing Ma 1, Wei Gao 2, Kam-Fai Wong 1,3 1 The Chinese University of Hong Kong 2 Victoria University of Wellington, New Zealand 3 MoE Key Laboratory of High Confidence Software Technologies, China July 15-20, 2018 ACL 2018 @ Melboume, Australia

Outline Introduction Related Work Problem Statement RvNN-based Rumor Detection Evaluation Conclusion and Future Work 2

Introduction What are rumors? A story or statement whose truth value is unverified or deliberately false 3

Introduction How the fake news propagated? denial Ø people tend to stop spreading a rumor if it is known as false. (Zubiaga et al., 2016b) Ø Previous studies focused on text mining from sequential microblog streams, we want to bridge the content semantics and propagation clues. supportive 4 2016/7/1

Motivation We generally are not good at distinguishing rumors It is crucial to track and debunk rumors early to minimize their harmful effects. Online fact-checking services have limited topical coverage and long delay. Existing models use feature engineering over simplistic; or recently deep neural networks ignore propagation structures; Kernel-based method develop based on tree structure but cannot learn high-level feature representations automatically. 5

Observation & Hypothesis Existing works: Consider post representation or propagation structure Support Neutral Doubt (a) RNN-based model (Ma et al. 2016) (b) Tree kernel-based model (Ma et al. 2017) IDEA: Combining the two models, leveraging propagation structure by representation learning algorithm 6

Observation & Hypothesis Why such model do better? (a) False rumor Polarity stances (b) True rumor Local characteristic: A reply usually respond to its immediate ancestor rather than the root tweet. Repliers tend to disagree with (or question) who support a false rumor or deny a true rumor; repliers tend to agree with who deny a false rumor or support a true rumor. 7

Contributions The first study that deeply integrates both structure and content semantics based on tree-structured recursive neural networks for detecting rumors from microblog posts Propose two variants of RvNN models based on bottom-up and top-down tree structures, to generate better integrated representations for a claim by capturing both structural and textural properties signaling rumors. Our experiments based on two real-world Twitter datasets achieve superior improvements over state-of-the-art baselines on both rumor classification and early detection tasks. We make the source codes in our experiments publicly accessible at https://github.com/majingcuhk/rumor_rvnn 8

Outline Introduction Related Work Problem Statement RvNN-based Rumor Detection Evaluation Conclusion and Future Work 9

Related Work Systems based on common sense and investigative journalism, e.g., snopes.com factcheck.org Learning-based models for rumor detection Information credibility: Castillo et al. (2011), Yang et al. (2012) Using handcrafted and temporal features: Liu et al. (2015), Ma et al. (2015), Kwon et al. (2013, 2017) Using cue terms: Zhao et al. (2015) Using recurrent neural networks: Ma et al. (2016, 2018) Tree-kernel-based model: Ma et al. (2017), Wu et al. (2015) RvNN-based works images segmentation (Socher et al, 2011) phrase representation from word vectors (Socher et al, 2012) Sentiment analysis (Socher et al, 2013) etc Without handcrafted features 10

Outline Introduction Related Work Problem Statement RvNN-based Rumor Detection Evaluation Conclusion and Future Work 11

Problem Statement Given a set of microblog posts R = {r}, model each source tweet as a tree structure T r = < V, E >, where each node v provide the text content of each post. And E is directed edges corresponding to response relation. Task 1 finer-grained classification for each source post false rumor, true rumor, non-rumor, unverified rumor Task 2 detect rumor as early as possible 12

Tweet Structure Root tweet bottom-up tree x 1 : #Walmart donates $10,000 to #DarrenWilson fund to continue police racial profiling x 2 : 1:30 Idc if they killed a mf foreal. Ima always shop with @Walmart. I'm just bein honest x 3 : NEED SOURCE. have a feeling this is just hearsay... replies x 4 : I agree. I have been hearing this all day but no source 1:12 x 5 : Exactly, i don't think Wal-Mart would let everyone know this if they did!! 2:21 top-down tree x 1 : #Walmart donates $10,000 to #DarrenWilson fund to continue police racial profiling x 2 : 1:30 Idc if they killed a mf foreal. Ima always shop with @Walmart. I'm just bein honest x 3 : NEED SOURCE. have a feeling this is just hearsay... 13 x 4 : I agree. I have been hearing this all day but no source 1:12 x 5 : Exactly, i don't think Wal-Mart would let everyone know this if they did!! 2:21

Outline Introduction Related Work Problem Statement RvNN-based Rumor Detection Evaluation Conclusion and Future Work 14

Standard Recursive Neural Networks RvNN (tree-structured neural networks) utilize sentence parse trees: representation associated with each node of a parse tree is computed from its direct children, computed by p = f(w 9 c ; ; c = + b) p: the feature vector of a parent node whose children are c ; and c = computation is done recursively over all tree nodes 15

Bottom-up RvNN Ø Input: bottom-up tree (node: a post represented as a vector of words ) Ø Structure: recursively visit every node from the leaves at the bottom to the root at the top (a natural extension to the original RvNN ) Ø Intuition: local rumor indicative features are aggregated along different branches (e.g., subtrees having a denial parent and a set of supportive children) (generate a feature vector for each subtree) Ø GRU equation at node j Children node Own input x 1 : #Walmart donates $10,000 to #DarrenWilson fund to continue police racial profiling x 2 : 1:30 Idc if they killed a mf foreal. Ima always shop with @Walmart. I'm just bein honest x 3 : NEED SOURCE. have a feeling this is just hearsay... x 4 : I agree. I have been hearing this all day but no source 1:12 x 5 : Exactly, i don't think Wal-Mart would let everyone know this if they did!! 2:21 16

Top-down RvNN Ø Input: top-down tree Ø Structure: recursively visit from the root node to its children until reaching all leaf nodes. (reverse Bottom-up RvNN) Ø GRU transition equation at node j Own input Parent node Ø Intuition: rumor-indicative features are aggregated along the propagation path (e.g., if a post agree with its parent s stance, the parent s stance should be reinforced) (models how information flows from source post to the current node) x 1 : #Walmart donates $10,000 to #DarrenWilson fund to continue police racial profiling x 2 : 1:30 Idc if they killed a mf foreal. Ima always shop with @Walmart. I'm just bein honest x 3 : NEED SOURCE. have a feeling this is just hearsay... x 4 : I agree. I have been hearing this all day but no source 1:12 x 5 : Exactly, i don't think Wal-Mart would let everyone know this if they did!! 2:21 17

Model Training Ø Comparison: both of the two RvNN models aim to capture the structural properties by recursively visiting all nodes Bottom-up RvNN: the state of root node (i.e., source tweet) can be regard as the representation of the whole tree (can be used for supervised classification). Top-down RvNN: the representation of each path are eventually embedded into the hidden vector of all the leaf nodes. learned vector of root node Ø Output Layer Bottom-up RvNN: y = Softmax Vh K + b Top-down RvNN: y = Softmax Vh L + b the pooling vector over all leaf nodes Ø Objective Function: L = T = = US; y O yq O λ Θ = R OS; + prediction Ground truth Ø Training Procedure parameters are updated using efficient back-propagation through structure (Goller and Kuchler, 1996; Socher et al., 2013) 18

Outline Introduction Related Work Problem Statement RvNN-based Rumor Detection Evaluation Conclusion and Future Work 19

Data Collection Use two reference Tree datasets: URL of the datasets: https://www.dropbox.com/s/0jhsfwep3ywvpca/rumdetect2017.zip?dl=0 20

Approaches to compare with DTR: Decision tree-based ranking model using enquiry phrases to identify trending rumors (Zhao et al., 2015) DTC: Twitter information credibility model using Decision Tree Classifier (Castillo et al., 2011); RFC: Random Forest Classifier using three parameters to fit the temporal tweets volume curve (Kwon et al., 2013) SVM-TS: Linear SVM classifier using time-series structures to model the variation of social context features. (Ma et al., 2015) SVM-BOW: linear SVM classifier using bag-of-words. SVM-TK and SVM-HK: SVM classifier uses a Tree Kernel (Ma et al., 2017) and that uses a Hybrid Kernel (Wu et al., 2015), both model propagation structures with kernels. GRU-RNN: The RNN-based rumor detection model. (Ma et al., 2016) Ours (BU-RvNN and TD-RvNN): Our bottom-up and topdown recursive models. 21

Results on Twitter15 NR: Non-Rumor; FR: False Rumor; TR: True Rumor; UR: Unverified Rumor; Method Accu. NR FR TR UR F1 F1 F1 F1 DTR 0.409 0.501 0.311 0.364 0.473 DTC 0.454 0.733 0.355 0.317 0.415 RFC 0.565 0.810 0.422 0.401 0.543 SVM-TS 0.544 0.796 0.472 0.404 0.483 SVM-BOW 0.548 0.564 0.524 0.582 0.512 SVM-HK 0.493 0.650 0.439 0.342 0.336 SVM-TK 0.667 0.619 0.669 0.772 0.645 GRU-RNN 0.641 0.684 0.634 0.688 0.571 BU-RvNN 0.708 0.695 0.728 0.759 0.653 TD-RvNN 0.723 0.682 0.758 0.821 0.654 hand-crafted features (e.g., user info NR vs others) Structural info Linear chain input More info loss 22

Results on Twitter16 NR: Non-Rumor; FR: False Rumor; TR: True Rumor; UR: Unverified Rumor; Method Accu. NR FR TR UR F1 F1 F1 F1 DTR 0.414 0.394 0.273 0.630 0.344 DTC 0.465 0.643 0.393 0.419 0.403 RFC 0.585 0.752 0.415 0.547 0.563 SVM-TS 0.574 0.755 0.420 0.571 0.526 SVM-BOW 0.585 0.553 0.556 0.655 0.578 SVM-HK 0.511 0.648 0.434 0.473 0.451 SVM-TK 0.662 0.643 0.623 0.783 0.655 GRU-RNN 0.633 0.617 0.715 0.577 0.527 BU-RvNN 0.718 0.723 0.712 0.779 0.659 TD-RvNN 0.737 0.662 0.743 0.835 0.708 models without hand-crafted features 23

Results on Early Detection (a) Twitter15 DATASET In the first few hours, the accuracy of the RvNNbased methods climbs more rapidly and stabilize more quickly TD-RvNN and BURvNN only need around 8 hours or about 90 tweets to achieve the comparable performance of the best baseline model. (b) Twitter16 DATASET 24

Early Detection Example Example subtree of a rumor captured by the algorithm at early stage of propagation Ø Bottom-up RvNN: a set of responses supporting the parent posts that deny or question the source post. Ø Top-down RvNN: some patterns of propagation from the root to leaf nodes like support deny support Ø Baselines: sequential models may be confused because the supportive key terms such as be right, yeah, exactly! dominate the responses, and the SVM-TK may miss similar subtrees by just comparing the surface words. 25

Outline Introduction Related Work Problem Statement RvNN-based Rumor Detection Evaluation Conclusion and Future Work 26

Conclusion and future work Propose a bottom-up and a top-down tree-structured model based on recursive neural networks for rumor detection on Twitter. Using propagation tree to guide the learning of representations from tweets content, such as embedding various indicative signals hidden in the structure, for better identifying rumors. Results on two public Twitter datasets show that our method improves rumor detection performance in very large margins as compared to state-of-the-art baselines. Future work: Integrate other types of information such as user properties into the structured neural models to further enhance representation learning Develop unsupervised models due to massive unlabeled data from social media. 27

Q & A 28