Cross-Lingual Sentiment Analysis using modified BRAE

15
Komachi Lab M1 Ryosuke Miyazaki 2015/10/16 Cross-Lingual Sentiment Analysis using modified BRAE Sarthak Jain and Shashank Batra EMNLP 2015 EMNLP 2015 reading group All figures in this slide are cited from original paper

Transcript of Cross-Lingual Sentiment Analysis using modified BRAE

Page 1: Cross-Lingual Sentiment Analysis using modified BRAE

Komachi Lab

M1 Ryosuke Miyazaki

2015/10/16

Cross-Lingual Sentiment Analysis using modified BRAE Sarthak Jain and Shashank Batra EMNLP 2015

EMNLP 2015 reading group

※ All figures in this slide are cited from original paper

Page 2: Cross-Lingual Sentiment Analysis using modified BRAE

Komachi Lab

Abstract

✤ To perform Cross Lingual Sentiment Analysis

- They use parallel corpus that include resource rich (English) and resource poor (Hindi)

✤ They create new Movie Reviews Dataset in Hindi for evaluation

✤ Their model significantly outperforms state of the art, especially when labeled data is scarce

2

Page 3: Cross-Lingual Sentiment Analysis using modified BRAE

Komachi Lab

Model and Training

3

Page 4: Cross-Lingual Sentiment Analysis using modified BRAE

Komachi Lab

BRAE Model

4

Bilingually Constrained Recursive Auto-encoder

First, we consider standard Recursive Auto-encoder for each language respectively

construct parent vector reconstruct children vector

Minimize reconstruction errors (Euclidean distance)

c: child vectory, p: parent vector

Page 5: Cross-Lingual Sentiment Analysis using modified BRAE

Komachi Lab

BRAE Model

5

Loss Function

They also produce representation from another language

AssumptionA phrase and its correct translation should share the same semantic meaning

Loss Function about source language

Transforming loss

Like wise, they define for target language

Objective function

Page 6: Cross-Lingual Sentiment Analysis using modified BRAE

Komachi Lab

Training (Unsupervised)✤ Word embeddings are pre-trained by Word2Vec

✤ 1st: Pre-train ps, and pt respectively on RAE

6

✤ 2nd: Fix pt and train ps on BRAE

- Vice-versa for ps

- Set ps = p’s, pt = p’t when it reaching a local minima.

Page 7: Cross-Lingual Sentiment Analysis using modified BRAE

Komachi Lab

Training (Supervise)

✤ Modification for Classifying Sentiment✤ Adding Softmax and Cross entropy error functions

to only source language (resource rich language)✤ In this phase, penalty term is included in reconstruction error

7

✤ And, transformation weights (θts, θst) are not updated in this phase

Page 8: Cross-Lingual Sentiment Analysis using modified BRAE

Komachi Lab

Training (Supervise)✤ 1st: only update resource rich related parameters

8

ce: cross entropy

✤ 2nd: only update resource poor related parameters- Since the gold labels are only associated with resource rich,

they use transformation to obtain sentiment distribution

✤ Predict overall sentiment associated with the resource poor- concat pt, p’s then

train by softmax regression using weight matrix

Page 9: Cross-Lingual Sentiment Analysis using modified BRAE

Komachi Lab

Experiments

9

Page 10: Cross-Lingual Sentiment Analysis using modified BRAE

Komachi Lab

Experimental Settings

✤ HindMonoCorp 0.5 (44.49M sentences) and English Gigaword Corpus for word embeddings

✤ Bilingual sentence-aligned data from HindEnCrop (273.9k sentence pairs)

10

For Unsupervised phase

For Supervised phase (use MOSES to obtain bilingual phrase pairs)

✤ IMDB11 dataset (25000 pos, 25000 neg)

✤ Rotten Tomatoes Review dataset (4 documents, {0, 1, 2, 3})

✤ Their model was able to correctly infer word sense for polysemous words

Page 11: Cross-Lingual Sentiment Analysis using modified BRAE

Komachi Lab

Experimental Setting

✤ Rating Based Hindi Movie Review Dataset (2945 movie reviews, {1, 2, 3, 4}) they create this new dataset for evaluation

✤ Standard Movie Reviews Dataset (125 positive, 125 negative)

11

Evaluation Data set

✤ learning rate: 0.05

✤ word vector dimension: 80

✤ joint error of BRAE (α): 0.2

✤ λL: 0.001

✤ λBRAE: 0.0001

Tuning by Grid Search on Cross Validation✤ κ: 0.2, η: 0.35

✤ λp: 0.01

✤ λS: 0.1

✤ λT: 0.04

Page 12: Cross-Lingual Sentiment Analysis using modified BRAE

Komachi Lab

Results

✤ BRAE-U: neither include penalty term, nor fix the transformations weights

✤ BRAE-P: only include the penalty term

✤ BRAE-F: include both term

12

monolingual

cross lingual

monolingual

monolingual

monolingualcross lingualcross lingualcross lingual Confusion matrix (BRAE-F)

Page 13: Cross-Lingual Sentiment Analysis using modified BRAE

Komachi Lab

Results

13

Accuracy with amount of labeled training data used

✤ Their model achieve best performance even though data are 50% less than those of others.

Accuracy with amount of unlabeled training data used

Page 14: Cross-Lingual Sentiment Analysis using modified BRAE

Komachi Lab

Analysis✤ Since the movement in semantic vector space was restricted, their

model have an advantage about unknown words

14

“Her acting of a schizophrenic mother made our hearts weep”

base line classify as negative due to “weep”, but their model correctly predict positive

Example:

✤ Their model was able to correctly infer word sense for polysemous words

Page 15: Cross-Lingual Sentiment Analysis using modified BRAE

Komachi Lab

Error Analysis

✤ conflicting sentiments about two different aspects about the same object

✤ presence of subtle contextual references

15

Difficult situation

✤ “His poor acting generally destroys a movie, but this time it didn’t”

- correct is positive, predict rate is 2

✤ “This movie made his last one looked good”

- wrong prediction of rating 3

Example of latter case