Speech Project

Week 10 Report

b02901085   徐瑞陽

b02901054   方為

Task

  • Task 5 from SemEval 2016
  • SemEval (Semantic Evaluation):
    • an ongoing series of evaluations of computational semantic analysis systems

Task 5:

Aspect Based Sentiment Analysis (ABSA)

  • Rise of e-commerce leads to growth of review sites for a variety of services and products
  • ABSA: mining opinions from text about specific entities and their aspects

ABSA

  • Given a target of interest (e.g., Apple Mac mini), summarize the content of reviews in an aspect-sentiment table

ABSA

  • Aspect:  entity type (E) and an attribute type (A) pair, denoted as {E#A}
  • "E" can be:
    • the reviewed entity itself (ex. laptop)
    • a part or component (ex. battery, customer service)
    • other relevant entities (ex. manufacturer of laptop)
  • "A": a particular attribute (ex. durability, portability)
  • "E" and "A" do not necessarily occur in the review
    • ex: “They sent it back with a huge crack in it and it still didn't work; and that was the fourth time I’ve sent it to them to get fixed” 
      -----> {customer_support#quality}

 

Domains and Languages

  • Domains & Languages: 
    • Restaurants: English, Dutch, French, Russian, Spanish, Turkish
    • Hotels: English, Arabic
    • Consumer Electronics: 
      • Laptops: English
      • Mobile Phones: Chinese, Dutch
      • Digital Cameras: Chinese
    • Telecommunications: Turkish

 

Subtask 1: Sentence-level ABSA

Given a review text about a target entity (laptop, restaurant, etc.),

identify the following information:

  • Slot 1: Aspect Category
    • ex. ''It is extremely portable and easily connects to WIFI at the library and elsewhere''
      ----->{LAPTOP#PORTABILITY}, {LAPTOP#CONNECTIVITY}
  • Slot 2: Opinion Target Expression (OTE)
    • ​an expression used in the given text to refer to the reviewed E#A
    • ​ex. ''The fajitas were delicious, but expensive''
      ----->{FOOD#QUALITY, “fajitas”}, {FOOD#PRICES, “fajitas”}
  • Slot 3: Sentiment Polarity
    • ​label: (positive, negative, or neutral) 

 

Subtask 2: Text-level ABSA

Given a set of customer reviews about a target entity (ex. a restaurant), identify a set of {aspect#category, polarity} tuples that summarize the opinions expressed in each review.

Subtask 3: Out-of-domain ABSA

Test system in a previously unseen domain (hotel reviews in SemEval 2015) for which no training data was made available. The gold annotations for Slots 1 and 2 were provided and the teams had to return the sentiment polarity values (Slot 3).

Our Framework

Framework 1

Encode by Tree-LSTM or autoencoder

Framework 2

End-to-end MemNN

Paper Study

Memory Networks

1.\quad x \Rightarrow I(x)
1.xI(x) 1.\quad x \Rightarrow I(x)
2. \quad update~memories: \quad m_i = G(m_i,I(x),m), \forall i
2.update memories:mi=G(mi,I(x),m),i2. \quad update~memories: \quad m_i = G(m_i,I(x),m), \forall i
3. \quad o = O(I(x),m)
3.o=O(I(x),m)3. \quad o = O(I(x),m)
4. \quad r = R(o)
4.r=R(o)4. \quad r = R(o)

Memory Networks

  • I component:
    • Standard preprocessing (ex. parsing)
    • Could also encode into feature representation
  • G component:
    • Simple implementation: store I(x) in a slot in the memory
    • For huge memory: slot choosing function H
  • O and R components:
    • O: Reading from memory and performing inference
    • R: Produce final response given O
      • ex. R could be implemented by an RNN
  • IGRO are pre-trained NN => MemNN

Memory Networks

  • MemNN experiments
    • Large-scale QA
      • statements (subject, relation, object): ex. (sheep, be-afraid-of, wolf)
      • question: ex. "What is sheep afraid of?"
      • Good performance
      • Implemented hashing techniques for significant speedups
    • Simulated World QA
      • Joe went to the kitchen. Fred went to the kitchen. Joe picked up the milk. Joe travelled to the office. Joe left the milk. Joe went to the bathroom.
        • Where is the milk now? A: office
        • Where is Joe? A: bathroom
        • Where was Joe before the office? A: kitchen
      • Performs better on "before" questions and "actor+object" questions compared to RNNs and LSTMs
        • ​can store memory for longer period of time

End-to-End Memory Networks

  • When training        ,        , in previous model ,

      we need to know  the mapping
      between input sentence and best response

      => Hard  in real-world

s_O~s_R
sO sRs_O~s_R
  • This model accept whole facts ,query in one time ,
    just have one answer ,
    and it model the inference process as RNN (memory hop)
    to combine internal state and output in previous hop
s_O
sOs_O

is the scoring function for query and memory

s_R
sRs_R

is the scoring function for related memory and response

End-to-End Memory Networks

  • Modifying RNN's param is like find 
    but it doesn't need to find argmax directly but  softmax ,
    use this probability to update next hop hidden state
s_O~s_R
sO sRs_O~s_R
  • When training , it will backpropogate automatically , although there are many memory hop (inference step)
  • If well-trained , the ouput of each hop shows the inference step

Problem we encountered

We have no idea
how to achieve 
Opinion Target Expression (OTE)

It's not a classification problem...

SpeechProject-week10

By sunprinces

SpeechProject-week10

  • 546