SlideShare uma empresa Scribd logo
1 de 21
Human Interface Laboratory
Pay Attention to Categories:
Syntax-Based Sentence Modeling with
Metadata Projection Matrix
2020. 10. 24, @PACLIC 34
Won Ik Cho, Nam Soo Kim (SNU ECE & INMC)
Contents
• Introduction
• Proposed Method
• Experiment and Result
• Visualization
• Conclusion
1
Introduction
• Brief overview on sentence modeling
 Sparse encoding to dense word vectors
2
Introduction
• Brief overview on sentence modeling
 Deep learning techniques and attention model
3
Introduction
• Brief overview on sentence modeling
 Self-attentive models (which advances self attention)
• Still a useful approach to sentence classification
4
Introduction
• Motivation
 What if we need to pay more attention to some syntactic categories and
want to decide the intensity automatically?
• e.g., Oxymoron detection (Cho et al., 2017)
– Drink a sugar-free sweet tea
– When I’m low, I get high
 Attention mechanism is useful, but less reliable in the control regarding
syntactic categories
• Not mainly considered since the necessity is underestimated
• Minimal information (e.g., POS) can be of help for some tasks!
• How can we take into account that information beyond just attaching it to each
token?
5
Related Work
• Word embedding
 Dense embedding of words, in view of distributional semantics
 Projecting words to low dimensional space, with the objective of ‘making
the distribution that enables the prediction on the probable surrounding
words’
 Variations of word2vec
• Word2Vec (Mikolov et al., 2013)
• GloVe (Pennington et al., 2014)
• fastText (Bojanovsky et al., 2016)
6
Related Work
• Deep learning techniques
 Convolutional neural networks
• Primarily applied to the vision area
• Used in Kim (2014) as 1D convolution that represents the sentence as a
sequence of dense word vectors
 Recurrent neural networks
• Adopted to model the sequential data
• Long short-term memory is used to cope with the vanishing gradient
• Bidirectional LSTM to consider the directivity
7
Related Work
• Deep learning techniques
 Attention models
• First used to deal with the word order sensitivity and matching in machine
translation
• Evolved to various format, such as location-based
 Self-attentive sentence embedding (Lin et al., 2017)
• Related to self attention, but more applicable to BiLSTM format
8
Related Work
• Deep learning techniques
 Self-attentive sentence embedding
9
Proposed Method
• Overall description
 Sequential word
embedding
and BiLSTM
 Feature extraction
for attention source
• TF-IDF? BiLSTM?
 Attention source
activated with ReLU
 PAC structure with:
• Weight layer with
category-wise info
• Projection matrix
• Multiplication (𝛼1
𝐿
)
10
Proposed Method
• Overall description
 Sequential word
embedding
and BiLSTM
 Feature extraction
for attention source
• TF-IDF? BiLSTM?
 Attention source
activated with ReLU
 PAC structure with:
• Weight layer with
category-wise info
• Projection matrix
• Multiplication (𝛼1
𝐿
)
11
Proposed Method
• Overall description
 Sequential word
embedding
and BiLSTM
 Feature extraction
for attention source
• TF-IDF? BiLSTM?
 Attention source
activated with ReLU
 PAC structure with:
• Weight layer with
category-wise info
• Projection matrix
• Multiplication (𝛼1
𝐿
)
12
Proposed Method
• Overall description
 PAC structure with:
• Weight layer with
category-wise info
• Projection matrix
• Multiplication (𝛼1
𝐿
)
13
Experiment and Result
• Implementation
 Baseline features
• TF-IDFs and bigrams
– Dictionary size set to 3,000 (=30 * 100)
• GloVe pretrained with Twitter 27B
– Word vector dim. 100
– Padding max length 30
 Baseline classifiers
• SVM for TF-IDF
• NN for GloVe (averaged) with Adam(0.0005) and batch size 16
• CNN (32 filters, window 3) and BiLSTM (hidden dim. 64) for Glove (padded)
 Baeseline attention model
• Lin et al. (2017) with context vector dim. 64, alongwith the above BiLSTM
 The proposed
• 𝑛 𝑝 follows the NLTK POS tagging result
14
Experiment and Result
• Dataset
 Metalanguage detection (2,393)
• Investigates whether a sentence contains explicit mention terms (‘title’ or ‘name’)
• Contains 629 mentioned and 1,764 not-mentioned instances excerpted from
Wikipedia
 Irony detection (4,618)
• Distributed in SemEval 2018 Task 3 for ironic tweet detection (Van Hee et al., 2018).
• Binary label case was taken into account; 2,222 contain irony and 2,396 do not
 Subjectivity detection (10,000)
• Refers to Pang and Lee (2004); checks if the movie review contains a subjective
judgment
• Incorporates equally 5,000 instances for each of the subjective and objective reviews
 Stance classification (3,835)
• Part of distributed dataset from SemEval 2016 Task 6 (Mohammad et al., 2016)
• Labels corresponding to target, stance, opinion towards and sentiment information
• 1,205, 2,409, and 221 instances for favor, against, and none each
 Sentiment classification (20,632)
• Utilizes the test data released in SemEval 2017 Task 4 (Rosenthal et al., 2017)
• Consists of 7,059 positive, 3,231 negative and 10,342 neutral tweets
15
Experiment and Result
• Result
 The proposed surpasses the baseline results in META, IRONY, and SUBJ
• Which are expected to benefit from identifying the specific syntactic categories
• Relatively weak at discerning the latent information such as STANCE and SENT
 Dependency on the source information
• META highly prefers the word-level attention source
– Concerns explicit existence of certain lexical terms
16
Experiment and Result
• Result
 STANCE and IRONY requires contextual information as well, but works
better in IRONY?
• IRONY incorporates hashtagged information which matters in the prediction
 Expected suitable application
• Bitstream or symbolic music analysis (where the formatted/syntactic information
plays an important role)
17
• Two excerpts from SUBJ
 Enables to see which syntactic
category is the most important
 Will be effective if the constituency
tagging is more reliable!
Visualization
18
Conclusion
• Sentence modeling concerns recent language modeling and deep
learning techniques
• Among attention approaches, how model pays attention is
determined automatically, but hardly cares the syntactic categories
that are given as a prior information
• Incorporating such information in deciding the attention weight via
projection matrix brings advantage in some tasks that necessitate
the attention on lexical features
19
Thank you!
EndOfPresentation

Mais conteúdo relacionado

Mais procurados

NLP Bootcamp 2018 : Representation Learning of text for NLP
NLP Bootcamp 2018 : Representation Learning of text for NLPNLP Bootcamp 2018 : Representation Learning of text for NLP
NLP Bootcamp 2018 : Representation Learning of text for NLPAnuj Gupta
 
ON THE UTILITY OF A SYLLABLE-LIKE SEGMENTATION FOR LEARNING A TRANSLITERATION...
ON THE UTILITY OF A SYLLABLE-LIKE SEGMENTATION FOR LEARNING A TRANSLITERATION...ON THE UTILITY OF A SYLLABLE-LIKE SEGMENTATION FOR LEARNING A TRANSLITERATION...
ON THE UTILITY OF A SYLLABLE-LIKE SEGMENTATION FOR LEARNING A TRANSLITERATION...cscpconf
 
Arabic Question Answering: Challenges, Tasks, Approaches, Test-sets, Tools, A...
Arabic Question Answering: Challenges, Tasks, Approaches, Test-sets, Tools, A...Arabic Question Answering: Challenges, Tasks, Approaches, Test-sets, Tools, A...
Arabic Question Answering: Challenges, Tasks, Approaches, Test-sets, Tools, A...Ahmed Magdy Ezzeldin, MSc.
 
MaxEnt (Loglinear) Models - Overview
MaxEnt (Loglinear) Models - OverviewMaxEnt (Loglinear) Models - Overview
MaxEnt (Loglinear) Models - Overviewananth
 
Transformers to Learn Hierarchical Contexts in Multiparty Dialogue
Transformers to Learn Hierarchical Contexts in Multiparty DialogueTransformers to Learn Hierarchical Contexts in Multiparty Dialogue
Transformers to Learn Hierarchical Contexts in Multiparty DialogueJinho Choi
 
Nlp research presentation
Nlp research presentationNlp research presentation
Nlp research presentationSurya Sg
 
Sentence representations and question answering (YerevaNN)
Sentence representations and question answering (YerevaNN)Sentence representations and question answering (YerevaNN)
Sentence representations and question answering (YerevaNN)YerevaNN research lab
 
Extraction Based automatic summarization
Extraction Based automatic summarizationExtraction Based automatic summarization
Extraction Based automatic summarizationAbdelaziz Al-Rihawi
 
VOC real world enterprise needs
VOC real world enterprise needsVOC real world enterprise needs
VOC real world enterprise needsIvan Berlocher
 
BERT: Pre-training of Deep Bidirectional Transformers for Language Understanding
BERT: Pre-training of Deep Bidirectional Transformers for Language UnderstandingBERT: Pre-training of Deep Bidirectional Transformers for Language Understanding
BERT: Pre-training of Deep Bidirectional Transformers for Language UnderstandingYoung Seok Kim
 
2010 INTERSPEECH
2010 INTERSPEECH 2010 INTERSPEECH
2010 INTERSPEECH WarNik Chow
 
Meta-evaluation of machine translation evaluation methods
Meta-evaluation of machine translation evaluation methodsMeta-evaluation of machine translation evaluation methods
Meta-evaluation of machine translation evaluation methodsLifeng (Aaron) Han
 
BERT: Pre-training of Deep Bidirectional Transformers for Language Understanding
BERT: Pre-training of Deep Bidirectional Transformers for Language UnderstandingBERT: Pre-training of Deep Bidirectional Transformers for Language Understanding
BERT: Pre-training of Deep Bidirectional Transformers for Language UnderstandingSeonghyun Kim
 
Enriching Word Vectors with Subword Information
Enriching Word Vectors with Subword InformationEnriching Word Vectors with Subword Information
Enriching Word Vectors with Subword InformationSeonghyun Kim
 
NLP Project Presentation
NLP Project PresentationNLP Project Presentation
NLP Project PresentationAryak Sengupta
 

Mais procurados (20)

Word embedding
Word embedding Word embedding
Word embedding
 
NLP Bootcamp 2018 : Representation Learning of text for NLP
NLP Bootcamp 2018 : Representation Learning of text for NLPNLP Bootcamp 2018 : Representation Learning of text for NLP
NLP Bootcamp 2018 : Representation Learning of text for NLP
 
ON THE UTILITY OF A SYLLABLE-LIKE SEGMENTATION FOR LEARNING A TRANSLITERATION...
ON THE UTILITY OF A SYLLABLE-LIKE SEGMENTATION FOR LEARNING A TRANSLITERATION...ON THE UTILITY OF A SYLLABLE-LIKE SEGMENTATION FOR LEARNING A TRANSLITERATION...
ON THE UTILITY OF A SYLLABLE-LIKE SEGMENTATION FOR LEARNING A TRANSLITERATION...
 
Arabic Question Answering: Challenges, Tasks, Approaches, Test-sets, Tools, A...
Arabic Question Answering: Challenges, Tasks, Approaches, Test-sets, Tools, A...Arabic Question Answering: Challenges, Tasks, Approaches, Test-sets, Tools, A...
Arabic Question Answering: Challenges, Tasks, Approaches, Test-sets, Tools, A...
 
MaxEnt (Loglinear) Models - Overview
MaxEnt (Loglinear) Models - OverviewMaxEnt (Loglinear) Models - Overview
MaxEnt (Loglinear) Models - Overview
 
Language models
Language modelsLanguage models
Language models
 
Arabic question answering ‫‬
Arabic question answering ‫‬Arabic question answering ‫‬
Arabic question answering ‫‬
 
Transformers to Learn Hierarchical Contexts in Multiparty Dialogue
Transformers to Learn Hierarchical Contexts in Multiparty DialogueTransformers to Learn Hierarchical Contexts in Multiparty Dialogue
Transformers to Learn Hierarchical Contexts in Multiparty Dialogue
 
Nlp research presentation
Nlp research presentationNlp research presentation
Nlp research presentation
 
Text summarization
Text summarization Text summarization
Text summarization
 
Sentence representations and question answering (YerevaNN)
Sentence representations and question answering (YerevaNN)Sentence representations and question answering (YerevaNN)
Sentence representations and question answering (YerevaNN)
 
DLBLR talk
DLBLR talkDLBLR talk
DLBLR talk
 
Extraction Based automatic summarization
Extraction Based automatic summarizationExtraction Based automatic summarization
Extraction Based automatic summarization
 
VOC real world enterprise needs
VOC real world enterprise needsVOC real world enterprise needs
VOC real world enterprise needs
 
BERT: Pre-training of Deep Bidirectional Transformers for Language Understanding
BERT: Pre-training of Deep Bidirectional Transformers for Language UnderstandingBERT: Pre-training of Deep Bidirectional Transformers for Language Understanding
BERT: Pre-training of Deep Bidirectional Transformers for Language Understanding
 
2010 INTERSPEECH
2010 INTERSPEECH 2010 INTERSPEECH
2010 INTERSPEECH
 
Meta-evaluation of machine translation evaluation methods
Meta-evaluation of machine translation evaluation methodsMeta-evaluation of machine translation evaluation methods
Meta-evaluation of machine translation evaluation methods
 
BERT: Pre-training of Deep Bidirectional Transformers for Language Understanding
BERT: Pre-training of Deep Bidirectional Transformers for Language UnderstandingBERT: Pre-training of Deep Bidirectional Transformers for Language Understanding
BERT: Pre-training of Deep Bidirectional Transformers for Language Understanding
 
Enriching Word Vectors with Subword Information
Enriching Word Vectors with Subword InformationEnriching Word Vectors with Subword Information
Enriching Word Vectors with Subword Information
 
NLP Project Presentation
NLP Project PresentationNLP Project Presentation
NLP Project Presentation
 

Semelhante a 2010 PACLIC - pay attention to categories

Deep Learning for Information Retrieval: Models, Progress, & Opportunities
Deep Learning for Information Retrieval: Models, Progress, & OpportunitiesDeep Learning for Information Retrieval: Models, Progress, & Opportunities
Deep Learning for Information Retrieval: Models, Progress, & OpportunitiesMatthew Lease
 
Natural Language Processing, Techniques, Current Trends and Applications in I...
Natural Language Processing, Techniques, Current Trends and Applications in I...Natural Language Processing, Techniques, Current Trends and Applications in I...
Natural Language Processing, Techniques, Current Trends and Applications in I...RajkiranVeluri
 
Biomedical Word Sense Disambiguation presentation [Autosaved]
Biomedical Word Sense Disambiguation presentation [Autosaved]Biomedical Word Sense Disambiguation presentation [Autosaved]
Biomedical Word Sense Disambiguation presentation [Autosaved]akm sabbir
 
Topic extraction using machine learning
Topic extraction using machine learningTopic extraction using machine learning
Topic extraction using machine learningSanjib Basak
 
Intro to Deep Learning for Question Answering
Intro to Deep Learning for Question AnsweringIntro to Deep Learning for Question Answering
Intro to Deep Learning for Question AnsweringTraian Rebedea
 
Topic Extraction using Machine Learning
Topic Extraction using Machine LearningTopic Extraction using Machine Learning
Topic Extraction using Machine LearningSanjib Basak
 
Using a keyword extraction pipeline to understand concepts in future work sec...
Using a keyword extraction pipeline to understand concepts in future work sec...Using a keyword extraction pipeline to understand concepts in future work sec...
Using a keyword extraction pipeline to understand concepts in future work sec...Kai Li
 
French machine reading for question answering
French machine reading for question answeringFrench machine reading for question answering
French machine reading for question answeringAli Kabbadj
 
Philippe Langlais - 2017 - Users and Data: The Two Neglected Children of Bili...
Philippe Langlais - 2017 - Users and Data: The Two Neglected Children of Bili...Philippe Langlais - 2017 - Users and Data: The Two Neglected Children of Bili...
Philippe Langlais - 2017 - Users and Data: The Two Neglected Children of Bili...Association for Computational Linguistics
 
Embed, Encode, Attend, Predict – applying the 4 step NLP recipe for text clas...
Embed, Encode, Attend, Predict – applying the 4 step NLP recipe for text clas...Embed, Encode, Attend, Predict – applying the 4 step NLP recipe for text clas...
Embed, Encode, Attend, Predict – applying the 4 step NLP recipe for text clas...Sujit Pal
 
Classifying Non-Referential It for Question Answer Pairs
Classifying Non-Referential It for Question Answer PairsClassifying Non-Referential It for Question Answer Pairs
Classifying Non-Referential It for Question Answer PairsJinho Choi
 
Deep Attention Model for Triage of Emergency Department Patients - Djordje Gl...
Deep Attention Model for Triage of Emergency Department Patients - Djordje Gl...Deep Attention Model for Triage of Emergency Department Patients - Djordje Gl...
Deep Attention Model for Triage of Emergency Department Patients - Djordje Gl...Institute of Contemporary Sciences
 
2018.01.25 rune sætre_triallecture_xai_v2
2018.01.25 rune sætre_triallecture_xai_v22018.01.25 rune sætre_triallecture_xai_v2
2018.01.25 rune sætre_triallecture_xai_v2Rune Sætre
 
Natural Language Processing: From Human-Robot Interaction to Alzheimer’s Dete...
Natural Language Processing: From Human-Robot Interaction to Alzheimer’s Dete...Natural Language Processing: From Human-Robot Interaction to Alzheimer’s Dete...
Natural Language Processing: From Human-Robot Interaction to Alzheimer’s Dete...Jekaterina Novikova, PhD
 
Sujit Pal - Applying the four-step "Embed, Encode, Attend, Predict" framework...
Sujit Pal - Applying the four-step "Embed, Encode, Attend, Predict" framework...Sujit Pal - Applying the four-step "Embed, Encode, Attend, Predict" framework...
Sujit Pal - Applying the four-step "Embed, Encode, Attend, Predict" framework...PyData
 
Multimodal Learning Analytics
Multimodal Learning AnalyticsMultimodal Learning Analytics
Multimodal Learning AnalyticsXavier Ochoa
 
Towards Model-Based Optimisation: Using Domain Knowledge Explicitly (MELO 2016)
Towards Model-Based Optimisation: Using Domain Knowledge Explicitly (MELO 2016)Towards Model-Based Optimisation: Using Domain Knowledge Explicitly (MELO 2016)
Towards Model-Based Optimisation: Using Domain Knowledge Explicitly (MELO 2016)Steffen Zschaler
 

Semelhante a 2010 PACLIC - pay attention to categories (20)

Deep Learning for Information Retrieval: Models, Progress, & Opportunities
Deep Learning for Information Retrieval: Models, Progress, & OpportunitiesDeep Learning for Information Retrieval: Models, Progress, & Opportunities
Deep Learning for Information Retrieval: Models, Progress, & Opportunities
 
Natural Language Processing, Techniques, Current Trends and Applications in I...
Natural Language Processing, Techniques, Current Trends and Applications in I...Natural Language Processing, Techniques, Current Trends and Applications in I...
Natural Language Processing, Techniques, Current Trends and Applications in I...
 
Biomedical Word Sense Disambiguation presentation [Autosaved]
Biomedical Word Sense Disambiguation presentation [Autosaved]Biomedical Word Sense Disambiguation presentation [Autosaved]
Biomedical Word Sense Disambiguation presentation [Autosaved]
 
Topic extraction using machine learning
Topic extraction using machine learningTopic extraction using machine learning
Topic extraction using machine learning
 
Intro to Deep Learning for Question Answering
Intro to Deep Learning for Question AnsweringIntro to Deep Learning for Question Answering
Intro to Deep Learning for Question Answering
 
Topic Extraction using Machine Learning
Topic Extraction using Machine LearningTopic Extraction using Machine Learning
Topic Extraction using Machine Learning
 
Using a keyword extraction pipeline to understand concepts in future work sec...
Using a keyword extraction pipeline to understand concepts in future work sec...Using a keyword extraction pipeline to understand concepts in future work sec...
Using a keyword extraction pipeline to understand concepts in future work sec...
 
French machine reading for question answering
French machine reading for question answeringFrench machine reading for question answering
French machine reading for question answering
 
Philippe Langlais - 2017 - Users and Data: The Two Neglected Children of Bili...
Philippe Langlais - 2017 - Users and Data: The Two Neglected Children of Bili...Philippe Langlais - 2017 - Users and Data: The Two Neglected Children of Bili...
Philippe Langlais - 2017 - Users and Data: The Two Neglected Children of Bili...
 
Embed, Encode, Attend, Predict – applying the 4 step NLP recipe for text clas...
Embed, Encode, Attend, Predict – applying the 4 step NLP recipe for text clas...Embed, Encode, Attend, Predict – applying the 4 step NLP recipe for text clas...
Embed, Encode, Attend, Predict – applying the 4 step NLP recipe for text clas...
 
Word2vector
Word2vectorWord2vector
Word2vector
 
Classifying Non-Referential It for Question Answer Pairs
Classifying Non-Referential It for Question Answer PairsClassifying Non-Referential It for Question Answer Pairs
Classifying Non-Referential It for Question Answer Pairs
 
Deep Attention Model for Triage of Emergency Department Patients - Djordje Gl...
Deep Attention Model for Triage of Emergency Department Patients - Djordje Gl...Deep Attention Model for Triage of Emergency Department Patients - Djordje Gl...
Deep Attention Model for Triage of Emergency Department Patients - Djordje Gl...
 
Word 2 vector
Word 2 vectorWord 2 vector
Word 2 vector
 
2018.01.25 rune sætre_triallecture_xai_v2
2018.01.25 rune sætre_triallecture_xai_v22018.01.25 rune sætre_triallecture_xai_v2
2018.01.25 rune sætre_triallecture_xai_v2
 
Natural Language Processing: From Human-Robot Interaction to Alzheimer’s Dete...
Natural Language Processing: From Human-Robot Interaction to Alzheimer’s Dete...Natural Language Processing: From Human-Robot Interaction to Alzheimer’s Dete...
Natural Language Processing: From Human-Robot Interaction to Alzheimer’s Dete...
 
Sujit Pal - Applying the four-step "Embed, Encode, Attend, Predict" framework...
Sujit Pal - Applying the four-step "Embed, Encode, Attend, Predict" framework...Sujit Pal - Applying the four-step "Embed, Encode, Attend, Predict" framework...
Sujit Pal - Applying the four-step "Embed, Encode, Attend, Predict" framework...
 
Multimodal Learning Analytics
Multimodal Learning AnalyticsMultimodal Learning Analytics
Multimodal Learning Analytics
 
Towards Model-Based Optimisation: Using Domain Knowledge Explicitly (MELO 2016)
Towards Model-Based Optimisation: Using Domain Knowledge Explicitly (MELO 2016)Towards Model-Based Optimisation: Using Domain Knowledge Explicitly (MELO 2016)
Towards Model-Based Optimisation: Using Domain Knowledge Explicitly (MELO 2016)
 
2211 APSIPA
2211 APSIPA2211 APSIPA
2211 APSIPA
 

Mais de WarNik Chow

2206 FAccT_inperson
2206 FAccT_inperson2206 FAccT_inperson
2206 FAccT_inpersonWarNik Chow
 
2204 Kakao talk on Hate speech dataset
2204 Kakao talk on Hate speech dataset2204 Kakao talk on Hate speech dataset
2204 Kakao talk on Hate speech datasetWarNik Chow
 
2108 [LangCon2021] kosp2e
2108 [LangCon2021] kosp2e2108 [LangCon2021] kosp2e
2108 [LangCon2021] kosp2eWarNik Chow
 
2102 Redone seminar
2102 Redone seminar2102 Redone seminar
2102 Redone seminarWarNik Chow
 
2010 HCLT Hate Speech
2010 HCLT Hate Speech2010 HCLT Hate Speech
2010 HCLT Hate SpeechWarNik Chow
 
2009 DevC Seongnam - NLP
2009 DevC Seongnam - NLP2009 DevC Seongnam - NLP
2009 DevC Seongnam - NLPWarNik Chow
 
2008 [lang con2020] act!
2008 [lang con2020] act!2008 [lang con2020] act!
2008 [lang con2020] act!WarNik Chow
 
2007 CogSci 2020 poster
2007 CogSci 2020 poster2007 CogSci 2020 poster
2007 CogSci 2020 posterWarNik Chow
 

Mais de WarNik Chow (20)

2312 PACLIC
2312 PACLIC2312 PACLIC
2312 PACLIC
 
2311 EAAMO
2311 EAAMO2311 EAAMO
2311 EAAMO
 
2211 HCOMP
2211 HCOMP2211 HCOMP
2211 HCOMP
 
2211 AACL
2211 AACL2211 AACL
2211 AACL
 
2210 CODI
2210 CODI2210 CODI
2210 CODI
 
2206 FAccT_inperson
2206 FAccT_inperson2206 FAccT_inperson
2206 FAccT_inperson
 
2206 Modupop!
2206 Modupop!2206 Modupop!
2206 Modupop!
 
2204 Kakao talk on Hate speech dataset
2204 Kakao talk on Hate speech dataset2204 Kakao talk on Hate speech dataset
2204 Kakao talk on Hate speech dataset
 
2108 [LangCon2021] kosp2e
2108 [LangCon2021] kosp2e2108 [LangCon2021] kosp2e
2108 [LangCon2021] kosp2e
 
2106 PRSLLS
2106 PRSLLS2106 PRSLLS
2106 PRSLLS
 
2106 JWLLP
2106 JWLLP2106 JWLLP
2106 JWLLP
 
2106 ACM DIS
2106 ACM DIS2106 ACM DIS
2106 ACM DIS
 
2104 Talk @SSU
2104 Talk @SSU2104 Talk @SSU
2104 Talk @SSU
 
2103 ACM FAccT
2103 ACM FAccT2103 ACM FAccT
2103 ACM FAccT
 
2102 Redone seminar
2102 Redone seminar2102 Redone seminar
2102 Redone seminar
 
2011 NLP-OSS
2011 NLP-OSS2011 NLP-OSS
2011 NLP-OSS
 
2010 HCLT Hate Speech
2010 HCLT Hate Speech2010 HCLT Hate Speech
2010 HCLT Hate Speech
 
2009 DevC Seongnam - NLP
2009 DevC Seongnam - NLP2009 DevC Seongnam - NLP
2009 DevC Seongnam - NLP
 
2008 [lang con2020] act!
2008 [lang con2020] act!2008 [lang con2020] act!
2008 [lang con2020] act!
 
2007 CogSci 2020 poster
2007 CogSci 2020 poster2007 CogSci 2020 poster
2007 CogSci 2020 poster
 

Último

UNIT - IV - Air Compressors and its Performance
UNIT - IV - Air Compressors and its PerformanceUNIT - IV - Air Compressors and its Performance
UNIT - IV - Air Compressors and its Performancesivaprakash250
 
Extrusion Processes and Their Limitations
Extrusion Processes and Their LimitationsExtrusion Processes and Their Limitations
Extrusion Processes and Their Limitations120cr0395
 
Processing & Properties of Floor and Wall Tiles.pptx
Processing & Properties of Floor and Wall Tiles.pptxProcessing & Properties of Floor and Wall Tiles.pptx
Processing & Properties of Floor and Wall Tiles.pptxpranjaldaimarysona
 
(PRIYA) Rajgurunagar Call Girls Just Call 7001035870 [ Cash on Delivery ] Pun...
(PRIYA) Rajgurunagar Call Girls Just Call 7001035870 [ Cash on Delivery ] Pun...(PRIYA) Rajgurunagar Call Girls Just Call 7001035870 [ Cash on Delivery ] Pun...
(PRIYA) Rajgurunagar Call Girls Just Call 7001035870 [ Cash on Delivery ] Pun...ranjana rawat
 
Introduction to IEEE STANDARDS and its different types.pptx
Introduction to IEEE STANDARDS and its different types.pptxIntroduction to IEEE STANDARDS and its different types.pptx
Introduction to IEEE STANDARDS and its different types.pptxupamatechverse
 
Introduction to Multiple Access Protocol.pptx
Introduction to Multiple Access Protocol.pptxIntroduction to Multiple Access Protocol.pptx
Introduction to Multiple Access Protocol.pptxupamatechverse
 
(ANVI) Koregaon Park Call Girls Just Call 7001035870 [ Cash on Delivery ] Pun...
(ANVI) Koregaon Park Call Girls Just Call 7001035870 [ Cash on Delivery ] Pun...(ANVI) Koregaon Park Call Girls Just Call 7001035870 [ Cash on Delivery ] Pun...
(ANVI) Koregaon Park Call Girls Just Call 7001035870 [ Cash on Delivery ] Pun...ranjana rawat
 
CCS335 _ Neural Networks and Deep Learning Laboratory_Lab Complete Record
CCS335 _ Neural Networks and Deep Learning Laboratory_Lab Complete RecordCCS335 _ Neural Networks and Deep Learning Laboratory_Lab Complete Record
CCS335 _ Neural Networks and Deep Learning Laboratory_Lab Complete RecordAsst.prof M.Gokilavani
 
MANUFACTURING PROCESS-II UNIT-1 THEORY OF METAL CUTTING
MANUFACTURING PROCESS-II UNIT-1 THEORY OF METAL CUTTINGMANUFACTURING PROCESS-II UNIT-1 THEORY OF METAL CUTTING
MANUFACTURING PROCESS-II UNIT-1 THEORY OF METAL CUTTINGSIVASHANKAR N
 
Java Programming :Event Handling(Types of Events)
Java Programming :Event Handling(Types of Events)Java Programming :Event Handling(Types of Events)
Java Programming :Event Handling(Types of Events)simmis5
 
Booking open Available Pune Call Girls Koregaon Park 6297143586 Call Hot Ind...
Booking open Available Pune Call Girls Koregaon Park  6297143586 Call Hot Ind...Booking open Available Pune Call Girls Koregaon Park  6297143586 Call Hot Ind...
Booking open Available Pune Call Girls Koregaon Park 6297143586 Call Hot Ind...Call Girls in Nagpur High Profile
 
University management System project report..pdf
University management System project report..pdfUniversity management System project report..pdf
University management System project report..pdfKamal Acharya
 
Call for Papers - African Journal of Biological Sciences, E-ISSN: 2663-2187, ...
Call for Papers - African Journal of Biological Sciences, E-ISSN: 2663-2187, ...Call for Papers - African Journal of Biological Sciences, E-ISSN: 2663-2187, ...
Call for Papers - African Journal of Biological Sciences, E-ISSN: 2663-2187, ...Christo Ananth
 
High Profile Call Girls Nagpur Meera Call 7001035870 Meet With Nagpur Escorts
High Profile Call Girls Nagpur Meera Call 7001035870 Meet With Nagpur EscortsHigh Profile Call Girls Nagpur Meera Call 7001035870 Meet With Nagpur Escorts
High Profile Call Girls Nagpur Meera Call 7001035870 Meet With Nagpur EscortsCall Girls in Nagpur High Profile
 
Booking open Available Pune Call Girls Pargaon 6297143586 Call Hot Indian Gi...
Booking open Available Pune Call Girls Pargaon  6297143586 Call Hot Indian Gi...Booking open Available Pune Call Girls Pargaon  6297143586 Call Hot Indian Gi...
Booking open Available Pune Call Girls Pargaon 6297143586 Call Hot Indian Gi...Call Girls in Nagpur High Profile
 
Glass Ceramics: Processing and Properties
Glass Ceramics: Processing and PropertiesGlass Ceramics: Processing and Properties
Glass Ceramics: Processing and PropertiesPrabhanshu Chaturvedi
 
BSides Seattle 2024 - Stopping Ethan Hunt From Taking Your Data.pptx
BSides Seattle 2024 - Stopping Ethan Hunt From Taking Your Data.pptxBSides Seattle 2024 - Stopping Ethan Hunt From Taking Your Data.pptx
BSides Seattle 2024 - Stopping Ethan Hunt From Taking Your Data.pptxfenichawla
 
Online banking management system project.pdf
Online banking management system project.pdfOnline banking management system project.pdf
Online banking management system project.pdfKamal Acharya
 

Último (20)

UNIT - IV - Air Compressors and its Performance
UNIT - IV - Air Compressors and its PerformanceUNIT - IV - Air Compressors and its Performance
UNIT - IV - Air Compressors and its Performance
 
Extrusion Processes and Their Limitations
Extrusion Processes and Their LimitationsExtrusion Processes and Their Limitations
Extrusion Processes and Their Limitations
 
Processing & Properties of Floor and Wall Tiles.pptx
Processing & Properties of Floor and Wall Tiles.pptxProcessing & Properties of Floor and Wall Tiles.pptx
Processing & Properties of Floor and Wall Tiles.pptx
 
(PRIYA) Rajgurunagar Call Girls Just Call 7001035870 [ Cash on Delivery ] Pun...
(PRIYA) Rajgurunagar Call Girls Just Call 7001035870 [ Cash on Delivery ] Pun...(PRIYA) Rajgurunagar Call Girls Just Call 7001035870 [ Cash on Delivery ] Pun...
(PRIYA) Rajgurunagar Call Girls Just Call 7001035870 [ Cash on Delivery ] Pun...
 
Introduction to IEEE STANDARDS and its different types.pptx
Introduction to IEEE STANDARDS and its different types.pptxIntroduction to IEEE STANDARDS and its different types.pptx
Introduction to IEEE STANDARDS and its different types.pptx
 
Introduction to Multiple Access Protocol.pptx
Introduction to Multiple Access Protocol.pptxIntroduction to Multiple Access Protocol.pptx
Introduction to Multiple Access Protocol.pptx
 
(ANVI) Koregaon Park Call Girls Just Call 7001035870 [ Cash on Delivery ] Pun...
(ANVI) Koregaon Park Call Girls Just Call 7001035870 [ Cash on Delivery ] Pun...(ANVI) Koregaon Park Call Girls Just Call 7001035870 [ Cash on Delivery ] Pun...
(ANVI) Koregaon Park Call Girls Just Call 7001035870 [ Cash on Delivery ] Pun...
 
CCS335 _ Neural Networks and Deep Learning Laboratory_Lab Complete Record
CCS335 _ Neural Networks and Deep Learning Laboratory_Lab Complete RecordCCS335 _ Neural Networks and Deep Learning Laboratory_Lab Complete Record
CCS335 _ Neural Networks and Deep Learning Laboratory_Lab Complete Record
 
MANUFACTURING PROCESS-II UNIT-1 THEORY OF METAL CUTTING
MANUFACTURING PROCESS-II UNIT-1 THEORY OF METAL CUTTINGMANUFACTURING PROCESS-II UNIT-1 THEORY OF METAL CUTTING
MANUFACTURING PROCESS-II UNIT-1 THEORY OF METAL CUTTING
 
Java Programming :Event Handling(Types of Events)
Java Programming :Event Handling(Types of Events)Java Programming :Event Handling(Types of Events)
Java Programming :Event Handling(Types of Events)
 
Booking open Available Pune Call Girls Koregaon Park 6297143586 Call Hot Ind...
Booking open Available Pune Call Girls Koregaon Park  6297143586 Call Hot Ind...Booking open Available Pune Call Girls Koregaon Park  6297143586 Call Hot Ind...
Booking open Available Pune Call Girls Koregaon Park 6297143586 Call Hot Ind...
 
University management System project report..pdf
University management System project report..pdfUniversity management System project report..pdf
University management System project report..pdf
 
Call for Papers - African Journal of Biological Sciences, E-ISSN: 2663-2187, ...
Call for Papers - African Journal of Biological Sciences, E-ISSN: 2663-2187, ...Call for Papers - African Journal of Biological Sciences, E-ISSN: 2663-2187, ...
Call for Papers - African Journal of Biological Sciences, E-ISSN: 2663-2187, ...
 
High Profile Call Girls Nagpur Meera Call 7001035870 Meet With Nagpur Escorts
High Profile Call Girls Nagpur Meera Call 7001035870 Meet With Nagpur EscortsHigh Profile Call Girls Nagpur Meera Call 7001035870 Meet With Nagpur Escorts
High Profile Call Girls Nagpur Meera Call 7001035870 Meet With Nagpur Escorts
 
Booking open Available Pune Call Girls Pargaon 6297143586 Call Hot Indian Gi...
Booking open Available Pune Call Girls Pargaon  6297143586 Call Hot Indian Gi...Booking open Available Pune Call Girls Pargaon  6297143586 Call Hot Indian Gi...
Booking open Available Pune Call Girls Pargaon 6297143586 Call Hot Indian Gi...
 
(INDIRA) Call Girl Aurangabad Call Now 8617697112 Aurangabad Escorts 24x7
(INDIRA) Call Girl Aurangabad Call Now 8617697112 Aurangabad Escorts 24x7(INDIRA) Call Girl Aurangabad Call Now 8617697112 Aurangabad Escorts 24x7
(INDIRA) Call Girl Aurangabad Call Now 8617697112 Aurangabad Escorts 24x7
 
DJARUM4D - SLOT GACOR ONLINE | SLOT DEMO ONLINE
DJARUM4D - SLOT GACOR ONLINE | SLOT DEMO ONLINEDJARUM4D - SLOT GACOR ONLINE | SLOT DEMO ONLINE
DJARUM4D - SLOT GACOR ONLINE | SLOT DEMO ONLINE
 
Glass Ceramics: Processing and Properties
Glass Ceramics: Processing and PropertiesGlass Ceramics: Processing and Properties
Glass Ceramics: Processing and Properties
 
BSides Seattle 2024 - Stopping Ethan Hunt From Taking Your Data.pptx
BSides Seattle 2024 - Stopping Ethan Hunt From Taking Your Data.pptxBSides Seattle 2024 - Stopping Ethan Hunt From Taking Your Data.pptx
BSides Seattle 2024 - Stopping Ethan Hunt From Taking Your Data.pptx
 
Online banking management system project.pdf
Online banking management system project.pdfOnline banking management system project.pdf
Online banking management system project.pdf
 

2010 PACLIC - pay attention to categories

  • 1. Human Interface Laboratory Pay Attention to Categories: Syntax-Based Sentence Modeling with Metadata Projection Matrix 2020. 10. 24, @PACLIC 34 Won Ik Cho, Nam Soo Kim (SNU ECE & INMC)
  • 2. Contents • Introduction • Proposed Method • Experiment and Result • Visualization • Conclusion 1
  • 3. Introduction • Brief overview on sentence modeling  Sparse encoding to dense word vectors 2
  • 4. Introduction • Brief overview on sentence modeling  Deep learning techniques and attention model 3
  • 5. Introduction • Brief overview on sentence modeling  Self-attentive models (which advances self attention) • Still a useful approach to sentence classification 4
  • 6. Introduction • Motivation  What if we need to pay more attention to some syntactic categories and want to decide the intensity automatically? • e.g., Oxymoron detection (Cho et al., 2017) – Drink a sugar-free sweet tea – When I’m low, I get high  Attention mechanism is useful, but less reliable in the control regarding syntactic categories • Not mainly considered since the necessity is underestimated • Minimal information (e.g., POS) can be of help for some tasks! • How can we take into account that information beyond just attaching it to each token? 5
  • 7. Related Work • Word embedding  Dense embedding of words, in view of distributional semantics  Projecting words to low dimensional space, with the objective of ‘making the distribution that enables the prediction on the probable surrounding words’  Variations of word2vec • Word2Vec (Mikolov et al., 2013) • GloVe (Pennington et al., 2014) • fastText (Bojanovsky et al., 2016) 6
  • 8. Related Work • Deep learning techniques  Convolutional neural networks • Primarily applied to the vision area • Used in Kim (2014) as 1D convolution that represents the sentence as a sequence of dense word vectors  Recurrent neural networks • Adopted to model the sequential data • Long short-term memory is used to cope with the vanishing gradient • Bidirectional LSTM to consider the directivity 7
  • 9. Related Work • Deep learning techniques  Attention models • First used to deal with the word order sensitivity and matching in machine translation • Evolved to various format, such as location-based  Self-attentive sentence embedding (Lin et al., 2017) • Related to self attention, but more applicable to BiLSTM format 8
  • 10. Related Work • Deep learning techniques  Self-attentive sentence embedding 9
  • 11. Proposed Method • Overall description  Sequential word embedding and BiLSTM  Feature extraction for attention source • TF-IDF? BiLSTM?  Attention source activated with ReLU  PAC structure with: • Weight layer with category-wise info • Projection matrix • Multiplication (𝛼1 𝐿 ) 10
  • 12. Proposed Method • Overall description  Sequential word embedding and BiLSTM  Feature extraction for attention source • TF-IDF? BiLSTM?  Attention source activated with ReLU  PAC structure with: • Weight layer with category-wise info • Projection matrix • Multiplication (𝛼1 𝐿 ) 11
  • 13. Proposed Method • Overall description  Sequential word embedding and BiLSTM  Feature extraction for attention source • TF-IDF? BiLSTM?  Attention source activated with ReLU  PAC structure with: • Weight layer with category-wise info • Projection matrix • Multiplication (𝛼1 𝐿 ) 12
  • 14. Proposed Method • Overall description  PAC structure with: • Weight layer with category-wise info • Projection matrix • Multiplication (𝛼1 𝐿 ) 13
  • 15. Experiment and Result • Implementation  Baseline features • TF-IDFs and bigrams – Dictionary size set to 3,000 (=30 * 100) • GloVe pretrained with Twitter 27B – Word vector dim. 100 – Padding max length 30  Baseline classifiers • SVM for TF-IDF • NN for GloVe (averaged) with Adam(0.0005) and batch size 16 • CNN (32 filters, window 3) and BiLSTM (hidden dim. 64) for Glove (padded)  Baeseline attention model • Lin et al. (2017) with context vector dim. 64, alongwith the above BiLSTM  The proposed • 𝑛 𝑝 follows the NLTK POS tagging result 14
  • 16. Experiment and Result • Dataset  Metalanguage detection (2,393) • Investigates whether a sentence contains explicit mention terms (‘title’ or ‘name’) • Contains 629 mentioned and 1,764 not-mentioned instances excerpted from Wikipedia  Irony detection (4,618) • Distributed in SemEval 2018 Task 3 for ironic tweet detection (Van Hee et al., 2018). • Binary label case was taken into account; 2,222 contain irony and 2,396 do not  Subjectivity detection (10,000) • Refers to Pang and Lee (2004); checks if the movie review contains a subjective judgment • Incorporates equally 5,000 instances for each of the subjective and objective reviews  Stance classification (3,835) • Part of distributed dataset from SemEval 2016 Task 6 (Mohammad et al., 2016) • Labels corresponding to target, stance, opinion towards and sentiment information • 1,205, 2,409, and 221 instances for favor, against, and none each  Sentiment classification (20,632) • Utilizes the test data released in SemEval 2017 Task 4 (Rosenthal et al., 2017) • Consists of 7,059 positive, 3,231 negative and 10,342 neutral tweets 15
  • 17. Experiment and Result • Result  The proposed surpasses the baseline results in META, IRONY, and SUBJ • Which are expected to benefit from identifying the specific syntactic categories • Relatively weak at discerning the latent information such as STANCE and SENT  Dependency on the source information • META highly prefers the word-level attention source – Concerns explicit existence of certain lexical terms 16
  • 18. Experiment and Result • Result  STANCE and IRONY requires contextual information as well, but works better in IRONY? • IRONY incorporates hashtagged information which matters in the prediction  Expected suitable application • Bitstream or symbolic music analysis (where the formatted/syntactic information plays an important role) 17
  • 19. • Two excerpts from SUBJ  Enables to see which syntactic category is the most important  Will be effective if the constituency tagging is more reliable! Visualization 18
  • 20. Conclusion • Sentence modeling concerns recent language modeling and deep learning techniques • Among attention approaches, how model pays attention is determined automatically, but hardly cares the syntactic categories that are given as a prior information • Incorporating such information in deciding the attention weight via projection matrix brings advantage in some tasks that necessitate the attention on lexical features 19

Notas do Editor

  1. .