SlideShare uma empresa Scribd logo
1 de 39
Bhaskar Mitra, Microsoft (Bing Sciences)
Search Solutions, 26th November, 2015
A brief introduction…
• I have worked as a relevance engineer for Bing since 2007 (then
called as Live Search)
• Mostly on Web document ranking and query formulation
• Moved to an applied research scientist role in 2013
• In this talk I will present…
• Some personal learnings and takeaways from working on (neural and non-
neural) text embeddings for IR
• Highlight a few of my favourite insights/papers from the broader academic
community
*Thanks to Nick Craswell for suggesting the title “Vectorland”.
A recent trend…
Mikolov et. al. Efficient Estimation of Word Representations in Vector Space. (2013)
Huang et al. Learning deep structured semantic models for web search using clickthrough data. (CIKM, 2013)
Hong. Improving Paragraph2Vec. (2015)
Grbovic et. al. Context- and Content-aware Embeddings for Query Rewriting in Sponsored Search. (SIGIR, 2015)
Or as a learned RNN model once said*…
2vec or not 2vec that is…
Recaims alone of those the mercorrance down. Sir,
And let it be, if not, no, then; if you get
Great rebels most of a heaven, I cannot mose
where his hearts makes the Rome arrase.
And then it stands: fear them against your honour,
I am a sifel loved him; he swores.
My lord, yet most gentle in our ears?
Our ax I can respect of? If you
concear, and lend me to his punishment?
If I make upon thee. Let me see how after
Wortens of she: is it your sister, pardon! air,
I give my recair to depose?
*The text above was auto-generated using Andrej
Karpathy’s Char-RNN implementation trained on the
works of Shakespeare and then seeded with the starting
text “to vector or not to vector that is”. Special thanks to
Milad Shokouhi for his help with running the RNN model.
Learning to
represent
A lot of recent work in
neural models and
“Deep Learning” is
focused on learning
vector representations
for text, image, speech,
entities, and other
nuggets of information
Learning to
represent
From analogies over
words and short texts…. Mikolov et. al. Efficient Estimation of Word Representations in Vector Space. (2013)
Mitra. Exploring Session Context using Distributed Representations of Queries and Reformulations. (SIGIR, 2015)
Learning to
represent
…and automatically
generating natural
language captions for
images,
Vinyals et. al. Show and Tell: A Neural Image Caption Generator. (2015)
Fang et. al. From Captions to Visual Concepts and Back. (CVPR, 2015)
Learning to
represent
…to building automated
conversational agents.
Vinyals et. al. A Neural Conversational Model. (ICML, 2015)
The basics...
One-hot vectors
A sparse bit vector where all values are zeros, except one. Each
position corresponds to a different item. The vector dimension is
equal to the number of items that need to be represented.
0 1 0 0 0 0 0 1
Bag-of-* vectors
A sparse count vector of component units. The vector dimension is
equal to the vocabulary size (number of distinct components).
0 0 0 0 0 1 0 0 0 1 0 0
“web search”
(Bag of words)
search web
0 1 0 1 0 0 2 0 1 0 1 0
“banana”
(Bag of trigrams)
ana nan#ba na# ban
Embeddings
A dense vector of real values. The
vector dimension is typically
much smaller than the number of
items or the vocabulary size.
You can imagine the vectors as
coordinates for items in the
embedding space.
Some distance metric defines a
notion of relatedness between
items in this space.
Neighborhoods in an embedding space
(Example)
Song et. al. Unsupervised Learning of Word Semantic Embedding using the Deep Structured Semantic Model. (2014)
Transitions in an embedding space
(Example)
Mitra. Exploring Session Context using Distributed Representations of Queries and Reformulations. (SIGIR, 2015)
Using text embeddings in search
Example use-cases
for text embeddings
in search
Learning a joint query and
document (title) embedding
for document ranking
Shen et. al. Learning semantic representations using convolutional neural networks for web search. (WWW, 2014)
Example use-cases
for text embeddings
in search
Gao et. al. Modeling Interestingness with Deep Neural Networks. (EMNLP, 2014)
Entity detection in document
(unstructured) body text
Example use-cases
for text embeddings
in search
Mitra and Craswell. Query Auto-Completion for Rare Prefixes. (CIKM, 2015)
Predicting suffixes (or next
word) for query auto-
completion for rare prefixes
Example use-cases
for text embeddings
in search
Mitra. Exploring Session Context using Distributed Representations of Queries and Reformulations. (SIGIR, 2015)
Session modelling by
learning an embedding for
query (or intent) transitions
Example use-cases
for text embeddings
in search
Nalisnick et. al. Improving Document Ranking with Dual Word Embeddings. (Submitted to WWW, 2016)
Modelling the aboutness of a
document by capturing
evidences from document
terms that do no match the
query
Passage about Albuquerque
Passage not about Albuquerque
Example use-cases
for text embeddings
in search
Liu et. al. Representation Learning Using Multi-Task Deep Neural Networks for
Semantic Classification and Information Retrieval. (NAACL, 2015)
Multi-task embedding of
queries for classification and
document retrieval
How do you learn an embedding?
How do you (typically) learn an embedding?
• Setup a prediction task
Source Item → Target Item
• Input and Output vectors are sparse
• Learning the embedding
≈ Dimensionality reduction
(*The bottleneck trick for NNs)
• Many options for the actual model
• Neural networks, matrix factorization,
Pointwise Mutual Information, etc.
Target
Item
Source
Item
Source
Embedding
Target
Embedding
Distance
Metric
Some examples of text embeddings
Embedding for Source Item Target Item Learning Model
Latent Semantic Analysis
Deerwester et. al. (1990)
Single word
Word
(one-hot)
Document
(one-hot)
Matrix factorization
Word2vec
Mikolov et. al. (2013)
Single Word
Word
(one-hot)
Neighboring Word
(one-hot)
Neural Network (Shallow)
Glove
Pennington et. al. (2014)
Single Word
Word
(one-hot)
Neighboring Word
(one-hot)
Matrix factorization
Semantic Hashing (auto-encoder)
Salakhutdinov and Hinton (2007)
Multi-word text
Document
(bag-of-words)
Same as source
(bag-of-words)
Neural Network (Deep)
DSSM
Huang et. al. (2013), Shen et. al. (2014)
Multi-word text
Query text
(bag-of-trigrams)
Document title
(bag-of-trigrams)
Neural Network (Deep)
Session DSSM
Mitra (2015)
Multi-word text
Query text
(bag-of-trigrams)
Next query in session
(bag-of-trigrams)
Neural Network (Deep)
Language Model DSSM
Mitra and Craswell (2015)
Multi-word text
Query prefix
(bag-of-trigrams)
Query suffix
(bag-of-trigrams)
Neural Network (Deep)
My first*
embedding
model (2010)
Sampled a small Word-Context bi-
partite graph data from historical Bing
queries.
Compute Pointwise Mutual Information
score for every Word-Context pair.
Each word embedding is the PMI score
with every possible Context node on
the right.
*It’s an old well-known technique in NLP but I
ended up re-discovering it for myself from playing
with data.
My first
embedding
model (2010)
Here are nearest neighbors based on
cosine similarity between these high
dimensional word embeddings.
You don’t need a neural network to
learn an embedding.
In fact…
Levy et. al. (2014) demonstrated
that the Positive-PMI based
vector representation of words
can be used for analogy tasks
and gives comparable
performance to Word2vec!
Levy et. al. Linguistic regularities in sparse and explicit word representations. (CoNLL, 2015)
The elegance is in the (machine
learning) model, but the magic is in
the structure of the information we
model.
…but
Neural Networks do have certain favorable attributes that lend them
well to learning embeddings
• Embeddings are a by-product of every Neural Network model!
• The output of any intermediate layer is a vector of real numbers – voila,
embedding (of something)!
• Often easier to batch train on large datasets than big matrix
factorizations or graph based approaches
• May be better at modelling non-linearities in the input space
Not all embeddings are created
equal.
The allure of a universal embedding
• The source-target training pairs strictly dictate what notion of
relatedness will be modelled in the embedding space
Is eminem more similar to rihanna or rap?
Is yale more similar to harvard or alumni?
Is seahawks more similar to broncos or seattle?
• Be very careful of using pre-trained embeddings as inputs to a
different model – you may be better off using either one-hot
representations or random initializations!
Typical vs. Topical similarity
If you train a DSSM on query prefix-suffix pairs you get a notion of
relatedness that is based on Type, as opposed to the Topical model
you get by training on query-document pairs
Primary vs. sub-intent similarity
If you train a DSSM on query-answer pairs you get a notion of
relatedness focused more on sub-intents rather than the primary
intent compared to the query-document model
Query-Document DSSM Query-Answer DSSM
What if I told you that everyone
who uses Word2vec is throwing half
the model away?
Using Word2vec for document ranking
Nalisnick, Mitra, Craswell and Caruana.
Improving Document Ranking with Dual
Word Embeddings. Submitted to WWW.
(2016)
Think about…
What makes embedding vectors compose-able?
How can we go from word vectors to sentence
vectors to document vectors?
Are paths in the query/document
embedding space semantically useful?
(e.g., for modelling search sessions)
Single embedding spaces for multiple types of information objects
(e.g., queries, documents, entities, etc.)
Vs.
Multiple embeddings for the same information object
(e.g., typical and topical embeddings for queries).
What is there a difference between learning
embeddings for knowledge and embeddings for
text and other surface forms?
References
• Public code / toolkits I use
• Computational Network Toolkit (CNTK)
• Sent2vec (DSSM)
• Word2vec
• Random reading list
• Omar Levy’s presentation on analogies using non-neural embeddings
• Marek Rei’s Deep Learning Summer School notes
• Piotr Mirowski’s talk on Representation Learning for NLP
“A robot will be truly autonomous when you instruct it to go
to work and it decides to go to the beach instead.”
- Brad Templeton
Thank You for listening!
(Please send any questions to bmitra@microsoft.com)

Mais conteúdo relacionado

Mais procurados

Deep Learning With Python | Deep Learning And Neural Networks | Deep Learning...
Deep Learning With Python | Deep Learning And Neural Networks | Deep Learning...Deep Learning With Python | Deep Learning And Neural Networks | Deep Learning...
Deep Learning With Python | Deep Learning And Neural Networks | Deep Learning...Simplilearn
 
Textrank algorithm
Textrank algorithmTextrank algorithm
Textrank algorithmAndrew Koo
 
位置情報にまつわるデータ補間技術
位置情報にまつわるデータ補間技術位置情報にまつわるデータ補間技術
位置情報にまつわるデータ補間技術Hiroaki Sengoku
 
統計的学習理論チュートリアル: 基礎から応用まで (Ibis2012)
統計的学習理論チュートリアル: 基礎から応用まで (Ibis2012)統計的学習理論チュートリアル: 基礎から応用まで (Ibis2012)
統計的学習理論チュートリアル: 基礎から応用まで (Ibis2012)Taiji Suzuki
 
Deep Learning for Natural Language Processing: Word Embeddings
Deep Learning for Natural Language Processing: Word EmbeddingsDeep Learning for Natural Language Processing: Word Embeddings
Deep Learning for Natural Language Processing: Word EmbeddingsRoelof Pieters
 
Conceptual foundations of text mining and preprocessing steps nfaoui el_habib
Conceptual foundations of text mining and preprocessing steps nfaoui el_habibConceptual foundations of text mining and preprocessing steps nfaoui el_habib
Conceptual foundations of text mining and preprocessing steps nfaoui el_habibEl Habib NFAOUI
 
Image captioning with Keras and Tensorflow - Debarko De @ Practo
Image captioning with Keras and Tensorflow - Debarko De @ PractoImage captioning with Keras and Tensorflow - Debarko De @ Practo
Image captioning with Keras and Tensorflow - Debarko De @ PractoDebarko De
 
Ordinal Regression and Machine Learning: Applications, Methods, Metrics
Ordinal Regression and Machine Learning: Applications, Methods, MetricsOrdinal Regression and Machine Learning: Applications, Methods, Metrics
Ordinal Regression and Machine Learning: Applications, Methods, MetricsFrancesco Casalegno
 
Image histogram and histogram equalization 影像直方圖與直方圖均化
Image histogram and histogram equalization 影像直方圖與直方圖均化Image histogram and histogram equalization 影像直方圖與直方圖均化
Image histogram and histogram equalization 影像直方圖與直方圖均化YKLee3434
 
BigGAN: Large Scale GAN Training for High Fidelity Natural Image Synthesis
BigGAN: Large Scale GAN Training for High Fidelity Natural Image SynthesisBigGAN: Large Scale GAN Training for High Fidelity Natural Image Synthesis
BigGAN: Large Scale GAN Training for High Fidelity Natural Image SynthesisYoung Seok Kim
 
All about that pooling
All about that poolingAll about that pooling
All about that poolingDong Heon Cho
 
論文紹介資料「Quantum Deep Field : Data-Driven Wave Function ...」
論文紹介資料「Quantum Deep Field : Data-Driven Wave Function ...」論文紹介資料「Quantum Deep Field : Data-Driven Wave Function ...」
論文紹介資料「Quantum Deep Field : Data-Driven Wave Function ...」DaikiKoge
 
Attention mechanism 소개 자료
Attention mechanism 소개 자료Attention mechanism 소개 자료
Attention mechanism 소개 자료Whi Kwon
 

Mais procurados (15)

Deep Learning With Python | Deep Learning And Neural Networks | Deep Learning...
Deep Learning With Python | Deep Learning And Neural Networks | Deep Learning...Deep Learning With Python | Deep Learning And Neural Networks | Deep Learning...
Deep Learning With Python | Deep Learning And Neural Networks | Deep Learning...
 
Textrank algorithm
Textrank algorithmTextrank algorithm
Textrank algorithm
 
位置情報にまつわるデータ補間技術
位置情報にまつわるデータ補間技術位置情報にまつわるデータ補間技術
位置情報にまつわるデータ補間技術
 
R-CNN
R-CNNR-CNN
R-CNN
 
統計的学習理論チュートリアル: 基礎から応用まで (Ibis2012)
統計的学習理論チュートリアル: 基礎から応用まで (Ibis2012)統計的学習理論チュートリアル: 基礎から応用まで (Ibis2012)
統計的学習理論チュートリアル: 基礎から応用まで (Ibis2012)
 
Deep Learning for Natural Language Processing: Word Embeddings
Deep Learning for Natural Language Processing: Word EmbeddingsDeep Learning for Natural Language Processing: Word Embeddings
Deep Learning for Natural Language Processing: Word Embeddings
 
Conceptual foundations of text mining and preprocessing steps nfaoui el_habib
Conceptual foundations of text mining and preprocessing steps nfaoui el_habibConceptual foundations of text mining and preprocessing steps nfaoui el_habib
Conceptual foundations of text mining and preprocessing steps nfaoui el_habib
 
Image captioning with Keras and Tensorflow - Debarko De @ Practo
Image captioning with Keras and Tensorflow - Debarko De @ PractoImage captioning with Keras and Tensorflow - Debarko De @ Practo
Image captioning with Keras and Tensorflow - Debarko De @ Practo
 
Ordinal Regression and Machine Learning: Applications, Methods, Metrics
Ordinal Regression and Machine Learning: Applications, Methods, MetricsOrdinal Regression and Machine Learning: Applications, Methods, Metrics
Ordinal Regression and Machine Learning: Applications, Methods, Metrics
 
Image histogram and histogram equalization 影像直方圖與直方圖均化
Image histogram and histogram equalization 影像直方圖與直方圖均化Image histogram and histogram equalization 影像直方圖與直方圖均化
Image histogram and histogram equalization 影像直方圖與直方圖均化
 
BigGAN: Large Scale GAN Training for High Fidelity Natural Image Synthesis
BigGAN: Large Scale GAN Training for High Fidelity Natural Image SynthesisBigGAN: Large Scale GAN Training for High Fidelity Natural Image Synthesis
BigGAN: Large Scale GAN Training for High Fidelity Natural Image Synthesis
 
The Origin of Grad-CAM
The Origin of Grad-CAMThe Origin of Grad-CAM
The Origin of Grad-CAM
 
All about that pooling
All about that poolingAll about that pooling
All about that pooling
 
論文紹介資料「Quantum Deep Field : Data-Driven Wave Function ...」
論文紹介資料「Quantum Deep Field : Data-Driven Wave Function ...」論文紹介資料「Quantum Deep Field : Data-Driven Wave Function ...」
論文紹介資料「Quantum Deep Field : Data-Driven Wave Function ...」
 
Attention mechanism 소개 자료
Attention mechanism 소개 자료Attention mechanism 소개 자료
Attention mechanism 소개 자료
 

Destaque

Deep Learning & NLP: Graphs to the Rescue!
Deep Learning & NLP: Graphs to the Rescue!Deep Learning & NLP: Graphs to the Rescue!
Deep Learning & NLP: Graphs to the Rescue!Roelof Pieters
 
Cs231n 2017 lecture11 Detection and Segmentation
Cs231n 2017 lecture11 Detection and SegmentationCs231n 2017 lecture11 Detection and Segmentation
Cs231n 2017 lecture11 Detection and SegmentationYanbin Kong
 
Cs231n 2017 lecture12 Visualizing and Understanding
Cs231n 2017 lecture12 Visualizing and UnderstandingCs231n 2017 lecture12 Visualizing and Understanding
Cs231n 2017 lecture12 Visualizing and UnderstandingYanbin Kong
 
Deep Learning for Chatbot (3/4)
Deep Learning for Chatbot (3/4)Deep Learning for Chatbot (3/4)
Deep Learning for Chatbot (3/4)Jaemin Cho
 
Satoshi Sonoh - 2015 - Toshiba MT System Description for the WAT2015 Workshop
Satoshi Sonoh - 2015 - Toshiba MT System Description for the WAT2015 WorkshopSatoshi Sonoh - 2015 - Toshiba MT System Description for the WAT2015 Workshop
Satoshi Sonoh - 2015 - Toshiba MT System Description for the WAT2015 WorkshopAssociation for Computational Linguistics
 
John Richardson - 2015 - KyotoEBMT System Description for the 2nd Workshop on...
John Richardson - 2015 - KyotoEBMT System Description for the 2nd Workshop on...John Richardson - 2015 - KyotoEBMT System Description for the 2nd Workshop on...
John Richardson - 2015 - KyotoEBMT System Description for the 2nd Workshop on...Association for Computational Linguistics
 
Using Text Embeddings for Information Retrieval
Using Text Embeddings for Information RetrievalUsing Text Embeddings for Information Retrieval
Using Text Embeddings for Information RetrievalBhaskar Mitra
 
Satoshi Sonoh - 2015 - Toshiba MT System Description for the WAT2015 Workshop
Satoshi Sonoh - 2015 - Toshiba MT System Description for the WAT2015 WorkshopSatoshi Sonoh - 2015 - Toshiba MT System Description for the WAT2015 Workshop
Satoshi Sonoh - 2015 - Toshiba MT System Description for the WAT2015 WorkshopAssociation for Computational Linguistics
 
Advanced Node.JS Meetup
Advanced Node.JS MeetupAdvanced Node.JS Meetup
Advanced Node.JS MeetupLINAGORA
 
Neural Models for Document Ranking
Neural Models for Document RankingNeural Models for Document Ranking
Neural Models for Document RankingBhaskar Mitra
 
John Richardson - 2015 - KyotoEBMT System Description for the 2nd Workshop on...
John Richardson - 2015 - KyotoEBMT System Description for the 2nd Workshop on...John Richardson - 2015 - KyotoEBMT System Description for the 2nd Workshop on...
John Richardson - 2015 - KyotoEBMT System Description for the 2nd Workshop on...Association for Computational Linguistics
 
Zhongyuan Zhu - 2015 - Evaluating Neural Machine Translation in English-Japan...
Zhongyuan Zhu - 2015 - Evaluating Neural Machine Translation in English-Japan...Zhongyuan Zhu - 2015 - Evaluating Neural Machine Translation in English-Japan...
Zhongyuan Zhu - 2015 - Evaluating Neural Machine Translation in English-Japan...Association for Computational Linguistics
 
Matthew Marge - 2017 - Exploring Variation of Natural Human Commands to a Rob...
Matthew Marge - 2017 - Exploring Variation of Natural Human Commands to a Rob...Matthew Marge - 2017 - Exploring Variation of Natural Human Commands to a Rob...
Matthew Marge - 2017 - Exploring Variation of Natural Human Commands to a Rob...Association for Computational Linguistics
 
State of Blockchain 2017: Smartnetworks and the Blockchain Economy
State of Blockchain 2017:  Smartnetworks and the Blockchain EconomyState of Blockchain 2017:  Smartnetworks and the Blockchain Economy
State of Blockchain 2017: Smartnetworks and the Blockchain EconomyMelanie Swan
 
Technological Unemployment and the Robo-Economy
Technological Unemployment and the Robo-EconomyTechnological Unemployment and the Robo-Economy
Technological Unemployment and the Robo-EconomyMelanie Swan
 
Blockchain Smartnetworks: Bitcoin and Blockchain Explained
Blockchain Smartnetworks: Bitcoin and Blockchain ExplainedBlockchain Smartnetworks: Bitcoin and Blockchain Explained
Blockchain Smartnetworks: Bitcoin and Blockchain ExplainedMelanie Swan
 
Deep Learning, an interactive introduction for NLP-ers
Deep Learning, an interactive introduction for NLP-ersDeep Learning, an interactive introduction for NLP-ers
Deep Learning, an interactive introduction for NLP-ersRoelof Pieters
 
iPhone5c的最后猜测
iPhone5c的最后猜测iPhone5c的最后猜测
iPhone5c的最后猜测Yanbin Kong
 
Philosophy of Deep Learning
Philosophy of Deep LearningPhilosophy of Deep Learning
Philosophy of Deep LearningMelanie Swan
 

Destaque (20)

Deep Learning & NLP: Graphs to the Rescue!
Deep Learning & NLP: Graphs to the Rescue!Deep Learning & NLP: Graphs to the Rescue!
Deep Learning & NLP: Graphs to the Rescue!
 
Cs231n 2017 lecture11 Detection and Segmentation
Cs231n 2017 lecture11 Detection and SegmentationCs231n 2017 lecture11 Detection and Segmentation
Cs231n 2017 lecture11 Detection and Segmentation
 
Cs231n 2017 lecture12 Visualizing and Understanding
Cs231n 2017 lecture12 Visualizing and UnderstandingCs231n 2017 lecture12 Visualizing and Understanding
Cs231n 2017 lecture12 Visualizing and Understanding
 
Deep Learning for Chatbot (3/4)
Deep Learning for Chatbot (3/4)Deep Learning for Chatbot (3/4)
Deep Learning for Chatbot (3/4)
 
Satoshi Sonoh - 2015 - Toshiba MT System Description for the WAT2015 Workshop
Satoshi Sonoh - 2015 - Toshiba MT System Description for the WAT2015 WorkshopSatoshi Sonoh - 2015 - Toshiba MT System Description for the WAT2015 Workshop
Satoshi Sonoh - 2015 - Toshiba MT System Description for the WAT2015 Workshop
 
John Richardson - 2015 - KyotoEBMT System Description for the 2nd Workshop on...
John Richardson - 2015 - KyotoEBMT System Description for the 2nd Workshop on...John Richardson - 2015 - KyotoEBMT System Description for the 2nd Workshop on...
John Richardson - 2015 - KyotoEBMT System Description for the 2nd Workshop on...
 
Using Text Embeddings for Information Retrieval
Using Text Embeddings for Information RetrievalUsing Text Embeddings for Information Retrieval
Using Text Embeddings for Information Retrieval
 
Satoshi Sonoh - 2015 - Toshiba MT System Description for the WAT2015 Workshop
Satoshi Sonoh - 2015 - Toshiba MT System Description for the WAT2015 WorkshopSatoshi Sonoh - 2015 - Toshiba MT System Description for the WAT2015 Workshop
Satoshi Sonoh - 2015 - Toshiba MT System Description for the WAT2015 Workshop
 
Advanced Node.JS Meetup
Advanced Node.JS MeetupAdvanced Node.JS Meetup
Advanced Node.JS Meetup
 
Neural Models for Document Ranking
Neural Models for Document RankingNeural Models for Document Ranking
Neural Models for Document Ranking
 
Chenchen Ding - 2015 - NICT at WAT 2015
Chenchen Ding - 2015 - NICT at WAT 2015Chenchen Ding - 2015 - NICT at WAT 2015
Chenchen Ding - 2015 - NICT at WAT 2015
 
John Richardson - 2015 - KyotoEBMT System Description for the 2nd Workshop on...
John Richardson - 2015 - KyotoEBMT System Description for the 2nd Workshop on...John Richardson - 2015 - KyotoEBMT System Description for the 2nd Workshop on...
John Richardson - 2015 - KyotoEBMT System Description for the 2nd Workshop on...
 
Zhongyuan Zhu - 2015 - Evaluating Neural Machine Translation in English-Japan...
Zhongyuan Zhu - 2015 - Evaluating Neural Machine Translation in English-Japan...Zhongyuan Zhu - 2015 - Evaluating Neural Machine Translation in English-Japan...
Zhongyuan Zhu - 2015 - Evaluating Neural Machine Translation in English-Japan...
 
Matthew Marge - 2017 - Exploring Variation of Natural Human Commands to a Rob...
Matthew Marge - 2017 - Exploring Variation of Natural Human Commands to a Rob...Matthew Marge - 2017 - Exploring Variation of Natural Human Commands to a Rob...
Matthew Marge - 2017 - Exploring Variation of Natural Human Commands to a Rob...
 
State of Blockchain 2017: Smartnetworks and the Blockchain Economy
State of Blockchain 2017:  Smartnetworks and the Blockchain EconomyState of Blockchain 2017:  Smartnetworks and the Blockchain Economy
State of Blockchain 2017: Smartnetworks and the Blockchain Economy
 
Technological Unemployment and the Robo-Economy
Technological Unemployment and the Robo-EconomyTechnological Unemployment and the Robo-Economy
Technological Unemployment and the Robo-Economy
 
Blockchain Smartnetworks: Bitcoin and Blockchain Explained
Blockchain Smartnetworks: Bitcoin and Blockchain ExplainedBlockchain Smartnetworks: Bitcoin and Blockchain Explained
Blockchain Smartnetworks: Bitcoin and Blockchain Explained
 
Deep Learning, an interactive introduction for NLP-ers
Deep Learning, an interactive introduction for NLP-ersDeep Learning, an interactive introduction for NLP-ers
Deep Learning, an interactive introduction for NLP-ers
 
iPhone5c的最后猜测
iPhone5c的最后猜测iPhone5c的最后猜测
iPhone5c的最后猜测
 
Philosophy of Deep Learning
Philosophy of Deep LearningPhilosophy of Deep Learning
Philosophy of Deep Learning
 

Semelhante a Vectorland: Brief Notes from Using Text Embeddings for Search

5 Lessons Learned from Designing Neural Models for Information Retrieval
5 Lessons Learned from Designing Neural Models for Information Retrieval5 Lessons Learned from Designing Neural Models for Information Retrieval
5 Lessons Learned from Designing Neural Models for Information RetrievalBhaskar Mitra
 
Neural Models for Information Retrieval
Neural Models for Information RetrievalNeural Models for Information Retrieval
Neural Models for Information RetrievalBhaskar Mitra
 
Continuous bag of words cbow word2vec word embedding work .pdf
Continuous bag of words cbow word2vec word embedding work .pdfContinuous bag of words cbow word2vec word embedding work .pdf
Continuous bag of words cbow word2vec word embedding work .pdfdevangmittal4
 
Neural Models for Information Retrieval
Neural Models for Information RetrievalNeural Models for Information Retrieval
Neural Models for Information RetrievalBhaskar Mitra
 
Dual Embedding Space Model (DESM)
Dual Embedding Space Model (DESM)Dual Embedding Space Model (DESM)
Dual Embedding Space Model (DESM)Bhaskar Mitra
 
Bridging the gap between AI and UI - DSI Vienna - full version
Bridging the gap between AI and UI - DSI Vienna - full versionBridging the gap between AI and UI - DSI Vienna - full version
Bridging the gap between AI and UI - DSI Vienna - full versionLiad Magen
 
Using topic modelling frameworks for NLP and semantic search
Using topic modelling frameworks for NLP and semantic searchUsing topic modelling frameworks for NLP and semantic search
Using topic modelling frameworks for NLP and semantic searchDawn Anderson MSc DigM
 
COMPREHENSIVE ANALYSIS OF NATURAL LANGUAGE PROCESSING TECHNIQUE
COMPREHENSIVE ANALYSIS OF NATURAL LANGUAGE PROCESSING TECHNIQUECOMPREHENSIVE ANALYSIS OF NATURAL LANGUAGE PROCESSING TECHNIQUE
COMPREHENSIVE ANALYSIS OF NATURAL LANGUAGE PROCESSING TECHNIQUEJournal For Research
 
From Linked Data to Semantic Applications
From Linked Data to Semantic ApplicationsFrom Linked Data to Semantic Applications
From Linked Data to Semantic ApplicationsAndre Freitas
 
Neural Text Embeddings for Information Retrieval (WSDM 2017)
Neural Text Embeddings for Information Retrieval (WSDM 2017)Neural Text Embeddings for Information Retrieval (WSDM 2017)
Neural Text Embeddings for Information Retrieval (WSDM 2017)Bhaskar Mitra
 
Challenges in transfer learning in nlp
Challenges in transfer learning in nlpChallenges in transfer learning in nlp
Challenges in transfer learning in nlpLaraOlmosCamarena
 
Deep Learning for Information Retrieval: Models, Progress, & Opportunities
Deep Learning for Information Retrieval: Models, Progress, & OpportunitiesDeep Learning for Information Retrieval: Models, Progress, & Opportunities
Deep Learning for Information Retrieval: Models, Progress, & OpportunitiesMatthew Lease
 
Exploring Session Context using Distributed Representations of Queries and Re...
Exploring Session Context using Distributed Representations of Queries and Re...Exploring Session Context using Distributed Representations of Queries and Re...
Exploring Session Context using Distributed Representations of Queries and Re...Bhaskar Mitra
 
Top 5 MOST VIEWED LANGUAGE COMPUTING ARTICLE - International Journal on Natur...
Top 5 MOST VIEWED LANGUAGE COMPUTING ARTICLE - International Journal on Natur...Top 5 MOST VIEWED LANGUAGE COMPUTING ARTICLE - International Journal on Natur...
Top 5 MOST VIEWED LANGUAGE COMPUTING ARTICLE - International Journal on Natur...kevig
 
THE ABILITY OF WORD EMBEDDINGS TO CAPTURE WORD SIMILARITIES
THE ABILITY OF WORD EMBEDDINGS TO CAPTURE WORD SIMILARITIESTHE ABILITY OF WORD EMBEDDINGS TO CAPTURE WORD SIMILARITIES
THE ABILITY OF WORD EMBEDDINGS TO CAPTURE WORD SIMILARITIESkevig
 
THE ABILITY OF WORD EMBEDDINGS TO CAPTURE WORD SIMILARITIES
THE ABILITY OF WORD EMBEDDINGS TO CAPTURE WORD SIMILARITIESTHE ABILITY OF WORD EMBEDDINGS TO CAPTURE WORD SIMILARITIES
THE ABILITY OF WORD EMBEDDINGS TO CAPTURE WORD SIMILARITIESkevig
 
Semantic Search tutorial at SemTech 2012
Semantic Search tutorial at SemTech 2012Semantic Search tutorial at SemTech 2012
Semantic Search tutorial at SemTech 2012Peter Mika
 
Neural word embedding and language modelling
Neural word embedding and language modellingNeural word embedding and language modelling
Neural word embedding and language modellingRiddhi Jain
 

Semelhante a Vectorland: Brief Notes from Using Text Embeddings for Search (20)

5 Lessons Learned from Designing Neural Models for Information Retrieval
5 Lessons Learned from Designing Neural Models for Information Retrieval5 Lessons Learned from Designing Neural Models for Information Retrieval
5 Lessons Learned from Designing Neural Models for Information Retrieval
 
Neural Models for Information Retrieval
Neural Models for Information RetrievalNeural Models for Information Retrieval
Neural Models for Information Retrieval
 
Continuous bag of words cbow word2vec word embedding work .pdf
Continuous bag of words cbow word2vec word embedding work .pdfContinuous bag of words cbow word2vec word embedding work .pdf
Continuous bag of words cbow word2vec word embedding work .pdf
 
The Duet model
The Duet modelThe Duet model
The Duet model
 
Neural Models for Information Retrieval
Neural Models for Information RetrievalNeural Models for Information Retrieval
Neural Models for Information Retrieval
 
Dual Embedding Space Model (DESM)
Dual Embedding Space Model (DESM)Dual Embedding Space Model (DESM)
Dual Embedding Space Model (DESM)
 
Bridging the gap between AI and UI - DSI Vienna - full version
Bridging the gap between AI and UI - DSI Vienna - full versionBridging the gap between AI and UI - DSI Vienna - full version
Bridging the gap between AI and UI - DSI Vienna - full version
 
Using topic modelling frameworks for NLP and semantic search
Using topic modelling frameworks for NLP and semantic searchUsing topic modelling frameworks for NLP and semantic search
Using topic modelling frameworks for NLP and semantic search
 
COMPREHENSIVE ANALYSIS OF NATURAL LANGUAGE PROCESSING TECHNIQUE
COMPREHENSIVE ANALYSIS OF NATURAL LANGUAGE PROCESSING TECHNIQUECOMPREHENSIVE ANALYSIS OF NATURAL LANGUAGE PROCESSING TECHNIQUE
COMPREHENSIVE ANALYSIS OF NATURAL LANGUAGE PROCESSING TECHNIQUE
 
From Linked Data to Semantic Applications
From Linked Data to Semantic ApplicationsFrom Linked Data to Semantic Applications
From Linked Data to Semantic Applications
 
Neural Text Embeddings for Information Retrieval (WSDM 2017)
Neural Text Embeddings for Information Retrieval (WSDM 2017)Neural Text Embeddings for Information Retrieval (WSDM 2017)
Neural Text Embeddings for Information Retrieval (WSDM 2017)
 
Challenges in transfer learning in nlp
Challenges in transfer learning in nlpChallenges in transfer learning in nlp
Challenges in transfer learning in nlp
 
Deep Learning for Information Retrieval: Models, Progress, & Opportunities
Deep Learning for Information Retrieval: Models, Progress, & OpportunitiesDeep Learning for Information Retrieval: Models, Progress, & Opportunities
Deep Learning for Information Retrieval: Models, Progress, & Opportunities
 
Exploring Session Context using Distributed Representations of Queries and Re...
Exploring Session Context using Distributed Representations of Queries and Re...Exploring Session Context using Distributed Representations of Queries and Re...
Exploring Session Context using Distributed Representations of Queries and Re...
 
Top 5 MOST VIEWED LANGUAGE COMPUTING ARTICLE - International Journal on Natur...
Top 5 MOST VIEWED LANGUAGE COMPUTING ARTICLE - International Journal on Natur...Top 5 MOST VIEWED LANGUAGE COMPUTING ARTICLE - International Journal on Natur...
Top 5 MOST VIEWED LANGUAGE COMPUTING ARTICLE - International Journal on Natur...
 
THE ABILITY OF WORD EMBEDDINGS TO CAPTURE WORD SIMILARITIES
THE ABILITY OF WORD EMBEDDINGS TO CAPTURE WORD SIMILARITIESTHE ABILITY OF WORD EMBEDDINGS TO CAPTURE WORD SIMILARITIES
THE ABILITY OF WORD EMBEDDINGS TO CAPTURE WORD SIMILARITIES
 
THE ABILITY OF WORD EMBEDDINGS TO CAPTURE WORD SIMILARITIES
THE ABILITY OF WORD EMBEDDINGS TO CAPTURE WORD SIMILARITIESTHE ABILITY OF WORD EMBEDDINGS TO CAPTURE WORD SIMILARITIES
THE ABILITY OF WORD EMBEDDINGS TO CAPTURE WORD SIMILARITIES
 
Semantic Search tutorial at SemTech 2012
Semantic Search tutorial at SemTech 2012Semantic Search tutorial at SemTech 2012
Semantic Search tutorial at SemTech 2012
 
Neural word embedding and language modelling
Neural word embedding and language modellingNeural word embedding and language modelling
Neural word embedding and language modelling
 
Image captions.pptx
Image captions.pptxImage captions.pptx
Image captions.pptx
 

Mais de Bhaskar Mitra

Joint Multisided Exposure Fairness for Search and Recommendation
Joint Multisided Exposure Fairness for Search and RecommendationJoint Multisided Exposure Fairness for Search and Recommendation
Joint Multisided Exposure Fairness for Search and RecommendationBhaskar Mitra
 
What’s next for deep learning for Search?
What’s next for deep learning for Search?What’s next for deep learning for Search?
What’s next for deep learning for Search?Bhaskar Mitra
 
So, You Want to Release a Dataset? Reflections on Benchmark Development, Comm...
So, You Want to Release a Dataset? Reflections on Benchmark Development, Comm...So, You Want to Release a Dataset? Reflections on Benchmark Development, Comm...
So, You Want to Release a Dataset? Reflections on Benchmark Development, Comm...Bhaskar Mitra
 
Efficient Machine Learning and Machine Learning for Efficiency in Information...
Efficient Machine Learning and Machine Learning for Efficiency in Information...Efficient Machine Learning and Machine Learning for Efficiency in Information...
Efficient Machine Learning and Machine Learning for Efficiency in Information...Bhaskar Mitra
 
Multisided Exposure Fairness for Search and Recommendation
Multisided Exposure Fairness for Search and RecommendationMultisided Exposure Fairness for Search and Recommendation
Multisided Exposure Fairness for Search and RecommendationBhaskar Mitra
 
Neural Learning to Rank
Neural Learning to RankNeural Learning to Rank
Neural Learning to RankBhaskar Mitra
 
Neural Information Retrieval: In search of meaningful progress
Neural Information Retrieval: In search of meaningful progressNeural Information Retrieval: In search of meaningful progress
Neural Information Retrieval: In search of meaningful progressBhaskar Mitra
 
Conformer-Kernel with Query Term Independence @ TREC 2020 Deep Learning Track
Conformer-Kernel with Query Term Independence @ TREC 2020 Deep Learning TrackConformer-Kernel with Query Term Independence @ TREC 2020 Deep Learning Track
Conformer-Kernel with Query Term Independence @ TREC 2020 Deep Learning TrackBhaskar Mitra
 
Neural Learning to Rank
Neural Learning to RankNeural Learning to Rank
Neural Learning to RankBhaskar Mitra
 
Duet @ TREC 2019 Deep Learning Track
Duet @ TREC 2019 Deep Learning TrackDuet @ TREC 2019 Deep Learning Track
Duet @ TREC 2019 Deep Learning TrackBhaskar Mitra
 
Benchmarking for Neural Information Retrieval: MS MARCO, TREC, and Beyond
Benchmarking for Neural Information Retrieval: MS MARCO, TREC, and BeyondBenchmarking for Neural Information Retrieval: MS MARCO, TREC, and Beyond
Benchmarking for Neural Information Retrieval: MS MARCO, TREC, and BeyondBhaskar Mitra
 
Deep Neural Methods for Retrieval
Deep Neural Methods for RetrievalDeep Neural Methods for Retrieval
Deep Neural Methods for RetrievalBhaskar Mitra
 
Neural Learning to Rank
Neural Learning to RankNeural Learning to Rank
Neural Learning to RankBhaskar Mitra
 
Learning to Rank with Neural Networks
Learning to Rank with Neural NetworksLearning to Rank with Neural Networks
Learning to Rank with Neural NetworksBhaskar Mitra
 
Deep Learning for Search
Deep Learning for SearchDeep Learning for Search
Deep Learning for SearchBhaskar Mitra
 
Deep Learning for Search
Deep Learning for SearchDeep Learning for Search
Deep Learning for SearchBhaskar Mitra
 
Neural Learning to Rank
Neural Learning to RankNeural Learning to Rank
Neural Learning to RankBhaskar Mitra
 
Deep Learning for Search
Deep Learning for SearchDeep Learning for Search
Deep Learning for SearchBhaskar Mitra
 
Adversarial and reinforcement learning-based approaches to information retrieval
Adversarial and reinforcement learning-based approaches to information retrievalAdversarial and reinforcement learning-based approaches to information retrieval
Adversarial and reinforcement learning-based approaches to information retrievalBhaskar Mitra
 
A Simple Introduction to Neural Information Retrieval
A Simple Introduction to Neural Information RetrievalA Simple Introduction to Neural Information Retrieval
A Simple Introduction to Neural Information RetrievalBhaskar Mitra
 

Mais de Bhaskar Mitra (20)

Joint Multisided Exposure Fairness for Search and Recommendation
Joint Multisided Exposure Fairness for Search and RecommendationJoint Multisided Exposure Fairness for Search and Recommendation
Joint Multisided Exposure Fairness for Search and Recommendation
 
What’s next for deep learning for Search?
What’s next for deep learning for Search?What’s next for deep learning for Search?
What’s next for deep learning for Search?
 
So, You Want to Release a Dataset? Reflections on Benchmark Development, Comm...
So, You Want to Release a Dataset? Reflections on Benchmark Development, Comm...So, You Want to Release a Dataset? Reflections on Benchmark Development, Comm...
So, You Want to Release a Dataset? Reflections on Benchmark Development, Comm...
 
Efficient Machine Learning and Machine Learning for Efficiency in Information...
Efficient Machine Learning and Machine Learning for Efficiency in Information...Efficient Machine Learning and Machine Learning for Efficiency in Information...
Efficient Machine Learning and Machine Learning for Efficiency in Information...
 
Multisided Exposure Fairness for Search and Recommendation
Multisided Exposure Fairness for Search and RecommendationMultisided Exposure Fairness for Search and Recommendation
Multisided Exposure Fairness for Search and Recommendation
 
Neural Learning to Rank
Neural Learning to RankNeural Learning to Rank
Neural Learning to Rank
 
Neural Information Retrieval: In search of meaningful progress
Neural Information Retrieval: In search of meaningful progressNeural Information Retrieval: In search of meaningful progress
Neural Information Retrieval: In search of meaningful progress
 
Conformer-Kernel with Query Term Independence @ TREC 2020 Deep Learning Track
Conformer-Kernel with Query Term Independence @ TREC 2020 Deep Learning TrackConformer-Kernel with Query Term Independence @ TREC 2020 Deep Learning Track
Conformer-Kernel with Query Term Independence @ TREC 2020 Deep Learning Track
 
Neural Learning to Rank
Neural Learning to RankNeural Learning to Rank
Neural Learning to Rank
 
Duet @ TREC 2019 Deep Learning Track
Duet @ TREC 2019 Deep Learning TrackDuet @ TREC 2019 Deep Learning Track
Duet @ TREC 2019 Deep Learning Track
 
Benchmarking for Neural Information Retrieval: MS MARCO, TREC, and Beyond
Benchmarking for Neural Information Retrieval: MS MARCO, TREC, and BeyondBenchmarking for Neural Information Retrieval: MS MARCO, TREC, and Beyond
Benchmarking for Neural Information Retrieval: MS MARCO, TREC, and Beyond
 
Deep Neural Methods for Retrieval
Deep Neural Methods for RetrievalDeep Neural Methods for Retrieval
Deep Neural Methods for Retrieval
 
Neural Learning to Rank
Neural Learning to RankNeural Learning to Rank
Neural Learning to Rank
 
Learning to Rank with Neural Networks
Learning to Rank with Neural NetworksLearning to Rank with Neural Networks
Learning to Rank with Neural Networks
 
Deep Learning for Search
Deep Learning for SearchDeep Learning for Search
Deep Learning for Search
 
Deep Learning for Search
Deep Learning for SearchDeep Learning for Search
Deep Learning for Search
 
Neural Learning to Rank
Neural Learning to RankNeural Learning to Rank
Neural Learning to Rank
 
Deep Learning for Search
Deep Learning for SearchDeep Learning for Search
Deep Learning for Search
 
Adversarial and reinforcement learning-based approaches to information retrieval
Adversarial and reinforcement learning-based approaches to information retrievalAdversarial and reinforcement learning-based approaches to information retrieval
Adversarial and reinforcement learning-based approaches to information retrieval
 
A Simple Introduction to Neural Information Retrieval
A Simple Introduction to Neural Information RetrievalA Simple Introduction to Neural Information Retrieval
A Simple Introduction to Neural Information Retrieval
 

Último

Introduction to Viruses
Introduction to VirusesIntroduction to Viruses
Introduction to VirusesAreesha Ahmad
 
Asymmetry in the atmosphere of the ultra-hot Jupiter WASP-76 b
Asymmetry in the atmosphere of the ultra-hot Jupiter WASP-76 bAsymmetry in the atmosphere of the ultra-hot Jupiter WASP-76 b
Asymmetry in the atmosphere of the ultra-hot Jupiter WASP-76 bSérgio Sacani
 
Thyroid Physiology_Dr.E. Muralinath_ Associate Professor
Thyroid Physiology_Dr.E. Muralinath_ Associate ProfessorThyroid Physiology_Dr.E. Muralinath_ Associate Professor
Thyroid Physiology_Dr.E. Muralinath_ Associate Professormuralinath2
 
Locating and isolating a gene, FISH, GISH, Chromosome walking and jumping, te...
Locating and isolating a gene, FISH, GISH, Chromosome walking and jumping, te...Locating and isolating a gene, FISH, GISH, Chromosome walking and jumping, te...
Locating and isolating a gene, FISH, GISH, Chromosome walking and jumping, te...Silpa
 
Justdial Call Girls In Indirapuram, Ghaziabad, 8800357707 Escorts Service
Justdial Call Girls In Indirapuram, Ghaziabad, 8800357707 Escorts ServiceJustdial Call Girls In Indirapuram, Ghaziabad, 8800357707 Escorts Service
Justdial Call Girls In Indirapuram, Ghaziabad, 8800357707 Escorts Servicemonikaservice1
 
Kochi ❤CALL GIRL 84099*07087 ❤CALL GIRLS IN Kochi ESCORT SERVICE❤CALL GIRL
Kochi ❤CALL GIRL 84099*07087 ❤CALL GIRLS IN Kochi ESCORT SERVICE❤CALL GIRLKochi ❤CALL GIRL 84099*07087 ❤CALL GIRLS IN Kochi ESCORT SERVICE❤CALL GIRL
Kochi ❤CALL GIRL 84099*07087 ❤CALL GIRLS IN Kochi ESCORT SERVICE❤CALL GIRLkantirani197
 
Pests of mustard_Identification_Management_Dr.UPR.pdf
Pests of mustard_Identification_Management_Dr.UPR.pdfPests of mustard_Identification_Management_Dr.UPR.pdf
Pests of mustard_Identification_Management_Dr.UPR.pdfPirithiRaju
 
Sector 62, Noida Call girls :8448380779 Model Escorts | 100% verified
Sector 62, Noida Call girls :8448380779 Model Escorts | 100% verifiedSector 62, Noida Call girls :8448380779 Model Escorts | 100% verified
Sector 62, Noida Call girls :8448380779 Model Escorts | 100% verifiedDelhi Call girls
 
Forensic Biology & Its biological significance.pdf
Forensic Biology & Its biological significance.pdfForensic Biology & Its biological significance.pdf
Forensic Biology & Its biological significance.pdfrohankumarsinghrore1
 
❤Jammu Kashmir Call Girls 8617697112 Personal Whatsapp Number 💦✅.
❤Jammu Kashmir Call Girls 8617697112 Personal Whatsapp Number 💦✅.❤Jammu Kashmir Call Girls 8617697112 Personal Whatsapp Number 💦✅.
❤Jammu Kashmir Call Girls 8617697112 Personal Whatsapp Number 💦✅.Nitya salvi
 
GBSN - Biochemistry (Unit 1)
GBSN - Biochemistry (Unit 1)GBSN - Biochemistry (Unit 1)
GBSN - Biochemistry (Unit 1)Areesha Ahmad
 
High Class Escorts in Hyderabad ₹7.5k Pick Up & Drop With Cash Payment 969456...
High Class Escorts in Hyderabad ₹7.5k Pick Up & Drop With Cash Payment 969456...High Class Escorts in Hyderabad ₹7.5k Pick Up & Drop With Cash Payment 969456...
High Class Escorts in Hyderabad ₹7.5k Pick Up & Drop With Cash Payment 969456...chandars293
 
Conjugation, transduction and transformation
Conjugation, transduction and transformationConjugation, transduction and transformation
Conjugation, transduction and transformationAreesha Ahmad
 
Module for Grade 9 for Asynchronous/Distance learning
Module for Grade 9 for Asynchronous/Distance learningModule for Grade 9 for Asynchronous/Distance learning
Module for Grade 9 for Asynchronous/Distance learninglevieagacer
 
Digital Dentistry.Digital Dentistryvv.pptx
Digital Dentistry.Digital Dentistryvv.pptxDigital Dentistry.Digital Dentistryvv.pptx
Digital Dentistry.Digital Dentistryvv.pptxMohamedFarag457087
 
GBSN - Microbiology (Unit 1)
GBSN - Microbiology (Unit 1)GBSN - Microbiology (Unit 1)
GBSN - Microbiology (Unit 1)Areesha Ahmad
 
Factory Acceptance Test( FAT).pptx .
Factory Acceptance Test( FAT).pptx       .Factory Acceptance Test( FAT).pptx       .
Factory Acceptance Test( FAT).pptx .Poonam Aher Patil
 
Grade 7 - Lesson 1 - Microscope and Its Functions
Grade 7 - Lesson 1 - Microscope and Its FunctionsGrade 7 - Lesson 1 - Microscope and Its Functions
Grade 7 - Lesson 1 - Microscope and Its FunctionsOrtegaSyrineMay
 

Último (20)

+971581248768>> SAFE AND ORIGINAL ABORTION PILLS FOR SALE IN DUBAI AND ABUDHA...
+971581248768>> SAFE AND ORIGINAL ABORTION PILLS FOR SALE IN DUBAI AND ABUDHA...+971581248768>> SAFE AND ORIGINAL ABORTION PILLS FOR SALE IN DUBAI AND ABUDHA...
+971581248768>> SAFE AND ORIGINAL ABORTION PILLS FOR SALE IN DUBAI AND ABUDHA...
 
Introduction to Viruses
Introduction to VirusesIntroduction to Viruses
Introduction to Viruses
 
Asymmetry in the atmosphere of the ultra-hot Jupiter WASP-76 b
Asymmetry in the atmosphere of the ultra-hot Jupiter WASP-76 bAsymmetry in the atmosphere of the ultra-hot Jupiter WASP-76 b
Asymmetry in the atmosphere of the ultra-hot Jupiter WASP-76 b
 
Thyroid Physiology_Dr.E. Muralinath_ Associate Professor
Thyroid Physiology_Dr.E. Muralinath_ Associate ProfessorThyroid Physiology_Dr.E. Muralinath_ Associate Professor
Thyroid Physiology_Dr.E. Muralinath_ Associate Professor
 
Locating and isolating a gene, FISH, GISH, Chromosome walking and jumping, te...
Locating and isolating a gene, FISH, GISH, Chromosome walking and jumping, te...Locating and isolating a gene, FISH, GISH, Chromosome walking and jumping, te...
Locating and isolating a gene, FISH, GISH, Chromosome walking and jumping, te...
 
Justdial Call Girls In Indirapuram, Ghaziabad, 8800357707 Escorts Service
Justdial Call Girls In Indirapuram, Ghaziabad, 8800357707 Escorts ServiceJustdial Call Girls In Indirapuram, Ghaziabad, 8800357707 Escorts Service
Justdial Call Girls In Indirapuram, Ghaziabad, 8800357707 Escorts Service
 
Kochi ❤CALL GIRL 84099*07087 ❤CALL GIRLS IN Kochi ESCORT SERVICE❤CALL GIRL
Kochi ❤CALL GIRL 84099*07087 ❤CALL GIRLS IN Kochi ESCORT SERVICE❤CALL GIRLKochi ❤CALL GIRL 84099*07087 ❤CALL GIRLS IN Kochi ESCORT SERVICE❤CALL GIRL
Kochi ❤CALL GIRL 84099*07087 ❤CALL GIRLS IN Kochi ESCORT SERVICE❤CALL GIRL
 
Pests of mustard_Identification_Management_Dr.UPR.pdf
Pests of mustard_Identification_Management_Dr.UPR.pdfPests of mustard_Identification_Management_Dr.UPR.pdf
Pests of mustard_Identification_Management_Dr.UPR.pdf
 
Sector 62, Noida Call girls :8448380779 Model Escorts | 100% verified
Sector 62, Noida Call girls :8448380779 Model Escorts | 100% verifiedSector 62, Noida Call girls :8448380779 Model Escorts | 100% verified
Sector 62, Noida Call girls :8448380779 Model Escorts | 100% verified
 
Forensic Biology & Its biological significance.pdf
Forensic Biology & Its biological significance.pdfForensic Biology & Its biological significance.pdf
Forensic Biology & Its biological significance.pdf
 
❤Jammu Kashmir Call Girls 8617697112 Personal Whatsapp Number 💦✅.
❤Jammu Kashmir Call Girls 8617697112 Personal Whatsapp Number 💦✅.❤Jammu Kashmir Call Girls 8617697112 Personal Whatsapp Number 💦✅.
❤Jammu Kashmir Call Girls 8617697112 Personal Whatsapp Number 💦✅.
 
GBSN - Biochemistry (Unit 1)
GBSN - Biochemistry (Unit 1)GBSN - Biochemistry (Unit 1)
GBSN - Biochemistry (Unit 1)
 
High Class Escorts in Hyderabad ₹7.5k Pick Up & Drop With Cash Payment 969456...
High Class Escorts in Hyderabad ₹7.5k Pick Up & Drop With Cash Payment 969456...High Class Escorts in Hyderabad ₹7.5k Pick Up & Drop With Cash Payment 969456...
High Class Escorts in Hyderabad ₹7.5k Pick Up & Drop With Cash Payment 969456...
 
Conjugation, transduction and transformation
Conjugation, transduction and transformationConjugation, transduction and transformation
Conjugation, transduction and transformation
 
Module for Grade 9 for Asynchronous/Distance learning
Module for Grade 9 for Asynchronous/Distance learningModule for Grade 9 for Asynchronous/Distance learning
Module for Grade 9 for Asynchronous/Distance learning
 
Digital Dentistry.Digital Dentistryvv.pptx
Digital Dentistry.Digital Dentistryvv.pptxDigital Dentistry.Digital Dentistryvv.pptx
Digital Dentistry.Digital Dentistryvv.pptx
 
GBSN - Microbiology (Unit 1)
GBSN - Microbiology (Unit 1)GBSN - Microbiology (Unit 1)
GBSN - Microbiology (Unit 1)
 
Factory Acceptance Test( FAT).pptx .
Factory Acceptance Test( FAT).pptx       .Factory Acceptance Test( FAT).pptx       .
Factory Acceptance Test( FAT).pptx .
 
Grade 7 - Lesson 1 - Microscope and Its Functions
Grade 7 - Lesson 1 - Microscope and Its FunctionsGrade 7 - Lesson 1 - Microscope and Its Functions
Grade 7 - Lesson 1 - Microscope and Its Functions
 
Site Acceptance Test .
Site Acceptance Test                    .Site Acceptance Test                    .
Site Acceptance Test .
 

Vectorland: Brief Notes from Using Text Embeddings for Search

  • 1. Bhaskar Mitra, Microsoft (Bing Sciences) Search Solutions, 26th November, 2015
  • 2. A brief introduction… • I have worked as a relevance engineer for Bing since 2007 (then called as Live Search) • Mostly on Web document ranking and query formulation • Moved to an applied research scientist role in 2013 • In this talk I will present… • Some personal learnings and takeaways from working on (neural and non- neural) text embeddings for IR • Highlight a few of my favourite insights/papers from the broader academic community *Thanks to Nick Craswell for suggesting the title “Vectorland”.
  • 3. A recent trend… Mikolov et. al. Efficient Estimation of Word Representations in Vector Space. (2013) Huang et al. Learning deep structured semantic models for web search using clickthrough data. (CIKM, 2013) Hong. Improving Paragraph2Vec. (2015) Grbovic et. al. Context- and Content-aware Embeddings for Query Rewriting in Sponsored Search. (SIGIR, 2015)
  • 4. Or as a learned RNN model once said*… 2vec or not 2vec that is… Recaims alone of those the mercorrance down. Sir, And let it be, if not, no, then; if you get Great rebels most of a heaven, I cannot mose where his hearts makes the Rome arrase. And then it stands: fear them against your honour, I am a sifel loved him; he swores. My lord, yet most gentle in our ears? Our ax I can respect of? If you concear, and lend me to his punishment? If I make upon thee. Let me see how after Wortens of she: is it your sister, pardon! air, I give my recair to depose? *The text above was auto-generated using Andrej Karpathy’s Char-RNN implementation trained on the works of Shakespeare and then seeded with the starting text “to vector or not to vector that is”. Special thanks to Milad Shokouhi for his help with running the RNN model.
  • 5. Learning to represent A lot of recent work in neural models and “Deep Learning” is focused on learning vector representations for text, image, speech, entities, and other nuggets of information
  • 6. Learning to represent From analogies over words and short texts…. Mikolov et. al. Efficient Estimation of Word Representations in Vector Space. (2013) Mitra. Exploring Session Context using Distributed Representations of Queries and Reformulations. (SIGIR, 2015)
  • 7. Learning to represent …and automatically generating natural language captions for images, Vinyals et. al. Show and Tell: A Neural Image Caption Generator. (2015) Fang et. al. From Captions to Visual Concepts and Back. (CVPR, 2015)
  • 8. Learning to represent …to building automated conversational agents. Vinyals et. al. A Neural Conversational Model. (ICML, 2015)
  • 10. One-hot vectors A sparse bit vector where all values are zeros, except one. Each position corresponds to a different item. The vector dimension is equal to the number of items that need to be represented. 0 1 0 0 0 0 0 1
  • 11. Bag-of-* vectors A sparse count vector of component units. The vector dimension is equal to the vocabulary size (number of distinct components). 0 0 0 0 0 1 0 0 0 1 0 0 “web search” (Bag of words) search web 0 1 0 1 0 0 2 0 1 0 1 0 “banana” (Bag of trigrams) ana nan#ba na# ban
  • 12. Embeddings A dense vector of real values. The vector dimension is typically much smaller than the number of items or the vocabulary size. You can imagine the vectors as coordinates for items in the embedding space. Some distance metric defines a notion of relatedness between items in this space.
  • 13. Neighborhoods in an embedding space (Example) Song et. al. Unsupervised Learning of Word Semantic Embedding using the Deep Structured Semantic Model. (2014)
  • 14. Transitions in an embedding space (Example) Mitra. Exploring Session Context using Distributed Representations of Queries and Reformulations. (SIGIR, 2015)
  • 16. Example use-cases for text embeddings in search Learning a joint query and document (title) embedding for document ranking Shen et. al. Learning semantic representations using convolutional neural networks for web search. (WWW, 2014)
  • 17. Example use-cases for text embeddings in search Gao et. al. Modeling Interestingness with Deep Neural Networks. (EMNLP, 2014) Entity detection in document (unstructured) body text
  • 18. Example use-cases for text embeddings in search Mitra and Craswell. Query Auto-Completion for Rare Prefixes. (CIKM, 2015) Predicting suffixes (or next word) for query auto- completion for rare prefixes
  • 19. Example use-cases for text embeddings in search Mitra. Exploring Session Context using Distributed Representations of Queries and Reformulations. (SIGIR, 2015) Session modelling by learning an embedding for query (or intent) transitions
  • 20. Example use-cases for text embeddings in search Nalisnick et. al. Improving Document Ranking with Dual Word Embeddings. (Submitted to WWW, 2016) Modelling the aboutness of a document by capturing evidences from document terms that do no match the query Passage about Albuquerque Passage not about Albuquerque
  • 21. Example use-cases for text embeddings in search Liu et. al. Representation Learning Using Multi-Task Deep Neural Networks for Semantic Classification and Information Retrieval. (NAACL, 2015) Multi-task embedding of queries for classification and document retrieval
  • 22. How do you learn an embedding?
  • 23. How do you (typically) learn an embedding? • Setup a prediction task Source Item → Target Item • Input and Output vectors are sparse • Learning the embedding ≈ Dimensionality reduction (*The bottleneck trick for NNs) • Many options for the actual model • Neural networks, matrix factorization, Pointwise Mutual Information, etc. Target Item Source Item Source Embedding Target Embedding Distance Metric
  • 24. Some examples of text embeddings Embedding for Source Item Target Item Learning Model Latent Semantic Analysis Deerwester et. al. (1990) Single word Word (one-hot) Document (one-hot) Matrix factorization Word2vec Mikolov et. al. (2013) Single Word Word (one-hot) Neighboring Word (one-hot) Neural Network (Shallow) Glove Pennington et. al. (2014) Single Word Word (one-hot) Neighboring Word (one-hot) Matrix factorization Semantic Hashing (auto-encoder) Salakhutdinov and Hinton (2007) Multi-word text Document (bag-of-words) Same as source (bag-of-words) Neural Network (Deep) DSSM Huang et. al. (2013), Shen et. al. (2014) Multi-word text Query text (bag-of-trigrams) Document title (bag-of-trigrams) Neural Network (Deep) Session DSSM Mitra (2015) Multi-word text Query text (bag-of-trigrams) Next query in session (bag-of-trigrams) Neural Network (Deep) Language Model DSSM Mitra and Craswell (2015) Multi-word text Query prefix (bag-of-trigrams) Query suffix (bag-of-trigrams) Neural Network (Deep)
  • 25. My first* embedding model (2010) Sampled a small Word-Context bi- partite graph data from historical Bing queries. Compute Pointwise Mutual Information score for every Word-Context pair. Each word embedding is the PMI score with every possible Context node on the right. *It’s an old well-known technique in NLP but I ended up re-discovering it for myself from playing with data.
  • 26. My first embedding model (2010) Here are nearest neighbors based on cosine similarity between these high dimensional word embeddings.
  • 27. You don’t need a neural network to learn an embedding.
  • 28. In fact… Levy et. al. (2014) demonstrated that the Positive-PMI based vector representation of words can be used for analogy tasks and gives comparable performance to Word2vec! Levy et. al. Linguistic regularities in sparse and explicit word representations. (CoNLL, 2015)
  • 29. The elegance is in the (machine learning) model, but the magic is in the structure of the information we model.
  • 30. …but Neural Networks do have certain favorable attributes that lend them well to learning embeddings • Embeddings are a by-product of every Neural Network model! • The output of any intermediate layer is a vector of real numbers – voila, embedding (of something)! • Often easier to batch train on large datasets than big matrix factorizations or graph based approaches • May be better at modelling non-linearities in the input space
  • 31. Not all embeddings are created equal.
  • 32. The allure of a universal embedding • The source-target training pairs strictly dictate what notion of relatedness will be modelled in the embedding space Is eminem more similar to rihanna or rap? Is yale more similar to harvard or alumni? Is seahawks more similar to broncos or seattle? • Be very careful of using pre-trained embeddings as inputs to a different model – you may be better off using either one-hot representations or random initializations!
  • 33. Typical vs. Topical similarity If you train a DSSM on query prefix-suffix pairs you get a notion of relatedness that is based on Type, as opposed to the Topical model you get by training on query-document pairs
  • 34. Primary vs. sub-intent similarity If you train a DSSM on query-answer pairs you get a notion of relatedness focused more on sub-intents rather than the primary intent compared to the query-document model Query-Document DSSM Query-Answer DSSM
  • 35. What if I told you that everyone who uses Word2vec is throwing half the model away?
  • 36. Using Word2vec for document ranking Nalisnick, Mitra, Craswell and Caruana. Improving Document Ranking with Dual Word Embeddings. Submitted to WWW. (2016)
  • 37. Think about… What makes embedding vectors compose-able? How can we go from word vectors to sentence vectors to document vectors? Are paths in the query/document embedding space semantically useful? (e.g., for modelling search sessions) Single embedding spaces for multiple types of information objects (e.g., queries, documents, entities, etc.) Vs. Multiple embeddings for the same information object (e.g., typical and topical embeddings for queries). What is there a difference between learning embeddings for knowledge and embeddings for text and other surface forms?
  • 38. References • Public code / toolkits I use • Computational Network Toolkit (CNTK) • Sent2vec (DSSM) • Word2vec • Random reading list • Omar Levy’s presentation on analogies using non-neural embeddings • Marek Rei’s Deep Learning Summer School notes • Piotr Mirowski’s talk on Representation Learning for NLP
  • 39. “A robot will be truly autonomous when you instruct it to go to work and it decides to go to the beach instead.” - Brad Templeton Thank You for listening! (Please send any questions to bmitra@microsoft.com)