SlideShare uma empresa Scribd logo
1 de 23
Baixar para ler offline
predicting the age of social network users
from user-generated texts with word
embeddings
Anton Alekseev1
and Sergey Nikolenko1,2,3
AINL 2016
St. Petersburg, November 11, 2016
1
Steklov Institute of Mathematics at St. Petersburg
2
Kazan Federal University, Kazan, Russia
3
Laboratory for Internet Studies, NRU Higher School of Economics, St. Petersburg
deep learning for nlp
• Recent success on numerous tasks: from text classification to
syntax parsing.
• Various applications of word vectors pretrained on huge
corpora: word2vec, GloVe, etc.
• In this work we have tried a few approaches for social network
users profiling using full-text items.
2
dataset
• 868,126 users of the Odnoklassniki social network. In particular,
it contains the following data:
(1) demographic information on 868,126 users of the network:
gender, age, and region;
(2) “friendship” relation, several different type of links: “friend”, “love”,
“spouse”, “parent”, and so on;
(3) history of logins for individual users;
(4) data on the “likes” (“class” marks) a user has given to other users’
statuses and posts in various groups;
(5) texts of the posts for individual users;
(6) communities (“groups”) statuses that have been liked by these
selected users.
3
dataset
• The dataset has been created as follows:
(1) 486 seed users;
(2) for these users, their sets of friends have been extracted;
(3) then the friends of these friends; as a result — seed users
neighborhood of depth 2 in the social graph.
4
dataset: age
• Mean age — 31.39 years;
0 20 40 60 80 100 120
Age
100
101
102
103
104
105
Number of users
Figure 1: Users’ age in the Odnoklassniki dataset
• Note users with implausible ages (ages 2 and 3, age higher than
100 years).
• We have removed all users with ages below 10 and above 80.
5
dataset: friends
100 101 102 103 104
Number of friends
100
101
102
103
104
Number of users
Figure 2: Number of friends distribution in the Odnoklassniki dataset
• “Arc” (not a straight line!) on a log-log plot — probably due to
the collection method: smaller number of degree-0, degree-1,
etc. vertices (that is, users) than in the whole “social graph”.
6
dataset and setup for word embeddings
A large Russian-language corpus with about 14G tokens in 2.5M
documents. This corpus includes:
• Russian Wikipedia: 1.15M documents, 238M tokens;
• automated Web crawl data: 890K documents, 568M tokens;
• (main part) the huge lib.rus.ec library corpus: 234K documents,
12.9G tokens;
• user statuses and group posts from the Odnoklassniki social
network, as described above.
For more info, please see the article cited in the paper:
Arefyev N., Panchenko A., Lukanin A., Lesota O., Romanov P. Evaluating Three Corpus-Based Semantic Similarity Systems for
Russian. In Proceedings of International Conference on Computational Linguistics Dialogue 2015, 2015
7
dataset and setup for word embeddings
• 𝑤𝑜𝑟𝑑2𝑣𝑒𝑐 configurations we took into consideration
Table 1: Word2vec models trained on large Russian corpora. The
columns are: dimension 𝑑, window size 𝑤, number of negative samples
𝑛, vocabulary threshold 𝑣, and the resulting model size
Type 𝑑 𝑤 𝑛 𝑣 Size
CBOW 100 11 10 20 1.3G
CBOW 100 11 10 30 0.97G
skip-gram 100 11 10 30 0.97G
skip-gram 200 11 1 20 1.3G
CBOW 200 11 1 30 2.0G
skip-gram 200 11 1 30 2.0G
CBOW 300 11 1 30 2.9G
skip-gram 300 11 1 30 2.9G
8
chosen base approaches
• How well is user’s age approximated by mean friends age?
(1) MeanAge: predict age with the mean of friends’ ages;
(2) LinearRegr: linear regression with a single feature
(mean friends age);
(3) ElasticNet: elastic net regressor with a single feature
(mean friends age);
(4) GradBoost: gradient boosting with a single feature
(mean friends age).
9
algorithms based on word embeddings
• The sum and/or mean of word embeddings to represent a
sentence/paragraph.
• The idea is not as naive as it seems; successfully applied in a
few studies (please see the paper for references).
• Three basic algorithms:
(1) MeanVec: train on mean vectors of all statuses for a user;
(2) LargestCluster: train on the centroid of the largest cluster of
statuses;
(3) AllMeanV: train on every status independently, with the mean
vector of a specific status and the mean age of friends as features
and the user’s demography as target; at the testing stage, we
compute predictions for every status and average the predictions.
10
meanvec
(1) for every user 𝑢 ∈ 𝑈:
• for every status 𝑠 ∈ 𝑆 𝑢, compute its mean vector ̄v 𝑠 = 1
|𝑠|
∑ 𝑤∈𝑆
v 𝑤;
• compute the mean vector of all statuses ̄v 𝑢 = 1
|𝑆 𝑢|
∑ 𝑠∈𝑆 𝑢
̄v 𝑠;
(2) train ML with features (MAF 𝑢, ̄v 𝑢) for every 𝑢 ∈ 𝑈.
11
largestcluster
The LargestCluster algorithm operates as follows: for a machine
learning algorithm ML,
(1) for every user 𝑢 ∈ 𝑈:
• for every status 𝑠 ∈ 𝑆 𝑢, compute its mean vector ̄v 𝑠 = 1
|𝑠|
∑ 𝑤∈𝑆
v 𝑤;
• cluster the set of vectors { ̄v 𝑠 ∣ 𝑠 ∈ 𝑆} into two clusters with
agglomerative clustering; denote by 𝐶 ⊆ 𝑆 the larger cluster;
• compute the mean vector of statuses from 𝐶 ̄c 𝑢 = 1
|𝐶|
∑ 𝑠∈𝐶
̄v 𝑠;
(2) train ML with features (MAF 𝑢, ̄c 𝑢) for every 𝑢 ∈ 𝑈.
12
allmeanv
The AllMeanV algorithm operates as follows: for a machine learning
algorithm ML,
(1) for every user 𝑢 ∈ 𝑈 and every status 𝑠 ∈ 𝑆 𝑢, compute its mean
vector ̄v 𝑠 = 1
|𝑠|
∑ 𝑤∈𝑆
v 𝑤;
(2) train ML with features (MAF 𝑢, ̄v 𝑠) for every 𝑢 ∈ 𝑈 and 𝑠 ∈ 𝑆 𝑢;
(3) on the prediction stage, for a user 𝑢 ∈ 𝑈test:
• for every status 𝑠 ∈ 𝑆 𝑢, compute its mean vector ̄v 𝑠 = 1
|𝑠|
∑ 𝑤∈𝑆
v 𝑤;
• predict the age for this status, 𝑎 𝑠 = ML(MAF 𝑢, ̄v 𝑠);
• return the average predicted age, 𝑎 = 1
|𝑆 𝑢|
∑ 𝑠∈𝑆 𝑢
ML(MAF 𝑢, ̄v 𝑠).
13
evaluation: prepared data
Train/test sets were splitted 80:20.
Dataset Training set Test set
Users Statuses Users Statuses
661206 10880321 165301 2704883
14
evaluation: comparing meanvec models using word2vec with
different parameters
• Adding word2vec components do help all models.
• CBOW models outperform skip-gram models.
• Increasing dimensionality adds minor improvements.
• Vocabulary threshold tweaking doesn’t affect the results greatly.
At the same time, it increases training time when made less
strict. Thus, tuning the threshold allows to reduce
computational expenses without loss in model quality.
• The more expressive the classifier is, the better normalized
word2vec models are.
15
evaluation: meanvec vs largestcluster
• MeanVec outperforms LargestCluster in all experiments.
• Possibly: largest cluster consists of least meaningful items, e.g.
updates from online game, short texts (e.g. “:)”). Verified with
direct examination.
• Requires thorough research, analysis and experiments.
Variations on clustering may still prove to be useful.
16
evaluation: meanvec vs largestcluster vs allmeanv
• LargestCluster loses in almost all cases to both MeanVec and
AllMeanV .
• AllMeanV is on par with MeanVec with linear models, but not
with GradBoost: MeanVec wins. This can be interpreted as
showing that simple averaging works very well in the semantic
space (not surprising — many semantic relations are known to
become linear in the space of embeddings); even better than
building an ensemble of predictions from individual statuses.
17
summary
• Huge unlabeled text datasets processed. A number of word2vec
models trained.
• User age prediction algorithms suggested and compared with
each other.
• Improvements were not huge when compared to ’mean friends
age’, but the general approach still looks promising.
• Possible developments:
• other word embeddings, possibly including ones that do word
sense disambiguation,
• text embeddings (such as Paragraph Vectors),
• character-level embeddings.
18
thank you!
Thank you for your attention!
19
appendix
Table 2: A comparison of word2vec models on the extended dataset with the
MeanVec algorithm
Word2vec params Train Test Train, nonzero Test, nonzero
type 𝑑 𝑛 𝑣 RMSE MAE RMSE MAE RMSE MAE RMSE MAE
MeanAge
9.672 6.707 9.668 6.711 8.778 5.835 8.800 5.848
LinearRegr
cbow 100 10 20 10.401 7.776 10.397 7.785 8.514 5.792 8.541 5.810
cbow 100 10 30 10.402 7.777 10.396 7.784 8.515 5.791 8.539 5.808
skip 100 10 30 10.818 8.219 10.813 8.232 8.624 5.863 8.645 5.879
skip 200 1 20 10.738 8.146 10.726 8.151 8.593 5.847 8.613 5.859
cbow 200 1 30 10.355 7.737 10.349 7.743 8.497 5.782 8.520 5.798
skip 200 1 30 10.735 8.143 10.724 8.148 8.592 5.846 8.613 5.859
cbow 300 1 30 10.338 7.722 10.329 7.727 8.492 5.779 8.512 5.794
skip 300 1 30 10.689 8.088 10.675 8.096 8.583 5.837 8.601 5.854
ElasticNet
cbow 100 10 20 10.810 8.208 10.799 8.217 8.694 5.903 8.719 5.921
cbow 100 10 30 10.806 8.203 10.795 8.212 8.702 5.909 8.726 5.926
skip 100 10 30 11.239 8.641 11.229 8.653 8.741 5.938 8.766 5.956
skip 200 1 20 11.239 8.641 11.229 8.653 8.741 5.938 8.766 5.956
cbow 200 1 30 10.949 8.349 10.937 8.359 8.736 5.934 8.760 5.951
skip 200 1 30 11.239 8.641 11.229 8.653 8.741 5.938 8.766 5.956
cbow 300 1 30 11.026 8.433 11.017 8.445 8.741 5.938 8.766 5.956
skip 300 1 30 11.239 8.641 11.229 8.653 8.741 5.938 8.766 5.956
GradBoost
cbow 100 10 20 9.089 6.352 9.065 6.344 8.399 5.697 8.394 5.699
cbow 100 10 30 9.093 6.356 9.066 6.345 8.401 5.699 8.395 5.700
skip 100 10 30 9.294 6.527 9.277 6.529 8.495 5.766 8.491 5.770
skip 200 1 20 9.363 6.580 9.342 6.576 8.519 5.785 8.512 5.785
cbow 200 1 30 9.067 6.341 9.043 6.333 8.383 5.682 8.377 5.683
skip 200 1 30 9.365 6.583 9.344 6.580 8.520 5.784 8.512 5.785
cbow 300 1 30 9.048 6.323 9.025 6.316 8.380 5.683 8.371 5.681
skip 300 1 30 9.387 6.596 9.367 6.595 8.536 5.799 8.532 5.804
20
appendix
Table 3: A comparison of word2vec models with their normalized versions on
the extended dataset with the MeanVec algorithm
Word2vec params Train Test Train, nonzero Test, nonzero
type 𝑑 𝑛 𝑣 RMSE MAE RMSE MAE RMSE MAE RMSE MAE
LinearRegr
cbow 100 10 20 10.402 7.777 10.396 7.784 8.515 5.791 8.539 5.808
n.cbow 100 10 20 10.426 7.807 10.418 7.807 8.535 5.807 8.560 5.824
cbow 100 10 30 10.818 8.219 10.813 8.232 8.624 5.863 8.645 5.879
n.cbow 100 10 30 10.839 8.231 10.826 8.240 8.630 5.864 8.649 5.879
skip 100 10 30 10.738 8.146 10.726 8.151 8.593 5.847 8.613 5.859
n.skip 100 10 30 10.753 8.156 10.736 8.159 8.601 5.853 8.620 5.864
skip 200 1 20 10.355 7.737 10.349 7.743 8.497 5.782 8.520 5.798
n.skip 200 1 20 10.363 7.748 10.351 7.746 8.512 5.795 8.532 5.808
cbow 200 1 30 10.735 8.143 10.724 8.148 8.592 5.846 8.613 5.859
n.cbow 200 1 30 10.750 8.152 10.733 8.155 8.601 5.853 8.620 5.864
skip 200 1 30 10.338 7.722 10.329 7.727 8.492 5.779 8.512 5.794
n.skip 200 1 30 10.333 7.720 10.319 7.717 8.501 5.789 8.518 5.800
cbow 300 1 30 10.689 8.088 10.675 8.096 8.583 5.837 8.601 5.854
n.cbow 300 1 30 10.687 8.083 10.668 8.084 8.586 5.840 8.603 5.853
ElasticNet
cbow 100 10 20 10.806 8.203 10.795 8.212 8.702 5.909 8.726 5.926
n.cbow 100 10 20 11.239 8.641 11.229 8.653 8.741 5.938 8.766 5.956
cbow 100 10 30 11.239 8.641 11.229 8.653 8.741 5.938 8.766 5.956
n.cbow 100 10 30 11.239 8.641 11.229 8.653 8.741 5.938 8.766 5.956
skip 100 10 30 11.239 8.641 11.229 8.653 8.741 5.938 8.766 5.956
n.skip 100 10 30 11.239 8.641 11.229 8.653 8.741 5.938 8.766 5.956
skip 200 1 20 10.949 8.349 10.937 8.359 8.736 5.934 8.760 5.951
n.skip 200 1 20 11.239 8.641 11.229 8.653 8.741 5.938 8.766 5.956
cbow 200 1 30 11.239 8.641 11.229 8.653 8.741 5.938 8.766 5.956
n.cbow 200 1 30 11.239 8.641 11.229 8.653 8.741 5.938 8.766 5.956
skip 200 1 30 11.026 8.433 11.017 8.445 8.741 5.938 8.766 5.956
n.skip 200 1 30 11.239 8.641 11.229 8.653 8.741 5.938 8.766 5.956
cbow 300 1 30 11.239 8.641 11.229 8.653 8.741 5.938 8.766 5.956
n.cbow 300 1 30 11.239 8.641 11.229 8.653 8.741 5.938 8.766 5.956
GradBoost
cbow 100 10 20 9.093 6.356 9.066 6.345 8.401 5.699 8.395 5.700
n.cbow 100 10 20 9.043 6.320 9.020 6.313 8.375 5.680 8.367 5.677
cbow 100 10 30 9.294 6.527 9.277 6.529 8.495 5.766 8.491 5.770
n.cbow 100 10 30 9.241 6.487 9.216 6.480 8.481 5.760 8.472 5.758
skip 100 10 30 9.363 6.580 9.342 6.576 8.519 5.785 8.512 5.785
n.skip 100 10 30 9.204 6.458 9.177 6.449 8.456 5.742 8.448 5.742
skip 200 1 20 9.067 6.341 9.043 6.333 8.383 5.682 8.377 5.683
n.skip 200 1 20 8.998 6.290 8.973 6.281 8.350 5.660 8.337 5.658
cbow 200 1 30 9.365 6.583 9.344 6.580 8.520 5.784 8.512 5.785
n.cbow 200 1 30 9.204 6.457 9.176 6.447 8.455 5.743 8.448 5.743
skip 200 1 30 9.048 6.323 9.025 6.316 8.380 5.683 8.371 5.681
n.skip 200 1 30 8.983 6.274 8.965 6.271 8.341 5.656 8.333 5.656
cbow 300 1 30 9.387 6.596 9.367 6.595 8.536 5.799 8.532 5.804
n.cbow 300 1 30 9.172 6.438 9.150 6.430 8.442 5.732 8.430 5.730
21
appendix
Table 4: A comparison of the MeanVec and LargestCluster algorithms on
the extended dataset for various normalized word2vec models
Word2vec params Train Test Train, nonzero Test, nonzero
type 𝑑 𝑛 𝑣 RMSE MAE RMSE MAE RMSE MAE RMSE MAE
LinearRegr
MV cbow 100 10 20 10.401 7.776 10.397 7.785 8.514 5.792 8.541 5.810
LC cbow 100 10 20 10.548 7.926 10.542 7.929 8.573 5.834 8.601 5.852
MV cbow 100 10 30 10.402 7.777 10.396 7.784 8.515 5.791 8.539 5.808
LC cbow 100 10 30 10.553 7.933 10.551 7.939 8.574 5.832 8.603 5.853
MV skip 200 1 20 10.738 8.146 10.726 8.151 8.593 5.847 8.613 5.859
LC skip 200 1 20 10.849 8.251 10.833 8.255 8.629 5.870 8.650 5.883
MV cbow 200 1 30 10.355 7.737 10.349 7.743 8.497 5.782 8.520 5.798
LC cbow 200 1 30 10.493 7.878 10.494 7.886 8.554 5.822 8.581 5.839
ElasticNet
MV cbow 100 10 20 10.810 8.208 10.799 8.217 8.694 5.903 8.719 5.921
LC cbow 100 10 20 11.239 8.641 11.229 8.653 8.741 5.938 8.766 5.956
MV cbow 100 10 30 10.806 8.203 10.795 8.212 8.702 5.909 8.726 5.926
LC cbow 100 10 30 11.239 8.641 11.229 8.653 8.741 5.938 8.766 5.956
MV skip 200 1 20 11.239 8.641 11.229 8.653 8.741 5.938 8.766 5.956
LC skip 200 1 20 11.239 8.641 11.229 8.653 8.741 5.938 8.766 5.956
MV cbow 200 1 30 10.949 8.349 10.937 8.359 8.736 5.934 8.760 5.951
LC cbow 200 1 30 11.239 8.641 11.229 8.653 8.741 5.938 8.766 5.956
GradBoost
MV cbow 100 10 20 9.089 6.352 9.065 6.344 8.399 5.697 8.394 5.699
LC cbow 100 10 20 9.122 6.388 9.100 6.381 8.427 5.720 8.419 5.720
MV cbow 100 10 30 9.093 6.356 9.066 6.345 8.401 5.699 8.395 5.700
LC cbow 100 10 30 9.141 6.399 9.120 6.396 8.431 5.720 8.428 5.723
MV skip 200 1 20 9.363 6.580 9.342 6.576 8.519 5.785 8.512 5.785
LC skip 200 1 20 9.277 6.520 9.250 6.508 8.493 5.772 8.490 5.770
MV cbow 200 1 30 9.067 6.341 9.043 6.333 8.383 5.682 8.377 5.683
LC cbow 200 1 30 9.090 6.361 9.069 6.353 8.406 5.700 8.398 5.702
22
appendix
Table 5: A comparison of the MeanVec, LargestCluster, and AllMeanV
algorithms on the basic dataset for various normalized word2vec models
Word2vec params Train Test
type 𝑑 𝑛 𝑣 RMSE MAE RMSE MAE
LinearRegr
mv cbow 100 10 20 8.778 6.066 8.791 6.087
lc cbow 100 10 20 8.927 6.174 8.946 6.200
amv cbow 100 10 20 8.981 6.096 8.991 6.125
mv cbow 100 10 30 8.798 6.087 8.806 6.095
lc cbow 100 10 30 8.954 6.201 8.955 6.204
amv cbow 100 10 30 8.992 6.108 9.001 6.133
mv skip 100 10 30 9.051 6.256 9.054 6.268
lc skip 100 10 30 9.153 6.332 9.146 6.335
amv skip 100 10 30 9.186 6.249 9.189 6.270
mv skip 200 1 20 8.974 6.216 8.983 6.225
lc skip 200 1 20 9.084 6.292 9.091 6.299
amv skip 200 1 20 9.132 6.210 9.137 6.231
mv cbow 200 1 30 8.734 6.035 8.736 6.052
lc cbow 200 1 30 8.898 6.158 8.902 6.169
amv cbow 200 1 30 8.944 6.071 8.952 6.098
mv skip 200 1 20 8.976 6.217 8.985 6.226
lc skip 200 1 20 9.086 6.296 9.094 6.301
amv skip 200 1 20 9.133 6.211 9.139 6.232
ElasticNet
mv cbow 100 10 20 9.390 6.498 9.397 6.522
lc cbow 100 10 20 9.390 6.498 9.397 6.522
amv cbow 100 10 20 9.398 6.428 9.398 6.451
mv cbow 100 10 30 9.390 6.498 9.397 6.522
lc cbow 100 10 30 9.390 6.498 9.397 6.522
amv cbow 100 10 30 9.399 6.428 9.398 6.451
mv skip 100 10 30 9.390 6.498 9.397 6.522
lc skip 100 10 30 9.390 6.498 9.397 6.522
amv skip 100 10 30 9.399 6.428 9.398 6.451
mv skip 200 1 20 9.390 6.498 9.397 6.522
lc skip 200 1 20 9.390 6.498 9.397 6.522
amv skip 200 1 20 9.398 6.428 9.398 6.451
mv cbow 200 1 30 9.390 6.498 9.397 6.522
lc cbow 200 1 30 9.390 6.498 9.397 6.522
amv cbow 200 1 30 9.399 6.428 9.398 6.451
mv skip 200 1 20 9.390 6.498 9.397 6.522
lc skip 200 1 20 9.390 6.498 9.397 6.522
amv skip 200 1 20 9.399 6.428 9.398 6.451
GradBoost
mv cbow 100 10 20 8.075 5.398 8.068 5.399
lc cbow 100 10 20 8.163 5.456 8.172 5.467
amv cbow 100 10 20 8.228 5.447 8.275 5.486
mv cbow 100 10 30 8.086 5.404 8.077 5.403
lc cbow 100 10 30 8.194 5.479 8.185 5.476
amv cbow 100 10 30 8.240 5.456 8.286 5.493
mv skip 100 10 30 8.277 5.534 8.271 5.537
lc skip 100 10 30 8.333 5.572 8.336 5.577
amv skip 100 10 30 8.362 5.548 8.408 5.583
mv skip 200 1 20 8.242 5.526 8.231 5.513
lc skip 200 1 20 8.308 5.562 8.307 5.559
amv skip 200 1 20 8.347 5.541 8.392 5.572
mv cbow 200 1 30 8.032 5.368 8.026 5.366
lc cbow 200 1 30 8.142 5.447 8.144 5.446
amv cbow 200 1 30 8.220 5.443 8.264 5.479
mv skip 200 1 20 8.240 5.518 8.230 5.510
lc skip 200 1 20 8.308 5.560 8.302 5.558
amv skip 200 1 20 8.347 5.542 8.392 5.573
23

Mais conteúdo relacionado

Mais procurados

Enriching Word Vectors with Subword Information
Enriching Word Vectors with Subword InformationEnriching Word Vectors with Subword Information
Enriching Word Vectors with Subword InformationSeonghyun Kim
 
Crash Course in Natural Language Processing (2016)
Crash Course in Natural Language Processing (2016)Crash Course in Natural Language Processing (2016)
Crash Course in Natural Language Processing (2016)Vsevolod Dyomkin
 
DeepWalk: Online Learning of Representations
DeepWalk: Online Learning of RepresentationsDeepWalk: Online Learning of Representations
DeepWalk: Online Learning of RepresentationsBryan Perozzi
 
Exploring Session Context using Distributed Representations of Queries and Re...
Exploring Session Context using Distributed Representations of Queries and Re...Exploring Session Context using Distributed Representations of Queries and Re...
Exploring Session Context using Distributed Representations of Queries and Re...Bhaskar Mitra
 
Crash-course in Natural Language Processing
Crash-course in Natural Language ProcessingCrash-course in Natural Language Processing
Crash-course in Natural Language ProcessingVsevolod Dyomkin
 
Vectorland: Brief Notes from Using Text Embeddings for Search
Vectorland: Brief Notes from Using Text Embeddings for SearchVectorland: Brief Notes from Using Text Embeddings for Search
Vectorland: Brief Notes from Using Text Embeddings for SearchBhaskar Mitra
 
Neural Models for Document Ranking
Neural Models for Document RankingNeural Models for Document Ranking
Neural Models for Document RankingBhaskar Mitra
 
Dual Embedding Space Model (DESM)
Dual Embedding Space Model (DESM)Dual Embedding Space Model (DESM)
Dual Embedding Space Model (DESM)Bhaskar Mitra
 
Language Models for Information Retrieval
Language Models for Information RetrievalLanguage Models for Information Retrieval
Language Models for Information RetrievalNik Spirin
 
Dstc6 an introduction
Dstc6 an introductionDstc6 an introduction
Dstc6 an introductionhkh
 
5 Lessons Learned from Designing Neural Models for Information Retrieval
5 Lessons Learned from Designing Neural Models for Information Retrieval5 Lessons Learned from Designing Neural Models for Information Retrieval
5 Lessons Learned from Designing Neural Models for Information RetrievalBhaskar Mitra
 
AINL 2016: Bastrakova, Ledesma, Millan, Zighed
AINL 2016: Bastrakova, Ledesma, Millan, ZighedAINL 2016: Bastrakova, Ledesma, Millan, Zighed
AINL 2016: Bastrakova, Ledesma, Millan, ZighedLidia Pivovarova
 
Natural Language Processing in Practice
Natural Language Processing in PracticeNatural Language Processing in Practice
Natural Language Processing in PracticeVsevolod Dyomkin
 
Can functional programming be liberated from static typing?
Can functional programming be liberated from static typing?Can functional programming be liberated from static typing?
Can functional programming be liberated from static typing?Vsevolod Dyomkin
 
TextRank: Bringing Order into Texts
TextRank: Bringing Order into TextsTextRank: Bringing Order into Texts
TextRank: Bringing Order into TextsShubhangi Tandon
 
Frontiers of Natural Language Processing
Frontiers of Natural Language ProcessingFrontiers of Natural Language Processing
Frontiers of Natural Language ProcessingSebastian Ruder
 
Duet @ TREC 2019 Deep Learning Track
Duet @ TREC 2019 Deep Learning TrackDuet @ TREC 2019 Deep Learning Track
Duet @ TREC 2019 Deep Learning TrackBhaskar Mitra
 

Mais procurados (20)

Enriching Word Vectors with Subword Information
Enriching Word Vectors with Subword InformationEnriching Word Vectors with Subword Information
Enriching Word Vectors with Subword Information
 
Crash Course in Natural Language Processing (2016)
Crash Course in Natural Language Processing (2016)Crash Course in Natural Language Processing (2016)
Crash Course in Natural Language Processing (2016)
 
DeepWalk: Online Learning of Representations
DeepWalk: Online Learning of RepresentationsDeepWalk: Online Learning of Representations
DeepWalk: Online Learning of Representations
 
The Duet model
The Duet modelThe Duet model
The Duet model
 
Exploring Session Context using Distributed Representations of Queries and Re...
Exploring Session Context using Distributed Representations of Queries and Re...Exploring Session Context using Distributed Representations of Queries and Re...
Exploring Session Context using Distributed Representations of Queries and Re...
 
Crash-course in Natural Language Processing
Crash-course in Natural Language ProcessingCrash-course in Natural Language Processing
Crash-course in Natural Language Processing
 
NLP Project Full Cycle
NLP Project Full CycleNLP Project Full Cycle
NLP Project Full Cycle
 
Vectorland: Brief Notes from Using Text Embeddings for Search
Vectorland: Brief Notes from Using Text Embeddings for SearchVectorland: Brief Notes from Using Text Embeddings for Search
Vectorland: Brief Notes from Using Text Embeddings for Search
 
Aspects of NLP Practice
Aspects of NLP PracticeAspects of NLP Practice
Aspects of NLP Practice
 
Neural Models for Document Ranking
Neural Models for Document RankingNeural Models for Document Ranking
Neural Models for Document Ranking
 
Dual Embedding Space Model (DESM)
Dual Embedding Space Model (DESM)Dual Embedding Space Model (DESM)
Dual Embedding Space Model (DESM)
 
Language Models for Information Retrieval
Language Models for Information RetrievalLanguage Models for Information Retrieval
Language Models for Information Retrieval
 
Dstc6 an introduction
Dstc6 an introductionDstc6 an introduction
Dstc6 an introduction
 
5 Lessons Learned from Designing Neural Models for Information Retrieval
5 Lessons Learned from Designing Neural Models for Information Retrieval5 Lessons Learned from Designing Neural Models for Information Retrieval
5 Lessons Learned from Designing Neural Models for Information Retrieval
 
AINL 2016: Bastrakova, Ledesma, Millan, Zighed
AINL 2016: Bastrakova, Ledesma, Millan, ZighedAINL 2016: Bastrakova, Ledesma, Millan, Zighed
AINL 2016: Bastrakova, Ledesma, Millan, Zighed
 
Natural Language Processing in Practice
Natural Language Processing in PracticeNatural Language Processing in Practice
Natural Language Processing in Practice
 
Can functional programming be liberated from static typing?
Can functional programming be liberated from static typing?Can functional programming be liberated from static typing?
Can functional programming be liberated from static typing?
 
TextRank: Bringing Order into Texts
TextRank: Bringing Order into TextsTextRank: Bringing Order into Texts
TextRank: Bringing Order into Texts
 
Frontiers of Natural Language Processing
Frontiers of Natural Language ProcessingFrontiers of Natural Language Processing
Frontiers of Natural Language Processing
 
Duet @ TREC 2019 Deep Learning Track
Duet @ TREC 2019 Deep Learning TrackDuet @ TREC 2019 Deep Learning Track
Duet @ TREC 2019 Deep Learning Track
 

Destaque

Destaque (20)

AINL 2016: Bodrunova, Blekanov, Maksimov
AINL 2016: Bodrunova, Blekanov, MaksimovAINL 2016: Bodrunova, Blekanov, Maksimov
AINL 2016: Bodrunova, Blekanov, Maksimov
 
AINL 2016: Bugaychenko
AINL 2016: BugaychenkoAINL 2016: Bugaychenko
AINL 2016: Bugaychenko
 
AINL 2016: Rykov, Nagornyy, Koltsova, Natta, Kremenets, Manovich, Cerrone, Cr...
AINL 2016: Rykov, Nagornyy, Koltsova, Natta, Kremenets, Manovich, Cerrone, Cr...AINL 2016: Rykov, Nagornyy, Koltsova, Natta, Kremenets, Manovich, Cerrone, Cr...
AINL 2016: Rykov, Nagornyy, Koltsova, Natta, Kremenets, Manovich, Cerrone, Cr...
 
AINL 2016: Skornyakov
AINL 2016: SkornyakovAINL 2016: Skornyakov
AINL 2016: Skornyakov
 
AINL 2016: Romanova, Nefedov
AINL 2016: Romanova, NefedovAINL 2016: Romanova, Nefedov
AINL 2016: Romanova, Nefedov
 
AINL 2016: Yagunova
AINL 2016: YagunovaAINL 2016: Yagunova
AINL 2016: Yagunova
 
AINL 2016: Ustalov
AINL 2016: Ustalov AINL 2016: Ustalov
AINL 2016: Ustalov
 
AINL 2016: Kozerenko
AINL 2016: Kozerenko AINL 2016: Kozerenko
AINL 2016: Kozerenko
 
AINL 2016: Eyecioglu
AINL 2016: EyeciogluAINL 2016: Eyecioglu
AINL 2016: Eyecioglu
 
AINL 2016: Muravyov
AINL 2016: MuravyovAINL 2016: Muravyov
AINL 2016: Muravyov
 
AINL 2016: Proncheva
AINL 2016: PronchevaAINL 2016: Proncheva
AINL 2016: Proncheva
 
AINL 2016: Goncharov
AINL 2016: GoncharovAINL 2016: Goncharov
AINL 2016: Goncharov
 
AINL 2016: Panicheva, Ledovaya
AINL 2016: Panicheva, LedovayaAINL 2016: Panicheva, Ledovaya
AINL 2016: Panicheva, Ledovaya
 
AINL 2016: Fenogenova, Karpov, Kazorin
AINL 2016: Fenogenova, Karpov, KazorinAINL 2016: Fenogenova, Karpov, Kazorin
AINL 2016: Fenogenova, Karpov, Kazorin
 
AINL 2016: Boldyreva
AINL 2016: BoldyrevaAINL 2016: Boldyreva
AINL 2016: Boldyreva
 
AINL 2016: Moskvichev
AINL 2016: MoskvichevAINL 2016: Moskvichev
AINL 2016: Moskvichev
 
AINL 2016: Strijov
AINL 2016: StrijovAINL 2016: Strijov
AINL 2016: Strijov
 
AINL 2016: Khudobakhshov
AINL 2016: KhudobakhshovAINL 2016: Khudobakhshov
AINL 2016: Khudobakhshov
 
AINL 2016: Malykh
AINL 2016: MalykhAINL 2016: Malykh
AINL 2016: Malykh
 
AINL 2016: Just AI
AINL 2016: Just AIAINL 2016: Just AI
AINL 2016: Just AI
 

Semelhante a AINL 2016: Alekseev, Nikolenko

IEEE Datamining 2016 Title and Abstract
IEEE  Datamining 2016 Title and AbstractIEEE  Datamining 2016 Title and Abstract
IEEE Datamining 2016 Title and Abstracttsysglobalsolutions
 
Big Linked Data Federation - ExtremeEarth Open Workshop
Big Linked Data Federation - ExtremeEarth Open WorkshopBig Linked Data Federation - ExtremeEarth Open Workshop
Big Linked Data Federation - ExtremeEarth Open WorkshopExtremeEarth
 
Image processing-ieee-2014-projects
Image processing-ieee-2014-projectsImage processing-ieee-2014-projects
Image processing-ieee-2014-projectsVijay Karan
 
Image Processing IEEE 2014 Projects
Image Processing IEEE 2014 ProjectsImage Processing IEEE 2014 Projects
Image Processing IEEE 2014 ProjectsVijay Karan
 
Practice of Streaming Processing of Dynamic Graphs: Concepts, Models, and Sys...
Practice of Streaming Processing of Dynamic Graphs: Concepts, Models, and Sys...Practice of Streaming Processing of Dynamic Graphs: Concepts, Models, and Sys...
Practice of Streaming Processing of Dynamic Graphs: Concepts, Models, and Sys...Subhajit Sahu
 
Mining User Lifecycles from Online Community Platforms and their Application ...
Mining User Lifecycles from Online Community Platforms and their Application ...Mining User Lifecycles from Online Community Platforms and their Application ...
Mining User Lifecycles from Online Community Platforms and their Application ...Matthew Rowe
 
00 Automatic Mental Health Classification in Online Settings and Language Emb...
00 Automatic Mental Health Classification in Online Settings and Language Emb...00 Automatic Mental Health Classification in Online Settings and Language Emb...
00 Automatic Mental Health Classification in Online Settings and Language Emb...Duke Network Analysis Center
 
Linked science presentation 25
Linked science presentation 25Linked science presentation 25
Linked science presentation 25Francesco Osborne
 
Low Cost Business Intelligence Platform for MongoDB instances using MEAN stack
Low Cost Business Intelligence Platform for MongoDB instances using MEAN stackLow Cost Business Intelligence Platform for MongoDB instances using MEAN stack
Low Cost Business Intelligence Platform for MongoDB instances using MEAN stackAvinash Kaza
 
IEEE 2014 Matlab Projects
IEEE 2014 Matlab ProjectsIEEE 2014 Matlab Projects
IEEE 2014 Matlab ProjectsVijay Karan
 
IEEE 2014 Matlab Projects
IEEE 2014 Matlab ProjectsIEEE 2014 Matlab Projects
IEEE 2014 Matlab ProjectsVijay Karan
 
Sergey Nikolenko and Anton Alekseev User Profiling in Text-Based Recommende...
Sergey Nikolenko and  Anton Alekseev  User Profiling in Text-Based Recommende...Sergey Nikolenko and  Anton Alekseev  User Profiling in Text-Based Recommende...
Sergey Nikolenko and Anton Alekseev User Profiling in Text-Based Recommende...AIST
 
High-Performance Analysis of Streaming Graphs
High-Performance Analysis of Streaming Graphs High-Performance Analysis of Streaming Graphs
High-Performance Analysis of Streaming Graphs Jason Riedy
 
Automated Essay Scoring Using Efficient Transformer-Based Language Models
Automated Essay Scoring Using Efficient Transformer-Based Language ModelsAutomated Essay Scoring Using Efficient Transformer-Based Language Models
Automated Essay Scoring Using Efficient Transformer-Based Language ModelsNat Rice
 
Scalable constrained spectral clustering
Scalable constrained spectral clusteringScalable constrained spectral clustering
Scalable constrained spectral clusteringNishanth Harapanahalli
 
SemanticSVD++: Incorporating Semantic Taste Evolution for Predicting Ratings
SemanticSVD++: Incorporating Semantic Taste Evolution for Predicting RatingsSemanticSVD++: Incorporating Semantic Taste Evolution for Predicting Ratings
SemanticSVD++: Incorporating Semantic Taste Evolution for Predicting Ratings Matthew Rowe
 
OREChem Services and Workflows
OREChem Services and WorkflowsOREChem Services and Workflows
OREChem Services and Workflowsmarpierc
 
RESUME SCREENING USING LSTM
RESUME SCREENING USING LSTMRESUME SCREENING USING LSTM
RESUME SCREENING USING LSTMIRJET Journal
 

Semelhante a AINL 2016: Alekseev, Nikolenko (20)

IEEE Datamining 2016 Title and Abstract
IEEE  Datamining 2016 Title and AbstractIEEE  Datamining 2016 Title and Abstract
IEEE Datamining 2016 Title and Abstract
 
Big Linked Data Federation - ExtremeEarth Open Workshop
Big Linked Data Federation - ExtremeEarth Open WorkshopBig Linked Data Federation - ExtremeEarth Open Workshop
Big Linked Data Federation - ExtremeEarth Open Workshop
 
Image processing-ieee-2014-projects
Image processing-ieee-2014-projectsImage processing-ieee-2014-projects
Image processing-ieee-2014-projects
 
Image Processing IEEE 2014 Projects
Image Processing IEEE 2014 ProjectsImage Processing IEEE 2014 Projects
Image Processing IEEE 2014 Projects
 
Practice of Streaming Processing of Dynamic Graphs: Concepts, Models, and Sys...
Practice of Streaming Processing of Dynamic Graphs: Concepts, Models, and Sys...Practice of Streaming Processing of Dynamic Graphs: Concepts, Models, and Sys...
Practice of Streaming Processing of Dynamic Graphs: Concepts, Models, and Sys...
 
Mining User Lifecycles from Online Community Platforms and their Application ...
Mining User Lifecycles from Online Community Platforms and their Application ...Mining User Lifecycles from Online Community Platforms and their Application ...
Mining User Lifecycles from Online Community Platforms and their Application ...
 
00 Automatic Mental Health Classification in Online Settings and Language Emb...
00 Automatic Mental Health Classification in Online Settings and Language Emb...00 Automatic Mental Health Classification in Online Settings and Language Emb...
00 Automatic Mental Health Classification in Online Settings and Language Emb...
 
Poster (1)
Poster (1)Poster (1)
Poster (1)
 
Linked science presentation 25
Linked science presentation 25Linked science presentation 25
Linked science presentation 25
 
Low Cost Business Intelligence Platform for MongoDB instances using MEAN stack
Low Cost Business Intelligence Platform for MongoDB instances using MEAN stackLow Cost Business Intelligence Platform for MongoDB instances using MEAN stack
Low Cost Business Intelligence Platform for MongoDB instances using MEAN stack
 
IEEE 2014 Matlab Projects
IEEE 2014 Matlab ProjectsIEEE 2014 Matlab Projects
IEEE 2014 Matlab Projects
 
IEEE 2014 Matlab Projects
IEEE 2014 Matlab ProjectsIEEE 2014 Matlab Projects
IEEE 2014 Matlab Projects
 
Declarative data analysis
Declarative data analysisDeclarative data analysis
Declarative data analysis
 
Sergey Nikolenko and Anton Alekseev User Profiling in Text-Based Recommende...
Sergey Nikolenko and  Anton Alekseev  User Profiling in Text-Based Recommende...Sergey Nikolenko and  Anton Alekseev  User Profiling in Text-Based Recommende...
Sergey Nikolenko and Anton Alekseev User Profiling in Text-Based Recommende...
 
High-Performance Analysis of Streaming Graphs
High-Performance Analysis of Streaming Graphs High-Performance Analysis of Streaming Graphs
High-Performance Analysis of Streaming Graphs
 
Automated Essay Scoring Using Efficient Transformer-Based Language Models
Automated Essay Scoring Using Efficient Transformer-Based Language ModelsAutomated Essay Scoring Using Efficient Transformer-Based Language Models
Automated Essay Scoring Using Efficient Transformer-Based Language Models
 
Scalable constrained spectral clustering
Scalable constrained spectral clusteringScalable constrained spectral clustering
Scalable constrained spectral clustering
 
SemanticSVD++: Incorporating Semantic Taste Evolution for Predicting Ratings
SemanticSVD++: Incorporating Semantic Taste Evolution for Predicting RatingsSemanticSVD++: Incorporating Semantic Taste Evolution for Predicting Ratings
SemanticSVD++: Incorporating Semantic Taste Evolution for Predicting Ratings
 
OREChem Services and Workflows
OREChem Services and WorkflowsOREChem Services and Workflows
OREChem Services and Workflows
 
RESUME SCREENING USING LSTM
RESUME SCREENING USING LSTMRESUME SCREENING USING LSTM
RESUME SCREENING USING LSTM
 

Mais de Lidia Pivovarova

Classification and clustering in media monitoring: from knowledge engineering...
Classification and clustering in media monitoring: from knowledge engineering...Classification and clustering in media monitoring: from knowledge engineering...
Classification and clustering in media monitoring: from knowledge engineering...Lidia Pivovarova
 
Convolutional neural networks for text classification
Convolutional neural networks for text classificationConvolutional neural networks for text classification
Convolutional neural networks for text classificationLidia Pivovarova
 
Grouping business news stories based on salience of named entities
Grouping business news stories based on salience of named entitiesGrouping business news stories based on salience of named entities
Grouping business news stories based on salience of named entitiesLidia Pivovarova
 
Интеллектуальный анализ текста
Интеллектуальный анализ текстаИнтеллектуальный анализ текста
Интеллектуальный анализ текстаLidia Pivovarova
 
AINL 2016: Shavrina, Selegey
AINL 2016: Shavrina, SelegeyAINL 2016: Shavrina, Selegey
AINL 2016: Shavrina, SelegeyLidia Pivovarova
 

Mais de Lidia Pivovarova (7)

Classification and clustering in media monitoring: from knowledge engineering...
Classification and clustering in media monitoring: from knowledge engineering...Classification and clustering in media monitoring: from knowledge engineering...
Classification and clustering in media monitoring: from knowledge engineering...
 
Convolutional neural networks for text classification
Convolutional neural networks for text classificationConvolutional neural networks for text classification
Convolutional neural networks for text classification
 
Grouping business news stories based on salience of named entities
Grouping business news stories based on salience of named entitiesGrouping business news stories based on salience of named entities
Grouping business news stories based on salience of named entities
 
Интеллектуальный анализ текста
Интеллектуальный анализ текстаИнтеллектуальный анализ текста
Интеллектуальный анализ текста
 
AINL 2016: Shavrina, Selegey
AINL 2016: Shavrina, SelegeyAINL 2016: Shavrina, Selegey
AINL 2016: Shavrina, Selegey
 
AINL 2016:
AINL 2016: AINL 2016:
AINL 2016:
 
AINL 2016: Grigorieva
AINL 2016: GrigorievaAINL 2016: Grigorieva
AINL 2016: Grigorieva
 

Último

GBSN - Microbiology (Unit 2)
GBSN - Microbiology (Unit 2)GBSN - Microbiology (Unit 2)
GBSN - Microbiology (Unit 2)Areesha Ahmad
 
Recombination DNA Technology (Nucleic Acid Hybridization )
Recombination DNA Technology (Nucleic Acid Hybridization )Recombination DNA Technology (Nucleic Acid Hybridization )
Recombination DNA Technology (Nucleic Acid Hybridization )aarthirajkumar25
 
Nanoparticles synthesis and characterization​ ​
Nanoparticles synthesis and characterization​  ​Nanoparticles synthesis and characterization​  ​
Nanoparticles synthesis and characterization​ ​kaibalyasahoo82800
 
Nightside clouds and disequilibrium chemistry on the hot Jupiter WASP-43b
Nightside clouds and disequilibrium chemistry on the hot Jupiter WASP-43bNightside clouds and disequilibrium chemistry on the hot Jupiter WASP-43b
Nightside clouds and disequilibrium chemistry on the hot Jupiter WASP-43bSérgio Sacani
 
Pests of cotton_Sucking_Pests_Dr.UPR.pdf
Pests of cotton_Sucking_Pests_Dr.UPR.pdfPests of cotton_Sucking_Pests_Dr.UPR.pdf
Pests of cotton_Sucking_Pests_Dr.UPR.pdfPirithiRaju
 
Lucknow 💋 Russian Call Girls Lucknow Finest Escorts Service 8923113531 Availa...
Lucknow 💋 Russian Call Girls Lucknow Finest Escorts Service 8923113531 Availa...Lucknow 💋 Russian Call Girls Lucknow Finest Escorts Service 8923113531 Availa...
Lucknow 💋 Russian Call Girls Lucknow Finest Escorts Service 8923113531 Availa...anilsa9823
 
Presentation Vikram Lander by Vedansh Gupta.pptx
Presentation Vikram Lander by Vedansh Gupta.pptxPresentation Vikram Lander by Vedansh Gupta.pptx
Presentation Vikram Lander by Vedansh Gupta.pptxgindu3009
 
❤Jammu Kashmir Call Girls 8617697112 Personal Whatsapp Number 💦✅.
❤Jammu Kashmir Call Girls 8617697112 Personal Whatsapp Number 💦✅.❤Jammu Kashmir Call Girls 8617697112 Personal Whatsapp Number 💦✅.
❤Jammu Kashmir Call Girls 8617697112 Personal Whatsapp Number 💦✅.Nitya salvi
 
Recombinant DNA technology (Immunological screening)
Recombinant DNA technology (Immunological screening)Recombinant DNA technology (Immunological screening)
Recombinant DNA technology (Immunological screening)PraveenaKalaiselvan1
 
Chemistry 4th semester series (krishna).pdf
Chemistry 4th semester series (krishna).pdfChemistry 4th semester series (krishna).pdf
Chemistry 4th semester series (krishna).pdfSumit Kumar yadav
 
Asymmetry in the atmosphere of the ultra-hot Jupiter WASP-76 b
Asymmetry in the atmosphere of the ultra-hot Jupiter WASP-76 bAsymmetry in the atmosphere of the ultra-hot Jupiter WASP-76 b
Asymmetry in the atmosphere of the ultra-hot Jupiter WASP-76 bSérgio Sacani
 
GBSN - Microbiology (Unit 1)
GBSN - Microbiology (Unit 1)GBSN - Microbiology (Unit 1)
GBSN - Microbiology (Unit 1)Areesha Ahmad
 
fundamental of entomology all in one topics of entomology
fundamental of entomology all in one topics of entomologyfundamental of entomology all in one topics of entomology
fundamental of entomology all in one topics of entomologyDrAnita Sharma
 
Pests of mustard_Identification_Management_Dr.UPR.pdf
Pests of mustard_Identification_Management_Dr.UPR.pdfPests of mustard_Identification_Management_Dr.UPR.pdf
Pests of mustard_Identification_Management_Dr.UPR.pdfPirithiRaju
 
Botany 4th semester file By Sumit Kumar yadav.pdf
Botany 4th semester file By Sumit Kumar yadav.pdfBotany 4th semester file By Sumit Kumar yadav.pdf
Botany 4th semester file By Sumit Kumar yadav.pdfSumit Kumar yadav
 
Formation of low mass protostars and their circumstellar disks
Formation of low mass protostars and their circumstellar disksFormation of low mass protostars and their circumstellar disks
Formation of low mass protostars and their circumstellar disksSérgio Sacani
 
Pulmonary drug delivery system M.pharm -2nd sem P'ceutics
Pulmonary drug delivery system M.pharm -2nd sem P'ceuticsPulmonary drug delivery system M.pharm -2nd sem P'ceutics
Pulmonary drug delivery system M.pharm -2nd sem P'ceuticssakshisoni2385
 
Isotopic evidence of long-lived volcanism on Io
Isotopic evidence of long-lived volcanism on IoIsotopic evidence of long-lived volcanism on Io
Isotopic evidence of long-lived volcanism on IoSérgio Sacani
 
Zoology 4th semester series (krishna).pdf
Zoology 4th semester series (krishna).pdfZoology 4th semester series (krishna).pdf
Zoology 4th semester series (krishna).pdfSumit Kumar yadav
 
Physiochemical properties of nanomaterials and its nanotoxicity.pptx
Physiochemical properties of nanomaterials and its nanotoxicity.pptxPhysiochemical properties of nanomaterials and its nanotoxicity.pptx
Physiochemical properties of nanomaterials and its nanotoxicity.pptxAArockiyaNisha
 

Último (20)

GBSN - Microbiology (Unit 2)
GBSN - Microbiology (Unit 2)GBSN - Microbiology (Unit 2)
GBSN - Microbiology (Unit 2)
 
Recombination DNA Technology (Nucleic Acid Hybridization )
Recombination DNA Technology (Nucleic Acid Hybridization )Recombination DNA Technology (Nucleic Acid Hybridization )
Recombination DNA Technology (Nucleic Acid Hybridization )
 
Nanoparticles synthesis and characterization​ ​
Nanoparticles synthesis and characterization​  ​Nanoparticles synthesis and characterization​  ​
Nanoparticles synthesis and characterization​ ​
 
Nightside clouds and disequilibrium chemistry on the hot Jupiter WASP-43b
Nightside clouds and disequilibrium chemistry on the hot Jupiter WASP-43bNightside clouds and disequilibrium chemistry on the hot Jupiter WASP-43b
Nightside clouds and disequilibrium chemistry on the hot Jupiter WASP-43b
 
Pests of cotton_Sucking_Pests_Dr.UPR.pdf
Pests of cotton_Sucking_Pests_Dr.UPR.pdfPests of cotton_Sucking_Pests_Dr.UPR.pdf
Pests of cotton_Sucking_Pests_Dr.UPR.pdf
 
Lucknow 💋 Russian Call Girls Lucknow Finest Escorts Service 8923113531 Availa...
Lucknow 💋 Russian Call Girls Lucknow Finest Escorts Service 8923113531 Availa...Lucknow 💋 Russian Call Girls Lucknow Finest Escorts Service 8923113531 Availa...
Lucknow 💋 Russian Call Girls Lucknow Finest Escorts Service 8923113531 Availa...
 
Presentation Vikram Lander by Vedansh Gupta.pptx
Presentation Vikram Lander by Vedansh Gupta.pptxPresentation Vikram Lander by Vedansh Gupta.pptx
Presentation Vikram Lander by Vedansh Gupta.pptx
 
❤Jammu Kashmir Call Girls 8617697112 Personal Whatsapp Number 💦✅.
❤Jammu Kashmir Call Girls 8617697112 Personal Whatsapp Number 💦✅.❤Jammu Kashmir Call Girls 8617697112 Personal Whatsapp Number 💦✅.
❤Jammu Kashmir Call Girls 8617697112 Personal Whatsapp Number 💦✅.
 
Recombinant DNA technology (Immunological screening)
Recombinant DNA technology (Immunological screening)Recombinant DNA technology (Immunological screening)
Recombinant DNA technology (Immunological screening)
 
Chemistry 4th semester series (krishna).pdf
Chemistry 4th semester series (krishna).pdfChemistry 4th semester series (krishna).pdf
Chemistry 4th semester series (krishna).pdf
 
Asymmetry in the atmosphere of the ultra-hot Jupiter WASP-76 b
Asymmetry in the atmosphere of the ultra-hot Jupiter WASP-76 bAsymmetry in the atmosphere of the ultra-hot Jupiter WASP-76 b
Asymmetry in the atmosphere of the ultra-hot Jupiter WASP-76 b
 
GBSN - Microbiology (Unit 1)
GBSN - Microbiology (Unit 1)GBSN - Microbiology (Unit 1)
GBSN - Microbiology (Unit 1)
 
fundamental of entomology all in one topics of entomology
fundamental of entomology all in one topics of entomologyfundamental of entomology all in one topics of entomology
fundamental of entomology all in one topics of entomology
 
Pests of mustard_Identification_Management_Dr.UPR.pdf
Pests of mustard_Identification_Management_Dr.UPR.pdfPests of mustard_Identification_Management_Dr.UPR.pdf
Pests of mustard_Identification_Management_Dr.UPR.pdf
 
Botany 4th semester file By Sumit Kumar yadav.pdf
Botany 4th semester file By Sumit Kumar yadav.pdfBotany 4th semester file By Sumit Kumar yadav.pdf
Botany 4th semester file By Sumit Kumar yadav.pdf
 
Formation of low mass protostars and their circumstellar disks
Formation of low mass protostars and their circumstellar disksFormation of low mass protostars and their circumstellar disks
Formation of low mass protostars and their circumstellar disks
 
Pulmonary drug delivery system M.pharm -2nd sem P'ceutics
Pulmonary drug delivery system M.pharm -2nd sem P'ceuticsPulmonary drug delivery system M.pharm -2nd sem P'ceutics
Pulmonary drug delivery system M.pharm -2nd sem P'ceutics
 
Isotopic evidence of long-lived volcanism on Io
Isotopic evidence of long-lived volcanism on IoIsotopic evidence of long-lived volcanism on Io
Isotopic evidence of long-lived volcanism on Io
 
Zoology 4th semester series (krishna).pdf
Zoology 4th semester series (krishna).pdfZoology 4th semester series (krishna).pdf
Zoology 4th semester series (krishna).pdf
 
Physiochemical properties of nanomaterials and its nanotoxicity.pptx
Physiochemical properties of nanomaterials and its nanotoxicity.pptxPhysiochemical properties of nanomaterials and its nanotoxicity.pptx
Physiochemical properties of nanomaterials and its nanotoxicity.pptx
 

AINL 2016: Alekseev, Nikolenko

  • 1. predicting the age of social network users from user-generated texts with word embeddings Anton Alekseev1 and Sergey Nikolenko1,2,3 AINL 2016 St. Petersburg, November 11, 2016 1 Steklov Institute of Mathematics at St. Petersburg 2 Kazan Federal University, Kazan, Russia 3 Laboratory for Internet Studies, NRU Higher School of Economics, St. Petersburg
  • 2. deep learning for nlp • Recent success on numerous tasks: from text classification to syntax parsing. • Various applications of word vectors pretrained on huge corpora: word2vec, GloVe, etc. • In this work we have tried a few approaches for social network users profiling using full-text items. 2
  • 3. dataset • 868,126 users of the Odnoklassniki social network. In particular, it contains the following data: (1) demographic information on 868,126 users of the network: gender, age, and region; (2) “friendship” relation, several different type of links: “friend”, “love”, “spouse”, “parent”, and so on; (3) history of logins for individual users; (4) data on the “likes” (“class” marks) a user has given to other users’ statuses and posts in various groups; (5) texts of the posts for individual users; (6) communities (“groups”) statuses that have been liked by these selected users. 3
  • 4. dataset • The dataset has been created as follows: (1) 486 seed users; (2) for these users, their sets of friends have been extracted; (3) then the friends of these friends; as a result — seed users neighborhood of depth 2 in the social graph. 4
  • 5. dataset: age • Mean age — 31.39 years; 0 20 40 60 80 100 120 Age 100 101 102 103 104 105 Number of users Figure 1: Users’ age in the Odnoklassniki dataset • Note users with implausible ages (ages 2 and 3, age higher than 100 years). • We have removed all users with ages below 10 and above 80. 5
  • 6. dataset: friends 100 101 102 103 104 Number of friends 100 101 102 103 104 Number of users Figure 2: Number of friends distribution in the Odnoklassniki dataset • “Arc” (not a straight line!) on a log-log plot — probably due to the collection method: smaller number of degree-0, degree-1, etc. vertices (that is, users) than in the whole “social graph”. 6
  • 7. dataset and setup for word embeddings A large Russian-language corpus with about 14G tokens in 2.5M documents. This corpus includes: • Russian Wikipedia: 1.15M documents, 238M tokens; • automated Web crawl data: 890K documents, 568M tokens; • (main part) the huge lib.rus.ec library corpus: 234K documents, 12.9G tokens; • user statuses and group posts from the Odnoklassniki social network, as described above. For more info, please see the article cited in the paper: Arefyev N., Panchenko A., Lukanin A., Lesota O., Romanov P. Evaluating Three Corpus-Based Semantic Similarity Systems for Russian. In Proceedings of International Conference on Computational Linguistics Dialogue 2015, 2015 7
  • 8. dataset and setup for word embeddings • 𝑤𝑜𝑟𝑑2𝑣𝑒𝑐 configurations we took into consideration Table 1: Word2vec models trained on large Russian corpora. The columns are: dimension 𝑑, window size 𝑤, number of negative samples 𝑛, vocabulary threshold 𝑣, and the resulting model size Type 𝑑 𝑤 𝑛 𝑣 Size CBOW 100 11 10 20 1.3G CBOW 100 11 10 30 0.97G skip-gram 100 11 10 30 0.97G skip-gram 200 11 1 20 1.3G CBOW 200 11 1 30 2.0G skip-gram 200 11 1 30 2.0G CBOW 300 11 1 30 2.9G skip-gram 300 11 1 30 2.9G 8
  • 9. chosen base approaches • How well is user’s age approximated by mean friends age? (1) MeanAge: predict age with the mean of friends’ ages; (2) LinearRegr: linear regression with a single feature (mean friends age); (3) ElasticNet: elastic net regressor with a single feature (mean friends age); (4) GradBoost: gradient boosting with a single feature (mean friends age). 9
  • 10. algorithms based on word embeddings • The sum and/or mean of word embeddings to represent a sentence/paragraph. • The idea is not as naive as it seems; successfully applied in a few studies (please see the paper for references). • Three basic algorithms: (1) MeanVec: train on mean vectors of all statuses for a user; (2) LargestCluster: train on the centroid of the largest cluster of statuses; (3) AllMeanV: train on every status independently, with the mean vector of a specific status and the mean age of friends as features and the user’s demography as target; at the testing stage, we compute predictions for every status and average the predictions. 10
  • 11. meanvec (1) for every user 𝑢 ∈ 𝑈: • for every status 𝑠 ∈ 𝑆 𝑢, compute its mean vector ̄v 𝑠 = 1 |𝑠| ∑ 𝑤∈𝑆 v 𝑤; • compute the mean vector of all statuses ̄v 𝑢 = 1 |𝑆 𝑢| ∑ 𝑠∈𝑆 𝑢 ̄v 𝑠; (2) train ML with features (MAF 𝑢, ̄v 𝑢) for every 𝑢 ∈ 𝑈. 11
  • 12. largestcluster The LargestCluster algorithm operates as follows: for a machine learning algorithm ML, (1) for every user 𝑢 ∈ 𝑈: • for every status 𝑠 ∈ 𝑆 𝑢, compute its mean vector ̄v 𝑠 = 1 |𝑠| ∑ 𝑤∈𝑆 v 𝑤; • cluster the set of vectors { ̄v 𝑠 ∣ 𝑠 ∈ 𝑆} into two clusters with agglomerative clustering; denote by 𝐶 ⊆ 𝑆 the larger cluster; • compute the mean vector of statuses from 𝐶 ̄c 𝑢 = 1 |𝐶| ∑ 𝑠∈𝐶 ̄v 𝑠; (2) train ML with features (MAF 𝑢, ̄c 𝑢) for every 𝑢 ∈ 𝑈. 12
  • 13. allmeanv The AllMeanV algorithm operates as follows: for a machine learning algorithm ML, (1) for every user 𝑢 ∈ 𝑈 and every status 𝑠 ∈ 𝑆 𝑢, compute its mean vector ̄v 𝑠 = 1 |𝑠| ∑ 𝑤∈𝑆 v 𝑤; (2) train ML with features (MAF 𝑢, ̄v 𝑠) for every 𝑢 ∈ 𝑈 and 𝑠 ∈ 𝑆 𝑢; (3) on the prediction stage, for a user 𝑢 ∈ 𝑈test: • for every status 𝑠 ∈ 𝑆 𝑢, compute its mean vector ̄v 𝑠 = 1 |𝑠| ∑ 𝑤∈𝑆 v 𝑤; • predict the age for this status, 𝑎 𝑠 = ML(MAF 𝑢, ̄v 𝑠); • return the average predicted age, 𝑎 = 1 |𝑆 𝑢| ∑ 𝑠∈𝑆 𝑢 ML(MAF 𝑢, ̄v 𝑠). 13
  • 14. evaluation: prepared data Train/test sets were splitted 80:20. Dataset Training set Test set Users Statuses Users Statuses 661206 10880321 165301 2704883 14
  • 15. evaluation: comparing meanvec models using word2vec with different parameters • Adding word2vec components do help all models. • CBOW models outperform skip-gram models. • Increasing dimensionality adds minor improvements. • Vocabulary threshold tweaking doesn’t affect the results greatly. At the same time, it increases training time when made less strict. Thus, tuning the threshold allows to reduce computational expenses without loss in model quality. • The more expressive the classifier is, the better normalized word2vec models are. 15
  • 16. evaluation: meanvec vs largestcluster • MeanVec outperforms LargestCluster in all experiments. • Possibly: largest cluster consists of least meaningful items, e.g. updates from online game, short texts (e.g. “:)”). Verified with direct examination. • Requires thorough research, analysis and experiments. Variations on clustering may still prove to be useful. 16
  • 17. evaluation: meanvec vs largestcluster vs allmeanv • LargestCluster loses in almost all cases to both MeanVec and AllMeanV . • AllMeanV is on par with MeanVec with linear models, but not with GradBoost: MeanVec wins. This can be interpreted as showing that simple averaging works very well in the semantic space (not surprising — many semantic relations are known to become linear in the space of embeddings); even better than building an ensemble of predictions from individual statuses. 17
  • 18. summary • Huge unlabeled text datasets processed. A number of word2vec models trained. • User age prediction algorithms suggested and compared with each other. • Improvements were not huge when compared to ’mean friends age’, but the general approach still looks promising. • Possible developments: • other word embeddings, possibly including ones that do word sense disambiguation, • text embeddings (such as Paragraph Vectors), • character-level embeddings. 18
  • 19. thank you! Thank you for your attention! 19
  • 20. appendix Table 2: A comparison of word2vec models on the extended dataset with the MeanVec algorithm Word2vec params Train Test Train, nonzero Test, nonzero type 𝑑 𝑛 𝑣 RMSE MAE RMSE MAE RMSE MAE RMSE MAE MeanAge 9.672 6.707 9.668 6.711 8.778 5.835 8.800 5.848 LinearRegr cbow 100 10 20 10.401 7.776 10.397 7.785 8.514 5.792 8.541 5.810 cbow 100 10 30 10.402 7.777 10.396 7.784 8.515 5.791 8.539 5.808 skip 100 10 30 10.818 8.219 10.813 8.232 8.624 5.863 8.645 5.879 skip 200 1 20 10.738 8.146 10.726 8.151 8.593 5.847 8.613 5.859 cbow 200 1 30 10.355 7.737 10.349 7.743 8.497 5.782 8.520 5.798 skip 200 1 30 10.735 8.143 10.724 8.148 8.592 5.846 8.613 5.859 cbow 300 1 30 10.338 7.722 10.329 7.727 8.492 5.779 8.512 5.794 skip 300 1 30 10.689 8.088 10.675 8.096 8.583 5.837 8.601 5.854 ElasticNet cbow 100 10 20 10.810 8.208 10.799 8.217 8.694 5.903 8.719 5.921 cbow 100 10 30 10.806 8.203 10.795 8.212 8.702 5.909 8.726 5.926 skip 100 10 30 11.239 8.641 11.229 8.653 8.741 5.938 8.766 5.956 skip 200 1 20 11.239 8.641 11.229 8.653 8.741 5.938 8.766 5.956 cbow 200 1 30 10.949 8.349 10.937 8.359 8.736 5.934 8.760 5.951 skip 200 1 30 11.239 8.641 11.229 8.653 8.741 5.938 8.766 5.956 cbow 300 1 30 11.026 8.433 11.017 8.445 8.741 5.938 8.766 5.956 skip 300 1 30 11.239 8.641 11.229 8.653 8.741 5.938 8.766 5.956 GradBoost cbow 100 10 20 9.089 6.352 9.065 6.344 8.399 5.697 8.394 5.699 cbow 100 10 30 9.093 6.356 9.066 6.345 8.401 5.699 8.395 5.700 skip 100 10 30 9.294 6.527 9.277 6.529 8.495 5.766 8.491 5.770 skip 200 1 20 9.363 6.580 9.342 6.576 8.519 5.785 8.512 5.785 cbow 200 1 30 9.067 6.341 9.043 6.333 8.383 5.682 8.377 5.683 skip 200 1 30 9.365 6.583 9.344 6.580 8.520 5.784 8.512 5.785 cbow 300 1 30 9.048 6.323 9.025 6.316 8.380 5.683 8.371 5.681 skip 300 1 30 9.387 6.596 9.367 6.595 8.536 5.799 8.532 5.804 20
  • 21. appendix Table 3: A comparison of word2vec models with their normalized versions on the extended dataset with the MeanVec algorithm Word2vec params Train Test Train, nonzero Test, nonzero type 𝑑 𝑛 𝑣 RMSE MAE RMSE MAE RMSE MAE RMSE MAE LinearRegr cbow 100 10 20 10.402 7.777 10.396 7.784 8.515 5.791 8.539 5.808 n.cbow 100 10 20 10.426 7.807 10.418 7.807 8.535 5.807 8.560 5.824 cbow 100 10 30 10.818 8.219 10.813 8.232 8.624 5.863 8.645 5.879 n.cbow 100 10 30 10.839 8.231 10.826 8.240 8.630 5.864 8.649 5.879 skip 100 10 30 10.738 8.146 10.726 8.151 8.593 5.847 8.613 5.859 n.skip 100 10 30 10.753 8.156 10.736 8.159 8.601 5.853 8.620 5.864 skip 200 1 20 10.355 7.737 10.349 7.743 8.497 5.782 8.520 5.798 n.skip 200 1 20 10.363 7.748 10.351 7.746 8.512 5.795 8.532 5.808 cbow 200 1 30 10.735 8.143 10.724 8.148 8.592 5.846 8.613 5.859 n.cbow 200 1 30 10.750 8.152 10.733 8.155 8.601 5.853 8.620 5.864 skip 200 1 30 10.338 7.722 10.329 7.727 8.492 5.779 8.512 5.794 n.skip 200 1 30 10.333 7.720 10.319 7.717 8.501 5.789 8.518 5.800 cbow 300 1 30 10.689 8.088 10.675 8.096 8.583 5.837 8.601 5.854 n.cbow 300 1 30 10.687 8.083 10.668 8.084 8.586 5.840 8.603 5.853 ElasticNet cbow 100 10 20 10.806 8.203 10.795 8.212 8.702 5.909 8.726 5.926 n.cbow 100 10 20 11.239 8.641 11.229 8.653 8.741 5.938 8.766 5.956 cbow 100 10 30 11.239 8.641 11.229 8.653 8.741 5.938 8.766 5.956 n.cbow 100 10 30 11.239 8.641 11.229 8.653 8.741 5.938 8.766 5.956 skip 100 10 30 11.239 8.641 11.229 8.653 8.741 5.938 8.766 5.956 n.skip 100 10 30 11.239 8.641 11.229 8.653 8.741 5.938 8.766 5.956 skip 200 1 20 10.949 8.349 10.937 8.359 8.736 5.934 8.760 5.951 n.skip 200 1 20 11.239 8.641 11.229 8.653 8.741 5.938 8.766 5.956 cbow 200 1 30 11.239 8.641 11.229 8.653 8.741 5.938 8.766 5.956 n.cbow 200 1 30 11.239 8.641 11.229 8.653 8.741 5.938 8.766 5.956 skip 200 1 30 11.026 8.433 11.017 8.445 8.741 5.938 8.766 5.956 n.skip 200 1 30 11.239 8.641 11.229 8.653 8.741 5.938 8.766 5.956 cbow 300 1 30 11.239 8.641 11.229 8.653 8.741 5.938 8.766 5.956 n.cbow 300 1 30 11.239 8.641 11.229 8.653 8.741 5.938 8.766 5.956 GradBoost cbow 100 10 20 9.093 6.356 9.066 6.345 8.401 5.699 8.395 5.700 n.cbow 100 10 20 9.043 6.320 9.020 6.313 8.375 5.680 8.367 5.677 cbow 100 10 30 9.294 6.527 9.277 6.529 8.495 5.766 8.491 5.770 n.cbow 100 10 30 9.241 6.487 9.216 6.480 8.481 5.760 8.472 5.758 skip 100 10 30 9.363 6.580 9.342 6.576 8.519 5.785 8.512 5.785 n.skip 100 10 30 9.204 6.458 9.177 6.449 8.456 5.742 8.448 5.742 skip 200 1 20 9.067 6.341 9.043 6.333 8.383 5.682 8.377 5.683 n.skip 200 1 20 8.998 6.290 8.973 6.281 8.350 5.660 8.337 5.658 cbow 200 1 30 9.365 6.583 9.344 6.580 8.520 5.784 8.512 5.785 n.cbow 200 1 30 9.204 6.457 9.176 6.447 8.455 5.743 8.448 5.743 skip 200 1 30 9.048 6.323 9.025 6.316 8.380 5.683 8.371 5.681 n.skip 200 1 30 8.983 6.274 8.965 6.271 8.341 5.656 8.333 5.656 cbow 300 1 30 9.387 6.596 9.367 6.595 8.536 5.799 8.532 5.804 n.cbow 300 1 30 9.172 6.438 9.150 6.430 8.442 5.732 8.430 5.730 21
  • 22. appendix Table 4: A comparison of the MeanVec and LargestCluster algorithms on the extended dataset for various normalized word2vec models Word2vec params Train Test Train, nonzero Test, nonzero type 𝑑 𝑛 𝑣 RMSE MAE RMSE MAE RMSE MAE RMSE MAE LinearRegr MV cbow 100 10 20 10.401 7.776 10.397 7.785 8.514 5.792 8.541 5.810 LC cbow 100 10 20 10.548 7.926 10.542 7.929 8.573 5.834 8.601 5.852 MV cbow 100 10 30 10.402 7.777 10.396 7.784 8.515 5.791 8.539 5.808 LC cbow 100 10 30 10.553 7.933 10.551 7.939 8.574 5.832 8.603 5.853 MV skip 200 1 20 10.738 8.146 10.726 8.151 8.593 5.847 8.613 5.859 LC skip 200 1 20 10.849 8.251 10.833 8.255 8.629 5.870 8.650 5.883 MV cbow 200 1 30 10.355 7.737 10.349 7.743 8.497 5.782 8.520 5.798 LC cbow 200 1 30 10.493 7.878 10.494 7.886 8.554 5.822 8.581 5.839 ElasticNet MV cbow 100 10 20 10.810 8.208 10.799 8.217 8.694 5.903 8.719 5.921 LC cbow 100 10 20 11.239 8.641 11.229 8.653 8.741 5.938 8.766 5.956 MV cbow 100 10 30 10.806 8.203 10.795 8.212 8.702 5.909 8.726 5.926 LC cbow 100 10 30 11.239 8.641 11.229 8.653 8.741 5.938 8.766 5.956 MV skip 200 1 20 11.239 8.641 11.229 8.653 8.741 5.938 8.766 5.956 LC skip 200 1 20 11.239 8.641 11.229 8.653 8.741 5.938 8.766 5.956 MV cbow 200 1 30 10.949 8.349 10.937 8.359 8.736 5.934 8.760 5.951 LC cbow 200 1 30 11.239 8.641 11.229 8.653 8.741 5.938 8.766 5.956 GradBoost MV cbow 100 10 20 9.089 6.352 9.065 6.344 8.399 5.697 8.394 5.699 LC cbow 100 10 20 9.122 6.388 9.100 6.381 8.427 5.720 8.419 5.720 MV cbow 100 10 30 9.093 6.356 9.066 6.345 8.401 5.699 8.395 5.700 LC cbow 100 10 30 9.141 6.399 9.120 6.396 8.431 5.720 8.428 5.723 MV skip 200 1 20 9.363 6.580 9.342 6.576 8.519 5.785 8.512 5.785 LC skip 200 1 20 9.277 6.520 9.250 6.508 8.493 5.772 8.490 5.770 MV cbow 200 1 30 9.067 6.341 9.043 6.333 8.383 5.682 8.377 5.683 LC cbow 200 1 30 9.090 6.361 9.069 6.353 8.406 5.700 8.398 5.702 22
  • 23. appendix Table 5: A comparison of the MeanVec, LargestCluster, and AllMeanV algorithms on the basic dataset for various normalized word2vec models Word2vec params Train Test type 𝑑 𝑛 𝑣 RMSE MAE RMSE MAE LinearRegr mv cbow 100 10 20 8.778 6.066 8.791 6.087 lc cbow 100 10 20 8.927 6.174 8.946 6.200 amv cbow 100 10 20 8.981 6.096 8.991 6.125 mv cbow 100 10 30 8.798 6.087 8.806 6.095 lc cbow 100 10 30 8.954 6.201 8.955 6.204 amv cbow 100 10 30 8.992 6.108 9.001 6.133 mv skip 100 10 30 9.051 6.256 9.054 6.268 lc skip 100 10 30 9.153 6.332 9.146 6.335 amv skip 100 10 30 9.186 6.249 9.189 6.270 mv skip 200 1 20 8.974 6.216 8.983 6.225 lc skip 200 1 20 9.084 6.292 9.091 6.299 amv skip 200 1 20 9.132 6.210 9.137 6.231 mv cbow 200 1 30 8.734 6.035 8.736 6.052 lc cbow 200 1 30 8.898 6.158 8.902 6.169 amv cbow 200 1 30 8.944 6.071 8.952 6.098 mv skip 200 1 20 8.976 6.217 8.985 6.226 lc skip 200 1 20 9.086 6.296 9.094 6.301 amv skip 200 1 20 9.133 6.211 9.139 6.232 ElasticNet mv cbow 100 10 20 9.390 6.498 9.397 6.522 lc cbow 100 10 20 9.390 6.498 9.397 6.522 amv cbow 100 10 20 9.398 6.428 9.398 6.451 mv cbow 100 10 30 9.390 6.498 9.397 6.522 lc cbow 100 10 30 9.390 6.498 9.397 6.522 amv cbow 100 10 30 9.399 6.428 9.398 6.451 mv skip 100 10 30 9.390 6.498 9.397 6.522 lc skip 100 10 30 9.390 6.498 9.397 6.522 amv skip 100 10 30 9.399 6.428 9.398 6.451 mv skip 200 1 20 9.390 6.498 9.397 6.522 lc skip 200 1 20 9.390 6.498 9.397 6.522 amv skip 200 1 20 9.398 6.428 9.398 6.451 mv cbow 200 1 30 9.390 6.498 9.397 6.522 lc cbow 200 1 30 9.390 6.498 9.397 6.522 amv cbow 200 1 30 9.399 6.428 9.398 6.451 mv skip 200 1 20 9.390 6.498 9.397 6.522 lc skip 200 1 20 9.390 6.498 9.397 6.522 amv skip 200 1 20 9.399 6.428 9.398 6.451 GradBoost mv cbow 100 10 20 8.075 5.398 8.068 5.399 lc cbow 100 10 20 8.163 5.456 8.172 5.467 amv cbow 100 10 20 8.228 5.447 8.275 5.486 mv cbow 100 10 30 8.086 5.404 8.077 5.403 lc cbow 100 10 30 8.194 5.479 8.185 5.476 amv cbow 100 10 30 8.240 5.456 8.286 5.493 mv skip 100 10 30 8.277 5.534 8.271 5.537 lc skip 100 10 30 8.333 5.572 8.336 5.577 amv skip 100 10 30 8.362 5.548 8.408 5.583 mv skip 200 1 20 8.242 5.526 8.231 5.513 lc skip 200 1 20 8.308 5.562 8.307 5.559 amv skip 200 1 20 8.347 5.541 8.392 5.572 mv cbow 200 1 30 8.032 5.368 8.026 5.366 lc cbow 200 1 30 8.142 5.447 8.144 5.446 amv cbow 200 1 30 8.220 5.443 8.264 5.479 mv skip 200 1 20 8.240 5.518 8.230 5.510 lc skip 200 1 20 8.308 5.560 8.302 5.558 amv skip 200 1 20 8.347 5.542 8.392 5.573 23