SlideShare uma empresa Scribd logo
1 de 8
Baixar para ler offline
Towards Building Semantic Role Labeler for Indian Languages
Maaz Anwar Nomani♣
and Dipti Misra Sharma♣
FC Kohli Center on Intelligent Systems (KCIS), IIIT-Hyderabad, India♣
maaz.anwar@research.iiit.ac.in, dipti@iiit.ac.in,
Abstract
We present a statistical system for identifying the semantic relationships or semantic roles for two major Indian Languages, Hindi and
Urdu. Given an input sentence and a predicate/verb, the system first identifies the arguments pertaining to that verb and then classifies it
into one of the semantic labels which can either be a DOER, THEME, LOCATIVE, CAUSE, PURPOSE etc. The system is based on 2
statistical classifiers trained on roughly 130,000 words for Urdu and 100,000 words for Hindi that were hand-annotated with semantic
roles under the PropBank project for these two languages. Our system achieves an accuracy of 86% in identifying the arguments of
a verb for Hindi and 75% for Urdu. At the subsequent task of classifying the constituents into their semantic roles, the Hindi system
achieved 58% precision and 42% recall whereas Urdu system performed better and achieved 83% precision and 80% recall. Our study
also allowed us to compare the usefulness of different linguistic features and feature combinations in the semantic role labeling task. We
also examine the use of statistical syntactic parsing as feature in the role labeling task.
Keywords: PropBank, Semantic Role labeling, Treebank, Dependency relations
1. Introduction
We introduce a statistical semantic role labeler for Hindi
and Urdu, two major Indian languages. A Semantic Role
labeler (henceforth, SRL) automatically marks the argu-
ments/valency of a predicate in a sentence. The proposed
system is based on supervised machine learning approach
on Hindi and Urdu PropBanks which are being built for
these languages.
The approach is a 2-stage architecture in which, first the
arguments pertaining to a predicate in a sentence are iden-
tified by the system and then those identified arguments are
classified into one of the PropBank semantic labels. Our
system uses a basic Logistic Regression machine learn-
ing algorithm (Pedregosa et al., 2011) for identifying the
predicates and Support Vector Machines (Pedregosa et al.,
2011) to classify the arguments of a predicate into seman-
tic labels. We have used 10 linguistic features, 6 of which
have been derived from literature and 4 are novel features.
Among the new features, we have used the karaka rela-
tions/dependency relations (described later) which turns
out to be the most discriminative feature among all. We
have also experimented with the automatic parses which
are used as features and derived from state-of-art Hindi and
Urdu parsers. To the best of our knowledge, this is the first
such attempt of building a semantic role labeler for any In-
dian language.
This paper is arranged as follows: Section 2 gives a brief
introduction about Semantic Parsing. In Section 3, we give
the data statistics of the language resources used. Section 4
talks about related work in semantic role labeling. In sec-
tion 5, we briefly describe the Hindi and Urdu Propbanks
and Treebanks which are being built for these two Indian
languages. Section 6 presents the Semantic Role labeler in
detail which includes our approach, its architecture, classi-
fiers and features used. We also describe the baseline fea-
tures, their impact on Indian language datasets and new fea-
tures incorporated in the systems. In section 7, we describe
in detail the different experiments done and show their re-
sults. Section 8 throws light on the impact of using auto-
matic parses as features in the semantic role labeling task.
In section 9, we thoroughly put the error analysis forward
along with some corpus examples showing erroneous role
labeling and discuss possible reasons for it. We conclude
the paper in section 10.
2. Semantic Parsing
Semantic analysis of natural text and languages has always
been an intriguing research area in NLP. Automatic seman-
tic analysis of texts thus becomes a challenging as well as
interesting task which includes tasks like Semantic Pars-
ing, Shallow Semantic Parsing and Semantic Role label-
ing. Automatic and accurate tools that can predict naturally
occurring arguments for a given predicate in a predicate-
argument structure of a sentence can be efficiently used for
Semantic Parsing, Summarization, Information Extraction
tasks, wider NLP areas like Machine Translation and Ques-
tion Answering and modern NLP challenges like parsing
code-mixed data.
Semantic Parsing is essentially the research investigation of
identifying WHO did WHAT to WHOM, WHERE, HOW,
WHY and WHEN etc. in a sentence and adding a layer of
semantic annotation in a sentence produces such a struc-
ture. Two major Indian languages, Hindi and Urdu are hav-
ing such a resource in the form of PropBank (also called
Proposition Banks) which establishes a layer of semantic
representation in a Treebank already annotated with Depen-
dency labels.
Proposition Bank (Kingsbury and Palmer, 2003) is a corpus
in which the arguments of each verb predicate (simple or
complex) are marked with their semantic roles.
3. Data
Table 1 shows the data statistics of the language resources.
We have used a part of these language resources for SRL
task.
4588
In the experiments reported here, we first performed the ar-
gument identification task which is a binary classification
problem (either ‘argument’ or ‘not an argument’) by tak-
ing 130,000 tokens as training data and 30,000 as test data
from Urdu PropBank. We took around 100,000 tokens as
training data from Hindi PropBank and 20,000 as test data.
Subsequently, we trained a multi-class classifier, SVM on
the same training set, using a well defined feature-set de-
scribed later.
Language Resource Tokens Sentences Predicates pbrel
Urdu Treebank ˜200,000 ˜8,000 - -
Hindi Treebank ˜350,000 ˜14,000 - -
Urdu Propbank ˜180,000 ˜7,000 ˜2,200 ˜7,000
Hindi Propbank ˜300,000 ˜11,000 ˜3,200 ˜12,000
Table 1: Hindi and Urdu resources statistics
4. Related Work
The last decade has seen an exhaustive research in se-
mantic parsing for different languages. Gildea and Juraf-
sky (2002) started out the work on semantic role label-
ing on 2001 release of English Propbank. Surdeanu et al.
(2003) build a decision tree classifier for predicting the se-
mantic labels. Gildea and Hockenmaier (2003) uses fea-
tures from Combinatory Categorial Grammar (CCG) which
is a form of dependency grammar. Chen and Rambow
(2003) used a decision tree classifier and additional syntac-
tic and semantic representations extracted from Tree Ad-
joining Grammar (TAG). Swier and Stevenson (2004) talks
of a novel bootstrapping algorithm for identifying semantic
labels. (Cohn and Blunsom, 2005) applied conditional ran-
dom fields (CRFs) to the semantic role labeling task. Xue
and Palmer (2004) experimented with different linguistic
features related to role labeling task.
Xue and Palmer (2005) built a role labeler for Chinese
by showing that verb classes, induced from the predicate-
argument information in the frame files helps in semantic
role labeling. Johansson and Nugues (2006) came up with a
FrameNet-based semantic role labeling system for Swedish
text. Toutanova et al. (2005) used joint learning amd joint
modeling of argument frames of verbs to improve the over-
all accuracy of semantic role labeler.
5. The PropBank and The Treebank
Indian Languages are very morphologically rich languages
and Treebanks for Hindi and Urdu are available which in-
clude syntactico-semantic information in the form of de-
pendency annotations as well as lexical semantic infor-
mation in the form of predicate-argument structures, thus
forming PropBanks. Hindi and Urdu PropBanks are part
of a multi-dimensional and multi-layered resource creation
effort for the Hindi-Urdu language (Bhatt et al., 2009).
The Hindi Dependency Treebank (HDT) and Urdu Depen-
dency Treebank (UDT) are built following the CPG (Com-
putational Paninian Grammar) framework (Begum et al.,
2008). PropBank establishes a layer of semantic represen-
tation in the Treebank which is already annotated with De-
pendency labels or phrase structures (as in Penn Treebank
for English). Capturing the semantics through predicate-
argument structure involves quite a few challenges pecu-
liar to each predicate type because the syntactic notions in
which the verb’s arguments and adjuncts are realized can
vary based on the senses. Table 2 shows the PropBank la-
bels for Hindi and Urdu and these labels are also used for
building the Semantic Role labeler.
Propbank labels or semantic labels are closely associated
with the karaka relations (Vaidya et al., 2011) (described
later) in their structure though the former are defined on a
verb-by-verb basis.
Label Description
ARG0 Agent, Experiencer or
doer
ARG1 Patient or Theme
ARG2 Beneficiary
ARG3 Instrument
ARG2-ATR Attribute or Quality
ARG2-LOC Physical Location
ARG2-GOL Goal
ARG2-SOU Source
ARGM-PRX noun-verb construc-
tion
ARGM-ADV Adverb
ARGM-DIR Direction
ARGM-EXT Extent or Compari-
sion
ARGM-MNR Manner
ARGM-PRP Purpose
ARGM-DIS Discourse
ARGM-LOC Abstract Location
ARGM-MNS Means
ARGM-NEG Negation
ARGM-TMP Time
ARGM-CAU Cause or Reason
Table 2: Hindi and Urdu PropBank labels with defini-
tions
6. Semantic Role labeler
This section gives a thorough analysis on our approach to-
wards role labeling task. At first, one can comprehend
the semantic role labeling task as a simple straight forward
multi-class classification problem in which the semantic la-
bels are classified directly without identifying them. But
such a simple approach will not work due to various rea-
sons.
(Xue and Palmer, 2004) observes that for a given predicate,
many constituents in a syntactic tree are not its semantic
argument. So, non-argument count overwhelms the argu-
ment count for the given predicate and classifiers will not
be efficient in predicting the right argument or in classify-
ing them. This problem was encountered in both Hindi and
Urdu data-sets as well.
Therefore, we adopted a 2-stage approach of first iden-
tifying the arguments and subsequently classifying the
identified semantic arguments into one of the semantic
labels. The second approach of directly classifyng the
arguments was also experimented by us and we show by
4589
statistics that such a forthright approach will not work for
Hindi and Urdu as well.
• Approach I
• Argument Identification - This is the sub-task of iden-
tifying a constituent in a dependency tree which repre-
sents the argument for the predicate in the argument-
predicate structure in a Hindi or Urdu PropBank sen-
tence. The identification is done by a binary classifier
(‘argument’ and ‘not an argument’). The classifier has
to predict whether the constituent in the tree is an ar-
gument of that verb or not.
• Argument Classification - After identification of argu-
ments is done, the identified arguments pertaining to
that verb are assigned appropriate semantic labels by
a multi-class classifier which is trained on multiple se-
mantic labels present in the training data. We have
used Gold Parses as well as Automatic Parses from
Hindi and Urdu parsers as features in this task.
• Approach II
• Argument Identification and Argument Classification -
We performed direct classification of arguments in
Hindi and Urdu data and found that the results were
not effective. We found that due to overwhelming
number of non-arguments present in the corpus, the
classifier failed to classify the arguments correctly.
For example, in Urdu SRL system, the arguments for
which gold parses were used as feature, were getting
classified with a precision of only 20.98% and non-
arguments were getting classified with a precision of
79.71%. For automatic parses, arguments got classi-
fied with a precision of 17.12% and non-arguments got
classified with 82.87% precision.
6.1. SRL Architecture
Figure 1 is the flow diagram of Semantic Role Labeler for
Hindi and Urdu. It is a 2-stage architecture in which at first
the arguments of a predicate in a sentence are identified by a
binary classifier and then in the subsequent stage only these
identified arguments are classified as one of the semantic
labels by a multi-class classifier.
6.2. Features: A critical look
Hindi and Urdu are free-word order languages i.e. the word
order is somewhat flexible in these languages. For this rea-
son, they are sometimes called as SOV languages (subject,
object, verb). Hindi and Urdu employs postpositions and
not prepositions (as in English), the resulting word order of
postpositional phrases can be the reverse of prepositional
phrases in English. We decided to use linguistic features
in our Semantic Role Labeler based on our intuition that
they might help the labeler in taking decisions to identify
and classify the semantic arguments of a predicate in the
sentence.
Filtering of pbrole sentences
Argument Identifier
Identified Arguments projected on Test Data
Only Arguments
Argument Classifier
Classified/Predicted labels
Projected to Test Data of PropBank
Test Data containing automatic SR labels
Training (Binary)
Training (Multi-Class)
Extracted
Figure 1: SRL Architecture
6.2.1. Baseline Features
The task of identifying and classifying the arguments of a
verb is done at chunk/phrase level. Most of the previous
works done in Semantic Parsing have shown that ‘head’ of
the chunk is a very critical feature because chunks/phrases
with certain head words are likely to be arguments of the
verb and also certain types of arguments. This implies that
head word of a chunk is a very discriminative feature for
both the tasks.
For our Indian languages SRL, we have used a combination
of baseline features introduced by (Gildea and Jurafsky,
2002) and (Xue and Palmer, 2004) for English because
we wanted to see the impact of these standard features in
context of the Indian languages and how these features
perform w.r.t free word order languages.
Baseline Features for Indian Languages
Predicate - predicate itself
Head word - syntactic head of the chunk/phrase
Head-word POS - Its POS category
Phrase Type - syntactic category (NP, VP etc. of the phrase)
Predicate + Phrase Type - combinational feature
Predicate + Head word - combinational feature
Table 3: Baseline feature template
6.2.2. New Features
After analyzing the PropBanks of both the languages, we
came up with certain new features for which we had an
intuition that they will contribute significantly towards this
task. These are discussed below:
4590
• Dependency/karaka relation - We experimented with
the Dependency labels or karaka relations present in
the Hindi and Urdu PropBanks. A sentence is com-
posed of a primary modified which is the root in the
dependency tree of the sentence and this modified is
occasionally the main verb of the sentence. The words
modifying this verb participate in the action specified
by the verb. The participant relation of the verb with
these words/elements are called karaka relations.
There are in total 43 karaka relations for Hindi and
Urdu as specified by (Begum et al., 2008). Table 4
shows some of the major karaka or dependency re-
lations. These dependency relations helps to identify
the syntactico-semantic relations in a sentence. Since,
PropBank labels illustrates the semantic relations be-
tween the constituents, we had this intuition that using
the karaka relations as features in our classifier to pre-
dict the arguments of a verb will have an impact on our
system performance. Also, Vaidya et al. (2011) has
shown that there is a correlation between dependency
and predicate argument structures for Hindi. Since
Hindi and Urdu are linguistically similar languages,
we extracted the dependency/karaka relations of both
the language Treebanks and used them as feature in
isolation as well as in the system.
Dependency Labels Description
k1 karta - doer/agent/subject
k2 karma - object/patient
k1s noun complement
k4 sampradana - recipient
k3 karana - instrument
k4a experiencer
k2p goal, destination
k5 apadana - source
k7 location elsewhere
k7p location in space
k7t time
adv adverbs
rh reason
rd direction
rt purpose
Table 4: Major karaka relations or Dependency
labels of Hindi and Urdu Treebank
From Tables 5 and 6, it is clear that dependency rela-
tions are indeed the most discriminative feature among
all the features used by showing high precision and re-
call for both languages. The high percentage of preci-
sion and recall in the Argument Classification task can
be accounted from the fact that dependency structure
and predicate-argument structure are somewhat simi-
lar structures and dependency labels and propbank la-
bels are similar in orientation. Therefore, using de-
pendency structures as features to predict and classify
the arguments of a predicate of Hindi and Urdu is very
influential.
• Named Entities (NEs) - Named Entities also contribute
towards the identification task but it doesn’t con-
tributes heavily towards the classification task. The
reason being NEs are generally arguments of a predi-
cate. This feature has been used in the past by (Prad-
han et al., 2004) for English but we have used it dif-
ferently by combining it with the POS category of the
NE. As shown is Tables 5 and 6, NEs have a decent
precision of 63% and 61% for Urdu and Hindi respec-
tively when used in isolation for the argument identifi-
cation task. But it is not able to discriminate between
the class of semantic labels to which a identified argu-
ment belong.
• Head + Phrase type - A combination of head word of
the chunk/phrase and syntactic category of the phrase.
• Head POS + Phrase type - A combination of the POS
category of the head word of the chunk/phrase and
syntactic category of the phrase.
Abbreviations Used: ‘P’=Precision, ‘R’=Recall,
‘PT’=Phrase Type, ‘HW’=Head Word, ‘POS’=Part-
of-speech’
Feature
¯
Feature wise performance (Urdu)
¯
Argument Id.
¯
Argument Classification
¯
P R f-score P R f-score
Predicate 51 72 60 00 01 00
Head-word 71 73 72 25 14 18
Head-POS 51 72 60 06 18 09
PhraseType 51 72 60 08 08 08
HeadPOS-PT 61 72 66 06 18 09
Head-PT 71 74 72 25 14 18
Predicate-HW 72 74 73 01 01 01
Predicate-PT 64 69 66 02 00 00
NEs 63 65 64 23 18 20
Dependency 78 79 78 87 85 86
Table 5: Individual feature performance (Urdu)
• We also experimented with certain other features like chunk
boundary and sentence boundary of the sentences. We had
an intuition that these boundaries will help the classifier
learn more effectively and overlap of the 2 sentences will
be minimized. But it turned out that both the features had a
major negative impact on the overall results of the system,
so we scrapped these two features.
The reason for this dip can be accounted from the fact
that chunk boundary and sentence boundary are merely
integers for the classifier as there are ‘n’ numbers for ‘n’
sentences/chunks which lack any significant information.
6.2.3. Classifiers used
We used a basic Logistic Regression classifier and trained
it on Hindi and Urdu data-sets thus building a binary
classifier which identifies arguments as ‘Arguments’ or
‘Not-an-Argument’. For multi-class classification in
argument classification task, we implemented LinearSVC
class of SVM (Pedregosa et al., 2011) for performing the
‘one-vs-the-rest’ multi-class strategy. All the parametres
4591
of Linear SVC were set to default at the time of training of
Hindi and Urdu data-sets.
Feature
¯
Feature wise performance (Hindi)
¯
Argument Id.
¯
Argument Classification
¯
P R f-score P R f-score
Predicate 41 64 50 41 61 49
Head-word 70 70 70 20 04 07
Head-POS 34 58 43 07 08 07
PhraseType 34 58 43 05 07 06
HeadPOS-PT 34 58 43 14 13 13
Head-PT 67 66 66 10 04 06
Predicate-HW 65 65 65 10 04 06
Predicate-PT 41 64 50 04 06 05
NEs 61 64 62 20 16 17
Dependency 88 87 87 52 49 50
Table 6: Individual feature performance (Hindi)
7. Experiments and Results
In this section, we show the results and extent of the
baseline features (standard features) in Semantic Role
Labeling in the experiments on the Indian data sets. We
present results of baseline features for both the task and
add new features subsequently in next step. We have laid
out the experiments in a way that each experiment conveys
the extent of usefulness and information gain of the feature
in both the tasks. For each experiment, the settings of
the SVM and features used in the previous experiment
were retained. The baseline features performs well on
the argument identification task of both languages with a
precision of 71% and 67% for Hindi and Urdu respectively.
The classification task of the identified arguments is
not profited much by the baseline features for both the
languages as can be seen from Tables 7 and 8.
Feature
¯
Hindi system Feature Template
¯
Argument Id.
¯
Argument Classification
¯
P R f-score P R f-score
Baseline 71 71 66 29 14 15
+Dependency 86 86 86 56 38 41
+Head-PT 85 85 86 56 38 41
+HeadPOS-PT 84 84 84 57 40 42
+NEs 85 84 84 58 42 49
Table 7: Hindi system performance
As expected, the best result for both the tasks is obtained
on incorporating dependency relations information, which
further strengthens our conjecture on the importance of de-
pendency labels in semantic role labeling. All 4 experi-
ments shows the effectiveness of the features for Hindi and
Urdu data sets. By incorporating NEs in the system, there is
an increase in the identification task of Urdu and marginal
increase in Hindi. The classification task is also benefitted
by a marginal precision increment of 1% in both the data
sets and recall of 2% in Hindi.
However, ‘Head + Phrase Type’ and ‘HeadPOS + Phrase
Type’ features had a negative impact on the performance
of classification task for both the sets. The identification
task was benefitted by every new feature in Hindi with
dependency relations contributing the most to the final
result of the system.
Feature
¯
Urdu system Feature Template
¯
Argument Id.
¯
Argument Classification
¯
P R f-score P R f-score
Baseline 67 70 68 16 14 15
+Head-PT 65 69 67 15 12 13
+HeadPOS-PT 66 69 67 15 12 13
+Dependency 72 74 73 82 80 81
+NEs 75 71 73 83 80 81
Table 8: Urdu system performance
8. Using Automatic Parses
So far, we have shown and discussed results using hand-
corrected parses of the Hindi and Urdu PropBank. How-
ever, if we are building a NLP application, the SRL will
have to extract features from automatic parses of the sen-
tences. Also, after getting the insights of the influence of
hand-corrected dependency relations, we wanted to analyse
the same using automatic parses and evaluate our semantic
role labeler on it. These automatic parses are generated by
the parser of respective language.
To come up with such an evaluation, we used the Malt
Parser (Nivre et al., 2006) on the same test data described
above for both languages. Malt Parser requires data in
CONLL format. By using state-of-art Urdu Parser (Bhat et
al., 2012), we extracted the automatic parses from the Urdu
test data and used them as feature for our identification and
classification tasks.
Parsing accuracy is measured by LAS, UAS and LA which
stands for Labeled Attachment score, Unlabeled Attach-
ment score and Label Accuracy respectively. LAS is the
percentage of words that are assigned the correct head and
dependency label; UAS is the percentage of words with the
correct head, and the label accuracy (LA) is the percentage
of words with the correct dependency label.
Measure Percentage (%)
LAS 76.61
UAS 88.45
LA 80.03
Table 9: Parsing performance of the Urdu dependency
parser
By adding automatic parses to the baseline features of Urdu
system, we observed a marginal increase in the argument
identification task and comparable improvement in classi-
fication task in Table 10. The decent jump in the precision
of the classification task while using automatic parses as
compared to using gold parses can be attributed from the
fact that gold parses (Urdu PropBank) have been manually
validated by annotators whereas Urdu parser tends to give
certain incorrect parses as can be seen from Table 9.
4592
Feature
¯
Automatic Parses (Urdu)
¯
Argument Id. Argument Classification
P R f-score P R f-score
Baseline 67 70 68 16 14 15
+Automatic Parses 68 66 67 62 59 60
Table 10: Urdu system with Automatic Parses
The same process was repeated by using the state-of-art
Hindi Parser (Kosaraju et al., 2010) on Hindi test data.
Measure Percentage (%)
LAS 88.19
UAS 94.00
LA 89.77
Table 11: Parsing performance of the Hindi dependency
parser
The usage of automatic parses of Hindi have similar impact
on the Hindi SRL system. Table 12 shows the performance
devaluation when automatically generated parses are used.
Feature
¯
Automatic Parses (Hindi)
¯
Argument Id.
¯
Argument Classification
¯
P R f-score P R f-score
Baseline 71 71 66 29 14 15
+Automatic Parses 73 74 73 41 28 33
Table 12: Hindi system with Automatic Parses
9. Error Analysis and Discussion
In this section, we present a thorough error analysis of the
Semantic Role Labeling task for Hindi and Urdu by taking
a sub-set of test data for error analysis. Tables 14 and 15
are the confusion matrices for Hindi and Urdu SRL respec-
tively. As Hindi and Urdu have postpositional phrases, the
postpositions or case markers play a very significant role in
deciding the semantic role for an argument. These postpo-
sitions may not be helpful cues in identifying the argument
but they are when it comes to classifying the arguments.
There are 6 pre-dominant case markers in Hindu and Urdu
viz. ‘ne’, ‘ko’, ‘mein’, ‘se’, ‘ki’, ‘par’. Table 13 gives
the definition of Hindi and Urdu case markers. These case
markers provide important information about the token pre-
ceding it and the nature of the chunk/phrase in which they
are present. For e.g. in both Hindi and Urdu, the case
clitic ‘ne’ is often indicative of an agent, and as ‘ARG0’
is mostly associated with ‘agentivity’, this can sometimes
provide a clue about ARG0. Same is the case with ‘mein’
which most of the times gives the ‘locative’ information in
the chunk/phrase. Some post-positions are very ambigous
in nature and here is when error in classifying the argu-
ments starts. We show in the examples below how ‘ko’ and
‘mein’ post-position can take multiple types of semantic la-
bels and therefore the classifier is not very discriminative
during multi-class classification.
Urdu Case marker Meaning
ne Ergative
mein Locative
par Locative
ko Dative/Accusative
se Instrumental/Ablative
ki Genetive
Table 13: Hindi and Urdu Case markers
As can be seen from confusion matrix of Hindi, the pair
of labels in which the Hindi SRL is having the maximum
confusion is ‘ARG1’ and ‘ARG0’. ‘ARG0’ label is given
to those arguments of verb which shows agentivity or vola-
tionality to do work whereas ARG1 is the theme or object
arguments. Another frequent confusion which the Hindi
SRL has is between ‘ArgM-LOC’ and ‘ARG0’. The con-
fusion between ‘ARGM-LOC’ and ‘ARG2-LOC’ can be
attributed to the fact that both the labels depicts locative
information though ARGM-LOC shows abstract location
whereas ARG2-LOC has physical location information.
Arg0 Arg1 Arg2 Arg2LOC Arg2SOU Arg2ATR ArgADV ArgMLOC ArgDIR
Arg0 23 03 02
Arg1 21 21 04 05 02 03
Arg2 01 01 01
Arg2LOC 02 02
Arg2SOU 02
Arg2ATR 03 01 07
ArgMADV 01
ArgMLOC 05 01 04 01 16
ArgMDIR 02
Table 14: Confusion Matrix of Propbank labels for
Hindi SRL
For Urdu, similar confusion is observed between ‘ARG0’-
‘ARG1’ and ‘ARGM-LOC’-‘ARG2-LOC’ pairs. Such con-
fusions can be reduced by coming up with linguistic fea-
tures which are aimed at solving ARG0 and ARG1 con-
fusion because it is the most frequent erroneous pair for
Indian languages SRL. These errors also occur due to the
presence of ambigous case markers with the arguments of
a verb and therefore the classifier is unable to discriminate
between the two labels.
Another interesting confusion pair is ARG1:ARG2-ATR.
There are instances in the training set in which the argu-
ments of a verb are annotated with ‘ARG1’ or ‘ARG2-ATR’
based on the context and sentence structure.
Arg0 Arg1 Arg2 Arg2LOC Arg2SOU Arg2ATR ArgADV ArgMLOC ArgDIR
Arg0 45 02 02 01
Arg1 08 48 03 05 02 04
Arg2 01 07 01
Arg2LOC 10 01 01 03
Arg2SOU 04 02 01 01 04
Arg2ATR 01
ArgMADV 01 12
ArgMLOC 01 36
ArgMDIR 03
Table 15: Confusion Matrix of Propbank labels for
Urdu SRL
We extracted certain sentences from the test set of Hindi
and Urdu and analyzed how our SRL is performing in com-
parison with the gold test data. These sentences are shown
below:
4593
(1) (iss
this
dauran)
period
[gold:ARGM-TMP; SRL: ARGM-TMP]
his
(uski)
daughter
(putri
DAT
ko)
man
[gold: ARG1; SRL: ARG1]
roam
(yuwak)
live
[gold: ARG0; SRL: ARG0]
be-PAST.
(ghumata)
[predicate] (rehta tha).
‘During this period, the man used to roam around his
daughter.’ HINDI TEST
(2) (idhar)
here
[gold: ARGM-DIR; SRL: ARGM-DIS]
Assam
(Assam
LOC
mein)
BJP
[gold: ARGM-LOC; SRL: ARGM-LOC]
main
(BJP)
party
[gold: ARG1; SRL: ARG0]
manner
(mukhya
LOC
party
emerge
ke)
live
(roop
be-PRS.
mein)
[gold: ARGM-MNR; SRL: ARGM-MNR]
(ubhar) [predicate] (rahi) (hai). URDU TEST
‘Here, in Assam, BJP is emerging as the main party.’
(3) (Jwala
Jwala
ne)
ERG
(bataya)
said
(ki)
that
(central
central
market
market
ke
comparison
muqaable
LOC
mein)
here
[gold: ARGM-EXT; SRL: ARGM-LOC]
rate
(yahan)
economical
[gold: ARG2-LOC; SRL: ARG2:LOC]
live
(rate)
be-PRS
[gold: ARG1; SRL: ARG1] (munaasib)
[gold: ARGM-PRX; SRL: ARGM:PRX] (rehta
hai) [predicate]. HINDI TEST
‘Jwala said that rate here is economical as compared
to central market.’
(4) (Barack
Barack
Obama
Obama
ko)
DAT
[gold: ARG0-GOL; SRL: ARG2]
feel
(ehsaas)
be-PRS
[gold: ARG1; SRL: ARG2-ATR]
thathis
(hai)
party
(ki)
election
(unki)
LOC
(party)
2009
[gold: ARG0; SRL: ARG1]
of
(intekhabaat
miracle
mein)
not
[gold: ARGM-LOC; SRL: ARGM-LOC]
see
(2009
get.
ka) (karishma) (nahi) (dikha) [predicate]
(paaegi). URDU TEST
‘Barack Obama feels that his party will not be able to
repeat the miracle of 2009 in the election.’
In sentence 3, ‘central market ke muqaable mein’ is
‘ARGM-EXT’ in gold test data as it conveys the mean-
ing of ‘comparison’ with some entity. But Hindi SRL
gives it ARGM-LOC label. One possible reason for this
may be that ‘mein’ case marker generally occurs more fre-
quently with ARGM-LOC or ARG2-LOC label in the loca-
tive sense in the training data as compared to comparison.
In sentence 4, ‘Barack Obama ko’ gets ARG2 label by Urdu
SRL instead of ARG0-GOL because of the same reasons
cited in the above example. Here, ‘ko’ case marker more
frequently occurs with ARG2 label i.e. beneficiary in the
training data as compared to experiencer here.
10. Conclusion
In this work, we have presented a supervised semantic role
labeler for Hindi and Urdu. We have used linguistic fea-
tures like predicate, head, head-POS, phrase type etc. and
combinations of certain features. In all, 10 features have
been used to guide the classifiers in predicting, identifying
and classifiyng the arguments of a verb.
We further used the gold parses (dependency parses) as
a feature which had a solid impact on the overall accu-
racy of the systems. Subsequently, we extracted the au-
tomatic parses by using state-of-art Hindi and Urdu parsers
and used them as features in our SRL. We believe that de-
pendency relations are the best features for classifying the
semantic arguments of a predicate because of their close
proximity with semantic labels.
11. Acknowledgements
We would like to thank Mr. Mujadia Vandan for the discus-
sions which we had with him to get a better insight of the
task.
12. Bibliographical References
Begum, R., Husain, S., Dhwaj, A., Sharma, D. M., Bai, L.,
and Sangal, R. (2008). Dependency annotation scheme
for indian languages. In IJCNLP, pages 721–726. Cite-
seer.
Bhat, R. A., Jain, S., and Sharma, D. M. (2012). Exper-
iments on dependency parsing of urdu. In The 11th In-
ternational Workshop on Treebanks and Linguistic The-
ories.
Bhatt, R., Narasimhan, B., Palmer, M., Rambow,
O., Sharma, D. M., and Xia, F. (2009). A
multi-representational and multi-layered treebank for
hindi/urdu. In Proceedings of the Third Linguistic Anno-
tation Workshop, pages 186–189. Association for Com-
putational Linguistics.
4594
Chen, J. and Rambow, O. (2003). Use of deep linguistic
features for the recognition and labeling of semantic ar-
guments. In Proceedings of the 2003 conference on Em-
pirical methods in natural language processing, pages
41–48. Association for Computational Linguistics.
Cohn, T. and Blunsom, P. (2005). Semantic role labelling
with tree conditional random fields. In Proceedings of
the Ninth Conference on Computational Natural Lan-
guage Learning, pages 169–172. Association for Com-
putational Linguistics.
Gildea, D. and Hockenmaier, J. (2003). Identifying seman-
tic roles using combinatory categorial grammar. In Pro-
ceedings of the 2003 conference on Empirical methods
in natural language processing, pages 57–64. Associa-
tion for Computational Linguistics.
Gildea, D. and Jurafsky, D. (2002). Automatic labeling
of semantic roles. Computational linguistics, 28(3):245–
288.
Johansson, R. and Nugues, P. (2006). A framenet-based
semantic role labeler for swedish. In Proceedings of
the COLING/ACL on Main conference poster sessions,
pages 436–443. Association for Computational Linguis-
tics.
Kingsbury, P. and Palmer, M. (2003). Propbank: the next
level of treebank. In Proceedings of Treebanks and lexi-
cal Theories, volume 3. Citeseer.
Kosaraju, P., Kesidi, S. R., Ainavolu, V. B. R., and
Kukkadapu, P. (2010). Experiments on indian language
dependency parsing. Proceedings of the ICON10 NLP
Tools Contest: Indian Language Dependency Parsing.
Nivre, J., Hall, J., and Nilsson, J. (2006). Maltparser: A
data-driven parser-generator for dependency parsing. In
Proceedings of LREC, volume 6, pages 2216–2219.
Pedregosa, F., Varoquaux, G., Gramfort, A., Michel, V.,
Thirion, B., Grisel, O., Blondel, M., Prettenhofer, P.,
Weiss, R., Dubourg, V., et al. (2011). Scikit-learn: Ma-
chine learning in python. The Journal of Machine Learn-
ing Research, 12:2825–2830.
Pradhan, S. S., Ward, W., Hacioglu, K., Martin, J. H., and
Jurafsky, D. (2004). Shallow semantic parsing using
support vector machines. In HLT-NAACL, pages 233–
240.
Surdeanu, M., Harabagiu, S., Williams, J., and Aarseth,
P. (2003). Using predicate-argument structures for in-
formation extraction. In Proceedings of the 41st Annual
Meeting on Association for Computational Linguistics-
Volume 1, pages 8–15. Association for Computational
Linguistics.
Swier, R. S. and Stevenson, S. (2004). Unsupervised se-
mantic role labelling. In Proceedings of EMNLP, vol-
ume 95, page 102.
Toutanova, K., Haghighi, A., and Manning, C. D. (2005).
Joint learning improves semantic role labeling. In Pro-
ceedings of the 43rd Annual Meeting on Association for
Computational Linguistics, pages 589–596. Association
for Computational Linguistics.
Vaidya, A., Choi, J. D., Palmer, M., and Narasimhan, B.
(2011). Analysis of the hindi proposition bank using de-
pendency structure. In Proceedings of the 5th Linguis-
tic Annotation Workshop, pages 21–29. Association for
Computational Linguistics.
Xue, N. and Palmer, M. (2004). Calibrating features for
semantic role labeling. In EMNLP, pages 88–94.
Xue, N. and Palmer, M. (2005). Automatic semantic role
labeling for chinese verbs. In IJCAI, volume 5, pages
1160–1165. Citeseer.
4595

Mais conteúdo relacionado

Mais procurados

A New Approach to Parts of Speech Tagging in Malayalam
A New Approach to Parts of Speech Tagging in MalayalamA New Approach to Parts of Speech Tagging in Malayalam
A New Approach to Parts of Speech Tagging in Malayalam
ijcsit
 
Parts of Speect Tagging
Parts of Speect TaggingParts of Speect Tagging
Parts of Speect Tagging
theyaseen51
 

Mais procurados (20)

Automatic classification of bengali sentences based on sense definitions pres...
Automatic classification of bengali sentences based on sense definitions pres...Automatic classification of bengali sentences based on sense definitions pres...
Automatic classification of bengali sentences based on sense definitions pres...
 
Clustering Algorithm for Gujarati Language
Clustering Algorithm for Gujarati LanguageClustering Algorithm for Gujarati Language
Clustering Algorithm for Gujarati Language
 
Named Entity Recognition for Telugu Using Conditional Random Field
Named Entity Recognition for Telugu Using Conditional Random FieldNamed Entity Recognition for Telugu Using Conditional Random Field
Named Entity Recognition for Telugu Using Conditional Random Field
 
RULE BASED TRANSLITERATION SCHEME FOR ENGLISH TO PUNJABI
RULE BASED TRANSLITERATION SCHEME FOR ENGLISH TO PUNJABIRULE BASED TRANSLITERATION SCHEME FOR ENGLISH TO PUNJABI
RULE BASED TRANSLITERATION SCHEME FOR ENGLISH TO PUNJABI
 
Ijartes v1-i1-002
Ijartes v1-i1-002Ijartes v1-i1-002
Ijartes v1-i1-002
 
Suitability of naïve bayesian methods for paragraph level text classification...
Suitability of naïve bayesian methods for paragraph level text classification...Suitability of naïve bayesian methods for paragraph level text classification...
Suitability of naïve bayesian methods for paragraph level text classification...
 
Comparative performance analysis of two anaphora resolution systems
Comparative performance analysis of two anaphora resolution systemsComparative performance analysis of two anaphora resolution systems
Comparative performance analysis of two anaphora resolution systems
 
A Proposition Bank of Urdu
A Proposition Bank of UrduA Proposition Bank of Urdu
A Proposition Bank of Urdu
 
ijcai05_srl
ijcai05_srlijcai05_srl
ijcai05_srl
 
p138-jiang
p138-jiangp138-jiang
p138-jiang
 
A New Approach to Parts of Speech Tagging in Malayalam
A New Approach to Parts of Speech Tagging in MalayalamA New Approach to Parts of Speech Tagging in Malayalam
A New Approach to Parts of Speech Tagging in Malayalam
 
Usage of regular expressions in nlp
Usage of regular expressions in nlpUsage of regular expressions in nlp
Usage of regular expressions in nlp
 
Usage of regular expressions in nlp
Usage of regular expressions in nlpUsage of regular expressions in nlp
Usage of regular expressions in nlp
 
Pronominal anaphora resolution in
Pronominal anaphora resolution inPronominal anaphora resolution in
Pronominal anaphora resolution in
 
A survey of named entity recognition in assamese and other indian languages
A survey of named entity recognition in assamese and other indian languagesA survey of named entity recognition in assamese and other indian languages
A survey of named entity recognition in assamese and other indian languages
 
Nlp (1)
Nlp (1)Nlp (1)
Nlp (1)
 
Parts of Speect Tagging
Parts of Speect TaggingParts of Speect Tagging
Parts of Speect Tagging
 
A Survey of Various Methods for Text Summarization
A Survey of Various Methods for Text SummarizationA Survey of Various Methods for Text Summarization
A Survey of Various Methods for Text Summarization
 
Hw6 interpreter iterator GoF
Hw6 interpreter iterator GoFHw6 interpreter iterator GoF
Hw6 interpreter iterator GoF
 
Ijarcet vol-3-issue-1-9-11
Ijarcet vol-3-issue-1-9-11Ijarcet vol-3-issue-1-9-11
Ijarcet vol-3-issue-1-9-11
 

Semelhante a Towards Building Semantic Role Labeler for Indian Languages

Chinese Word Segmentation in MSR-NLP
Chinese Word Segmentation in MSR-NLPChinese Word Segmentation in MSR-NLP
Chinese Word Segmentation in MSR-NLP
Andi Wu
 
Doc format.
Doc format.Doc format.
Doc format.
butest
 
A semi-supervised method for efficient construction of statistical spoken lan...
A semi-supervised method for efficient construction of statistical spoken lan...A semi-supervised method for efficient construction of statistical spoken lan...
A semi-supervised method for efficient construction of statistical spoken lan...
Seokhwan Kim
 
Ijarcet vol-2-issue-2-323-329
Ijarcet vol-2-issue-2-323-329Ijarcet vol-2-issue-2-323-329
Ijarcet vol-2-issue-2-323-329
Editor IJARCET
 

Semelhante a Towards Building Semantic Role Labeler for Indian Languages (20)

COMPREHENSIVE ANALYSIS OF NATURAL LANGUAGE PROCESSING TECHNIQUE
COMPREHENSIVE ANALYSIS OF NATURAL LANGUAGE PROCESSING TECHNIQUECOMPREHENSIVE ANALYSIS OF NATURAL LANGUAGE PROCESSING TECHNIQUE
COMPREHENSIVE ANALYSIS OF NATURAL LANGUAGE PROCESSING TECHNIQUE
 
FIRE2014_IIT-P
FIRE2014_IIT-PFIRE2014_IIT-P
FIRE2014_IIT-P
 
D3 dhanalakshmi
D3 dhanalakshmiD3 dhanalakshmi
D3 dhanalakshmi
 
STATISTICAL FUNCTION TAGGING AND GRAMMATICAL RELATIONS OF MYANMAR SENTENCES
STATISTICAL FUNCTION TAGGING AND GRAMMATICAL RELATIONS OF MYANMAR SENTENCESSTATISTICAL FUNCTION TAGGING AND GRAMMATICAL RELATIONS OF MYANMAR SENTENCES
STATISTICAL FUNCTION TAGGING AND GRAMMATICAL RELATIONS OF MYANMAR SENTENCES
 
Chinese Word Segmentation in MSR-NLP
Chinese Word Segmentation in MSR-NLPChinese Word Segmentation in MSR-NLP
Chinese Word Segmentation in MSR-NLP
 
Automatic multiple choice question generation system for
Automatic multiple choice question generation system forAutomatic multiple choice question generation system for
Automatic multiple choice question generation system for
 
THE ABILITY OF WORD EMBEDDINGS TO CAPTURE WORD SIMILARITIES
THE ABILITY OF WORD EMBEDDINGS TO CAPTURE WORD SIMILARITIESTHE ABILITY OF WORD EMBEDDINGS TO CAPTURE WORD SIMILARITIES
THE ABILITY OF WORD EMBEDDINGS TO CAPTURE WORD SIMILARITIES
 
THE ABILITY OF WORD EMBEDDINGS TO CAPTURE WORD SIMILARITIES
THE ABILITY OF WORD EMBEDDINGS TO CAPTURE WORD SIMILARITIESTHE ABILITY OF WORD EMBEDDINGS TO CAPTURE WORD SIMILARITIES
THE ABILITY OF WORD EMBEDDINGS TO CAPTURE WORD SIMILARITIES
 
Implementation of Semantic Analysis Using Domain Ontology
Implementation of Semantic Analysis Using Domain OntologyImplementation of Semantic Analysis Using Domain Ontology
Implementation of Semantic Analysis Using Domain Ontology
 
Doc format.
Doc format.Doc format.
Doc format.
 
A semi-supervised method for efficient construction of statistical spoken lan...
A semi-supervised method for efficient construction of statistical spoken lan...A semi-supervised method for efficient construction of statistical spoken lan...
A semi-supervised method for efficient construction of statistical spoken lan...
 
Ijarcet vol-2-issue-2-323-329
Ijarcet vol-2-issue-2-323-329Ijarcet vol-2-issue-2-323-329
Ijarcet vol-2-issue-2-323-329
 
HMM BASED POS TAGGER FOR HINDI
HMM BASED POS TAGGER FOR HINDIHMM BASED POS TAGGER FOR HINDI
HMM BASED POS TAGGER FOR HINDI
 
Verb based manipuri sentiment analysis
Verb based manipuri sentiment analysisVerb based manipuri sentiment analysis
Verb based manipuri sentiment analysis
 
DEVELOPMENT OF ARABIC NOUN PHRASE EXTRACTOR (ANPE)
DEVELOPMENT OF ARABIC NOUN PHRASE EXTRACTOR (ANPE)DEVELOPMENT OF ARABIC NOUN PHRASE EXTRACTOR (ANPE)
DEVELOPMENT OF ARABIC NOUN PHRASE EXTRACTOR (ANPE)
 
Enhancing the Performance of Sentiment Analysis Supervised Learning Using Sen...
Enhancing the Performance of Sentiment Analysis Supervised Learning Using Sen...Enhancing the Performance of Sentiment Analysis Supervised Learning Using Sen...
Enhancing the Performance of Sentiment Analysis Supervised Learning Using Sen...
 
ENHANCING THE PERFORMANCE OF SENTIMENT ANALYSIS SUPERVISED LEARNING USING SEN...
ENHANCING THE PERFORMANCE OF SENTIMENT ANALYSIS SUPERVISED LEARNING USING SEN...ENHANCING THE PERFORMANCE OF SENTIMENT ANALYSIS SUPERVISED LEARNING USING SEN...
ENHANCING THE PERFORMANCE OF SENTIMENT ANALYSIS SUPERVISED LEARNING USING SEN...
 
IDENTIFYING THE SEMANTIC RELATIONS ON UNSTRUCTURED DATA
IDENTIFYING THE SEMANTIC RELATIONS ON UNSTRUCTURED DATAIDENTIFYING THE SEMANTIC RELATIONS ON UNSTRUCTURED DATA
IDENTIFYING THE SEMANTIC RELATIONS ON UNSTRUCTURED DATA
 
Identifying the semantic relations on
Identifying the semantic relations onIdentifying the semantic relations on
Identifying the semantic relations on
 
Sanskrit parser Project Report
Sanskrit parser Project ReportSanskrit parser Project Report
Sanskrit parser Project Report
 

Último

Call Girls In Bellandur ☎ 7737669865 🥵 Book Your One night Stand
Call Girls In Bellandur ☎ 7737669865 🥵 Book Your One night StandCall Girls In Bellandur ☎ 7737669865 🥵 Book Your One night Stand
Call Girls In Bellandur ☎ 7737669865 🥵 Book Your One night Stand
amitlee9823
 
Call Girls Jalahalli Just Call 👗 7737669865 👗 Top Class Call Girl Service Ban...
Call Girls Jalahalli Just Call 👗 7737669865 👗 Top Class Call Girl Service Ban...Call Girls Jalahalli Just Call 👗 7737669865 👗 Top Class Call Girl Service Ban...
Call Girls Jalahalli Just Call 👗 7737669865 👗 Top Class Call Girl Service Ban...
amitlee9823
 
Vip Mumbai Call Girls Marol Naka Call On 9920725232 With Body to body massage...
Vip Mumbai Call Girls Marol Naka Call On 9920725232 With Body to body massage...Vip Mumbai Call Girls Marol Naka Call On 9920725232 With Body to body massage...
Vip Mumbai Call Girls Marol Naka Call On 9920725232 With Body to body massage...
amitlee9823
 
Call Girls Indiranagar Just Call 👗 7737669865 👗 Top Class Call Girl Service B...
Call Girls Indiranagar Just Call 👗 7737669865 👗 Top Class Call Girl Service B...Call Girls Indiranagar Just Call 👗 7737669865 👗 Top Class Call Girl Service B...
Call Girls Indiranagar Just Call 👗 7737669865 👗 Top Class Call Girl Service B...
amitlee9823
 
Vip Mumbai Call Girls Thane West Call On 9920725232 With Body to body massage...
Vip Mumbai Call Girls Thane West Call On 9920725232 With Body to body massage...Vip Mumbai Call Girls Thane West Call On 9920725232 With Body to body massage...
Vip Mumbai Call Girls Thane West Call On 9920725232 With Body to body massage...
amitlee9823
 
Call Girls Bannerghatta Road Just Call 👗 7737669865 👗 Top Class Call Girl Ser...
Call Girls Bannerghatta Road Just Call 👗 7737669865 👗 Top Class Call Girl Ser...Call Girls Bannerghatta Road Just Call 👗 7737669865 👗 Top Class Call Girl Ser...
Call Girls Bannerghatta Road Just Call 👗 7737669865 👗 Top Class Call Girl Ser...
amitlee9823
 
➥🔝 7737669865 🔝▻ Bangalore Call-girls in Women Seeking Men 🔝Bangalore🔝 Esc...
➥🔝 7737669865 🔝▻ Bangalore Call-girls in Women Seeking Men  🔝Bangalore🔝   Esc...➥🔝 7737669865 🔝▻ Bangalore Call-girls in Women Seeking Men  🔝Bangalore🔝   Esc...
➥🔝 7737669865 🔝▻ Bangalore Call-girls in Women Seeking Men 🔝Bangalore🔝 Esc...
amitlee9823
 
Call Girls Begur Just Call 👗 7737669865 👗 Top Class Call Girl Service Bangalore
Call Girls Begur Just Call 👗 7737669865 👗 Top Class Call Girl Service BangaloreCall Girls Begur Just Call 👗 7737669865 👗 Top Class Call Girl Service Bangalore
Call Girls Begur Just Call 👗 7737669865 👗 Top Class Call Girl Service Bangalore
amitlee9823
 
Al Barsha Escorts $#$ O565212860 $#$ Escort Service In Al Barsha
Al Barsha Escorts $#$ O565212860 $#$ Escort Service In Al BarshaAl Barsha Escorts $#$ O565212860 $#$ Escort Service In Al Barsha
Al Barsha Escorts $#$ O565212860 $#$ Escort Service In Al Barsha
AroojKhan71
 
Probability Grade 10 Third Quarter Lessons
Probability Grade 10 Third Quarter LessonsProbability Grade 10 Third Quarter Lessons
Probability Grade 10 Third Quarter Lessons
JoseMangaJr1
 

Último (20)

Call Girls In Bellandur ☎ 7737669865 🥵 Book Your One night Stand
Call Girls In Bellandur ☎ 7737669865 🥵 Book Your One night StandCall Girls In Bellandur ☎ 7737669865 🥵 Book Your One night Stand
Call Girls In Bellandur ☎ 7737669865 🥵 Book Your One night Stand
 
Sampling (random) method and Non random.ppt
Sampling (random) method and Non random.pptSampling (random) method and Non random.ppt
Sampling (random) method and Non random.ppt
 
Call Girls Jalahalli Just Call 👗 7737669865 👗 Top Class Call Girl Service Ban...
Call Girls Jalahalli Just Call 👗 7737669865 👗 Top Class Call Girl Service Ban...Call Girls Jalahalli Just Call 👗 7737669865 👗 Top Class Call Girl Service Ban...
Call Girls Jalahalli Just Call 👗 7737669865 👗 Top Class Call Girl Service Ban...
 
Vip Mumbai Call Girls Marol Naka Call On 9920725232 With Body to body massage...
Vip Mumbai Call Girls Marol Naka Call On 9920725232 With Body to body massage...Vip Mumbai Call Girls Marol Naka Call On 9920725232 With Body to body massage...
Vip Mumbai Call Girls Marol Naka Call On 9920725232 With Body to body massage...
 
(NEHA) Call Girls Katra Call Now 8617697112 Katra Escorts 24x7
(NEHA) Call Girls Katra Call Now 8617697112 Katra Escorts 24x7(NEHA) Call Girls Katra Call Now 8617697112 Katra Escorts 24x7
(NEHA) Call Girls Katra Call Now 8617697112 Katra Escorts 24x7
 
Digital Advertising Lecture for Advanced Digital & Social Media Strategy at U...
Digital Advertising Lecture for Advanced Digital & Social Media Strategy at U...Digital Advertising Lecture for Advanced Digital & Social Media Strategy at U...
Digital Advertising Lecture for Advanced Digital & Social Media Strategy at U...
 
Call me @ 9892124323 Cheap Rate Call Girls in Vashi with Real Photo 100% Secure
Call me @ 9892124323  Cheap Rate Call Girls in Vashi with Real Photo 100% SecureCall me @ 9892124323  Cheap Rate Call Girls in Vashi with Real Photo 100% Secure
Call me @ 9892124323 Cheap Rate Call Girls in Vashi with Real Photo 100% Secure
 
Halmar dropshipping via API with DroFx
Halmar  dropshipping  via API with DroFxHalmar  dropshipping  via API with DroFx
Halmar dropshipping via API with DroFx
 
Thane Call Girls 7091864438 Call Girls in Thane Escort service book now -
Thane Call Girls 7091864438 Call Girls in Thane Escort service book now -Thane Call Girls 7091864438 Call Girls in Thane Escort service book now -
Thane Call Girls 7091864438 Call Girls in Thane Escort service book now -
 
Call Girls Indiranagar Just Call 👗 7737669865 👗 Top Class Call Girl Service B...
Call Girls Indiranagar Just Call 👗 7737669865 👗 Top Class Call Girl Service B...Call Girls Indiranagar Just Call 👗 7737669865 👗 Top Class Call Girl Service B...
Call Girls Indiranagar Just Call 👗 7737669865 👗 Top Class Call Girl Service B...
 
Vip Mumbai Call Girls Thane West Call On 9920725232 With Body to body massage...
Vip Mumbai Call Girls Thane West Call On 9920725232 With Body to body massage...Vip Mumbai Call Girls Thane West Call On 9920725232 With Body to body massage...
Vip Mumbai Call Girls Thane West Call On 9920725232 With Body to body massage...
 
Midocean dropshipping via API with DroFx
Midocean dropshipping via API with DroFxMidocean dropshipping via API with DroFx
Midocean dropshipping via API with DroFx
 
Call Girls Bannerghatta Road Just Call 👗 7737669865 👗 Top Class Call Girl Ser...
Call Girls Bannerghatta Road Just Call 👗 7737669865 👗 Top Class Call Girl Ser...Call Girls Bannerghatta Road Just Call 👗 7737669865 👗 Top Class Call Girl Ser...
Call Girls Bannerghatta Road Just Call 👗 7737669865 👗 Top Class Call Girl Ser...
 
➥🔝 7737669865 🔝▻ Bangalore Call-girls in Women Seeking Men 🔝Bangalore🔝 Esc...
➥🔝 7737669865 🔝▻ Bangalore Call-girls in Women Seeking Men  🔝Bangalore🔝   Esc...➥🔝 7737669865 🔝▻ Bangalore Call-girls in Women Seeking Men  🔝Bangalore🔝   Esc...
➥🔝 7737669865 🔝▻ Bangalore Call-girls in Women Seeking Men 🔝Bangalore🔝 Esc...
 
Call Girls Begur Just Call 👗 7737669865 👗 Top Class Call Girl Service Bangalore
Call Girls Begur Just Call 👗 7737669865 👗 Top Class Call Girl Service BangaloreCall Girls Begur Just Call 👗 7737669865 👗 Top Class Call Girl Service Bangalore
Call Girls Begur Just Call 👗 7737669865 👗 Top Class Call Girl Service Bangalore
 
Generative AI on Enterprise Cloud with NiFi and Milvus
Generative AI on Enterprise Cloud with NiFi and MilvusGenerative AI on Enterprise Cloud with NiFi and Milvus
Generative AI on Enterprise Cloud with NiFi and Milvus
 
Al Barsha Escorts $#$ O565212860 $#$ Escort Service In Al Barsha
Al Barsha Escorts $#$ O565212860 $#$ Escort Service In Al BarshaAl Barsha Escorts $#$ O565212860 $#$ Escort Service In Al Barsha
Al Barsha Escorts $#$ O565212860 $#$ Escort Service In Al Barsha
 
5CL-ADBA,5cladba, Chinese supplier, safety is guaranteed
5CL-ADBA,5cladba, Chinese supplier, safety is guaranteed5CL-ADBA,5cladba, Chinese supplier, safety is guaranteed
5CL-ADBA,5cladba, Chinese supplier, safety is guaranteed
 
Call Girls in Sarai Kale Khan Delhi 💯 Call Us 🔝9205541914 🔝( Delhi) Escorts S...
Call Girls in Sarai Kale Khan Delhi 💯 Call Us 🔝9205541914 🔝( Delhi) Escorts S...Call Girls in Sarai Kale Khan Delhi 💯 Call Us 🔝9205541914 🔝( Delhi) Escorts S...
Call Girls in Sarai Kale Khan Delhi 💯 Call Us 🔝9205541914 🔝( Delhi) Escorts S...
 
Probability Grade 10 Third Quarter Lessons
Probability Grade 10 Third Quarter LessonsProbability Grade 10 Third Quarter Lessons
Probability Grade 10 Third Quarter Lessons
 

Towards Building Semantic Role Labeler for Indian Languages

  • 1. Towards Building Semantic Role Labeler for Indian Languages Maaz Anwar Nomani♣ and Dipti Misra Sharma♣ FC Kohli Center on Intelligent Systems (KCIS), IIIT-Hyderabad, India♣ maaz.anwar@research.iiit.ac.in, dipti@iiit.ac.in, Abstract We present a statistical system for identifying the semantic relationships or semantic roles for two major Indian Languages, Hindi and Urdu. Given an input sentence and a predicate/verb, the system first identifies the arguments pertaining to that verb and then classifies it into one of the semantic labels which can either be a DOER, THEME, LOCATIVE, CAUSE, PURPOSE etc. The system is based on 2 statistical classifiers trained on roughly 130,000 words for Urdu and 100,000 words for Hindi that were hand-annotated with semantic roles under the PropBank project for these two languages. Our system achieves an accuracy of 86% in identifying the arguments of a verb for Hindi and 75% for Urdu. At the subsequent task of classifying the constituents into their semantic roles, the Hindi system achieved 58% precision and 42% recall whereas Urdu system performed better and achieved 83% precision and 80% recall. Our study also allowed us to compare the usefulness of different linguistic features and feature combinations in the semantic role labeling task. We also examine the use of statistical syntactic parsing as feature in the role labeling task. Keywords: PropBank, Semantic Role labeling, Treebank, Dependency relations 1. Introduction We introduce a statistical semantic role labeler for Hindi and Urdu, two major Indian languages. A Semantic Role labeler (henceforth, SRL) automatically marks the argu- ments/valency of a predicate in a sentence. The proposed system is based on supervised machine learning approach on Hindi and Urdu PropBanks which are being built for these languages. The approach is a 2-stage architecture in which, first the arguments pertaining to a predicate in a sentence are iden- tified by the system and then those identified arguments are classified into one of the PropBank semantic labels. Our system uses a basic Logistic Regression machine learn- ing algorithm (Pedregosa et al., 2011) for identifying the predicates and Support Vector Machines (Pedregosa et al., 2011) to classify the arguments of a predicate into seman- tic labels. We have used 10 linguistic features, 6 of which have been derived from literature and 4 are novel features. Among the new features, we have used the karaka rela- tions/dependency relations (described later) which turns out to be the most discriminative feature among all. We have also experimented with the automatic parses which are used as features and derived from state-of-art Hindi and Urdu parsers. To the best of our knowledge, this is the first such attempt of building a semantic role labeler for any In- dian language. This paper is arranged as follows: Section 2 gives a brief introduction about Semantic Parsing. In Section 3, we give the data statistics of the language resources used. Section 4 talks about related work in semantic role labeling. In sec- tion 5, we briefly describe the Hindi and Urdu Propbanks and Treebanks which are being built for these two Indian languages. Section 6 presents the Semantic Role labeler in detail which includes our approach, its architecture, classi- fiers and features used. We also describe the baseline fea- tures, their impact on Indian language datasets and new fea- tures incorporated in the systems. In section 7, we describe in detail the different experiments done and show their re- sults. Section 8 throws light on the impact of using auto- matic parses as features in the semantic role labeling task. In section 9, we thoroughly put the error analysis forward along with some corpus examples showing erroneous role labeling and discuss possible reasons for it. We conclude the paper in section 10. 2. Semantic Parsing Semantic analysis of natural text and languages has always been an intriguing research area in NLP. Automatic seman- tic analysis of texts thus becomes a challenging as well as interesting task which includes tasks like Semantic Pars- ing, Shallow Semantic Parsing and Semantic Role label- ing. Automatic and accurate tools that can predict naturally occurring arguments for a given predicate in a predicate- argument structure of a sentence can be efficiently used for Semantic Parsing, Summarization, Information Extraction tasks, wider NLP areas like Machine Translation and Ques- tion Answering and modern NLP challenges like parsing code-mixed data. Semantic Parsing is essentially the research investigation of identifying WHO did WHAT to WHOM, WHERE, HOW, WHY and WHEN etc. in a sentence and adding a layer of semantic annotation in a sentence produces such a struc- ture. Two major Indian languages, Hindi and Urdu are hav- ing such a resource in the form of PropBank (also called Proposition Banks) which establishes a layer of semantic representation in a Treebank already annotated with Depen- dency labels. Proposition Bank (Kingsbury and Palmer, 2003) is a corpus in which the arguments of each verb predicate (simple or complex) are marked with their semantic roles. 3. Data Table 1 shows the data statistics of the language resources. We have used a part of these language resources for SRL task. 4588
  • 2. In the experiments reported here, we first performed the ar- gument identification task which is a binary classification problem (either ‘argument’ or ‘not an argument’) by tak- ing 130,000 tokens as training data and 30,000 as test data from Urdu PropBank. We took around 100,000 tokens as training data from Hindi PropBank and 20,000 as test data. Subsequently, we trained a multi-class classifier, SVM on the same training set, using a well defined feature-set de- scribed later. Language Resource Tokens Sentences Predicates pbrel Urdu Treebank ˜200,000 ˜8,000 - - Hindi Treebank ˜350,000 ˜14,000 - - Urdu Propbank ˜180,000 ˜7,000 ˜2,200 ˜7,000 Hindi Propbank ˜300,000 ˜11,000 ˜3,200 ˜12,000 Table 1: Hindi and Urdu resources statistics 4. Related Work The last decade has seen an exhaustive research in se- mantic parsing for different languages. Gildea and Juraf- sky (2002) started out the work on semantic role label- ing on 2001 release of English Propbank. Surdeanu et al. (2003) build a decision tree classifier for predicting the se- mantic labels. Gildea and Hockenmaier (2003) uses fea- tures from Combinatory Categorial Grammar (CCG) which is a form of dependency grammar. Chen and Rambow (2003) used a decision tree classifier and additional syntac- tic and semantic representations extracted from Tree Ad- joining Grammar (TAG). Swier and Stevenson (2004) talks of a novel bootstrapping algorithm for identifying semantic labels. (Cohn and Blunsom, 2005) applied conditional ran- dom fields (CRFs) to the semantic role labeling task. Xue and Palmer (2004) experimented with different linguistic features related to role labeling task. Xue and Palmer (2005) built a role labeler for Chinese by showing that verb classes, induced from the predicate- argument information in the frame files helps in semantic role labeling. Johansson and Nugues (2006) came up with a FrameNet-based semantic role labeling system for Swedish text. Toutanova et al. (2005) used joint learning amd joint modeling of argument frames of verbs to improve the over- all accuracy of semantic role labeler. 5. The PropBank and The Treebank Indian Languages are very morphologically rich languages and Treebanks for Hindi and Urdu are available which in- clude syntactico-semantic information in the form of de- pendency annotations as well as lexical semantic infor- mation in the form of predicate-argument structures, thus forming PropBanks. Hindi and Urdu PropBanks are part of a multi-dimensional and multi-layered resource creation effort for the Hindi-Urdu language (Bhatt et al., 2009). The Hindi Dependency Treebank (HDT) and Urdu Depen- dency Treebank (UDT) are built following the CPG (Com- putational Paninian Grammar) framework (Begum et al., 2008). PropBank establishes a layer of semantic represen- tation in the Treebank which is already annotated with De- pendency labels or phrase structures (as in Penn Treebank for English). Capturing the semantics through predicate- argument structure involves quite a few challenges pecu- liar to each predicate type because the syntactic notions in which the verb’s arguments and adjuncts are realized can vary based on the senses. Table 2 shows the PropBank la- bels for Hindi and Urdu and these labels are also used for building the Semantic Role labeler. Propbank labels or semantic labels are closely associated with the karaka relations (Vaidya et al., 2011) (described later) in their structure though the former are defined on a verb-by-verb basis. Label Description ARG0 Agent, Experiencer or doer ARG1 Patient or Theme ARG2 Beneficiary ARG3 Instrument ARG2-ATR Attribute or Quality ARG2-LOC Physical Location ARG2-GOL Goal ARG2-SOU Source ARGM-PRX noun-verb construc- tion ARGM-ADV Adverb ARGM-DIR Direction ARGM-EXT Extent or Compari- sion ARGM-MNR Manner ARGM-PRP Purpose ARGM-DIS Discourse ARGM-LOC Abstract Location ARGM-MNS Means ARGM-NEG Negation ARGM-TMP Time ARGM-CAU Cause or Reason Table 2: Hindi and Urdu PropBank labels with defini- tions 6. Semantic Role labeler This section gives a thorough analysis on our approach to- wards role labeling task. At first, one can comprehend the semantic role labeling task as a simple straight forward multi-class classification problem in which the semantic la- bels are classified directly without identifying them. But such a simple approach will not work due to various rea- sons. (Xue and Palmer, 2004) observes that for a given predicate, many constituents in a syntactic tree are not its semantic argument. So, non-argument count overwhelms the argu- ment count for the given predicate and classifiers will not be efficient in predicting the right argument or in classify- ing them. This problem was encountered in both Hindi and Urdu data-sets as well. Therefore, we adopted a 2-stage approach of first iden- tifying the arguments and subsequently classifying the identified semantic arguments into one of the semantic labels. The second approach of directly classifyng the arguments was also experimented by us and we show by 4589
  • 3. statistics that such a forthright approach will not work for Hindi and Urdu as well. • Approach I • Argument Identification - This is the sub-task of iden- tifying a constituent in a dependency tree which repre- sents the argument for the predicate in the argument- predicate structure in a Hindi or Urdu PropBank sen- tence. The identification is done by a binary classifier (‘argument’ and ‘not an argument’). The classifier has to predict whether the constituent in the tree is an ar- gument of that verb or not. • Argument Classification - After identification of argu- ments is done, the identified arguments pertaining to that verb are assigned appropriate semantic labels by a multi-class classifier which is trained on multiple se- mantic labels present in the training data. We have used Gold Parses as well as Automatic Parses from Hindi and Urdu parsers as features in this task. • Approach II • Argument Identification and Argument Classification - We performed direct classification of arguments in Hindi and Urdu data and found that the results were not effective. We found that due to overwhelming number of non-arguments present in the corpus, the classifier failed to classify the arguments correctly. For example, in Urdu SRL system, the arguments for which gold parses were used as feature, were getting classified with a precision of only 20.98% and non- arguments were getting classified with a precision of 79.71%. For automatic parses, arguments got classi- fied with a precision of 17.12% and non-arguments got classified with 82.87% precision. 6.1. SRL Architecture Figure 1 is the flow diagram of Semantic Role Labeler for Hindi and Urdu. It is a 2-stage architecture in which at first the arguments of a predicate in a sentence are identified by a binary classifier and then in the subsequent stage only these identified arguments are classified as one of the semantic labels by a multi-class classifier. 6.2. Features: A critical look Hindi and Urdu are free-word order languages i.e. the word order is somewhat flexible in these languages. For this rea- son, they are sometimes called as SOV languages (subject, object, verb). Hindi and Urdu employs postpositions and not prepositions (as in English), the resulting word order of postpositional phrases can be the reverse of prepositional phrases in English. We decided to use linguistic features in our Semantic Role Labeler based on our intuition that they might help the labeler in taking decisions to identify and classify the semantic arguments of a predicate in the sentence. Filtering of pbrole sentences Argument Identifier Identified Arguments projected on Test Data Only Arguments Argument Classifier Classified/Predicted labels Projected to Test Data of PropBank Test Data containing automatic SR labels Training (Binary) Training (Multi-Class) Extracted Figure 1: SRL Architecture 6.2.1. Baseline Features The task of identifying and classifying the arguments of a verb is done at chunk/phrase level. Most of the previous works done in Semantic Parsing have shown that ‘head’ of the chunk is a very critical feature because chunks/phrases with certain head words are likely to be arguments of the verb and also certain types of arguments. This implies that head word of a chunk is a very discriminative feature for both the tasks. For our Indian languages SRL, we have used a combination of baseline features introduced by (Gildea and Jurafsky, 2002) and (Xue and Palmer, 2004) for English because we wanted to see the impact of these standard features in context of the Indian languages and how these features perform w.r.t free word order languages. Baseline Features for Indian Languages Predicate - predicate itself Head word - syntactic head of the chunk/phrase Head-word POS - Its POS category Phrase Type - syntactic category (NP, VP etc. of the phrase) Predicate + Phrase Type - combinational feature Predicate + Head word - combinational feature Table 3: Baseline feature template 6.2.2. New Features After analyzing the PropBanks of both the languages, we came up with certain new features for which we had an intuition that they will contribute significantly towards this task. These are discussed below: 4590
  • 4. • Dependency/karaka relation - We experimented with the Dependency labels or karaka relations present in the Hindi and Urdu PropBanks. A sentence is com- posed of a primary modified which is the root in the dependency tree of the sentence and this modified is occasionally the main verb of the sentence. The words modifying this verb participate in the action specified by the verb. The participant relation of the verb with these words/elements are called karaka relations. There are in total 43 karaka relations for Hindi and Urdu as specified by (Begum et al., 2008). Table 4 shows some of the major karaka or dependency re- lations. These dependency relations helps to identify the syntactico-semantic relations in a sentence. Since, PropBank labels illustrates the semantic relations be- tween the constituents, we had this intuition that using the karaka relations as features in our classifier to pre- dict the arguments of a verb will have an impact on our system performance. Also, Vaidya et al. (2011) has shown that there is a correlation between dependency and predicate argument structures for Hindi. Since Hindi and Urdu are linguistically similar languages, we extracted the dependency/karaka relations of both the language Treebanks and used them as feature in isolation as well as in the system. Dependency Labels Description k1 karta - doer/agent/subject k2 karma - object/patient k1s noun complement k4 sampradana - recipient k3 karana - instrument k4a experiencer k2p goal, destination k5 apadana - source k7 location elsewhere k7p location in space k7t time adv adverbs rh reason rd direction rt purpose Table 4: Major karaka relations or Dependency labels of Hindi and Urdu Treebank From Tables 5 and 6, it is clear that dependency rela- tions are indeed the most discriminative feature among all the features used by showing high precision and re- call for both languages. The high percentage of preci- sion and recall in the Argument Classification task can be accounted from the fact that dependency structure and predicate-argument structure are somewhat simi- lar structures and dependency labels and propbank la- bels are similar in orientation. Therefore, using de- pendency structures as features to predict and classify the arguments of a predicate of Hindi and Urdu is very influential. • Named Entities (NEs) - Named Entities also contribute towards the identification task but it doesn’t con- tributes heavily towards the classification task. The reason being NEs are generally arguments of a predi- cate. This feature has been used in the past by (Prad- han et al., 2004) for English but we have used it dif- ferently by combining it with the POS category of the NE. As shown is Tables 5 and 6, NEs have a decent precision of 63% and 61% for Urdu and Hindi respec- tively when used in isolation for the argument identifi- cation task. But it is not able to discriminate between the class of semantic labels to which a identified argu- ment belong. • Head + Phrase type - A combination of head word of the chunk/phrase and syntactic category of the phrase. • Head POS + Phrase type - A combination of the POS category of the head word of the chunk/phrase and syntactic category of the phrase. Abbreviations Used: ‘P’=Precision, ‘R’=Recall, ‘PT’=Phrase Type, ‘HW’=Head Word, ‘POS’=Part- of-speech’ Feature ¯ Feature wise performance (Urdu) ¯ Argument Id. ¯ Argument Classification ¯ P R f-score P R f-score Predicate 51 72 60 00 01 00 Head-word 71 73 72 25 14 18 Head-POS 51 72 60 06 18 09 PhraseType 51 72 60 08 08 08 HeadPOS-PT 61 72 66 06 18 09 Head-PT 71 74 72 25 14 18 Predicate-HW 72 74 73 01 01 01 Predicate-PT 64 69 66 02 00 00 NEs 63 65 64 23 18 20 Dependency 78 79 78 87 85 86 Table 5: Individual feature performance (Urdu) • We also experimented with certain other features like chunk boundary and sentence boundary of the sentences. We had an intuition that these boundaries will help the classifier learn more effectively and overlap of the 2 sentences will be minimized. But it turned out that both the features had a major negative impact on the overall results of the system, so we scrapped these two features. The reason for this dip can be accounted from the fact that chunk boundary and sentence boundary are merely integers for the classifier as there are ‘n’ numbers for ‘n’ sentences/chunks which lack any significant information. 6.2.3. Classifiers used We used a basic Logistic Regression classifier and trained it on Hindi and Urdu data-sets thus building a binary classifier which identifies arguments as ‘Arguments’ or ‘Not-an-Argument’. For multi-class classification in argument classification task, we implemented LinearSVC class of SVM (Pedregosa et al., 2011) for performing the ‘one-vs-the-rest’ multi-class strategy. All the parametres 4591
  • 5. of Linear SVC were set to default at the time of training of Hindi and Urdu data-sets. Feature ¯ Feature wise performance (Hindi) ¯ Argument Id. ¯ Argument Classification ¯ P R f-score P R f-score Predicate 41 64 50 41 61 49 Head-word 70 70 70 20 04 07 Head-POS 34 58 43 07 08 07 PhraseType 34 58 43 05 07 06 HeadPOS-PT 34 58 43 14 13 13 Head-PT 67 66 66 10 04 06 Predicate-HW 65 65 65 10 04 06 Predicate-PT 41 64 50 04 06 05 NEs 61 64 62 20 16 17 Dependency 88 87 87 52 49 50 Table 6: Individual feature performance (Hindi) 7. Experiments and Results In this section, we show the results and extent of the baseline features (standard features) in Semantic Role Labeling in the experiments on the Indian data sets. We present results of baseline features for both the task and add new features subsequently in next step. We have laid out the experiments in a way that each experiment conveys the extent of usefulness and information gain of the feature in both the tasks. For each experiment, the settings of the SVM and features used in the previous experiment were retained. The baseline features performs well on the argument identification task of both languages with a precision of 71% and 67% for Hindi and Urdu respectively. The classification task of the identified arguments is not profited much by the baseline features for both the languages as can be seen from Tables 7 and 8. Feature ¯ Hindi system Feature Template ¯ Argument Id. ¯ Argument Classification ¯ P R f-score P R f-score Baseline 71 71 66 29 14 15 +Dependency 86 86 86 56 38 41 +Head-PT 85 85 86 56 38 41 +HeadPOS-PT 84 84 84 57 40 42 +NEs 85 84 84 58 42 49 Table 7: Hindi system performance As expected, the best result for both the tasks is obtained on incorporating dependency relations information, which further strengthens our conjecture on the importance of de- pendency labels in semantic role labeling. All 4 experi- ments shows the effectiveness of the features for Hindi and Urdu data sets. By incorporating NEs in the system, there is an increase in the identification task of Urdu and marginal increase in Hindi. The classification task is also benefitted by a marginal precision increment of 1% in both the data sets and recall of 2% in Hindi. However, ‘Head + Phrase Type’ and ‘HeadPOS + Phrase Type’ features had a negative impact on the performance of classification task for both the sets. The identification task was benefitted by every new feature in Hindi with dependency relations contributing the most to the final result of the system. Feature ¯ Urdu system Feature Template ¯ Argument Id. ¯ Argument Classification ¯ P R f-score P R f-score Baseline 67 70 68 16 14 15 +Head-PT 65 69 67 15 12 13 +HeadPOS-PT 66 69 67 15 12 13 +Dependency 72 74 73 82 80 81 +NEs 75 71 73 83 80 81 Table 8: Urdu system performance 8. Using Automatic Parses So far, we have shown and discussed results using hand- corrected parses of the Hindi and Urdu PropBank. How- ever, if we are building a NLP application, the SRL will have to extract features from automatic parses of the sen- tences. Also, after getting the insights of the influence of hand-corrected dependency relations, we wanted to analyse the same using automatic parses and evaluate our semantic role labeler on it. These automatic parses are generated by the parser of respective language. To come up with such an evaluation, we used the Malt Parser (Nivre et al., 2006) on the same test data described above for both languages. Malt Parser requires data in CONLL format. By using state-of-art Urdu Parser (Bhat et al., 2012), we extracted the automatic parses from the Urdu test data and used them as feature for our identification and classification tasks. Parsing accuracy is measured by LAS, UAS and LA which stands for Labeled Attachment score, Unlabeled Attach- ment score and Label Accuracy respectively. LAS is the percentage of words that are assigned the correct head and dependency label; UAS is the percentage of words with the correct head, and the label accuracy (LA) is the percentage of words with the correct dependency label. Measure Percentage (%) LAS 76.61 UAS 88.45 LA 80.03 Table 9: Parsing performance of the Urdu dependency parser By adding automatic parses to the baseline features of Urdu system, we observed a marginal increase in the argument identification task and comparable improvement in classi- fication task in Table 10. The decent jump in the precision of the classification task while using automatic parses as compared to using gold parses can be attributed from the fact that gold parses (Urdu PropBank) have been manually validated by annotators whereas Urdu parser tends to give certain incorrect parses as can be seen from Table 9. 4592
  • 6. Feature ¯ Automatic Parses (Urdu) ¯ Argument Id. Argument Classification P R f-score P R f-score Baseline 67 70 68 16 14 15 +Automatic Parses 68 66 67 62 59 60 Table 10: Urdu system with Automatic Parses The same process was repeated by using the state-of-art Hindi Parser (Kosaraju et al., 2010) on Hindi test data. Measure Percentage (%) LAS 88.19 UAS 94.00 LA 89.77 Table 11: Parsing performance of the Hindi dependency parser The usage of automatic parses of Hindi have similar impact on the Hindi SRL system. Table 12 shows the performance devaluation when automatically generated parses are used. Feature ¯ Automatic Parses (Hindi) ¯ Argument Id. ¯ Argument Classification ¯ P R f-score P R f-score Baseline 71 71 66 29 14 15 +Automatic Parses 73 74 73 41 28 33 Table 12: Hindi system with Automatic Parses 9. Error Analysis and Discussion In this section, we present a thorough error analysis of the Semantic Role Labeling task for Hindi and Urdu by taking a sub-set of test data for error analysis. Tables 14 and 15 are the confusion matrices for Hindi and Urdu SRL respec- tively. As Hindi and Urdu have postpositional phrases, the postpositions or case markers play a very significant role in deciding the semantic role for an argument. These postpo- sitions may not be helpful cues in identifying the argument but they are when it comes to classifying the arguments. There are 6 pre-dominant case markers in Hindu and Urdu viz. ‘ne’, ‘ko’, ‘mein’, ‘se’, ‘ki’, ‘par’. Table 13 gives the definition of Hindi and Urdu case markers. These case markers provide important information about the token pre- ceding it and the nature of the chunk/phrase in which they are present. For e.g. in both Hindi and Urdu, the case clitic ‘ne’ is often indicative of an agent, and as ‘ARG0’ is mostly associated with ‘agentivity’, this can sometimes provide a clue about ARG0. Same is the case with ‘mein’ which most of the times gives the ‘locative’ information in the chunk/phrase. Some post-positions are very ambigous in nature and here is when error in classifying the argu- ments starts. We show in the examples below how ‘ko’ and ‘mein’ post-position can take multiple types of semantic la- bels and therefore the classifier is not very discriminative during multi-class classification. Urdu Case marker Meaning ne Ergative mein Locative par Locative ko Dative/Accusative se Instrumental/Ablative ki Genetive Table 13: Hindi and Urdu Case markers As can be seen from confusion matrix of Hindi, the pair of labels in which the Hindi SRL is having the maximum confusion is ‘ARG1’ and ‘ARG0’. ‘ARG0’ label is given to those arguments of verb which shows agentivity or vola- tionality to do work whereas ARG1 is the theme or object arguments. Another frequent confusion which the Hindi SRL has is between ‘ArgM-LOC’ and ‘ARG0’. The con- fusion between ‘ARGM-LOC’ and ‘ARG2-LOC’ can be attributed to the fact that both the labels depicts locative information though ARGM-LOC shows abstract location whereas ARG2-LOC has physical location information. Arg0 Arg1 Arg2 Arg2LOC Arg2SOU Arg2ATR ArgADV ArgMLOC ArgDIR Arg0 23 03 02 Arg1 21 21 04 05 02 03 Arg2 01 01 01 Arg2LOC 02 02 Arg2SOU 02 Arg2ATR 03 01 07 ArgMADV 01 ArgMLOC 05 01 04 01 16 ArgMDIR 02 Table 14: Confusion Matrix of Propbank labels for Hindi SRL For Urdu, similar confusion is observed between ‘ARG0’- ‘ARG1’ and ‘ARGM-LOC’-‘ARG2-LOC’ pairs. Such con- fusions can be reduced by coming up with linguistic fea- tures which are aimed at solving ARG0 and ARG1 con- fusion because it is the most frequent erroneous pair for Indian languages SRL. These errors also occur due to the presence of ambigous case markers with the arguments of a verb and therefore the classifier is unable to discriminate between the two labels. Another interesting confusion pair is ARG1:ARG2-ATR. There are instances in the training set in which the argu- ments of a verb are annotated with ‘ARG1’ or ‘ARG2-ATR’ based on the context and sentence structure. Arg0 Arg1 Arg2 Arg2LOC Arg2SOU Arg2ATR ArgADV ArgMLOC ArgDIR Arg0 45 02 02 01 Arg1 08 48 03 05 02 04 Arg2 01 07 01 Arg2LOC 10 01 01 03 Arg2SOU 04 02 01 01 04 Arg2ATR 01 ArgMADV 01 12 ArgMLOC 01 36 ArgMDIR 03 Table 15: Confusion Matrix of Propbank labels for Urdu SRL We extracted certain sentences from the test set of Hindi and Urdu and analyzed how our SRL is performing in com- parison with the gold test data. These sentences are shown below: 4593
  • 7. (1) (iss this dauran) period [gold:ARGM-TMP; SRL: ARGM-TMP] his (uski) daughter (putri DAT ko) man [gold: ARG1; SRL: ARG1] roam (yuwak) live [gold: ARG0; SRL: ARG0] be-PAST. (ghumata) [predicate] (rehta tha). ‘During this period, the man used to roam around his daughter.’ HINDI TEST (2) (idhar) here [gold: ARGM-DIR; SRL: ARGM-DIS] Assam (Assam LOC mein) BJP [gold: ARGM-LOC; SRL: ARGM-LOC] main (BJP) party [gold: ARG1; SRL: ARG0] manner (mukhya LOC party emerge ke) live (roop be-PRS. mein) [gold: ARGM-MNR; SRL: ARGM-MNR] (ubhar) [predicate] (rahi) (hai). URDU TEST ‘Here, in Assam, BJP is emerging as the main party.’ (3) (Jwala Jwala ne) ERG (bataya) said (ki) that (central central market market ke comparison muqaable LOC mein) here [gold: ARGM-EXT; SRL: ARGM-LOC] rate (yahan) economical [gold: ARG2-LOC; SRL: ARG2:LOC] live (rate) be-PRS [gold: ARG1; SRL: ARG1] (munaasib) [gold: ARGM-PRX; SRL: ARGM:PRX] (rehta hai) [predicate]. HINDI TEST ‘Jwala said that rate here is economical as compared to central market.’ (4) (Barack Barack Obama Obama ko) DAT [gold: ARG0-GOL; SRL: ARG2] feel (ehsaas) be-PRS [gold: ARG1; SRL: ARG2-ATR] thathis (hai) party (ki) election (unki) LOC (party) 2009 [gold: ARG0; SRL: ARG1] of (intekhabaat miracle mein) not [gold: ARGM-LOC; SRL: ARGM-LOC] see (2009 get. ka) (karishma) (nahi) (dikha) [predicate] (paaegi). URDU TEST ‘Barack Obama feels that his party will not be able to repeat the miracle of 2009 in the election.’ In sentence 3, ‘central market ke muqaable mein’ is ‘ARGM-EXT’ in gold test data as it conveys the mean- ing of ‘comparison’ with some entity. But Hindi SRL gives it ARGM-LOC label. One possible reason for this may be that ‘mein’ case marker generally occurs more fre- quently with ARGM-LOC or ARG2-LOC label in the loca- tive sense in the training data as compared to comparison. In sentence 4, ‘Barack Obama ko’ gets ARG2 label by Urdu SRL instead of ARG0-GOL because of the same reasons cited in the above example. Here, ‘ko’ case marker more frequently occurs with ARG2 label i.e. beneficiary in the training data as compared to experiencer here. 10. Conclusion In this work, we have presented a supervised semantic role labeler for Hindi and Urdu. We have used linguistic fea- tures like predicate, head, head-POS, phrase type etc. and combinations of certain features. In all, 10 features have been used to guide the classifiers in predicting, identifying and classifiyng the arguments of a verb. We further used the gold parses (dependency parses) as a feature which had a solid impact on the overall accu- racy of the systems. Subsequently, we extracted the au- tomatic parses by using state-of-art Hindi and Urdu parsers and used them as features in our SRL. We believe that de- pendency relations are the best features for classifying the semantic arguments of a predicate because of their close proximity with semantic labels. 11. Acknowledgements We would like to thank Mr. Mujadia Vandan for the discus- sions which we had with him to get a better insight of the task. 12. Bibliographical References Begum, R., Husain, S., Dhwaj, A., Sharma, D. M., Bai, L., and Sangal, R. (2008). Dependency annotation scheme for indian languages. In IJCNLP, pages 721–726. Cite- seer. Bhat, R. A., Jain, S., and Sharma, D. M. (2012). Exper- iments on dependency parsing of urdu. In The 11th In- ternational Workshop on Treebanks and Linguistic The- ories. Bhatt, R., Narasimhan, B., Palmer, M., Rambow, O., Sharma, D. M., and Xia, F. (2009). A multi-representational and multi-layered treebank for hindi/urdu. In Proceedings of the Third Linguistic Anno- tation Workshop, pages 186–189. Association for Com- putational Linguistics. 4594
  • 8. Chen, J. and Rambow, O. (2003). Use of deep linguistic features for the recognition and labeling of semantic ar- guments. In Proceedings of the 2003 conference on Em- pirical methods in natural language processing, pages 41–48. Association for Computational Linguistics. Cohn, T. and Blunsom, P. (2005). Semantic role labelling with tree conditional random fields. In Proceedings of the Ninth Conference on Computational Natural Lan- guage Learning, pages 169–172. Association for Com- putational Linguistics. Gildea, D. and Hockenmaier, J. (2003). Identifying seman- tic roles using combinatory categorial grammar. In Pro- ceedings of the 2003 conference on Empirical methods in natural language processing, pages 57–64. Associa- tion for Computational Linguistics. Gildea, D. and Jurafsky, D. (2002). Automatic labeling of semantic roles. Computational linguistics, 28(3):245– 288. Johansson, R. and Nugues, P. (2006). A framenet-based semantic role labeler for swedish. In Proceedings of the COLING/ACL on Main conference poster sessions, pages 436–443. Association for Computational Linguis- tics. Kingsbury, P. and Palmer, M. (2003). Propbank: the next level of treebank. In Proceedings of Treebanks and lexi- cal Theories, volume 3. Citeseer. Kosaraju, P., Kesidi, S. R., Ainavolu, V. B. R., and Kukkadapu, P. (2010). Experiments on indian language dependency parsing. Proceedings of the ICON10 NLP Tools Contest: Indian Language Dependency Parsing. Nivre, J., Hall, J., and Nilsson, J. (2006). Maltparser: A data-driven parser-generator for dependency parsing. In Proceedings of LREC, volume 6, pages 2216–2219. Pedregosa, F., Varoquaux, G., Gramfort, A., Michel, V., Thirion, B., Grisel, O., Blondel, M., Prettenhofer, P., Weiss, R., Dubourg, V., et al. (2011). Scikit-learn: Ma- chine learning in python. The Journal of Machine Learn- ing Research, 12:2825–2830. Pradhan, S. S., Ward, W., Hacioglu, K., Martin, J. H., and Jurafsky, D. (2004). Shallow semantic parsing using support vector machines. In HLT-NAACL, pages 233– 240. Surdeanu, M., Harabagiu, S., Williams, J., and Aarseth, P. (2003). Using predicate-argument structures for in- formation extraction. In Proceedings of the 41st Annual Meeting on Association for Computational Linguistics- Volume 1, pages 8–15. Association for Computational Linguistics. Swier, R. S. and Stevenson, S. (2004). Unsupervised se- mantic role labelling. In Proceedings of EMNLP, vol- ume 95, page 102. Toutanova, K., Haghighi, A., and Manning, C. D. (2005). Joint learning improves semantic role labeling. In Pro- ceedings of the 43rd Annual Meeting on Association for Computational Linguistics, pages 589–596. Association for Computational Linguistics. Vaidya, A., Choi, J. D., Palmer, M., and Narasimhan, B. (2011). Analysis of the hindi proposition bank using de- pendency structure. In Proceedings of the 5th Linguis- tic Annotation Workshop, pages 21–29. Association for Computational Linguistics. Xue, N. and Palmer, M. (2004). Calibrating features for semantic role labeling. In EMNLP, pages 88–94. Xue, N. and Palmer, M. (2005). Automatic semantic role labeling for chinese verbs. In IJCAI, volume 5, pages 1160–1165. Citeseer. 4595