1. NTNU@MediaEval 2011: Social Event Detection Task (SED) Massimiliano Ruocco, Heri Ramampiaro Data and Information Management Group Department Of Computer and Information Science Norwegian University of Science and Technology [email_address] MediaEval 2011 Workshop - Pisa
17. Thanks for the attention Questions? http://www.idi.ntnu.no/~ruocco/
Notas do Editor
For each venue name in each different language the related spatial location a query is built. The tetxual terms of the venue name compose a boolean query by using the OR operator. A spatial and textual search is performed by the search engine for each input query related to each venue. A threshold on the search engine integrated score for the results occurrencies is used to select the results. The categorization is performed by using the three textual metadata surrounding every picture. Output is a set of flickr pictures grouped by venues which has SOCCER as topic.
Then the similarity between two clusters is based on the number of shared entity names.
In both refinement step is not in the process Giving a look over all the measures we can see a benefit using the refinement step and even a benefit using entity names instead of pure most frequent tags to refine the clusters. (observe the interesting increase of the recall measure) I would even remarks some point. Clustering evaluation is not an easy task since there are different metrics and different constraint can be verified. In particular two constraint that an evaluation metrics shoould evaluate are COMPLETENESS () and HOMOGEINITY (). For the first, Recall is a good indicator ad for the second recall and NMI are good indicator. With NMI it seems even to have increasing performarmce in term of completeness from the refinement by using top-100 tags and entity names
In both refinement step is not in the process Giving a look over all the measures we can see a benefit using the refinement step and even a benefit using entity names instead of pure most frequent tags to refine the clusters. (observe the interesting increase of the recall measure) I would even remarks some point. Clustering evaluation is not an easy task since there are different metrics and different constraint can be verified. In particular two constraint that an evaluation metrics shoould evaluate are COMPLETENESS () and HOMOGEINITY (). For the first, Recall is a good indicator ad for the second recall and NMI are good indicator. With NMI it seems even to have increasing performarmce in term of completeness from the refinement by using top-100 tags and entity names
In both refinement step is not in the process Giving a look over all the measures we can see a benefit using the refinement step and even a benefit using entity names instead of pure most frequent tags to refine the clusters. (observe the interesting increase of the recall measure) I would even remarks some point. Clustering evaluation is not an easy task since there are different metrics and different constraint can be verified. In particular two constraint that an evaluation metrics shoould evaluate are COMPLETENESS () and HOMOGEINITY (). For the first, Recall is a good indicator ad for the second recall and NMI are good indicator. With NMI it seems even to have increasing performarmce in term of completeness from the refinement by using top-100 tags and entity names