This document discusses using Watson Natural Language Classifier to tag ideas from an M-CAFE dataset with topics. 106 ideas were split randomly into a training set of 86 tagged ideas and a test set of 20 untagged ideas. Watson was trained on the training set and tested on the test set, achieving an accuracy of 80%. Examples of correctly and incorrectly classified ideas are provided. Questions are also included about how the classifier is trained and whether unspervised classification is available.
8. Idea Topic
Slower pace. Lectures
Add Lecture overview Resources
I want more practice with Relational Algebra and eventually SQL. Homework
The last few lectures have been very mathematically precise in
notation which can make it a bit tricky to wrap your head around.
Specific questions/examples (like what might be on hw) would be
great to help us make sure we understand it moving forward.
Lectures
The project seems a little stop and go. We haven't been able to
work on it for a week or so but I feel like we'll soon be expected
to do a bunch of work for DP2. It would be helpful if we could
have the tools to have a more constant level of work on the
project.
Projects
Please try and post the labs earlier so that we can get a head
start reading and understanding them.
Labs
Homework 2 only has database questions, maybe put some
connectives?
Homework
Incorporate a short question and answer period midway of
lecture to assess participating students' understanding of the
lecture/topics being presented.
Lectures
Examples of ideas which are correctly classified:
10. Idea True Tag Pred Tag Confidence
3. I would like have some
implantation problems
using SQL
Homework New Topics
New Topics:
0.803;
Homework:
0.076
4. More hands on
experiences on Databases
Homework New Topics
New Topics:
0.786;
Homework:
0.117
Misclassifications Contd…
• The true tag is among the top two tags suggested by the
classifier.
• Misclassification occurs when an idea is arbitrarily tagged
or with lack of context.