Abstract: The rise of data science has been largely fueled by the promise of changing the business landscape - enhancing one's competitive advantage, increasing business optimization and efficiency, and ultimately delivering a better bottom-line. This promise reaches across sectors as machine learning methods are getting better, data access continues to grow, and computation power is easily accessible. However, because the practice of doing data science can be expensive, there is a danger that this so-called promise of data science may only be available to the most well-resourced organizations with sophisticated data capabilities and staff. For the past five years, DataKind has been working to ensure social change organizations too have access to data science, teaming them up with data scientists to build machine learning and artificial intelligence solutions that aim to reduce human suffering. In doing so, DataKind has learned what it takes to apply data science in the social sector and the many applications it has for creating positive change in the world. This session presents DataKind projects showcasing the wide range of applications for ML/AI for social good. From using satellite imagery and remote sensing techniques to detect wheat farm boundaries to protect livelihoods in Ethiopia, to leveraging NLP to automate the time consuming process of synthesizing findings from academic studies to inform conservation efforts and to classifying text records to better understand human rights conditions across the world to using machine learning to reduce traffic fatalities in U.S. cities, learn about some of the latest breakthroughs and findings in the data science for social good space and learn how you can get involved
3. Our Purpose: We amplify the impact of social organizations.
Our Customer: Social organizations that have a clear theory of
change for reducing human suffering
Our Competitive Advantage: Our network of pro bono data scientists
Our Product: Data science services, i.e. predictive analytics, machine
learning, AI
Our Style: Human-centered design, jargon-free, accessible
What We Do
17. ICAAD: International Center for Advocates
Against Discrimination
Non-profit organization that combats structural discrimination through
monitoring global trends, fostering research and designing interventions
Promote religious freedom in France
Combat gender-based violence in the Pacific Islands
Better documentation of hate crimes
Mapping discrimination
23. What was the DataCorps problem?
How do we leverage these UPR records to better
understand human rights conditions across the world?
24. Labeling with Sustainable Development Goals
Adopted in 2015, the SDGs are a set of seventeen aspirational goals that all UN
member states are committed to achieve, covering a broad range of human
rights and development issues
Successor to the Millennium Development Goals
26. Deliverables
1) Build an MVP algorithm that systematically classifies Universal Periodic
Review (UPR) records using Sustainable Development Goals
2) Using the results from the algorithm, create a dashboard that visualizes
global patterns of discrimination
These two tools will enable ICAAD to better allocate their resources towards
the most important human rights interventions, as well as better disseminate
their findings to other related organizations.
28. UPR Data
Source Number of Records Number of Labels
ICAAD labeled 1247 2351
DataKind labeled 349 628
All-organic, self-harvested and hand-labeled...
29. Data Prep
1) Each UPR = (very short) document
2) Clean, tokenize and create (1,2)-grams
3) Create term-document matrix
4) Feed bag-of-words matrix into ML model
5) ML model = two step “ensemble”
30. Machine Learning Layer: Multi-Label SVM
Support Vector Machine
Linear (no kernel)
Loss function: Squared hinge
Penalty type: L1
Regularization constant: 2.0
31. Keyword Lookup Layer
If UPR text contains the word “corruption” → SDG #16
If UPR text contains the word “HIV” or “AIDS” → SDG #3
If UPR text contains the word “ICRMW” → SDG #10
And so on...
32. Final Ensemble Model: CV Metrics
ML Layer ML + Keyword Lookup
Precision 0.827 0.772
Recall 0.758 0.848
F1-Score 0.787 0.802
ML Layer by itself does very well, but by adding the Keyword layer,
we can sacrifice a little bit of precision for a large gain in recall,
and get overall better performance.
34. The Aftermath Part 1
● Proof of concept algorithm delivered last October
● Demonstrated and implicated among various project
partners
35. The Aftermath Part 2
● Team from Xerox brought in to build v2 of algorithm
○ Main SDG category contains 169 additional sub-goals
○ ICAAD wants to classify UPR records using these sub-goals
○ Army of volunteer lawyers doing a lot of manual labeling
● Something concrete by next summer!
36. What We Learned (Parting Shots)
1. Easier = better
2. Small data is hard
3. Simple Boolean logic works surprisingly well
4. Data scientists are paid (and sometimes not) to do the
dirty work
37. DataCorps Team
Ben Cohen: Software Engineer @ Warby Parker
Rebecca Wei: PhD Student @ Northwestern
Karry Lu: Senior Data Scientist @ Plated
46. Respondents from UK
conservation community
indicate desire to use
evidence but:
Lacked a support framework
to quickly sort and evaluate
evidence
Experience-
based
Evidence-
based
modified from Pullin et al. 2004
Evidence gap
48. Need for knowledge on
effectiveness
Evidence-based decision making:
Using findings to inform actions
Desired outcomes achieved
Research project
Communicate findings
Monitor and evaluate
progress and outcomes
Identify
knowledge gaps
Synthesize
knowledge gluts
Determine
indicators
Adjust actions
RESEARCHERS
PRACTITIONERS
Theory of change
49. The need
Practitioners need standardized
storage and access to research
insights from academic and grey
literature for evidence-based
decision making
Researchers need a framework to
follow to create these resources
Best
Science
Expert
Opinion
Society’s needs
and preferences
Evidence based
Decision-making
52. Problem #1: interactivity
Thorn, Jessica PR, et al. "What evidence exists for the
effectiveness of on-farm conservation land
management strategies for preserving ecosystem
services in developing countries? A systematic map."
Environmental Evidence 5.1 (2016): 13.
63. System 1: relevance ranking
• Citations are ranked by expected relevance
depending on the availability and number of
user-labeled examples
– 1st uses search terms from review planning:
computes the amount of overlap between those
terms and citations' title + abstract + keywords
– 2nd after enough examples have been labeled,
uses distributional word vectors (word2vec) as
features for a support vector classifier that predicts
inclusion or exclusion; use confidence of that
classification as expected relevance
• Citations are randomly sampled each time, to avoid
hasty generalization
65. System 2: extraction and tagging
• A better methodology might be to use the training
data to find sentences in the document that might
indicate a label. (provide provenance)
• We can train the system to over-predict (predict
sentences from a large number of the labels), so
that the system can focus on recall, while human
annotators can focus on precision
• For locations we can use a "Named Entity
Recognition" system to find mentioned locations
in the document, and suggest these as labels
• For other metadata, we can train a model which
predicts the relevance of sentences to a label
• We show the sentences that best predict labels to
the user, who can then use that information to
pick the correct labels
68. Interaction with the data portal
•Output CSV file with
individual citations and
factor tags
•CSV ingested by receiving
system
http://natureandpeopleevidence.org/
69. Measuring evidence synthesis and
dissemination on the “T” impact model
Sector (diffuse) impact as
measured by
• Access and operability
• Common vs. uncommon solution
• Dissemination framework
Organization (deep) impact as
measured by
• Operational efficiency
• Increased productivity
• Expanded service