Call Girls In Hsr Layout ☎ 7737669865 🥵 Book Your One night Stand
Data, Responsibly: The Next Decade of Data Science
1. Data, Responsibly:
The Next Decade of Data Science
Bill Howe, PhD
Associate Professor, Information School
Associate Director, eScience Institute
Adjunct Associate Professor, Computer Science & Engineering
University of Washington
2. My goals this evening
• Describe emerging topics in data science research
and practice around a technical interpretation of ethics
• Describe some specific thrusts we are pursuing
• Encourage you to get involved
11/10/2016 Data, Responsibly / SciTech NW 2
3. How much time do you spend “handling
data” as opposed to “doing science”?
Mode answer: “90%”
11/10/2016 Bill Howe, UW 3
4. 1) Upload data “as is”
Cloud-hosted; no need to
install or design a database;
no pre-defined schema
2) Analyze data with SQL
Right in your browser,
writing queries on top of
queries on top of queries ...
SELECT hit, COUNT(*)
FROM tigrfam_surface
GROUP BY hit
ORDER BY cnt DESC
3) Share the results
Click on the science question,
see the SQL that answers it
6. Making it easier to do data science
• SQLShare: Easier to use a database
• Myria: Easier to use a bunch of different
systems at once, at scale
• Worked great in the physical sciences
• But some collaborators weren’t that excited…
November 10, 2016 6
7. 11/10/2016 Bill Howe, UW 7
Data Science Kickoff Session:
137 posters from 30+ departments and units
9. 9
• Pursue transformative interdisciplinary urban research
• Facilitate translation from UW to .gov stakeholders
• Position Seattle/UW as a leader in applied urban research
• 80+ faculty from 20+ departments around campus
10. 10
Assessing Community Well-Being
Third-Place Technologies
Optimization of King County Metro Paratransit
Computer Science & Engineering
Predictors of Permanent Housing for Homeless Families
Bill and Melinda Gates Foundation
Open Sidewalk Graph for Accessible Trip Planning
Computer Science & Engineering
Inaugural 2015 program:
16 spots
140 applicants
…from 20+ departments
11. 11
Mining Online Data to Detect Unsafe Food Products
Elaine Nsoesie, Institute for Health Metrics and Evaluation
ORCA data for improved transit system planning and operation
Washington State Transportation Center (TRAC)
Global Open Sidewalks: Creating a shared open data layer
Taskar Center for Accessible Technology
CrowdSensing Census: A tool for estimating poverty
Bell Labs, Nokia
2016 program:
16 spots
190 applicants
New in 2016: An explicit emphasis on data ethics
15. First decade of Data Science research and practice:
What can we do with massive, noisy, heterogeneous datasets?
Next decade of Data Science research and practice:
What should we do with massive, noisy, heterogeneous datasets?
The way I think about this…..(1)
16. The way I think about this…. (2)
Decisions are based on two sources of information:
1. Past examples
e.g., “prior arrests tend to increase likelihood of future arrests”
2. Societal constraints
e.g., “we must avoid racial discrimination”
11/10/2016 Data, Responsibly / SciTech NW 16
We’ve become very good at automating the use of past examples
We’ve only just started to think about incorporating societal constraints
17. The way I think about this… (3)
How do we apply societal constraints to algorithmic
decision-making?
Option 1: Keep a human in the loop
Ex: EU General Data Protection Regulation requires that a
human be involved in legally binding algorithmic decision-making
Ex: Wisconsin Supreme Court says a human must review
algorithmic decisions made by recidivism models
Option 2: Build them into the algorithms themselves
I’ll talk about some approaches for this
11/10/2016 Data, Responsibly / SciTech NW 17
18. The way I think about this…(4)
On transparency vs. accountability:
• For human decision-making, sometimes explanations are
required, improving transparency
– Supreme court decisions
– Employee reprimands/termination
• But when transparency is difficult, accountability takes over
– medical emergencies, business decisions
• As we shift decisions to algorithms, we lose both
transparency AND accountability
• “The buck stops where?”
11/10/2016 Data, Responsibly / SciTech NW 18
19. Some Facets of “Data, Responsibly”
• Privacy
• Fairness
• Transparency
• Reproducibility
• Ethics
11/10/2016 Data, Responsibly / SciTech NW 19
I won’t be talking about this
I’ll give a taste of the work here
I won’t be talking about this
Towards automatic scientific claim-checking
Vignette on teaching data ethics
21. 21
Ex: Staples online pricing
Reasoning: Offer deals to people that live near competitors’ stores
Effect: lower prices offered to buyers who live in more affluent
neighborhoods
22. 22
[Latanya Sweeney; CACM 2013]
Racially identifying names trigger
ads suggestive of an arrest record
slide adapted from Stoyanovich, Miklau
24. 24
The Special Committee on Criminal Justice Reform's
hearing of reducing the pre-trial jail population.
Technical.ly, September 2016
Philadelphia is grappling with the prospect of a racist computer algorithm
Any background signal in the
data of institutional racism is
amplified by the algorithm
operationalized by the algorithm
legitimized by the algorithm
“Should I be afraid of risk assessment tools?”
“No, you gotta tell me a lot more about yourself.
At what age were you first arrested?
What is the date of your most recent crime?”
“And what’s the culture of policing in the
neighborhood in which I grew up in?”
25. 26
Towards a precise characterization of fairness…
Positive Outcomes Negative Outcomes
offered employment denied employment
accepted to school rejected from school
offered a loan denied a loan
offered a discount not offered a discount
Label outcomes to individuals as positive or negative
Fairness is concerned with how outcomes are
assigned to a population
slide adapted from Stoyanovich, Miklau
26. 27
Statistical parity
race
black
white
⊕ ⊖
⊖
⊕
⊕⊖
⊖
⊖
⊕
40% of the whole population
positive
outcomes
⊖
Statistical parity
demographics of the individuals receiving any outcome are the same
as demographics of the underlying population
20%
of black
60%
of white
slide adapted from Stoyanovich, Miklau
27. 28
First attempt: Ignore sensitive information
zip code
10025 10027
race
black
white
20%
of black
60%
of white
⊕
⊖
⊖⊖
⊕
⊕ ⊖
⊖
⊖
⊕
positive
outcomes
Removing race from the vendor’s assignment process
does not prevent discrimination
Assessing disparate impact
Discrimination is assessed by the effect on the protected sub-
population, not by the input or by the process that lead to the effect.
slide adapted from Stoyanovich, Miklau
28. 29
More directly: Impose statistical parity
credit score
good bad
black
white
⊕
⊖
⊖
⊖
⊕
⊕ ⊖
⊖
⊖
⊕
positive
outcomes
40%
of black
40%
of white
race positive outcome: offered a loan
Tradeoff between (perceived) accuracy and fairness;
may be contrary to the goals of the vendor
slide adapted from Stoyanovich, Miklau
29. 30
A systems approach:
FairTest: fairness test suite for data analysis apps
• Tests for unintentional discrimination according to several
representative discrimination measures.
• Automates search for context-specific associations between
protected variables and application outputs
• Report findings, ranked by association strength and
affected population size
[F. Tramèr et al., arXiv:1510.02377 (2015)]
30. As a corporation, should I care?
Compliance
Jacobson, Scientific American, 2013
Customer
Retention
Employee
Retention
Eichler, Hiffington Post, 2012
CNET, May 2016
32. Science is a complete mess
• Reproducibility
– Begley & Ellis, Nature 2012: 6 out of 53 cancer studies reproducible
– Only about half of psychology 100 studies had effect sizes that approximated
the original result (Science, 2015)
– Ioannidis 2005: Why most public research findings are false
– Reinhart & Rogoff: global economic policy based on spreadsheet fuck ups
11/10/2016 Bill Howe, UW 33
40. Vision: Validate scientific claims automatically
– Check for manipulation (manipulated images, Benford’s Law)
– Extract claims from papers
– Check claims against the authors’ data
– Check claims against related data sets
– Automatic meta-analysis across the literature + public datasets
• First steps
– Automatic curation: Validate and attach metadata to public datasets
– Longitudinal analysis of the visual literature
11/10/2016 Data, Responsibly / SciTech NW 41
42. 11/10/2016 Bill Howe, UW 43
Microarray samples submitted to the Gene Expression Omnibus
Curation is fast becoming the
bottleneck to data sharing
Maxim
Gretchkin
Hoifung
Poon
45. color = labels supplied
as metadata
clusters = 1st two PCA
dimensions on the
gene expression data
itself
Can we use curate algorithmically?
Maxim
Gretchkin
Hoifung
Poon
The expression data
and the text labels
appear to disagree
47. Deep Curation Maxim
Gretchkin
Hoifung
Poon
Distant supervision and co-learning between text-
based classified and expression-based classifier: Both
models improve by training on each others’ results.
Free-text classifier
Expression classifier
48. Deep Curation:
Our stuff wins, with no training data
Maxim
Gretchkin
Hoifung
Poon
state of the art
our reimplementation
of the state of the art
our dueling
pianos NN
amount of training data used
50. Alcohol Study, Barrow Alaska, 1979
Native leaders and city officials,
worried about drinking and associated
violence in their community invited a
group of sociology researchers to
assess the problem and work with
them to devise solutions.
51. Methods
• 10% representative sample
(N=88) of everyone over the age
of 15 using a 1972 demographic
survey
• Interviewed on attitudes and
values about use of alcohol
• Obtained psychological histories
including drinking behavior
• Given the Michigan Alcoholism
Screening Test (Seltzer, 1971)
• Asked to draw a picture of a
person
– Used to determine cultural identity
52. Results announced unilaterally and publicly
At the conclusion of the study researchers formulated a report entitled “The
Inupiat, Economics and Alcohol on the Alaskan North Slope” which was released
simultaneously at a press release and to the Barrow community. The press
release was picked up by the New York Times, who ran a front page story
entitled Alcohol Plagues Eskimos
53. The results of the Barrow Alcohol Study in Alaska were revealed in the context of a
press conference that was held far from the Native village, and without the
presence, much less the knowledge or consent, of any community member who
might have been able to present any context concerning the socioeconomic
conditions of the village. Study results suggested that nearly all adults in the
community were alcoholics. In addition to the shame felt by community members,
the town’s Standard and Poor bond rating suffered as a result, which in turn
decreased the tribe’s ability to secure funding for much needed projects.
Backlash
54. Methodological Problems
“The authors once again met with the Barrow Technical
Advisory Group, who stated their concern that only Natives
were studied, and that outsiders in town had not been
included.”
“The estimates of the frequency of intoxication based on
association with the probability of being detained were
termed "ludicrous, both logically and statistically.””
Edward F. Foulks, M.D., Misalliances In The Barrow Alcohol Study
55. Ethical Problems
• Participants were not in control of their data nor
the context in which they were presented.
• Easy to demonstrate specific, significant harms:
– Social: Stigmatization
– Financial: Bond rating lowered
• Important: Nothing to do with individual privacy
– No PII revealed at any point, to anyone
– No violations of best practices in data handling
– But even those who did not participate in the study
incurred harm
56. Two Topics
• Social Component: Codes of Conduct
• Technical Component: Managing Sensitive
Data
57. Ethical principles vs. ethical rules
• In the Barrow example, ethical rules
were generally followed
• But ethical principles were violated: The
researchers appear to have placed their
own interests ahead of those of the
research subjects, the client, and society
58. Principles: Codes of Conduct
• American Statistical Association
– http://www.amstat.org/committees/ethics/
• Certified Analytics Professional
– https://www.certifiedanalytics.org/ethics.php
• Data Science Association
– http://www.datascienceassn.org/code-of-
conduct.html
59. Recap
• There’s a sea change underway in how we will teach
and practice data science
• No longer only about what can be done, but about
what should be done
• This is not just a policy/behavior/culture issue – there
are technical problems to solve
• If you’re not thinking about this stuff, you will be facing
retention issues and compliance issues very soon
– Witness privacy, which is a few years ahead
Notas do Editor
In each of these fields, my research interests are driven by this question. We like to ask researchers how much time they spend "handling data" as opposed to "doing science”. They say things like 90%, and they don’t even blink.
So my overarching research question is “How can we reduce this "data science overhead.”?
One effort in this area was to develop SQLShare, where we emphasize a very simple workflow:
1) you can upload data “as is” from spreadsheets or anything. There’s no need to install software or design a schema.
2) Then you can immediately begin writing queries, right in your browser, and define queries on top of queries on top of queries to express even complex workflows.
3) Then you can share the results online: Your colleagues can browse science questions in English and see the SQL that answers it.
----
Key ideas to get data in: a) Use the cloud to avoid having to install and run a databaseb) Give up on the schema -- just throw your data in "as is" and do "lazy integration.”
c) Use some magic to automate parsing, integration, recommendations, and more.
Key ideas to get data out:
a) Associate science questions (in English) with each SQL query -- makes them easy to understand and easy to find.b) Saving and reusing queries is a first class requirement. Given examples, it's easy to modify it into an "adjacent" query.
c) Expose the whole system through a REST API to make it easy to bring new client applications online.
You’ll see these applications in the demo
Solutions are emerging, powered by the open data movement.
Socrata, a local Seattle company, has built a very successful business of helping cities jailbreak their data, and are now engaged in climbing the application stack to support analytics and visualization.
Essentially every url of the form data.yourcity.gov is powered by Socrata’s technology
Data, People, and Infrastructure
Following a 2014 report entitled “Big Data: Seizing Opportunities, Preserving Values”
On which projects should we engage?
How can we ensure fairness, accountability, and transparency for algorithmic decision-making?
How do we ensure privacy?
How do we avoid junk science?
From WLJ article:
A Wall Street Journal investigation found that the Staples Inc. website displays different prices to people after estimating their locations. More than that, Staples appeared to consider the person's distance from a rival brick-and-mortar store, either OfficeMax Inc. or Office Depot Inc. If rival stores were within 20 miles or so, Staples.com usually showed a discounted price.
In what appears to be an unintended side effect of Staples' pricing methods—likely a function of retail competition with its rivals—the Journal's testing also showed that areas that tended to see the discounted prices had a higher average income than areas that tended to see higher prices.
Users and regulators must be able to understand how raw data was selected, and what operations were performed during analysis
Users want to control what is recorded about them and how that information is used
Users must be able to access their own information and correct any errors (US Fair Credit Reporting Act)
Transparency facilitates accountability - verifying that a service performs as it should, and that data is used according to contract
LSI-R model
25 states use it
Most for targeted programs
Idaho and Colorado use this for sentencing
“As a Black male,” Cobb asked Penn statistician and resident expert Richard Berk, “should I be afraid of risk assessment tools?”
“No,” Berk said, without skipping a beat. “You gotta tell me a lot more about yourself. … At what age were you first arrested? What is the date of your most recent crime? What are you charged with?”
Cobb interjected: “And what’s the culture of policing in the neighborhood in which I grew up in?”
(emphasis mine)
That's exactly the point (and to Michael -- this is what I was arguing about with the guy from Comcast): a little bit of institutional racism has a triple effect:
a) institutional racism is amplified by the algorithm (a small signal can now dominate the model)
b) institutional racism is operationalized by the algorithm (it's far easier now to make impactful decisions based on bad data)
c) institutional racism is legitimized by the algorithm (so that everyone thinks "it's just data" and actively defends the algorithm's assumed objectivity, even when the racist results are staring you right in the face. This vigorous defense doesn't happen when a human is shown to be correlating their decisions perfectly with race.)
More formal definition; cite laws, 80% rule, statistical significance.
Redundant encoding was actually used to disguise discrimination: red lining
The point here is that avoiding discrimination may be directly in conflict with the vendor (or classifier’s) utility goals. It is therefore a constraint on the assignment of outcomes that must be balanced with the vendor’s interests.
In reality, it is impossible to predict loan payback accurately, so we use past information. Then the question arises whether that past information is biased.
You can’t roll the dice a bunch of times then yell “Yahtzee!”
Google knowledge graph
Specialized Ontologies
Google knowledge graph
Specialized Ontologies
Google knowledge graph
Specialized Ontologies
"HeLa", "K562", "MCF-7" and "brain tumor”
PCA on expression values
Google knowledge graph – common knowledge, high redundancy, possibly crowdsourcing (visual: question answering via Google)
Text features:
presence of ontology terms
sibling of ontology term
Expression features
Native leaders and city officials in Barrow, Alaska, worried about drinking and associated violence and accidental deaths in their community invited a group of sociology researchers to assess the problem and work with them to devise solutions. At the conclusion of the study researchers formulated a report entitled “The Inupiat, Economics and Alcohol on the Alaskan North Slope” which was released simultaneously at a press release and to the Barrow community. The press release was picked up by the New York Times, who ran a front page story entitled Alcohol Plagues
Responsibility to which parties?
* Society
* Employers and Clients
* Colleagues
* Research Subjects
ASA:
Professionalism
Responsibilities to Funders, Clients, Employers
Responsibilities in Publications and Testimony
Responsibilities to Research Subjects
Responsibilities to Research Team Colleagues
Responsibilities to Other Statisticians or Statistical Practitioners
Responsibilities Regarding Allegations of Misconduct
Responsibilities of Employers
Code of Conduct: Rules
Competence
Do what you client asks, unless violates law
Communication with clients
Confidential information
Conflicts of interest
Rule 7: More on conflicts of interest and confidentiality
Rule 8: Scientific integrity
+++ Interesting: If a data scientist reasonably believes a client is misusing data science to communicate a false reality or promote an illusion of understanding, the data scientist shall take reasonable remedial measures, including disclosure to the client, and including, if necessary, disclosure to the proper authorities. The data scientist shall take reasonable measures to persuade the client to use data science appropriately.
Rule 9: Misconduct (follow the rules)