Ethics
Learning outcomes
These are the three learning outcomes for this lecture:
- Identify pros and cons of using given data to solve the desired problem
- Identify and defend privacy issues
- Eventually, do this without being told there is a privacy issue to identify
- Identify stakeholders, and values and interests at stake for each stakeholder
- Identify and discuss conflicts of values
- Identify modifications that would mitigate conflicts of values
Here is a diagram of the overall Systematic Software Program Design and Implementation Process that we generally follow when designing and implementing a software product:
We will revisit it multiple times this semester, especially the first three sections (Requirements, Design, and Implementation), and dive deeper into all five sections in CS 3100. These five phases are generally done in that order, though it is possible to revisit an earlier phase. Each phase also has an optional pathway to "Abandon Project" for various reasons.
Today, we will focus on the first phase, Requirements, specifically the stakeholder-value matrix.
Why Value-Sensitive Design?
Computer science is a group activity. In this course, we are teaching the vocabulary for moral language to be able to talk with your teammates and advocate for what you think is right.
We are also choosing to focus specifically on "Identifying and defending privacy issues" to debunk the common excuse of "I have nothing to hide" 1 using Value-Sensitive Design. We will be building stakeholder-value matrices, which help us realize which stakeholders might be put at risk because they "have something to hide."
Our steps for value-sensitive design happen at the beginning of the overall Systematic Program Design and Implementation Process. They help shape the core functionality of a program before we dive into the code and data.
Here are two examples that may help to illustrate the usefulness of the process:
-
Social media apps were built to connect communities and bring users closer together. But there are other stakeholders with conflicting values, namely, the social media platform company itself: they make more money when more users look at their content (and advertisements) for longer. Their financial values can lead them to optimize the platform to keep users looking at the screen for longer, which can conflict with the users' value of mental health.
-
reCAPTCHA and Duolingo were built to verify that users are human and to teach new languages, respectively. But, they both also use the data generated by users to train AI models.2 You may form your own opinions about your unpaid labor. But, similar to social media apps, Duolingo's financial values also compel them to make users view their app more frequently, leading them to prioritize viewing rather than learning.
Steps for Value-Sensitive Design
These are the steps you will take to demonstrate Value-Sensitive Design in the homework assignments for this course:
- Identify the stakeholders: identify people who are affected by the software in any way
- Identify the values: identify the values at stake for those stakeholders when considering the software
- Fill out a stakeholder-value matrix: create a table, where the columns are the values, and the rows are the stakeholders, and write in how each stakeholder's value relates to the software
- Identify and analyze conflicts in the matrix: for each conflict in the matrix, identify whether and how to address it
In the CS 2000, 2100, and 3100 courses you will demonstrate Value-Sensitive Design in the assignments with each of these steps. The next section will expand and demonstrate on each of these steps using a case study of a dating app.
Value-Sensitive Design Case study: Finder, a dating app
Let’s consider the example of “Finder,” a fictitious dating app (such as Hinge or Grindr). Let's go through the steps of the Value-Sensitive Design process above.
Step 1: Identify the stakeholders
Users are only a few of many people who will be affected by a proposed technology.
Finder’s users include nonbinary people, transgender people, people who date people of all genders, people whose sexual orientation is criminalized in their geographic location, people who are not legally allowed to be in their geographic location altogether, sex workers, survivors of intimate partner violence, people who are blind, people who don’t speak English, people whose names have apostrophes in them, and so many more.
There are many more stakeholders. Often, if a platform enforces rules for its users, then the content moderators who check for disallowed content are subject to viewing harmful content. They are stakeholders, along with the software engineers and all the other employees of the company which is building Finder.
If Finder’s financial model includes advertising, then the advertisers are key stakeholders, along with the other customers of the advertising companies. And let’s not forget the actors and influencers doing the advertising.
Governments and regulators have stakes involving following laws and regulations, and protecting the public.
Anyone who dates is a stakeholder, since dating apps have fundamentally changed the way we date. And in that vein, relationships formed with Finder will impact many more stakeholders, including friends and family.
There are undoubtedly stakeholders who are not listed, and more stakeholders who will get created3 over time.
Step 2: Identify the values
Values inform the design, development, and deployment of technology, whether or not the technologists acknowledge it. And, once implemented, technology has the ability to change our practices and values.
Privacy is a key value highlighted throughout this course. Many of the stakeholders listed as users will value privacy, whether it is to hide their identity from the government or hide their location from an abusive person. Safety can also be listed as a separate value, or as a subset of privacy.
Accessibility as a value will entail different things to different users. Nonbinary people may not have access to Finder if we don’t include them as a gender option. People who are blind or who don’t speak English may need us to design Finder so that screen readers and translation apps can work. Sex workers are often unfairly banned from dating apps, even if they are not doing sex work or soliciting clients through the dating app. And of course, we need to support a large variety of users’ names so that they can create profiles and access Finder.
Fairness will be valued if we prioritize it. We only want to take into account users’ personal preferences, and nothing else, when recommending people as potential dating partners.
Financial cost will be pertinent to advertisers and users, depending on Finder’s financial model. Reputation is heavily valued, too. And of course, many people value laws and regulations.
Step 3: Fill out a stakeholder-value matrix
Here, we set up a matrix where the columns represent the values, and the rows represent the stakeholders. The cells describe how each stakeholder interacts with each value.
Privacy | Accessibility | Fairness | Financial cost | Reputation | Laws and regulations | Hope of finding connection | |
---|---|---|---|---|---|---|---|
Nonbinary people | May want to keep gender identity hidden | Need nonbinary option for profile | Like all users, they should be recommended fairly | Like all users, they may want to keep costs reasonable | Generally want to exist legally | Like all users, they want to find connection | |
Transgender people | May want to keep gender identity hidden | Need to be allowed to “change” profile’s gender if wanted | |||||
People who date people of all genders | May want to keep their sexual orientation hidden | Need to be able to search dating profiles of all genders | |||||
People whose sexual orientation is criminalized in their geographic location | Likely want to keep their sexual orientation hidden from the government | Need to be able to search dating profiles of appropriate genders from their location | |||||
People who are not legally allowed to be in their geographic location | Likely want to keep their location hidden from the government | Need to be able to set up a profile without government identification | |||||
Sex workers | Likely want to keep their industry hidden from the government | Must be allowed on the platform | |||||
Survivors of intimate partner violence | Likely want to keep their location hidden from certain people of concern | Need 2FA or other methods to prevent account takeover | Hope we go above minimum regulations for accommodation and protection | ||||
People who are blind | May want to keep disability status hidden | Need to be able to use screen readers | |||||
People who don’t speak English | Like all users, they may want to avoid surveillance | Need to be able to use translation apps | |||||
People whose names have apostrophes | Like all users, they may want to avoid surveillance | Must be able to set up a profile with their name | |||||
Content moderators | Likely want to avoid leaking their identities to users | Need access to everything they need, and nothing more (no personal data or direct messages) | Need training to ensure consistency | Need to be well-compensated for experiencing harmful content | Want to uphold good reputation of Finder and its company | Must follow laws about allowed speech on platforms and personal data protection | |
Software engineers | Likely want to avoid leaking their identities to users | Keep costs low (API usage, cloud storage, etc.) | Want to design a product to help people find connection | ||||
Other employees of Finder’s company | Likely want to avoid leaking their identities to users | Company needs to make money to survive | |||||
Actors and influencers | Likely want to avoid leaking personal info to users | Should be shown to users fairly | Likely want good compensation | Want to be seen advertising good products on good platforms | |||
Advertisers | Probably have some data to share and some to keep private to avoid abuse | Want access to users’ contact information, but we shouldn’t give it (they’ll stop using Finder if they can directly email users) | Want to reach more audiences for less money | Want to be seen on good platforms | Must follow ad regulations | ||
Other customers of advertisers | Likely want to avoid leaking personal info to users | If the ads have special deals, then the other customers lose out on those deals | |||||
Governments and regulators | Likely want to avoid leaking their identities to users | Need access to everything they need, and nothing more | Need to collect accurate taxes | Must enforce the law | |||
General public | Likely want to avoid leaking personal info to users | Probably no access to the app unless they sign up as users | Competition drives costs up |
Step 4: Identify and analyze conflicts in the matrix
Some stakeholders’ values conflict.
For example, advertisers want to keep financial costs low, but that may push the financial cost onto the users or the company building Finder. Choosing the financial model will require choosing one over the other.
Of course, many users want to keep their identities or locations private from the government, but the government may want access to that information to enforce laws and protect the public. There may be mitigation strategies such as keeping all information private until regulators provide a subpoena, or only providing specific information to the government (not location or sexual orientation).
Fairness or accessibility can sometimes unfortunately require risking Finder’s reputation, depending on the audience giving reputation-related opinions, and how strongly they feel that others don’t belong. Also, ensuring fairness for Finder’s users can require collecting personal information, which can then compromise privacy, especially in case of a subpoena.
There is also often a conflict when criminalized populations simply want to exist legally and peacefully despite being targeted by governments.
Throughout the VSD Process: Bias and Unfairness
Bias and unfairness are baked into our society in a way that counteracting them is a challenging, continuous task which does not end with a VSD matrix. However, there are some concrete places we can keep in mind to mitigate bias and unfairness in our technology.
Bias and unfairness can enter our technology in a variety of ways. Here are some examples:
- The technology might rely on a dataset that doesn’t accurately reflect the statistics of the population it represents. For example, a dataset of “the general population” can forget to include women.
- The tools used to collect the data might be biased. For example, a survey might include multiple-choice questions that are impossible to answer correctly.
- The technologists building the tool may misunderstand the data. For example, they might not understand the context in which the data was collected, leading to dismissing important data points or selecting the wrong variables for use.
- The technologists building the tool may make mistakes, such as using a correlation between variables to imply causation. For example, if there are many reports of fraud in September, they may assume that the start of a new academic year causes students to commit fraud.
- The technology can be used for a different purpose than the one for which it was built. For example, a platform built for vital communication in vulnerable communities can also be used to spam and harass those vulnerable people.
Bias and unfairness can be created or amplified once the technology is implemented in a particular societal context.
- Technology reflects historical injustices as they unfold and compound. For example, word representations (the numbers used by computers to represent words in a natural language such as English) reflect the bias of text on which they are trained.4
- People using the technology have their own implicit biases, and technology can exacerbate the impact of those biases. Technology can also reinforce those stereotypes to those users.
- Technology can have disparate impact given the social context and features outside the model. For example, a tool that is very useful to one population may be harmful or inaccessible to another population.
- This unfairness is also compounded through feedback loops. For example, social media platforms often highlight posts which have already received a lot of positive attention, which in turn gives those posts even more positive attention, reinforcing society’s standards for which types of posts should receive positive attention.
Privacy
We have been using the term "privacy" throughout this lecture as one of the values in VSD. Privacy is "the ability to determine for ourselves when, how, and to what extent information about us is communicated to others" (Westin, 1967, as summarized in DeCew 2018). But, as Dr. Katie Creel puts it, "individual privacy often appears to be in conflict with the interests of society or the state: a balance must be struck between public and private interests. For example, people describe a 'tradeoff' between privacy and national security. If privacy is only understood in this way, privacy often loses." So, decisions about privacy are often made using social models and context, to help us determine the right tradeoffs.
Here are the questions that we use throughout this course to help us make those decisions:
Question | Answer |
---|---|
What type of information is shared? | |
Who is the subject of the information? | |
Who is the sender of the information? | |
Who are the potential recipients of the information? | |
What principles govern the collection and transmission of this information? |
Privacy Case study: Gaggle, a messaging app for schools
Gaggle, an online platform designed for use in the classroom that seeks to replace communcation tools such as blogging software and email clients with similar software equipped with content filters, states that "Gaggle will not distribute to third parties any staff data or student data without the consent of either a parent/guardian or a qualified educational institution except in cases of Possible Student Situations (PSS), which may be reported to law enforcement."
Imagine that a student sent a message to another student at 8pm on a Saturday and Gaggle flagged it as a potential indicator that the student is depressed.
Question | Answer |
---|---|
What type of information is shared? | Student data (the message, sender, recipient, timestamp, location) |
Who is the subject of the information? | The student and their mental health concerns |
Who is the sender of the information? | The student |
Who are the potential recipients of the information? | The other student and Gaggle. If Gaggle alerts the parents, school administration, or law enforcement, then they will also become recipients. |
What principles govern the collection and transmission of this information? | "Gaggle will not distribute to third parties any staff data or student data without the consent of either a parent/guardian or a qualified educational institution except in cases of Possible Student Situations (PSS), which may be reported to law enforcement." |
Often, as with our Homework assignments, the potential recipients of the information could also include unintended recipients, such as people looking over the shoulder of school administrators, or any students or teachers who notice the contacting of law enforcement.
Your Case Study: Algorithmic hiring
Now for some practice. Consider this scenario: “Shamazon” (a fictitious company) is looking to hire software engineers, and you have been tasked with designing a tool to filter the submitted resumes and select the ideal candidates for hire.
- Who are the stakeholders?
- What are the values?
- What is in the stakeholder-value matrix?
- What are the conflicts in the matrix? And how can we mitigate them?
- Where are bias and unfairness entering the product? And how can we mitigate them?
As you can see, we are relying on you to design our future. Good luck.
Footnotes
-
Nothing To Hide:The False Tradeoff Between Privacy and Security and 'I've Got Nothing to Hide' and Other Misunderstandings of Privacy by Daniel J. Solove are recommended further reading for those interested by the "I have nothing to hide" argument. ↩
-
This is not an endorsement of CheaterBuster. ↩