An Ethical Framework for AI
I've been teaching a course at UT-Austin called Law, Ethics, and Artificial Intelligence. As I developed the content, I found it helpful to create a framework for the students and I to use when analyzing assorted topics (e.g., AI and medicine, autonomous vehicles, AI and war, AI and finance, etc.). The following is what has been a useful tool to pick apart issues in a systematic way, exploring how each element applies to the topic at hand. I'm sharing it here with the hope that others will find it helpful as well. The most recent iteration of the mnemonic we use is: "Separate DB tabs".
For example, when we analyze whether the AI used to detect cancer in radiology images is ethical, I encourage the students to ask themselves questions like:
- Where did the data come from that was used to train the AI model?
- Who is liable for misdiagnoses by the AI? Is it the doctor? The hospital? The company that created the AI?
- Has the AI company been thoroughly audited to affirm that the AI performs as advertised?
A Deeper Dive: "Separate DB Tabs"
Source of Data
Understanding how the organization that created the AI collected the data it used to train the AI model is important. It can have both ethical and legal implications.
- Where did the data come from and how did they get it?
- Is the data representative of the population it's trying to model, or did the organization only gather data that was convenient to collect?
- Who participated in the data creation?
- Who is left out of the datasets?
- Who chose which data to collect?
- Did the sources provide informed consent to use their data for the purposes of the model?
Environment
The pros of using AI can't be overlooked, but the cost it has on the environment is unavoidable. The good of a modern technology is only a temporary benefit if it's significantly outweighed by the damage done to the planet.
- What impact will the AI have on the environment?
- How much electricity will the AI need?
- Where will the electricity come from (oil, gas, solar, wind, etc.)
- How can energy usage be reduced?
- How much pollution does the AI generate?
- How much water waste does the AI create?
Privacy
If AI models are cultivated using datasets containing private information, further inspection is necessary to preserve peoples' privacy.
- What security is in place to prevent Personal Identifying Information (PII), health information, and/or financial information from leaking?
- Who can access the PII?
- Do people have a write to having their PII removed from the model?
- Does the AI contribute to what may be privacy-related torts (intrusion on seclusion, false light, public disclosure of private facts, and appropriation of name or likeness?
Accountability
Regardless of whether the regulatory scheme is based on self-regulation or one imposed by a government, there must be clear accountability for the scheme to be effective.
- Who should ultimately be responsible for the AI’s actions? Should it be the user? The programmers of the AI? The company? The executives? The board of directors?
- How will the user get justice if the AI harms them?
- Is there someone who oversees and "owns" the AI development within the company?
Responsibility
Society must determine who should get to make decisions regarding AI that impacts society. For example, should it belong solely to profit-driven private individuals? Or should there be more democratic and public input?
- Who should decide what AI should and should not be allowed to do? The engineers? The executives at the company? The Board of the company? The government?
- Who should have access to advanced AI? Should everyone be trusted equally?
- How do we ensure it's not used for harmful activities?
Anthropomorphism
As numerous studies and common interactions with smart speakers and chatbots reveal, humans are quick to anthropomorphize AI by giving them pronouns and interpreting its actions as being driven by intent. It's not difficult to think how this tendency could be used to convince people to take actions they may not have otherwise taken.
- Does the AI take advantage of user emotions? For example, does it try to come across as human or as a sympathetic creature so the user will spend more money on something or take some action they likely would not have otherwise taken?
- Is there a reason the AI company chooses to make the AI seem human? For example, why give a speaker a human voice rather than one that is more alien or robotic? Is that the best alternative?
Trust
One of the primary purposes for using AI is for it to give us reliable results. However, some advanced AI systems, like LLMs, have shown that a response from a complex AI doesn't necessarily equate to precision, as they are capable of creating fictional facts and URLs.
- Are the outputs of the AI accurate?
- How can we be sure an LLM isn't hallucinating?
Explainable / Interpretable
It's vital the AI can explain in sufficient detail how it reached its conclusion. This is especially relevant when AIs make decisions that can impact others' livelihoods, such as with finances or healthcare. A black box that humans are told to place their faith in is not sufficient.
- Can we interrogate the AI?
- Can the AI accurately explain why it reached its output?
- Can the AI explain how it reached its output?
- Explainable AI should consider who it should be explainable for. For example, AI for medicine should be explainable to both doctors and patients.
- Are doctors using risk scores to triage patients able to interpret a risk score correctly? Are bankers able to interpret AI ratings of loan applicants?
- Is the system output only a single score or does it also output potential decision explanations and guidance for how to use the score?
Data Quality
Apart from the method of collecting data (reference: Source of data, above), it is imperative to take the data quality into consideration. Entering bad data into a model will result in an unreliable model that could have dangerous implications. Garbage in, garbage out.
- Is the data reliable and accurate?
- Is there any measurement error (for example, from incomplete or incorrect data)?
- Did selection bias play a role in data collection? Selection bias can occur if the subset of individuals represented in the training data is not representative of the patient population of interest.
- Does the AI model overfit because it was trained on too little data?
- Are historical biases being kept alive via the AI?
- Are some ethnicities, races, nationalities, or other populations inappropriately missing from the data? For example, text scraped from the web may overrepresent people who can afford internet access and have the time to post a comment.
- Check to ensure the data between groups is substantially similar. If it’s not, you may get inaccurate and unfair results. This happens with credit ratings when poorer folks have less credit history, so they are given lower scores, which makes it harder to work up to higher scores.
Beneficiaries
Because the most advanced AI models require many millions of dollars to create, only the already-wealthy can participate in its creation. This severely limits who can create the AI and naturally forms a dominant oligopoly as the success of the advanced AI makes the already-wealthy companies wealthier. The same effect likely happens on every scale regarding use of the AI as well. Entities that can afford to use the AI will benefit more than those who can't use the AI, which extends the advantage of the entities that are already relatively wealthy.
- Do the rich get richer at the expense of the less fortunate? Do individuals benefit? Or certain communities? Or certain nations?
- Is there inequitable access to the AI (e.g., low SES people may go to hospitals that don’t have the latest and greatest AI; low SES students may not have the best AI in their schools)?
- Is there any issue with Fortune 500 corporations growing their profits to benefit a relatively small number of people (investors) by replacing human workers with AI? Or is this a feature of capitalism, not a bug?
Transparency
Transparency is being open and clear with the user about how the AI should be used, when the AI is being used, the limitations of the AI, potential benefits and harms, etc. In short, the AI should not be a secret to the user.
- Are the kinds of harms an AI system might cause, the shortcomings the AI may suffer from, the varying degrees of severity of the harms, the limitations of the technology, how and when it’s being used, and so on communicated to the user?
- If the AI creates visualizations, like charts, does it include context, known unknowns and uncertainties, the source of the data, and how the source data was collected?
- Examples of limitations the AI may need to disclose: (a) AI systems are parametric, meaning the same data run through the same system could yield a different result each time. (b) The AI can only see what’s in the data (text, video, image, rows and columns, etc.). It doesn’t know the context around it (e.g., maybe someone had high blood pressure because a loved one recently died).
- Do all interfaces of the AI that make decisions about users clearly and succinctly inform users what is going on, how it works, how long the outputs are useful, and why the AI was needed to reach the output as opposed to any other method?
- Does the AI system allow users to obtain the information it has about them, and let the users dispute that information if the users believe it's inaccurate?
Audited
It's one thing for a company to make claims about the safety, security, accuracy, and impact of their AI. It's another to have an independent third-party audit and verify those claims. With advanced AI it may be better to verify, then trust.
- Have independent auditors reviewed the data and source code to verify it is and does what it says it is/will do?
- Is the AI model independently audited on at least an annual basis?
- Can we be certain the claims made by the developers are accurate and reliable? For example, that the data isn't biased, that the outputs are trustworthy, and that the environmental impact is no greater than claimed?
Bias
As seen in hiring, promotions, image recognition, allocation of resources, and in criminal justice, among many, many other areas, AI is fallible. The developers may have the best intentions and the algorithm may appear entirely neutral, but if the outcomes are biased, or if the data used to train the data was gathered during more biased times in history (against people of certain races, genders, sexual orientations, etc.), then the AI should face greater scrutiny.
- Is the data biased to favor one group over another based on an irrelevant or protected class?
- Were various forms of bias explored (e.g., disparate impact, statistical parity, equal opportunity, and/or average odds), or did the company only publish the most flattering analyses?
- Does the AI perpetuate societal biases?
Stakeholders
Representatives from all groups who may be affected by the AI should be consulted during the AI's development prior to its release. Stakeholders should include even those the AI may affect who won't even know of the AI's existence and will never interact with it directly becasue they could be affected indirectly.
- Did the AI company get feedback from a representative sample of everyone the AI will affect?
- Are the values of the users, the developers, and the society in which the AI is used align?
Data & AI fueled leadership: delivering measurable business outcomes
10moSunita Kannan Jack Lee Bechir Sellami
People Ops Leader | Lawyer | Tech | Start-Ups | Data-Obsessed Economist
1yI think something key is missing - namely, what do we do in the event of a robot apocalypse?!?
Research Scientist | Linguistics Expert | AI Enthusiast | Data Wrangler
1yShared it with my Fairness in AI class. Great resource!
Chief Strategy Officer at vLex | Entrepreneur | Law Professor | Keynote Speaker
1yA very helpful framework - rigorous and thorough!