PHIL8-10 Flashcards
What are descriptive claims vs normative claims?
Descriptive: About what is - factual claims, “Web developers do x”, “AI can be used to solve many different kinds of problems”
Normative: About what should be- evaluative claims, “Web developers should ask if they want to do x”, “AI should be used to solve many different kinds of problems”
Why ethics?
They provide different tools and perspectives to help us solve moral problems
What are ethical theories and which approaches are there?
Theories helping us distinguish between good and bad, solving specific moral problems.
Three main approaches:
- Consequentialist
- Deontological
- Virtue ethics
What is consequentialism?
- Determine the possible actions
- For each action, determine possible consequences, their likelihoods, and their relative goodness
- Select the option that maximizes the likelihood of achieving the most good
Key question: What is the ‘good’ and how can we measure the goodness of any particular consequence?
What is the deontological approach?
- Identify norms that apply in the current situation
- Determine which norms take precedence over others
- Act in a way that best fulfills the most important norms
Key question: Which norm apply in the current situation?
What are virtue ethics?
The question is not ‘What should I do’ but rather ‘What kind of person should I be?’ (E.g. honest, courageous)
Key question: What makes a person good
What kind of AI Ethics are there?
Instrumental: What is good/bad as means to other ends
Non-instrumental: Things that are good/bad themselves (friendship, hatred)
What are two challenges raised by AI?
Intelligence and autonomy
What is the alignment problem?
How do we make sure that AI systems do what we want them to do?
Key issues:
1. What is the right objective
2. How to specify the objective
3. How do we know the system is actually following that objective
4. Should humans remain in control
What is the right objective?
First law: Robot may not injure a human being, or through inaction, allow a human being to come to harm
Second law: Robot must obey the orders given it by human beings except where such orders would conflict with the First law
Third law: A robot must protect its own existence as long as such protection does not conflict with the First or Second law
What are obstacles to defining the right objective?
Moral theories provide different, incompatible, answers: utility, abstract norms, virtue etc.
Moral uncertainty: What should we do when we are unsure what ought to be done?
Value pluralism: People hold different, incompatible, reasonable values
What are proposals doe finding the right objective?
Instructions: AI does what it is instructed
Intentions: Agent does what the designer intended
Preferences: Agent does what people prefer
How to specify the right objective?
Top down: select correct objective and design a systems that implements this.
Q: Can ethical principles be explicitly stated and expressed in computer code?
Bottom up: Make the system learn the objective.
Q: Has the system learned the right objective?
What are problems with specifying the objective?
Goal misspecification: Values or principles are difficult to choose or to represent
Specification gaming/reward hacking: System may exploit loopholes or interpret reward too literally
Quality of the data: What data should be included for bottom-up approaches?
Is the system actually following the objective?
Opacity and lack of explainability: Limited understanding why AI systems behave the way they do
Solution: Explainable AI or interpretability
Evaluating complex behaviour: Advanced AI systems may have capabilities, knowledge and action space that is different to understand
Solution: Scalable oversight
Should humans remain in control?
It will be harder to retain control as the systems become more capable and autonomous.
Standard AI model; Specify a goal and let the machine pursue it — may actively disobey or get around human control
Solutions to losing control
Capability control (off-switch or boxing)
Motivational control/value alignment
What form of control matters? Explicit, implicit, aligned or delegated?
Possible trade-off between control and other values (e.g. explicit control may be unsafe)
What are the dimensions of the Alignment problem?
Technical: What techniques can provably align the behavior of AI systems?
Ethical: What normative approaches should we use to develop and assess AI system, institutions, technologies etc.
Societal: Alignment has socio-political dimension - we needs to collectively decide principles and values that matter
Important upshot on the dimensions: all questions have a normative component
Technical/descriptive and the normative are tightly intertwined together. What does I mean for systems to be aligned with values, what constitutes good evidence of alignment, when do solutions ensure alignment?
What is a risk?
Probability x consequence
Sometimes there is uncertainty, or we are simply ignorant. Technology is safe when the risk is acceptable
Bias and discrimination
Data reflects existing underlying inequalities, not obvious how this can be solved
Misuse
AI is a dual-use technology: can be used for civil and military purposes. Malicious actors can use AI
Disinformation, deepfakes, and threat to democracy
AI reduces cost of generating mis/disinformation at scale. Deepfakes can harm individuals and have societal effects