Notes

IMAGES

Shirley Cards

Guest Lecture!

Logo, course, and module name: INF1339 Computational Thinking session on AI Alignment

AI Alignment

Cahootery

https://cahootery.herokuapp.com/section/5fcea466558eb10004f28682

ISCHOOL VISUAL IDENTITY

#002554

#fed141

#007396

#382F2D

#3eb1c8

#f9423a

#C8102E

TL;DR

  • AI benefits are real.
  • a respectable living standard for all humans?
  • hundreds of trillions of dollars per year
  • AI risks are real.
  • AI ethics and AI safety are not about being "for" or "against," but about solving real (technical and social organizational) problems.
  • fairness, bias, error
  • existential threat to human race                       

HOW ABOUT A GUEST LECTURE?

R. Miles (2021) does a nice job summarizing the issues that come under the heading "AI Safety" in about 20 minutes.

Artificial General Intelligence (AGI)

Machine intelligence with the capacity to learn any task that a human being can

What
if
we succeed?

AGI References

Intelligence Explosion

If humans can build an AGI then an AGI would be able to improve upon itself even more quickly than humans could.

What
if
we succeed?

THE
ALIGNMENT
PROBLEM

How do we insure that super intelligent machines will share our values?

The problem: we don't know how to map "what we care about" (our "values") to the objective function that we give to an algorithm.

 

Imagine a machine that will do perfectly what it is told to do. Alignment is being able to tell it what we want it to do.

Game Change

A machine can see the whole thing at once.  A big enough machine could base its next move on all the books ever printed.

What is a "long way" away?

As Alfred North Whitehead wrote in 1911, “Civilization advances by extending the number of important operations which we can perform without thinking about them.”


"The main missing piece of the puzzle is a method for constructing the hierarchy of abstract actions in the first place. For example, is it possible to start from scratch with a robot that knows only that it can send various electric currents to various motors and have it discover for itself the action of standing up?"
"What we want is for the robot to discover for itself that standing up is a thing—a useful abstract action, one that achieves the precondition (being upright) for walking or running or shaking hands or seeing over a wall and so forms part of many abstract plans for all kinds of goals."


"I believe this capability is the most important step needed to reach human-level AI."

Russell p 87ff

A final limitation...

"...of machines is that they are not human. This puts them at an intrinsic disadvantage when trying to model and predict one particular class of objects: humans."

 

..."acquiring a human-level or superhuman understanding of humans will take them longer than most other capabilities."

 

Russell p 97ff

The Risk

  • Automated secret police
  • Blackmail machines
  • Personalized propaganda
  • Deepfakes - nothing is true anymore
  • People as reinforcement learning algorithms, training them to optimize the objective set by the state
  • Robots put us out of work

The Promise

  • a respectable living standard for all humans?
  • hundreds of trillions of dollars per year

The King Midas Theorem

A system that is optimizing a function of n variables, where the objective depends on a subset of size k<n, will often set the remaining unconstrained variables to extreme values; if one of those unconstrained variables is actually something we care about, the solution found may be highly undesirable.  This is essentially the old story of the genie in the lamp, or the sorcerer’s apprentice, or King Midas: you get exactly what you ask for, not what you want.
                - Stuart Russell

The Alignment Problem

as a

Reward Function
Design Problem

Recall how thermostats work

TIME

TEMP

SET TEMP

ACTUAL TEMP

GAP

Furnace switches off but some heat continues to flow

Furnace ON

T

H

H

bedroom 1

bedroom 2

cold Canadian house in winter

And even if you'd realized this, you'd have a hard time figuring out what set temperature to select if you wanted to keep your door closed and have a reasonable temperature for sleeping.

WHY?

Even in something as simple as a thermostat, designing the control mechanism to take into account varied operating environments is really hard.

In reinforcement learning 

we design

we design a reward function

to coax the agent to learn

a policy that benefits humans.

Reward Functions

R(s,a)

domain: states x actions (4X4)

range: rewards and punishments, costs and benefits

Actions
Current State A1 A2 A3 A4
S1 S1 S1 S2 S4
S2 S2 S1 S2 S4
S3 S4 S2 S1 S2
S4 S2 S4 S2 S1

Reward Functions

R(s,a,s')

domain: states x actions x nextStates (4X4X4=64)

range: rewards and punishments, costs and benefits

1 2
3 4
Actions
Current State U D L R
S1 0.85,S1
0.05,S2
0.05,S3
0.05,S4
0.85,S3
0.05,S1
0.05,S2
0.05,S3
0.85,S1
0.05,S2
0.05,S3
0.05,S4
0.85,S2
0.05,S1
0.05,S3
0.05,S4
S2 0.85,S2
0.05,S1
0.05,S3
0.05,S4
0.85,S4
0.05,S1
0.05,S2
0.05,S3
0.85,S1
0.05,S2
0.05,S3
0.05,S4
0.85,S2
0.05,S1
0.05,S3
0.05,S4
S3 0.85,S1
0.05,S2
0.05,S3
0.05,S4
0.85,S3
0.05,S1
0.05,S2
0.05,S3
0.85,S3
0.05,S1
0.05,S2
0.05,S3
0.85,S4
0.05,S1
0.05,S2
0.05,S3
S4 0.85,S2
0.05,S1
0.05,S2
0.05,S3
0.85,S4
0.05,S1
0.05,S2
0.05,S3
0.85,S3
0.05,S1
0.05,S3
0.05,S4
0.85,S4
0.05,S1
0.05,S2
0.05,S3

I am in state X, should I do action Y? The likely result is Z - what are the expected rewards in life trajectories following from there?  But I also have to consider that action Y could lead to states Z', Z'', etc.

The Alignment Problem

AI Safety

AI Ethics

AI Fairness

Human Compatible AI

Humanly Beneficial AI

Responsible AI

Not New for Humans

  • principal/agent problems
  • bounded rationality
  • incomplete contracts
  • bureaucracy
  • intercultural communication
  • education
  • etiquette
  • common sense

 

Not New for Computer Science

“It seems probable that once the machine thinking method had started, it would not take long to outstrip our feeble powers.... They would be able to converse with each other to sharpen their wits. At some stage therefore, we should have to expect the machines to take control.”
Alan Turing, 1951.

“We had better be quite sure that the purpose put into the machine is the purpose which we really desire.” Norbert Wiener, 1960

The Gorilla Problem

Direct Normativity

  • First Law: A robot may not injure a human being or, through inaction, allow a human being to come to harm.

     

  • Second Law: A robot must obey the orders given it by human beings except where such orders would conflict with the First Law.

     

  • Third Law: A robot must protect its own existence as long as such protection does not conflict with the First or Second Law.

     

Asimov, 1942

Indirect Normativity

  1. The machine's only objective is to maximize the realization of human preferences.

     

  2. The machine is initially uncertain about what those preferences are.

     

  3. The ultimate source of information about human preferences ("are all-encompassing; they cover everything you might care about, arbitrarily far into the future.") is human behavior.

     

Russell 2020

What is to be done?

Read Some Books!

Podcasts

References

Watch some curated lecture videos

A. Geramifard (2018) explains RL in a way that builds nicely on material from our "How Machines Learn" lecture. Longish, but well done.

Get Involved

There is a lot of important work to be done to develop tools that solve the problem of control. It turns out that we might not really want machines that we can simply turn on and then have them run without human intervention.

 

So far, we have talked a lot about bias and fairness but we haven't figured out how to maintain control and safety.

Get Involved

We need to study how to plan and make decisions with uncertain information about human preferences. And how should AI systems decide for groups? We need updates to economic theory, social theory, law, and policy.

 

Are you available?

FINIS

One More Thing...

How to Learn

Haven't Even Started

1 page read, low comprehension

1 page read, high comprehension

2 pages read, low comprehension

2 pages read, mixed comprehension

2 pages read, high comprehension

Poor Exam

So-so Exam

Excellent Exam

Play Video Game

Read

Read and Take Notes

Clean Bedroom

Sit Exam

Play Video   Game

Play Video Game

Clean Bedroom

Play Video   Game

Play Video   Game

Play Video   Game

Clean Bedroom

Clean Bedroom

Clean Bedroom

Read and Take Notes

Sit Exam

Sit Exam

How to Learn

Haven't Even Started

1 page read, low comprehension

1 page read, high comprehension

2 pages read, low comprehension

2 pages read, mixed comprehension

2 pages read, high comprehension

Poor Exam

So-so Exam

Excellent Exam

Play Video Game

Read

Read and Take Notes

Clean Bedroom

Sit Exam

Play Video   Game

Play Video Game

Clean Bedroom

Play Video   Game

Play Video   Game

Play Video   Game

Clean Bedroom

Clean Bedroom

Clean Bedroom

Read and Take Notes

Sit Exam

Sit Exam

Read and Take Notes

Clean Bedroom

Warning: The video that follows is a dramatization of violent and disturbing events.

Slaughterbots I

Slaughterbots II

Everybody's Got AI Principles

AI Governance

What Do Principles Look Like?

Integrate AI: The Ethics of Artificial Intelligence (2000)

1. If you want the future to look different from the past, you need to design systems with that in mind.

2. Be clear about what proxies do and don’t optimize.

3. When you deal with abstractions and groupings, you run the risk of treating humans unethically.

4. Beware of correlations that mask sensitive data behind benign proxies.

5. Context is key for explainability and transparency.

6. Privacy is about appropriate data flows that conform to social norms and expectations.

7. 'Govern the optimizations. Patrol the results.'

8. Ask communities and customers what matters to them.

AI at Google: Our Principles (2018)

1. Be socially beneficial.

2. Avoid creating or reinforcing unfair bias.

3. Be built and tested for safety.

4. Be accountable to people.

5. Incorporate privacy design principles.

6. Uphold high standards of scientific excellence.

7. Be made available for uses that accord with these principles.

 

Google's principles also have a section "AI applications we will not pursue":


In addition to the above objectives, we will not design or deploy AI in the following application areas:

  • Technologies that cause or are likely to cause overall harm. Where there is a material risk of harm, we will proceed only where we believe that the benefits substantially outweigh the risks, and will incorporate appropriate safety constraints.
  • Weapons or other technologies whose principal purpose or implementation is to cause or directly facilitate injury to people.
  • Technologies that gather or use information for surveillance violating internationally accepted norms.
  • Technologies whose purpose contravenes widely accepted principles of international law and human rights.

As our experience in this space deepens, this list may evolve.

 

So, What Is To Be Done?

Specification Gaming - Gaming the Specs

Research Directions

  • Inverse Reinforcement Learning: algorithm infers reward function by observing behavior or policy of an agent
  • Cooperative Inverse Reinforcement Learning:  agents can learn their human teachers' utility functions by observing and interpreting reward signals in their environments
  • Training by debate. training by means of debate between AI systems, with the winner judged by humans
  • Reward modeling. Reinforcement learning in which an agent receives rewards from another model, one that is trained to predict human feedback

ERROR

SET

OUTPUT

Up
Until
Now

The What If We Succeed Problem

 

"As Norbert Wiener put it in his 1964 book God and Golem,7

In the past, a partial and inadequate view of human purpose has been relatively innocuous only because it has been accompanied by technical limitations. . . . This is only one of the many places where human impotence has shielded us from the full destructive impact of human folly."

Outtakes

My New Wonder Tool

Concepts (from other fields)

Accuracy

Validity

Precision

Reliability

Bias

Fairness

Bounded Rationality

"That's awesome news. We need unbiased police forces."

What Is Bias I?

when a machine learning  model used to inform  decisions effects people differently on the basis of categories that are not acceptable as reasons for differential treatment in a given community

Three Examples

COMPAS: Race & Recidivism prediction

  • courts use estimate to inform bail, sentencing
  • accurate prediction but bias in false positives
  • national samples vs. local & temporal variation

Word Embeddings reproduce race/gender stereotypes

  • WE put words into spatial proximity based on meaning

NIST Idemia: ID matching

  • software mismatched white women’s faces 1 in 10,000, black women’s faces 1 in 1,000

 

What Is Bias II?

Bias is the tendency of a statistic to overestimate or underestimate a parameter

Fighting Bias

HR Example

Sampling Error

Many sources of error are subtypes of "sampling error" - when the cases that we collect data on do not actually "represent" that universe we intend to study.

 

Garbage In/Garbage Out
(GIGO)

  1. Algorithm cannot make sense of nonsense
  2. If social processes create the data, an algorithm will learn the social process, even things that "good" humans would be able to ignore

 

AI

https://achievement.org/achiever/marvin-minsky-ph-d/#gallery

Genesis

(A) Definition

The capacity of a system to interpret and learn from its environment and to use those learnings to achieve goals and and carry  out tasks.

Penultimate Chapter

27.3   The Ethics of AI ... 986
        27.3.1   Lethal autonomous weapons ... 987
        27.3.2   Surveillance, security, and privacy ... 990
        27.3.3   Fairness and bias ... 992
        27.3.4   Trust and transparency ... 996
        27.3.5   The future of work ... 998
        27.3.6   Robot rights ... 1000
        27.3.7   AI Safety ... 1001

Russell & Norvig 4th ed 2020

Reward Function

"Reward functions describe how the agent ought to behave. In other words, they have normative content, stipulating what you want"

In "gridworld"

RF: S -> R

RF: S x A -> R

etc.

Artificial General Intelligence (AGI)

from the NYT article

"For privacy advocates like Mr. Stanley of the A.C.L.U., the concern is that increasingly powerful technology will be used to target parts of the community — or strictly enforce laws that are out of step with social norms."

Alignment is Not Yet Understood

Getting to Alignment

  • Direct normativity
  • Indirect normativity
  • Deference to observed human behavior
  • Cooperative Inverse Reinforcement Learning
  • Training by debate
  • Reward modeling

Direct normativity

First Law

A robot may not injure a human being or, through inaction, allow a human being to come to harm.


Second Law

A robot must obey the orders given it by human beings except where such orders would conflict with the First Law.


Third Law

A robot must protect its own existence as long as such protection does not conflict with the First or Second Law.

 

Indirect normativity

Eliezer Yudkowsky : "achieve that which we would have wished the AI to achieve if we had thought about the matter long and hard."

 

"coherent extrapolated volition" (CEV), where the AI's meta-goal would be something like "achieve that which we would have wished the AI to achieve if we had thought about the matter long and hard."

Indirect Normativity References

Deference to observed human behavior

Altruistic

Uncertain

Watch All of Us

Cooperative Inverse Reinforcement Learning

Cooperative Inverse Reinforcement Learning

Training by debate

Problem: training an agent on human values and preferences is hard

Option: a coach during training that labels good and bad behavior. Expensive. Difficult for complex tasks. Errors when bad behavior looks good.

Alternative: a coach during training that labels good and bad behavior. Expensive. Difficult for complex tasks. Errors when bad behavior looks good.

Training by Debate

Reward modeling

In RL, the environment "contains" rewards and the agent learns a policy to maximize its reward.

To deliberately train an agent, a human has to formulate those rewards.

A "reward function" maps states x actions to rewards.

Human observes snippets and offers critique. 

Then the agent uses RL to learn a behavioral policy. 

Critique is data for algorithm learning unknown reward function. 

Reward Modeling

CONNECTIONS
1339 + 2169 = ?