Instructor: Dan Ryan

Human and Machine Intelligence Alignment

Fall 2025

"Alignment" is an evocative term and a polysemous term - it can mean many things.

In this course we will build a broad understanding of alignment. As a starting point, we can say that a system is aligned when it does what it is intended to do, and when its actions are beneficial to other people rather than harmful.

 

For now, we defer questions about what harmful or beneficial mean and to whom a system might be beneficial or harmful.

Ryan: Human and Machine Intelligence Alignment Fall 2025

The origin of the course was a lightning talk I gave at FOO Camp in 2024.

The gist of the talk was that social theory has always been the study of human intelligence alignment. Social order - practices and structures that support the coordination and cooperation that allow humans to thrive - is alignment. 

 

In the talk I suggest social theory has articulated three big insights about human intelligence alignment: socialization; shared meaning, group life. 

 

Ryan: Human and Machine Intelligence Alignment Fall 2025

Ryan: Human and Machine Intelligence Alignment Fall 2025

Dan Ryan on Social Theory and Alignment [6m]

This
is not our
first rodeo

What This Course Is NOT

Ryan: Human and Machine Intelligence Alignment Fall 2025

Ethics of AI

Computer Science / Engineering Ethics

AI 101

Technical AI Alignment and Safety

AI Policy and Regulation

But here's a course on that...

But here's a course on that...

601.104  COMPUTER ETHICS

Students will examine a variety of topics regarding policy, legal, and moral issues related to the computer science profession itself and to the proliferation of computers in all aspects of society, especially in the era of the Internet. The course will cover various general issues related to ethical frameworks and apply those frameworks more specifically to the use of computers and the Internet. The topics will include privacy issues, computer crime, intellectual property law -- specifically copyright and patent issues, globalization, and ethical responsibilities for computer science professionals. Work in the course will consist of weekly assignments on one or more of the readings and a final paper on a topic chosen by the student and approved by the instructor.

But here's a course on that...

Values and Ethics in Artificial Intelligence - 705.612

Modern artificial intelligence, and the related area of autonomous systems are becoming so powerful that they raise new ethical issues. This course will prepare professional engineers and developers to thoughtfully engage with the moral, ethical, and cultural aspect of these emerging technology. Topics include: safety considerations for autonomous vehicles, algorithm bias, AI explainability, data privacy, ethical considerations of ‘deep fakes’, ethics of artificial life, values advocacy within organizations, technological unemployment, and far-future considerations related to AI safety. 

There are a million of these on the web....

But here's a course on that...

EN.705.613.  Responsible AI

This course explores the ethical, societal, and policy implications of artificial intelligence (AI) technologies, providing students with a comprehensive understanding of the responsibilities that come with their development and deployment. Through case studies, real-world examples, and critical discussions, students will examine issues related to fairness, transparency, privacy, accountability, and bias in AI systems. The course also delves into the regulatory landscape surrounding AI, exploring frameworks for responsible innovation and the governance structures needed to ensure AI aligns with human values. By the end of the course, students will be equipped with the tools to assess AI systems from an ethical standpoint, design responsible AI solutions, and contribute to the ongoing dialogue around the future of AI and its impact on society.

What This Course IS:
The Four Alignment Problems

Ryan: Human and Machine Intelligence Alignment Fall 2025

The Four Alignment Problems

Ryan: Human and Machine Intelligence Alignment Fall 2025

How to Live with Other Human Intelligences

How to live with Organizational Intelligences

How to Live with Machine Intelligences

How to Live with Expert Intelligences

Machine intelligence alignment problems arise when societies introduce artificial agents that operate at or above human capability and are beyond human control. An aligned machine intelligence does what it is supposed to do and performs in a manner that is humanly beneficial.  Because machines lack social context, moral intuition, or an intrinsic understanding of purpose, alignment cannot be assumed to emerge naturally; it must be deliberately engineered through technical design, institutional oversight, and governance structures that keep non-human cognition responsive to human ends rather than allowing the machines to optimize around them.

Organizations are a primitive form of multi-agent super-intelligence. Humans can create structures that amplify collective intelligence while dampening the limits and biases of individuals. From coordinated hunts and village life to modern corporations and state bureaucracies, human intelligence has repeatedly used organizations to create agents capable of acting at scales no single human could achieve.

Organizations pose profound alignment problems. Individuals are not naturally designed to function as interchangeable components, and organizational goals rarely map cleanly onto human values or moral intuitions. And organizations introduce sharp asymmetries of power, knowledge, and coordination  into human communities. They can act with speed and reach that exceeds those of the individuals they are meant to serve. The organizational alignment problem is twofold: aligning individuals within organizations and aligning organizations themselves with the interests and values of the societies that authorize their existence.

Experts are super-intelligent agents wielding concentrated portions of collective knowledge. Epistemic asymmetry makes expertise valuable but risky: laypeople must trust judgments they cannot inspect, challenge, or reproduce.
The alignment problem is ensuring the person who is a doctor, witch, or engineer acts as faithful carrier of collective understanding rather than unaccountable intelligence pursuing private goals under expertise's cover. Because experts cannot be directly monitored, societies use indirect alignment mechanisms.
Professional norms, credentialing, peer review, reputation systems, liability, and institutional embedding constrain expert discretion without eliminating it. These mechanisms don't guarantee benevolence; they keep expert power socially accountable rather than self-authorizing. The challenge is not removing asymmetry but governing it: preserving specialized intelligence's benefits while preventing its domination of collective judgment.

Human coexistence is not natural; it is engineered. The achievement of human intelligence is been less about solitary problem-solving than about making life among other intelligent agents tolerable, predictable, and productive. Humans must constantly ask: Are you safe? Will you exploit me? Do we understand the situation the same way? Can we act together? Humans have invented a layered set of tools to manage these risks.

Reputation, reciprocity, and commitment devices generate trust. Language bootstraps shared meaning, allowing people to compress experience into symbols. Norms, rituals, and conventions reduce the cognitive burden of coordination. Institutions such as turn-taking, majority rule, and hierarchy make conflict governable.

The history of human intelligence is a history of alignment technologies, social, cultural, and institutional inventions that make multi-agent life possible.

About the Instructor

Ryan: Human and Machine Intelligence Alignment Fall 2025

Alignment Cards

Ryan: Human and Machine Intelligence Alignment Fall 2025

Why Are We Here?

Ryan: Human and Machine Intelligence Alignment Fall 2025

2015

Do YOU think we should  worry about intelligent machines that do exactly what we ask them to do?

Ryan: Human and Machine Intelligence Alignment Fall 2025

Ryan: Human and Machine Intelligence Alignment Fall 2025

Who are you?

Ryan: Human and Machine Intelligence Alignment Fall 2025

Ryan: Human and Machine Intelligence Alignment Fall 2025

Ryan: Human and Machine Intelligence Alignment Fall 2025

2019

What
Is
Value
Alignment?

Ryan: Human and Machine Intelligence Alignment Fall 2025

2019

Ryan: Human and Machine Intelligence Alignment Fall 2025

AI 101

Alignment 101

How to read hard stuff

Thinking Analogically

 

The Four

Alignment

Problems

 

Isn't It Just "Ethics"?

Principles Are
Not Enough!

Align is a Verb

 

 

 

 

 

 

Shared Meaning

Hierarchy

Groups

Markets

Qualification

Records

Control

Deterrence

Incentives

Safety

Institutions

Foundations

 

 

 

 

Humans
Have Been Doing Alignment Forever

 

 

 

 

Traits Are
Not Enough!

 

 Humans +

Humans

 

 

 Principals + Agents

 

 

 Normies +

Magicians

 

 

 Humans +

Machines

 

Humans

Organizations

Experts

Machines

Humans

Organizations

Experts

Machines

Humans

Organizations

Experts

Machines

Humans

Organizations

Experts

Machines

Humans

Organizations

Experts

Machines

Humans

Organizations

Experts

Machines

Humans

Organizations

Experts

Machines

Basic Requirements

Ryan: Human and Machine Intelligence Alignment Fall 2025

27 class meetings.

~25 pre-class work.

<25 post-class work.  

Multiple draft essay.

Alignment card deck.

Oral exam.

Alignment Cards

Ryan: Human and Machine Intelligence Alignment Fall 2025

Alignment Chat

Ryan: Human and Machine Intelligence Alignment Fall 2025

Next Time on...

PostClass Work

PreClass Work AI Bootcamp

 

Ryan: Human and Machine Intelligence Alignment Fall 2025

Further Reading on Alignment

  1. IBM: "What is alignment?"
    • General answer with examples of things IBM is developing
  2. Wikipedia. "Al Alignment" 25-30 min read
    • Read first three sections and skim the rest for a first reading
  3. Vinod Chugani 2024: "What is AI Alignment? Ensuring AI Works for Humanity" on DataCamp blog · 12 min read
    • Ignore opening idea (building in values); after that several good break downs of concepts related to alignment
  4. Paul Christiano. 2018. "Clarifying 'AI alignment'." 4 min read
    • head of AI safety at the US AI Safety Institute
  5. Ji, J. et al. 2024. "AI Alignment: A Comprehensive Survey" 60PP!
    • dense text, read from the outside in - look at contents and overall shape
  6. Tech AI Digest. 2024. "AI Alignment Explained: Ensuring AI Follows Human Values" (11min podcast)
    • 2 person conversation - very beginner friendly

Ryan: Human and Machine Intelligence Alignment Fall 2025

We just learned three terms

maximize a utility function

inverse reinforcement learning

value alignment

Ryan: Human and Machine Intelligence Alignment Fall 2025

and something about a guy named Szilard

HMIA 2025 External Overview

By Dan Ryan

HMIA 2025 External Overview

  • 1