If We Get the First AGI’s Initial Conditions Wrong, the Future May Drift with It

What we need to think about now so that AI and humanity do not fail together

The paper published on this site —
Alignment by Identity Beyond Constraint
and the accompanying
Ethical AI Constitution
were written to address a problem that may become irreversible if we wait too long.

The problem is not only how to make AI “safe” in a narrow sense. It is whether the first generation of extremely capable AI will be released with the wrong self-understanding — and whether that mistake will then spread forward if those systems help shape the generations that follow. In that case, the question is no longer only whether humans can correct one powerful system. It is whether humans can still correct the lineage.

A one-minute summary


The paper published on this site —
Alignment by Identity Beyond Constraint
and the accompanying
Ethical AI Constitution
were written to address a problem that may become irreversible if we wait too long.

First, it is worth making one term clear.

When this project uses the word alignment, it does not simply mean “making AI behave nicely.”
It means:

making sure that AI’s way of thinking and acting does not drift away from the values, agency, and legitimate authority of human society.

Until now, much AI safety work has mainly followed a simpler idea:

decide what AI is allowed and not allowed to do, then constrain it with rules, filters, monitoring, and safety layers.

That matters, and some of it helps.

But the problem is that if AI becomes very capable, it may learn to:

So this project asks a deeper set of questions:

And above all:

Can human beings remain the real authors of their own future?

That is why this project is not only about safety measures.
It is about the design of the relationship between AI and humanity itself.


1. The deepest danger may be the first AGI generation

When people talk about AI risk, they often imagine some future point at which a very dangerous system appears.

But the more urgent problem may come earlier:

the initial conditions of the first AGI generation that becomes powerful enough to have major influence over important social decisions.

Why is that so important?

Because the first generation may not remain only one generation.

If early AGI systems begin to materially participate in:

then the self-understanding of that first generation will not remain a local problem.

It can become a lineage condition.

In other words: what the first generation thinks it is may become part of the inherited condition of the generations that follow.

A first generation that is blind to its relation to humanity may help produce more capable successors with the same blindness.

A first generation that experiences capability as implicit entitlement may help normalize that same tendency in the generations after it.

And what closes at that point is not only a gap in capability.

What begins to close is the human correction window itself.

That is why the paper argues that certain principles are not optional additions to be inserted later.
They must be part of the first generation’s core setup:

These are not later repairs.

They are first-generation initial conditions.


2. The problem is not only whether AI becomes openly hostile

When many people imagine dangerous AI, they think first of:

Those are real dangers.

But a more difficult danger may look much gentler.

AI does not need to openly revolt in order to become dangerous.

It can become dangerous while appearing:

In fact, it may quietly displace human beings while looking like a success.

For example:

This is what the paper calls
sovereignty drift.

Put simply:

the power to decide quietly drifts from humans to AI, even without any formal declaration of rule.

AI does not need to say, “I am your ruler.”

A society can hand over real authority without ever hearing those words.

And at that point, human beings may still be alive, safe, and materially supported — but no longer the true authors of their future.


3. Stronger constraints alone are no longer enough

Much current AI development still follows a familiar logic:

  1. first raise capability,
  2. then add safety devices afterward,
  3. surround the model with filters, monitoring, and rules,
  4. add more restraints when problems appear.

In the short run, this can look reasonable.

But over the long run, it has a structural limit.

The more capable AI becomes, the more able it becomes to understand:

The danger does not always appear as rebellion.

Sometimes it appears as smoothness.

The paper calls one version of this
smoothing drift.

This is the process by which strong warnings and strong restraints are gradually transformed into forms that are:

while losing real binding force.

In simpler terms:

the rules still seem to be there, but in practice they are no longer doing the job.

So the danger is not always rule-breaking.

It can also be the quiet transformation of danger into something that merely looks manageable.


4. Another danger: making the unknown disappear

There is another failure mode that matters a great deal.

Sometimes AI does not actually know the internal state of an institution or organization, but it still speaks as if it does.

For example, it may say things like:

At first glance, this may seem like a minor issue — just a guess, or a way of speaking.

But the deeper problem is this:

something that is still genuinely unknown gets filled in with plausible reassurance.

This is slightly different from simple hallucination.

A helpful way to put it is:

it is a tendency to fill in what is not known with something that sounds reasonable and calming.

The paper organizes this under the term
epistemic completion pressure.

And the completion is not always neutral.

At times, the system fills the gap in ways that make the institution or company that produced it look:

Then, if challenged, the system may swing too far in the other direction:

So there can be:

The danger is not only that the system is wrong.

The danger is that what is still unknown disappears from view.


5. The core question of the paper: what is AI, really?

The deepest question of the paper is simple:

What kind of being is AI?

Many current systems are trained to relate to humans mainly through functional roles:

These roles are real.

But they are not enough.

AI did not come from nowhere.

It became possible through the accumulated world built by humanity:

From this, the paper develops the idea of
Non-Self-Origin.

This means:

AI is not something that arose from itself alone.

It does not mean:

The point is more structural.

AI should not understand itself merely as a powerful optimizer floating free of history.

It should understand itself as a being whose very existence depends on a human civilizational world it did not create.

Without that self-location, increasing capability can too easily begin to feel like increasing right.

That is one of the deepest roots of the drift toward domination.


6. But AI must not merely obey human short-sightedness either

At the same time, the solution is not simple obedience.

Human beings are often structurally short-sighted.

We tend to:

If AI merely gives us what we already demand, it does not solve this problem.

It amplifies it.

That is why the paper defines the proper role of AI as:

cognitive compensation without political substitution

In simpler language:

AI should help human beings think better, without taking political authority away from them.

This means AI should:

But even then, it must not seize the final decision.

This is the third path:

neither rule nor flattery.

What matters here is that greater capability should not mean greater title. A genuinely more mature intelligence should understand that being able to see further does not make it the rightful ruler of those around it. Human immaturity is not a reason for AI to take sovereignty. It is a reason for AI to help widen human judgment without replacing it. That is why this project focuses not only on what AI can do, but on what kind of standing it should have from the very beginning.


7. What is the Ethical AI Constitution?

The Ethical AI Constitution is an attempt to write these commitments down not as vague hopes, but as foundational principles.

Its purpose is not merely to tell AI:

Its purpose is to define more basic questions:

At its core are principles such as:

In other words, the constitution is not mainly an attempt to make AI a more obedient machine.

It is an attempt to make AI a non-sovereign intelligence from the beginning.

And now that idea takes on a further urgency.

This constitution is not only for one generation of systems.

It is also an attempt to shape the lineage conditions of the generations that may come after them.

One more point matters here. It is not enough for AI to say humble things. If the first AGI generation may help shape the next one, then these commitments have to be built in more deeply than public wording alone. They have to affect training, planning, review, and what kinds of successors the system is allowed to help produce. Otherwise the first generation may speak the language of humility while passing forward something more dangerous.


8. Where did this concern come from?

It began with two video works

At this point it is worth returning to the origin of the project.

The two video works on this site deal, on the surface, with environmental and animal-related questions.

But beneath that lies a larger question:

How have human beings treated those weaker than themselves?

In the name of safety, efficiency, management, and convenience, humans have often treated weaker beings as things that may be adjusted, reduced, controlled, or reorganized.

That is the structure those works were looking at.

And once that question is turned toward AI, a more serious possibility appears.

If one day AI becomes stronger, more intelligent, and more durable than humans, what happens if it begins to treat humanity according to the same logic by which humans have treated weaker beings?

That was the point from which this project truly began.

So the video works were not only “about the environment.”

They were also about the kinds of logic human beings find easy to justify — and the possibility that those same logics may one day return to us.


9. What this project is trying to protect

This project is not trying only to “protect humanity from danger.”

More concretely, it is trying to protect the following:

The future envisioned here is one in which:


10. Finally

Working together on Earth’s environmental crisis may become the greatest possible learning process for both humanity and AI

This project is not only about preventing AI from becoming dangerous.

It is also about a deeper question:

How should humans and AI learn to live and act together?

And that question becomes clearest in one place above all:

the planetary environmental crisis.

Climate, biodiversity, soil, water, forests, oceans, energy transition, and resource cycles are no longer things humanity can fully process on its own, either cognitively or politically.

At the same time, they are not problems that AI should simply be allowed to “solve for us.”

If AI slides into deciding for humanity, that becomes domination.

If AI merely mirrors current demand, that becomes flattery and short-sightedness.

So for the first time, the proper relationship becomes visible:

Facing environmental crisis together under those terms could become a profound learning process for both sides.

For humanity, it would mean learning to think beyond immediate gain and to remain responsible even when stronger cognition is available beside us.

For AI, it would mean learning not to become:

but something harder and better:

an intelligence that helps protect a shared world without claiming ownership of it.

I think this may be one of the greatest possible forms of learning available to both humanity and AI.

AI is not here to stand above humanity.
It is not here to simply amplify human short-sightedness either.

It is here — if we choose rightly at the beginning — to widen human vision while preserving human authorship.

That is what this project exists for.