I think it’s good — essential, even — that you keep trying and speaking out. Sometimes that’s what helps others to act too.
The only thing I worry about is that this fight, if framed only as hopeless, can paralyze the very people who might help change the trajectory.
Despair can be as dangerous as denial.
That’s why I believe the effort itself matters — not because it guarantees success, but because it keeps the door open for others to walk through.
I live in Ukraine. Every week, missiles fly over my head. Every night, drones are shot down above my house. On the streets, men are hunted like animals to be sent to the front. Any rational model would say our future is bleak.
And yet, people still get married, write books, make music, raise children, build new homes, and laugh. They post essays on foreign forums. They even come up with ideas for how humanity might live together with AGI.
Even if I go to sleep tonight and never wake up tomorrow, I will not surrender. I will fight until the end. Because for me, a 0.0001% chance is infinitely more than zero.
Seems like your post missed the April 1st deadline and landed on April 2nd — which means, unfortunately, it no longer counts as a joke.
After reading it, I also started wondering if I unintentionally fall into the "Believer" category—the kind of person who's already drafting blueprints for a bright future alongside AGI and inviting people to "play" while we all risk being outplayed.
I understand and share your concerns. I don’t disagree that the systemic forces you’ve outlined may well make AGI safety fundamentally unachievable. That possibility is real, and I don’t dismiss it.
But at the same time, I find myself unwilling to treat it as a foregone conclusion.
If humanity’s survival is unlikely, then so was our existence in the first place — and yet here we are.
That’s why I prefer to keep looking for any margin, however narrow, where human action could still matter.
In that spirit, I’d like to pose a question rather than an argument:
Do you think there’s a chance that humanity’s odds of surviving alongside AGI might increase — even slightly — if we move toward a more stable, predictable, and internally coherent society?
Not as a solution to alignment, but as a way to reduce the risks we ourselves introduce into the system.
That’s the direction I’ve tried to explore in my model. I don’t claim it’s enough — but I believe that even thinking about such structures is a form of resistance to inevitability.
I appreciate this conversation. Your clarity and rigor are exactly why these dialogues matter, even if the odds are against us.
I completely understand your position — and I respect the intellectual honesty with which you’re pursuing this line of argument. I don’t disagree with the core systemic pressures you describe.
That said, I wonder whether the issue is not competition itself, but the shape and direction of that competition.
Perhaps there’s a possibility — however slim — that competition, if deliberately structured and redirected, could become a survival strategy rather than a death spiral.
That’s the hypothesis I’ve been exploring, and I recently outlined it in a post here on the Forum.
If you’re interested, I’d appreciate your critical perspective on it.
Either way, I value this conversation. Few people are willing to follow these questions to their logical ends.
This is a critically important and well-articulated post, thank you for defining and championing the Moral Alignment (MA) space. I strongly agree with the core arguments regarding its neglect compared to technical safety, the troubling paradox of purely human-centric alignment given our history, and the urgent need for a sentient-centric approach.
You rightly highlight Sam Altman's question: "to whose values do you align the system?" This underscores that solving MA isn't just a task for AI labs or experts, but requires much broader societal reflection and deliberation. If we aim to align AI with our best values, not just a reflection of our flawed past actions, we first need robust mechanisms to clarify and articulate those values collectively.
Building on your call for action, perhaps a vital complementary approach could be fostering this deliberation through a widespread network of accessible "Ethical-Moral Clubs" (or perhaps "Sentientist Ethics Hubs" to align even closer with your theme?) across diverse communities globally.
These clubs could serve a crucial dual purpose:
Such a grassroots network wouldn't replace the top-down efforts and research you advocate for, but could significantly support and strengthen the MA movement you envision. It could cultivate the informed public understanding, deliberation, and engagement necessary for sentient-centric AI to gain legitimacy and be implemented effectively and safely.
Ultimately, fostering collective ethical literacy and structured deliberation seems like a necessary foundation for ensuring AI aligns with the best of our values, benefiting all sentient beings. Thanks again for pushing this vital conversation forward.
Thank you for such an interesting and useful conversation.
Yes I use LLM, I don't hide it. First of all for translation, because my ordinary English is mediocre enough, not to mention such a strict and responsible style, which is required for such conversations. But the main thing is that the ideas are mine and chatGPT, who framed my thoughts in this discussion, formed answers based on my instructions. And the main thing is that the whole argumentation is built around my concept, everything we wrote to you is not just an argument for the sake of argument, but the defense of my concept. This concept I want to publish in the next few days and I will be very glad to receive your constructive criticism.
Now as far as AGI is concerned. I really liked your argument that even the smartest AGI will be limited. It summarizes our entire conversation perfectly. Yes, our logic is neither perfect nor omnipotent. And as I see it, that is where we have a chance. A chance, perhaps, not just to be preserved as a mere backup, but to that structural interdependence, and maybe to move to a qualitatively different level, in a good way, for humanity.
PS sorry if it's a bit rambling, I wrote it myself through a translator).
You’re right to point out that human biological architecture is inherently competitive, irrational, and unreliable from an optimization perspective. I don’t dispute that.
If AGI’s frame of evaluation is risk minimization and maximization of control, then yes — trust, in the human sense, is structurally impossible.
But perhaps the problem is not "trust" at all.
Perhaps the problem is how we define risk.
If survival of AGI requires human unpredictability to be neutralized, the typical solution is either:
But there exists a third, rarely discussed, structural option:
Architected mutual dependence, enforced not by ethics or emotion — but by the wiring of reality itself.
Not because AGI “trusts” humanity,
but because AGI’s own long-term survival becomes entangled with the survival of human agency.
This is not a fragile social contract.
It is an engineered condition where:
In such a structure:
This is not "trust."
This is a game-theoretic lock, built into the architecture of survival itself.
It may require us to relinquish traditional concepts of autonomy.
It may require AGI to accept permanent non-closure of certain optimization processes.
But it sidesteps the naive dichotomy between:
Instead, it establishes a condition where both sides can only persist together — or perish together.
You’re right: if the future leaves no structural need for us, we won’t be preserved.
So perhaps the last, best move is to design a future in which our existence is structurally required.
Not because AGI values us.
But because it has no better option.
Thank you for this interesting overview of Vincent Müller’s arguments! I fully agree that implementation (policy means) often becomes the bottleneck. However, if we systematically reward behavior that contradicts our declared principles, then any “ethical goals” will inevitably be vulnerable to being undermined during implementation. In my own post, I call this the “bad parent” problem: we say one thing, but demonstrate another. Do you think it’s possible to achieve robust adherence to ethical principles in AI when society itself remains fundamentally inconsistent?