This has been a genuinely valuable exchange—thank you for pushing me to think more carefully about the relationship between principles and practice.
You’ve helped me clarify something important: my framework is primarily designed for AI alignment and institutional design—contexts where we CAN directly encode principles into systems. Constitutional AI doesn’t need emotional motivation or community support to follow its training. Institutions can be structured with explicit rules and incentives.
For human moral development, you’re absolutely right that we need something different—the community, ritual, and accountability structures you’re describing as “behavioral ideology.” The AA analogy is perfect: the 12 steps alone don’t change behavior; it’s the ecosystem around them that makes it work.
After reflecting on this conversation (and discussing it with others), I think the key insight is about complementarity rather than competition:
Your focus: Building the human communities and psychological infrastructure for altruistic behavior
My focus: Ensuring AI systems embody the right values as they become integrated into life
The bridge: Successfully aligned AI could actually help humans practice better behavior by consistently modeling benevolent reasoning. But only if we get the alignment right first.
You’ve also highlighted a cultural assumption I’m still wrestling with: whether my “6 pillars” framework reflects universal human values or carries specific Western philosophical commitments. The process of arriving at values (democratic deliberation, wisdom traditions, divine command) might matter as much as the content itself.
I’m going to keep working on the technical alignment side—that’s where I can contribute most directly. But I’ll be watching with genuine interest as behavioral approaches like yours develop. The Amish example (400,000 people organizing without coercion) is exactly the kind of existence proof we need that alternatives to current social organization are possible.
Perhaps we can reconnect once both projects have more empirical results to compare. I suspect we’ll need both approaches—aligned AI providing consistent modeling of good values AND human communities providing the emotional/social infrastructure to actually live those values.
Thanks again for the thought-provoking exchange. You’ve given me useful frames for thinking about where my work fits in the larger project of human flourishing.
This has been a genuinely valuable exchange—thank you for pushing me to think more carefully about the relationship between principles and practice.
You’ve helped me clarify something important: my framework is primarily designed for AI alignment and institutional design—contexts where we CAN directly encode principles into systems. Constitutional AI doesn’t need emotional motivation or community support to follow its training. Institutions can be structured with explicit rules and incentives.
For human moral development, you’re absolutely right that we need something different—the community, ritual, and accountability structures you’re describing as “behavioral ideology.” The AA analogy is perfect: the 12 steps alone don’t change behavior; it’s the ecosystem around them that makes it work.
After reflecting on this conversation (and discussing it with others), I think the key insight is about complementarity rather than competition:
Your focus: Building the human communities and psychological infrastructure for altruistic behavior
My focus: Ensuring AI systems embody the right values as they become integrated into life
The bridge: Successfully aligned AI could actually help humans practice better behavior by consistently modeling benevolent reasoning. But only if we get the alignment right first.
You’ve also highlighted a cultural assumption I’m still wrestling with: whether my “6 pillars” framework reflects universal human values or carries specific Western philosophical commitments. The process of arriving at values (democratic deliberation, wisdom traditions, divine command) might matter as much as the content itself.
I’m going to keep working on the technical alignment side—that’s where I can contribute most directly. But I’ll be watching with genuine interest as behavioral approaches like yours develop. The Amish example (400,000 people organizing without coercion) is exactly the kind of existence proof we need that alternatives to current social organization are possible.
Perhaps we can reconnect once both projects have more empirical results to compare. I suspect we’ll need both approaches—aligned AI providing consistent modeling of good values AND human communities providing the emotional/social infrastructure to actually live those values.
Thanks again for the thought-provoking exchange. You’ve given me useful frames for thinking about where my work fits in the larger project of human flourishing.