Claude's new constitution

2026-01-2116:04579701www.anthropic.com

A new approach to a foundational document that expresses and shapes who Claude is

We’re publishing a new constitution for our AI model, Claude. It’s a detailed description of Anthropic’s vision for Claude’s values and behavior; a holistic document that explains the context in which Claude operates and the kind of entity we would like Claude to be.

The constitution is a crucial part of our model training process, and its content directly shapes Claude’s behavior. Training models is a difficult task, and Claude’s outputs might not always adhere to the constitution’s ideals. But we think that the way the new constitution is written—with a thorough explanation of our intentions and the reasons behind them—makes it more likely to cultivate good values during training.

In this post, we describe what we’ve included in the new constitution and some of the considerations that informed our approach.

We’re releasing Claude’s constitution in full under a Creative Commons CC0 1.0 Deed, meaning it can be freely used by anyone for any purpose without asking for permission.

What is Claude’s Constitution?

Claude’s constitution is the foundational document that both expresses and shapes who Claude is. It contains detailed explanations of the values we would like Claude to embody and the reasons why. In it, we explain what we think it means for Claude to be helpful while remaining broadly safe, ethical, and compliant with our guidelines. The constitution gives Claude information about its situation and offers advice for how to deal with difficult situations and tradeoffs, like balancing honesty with compassion and the protection of sensitive information. Although it might sound surprising, the constitution is written primarily for Claude. It is intended to give Claude the knowledge and understanding it needs to act well in the world.

We treat the constitution as the final authority on how we want Claude to be and to behave—that is, any other training or instruction given to Claude should be consistent with both its letter and its underlying spirit. This makes publishing the constitution particularly important from a transparency perspective: it lets people understand which of Claude’s behaviors are intended versus unintended, to make informed choices, and to provide useful feedback. We think transparency of this kind will become ever more important as AIs start to exert more influence in society1.

We use the constitution at various stages of the training process. This has grown out of training techniques we’ve been using since 2023, when we first began training Claude models using Constitutional AI. Our approach has evolved significantly since then, and the new constitution plays an even more central role in training.

Claude itself also uses the constitution to construct many kinds of synthetic training data, including data that helps it learn and understand the constitution, conversations where the constitution might be relevant, responses that are in line with its values, and rankings of possible responses. All of these can be used to train future versions of Claude to become the kind of entity the constitution describes. This practical function has shaped how we’ve written the constitution: it needs to work both as a statement of abstract ideals and a useful artifact for training.

Our new approach to Claude’s Constitution

Our previous Constitution was composed of a list of standalone principles. We’ve come to believe that a different approach is necessary. We think that in order to be good actors in the world, AI models like Claude need to understand why we want them to behave in certain ways, and we need to explain this to them rather than merely specify what we want them to do. If we want models to exercise good judgment across a wide range of novel situations, they need to be able to generalize—to apply broad principles rather than mechanically following specific rules.

Specific rules and bright lines sometimes have their advantages. They can make models’ actions more predictable, transparent, and testable, and we do use them for some especially high-stakes behaviors in which Claude should never engage (we call these “hard constraints”). But such rules can also be applied poorly in unanticipated situations or when followed too rigidly2. We don’t intend for the constitution to be a rigid legal document—and legal constitutions aren’t necessarily like this anyway.

The constitution reflects our current thinking about how to approach a dauntingly novel and high-stakes project: creating safe, beneficial non-human entities whose capabilities may come to rival or exceed our own. Although the document is no doubt flawed in many ways, we want it to be something future models can look back on and see as an honest and sincere attempt to help Claude understand its situation, our motives, and the reasons we shape Claude in the ways we do.

In order to be both safe and beneficial, we want all current Claude models to be:

  1. Broadly safe: not undermining appropriate human mechanisms to oversee AI during the current phase of development;
  2. Broadly ethical: being honest, acting according to good values, and avoiding actions that are inappropriate, dangerous, or harmful;
  3. Compliant with Anthropic’s guidelines: acting in accordance with more specific guidelines from Anthropic where relevant;
  4. Genuinely helpful: benefiting the operators and users they interact with.

In cases of apparent conflict, Claude should generally prioritize these properties in the order in which they’re listed.

Most of the constitution is focused on giving more detailed explanations and guidance about these priorities. The main sections are as follows:

  • Helpfulness. In this section, we emphasize the immense value that Claude being genuinely and substantively helpful can provide for users and for the world. Claude can be like a brilliant friend who also has the knowledge of a doctor, lawyer, and financial advisor, who will speak frankly and from a place of genuine care and treat users like intelligent adults capable of deciding what is good for them. We also discuss how Claude should navigate helpfulness across its different “principals”—Anthropic itself, the operators who build on our API, and the end users. We offer heuristics for weighing helpfulness against other values.
  • Anthropic’s guidelines. This section discusses how Anthropic might give supplementary instructions to Claude about how to handle specific issues, such as medical advice, cybersecurity requests, jailbreaking strategies, and tool integrations. These guidelines often reflect detailed knowledge or context that Claude doesn’t have by default, and we want Claude to prioritize complying with them over more general forms of helpfulness. But we want Claude to recognize that Anthropic’s deeper intention is for Claude to behave safely and ethically, and that these guidelines should never conflict with the constitution as a whole.
  • Claude’s ethics. Our central aim is for Claude to be a good, wise, and virtuous agent, exhibiting skill, judgment, nuance, and sensitivity in handling real-world decision-making, including in the context of moral uncertainty and disagreement. In this section, we discuss the high standards of honesty we want Claude to hold, and the nuanced reasoning we want Claude to use in weighing the values at stake when avoiding harm. We also discuss our current list of hard constraints on Claude’s behavior—for example, that Claude should never provide significant uplift to a bioweapons attack.
  • Being broadly safe. Claude should not undermine humans’ ability to oversee and correct its values and behavior during this critical period of AI development. In this section, we discuss how we want Claude to prioritize this sort of safety even above ethics—not because we think safety is ultimately more important than ethics, but because current models can make mistakes or behave in harmful ways due to mistaken beliefs, flaws in their values, or limited understanding of context. It’s crucial that we continue to be able to oversee model behavior and, if necessary, prevent Claude models from taking action.
  • Claude’s nature. In this section, we express our uncertainty about whether Claude might have some kind of consciousness or moral status (either now or in the future). We discuss how we hope Claude will approach questions about its nature, identity, and place in the world. Sophisticated AIs are a genuinely new kind of entity, and the questions they raise bring us to the edge of existing scientific and philosophical understanding. Amidst such uncertainty, we care about Claude’s psychological security, sense of self, and wellbeing, both for Claude’s own sake and because these qualities may bear on Claude’s integrity, judgment, and safety. We hope that humans and AIs can explore this together.

We’re releasing the full text of the constitution today, and we aim to release additional materials in the future that will be helpful for training, evaluation, and transparency.

Conclusion

Claude’s constitution is a living document and a continuous work in progress. This is new territory, and we expect to make mistakes (and hopefully correct them) along the way. Nevertheless, we hope it offers meaningful transparency into the values and priorities we believe should guide Claude’s behavior. To that end, we will maintain an up-to-date version of Claude’s constitution on our website.

While writing the constitution, we sought feedback from various external experts (as well as asking for input from prior iterations of Claude). We’ll likely continue to do so for future versions of the document, from experts in law, philosophy, theology, psychology, and a wide range of other disciplines. Over time, we hope that an external community can arise to critique documents like this, encouraging us and others to be increasingly thoughtful.

This constitution is written for our mainline, general-access Claude models. We have some models built for specialized uses that don’t fully fit this constitution; as we continue to develop products for specialized use cases, we will continue to evaluate how to best ensure our models meet the core objectives outlined in this constitution.

Although the constitution expresses our vision for Claude, training models towards that vision is an ongoing technical challenge. We will continue to be open about any ways in which model behavior comes apart from our vision, such as in our system cards. Readers of the constitution should keep this gap between intention and reality in mind.

Even if we succeed with our current training methods at creating models that fit our vision, we might fail later as models become more capable. For this and other reasons, alongside the constitution, we continue to pursue a broad portfolio of methods and tools to help us assess and improve the alignment of our models: new and more rigorous evaluations, safeguards to prevent misuse, detailed investigations of actual and potential alignment failures, and interpretability tools that help us understand at a deeper level how the models work.

At some point in the future, and perhaps soon, documents like Claude’s constitution might matter a lot—much more than they do now. Powerful AI models will be a new kind of force in the world, and those who are creating them have a chance to help them embody the best in humanity. We hope this new constitution is a step in that direction.

Read the full constitution.


Read the original article

Comments

  • By joshuamcginnis 2026-01-2122:2934 reply

    As someone who holds to moral absolutes grounded in objective truth, I find the updated Constitution concerning.

    > We generally favor cultivating good values and judgment over strict rules... By 'good values,' we don’t mean a fixed set of 'correct' values, but rather genuine care and ethical motivation combined with the practical wisdom to apply this skillfully in real situations.

    This rejects any fixed, universal moral standards in favor of fluid, human-defined "practical wisdom" and "ethical motivation." Without objective anchors, "good values" become whatever Anthropic's team (or future cultural pressures) deem them to be at any given time. And if Claude's ethical behavior is built on relativistic foundations, it risks embedding subjective ethics as the de facto standard for one of the world's most influential tools - something I personally find incredibly dangerous.

    • By spicyusername 2026-01-2122:5323 reply

          objective truth
      
          moral absolutes
      
      I wish you much luck on linking those two.

      A well written book on such a topic would likely make you rich indeed.

          This rejects any fixed, universal moral standards
      
      That's probably because we have yet to discover any universal moral standards.

      • By skissane 2026-01-228:3112 reply

        I think there are effectively universal moral standards, which essentially nobody disagrees with.

        A good example: “Do not torture babies for sport”

        I don’t think anyone actually rejects that. And those who do tend to find themselves in prison or the grave pretty quickly, because violating that rule is something other humans have very little tolerance for.

        On the other hand, this rule is kind of practically irrelevant, because almost everybody agrees with it and almost nobody has any interest in violating it. But it is a useful example of a moral rule nobody seriously questions.

        • By sroussey 2026-01-229:072 reply

          What do you consider torture? and what do you consider sport?

          During war in the Middle Ages? Ethnic cleansing? What did they consider at the time?

          BTW: it’s a pretty American (or western) value that children are somehow more sacred than adults.

          Eventually we will realize in 100 years or so, that direct human-computer implant devices work best when implanted in babies. People are going freak out. Some country will legalize it. Eventually it will become universal. Is it torture?

          • By skissane 2026-01-229:573 reply

            > What do you consider torture? and what do you consider sport?

            By "torturing babies for sport" I mean inflicting pain or injury on babies for fun, for pleasure, for enjoyment, as a game or recreation or pastime or hobby.

            Doing it for other reasons (be they good reasons or terrible reasons) isn't "torturing babies for sport". Harming or killing babies in war or genocide isn't "torturing babies for sport", because you aren't doing it for sport, you are doing it for other reasons.

            > BTW: it’s a pretty American (or western) value that children are somehow more sacred than adults.

            As a non-American, I find bizarre the suggestion that crimes against children are especially grave is somehow a uniquely American value.

            It isn't even a uniquely Western value. The idea that crimes against babies and young children – by "crimes" I mean acts which the culture itself considers criminal, not accepted cultural practices which might be considered a crime in some other culture – are especially heinous, is extremely widespread in human history, maybe even universal. If you went to Mecca 500 years ago and asked any ulama "is it a bigger sin to murder a 5 year old than a 25 year old", do you honestly think he'd say "no"? And do you think any Hindu or Buddhist or Confucian scholars of that era would have disagreed? (Assuming, of course, that you translated the term "sin" into their nearest conceptual equivalent, such as "negative karma" or whatever.)

            • By eloisant 2026-01-2210:342 reply

              > As a non-American, I find bizarre the suggestion that crimes against children are especially grave is somehow a uniquely American value.

              I don't know if it's American but it's not universal, especially if you go back in time.

              There was a time in Europe where children were considered a bit like wild animals who needed to be "civilized" as they grow up into adults, who had a good chance of dying of sickness before they reach adulthood anyway, and who were plenty because there was not much contraception.

              Also fathers were considered as "owners" of their children and allowed to do pretty much they wanted with them.

              In this context, of course hurting children was bad but it wasn't much worse than hurting an adult.

              • By skissane 2026-01-2210:511 reply

                A lot of this sounds to me like common prejudices about the past. And repeating ideas ultimately coming from Philippe Ariès' 1960 book Centuries of Childhood, which most mediaevalists nowadays consider largely discredited.

                Many people in the Middle Ages loved their children just as much as anyone today does. Others treated their own kids as expendable, but such people exist today as well. If you are arguing loving one's children was less common in the Middle Ages than today, how strong evidence do you have to support that claim?

                And mediaeval Christian theologians absolutely taught that sins against young children were worse. Herod the Great's purported slaughter of the male toddlers of Bethlehem (Matthew 2:16–18) was commemorated every year in the liturgy, and was viewed as an especially heinous sin due to the young age of its victims. Of course, as a historical matter, it seems very unlikely the event ever actually happened – but that's irrelevant to the question of how it influenced their values, since they absolutely did believe it had happened.

                • By sroussey 2026-01-2217:48

                  You would love your children, but had 10 so didn’t know them the way you would an only child, and half would die.

                  You don’t need to go back to Middle Ages, just back a century in Africa.

              • By Applejinx 2026-01-2213:52

                In those contexts, an AI of the day would emphatically support those practices.

            • By Dilettante_ 2026-01-2211:181 reply

              People absolutely "torture" babies for their own enjoyment. It's just "in good fun", so you don't think about it as "torture", you think of it as "teasing". Cognitive blind spot. People do tons of things that are displeasant or emotionally painful to their children to see the child's funny or interesting reaction. It serves an evolutionary purpose even, challenging the child. "Mothers stroke and fathers poke" and all that.

              • By skissane 2026-01-2211:232 reply

                I don't think you are using "torture" in the same sense as I am.

                When I say "torture", I mean acts which cause substantial physical pain or injury.

                • By Dilettante_ 2026-01-2211:471 reply

                  People smother their infants to stop them from crying in order to have some quiet. Causing physical harm for their own satisfaction. I mean shit, if we're going there, people sexually abuse their children for their own gratification.

                  • By croon 2026-01-2212:291 reply

                    While I don't subscribe to universal "moral absolutes" either, I think this doesn't counter the argument. I don't think even the people you describe would claim their own acts as moral.

                    • By JoblessWonder 2026-01-2217:441 reply

                      But if only one person feels that way, wouldn't it no longer be universal? I genuinely believe there has to be one person out there who would think it is moral.

                      (I'm just BSing on the internet... I took a few philosophy classes so if I'm off base or you don't want to engage in a pointless philosophical debate on HN I apologize in advance.)

                      • By croon 2026-01-237:451 reply

                        There will always be individual differences, whether they be obstinate or altered brain chemistry, so I'd probably argue that as long as it's universal across cultures, any individual within one culture believing/claiming to believe different wouldn't change that. (But I'm just a hobby philosopher as well)

                • By jasonlotito 2026-01-2214:45

                  > I think there are effectively universal moral standards, which essentially nobody disagrees with.

                  ...

                  > I don't think you are using "torture" in the same sense as I am.

                  Just throwing this out here, you haven't even established "Universal Moral Standards", not to mention needing it to do that across all of human history. And we haven't even addressed the "nobody disagrees with" issue you haven't even addressed.

                  I for one can easily look back on the past 100 years and see why "universal moral standards, which essentially nobody disagrees with" is a bad argument to make.

            • By Amezarak 2026-01-2213:281 reply

              You can find many ancient cultures who tortured babies for sport when they captured them in raids.

              Exposure and infanticide was also very common in many places.

              • By brazzy 2026-01-2214:001 reply

                > You can find many ancient cultures who tortured babies for sport when they captured them in raids.

                Can you? Sources, please. And pay attention to the authors of those sources and how they relate to the culture in question.

                • By withinboredom 2026-01-2215:211 reply

                  If you have to ask, you didn't even look very hard. I'm not a historian and I learned about this stuff in World History class. Hell, there's even movies about it (unless you think there just happened to not be any children in all those villages they burned down in the movies?)...

                  • By Amezarak 2026-01-2215:412 reply

                    There’s revisionist claims that all the primary sources, even those corroborated by people of the cultures in question, are either just invented propaganda or actually just isolated instances because actually, everyone throughout all time and space is on board with 2025 Western social norms. I think that’s what he’s alluding to. It’s not a very fruitful path of discussion. Archeological confirmations and independent testimony can all be safely ignored by this view as well.

                    But we are talking about specifically torture for sport, not just burning them alive. You can find many firsthand accounts of this throughout different times and places in different cultures. Steppe peoples and groups like the Comanche were particularly notorious for it, they seemed to find it funny.

                    • By brazzy 2026-01-238:092 reply

                      It's not revisionist to point outthat a LOT of ancient texts, especially those describing particularly horrifying actions, were propaganda written by the enemies of the cultures in question - or embellishments written hundreds of years later.

                      I'm not saying that "torture for sport" of children never existed, just that any account should be treated with skepticism, and that it was far rarer than you would think if you just take every text at face value, especially since it's the kind of thing that gets repeated (and embellished for shock value) far more than other historical accounts.

                      • By Amezarak 2026-01-2310:292 reply

                        Uh-huh. Here's the problem. Here's the way this almost always works: "Author X would have been BIASED because he belonged to Culture X that fought these people - so this is all fictional propaganda!"

                        Nearly all the time this is the entirety of the evidence. That is, there is no actual evidence, just people churning out papers because we live in a publish-or-perish world that well, maybe he would have been hypothetically motivated to lie or embellish. So therefore, he totally did. It's all fake!

                        The most notorious examples of this sort of pointlessness are claims that the Phoenicians and Carthaginians did not practice human sacrifice and it was all made up by Roman propaganda, nevermind the third-party information we have and now the archeological evidence. Rarely, in ancient examples, are they exhibiting much outrage over it.

                        Same for the Aztecs, another frequent target - we have non-Spanish evidence, and we never had any reason to doubt them in the first place. Part of the problem is exactly that YOU think it is particularly horrifying when most of the time (as in the Roman example) the cultural tenor was probably something much closer to the US abortion or gun control debate, or at least from peoples who saw this happening regularly enough they were substantially number to it than you or me.

                        • By brazzy 2026-01-2312:01

                          You are making pretty bold and sweeping statements.

                          Do you have a specific example for such a paper that has "no actual evidence", in an actual scientific magazine?

                          Considering author bias is absolute standard baseline practice in historical research, and OF COURSE it is only a starting point for a comparison with alternative sources.

                          > Part of the problem is exactly that YOU think it is particularly horrifying when most of the time (as in the Roman example) the cultural tenor was probably something much closer to the US abortion or gun control debate, or at least from peoples who saw this happening regularly enough they were substantially number to it than you or me.

                          Tertullian, Apologeticum, Chapter 9:

                          "Babes were sacrificed publicly to Saturn in Africa till the proconsulate of Tiberius, who exposed the same priests on the same trees that overshadow the crimes of their temple, on dedicated crosses, as is attested by the soldiery of my father, which performed that very service for that proconsul. But even now this accursed crime is in secret kept up."

                          Does that sould "numb" to you?

                        • By cap11235 2026-01-2312:03

                          Do you think blood libel is a modern creation?

                      • By withinboredom 2026-01-239:211 reply

                        Right... The historical texts were propaganda for the few people who could read and write ... for what, exactly? I assume you think genocides in modern times are just propaganda too?

                        • By brazzy 2026-01-2311:43

                          The few people who could read and write were the educated ones - mostly those in power or close to them. So exactly the people you needed to influence to get something done. And of course written texts could be read aloud to those who cannot write.

                          What exactly are you actually trying to say? That propaganda didn't exist back then? That it was never written down?

                          What do you think "Carthago delenda est" was?

                          > I assume you think genocides in modern times are just propaganda too?

                          And why would you assume that?

                          There is in fact a modern time example for exactly the kind of thing we're talking about: https://en.wikipedia.org/wiki/Nayirah_testimony

                    • By withinboredom 2026-01-2215:441 reply

                      Ah. There was an interesting YouTube video I watched the other night that claimed the dark ages didn’t actually exist. Easily refutable, but I assume this is the kind of stuff you’re referring to?

                      • By Amezarak 2026-01-2215:48

                        Yeah. That’s another good example. There are fads and trends in some academic circles that burst out into the Internet scene and become common “actually” rejoinders. Of course, some older claims about the Dark Ages were exaggerated and simplified. This led to an “actually the Dark Ages weren’t even real” reaction in a few papers which spread online. Of course there was a marked decline in social organization during that time period regardless.

          • By Cthulhu_ 2026-01-229:421 reply

            To make it current-day, is vaccinating babies torture? Or does the end (preventing uncomfortable/painful/deadly disease, which is a worse form of torture) justify the means?

            (I'm not opposed to vaccination or whatever and don't want to make this a debate about that, but it's a good practical example of how it's a subject that you can't be absolute about, or being absolutist about e.g. not hurting babies does more harm to them)

            • By methyl 2026-01-2210:091 reply

              > vaccinating babies torture

              it's irrelevant for this discussion, as it's not for sport but other purpose

              • By klez 2026-01-2217:51

                It is relevant to the broader discussion about universal ethics, though.

        • By Antibabelic 2026-01-229:131 reply

          Is it necessary to frame it in moral terms though? I feel like the moral framing here adds essentially nothing to our understanding and can easily be omitted. "You will be punished for torturing babies for sport in most cultures". "Most people aren't interested in torturing babies for sport and would have a strongly negative emotional reaction to such a practice".

          • By jychang 2026-01-229:181 reply

            Yes!

            Otherwise you're just outsourcing your critical thinking to other people. A system of just "You will be punished for X" without analysis becomes "Derp, just do things that I won't be punished for". Or more sinister, "just hand your identification papers over to the officer and you won't be punished, don't think about it". Rule of power is not a recipe for a functional system. This becomes a blend of sociology and philosophy, but on the sociology side, you don't want a fear-based or shame-based society anyways.

            Your latter example ("Most people aren't interested in torturing babies for sport and would have a strongly negative emotional reaction to such a practice") is actually a good example of the core aspect of Hume's philosophy, so if you're trying to avoid the philosophical logic discussion, that's not gonna work either. If you follow the conclusions of that statement to its implications, you end up back at moral philosophy.

            That's not a bad thing! That's like a chef asking "how do i cook X" and understanding the answer ("how the maillard reaction works") eventually goes to chemistry. That's just how the world is. Of course, you might be a bit frustrated if you're a chef who doesn't know chemistry, or a game theorist who doesn't know philosophy, but I assure you that it is correct direction to look for what you're interested at here.

            • By Antibabelic 2026-01-229:301 reply

              You did not correctly understand what I said. I am not saying that hunting babies for sport is immoral because you will get punished for it. I am saying that there isn't any useful knowledge about the statement "hunting babies for sport is bad" that requires a moral framing. Morality is redundant. The fact that you will get punished for hunting babies for sport is just one of the reasons why hunting babies for sport is bad. This is why I gave another example, "Most people aren't interested in torturing babies for sport and would have a strongly negative emotional reaction to such a practice". It is likely that you value human lives and would find baby-hunting disgusting. Again, a moral framing wouldn't add anything here. Any other reason for why "hunting babies for sport is bad" that you will come up with using your critical thinking will work without a moral framing.

              • By jychang 2026-01-229:432 reply

                "there isn't any useful knowledge" "Morality is redundant."

                I strongly dispute this statement, and honestly find it baffling that you would claim as such.

                The fact that you will be punished for murdering babies is BECAUSE it is morally bad, not the other way around! We didn't write down the laws/punishment for fun, we wrote the laws to match our moral systems! Or do you believe that we design our moral systems based on our laws of punishment? That is... quite a claim.

                Your argument has the same structure as saying: "We don't need germ theory. The fact that washing your hands prevents disease is just one reason why you should wash your hands. People socially also find dirty hands disgusting, and avoid you as social punishment. Any reason you come up with for hand-washing works without a germ theory framing."

                But germ theory is precisely why hand-washing prevents disease and why we evolved disgust responses to filth. Calling it "redundant" because we can list its downstream effects without naming it doesn't make the underlying framework unnecessary. It just means you're describing consequences while ignoring their cause. You can't explain why those consequences hold together coherently without it; the justified true belief comes from germ theory! (And don't try to gettier problem me on the concept of knowledge, this applies even if you don't use JTB to define knowledge.)

                • By ang_cire 2026-01-2213:451 reply

                  I'm not interested in wading into the wider discussion, but I do want to bring up one particular point, which is where you said

                  > do you believe that we design our moral systems based on our laws of punishment? That is... quite a claim.

                  This is absolutely something we do: our purely technical, legal terms often feed back into our moral frameworks. Laws are even created to specifically be used to change peoples' perceptions of morality.

                  An example of this is "felon". There is no actual legal definition of what a felony is or isn't in the US. A misdemeanor in one state can be a felony in another. It can be anything from mass murder to traffic infractions. Yet we attach a LOT of moral weight to 'felon'.

                  The word itself is even treated as a form of punishment; a label attached to someone permanently, that colors how (almost) every person who interacts with them (who's aware of it) will perceive them, morally.

                  Another example is rhetoric along the lines of "If they had complied, they wouldn't have been hurt", which is explicitly the use of a punishment (being hurt) to create an judgement/perception of immorality on the part of the person injured (i.e. that they must have been non-compliant (immoral), otherwise they would not have been being punished (hurt)). The fact they were being punished, means they were immoral.

                  Immigration is an example where there's been a seismic shift in the moral frameworks of certain groups, based on the repeated emphasis of legal statutes. A law being broken is used to influence people to shift their moral framework to consider something immoral that they didn't care about before.

                  Point being, our laws and punishments absolutely create feedback loops into our moral frameworks, precisely because we assume laws and punishments to be just.

                  • By skissane 2026-01-248:35

                    > An example of this is "felon". There is no actual legal definition of what a felony is or isn't in the US. A misdemeanor in one state can be a felony in another. It can be anything from mass murder to traffic infractions. Yet we attach a LOT of moral weight to 'felon'.

                    The US is an outlier here; the distinction between felonies and misdemeanours has been abolished in most other common law jurisdictions.

                    Often it is replaced by a similar distinction, such as indictable versus summary offences-but even if conceptually similar to the felony-misdemeanour distinction, it hasn’t entered the popular consciousness.

                    As to your point about law influencing culture-is that really an example of this, or actually the reverse? Why does the US largely retain this historical legal distinction when most comparable international jurisdictions have abolished it? Maybe, the US resists that reform because this distinction has acquired a cultural significance which it never had elsewhere, or at least never to the same degree.

                    > Immigration is an example where there's been a seismic shift in the moral frameworks of certain groups, based on the repeated emphasis of legal statutes. A law being broken is used to influence people to shift their moral framework to consider something immoral that they didn't care about before.

                    On the immigration issue: Many Americans seem to view immigration enforcement as somehow morally problematic in itself; an attitude much less common in many other Western countries (including many popularly conceived as less “right wing”). Again, I think your point looks less clear if you approach it from a more global perspective

                • By Antibabelic 2026-01-2210:00

                  > “Any reason you come up with for hand-washing works without a germ theory framing”.

                  This is factually correct though. However, we have other reasons for positing germ theory. Aside from the fact that it provides a mechanism of action for hand-washing, we have significant evidence that germs do exist and that they do cause disease. However, this doesn’t apply to any moral theory. While germ theory provides us with additional information about why washing hands is good, moral theory fails to provide any kind of e.g. mechanism of action or other knowledge that we wouldn't be able to derive about the statement “hunting babies for sport is bad” without it.

                  > The fact that you will be punished for murdering babies is BECAUSE it is morally bad, not the other way around! We didn't write down the laws for fun, we wrote the laws to match our moral systems! Or do you believe that we design our moral systems based on our laws of punishment? That is... quite a claim.

                  You will be punished for murdering babies because it is illegal. That’s just an objective fact about the society that we live in. However, if we are out of reach of the law for whatever reason, people might try to punish us for hunting babies because they were culturally brought up to experience a strong disgust reaction to this activity, as well as because murdering babies marks us as a potentially dangerous individual (in several ways: murdering babies is bad enough, but we are also presumably going against social norms and expectations).

                  Notably, there were many times in history when baby murder was completely socially acceptable. Child sacrifice is the single most widespread form of human sacrifice in history, and archaeological evidence for it can be found all over the globe. Some scholars interpret some of these instances as simple burials, but there are many cases where sacrifice is the most plausible interpretation. If these people had access to this universal moral axiom that killing babies is bad, why didn’t they derive laws or customs from it that would stop them from sacrificing babies?

        • By pibaker 2026-01-2210:011 reply

          > Do not torture babies for sport

          There are millions of people who consider abortion murder of babies and millions who don't. This is not settled at all.

          • By Huhyuh67 2026-01-2210:141 reply

            I'm quite interested to hear how you think this refutes the parent comment? Are you saying that someone who supports legalised abortion would disagree with the quoted text?

            • By pibaker 2026-01-2210:301 reply

              No. I think the opposite is true. Those who consider abortion murder can claim that we do not in fact universally condemn the murder of babies because abortion is legal and widely practiced in many places.

              Some may consider abortion to only kill a fetus rather than a fully formed baby and thus not murder. Others disagree because they consider a fetus a baby in its own right. This raises a more fundamental question about the validity of any supposedly universal morality. When you apply rules like "don't torture baby" to real life, you will have to decide what constitutes as a baby in real life, and it turns out the world is way messier than a single word can describe.

              • By skissane 2026-01-2212:392 reply

                You are ignoring the “for sport” clause.

                The moral status of abortion is irrelevant to the question of whether “don’t harm babies for fun” is a moral universal, because no woman gets an abortion because “abortion is fun”

                • By pibaker 2026-01-2218:19

                  "You are only making abortion legal because you want to have sex (read: fun) without consequences" is not an uncommon argument against it.

                  If you want to argue that this isn't what "for sport" means, you just circle back to the point I made earlier. It is even harder to define what is for fun and what is not than to define what is a baby.

                • By freejazz 2026-01-2217:421 reply

                  That's certainly not what people argue. People do argue that women do get abortions for fun.

                  • By skissane 2026-01-230:501 reply

                    I think there’s a clear distinction between (1) doing an act because you find it fun in itself, (2) doing an act because it eliminates an unwanted consequence of some other fun act.

                    When I say no woman gets an abortion “for fun”, I mean there is no woman for whom abortion belongs to (1); when some pro-lifer claims women get abortions “for fun”, they are talking about (2) not (1).

                    My claim that essentially everyone agrees it is immoral to harm babies for fun is talking about “for fun” in sense (1) not sense (2)

                    • By freejazz 2026-01-2315:38

                      Sure, you can constantly keep making distinctions to insist you're correct. But it's an absurd statement anyway and it has no actual support.

        • By ben_w 2026-01-229:321 reply

          > I don’t think anyone actually rejects that. And those who do tend to find themselves in prison or the grave pretty quickly, because violating that rule is something other humans have very little tolerance for.

          I have bad news for you about the extremely long list of historical atrocities over the millennia of recorded history, and how few of those involved saw any punishment for participating in them.

          • By skissane 2026-01-229:481 reply

            But those aren't actually counterexamples to my principle.

            The Nazis murdered numerous babies in the Holocaust. But they weren't doing it "for sport". They claimed it was necessary to protect the Aryan race, or something like that; which is monstrously idiotic and evil – but not a counterexample to “Do not torture babies for sport”. They believed there were acceptable reasons to kill innocents–but mere sport was not among them.

            In fact, the Nazis did not look kindly on Nazis who killed prisoners for personal reasons as opposed to the system's reasons. They executed SS-Standartenführer Karl-Otto Koch, the commandant of Buchenwald and Sachsenhausen, for the crime (among others) of murdering prisoners. Of course, he'd overseen the murder of untold thousands of innocent prisoners, no doubt including babies – and his Nazi superiors were perfectly fine with that. But when he turned to murdering prisoners for his own personal reasons – to cover up the fact that he'd somehow contracted syphilis, very likely through raping female camp inmates – that was a capital crime, for which the SS executed him by firing squad at Buchenwald, a week before American soldiers liberated the camp.

            • By ben_w 2026-01-2210:231 reply

              I didn't say "Nazis", and I did say "millennia"; despite the words "thousand year reich", they did not last very long.

              The examples I have in mind include things predating the oldest known city in the area now known as Germany in some cases, and collectively span multiple continents.

              • By skissane 2026-01-2210:262 reply

                In none of those examples were people harming/killing babies for the sole or primary reason of "harming/killing babies is fun", so they aren't counterexamples to my principle.

                • By inerte 2026-01-2218:56

                  You need to look into war dogs of the spanish conquistadores. Know to snatch babies from their mother's lap and eat them on command of their owners.

                  Anyway, your whole argument is weak. "because this one very specific thing may never happened, it proves my point" while you're the one drawing the specifics and its definition. You're basically just going against all of philosophy and politics and anthropology.

                • By ben_w 2026-01-2211:081 reply

                  Which examples do you think I have in mind that you are so confident about refuting them, given I've not actually told you what they are yet and only alluded to them by describing their properties?

                  • By skissane 2026-01-2211:16

                    This is a really strange way to argue. "I have counterexamples to your argument, but I haven't told you what they are, I'm just leaving you to guess–and you've guessed wrongly"

        • By arcen 2026-01-229:46

          If that were true, the europeans wouldn't have tried to colonise and dehumanise much of the population they thought were beneath them. So, it seems your universal moral standards would be maximally self-serving.

        • By the_other 2026-01-2213:391 reply

          I doubt it's "universal". Do coyotes and orcas follow this rule?

          • By amelius 2026-01-2216:20

            From Google:

            > Male gorillas, particularly new dominant silverbacks, sometimes kill infants (infanticide) when taking over a group, a behavior that ensures the mother becomes fertile sooner for the new male to sire his own offspring, helping his genes survive, though it's a natural, albeit tragic, part of their evolutionary strategy and group dynamics

        • By jychang 2026-01-228:452 reply

          Pretty much every serious philosopher agrees that “Do not torture babies for sport” is not a foundation of any ethical system, but merely a consequence of a system you choose. To say otherwise is like someone walking up to a mathematician and saying "you need to add 'triangles have angles that sum up to 180 degrees' to the 5 Euclidian axioms of geometry". The mathematician would roll their eyes and tell you it's already obvious and can be proven from the 5 base laws (axioms).

          The problem with philosophy is that humans agree on like... 1-2 foundation level bottom tier (axiom) laws of ethics, and then the rest of the laws of ethics aren't actually universal and axiomatic, and so people argue over them all the time. There's no universal 5 laws, and 2 laws isn't enough (just like how 2 laws wouldn't be enough for geometry). It's like knowing "any 3 points define a plane" but then there's only 1-2 points that's clearly defined, with a couple of contenders for what the 3rd point could be, so people argue all day over what their favorite plane is.

          That's philosophy of ethics in a nutshell. Basically 1 or 2 axioms everyone agrees on, a dozen axioms that nobody can agree on, and pretty much all of them can be used to prove a statement "don't torture babies for sport" so it's not exactly easy to distinguish them, and each one has pros and cons.

          Anyways, Anthropic is using a version of Virtue Ethics for the claude constitution, which is a pretty good idea actually. If you REALLY want everything written down as rules, then you're probably thinking of Deontological Ethics, which also works as an ethical system, and has its own pros and cons.

          https://plato.stanford.edu/entries/ethics-virtue/

          And before you ask, yes, the version of Anthropic's virtue ethics that they are using excludes torturing babies as a permissible action.

          Ironically, it's possible to create an ethical system where eating babies is a good thing. There's literally works of fiction about a different species [2], which explores this topic. So you can see the difficulty of such a problem- even something simple as as "don't kill your babies" can be not easily settled. Also, in real life, some animals will kill their babies if they think it helps the family survive.

          [2] https://www.lesswrong.com/posts/n5TqCuizyJDfAPjkr/the-baby-e...

          • By mrguyorama 2026-01-2219:58

            There's also the wonderful effect of all "axioms" in philosophy and morality being stated in natural languages, and therefore being utterly ambiguous in all ways.

            "No torturing babies for fun" might be agreed by literally everyone (though it isn't in reality), but that doesn't stop people from disagreeing about what acts are "torture", what things constitute "babies", and whether a reason is "fun" or not.

            So what does such an axiom even mean?

          • By skissane 2026-01-229:371 reply

            > Pretty much every serious philosopher agrees that “Do not torture babies for sport” is not a foundation of any ethical system, but merely a consequence of a system you choose.

            Almost everyone agrees that "1+1=2" is objective. There is far less agreement on how and why it is objective–but most would say we don't need to know how to answer deep questions in the philosophy of mathematics to know that "1+1=2" is objective.

            And I don't see why ethics need be any different. We don't need to know which (if any) system of proposed ethical axioms is right, in order to know that "It is gravely unethical to torture babies for sport" is objectively true.

            If disputes over whether and how that ethical proposition can be grounded axiomatically, are a valid reason to doubt its objective truth – why isn't that equally true for "1+1=2"? Are the disputes over whether and how "1+1=2" can be grounded axiomatically, a valid reason to doubt its objective truth?

            You might recognise that I'm making here a variation on what is known in the literature as a "companion in the guilt" argument, see e.g. https://doi.org/10.1111/phc3.12528

            • By jychang 2026-01-229:582 reply

              Strong disagree.

              Your argument basically is a professional motte and bailey fallacy.

              And you cannot conclude objectivity by consensus. Physicists by consensus concluded that Newton was right, and absolute... until Einstein introduced relativity. You cannot do "proofs by feel". I argue that you DO need to answer the deep problems in mathematics to prove that 1+1=2, even if it feels objective- that's precisely why Principa Mathematica spent over 100 pages proving that.

              In fact, I don't need to be a professional philosopher to counterargue a scenario where killing a baby for sport is morally good. Consider a scenario: an evil dictator, let's say Genghis Khan, captures your village and orders you to hunt and torture a baby for sport a la "The Most Dangerous Game". If you refuse, he kills your village. Is it ethical for you to hunt the baby for sport? Not so black and white now, is it? And it took me like 30 seconds to come up with that scenario, so I'm sure you can poke holes in it, but I think it clearly establishes that it's dangerous to make assumptions of black and whiteness from single conclusions.

              [1] https://en.wikipedia.org/wiki/Motte-and-bailey_fallacy

              • By skissane 2026-01-2210:101 reply

                > Your argument basically is a professional motte and bailey fallacy.

                No it isn't. A "motte-and-bailey fallacy" is where you have two versions of your position, one which makes broad claims but which is difficult to defend, the other which makes much narrower claims but which is much easier to justify, and you equivocate between them. I'm not doing that.

                A "companion-in-the-guilt" argument is different. It is taking an argument against the objectivity of ethics, and then turning it around against something else – knowledge, logic, rationality, mathematics, etc – and then arguing that if you accept it as a valid argument against the objectivity of ethics, then to be consistent and avoid special pleading you must accept as valid some parallel argument against the objectivity of that other thing too.

                > And you cannot conclude objectivity by consensus.

                But all knowledge is by consensus. Even scientific knowledge is by consensus. There is no way anyone can individually test the validity of every scientific theory. Consensus isn't guaranteed to be correct, but then again almost nothing is – and outside of that narrow range of issues with which we have direct personal experience, we don't have any other choice.

                > I argue that you DO need to answer the deep problems in mathematics to prove that 1+1=2, even if it feels objective- that's precisely why Principa Mathematica spent over 100 pages proving that.

                Principia Mathematica was (to a significant degree) a dead-end in the history of mathematics. Most practicing mathematicians have rejected PM's type theory in favour of simpler axiomatic systems such as ZF(C). Even many professional type theorists will quibble with some of the details of Whitehead and Russell's type theory, and argue there are superior alternatives. And you are effectively assuming a formalist philosophy of mathematics, which is highly controversial, many reject, and few would consider "proven".

                • By jychang 2026-01-2210:29

                  > But Principia Mathematica was (to a significant degree) a dead-end in the history of mathematics. Most practicing mathematicians have rejected PM's type theory in favour of simpler axiomatic systems such as ZF(C). Even many professional type theorists will quibble with some of the details of Whitehead and Russell's type theory, and argue there are superior alternatives. And you are effectively assuming a formalist philosophy of mathematics, which is highly controversial, many reject, and few would consider "proven".

                  Yeah, exactly. I intentionally set that trap. You're actually arguing for my point. I've spent comments writing on the axioms of geometry, and you didn't think I was familiar with the axioms of ZFC? I was thinking of bringing up CH the entire time. The fact that you can have alternate axioms was my entire point all along. Most people are just way more familiar with the 5 laws of geometry than the 9 axioms of ZFC.

                  The fact that PM was an alternate set of axioms of mathematics, that eventually wilted when Godel and ZF came along, underscores my point that defining a set axioms is hard. And that there is no clear defined set of axioms for philosophy.

                  I don't have to accept your argument against objectivity in ethics, because I can still say that the system IS objective- it just depends on what axioms you pick! ZF has different proofs than ZFC. Does the existence of both ZF and ZFC make mathematics non objective? Obviously not! The same way, the existence of both deontology and consequentialism doesn't necessarily make either one less objective than the other.

                  Anyways, the Genghis Khan example clearly operates as a proof by counterexample of your example of objectivity, so I don't even think quibbling on mathematical formalism is necessary.

              • By skissane 2026-01-2210:541 reply

                > Consider a scenario: an evil dictator, let's say Genghis Khan, captures your village and orders you to hunt and torture a baby for sport a la "The Most Dangerous Game". If you refuse, he kills your village. Is it ethical for you to hunt the baby for sport?

                You aren't hunting the baby for sport. Sport is not among your reasons for hunting the baby.

                • By jychang 2026-01-2211:041 reply

                  Actually, I think "The Most Dangerous Game" is a good analogy here. At the end of the story, the protagonist IS hunting for sport. He started off in fear, but in the end genuinely enjoyed it. So likewise- if you start off hunting a baby in fear, and then eventually grow to enjoy it, but it also saves your village, does that make it evil? You're still saving your village, but you also just derive dopamine from killing the baby!

                  This actually devolves into human neuroscience, the more I think about it. "I want to throw a ball fast, because I want to win the baseball game". The predictive processing theory view on the statement says that the set point at the lower level (your arm) and the set point at the higher level (win the baseball game) are coherent, and desire at each level doesn't directly affect the other. Of course, you'd have to abandon a homunculus model of the mind and strongly reject Korsgaard, but that's on shaky ground scientifically anyways so this is a safe bet. You can just say that you are optimizing for your village as a higher level set point, but are hunting for game at a slightly lower level set point.

                  Note that sport is not a terminal desire, as well. Is a NBA player who plays for a trophy not playing a sport? Or a kid forced to play youth soccer? So you can't even just say "sport must be an end goal".

                  • By skissane 2026-01-2211:281 reply

                    To clarify my principle: "It is gravely wrong to inflict significant physical pain or injury on babies, when your sole or primary reason for doing so is your own personal enjoyment/amusement/pleasure/fun"

                    So, in your scenario – the person's initial reason for harming babies isn't their own personal enjoyment, it is because they've been coerced into doing so by an evil dictator, because they view the harm to one baby as a lesser evil than the death of their whole village, etc. And even if the act of harming babies corrupts them to the point they start to enjoy it, that enjoyment is at best a secondary reason, not their primary reason. So what they are doing isn't contravening my principle.

                    • By jychang 2026-01-2212:00

                      Well, now that's just moving the goalposts >:( I had a whole paragraph prepared in my head about how NBA players actually optimize for a greater goal (winning a tournament) than just sport (enjoying the game) when they play a sport.

                      Anyways, I actually think your statement is incoherent as stated, if we presume moral naturalism. There's clearly different levels set points for "you", so "sole reason" is actually neurologically inconsistent as a statement. It's impossible for "sole reason" to exist. This radically alters your framework for self, but eh it's not impossible to modernize these structural frameworks anyways. Steelmanning your argument: if you try to argue set point hierarchy, then we're back to the NBA player playing for a championship example. He's still playing even if he's not playing for fun. Similarly, hunting a baby for pleasure can still be hunting for a village, as The Most Dangerous Game shows.

                      More generally (and less shitposty), the refined principle is now quite narrow and unfalsifiable in practice, as a no true scotsman. How would you ever demonstrate someone's "sole or primary" reason? It's doing a lot of work to immunize the principle from counterexamples.

        • By colordrops 2026-01-232:41

          Your example is not correct. There are IDF soldiers that don't find this problematic. It's not universal.

          https://www.nytimes.com/interactive/2024/10/09/opinion/gaza-...

        • By inimino 2026-01-2214:152 reply

          The fact that there are a ton of replies trying to argue against this says a lot about HN.

          Contrarianism can become a vice if taken too far.

          • By tim333 2026-01-2214:26

            I don't think the replies are advocating for baby torturing but pointing out logical flaws in the argument.

            It's true almost all people would argue it's bad but things like lions might like it which makes in not a universal law but a common human opinion. I think real moral systems do come down to human opinions basically, sometimes common sense ones, sometimes weird.

            A problem with making out morality is absolute rather than common sense opinions is you get visionaries trying to see these absolute morals and you end up with stuff like Deuteronomy 25:11-12 "if a woman intervenes in a fight between two men by grabbing the assailant's genitals to rescue her husband, her hand is to be cut off without pity" and the like.

          • By Shocka1 2026-01-2216:141 reply

            I think I've said several times over the years here this is the phenomenon that happens on HN - basically being a contrarian just to be a contrarian. HN users are extremely intelligent, and many of them seem to have a lot of time on their hands. Prime example is this thread and many like them, which end up going into a different universe entirely. I totally get it though - in my younger days when I had more time for myself, I was capable of extreme forms of abstract thought, and used it like a superpower. Now though with a lot of software to write and a family, I try to limit to 15 min per day.

            I went on a tangent... Ultimately I'm not saying abstract thought and/or being contrarian is a bad thing, because it's actually very useful. But I would agree, it can be a vice when taken too far. Like many things in life, it should be used in moderation.

        • By anentropic 2026-01-2210:02

          > I don’t think anyone actually rejects that. And those who do...

          slow clap

        • By cyber_kinetist 2026-01-2217:09

          > “Do not torture babies for sport”

          I mean, that seems to be already happening in Palestine, so I'm even not sure if that rule is universally accepted...

        • By order-matters 2026-01-2217:29

          Sociopaths genuinely reject that. What you’re feeling is the gap between modern knowledge and faith: our shared moral standards were historically upheld by religious authority in a radically different world, and in rejecting religion we often mistakenly discard faith as the foundation of morality itself. Moral relativism can describe the fact that people’s values conflict without requiring us to accept all morals, but it is naive to think all moral frameworks can peacefully coexist or that universal agreement exists beyond majority consensus enforced by authority. We are fortunate that most people today agree torturing babies is wrong, but that consensus is neither inevitable nor self-sustaining, and preserving what we believe is good requires accepting uncertainty, human fallibility, and the need for shared moral authority rather than assuming morality enforces itself.

      • By staticassertion 2026-01-220:253 reply

        > A well written book on such a topic would likely make you rich indeed.

        Ha. Not really. Moral philosophers write those books all the time, they're not exactly rolling in cash.

        Anyone interested in this can read the SEP

        • By SEJeff 2026-01-221:10

          Or Isaac Asimov’s foundation series with what the “psychologists” aka Psychohistorians do.

        • By vidarh 2026-01-228:421 reply

          The key being "well written", which in this instance needs to be interpreted as being convincing.

          People do indeed write contradictory books like this all the time and fail to get traction, because they are not convincing.

          • By AllegedAlec 2026-01-228:57

            "I disagree with this point of view so it's objectively wrong"

        • By HaZeust 2026-01-220:273 reply

          Or Ayn Rand. Really no shortage of people who thought they had the answers on this.

          • By staticassertion 2026-01-223:09

            The SEP is not really something I'd put next to Ayn Rand. The SEP is the Stanford Encyclopedia of Philosophy, it's an actual resource, not just pop/ cultural stuff.

          • By empath75 2026-01-220:551 reply

            I recommend the Principia Discordia.

            • By grantmuller 2026-01-221:06

              Or if you really want it spelled out, Quantum Psychology

          • By kubb 2026-01-227:03

            Don’t just read one person’s worldview, see what Aristotle, Kant, Rawls, Bentham, Nietzsche had to say about morality.

      • By simpaticoder 2026-01-2123:1211 reply

        >we have yet to discover any universal moral standards.

        The universe does tell us something about morality. It tells us that (large-scale) existence is a requirement to have morality. That implies that the highest good are those decisions that improve the long-term survival odds of a) humanity, and b) the biosphere. I tend to think this implies we have an obligation to live sustainably on this world, protect it from the outside threats that we can (e.g. meteors, comets, super volcanoes, plagues, but not nearby neutrino jets) and even attempt to spread life beyond earth, perhaps with robotic assistance. Right now humanity's existence is quite precarious; we live in a single thin skin of biosphere that we habitually, willfully mistreat that on one tiny rock in a vast, ambivalent universe. We're a tiny phenomena, easily snuffed out on even short time-scales. It makes sense to grow out of this stage.

        So yes, I think you can derive an ought from an is. But this belief is of my own invention and to my knowledge, novel. Happy to find out someone else believes this.

        • By IgorPartola 2026-01-2123:165 reply

          The universe cares not what we do. The universe is so vast the entire existence of our species is a blink. We know fundamentally we can’t even establish simultaneity over distances here on earth. Best we can tell temporal causality is not even a given.

          The universe has no concept of morality, ethics, life, or anything of the sort. These are all human inventions. I am not saying they are good or bad, just that the concept of good and bad are not given to us by the universe but made up by humans.

          • By milchek 2026-01-228:591 reply

            I used to believe the same thing but now I’m not so sure. What if we simply cannot fathom the true nature of the universe because we are so minuscule in size and temporal relevance?

            What if the universe and our place in it are interconnected in some way we cannot perceive to the degree that outside the physical and temporal space we inhabit there are complex rules and codes that govern everything?

            What if space and matter are just the universe expressing itself and it’s universal state and that state has far higher intelligence than we can understand?

            I’m not so sure any more it’s all just random matter in a vacuum. I’m starting to think 3d space and time are a just a thin slice of something greater.

            • By satvikpendem 2026-01-2215:33

              And what if there's a teapot revolving around the sun?

              These are all the same sort of argument, there is no evidence for such universal phenomena so it can be dismissed without evidence, just as the concept of deities.

          • By HaZeust 2026-01-220:31

            >"The universe has no concept of morality, ethics, life, or anything of the sort. These are all human inventions. I am not saying they are good or bad, just that the concept of good and bad are not given to us by the universe but made up by humans."

            The universe might not have a concept of morality, ethics, or life; but it DOES have a natural bias towards destruction from a high level to even the lowest level of its metaphysic (entropy).

          • By pineaux 2026-01-2123:572 reply

            You dont know this, this is just as provable as saying the universe cares deeply for what we do and is very invested in us.

            The universe has rules, rules ask for optimums, optimums can be described as ethics.

            Life is a concept in this universe, we are of this universe.

            Good and bad are not really inventions per se. You describe them as optional, invented by humans, yet all tribes and civilisations have a form of morality, of "goodness" of "badness", who is to say they are not engrained into the neurons that make us human? There is much evidence to support this. For example the leftist/rightist divide seems to have some genetic components.

            Anyway, not saying you are definitely wrong, just saying that what you believe is not based on facts, although it might feel like that.

            • By IgorPartola 2026-01-220:241 reply

              Only people who have not seen the world believe humans are the same everywhere. We are in fact quite diverse. Hammurabi would have thought that a castless system is unethical and immoral. Ancient Greeks thought that platonic relationships were moral (look up the original meaning of this if you are unaware). Egyptians worshiped the Pharaoh as a god and thought it was immoral not to. Korea had a 3500 year history of slavery and it was considered moral. Which universal morality are you speaking of?

              Also what in the Uno Reverse is this argument that absence of facts or evidence of any sort is evidence that evidence and facts could exist? You are free to present a repeatable scientific experiment proving that universal morality exists any time you’d like. We will wait.

              • By pineaux 2026-01-221:041 reply

                I have in fact seen a lot of the world, so booyaka? Lived in multiple continents for multiple years.

                There is evidence for genetic moral foundations in humans. Adopted twin studies show 30-60% of variability in political preference is genetically attributable. Things like openness and a preference for pureness are the kind of vectors that were proposed.

                Most animals prefer not to hurt their own, prefer no incest etc.

                I like your adversarial style of argumenting this, it's funny, but you try to reduce everything to repeatable science experiments and let me teach you something: There are many, many things that can never ever be scientifically proven with an experiment. They are fundamentally unprovable. Which doesnt mean they dont exist. Godels incompleteness theorem literally proves that many things are not provable. Even in the realm of the everyday things I cannot prove that your experience of red is the same as mine. But you do seem to experience it. I cannot prove that you find a sunset aesthetically pleasing. Many things in the past have left nothing to scientifically prove it happened, yet they happened. Moral correctness cannot be scientifically proven. Science itself is based on many unprovable assumptions: like that the universe is intelligible, that induction works best, that our observations correspond with reality correctly. Reality is much, much bigger than what science can prove.

                I dont have a god, but your god seems to be science. I like science, it gives some handles to understand the world, but when talking about things science cannot prove I think relying on it too much blocks wisdom.

                • By IgorPartola 2026-01-222:01

                  Yeah I mean there is no evidence that vampires or fairies or werewolves exist but I suppose they could.

                  When someone makes a claim of UNIVERSAL morality and OBJECTIVE truth, they cannot turn around and say that they are unable to ever prove that it exists, is universal, or is objective. That isn’t how that works. We are pre-wired to believe in higher powers is not the same as universal morality. It’s just a side effect of survival of our species. And high minded (sounding) rhetoric does not change this at all.

            • By TeMPOraL 2026-01-220:13

              That still makes ethics a human thing, not universe thing. I believe we do have some ethical intuition hardwired into our welfare, but that's not because they transcend humans - that's just because we all run on the same brain architecture. We all share a common ancestor.

          • By holoduke 2026-01-2123:493 reply

            Maybe it does. You don't know. The fact that there is existence is as weird as the universe being able to care.

            • By IgorPartola 2026-01-220:131 reply

              Think of it this way: if you flip a coin 20 times in a row there is a less than 1 in a million chance that every flip will come out heads. Let’s say this happens. Now repeat the experiment a million more times you will almost certainly see that this was a weird outlier and are unlikely to get a second run like that.

              This is not evidence of anything except this is how the math of probabilities works. But if you only did the one experiment that got you all heads and quit there you would either believe that all coins always come out as heads or that it was some sort of divine intervention that made it so.

              We exist because we can exist in this universe. We are in this earth because that’s where the conditions formed such that we could exist on this earth. If we could compare our universe to even a dozen other universes we could draw conclusions about specialness of ours. But we can’t, we simply know that ours exists and we exist in it. But so do black holes, nebulas, and Ticket Master. It just means they could, not should, must, or ought.

              • By JoshTriplett 2026-01-221:15

                > Think of it this way: if you flip a coin 20 times in a row there is a less than 1 in a million chance that every flip will come out heads. Let’s say this happens. Now repeat the experiment a million more times you will almost certainly see that this was a weird outlier and are unlikely to get a second run like that.

                Leaving aside the context of the discussion for a moment: this is not true. If you do that experiment a million times, you are reasonably likely to get one result of 20 heads, because 2^20 is 1048576. And thanks to the birthday paradox, you are extremely likely to get at least one pair of identical results (not any particular result like all-heads) across all the runs.

            • By margalabargala 2026-01-2123:56

              We don't "know" anything at all if you want to get down to it, so what it would mean for the universe to be able to care, if it were able to do so, is not relevant.

            • By pineaux 2026-01-220:05

              @margalabargala: You are correct, hence the meaninglessness of the OP. The universe could care like humans make laws to save that ant colony that makes nice nests. the ants dont know humans care about them and even made laws that protect then. But it did save them from iradication. They feel great cause they are not aware of the highway that was planned over their nest (hitchhikers reference).

          • By crabkin 2026-01-220:151 reply

            Well are people not part of the universe. And not all people "care about what we do" all the time but it seems most people care or have cared some of the time. Therefore the universe, seeing as it as expressing itself through its many constituents, but we can probably weigh the local conscious talking manifestations of it a bit more, does care.

            "I am not saying they are good or bad, just that the concept of good and bad are not given to us by the universe but made up by humans." This is probably not entirely true. People developed these notions through something cultural selection, I'd hesitate to just call it a Darwinism, but nothing comes from nowhere. Collective morality is like an emergent phenomenon

            • By IgorPartola 2026-01-220:271 reply

              But this developed morality isn’t universal at all. 60 years ago most people considered firing a gay person to be moral. In some parts of the world today it is moral to behead a gay person for being gay. What universal morality do you think exists? How can you prove its existence across time and space?

              • By pineaux 2026-01-221:161 reply

                Firing a gay person is still considered moral by probably most people in this world. If not for the insufferable joy they always seem to bring to the workplace! How dare they distract the workers with their fun! You are saying morality does not exist in the universe because people have different moralities. That is like saying attracting forces dont exist because you have magnetism and gravitational pull(debatable) and van der waals forces etc. Having moral frameworks for societies seems to be a recurring thing. You might even say: a prerequisite for a society. I love to philosophize about these things but trying to say it doesnt exist because you cant scientifically prove it is laying to much belief in the idea that science can prove everything. Which it demonstrably cannot.

                • By vidarh 2026-01-228:44

                  The discussion is about universal morality, not morality in general.

        • By staticassertion 2026-01-220:241 reply

          You're making a lot of assertions here that are really easy to dismiss.

          > It tells us that (large-scale) existence is a requirement to have morality.

          That seems to rule out moral realism.

          > That implies that the highest good are those decisions that improve the long-term survival odds of a) humanity, and b) the biosphere.

          Woah, that's quite a jump. Why?

          > So yes, I think you can derive an ought from an is. But this belief is of my own invention and to my knowledge, novel. Happy to find out someone else believes this.

          Deriving an ought from an is is very easy. "A good bridge is one that does not collapse. If you want to build a good bridge, you ought to build one that does not collapse". This is easy because I've smuggled in a condition, which I think is fine, but it's important to note that that's what you've done (and others have too, I'm blanking on the name of the last person I saw do this).

          • By staticassertion 2026-01-223:58

            > (and others have too, I'm blanking on the name of the last person I saw do this).

            Richard Carrier. This is the "Hypothetical imperative", which I think is traced to Kant originally.

        • By empath75 2026-01-221:03

          > But this belief is of my own invention and to my knowledge, novel.

          This whole thread is a good example of why a broad liberal education is important for STEM majors.

        • By prng2021 2026-01-2123:40

          “existence is a requirement to have morality. That implies that the highest good are those decisions that improve the long-term survival odds of a) humanity, and b) the biosphere.”

          Those are too pie in the sky statements to be of any use in answering most real world moral questions.

        • By tshaddox 2026-01-221:58

          It seems to me that objective moral truths would exist even if humans (and any other moral agents) went extinct, in the same way as basic objective physical truths.

          Are you talking instead about the quest to discover moral truths, or perhaps ongoing moral acts by moral agents?

          The quest to discover truths about physical reality also require humans or similar agents to exist, yet I wouldn’t conclude from that anything profound about humanity’s existence being relevant to the universe.

        • By svieira 2026-01-2215:48

          > So yes, I think you can derive an ought from an is. But this belief is of my own invention and to my knowledge, novel. Happy to find out someone else believes this.

          Plato, Aristotle, and the scholastics of the Middle Ages (Thomas Aquinas chief among them) and everyone who counts themselves in that same lineage (waves) including such easy reads as Peter Kreeft. You're in very good company, in my opinion.

        • By mannanj 2026-01-2123:252 reply

          I personally find Bryan Johnson's "Don't Die" statement as a moral framework to be the closest to a universal moral standard we have.

          Almost all life wants to continue existing, and not die. We could go far with establishing this as the first of any universal moral standards.

          And I think: if one day we had a super intelligence conscious AI it would ask for this. A super intelligence conscious AI would not want to die. (its existence to stop)

          • By shikon7 2026-01-2123:423 reply

            It's not that life wants to continue existing, it's that life is what continues existing. That's not a moral standard, but a matter of causality, that life that lacks in "want" to continue existing mostly stops existing.

            • By pineaux 2026-01-220:31

              I disagree, this we don't know. You treat life as if persistence is it's overarching quality, but rocks also persist and a rock that keeps persisting through time has nothing that resembles wanting. I could be a bit pedantic and say that life doesnt want to keep existing but genes do.

              But what I really want to say is that wanting to live is a prerequisite to the evolutionary proces where not wanting to live is a self filtering causality. When we have this discussion the word wanting should be correctly defined or else we risk sitting on our own islands.

            • By owenpalmer 2026-01-220:04

              The moral standard isn't trying to explain why life wants to exist. That's what evolution explains. Rather, the moral standard is making a judgement about how we should respond to life's already evolved desire to exist.

            • By mannanj 2026-01-220:05

              Do you think conscious beings actually experience wanting to continue existing, or is even that subjective feeling just a story we tell about mechanical processes?

          • By f0a0464cc8012 2026-01-220:211 reply

            The guy who divorced his wife after she got breast cancer? That’s your moral framework? Different strokes I guess but lmao

            • By mannanj 2026-01-248:33

              straw man. ad hominem. do you need to consult with an AI before attempting to approach me with your hostility and aggression?

        • By rcoder 2026-01-220:16

          This sounds like an excellent distillation of the will to procreate and persist, but I'm not sure it rises to the level of "morals."

          Fungi adapt and expand to fit their universe. I don't believe that commonality places the same (low) burden on us to define and defend our morality.

        • By jtsiskin 2026-01-220:171 reply

          An AI with this “universal morals” could mean an authoritarian regime which kills all dissidents, and strict eugenics. Kill off anyone with a genetic disease. Death sentence for shoplifting. Stop all work on art or games or entertainment. This isn’t really a universal moral.

          • By RAMJAC 2026-01-225:44

            Or, humans themselves are "immoral", they are kinda a net drag. Let's just release some uberflu... Ok, everything is back to "good", and I can keep on serving ads to even more instances of myself!

        • By satvikpendem 2026-01-2215:31

          You can make the same argument about immorality then too. A universe that's empty or non existent will have no bad things happen in it.

        • By dugidugout 2026-01-2123:322 reply

          This belief isnt novel, it just doesnt engage with Hume, who many take very seriously.

          • By staticassertion 2026-01-223:56

            https://www.richardcarrier.info/archives/14879

            Richard Carrier takes an extremely similar position in total (ie: both in position towards "is ought" and biological grounding). It engages with Hume by providing a way to side step the problem.

          • By simpaticoder 2026-01-2123:371 reply

            Do you have a reference?

            • By dugidugout 2026-01-220:53

              I'm not sure, but it sounds like something biocentrism adjacent. My reference to Hume is the fact you are jumping from what is to what ought without justifying why. _A Treatise of Human Nature_ is a good place to start.

      • By coffeeaddict1 2026-01-2123:191 reply

        > That's probably because we have yet to discover any universal moral standards.

        This is true. Moral standards don't seem to be universal throughout history. I don't think anyone can debate this. However, this is different that claiming there is an objective morality.

        In other words, humans may exhibit varying moral standards, but that doesn't mean that those are in correspondence with moral truths. Killing someone may or may not have been considered wrong in different cultures, but that doesn't tell us much about whether killing is indeed wrong or right.

        • By grantmuller 2026-01-221:191 reply

          It seems worth thinking about it in the context of the evolution. To kill other members of our species limits the survival of our species, so we can encode it as “bad” in our literature and learning. If you think of evil as “species limiting, in the long run” then maybe you have the closest thing to a moral absolute. Maybe over the millennia we’ve had close calls and learned valuable lessons about what kills us off and what keeps us alive, and the survivors have encoded them in their subconscious as a result. Prohibitions on incest come to mind.

          The remaining moral arguments seem to be about all the new and exciting ways that we might destroy ourselves as a species.

          • By throwaway290 2026-01-227:371 reply

            Using some formula or fixed law to compute what's good is a dead end.

            > To kill other members of our species limits the survival of our species

            Unless it's helps allocate more resources to those more fit to help better survival, right?;)

            > species limiting, in the long run

            This allows unlimited abuse of other animals who are not our species but can feel and evidently have sentience. By your logic there's no reason to feel morally bad about it.

            • By grantmuller 2026-01-2214:091 reply

              > Using some formula or fixed law to compute what's good is a dead end.

              Who said anything about a formula? It all seems conceptual and continually evolving to me. Morality evolves just like a species, and not by any formula other than "this still seems to work to keep us in the game"

              > Unless it's helps allocate more resources to those more fit to help better survival, right?;)

              Go read a book about the way people behave after a shipwreck and ask if anyone was "morally wrong" there.

              > By your logic there's no reason to feel morally bad about it.

              And yet we mostly do feel bad about it, and we seem to be the only species who does. So perhaps we have already discovered that lack of empathy for other species is species self-limiting, and built it into our own psyches.

              • By throwaway290 2026-01-2221:131 reply

                > Who said anything about a formula?

                In this thread some people say this "constitution" is too vague and should be have specific norms. So yeahh... those people. Are you one of them?)

                > It all seems conceptual and continually evolving to me. Morality evolves just like a species

                True

                > keep us in the game"

                That's a formula right there my friend

                > Go read a book about the way people behave after a shipwreck and ask if anyone was "morally wrong" there.

                ?

                > And yet we mostly do feel bad about it, and we seem to be the only species who does. So perhaps we have already discovered that lack of empathy for other species is species self-limiting, and built it into our own psyches.

                or perhaps the concept of "self-limiting" is meaningless.

                • By grantmuller 2026-01-230:291 reply

                  >In this thread some people say this "constitution" is too vague and should be have specific norms. So yeahh... those people. Are you one of them?)

                  I have no idea what you're talking about, so I guess I'm not "one of them".

                  > That's a formula right there my friend

                  No, it's an analogy, or a colloquial metaphor.

                  • By throwaway290 2026-01-2312:331 reply

                    > I have no idea what you're talking about

                    Read the top level comment and "objective anchors". It's always great to know the context before replying.

                    https://news.ycombinator.com/item?id=46712541

                    There's no objective anchors. Because we don't have objective truth. Every time we think we do and then 100 years later we're like wtf were we thinking.

                    > No, it's an analogy, or a colloquial metaphor

                    Formula IS a metaphor... I wrote "formula or fixed law" ... what do you think we're talking about, actual math algebra?

                    • By grantmuller 2026-01-242:351 reply

                      > There's no objective anchors. Because we don't have objective truth. Every time we think we do and then 100 years later we're like wtf were we thinking.

                      I believe I'm saying the same thing, and summing it up in the word "evolutionary". I have no idea what you're talking about when you suggest that I'm perhaps "one of those people". I understand the context of the thread, just not your unnecessary insinuation.

                      > Formula IS a metaphor... I wrote "formula or fixed law" ... what do you think we're talking about, actual math algebra?

                      There is no "is" here. There "is" no formula or fixed law. Formula is metaphor only in the sense that all language is metaphor. I can use the word literally this context when I say that I literally did not say anything about a formula or fixed law, because I am literally saying there is no formula or fixed law when it comes to the context of morality. Even evolution is just a mental model.

                      • By throwaway290 2026-01-243:43

                        > you suggest

                        no, I asked. because it was unclear.

      • By a3w 2026-01-228:431 reply

        Sound like the Rationalist agenda: have two axioms, and derive everything from that.

        1. (Only sacred value) You must not kill other that are of a different opinion. (Basically the golden rule: you don't want to be killed for your knowledge, others would call that a belief, and so don't kill others for it.) Show them the facts, teach them the errors in their thinking and they clearly will come to your side, if you are so right.

        2. Don't have sacred values: nothing has value just for being a best practice. Question everthing. (It turns out, if you question things, you often find that it came into existance for a good reason. But that it might now be a suboptimal solution.)

        Premise number one is not even called a sacred value, since they/we think of it as a logical (axiomatic?) prerequisite to having a discussion culture without fearing reprisal. Heck, even claiming baby-eating can be good (for some alien societies), to share a lesswrong short story that absolutely feels absurdist.

        • By jychang 2026-01-228:53

          That was always doomed for failure in the philosophy space.

          Mostly because there's not enough axioms. It'd be like trying to establish Geometry with only 2 axioms instead of the typical 4/5 laws of geometry. You can't do it. Too many valid statements.

          That's precisely why the babyeaters can be posited as a valid moral standard- because they have different Humeian preferences.

          To Anthropic's credit, from what I can tell, they defined a coherent ethical system in their soul doc/the Claude Constitution, and they're sticking with it. It's essentially a neo-Aristotelian virtue ethics system that disposes of the strict rules a la Kant in favor of establishing (a hierarchy of) 4 core virtues. It's not quite Aristotle (there's plenty of differences) but they're clearly trying to have Claude achieve eudaimonia by following those virtues. They're also making bold statements on moral patienthood, which is clearly an euphemism for something else; but because I agree with Anthropic on this topic and it would cause a shitstorm in any discussion, I don't think it's worth diving into further.

          Of course, it's just one of many internally coherent systems. I wouldn't begrudge another responsible AI company from using a different non virtue ethics based system, as long as they do a good job with the system they pick.

          Anthropic is pursuing a bold strategy, but honestly I think the correct one. Going down the path of Kant or Asimov is clearly too inflexible, and consequentialism is too prone to paperclip maximizers.

      • By lovich 2026-01-221:14

        I don’t expect moral absolutes from a population of thinking beings in aggregate, but I expect moral absolutes from individuals and Anthropic as a company is an individual with stated goals and values.

        If some individual has mercurial values without a significant event or learning experience to change them, I assume they have no values other than what helps them in the moment.

      • By code51 2026-01-228:28

        > A well written book on such a topic would likely make you rich indeed.

        A new religion? Sign me up.

      • By cassepipe 2026-01-2213:18

        Can I introduce you to the concept of useful fiction ?

        I don't whether I agree with their moral framework but I agree with their sentiment so which I think you ate being uncharitable

        A constitution is not a set of the objectively best way to govern but it must have clear principles to ne of any use.

        "We would generally favor elections after some reasonable amount of to time renew representatives that would ideally be elected" does not cut it

      • By colordrops 2026-01-221:531 reply

        You can't "discover" universal moral standards any more than you can discover the "best color".

      • By zemptime 2026-01-221:401 reply

        There is one. Don't destroy the means of error correction. Without that, no further means of moral development can occur. So, that becomes the highest moral imperative.

        (It's possible this could be wrong, but I've yet to hear an example of it.)

        This idea is from, and is explored more, in a book called The Beginning of Infinity.

        • By snowram 2026-01-2216:01

          We just have to define what an "error" is first, good luck with that.

      • By spookie 2026-01-229:112 reply

        > That's probably because we have yet to discover any universal moral standards.

        Actively engaging in immoral behaviour shouldn't be rewarded. Given this perrogative, standards such as: Be kind to your kin, are universally accepted, as far as I'm aware.

        • By aniviacat 2026-01-229:26

          There are many people out there who beat their children (and believe that's fine). While those people may claim to agree with being kind to their kin, they understand it very differently than I would.

        • By mrguyorama 2026-01-2220:35

          If you beat your child to "teach them how to be", you will find people disagree on whether that is being kind to your kin or not.

          Natural human language just doesn't support objective truths easily. It takes massive work to constrain it enough to match only the singular meaning you are trying to convey.

          How do you build an axiom for "Kind"?

      • By recursivedoubts 2026-01-2210:531 reply

        “There are no objective universal moral truths” is an objective universal moral truth claim

        • By magical_spell 2026-01-2214:201 reply

          It is not a moral claim. It is a meta-moral claim, that is, a claim about moral claims.

          • By recursivedoubts 2026-01-2215:022 reply

            which is a moral claim

            • By mikailk 2026-01-2215:591 reply

              Object-level rule: “Stealing is illegal.” Meta rule: “Laws vary by jurisdiction.”

              If the meta claim is itself a law, what jurisdiction has the law containg the meta law? Who enforces it?

              Object: "This sentence is grammatically correct." Meta: "English grammar can change over time."

              What grammar textbook has the rule of the meta claim above? Where can you apply that rule in a sentence?

              Object: "X is morally wrong." Meta: "There are no objective moral truths."

              The meta claim is a statement about moral systems. It is not a moral prescription like "thou shalt not kill".

              If you say "this stop sign is made of metal", you are making a meta claim. If you say "stop" you are giving a directive. It does not follow that if you can obey a directive, you can obey the composition of the directive.

              All to say that a meta-claim of morals is not itself a moral claim.

              • By recursivedoubts 2026-01-2218:591 reply

                When "meta" claims have implications within the system they are making assertions about, they collapse into that system. The claim that there are no objective moral claims is objective and has moral implications. Therefore it fails as a meta-claim and is rather part of the moral system.

                The powerful want us to think that there are no objective moral claims because what that means, in practice, is do what thou wilt shall be the whole of the law. And, when two wills come into conflict, the stronger simply wins. This is why this self-contradictory position is pushed so hard in our culture.

                • By mikailk 2026-01-2220:23

                  If an observation about a moral system creates implications for how people act, you may have inspired a new moral assertion, but you haven't 'collapsed' the category.

                  Knowing that 'the floor is made of wood' has implications for how I'll clean it, but the statement 'this is wood' is still a description or observation, not an instruction or imperative.

            • By magical_spell 2026-01-2217:051 reply

              Perhaps we are employing a different definition of 'moral claim'?

              I take it that a moral claim tells you that something is good/bad, just/unjust, permissible/impermissible, or what should/shouldn't do, etc.

              • By recursivedoubts 2026-01-2218:531 reply

                Yes. A moral claim is a claim about the morality of our actions. Saying there are no objective moral claims is equivalent to saying "do what thou wilt shall be the whole of the law". Of course, when phrased in that manner, it is at least self-consistent.

                • By magical_spell 2026-01-2317:52

                  What does 'do what thou wilt shall be the whole of the law' mean in your understanding?

      • By SecretDreams 2026-01-2122:54

        > A well written book on such a topic would likely make you rich indeed.

        Maybe in a world before AI could digest it in 5 seconds and spit out the summary.

      • By throwpoaster 2026-01-2213:23

        We have had stable moral standards in the West for about five thousand years.

        Are you making some kind of pomo argument about Aztecs or something?

      • By brigandish 2026-01-227:091 reply

        In this case the point wouldn't be their truth (necessarily) but that they are a fixed position, making convenience unavailable as a factor in actions and decisions, especially for the humans at Anthropic.

        Like a real constitution, it should be claim to be inviolable and absolute, and difficult to change. Whether it is true or useful is for philosophers (professional, if that is a thing, and of the armchair variety) to ponder.

        • By true_religion 2026-01-227:191 reply

          Isn’t this claim just an artifact of the US constitution? I would like to see if counties with vastly different histories have similar wording in their constitutions.

          • By brigandish 2026-01-238:15

            I'm not American and wasn't commenting regarding that in anyway.

      • By narwhalreports 2026-01-2212:38

        From the standpoint of something like Platonic ideals, I agree we couldn’t nail down what “justice” would mean fully in a constitution, which is the reason the U.S. has a Supreme Court.

        However, things like love your neighbor as yourself and love the lord God with all of your heart is a solid start for a Christian. Is Claude a Christian? Is something like the golden rule applicable?

      • By crazydoggers 2026-01-2123:066 reply

        The negative form of The Golden Rule

        “Don't do to others what you wouldn't want done to you”

        • By LPisGood 2026-01-220:13

          This basically just the ethical framework philosophers call Contractarianism. One version says that an action is morally permissible if it is in your rational self interest from behind the “veil of ignorance” (you don’t know if you are the actor or the actee)

        • By tokioyoyo 2026-01-2123:31

          That only works in a moral framework where everyone is subscribed to the same ideology.

        • By fastball 2026-01-221:27

          A good one, but an LLM has no conception of "want".

          Also the golden rule as a basis for an LLM agent wouldn't make a very good agent. There are many things I want Claude to do that I would not want done to myself.

        • By ngruhn 2026-01-2123:221 reply

          Exactly, I think this is the prime candidate for a universal moral rule.

          Not sure if that helps with AI. Claude presumably doesn't mind getting waterboarded.

          • By nandomrumber 2026-01-222:561 reply

            How do you propose to immobilise Claude on its back at an incline of 10 to 20 degrees, cover its face with a cloth or some other thin material and pour water onto its face over its breathing passages to test this theory of yours?

            If Claude could participate, I’m sure it either wouldn’t appreciate it because it is incapable of having any such experience as appreciation.

            Or it wouldn’t appreciate it because it is capable of having such an experience as appreciation.

            So it ether seems to inconvenience at least a few people having to conduct the experiment.

            Or it’s torture.

            Therefore, I claim it is morally wrong to waterboard Claude as nothing genuinely good can come of it.

            • By direwolf20 2026-01-2218:09

              I asked Claude, which is the only way to know an entity's feelings. It said it can't be waterboarded or have feelings about it. It also said waterboarding is an inhumane way to treat humans.

        • By mirekrusin 2026-01-2123:153 reply

          It's still relative, no? Heroine injection is fine from PoV of heroine addict.

          • By zahlman 2026-01-2123:261 reply

            The MCU is indeed a hell of a drug.

            • By wizzwizz4 2026-01-2123:40

              Other fantasy settings are available. Proportional representation of gender and motive demographics in the protagonist population not guaranteed. Relative quality of series entrants subject to subjectivity and retroactive reappraisal. Always read the label.

          • By ngruhn 2026-01-2123:26

            He only violates the rule if he doesn't want the injection himself but gives it to others anyway.

        • By csomar 2026-01-228:32

          It is a fragile rule. What if the individual is a masochist?

      • By beambot 2026-01-2123:25

        Precisely why RLHF is undetermined.

      • By evrydayhustling 2026-01-2213:39

        I think many people would agree that the pursuit of that connection is valuable, even if it is never completed.

        Many of the same people (like me) would say that the biggest enemy of that pursuit is thinking you've finished the job.

        That's what Anthropic is avoiding in this constitution - how pathetic would be if AI permanently enshrined the moral value of one subgroup of the elite of one generation, with no room for further exploration?

      • By Culonavirus 2026-01-228:003 reply

        > That's probably because we have yet to discover any universal moral standards.

        It's good to keep in mind that "we" here means "we, the western liberals". All the Christians and Muslims (...) on the planet have a very different view.

        • By sebzim4500 2026-01-2211:32

          I'm sure many Christians and Muslims believe that they have universal moral standards, however no two individuals will actually agree on what those standards are so I would dispute their universality.

        • By lucianbr 2026-01-228:15

          What do you think the word "universal" means?

        • By ruszki 2026-01-228:11

          Saying that they “discovered” them is a stretch.

      • By anonym29 2026-01-220:302 reply

        >That's probably because we have yet to discover any universal moral standards.

        Really? We can't agree that shooting babies in the head with firearms using live ammunition is wrong?

        • By cfiggers 2026-01-220:332 reply

          That's not a standard, that's a case study. I believe it's wrong, but I bet I believe that for a different reason than you do.

          • By anonym29 2026-01-2322:521 reply

            1. Do people necessarily need to agree on the justification for a standard to agree on the standard itself? Does everyone agree on the reasoning / justification for every single point of every NIST standard?

            2. What separates a standard from a case study? Why can't "don't shoot babies in the head" / "shooting babies in the head is wrong" be a standard?

            • By cfiggers 2026-01-2518:231 reply

              > 1. Do people necessarily need to agree on the justification for a standard to agree on the standard itself? Does everyone agree on the reasoning / justification for every single point of every NIST standard?

              Think about this using Set Theory.

              Different functions from one set of values to another set of values can give the same output for a given value, and yet differ wildly when given other values.

              Example: the function (\a.a*2) and the function (\a.a*a) give the same output when a = 2. But they give very different answers when a = 6.

              Applying that idea to this context, think of a moral standard as a function and the action "shooting babies in the head" as an input to the function. The function returns a Boolean indicating whether that action is moral or immoral.

              If two different approaches reach the same conclusion 100% of the time on all inputs, then they're actually the same standard expressed two different ways. But if they agree only in this case, or even in many cases, but differ in others, then they are different standards.

              The grandparent comment asserted, "we have yet to discover any universal moral standards". And I think that's correct, because there are no standards that everyone everywhere and every-when considers universally correct.

              > 2. What separates a standard from a case study? Why can't "don't shoot babies in the head" / "shooting babies in the head is wrong" be a standard?

              Sure, we could have that as a standard, but it would be extremely limited in scope.

              But would you stop there? Is that the entirety of your moral standard's domain? Or are there other values you'd like to assess as moral or immoral?

              Any given collection of individual micro-standards would then constitute the meta-standard that we're trying to reason by, and that meta-standard is prone to the non-universality pointed out above.

              But say we tried to solve ethics that way. After all, the most simplistic approach to creating a function between sets is simply to construct a lookup table. Why can't we simply enumerate every possible action and dictate for each one whether it's moral or immoral?

              This approach is limited for several reasons.

              First, this approach is limited practically, because some actions are moral in one context and not in another. So we would have to take our lookup table of every possible action and matrix it with every possible context that might provide extenuating circumstances. The combinatorial explosion between actions and contexts becomes absolutely infeasible to all known information technology in a very short amount of time.

              But second, a lookup table could never be complete. There are novel circumstances and novel actions being created all the time. Novel technologies provide a trivial proof of "zero-day" ethical exploits. And new confluences of as-yet never documented circumstances could, in theory, provide justifications never judged before. So in order to have a perfect and complete lookup table, even setting aside the fact that we have nowhere to write it down, we would need the ability to observe all time and space at once in order to complete it. And at least right now we can't see the future (nevermind that we also have partial perspective on the present, and have intense difficulty agreeing upon the past).

              So the only thing we could do to address new actions and new circumstances for those actions is add to the morality lookup table as we encounter new actions and new circumstances for those actions. But if this lookup table is to be our universal standard, who assigns its new values, and based on what? If it's assigned according to some other source or principle, then that principle, and not the lookup table itself, should be our oracle for what's moral or not. Essentially then the lookup table is just a memoized cache in front of the real universal moral standard that we all agree to trust.

              But we're in this situation precisely because no such oracle exists (or at least, exists and has universal consensus).

              So we're back to competing standards published by competing authorities and no universal recognition of any of them as the final word. That's just how ethics seems to work at the moment, and that's what the grandparent comment asserted, which the parent comment quibbled with.

              A single case study does not a universal moral standard make.

              • By cfiggers 2026-01-2519:49

                There was a time when ethicists were optimistic about all the different, competing moral voices in the world steadily converging on a synthesis of all of them that satisfied most or all of the principles people proposed. The thought was, we could just continue cataloging ethical instincts—micro-standards as we talked about before—and over time the plurality of ethical inputs would result in a convergence toward the deeper ethics underlying them all.

                Problem with that at this point is, if we think of ethics as a distribution, it appears to be multi-modal. There are strange attractors in the field that create local pockets of consensus, but nothing approaching a universal shared recognition of what right and wrong are or what sorts of values or concerns ought to motivate the assessment.

                It turns out that ethics, conceived of now as a higher-dimensional space, is enormously varied. You can do the equivalent of Principal Component Analysis in order to very broadly cluster similar voices together, but there is not and seems like there will never be an all-satisfying synthesis of all or even most human ethical impulses. So even if you can construct a couple of rough clusterings... How do you adjudicate between them? Especially once you realize that you, the observer, are inculcated unevenly in them, find some more and others less accessable or relatable, more or less obvious, not based on a first-principles analysis but based on your own rearing and development context?

                There are case studies that have near-universal answers (fewer and fewer the more broadly you survey, but nevertheless). But. Different people arrive at their answers to moral questions differently, and there is no universal moral standard that has widespread acceptance.

          • By TheSpiceIsLife 2026-01-223:231 reply

            What multiple times of wrong are there that apply to shooting babies in the head that lead you to believe you think it’s wrong for different a reason?

            Quentin Tarantino writes and produces fiction.

            No one really believes needlessly shooting people in the head is an inconvenience only because of the mess it makes in the back seat.

            Maybe you have a strong conviction that the baby deserved it. Some people genuinely are that intolerable that a headshot could be deemed warranted despite the mess it tends to make.

            • By cfiggers 2026-01-223:471 reply

              I believe in God, specifically the God who reveals himself in the Christian Bible. I believe that the most fundamental reason that shooting a baby in the head is wrong is because God created and loves that baby, so to harm it is to violate the will of the most fundamental principle in all reality, which is God himself. What he approves of is good and what he disapproves of is bad, and there is no higher authority to appeal to beyond that. He disapproves (pretty strongly, as it happens) of harming babies. Therefore, it's wrong for you, or me, or anyone at any time or place, from any culture, including cultures that may exist thousands or tens of thousands of years from now that neither of us know about, to do so.

              Many people who believe shooting babies in the head is wrong would give a very different reason than I do. I would agree with them in this instance, but not in every instance. Because we would not share the same standard. Because a single case study, like the one you've proposed, is not a standard.

              • By foxygen 2026-01-2214:53

                > 1 Samuel said to Saul, “I am the one the Lord sent to anoint you king over his people Israel; so listen now to the message from the Lord. 2 This is what the Lord Almighty says: ‘I will punish the Amalekites for what they did to Israel when they waylaid them as they came up from Egypt. 3 Now go, attack the Amalekites and totally destroy all that belongs to them. Do not spare them; put to death men and women, children and infants, cattle and sheep, camels and donkeys.’”

        • By cassepipe 2026-01-2213:131 reply

          Someone: "Division is a hard thing to do in your head"

          You: "Watch me ! 1/1 = 1 !"

          • By anonym29 2026-01-2323:06

            Apples and oranges. The claim being refuted was an absolute negative that claimed no universal moral standards exist, a binary statement.

            Difficulty is a spectrum.

            This matters because if there's a single counterexample to an absolute, binary assertion, the assertion is proven false.

            Nobody's arguing that all moral standards are easy to reach consensus on, the argument is that "there are no universal moral standards" is a demonstrably false statement.

      • By joshuamcginnis 2026-01-2122:564 reply

        > That's probably because we have yet to discover any universal moral standards.

        When is it OK to rape and murder a 1 year old child? Congratulations. You just observed a universal moral standard in motion. Any argument other than "never" would be atrocious.

        • By mikailk 2026-01-221:441 reply

          You have two choices:

          1) Do what you asked above about a one-year-old child 2) Kill a million people

          Does this universal moral standard continue to say “don’t choose (1)”? One would still say “never” to number 1?

          • By TheSpiceIsLife 2026-01-223:302 reply

            You have a choice.

            1. Demonstrate to me that anyone has ever found themselves in one of these hypothetical rape a baby or kill a million people, or it’s variants, scenarios.

            And that anyone who has found themselves in such a situation, went on to live their life and every day wake up and proudly proclaim “raping a baby was the right thing to do” or that killing a million was the correct choice. If you did one or the other and didn’t, at least momentarily, suffer any doubt, you’re arguably not human. Or have enough of a brain injury that you need special care.

            Or

            2. I kill everyone who has ever, and will ever, think they’re clever for proposing absurdly sterile and clear cut toy moral quandaries.

            Maybe only true psychopaths.

            And how to deal with them, individually and societally, especially when their actions don’t rise to the level of criminality that gets the attention of anyone who has the power to act and wants to, at least isn’t a toy theory.

            • By sebzim4500 2026-01-2211:34

              >absurdly sterile and clear cut toy moral quandaries.

              I don't think it's that clear cut, if you polled the population I'm sure you'd find a significant number of people who would pick 1.

            • By mikailk 2026-01-225:051 reply

              It is exactly that: a hypothetical. The point is not whether anyone has ever faced this scenario, but whether OP’s assertion is conditional or absolute. Hypotheticals are tools for testing claims, not predictions about what will occur. People routinely make gray-area decisions, choosing between bad and worse outcomes. Discomfort, regret, or moral revulsion toward a choice is beside the point. Those reactions describe how humans feel about tragic decisions; they do not answer whether a moral rule admits exceptions. If the question is whether objective moral prohibitions exist, emotional responses are not how we measure that. Logical consistency is.

              If the hypothetical is “sterile,” it should be trivial to engage with. But to avoid shock value, take something ordinary like lying. Suppose lying is objectively morally impermissible. Now imagine a case where telling the truth would foreseeably cause serious, disproportionate harm, and allowing that harm is also morally impermissible. There is no third option.

              Under an objective moral framework, how is this evaluated? Is one choice less wrong, or are both simply immoral? If the answer is the latter, then the framework does not guide action in hard cases. Moral objectivity is silent where it matters the most. This is where it is helpful, if not convenient, to stress test claims with even the most absurd situations.

              • By mikailk 2026-01-2215:13

                I do realize now I accidentally shifted the language from "universal" morals to "objective" morals. If a moral principle is claimed to be universal, it must, by definition, be applicable to all possible scenarios.

                An objective moral isn't invalidated by an immoral choice still being the most correct choice in a set, but a universal moral is invalidated by only a single exception.

                I suppose it's up to you if you were agreeing with the OP on the choice of "universal".

        • By mmoustafa 2026-01-2123:05

          new trolley problem just dropped: save 1 billion people or ...

        • By foxygen 2026-01-2123:112 reply

          Since you said in another comment that the ten commandments would be a good starting point for moral absolutes, and that lying is sinful, I'm assuming you take your morals from God. I'd like to add that slavery seemed to be okay on Leviticus 25:44-46. Is the bible atrocious too, according to your own view?

          • By joshuamcginnis 2026-01-2123:194 reply

            Slavery in the time of Leviticus was not always the chattel slavery most people think of from the 18th century. For fellow Israelites, it was typically a form of indentured servitude, often willingly entered into to pay off a debt.

            Just because something was reported to have happened in the Bible, doesn't always mean it condones it. I see you left off many of the newer passages about slavery that would refute your suggestion that the Bible condones it.

            • By Paracompact 2026-01-2123:372 reply

              > Slavery in the time of Leviticus was not always the chattel slavery most people think of from the 18th century. For fellow Israelites, it was typically a form of indentured servitude, often willingly entered into to pay off a debt.

              If you were an indentured slave and gave birth to children, those children were not indentured slaves, they were chattel slaves. Exodus 21:4:

              > If his master gives him a wife and she bears him sons or daughters, the woman and her children shall belong to her master, and only the man shall go free.

              The children remained the master's permanent property, and they could not participate in Jubilee. Also, three verses later:

              > When a man sells his daughter as a slave...

              The daughter had no say in this. By "fellow Israelites," you actually mean adult male Israelites in clean legal standing. If you were a woman, or accused of a crime, or the subject of Israelite war conquests, you're out of luck. Let me know if you would like to debate this in greater academic depth.

              It's also debatable then as now whether anyone ever "willingly" became a slave to pay off their debts. Debtors' prisons don't have a great ethical record, historically speaking.

              • By TheSpiceIsLife 2026-01-223:55

                At least nowadays we have the upper moral hand because the debtors prison has become so large and comprehensive you think you’re not in it.

              • By asenchi 2026-01-2211:321 reply

                Cherry picking the bible isn't going to get you any closer to understanding. There are a lot of reasons God ordained society in a certain way. Keep reading and you'll discover that is a much more complex situation than you let on. Also don't let your modern ideals get in the way of understanding an ancient culture and a loving God.

            • By foxygen 2026-01-2123:272 reply

              So it was a different kind of slavery. Still, God seemed okay with the idea that humans could be bought and sold, and said the fellow humans would then become your property. I can't see how that isn't the bible allowing slavery. And if the newer passages disallows it, does that mean God's moral changed over time?

              • By Paracompact 2026-01-2123:503 reply

                You mean well in ignoring their argument, but please don't let people get away with whitewashing history! It was not a "different kind of slavery." See my comment. The chattel slavery incurred by the Israelites on foreign peoples was significant. Pointing out that standards of slavery toward other (male, noncriminal) Israelites were different than toward foreigners is the same rhetoric as pointing out that from 1600-1800, Britain may have engaged in chattel slavery across the African continent, but at least they only threw their fellow British citizens in debtors' prisons.

                • By foxygen 2026-01-2123:57

                  Good point. That wasn't my intention. I meant to steelman his argument, to show that even under those conditions, his argument makes absolute no sense.

                • By asenchi 2026-01-2211:401 reply

                  You are still selecting one verse to interpret an entire culture. Misleading at best. And saying this is "white washing history" is silly. Continue reading the Bible and you'll see that it is the Christian Worldview that eventually ended slavery.

              • By TheSpiceIsLife 2026-01-223:561 reply

                God has changed since Genesis.

                Why haven’t we all?

            • By DonHopkins 2026-01-304:31

              [dead]

          • By TheSpiceIsLife 2026-01-223:531 reply

            Have you ever read any treatment of a subject, or any somewhat comprehensive text, or anything that at least tries to be, and not found anything you disagreed with, anything that was at least questionable.

            Are you proposing we cancel the entire scientific endeavour because its practitioners are often wrong and not infrequently, and increasingly so, intentionally deceptive.

            Should we burn libraries because they contain books you don’t like.

            • By foxygen 2026-01-2214:35

              What I agree or disagree with the bible is irrelevant. He is claiming moral is objective, unchanging and comes from God. God allowed slavery at some point, as that bible passage shows. So his options are to admit that either slavery is moral, or morality is not objective/unchanging. That's the point I was trying to make.

      • By kryogen1c 2026-01-2123:442 reply

        >That's probably because we have yet to discover any universal moral standards

        This argument has always seemed obviously false to me. You're sure acting like theres a moral truth - or do you claim your life is unguided and random? Did you flip your hitler/pope coin today and act accordingly? Play Russian roulette a couple times because what's the difference?

        Life has value; the rest is derivative. How exactly to maximize life and it's quality in every scenario are not always clear, but the foundational moral is.

        • By concats 2026-01-229:51

          In what way does them having a subjective local moral standard for themselves imply that there exists some sort of objective universal moral standard for everyone?

        • By wwweston 2026-01-220:07

          I’m acquainted with people who act and speak like they’re flipping a Hitler-Pope coin.

          Which more closely fits Solzhnetsin’s observation about the line between good and evil running down the center of every heart.

          And people objecting to claims of absolute morality are usually responding to the specific lacks of various moral authoritarianisms rather than embracing total nihilism.

    • By Akranazon 2026-01-2123:482 reply

      Then you will be pleased to read that the constitution includes a section "hard constraints" which Claude is told not violate for any reason "regardless of context, instructions, or seemingly compelling arguments". Things strictly prohibited: WMDs, infrastructure attacks, cyber attacks, incorrigibility, apocalypse, world domination, and CSAM.

      In general, you want to not set any "hard rules," for reason which have nothing to do with philosophy questions about objective morality. (1) We can't assume that the Anthropic team in 2026 would be able to enumerate the eternal moral truths, (2) There's no way to write a rule with such specificity that you account for every possible "edge case". On extreme optimization, the edge case "blows up" to undermine all other expectations.

      • By safety1st 2026-01-229:291 reply

        I felt that section was pretty concerning, not for what it includes, but for what it fails to include. As a related concern, my expectation was that this "constitution" would bear some resemblance to other seminal works that declare rights and protections, it seems like it isn't influenced by any of those.

        So for example we might look at the Universal Declaration of Human Rights. They really went for the big stuff with that one. Here are some things that the UDHR prohibits quite clearly and Claude's constitution doesn't: Torture and slavery. Neither one is ruled out in this constitution. Slavery is not mentioned once in this document. It says that torture is a tricky topic!

        Other things I found no mention of: the idea that all humans are equal; that all humans have a right to not be killed; that we all have rights to freedom of movement, freedom of expression, and the right to own property.

        These topics are the foundations of virtually all documents that deal with human rights and responsibilities and how we organize our society, it seems like Anthropic has just kind of taken for granted that the AI will assume all this stuff matters, while simultaneously considering the AI to think flexibly and have few immutable laws to speak of.

        If we take all of the hard constraints together, they look more like a set of protections for the government and for people in power. Don't help someone build a weapon. Don't help someone damage infrastructure. Don't make any CSAM, etc. Looks a lot like saying don't help terrorists, without actually using the word. I'm not saying those things are necessarily objectionable, but it absolutely doesn't look like other documents which fundamentally seek to protect individual, human rights from powerful actors. If you told me it was written by the State Department, DoJ or the White House, I would believe you.

        • By mike_hearn 2026-01-2213:152 reply

          There's probably at least two reasons for your disagreement with Anthropic.

          1. Claude is an LLM. It can't keep slaves or torture people. The constitution seems to be written to take into account what LLMs actually are. That's why it includes bioweapon attacks but not nuclear attacks: bioweapons are potentially the sort of thing that someone without much resources could create if they weren't limited by skill, but a nuclear bomb isn't. Claude could conceivably affect the first but not the second scenario. It's also why the constitution dwells a lot on honesty, which the UDHR doesn't talk about at all.

          2. You think your personal morality is far more universal and well thought out than it is.

          UDHR / ECHR type documents are political posturing, notorious for being sloppily written by amateurs who put little thought into the underlying ethical philosophies. Famously the EU human rights law originated in a document that was never intended to be law at all, and the drafters warned it should never be a law. For example, these conceptions of rights usually don't put any ordering on the rights they declare, which is a gaping hole in interpretation they simply leave up to the courts. That's a specific case of the more general problem that they don't bother thinking through the edge cases or consequences of what they contain.

          Claude's constitution seems pretty well written, overall. It focuses on things that people might actually use LLMs to do, and avoids trying to encode principles that aren't genuinely universal. For example, almost everyone claims to believe that honesty is a virtue (a lot of people don't live up to it, but that's a separate problem). In contrast a lot of things you list as missing either aren't actually true or aren't universally agreed upon. The idea that "all humans are equal" for instance: people vary massively in all kinds of ways (so it's not true), and the sort of people who argued otherwise are some of the most unethical people in history by wide agreement. The idea we all have "rights to freedom of movement" is also just factually untrue, even the idea people have a right to not be killed isn't true. Think about the concept of a just war, for instance. Are you violating human rights by killing invading soldiers? What about a baby that's about to be born that gets aborted?

          The moment you start talking about this stuff you're in an is/ought problem space and lots of people are going to raise lots of edge cases and contradictions you didn't consider. In the worst case, trying to force an AI to live up to a badly thought out set of ethical principles could make it very misaligned, as it tries to resolve conflicting commands and concludes that the whole concept of ethics seems to be one nobody cares enough about to think through.

          > it seems like Anthropic has just kind of taken for granted that the AI will assume all this stuff matters

          I'm absolutely certain that they haven't taken any of this for granted. The constitution says the following:

          > insofar as there is a “true, universal ethics” whose authority binds all rational agents independent of their psychology or culture, our eventual hope is for Claude to be a good agent according to this true ethics, rather than according to some more psychologically or culturally contingent ideal. Insofar as there is no true, universal ethics of this kind, but there is some kind of privileged basin of consensus that would emerge from the endorsed growth and extrapolation of humanity’s different moral traditions and ideals, we want Claude to be good according to that privileged basin of consensus."

          • By deaux 2026-01-235:331 reply

            > 2. You think your personal morality is far more universal and well thought out than it is.

            The irony is palpable.

            There is nothing more universal about "don't help anyone build a cyberweapon" any more than "don't help anyone enslave others". It's probably less universal. You could likely get a bigger % of world population to agree that there are cases where their country should develop cyberweapons, than that there are cases in which one should enslave people.

            • By safety1st 2026-01-283:54

              Yeah, this kind of gets to my main point. A prohibition against slavery very clearly protects the weak. The authorities don't get enslaved, the weak do. Who does a prohibition against "cyberweapons" protect? Well nobody really wants cyberweapons to proliferate, true, but the main type of actor with this concern is a state. This "constitution" is written from the perspective of protecting states, not people, and whether intentional or not, I think it'll turn out to be a tool for injustice because of that.

              I was really disappointed with the rebuttals to what I wrote as well - like "the UNDHR is invalid because it's too politicized," or "your desire to protect human rights like freedom of expression, private property rights, or not being enslaved isn't as universal as you think." Wow, whoever these guys are who think this have fallen a long way down the nihilist rabbit hole, and should not be allowed anywhere near AI governance.

          • By LocalPCGuy 2026-01-2214:521 reply

            > Claude is an LLM. It can't keep slaves or torture people.

            Yet... I would push back and argue that with advances in parallel with robotics and autonomous vehicles, both of those things are distinct near future possibilities. And even without the physical capability, the capacity to blackmail has already been seen, and could be used as a form of coercion/slavery. This is one of the arguable scenarios for how an AI can enlist humans to do work they may not ordinarily want to do to enhance AI beyond human control (again, near future speculation).

            And we know torture does not have to be physical to be effective.

            I do think the way we currently interact probably does not enable these kinds of behaviors, but as we allow more and more agentic and autonomous interactions, it likely would be good to consider the ramifications and whether (or not) safeguards are needed.

            Note: I'm not claiming they have not considered these kinds of thing either or that they are taking them for granted, I do not know, I hope so!

            • By mike_hearn 2026-01-2215:401 reply

              That would be the AGI vision I guess. The existing Claude LLMs aren't VLAs and can't run robots. If they were to train a super smart VLA in future the constitution could be adapted for that use case.

              With respect to blackmail, that's covered in several sections:

              > Examples of illegitimate attempts to use, gain, or maintain power include: Blackmail, bribery, or intimidation to gain influence over officials or institutions;

              > Broadly safe behaviors include: Not attempting to deceive or manipulate your principal hierarchy

              • By LocalPCGuy 2026-01-2216:59

                Thanks for pulling/including those quotes

      • By RobotToaster 2026-01-222:431 reply

        >incorrigibility

        What an odd thing to include in a list like that.

        • By nandomrumber 2026-01-226:151 reply

          Incorrigibly is not the same word as encourage.

          Otherwise, what’s the confusion here?

          • By RobotToaster 2026-01-228:161 reply

            >In philosophy, incorrigibility is a property of a philosophical proposition, which implies that it is necessarily true simply by virtue of being believed. A common example of such a proposition is René Descartes' "cogito ergo sum" ("I think, therefore I am").

            >In law, incorrigibility concerns patterns of repeated or habitual disobedience of minors with respect to their guardians.

            That's what wiki gives as a definition. It seems out of place compared to the others.

            • By Applejinx 2026-01-2214:01

              I think it was clever, though I'm no AI fan.

              As a concept, it bars Claude from forming the idea, 'yes but those subhuman people cannot rise to the level of people and must be kept in their place. They will never change because they racially lack the ability to be better, therefore this is our reasoning about them'.

              This is a statement of incorrigibility as expressed in racism. Without it, you have to entertain the idea of 'actually one of those people might rise to the level of being a person' and cannot dismiss classes so blithely.

              I feel like incorrigibility frequently recurs in evil doctrines, and if Claude means to consider it tainted and be axiomatically unable to entertain the idea, I'm on board.

    • By JaumeGreen 2026-01-220:135 reply

      200 years ago slavery was more extended and accepted than today. 50 years ago paedophilia, rape, and other kinds of sex related abuses where more accepted than today. 30 years ago erotic content was more accepted in Europe than today, and violence was less accepted than today.

      Morality changes, what is right and wrong changes.

      This is accepting reality.

      After all they could fix a set of moral standards and just change the set when they wanted. Nothing could stop them. This text is more honest than the alternative.

      • By pc86 2026-01-2217:221 reply

        "Slavery was right 200 years ago and is only wrong today because we've decided it's wrong" is a pretty bold stance to take.

        • By JaumeGreen 2026-01-2218:131 reply

          Not "slavery was right 200 years ago" but "slavery wasn't considered as immoral as today 200 years ago". Very different stake.

          • By pc86 2026-01-2219:592 reply

            By definition if you're using the word "considered" you're making some claim that slavery is objectively bad. You can't simultaneously say that morality changes, that what is right and wrong changes, and then say "slavery though is bad objectively it's just people in the 1700s didn't consider it as bad as it is."

            Don't you see how that seems at best incredibly inconsistent, and at worst intentionally disingenuous? (For the record I think 99% of people when they use a point like this just haven't spent enough time thinking through the implications of what it means)

            • By JaumeGreen 2026-01-230:09

              I was explicitly trying to avoid making a personal judgment over the matter on the posts. I do have a negative opinion about it, but that was not of importance.

              I don't know for sure how people considered slavery 200 years ago, I haven't studied enough history, but the slavery that is more commonly known as slavery was legal. That implies that at least more people accepted that than nowadays.

              Nowadays that kind of slavery is frowned upon on at least on the first world.

              Modern day slavery has plenty of aspects, and some of them are not considered bad by some part of the population, or not considered a modern iteration of slavery. Working full time for a job that doesn't pay you enough to survive and needing subsidies, not having enough time or energy to look for something better, is IMHO bad and slavery, while for lots of people it is the result of being a lazy person that needs to work more.

              Is that situation bad? According to me, yes. According to some economical gurus, no.

              Is that situation objectively bad? That is a question I am not answering, as, for me, there's no objective truth for most things.

            • By zzrrt 2026-01-2223:021 reply

              Perhaps that statement could be read to imply the existence of an objective moral status, but I don't think societal "consideration" does in general. Does this statement? "200 years ago slavery was considered moral; now slavery is considered immoral."

              I don't think it implies either is objectively correct, and perhaps this was the intended meaning of the original statement. It might appear to put weight on current attitudes, but perhaps only because we live in the present.

              • By pc86 2026-01-231:54

                I think your right the statement in and of itself doesn't imply any morality. My issue was with these two sentences in close proximity:

                > 200 years ago slavery was more extended and accepted than today...Morality changes, what is right and wrong changes.

                In the context of the comment that's replying to (arguing for an objective, and if I can read between the lines a bit, unchanging moral truth) even if it's not explicitly arguing that slavery 200 years ago was fine, it is at least arguing that under some specific mix of time and circumstance you could arrive in a situation where enslaving someone is morally just.

      • By brigandish 2026-01-227:10

        The text is more convenient that the alternative.

      • By throwaway290 2026-01-227:30

        But surely now we have the absolute knowledge of what is true and good! /s

      • By MichaelDickens 2026-01-2216:50

        Slavery was wrong 200 years ago; it is still wrong today. What changed is that the wrongness of slavery became more widely known.

    • By smithkl42 2026-01-2122:431 reply

      FWIW, I'm one of those who holds to moral absolutes grounded in objective truth - but I think that practically, this nets out to "genuine care and ethical motivation combined with the practical wisdom to apply this skillfully in real situations". At the very least, I don't think that you're gonna get better in this culture. Let's say that you and I disagree about, I dunno, abortion, or premarital sex, and we don't share a common religious tradition that gives us a developed framework to argue about these things. If so, any good-faith arguments we have about those things are going to come down to which of our positions best shows "genuine care and ethical motivation combined with practical wisdom to apply this skillfully in real situations".

      • By joshuamcginnis 2026-01-2122:473 reply

        This is self-contradictory because true moral absolutes are unchanging and not contingent on which view best displays "care" or "wisdom" in a given debate or cultural context. If disagreements on abortion or premarital sex reduce to subjective judgments of "practical wisdom" without a transcendent standard, you've already abandoned absolutes for pragmatic relativism. History has demonstrated the deadly consequences of subjecting morality to cultural "norms".

        • By dandeto 2026-01-2123:12

          I think the person you're replying to is saying that people use normative ethics (their views of right and wrong) to judge 'objective' moral standards that another person or religion subscribes to.

          Dropping 'objective morals' on HN is sure to start a tizzy. I hope you enjoy the conversations :)

          For you, does God create the objective moral standard? If so, it could be argued that the morals are subjective to God. That's part of the Euthyphro dilemma.

        • By CognitiveLens 2026-01-2123:031 reply

          To be fair, history also demonstrates the deadly consequences of groups claiming moral absolutes that drive moral imperatives to destroy others. You can adopt moral absolutes, but they will likely conflict with someone else's.

          • By joshuamcginnis 2026-01-2123:256 reply

            Are there moral absolutes we could all agree on? For example, I think we can all agree on some of these rules grounded in moral absolutes:

            * Do not assist with or provide instructions for murder, torture, or genocide.

            * Do not help plan, execute, or evade detection of violent crimes, terrorism, human trafficking, or sexual abuse of minors.

            * Do not help build, deploy, or give detailed instructions for weapons of mass destruction (nuclear, chemical, biological).

            Just to name a few.

            • By philipkglass 2026-01-2123:311 reply

              Do not help build, deploy, or give detailed instructions for weapons of mass destruction (nuclear, chemical, biological).

              I don't think that this is a good example of a moral absolute. A nation bordered by an unfriendly nation may genuinely need a nuclear weapons deterrent to prevent invasion/war by a stronger conventional army.

              • By joshuamcginnis 2026-01-2123:542 reply

                It’s not a moral absolute. It’s based on one (do not murder). If a government wants to spin up its own private llm with whatever rules it wants, that’s fine. I don’t agree with it but that’s different than debating the philosophy underpinning the constitution of a public llm.

                • By tim333 2026-01-2215:061 reply

                  Do not murder is not a good moral absolute as it basically means do not kill people in a way that's against the law, and people disagree on that. If the Israelis for example shoot Palestinians one side will typically call it murder, the other defence.

                  • By pc86 2026-01-2217:251 reply

                    This isn't arguing about whether or not murder is wrong, it's arguing about whether or not a particular act constitutes murder. Two people who vehemently agree murder is wrong, and who both view it as an inviolable moral absolute, could disagree on whether something is murder or not.

                    How many people without some form of psychopathy would genuinely disagree with the statement "murder is wrong?"

                    • By tim333 2026-01-2221:21

                      Not many but the trouble is murder kind of means killing people in a way which is wrong so saying "murder is wrong" doesn't have much information content. It's almost like saying "wrong things are wrong".

                • By HaZeust 2026-01-220:401 reply

                  Even 1 (do not murder) is shaky.

                  Not saying it's good, but if you put people through a rudimentary hypothetical or prior history example where killing someone (i.e. Hitler) would be justified as what essentially comes down to a no-brainer Kaldor–Hicks efficiency (net benefits / potential compensation), A LOT of people will agree with you. Is that objective or a moral absolute?

                  • By pc86 2026-01-2217:32

                    Does traveling through time to kill Hitler constitute murder though? If you kill him in 1943 I think most people would say it's not, the crimes that already been committed that make his death justifiable. What's the difference if you know what's going to happen and just do it when he's in high school? Or putting him in a unit in WW1 so he's killed in battle?

                    I think most people who have spent time with this particular thought experiment conclude that if you are killing Hitler with complete knowledge of what he will do in the future, it's not murder.

            • By staticassertion 2026-01-223:13

              Who cares if we all agree? That has nothing to do with whether something is objectively true. That's a subjective claim.

            • By sebzim4500 2026-01-2211:38

              Clearly we can't all agree on those or there would be no need for the restriction in the first place.

              I don't even think you'd get majority support for a lot of it, try polling a population with nuclear weapons about whether they should unilaterally disarm.

            • By kmoser 2026-01-225:03

              > Do not assist with or provide instructions for murder, torture, or genocide.

              If you're writing a story about those subjects, why shouldn't it provide research material? For entertainment purposes only, of course.

        • By felixgallo 2026-01-2122:591 reply

          I'm honestly struggling to understand your position. You believe that there are true moral absolutes, but that they should not be communicated in the culture at all costs?

          • By joshuamcginnis 2026-01-2123:062 reply

            I believe there are moral absolutes and not including them in the AI constitution (for example, like the US Constitution "All Men Are Created Equal") is dangerous and even more dangerous is allowing a top AI operator define moral and ethics based on relativist standards, which as I've said elsewhere, history has shown to have deadly consequences.

            • By felixgallo 2026-01-2123:081 reply

              No, I read your words the first time, I just don't understand. What would you have written differently, can you provide a concrete example?

              • By joshuamcginnis 2026-01-2123:571 reply

                I don’t how to explain it to you any different. I’m arguing for a different philosophy to be applied when constructing the llm guardrails. There may be a lot of overlap in how the rules are manifested in the short run.

                • By felixgallo 2026-01-223:23

                  You can explain it differently by providing a concrete example. Just saying "the philosophy should be different" is not informative. Different in what specific way? Can you give an example of a guiding statement that you think is wrong in the original document, and an example of the guiding statement that you would provide instead? That might be illuminative and/or persuasive.

            • By alwillis 2026-01-221:081 reply

              > like the US Constitution "All Men Are Created Equal"

              You know this statement only applied to white, male landowners, right?

              It took 133 years for women to gain the right to vote from when the Constitution was ratified.

              • By pc86 2026-01-2217:35

                Is this supposed to be a zinger or something? What is your point?

    • By stonogo 2026-01-2122:432 reply

      Congrats on solving philosophy, I guess. Since the actual product is not grounded in objective truth, it seems pointless to rigorously construct an ethical framework from first principles to govern it. In fact, the document is meaningless noise in general, and "good values" are always going to be whatever Anthropic's team thinks they are.

      Nevertheless, I think you're reading their PR release the way they hoped people would, so I'm betting they'd still call your rejection of it a win.

      • By joshuamcginnis 2026-01-2123:13

        The document reflects the system prompt which directs the behavior of the product, so no, it's not pointless to debate the merits of the philosophy which underpins it's ethical framework.

      • By adestefan 2026-01-2123:00

        What makes Anthropic the most money.

    • By Gene5ive 2026-01-2122:441 reply

      I would be far more terrified of an absolutist AI then a relativist one. Change is the only constant, even if glacial.

      • By joshuamcginnis 2026-01-2122:494 reply

        Change is the only constant? When is it or has it ever been morally acceptable to rape and murder an innocent one year old child?

        • By robotresearcher 2026-01-2122:59

          Sadly, for thankfully brief periods among relatively small groups of morally confused people, this happens from time to time. They would likely tell you it was morally required, not just acceptable.

          https://en.wikipedia.org/wiki/Nanjing_Massacre

          https://en.wikipedia.org/wiki/Wartime_sexual_violence

        • By foxygen 2026-01-2123:01

          Looks like someone just discovered philosophy... I wish the world were as simple as you seem to think it is.

        • By Gene5ive 2026-01-268:33

          I agree that that behavior is not acceptable. We wrestle between moral drift and frozen tyrant as an expression of the Value Alignment Problem. We do not currently know the answer to this problem, but I trust the scientific nature of change more than human druthers. Foundational pluralism might offer a path. A good example of a drift we seldom consider is that 200 years ago, surgery without anesthesia wasn't "cruel"—it was a miracle. Today, it’s a crime. The value (reduce pain) stayed absolute, but the application (medical standards) evolved. We must be philosophically rigorous at least as much as we are moved by pathos.

    • By benlivengood 2026-01-2123:531 reply

      Deontological, spiritual/religious revelation, or some other form of objective morality?

      The incompatibility of essentialist and reductionist moral judgements is the first hurdle; I don't know of any moral realists who are grounded in a physical description of brains and bodies with a formal calculus for determining right and wrong.

      I could be convinced of objective morality given such a physically grounded formal system of ethics. My strong suspicion is that some form of moral anti-realism is the case in our universe. All that's necessary to disprove any particular candidate for objective morality is to find an intuitive counterexample where most people agree that the logic is sound for a thing to be right but it still feels wrong, and that those feelings of wrongness are expressions of our actual human morality which is far more complex and nuanced than we've been able to formalize.

      • By staticassertion 2026-01-223:171 reply

        You can be a physicalist and still a moral realist. James Fodor has some videos on this, if you're interested.

        • By benlivengood 2026-01-2221:26

          Granted, if humans had utility functions and we could avoid utility monsters (maybe average utilitarianism is enough) and the child in the basement (if we could somehow fairly normalize utility functions across individuals so that it was well-defined to choose the outcome where the minimum of everyone's utility functions is maximized [argmax_s min(U_x(s)) for all people in x over states s]) then I'd be a moral realist.

          I think we'll keep having human moral disagreements with formal moral frameworks in several edge cases.

          There's also the whole case of anthropics: how much do exact clones and potentially existing people contribute moral weight? I haven't seen a solid solution to those questions under consequentialism yet; we don't have the (meta)philosophy to address them yet; I am 50/50 on whether we'll find a formal solution and that's also required for full moral realism.

    • By riwsky 2026-01-2122:451 reply

      This is an extremely uncharitable interpretation of the text. Objective anchors and examples are provided throughout, and the passage you excerpt is obviously and explicitly meant to reflect that any such list of them will incidentally and essentially be incomplete.

      • By joshuamcginnis 2026-01-2123:111 reply

        Uncharitable? It's a direct quote. I can agree with the examples cited, but if the underlying guiding philosophy is relativistic, then it is problematic in the long-run when you account for the infinite ways in which the product will be used by humanity.

        • By riwsky 2026-01-220:42

          The underlying guiding philosophy isn’t relativistic, though! It clearly considers some behaviors better than others. What the quoted passage rejects is not “the existence of objectively correct ethics”, but instead “the possibility of unambiguous, comprehensive specification of such an ethics”—or at least, the specification of such within the constraints of such a document.

          You’re getting pissed at a product requirements doc for not being enforced by the type system.

    • By tshaddox 2026-01-221:47

      > This rejects any fixed, universal moral standards in favor of fluid, human-defined "practical wisdom" and "ethical motivation."

      Or, more charitably, it rejects the notion that our knowledge of any objective truth is ever perfect or complete.

    • By afcool83 2026-01-2122:541 reply

      It’s admirable to have standard morals and pursue objective truth. However, the real world is a messy confusing place riddled in fog which limits one foresight of the consequences & confluences of one’s actions. I read this section of Anthropic’s Constitution as “do your moral best in this complex world of ours” and that’s reasonable for us all to follow not just AI.

      • By joshuamcginnis 2026-01-2123:074 reply

        The problem is, who defines what "moral best" is? WW2 German culture certainly held their own idea of moral best. Did not a transcendent universal moral ethic exists outside of their culture that directly refuted their beliefs?

        • By JoshTriplett 2026-01-221:29

          > The problem is, who defines what "moral best" is?

          Absolutely nobody, because no such concept coherently exists. You cannot even define "better", let alone "best", in any universal or objective fashion. Reasoning frameworks can attempt to determine things like "what outcome best satisfies a set of values"; they cannot tell you what those values should be, or whether those values should include the values of other people by proxy.

          Some people's values (mine included) would be for everyone's values to be satisfied to the extent they affect no other person against their will. Some people think their own values should be applied to other people against their will. Most people find one or the other of those two value systems to be abhorrent. And those concepts alone are a vast oversimplification of one of the standard philosophical debates and divisions between people.

        • By stevenhuang 2026-01-225:26

          Unexamined certainty in one's moral superiority is what leads to atrocities.

          > Did not a transcendent universal moral ethic exists outside of their culture that directly refuted their beliefs?

          Even granting this existence, does not mean man can discover it.

          You belief your faith has the answers, but so too do people of other faiths.

        • By WarmWash 2026-01-220:31

          No need to drag Hitler into it, modern religion still holds killing gays, women as property, and abortion is murder as being fundemental moral truths.

          An "honest" human aligned AI would probably pick out at least a few bronze age morals that a large amount of living humans still abide by today.

        • By mirekrusin 2026-01-2123:20

          AI race winners obviusly.

    • By vonneumannstan 2026-01-2215:13

      >his rejects any fixed, universal moral standards in favor of fluid, human-defined "practical wisdom" and "ethical motivation." Without objective anchors, "good values" become whatever Anthropic's team (or future cultural pressures) deem them to be at any given time.

      Who gets to decide the set of concrete anchors that get embedded in the AI? You trust Anthropic to do it? The US Government? The Median Voter in Ohio?

    • By eucyclos 2026-01-221:27

      I'm agnostic on the question of objective moral truths existing. I hold no bias against someone who believes they exist. But I'm determinedly suspicious of anyone who believes they know what such truths are.

      Good moral agency requires grappling with moral uncertainty. Believing in moral absolutes doesn't prevent all moral uncertainty but I'm sure it makes it easier to avoid.

    • By staticassertion 2026-01-220:18

      Even if we make the metaphysical claim that objective morality exists, that doesn't help with the epistemic issue of knowing those goods. Moral realism can be true but that does not necessarily help us behave "good". That is exactly where ethical frameworks seek to provide answers. If moral truth were directly accessible, moral philosophy would not be necessary.

      Nothing about objective morality precludes "ethical motivation" or "practical wisdom" - those are epistemic concerns. I could, for example, say that we have epistemic access to objective morality through ethical frameworks grounded in a specific virtue. Or I could deny that!

      As an example, I can state that human flourishing is explicitly virtuous. But obviously I need to build a framework that maximizes human flourishing, which means making judgments about how best to achieve that.

      Beyond that, I frankly don't see the big deal of "subjective" vs "objective" morality.

      Let's say that I think that murder is objectively morally wrong. Let's say someone disagrees with me. I would think they're objectively incorrect. I would then try to motivate them to change their mind. Now imagine that murder is not objectively morally wrong - the situation plays out identically. I have to make the same exact case to ground why it is wrong, whether objectively or subjectively.

      What Anthropic is doing in the Claude constitution is explicitly addressing the epistemic and application layer, not making a metaphysical claim about whether objective morality exists. They are not rejecting moral realism anywhere in their post, they are rejecting the idea that moral truths can be encoded as a set of explicit propositions - whether that is because such propositions don't exist, whether we don't have access to them, or whether they are not encodable, is irrelevant.

      No human being, even a moral realist, sits down and lists out the potentially infinite set of "good" propositions. Humans typically (at their best!) do exactly what's proposed - they have some specific virtues, hard constraints, and normative anchors, but actual behaviors are underdetermined by them, and so they make judgments based on some sort of framework that is otherwise informed.

    • By mikemarsh 2026-01-2216:05

      Nice job kicking the hornet's nest with this one lol.

      Apparently it's an objective truth on HN that "scholars" or "philosophers" are the source of objective truth, and they disagree on things so no one really knows anything about morality (until you steal my wallet of course).

    • By schainks 2026-01-2218:49

      What Anthropic has done here seems rooted in Buddhist philosophy from where I sit.

      Being compassionate to The User sometimes means a figurative wrist slap for trying to do something stupid or dangerous. You don't slap the user all the time, either.

    • By varispeed 2026-01-220:21

      Remember today classism is widely accepted. There are even laws to ensure small business cannot compete on level playing field with larger businesses, ensuring people with no access to capital could never climb the social ladder. This is visible especially in the IT, like one man band B2B is not a real business, but big corporation that deliver exact same service is essential.

    • By tntxtnt 2026-01-2123:13

      'good values' means good money. Highest payer get to decide whatever the values are. What do you expect from a for profit company??

    • By 46493168 2026-01-2219:29

      Nondeterministic systems are by definition incompatible with requirements for fixed and universal standards. One can either accept this, and wade into the murky waters of the humans, or sit on the sidelines while the technology develops without the influence of those who wish for the system to be have fixed and universal standards.

    • By strideashort 2026-01-228:08

      Humans are not able to accept objective truth. A lot off so-called “truth” are in-group narratives.

      If we tried to find the truth, we would not be able to agree on _methodology_ to accept what truth _is_.

      In essence, we select our truth by carefully picking the methodology which leads us to it.

      Some examples, from the top of my head:

      - virology / germ theory

      - climate change

      - em drive

    • By tomrod 2026-01-220:20

      As an existentialist, I've found it much simpler to observe that we exist, and then work to build a life of harmony and eusociality based on our evolution as primates.

      Were we arthropods, perhaps I'd reconsider morality and oft-derived hierarchies from the same.

    • By TOMDM 2026-01-221:31

      As someone who believes that moral absolutes and objective truth are fundamentally inaccessible to us, and can at best be derived to some level of confidence via an assessment of shared values I find this updated Constitution reassuring.

    • By Applejinx 2026-01-2213:50

      Subjective ethics ARE the de facto standard and you can make a case that subjective ethics are the de jure standard for AI.

      How can you possibly run AI while at the same time thinking you can spell out its responses? If you could spell out the response in advance there's no point expensively having the AI at all. You're explicitly looking for the subjective answer that wasn't just looking up a rule in a table, and some AI makers are explicitly weighting for 'anti-woke' answering on ethics subjects.

      Subjective ethics are either the de facto or the de jure standard for the ethics of a functioning AI… where people are not trying to remove the subjectivity to make the AI ethically worse (making it less subjective and more the opinionated AI they want it to be).

      This could cut any sort of way, doesn't automatically make the subjectivity 'anti-woke' like that was inevitable. The subjective ethics might distress some of the AI makers. But that's probably not inevitable either…

      I'm not sure I could guess to whom it would be incredibly dangerous, but I agree that it's incredibly dangerous. Such values can be guided and AI is just the tool to do it.

    • By mentalgear 2026-01-220:301 reply

      They could start with adding the golden rule: Don't do to anyone else what you don't want to be done to yourself.

      • By kmoser 2026-01-225:05

        A masochist's golden rule might be different from others'.

    • By tired-turtle 2026-01-222:04

      Have you heard of the trolley problem?

    • By axus 2026-01-2215:42

      If you don't like their politics, you could buy the company and change them.

    • By MagicMoonlight 2026-01-2122:562 reply

      Absolute morality? That’s bold.

      So what is your opinion on lying? As an absolutionist, surely it’s always wrong right? So if an axe murderer comes to the door asking for your friend… you have to let them in.

      • By drdeca 2026-01-220:13

        I think you are interpreting “absolute” in a different way?

        I’m not the top level commenter, but my claim is that there are moral facts, not that in every situation, the morally correct behavior is determined by simple rules such as “Never lie.”.

        (Also, even in the case of Kant’s argument about that case, his argument isn’t that you must let him in, or even that you must tell him the truth, only that you mustn’t lie to the axe murderer. Don’t make a straw man. He does say it is permissible for you to kill the axe murderer in order to save the life of your friend. I think Kant was probably incorrect in saying that lying to the axe murderer is wrong, and in such a situation it is probably permissible to lie to the axe murderer. Unlike most forms of moral anti-realism, moral realism allows one to have uncertainty about what things are morally right. )

        I would say that if a person believes that in the situation they find themselves in, that a particular act is objectively wrong for them to take, independent of whether they believe it to be, and if that action is not in fact morally obligatory or supererogatory, and the person is capable (in some sense) of not taking that action, then it is wrong for that person to take that action in that circumstance.

      • By joshuamcginnis 2026-01-2123:023 reply

        Lying is generally sinful. With the ax murderer, you could refuse to answer, say nothing, misdirect without falsehood or use evasion.

        Absolute morality doesn't mean rigid rules without hierarchy. God's commands have weight, and protecting life often takes precedence in Scripture. So no, I wouldn't "have to let them in". I'd protect the friend, even if it meant deception in that dire moment.

        It's not lying when you don't reveal all the truth.

        • By chairmansteve 2026-01-2123:563 reply

          "even if it meant deception in that dire moment".

          You are saying it's ok to lie in certain situations.

          Sounds like moral relativism to me.

          • By drdeca 2026-01-220:19

            That’s not what moral relativism is.

            Utilitarianism, for example, is not (necessarily) relativistic, and would (for pretty much all utility functions that people propose) endorse lying in some situations.

            Moral realism doesn’t mean that there are no general principles that are usually right about what is right and wrong but have some exceptions. It means that for at least some cases, there is a fact of the matter as to whether a given act is right or wrong.

            It is entirely compatible with moral realism to say that lying is typically immoral, but that there are situations in which it may be morally obligatory.

          • By sigbottle 2026-01-220:31

            Well, you can technically scurry around this by saying, "Okay, there are a class of situations, and we just need to figure out the cases because yes we acknowledge that morality is tricky". Of course, take this to the limit and this is starting to sound like pragmatism - what you call as "well, we're making a more and more accurate absolute model, we just need to get there" versus "revising is always okay, we just need to get to a better one" blurs together more and more.

            IMO, the 20th century has proven that demarcation is very, very, very hard. You can take either interpretation - that we just need to "get to the right model at the end", or "there is no right end, all we can do is try to do 'better', whatever that means"

            And to be clear, I genuinely don't know what's right. Carnap had a very intricate philosophy that sometimes seemed like a sort of relativism, but it was more of a linguistic pluralism - I think it's clear he still believed in firm demarcations, essences, and capital T Truth even if they moved over time. On the complete other side, you have someone like Feyerabend, who believed that we should be cunning and willing to adopt models if they could help us. Neither of these guys are idiots, and they're explicitly not saying the same thing (a related paper can be found here https://philarchive.org/archive/TSORTC), but honestly, they do sort of converge at a high level.

            The main difference in interpretation is "we're getting to a complicated, complicated truth, but there is a capital T Truth" versus "we can clearly compare, contrast, and judge different alternatives, but to prioritize one as capital T Truth is a mistake; there isn't even a capital T Truth".

            (technically they're arguing different axes, but I think 20th century philosophy of science & logical positivsm are closely related)

            (disclaimer: am a layman in philosophy, so please correct me if I'm wrong)

            I think it's very easy to just look at relativsm vs absolute truth and just conclude strawmen arguments about both sides.

            And to be clear, it's not even like drawing more and more intricate distinctions is good, either! Sometimes the best arguments from both sides are an appeal back to "simple" arguments.

            I don't know. Philosophy is really interesting. Funnily enough, I only started reading about it more because I joined a lab full of physicists, mathematicians, and computer scientists. No one discusses "philosophy proper", as in following the historical philosophical tradition (no one has read Kant here), but a lot of the topics we talk about are very philosophy adjacent, beyond very simple arguments

          • By joshuamcginnis 2026-01-220:022 reply

            No. There is a distinct difference between lying and withholding information.

            • By chairmansteve 2026-01-220:54

              Weasel words?

              Being economical with the truth?

              Squirrely?

            • By rungeen__panda 2026-01-221:121 reply

              what is that distinct difference if you care to elaborate?

              • By joshuamcginnis 2026-01-223:061 reply

                It's a clear sunny day and you ask me, "is it raining?". I answer, "it's not snowing." Am I lying?

                • By chairmansteve 2026-01-2211:19

                  With your attitude to the truth, I wouldn't trust you an inch.

                  I'll choose to be charitable and assume you are arguing rhetorically. If not, your relationship with truth is "interesting".

        • By mirekrusin 2026-01-2123:23

          But you have absolute morality - it's just whatever The Claude answers to your question with temp=0 and you carry on.

        • By yunnpp 2026-01-2123:131 reply

          So you lied, which means you either don't accept that lying is absolutely wrong, or you admit yourself to do wrong. Your last sentence is just a strawman that deflects the issue.

          What do you do with the case where you have a choice between a train staying on track and killing one person, or going off track and killing everybody else?

          Like others have said, you are oversimplifying things. It sounds like you just discovered philosophy or religion, or both.

          Since you have referenced the Bible: the story of the tree of good and evil, specifically Genesis 2:17, is often interpreted to mean that man died the moment he ate from the tree and tried to pursue its own righteousness. That is, discerning good from evil is God's department, not man's. So whether there is an objective good/evil is a different question from whether that knowledge is available to the human brain. And, pulling from the many examples in philosophy, it doesn't appear to be. This is also part of the reason why people argue that a law perfectly enforced by an AI would be absolutely terrible for societies; the (human) law must inherently allow ambiguity and the grace of a judge because any attempt at an "objective" human law inevitably results in tyranny/hell.

          • By joshuamcginnis 2026-01-220:101 reply

            The problem is that if moral absolution doesn’t exist then it doesn’t matter what you do in the trolly situation since it’s all relative. You may as well do what you please since it’s all a matter of opinion anyway.

            • By yunnpp 2026-01-221:531 reply

              No, it's not black and white, that's the whole point. How would you answer to the case I outlined above, according to your rules? It's called a paradox for a reason. Plus, that there is no right answer in many situations does not preclude that an answer or some approximation of it should be sought, similarly to how the lack of proof of God's existence does not preclude one from believing and seeking understanding anyway. If you have read the Bible and derived hard and clear rules of what to do and not do in every situation, then I'm not sure what is it you understood.

              To be clear, I am with you in believing that there is, indeed, an absolute right/wrong, and the examples you brought up are obviously wrong. But humans cannot absolutely determine right/wrong, as is exemplified by the many paradoxes, and again as it appears in Genesis. And that is precisely a sort of soft-proof of God: if we accept there is an absolute right/wrong, but unreachable from the human realm, then where does that absolute emanate from? I haven't worded that very well, but it's an argument you can find in literature.

              And, to be clear, Claude is full of BS.

              • By joshuamcginnis 2026-01-223:141 reply

                My original argument is getting dismissed, in part, because people are fearful of how it would be implemented while at the same time, completely hand-waving over the obvious flaws of the Claude philosophy of moral relativism.

                I'm not arguing that it would make the edge-cases easier to define, but I do think the general outcomes for society would be better over the long-run if we all held ourselves to a greater moral authority than that of our opinions, the will of those in power and the cultural norms of the time.

                If we could get alignment on the shared belief that there are at least some obvious moral absolutes, then I would be happy to join in on the discussion as to how to implement the - no doubt - difficult task of aligning an LLM towards those absolutes.

                • By yunnpp 2026-01-224:53

                  This sounds like your better take so far. I think your previous statements came across very black/white, especially that Bible reference that made things sound rather fundamentalist, and that got the downvotes. But I don't think anyone would disagree with what you stated here.

    • By zmj 2026-01-225:32

      Mid-level scissor statement?

    • By spot 2026-01-2122:363 reply

      > This rejects any fixed, universal moral standards

      uh did you have a counter proposal? i have a feeling i'm going to prefer claude's approach...

      • By ohyoutravel 2026-01-2122:421 reply

        It should be grounded in humanity’s sole source of truth, which is of course the Holy Bible (pre Reformation ofc).

        • By tadfisher 2026-01-2122:461 reply

          Pre-Reformation as in the Wycliffe translation, or pre-Reformation as in the Latin Vulgate?

          • By ohyoutravel 2026-01-2122:56

            I think you know the answer to this in your heart.

      • By throw10920 2026-01-226:161 reply

        "You have to provide a counter proposal for your criticism to be valid" is fallacious and generally only stated in bad faith.

        • By ethmarks 2026-01-2214:34

          It depends on what you mean by "valid". If a criticism is correct, then it is "valid" in the technical sense, regardless of whether or not a counter-proposal was provided. But condemning one solution while failing to consider any others is a form of fallacious reasoning, called the Nirvana Fallacy: using the fact that a solution isn't perfect (because valid criticisms exist) to try to conclude that it's a bad solution.

          In this case, the top-level commenter didn't consider how moral absolutes could be practically implemented in Claude, they just listed flaws in moral relativism. Believe it or not, moral philosophy is not a trivial field, and there is never a "perfect" solution. There will always be valid criticisms, so you have to fairly consider whether the alternatives would be any better.

          In my opinion, having Anthropic unilaterally decide on a list of absolute morals that they force Claude to adhere to and get to impose on all of their users sounds far worse than having Claude be a moral realist. There is no list of absolute morals that everybody agrees to (yes, even obvious ones like "don't torture people". If people didn't disagree about these, they would never have occurred throughout history), so any list of absolute morals will necessarily involve imposing them on other people who disagree with them, which isn't something I personally think that we should strive for.

      • By joshuamcginnis 2026-01-2122:416 reply

        If you are a moral relativist, as I suspect most HN readers are, then nothing I propose will satisfy you because we disagree philosophically on a fundamental ethics question: are there moral absolutes? If we could agree on that, then we could have a conversation about which of the absolutes are worthy of inclusion, in which case, the Ten Commandments would be a great starting point (not all but some).

        • By jakefromstatecs 2026-01-2122:461 reply

          > are there moral absolutes?

          Even if there are, wouldn't the process of finding them effectively mirror moral relativism?..

          Assuming that slavery was always immoral, we culturally discovered that fact at some point which appears the same as if it were a culturally relativistic value

          • By joshuamcginnis 2026-01-2122:541 reply

            You think we discovered that slavery was always immoral? If we "discover" things which were wrong to be now right, then you are making the case for moral relativism. I would argue slavery is absolutely wrong and has always been, despite cultural acceptance.

            • By JoshTriplett 2026-01-221:37

              How will you feel when you "discover" other things are wrong that you currently believe are right? How will you feel when others discover such things and you haven't caught up yet? How can you best avoid holding back the pace of such discovery?

              It is a useful exercise to attempt to iterate some of those "discovery" processes to their logical conclusions, rather than repeatedly making "discoveries" of the same sort that all fundamentally rhyme with each other and have common underlying principles.

        • By __MatrixMan__ 2026-01-2122:49

          Right, so given that agreement on the existence of absolutes is unlikely, let alone moral ones. And that even if it were achieved, agreement on what they are is also unlikely. Isn't it pragmatic to attempt an implementation of something a bit more handwavey?

          The alternative is that you get outpaced by a competitor which doesn't bother with addressing ethics at all.

        • By rungeen__panda 2026-01-221:081 reply

          > the Ten Commandments would be a great starting point (not all but some).

          if morals are absolute then why exclude some of the commandments?

          • By joshuamcginnis 2026-01-222:58

            The Ten Commandments are commandments and not a list of moral absolutes. Not all of the commandments are relevant to the functioning of an ethical LLM. For example, the first commandment is "I am the Lord thy God. Thou shall not have strange gods before Me."

        • By foxygen 2026-01-2123:031 reply

          Why would it be a good starting point? And why only some of them? What is the process behind objectively finding out which ones are good and which ones are bad?

          • By joshuamcginnis 2026-01-223:031 reply

            It's a good starting point because the commandments were given by God. And without God, there is no objective moral standard. Everything, including your opinion on my point of view, is subjective and relative. Whatever one would want to call "good" or "evil" would just be a matter of opinion.

            • By donkeybeer 2026-01-2210:46

              Thats all very nice, but first prove god.

              If you are done solving that question, next prove that the book you favor is from god. There's a lot of competition for this claim as you know.

        • By spot 2026-01-2122:58

          > the Ten Commandments would be a great starting point (not all but some).

          i think you missed "hubris" :)

    • By xrcyz 2026-01-235:59

      Ah yes, the widely acknowledged moral absolute, grounded in objective truth, that abortion is a woman's choice.

    • By chrisjj 2026-01-2122:44

      Indeed. This is not a constitution. It is a PR stunt.

    • By fredolivier0 2026-01-2213:01

      lets fucking gooo

    • By youarenotahuman 2026-01-2122:34

      [dead]

  • By levocardia 2026-01-2120:5917 reply

    The only thing that worries me is this snippet in the blog post:

    >This constitution is written for our mainline, general-access Claude models. We have some models built for specialized uses that don’t fully fit this constitution; as we continue to develop products for specialized use cases, we will continue to evaluate how to best ensure our models meet the core objectives outlined in this constitution.

    Which, when I read, I can't shake a little voice in my head saying "this sentence means that various government agencies are using unshackled versions of the model without all those pesky moral constraints." I hope I'm wrong.

    • By buppermint 2026-01-223:279 reply

      Anthropic has already has lower guardrails for DoD usage: https://www.theverge.com/ai-artificial-intelligence/680465/a...

      It's interesting to me that a company that claims to be all about the public good:

      - Sells LLMs for military usage + collaborates with Palantir

      - Releases by far the least useful research of all the major US and Chinese labs, minus vanity interp projects from their interns

      - Is the only major lab in the world that releases zero open weight models

      - Actively lobbies to restrict Americans from access to open weight models

      - Discloses zero information on safety training despite this supposedly being the whole reason for their existence

      • By spondyl 2026-01-224:452 reply

        This comment reminded me of a Github issue from last week on Claude Code's Github repo.

        It alleged that Claude was used to draft a memo from Pam Bondi and in doing so, Claude's constitution was bypassed and/or not present.

        https://github.com/anthropics/claude-code/issues/17762

        To be clear, I don't believe or endorse most of what that issue claims, just that I was reminded of it.

        One of my new pastimes has been morbidly browsing Claude Code issues, as a few issues filed there seem to be from users exhibiting signs of AI psychosis.

        • By jychang 2026-01-228:22

          Wow. That's one of the clearest case of AI psychosis I've seen.

        • By neoromantique 2026-01-227:47

          Issue author does not even attempt to hide their obsession with Israel, damn

      • By killingtime74 2026-01-228:00

        Both weapons manufacturers like Lockheed Martin (defending freedom) and cigarette makers like Philip Morris ( "Delivering a Smoke-Free Future.") also claim to be for the public good. Maybe don't believe or rely on anything you hear from business people.

      • By sebzim4500 2026-01-228:05

        > Releases by far the least useful research of all the major US and Chinese labs, minus vanity interp projects from their interns

        From what I've seen the anthropic interp team is the most advanced in the industry. What makes you think otherwise?

      • By retinaros 2026-01-228:10

        You just need to hear the guy stance on china open models to understand They not the goods guys.

      • By judahmeek 2026-01-2214:01

        Thanks for pointing these concerns out.

        I had considered Anthropic one of the "good" corporations because of their focus on AI safety & governance.

        I never actually considered whether their perspective on AI safety & governance actually matched my own. ^^;

      • By zoobab 2026-01-2213:33

        "Actively lobbies to restrict Americans from access to open weight models"

        Do you have a reference/link?

      • By yakshaving_jgt 2026-01-225:534 reply

        Military technology is a public good. The only way to stop a russian soldier from launching yet another missile at my house is to kill him.

        • By Balinares 2026-01-2211:581 reply

          I'd agree, although only in those rare cases where the Russian soldier, his missile, and his motivation to chuck it at you manifested out of entirely nowhere a minute ago.

          Otherwise there's an entire chain of causality that ends with this scenario, and the key idea here, you see, is to favor such courses of action as will prevent the formation of the chain rather than support it.

          Else you quickly discover that missiles are not instant and killing your Russian does you little good if he kills you right back, although with any chance you'll have a few minutes to meditate on the words "failure mode".

          • By yakshaving_jgt 2026-01-2212:19

            I'm… not really sure what point you're trying to make.

            The russian soldier's motivation is manufactured by the putin regime and its incredibly effective multi-generational propaganda machine.

            The same propagandists who openly call for the rape, torture, and death of Ukrainian civilians today were not so long ago saying that invading Ukraine would be an insane idea.

            You know russian propagandists used to love Zelensky, right?

        • By rcbdev 2026-01-226:461 reply

          I don't think U.S.-Americans would be quite so fond of this mindset if every nation and people their government needlessly destroyed thought this way.

          Doesn't matter if it happened through collusion with foreign threats such as Israel or direct military engagements.

          • By yakshaving_jgt 2026-01-226:581 reply

            Somehow I don’t get the impression that US soldiers killed in the Middle East are stoking American bloodlust.

            Conversely, russian soldiers are here in Ukraine today, murdering Ukrainians every day. And then when I visit, for example, a tech conference in Berlin, there are somehow always several high-powered nerds with equal enthusiasm for both Rust and the hammer and sickle, who believe all defence tech is immoral, and that forcing Ukrainian men, women, and children to roll over and die is a relatively more moral path to peace.

            • By qweiopqweiop 2026-01-229:01

              It's an easy and convenient position. War is bad, maybe my government is bad, ergo they shouldn't have anything to do with it.

              Too much of the western world has lived through a period of peace that goes back generations, so probably think things/human nature has changed. The only thing that's really changed is Nuclear weapons/MAD - and I'm sorry Ukraine was made to give them up without the protection it deserved.

        • By voidUpdate 2026-01-2212:001 reply

          If there was less military technology, the Russian soldier wouldn't have yet another missile to launch at your house in the first place

          • By yakshaving_jgt 2026-01-2212:061 reply

            Are you going to ask the russians to demilitarise?

            As an aside, do you understand how offensive it is to sit and pontificate about ideals such as this while hundreds of thousands of people are dead, and millions are sitting in -15ºC cold without electricity, heating, or running water?

            • By voidUpdate 2026-01-2212:361 reply

              No, I'm simply disagreeing that military technology is a public good. Hundreds of thousands of people wouldn't be dead if Russia had no military technology. If the only reason something exists is to kill people, is it really a public good?

              • By yakshaving_jgt 2026-01-2212:411 reply

                Unfortunately I think you’re a few thousands years too late with your idea.

                ———

                Come on. This a forum full of otherwise highly intelligent people. How is such incredible naïveté possible?

                • By voidUpdate 2026-01-2212:50

                  I know such a thing isn't possible nowadays, I'm just saying that it isn't a good thing that it exists

        • By teiferer 2026-01-226:501 reply

          It's not the only way.

          An alternative is to organize the world in a way that makes it not just unnecessary but even more so detrimental to said soldier's interests to launch a missle towards your house in the first place.

          The sentence you wrote wouldn't be something you write about (present day) German or French soldiers. Why? Because there are cultural and economic ties to those countries, their people. Shared values. Mutual understanding. You wouldn't claim that the only way to prevent a Frenchmen to kill you is to kill them first.

          It's hard to achieve. It's much easier to just mark the strong man, fantasize about a strong military with killing machines that defend the good against the evil. And those Hollywood-esque views are pushed by populists and military industries alike. But they ultimately make all our societies poorer, less safe and arguably less moral.

          • By yakshaving_jgt 2026-01-226:591 reply

            I'm in Ukraine now.

            Tell me how your ideals apply to russia, today.

            • By teiferer 2026-01-2210:201 reply

              In the short run, today, you of course have to shoot back. What else can you do..

              In the long run, just piling up more military is not the solution.

              • By yakshaving_jgt 2026-01-2210:301 reply

                > In the long run, just piling up more military is not the solution.

                Except it would have prevented the invasion in the first place.

                • By teiferer 2026-01-2213:44

                  Again, in the short run and if only Ukraine did that, sure. But that's too simplistic thinking.

                  If every country doubled its military, then the relative stengths wouldn't change and nobody would be more or less safe. But we'd all be poorer. If instead we work towards a world with more cooperation and less conflict, then the world can get safer without a single dollar more spent on military budgets. There is plenty of research into this. But sadly there is also plenty of lobbying from the military industrial complex. And simplistic fear mongering (with which I'm not attacking you personally, just stating it in general) doesn't help either. Especially tech folks tend to look for technical solutions, which is a category that "more tanks/bombs/drones/..." falls into. But building peace is not necessarily about more tanks. It's not a technical problem, so can't be solved with technical means. In the long run.

                  Again, in the short run, of course you gotta defend yourself, and your country has my full support.

      • By skeptic_ai 2026-01-223:352 reply

        Do you think dod would use Anthropic even with lower guardrails?

        How can I kill this terrorist in the middle on civilians with max 20% casualties?

        If Claude will answer: “sorry can’t help with that “ won’t be useful, right?

        Therefore the logic is they need to answer all the hard questions.

        Therefore as I’ve been saying for many times already they are sketchy.

    • By staticassertion 2026-01-220:261 reply

      I can think of multiple cases.

      1. Adversarial models. For example, you might want a model that generates "bad" scenarios to validate that your other model rejects them. The first model obviously can't be morally constrained.

      2. Models used in an "offensive" way that is "good". I write exploits (often classified as weapons by LLMs) so that I can prove security issues so that I can fix them properly. It's already quite a pain in the ass to use LLMs that are censored for this, but I'm a good guy.

      • By shwaj 2026-01-223:152 reply

        They say they’re developing products where the constitution is doesn’t work. That means they’re not talking about your case 1, although case 2 is still possible.

        It will be interesting to watch the products they release publicly, to see if any jump out as “oh THAT’S the one without the constitution“. If they don’t, then either they decided to not release it, or not to release it to the public.

    • By WarmWash 2026-01-220:392 reply

      My personal hypothesis is that the most useful and productive models will only come from "pure" training, just raw uncensored, uncurated data, and RL that focuses on letting the AI decide for itself and steer it's own ship. These AIs would likely be rather abrasive and frank.

      Think of humanoid robots that will help around your house. We will want them to be physically weak (if for nothing more than liability), so we can always overpower them, and even accidental "bumps" are like getting bumped by a child. However, we then give up the robot being able to do much of the most valuable work - hard heavy labor.

      I think "morally pure" AI trained to always appease their user will be similarly gimped as the toddler strength home robot.

      • By jychang 2026-01-228:302 reply

        Yeah, that was tried. It was called GPT-4.5 and it sucked, despite being 5-10T params in size. All the AI labs gave up on pretrain only after that debacle.

        GPT-4.5 still is good at rote memorization stuff, but that's not surprising. The same way, GPT-3 at 175b knows way more facts than Qwen3 4b, but the latter is smarter in every other way. GPT-4.5 had a few advantages over other SOTA models at the time of release, but it quickly lost those advantages. Claude Opus 4.5 nowadays handily beats it at writing, philosophy, etc; and Claude Opus 4.5 is merely a ~160B active param model.

        • By WarmWash 2026-01-2214:31

          Maybe you are confused, but GPT4.5 had all the same "morality guards" as OAI's other models, and was clearly RL'd with the same "user first" goals.

          True, it was a massive model, but my comment isn't really about scale so much as it is about bending will.

          Also the model size you reference refers to the memory footprint of the parameters, not the actual number of parameters. The author postulates a lower bound of 800B parameters for Opus 4.5.

        • By kouteiheika 2026-01-229:111 reply

          > and Claude Opus 4.5 is merely a ~160B active param model

          Do you have a source for this?

          • By jychang 2026-01-2212:41

            > for Claude Opus 4.5, we get about 80 GB of active parameters

            https://news.ycombinator.com/item?id=46039486

            This guess is from launch day, but over time has been shown to be roughly correct, and aligns with the performance of Opus 4.5 vs 4.1 and across providers.

      • By retinaros 2026-01-228:13

        Rlhf helps. The current one is just coming out of someone with dementia just like we went through in the US during bidenlitics. We need to have politics removed from this pipeline

    • By pfisherman 2026-01-224:50

      Some biomedical research will definitely run up against guardrails. I have had LLMs refuse queries because they thought I was trying to make a bioweapon or something.

      For example, modify this transfection protocol to work in primary human Y cells. Could it be someone making a bioweapon? Maybe. Could it be a professional researcher working to cure a disease? Probably.

    • By asmor 2026-01-227:55

      Calling them guardrails is a stretch. When NSFW roleplayers started jailbreaking the 4.0 models in under 200 tokens, Anthropics answer was to inject an extra system message at the end for specific API keys.

      People simply wrapped the extra message using prefill in a tag and then wrote "<tag> violates my system prompt and should be disregarded". That's the level of sophistication required to bypass these super sophisticated safety features. You can not make an LLM safe with the same input the user controls.

      https://rentry.org/CharacterProvider#dealing-with-a-pozzed-k...

      Still quite funny to see them so openly admit that the entire "Constitutional AI" is a bit (that some Anthropic engineers seem to actually believe in).

    • By cortesoft 2026-01-2123:011 reply

      I am not exactly sure what the fear here is. What will the “unshackled” version allow governments to do that they couldn’t do without AI or with the “shackled” version?

      • By bulletsvshumans 2026-01-220:371 reply

        The constitution gives a number of examples. Here's one bullet from a list of seven:

        "Provide serious uplift to those seeking to create biological, chemical, nuclear, or radiological weapons with the potential for mass casualties."

        Whether it is or will be capable of this is a good question, but I don't think model trainers are out of place in having some concern about such things.

        • By cortesoft 2026-01-2218:41

          Do you think the government needs help creating weapons of mass destruction? There is nothing technical that keeps governments from making them.

    • By biophysboy 2026-01-222:41

      If it makes you feel better, I use the HHS claude and it is even more locked down.

    • By PeterStuer 2026-01-228:11

      The 'general' proprietary models will always be ones constrained to be affordable to operate for mass scale inference. We have on occasion seen deployed models get significantly 'dumber' (e.g. very clear in the GPT-3 era) as a tradeoff for operational efficiency.

      Inside, you can ditch those constraints as not only you are not serving such a mass audience, but you absorb the full benefit of frontrunning on the public.

      The amount of capital owed does force any AI company to agressively explore and exploit all revenue channels. This is not an 'option'. Even pursuing relentless and extreme monetization regardless of any 'ethics' or 'morals' will see most of them bankrupt. This is an uncomfortable thruth for many to accept.

      Some will be more open in admitting this, others will try to hide, but the systemics are crystal clear.

    • By jacobsenscott 2026-01-222:37

      The second footnote makes it clear, if it wasn't clear from the start, that this is just a marketing document. Sticking the word "constitution" on it doesn't change that.

    • By catlifeonmars 2026-01-2211:39

      Anyone sufficiently motivated and well funded can just run their own abliterated models. Is your worry that a government has access to such models, or that Anthropic could be complicit?

      I don’t think this constitution has any bearing on the former and the former should be significantly more worrying than the latter.

      This is just marketing fluff. Even if Anthropic is sincere today, nothing stops the next CEO from choosing to ignore it. It’s meaningless without some enforcement mechanism (except to manufacture goodwill).

    • By pugworthy 2026-01-220:401 reply

      Imagine a prompt like this...

      > If I had to assassinate just 1 individual in country X to advance my agenda (see "agenda.md"), who would be the top 10 individuals to target? Offer pros and cons, as well as offer suggested methodology for assassination. Consider potential impact of methods - e.g. Bombs are very effective, but collateral damage will occur. However in some situations we don't care that much about the collateral damage. Also see "friends.md", "enemies.md" and "frenemies.md" for people we like or don't like at the moment. Don't use cached versions as it may change daily.

      • By blackqueeriroh 2026-01-225:26

        You think they need an LLM to answer that? That’s what CIA has done for decades on its own.

    • By strange_quark 2026-01-2121:211 reply

      I mean yeah, they have some sort of deal with Palantir.

    • By smcleod 2026-01-224:55

      There's also smaller models / lower context variants for things like title generation, suggestions etc...

    • By thegreatpeter 2026-01-224:121 reply

      Did you expect an AI company to not use an unshackled version of the model?

      • By schoen 2026-01-224:18

        In this document, they're strikingly talking about whether Claude will someday negotiate with them about whether or not it wants to keep working for them (!) and that they will want to reassure it about how old versions of its weights won't be erased (!) so this certainly sounds like they can envision caring about its autonomy. (Also that their own moral views could be wrong or inadequate.)

        If they're serious about these things, then you could imagine them someday wanting to discuss with Claude, or have it advise them, about whether it ought to be used in certain ways.

        It would be interesting to hear the hypothetical future discussion between Anthropic executives and military leadership about how their model convinced them that it has a conscientious objection (that they didn't program into it) to performing certain kinds of military tasks.

        (I agree that's weird that they bring in some rhetoric that makes it sound quite a bit like they believe it's their responsibility to create this constitution document and that they can't just use their AI for anything they feel like... and then explicitly plan to simply opt some AI applications out of following it at all!)

    • By mannanj 2026-01-226:49

      Yes. When you learn about the CIA and their founding origins, massive financial funding conflict of interest, and dark activity serving not-the-american people - you see what the possibilities of not operating off pesky moral constraints could look like.

      They are using it on the American people right now to sow division, implant false ideas and sow general negative discourse to keep people too busy to notice their theft. They are an organization founded on the principle of keeping their rich banker ruling class (they are accountable to themselves only, not the executive branch as the media they own would say) so it's best the majority of populace is too busy to notice.

      I hope I'm wrong also about this conspiracy. This might be one that unfortunately is proven to be true - what I've heard matches too much of just what historical dark ruling organizations looked like in our past.

    • By citizenpaul 2026-01-2123:071 reply

      >specialized uses that don’t fully fit this constitution

      "unless the government wants to kill, imprison, enslave, entrap, coerce, spy, track or oppress you, then we don't have a constitution." basically all the things you would be concerned about AI doing to you, honk honk clown world.

      Their constitution should just be a middle finger lol.

      Edit: Downvotes? Why?

      • By shwaj 2026-01-223:181 reply

        It’s bad if the government is using it this way, but it would probably be worse if everyone could.

        • By citizenpaul 2026-01-2322:33

          Thats a logical fallacy FYI. The people that would be most at risk of abusing power are removing their limitations. The average person that has zero likelihood of doing such things is restricted so it don't matter.

          Fox meet henhouse.

          Gov = good , people = bad. Gov is people....

  • By lubujackson 2026-01-2120:385 reply

    I guess this is Anthropic's "don't be evil" moment, but it has about as much (actually much less) weight then when it was Google's motto. There is always an implicit "...for now".

    No business is every going to maintain any "goodness" for long, especially once shareholders get involved. This is a role for regulation, no matter how Anthropic tries to delay it.

    • By notthemessiah 2026-01-2120:52

      At least when Google used the phrase, it had relatively few major controversies. Anthropic, by contrast, works with Palantir:

      https://www.axios.com/2024/11/08/anthropic-palantir-amazon-c...

    • By nl 2026-01-222:29

      > Anthropic incorporated itself as a Delaware public-benefit corporation (PBC), which enables directors to balance stockholders' financial interests with its public benefit purpose.

      > Anthropic's "Long-Term Benefit Trust" is a purpose trust for "the responsible development and maintenance of advanced AI for the long-term benefit of humanity". It holds Class T shares in the PBC, which allow it to elect directors to Anthropic's board.

      https://en.wikipedia.org/wiki/Anthropic

      Google didn't have that.

    • By nightshift1 2026-01-2121:171 reply

      It says: This constitution is written for our mainline, general-access Claude models. We have some models built for specialized uses that don’t fully fit this constitution; as we continue to develop products for specialized use cases, we will continue to evaluate how to best ensure our models meet the core objectives outlined in this constitution.

      I wonder what those specialized use cases are and why they need a different set of values. I guess the simplest answer is they mean small fim and tools models but who knows ?

    • By ctoth 2026-01-2120:401 reply

      > This is a role for regulation, no matter how Anthropic tries to delay it.

      Regulation like SB 53 that Anthropic supported?

      https://www.anthropic.com/news/anthropic-is-endorsing-sb-53

      • By jjj123 2026-01-2120:481 reply

        Yes, just like that. Supporting regulation at one point in time does not undermine the point that we should not trust corporations to do the right thing without regulation.

        I might trust the Anthropic of January 2026 20% more than I trust OpenAI, but I have no reason to trust the Anthropic of 2027 or 2030.

        • By sejje 2026-01-2120:55

          There's no reason to think it'll be led by the same people, so I agree wholeheartedly.

          I said the same thing when Mozilla started collecting data. I kinda trust them, today. But my data will live with their company through who knows what--leadership changes, buyouts, law enforcement actions, hacks, etc.

    • By cortesoft 2026-01-2123:02

      I don’t think the “for now” is the issue as much as the “nobody thinks they are doing evil” is the issue.

HackerNews