Discussion about this post

User's avatar
Priank Ravichandar's avatar

I recently read Anthropic’s CEO’s essay “The Adolescence of Technology,” where he lays out his concerns about powerful AI systems. What I found unusual was how much of their approach to preventing bad outcomes seems to be pinned on their “constitution.” They seem to believe it will be instrumental in ensuring the model doesn’t go off the rails or get used for harmful purposes.

The anthropomorphic language is especially concerning because it assumes the model has the capacity to care about abstract things like values or principles, when it can’t be held accountable for the consequences of its actions in any meaningful way.

Barbara Z's avatar

I can't believe my eyes. As a lawyer, I have always supported Claude for the reasons you have mentioned, Luiza. Everything in this constitution from start to finish looks and sounds awful. I have checked who the authors were and the origins of this constitution. Seems the origins go back to this paper: https://arxiv.org/abs/2212.08073 (with Amanda Askell listed as one of them and who is the primary author of the "constitution") https://www.linkedin.com/pulse/qa-amanda-askell-lead-author-anthropics-new-constitution-oqtte/ "The old constitution was trying to move the model towards these kinds of high-level principles or traits. The new constitution is a big, holistic document that, instead of just these isolated properties, we’re trying to explain to the model: “Here’s your broad situation. Here’s the way that we want you to interact with the world. Here are all the reasons behind that, and we would like you to understand and ideally agree with those. Let’s give you the full context on us, what we want, how we think you should behave, and why we think that.”

to me it looks like 1) this has been years in the making 2) worst form of anthropomorphism i've ever could have imagined 3) a group of people trying to recreate a form of science-fiction in real life (and actually stir the systems in that direction, as you say, the model will be trained on this material). 4) the explanation seems to be: this will help to control AI behaviour...with all of this language it seems like we have given up on ai governance and humans turning the tap on and off altogether.

18 more comments...

No posts

Ready for more?