Morality in the tool box

A new study has some interesting results pointing to resistance to AI stemming from moral and not just “practical” concerns. Pre-publication article here

While I might disagree that moral concerns are separable from “practical” or “pragmatic” ones on the grounds that the practical results of one’s actions have a moral valence, I do think this is an interesting set of findings and hope to see more research on moral, ethical, and even religious reactions to advanced technologies.

I’ll offer one way to think about this. Tools (technologies, things) aren’t moral actors in their own right, but their creation and use has moral implications. The more specific a tool, the more the morality that led to its development comes into focus.

Anthropologist Margaret Mead is widely (and probably incorrectly) credited with saying that the first sign of civilization in an ancient culture is a healed femur. Whoever actually made this claim should’ve taken credit because they created an interesting thought experiment. There’s a moral claim being made that places the beginning of human civilization at the ability and willingness to share someone else’s burden and care for a community member who, left to the course of nature, would surely die before their bone could heal. There is certainly something appealing about implicitly saying the dawn of human civilization began with a very specific kind of tool: a splint. This new technology made such healing possible. The splint, a tool whose only purpose is to heal an injury and, on top of that, assumes someone else is able and willing to take on the survival load of the injured person has moral urgency. As a fable of community, civilization, and technology, this is quite compelling. For the purposes of morality and technology, this story creates one pole for moral technology: The morality of the splint is one of community, healing, and shared endeavors.

The morality of the sword is quite different and a good example of the other pole of morality and technology. As much as I like swords (I was a collegiate fencer and I do enjoy DnD as well as medieval books and movies) it is hard to avoid understanding their purpose. The morality of the sword stems from the idea that killing other humans is permissible and even sometimes desirable. At inception, a sword had no other purpose, there are far better, cheaper, and easier tools for building a house, excavating valuable resources, chopping wood or any other creative endeavor. Even for defense against animals or people, spears, clubs, and similar simpler tools can work better for most folks. The sword requires specialization to produce and to use effectively. It would seem to encourage, if not outright demand, hierarchy. To put it simply, swords are for specialists to use to kill people. The morality of the sword is one that elevates aggression in a zero sum game of survival and conquest.

Stark contrasts here for single-purpose, specialized tools. What does that mean for us today?

Many of our digital tools have multiple purposes and even their developers had multiple aims in their creation. So the morality, then, shifts to the use of the tool. Just like with humans, tools of general utility can widely vary in their moral implications. We, as people, project our morality on to these tools by how we use them. The most effective users will, intentionally or not, imprint their moral agendas onto these tools. Use, over time, will also create specializations in these tools that show off the morality that “won” among their various users. As the tools seem to do more “thinking” or at least have the autonomy to execute on decisions in ambiguous situations, the moral programming that guided their actual programming begins to have more “practical” or “pragmatic” import.

It’s no wonder, then, as we create tools that are more an more like us, more autonomous and wide ranging, we struggle with the moral implications of why our tools get created and what they will be allowed to do.

Title photo by Barn Images on Unsplash

Now nobody wants to steer the bus?!

Anthropic, the AI company, has long been trying to differentiate itself as the “safe” and “responsible” choice in AI. Where companies like OpenAI have long made it clear that safety or responsibility are not a priority, or maybe not even a concern, and AI efforts like Grok seem to be premised on the idea that irresponsibility is the right way to go, Anthropic sold itself as the company bucking that trend.

From Dario Amodei’s articles on the importance of interpretability to picking a fight with the Pentagon over surveillance and autonomous weapons (where I now think Antropic is more likely than not to cave), it seemed like Anthropic would be different. So it comes as a shock when Time Magazine reported that Anthropic was dropping its longstanding safety pledge. At its heart, the Responsible Scaling Policy (RSP) promised that Anthropic would never train an AI system without an advance guarantee that the company’s safety measures could compensate for anticipated risks from that system. The company hoped other developers would follow suit, but they didn’t. Now Anthropic has adopted the ‘if you can’t beat ’em, join ’em” approach.

It shouldn’t be too surprising that Anthropic chose to follow the money, with the rumor of an IPO this year, but it is disappointing. This is doubly true when there aren’t a lot of others willing to bear the pro-safety flag. The US, under the Trump administration, has declared a holiday from AI responsibility, even the EU seems to be backtracking. International organizations have followed suit as well. At this point, we’re left with the UN and the Pope demanding safety, but neither of them actually develop AI.

In his important article “The Urgency of Interpretability”, Anthropic leader Dario Amodei said:

perhaps the most important lesson I’ve learned is this: the progress of the underlying technology is inexorable, driven by forces too powerful to stop, but the way in which it happens—the order in which things are built, the applications we choose, and the details of how it is rolled out to society—are eminently possible to change, and it’s possible to have great positive impact by doing so.  We can’t stop the bus, but we can steer it.

It’s still true that we have the opportunity, privilege and responsibility to steer the future of technology, but that presupposes there’s someone with the courage to steer that bus.