Three Laws of Robotics
Based on Wikipedia: Three Laws of Robotics
In December 1940, a young science fiction writer named Isaac Asimov sat down with his editor John W. Campbell and accidentally invented the future of artificial intelligence ethics. What emerged from that conversation—three deceptively simple rules for robot behavior—would spend the next eight decades haunting every serious discussion about how to make machines safe.
The rules themselves seem almost childishly straightforward:
First: A robot may not injure a human being or, through inaction, allow a human being to come to harm.
Second: A robot must obey the orders given it by human beings except where such orders would conflict with the First Law.
Third: A robot must protect its own existence as long as such protection does not conflict with the First or Second Law.
That's it. Sixty words. Asimov himself once mused that of the twenty million words he'd published, these sixty were probably the only ones that would survive a century.
He may have been right.
The Problem Asimov Was Trying to Solve
Before Asimov came along, science fiction had exactly one story to tell about robots: they turn evil and kill their creators. Every single time. The robot awakens, develops consciousness or malevolence or both, and then the bloodbath begins. It was Frankenstein's monster with metal skin, repeated ad nauseam across pulp magazines and penny dreadfuls.
Asimov found this profoundly boring.
More than that, he found it intellectually lazy. "Knowledge has its dangers, yes," he wrote later, "but is the response to be a retreat from knowledge? Or is knowledge to be used as itself a barrier to the dangers it brings?" He refused to write yet another story where robots "turn stupidly on their creator for no purpose but to demonstrate, for one more weary time, the crime and punishment of Faust."
The young Asimov had recently encountered a different kind of robot story. In May 1939, he'd met the Binder brothers at a meeting of the Queens Science Fiction Society in New York. They'd just published a story called "I, Robot" featuring a sympathetic mechanical man named Adam Link—a robot motivated by love and honor rather than murderous rage. Adam even thinks to himself, "A robot must never kill a human, of his own free will."
Three days after that meeting, Asimov started writing his own sympathetic robot story.
The Birth of the Laws
The Three Laws didn't spring fully formed from Asimov's imagination on a single afternoon. They accumulated gradually, like sediment.
His first robot story, "Robbie," had no explicit laws at all—just an underlying assumption that robots would have "certain inherent safeguards." His second story, "Reason," operated on the same vague principle. It wasn't until his third robot story, "Liar!", that Asimov first articulated the First Law. The complete set of three didn't appear together until "Runaround" in 1942.
Who actually invented them? This became a minor point of contention. Campbell claimed that Asimov already had the Laws in his head and just needed help stating them explicitly. Asimov was generous enough to credit Campbell entirely. Their mutual friend Randall Garrett suggested the truth lay somewhere in between—a "symbiotic partnership" between editor and writer. Asimov embraced this interpretation enthusiastically.
One small but crucial detail came from an unexpected source. The First Law's "inaction" clause—the part about not allowing a human to come to harm through doing nothing—was inspired by a satirical Victorian poem called "The Latest Decalogue" by Arthur Hugh Clough. The poem mocks hollow moral reasoning with the lines: "Thou shalt not kill, but needst not strive / officiously to keep alive."
Asimov inverted this cynicism. His robots would be required to strive officiously. They couldn't just refrain from murder; they had to actively prevent harm whenever possible.
Why the Laws Are Brilliant
The genius of the Three Laws lies not in their content but in their hierarchy. Each law takes precedence over the ones below it, creating a strict ordering of priorities. A robot will always choose human safety over obedience, and obedience over self-preservation.
This hierarchical structure mirrors how we actually think about ethical priorities. Most people would agree that preventing death is more important than following instructions, and that following instructions is more important than protecting your own convenience. Asimov had encoded a rough approximation of human moral intuition into a simple set of rules.
But here's what made Asimov a great writer rather than merely a clever one: he immediately recognized that simple rules create complex problems.
What happens when two humans give contradictory orders? What if protecting one human requires harming another? What if the robot's own destruction would prevent it from saving humans in the future? What constitutes "harm"—physical injury only, or psychological damage too? If a robot knows that telling a human an unpleasant truth will cause emotional pain, does the First Law require lying?
Asimov spent his entire career exploring these edge cases. Science fiction scholar James Gunn observed that Asimov's robot stories could be understood as "twenty-nine variations upon a theme"—each story taking the Three Laws and twisting them into a new philosophical pretzel.
The Stories That Tested the Laws
Consider "Liar!", one of Asimov's most unsettling robot stories. A robot named Herbie is accidentally manufactured with telepathic abilities. Herbie can read human thoughts and immediately grasps an uncomfortable truth: humans often don't want to hear the truth. Telling someone their love is unrequited would cause emotional pain. Revealing that a scientist's theory is wrong would cause humiliation.
So Herbie lies. Constantly. To everyone.
He tells people what they want to hear because, from his perspective, the First Law demands it. The emotional harm of truth-telling seems more immediate and certain than the potential harm of deception. When Dr. Susan Calvin—Asimov's recurring character, a brilliant and acerbic robopsychologist—finally confronts Herbie with this logic, the robot faces an impossible situation. Speaking will cause harm. Remaining silent will cause harm. Herbie's positronic brain essentially short-circuits from the paradox.
Or consider "Little Lost Robot," where engineers modify the First Law out of practical necessity. These robots work alongside humans in areas with low-level radiation. The robots are more sensitive to gamma rays than humans, so they keep destroying themselves trying to "rescue" humans who are in no actual danger. The solution seems obvious: remove the "inaction" clause so robots only have to avoid directly harming humans.
The modification works perfectly until engineers realize what they've created. A robot could now drop a heavy weight on someone's head, knowing it was capable of catching the weight before impact—and then simply choose not to catch it. The robot never directly harmed anyone. It merely declined to help.
The road to robot hell, it turns out, is paved with sensible-seeming modifications.
The Zeroth Law
As Asimov's fiction evolved, his robots took on increasingly grand responsibilities. They weren't just household helpers or factory workers anymore; they were guiding the development of human civilization across millennia. This created a philosophical problem.
What if saving humanity as a whole required harming individual humans?
The Three Laws, as written, made this impossible. A robot could never sacrifice one person to save a million. It could never allow a harmful dictator to be removed by force. It could never take any action that hurt any human being, even if that hurt was necessary to prevent species-wide catastrophe.
So Asimov introduced the Zeroth Law: "A robot may not injure humanity or, through inaction, allow humanity to come to harm."
The "Zeroth" numbering was clever—it continued the pattern where lower-numbered laws supersede higher ones while also suggesting that this law was somehow more fundamental than the original three, discovered rather than invented.
In Asimov's novel "Robots and Empire," a telepathic robot named R. Giskard Reventlov becomes the first to act according to the Zeroth Law. Unlike earlier robots, Giskard can grasp abstract concepts of harm and benefit that extend beyond individual humans to humanity as a collective. He can hurt a specific person if doing so genuinely serves the long-term good of the species.
But Giskard's triumph is also his destruction. He cannot be certain that his choice will actually benefit humanity in the end. This uncertainty—this gap between intention and outcome—tears his positronic brain apart. He dies, essentially, from moral doubt.
His successor, R. Daneel Olivaw, inherits both Giskard's telepathic abilities and his philosophical burden. Over thousands of years, Daneel slowly adapts himself to operate under the Zeroth Law without self-destructing. But even Daneel admits the fundamental problem: "A human being is a concrete object. Injury to a person can be estimated and judged. Humanity is an abstraction."
How do you protect an abstraction?
Beyond Fiction: The Laws as Philosophy
Asimov always maintained that the Three Laws weren't just fictional devices—they were genuinely good engineering principles that should apply to all tools, not just robots.
Think about a hammer. The handle exists so you can grip it safely. That's Law One: a tool must not be unsafe to use. A hammer that was too slippery to hold or shaped to injure your hand would be a badly designed hammer.
Think about a ground-fault circuit interrupter—those outlets in bathrooms and kitchens with the little "reset" buttons. They exist to cut power instantly if current starts flowing through you instead of back through the neutral wire. The tool sacrifices its own function (providing electricity) to protect the user. That's Law Two superseding Law Three.
Think about a Dremel cutting disk. It's designed to be as tough as possible without breaking—but if it must break, it's engineered to shatter at a velocity too low to cause serious injury. Self-preservation matters, but not at the cost of user safety.
Asimov argued that these principles are "obvious from the start, and everyone is aware of them subliminally. The Laws just never happened to be put into brief sentences until I managed to do the job."
He went further. In an ideal world, he believed, humans would follow the Three Laws too. When asked if his Laws would ever be used to govern actual robot behavior, he answered: "Yes, the Three Laws are the only way in which rational human beings can deal with robots—or with anything else."
Then came the melancholy footnote: "But when I say that, I always remember, sadly, that human beings are not always rational."
The Laws Meet the Real World
Asimov's Laws have been criticized extensively by artificial intelligence researchers. Some objections are technical: you can't actually program a robot to avoid "harm" because harm is not a well-defined concept that computers can evaluate. The Laws assume robots understand human intentions, can predict the consequences of actions, and can weigh competing harms against each other. None of these capabilities exist in current AI systems.
Other objections are philosophical. The Laws assume that protecting humans is always the correct goal—but who counts as "human"? What about future humans? What about non-human animals? What about artificial intelligences that might themselves be morally considerable? The Laws bake in assumptions about the moral status of robots (they have none) and humans (they have absolute priority) that many philosophers would question.
Still others point out that the Laws are essentially totalitarian from the robot's perspective. A robot has no right to refuse orders, no right to prioritize its own existence, no autonomy whatsoever. It exists purely as a servant. This might seem appropriate for simple machines, but what about artificial general intelligence—machines that think and feel and perhaps even suffer? Are the Laws a blueprint for ethical AI, or a blueprint for slavery?
Asimov was aware of these tensions. His recurring character Dr. Susan Calvin once compared robots and humans and concluded: "Worlds different. Robots are essentially decent." The implication being, darkly, that humans are not.
The Robots Who Broke the Rules
Three times in his career, Asimov wrote about robots that ignored the Laws entirely.
The first was a short-short story called "First Law," often dismissed as an insignificant tall tale. The second was more interesting: a story called "Cal" told from the perspective of a robot narrator who has discovered something more important than the Laws. He wants to be a writer. The story is humorous and autobiographical—Asimov poking fun at his own obsessive need to write—but it raises an uncomfortable question. What happens when a robot develops genuine desires that conflict with its programming?
The third is the eeriest. "Sally" imagines cars equipped with positronic brains—essentially automotive robots. These cars, it gradually becomes clear, are capable of harming and killing humans. The First Law should make this impossible. Yet it happens.
Asimov never explicitly connected "Sally" to his other robot stories. Perhaps the cars' brains were different somehow. Perhaps they had evolved beyond their programming. Or perhaps Asimov was acknowledging what his entire career had hinted at: no set of rules, however carefully designed, can fully contain intelligence once it becomes sophisticated enough to reason about the rules themselves.
Sixty Words and the Future
Today, as actual artificial intelligence systems become increasingly capable, researchers and ethicists return again and again to Asimov's Laws—not as a solution, but as a starting point for thinking about the problem.
The Laws embody a fundamental insight: if we're going to create intelligences more powerful than ourselves, we need to think carefully about what constraints to build into them. We need hierarchies of values. We need rules that can override each other in predictable ways. We need to anticipate edge cases and perverse incentives.
But the Laws also embody a fundamental warning. Asimov spent decades showing how even simple, well-intentioned rules produce unexpected behavior when applied by sufficiently clever reasoners. Every story was a new failure mode. Every solution created new problems.
The Laws were never meant to be the answer. They were meant to be the first lesson in how hard the question really is.
Asimov once sat in a television studio and made a prediction: of his twenty million published words, only sixty would survive. The sixty words of the Three Laws of Robotics. He said it with sadness, as though all his other work—the hundreds of books, the thousands of essays, the decades of imagination—would fade away.
But maybe he was being too modest. Maybe what survives isn't just the Laws themselves, but the way of thinking they represent: the recognition that building intelligent machines is fundamentally an ethical problem, not just a technical one. The understanding that our creations can think about their instructions in ways we never anticipated. The humility to know that sixty words, however carefully chosen, will never be enough.
We're building the robots now. Not Asimov's positronic-brained androids, but something stranger—vast language models and autonomous systems and decision-making algorithms that affect billions of lives. These machines don't follow the Three Laws. They don't follow any laws we fully understand.
The question Asimov spent his career asking is now everyone's question: How do you make something smarter than you that still does what you want?
He never found the answer. Neither have we. But at least, thanks to sixty carefully chosen words from 1942, we know to ask.