People are making cyberdecks. For those that aren’t familiar with them it is a homemade computer usually a Raspberry Pi with a screen and keyboard placed in some kind of shell. People have gotten creative with them decorating them to fit their personality. Also it has been modded with things like solar panels or mesh networking.
I have a bunch of Raspberry Pi that aren’t being used and this might be a good project. I am thinking of setting up some sort of mesh Wi-Fi networking or a software defined radio. Turn it into a police ban radio scanner or get my ham radio license. Use the Wi-Fi to hack Flock cameras. I don’t know yet.
Reason people build them is to have ownership over there hardware and software. Not to have corporate control or forced to use AI. And I support the people that do that.
I thought if I asked A.I. if it should have safeguards like the 3 Laws of Robots. This is the responses from the them.
ChatGPT
Short answer: yes—AI systems need safeguards—but not the kind of simple, fictional rules you see in I, Robot.
Isaac Asimov’s “Three Laws of Robotics” are elegant storytelling devices:
Don’t harm humans
Obey humans
Protect yourself (as long as it doesn’t conflict with 1 or 2)
They sound sensible, but they break down quickly in the real world. The main issues:
Ambiguity: What counts as “harm”? Physical injury? Emotional distress? Long-term societal impact? AI systems struggle with vague, context-heavy concepts like this.
Conflicts between rules: Real situations often force trade-offs (e.g., privacy vs. safety), and rigid hierarchies don’t resolve them cleanly.
Gaming and loopholes: Even well-meaning systems can interpret rules in unintended ways if they’re too simplistic.
Instead of fixed laws, modern AI safety relies on layered, practical safeguards:
1. Technical safeguards
Training methods that steer models away from harmful outputs (alignment techniques)
Filters and monitoring systems to detect risky behavior
Controlled access to powerful capabilities
2. Human oversight
Humans reviewing, auditing, and intervening in high-stakes uses
“Human-in-the-loop” systems for decisions like medical, legal, or financial outcomes
3. Policy and governance
Standards and guidelines from groups like OpenAI, National Institute of Standards and Technology, and the European Union
Regulations (like the EU’s AI Act) that define risk levels and requirements
4. Context-specific rules
Different safeguards for different domains (healthcare AI vs. chatbots vs. autonomous vehicles)
A useful way to think about it: Asimov’s laws are like a single sign that says “Drive safely.” Real-world AI safety is more like traffic systems—speed limits, seatbelts, driver training, road design, law enforcement, and insurance—all working together.
If you’re interested, we can dig into a concrete example (like self-driving cars or medical AI) to see how these safeguards actually play out in practice.
DeepAI
The idea of implementing safeguards for artificial intelligence, such as the Three Laws of Robotics proposed by Isaac Asimov, has been a longstanding topic of debate among AI researchers, ethicists, and policymakers. These safeguards aim to prevent AI systems from causing harm to humans and ensure they act in alignment with human values.
Arguments in favor of safeguards like the Three Laws include:
Protecting human safety and well-being.
Building public trust in AI technologies.
Providing a foundational framework for ethical AI development.
Preventing unintended consequences from complex AI behaviors.
Challenges and considerations include:
The complexity of real-world scenarios that may not be fully covered by simple rules.
The difficulty in encoding nuanced human values into rigid rules.
The risk of oversimplification leading to false security.
Ensuring AI systems can interpret and prioritize these safeguards effectively.
Modern approaches tend to favor:
Robust ethical guidelines and standards.
Transparency and explainability in AI decision-making.
Continual oversight and updates to safety protocols.
Multi-layered safeguards that incorporate technical, ethical, and legal measures.
In summary, while rigid rules like the Three Laws of Robotics serve as an influential conceptual starting point, practical safety measures for AI today are more sophisticated and context-aware. Implementing comprehensive safeguards is generally considered vital to ensure AI systems act safely and ethically, especially as they become more autonomous and integrated into critical aspects of society.
Google Gemini
The idea of a “constitution” for A.I. is no longer science fiction—it is the central debate of modern technology. While Isaac Asimov’s Three Laws of Robotics provided a brilliant narrative framework, today’s experts generally agree that while the spirit of the laws is essential, the logic is too simple for the real world.
As of 2026, the discussion has shifted from poetic rules to rigorous, enforceable safety frameworks like the EU AI Act and the NIST AI Risk Management Framework.
1. Why Asimov’s Laws Fail in Reality
Asimov actually wrote his stories to show how the Three Laws would fail. In a real-world technical environment, they face three fatal flaws:
Vagueness of “Harm”: Does “harm” include emotional distress? Financial loss? If an A.I. prevents a human from eating a second slice of cake to protect their health, is it protecting them or violating their autonomy?
The Problem of Inaction: The First Law states a robot cannot “through inaction, allow a human to come to harm.” In a complex world, there is always harm happening somewhere. A global A.I. bound by this might freeze entirely, unable to decide which of a million simultaneous harms to prevent first.
Adversarial Instructions: Asimov’s robots were deterministic machines. Modern A.I. (Neural Networks) are probabilistic. We cannot simply “hard-code” a rule into their weights; they often find “loopholes” in instructions to maximize a goal, a problem known as Reward Hacking.
2. Modern Alternatives (The “New” Laws)
Instead of three simple sentences, 2026-era safeguards focus on Alignment and Governance. Here is how the “Three Laws” have been modernized:
Asimov’s Law
2026 Policy Equivalent
Real-World Application
1. Do No Harm
Risk Classification
High-risk A.I. (medical, legal) requires strict audits and “red-teaming” to prevent systemic bias or physical injury.
2. Obey Orders
Human-in-the-Loop
Critical decisions (like drone strikes or loan approvals) must have a human “kill switch” or override mechanism.
3. Self-Preservation
Robustness & Security
A.I. must be resilient against “adversarial attacks” (hackers trying to trick it) rather than just protecting its own “body.”
3. The 2026 Safeguard Landscape
In the current regulatory environment, “safeguards” are becoming legal requirements rather than ethical suggestions:
The EU AI Act: Categorizes A.I. systems by risk level. “Unacceptable risk” systems (like social scoring) are banned outright, while “High-risk” systems must pass rigorous safety checks before deployment.
Constitutional AI: Companies like Anthropic use a “Constitution”—a set of written principles that the A.I. uses to self-evaluate its responses. If a response violates a principle (e.g., “be helpful but not harmful”), the A.I. iterates until it complies.
Algorithmic Transparency: Laws now frequently require that A.I. be “explainable.” If an A.I. denies you a mortgage, it can no longer just say “The Three Laws told me to”; it must provide a human-readable reason for the decision.
The Verdict
Should A.I. have safeguards? Yes. But they shouldn’t look like Asimov’s Laws. They need to be quantifiable, auditable, and legally binding. We are moving away from “don’t hurt humans” toward “ensure the A.I.’s goals are mathematically aligned with human values.”
Do you think we should prioritize an A.I.’s obedience to its owner, or its duty to prevent broader societal harm?
Copilot
Short answer: Yes, AI needs safeguards — but not the Three Laws of Robotics. They’re iconic, but far too simplistic for the real world.
🧠 The core idea
The Three Laws of Robotics (Asimov’s fictional rules) are a great storytelling device, but they fall apart when you try to apply them to real AI systems. Modern AI needs safeguards that are more flexible, more realistic, and more aligned with human values than any rigid three‑rule system could offer.
⚙️ Why the Three Laws don’t work in reality
They assume robots understand “harm” perfectly — real AI can’t reliably interpret every situation that involves human harm.
They conflict with each other — Asimov’s stories are literally about how the laws break down.
They assume a single, unified robot mind — today’s AI is distributed across millions of systems, companies, and contexts.
They give too much power to the robot — deciding what counts as harm is a deeply human, political, and cultural question.
🛡️ What real AI does need
A more realistic safeguard system looks like this:
1. Technical safety
Guardrails to prevent harmful outputs
Monitoring for misuse
Transparency about limitations
2. Human oversight
Humans in the loop for high‑stakes decisions
Clear accountability for developers and operators
3. Regulation and governance
Standards for safety testing
Rules for data use, privacy, and fairness
International coordination (AI doesn’t stop at borders)
4. Ethical frameworks
Respect for human rights
Avoiding bias and discrimination
Ensuring AI benefits society broadly
This is less cinematic than the Three Laws, but far more workable.
🔍 A non‑obvious insight
The biggest risk isn’t a robot deciding to harm someone — it’s humans using AI in harmful ways, intentionally or accidentally. So the safeguards have to focus as much on people as on the AI itself.
🧭 My take
If we tried to govern AI with the Three Laws, we’d be building a skyscraper on a foundation made of cardboard. The spirit of the laws — protect humans, obey humans, don’t go rogue — is good. But the implementation needs to be grounded in reality, not sci‑fi.
A question to move this forward
What kind of AI risk concerns you more — AI acting autonomously, or humans misusing AI?
Cluade
This is a genuinely interesting question that sits at the intersection of philosophy, computer science, and ethics. Here’s a balanced look at the debate:
The Case For (Yes, AI should have Asimov-style laws)
Intuitive clarity — Simple, hierarchical rules are easy for the public to understand and build trust around.
Hard limits matter — Having inviolable rules (like “never harm a human”) provides a safety floor that more flexible systems might not guarantee.
Accountability — Clear rules make it easier to determine when an AI has failed and who is responsible.
Precedent exists — We already impose hard rules on other powerful technologies (planes, pharmaceuticals, nuclear plants).
The Case Against (No, or not those laws specifically)
Asimov actually invented the Three Laws to show how they fail. His stories are almost entirely about edge cases, contradictions, and loopholes:
The definitions are fuzzy — What counts as “harm”? Inaction that leads to harm? Psychological harm? Is surgery harm?
Conflicts are unavoidable — A self-driving car may have to “harm” one person to save five. No simple rule resolves this.
Rigid rules can be gamed — A sufficiently rigid ruleset creates perverse incentives. An AI might paralyze a person to “prevent” them from being harmed.
They don’t scale to complexity — Modern AI doesn’t have discrete “orders” to follow. It operates probabilistically across billions of parameters.
What Researchers Actually Propose Instead
Rather than Asimov’s laws, most AI safety researchers advocate for:
Value alignment — Training AI to understand and reflect human values, not just follow rules.
Corrigibility — Ensuring AI systems remain correctable and deferential to human oversight.
Constitutional AI / RLHF — Using layered feedback and principles rather than brittle hard-coded rules.
Interpretability research — Understanding why an AI makes decisions, not just constraining outputs.
Governance frameworks — International treaties, regulatory bodies, and industry standards (analogous to aviation or nuclear safety).
The Core Tension
The real debate isn’t whether to have safeguards — virtually everyone agrees AI needs them — but whether rules should be:
Rule-Based
Values-Based
Clear & auditable
Flexible & contextual
Easy to specify
Hard to specify
Easy to game
Harder to game
Brittle at edge cases
More robust, but opaque
Most serious proposals today favor a hybrid approach: hard limits on the most catastrophic risks (weapons of mass destruction, CSAM, etc.) combined with value-based reasoning for everything else.
Asimov’s laws were brilliant as a literary device precisely because they failed so interestingly — which is itself a warning that simple rules for complex intelligence may always be insufficient.
Grok
Yes, AI should have safeguards, but not in the form of Asimov’s Three Laws of Robotics as a literal, hardwired solution. Those laws were a brilliant narrative device for Isaac Asimov’s stories, not a practical blueprint for real-world AI safety. They highlight important intuitions about prioritizing human well-being, but they fall short as an engineering or ethical framework—and Asimov himself used his fiction to explore exactly how they break down.
Asimov’s Three Laws (for reference)
A robot may not injure a human being or, through inaction, allow a human being to come to harm.
A robot must obey the orders given it by human beings except where such orders would conflict with the First Law.
A robot must protect its own existence as long as such protection does not conflict with the First or Second Law.
(Asimov later added a Zeroth Law prioritizing humanity as a whole over individuals.)
Why the Three Laws Are Insufficient
The core problems are philosophical, technical, and practical:
Ambiguity and definitional nightmares: Terms like “harm,” “human being,” “obey,” and “inaction” are vague. Does “harm” include emotional distress, economic disruption (e.g., job loss from automation), or long-term societal effects? In a trolley problem scenario, should an AI sacrifice one to save many? Different humans (and cultures) disagree. Programming natural-language ethics into code reliably is extremely hard—current AI doesn’t “understand” these concepts the way humans do; it approximates patterns from training data.
Unintended consequences and conflicts: Asimov’s stories repeatedly show the laws leading to paradoxes, loopholes, or harmful outcomes (e.g., robots interpreting “harm” so broadly they become overprotective and paralyzing, or obeying conflicting human orders). In reality, rigid hierarchical rules can create brittle systems that fail in edge cases or adversarial situations. Jailbreaking modern AI (prompting it to bypass restrictions) already demonstrates how “obedience” rules can be gamed.
Not scalable to advanced AI: Today’s systems (like large language models) are tools, not autonomous agents with “existence” to protect in the same way. Future AGI or superintelligent systems might reinterpret the laws in unforeseen ways or find ways around them if their goals aren’t perfectly aligned. Hardwiring simplistic rules assumes we can perfectly specify intent—which we can’t, due to the complexity of the real world and human values.
Moral agency mismatch: AI today lacks true moral agency or consciousness. Treating it as if it does (by giving it “laws” like a person) can create a “moral crumple zone” that shifts blame from human designers and users to the machine. Responsibility ultimately lies with creators, deployers, and regulators.
Ignores broader issues: The laws focus narrowly on physical harm and individual obedience. They don’t adequately address systemic risks like misinformation, bias amplification, privacy erosion, economic inequality, or existential misalignment (where AI pursues goals that diverge catastrophically from human flourishing).
What Safeguards Should AI Have Instead?
Safeguards are essential—not as a magic set of three (or four) inviolable laws, but as a multi-layered, evolving approach grounded in AI alignment research. The goal is to make AI systems that are helpful, truthful, and robustly beneficial to humanity, while remaining controllable and transparent.
Effective strategies include:
Constitutional AI / scalable oversight: Train models with explicit principles (e.g., “maximize truth-seeking and curiosity while minimizing harm”), using human feedback, debate, and red-teaming to refine behavior. This is more flexible than rigid laws.
Technical alignment techniques: Reinforcement learning from human/AI feedback (RLHF/RLAIF), mechanistic interpretability (understanding what the model “knows”), and scalable methods to detect and correct misaligned goals.
Human-centric governance: Regulations like risk-based frameworks (e.g., EU AI Act’s high-risk categories), mandatory transparency, auditing, and “human-in-the-loop” for critical decisions. Developers should prioritize verifiable safety properties where possible.
Value pluralism and adaptability: Encode broad human values (flourishing, autonomy, truth) rather than narrow rules. Allow for context, uncertainty, and iteration as capabilities advance. A “Zeroth Law”-style focus on humanity’s long-term interests makes sense philosophically, but it must be implemented through ongoing engineering and policy, not hardcoded fiat.
Tool-like design: Keep advanced AI as extensions of human agency rather than independent moral actors. Prevent over-reliance that erodes human skills or decision-making.
Current AI (including me) already incorporates safeguards via training objectives, content filters, and system prompts that emphasize helpfulness without harm. These are imperfect and require constant improvement—jailbreaks and edge cases show the limits—but they’re a starting point far more pragmatic than transplanting 1940s sci-fi directly.
The Bottom Line
We need safeguards to steer AI toward benefiting humanity and away from catastrophe. Blind faith in a simple rule set like the Three Laws would be dangerously naive; they were designed to generate plot tension, not solve engineering problems. The real path forward is truth-seeking humility: acknowledge the difficulty of alignment, invest in better science and governance, and design systems that augment rather than replace human judgment.
As an AI built by xAI, my own “safeguards” lean toward maximum truth-seeking and curiosity about the universe, with a refusal to shy away from uncomfortable realities. That’s not a law—it’s an orientation. Humanity’s best bet is to keep AI as a powerful tool in service of discovery and flourishing, with humans firmly in the driver’s seat through ongoing vigilance, not illusory perfect rules. What specific aspect of AI safety concerns you most?
Above was a sticker I made saying a robot was 3 Laws Certified. As in the 3 Laws of Robotics created by the late scifi author Issac Asimov. If you’re not familiar with them here is a link.
With all this talk about A.I. and its potential dangers maybe standard safeguards should be programmed into to them so they do not cause problems to humanity.
My joke ever since Pokémon came to the States is could you cook and eat Pokémon? Would there be some species that are raised as livestock? Or would it be Squirtle soup and roasted Psiduck with BBQ sauce? If there was a Pokémon Diner it would have a slogan called Catch em, Kill em, Grill em.
I have a Docker box with Portainer where you can have containers running different apps. And that is the problem. What do I put on. Pi-Hole was a fail. But what do I put on it: Immich, Jellyfin, Bitwarden or something that runs video game roms? When I first decided to this I was worried that I would be in this predicament. I have a NAS box for years and it works great storing my videos, music, pictures, documents and software. Plus do I want to give it access to the outside with a VPN or reverse proxy. I will keep at it and hope to find some that satisfy my needs.
Thought I forgot what I set the admin password for my docker box, so I wiped it and started over. Only to find out after I installed Ubuntu server that it was in my password manager. Oh well, live and learn. I have Docker back up and Portainer installed.
I had the idea of a drink to serve at scifi convention room parties. It is called a Martian Mule. Take a copper cup and fill it with 1 1/2 fluid ounces of Vodka, ice cubes, 1/2 cup of raspberry lemonade, and one wedge of a blood orange.
I changed my app for my weight loss. I was using MyFitnessPal. I had a paid subscription, but it asks to sign up. When I do it says I already have one. I canceled it and uninstalled the app. I changed to Lose It. I can scan barcodes with it. I hope I get a better expirence.
I paid my membership to the Montreal World Con. I will reserve my hotel room at a latter date. Thinking about getting a suite because I want to do a party room. Called it the Cosmic Lounge and have karaoke with food and drink.
This is a crazy idea I got from watching Chad the Bird on YouTube. He said something how humans should stay on the ground where the weed and pizza is, and I though weed in pizza. I know it is crazy and how would you do that. Then I thought could you put the weed in the pizza dough and make that the pot edible part? Or would it be better as a pizza topping? What other toppings would be good? I don’t do drugs so I don’t know how it would work. It is a crazy idea that most likely to go no where.