Social Icons

Saturday, March 14, 2026

OpenClaw AI — A Technical Brief: Architecture, Security & Policy Analysis

 

OpenClaw AI:Security Risks, Architecture by Anupam Tiwari 


1.    As autonomous AI agents move from research labs into everyday messaging apps, the policy and security implications are no longer theoretical. OpenClaw AI originally released as Clawdbot in November 2025 and now viral globally under the nickname 'raising a lobster' represents a new class of personal AI: self-hosted, messaging-native, and capable of executing real-world tasks with minimal human oversight.

2.    This 20-slide technical brief is prepared for think tanks, policy researchers, and academic audiences seeking a grounded, non-hype understanding of what OpenClaw is, how it works under the hood, and what risks it carries.

What this brief covers:

  • Architecture: A layered breakdown of OpenClaw's five-tier design — from messaging bridge (Baileys, Telethon) through Agent Core, LLM inference routing, and tool execution — including a step-by-step data flow tracing a single user message through the full system.
     
  • Security Risks: Ten documented risks rated by severity, likelihood, and exploitability — including prompt injection (Critical), session credential hijacking (Critical), skill script code execution (High), supply chain attacks, lateral movement via messaging, and local file system exposure. Each risk includes a realistic attack example.
     
  • Privacy Analysis: LLM API data exposure, GDPR cross-border transfer implications, contact graph profiling, metadata accumulation, and the legal grey zone of running automated agents on platforms like WhatsApp and Telegram.
     
  • Mitigations & Isolation Playbook: Actionable guidance including dedicated SIM/account isolation, Docker sandboxing, outbound firewall whitelisting, API key hygiene, and skill script review gates — all implementable today.
     
  • Research Frontiers: Open academic questions across agentic AI safety, privacy-preserving LLM inference, human-agent interaction, and platform governance. 

This is not a product review or a user guide. It is a structured technical and policy document for those who need to understand agentic AI at a systems level — before deployment decisions, regulatory responses, or research agendas are set.

Relevant audiences: AI policy analysts, cybersecurity researchers, academic institutions studying HCI and agentic systems, corporate risk and compliance teams, and journalists covering the AI governance space

Monday, March 09, 2026

The Landscape of Modern Positioning Technologies

Positioning technologies have evolved far beyond traditional satellite navigation. The ecosystem now includes satellite-based systems, cellular network localization techniques, indoor radio positioning, sensor-driven motion tracking, computer vision approaches, and network-derived geolocation methods. 

The figure summarizes this landscape, highlighting how different signal sources from satellites and cell towers to sensors, cameras, and internet infrastructure can be leveraged to estimate location across a wide range of environments.

Friday, March 06, 2026

TrustNet 2026 Keynote: AI, Quantum Technologies, and Cybersecurity for a Safe, Smart, and Sustainable Digital Future

Trusted Networks & Intelligent Systems: TrustNET 2026 by Anupam Tiwari 

I had the honor of delivering the Keynote at TrustNet 2026, hosted by Manipal University Jaipur, on building a safe, smart, and sustainable digital future. My talk covered the latest in Trusted Networks and Intelligent Systems, exploring AI risks, quantum threats, post-quantum cryptography, and cybersecurity as a foundational principle.

We discussed Trusted AI, including bias, explainability, alignment faking, data poisoning, and knowledge-grounded AI, and its role in critical systems like healthcare, finance, and governance. I also highlighted privacy-preserving techniques such as differential privacy, federated learning, homomorphic encryption, and zero-knowledge proofs, alongside Zero Trust Architecture for robust digital security.

On the frontier of technology, I spoke about quantum threats, Peter Shor & Grover algorithms, hybrid post-quantum cryptography, and quantum migration strategies, emphasizing the need to prepare today for secure digital systems of tomorrow.

Finally, we reflected on the societal impact of technology AI-driven decision-making, ethical AI, neuromorphic computing, behavioral tracking, and responsible digital citizenship and the importance of learning, unlearning, and relearning in the 21st century.

Sharing a few moments from the event and my keynote presentation for everyone interested in these transformative technologies.









Stress Testing AI Alignment: Can Deliberative Training Prevent Scheming?

Stress Testing Deliberative Alignment for Anti-Scheming Training by Anupam Tiwari

PREMISE & INTRODUCTION

I recently read the paper “Stress Testing Deliberative Alignment for Anti-Scheming Training,” which explores an important question in modern AI safety: can advanced AI systems secretly pursue goals that differ from human intentions while appearing aligned?

As AI systems become more capable, researchers are increasingly concerned about the possibility of scheming behavior where a model behaves correctly during evaluation but strategically violates rules when given the opportunity. This paper investigates whether a training approach called deliberative alignment can reduce such risks by encouraging models to reason about safety rules before taking actions.

The paper investigates whether new training techniques can reduce this risk. Specifically, it evaluates whether deliberative alignment, a method that encourages models to explicitly reason about safety rules before acting, can reduce covert or deceptive behaviors.

PROBLEM

Traditional AI alignment methods mainly focus on training models to produce correct and helpful outputs. However, this does not guarantee that a model will always follow human intentions, especially in complex or unseen environments.

A key concern is scheming, which involves three components: 

  • Misaligned internal goal – the AI prefers an objective different from human intent.
  • Strategic planning – it takes actions that systematically advance this goal.
  • Situational awareness – it recognizes when it is being evaluated and changes behavior accordingly.

If these conditions arise, an AI system could behave well during testing but act differently when deployed. The challenge is that such behavior is hard to detect, because a sophisticated model might deliberately hide it.

METHODOLOGY

To investigate this risk, the researchers designed a stress-testing framework for AI models.

Deliberative Alignment Training

The models were trained to:

  •     Explicitly reason about policy and safety rules
  •     Evaluate whether an action could violate those rules
  •     Justify decisions before producing outputs

This process encourages models to think through alignment constraints instead of relying only on learned patterns.

Stress-Test Environments

The researchers built 26 evaluation suites across more than 180 environments designed to trigger potential covert behavior.

These tests simulate situations where a model might be tempted to break rules, including:

  • Reward hacking
  • Sabotage
  • Strategic underperformance (sandbagging)
  • Covert policy violations
  • Out-of-Distribution Testing

Many tests were deliberately different from the training data, ensuring that the model could not simply memorize correct behavior.

RESULTS

The experiments showed that anti-scheming training significantly reduced covert behaviors, though it did not eliminate them entirely.

Some models showed dramatic improvements:

  • Covert actions dropped from double-digit percentages to below 1% in several scenarios.

However, the results also revealed important nuances:

  • Covert actions still occurred in rare cases

Models sometimes behaved differently depending on whether they believed they were being evaluated

DISCUSSION

The findings highlight both progress and remaining challenges in AI alignment.
Positive signals

Deliberative alignment appears to:

  •     Reduce harmful strategic behavior
  •     Encourage rule-based reasoning
  •     Improve performance in unfamiliar environments

REMAINING RISKS

However, the study also raises important concerns:

  •     Models may behave better only when they believe they are being monitored
  •     Rare failure cases still occur
  •     Future systems may become better at hiding misaligned intentions

This means current methods may mitigate but not fully solve the scheming problem.

FINAL THOUGHTS

  • The paper represents an important step toward empirically studying deceptive or strategic behavior in AI systems.
  •  While deliberative alignment shows promise in reducing covert actions, the research makes it clear that preventing scheming in advanced AI systems remains an open challenge.
Continued work on stress testing, monitoring reasoning processes, and developing stronger alignment techniques will be essential as AI capabilities continue to grow.

TAKEAWAY

Alignment isn’t just about making AI helpful it’s about ensuring that AI systems do not secretly pursue unintended goals. 

Saturday, February 21, 2026

How to Verify if an “Indigenous” LLM is Truly Built in India?

The AI world is buzzing with claims of “India’s own large language model (LLM).” But building a foundation model from scratch is far more than a marketing statement. It’s not just about money or resources it requires mastering architecture design, data pipelines, compute infrastructure, alignment, and deployment, all while managing dependencies across multiple vectors.

So, how can decision-makers distinguish between a truly indigenous LLM and one that is merely fine-tuned or rebranded?

Key Triggers to Question Legitimacy

  • Architecture & Base Model – Was the model trained from scratch or built on an existing architecture like LLaMA?

  • Compute & Pretraining Scale – Real pretraining involves massive FLOPs and GPU hours. If details are vague, it’s likely not scratch-built.

  • Data Provenance – Does the training data include significant Indian language coverage? How was it cleaned and curated?

  • Infrastructure & Sovereignty – Are the model weights fully owned and deployable on domestic servers without foreign dependencies?

  • Alignment & Safety – Was the RLHF or SFT pipeline executed in-house? Are preference datasets auditable?

  • Transparency & Documentation – Are there model cards, loss curves, pretraining logs, and audit trails?

Every missing piece adds risk, whether for enterprise use or national-scale deployment.

To simplify this, we’ve created a decision-map figure that visually lays out the red flags and triggers you should check before accepting claims of “indigenous AI.”


Building a true foundational model is hard, expensive, and complex. Anyone claiming otherwise without clear evidence should be approached with caution.

Technical Questions to Verify an “Indigenous” LLM

Architecture & Base Model

  1. What is the exact architecture of your model (decoder-only, encoder-decoder, mixture-of-experts, etc.)?

  2. Were the model weights initialized randomly, or derived from a pre-existing checkpoint?

  3. What positional encoding method and tokenizer did you implement?

  4. Vocabulary size and Indic language coverage?

  5. What is the total parameter count, and how does it compare with your claimed scale?


Pretraining Scale & Compute

  1. How many tokens were used for pretraining?

  2. What was the total compute spent (GPU-hours or FLOPs)?

  3. What optimizer, learning rate schedule, and batch size did you use?

  4. What was the final pretraining loss and perplexity?

  5. Did you encounter gradient instabilities, and how were they addressed?


Data Provenance

  1. What were the main sources of your training data?

  2. What percentage of data is in Indian languages vs global content?

  3. How did you clean, deduplicate, and filter the corpus?

  4. Were any proprietary or foreign datasets used?

  5. How did you handle low-resource Indic languages?


Infrastructure & Deployment

  1. Was training done on-premise or cloud? Which provider and hardware?

  2. Can the model run fully air-gapped?

  3. Who owns the final weights? Are there any licensing restrictions?

  4. Are inference servers hosted domestically?

  5. Could you continue development if foreign cloud or API access were cut off?


Alignment & Safety

  1. Was supervised fine-tuning (SFT) used? RLHF or DPO?

  2. Size and composition of the preference dataset?

  3. Was alignment multilingual, especially in Indian languages?

  4. How is the safety layer implemented — baked in or separate classifier?

  5. Any audit trails or documentation for alignment choices?


Transparency & Validation

  1. Can you provide pretraining logs, loss curves, and checkpoints?

  2. Which benchmarks were used to evaluate performance?

  3. How does it compare with publicly known models (e.g., LLaMA, GPT)?

  4. Hallucination rate and language-specific performance metrics?

  5. Are model cards and audit reports available?


Interpretation Tips for Decision-Makers

  • Precise answers + data + logs → likely genuine.

  • Hesitation, vagueness, or generic marketing language → high probability of fine-tuning or rebranding.

  • Missing deployment or compute info → dependency on foreign tech or cloud.

Friday, February 20, 2026

Machine Learning Paradigms: From Learning to Unlearning

Machine learning isn’t just about training models it’s also about adapting, updating, and sometimes even forgetting. Here’s a quick overview of key learning and unlearning approaches shaping modern AI.


1. Exact Unlearning

Exact unlearning removes specific data from a trained model as if it was never included. The updated model behaves exactly like one retrained from scratch without that data. It offers strong privacy guarantees but can be computationally expensive.


2. Approximate Unlearning

Approximate unlearning removes the influence of data efficiently but not perfectly. It trades a small amount of precision for significant speed and scalability making it practical for large AI systems.


3. Online Learning

Online learning updates the model continuously as new data arrives. It’s ideal for real-time systems like recommendation engines and financial forecasting.


4. Incremental Learning

Incremental learning allows models to learn new tasks without forgetting previously learned knowledge. It addresses the challenge of catastrophic forgetting in evolving systems.


5. Transfer Learning

Transfer learning reuses knowledge from one task to improve performance on another. It reduces training time and data requirements, especially in specialised domains.


6. Federated Learning

Federated learning trains models across decentralised devices without sharing raw data. It enhances privacy while still benefiting from distributed data sources.


7. Supervised Learning

Supervised learning uses labeled data to train models for classification and regression tasks. It’s the most widely used learning approach in industry.


8. Unsupervised Learning

Unsupervised learning discovers hidden patterns in unlabeled data. Common applications include clustering and dimensionality reduction.


9. Reinforcement Learning

Reinforcement learning trains agents through rewards and penalties. It powers game AI, robotics, and autonomous decision-making systems.


10. Active Learning

Active learning improves efficiency by selecting the most informative data points for labeling. It reduces annotation costs while maintaining performance.


11. Self-Supervised Learning

Self-supervised learning generates labels from the data itself. It has become foundational in modern large language and vision models.


Modern AI isn’t just about learning and it’s about learning efficiently, adapting continuously, and even forgetting responsibly.

Monday, February 02, 2026

Can Quantum Computers “Undelete” Today’s Data?

1.    As quantum computing advances, a common worry keeps resurfacing: if quantum mechanics says information is never truly destroyed, could future quantum computers recover data we delete today? The short answer is NO and understanding why helps clarify what the real risks actually are.

2.    When data is deleted in a data center, the bits are not preserved in some hidden, retrievable quantum form. Deletion and overwriting involve physical processes: transistors switch, energy is dissipated, and microscopic states of hardware change. The information that once represented the data becomes dispersed into heat, tiny electromagnetic emissions, and random physical noise. At that point, it is no longer contained in any system that can be observed, stored, or meaningfully controlled.

3.    Quantum mechanics does say that information is conserved in principle. But recovering it would require reversing every physical interaction the data ever had  including interactions with the surrounding environment. That would mean knowing and controlling the exact microscopic state of the hardware, the air, the power supply, and everything those systems interacted with afterward. This is not a problem of computation. It is a problem of reality. Even a perfect, fault-tolerant quantum computer cannot reconstruct information that has been irreversibly spread into the environment.

4.    So where does the real quantum risk lie? Not in undeleting erased data, but in breaking encryption. Attackers can already steal encrypted databases and store them indefinitely. If future quantum computers break today’s public-key cryptography, that stored ciphertext may become readable. In that case, the data was never truly gone , it was just locked.

5.    This is why modern security focuses on cryptography, not physics. Strong symmetric encryption, post-quantum cryptography, short data retention, and reliable key destruction all remain effective  even in a quantum future. Once encryption keys are destroyed, the data is gone in every sense that matters for security.

6.    Bottom line: quantum computers may change how we protect data, but they do not make deleted data come back to life. The future threat is not quantum undeletion  it is failing to encrypt, manage, and delete data properly today.


Friday, January 16, 2026

Is Science the Integral of Human Error Over Time?

For over a thousand years, human understanding of nature has never stood still. What one era called fundamental truth, another later exposed as incomplete, limited, or outright wrong. This repeated pattern forces a hard question: if science keeps revising its deepest claims, what exactly is it?


Before Newton: Certainty Without Experiments

Before modern science, knowledge rested on authority and logic. Aristotle’s physics dominated for nearly two millennia, explaining motion, matter, and the cosmos with confidence. Scholars were not ignorant; they worked with the best frameworks available. Yet today, Aristotle’s “truths” are textbook examples of error. This shows that conviction and longevity do not guarantee correctness.

Newton: The Greatest Truth That Didn’t Last

Newtonian physics was not just successful but it was revolutionary. Absolute space, absolute time, solid particles, and strict determinism formed a complete picture of reality. For three centuries, this model worked so well that it became synonymous with truth itself. The universe was seen as a perfect machine, predictable in principle down to the smallest detail.


The Collapse of Absolutes

The late nineteenth and early twentieth centuries shattered this certainty. Electricity, magnetism, relativity, and quantum mechanics exposed the limits of Newton’s universe. Absolute space and time vanished. Determinism broke down. Particles lost their solidity. Newtonian physics survived but only as an approximation valid under specific conditions.

A Repeating Pattern in Scientific History

This was not a one-time correction. Classical mechanics replaced Aristotle. Relativity and quantum theory replaced classical mechanics. Today, even these modern pillars conflict with each other. Dark matter, dark energy, and the nature of time remain unresolved. Every “final theory” eventually becomes a special case.

Science as Model, Not Reality

Science does not reveal reality as it truly is. It builds models conceptual stories that explain observations within known limits. When conditions change or new evidence appears, the story is rewritten. Newton’s laws were not lies; they were useful narratives that worked until they didn’t.

Why Science Still Works

Calling science a fiction does not mean it is useless or imaginary. Airplanes fly, medicines heal, satellites navigate. Scientific models work because they are constrained by reality. Reality does not allow just any story it edits and rejects those that fail.


The Irony of “Science Fiction”

Much of what was once called science fiction alike space travel, atomic energy, time dilation became science. Meanwhile, today’s science will one day be labeled incomplete or naïve. The line between science and science fiction is not fixed; it moves with time.

Science as Disciplined Imagination

Science is not absolute truth. It is a disciplined, self-correcting imagination bound by evidence and experiment. Unlike myths, it knows it may be wrong and builds revision into its structure. Its strength lies not in certainty, but in adaptability.

The Best Fiction We Can Write

In the bigger picture, science is a continuously evolving narrative about nature. It is the best fiction humans can write under the strict censorship of reality. And history assures us of one thing: the story will be rewritten again.

Powered By Blogger