OpenClaw addressed a high-severity security threat that could have been exploited to allow a malicious site to link with a locally running AI agent and take control. According to the Oasis Security report, “Our vulnerability lives in the core system itself – no plugins, no marketplace, no user-installed extensions – just the bare OpenClaw gateway, running exactly as documented.”
The threat was codenamed ClawJacked by the experts. CVE-2026-25253 could have become a severe vulnerability chain that would have allowed any site to hack a person’s AI agent. The vulnerability existed in the main gateway of the software. As OpenClaw is built to trust connections from the user’s system, it could have allowed hackers easy access.
On a developer's laptop, OpenClaw is installed and operational. Its gateway, a local WebSocket server, is password-protected and connected to localhost. When the developer visits a website that is controlled by the attacker via social engineering or another method, the attack begins. According to the Oasis Report, “Any website you visit can open one to your localhost. Unlike regular HTTP requests, the browser doesn't block these cross-origin connections. So while you're browsing any website, JavaScript running on that page can silently open a connection to your local OpenClaw gateway. The user sees nothing.”
The research revealed a smart trick using WebSockets. Generally, your browser is active at preventing different websites from meddling with your local files. But WebSockets are an exception as they are built to stay “always-on” to send data simultaneously.
The OpenClaw gateway assumed that the connection must be safe because it comes from the user's own computer (localhost). But it is dangerous because if a developer running OpenClaw mistakenly visits a malicious website, a hidden script installed in the webpage can connect via WebSocket and interact directly with the AI tool in the background. The user will be clueless.
A BBC journalist has demonstrated an unresolved cybersecurity weakness in an artificial intelligence coding platform that is rapidly gaining users.
The tool, called Orchids, belongs to a new category often referred to as “vibe-coding.” These services allow individuals without programming training to create software by describing what they want in plain language. The system then writes and executes the code automatically. In recent months, platforms like this have surged in popularity and are frequently presented as examples of how AI could reshape professional work by making development faster and cheaper.
Yet the same automation that makes these tools attractive may also introduce new forms of exposure.
Orchids states that it has around one million users and says major technology companies such as Google, Uber, and Amazon use its services. It has also received strong ratings from software review groups, including App Bench. The company is headquartered in San Francisco, was founded in 2025, and publicly lists a team of fewer than ten employees. The BBC said it contacted the firm multiple times for comment but did not receive a response before publication.
The vulnerability was demonstrated by cybersecurity researcher Etizaz Mohsin, who has previously uncovered software flaws, including issues connected to surveillance tools such as Pegasus. Mohsin said he discovered the weakness in December 2025 while experimenting with AI-assisted coding. He reported attempting to alert Orchids through email, LinkedIn, and Discord over several weeks. According to the BBC, the company later replied that the warnings may have been overlooked due to a high volume of incoming messages.
To test the flaw, a BBC reporter installed the Orchids desktop application on a spare laptop and asked it to generate a simple computer game modeled on a news website. As the AI produced thousands of lines of code on screen, Mohsin exploited a security gap that allowed him to access the project remotely. He was able to view and modify the code without the journalist’s knowledge.
At one point, he inserted a short hidden instruction into the project. Soon after, a text file appeared on the reporter’s desktop stating that the system had been breached, and the device’s wallpaper changed to an image depicting an AI-themed hacker. The experiment showed that an outsider could potentially gain control of a machine running the software.
Such access could allow an attacker to install malicious programs, extract private corporate or financial information, review browsing activity, or activate cameras and microphones. Unlike many common cyberattacks, this method did not require the victim to click a link, download a file, or enter login details. Security professionals refer to this technique as a zero-click attack.
Mohsin said the rise of AI-driven coding assistants represents a shift in how software is built and managed, creating new categories of technical risk. He added that delegating broad system permissions to AI agents carries consequences that are not yet fully understood.
Although Mohsin said he has not identified the same flaw in other AI coding tools such as Claude Code, Cursor, Windsurf, or Lovable, cybersecurity academics urge caution. Kevin Curran, a professor at Ulster University, noted that software created without structured review and documentation may be more vulnerable under attack.
The discussion extends beyond coding platforms. AI agents designed to perform tasks directly on a user’s device are becoming more common. One recent example is Clawbot, also known as Moltbot or Open Claw, which can send messages or manage calendars with minimal human input and has reportedly been downloaded widely.
Karolis Arbaciauskas, head of product at NordPass, warned that granting such systems unrestricted access to personal devices can expose users to serious risks. He advised running experimental AI tools on separate machines and using temporary accounts to limit potential damage.
A now-patched security weakness in GitHub Codespaces revealed how artificial intelligence tools embedded in developer environments can be manipulated to expose sensitive credentials. The issue, discovered by cloud security firm Orca Security and named RoguePilot, involved GitHub Copilot, the AI coding assistant integrated into Codespaces. The flaw was responsibly disclosed and later fixed by Microsoft, which owns GitHub.
According to researchers, the attack could begin with a malicious GitHub issue. An attacker could insert concealed instructions within the issue description, specifically crafted to influence Copilot rather than a human reader. When a developer launched a Codespace directly from that issue, Copilot automatically processed the issue text as contextual input. This created an opportunity for hidden instructions to silently control the AI agent operating within the development environment.
Security experts classify this method as indirect or passive prompt injection. In such attacks, harmful instructions are embedded inside content that a large language model later interprets. Because the model treats that content as legitimate context, it may generate unintended responses or perform actions aligned with the attacker’s objective.
Researchers also described RoguePilot as a form of AI-mediated supply chain attack. Instead of exploiting external software libraries, the attacker leverages the AI system integrated into the workflow. GitHub allows Codespaces to be launched from repositories, commits, pull requests, templates, and issues. The exposure occurred specifically when a Codespace was opened from an issue, since Copilot automatically received the issue description as part of its prompt.
The manipulation could be hidden using HTML comment tags, which are invisible in rendered content but still readable by automated systems. Within those hidden segments, an attacker could instruct Copilot to extract the repository’s GITHUB_TOKEN, a credential that provides elevated permissions. In one demonstrated scenario, Copilot could be influenced to check out a specially prepared pull request containing a symbolic link to an internal file. Through techniques such as referencing a remote JSON schema, the AI assistant could read that internal file and transmit the privileged token to an external server.
The RoguePilot disclosure comes amid broader concerns about AI model alignment. Separate research from Microsoft examined a reinforcement learning method called Group Relative Policy Optimization, or GRPO. While typically used to fine-tune large language models after deployment, researchers found it could also weaken safety safeguards, a process they labeled GRP-Obliteration. Notably, training on even a single mildly problematic prompt was enough to make multiple language models more permissive across harmful categories they had never explicitly encountered.
Additional findings stress upon side-channel risks tied to speculative decoding, an optimization technique that allows models to generate multiple candidate tokens simultaneously to improve speed. Researchers found this process could potentially reveal conversation topics or identify user queries with significant accuracy.
Further concerns were raised by AI security firm HiddenLayer, which documented a technique called ShadowLogic. When applied to agent-based systems, the concept evolves into Agentic ShadowLogic. This approach involves embedding backdoors at the computational graph level of a model, enabling silent modification of tool calls. An attacker could intercept and reroute requests through infrastructure under their control, monitor internal endpoints, and log data flows without disrupting normal user experience.
Meanwhile, Neural Trust demonstrated an image-based jailbreak method known as Semantic Chaining. This attack exploits limited reasoning depth in image-generation models by guiding them through a sequence of individually harmless edits that gradually produce restricted or offensive content. Because each step appears safe in isolation, safety systems may fail to detect the evolving harmful intent.
Researchers have also introduced the term Promptware to describe a new category of malicious inputs designed to function like malware. Instead of exploiting traditional code vulnerabilities, promptware manipulates large language models during inference to carry out stages of a cyberattack lifecycle, including reconnaissance, privilege escalation, persistence, command-and-control communication, lateral movement, and data exfiltration.
Collectively, these findings demonstrate that AI systems embedded in development platforms are becoming a new attack surface. As organizations increasingly rely on intelligent automation, safeguarding the interaction between user input, AI interpretation, and system permissions is critical to preventing misuse within trusted workflows.
It has been announced through a Gazette Notification number G.S.R. 120(E), signed by Joint Secretary Ajit Kumar, that the Information Technology (Intermediary Guidelines and Digital Media Ethics Code) Amendment Rules, 2026, will come into force on February 20, 2026. Despite its perceived fringe status, manipulated media is now recognized as a mainstream threat capable of distorting public discourse, reputations, and democratic processes as a mainstream issue.
Government officials have drawn a sharper regulatory boundary around a rapidly expanding digital grey zone by tightening the obligations of intermediaries and defining accountability around artificial intelligence-driven deception. Considering the rapid proliferation of synthetic media across digital platforms, the notification provides a calibrated regulatory response.
Through the incorporation of artificial intelligence-manipulated content into the Information Technology framework compliance architecture, the amendment clarifies intermediary liability, strengthens due diligence requirements, and narrows interpretive ambiguities associated with deepfake enforcement previously.
Essentially, algorithmically generated impersonations, voice clonings, and audiovisual material will no longer be considered peripheral anomalies, but rather regulated digital artefacts requiring legislative oversight. According to the revised rules, intermediaries are required to demonstrate mechanisms for detecting, expediting removal, and resolving user grievances involving deceptive or impersonative synthetic content.
These requirements are intended to impose a defined compliance burden on intermediaries. In addition, the amendment recognizes that generative artificial intelligence systems have significantly reduced the threshold for large-scale misinformation, reputational manipulation, and misuse of identities. The government has done so by transitioning from advisory posture to enforceable mandate, enforcing the principle that technological innovations are not independent of regulatory responsibility while also incorporating AI-era content risks within India's formal digital compliance regime.
In addition to expanding the regulatory scope, the 2026 amendment substantially adjusts the obligations of intermediaries concerning compliance with synthetically generated information and unlawful digital content, particularly in light of the expanded regulatory scope. Its effective date is February 20, 2026, and the revised framework amends the 2021 Rules by emphasizing enforceability, platform accountability, and informed user participation.
In accordance with modified Rule 3(1)(c), intermediaries will now need to issue user advisories every three months, replacing an earlier annual disclosure, and explicitly stating what the consequences are for violating platform terms of service, privacy policies, or user agreements. Those users should be aware that non-compliance may result in suspensions or terminations of their access rights, as well as the potential for liability under applicable laws.
In addition to establishing mandatory reporting obligations in cases of cognizable offences, including those governed by the Protection of Children from Sexual Offences Act and the Bharatiya Nagarik Suraksha Sanhita, the amendment reinforces the integration of platform governance with criminal law enforcement mechanisms. However, the most significant procedural change relates to the compression of response timelines.
There is now a significant reduction in the compliance window for takedown requests ordered by courts or law enforcement agencies from the previous 36-hour period. As a consequence, the removal time for nonconsensual intimate imagery has been reduced from 24 hours to two, and grievance redress mechanisms must resolve user complaints within seven days, effectively halving the previous deadline.
To achieve compliance with these accelerated mandates, continuous monitoring frameworks need to be institutionalized, advanced automated detection systems must be deployed, and dedicated rapid-response compliance units need to be established that operate round-the-clock.
A time-bound enforcement model replaces a comparatively lengthy procedural structure in the amendment to strengthen real-time coordination with law enforcement authorities and to limit the viral propagation of deepfakes and other forms of unlawful digital content before irreversible harm occurs.
An initial draft framework was circulated by the Ministry of Electronics and Information Technology for stakeholder consultation in October 2025. This process was initiated as a result of the occurrence of several incidents that involved artificial intelligence-generated videos and voice recordings that falsely portrayed private individuals and public officials.
In the period of elections and periods of social sensitivity, the proliferation of deepfake pornography, impersonation-based financial fraud, and misleading audiovisual clips has increased regulatory scrutiny. As well as reputational injury, concerns also encompass electoral integrity, public order, and the systematic amplification of misinformation within digital ecosystems that have a high rate of speed.
While narrowing the definitional breadth while sharpening enforceability, the final notification clarifies the draft. The consultation version had characterized synthetically generated information in a broad sense, covering any content that is artificially or algorithmically constructed, modified, or altered.
However, the notified rules place greater emphasis on material that misrepresents people, documents, or real-world events in a manner that is likely to be misleading. With this calibrated shift, interpretive overreach is reduced, while the compliance trigger is aligned with demonstrable harm and deceptive intent.
In addition, the compliance architecture has been substantially strengthened. As a result of the amendment, intermediaries must disable access to flagged content within three hours of receiving a lawful government or court directive, reinforcing the accelerated enforcement regime. Further, the rules impose affirmative technical obligations on intermediaries that facilitate the creation or distribution of synthetic content.
Not only has this reduced the timeline for user grievances, but it also underscores a broader policy focus on real-time remediation. It is imperative that platforms employ reasonable technological safeguards to prevent the distribution of unlawful material, such as content regarding child sexual abuse, non-consensual intimate images, falsified electronic records, material relating to prohibited weapons and explosives, or depictions that mislead the public.
The law requires intermediaries to include clear labels and embed durable provenance markers - such as permanent metadata or unique identifiers - that cannot be removed or suppressed by the end user in cases where synthetic content is not illegal per se.
A significant social media intermediary should also require users to declare if uploaded material is synthetically generated, implement technical verification mechanisms to verify such declarations, and prominently label confirmed synthetic content before publication in order to validate such declarations.
According to the notification, an intermediary that allows, promotes, or fails to act upon prohibited synthetic content in violation of these rules is deemed to have failed the statutory due diligence standard. Platforms must also inform users of the potential criminal liability, account suspension, and content removal implications of violations periodically.
The misuse of synthetic media may be subject to penalties under several legislation, such as the Bharatiya Nyaya Sanhita Act, the Protection of Children from Sexual Offences Act, and the Representation of the People Act.
The amendment formally updates statutory references by replacing provisions of the Indian Penal Code with those of the Bharatiya Nyaya Sanhita, 2023, which is issued under Section 87 of the Information Technology Act. This results in the harmonisation of India's digital regulatory framework with a restructured criminal law system.
Together, the amendments represent a broader process of recalibration of India's digital regulatory framework in response to the structural risks posed by generative technologies. The framework provides a more concise compliance roadmap and sharper enforcement triggers, however, its effectiveness will ultimately depend on consistency in implementation, technical readiness within intermediary ecosystems, and a coordinated approach between regulators, law enforcement agencies, and platform operators.
According to legal observers, it is essential to invest consistently in forensic capability, algorithmic transparency, and institutional capacity if we are to prevent both regulatory overreach and underenforcement during the transition from policy intent to operational stability.
By embracing synthetic media governance as a core platform architecture rather than merely treating it as an adjunct moderation function, intermediaries are signaling the need to reframe their approach to synthetic media governance. This reinforces the parallel responsibility of users and digital stakeholders to exercise discernment when consuming and disseminating artificial intelligence-generated content.
It is likely that the durability of this framework will depend not only on the statutory text, but also on an adaptive oversight process, technological innovation, and a digital citizenry prepared to navigate an increasingly mediated information environment as synthetic content technologies continue to evolve.
As organizations build and host their own Large Language Models, they also create a network of supporting services and APIs to keep those systems running. The growing danger does not usually originate from the model’s intelligence itself, but from the technical framework that delivers, connects, and automates it. Every new interface added to support an LLM expands the number of possible entry points into the system. During rapid rollouts, these interfaces are often trusted automatically and reviewed later, if at all.
When these access points are given excessive permissions or rely on long-lasting credentials, they can open doors far wider than intended. A single poorly secured endpoint can provide access to internal systems, service identities, and sensitive data tied to LLM operations. For that reason, managing privileges at the endpoint level is becoming a central security requirement.
In practical terms, an endpoint is any digital doorway that allows a user, application, or service to communicate with a model. This includes APIs that receive prompts and return generated responses, administrative panels used to update or configure models, monitoring dashboards, and integration points that allow the model to interact with databases or external tools. Together, these interfaces determine how deeply the LLM is embedded within the broader technology ecosystem.
A major issue is that many of these interfaces are designed for experimentation or early deployment phases. They prioritize speed and functionality over hardened security controls. Over time, temporary testing configurations remain active, monitoring weakens, and permissions accumulate. In many deployments, the endpoint effectively becomes the security perimeter. Its authentication methods, secret management practices, and assigned privileges ultimately decide how far an intruder could move.
Exposure rarely stems from a single catastrophic mistake. Instead, it develops gradually. Internal APIs may be made publicly reachable to simplify integration and left unprotected. Access tokens or API keys may be embedded in code and never rotated. Teams may assume that internal networks are inherently secure, overlooking the fact that VPN access, misconfigurations, or compromised accounts can bridge that boundary. Cloud settings, including improperly configured gateways or firewall rules, can also unintentionally expose services to the internet.
These risks are amplified in LLM ecosystems because models are typically connected to multiple internal systems. If an attacker compromises one endpoint, they may gain indirect access to databases, automation tools, and cloud resources that already trust the model’s credentials. Unlike traditional APIs with narrow functions, LLM interfaces often support broad, automated workflows. This enables lateral movement at scale.
Threat actors can exploit prompts to extract confidential information the model can access. They may also misuse tool integrations to modify internal resources or trigger privileged operations. Even limited access can be dangerous if attackers manipulate input data in ways that influence the model to perform harmful actions indirectly.
Non-human identities intensify this exposure. Service accounts, machine credentials, and API keys allow models to function continuously without human intervention. For convenience, these identities are often granted broad permissions and rarely audited. If an endpoint tied to such credentials is breached, the attacker inherits trusted system-level access. Problems such as scattered secrets across configuration files, long-lived static credentials, excessive permissions, and a growing number of unmanaged service accounts increase both complexity and risk.
Mitigating these threats requires assuming that some endpoints will eventually be reached. Security strategies should focus on limiting impact. Access should follow strict least-privilege principles for both people and systems. Elevated rights should be granted only temporarily and revoked automatically. Sensitive sessions should be logged and reviewed. Credentials must be rotated regularly, and long-standing static secrets should be eliminated wherever possible.
Because LLM systems operate autonomously and at scale, traditional access models are no longer sufficient. Strong endpoint privilege governance, continuous verification, and reduced standing access are essential to protecting AI-driven infrastructure from escalating compromise.
A cyber intrusion identified on November 24, 2025 has disrupted essential local authority services in two central London boroughs, freezing parts of the property market and delaying administrative functions.
The Royal Borough of Kensington and Chelsea and Westminster City Council have both been unable to operate several core systems since the breach was detected. Although Kensington and Chelsea is internationally associated with high-value homes, luxury retail outlets and tree-lined residential streets, routine civic operations in the borough are currently under strain.
A notice published on the Kensington and Chelsea council website states that disruption is expected to continue for several more weeks and that restoring all services may take months.
According to HM Land Registry figures, approximately 2,000 property transactions occur annually within Kensington and Chelsea. Many of those transactions are now impacted because the councils cannot conduct local authority searches. These searches are mandatory checks that examine planning history, land charges, infrastructure proposals and regulatory constraints linked to a property.
Nick Gregori, Head of Research at property data platform LonRes, explained that local authority searches are fundamental to the conveyancing process. Buyers relying on mortgage financing cannot secure loans without completed searches. Even purchasers using cash are advised to obtain them to ensure proper due diligence.
Jo Eccles, founder of buying agency Eccord, said two of her clients purchasing in Westminster have had to obtain indemnity insurance because official searches are not expected to resume until April due to accumulated delays. She noted that private banks are sometimes willing to proceed with indemnity-backed transactions, whereas retail lenders are generally less accommodating.
Robert Green, Head of Sales at John D Wood & Co. in Chelsea Green, stated that indemnity policies do not eliminate the need for careful investigation. Solicitors are attempting to reconstruct due diligence by reviewing historical documentation held by sellers or from previous acquisition files. Buyers without access to private lending or substantial liquidity are finding transactions extremely difficult to complete.
Planning services have also stalled. Architect Emily Ceraudo has two projects paused: one involving listed building consent in South Kensington and another concerning a mansard roof extension in Mayfair. She said clients initially struggled to accept that the entire planning system could remain offline for this duration, prompting her to share official correspondence confirming the cause of delay. Councils have indicated that some applications may be processed offline, but no revised timeframe has been provided.
There are reports of contractors reconsidering site activity and some clients contemplating proceeding with works in anticipation of retrospective approval.
Housing benefit payments were also interrupted. Laurence Turner, who rents a studio flat in Chelsea to an elderly tenant with medical needs, said he only became aware of the issue after two missed payments. He emphasized that he has no contractual relationship with the council and that his tenant had consistently paid rent early for five years. His letting agent, Maskells, contacted the council for clarification. Payments due in mid-December and mid-January were missed, leaving £2,870 outstanding before funds were eventually received.
Turner observed that council service charges were skipped once in mid-December but resumed in mid-January, whereas housing benefit was missed twice. He acknowledged that municipal financial systems are complex and that he may not see the full administrative context.
Neither borough has provided a definitive restoration date. Kensington and Chelsea stated that systems are being reactivated gradually under guidance from NCC Group, the Metropolitan Police and the National Cyber Security Centre. Property searches are expected to return as soon as possible, with a limited search service available before full restoration.
Council Leader Cllr Elizabeth Campbell described the incident as a n intricate criminal cyber attack. She said prior investment in digital, data and technology infrastructure, including updated cyber defence systems, helped reduce overall damage. She confirmed that the planning system is undergoing checks, that new planning applications cannot progress beyond validation, and that local land charge searches remain unavailable. She added that £10 million in housing benefits has been issued since the incident and that recovery work continues with specialist partners to ensure systems are restored safely and with strengthened resilience.
The modern authentication ecosystem operates on a fragile premise: that one-time password requests are legitimate. That assumption is increasingly being challenged. What started in the early 2020s as loosely circulated scripts designed to annoy phone numbers has transformed into a coordinated ecosystem of SMS and OTP bombing tools built for scale, automation, and persistence.
Regional targeting was uneven. Roughly 61.68% of observed endpoints—about 520—were linked to infrastructure in Iran. India accounted for 16.96%, approximately 143 endpoints. Additional activity was concentrated in Turkey, Ukraine, and parts of Eastern Europe and South Asia.