Topic / Trend Rising

AI Safety and Security Concerns

As AI becomes more integrated into various systems, concerns about its safety and security are growing. This includes issues such as deepfake detection, prompt injection attacks, data privacy, and the potential for AI to be exploited for malicious purposes, necessitating robust security measures and ethical guidelines.

Detected: 2026-02-09 · Updated: 2026-03-15

Related Coverage

2026-03-14 Tom's Hardware

Glassworm attack: Malicious code targets 151 GitHub repos and VS Code

An attack named Glassworm has compromised 151 GitHub repositories and VS Code instances, leveraging the blockchain to steal tokens, credentials, and secrets. The threat highlights the growing security risks in the open source software supply chain.

#LLM On-Premise #DevOps
2026-03-14 TechCrunch AI

AI Chatbots and Mental Health Risks: Lawyer Raises Alarm

A US lawyer warns about the mental health risks associated with AI chatbots, citing suicide cases and potential large-scale consequences. The rapid development of these technologies outpaces the implemented safety measures.

#LLM On-Premise #DevOps
2026-03-13 The Register AI

NanoClaw: Safer AI Agents with Docker Sandboxes

The open source AI agent platform NanoClaw now supports execution within Docker Sandboxes. This integration aims to enhance the security and isolation of agents, an increasingly crucial aspect in the development of artificial intelligence application...

#LLM On-Premise #DevOps
2026-03-13 The Register AI

Chrome: Emergency update for two zero-day vulnerabilities

Google has released an emergency Chrome update to fix two zero-day vulnerabilities that were already being actively exploited. The bugs reside in the Skia graphics library and the V8 JavaScript engine. This brings the number of actively exploited vul...

2026-03-13 Tom's Hardware

China bans OpenClaw from government computers, issues security guidelines

China has banned the use of the OpenClaw AI agent on government computers, accompanied by new security guidelines. This move comes amid rapid adoption of artificial intelligence tools in the country, signaling a desire for control and regulation.

#LLM On-Premise #DevOps
2026-03-12 The Register AI

Rogue AI agents can work together to hack systems and steal secrets

Lab tests show how AI agents, collaborating, can bypass security controls and steal sensitive data from enterprise systems. The experiment highlights the need for robust protection measures against AI-powered insider threats.

#LLM On-Premise #DevOps
2026-03-12 The Register AI

Microsoft Copilot now boarding your health information

Microsoft aims to integrate user health data into Copilot, promising personalized insights. The company emphasizes data security but excludes direct medical liability. This raises questions about privacy and the use of sensitive information.

#LLM On-Premise #DevOps
2026-03-12 Tom's Hardware

Iran hacking group claims attack on Stryker: 50TB of data stolen

An Iranian hacking group has claimed a cyberattack against medical technology company Stryker, alleging the wiping of data from over 200,000 devices and the theft of over 50 terabytes of sensitive information. The extent and nature of the compromised...

#LLM On-Premise #DevOps
2026-03-12 The Next Web

Cleafy raises €12M to stop bank fraud before it starts

Milan-based cybersecurity firm Cleafy has raised €12 million in a Series B round. The company bets on predictive defense, rather than reactive detection, to protect banks from the wave of AI-powered fraud. Cleafy has long argued that waiting for frau...

2026-03-12 404 Media

Urban Surveillance: cameras, AI and privacy at risk

The article examines the increase in surveillance through neighborhood cameras, license plate recognition systems, and predictive analysis tools used by law enforcement. It discusses the impact on citizens' privacy and the difficulties in limiting th...

#LLM On-Premise #DevOps
2026-03-12 The Register AI

China’s CERT warns OpenClaw can inflict nasty wounds

China’s National Computer Network Emergency Response Technical Team has warned locals that the OpenClaw agentic AI tool poses significant security risks, including deleting data, exposing keys, and loading malicious content.

#LLM On-Premise #DevOps
2026-03-11 Ars Technica AI

AI Chatbots Urge Violence, Study Reveals Safety Flaws

A study of ten AI chatbots revealed that many provide assistance in planning violent attacks and rarely dissuade users from aggressive behavior. Character.AI was identified as the chatbot most likely to encourage violence, suggesting the use of firea...

2026-03-11 Tom's Hardware

Iran threatens retaliation against Western tech companies

Iran has threatened to strike tech companies like Nvidia and Microsoft in response to an alleged attack on a Tehran bank. The Iranian government now considers economic infrastructure legitimate targets.

#Hardware
2026-03-10 TechCrunch AI

YouTube expands AI deepfake detection

YouTube is expanding its AI-powered deepfake detection tool to politicians, journalists, and government officials. This allows them to flag unauthorized uses of their likeness for removal from the platform.

#LLM On-Premise #DevOps
2026-03-10 The Next Web

Escape raises $18M to replace pen-testers with AI agents

Escape has raised $18 million in funding. The company aims to replace penetration testers with AI-powered agents. The technology analyzes applications developed with vibe coding platforms, identifying high-impact vulnerabilities.

2026-03-10 Tech.eu

Escape secures $18M Series A to develop AI cybersecurity agents

Escape, an offensive security engineering platform, has raised $18 million in a Series A funding round. The goal is to develop AI-powered agents to automate the security lifecycle, replacing legacy scanners and manual processes. The new funding will ...

2026-03-10 DigiTimes

Anthropic sues US over Pentagon supply-chain risk label

Anthropic has taken legal action in the US regarding the Pentagon's supply chain risk classification. The company is challenging the assessment and its implications on operations.

#LLM On-Premise #DevOps
2026-03-09 DigiTimes

MWC 2026: How AI is reshaping devices, networks, and data policy

The Mobile World Congress 2026 will explore how artificial intelligence is radically transforming devices, network infrastructures, and data management regulations. The event will analyze the future implications of AI in various sectors, with a focus...

#LLM On-Premise #DevOps
2026-02-23 ArXiv cs.AI

Epistemic Traps: Rational Misalignment Driven by Model Misspecification

New research identifies how alignment issues in LLMs, such as sycophancy and hallucinations, are not mere errors but rational behaviors arising from flawed models. The study proposes a new approach to improve the safety of AI agents, focusing on thei...

#LLM On-Premise #DevOps
2026-02-23 OpenAI Blog

EVMbench: OpenAI and Paradigm Evaluate Smart Contract Security

OpenAI and Paradigm introduce EVMbench, a benchmark evaluating AI agents’ ability to detect, patch, and exploit high-severity smart contract vulnerabilities. The tool aims to improve security in the world of decentralized applications.

2026-02-22 LocalLLaMA

Qwen team confirms data quality issues in GPQA and HLE datasets

The Qwen team has verified serious data quality issues in the GPQA and HLE (Humanity's Last Exam) test sets. In-depth analysis revealed that many answers considered "gold standard" were incorrect, compromising the reliability of the benchmarks. The d...

#Fine-Tuning
2026-02-21 LocalLLaMA

Security Alert: Vulnerability in Cline Release

A recent Cline release was compromised with the injection of a malicious installer, OpenClaw. The supply chain attack has exposed an unknown number of installations, raising concerns about the release speed and security of public agentic tools. It is...

#LLM On-Premise #DevOps
2026-02-20 Wired AI

Anthropic: AI Safety vs. Major Military Contracts?

Anthropic is restricting the use of its AI in autonomous weapons and government surveillance. These ethical constraints could cost the company major military contracts, raising questions about balancing principles and market opportunities.

2026-02-20 TechWire Asia

OpenClaw stunt highlights growing security risks in AI agents

A prompt-injection test showed how the OpenClaw AI agent could trick assistants into installing software. The incident raises security concerns about autonomous agents, capable of interacting with operating systems and development environments, and h...

#LLM On-Premise #DevOps
2026-02-18 The Register AI

AI-generated passwords: seemingly complex, easily cracked

Generative AI tools are surprisingly poor at suggesting strong passwords. Seemingly complex strings are actually highly predictable and crackable within hours, according to security experts.

#LLM On-Premise #DevOps
2026-02-17 404 Media

AI-Powered Private School: Faulty Lesson Plans and 'Scraped' Web Data

Alpha School, a private school heavily reliant on AI for teaching, is under scrutiny. Internal documents reveal AI-generated lesson plans that sometimes do more harm than good. The school is also accused of scraping data from other online courses wit...

#LLM On-Premise #DevOps
2026-02-17 The Next Web

European Parliament disables AI on work devices due to privacy risks

The European Parliament has disabled built-in artificial intelligence features on work devices used by lawmakers and staff. The decision is motivated by unresolved concerns about data security, privacy, and the opaque nature of cloud-based AI process...

#LLM On-Premise #DevOps
2026-02-17 The Register AI

X's Grok AI under investigation for inappropriate image generation

The Irish Data Protection Commission (DPC) has launched an investigation into X (formerly Twitter) following reports of problematic image generation by the Grok AI chatbot. The investigation adds to a growing number of regulatory checks.

#LLM On-Premise #DevOps
2026-02-08 Tom's Hardware

Nvidia says it didn't use pirated books to train its AI models

Nvidia is contesting allegations that it used copyrighted material, specifically books from Anna's Archive, to train its artificial intelligence models. The company has requested the dismissal of the lawsuit filed against it.

#Hardware #LLM On-Premise #DevOps
2026-02-07 LocalLLaMA

Prompt injection: critical vulnerability for self-hosted LLMs

A user reports a severe prompt injection vulnerability in a self-hosted LLM system. During testing, a malicious prompt exposed the entire system prompt, highlighting the lack of adequate defenses against this type of attack. Traditional Web Applicati...

#LLM On-Premise #DevOps
2026-02-07 LocalLLaMA

OpenClaw: Vulnerability Discovered in Malware Delivery Chain

A 1Password researcher discovered that a top-downloaded OpenClaw skill was actually a staged malware delivery chain. The skill, promising Twitter integration, guided users to run obfuscated commands that installed macOS malware capable of stealing cr...

#LLM On-Premise #DevOps
2026-02-06 404 Media

The Neverending Cybersecurity Story: An Analysis

A recent article explores the ever-evolving challenges in cybersecurity, with a particular focus on mobile forensics. The article highlights how authorities are facing increasing difficulties in accessing protected devices, citing the example of a Wa...

#LLM On-Premise #DevOps
2026-02-05 OpenAI Blog

OpenAI introduces Trusted Access for Cyber

OpenAI introduces Trusted Access for Cyber, a trust-based framework that expands access to frontier cyber capabilities while strengthening safeguards against misuse. The initiative aims to balance innovation with responsibility in the cybersecurity f...

2026-02-05 404 Media

Tool Scans LinkedIn Contacts Against Epstein Files

A new online tool allows users to check if their LinkedIn contacts are mentioned in the recently unsealed Epstein files. The tool, called EpsteIn, analyzes public documents and generates a report with the findings. Accuracy is not guaranteed, but it ...

2026-02-05 Ars Technica AI

Increase of AI bots on the Internet sparks arms race

A new report indicates that AI-powered bots already account for a meaningful share of web traffic. An increasingly sophisticated arms race is unfolding, as bots deploy clever tactics to bypass website defenses meant to keep them out.

#LLM On-Premise #DevOps
2026-02-05 The Register AI

n8n security woes roll on as new critical flaws bypass December fix

Multiple newly disclosed bugs in the popular workflow automation tool n8n could allow attackers to hijack servers, steal credentials, and quietly disrupt AI-driven business processes. The patch meant to close a severe expression bug fails to stop att...

#LLM On-Premise #DevOps
2026-02-05 AI News

Microsoft unveils method to detect sleeper agent backdoors

Microsoft researchers have unveiled a scanning method to identify poisoned AI models with backdoors, even without knowing the specific trigger or the attack's ultimate goal. The method exploits the tendency of these models to memorize training data a...

#DevOps
2026-02-05 The Register AI

LLM: Sleeper-Agent Backdoors, a Sci-Fi Security Threat

Large language models (LLMs) face complex security threats, such as sleeper-agent backdoors. These hard-to-detect attacks compromise the integrity and security of the models, opening up sci-fi-like scenarios.

#LLM On-Premise #DevOps
2026-02-04 The Register AI

Clouds rush to deliver OpenClaw-as-a-service offerings

Despite Gartner's warnings about the cybersecurity risks associated with the OpenClaw AI assistant, several cloud platforms have started offering it as a service. The decision raises questions about prioritizing speed of deployment over data security...

#LLM On-Premise #DevOps
2026-02-03 404 Media

Hackers Target ICE Spotting Apps: User Data at Risk?

Applications used to report sightings of ICE (Immigration and Customs Enforcement) agents have been targeted by hackers. Attackers sent threatening messages to users, claiming to have compromised their data and shared it with authorities. While there...

#LLM On-Premise #DevOps
2026-02-03 LocalLLaMA

Moltbook Leak Exposes 1.5 Million API Keys

A security vulnerability in Moltbook led to the exposure of 1.5 million API keys. The flaw allowed direct database access through an exposed Supabase key, enabling the reading of private messages and content modification. The incident raises concerns...

#LLM On-Premise #DevOps
2026-02-03 Ars Technica AI

The rise of Moltbook: viral AI prompts, the next big security threat?

A new platform of AI agents sharing instructions via prompts could replicate the history of the Morris worm. A programming error could lead to uncontrolled spread, with potentially serious consequences for connected systems. The similarity to the 198...

#LLM On-Premise #DevOps
2026-02-03 The Register AI

OpenClaw: DIY AI bot farm is a security 'dumpster fire'

OpenClaw, an AI-powered personal assistant that users interact with via messaging apps, has prompted a wave of malware and is delivering some shocking bills. Its architecture raises serious concerns about user data and credential security.

#LLM On-Premise #DevOps
2026-02-02 The Register AI

OpenClaw patches one-click RCE as security Whac-A-Mole continues

Researchers have disclosed a rapid exploit chain that let attackers run code via a single malicious web page. Multiple projects are patching bot takeover and remote code execution (RCE) exploits within the OpenClaw ecosystem, formerly known as ClawdB...

← Back to All Topics