• Industry News
  • CXO Spotlight
  • AI
  • Enterprise Security

 Back to New Tab

New Report Says Workers and Execs Alike are Breaking Their Own Rules on AI Usage

New Tab News Team
September 18, 2025
Industry News

According to a recent report, over half of U.S. office workers are willing to violate company AI policies to ease their workload, posing security risks.

Credit: Outlever

More than half of U.S. office workers are willing to violate company AI policy to make their jobs easier, creating significant internal security risks, according to a new report from AI security firm CalypsoAI. The findings reveal a workforce that is now routinely circumventing the very guardrails put in place to manage the technology.

  • Trust in the machine: Workplace trust is fundamentally shifting away from people and toward AI, with nearly half of employees now trusting algorithms more than their own coworkers. The preference is so strong that over a third would rather report to an AI manager, and a third say they would quit if their employer banned AI altogether.

  • Do as I say, not as I do: The disregard for protocol extends all the way to the top. More than two-thirds of C-suite executives would bypass internal policies to use AI, yet many are operating in the dark. Nearly 40% of business leaders also admitted they don’t know what an AI agent is.

  • Fox guarding the henhouse: Perhaps most alarmingly, a similar dynamic is playing out among the people in charge of enforcement. Nearly 60% of security professionals say they trust AI more than their colleagues—a sentiment that helps explain why more than 40% of them knowingly use AI against company policy.

"We know inappropriate use of AI can be catastrophic for enterprises, and this isn't a future threat — it's already happening inside organizations today," said CalypsoAI CEO Donnchadh Casey in a release. The problem of "shadow AI" is particularly acute in highly regulated industries like finance and healthcare, echoing the early days of cloud adoption where unsanctioned tech use was rampant. But it's not all defiance; many workers are simply looking for guidance, with a majority indicating they would welcome more education on AI risks.

Related content

An Insider's Guide to Rewiring Orgs as Agents Move From Tools to Core Operators

Omer Grossman, former Chief Trust Officer and Head of the CYBR Unit at CyberArk, explains why nearly every enterprise claims to use AI but almost none have transformed the way their organizations actually operate.

Shadow AI and Departmental Silos Force Enterprises to Rethink Resilience

Nethusha Ravisuthan, Sales Support and Operations Manager at Microsoft, argues that Shadow AI, departmental silos, and ungoverned AI agents are compounding enterprise risk, and that operational trust and holistic system resilience must become foundational to AI deployment.

How Higher Education Puts Boundaries Around AI Agents With Sanctioned Access Models

Vijay Samtani, CISO at Cambridge University, discusses how blocking AI agents is a losing battle for security leaders. Their best course of action is to build clear rules and guidelines for AI access to control vulnerable surfaces.

You might also like

See all →

Apple Doubles Top Bug Bounty to $2M in Spyware Arms Race

Report says majority of employees embrace AI unsupervised, leaving companies vulnerable

New Report Says Workers and Execs Alike are Breaking Their Own Rules on AI Usage

Powered by Island.
ISLAND, All rights reserved ©