• About
  • Subscribe
  • Contact
Thursday, July 10, 2025
    Login
FutureCISO
  • People
  • Process
  • Technology
  • Resources
    • White Papers
    • PodChats
No Result
View All Result
  • People
  • Process
  • Technology
  • Resources
    • White Papers
    • PodChats
No Result
View All Result
FutureCISO
No Result
View All Result
Home Technology Data Protection

Pitting AI agents against each other to uncover security holes

Eileen Yu by Eileen Yu
July 10, 2025
Photo by Tima Miroshnichenko: https://www.pexels.com/photo/close-up-photo-of-a-wooden-chessboard-6599617/

Photo by Tima Miroshnichenko: https://www.pexels.com/photo/close-up-photo-of-a-wooden-chessboard-6599617/

Share on FacebookShare on Twitter

With the volume and sophistication of cyber threats increasing exponentially, fuelled by artificial intelligence (AI), one solution that has emerged involves letting non-human agents compete against each other for rewards.

Give one team of AI agents (the blue team) the role of defenders, while a second team (the red team) is instructed to break down whatever defences the other team puts up.

Both are then given free rein in their respective roles, learning as the other evolves its tactics -- either in defending or attacking -- so each team continues to find ways to improve its own techniques in response.

That is what ST Engineering has done, creating digital twins in which these two teams of AI agents set off on their given task, according to Ivan Jacobs, vice president and head of capability development for AI and cyber at ST Engineering, a Singapore government-linked tech conglomerate.

It went with this approach because the cyber threat landscape changes rapidly, with threat actors constantly evolving their attack vectors.

Ivan Jacobs

There inevitably will be blind spots, Jacobs said in an interview with FutureCISO, on the sidelines of ST Engineering Cybersecurity Summit in Singapore. He leads the company’s research efforts in AI for cybersecurity and works with its business teams on AI-related product roadmaps.

Training AI models on labelled data, commonly done today, is limiting and not optimal to build effective cybersecurity solutions, he noted.

Most companies also are unwilling to share data to train AI models, he added.

His team, hence, refrained from looking at ways that needed data to be acquired or AI models that are trained on labelled data.

And while there may be some benefits to training AI on vertical- or purpose-specific large language models (LLMs), a blackbox learning method would enable the AI agents to achieve better results, Jacobs said.

It is akin to teaching a beginner in chess by showing them the basic rules of the game, rather than how to win via a certain strategy.

“In the digital twins…you don’t need to be familiar with a particular attack. You let the agents generate those attacks,” Jacobs said.

He noted, though, that foundation models were used as a starting point on which the digital twins were built. ST Engineering’s own AI models also were added to the mix, he said.

AI agents in competition evolve at speed

The AI agents then are allowed to evolve autonomously, unscripted, in the digital twins, running in an environment that combines competition, feedback, and reward mechanisms.

Essentially, the AI agents assess their environment, take actions, receive feedback via penalties or rewards, and tweak their tactics based on the feedback. The loop continues, with AI agents from each team constantly finetuning and optimising their strategy.

Related:  Mature, measurable zero-trust in place by 2026

The concept is not new and is frequently used in the gaming industry, Jacobs said. Nvidia, too, uses this approach for its robotics, creating foundational agents that run within a digital twin and deployed into the physical world only when they are deemed ready.

In ST Engineering’s digital twins, the red- and blue-team AI agents evolve rapidly because they are pitted against each other in a competition, he said.

In fact, the intelligence and speed of the attackers, or the red team, have been impressive, he noted.

It highlights the difficulty organisations that employ a human-only cyberdefence team will face in combating increasingly sophisticated AI-powered attacks, he said.

Businesses need agentic capabilities, including AI agents capable of acting autonomously, to be able to boost their cyber resilience, Jacobs said.

“A solely human defence or technical approach [most have now] isn’t going to be enough,” he said.

His team has been running the digital twins for a year and is looking to achieve some “equilibrium” between the red and blue teams, before it assesses how the learnings can be applied to build actual cybersecurity products.

When asked, he declined to say which team was currently “winning”.

Robert Hannigan

Need to safeguard AI development lifecycle

With AI projected to account for 75% of cyberattacks by end-2025, organisations need to leverage AI for cyberdefence, said Robert Hannigan, BlueVoyant’s EMEA chairman and former director of the UK government’s communications headquarters.

This can span different areas, including in threat detection and prediction, automated incident response, and refactoring insecure software codes, said Hannigan, during his keynote at the cybersecurity summit.

AI is fuelling the growth in attack volume and scale, as cybercriminals tap the technology to make even modest changes to malware. This was not possible just a couple of years ago, he said.

Like ST Engineering, Hannigan also pitched the use of agentic AI in cyberdefence and reduce the time to respond to cyber incidents.

LLM-powered agents have an 80% accuracy rate in detecting threats, compared to a human analyst’s 60%, he said, adding that the AI agents were able to improve on their initially lower accuracy rate through constant training.

He suggested that future SIEM (Security Information and Event Management) infrastructures and SOCs (security operations centres) will comprise multiple AI agents working together.

He added that agentic AI can used across several cyberdefence functions, including autonomous investigation guidance as well as recommendations, incident summarisation, and securing software development lifecycle.

Related:  98% of CEOs believe in the benefit of AI, study finds

As AI agents continue to advance, they also can be applied to AI-powered red-teaming, incident response orchestration, and reverse engineering of malware, he said.

Hannigan further mooted the need for a bill of materials (BOM) for AI systems, as these are largely dependent on two critical components: data and software.

Both are susceptible to vulnerabilities, driving the need to protect AI deployments against a range of risks, including data poisoning, prompt injection attacks, supply chain and third-party attacks, and hallucinations, he said.

He highlighted the need to defend the AI lifecycle, stressing the importance of securing its design, development, development, and operations and maintenance.

In securing its development, for instance, he recommended that organisations carry out an in-depth review of their software BOMs.

A software BOM lists all components that are in an application, including direct and third-party dependencies, libraries, unique identifiers, authors, and known vulnerabilities.

The document is deemed critical for users to manage the software’s development lifecycle, including its security.

Hannigan called for the need to more closely analyse software BOMs in procurement as well as the possibility of establishing a data BOM to address AI risks.

The latter will be critical as data integrity is increasingly challenged, with the rise of ransomware attacks and the risk of data poisoning, he said.

Potential need for personalised security tools

Hyper-personalised security tools may also be critical in combating sophisticated AI-powered threats, particularly, deepfakes, according to Ong Chen Hui, assistant chief executive of business and technology group, Infocomm and Media Development Authority (IMDA).

She highlighted the need to better understand what it takes to build secured systems, with AI developers now just as concerned about security and safety.

This was not the case just a few years ago, Ong said during a panel discussion at the cybersecurity summit.

The industry needs to get better at harnessing AI in cyberdefence, especially as a critical cognitive gap has emerged, she said.

She pointed to the use of GenAI to mimic humans, which has opened up risks that must be addressed and managed differently.

As security vendors push out products to detect deepfakes, Ong suggested the need to treat certain key individuals in an organisation differently.

Deepfake detection tools currently analyse humans uniformly, looking for signals such as lip synching and liveness, to detect AI-generated content.

With certain individuals in an organisation, such as the CFO or HR director, common targets in deepfake attacks, rules can be created to treat such personnel with more personalised authentication or profile details.

This can enable detection tools to more accurately detect deepfakes for specific individuals, Ong said.

Tags: agentic AIAI AgentsArtificial Intelligencecybersecuritydigital twinsgenerative AI
Eileen Yu

Eileen Yu

Eileen is currently an independent tech journalist and content specialist, providing analysis of key market developments across the Asian region and helping enterprises craft their communications plan. She also moderates panel discussions and roundtables, as well as provides media training to help senior executives better manage press interviews. Eileen has worked with corporate clients in markets, such as cybersecurity and enterprise software, and non-tech including financial services and logistics. She also has planned high-level panel and roundtable discussions and has been an invited speaker on online media. On CXOCIETY, she contributes articles across the four CXOCIETY brands -- FutureCIO, FutureCISO, FutureIoT, and FutureCFO -- covering key industry developments impacting the Asia-Pacific region, including cybersecurity, AI, data management, governance, workforce modernisation, and supply chain. Eileen has more than 25 years of industry experience at established media platforms, including ZDNET in Singapore, where she led the tech site's Asian editorial team and blogger network. Before her stint at ZDNET, she was assistant editor at Computer Times for Singapore Press Holdings and deputy editor of Computerworld Singapore. With her extensive industry experience, Eileen has navigated discussions on key trending topics including cybersecurity, artificial intelligence, quantum computing, edge/cloud computing, and regulatory policies. Eileen trained under the Journalism department at The University of Queensland, Australia. There, she earned a Bachelor of Arts (Honours) degree in Journalism, with a thesis titled, To Censor or Not: The Great Singapore Dilemma.

No Result
View All Result

Recent Posts

  • Pitting AI agents against each other to uncover security holes
  • Construction under cyber siege: SEA targeted
  • PodChats for FutureCISO: Navigating the AI imperative
  • Sensitive data exposed: Calling for API protections
  • AI spending displaces traditional security budgets in APAC

Categories

  • Blogs
  • Compliance and Governance
  • Culture and Behaviour
  • Cybersecurity careers
  • Data Protection
  • Endpoint Security
  • Incident Response
  • Network Security
  • People
  • Process
  • Resources
  • Risk Management
  • Technology
  • Training and awarenes
  • Videos
  • Webinars and PodChats
  • White Papers

Strategic Insights for Chief Information Officers

FutureCISO serves the interests of the Chief Information Security Officer (CISO) and the information security profession. Its purpose is to provide relevant and timely industry insights around all things important to security professionals and organisations that recognize and value the importance of protecting the organisation’s data and its customers’ privacy.

Cxociety Media Brands

  • FutureIoT
  • FutureCFO
  • FutureCIO

Categories

  • Privacy Policy
  • Terms of Use
  • Cookie Policy

Copyright © 2024 Cxociety Pte Ltd | Designed by Pixl

Login to your account below

or

Not a member yet? Register here

Forgotten Password?

Fill the forms bellow to register

All fields are required. Log In

Retrieve your password

Please enter your username or email address to reset your password.

Log In
No Result
View All Result
  • People
  • Process
  • Technology
  • Resources
    • White Papers
    • PodChats
Login

Copyright © 2024 Cxociety Pte Ltd | Designed by Pixl