Close Menu
RoboNewsWire – Latest Insights on AI, Robotics, Crypto and Tech Innovations
  • Home
  • AI
  • Crypto
  • Cybersecurity
  • IT
  • Energy
  • Robotics
  • TechCrunch
  • Technology
What's Hot

Investors trust Google more than Meta when comes to spending on AI

April 30, 2026

Paragon is not collaborating with Italian authorities probing spyware attacks, report says

April 28, 2026

Microsoft cuts OpenAI revenue share as their AI alliance loosens

April 28, 2026
Facebook X (Twitter) Instagram
Trending
  • Investors trust Google more than Meta when comes to spending on AI
  • Paragon is not collaborating with Italian authorities probing spyware attacks, report says
  • Microsoft cuts OpenAI revenue share as their AI alliance loosens
  • Robotically assembled building blocks could make construction more efficient and sustainable | MIT News
  • AI showdown: Musk and Altman go to trial in fight over OpenAI’s beginnings
  • U.S., Iran seize ships as war evolves into standoff over Strait of Hormuz
  • Google launches training and inference TPUs in latest shot at Nvidia
  • Zoom teams up with World to verify humans in meetings
  • Home
  • About Us
  • Advertise
  • Contact Us
  • DMCA
  • Privacy Policy
  • Terms & Conditions
Facebook X (Twitter) Instagram
RoboNewsWire – Latest Insights on AI, Robotics, Crypto and Tech InnovationsRoboNewsWire – Latest Insights on AI, Robotics, Crypto and Tech Innovations
Thursday, May 7
  • Home
  • AI
  • Crypto
  • Cybersecurity
  • IT
  • Energy
  • Robotics
  • TechCrunch
  • Technology
RoboNewsWire – Latest Insights on AI, Robotics, Crypto and Tech Innovations
Home » Adversarial learning breakthrough enables real-time AI security

Adversarial learning breakthrough enables real-time AI security

GTBy GTNovember 26, 2025 AI No Comments5 Mins Read
Share
Facebook Twitter LinkedIn Pinterest Email


The ability to execute adversarial learning for real-time AI security offers a decisive advantage over static defence mechanisms.

The emergence of AI-driven attacks – utilising reinforcement learning (RL) and Large Language Model (LLM) capabilities – has created a class of “vibe hacking” and adaptive threats that mutate faster than human teams can respond. This represents a governance and operational risk for enterprise leaders that policy alone cannot mitigate.

Attackers now employ multi-step reasoning and automated code generation to bypass established defences. Consequently, the industry is observing a necessary migration toward “autonomic defence” (i.e. systems capable of learning, anticipating, and responding intelligently without human intervention.)

Transitioning to these sophisticated defence models, though, has historically hit a hard operational ceiling: latency.

Applying adversarial learning, where threat and defence models are trained continuously against one another, offers a method for countering malicious AI security threats. Yet, deploying the necessary transformer-based architectures into a live production environment creates a bottleneck.

Abe Starosta, Principal Applied Research Manager at Microsoft NEXT.ai, said: “Adversarial learning only works in production when latency, throughput, and accuracy move together. 

Computational costs associated with running these dense models previously forced leaders to choose between high-accuracy detection (which is slow) and high-throughput heuristics (which are less accurate).

Engineering collaboration between Microsoft and NVIDIA shows how hardware acceleration and kernel-level optimisation remove this barrier, making real-time adversarial defence viable at enterprise scale.

Operationalising transformer models for live traffic required the engineering teams to target the inherent limitations of CPU-based inference. Standard processing units struggle to handle the volume and velocity of production workloads when burdened with complex neural networks.

In baseline tests conducted by the research teams, a CPU-based setup yielded an end-to-end latency of 1239.67ms with a throughput of just 0.81req/s. For a financial institution or global e-commerce platform, a one-second delay on every request is operationally untenable.

By transitioning to a GPU-accelerated architecture (specifically utilising NVIDIA H100 units), the baseline latency dropped to 17.8ms. Hardware upgrades alone, though, proved insufficient to meet the strict requirements of real-time AI security.

Through further optimisation of the inference engine and tokenisation processes, the teams achieved a final end-to-end latency of 7.67ms—a 160x performance speedup compared to the CPU baseline. Such a reduction brings the system well within the acceptable thresholds for inline traffic analysis, enabling the deployment of detection models with greater than 95 percent accuracy on adversarial learning benchmarks.

One operational hurdle identified during this project offers valuable insight for CTOs overseeing AI integration. While the classifier model itself is computationally heavy, the data pre-processing pipeline – specifically tokenisation – emerged as a secondary bottleneck.

Standard tokenisation techniques, often relying on whitespace segmentation, are designed for natural language processing (e.g. articles and documentation). They prove inadequate for cybersecurity data, which consists of densely packed request strings and machine-generated payloads that lack natural breaks.

To address this, the engineering teams developed a domain-specific tokeniser. By integrating security-specific segmentation points tailored to the structural nuances of machine data, they enabled finer-grained parallelism. This bespoke approach for security delivered a 3.5x reduction in tokenisation latency, highlighting that off-the-shelf AI components often require domain-specific re-engineering to function effectively in niche environments.

Achieving these results required a cohesive inference stack rather than isolated upgrades. The architecture utilised NVIDIA Dynamo and Triton Inference Server for serving, coupled with a TensorRT implementation of Microsoft’s threat classifier.

The optimisation process involved fusing key operations – such as normalisation, embedding, and activation functions – into single custom CUDA kernels. This fusion minimises memory traffic and launch overhead, which are frequent silent killers of performance in high-frequency trading or security applications. TensorRT automatically fused normalisation operations into preceding kernels, while developers built custom kernels for sliding window attention.

The result of these specific inference optimisations was a reduction in forward-pass latency from 9.45ms to 3.39ms, a 2.8x speedup that contributed the majority of the latency reduction seen in the final metrics.

Rachel Allen, Cybersecurity Manager at NVIDIA, explained: “Securing enterprises means matching the volume and velocity of cybersecurity data and adapting to the innovation speed of adversaries.

“Defensive models need the ultra-low latency to run at line-rate and the adaptability to protect against the latest threats. The combination of adversarial learning with NVIDIA TensorRT accelerated transformer-based detection models does just that.”

Success here points to a broader requirement for enterprise infrastructure. As threat actors leverage AI to mutate attacks in real-time, security mechanisms must possess the computational headroom to run complex inference models without introducing latency.

Reliance on CPU compute for advanced threat detection is becoming a liability. Just as graphics rendering moved to GPUs, real-time security inference requires specialised hardware to maintain throughput >130 req/s while ensuring robust coverage.

Furthermore, generic AI models and tokenisers often fail on specialised data. The “vibe hacking” and complex payloads of modern threats require models trained specifically on malicious patterns and input segmentations that reflect the reality of machine data.

Looking ahead, the roadmap for future security involves training models and architectures specifically for adversarial robustness, potentially using techniques like quantisation to further enhance speed.

By continuously training threat and defence models in tandem, organisations can build a foundation for real-time AI protection that scales with the complexity of evolving security threats. The adversarial learning breakthrough demonstrates the technology to achieve this – balancing latency, throughput, and accuracy – is now capable of being deployed today.

See also: ZAYA1: AI model using AMD GPUs for training hits milestone

Banner for AI & Big Data Expo by TechEx events.

Want to learn more about AI and big data from industry leaders? Check out AI & Big Data Expo taking place in Amsterdam, California, and London. The comprehensive event is part of TechEx and is co-located with other leading technology events including the Cyber Security Expo. Click here for more information.

AI News is powered by TechForge Media. Explore other upcoming enterprise technology events and webinars here.



Source link

GT
  • Website

Keep Reading

Enterprise users swap AI pilots for deep integrations

Google, Sony Innovation Fund, and Okta back Resemble AI deepfake detection plan

Platform corrects AI algorithmic bias for eKYC

What ByteDance’s Launch Means for Enterprise

UK and Germany plan to commercialise quantum supercomputing

Frontier AI agents replace chatbots

Add A Comment
Leave A Reply Cancel Reply

Editors Picks

Investors trust Google more than Meta when comes to spending on AI

April 30, 2026

Google launches training and inference TPUs in latest shot at Nvidia

April 27, 2026

Meta tracks employee usage on Google, LinkedIn AI training project

April 25, 2026

Meta will cut 10% of workforce as company pushes deeper into AI

April 24, 2026
Latest Posts

Malicious Chrome Extension Steal ChatGPT and DeepSeek Conversations from 900K Users

April 1, 2026

Top 10 Best Server Monitoring Tools

April 1, 2026

10 Best Cybersecurity Risk Management Tools

March 31, 2026

Subscribe to News

Subscribe to our newsletter and never miss our latest news

Subscribe my Newsletter for New Posts & tips Let's stay updated!

Welcome to RoboNewsWire, your trusted source for cutting-edge news and insights in the world of technology. We are dedicated to providing timely and accurate information on the most important trends shaping the future across multiple sectors. Our mission is to keep you informed and ahead of the curve with deep dives, expert analysis, and the latest updates in key industries that are transforming the world.

Subscribe to Updates

Subscribe to our newsletter and never miss our latest news

Subscribe my Newsletter for New Posts & tips Let's stay updated!

Facebook X (Twitter) Instagram
  • Home
  • About Us
  • Advertise
  • Contact Us
  • DMCA
  • Privacy Policy
  • Terms & Conditions
© 2026 Robonewswire. Designed by robonewswire.

Type above and press Enter to search. Press Esc to cancel.