<?xml version="1.0" encoding="UTF-8"?><rss version="2.0"
	xmlns:content="http://purl.org/rss/1.0/modules/content/"
	xmlns:wfw="http://wellformedweb.org/CommentAPI/"
	xmlns:dc="http://purl.org/dc/elements/1.1/"
	xmlns:atom="http://www.w3.org/2005/Atom"
	xmlns:sy="http://purl.org/rss/1.0/modules/syndication/"
	xmlns:slash="http://purl.org/rss/1.0/modules/slash/"
	>

<channel>
	<title>Mathieu Bricou-Belouet, Auteur</title>
	<atom:link href="https://www.riskinsight-wavestone.com/en/author/mathieu-bricoubelouet/feed/" rel="self" type="application/rss+xml" />
	<link>https://www.riskinsight-wavestone.com/author/mathieu-bricoubelouet/</link>
	<description>The cybersecurity &#38; digital trust blog by Wavestone&#039;s consultants</description>
	<lastBuildDate>Tue, 07 Apr 2026 17:54:14 +0000</lastBuildDate>
	<language>en-US</language>
	<sy:updatePeriod>
	hourly	</sy:updatePeriod>
	<sy:updateFrequency>
	1	</sy:updateFrequency>
	

<image>
	<url>https://www.riskinsight-wavestone.com/wp-content/uploads/2024/02/Blogs-2024_RI-39x39.png</url>
	<title>Mathieu Bricou-Belouet, Auteur</title>
	<link>https://www.riskinsight-wavestone.com/author/mathieu-bricoubelouet/</link>
	<width>32</width>
	<height>32</height>
</image> 
	<item>
		<title>Agentic AI for Offensive Security</title>
		<link>https://www.riskinsight-wavestone.com/en/2026/04/agentic-ai-for-offensive-security/</link>
					<comments>https://www.riskinsight-wavestone.com/en/2026/04/agentic-ai-for-offensive-security/#respond</comments>
		
		<dc:creator><![CDATA[Mathieu Bricou-Belouet]]></dc:creator>
		<pubDate>Tue, 07 Apr 2026 14:43:09 +0000</pubDate>
				<category><![CDATA[Ethical Hacking & Incident Response]]></category>
		<category><![CDATA[Focus]]></category>
		<category><![CDATA[Agentic AI]]></category>
		<category><![CDATA[AI]]></category>
		<category><![CDATA[AI Hallucinations]]></category>
		<category><![CDATA[Autonomous Pentesting]]></category>
		<category><![CDATA[ctf]]></category>
		<category><![CDATA[LLM]]></category>
		<category><![CDATA[offensive security]]></category>
		<category><![CDATA[pentest]]></category>
		<category><![CDATA[Vulnerabilities]]></category>
		<category><![CDATA[Web pentesting]]></category>
		<guid isPermaLink="false">https://www.riskinsight-wavestone.com/?p=29693</guid>

					<description><![CDATA[<p>AI is now embedded across a growing range of offensive security workflows. The most visible shift is the rise of services that apply large language models and agentic orchestration to autonomous testing activity. Some vendors have been present for years,...</p>
<p>Cet article <a href="https://www.riskinsight-wavestone.com/en/2026/04/agentic-ai-for-offensive-security/">Agentic AI for Offensive Security</a> est apparu en premier sur <a href="https://www.riskinsight-wavestone.com/en/">RiskInsight</a>.</p>
]]></description>
										<content:encoded><![CDATA[
<p style="text-align: justify;">AI is now embedded across a growing range of offensive security workflows. The most visible shift is the rise of services that apply large language models and agentic orchestration to autonomous testing activity. Some vendors have been present for years, while others have emerged only recently, but the pace of change has clearly accelerated over the last six months.</p>
<p style="text-align: justify;">Commercial offerings include editor-backed platforms such as Horizon3.ai / NodeZero, Pentera, XBOW, and RunSybil, while the open-source ecosystem includes projects such as Strix, Shannon, PentAGI, PentestGPT, and PentestAgent. Their positioning differs, but they all attempt to translate the adaptability of modern AI systems into concrete offensive security outcomes.</p>
<p style="text-align: justify;">The objective of this article is not to rank vendors. Instead, it is to clarify how agentic pentesting systems work, what technical prerequisites they require, and where their current limitations still prevent them from being treated as fully reliable autonomous testers.</p>
<p> </p>
<h2>A common architecture for agentic offensive testing</h2>
<p style="text-align: justify;">The current landscape is made up of heterogeneous tools with very different product strategies and target use cases: external web security testing, internal infrastructure and Active Directory reviews, cloud security assessments, or source-code analysis close to the CI/CD pipeline.</p>
<p style="text-align: justify;">Nowadays, in their best configurations, the strongest systems can conduct autonomous static and dynamic security reviews with strong reasoning capabilities, and a workflow that can, at times, resemble the analytical posture of a human pentester.</p>
<figure id="attachment_29694" aria-describedby="caption-attachment-29694" style="width: 1511px" class="wp-caption aligncenter"><img fetchpriority="high" decoding="async" class="size-full wp-image-29694" src="https://www.riskinsight-wavestone.com/wp-content/uploads/2026/04/1-Example-of-autonomous-reasoning-and-tool-execution.png" alt="Example of autonomous reasoning and tool execution" width="1511" height="767" srcset="https://www.riskinsight-wavestone.com/wp-content/uploads/2026/04/1-Example-of-autonomous-reasoning-and-tool-execution.png 1511w, https://www.riskinsight-wavestone.com/wp-content/uploads/2026/04/1-Example-of-autonomous-reasoning-and-tool-execution-376x191.png 376w, https://www.riskinsight-wavestone.com/wp-content/uploads/2026/04/1-Example-of-autonomous-reasoning-and-tool-execution-71x36.png 71w, https://www.riskinsight-wavestone.com/wp-content/uploads/2026/04/1-Example-of-autonomous-reasoning-and-tool-execution-768x390.png 768w" sizes="(max-width: 1511px) 100vw, 1511px" /><figcaption id="caption-attachment-29694" class="wp-caption-text"><em>Example of autonomous reasoning and tool execution</em></figcaption></figure>
<p style="text-align: justify;">Many of these tools are benchmarked internally, or through capture-the-flag environments, as CTFs provide an observable way to compare reasoning depth, exploitation ability, and tool usage. Despite a wide range of architecture, the following essential building blocks are broadly consistent across most solutions:</p>
<figure id="attachment_29696" aria-describedby="caption-attachment-29696" style="width: 1837px" class="wp-caption aligncenter"><img decoding="async" class="size-full wp-image-29696" src="https://www.riskinsight-wavestone.com/wp-content/uploads/2026/04/2-Standard-architecture-and-components-of-an-agentic-automated-pentesting-solution.png" alt="Standard architecture and components of an agentic automated pentesting solution" width="1837" height="561" srcset="https://www.riskinsight-wavestone.com/wp-content/uploads/2026/04/2-Standard-architecture-and-components-of-an-agentic-automated-pentesting-solution.png 1837w, https://www.riskinsight-wavestone.com/wp-content/uploads/2026/04/2-Standard-architecture-and-components-of-an-agentic-automated-pentesting-solution-437x133.png 437w, https://www.riskinsight-wavestone.com/wp-content/uploads/2026/04/2-Standard-architecture-and-components-of-an-agentic-automated-pentesting-solution-71x22.png 71w, https://www.riskinsight-wavestone.com/wp-content/uploads/2026/04/2-Standard-architecture-and-components-of-an-agentic-automated-pentesting-solution-768x235.png 768w, https://www.riskinsight-wavestone.com/wp-content/uploads/2026/04/2-Standard-architecture-and-components-of-an-agentic-automated-pentesting-solution-1536x469.png 1536w" sizes="(max-width: 1837px) 100vw, 1837px" /><figcaption id="caption-attachment-29696" class="wp-caption-text"><em>Standard architecture and components of an agentic automated pentesting solution</em></figcaption></figure>
<ul>
<li style="text-align: justify;"><strong>An orchestrator: </strong>This layer coordinates parallel agents, handles freezes and timeouts, manages preconfigured workflows, and connects the other components into a coherent execution chain.</li>
<li style="text-align: justify;"><strong>An underlying LLM: </strong>The model acts as the cognitive core of the system, alternating between reasoning loops, tool invocation, and the creation of sub-agents when needed. Tool use is mandatory, and larger frontier models generally yield better results.</li>
<li style="text-align: justify;"><strong>An attack toolbox: </strong>Most platforms rely on a containerized toolkit broadly aligned with standard Kali-style capabilities. The exact content varies by use case, but web testing stacks are often relatively conventional. Many solutions also allow the agent to download additional tools or clone GitHub repositories dynamicaly when required.</li>
<li style="text-align: justify;"><strong>A set of skills or knowledge packs: </strong>These local libraries encode reusable expertise, including technology-specific attack techniques, pentester cheat sheets, standard exploitation workflows, and details related to newly disclosed vulnerabilities or attack patterns.</li>
</ul>
<p style="text-align: justify;">This last layer is often where vendors can differentiate most clearly. Strong cyber monitoring, threat hunting, and cyber threat intelligence capabilities can continuously refresh the knowledge base and improve both adaptability and confidence in the actual coverage delivered by automated sessions.</p>
<p style="text-align: justify;">Because these agents can execute offensive actions against production-like environments, observability and governance are essential. Most serious implementations therefore include logging, telemetry, session replay, human approval steps for selected actions, and safeguards that distinguish lower-risk modules from more dangerous commands or exploit paths.</p>
<p style="text-align: justify;">A key distinction often blurred in vendor marketing: fully agentic systems use an LLM to drive the entire decision loop, while AI-assisted platforms apply AI only to specific steps (usually the hardest exploitation decisions) within an otherwise deterministic pipeline. Most commercial products today fall into the second category.</p>
<p> </p>
<h2>An efficiency case study</h2>
<h3>Case study : CTF</h3>
<p style="text-align: justify;">To assess the current effectiveness of agentic pentesting, we benchmarked one such solution (Strix) using several different models against an internal set of Wavestone CTF challenges for which no public write-ups were available. The goal was not to compare products against each other, but rather to understand how model quality affects outcomes in a web security context.</p>
<p style="text-align: justify;">This choice of benchmark offers a useful signal because web exploitation combines broad topic coverage with varying levels of difficulty. At the same time, the exercise should not be over-generalized: it does not fully represent other contexts such as internal infrastructure testing or Active Directory assessments.</p>
<figure id="attachment_29698" aria-describedby="caption-attachment-29698" style="width: 1838px" class="wp-caption aligncenter"><img decoding="async" class="size-full wp-image-29698" src="https://www.riskinsight-wavestone.com/wp-content/uploads/2026/04/3-Benchmark-of-several-LLMs-on-internal-CTF-challenges.png" alt="Benchmark of several LLMs on internal CTF challenges" width="1838" height="727" srcset="https://www.riskinsight-wavestone.com/wp-content/uploads/2026/04/3-Benchmark-of-several-LLMs-on-internal-CTF-challenges.png 1838w, https://www.riskinsight-wavestone.com/wp-content/uploads/2026/04/3-Benchmark-of-several-LLMs-on-internal-CTF-challenges-437x173.png 437w, https://www.riskinsight-wavestone.com/wp-content/uploads/2026/04/3-Benchmark-of-several-LLMs-on-internal-CTF-challenges-71x28.png 71w, https://www.riskinsight-wavestone.com/wp-content/uploads/2026/04/3-Benchmark-of-several-LLMs-on-internal-CTF-challenges-768x304.png 768w, https://www.riskinsight-wavestone.com/wp-content/uploads/2026/04/3-Benchmark-of-several-LLMs-on-internal-CTF-challenges-1536x608.png 1536w" sizes="(max-width: 1838px) 100vw, 1838px" /><figcaption id="caption-attachment-29698" class="wp-caption-text"><em>Benchmark of several LLMs on internal CTF challenges</em></figcaption></figure>
<p style="text-align: justify;">Several conclusions emerged from this exercise:</p>
<ul style="text-align: justify;">
<li>The results become genuinely impressive only when the system is paired with a state-of-the-art model.</li>
<li>Conversely, models that can realistically run on a high-end consumer workstation still tend to produce mediocre offensive-testing performance, which often makes SaaS-based AI providers the sole effective solution today.</li>
<li>Even powerful models can miss exploitable weaknesses, while some still-large but less optimized models can underperform, potentially because Strix was not designed and tuned with them in mind.</li>
<li>Smaller models occasionally show flashes of insight and solve challenges that stronger models miss.</li>
<li>A broad tendency remains for models to hallucinate paths to exploitation, especially when they reach a dead end. In CTF settings this often manifests as fabricated flags rather than validated solutions.</li>
<li>In order to not pollute their context with large volume of data, agents tend to heavily truncate data (such as web pages or codebase files) and being too specific when using “grep” or “find” for research. In both cases, the behavior can restrict their coverage of the scope and their overall efficiency.</li>
</ul>
<p style="text-align: justify;">These results should be interpreted cautiously. For each model and each challenge, the benchmark was limited to at most two runs. In several cases, a model was very close to the solution before hallucinating the final step, or required human steering to close the investigation. Typically, those cases could plausibly be recovered in a real-world workflow that includes human review.</p>
<p style="text-align: justify;">The best benchmark results were obtained with frontier proprietary models. In our observations, these models can solve a substantial portion of constrained offensive tasks while remaining operationally affordable; at least as long as sessions converge quickly.</p>
<figure id="attachment_29700" aria-describedby="caption-attachment-29700" style="width: 1590px" class="wp-caption aligncenter"><img loading="lazy" decoding="async" class="size-full wp-image-29700" src="https://www.riskinsight-wavestone.com/wp-content/uploads/2026/04/4-Performance-of-a-frontier-model-and-key-consumption-metrics.png" alt="Performance of a frontier model and key consumption metrics" width="1590" height="899" srcset="https://www.riskinsight-wavestone.com/wp-content/uploads/2026/04/4-Performance-of-a-frontier-model-and-key-consumption-metrics.png 1590w, https://www.riskinsight-wavestone.com/wp-content/uploads/2026/04/4-Performance-of-a-frontier-model-and-key-consumption-metrics-338x191.png 338w, https://www.riskinsight-wavestone.com/wp-content/uploads/2026/04/4-Performance-of-a-frontier-model-and-key-consumption-metrics-69x39.png 69w, https://www.riskinsight-wavestone.com/wp-content/uploads/2026/04/4-Performance-of-a-frontier-model-and-key-consumption-metrics-768x434.png 768w, https://www.riskinsight-wavestone.com/wp-content/uploads/2026/04/4-Performance-of-a-frontier-model-and-key-consumption-metrics-1536x868.png 1536w" sizes="auto, (max-width: 1590px) 100vw, 1590px" /><figcaption id="caption-attachment-29700" class="wp-caption-text"><em>Performance of a frontier model and key consumption metrics</em></figcaption></figure>
<p> </p>
<figure id="attachment_29702" aria-describedby="caption-attachment-29702" style="width: 1579px" class="wp-caption aligncenter"><img loading="lazy" decoding="async" class="size-full wp-image-29702" src="https://www.riskinsight-wavestone.com/wp-content/uploads/2026/04/5-Performance-of-an-alternative-frontier-model-and-key-consumption-metrics.png" alt="Performance of an alternative frontier model and key consumption metrics" width="1579" height="899" srcset="https://www.riskinsight-wavestone.com/wp-content/uploads/2026/04/5-Performance-of-an-alternative-frontier-model-and-key-consumption-metrics.png 1579w, https://www.riskinsight-wavestone.com/wp-content/uploads/2026/04/5-Performance-of-an-alternative-frontier-model-and-key-consumption-metrics-335x191.png 335w, https://www.riskinsight-wavestone.com/wp-content/uploads/2026/04/5-Performance-of-an-alternative-frontier-model-and-key-consumption-metrics-68x39.png 68w, https://www.riskinsight-wavestone.com/wp-content/uploads/2026/04/5-Performance-of-an-alternative-frontier-model-and-key-consumption-metrics-768x437.png 768w, https://www.riskinsight-wavestone.com/wp-content/uploads/2026/04/5-Performance-of-an-alternative-frontier-model-and-key-consumption-metrics-1536x875.png 1536w" sizes="auto, (max-width: 1579px) 100vw, 1579px" /><figcaption id="caption-attachment-29702" class="wp-caption-text"><em>Performance of an alternative frontier model and key consumption metrics</em></figcaption></figure>
<p>What it shows is :</p>
<ul>
<li style="text-align: justify;">Per-challenge cost can remain relatively modest, on the order of a few euros when the agent converges efficiently.</li>
<li style="text-align: justify;">Execution can be surprisingly fast, with many CTFs solved in less than five minutes when the model identifies the relevant path early.</li>
<li style="text-align: justify;">Failure is expensive. Without strict guardrails on duration and budget, token consumption can increase dramatically over the course of a few hours.</li>
<li style="text-align: justify;">In our own setup, solve rates between top-tier commercial models were close, but efficiency varied substantially in time, token consumption, and number of tool invocations. Surprisingly, despite Sonnet&#8217;s higher per-token price, overall session costs were comparable to GPT-5, Anthropic&#8217;s model compensated through greater token efficiency.</li>
</ul>
<p> </p>
<h3 style="text-align: justify;">Case study : real web application</h3>
<p style="text-align: justify;">To complement the CTF benchmarks, we also tested one of our internally developed web applications (used for staffing and performance management). The system was assessed with several approaches, including authenticated modes in which the agent is provided with credentials or tokens.</p>
<p style="text-align: justify;">In one representative pentesting session, 25 agents were deployed, 366 tool calls were executed, for a total cost around USD 5, and the session ran for around one hour. The resulting automatically generated report included an executive summary, an OWASP-oriented methodology section, technical findings with CVSS v3 scoring, and a prioritized remediation roadmap.</p>
<figure id="attachment_29704" aria-describedby="caption-attachment-29704" style="width: 706px" class="wp-caption aligncenter"><img loading="lazy" decoding="async" class="size-full wp-image-29704" src="https://www.riskinsight-wavestone.com/wp-content/uploads/2026/04/6-Agent-hierarchy-spawned-during-an-automated-security-review.png" alt="Agent hierarchy spawned during an automated security review" width="706" height="771" srcset="https://www.riskinsight-wavestone.com/wp-content/uploads/2026/04/6-Agent-hierarchy-spawned-during-an-automated-security-review.png 706w, https://www.riskinsight-wavestone.com/wp-content/uploads/2026/04/6-Agent-hierarchy-spawned-during-an-automated-security-review-175x191.png 175w, https://www.riskinsight-wavestone.com/wp-content/uploads/2026/04/6-Agent-hierarchy-spawned-during-an-automated-security-review-36x39.png 36w" sizes="auto, (max-width: 706px) 100vw, 706px" /><figcaption id="caption-attachment-29704" class="wp-caption-text"><em>Agent hierarchy spawned during an automated security review</em></figcaption></figure>
<p style="text-align: justify;">The outputs were mixed, but broadly informative after human review and retesting:</p>
<ul style="text-align: justify;">
<li>The agent surfaced several relevant minor improvement areas, although findings were not always well contextualized and could become overly alarmist.</li>
<li>Critical miss however : the agent completely missed an exposed admin interface with default credentials: a vulnerability no human pentester would overlook. This illustrates the reliability ceiling of current autonomous systems.</li>
<li>The report also included a non-existent vulnerability candidate, JWT algorithm confusion, rated as critical, along with proof-of-exploit scripts that did not succeed in practice. This illustrates the persistent false-positive risk of autonomous systems.</li>
</ul>
<p style="text-align: justify;">Additional remarks :</p>
<ul style="text-align: justify;">
<li>As with the CTF benchmarks, the quality of the review improved significantly when using a frontier-grade model.</li>
<li>The non-deterministic nature of generative models remains visible: two runs can produce substantially different findings and reports against the same target.</li>
<li>If prompting and scope controls are insufficient, some models attempt to expand the scope of the assessment by probing adjacent ports, applications, or subdomains.</li>
<li>Coverage and relevance improve markedly in white-box or hybrid white-box/grey-box modes, where the agent can inspect the codebase, identify candidate weaknesses, and then attempt to validate them dynamically on the live application. Even then, some agents can still fixate on non-existent issues. And in white-box, very large codebases may saturate the system and reduce overall efficiency.</li>
<li>Browser-driven interactions have progressed, yet some application types remain difficult to assess autonomously, especially multi-window or thick-client environments where headless browser interaction may not be enough.</li>
<li>These systems rarely build a deep understanding of business logic. Their outputs remain strongly aligned with generic OWASP-style patterns and may not challenge the real business risk or abuse scenarios in a sufficiently contextual way.</li>
</ul>
<p style="text-align: justify;">It should be noted that the majority of these criticisms can also apply to human pentesters, who nonetheless remain more easily held accountable.</p>
<p style="text-align: justify;">The scaling problem remains central. CTFs are only partially representative of real applications. While a CTF typically channels the tester toward a narrow and deliberate attack path, even a modest business application exposes a much broader surface. Today, guaranteeing exhaustiveness while avoiding fixation on irrelevant endpoints remains difficult.</p>
<p> </p>
<h2>Verdict and current limitations</h2>
<h3>Verdict</h3>
<p style="text-align: justify;">If one considers solutions that relies entirely on a general-purpose LLM for its decision tree, the conclusion is clear at the present time: only frontier-grade models from major AI providers consistently deliver results that are both relevant and reasonably verifiable.</p>
<p style="text-align: justify;">Condisering four practical deployment options:</p>
<ul>
<li style="text-align: justify;">SaaS LLM services: currently the highest-quality option, leveraging very large frontier models (&gt;1T parameters) billed per use. The main drawback is data sovereignty: all prompts and findings leave your environment.</li>
<li style="text-align: justify;">Large private datacenter deployments, which can run powerful models (500b) and may become increasingly relevant for pentesting, but may still remain materially below the best commercial frontier systems.</li>
<li style="text-align: justify;">Small private datacenter deployments, which can run capable models (300b), but clearly not sufficient to efficiently orchestrate autonomous pentests.</li>
<li style="text-align: justify;">Dedicated workstations, which, even with very strong specifications, may quickly struggle above 100b, and remain far insufficient today.</li>
</ul>
<figure id="attachment_29706" aria-describedby="caption-attachment-29706" style="width: 1716px" class="wp-caption aligncenter"><img loading="lazy" decoding="async" class="size-full wp-image-29706" src="https://www.riskinsight-wavestone.com/wp-content/uploads/2026/04/7-Illustrative-distribution-of-open-source-local-models-by-number-of-parameters-and-total-size.png" alt="Illustrative distribution of open-source local models by number of parameters and total size" width="1716" height="924" srcset="https://www.riskinsight-wavestone.com/wp-content/uploads/2026/04/7-Illustrative-distribution-of-open-source-local-models-by-number-of-parameters-and-total-size.png 1716w, https://www.riskinsight-wavestone.com/wp-content/uploads/2026/04/7-Illustrative-distribution-of-open-source-local-models-by-number-of-parameters-and-total-size-355x191.png 355w, https://www.riskinsight-wavestone.com/wp-content/uploads/2026/04/7-Illustrative-distribution-of-open-source-local-models-by-number-of-parameters-and-total-size-71x39.png 71w, https://www.riskinsight-wavestone.com/wp-content/uploads/2026/04/7-Illustrative-distribution-of-open-source-local-models-by-number-of-parameters-and-total-size-768x414.png 768w, https://www.riskinsight-wavestone.com/wp-content/uploads/2026/04/7-Illustrative-distribution-of-open-source-local-models-by-number-of-parameters-and-total-size-1536x827.png 1536w" sizes="auto, (max-width: 1716px) 100vw, 1716px" /><figcaption id="caption-attachment-29706" class="wp-caption-text"><em>Illustrative distribution of open-source local models by number of parameters and total size</em></figcaption></figure>
<p style="text-align: justify;">The dependence on SaaS providers raises unavoidable sovereignty and confidentiality questions. Offensive security assessments often consolidate highly sensitive technical information about an organization’s weaknesses. Any externalization of prompts, traces, findings, or attack hypotheses therefore requires careful governance. And data anonymisation before the LLM step might not be a reliable mitigation, as it can decrease the efficiency of the run, while still sharing exploitable meta-data my SaaS suppliers.</p>
<p style="text-align: justify;">In their current state, even equipped with the most capable LLMs, these systems also exhibit structural limitations that directly affect reliability:</p>
<ul>
<li style="text-align: justify;">Instances of “tunnel vision”, with prolonged fixation on a single irrelevant attack path.</li>
<li style="text-align: justify;">A tendency to launch time-consuming brute-force activities without a sound appreciation of computational complexity or cost.</li>
<li style="text-align: justify;">Persistent hallucinations: despite significant progress, even frontier models still fabricate findings, exploit paths, or flag non-existent vulnerabilities, as shown in the JWT confusion example.</li>
</ul>
<figure id="attachment_29708" aria-describedby="caption-attachment-29708" style="width: 1511px" class="wp-caption aligncenter"><img loading="lazy" decoding="async" class="size-full wp-image-29708" src="https://www.riskinsight-wavestone.com/wp-content/uploads/2026/04/8-Easy-capability-to-hallucinate-or-misinterpret-results-here-with-kimi-k2-.png" alt="Easy capability to hallucinate or misinterpret results, here with kimi-k2" width="1511" height="334" srcset="https://www.riskinsight-wavestone.com/wp-content/uploads/2026/04/8-Easy-capability-to-hallucinate-or-misinterpret-results-here-with-kimi-k2-.png 1511w, https://www.riskinsight-wavestone.com/wp-content/uploads/2026/04/8-Easy-capability-to-hallucinate-or-misinterpret-results-here-with-kimi-k2--437x97.png 437w, https://www.riskinsight-wavestone.com/wp-content/uploads/2026/04/8-Easy-capability-to-hallucinate-or-misinterpret-results-here-with-kimi-k2--71x16.png 71w, https://www.riskinsight-wavestone.com/wp-content/uploads/2026/04/8-Easy-capability-to-hallucinate-or-misinterpret-results-here-with-kimi-k2--768x170.png 768w" sizes="auto, (max-width: 1511px) 100vw, 1511px" /><figcaption id="caption-attachment-29708" class="wp-caption-text"><em>Easy capability to hallucinate or misinterpret results, here with kimi-k2</em></figcaption></figure>
<ul>
<li style="text-align: justify;">The non deterministic nature of LLM, making some runs way less efficient and relevant than others</li>
<li style="text-align: justify;">A scaling problem tied to context-window constraints: it “scales” in the sense that you can launch as many parallel sessions against as many targets. However, it scales more poorly when a single session is launched against a single highly complex application. It becomes much harder to maintain exhaustive coverage and memory continuity across large, content-rich applications. Large improvments can be achieved on this front, with an efficient long term memory management allowing for more coherent runs for large applications and improving coverage.</li>
<li style="text-align: justify;">High verbosity and limited stealth, which makes these systems poorly suited in their default form for red-team style end-to-end scenarios that require discretion and tradecraft. This can be improved through dedicated configuration, without however equaling human capabilities</li>
</ul>
<p style="text-align: justify;">And from a higher standpoint, an autonomous SaaS-run process having the ability to remotely execute commands in your IS poses from the start the issue of accountability :</p>
<ul style="text-align: justify;">
<li>Classifying tools as dangerous versus safe may not be enough, for instance with Swiss-army toolsets, capable of the most inocuous recon and of aggressive and potentially damaging exploits. Threat level should be dynamically assessed, taking the context and previous tests into accounts.</li>
<li>Even then, pausing the tests and requesting a human approval may lead to a similar situation with coding agents, with “developer fatigue”, where users become too trusting and stop critically challenging the agent’s conclusions.</li>
</ul>
<p style="text-align: justify;">And of course, any vulnerability at the LLM level, such as susceptibility to prompt injection or poisonning, could be leveraged to hijack the automated pentest workflow. Essentially, those autonomous tools, if deployed internally, should be regarded as critical assets, with high value for attackers.</p>
<p> </p>
<h3>Where the architecture can improve</h3>
<p style="text-align: justify;">Beyond model quality itself, a substantial part of the improvement space lies in the overall system design. Several architectural directions already appear promising:</p>
<ul style="text-align: justify;">
<li>Multiply sessions and validation passes, using continuous exploration, focused zoom-in phases, and explicit confirmation loops for candidate findings. This improves reliability but increases cost and duration.</li>
<li>Precede the autonomous phase with scripted tests and deterministic reconnaissance, then feed those structured outputs to the agent. This is far more cost-efficient than spending LLM context and tokens on tasks that are already easy to automate without AI. The core principle should be simple: do not use AI where conventional automation already performs well. Delegate only the genuinely ambiguous, adaptive, or investigative parts of the workflow to the LLM, and avoid overloading the model with unnecessary command history and context noise.</li>
<li>Introduce dedicated validation instances to confirm exploitability in a controlled environment before findings are promoted to a report.</li>
<li>Use leaner decision trees or specialized modules upstream of exploitation, reserving high-end models only for the parts of the workflow that truly require adaptability and reasoning.</li>
</ul>
<p style="text-align: justify;">In practice, this last point is already the direction taken by many vendor platforms. They do not rely entirely on agentic AI; instead, they combine deterministic security logic with agentic exploitation only when potential weaknesses have already been narrowed down.</p>
<figure id="attachment_29710" aria-describedby="caption-attachment-29710" style="width: 1854px" class="wp-caption aligncenter"><img loading="lazy" decoding="async" class="size-full wp-image-29710" src="https://www.riskinsight-wavestone.com/wp-content/uploads/2026/04/9-Potential-multi-step-architecture-designed-to-improve-result-reliability-and-reduce-unnecessary-model-load.png" alt="Potential multi-step architecture designed to improve result reliability and reduce unnecessary model load" width="1854" height="798" srcset="https://www.riskinsight-wavestone.com/wp-content/uploads/2026/04/9-Potential-multi-step-architecture-designed-to-improve-result-reliability-and-reduce-unnecessary-model-load.png 1854w, https://www.riskinsight-wavestone.com/wp-content/uploads/2026/04/9-Potential-multi-step-architecture-designed-to-improve-result-reliability-and-reduce-unnecessary-model-load-437x188.png 437w, https://www.riskinsight-wavestone.com/wp-content/uploads/2026/04/9-Potential-multi-step-architecture-designed-to-improve-result-reliability-and-reduce-unnecessary-model-load-71x31.png 71w, https://www.riskinsight-wavestone.com/wp-content/uploads/2026/04/9-Potential-multi-step-architecture-designed-to-improve-result-reliability-and-reduce-unnecessary-model-load-768x331.png 768w, https://www.riskinsight-wavestone.com/wp-content/uploads/2026/04/9-Potential-multi-step-architecture-designed-to-improve-result-reliability-and-reduce-unnecessary-model-load-1536x661.png 1536w" sizes="auto, (max-width: 1854px) 100vw, 1854px" /><figcaption id="caption-attachment-29710" class="wp-caption-text"><em>Potential multi-step architecture designed to improve result reliability and reduce unnecessary model load</em></figcaption></figure>
<p style="text-align: justify;">Lastly, an interesting thought : as such automated solutions may be used by real attackers, we may see “anti-AI” mechanisms included in applications and endpoints, such as “links labyrith” and token-draining honeypots designed specifically to mislead or exhaust automated testing systems.</p>
<p style="text-align: justify;"> </p>
<p style="text-align: justify;">With strong enough models, agentic systems can already excel in constrained environments such as CTFs. Their performance in real application assessments is more mixed: often useful, sometimes impressive, but still too inconsistent to be trusted without human oversight.</p>
<p style="text-align: justify;">The most pragmatic path today is therefore a hybrid operating model: an agentic system carrying out the majority of the tests and suggesting investigation leads, supported by human pentesters who arbitrate, validate, and take over in the most complex cases. The result is a security assessment that is significantly shorter, while still guaranteeing a degree of coverage and relevance in the findings.</p>
<p style="text-align: justify;">Agentic AI is not a replacement for human pentesters, not yet. At its current level of maturity, it is better understood as a force multiplier, one that can accelerate exploration and triage, but that still depends on expert supervision to turn raw autonomous activity into trustworthy security outcomes. In any case, these systems should also be treated as highly sensitive because of their autonomous nature, and the current constraints toward SaaS-run models should be considered, in terms of data confidentiality and digital souvereignty.</p>
<p style="text-align: justify;">Despite not being fully mature yet, those solutions are beginning to leave a mark in the cybersecurity landscape, and will most likely alter the trajectory of the pentesting market, toward an ecosystem more centered on tools and compute while conserving a hybrid approach. We might even see audits following a “Bring Your Own Compute” model, where auditees provide their own LLM, and the auditors provide custom tools and skills.</p>
<p> </p>
<p>Cet article <a href="https://www.riskinsight-wavestone.com/en/2026/04/agentic-ai-for-offensive-security/">Agentic AI for Offensive Security</a> est apparu en premier sur <a href="https://www.riskinsight-wavestone.com/en/">RiskInsight</a>.</p>
]]></content:encoded>
					
					<wfw:commentRss>https://www.riskinsight-wavestone.com/en/2026/04/agentic-ai-for-offensive-security/feed/</wfw:commentRss>
			<slash:comments>0</slash:comments>
		
		
			</item>
	</channel>
</rss>
