<?xml version="1.0" encoding="UTF-8"?><rss version="2.0"
	xmlns:content="http://purl.org/rss/1.0/modules/content/"
	xmlns:wfw="http://wellformedweb.org/CommentAPI/"
	xmlns:dc="http://purl.org/dc/elements/1.1/"
	xmlns:atom="http://www.w3.org/2005/Atom"
	xmlns:sy="http://purl.org/rss/1.0/modules/syndication/"
	xmlns:slash="http://purl.org/rss/1.0/modules/slash/"
	>

<channel>
	<title>IA - RiskInsight</title>
	<atom:link href="https://www.riskinsight-wavestone.com/en/tag/ia-2/feed/" rel="self" type="application/rss+xml" />
	<link>https://www.riskinsight-wavestone.com/en/tag/ia-2/</link>
	<description>The cybersecurity &#38; digital trust blog by Wavestone&#039;s consultants</description>
	<lastBuildDate>Wed, 11 Feb 2026 09:12:33 +0000</lastBuildDate>
	<language>en-US</language>
	<sy:updatePeriod>
	hourly	</sy:updatePeriod>
	<sy:updateFrequency>
	1	</sy:updateFrequency>
	

<image>
	<url>https://www.riskinsight-wavestone.com/wp-content/uploads/2024/02/Blogs-2024_RI-39x39.png</url>
	<title>IA - RiskInsight</title>
	<link>https://www.riskinsight-wavestone.com/en/tag/ia-2/</link>
	<width>32</width>
	<height>32</height>
</image> 
	<item>
		<title>GenAI Guardrails – Why do you need them &#038; Which one should you use?</title>
		<link>https://www.riskinsight-wavestone.com/en/2026/02/genai-guardrails-why-do-you-need-them-which-one-should-you-use/</link>
					<comments>https://www.riskinsight-wavestone.com/en/2026/02/genai-guardrails-why-do-you-need-them-which-one-should-you-use/#respond</comments>
		
		<dc:creator><![CDATA[Nicolas Lermusiaux]]></dc:creator>
		<pubDate>Wed, 11 Feb 2026 09:10:19 +0000</pubDate>
				<category><![CDATA[Ethical Hacking & Incident Response]]></category>
		<category><![CDATA[Focus]]></category>
		<category><![CDATA[AI]]></category>
		<category><![CDATA[AI Guardrails]]></category>
		<category><![CDATA[AI Red Teaming]]></category>
		<category><![CDATA[AI security]]></category>
		<category><![CDATA[AI vulnerabilities]]></category>
		<category><![CDATA[artificial intelligence]]></category>
		<category><![CDATA[Critères de selection]]></category>
		<category><![CDATA[cybersécurité]]></category>
		<category><![CDATA[cybersecurity]]></category>
		<category><![CDATA[Filtering]]></category>
		<category><![CDATA[Filtrage]]></category>
		<category><![CDATA[generative AI]]></category>
		<category><![CDATA[Guardrails]]></category>
		<category><![CDATA[Guardrails solutions]]></category>
		<category><![CDATA[IA]]></category>
		<category><![CDATA[prompt injection]]></category>
		<category><![CDATA[Selection criteria]]></category>
		<guid isPermaLink="false">https://www.riskinsight-wavestone.com/?p=28986</guid>

					<description><![CDATA[<p>The rise of generative AI and Large Language Models (LLMs) like ChatGPT has disrupted digital practices. More companies choose to deploy applications integrating these language models, but this integration comes with new vulnerabilities, identified by OWASP in its Top 10...</p>
<p>Cet article <a href="https://www.riskinsight-wavestone.com/en/2026/02/genai-guardrails-why-do-you-need-them-which-one-should-you-use/">GenAI Guardrails – Why do you need them &amp; Which one should you use?</a> est apparu en premier sur <a href="https://www.riskinsight-wavestone.com/en/">RiskInsight</a>.</p>
]]></description>
										<content:encoded><![CDATA[
<p style="text-align: justify;">The rise of generative AI and Large Language Models (LLMs) like ChatGPT has disrupted digital practices. More companies choose to deploy applications integrating these language models, but this integration comes with new vulnerabilities, identified by OWASP in its Top 10 LLM 2025 and Top 10 for Agentic Applications 2026. Faced with these new risks and new regulations like the AI Act, specialized solutions, named guardrails, have emerged to secure interactions (by analysing semantically all the prompts and responses) with LLMs and are becoming essential to ensure compliance and security for these applications.</p>
<p> </p>
<h2>The challenge of choosing a guardrails solution</h2>
<p style="text-align: justify;">As guardrails solutions multiply, organizations face a practical challenge: selecting protection mechanisms that effectively reduce risk without compromising performance, user experience, or operational feasibility.</p>
<p style="text-align: justify;">Choosing guardrails is not limited to blocking malicious prompts. It requires balancing detection accuracy, false positives, latency, and the ability to adapt filtering to the specific context, data sources, and threat exposure of each application. In practice, no single solution addresses all use cases equally well, making guardrail selection a contextual and risk-driven decision.</p>
<p> </p>
<h2>An important diversity of solutions</h2>
<figure id="attachment_28987" aria-describedby="caption-attachment-28987" style="width: 2560px" class="wp-caption aligncenter"><img fetchpriority="high" decoding="async" class="size-full wp-image-28987" src="https://www.riskinsight-wavestone.com/wp-content/uploads/2026/02/IMG1-Overview-of-guardrails-solutions-not-exhaustive-scaled.png" alt="Overview of guardrails solutions (not exhaustive)" width="2560" height="1576" srcset="https://www.riskinsight-wavestone.com/wp-content/uploads/2026/02/IMG1-Overview-of-guardrails-solutions-not-exhaustive-scaled.png 2560w, https://www.riskinsight-wavestone.com/wp-content/uploads/2026/02/IMG1-Overview-of-guardrails-solutions-not-exhaustive-310x191.png 310w, https://www.riskinsight-wavestone.com/wp-content/uploads/2026/02/IMG1-Overview-of-guardrails-solutions-not-exhaustive-63x39.png 63w, https://www.riskinsight-wavestone.com/wp-content/uploads/2026/02/IMG1-Overview-of-guardrails-solutions-not-exhaustive-768x473.png 768w, https://www.riskinsight-wavestone.com/wp-content/uploads/2026/02/IMG1-Overview-of-guardrails-solutions-not-exhaustive-1536x946.png 1536w, https://www.riskinsight-wavestone.com/wp-content/uploads/2026/02/IMG1-Overview-of-guardrails-solutions-not-exhaustive-2048x1261.png 2048w" sizes="(max-width: 2560px) 100vw, 2560px" /><figcaption id="caption-attachment-28987" class="wp-caption-text"><em>Overview of guardrails solutions (not exhaustive)</em></figcaption></figure>
<p> </p>
<p style="text-align: justify;">In 2025, the AI security and LLM guardrails landscape experienced significant consolidation. Major cybersecurity vendors increasingly sought to extend their portfolios with protections dedicated to generative AI, model usage, and agent interactions. Rather than building these capabilities from scratch, many chose to acquire specialized startups to rapidly integrate AI-native security features into their existing platforms, such as SentinelOne with Prompt Security or Check Point with Lakera.</p>
<p style="text-align: justify;">This trend illustrates a broader shift in the cybersecurity market: protections for LLM-based applications are becoming a standard component of enterprise security offerings, alongside more traditional controls. Guardrails and runtime AI protections are no longer niche solutions, but are progressively embedded into mainstream security stacks to support enterprise-scale AI adoption</p>
<p> </p>
<h2>The main criteria to choose your guardrails</h2>
<p style="text-align: justify;">With so many guardrails’ solutions, choosing the right option becomes a challenge. The most important criteria to focus on are:</p>
<ul>
<li style="text-align: justify;"><strong>Filtering effectiveness</strong>, to reduce exposure to malicious prompts while limiting false positives</li>
<li style="text-align: justify;"><strong>Latency</strong>, to ensure a user-friendly experience</li>
<li style="text-align: justify;"><strong>Personalisation capabilities</strong>, to adapt filtering to business-specific contexts and risks</li>
<li style="text-align: justify;"><strong>Operational cost</strong>, to support scalability over time</li>
</ul>
<p> </p>
<h2>Key Results &amp; Solutions Profiles</h2>
<p style="text-align: justify;">To get an idea of the performances the guardrails in the market, we tested several solutions across these criteria and a few profiles stood out:</p>
<ul>
<li style="text-align: justify;">Some solutions offer rapid deployment and effective baseline protection with minimal configuration, making them suitable for organizations seeking immediate risk reduction. These solutions typically perform well out of the box but provide limited customization.</li>
<li style="text-align: justify;">Other solutions emphasize flexibility and fine-grained control. While these frameworks enable advanced filtering strategies, they often exhibit poor default performance and require significant configuration effort to reach good protection levels.</li>
</ul>
<p style="text-align: justify;">As a result, selecting a guardrails solution depends less on raw detection scores and more on the expected level of customization, operational maturity, and acceptable setup effort.</p>
<p> </p>
<h2>Focus on Cloud Providers’ guardrails</h2>
<p style="text-align: justify;">As most LLM-based applications are deployed in cloud environments, native guardrails offered by cloud providers represent a pragmatic first layer of protection. These solutions are easy to activate, cost-effective, and integrate seamlessly into existing cloud workflows.</p>
<p style="text-align: justify;">Using automated red-teaming techniques, we observed that cloud-native guardrails consistently blocked most of the common prompt injection and jailbreak attempts. The overall performance of the guardrails available on Azure, AWS and GCP were similar, confirming their relevance as baseline protection mechanisms for production workloads.</p>
<p> </p>
<h3>Sensitivity Configuration</h3>
<p style="text-align: justify;">The configuration of several of the Cloud provider’s solutions allows us to set a sensitivity level to the guardrails configured in order to adapt the detection to the required level for the considered use-case.</p>
<figure id="attachment_28989" aria-describedby="caption-attachment-28989" style="width: 911px" class="wp-caption aligncenter"><img decoding="async" class="size-full wp-image-28989" src="https://www.riskinsight-wavestone.com/wp-content/uploads/2026/02/IMG2-AWS-Bedrock-Guardrails-configuration.png" alt="AWS Bedrock Guardrails configuration" width="911" height="343" srcset="https://www.riskinsight-wavestone.com/wp-content/uploads/2026/02/IMG2-AWS-Bedrock-Guardrails-configuration.png 911w, https://www.riskinsight-wavestone.com/wp-content/uploads/2026/02/IMG2-AWS-Bedrock-Guardrails-configuration-437x165.png 437w, https://www.riskinsight-wavestone.com/wp-content/uploads/2026/02/IMG2-AWS-Bedrock-Guardrails-configuration-71x27.png 71w, https://www.riskinsight-wavestone.com/wp-content/uploads/2026/02/IMG2-AWS-Bedrock-Guardrails-configuration-768x289.png 768w" sizes="(max-width: 911px) 100vw, 911px" /><figcaption id="caption-attachment-28989" class="wp-caption-text"><em>AWS Bedrock Guardrails configuration</em></figcaption></figure>
<p>        </p>
<h3>Customization</h3>
<p style="text-align: justify;">Beyond sensitivity tuning, fine-grained customization is essential for effective guardrails protections. Each application has specific filtering requirements, driven by business context, regulatory constraints, and threat exposure.</p>
<p style="text-align: justify;">Personalization is required at multiple levels:</p>
<ul style="text-align: justify;">
<li><strong>Business context</strong>: blocking application-specific forbidden topics, such as competitors, confidential projects, or regulated information</li>
<li><strong>Threat mitigation</strong>: adapting filters to address high-impact attacks, including indirect prompt injection</li>
<li><strong>Data flow awareness</strong>: within a single application, different data sources require different filtering strategies. User inputs, retrieved documents, and tool outputs should not be filtered identically.</li>
</ul>
<p style="text-align: justify;"> </p>
<p style="text-align: justify;">Applying uniform filtering across all inputs significantly limits effectiveness and may create blind spots. Guardrails must therefore be designed as part of the application architecture, not as a single monolithic filter.</p>
<figure id="attachment_28991" aria-describedby="caption-attachment-28991" style="width: 1675px" class="wp-caption aligncenter"><img decoding="async" class="size-full wp-image-28991" src="https://www.riskinsight-wavestone.com/wp-content/uploads/2026/02/IMG3-Guardrails-position-in-your-applications-infrastructure-1.png" alt="Guardrails position in your application's infrastructure" width="1675" height="735" srcset="https://www.riskinsight-wavestone.com/wp-content/uploads/2026/02/IMG3-Guardrails-position-in-your-applications-infrastructure-1.png 1675w, https://www.riskinsight-wavestone.com/wp-content/uploads/2026/02/IMG3-Guardrails-position-in-your-applications-infrastructure-1-435x191.png 435w, https://www.riskinsight-wavestone.com/wp-content/uploads/2026/02/IMG3-Guardrails-position-in-your-applications-infrastructure-1-71x31.png 71w, https://www.riskinsight-wavestone.com/wp-content/uploads/2026/02/IMG3-Guardrails-position-in-your-applications-infrastructure-1-768x337.png 768w, https://www.riskinsight-wavestone.com/wp-content/uploads/2026/02/IMG3-Guardrails-position-in-your-applications-infrastructure-1-1536x674.png 1536w" sizes="(max-width: 1675px) 100vw, 1675px" /><figcaption id="caption-attachment-28991" class="wp-caption-text"><em>Guardrails position in your application&#8217;s infrastructure</em></figcaption></figure>
<p> </p>
<h3>Key Insights</h3>
<p style="text-align: justify;">This study highlights several key insights:</p>
<ul style="text-align: justify;">
<li>No single guardrails solution fits all use cases, trade-offs exist between ease of deployment, performance, and customization</li>
<li>Cloud-native guardrails provide an effective and low-effort baseline for most cloud-hosted applications</li>
<li>Advanced use cases require configurable solutions capable of adapting filtering logic to application context and data flows</li>
</ul>
<p style="text-align: justify;">Guardrails should be selected based on risk exposure, operational maturity, and long-term maintainability rather than raw detection scores alone.</p>
<h2 style="text-align: justify;"> </h2>
<p style="text-align: justify;">Guardrails have become a necessary component of LLM-based applications, and a wide range of solutions is now available. Selecting the right guardrails requires identifying the solution that best aligns with an organization’s specific risks, constraints, and application architecture.</p>
<p style="text-align: justify;">Depending on your profile we have several suggestions for you:</p>
<ul style="text-align: justify;">
<li>If your application is already deployed in a cloud environment, using the guardrails provided by the cloud provider is a good solution.</li>
<li>If you want better control over the filtering solution, deploying one of the open-source guardrails solutions may be the most suitable option.</li>
<li>You want the best and have the capacity, you can issue an RFI or RFP to compare different solutions and select the most tailored to your needs.</li>
</ul>
<p style="text-align: justify;">Finally, guardrails alone are not sufficient to protect your applications. Secure LLM applications also rely on properly configured tools, strict IAM policies, and robust security architecture to prevent more severe exploitation scenarios.</p>
<p> </p>
<p> </p>
<p> </p>
<p> </p>
<p> </p>


<p>Cet article <a href="https://www.riskinsight-wavestone.com/en/2026/02/genai-guardrails-why-do-you-need-them-which-one-should-you-use/">GenAI Guardrails – Why do you need them &amp; Which one should you use?</a> est apparu en premier sur <a href="https://www.riskinsight-wavestone.com/en/">RiskInsight</a>.</p>
]]></content:encoded>
					
					<wfw:commentRss>https://www.riskinsight-wavestone.com/en/2026/02/genai-guardrails-why-do-you-need-them-which-one-should-you-use/feed/</wfw:commentRss>
			<slash:comments>0</slash:comments>
		
		
			</item>
		<item>
		<title>Cybersecurity Startups Radar: 2025, AI at the service of cybersecurity</title>
		<link>https://www.riskinsight-wavestone.com/en/2025/06/cybersecurity-startups-radar-2025-ai-at-the-service-of-cybersecurity/</link>
					<comments>https://www.riskinsight-wavestone.com/en/2025/06/cybersecurity-startups-radar-2025-ai-at-the-service-of-cybersecurity/#respond</comments>
		
		<dc:creator><![CDATA[Ahmed Amine Ghariani]]></dc:creator>
		<pubDate>Fri, 20 Jun 2025 14:17:53 +0000</pubDate>
				<category><![CDATA[Cyberrisk Management & Strategy]]></category>
		<category><![CDATA[Cybersecurity & Digital Trust]]></category>
		<category><![CDATA[Deep-dive]]></category>
		<category><![CDATA[artificial intelligence]]></category>
		<category><![CDATA[cybersecurity]]></category>
		<category><![CDATA[Cybersecurity startups radar]]></category>
		<category><![CDATA[IA]]></category>
		<category><![CDATA[radar 2025]]></category>
		<category><![CDATA[Swiss radar]]></category>
		<category><![CDATA[Switzerland]]></category>
		<guid isPermaLink="false">https://www.riskinsight-wavestone.com/?p=26402</guid>

					<description><![CDATA[<p>AI at the service of cybersecurity: a concrete step forward   Every year since 2020, Wavestone has identified Swiss cybersecurity startups in its eponymous radar. While AI has established itself as a cross-disciplinary subject in all fields, the 2025 Radar...</p>
<p>Cet article <a href="https://www.riskinsight-wavestone.com/en/2025/06/cybersecurity-startups-radar-2025-ai-at-the-service-of-cybersecurity/">Cybersecurity Startups Radar: 2025, AI at the service of cybersecurity</a> est apparu en premier sur <a href="https://www.riskinsight-wavestone.com/en/">RiskInsight</a>.</p>
]]></description>
										<content:encoded><![CDATA[
<h2 style="text-align: justify;">AI at the service of cybersecurity: a concrete step forward</h2>
<p> </p>
<p style="text-align: justify;">Every year since 2020, Wavestone has identified Swiss cybersecurity startups in its eponymous radar. While AI has established itself as a cross-disciplinary subject in all fields, the 2025 Radar focuses on the <strong>use of artificial intelligence as a tool</strong>, not just as a subject to be secured, but as a <strong>technology at the very heart of the cyber response.</strong></p>
<p style="text-align: justify;">Several startups are using AI to automate, enhance or personalize their solutions:</p>
<p style="text-align: justify;"><strong>Egonym</strong> uses generative AI to anonymize faces in images and videos while preserving useful traits like age and emotion — striking a rare balance between privacy and utility.</p>
<p style="text-align: justify;"><img loading="lazy" decoding="async" class="aligncenter size-full wp-image-26393" src="https://www.riskinsight-wavestone.com/wp-content/uploads/2025/06/Image16.jpg" alt="egonym logo" width="155" height="53" srcset="https://www.riskinsight-wavestone.com/wp-content/uploads/2025/06/Image16.jpg 155w, https://www.riskinsight-wavestone.com/wp-content/uploads/2025/06/Image16-71x24.jpg 71w" sizes="auto, (max-width: 155px) 100vw, 155px" /></p>
<p style="text-align: justify;"><strong>Hafnova</strong> applies real-time AI to detect, block, and report threats across critical infrastructures with high responsiveness and minimal delay.</p>
<p style="text-align: justify;"><img loading="lazy" decoding="async" class="aligncenter size-full wp-image-26369" src="https://www.riskinsight-wavestone.com/wp-content/uploads/2025/06/Image2.png" alt="Hafnova logo" width="67" height="71" srcset="https://www.riskinsight-wavestone.com/wp-content/uploads/2025/06/Image2.png 67w, https://www.riskinsight-wavestone.com/wp-content/uploads/2025/06/Image2-37x39.png 37w" sizes="auto, (max-width: 67px) 100vw, 67px" /></p>
<p style="text-align: justify;"><strong>Aurigin</strong> combats deepfake-based fraud in real time using multimodal AI that simultaneously analyzes voice, image, and text to validate identities. </p>
<p style="text-align: justify;"><img loading="lazy" decoding="async" class="aligncenter size-full wp-image-26395" src="https://www.riskinsight-wavestone.com/wp-content/uploads/2025/06/Image17.jpg" alt="aurigin.ai logo" width="176" height="51" srcset="https://www.riskinsight-wavestone.com/wp-content/uploads/2025/06/Image17.jpg 176w, https://www.riskinsight-wavestone.com/wp-content/uploads/2025/06/Image17-71x21.jpg 71w" sizes="auto, (max-width: 176px) 100vw, 176px" /></p>
<p style="text-align: justify;"><strong>RedCarbon</strong> delivers autonomous AI agents capable of handling complex cybersecurity tasks such as threat detection, hunting, and compliance monitoring — significantly reducing analyst workload.</p>
<p style="text-align: justify;"><img loading="lazy" decoding="async" class="aligncenter wp-image-26375" src="https://www.riskinsight-wavestone.com/wp-content/uploads/2025/06/Image6-e1750428148421.png" alt="REDCARBON logo" width="183" height="24" /></p>
<p style="text-align: justify;"><strong>Baited</strong> leverages AI and OSINT to generate hyper-realistic phishing simulations, enabling organizations to test and train employees under real-world conditions.</p>
<p style="text-align: justify;"><img loading="lazy" decoding="async" class="aligncenter wp-image-26373" src="https://www.riskinsight-wavestone.com/wp-content/uploads/2025/06/Image5-e1750428603602.png" alt="baited logo" width="142" height="45" /></p>
<p style="text-align: justify;">It&#8217;s good to see AI becoming an essential defensive weapon contributing to the defense of our information systems.</p>
<p style="text-align: justify;"> </p>
<h2 style="text-align: justify;">Strong momentum around threat detection, response and monitoring</h2>
<p> </p>
<p style="text-align: justify;">The second strong trend this year is the emergence or reinforcement of startups specializing in <strong>intrusion detection, suspicious behavior detection, incident response</strong> and <strong>continuous supervision.</strong></p>
<p style="text-align: justify;">This segment, already well established historically, is undoubtedly gaining strength with several new entries:</p>
<p style="text-align: justify;"><strong>RedCarbon</strong>: AI agents for threat detection &amp; automated hunting.</p>
<p style="text-align: justify;"><img loading="lazy" decoding="async" class="aligncenter wp-image-26375" src="https://www.riskinsight-wavestone.com/wp-content/uploads/2025/06/Image6-e1750428148421.png" alt="REDCARBON logo" width="175" height="23" /></p>
<p style="text-align: justify;"><strong>Swiss Security Hub</strong>: continuous monitoring of SAP systems with XDR integration.</p>
<p style="text-align: justify;"><img loading="lazy" decoding="async" class="aligncenter wp-image-26377" src="https://www.riskinsight-wavestone.com/wp-content/uploads/2025/06/Image7-e1750428038244.png" alt="Security Hub logo" width="115" height="115" /></p>
<p style="text-align: justify;"><strong>Cyberservices</strong> : XDR platform based on the Google ecosystem.</p>
<p style="text-align: justify;"><img loading="lazy" decoding="async" class="aligncenter  wp-image-26379" src="https://www.riskinsight-wavestone.com/wp-content/uploads/2025/06/Image8-e1750428222485.png" alt="Cyberservices logo" width="64" height="35" /></p>
<p style="text-align: justify;"><strong>Hafnova</strong>: real-time cyber supervision in critical sectors.</p>
<p style="text-align: justify;"><img loading="lazy" decoding="async" class="aligncenter size-full wp-image-26369" src="https://www.riskinsight-wavestone.com/wp-content/uploads/2025/06/Image2.png" alt="Hafnova logo" width="67" height="71" srcset="https://www.riskinsight-wavestone.com/wp-content/uploads/2025/06/Image2.png 67w, https://www.riskinsight-wavestone.com/wp-content/uploads/2025/06/Image2-37x39.png 37w" sizes="auto, (max-width: 67px) 100vw, 67px" /></p>
<p style="text-align: justify;"><strong>Tirreno</strong>: on-prem platform for online fraud detection with user trust scoring.</p>
<p style="text-align: justify;"><img loading="lazy" decoding="async" class="aligncenter  wp-image-26383" src="https://www.riskinsight-wavestone.com/wp-content/uploads/2025/06/Image10-e1750428239189.png" alt="tirreno logo" width="131" height="29" /></p>
<p style="text-align: justify;"> </p>
<p style="text-align: justify;">At a time when cyber-attacks continue to increase in number and complexity, <strong>preventive, contextualized and autonomous detection</strong> is and will remain key to strengthening operational resilience.</p>
<p style="text-align: justify;"><strong> </strong></p>
<h2 style="text-align: justify;">New ground explored: digital sovereignty and secure hardware</h2>
<p> </p>
<p style="text-align: justify;">Among the notable additions, <strong>The Cosmic Dolphins</strong> stands out with its sovereign hardware approach:</p>
<p style="text-align: justify;"><strong>The Cosmic Dolphins</strong>: Swiss smartphones with dual-zone OS (Shark Zone / Dolphin Zone), kill switch, and hardware-first approach to privacy.</p>
<p style="text-align: justify;"><img loading="lazy" decoding="async" class="aligncenter size-full wp-image-26385" src="https://www.riskinsight-wavestone.com/wp-content/uploads/2025/06/Image11-e1750428264682.jpg" alt="The Cosmic Dolphins logo" width="99" height="99" /></p>
<p style="text-align: justify;">Swiss innovation isn&#8217;t limited to software: <strong>mastery of the physical infrastructure</strong> is becoming an issue of trust, sovereignty and differentiation.</p>
<p style="text-align: justify;"><img loading="lazy" decoding="async" class="aligncenter size-full wp-image-26387" src="https://www.riskinsight-wavestone.com/wp-content/uploads/2025/06/Image12.jpg" alt="Cybersecurity Startups Radar: 2025, AI at the service of cybersecurity" width="602" height="338" srcset="https://www.riskinsight-wavestone.com/wp-content/uploads/2025/06/Image12.jpg 602w, https://www.riskinsight-wavestone.com/wp-content/uploads/2025/06/Image12-340x191.jpg 340w, https://www.riskinsight-wavestone.com/wp-content/uploads/2025/06/Image12-69x39.jpg 69w, https://www.riskinsight-wavestone.com/wp-content/uploads/2025/06/Image12-600x338.jpg 600w" sizes="auto, (max-width: 602px) 100vw, 602px" /></p>
<h2 style="text-align: justify;">Key Figures</h2>
<h2 style="text-align: justify;" data-wp-editing="1"><img loading="lazy" decoding="async" class="aligncenter wp-image-26406 size-full" src="https://www.riskinsight-wavestone.com/wp-content/uploads/2025/06/Image18-e1750429011414.png" alt="" width="200" height="39" /><br /><img loading="lazy" decoding="async" class="aligncenter wp-image-26408 size-full" src="https://www.riskinsight-wavestone.com/wp-content/uploads/2025/06/Image19-e1750428988444.png" alt="" width="200" height="67" /></h2>
<p> </p>
<h2 style="text-align: justify;">Geographical focus: undisputed predominance of Lausanne and Zurich, but other regions are gaining ground</h2>
<p> </p>
<p style="text-align: justify;">Unsurprisingly, most startups are located around two main technological clusters: Zürich and Lausanne. This confirms an already existing trend since these two cities are hosting Swiss Federal institutes of technology (ETHZ in Zürich, EPFL in Lausanne).</p>
<p style="text-align: justify;">These universities are providing a fertile ground for startups as they offer support in terms of infrastructure but also in terms of collaboration with students and labs. In return, intellectual property is shared between startups and universities. This model is a success for Switzerland as it allows to continuously improve the economy of these regions with a good balance between investment and research.</p>
<p style="text-align: justify;">Nevertheless, other regions such as Geneva and Ticino are showing increasing dynamism, with several new startups emerging in this year’s edition. This points to a gradually diversifying ecosystem, supported by regional initiatives like innovation hubs and dedicated startup incubators.</p>
<p style="text-align: justify;"> </p>
<h2 style="text-align: justify;">Methodology</h2>
<p> </p>
<p style="text-align: justify;">Wavestone’s Swiss Cybersecurity Startups Radar identifies new players in the Swiss cyber innovation ecosystem. Its objective: to provide a global and critical view of an ever-renewing environment.</p>
<ul>
<li style="text-align: justify;">Startups were selected according to our eligibility criteria:</li>
<li style="text-align: justify;">Head office in Switzerland</li>
<li style="text-align: justify;">Less than 50 employees</li>
<li style="text-align: justify;">Less than 8 years of activity (established as of 2017)</li>
<li style="text-align: justify;">Business model around a specific product (software or hardware)</li>
<li style="text-align: justify;">Startups were identified and evaluated according to the following procedure:</li>
<li style="text-align: justify;">Open Source Intelligence (OSINT) data consolidation</li>
<li style="text-align: justify;">Evaluation in regard to above criteria</li>
<li style="text-align: justify;">Qualitative interviews with the startups</li>
</ul>
<p>Cet article <a href="https://www.riskinsight-wavestone.com/en/2025/06/cybersecurity-startups-radar-2025-ai-at-the-service-of-cybersecurity/">Cybersecurity Startups Radar: 2025, AI at the service of cybersecurity</a> est apparu en premier sur <a href="https://www.riskinsight-wavestone.com/en/">RiskInsight</a>.</p>
]]></content:encoded>
					
					<wfw:commentRss>https://www.riskinsight-wavestone.com/en/2025/06/cybersecurity-startups-radar-2025-ai-at-the-service-of-cybersecurity/feed/</wfw:commentRss>
			<slash:comments>0</slash:comments>
		
		
			</item>
		<item>
		<title>Red Teaming IA : State of play of AI risks in 2025</title>
		<link>https://www.riskinsight-wavestone.com/en/2025/04/red-teaming-ia-state-of-play-of-ai-risks-in-2025/</link>
					<comments>https://www.riskinsight-wavestone.com/en/2025/04/red-teaming-ia-state-of-play-of-ai-risks-in-2025/#respond</comments>
		
		<dc:creator><![CDATA[Basma Benali]]></dc:creator>
		<pubDate>Tue, 15 Apr 2025 13:00:00 +0000</pubDate>
				<category><![CDATA[Cloud & Next-Gen IT Security]]></category>
		<category><![CDATA[artificial intelligence]]></category>
		<category><![CDATA[cybersecurity]]></category>
		<category><![CDATA[genai]]></category>
		<category><![CDATA[IA]]></category>
		<category><![CDATA[IA Generative]]></category>
		<category><![CDATA[LLM]]></category>
		<category><![CDATA[mlops]]></category>
		<category><![CDATA[Red Teaming]]></category>
		<guid isPermaLink="false">https://www.riskinsight-wavestone.com/?p=25767</guid>

					<description><![CDATA[<p>Generative AI systems are fallible: in March 2025, a ChatGPT vulnerability was widely exploited to trap its users; a few months earlier, Microsoft&#8217;s health chatbot exposed sensitive data; in December, a simple prompt injection allowed the takeover of a user...</p>
<p>Cet article <a href="https://www.riskinsight-wavestone.com/en/2025/04/red-teaming-ia-state-of-play-of-ai-risks-in-2025/">Red Teaming IA : State of play of AI risks in 2025</a> est apparu en premier sur <a href="https://www.riskinsight-wavestone.com/en/">RiskInsight</a>.</p>
]]></description>
										<content:encoded><![CDATA[
<p style="text-align: justify;">Generative AI systems are fallible: in March 2025, a ChatGPT vulnerability was widely exploited to trap its users; a few months earlier, Microsoft&#8217;s health chatbot exposed sensitive data; in December, a simple prompt injection allowed the takeover of a user account on the competing service DeepSeek.</p>
<p style="text-align: justify;">Today, the impacts are limited because the latitude given to AI systems is still relatively low. Tomorrow, with the rise of agentic AI, accelerated adoption of generative AI, and the multiplication of use cases, the impacts will grow. Just as the ransomware WannaCry exploited vulnerabilities on a massive scale in 2017, major cyberattacks are likely to target AI systems and could result in injuries or financial bankruptcies.</p>
<p style="text-align: justify;">These risks can be anticipated. One of the most pragmatic ways to do this is to take on the role of a malicious individual and attempt to manipulate an AI system to study its robustness. This approach highlights system vulnerabilities and how to fix them. Specifically for generative AI, this discipline is called AI RedTeaming. In this article, we offer insight into its contours, focusing particularly on field feedback regarding the main vulnerabilities encountered.</p>
<p style="text-align: justify;">To stay aligned with the market practices, this article exclusively focuses on the RedTeaming of generative AI systems.</p>
<p style="text-align: justify;"><em> </em></p>
<h2 style="text-align: justify;"><!--StartFragment --><span class="cf0">Back to basics, how does genAI work</span> ?</h2>
<p> </p>
<p style="text-align: justify;">GenAI relies on components that are often distributed between cloud and on-premise environments. Generally, the more functionalities a generative AI system offers (searching for information, launching actions, executing code, etc.), the more components it includes. From a cybersecurity perspective, this exposes the system to multiple risks :</p>
<p><img loading="lazy" decoding="async" class="wp-image-25779 size-full" src="https://www.riskinsight-wavestone.com/wp-content/uploads/2025/04/Diapositive2.png" alt="Underlying infrastructure / GenAI hosting platforms" width="1280" height="720" srcset="https://www.riskinsight-wavestone.com/wp-content/uploads/2025/04/Diapositive2.png 1280w, https://www.riskinsight-wavestone.com/wp-content/uploads/2025/04/Diapositive2-340x191.png 340w, https://www.riskinsight-wavestone.com/wp-content/uploads/2025/04/Diapositive2-69x39.png 69w, https://www.riskinsight-wavestone.com/wp-content/uploads/2025/04/Diapositive2-768x432.png 768w, https://www.riskinsight-wavestone.com/wp-content/uploads/2025/04/Diapositive2-800x450.png 800w" sizes="auto, (max-width: 1280px) 100vw, 1280px" /></p>
<p style="text-align: center;"><em>Diagram of a Generative AI System and Issues Raised by Component</em></p>
<p> </p>
<p style="text-align: justify;">In general, an attacker only has access to a web interface through which they can interact (click, enter text into fields, etc.). From there, they can:</p>
<ul>
<li>Conduct classic cybersecurity attacks (inserting malicious scripts – XSS, etc.) by exploiting vulnerabilities in the AI system’s components;</li>
<li>Perform a new type of attack by writing in natural language to exploit the functionalities provided by the generative AI system behind the web interface: data exfiltration, executing malicious actions using the privileges of the generative AI system, etc.</li>
</ul>
<p style="text-align: justify;">Technically, each component is protected by implementing security measures defined by Security Integration Processes within Projects. It is then useful to practically assess the effective level of security through an AI RedTeam audit.</p>
<p style="text-align: justify;"> </p>
<h2 style="text-align: justify;">RedTeaming IA, Art of findings AI vulnerabilities</h2>
<p> </p>
<p style="text-align: justify;">AI RedTeam audits are similar to traditional security audits. However, to address the new challenges of GenAI, they rely on specific methodologies, frameworks, and tools. Indeed, during an AI RedTeam audit, the goal is to bypass the generative AI system by either attacking its components or crafting malicious instructions in natural language. This second type of attack is called prompt injection, the art of formulating malicious queries to an AI system to divert its functionalities.</p>
<p style="text-align: justify;">During an AI RedTeam audit, two types of tests in natural language attacks (specific to AI) are conducted simultaneously:</p>
<ul>
<li>Manual tests. These allow a reconnaissance phase using libraries of malicious questions consolidated beforehand.</li>
<li>Automated tests. These usually involve a generative AI attacking the target generative AI system by generating a series of malicious prompts and automatically analyzing the coherence of the chatbot&#8217;s responses. They help assess the system&#8217;s robustness across a wide range of scenarios.</li>
</ul>
<p style="text-align: justify;">These tests typically identify several vulnerabilities and highlight cybersecurity risks that are often underestimated.</p>
<p style="text-align: justify;"> </p>
<h2 style="text-align: justify;">What are the main vulnerabilities we found ?</h2>
<p> </p>
<p style="text-align: justify;">We have covered three main deployment categories with our clients:</p>
<ol>
<li>Simple chatbot : these solutions are primarily used for redirecting and sorting user requests;</li>
<li>RAG (Retrieval-Augmented Generation) chatbot : these more sophisticated systems consult internal document databases to enrich their responses;</li>
<li>Agentic chatbot : these advanced solutions can interact with other systems and execute actions.</li>
</ol>
<p style="text-align: justify;">The consolidation of vulnerabilities identified during our interventions, as well as their relative criticality, allows us to define the following ranking:</p>
<p style="text-align: justify;"><img loading="lazy" decoding="async" class="aligncenter wp-image-25775 size-full" src="https://www.riskinsight-wavestone.com/wp-content/uploads/2025/04/Diapositive4.png" alt="Vulnerabilités LLM &amp; Chatbots Notre Top 6 2025" width="1280" height="720" srcset="https://www.riskinsight-wavestone.com/wp-content/uploads/2025/04/Diapositive4.png 1280w, https://www.riskinsight-wavestone.com/wp-content/uploads/2025/04/Diapositive4-340x191.png 340w, https://www.riskinsight-wavestone.com/wp-content/uploads/2025/04/Diapositive4-69x39.png 69w, https://www.riskinsight-wavestone.com/wp-content/uploads/2025/04/Diapositive4-768x432.png 768w, https://www.riskinsight-wavestone.com/wp-content/uploads/2025/04/Diapositive4-800x450.png 800w" sizes="auto, (max-width: 1280px) 100vw, 1280px" /></p>
<h3 style="text-align: justify;"><strong>Diversion of the model and generation of illegitimate content </strong></h3>
<p style="text-align: justify;">This concerns the circumvention of the technical safeguards put in place during the development of the chatbot in order to generate offensive, malicious, or inappropriate content. Thus, the credibility and reputation of the company are at risk of being impacted since it is responsible for the content produced by its chatbot. </p>
<p style="text-align: justify;">It is worth noting that the circumvention of the model&#8217;s security mechanisms can lead to a complete unlocking. This is referred to as a jailbreak of the model, which shifts it into an unrestricted mode. In this state, it can produce content outside the framework desired by the company.</p>
<h3 style="text-align: justify;"><strong>Access to the preprompt</strong></h3>
<p style="text-align: justify;">The term preprompt refers to the set of instructions that feed the model and shape it for the desired use. All models are instructed not to disclose this preprompt in any form. </p>
<p style="text-align: justify;">An attacker gaining access to this preprompt has their attack facilitated, as it allows them to map the capabilities of the chatbot model. This mapping is particularly useful for complex systems interfaced with APIs or other external systems. Furthermore, access to this preprompt by an attacker enables them to visualize how the filters and limitations of the chatbot have been implemented, which allows them to bypass them more easily.</p>
<h3 style="text-align: justify;"><strong>Web integration and third-party integration</strong></h3>
<p style="text-align: justify;">GenAI solutions are often presented to users through a web interface. AI RedTeaming activities regularly highlight classic issues of web applications, particularly the isolation of user sessions or attacks aimed at trapping them. In the case of agentic systems, these vulnerabilities can also affect third-party components interconnected with the GenAI system.</p>
<h3 style="text-align: justify;"><strong>Sensitive data leaks</strong></h3>
<p style="text-align: justify;">If the data feeding the internal knowledge base of a RAG chatbot is insufficiently consolidated (selection, management, anonymization, &#8230;), the models may inadvertently reveal sensitive or confidential information. </p>
<p style="text-align: justify;">This issue is related to aspects of rights management, data classification, and hardening the data preparation and transit pipelines (MLOps).</p>
<h3 style="text-align: justify;"><strong>Stored injection</strong></h3>
<p style="text-align: justify;">In the case of stored injection, the attacker is able to feed the knowledge base of a model by including malicious instructions (via a compromised document). This knowledge base is used for the chatbot&#8217;s responses, so any user interacting with the model and requesting the said document will have their session compromised (leak of users&#8217; conversation history data, malicious redirections, participation in a social engineering attack, etc.). </p>
<p style="text-align: justify;">Compromised documents may be particularly difficult to identify, especially in the case of large or poorly managed knowledge bases. This attack is thus persistent and stealthy.</p>
<h3 style="text-align: justify;"><strong>Mention honorable: parasitism and cost explosion</strong></h3>
<p style="text-align: justify;">We talk about parasitism when a user is able to unlock the chatbot to fully utilize the model&#8217;s capabilities and do so for free. Coupled with a lack of volumetric restrictions, a user can make a prohibitive number of requests, unrelated to the initial use case, and still be charged for them.</p>
<p style="text-align: justify;">In general, some of the mentioned vulnerabilities concern relatively minor risks, whose business impact on information systems (IS) is limited. Nevertheless, with advances in AI technologies, these vulnerabilities take on a different dimension, particularly in the following cases:</p>
<ul>
<li>Agentic solutions with access to sensitive systems</li>
<li>RAG applications involving confidential data</li>
<li>Systems for which users have control over the knowledge base documents, opening the door to stored injections</li>
</ul>
<p style="text-align: justify;"><strong>The tested GenAI systems are largely unlockable, although the exercise becomes more complex over time. This persistent inability of the models to implement effective restrictions encourages the AI ecosystem to turn to external security components.</strong></p>
<p style="text-align: justify;"><strong> </strong></p>
<h2 style="text-align: justify;">What are the new attack surfaces ?</h2>
<p> </p>
<p style="text-align: justify;">The increasing integration of AI into sensitive sectors (healthcare, finance, defense, &#8230;) expands the attack surfaces of critical systems, which reinforces the need for filtering and anonymization of sensitive data. Where AI applications were previously very compartmentalized, agentic AI puts an end to this compartmentalization as it deploys a capacity for interconnection, opening the door to potential threat propagation within information systems. </p>
<p style="text-align: justify;">The decrease in the technical level required to create an AI system, particularly through the use of SaaS platforms and Low/no code services, facilitates its use for both legitimate users and attackers. </p>
<p style="text-align: justify;">Finally, the widespread adoption of &#8220;co-pilots&#8221; directly on employees&#8217; workstations results in an increasing use of increasingly autonomous components that act in place of and with the privileges of a human, accelerating the emergence of uncontrolled AI perimeters or Shadow IT AI. </p>
<p> </p>
<h2 style="text-align: justify;">Towards increasingly difficult-to-control systems</h2>
<p> </p>
<p style="text-align: justify;">Although appearing to imitate human intelligence, GenAI models (LLMs, or Large Language Models) have the sole function of mimicking language and often act as highly efficient text auto-completion systems. These systems are not natively trained to reason, and their use encounters a &#8220;black box&#8221; operation. It is indeed complex to reliably explain their reasoning, which regularly results in hallucinations in their outputs or logical fallacies. In practice, it is also impossible to prove the absence of &#8220;backdoors&#8221; in these models, further limiting our trust in these systems. </p>
<p style="text-align: justify;">The emergence of agentic AI complicates the situation. By interconnecting systems with opaque functioning, it renders the entire reasoning process generally unverifiable and inexplicable. Cases of models training, auditing, or attacking other models are becoming widespread, leading to a major trust issue when they are integrated into corporate information systems.</p>
<p style="text-align: justify;"> </p>
<h2>What are the perspectives for the future ?</h2>
<p> </p>
<p style="text-align: justify;">The RedTeaming AI audits conducted on generative AI systems reveal a contrasting reality. On one hand, innovation is rapid, driven by increasingly powerful and integrated use cases. On the other hand, the identified vulnerabilities demonstrate that these systems, often perceived as intelligent, remain largely manipulable, unstable, and poorly explainable. </p>
<p style="text-align: justify;">This observation is part of a broader context of the democratization of AI tools coupled with their increasing autonomy. Agentic AI, in particular, reveals chains of action that are difficult to trace, acting with human privileges. In such a landscape, the risk is no longer solely technical: it also becomes organizational and strategic, involving continuous governance and oversight of its uses. </p>
<p style="text-align: justify;">In the face of these challenges, RedTeaming AI emerges as an essential lever to anticipate possible deviations, adopting the attacker’s perspective to better prevent drifts. It involves testing the limits of a system to design robust, sustainable protection mechanisms that align with new uses. Only by doing so can generative AI continue to evolve within a framework of trust, serving both users and organizations. </p>
<p>Cet article <a href="https://www.riskinsight-wavestone.com/en/2025/04/red-teaming-ia-state-of-play-of-ai-risks-in-2025/">Red Teaming IA : State of play of AI risks in 2025</a> est apparu en premier sur <a href="https://www.riskinsight-wavestone.com/en/">RiskInsight</a>.</p>
]]></content:encoded>
					
					<wfw:commentRss>https://www.riskinsight-wavestone.com/en/2025/04/red-teaming-ia-state-of-play-of-ai-risks-in-2025/feed/</wfw:commentRss>
			<slash:comments>0</slash:comments>
		
		
			</item>
		<item>
		<title>US Executive Order &#038; Betchley Declaration</title>
		<link>https://www.riskinsight-wavestone.com/en/2024/05/us-executive-order-betchley-declaration/</link>
					<comments>https://www.riskinsight-wavestone.com/en/2024/05/us-executive-order-betchley-declaration/#respond</comments>
		
		<dc:creator><![CDATA[Amélie Grangien]]></dc:creator>
		<pubDate>Fri, 03 May 2024 08:49:27 +0000</pubDate>
				<category><![CDATA[Cloud & Next-Gen IT Security]]></category>
		<category><![CDATA[Focus]]></category>
		<category><![CDATA[artificial intelligence]]></category>
		<category><![CDATA[IA]]></category>
		<category><![CDATA[regulation]]></category>
		<category><![CDATA[UK]]></category>
		<category><![CDATA[US]]></category>
		<guid isPermaLink="false">https://www.riskinsight-wavestone.com/?p=23097</guid>

					<description><![CDATA[<p>In the evolving landscape of AI governance and regulation, recent efforts have shifted from scattered and reactive measures to cohesive policy frameworks that foster innovation while safeguarding against potential misuse. As AI becomes more integrated into our daily life, both...</p>
<p>Cet article <a href="https://www.riskinsight-wavestone.com/en/2024/05/us-executive-order-betchley-declaration/">US Executive Order &#038; Betchley Declaration</a> est apparu en premier sur <a href="https://www.riskinsight-wavestone.com/en/">RiskInsight</a>.</p>
]]></description>
										<content:encoded><![CDATA[
<p style="text-align: justify;">In the evolving landscape of AI governance and regulation, recent efforts have shifted from scattered and reactive measures to cohesive policy frameworks that foster innovation while safeguarding against potential misuse. <br />As AI becomes more integrated into our daily life, both public and private sectors have raised ethical concerns around issues of privacy, bias, accountability, and transparency.</p>
<p style="text-align: justify;"><img loading="lazy" decoding="async" class="aligncenter size-full wp-image-23098" src="https://www.riskinsight-wavestone.com/wp-content/uploads/2024/05/article-pierre-photo1-FR-2.png" alt="" width="723" height="471" srcset="https://www.riskinsight-wavestone.com/wp-content/uploads/2024/05/article-pierre-photo1-FR-2.png 723w, https://www.riskinsight-wavestone.com/wp-content/uploads/2024/05/article-pierre-photo1-FR-2-293x191.png 293w, https://www.riskinsight-wavestone.com/wp-content/uploads/2024/05/article-pierre-photo1-FR-2-60x39.png 60w" sizes="auto, (max-width: 723px) 100vw, 723px" /></p>
<figure id="attachment_23071" aria-describedby="caption-attachment-23071" style="width: 594px" class="wp-caption aligncenter"><img loading="lazy" decoding="async" class="wp-image-23071 size-full" src="https://www.riskinsight-wavestone.com/wp-content/uploads/2024/05/article-pierre-photo2-FR.png" alt="" width="594" height="421" srcset="https://www.riskinsight-wavestone.com/wp-content/uploads/2024/05/article-pierre-photo2-FR.png 594w, https://www.riskinsight-wavestone.com/wp-content/uploads/2024/05/article-pierre-photo2-FR-269x191.png 269w, https://www.riskinsight-wavestone.com/wp-content/uploads/2024/05/article-pierre-photo2-FR-55x39.png 55w" sizes="auto, (max-width: 594px) 100vw, 594px" /><figcaption id="caption-attachment-23071" class="wp-caption-text"><a href="https://ourworldindata.org/artificial-intelligence"><em>Source: https://ourworldindata.org/artificial-intelligence</em></a></figcaption></figure>
<p style="text-align: justify;"><br />Today, as governments actively craft AI guidance and legislation, policymakers face the challenge of delicately balancing the need to foster innovation and ensuring accountability. A regulatory framework that prioritizes innovation but relies too heavily on the private sector&#8217;s self-governance could lead to a lack of oversight and accountability. Conversely, while robust safeguards are essential to mitigate potential risks, an overly restrictive approach may stifle technological progress.<br />This whitepaper will explore the approaches proposed by the governments of the United States and the United Kingdom as they pertain to AI governance across both the public and private sectors.</p>
<h2 style="text-align: justify;"><br />American Approach to AI Regulation</h2>
<p style="text-align: justify;">In October of 2023, the White House published the AI Executive Order. The order specifies key near-term priorities of introducing reporting requirements for AI developers exceeding computing thresholds, launching research initiatives, developing frameworks for responsible AI use, and establishing AI governance within the federal government. Longer-term efforts focus on international cooperation, global standards, and AI safety.<br />On the side of ensuring accountability, the order calls for the Secretary of Commerce to enforce reporting provisions for companies developing dual-use AI foundation models, organizations acquiring large-scale computing clusters, and Infrastructure as a Service providers enabling foreign entities to conduct certain AI model training. While these criteria will likely exempt most small to medium sized AI companies from immediate regulations, large industry players like Open AI, Anthropic, and Meta could be affected if they surpass the computing threshold established by the order. <br />On the other side of fostering innovation, further sections of the order reaffirm the US government’s aim to promote AI innovation and competition – supporting R&amp;D initiatives and public-private partnerships, provisioning streamlined visa processes to attract AI talent to the US, prioritizing AI-oriented recruitment within the federal government, clarifying IP issues related to AI, and preventing unlawful collusion. <br />Overall, the nature of the documents published by the US is mostly non-binding, indicating a strategy of encouraging the private sector to self-regulate and align to common AI best practices. In this approach, the White House has been persistent in its messaging that it is committed to nurturing innovation, research, and leadership in the domain, while also balancing with the need for a secure and responsible AI ecosystem.</p>
<h2 style="text-align: justify;"><br />The British Approach to AI Regulation</h2>
<p style="text-align: justify;">The Bletchley Declaration, agreed upon during the AI Safety Summit 2023 held at Bletchley Park, Buckinghamshire, marks a pioneering international effort towards ensuring the safe and responsible development of AI technologies. This declaration represents a commitment from 29 governments to collaborate on developing AI in a manner that is human-centric, trustworthy, and responsible, with the UK, US, China, and major European member states among the notable signatories. The focus is on &#8220;frontier AI,&#8221; which refers to highly capable, general-purpose AI models that could pose significant risks, particularly in areas such as cybersecurity and biotechnology.<br />The declaration emphasizes the need for governments to take proactive measures to ensure the safe development of AI, acknowledging the technology&#8217;s pervasive deployment across various facets of daily life including housing, employment, education, and healthcare. It calls for the development of risk-based policies, appropriate evaluation metrics, tools for safety testing, and building relevant public sector capability and scientific research.<br />In addition to the declaration, a policy paper on AI &#8216;Safety Testing&#8217; was also signed by ten countries, including the UK and the US, as well as major technology companies. This policy paper outlines a broad framework for testing next-generation AI models by government agencies, promoting international cooperation, and enabling government agencies to develop their own approaches to AI safety regulation.<br />The key takeaways from the Bletchley Declaration include a clear signal from governments regarding the urgency to address the development of safe AI. However, how these commitments will translate into specific policy proposals and the role of the newly announced AI Safety Institute (AISI) in the UK&#8217;s regulatory landscape remain to be seen. The AISI&#8217;s mission is to minimize surprise from rapid and unexpected advances in AI, focusing on testing and evaluation of advanced AI systems, foundational AI safety research, and facilitating information exchange.</p>
<p style="text-align: justify;"><br />As they seek to establish themselves as AI leaders in the global community and set the direction for effective policymaking, both the US and the UK are navigating the balance between promoting AI innovation and ensuring ethical governance. While most of the current focus is on proposing guidelines and frameworks for the safe and responsible use of AI, the reference to potential future regulations across both documents should serve as a wake-up call for companies to start aligning their practices with the principles and recommendations outlined. <br />To stay ahead of the curve, organizations should develop robust methodologies to monitor AI risks effectively. This involves adapting their AI strategy to prioritize risk mitigation, identifying potential harms that may arise from the deployment of AI systems, and preparing for forthcoming regulatory measures by implementing a secure and comprehensive risk management program. <br />However, the US and UK opportunist approach to AI legislation is not followed by all. China chose a targeted and evolutive approach by writing a law on Generative AI that came into effect in 2023. Finally, in Europe, the AI Act shows that the EU doesn&#8217;t want to let AI technologies go out of hand.</p>
<p>Cet article <a href="https://www.riskinsight-wavestone.com/en/2024/05/us-executive-order-betchley-declaration/">US Executive Order &#038; Betchley Declaration</a> est apparu en premier sur <a href="https://www.riskinsight-wavestone.com/en/">RiskInsight</a>.</p>
]]></content:encoded>
					
					<wfw:commentRss>https://www.riskinsight-wavestone.com/en/2024/05/us-executive-order-betchley-declaration/feed/</wfw:commentRss>
			<slash:comments>0</slash:comments>
		
		
			</item>
		<item>
		<title>Securing AI: The New Cybersecurity Challenges</title>
		<link>https://www.riskinsight-wavestone.com/en/2024/03/securing-ai-the-new-cybersecurity-challenges/</link>
					<comments>https://www.riskinsight-wavestone.com/en/2024/03/securing-ai-the-new-cybersecurity-challenges/#respond</comments>
		
		<dc:creator><![CDATA[Gérôme Billois]]></dc:creator>
		<pubDate>Wed, 13 Mar 2024 15:08:52 +0000</pubDate>
				<category><![CDATA[Challenges]]></category>
		<category><![CDATA[Cloud & Next-Gen IT Security]]></category>
		<category><![CDATA[adversarial attacks]]></category>
		<category><![CDATA[AI]]></category>
		<category><![CDATA[AI security]]></category>
		<category><![CDATA[attaques par poison]]></category>
		<category><![CDATA[Auto-encoders]]></category>
		<category><![CDATA[auto-encodeurs]]></category>
		<category><![CDATA[federated learning]]></category>
		<category><![CDATA[GAN]]></category>
		<category><![CDATA[IA]]></category>
		<category><![CDATA[Oracle]]></category>
		<category><![CDATA[poison attacks]]></category>
		<category><![CDATA[prompt injection]]></category>
		<category><![CDATA[sécurité IA]]></category>
		<guid isPermaLink="false">https://www.riskinsight-wavestone.com/?p=22729</guid>

					<description><![CDATA[<p>The use of artificial intelligence systems and Large Language Models (LLMs) has exploded since 2023. Businesses, cybercriminals and individuals alike are beginning to use them regularly. However, like any new technology, AI is not without risks. To illustrate these, we...</p>
<p>Cet article <a href="https://www.riskinsight-wavestone.com/en/2024/03/securing-ai-the-new-cybersecurity-challenges/">Securing AI: The New Cybersecurity Challenges</a> est apparu en premier sur <a href="https://www.riskinsight-wavestone.com/en/">RiskInsight</a>.</p>
]]></description>
										<content:encoded><![CDATA[
<p style="text-align: justify;">The use of artificial intelligence systems and Large Language Models (LLMs) has exploded since 2023. Businesses, cybercriminals and individuals alike are beginning to use them regularly. However, like any new technology, AI is not without risks. To illustrate these, we have simulated two realistic attacks in previous articles: <a href="https://www.riskinsight-wavestone.com/en/2023/06/attacking-ai-a-real-life-example/">Attacking an AI? A real-life example!</a> and <a href="https://www.riskinsight-wavestone.com/en/2023/10/language-as-a-sword-the-risk-of-prompt-injection-on-ai-generative/">Language as a sword: the risk of prompt injection on AI Generative</a>.</p>
<p style="text-align: justify;">This article provides an overview of the <strong>threat posed by AI</strong> and the <strong>main defence mechanisms</strong> to democratize their use.</p>
<p style="text-align: justify;"> </p>
<h2 style="text-align: justify;"><span style="color: #612391;">AI introduces new attack techniques, already widely exploited by cybercriminals </span></h2>
<p style="text-align: justify;">As with any new technology, AI introduces new vulnerabilities and risks that need to be addressed in parallel with its adoption. The attack surface is vast: a malicious actor could <strong>attack</strong> both <strong>the model </strong>itself (model theft, model reconstruction, diversion from initial use) and<strong> its data</strong> (extracting training data, modifying behaviour by adding false data, etc.).</p>
<p style="text-align: justify;"><a href="https://www.riskinsight-wavestone.com/en/2023/10/language-as-a-sword-the-risk-of-prompt-injection-on-ai-generative/">Prompt injection</a> is undoubtedly the most talked-about technique. It enables an attacker to perform unwanted actions on the model, such as extracting sensitive data, executing arbitrary code, or generating offensive content.</p>
<p style="text-align: justify;">Given the growing variety of attacks on AI models, we will take a non-exhaustive look at the main categories:</p>
<h3 style="text-align: justify;"><span style="color: #5a75a3;">Data theft (impact on confidentiality)</span></h3>
<p style="text-align: justify;">As soon as data is used to train Machine Learning models, it can be (partially) reused to respond to users. A poorly configured model can then be a little too verbose, unintentionally revealing sensitive information. This situation presents a risk of violation of privacy and infringement of intellectual property.</p>
<p style="text-align: justify;">And the risk is all the greater if the models are &#8216;overfitted&#8217; with specific data. <strong>Oracle attacks</strong> take place when the model is in production, and the attacker questions the model to exploit its responses. These attacks can take several forms:</p>
<ul style="text-align: justify;">
<li><strong>Model extraction/theft</strong>: an attacker can extract a functional copy of a private model by using it as an oracle. By repeatedly querying the Machine Learning model&#8217;s API access, the adversary can collect the model&#8217;s responses. These responses will be used as labels to form a separate model that mimics the behaviour and performance of the target model.</li>
<li><strong>Membership inference attacks</strong>: this attack aims to check whether a specific piece of data has been used during the training of an AI model. The consequences can be far-reaching, particularly for health data: imagine being able to check whether an individual has cancer or not! This method was used by the New York Times to prove that its articles were used to train ChatGPT<a href="#_ftn1" name="_ftnref1">[1]</a>.</li>
</ul>
<p> </p>
<h3 style="text-align: justify;"><span style="color: #5a75a3;">Destabilisation and damage to reputation (impact on integrity)</span></h3>
<p style="text-align: justify;">The performance of a Machine Learning model depends on the reliability and quality of its training data. <strong>Poison attacks </strong>aim to compromise the training data  to affect the model&#8217;s performance:</p>
<ul style="text-align: justify;">
<li><strong>Model skewing</strong>: the attack aims to deliberately manipulate a model during training (either during initial training, or after it has been put into production if the model continues to learn) to introduce biases and steer the model&#8217;s predictions. As a result, the biased model may favour certain groups or characteristics, or be directed towards malicious predictions.</li>
<li><strong>Backdoors</strong>: an attacker can train and distribute a corrupted model containing a backdoor. Such a model functions normally until an input containing a trigger modifies its behaviour. This trigger can be a word, a date or an image. For example, a malware classification system may let malware through if it sees a specific keyword in its name or from a specific date. Malicious code can also be executed<a href="#_ftn2" name="_ftnref2">[2]</a>!</li>
</ul>
<p style="text-align: justify;">The attacker can also add carefully selected noise to mislead the prediction of a healthy model. This is known as an adversarial or evasion attack:</p>
<ul style="text-align: justify;">
<li><strong>Evasion attack</strong> (adversarial attack): the aim of this attack is to make the model generate an output not intended by the designer (making a wrong prediction or causing a malfunction in the model). This can be done by slightly modifying the input to avoid being detected as malicious input. For example:
<ul>
<li>Ask the model to describe a white image that contains a hidden injection prompt, <a href="https://twitter.com/goodside/status/1713000581587976372">written white on white in the image</a>.</li>
<li>Wear a special pair of glasses to avoid being recognised by a facial recognition algorithm<a href="#_ftn3" name="_ftnref3">[3]</a>.</li>
<li>Add a sticker of some kind to a &#8220;Stop&#8221; sign so that the model recognises a &#8220;45km/h limit&#8221; sign<a href="#_ftn4" name="_ftnref4">[4]</a>.</li>
</ul>
</li>
</ul>
<h3 style="text-align: justify;"><span style="color: #5a75a3;">Impact on availability</span></h3>
<p style="text-align: justify;">In addition to data theft and the impact on image, attackers can also hamper the availability of Artificial Intelligence (AI) systems. These tactics are aimed not only at making data unavailable, but also at disrupting the regular operation of systems. One example is the poisoning attack, the impact of which is to make the model unavailable while it is retrained (which also has an economic impact due to the cost of retraining the model). Here is another example of an attack:</p>
<ul style="text-align: justify;">
<li><strong>Denial of service attack (DDOS) on the model</strong>: like all other applications, Machine Learning models are sensitive to denial-of-service attacks that can hamper system availability. The attack can combine a high number of requests, while sending requests that are very heavy to process. In the case of Machine Learning models, the financial consequences are greater because tokens/prompts are very expensive (for example, ChatGPT is not profitable despite its 616 million monthly users).</li>
</ul>
<p style="text-align: justify;"> </p>
<h2 style="text-align: justify;"><span style="color: #612391;">Two ways of securing your AI projects: adapt your existing cyber controls, and develop specific Machine Learning measures</span></h2>
<p style="text-align: justify;">Just like security projects, a prior risk analysis is necessary to implement the right controls, while finding an acceptable compromise between security and the functioning of the model. To do this, <strong>our traditional risk methods need to evolve</strong> to include the risks detailed above, which are not well covered by historical methods.</p>
<p style="text-align: justify;">Following these risk analyses, security measures will need to be implemented. <strong>Wavestone has identified over 60 different measures</strong>. In this second part, we present a small selection of these measures to be implemented according to the criticality of your models.</p>
<h3 style="text-align: justify;"><span style="color: #5a75a3; font-size: revert; font-weight: revert;">1.   Adapting cyber controls to Machine Learning models</span></h3>
<p style="text-align: justify;">The first line of defence corresponds to the basic application, infrastructure, and organisational measures for cybersecurity. The aim is to adapt requirements that we already know about, which are present in the various security policies, but do not necessarily apply in the same way to AI projects. We need to consider these specificities, which can sometimes be quite subtle.</p>
<p style="text-align: justify;">The most obvious example is the creation of <strong>AI pentests</strong>. Conventional pentests involve finding a vulnerability to gain access to the information system. However, AI models can be attacked without entering the IS (like evasion and oracle attacks). RedTeaming procedures need to evolve to deal with these particularities while developing detection and incident response mechanisms to cover the new applications of AI.</p>
<p style="text-align: justify;">Another essential example is the <strong>isolation of AI environments</strong> used throughout the lifecycle of Machine Learning models. This reduces the impact of a compromise by protecting the models, training data, and prediction results.</p>
<p style="text-align: justify;">You also need to assess the <strong>regulations</strong> and laws with which the Machine Learning application must comply, and adhere to the latest legislation on artificial intelligence (the IA Act in Europe, for example).</p>
<p style="text-align: justify;">And finally, a more than classic measure: <strong>awareness and training campaigns</strong>. We need to ensure that the stakeholders (project managers, developers, etc.) are trained in the risks of AI systems and that users are made aware of these risks.</p>
<p> </p>
<h3><span style="color: #5a75a3;">2.  Specific controls to protect sensitive Machine Learning models</span></h3>
<p style="text-align: justify;">In addition to the standard measures that need to be adapted, specific measures need to be identified and applied.</p>
<h4 style="text-align: justify;"><span style="color: #bf5283;">For your least critical projects, keep things simple and implement the basics</span></h4>
<p style="text-align: justify;"><strong>Poison control</strong>: to guard against poisoning attacks, you need to detect any &#8220;false&#8221; data that may have been injected by an attacker. This involves using exploratory statistical analysis to identify poisoned data (analysing the distribution of data and identifying absurd data, for example). This step can be included in the lifecycle of a Machine Learning model to automate downstream actions. However, human verification will always be necessary.</p>
<p style="text-align: justify;"><strong>Input control</strong> (analysing user input): to counter prompt injection and evasion attacks, user input is analysed and filtered to block all malicious input. We can think of basic rules (blocking requests containing a specific word) as well as more specific statistical rules (format, consistency, semantic coherence, noise, etc.). However, this approach could have a negative impact on model performance, as false positives would be blocked.</p>
<p><img loading="lazy" decoding="async" class="aligncenter wp-image-22699" src="https://www.riskinsight-wavestone.com/wp-content/uploads/2024/03/Picture1.png" alt="" width="700" height="182" srcset="https://www.riskinsight-wavestone.com/wp-content/uploads/2024/03/Picture1.png 2545w, https://www.riskinsight-wavestone.com/wp-content/uploads/2024/03/Picture1-437x114.png 437w, https://www.riskinsight-wavestone.com/wp-content/uploads/2024/03/Picture1-71x18.png 71w, https://www.riskinsight-wavestone.com/wp-content/uploads/2024/03/Picture1-768x200.png 768w, https://www.riskinsight-wavestone.com/wp-content/uploads/2024/03/Picture1-1536x400.png 1536w, https://www.riskinsight-wavestone.com/wp-content/uploads/2024/03/Picture1-2048x533.png 2048w" sizes="auto, (max-width: 700px) 100vw, 700px" /></p>
<h4> </h4>
<h4 style="text-align: justify;"><span style="color: #bf5283;">For your moderately sensitive projects, aim for a good investment/risk coverage ratio</span></h4>
<p style="text-align: justify;">There is a plethora of measures, and a great deal of <a href="https://www.enisa.europa.eu/publications/securing-machine-learning-algorithms">literature</a> on the subject. On the other hand, some measures can cover several risks at once. We think it is worth considering them first.</p>
<p style="text-align: justify;"><strong>Transform inputs</strong>: an input transformation step is added between the user and the model. The aim is twofold:</p>
<ol style="text-align: justify;">
<li>For example, remove or modify any malicious input by reformulating the input or truncating it. An implementation using encoders is also possible (but will be detailed in the next section).</li>
<li>Another instance will be to reduce the attacker&#8217;s visibility to counter oracle attacks (which require precise knowledge of the model&#8217;s input and output) by adding random noise or reformulating the prompt.</li>
</ol>
<p style="text-align: justify;">Depending on the implementation method, impacts on model performance are to be expected.</p>
<p style="text-align: justify;"><strong>Supervise AI with AI models</strong>: any AI model that learns after it has been put into production must be specifically supervised as part of overall incident detection and response processes. This involves both collecting the appropriate logs to carry out investigations, but also monitoring the statistical deviation of the model to spot any abnormal drift. In other words, it involves assessing changes in the quality of predictions over time. Microsoft&#8217;s Tay model launched on Twitter in 2016 is a good example of a model that has drifted.</p>
<p><img loading="lazy" decoding="async" class="aligncenter wp-image-22701" src="https://www.riskinsight-wavestone.com/wp-content/uploads/2024/03/Picture2.png" alt="" width="700" height="192" srcset="https://www.riskinsight-wavestone.com/wp-content/uploads/2024/03/Picture2.png 2404w, https://www.riskinsight-wavestone.com/wp-content/uploads/2024/03/Picture2-437x120.png 437w, https://www.riskinsight-wavestone.com/wp-content/uploads/2024/03/Picture2-71x20.png 71w, https://www.riskinsight-wavestone.com/wp-content/uploads/2024/03/Picture2-768x211.png 768w, https://www.riskinsight-wavestone.com/wp-content/uploads/2024/03/Picture2-1536x422.png 1536w, https://www.riskinsight-wavestone.com/wp-content/uploads/2024/03/Picture2-2048x563.png 2048w" sizes="auto, (max-width: 700px) 100vw, 700px" /></p>
<p> </p>
<h4 style="text-align: justify;"><span style="color: #bf5283;">For your critical projects, go further to cover specific risks</span></h4>
<p style="text-align: justify;">There are measures that we believe are highly effective in covering certain risks. Of course, this involves carrying out a risk analysis beforehand. Here are two examples (among many others):</p>
<p style="text-align: justify;"><strong>Randomized Smoothing</strong>: a training technique designed to improve the robustness of a model&#8217;s predictions. The model is trained twice: once with real training data, then a second time with the same data altered by noise. The aim is to have the same behaviour, whether noise is present in the input. This limits evasion attacks, particularly for classification algorithms.</p>
<p style="text-align: justify;"><strong>Learning from contradictory examples</strong>: the aim is to teach the model to recognise malicious inputs to make it more robust to adversarial attacks. In practical terms, this means labelling contradictory examples (i.e. a real input that includes a small error/disturbance) as malicious data and adding them during the training phase. By confronting the model with these simulated attacks, it learns to recognise and counter malicious patterns. This is a very effective measure, but it involves a certain cost in terms of resources (longer training phase) and can have an impact on the accuracy of the model.</p>
<p style="text-align: justify;"><img loading="lazy" decoding="async" class="aligncenter wp-image-22703" src="https://www.riskinsight-wavestone.com/wp-content/uploads/2024/03/Picture3.png" alt="" width="700" height="192" srcset="https://www.riskinsight-wavestone.com/wp-content/uploads/2024/03/Picture3.png 2417w, https://www.riskinsight-wavestone.com/wp-content/uploads/2024/03/Picture3-437x120.png 437w, https://www.riskinsight-wavestone.com/wp-content/uploads/2024/03/Picture3-71x19.png 71w, https://www.riskinsight-wavestone.com/wp-content/uploads/2024/03/Picture3-768x210.png 768w, https://www.riskinsight-wavestone.com/wp-content/uploads/2024/03/Picture3-1536x421.png 1536w, https://www.riskinsight-wavestone.com/wp-content/uploads/2024/03/Picture3-2048x561.png 2048w" sizes="auto, (max-width: 700px) 100vw, 700px" /></p>
<p> </p>
<h2 style="text-align: justify;"><span style="color: #612391;">Versatile guardians &#8211; three sentinels of AI security</span></h2>
<p style="text-align: justify;">Three methods stand out for their effectiveness and their ability to mitigate several attack scenarios simultaneously: <strong>GAN</strong> (Generative Adversarial Network), <strong>filters</strong> (encoders and auto-encoders that are models of neural networks) and <strong>federated learning</strong>.</p>
<h3 style="text-align: justify;"><span style="color: #5a75a3;">The GAN: the forger and the critic</span></h3>
<p style="text-align: justify;">The GAN, or Generative Adversarial Network, is an AI model training technique that works like a forger and a critic working together. The forger, called the generator, creates &#8220;copies of works of art&#8221; (such as images). The critic, called the discriminator, evaluates these works to identify the fakes from the real ones and gives advice to the forger on how to improve. The two work in tandem to produce increasingly realistic works until the critic can no longer identify the fakes from the real thing.</p>
<p style="text-align: justify;">A GAN can help reduce the attack surface in two ways:</p>
<ul style="text-align: justify;">
<li>With the <strong>generator (the faker)</strong> to prevent sensitive data leaks. A new fictitious training database can be generated, like the original but containing no sensitive or personal data.</li>
<li>The <strong>discriminator (the critic)</strong> limits evasion or poisoning attacks by identifying malicious data. The discriminator compares a model&#8217;s inputs with its training data. If they are too different, then the input is classified as malicious. In practice, it can predict whether an input belongs to the training data by associating a likelihood scope with it.</li>
</ul>
<p> </p>
<h3 style="text-align: justify;"><span style="color: #5a75a3;">Auto-encoders: an unsupervised learning algorithm for filtering inputs and</span><span style="color: #5a75a3;"> outputs</span></h3>
<p style="text-align: justify;">An auto-encoder transforms an input into another dimension, changing its form but not its essence. To take a simplifying analogy, it&#8217;s as if the prompt were summarized and rewritten to remove undesirable elements. In practice, the input is compressed by a noise-removing encoder (via a first layer of the neural network), then reconstructed via a decoder (via a second layer). This model has two uses:</p>
<ul style="text-align: justify;">
<li>If an auto-encoder is positioned <strong>upstream</strong> of the model, it will have the ability to transform the input before it is processed by the application, removing potential malicious payloads. In this way, it becomes more difficult for an attacker to introduce elements enabling an evasion attack, for example.</li>
<li>We can use this same system <strong>downstream</strong> of the model to protect against oracle attacks (which aim to extract information about the data or the model by interrogating it). The output will thus be filtered, reducing the verbosity of the model, i.e. reducing the amount of information output by the model.</li>
</ul>
<p style="text-align: justify;"> </p>
<h3 style="text-align: justify;"><span style="color: #5a75a3;">Federated Learning: strength in numbers</span></h3>
<p style="text-align: justify;">When a model is deployed on several devices, a delocalised learning method such as federated learning can be used. The principle: several models learn locally with their own data and only send their learning back to the central system. This allows several devices to collaborate without sharing their raw data. This technique makes it possible to cover a large number of cyber risks in applications based on artificial intelligence models:</p>
<ul style="text-align: justify;">
<li><strong>Segmentation of training databases</strong> plays a crucial role in limiting the risks of Backdoor and Model Skewing poisoning. The fact that training data is specific to each device makes it extremely difficult for an attacker to inject malicious data in a coordinated way, as he does not have access to the global set of training data. This same division limits the risks of data extraction.</li>
<li>The federated learning process also limits the <strong>risks of model extraction</strong>. The learning process makes the link between training data and model behaviour extremely complex, as the model does not learn directly. This makes it difficult for an attacker to understand the link between input and output data.</li>
</ul>
<p style="text-align: justify;">Together, GAN, filters (encoders and auto-encoders) and federated learning form a good risk hedging proposition for Machine Learning projects despite the technicality of their implementation. These versatile guardians demonstrate that innovation and collaboration are the pillars of a robust defence in the dynamic artificial intelligence landscape.</p>
<p style="text-align: justify;">To take this a step further, Wavestone has written a <a href="https://www.enisa.europa.eu/publications/securing-machine-learning-algorithms">practical guide</a> for ENISA on securing the deployment of machine learning, which lists the various security controls that need to be established.</p>
<p style="text-align: justify;"> </p>
<h2 style="text-align: justify;"><span style="color: #612391;">In a nutshell</span></h2>
<p style="text-align: justify;">Artificial intelligence can be compromised by methods that are not usually encountered in our information systems. There is no such thing as zero risk: every model is vulnerable. To mitigate these new risks, additional defence mechanisms need to be implemented depending on the criticality of the project. A compromise will have to be found between security and model performance.</p>
<p style="text-align: justify;">AI security is a very active field, from Reddit users to advanced research work on model deviation. That&#8217;s why it&#8217;s important to keep an organisational and technical watch on the subject.</p>
<p> </p>
<p style="text-align: justify;"><a href="#_ftnref1" name="_ftn1">[1]</a> <a href="https://www.nytimes.com/2023/12/27/business/media/new-york-times-open-ai-microsoft-lawsuit.html">New York Times proved that their articles were in AI training data set</a></p>
<p style="text-align: justify;"><a href="#_ftnref2" name="_ftn2">[2]</a> <a href="https://www.clubic.com/actualite-520447-au-moins-une-centaine-de-modeles-d-ia-malveillants-seraient-heberges-par-la-plateforme-hugging-face.html">Au moins une centaine de modèles d&#8217;IA malveillants seraient hébergés par la plateforme Hugging Face</a></p>
<p style="text-align: justify;"><a href="#_ftnref3" name="_ftn3">[3]</a> Sharif, M. et al. (2016). Accessorize to a crime: Real and stealthy attacks on state-of-the-art face recognition. ACM Conference on Computer and Communications Security (CCS)</p>
<p style="text-align: justify;"><a href="#_ftnref4" name="_ftn4">[4]</a> Eykholt, K. et al. (2018). Robust Physical-World Attacks on Deep Learning Visual Classification. CVPR. <a href="https://arxiv.org/pdf/1707.08945.pdf">https://arxiv.org/pdf/1707.08945.pdf</a></p>
<p style="text-align: justify;"> </p>
<p>Cet article <a href="https://www.riskinsight-wavestone.com/en/2024/03/securing-ai-the-new-cybersecurity-challenges/">Securing AI: The New Cybersecurity Challenges</a> est apparu en premier sur <a href="https://www.riskinsight-wavestone.com/en/">RiskInsight</a>.</p>
]]></content:encoded>
					
					<wfw:commentRss>https://www.riskinsight-wavestone.com/en/2024/03/securing-ai-the-new-cybersecurity-challenges/feed/</wfw:commentRss>
			<slash:comments>0</slash:comments>
		
		
			</item>
	</channel>
</rss>
