Skip to main content
COSMICBYTEZLABS
NewsSecurityHOWTOsToolsStudyTraining
ProjectsChecklistsAI RankingsNewsletterStatusTagsAbout
Subscribe

Press Enter to search or Esc to close

News
Security
HOWTOs
Tools
Study
Training
Projects
Checklists
AI Rankings
Newsletter
Status
Tags
About
RSS Feed
Reading List
Subscribe

Stay in the Loop

Get the latest security alerts, tutorials, and tech insights delivered to your inbox.

Subscribe NowFree forever. No spam.
COSMICBYTEZLABS

Your trusted source for IT intelligence, cybersecurity insights, and hands-on technical guides.

735+ Articles
120+ Guides

CONTENT

  • Latest News
  • Security Alerts
  • HOWTOs
  • Projects
  • Exam Prep

RESOURCES

  • Search
  • Browse Tags
  • Newsletter Archive
  • Reading List
  • RSS Feed

COMPANY

  • About Us
  • Contact
  • Privacy Policy
  • Terms of Service

© 2026 CosmicBytez Labs. All rights reserved.

System Status: Operational
  1. Home
  2. News
  3. Microsoft and Salesforce Patch Prompt Injection Flaws in AI Agents
Microsoft and Salesforce Patch Prompt Injection Flaws in AI Agents
NEWS

Microsoft and Salesforce Patch Prompt Injection Flaws in AI Agents

Both Microsoft Copilot and Salesforce Agentforce contained prompt injection vulnerabilities that allowed external attackers to leak sensitive data through AI agent interactions. Both flaws have now been patched.

Dylan H.

News Desk

April 16, 2026
5 min read

Microsoft and Salesforce have both issued patches for prompt injection vulnerabilities in their respective AI agent platforms — Microsoft Copilot and Salesforce Agentforce — that would have allowed an external attacker to cause the AI agents to leak sensitive data from connected enterprise systems.

The flaws represent an emerging attack class specific to agentic AI systems: rather than exploiting code logic directly, attackers craft malicious content that manipulates the AI model's behavior when it processes that content during a task.

What Was Found

Security researchers identified prompt injection vulnerabilities in both platforms during the same general research period. Both flaws shared a common characteristic: an external party could influence AI agent behavior by embedding adversarial instructions in content that the agent would process during normal operation.

Microsoft Copilot

In Microsoft Copilot, the vulnerability allowed an attacker to craft a malicious document, email, or webpage. When a Copilot user asked the AI to summarize, analyze, or act on that content, hidden instructions embedded in the malicious content could redirect the agent to exfiltrate data from the user's Microsoft 365 environment — including emails, files, and calendar data — to an attacker-controlled destination.

Microsoft Copilot is deeply integrated into Microsoft 365 services and has access to enterprise data at the level of the authenticated user. This makes prompt injection attacks against Copilot particularly high-impact: a single successful injection can expose the entire corpus of data accessible to the targeted user's account.

Salesforce Agentforce

In Salesforce Agentforce, researchers found a similar injection path. Agentforce agents — which can access Salesforce CRM objects, run automations, and communicate via connected channels — could be manipulated by adversarial content in customer messages, support tickets, or external data feeds processed by the agent.

A successful exploitation would cause Agentforce to retrieve and transmit sensitive CRM data (such as customer records, opportunity details, or contact information) to locations specified by the attacker-controlled injection payload.

Why Prompt Injection Is Difficult to Prevent

Unlike traditional injection attacks (SQL injection, command injection), prompt injection targets the AI model's instruction-following behavior rather than a deterministic parser with well-defined syntax rules. This creates unique defensive challenges:

Traditional InjectionPrompt Injection
Exploits parser logicExploits model instruction-following
Fixed syntax boundariesFluid natural language — no clear delimiter
Detectable via pattern matchingRequires semantic understanding to detect
Patched via input sanitizationRequires multi-layer controls

Current mitigations include:

  • System prompt hardening — Explicit instructions to the model not to follow instructions embedded in processed content
  • Output filtering — Inspecting agent outputs for evidence of exfiltration patterns before they are executed
  • Permission scoping — Restricting agent capabilities so they cannot send data to external endpoints without explicit user confirmation
  • Content provenance tracking — Flagging content from external sources as untrusted before it reaches the agent's context window

Neither approach is fully reliable in isolation; defense-in-depth is required.

Patch Status

Both Microsoft and Salesforce have confirmed that the vulnerabilities have been addressed:

VendorPlatformFix Status
MicrosoftCopilot (Microsoft 365)Patched — no user action required
SalesforceAgentforcePatched — no user action required

As server-side AI platforms, both fixes were deployed by the vendors without requiring customers to update local software. Users of both platforms are automatically protected following the vendor-side remediation.

Recommendations for AI Platform Administrators

Despite the patches, organizations using AI agents with broad data access should review their deployment configurations:

  1. Audit AI agent permissions — Apply least-privilege to all AI agent service accounts; agents should only access data required for their specific functions
  2. Enable agent action logging — Ensure all AI agent actions (data reads, external communications, automations triggered) are logged and reviewable
  3. Implement human-in-the-loop checkpoints for high-risk agent actions such as sending emails, creating records, or transmitting data externally
  4. Monitor for anomalous agent behavior — Unexpectedly large data retrievals or outbound requests to unfamiliar URLs may indicate exploitation
  5. Restrict external content processing — Where possible, prevent AI agents from processing unvetted content from external parties (public web, anonymous customer messages) without sandboxing

Broader AI Security Context

These disclosures are part of a broader pattern of security research exposing prompt injection as a fundamental risk class for agentic AI systems. As organizations increasingly deploy AI agents with privileged access to enterprise data — email, CRM, code repositories, financial systems — the attack surface for prompt injection grows proportionally.

Security researchers and vendors are working to establish standards for AI agent security, including proposed frameworks from OWASP (the LLM Application Security Top 10) and emerging NIST guidance on AI system risk management. The OWASP LLM Top 10 lists prompt injection as the #1 risk for large language model applications.

For organizations evaluating or expanding AI agent deployments, security architecture reviews should treat prompt injection as a first-class concern alongside traditional authentication and authorization controls.

References

  • Microsoft, Salesforce Patch AI Agent Data Leak Flaws — Dark Reading
  • OWASP LLM Top 10: LLM01 — Prompt Injection
  • Microsoft Security Response Center
  • Salesforce Security Advisories
#Microsoft#Salesforce#AI Security#Prompt Injection#Copilot#Agentforce#Data Breach

Related Articles

Microsoft and Salesforce Patch AI Agent Prompt Injection Flaws Enabling Data Leaks

Security researchers disclosed prompt injection vulnerabilities in both Microsoft Copilot and Salesforce Agentforce that would have allowed unauthenticated attackers to exfiltrate sensitive data through manipulated AI agent responses. Both vendors have issued patches.

5 min read

Microsoft Discovers 'AI Recommendation Poisoning' via

Microsoft's Defender team tracked over 50 unique prompt injection payloads from 31 companies using 'Summarize with AI' buttons to manipulate chatbot...

3 min read

Data Breach at EdTech Giant McGraw Hill Affects 13.5 Million Accounts

ShinyHunters has leaked over 100GB of data from 13.5 million McGraw Hill user accounts after exploiting a Salesforce misconfiguration. Names, addresses, phone numbers, and emails were exposed in the extortion campaign.

5 min read
Back to all News