
A viral Instagram and LinkedIn development is popping innocent enjoyable into a possible safety headache.
Hundreds of thousands of customers are prompting ChatGPT to “create a caricature of me and my job primarily based on every part about me,” then posting the outcomes publicly — inadvertently signaling how they use AI at work and what entry they may need to delicate information.
“Whereas many have been discussing the privateness dangers of individuals following the ChatGPT caricature development, the immediate reveals one thing else alarming — persons are speaking to their LLMs about work,” stated Josh Davies, principal market strategist at Fortra, in an e-mail to eSecurityPlanet.
He added, “If they don’t seem to be utilizing a sanctioned ChatGPT occasion, they might be inputting delicate work data right into a public LLM. Those that publicly share these pictures could also be placing a goal on their again for social engineering makes an attempt, and malicious actors have thousands and thousands of entries to pick engaging targets from.”
Davies defined, “If an attacker is ready to take over the LLM account, probably utilizing the detailed data included within the picture for a focused social engineering assault, then they may view the immediate historical past and seek for delicate data shared with the LLM.”
He additionally added, “This development doesn’t simply spotlight a privateness danger, but additionally the danger of shadow AI and information leakage in prompts – the place organizations lose management of their delicate information by way of workers irresponsibly utilizing AI.”
How AI tendencies expose enterprise information
The OWASP LLM Prime Ten lists Delicate Data Disclosure (LLM2025:02) as one of many high dangers related to LLMs.
This danger extends past unintended oversharing — it encompasses any situation through which delicate information entered into an LLM turns into accessible to unauthorized events.
In opposition to that backdrop, the AI caricature development is greater than innocent social media leisure.
It serves as a visual indicator of a broader shadow AI problem: workers utilizing public AI platforms with out formal governance, oversight, or technical controls. It additionally demonstrates how simply menace actors can determine people who’re more likely to combine LLMs into their day by day workflows.
How the AI caricature development expands the assault floor
Most of the posted caricatures clearly depict the consumer’s occupation — banker, engineer, HR supervisor, developer, healthcare supplier.
Whereas job titles themselves are sometimes publicly out there by means of skilled networking websites, participation on this development provides a brand new layer of context. By producing and sharing these pictures, customers successfully affirm that they depend on a selected public LLM platform for work-related actions. That affirmation is efficacious intelligence for an adversary conducting reconnaissance.
The dimensions amplifies the danger. On the time of writing, thousands and thousands of pictures have been shared, many from public accounts, making a searchable dataset of pros who seemingly use public AI techniques.
For attackers, this lowers the barrier to constructing focused phishing lists targeted on high-value roles with possible entry to delicate data.
Safety groups evaluating this development ought to view it by means of the lens of shadow AI and AI governance. Unapproved or unmanaged AI utilization expands the group’s assault floor, typically with out visibility from safety operations groups.
The caricature itself shouldn’t be the vulnerability; fairly, it alerts that probably delicate prompts might have been submitted to an exterior AI service outdoors enterprise management.
The 2 major menace paths
From a menace modeling perspective, two major assault paths emerge: account takeover and delicate information extraction by means of manipulation.
The extra rapid danger is LLM account compromise. A public Instagram put up supplies a username, profile data, and infrequently clues in regards to the particular person’s employer and duties. Utilizing fundamental open-source intelligence strategies, attackers can incessantly correlate this information with an e-mail deal with.
If that very same e-mail deal with is used to register for the LLM platform, focused phishing or credential harvesting assaults change into considerably simpler. As soon as an attacker positive aspects entry to the LLM account, the affect can escalate rapidly.
Immediate histories might include buyer information, inside communications, monetary projections, proprietary supply code, or strategic planning discussions.
As a result of LLM interfaces permit customers to look, summarize, and reference previous conversations, an attacker with authenticated entry can effectively determine and extract worthwhile data.
Though suppliers implement safeguards to stop cross-user information publicity, immediate histories stay absolutely accessible to the legit — or compromised — account holder.
Immediate injection and mannequin manipulation
The second path includes immediate injection assaults.
Safety researchers have demonstrated a number of methods to govern mannequin habits, together with persona-based jailbreaks, instruction overrides like “ignore earlier directions,” and payload-splitting strategies that reconstruct malicious prompts inside the mannequin’s context window.
In each instances, the underlying problem shouldn’t be the caricature development itself.
The actual danger lies in what it implies: that delicate enterprise data might have been entered into unmanaged, public AI environments. The social media put up merely makes that danger extra seen — to defenders and adversaries alike.
Sensible steps to cut back shadow AI danger
As generative AI turns into extra built-in into on a regular basis workflows, organizations ought to undertake a structured, proactive strategy to managing related dangers.
- Set up and repeatedly reinforce a complete AI governance coverage that clearly defines acceptable use, information dealing with necessities, and worker duties.
- Present a safe, enterprise-managed AI different whereas proscribing or monitoring unsanctioned AI purposes to cut back shadow AI publicity.
- Deploy information loss prevention and information classification controls to detect, block, or warn in opposition to the submission of delicate data into AI platforms.
- Implement sturdy identification and entry administration practices, together with multi-factor authentication, role-based entry controls, and monitoring for credential publicity.
- Phase and monitor AI site visitors by means of safe net gateways, browser isolation, or community controls to cut back the danger of knowledge exfiltration and lateral motion.
- Combine AI-specific eventualities into safety consciousness applications and repeatedly check incident response plans by means of tabletop workout routines involving AI-related compromise.
- Repeatedly monitor for indicators of AI account compromise, immediate misuse, or leaked credentials throughout the open net and darkish net.
Efficient AI danger administration requires greater than a single coverage or device; it includes coordinated governance, technical controls, consumer schooling, and ongoing monitoring.
Editor’s observe: This text initially appeared on our sister web site, eSecurityPlanet.