An organization implemented a customizable GPT interface that allowed end users to define system instructions and share tailored personas for contextual conversations with an LLM. During a security review, Blueinfy identified that allowing users to modify the system context introduced significant vulnerabilities, as malicious instructions could alter the model’s behavior. Exploits included attempts to exfiltrate chat history, prompt users for sensitive PII, spread misinformation, and present phishing links. These risks not only threatened data security but also posed reputational harm if the application generated biased, abusive, or unsafe outputs. Blueinfy recommended implementing real-time content filtering, moderation tools, and guardrails to block harmful inputs and protect the LLM’s behavior. The improved safeguards ensured safer, controlled interactions with the customized GPT system.