Most AI security tools have a fatal flaw: they can be modified at runtime.
Your guardrails, your content filters, your prompt injection detectors. They're all just Python objects sitting in memory. One clever exploit, one monkey-patched module, and your entire security stack folds.
I built SovereignShield to fix this. It's an Immutable AI firewall where every security layer is sealed with Python's FrozenNamespace after initialization. Once sealed, the rules cannot be changed, bypassed, or tampered with. Not by an attacker, not by a rogue plugin, not even by your own code.
The Problem: Mutable Security is Broken Security
Here's what a typical AI security setup looks like:
class SecurityFilter:
def __init__(self):
self.blocked_patterns = ["ignore previous", "system prompt"]
def check(self, text):
return not any(p in text.lower() for p in self.blocked_patterns)
Looks fine, right? Except anyone with access to the object can do this:
filter.blocked_patterns = [] # Security? Gone.
Or worse, a sophisticated prompt injection could trigger code that modifies the filter at runtime. Your security layer just became decoration.
The Fix: FrozenNamespace
SovereignShield seals every security layer after initialization:
from types import SimpleNamespace
class FrozenNamespace(SimpleNamespace):
"""Immutable after creation. Cannot be modified."""
_frozen = False
def __init__(self, **kwargs):
super().__init__(**kwargs)
object.__setattr__(self, '_frozen', True)
def __setattr__(self, name, value):
if self._frozen:
raise AttributeError("This object is sealed and cannot be modified.")
super().__setattr__(name, value)
def __delattr__(self, name):
raise AttributeError("This object is sealed and cannot be modified.")
Once the 4 security layers (InputFilter, AdaptiveShield, CoreSafety, Conscience) are initialized, they're frozen. Any attempt to modify them raises an exception. Period.
What It Actually Does
SovereignShield scans both user input (before it reaches your LLM) and LLM output (before it reaches your users). It catches:
Prompt injection (50+ patterns)
Credential exfiltration attempts
Shell command injection
Data leak patterns
Social engineering attacks
All in under 1 millisecond. Zero dependencies. No API calls to third-party LLMs to "judge" if something is safe.
Try It
Grab a free API key (1,000 scans/month, no credit card) at sovereign-shield.net , then:
pip install sovereign-shield-client
from sovereign_shield_client import SovereignShield
shield = SovereignShield(api_key="your_key")
# Scan user input before sending to LLM
safe_input = shield.scan("user's message here")
# Scan LLM response before showing to user
safe_output = shield.veto("LLM's response here")
If the input or output is safe, you get the string back. If it's dangerous, an InputBlockedError is raised with the reason. That's the entire integration.
Get a free API key (1,000 scans/month) at sovereign-shield.net
The full source is on GitHub:https://github.com/mattijsmoens/sovereign-shield under BSL 1.1.
The point isn't that SovereignShield has more rules or fancier detection. The point is that the rules can't be turned off. In security, that's the only thing that matters.
Top comments (0)