We are using AI tools everywhere — from coding assistants to resume builders to business workflows.
But there’s a problem most people ignore:
We are pasting sensitive data into AI systems without thinking twice.
Emails. API keys. Client details. Internal documents.
And once that data is sent, we lose control over how it’s processed.
The Problem Nobody Talks About
AI tools are powerful, but they are not designed with user-side privacy protection in mind.
Most users either:
• Manually remove sensitive data before using AI
• Or ignore the risk completely
Neither approach is reliable.
Manual editing is slow and error-prone. Ignoring the risk can lead to serious consequences.
The Idea: A Privacy Layer for AI
Instead of changing how AI works, I asked:
What if we add a security layer before the data reaches the model?
(LLM MODEL)
That’s how ARGUS OBSIDIAN was built.
How ARGUS Works
ARGUS sits between the user and the AI model and processes data in real time.
The system follows a simple pipeline:
Input → Detect → Mask → Send → Restore → Display
- Detection
The system identifies sensitive data such as:
• API keys
• Emails
• Phone numbers
• Passwords
• Addresses
- Masking
Sensitive data is replaced with placeholders:
[EMAIL_1], [API_KEY_1]
A mapping is stored internally.
- Secure Processing
Only the masked version is sent to the AI model.
This ensures that raw sensitive data never leaves the system.
- Restoration
After receiving the response, ARGUS restores the original data seamlessly.
The user sees a clean, natural output without any loss of meaning.
The Experience
To make the system usable, I designed it as a chat interface.
The goal was simple:
The user should not have to think about privacy.
They type normally. The system protects automatically.
Why This Matters
AI adoption is increasing rapidly, but privacy practices are not keeping up.
If we want AI to be trusted in real-world workflows, we need systems that protect users by default.
ARGUS is a step in that direction.
What’s Next
• Support for more sensitive data types
• Local model integration for full privacy
• Browser-level protection for all AI tools
Final Thought
AI should not force users to choose between convenience and privacy.
It should give both.
ARGUS is an attempt to make that possible.
Top comments (0)