There is a dangerous trend in AI right now: Context Gluttony.
We are cheering for 1 million token context windows. We are dumping entire wikis into vector databases and calling it "RAG." We are feeding our LLMs thousands of documents and hoping they can find the needle in the haystack.
But in enterprise engineering, more context is often less intelligence.
If I ask my AI agent, "What is blocking my team right now?", and it reads 5,000 Slack messages, 200 Jira tickets, and the last month of git logs, it will hallucinate. It will find a pattern that isn't there. It will confuse a "blocking" issue from three weeks ago with a "blocking" issue from today.
I propose a different approach: Scale by Subtraction.
Instead of building bigger pipes to feed the AI, we should build smarter filters to starve it of noise. We don't need a bigger haystack; we need a magnet.
I call this architecture the Multidimensional Knowledge Graph.
The Problem with Flat RAG
Traditional RAG (Retrieval-Augmented Generation) is essentially flat. It converts text to vectors and looks for semantic similarity.
If I query: "What critical bugs are assigned to me?"
- Vector Search looks for the words "critical," "bug," and my name.
- The Result: It retrieves a "critical" bug I fixed last year, a "bug" discussion in Slack where someone mentioned my name, and a "critical" feature request that belongs to another team.
The LLM receives all this garbage and tries to make sense of it. This is expensive, slow, and prone to error.
The Solution: Semantic Physics
A Multidimensional Knowledge Graph doesn't just store data; it enforces the "physics" of your organization. It understands that:
- Time Matters: A ticket from yesterday is heavier than a ticket from last month.
- Authority Matters: A Jira ticket is heavier than a Slack rumor.
- Hierarchy Matters: If my direct report has a P0 outage, that is my problem too.
By applying these dimensions as constraints, we can subtract 99% of the noise before the LLM ever sees a single token.
Visualizing the Physics of Time
Most vector stores treat a document from 2020 the same as a document from 2025 if the keywords match. The Multidimensional Graph applies an Exponential Decay Function to the relevance score.
In this model, a "Critical" bug from 6 months ago (Score 0.05) will never outweigh a "Medium" issue from today (Score 1.0). The graph physically suppresses old noise.
The Code: Implementing the "Semantic Firewall"
I built a prototype to demonstrate this. The core logic isn't in the prompt; it's in a filtering pipeline that acts as a semantic firewall.
Here is the Python implementation of the multidimensional query engine:
def query_with_constraints(self, query: str, user_context: UserContext) -> List[WorkItem]:
"""
The Semantic Firewall: Subtracting noise deterministically.
"""
candidates = self.work_items.copy()
# Dimension 1: Identity (Who am I?)
# If I am a Manager, show me critical/strategic items.
candidates = self.dimensions['identity'].apply_filter(user, candidates)
# Dimension 2: Org Hierarchy (Who works for me?)
# Expand scope to include P0s from my direct reports.
candidates = self.dimensions['organizational'].expand_scope(user, candidates)
# Dimension 3: Service Ownership (What do we own?)
# Filter to services my team actually owns.
candidates = self.dimensions['service_ownership'].filter_by_ownership(user, candidates)
# Dimension 4: Dependencies (What breaks us?)
# Include critical issues from partners we depend on.
candidates = self.dimensions['dependencies'].include_dependencies(team_services, candidates)
# Dimension 5: Temporal Physics (Time Decay)
# Apply exponential decay. Old news fades away.
candidates = self.dimensions['temporal'].apply_temporal_weight(candidates, current_time)
# Dimension 6: Authority (Source of Truth)
# Jira > Slack. Logs > Email.
candidates = self.dimensions['authority'].apply_authority_weight(candidates)
return candidates
Visualizing the Subtraction
Here is what happens to the data when you apply these dimensional filters. We aren't searching; we are whittling.
The Results: 99% Noise Reduction
When I ran a simulation comparing this approach against standard Vector Search for the query "What pending items do I have on my plate?", the difference was stark.
Traditional RAG:
- Items Returned: 14 mixed items (Slack rumors, old bugs, other teams' tickets).
- Token Cost: ~8,000 tokens.
- Hallucination Risk: High (It might report the Slack rumor as a fact).
Multidimensional Graph:
- Items Returned: 4 highly relevant items.
- Token Cost: ~400 tokens.
- Hallucination Risk: Near Zero.
By the time the LLM sees the data, the graph has already done the heavy lifting. The prompt changes from "Read this mess and tell me what matters" to "Summarize these 4 confirmed critical facts."
The Startup Opportunity: The "Context Operating System"
This architecture reveals a massive gap in the current AI stack. Everyone is trying to build the "Brain" (the Agent/LLM). Very few are building the "Central Nervous System."
The real opportunity is to combine the Universal Signal Bus (which I wrote about previously) with this Multidimensional Knowledge Graph.
- The Bus (The Ears): Ingests wild signals from Logs, Audio, and Code.
- The Graph (The Memory): Applies physics (Time, Authority, Hierarchy) to filter that noise.
- The Agent (The Brain): Receives a clean, curated signal.
If you can build a managed service that accepts raw enterprise noise and outputs Graph-Curated Context, you are effectively building the operating system for Enterprise AI. You stop the Agent from hallucinating not by prompt engineering, but by strictly controlling what it's allowed to see.
Don't just scale your context window. Scale your filters.


Top comments (0)