Originally published on Truthlocks Blog
Think about how you onboard a new hire. On day one, they get a badge, a laptop, and access to the tools they need for their specific role. They do not get the root password to the production database. Over time, as they prove themselves reliable and competent, their access expands. If they make a serious mistake, access gets dialed back. This is common sense for humans. For AI agents, we have been doing the opposite.
Most organizations deploying AI agents today hand them a shared API key and let them loose. The agent that was spun up five minutes ago to test a new prompt template gets the same level of access as the agent that has been running flawlessly in production for six months. There is no mechanism to distinguish between them. There is no way to say "this agent has earned our trust" or "this agent is brand new and should be on a short leash."
That is what trust scores fix.
What a Trust Score Actually Is
A trust score is a number between 0 and 100 that represents how much confidence you should place in a specific AI agent. It is not a static label assigned at registration. It is a living metric that goes up when the agent behaves well and goes down when it does not.
The score is computed from five signals that we call trust factors:
Behavioral compliance measures whether the agent stays within its expected operating patterns. An agent that consistently completes its assigned tasks without errors or unexpected actions scores high. An agent that starts making API calls it has never made before, or that suddenly increases its request volume tenfold, scores low.
Scope adherence tracks whether the agent respects the boundaries it was given. Every agent in the Truthlocks system has a defined set of scopes that describe what it is allowed to do. An agent authorized for customers:read that never attempts a write operation scores high. An agent that repeatedly tries to access resources outside its scopes scores low, and those attempts are logged.
Anomaly score is the inverse of how unusual the agent's recent behavior looks compared to its historical baseline. Machine learning models analyze request patterns, timing, payload structures, and resource access sequences. The more normal the behavior, the higher the score.
Peer attestations captures trust signals from other systems and agents. If an agent's outputs are consistently accepted and acted upon by downstream systems without errors or rejections, that is a positive signal. If downstream systems are frequently rejecting or rolling back the agent's work, that is a negative signal.
Session hygiene evaluates whether the agent manages its sessions properly. Does it authenticate cleanly? Does it respect session timeouts? Does it request only the scopes it needs rather than asking for everything? Good session management indicates a well engineered agent.
How Trust Scores Change Real Decisions
Trust scores are not just a dashboard metric. They are an input to authorization decisions that happen in real time.
Consider a financial services company that uses AI agents to process customer data. They can set a policy that says: any agent with a trust score below 60 can only access anonymized data. Agents scoring between 60 and 80 can access full customer records but cannot make changes. Only agents scoring above 80 can modify customer data.
This means a newly registered agent starts with restricted access. As it operates cleanly over days and weeks, its trust score rises and its capabilities expand automatically. If something goes wrong and the agent starts behaving erratically, the score drops and access is immediately constrained. No human intervention required.
This is not theoretical. This is how the Truthlocks trust score system works today.
The Kill Switch Connection
Trust scores also connect to the kill switch. Organizations can set automated policies: if any agent's trust score drops below 20, revoke its identity immediately. All active sessions are terminated, all tokens are invalidated, and a revocation event is broadcast to every connected system. The agent is effectively shut down in seconds.
This is critical for the scenario every security team worries about: a compromised agent. Whether the compromise comes from a prompt injection attack, a stolen key, or a bug that causes the agent to go haywire, the trust score will detect the abnormal behavior and the kill switch can activate automatically.
Building Trust Into Your Agent Architecture
If you are building AI agents today, start thinking about trust as a first class architectural concern. Register every agent with a unique identity. Define explicit scopes for what each agent is allowed to do. Monitor behavioral signals and use them to gate access. Have a plan for what happens when an agent goes wrong.
The agents you deploy today will multiply. The access they have today will expand. The damage they can do if things go wrong will grow. Trust scores give you a systematic way to manage that risk as your agent fleet scales.
To start building with trust scores, visit the Machine Identity documentation or sign in to the Truthlocks Console to register your first agent.
Truthlocks provides machine identity infrastructure for AI agents.
Top comments (0)