<?xml version="1.0" encoding="UTF-8"?>
<rss version="2.0" xmlns:atom="http://www.w3.org/2005/Atom" xmlns:dc="http://purl.org/dc/elements/1.1/">
  <channel>
    <title>DEV Community: Tyler</title>
    <description>The latest articles on DEV Community by Tyler (@tylerloveamber).</description>
    <link>https://dev.to/tylerloveamber</link>
    <image>
      <url>https://media2.dev.to/dynamic/image/width=90,height=90,fit=cover,gravity=auto,format=auto/https:%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Fuser%2Fprofile_image%2F3862245%2F389121e1-79c5-496b-a055-8e604259743a.png</url>
      <title>DEV Community: Tyler</title>
      <link>https://dev.to/tylerloveamber</link>
    </image>
    <atom:link rel="self" type="application/rss+xml" href="https://dev.to/feed/tylerloveamber"/>
    <language>en</language>
    <item>
      <title>Claude Code Leaked. Here's What It Means for Your Team's Security Policy.</title>
      <dc:creator>Tyler</dc:creator>
      <pubDate>Sun, 05 Apr 2026 14:59:56 +0000</pubDate>
      <link>https://dev.to/tylerloveamber/claude-code-leaked-heres-what-it-means-for-your-teams-security-policy-115i</link>
      <guid>https://dev.to/tylerloveamber/claude-code-leaked-heres-what-it-means-for-your-teams-security-policy-115i</guid>
      <description>&lt;p&gt;In April 2026, Anthropic accidentally published the full source code of Claude Code — roughly 512,000 lines of TypeScript — inside an npm package update. A researcher found it within hours and posted it publicly on GitHub.&lt;/p&gt;

&lt;p&gt;If your team uses Claude Code, here's what you need to know — without the technical jargon.&lt;/p&gt;




&lt;h2&gt;
  
  
  What actually leaked
&lt;/h2&gt;

&lt;p&gt;&lt;strong&gt;The short version: Anthropic's own internal code. Not your data.&lt;/strong&gt;&lt;/p&gt;

&lt;p&gt;Anthropic accidentally included a &lt;code&gt;.map&lt;/code&gt; file that contained every line of the original source code. Think of it like accidentally shipping your product with all your internal engineering notes printed on the back of the box.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;What was exposed:&lt;/strong&gt;&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;How Claude Code works internally — the instructions it follows, how it decides what to do&lt;/li&gt;
&lt;li&gt;Features your team didn't know existed — including a memory system, multi-agent coordination, and an autonomous permissions mode internally called "YOLO classifier"&lt;/li&gt;
&lt;li&gt;Security mechanisms Anthropic built to protect API credentials on your machine&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;&lt;strong&gt;What was NOT exposed:&lt;/strong&gt;&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;Your company's code&lt;/li&gt;
&lt;li&gt;Your API keys or conversation history&lt;/li&gt;
&lt;li&gt;Anthropic's AI models or training data&lt;/li&gt;
&lt;/ul&gt;




&lt;h2&gt;
  
  
  Why it matters for your security policy
&lt;/h2&gt;

&lt;h3&gt;
  
  
  1. Your vendor's operational security is part of your risk
&lt;/h3&gt;

&lt;p&gt;Shipping a source map in a production npm package is a basic release hygiene mistake. The kind a pre-publish checklist or CI check catches automatically.&lt;/p&gt;

&lt;p&gt;When you approve an AI tool for your team, you're implicitly trusting the vendor's internal processes. This is a data point about Anthropic's release process maturity.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;What to add to your AI tool policy:&lt;/strong&gt; Ask vendors — have they had prior security incidents? Do they have a disclosed security program? How do they notify customers when something goes wrong?&lt;/p&gt;

&lt;h3&gt;
  
  
  2. There are features in the tool your team didn't approve
&lt;/h3&gt;

&lt;p&gt;The leaked code reveals production-ready features hidden behind feature flags — not visible, but running on your developers' machines:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;An autonomous permissions mode that makes decisions without asking the user&lt;/li&gt;
&lt;li&gt;A memory system that persists information across sessions&lt;/li&gt;
&lt;li&gt;Multi-agent coordination&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;Your AI acceptable use policy was written against features you knew about. If these get enabled quietly in a future update, your policy doesn't cover them.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;What to add:&lt;/strong&gt; Review AI tools quarterly for new default features. Add a clause requiring vendor notification before significant new capabilities are activated.&lt;/p&gt;

&lt;h3&gt;
  
  
  3. You need a process for third-party AI tool incidents
&lt;/h3&gt;

&lt;p&gt;Most small teams don't have this. When a tool your team uses has a security incident — even one that doesn't affect your data — you need to be able to answer quickly:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;Does this change our threat model?&lt;/li&gt;
&lt;li&gt;Do we need to rotate credentials?&lt;/li&gt;
&lt;li&gt;Do we need to notify clients?&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;For this incident, the answer is mostly "no immediate action, but monitor." But having the process matters.&lt;/p&gt;




&lt;h2&gt;
  
  
  What to do right now (30 minutes)
&lt;/h2&gt;

&lt;p&gt;&lt;strong&gt;If your team uses Claude Code:&lt;/strong&gt;&lt;/p&gt;

&lt;ol&gt;
&lt;li&gt;
&lt;strong&gt;Rotate your Anthropic API keys&lt;/strong&gt; — 5 minutes in the Anthropic console. Not strictly required, but good practice after any vendor incident.&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;Check audit logs&lt;/strong&gt; — do you have a record of what Claude Code accessed and when?&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;Brief your tech lead&lt;/strong&gt; — make sure they've reviewed Anthropic's official response.&lt;/li&gt;
&lt;/ol&gt;

&lt;p&gt;&lt;strong&gt;If your team uses other AI coding tools (Cursor, Copilot, etc.):&lt;/strong&gt;&lt;/p&gt;

&lt;p&gt;Use this as a trigger to run the same questions for every AI tool. I built a checklist for exactly this: &lt;a href="https://www.aipolicydesk.com/blog/ceo-ai-tool-approval-checklist" rel="noopener noreferrer"&gt;CEO AI Tool Approval Checklist&lt;/a&gt;&lt;/p&gt;




&lt;h2&gt;
  
  
  The broader lesson
&lt;/h2&gt;

&lt;p&gt;AI tools are software. Software vendors make operational mistakes.&lt;/p&gt;

&lt;p&gt;The question isn't "is this AI tool perfectly secure?" — nothing is. The question is: &lt;strong&gt;do you have enough visibility and process to respond when something goes wrong?&lt;/strong&gt;&lt;/p&gt;

&lt;p&gt;The Claude Code leak is a low-severity incident with a high-visibility lesson.&lt;/p&gt;




&lt;p&gt;How does your team handle third-party AI tool incidents? Do you have a process, or would this have caught you off guard?&lt;/p&gt;

</description>
      <category>security</category>
      <category>startup</category>
      <category>devtools</category>
      <category>claudecode</category>
    </item>
    <item>
      <title>CEO Checklist: 10 Questions Before Approving Cursor, ChatGPT, or Claude for Your Team</title>
      <dc:creator>Tyler</dc:creator>
      <pubDate>Sun, 05 Apr 2026 12:58:19 +0000</pubDate>
      <link>https://dev.to/tylerloveamber/ceo-checklist-10-questions-before-approving-cursor-chatgpt-or-claude-for-your-team-22d8</link>
      <guid>https://dev.to/tylerloveamber/ceo-checklist-10-questions-before-approving-cursor-chatgpt-or-claude-for-your-team-22d8</guid>
      <description>&lt;p&gt;I'm a PM. My CEO kept asking: &lt;em&gt;"Are these AI tools actually safe?"&lt;/em&gt;&lt;/p&gt;

&lt;p&gt;I didn't have a good answer. So I went and built one.&lt;/p&gt;

&lt;p&gt;Here's what I found — and what most startups get wrong.&lt;/p&gt;




&lt;h2&gt;
  
  
  The core problem
&lt;/h2&gt;

&lt;p&gt;Most teams adopt AI coding tools like this:&lt;/p&gt;

&lt;ol&gt;
&lt;li&gt;Dev asks "can I use Cursor?"&lt;/li&gt;
&lt;li&gt;CEO Googles it briefly&lt;/li&gt;
&lt;li&gt;CEO says yes&lt;/li&gt;
&lt;/ol&gt;

&lt;p&gt;No one checks training defaults. No one verifies whether source code leaves the environment. No one sets up audit logs.&lt;/p&gt;

&lt;p&gt;Then something goes wrong and there's no paper trail.&lt;/p&gt;




&lt;h2&gt;
  
  
  10 questions to run before you approve any AI tool
&lt;/h2&gt;

&lt;h3&gt;
  
  
  1. Is this tool training on our code?
&lt;/h3&gt;

&lt;div class="table-wrapper-paragraph"&gt;&lt;table&gt;
&lt;thead&gt;
&lt;tr&gt;
&lt;th&gt;Tool&lt;/th&gt;
&lt;th&gt;Default&lt;/th&gt;
&lt;th&gt;How to opt out&lt;/th&gt;
&lt;/tr&gt;
&lt;/thead&gt;
&lt;tbody&gt;
&lt;tr&gt;
&lt;td&gt;Cursor Personal&lt;/td&gt;
&lt;td&gt;ON&lt;/td&gt;
&lt;td&gt;Upgrade to Business&lt;/td&gt;
&lt;/tr&gt;
&lt;tr&gt;
&lt;td&gt;ChatGPT Free/Plus&lt;/td&gt;
&lt;td&gt;ON&lt;/td&gt;
&lt;td&gt;Settings → Data Controls → toggle off&lt;/td&gt;
&lt;/tr&gt;
&lt;tr&gt;
&lt;td&gt;Claude API / Team&lt;/td&gt;
&lt;td&gt;OFF&lt;/td&gt;
&lt;td&gt;Already off&lt;/td&gt;
&lt;/tr&gt;
&lt;tr&gt;
&lt;td&gt;Notion AI&lt;/td&gt;
&lt;td&gt;OFF&lt;/td&gt;
&lt;td&gt;Per policy&lt;/td&gt;
&lt;/tr&gt;
&lt;/tbody&gt;
&lt;/table&gt;&lt;/div&gt;

&lt;p&gt;&lt;strong&gt;Action:&lt;/strong&gt; Verify each tool. Screenshot the setting. Save it in a doc.&lt;/p&gt;




&lt;h3&gt;
  
  
  2. Does source code leave our environment?
&lt;/h3&gt;

&lt;p&gt;Cursor and Copilot send code context to servers on every completion. That's how they work — there's no offline mode.&lt;/p&gt;

&lt;p&gt;ChatGPT: only if your dev manually pastes code into the chat.&lt;/p&gt;

&lt;p&gt;This matters even if training is OFF. If the vendor gets breached, your code is exposed.&lt;/p&gt;




&lt;h3&gt;
  
  
  3. How long is data retained?
&lt;/h3&gt;

&lt;ul&gt;
&lt;li&gt;
&lt;strong&gt;Cursor:&lt;/strong&gt; ~30 days, deletion available on request&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;ChatGPT:&lt;/strong&gt; until you delete the conversation&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;Claude API:&lt;/strong&gt; not retained after the request completes&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;If you can't find "Data Retention" in their Privacy Policy in 5 minutes — that's a red flag.&lt;/p&gt;




&lt;h3&gt;
  
  
  4. If there's a breach, how fast do they notify you?
&lt;/h3&gt;

&lt;p&gt;90% of CEOs never ask this. GDPR requires 72 hours — but only if you're in scope.&lt;/p&gt;

&lt;p&gt;Search "incident notification" in their Terms of Service. No clause = no contractual obligation to tell you anything.&lt;/p&gt;




&lt;h3&gt;
  
  
  5. Are your devs using personal accounts for work?
&lt;/h3&gt;

&lt;p&gt;Personal ChatGPT free = training ON, no audit logs, no way to revoke access when they leave.&lt;/p&gt;

&lt;p&gt;This is the most common problem. Most teams have it and don't know.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;Fix:&lt;/strong&gt; Mandate company plans. No personal accounts for work AI tools.&lt;/p&gt;




&lt;h3&gt;
  
  
  6. Do you have audit logs?
&lt;/h3&gt;

&lt;ul&gt;
&lt;li&gt;
&lt;strong&gt;Personal/Free plans:&lt;/strong&gt; No logs&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;Team/Business plans:&lt;/strong&gt; Basic logs&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;Enterprise plans:&lt;/strong&gt; Full logs&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;No logs = no way to reconstruct what happened during an incident.&lt;/p&gt;




&lt;h3&gt;
  
  
  7. Can you delete your data if you leave?
&lt;/h3&gt;

&lt;p&gt;Test this before committing. Create an account → use it → request deletion → see if they confirm clearly.&lt;/p&gt;

&lt;p&gt;Some vendors confirm in days. Others are vague. If you can't get written confirmation — assume the data stays forever.&lt;/p&gt;




&lt;h3&gt;
  
  
  8. Do they have SOC 2?
&lt;/h3&gt;

&lt;ul&gt;
&lt;li&gt;
&lt;strong&gt;Cursor:&lt;/strong&gt; Yes (SOC 2 Type II)&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;ChatGPT / OpenAI:&lt;/strong&gt; Yes&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;Claude / Anthropic:&lt;/strong&gt; Yes&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;Notion:&lt;/strong&gt; Yes&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;No certification = you're trusting their self-assessment, not an external audit. Not automatically a dealbreaker — but you should know.&lt;/p&gt;




&lt;h3&gt;
  
  
  9. Who owns the AI-generated output?
&lt;/h3&gt;

&lt;p&gt;Cursor, ChatGPT, Claude: you own the output per current Terms of Service.&lt;/p&gt;

&lt;p&gt;Unresolved edge case: if two companies generate nearly identical code from the same prompt — who owns it? No clear case law yet.&lt;/p&gt;




&lt;h3&gt;
  
  
  10. Does your team know any of this?
&lt;/h3&gt;

&lt;p&gt;The best policy is useless if only the CEO has read it.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;Fix:&lt;/strong&gt; 30-minute team briefing. One doc listing approved tools, prohibited tools, and required account type. Add it to your onboarding checklist.&lt;/p&gt;




&lt;h2&gt;
  
  
  What to document after the checklist
&lt;/h2&gt;

&lt;p&gt;For each approved tool, record:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;Tool name&lt;/li&gt;
&lt;li&gt;Plan/tier (Personal, Team, Business, Enterprise)&lt;/li&gt;
&lt;li&gt;Whether training is OFF&lt;/li&gt;
&lt;li&gt;Whether audit logs are available&lt;/li&gt;
&lt;li&gt;Date last reviewed&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;Keep this in a shared doc and review every 6 months or when a vendor changes their policy.&lt;/p&gt;




&lt;p&gt;Full version with FAQ and more detail: &lt;a href="https://www.aipolicydesk.com/blog/ceo-ai-tool-approval-checklist" rel="noopener noreferrer"&gt;aipolicydesk.com/blog/ceo-ai-tool-approval-checklist&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;How does your team handle AI tool approvals? Is there a process, or is it mostly "dev asks, CEO says yes"?&lt;/p&gt;

</description>
      <category>security</category>
      <category>startup</category>
      <category>devtools</category>
      <category>productivity</category>
    </item>
  </channel>
</rss>
