<?xml version="1.0" encoding="UTF-8"?>
<rss version="2.0" xmlns:atom="http://www.w3.org/2005/Atom" xmlns:dc="http://purl.org/dc/elements/1.1/">
  <channel>
    <title>DEV Community: tylerbrennan</title>
    <description>The latest articles on DEV Community by tylerbrennan (@tylerbrennan).</description>
    <link>https://dev.to/tylerbrennan</link>
    <image>
      <url>https://media2.dev.to/dynamic/image/width=90,height=90,fit=cover,gravity=auto,format=auto/https:%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Fuser%2Fprofile_image%2F3917675%2Fe6cfaca6-c850-4a26-ae95-20e70f089ff6.png</url>
      <title>DEV Community: tylerbrennan</title>
      <link>https://dev.to/tylerbrennan</link>
    </image>
    <atom:link rel="self" type="application/rss+xml" href="https://dev.to/feed/tylerbrennan"/>
    <language>en</language>
    <item>
      <title>1 of 57 pages indexed. The other 56 are technically perfect. So why?</title>
      <dc:creator>tylerbrennan</dc:creator>
      <pubDate>Thu, 30 Apr 2026 00:00:00 +0000</pubDate>
      <link>https://dev.to/tylerbrennan/1-of-57-pages-indexed-the-other-56-are-technically-perfect-so-why-34lb</link>
      <guid>https://dev.to/tylerbrennan/1-of-57-pages-indexed-the-other-56-are-technically-perfect-so-why-34lb</guid>
      <description>&lt;p&gt;I caught instayolo.com on the drop in March 2025. 13 months later, 1 of 57 indexable pages on the site is in Google. The other 56 are sitting in two GSC buckets despite every technical signal being correct. The real bottleneck isn't crawlability or schema — it's the prior owner's reputation drag, and the only way out is one external high-trust link.&lt;/p&gt;

&lt;h2&gt;
  
  
  Caught a dropped domain in March 2025
&lt;/h2&gt;

&lt;p&gt;I caught instayolo.com on the drop in March 2025. The previous owner had let registration lapse. Short name, two real words, reads as a downloader. Bought, sat on it for a couple weeks, then started building.&lt;/p&gt;

&lt;h2&gt;
  
  
  1 of 57 indexed
&lt;/h2&gt;

&lt;p&gt;13 months later, 1 of 57 indexable pages on the site is in Google. Just the homepage.&lt;/p&gt;

&lt;p&gt;Of the 56 not indexed, 38 are in GSC's "Discovered – currently not indexed" bucket — Google saw the URL in the sitemap and decided not to bother crawling. 17 are in "Crawled – currently not indexed" — Google crawled the page, then quietly tossed the result. One straggler is "Page with redirect" (&lt;code&gt;/story-viewer&lt;/code&gt; → &lt;code&gt;/story-downloader&lt;/code&gt;, a 308 shipped after merging two near-duplicate URLs into one).&lt;/p&gt;

&lt;h2&gt;
  
  
  Technical state, by category
&lt;/h2&gt;

&lt;p&gt;Sitemap.xml gets served from a real &lt;code&gt;app/sitemap.ts&lt;/code&gt; route with &lt;code&gt;lastmod&lt;/code&gt; derived from &lt;code&gt;git log&lt;/code&gt; instead of the lazy &lt;code&gt;now&lt;/code&gt; value everybody ships. Robots.txt allows &lt;code&gt;/&lt;/code&gt;, disallows only &lt;code&gt;/api/&lt;/code&gt;. One H1 per page. JSON-LD spans Organization, WebSite (with SearchAction so sitelinks-searchbox is unlocked), SoftwareApplication, BreadcrumbList, Article, Person. Every block validates in Rich Results Test. hreflang &lt;code&gt;x-default&lt;/code&gt; everywhere. Canonical URLs absolute. X-Robots-Tag is never set to &lt;code&gt;noindex&lt;/code&gt; on an indexable host. Googlebot UA gets the same 200 and same byte size as default UA — no cloaking, no soft 404s.&lt;/p&gt;

&lt;p&gt;Two different LLM agents audited the stack independently. Both came back clean.&lt;/p&gt;

&lt;p&gt;So why.&lt;/p&gt;

&lt;h2&gt;
  
  
  The domain wasn't fresh
&lt;/h2&gt;

&lt;p&gt;The previous owner had run a Shopify-flavored e-commerce site under instayolo.com — &lt;code&gt;/products/...&lt;/code&gt;, &lt;code&gt;/collections/...&lt;/code&gt;, &lt;code&gt;/cart&lt;/code&gt;, &lt;code&gt;/account/login&lt;/code&gt; — and every one of those was cached in Google's index when I took ownership. Search "instayolo" today and the brand entity in Google's knowledge graph is partially mine, partially the prior owner's. Google's quality classifier is still working out what this domain &lt;em&gt;is&lt;/em&gt;, given the gap between its memory of the domain and what my actual content says it does.&lt;/p&gt;

&lt;p&gt;This is the part SEO writeups skip. Fresh domains are predictable — sandbox, ride it out, build links, climb. &lt;em&gt;Inherited&lt;/em&gt; domains with an active prior-owner footprint are different. The crawl scheduler treats you as worse than new: not unproven, but established-and-radically-changed. Authority signals from the prior site decay slowly, but they decay onto &lt;em&gt;your&lt;/em&gt; pages now. The quality classifier eyes your content sideways for not matching the brand's history. Nobody warned me.&lt;/p&gt;

&lt;p&gt;A week of compressed work follows.&lt;/p&gt;

&lt;h2&gt;
  
  
  Sweep the ghost URLs
&lt;/h2&gt;

&lt;p&gt;GSC URL Removals tool, Directory mode (collapses 12 paths into 4 prefix patterns). Same patterns into Bing Webmaster's Block URLs. Removals buys roughly 6 months of suppression. Without it, plan on Google taking 12+ months to drop the prior owner's URLs, with ghost results haunting your brand SERP the whole time.&lt;/p&gt;

&lt;h2&gt;
  
  
  Eliminate every false signal
&lt;/h2&gt;

&lt;p&gt;Tedious but earns the keep. The Article schema's &lt;code&gt;publisher.logo&lt;/code&gt; pointed at &lt;code&gt;/icon-512.png&lt;/code&gt; — that URL 404'd, because Next.js App Router generates &lt;code&gt;/icon&lt;/code&gt; dynamically without the &lt;code&gt;-512.png&lt;/code&gt; suffix. Fixed to the real route, explicit 256×256 dimensions. Nginx was returning &lt;code&gt;&amp;lt;h1&amp;gt;301 Moved Permanently&amp;lt;/h1&amp;gt;&lt;/code&gt; as the response body on www→apex redirects. Google ignores those bodies. Bing was treating it as page content, and reporting the www host as "missing meta description." Added &lt;code&gt;X-Robots-Tag: noindex, follow&lt;/code&gt; on the www server block — tells engines to never index www separately and consolidate everything into apex. 38 &lt;code&gt;meta_title&lt;/code&gt; strings rendered over 60 chars in real SERPs; trimmed. 12 &lt;code&gt;meta_description&lt;/code&gt; strings over 160; trimmed.&lt;/p&gt;

&lt;h2&gt;
  
  
  Performance
&lt;/h2&gt;

&lt;p&gt;PSI mobile was Perf 78, TBT 530ms. The TBT was Google Tag Manager loading via &lt;code&gt;afterInteractive&lt;/code&gt;. Switched it to &lt;code&gt;lazyOnload&lt;/code&gt;. Perf 78 → 98, TBT 530 → 56ms. Then a Cloudflare Cache Rule for HTML — Cloudflare's default does not cache HTML, which I had assumed it did. (You have to opt in, via a Page Rule or Cache Rule.) TTFB 1.6s → 0.64s.&lt;/p&gt;

&lt;h2&gt;
  
  
  Author + entity hookup
&lt;/h2&gt;

&lt;p&gt;Built &lt;code&gt;/authors/torrance&lt;/code&gt; with a proper Person schema. Switched every Article schema's &lt;code&gt;author.url&lt;/code&gt; from &lt;code&gt;/about&lt;/code&gt; to &lt;code&gt;/authors/{slug}&lt;/code&gt;. Added an "About the author" block at the foot of every blog post — template change, propagates to all 16 posts at once, no manual loop. Cross-linked old posts to newer ones. The "Crawled – not indexed" bucket was almost entirely older posts with no inbound internal links from fresh content, so the working theory is Google saw orphans and dropped them.&lt;/p&gt;

&lt;p&gt;Indexed count after all of that: still 1.&lt;/p&gt;

&lt;h2&gt;
  
  
  What nobody warned me
&lt;/h2&gt;

&lt;p&gt;What nobody warned me about, and the reason this post exists, is that &lt;strong&gt;none of the technical work moves crawl budget on a reputation-drag domain.&lt;/strong&gt; What moves it is one external high-trust link. Technical work matters for &lt;em&gt;quality&lt;/em&gt; once Google decides to crawl. It doesn't move &lt;em&gt;whether&lt;/em&gt; Google decides to crawl. Authority does that. Specifically, authority that is observably new and attached to current content. A single backlink from a domain Google trusts probably breaks the purgatory faster than two more weeks of internal optimization.&lt;/p&gt;

&lt;p&gt;Hence this post.&lt;/p&gt;

&lt;p&gt;If you've inherited a dropped domain and Google is still stuck on the prior owner, feel free to reach out — receipts on the small fixes (commits, before/after PSI, GSC screenshots) are happy to share. Compare notes welcome.&lt;/p&gt;

&lt;p&gt;Indexed pages: 1.&lt;/p&gt;

</description>
      <category>seo</category>
      <category>indexation</category>
      <category>sitenotes</category>
      <category>google</category>
    </item>
  </channel>
</rss>
