<?xml version="1.0" encoding="UTF-8"?>
<rss version="2.0" xmlns:atom="http://www.w3.org/2005/Atom" xmlns:dc="http://purl.org/dc/elements/1.1/">
  <channel>
    <title>DEV Community: Cara Jung</title>
    <description>The latest articles on DEV Community by Cara Jung (@carasjung).</description>
    <link>https://dev.to/carasjung</link>
    <image>
      <url>https://media2.dev.to/dynamic/image/width=90,height=90,fit=cover,gravity=auto,format=auto/https:%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Fuser%2Fprofile_image%2F3448967%2F4a49be6a-32ba-4490-aa57-3617e659db09.png</url>
      <title>DEV Community: Cara Jung</title>
      <link>https://dev.to/carasjung</link>
    </image>
    <atom:link rel="self" type="application/rss+xml" href="https://dev.to/feed/carasjung"/>
    <language>en</language>
    <item>
      <title>What Predicts a Hit? I Trained 3 ML Models to Find Out</title>
      <dc:creator>Cara Jung</dc:creator>
      <pubDate>Mon, 06 Apr 2026 07:00:00 +0000</pubDate>
      <link>https://dev.to/carasjung/what-predicts-a-hit-i-trained-3-ml-models-to-find-out-31mj</link>
      <guid>https://dev.to/carasjung/what-predicts-a-hit-i-trained-3-ml-models-to-find-out-31mj</guid>
      <description>&lt;p&gt;In many entertainment adaptation decisions, content selections are still instinct-driven. Maybe a producer was vibing with a story or overheard their Gen Alpha nephew mentioning a GOAT title. This subjective approach has often led to expensive missteps and wasted resources for studios when the feature or show turns into a flop. &lt;/p&gt;

&lt;p&gt;As someone who has worked in the breeding ground of popular webcomics, I asked: what if there was a system that could measure “success potential” of IPs based on real user behavior? Using ML, I wanted to see if I could build a forecasting model that could rank unadapted titles by their predicted commercial success. &lt;/p&gt;

&lt;p&gt;&lt;strong&gt;The Data&lt;/strong&gt;&lt;/p&gt;

&lt;p&gt;For my endeavor, I worked with three datasets:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;Source material metadata of roughly 1,500 titles that included engagement metrics such as views, likes, subscribers, genre, release schedule, and creator usernames&lt;/li&gt;
&lt;li&gt;Produced show metadata of 1,977 titles including ratings, watcher counts, genre, episode count, and cast&lt;/li&gt;
&lt;li&gt;Historical webcomic adaptation records of 424 cross-referenced titles that went from source material to screen, with data pulled from both sides&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;Before any modeling, I ran exploratory data analysis on all three and found a few things:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;Engagement metrics (likes, views, subscribers) were strongly correlated with each other and overall popularity&lt;/li&gt;
&lt;li&gt;Genre and tags correlated with watcher counts in the produced show data&lt;/li&gt;
&lt;li&gt;Creator frequency showed no statistically significant impact on adaptation success, which directly contradicted what studios commonly assume&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fhirqnr4guj1s9aabpv0v.png" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fhirqnr4guj1s9aabpv0v.png" alt="Modeling Pipeline" width="800" height="853"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;Engineering the Target Variable&lt;/strong&gt;&lt;/p&gt;

&lt;p&gt;One hurdle I ran into was that I couldn't directly measure adaptation “success" from the source material side alone. So I engineered a composite Popularity Score by normalizing and combining views, likes, and subscribers into a single metric representing audience appeal, which became the target variable for prediction.&lt;/p&gt;

&lt;p&gt;For the produced show data, I created a parallel score using rating and watcher count.&lt;/p&gt;

&lt;p&gt;Since correlation analysis confirmed that source popularity and show popularity moved together in historical adaptations, I used source popularity as a proxy target.&lt;/p&gt;

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fp4zeu7ke58soe0baxo2z.png" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fp4zeu7ke58soe0baxo2z.png" alt="Close overlap between actual and predicted curves" width="800" height="467"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;Simple vs Complex Models&lt;/strong&gt;&lt;/p&gt;

&lt;p&gt;I implemented three models: Random Forest, XGBoost, and Ridge Regression. If you worked with ML models, there’s an expectation that the more complex models will win. However, this wasn’t the case. Ridge Regression became the unexpected underdog model that won:&lt;/p&gt;

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2F0dqerf5x3rq1u3tnxs0d.png" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2F0dqerf5x3rq1u3tnxs0d.png" alt="Cross-validation applied across all three models to reduce overfitting risk and validate stability on the adaptation dataset." width="800" height="245"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;I cross-validated all three models to reduce overfitting and validate stability.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;Likes = Success&lt;/strong&gt;&lt;/p&gt;

&lt;p&gt;Using standardized coefficients for feature importance in the Ridge model, the ranking was as follows:&lt;/p&gt;

&lt;ol&gt;
&lt;li&gt;Likes (strongest predictor by a significant margin)&lt;/li&gt;
&lt;li&gt;Views&lt;/li&gt;
&lt;li&gt;Subscribers&lt;/li&gt;
&lt;/ol&gt;

&lt;p&gt;The factors that studios often focus on such as creator reputation, genre, rating, and engagement rate showed weak or no statistical significance.&lt;/p&gt;

&lt;p&gt;I validated this further using Mann-Whitney U tests comparing adapted titles against the general pool. Adapted titles showed significantly higher “likes” than non-adapted ones and the difference was meaningful.&lt;/p&gt;

&lt;p&gt;Feature Importance for Ridge regression(standardized coefficients)&lt;/p&gt;

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2F8pnlaapyh3eq148fjh7m.png" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2F8pnlaapyh3eq148fjh7m.png" alt="Creator, genre, and rating showed no statistically significant impact and were excluded from the final model" width="800" height="442"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;So why “likes”? &lt;/p&gt;

&lt;p&gt;One interpretation is that likes are intentional. A view can be passive while a subscription can be habitual. But giving a “like” is an act of emotional investment and this behavior is exactly what translates from IP to screen.&lt;/p&gt;

&lt;p&gt;The Output&lt;/p&gt;

&lt;p&gt;The final model produced a ranked list of the top 10 unadapted webcomic titles by predicted success, along with contextual signals for each including genre appeal, subscriber trends, engagement consistency, and creator track record.&lt;/p&gt;

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2F6fj4jgpy0pa8lwvh94wa.png" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2F6fj4jgpy0pa8lwvh94wa.png" alt="Top unadapted titles" width="800" height="603"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;Qualitative review of the top 10 confirmed alignment with the engagement patterns seen in historically successful adaptations. Cliff's Delta calculations showed that the predicted top titles had significantly higher likes than past adaptations.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;Limitations on the Model&lt;/strong&gt;&lt;/p&gt;

&lt;p&gt;Part of doing good data work is being honest about the limitations. There were a few things that fell short:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;Small adaptation dataset. 424 entries is workable, but more data would reduce overfitting risk and better generalization.&lt;/li&gt;
&lt;li&gt;Proxy target variable. Using source popularity instead of actual show performance is a justified simplification, but it means the model can't fully capture real-world production quality, casting, or distribution reach.&lt;/li&gt;
&lt;li&gt;Categorical features dropped. Creator and genre have too many levels and their coefficients dominated the model without adding significance. Excluding them improved interpretability but at the cost of losing some nuance.&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;&lt;strong&gt;What I'd Do Next&lt;/strong&gt;&lt;/p&gt;

&lt;p&gt;If I extended this project, I'd rethink how signal is captured and focus on the following:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;Use NLP for deeper context

&lt;ul&gt;
&lt;li&gt;Synopsis embeddings or sentiment analysis on reader reviews could capture thematic richness that raw engagement metrics miss.&lt;/li&gt;
&lt;/ul&gt;


&lt;/li&gt;

&lt;li&gt;Take a hybrid ranking approach

&lt;ul&gt;
&lt;li&gt;Combining regression with a learning-to-rank algorithm could improve recommendation quality at the top of the list, where small differences actually matter.&lt;/li&gt;
&lt;/ul&gt;


&lt;/li&gt;

&lt;li&gt;Longitudinal validation

&lt;ul&gt;
&lt;li&gt;The real test is tracking what happens when predicted titles actually get produced. Building a feedback loop into the model would sharpen it over time.&lt;/li&gt;
&lt;/ul&gt;


&lt;/li&gt;

&lt;/ul&gt;

&lt;p&gt;&lt;strong&gt;Final Thoughts&lt;/strong&gt;&lt;/p&gt;

&lt;p&gt;The core insight here doesn’t only strictly apply to entertainment. It can apply to decisions that are being made by intuition or legacy practice. As the models showed, behavioral signals from real users outperform assumptions about what will succeed.&lt;/p&gt;

&lt;p&gt;Likes beat creator prestige. Engagement beat genre conventions. The audience’s preferences, not the ones from industry decision makers, predicted outcomes more reliably.&lt;/p&gt;

&lt;p&gt;Whether you're choosing which content to produce, which features to build, or which markets to enter, the same principle applies. The answers are within the data, but we often overlook the right signals. &lt;/p&gt;

</description>
      <category>machinelearning</category>
      <category>ai</category>
      <category>python</category>
      <category>datascience</category>
    </item>
  </channel>
</rss>
