DEV Community

Cover image for Time-Series LLMs: The $50k Health Intelligence Gap
Dr Hernani Costa
Dr Hernani Costa

Posted on • Originally published at radar.firstaimovers.com

Time-Series LLMs: The $50k Health Intelligence Gap

Your wearables are generating 10,000 data points daily. Your doctor sees one number per quarter. That gap is costing you—and your organization—millions in preventable health crises.

Time-series LLMs are AIs that learn from your body's data over time, not just single snapshots. Think of it this way: your body isn't a photograph—it's a Netflix series. Wearables and lab tests aren't random images; they're episodes and scenes unfolding across days, weeks, and months. Time-series LLMs are models trained to understand the entire show, spotting patterns, character arcs, and plot twists that you'd miss if you only looked at one frame.

I'll walk you through this step by step, no PhD required. By the end, you'll understand how these systems work and how to start building with them.

1. What is "Time-Series" and Why Does Health Care About It?

Time-series = data that changes over time.

In health tech, that looks like:

  • Your heart rate every minute from a chest strap or watch
  • Your sleep stages every 30-second epoch throughout the night
  • Your glucose reading every 5 minutes from a continuous glucose monitor (CGM)
  • Your lab tests every few months: cholesterol, HbA1c, vitamin D, inflammation markers

Why Single Points Are Almost Useless

When your doctor says, "Your HRV is 42 ms," that number tells you almost nothing without context.

But if you see:

  • Last month: HRV averaged 55 ms
  • Two weeks ago: HRV dropped to 50 ms
  • This week: HRV is now 42 ms
  • Alongside: Your sleep quality degraded from 85% to 68%, and you started waking 3 times per night instead of once

Now you have a story. You're trending toward overtraining, chronic stress, or early illness.

What Time-Series Models Detect

Time-series models are built to see patterns like:

  • Trends: Your resting heart rate has slowly climbed 8 bpm over 3 weeks—maybe you're getting sick or more stressed
  • Seasonality: Your glucose always spikes after dinner, but not breakfast
  • Anomalies: Last night's HRV was 30% lower than your 90-day baseline for no obvious reason
  • Correlations: When your bedtime drifts later, your HRV drops and your next-day glucose variability increases

This is longitudinal intelligence—understanding how your body behaves across time, not just in a moment.

2. What's an LLM Doing in the Middle of All This?

Traditional machine learning models excel at numeric predictions: "You have a 73% risk of metabolic syndrome based on these 12 lab values."

LLMs (Large Language Models) excel at:

  • Understanding and generating human language
  • Explaining complex patterns in plain English
  • Following instructions like "talk to me like I'm a college athlete" or "give me 3 actionable steps"
  • Reasoning over context—connecting dots across multiple data sources

The Magic Combo: Numbers + Language

In health tech, the breakthrough is combining both:

  1. A numeric model (or a time-series-aware LLM) processes your raw physiological curves: heart rate, HRV, sleep architecture, glucose, activity, temperature
  2. An LLM translates that into coaching language you actually understand:

"Over the past 7 days, your bedtime drifted 90 minutes later, your HRV dropped 18%, and your glucose swings got bigger. That pattern usually means your nervous system is under stress. Let's focus on sleep regularity for the next 3 days—aim for lights out by 10:30 PM."

The LLM is your health translator + coach, sitting on top of the raw data intelligence.

3. Time-Series LLMs (Health-LLM, OpenTSLM, PH-LLM): What Makes Them Special?

Classic LLMs like GPT-4 or Claude are trained mostly on text—books, articles, conversations.

Time-series LLMs are architecturally adapted to also "read" sequences of numbers, like:

[70, 72, 75, 90, 110, 130, 145, 120, 100, 80]

This could represent:

  • Heart rate during a 10-minute workout
  • Glucose after eating a bagel
  • HRV sampled every hour overnight

How They Work Under the Hood

Models like Health-LLM, OpenTSLM, PH-LLM, and MedTsLLM do several clever things:

  1. Tokenize time-series data—convert numeric sequences into "tokens" (like words, but for numbers) that the LLM can process
  2. Mix numeric patterns + text context in one unified model—the same architecture that reads "Your HRV dropped" can also read [55, 50, 42] directly
  3. Learn health-specific tasks through fine-tuning: sleep staging, arrhythmia detection, anomaly flagging, glucose forecasting, fatigue prediction

What This Feels Like in Practice

From a user's perspective, you get an AI that can:

  • Analyze 7–30 days of continuous wearable data
  • Spot subtle patterns humans would miss (e.g., "Your HRV drops every Thursday night—what's different on Thursdays?")
  • Explain those patterns in natural language
  • Provide contextual recommendations that account for your recent trends

It's like having a very patient, obsessive health nerd reading your wrist data 24/7 and summarizing it into actionable insights.

Why OpenTSLM is a Big Deal

OpenTSLM, developed at Stanford, is particularly interesting because it integrates temporal sensor data directly into LLM reasoning. You can ask it:

"What patterns do you see in my ECG that might explain my symptoms?"

And it doesn't just spit out a diagnosis—it provides detailed reasoning about temporal patterns it observed, explains how they relate to your question, and contextualizes findings with your patient-specific data. Cardiologists rated OpenTSLM's reasoning as correct or partially correct in 92.9% of cases, with particularly strong clinical context integration.

This is huge: it's not just accuracy—it's interpretability.

4. RAG vs Fine-Tuning: Two Ways to Make LLMs Smart About Health

When building a health AI readiness assessment, you have two core strategies:

RAG (Retrieval-Augmented Generation)

What it is: The LLM is a generalist. When you ask a question, it retrieves relevant information from an external knowledge base—medical guidelines, research papers, your past health records—and uses that context to answer.

Example workflow:

  1. User asks: "What does low HRV mean?"
  2. System retrieves: Relevant paragraphs from UpToDate, Mayo Clinic, recent research on autonomic function
  3. LLM synthesizes: "Low HRV typically indicates reduced autonomic flexibility, often associated with stress, overtraining, or illness..."

Best for:

  • Education and explanations—"What is insulin resistance?"
  • Latest medical guidelines—"What's the updated CDC recommendation on HbA1c?"
  • Contextual lookups—"Show me studies on magnesium and sleep quality"

Fine-Tuning

What it is: You actually retrain the LLM on thousands of health-specific examples, so it learns patterns like:

"When HRV drops + sleep degrades + glucose variability increases → suggest a recovery day and explain physiological reasoning."

The model internalizes these cause-effect patterns and develops a "health coaching personality".

Example workflow:

  1. You train the model on 10,000 examples of {wearable data → expert coach response}
  2. User's data shows: HRV down, sleep fragmented, glucose erratic
  3. Model generates: "Your body is showing clear signs of overload. Let's prioritize: (a) Fix a consistent bedtime this week, (b) Add one short walk after dinner, (c) Delay intense workouts until your HRV climbs back toward baseline."

Best for:

  • Pattern detection over your personal timeline
  • Personalized behavior coaching
  • Consistent tone and style
  • Judgement calls based on multi-signal integration

Rule of Thumb

Use RAG for "knowledge about health"—definitions, guidelines, research.

Use fine-tuning for "judgement over your data over time"—pattern recognition, personalized coaching, longitudinal recommendations.

Many production systems use both: RAG for educational content, fine-tuned models for personalized insights.

5. How Wearables + Lab Tests Come Together

Think of your health data stack in three layers:

Layer 1: Raw Data (Sensors + Tests + Self-Reports)

  • Wearables: heart rate, HRV, steps, sleep stages, body temperature, SpO₂, respiratory rate
  • Lab tests: HbA1c (average glucose), lipid panel, vitamin levels, inflammation markers (CRP, homocysteine), hormone levels
  • Self-reports: mood, perceived stress, pain levels, energy, food intake, menstrual cycle

Layer 2: Timeline Construction (Episodes, Not Random Points)

Instead of dumping raw timestamps into the model, you compress them into human-readable summaries:

Example 14-day summary:

  • Bedtime drifted 1.5 hours later (from 10:15 PM → 11:45 PM average)
  • Average HRV dropped from 55 ms → 40 ms
  • Wake-ups per night increased from 1.2 → 3.1
  • Morning fasting glucose higher on 9 out of 14 days
  • Lab context: HbA1c borderline high (5.9%), vitamin D low (22 ng/mL)

This becomes the input context for your model.

Layer 3: Model Intelligence (Time-Series LLM + Coach LLM)

  1. A time-series model or time-series LLM handles the pattern math—forecasting, anomaly detection, trend analysis
  2. A fine-tuned health coach LLM turns it into actionable guidance:

"Your body is showing clear signs of overload. Here's the plan:
(a) Fix a consistent bedtime this week—aim for 10:30 PM ±15 minutes
(b) Add one 15-minute walk after dinner to stabilize evening glucose
(c) Delay high-intensity workouts until your HRV climbs back toward 50 ms
(d) Consider vitamin D supplementation—discuss with your doctor"

Why Lab Tests + Wearables Are Better Together

  • Lab tests provide slow, deep markers—months of metabolic behavior condensed into one blood draw
  • Wearables provide fast, continuous signals—daily fluctuations in autonomic tone, sleep, activity
  • The LLM is the brain that synthesizes both timescales into coherent, personalized recommendations

6. How to Start Learning This as a Beginner

If you want to get hands-on, here's a realistic, Karpathy-style learning path—building from simple to complex.

Step 1: Get Comfortable with the Basics

Learn Python fundamentals:

  • Variables, loops, functions, lists, dictionaries
  • File I/O, string manipulation

Learn basic data handling with pandas:

  • Read CSVs
  • Filter rows, select columns
  • Plot simple graphs with matplotlib or plotly

Time investment: 2–4 weeks if you code 1–2 hours daily.

Step 2: Play with Your Own (or Sample) Time-Series Data

Export data from a wearable (Garmin, Oura, Fitbit, Apple Health, Whoop) or use open datasets from health research (MIMIC-III, PhysioNet).

Do simple experiments:

  1. Plot your resting heart rate over time
  2. Plot HRV vs. bedtime
  3. Plot glucose (if available) before vs. after meals
  4. Calculate 7-day rolling averages
  5. Flag days where HRV dropped >20% from your baseline

What you're learning: Developing intuition for time-series—what trends look like, what noise looks like, what anomalies feel like.

Step 3: Learn "Classic" Time-Series Tools

Before touching LLMs, understand the foundational ideas:

  • Moving averages (smoothing noisy signals)
  • Rolling windows (e.g., "last 7 days average HRV")
  • Anomaly detection (is today very different from your typical range?)
  • Simple forecasting (linear regression, exponential smoothing)

Why this matters: Time-series LLMs feel like a natural extension once you understand these basics—they're just far more powerful at capturing complex, non-linear patterns.

Step 4: Understand LLMs Conceptually

You don't need the full math, but you should know:

  • LLMs read tokens (pieces of text) and predict the next token
  • Fine-tuning = retraining the model on new examples so it behaves differently (e.g., becomes a health coach)
  • RAG = giving the model extra documents to read while answering (like open-book exam)
  • Prompting = instructing the model to behave a certain way ("Explain this like I'm 16", "Be concise")

Link to health:

  • Instead of only text tokens, time-series LLMs also get "tokens" that encode numeric sequences
  • The model learns to "read" patterns in those sequences the same way it reads sentences

Step 5: Read High-Level Summaries of Health-LLM / OpenTSLM / PH-LLM

You're not expected to fully understand the research papers yet. Look for:

  • What problems they solve: sleep staging, ECG classification, glucose forecasting, fatigue prediction
  • How they mix time-series and text: tokenizing numeric sequences, multi-modal architectures
  • What data they needed: wearables (Oura, Garmin), EHR records, lab results

Your goal: Think, "Ah, so this is like giving the model a compressed playlist of my body signals + text notes, and it learns to make predictions and explanations".

Step 6: Build a Tiny Prototype

Project idea: Build a simple "HRV trend explainer"

  1. Export 30 days of your HRV data
  2. Calculate 7-day rolling average
  3. Flag days where HRV dropped >15% from average
  4. Use OpenAI API or Claude API with a prompt like:

You are a health coach. Here is the user's HRV data for the past 30 days:
[paste data]

Days flagged as low: Day 8, Day 15, Day 22.

Provide a brief, friendly explanation of what might be happening and suggest 2–3 recovery strategies.

What you're learning: How to combine numeric analysis (simple time-series logic) with LLM reasoning (explanations and coaching).

This is the core pattern of real health AI products.

7. How All of This Becomes a Real Product

In a production health tech app that combines wearables + lab tests, the architecture typically looks like this:

System Architecture

  1. Data ingestion: APIs from Garmin/Oura/Fitbit/LabCorp pull your data regularly
  2. Feature & timeline builder: Raw streams are transformed into summaries and episodes (e.g., 7-day windows, 30-day trends). This step is a core part of any effective Workflow Automation Design in health tech.
  3. Time-series / numeric model: Predicts risk scores, flags anomalies, forecasts future states (e.g., "likely to have fragmented sleep tonight")
  4. Fine-tuned coach LLM: Explains results, suggests next steps, maintains consistent tone and personality
  5. Guardrails: Blocks medical diagnosis, urgent advice, medication changes; escalates emergencies to humans
  6. UI/UX: Daily insight cards, weekly reviews, push notifications, educational content

What the User Sees

From your perspective, you just see:

Tonight: Go to bed 30 minutes earlier than yesterday.
Why: Your recent pattern suggests your nervous system needs recovery—HRV has dropped 12% over 5 days while sleep latency increased.

Under the hood: A time-series LLM analyzed your 7-day physiological curves, detected a stress pattern, and a coach LLM translated that into friendly, actionable language.

Example: Real-World Implementation

A recent study demonstrated a Selective RAG-Enhanced Hybrid ML-LLM framework for wearable-based fatigue prediction:

  • ML model (logistic regression) handled fast, efficient classification
  • LLM reasoning provided interpretable explanations when ML confidence was low
  • SHAP-based interpretation + LLM analysis both identified short-term sleep duration and HRV as dominant predictors

This hybrid approach achieved robustness, interpretability, and efficiency—exactly what you need for real-world health monitoring.

8. Mental Model to Keep in Your Head

If you remember only this, it's enough to build from:

  • Wearables + labs = your body's timeline (episodes, not snapshots)
  • Time-series models = pattern detectors over that timeline (trends, anomalies, forecasts)
  • LLMs = explainers + coaches (turn numbers into language and actions)
  • Time-series LLMs (like Health-LLM / OpenTSLM / PH-LLM) = models that can do both: read the curves AND talk about them

Once you're solid on Python, basic ML intuition, and time-series fundamentals, these models stop being mysterious black boxes and start feeling like powerful tools you can actually build with.

9. Next Steps: From Learning to Building

If You're Just Starting (0–6 Months)

  • Focus on Python + pandas + basic time-series visualization
  • Export your own wearable data and explore it
  • Build simple experiments: "What's my average HRV on days I sleep >8 hours vs. <7 hours?"

If You're Intermediate (6–12 Months)

  • Learn basic ML (scikit-learn, simple regression, classification)
  • Experiment with LLM APIs (OpenAI, Anthropic) for text generation
  • Build a simple health coach bot that reads your exported CSV and gives personalized feedback using prompts

If You're Advanced (12+ Months)

  • Study time-series LLM architectures (Health-LLM, OpenTSLM, MedTsLLM papers)
  • Experiment with fine-tuning smaller models (Llama, Mistral) on health coaching examples
  • Build a RAG + fine-tuned hybrid system that combines medical knowledge retrieval with personalized pattern detection

Resources to Explore

  • Andrej Karpathy's Neural Networks: Zero to Hero (YouTube series teaching LLMs from scratch)
  • OpenTSLM GitHub repo (Stanford's open-source time-series LLM)
  • PhysioNet datasets (open health data for practice)
  • Google's PH-LLM research (case studies on wearable-based health reasoning)

Final Thoughts: Why This Matters Now

We're at an inflection point where personalized health AI is transitioning from research labs to real products. Time-series LLMs enable something that was impossible before: continuous, interpretable, personalized health intelligence that explains itself in plain language.

The key insight: Your body is a dynamic system, not a static snapshot. Time-series LLMs finally give us AI that understands timelines, not just moments.

And the best part? You can start learning this today—no PhD required, just curiosity and patience.

Further Reading


*Written by Dr Hernani Costa | Powered by Core Ventures

Originally published at First AI Movers.

Technology is easy. Mapping it to P&L is hard. At First AI Movers, we don't just write code; we build the 'Executive Nervous System' for EU SMEs.

Is your architecture creating technical debt or business equity?

👉 Get your AI Readiness Score (Free Company Assessment)

Our AI Readiness Assessment evaluates your organization's capability to deploy time-series LLMs and other advanced health AI systems. We combine AI Strategy Consulting with Operational AI Implementation to ensure your Digital Transformation Strategy delivers measurable ROI.

Top comments (0)