<?xml version="1.0" encoding="UTF-8"?>
<rss version="2.0" xmlns:atom="http://www.w3.org/2005/Atom" xmlns:dc="http://purl.org/dc/elements/1.1/">
  <channel>
    <title>DEV Community: Ege Pakten</title>
    <description>The latest articles on DEV Community by Ege Pakten (@egepakten).</description>
    <link>https://dev.to/egepakten</link>
    <image>
      <url>https://media2.dev.to/dynamic/image/width=90,height=90,fit=cover,gravity=auto,format=auto/https:%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Fuser%2Fprofile_image%2F2054823%2Fa7415b5b-d68c-4462-a993-62c029abd337.png</url>
      <title>DEV Community: Ege Pakten</title>
      <link>https://dev.to/egepakten</link>
    </image>
    <atom:link rel="self" type="application/rss+xml" href="https://dev.to/feed/egepakten"/>
    <language>en</language>
    <item>
      <title>The Machine Learning Lifecycle: 10 Steps From Problem to Production (And Why Most Projects Fail at Step 3)</title>
      <dc:creator>Ege Pakten</dc:creator>
      <pubDate>Sat, 25 Apr 2026 09:15:35 +0000</pubDate>
      <link>https://dev.to/egepakten/the-machine-learning-lifecycle-10-steps-from-problem-to-production-and-why-most-projects-fail-at-b38</link>
      <guid>https://dev.to/egepakten/the-machine-learning-lifecycle-10-steps-from-problem-to-production-and-why-most-projects-fail-at-b38</guid>
      <description>&lt;blockquote&gt;
&lt;p&gt;Every ML tutorial jumps straight to model training. But in the real world, training is step 7 out of 10 — and the steps before it are where projects succeed or fail. This post walks through the full Machine Learning Lifecycle, from defining your problem to keeping your model healthy in production, with real examples and practical advice at every stage.&lt;/p&gt;
&lt;/blockquote&gt;




&lt;h2&gt;
  
  
  The Big Picture
&lt;/h2&gt;

&lt;p&gt;Machine Learning is an &lt;strong&gt;iterative and structured process&lt;/strong&gt;. It's not "throw data at an algorithm and hope for magic." It's a cycle — and most teams loop through it multiple times before they get something that works in production.&lt;/p&gt;

&lt;p&gt;Here are the 10 stages:&lt;br&gt;
&lt;/p&gt;

&lt;div class="highlight js-code-highlight"&gt;
&lt;pre class="highlight plaintext"&gt;&lt;code&gt;1. Problem Definition → 2. Data Collection → 3. Data Cleaning &amp;amp; Preprocessing
→ 4. Exploratory Data Analysis (EDA) → 5. Feature Engineering &amp;amp; Selection
→ 6. Model Selection → 7. Model Training → 8. Model Evaluation &amp;amp; Tuning
→ 9. Model Deployment → 10. Monitoring &amp;amp; Maintenance → (back to 1)
&lt;/code&gt;&lt;/pre&gt;

&lt;/div&gt;



&lt;p&gt;Let's go through each one.&lt;/p&gt;




&lt;h2&gt;
  
  
  1. Problem Definition — "What are we actually solving?"
&lt;/h2&gt;

&lt;p&gt;This is where most failed ML projects go wrong. Before touching any data or code, you need to answer:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;
&lt;strong&gt;What business problem am I solving?&lt;/strong&gt; Not "I want to use AI," but "I want to reduce customer churn by 15%" or "I want to detect fraudulent transactions in real time."&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;Is ML even the right tool?&lt;/strong&gt; Sometimes a simple rule-based system or a SQL query is better. ML is expensive overkill for problems that have clear, deterministic rules.&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;What does success look like?&lt;/strong&gt; Define a measurable metric: accuracy, precision, recall, revenue impact, latency requirements.&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;What type of ML problem is this?&lt;/strong&gt; Classification? Regression? Clustering? Recommendation? This dictates everything downstream.&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;&lt;strong&gt;The problem definition dictates the type of data you need.&lt;/strong&gt; If you define the problem wrong, you'll collect the wrong data, build the wrong model, and ship something nobody wanted.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;Example:&lt;/strong&gt; A bank wants to "use AI." That's not a problem definition. "Predict which credit card transactions are fraudulent with less than 0.1% false positive rate and under 200ms latency" — &lt;em&gt;that's&lt;/em&gt; a problem definition.&lt;/p&gt;




&lt;h2&gt;
  
  
  2. Data Collection — "Do we have enough?"
&lt;/h2&gt;

&lt;p&gt;Once you know what you're solving, you need data. This step is about gathering enough high-quality, relevant data to train a model.&lt;/p&gt;

&lt;p&gt;Key questions:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;
&lt;strong&gt;Where does the data come from?&lt;/strong&gt; Internal databases, APIs, web scraping, third-party vendors, public datasets, user-generated content?&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;How much data do I need?&lt;/strong&gt; Depends on complexity. A simple classifier might need 1,000 examples. A computer vision model might need 100,000+ labeled images. An LLM needs billions of tokens.&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;Is the data labeled?&lt;/strong&gt; For supervised learning, you need labels (the "right answers"). Labeling is often the most expensive and time-consuming part.&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;Is the data representative?&lt;/strong&gt; If you train a facial recognition system only on photos of one demographic, it will fail on others. Your data must represent the real-world distribution.&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;&lt;strong&gt;Common pitfalls:&lt;/strong&gt;&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;Assuming you have "big data" when you actually have big &lt;em&gt;noise&lt;/em&gt;
&lt;/li&gt;
&lt;li&gt;Not checking for sampling bias&lt;/li&gt;
&lt;li&gt;Ignoring data privacy regulations (GDPR, KVKK, HIPAA)&lt;/li&gt;
&lt;li&gt;Collecting too many features and not enough samples&lt;/li&gt;
&lt;/ul&gt;




&lt;h2&gt;
  
  
  3. Data Cleaning and Preprocessing — "Garbage in, garbage out"
&lt;/h2&gt;

&lt;p&gt;This is where you spend &lt;strong&gt;60-80% of your actual project time.&lt;/strong&gt; Raw data is messy. Always.&lt;/p&gt;

&lt;p&gt;What you're doing here:&lt;/p&gt;

&lt;h3&gt;
  
  
  Handling Missing Values
&lt;/h3&gt;

&lt;ul&gt;
&lt;li&gt;Some rows have blank fields. Do you fill them with the mean? The median? A prediction? Or drop them entirely?&lt;/li&gt;
&lt;li&gt;The right answer depends on &lt;em&gt;why&lt;/em&gt; the data is missing. "Random missing" and "systematically missing" require different approaches.&lt;/li&gt;
&lt;/ul&gt;

&lt;h3&gt;
  
  
  Removing Duplicates
&lt;/h3&gt;

&lt;ul&gt;
&lt;li&gt;Duplicate records distort your model's understanding of the distribution.&lt;/li&gt;
&lt;/ul&gt;

&lt;h3&gt;
  
  
  Fixing Inconsistent Data
&lt;/h3&gt;

&lt;ul&gt;
&lt;li&gt;"New York", "new york", "NY", "N.Y." are the same city but four different strings.&lt;/li&gt;
&lt;li&gt;Date formats: "04/20/2026" vs "2026-04-20" vs "20 April 2026"&lt;/li&gt;
&lt;li&gt;Units: meters vs feet, Celsius vs Fahrenheit&lt;/li&gt;
&lt;/ul&gt;

&lt;h3&gt;
  
  
  Handling Outliers
&lt;/h3&gt;

&lt;ul&gt;
&lt;li&gt;A salary dataset where most values are $40K-$120K but one entry says $99,999,999. Is it real or a typo? Outliers can destroy model performance or provide critical signal — you have to decide.&lt;/li&gt;
&lt;/ul&gt;

&lt;h3&gt;
  
  
  Data Type Conversions
&lt;/h3&gt;

&lt;ul&gt;
&lt;li&gt;Categorical variables need encoding (one-hot, label encoding)&lt;/li&gt;
&lt;li&gt;Text needs tokenization&lt;/li&gt;
&lt;li&gt;Images need resizing, normalization&lt;/li&gt;
&lt;li&gt;Dates need feature extraction (day of week, month, holiday flag)&lt;/li&gt;
&lt;/ul&gt;

&lt;h3&gt;
  
  
  Normalization and Scaling
&lt;/h3&gt;

&lt;ul&gt;
&lt;li&gt;Features on different scales (age: 0-100, salary: 20,000-500,000) can bias models that use distance calculations. Standard scaling (z-score) or min-max scaling fixes this.&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;&lt;strong&gt;The motto: garbage in, garbage out.&lt;/strong&gt; No model, no matter how sophisticated, can learn good patterns from bad data.&lt;/p&gt;




&lt;h2&gt;
  
  
  4. Exploratory Data Analysis (EDA) — "What does the data actually look like?"
&lt;/h2&gt;

&lt;p&gt;Before building any model, you need to &lt;strong&gt;understand your data&lt;/strong&gt;. EDA is about getting the big picture.&lt;/p&gt;

&lt;p&gt;What you're looking for:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;
&lt;strong&gt;Distributions&lt;/strong&gt; — Is your target variable balanced? If 99% of transactions are legitimate and 1% are fraud, you have a class imbalance problem.&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;Correlations&lt;/strong&gt; — Which features are related to each other? Which features predict your target?&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;Patterns and trends&lt;/strong&gt; — Seasonal effects? Time-based shifts? Geographic clusters?&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;Data quality issues&lt;/strong&gt; you missed in step 3 — Sometimes problems only become visible in visualization.&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;Tools: histograms, scatter plots, correlation matrices, box plots, pair plots. Libraries: Pandas, Matplotlib, Seaborn, Plotly.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;Example:&lt;/strong&gt; You're building a house price predictor. EDA reveals that "number of bedrooms" and "square footage" are highly correlated (0.92). Including both might cause multicollinearity. You might drop one or combine them.&lt;/p&gt;

&lt;p&gt;EDA often sends you &lt;strong&gt;back to step 2 or 3&lt;/strong&gt; — you realize you need more data, or your data has problems you didn't see before.&lt;/p&gt;




&lt;h2&gt;
  
  
  5. Feature Engineering and Selection — "Good features &amp;gt; fancy models"
&lt;/h2&gt;

&lt;p&gt;This is often the difference between a mediocre model and a great one. Feature engineering is the art of &lt;strong&gt;creating new input variables&lt;/strong&gt; that help the model learn patterns better.&lt;/p&gt;

&lt;h3&gt;
  
  
  Feature Engineering (Creating)
&lt;/h3&gt;

&lt;ul&gt;
&lt;li&gt;
&lt;strong&gt;From dates:&lt;/strong&gt; extract day_of_week, is_weekend, month, quarter, days_since_last_event&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;From text:&lt;/strong&gt; word count, sentiment score, TF-IDF values, embeddings&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;From location:&lt;/strong&gt; distance to nearest city, population density, latitude buckets&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;Combining features:&lt;/strong&gt; price_per_sqft = price / square_footage, BMI = weight / height²&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;Domain knowledge:&lt;/strong&gt; a doctor knows that "blood pressure × age" interaction matters; encode that&lt;/li&gt;
&lt;/ul&gt;

&lt;h3&gt;
  
  
  Feature Selection (Removing)
&lt;/h3&gt;

&lt;p&gt;Not all features help. Some add noise. Too many features cause overfitting and slow training. Techniques:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;
&lt;strong&gt;Correlation analysis&lt;/strong&gt; — drop features that are highly correlated with each other&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;Feature importance&lt;/strong&gt; from tree-based models (Random Forest, XGBoost)&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;Recursive Feature Elimination (RFE)&lt;/strong&gt; — iteratively remove the least important feature&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;L1 Regularization (Lasso)&lt;/strong&gt; — automatically zeroes out unimportant features during training&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;&lt;strong&gt;Key insight: a simple model with great features almost always beats a complex model with bad features.&lt;/strong&gt;&lt;/p&gt;




&lt;h2&gt;
  
  
  6. Model Selection — "Choose the right tool for the job"
&lt;/h2&gt;

&lt;p&gt;Now you pick which algorithm(s) to try. This depends on:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;
&lt;strong&gt;Problem type:&lt;/strong&gt; Classification → Logistic Regression, Random Forest, SVM, Neural Network. Regression → Linear Regression, XGBoost, Neural Network. Clustering → K-Means, DBSCAN. Sequence → RNN, LSTM, Transformer.&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;Data size:&lt;/strong&gt; Small data → simpler models (logistic regression, SVM). Large data → deep learning can shine.&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;Interpretability needs:&lt;/strong&gt; Healthcare and finance often need explainable models (decision trees, linear models). Recommendation engines can afford black boxes (deep learning).&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;Latency requirements:&lt;/strong&gt; Real-time inference needs fast models. Batch processing can afford slower ones.&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;&lt;strong&gt;Best practice:&lt;/strong&gt; Start simple. Try logistic regression or a decision tree first. If it gets 85% accuracy, you have a strong baseline. Then try more complex models and see if the improvement justifies the complexity.&lt;/p&gt;

&lt;p&gt;You often try &lt;strong&gt;3-5 different models&lt;/strong&gt; and compare their performance.&lt;/p&gt;




&lt;h2&gt;
  
  
  7. Model Training — "The model learns about the data"
&lt;/h2&gt;

&lt;p&gt;This is the step everyone thinks ML is about — but as you've seen, it's step 7 of 10.&lt;/p&gt;

&lt;p&gt;Training means:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;
&lt;strong&gt;Feed data into the algorithm&lt;/strong&gt; — the model sees examples and adjusts its internal parameters (weights) to minimize error&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;Split data into train/validation/test sets&lt;/strong&gt; — typically 70/15/15 or 80/10/10. Never evaluate on data the model trained on.&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;Choose a loss function&lt;/strong&gt; — the mathematical definition of "what is wrong." Cross-entropy for classification, MSE for regression, etc.&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;Set hyperparameters&lt;/strong&gt; — learning rate, batch size, epochs, regularization strength. These are not learned by the model; you set them.&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;Iterate&lt;/strong&gt; — training is rarely one-shot. You train, look at results, adjust, retrain.&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;&lt;strong&gt;Key concept: train/test split.&lt;/strong&gt; If you evaluate your model on the same data it trained on, you get misleadingly high scores. It's like grading a student using the exact exam questions they practiced on.&lt;/p&gt;




&lt;h2&gt;
  
  
  8. Model Evaluation and Tuning — "How is your model doing?"
&lt;/h2&gt;

&lt;p&gt;Training is done. Now: is the model actually good?&lt;/p&gt;

&lt;h3&gt;
  
  
  Evaluation Metrics
&lt;/h3&gt;

&lt;p&gt;Different problems need different metrics:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;
&lt;strong&gt;Accuracy&lt;/strong&gt; — % of correct predictions. Misleading with imbalanced data (99% accuracy on fraud detection means nothing if you just predict "not fraud" every time).&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;Precision&lt;/strong&gt; — Of all things the model flagged as positive, how many were actually positive?&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;Recall&lt;/strong&gt; — Of all actual positives, how many did the model catch?&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;F1 Score&lt;/strong&gt; — Harmonic mean of precision and recall. Good when you need to balance both.&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;AUC-ROC&lt;/strong&gt; — Area under the curve. Measures how well the model separates classes across all thresholds.&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;MSE / RMSE / MAE&lt;/strong&gt; — For regression: how far off are predictions from actual values?&lt;/li&gt;
&lt;/ul&gt;

&lt;h3&gt;
  
  
  Hyperparameter Tuning
&lt;/h3&gt;

&lt;p&gt;If results aren't good enough, adjust hyperparameters:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;
&lt;strong&gt;Grid Search&lt;/strong&gt; — try every combination of predefined values&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;Random Search&lt;/strong&gt; — randomly sample combinations (often faster than grid search)&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;Bayesian Optimization&lt;/strong&gt; — smart search that learns from previous trials&lt;/li&gt;
&lt;/ul&gt;

&lt;h3&gt;
  
  
  Dealing with Problems
&lt;/h3&gt;

&lt;ul&gt;
&lt;li&gt;
&lt;strong&gt;Overfitting&lt;/strong&gt; (training score high, test score low) → more data, simpler model, regularization, dropout&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;Underfitting&lt;/strong&gt; (both scores low) → more complex model, more features, longer training&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;Class imbalance&lt;/strong&gt; → oversampling (SMOTE), undersampling, class weights, different metrics&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;&lt;strong&gt;This step often sends you back to steps 3, 4, 5, or 6.&lt;/strong&gt; That's the iterative nature of ML.&lt;/p&gt;




&lt;h2&gt;
  
  
  9. Model Deployment — "Integrate model to the real world"
&lt;/h2&gt;

&lt;p&gt;Your model works in a Jupyter notebook. Now it needs to work in production — handling real users, real data, and real scale.&lt;/p&gt;

&lt;p&gt;Deployment means:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;
&lt;strong&gt;Packaging the model&lt;/strong&gt; — save weights, serialize with ONNX, TorchScript, or pickle&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;Creating an API&lt;/strong&gt; — wrap the model in a REST API (Flask, FastAPI) or gRPC endpoint&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;Infrastructure&lt;/strong&gt; — where does it run? AWS SageMaker, Google Vertex AI, Azure ML, self-hosted Kubernetes, or edge devices?&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;Scaling&lt;/strong&gt; — handle 10 requests/second? 10,000? Auto-scaling, load balancing, caching&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;CI/CD for ML&lt;/strong&gt; — automated testing, model versioning, rollback capabilities&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;&lt;strong&gt;Common deployment patterns:&lt;/strong&gt;&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;
&lt;strong&gt;Real-time inference&lt;/strong&gt; — API call, response in milliseconds (fraud detection, chatbot)&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;Batch inference&lt;/strong&gt; — process large datasets periodically (weekly churn predictions, nightly recommendations)&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;Edge deployment&lt;/strong&gt; — model runs on device (mobile app, IoT sensor, self-driving car)&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;Deployment is NOT the finish line. It's where the real work begins.&lt;/p&gt;




&lt;h2&gt;
  
  
  10. Monitoring and Maintenance — "Keep model healthy"
&lt;/h2&gt;

&lt;p&gt;A deployed model is a living system. It degrades over time because the real world changes.&lt;/p&gt;

&lt;h3&gt;
  
  
  What to Monitor
&lt;/h3&gt;

&lt;ul&gt;
&lt;li&gt;
&lt;strong&gt;Model performance&lt;/strong&gt; — are accuracy/precision/recall staying stable?&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;Data drift&lt;/strong&gt; — is incoming data different from training data? (seasonal changes, new user demographics, market shifts)&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;Concept drift&lt;/strong&gt; — has the relationship between features and target changed? (what predicted churn in 2023 might not in 2026)&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;Latency&lt;/strong&gt; — is inference speed within requirements?&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;Resource usage&lt;/strong&gt; — CPU, memory, cost&lt;/li&gt;
&lt;/ul&gt;

&lt;h3&gt;
  
  
  When to Retrain
&lt;/h3&gt;

&lt;ul&gt;
&lt;li&gt;Performance drops below a threshold&lt;/li&gt;
&lt;li&gt;Data distribution shifts significantly&lt;/li&gt;
&lt;li&gt;Business requirements change&lt;/li&gt;
&lt;li&gt;New data categories appear that the model has never seen&lt;/li&gt;
&lt;/ul&gt;

&lt;h3&gt;
  
  
  Best Practices
&lt;/h3&gt;

&lt;ul&gt;
&lt;li&gt;Set up &lt;strong&gt;automated alerts&lt;/strong&gt; for performance degradation&lt;/li&gt;
&lt;li&gt;Keep a &lt;strong&gt;champion/challenger&lt;/strong&gt; system: new model version runs alongside the old one; switch only when the new one proves better&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;Log everything&lt;/strong&gt;: predictions, input data, confidence scores. You'll need this for debugging.&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;Version your models&lt;/strong&gt; like you version code. Know exactly which model version produced which prediction.&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;&lt;strong&gt;Monitoring is vital.&lt;/strong&gt; A model that was 95% accurate at launch can silently drop to 70% if nobody's watching.&lt;/p&gt;




&lt;h2&gt;
  
  
  Summary — The Key Takeaways
&lt;/h2&gt;

&lt;p&gt;&lt;strong&gt;ML is an iterative and structured process.&lt;/strong&gt; It's a cycle, not a line. You will loop back to earlier steps repeatedly — that's normal, not failure.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;Data quality and feature engineering are critical.&lt;/strong&gt; Steps 3 and 5 have more impact on final model performance than the choice of algorithm at step 6. Good features beat fancy models.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;Evaluation and tuning improve model performance.&lt;/strong&gt; Don't ship the first model that trains. Rigorously evaluate, tune hyperparameters, and test on data the model has never seen.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;Deployment isn't the end; monitoring is vital.&lt;/strong&gt; The real world changes. Your model will degrade. Monitor, retrain, and iterate continuously.&lt;/p&gt;

&lt;p&gt;The lifecycle is a loop. The best ML teams are the ones that spin through it fastest — not the ones with the fanciest models.&lt;/p&gt;




&lt;p&gt;*If this helped you see the full picture of ML beyond "just training," drop a reaction.&lt;/p&gt;

</description>
      <category>ai</category>
      <category>datascience</category>
      <category>machinelearning</category>
      <category>tutorial</category>
    </item>
    <item>
      <title>Two Main Sources of ML Models: Pre-trained vs Custom — Which One Should You Use?</title>
      <dc:creator>Ege Pakten</dc:creator>
      <pubDate>Mon, 20 Apr 2026 09:07:42 +0000</pubDate>
      <link>https://dev.to/egepakten/two-main-sources-of-ml-models-pre-trained-vs-custom-which-one-should-you-use-24m6</link>
      <guid>https://dev.to/egepakten/two-main-sources-of-ml-models-pre-trained-vs-custom-which-one-should-you-use-24m6</guid>
      <description>&lt;blockquote&gt;
&lt;p&gt;You want to build an AI-powered feature. Maybe a chatbot, an image classifier, or a recommendation engine. The first decision you'll face is: &lt;strong&gt;should I use someone else's model or train my own?&lt;/strong&gt; This post breaks down both paths, when to pick which, and the real-world tradeoffs nobody tells you about upfront.&lt;/p&gt;
&lt;/blockquote&gt;




&lt;h2&gt;
  
  
  Why This Matters
&lt;/h2&gt;

&lt;p&gt;Before you write a single line of code, this one decision determines three things: how fast you ship, how much you spend, and how well your product works. Pick wrong and you'll either waste months reinventing something that already exists, or ship a generic solution that doesn't actually solve your problem.&lt;/p&gt;




&lt;h2&gt;
  
  
  Source 1: Open Source Pre-trained Models
&lt;/h2&gt;

&lt;h3&gt;
  
  
  What Are They?
&lt;/h3&gt;

&lt;p&gt;Pre-trained models are models that someone else (usually a big tech company or research lab) has &lt;strong&gt;already trained on massive datasets&lt;/strong&gt;. They've spent the GPU hours, the engineering time, and the data collection effort. You just download and use.&lt;/p&gt;

&lt;p&gt;Think of it like buying a car vs building one from scratch. The car (pre-trained model) already works — you just need to learn to drive it and maybe customize the seats.&lt;/p&gt;

&lt;h3&gt;
  
  
  Real Examples
&lt;/h3&gt;

&lt;p&gt;Here are some of the most well-known pre-trained models and what they do:&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;Natural Language Processing (NLP):&lt;/strong&gt;&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;
&lt;strong&gt;BERT&lt;/strong&gt; (Google) — understands text, great for search, Q&amp;amp;A, classification&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;GPT&lt;/strong&gt; (OpenAI) — generates text, powers ChatGPT&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;LLaMA&lt;/strong&gt; (Meta) — open-weight LLM, fine-tunable for custom use&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;T5&lt;/strong&gt; (Google) — text-to-text framework, versatile for many NLP tasks&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;&lt;strong&gt;Computer Vision:&lt;/strong&gt;&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;
&lt;strong&gt;ResNet&lt;/strong&gt; — image classification, object detection&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;YOLO&lt;/strong&gt; (You Only Look Once) — real-time object detection&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;CLIP&lt;/strong&gt; (OpenAI) — connects text and images in the same space&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;&lt;strong&gt;Speech AI:&lt;/strong&gt;&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;
&lt;strong&gt;Whisper&lt;/strong&gt; (OpenAI) — speech-to-text in 99 languages&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;Wav2Vec&lt;/strong&gt; (Meta) — speech recognition with minimal labeled data&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;&lt;strong&gt;Healthcare &amp;amp; Science:&lt;/strong&gt;&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;
&lt;strong&gt;AlphaFold&lt;/strong&gt; (DeepMind) — protein structure prediction&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;BioGPT&lt;/strong&gt; (Microsoft) — biomedical text generation&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;&lt;strong&gt;Art &amp;amp; Creative:&lt;/strong&gt;&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;
&lt;strong&gt;Stable Diffusion&lt;/strong&gt; — text-to-image generation&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;MusicGen&lt;/strong&gt; (Meta) — text-to-music&lt;/li&gt;
&lt;/ul&gt;

&lt;h3&gt;
  
  
  Benefits
&lt;/h3&gt;

&lt;p&gt;&lt;strong&gt;Saves development time.&lt;/strong&gt; Training a model like BERT from scratch would take weeks on expensive GPUs. Downloading it takes 30 seconds.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;Great for transfer learning.&lt;/strong&gt; This is a key concept. The model has already learned general patterns (what a face looks like, how sentences are structured, etc.). You can then &lt;strong&gt;fine-tune&lt;/strong&gt; it on your specific data to adapt it for your use case — often with very little data.&lt;/p&gt;

&lt;p&gt;For example: BERT was trained on all of Wikipedia and BookCorpus. If you fine-tune it on 1,000 customer support tickets, it can classify support requests with 90%+ accuracy — because it already understands language, it just needs to learn your categories.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;Community and ecosystem.&lt;/strong&gt; Popular pre-trained models have huge communities, tutorials, and tooling. Platforms like Hugging Face host thousands of pre-trained models you can try in minutes.&lt;/p&gt;

&lt;h3&gt;
  
  
  When to Use Pre-trained Models
&lt;/h3&gt;

&lt;ul&gt;
&lt;li&gt;You're solving a &lt;strong&gt;common problem&lt;/strong&gt; (text classification, image recognition, translation, Q&amp;amp;A)&lt;/li&gt;
&lt;li&gt;You need to &lt;strong&gt;ship fast&lt;/strong&gt; (prototype in days, not months)&lt;/li&gt;
&lt;li&gt;You have &lt;strong&gt;limited data&lt;/strong&gt; (transfer learning works with small datasets)&lt;/li&gt;
&lt;li&gt;You have &lt;strong&gt;limited budget&lt;/strong&gt; (no GPU cluster needed)&lt;/li&gt;
&lt;li&gt;Your domain isn't extremely specialized&lt;/li&gt;
&lt;/ul&gt;




&lt;h2&gt;
  
  
  Source 2: Training Custom Models
&lt;/h2&gt;

&lt;h3&gt;
  
  
  What Are They?
&lt;/h3&gt;

&lt;p&gt;You collect your own data and &lt;strong&gt;train a model from scratch&lt;/strong&gt; — or from a very early starting point. You design the architecture, curate the dataset, define the training process, and iterate until it works.&lt;/p&gt;

&lt;h3&gt;
  
  
  When to Use Custom Models
&lt;/h3&gt;

&lt;p&gt;&lt;strong&gt;When pre-trained models aren't specific enough.&lt;/strong&gt; If you're doing something that generic models haven't seen — like detecting manufacturing defects on a very specific assembly line, or classifying rare medical conditions from proprietary scan data — a pre-trained model won't cut it.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;For proprietary, domain-specific problems.&lt;/strong&gt; Some industries have data that no public model has ever seen: financial fraud patterns unique to your bank, satellite imagery of your specific crop types, sonar data for underwater pipeline inspection. These need custom models.&lt;/p&gt;

&lt;h3&gt;
  
  
  Benefits
&lt;/h3&gt;

&lt;p&gt;&lt;strong&gt;Tailored performance.&lt;/strong&gt; A custom model trained on &lt;em&gt;your&lt;/em&gt; data for &lt;em&gt;your&lt;/em&gt; problem will almost always outperform a generic one — if you have enough data and expertise.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;Control over training data and bias handling.&lt;/strong&gt; You decide exactly what goes in. You can audit, clean, balance, and de-bias your dataset. With a pre-trained model, you inherit whatever biases were in &lt;em&gt;their&lt;/em&gt; training data — and you often can't even see what that data was.&lt;/p&gt;

&lt;h3&gt;
  
  
  The Real Cost
&lt;/h3&gt;

&lt;p&gt;Training custom models is expensive in multiple ways:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;
&lt;strong&gt;Data collection and labeling&lt;/strong&gt; — You need thousands to millions of labeled examples. Getting quality labels is slow and costly (think: paying domain experts to annotate medical images).&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;Compute&lt;/strong&gt; — Training from scratch needs powerful GPUs/TPUs, often for days or weeks. A single training run for a large model can cost $10,000–$100,000+.&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;Expertise&lt;/strong&gt; — You need ML engineers who understand model architecture, loss functions, hyperparameter tuning, evaluation metrics, overfitting, and deployment.&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;Time&lt;/strong&gt; — Months of iteration before you have something production-ready.&lt;/li&gt;
&lt;/ul&gt;




&lt;h2&gt;
  
  
  The Decision Framework
&lt;/h2&gt;

&lt;p&gt;Here's how to actually choose. Ask yourself these questions in order:&lt;/p&gt;

&lt;h3&gt;
  
  
  Question 1: Has someone already solved this?
&lt;/h3&gt;

&lt;p&gt;Search Hugging Face, TensorFlow Hub, PyTorch Hub. If a pre-trained model exists for your exact task — use it. Don't reinvent the wheel.&lt;/p&gt;

&lt;h3&gt;
  
  
  Question 2: Is my problem &lt;em&gt;close&lt;/em&gt; to something that's been solved?
&lt;/h3&gt;

&lt;p&gt;If yes, &lt;strong&gt;start with a pre-trained model and fine-tune it&lt;/strong&gt;. This is the sweet spot for most companies. You get 80% of the benefit of custom training at 10% of the cost.&lt;/p&gt;

&lt;p&gt;Fine-tuning means: take a pre-trained model, freeze most of its layers, and retrain just the last few layers on your specific data. The model keeps its general knowledge but adapts to your task.&lt;/p&gt;

&lt;h3&gt;
  
  
  Question 3: Is my problem truly unique?
&lt;/h3&gt;

&lt;p&gt;If your data is proprietary, your domain is narrow, no pre-trained model comes close, and you have the budget and expertise — then train custom.&lt;/p&gt;

&lt;p&gt;But even here, most teams &lt;strong&gt;start from a pre-trained base&lt;/strong&gt; and do heavy fine-tuning rather than literally training from a random weight initialization.&lt;/p&gt;

&lt;h3&gt;
  
  
  The Spectrum
&lt;/h3&gt;

&lt;p&gt;In practice, it's rarely a binary choice. Think of it as a spectrum:&lt;br&gt;
&lt;/p&gt;

&lt;div class="highlight js-code-highlight"&gt;
&lt;pre class="highlight plaintext"&gt;&lt;code&gt;Use as-is ← Fine-tune ← Heavy fine-tune ← Train from scratch
(cheapest)                                    (most expensive)
&lt;/code&gt;&lt;/pre&gt;

&lt;/div&gt;



&lt;p&gt;Most real-world ML projects land somewhere in the middle.&lt;/p&gt;




&lt;h2&gt;
  
  
  Key Terms You Should Know
&lt;/h2&gt;

&lt;p&gt;&lt;strong&gt;Transfer Learning&lt;/strong&gt; — Taking a model trained on one task and adapting it for a different (but related) task. The core technique that makes pre-trained models so powerful.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;Fine-tuning&lt;/strong&gt; — Retraining some or all layers of a pre-trained model on your specific dataset. Keeps general knowledge, adds specialized knowledge.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;Feature Extraction&lt;/strong&gt; — Using a pre-trained model as a fixed "feature extractor" — you feed data through it, take the intermediate representations, and train a simple classifier on top. Even cheaper than fine-tuning.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;Hyperparameters&lt;/strong&gt; — Settings you choose &lt;em&gt;before&lt;/em&gt; training starts: learning rate, batch size, number of epochs, model architecture. Getting these right is critical and often requires experimentation.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;Overfitting&lt;/strong&gt; — When a model performs great on training data but poorly on new data. More common with small datasets and complex models. Regularization, dropout, and data augmentation help prevent it.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;Data Augmentation&lt;/strong&gt; — Artificially expanding your dataset by creating modified versions of existing data (rotating images, adding noise, paraphrasing text). Especially useful when you have limited training data.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;Model Hub&lt;/strong&gt; — Platforms that host thousands of pre-trained models ready to download: Hugging Face, TensorFlow Hub, PyTorch Hub, AWS SageMaker JumpStart.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;Inference&lt;/strong&gt; — Using a trained model to make predictions on new data. This is what happens in production. Different from training — inference is cheap and fast.&lt;/p&gt;




&lt;h2&gt;
  
  
  Common Mistakes
&lt;/h2&gt;

&lt;p&gt;&lt;strong&gt;Mistake 1: Training from scratch when a pre-trained model exists.&lt;/strong&gt; This is the most common waste of time and money in ML projects. Always search first.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;Mistake 2: Using a pre-trained model without evaluation.&lt;/strong&gt; Just because a model is on Hugging Face doesn't mean it works for &lt;em&gt;your&lt;/em&gt; data. Always benchmark it on your own test set before shipping.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;Mistake 3: Ignoring bias.&lt;/strong&gt; Pre-trained models inherit biases from their training data. If you're deploying in a sensitive domain (hiring, lending, healthcare), you must audit for fairness.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;Mistake 4: Underestimating data quality.&lt;/strong&gt; A custom model is only as good as its training data. Garbage in, garbage out — no amount of GPU power fixes bad labels.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;Mistake 5: Skipping the "boring" baseline.&lt;/strong&gt; Before reaching for deep learning, try a simple approach (logistic regression, decision tree, keyword matching). Sometimes it's 90% as good at 1% of the complexity.&lt;/p&gt;




&lt;h2&gt;
  
  
  Wrapping Up
&lt;/h2&gt;

&lt;p&gt;&lt;strong&gt;Pre-trained models&lt;/strong&gt; are great for common tasks and faster deployment. They save you time, money, and let you benefit from massive datasets and research you couldn't replicate yourself.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;Custom models&lt;/strong&gt; offer flexibility for solving unique or proprietary problems. They give you full control over performance and bias, but at significantly higher cost and complexity.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;Choosing the right source&lt;/strong&gt; depends on your goal, your data, and your resources. Most teams start with pre-trained, fine-tune to their needs, and only train from scratch when they absolutely must.&lt;/p&gt;

&lt;p&gt;The best ML engineers aren't the ones who build everything from zero — they're the ones who know &lt;em&gt;when not to&lt;/em&gt;.&lt;/p&gt;




&lt;p&gt;&lt;em&gt;If this helped, drop a reaction. More AWS notes coming soon.&lt;/em&gt;&lt;/p&gt;

</description>
      <category>ai</category>
      <category>beginners</category>
      <category>machinelearning</category>
      <category>opensource</category>
    </item>
    <item>
      <title>What is RAG? A Beginner's Guide to Retrieval-Augmented Generation (With a Full Pipeline Walkthrough)</title>
      <dc:creator>Ege Pakten</dc:creator>
      <pubDate>Sat, 18 Apr 2026 09:21:50 +0000</pubDate>
      <link>https://dev.to/egepakten/what-is-rag-a-beginners-guide-to-retrieval-augmented-generation-with-a-full-pipeline-walkthrough-3djm</link>
      <guid>https://dev.to/egepakten/what-is-rag-a-beginners-guide-to-retrieval-augmented-generation-with-a-full-pipeline-walkthrough-3djm</guid>
      <description>&lt;blockquote&gt;
&lt;p&gt;If you've ever wondered &lt;em&gt;how&lt;/em&gt; ChatGPT-style apps can suddenly "know" about your company's internal documents, product manuals, or legal files without being retrained, the answer is almost always &lt;strong&gt;RAG&lt;/strong&gt; — Retrieval-Augmented Generation. In this post, we'll break down what RAG is, why it exists, and walk through the full pipeline step-by-step with a real example.&lt;/p&gt;
&lt;/blockquote&gt;




&lt;h2&gt;
  
  
  1. What is RAG?
&lt;/h2&gt;

&lt;p&gt;&lt;strong&gt;Retrieval-Augmented Generation (RAG)&lt;/strong&gt; is an AI framework that integrates an &lt;em&gt;information retrieval&lt;/em&gt; component into the generation process of Large Language Models (LLMs) to improve &lt;strong&gt;factuality&lt;/strong&gt; and &lt;strong&gt;relevance&lt;/strong&gt;.&lt;/p&gt;

&lt;p&gt;In plain English:&lt;/p&gt;

&lt;blockquote&gt;
&lt;p&gt;Instead of making the LLM &lt;em&gt;remember&lt;/em&gt; everything, we let it &lt;strong&gt;look things up&lt;/strong&gt; in a knowledge base right before answering.&lt;/p&gt;
&lt;/blockquote&gt;

&lt;p&gt;The term RAG was coined in a &lt;strong&gt;2020 research paper&lt;/strong&gt; by Patrick Lewis et al. ("Retrieval-Augmented Generation for Knowledge-Intensive NLP Tasks") published on arXiv. The core insight: combine a &lt;em&gt;parametric&lt;/em&gt; memory (the LLM's weights) with a &lt;em&gt;non-parametric&lt;/em&gt; memory (a searchable document store) — and you get the best of both worlds.&lt;/p&gt;




&lt;h2&gt;
  
  
  2. Why RAG? The Motivation
&lt;/h2&gt;

&lt;p&gt;Three big problems drove the invention of RAG:&lt;/p&gt;

&lt;h3&gt;
  
  
  LLM Limitations
&lt;/h3&gt;

&lt;p&gt;LLMs are frozen snapshots. Once a model is trained, it only knows what was in its training data. It doesn't know:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;What your company policies say&lt;/li&gt;
&lt;li&gt;What happened after its training cutoff&lt;/li&gt;
&lt;li&gt;What's in your private documents&lt;/li&gt;
&lt;li&gt;What yesterday's sales numbers were&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;And even with what it &lt;em&gt;does&lt;/em&gt; know, it can hallucinate confidently.&lt;/p&gt;

&lt;h3&gt;
  
  
  Cost of Retraining vs. Dynamic Retrieval
&lt;/h3&gt;

&lt;p&gt;You &lt;em&gt;could&lt;/em&gt; retrain or fine-tune the model every time your data changes. But:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;Retraining a large model can cost tens of thousands to millions of dollars&lt;/li&gt;
&lt;li&gt;It takes days or weeks&lt;/li&gt;
&lt;li&gt;You have to do it &lt;em&gt;again&lt;/em&gt; every time the data updates&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;Dynamic retrieval (looking things up at query time) is &lt;strong&gt;vastly&lt;/strong&gt; cheaper and always up-to-date.&lt;/p&gt;

&lt;h3&gt;
  
  
  Need for Grounded, Up-to-Date Knowledge
&lt;/h3&gt;

&lt;p&gt;For regulated industries (finance, healthcare, legal), you can't ship answers that come from "the model's memory." You need answers backed by &lt;strong&gt;sources&lt;/strong&gt; you can cite and audit.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;RAG addresses all three challenges by decoupling knowledge from the model.&lt;/strong&gt;&lt;/p&gt;




&lt;h2&gt;
  
  
  3. The RAG Pipeline — Step-by-Step With a Real Example
&lt;/h2&gt;

&lt;p&gt;This is the part most tutorials rush through. We're going to slow down.&lt;/p&gt;

&lt;p&gt;Let's use a concrete example. Imagine you're building an &lt;strong&gt;internal developer assistant&lt;/strong&gt; at a company called &lt;em&gt;Acme Corp&lt;/em&gt;. Employees can ask it questions about the engineering handbook, API docs, and on-call runbooks.&lt;/p&gt;

&lt;p&gt;A developer asks:&lt;/p&gt;

&lt;blockquote&gt;
&lt;p&gt;"How do I rotate the database credentials for the billing service?"&lt;/p&gt;
&lt;/blockquote&gt;

&lt;p&gt;Here's exactly what happens behind the scenes.&lt;/p&gt;




&lt;h3&gt;
  
  
  Phase 1: Indexing (Done Once, Ahead of Time)
&lt;/h3&gt;

&lt;p&gt;Before anyone can ask anything, we need to prepare the knowledge base.&lt;/p&gt;

&lt;h4&gt;
  
  
  Step 1a — Knowledge Corpus
&lt;/h4&gt;

&lt;p&gt;First, we gather &lt;strong&gt;every document&lt;/strong&gt; we want the assistant to know about:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;The engineering handbook (Markdown files)&lt;/li&gt;
&lt;li&gt;API documentation (HTML + Swagger specs)&lt;/li&gt;
&lt;li&gt;Runbooks (Confluence pages)&lt;/li&gt;
&lt;li&gt;Past incident post-mortems (Google Docs)&lt;/li&gt;
&lt;li&gt;Security policies (PDFs)&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;Let's say this gives us &lt;strong&gt;8,000 documents&lt;/strong&gt;.&lt;/p&gt;

&lt;h4&gt;
  
  
  Step 1b — Document Chunking
&lt;/h4&gt;

&lt;p&gt;An LLM can't efficiently search through a 50-page PDF. And you don't want to return a whole 50-page PDF to the user either — you want the &lt;em&gt;one paragraph&lt;/em&gt; that actually answers their question.&lt;/p&gt;

&lt;p&gt;So we &lt;strong&gt;chunk&lt;/strong&gt; each document into smaller pieces. A common approach:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;500 tokens per chunk (~300 words)&lt;/li&gt;
&lt;li&gt;50 token overlap between chunks (so we don't split an idea across a boundary)&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;One chunk in our knowledge base might look like this:&lt;br&gt;
&lt;/p&gt;

&lt;div class="highlight js-code-highlight"&gt;
&lt;pre class="highlight plaintext"&gt;&lt;code&gt;[Chunk #4729 — Source: runbooks/billing-service.md]
"To rotate database credentials for the billing service:
1. Generate a new password in AWS Secrets Manager.
2. Update the 'billing-db' secret with the new value.
3. Trigger a rolling restart via: kubectl rollout restart deploy/billing.
4. Verify health endpoints return 200 OK.
5. Revoke the old credentials after 24h grace period."
&lt;/code&gt;&lt;/pre&gt;

&lt;/div&gt;



&lt;p&gt;After chunking, our 8,000 documents become maybe &lt;strong&gt;120,000 chunks&lt;/strong&gt;.&lt;/p&gt;

&lt;h4&gt;
  
  
  Step 1c — Vector Embeddings
&lt;/h4&gt;

&lt;p&gt;For each chunk, we call an &lt;strong&gt;embedding model&lt;/strong&gt; (like BERT, OpenAI's &lt;code&gt;text-embedding-3-small&lt;/code&gt;, or Cohere's embedder). This turns each chunk into a &lt;strong&gt;vector&lt;/strong&gt; — a list of ~1,536 numbers that represents the meaning of that chunk.&lt;br&gt;
&lt;/p&gt;

&lt;div class="highlight js-code-highlight"&gt;
&lt;pre class="highlight json"&gt;&lt;code&gt;&lt;span class="err"&gt;Chunk&lt;/span&gt;&lt;span class="w"&gt; &lt;/span&gt;&lt;span class="err"&gt;#&lt;/span&gt;&lt;span class="mi"&gt;4729&lt;/span&gt;&lt;span class="w"&gt; &lt;/span&gt;&lt;span class="err"&gt;→&lt;/span&gt;&lt;span class="w"&gt; &lt;/span&gt;&lt;span class="p"&gt;[&lt;/span&gt;&lt;span class="mf"&gt;0.12&lt;/span&gt;&lt;span class="p"&gt;,&lt;/span&gt;&lt;span class="w"&gt; &lt;/span&gt;&lt;span class="mf"&gt;-0.08&lt;/span&gt;&lt;span class="p"&gt;,&lt;/span&gt;&lt;span class="w"&gt; &lt;/span&gt;&lt;span class="mf"&gt;0.44&lt;/span&gt;&lt;span class="p"&gt;,&lt;/span&gt;&lt;span class="w"&gt; &lt;/span&gt;&lt;span class="err"&gt;...&lt;/span&gt;&lt;span class="p"&gt;,&lt;/span&gt;&lt;span class="w"&gt; &lt;/span&gt;&lt;span class="mf"&gt;0.91&lt;/span&gt;&lt;span class="p"&gt;]&lt;/span&gt;&lt;span class="w"&gt;   &lt;/span&gt;&lt;span class="err"&gt;(&lt;/span&gt;&lt;span class="mi"&gt;1&lt;/span&gt;&lt;span class="err"&gt;,&lt;/span&gt;&lt;span class="mi"&gt;536&lt;/span&gt;&lt;span class="w"&gt; &lt;/span&gt;&lt;span class="err"&gt;numbers)&lt;/span&gt;&lt;span class="w"&gt;
&lt;/span&gt;&lt;/code&gt;&lt;/pre&gt;

&lt;/div&gt;



&lt;h4&gt;
  
  
  Step 1d — Vector Database
&lt;/h4&gt;

&lt;p&gt;We store all 120,000 of these vectors in a &lt;strong&gt;vector database&lt;/strong&gt; — something like FAISS, Pinecone, Weaviate, Milvus, or Qdrant. The database indexes them so we can search across all of them in &lt;em&gt;milliseconds&lt;/em&gt;.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;Indexing is done.&lt;/strong&gt; This usually runs as a background job, and you only re-run it when documents change.&lt;/p&gt;




&lt;h3&gt;
  
  
  Phase 2: Retrieval (Happens at Query Time)
&lt;/h3&gt;

&lt;p&gt;Now a developer types:&lt;/p&gt;

&lt;blockquote&gt;
&lt;p&gt;"How do I rotate the database credentials for the billing service?"&lt;/p&gt;
&lt;/blockquote&gt;

&lt;h4&gt;
  
  
  Step 2a — User Query
&lt;/h4&gt;

&lt;p&gt;The question comes in as plain text.&lt;/p&gt;

&lt;h4&gt;
  
  
  Step 2b — Query Embedding
&lt;/h4&gt;

&lt;p&gt;We run the &lt;strong&gt;same embedding model&lt;/strong&gt; on the question, producing a query vector:&lt;br&gt;
&lt;/p&gt;

&lt;div class="highlight js-code-highlight"&gt;
&lt;pre class="highlight json"&gt;&lt;code&gt;&lt;span class="err"&gt;Query&lt;/span&gt;&lt;span class="w"&gt; &lt;/span&gt;&lt;span class="err"&gt;→&lt;/span&gt;&lt;span class="w"&gt; &lt;/span&gt;&lt;span class="p"&gt;[&lt;/span&gt;&lt;span class="mf"&gt;0.15&lt;/span&gt;&lt;span class="p"&gt;,&lt;/span&gt;&lt;span class="w"&gt; &lt;/span&gt;&lt;span class="mf"&gt;-0.11&lt;/span&gt;&lt;span class="p"&gt;,&lt;/span&gt;&lt;span class="w"&gt; &lt;/span&gt;&lt;span class="mf"&gt;0.48&lt;/span&gt;&lt;span class="p"&gt;,&lt;/span&gt;&lt;span class="w"&gt; &lt;/span&gt;&lt;span class="err"&gt;...&lt;/span&gt;&lt;span class="p"&gt;,&lt;/span&gt;&lt;span class="w"&gt; &lt;/span&gt;&lt;span class="mf"&gt;0.87&lt;/span&gt;&lt;span class="p"&gt;]&lt;/span&gt;&lt;span class="w"&gt;
&lt;/span&gt;&lt;/code&gt;&lt;/pre&gt;

&lt;/div&gt;



&lt;p&gt;This is critical: you must embed the query with the &lt;em&gt;same model&lt;/em&gt; you used to embed the chunks, otherwise the vectors live in different spaces and similarity becomes meaningless.&lt;/p&gt;

&lt;h4&gt;
  
  
  Step 2c — Similarity Search
&lt;/h4&gt;

&lt;p&gt;Now we ask the vector database: &lt;strong&gt;"Which chunks have vectors closest to this query vector?"&lt;/strong&gt;&lt;/p&gt;

&lt;p&gt;Closeness is measured with a &lt;strong&gt;similarity metric&lt;/strong&gt;, most commonly &lt;strong&gt;cosine similarity&lt;/strong&gt; — it measures the angle between two vectors. The smaller the angle, the more similar the meaning.&lt;/p&gt;

&lt;p&gt;Under the hood, the database uses &lt;strong&gt;Approximate Nearest Neighbors (ANN)&lt;/strong&gt; tricks to search 120,000 vectors in ~5 milliseconds instead of comparing one by one.&lt;/p&gt;

&lt;h4&gt;
  
  
  Step 2d — Relevant Passages
&lt;/h4&gt;

&lt;p&gt;The database returns the &lt;strong&gt;top-k&lt;/strong&gt; most similar chunks (typically k=3 to k=10). For our query, we might get:&lt;br&gt;
&lt;/p&gt;

&lt;div class="highlight js-code-highlight"&gt;
&lt;pre class="highlight plaintext"&gt;&lt;code&gt;1. Chunk #4729 (score 0.94) — billing-service runbook, credential rotation
2. Chunk #3180 (score 0.89) — AWS Secrets Manager general guide
3. Chunk #5512 (score 0.85) — rolling restart playbook
&lt;/code&gt;&lt;/pre&gt;

&lt;/div&gt;



&lt;p&gt;These are the passages most likely to contain the answer.&lt;/p&gt;




&lt;h3&gt;
  
  
  Phase 3: Augmentation
&lt;/h3&gt;

&lt;p&gt;Now we have relevant chunks, but we don't just &lt;em&gt;show&lt;/em&gt; them to the user. We want the LLM to write a nice, synthesized answer using them.&lt;/p&gt;

&lt;h4&gt;
  
  
  Step 3a — Original Prompt
&lt;/h4&gt;

&lt;p&gt;The user's raw question:&lt;br&gt;
&lt;/p&gt;

&lt;div class="highlight js-code-highlight"&gt;
&lt;pre class="highlight plaintext"&gt;&lt;code&gt;"How do I rotate the database credentials for the billing service?"
&lt;/code&gt;&lt;/pre&gt;

&lt;/div&gt;



&lt;h4&gt;
  
  
  Step 3b — Augmented Prompt
&lt;/h4&gt;

&lt;p&gt;We wrap it in a &lt;strong&gt;prompt template&lt;/strong&gt; that injects the retrieved chunks as context:&lt;br&gt;
&lt;/p&gt;

&lt;div class="highlight js-code-highlight"&gt;
&lt;pre class="highlight plaintext"&gt;&lt;code&gt;You are Acme Corp's internal engineering assistant.
Answer the user's question using ONLY the context below.
If the answer isn't in the context, say you don't know.

---CONTEXT---
[Chunk #4729]: To rotate database credentials for the billing
service: 1. Generate a new password in AWS Secrets Manager...
[Chunk #3180]: AWS Secrets Manager allows you to store and
rotate database credentials...
[Chunk #5512]: Rolling restarts can be triggered via kubectl
rollout restart deploy/&amp;lt;service-name&amp;gt;...
---END CONTEXT---

USER QUESTION: How do I rotate the database credentials
for the billing service?
&lt;/code&gt;&lt;/pre&gt;

&lt;/div&gt;



&lt;p&gt;This is the "A" in RAG — &lt;strong&gt;Augmentation&lt;/strong&gt;. We're augmenting the prompt with retrieved knowledge.&lt;/p&gt;




&lt;h3&gt;
  
  
  Phase 4: Generation
&lt;/h3&gt;

&lt;h4&gt;
  
  
  Step 4a — LLM Call
&lt;/h4&gt;

&lt;p&gt;We send the augmented prompt to an LLM (GPT-3/4/5, Claude, LLaMA, T5, Mistral — your choice).&lt;/p&gt;

&lt;p&gt;Because the model has the actual documentation &lt;em&gt;right there in its context window&lt;/em&gt;, it doesn't need to remember anything — it just reads and synthesizes.&lt;/p&gt;

&lt;h4&gt;
  
  
  Step 4b — Final Response
&lt;/h4&gt;

&lt;p&gt;The model produces a grounded answer:&lt;/p&gt;

&lt;blockquote&gt;
&lt;p&gt;"To rotate the database credentials for the billing service:&lt;/p&gt;

&lt;ol&gt;
&lt;li&gt;Generate a new password in AWS Secrets Manager.&lt;/li&gt;
&lt;li&gt;Update the &lt;code&gt;billing-db&lt;/code&gt; secret with the new value.&lt;/li&gt;
&lt;li&gt;Trigger a rolling restart with &lt;code&gt;kubectl rollout restart deploy/billing&lt;/code&gt;.&lt;/li&gt;
&lt;li&gt;Verify that the health endpoints return 200 OK.&lt;/li&gt;
&lt;li&gt;Revoke the old credentials after a 24-hour grace period.&lt;/li&gt;
&lt;/ol&gt;

&lt;p&gt;Source: runbooks/billing-service.md"&lt;/p&gt;
&lt;/blockquote&gt;

&lt;p&gt;Notice three things:&lt;/p&gt;

&lt;ol&gt;
&lt;li&gt;The answer is &lt;strong&gt;accurate&lt;/strong&gt; — it comes from your actual docs.&lt;/li&gt;
&lt;li&gt;The answer is &lt;strong&gt;current&lt;/strong&gt; — if you update the runbook, the next query uses the new version. No retraining needed.&lt;/li&gt;
&lt;li&gt;The answer can be &lt;strong&gt;cited&lt;/strong&gt; — you know exactly which document it came from.&lt;/li&gt;
&lt;/ol&gt;

&lt;p&gt;&lt;strong&gt;That's the whole RAG pipeline.&lt;/strong&gt; Indexing → Retrieval → Augmentation → Generation.&lt;/p&gt;




&lt;h2&gt;
  
  
  4. The Retrieval Component in Detail
&lt;/h2&gt;

&lt;p&gt;Three pieces make retrieval work:&lt;/p&gt;

&lt;h3&gt;
  
  
  Embedding Models
&lt;/h3&gt;

&lt;p&gt;The model that turns text into vectors. Examples: &lt;strong&gt;BERT&lt;/strong&gt;, &lt;code&gt;text-embedding-3-small&lt;/code&gt;, Cohere Embed, Sentence-BERT. Choose one that's trained well for your language and domain.&lt;/p&gt;

&lt;h3&gt;
  
  
  Vector Stores
&lt;/h3&gt;

&lt;p&gt;Databases optimized for vector similarity search. Popular options: &lt;strong&gt;FAISS&lt;/strong&gt; (local, Facebook), &lt;strong&gt;Pinecone&lt;/strong&gt; (managed), &lt;strong&gt;Weaviate&lt;/strong&gt;, &lt;strong&gt;Milvus&lt;/strong&gt;, &lt;strong&gt;Qdrant&lt;/strong&gt;, and pgvector (Postgres extension).&lt;/p&gt;

&lt;h3&gt;
  
  
  Similarity Metrics
&lt;/h3&gt;

&lt;p&gt;How we measure "closeness" between vectors. The go-to is &lt;strong&gt;cosine similarity&lt;/strong&gt;, but Euclidean distance and dot product also show up. Cosine similarity is popular because it ignores vector length and focuses on &lt;em&gt;direction&lt;/em&gt; — which is what semantic meaning lives in.&lt;/p&gt;




&lt;h2&gt;
  
  
  5. Augmentation &amp;amp; Generation in Detail
&lt;/h2&gt;

&lt;h3&gt;
  
  
  Prompt Templates
&lt;/h3&gt;

&lt;p&gt;The structure that tells the LLM how to use the retrieved context. Good templates specify:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;The assistant's role&lt;/li&gt;
&lt;li&gt;What to do if context is missing&lt;/li&gt;
&lt;li&gt;Output format (JSON, bullet points, prose)&lt;/li&gt;
&lt;li&gt;Citation rules&lt;/li&gt;
&lt;/ul&gt;

&lt;h3&gt;
  
  
  Managing Model Context
&lt;/h3&gt;

&lt;p&gt;The LLM only has so much context window. If retrieval returns 30 chunks but each chunk is 500 tokens, that's 15,000 tokens just for context. You have to:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;Pick top-k carefully (more isn't always better)&lt;/li&gt;
&lt;li&gt;Rerank retrieved chunks&lt;/li&gt;
&lt;li&gt;Sometimes summarize chunks before injection&lt;/li&gt;
&lt;/ul&gt;

&lt;h3&gt;
  
  
  LLM Choices
&lt;/h3&gt;

&lt;p&gt;Any generative LLM can work: &lt;strong&gt;GPT-3/4/5&lt;/strong&gt;, &lt;strong&gt;T5&lt;/strong&gt;, &lt;strong&gt;LLaMA&lt;/strong&gt;, &lt;strong&gt;Claude&lt;/strong&gt;, &lt;strong&gt;Mistral&lt;/strong&gt;, &lt;strong&gt;Gemini&lt;/strong&gt;. The RAG pipeline is mostly model-agnostic.&lt;/p&gt;




&lt;h2&gt;
  
  
  6. Applications and Benefits
&lt;/h2&gt;

&lt;p&gt;RAG is behind a huge number of real-world AI products:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;
&lt;strong&gt;Knowledge-centric chatbots&lt;/strong&gt; — customer support bots grounded in your docs&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;Document summarization &amp;amp; Q&amp;amp;A&lt;/strong&gt; — ask questions about contracts, research papers, medical records&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;Enterprise search &amp;amp; knowledge management&lt;/strong&gt; — "Glean for your company" style tools&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;Benefits:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;No retraining required when data changes&lt;/li&gt;
&lt;li&gt;Answers are traceable back to sources&lt;/li&gt;
&lt;li&gt;Private data stays in your vector DB — never baked into model weights&lt;/li&gt;
&lt;li&gt;Cheaper than fine-tuning for most use cases&lt;/li&gt;
&lt;li&gt;Can mix multiple knowledge bases with one model&lt;/li&gt;
&lt;/ul&gt;




&lt;h2&gt;
  
  
  7. Challenges and Future Directions
&lt;/h2&gt;

&lt;p&gt;RAG isn't magic. Here are the real tradeoffs.&lt;/p&gt;

&lt;h3&gt;
  
  
  Source Reliability &amp;amp; Bias
&lt;/h3&gt;

&lt;p&gt;Garbage in, garbage out. If your knowledge base has outdated or biased content, your RAG system will confidently repeat it. Curation matters.&lt;/p&gt;

&lt;h3&gt;
  
  
  Latency &amp;amp; System Complexity
&lt;/h3&gt;

&lt;p&gt;A RAG query is actually: embed → ANN search → rerank → build prompt → LLM call. That's a lot of moving parts. Each step adds latency, and each step can fail. Production RAG systems require serious observability.&lt;/p&gt;

&lt;h3&gt;
  
  
  Privacy &amp;amp; Security Safeguards
&lt;/h3&gt;

&lt;p&gt;Your vector DB now contains sensitive content. Access control, encryption, and embedding leakage (yes, embeddings can leak information) all need attention.&lt;/p&gt;

&lt;h3&gt;
  
  
  Research Frontiers
&lt;/h3&gt;

&lt;p&gt;Where RAG is heading:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;
&lt;strong&gt;Multi-hop retrieval&lt;/strong&gt; — answering questions that need multiple retrieval rounds (e.g., "Find X, then look up Y for X")&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;Adapters&lt;/strong&gt; — lightweight modules that specialize the LLM for using retrieved content better&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;Self-improving retrieval&lt;/strong&gt; — the model learns which retrievals helped and which didn't&lt;/li&gt;
&lt;/ul&gt;




&lt;h2&gt;
  
  
  Wrapping Up — The TL;DR
&lt;/h2&gt;

&lt;p&gt;&lt;strong&gt;What is RAG?&lt;/strong&gt;&lt;br&gt;
A framework that lets an LLM look things up in a knowledge base before answering, so its responses are grounded in real, current, specific information.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;Why does it exist?&lt;/strong&gt;&lt;br&gt;
Because retraining is expensive, LLMs hallucinate, and most real-world apps need to answer from &lt;em&gt;your&lt;/em&gt; data — not what the model memorized during training.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;How does the pipeline work?&lt;/strong&gt;&lt;/p&gt;

&lt;ol&gt;
&lt;li&gt;
&lt;strong&gt;Indexing&lt;/strong&gt; — Chunk documents → embed each chunk → store vectors in a vector DB.&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;Retrieval&lt;/strong&gt; — Embed the user query → find nearest chunks with cosine similarity.&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;Augmentation&lt;/strong&gt; — Inject retrieved chunks into a prompt template.&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;Generation&lt;/strong&gt; — Send augmented prompt to an LLM → return grounded answer.&lt;/li&gt;
&lt;/ol&gt;

&lt;p&gt;&lt;strong&gt;Where do you use it?&lt;/strong&gt;&lt;br&gt;
Anywhere you need an AI that can answer from your own content: internal docs, product support, legal research, medical Q&amp;amp;A, academic search, knowledge management, developer assistants, and more.&lt;/p&gt;

&lt;p&gt;Once you understand the four-phase pipeline — &lt;strong&gt;Indexing, Retrieval, Augmentation, Generation&lt;/strong&gt; — every RAG system you encounter becomes a variation on the same theme.&lt;/p&gt;




&lt;p&gt;&lt;em&gt;If this helped you finally "get" RAG, drop a reaction. More notes coming soon.&lt;/em&gt;&lt;/p&gt;

</description>
      <category>ai</category>
      <category>beginners</category>
      <category>rag</category>
      <category>tutorial</category>
    </item>
    <item>
      <title>Embeddings Explained: The Secret Language AI Uses to Understand the World</title>
      <dc:creator>Ege Pakten</dc:creator>
      <pubDate>Sat, 18 Apr 2026 09:20:32 +0000</pubDate>
      <link>https://dev.to/egepakten/embeddings-explained-the-secret-language-ai-uses-to-understand-the-world-3e0o</link>
      <guid>https://dev.to/egepakten/embeddings-explained-the-secret-language-ai-uses-to-understand-the-world-3e0o</guid>
      <description>&lt;blockquote&gt;
&lt;p&gt;If you've ever wondered how ChatGPT "knows" that &lt;em&gt;king&lt;/em&gt; and &lt;em&gt;queen&lt;/em&gt; are related, or how Spotify recommends songs you actually like, the answer is almost always the same: &lt;strong&gt;embeddings&lt;/strong&gt;. This post breaks down what embeddings are, how they work, where they're used, and what you can actually &lt;em&gt;do&lt;/em&gt; with them — no PhD required.&lt;/p&gt;
&lt;/blockquote&gt;




&lt;h2&gt;
  
  
  1. What Are Embeddings?
&lt;/h2&gt;

&lt;p&gt;At their core, &lt;strong&gt;embeddings are just numbers&lt;/strong&gt; — more specifically, a list of numbers (a vector) that represents something like a word, a sentence, an image, or even a user.&lt;/p&gt;

&lt;p&gt;Computers don't understand the word "cat." They understand numbers. So we need a way to turn "cat" into numbers &lt;em&gt;in a way that preserves its meaning&lt;/em&gt;. That's what an embedding does.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;Simple example:&lt;/strong&gt;&lt;br&gt;
&lt;/p&gt;

&lt;div class="highlight js-code-highlight"&gt;
&lt;pre class="highlight json"&gt;&lt;code&gt;&lt;span class="s2"&gt;"cat"&lt;/span&gt;&lt;span class="w"&gt;  &lt;/span&gt;&lt;span class="err"&gt;→&lt;/span&gt;&lt;span class="w"&gt; &lt;/span&gt;&lt;span class="p"&gt;[&lt;/span&gt;&lt;span class="mf"&gt;0.21&lt;/span&gt;&lt;span class="p"&gt;,&lt;/span&gt;&lt;span class="w"&gt; &lt;/span&gt;&lt;span class="mf"&gt;-0.44&lt;/span&gt;&lt;span class="p"&gt;,&lt;/span&gt;&lt;span class="w"&gt; &lt;/span&gt;&lt;span class="mf"&gt;0.89&lt;/span&gt;&lt;span class="p"&gt;,&lt;/span&gt;&lt;span class="w"&gt; &lt;/span&gt;&lt;span class="mf"&gt;0.12&lt;/span&gt;&lt;span class="p"&gt;,&lt;/span&gt;&lt;span class="w"&gt; &lt;/span&gt;&lt;span class="err"&gt;...&lt;/span&gt;&lt;span class="p"&gt;,&lt;/span&gt;&lt;span class="w"&gt; &lt;/span&gt;&lt;span class="mf"&gt;0.03&lt;/span&gt;&lt;span class="p"&gt;]&lt;/span&gt;&lt;span class="w"&gt;   &lt;/span&gt;&lt;span class="err"&gt;(e.g.,&lt;/span&gt;&lt;span class="w"&gt; &lt;/span&gt;&lt;span class="mi"&gt;768&lt;/span&gt;&lt;span class="w"&gt; &lt;/span&gt;&lt;span class="err"&gt;numbers)&lt;/span&gt;&lt;span class="w"&gt;
&lt;/span&gt;&lt;span class="s2"&gt;"dog"&lt;/span&gt;&lt;span class="w"&gt;  &lt;/span&gt;&lt;span class="err"&gt;→&lt;/span&gt;&lt;span class="w"&gt; &lt;/span&gt;&lt;span class="p"&gt;[&lt;/span&gt;&lt;span class="mf"&gt;0.19&lt;/span&gt;&lt;span class="p"&gt;,&lt;/span&gt;&lt;span class="w"&gt; &lt;/span&gt;&lt;span class="mf"&gt;-0.41&lt;/span&gt;&lt;span class="p"&gt;,&lt;/span&gt;&lt;span class="w"&gt; &lt;/span&gt;&lt;span class="mf"&gt;0.85&lt;/span&gt;&lt;span class="p"&gt;,&lt;/span&gt;&lt;span class="w"&gt; &lt;/span&gt;&lt;span class="mf"&gt;0.15&lt;/span&gt;&lt;span class="p"&gt;,&lt;/span&gt;&lt;span class="w"&gt; &lt;/span&gt;&lt;span class="err"&gt;...&lt;/span&gt;&lt;span class="p"&gt;,&lt;/span&gt;&lt;span class="w"&gt; &lt;/span&gt;&lt;span class="mf"&gt;0.06&lt;/span&gt;&lt;span class="p"&gt;]&lt;/span&gt;&lt;span class="w"&gt;
&lt;/span&gt;&lt;span class="s2"&gt;"car"&lt;/span&gt;&lt;span class="w"&gt;  &lt;/span&gt;&lt;span class="err"&gt;→&lt;/span&gt;&lt;span class="w"&gt; &lt;/span&gt;&lt;span class="p"&gt;[&lt;/span&gt;&lt;span class="mf"&gt;-0.72&lt;/span&gt;&lt;span class="p"&gt;,&lt;/span&gt;&lt;span class="w"&gt; &lt;/span&gt;&lt;span class="mf"&gt;0.31&lt;/span&gt;&lt;span class="p"&gt;,&lt;/span&gt;&lt;span class="w"&gt; &lt;/span&gt;&lt;span class="mf"&gt;-0.12&lt;/span&gt;&lt;span class="p"&gt;,&lt;/span&gt;&lt;span class="w"&gt; &lt;/span&gt;&lt;span class="mf"&gt;0.88&lt;/span&gt;&lt;span class="p"&gt;,&lt;/span&gt;&lt;span class="w"&gt; &lt;/span&gt;&lt;span class="err"&gt;...&lt;/span&gt;&lt;span class="p"&gt;,&lt;/span&gt;&lt;span class="w"&gt; &lt;/span&gt;&lt;span class="mf"&gt;-0.44&lt;/span&gt;&lt;span class="p"&gt;]&lt;/span&gt;&lt;span class="w"&gt;
&lt;/span&gt;&lt;/code&gt;&lt;/pre&gt;

&lt;/div&gt;



&lt;p&gt;Notice how &lt;code&gt;cat&lt;/code&gt; and &lt;code&gt;dog&lt;/code&gt; have similar-looking numbers, while &lt;code&gt;car&lt;/code&gt; looks very different. That's not an accident — it's the whole point. &lt;strong&gt;Similar meanings produce similar vectors.&lt;/strong&gt;&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;The key idea:&lt;/strong&gt; Embeddings are a way of placing concepts on a giant invisible map, where things that mean similar things end up close together, and things that mean different things end up far apart.&lt;/p&gt;




&lt;h2&gt;
  
  
  2. How Do Embeddings Work?
&lt;/h2&gt;

&lt;p&gt;Embeddings don't appear out of nowhere. They're &lt;em&gt;learned&lt;/em&gt; by a model during training. There are three core mechanisms worth understanding:&lt;/p&gt;

&lt;h3&gt;
  
  
  a) Self-Supervised Contrastive Learning
&lt;/h3&gt;

&lt;p&gt;The model looks at massive amounts of raw data (text, images, etc.) and learns by playing a game: &lt;strong&gt;"pull similar things together, push dissimilar things apart."&lt;/strong&gt;&lt;/p&gt;

&lt;p&gt;For example, during training:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;A sentence and a slightly rephrased version of it → should be &lt;em&gt;close&lt;/em&gt;
&lt;/li&gt;
&lt;li&gt;A sentence about cats and a sentence about quantum physics → should be &lt;em&gt;far apart&lt;/em&gt;
&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;No human has to label anything. The model figures it out from the structure of the data itself. That's the "self-supervised" part.&lt;/p&gt;

&lt;h3&gt;
  
  
  b) Contextual Embeddings
&lt;/h3&gt;

&lt;p&gt;Older embeddings gave every word a &lt;em&gt;single&lt;/em&gt; fixed vector. That's a problem, because words can mean different things in different contexts:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;"I deposited money at the &lt;strong&gt;bank&lt;/strong&gt;." (financial institution)&lt;/li&gt;
&lt;li&gt;"We had a picnic by the river &lt;strong&gt;bank&lt;/strong&gt;." (side of a river)&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;Modern embeddings (like those from BERT or GPT) generate a &lt;em&gt;different&lt;/em&gt; vector depending on the surrounding words. The model reads the whole sentence first, then decides what "bank" means &lt;em&gt;here&lt;/em&gt;.&lt;/p&gt;

&lt;h3&gt;
  
  
  c) Dimensionality Reduction
&lt;/h3&gt;

&lt;p&gt;Raw data (like a full image or a giant sparse word matrix) has way too many numbers. Embeddings compress this into a smaller, dense, meaningful representation — typically &lt;strong&gt;256, 512, 768, or 1536 dimensions&lt;/strong&gt;.&lt;/p&gt;

&lt;p&gt;Think of it like writing a movie review: instead of describing every pixel in every frame, you capture the essence in a paragraph.&lt;/p&gt;




&lt;h2&gt;
  
  
  3. Embeddings Deep Dive
&lt;/h2&gt;

&lt;p&gt;Let's go one layer deeper. Three properties make embeddings actually useful:&lt;/p&gt;

&lt;h3&gt;
  
  
  Mapping to a Vector Space
&lt;/h3&gt;

&lt;p&gt;Every piece of data becomes a point in a multi-dimensional space. You can't visualize 768 dimensions, but you can imagine a 3D version:&lt;br&gt;
&lt;/p&gt;

&lt;div class="highlight js-code-highlight"&gt;
&lt;pre class="highlight plaintext"&gt;&lt;code&gt;        cat •
   dog •
         • kitten
                               • airplane
                                      • rocket
&lt;/code&gt;&lt;/pre&gt;

&lt;/div&gt;



&lt;p&gt;Cats, dogs, and kittens cluster together. Airplanes and rockets cluster together. The space itself has &lt;strong&gt;meaning baked into distance and direction&lt;/strong&gt;.&lt;/p&gt;

&lt;h3&gt;
  
  
  Preserving Semantic Relationships
&lt;/h3&gt;

&lt;p&gt;The famous example:&lt;br&gt;
&lt;/p&gt;

&lt;div class="highlight js-code-highlight"&gt;
&lt;pre class="highlight plaintext"&gt;&lt;code&gt;king - man + woman ≈ queen
&lt;/code&gt;&lt;/pre&gt;

&lt;/div&gt;



&lt;p&gt;You can literally do math on meanings. This works because "royalty," "gender," and other concepts become &lt;em&gt;directions&lt;/em&gt; in the embedding space.&lt;/p&gt;

&lt;h3&gt;
  
  
  Efficient Processing
&lt;/h3&gt;

&lt;p&gt;Once everything is a vector, you can do fast operations on millions or billions of items:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;Compare two things? → compute cosine similarity (one quick math operation)&lt;/li&gt;
&lt;li&gt;Find the nearest match? → use Approximate Nearest Neighbors (ANN)&lt;/li&gt;
&lt;li&gt;Cluster similar items? → run k-means&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;This is why embeddings power huge real-world systems.&lt;/p&gt;




&lt;h2&gt;
  
  
  4. Types of Embeddings
&lt;/h2&gt;

&lt;p&gt;Not all embeddings are created equal. Here are the major families:&lt;/p&gt;

&lt;h3&gt;
  
  
  Static Word Embeddings — Word2Vec, GloVe
&lt;/h3&gt;

&lt;p&gt;These were the breakthrough that started it all. Each word gets exactly one vector, learned from how words co-occur in giant text corpora.&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;
&lt;strong&gt;Pros:&lt;/strong&gt; Fast, simple, very cheap to use.&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;Cons:&lt;/strong&gt; Can't handle context ("bank" is always the same vector).&lt;/li&gt;
&lt;/ul&gt;

&lt;h3&gt;
  
  
  Contextual Embeddings — ELMo, BERT
&lt;/h3&gt;

&lt;p&gt;These read the whole sentence and produce a vector for each word &lt;em&gt;in context&lt;/em&gt;.&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;
&lt;strong&gt;Pros:&lt;/strong&gt; Much more accurate for real language understanding.&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;Cons:&lt;/strong&gt; Heavier to compute, need a bigger model.&lt;/li&gt;
&lt;/ul&gt;

&lt;h3&gt;
  
  
  Sentence / Document Embeddings — Universal Sentence Encoder, Sentence-BERT
&lt;/h3&gt;

&lt;p&gt;Instead of one vector per word, you get one vector for an entire sentence, paragraph, or document. Super useful for search, clustering, and classification.&lt;/p&gt;

&lt;h3&gt;
  
  
  Multimodal Embeddings — CLIP
&lt;/h3&gt;

&lt;p&gt;These put text &lt;em&gt;and&lt;/em&gt; images in the &lt;em&gt;same&lt;/em&gt; vector space. A photo of a beach and the sentence "a sunny day at the ocean" end up close together. This is what powers most modern image search and text-to-image tools.&lt;/p&gt;




&lt;h2&gt;
  
  
  5. Key Use Cases — Where Embeddings Actually Shine
&lt;/h2&gt;

&lt;p&gt;This is the "so what" section. Here's what you can &lt;em&gt;build&lt;/em&gt; with embeddings.&lt;/p&gt;

&lt;h3&gt;
  
  
  Semantic Search
&lt;/h3&gt;

&lt;p&gt;Forget keyword matching. With embeddings, a user can search for:&lt;/p&gt;

&lt;blockquote&gt;
&lt;p&gt;"How do I stop my laptop from overheating?"&lt;/p&gt;
&lt;/blockquote&gt;

&lt;p&gt;…and you can return a document that says:&lt;/p&gt;

&lt;blockquote&gt;
&lt;p&gt;"Thermal management tips for portable computers"&lt;/p&gt;
&lt;/blockquote&gt;

&lt;p&gt;No shared keywords, but the meaning is almost identical — and the vectors are close. This is the foundation of modern search, documentation bots, and &lt;strong&gt;RAG (Retrieval Augmented Generation)&lt;/strong&gt;.&lt;/p&gt;

&lt;h3&gt;
  
  
  Clustering and Recommendation
&lt;/h3&gt;

&lt;p&gt;Group similar items automatically. Examples:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;Netflix grouping movies you'd like based on what you've watched&lt;/li&gt;
&lt;li&gt;Spotify building "Discover Weekly" playlists&lt;/li&gt;
&lt;li&gt;Customer segmentation for marketing&lt;/li&gt;
&lt;li&gt;Automatically grouping support tickets by topic&lt;/li&gt;
&lt;/ul&gt;

&lt;h3&gt;
  
  
  Anomaly Detection
&lt;/h3&gt;

&lt;p&gt;If everything "normal" clusters in one region of the vector space, then anything far away from that cluster is probably weird. This is used for:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;Credit card fraud detection&lt;/li&gt;
&lt;li&gt;Network intrusion detection&lt;/li&gt;
&lt;li&gt;Spotting defective products on factory lines&lt;/li&gt;
&lt;li&gt;Finding unusual user behavior&lt;/li&gt;
&lt;/ul&gt;

&lt;h3&gt;
  
  
  Classification
&lt;/h3&gt;

&lt;p&gt;Train a lightweight classifier on top of embeddings for things like spam detection, sentiment analysis, or intent recognition. You get great accuracy with almost no data.&lt;/p&gt;




&lt;h2&gt;
  
  
  6. Properties and Best Practices
&lt;/h2&gt;

&lt;p&gt;If you're going to actually use embeddings, here are the things that matter in practice.&lt;/p&gt;

&lt;h3&gt;
  
  
  Normalize Your Vectors
&lt;/h3&gt;

&lt;p&gt;Most similarity math works better when vectors are normalized to length 1. This means you're comparing &lt;strong&gt;direction&lt;/strong&gt;, not magnitude — which is usually what you want semantically.&lt;/p&gt;

&lt;h3&gt;
  
  
  Pick the Right Dimensionality
&lt;/h3&gt;

&lt;ul&gt;
&lt;li&gt;
&lt;strong&gt;Smaller (128–384):&lt;/strong&gt; Faster, cheaper storage, less memory. Good for mobile or massive-scale systems.&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;Larger (768–1536+):&lt;/strong&gt; More expressive, better accuracy, higher cost.&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;There's no free lunch. Start small, go bigger only if quality suffers.&lt;/p&gt;

&lt;h3&gt;
  
  
  Use Proper Indexing
&lt;/h3&gt;

&lt;p&gt;If you have millions of vectors, you can't compare them one by one. Use a vector database or library:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;
&lt;strong&gt;FAISS&lt;/strong&gt; (Facebook)&lt;/li&gt;
&lt;li&gt;&lt;strong&gt;Pinecone&lt;/strong&gt;&lt;/li&gt;
&lt;li&gt;&lt;strong&gt;Weaviate&lt;/strong&gt;&lt;/li&gt;
&lt;li&gt;&lt;strong&gt;Milvus&lt;/strong&gt;&lt;/li&gt;
&lt;li&gt;&lt;strong&gt;Qdrant&lt;/strong&gt;&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;These use tricks like ANN (Approximate Nearest Neighbors) to search billions of vectors in milliseconds.&lt;/p&gt;

&lt;h3&gt;
  
  
  Match the Embedding Model to Your Task
&lt;/h3&gt;

&lt;p&gt;A general-purpose embedding model is fine to start. But for specialized domains (medical, legal, code), a fine-tuned or domain-specific model will often double your accuracy.&lt;/p&gt;




&lt;h2&gt;
  
  
  7. Challenges and Limitations
&lt;/h2&gt;

&lt;p&gt;Embeddings are powerful, but they are not magic. Know the tradeoffs.&lt;/p&gt;

&lt;h3&gt;
  
  
  Memory and Compute
&lt;/h3&gt;

&lt;p&gt;Storing a billion 1536-dimensional float vectors is not cheap. High-dimensional search can get expensive quickly. You'll eventually need to think about quantization, sharding, and cost.&lt;/p&gt;

&lt;h3&gt;
  
  
  Privacy and Data Leakage
&lt;/h3&gt;

&lt;p&gt;Here's something that surprises most people: &lt;strong&gt;embeddings can leak information&lt;/strong&gt;. Even though a vector looks like "just numbers," research has shown attackers can sometimes &lt;em&gt;reconstruct&lt;/em&gt; or &lt;em&gt;infer&lt;/em&gt; parts of the original text from an embedding ("embedding inversion attacks").&lt;/p&gt;

&lt;p&gt;If you're embedding sensitive data (medical records, private messages, internal docs), treat the embeddings themselves as sensitive and protect them like you would the raw data.&lt;/p&gt;

&lt;h3&gt;
  
  
  Interpretability
&lt;/h3&gt;

&lt;p&gt;A 1536-dimensional vector is a black box. You can't easily explain &lt;em&gt;why&lt;/em&gt; two things are close. For regulated industries (finance, healthcare, EU AI Act compliance), this is a real concern.&lt;/p&gt;

&lt;h3&gt;
  
  
  Bias
&lt;/h3&gt;

&lt;p&gt;Embeddings learn from data, and data contains human biases. If your training text associates certain jobs with certain genders, your embeddings will too — and any downstream system will inherit that bias.&lt;/p&gt;




&lt;h2&gt;
  
  
  8. Future Directions
&lt;/h2&gt;

&lt;p&gt;Where is this all heading?&lt;/p&gt;

&lt;h3&gt;
  
  
  Hierarchical Embeddings
&lt;/h3&gt;

&lt;p&gt;Instead of one flat vector, future systems will learn representations at multiple levels — word → sentence → paragraph → document — all connected, all meaningful.&lt;/p&gt;

&lt;h3&gt;
  
  
  Continual and Federated Learning
&lt;/h3&gt;

&lt;p&gt;Today, most embedding models are trained once and frozen. The future is models that &lt;strong&gt;keep learning safely&lt;/strong&gt;, updating over time without forgetting old knowledge — and learning across devices (federated learning) without centralizing private data.&lt;/p&gt;

&lt;h3&gt;
  
  
  Richer Multimodal Embeddings
&lt;/h3&gt;

&lt;p&gt;Text + image is just the beginning. Expect models that unify &lt;strong&gt;text, image, audio, video, sensor data, and 3D scenes&lt;/strong&gt; all in the same space. Search "the sound of rain on a metal roof" and get back audio clips &lt;em&gt;and&lt;/em&gt; matching videos.&lt;/p&gt;




&lt;h2&gt;
  
  
  Wrapping Up — The TL;DR
&lt;/h2&gt;

&lt;p&gt;Let's tie it all together.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;What is an embedding?&lt;/strong&gt;&lt;br&gt;
A list of numbers that represents the &lt;em&gt;meaning&lt;/em&gt; of something (a word, image, sentence, user, product) in a way a computer can work with.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;Where are embeddings used?&lt;/strong&gt;&lt;br&gt;
Semantic search, RAG systems, recommendations, clustering, anomaly detection, fraud detection, classification, and multimodal search — basically anywhere you need a machine to understand "similarity" or "meaning."&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;What can you actually do with them?&lt;/strong&gt;&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;Build a search engine that understands meaning, not just keywords&lt;/li&gt;
&lt;li&gt;Power a chatbot with RAG using your own documents&lt;/li&gt;
&lt;li&gt;Detect fraud, spam, or defects&lt;/li&gt;
&lt;li&gt;Group customers, songs, movies, or articles automatically&lt;/li&gt;
&lt;li&gt;Search images with text, or text with images&lt;/li&gt;
&lt;li&gt;Add semantic understanding to almost any existing product&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;Embeddings are the quiet backbone of almost every modern AI system. You won't see them in the UI — but they're doing most of the real work behind the scenes. Once you understand embeddings, a huge amount of what seems "magical" about modern AI suddenly makes sense.&lt;/p&gt;




&lt;p&gt;*If this helped you click on what embeddings really are, drop a reaction. &lt;/p&gt;

</description>
      <category>ai</category>
      <category>webdev</category>
      <category>programming</category>
      <category>vectordatabase</category>
    </item>
    <item>
      <title>AI Agents Explained: 5 Types, Components, Frameworks, and Real-World Use Cases</title>
      <dc:creator>Ege Pakten</dc:creator>
      <pubDate>Sun, 12 Apr 2026 09:55:27 +0000</pubDate>
      <link>https://dev.to/egepakten/ai-agents-explained-5-types-components-frameworks-and-real-world-use-cases-52i4</link>
      <guid>https://dev.to/egepakten/ai-agents-explained-5-types-components-frameworks-and-real-world-use-cases-52i4</guid>
      <description>&lt;blockquote&gt;
&lt;h2&gt;
  
  
  AI agents are no longer just chatbots. They think, plan, use tools, and work together to solve complex problems autonomously. In this post, I break down the 5 types of AI agents, how they work, ReAct vs ReWOO frameworks, multi-agent systems, and guardrails — written so anyone can understand.
&lt;/h2&gt;
&lt;/blockquote&gt;

&lt;h2&gt;
  
  
  Introduction: AI Doesn't Just "Answer" Anymore
&lt;/h2&gt;

&lt;p&gt;For most of us, our first experience with AI was simple: ask a question, get an answer, done. But there is a quiet revolution happening. &lt;strong&gt;AI agents&lt;/strong&gt; are systems that don't just respond — they &lt;strong&gt;think&lt;/strong&gt;, &lt;strong&gt;plan&lt;/strong&gt;, &lt;strong&gt;use tools&lt;/strong&gt;, and even &lt;strong&gt;talk to each other&lt;/strong&gt; to solve complex tasks without human hand-holding.&lt;/p&gt;

&lt;p&gt;This post is based on notes I took from IBM's "What Are AI Agents?" page. My goal is to write something clear enough that even someone who has never heard the term "AI agent" can walk away understanding the full picture.&lt;/p&gt;

&lt;h2&gt;
  
  
  What Is an AI Agent?
&lt;/h2&gt;

&lt;p&gt;The simplest definition:&lt;/p&gt;

&lt;blockquote&gt;
&lt;p&gt;An &lt;strong&gt;AI agent&lt;/strong&gt; is an AI tool that can &lt;strong&gt;autonomously&lt;/strong&gt; perform complex tasks that would otherwise require human involvement.&lt;/p&gt;
&lt;/blockquote&gt;

&lt;p&gt;The key word is "autonomously." A regular chatbot needs you to tell it what to do at every step. An AI agent takes a &lt;strong&gt;goal&lt;/strong&gt;, &lt;strong&gt;plans&lt;/strong&gt; how to reach it on its own, &lt;strong&gt;selects&lt;/strong&gt; the right tools, and delivers the result.&lt;/p&gt;

&lt;h2&gt;
  
  
  Agentic vs Non-Agentic: What's the Difference?
&lt;/h2&gt;

&lt;p&gt;Not every AI is an "agent." This distinction matters.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;Non-Agentic AI:&lt;/strong&gt;&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;No tools — can only say what it already knows&lt;/li&gt;
&lt;li&gt;No memory — doesn't remember previous conversations&lt;/li&gt;
&lt;li&gt;Limited reasoning — can't plan ahead&lt;/li&gt;
&lt;li&gt;Needs constant human input to function&lt;/li&gt;
&lt;li&gt;Example: A basic FAQ chatbot&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;&lt;strong&gt;Agentic AI:&lt;/strong&gt;&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;Has access to tools — can search the web, call APIs, read files&lt;/li&gt;
&lt;li&gt;Has memory — remembers past interactions and learns from them&lt;/li&gt;
&lt;li&gt;Performs reasoning — creates step-by-step plans to reach goals&lt;/li&gt;
&lt;li&gt;Works autonomously — minimal human intervention needed&lt;/li&gt;
&lt;li&gt;Example: A hospital agent that handles insurance authorizations end to end&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;Think of it this way: non-agentic AI is a &lt;strong&gt;calculator&lt;/strong&gt; — press a button, get a result. Agentic AI is an &lt;strong&gt;accountant&lt;/strong&gt; — say "prepare my tax return" and it figures out the rest.&lt;/p&gt;

&lt;h2&gt;
  
  
  The 3 Core Components of Every AI Agent
&lt;/h2&gt;

&lt;p&gt;No matter how simple or complex, every AI agent runs on three building blocks:&lt;/p&gt;

&lt;h3&gt;
  
  
  1. Goal (from the User)
&lt;/h3&gt;

&lt;p&gt;Everything starts with a goal. The user tells the agent what needs to happen: "get insurance approval for this patient" or "find broken links on my website and report them." The agent takes this goal and performs &lt;strong&gt;task decomposition&lt;/strong&gt; — breaking it into smaller, manageable subtasks.&lt;/p&gt;

&lt;h3&gt;
  
  
  2. Tools
&lt;/h3&gt;

&lt;p&gt;Agents rarely have all the information they need on their own. So they reach out to external tools. These can be web search engines, APIs (weather, stock market, health records), databases, or even other AI agents (yes, an agent can use another agent as a tool).&lt;/p&gt;

&lt;h3&gt;
  
  
  3. Agentic Reasoning
&lt;/h3&gt;

&lt;p&gt;This is the agent's "brain." It evaluates the information it perceives, uses its memory, and selects the best action to move toward the goal. It uses conditional logic, heuristics, and feedback loops to make decisions continuously.&lt;/p&gt;

&lt;h2&gt;
  
  
  The 5 Types of AI Agents
&lt;/h2&gt;

&lt;p&gt;According to IBM, there are 5 distinct types, from simplest to most sophisticated:&lt;/p&gt;

&lt;h3&gt;
  
  
  1. Simple Reflex Agent
&lt;/h3&gt;

&lt;p&gt;The most basic type. It works on "if X happens, do Y" rules. No memory, no planning — just pre-programmed reflexes.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;Real-world analogy:&lt;/strong&gt; A night light that turns on automatically when the room gets dark. Sense → react, nothing else.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;Example:&lt;/strong&gt; A temperature sensor that turns on the AC when it hits 30°C. It doesn't know &lt;em&gt;why&lt;/em&gt; the temperature rose — it just follows the rule.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;Limitation:&lt;/strong&gt; If it encounters a situation it has no rule for, it's stuck.&lt;/p&gt;

&lt;h3&gt;
  
  
  2. Model-Based Reflex Agent
&lt;/h3&gt;

&lt;p&gt;One step above simple reflex. The key difference: it &lt;strong&gt;has memory&lt;/strong&gt;. It maintains a model of its environment and can operate in changing, partially observable settings.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;Real-world analogy:&lt;/strong&gt; A robot vacuum. It maps the room, remembers which areas it already cleaned, and navigates around furniture. It won't re-clean the same spot because it &lt;em&gt;remembers&lt;/em&gt; it was already done.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;Limitation:&lt;/strong&gt; Still rule-based — smarter, but can't go beyond its predefined rules.&lt;/p&gt;

&lt;h3&gt;
  
  
  3. Goal-Based Agent
&lt;/h3&gt;

&lt;p&gt;This is where things get serious. This agent doesn't just react — it has a &lt;strong&gt;goal&lt;/strong&gt; and uses &lt;strong&gt;planning&lt;/strong&gt; and &lt;strong&gt;reasoning&lt;/strong&gt; to achieve it.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;Real-world analogy:&lt;/strong&gt; GPS navigation. When you say "take me to the airport," it doesn't just look at the current street. It examines the map, calculates traffic, evaluates alternative routes, and picks the best path.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;Example:&lt;/strong&gt; A customer service agent. Goal: "resolve the customer's issue." The agent understands the problem, searches a knowledge base, escalates if needed — always moving toward the goal.&lt;/p&gt;

&lt;h3&gt;
  
  
  4. Utility-Based Agent
&lt;/h3&gt;

&lt;p&gt;Takes goal-based thinking one step further. This agent doesn't just achieve the goal — it aims to achieve it in the &lt;strong&gt;best possible way&lt;/strong&gt;. It uses a &lt;strong&gt;utility function&lt;/strong&gt; to evaluate different actions and pick the one that maximizes overall benefit.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;Real-world analogy:&lt;/strong&gt; An investment advisor. Doesn't just say "make money." It considers your risk tolerance, market conditions, time horizon, and recommends the strategy with the &lt;strong&gt;best balance&lt;/strong&gt; of risk and reward.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;Example:&lt;/strong&gt; A logistics optimization agent. There are many ways to deliver a package, but this agent calculates the cheapest, fastest, and least risky route.&lt;/p&gt;

&lt;h3&gt;
  
  
  5. Learning Agent
&lt;/h3&gt;

&lt;p&gt;The most advanced type. It has all the capabilities of the other types, plus the ability to &lt;strong&gt;learn&lt;/strong&gt;. It improves with every experience, updates its knowledge base, and makes better decisions over time.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;Real-world analogy:&lt;/strong&gt; A medical resident. On day one, they're inexperienced, but with every patient interaction they learn. Years later, their diagnoses are far more accurate because of accumulated experience.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;Example:&lt;/strong&gt; An e-commerce recommendation engine. At first, it recommends the same products to everyone. Over time, it learns individual preferences and delivers personalized suggestions.&lt;/p&gt;

&lt;h2&gt;
  
  
  AI Agent Frameworks: ReAct vs ReWOO
&lt;/h2&gt;

&lt;p&gt;How do agents actually "think"? Different architectural approaches have been developed. Two stand out:&lt;/p&gt;

&lt;h3&gt;
  
  
  ReAct (Reasoning + Acting)
&lt;/h3&gt;

&lt;p&gt;ReAct follows a "Think → Act → Observe" loop. The agent pauses at each step, thinks, and plans its next move.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;Flow:&lt;/strong&gt;&lt;/p&gt;

&lt;ol&gt;
&lt;li&gt;Agent looks at the goal and thinks: "I need this information to solve this"&lt;/li&gt;
&lt;li&gt;Uses a tool (e.g., web search)&lt;/li&gt;
&lt;li&gt;Observes the tool's result&lt;/li&gt;
&lt;li&gt;Thinks again: "Is this enough, or do I need more?"&lt;/li&gt;
&lt;li&gt;Loop continues until the goal is met&lt;/li&gt;
&lt;/ol&gt;

&lt;p&gt;&lt;strong&gt;Advantage:&lt;/strong&gt; Highly flexible — can adapt its strategy at every step and handle unexpected results.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;Disadvantage:&lt;/strong&gt; The "thinking" at each step burns extra tokens and takes time. Higher cost and latency.&lt;/p&gt;

&lt;h3&gt;
  
  
  ReWOO (Reasoning Without Observation)
&lt;/h3&gt;

&lt;p&gt;ReWOO follows a "plan first, execute later" approach. The agent does all its thinking &lt;strong&gt;upfront&lt;/strong&gt;, creates a complete plan, and then executes the steps sequentially.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;Flow:&lt;/strong&gt;&lt;/p&gt;

&lt;ol&gt;
&lt;li&gt;Agent looks at the goal and generates a &lt;strong&gt;complete plan&lt;/strong&gt;: "Step 1 is this, Step 2 is this, Step 3 is this"&lt;/li&gt;
&lt;li&gt;Executes the plan in order — doesn't rethink at each step&lt;/li&gt;
&lt;li&gt;Collects results&lt;/li&gt;
&lt;/ol&gt;

&lt;p&gt;&lt;strong&gt;Advantage:&lt;/strong&gt; Uses approximately &lt;strong&gt;80% fewer tokens&lt;/strong&gt; than ReAct. Much cheaper and faster.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;Disadvantage:&lt;/strong&gt; Struggles when it has limited context about its environment. Harder to handle unexpected situations since the plan was made upfront.&lt;/p&gt;

&lt;h3&gt;
  
  
  Which Should You Use?
&lt;/h3&gt;

&lt;ul&gt;
&lt;li&gt;
&lt;strong&gt;Dynamic, unpredictable environments&lt;/strong&gt; (customer service, chat, research) → &lt;strong&gt;ReAct&lt;/strong&gt;
&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;Well-defined, repetitive tasks&lt;/strong&gt; (data processing, report generation, batch operations) → &lt;strong&gt;ReWOO&lt;/strong&gt;
&lt;/li&gt;
&lt;/ul&gt;

&lt;h2&gt;
  
  
  Multi-Agent Systems
&lt;/h2&gt;

&lt;p&gt;A single agent is powerful, but some problems need a team. In &lt;strong&gt;multi-agent&lt;/strong&gt; systems, multiple specialized agents work together.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;Why isn't one agent enough?&lt;/strong&gt;&lt;/p&gt;

&lt;p&gt;Think of a human team: one person can't be the researcher, editor, &lt;em&gt;and&lt;/em&gt; graphic designer efficiently. Same logic applies to AI agents.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;Multi-agent structure example:&lt;/strong&gt;&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;
&lt;strong&gt;Researcher Agent&lt;/strong&gt;: Gathers data and finds sources&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;Critic Agent&lt;/strong&gt;: Audits data quality and fact-checks&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;Writer Agent&lt;/strong&gt;: Synthesizes results and produces output&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;Each agent focuses on its specialty. According to IBM, multi-agent systems produce &lt;strong&gt;higher quality and more reliable&lt;/strong&gt; outcomes than single agents.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;Real-world example:&lt;/strong&gt; A supply chain management system. One agent monitors inventory levels, another forecasts demand, a third optimizes logistics routes. They communicate with each other to make holistic decisions.&lt;/p&gt;

&lt;h2&gt;
  
  
  AI Agent Memory: Why It Matters
&lt;/h2&gt;

&lt;p&gt;The real power of agents lies in their memory. Traditional AI processes each task independently — it doesn't remember yesterday's conversation. Agents use multiple types of memory:&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;Working Memory:&lt;/strong&gt; Holds information about the active task. Like knowing which sources you've already checked during a research assignment.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;Long-term Memory:&lt;/strong&gt; Stores past experiences and learnings. When a similar task comes up again, the agent draws on previous experience.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;Procedural Memory:&lt;/strong&gt; Stores learned skills and automated behaviors. When an agent repeats a complex operation it has done before, it doesn't need to reason through every step from scratch.&lt;/p&gt;

&lt;h2&gt;
  
  
  AI Guardrails: The Safety Barriers
&lt;/h2&gt;

&lt;p&gt;Agents are powerful, but powerful tools can be dangerous. &lt;strong&gt;Guardrails&lt;/strong&gt; are boundaries an agent should never cross.&lt;/p&gt;

&lt;p&gt;IBM defines guardrails like highway barriers: they don't slow the car down, but they keep it from going off the road.&lt;/p&gt;

&lt;p&gt;What guardrails protect against:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;
&lt;strong&gt;Harmful content generation&lt;/strong&gt; — preventing offensive, misleading, or dangerous outputs&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;Sensitive data exposure&lt;/strong&gt; — preventing the agent from leaking personal or confidential information&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;Authority overreach&lt;/strong&gt; — preventing the agent from making decisions outside its scope&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;With proper guardrails, agents can improve continuously while staying safe.&lt;/p&gt;

&lt;h2&gt;
  
  
  Final Thoughts
&lt;/h2&gt;

&lt;p&gt;The thing that struck me most reading IBM's page was this: AI agents are no longer a technical concept sitting in research papers. They are &lt;strong&gt;real operational systems&lt;/strong&gt; running in hospitals, supply chains, customer service centers, and more.&lt;/p&gt;

&lt;p&gt;If you're new to this space, here's my suggested learning path:&lt;/p&gt;

&lt;ol&gt;
&lt;li&gt;
&lt;strong&gt;Understand the 5 agent types&lt;/strong&gt; — know which one fits which problem&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;Learn the ReAct vs ReWOO difference&lt;/strong&gt; — framework choice directly affects cost and performance&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;Never skip guardrails&lt;/strong&gt; — a powerful agent without controls is a dangerous agent&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;Think multi-agent&lt;/strong&gt; — a single agent can do everything, but a team always does it better&lt;/li&gt;
&lt;/ol&gt;

&lt;p&gt;I study one new AI topic per day from IBM's resources. This approach has been great for building deep understanding fast. I highly recommend the same: read, watch, but most importantly &lt;strong&gt;write&lt;/strong&gt;. Writing forces you to truly understand the material — far more than passive consumption.&lt;/p&gt;




&lt;p&gt;&lt;em&gt;This post is based on IBM's "What Are AI Agents?" page. For more detail, I recommend checking the original source.&lt;/em&gt;&lt;/p&gt;

</description>
      <category>ai</category>
      <category>agents</category>
      <category>llm</category>
      <category>webdev</category>
    </item>
    <item>
      <title>What Is AgentOps? A Beginner-Friendly Guide Using a Real Hospital Use Case</title>
      <dc:creator>Ege Pakten</dc:creator>
      <pubDate>Sat, 11 Apr 2026 08:20:07 +0000</pubDate>
      <link>https://dev.to/egepakten/what-is-agentops-a-beginner-friendly-guide-using-a-real-hospital-use-case-3okp</link>
      <guid>https://dev.to/egepakten/what-is-agentops-a-beginner-friendly-guide-using-a-real-hospital-use-case-3okp</guid>
      <description>&lt;p&gt;AI agents are no longer just chatbots; they are autonomous workers running real operations. But how do we know they are doing a good job? AgentOps is the answer. In this post, I break down the three layers of AgentOps using a hospital scenario anyone can follow.&lt;/p&gt;

&lt;h2&gt;
  
  
  Introduction: AI Agents Have Grown Up
&lt;/h2&gt;

&lt;p&gt;A few years ago, when most people heard "AI", they pictured a simple chatbot: you ask a question, you get an answer, end of story. That world is gone.&lt;/p&gt;

&lt;p&gt;Today's &lt;strong&gt;AI agents&lt;/strong&gt; can think for themselves, talk to other software systems, read and write files, call APIs, and even hand off tasks to other agents. They are starting to behave less like a search box and more like a junior employee who shows up every day, takes assignments, and tries to get things done.&lt;/p&gt;

&lt;p&gt;This is exciting. But it raises a serious question: &lt;strong&gt;how do we know these "employees" are doing a good job?&lt;/strong&gt; What happens if an agent makes a mistake — especially in a setting like a hospital where mistakes can affect a patient's life?&lt;/p&gt;

&lt;p&gt;This is exactly the gap that &lt;strong&gt;AgentOps&lt;/strong&gt; fills. This post is built on notes I took while watching an IBM Technology video on the topic. My goal is to write something that even someone who has never heard the term "AI agent" can read and walk away understanding.&lt;/p&gt;

&lt;h2&gt;
  
  
  So, What Is AgentOps?
&lt;/h2&gt;

&lt;p&gt;Here is the simplest way to define it:&lt;/p&gt;

&lt;blockquote&gt;
&lt;p&gt;&lt;strong&gt;AgentOps&lt;/strong&gt; is the discipline of &lt;strong&gt;managing&lt;/strong&gt;, &lt;strong&gt;improving&lt;/strong&gt;, and &lt;strong&gt;monitoring&lt;/strong&gt; AI agents in production.&lt;/p&gt;
&lt;/blockquote&gt;

&lt;p&gt;If you have a software background, you may know "DevOps". DevOps is about continuously running, watching, and improving software systems. AgentOps is the same idea, but for AI agents. The difference is that the thing you are monitoring is not a static piece of code — it's a non-deterministic, decision-making "AI worker".&lt;/p&gt;

&lt;p&gt;AgentOps is built on three layers:&lt;/p&gt;

&lt;ol&gt;
&lt;li&gt;Observability&lt;/li&gt;
&lt;li&gt;Evaluation&lt;/li&gt;
&lt;li&gt;Optimization&lt;/li&gt;
&lt;/ol&gt;

&lt;p&gt;We will walk through each layer using one running example so the concepts don't feel abstract.&lt;/p&gt;

&lt;h2&gt;
  
  
  Our Running Example: Two AI Agents in a Hospital
&lt;/h2&gt;

&lt;p&gt;Imagine a hospital. A doctor prescribes a new medication to a patient. Before the patient can pick it up, the insurance company needs to approve it. Traditionally this is a long, painful process involving phone calls, faxes, paperwork, and lots of human waiting.&lt;/p&gt;

&lt;p&gt;Now imagine we automate it with two AI agents:&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;1. Clinical Documentation Agent&lt;/strong&gt;&lt;/p&gt;

&lt;p&gt;This agent connects to the hospital's Electronic Health Record (EHR) system. It pulls the doctor's notes, lab results, the patient's medical history, and any prior treatments. Then it bundles all the relevant information into a clean package that an insurance company would expect.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;2. Payer Authorization Agent&lt;/strong&gt;&lt;/p&gt;

&lt;p&gt;This second agent takes that package, logs into the insurance company's portal, fills out the authorization form, submits it, and waits for approval. Once it gets a green light, it notifies the pharmacy and the doctor.&lt;/p&gt;

&lt;p&gt;These two agents talk to each other (we call this &lt;strong&gt;A2A&lt;/strong&gt;, short for &lt;em&gt;agent-to-agent&lt;/em&gt;) and they each call out to external systems like the EHR, the insurance portal, and the pharmacy.&lt;/p&gt;

&lt;p&gt;Sounds great. But how reliable is this system? How fast? How expensive? What if it makes a mistake one day? That is why we need AgentOps.&lt;/p&gt;

&lt;h2&gt;
  
  
  Layer 1: Observability
&lt;/h2&gt;

&lt;p&gt;Observability answers the question: &lt;strong&gt;"What is my agent doing right now, and how long is it taking?"&lt;/strong&gt; Instead of treating the agent like a mysterious black box, you turn it into a glass box.&lt;/p&gt;

&lt;p&gt;In our hospital example, there are four key signals we want to watch.&lt;/p&gt;

&lt;h3&gt;
  
  
  End-to-End (E2E) Trace Duration
&lt;/h3&gt;

&lt;p&gt;This is the total time from the moment the user makes a request to the moment they get an answer back. In our case, from the second the doctor says "get insurance approval for this patient" to the second that approval is confirmed.&lt;/p&gt;

&lt;p&gt;If that takes 10 seconds, fantastic. If it takes 4 hours, something is wrong somewhere in the pipeline and you need to find it.&lt;/p&gt;

&lt;h3&gt;
  
  
  Agent-to-Agent (A2A) Handoff Latency
&lt;/h3&gt;

&lt;p&gt;In our example, the Clinical Documentation Agent finishes its work and hands the task off to the Payer Authorization Agent. &lt;strong&gt;How long does that handoff take?&lt;/strong&gt;&lt;/p&gt;

&lt;p&gt;This matters more than people realize. Sometimes the time spent passing tasks between agents is longer than the time spent doing the actual work. A clean handoff protocol can save you minutes per request.&lt;/p&gt;

&lt;h3&gt;
  
  
  Tool Execution Latency
&lt;/h3&gt;

&lt;p&gt;Agents rarely work alone. They use &lt;strong&gt;tools&lt;/strong&gt; — calling the EHR system is a tool, opening the insurance portal is a tool, sending a message to the pharmacy is a tool.&lt;/p&gt;

&lt;p&gt;How long does each tool take to respond? Maybe the insurance portal is slow and bottlenecking everything. Tool execution latency lets you spot exactly where the slowdown is, instead of blaming "the AI" in general.&lt;/p&gt;

&lt;h3&gt;
  
  
  Cost per Authorization
&lt;/h3&gt;

&lt;p&gt;How much does a single insurance approval actually cost us? AI agents are not free. Every model call burns tokens, every tool call costs money. If a single approval costs $50 to run while a human staff member could do the same job for $10, the math doesn't work.&lt;/p&gt;

&lt;p&gt;This is the metric that tells you whether your AI investment is actually paying off — or quietly bleeding the budget.&lt;/p&gt;

&lt;h2&gt;
  
  
  Layer 2: Evaluation
&lt;/h2&gt;

&lt;p&gt;Observability tells you &lt;strong&gt;what is happening&lt;/strong&gt;. Evaluation tells you whether &lt;strong&gt;what's happening is actually good&lt;/strong&gt;. An agent that responds in two seconds but gives wrong answers is worse than useless — it's dangerous.&lt;/p&gt;

&lt;h3&gt;
  
  
  Task Completion Rate
&lt;/h3&gt;

&lt;p&gt;Out of every 100 requests, how many does my agent finish successfully &lt;strong&gt;without a human having to step in&lt;/strong&gt;? If that number is 95%, great. If it's 40%, your humans are still doing more than half the work and the automation barely exists.&lt;/p&gt;

&lt;h3&gt;
  
  
  Factual Accuracy
&lt;/h3&gt;

&lt;p&gt;Are the things your agent says &lt;strong&gt;actually true&lt;/strong&gt;? In healthcare this is not optional — it's life or death.&lt;/p&gt;

&lt;p&gt;If the Clinical Documentation Agent records "no penicillin allergy" when the patient actually has one, the consequences can be catastrophic. Factual accuracy measures whether the agent's outputs match the real underlying data.&lt;/p&gt;

&lt;h3&gt;
  
  
  Guardrail Violations
&lt;/h3&gt;

&lt;p&gt;Think of "guardrails" like the railings on a highway — boundaries the agent should never cross. For a hospital, those might be:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;Never leak patient information to anyone unauthorized&lt;/li&gt;
&lt;li&gt;Always comply with HIPAA and similar privacy laws&lt;/li&gt;
&lt;li&gt;Never make decisions outside its authority&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;The guardrail violation rate measures how often the agent crosses one of those lines. The closer this number is to zero, the safer your system.&lt;/p&gt;

&lt;h3&gt;
  
  
  Clinical Appropriateness
&lt;/h3&gt;

&lt;p&gt;This is a healthcare-specific check. Are the agent's decisions &lt;strong&gt;medically reasonable&lt;/strong&gt;? For example, approving an adult dosage for a pediatric patient is technically a "decision" — but it's not clinically appropriate. This is usually scored using rules designed by actual clinicians.&lt;/p&gt;

&lt;h3&gt;
  
  
  First-pass Approval Rate
&lt;/h3&gt;

&lt;p&gt;Of all the insurance authorizations the agent submits, what percentage get approved on the &lt;strong&gt;first try&lt;/strong&gt;? This single metric secretly measures two things at once:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;How well the agent prepares documentation&lt;/li&gt;
&lt;li&gt;How well the agent understands the insurance company's rules&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;If your first-pass approval rate is 85%, the other 15% have to be redone, which means lost time and lost money.&lt;/p&gt;

&lt;h2&gt;
  
  
  Layer 3: Optimization
&lt;/h2&gt;

&lt;p&gt;The first two layers tell you what is happening and whether it's good. The third layer answers: &lt;strong&gt;"How do we make it better?"&lt;/strong&gt;&lt;/p&gt;

&lt;h3&gt;
  
  
  Prompt Token Efficiency
&lt;/h3&gt;

&lt;p&gt;Language models are billed by "tokens". Every token is money. This metric asks: &lt;strong&gt;"How much output quality am I getting for each token I spend?"&lt;/strong&gt;&lt;/p&gt;

&lt;p&gt;If you're shoving a 5000-token mega-prompt into the model and only getting a one-line answer back, you are wasting money. If you can get the same quality with 500 tokens, you just made the system 10x cheaper. Multiplied over thousands of requests a day, this is a huge deal.&lt;/p&gt;

&lt;h3&gt;
  
  
  Flow Step Efficiency
&lt;/h3&gt;

&lt;p&gt;How many steps does the agent take to complete a task? Sometimes agents loop, double-check things they already know, or ask the same question twice. For example, an agent might query the patient's name five times when it could have stored it once and reused it.&lt;/p&gt;

&lt;p&gt;Cutting unnecessary steps makes the agent faster and cheaper at the same time.&lt;/p&gt;

&lt;h3&gt;
  
  
  Retrieval Precision (at K)
&lt;/h3&gt;

&lt;p&gt;This one is a little technical, but it's important. Most agents don't know everything — they pull information from a knowledge base on demand. This is called &lt;strong&gt;RAG&lt;/strong&gt;, short for &lt;em&gt;Retrieval Augmented Generation&lt;/em&gt;. For example, an agent might pull the patient's previous lab reports to figure out their condition.&lt;/p&gt;

&lt;p&gt;"Retrieval Precision at K" measures: out of the K documents the agent pulled, &lt;strong&gt;how many were actually relevant?&lt;/strong&gt; If the agent grabs 10 documents but only 2 are useful, the other 8 are &lt;strong&gt;noise&lt;/strong&gt;. Noise slows the agent down, confuses the model, and can lead to wrong decisions.&lt;/p&gt;

&lt;h3&gt;
  
  
  Handoff Success Rate
&lt;/h3&gt;

&lt;p&gt;How often do handoffs between agents succeed cleanly? The Clinical Documentation Agent prepares a file, but does the Payer Authorization Agent actually receive it correctly? Or does it get a half-broken version? Failed handoffs are silent killers — the system looks like it's running but the wrong things are flowing through it.&lt;/p&gt;

&lt;h3&gt;
  
  
  Improvement Velocity
&lt;/h3&gt;

&lt;p&gt;How fast is your agent actually getting better over time? This is a meta-metric. Are you continuously testing, measuring, tweaking, and re-deploying? Or is the agent today exactly as good as it was on day one? The real power of AgentOps is creating a tight feedback loop where the system improves itself week after week.&lt;/p&gt;

&lt;h2&gt;
  
  
  Back to the Hospital: What the Numbers Could Look Like
&lt;/h2&gt;

&lt;p&gt;Imagine we apply all three layers in our hospital. After a few months of running this system with proper AgentOps, we might see results like:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;Time to get an insurance approval reduced by &lt;strong&gt;85%&lt;/strong&gt; (from days to hours)&lt;/li&gt;
&lt;li&gt;Cases requiring human intervention dropped by &lt;strong&gt;50%&lt;/strong&gt;
&lt;/li&gt;
&lt;li&gt;Cost per authorization down to &lt;strong&gt;$0.47&lt;/strong&gt;
&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;These are not just cute numbers on a dashboard. They translate into:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;Patients getting their medications faster&lt;/li&gt;
&lt;li&gt;Nurses and doctors spending less time on paperwork and more time with patients&lt;/li&gt;
&lt;li&gt;The hospital running more efficiently as a business&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;That's the real promise of agents in healthcare — and AgentOps is the discipline that makes it possible.&lt;/p&gt;

&lt;h2&gt;
  
  
  Why You Cannot Skip AgentOps
&lt;/h2&gt;

&lt;p&gt;Putting an AI agent into a hospital, a bank, or any other critical environment is like handing the car keys to a teenager. Can they drive? Maybe. Would you take your eyes off them? Absolutely not.&lt;/p&gt;

&lt;p&gt;AgentOps is the "eyes" of your AI system. It watches, it grades, it improves. Running an AI agent without AgentOps is like:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;Driving a car with no speedometer&lt;/li&gt;
&lt;li&gt;Running a company with no financial reports&lt;/li&gt;
&lt;li&gt;Treating a patient without a thermometer&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;It's technically possible, but it is asking for trouble.&lt;/p&gt;

&lt;h2&gt;
  
  
  Final Thoughts
&lt;/h2&gt;

&lt;p&gt;The thing that hit me the hardest watching the IBM video was this: AI agents are no longer "experiments". They are becoming &lt;strong&gt;real operational systems&lt;/strong&gt;. And real systems need real metrics.&lt;/p&gt;

&lt;p&gt;To recap the three layers in plain English:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;
&lt;strong&gt;Observability:&lt;/strong&gt; What is happening, how long is it taking, and how much does it cost?&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;Evaluation:&lt;/strong&gt; Is what's happening correct, safe, and useful?&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;Optimization:&lt;/strong&gt; How do we make it better, cheaper, and faster?&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;If you ever deploy your own AI agent into something that matters, do not skip these three layers. Building a flashy demo is easy. Building an agent that runs reliably every hour of every day is only possible with a solid AgentOps practice behind it.&lt;/p&gt;

&lt;p&gt;I'm trying to learn one new AI concept per day by watching IBM Technology videos. If you're doing something similar, I highly recommend turning what you watch into written notes like this one. Writing forces you to actually understand the material — much more than just hitting play and nodding along.&lt;/p&gt;




&lt;p&gt;&lt;em&gt;This post is based on notes I took from an IBM Technology video on AgentOps. If you want to go deeper, I recommend watching the original.&lt;/em&gt;&lt;/p&gt;

</description>
      <category>ai</category>
      <category>agents</category>
      <category>webdev</category>
      <category>devops</category>
    </item>
    <item>
      <title>AWS Basics But Needs to Be Known Before You Start Your Certification</title>
      <dc:creator>Ege Pakten</dc:creator>
      <pubDate>Fri, 27 Mar 2026 16:30:23 +0000</pubDate>
      <link>https://dev.to/egepakten/aws-basics-but-needs-to-be-known-before-you-start-your-certification-3201</link>
      <guid>https://dev.to/egepakten/aws-basics-but-needs-to-be-known-before-you-start-your-certification-3201</guid>
      <description>&lt;blockquote&gt;
&lt;p&gt;You don't need to memorize 200 services to start your AWS journey. But you &lt;strong&gt;do&lt;/strong&gt; need to understand the foundations. Here's everything I wish someone had explained to me before I started studying for my AWS certification.&lt;/p&gt;
&lt;/blockquote&gt;




&lt;h2&gt;
  
  
  1. What Even Is Cloud Computing?
&lt;/h2&gt;

&lt;p&gt;Let's start from the very beginning. Cloud computing sounds fancy, but the core idea is simple: &lt;strong&gt;there's a physical server somewhere, and you're renting it over the internet.&lt;/strong&gt;&lt;/p&gt;

&lt;p&gt;Instead of buying your own hardware, setting up a server room, and hiring people to maintain it — you use someone else's infrastructure (like AWS) and pay only for what you use.&lt;/p&gt;

&lt;p&gt;Here's the official NIST definition:&lt;/p&gt;

&lt;blockquote&gt;
&lt;p&gt;&lt;em&gt;"A model for enabling ubiquitous, convenient, on-demand network access to a shared pool of **configurable computing resources&lt;/em&gt;* that can be rapidly provisioned and released with minimal management effort or service provider interaction."*&lt;/p&gt;
&lt;/blockquote&gt;

&lt;p&gt;Let's break that down into human language:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;
&lt;strong&gt;On-demand&lt;/strong&gt; → You get resources whenever you want them&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;Shared pool&lt;/strong&gt; → Multiple customers share the same physical infrastructure&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;Configurable&lt;/strong&gt; → You choose how much CPU, RAM, storage you need&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;Rapidly provisioned and released&lt;/strong&gt; → Spin up a server in minutes, shut it down when you're done&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;Minimal management effort&lt;/strong&gt; → No need to physically touch any hardware&lt;/li&gt;
&lt;/ul&gt;




&lt;h2&gt;
  
  
  2. What's Actually Behind the "Cloud"?
&lt;/h2&gt;

&lt;p&gt;Behind the cloud, there are real, physical server racks sitting in massive data centers around the world. AWS has these data centers spread across the globe, and the infrastructure is organized in a clear hierarchy:&lt;/p&gt;

&lt;h3&gt;
  
  
  Region
&lt;/h3&gt;

&lt;p&gt;A large geographic area (e.g., &lt;code&gt;eu-west-1&lt;/code&gt; for Ireland, &lt;code&gt;us-east-1&lt;/code&gt; for N. Virginia). You choose a region based on where your users are, compliance requirements, or pricing.&lt;/p&gt;

&lt;h3&gt;
  
  
  Availability Zone (AZ)
&lt;/h3&gt;

&lt;p&gt;Each Region contains multiple AZs (e.g., &lt;code&gt;eu-west-1a&lt;/code&gt;, &lt;code&gt;eu-west-1b&lt;/code&gt;, &lt;code&gt;eu-west-1c&lt;/code&gt;). An AZ is one or more data centers with independent power, networking, and connectivity. Multiple AZs exist for &lt;strong&gt;redundancy&lt;/strong&gt; — if one goes down, the others keep running.&lt;/p&gt;

&lt;h3&gt;
  
  
  Edge Locations / Points of Presence (PoP)
&lt;/h3&gt;

&lt;p&gt;These are smaller, lightweight locations spread even more widely than Regions. They're used primarily for &lt;strong&gt;caching content closer to end users&lt;/strong&gt;.&lt;/p&gt;




&lt;h2&gt;
  
  
  3. Edge Networks and CDN — Speed Matters
&lt;/h2&gt;

&lt;p&gt;Imagine your origin server is in the US, but a user in Southeast Asia wants to load your website. Without any optimization, every single request travels across the Pacific Ocean and back. That's slow.&lt;/p&gt;

&lt;p&gt;This is where &lt;strong&gt;Edge Networks&lt;/strong&gt; and &lt;strong&gt;CDN (Content Delivery Network)&lt;/strong&gt; come in.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;How it works:&lt;/strong&gt;&lt;/p&gt;

&lt;ol&gt;
&lt;li&gt;Your original content lives on the &lt;strong&gt;Origin Server&lt;/strong&gt; (e.g., in &lt;code&gt;us-east-1&lt;/code&gt;)&lt;/li&gt;
&lt;li&gt;AWS caches copies of your static content (images, CSS, JS, videos) at &lt;strong&gt;PoP / Edge Locations&lt;/strong&gt; around the world&lt;/li&gt;
&lt;li&gt;When a user in Singapore requests your site, they get the cached version from the nearest PoP — not from the US&lt;/li&gt;
&lt;li&gt;Result: &lt;strong&gt;dramatically lower latency&lt;/strong&gt;
&lt;/li&gt;
&lt;/ol&gt;

&lt;p&gt;&lt;strong&gt;CDN vs Edge Network — are they the same thing?&lt;/strong&gt;&lt;/p&gt;

&lt;p&gt;Not exactly. The &lt;strong&gt;Edge Network&lt;/strong&gt; is the infrastructure — the distributed network of servers worldwide. &lt;strong&gt;CDN&lt;/strong&gt; is the most well-known &lt;em&gt;use case&lt;/em&gt; of that infrastructure. In practice, people use the terms interchangeably, and that's mostly fine.&lt;/p&gt;

&lt;p&gt;AWS's CDN service is called &lt;strong&gt;CloudFront&lt;/strong&gt;, and it leverages these Edge Locations.&lt;/p&gt;




&lt;h2&gt;
  
  
  4. Virtualization and the Hypervisor
&lt;/h2&gt;

&lt;p&gt;Here's where it gets interesting. When you launch an EC2 instance, you're &lt;strong&gt;not&lt;/strong&gt; getting a dedicated physical server. Here's what's actually happening:&lt;/p&gt;

&lt;ol&gt;
&lt;li&gt;
&lt;strong&gt;Host Computer&lt;/strong&gt; → The physical AWS server rack in a data center&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;Host OS&lt;/strong&gt; → The operating system running on that physical machine&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;Hypervisor&lt;/strong&gt; → Software that sits on top of the Host OS and splits the physical machine into multiple &lt;strong&gt;virtual machines&lt;/strong&gt;
&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;Your EC2 Instance&lt;/strong&gt; → One of those virtual machines&lt;/li&gt;
&lt;/ol&gt;

&lt;p&gt;So if a physical server has 64 CPUs and 256 GB RAM, the hypervisor might split it into four isolated instances of 16 CPUs / 64 GB RAM each. Different customers can be using the same physical hardware without ever knowing about each other — completely isolated.&lt;/p&gt;

&lt;p&gt;AWS built their own hypervisor called &lt;strong&gt;Nitro&lt;/strong&gt;, which operates at the hardware level for minimal performance overhead.&lt;/p&gt;

&lt;blockquote&gt;
&lt;p&gt;&lt;strong&gt;Key takeaway:&lt;/strong&gt; EC2 instances are virtual slices of physical servers, managed by a hypervisor. Multiple instances from different customers can live on the same physical machine, fully isolated from each other.&lt;/p&gt;
&lt;/blockquote&gt;




&lt;h2&gt;
  
  
  5. The Shared Responsibility Model — Who's Responsible for What?
&lt;/h2&gt;

&lt;p&gt;This is probably the &lt;strong&gt;most important concept&lt;/strong&gt; for your certification exam. AWS and the customer share security responsibilities, but the line between them is very clear.&lt;/p&gt;

&lt;h3&gt;
  
  
  Security &lt;strong&gt;OF&lt;/strong&gt; the Cloud → AWS's Job
&lt;/h3&gt;

&lt;p&gt;AWS is responsible for protecting the infrastructure that runs all the services:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;
&lt;strong&gt;Physical security&lt;/strong&gt; — data center access, cameras, biometric entry&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;Hardware&lt;/strong&gt; — servers, storage, networking equipment&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;Software&lt;/strong&gt; — compute, storage, database, networking services&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;Global infrastructure&lt;/strong&gt; — Regions, Availability Zones, Edge Locations&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;Host OS &amp;amp; Hypervisor&lt;/strong&gt; — you can't even access these&lt;/li&gt;
&lt;/ul&gt;

&lt;h3&gt;
  
  
  Security &lt;strong&gt;IN&lt;/strong&gt; the Cloud → Your Job
&lt;/h3&gt;

&lt;p&gt;You are responsible for everything you put in and configure on the cloud:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;
&lt;strong&gt;Customer data&lt;/strong&gt; — whatever you upload, store, or process&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;Platform, applications &amp;amp; IAM&lt;/strong&gt; — who has access to what, roles, permissions, secret keys&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;OS &amp;amp; firewall configuration&lt;/strong&gt; — if you launched an EC2 with Ubuntu, patching it is on you&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;Encryption&lt;/strong&gt; — client-side, server-side, in-transit, at-rest decisions&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;Network traffic protection&lt;/strong&gt; — security groups, NACLs, VPN configurations&lt;/li&gt;
&lt;/ul&gt;

&lt;h3&gt;
  
  
  The Three Control Categories
&lt;/h3&gt;

&lt;div class="table-wrapper-paragraph"&gt;&lt;table&gt;
&lt;thead&gt;
&lt;tr&gt;
&lt;th&gt;Category&lt;/th&gt;
&lt;th&gt;Who?&lt;/th&gt;
&lt;th&gt;Examples&lt;/th&gt;
&lt;/tr&gt;
&lt;/thead&gt;
&lt;tbody&gt;
&lt;tr&gt;
&lt;td&gt;&lt;strong&gt;Inherited&lt;/strong&gt;&lt;/td&gt;
&lt;td&gt;AWS only&lt;/td&gt;
&lt;td&gt;Physical &amp;amp; environmental security, host OS, physical servers&lt;/td&gt;
&lt;/tr&gt;
&lt;tr&gt;
&lt;td&gt;&lt;strong&gt;Shared&lt;/strong&gt;&lt;/td&gt;
&lt;td&gt;Both&lt;/td&gt;
&lt;td&gt;Patch management, configuration management, awareness &amp;amp; training&lt;/td&gt;
&lt;/tr&gt;
&lt;tr&gt;
&lt;td&gt;&lt;strong&gt;Customer-Specific&lt;/strong&gt;&lt;/td&gt;
&lt;td&gt;Customer only&lt;/td&gt;
&lt;td&gt;Guest OS, custom applications, data encryption strategies&lt;/td&gt;
&lt;/tr&gt;
&lt;/tbody&gt;
&lt;/table&gt;&lt;/div&gt;

&lt;h3&gt;
  
  
  Quick Quiz — Test Yourself
&lt;/h3&gt;

&lt;div class="table-wrapper-paragraph"&gt;&lt;table&gt;
&lt;thead&gt;
&lt;tr&gt;
&lt;th&gt;Question&lt;/th&gt;
&lt;th&gt;Answer&lt;/th&gt;
&lt;/tr&gt;
&lt;/thead&gt;
&lt;tbody&gt;
&lt;tr&gt;
&lt;td&gt;Patching your EC2 instance's operating system?&lt;/td&gt;
&lt;td&gt;&lt;strong&gt;You&lt;/strong&gt;&lt;/td&gt;
&lt;/tr&gt;
&lt;tr&gt;
&lt;td&gt;Guest OS security patches?&lt;/td&gt;
&lt;td&gt;&lt;strong&gt;You&lt;/strong&gt;&lt;/td&gt;
&lt;/tr&gt;
&lt;tr&gt;
&lt;td&gt;Running the host OS and virtualization layer?&lt;/td&gt;
&lt;td&gt;&lt;strong&gt;AWS&lt;/strong&gt;&lt;/td&gt;
&lt;/tr&gt;
&lt;tr&gt;
&lt;td&gt;Managing IAM user access and secret keys?&lt;/td&gt;
&lt;td&gt;&lt;strong&gt;You&lt;/strong&gt;&lt;/td&gt;
&lt;/tr&gt;
&lt;tr&gt;
&lt;td&gt;Maintaining the server under your Lambda functions?&lt;/td&gt;
&lt;td&gt;&lt;strong&gt;AWS&lt;/strong&gt;&lt;/td&gt;
&lt;/tr&gt;
&lt;tr&gt;
&lt;td&gt;Physical security of data centers?&lt;/td&gt;
&lt;td&gt;&lt;strong&gt;AWS&lt;/strong&gt;&lt;/td&gt;
&lt;/tr&gt;
&lt;tr&gt;
&lt;td&gt;Encryption-at-rest strategy for your RDS database?&lt;/td&gt;
&lt;td&gt;&lt;strong&gt;You&lt;/strong&gt;&lt;/td&gt;
&lt;/tr&gt;
&lt;/tbody&gt;
&lt;/table&gt;&lt;/div&gt;

&lt;blockquote&gt;
&lt;p&gt;&lt;strong&gt;Simple rule of thumb:&lt;/strong&gt; If you can configure it in the AWS Console or CLI → it's your responsibility. If you can't even touch it → it's AWS's responsibility. If both sides need to do their part → it's shared.&lt;/p&gt;
&lt;/blockquote&gt;

&lt;p&gt;Think of it like renting an apartment: AWS gives you a secure building (locked doors, fire alarms, security guards). But whether you lock your own door, put your valuables in a safe, or leave your windows open — that's entirely on you.&lt;/p&gt;




&lt;h2&gt;
  
  
  6. AWS Support Plans — Developer vs Business
&lt;/h2&gt;

&lt;p&gt;When you create an AWS account, you'll need to choose a support plan. Here's the practical difference between the two most common ones:&lt;/p&gt;

&lt;div class="table-wrapper-paragraph"&gt;&lt;table&gt;
&lt;thead&gt;
&lt;tr&gt;
&lt;th&gt;Feature&lt;/th&gt;
&lt;th&gt;Developer (~$29/mo)&lt;/th&gt;
&lt;th&gt;Business (~$100/mo or % of usage)&lt;/th&gt;
&lt;/tr&gt;
&lt;/thead&gt;
&lt;tbody&gt;
&lt;tr&gt;
&lt;td&gt;Who can open tickets&lt;/td&gt;
&lt;td&gt;1 person (primary contact)&lt;/td&gt;
&lt;td&gt;Unlimited team members&lt;/td&gt;
&lt;/tr&gt;
&lt;tr&gt;
&lt;td&gt;Response time (critical)&lt;/td&gt;
&lt;td&gt;12 hours&lt;/td&gt;
&lt;td&gt;&lt;strong&gt;1 hour&lt;/strong&gt;&lt;/td&gt;
&lt;/tr&gt;
&lt;tr&gt;
&lt;td&gt;Support channels&lt;/td&gt;
&lt;td&gt;Email only&lt;/td&gt;
&lt;td&gt;Email + &lt;strong&gt;Phone + Chat&lt;/strong&gt;
&lt;/td&gt;
&lt;/tr&gt;
&lt;tr&gt;
&lt;td&gt;Trusted Advisor&lt;/td&gt;
&lt;td&gt;Limited checks&lt;/td&gt;
&lt;td&gt;&lt;strong&gt;Full access&lt;/strong&gt;&lt;/td&gt;
&lt;/tr&gt;
&lt;tr&gt;
&lt;td&gt;Third-party software support&lt;/td&gt;
&lt;td&gt;No&lt;/td&gt;
&lt;td&gt;Yes&lt;/td&gt;
&lt;/tr&gt;
&lt;/tbody&gt;
&lt;/table&gt;&lt;/div&gt;

&lt;h3&gt;
  
  
  Which one should you pick?
&lt;/h3&gt;

&lt;ul&gt;
&lt;li&gt;
&lt;strong&gt;Developer&lt;/strong&gt; → Great for learning, experimenting, and building prototypes. Start here if you're just getting started.&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;Business&lt;/strong&gt; → Essential when you're running production workloads with real users. The 1-hour critical response time alone is worth it when something goes down.&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;&lt;strong&gt;Good news:&lt;/strong&gt; You can start with Developer and &lt;strong&gt;upgrade to Business anytime&lt;/strong&gt; from the AWS Console (&lt;code&gt;Support Center &amp;gt; Change Plan&lt;/code&gt;). Billing is prorated, so you only pay for what you use. You can also downgrade later if needed.&lt;/p&gt;




&lt;h2&gt;
  
  
  Wrapping Up
&lt;/h2&gt;

&lt;p&gt;Before diving into specific services like S3, Lambda, or DynamoDB, make sure these foundational concepts are solid:&lt;/p&gt;

&lt;ol&gt;
&lt;li&gt;
&lt;strong&gt;Cloud = renting virtual resources&lt;/strong&gt; from physical infrastructure over the internet&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;Regions → AZs → Edge Locations&lt;/strong&gt; form the backbone of AWS's global presence&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;CDN / Edge Networks&lt;/strong&gt; cache content close to users for speed&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;Hypervisors&lt;/strong&gt; split physical servers into isolated virtual machines (EC2 instances)&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;Shared Responsibility Model&lt;/strong&gt; — know what's yours vs what's AWS's (this &lt;em&gt;will&lt;/em&gt; be on the exam)&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;Support Plans&lt;/strong&gt; can be upgraded/downgraded as your needs evolve&lt;/li&gt;
&lt;/ol&gt;

&lt;p&gt;These aren't glamorous topics, but they're the bedrock everything else is built on. Nail these, and the rest of your certification journey will make a lot more sense.&lt;/p&gt;




&lt;p&gt;&lt;em&gt;Happy cloud learning! If you found this helpful, drop a reaction or follow for more AWS certification notes.&lt;/em&gt;&lt;/p&gt;

</description>
      <category>aws</category>
      <category>infrastructure</category>
      <category>cloud</category>
      <category>programming</category>
    </item>
    <item>
      <title>What I Learned from "The Mom Test" - Chapter 8: Running the Process</title>
      <dc:creator>Ege Pakten</dc:creator>
      <pubDate>Wed, 11 Mar 2026 09:59:08 +0000</pubDate>
      <link>https://dev.to/egepakten/what-i-learned-from-the-mom-test-chapter-8-running-the-process-242h</link>
      <guid>https://dev.to/egepakten/what-i-learned-from-the-mom-test-chapter-8-running-the-process-242h</guid>
      <description>&lt;p&gt;&lt;em&gt;A developer's guide to turning customer conversations from a solo guessing game into a team-wide learning machine&lt;/em&gt;&lt;/p&gt;




&lt;p&gt;You've learned the skills: asking good questions, avoiding compliments, pushing for commitments, finding conversations, and choosing your customers. But skills alone aren't enough. &lt;strong&gt;If you don't have a process to capture and share what you learn, it all falls apart.&lt;/strong&gt;&lt;/p&gt;

&lt;p&gt;Chapter 8 is the final chapter of The Mom Test, and it ties everything together. It's about the nuts and bolts of actually running customer conversations as a repeatable, team-wide process — not just something the "business person" does while everyone else codes.&lt;/p&gt;




&lt;h2&gt;
  
  
  Outline
&lt;/h2&gt;

&lt;p&gt;In this post, I'll break down Chapter 8 into the following sections:&lt;/p&gt;

&lt;ol&gt;
&lt;li&gt;
&lt;strong&gt;Prepping for the Conversation&lt;/strong&gt; — Why going in with a clear list of your riskiest assumptions saves you from aimless chats.&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;Who Should Show Up&lt;/strong&gt; — Why the whole founding team needs to hear customer conversations firsthand, and what goes wrong when they don't.&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;How to Write It Down&lt;/strong&gt; — The note-taking system that actually works: exact quotes, emotions, and constraints — not just "they liked it."&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;Reviewing with Your Team&lt;/strong&gt; — How to turn a pile of messy notes into clear decisions, and why reviewing together prevents the "telephone game" problem.&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;Talking to Customers Is a Team Sport&lt;/strong&gt; — Why delegating all customer conversations to one person is a recipe for disaster.&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;The Process&lt;/strong&gt; — Fitzpatrick's lightweight, practical process for weaving customer learning into your weekly rhythm.&lt;/li&gt;
&lt;/ol&gt;




&lt;h2&gt;
  
  
  Prepping for the Conversation
&lt;/h2&gt;

&lt;p&gt;Before you walk into a customer conversation, you need to know what you're trying to learn. This sounds obvious, but most people skip it. They just show up and "see what happens."&lt;/p&gt;

&lt;p&gt;Fitzpatrick recommends sitting down with your team before any batch of conversations and asking: &lt;strong&gt;"What are the three biggest questions we need to answer right now?"&lt;/strong&gt;&lt;/p&gt;

&lt;p&gt;These should be your scariest, riskiest assumptions — the ones that could sink the whole business if you're wrong. Maybe it's "Do people actually switch tools for this?" or "Will managers pay for something their team uses?" or "Is this problem painful enough to justify the effort?"&lt;/p&gt;

&lt;p&gt;If you go in without a clear list, you'll default to comfortable topics. You'll talk about features you're excited about instead of risks you're afraid of. And you'll come out feeling good but learning nothing.&lt;/p&gt;

&lt;p&gt;You don't need a rigid script — that kills the natural flow of conversation. But you need a cheat sheet of the big questions. Glance at it before the meeting. Make sure you hit the important stuff before the conversation ends.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;Rule of Thumb:&lt;/strong&gt; If you don't know what you're trying to learn, you're not ready for the conversation. Prep your three big questions before every batch of meetings.&lt;/p&gt;




&lt;h2&gt;
  
  
  Who Should Show Up
&lt;/h2&gt;

&lt;p&gt;This one surprised me. Fitzpatrick is adamant: &lt;strong&gt;the whole founding team should be in customer meetings.&lt;/strong&gt; Not just the CEO. Not just the "business person." Everyone.&lt;/p&gt;

&lt;p&gt;Why? Because customer learning that passes through a middleman always gets distorted. It's like a game of telephone. The person who was in the meeting says "they seemed really excited about feature X" and everyone else nods along. But were they actually excited? Or were they just being polite? Only the person in the room can judge the body language, the tone, the hesitations.&lt;/p&gt;

&lt;p&gt;When the whole team hears the same thing at the same time, two things happen:&lt;/p&gt;

&lt;ol&gt;
&lt;li&gt;
&lt;strong&gt;You avoid arguments about what the customer "really meant."&lt;/strong&gt; Everyone heard it themselves.&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;The technical co-founders hear constraints and problems directly&lt;/strong&gt;, which means they can often come up with better solutions than what the customer (or the business co-founder) would have suggested.&lt;/li&gt;
&lt;/ol&gt;

&lt;p&gt;Obviously, this doesn't scale forever. You can't have five people show up to every coffee chat. But in the early days — when every conversation shapes your product direction — it's worth the investment.&lt;/p&gt;

&lt;p&gt;If you absolutely can't have everyone attend, rotate who goes. But NEVER let one person become the sole keeper of customer knowledge. That's a single point of failure, and when they're wrong about something, the whole team is wrong.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;Rule of Thumb:&lt;/strong&gt; Everyone on the team who is making big product or design decisions should be sitting in on at least some customer conversations. Don't let secondhand summaries replace firsthand experience.&lt;/p&gt;




&lt;h2&gt;
  
  
  How to Write It Down
&lt;/h2&gt;

&lt;p&gt;Taking notes during customer conversations is critical, but most people do it wrong. They write down summaries, interpretations, and vague impressions: "She seemed interested in the analytics dashboard" or "He thought the pricing was fair."&lt;/p&gt;

&lt;p&gt;The problem? Those aren't facts — they're your interpretations. And interpretations are where bias creeps in.&lt;/p&gt;

&lt;p&gt;Fitzpatrick recommends a specific note-taking format. Write down:&lt;/p&gt;

&lt;ol&gt;
&lt;li&gt;&lt;p&gt;&lt;strong&gt;Exact quotes&lt;/strong&gt; — The actual words they used. Not your paraphrase. Not your interpretation. Their exact words. Put them in quotation marks so you can distinguish them from your own notes later.&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;&lt;strong&gt;Emotions&lt;/strong&gt; — Did they light up when talking about a particular problem? Did they seem bored? Angry? Dismissive? Emotions reveal what matters to people in a way that words sometimes don't.&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;&lt;strong&gt;Constraints&lt;/strong&gt; — Hard facts about their situation. "Their team has 12 people." "They spend $2,000/month on the current tool." "They've been looking for a solution for 6 months." These are the concrete details that help you make decisions later.&lt;/p&gt;&lt;/li&gt;
&lt;/ol&gt;

&lt;p&gt;What NOT to write down: your own ideas, feature requests you thought of during the conversation, or your emotional reactions. Those go in a separate section. Keep the raw customer data clean and uncontaminated.&lt;/p&gt;

&lt;p&gt;Fitzpatrick also suggests using shorthand symbols to quickly flag important moments. For instance, use a smiley face or an exclamation mark for strong emotions, a dollar sign for anything related to money or budget, and a star for particularly important quotes.&lt;/p&gt;

&lt;p&gt;After the meeting, &lt;strong&gt;spend five minutes cleaning up your notes while the conversation is still fresh.&lt;/strong&gt; Fill in the gaps, clarify the shorthand, and highlight the most important bits. If you wait until the next day, you'll have forgotten half the nuance.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;Rule of Thumb:&lt;/strong&gt; Write down exact quotes, emotions, and constraints. Keep your interpretations separate from the raw data. Spend five minutes cleaning up notes immediately after every conversation.&lt;/p&gt;




&lt;h2&gt;
  
  
  Reviewing with Your Team
&lt;/h2&gt;

&lt;p&gt;Raw notes from individual conversations aren't very useful on their own. The magic happens when you &lt;strong&gt;review them together as a team.&lt;/strong&gt;&lt;/p&gt;

&lt;p&gt;Fitzpatrick recommends a regular review session — ideally weekly — where the team sits down and goes through the recent conversation notes together. The goal isn't just to share what you heard. It's to look for patterns, update your beliefs, and decide what to do next.&lt;/p&gt;

&lt;p&gt;Here's how a good review works:&lt;/p&gt;

&lt;ol&gt;
&lt;li&gt;&lt;p&gt;&lt;strong&gt;Share the raw notes.&lt;/strong&gt; Read out the exact quotes, the emotions, the constraints. Don't editorialize. Let the data speak for itself first.&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;&lt;strong&gt;Look for patterns.&lt;/strong&gt; Are multiple people saying the same thing? Are you hearing the same pain point from different customers? That's a strong signal. Are you hearing wildly different things? That might mean your segment is too broad (go back to Chapter 7).&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;&lt;strong&gt;Update your assumptions.&lt;/strong&gt; Remember those three big questions you prepped? Based on what you've heard, which ones have been answered? Which ones are still open? What new questions have emerged?&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;&lt;strong&gt;Decide on next steps.&lt;/strong&gt; What conversations do you need to have next? Do you need to talk to more of the same type of customer, or a different segment? Do you have enough signal to start building something?&lt;/p&gt;&lt;/li&gt;
&lt;/ol&gt;

&lt;p&gt;The key is that this review should be a conversation, not a presentation. If one person just summarises what they heard and everyone else nods, you've missed the point. Everyone should engage with the raw data and form their own conclusions.&lt;/p&gt;

&lt;p&gt;This process also catches mistakes. If one team member misinterpreted a conversation, the others can catch it during the review. If someone got excited about a compliment disguised as a commitment (Chapter 5 flashback!), the team can flag it.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;Rule of Thumb:&lt;/strong&gt; Don't let customer learnings sit in one person's notebook. Review notes together as a team regularly, look for patterns, and update your beliefs based on the evidence.&lt;/p&gt;




&lt;h2&gt;
  
  
  Talking to Customers Is a Team Sport
&lt;/h2&gt;

&lt;p&gt;Fitzpatrick keeps hammering this point home because it's so often ignored: &lt;strong&gt;customer learning is not one person's job.&lt;/strong&gt;&lt;/p&gt;

&lt;p&gt;In many startups, there's a division of labour: the "business person" talks to customers, and the "technical person" builds the product. This feels efficient, but it's actually a disaster.&lt;/p&gt;

&lt;p&gt;Here's why: the technical co-founder ends up building based on secondhand information. They hear "customers want feature X" and build it. But if they had been in the room, they might have heard the underlying problem behind the request and come up with a completely different (and better) solution.&lt;/p&gt;

&lt;p&gt;Engineers tend to think of customer conversations as someone else's responsibility — something that takes them away from "real work." But talking to customers IS real work. It's arguably the most important work in the early stages of a startup, because it determines whether you're building the right thing.&lt;/p&gt;

&lt;p&gt;Fitzpatrick even goes so far as to say: if someone on your team refuses to participate in customer conversations, that's a serious red flag. It means they're comfortable building in the dark, which is a recipe for building something nobody wants.&lt;/p&gt;

&lt;p&gt;This doesn't mean every engineer needs to lead conversations. Some people are naturally better at it than others. But everyone should at least be present for some conversations, take turns leading when possible, and participate actively in the team review sessions.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;Rule of Thumb:&lt;/strong&gt; Customer conversations are everyone's responsibility. If only one person on the team talks to customers, you have a bottleneck that will eventually break.&lt;/p&gt;




&lt;h2&gt;
  
  
  The Process
&lt;/h2&gt;

&lt;p&gt;So how does all of this come together in practice? Fitzpatrick outlines a lightweight process that you can adapt to your situation. It's not complicated, but it requires discipline:&lt;/p&gt;

&lt;h3&gt;
  
  
  Before a Batch of Conversations:
&lt;/h3&gt;

&lt;ul&gt;
&lt;li&gt;Sit down as a team and identify your &lt;strong&gt;three biggest learning goals&lt;/strong&gt; (your scariest assumptions)&lt;/li&gt;
&lt;li&gt;Decide &lt;strong&gt;who you're going to talk to&lt;/strong&gt; (your customer segment from Chapter 7)&lt;/li&gt;
&lt;li&gt;Decide &lt;strong&gt;where to find them&lt;/strong&gt; (your "who-where" pair)&lt;/li&gt;
&lt;/ul&gt;

&lt;h3&gt;
  
  
  During Each Conversation:
&lt;/h3&gt;

&lt;ul&gt;
&lt;li&gt;
&lt;strong&gt;One person leads&lt;/strong&gt; the conversation, asking questions and steering the discussion&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;Another person takes notes&lt;/strong&gt; — exact quotes, emotions, constraints&lt;/li&gt;
&lt;li&gt;Keep it casual (Chapter 4) and follow the Mom Test rules (Chapters 1-3)&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;Push for commitment&lt;/strong&gt; at the end (Chapter 5)&lt;/li&gt;
&lt;/ul&gt;

&lt;h3&gt;
  
  
  Immediately After:
&lt;/h3&gt;

&lt;ul&gt;
&lt;li&gt;Spend &lt;strong&gt;five minutes&lt;/strong&gt; reviewing and cleaning up notes together&lt;/li&gt;
&lt;li&gt;Flag anything surprising or particularly important&lt;/li&gt;
&lt;li&gt;Note any follow-up actions or commitments you made&lt;/li&gt;
&lt;/ul&gt;

&lt;h3&gt;
  
  
  Weekly (or After Every Batch):
&lt;/h3&gt;

&lt;ul&gt;
&lt;li&gt;
&lt;strong&gt;Team review session&lt;/strong&gt; — go through notes together&lt;/li&gt;
&lt;li&gt;Look for patterns across conversations&lt;/li&gt;
&lt;li&gt;Update your assumptions and beliefs&lt;/li&gt;
&lt;li&gt;Decide: do you need more conversations, or is it time to build?&lt;/li&gt;
&lt;li&gt;Identify your &lt;strong&gt;next three big questions&lt;/strong&gt; for the next batch&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;The beauty of this process is that it's fast. Each cycle takes about a week. In a single week, you can have 5-10 conversations, review them as a team, and make a clear decision about what to do next. Compare that to spending months building something in isolation and then discovering nobody wants it.&lt;/p&gt;

&lt;p&gt;Fitzpatrick emphasises that the process should be lightweight. If it feels like bureaucracy, you've over-complicated it. The notes should be quick. The reviews should be short. The goal is to learn fast and stay nimble.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;Rule of Thumb:&lt;/strong&gt; Keep the process simple. Prep → Conversations → Notes → Review → Decide → Repeat. If you're spending more time on process than on actual conversations, something's wrong.&lt;/p&gt;




&lt;h2&gt;
  
  
  Key Takeaways from Chapter 8
&lt;/h2&gt;

&lt;ol&gt;
&lt;li&gt;&lt;p&gt;&lt;strong&gt;Prep before every batch of conversations.&lt;/strong&gt; Identify your three scariest assumptions and make sure you address them. Don't just "wing it."&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;&lt;strong&gt;The whole team should hear customer conversations.&lt;/strong&gt; Secondhand summaries always lose nuance. Everyone who makes product decisions needs firsthand exposure.&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;&lt;strong&gt;Take notes the right way.&lt;/strong&gt; Exact quotes, emotions, and constraints — not interpretations. Keep the raw data separate from your opinions.&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;&lt;strong&gt;Review notes as a team.&lt;/strong&gt; Look for patterns, catch misinterpretations, update your beliefs, and decide on next steps together.&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;&lt;strong&gt;Customer learning is everyone's job.&lt;/strong&gt; Don't delegate it to one person. Engineers, designers, and founders all benefit from being in the room.&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;&lt;strong&gt;Keep the process lightweight.&lt;/strong&gt; Prep → Talk → Note → Review → Decide → Repeat. One cycle per week. Don't let process become bureaucracy.&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;&lt;strong&gt;Learning is the goal, not meetings.&lt;/strong&gt; Every conversation should move you closer to a clear picture of what to build and who to build it for.&lt;/p&gt;&lt;/li&gt;
&lt;/ol&gt;




&lt;p&gt;&lt;em&gt;Previously: &lt;a href="https://dev.to/egepakten/what-i-learned-from-the-mom-test-chapter-7-choosing-your-customers-6l6"&gt;Chapter 7 - Choosing Your Customers&lt;/a&gt;&lt;/em&gt;&lt;/p&gt;

&lt;p&gt;&lt;em&gt;That's a wrap on The Mom Test! If you've been following along, I hope these posts have been as useful to you as writing them was for me. Now go talk to your customers — the right way.&lt;/em&gt;&lt;/p&gt;

</description>
      <category>webdev</category>
      <category>startup</category>
      <category>speaking</category>
      <category>learning</category>
    </item>
    <item>
      <title>What I Learned from "The Mom Test" - Chapter 7: Choosing Your Customers</title>
      <dc:creator>Ege Pakten</dc:creator>
      <pubDate>Mon, 09 Mar 2026 10:58:49 +0000</pubDate>
      <link>https://dev.to/egepakten/what-i-learned-from-the-mom-test-chapter-7-choosing-your-customers-6l6</link>
      <guid>https://dev.to/egepakten/what-i-learned-from-the-mom-test-chapter-7-choosing-your-customers-6l6</guid>
      <description>&lt;p&gt;&lt;em&gt;A developer's guide to why talking to "everyone" is the fastest way to learn nothing&lt;/em&gt;&lt;/p&gt;




&lt;p&gt;There's a saying in the startup world: &lt;strong&gt;startups don't starve, they drown.&lt;/strong&gt; You never have too few options, too few leads, or too few ideas — you have too many. You get overwhelmed. You do a little bit of everything and make progress on nothing.&lt;/p&gt;

&lt;p&gt;Chapter 7 is about the antidote: &lt;strong&gt;customer segmentation.&lt;/strong&gt; Choosing who to focus on so you can actually learn something useful instead of drowning in mixed signals.&lt;/p&gt;




&lt;h2&gt;
  
  
  Outline
&lt;/h2&gt;

&lt;p&gt;In this post, I'll break down Chapter 7 into the following sections:&lt;/p&gt;

&lt;ol&gt;
&lt;li&gt;
&lt;strong&gt;Segmentation: Why Starting Broad Kills You&lt;/strong&gt; — How Google, Paypal, and Evernote all started narrow, and why you should too.&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;Babies or Body Builders?&lt;/strong&gt; — A real-world example of what happens when you try to serve everyone with one product.&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;Big Brands or Mom &amp;amp; Pop?&lt;/strong&gt; — Fitzpatrick's own painful lesson of choosing a customer segment that was way too broad.&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;But What Does It Mean?&lt;/strong&gt; — Why "students" or "advertisers" aren't real customer segments, and what happens when you treat them like they are.&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;Customer Slicing&lt;/strong&gt; — A practical step-by-step technique for narrowing down your segment until you know exactly who to talk to and where to find them.&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;Talking to the Wrong People&lt;/strong&gt; — The three traps that lead you to waste time on the wrong conversations.&lt;/li&gt;
&lt;/ol&gt;




&lt;h2&gt;
  
  
  Segmentation: Why Starting Broad Kills You
&lt;/h2&gt;

&lt;p&gt;When we look at the big tech successes, they seem to serve the whole world. Google lets anyone find anything. Paypal helps anyone send money anywhere. Evernote backs up everyone's notes.&lt;/p&gt;

&lt;p&gt;But here's the thing — &lt;strong&gt;they didn't start there.&lt;/strong&gt;&lt;/p&gt;

&lt;p&gt;In its early days, Google helped PhD students find obscure bits of code. Paypal helped collectors buy and sell Pez dispensers and Beanie Babies more efficiently. Evernote helped moms save and share recipes.&lt;/p&gt;

&lt;p&gt;Every one of them started with a tiny, specific group. They nailed it for that group, then expanded. If you try to start with "everyone" as your customer, three things go wrong:&lt;/p&gt;

&lt;ol&gt;
&lt;li&gt;
&lt;strong&gt;You get overwhelmed by options&lt;/strong&gt; and don't know where to start&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;You aren't moving forward&lt;/strong&gt; but can't prove yourself wrong either&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;You receive mixed feedback&lt;/strong&gt; and can't make sense of it&lt;/li&gt;
&lt;/ol&gt;

&lt;p&gt;The feedback from a Fortune 500 CFO and a freelance designer will be wildly different — even if they're both technically "potential customers." If you're talking to both at once, their conflicting needs cancel each other out and you end up paralysed.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;Rule of Thumb:&lt;/strong&gt; If you aren't finding consistent problems and goals, you don't have a specific enough customer segment.&lt;/p&gt;




&lt;h2&gt;
  
  
  Babies or Body Builders?
&lt;/h2&gt;

&lt;p&gt;Fitzpatrick shares a brilliant example. A woman had developed a powdered condiment — sweet like cinnamon brown sugar but packed with the nutrition of a multivitamin. An all-natural superfood you could survive on indefinitely.&lt;/p&gt;

&lt;p&gt;She said it had countless uses: moms could sprinkle it on breakfast to trick their kids into being healthy, restaurants could leave it on tables as a sugar alternative, bodybuilders could mix it into protein shakes.&lt;/p&gt;

&lt;p&gt;Sounds like a huge market, right? Wrong. She was running in circles. The bodybuilders wanted one thing, the restaurants wanted another, and the moms needed a third. Making one group happy always disappointed the others. She didn't know how to start. Even simple decisions — like what colour to use for the label — were impossible to answer because every group had different preferences.&lt;/p&gt;

&lt;p&gt;The fix? She had to choose. She focused on &lt;strong&gt;moms with young kids who were already shopping at health food stores.&lt;/strong&gt; Now she knew who to talk to, where to find them, and what they cared about.&lt;/p&gt;

&lt;p&gt;Her next move was clever: she went to small, independent health food stores and asked them to place a few bottles of her condiment beside the breakfast foods. This was a great commitment to ask for — it gave her real shelf space and helped distinguish between stores that were just being polite and ones that were genuinely interested. She'd return in a week to check if the product sold and to talk to store owners about their experience.&lt;/p&gt;

&lt;p&gt;Choosing a specific segment felt like losing all the other options. But it was the only way to actually make progress. As Fitzpatrick puts it: before we can serve &lt;em&gt;everyone&lt;/em&gt;, we have to serve &lt;em&gt;someone&lt;/em&gt;.&lt;/p&gt;




&lt;h2&gt;
  
  
  Big Brands or Mom &amp;amp; Pop?
&lt;/h2&gt;

&lt;p&gt;Fitzpatrick shares his own painful lesson here. In one of his startups, he was thrilled that his customer segment was "advertisers." Everyone advertises somehow, so the market was practically infinite!&lt;/p&gt;

&lt;p&gt;He talked to mom-and-pop shops, e-tailers, big brands, creative agencies, SMEs, music labels — anyone who spent money on advertising. And the result? Complete chaos.&lt;/p&gt;

&lt;p&gt;Everything they tried &lt;em&gt;sort of&lt;/em&gt; worked. Everything was &lt;em&gt;somewhat&lt;/em&gt; promising. Some people were talking about paying $10,000/month while others scoffed at $10. Every new feature was moderately popular. But if they tried to cut any feature, someone would scream because it was their favourite part.&lt;/p&gt;

&lt;p&gt;The fundamental problem: &lt;strong&gt;they couldn't prove themselves right or wrong.&lt;/strong&gt; They were paying attention to so many customer types that there was pretty much always &lt;em&gt;someone&lt;/em&gt; who liked a new idea. But making a so-so product for a bunch of audiences isn't the same as making an incredible product for one.&lt;/p&gt;

&lt;p&gt;Eventually they noticed unusually strong signals from creative agencies who wanted to be edgy. They ignored everyone else, cut a bunch of features, and were finally able to get a clear picture of what was working and what wasn't.&lt;/p&gt;




&lt;h2&gt;
  
  
  But What Does It Mean?
&lt;/h2&gt;

&lt;p&gt;This section is one of the most eye-opening in the entire book. Fitzpatrick describes two founders who were doing everything right — asking good Mom Test questions, pushing for commitments, using every sales meeting as a learning opportunity. And yet they were still completely confused.&lt;/p&gt;

&lt;p&gt;After 20 conversations, they had 20 different must-have features and 20 separate must-solve problems. The more people they talked to, the more confused they got. What was going on?&lt;/p&gt;

&lt;p&gt;Their customer segment was too broad, but in a sneaky way. They were building something for "students." Sounds specific enough, right? But think about what "students" actually means:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;A PhD student at a research university&lt;/li&gt;
&lt;li&gt;An ambitious teenager at a prep school&lt;/li&gt;
&lt;li&gt;A homeschooling parent who wants to use it with her kid&lt;/li&gt;
&lt;li&gt;A child in a rural Indian village self-educating through a shared computer&lt;/li&gt;
&lt;li&gt;A student in Africa running the app off a shaky cellphone connection&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;All are technically "students." But they have completely different needs, workflows, devices, budgets, and goals. These founders weren't having 20 conversations with their customers — they were having &lt;strong&gt;one conversation each with 20 different types of customers.&lt;/strong&gt; That's why the feedback was so inconsistent.&lt;/p&gt;

&lt;p&gt;The same thing happens with segments like "small businesses," "developers," or "sales organisations." They sound specific but contain enormous variation underneath.&lt;/p&gt;




&lt;h2&gt;
  
  
  Customer Slicing
&lt;/h2&gt;

&lt;p&gt;So how do you narrow down a broad segment into something useful? Fitzpatrick introduces a technique called &lt;strong&gt;Customer Slicing.&lt;/strong&gt; It's simple but powerful.&lt;/p&gt;

&lt;p&gt;Start with a broad segment and keep slicing it into smaller and smaller sub-sets by asking these questions:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;Within this group, which type of person would want it most?&lt;/li&gt;
&lt;li&gt;Would everyone in this group buy/use it, or only some?&lt;/li&gt;
&lt;li&gt;Why does that sub-set want it? What is their specific problem?&lt;/li&gt;
&lt;li&gt;Does everyone in the group have that motivation, or only some?&lt;/li&gt;
&lt;li&gt;What additional motivations are there?&lt;/li&gt;
&lt;li&gt;Which other types of people have these motivations?&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;You repeat this process until you end up with a segment that is specific enough to &lt;strong&gt;find and reach&lt;/strong&gt; in the real world.&lt;/p&gt;

&lt;p&gt;Here's Fitzpatrick's example: Say you're building a high-end fitness gadget for busy professionals. Your first instinct might be: "finance professionals, age 25-35, living in a major city." That sounds specific, but it's actually useless — it doesn't help you make product decisions and it doesn't help you find them.&lt;/p&gt;

&lt;p&gt;Slice further: finance professionals in London who are currently training for a marathon. Better! Now slice again: the sub-sub-subset who go to the gym during their lunch hour. Now you know exactly where to find them — at a gym in London's financial district during lunch time. You can have all the customer conversations you want for the price of a gym membership.&lt;/p&gt;

&lt;p&gt;The key test: &lt;strong&gt;if there isn't a clear physical or digital location where you can find your customer segment, it's probably still too broad.&lt;/strong&gt; Go back and slice it into finer pieces.&lt;/p&gt;

&lt;p&gt;Once you have a bunch of "who-where" pairs, decide who to start with based on three criteria:&lt;/p&gt;

&lt;ol&gt;
&lt;li&gt;
&lt;strong&gt;Profitable or big&lt;/strong&gt; — Is this segment worth pursuing financially?&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;Easy to reach&lt;/strong&gt; — Can you actually find and talk to these people?&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;Personally rewarding&lt;/strong&gt; — Do you enjoy working with this group?&lt;/li&gt;
&lt;/ol&gt;

&lt;p&gt;Don't overthink it. Spend a few minutes to reach a concrete initial segment, find a few of them, and start learning. You can always broaden your segment later once you've nailed it for the first group.&lt;/p&gt;

&lt;p&gt;Fitzpatrick adds a personal note that I really appreciate: the third factor — personally rewarding — matters more than people think. This stuff is hard work, and it can become a real grind if you're cynical about the people or the industry you're serving. Choose customers you admire and enjoy being around.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;Rule of Thumb:&lt;/strong&gt; Good customer segments are a who-where pair. If you don't know where to go to find your customers, keep slicing your segment into smaller pieces until you do.&lt;/p&gt;




&lt;h2&gt;
  
  
  Talking to the Wrong People
&lt;/h2&gt;

&lt;p&gt;Even with good segmentation, you can still end up talking to the wrong people. Fitzpatrick identifies three ways this happens:&lt;/p&gt;

&lt;h3&gt;
  
  
  1. Your segment is too broad and you're talking to everyone
&lt;/h3&gt;

&lt;p&gt;We've covered this extensively. If your feedback is all over the place, you're probably mixing multiple customer types together. The fix: slice your segment narrower.&lt;/p&gt;

&lt;h3&gt;
  
  
  2. You have multiple customer segments and missed some of them
&lt;/h3&gt;

&lt;p&gt;Sometimes it's obvious — if you're a two-sided marketplace (like Airbnb), you clearly have hosts and guests as separate segments. But sometimes it's sneakier. If you're building an app for kids, you need to understand both the kids AND their parents. If you're building for public schools, you need the teachers, the students, the administration, and potentially even the parent-teacher association and tax payers.&lt;/p&gt;

&lt;p&gt;You'll also need to worry about important partners — whether for manufacturing, distribution, or promotion. If your business relies on them, you need to understand their goals and constraints just as well as your customers'.&lt;/p&gt;

&lt;h3&gt;
  
  
  3. You're selling to businesses with complicated buying processes and have overlooked some stakeholders
&lt;/h3&gt;

&lt;p&gt;This is the B2B trap. Don't fall into only talking to the most senior or impressive people you can find. You want to talk to people who are &lt;strong&gt;representative of your actual customers&lt;/strong&gt;, not the ones who sound impressive on your status report.&lt;/p&gt;

&lt;p&gt;Fitzpatrick admits making this mistake himself: when building interactive advertising products, he spent lots of time talking to executives and none talking to the kids who were supposed to actually love the products.&lt;/p&gt;

&lt;p&gt;If you're in a multi-sided marketplace, yes, you need to run customer conversations separately for all the various segments. But hopefully that isn't as scary now that you know how to keep conversations casual and efficient.&lt;/p&gt;




&lt;h2&gt;
  
  
  Key Takeaways from Chapter 7
&lt;/h2&gt;

&lt;ol&gt;
&lt;li&gt;&lt;p&gt;&lt;strong&gt;Startups drown, they don't starve.&lt;/strong&gt; Too many options is the real killer. Segmentation is your life raft.&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;&lt;strong&gt;Every big company started small.&lt;/strong&gt; Google, Paypal, Evernote — they all began by serving a tiny, specific group before expanding to the world.&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;&lt;strong&gt;"Everyone" is not a customer segment.&lt;/strong&gt; If your feedback is inconsistent, your segment is too broad. Slice it down.&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;&lt;strong&gt;Use Customer Slicing.&lt;/strong&gt; Keep asking "within this group, who wants it most?" until you have a who-where pair — a specific person you can physically go find and talk to.&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;&lt;strong&gt;Choose segments that are profitable, reachable, and personally rewarding.&lt;/strong&gt; Don't just pick the biggest market — pick one you can actually serve and enjoy serving.&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;&lt;strong&gt;Watch out for hidden segments.&lt;/strong&gt; Multi-sided marketplaces, products for kids (parents are the buyers), B2B with complex buying processes — make sure you're talking to all the relevant groups.&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;&lt;strong&gt;Don't over-plan.&lt;/strong&gt; Spend a few minutes choosing an initial segment, go talk to them, and adjust as you learn. You can always broaden later.&lt;/p&gt;&lt;/li&gt;
&lt;/ol&gt;




&lt;p&gt;&lt;em&gt;This is part of my series where I break down each chapter of The Mom Test by Rob Fitzpatrick. If you're building a product and talking to customers, this book is essential reading.&lt;/em&gt;&lt;/p&gt;

&lt;p&gt;&lt;em&gt;Previously: &lt;a href="https://dev.to/egepakten/what-i-learned-from-the-mom-test-chapter-6-finding-conversations-4bjc"&gt;Chapter 6 - Finding Conversations&lt;/a&gt;&lt;/em&gt;&lt;br&gt;
&lt;em&gt;Next up: Chapter 8 - Running the Process&lt;/em&gt;&lt;/p&gt;

</description>
      <category>webdev</category>
      <category>startup</category>
      <category>speaking</category>
      <category>learning</category>
    </item>
    <item>
      <title>What I Learned from "The Mom Test" - Chapter 6: Finding Conversations</title>
      <dc:creator>Ege Pakten</dc:creator>
      <pubDate>Fri, 06 Mar 2026 10:16:10 +0000</pubDate>
      <link>https://dev.to/egepakten/what-i-learned-from-the-mom-test-chapter-6-finding-conversations-4bjc</link>
      <guid>https://dev.to/egepakten/what-i-learned-from-the-mom-test-chapter-6-finding-conversations-4bjc</guid>
      <description>&lt;p&gt;&lt;em&gt;A developer's guide to actually finding people to talk to — and making them want to talk to you&lt;/em&gt;&lt;/p&gt;




&lt;p&gt;So you know how to ask good questions (Chapter 1-3), keep things casual (Chapter 4), and push for real commitments (Chapter 5). Great. But there's one problem nobody talks about enough: &lt;strong&gt;where do you actually find these people to talk to?&lt;/strong&gt;&lt;/p&gt;

&lt;p&gt;Chapter 6 is Fitzpatrick's playbook for finding conversations. And the core insight surprised me: you don't need to be a networking wizard. You just need to be strategic about where you show up and how you ask.&lt;/p&gt;




&lt;h2&gt;
  
  
  Outline
&lt;/h2&gt;

&lt;p&gt;In this post, I'll break down Chapter 6 into the following sections:&lt;/p&gt;

&lt;ol&gt;
&lt;li&gt;
&lt;strong&gt;Going to Them: Cold Outreach&lt;/strong&gt; — Why cold conversations are a necessary evil and how to make the most of them (cold calls, serendipity, finding excuses, immersing yourself, landing pages, and getting clever).&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;Bringing Them to You&lt;/strong&gt; — How to flip the dynamic so customers find you instead (organising meetups, speaking &amp;amp; teaching, industry blogging).&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;Creating Warm Intros&lt;/strong&gt; — The gold standard for conversations and how to manufacture them (7 degrees of bacon, industry advisors, universities, investors, and cashing in favours).&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;Asking For and Framing the Meeting&lt;/strong&gt; — The 5-element framework (Vision/Framing/Weakness/Pedestal/Ask) for getting meetings without sounding salesy.&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;To Commute or to Call&lt;/strong&gt; — Why in-person beats phone calls, and when exceptions make sense.&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;The Advisory Flip&lt;/strong&gt; — A powerful mindset shift that makes customer conversations feel natural instead of desperate.&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;How Many Meetings Do You Actually Need?&lt;/strong&gt; — When to stop talking and start building.&lt;/li&gt;
&lt;/ol&gt;




&lt;h2&gt;
  
  
  Going to Them: Cold Outreach
&lt;/h2&gt;

&lt;p&gt;Let's be real — cold outreach sucks. Nobody likes cold calls, and nobody likes receiving them. But sometimes, especially at the very beginning, you have no choice. You don't know anyone in the industry yet, and you need to start somewhere.&lt;/p&gt;

&lt;p&gt;The key insight Fitzpatrick shares is this: &lt;strong&gt;the goal of cold conversations is to stop having them.&lt;/strong&gt; You hustle together the first one or two from wherever you can, treat people's time respectfully, genuinely try to solve their problem, and those cold conversations start turning into warm intros. The snowball starts rolling.&lt;/p&gt;

&lt;h3&gt;
  
  
  Cold Calls
&lt;/h3&gt;

&lt;p&gt;If you reach out to 100 people and 98 of them ignore you, what does that mean? It means you now have 2 conversations in play. That's it. Unless your entire business model depends on cold outreach, the rejection rate is irrelevant. You only need a few yeses to get the ball rolling.&lt;/p&gt;

&lt;p&gt;One team successfully used cold LinkedIn messages to reach C-level executives at major UK retailers. They were ignored by practically everyone, but they only needed one "yes" to get started. And that one "yes" led to warm intros to others.&lt;/p&gt;

&lt;h3&gt;
  
  
  Seizing Serendipity
&lt;/h3&gt;

&lt;p&gt;Beyond cold outreach, stay open to unexpected opportunities. Fitzpatrick shares a story about being at an engagement party and overhearing someone mention a speaking engagement. He walked over, started a genuine conversation about her career, and she ended up becoming his first committed alpha user.&lt;/p&gt;

&lt;p&gt;The lesson? If you stop thinking of customer conversations as formal "interviews" and start thinking of them as genuine conversations about people's lives and problems, opportunities are everywhere. People love talking about their problems — by showing genuine interest, you become more interesting than 99% of people they've met.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;Rule of Thumb:&lt;/strong&gt; If it's not a formal meeting, you don't need to make excuses about why you're there or even mention that you're starting a business. Just ask about their life.&lt;/p&gt;

&lt;h3&gt;
  
  
  Find a Good Excuse
&lt;/h3&gt;

&lt;p&gt;Sometimes you need a reason to start a conversation with a stranger. Fitzpatrick tells a great story about an entrepreneur who was building a product for cafe owners. He'd been hitting the pavement for weeks, getting turned away from cafe after cafe. The fix? Walking into a cafe and saying: "This coffee is amazing — I wanted to ask about the story behind the beans." Suddenly, the conversation was natural. The owner wasn't around, but the manager gave them the owner's contact details.&lt;/p&gt;

&lt;p&gt;The trade-off with using an excuse is that it's hard to transition into a product or sales conversation later without revealing the initial deception. So think of these as one-time learning opportunities, not the start of an ongoing relationship.&lt;/p&gt;

&lt;p&gt;The ultimate excuse? Having a PhD student on your team. "I'm doing research on X for my dissertation" opens almost any door. If you're really desperate, you can always say you're "writing a book" on the topic.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;Rule of Thumb:&lt;/strong&gt; If it's a topic you both care about, find an excuse to talk about it. Your idea never needs to enter the equation and you'll both enjoy the chat.&lt;/p&gt;

&lt;h3&gt;
  
  
  Immerse Yourself in Where They Are
&lt;/h3&gt;

&lt;p&gt;This is one of the most underrated strategies. Instead of trying to find customers from the outside, go to where they already gather.&lt;/p&gt;

&lt;p&gt;Fitzpatrick wanted to build tools for conference organisers and professional speakers. He didn't know any of the big names. So he hit the conference circuit and gave free talks everywhere he could. The speakers' lounge became his personal customer conversation machine. By immersing himself in the community, he met a load of people and had all the connections and conversations he could handle.&lt;/p&gt;

&lt;p&gt;Interestingly, he ultimately decided that big speakers and big conferences were a bad customer segment and walked away. Not every conversation has to validate your idea — sometimes the most valuable learning is discovering what NOT to build.&lt;/p&gt;

&lt;h3&gt;
  
  
  Landing Pages
&lt;/h3&gt;

&lt;p&gt;Joel Gascoigne did a classic landing page test with his startup Buffer. He described the value proposition and collected emails. But here's what most people miss: it wasn't the conversion rate metrics that convinced him to move forward. It was the conversations that resulted from him emailing every single person who signed up and saying hello.&lt;/p&gt;

&lt;p&gt;Landing pages are a great way to collect qualified leads that you can then reach out to and have real conversations with. Paul Graham suggests a similar approach: get your product out there, see who seems to like it most, and then reach out to those types of users for deeper learning.&lt;/p&gt;

&lt;h3&gt;
  
  
  Get Clever
&lt;/h3&gt;

&lt;p&gt;Every business is different, and sometimes you need a creative approach. One guy wanted to sell to top-tier universities like Stanford and Harvard. He needed to understand their problems and be taken seriously by decision-makers. His solution? He organised a semi-monthly "knowledge exchange" call between department heads of top universities to discuss shared challenges. By simply organising the call and playing host, he immediately absorbed all the credibility of the participating universities and got direct phone access to a pile of great leads.&lt;/p&gt;

&lt;p&gt;The point is: don't just copy what someone else is doing. Consider your unique situation and get creative about how to manufacture conversations.&lt;/p&gt;




&lt;h2&gt;
  
  
  Bringing Them to You
&lt;/h2&gt;

&lt;p&gt;When you're the one initiating conversations, you're always on the back foot. The other person is suspicious and trying to figure out if you're wasting their time. The better approach? Find ways to make them come to you. This saves time, reduces friction, and makes people take you more seriously.&lt;/p&gt;

&lt;h3&gt;
  
  
  Organise Meetups
&lt;/h3&gt;

&lt;p&gt;For marginally more effort than attending an event, you can organise your own and benefit from being the centre of attention. Want to understand HR professionals' problems? Organise an "HR professionals happy hour." People will assume you're credible just because you're the one who sent the invite emails. It's the fastest and most unfair trick for rapid customer learning, and it also bootstraps your industry credibility.&lt;/p&gt;

&lt;h3&gt;
  
  
  Speaking &amp;amp; Teaching
&lt;/h3&gt;

&lt;p&gt;Teaching is massively under-valued as both a learning and selling tool. If you're building better project management software, you probably have expertise and strong opinions about how things could be better. That's the magic combination for being an effective teacher.&lt;/p&gt;

&lt;p&gt;Teach at conferences, workshops, through online videos, blogging, and by doing free consulting or office hours. You'll refine your message, connect with potential customers who take you seriously, and learn which parts of your offering resonate — all before you've even built it. Then simply chat up the attendees who are most keen.&lt;/p&gt;

&lt;h3&gt;
  
  
  Industry Blogging
&lt;/h3&gt;

&lt;p&gt;Even if you have no audience, blogging about your industry is incredibly helpful. When Fitzpatrick sent cold emails from his blog email address, people would often meet with him because they had checked his domain, seen his industry blog, and figured he was an interesting person to talk to. The traffic and audience were almost irrelevant — the blog served as a credibility signal.&lt;/p&gt;

&lt;p&gt;Blogging is also a great exercise for getting your thoughts in a row. It makes you a better customer conversationalist because you've already spent time thinking deeply about the industry's problems.&lt;/p&gt;




&lt;h2&gt;
  
  
  Creating Warm Intros
&lt;/h2&gt;

&lt;p&gt;Warm intros are the gold standard. Conversations are infinitely easier when you get an introduction through a mutual friend that establishes your credibility and reason for being there.&lt;/p&gt;

&lt;h3&gt;
  
  
  7 Degrees of Bacon
&lt;/h3&gt;

&lt;p&gt;The world is smaller than you think. Everyone knows someone. You just have to ask.&lt;/p&gt;

&lt;p&gt;Fitzpatrick tells a story about a team of recent graduates who needed to reach McKinsey-style consultants. They were at a co-working space, so one of them stood on a chair and yelled: "Does anyone here know anyone who works at McKinsey? Can we talk to you for a second? We'll buy you a beer!" They bought three beers, had three quick chats, and left with a diary full of intros.&lt;/p&gt;

&lt;p&gt;For consumer products, it's even easier — everyone knows a recent mom, an amateur athlete, or a theatre enthusiast.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;Rule of Thumb:&lt;/strong&gt; You can find anyone you need if you ask for it a couple of times. Kevin Bacon's 7 degrees of separation applies to customer conversations too.&lt;/p&gt;

&lt;h3&gt;
  
  
  Industry Advisors
&lt;/h3&gt;

&lt;p&gt;Advisors can be a goldmine for intros. In his first company, Fitzpatrick relied heavily on 5 advisors who each had around half a percent of equity. Their main job was to make credible introductions. He met with each one once per month and got a fresh batch of intros weekly without it being a huge time burden for anyone.&lt;/p&gt;

&lt;p&gt;You'd also be surprised by the quality of people willing to join your advisory board. The first conversation with a good advisor looks a lot like the first conversation with a flagship customer — you're both talking passionately about a space you care about. You can sometimes recruit great advisors directly from your early customer conversations.&lt;/p&gt;

&lt;h3&gt;
  
  
  Universities
&lt;/h3&gt;

&lt;p&gt;If you're still in or recently out of university, professors are a goldmine. They get their grant funding from friendly, high-level industry folks, and since they're investing in research, those industry contacts are self-selected to be excited about new projects. Plus, professors are easy to reach — they post their emails publicly and you can generally just wander into their office.&lt;/p&gt;

&lt;h3&gt;
  
  
  Investors
&lt;/h3&gt;

&lt;p&gt;Top-tier investors are awesome for B2B intros. Beyond their own rolodex and company portfolio, they can usually pull off cold intros to practically any industry. They can also help you close better advisors and directors than you'd be able to wrangle on your own. This applies to anyone who is a "big deal" and has already bought into your idea — always ask: who can they connect you to?&lt;/p&gt;

&lt;h3&gt;
  
  
  Cash in Favours
&lt;/h3&gt;

&lt;p&gt;Remember all those people who brushed you off saying "Sounds great, keep me in the loop and let me know how I can help"? Now's the time to call in those favours. Reply to that old email and tell them you're ready for an intro to that person they know. Use the framing format from the next section to make their lives easy and reassure them you won't waste anyone's time.&lt;/p&gt;

&lt;p&gt;You'll get ignored a lot, but who cares? You're not trying to minimise your failure rate — you're trying to get a few conversations going. Don't make a habit of it though, since it can burn bridges.&lt;/p&gt;




&lt;h2&gt;
  
  
  Asking For and Framing the Meeting
&lt;/h2&gt;

&lt;p&gt;Sometimes a proper meeting can't be avoided — you need the full hour with someone senior. In those cases, how you frame the meeting request makes all the difference.&lt;/p&gt;

&lt;p&gt;If you don't frame it properly, it becomes a sales meeting by default, which is bad for three reasons: the customer closes up about pricing, attention shifts to you instead of them, and it's going to be the worst sales meeting ever because you aren't ready.&lt;/p&gt;

&lt;h3&gt;
  
  
  The Five Key Elements
&lt;/h3&gt;

&lt;p&gt;Fitzpatrick outlines a framework for requesting meetings that works incredibly well. The five elements are: &lt;strong&gt;Vision / Framing / Weakness / Pedestal / Ask&lt;/strong&gt;&lt;/p&gt;

&lt;p&gt;The mnemonic is: &lt;strong&gt;"Very Few Wizards Properly Ask [for help]."&lt;/strong&gt;&lt;/p&gt;

&lt;ol&gt;
&lt;li&gt;
&lt;strong&gt;Vision&lt;/strong&gt; — You're an entrepreneur trying to solve a problem or achieve a vision. Don't mention your specific idea.&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;Framing&lt;/strong&gt; — Set expectations about what stage you're at and, if true, that you don't have anything to sell.&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;Weakness&lt;/strong&gt; — Show vulnerability by mentioning a specific problem you're struggling with. This also clarifies you're not a time waster.&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;Pedestal&lt;/strong&gt; — Put them on a pedestal by showing how much they, in particular, can help.&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;Ask&lt;/strong&gt; — Explicitly ask for help.&lt;/li&gt;
&lt;/ol&gt;

&lt;p&gt;Here's an example of what this looks like in practice:&lt;/p&gt;

&lt;p&gt;As Fitzpatrick suggests, a good meeting request might look like this:&lt;/p&gt;

&lt;blockquote&gt;
&lt;p&gt;Hey Pete,&lt;/p&gt;

&lt;p&gt;I'm trying to make desk &amp;amp; office rental less of a pain for new businesses (&lt;em&gt;vision&lt;/em&gt;). We're just starting out and don't have anything to sell, but want to make sure we're building something that actually helps (&lt;em&gt;framing&lt;/em&gt;).&lt;/p&gt;

&lt;p&gt;I've only ever come at it from the tenant's side and I'm having a hard time understanding how it all works from the landlord's perspective (&lt;em&gt;weakness&lt;/em&gt;). You've been renting out desks for a while and could really help me cut through the fog (&lt;em&gt;pedestal&lt;/em&gt;).&lt;/p&gt;

&lt;p&gt;Do you have time in the next couple weeks to meet up for a chat? (&lt;em&gt;ask&lt;/em&gt;)&lt;/p&gt;

&lt;p&gt;— Rob Fitzpatrick, &lt;em&gt;The Mom Test&lt;/em&gt;&lt;/p&gt;
&lt;/blockquote&gt;

&lt;p&gt;People like to help entrepreneurs, but they also hate wasting their time. This kind of opening tells them you know what you need and that they'll be able to make a real difference.&lt;/p&gt;

&lt;p&gt;Once the meeting starts, you need to grab the reins quickly. Repeat what you said in the email and immediately drop into your first question. If someone else made the introduction, use them as a voice of authority. Have a plan for the meeting and be assertive about keeping it on track.&lt;/p&gt;




&lt;h2&gt;
  
  
  To Commute or to Call
&lt;/h2&gt;

&lt;p&gt;One common shortcut is to move conversations to phone or video calls. Fitzpatrick is not a fan, and for good reason.&lt;/p&gt;

&lt;p&gt;When you're in person, you get access to body language, facial expressions, and the natural rapport that comes from sharing a physical space. On the phone, people are trying to squeeze calls between other activities, wondering when they can hang up, and the whole thing ends up feeling more like a scripted interview than a natural conversation.&lt;/p&gt;

&lt;p&gt;In-person meetings also have a huge advantage for building ongoing relationships. Nobody becomes friends over the phone. And those friendships are what lead to warm intros and future meetings.&lt;/p&gt;

&lt;p&gt;That said, some experienced practitioners in the field do recommend phone calls, and they can work. But Fitzpatrick's advice is to &lt;strong&gt;start in person&lt;/strong&gt; first. It's too easy to use phone calls as an excuse to skip the awkwardness of meeting someone face-to-face, rather than as a considered trade-off.&lt;/p&gt;




&lt;h2&gt;
  
  
  The Advisory Flip
&lt;/h2&gt;

&lt;p&gt;This is a subtle but powerful mindset shift. Instead of going into conversations thinking &lt;em&gt;"I need to find customers"&lt;/em&gt;, think &lt;em&gt;"I'm looking for industry advisors."&lt;/em&gt;&lt;/p&gt;

&lt;p&gt;When you're looking for customers, you feel needy. You're on the back foot. You're basically asking: "Please buy my thing." But when you're looking for advisors, the whole dynamic flips. You're evaluating them. You're the one deciding if they're a good fit. Even if the topics you discuss are the same, both you and the other person will notice the difference.&lt;/p&gt;

&lt;p&gt;This isn't about explicitly telling people you're looking for advisors (unless you already like them and it comes up naturally). It's about orienting your state of mind to give you a helpful internal narrative and consistent front.&lt;/p&gt;

&lt;p&gt;The sales-advisor switch also puts you firmly in control of the meeting, since you're now evaluating them. You set the agenda, you keep it on topic, and you propose next steps.&lt;/p&gt;




&lt;h2&gt;
  
  
  How Many Meetings Do You Actually Need?
&lt;/h2&gt;

&lt;p&gt;Every meeting has an opportunity cost. When you're travelling to that meeting, you aren't writing code or building your product. So how many conversations is enough?&lt;/p&gt;

&lt;p&gt;The UX community says: &lt;strong&gt;keep talking to people until you stop hearing new information.&lt;/strong&gt;&lt;/p&gt;

&lt;p&gt;In practice, if your initial assumptions are mostly correct and you're in a simple industry, it might only take 3-5 conversations to confirm what you already believe. But you usually won't get that lucky. It often takes more until you start hearing a consistent message.&lt;/p&gt;

&lt;p&gt;Here's a useful diagnostic: if you've run more than 10 conversations and the results are all over the map, your customer segment is probably too vague. You might be mashing together feedback from multiple different types of customers.&lt;/p&gt;

&lt;p&gt;The goal isn't to have a thousand meetings. It's about quickly learning what you need, and then getting back to building. In most cases, you should be able to answer almost any question about your business or customers within a week.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;Rule of Thumb:&lt;/strong&gt; Keep having conversations until you stop hearing new stuff. If you're still getting wildly different answers after 10+ conversations, your customer segment is too broad.&lt;/p&gt;




&lt;h2&gt;
  
  
  Key Takeaways from Chapter 6
&lt;/h2&gt;

&lt;ol&gt;
&lt;li&gt;&lt;p&gt;&lt;strong&gt;The goal of cold outreach is to stop doing cold outreach.&lt;/strong&gt; Hustle the first few conversations, then convert them into warm intros.&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;&lt;strong&gt;Go where your customers already are.&lt;/strong&gt; Conferences, meetups, online communities — immerse yourself in their world instead of trying to pull them into yours.&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;&lt;strong&gt;Make them come to you.&lt;/strong&gt; Organise events, teach, blog. Being the host or the expert flips the power dynamic in your favour.&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;&lt;strong&gt;Warm intros beat everything.&lt;/strong&gt; Ask your network, advisors, investors, and even previous contacts for introductions. Everyone knows someone.&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;&lt;strong&gt;Frame your meeting request properly.&lt;/strong&gt; Use the VFWPA framework (Vision/Framing/Weakness/Pedestal/Ask) to get meetings without sounding salesy.&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;&lt;strong&gt;Start in person.&lt;/strong&gt; Phone calls are a shortcut that often backfires. In-person conversations build better relationships and give you better data.&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;&lt;strong&gt;Flip your mindset.&lt;/strong&gt; You're not looking for customers — you're looking for advisors. This one mental switch changes everything about how the conversation feels.&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;&lt;strong&gt;Know when to stop.&lt;/strong&gt; Keep talking until you stop learning new things, then get back to building.&lt;/p&gt;&lt;/li&gt;
&lt;/ol&gt;




&lt;p&gt;&lt;em&gt;This is part of my series where I break down each chapter of The Mom Test by Rob Fitzpatrick. If you're building a product and talking to customers, this book is essential reading.&lt;/em&gt;&lt;/p&gt;

&lt;p&gt;&lt;em&gt;Previously: &lt;a href="https://dev.to/egepakten/the-mom-test-chapter-5-commitment-and-advancement-1k27"&gt;Chapter 5 - Commitment and Advancement&lt;/a&gt;&lt;/em&gt;&lt;br&gt;
&lt;em&gt;Next up: Chapter 7 - Choosing Your Customers&lt;/em&gt;&lt;/p&gt;

</description>
      <category>webdev</category>
      <category>startup</category>
      <category>speaking</category>
      <category>learning</category>
    </item>
    <item>
      <title>The Mom Test - Chapter 5: Commitment and Advancement</title>
      <dc:creator>Ege Pakten</dc:creator>
      <pubDate>Mon, 02 Mar 2026 11:03:54 +0000</pubDate>
      <link>https://dev.to/egepakten/the-mom-test-chapter-5-commitment-and-advancement-1k27</link>
      <guid>https://dev.to/egepakten/the-mom-test-chapter-5-commitment-and-advancement-1k27</guid>
      <description>&lt;p&gt;In the previous chapters, we learned how to have proper customer conversations — avoiding compliments, digging into specifics, and not pitching too early. But here's a question that kept bugging me: &lt;strong&gt;How do I know if a meeting actually went well?&lt;/strong&gt;&lt;/p&gt;

&lt;p&gt;Chapter 5 answers exactly that. And the answer is brutally simple: &lt;strong&gt;a meeting went well only if it ends with a commitment.&lt;/strong&gt;&lt;/p&gt;




&lt;h2&gt;
  
  
  Outline
&lt;/h2&gt;

&lt;p&gt;In this post, I'll break down Chapter 5 into the following sections:&lt;/p&gt;

&lt;ol&gt;
&lt;li&gt;
&lt;strong&gt;There's No Such Thing as a Meeting That "Went Well"&lt;/strong&gt; — Why every meeting either succeeds or fails, and how compliments trick you into thinking you're making progress.&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;Commitment and Advancement: Two Sides of the Same Coin&lt;/strong&gt; — The two key concepts of the chapter and why they always come together.&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;The Currencies of Commitment&lt;/strong&gt; — The three types of commitment (Time, Reputation, Money) and how they escalate in seriousness.&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;The Spectrum: From Zombie Lead to Committed Customer&lt;/strong&gt; — How to read the signals and know exactly where you stand with a potential customer.&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;Why We Don't Ask for Commitments (And Why We Should)&lt;/strong&gt; — The two traps that prevent us from getting real signals: fishing for compliments and not asking for next steps.&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;The "Crazy" First Customers: Your Early Evangelists&lt;/strong&gt; — Why your first customers won't be "normal" buyers, and why that's a feature, not a bug.&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;How to Push for Commitment Without Being a Used Car Salesman&lt;/strong&gt; — A practical framework for asking for commitments without feeling pushy.&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;Don't Ask for Commitment Too Early&lt;/strong&gt; — Why timing matters and how to match your ask to the stage of the relationship.&lt;/li&gt;
&lt;/ol&gt;




&lt;h2&gt;
  
  
  There's No Such Thing as a Meeting That "Went Well"
&lt;/h2&gt;

&lt;p&gt;This was a mindset shift for me. I used to walk out of meetings thinking &lt;em&gt;"That went great! They loved the idea!"&lt;/em&gt; — and then... nothing happened. No follow-up, no next steps, just silence.&lt;/p&gt;

&lt;p&gt;Fitzpatrick puts it bluntly:&lt;/p&gt;

&lt;blockquote&gt;
&lt;p&gt;Every meeting either succeeds or fails.&lt;/p&gt;
&lt;/blockquote&gt;

&lt;p&gt;A meeting &lt;strong&gt;fails&lt;/strong&gt; when you leave with:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;A compliment: &lt;em&gt;"That's a really cool idea!"&lt;/em&gt;
&lt;/li&gt;
&lt;li&gt;A stalling tactic: &lt;em&gt;"Let's circle back after the holidays."&lt;/em&gt;
&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;A meeting &lt;strong&gt;succeeds&lt;/strong&gt; when you leave with:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;A &lt;strong&gt;commitment&lt;/strong&gt; to the next step&lt;/li&gt;
&lt;li&gt;Something concrete that &lt;strong&gt;advances&lt;/strong&gt; the relationship forward&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;The tricky part? The subtle stalls don't feel like rejection. &lt;em&gt;"We should definitely talk again soon"&lt;/em&gt; sounds positive, but it's just a polished version of &lt;em&gt;"Don't call me, I'll call you."&lt;/em&gt;&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;Rule of Thumb:&lt;/strong&gt; If you leave a meeting feeling good but without a concrete next step, you probably got played by a compliment, not a commitment.&lt;/p&gt;




&lt;h2&gt;
  
  
  Commitment and Advancement: Two Sides of the Same Coin
&lt;/h2&gt;

&lt;p&gt;Fitzpatrick introduces two key concepts:&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;Commitment&lt;/strong&gt; — When someone gives you something they value. This proves they're serious and not just being polite.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;Advancement&lt;/strong&gt; — When the relationship moves to the next concrete step in your sales or learning process.&lt;/p&gt;

&lt;p&gt;These two almost always come together. To advance to the next step, someone has to commit something. And if someone commits something, the process naturally advances.&lt;/p&gt;

&lt;p&gt;For example: You want to demo your product to a company's decision-maker. To get that meeting (advancement), your current contact needs to introduce you to their boss (reputation commitment). One doesn't happen without the other.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;Rule of Thumb:&lt;/strong&gt; Commitment and advancement are functionally the same thing. If you're getting one, you're usually getting both. If you're getting neither, the meeting failed.&lt;/p&gt;




&lt;h2&gt;
  
  
  The Currencies of Commitment
&lt;/h2&gt;

&lt;p&gt;Not all commitments are created equal. Fitzpatrick breaks them down into three "currencies" — and they escalate in seriousness:&lt;/p&gt;

&lt;h3&gt;
  
  
  1. Time Commitment
&lt;/h3&gt;

&lt;p&gt;This is the lightest form. The person is investing their time to engage with you further.&lt;/p&gt;

&lt;p&gt;Examples:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;Agreeing to a follow-up meeting with clear next steps&lt;/li&gt;
&lt;li&gt;Sitting down for a longer, deeper conversation&lt;/li&gt;
&lt;li&gt;Trying out your prototype or beta and giving feedback&lt;/li&gt;
&lt;li&gt;Coming to your office (or going out of their way) for a meeting&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;If someone won't even give you another 30 minutes of their time, that's a pretty clear signal.&lt;/p&gt;

&lt;h3&gt;
  
  
  2. Reputation Commitment
&lt;/h3&gt;

&lt;p&gt;This is heavier. The person is putting their name and credibility on the line for you.&lt;/p&gt;

&lt;p&gt;Examples:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;Introducing you to their boss or a decision-maker&lt;/li&gt;
&lt;li&gt;Introducing you to a peer or potential customer&lt;/li&gt;
&lt;li&gt;Giving you a public testimonial or case study&lt;/li&gt;
&lt;li&gt;Posting about you on social media or their company Slack&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;When someone introduces you to their boss, they're essentially saying &lt;em&gt;"I believe in this enough to risk looking stupid if it doesn't work out."&lt;/em&gt; That's real skin in the game.&lt;/p&gt;

&lt;h3&gt;
  
  
  3. Financial Commitment
&lt;/h3&gt;

&lt;p&gt;The ultimate signal. Money talks, everything else walks.&lt;/p&gt;

&lt;p&gt;Examples:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;A letter of intent (LOI) or pre-order&lt;/li&gt;
&lt;li&gt;A deposit or partial payment&lt;/li&gt;
&lt;li&gt;Pre-paying for the product before it's built&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;If someone says &lt;em&gt;"I'd definitely pay for that"&lt;/em&gt; — that means nothing. If someone says &lt;em&gt;"Here's $500, let me know when it's ready"&lt;/em&gt; — that means everything.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;Rule of Thumb:&lt;/strong&gt; The more someone gives you (time → reputation → money), the more seriously you can take their signal. Compliments cost nothing. Commitments cost something. That's the whole difference.&lt;/p&gt;




&lt;h2&gt;
  
  
  The Spectrum: From Zombie Lead to Committed Customer
&lt;/h2&gt;

&lt;p&gt;Fitzpatrick describes a spectrum of signals you might get from potential customers, and it's incredibly useful for figuring out where you actually stand:&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;Cold signals (the meeting failed):&lt;/strong&gt;&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;
&lt;em&gt;"That's cool, I like it"&lt;/em&gt; → compliment, worthless&lt;/li&gt;
&lt;li&gt;
&lt;em&gt;"Looks interesting, keep me in the loop"&lt;/em&gt; → polite brush-off&lt;/li&gt;
&lt;li&gt;
&lt;em&gt;"Let's grab coffee sometime"&lt;/em&gt; → stalling, no specifics&lt;/li&gt;
&lt;li&gt;No follow-up after the meeting → they forgot you exist&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;&lt;strong&gt;Warm signals (getting somewhere):&lt;/strong&gt;&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;
&lt;em&gt;"Can you show this to my team next Tuesday?"&lt;/em&gt; → time + reputation commitment&lt;/li&gt;
&lt;li&gt;
&lt;em&gt;"Send me the beta link, I'll try it this week"&lt;/em&gt; → time commitment with a deadline&lt;/li&gt;
&lt;li&gt;
&lt;em&gt;"Let me introduce you to our Head of Product"&lt;/em&gt; → reputation commitment&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;&lt;strong&gt;Hot signals (you're onto something):&lt;/strong&gt;&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;
&lt;em&gt;"How much would this cost? Can we do a pilot?"&lt;/em&gt; → moving toward financial commitment&lt;/li&gt;
&lt;li&gt;
&lt;em&gt;"We'd like to pre-order 50 licenses"&lt;/em&gt; → money on the table&lt;/li&gt;
&lt;li&gt;
&lt;em&gt;"Here's a deposit, build it"&lt;/em&gt; → they're all in&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;&lt;strong&gt;Rule of Thumb:&lt;/strong&gt; If you can't tell where someone falls on this spectrum, you didn't push hard enough for a commitment at the end of the meeting.&lt;/p&gt;




&lt;h2&gt;
  
  
  Why We Don't Ask for Commitments (And Why We Should)
&lt;/h2&gt;

&lt;p&gt;So if commitments are so important, why don't we ask for them? Fitzpatrick identifies two main traps:&lt;/p&gt;

&lt;h3&gt;
  
  
  Trap 1: You're Fishing for Compliments
&lt;/h3&gt;

&lt;p&gt;Instead of asking &lt;em&gt;"Would you be willing to pay for this?"&lt;/em&gt; or &lt;em&gt;"Can I show this to your boss?"&lt;/em&gt;, we ask soft questions like:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;&lt;em&gt;"What do you think of the idea?"&lt;/em&gt;&lt;/li&gt;
&lt;li&gt;&lt;em&gt;"Would you use something like this?"&lt;/em&gt;&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;These questions are begging for a compliment, not a commitment. And guess what? People are happy to give you a compliment because it costs them nothing and gets you out of the room.&lt;/p&gt;

&lt;h3&gt;
  
  
  Trap 2: You're Not Asking for Next Steps
&lt;/h3&gt;

&lt;p&gt;The meeting is going well. You're vibing. You're having a great conversation. And then... you just let it end. No ask. No push. You walk away with warm feelings and zero concrete progress.&lt;/p&gt;

&lt;p&gt;This is fear dressed up as politeness. We don't want to be "pushy" so we don't ask. But here's the thing — if your product is genuinely solving their problem, asking for a next step isn't pushy. It's helpful.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;Rule of Thumb:&lt;/strong&gt; Always know what commitment you want before the meeting starts. Then ask for it before the meeting ends. If you don't ask, you won't get it. Period.&lt;/p&gt;




&lt;h2&gt;
  
  
  The "Crazy" First Customers: Your Early Evangelists
&lt;/h2&gt;

&lt;p&gt;Fitzpatrick makes an important point about who your first customers will be. They won't be normal, rational, cautious buyers. Your first customers will be a little bit "crazy" — and that's a good thing.&lt;/p&gt;

&lt;p&gt;Your early evangelists typically:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;
&lt;strong&gt;Have the problem right now&lt;/strong&gt;, not "someday"&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;Know they have the problem&lt;/strong&gt; — they're not in denial&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;Have already tried to solve it&lt;/strong&gt; (maybe with spreadsheets, duct tape, or a competitor)&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;Have the budget&lt;/strong&gt; or authority to actually pay for a solution&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;Are desperate enough&lt;/strong&gt; to try an unfinished, unpolished product from an unknown startup&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;Think about it: a normal person wouldn't use a half-built product from two people in a garage. But someone who's in pain RIGHT NOW and has been looking for a solution? They'll tolerate bugs, missing features, and a terrible UI — because you're solving their burning problem.&lt;/p&gt;

&lt;p&gt;These people are gold. They give you real feedback, real money, and real validation.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;Rule of Thumb:&lt;/strong&gt; If you can't find anyone who's desperate enough to use your product in its current state, you either haven't found your real customer segment, or you're not solving a painful enough problem.&lt;/p&gt;




&lt;h2&gt;
  
  
  How to Push for Commitment Without Being a Used Car Salesman
&lt;/h2&gt;

&lt;p&gt;A common fear: &lt;em&gt;"But I don't want to be pushy!"&lt;/em&gt;&lt;/p&gt;

&lt;p&gt;Fitzpatrick's answer: you're not being pushy if you're genuinely trying to help. Here's his framework:&lt;/p&gt;

&lt;ol&gt;
&lt;li&gt;&lt;p&gt;&lt;strong&gt;Know your ask before the meeting.&lt;/strong&gt; What's the ideal next step? An intro to the boss? A pilot program? A pre-order? Know this going in.&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;&lt;strong&gt;Ask at the end of the meeting.&lt;/strong&gt; Don't let the meeting fizzle out. Before wrapping up, clearly state what you'd like to happen next.&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;&lt;strong&gt;Accept the answer gracefully.&lt;/strong&gt; If they say no, that's actually great information. A clear "no" is infinitely more useful than a wishy-washy "maybe." At least now you know where you stand.&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;&lt;strong&gt;Interpret the response honestly.&lt;/strong&gt; If they dodge, stall, or give you a compliment instead of a commitment — recognize it for what it is. Don't lie to yourself.&lt;/p&gt;&lt;/li&gt;
&lt;/ol&gt;

&lt;p&gt;Examples of good asks:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;&lt;em&gt;"Would you be willing to do a trial run with your team next month?"&lt;/em&gt;&lt;/li&gt;
&lt;li&gt;&lt;em&gt;"Could you introduce me to [decision-maker] so I can understand their perspective?"&lt;/em&gt;&lt;/li&gt;
&lt;li&gt;&lt;em&gt;"If we build this by March, would you commit to being a pilot customer?"&lt;/em&gt;&lt;/li&gt;
&lt;li&gt;&lt;em&gt;"Can I get a letter of intent so we can prioritize building this feature for you?"&lt;/em&gt;&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;&lt;strong&gt;Rule of Thumb:&lt;/strong&gt; If you're afraid to ask for a commitment because you think the person will say no — that's exactly why you need to ask. A "no" now saves you months of chasing a dead lead.&lt;/p&gt;




&lt;h2&gt;
  
  
  Don't Ask for Commitment Too Early
&lt;/h2&gt;

&lt;p&gt;Here's the balance: pushing for commitment is essential, but timing matters.&lt;/p&gt;

&lt;p&gt;If you push for money or a huge commitment during what's supposed to be an early learning conversation, you'll scare people away. The first few conversations should be about &lt;strong&gt;learning&lt;/strong&gt; — understanding their problem, their workflow, their pain.&lt;/p&gt;

&lt;p&gt;Once you've validated the problem and have something to show (even a rough prototype), THEN you start pushing for commitments.&lt;/p&gt;

&lt;p&gt;The progression looks like this:&lt;/p&gt;

&lt;ol&gt;
&lt;li&gt;
&lt;strong&gt;Early conversations:&lt;/strong&gt; Learn about the problem. No pitch, no ask. Just listen.&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;Problem validated:&lt;/strong&gt; Start showing your solution concept. Ask for time commitments (follow-up meetings, beta testing).&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;Solution takes shape:&lt;/strong&gt; Push for reputation commitments (introductions, referrals).&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;Product is tangible:&lt;/strong&gt; Push for financial commitments (pre-orders, deposits, LOIs).&lt;/li&gt;
&lt;/ol&gt;

&lt;p&gt;Skipping steps or pushing too hard too early is just as bad as never pushing at all.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;Rule of Thumb:&lt;/strong&gt; Match your ask to the stage of the relationship. Early = learn. Middle = time and reputation. Late = money.&lt;/p&gt;




&lt;h2&gt;
  
  
  Key Takeaways from Chapter 5
&lt;/h2&gt;

&lt;p&gt;Let me sum up the core lessons:&lt;/p&gt;

&lt;ol&gt;
&lt;li&gt;&lt;p&gt;&lt;strong&gt;Meetings don't "go well."&lt;/strong&gt; They either produce a commitment or they fail. Stop fooling yourself with compliments.&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;&lt;strong&gt;Commitments come in three currencies:&lt;/strong&gt; Time, Reputation, and Money — in escalating order of seriousness.&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;&lt;strong&gt;Always push for a next step.&lt;/strong&gt; Know your ask before the meeting and make it before the meeting ends.&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;&lt;strong&gt;Compliments ≠ Commitments.&lt;/strong&gt; "That's a great idea" is worthless. "Here's my credit card" is priceless.&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;&lt;strong&gt;Your first customers will be "crazy."&lt;/strong&gt; They have the problem now, they know it, and they're desperate enough to use your unfinished product.&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;&lt;strong&gt;A "no" is better than a "maybe."&lt;/strong&gt; Rejection gives you clarity. Wishy-washiness wastes your time.&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;&lt;strong&gt;Match your ask to the stage.&lt;/strong&gt; Don't ask for money when you should be asking questions. Don't ask for opinions when you should be asking for money.&lt;/p&gt;&lt;/li&gt;
&lt;/ol&gt;




&lt;p&gt;&lt;em&gt;This is part of my series where I break down each chapter of The Mom Test by Rob Fitzpatrick. If you're building a product and talking to customers, this book is essential reading.&lt;/em&gt;&lt;/p&gt;

&lt;p&gt;&lt;em&gt;Previously: &lt;a href="https://dev.to/egepakten/what-i-learned-from-the-mom-test-chapter-4-why-you-should-keep-customer-conversations-casual-40b"&gt;Chapter 4 - Why You Should Keep Customer Conversations Casual&lt;/a&gt;&lt;/em&gt;&lt;br&gt;
&lt;em&gt;Next up: Chapter 6 - Finding Conversations&lt;/em&gt;&lt;/p&gt;

</description>
      <category>webdev</category>
      <category>mentorship</category>
      <category>speaking</category>
      <category>startup</category>
    </item>
    <item>
      <title>What I Learned from "The Mom Test" - Chapter 4: Why You Should Keep Customer Conversations Casual</title>
      <dc:creator>Ege Pakten</dc:creator>
      <pubDate>Sat, 24 Jan 2026 12:50:12 +0000</pubDate>
      <link>https://dev.to/egepakten/what-i-learned-from-the-mom-test-chapter-4-why-you-should-keep-customer-conversations-casual-40b</link>
      <guid>https://dev.to/egepakten/what-i-learned-from-the-mom-test-chapter-4-why-you-should-keep-customer-conversations-casual-40b</guid>
      <description>&lt;p&gt;&lt;em&gt;A developer's guide to learning from customers without the meeting overhead&lt;/em&gt;&lt;/p&gt;




&lt;p&gt;If you've been following my series on "The Mom Test" by Rob Fitzpatrick, you know we've been exploring how to have better conversations with customers. Today, I want to share what I learned from Chapter 4: &lt;strong&gt;Keeping It Casual&lt;/strong&gt;.&lt;/p&gt;

&lt;p&gt;Here's the big idea: &lt;strong&gt;formal meetings can actually hurt your customer learning.&lt;/strong&gt;&lt;/p&gt;

&lt;p&gt;Sounds counterintuitive, right? Let me explain.&lt;/p&gt;




&lt;h2&gt;
  
  
  The Problem with Formal Meetings
&lt;/h2&gt;

&lt;p&gt;We developers love structure. When we want to learn something from customers, our first instinct is often to schedule a meeting. But Rob Fitzpatrick argues this is actually a trap.&lt;/p&gt;

&lt;p&gt;Think about it. When you schedule a formal meeting:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;The customer knows you want something from them&lt;/li&gt;
&lt;li&gt;They put on their "meeting behavior"&lt;/li&gt;
&lt;li&gt;They try to be polite instead of honest&lt;/li&gt;
&lt;li&gt;Everything becomes awkward and formal&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;The author shares that Steve Blank, in his book "4 Steps to the Epiphany," recommends &lt;strong&gt;3 separate meetings&lt;/strong&gt;: one about the customer's problem, one about your solution, and one to sell the product. The idea is to avoid what Fitzpatrick calls the "Pathos Problem"—where you jump too quickly into emotional selling mode.&lt;/p&gt;

&lt;p&gt;But here's the thing: setting up 3 formal meetings is incredibly time-consuming. Once you factor in scheduling, travel, preparation, and the actual meeting, a single 1-hour meeting can cost you 4+ hours of your time.&lt;/p&gt;

&lt;p&gt;And as the author puts it: &lt;strong&gt;"The most precious resource in a startup is its founders' time."&lt;/strong&gt;&lt;/p&gt;




&lt;h2&gt;
  
  
  The Solution: Keep It Casual
&lt;/h2&gt;

&lt;p&gt;So what's the alternative? Keep your conversations casual and informal.&lt;/p&gt;

&lt;p&gt;Instead of sending a calendar invite, just have a chat. Instead of a conference room, talk at a coffee shop. Instead of an interview, have a conversation.&lt;/p&gt;

&lt;p&gt;Here's a great example from the book. Imagine you're at a conference and you bump into someone in your target industry. Instead of saying:&lt;/p&gt;

&lt;blockquote&gt;
&lt;p&gt;"Can I schedule a meeting to interview you about your problems?"&lt;/p&gt;
&lt;/blockquote&gt;

&lt;p&gt;You say:&lt;/p&gt;

&lt;blockquote&gt;
&lt;p&gt;"Hey, I'm curious—how did you end up getting this gig?"&lt;/p&gt;
&lt;/blockquote&gt;

&lt;p&gt;See the difference? One feels like work. The other feels like genuine human curiosity.&lt;/p&gt;

&lt;p&gt;When you &lt;strong&gt;strip all the formality&lt;/strong&gt; from the process, something magical happens:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;No meetings to schedule&lt;/li&gt;
&lt;li&gt;No "interviews" to conduct&lt;/li&gt;
&lt;li&gt;Conversations become fast and lightweight&lt;/li&gt;
&lt;li&gt;You can talk to a dozen people at a single industry meetup&lt;/li&gt;
&lt;/ul&gt;




&lt;h2&gt;
  
  
  The Meeting Anti-Pattern
&lt;/h2&gt;

&lt;p&gt;The author introduces a concept called "The Meeting Anti-Pattern." This is the &lt;strong&gt;tendency to relegate&lt;/strong&gt; every customer conversation into a calendar block.&lt;/p&gt;

&lt;p&gt;Beyond being a bad use of time, our &lt;strong&gt;over-reliance&lt;/strong&gt; on formal meetings makes us miss &lt;strong&gt;serendipitous&lt;/strong&gt; learning opportunities.&lt;/p&gt;

&lt;p&gt;Here's a funny example from the book: Imagine you're at a café, and your dream customer sits next to you. Instead of just starting a natural conversation, you &lt;strong&gt;psych yourself up&lt;/strong&gt; and then &lt;strong&gt;fumble&lt;/strong&gt; through an awkward pitch asking if maybe they want to get coffee sometime... at a different place.&lt;/p&gt;

&lt;p&gt;That's ridiculous, right? You're already having coffee together! Just talk to them like a normal human being.&lt;/p&gt;




&lt;h2&gt;
  
  
  How Formal Is Too Formal?
&lt;/h2&gt;

&lt;p&gt;Here are some warning signs that your conversation is too formal:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;"So, first off, thanks for agreeing to this interview..."&lt;/li&gt;
&lt;li&gt;"On a scale of 1 to 5, how much would you say..."&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;When you start with phrases like these, you immediately put the other person in "interview mode." They feel like they're doing you a favor, which creates an &lt;strong&gt;ominous&lt;/strong&gt; atmosphere where they'll say whatever they think you want to hear.&lt;/p&gt;

&lt;p&gt;Learning from customers doesn't mean wearing a suit and sipping boardroom coffee. The right questions are fast, interesting, and touch on topics people genuinely enjoy discussing.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;Rule of thumb from the book: If it feels like they're doing you a favor by talking to you, it's probably too formal.&lt;/strong&gt;&lt;/p&gt;

&lt;p&gt;At their best, these conversations are a pleasure for both parties. You're probably the first person in a long time to be truly interested in the &lt;strong&gt;petty&lt;/strong&gt; annoyances of their daily work.&lt;/p&gt;




&lt;h2&gt;
  
  
  How Long Should These Conversations Be?
&lt;/h2&gt;

&lt;p&gt;This surprised me: early conversations can be incredibly short.&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;
&lt;strong&gt;5 minutes&lt;/strong&gt; is enough to learn whether a problem exists&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;10-15 minutes&lt;/strong&gt; gets you into their workflow, time usage, and what they've tried before&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;30+ minutes&lt;/strong&gt; happens naturally when you hit a topic they love (people enjoy talking about themselves!)&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;The longer conversations are easier to &lt;strong&gt;facilitate&lt;/strong&gt; because once someone starts explaining their work, they can go into a monologue. You just need to point them in the right direction.&lt;/p&gt;

&lt;p&gt;But here's the catch with formal B2B meetings: the duration is often determined by the arbitrary calendar block (usually 30 minutes or 1 hour), not by what you actually need to learn. You might lose:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;5 minutes to &lt;strong&gt;miscellaneous tardiness&lt;/strong&gt;
&lt;/li&gt;
&lt;li&gt;5 minutes to saying hello and small talk&lt;/li&gt;
&lt;li&gt;10 minutes to product demos&lt;/li&gt;
&lt;li&gt;5 minutes to figuring out next steps&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;That's half your meeting gone before you even start learning!&lt;/p&gt;




&lt;h2&gt;
  
  
  Putting It Together: A Real Example
&lt;/h2&gt;

&lt;p&gt;The author shares a great story about trying to get feedback from busy investors who manage their &lt;strong&gt;dealflow&lt;/strong&gt; with hundreds of meetings per month.&lt;/p&gt;

&lt;p&gt;Instead of trying to schedule yet another meeting in their packed calendars, he showed up to an industry meeting and, during casual small talk, mentioned: "Our analysts kill most of them before they ever reach us."&lt;/p&gt;

&lt;p&gt;That was it. One sentence during small talk. The investor responded with a bunch more details about how they deal with applications, pointed at some sticky notes on the wall, and shared exactly how their process worked.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;It took 5 minutes.&lt;/strong&gt; No formal meeting. No biases. No compliments. Just concrete facts.&lt;/p&gt;

&lt;p&gt;Now compare that to someone who takes a 2-hour &lt;strong&gt;commute&lt;/strong&gt; to attend a formal meeting. They might get the same information (or worse), but at a much higher cost.&lt;/p&gt;

&lt;p&gt;The lesson? Sometimes a casual 5-minute chat is worth more than an hour-long formal meeting.&lt;/p&gt;




&lt;h2&gt;
  
  
  The Visionary Leap
&lt;/h2&gt;

&lt;p&gt;Here's the beautiful part: once you've collected all these casual insights, you can take what the author calls the "visionary &lt;strong&gt;leap&lt;/strong&gt;."&lt;/p&gt;

&lt;p&gt;You take everything you've learned—all the problems, all the frustrations, all the workarounds—and you come up with a specific offering that makes your customers' lives better. Then you ask them to commit to it.&lt;/p&gt;

&lt;p&gt;But you can only make this leap if you've been listening properly. And proper listening happens in casual, honest conversations—not in formal interviews where people tell you what you want to hear.&lt;/p&gt;




&lt;h2&gt;
  
  
  Key Takeaways
&lt;/h2&gt;

&lt;p&gt;Let me summarize the main lessons from this chapter:&lt;/p&gt;

&lt;ol&gt;
&lt;li&gt;&lt;p&gt;&lt;strong&gt;Formal meetings create bias.&lt;/strong&gt; People behave differently when they know they're being "interviewed."&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;&lt;strong&gt;Casual conversations are faster and more honest.&lt;/strong&gt; You can learn in 5 minutes what might take an hour in a formal setting.&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;&lt;strong&gt;The first conversation doesn't need to be a meeting.&lt;/strong&gt; It works better as a casual chat.&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;&lt;strong&gt;Your time is precious.&lt;/strong&gt; Don't waste 4 hours on meeting overhead when a 10-minute conversation would give you the same insights.&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;&lt;strong&gt;Serendipity is your friend.&lt;/strong&gt; Be open to learning opportunities everywhere—conferences, coffee shops, random encounters.&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;&lt;strong&gt;If they feel like they're doing you a favor, it's too formal.&lt;/strong&gt; The best conversations are enjoyable for both parties.&lt;/p&gt;&lt;/li&gt;
&lt;/ol&gt;




&lt;h2&gt;
  
  
  Rule of Thumb
&lt;/h2&gt;

&lt;p&gt;I'll leave you with the book's main rule for this chapter:&lt;/p&gt;

&lt;blockquote&gt;
&lt;p&gt;&lt;strong&gt;"Learning about a customer and their problems works better as a quick and casual chat than a long, formal meeting."&lt;/strong&gt;&lt;/p&gt;
&lt;/blockquote&gt;

&lt;p&gt;And one more:&lt;/p&gt;

&lt;blockquote&gt;
&lt;p&gt;&lt;strong&gt;"Give as little information as possible about your idea while still nudging the discussion in a useful direction."&lt;/strong&gt;&lt;/p&gt;
&lt;/blockquote&gt;




&lt;p&gt;That's it for Chapter 4! The message is clear: stop hiding behind formal meetings. Get out there, talk to people like a normal human being, and you'll learn so much more.&lt;/p&gt;

&lt;p&gt;In the next chapter, we'll explore more techniques for having effective customer conversations. Stay tuned!&lt;/p&gt;




&lt;p&gt;&lt;em&gt;This post is part of my series on "The Mom Test" by Rob Fitzpatrick. I'm a developer learning product management and customer validation, sharing my learnings with fellow developers who want to build products that people actually want. All examples and stories are credited to the original author.&lt;/em&gt;&lt;/p&gt;

&lt;p&gt;&lt;em&gt;What do you think? Have you experienced the difference between casual and formal customer conversations? Let me know in the comments!&lt;/em&gt;&lt;/p&gt;

</description>
      <category>webdev</category>
      <category>startup</category>
      <category>speaking</category>
      <category>learning</category>
    </item>
  </channel>
</rss>
