<?xml version="1.0" encoding="UTF-8"?>
<rss version="2.0" xmlns:atom="http://www.w3.org/2005/Atom" xmlns:dc="http://purl.org/dc/elements/1.1/">
  <channel>
    <title>DEV Community: Maureen Mukami</title>
    <description>The latest articles on DEV Community by Maureen Mukami (@maureen_mukami_4268d10eac).</description>
    <link>https://dev.to/maureen_mukami_4268d10eac</link>
    <image>
      <url>https://media2.dev.to/dynamic/image/width=90,height=90,fit=cover,gravity=auto,format=auto/https:%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Fuser%2Fprofile_image%2F3252374%2Fcf28f28a-cb32-462a-a923-746b5237d175.PNG</url>
      <title>DEV Community: Maureen Mukami</title>
      <link>https://dev.to/maureen_mukami_4268d10eac</link>
    </image>
    <atom:link rel="self" type="application/rss+xml" href="https://dev.to/feed/maureen_mukami_4268d10eac"/>
    <language>en</language>
    <item>
      <title>RAGs for Dummies: The Game-Changing Power of RAG</title>
      <dc:creator>Maureen Mukami</dc:creator>
      <pubDate>Sun, 21 Sep 2025 17:01:06 +0000</pubDate>
      <link>https://dev.to/maureen_mukami_4268d10eac/rags-for-dummies-the-game-changing-power-of-rag-2j86</link>
      <guid>https://dev.to/maureen_mukami_4268d10eac/rags-for-dummies-the-game-changing-power-of-rag-2j86</guid>
      <description>&lt;p&gt;&lt;strong&gt;What is RAG?&lt;/strong&gt;&lt;br&gt;
RAG stands for Retrieval-Augmented Generation. It’s a powerful framework that combines large language models (LLMs)  such as GPT-4, LLaMA, or Mistral with external knowledge sources like company documents, databases, or the web. If we think of LLM as a smart student with a huge memory, then RAG is like giving that student access to a library full of the latest books and notes. This way, they don’t just rely on what they memorized years ago; they can also look things up before answering.&lt;br&gt;
LLMs are great, but they come with limitations:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;Limited knowledge as they only know what they were trained on, so recent updates or niche details might be missing.&lt;/li&gt;
&lt;li&gt;Hallucinations where by sometimes they confidently make things up, even if they sound convincing.&lt;/li&gt;
&lt;li&gt;Generic answers in that without specific context, they may respond vaguely.&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;RAG solves these problems by letting the model fetch real information first, then generate an informed answer.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;Steps for setting up RAGs framework&lt;/strong&gt;&lt;br&gt;
&lt;strong&gt;&lt;em&gt;1. Data Collection&lt;/em&gt;&lt;/strong&gt;&lt;br&gt;
This is gathering the knowledge your system will rely on, such as company policies, SOPs, HR manuals, FAQs, product guides, business reports, or even external sources like research papers and regulations. The data can come in different formats including PDFs, Word docs, spreadsheets, databases, or web pages. It’s important to collect only relevant information, clean and update it regularly, organize it by category, and secure sensitive files. For example, a phone company might collect manuals, warranty documents, and troubleshooting guides to help a RAG bot provide accurate customer support.&lt;br&gt;
&lt;strong&gt;&lt;em&gt;2. Data Chunking&lt;/em&gt;&lt;/strong&gt;&lt;br&gt;
This process entails splitting large documents into smaller, manageable sections (e.g., 500–1000 words) so the system can retrieve only the most relevant parts. To avoid cutting off important context mid-sentence or paragraph, chunks are often created with a slight overlap (like 50–100 words) between them. This overlap ensures smoother continuity, prevents loss of meaning, and improves the quality of answers generated by the LLM.&lt;br&gt;
&lt;strong&gt;&lt;em&gt;3. Document Embeddings&lt;/em&gt;&lt;/strong&gt;&lt;br&gt;
It involves converting text chunks into numerical vectors that capture their meaning, making them searchable by similarity rather than exact words. These embeddings are stored in a vector database (like FAISS, Pinecone, or Chroma), which allows the retriever to quickly find the most relevant chunks when a user asks a question. When generating embeddings, you can pass kwargs (keyword arguments) such as model name, batch size, or normalization options to fine-tune how the embeddings are created. This step is crucial because high-quality embeddings directly determine how accurately your RAG system retrieves and ranks information.&lt;br&gt;
&lt;strong&gt;&lt;em&gt;4. User Queries&lt;/em&gt;&lt;/strong&gt;&lt;br&gt;
The user queries are transformed into embeddings so they can be compared with document embeddings in the vector database. The retriever then selects the most relevant chunks based on similarity, often using parameters like k to control how many results are returned. These chunks are combined with the query and sent to the LLM, which generates a context-aware answer.&lt;br&gt;
&lt;strong&gt;&lt;em&gt;5. Generate a Response&lt;/em&gt;&lt;/strong&gt;&lt;br&gt;
The query and the selected chunks are passed into the LLM, which uses both to craft an informed and accurate answer. You can fine-tune this step with parameters like temperature (controls creativity), max_tokens (limits response length), and other kwargs for custom behavior. This ensures the final response is not only factually grounded but also clear, coherent, and aligned with the user’s needs.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;Real-World Power of RAG&lt;/strong&gt;&lt;br&gt;
i. Customer Support&lt;br&gt;
RAGs answer customer questions with precision.&lt;br&gt;
For example if a customer asks, “What’s the warranty on Model X200?” → RAG fetches the warranty policy and answers: “The X200 has a 2-year warranty covering manufacturing defects.”&lt;br&gt;
ii. Market Research&lt;br&gt;
RAGs summarize customer reviews, social media, or industry reports.&lt;br&gt;
For example a question of  “What are customers saying about our new app update?” → RAG analyzes feedback and gives a sentiment breakdown.&lt;br&gt;
iii. Content Generation&lt;br&gt;
RAGs automatically create product descriptions, wikis, or reports.&lt;br&gt;
An  example is when RAG generates a sales report by pulling the latest figures from company databases.&lt;br&gt;
iv. Data Analysis &amp;amp; Business Intelligence&lt;br&gt;
RAGs extract insights from huge datasets.&lt;br&gt;
An example is a question “What were the top 5 reasons for customer complaints last quarter?” → RAG scans logs and summarizes findings.&lt;br&gt;
v. Knowledge Management&lt;br&gt;
RAGs make company policies and procedures easy to access.&lt;br&gt;
For example new employees can ask, “What’s the leave policy?” and instantly get the official HR response.&lt;/p&gt;

&lt;p&gt;In conclusion, RAG is a game changer as it: Keeps responses accurate and current, reduces hallucinations from LLMs, makes AI useful for real-world business problems, and is flexible in that it works with cloud APIs (like GPT-4, Claude) or local open-source models (like Mistral, LLaMA).&lt;br&gt;
At its core, RAG = LLM + Your Data = Smart, Reliable Assistant.&lt;br&gt;
It gives AI the brains of a language model plus the memory of a search engine, making it one of the most powerful tools for businesses, researchers, and everyday users alike.&lt;/p&gt;

</description>
      <category>rag</category>
      <category>vectordatabase</category>
      <category>datascience</category>
    </item>
    <item>
      <title>Clustering as a Method of Unveiling Hidden Patterns in Data</title>
      <dc:creator>Maureen Mukami</dc:creator>
      <pubDate>Fri, 05 Sep 2025 13:04:29 +0000</pubDate>
      <link>https://dev.to/maureen_mukami_4268d10eac/clustering-as-a-method-of-unveiling-hidden-patterns-in-data-3c09</link>
      <guid>https://dev.to/maureen_mukami_4268d10eac/clustering-as-a-method-of-unveiling-hidden-patterns-in-data-3c09</guid>
      <description>&lt;p&gt;Unsupervised learning is a type of machine learning that deals with unlabeled data. While supervised learning relies on labeled data to make predictions, unsupervised learning works with data that has no predefined labels or outputs. This makes it particularly powerful for uncovering hidden patterns, relationships, and structures in data without human intervention. Unsupervised learning algorithms do not rely on direct input-to-output mappings. Instead, they autonomously explore data to find meaningful organization. Over the years, these algorithms have become increasingly efficient in discovering the underlying structures of complex, unlabeled datasets.&lt;br&gt;
&lt;em&gt;How Does Unsupervised Learning Work?&lt;/em&gt;&lt;br&gt;
Unsupervised learning works by:&lt;br&gt;
-Analyzing unlabeled data to identify similarities, differences, and relationships.&lt;br&gt;
-Grouping or transforming data into structures that highlight hidden patterns.&lt;br&gt;
-Providing insights that may not be obvious through human observation.&lt;br&gt;
&lt;strong&gt;&lt;em&gt;Main Models in Unsupervised Learning&lt;/em&gt;&lt;/strong&gt;&lt;br&gt;
There are three primary methods in unsupervised learning:&lt;br&gt;
&lt;em&gt;Clustering&lt;/em&gt;&lt;br&gt;
Involves grouping untagged data based on similarities and differences.&lt;br&gt;
Items in the same group (cluster) share common properties, while items in different groups are dissimilar.&lt;br&gt;
&lt;em&gt;Association Rules&lt;/em&gt;&lt;br&gt;
A rule-based approach for discovering interesting relationships between features in a dataset.&lt;br&gt;
Uses statistical measures (support, confidence, lift) to identify strong associations.&lt;br&gt;
&lt;em&gt;Dimensionality Reduction&lt;/em&gt;&lt;br&gt;
Transforms data from high-dimensional spaces into low-dimensional spaces without losing important information. Common techniques: PCA (Principal Component Analysis), t-SNE. Useful for visualization, noise reduction, and computational efficiency.&lt;br&gt;
&lt;strong&gt;Clustering&lt;/strong&gt;&lt;br&gt;
Clustering is the most widely applied technique in unsupervised learning.&lt;br&gt;
Clustering is the process of organizing data into groups so that objects within the same group (cluster) are more similar to each other than to those in other groups. It is often used to reveal natural structures within datasets where no prior labels exist.&lt;br&gt;
Clustering answers the question:&lt;br&gt;
&lt;em&gt;Which data points naturally belong together?”&lt;/em&gt;&lt;br&gt;
Common Clustering Algorithms include:&lt;br&gt;
&lt;strong&gt;&lt;em&gt;a.K-Means Clustering&lt;/em&gt;&lt;/strong&gt;&lt;br&gt;
K- Means clustering divides data into k clusters, where k is predefined. Each cluster is represented by a centroid, and data points are assigned to the nearest centroid. It is an iterative algorithm that creates non-overlapping clusters meaning each instance in your dataset can only belong to one cluster exclusively&lt;br&gt;
Pros: Efficient on large datasets.&lt;br&gt;
Cons: Requires pre-selecting the number of clusters (k) and is sensitive to outliers.&lt;br&gt;
&lt;strong&gt;&lt;em&gt;b.Hierarchical Clustering&lt;/em&gt;&lt;/strong&gt;&lt;br&gt;
Hierarchical clustering is a method of clustering that builds a hierarchy of clusters which are visualized through a dendrogram. There are two types of this method:&lt;br&gt;
i.Agglomerative (Bottom-Up): Starts with each point as its own cluster and merges them step by step.&lt;br&gt;
ii.Divisive (Top-Down): Starts with one cluster and splits it recursively.&lt;br&gt;
Pros: No need to specify the number of clusters beforehand.&lt;br&gt;
Cons: Computationally expensive for large datasets.&lt;br&gt;
&lt;strong&gt;&lt;em&gt;c.DBSCAN (Density-Based Spatial Clustering of Applications with Noise)&lt;/em&gt;&lt;/strong&gt;&lt;br&gt;
DBSCAN groups points that are closely packed together and labels isolated points as noise. Unlike K-Means, the number of clusters does not need to be specified.&lt;br&gt;
Pros: Can detect arbitrarily shaped clusters and handle outliers.&lt;br&gt;
Cons: Performance drops with clusters of varying densities.&lt;br&gt;
&lt;strong&gt;&lt;em&gt;d.Mean Shift Clustering&lt;/em&gt;&lt;/strong&gt;&lt;br&gt;
Mean Shift clustering identifies clusters by shifting data points toward areas of higher density.&lt;br&gt;
Automatically determines the number of clusters based on data distribution.&lt;br&gt;
Pros: No need for k value.&lt;br&gt;
Cons: Can be computationally heavy.&lt;br&gt;
&lt;strong&gt;&lt;em&gt;Applications of Clustering&lt;/em&gt;&lt;/strong&gt;&lt;br&gt;
Customer Segmentation – grouping customers based on behavior for personalized marketing.&lt;br&gt;
Anomaly Detection – spotting unusual data points such as fraudulent transactions.&lt;br&gt;
Document/Text Clustering – organizing news articles, research papers, or emails into categories.&lt;br&gt;
Image Segmentation – dividing images into regions for medical or computer vision tasks.&lt;br&gt;
Recommendation Systems – grouping users with similar preferences for product or content suggestions.&lt;/p&gt;

&lt;p&gt;My key insights is that, unlike in supervised learning, where models rely on labeled data and clear input-output mappings, unsupervised learning and clustering in particular thrives in situations where labels are absent. To me, this is its greatest strength because most real-world data is unstructured and unlabeled. I find clustering especially valuable because it reveals hidden structures without requiring human guidance. For example, while supervised models can classify emails as spam or not spam, clustering can go further by discovering new, previously unseen patterns in user behavior or anomalies that no one thought to label. That said, I also recognize the challenges. Unlike supervised learning, the results of clustering are not always straightforward to evaluate. Determining the right number of clusters, the right algorithm, or even whether the clusters are meaningful can sometimes feel subjective. In my view, this makes clustering as much an art as it is a science.&lt;/p&gt;

</description>
      <category>machinelearning</category>
      <category>datascience</category>
    </item>
    <item>
      <title>A Simple Guide to Classification in Machine Learning</title>
      <dc:creator>Maureen Mukami</dc:creator>
      <pubDate>Mon, 25 Aug 2025 09:36:05 +0000</pubDate>
      <link>https://dev.to/maureen_mukami_4268d10eac/a-simple-guide-to-classification-in-machine-learning-37aj</link>
      <guid>https://dev.to/maureen_mukami_4268d10eac/a-simple-guide-to-classification-in-machine-learning-37aj</guid>
      <description>&lt;p&gt;Supervised learning is a machine learning approach where models are trained with labeled data. The data is split into training sets which  teach the model and testing sets which check accuracy of the model. The goal is for predictions to match the true outcomes as closely as possible.&lt;br&gt;
&lt;strong&gt;&lt;em&gt;How Classification Works&lt;/em&gt;&lt;/strong&gt;&lt;br&gt;
Classification works by teaching a model how to recognize patterns using examples that already have answers. Classification deals with discrete data, assigning items to specific categories. It can be into three types: &lt;strong&gt;Binary classification&lt;/strong&gt; where data belongs to one of two classes, such as “pass or fail” or “true or false.” &lt;strong&gt;Multiclass classification&lt;/strong&gt; where data can fall into one class among many, for example identifying different species of plants. &lt;strong&gt;Multilabel classification&lt;/strong&gt; where data can belong to multiple classes at once, such as classifying a movie as both “comedy” and “romance” (romcom).&lt;br&gt;
Different models which are different algorithms or techniques that are applied to sort data into categories (classes) include:&lt;br&gt;
&lt;strong&gt;Logistic Regression&lt;/strong&gt; It predicts probabilities using a sigmoid function.&lt;br&gt;
&lt;strong&gt;Support Vector Machines(SVMs)&lt;/strong&gt; which find the best boundary between classes.&lt;br&gt;
&lt;strong&gt;Decision Trees&lt;/strong&gt; which split data into branches for decisions.&lt;br&gt;
&lt;strong&gt;Random Forests&lt;/strong&gt; which combine multiple trees for better accuracy.&lt;br&gt;
&lt;strong&gt;Naive Bayes&lt;/strong&gt; which uses probability with independence assumptions.&lt;br&gt;
&lt;strong&gt;KNN&lt;/strong&gt; which classifies based on nearest neighbors.&lt;br&gt;
&lt;strong&gt;Neural Networks&lt;/strong&gt; which use layers of neurons to learn complex patterns.&lt;/p&gt;

&lt;p&gt;From my perspective, classification makes handling large datasets easier by grouping and labeling information in meaningful ways. This allows people to analyze patterns more effectively and make smart, data-driven decisions. What excites me most is the impact of classification in real life. For example, advanced systems can help detect diseases early in healthcare, or spot fraudulent transactions in banking, which protects people’s money. Classification also powers personalization—recommendation systems suggest movies, music, and products that match individual preferences, creating unique user experiences. I also see classification as an assistant for decision-making. Instead of replacing human judgment, it supports people by providing fast, evidence-based insights. That combination of speed and accuracy is transformative across industries.&lt;/p&gt;

&lt;p&gt;While I’ve struggled to fully grasp some algorithms at first, constant practice with Python has made concepts clearer. I’m also learning that success is not only about knowing algorithms but also about choosing the right one for the right problem.&lt;/p&gt;

</description>
    </item>
    <item>
      <title>Lupus and the Trade-off between Type I and Type II Errors in Diagnosis</title>
      <dc:creator>Maureen Mukami</dc:creator>
      <pubDate>Wed, 13 Aug 2025 10:15:13 +0000</pubDate>
      <link>https://dev.to/maureen_mukami_4268d10eac/lupus-and-the-trade-off-between-type-i-and-type-ii-errors-in-diagnosis-4n7f</link>
      <guid>https://dev.to/maureen_mukami_4268d10eac/lupus-and-the-trade-off-between-type-i-and-type-ii-errors-in-diagnosis-4n7f</guid>
      <description>&lt;p&gt;Lupus, also known as systemic lupus erythematosus (SLE), is a chronic autoimmune disease in which the immune system mistakenly attacks healthy tissues and organs. The exact cause is unknown, but potential triggers include genetics, hormones (it is more common in women of childbearing age), and environmental factors such as sunlight, infections, stress, and certain medications. While there is no cure, treatment can control symptoms and prevent organ damage. &lt;br&gt;
In hypothesis testing, the process of diagnosing lupus can be expressed as:&lt;br&gt;
  • &lt;strong&gt;&lt;em&gt;H₀ (Null hypothesis):&lt;/em&gt;&lt;/strong&gt; The patient does not have lupus.&lt;br&gt;
  • &lt;strong&gt;&lt;em&gt;H₁ (Alternative hypothesis)&lt;/em&gt;&lt;/strong&gt;: The patient has lupus.&lt;br&gt;
In medical diagnosis, two types of errors can occur:&lt;br&gt;
&lt;strong&gt;&lt;em&gt;Type I error (false positive)&lt;/em&gt;&lt;/strong&gt; happens when someone is diagnosed with lupus even though they don’t actually have it. This can lead to unnecessary treatment, which may harm the body. Medicines used for lupus are often strong, so they can overwork the liver as it processes them, increasing the risk of liver damage. The kidneys can also be strained from filtering drugs that aren’t needed. Strong medications like steroids may weaken the immune system, making infections more likely, and long-term use can cause bone loss, weight gain, high blood pressure, or diabetes. Other drugs may damage the eyes (retinopathy), and anti-inflammatory medicines can cause stomach ulcers or bleeding.&lt;br&gt;
&lt;strong&gt;&lt;em&gt;Type II error (false negative)&lt;/em&gt;&lt;/strong&gt; is when someone actually has lupus but is told they don’t. Without proper treatment, the disease keeps progressing and can cause serious organ damage. Kidneys may develop lupus nephritis, leading to kidney failure. The heart can become inflamed or develop a higher risk of heart attack. The lungs may be damaged or inflamed, causing breathing problems. Lupus can also affect the brain and nervous system, leading to seizures, strokes, memory loss, or mood changes. It may damage the blood and bone marrow, causing anemia or clotting problems. On the outside, it can cause ongoing skin rashes or ulcers, and in the joints, it can lead to painful, chronic arthritis.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;Trade-off between Type I and Type II Errors in Lupus Diagnosis&lt;/strong&gt;&lt;br&gt;
In medicine, doctors must balance the risk of over-diagnosing (Type I error) with the risk of missing the diagnosis (Type II error). While Type I errors can result in unpleasant side effects, financial costs, and temporary health risks from unnecessary treatment, they are often reversible. A patient can return for follow-up visits, seek a second opinion, and stop treatment if the misdiagnosis is identified. Type II errors, however, can be far more dangerous in the case of lupus. Because the disease may progress silently, patients might feel healthy while lupus continues to damage vital organs such as the kidneys, heart, lungs, and brain. By the time the correct diagnosis is made, the damage could be irreversible, leading to lifelong disability or even death. For this reason, in diseases like lupus where early treatment is critical, it is generally safer to minimize Type II errors even if that means accepting a slightly higher risk of Type I errors. In other words, it is often better to risk over-treating a healthy patient than to miss lupus in someone who truly has it.&lt;/p&gt;

</description>
    </item>
    <item>
      <title>Who Has the Upper Hand? Data-Driven Prediction for the 2025/2026 EPL Title Race</title>
      <dc:creator>Maureen Mukami</dc:creator>
      <pubDate>Thu, 31 Jul 2025 11:15:29 +0000</pubDate>
      <link>https://dev.to/maureen_mukami_4268d10eac/who-has-the-upper-hand-data-driven-prediction-for-the-20252026-epl-title-race-134c</link>
      <guid>https://dev.to/maureen_mukami_4268d10eac/who-has-the-upper-hand-data-driven-prediction-for-the-20252026-epl-title-race-134c</guid>
      <description>&lt;p&gt;Predicting the winner of the English Premier League (EPL) is always a complex challenge, given the intense competition and frequent surprises that define the league. Yet, by applying a data-driven approach to team performance across the last two seasons, a strong case can be made for Liverpool FC as the top contender for the 2025/2026 title.&lt;br&gt;
This prediction is based on an analysis of match outcomes from the 2023/2024 and 2024/2025 seasons, specifically looking at wins, draws, and losses for each team. Using the standard EPL point system which is 3 points for a win, 1 point for a draw and 0 for a loss. I calculated the total points each team accumulated. I then determined each team's performance efficiency by dividing their total earned points by the maximum possible points (calculated as matches played × 3 × total number of teams). This yielded a probability score reflecting each team's ability to consistently secure points.&lt;br&gt;
According to this analysis, Liverpool FC emerged with a probability score of 0.036404 being the highest among all EPL teams over the two-year period. While the figure may appear modest on its own, its value lies in how it compares with the rest of the league. Liverpool's score positions them at the top of the performance index, highlighting their consistent form and competitive edge.&lt;br&gt;
What does this mean going into the new season? Simply put, Liverpool has shown the strongest efficiency in turning matches into points across two full seasons. Their consistency, combined with tactical discipline, squad depth, and a winning mindset, makes them statistically the most likely team to lift the 2025/2026 trophy. In a sport where emotions, media narratives, and unpredictable transfers often influence expectations, a data-led forecast offers a refreshing, objective lens. As the EPL gears up for another thrilling season, the numbers suggest one thing clearly: Liverpool FC is the team to watch.&lt;/p&gt;

</description>
    </item>
    <item>
      <title>Measures of central tendencies</title>
      <dc:creator>Maureen Mukami</dc:creator>
      <pubDate>Thu, 24 Jul 2025 14:30:06 +0000</pubDate>
      <link>https://dev.to/maureen_mukami_4268d10eac/measures-of-central-tendencies-41g7</link>
      <guid>https://dev.to/maureen_mukami_4268d10eac/measures-of-central-tendencies-41g7</guid>
      <description>&lt;p&gt;In the world of data analysis, making sense of large volumes of information is crucial. One of the foundational concepts that enable this is measures of central tendency. These are statistical tools used to describe the center point or typical value of a dataset, helping analysts and data scientists summarize data in a meaningful way. The three most common measures are the mean, median, and mode each serving a unique purpose depending on the data context.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;&lt;em&gt;The Mean: The Arithmetic Average&lt;/em&gt;&lt;/strong&gt;&lt;br&gt;
The mean often referred to as the average is calculated by summing all values in a dataset and dividing by the total number of values. It is widely used due to its simplicity and interpretability. However, one of its limitations is its sensitivity to extreme values, also known as outliers. In skewed datasets, even a single unusually high or low value can distort the mean, making it less representative of the overall data distribution.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;&lt;em&gt;The Median: The Middle Ground&lt;/em&gt;&lt;/strong&gt;&lt;br&gt;
The median represents the middle value in a sorted dataset. If the number of observations is even, the median is the average of the two central values. One of the key advantages of the median is its resistance to outliers. It offers a better sense of central location in skewed datasets.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;The Mode: The Most Frequent Value&lt;/strong&gt;&lt;br&gt;
The mode is the value that appears most frequently in a dataset. Unlike the mean and median, the mode is particularly useful for categorical data, where understanding the most common category is important. A dataset can be unimodal, bimodal, multimodal, or even have no mode if all values are unique.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;&lt;em&gt;Why Measures of Central Tendency Matter in Data Science&lt;/em&gt;&lt;/strong&gt;&lt;br&gt;
In data science, understanding the central tendency of a dataset is more than just a basic statistical exercise. It has broad applications across different stages of data analysis and model development.&lt;/p&gt;

&lt;ol&gt;
&lt;li&gt;&lt;p&gt;Data Summarization and Exploration&lt;br&gt;
During exploratory data analysis (EDA), central tendency offers a quick overview of the data, allowing analysts to identify trends and patterns without diving into every individual data point.&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;Understanding Data Distributions&lt;br&gt;
The relationship between the mean, median, and mode provides insights into the shape of the distribution. In normally distributed data, all three measures are close. In skewed data, significant differences can highlight asymmetry, helping determine the appropriate statistical techniques or transformations.&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;Outlier Detection&lt;br&gt;
A large gap between the mean and median can suggest the presence of outliers—unusual data points that may impact analysis or model accuracy.&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;Feature Engineering and Preprocessing&lt;br&gt;
These measures are often used to fill in missing values or create derived features. They also guide data transformation decisions, especially when preparing data for algorithms that assume normality.&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;Communication and Reporting&lt;br&gt;
Finally, central tendency measures make it easier to communicate findings to stakeholders. Saying “the average customer spends $50” is more impactful and digestible than listing all individual transactions.&lt;/p&gt;&lt;/li&gt;
&lt;/ol&gt;

</description>
    </item>
    <item>
      <title>Relationships in Power BI</title>
      <dc:creator>Maureen Mukami</dc:creator>
      <pubDate>Sun, 22 Jun 2025 06:13:43 +0000</pubDate>
      <link>https://dev.to/maureen_mukami_4268d10eac/relationships-in-power-bi-44a4</link>
      <guid>https://dev.to/maureen_mukami_4268d10eac/relationships-in-power-bi-44a4</guid>
      <description>&lt;p&gt;Power BI is a powerful tool for data analysis and visualization, but its real strength lies in how it connects different tables through relationships. These relationships allow you to combine data, filter across tables, and create interactive reports with ease.&lt;/p&gt;

&lt;p&gt;&lt;em&gt;📘 What Are Relationships&lt;/em&gt;&lt;br&gt;
In Power BI, relationships refer to the logical connections between two or more tables based on a common column. These connections enable Power BI to combine data from multiple tables and perform calculations across them, just like joins in traditional databases.&lt;br&gt;
Relationships determine how data from different sources interacts by ensuring that slicers, filters, and visuals all work harmoniously when pulling from multiple tables.&lt;/p&gt;

&lt;p&gt;&lt;em&gt;🧩 How Relationships Are Categorized&lt;/em&gt;&lt;br&gt;
In Power BI, every relationship between tables is defined by two key properties:&lt;/p&gt;

&lt;ol&gt;
&lt;li&gt;&lt;p&gt;Cardinality&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;Cross-filter direction&lt;/p&gt;&lt;/li&gt;
&lt;/ol&gt;

&lt;p&gt;&lt;em&gt;1️⃣ Cardinality&lt;/em&gt;&lt;br&gt;
Cardinality refers to the nature of the relationship between two tables specifically, how many unique values in one column relate to values in another column. Power BI supports four types of cardinality:&lt;/p&gt;

&lt;p&gt;a. One to Many (1:*)&lt;br&gt;
This is the most common relationship type. One record in the first table is related to multiple records in the second table.&lt;br&gt;
Example: One product in the Products table can appear in many rows in the Sales table.&lt;/p&gt;

&lt;p&gt;b. Many to One (*:1)&lt;br&gt;
This is essentially the reverse of one-to-many. Multiple records in the first table relate to one record in the second table.&lt;br&gt;
Example: Many sales entries can point back to one customer in the Customers table.&lt;/p&gt;

&lt;p&gt;c. One to One (1:1)&lt;br&gt;
Each record in the first table corresponds to exactly one matching record in the second table, and vice versa.&lt;br&gt;
Example: Each employee in an Employees table has one matching record in a Payroll table.&lt;/p&gt;

&lt;p&gt;d. Many to Many (&lt;em&gt;:&lt;/em&gt;)&lt;br&gt;
This relationship exists when multiple values in one table relate to multiple values in another table often used for complex models.&lt;br&gt;
Example: A student can enroll in many courses, and each course can have many students.&lt;/p&gt;

&lt;p&gt;&lt;em&gt;2️⃣ Cross-filter Direction&lt;/em&gt;&lt;br&gt;
Cross-filter direction defines how filters flow between the related tables when interacting with visuals.&lt;/p&gt;

&lt;p&gt;a. Single Direction&lt;br&gt;
Filters flow from one table to the other only. Recommended for simpler models for better performance.&lt;br&gt;
Example: Filters flow from the Date table to the Sales table, but not the other way around.&lt;br&gt;
b. Both Directions (Bi-directional)&lt;br&gt;
Filters can flow in both directions between tables. Useful in complex scenarios like many-to-many relationships or when multiple slicers are involved. Should be used cautiously to avoid ambiguous relationships and performance issues.&lt;/p&gt;

&lt;p&gt;As I conclude, I’d like to emphasize that defining relationships correctly in Power BI is critical to building a robust and accurate data model. By understanding cardinality and cross-filter direction, you can ensure your data behaves as expected across dashboards and reports. Take the time to plan your relationships carefully they are the backbone of meaningful, interactive insights in Power BI.&lt;/p&gt;

</description>
    </item>
    <item>
      <title>How Excel is Used in Real-World Data Analysis</title>
      <dc:creator>Maureen Mukami</dc:creator>
      <pubDate>Wed, 11 Jun 2025 07:42:51 +0000</pubDate>
      <link>https://dev.to/maureen_mukami_4268d10eac/how-excel-is-used-in-real-world-data-analysis-6c4</link>
      <guid>https://dev.to/maureen_mukami_4268d10eac/how-excel-is-used-in-real-world-data-analysis-6c4</guid>
      <description>&lt;p&gt;&lt;strong&gt;Excel&lt;/strong&gt; is a powerful spreadsheet software that allows users to input, organize, and manipulate data using rows and columns. While it's commonly associated with basic calculations Excel is widely used in the real-world data analysis across many industries such as manufacturing, healthcare, logistics and transportation &lt;/p&gt;

&lt;p&gt;&lt;em&gt;&lt;strong&gt;Real-World Applications of Excel&lt;/strong&gt;&lt;/em&gt;&lt;br&gt;
One key area where Excel is used is in &lt;strong&gt;sales analysis&lt;/strong&gt;. Businesses rely on it to monitor product sales across different regions, track performance over time, and identify best selling items. With charts and PivotTables, trends become visible, helping guide better business decisions.&lt;/p&gt;

&lt;p&gt;Excel is also used in &lt;strong&gt;inventory management&lt;/strong&gt;, it helps companies track stock levels and monitor reorder points. Using simple formulas and conditional formatting, it becomes easy to highlight items that are running low or overstocked.&lt;/p&gt;

&lt;p&gt;Another common use is in &lt;strong&gt;customer data analysis&lt;/strong&gt;. Excel enables organizations to study customer demographics, purchase behavior, and satisfaction scores. This helps tailor marketing strategies and improve customer experiences.&lt;/p&gt;

&lt;p&gt;&lt;em&gt;&lt;strong&gt;Excel Features and Formulas&lt;/strong&gt;&lt;/em&gt;&lt;br&gt;
One useful feature I’ve explored is &lt;strong&gt;Conditional formatting&lt;/strong&gt;. It automatically highlights cells based on specific criteria, making trends and outliers stand out at a glance. For instance, low sales figures can be marked in red, while high performers can be green.&lt;/p&gt;

&lt;p&gt;Another powerful tool is the &lt;strong&gt;PivotTable&lt;/strong&gt;. This feature allows one to summarize and analyze large datasets quickly. It’s incredibly helpful for grouping and comparing data like total sales by region or by product without writing complex formulas.&lt;/p&gt;

&lt;p&gt;I also learned the &lt;strong&gt;VLOOKUP&lt;/strong&gt; function, which allows you to search for a value in one table and return related information from another column. This is especially useful when working with multiple datasets, like matching customer names with their corresponding purchase history.&lt;/p&gt;

&lt;p&gt;&lt;em&gt;&lt;strong&gt;My Reflection&lt;/strong&gt;&lt;/em&gt;&lt;br&gt;
Learning Excel has completely transformed the way I perceive data. I used to think that data analysis required advanced tools or programming knowledge. However, I've come to realize that Excel is a powerful and versatile tool capable of handling a wide range of data tasks such as cleaning, organizing, analyzing and visualizing.&lt;/p&gt;

&lt;p&gt;What impressed me most is Excel’s flexibility. It enables me to carry out various functions in one platform, whether it's performing calculations, identifying trends, or generating visual insights. As data continues to play a critical role in decision-making across all industries, mastering Excel provides me with a strong foundation as I advance in my data science journey.&lt;/p&gt;

</description>
    </item>
  </channel>
</rss>
