<?xml version="1.0" encoding="UTF-8"?>
<rss version="2.0" xmlns:atom="http://www.w3.org/2005/Atom" xmlns:dc="http://purl.org/dc/elements/1.1/">
  <channel>
    <title>DEV Community: Raymond</title>
    <description>The latest articles on DEV Community by Raymond (@raymond_79cf41a94e682b4a3).</description>
    <link>https://dev.to/raymond_79cf41a94e682b4a3</link>
    <image>
      <url>https://media2.dev.to/dynamic/image/width=90,height=90,fit=cover,gravity=auto,format=auto/https:%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Fuser%2Fprofile_image%2F3256612%2F487c1416-a3db-483a-a8ba-f3fafdd3719f.png</url>
      <title>DEV Community: Raymond</title>
      <link>https://dev.to/raymond_79cf41a94e682b4a3</link>
    </image>
    <atom:link rel="self" type="application/rss+xml" href="https://dev.to/feed/raymond_79cf41a94e682b4a3"/>
    <language>en</language>
    <item>
      <title>RAG FOR DUMMIES 🤪</title>
      <dc:creator>Raymond</dc:creator>
      <pubDate>Wed, 17 Sep 2025 15:26:32 +0000</pubDate>
      <link>https://dev.to/raymond_79cf41a94e682b4a3/rag-for-dummies-4cbf</link>
      <guid>https://dev.to/raymond_79cf41a94e682b4a3/rag-for-dummies-4cbf</guid>
      <description>&lt;p&gt;&lt;strong&gt;RAG&lt;/strong&gt; stands for &lt;strong&gt;Retrieval-Augmented Generation&lt;/strong&gt;. It's a method for improving the accuracy and relevance of Large Language Models. An LLM is a type of AI that can generate human-like text, but its knowledge is limited to the data it was trained on. This means it can't access real-time information and may sometimes  make up information.&lt;/p&gt;

&lt;p&gt;RAG solves this problem by giving the LLM an external knowledge base to work from. When a user asks a question, the RAG system first searches a specific, reliable source like a company's internal documents or a recent database to find relevant information. Then, it uses this retrieved information to guide the LLM's response. This ensures the answer is both accurate and current, reducing the chance of errors.  &lt;/p&gt;

&lt;p&gt;The process works in two main steps. First, the&lt;u&gt; "retrieval" &lt;/u&gt;step is all about finding information. The user's query is used to search a database of documents, and the most relevant ones are pulled out. Second, the&lt;u&gt; "generation" &lt;/u&gt;step is where the LLM does its work. The retrieved documents, along with the original question, are fed into the LLM as a single prompt. The model then synthesizes this information to create a coherent answer, based on the facts from the retrieved data. This two-step process allows the AI to provide verifiable answers without the need for constant, expensive retraining on new data. &lt;/p&gt;

</description>
    </item>
    <item>
      <title>UNSUPERVISED ML</title>
      <dc:creator>Raymond</dc:creator>
      <pubDate>Fri, 05 Sep 2025 17:07:36 +0000</pubDate>
      <link>https://dev.to/raymond_79cf41a94e682b4a3/unsupervised-ml-e0a</link>
      <guid>https://dev.to/raymond_79cf41a94e682b4a3/unsupervised-ml-e0a</guid>
      <description>&lt;p&gt;&lt;strong&gt;What is Unsupervised ml ???&lt;/strong&gt;&lt;/p&gt;

&lt;p&gt;Unsupervised ML is a branch of ML that  operates without the luxury of labeled data like supervised ML, making it a powerful tool for discovery. I’d say the main principle behind unsupervised learning is pattern recognition. The algorithm must identify hidden structures, group similar data points, reduce dimensionality, or detect anomalies based only on the characteristics of the data. Among the various techniques within unsupervised ML, clustering emerges as maybe the most intuitive and widely applicable approach to understanding data. &lt;/p&gt;

&lt;p&gt;&lt;strong&gt;How does it work ??? **&lt;br&gt;
Unsupervised ML achieves it tasks by first of all **Clustering&lt;/strong&gt;. Clustering is a very common technique that groups data points that are similar to each other, such as when a company sorts its customers into different groups based on their buying habits without anyone telling the algorithm what a "frequent shopper" looks like. Next it uses &lt;strong&gt;Association methods&lt;/strong&gt; to find rules that show relationships between different items, with market basket analysis being a well-known example that might discover people who buy product A are also likely to buy product B. &lt;strong&gt;Dimensionality reduction techniques&lt;/strong&gt; then reduces the number of features or variables in a dataset, which helps make complex data easier to visualize and work with without losing the important information.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;Common Models and Examples&lt;/strong&gt;&lt;/p&gt;

&lt;p&gt;&lt;em&gt;K-Means Clustering&lt;/em&gt;&lt;br&gt;
K-means is the most popular clustering method because it's simple and works well. It splits data into k groups by finding the best center points for each group and putting data points with their nearest center. The algorithm keeps moving these center points and reassigning data until it finds the best arrangement. K-means works fast, but it works best with round-shaped groups and you need to decide how many groups you want beforehand.&lt;/p&gt;

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fv56os4hhqlg4b7aj0gjn.png" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fv56os4hhqlg4b7aj0gjn.png" alt=" " width="663" height="437"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;&lt;em&gt;Hierarchical Clustering&lt;/em&gt;&lt;br&gt;
This method creates a family tree of groups, working either from bottom-up or top-down. The bottom-up approach starts by treating each data point as its own group, then combines the closest groups together step by step. The top-down approach does the opposite - it starts with all data in one big group and keeps splitting it into smaller groups. This creates a tree diagram that shows how groups relate to each other at different levels.&lt;/p&gt;

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2F1j8syg1w58cab8c1wclf.png" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2F1j8syg1w58cab8c1wclf.png" alt=" " width="305" height="165"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;&lt;em&gt;Gaussian Mixture Models (GMM)&lt;/em&gt;&lt;br&gt;
GMM assumes that data comes from a mix of bell-shaped patterns and uses a special method to figure out the best settings. Unlike other clustering methods that put each data point in just one group, GMM gives each point a probability of belonging to different groups. This "soft" grouping approach is really helpful when it's unclear which group a data point should belong to, since it can partly belong to multiple groups at once.&lt;/p&gt;

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fjsc2s4jwclnwk3iocqc0.png" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fjsc2s4jwclnwk3iocqc0.png" alt=" " width="445" height="218"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;&lt;em&gt;My two cents(opinions)&lt;/em&gt;&lt;/strong&gt;&lt;/p&gt;

&lt;p&gt;As someone who's just started exploring clustering, I'm genuinely amazed by what these algorithms can do. What strikes me most is how clustering reveals hidden connections. I've noticed that the groups it creates often make perfect sense once you see them, but would have been nearly impossible to identify manually, especially with lots of data points. I've learned that choosing the right number of clusters is trickier than it first appears. Sometimes the "best" mathematical answer doesn't match what makes sense in the real world, and that's taught me that these tools need human judgment too. The results aren't always perfect, but they consistently give me new ways to think about my data.&lt;/p&gt;

</description>
    </item>
    <item>
      <title>A BRIEF INTRO TO CLASSIFICATION IN MACHINE LEARNING</title>
      <dc:creator>Raymond</dc:creator>
      <pubDate>Mon, 25 Aug 2025 02:24:56 +0000</pubDate>
      <link>https://dev.to/raymond_79cf41a94e682b4a3/a-brief-intro-to-classification-in-machine-learning-bae</link>
      <guid>https://dev.to/raymond_79cf41a94e682b4a3/a-brief-intro-to-classification-in-machine-learning-bae</guid>
      <description>&lt;p&gt;Classification in ML is a type of supervised learning where the goal is to predict which class a given input belongs to. These categorical predictions are derived from the labeled training data and fed to the algorithm. Some examples of where classification can be applied include trying to predict what’s in a given image or whether or not one would default on the credit card bills. &lt;/p&gt;

&lt;p&gt;&lt;strong&gt;HOW IT WORKS&lt;/strong&gt;&lt;/p&gt;

&lt;p&gt;In the training phase, the algorithm receives a dataset with input features (e.g, age, income) and known labels. For example, approved/denied loans. The data is then analysed to find patterns or relationships between the features. These patterns are then captured through a mathematical model. In the pattern recognition phase, the algorithm looks for decision boundaries that separate different classes. Let's say: If income &amp;gt; 50k AND credit score &amp;gt; 700, then likely approved. &lt;/p&gt;

&lt;p&gt;&lt;strong&gt;MODELS USED FOR CLASSIFICATION&lt;/strong&gt;&lt;br&gt;
&lt;em&gt;Decision Trees&lt;/em&gt;&lt;br&gt;
This is simply a tree of yes/no questions. It will make decisions based off a series of questions based on the data’s features. It’s easy to understand but can overfit. &lt;/p&gt;

&lt;p&gt;&lt;em&gt;K-Nearest Neighbours&lt;/em&gt;&lt;br&gt;
A simple algorithm that will classify a datapoint based on the majority class of their “k” nearest neighbours.&lt;/p&gt;

&lt;p&gt;&lt;em&gt;Random Forest&lt;/em&gt;&lt;br&gt;
This is an ensemble technique that will basically combine multiple decision trees to improve on the accuracy and reduce the risk of over fitting.&lt;/p&gt;

&lt;p&gt;&lt;em&gt;Logistics Regression&lt;/em&gt;&lt;br&gt;
Typically used for binary classification. This algorithm uses probability curves to estimate likelihood of each class. For instance, probabilities could be a 0.8 chance of approval, 0.2 chance of denial The final prediction based on highest probability. &lt;/p&gt;

&lt;p&gt;&lt;strong&gt;Applications of classification&lt;/strong&gt; algorithms can be found in spam filtering, sentiment analysis, image recognition, fraud detection, and medical diagnosis, among others.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;PERSONAL THOUGHTS AND INSIGHTS&lt;/strong&gt;&lt;/p&gt;

&lt;p&gt;For a long time, I've been fascinated by how different models work to predict outcomes and boost productivity. I had no idea there were so many approaches to this, like classification, regression, and more. Exploring them, both in my academic classes and for personal projects, has helped me understand how to efficiently use each method. I've learned that the key isn't to just use the latest or most complex model, but to choose the right one for the specific problem at hand.&lt;/p&gt;

&lt;p&gt;I believe that the next major trend in AI is its role as a "second brain". My view is that while our brains are exceptional at creative leaps and intuition, they aren't built to store and instantly recall every piece of information we encounter. This is where AI, particularly Generative AI, comes in as the perfect complement.&lt;/p&gt;

&lt;p&gt;It's not about replacing our minds but about offloading the mundane tasks of information retrieval and organization. It's an external memory that doesn't just hold data; it actively helps you connect ideas, synthesize information, and spark new insights.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;CHALLENGES FACED SO FAR&lt;/strong&gt;&lt;br&gt;
While i’m still a newbie and haven’t had as much experience working with classification models yet,   some of the challenges i’ve faced include: &lt;br&gt;
Dealing with imbalanced datasets&lt;br&gt;
I was trying to build a model to detect a rare event, and I quickly learned that if my training data had way more examples of one class than another, the model would become biased. This forced me to explore techniques like oversampling and undersampling to balance the classes&lt;br&gt;
Deciding on the right Features&lt;br&gt;
Another challenge was realizing that the raw data often isn't enough. It's not just about having a lot of data; it's about having the right data, and sometimes you have to create it yourself&lt;/p&gt;

</description>
    </item>
    <item>
      <title>A Simple Guide to Calculating Win Probabilities</title>
      <dc:creator>Raymond</dc:creator>
      <pubDate>Thu, 31 Jul 2025 17:43:22 +0000</pubDate>
      <link>https://dev.to/raymond_79cf41a94e682b4a3/a-simple-guide-to-calculating-win-probabilities-1e4l</link>
      <guid>https://dev.to/raymond_79cf41a94e682b4a3/a-simple-guide-to-calculating-win-probabilities-1e4l</guid>
      <description>&lt;p&gt;&lt;strong&gt;UNVEILING THE ODDS&lt;/strong&gt;&lt;br&gt;
In a recent assignment, I was tasked to find out the win probabilities of the most recent premier league teams for the next season. My methodology, grounded in python, utilizes the effective &lt;a href=""&gt;football-data.org&lt;/a&gt; API, which provides a clear, data-driven source of each team's performance. By simply making an API call, we can retrieve all the data we need for this simple model. I focused on the 2023 season of the Premier League for this project.&lt;br&gt;
The python script I ran uses a secure API key, loaded from an environment file for privacy and security. The core of our data retrieval is a GET request to the following URL, which fetches all matches from the specified league and season: &lt;a href="https://api.football-data.org/v4/competitions/PL/matches?season=2023" rel="noopener noreferrer"&gt;https://api.football-data.org/v4/competitions/PL/matches?season=2023&lt;/a&gt;.  This returns all the relevant data required to accomplish the task at hand. &lt;/p&gt;

&lt;p&gt;&lt;strong&gt;THE CALCULATION METHODOLOGY&lt;/strong&gt;&lt;br&gt;
The approach here was quite simple. Fetch the total wins for a particular team and the total number of matches played in that season. The win percentage is simply the ratio between the two expressed in percentage form. The python script processes the data in the following steps:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;
&lt;em&gt;Setup and Data Retrieval&lt;/em&gt; – The API credentials are setup and a request is made to the football-data.org endpoint. Whatever we retrieve is parsed using a dictionary.&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fs1isiip906ytnx5h4b8t.png" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fs1isiip906ytnx5h4b8t.png" alt=" " width="774" height="436"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;
&lt;em&gt;Filtering for Final Results&lt;/em&gt;: The API provides data for matches in various statuses (e.g., scheduled, live, finished). Our analysis is based on past performance, so we iterate through the list of matches and only consider those with the status “FINISHED”&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fe6wq50276ikc267kg6js.png" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fe6wq50276ikc267kg6js.png" alt=" " width="700" height="256"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;
&lt;em&gt;Aggregating Team Statistics&lt;/em&gt;: We create an empty dictionary called teams to store our performance data. For each finished match, we identify the home team, the away team, and the winner. We then update our teams dictionary. For both the home and away team, we increment the played counter. Based on the winner field ("HOME_TEAM" or "AWAY_TEAM"), we increment the wins counter for the corresponding team.&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fm2b174sn6q15j3tsexve.png" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fm2b174sn6q15j3tsexve.png" alt=" " width="700" height="334"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;
&lt;em&gt;Calculating and Displaying Probabilities&lt;/em&gt;: After iterating through all the finished matches, the teams dictionary contains the total number of games played and won for every team in the league. The final step is to loop through this dictionary and calculate the win probability using the formula:&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;&lt;strong&gt;- Win Probability= (Total Matches Played / Total Wins) × 100&lt;/strong&gt;&lt;/p&gt;

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fogi7pqr5xncxin6txdbq.png" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fogi7pqr5xncxin6txdbq.png" alt=" " width="744" height="266"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;OBSERVATION/CONCLUSION&lt;/strong&gt;&lt;/p&gt;

&lt;p&gt;Despite the fact that the method used fails to factor in opponent strength, home/away field or other parameters, it provides a solid baseline for understanding team performance and is a quick and efficient way to summarize a team's historical success within a season.&lt;/p&gt;

</description>
    </item>
    <item>
      <title>Excel for Data Analysis: From Chaos to Clarity.</title>
      <dc:creator>Raymond</dc:creator>
      <pubDate>Wed, 11 Jun 2025 18:01:29 +0000</pubDate>
      <link>https://dev.to/raymond_79cf41a94e682b4a3/excel-for-data-analysis-from-chaos-to-clarity-5h0e</link>
      <guid>https://dev.to/raymond_79cf41a94e682b4a3/excel-for-data-analysis-from-chaos-to-clarity-5h0e</guid>
      <description>&lt;p&gt;&lt;strong&gt;What is Excel?&lt;/strong&gt;&lt;/p&gt;

&lt;p&gt;Microsoft Excel is a popular tool that helps people organize, analyze, and make sense of data. While it might look like just a grid of rows and columns, Excel is much more than that—it’s a flexible platform for working with numbers, spotting trends, and turning raw information into clear, actionable insights using formulas, charts, and interactive tools.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;How Excel is Used in Real Life for Data Analysis&lt;/strong&gt;&lt;/p&gt;

&lt;p&gt;Tracking Business Performance&lt;br&gt;
Companies often use Excel to keep an eye on how their business is doing. It helps track key numbers like sales, customer reviews, and discounts. With this information, businesses can figure out which products are performing well and make smarter decisions about pricing and inventory.&lt;/p&gt;

&lt;p&gt;Handling Finances and Budgets&lt;br&gt;
Excel is a go-to tool for financial planning. Whether it’s creating budgets, keeping track of expenses, or putting together detailed financial reports, Excel makes it easier. Its powerful formulas can guide important financial choices.&lt;/p&gt;

&lt;p&gt;Analyzing Marketing Campaigns&lt;br&gt;
Marketing teams use Excel to see how well their campaigns are working. By looking at data like click-through rates, customer interest, and how much return they get on their ad spending, teams can spot what’s working and where to focus future efforts. Excel’s charts and graphs make this data easier to understand and share.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;Excel Features and Formulas I've learned&lt;/strong&gt;&lt;/p&gt;

&lt;p&gt;Pivot Charts and Tables&lt;br&gt;
Pivot tables and charts are great for quickly making sense of large amounts of data. With just a few clicks—dragging and dropping columns—you can spot your best-selling products, break down trends by category, or compare performance over time. Pivot charts help turn complex data into clear visuals that are easy to understand.&lt;/p&gt;

&lt;p&gt;Conditional Formatting&lt;br&gt;
This feature helps important data stand out automatically. For example, you can set it to highlight ratings above 4.5 in green and those below 3.0 in red. It’s a simple way to quickly spot what’s doing well and what might need attention—no manual checking required.&lt;/p&gt;

&lt;p&gt;INDEX and MATCH Functions&lt;br&gt;
This combo is a smarter alternative to VLOOKUP. It lets you pull data from anywhere in a table—not just from left to right. It’s especially handy when working with big datasets, like finding the top-rated products from a full list without having to rearrange your columns.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;MY PERSONAL REFLECTION&lt;/strong&gt;&lt;/p&gt;

&lt;p&gt;I used to think data was messy - this overwhelming block of numbers and text that was incredibly hard to interpret. Raw datasets felt like puzzles with missing pieces, and drawing meaningful conclusions seemed impossible. However, learning Excel's features this week has completely transformed my perspective.&lt;br&gt;
Now I can see how straightforward data analysis can become with proper tools and techniques. Good data cleaning establishes a solid foundation, while visualization features like charts and conditional formatting reveal patterns that were previously hidden. What once appeared as chaotic information now presents clear stories and actionable insights.&lt;br&gt;
Most importantly, I've learned that effective data analysis isn't about having the most sophisticated tools - it's about asking the right questions and using available features strategically. Excel has shown me that even seemingly simple tools can unlock powerful insights when applied thoughtfully.&lt;/p&gt;

</description>
    </item>
  </channel>
</rss>
