<?xml version="1.0" encoding="UTF-8"?>
<rss version="2.0" xmlns:atom="http://www.w3.org/2005/Atom" xmlns:dc="http://purl.org/dc/elements/1.1/">
  <channel>
    <title>DEV Community: Leon Mutisya</title>
    <description>The latest articles on DEV Community by Leon Mutisya (@leonmutisya).</description>
    <link>https://dev.to/leonmutisya</link>
    <image>
      <url>https://media2.dev.to/dynamic/image/width=90,height=90,fit=cover,gravity=auto,format=auto/https:%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Fuser%2Fprofile_image%2F1867087%2Ffb9e4c91-21b3-46a5-a0d9-03642c51ab2b.png</url>
      <title>DEV Community: Leon Mutisya</title>
      <link>https://dev.to/leonmutisya</link>
    </image>
    <atom:link rel="self" type="application/rss+xml" href="https://dev.to/feed/leonmutisya"/>
    <language>en</language>
    <item>
      <title>Feature Engineering: The Ultimate Guide</title>
      <dc:creator>Leon Mutisya</dc:creator>
      <pubDate>Sat, 17 Aug 2024 12:06:45 +0000</pubDate>
      <link>https://dev.to/leonmutisya/feature-engineering-the-ultimate-guide-48ip</link>
      <guid>https://dev.to/leonmutisya/feature-engineering-the-ultimate-guide-48ip</guid>
      <description>&lt;p&gt;&lt;strong&gt;Feature Engineering&lt;/strong&gt;&lt;br&gt;
Feature Engineering is described as a preprocessing step in machine learning which transforms raw data into a more effective set of inputs which have several attributes known as &lt;strong&gt;features.&lt;/strong&gt;&lt;/p&gt;

&lt;p&gt;The success of machine learning models heavily depends on the quality of the features used to train them. Feature engineering involves a set of techniques that enable us to create new features by combining or transforming the existing ones. These techniques help highlight the most important patterns and relationships in the data, which in turn helps the machine learning model to learn from the data more effectively.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;Key Techniques in Feature Engineering&lt;/strong&gt;&lt;br&gt;
Feature Engineering can be classified into two key steps namely;&lt;/p&gt;

&lt;ol&gt;
&lt;li&gt;&lt;p&gt;Data Pre-processing &lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;Business Understanding(Domain Knowledge)&lt;/p&gt;&lt;/li&gt;
&lt;/ol&gt;

&lt;p&gt;&lt;strong&gt;Data Pre-processing&lt;/strong&gt;&lt;br&gt;
This is usually a step in feature engineering and involves preparing and manipulating the data to the current machine language needs. Various techniques are used here among them;&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;&lt;p&gt;Handling Missing Value where techniques like imputation (mean, median, mode), or using algorithms that handle missing values natively can be employed.&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;Encoding Categorical Variables where categorical data must be converted into numerical form for most algorithms using common methods like one-hot encoding, label encoding, and target encoding.&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;Scaling and Normalization where scaling features ensures that they contribute equally to the model. Techniques include standardization (z-score)&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;Feature Interaction &amp;amp; Feature Creation where existing features are combined to create new features thus creating complex relationships with the data&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;Dimensionality Reduction where techniques like PCA (Principal Component Analysis) or t-SNE reduce the number of features while retaining the most important information.&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;&lt;a href="https://dev.to/leonmutisya/understanding-your-data-the-essentials-of-exploratory-data-analysis-3o8e"&gt;EDA&lt;/a&gt; can also be utilized in feature engineering and is usually a precursor to feature engineering.&lt;/p&gt;&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;&lt;strong&gt;Domain Knowledge&lt;/strong&gt;&lt;br&gt;
Domain knowledge refers to the understanding and expertise in a specific field or industry. In feature engineering, it involves applying insights and understanding of the data's context and relationships to create meaningful features that can enhance model performance.&lt;/p&gt;

&lt;p&gt;It helps in identifying which features are relevant to the problem at hand and understand data relationships.&lt;/p&gt;

</description>
      <category>beginners</category>
      <category>python</category>
      <category>datascience</category>
    </item>
    <item>
      <title>Understanding Your Data: The Essentials of Exploratory Data Analysis</title>
      <dc:creator>Leon Mutisya</dc:creator>
      <pubDate>Sun, 11 Aug 2024 17:48:04 +0000</pubDate>
      <link>https://dev.to/leonmutisya/understanding-your-data-the-essentials-of-exploratory-data-analysis-3o8e</link>
      <guid>https://dev.to/leonmutisya/understanding-your-data-the-essentials-of-exploratory-data-analysis-3o8e</guid>
      <description>&lt;h2&gt;
  
  
  Exploratory Data Analysis
&lt;/h2&gt;

&lt;p&gt;Data analysis is the process of inspecting, cleansing, transforming, and modeling data with the goal of discovering useful information, informing conclusions, and supporting decision-making. &lt;/p&gt;

&lt;p&gt;Exploratory data analysis commonly known as (EDA) is the approach of analyzing data sets to summarize their main characteristics, often using statistical graphics and other data visualization methods. We can therefore say that EDA supports the main function of data analysis. &lt;/p&gt;

&lt;p&gt;Look at it as a professional athlete running track. Before they take to competition they must first ensure their spikes are well functioning ,scope out the track and partake in warmups. Simply, EDA is the warmup session to the Data Analysis' running.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;Importance of EDA&lt;/strong&gt;&lt;br&gt;
The main purpose of EDA is to help look at data before making any assumptions. It helps identify obvious errors, as well as better understand patterns within the data, detect outliers and find interesting relations among variables. EDA also helps stakeholders by confirming they are asking the right questions&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;Essentials of EDA&lt;/strong&gt;&lt;br&gt;
EDA involves a range of activities, including data integration, analysis, cleaning, transformation, and dimension reduction. In this article we will highlight some key steps in EDA.&lt;/p&gt;

&lt;ol&gt;
&lt;li&gt;&lt;p&gt;Data Cleaning&lt;br&gt;
Begin by checking for missing values, duplicates, and inconsistent data types. Clean the data to ensure it's ready for analysis. we usually import necessary python libraries such as pandas, NumPy and matplotlib.&lt;br&gt;
               &lt;code&gt;import pandas as pd&lt;/code&gt;&lt;br&gt;
               &lt;code&gt;import numpy as np&lt;/code&gt;&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;Descriptive Statistics&lt;br&gt;
Here we will calculate basic statistics like mean, median, mode, standard deviation, and variance to get a sense of the data distribution. This is usually supported by the imported libraries that enable the aforementioned mathematical functions.&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;Data Visualization&lt;br&gt;
Once we have calculated the mean, median ,std etc. of the data we will use visual tools like histograms, box plots, and scatter plots to visualize data distributions, relationships, and patterns. These visuals reveal trends over time in a way which cannot be seen from raw data. This can be broken down to;&lt;/p&gt;&lt;/li&gt;
&lt;/ol&gt;

&lt;ul&gt;
&lt;li&gt;&lt;p&gt;Univariate Analysis which is the use of histograms, box plots, and density plots to examine the distribution of individual variables.&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;Bivariate Analysis which is the creation of  scatter plots, pair plots, and bar plots to explore relationships between two variables and;&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;Multivariate Analysis which employs heatmaps, correlation matrices, and pair plots to investigate interactions among multiple variables.&lt;/p&gt;&lt;/li&gt;
&lt;/ul&gt;

&lt;ol&gt;
&lt;li&gt;&lt;p&gt;Correlation Analysis&lt;br&gt;
In correlation analysis we compute correlation matrices and heatmaps to explore relationships between variables. This helps identify which variables are related to inform further modelling.&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;
&lt;/li&gt;
&lt;/ol&gt;

&lt;p&gt;Handling Outliers&lt;br&gt;
Detecting and analyzing outliers using methods like the Z-score or Interquartile Range (IQR) is usually done where decisions are made whether to keep ,transform, or remove outliers based on their impact on the analysis.&lt;/p&gt;

&lt;ol&gt;
&lt;li&gt;&lt;p&gt;Understanding Distribution where we analyze the shape of data distributions to determine if they are skewed or normally distributed. This informs decisions about data transformation and the suitability of statistical tests&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;Dimension Reduction or Addition where we can either reduce or impute the number of variables while preserving essential information. EDA can inform this through the use of statistics methods in filling our data. This needs domain knowledge as it should and will add value to the data.&lt;/p&gt;&lt;/li&gt;
&lt;/ol&gt;

&lt;p&gt;&lt;strong&gt;Conclusions&lt;/strong&gt;&lt;br&gt;
In conclusion, EDA is crucial for understanding datasets, identifying patterns, and informing subsequent analysis.&lt;/p&gt;

</description>
      <category>python</category>
      <category>beginners</category>
      <category>datascience</category>
    </item>
    <item>
      <title>Data Analysis: The Ultimate Guide to Data Analytics Techniques and Tools</title>
      <dc:creator>Leon Mutisya</dc:creator>
      <pubDate>Sun, 04 Aug 2024 09:25:38 +0000</pubDate>
      <link>https://dev.to/leonmutisya/data-analysis-the-ultimate-guide-to-data-analytics-techniques-and-tools-18gm</link>
      <guid>https://dev.to/leonmutisya/data-analysis-the-ultimate-guide-to-data-analytics-techniques-and-tools-18gm</guid>
      <description>&lt;h2&gt;
  
  
  Introduction
&lt;/h2&gt;

&lt;p&gt;In this article, we look into what is data analysis and techniques/tools commonly used in the domain. Understanding these concepts is essential to cope up with humongous data generated in many industries today and for its effective management and retrieval of useful information from it.&lt;/p&gt;

&lt;h2&gt;
  
  
  Data Analysis
&lt;/h2&gt;

&lt;p&gt;Data analysis is the process of inspecting, cleansing, transforming, and modeling data with the goal of discovering useful information, informing conclusions, and supporting decision-making. Data analysis therefore plays a crucial role in the growth of diverse industries  as we need to manage the countless bits of data that are streaming in to make informed business decisions.&lt;/p&gt;

&lt;h3&gt;
  
  
  Types of Data Analytics
&lt;/h3&gt;

&lt;p&gt;Data Analysis can be divided into four main types among them;&lt;/p&gt;

&lt;ol&gt;
&lt;li&gt;&lt;p&gt;Descriptive Data Analysis : which is the simplest type of analytics and the foundation the other types are built on by involving understanding past data&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;Diagnostic Analytics : which asks why a particular thing happened in regards to data and thus analyses past data&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;Predictive Analytics : which makes predictions as to future events by use of historical data and;&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;Prescriptive Analytics which deals with what next in regards to data recommending actions based on predictive analytics&lt;/p&gt;&lt;/li&gt;
&lt;/ol&gt;

&lt;h3&gt;
  
  
  Data Analytics Tools
&lt;/h3&gt;

&lt;p&gt;1.Programming Languages&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;&lt;p&gt;Python which is a popular close to human programming language with libraries like Pandas, NumPy, and SciPy&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;SQL which queries and manages databases&lt;/p&gt;&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;2.Data Visualization Tools &amp;amp; Statistical Analysis Tools&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;&lt;p&gt;R which is a  language tailored for statistical analysis and data visualization&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;Tableau which creates interactive dashboards&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;Power BI which is Microsoft's business analytics tool&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;Matplotlib and Seaborn which are python libraries for creating static, animated, and interactive visualizations&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;MS Excel which is a widely used spreadsheet software that offers basic statistical tool&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;SAS which is a premium statistical analysis platform offering GUI and scripting options for advanced analyses and publication of worthy graphics and charts&lt;/p&gt;&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;3.Machine Learning Libraries&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;Python Libraries such as NumPy for  high level mathematical functions&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;4.Big Data Tools such as NoSQL Databases like MongoDB which is designed for storing, retrieving, and managing big data&lt;/p&gt;

&lt;p&gt;5.Business Intelligence Tools&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;&lt;p&gt;QlikView which is a BI tool for transforming raw data into knowledge.&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;Looker which is a modern data platform that creates real-time dashboards and reports&lt;/p&gt;&lt;/li&gt;
&lt;/ul&gt;

&lt;h2&gt;
  
  
  Data Analytics Techniques
&lt;/h2&gt;

&lt;p&gt;1.Data Collection which is where data is collected from a variety of sources either through interviews, downloads from online sources, or reading documentation and in different file formats and datasets&lt;br&gt;
2.Data Processing where data is processed for analysis and may involve putting data into rows and columns &lt;br&gt;
3.Data Cleaning where data is cleaned and processed . Here missing values are handled and we ensure that data is consistent.&lt;br&gt;
4.Exploratory Data Analysis which is a method of taking a look at a dataset, summing up the essential elements; it often employs statistical graphics, along with other data visualization techniques. The additional cleaning of the data or some further transformations may be required in this step based on the preliminary findings.&lt;br&gt;
In this step&lt;a href="https://en.wikipedia.org/wiki/Exploratory_data_analysis" rel="noopener noreferrer"&gt;EDA&lt;/a&gt; can have further cleaning and requests to the data and &lt;strong&gt;data visualization&lt;/strong&gt; is also a technique used.&lt;br&gt;
5.Data Classification and Clustering where this method identifies structures within a dataset. It’s like sorting objects into different boxes (clusters) based on their similarities. The data points within a similar group are similar to each other (homogeneous). Cluster analysis aims to find hidden patterns in the data.&lt;br&gt;
6.Time Series Analysis&lt;br&gt;
This process is used for those data points that are collected or recorded at regular time intervals. Analysis of time series enables identification of trends, cycles, and patterns over some time, thus very useful in projecting future events.&lt;/p&gt;

&lt;h3&gt;
  
  
  Conclusion
&lt;/h3&gt;

&lt;p&gt;In today's data-driven world, a person has to understand data analysis and its processes. Data analysis plays a significant role in current operations, beginning with business and proceeding further to sports, medicine and marketing among other fields . This information enables an organization to make informed decisions that better equip it to forge ahead into the ever-changing global environment &lt;/p&gt;

</description>
    </item>
  </channel>
</rss>
