<?xml version="1.0" encoding="UTF-8"?>
<rss version="2.0" xmlns:atom="http://www.w3.org/2005/Atom" xmlns:dc="http://purl.org/dc/elements/1.1/">
  <channel>
    <title>DEV Community: Isaac</title>
    <description>The latest articles on DEV Community by Isaac (@130).</description>
    <link>https://dev.to/130</link>
    <image>
      <url>https://media2.dev.to/dynamic/image/width=90,height=90,fit=cover,gravity=auto,format=auto/https:%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Fuser%2Fprofile_image%2F1181318%2F3ce1bc8b-159f-4b0e-9498-bd4a2fc37147.jpg</url>
      <title>DEV Community: Isaac</title>
      <link>https://dev.to/130</link>
    </image>
    <atom:link rel="self" type="application/rss+xml" href="https://dev.to/feed/130"/>
    <language>en</language>
    <item>
      <title>DATA SCIENCE 2024 ROADMAP</title>
      <dc:creator>Isaac</dc:creator>
      <pubDate>Tue, 06 Aug 2024 00:29:36 +0000</pubDate>
      <link>https://dev.to/130/data-science-2024-roadmap-331p</link>
      <guid>https://dev.to/130/data-science-2024-roadmap-331p</guid>
      <description>&lt;p&gt;In the current growing technology industry, organizations are generating and storing more and more data and are looking to hire professionals to derive valuable insights from said data to help drive business decisions. Here, data science plays a big role and has actually been considered “the sexiest job of the 21st century” according to Harvard Business Review. Whether you're a recent graduate, a career changer, or simply curious about the world of data, this roadmap is designed to guide you through your journey to achieve a desired objective or goal within the timeframe of a year.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;1. Understand the Basics&lt;/strong&gt;&lt;br&gt;
&lt;strong&gt;A. What is Data Science?&lt;/strong&gt;&lt;br&gt;
Before diving into data science, it's crucial to understand its essence. Data science is the practice of extracting meaningful insights and knowledge from data using various techniques, including statistical analysis, machine learning, and data visualization. Hence, briefly it can be said that data science involves;&lt;br&gt;
• Statistics, computer science, mathematics&lt;br&gt;
• Data cleaning and formatting&lt;br&gt;
• Data visualization&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;2. Learning the Fundamentals ( 3 – 5 months)&lt;/strong&gt;&lt;br&gt;
&lt;strong&gt;A. Mathematics and Statistics ( 1 – 2 months)&lt;/strong&gt;&lt;br&gt;
Linear Algebra and Calculus are very important as they help in understanding various machine learning algorithms that play an important role in data science. Similarly, statistics is very significant as it is a part of data analysis. Descriptive Statistics is a powerful method to summarize data while Inferential Statistics is applicable in hypothesis testing.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;B. Programming Skills. (2 – 3 months)&lt;/strong&gt;&lt;br&gt;
If you are a beginner, learning Python is strongly recommended for data science. Python is a favorite among data scientists for its simplified syntax. One can also access a lot of open-source libraries, including NumPy, pandas, and scikit-learn for the implementation of various data science tasks.&lt;/p&gt;

&lt;ol&gt;
&lt;li&gt;
&lt;strong&gt;Data Manipulation and Analysis ( 2 – 3 months)&lt;/strong&gt;
A. &lt;strong&gt;Data Collection and Wrangling (1 months)&lt;/strong&gt;
Data collection is the process of gathering relevant data for further analysis from a variety of sources while data wrangling is the preparing and transforming data to an easier format for further analysis.&lt;/li&gt;
&lt;/ol&gt;

&lt;p&gt;B. &lt;strong&gt;Exploratory Data Analysis (EDA) ( 1 – 2 months)&lt;/strong&gt;&lt;br&gt;
Master the art of EDA to gain insights from your data. EDA involves exploring the data using various statistical models like mean, median etc and come up with hypotheses and perform analyses. Data visualization tools like Matplotlib and Seaborn will be your best friends during this stage and will include data exploration using visual methods like histograms, bar charts and pie charts to identify trends and patterns within data.&lt;/p&gt;

&lt;ol&gt;
&lt;li&gt;
&lt;strong&gt;Machine Learning and AI ( 3 – 4 months)&lt;/strong&gt;
&lt;strong&gt;A. Introduction to Machine Learning&lt;/strong&gt;
Understand the core concepts of machine learning, including supervised learning which includes regression and classification problems and unsupervised learning whose applications are clustering and dimensionality reduction.&lt;/li&gt;
&lt;/ol&gt;

&lt;p&gt;&lt;strong&gt;B. Model Building&lt;/strong&gt;&lt;br&gt;
Learn to build, train, and evaluate machine learning models. Scikit-learn provides an extensive toolkit for this purpose.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;C. Deep Learning (Optional)&lt;/strong&gt;&lt;br&gt;
If you're interested in more advanced techniques, consider delving into deep learning using libraries like Tensor Flow or PyTorch.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;5.Data Engineering (2 – 3 months)&lt;/strong&gt;&lt;br&gt;
Data engineering is the field of building data infrastructure by designing, building and maintaining ETL data pipelines. This is not mandatory for data scientists to learn but having a good understanding is a big plus in the job market.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;Points to Remember&lt;/strong&gt;&lt;br&gt;
• &lt;strong&gt;No Degree Requirement&lt;/strong&gt;: While a degree in computer science can be beneficial, it's not mandatory for a career in data science. What matters most are the skills you acquire and master.&lt;br&gt;
• &lt;strong&gt;Domain Expertise&lt;/strong&gt;: Having expertise in a specific domain or industry can be an advantage as it enables you to leverage data effectively for solving domain-specific problems.&lt;br&gt;
• &lt;strong&gt;Communication Skills&lt;/strong&gt;: Good verbal and written communication skills are essential for collaborating with various stakeholders and effectively conveying your data findings and recommendations.&lt;br&gt;
• Focus on Fundamentals: Data science is vast, so it's important to start by understanding the basics before delving into advanced concepts. Building a strong foundation is key.&lt;br&gt;
• &lt;strong&gt;Practical Applications&lt;/strong&gt;: Practical skills gained through working on real-world projects are highly valued by organizations. Practical application of knowledge is often more important than theoretical knowledge alone.&lt;br&gt;
• &lt;strong&gt;Track Your Progress&lt;/strong&gt;: Monitoring your learning progress is crucial. Assignments and assessments can help you gauge whether you are grasping concepts effectively and moving in the right direction.&lt;br&gt;
•** Stay Updated**: Data science is an evolving field. Keeping up with the latest research and developments will help you remain competitive and stand out in your career.&lt;br&gt;
These points provide valuable guidance for individuals looking to embark on a data science journey or advance their existing data science skills. They emphasize the importance of a balanced approach that combines theoretical knowledge with practical experience and ongoing learning.&lt;/p&gt;

</description>
    </item>
    <item>
      <title>DATA ENGINEERING FOR BEGINNERS -A STEP BY STEP GUIDE</title>
      <dc:creator>Isaac</dc:creator>
      <pubDate>Sun, 05 Nov 2023 12:37:01 +0000</pubDate>
      <link>https://dev.to/130/data-engineering-for-beginners-a-step-by-step-guide-1c3a</link>
      <guid>https://dev.to/130/data-engineering-for-beginners-a-step-by-step-guide-1c3a</guid>
      <description>&lt;p&gt;What is data engineering? This is probably the first question that comes to your mind after reading this title. Data engineering is simply the acquisition, storage, transformation and management of data. This ranges from simple data analysis to writing complex data processing models and algorithms to manage Big Data. Data engineers are the most technical profile in the field of data science, serving as a critical bridge between software and application developers. They are responsible for the first step of the traditional data science workflow, namely the process of collecting and storing data. They ensure that large amounts of data collected from various sources become resources that can be accessed by other data science professionals, such as data analysts and data scientists. The most important skills of a data engineer include:&lt;/p&gt;

&lt;ol&gt;
&lt;li&gt;- Be proficient in programming in languages ​​such as Python and Scala. - Learn automation and scripts&lt;/li&gt;
&lt;li&gt;- Understand database management and develop your SQL skills&lt;/li&gt;
&lt;li&gt;- Master data processing system&lt;/li&gt;
&lt;li&gt;- Learn how to organize your workflow&lt;/li&gt;
&lt;li&gt;- Expand knowledge of cloud computing and platforms such as AWS&lt;/li&gt;
&lt;li&gt;- Increase your knowledge of infrastructure such as Docker and Kubernetes&lt;/li&gt;
&lt;/ol&gt;

&lt;p&gt;Programming skills in Python, R and/or Scala and database management are the core skills of any data engineer, as you will use them 75% of your time. Therefore, it is very important to learn these skills. According to many studies, data engineering is a growing field that holds great promise for those willing to enter. Below is a review of recent studies on advancements in technology.&lt;/p&gt;

&lt;p&gt;&lt;a href="https://media.dev.to/cdn-cgi/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Frzm7w2g2y7lryb9o66k8.png" class="article-body-image-wrapper"&gt;&lt;img src="https://media.dev.to/cdn-cgi/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Frzm7w2g2y7lryb9o66k8.png" alt="Career growths in TEC" width="768" height="462"&gt;&lt;/a&gt; &lt;/p&gt;

&lt;p&gt;Data engineering is an emerging profession and it is not always easy for recruiters to find the right candidates. Competition for this hard-to-find talent is high among companies, resulting in some of the highest salaries among data science positions. Based on several career portals, the average salary for data engineers in the United States is $114,564. However, these numbers differ depending on where you work in the country. So, for example, according to Glassdoor, the average salary for a data analyst in New York is $120,637, while in Chicago it is $113,610 and in Houston it is $94,416.&lt;/p&gt;

</description>
    </item>
    <item>
      <title>TIME SERIES ANALYSIS</title>
      <dc:creator>Isaac</dc:creator>
      <pubDate>Wed, 01 Nov 2023 10:10:43 +0000</pubDate>
      <link>https://dev.to/130/time-series-analysis-4138</link>
      <guid>https://dev.to/130/time-series-analysis-4138</guid>
      <description>&lt;p&gt;A time series analysis is a method of data analysis that involves the observation of a variable's trend over time. Using this data organizations are able to identify how a certain product has been doing over time as well as identify the trend of its rise and falls over time and be able to answer the "why" question. This could help know at what season a product has the best sales so at to ensure they have it stock by the next season sales are expected to rise. This is important important as it  helps the company do timely re-stocking and hence they are able to maximize sales and profits.&lt;br&gt;
Additionally, a time-series analysis helps an organization detect when a commodity was at its lowest and this can help them understand why it was so by reviewing the timelines and see circumstances that might have led to that. This makes a time-series analysis an important tool for business forecasting.&lt;/p&gt;

&lt;p&gt;The best way to get the most of a time-series analysis is by visualizing it using eg with a scatter-plot, line chart or any other appropriate data visualizing tool at your disposal&lt;/p&gt;

&lt;h2&gt;
  
  
  Steps to a Time-Series Analysis
&lt;/h2&gt;

&lt;p&gt;&lt;strong&gt;1. Data Collection&lt;/strong&gt;&lt;br&gt;
Gather all the relevant information you need for your analysis may it be sales records, website logs, order histories, date and time stamps any relevant information for your analysis should be available&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;2. Data Cleaning and Preprocessing&lt;/strong&gt;&lt;br&gt;
This is the most important stage as this helps prepare the data you will use in you analysis. Data pre-processing involves the filling of null values, removal of duplicate values, data type conversion, dealing with outliers and standardizing the data to the best workable format possible. This helps to prevent outputting wrong analysis during your analysis based on the rule of &lt;strong&gt;&lt;em&gt;Garbage In Garbage Out&lt;/em&gt;&lt;/strong&gt;&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;3. Data Visualization&lt;/strong&gt; &lt;/p&gt;

&lt;p&gt;After you data is ready, use data visualizing tools to plot your data to visuals that will help you get actionable insights from your data. This will enable you notice major trends, outliers as well as patterns that are important in answering your questions&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;4. Time Series Decomposition&lt;/strong&gt;&lt;br&gt;
A time series has 3 components : Trend, Seasonality and Residue(noise)&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;Trend=&amp;gt;&lt;/strong&gt; Identifiable pattern of frequency of events. &lt;strong&gt;Seasonal=&amp;gt;&lt;/strong&gt; Trends that happen after specific amount of time  &lt;strong&gt;Residual=&amp;gt;&lt;/strong&gt; Patterns that are out of trends and not in season&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;5. Model Selection&lt;/strong&gt;&lt;br&gt;
Choose an appropriate Time-Series model to help model your data. The major models include:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;
&lt;strong&gt;Autoregressive Integrated Moving Average (ARIMA)&lt;/strong&gt;: For stationary data.&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;Seasonal Decomposition of Time Series (STL)&lt;/strong&gt;: For data with seasonality.&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;Exponential Smoothing (ETS)&lt;/strong&gt;: Another method for forecasting time series data.&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;Data from the model should be evaluated so as to determine its correctness.&lt;/p&gt;

</description>
    </item>
    <item>
      <title>DIVE INTO RFM DATA ANALYSIS</title>
      <dc:creator>Isaac</dc:creator>
      <pubDate>Wed, 11 Oct 2023 06:02:50 +0000</pubDate>
      <link>https://dev.to/130/dive-into-rfm-data-analysis-18ak</link>
      <guid>https://dev.to/130/dive-into-rfm-data-analysis-18ak</guid>
      <description>&lt;p&gt;RFM is an acronym  for &lt;strong&gt;R&lt;/strong&gt;ecency &lt;strong&gt;F&lt;/strong&gt;requency &lt;strong&gt;M&lt;/strong&gt;onetary Data Analysis. It is a technique used by most data scientist in the e-commerce industry to rank customers based on how lately they bought goods from their store, how frequently they make purchases and how much in total they spend &lt;/p&gt;

&lt;p&gt;RFM can be used to detect customers usage behavior and  patterns, the number of customers who are frequent buyers or high spenders. This helps the company to come up with with new custom marketing strategies for each target group with  an aim to accelerate sales and get the most out of each group. The Analysis is thus  very important in coming up with better marketing strategies of a particular company.&lt;/p&gt;

&lt;p&gt;In this article we will dive into a sample case scenario to explore some uses of RFM and how to leverage it in your industry. &lt;strong&gt;This is case study for learning purposes and not a full grade RFM Model&lt;/strong&gt;&lt;br&gt;
In an attempt to solve &lt;a href="https://statso.io/rfm-analysis-case-study/"&gt;this&lt;/a&gt;  I came up with this model:&lt;br&gt;
&lt;/p&gt;

&lt;div class="highlight js-code-highlight"&gt;
&lt;pre class="highlight plaintext"&gt;&lt;code&gt;import pandas as pd
import numpy as np
from matplotlib import pyplot as plt

#Read csv File with raw data

myData= pd.read_csv("rfm_data.csv")

#Remove duplicates by making a copy of original Data
data = myData.copy(deep=True).drop_duplicates(subset=["CustomerID"],keep="last")


#Calculate frequency of each CustomerID
frequencySeries =myData.CustomerID.value_counts().reset_index(name="Frequency")
print(frequencySeries.loc[frequencySeries["CustomerID"]==8317])



#Appending frequency column to DataFrame
frequencySeries.rename({"index":"CustomerID"},inplace=True)
data = data.merge(frequencySeries,on="CustomerID",how="left")


#Total Money spent per user
moneySpent = myData.groupby("CustomerID")['TransactionAmount'].sum().reset_index(name="Total Spent")
data= data.merge(moneySpent,on="CustomerID",how="left")




recencyScore = [5,4,3,2,1,]
frequencyScore = [1,2,3,4,5]
monetaryScore = [1,2,3,4,5]

#Grading users 
data["recencyScore"] =pd.cut(data["recency"],bins=5,labels=recencyScore).astype(int)

data["FrequencyScore"] = pd.cut(data["Frequency"],bins=5,labels=frequencyScore).astype(int)

data["Monetary Score"] = pd.cut(data["Total Spent"],bins=5,labels=monetaryScore).astype(int)

data['totalScore'] = data["FrequencyScore"] + data["Monetary Score"] + data["recencyScore"]

#Ranking users based on ttal RFM score

myLabels = ["Beginner","Intermediate","PRO"]
data["Rank"] = pd.cut(data["totalScore"],bins=3,labels=myLabels)
&lt;/code&gt;&lt;/pre&gt;

&lt;/div&gt;



&lt;p&gt;I then used &lt;em&gt;matplotlib&lt;/em&gt; to show the various statistics of users&lt;br&gt;
 Beginner- Users with Low RFM Score&lt;br&gt;
 Intermediate - Users with Average RFM Score&lt;br&gt;
 Expert - Users with High RFM Score&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;1. Show the Number of users per ranking&lt;/strong&gt;&lt;/p&gt;

&lt;p&gt;&lt;code&gt;plt.pie([len(typeBeginners),len(typeInter),len(typePro)],labels=myLabels,autopct='%1.1f%%')&lt;br&gt;
plt.axis('equal')&lt;br&gt;
plt.legend(labels=myLabels)&lt;br&gt;
plt.show()&lt;/code&gt;&lt;/p&gt;

&lt;p&gt;&lt;a href="https://res.cloudinary.com/practicaldev/image/fetch/s--4v4eR51o--/c_limit%2Cf_auto%2Cfl_progressive%2Cq_auto%2Cw_800/https://dev-to-uploads.s3.amazonaws.com/uploads/articles/e56j012ufnz9ra5w45qz.png" class="article-body-image-wrapper"&gt;&lt;img src="https://res.cloudinary.com/practicaldev/image/fetch/s--4v4eR51o--/c_limit%2Cf_auto%2Cfl_progressive%2Cq_auto%2Cw_800/https://dev-to-uploads.s3.amazonaws.com/uploads/articles/e56j012ufnz9ra5w45qz.png" alt="Pie chart to show users distribution" width="701" height="621"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;2. Show user distribution across major cities&lt;/strong&gt;&lt;br&gt;
&lt;/p&gt;

&lt;div class="highlight js-code-highlight"&gt;
&lt;pre class="highlight plaintext"&gt;&lt;code&gt;def getType(myType,myCity) :
    return len(data.loc[(data["Rank"]==myType) &amp;amp; (data["Location"]==myCity)])

myCityStats={}
for i in myLabels:
    for j in locationLabels:
        myCityStats[i]= [getType(i,j) for j in locationLabels]

print(myCityStats)

x=np.arange(len(locationLabels))
width = 0.3
multiplier = 0

fig,ax = plt.subplots(layout='constrained')

for att,val in myCityStats.items():
    offset = width*multiplier
    bars = ax.bar(x+offset,val,width,label=att)
    ax.bar_label(bars,padding=3)
    multiplier+=1


# Add some text for labels, title and custom x-axis tick labels, etc.
ax.set_ylabel('Length (mm)')
ax.set_title('Number of users in a city per RFM Rank')
ax.set_xticks(x + width, locationLabels)
ax.legend(loc='upper left', ncols=3)
ax.set_ylim(0, 250)

plt.show()


&lt;/code&gt;&lt;/pre&gt;

&lt;/div&gt;



&lt;p&gt;&lt;em&gt;Group bar charts for Rank distribution in Major Cities&lt;/em&gt;&lt;br&gt;
&lt;a href="https://res.cloudinary.com/practicaldev/image/fetch/s--edCehK8M--/c_limit%2Cf_auto%2Cfl_progressive%2Cq_auto%2Cw_800/https://dev-to-uploads.s3.amazonaws.com/uploads/articles/vrv0q7ryqozjo1rmoxf5.png" class="article-body-image-wrapper"&gt;&lt;img src="https://res.cloudinary.com/practicaldev/image/fetch/s--edCehK8M--/c_limit%2Cf_auto%2Cfl_progressive%2Cq_auto%2Cw_800/https://dev-to-uploads.s3.amazonaws.com/uploads/articles/vrv0q7ryqozjo1rmoxf5.png" alt="Group bar charts for Rank distribution in Major Cities" width="800" height="469"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;This are just but a few visualizations more can be done based on the questions you focus on answering based on your organization&lt;/p&gt;

</description>
      <category>datascience</category>
      <category>dataanalysis</category>
      <category>python</category>
      <category>pandas</category>
    </item>
    <item>
      <title>DATA EXPLORATORY ANALYSIS</title>
      <dc:creator>Isaac</dc:creator>
      <pubDate>Tue, 10 Oct 2023 08:39:40 +0000</pubDate>
      <link>https://dev.to/130/data-exploratory-analysis-54ol</link>
      <guid>https://dev.to/130/data-exploratory-analysis-54ol</guid>
      <description>&lt;p&gt;Just as the name might suggest ,Data exploratory analysis is a method of data analysis that involves the highlighting of key insights from a dataset and incorporating data visualizing tools to display them.&lt;/p&gt;

&lt;p&gt;Most datasets you will find you need to work on will need to be worked on before you can actually use them for the data analysis itself. This will mainly include the removal of row or columns with missing values, removal of duplicate value and other necessary data cleaning practices needed. First we will explore the pre-EDA activities and then the  post-EDA activities.&lt;/p&gt;

&lt;h2&gt;
  
  
  &lt;strong&gt;PRE-EDA Activities&lt;/strong&gt;
&lt;/h2&gt;

&lt;p&gt;&lt;strong&gt;1. Data Cleaning&lt;/strong&gt;&lt;br&gt;
   Raw data is mainly 'raw data' and needed to be 'cooked' before its actually ready for consumption :) . Data analysis is actually 80% data cleaning and data analysis  that means that this is the most crucial step and need lots of attention so as to make sure the data being used is free from anomalies. Just as stated earlier data cleaning mainly involves the removal or correction of  inaccurate records. The major data cleaning practices include:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;&lt;p&gt;&lt;strong&gt;Parsing&lt;/strong&gt;-&amp;gt; Converting data to the required format acceptable to the application in use&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;&lt;strong&gt;Duplicate Elimination&lt;/strong&gt; -&amp;gt; Removal of repeating entries of a key in the dataset&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;&lt;strong&gt;Removal or Updating of missing values&lt;/strong&gt; -&amp;gt; Some entries might be blank and may need to be removed or filled systematically to ensure a true representation of the collected dataset as much as possible. If the missing values are too many  you may consider removing them but if the number is minimal use various methods to provide them eg. use the median or the mode&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;&lt;strong&gt;Data Transformation&lt;/strong&gt; -&amp;gt; This involve reformatting hoe the data had been represented to away that can be used by the analytics tool.&lt;/p&gt;&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;Data cleaning is mainly done using major programming languages like Python using Numpy and Pandas , MS Excel, Tableau and many more.&lt;/p&gt;

&lt;h2&gt;
  
  
  &lt;strong&gt;POST EDA ACTIVITIES&lt;/strong&gt;
&lt;/h2&gt;

&lt;p&gt;When now using EDA the analysis can be graphical or non-graphical in that it may contain a table with an analysis of lets say measures of central tendency  such as mean, mode, median, or the quartiles of the data. Graphical ones can contains graphs, pie charts, histograms, line-graphs, funnels, scatter plots or any other graphical representation of the data.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;NON-GRAPHICAL EDA&lt;/strong&gt;&lt;br&gt;
&lt;a href="https://res.cloudinary.com/practicaldev/image/fetch/s--AOBRThAs--/c_limit%2Cf_auto%2Cfl_progressive%2Cq_auto%2Cw_800/https://dev-to-uploads.s3.amazonaws.com/uploads/articles/on93ou330nlm36spg3yz.png" class="article-body-image-wrapper"&gt;&lt;img src="https://res.cloudinary.com/practicaldev/image/fetch/s--AOBRThAs--/c_limit%2Cf_auto%2Cfl_progressive%2Cq_auto%2Cw_800/https://dev-to-uploads.s3.amazonaws.com/uploads/articles/on93ou330nlm36spg3yz.png" alt="Non-Graphical EDA" width="292" height="162"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;Graphical EDA&lt;/strong&gt;&lt;br&gt;
&lt;a href="https://res.cloudinary.com/practicaldev/image/fetch/s--Mpc-MhBp--/c_limit%2Cf_auto%2Cfl_progressive%2Cq_auto%2Cw_800/https://dev-to-uploads.s3.amazonaws.com/uploads/articles/zknq0zdautlk696pv8zm.png" class="article-body-image-wrapper"&gt;&lt;img src="https://res.cloudinary.com/practicaldev/image/fetch/s--Mpc-MhBp--/c_limit%2Cf_auto%2Cfl_progressive%2Cq_auto%2Cw_800/https://dev-to-uploads.s3.amazonaws.com/uploads/articles/zknq0zdautlk696pv8zm.png" alt="Graphical EDA" width="235" height="214"&gt;&lt;/a&gt;&lt;/p&gt;

</description>
    </item>
  </channel>
</rss>
