<?xml version="1.0" encoding="UTF-8"?>
<rss version="2.0" xmlns:atom="http://www.w3.org/2005/Atom" xmlns:dc="http://purl.org/dc/elements/1.1/">
  <channel>
    <title>DEV Community: Judy</title>
    <description>The latest articles on DEV Community by Judy (@kesh).</description>
    <link>https://dev.to/kesh</link>
    <image>
      <url>https://media2.dev.to/dynamic/image/width=90,height=90,fit=cover,gravity=auto,format=auto/https:%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Fuser%2Fprofile_image%2F1173616%2Fd6ea12f1-07a2-4c23-81a5-31cc18adf0c8.png</url>
      <title>DEV Community: Judy</title>
      <link>https://dev.to/kesh</link>
    </image>
    <atom:link rel="self" type="application/rss+xml" href="https://dev.to/feed/kesh"/>
    <language>en</language>
    <item>
      <title>Data Engineering for Beginners: A Step-by-Step Guide</title>
      <dc:creator>Judy</dc:creator>
      <pubDate>Mon, 06 Nov 2023 10:00:21 +0000</pubDate>
      <link>https://dev.to/kesh/data-engineering-for-beginners-a-step-by-step-guide-2dg1</link>
      <guid>https://dev.to/kesh/data-engineering-for-beginners-a-step-by-step-guide-2dg1</guid>
      <description>&lt;p&gt;Data engineering has become critical to the data ecosystem due to the influx of massive amounts of data from a variety of sources. Organizations are also looking to establish and expand their data engineering teams. &lt;br&gt;
Some data professions, such as analyst, do not require prior expertise in the industry if you have excellent SQL and programming skills. However, prior knowledge in data analytics or software engineering is often beneficial for breaking into data engineering.&lt;/p&gt;

&lt;h2&gt;
  
  
  What is Data engineering?
&lt;/h2&gt;

&lt;p&gt;Data engineering is a subfield of data science concerned with the practical applications of data analysis and acquisition. Like other areas of engineering, it is concerned with the application of data science in the actual world. &lt;br&gt;
Data engineering has nothing to do with experimental design. It is more concerned with establishing systems for improved information flow and access. &lt;/p&gt;

&lt;h2&gt;
  
  
  What Does a Data Engineer Do?
&lt;/h2&gt;

&lt;p&gt;A data engineer is responsible for creating and maintaining data architectures (such as a database). They are in charge of data collecting and the processing of raw data into useable data. &lt;br&gt;
You can't collect data until you have a data engineer. Companies demand data engineers to be knowledgeable with SQL, Java, AWS, Scala, and other programming languages. &lt;br&gt;
A background in backend development or programming is required for data engineering.&lt;br&gt;
As a data engineer, you'll be responsible for managing data collection, storage, and processing for future use.&lt;/p&gt;

&lt;h2&gt;
  
  
  Data Engineering Concepts
&lt;/h2&gt;

&lt;p&gt;&lt;strong&gt;Data Sources and Types&lt;/strong&gt;&lt;/p&gt;

&lt;p&gt;The data coming from these sources can be classified into one of the three broad categories:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;&lt;p&gt;&lt;strong&gt;Unstructured data&lt;/strong&gt;&lt;br&gt;
Lacks a well-defined schema. For example; Images, videos and other multimedia files, website data&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;&lt;strong&gt;Semi-structured data&lt;/strong&gt;&lt;br&gt;
Has some structure but no rigid schema. Typically has metadata tags that provide additional information. For example; JSON,XML data, emails, and zip files.&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;&lt;strong&gt;Structured data&lt;/strong&gt;&lt;br&gt;
Has a well-defined schema. For example; spreadsheets.&lt;/p&gt;&lt;/li&gt;
&lt;/ul&gt;

&lt;h2&gt;
  
  
  Data Repositories: Data Warehouses, Data Lakes, and Data Marts
&lt;/h2&gt;

&lt;p&gt;The raw data collected from various sources is staged in a suitable repository.&lt;br&gt;
There are two data processing systems, OLTP and OLAP systems:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;&lt;p&gt;OLTP or Online Transactional Processing systems are used to store day-to-day operational data for applications such as inventory management. OLTP systems include relational databases that store data that can be used for analysis and deriving business insights. &lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;OLAP or Online Analytical Processing systems are used to store large volumes of historical data for carrying out complex analytics. In addition to databases, OLAP systems also include data warehouses and data lakes (more on this shortly).&lt;/p&gt;&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;The source and type of data frequently influence the choice of data store.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;- Data warehouses&lt;/strong&gt;: A data warehouse is a centralized repository for receiving data.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;- Data lakes&lt;/strong&gt;: Data lakes enable the storage of all data kinds in their raw form, including semi-structured and unstructured data. ELT processes (which we will explain subsequently) frequently end up in data lakes.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;- Data mart&lt;/strong&gt;: A data mart is a smaller subset of a data warehouse that is targeted to a certain business use case. &lt;/p&gt;

&lt;p&gt;&lt;strong&gt;- Data lake houses&lt;/strong&gt;: Recently, data lake homes have gained popularity because they combine the freedom of data lakes with the structure and organization of data warehouses.&lt;/p&gt;

&lt;h2&gt;
  
  
  Data Pipelines: ETL and ELT Processes
&lt;/h2&gt;

&lt;p&gt;Data pipelines encompass the data's travel from source to destination systems via ETL and ELT operations. &lt;/p&gt;

&lt;p&gt;&lt;strong&gt;The ETL&lt;/strong&gt; (Extract, Transform, and Load) process consists of the following steps:&lt;br&gt;
Data extraction from several sources &lt;br&gt;
Clean, validate, and standardize data before transforming it.&lt;br&gt;
Load the data into a database or a destination application.&lt;br&gt;
The destination of ETL processes is frequently a data warehouse.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;ELT&lt;/strong&gt;(Extract, Load, and Transform) is a variant of the ETL process in which the phases are reversed: extract, load, and transform instead of extract, transform, and load.&lt;br&gt;
That is, before any modification is conducted, the raw data gathered from the source is loaded into the data repository. This enables us to apply modifications tailored to a certain application. Data lakes are the final endpoint of ELT procedures.&lt;/p&gt;

&lt;h2&gt;
  
  
  Tools Data Engineers Should Know
&lt;/h2&gt;

&lt;ul&gt;
&lt;li&gt;&lt;p&gt;Dbt (data build tool) for analytics engineering&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;Apache Spark for big data analytics and distributed data processing framework.&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;Airflow for data pipeline orchestration.&lt;/p&gt;&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;-Cloud computing fundamentals and experience working with at least one cloud provider, such as AWS or Microsoft Azure.&lt;/p&gt;

&lt;p&gt;In conclusion, Data engineering is a vast field. And there is a high need for persons with this skill set. It only takes one step, so begin your learning adventure right away.&lt;/p&gt;

</description>
      <category>dataengineering</category>
      <category>datascience</category>
      <category>data</category>
      <category>datastructures</category>
    </item>
    <item>
      <title>The Complete Guide to Time Series Models.</title>
      <dc:creator>Judy</dc:creator>
      <pubDate>Fri, 27 Oct 2023 18:57:56 +0000</pubDate>
      <link>https://dev.to/kesh/the-complete-guide-to-time-series-models-4e13</link>
      <guid>https://dev.to/kesh/the-complete-guide-to-time-series-models-4e13</guid>
      <description>&lt;p&gt;Time series data is everywhere in our lives. It can be found in almost any domain: monitoring, sensors, stock prices, weather forecasts, exchange rates, application performance, and a plethora of other metrics on which we rely in our professional and personal life.&lt;/p&gt;

&lt;h2&gt;
  
  
  What is time series data?
&lt;/h2&gt;

&lt;p&gt;Time-series analysis refers to the technique and mathematical tools used to examine time-series data in order to discover not only what happened but also when and why it happened, as well as what is most likely to happen in the future.&lt;/p&gt;

&lt;h2&gt;
  
  
  ‌‌Types of Time-Series Analysis
&lt;/h2&gt;

&lt;ul&gt;
&lt;li&gt;&lt;p&gt;&lt;strong&gt;Exploratory analysis&lt;/strong&gt;&lt;br&gt;
When you wish to describe what you observe and why you see it in a given time series, an exploratory analysis can help. It requires breaking down the data into trends, seasonality, cyclicity, and abnormalities. ‌‌&lt;br&gt;
We can explain what each component symbolizes in the real world and, maybe, what caused it once the series has been deconstructed. This is not as simple as it appears and frequently involves spectrum decomposition to identify any specific frequency of recurrences and autocorrelation analysis to determine whether present values are dependent on prior values.&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;&lt;strong&gt;Curve fitting&lt;/strong&gt;&lt;br&gt;
Because time series is a discrete set, you can always tell how many data points are in it.&lt;br&gt;
But what if you want to know the value of a time-series parameter at a point in time that your data does not cover? &lt;/p&gt;&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;To address this question, we need to add a continuous set—a curve—to our data. This can be accomplished in a variety of methods, including interpolation and regression. The former is a perfect match for parts of the given time series and is mostly useful for guessing missing data points. The latter, on the other hand, is a "best-fit" curve, which requires you to make an informed guess about the form of the function to be fitted (e.g., linear) and then modify the parameters until your best-fit criteria are met. &lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;&lt;p&gt;&lt;strong&gt;Forecasting&lt;/strong&gt;&lt;br&gt;
The process of generalization from sample to whole is known as statistical inference. It can be done over time with time-series data, allowing for future predictions or forecasting: from extrapolating regression models to more complex techniques involving stochastic simulations and machine learning.&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;&lt;strong&gt;Classification and segmentation&lt;/strong&gt;&lt;br&gt;
Classification is the process of identifying patterns in a series and assigning them to one of several classes. Segmentation, on the other hand, is the process of dividing a time series into a number of segments based on some specified criterion. ‌‌‌‌‌‌&lt;/p&gt;&lt;/li&gt;
&lt;/ul&gt;

&lt;h2&gt;
  
  
  Time series data visualization
&lt;/h2&gt;

&lt;p&gt;Time series data visualization is often conducted with specialist tools that offer users a variety of visualization kinds and formats from which to pick. Let's look at some of the most popular data visualization methods.&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;&lt;p&gt;&lt;strong&gt;Time series graph&lt;/strong&gt;&lt;br&gt;
Time series graphs, also known as time series plots, are the most commonly used data visualization tool for illustrating data points at a temporal scale where each point corresponds to both time and the unit of measurement.&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;&lt;strong&gt;Real time graph&lt;/strong&gt;&lt;br&gt;
Time series data is displayed in real time using real time graphs, often known as data streaming charts. This means that a real-time graph will refresh automatically every few seconds or when a new data point is received from the server.&lt;/p&gt;&lt;/li&gt;
&lt;/ul&gt;

&lt;h2&gt;
  
  
  Data models used for time series data
&lt;/h2&gt;

&lt;ul&gt;
&lt;li&gt;&lt;strong&gt;Autoregressive (AR) models&lt;/strong&gt;&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;AR model is a representation of a form of random process, it is used to describe data reflecting time-varying processes such as changes in weather, economics, and so on.&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;&lt;strong&gt;Integrated (I) models&lt;/strong&gt;&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;Integrated models are made up of a series of random walk components. These series are called integrated because they are the summation of weakly steady components.&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;&lt;strong&gt;Moving-average (MA) models&lt;/strong&gt;&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;Univariate time series are modeled using moving-average models. The output variable in MA models is linearly dependent on the current and various historical values of an imperfectly predicted (stochastic) factor.&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;&lt;strong&gt;Autoregressive moving average (ARMA) models&lt;/strong&gt;&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;ARMA models combine the AR and MA classes, with the AR part regressing the variable on its own historical values and the MA part modeling the error term as a linear mixture of error terms happening concurrently and at different times in the past. ARMA models are commonly employed in analytics for forecasting future values in a series.&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;&lt;strong&gt;Autoregressive integrated moving average (ARIMA) models&lt;/strong&gt;&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;ARIMA models are a generalization of an ARMA model and are used when data show evidence of non-stationarity, where an initial differencing step, corresponding to the integrated part of the model, can be applied one or more times to eliminate the mean function's non-stationarity.&lt;/p&gt;

&lt;p&gt;Both the ARMA and ARIMA models are commonly employed for analytics and forecasting future values in a series.&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;&lt;strong&gt;Autoregressive fractionally integrated moving average (ARFIMA)&lt;/strong&gt;&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;ARFIMA models, in turn, generalize ARIMA models (or, more broadly, all three fundamental types) by allowing non-integer differencing parameter values. ARFIMA models are commonly used to simulate so-called long memory time series, in which deviations from the long-run mean dissipate more slowly than exponential decay.&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;&lt;strong&gt;Autoregressive conditional heteroscedasticity (ARCH)&lt;/strong&gt;&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;The ARCH model, for example, describes the variance of the present error term or innovation as a function of the actual sizes of error terms in earlier time periods.&lt;/p&gt;

&lt;h2&gt;
  
  
  Challenges in Handling Time-Series Data
&lt;/h2&gt;

&lt;p&gt;While time series data provides great insights, it also brings distinct obstacles that must be handled during analysis.&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;&lt;p&gt;&lt;strong&gt;Dealing with missing values&lt;/strong&gt;&lt;br&gt;
Time-series data frequently contains missing or partial values, which can impair analysis and modeling accuracy. Depending on the nature of the data and the level of missing values, several techniques such as interpolation or imputation can be used to handle missing values.&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;&lt;strong&gt;Overcoming noise in time-series data&lt;/strong&gt;&lt;br&gt;
Noise in time series data refers to random fluctuations or anomalies that can conceal underlying patterns and trends. Moving averages and wavelet treatments, for example, can assist minimize noise and extract the most important information from data.&lt;/p&gt;&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;In conclusion, time series analysis attempts to understand how patterns develop over time. These patterns aid in the generation of exact estimates for things like future sales, GDP, and global temperatures.&lt;br&gt;
One thing to keep in mind is that time series models take into account the fact that time flows in only one direction.&lt;br&gt;
Events that are near in time often have a greater link than more distant discoveries.&lt;br&gt;
Time-series data, like all data, has random oscillations. This randomness has the potential to hide the underlying patterns. Smoothing techniques help to wipe out these swings, revealing the trends and cycles more clearly.&lt;br&gt;
‌‌‌‌‌‌&lt;/p&gt;

</description>
      <category>beginners</category>
      <category>data</category>
      <category>datascience</category>
    </item>
    <item>
      <title>How to use python in data visualization for credit risk assessment.</title>
      <dc:creator>Judy</dc:creator>
      <pubDate>Thu, 19 Oct 2023 08:07:42 +0000</pubDate>
      <link>https://dev.to/kesh/how-to-use-python-in-data-visualization-for-credit-risk-assessment-2k89</link>
      <guid>https://dev.to/kesh/how-to-use-python-in-data-visualization-for-credit-risk-assessment-2k89</guid>
      <description>&lt;p&gt;Most individuals rely on credit to finance vehicles, real estate, student loans, and the start-up of small enterprises. Assessing credit risk data is crucial for financial institutions when deciding whether to offer the loans.&lt;/p&gt;

&lt;p&gt;Dataset used for credit assessment was sourced from &lt;a href="https://www.kaggle.com/datasets?fileType=csv"&gt;kaggle.com.com&lt;/a&gt;. Therefore proceed to load the relevant libraries in python which will be used for credit risk assessment.&lt;br&gt;
&lt;/p&gt;

&lt;div class="highlight js-code-highlight"&gt;
&lt;pre class="highlight plaintext"&gt;&lt;code&gt;import pandas as pd
import numpy as np
import matplotlib.pyplot as plt
import seaborn as sns
&lt;/code&gt;&lt;/pre&gt;

&lt;/div&gt;



&lt;p&gt;Data is then loaded on python as csv file;&lt;br&gt;
&lt;/p&gt;

&lt;div class="highlight js-code-highlight"&gt;
&lt;pre class="highlight plaintext"&gt;&lt;code&gt;data = pd.read_csv('credit_risk_dataset_test.csv')
&lt;/code&gt;&lt;/pre&gt;

&lt;/div&gt;



&lt;p&gt;After the data is loaded, data cleaning is done next to format any inconsistencies and ensure the data is in order to avoid any errors.&lt;br&gt;
First check the data types for the columns as&lt;br&gt;
&lt;/p&gt;

&lt;div class="highlight js-code-highlight"&gt;
&lt;pre class="highlight plaintext"&gt;&lt;code&gt;print(data.dtypes)
&lt;/code&gt;&lt;/pre&gt;

&lt;/div&gt;



&lt;p&gt;In this data set there are a mix types of data which may make data manipulation a bit hard hence convert integers to float. I prefer using floats as it allows me to represent data on plots accurately and ensures compatibility with various libraries.&lt;br&gt;
&lt;/p&gt;

&lt;div class="highlight js-code-highlight"&gt;
&lt;pre class="highlight plaintext"&gt;&lt;code&gt;data['person_income'] = data['person_income'].astype(str)
data['loan_amount'] = data['loan_amount'].astype(str)
data['loan_int_rate'] = data['loan_int_rate'].astype(str)
data['debt_to_income_ratio'] = data['debt_to_income_ratio'].astype(str)
&lt;/code&gt;&lt;/pre&gt;

&lt;/div&gt;



&lt;p&gt;Once the data is cleaned, data visualisation is next. Histograms will be used to show visual representation of ages. In this data set, shows most people who have loans range 20-40 years. Highest number are in their 20s.&lt;br&gt;
&lt;/p&gt;

&lt;div class="highlight js-code-highlight"&gt;
&lt;pre class="highlight plaintext"&gt;&lt;code&gt;plt.figure(figsize=(8, 6))
sns.histplot(data['person_age'], bins=20, kde=True)
plt.title('Distribution of Person Age')
plt.xlabel('Age')
plt.ylabel('Frequency')
plt.show()
&lt;/code&gt;&lt;/pre&gt;

&lt;/div&gt;



&lt;p&gt;Next, creation of boxplot to visualize the distribution of loan amounts by loan grade.Each box in the plot represents a specific loan grade, and it shows the distribution of loan amounts for that grade. The boxplot provides information about the median, quartiles, and any potential outliers in the data for each loan grade, making it a useful tool for understanding the distribution of loan amounts across different grades. Grade F loans have a median income that is superior to other grades.&lt;br&gt;
&lt;/p&gt;

&lt;div class="highlight js-code-highlight"&gt;
&lt;pre class="highlight plaintext"&gt;&lt;code&gt;plt.figure(figsize=(10, 6))
sns.boxplot(data=data, x='loan_grade', y='loan_amount')
plt.title('Loan Amount Distribution by Loan Grade')
plt.xlabel('Loan Grade')
plt.ylabel('Loan Amount')
plt.show()
&lt;/code&gt;&lt;/pre&gt;

&lt;/div&gt;



&lt;p&gt;The next step,use of scatter plots to provide insights into the relationship between a borrower's debt-to-income ratio, interest rate, and whether they defaulted on their loan. A higher debt-to-income ratio indicates that the borrower has a larger proportion of their income committed to debt payments. In the data set, loans with lower interest rate, lower debt to income ratio have not defaulted while loans with higher interest rate have defaulted.&lt;br&gt;
&lt;/p&gt;

&lt;div class="highlight js-code-highlight"&gt;
&lt;pre class="highlight plaintext"&gt;&lt;code&gt;plt.figure(figsize=(8, 6))
sns.scatterplot(data=data, x='debt_to_income_ratio', y='loan_int_rate', hue='cb_person_default_on_file')
plt.title('Debt-to-Income Ratio vs. Interest Rate')
plt.xlabel('Debt-to-Income Ratio')
plt.ylabel('Interest Rate')
plt.legend(title='Default')
plt.show()
&lt;/code&gt;&lt;/pre&gt;

&lt;/div&gt;



&lt;p&gt;Count plots of homeownership will be done to provide a visual representation of different types of home ownership among borrowers. In this data set the "RENT" is the most common home ownership type among borrowers, followed by "MORTGAGE." The "OWN" category has the fewest borrowers.&lt;br&gt;
This plot helps one understand the characteristics of borrowers and can help identify potential factors that impact credit risk.&lt;br&gt;
In this case, most of the borrowers are renters, followed by those with mortgage. The least borrowers are home owners&lt;br&gt;
&lt;/p&gt;

&lt;div class="highlight js-code-highlight"&gt;
&lt;pre class="highlight plaintext"&gt;&lt;code&gt;plt.figure(figsize=(8, 6))
sns.countplot(data=data, x='person_home_ownership')
plt.title('Count of Home Ownership Types')
plt.xlabel('Home Ownership')
plt.ylabel('Count')
plt.xticks(rotation=45)
plt.show()
&lt;/code&gt;&lt;/pre&gt;

&lt;/div&gt;



&lt;p&gt;Lastly, relationship between person income and loan amount is established by using scatterplot. In the data set, most borrowers are low income earners and have different intent use for the loans.&lt;br&gt;
&lt;/p&gt;

&lt;div class="highlight js-code-highlight"&gt;
&lt;pre class="highlight plaintext"&gt;&lt;code&gt;plt.figure(figsize=(10, 6))
sns.scatterplot(x='person_income', y='loan_amount', data=data, hue='loan_intent', palette='Dark2')
plt.xlabel('Person Income')
plt.ylabel('Loan Amount')
plt.title('Person Income vs. Loan Amount')
plt.show()
&lt;/code&gt;&lt;/pre&gt;

&lt;/div&gt;



&lt;p&gt;Python being an open source gives one access to different libraries that enables you to handle large data sets and easily customize to what you prefer.&lt;/p&gt;

&lt;p&gt;To view the output for the above, visit &lt;a href="https://github.com/K-erubo/How-to-use-python-for-data-visualization-in-credit-risk-assessment./blob/main/Data%20visualization%20in%20credit%20risk%20assessment.ipynb"&gt;github.com&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;Thanks for reading!&lt;/p&gt;

</description>
      <category>python</category>
      <category>data</category>
      <category>analytics</category>
    </item>
    <item>
      <title>Exploratory Data Analysis using Data Visualization Techniques.</title>
      <dc:creator>Judy</dc:creator>
      <pubDate>Tue, 10 Oct 2023 18:19:16 +0000</pubDate>
      <link>https://dev.to/kesh/exploratory-data-analysis-using-data-visualization-techniques-3pdn</link>
      <guid>https://dev.to/kesh/exploratory-data-analysis-using-data-visualization-techniques-3pdn</guid>
      <description>&lt;h2&gt;
  
  
  What is Exploratory Data Analysis?
&lt;/h2&gt;

&lt;p&gt;Exploratory data analysis is a process used to analyze and summarize datasets.&lt;br&gt;
Using various statistical and graphical tools, EDA tries to find patterns, detect anomalies, test hypotheses, and validate assumptions. It is an important part of the data science process because it allows analysts to obtain a thorough grasp of their data before going on to more advanced modeling and machine learning tasks.&lt;/p&gt;

&lt;h2&gt;
  
  
  Steps involved in Exploratory Data analysis
&lt;/h2&gt;

&lt;ol&gt;
&lt;li&gt;&lt;p&gt;Data Collection&lt;br&gt;
The data that will be evaluated is collected as the first step in the EDA process. Data can be obtained from a variety of sources, including structured databases, APIs, and even web scraping. To ensure compatibility with the analysis tools you intend to employ, it is critical to understand the many types of data sources, as well as their formats and architectures.&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;Data Cleaning&lt;br&gt;
Once the data has been acquired, it must be cleaned and preprocessed to assure its quality and dependability. Handling missing numbers, deleting duplicates, changing data types, and detecting and addressing outliers are all examples of this process.&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;Data Exploration&lt;br&gt;
Data exploration is the process of studying a dataset using various statistical and graphical approaches in order to discover the underlying structure, relationships, and trends in the data. This process is divided into three parts: univariate, bivariate, and multivariate analysis.&lt;/p&gt;&lt;/li&gt;
&lt;/ol&gt;

&lt;h2&gt;
  
  
  Data Visualization
&lt;/h2&gt;

&lt;p&gt;Data visualization facilitates the efficient communication of insights and patterns identified throughout the exploratory data analysis process. Choosing the correct style of visualization, following best practices, and utilizing popular visualization libraries can all help to increase the impact of your study.&lt;/p&gt;

&lt;h2&gt;
  
  
  Choosing the Right Type of Visualization
&lt;/h2&gt;

&lt;p&gt;The suitable visualization is determined by the nature of the data and the insights you wish to express. Bar charts, line charts, pie charts, scatter plots, and heatmaps are examples of common visualization types. Each of these visualizations has a specific function, such as comparing categories, exhibiting trends through time, or demonstrating relationships between variables.&lt;/p&gt;

&lt;h2&gt;
  
  
  Visualization Techniques Used for Exploratory Data Analysis.
&lt;/h2&gt;

&lt;p&gt;Several visualization tools and techniques are in use. Here are the mostly used for the same.&lt;/p&gt;

&lt;h2&gt;
  
  
  &lt;strong&gt;Univariate Analysis&lt;/strong&gt;
&lt;/h2&gt;

&lt;p&gt;Univariate analysis examines the distribution, central tendency, and dispersion of a single variable. This study aids in comprehending the unique properties of each variable in the dataset.&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;&lt;p&gt;&lt;strong&gt;Histograms&lt;/strong&gt;&lt;br&gt;
Histograms are graphical representations of a variable's distribution that divide data points into bins based on their values. Histograms aid in the identification of the distribution's form, any gaps or clusters, and probable outliers. &lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;&lt;strong&gt;Box plots&lt;/strong&gt;&lt;br&gt;
Box plots show the quartile distribution of a variable, showing the median, interquartile range (IQR), and any outliers. Box plots show the dispersion of a variable in a compact manner.&lt;/p&gt;&lt;/li&gt;
&lt;/ul&gt;

&lt;h2&gt;
  
  
  &lt;strong&gt;Bivariate Analysis&lt;/strong&gt;
&lt;/h2&gt;

&lt;p&gt;Involves examining the relationship between two variables, exploring potential correlations, trends, or patterns between them.&lt;br&gt;
Below are bivalents plots used for EDA;&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;&lt;p&gt;&lt;strong&gt;Correlation plots or Heatmaps&lt;/strong&gt;&lt;br&gt;
Heatmaps use color intensity to depict the strength of a relationship between many variables in a matrix format. This visualization aids in quickly recognizing groups of similar variables.&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;&lt;strong&gt;Bar Graphs&lt;/strong&gt;&lt;br&gt;
They are used to compare nominal or ordinal data. They are helpful for recognizing trends.&lt;/p&gt;&lt;/li&gt;
&lt;/ul&gt;

&lt;h2&gt;
  
  
  &lt;strong&gt;Multivariate Analysis&lt;/strong&gt;
&lt;/h2&gt;

&lt;p&gt;Multivariate analysis examines the relationships among three or more variables simultaneously, providing a more holistic view of the data.&lt;br&gt;
Below are multivariate analysis techniques&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;&lt;p&gt;&lt;strong&gt;Multiple linear regression&lt;/strong&gt;&lt;br&gt;
A dependence method that examines the relationship between one dependent variable and two or more independent variables. A multiple regression model will tell you how well each independent variable correlates with the dependent variable. This is useful because it allows you to forecast future outcomes by understanding which elements are likely to impact a specific event.&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;&lt;strong&gt;Multiple logistic regression&lt;/strong&gt;&lt;br&gt;
The chance of a binary event occurring is calculated (and predicted) using logistic regression analysis. A binary result has only two possible outcomes: the event occurs (1) or it does not occur (0). So, logistic regression can forecast the likelihood of a given situation based on a set of independent factors. It is also employed in classification. &lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;&lt;strong&gt;Multivariate analysis of variance (MANOVA)&lt;/strong&gt;&lt;br&gt;
MANOVA is a statistical method for determining the effect of many independent factors on two or more dependent variables. It's vital to remember that the independent variables in MANOVA are categorical, whereas the dependent variables are metric.&lt;br&gt;
In MANOVA analysis, you look at different combinations of independent variables to see how they differ in their influence on the dependent variable.&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;&lt;strong&gt;Factor analysis&lt;/strong&gt;&lt;br&gt;
Factor analysis is an interdependent technique for reducing the number of variables in a dataset. Finding patterns in data might be tough if you have too many variables. Simultaneously, models built with too many variables are vulnerable to overfitting. Overfitting is a modeling error that occurs when a model fits a dataset too closely and specifically, making it less generalizable to future datasets and potentially less accurate in its predictions.&lt;br&gt;
Factor analysis works by identifying groups of variables that have a high correlation with one another. These variables can then be merged to form a single variable. Factor analysis is frequently used by data analysts to prepare data for further examination.&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;&lt;strong&gt;Cluster analysis&lt;/strong&gt;&lt;br&gt;
Cluster analysis is used to group similar items within a dataset into clusters. When grouping data into clusters, the aim is for the variables in one cluster to be more similar to each other than they are to variables in other clusters. &lt;br&gt;
Cluster analysis helps you to understand how data in your sample is distributed, and to find patterns.&lt;/p&gt;&lt;/li&gt;
&lt;/ul&gt;

&lt;h2&gt;
  
  
  Exploratory Data Analysis tools and libraries
&lt;/h2&gt;

&lt;h2&gt;
  
  
  Python Libraries
&lt;/h2&gt;

&lt;p&gt;Commonly used python libraries are:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;&lt;p&gt;&lt;strong&gt;Pandas&lt;/strong&gt;&lt;br&gt;
Referred as pd. It used for data analysis and manipulation&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;&lt;strong&gt;Numpy&lt;/strong&gt;&lt;br&gt;
Referred as np. It is used for numerical computing, offering support for arrays and mathematical functions.&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;&lt;strong&gt;Matplotlib&lt;/strong&gt;&lt;br&gt;
Referred as plt. It used for creating static, interactive, and animated visualizations.&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;&lt;strong&gt;Seaborn&lt;/strong&gt;&lt;br&gt;
Referred as sns. It is a statistical data visualization library based on Matplotlib, providing a high-level interface for creating informative and attractive visualizations.&lt;/p&gt;&lt;/li&gt;
&lt;/ul&gt;

&lt;h2&gt;
  
  
  R Libraries
&lt;/h2&gt;

&lt;p&gt;R is another popular programming language for data analysis that has a robust ecosystem of EDA modules. R libraries that are regularly used include:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;&lt;p&gt;&lt;strong&gt;dplyr&lt;/strong&gt;&lt;br&gt;
A data manipulation package that provides a consistent collection of methods for filtering, sorting, and aggregating data.&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;&lt;strong&gt;ggplot2&lt;/strong&gt;&lt;br&gt;
A sophisticated and versatile data visualization library based on the Grammar of Graphics, allowing complex visualizations to be created with minimal code.&lt;/p&gt;&lt;/li&gt;
&lt;/ul&gt;

&lt;h2&gt;
  
  
  Data Visualization Tools
&lt;/h2&gt;

&lt;p&gt;Aside from programming languages and libraries, EDA use a variety of data visualization tools. These tools provide an easier-to-use interface for producing and customizing visualizations. Among the most popular data visualization tools are:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;&lt;p&gt;&lt;strong&gt;Tableau&lt;/strong&gt;&lt;br&gt;
A popular data visualization software that enables users to create interactive and shareable dashboards.&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;&lt;strong&gt;Power BI&lt;/strong&gt;&lt;br&gt;
A Microsoft business analytics service that provides data visualization and reporting capabilities.&lt;/p&gt;&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;In conclusion, EDA is an important step in the data analytics process because it allows data scientists to understand the features of the data, uncover patterns and relationships, and make educated judgments in following analysis or modeling stages. Data scientists may extract useful insights from their data and construct more effective and interpretable machine learning models by utilizing various EDA approaches and tools.&lt;/p&gt;

&lt;p&gt;Always keep in mind, EDA is an iterative and exploratory process, and that constantly improving your research will result in a greater knowledge and more robust findings. You will develop a deeper understanding for the data and its underlying patterns as you gain expertise in EDA, making you a more effective data scientist and analyst.&lt;/p&gt;

</description>
      <category>dataanalysis</category>
      <category>visualization</category>
      <category>python</category>
      <category>datascience</category>
    </item>
    <item>
      <title>Data Science for Beginners: 2023 - 2024 Complete Roadmap</title>
      <dc:creator>Judy</dc:creator>
      <pubDate>Sat, 30 Sep 2023 08:52:47 +0000</pubDate>
      <link>https://dev.to/kesh/data-science-for-beginners-2023-2024-complete-roadmap-2b3i</link>
      <guid>https://dev.to/kesh/data-science-for-beginners-2023-2024-complete-roadmap-2b3i</guid>
      <description>&lt;p&gt;Data science is an emerging field that has blended statistical, programming, and domain expertise to get insights from data.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;What is Data Science?&lt;/strong&gt;&lt;br&gt;
Data science is a field of study that involves extracting knowledge and insights from large and complex datasets using techniques like data mining, statistical analysis, machine learning and visualization.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;Who is a Data Scientist?&lt;/strong&gt;&lt;br&gt;
A person responsible for collecting, cleaning and analyzing large data sets to extract valuable insights. They work with unstructured, semi structured and structured data to find patterns that help them solve problems and predict future events.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;Why Data Science?&lt;/strong&gt;&lt;br&gt;
The pace of change globally is really rapid. There are new technologies being introduced every day. To be competitive in such a situation, businesses must employ innovative strategies. Big data and data science are now important instruments for corporate growth as a result of this.&lt;br&gt;
Data science is applicable to all industries, but most significant in those that generate enormous amounts of raw data on a regular basis, such as healthcare, retail, and finance. &lt;br&gt;
Companies are now collecting more data on their customers to help them gain insights into their customer behaviors and preferences that aid in making better decisions.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;How easy is Data science for beginners?&lt;/strong&gt;&lt;br&gt;
Data science is an exciting and rewarding field for a beginner to learn however it has some challenges;&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;&lt;p&gt;Understanding the mathematical and statistical concepts that underpin many of the approaches used in data science like probability, statistics, linear algebra, and optimization can be difficult to grasp, especially for people who have never studied them before.&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;Learning the numerous tools and technology used in data science including programming languages like Python and R, as well as libraries and frameworks like NumPy, Pandas, and sci-kit-learn. These tools can be overwhelming for those who are new to programming and are unsure how to begin with data science. Programming language tools necessitate familiarity with concepts such as variables, functions, and loops.&lt;/p&gt;&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;Despite these challenges,beginner in data science can still succeed by being persistent, curious, eager to learn and practice. With time they become proficient at it.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;How do you start learning Data science as a beginner?&lt;/strong&gt;&lt;br&gt;
As a beginner, learning data science entails learning the essential tools and technologies, comprehending the underlying concepts, and applying and implementing what you have learned. You may create a good foundation in data science and become adept in the discipline with perseverance and devotion. So, if you're not sure where to start learning data science, here's a step-by-step data science roadmap for beginners to get you started.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;1.Learning Fundamentals and SQL&lt;/strong&gt;&lt;br&gt;
If you are a beginner with no background in statistics or mathematics, you can familiarize yourself with probability and statistical concepts.&lt;br&gt;
You then can proceed to learn a query language such as SQL. SQL (Structured Query Language) is a database management and manipulation programming language. It is an essential ability for any data scientist since it helps you to obtain, filter, and combine data from a variety of sources. SQL learning tools for beginners include online classes, tutorials, and textbooks. You can also hone your skills by completing SQL exercises and projects. You can go to the following phase once you have a solid foundation in SQL. &lt;/p&gt;

&lt;p&gt;&lt;strong&gt;2.Learning Programming language like Python/R&lt;/strong&gt;&lt;br&gt;
The next stage in data science for beginners is to learn a computer language such as R or Python. R and Python are popular data science programming languages for data manipulation, visualization, and machine learning. &lt;br&gt;
To begin, select one of the languages and begin studying the fundamentals. Variables, data types, loops, and functions are examples of such ideas. There are numerous tools accessible for studying R or Python, including online courses and tutorials on the greatest data science websites. As you continue, you will be able to delve into more sophisticated themes and improve your talents. You can access online courses on Microsoft Azure Learning, AWS learning platform udemy and coursera.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;3.Learning Data visualisation tool-Tableau/Power BI&lt;/strong&gt;&lt;br&gt;
Once you've mastered programming and data manipulation, the next step as a data science beginner student is to study a visualization tool such as Power BI or Tableau. To share your data insights, you can use these tools to create dynamic and visually appealing charts, graphs, and dashboards.&lt;br&gt;
To get started, select one of these tools and begin studying the fundamentals. Topics like as producing charts and graphs, constructing dashboards, and connecting to data sources may be covered. There are numerous resources accessible for learning visualization technologies, such as online classes, tutorials, and documentation. As you continue, you will be able to explore more advanced features and approaches.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;4.Learning statistics for Machine Learning&lt;/strong&gt;&lt;br&gt;
After learning a programming language and a visualization tool, the next stage is to understand fundamental statistics for machine learning. Machine learning is a data science subject that involves utilizing algorithms to learn from and predict data. To begin, you should understand fundamental concepts such as probability, statistics, and linear regression. There are numerous resources accessible for understanding the fundamental statistics of machine learning. Data science online courses, tutorials, and textbooks are examples of resources. As you improve, you can delve into more advanced topics and hone your machine learning skills. &lt;/p&gt;

&lt;p&gt;&lt;strong&gt;5.Learning Machine Learning Algorithms&lt;/strong&gt;&lt;br&gt;
After you've mastered fundamental statistics, the next step is to learn about machine learning methods. There are numerous algorithms used in machine learning, each with its own set of strengths and disadvantages. To begin, familiarize yourself with supervised and unsupervised learning like as decision trees, linear regression, and k-means clustering. &lt;/p&gt;

&lt;p&gt;&lt;strong&gt;6. Practice and Implement your skills&lt;/strong&gt;&lt;br&gt;
The final stage in learning data science as a beginner is to put what you've learnt into practice. Working on projects and exercises to put your skills to use, as well as engaging in online groups and forums to learn from others and get feedback on your work, can all be part of it. You should also think about joining a data science group or club, which will provide you more possibilities to study and work with others. You can use real-world data sets to practice and apply your skills by exploring, visualizing, and analyzing the data with the tools and techniques you've learned. You can also experiment with creating machine-learning models and testing them on various data sets. &lt;/p&gt;

&lt;p&gt;In conclusion, data science provides an amazing potential to have a significant influence in a variety of areas, ranging from healthcare to finance and beyond. Whether you want to be a data scientist, data analyst, machine learning engineer, or specialize in a particular topic, your commitment to mastering this profession will open doors to a world of opportunities. So, enjoy the journey of data science, adapt to new difficulties, and never stop learning. As a data science expert, your future is full with promise, innovation, and the opportunity to shape a data-driven world.&lt;/p&gt;

</description>
      <category>beginners</category>
      <category>datascience</category>
    </item>
  </channel>
</rss>
