<?xml version="1.0" encoding="UTF-8"?>
<rss version="2.0" xmlns:atom="http://www.w3.org/2005/Atom" xmlns:dc="http://purl.org/dc/elements/1.1/">
  <channel>
    <title>DEV Community: Vee</title>
    <description>The latest articles on DEV Community by Vee (@njengavee).</description>
    <link>https://dev.to/njengavee</link>
    <image>
      <url>https://media2.dev.to/dynamic/image/width=90,height=90,fit=cover,gravity=auto,format=auto/https:%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Fuser%2Fprofile_image%2F1173787%2F6e938641-bd94-4692-888c-881bb3a05e23.jpeg</url>
      <title>DEV Community: Vee</title>
      <link>https://dev.to/njengavee</link>
    </image>
    <atom:link rel="self" type="application/rss+xml" href="https://dev.to/feed/njengavee"/>
    <language>en</language>
    <item>
      <title>Data Engineering for Beginners: A Step-by-Step Guide</title>
      <dc:creator>Vee</dc:creator>
      <pubDate>Mon, 30 Oct 2023 21:00:25 +0000</pubDate>
      <link>https://dev.to/njengavee/data-engineering-for-beginners-a-step-by-step-guide-i7d</link>
      <guid>https://dev.to/njengavee/data-engineering-for-beginners-a-step-by-step-guide-i7d</guid>
      <description>&lt;p&gt;In today's data-driven world, the effective management and processing of data are critical for organizations and individuals alike. Data engineering plays a crucial role in this process, enabling the collection, storage, and transformation of data into valuable insights. If you're a beginner eager to dive into the world of data engineering, this step-by-step guide is here to help you get started.&lt;/p&gt;

&lt;h2&gt;
  
  
  Data Engineering
&lt;/h2&gt;

&lt;p&gt;Data engineering is the foundation of data-driven decision-making. According to Wikipedia, it refers to the building of systems to enable the collection and usage of data. It involves designing, building and maintaining data infrastructure and platforms, and making data accessible and usable for data scientists, analysts, and decision-makers. Without data engineering, raw data remains untamed and untapped, limiting the potential for valuable insights.&lt;/p&gt;

&lt;p&gt;Data engineers play an important role in an organization’s success through providing easier access to data that data scientists, analysts, and decision-makers need to do their jobs. To create scalable solutions, data engineers mostly require programming and problem-solving skills.&lt;/p&gt;

&lt;h2&gt;
  
  
  How to develop a data engineering career
&lt;/h2&gt;

&lt;p&gt;To become a data engineer, you need to be conversant with the following fundamentals:&lt;/p&gt;

&lt;ol&gt;
&lt;li&gt;&lt;p&gt;Programming basics&lt;br&gt;
You need to understand the basic of python programming including the syntax, operators, variables, data types, loops and conditional statements, data structures and standard libraries such as Numpy and Pandas. SQL is also fundamental when working with databases. Other programming languages you will need as you build on your skillset are Java and Scala which are also used in data processing.&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;Database Knowledge&lt;br&gt;
Databases rank among the most common solutions for data storage. You should be familiar with both relational and non-relational databases, and how they work. For relational databases, you need to learn the querying syntax and commands in SQL including the keys, joins and subqueries, window functions and normalization. For non-relational databases that deal with unstructured data, MongoDB and Cassandra are vital to learn.&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;ETL (extract, transform, and load) systems&lt;br&gt;
ETL is the process by which you’ll move data from databases and other sources into a single repository, like a data warehouse. Common ETL tools include Xplenty, Stitch, Alooma, and Talend.&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;Data processing with Apache Spark&lt;br&gt;
Data Processing refers to converting raw data into meaningful information that is machine readable. Apache® Spark™ is a fast, flexible, and developer-friendly open-source platform for large-scale SQL, batch processing, stream processing, and machine learning. Data engineers constantly work with big data and therefore incorporating Spark into their applications helps them rapidly query, analyze, and transform data at scale. As a data engineer it will then be vital to comprehend Spark architecture, RDDs in spark, working with Spark data frames, understand Spark execution, Spark SQL, broadcast and accumulators&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;Apache Hadoop-Based Analytics&lt;br&gt;
Apache Hadoop is an open-source platform that is used to compute distributed processing and storage against datasets. They assist in a wide range of operations, such as data processing, access, storage, governance, security, and operations. You'll need to understand MapReduce architecture, working with YARN and how to use Hadoop on the cloud for example, AWS with EMR.&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;Data Warehousing with Apache Hive&lt;br&gt;
Data warehousing helps data engineers to aggregate unstructured data, collected from multiple sources. It is then compared and assessed to improve the efficiency of business operations. Apache Hive is a data warehouse infrastructure built on top of Hadoop. It provides tools to enable easy data ETL, a mechanism to put structures on the data, and the capability for querying and analysis of large data sets stored in Hadoop files. It is important to learn the Hive querying language, managed visa vis extensional tables, partitioning and bucketing, and types of file formats.&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;Automation and scripting: Automation is a necessary part of working with big data simply because organizations are able to collect so much information. You should be able to write scripts to automate repetitive tasks.&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;Cloud computing&lt;br&gt;
Cloud computing stores data remotely, accessible from nearly any internet connection. This makes it a flexible and scalable environment for businesses and professionals to operate without the overheads of maintaining physical infrastructure. Cloud computing also make collaboration in data science teams possible. It is therefore vital to understand cloud storage and cloud computing as companies are increasingly shifting to cloud services. Beginners may consider a course in &lt;a href="https://www.coursera.org/learn/aws-fundamentals-going-cloud-native"&gt;Amazon Web Services (AWS)&lt;/a&gt; or &lt;a href="https://www.coursera.org/learn/gcp-big-data-ml-fundamentals?specialization=gcp-data-engineering"&gt;Google Cloud&lt;/a&gt;.&lt;/p&gt;&lt;/li&gt;
&lt;/ol&gt;

&lt;h2&gt;
  
  
  Conclusion
&lt;/h2&gt;

&lt;p&gt;Data engineering is the backbone of successful data analysis and decision-making. As a beginner, you now have a solid foundation to start your data engineering journey. Remember to continually explore new tools, technologies, and best practices as the field evolves. With dedication and a curious mindset, you'll be well on your way to becoming a proficient data engineer.&lt;/p&gt;

</description>
      <category>dataengineering</category>
      <category>beginners</category>
      <category>python</category>
      <category>learning</category>
    </item>
    <item>
      <title>The Complete Guide to Time Series Models</title>
      <dc:creator>Vee</dc:creator>
      <pubDate>Tue, 24 Oct 2023 16:54:49 +0000</pubDate>
      <link>https://dev.to/njengavee/the-complete-guide-to-time-series-models-12b6</link>
      <guid>https://dev.to/njengavee/the-complete-guide-to-time-series-models-12b6</guid>
      <description>&lt;p&gt;Have you ever wondered how meteorologists are able to predict future weather, or how businesses predict future trends, and easily detect changes in the market? Those are just a few eye openers of how Time Series works in various aspects of our daily lives. In this guide, we are going to get an understanding of Time series models work and the way they apply in our daily life.&lt;/p&gt;

&lt;h2&gt;
  
  
  Understanding Time series
&lt;/h2&gt;

&lt;p&gt;According to Wikipedia, a time series is a series of data points indexed in time order. A time series model is a type of machine learning model used to analyze and forecast the future based on verified previous data observed at regular intervals (Engineering Statistics Handbook, 2010). In this model, time is usually the independent variable.&lt;/p&gt;

&lt;p&gt;Time series models use non-stationary data. Non-stationary data is data that keeps fluctuating with time. The data is analyzed to give insights on different trends over time. This is what makes time series models part of &lt;a href="https://www.tableau.com/learn/articles/what-is-predictive-analytics"&gt;predictive analytics&lt;/a&gt;&lt;/p&gt;

&lt;h2&gt;
  
  
  Characteristics of time series models
&lt;/h2&gt;

&lt;ul&gt;
&lt;li&gt;Seasonality: Refers to patterns that occur regularly ( weekly, quarterly, monthly or annually) in the data due to various seasonal factors. For example, ice cream sales are higher during summer and low during winter. The figure below is an illustration of seasonality&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;&lt;a href="https://res.cloudinary.com/practicaldev/image/fetch/s--Jw4LqR4_--/c_limit%2Cf_auto%2Cfl_progressive%2Cq_auto%2Cw_800/https://dev-to-uploads.s3.amazonaws.com/uploads/articles/crta7rvl1pb3lfu0pjkx.jpg" class="article-body-image-wrapper"&gt;&lt;img src="https://res.cloudinary.com/practicaldev/image/fetch/s--Jw4LqR4_--/c_limit%2Cf_auto%2Cfl_progressive%2Cq_auto%2Cw_800/https://dev-to-uploads.s3.amazonaws.com/uploads/articles/crta7rvl1pb3lfu0pjkx.jpg" alt="seasonality" width="512" height="165"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;Image: Marco Peixeiro&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;Stationarity: This is the state of statistical properties of a time series not changing over time. This means a constant mean, variance and covariance is independent of time. Stationarity is very crucial as it influences how data is perceived and predicted. The following diagram shows an example of a stationary process:&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;&lt;a href="https://res.cloudinary.com/practicaldev/image/fetch/s--DlKY-oJb--/c_limit%2Cf_auto%2Cfl_progressive%2Cq_auto%2Cw_800/https://dev-to-uploads.s3.amazonaws.com/uploads/articles/bed2ll25g9suv5juqt7u.jpg" class="article-body-image-wrapper"&gt;&lt;img src="https://res.cloudinary.com/practicaldev/image/fetch/s--DlKY-oJb--/c_limit%2Cf_auto%2Cfl_progressive%2Cq_auto%2Cw_800/https://dev-to-uploads.s3.amazonaws.com/uploads/articles/bed2ll25g9suv5juqt7u.jpg" alt="Stationarity" width="512" height="155"&gt;&lt;/a&gt;           &lt;/p&gt;

&lt;p&gt;Image: Marco Peixeiro&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;Autocorrelation: This is the degree of similarity between a given time series and a lagged version of itself over a certain period of time. It measures the relationship between a variable's current value and its past values. An autocorrelation of +1 represents a perfect positive correlation, while an autocorrelation of -1 represents a perfect negative correlation. For instance, in businesses, autocorrelation can be used to evaluate how past prices are likely to influence future prices.&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;&lt;a href="https://res.cloudinary.com/practicaldev/image/fetch/s--wFRRq3cG--/c_limit%2Cf_auto%2Cfl_progressive%2Cq_auto%2Cw_800/https://dev-to-uploads.s3.amazonaws.com/uploads/articles/5k73iyz163695rfnpfnv.jpg" class="article-body-image-wrapper"&gt;&lt;img src="https://res.cloudinary.com/practicaldev/image/fetch/s--wFRRq3cG--/c_limit%2Cf_auto%2Cfl_progressive%2Cq_auto%2Cw_800/https://dev-to-uploads.s3.amazonaws.com/uploads/articles/5k73iyz163695rfnpfnv.jpg" alt="Autocorrelation" width="512" height="260"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;Image: Marco Peixeiro&lt;/p&gt;

&lt;h2&gt;
  
  
  Types of Time series analysis
&lt;/h2&gt;

&lt;p&gt;There are various types of time series analysis used for different purposes. They include:&lt;/p&gt;

&lt;ol&gt;
&lt;li&gt;&lt;p&gt;Forecasting: It is used to predict the future. It utilizes past data as a model for future data to predict future events.&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;Explanative: It strives to understand the data and the relationships within it including cause and effect.&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;Curve fitting: It plots the data along a curve to examine the relationships between variables in data.&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;Segmentation: It attempts to understand the underlying properties of the source information by splitting the data into segments.&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;Classification: It identifies and assigns categories to the data.&lt;/p&gt;&lt;/li&gt;
&lt;/ol&gt;

&lt;h2&gt;
  
  
  Time series models
&lt;/h2&gt;

&lt;p&gt;Time series models are classified into three broad categories:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;&lt;p&gt;&lt;a href="https://en.wikipedia.org/wiki/Autoregressive"&gt;Autoregressive (AR) models&lt;/a&gt;&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;Integrated (I) models&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;&lt;a href="https://en.wikipedia.org/wiki/Moving_average_model"&gt;Moving average (MA) models&lt;/a&gt;&lt;/p&gt;&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;Combinations of these three models brings forth autoregressive moving average (ARMA) and autoregressive integrated moving average (ARIMA) models. &lt;/p&gt;

&lt;p&gt;&lt;strong&gt;Autoregressive moving average (ARMA) model&lt;/strong&gt;&lt;/p&gt;

&lt;p&gt;The ARMA model is a regression model where the dependent variable is a linear function of past values of both the dependent variable and the error term. The order of an ARMA model is represented by ‘p’ for the autoregressive part and ‘q’ for the moving average part. For example, if p=0 and q=0, then it means that we are predicting time-step (t) based on time-step (t) only. If p=n and q=m, then we are predicting time-step (t) based on n past time-steps of the dependent/response variable and m past time-steps of the error term.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;Autoregressive integrated moving average (ARIMA) model&lt;/strong&gt;&lt;br&gt;
The autoregressive integrated moving average (ARIMA) model is a generalization of the ARMA model. The ARIMA model is a regression model in which the dependent variable is a linear function of past values of both the dependent variable and the error term, where the error term has been differentiated ‘d’ times. The model's goal is to predict future securities or financial market moves by examining the differences between values in the series instead of through actual values.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;Seasonal autoregressive integrated moving average (SARIMA) model&lt;/strong&gt;&lt;br&gt;
SARIMA is a type of time-series forecasting model that takes into account both seasonality and autocorrelation. SARIMA models are based on a combination of differencing I(d), autoregression model AR(p), moving average model MA(q) and seasonality S(P, D, Q, s), where s is simply the season’s length.&lt;/p&gt;

&lt;p&gt;&lt;a href="https://res.cloudinary.com/practicaldev/image/fetch/s--MCZ-982U--/c_limit%2Cf_auto%2Cfl_progressive%2Cq_auto%2Cw_800/https://dev-to-uploads.s3.amazonaws.com/uploads/articles/b4val2llb7ri3g4fmq3m.png" class="article-body-image-wrapper"&gt;&lt;img src="https://res.cloudinary.com/practicaldev/image/fetch/s--MCZ-982U--/c_limit%2Cf_auto%2Cfl_progressive%2Cq_auto%2Cw_800/https://dev-to-uploads.s3.amazonaws.com/uploads/articles/b4val2llb7ri3g4fmq3m.png" alt="SARIMA Model" width="602" height="130"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;Image: Utkarsh Soni,Helical IT Solutions(2023)&lt;/p&gt;

&lt;p&gt;SARIMA models are generally considered to be more accurate than other types of time-series forecasting models, such as ARIMA models. SARIMA models are also relatively easy to interpret and use. The SARIMA model can be used to forecast demand for a product or service over the course of a year, forecasting stock prices and weather patterns.  &lt;/p&gt;

&lt;h2&gt;
  
  
  Applications of time series models
&lt;/h2&gt;

&lt;p&gt;Finance − It includes sales forecasting, inventory analysis, stock market analysis, price estimation.&lt;/p&gt;

&lt;p&gt;Retail - Retailers may apply time series models to study how other companies’ prices and the number of customer purchases change over time, helping them optimize prices. &lt;/p&gt;

&lt;p&gt;Meteorology − Time series models can be used in temperature estimation, climate change, seasonal shift recognition, and weather forecasting.&lt;/p&gt;

&lt;p&gt;Healthcare - Time series models can be used to monitor the spread of diseases by observing how many people transmit a disease and how many people die after being infected.&lt;/p&gt;

&lt;h2&gt;
  
  
  Conclusion
&lt;/h2&gt;

&lt;p&gt;In this article, you have had an introduction to the fascinating world of time series models. Understanding time series data, its components, and the models used to analyze it is a crucial skill in today's data-driven world. For further study, consider the following resources: &lt;br&gt;
&lt;a href="https://builtin.com/data-science/time-series-model"&gt;The complete guide to Time series models&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;&lt;a href="https://vitalflux.com/different-types-of-time-series-forecasting-models/#Autoregressive_AR_model"&gt;Different types of Time-series Forecasting Models&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;&lt;a href="https://www.simplilearn.com/tutorials/statistics-tutorial/what-is-time-series-analysis#why_do_we_need_timeseries_analysis"&gt;Time series analysis&lt;/a&gt;&lt;/p&gt;

</description>
      <category>datascience</category>
      <category>machinelearning</category>
      <category>algorithms</category>
      <category>data</category>
    </item>
    <item>
      <title>Exploratory Data Analysis using Data Visualization techniques</title>
      <dc:creator>Vee</dc:creator>
      <pubDate>Wed, 11 Oct 2023 23:42:56 +0000</pubDate>
      <link>https://dev.to/njengavee/exploratory-data-analysis-using-data-visualization-techniques-34c6</link>
      <guid>https://dev.to/njengavee/exploratory-data-analysis-using-data-visualization-techniques-34c6</guid>
      <description>&lt;p&gt;Exploratory data analysis (EDA) is the process of studying data using visualization and statistical methods to understand the data. In other words, it is the first look on your data. It is a vital step before you begin with the actual data analysis. EDA helps to discover relationships within data, identify patterns and outliers that may exist within the dataset. Data scientists use EDA to ensure the results they produce are valid and applicable to any desired business outcomes and goals.&lt;/p&gt;

&lt;h2&gt;
  
  
  Objectives of EDA
&lt;/h2&gt;

&lt;p&gt;The main objectives of EDA are to:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;Confirm if the data is making sense in the context of the problem being solved. In the case where it doesn't, we come up with other strategies such as collecting more data.&lt;/li&gt;
&lt;li&gt;Uncover and resolve issues on data quality such as, duplicates, missing values, incorrect data types and incorrect values.&lt;/li&gt;
&lt;li&gt;Get insights about the data, for example, descriptive statistics.&lt;/li&gt;
&lt;li&gt;Detect anomalies and outliers that may cause problems during data analysis. Outliers are values that lie too far from the standard values.&lt;/li&gt;
&lt;li&gt;Uncover data patterns and correlations between variables.&lt;/li&gt;
&lt;/ul&gt;

&lt;h2&gt;
  
  
  Types of EDA
&lt;/h2&gt;

&lt;p&gt;Exploratory data analysis is classified into three broad categories namely:&lt;br&gt;
&lt;a href="https://www.ibm.com/topics/exploratory-data-analysis#Exploratory+data+analysis+tools" rel="noopener noreferrer"&gt;Univariate&lt;/a&gt;&lt;br&gt;
&lt;a href="https://www.ibm.com/topics/exploratory-data-analysis#Exploratory+data+analysis+tools" rel="noopener noreferrer"&gt;Bivariate&lt;/a&gt;&lt;br&gt;
&lt;a href="https://www.ibm.com/topics/exploratory-data-analysis#Exploratory+data+analysis+tools" rel="noopener noreferrer"&gt;Multivariate&lt;/a&gt;&lt;/p&gt;

&lt;h2&gt;
  
  
  Steps in EDA
&lt;/h2&gt;

&lt;p&gt;The following is a step-by-step approach to undertaking an exploratory data analysis:&lt;/p&gt;

&lt;ol&gt;
&lt;li&gt;Data Collection&lt;/li&gt;
&lt;/ol&gt;

&lt;p&gt;Gather relevant and sufficient data for your project. There are various sites online you could get your data from irrespective of the sector you're in. Here are a few examples to check out: &lt;a href="https://www.kaggle.com/datasets" rel="noopener noreferrer"&gt;Kaggle&lt;/a&gt;, &lt;a href="https://datahub.io/collections" rel="noopener noreferrer"&gt;Datahub.io&lt;/a&gt;, &lt;a href="https://www.bfi.org.uk/education-research/film-industry-statistics-research" rel="noopener noreferrer"&gt;BFI film industry statistics&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;2.Familiarize with data&lt;/p&gt;

&lt;p&gt;This step is important as it helps you to determine whether the data is adequate for the analysis about to be done.&lt;/p&gt;

&lt;p&gt;3.Data cleaning&lt;/p&gt;

&lt;p&gt;This where any missing values, outliers and duplicates are identified and removed from the dataset. Also, data that is irrelevant for the anticipated analysis is removed at this stage.&lt;/p&gt;

&lt;p&gt;4.Identify associations in the dataset&lt;/p&gt;

&lt;p&gt;Look for any correlations between variables. You can use a heatmap or scatterplots to make it easier for you to identify the correlations.&lt;/p&gt;

&lt;h2&gt;
  
  
  Example: Exploratory Data Analysis using NYC Citi Bike data.
&lt;/h2&gt;

&lt;p&gt;We will now perform an exploratory data analysis on NYC Citi Bike data to get a better understanding of the process. You can access the data &lt;a href="https://www.youtube.com/redirect?event=video_description&amp;amp;redir_token=QUFFLUhqbXZvb2w3b1dqRzJuNjRxdEFuX016Q0FzY0F6UXxBQ3Jtc0trMzJfdUk5M25SbEYtWVloVUw2TGdBUlhBSG9TOFp1R29MY3VNRFdCdGpUZzFfR09GT0Y3bXhNQXFUajFtS2c2RThkTks4Rm5kTFVkckpkM3RqTmFlUDlROU9HQzBrOThrVDNzbzNqQkpiUnJxMnZjQQ&amp;amp;q=https%3A%2F%2Fgithub.com%2Fmamaggie%2FTutorialProjects%2Ftree%2Fmain%2Fdata&amp;amp;v=pkYtQjHhi3Q" rel="noopener noreferrer"&gt;here&lt;/a&gt;.&lt;/p&gt;

&lt;p&gt;1.Import data&lt;/p&gt;

&lt;p&gt;The first step is to import all the modules you are going to use in your project. In this case, we will need pandas for data wrangling, seaborn for data visualization. This is how I would do it:&lt;/p&gt;

&lt;p&gt;`import pandas as pd&lt;br&gt;
import seaborn as sns&lt;/p&gt;

&lt;p&gt;Then import your dataset. If you're using Google colab, this is how you would load the data:&lt;/p&gt;

&lt;p&gt;&lt;code&gt;from google.colab import files&lt;br&gt;
uploaded = files.upload()&lt;/code&gt;&lt;/p&gt;

&lt;p&gt;You will then read in the data as a pandas data frame like this:&lt;/p&gt;

&lt;p&gt;&lt;a href="https://media.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2F7nj1fviwpmh6tdevxou4.PNG" class="article-body-image-wrapper"&gt;&lt;img src="https://media.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2F7nj1fviwpmh6tdevxou4.PNG" alt="Image description"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;2.Get an overview of the data&lt;/p&gt;

&lt;p&gt;You can approach this in various ways. For example, using .info() helps us to know the data types, number of columns, column names, and number of values in the data frame. The following is an example:&lt;/p&gt;

&lt;p&gt;&lt;a href="https://media.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2F9u7o7xk595nfopug0ypl.PNG" class="article-body-image-wrapper"&gt;&lt;img src="https://media.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2F9u7o7xk595nfopug0ypl.PNG" alt="Image description"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;The other alternative is to use .describe(). This gives you a statistical summary of the data. Here's an example:&lt;/p&gt;

&lt;p&gt;&lt;a href="https://media.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fuhesfsar6ll7ufrp8jhv.PNG" class="article-body-image-wrapper"&gt;&lt;img src="https://media.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fuhesfsar6ll7ufrp8jhv.PNG" alt="Image description"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;3.Visualize the distribution for trip duration&lt;br&gt;
This will help us to have a glimpse of how long most trips took. Using seaborn, this is how I would do it:&lt;/p&gt;

&lt;p&gt;&lt;code&gt;# visualize distribution for trip duration&lt;br&gt;
sns.histplot(data['tripduration']) &lt;/code&gt;&lt;br&gt;
Here's the sample output:&lt;/p&gt;

&lt;p&gt;&lt;a href="https://media.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fyilo4xxrrbsh1wu39oob.PNG" class="article-body-image-wrapper"&gt;&lt;img src="https://media.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fyilo4xxrrbsh1wu39oob.PNG" alt="Image description"&gt;&lt;/a&gt;&lt;br&gt;
From the output, it is evident that most trips were ranging within 10 minutes.&lt;/p&gt;

&lt;p&gt;4.Visualize correlation between gender and trip duration&lt;/p&gt;

&lt;p&gt;&lt;code&gt;# checking for association between tripduration and gender using scatterplots&lt;br&gt;
sns.pairplot(data[['tripduration', 'gender']])&lt;/code&gt;&lt;/p&gt;

&lt;p&gt;Sample output is as follows:&lt;/p&gt;

&lt;p&gt;&lt;a href="https://media.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fyl6ik2th50x7xnflq9i7.PNG" class="article-body-image-wrapper"&gt;&lt;img src="https://media.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fyl6ik2th50x7xnflq9i7.PNG" alt="Image description"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;5.Calculate the percentage of subscribers&lt;/p&gt;

&lt;p&gt;we need to find out the share of subscribers from the total number of riders in New York city. Here's how to find out:&lt;/p&gt;

&lt;p&gt;&lt;a href="https://media.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fujw0l2nijaj4n4wslkwd.PNG" class="article-body-image-wrapper"&gt;&lt;img src="https://media.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fujw0l2nijaj4n4wslkwd.PNG" alt="Image description"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;6.Evaluate how trip length varies based on trip start time&lt;/p&gt;

&lt;p&gt;&lt;code&gt;data['hour'] = data.starttime.apply(lambda x: x[11:13]).astype('str')&lt;br&gt;
data&lt;/code&gt;&lt;/p&gt;

&lt;p&gt;&lt;code&gt;# visualize correlation&lt;br&gt;
sns.scatterplot(x= 'hour', y= 'tripduration', data = data, hue= 'usertype')&lt;/code&gt;&lt;/p&gt;

&lt;p&gt;The output is as follows:&lt;/p&gt;

&lt;p&gt;&lt;a href="https://media.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fnq545n9kfzs1tna2mebw.PNG" class="article-body-image-wrapper"&gt;&lt;img src="https://media.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fnq545n9kfzs1tna2mebw.PNG" alt="Image description"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;7.Determine the bike stations where most trips starts&lt;/p&gt;

&lt;p&gt;First we get the count of bike stations and store the output as a new data frame. We we then drop the duplicates from the original data frame then merge the two new data frames for visualization. &lt;/p&gt;

&lt;p&gt;&lt;code&gt;# Get the count of trips from each station&lt;br&gt;
new_data = data.groupby(['start station id']).size().reset_index(name= 'counts')&lt;/code&gt;&lt;/p&gt;

&lt;p&gt;&lt;code&gt;#remove duplicate values from the start station id column&lt;br&gt;
temp_data = data.drop_duplicates('start station id')&lt;/code&gt;&lt;/p&gt;

&lt;p&gt;&lt;code&gt;# left join to merge new_data and temp_data dataframes&lt;br&gt;
newdata2 = pd.merge(new_data, temp_data[['start station id', 'start station name', 'start station latitude', 'start station longitude']], how= 'left', on= ['start station id'])&lt;/code&gt;&lt;/p&gt;

&lt;p&gt;&lt;code&gt;#install folium&lt;br&gt;
!pip install folium&lt;br&gt;
import folium&lt;/code&gt;&lt;/p&gt;

&lt;p&gt;&lt;code&gt;# initialize a map&lt;br&gt;
m = folium.Map(location=[40.691966, -73.981302],tiles= 'OpenstreetMap', zoom_start= 12)&lt;br&gt;
m&lt;/code&gt;&lt;/p&gt;

&lt;p&gt;&lt;a href="https://media.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fvuvm8ag8bonjhkbpx1hi.PNG" class="article-body-image-wrapper"&gt;&lt;img src="https://media.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fvuvm8ag8bonjhkbpx1hi.PNG" alt="Image description"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;The output is as follows:&lt;/p&gt;

&lt;p&gt;&lt;a href="https://media.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fca0qilqs77kpcssihzud.PNG" class="article-body-image-wrapper"&gt;&lt;img src="https://media.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fca0qilqs77kpcssihzud.PNG" alt="Image description"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;h2&gt;
  
  
  Conclusion
&lt;/h2&gt;

&lt;p&gt;EDA is very crucial as it affects the quality of the findings in the final analysis. The success of any EDA is dependent on the quality and quantity of data, the type of tools and visualization used, and proper interpretation by a data scientist. &lt;/p&gt;

</description>
      <category>datascience</category>
      <category>beginners</category>
      <category>python</category>
      <category>luxdatanerds</category>
    </item>
    <item>
      <title>Understanding data science. A complete beginner roadmap</title>
      <dc:creator>Vee</dc:creator>
      <pubDate>Sun, 01 Oct 2023 16:52:32 +0000</pubDate>
      <link>https://dev.to/njengavee/understanding-data-science-a-complete-beginner-roadmap-3cn2</link>
      <guid>https://dev.to/njengavee/understanding-data-science-a-complete-beginner-roadmap-3cn2</guid>
      <description>&lt;p&gt;Imagine you are a CEO in an ecommerce platform and you want to understand what your customers think about your services. First you need to do a survey to get views and opinions from your customers. Next step will be to explore your data to draw meaningful insights. The insights will now inform how the customer service will be improved.&lt;br&gt;
That is data science for you. Let's get started and understand what this term "&lt;em&gt;data science&lt;/em&gt;" is all about.&lt;/p&gt;

&lt;h2&gt;
  
  
  What is data science
&lt;/h2&gt;

&lt;p&gt;Data science is the process of drawing meaningful insights from large data sets. It acts as a third eye by foreseeing problems and creating solutions before they occur. This is how it acts as a third eye.&lt;br&gt;&lt;br&gt;
Based on the field you're in, data can help you find solutions to your problems. For instance, in an ecommerce company, data can help you know: &lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;What products are more preferred by customers&lt;/li&gt;
&lt;li&gt;How to improve customer service&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;&lt;strong&gt;The roles of a Data Scientist&lt;/strong&gt;&lt;br&gt;
As a data scientist in an organization, your roles will range from: &lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;identifying crucial areas for research &lt;/li&gt;
&lt;li&gt;knowing where to get your data from&lt;/li&gt;
&lt;li&gt;data cleaning, data exploration&lt;/li&gt;
&lt;li&gt;creating statistical models&lt;/li&gt;
&lt;li&gt;presenting the findings to stakeholders&lt;/li&gt;
&lt;/ul&gt;

&lt;h2&gt;
  
  
  Application areas
&lt;/h2&gt;

&lt;p&gt;Data science is crucial in the decision making in all aspects of our daily lives. The following are some applications of data science in various sectors:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;&lt;p&gt;Education: Data science is being used by teachers to assess the comprehension of students in various units in order to come up with better teaching mechanisms to improve the students' performance.&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;Business: Data science is used in businesses to predict future market trends, improve quality of products based on customer preferences and enhance product marketing based on customer previous purchase history and browsing behavior.&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;Meteorology: Data science is applied in this sector to improve the level of accuracy in weather forecasting to help save lives from extreme weather events.&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;Transportation: In the transport sector, data science is applied to help optimize routing, improve safety, and reduce emissions.&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;Energy: Data science is used in energy to optimize energy production and distribution, reduce costs, and improve efficiency.&lt;/p&gt;&lt;/li&gt;
&lt;/ul&gt;

&lt;h2&gt;
  
  
  Skills required in Data Science
&lt;/h2&gt;

&lt;p&gt;To be a Data Scientist, you require skills in the following fields:&lt;br&gt;
Programming languages :As a beginner you need to learn the basics of programming. In data science some of the programming languages include:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;&lt;a href="https://www.geeksforgeeks.org/python-programming-language/?ref=lbp"&gt;Python&lt;/a&gt;&lt;/li&gt;
&lt;li&gt;&lt;a href="https://www.geeksforgeeks.org/r-tutorial/"&gt;R-Propramming&lt;/a&gt;&lt;/li&gt;
&lt;li&gt;&lt;a href="https://www.geeksforgeeks.org/sql-tutorial/"&gt;SQL&lt;/a&gt;&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;&lt;strong&gt;Statistics&lt;/strong&gt;&lt;br&gt;
You need have a foundation in calculus, linear algebra and statistics. Calculus is useful in learning how to design optimization algorithms for machine learning. Learning linear algebra helps you work with vectors and matrices. This will be very crucial in conducting analysis on data.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;Machine Learning&lt;/strong&gt;&lt;br&gt;
This is a subfield of artificial intelligence (AI) and computer science that focuses on using data and algorithms to mimic how humans learn, gradually improving accuracy.&lt;/p&gt;

&lt;h2&gt;
  
  
  Learning Guide
&lt;/h2&gt;

&lt;p&gt;You might wonder where to get the resources to start learning. There are numerous resources available to get you started, including online courses, YouTube videos and Bootcamps. The following are a few resources you can check out:&lt;br&gt;
&lt;a href="https://youtu.be/ua-CiDNNj30?si=Ddwk1opxqIMbeeJd"&gt;https://youtu.be/ua-CiDNNj30?si=Ddwk1opxqIMbeeJd&lt;/a&gt; &lt;br&gt;
&lt;a href="https://youtu.be/rGx1QNdYzvs?si=ZvMVq6r9o8czTjFe"&gt;https://youtu.be/rGx1QNdYzvs?si=ZvMVq6r9o8czTjFe&lt;/a&gt;&lt;br&gt;
Here are a few tips for you to accelerate your learning:&lt;br&gt;
&lt;strong&gt;Start with the basics-&lt;/strong&gt; Try to learn only some things at a time. Focus on learning the basics of programming, statistics, and machine learning first.&lt;br&gt;
&lt;strong&gt;Practice regularly-&lt;/strong&gt; The best way to learn data science is by practicing. You can check out &lt;a href="https://www.hackerrank.com/"&gt;Hackerrank&lt;/a&gt; and &lt;a href="https://www.kaggle.com/datasets"&gt;kaggle&lt;/a&gt; for practice.&lt;br&gt;
&lt;strong&gt;Always reach out for help-&lt;/strong&gt; There are many online communities and forums where you can ask questions and get help from other data scientists. &lt;/p&gt;

&lt;h2&gt;
  
  
  Conclusion
&lt;/h2&gt;

&lt;p&gt;Data science is a dynamic and rewarding field that opens up countless opportunities. Whether you're driven by curiosity, the potential for a lucrative career, or the desire to make data-driven decisions, data science has something to offer. With a growing job market and diverse roles, such as data wrangler, machine learning engineer, and business intelligence analyst, there's a niche for everyone in this exciting field.&lt;br&gt;
As you embark on your data science journey, remember that continuous learning and hands-on experience will be your best allies. Enjoy the adventure!&lt;/p&gt;

</description>
      <category>beginners</category>
      <category>datascience</category>
      <category>python</category>
    </item>
  </channel>
</rss>
