<?xml version="1.0" encoding="UTF-8"?>
<rss version="2.0" xmlns:atom="http://www.w3.org/2005/Atom" xmlns:dc="http://purl.org/dc/elements/1.1/">
  <channel>
    <title>DEV Community: Simon Ngotho</title>
    <description>The latest articles on DEV Community by Simon Ngotho (@ngotho).</description>
    <link>https://dev.to/ngotho</link>
    <image>
      <url>https://media2.dev.to/dynamic/image/width=90,height=90,fit=cover,gravity=auto,format=auto/https:%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Fuser%2Fprofile_image%2F1174314%2Fcc02f997-c76f-4d88-8d0c-9ac96d1c790c.png</url>
      <title>DEV Community: Simon Ngotho</title>
      <link>https://dev.to/ngotho</link>
    </image>
    <atom:link rel="self" type="application/rss+xml" href="https://dev.to/feed/ngotho"/>
    <language>en</language>
    <item>
      <title>Article 4: A step-by-step guide for Data Engineering beginners</title>
      <dc:creator>Simon Ngotho</dc:creator>
      <pubDate>Thu, 02 Nov 2023 06:44:33 +0000</pubDate>
      <link>https://dev.to/ngotho/article-4-a-step-by-step-guide-for-data-engineering-beginners-1ja8</link>
      <guid>https://dev.to/ngotho/article-4-a-step-by-step-guide-for-data-engineering-beginners-1ja8</guid>
      <description>&lt;h3&gt;
  
  
  &lt;strong&gt;A step-by-step guide for Data Engineering beginners&lt;/strong&gt;
&lt;/h3&gt;

&lt;p&gt;&lt;strong&gt;Introduction&lt;/strong&gt;&lt;/p&gt;

&lt;p&gt;In order to build a successful data engineering career and profession, it is important to understand what it entails as well as what it takes to be a data engineer.&lt;br&gt;
In 2006, the British mathematician, Clive Humby declared that “data is the new oil”. The mathematics genius was right as he meant that data, just crude like oil, isn't useful in its raw state. It needs to be refined, processed and turned into something useful since its value lies in its potential. Sure enough, it is only refined oil that is able to run the world and so is data in the current world.  &lt;/p&gt;

&lt;p&gt;&lt;em&gt;&lt;strong&gt;Data engineering&lt;/strong&gt;&lt;/em&gt; is the process of building and designing systems that to help people and entities collect and analyze raw data from various sources and formats. These systems are instrumental in helping them manipulate data for use by businesses to make critical decisions. &lt;/p&gt;

&lt;p&gt;&lt;em&gt;&lt;strong&gt;Data engineers&lt;/strong&gt;&lt;/em&gt; design, build, and optimize systems for data collection, storage, access, and analytics at scale. They come before data scientist who need these pipelines to convert raw data into usable formats either through data-centric applications or by other data consumers. &lt;/p&gt;

&lt;p&gt;In this article, I have explored my data engineering learning path to help me build my profession in a seamless manner. &lt;/p&gt;

&lt;p&gt;&lt;a href="https://res.cloudinary.com/practicaldev/image/fetch/s--VpGqzg_n--/c_limit%2Cf_auto%2Cfl_progressive%2Cq_auto%2Cw_800/https://dev-to-uploads.s3.amazonaws.com/uploads/articles/dtimtzhbb1k1c7xcvbtt.png" class="article-body-image-wrapper"&gt;&lt;img src="https://res.cloudinary.com/practicaldev/image/fetch/s--VpGqzg_n--/c_limit%2Cf_auto%2Cfl_progressive%2Cq_auto%2Cw_800/https://dev-to-uploads.s3.amazonaws.com/uploads/articles/dtimtzhbb1k1c7xcvbtt.png" alt="Image description" width="800" height="450"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;Understanding the key roles and responsibilities for Data Engineer jobs&lt;/strong&gt;&lt;/p&gt;

&lt;ol&gt;
&lt;li&gt;Creating and maintaining optimal data pipeline architecture.&lt;/li&gt;
&lt;li&gt;Building analytics tools that utilize the data pipeline to provide actionable insights into customer acquisition, operational efficiency and other key business performance metrics.&lt;/li&gt;
&lt;li&gt;Identifying, designing, and implementing internal process improvements: automating manual processes, optimizing data delivery, re-designing infrastructure for greater scalability, etc.&lt;/li&gt;
&lt;li&gt;Assembling large, complex data sets that meet functional / non-functional business requirements.&lt;/li&gt;
&lt;li&gt;Building the infrastructure required for optimal extraction, transformation, and loading of data from a wide variety of data sources using SQL and other ‘big data’ technologies.&lt;/li&gt;
&lt;li&gt;Monitoring and troubleshooting data systems and pipelines, ensuring that they are reliable, secure, and scalable, and resolving any issues or errors that may occur.&lt;/li&gt;
&lt;li&gt;Creating data tools for analytics and data scientist team members that assist them in building and optimizing product into an innovative industry leader.&lt;/li&gt;
&lt;li&gt;Collaborating and communicating with other data professionals, such as data scientists, data analysts, data architects, etc., to understand the data needs and provide the data solutions.&lt;/li&gt;
&lt;/ol&gt;

&lt;p&gt;&lt;strong&gt;Skills and tools needed to build strong data engineering career.&lt;/strong&gt;&lt;/p&gt;

&lt;p&gt;To become a data engineer, I will need to have a strong background in;&lt;/p&gt;

&lt;p&gt;a.  Computer science, &lt;br&gt;
b.  Mathematics, and &lt;br&gt;
c.  Statistics&lt;br&gt;
.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;Technical skills and tools that I need to equip myself with include;&lt;/strong&gt; &lt;/p&gt;

&lt;p&gt;&lt;strong&gt;Programming languages&lt;/strong&gt;: To be used for data manipulation, analysis, and automation. Python, Java, Scala, etc &lt;/p&gt;

&lt;p&gt;&lt;strong&gt;Databases&lt;/strong&gt;: To store and query structured, semi-structured, or unstructured data.&lt;br&gt;
 SQL, NoSQL, or graph databases, that can.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;Cloud computing&lt;/strong&gt;. For scalable and cost-effective data storage services.&lt;br&gt;
AWS, Azure, or Google Cloud,&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;Data warehouse platforms&lt;/strong&gt;. To provide data warehousing and analytics capabilities. Snowflake, Redshift, or BigQuery.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;Big-Data tools&lt;/strong&gt;. To handle distributed and parallel data processing and streaming.  Hadoop, Spark, MapReduce, Kafka, etc.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;Data visualization tools&lt;/strong&gt;. To create interactive and informative data dashboards and reports. Tableau, Power BI, or Dash&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;Orchestration tools&lt;/strong&gt;. To orchestrate and schedule data pipelines and workflows. Airflow, Luigi, or Prefect&lt;/p&gt;

&lt;p&gt;&lt;a href="https://res.cloudinary.com/practicaldev/image/fetch/s--i8byBrqT--/c_limit%2Cf_auto%2Cfl_progressive%2Cq_auto%2Cw_800/https://dev-to-uploads.s3.amazonaws.com/uploads/articles/nqxr0xy7kbpp7e3m9hu4.JPG" class="article-body-image-wrapper"&gt;&lt;img src="https://res.cloudinary.com/practicaldev/image/fetch/s--i8byBrqT--/c_limit%2Cf_auto%2Cfl_progressive%2Cq_auto%2Cw_800/https://dev-to-uploads.s3.amazonaws.com/uploads/articles/nqxr0xy7kbpp7e3m9hu4.JPG" alt="Image description" width="800" height="594"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;Real-world Projects&lt;/strong&gt;&lt;/p&gt;

&lt;p&gt;Real-world projects are an excellent way for me to apply my skills and gain practical experience. I intend to join Hackathons and Competitions, data engineering forums and communities like Stack Overflow, Reddit's r/dataengineering, LinkedIn groups, and Data Engineering Club etc. These will not only help me build a strong portfolio but also deepen my understanding of data engineering concepts. When working on these projects, I will focus on best practices in data engineering, data quality, scalability, and automation. I will strive to document my projects and endeavor to share my work on platforms like GitHub, Dev.to, medium.com etc. to showcase my skills to potential employers and data science communities. &lt;/p&gt;

&lt;p&gt;&lt;strong&gt;Conclusion&lt;/strong&gt;&lt;/p&gt;

&lt;p&gt;Data engineering is a dynamic and evolving field that requires constant learning and adaptation to new technologies and trends. &lt;br&gt;
This roadmap will provide me with a strong foundation, and I believe I can expand my knowledge from here based on my specific data engineering career interest.&lt;br&gt;&lt;br&gt;
It provides a structured path to follow, making it easier to understand the field's complexities and where to start and thus help me acquire the most important skills without wasting time on less relevant topics.&lt;br&gt;
By including projects and hands-on practice, this guide will encourage me to apply my knowledge to real-world scenarios, making me job-ready.&lt;/p&gt;

</description>
    </item>
    <item>
      <title>Article 3: Complete Guide to Time Series Models</title>
      <dc:creator>Simon Ngotho</dc:creator>
      <pubDate>Sat, 28 Oct 2023 08:45:13 +0000</pubDate>
      <link>https://dev.to/ngotho/article-3-complete-guide-to-time-series-models-3akl</link>
      <guid>https://dev.to/ngotho/article-3-complete-guide-to-time-series-models-3akl</guid>
      <description>&lt;h2&gt;
  
  
  &lt;strong&gt;COMPLETE GUIDE TO TIME SERIES MODELS&lt;/strong&gt;
&lt;/h2&gt;

&lt;p&gt;&lt;strong&gt;PRELIMINARIES&lt;/strong&gt;&lt;/p&gt;

&lt;p&gt;Time series analysis is a detailed way of analysing a sequence of data points collected over a period of time. To carry this analysis, analysts record data points at consistent intervals over a given period of time rather than just recording the data points at random or sporadically.&lt;br&gt;
Time series data is pervasive in various fields like financial markets, economics, energy, healthcare, environmental sciences and many more. &lt;br&gt;
To measure time series, it requires one to build a time series model that helps analyse and forecast the future. In this models, time is often the independent variable, and the goal is usually to make a prediction for the future. Understanding and effectively modeling time-dependent data is crucial for making informed decisions and predictions. &lt;br&gt;
In this comprehensive guide, we will explore different characteristics and types of time series models in data science. &lt;/p&gt;

&lt;p&gt;&lt;strong&gt;CHARACTERISTICS OF TIME SERIES MODELS&lt;/strong&gt;&lt;br&gt;
In order to have an understanding of how time series models work, it is paramount to explore their main characteristics as detailed below; &lt;/p&gt;

&lt;p&gt;&lt;strong&gt;1.    Stationarity&lt;/strong&gt;&lt;br&gt;
It refers to the statistical properties of a time series remaining constant over time. It has three main statistical characteristics, i.e. mean, variance, and autocorrelation, which do not exhibit significant changes with respect to time. Stationarity is crucial because many time series models and statistical techniques assume or work better with stationary data. &lt;br&gt;
If a time series is found to be non-stationary, appropriate transformations like differencing or using models designed for non-stationary data (e.g., vector autoregression for integrated time series) may be applied to address the non-stationarity and make the data amenable to analysis and forecasting.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;2.    Seasonality&lt;/strong&gt;&lt;br&gt;
It refers to recurring, predictable patterns or fluctuations that occur at regular intervals within a given time frame (i.e daily, weekly, monthly, or yearly patterns). These patterns are often associated with seasonal, environmental, or calendar-related factors and can have a significant impact on the behaviour of the data.&lt;br&gt;
By accounting for seasonality, analysts can better capture the true underlying dynamics of the data and make more informed decisions in various applications, including business forecasting, economic analysis, and environmental monitoring.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;3.    Autocorrelation&lt;/strong&gt;&lt;br&gt;
Also known as serial correlation, is a statistical concept in time series that quantitifies the degree of similarity between a time series and a lagged version of itself. In short, it assesses the correlation between a data point and previous data points in the same time series.&lt;br&gt;
It is used to identify patterns and relationships within a time series. It is an essential concept in time series analysis because it helps to detect and understand underlying structures, trends, and seasonality in the data.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;TYPES OF TIME SERIES MODELS&lt;/strong&gt;&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;1.    Autoregressive Integrated Moving Average (ARIMA) Models:&lt;/strong&gt; &lt;/p&gt;

&lt;p&gt;ARIMA models usually combine autoregressive, differencing, and moving average components to model a wide range of time series data. ARIMA (p, d, q) models are useful for handling non-stationary data and capturing both short-term and long-term dependencies.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;Note:&lt;/strong&gt; &lt;br&gt;
&lt;em&gt;&lt;strong&gt;p (Autoregressive Order):&lt;/strong&gt;&lt;/em&gt; The autoregressive order, denoted as 'p,' refers to the number of lagged observations included in the model to predict the current value.&lt;/p&gt;

&lt;p&gt;&lt;em&gt;&lt;strong&gt;d (Integrated Order):&lt;/strong&gt;&lt;/em&gt; The differencing order, denoted as 'd,' represents the number of differences required to make the time series data stationary.&lt;/p&gt;

&lt;p&gt;&lt;em&gt;&lt;strong&gt;q (Moving Average Order):&lt;/strong&gt;&lt;/em&gt; The moving average order, denoted as 'q,' indicates the number of lagged forecast errors included in the model to predict the current value.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;2.    Autoregressive (AR) Models:&lt;/strong&gt;&lt;br&gt;
AR are a class of time series models that rely on the linear relationship between a data point and its past values. In AR(p) models, the current value is a linear combination of the previous p values. AR models are used when the time series exhibits autocorrelation and may have a stationary behavior.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;3.    Moving Average (MA) Models:&lt;/strong&gt;&lt;br&gt;
MA are other classes of linear time series models that focus on the relationship between a data point and past forecast errors. In MA(q) models, the current value depends on the previous q forecast errors. They are used to capture short-term variations in a time series.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;4.    Vector Autoregression (VAR) Models:&lt;/strong&gt;&lt;br&gt;
VAR models are used for multivariate time series data, where multiple variables are interrelated. They use the past values of all variables in the system to make predictions.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;5.    Seasonal Autoregressive Integrated Moving Average (SARIMA) Models:&lt;/strong&gt;&lt;br&gt;
SARIMA models extend ARIMA models that incorporate seasonal components. They are designed to handle time series data with seasonal patterns or periodic, like weekly, monthly, or annual cycles.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;6.    Exponential Smoothing Models:&lt;/strong&gt;&lt;br&gt;
These methods include the simple exponential smoothing, Holt-Winters exponential smoothing and Holt's linear exponential smoothing. They are used to record different levels of trend and seasonality in time series data.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;7.    Long Short-Term Memory (LSTM) and Recurrent Neural Networks (RNNs):&lt;/strong&gt; &lt;br&gt;
Deep learning models like LSTMs and RNNs are used capture complex temporal dependencies in time series data. They are effective when dealing with non-linear and high-dimensional time series.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;8.    Bayesian Structural Time Series (BSTS) Models:&lt;/strong&gt; &lt;br&gt;
BSTS are Bayesian state space models that provide a powerful framework for time series decomposition, forecasting, and uncertainty estimation. They offer a flexible and powerful framework for decomposing time series data into its constituent components, modeling complex dependencies, and generating forecasts while accounting for uncertainty.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;9.    State Space Time Series Models (SSTSM):&lt;/strong&gt; &lt;br&gt;
They are a class of statistical models that combine elements of both state space models and time series models. These models are designed to handle complex time-dependent data with the added capability of capturing the underlying dynamics of the system generating the data.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;10.   Prophet Time Series Model:&lt;/strong&gt; &lt;br&gt;
It is an open-source forecasting tool developed by Facebook. It is designed for time series data with strong seasonal and holiday patterns. It is designed to make forecasting easy and approachable, especially for users without advanced expertise in time series analysis. It can handle missing data and outliers and is easy to use for quick forecasting tasks.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;11.   Generalized Autoregressive Conditional Heteroskedasticity (GARCH) Models:&lt;/strong&gt; &lt;br&gt;
They are a class of time series models used to analyze and forecast volatility in financial time series data. These models are especially important in the field of finance, where understanding and forecasting volatility is crucial for risk management, option pricing, portfolio optimization, and other financial applications.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;APPLICATIONS OF TIME SERIES MODELS&lt;/strong&gt;&lt;/p&gt;

&lt;p&gt;Time series models are suitably used in a wide range of fields;&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;1.    Finance&lt;/strong&gt;&lt;br&gt;
Financial analysts can leverage time series models to record sales, Stock Price Forecasting, risk management, Interest Rate Forecasting, Credit Scoring, asset allocation and many.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;2.    Healthcare&lt;/strong&gt;&lt;br&gt;
They can help analyze historical data, detect patterns, forecast future trends, and support decision-making in various aspects of the healthcare industry. They contribute to more efficient resource allocation, improved patient care, better financial planning, and enhanced public health responses.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;3.    Energy&lt;/strong&gt;&lt;br&gt;
They are used to analyze, forecast, and optimize various aspects of energy production, consumption, and distribution. They are a vital tool for the energy sector, contributing to more efficient energy production, distribution, and consumption.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;4.    Agriculture&lt;/strong&gt;&lt;br&gt;
They to analyse historical data, make predictions, and support decision-making for crop management, resource allocation, and sustainable farming practices. E.g they can take into account seasonal temperatures, the number of rainy days each month and other variables over the course of years, allowing agricultural workers to assess environmental conditions and ensure a successful harvest.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;5.    Cybersecurity&lt;/strong&gt;&lt;br&gt;
They are essential for identifying, mitigating, and responding to cyber threats. IT and cybersecurity teams can develop patterns in user behavior with time series models, allowing them to be aware of when behavior doesn’t align with normal trends.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;CONCLUSION&lt;/strong&gt;&lt;br&gt;
Time series modeling is an indispensable skill in the data scientist's toolkit. This guide has provided a comprehensive overview of time series models in data science, covering data characteristis, types and real world applications.  With use of this knowledge, data scientists can take advantage of the power of time series data to make informed decisions, identify trends, and make accurate predictions in a wide range of applications. Whether you are working in finance, energy, healthcare, or any other industry, mastering time series analysis is a valuable asset in your data science journey.&lt;/p&gt;

</description>
    </item>
    <item>
      <title>Article 2: Exploratory-Data-Analysis-using-Data-Visualization-Techniques</title>
      <dc:creator>Simon Ngotho</dc:creator>
      <pubDate>Mon, 16 Oct 2023 17:16:13 +0000</pubDate>
      <link>https://dev.to/ngotho/article-2-exploratory-data-analysis-using-data-visualization-techniques-3ebb</link>
      <guid>https://dev.to/ngotho/article-2-exploratory-data-analysis-using-data-visualization-techniques-3ebb</guid>
      <description>&lt;p&gt;Exploratory Data Analysis using Data Visualization Techniques&lt;/p&gt;

&lt;h1&gt;
  
  
  &lt;strong&gt;Preliminaries&lt;/strong&gt;
&lt;/h1&gt;

&lt;p&gt;Exploratory Data Analysis (EDA) is very crucial step in data analysis as it helps analysts and data scientists understand the datasets by looking at its structure and characteristics. It involves calculating summary statistics, identifying outliers, visualizing data distributions, exploring relationships between variables and performing hypothesis testing. This helps in uncovering insights that can help business make solid decisions of further data analysis. Data visualization tools helps present complex data in simpler and consumable manner even to no tech-savvy professionals. &lt;/p&gt;

&lt;p&gt;This article explores difference data visualization techniques to help get deeper insights on their application and significance. &lt;/p&gt;

&lt;h1&gt;
  
  
  &lt;strong&gt;Common Data Visualization Techniques for EDA&lt;/strong&gt;
&lt;/h1&gt;

&lt;h1&gt;
  
  
  &lt;strong&gt;1. Scatter Plots&lt;/strong&gt;
&lt;/h1&gt;

&lt;p&gt;They help visualize the relationship between two numerical variables. They key tools for identifying correlations, outliers, clusters or trends. A scatter plot with a regression line can also reveal linear relationships.&lt;br&gt;
&lt;/p&gt;

&lt;div class="highlight js-code-highlight"&gt;
&lt;pre class="highlight python"&gt;&lt;code&gt;&lt;span class="c1"&gt;#using matplotlib
&lt;/span&gt;
&lt;span class="kn"&gt;import&lt;/span&gt; &lt;span class="nn"&gt;matplotlib.pyplot&lt;/span&gt; &lt;span class="k"&gt;as&lt;/span&gt; &lt;span class="n"&gt;plt&lt;/span&gt;
&lt;span class="kn"&gt;import&lt;/span&gt; &lt;span class="nn"&gt;numpy&lt;/span&gt; &lt;span class="k"&gt;as&lt;/span&gt; &lt;span class="n"&gt;np&lt;/span&gt;

&lt;span class="n"&gt;x&lt;/span&gt; &lt;span class="o"&gt;=&lt;/span&gt; &lt;span class="n"&gt;np&lt;/span&gt;&lt;span class="p"&gt;.&lt;/span&gt;&lt;span class="n"&gt;random&lt;/span&gt;&lt;span class="p"&gt;.&lt;/span&gt;&lt;span class="n"&gt;rand&lt;/span&gt;&lt;span class="p"&gt;(&lt;/span&gt;&lt;span class="mi"&gt;50&lt;/span&gt;&lt;span class="p"&gt;)&lt;/span&gt; &lt;span class="c1"&gt;# random data
&lt;/span&gt;&lt;span class="n"&gt;y&lt;/span&gt; &lt;span class="o"&gt;=&lt;/span&gt; &lt;span class="n"&gt;x&lt;/span&gt; &lt;span class="o"&gt;+&lt;/span&gt; &lt;span class="mi"&gt;2&lt;/span&gt; &lt;span class="o"&gt;+&lt;/span&gt; &lt;span class="n"&gt;np&lt;/span&gt;&lt;span class="p"&gt;.&lt;/span&gt;&lt;span class="n"&gt;random&lt;/span&gt;&lt;span class="p"&gt;.&lt;/span&gt;&lt;span class="n"&gt;randn&lt;/span&gt;&lt;span class="p"&gt;(&lt;/span&gt;&lt;span class="mi"&gt;50&lt;/span&gt;&lt;span class="p"&gt;)&lt;/span&gt;  &lt;span class="c1"&gt;# using linear relationship simulation
&lt;/span&gt;
&lt;span class="c1"&gt;#Creating scatter plot
&lt;/span&gt;
&lt;span class="n"&gt;plt&lt;/span&gt;&lt;span class="p"&gt;.&lt;/span&gt;&lt;span class="n"&gt;scatter&lt;/span&gt;&lt;span class="p"&gt;(&lt;/span&gt;&lt;span class="n"&gt;x&lt;/span&gt;&lt;span class="p"&gt;,&lt;/span&gt; &lt;span class="n"&gt;y&lt;/span&gt;&lt;span class="p"&gt;,&lt;/span&gt; &lt;span class="n"&gt;c&lt;/span&gt;&lt;span class="o"&gt;=&lt;/span&gt;&lt;span class="s"&gt;'yellow'&lt;/span&gt;&lt;span class="p"&gt;,&lt;/span&gt; &lt;span class="n"&gt;edgecolor&lt;/span&gt;&lt;span class="o"&gt;=&lt;/span&gt;&lt;span class="s"&gt;'red'&lt;/span&gt;&lt;span class="p"&gt;,&lt;/span&gt; &lt;span class="n"&gt;marker&lt;/span&gt;&lt;span class="o"&gt;=&lt;/span&gt;&lt;span class="s"&gt;'.'&lt;/span&gt;&lt;span class="p"&gt;,&lt;/span&gt; &lt;span class="n"&gt;label&lt;/span&gt;&lt;span class="o"&gt;=&lt;/span&gt;&lt;span class="s"&gt;'Points'&lt;/span&gt;&lt;span class="p"&gt;)&lt;/span&gt;

&lt;span class="c1"&gt;#title the scatter plots
&lt;/span&gt;
&lt;span class="n"&gt;plt&lt;/span&gt;&lt;span class="p"&gt;.&lt;/span&gt;&lt;span class="n"&gt;xlabel&lt;/span&gt;&lt;span class="p"&gt;(&lt;/span&gt;&lt;span class="s"&gt;'X-Axis'&lt;/span&gt;&lt;span class="p"&gt;)&lt;/span&gt;
&lt;span class="n"&gt;plt&lt;/span&gt;&lt;span class="p"&gt;.&lt;/span&gt;&lt;span class="n"&gt;ylabel&lt;/span&gt;&lt;span class="p"&gt;(&lt;/span&gt;&lt;span class="s"&gt;'Y-Axis'&lt;/span&gt;&lt;span class="p"&gt;)&lt;/span&gt;
&lt;span class="n"&gt;plt&lt;/span&gt;&lt;span class="p"&gt;.&lt;/span&gt;&lt;span class="n"&gt;title&lt;/span&gt;&lt;span class="p"&gt;(&lt;/span&gt;&lt;span class="s"&gt;'Lux Tech Practice Scatter Plot1'&lt;/span&gt;&lt;span class="p"&gt;)&lt;/span&gt;


&lt;span class="c1"&gt;#using seaborn
&lt;/span&gt;
&lt;span class="kn"&gt;import&lt;/span&gt; &lt;span class="nn"&gt;seaborn&lt;/span&gt; &lt;span class="k"&gt;as&lt;/span&gt; &lt;span class="n"&gt;sns&lt;/span&gt;
&lt;span class="kn"&gt;import&lt;/span&gt; &lt;span class="nn"&gt;numpy&lt;/span&gt; &lt;span class="k"&gt;as&lt;/span&gt; &lt;span class="n"&gt;np&lt;/span&gt;

&lt;span class="n"&gt;x&lt;/span&gt; &lt;span class="o"&gt;=&lt;/span&gt; &lt;span class="n"&gt;np&lt;/span&gt;&lt;span class="p"&gt;.&lt;/span&gt;&lt;span class="n"&gt;random&lt;/span&gt;&lt;span class="p"&gt;.&lt;/span&gt;&lt;span class="n"&gt;rand&lt;/span&gt;&lt;span class="p"&gt;(&lt;/span&gt;&lt;span class="mi"&gt;50&lt;/span&gt;&lt;span class="p"&gt;)&lt;/span&gt; &lt;span class="c1"&gt;# random data
&lt;/span&gt;&lt;span class="n"&gt;y&lt;/span&gt; &lt;span class="o"&gt;=&lt;/span&gt; &lt;span class="n"&gt;x&lt;/span&gt; &lt;span class="o"&gt;+&lt;/span&gt; &lt;span class="mi"&gt;2&lt;/span&gt; &lt;span class="o"&gt;+&lt;/span&gt; &lt;span class="n"&gt;np&lt;/span&gt;&lt;span class="p"&gt;.&lt;/span&gt;&lt;span class="n"&gt;random&lt;/span&gt;&lt;span class="p"&gt;.&lt;/span&gt;&lt;span class="n"&gt;randn&lt;/span&gt;&lt;span class="p"&gt;(&lt;/span&gt;&lt;span class="mi"&gt;50&lt;/span&gt;&lt;span class="p"&gt;)&lt;/span&gt;  &lt;span class="c1"&gt;# using linear relationship simulation
&lt;/span&gt;
&lt;span class="c1"&gt;#Creating scatter plot
&lt;/span&gt;
&lt;span class="n"&gt;Sns&lt;/span&gt;&lt;span class="p"&gt;.&lt;/span&gt;&lt;span class="n"&gt;scatterplot&lt;/span&gt;&lt;span class="p"&gt;(&lt;/span&gt;&lt;span class="n"&gt;x&lt;/span&gt;&lt;span class="p"&gt;,&lt;/span&gt; &lt;span class="n"&gt;y&lt;/span&gt;&lt;span class="p"&gt;,&lt;/span&gt; &lt;span class="n"&gt;color&lt;/span&gt;&lt;span class="o"&gt;=&lt;/span&gt;&lt;span class="s"&gt;'yellow'&lt;/span&gt;&lt;span class="p"&gt;,&lt;/span&gt; &lt;span class="n"&gt;edgecolor&lt;/span&gt;&lt;span class="o"&gt;=&lt;/span&gt;&lt;span class="s"&gt;'red'&lt;/span&gt;&lt;span class="p"&gt;,&lt;/span&gt; &lt;span class="n"&gt;marker&lt;/span&gt;&lt;span class="o"&gt;=&lt;/span&gt;&lt;span class="s"&gt;'.'&lt;/span&gt;&lt;span class="p"&gt;,&lt;/span&gt; &lt;span class="n"&gt;label&lt;/span&gt;&lt;span class="o"&gt;=&lt;/span&gt;&lt;span class="s"&gt;'Points'&lt;/span&gt;&lt;span class="p"&gt;)&lt;/span&gt;

&lt;span class="c1"&gt;#title the scatter plots
&lt;/span&gt;
&lt;span class="n"&gt;plt&lt;/span&gt;&lt;span class="p"&gt;.&lt;/span&gt;&lt;span class="n"&gt;xlabel&lt;/span&gt;&lt;span class="p"&gt;(&lt;/span&gt;&lt;span class="s"&gt;'X-Axis'&lt;/span&gt;&lt;span class="p"&gt;)&lt;/span&gt;
&lt;span class="n"&gt;plt&lt;/span&gt;&lt;span class="p"&gt;.&lt;/span&gt;&lt;span class="n"&gt;ylabel&lt;/span&gt;&lt;span class="p"&gt;(&lt;/span&gt;&lt;span class="s"&gt;'Y-Axis'&lt;/span&gt;&lt;span class="p"&gt;)&lt;/span&gt;
&lt;span class="n"&gt;plt&lt;/span&gt;&lt;span class="p"&gt;.&lt;/span&gt;&lt;span class="n"&gt;title&lt;/span&gt;&lt;span class="p"&gt;(&lt;/span&gt;&lt;span class="s"&gt;'Lux Tech Practice Scatter Plot2'&lt;/span&gt;&lt;span class="p"&gt;)&lt;/span&gt;
&lt;/code&gt;&lt;/pre&gt;

&lt;/div&gt;



&lt;h1&gt;
  
  
  &lt;strong&gt;2. Histograms&lt;/strong&gt;
&lt;/h1&gt;

&lt;p&gt;Histograms help understand the distribution of numerical data. They display the frequency or count of data points within specified ranges or bins. Histograms helps in checking whether the data is normally distributed or how skewed it is.&lt;br&gt;
In Python, one can use Matplotlib or Seaborn as shown below;&lt;br&gt;
&lt;/p&gt;

&lt;div class="highlight js-code-highlight"&gt;
&lt;pre class="highlight python"&gt;&lt;code&gt;&lt;span class="c1"&gt;#Using matplotlib
&lt;/span&gt;
&lt;span class="kn"&gt;import&lt;/span&gt; &lt;span class="nn"&gt;matplotlib.pyplot&lt;/span&gt; &lt;span class="k"&gt;as&lt;/span&gt; &lt;span class="n"&gt;plt&lt;/span&gt;
&lt;span class="kn"&gt;import&lt;/span&gt; &lt;span class="nn"&gt;numpy&lt;/span&gt; &lt;span class="k"&gt;as&lt;/span&gt; &lt;span class="n"&gt;np&lt;/span&gt;

&lt;span class="n"&gt;data&lt;/span&gt; &lt;span class="o"&gt;=&lt;/span&gt; &lt;span class="n"&gt;np&lt;/span&gt;&lt;span class="p"&gt;.&lt;/span&gt;&lt;span class="n"&gt;random&lt;/span&gt;&lt;span class="p"&gt;.&lt;/span&gt;&lt;span class="n"&gt;randn&lt;/span&gt;&lt;span class="p"&gt;(&lt;/span&gt;&lt;span class="mi"&gt;1000&lt;/span&gt;&lt;span class="p"&gt;)&lt;/span&gt;  &lt;span class="c1"&gt;# use a Sample data – this random data
&lt;/span&gt;

&lt;span class="c1"&gt;#Creating histogram
&lt;/span&gt;
&lt;span class="n"&gt;plt&lt;/span&gt;&lt;span class="p"&gt;.&lt;/span&gt;&lt;span class="n"&gt;hist&lt;/span&gt;&lt;span class="p"&gt;(&lt;/span&gt;&lt;span class="n"&gt;data&lt;/span&gt;&lt;span class="p"&gt;,&lt;/span&gt; &lt;span class="n"&gt;bins&lt;/span&gt;&lt;span class="o"&gt;=&lt;/span&gt;&lt;span class="mi"&gt;15&lt;/span&gt;&lt;span class="p"&gt;,&lt;/span&gt; &lt;span class="n"&gt;edgecolor&lt;/span&gt;&lt;span class="o"&gt;=&lt;/span&gt;&lt;span class="s"&gt;'black'&lt;/span&gt;&lt;span class="p"&gt;,&lt;/span&gt; &lt;span class="n"&gt;alpha&lt;/span&gt;&lt;span class="o"&gt;=&lt;/span&gt;&lt;span class="mf"&gt;0.7&lt;/span&gt;&lt;span class="p"&gt;,&lt;/span&gt; &lt;span class="n"&gt;color&lt;/span&gt;&lt;span class="o"&gt;=&lt;/span&gt;&lt;span class="s"&gt;'yellow'&lt;/span&gt;&lt;span class="p"&gt;)&lt;/span&gt;

&lt;span class="c1"&gt;#Title the Histogram
&lt;/span&gt;
&lt;span class="n"&gt;plt&lt;/span&gt;&lt;span class="p"&gt;.&lt;/span&gt;&lt;span class="n"&gt;xlabel&lt;/span&gt;&lt;span class="p"&gt;(&lt;/span&gt;&lt;span class="s"&gt;'Values'&lt;/span&gt;&lt;span class="p"&gt;)&lt;/span&gt;
&lt;span class="n"&gt;plt&lt;/span&gt;&lt;span class="p"&gt;.&lt;/span&gt;&lt;span class="n"&gt;ylabel&lt;/span&gt;&lt;span class="p"&gt;(&lt;/span&gt;&lt;span class="s"&gt;'Frequency'&lt;/span&gt;&lt;span class="p"&gt;)&lt;/span&gt;
&lt;span class="n"&gt;plt&lt;/span&gt;&lt;span class="p"&gt;.&lt;/span&gt;&lt;span class="n"&gt;title&lt;/span&gt;&lt;span class="p"&gt;(&lt;/span&gt;&lt;span class="s"&gt;'Lux Tech Practice Histogram1'&lt;/span&gt;&lt;span class="p"&gt;)&lt;/span&gt;


&lt;span class="c1"&gt;#Using seaborn
&lt;/span&gt;
&lt;span class="kn"&gt;import&lt;/span&gt; &lt;span class="nn"&gt;seaborn&lt;/span&gt; &lt;span class="k"&gt;as&lt;/span&gt; &lt;span class="n"&gt;sns&lt;/span&gt;
&lt;span class="kn"&gt;import&lt;/span&gt; &lt;span class="nn"&gt;numpy&lt;/span&gt; &lt;span class="k"&gt;as&lt;/span&gt; &lt;span class="n"&gt;np&lt;/span&gt;

&lt;span class="n"&gt;data&lt;/span&gt; &lt;span class="o"&gt;=&lt;/span&gt; &lt;span class="n"&gt;np&lt;/span&gt;&lt;span class="p"&gt;.&lt;/span&gt;&lt;span class="n"&gt;random&lt;/span&gt;&lt;span class="p"&gt;.&lt;/span&gt;&lt;span class="n"&gt;randn&lt;/span&gt;&lt;span class="p"&gt;(&lt;/span&gt;&lt;span class="mi"&gt;500&lt;/span&gt;&lt;span class="p"&gt;)&lt;/span&gt; &lt;span class="c1"&gt;# use a Sample data – this random data
&lt;/span&gt;
&lt;span class="c1"&gt;#Creating histogram
&lt;/span&gt;
&lt;span class="n"&gt;sns&lt;/span&gt;&lt;span class="p"&gt;.&lt;/span&gt;&lt;span class="n"&gt;histplot&lt;/span&gt; &lt;span class="p"&gt;(&lt;/span&gt;&lt;span class="n"&gt;data&lt;/span&gt;&lt;span class="p"&gt;,&lt;/span&gt; &lt;span class="n"&gt;bins&lt;/span&gt;&lt;span class="o"&gt;=&lt;/span&gt;&lt;span class="mi"&gt;15&lt;/span&gt;&lt;span class="p"&gt;,&lt;/span&gt; &lt;span class="n"&gt;color&lt;/span&gt;&lt;span class="o"&gt;=&lt;/span&gt;&lt;span class="s"&gt;'yellow'&lt;/span&gt;&lt;span class="p"&gt;,&lt;/span&gt; &lt;span class="n"&gt;edgecolor&lt;/span&gt;&lt;span class="o"&gt;=&lt;/span&gt;&lt;span class="s"&gt;'red'&lt;/span&gt;&lt;span class="p"&gt;)&lt;/span&gt;

&lt;span class="c1"&gt;#Title the Histogram
&lt;/span&gt;
&lt;span class="n"&gt;plt&lt;/span&gt;&lt;span class="p"&gt;.&lt;/span&gt;&lt;span class="n"&gt;xlabel&lt;/span&gt; &lt;span class="p"&gt;(&lt;/span&gt;&lt;span class="s"&gt;'values'&lt;/span&gt;&lt;span class="p"&gt;)&lt;/span&gt;
&lt;span class="n"&gt;plt&lt;/span&gt;&lt;span class="p"&gt;.&lt;/span&gt;&lt;span class="n"&gt;ylabel&lt;/span&gt; &lt;span class="p"&gt;(&lt;/span&gt;&lt;span class="s"&gt;'frequency'&lt;/span&gt;&lt;span class="p"&gt;)&lt;/span&gt;
&lt;span class="n"&gt;plt&lt;/span&gt;&lt;span class="p"&gt;.&lt;/span&gt;&lt;span class="n"&gt;title&lt;/span&gt; &lt;span class="p"&gt;(&lt;/span&gt;&lt;span class="s"&gt;'Lux Tech Practice Histogram2'&lt;/span&gt;&lt;span class="p"&gt;)&lt;/span&gt;
&lt;/code&gt;&lt;/pre&gt;

&lt;/div&gt;



&lt;h1&gt;
  
  
  &lt;strong&gt;3. Bar Charts&lt;/strong&gt;
&lt;/h1&gt;

&lt;p&gt;They show the frequency or count of different categories for easy identification of the most common/rare values. Stacked bar charts are also used to show the distribution of categories within subgroups.&lt;br&gt;
&lt;/p&gt;

&lt;div class="highlight js-code-highlight"&gt;
&lt;pre class="highlight python"&gt;&lt;code&gt;&lt;span class="c1"&gt;#Using matplotlib
&lt;/span&gt;&lt;span class="kn"&gt;import&lt;/span&gt; &lt;span class="nn"&gt;matplotlib.pyplot&lt;/span&gt; &lt;span class="k"&gt;as&lt;/span&gt; &lt;span class="n"&gt;plt&lt;/span&gt;
&lt;span class="n"&gt;categories&lt;/span&gt; &lt;span class="o"&gt;=&lt;/span&gt; &lt;span class="p"&gt;[&lt;/span&gt;&lt;span class="s"&gt;'Nakuru'&lt;/span&gt;&lt;span class="p"&gt;,&lt;/span&gt; &lt;span class="s"&gt;'Kisumu'&lt;/span&gt;&lt;span class="p"&gt;,&lt;/span&gt; &lt;span class="s"&gt;'Mombasa'&lt;/span&gt;&lt;span class="p"&gt;,&lt;/span&gt; &lt;span class="s"&gt;'Nairobi'&lt;/span&gt;&lt;span class="p"&gt;]&lt;/span&gt;
&lt;span class="n"&gt;values&lt;/span&gt; &lt;span class="o"&gt;=&lt;/span&gt; &lt;span class="p"&gt;[&lt;/span&gt;&lt;span class="mf"&gt;1.9&lt;/span&gt;&lt;span class="p"&gt;,&lt;/span&gt; &lt;span class="mf"&gt;2.4&lt;/span&gt;&lt;span class="p"&gt;,&lt;/span&gt; &lt;span class="mf"&gt;3.2&lt;/span&gt;&lt;span class="p"&gt;,&lt;/span&gt; &lt;span class="mf"&gt;5.5&lt;/span&gt;&lt;span class="p"&gt;]&lt;/span&gt; &lt;span class="c1"&gt;# Sample data
&lt;/span&gt;
&lt;span class="c1"&gt;#Creating bar chart
&lt;/span&gt;
&lt;span class="n"&gt;plt&lt;/span&gt;&lt;span class="p"&gt;.&lt;/span&gt;&lt;span class="n"&gt;bar&lt;/span&gt;&lt;span class="p"&gt;(&lt;/span&gt;&lt;span class="n"&gt;categories&lt;/span&gt;&lt;span class="p"&gt;,&lt;/span&gt; &lt;span class="n"&gt;values&lt;/span&gt;&lt;span class="p"&gt;,&lt;/span&gt; &lt;span class="n"&gt;color&lt;/span&gt;&lt;span class="o"&gt;=&lt;/span&gt;&lt;span class="s"&gt;'yellow'&lt;/span&gt;&lt;span class="p"&gt;,&lt;/span&gt; &lt;span class="n"&gt;edgecolor&lt;/span&gt;&lt;span class="o"&gt;=&lt;/span&gt;&lt;span class="s"&gt;'red'&lt;/span&gt;&lt;span class="p"&gt;)&lt;/span&gt;

&lt;span class="c1"&gt;#titling and labelling 
&lt;/span&gt;
&lt;span class="n"&gt;plt&lt;/span&gt;&lt;span class="p"&gt;.&lt;/span&gt;&lt;span class="n"&gt;xlabel&lt;/span&gt;&lt;span class="p"&gt;(&lt;/span&gt;&lt;span class="s"&gt;'Cities'&lt;/span&gt;&lt;span class="p"&gt;)&lt;/span&gt;
&lt;span class="n"&gt;plt&lt;/span&gt;&lt;span class="p"&gt;.&lt;/span&gt;&lt;span class="n"&gt;ylabel&lt;/span&gt;&lt;span class="p"&gt;(&lt;/span&gt;&lt;span class="s"&gt;'Population in Millions'&lt;/span&gt;&lt;span class="p"&gt;)&lt;/span&gt;
&lt;span class="n"&gt;plt&lt;/span&gt;&lt;span class="p"&gt;.&lt;/span&gt;&lt;span class="n"&gt;title&lt;/span&gt;&lt;span class="p"&gt;(&lt;/span&gt;&lt;span class="s"&gt;'Lux Tech Practice Bar Chart1'&lt;/span&gt;&lt;span class="p"&gt;)&lt;/span&gt;
&lt;/code&gt;&lt;/pre&gt;

&lt;/div&gt;



&lt;h1&gt;
  
  
  &lt;strong&gt;4. Box Plots&lt;/strong&gt;
&lt;/h1&gt;

&lt;p&gt;Box plots shows the median, quartiles, and potential outliers and are useful for comparing distributions between different categories or groups within the data.&lt;br&gt;
&lt;/p&gt;

&lt;div class="highlight js-code-highlight"&gt;
&lt;pre class="highlight python"&gt;&lt;code&gt;&lt;span class="c1"&gt;#Using matplotlib
&lt;/span&gt;
&lt;span class="kn"&gt;import&lt;/span&gt; &lt;span class="nn"&gt;matplotlib.pyplot&lt;/span&gt; &lt;span class="k"&gt;as&lt;/span&gt; &lt;span class="n"&gt;plt&lt;/span&gt;
&lt;span class="kn"&gt;import&lt;/span&gt; &lt;span class="nn"&gt;numpy&lt;/span&gt; &lt;span class="k"&gt;as&lt;/span&gt; &lt;span class="n"&gt;np&lt;/span&gt;

&lt;span class="n"&gt;data&lt;/span&gt; &lt;span class="o"&gt;=&lt;/span&gt; &lt;span class="p"&gt;[&lt;/span&gt;&lt;span class="n"&gt;np&lt;/span&gt;&lt;span class="p"&gt;.&lt;/span&gt;&lt;span class="n"&gt;random&lt;/span&gt;&lt;span class="p"&gt;.&lt;/span&gt;&lt;span class="n"&gt;normal&lt;/span&gt;&lt;span class="p"&gt;(&lt;/span&gt;&lt;span class="mi"&gt;0&lt;/span&gt;&lt;span class="p"&gt;,&lt;/span&gt; &lt;span class="mi"&gt;1&lt;/span&gt;&lt;span class="p"&gt;,&lt;/span&gt; &lt;span class="mi"&gt;50&lt;/span&gt;&lt;span class="p"&gt;),&lt;/span&gt; &lt;span class="n"&gt;np&lt;/span&gt;&lt;span class="p"&gt;.&lt;/span&gt;&lt;span class="n"&gt;random&lt;/span&gt;&lt;span class="p"&gt;.&lt;/span&gt;&lt;span class="n"&gt;normal&lt;/span&gt;&lt;span class="p"&gt;(&lt;/span&gt;&lt;span class="mi"&gt;2&lt;/span&gt;&lt;span class="p"&gt;,&lt;/span&gt; &lt;span class="mi"&gt;1&lt;/span&gt;&lt;span class="p"&gt;,&lt;/span&gt; &lt;span class="mi"&gt;50&lt;/span&gt;&lt;span class="p"&gt;),&lt;/span&gt; &lt;span class="n"&gt;np&lt;/span&gt;&lt;span class="p"&gt;.&lt;/span&gt;&lt;span class="n"&gt;random&lt;/span&gt;&lt;span class="p"&gt;.&lt;/span&gt;&lt;span class="n"&gt;normal&lt;/span&gt;&lt;span class="p"&gt;(&lt;/span&gt;&lt;span class="mi"&gt;4&lt;/span&gt;&lt;span class="p"&gt;,&lt;/span&gt; &lt;span class="mi"&gt;1&lt;/span&gt;&lt;span class="p"&gt;,&lt;/span&gt; &lt;span class="mi"&gt;50&lt;/span&gt;&lt;span class="p"&gt;)]&lt;/span&gt; &lt;span class="c1"&gt;# Sample data
&lt;/span&gt;
&lt;span class="c1"&gt;#Creating a box plot
&lt;/span&gt;
&lt;span class="n"&gt;plt&lt;/span&gt;&lt;span class="p"&gt;.&lt;/span&gt;&lt;span class="n"&gt;boxplot&lt;/span&gt;&lt;span class="p"&gt;(&lt;/span&gt;&lt;span class="n"&gt;data&lt;/span&gt;&lt;span class="p"&gt;,&lt;/span&gt; &lt;span class="n"&gt;labels&lt;/span&gt;&lt;span class="o"&gt;=&lt;/span&gt;&lt;span class="p"&gt;[&lt;/span&gt;&lt;span class="s"&gt;'Category A'&lt;/span&gt;&lt;span class="p"&gt;,&lt;/span&gt; &lt;span class="s"&gt;'Category B'&lt;/span&gt;&lt;span class="p"&gt;,&lt;/span&gt; &lt;span class="s"&gt;'Category C'&lt;/span&gt;&lt;span class="p"&gt;])&lt;/span&gt;

&lt;span class="c1"&gt;#Title and label
&lt;/span&gt;
&lt;span class="n"&gt;plt&lt;/span&gt;&lt;span class="p"&gt;.&lt;/span&gt;&lt;span class="n"&gt;xlabel&lt;/span&gt;&lt;span class="p"&gt;(&lt;/span&gt;&lt;span class="s"&gt;'Categories'&lt;/span&gt;&lt;span class="p"&gt;)&lt;/span&gt;
&lt;span class="n"&gt;plt&lt;/span&gt;&lt;span class="p"&gt;.&lt;/span&gt;&lt;span class="n"&gt;ylabel&lt;/span&gt;&lt;span class="p"&gt;(&lt;/span&gt;&lt;span class="s"&gt;'Values'&lt;/span&gt;&lt;span class="p"&gt;)&lt;/span&gt;
&lt;span class="n"&gt;plt&lt;/span&gt;&lt;span class="p"&gt;.&lt;/span&gt;&lt;span class="n"&gt;title&lt;/span&gt;&lt;span class="p"&gt;(&lt;/span&gt;&lt;span class="s"&gt;'Lux Tech Practice Box Plot1'&lt;/span&gt;&lt;span class="p"&gt;)&lt;/span&gt;
&lt;/code&gt;&lt;/pre&gt;

&lt;/div&gt;



&lt;h1&gt;
  
  
  &lt;strong&gt;5. Time Series Plots&lt;/strong&gt;
&lt;/h1&gt;

&lt;p&gt;They are helpful to analysts as they help them understand how data evolves over time, identify trends, seasonality, and anomalies, and make predictions based on historical data.&lt;br&gt;
&lt;/p&gt;

&lt;div class="highlight js-code-highlight"&gt;
&lt;pre class="highlight python"&gt;&lt;code&gt;&lt;span class="kn"&gt;import&lt;/span&gt; &lt;span class="nn"&gt;matplotlib.pyplot&lt;/span&gt; &lt;span class="k"&gt;as&lt;/span&gt; &lt;span class="n"&gt;plt&lt;/span&gt;
&lt;span class="kn"&gt;import&lt;/span&gt; &lt;span class="nn"&gt;numpy&lt;/span&gt; &lt;span class="k"&gt;as&lt;/span&gt; &lt;span class="n"&gt;np&lt;/span&gt;
&lt;span class="kn"&gt;import&lt;/span&gt; &lt;span class="nn"&gt;pandas&lt;/span&gt; &lt;span class="k"&gt;as&lt;/span&gt; &lt;span class="n"&gt;pd&lt;/span&gt;

&lt;span class="n"&gt;time_period&lt;/span&gt; &lt;span class="o"&gt;=&lt;/span&gt; &lt;span class="n"&gt;pd&lt;/span&gt;&lt;span class="p"&gt;.&lt;/span&gt;&lt;span class="n"&gt;date_range&lt;/span&gt; &lt;span class="p"&gt;(&lt;/span&gt;&lt;span class="n"&gt;start&lt;/span&gt;&lt;span class="o"&gt;=&lt;/span&gt;&lt;span class="s"&gt;'2000-01-01'&lt;/span&gt;&lt;span class="p"&gt;,&lt;/span&gt; &lt;span class="n"&gt;periods&lt;/span&gt;&lt;span class="o"&gt;=&lt;/span&gt;&lt;span class="mi"&gt;30&lt;/span&gt;&lt;span class="p"&gt;,&lt;/span&gt; &lt;span class="n"&gt;freq&lt;/span&gt;&lt;span class="o"&gt;=&lt;/span&gt;&lt;span class="s"&gt;'M'&lt;/span&gt;&lt;span class="p"&gt;)&lt;/span&gt;
&lt;span class="n"&gt;data&lt;/span&gt; &lt;span class="o"&gt;=&lt;/span&gt; &lt;span class="n"&gt;np&lt;/span&gt;&lt;span class="p"&gt;.&lt;/span&gt;&lt;span class="n"&gt;cumsum&lt;/span&gt;&lt;span class="p"&gt;(&lt;/span&gt;&lt;span class="n"&gt;np&lt;/span&gt;&lt;span class="p"&gt;.&lt;/span&gt;&lt;span class="n"&gt;random&lt;/span&gt;&lt;span class="p"&gt;.&lt;/span&gt;&lt;span class="n"&gt;randn&lt;/span&gt;&lt;span class="p"&gt;(&lt;/span&gt;&lt;span class="mi"&gt;30&lt;/span&gt;&lt;span class="p"&gt;))&lt;/span&gt; &lt;span class="c1"&gt;# Generate sample time series data
&lt;/span&gt;

&lt;span class="c1"&gt;#Creating time series plot
&lt;/span&gt;
&lt;span class="n"&gt;plt&lt;/span&gt;&lt;span class="p"&gt;.&lt;/span&gt;&lt;span class="n"&gt;figure&lt;/span&gt;&lt;span class="p"&gt;(&lt;/span&gt;&lt;span class="n"&gt;figsize&lt;/span&gt;&lt;span class="o"&gt;=&lt;/span&gt;&lt;span class="p"&gt;(&lt;/span&gt;&lt;span class="mi"&gt;14&lt;/span&gt;&lt;span class="p"&gt;,&lt;/span&gt; &lt;span class="mi"&gt;7&lt;/span&gt;&lt;span class="p"&gt;))&lt;/span&gt;
&lt;span class="n"&gt;plt&lt;/span&gt;&lt;span class="p"&gt;.&lt;/span&gt;&lt;span class="n"&gt;plot&lt;/span&gt;&lt;span class="p"&gt;(&lt;/span&gt;&lt;span class="n"&gt;time_period&lt;/span&gt;&lt;span class="p"&gt;,&lt;/span&gt; &lt;span class="n"&gt;data&lt;/span&gt;&lt;span class="p"&gt;,&lt;/span&gt; &lt;span class="n"&gt;color&lt;/span&gt;&lt;span class="o"&gt;=&lt;/span&gt;&lt;span class="s"&gt;'yellow'&lt;/span&gt;&lt;span class="p"&gt;,&lt;/span&gt; &lt;span class="n"&gt;marker&lt;/span&gt;&lt;span class="o"&gt;=&lt;/span&gt;&lt;span class="s"&gt;'d'&lt;/span&gt;&lt;span class="p"&gt;,&lt;/span&gt; &lt;span class="n"&gt;linestyle&lt;/span&gt;&lt;span class="o"&gt;=&lt;/span&gt;&lt;span class="s"&gt;'-'&lt;/span&gt;&lt;span class="p"&gt;)&lt;/span&gt;

&lt;span class="c1"&gt;#labels and title
&lt;/span&gt;
&lt;span class="n"&gt;plt&lt;/span&gt;&lt;span class="p"&gt;.&lt;/span&gt;&lt;span class="n"&gt;grid&lt;/span&gt;&lt;span class="p"&gt;(&lt;/span&gt;&lt;span class="bp"&gt;True&lt;/span&gt;&lt;span class="p"&gt;)&lt;/span&gt;
&lt;span class="n"&gt;plt&lt;/span&gt;&lt;span class="p"&gt;.&lt;/span&gt;&lt;span class="n"&gt;xlabel&lt;/span&gt;&lt;span class="p"&gt;(&lt;/span&gt;&lt;span class="s"&gt;'Time'&lt;/span&gt;&lt;span class="p"&gt;)&lt;/span&gt;
&lt;span class="n"&gt;plt&lt;/span&gt;&lt;span class="p"&gt;.&lt;/span&gt;&lt;span class="n"&gt;ylabel&lt;/span&gt;&lt;span class="p"&gt;(&lt;/span&gt;&lt;span class="s"&gt;'Values'&lt;/span&gt;&lt;span class="p"&gt;)&lt;/span&gt;
&lt;span class="n"&gt;plt&lt;/span&gt;&lt;span class="p"&gt;.&lt;/span&gt;&lt;span class="n"&gt;title&lt;/span&gt;&lt;span class="p"&gt;(&lt;/span&gt;&lt;span class="s"&gt;'Time Series Plot Example'&lt;/span&gt;&lt;span class="p"&gt;)&lt;/span&gt;
&lt;/code&gt;&lt;/pre&gt;

&lt;/div&gt;



&lt;h1&gt;
  
  
  &lt;strong&gt;6. Heatmaps&lt;/strong&gt;
&lt;/h1&gt;

&lt;p&gt;Heatmaps help visualize relationships between variables by use of colour intensity to represent the strength and direction of correlations between pairs of variables. This way, they make it easier to identify patterns in large datasets.&lt;br&gt;
&lt;/p&gt;

&lt;div class="highlight js-code-highlight"&gt;
&lt;pre class="highlight python"&gt;&lt;code&gt;&lt;span class="c1"&gt;#using seaborn
&lt;/span&gt;
&lt;span class="kn"&gt;import&lt;/span&gt; &lt;span class="nn"&gt;seaborn&lt;/span&gt; &lt;span class="k"&gt;as&lt;/span&gt; &lt;span class="n"&gt;sns&lt;/span&gt;
&lt;span class="kn"&gt;import&lt;/span&gt; &lt;span class="nn"&gt;numpy&lt;/span&gt; &lt;span class="k"&gt;as&lt;/span&gt; &lt;span class="n"&gt;np&lt;/span&gt;

&lt;span class="n"&gt;data&lt;/span&gt; &lt;span class="o"&gt;=&lt;/span&gt; &lt;span class="n"&gt;np&lt;/span&gt;&lt;span class="p"&gt;.&lt;/span&gt;&lt;span class="n"&gt;random&lt;/span&gt;&lt;span class="p"&gt;.&lt;/span&gt;&lt;span class="n"&gt;rand&lt;/span&gt;&lt;span class="p"&gt;(&lt;/span&gt;&lt;span class="mi"&gt;10&lt;/span&gt;&lt;span class="p"&gt;,&lt;/span&gt; &lt;span class="mi"&gt;5&lt;/span&gt;&lt;span class="p"&gt;)&lt;/span&gt;  &lt;span class="c1"&gt;# sample data
&lt;/span&gt;
&lt;span class="c1"&gt;#Creating a heatmap 
&lt;/span&gt;
&lt;span class="n"&gt;sns&lt;/span&gt;&lt;span class="p"&gt;.&lt;/span&gt;&lt;span class="n"&gt;heatmap&lt;/span&gt;&lt;span class="p"&gt;(&lt;/span&gt;&lt;span class="n"&gt;data&lt;/span&gt;&lt;span class="p"&gt;,&lt;/span&gt; &lt;span class="n"&gt;annot&lt;/span&gt;&lt;span class="o"&gt;=&lt;/span&gt;&lt;span class="bp"&gt;True&lt;/span&gt;&lt;span class="p"&gt;,&lt;/span&gt; &lt;span class="n"&gt;cmap&lt;/span&gt;&lt;span class="o"&gt;=&lt;/span&gt;&lt;span class="s"&gt;"YlGnBu"&lt;/span&gt;&lt;span class="p"&gt;)&lt;/span&gt;

&lt;span class="c1"&gt;#label and title
&lt;/span&gt;
&lt;span class="n"&gt;plt&lt;/span&gt;&lt;span class="p"&gt;.&lt;/span&gt;&lt;span class="n"&gt;xlabel&lt;/span&gt;&lt;span class="p"&gt;(&lt;/span&gt;&lt;span class="s"&gt;'Variables'&lt;/span&gt;&lt;span class="p"&gt;)&lt;/span&gt;
&lt;span class="n"&gt;plt&lt;/span&gt;&lt;span class="p"&gt;.&lt;/span&gt;&lt;span class="n"&gt;ylabel&lt;/span&gt;&lt;span class="p"&gt;(&lt;/span&gt;&lt;span class="s"&gt;'Variables'&lt;/span&gt;&lt;span class="p"&gt;)&lt;/span&gt;
&lt;span class="n"&gt;plt&lt;/span&gt;&lt;span class="p"&gt;.&lt;/span&gt;&lt;span class="n"&gt;title&lt;/span&gt;&lt;span class="p"&gt;(&lt;/span&gt;&lt;span class="s"&gt;'Lux Tech Practice Heatmap1’)
&lt;/span&gt;&lt;/code&gt;&lt;/pre&gt;

&lt;/div&gt;



&lt;h1&gt;
  
  
  &lt;strong&gt;Conclusion&lt;/strong&gt;
&lt;/h1&gt;

&lt;p&gt;Data visualization techniques make complex data more interpretable, enabling the identification of patterns, trends, outliers, and data quality issues. It becomes easy for users to interpret regardless of their background and hence helping them make data-driven decisions to solve problems. &lt;/p&gt;

</description>
    </item>
    <item>
      <title>Article 1: Data Science for Beginners: 2023 - 2024 Complete Roadmap</title>
      <dc:creator>Simon Ngotho</dc:creator>
      <pubDate>Sun, 01 Oct 2023 12:17:49 +0000</pubDate>
      <link>https://dev.to/ngotho/data-science-for-beginners-2023-2024-complete-roadmap-2g2m</link>
      <guid>https://dev.to/ngotho/data-science-for-beginners-2023-2024-complete-roadmap-2g2m</guid>
      <description>&lt;p&gt;&lt;strong&gt;Preliminaries&lt;/strong&gt;&lt;br&gt;
Establish what it takes to be a data scientist. By joining LUV Tech Academy, I believe that I will get adequate knowledge that will help me kick-start my career in data science/data analytics. &lt;br&gt;
I am looking forward to Iearning programming languages that will help me navigate data science i.e. Python, SQL.&lt;br&gt;
I will also familiarize myself with knowledge in Mathematics, Probability and Statistics on areas applicable in data science. &lt;/p&gt;

&lt;p&gt;&lt;strong&gt;Learn SQL&lt;/strong&gt;&lt;br&gt;
I intend to learn SQL Language through Online Resources/sites e.g Udemy. I also intend to listen to you tube contents to help me improve on my proficiency. &lt;/p&gt;

&lt;p&gt;&lt;strong&gt;Learn Python&lt;/strong&gt;&lt;br&gt;
I intend to learn Python Language and related libraries like Pandas, Numpy, through Online Resources/sites e.g Udemy. I also intend to listen to youtube contents to help me improve on my proficiency.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;Learn Mathematics, Statistics and Probability&lt;/strong&gt;&lt;br&gt;
I understand this knowledge is live and applicable in data science and as such it will be crucial to grasp the concepts. I intend to use online resources as well as materials shared in the data camp to improve on my proficiency.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;Data Manipulation&lt;/strong&gt;&lt;br&gt;
I plan to learn data manipulation through SQL and Python. I will do a lot of practices using MySql, DB Browser for SQLite, Anaconda (Jupiter Notebook) etc. I will supplement my knowledge through Youtube Content as well.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;Data Analysis&lt;/strong&gt;&lt;br&gt;
I will use SQL and Python to carry out data analysis. I will get content from sites like Udemy, Codecademy, Cisco, SQLZoo and other internet sources. &lt;/p&gt;

&lt;p&gt;&lt;strong&gt;Data Visualization/Presentation&lt;/strong&gt;&lt;br&gt;
I intend to learn visualization through different tools like Power BI, Tableau and Matplotlib from online sources.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;Machine Learning&lt;/strong&gt;&lt;br&gt;
I intend to use knowledge learnt in the data-camp to set base for machine learning. I will supplement the learning with online contents from Udemy, Coursera and others. &lt;/p&gt;

&lt;p&gt;&lt;strong&gt;Deep Learning&lt;/strong&gt;&lt;br&gt;
I intend to get basic knowledge of deep learning since I am likely to interact with it in course of my work. I will use Udemy, Coursera, Udacity or Edx.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;Real Projects&lt;/strong&gt;&lt;br&gt;
I will purpose to undertake real-world project on data science using the Data-camp materials and other sourced resources. This will help me get prepared for real work assignments.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;Continuous learning&lt;/strong&gt;&lt;br&gt;
I will purpose to practice data science programming everyday using Python and SQL. I will source for projects from Online sites and join Online communities to enable me network with professionals. I will also participate in Data Seminars, Webinars etc. &lt;/p&gt;

</description>
    </item>
  </channel>
</rss>
