<?xml version="1.0" encoding="UTF-8"?>
<rss version="2.0" xmlns:atom="http://www.w3.org/2005/Atom" xmlns:dc="http://purl.org/dc/elements/1.1/">
  <channel>
    <title>DEV Community: Victor Alando</title>
    <description>The latest articles on DEV Community by Victor Alando (@victoralando).</description>
    <link>https://dev.to/victoralando</link>
    <image>
      <url>https://media2.dev.to/dynamic/image/width=90,height=90,fit=cover,gravity=auto,format=auto/https:%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Fuser%2Fprofile_image%2F1175714%2F6abc69ac-9f44-4e45-b739-d499703d29a4.jpg</url>
      <title>DEV Community: Victor Alando</title>
      <link>https://dev.to/victoralando</link>
    </image>
    <atom:link rel="self" type="application/rss+xml" href="https://dev.to/feed/victoralando"/>
    <language>en</language>
    <item>
      <title>K-means Clustering Using the Elbow Method.</title>
      <dc:creator>Victor Alando</dc:creator>
      <pubDate>Mon, 06 Jan 2025 07:51:17 +0000</pubDate>
      <link>https://dev.to/victoralando/k-means-clustering-using-the-elbow-method-235j</link>
      <guid>https://dev.to/victoralando/k-means-clustering-using-the-elbow-method-235j</guid>
      <description>&lt;h2&gt;
  
  
  Introduction
&lt;/h2&gt;

&lt;p&gt;Clustering or cluster analysis is machine learning technique, which groups the unlabeled dataset. It can be said that as "&lt;em&gt;way of grouping the data points into different clusters, consisting of similar data points. The objects with the possible similarities remain in a group and those that have less or no similarities with another group&lt;/em&gt;"&lt;/p&gt;

&lt;p&gt;Let's understand the clustering technique with the real-world example of Mall. When customers visit any shopping mall, we can observe that the things with similar usage are grouped together. Such as the t-shirts are grouped in one section, and trousers are at other sections, similarly, at vegetable sections, apple, bananas, Mangoes, e.t.c are grouped in a separate section, so that customers can easily find out the things. The clustering technique also works in the same way. Other examples of clustering are grouping of documents according to topics.&lt;/p&gt;

&lt;h2&gt;
  
  
  Python Implementation of K-means Clustering Algorithm.
&lt;/h2&gt;

&lt;h2&gt;
  
  
  Prerequisites
&lt;/h2&gt;

&lt;ul&gt;
&lt;li&gt;&lt;strong&gt;What is K-means Clustering Algorithm.&lt;/strong&gt;&lt;/li&gt;
&lt;li&gt;&lt;strong&gt;How does the k-means algorithm work?&lt;/strong&gt;&lt;/li&gt;
&lt;li&gt;&lt;strong&gt;How to find and choose the value of "k: number of clusters in k-means clustering.&lt;/strong&gt;&lt;/li&gt;
&lt;li&gt;&lt;strong&gt;Data preprocessing.&lt;/strong&gt;&lt;/li&gt;
&lt;li&gt;&lt;strong&gt;Standardization and feature scaling.&lt;/strong&gt;&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;Fitting the training and Data Transformation&lt;/strong&gt;.&lt;/li&gt;
&lt;li&gt;&lt;strong&gt;Training the K-means Algorithm on the Training Dataset.&lt;/strong&gt;&lt;/li&gt;
&lt;li&gt;&lt;strong&gt;Make Predictions.&lt;/strong&gt;&lt;/li&gt;
&lt;li&gt;&lt;strong&gt;Inspect the coordinates of the 5 centroids&lt;/strong&gt;&lt;/li&gt;
&lt;li&gt;&lt;strong&gt;Finding the Optimal (k) number of clusters using the Elbow Method.&lt;/strong&gt;&lt;/li&gt;
&lt;li&gt;&lt;strong&gt;Visualizing the Clusters&lt;/strong&gt;&lt;/li&gt;
&lt;li&gt;&lt;strong&gt;Summary Findings&lt;/strong&gt;&lt;/li&gt;
&lt;/ul&gt;

&lt;h2&gt;
  
  
  What is K-means Clustering Algorithm?
&lt;/h2&gt;

&lt;p&gt;K-Means clustering is an unsupervised learning algorith, which groups the unlabeled dataset into different clusters. Here k defines the number of pre-defined clusters that need to be created in the line process, as if K=2, there will be two clusters, and K=3 it means that there will be clusters, and so on.&lt;/p&gt;

&lt;p&gt;It is an iterative algorithm that divides the unlabeled dataset into k different clusters in such a way that each dataset belongs only to only one group that has similar properties&lt;/p&gt;

&lt;p&gt;It allows us to cluster the data into different groups and a convenient way to discover the categories of groups in the unlabeled dataset on its own without the need for any training.&lt;/p&gt;

&lt;p&gt;It is a centroid-based algorithm, where clusters are associated with centroid. The main aim of this algorithm is to minimize the sum of distances between the data point and their correspond clusters.&lt;/p&gt;

&lt;p&gt;The algorithm takes the unlabeled dataset as input, divides the dataset into k-number of clusters, and repeats the process until it does not find the best clusters. The value of k should be predetermined in this algorithm.&lt;/p&gt;

&lt;p&gt;The k-means clustering algorithm mainly performs two tasks: - Determines the best value for k center points or centroids by an iterative process. - Assigns each data point to its closest k-center. Those data points which are near to the particular k-center, create a cluster.&lt;/p&gt;

&lt;p&gt;Hence each cluster has data points with some commonalities, and it is away from other clusters. &lt;/p&gt;

&lt;p&gt;Consider the below diagram that explains the working of the K-means Clustering Algorithm.&lt;/p&gt;

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fyqorsy60pnxb1npr33qb.png" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fyqorsy60pnxb1npr33qb.png" alt=" " width="800" height="331"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;How does the k-means algorithm work?&lt;/strong&gt;&lt;/p&gt;

&lt;p&gt;The workings of the Kmeans algorithm can explained in the below steps:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;&lt;p&gt;Select the number of k to decide the number of clusters.&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;Select random K points or centroids (it can be from the input data set).&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;Assign each data point to their closet centroid, which will form the predefined &lt;strong&gt;k&lt;/strong&gt; clusters.&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;Calculate the variance and place a new centroid of each cluster.&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;Repeat the third step which means, reassign each data point to the new closet centroid of each cluster.&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;If any reassignment occurs, then go to step 4 else go to finish.&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;The model is ready.&lt;/p&gt;&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;Let’s now understand the above steps with the help of a visual plots: Suppose we have two variables &lt;strong&gt;M1&lt;/strong&gt; and &lt;strong&gt;M2.&lt;/strong&gt;&lt;/p&gt;

&lt;p&gt;The x-y axis scatter plot of these two variables is given below:&lt;/p&gt;

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fguh62xhyx8pw06nnpkn7.png" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fguh62xhyx8pw06nnpkn7.png" alt=" " width="800" height="637"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;&lt;p&gt;Let’s take number k of clusters, i.e. K=2, to identify the dataset and to put them into different clusters. It means that we will try to group these datasets into two different clusters.&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;We need to choose some random k points or centroids to form the cluster. These points can be either or any other point. So, here wea re selecting the below two points as k points, which are not the part of our dataset.&lt;/p&gt;&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;Consider the below visual plots:&lt;/p&gt;

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Ftac67gpl1m7fnjn9t8xi.png" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Ftac67gpl1m7fnjn9t8xi.png" alt=" " width="800" height="622"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;Now we will assign each data point of the scatter plots to its closest K-point or centroids. We will compute it by applying some mathematics that we have studied to calculate the distance between two points. So, we will draw a median between both the centroids. See the visual plot below:&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fmrywa0u2emgj8zdm8io6.png" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fmrywa0u2emgj8zdm8io6.png" alt=" " width="800" height="612"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;From the above scatter visualization, it is clear that points left side of the line is near to the K1 or blue centroids, and points to the right of line are close to the orange centroid. Let's color them as blue and orange for clear visualization.&lt;/p&gt;

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2F4qv2ginpir6xa10z3f8q.png" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2F4qv2ginpir6xa10z3f8q.png" alt=" " width="698" height="551"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;We need to find the closest cluster, so we will repeat the process by choosing a new centroid. To choose the new centroid, we will compute the center of gravity of these centroids and will find new centroids as shown in the below image.&lt;/p&gt;

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fgvurhsc0jwcam7r2x2jx.png" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fgvurhsc0jwcam7r2x2jx.png" alt=" " width="672" height="575"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;Next, we will reassign each data point to the new centroid. For this, we will repeat the same process of finding a median line. The median will be look like the one below.&lt;/p&gt;

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fv0n37eeactunvs2mzoer.png" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fv0n37eeactunvs2mzoer.png" alt=" " width="679" height="535"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;From the above image, we can see one orange point is on the left side of the line, and two blue points are right to the line. So, these three points will be assigned to new centroids.&lt;/p&gt;

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fd0qx4gofhu5ttx2tvvan.png" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fd0qx4gofhu5ttx2tvvan.png" alt=" " width="611" height="528"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;As reassignment has taken place, so we will again go to step 4, which is finding new centroids or k-points.&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;We will repeat the process by finding the center of gravity of centroids, so the new centroids will be as shown in the below image:&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fs3ivt2kv6nfyq8bsaqgo.png" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fs3ivt2kv6nfyq8bsaqgo.png" alt=" " width="665" height="556"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;As we got the new centroids so again, we will draw the median line and reassign the data points. So, the image will be;&lt;/p&gt;

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fyjy8zonc2k2piuvhgsct.png" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fyjy8zonc2k2piuvhgsct.png" alt=" " width="658" height="594"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;We can see in the above image, there are no dissimilar data points on either side of the line, which means our model is formed. Consider the below image.&lt;/p&gt;

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2F33qju1kba1cqv1m98t7m.png" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2F33qju1kba1cqv1m98t7m.png" alt=" " width="654" height="453"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;As our model is ready, so we can now remove the assumed centroids, and the two final clusters will be as shown in the image below:&lt;/p&gt;

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2F0fc9cettfxr89gn47j6u.png" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2F0fc9cettfxr89gn47j6u.png" alt=" " width="613" height="479"&gt;&lt;/a&gt;&lt;/p&gt;

</description>
      <category>machinelearning</category>
      <category>datascience</category>
      <category>clustering</category>
      <category>python</category>
    </item>
    <item>
      <title>Association Rule Learning</title>
      <dc:creator>Victor Alando</dc:creator>
      <pubDate>Fri, 03 Jan 2025 13:29:46 +0000</pubDate>
      <link>https://dev.to/victoralando/association-rule-learning-27k2</link>
      <guid>https://dev.to/victoralando/association-rule-learning-27k2</guid>
      <description>&lt;h2&gt;
  
  
  Introduction
&lt;/h2&gt;

&lt;p&gt;The association rule learning is one of the very important concepts of machine learning and it is employed in Market Basket Analysis, Web usage mining, Continous production. Here market basket analysis is a technique used by the various big retailer to discover the associations between items.&lt;/p&gt;

&lt;p&gt;We can understand it by taking an example of a supermarket, as in a supermarket, all products that are purchased together are put together. For example, if a customer buys bread, he most likely can also buy butter, eggs, or milk so these products are stored within a shelf or mostly nearby.&lt;/p&gt;

&lt;h2&gt;
  
  
  Prerequisites
&lt;/h2&gt;

&lt;ul&gt;
&lt;li&gt;What is Association Rule Learning.&lt;/li&gt;
&lt;li&gt;How does Association Rule Work?&lt;/li&gt;
&lt;li&gt;Types of association Rule.&lt;/li&gt;
&lt;li&gt;Metrics of Association Rule.&lt;/li&gt;
&lt;li&gt;Types of Association Rule Algorithms.&lt;/li&gt;
&lt;li&gt;Types of Association Rule Learning.&lt;/li&gt;
&lt;/ul&gt;

&lt;blockquote&gt;
&lt;p&gt;People usually say that people who buy diapers must also buy juice ~ Anonymous&lt;/p&gt;
&lt;/blockquote&gt;

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2F7u8o31hru88jclp1tind.png" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2F7u8o31hru88jclp1tind.png" alt=" " width="800" height="418"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;h2&gt;
  
  
  What is Association Rule Learning?
&lt;/h2&gt;

&lt;p&gt;Association rule learning is a type of unsupervised learning technique that checks for the dependency of one data item on another data item and maps accordingly so that it can be more profitable. It tries to find some interesting relations or associations among the variables of dataset. It is based on different rules to discover the interesting relations between variables in the database.&lt;/p&gt;

&lt;p&gt;Association rule learning can be divided into three types of algorithms:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;&lt;strong&gt;Apriori.&lt;/strong&gt;&lt;/li&gt;
&lt;li&gt;&lt;strong&gt;Eclat.&lt;/strong&gt;&lt;/li&gt;
&lt;li&gt;&lt;strong&gt;F-P Growth Algorithm.&lt;/strong&gt;&lt;/li&gt;
&lt;/ul&gt;

&lt;h2&gt;
  
  
  How does Association Rule Learning Work?
&lt;/h2&gt;



&lt;br&gt;
Association rule learning works on the concept of If and Else Statement, such as if A then B.

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fkgd7p51629b7lxaycqwa.png" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fkgd7p51629b7lxaycqwa.png" alt=" " width="800" height="193"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;Here the if element is called antecedent, and then the statement is called as Consequent. These types of relationships where we can find out some association or relation between two items is known as Single Cardinality. It is all about creating rules, and if the number of items increases, then cardinality also increases accordingly. So, to measure the associations between thousands of data items, there are several metrics to follow:&lt;/p&gt;

&lt;h2&gt;
  
  
  Metrics of association Rule
&lt;/h2&gt;

&lt;ul&gt;
&lt;li&gt;&lt;strong&gt;Support&lt;/strong&gt;&lt;/li&gt;
&lt;li&gt;&lt;strong&gt;Confidence&lt;/strong&gt;&lt;/li&gt;
&lt;li&gt;&lt;strong&gt;Lift&lt;/strong&gt;&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;Let’s understand each of them:&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;Support&lt;/strong&gt;&lt;/p&gt;

&lt;p&gt;Support is the frequency of A or how frequently an item appears in the dataset. It is defined as the fraction of the transaction T that contains the item set X datasets, then for transactions &lt;strong&gt;T&lt;/strong&gt;, it can be written as&lt;/p&gt;

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fkd6n1d9d8mjdz5x15hse.png" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fkd6n1d9d8mjdz5x15hse.png" alt=" " width="800" height="102"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;Confidence&lt;/strong&gt;&lt;/p&gt;

&lt;p&gt;Confidence indicates how often the rules has been found to be true. Or how often the terms X and Y occur together in the dataset when the occurrence of X is already given. It is the ration of the transaction that contains X and Y to the number of records that contain X.&lt;/p&gt;

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fwga9nmpgi0nrvnx32aeg.png" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fwga9nmpgi0nrvnx32aeg.png" alt=" " width="800" height="95"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;Lift&lt;/strong&gt;&lt;/p&gt;

&lt;p&gt;It is the strength of any rule, which can be defined as:- It is the ration of the observed support measure and expected support if X and Y are independent of each other. It has three possible values:&lt;/p&gt;

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Ffaetxsf879wuqymd1pqp.png" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Ffaetxsf879wuqymd1pqp.png" alt=" " width="800" height="114"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;h2&gt;
  
  
  Types of Association Rule Algorithms
&lt;/h2&gt;

&lt;p&gt;Association rule learning can be divided into three algorithms:&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;Apriori Algorithm&lt;/strong&gt;&lt;/p&gt;

&lt;p&gt;This algorithm uses frequent datasets to generate association rules. It is designed to work on the datasets that contain transactions. This algorithm uses a breadth-first search and Hash Tree to calculate the itemset efficiently.&lt;/p&gt;

&lt;p&gt;It is mainly used for market basket analysis and helps to understand the products that can be bought together. it can also be used in the healthcare industry to find drug reactions for patients.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;Eclat Algorithm&lt;/strong&gt;&lt;/p&gt;

&lt;p&gt;Eclat algorithm stands for Equivalence Class Transformation. This algorithm uses a depth-first search technique to find frequent itemsets in a transaction database. It performs faster execution than Apriori Algorithm.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;F-P Growth Algorithm&lt;/strong&gt;&lt;/p&gt;

&lt;p&gt;The F-P growth algorithm stands for Frequent Pattern, and it is the improved version the Apriori Algorithm. It represents the database in the form of a tree structure that is known as a frequent pattern or tree. The purpose of this frequent tree is to extract the most frequent patterns.&lt;/p&gt;

&lt;h2&gt;
  
  
  Applications of Association Rule Learning
&lt;/h2&gt;

&lt;p&gt;It has various applications in machine learning and data mining. Below are some of the popular applications of association rule learning:&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;Market Basket Analysis&lt;/strong&gt;: It is one of the popular examples and applications association rule mining. This technique is commonly used by big retailers to determine the association between items.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;Medical Diagnosis&lt;/strong&gt;: With the help of association rules, patients can be cured easily, as it helps in identifying the probability of illness for a particular disease.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;Protein Sequence&lt;/strong&gt;: The association rules help in determining the synthesis of artificial proteins.&lt;/p&gt;

&lt;p&gt;It is also used for the &lt;strong&gt;Catalog Design and Loss-leader Analysis&lt;/strong&gt; and many ore other applications.&lt;/p&gt;

&lt;p&gt;For Python Implementation of Association Rule &lt;a href="https://dev.tourl"&gt;Click Here&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;************&lt;strong&gt;&lt;em&gt;Thanks for Reading Give me a Thumb&lt;/em&gt;&lt;/strong&gt;*******&lt;/p&gt;

</description>
      <category>datascience</category>
      <category>machinelearning</category>
      <category>datamining</category>
      <category>associationrule</category>
    </item>
    <item>
      <title>The Ultimate Guide to Data Analytics</title>
      <dc:creator>Victor Alando</dc:creator>
      <pubDate>Thu, 29 Aug 2024 10:18:36 +0000</pubDate>
      <link>https://dev.to/victoralando/the-ultimate-guide-to-data-analytics-14jj</link>
      <guid>https://dev.to/victoralando/the-ultimate-guide-to-data-analytics-14jj</guid>
      <description>&lt;h2&gt;
  
  
  Introduction
&lt;/h2&gt;

&lt;p&gt;Data is shorthand for “information,” and whether you are collecting, reviewing, and/or analyzing data this process has always been part of Head Start program operations. Students' enrollment into the program requires many pieces of information. The provision of health and dental services includes information from screening and any follow-up services that are provided. All areas of a Head Start program – content and management – involve the collection and use of substantial amounts of information.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;Prerequisites&lt;/strong&gt;&lt;/p&gt;

&lt;ol&gt;
&lt;li&gt;What is Data Analysis&lt;/li&gt;
&lt;li&gt;Who is a Data Analyst.&lt;/li&gt;
&lt;li&gt;Data Analysis Life Cycle&lt;/li&gt;
&lt;li&gt;Improving the model functionality.&lt;/li&gt;
&lt;/ol&gt;

&lt;h2&gt;
  
  
  What Is Data Analysis
&lt;/h2&gt;

&lt;p&gt;Data analysis is the processing of data to yield useful insights or knowledge.&lt;br&gt;
 • Data processing involves finding, loading, cleaning, manipulating, transforming, modeling, and visualizing the data.&lt;br&gt;
 • The knowledge may be used for scientific discovery, business decision-making, or a variety of other applications.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;A data analyst&lt;/strong&gt; is a person who uses tools and applications to transform raw data into a form that will be useful.&lt;/p&gt;

&lt;p&gt;From this perspective, we present a data analysis process that includes the following key components:&lt;br&gt;
• Purpose&lt;br&gt;
• Questions&lt;br&gt;
• Data Collection&lt;br&gt;
• Data Analysis Procedures and Methods&lt;br&gt;
• Interpretation/Identification of Findings&lt;br&gt;
• Writing, Reporting, and Dissemination; and&lt;br&gt;
• Evaluation&lt;br&gt;
We have also found, from our review of the literature, that there are many different ways&lt;br&gt;
of conceptualizing the data analysis process. We can make a basic distinction between a&lt;br&gt;
linear approach and a cyclical approach; in this Handbook we provide examples of both.&lt;/p&gt;

&lt;h2&gt;
  
  
  Data Analytics Life Circle
&lt;/h2&gt;

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2F6sv1dmpmy7k1xqo7vpxp.jpg" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2F6sv1dmpmy7k1xqo7vpxp.jpg" alt=" " width="800" height="400"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;Data is precious in today’s digital world environment. It goes through several life stages, including creation, testing, preprocessing, consumption, and reuse. &lt;/p&gt;

&lt;p&gt;These stages are mapped out in the Data Analytics Life Cycle for professionals working on data analytics initiatives. Each stage has its own significance and characteristics.&lt;/p&gt;

&lt;ol&gt;
&lt;li&gt;
&lt;strong&gt;Define the Problem&lt;/strong&gt;
In the data analysis process, the most challenging phase is to define the problem that needs to be solved. Deciphering the root cause of an issue requires a profound understanding of a business’ needs and aspirations, and involves a deep dive into metrics, KPIs, and other crucial indicators.&lt;/li&gt;
&lt;/ol&gt;

&lt;p&gt;This stage involves conducting initial analyses in order to gain valuable insights. It is crucial that this stage is done properly, as it lays a strong foundation for the entire data analysis process.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;2. Data Collection&lt;/strong&gt;&lt;br&gt;
After defining the problem, a data analyst then determines the most suitable data to address that question. The types of data they usually collect here include &lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;Quantitative data – like marketing figures – - or Qualitative data – like customer reviews.&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;Data types can be further categorized into 3 main groups: &lt;br&gt;
&lt;strong&gt;first-party data&lt;/strong&gt; (or data collected directly by an organization)&lt;br&gt;
&lt;strong&gt;second-party data&lt;/strong&gt; (or first-party data collected by one organization used by another), or &lt;strong&gt;third-party data&lt;/strong&gt; (or data aggregated from multiple sources by a third party).&lt;/p&gt;

&lt;p&gt;If the necessary data is incomplete or missing, a data analyst will be responsible in this step for devising a strategy for data collection. This includes different methods like surveys, social media monitoring, website analytics tracking, and online tracking in general.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;3. Data Cleaning&lt;/strong&gt;&lt;br&gt;
Freshly-collected data in its raw form is typically unorganized and messy. Before proceeding with the necessary analysis, that data must be cleaned up. In order to clean data, errors, duplicates, and outliers must be removed, along with any irrelevant data that does not contribute to the analysis being done.&lt;/p&gt;

&lt;p&gt;Additionally, the data must be restructured in a more meaningful manner depending on the type of analysis being done. Missing values must be filled in, too, in order to make the data more accurate. Data that is highly accurate can provide more valuable insights in the data analysis process.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;4. Data validation&lt;/strong&gt;&lt;br&gt;
After it is cleaned, the data must be validated. This process involves verifying whether the data meets the specific requirements of the analysis being performed.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;5. Perform Exploratory Data Analysis (EDA)&lt;/strong&gt;&lt;br&gt;
Exploratory Data Analysis(EDA) is the main step in the process of various data analysis. It helps data to visualize the patterns, characteristics, and relationships between variables. Python provides various libraries used for EDA such as &lt;strong&gt;NumPy&lt;/strong&gt;, &lt;strong&gt;Pandas&lt;/strong&gt;, &lt;strong&gt;Matplotlib&lt;/strong&gt;, &lt;strong&gt;Seaborn&lt;/strong&gt;, and &lt;strong&gt;Plotly&lt;/strong&gt;.&lt;/p&gt;

&lt;p&gt;6.&lt;strong&gt;Build the Model&lt;/strong&gt;&lt;/p&gt;

&lt;p&gt;Model building is an essential part of data analytics and is used to extract insights and knowledge from the data to make business decisions and strategies. In this phase of the project data science team needs to develop data sets for training, testing, and production purposes. To do this, dataset needs to be divided into two parts;&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;Training dataset &lt;/li&gt;
&lt;li&gt;Test dataset
&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;Note: Based on the dataset quality and quantity of the data one may choose to divide his dataset into three parts training and testing and validation data.&lt;/p&gt;

&lt;p&gt;To divide the dataset, Python  sklearn library which helps  in dividing the dataset into training and testing datasets is used to perform the train and test split. Here a data analyst will choose the ratio by which he/she want to divide the dataset by default it 8:2 meaning 80% and 20% for training and testing respectively.    &lt;/p&gt;

&lt;p&gt;&lt;strong&gt;7. Share the outcome&lt;/strong&gt;&lt;br&gt;
After conducting the analysis and extracting important insights, the final step lies in effectively communicating these findings to those who initiated the project in the first place.&lt;/p&gt;

&lt;p&gt;While it is essential to interpret the data accurately, it is equally important to be able to present those findings clearly and concisely. A data analyst is often working with marketing executives or stakeholders who are on time constraints and may not possess much technical expertise.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;8. Model Deployment&lt;/strong&gt;&lt;br&gt;
After performing a success EDA, the next final stage is to deploy the model into a real-world system or application to automatically generate predictions or perform specific tasks.&lt;/p&gt;

&lt;p&gt;*************Thank you for Reading########&lt;/p&gt;

</description>
      <category>python</category>
      <category>datascience</category>
      <category>analytics</category>
      <category>machinelearning</category>
    </item>
    <item>
      <title>Understanding Your Data: The Essentials of Exploratory Data Analysis</title>
      <dc:creator>Victor Alando</dc:creator>
      <pubDate>Tue, 20 Aug 2024 17:18:18 +0000</pubDate>
      <link>https://dev.to/victoralando/understanding-your-data-the-essentials-of-exploratory-data-analysis-4o0b</link>
      <guid>https://dev.to/victoralando/understanding-your-data-the-essentials-of-exploratory-data-analysis-4o0b</guid>
      <description>&lt;h2&gt;
  
  
  Introduction
&lt;/h2&gt;

&lt;p&gt;Exploratory data analysis (EDA) has been used by data scientists to analyze and investigate data sets and summarize their main characteristics, by applying  data visualization methods. It is a vital process in Data science since it helps in understanding the data you are dealing with and also making conclusions about it. EDA serves as a bridge between the process of data collection and the processes of building machine learning models.&lt;/p&gt;

&lt;h2&gt;
  
  
  Prerequisites
&lt;/h2&gt;

&lt;ol&gt;
&lt;li&gt;What is Exploratory Data Analysis&lt;/li&gt;
&lt;li&gt;Data Preprocessing and Feature Engineering in data science.
3  Types of Exploratory Data Analysis&lt;/li&gt;
&lt;li&gt;EDA Python Libraries.&lt;/li&gt;
&lt;/ol&gt;

&lt;p&gt;&lt;strong&gt;1. What Is Exploratory Data Analysis&lt;/strong&gt;&lt;/p&gt;

&lt;p&gt;Exploratory data analysis (EDA) is a critical initial step in the data science workflow. It involves using Python libraries to inspect, summarize, and visualize data to uncover trends, patterns, and relationships.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;2. Data Preprocessing and Feature Engineering in data science.&lt;/strong&gt;&lt;/p&gt;

&lt;p&gt;Data preprocessing and feature engineering are crucial steps in preparing datasets for effective model training.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;Data Preprocessing&lt;/strong&gt;&lt;/p&gt;

&lt;p&gt;Data pre-processing involves cleaning and preparing raw data to facilitate effective analysis and also prove the validity of a model. Data preprocessing include:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;Handling Missing Values.&lt;/li&gt;
&lt;li&gt;Detecting Outliers in a given dataset.&lt;/li&gt;
&lt;li&gt;Perform Encoding for categorical variables such as gender, country names needs to be converted into numerical format for machine learning algorithms. Encoding techniques like &lt;strong&gt;one-hot encoding&lt;/strong&gt; or &lt;strong&gt;label encoding&lt;/strong&gt; transform categorical ****variables into a format that algorithms can understand.&lt;/li&gt;
&lt;li&gt;Checking for Duplicate entries in given a dataset.&lt;/li&gt;
&lt;li&gt; Performing Train/Test Split to have dataset divided into two sets for further training the model and also for testing.&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;&lt;strong&gt;Feature Engineering&lt;/strong&gt;&lt;/p&gt;

&lt;p&gt;Feature engineering is a critical task that significantly influences the outcome of a model. It involves crafting new features based on existing data. This task is called &lt;strong&gt;Creation of Derived Features&lt;/strong&gt;. For example, extracting the day of the week from a date or creating interaction terms between existing features can provide valuable information.&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;&lt;p&gt;&lt;strong&gt;Dimensionality Reduction&lt;/strong&gt; is also another method of feature engineering. High-dimensional datasets may suffer from the curse of dimensionality, leading to increased computational complexity and potential overfitting.&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;Handling Outliers Outliers can distort model training, and addressing them is crucial. Techniques such as trimming, minorizing, or transforming features can mitigate the impact of outliers on model performance.&lt;/p&gt;&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;&lt;strong&gt;3. Types of Exploratory Analysis.&lt;/strong&gt;&lt;/p&gt;

&lt;p&gt;There are three main types of Exploratory Data Analysis (EDA):&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;Univariate (Non - Graphical).&lt;/li&gt;
&lt;li&gt;Univariate (Graphical).&lt;/li&gt;
&lt;li&gt;Bivariate.&lt;/li&gt;
&lt;li&gt;Multivariate (Non-Graphical).&lt;/li&gt;
&lt;li&gt;Multivariate (Graphical).&lt;/li&gt;
&lt;/ul&gt;

&lt;h4&gt;
  
  
  Univariate non-graphical
&lt;/h4&gt;

&lt;p&gt;This is simplest form of data analysis, where the data being analyzed consists of just one variable. Since it’s a single variable, it doesn’t deal with causes or relationships. The main purpose of univariate analysis is to describe the data and find patterns that exist within it.&lt;/p&gt;

&lt;h4&gt;
  
  
  Univariate graphical
&lt;/h4&gt;

&lt;p&gt;Non-graphical methods don’t provide a full picture of the data. Graphical methods are therefore required. Common types of univariate graphics include: Stem-and-leaf plots, which show all data values and the shape of the distribution. &lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;Histograms, a bar plot in which each bar represents the frequency (count) or proportion (count/total count) of cases for a range of values. &lt;/li&gt;
&lt;li&gt;Box plots, which graphically depict the five-number summary of minimum, first quartile, median, third quartile, and maximum.&lt;/li&gt;
&lt;/ul&gt;

&lt;h4&gt;
  
  
  Multivariate Non-graphical
&lt;/h4&gt;

&lt;p&gt;Multivariate data arises from more than one variable. Multivariate non-graphical EDA techniques generally show the relationship between two or more variables of the data through cross-tabulation or statistics.&lt;/p&gt;

&lt;h4&gt;
  
  
  Multivariate graphical
&lt;/h4&gt;

&lt;p&gt;Multivariate data uses graphics to display relationships between two or more sets of data. The most used graphic is a grouped bar plot or bar chart with each group representing one level of one of the variables and each bar within a group representing the levels of the other variable.&lt;/p&gt;

&lt;p&gt;Other common types of multivariate graphics include:&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;Scatter plots&lt;/strong&gt;: which is used to plot data points on a horizontal and a vertical axis to show how much one variable is affected by another.&lt;br&gt;
&lt;strong&gt;Multivariate chart&lt;/strong&gt;: which is a graphical representation of the relationships between factors and a response.&lt;br&gt;
&lt;strong&gt;Run chart&lt;/strong&gt;: which is a line graph of data plotted over time.&lt;br&gt;
&lt;strong&gt;Bubble chart&lt;/strong&gt;: which is a data visualization that displays multiple circles (bubbles) in a two-dimensional plot.&lt;br&gt;
&lt;strong&gt;Heat map:&lt;/strong&gt; which is a graphical representation of data where values are depicted by color.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;4. EDA Python Libraries&lt;/strong&gt;.&lt;/p&gt;

&lt;p&gt;Python  has top libraries for EDA which include;&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;Pandas for data manipulation.&lt;/li&gt;
&lt;li&gt;Matplotlib and Seaborn for visualisations. &lt;/li&gt;
&lt;li&gt;Plotly for interactive plots and &lt;/li&gt;
&lt;li&gt;Dask for scalable computing. &lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;Example:&lt;/p&gt;

&lt;p&gt;&lt;code&gt;import pandas as pd&lt;/code&gt;&lt;br&gt;
&lt;code&gt;import numpy as np&lt;/code&gt;&lt;br&gt;
&lt;code&gt;import matplotlib.pyplot as plt&lt;/code&gt;&lt;br&gt;
&lt;code&gt;import seaborn as sns&lt;/code&gt;&lt;br&gt;
&lt;code&gt;#to ignore warnings&lt;/code&gt;&lt;br&gt;
&lt;code&gt;import warnings&lt;/code&gt;&lt;br&gt;
&lt;code&gt;warnings.filterwarnings('ignore')&lt;/code&gt;&lt;/p&gt;

&lt;p&gt;These libraries enhance data analysis by offering powerful tools for summarizing, visualizing, and managing large datasets effectively.&lt;/p&gt;

</description>
      <category>machinelearning</category>
      <category>datascience</category>
      <category>python</category>
    </item>
    <item>
      <title>Feature Engineering</title>
      <dc:creator>Victor Alando</dc:creator>
      <pubDate>Mon, 19 Aug 2024 10:11:50 +0000</pubDate>
      <link>https://dev.to/victoralando/feature-engineering-304p</link>
      <guid>https://dev.to/victoralando/feature-engineering-304p</guid>
      <description></description>
    </item>
    <item>
      <title>Build an LSTM Stock Market model for Time Series Prediction:</title>
      <dc:creator>Victor Alando</dc:creator>
      <pubDate>Mon, 29 Apr 2024 10:13:44 +0000</pubDate>
      <link>https://dev.to/victoralando/build-an-lstm-model-for-time-series-prediction-using-python-with-tensorflowkeras-2h7c</link>
      <guid>https://dev.to/victoralando/build-an-lstm-model-for-time-series-prediction-using-python-with-tensorflowkeras-2h7c</guid>
      <description>&lt;h2&gt;
  
  
  Introduction
&lt;/h2&gt;

&lt;p&gt;These import statements lay the groundwork for conducting time series analysis and building a LSTM neural network model for stock price prediction. Each imported library serves a specific purpose in the data fetching, preprocessing, modeling, and visualization stages of the analysis.&lt;/p&gt;

&lt;h2&gt;
  
  
  Prerequisites
&lt;/h2&gt;

&lt;ul&gt;
&lt;li&gt;Data Retrieval&lt;/li&gt;
&lt;li&gt;Data Preprocessing&lt;/li&gt;
&lt;li&gt;Modeling&lt;/li&gt;
&lt;li&gt;Visualization&lt;/li&gt;
&lt;li&gt;Predictions&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;In this tutorial, we are going to learn about how to build LSTM models for time series predictions. LSTM stands for Long-Short-Term Memory&lt;/p&gt;

&lt;p&gt;These are models built on recurrent neural network (RNN) that are particularly effective for sequence prediction problems, such as time series forecasting.&lt;/p&gt;

&lt;p&gt;You need good machine learning models that can look at the history of a sequence of data and correctly predict what the future elements of the sequence are going to be.&lt;/p&gt;

&lt;p&gt;We are going  first to load the data from Alpha Vantage. The stock Market is for American Airlines stock market prices to make your  predictions, we are going to set the ticker to "AAL". &lt;/p&gt;

&lt;p&gt;Additionally, you also define a url_string, which will return a JSON file with all the stock market data for American Airlines within the last 20 years, and a file_to_save, which will be the file to which you save the data. You'll use the ticker variable that you defined beforehand to help name this file.&lt;/p&gt;

</description>
      <category>deeplearning</category>
      <category>python</category>
      <category>tensorflow</category>
      <category>keras</category>
    </item>
    <item>
      <title>K-Nearest Neighbor(K-NN) Algorithms for Machine Learning</title>
      <dc:creator>Victor Alando</dc:creator>
      <pubDate>Wed, 10 Apr 2024 10:00:31 +0000</pubDate>
      <link>https://dev.to/victoralando/k-nearest-neighborknn-algorithms-for-machine-learning-2n1b</link>
      <guid>https://dev.to/victoralando/k-nearest-neighborknn-algorithms-for-machine-learning-2n1b</guid>
      <description>&lt;p&gt;&lt;strong&gt;Introduction&lt;/strong&gt;&lt;/p&gt;

&lt;p&gt;In this tutorial, you are going to learn about how K-Nearest Neighbors (K-NN) as applied in Machine Learning models and also in Classification.&lt;/p&gt;

&lt;h2&gt;
  
  
  What is K-Nearest Neighbor (K-NN)
&lt;/h2&gt;

&lt;ul&gt;
&lt;li&gt;&lt;p&gt;K-Nearest Neighbor is one of the simplest Machine Learning algorithm based on supervised learning techniques.&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;K-Nearest Neighbor assumes the similarity between the new case and data available cases and put the new case into the category that is most similar to the available categories.&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;K-NN algorithm stores all the available data and classifies a new data point based on the similarity. This means that when new data point appears then it can be easily classified into a well suite category by using K-NN algorithm.&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;K-NN algorithm can be used for Regression as well as for classification problems.&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;K-NN is non-parametric algorithm because it does not learn from the training set immediately instead it stores the dataset and at the time of classification, it performs an action on the dataset.&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;K-NN algorithm at the training phase just stores the dataset and when it gets new data, then it classifies that data into category that is much similar to the new data.&lt;/p&gt;&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;&lt;strong&gt;Example&lt;/strong&gt; &lt;br&gt;
Suppose we have an image of a creature that looks similar to a Cheater and a Leopard, but we want to know either it is a Cheater or a Leopard. So for this identification, we can use the KNN algorithm, as it works on similarity measure.&lt;/p&gt;

&lt;p&gt;Our KNN Model will find the similar features of the new dataset to the Cheaters and Leopards images and based on the similar features it will put it either Cheater or a Leopard category.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;Why do we need a K-NN Algorithm?&lt;/strong&gt;&lt;/p&gt;

</description>
      <category>machinelearning</category>
      <category>algorithms</category>
      <category>datascience</category>
    </item>
    <item>
      <title>Understanding Data Warehousing</title>
      <dc:creator>Victor Alando</dc:creator>
      <pubDate>Tue, 09 Apr 2024 08:21:16 +0000</pubDate>
      <link>https://dev.to/victoralando/data-warehousing-3l44</link>
      <guid>https://dev.to/victoralando/data-warehousing-3l44</guid>
      <description>&lt;h2&gt;
  
  
  Introduction
&lt;/h2&gt;

&lt;p&gt;Data warehouse works like a relational database designed for data analytical needs. It functions on the basis of OLAP (Online Analytical Processing). It is a central location where consolidated data from multiple locations (databases) are stored.&lt;/p&gt;

&lt;h2&gt;
  
  
  Prerequisites
&lt;/h2&gt;

&lt;p&gt;In this tutorial, we are going to learn the following concepts in Data Warehousing.&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;What is Data Warehousing.&lt;/li&gt;
&lt;li&gt;Data Warehouse Process.&lt;/li&gt;
&lt;li&gt;Data Warehousing Architecture.&lt;/li&gt;
&lt;li&gt;Data Warehouse Characteristics.&lt;/li&gt;
&lt;li&gt;Modern Data Warehousing Examples.&lt;/li&gt;
&lt;/ul&gt;

&lt;h2&gt;
  
  
  1. What is Data Warehousing
&lt;/h2&gt;

&lt;p&gt;Data warehousing is the act of organizing &amp;amp; storing data in a way so as to make its retrieval efficient and insightful. It is also called as the process of transforming data into information for future business intelligence.&lt;/p&gt;

&lt;h2&gt;
  
  
  2. Data Warehouse Process.
&lt;/h2&gt;

&lt;p&gt;The data warehousing process refers to the sequence of steps involved in collecting, preparing, storing, and delivering data for analytics and business intelligence.&lt;/p&gt;

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2F7tq980pfd5rs4dfgluga.png" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2F7tq980pfd5rs4dfgluga.png" alt=" " width="800" height="303"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;The process of data warehousing can involve the following tools:&lt;br&gt;
i) Airbyte&lt;br&gt;
ii) Fivetran&lt;br&gt;
iii) Talend&lt;br&gt;
iv) Informatica&lt;br&gt;
v) custom ETL scripts (Python/Spark)&lt;/p&gt;

&lt;h2&gt;
  
  
  3. Architecture
&lt;/h2&gt;

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fyhupmu37hbvx2w2pne1q.png" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fyhupmu37hbvx2w2pne1q.png" alt=" " width="682" height="287"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;h4&gt;
  
  
  Data Sources
&lt;/h4&gt;

&lt;p&gt;In data warehousing architecture, data sources are key to the preparation of data warehouse. The goal of data extraction is to collect data from multiple sources like:&lt;/p&gt;

&lt;p&gt;i) Transactional databases (e.g., MySQL, PostgreSQL)&lt;br&gt;
ii) ERP/CRM systems (e.g., SAP, Salesforce)&lt;br&gt;
iii) Flat files (CSV, Excel)&lt;br&gt;
iv) APIs or web services&lt;br&gt;
v) Logs and IoT devices&lt;/p&gt;

&lt;h4&gt;
  
  
  ETL / ELT Process
&lt;/h4&gt;

&lt;p&gt;&lt;strong&gt;ETL = Extract → Transform → Load.&lt;/strong&gt;&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;Extract data from sources&lt;/li&gt;
&lt;li&gt;Transform (clean, merge, format, validate)&lt;/li&gt;
&lt;li&gt;Load into the warehouse.&lt;/li&gt;
&lt;/ul&gt;

&lt;h4&gt;
  
  
  Storage Layers
&lt;/h4&gt;

&lt;ul&gt;
&lt;li&gt;
&lt;strong&gt;Staging area&lt;/strong&gt;: temporary holding zone before processing&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;Data warehouse&lt;/strong&gt;: structured, cleaned, integrated data&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;Data marts&lt;/strong&gt;: subsets of warehouse data for specific teams (e.g. sales, finance)&lt;/li&gt;
&lt;/ul&gt;

&lt;h4&gt;
  
  
  BI Tools and Analytics
&lt;/h4&gt;

&lt;ul&gt;
&lt;li&gt;Power BI&lt;/li&gt;
&lt;li&gt;Tableau&lt;/li&gt;
&lt;li&gt;Looker&lt;/li&gt;
&lt;li&gt;Superset&lt;/li&gt;
&lt;li&gt;Grafana&lt;/li&gt;
&lt;li&gt;Metabase&lt;/li&gt;
&lt;/ul&gt;

&lt;h2&gt;
  
  
  4. Data Warehouse Characteristics.
&lt;/h2&gt;

&lt;p&gt;For decision making process to happen,  data warehouse must be a subject-oriented, integrated, time variant and non-volatile collection of data in support of management’s goals and business improvement.&lt;/p&gt;

&lt;h3&gt;
  
  
  Subject - Oriented.
&lt;/h3&gt;

&lt;p&gt;A Data warehouse can be used to analyze a particular subject area target oriented business analysis for example: “&lt;strong&gt;Sales&lt;/strong&gt;” can be particular subject-oriented type of analysis.&lt;/p&gt;

</description>
      <category>python</category>
      <category>datascience</category>
      <category>cloud</category>
      <category>azure</category>
    </item>
    <item>
      <title>Data Science for Beginners: 2023 - 2024 Complete Road Map</title>
      <dc:creator>Victor Alando</dc:creator>
      <pubDate>Fri, 05 Jan 2024 16:36:15 +0000</pubDate>
      <link>https://dev.to/victoralando/data-science-for-beginners-2023-2024-complete-road-map-1ad8</link>
      <guid>https://dev.to/victoralando/data-science-for-beginners-2023-2024-complete-road-map-1ad8</guid>
      <description>&lt;p&gt;Data science is the study of data, much like marine biology is the study of sea-dwelling biological life forms. Data scientists construct questions around specific data sets and then use data analytics and advanced analytics to find patterns, create predictive models, and develop insights that guide decision-making within business logics.&lt;/p&gt;

&lt;p&gt;Roadmaps are strategic plans that determine a goal or the desired outcome and feature the significant steps or milestones required to reach it.&lt;/p&gt;

&lt;p&gt;A data science roadmap is a visual representation of a strategic plan designed to help aspiring those who aspire aspiring to learn and succeed in the field of data science.&lt;/p&gt;

&lt;h2&gt;
  
  
  Key Tools used in Data Science
&lt;/h2&gt;

&lt;p&gt;we’ll take a look at  key data science tools that will help on your data science roadmap journey successful.&lt;/p&gt;

&lt;ol&gt;
&lt;li&gt; &lt;strong&gt;Programming Languages&lt;/strong&gt; - Here we have different programming languages that you need to master and know how to use them. Examples are;&lt;/li&gt;
&lt;/ol&gt;

&lt;p&gt;&lt;strong&gt;R-Language&lt;/strong&gt; - is similar to python and a famous programming language for working with data. it's a powerful language for performing data wrangling with &lt;strong&gt;dplyr&lt;/strong&gt;  and  &lt;strong&gt;ggplot2&lt;/strong&gt; to create any kind of chart you might need.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;Python&lt;/strong&gt; - Python is one of the greatest options available to you. In python, you can take advantage of the following libraries under a package known as Jupyter Notebook; &lt;br&gt;
a) Pandas&lt;br&gt;
b) Matplotlib&lt;br&gt;
c) Scikit-learn&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;SQL&lt;/strong&gt; - Stands for Structured Query Language). SQL allows the user to insert, update, delete, and select data from databases and to create new tables. The most common way to interact with these databases — called relational databases--is through Structured Query Language, or simply SQL. &lt;/p&gt;

&lt;p&gt;2.&lt;strong&gt;Machine Learning Libraries&lt;/strong&gt;&lt;br&gt;
  In ML  there are libraries you need to get familiar with like &lt;strong&gt;TensorFlow&lt;/strong&gt;, &lt;strong&gt;Scikit learn&lt;/strong&gt;, &lt;strong&gt;Pandas&lt;/strong&gt;, &lt;strong&gt;Matplotlib&lt;/strong&gt;, &lt;strong&gt;Numpy&lt;/strong&gt; and &lt;strong&gt;NLTK&lt;/strong&gt;.&lt;/p&gt;

&lt;p&gt;3.&lt;strong&gt;Data Visualization Tools&lt;/strong&gt;&lt;br&gt;
Data visualization tools are software applications that render information in a visual format such as a graph, chart, or heat map for data analysis purposes. Such tools make it easier to understand and work with massive amounts of data. Examples are &lt;strong&gt;PowerBI&lt;/strong&gt;, &lt;strong&gt;Tableau&lt;/strong&gt; and &lt;strong&gt;Matplotlib&lt;/strong&gt;.&lt;br&gt;
4.&lt;strong&gt;Data Storage Software&lt;/strong&gt;&lt;br&gt;
Learn about data storage software like &lt;strong&gt;SQL&lt;/strong&gt;, &lt;strong&gt;MySQL&lt;/strong&gt;, &lt;strong&gt;PostgreSQL&lt;/strong&gt; and &lt;strong&gt;MongoDB&lt;/strong&gt;&lt;br&gt;
5.&lt;strong&gt;Cloud Computing Platforms&lt;/strong&gt;&lt;br&gt;
These includes &lt;em&gt;&lt;strong&gt;AWS - Amazon Cloud Services&lt;/strong&gt;&lt;/em&gt;, &lt;em&gt;&lt;strong&gt;Microsoft Azure&lt;/strong&gt;&lt;/em&gt; and &lt;em&gt;&lt;strong&gt;Google cloud services&lt;/strong&gt;&lt;/em&gt;. by learning these, you will be able to interact with cloud storage services with your local stored data.&lt;/p&gt;

&lt;h2&gt;
  
  
  &lt;strong&gt;Learn about programming software Engineering&lt;/strong&gt;
&lt;/h2&gt;

&lt;p&gt;When you begin your data science roadmap, you must have a solid foundation. The data science field requires skills and experience in either software engineering or programming. You should learn a minimum of one programming language, such as &lt;strong&gt;&lt;em&gt;Python&lt;/em&gt;&lt;/strong&gt;, &lt;strong&gt;&lt;em&gt;SQL&lt;/em&gt;&lt;/strong&gt;, &lt;strong&gt;&lt;em&gt;Scala&lt;/em&gt;&lt;/strong&gt;, &lt;em&gt;&lt;strong&gt;Java&lt;/strong&gt;&lt;/em&gt;, or &lt;strong&gt;_ R._&lt;/strong&gt;&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;Example Programming Topics to learn&lt;/strong&gt;&lt;br&gt;
 Data scientists should learn about common data structures (e.g., dictionaries, data types, lists, sets, tuples), searching and sorting algorithms, logic, control flow, writing functions, object-oriented programming, and how to work with external libraries.&lt;/p&gt;

&lt;p&gt;Additionally, aspiring data scientists should be familiar with using Git and GitHub-related elements such as terminals and version control.&lt;/p&gt;

&lt;p&gt;Finally, data scientists should enjoy a familiarity with SQL scripting.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;Learn Git and Github&lt;/strong&gt;&lt;br&gt;
Git and Github allows you as a data scientist to push your ready-made projects. This will make you share with the outside world and learn more about git concepts and rules of   writing Git files and also sharing with others.&lt;/p&gt;

&lt;blockquote&gt;
&lt;blockquote&gt;
&lt;blockquote&gt;
&lt;blockquote&gt;
&lt;blockquote&gt;
&lt;blockquote&gt;
&lt;blockquote&gt;
&lt;blockquote&gt;
&lt;blockquote&gt;
&lt;blockquote&gt;
&lt;blockquote&gt;
&lt;blockquote&gt;
&lt;blockquote&gt;
&lt;blockquote&gt;
&lt;blockquote&gt;
&lt;blockquote&gt;
&lt;blockquote&gt;
&lt;/blockquote&gt;


&lt;/blockquote&gt;
&lt;br&gt;
&lt;/blockquote&gt;
&lt;br&gt;
&lt;/blockquote&gt;
&lt;br&gt;
&lt;/blockquote&gt;
&lt;br&gt;
&lt;/blockquote&gt;
&lt;br&gt;
&lt;/blockquote&gt;
&lt;br&gt;
&lt;/blockquote&gt;
&lt;br&gt;
&lt;/blockquote&gt;
&lt;br&gt;
&lt;/blockquote&gt;
&lt;br&gt;
&lt;/blockquote&gt;
&lt;br&gt;
&lt;/blockquote&gt;
&lt;br&gt;
&lt;/blockquote&gt;
&lt;br&gt;
&lt;/blockquote&gt;
&lt;br&gt;
&lt;/blockquote&gt;
&lt;br&gt;
&lt;/blockquote&gt;
&lt;br&gt;
&lt;/blockquote&gt;

</description>
      <category>python</category>
      <category>machinelearning</category>
      <category>visualization</category>
      <category>beginners</category>
    </item>
    <item>
      <title>Database Normalization In DBMS</title>
      <dc:creator>Victor Alando</dc:creator>
      <pubDate>Tue, 14 Nov 2023 10:35:16 +0000</pubDate>
      <link>https://dev.to/victoralando/database-normalization-24k0</link>
      <guid>https://dev.to/victoralando/database-normalization-24k0</guid>
      <description>&lt;h2&gt;
  
  
  What is Normalization?
&lt;/h2&gt;

&lt;p&gt;Normalization is the process of organizing the data and the attributes of a database.&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;&lt;p&gt;It is performed to reduce the data redundancy in a database and ensure that data is stored logically.&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;Normalization is systematic approach of decomposing table to eliminate data redundancy and undesirable characteristics like insertion, update and delete.&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;Normalization is multi-step process that puts data in tabular form and remove duplicate data from relation tables.&lt;/p&gt;&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;&lt;strong&gt;Example Employees Table&lt;/strong&gt;&lt;/p&gt;

&lt;div class="table-wrapper-paragraph"&gt;&lt;table&gt;
&lt;thead&gt;
&lt;tr&gt;
&lt;th&gt;Id&lt;/th&gt;
&lt;th&gt;Name&lt;/th&gt;
&lt;th&gt;Address&lt;/th&gt;
&lt;th&gt;Profession&lt;/th&gt;
&lt;/tr&gt;
&lt;/thead&gt;
&lt;tbody&gt;
&lt;tr&gt;
&lt;td&gt;101&lt;/td&gt;
&lt;td&gt;Mary&lt;/td&gt;
&lt;td&gt;1245&lt;/td&gt;
&lt;td&gt;Developer&lt;/td&gt;
&lt;/tr&gt;
&lt;tr&gt;
&lt;td&gt;102&lt;/td&gt;
&lt;td&gt;David&lt;/td&gt;
&lt;td&gt;5234&lt;/td&gt;
&lt;td&gt;Accountant&lt;/td&gt;
&lt;/tr&gt;
&lt;tr&gt;
&lt;td&gt;103&lt;/td&gt;
&lt;td&gt;Juliet&lt;/td&gt;
&lt;td&gt;1444&lt;/td&gt;
&lt;td&gt;Salesperson&lt;/td&gt;
&lt;/tr&gt;
&lt;tr&gt;
&lt;td&gt;104&lt;/td&gt;
&lt;td&gt;Elizabeth&lt;/td&gt;
&lt;td&gt;8745&lt;/td&gt;
&lt;td&gt;Manager&lt;/td&gt;
&lt;/tr&gt;
&lt;tr&gt;
&lt;td&gt;105&lt;/td&gt;
&lt;td&gt;Haskell&lt;/td&gt;
&lt;td&gt;3251&lt;/td&gt;
&lt;td&gt;Operation&lt;/td&gt;
&lt;/tr&gt;
&lt;/tbody&gt;
&lt;/table&gt;&lt;/div&gt;

&lt;p&gt;In this table, we have data of office employees.&lt;/p&gt;




&lt;p&gt;&lt;strong&gt;1. Insertion Anomaly&lt;/strong&gt;&lt;br&gt;
An insertion anomaly occurs in the relation database when some attributes or data items are inserted into database without existence of other attributes.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;2. Updation Anomaly&lt;/strong&gt;&lt;br&gt;
Updation Anomaly occurs when the same data item is repeated with the same values are not linked to each other.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;3. Deletion Anomalies&lt;/strong&gt;&lt;/p&gt;

&lt;p&gt;Deletion Anomalies occurs when deleting one part of the data deletes the other necessary information from the database.&lt;/p&gt;




&lt;h2&gt;
  
  
  Types of Normalization
&lt;/h2&gt;

&lt;ul&gt;
&lt;li&gt;&lt;p&gt;1NF&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;2NF&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;3NF&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;BCNF&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;4NF&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;5NF&lt;/p&gt;&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;&lt;strong&gt;Diagram:&lt;/strong&gt;&lt;/p&gt;

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fm33j9bawmz0elwer4wa0.png" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fm33j9bawmz0elwer4wa0.png" alt=" " width="610" height="427"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;1. 1NF (First Normal Form)&lt;/strong&gt;&lt;br&gt;
In 1NF relation, each table cell should contain a single value. Each record looks like unique.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;Example&lt;/strong&gt;&lt;/p&gt;

&lt;div class="table-wrapper-paragraph"&gt;&lt;table&gt;
&lt;thead&gt;
&lt;tr&gt;
&lt;th&gt;CouserId&lt;/th&gt;
&lt;th&gt;Course Name&lt;/th&gt;
&lt;th&gt;Framework&lt;/th&gt;
&lt;/tr&gt;
&lt;/thead&gt;
&lt;tbody&gt;
&lt;tr&gt;
&lt;td&gt;JAV101&lt;/td&gt;
&lt;td&gt;Java&lt;/td&gt;
&lt;td&gt;NetBeans&lt;/td&gt;
&lt;/tr&gt;
&lt;tr&gt;
&lt;td&gt;SQL102&lt;/td&gt;
&lt;td&gt;SQL&lt;/td&gt;
&lt;td&gt;MySQL, PostgreSQL&lt;/td&gt;
&lt;/tr&gt;
&lt;tr&gt;
&lt;td&gt;PY214&lt;/td&gt;
&lt;td&gt;Python&lt;/td&gt;
&lt;td&gt;Flask&lt;/td&gt;
&lt;/tr&gt;
&lt;/tbody&gt;
&lt;/table&gt;&lt;/div&gt;

&lt;p&gt;Here in the Framework row, we stored two frameworks of course Name MySQL, PostgreSQL so it is *&lt;em&gt;multi-valued attribute. *&lt;/em&gt; it is not 1NF relation. We need to convert it into 1NF.  &lt;/p&gt;

&lt;p&gt;&lt;strong&gt;Convert it into 1NF&lt;/strong&gt;&lt;/p&gt;

&lt;div class="table-wrapper-paragraph"&gt;&lt;table&gt;
&lt;thead&gt;
&lt;tr&gt;
&lt;th&gt;CourseId&lt;/th&gt;
&lt;th&gt;Course Name&lt;/th&gt;
&lt;th&gt;Framework&lt;/th&gt;
&lt;/tr&gt;
&lt;/thead&gt;
&lt;tbody&gt;
&lt;tr&gt;
&lt;td&gt;JAV101&lt;/td&gt;
&lt;td&gt;Java&lt;/td&gt;
&lt;td&gt;NetBeans&lt;/td&gt;
&lt;/tr&gt;
&lt;tr&gt;
&lt;td&gt;SQL102&lt;/td&gt;
&lt;td&gt;SQL&lt;/td&gt;
&lt;td&gt;MySQL&lt;/td&gt;
&lt;/tr&gt;
&lt;tr&gt;
&lt;td&gt;SQL102&lt;/td&gt;
&lt;td&gt;SQL&lt;/td&gt;
&lt;td&gt;PostgreSQL&lt;/td&gt;
&lt;/tr&gt;
&lt;tr&gt;
&lt;td&gt;PY214&lt;/td&gt;
&lt;td&gt;Python&lt;/td&gt;
&lt;td&gt;Flask&lt;/td&gt;
&lt;/tr&gt;
&lt;/tbody&gt;
&lt;/table&gt;&lt;/div&gt;

&lt;p&gt;It's a simple method to store Framework separately in 1NF. Now this is &lt;strong&gt;First Normal Form&lt;/strong&gt;. 1NF wants to store unique information in table without data repetition.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;2. 2NF (Second Normal Form)&lt;/strong&gt;&lt;/p&gt;

&lt;p&gt;In &lt;strong&gt;2NF&lt;/strong&gt;, relation must in &lt;strong&gt;1NF&lt;/strong&gt;. In the Second Normal Form, all non-key attributes are fully functionally dependent on the primary key.&lt;/p&gt;

&lt;div class="table-wrapper-paragraph"&gt;&lt;table&gt;
&lt;thead&gt;
&lt;tr&gt;
&lt;th&gt;StudentID&lt;/th&gt;
&lt;th&gt;Specialization&lt;/th&gt;
&lt;th&gt;Student Age&lt;/th&gt;
&lt;/tr&gt;
&lt;/thead&gt;
&lt;tbody&gt;
&lt;tr&gt;
&lt;td&gt;501&lt;/td&gt;
&lt;td&gt;Data Analyst&lt;/td&gt;
&lt;td&gt;22&lt;/td&gt;
&lt;/tr&gt;
&lt;tr&gt;
&lt;td&gt;501&lt;/td&gt;
&lt;td&gt;Data Engineer&lt;/td&gt;
&lt;td&gt;22&lt;/td&gt;
&lt;/tr&gt;
&lt;tr&gt;
&lt;td&gt;502&lt;/td&gt;
&lt;td&gt;Full Stack Developer&lt;/td&gt;
&lt;td&gt;24&lt;/td&gt;
&lt;/tr&gt;
&lt;tr&gt;
&lt;td&gt;503&lt;/td&gt;
&lt;td&gt;Web Developer&lt;/td&gt;
&lt;td&gt;23&lt;/td&gt;
&lt;/tr&gt;
&lt;/tbody&gt;
&lt;/table&gt;&lt;/div&gt;

</description>
      <category>database</category>
      <category>datascience</category>
      <category>python</category>
      <category>programming</category>
    </item>
    <item>
      <title>Modern Data Stacks</title>
      <dc:creator>Victor Alando</dc:creator>
      <pubDate>Thu, 09 Nov 2023 07:22:03 +0000</pubDate>
      <link>https://dev.to/victoralando/modern-data-stacks-1lge</link>
      <guid>https://dev.to/victoralando/modern-data-stacks-1lge</guid>
      <description>&lt;p&gt;Modern data stacks, often referred to as data technology stacks or data toolchains, are the combination of software and technologies used to collect, store, process, and analyze data in contemporary data-driven organizations. These stacks have evolved significantly in recent years, incorporating a variety of open-source and proprietary tools to meet the growing demands of data analytics and data-driven decision-making. Here's a high-level overview of components commonly found in modern data stacks:&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;1. Data Ingestion:&lt;/strong&gt;&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;&lt;p&gt;Apache Kafka: For real-time data streaming.&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;Apache Nifi, Flume, or Logstash: For data collection and ETL (Extract, Transform, Load) processes&lt;/p&gt;&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;&lt;strong&gt;2. Data Storage:&lt;/strong&gt;&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;&lt;p&gt;Cloud-based data warehouses like Amazon Redshift, Google BigQuery, or Snowflake.&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;Distributed file systems like Hadoop HDFS.&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;NoSQL databases like MongoDB, Cassandra, or Elasticsearch for unstructured or semi-structured data.&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;Traditional relational databases such as PostgreSQL or MySQL.&lt;/p&gt;&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;&lt;strong&gt;3.Data Processing and Transformation:&lt;/strong&gt;&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;&lt;p&gt;Apache Spark: For distributed data processing and ETL.&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;Apache Flink: For real-time stream processing.&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;Apache Beam: For unified batch and stream data processing. Data processing frameworks like Apache Airflow for workflow management.&lt;/p&gt;&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;&lt;strong&gt;4. Data Query and Analysis:&lt;/strong&gt;&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;&lt;p&gt;SQL-based query engines for data warehousing solutions.&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;Business intelligence tools like Tableau, Power BI, or Looker.&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;Jupyter notebooks with Python or R for data analysis.&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;Custom dashboards using frameworks like Superset or Redash.&lt;/p&gt;&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;&lt;strong&gt;5. Data Visualization:&lt;/strong&gt;&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;&lt;p&gt;Tools like Tableau, Power BI, or Qlik for interactive data visualization.&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;Libraries like D3.js, Plotly, or Matplotlib for custom visualizations.&lt;/p&gt;&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;&lt;strong&gt;6. Data Governance and Security:&lt;/strong&gt;&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;&lt;p&gt;Data catalog and metadata management tools.&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;Access control and encryption solutions.&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;Data lineage and auditing tools for compliance.&lt;/p&gt;&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;&lt;strong&gt;7. Machine Learning and AI:&lt;/strong&gt;&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;&lt;p&gt;Machine learning frameworks like TensorFlow and PyTorch.&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;ML platforms like MLflow for model tracking and management.&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;AutoML tools for automated model building and deployment.&lt;/p&gt;&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;&lt;strong&gt;8. Cloud Services:&lt;/strong&gt;&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;Leveraging cloud platforms like AWS, Azure, or Google Cloud for scalable and cost-effective data storage and processing.&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;&lt;strong&gt;9. DevOps and Infrastructure:&lt;/strong&gt;&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;&lt;p&gt;Containers and orchestration tools like Docker and Kubernetes.&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;Infrastructure as code (IaC) for managing and scaling data infrastructure.&lt;/p&gt;&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;&lt;strong&gt;10. Monitoring and Management:&lt;/strong&gt;&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;&lt;p&gt;Tools for logging, monitoring, and alerting, such as Prometheus, Grafana, or ELK stack.&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;Data pipeline orchestration and job scheduling using tools like Apache Oozie or Luigi.&lt;/p&gt;&lt;/li&gt;
&lt;/ul&gt;

&lt;blockquote&gt;
&lt;p&gt;The specific components and technologies in a data stack can vary based on the organization's needs, data volume, and budget. Modern data stacks are often designed to be flexible, scalable, and capable of handling both batch and real-time data processing, making them a crucial part of any data-driven enterprise.&lt;/p&gt;
&lt;/blockquote&gt;

</description>
    </item>
    <item>
      <title>Data Engineering for Beginners Step-by-Step</title>
      <dc:creator>Victor Alando</dc:creator>
      <pubDate>Thu, 09 Nov 2023 07:08:48 +0000</pubDate>
      <link>https://dev.to/victoralando/data-engineering-for-beginners-step-by-step-120p</link>
      <guid>https://dev.to/victoralando/data-engineering-for-beginners-step-by-step-120p</guid>
      <description>&lt;p&gt;In this article, We are going to look at the skills and qualifications you need to become a data engineer and provide you with some tips to help you land your first position in the industry. &lt;/p&gt;

&lt;h2&gt;
  
  
  Who is a data Engineer?
&lt;/h2&gt;

&lt;p&gt;A Data engineer is responsible for laying the foundations for storage, transformation, and management of data in an organization. They manage the design, creation, and maintenance of database architecture and data processing systems; this ensures that the subsequent work of data analysis, visualization, and machine learning models development can be carried out seamlessly, continuously, securely, and effectively. &lt;/p&gt;

&lt;h2&gt;
  
  
  Responsibilities of a Data Engineer
&lt;/h2&gt;

&lt;ul&gt;
&lt;li&gt;&lt;p&gt;They ensure that the large volume of data collected from different sources becomes accessible raw material for other data science specialists, such as data analysts and data scientists.&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;Serve as a data resource expert for the organization.&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;Build and execute data ETL solution pipelines for multiple clients in different industries.&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;Independently create data-driven solutions that are accurate and informative.&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;Interact with the data scientists team and assist them in providing suitable datasets for analysis.&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;Leverage various big data engineering tools and cloud service providing platforms to create data extractions and storage pipelines.&lt;/p&gt;&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;.&lt;/p&gt;

&lt;h2&gt;
  
  
  Exploring the basics of Data Engineering
&lt;/h2&gt;

&lt;p&gt;To become a data Engineer, below is the path of the leaning path.&lt;/p&gt;

&lt;ol&gt;
&lt;li&gt;Know programming languages like Python and Scala, R and Java.&lt;/li&gt;
&lt;li&gt;Learn the basics of Automation and scripting.&lt;/li&gt;
&lt;li&gt;High efficiency in advanced probability and statistics&lt;/li&gt;
&lt;li&gt;Ability to demonstrate expertise in database management systems.&lt;/li&gt;
&lt;li&gt;Experience with using cloud services providing platforms like AWS/GCP/Azure.&lt;/li&gt;
&lt;li&gt;Good knowledge of various machine learning and deep learning algorithms will be a bonus. &lt;/li&gt;
&lt;li&gt;Knowledge of popular big data tools like Apache Spark, Apache Hadoop, etc.&lt;/li&gt;
&lt;li&gt;Good communication skills as a data engineer directly works with the different teams.&lt;/li&gt;
&lt;/ol&gt;

</description>
      <category>python</category>
      <category>database</category>
    </item>
  </channel>
</rss>
