<?xml version="1.0" encoding="UTF-8"?>
<rss version="2.0" xmlns:atom="http://www.w3.org/2005/Atom" xmlns:dc="http://purl.org/dc/elements/1.1/">
  <channel>
    <title>DEV Community: Michael Obed </title>
    <description>The latest articles on DEV Community by Michael Obed  (@obedm16).</description>
    <link>https://dev.to/obedm16</link>
    <image>
      <url>https://media2.dev.to/dynamic/image/width=90,height=90,fit=cover,gravity=auto,format=auto/https:%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Fuser%2Fprofile_image%2F1126919%2Ffec3859d-8ce9-4041-874f-5ee4ae8c9cab.jpeg</url>
      <title>DEV Community: Michael Obed </title>
      <link>https://dev.to/obedm16</link>
    </image>
    <atom:link rel="self" type="application/rss+xml" href="https://dev.to/feed/obedm16"/>
    <language>en</language>
    <item>
      <title>Creating and Integrating Data Pipeline Using Amazon S3 and Snowflake Data Warehouse. (Using SnowSQL).</title>
      <dc:creator>Michael Obed </dc:creator>
      <pubDate>Thu, 11 Jul 2024 15:38:57 +0000</pubDate>
      <link>https://dev.to/obedm16/creating-and-integrating-data-pipeline-using-amazon-s3-and-snowflake-data-warehouse-using-snowsql-4ekb</link>
      <guid>https://dev.to/obedm16/creating-and-integrating-data-pipeline-using-amazon-s3-and-snowflake-data-warehouse-using-snowsql-4ekb</guid>
      <description>&lt;p&gt;Data pipeline is widely used in the field of Data Engineering and Analytics to fetch data from external sources for instance AWS Redshift, S3 (Simple Storage Service), GCP (Google Cloud Platform), Oracle, Azure and much more industry –used technologies.&lt;/p&gt;

&lt;p&gt;&lt;a href="https://media.dev.to/cdn-cgi/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fet05epkp5vyqkijv4pci.png" class="article-body-image-wrapper"&gt;&lt;img src="https://media.dev.to/cdn-cgi/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fet05epkp5vyqkijv4pci.png" alt="Image description" width="714" height="318"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;Data Pipeline – according to Snowflake is concerned with moving data from one place to a destination (such as a data warehouse or large storage service) while simultaneously optimizing and transforming the data. As a result, the data arrives in a state that can be analyzed and used to develop business insights.&lt;/p&gt;

&lt;p&gt;This article aims to cover how integrate AWS S3 with Snowflake which is convenient when dealing and working with large data sources.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;Requirements:&lt;/strong&gt;&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;Knowledge of SQL for working in Snowflake.&lt;/li&gt;
&lt;li&gt;Snowflake account and AWS&lt;/li&gt;
&lt;li&gt;Data file in preferably CSV (Comma Separated Value), it can be large up to 160 GB of size. Beyond that AWS requires the Cloud developer to use other tools – this is beyond this article’s scope.&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;&lt;strong&gt;Step 1&lt;/strong&gt;&lt;br&gt;
Setup an AWS account as a root user.&lt;/p&gt;

&lt;p&gt;&lt;a href="https://media.dev.to/cdn-cgi/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fmzoxwua3s62q596a1f63.png" class="article-body-image-wrapper"&gt;&lt;img src="https://media.dev.to/cdn-cgi/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fmzoxwua3s62q596a1f63.png" alt="Image description" width="629" height="331"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;Step 2&lt;/strong&gt;&lt;br&gt;
After a successful login into your account go to &lt;strong&gt;services&lt;/strong&gt; as shown. For new users locate &lt;strong&gt;storage&lt;/strong&gt; by scrolling downward then &lt;strong&gt;select S3.&lt;/strong&gt;&lt;/p&gt;

&lt;p&gt;&lt;a href="https://media.dev.to/cdn-cgi/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fl2pa5ixoebwj0oo6byox.png" class="article-body-image-wrapper"&gt;&lt;img src="https://media.dev.to/cdn-cgi/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fl2pa5ixoebwj0oo6byox.png" alt="Image description" width="577" height="271"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;&lt;a href="https://media.dev.to/cdn-cgi/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fmtmecertyko0zclny36a.png" class="article-body-image-wrapper"&gt;&lt;img src="https://media.dev.to/cdn-cgi/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fmtmecertyko0zclny36a.png" alt="Image description" width="623" height="335"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;Step 3&lt;/strong&gt;&lt;br&gt;
Upon selecting &lt;strong&gt;S3&lt;/strong&gt; we expect the display as follows – design might change later.&lt;/p&gt;

&lt;p&gt;&lt;a href="https://media.dev.to/cdn-cgi/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fnoibxrcuy1wg9flwf1qi.png" class="article-body-image-wrapper"&gt;&lt;img src="https://media.dev.to/cdn-cgi/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fnoibxrcuy1wg9flwf1qi.png" alt="Image description" width="624" height="335"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;Click &lt;strong&gt;Create bucket&lt;/strong&gt; then give it a name – in our case &lt;strong&gt;mybootcampbucket&lt;/strong&gt;:&lt;/p&gt;

&lt;p&gt;&lt;a href="https://media.dev.to/cdn-cgi/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2F2zhqvuyvhqug1uwon4px.png" class="article-body-image-wrapper"&gt;&lt;img src="https://media.dev.to/cdn-cgi/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2F2zhqvuyvhqug1uwon4px.png" alt="Image description" width="624" height="337"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;&lt;a href="https://media.dev.to/cdn-cgi/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2F7979ssb0ngwsayv8wkx1.png" class="article-body-image-wrapper"&gt;&lt;img src="https://media.dev.to/cdn-cgi/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2F7979ssb0ngwsayv8wkx1.png" alt="Image description" width="624" height="333"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;Click &lt;strong&gt;Create bucket&lt;/strong&gt; to finish the process.&lt;/p&gt;

&lt;p&gt;&lt;a href="https://media.dev.to/cdn-cgi/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2F3m9hhxiijywokv3zyiu1.png" class="article-body-image-wrapper"&gt;&lt;img src="https://media.dev.to/cdn-cgi/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2F3m9hhxiijywokv3zyiu1.png" alt="Image description" width="624" height="333"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;To upload the file select the newly created bucket (highlighted).&lt;/p&gt;

&lt;p&gt;&lt;a href="https://media.dev.to/cdn-cgi/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2F2re1gshm3fpblcyy10k2.png" class="article-body-image-wrapper"&gt;&lt;img src="https://media.dev.to/cdn-cgi/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2F2re1gshm3fpblcyy10k2.png" alt="Image description" width="624" height="335"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;Upload the file now as depicted below.&lt;/p&gt;

&lt;p&gt;&lt;a href="https://media.dev.to/cdn-cgi/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fyzxy4ixk1ex68ern3lrb.png" class="article-body-image-wrapper"&gt;&lt;img src="https://media.dev.to/cdn-cgi/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fyzxy4ixk1ex68ern3lrb.png" alt="Image description" width="624" height="337"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;&lt;a href="https://media.dev.to/cdn-cgi/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2F6goja0sy9d3xckkim1s7.png" class="article-body-image-wrapper"&gt;&lt;img src="https://media.dev.to/cdn-cgi/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2F6goja0sy9d3xckkim1s7.png" alt="Image description" width="624" height="336"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;Step 4&lt;/strong&gt;&lt;br&gt;
After uploading the file/dataset. Policy need to be set up. The policy facilitates permission for sharing to external integration when associated with identity and compliance to respective AWS account.&lt;br&gt;
To setup policy click &lt;strong&gt;services&lt;/strong&gt; then head to &lt;strong&gt;Security&lt;/strong&gt;, &lt;strong&gt;Identity&lt;/strong&gt; and &lt;strong&gt;Compliance&lt;/strong&gt;. See below:&lt;/p&gt;

&lt;p&gt;&lt;a href="https://media.dev.to/cdn-cgi/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fw9caa7t13tqvbxn4jpvk.png" class="article-body-image-wrapper"&gt;&lt;img src="https://media.dev.to/cdn-cgi/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fw9caa7t13tqvbxn4jpvk.png" alt="Image description" width="624" height="333"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;Select &lt;strong&gt;IAM&lt;/strong&gt; (Identity Access Management) then open &lt;strong&gt;Policies&lt;/strong&gt;.&lt;/p&gt;

&lt;p&gt;&lt;a href="https://media.dev.to/cdn-cgi/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fmp53mu6cbnd6undn83t0.png" class="article-body-image-wrapper"&gt;&lt;img src="https://media.dev.to/cdn-cgi/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fmp53mu6cbnd6undn83t0.png" alt="Image description" width="624" height="334"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;&lt;a href="https://media.dev.to/cdn-cgi/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fx6446gcept50puuew41m.png" class="article-body-image-wrapper"&gt;&lt;img src="https://media.dev.to/cdn-cgi/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fx6446gcept50puuew41m.png" alt="Image description" width="624" height="335"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;After clicking ‘&lt;strong&gt;Create policy&lt;/strong&gt;’, allocate the name to the policy to be created. Our case is &lt;strong&gt;Bootcamp2023&lt;/strong&gt;.&lt;/p&gt;

&lt;p&gt;&lt;a href="https://media.dev.to/cdn-cgi/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2F0ci0zcdct6k9w0mcth3f.png" class="article-body-image-wrapper"&gt;&lt;img src="https://media.dev.to/cdn-cgi/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2F0ci0zcdct6k9w0mcth3f.png" alt="Image description" width="624" height="352"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;Under policies, click &lt;strong&gt;JSON Policy&lt;/strong&gt;, this is because Policies are written in JavaScript Object Notation (JSON) which is a data representation Key – Value pair syntax.&lt;/p&gt;

&lt;p&gt;&lt;a href="https://media.dev.to/cdn-cgi/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fug53mog0ni9o3ir6g06n.png" class="article-body-image-wrapper"&gt;&lt;img src="https://media.dev.to/cdn-cgi/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fug53mog0ni9o3ir6g06n.png" alt="Image description" width="624" height="335"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;&lt;a href="https://media.dev.to/cdn-cgi/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fx2fa79bzaw3bjyv0mk49.png" class="article-body-image-wrapper"&gt;&lt;img src="https://media.dev.to/cdn-cgi/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fx2fa79bzaw3bjyv0mk49.png" alt="Image description" width="624" height="333"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;&lt;a href="https://media.dev.to/cdn-cgi/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2F3wtr93ypdgdvkuh5kavw.png" class="article-body-image-wrapper"&gt;&lt;img src="https://media.dev.to/cdn-cgi/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2F3wtr93ypdgdvkuh5kavw.png" alt="Image description" width="624" height="335"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;After setting up the policy, next we create the &lt;strong&gt;Role&lt;/strong&gt;.&lt;/p&gt;

&lt;p&gt;&lt;a href="https://media.dev.to/cdn-cgi/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fyzf7lenk56ivm24q5dw6.png" class="article-body-image-wrapper"&gt;&lt;img src="https://media.dev.to/cdn-cgi/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fyzf7lenk56ivm24q5dw6.png" alt="Image description" width="624" height="321"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;Give your role a name, in this scenario – our role is named &lt;strong&gt;Bootcamp_2023&lt;/strong&gt;. &lt;/p&gt;

&lt;p&gt;&lt;em&gt;&lt;strong&gt;Setup the permission&lt;/strong&gt;&lt;/em&gt;.&lt;/p&gt;

&lt;p&gt;&lt;a href="https://media.dev.to/cdn-cgi/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fk22161pqjn62jp05igsx.png" class="article-body-image-wrapper"&gt;&lt;img src="https://media.dev.to/cdn-cgi/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fk22161pqjn62jp05igsx.png" alt="Image description" width="624" height="334"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;&lt;a href="https://media.dev.to/cdn-cgi/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Frlcersegsc04i64u0tj9.png" class="article-body-image-wrapper"&gt;&lt;img src="https://media.dev.to/cdn-cgi/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Frlcersegsc04i64u0tj9.png" alt="Image description" width="624" height="334"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;I have highlighted to note that we deal with &lt;strong&gt;This account&lt;/strong&gt; in particular.&lt;br&gt;
Next select the policy to setup a role for.&lt;/p&gt;

&lt;p&gt;&lt;a href="https://media.dev.to/cdn-cgi/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2F3zilbksdo8dczyh75sp2.png" class="article-body-image-wrapper"&gt;&lt;img src="https://media.dev.to/cdn-cgi/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2F3zilbksdo8dczyh75sp2.png" alt="Image description" width="624" height="335"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;Finish up with the role creation.&lt;/p&gt;

&lt;p&gt;&lt;a href="https://media.dev.to/cdn-cgi/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fxlawolmmxui5ad212hdy.png" class="article-body-image-wrapper"&gt;&lt;img src="https://media.dev.to/cdn-cgi/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fxlawolmmxui5ad212hdy.png" alt="Image description" width="624" height="335"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;Now copy the &lt;strong&gt;ARN&lt;/strong&gt; (Amazon Resource Names) that will help in identifying resources uniquely.&lt;/p&gt;

&lt;p&gt;&lt;a href="https://media.dev.to/cdn-cgi/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fg6q00svh5ahmiv57jc2c.png" class="article-body-image-wrapper"&gt;&lt;img src="https://media.dev.to/cdn-cgi/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fg6q00svh5ahmiv57jc2c.png" alt="Image description" width="623" height="332"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;Step 5 – Setting Up/Creating a Snowflake Account.&lt;/strong&gt;&lt;br&gt;
To create a Snowflake account head to &lt;a href="https://signup.snowflake.com/?utm_cta=website-homepage-hero-button-start-for-free" rel="noopener noreferrer"&gt;Snowflake&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;&lt;a href="https://media.dev.to/cdn-cgi/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fwk0ac5jtxsz9h8fz91e4.png" class="article-body-image-wrapper"&gt;&lt;img src="https://media.dev.to/cdn-cgi/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fwk0ac5jtxsz9h8fz91e4.png" alt="Image description" width="624" height="336"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;After creating an account, we need to create our &lt;strong&gt;Warehouse&lt;/strong&gt; – as shown below.&lt;/p&gt;

&lt;p&gt;&lt;a href="https://media.dev.to/cdn-cgi/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fq6fm86803gb08etdc7zo.png" class="article-body-image-wrapper"&gt;&lt;img src="https://media.dev.to/cdn-cgi/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fq6fm86803gb08etdc7zo.png" alt="Image description" width="624" height="336"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;Displaying the pipeline dataset loaded from the AWS S3.&lt;/p&gt;

&lt;p&gt;&lt;a href="https://media.dev.to/cdn-cgi/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Facwgxwf8ibvjormqtv9c.png" class="article-body-image-wrapper"&gt;&lt;img src="https://media.dev.to/cdn-cgi/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Facwgxwf8ibvjormqtv9c.png" alt="Image description" width="624" height="335"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;That is all for this article.&lt;/p&gt;

</description>
    </item>
    <item>
      <title>Data Engineering for Beginners: A Step - by - Step Guide.</title>
      <dc:creator>Michael Obed </dc:creator>
      <pubDate>Tue, 31 Oct 2023 05:24:24 +0000</pubDate>
      <link>https://dev.to/obedm16/data-engineering-for-beginners-a-step-by-step-guide-1ane</link>
      <guid>https://dev.to/obedm16/data-engineering-for-beginners-a-step-by-step-guide-1ane</guid>
      <description>&lt;p&gt;&lt;a href="https://media.dev.to/cdn-cgi/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Foxy15091w1cygi1v0a0p.png" class="article-body-image-wrapper"&gt;&lt;img src="https://media.dev.to/cdn-cgi/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Foxy15091w1cygi1v0a0p.png" alt="Image description" width="800" height="451"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;Data Engineering&lt;/strong&gt; is one of the fastest rising career position in the data technology field. Data Engineering is one of the fastest rising career position in the data technology field. This is a complex field with task of making raw data usable to data scientists and groups within an organization. Data engineering encompasses numerous specialties of data science.&lt;br&gt;
A data engineer is a type of software engineer who creates big data ETL pipelines to manage the flow of data through the organization.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;Responsibilities of a Data Engineer&lt;/strong&gt;&lt;br&gt;
• Develops and maintains data pipelines&lt;br&gt;
• Ensures Data quality &amp;amp; accuracy &lt;br&gt;
• Design, develop and maintain database architecture&lt;br&gt;
• More of ETL process – (Extract Transform Load) This is an automated process to get data into a database from the source.&lt;br&gt;
• As long as the data engineer will be getting data from a given source, they will need to ensure data quality and accuracy.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;Skills Required to Become a Data Engineer&lt;/strong&gt;&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;&lt;em&gt;1. SQL/NoSQL&lt;/em&gt;&lt;/strong&gt;&lt;br&gt;
Structured Query Language or simply SQL (‘sequel’) is the cream when it comes to fundamental of skill-set for Data Engineers and any other related field of data and databases. Modern data is greatly stored in Relational Database Management Systems (RDBMS) and hence this demands highly aspiring Data Engineers to master the SQL. There are a lot of Relational Database Systems including &lt;em&gt;&lt;strong&gt;Oracle 23C, MySQL, SQL Server, PostgreSQL&lt;/strong&gt;&lt;/em&gt; just but to mention a few. &lt;/p&gt;

&lt;p&gt;For NoSQL, - This means that data is not stored in table format hence the data is in a non-relational database design does not require a schema, it offers rapid scalability to manage large and typically unstructured data sets. Examples of NoSQL database systems are &lt;em&gt;MongoDB&lt;/em&gt; and &lt;em&gt;Cassandra&lt;/em&gt;&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;&lt;em&gt;2. Python/R&lt;/em&gt;&lt;/strong&gt;&lt;br&gt;
To be able to work with data from source and create automation tasks around data, the data engineer need to have knowledge of a programming language with respective packages or libraries. Python so far is highly considered given to its versatility in the libraries like Pandas, Airflow. For R, a Data Engineer have at their disposal Tidyverse is a collection of R packages, primarily for data engineering and analytics. These packages are &lt;em&gt;ggplot2, purrr, tibble, dplyr, tidyr, stringr, readr, and forcats&lt;/em&gt;&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;&lt;em&gt;3. Apache Kafka&lt;/em&gt;&lt;/strong&gt;&lt;br&gt;
This is an open-source distributed event streaming platform widely used in data engineering for building real-time data pipelines and streaming applications. It is designed to handle high-throughput, fault-tolerant, and scalable data streaming, making it a popular choice for processing and managing large volumes of data in real time.&lt;br&gt;
It is used to design and implement robust, real-time data pipelines, enabling the efficient processing, storage, and analysis of streaming data.&lt;/p&gt;

&lt;p&gt;&lt;em&gt;&lt;strong&gt;4. AWS/Azure/GCP&lt;/strong&gt;&lt;/em&gt;&lt;br&gt;
As a Data Engineer, having a knowledge of Cloud Computing is an important part of you. Big Data which cannot be  &lt;/p&gt;

&lt;p&gt;&lt;em&gt;&lt;strong&gt;5. Apache Spark&lt;/strong&gt;&lt;/em&gt;&lt;br&gt;
Apache Spark is a powerful open-source distributed computing system that is widely used in the field of data engineering. It is designed to handle large-scale data processing tasks and is known for its speed, ease of use, and versatility in supporting a wide range of applications. &lt;br&gt;
It serves as a critical tool for performing data transformation, data integration, and data analytics. It provides a unified engine for big data processing that supports Java, Scala, Python, and R. This flexibility makes it easier for data engineers to work with different data formats and leverage their preferred programming languages for data processing tasks.&lt;br&gt;
Apache Spark is particularly valuable in data engineering for the following reasons:&lt;/p&gt;

&lt;ol&gt;
&lt;li&gt; &lt;strong&gt;&lt;em&gt;Speed and efficiency:&lt;/em&gt;&lt;/strong&gt; Apache Spark can process large datasets much faster than traditional data processing systems, thanks to its in-memory processing capabilities and optimized execution engine.&lt;/li&gt;
&lt;li&gt; &lt;em&gt;&lt;strong&gt;Scalability:&lt;/strong&gt;&lt;/em&gt; It can handle massive datasets and scale seamlessly to accommodate growing data volumes, making it suitable for handling big data challenges.&lt;/li&gt;
&lt;li&gt; &lt;em&gt;&lt;strong&gt;Fault tolerance:&lt;/strong&gt;&lt;/em&gt; It ensures fault tolerance by storing intermediate results during processing, enabling the system to recover from failures without losing data.&lt;/li&gt;
&lt;li&gt; &lt;em&gt;&lt;strong&gt;Versatility:&lt;/strong&gt;&lt;/em&gt; Apache Spark supports a wide range of data processing tasks, including batch processing, real-time streaming, machine learning, and graph processing, making it a versatile solution for diverse data engineering needs.
Data engineers utilize Apache Spark to build and maintain data pipelines, process and analyze large datasets, and perform complex data transformations. By leveraging its capabilities, data engineers can efficiently manage and process big data, extract valuable insights, and build robust data-driven applications and systems.&lt;/li&gt;
&lt;/ol&gt;

&lt;p&gt;&lt;strong&gt;&lt;em&gt;6. Data Structures and Algorithms.&lt;/em&gt;&lt;/strong&gt;&lt;br&gt;
This is crucial in data engineering, which are fundamental building blocks for the efficient management and processing of data.&lt;br&gt;
Data structures refer to the specific ways data is organized and stored in a memory of a computer, enabling efficient access, modification, and deletion of data. Algorithms are step-by-step procedures or formulas for performing specific tasks, such as searching, sorting, and data manipulation. Data engineering heavily relies on various algorithms to efficiently handle data operations, such as data transformation, data integration, and data aggregation.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;&lt;em&gt;7. Data Visualization&lt;/em&gt;&lt;/strong&gt;&lt;br&gt;
This is graphical representation of data and information. It plays a crucial role in presenting complex data in a visual format that is easy to understand and interpret. It enables data engineers to communicate insights and findings effectively to stakeholders, thus facilitating data-driven decision-making and understanding of complex data relationships.&lt;/p&gt;

&lt;p&gt;Tools of Business Intelligence used for Data Visualization are &lt;em&gt;&lt;strong&gt;Matplotlib, Power BI, Tableau, Seaborn&lt;/strong&gt;&lt;/em&gt; just to mention a few.&lt;/p&gt;

</description>
    </item>
    <item>
      <title>The Complete Guide to Time Series Model</title>
      <dc:creator>Michael Obed </dc:creator>
      <pubDate>Thu, 26 Oct 2023 06:15:44 +0000</pubDate>
      <link>https://dev.to/obedm16/the-complete-guide-to-time-series-model-4fcg</link>
      <guid>https://dev.to/obedm16/the-complete-guide-to-time-series-model-4fcg</guid>
      <description>&lt;p&gt;&lt;strong&gt;Time series analysis&lt;/strong&gt; comprises methods for analyzing time series data in order to extract meaningful statistics and other characteristics of the data while &lt;strong&gt;time series forecasting&lt;/strong&gt; is the use of a model to predict future values based on previously observed values.&lt;/p&gt;

&lt;p&gt;Combining time series analysis and time series forecasting we derive &lt;strong&gt;time series model&lt;/strong&gt; which refers to data points ordered in time used in forecasting the future.&lt;/p&gt;

&lt;h2&gt;
  
  
  Forecasting Models
&lt;/h2&gt;

&lt;p&gt;The tool used in conducting time series model is referred to as ARIMA an acronym for Auto Regressive Integrated Moving Average.&lt;/p&gt;

&lt;p&gt;&lt;em&gt;&lt;strong&gt;How ARIMA is applied&lt;/strong&gt;&lt;/em&gt;&lt;br&gt;
It is specified by three order parameters: &lt;strong&gt;p&lt;/strong&gt;, &lt;strong&gt;d&lt;/strong&gt;, &lt;strong&gt;q&lt;/strong&gt;.&lt;br&gt;
&lt;em&gt;&lt;strong&gt;P&lt;/strong&gt;&lt;/em&gt; refers to the number of autoregressive terms (&lt;strong&gt;AR&lt;/strong&gt;)&lt;br&gt;
&lt;strong&gt;&lt;em&gt;d&lt;/em&gt;&lt;/strong&gt; refers to how many non – seasonal differences are needed to achieve stationarity (&lt;strong&gt;I&lt;/strong&gt;)&lt;br&gt;
&lt;strong&gt;&lt;em&gt;q&lt;/em&gt;&lt;/strong&gt; refers to the number of lagged forecast errors in the prediction equation (&lt;strong&gt;MA&lt;/strong&gt;)&lt;/p&gt;

&lt;h2&gt;
  
  
  Stationarity of Time Series
&lt;/h2&gt;

&lt;p&gt;The stationarity of time series depends on:&lt;br&gt;
• Mean&lt;br&gt;
• Variance&lt;br&gt;
• Co-variance&lt;/p&gt;

&lt;p&gt;&lt;a href="https://media.dev.to/cdn-cgi/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Ficov4cvzwoef5f3l7prq.jpg" class="article-body-image-wrapper"&gt;&lt;img src="https://media.dev.to/cdn-cgi/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Ficov4cvzwoef5f3l7prq.jpg" alt="Image description" width="800" height="306"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;The variance of the series should not be a function of time.&lt;/p&gt;

&lt;p&gt;&lt;a href="https://media.dev.to/cdn-cgi/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2F6e0dqhrai25n4xc52vm5.jpg" class="article-body-image-wrapper"&gt;&lt;img src="https://media.dev.to/cdn-cgi/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2F6e0dqhrai25n4xc52vm5.jpg" alt="Image description" width="800" height="393"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;The covariance of the &lt;strong&gt;i th&lt;/strong&gt; term and the (&lt;strong&gt;i  + m&lt;/strong&gt;) th term should not be a function of time.&lt;/p&gt;

&lt;p&gt;&lt;em&gt;&lt;strong&gt;When the values are constant over a period of time.&lt;/strong&gt;&lt;/em&gt;&lt;/p&gt;

&lt;p&gt;&lt;a href="https://media.dev.to/cdn-cgi/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fj4mk8d59eboielffo8xv.jpg" class="article-body-image-wrapper"&gt;&lt;img src="https://media.dev.to/cdn-cgi/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fj4mk8d59eboielffo8xv.jpg" alt="Image description" width="563" height="281"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;h2&gt;
  
  
  &lt;strong&gt;Components Affecting Time Series&lt;/strong&gt;
&lt;/h2&gt;

&lt;p&gt;&lt;strong&gt;&lt;em&gt;Trend&lt;/em&gt;&lt;/strong&gt;&lt;br&gt;
This is the increase or decrease in the series over a period of time. It persists over a long period of time. &lt;/p&gt;

&lt;p&gt;&lt;em&gt;Example: Population growth over the years can be seen as an upward trend.&lt;/em&gt;&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;&lt;em&gt;Seasonality&lt;/em&gt;&lt;/strong&gt;&lt;br&gt;
Seasonality refers to the regular pattern of up and down fluctuations. It is a short –term variation occurring due to seasonal factors.&lt;/p&gt;

&lt;p&gt;&lt;em&gt;Example: sale of ice-cream increases on summer.&lt;/em&gt;&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;&lt;em&gt;Cyclicity&lt;/em&gt;&lt;/strong&gt;&lt;br&gt;
This is a medium – term variation caused by circumstances, which repeat in irregular intervals.&lt;/p&gt;

&lt;p&gt;&lt;em&gt;Example: 5 years of economic growth, followed by 2 years of economic recession, followed by 7 years of economic growth followed by 1 year of economic recession.&lt;/em&gt;&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;&lt;em&gt;Irregularity&lt;/em&gt;&lt;/strong&gt;&lt;br&gt;
This refers to the variations that occur due to unpredictable factors and also do not repeat in particular patterns.&lt;/p&gt;

&lt;p&gt;&lt;em&gt;Example: Variations caused by incidents like earthquakes, war, floods and other variations that have unpredictable factors.&lt;/em&gt;&lt;/p&gt;

&lt;p&gt;**&lt;/p&gt;

&lt;h2&gt;
  
  
  Moving Average
&lt;/h2&gt;

&lt;p&gt;**&lt;br&gt;
The moving average model is probably the most simple approach to time series modeling. This model simply states that the next observation is the mean of all past observations.&lt;/p&gt;

&lt;p&gt;&lt;em&gt;&lt;strong&gt;Exponential Smoothing&lt;/strong&gt;&lt;/em&gt;&lt;br&gt;
Exponential smoothing uses similar logic to moving average, but this time, a different decreasing weight is assigned to each observation. In other words, less importance is given to observations as we move further from the present.&lt;br&gt;
Mathematically, exponential smoothing is expressed as: &lt;/p&gt;

&lt;p&gt;&lt;a href="https://media.dev.to/cdn-cgi/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fcgp3onov1urznaub99jx.jpg" class="article-body-image-wrapper"&gt;&lt;img src="https://media.dev.to/cdn-cgi/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fcgp3onov1urznaub99jx.jpg" alt="Image description" width="512" height="56"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;Alpha is a smoothing factor that takes values between zero and one. It determines how fast the weight decreases for previous observations.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;&lt;em&gt;Double Exponential Smoothing&lt;/em&gt;&lt;/strong&gt;&lt;br&gt;
Double exponential smoothing is used when there is a trend in the time series. In that case, we use this technique, which is simply a recursive use of exponential smoothing twice.&lt;br&gt;
Mathematically:&lt;/p&gt;

&lt;p&gt;&lt;a href="https://media.dev.to/cdn-cgi/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2F50yf0yfat8e6ep2ihcta.jpg" class="article-body-image-wrapper"&gt;&lt;img src="https://media.dev.to/cdn-cgi/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2F50yf0yfat8e6ep2ihcta.jpg" alt="Image description" width="512" height="108"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;Here, beta is the trend smoothing factor, and it takes values between zero and one.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;&lt;em&gt;Triple Exponential Smoothing&lt;/em&gt;&lt;/strong&gt;&lt;br&gt;
This method extends double exponential smoothing by adding a seasonal smoothing factor. Of course, this is useful if you notice seasonality in your time series.&lt;br&gt;
Mathematically, triple exponential smoothing is expressed as:&lt;/p&gt;

&lt;p&gt;&lt;a href="https://media.dev.to/cdn-cgi/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fm04u8x2xgdhqndbbtey7.jpg" class="article-body-image-wrapper"&gt;&lt;img src="https://media.dev.to/cdn-cgi/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fm04u8x2xgdhqndbbtey7.jpg" alt="Image description" width="512" height="234"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;h2&gt;
  
  
  Why Time Series Models?
&lt;/h2&gt;

&lt;p&gt;There are a number of reasons why businesses and organizations through their respective Data personal conduct Time series analysis. The following are some of the applications of the Time Series Modeling.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;_Forecasting Future Trends _&lt;/strong&gt;&lt;br&gt;
Based on previous data collected through time series models, businesses can predict how future trends may develop to protect their financial resources, explore new markets, restock inventory and perform other tasks. &lt;/p&gt;

&lt;p&gt;&lt;strong&gt;&lt;em&gt;Detecting Anomalies&lt;/em&gt;&lt;/strong&gt;&lt;br&gt;
Time series models also allow organizations to more easily spot data shifts that may signal unusual behavior or changes in the market.&lt;/p&gt;

&lt;p&gt;&lt;em&gt;&lt;strong&gt;Determining Patterns&lt;/strong&gt;&lt;/em&gt;&lt;br&gt;
Businesses that rely on seasonal sales, monthly online traffic spikes and other repetitive behavior can establish expectations based on time series models, gauging their overall health and performance.&lt;/p&gt;

&lt;h2&gt;
  
  
  Forecasting with Time Series Models
&lt;/h2&gt;

&lt;p&gt;&lt;strong&gt;&lt;em&gt;Healthcare&lt;/em&gt;&lt;/strong&gt;&lt;br&gt;
Time series models can be used to monitor the spread of diseases by observing how many people transmit a disease and how many people die after being infected.&lt;/p&gt;

&lt;p&gt;&lt;em&gt;&lt;strong&gt;Agriculture&lt;/strong&gt;&lt;/em&gt;&lt;br&gt;
Time series models take into account seasonal temperatures, the number of rainy days each month and other variables over the course of years, allowing agricultural workers to assess environmental conditions and ensure a successful harvest. &lt;/p&gt;

&lt;p&gt;&lt;strong&gt;&lt;em&gt;Finance&lt;/em&gt;&lt;/strong&gt;&lt;br&gt;
Financial analysts can leverage time series models to record sales numbers for each month and predict potential stock market behavior.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;&lt;em&gt;Cybersecurity&lt;/em&gt;&lt;/strong&gt;&lt;br&gt;
IT and cybersecurity teams can develop patterns in user behavior with time series models, allowing them to be aware of when behavior doesn’t align with normal trends.&lt;/p&gt;

&lt;p&gt;&lt;em&gt;&lt;strong&gt;Retail&lt;/strong&gt;&lt;/em&gt;&lt;br&gt;
Retailers may apply time series models to study how other companies’ prices and the number of customer purchases change over time, helping them optimize prices.&lt;/p&gt;

</description>
    </item>
    <item>
      <title>Exploratory Data Analysis Using Data Visualization Techniques</title>
      <dc:creator>Michael Obed </dc:creator>
      <pubDate>Mon, 16 Oct 2023 14:03:39 +0000</pubDate>
      <link>https://dev.to/obedm16/exploratory-data-analysis-using-data-visualization-techniques-57om</link>
      <guid>https://dev.to/obedm16/exploratory-data-analysis-using-data-visualization-techniques-57om</guid>
      <description>&lt;p&gt;&lt;a href="https://media.dev.to/cdn-cgi/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2F29xbggxej9b3k4tqgmfs.jpg" class="article-body-image-wrapper"&gt;&lt;img src="https://media.dev.to/cdn-cgi/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2F29xbggxej9b3k4tqgmfs.jpg" alt="Image description" width="800" height="450"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;Exploratory Data Analysis or EDA refers to the process of analyzing and summarizing datasets to gain insights into the data. This helps in understanding the data and identify patterns, relationships and anomalies.&lt;br&gt;
EDA is a crucial step in any data analysis project. Some of the techniques used in EDA include Python, R, SQL, Excel &lt;br&gt;
 The objectives of EDA is to:&lt;/p&gt;

&lt;p&gt;-Enable unexpected discoveries in the data&lt;br&gt;
-Suggest hypotheses about the causes of observed phenomena&lt;br&gt;
-Assess assumptions on which statistical inference will be based&lt;br&gt;
-Support the selection of appropriate statistical tools and techniques&lt;br&gt;
-Provide a basis for further data collection through surveys or experiments&lt;br&gt;
To ease more on analyzing data Python is discussed more in details.&lt;/p&gt;

&lt;div class="highlight js-code-highlight"&gt;
&lt;pre class="highlight plaintext"&gt;&lt;code&gt;              **Why Python for EDA?**
&lt;/code&gt;&lt;/pre&gt;

&lt;/div&gt;

&lt;ul&gt;
&lt;li&gt;Python is a general yet powerful and versatile programming language widely used in Data Analysis.&lt;/li&gt;
&lt;li&gt;Python is rich in libraries and tools that make it easy to perform the EDA tasks. To mention these libraries are Pandas, NumPy, Matplotlib, Seaborn and Plotly – to be discussed further.&lt;/li&gt;
&lt;li&gt;
&lt;p&gt;Python is easy to learn and to be used in EDA tasks hence it is an ideal choice for beginners and experts alike.&lt;/p&gt;
&lt;pre class="highlight plaintext"&gt;&lt;code&gt;        **Introduction to Data Visualization**
&lt;/code&gt;&lt;/pre&gt;
&lt;/li&gt;
&lt;li&gt;&lt;p&gt;Data Visualization helps in the communication of insights derived from complex datasets.&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;Python is preferred to do the task since it comes in handy with libraries that enables a Data Analyst or Data Scientist to create visualization for the data prepared.&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;The libraries are matplotlib, seaborn, and plotly help convey the findings effectively.&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;Matplotlib helps to create a 2d plot in Python.&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;
&lt;p&gt;Plotly helps to offer interactive capabilities to visualizations.&lt;/p&gt;
&lt;pre class="highlight plaintext"&gt;&lt;code&gt;               **Steps in EDA**
&lt;/code&gt;&lt;/pre&gt;
&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;To conduct EDA appropriately there are steps to be followed to ensure the Data Scientist in the end has a clean data. The steps are: &lt;/p&gt;

&lt;p&gt;Before we dive into EDA lets prepare our environment.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;&lt;em&gt;Note:&lt;/em&gt;&lt;/strong&gt; I used VS Code studio and Anaconda to conduct the Exploratory Data Analysis for the dataset provided. &lt;/p&gt;

&lt;p&gt;Also, your PC needs to have a core i5 processor and at least 8GB RAM due to high data analysis processing power requirements. &lt;br&gt;
Download VS code here and Anaconda Error! Hyperlink reference not valid.&lt;/p&gt;

&lt;p&gt;After Downloading Anaconda and installing the interface should appear as follows:&lt;/p&gt;

&lt;p&gt;&lt;a href="https://media.dev.to/cdn-cgi/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2F1k57pgyvl3okb7k6wtnl.jpg" class="article-body-image-wrapper"&gt;&lt;img src="https://media.dev.to/cdn-cgi/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2F1k57pgyvl3okb7k6wtnl.jpg" alt="Image description" width="800" height="428"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;Now click &lt;strong&gt;Launch&lt;/strong&gt; - &lt;em&gt;Notebook(Jupyter)&lt;/em&gt;. The interface of Jupyter is as shown below.&lt;/p&gt;

&lt;p&gt;&lt;a href="https://media.dev.to/cdn-cgi/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fa0g79oo2z1u8ocerpyip.jpg" class="article-body-image-wrapper"&gt;&lt;img src="https://media.dev.to/cdn-cgi/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fa0g79oo2z1u8ocerpyip.jpg" alt="Image description" width="800" height="426"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;Now the remaining part is to conduct the EDA process.&lt;/p&gt;

&lt;p&gt;To Use VS Code – after downloading. Create a folder for your project.&lt;/p&gt;

&lt;p&gt;&lt;a href="https://media.dev.to/cdn-cgi/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Feny4m3vs4sqtk25ofwzq.jpg" class="article-body-image-wrapper"&gt;&lt;img src="https://media.dev.to/cdn-cgi/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Feny4m3vs4sqtk25ofwzq.jpg" alt="Image description" width="800" height="428"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;To run the Python file open terminal via keyboard shortcut &lt;strong&gt;CTRL + `&lt;/strong&gt; (called back-tick) or locate the 3 dots on tab bar. &lt;/p&gt;

&lt;p&gt;&lt;a href="https://media.dev.to/cdn-cgi/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fl1efiwa9fldkytdt3s4o.jpg" class="article-body-image-wrapper"&gt;&lt;img src="https://media.dev.to/cdn-cgi/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fl1efiwa9fldkytdt3s4o.jpg" alt="Image description" width="800" height="426"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;Run this command C:/ProgramData/anaconda3/python.exe "d:/Data Science BootCamp/Week 2/eda.py" (replace path with your own path to the file where scripts are written)&lt;/p&gt;

&lt;div class="highlight js-code-highlight"&gt;
&lt;pre class="highlight plaintext"&gt;&lt;code&gt;                  **Steps in EDA**
&lt;/code&gt;&lt;/pre&gt;

&lt;/div&gt;

&lt;p&gt;To conduct EDA appropriately there are steps to be followed to ensure the Data Scientist in the end has a clean data. The steps are: &lt;/p&gt;

&lt;ol&gt;
&lt;li&gt;Understand the data at hand – the dataset to be analyzed. This can be done by first importing the Python Libraries.&lt;/li&gt;
&lt;/ol&gt;

&lt;p&gt;&lt;a href="https://media.dev.to/cdn-cgi/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fqh0jtjfdvcy5ekkzc9j5.jpg" class="article-body-image-wrapper"&gt;&lt;img src="https://media.dev.to/cdn-cgi/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fqh0jtjfdvcy5ekkzc9j5.jpg" alt="Image description" width="800" height="426"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;ol&gt;
&lt;li&gt;Clean the data using the imported libraries. This is done by inspecting the data.&lt;/li&gt;
&lt;li&gt;Dirty data can contain null values, duplicates, inconsistent column names, spaces in between the data.&lt;/li&gt;
&lt;li&gt;Then analyze the relationship in data variables. &lt;/li&gt;
&lt;/ol&gt;

</description>
    </item>
    <item>
      <title>Data Science for Beginners: 2023 - 2024 Complete Roadmap</title>
      <dc:creator>Michael Obed </dc:creator>
      <pubDate>Fri, 06 Oct 2023 10:23:34 +0000</pubDate>
      <link>https://dev.to/obedm16/data-science-for-beginners-2023-2024-complete-roadmap-3j6g</link>
      <guid>https://dev.to/obedm16/data-science-for-beginners-2023-2024-complete-roadmap-3j6g</guid>
      <description>&lt;p&gt;**&lt;/p&gt;

&lt;h2&gt;
  
  
  Data Science, Data Analysis, Data Engineering and Analytical Engineering
&lt;/h2&gt;

&lt;p&gt;**&lt;/p&gt;

&lt;p&gt;&lt;a href="https://media.dev.to/cdn-cgi/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fycowstgcrc4b9p3mx6zp.jpg" class="article-body-image-wrapper"&gt;&lt;img src="https://media.dev.to/cdn-cgi/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fycowstgcrc4b9p3mx6zp.jpg" alt="Image description" width="613" height="334"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;Data is the new gold and research done The Economist recently called data the world's most valuable resource and now; let’s see major fields related to data specialization covering Data Science, Data Analysis, Data Engineering and Analytics Engineering with roadmap in Data Science.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;Data Science&lt;/strong&gt; – This is a field of study that combines domain expertise, programming skills and knowledge of mathematics and statistics to extract meaningful insights from data. Data Science is a new field in that its existence paved in the way from the late 1980 and around 2015, this lucrative and promising field gained popularity and few years later; it became a very popular and regarded among high – paying industry.&lt;br&gt;
Regarding data – this applies to all field, the stages are: capture, maintain, process, analyze communicate. After conducting the data stages all expected now is, through data – facilitate strategic business decisions.&lt;/p&gt;

&lt;p&gt;To build a career in Data Science an individual has either go for formal path – earning a degree in Computer Science, Business Information Technology, degree or Masters in Statistics/Mathematics, or apprenticeships, boot camps and certifications or self-taught.&lt;/p&gt;

&lt;p&gt;&lt;em&gt;&lt;strong&gt;Courses and Certifications for career in Data Science&lt;/strong&gt;&lt;/em&gt;&lt;br&gt;
&lt;strong&gt;&lt;a href="https://www.coursera.org/professional-certificates/ibm-data-science" rel="noopener noreferrer"&gt;IBM Data Science Professional Certificate (Coursera)&lt;/a&gt;&lt;/strong&gt;&lt;/p&gt;

&lt;p&gt;&lt;a href="https://media.dev.to/cdn-cgi/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fq60drtayn0nf6obv010j.jpg" class="article-body-image-wrapper"&gt;&lt;img src="https://media.dev.to/cdn-cgi/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fq60drtayn0nf6obv010j.jpg" alt="Image description" width="489" height="100"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;This is a beginner level if paced at 10 hours a week will take the learner 5 months to finish with a capstone project for their portfolio and cement the Data Science skills.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;&lt;a href="https://www.udemy.com/course/data-science-machine-learning-data-analysis-python-r/?utm_source=adwords&amp;amp;utm_medium=udemyads&amp;amp;utm_campaign=DSA_Catchall_la.EN_cc.ROW&amp;amp;utm_content=deal4584&amp;amp;utm_term=_._ag_88010211481_._ad_535397282061_._kw__._de_c_._dm__._pl__._ti_dsa-392284169515_._li_1009824_._pd__._&amp;amp;matchtype=&amp;amp;gclid=CjwKCAjw4P6oBhBsEiwAKYVkqxBfMP2S4puPCJPHJfUXXnYw01zxGeghJS6-fKGLc2bFydhYkOoOYxoCXHwQAvD_BwE" rel="noopener noreferrer"&gt;Data Science, Machine Learning, Data Analysis, Python &amp;amp; R (Udemy)&lt;/a&gt;&lt;/strong&gt;&lt;br&gt;
&lt;a href="https://media.dev.to/cdn-cgi/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Ffbjyq5wsvt32xzgqhrhi.jpg" class="article-body-image-wrapper"&gt;&lt;img src="https://media.dev.to/cdn-cgi/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Ffbjyq5wsvt32xzgqhrhi.jpg" alt="Image description" width="471" height="223"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;This is free Data Science, Machine Learning, Data Analysis, Data Visualization using Python and R Programming that the learner in 8 hours will get the nitty gritty in Data Science and Machine Learning.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;&lt;a href="https://www.freecodecamp.org/learn/scientific-computing-with-python/" rel="noopener noreferrer"&gt;FreeCodeCamp&lt;/a&gt;&lt;/strong&gt;&lt;br&gt;
&lt;a href="https://media.dev.to/cdn-cgi/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fc1v7d9r57x1n4xghtlmf.jpg" class="article-body-image-wrapper"&gt;&lt;img src="https://media.dev.to/cdn-cgi/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fc1v7d9r57x1n4xghtlmf.jpg" alt="Image description" width="575" height="276"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;Python for everybody&lt;/strong&gt; is a free video course series that teaches the basics of using Python 3.&lt;br&gt;
The courses were created by Dr. Charles Severance (also known as Dr. Chuck). He is a Clinical Professor at the University of Michigan School of Information, where he teaches various technology-oriented courses including programming, database design, and web development.&lt;br&gt;
This course is self-paced. You’ll be able to learn consistently at your own convenience.&lt;/p&gt;

&lt;p&gt;In terms of job growth, all jobs in tech are expected to grow by 13% over the next 10 years, that is according to &lt;strong&gt;Shane Hummus&lt;/strong&gt; - a renown Data - content creator on YouTube: &lt;a href="https://youtu.be/O9nf1CqjGzI?si=2E7U8UXHAlguavqv" rel="noopener noreferrer"&gt;https://youtu.be/O9nf1CqjGzI?si=2E7U8UXHAlguavqv&lt;/a&gt; &lt;/p&gt;

&lt;p&gt;A &lt;strong&gt;Data Scientist&lt;/strong&gt; is a professional responsible for collecting, analyzing and interpreting extremely large amounts of data.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;Data Analysis&lt;/strong&gt; – The process of inspecting, cleansing, transforming and modeling data with the goal of discovering useful information, informing conclusions, and supporting decision-making.&lt;br&gt;
Data analysis involves working with smaller, structured datasets to answer specific questions or solve specific problems.&lt;/p&gt;

&lt;p&gt;&lt;em&gt;&lt;strong&gt;Courses and Certifications for career in Data Analysis&lt;/strong&gt;&lt;/em&gt;&lt;br&gt;
&lt;strong&gt;&lt;a href="https://www.coursera.org/professional-certificates/google-data-analytics" rel="noopener noreferrer"&gt;Google Data Analytics&lt;/a&gt;&lt;/strong&gt;&lt;br&gt;
&lt;a href="https://media.dev.to/cdn-cgi/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fobcldg97lk1sl7uncqjj.jpg" class="article-body-image-wrapper"&gt;&lt;img src="https://media.dev.to/cdn-cgi/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fobcldg97lk1sl7uncqjj.jpg" alt="Image description" width="705" height="113"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;This is a paid course fit for new learners in the world of Data Analysis. Even though it is a paid – subscription based, good news is that it has Financial Aid request available to cater for those who are limited to raise the per-module cost of around $ 39 USD. This course covers Excel, R programming, SQL, Tableau, working with GCP – Google Cloud Platform and Foundations of Data Analysis and much more.&lt;/p&gt;

&lt;p&gt;Majority of what is learnt by in Data Science is also covered in Data Analysis. It can be noted that a Data Scientist can only work around the dataset where a Data Analyst is not able.&lt;/p&gt;

&lt;p&gt;A &lt;strong&gt;Data Analyst&lt;/strong&gt; is a professional able to use statistical methods to test hypotheses and draw conclusions from the data.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;Data Engineering&lt;/strong&gt; – This refers to the art of designing and building systems for collecting, storing, and analyzing data at scale. It is a broad field with applications in just about every industry.&lt;br&gt;
A &lt;strong&gt;Data Engineer&lt;/strong&gt; is a professional responsible for laying the foundations for the acquisition, storage, transformation, and management of data in an organization. They manage the design, creation, and maintenance of database architecture and data processing systems; this ensures that the subsequent work of analysis, visualization, and machine learning models development can be carried out seamlessly, continuously, securely, and effectively. &lt;/p&gt;

&lt;p&gt;&lt;strong&gt;Analytics Engineering&lt;/strong&gt; –  This is a multi-disciplinary role and can be defined as the discipline of engineering applied to the practice of analytics and big data.&lt;br&gt;
An Analytical Engineer refers to professional who blends technical expertise with domain knowledge to craft meaningful insights from data and deliver them to users in a timely manner.&lt;br&gt;
(Separate article coming up)&lt;br&gt;
**&lt;/p&gt;

&lt;h2&gt;
  
  
  Roadmap to Data Science
&lt;/h2&gt;

&lt;p&gt;**&lt;br&gt;
Roadmaps refers to procedures that determine a goal or the desired outcome and feature the significant steps or milestones required to reach it.&lt;br&gt;
&lt;em&gt;Learn the fundamentals of data science.&lt;/em&gt;&lt;br&gt;
&lt;strong&gt;Programming&lt;/strong&gt;: Having a good knowledge in programming is a key feature to any aspiring Data Scientist. Top languages are Python and R. Choose one tool and build skills to help you do the job.&lt;br&gt;
&lt;strong&gt;Statistics&lt;/strong&gt;: Learn probability, descriptive statistics, and inferential statistics.&lt;br&gt;
&lt;strong&gt;Machine learning&lt;/strong&gt;: a subfield of Computer Science that gives computers the ability to learn without being explicitly programmed. Machine Learning is used image recognition, natural language processing, and fraud detection.&lt;br&gt;
&lt;strong&gt;Visualization&lt;/strong&gt; – When all is done with data preparation, A Data Scientist might go step ahead to create visualization using Tableau, Power BI and other industry – standard tools and softwares.&lt;/p&gt;

&lt;p&gt;&lt;em&gt;&lt;strong&gt;Gain experience with data science tools&lt;/strong&gt;&lt;/em&gt;&lt;br&gt;
&lt;strong&gt;Python libraries&lt;/strong&gt;: NumPy, Pandas, Matplotlib, Seaborn, and Scikit-learn are some of the most popular Python libraries for Data Science. They provide a wide range of functions for data manipulation, visualization and Machine Learning.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;SQL&lt;/strong&gt;: Structured Query Language is used to query and manage relational databases. MySQL, PostgreSQL, and Oracle are some of the most popular SQL databases.&lt;br&gt;
&lt;strong&gt;Cloud computing platforms&lt;/strong&gt;: Cloud computing platforms like AWS, Azure, and Google Cloud Platform offer a variety of services for data science, such as data storage, data processing, and machine learning.&lt;br&gt;
&lt;strong&gt;&lt;em&gt;Build a portfolio of data science projects&lt;/em&gt;&lt;/strong&gt;&lt;br&gt;
• Choose projects that are challenging yet realistic.&lt;br&gt;
• Document your work thoroughly – write article about the project.&lt;br&gt;
• Make your code clean, reusable and concise.&lt;br&gt;
• Host your projects on a public platform like GitHub.&lt;/p&gt;

&lt;p&gt;&lt;em&gt;&lt;strong&gt;Network with those in the field of Data Science&lt;/strong&gt;&lt;/em&gt;&lt;br&gt;
Tips for networking:&lt;br&gt;
• Be active on social media platforms like LinkedIn and Twitter.&lt;br&gt;
• Attend data science meetups and conferences.&lt;br&gt;
• Connect with other data scientists on LinkedIn.&lt;br&gt;
• Reach out to data scientists you admire and ask for advice.&lt;/p&gt;

&lt;p&gt;&lt;em&gt;&lt;strong&gt;Apply for data science jobs&lt;/strong&gt;&lt;/em&gt;&lt;br&gt;
Tailor your resume and cover letter to each job you apply for.&lt;br&gt;
Highlight your relevant skills and experience in your resume and cover letter.&lt;br&gt;
Practice answering Data Science interviews.&lt;/p&gt;

&lt;p&gt;That is all for this work. Happy kick-start Data Science Career.&lt;/p&gt;

</description>
    </item>
  </channel>
</rss>
