<?xml version="1.0" encoding="UTF-8"?>
<rss version="2.0" xmlns:atom="http://www.w3.org/2005/Atom" xmlns:dc="http://purl.org/dc/elements/1.1/">
  <channel>
    <title>DEV Community: SeattleDataGuy</title>
    <description>The latest articles on DEV Community by SeattleDataGuy (@seattledataguy).</description>
    <link>https://dev.to/seattledataguy</link>
    <image>
      <url>https://media2.dev.to/dynamic/image/width=90,height=90,fit=cover,gravity=auto,format=auto/https:%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Fuser%2Fprofile_image%2F177966%2F0f8eb580-5b46-4ba7-8390-18d85b04c7be.jpg</url>
      <title>DEV Community: SeattleDataGuy</title>
      <link>https://dev.to/seattledataguy</link>
    </image>
    <atom:link rel="self" type="application/rss+xml" href="https://dev.to/feed/seattledataguy"/>
    <language>en</language>
    <item>
      <title>Let’s Move Fast And Get Rid Of Data Engineers</title>
      <dc:creator>SeattleDataGuy</dc:creator>
      <pubDate>Sat, 17 Sep 2022 20:57:53 +0000</pubDate>
      <link>https://dev.to/seattledataguy/lets-move-fast-and-get-rid-of-data-engineers-4d4e</link>
      <guid>https://dev.to/seattledataguy/lets-move-fast-and-get-rid-of-data-engineers-4d4e</guid>
      <description>&lt;p&gt;We live in a data world that wants to move fast.&lt;/p&gt;

&lt;p&gt;Just shove your data into a data lake and we'll figure it out later.&lt;/p&gt;

&lt;p&gt;Create this table for this one dashboard that we will only look at once and then it will join the hundreds(if not thousands) of other dashboards that are ignored.&lt;/p&gt;

&lt;p&gt;All of which can work in the short term but often leaves behind a lot of technical debt. One project I recently worked on literally dumped all of their raw data into an S3 bucket with no form of breakdown source or timing and it was quite chaotic to grasp what was going on.&lt;/p&gt;

&lt;p&gt;All of this is driven by executives who need their data yesterday and don't want to wait as well as by the software and analysts teams who are often driven by very different motivations.&lt;/p&gt;

&lt;p&gt;With all this pressure to move fast coming from all sides, an interesting solution I have come across a few times is...Let's just get rid of data engineering and governance.&lt;/p&gt;

&lt;h2&gt;
  
  
  Let's Just Get Rid Of Data Engineers?
&lt;/h2&gt;

&lt;p&gt;&lt;a href="https://media.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fmiro.medium.com%2Fmax%2F1400%2F0%2AfMupendJH50wJaFu.png" class="article-body-image-wrapper"&gt;&lt;img src="https://media.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fmiro.medium.com%2Fmax%2F1400%2F0%2AfMupendJH50wJaFu.png" alt="data engineering diagram"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;&lt;a href="https://youtu.be/QU7jqnSbZQQ?t=448" rel="noopener noreferrer"&gt;Inspired By Codestrap&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;A recent conversation I had with a head of data forced me to pause.&lt;/p&gt;

&lt;p&gt;They brought up how at several organizations they had worked for made the decision to cut data engineering. This allowed their analysts and data scientists to use dbt and &lt;a href="https://www.theseattledataguy.com/scaling-airflow-astronomer-vs-cloud-composer-vs-managed-workflows-for-apache-airflow/" rel="noopener noreferrer"&gt;cloud composer&lt;/a&gt; to build tables as need be. After all these end-users knew what data they needed and were now empowered to get it.&lt;/p&gt;

&lt;p&gt;&lt;a href="https://medium.com/airbnb-engineering/data-quality-at-airbnb-e582465f3ef7#:~:text=The%20Data%20Quality%20Initiative,-In%20early%202019&amp;amp;text=Ensure%20clear%20ownership%20for%20all,is%20trustworthy%20and%20routinely%20validated" rel="noopener noreferrer"&gt;Now in a world where companies like Airbnb recently realized why they &lt;/a&gt;&lt;a href="https://medium.com/airbnb-engineering/data-quality-at-airbnb-e582465f3ef7#:~:text=The%20Data%20Quality%20Initiative,-In%20early%202019&amp;amp;text=Ensure%20clear%20ownership%20for%20all,is%20trustworthy%20and%20routinely%20validated" rel="noopener noreferrer"&gt;need &lt;/a&gt;&lt;a href="https://medium.com/airbnb-engineering/data-quality-at-airbnb-e582465f3ef7#:~:text=The%20Data%20Quality%20Initiative,-In%20early%202019&amp;amp;text=Ensure%20clear%20ownership%20for%20all,is%20trustworthy%20and%20routinely%20validated" rel="noopener noreferrer"&gt;to increase their data engineering initiatives this is always odd to hear.&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;Of course the end result for these companies was they did need to re-invest in data engineering.&lt;/p&gt;

&lt;p&gt;But why cut data engineering in the in the first place?&lt;/p&gt;

&lt;p&gt;Costs aside. Data engineering is often viewed as a bottleneck.&lt;/p&gt;

&lt;p&gt;If you're a software engineer you're likely getting chased down by data engineers who want you to tell them before you update tables on the application. Maybe, they are even asking you to add in &lt;a href="https://dataproducts.substack.com/p/the-rise-of-data-contracts?r=7hcrb&amp;amp;s=w&amp;amp;utm_campaign=post&amp;amp;utm_medium=web" rel="noopener noreferrer"&gt;extra layers like data contracts&lt;/a&gt;.&lt;/p&gt;

&lt;p&gt;Yikes. You don't want that. This means you won't be able to deliver that feature before your next review cycle.&lt;/p&gt;

&lt;p&gt;If you're on the data science and analyst side, well then you often have to wait for &lt;a href="https://www.youtube.com/watch?v=w24hpnPg-sk&amp;amp;t=1s" rel="noopener noreferrer"&gt;data engineering&lt;/a&gt; to find time to pull your data into "core tables". Some of these&lt;a href="https://seattledataguy.substack.com/p/setting-standards-for-your-data-team" rel="noopener noreferrer"&gt; data engineers might even insist on making sure to implement some level of standardizatio&lt;/a&gt;n and data governance.&lt;/p&gt;

&lt;p&gt;All of which means slower access to data.&lt;/p&gt;

&lt;p&gt;There is a reason that data engineering and data governance exist. But if you really do want to get rid of your data engineering teams then here are some points you must consider.&lt;/p&gt;

&lt;h2&gt;
  
  
  Tracking Historical Data
&lt;/h2&gt;

&lt;p&gt;On a few projects I completed earlier this year I noted there was a lack of tracking historical data. This will eventually lead to problems when your management asks:&lt;/p&gt;

&lt;blockquote&gt;
&lt;p&gt;How much revenue did we have per customer in NYC year over year?&lt;/p&gt;
&lt;/blockquote&gt;

&lt;p&gt;Because all you will be reporting on is the current state of the data and not how it has changed over time.&lt;/p&gt;

&lt;p&gt;The classic way to track changes in dimensions and entities is to use SCD or &lt;a href="https://towardsdatascience.com/data-analysts-primer-to-slowly-changing-dimensions-d087c8327e08" rel="noopener noreferrer"&gt;slowly changing dimensions&lt;/a&gt;.&lt;/p&gt;

&lt;p&gt;But there are other options.&lt;/p&gt;

&lt;p&gt;&lt;a href="https://www.linkedin.com/in/ACoAAA5yYyYBYUSJbRhZcqtR3-YIXUwGM6vz-aE" rel="noopener noreferrer"&gt;Ahmed Elsamadisi&lt;/a&gt; has discussed the idea of an &lt;a href="https://www.activityschema.com/" rel="noopener noreferrer"&gt;activity schema&lt;/a&gt; that can help track changes on entities in a streaming fashion.&lt;/p&gt;

&lt;p&gt;And at some other companies just store snapshots of every day in separate partitions in a table(not my favorite). The result of all of these is the ability for an end-user to track historical data over time. Meaning that when an analyst is asked about a year over year comparisons, they can answer accurately.&lt;/p&gt;

&lt;p&gt;All that said, I would say there is a far more difficult problem that most companies need to deal with.&lt;/p&gt;

&lt;h2&gt;
  
  
  Data Integration
&lt;/h2&gt;

&lt;p&gt;As pointed out by &lt;a href="https://www.linkedin.com/pulse/snowflake-critique-bill-inmon/?trackingId=2BENsSmxTgWxcc7v6MOyTQ%3D%3D" rel="noopener noreferrer"&gt;Bill Inmon earlier this year&lt;/a&gt;:&lt;/p&gt;

&lt;blockquote&gt;
&lt;p&gt;In order to be a data warehouse an organization must integrate data. If you don't do integration of data then you are not a data warehouse...&lt;/p&gt;

&lt;p&gt;...But when it comes to integrating data I don't see that Snowflake understands that at all. If anything Snowflake is an embodiment of the Kimball model.&lt;/p&gt;
&lt;/blockquote&gt;

&lt;p&gt;There are many cases where in order to move fast a team might load data from each data source as is without thinking about how it integrates with all the other various sources.&lt;/p&gt;

&lt;p&gt;While working at Facebook, yes we were spoiled by the free lunches.&lt;/p&gt;

&lt;p&gt;But truthfully, as a data engineer, I felt spoiled by how well integrated all the data was. Compared to when I worked at a hospital and was trying to integrate a finance system into a project management system where the project id field was an open text field that allowed for users to put multiple project ids into it(as well as the occasional budget number), Facebook was amazing.&lt;/p&gt;

&lt;p&gt;&lt;a href="https://media.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fmiro.medium.com%2Fmax%2F1400%2F0%2A_PpNr1dewq8RGlsk.png" class="article-body-image-wrapper"&gt;&lt;img src="https://media.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fmiro.medium.com%2Fmax%2F1400%2F0%2A_PpNr1dewq8RGlsk.png" alt="data engineering integration"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;It probably didn't help that the field was called project number not project id...&lt;/p&gt;

&lt;p&gt;This was because the data sources all interacted with each other on an application level. Meaning there had to be ids that the sources themselves shared.&lt;/p&gt;

&lt;p&gt;In fact, instead of having to figure out how we would join data together, we often had to figure out which IDs we would remove to avoid confusing analysts on which ID is the core ID to join on.&lt;/p&gt;

&lt;p&gt;Data integration is often skipped when your team is just ingesting data via EL. Why think about integration? There is a current report that needs to be answered and it doesn't require bringing in another data source. However, the problem is eventually it will. Eventually, someone will want to ask about data from multiple sources.&lt;/p&gt;

&lt;p&gt;In the end, you not only start running into integration issues but governance issues.&lt;/p&gt;

&lt;h2&gt;
  
  
  Data Governance
&lt;/h2&gt;

&lt;p&gt;Finally governance. Now at large companies, data governance is usually pretty obvious in terms of necessity. There are committees that spend hours deciding how they will implement the various processes and policies to protect, standardize and better manage the use of data.&lt;/p&gt;

&lt;p&gt;Entire departments are dedicated to these task. In these cases, the data engineering team is usually not involved in these decisions or at the very least they are not the SME. They might be the ones that programmatically implement the policies made by data governance. But are not the actual expert.&lt;/p&gt;

&lt;p&gt;Of course, these companies are also dealing with tens if not hundreds of applications(many duplicates).&lt;/p&gt;

&lt;p&gt;But there are a lot of data systems in the SMB and mid-market space too and not having some form of data governance strategy in a world that is becoming more data-aware poses a lot of risks.&lt;/p&gt;

&lt;h2&gt;
  
  
  Everything Has A Cost
&lt;/h2&gt;

&lt;p&gt;Moving fast can work out early on because much of the modeling and data pipeline debt isn't apparent(and I didn't even reference data quality). However, as a company grows and its data needs mature limitations will bubble to the surface.&lt;/p&gt;

&lt;p&gt;An executive will ask a question and the data team won't be able to answer it.&lt;/p&gt;

&lt;p&gt;The ML team will invest 500k into a model only to figure out all of the data was wrong or from a table, which no one supports anymore.&lt;/p&gt;

&lt;p&gt;And&lt;a href="https://seattledataguy.substack.com/p/why-are-we-still-struggling-to-answer" rel="noopener noreferrer"&gt; we will just add to the reasons that data teams fail&lt;/a&gt;.&lt;/p&gt;

&lt;p&gt;All of which will expose the true cost of moving fast.&lt;/p&gt;

</description>
      <category>database</category>
      <category>datascience</category>
      <category>watercooler</category>
    </item>
    <item>
      <title>How To Start Your Next Data Engineering Project</title>
      <dc:creator>SeattleDataGuy</dc:creator>
      <pubDate>Sat, 16 Apr 2022 21:08:30 +0000</pubDate>
      <link>https://dev.to/seattledataguy/how-to-start-your-next-data-engineering-project-p7c</link>
      <guid>https://dev.to/seattledataguy/how-to-start-your-next-data-engineering-project-p7c</guid>
      <description>&lt;p&gt;Photo by &lt;a href="https://unsplash.com/@sigmund?utm_source=medium&amp;amp;utm_medium=referral" rel="noopener noreferrer"&gt;Sigmund&lt;/a&gt; on &lt;a href="https://unsplash.com/?utm_source=medium&amp;amp;utm_medium=referral" rel="noopener noreferrer"&gt;Unsplash&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;Many programmers who are just starting out struggle with starting new data engineering projects. In our recent poll on YouTube, most viewers admitted that they have the most difficulty with even starting a data engineering project. The most common reasons noted in the poll were:&lt;/p&gt;

&lt;ol&gt;
&lt;li&gt; Finding the right data sets for my project&lt;/li&gt;
&lt;li&gt; Which tools should you use?&lt;/li&gt;
&lt;li&gt; What do I do with the data once I have it?&lt;/li&gt;
&lt;/ol&gt;

&lt;p&gt;Let's talk about each of these points, starting with the array of tools you have at your disposal.&lt;/p&gt;

&lt;h1&gt;
  
  
  Picking the Right Tools and Skills
&lt;/h1&gt;

&lt;p&gt;For the first part of this project I am going to borrow from &lt;a href="https://www.youtube.com/watch?v=XYKuslcJp7A" rel="noopener noreferrer"&gt;Thu Vu's advice&lt;/a&gt; for starting a data analytics project.&lt;/p&gt;

&lt;p&gt;Why not look at a data engineering job description on a head-hunting site and figure out what tools people are asking for? Nothing could be easier. We pulled up a job description from Smart Sheets Data Engineering. It is easy to see the exact tools and skills they are looking for and match them to the skills you can offer.&lt;/p&gt;

&lt;p&gt;Below are several key tools that you will want to include.&lt;/p&gt;

&lt;p&gt;&lt;a href="https://media.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fmiro.medium.com%2Fmax%2F1400%2F0%2A7xlLzKWa2rzlTZwH.png" class="article-body-image-wrapper"&gt;&lt;img src="https://media.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fmiro.medium.com%2Fmax%2F1400%2F0%2A7xlLzKWa2rzlTZwH.png"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;h1&gt;
  
  
  Cloud Platforms
&lt;/h1&gt;

&lt;p&gt;One of the first things you'll need for your &lt;a href="https://www.youtube.com/watch?v=385mKftVr3I&amp;amp;" rel="noopener noreferrer"&gt;data engineering project&lt;/a&gt; a cloud platform like Amazon Web Services (AWS). We always recommend that beginning programmers learn cloud platforms early. A lot of a data engineer's work is not completed on-premise(anymore). We do almost all work in the cloud.&lt;/p&gt;

&lt;p&gt;So, pick a platform you prefer: Google Cloud Platform (GCP), AWS, or Microsoft Azure. It's not that important what platform you pick out of these three; it's important that you pick one of these three because people understand that if you have used one, it is likely you can easily pick up another.&lt;/p&gt;

&lt;h1&gt;
  
  
  Workflow Management Platforms And Data Storage Analytic Systems
&lt;/h1&gt;

&lt;p&gt;In addition to picking a cloud provider, you will want to pick a tool to manage your automated workflows and a tool to manage the data itself. From the job description Airflow and Snowflake were both referenced.&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;  &lt;a href="https://www.youtube.com/watch?v=eTtrsGc4Wd4" rel="noopener noreferrer"&gt;Airflow --- It also would be a great idea to pick a cloud managed service like MWAA or Cloud Composer&lt;/a&gt;
&lt;/li&gt;
&lt;li&gt;  &lt;a href="https://seattledataguy.substack.com/p/snowflake-vs-bigquery-two-cloud-data?s=w" rel="noopener noreferrer"&gt;Snowflake --- Of course BigQuery is another solid choice&lt;/a&gt;
&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;These are not the only choices, by any stretch of the imagination. Other popular options for orchestration are &lt;a href="https://dagster.io/" rel="noopener noreferrer"&gt;&lt;em&gt;Dagster&lt;/em&gt;&lt;/a&gt;&lt;em&gt;, *and &lt;a href="https://www.prefect.io/" rel="noopener noreferrer"&gt;&lt;em&gt;Prefect&lt;/em&gt;&lt;/a&gt;. We actually recommend starting with *Airflow&lt;/em&gt; and then looking to others as you get more familiar with the processes.&lt;/p&gt;

&lt;p&gt;Don't be concerned with learning all these tools at first; keep it simple. Just concentrate on one or two, since this is not our primary focus.&lt;/p&gt;

&lt;h1&gt;
  
  
  Picking Data Sets
&lt;/h1&gt;

&lt;p&gt;Data sets come in every color of the rainbow. Your focus shouldn't be to work with already processed data; your attention should center on developing a data pipeline and finding raw data sources. As the &lt;a href="https://www.theseattledataguy.com/is-data-engineering-for-you-maybe-youre-a-recovering-data-scientist/" rel="noopener noreferrer"&gt;data engineer,&lt;/a&gt; you will need to know how to set up a data pipeline and pull the data you need from a raw source. Some &lt;a href="https://www.youtube.com/watch?v=LJkVvNWlO0g" rel="noopener noreferrer"&gt;great sources of raw data sets&lt;/a&gt; are:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;  &lt;a href="https://openskynetwork.github.io/opensky-api/" rel="noopener noreferrer"&gt;OpenSky&lt;/a&gt;: provides information on where flights currently are and where they are going, and much, much more.&lt;/li&gt;
&lt;li&gt;  &lt;a href="https://www.space-track.org/" rel="noopener noreferrer"&gt;Spacetrack&lt;/a&gt;: is a tracking system that will track, identify, and catalog all artificial satellites orbiting the Earth.&lt;/li&gt;
&lt;li&gt;  Other data sets: &lt;a href="https://catalog.ldc.upenn.edu/LDC2008T19" rel="noopener noreferrer"&gt;New York Times Annotated Corpus&lt;/a&gt;, &lt;a href="https://guides.loc.gov/datasets/repositories" rel="noopener noreferrer"&gt;The Library of Congress Dataset Repository&lt;/a&gt;, the &lt;a href="https://data.gov/" rel="noopener noreferrer"&gt;U.S. government data sets&lt;/a&gt;, and these &lt;a href="https://www.tableau.com/learn/articles/free-public-data-sets" rel="noopener noreferrer"&gt;free public data sets&lt;/a&gt; at Tableau.&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;Pulling data involves a lot of work, so when you are picking your data sets, you will probably find an API to help you extract the data into a comma-separated value (CSV) or parquet file to load into your data warehouse.&lt;/p&gt;

&lt;p&gt;Now you know how to find your data sets and manipulate them. So, what do you do with all this raw data that you've got?&lt;/p&gt;

&lt;h1&gt;
  
  
  Visualize Your Data
&lt;/h1&gt;

&lt;p&gt;An easy way to display your data engineering work is to create dashboards with metrics. Even if you won't be building too many dashboards in the future, you will want to create some type of final project.&lt;/p&gt;

&lt;p&gt;Dashboards are an easy way to do so.&lt;/p&gt;

&lt;p&gt;Here are a few tools you can look into:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;  &lt;a href="https://www.tableau.com/" rel="noopener noreferrer"&gt;Tableau&lt;/a&gt;
&lt;/li&gt;
&lt;li&gt;  &lt;a href="https://powerbi.microsoft.com/en-us/landing/free-account/?&amp;amp;ef_id=CjwKCAjwo8-SBhAlEiwAopc9W9kbHliQ_kOFZcScsos7JNyPDooe2wulwb1L__QfHuuPxgmkZbJKNBoCtboQAvD_BwE:G:s&amp;amp;OCID=AID2202141_SEM_CjwKCAjwo8-SBhAlEiwAopc9W9kbHliQ_kOFZcScsos7JNyPDooe2wulwb1L__QfHuuPxgmkZbJKNBoCtboQAvD_BwE:G:s&amp;amp;gclid=CjwKCAjwo8-SBhAlEiwAopc9W9kbHliQ_kOFZcScsos7JNyPDooe2wulwb1L__QfHuuPxgmkZbJKNBoCtboQAvD_BwE" rel="noopener noreferrer"&gt;PowerBI&lt;/a&gt;
&lt;/li&gt;
&lt;li&gt;  &lt;a href="https://d3js.org/" rel="noopener noreferrer"&gt;D3.js&lt;/a&gt;
&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;With your data visualization tool selected you can now start to pick some metrics and questions you would like to track.&lt;/p&gt;

&lt;p&gt;Maybe you would like to know how many flights occur in a single day. To build on that, you may want to know destinations by flight, time, or length and distance of travel. Discern some basic metrics and compile a graph.&lt;/p&gt;

&lt;p&gt;Anything basic like this will help you get comfortable figuring out your question of "why?" This question needs to be answered before you begin your real project. Consider it a warm-up to get the juices flowing.&lt;/p&gt;

&lt;p&gt;Now let's go over a fe w project ideas that you could try out.&lt;/p&gt;

&lt;h1&gt;
  
  
  3 Data Engineering Projects: Ideas
&lt;/h1&gt;

&lt;p&gt;&lt;em&gt;Beginning Data Engineering Projects Example&lt;/em&gt;: Consider using tools like &lt;a href="https://cloud.google.com/composer" rel="noopener noreferrer"&gt;Cloud Composer&lt;/a&gt; or &lt;a href="https://www.theseattledataguy.com/what-is-managed-workflows-for-apache-airflow-on-aws-and-why-companies-should-migrate-to-it/#page-content" rel="noopener noreferrer"&gt;Amazon Managed Workflows for Apache Airflow (MWAA)&lt;/a&gt;. These tools let you circumvent setting up &lt;em&gt;Airflow&lt;/em&gt; from scratch, which allows you more time to learn the functions of &lt;em&gt;Airflow&lt;/em&gt; without the hassle of figuring out how to set it up. From there, use an API such as &lt;a href="https://www.predictit.org/" rel="noopener noreferrer"&gt;PredictIt&lt;/a&gt; to scrape the data and return it in eXtensible markup language (XML).&lt;/p&gt;

&lt;p&gt;Maybe you are looking for data on massive swings in trading over a day. You could create a model where you identify certain patterns or swings over the day. If you created a Twitter bot and posted about these swings day after day, some traders would definitely see the value in your posts and data.&lt;/p&gt;

&lt;p&gt;If you wanted to upgrade that idea, track down articles relating to that swing for discussion and post those. There is definite value in that data, and it is a pretty simple thing to do. You are just using a Cloud Composer to ingest the data and storing it in a data warehouse like &lt;a href="https://cloud.google.com/bigquery" rel="noopener noreferrer"&gt;BigQuery&lt;/a&gt; or &lt;em&gt;Snowflake&lt;/em&gt;, creating a Twitter bot to post outputs to Twitter using something like Airflow.&lt;/p&gt;

&lt;p&gt;It is a fun and simple project because you do not have to reinvent the wheel or invest time in setting up infrastructure.&lt;/p&gt;

&lt;p&gt;&lt;em&gt;Intermediate Example&lt;/em&gt;: This &lt;a href="https://www.startdataengineering.com/post/data-engineering-project-for-beginners-batch-edition/" rel="noopener noreferrer"&gt;data engineering project&lt;/a&gt; is brought to us by Start Data Engineering(SDE). While they seem to reference just a basic CSV file about movie reviews, a better option might be to go to the &lt;a href="https://developer.nytimes.com/" rel="noopener noreferrer"&gt;New York Times Developers Portal&lt;/a&gt; and use their API to pull live movie reviews. Use SDE's framework and customize it for your own needs.&lt;/p&gt;

&lt;p&gt;SDE has done a superior job of breaking this project down like a recipe. They tell you exactly what tools you need and when they are going to be required for your project. They list out the prerequisites you need:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;  &lt;a href="https://www.docker.com/" rel="noopener noreferrer"&gt;Docker&lt;/a&gt;
&lt;/li&gt;
&lt;li&gt;  &lt;a href="http://aws.amazon.com/" rel="noopener noreferrer"&gt;Amazon Web Services (AWS) account&lt;/a&gt;
&lt;/li&gt;
&lt;li&gt;  &lt;a href="https://airflow.apache.org/" rel="noopener noreferrer"&gt;Airflow&lt;/a&gt;
&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;In this example, SDE shows you how to set up Apache Airflow from scratch rather than using the presets. You will also be exposed to tools such as:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;  &lt;a href="https://aws.amazon.com/s3/" rel="noopener noreferrer"&gt;Amazon S3&lt;/a&gt;
&lt;/li&gt;
&lt;li&gt;  &lt;a href="https://aws.amazon.com/redshift/" rel="noopener noreferrer"&gt;Amazon Redshift&lt;/a&gt;
&lt;/li&gt;
&lt;li&gt;  &lt;a href="https://theiam.org/" rel="noopener noreferrer"&gt;IAM&lt;/a&gt;
&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;There are many components offered, so when you are out there catching the attention of potential future employers, this project will help you detail the in-demand skills employers are looking for.&lt;/p&gt;

&lt;p&gt;&lt;em&gt;Advanced Example&lt;/em&gt;: For our advanced example, we will use mostly open-source tools. &lt;a href="https://www.sspaeti.com/blog/data-engineering-project-in-twenty-minutes/" rel="noopener noreferrer"&gt;Sspaeti's website gives a lot &lt;/a&gt;of inspiration for fantastic projects. For this project, they use a kitchen-sink variety of every open-source tool imaginable, such as:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;  Amazon S3&lt;/li&gt;
&lt;li&gt;  &lt;a href="https://spark.apache.org/" rel="noopener noreferrer"&gt;Apache Spark&lt;/a&gt;
&lt;/li&gt;
&lt;li&gt;  &lt;a href="https://delta.io/" rel="noopener noreferrer"&gt;Delta Lake&lt;/a&gt;
&lt;/li&gt;
&lt;li&gt;  &lt;a href="https://druid.apache.org/" rel="noopener noreferrer"&gt;Apache Druid&lt;/a&gt;
&lt;/li&gt;
&lt;li&gt;  &lt;a href="https://dagster.io/" rel="noopener noreferrer"&gt;Dagster&lt;/a&gt;
&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;In this project, you will scrape real estate data from the actual site, so it's going to be a marriage between an API and a website. You will scrape the website for data and clean up the HTML. You will implement a change-data-capture (CDC), data science, and visualizations with supersets.&lt;/p&gt;

&lt;p&gt;This can be a basic framework from which you can expand your idea. These are challenges to make you push and stretch yourself and your boundaries. But, if you decide to build this, just pick a few tools and try it. Your project doesn't have to be exactly the same.&lt;/p&gt;

&lt;h1&gt;
  
  
  Now It's Your Turn
&lt;/h1&gt;

&lt;p&gt;If you are struggling to begin a project, hopefully, we've inspired you and given you the tools and direction in which to take that first step.&lt;/p&gt;

&lt;p&gt;Don't let hesitation or a lack of a clear premise keep you from starting. The first step is to commit to an idea, then execute it without delay, even if it is something as simple as a Twitter bot. That Twitter bot could be the inspiration for bigger and better ideas!&lt;/p&gt;

</description>
      <category>datascience</category>
      <category>database</category>
      <category>beginners</category>
      <category>programming</category>
    </item>
    <item>
      <title>Data Platforms Vs Data Warehouses Vs Data Virtualization And Federated SQL Engine</title>
      <dc:creator>SeattleDataGuy</dc:creator>
      <pubDate>Thu, 31 Mar 2022 03:26:09 +0000</pubDate>
      <link>https://dev.to/seattledataguy/data-platforms-vs-data-warehouses-vs-data-virtualization-and-federated-sql-engine-11i4</link>
      <guid>https://dev.to/seattledataguy/data-platforms-vs-data-warehouses-vs-data-virtualization-and-federated-sql-engine-11i4</guid>
      <description>&lt;p&gt;There is a lot of discussion on how you should implement your data stacks today. Should you bundle or unbundle, should you use a traditional data warehouse, lakes, lake houses, or mesh?&lt;/p&gt;

&lt;p&gt;Or a hybrid of all of them?&lt;/p&gt;

&lt;p&gt;With all these different ways you can put together your data stack, the question becomes where to even start? How can your team set up a successful data stack and set of processes that ensure you can build reliable reporting quickly?&lt;/p&gt;

&lt;p&gt;Well, it all starts with data. Data is the base layer for a company's analytics projects. Whether you're building data science models, standardized reporting, dynamic pricing API end-points, or KPIs, these are the appliances that will run on the electricity that is data.&lt;/p&gt;

&lt;p&gt;With all these different ways companies can set up their data, tools to help them set up their core data layer have risen to answer every problem. &lt;/p&gt;

&lt;p&gt;In this article, we will discuss some of the popular routes data solution providers are talking in terms of types of data layers.&lt;/p&gt;

&lt;h3&gt;
  
  
  Strategies for developing your data layer
&lt;/h3&gt;

&lt;p&gt;Looking back a little over a decade ago, there were only one or two ways that were considered best practices in terms of storing data for analytics. &lt;/p&gt;

&lt;p&gt;Often most companies would use some version of a standardized EDW. However, much of the standard best practices are being challenged. For example, data meshes have risen in popularity under the perception that companies can reduce the time from raw data to reportable data.&lt;/p&gt;

&lt;p&gt;Besides data meshes, there are also streaming data-warehouses such as Matarlized and Vectorized, virtual data layers, and federated SQL engines like &lt;a href="https://www.theseattledataguy.com/what-is-data-virtualization/#page-content"&gt;Denodo&lt;/a&gt;, &lt;a href="https://seattledataguy.substack.com/p/what-is-trino-and-why-is-it-great?s=w"&gt;Trino&lt;/a&gt;, and &lt;a href="https://varada.io/"&gt;Varada&lt;/a&gt;.&lt;/p&gt;

&lt;p&gt;Not to mention that concept of a data platform becoming more prominent as tools like Snowflake and Databricks fight for dominance in the data world.&lt;/p&gt;

&lt;h3&gt;
  
  
  Data Platforms
&lt;/h3&gt;

&lt;p&gt;When Snowflake first came out they described themselves as the Cloud's first data warehouse. But that was a decade ago. Now Snowflake needed to change with the times and they described themselves  as a Cloud Data Platform. Their data warehouse is still the core of their offerings. However, their recent purchase of Streamlit is a great example of how they are trying to morph into a data platform. They don't just want to own the storage and management, they want to own the value add as well.&lt;/p&gt;

&lt;p&gt;This is likely due to several large competitors coming into the market. Databricks being one of those major players.&lt;/p&gt;

&lt;p&gt;On the other side Databricks originated in academia and the open-source community, the company was founded in 2013. The original creators of Apache Spark™, Delta Lake, and MLflow wanted to create a unified, open platform for all your data.&lt;/p&gt;

&lt;p&gt;The goal of both of these tools is to go beyond just a data warehouse or data lakehouse. Instead, they want to be the platform everyone else will build their applications and data stacks off of. Essentially the Microsoft or iPhone of data. Everything builds on their platform layer. Whether it be internal company data products or third party solutions like Fivetran and Rudderstack.&lt;/p&gt;

&lt;p&gt;But neither Snowflake or Databricks would be anywhere without the classic data warehouse.&lt;/p&gt;

&lt;h3&gt;
  
  
  Data Warehouse
&lt;/h3&gt;

&lt;p&gt;Data warehouses remain a standard choice when picking how to store your company's data. In particular, companies like FireboltDB have decided to not get involved in all the new buzzwords and decided to make a stand as a data warehouse. &lt;/p&gt;

&lt;p&gt;In addition, there are plenty of other solutions that can be utilized in a data warehouse context. Some are more standard databases like, Postgres, others are possibly modern cloud data warehouses like BigQuery, Teradata, or Redshift.  &lt;/p&gt;

&lt;p&gt;In general, all of these systems have their nuances and some aren't necessarily a "data warehouse" but they are often built like them, and for now, none of them have visions to go much beyond the data warehouse. Data warehouses remain popular as a design choice due to the fact that we, as a community, have been building them forever. Most of us understand concepts such as fact and dimension tables as well as slowly changing dimensions. Even with most modern data warehouses taking a less strict approach to much of these design concepts they still play a role in developing many company's core data layers.&lt;/p&gt;

&lt;p&gt;But these are usually built as a more classic data warehouse (even if BigQuery isn't a fan of standard data warehouses). &lt;/p&gt;

&lt;h3&gt;
  
  
  Streaming Data Warehouses
&lt;/h3&gt;

&lt;p&gt;Still, another form of analytical data storage systems is utilizing streaming databases. For example, &lt;a href="https://materialize.com/"&gt;Materialize&lt;/a&gt;, is a SQL streaming database startup built on top of the open-source Timely Dataflow project. It allows users to ask questions of living, streaming data, connecting directly to existing event streaming infrastructure, like Kafka, and to client applications.&lt;/p&gt;

&lt;p&gt;Engineers can interact with Materialize using a standard PostgreSQL interface, enabling plug-and-play integration of existing tooling. From there data professionals can build data products such as operational dashboards without ETLs.&lt;/p&gt;

&lt;p&gt;In addition, &lt;a href="https://vectorized.io/"&gt;Vectorized&lt;/a&gt; and Rockset are two other similar streaming database systems that allow developers to connect directly to data sources and run queries on top of them.&lt;/p&gt;

&lt;p&gt;All of these options are attempting to approach data management and reporting in a very different way.&lt;/p&gt;

&lt;h3&gt;
  
  
  Data Virtualization/Federated SQL Engine
&lt;/h3&gt;

&lt;p&gt;Some tools are working to directly query data across multiple data sources. &lt;/p&gt;

&lt;p&gt;One such example is &lt;a href="https://varada.io/"&gt;Varada&lt;/a&gt;. &lt;a href="https://varada.io/blog/data-lake/data-lake-economics-indexing/"&gt;Varada indexes data in Trino&lt;/a&gt; in such a way that reduces the time the CPU is used for data scanning (via indexes) and frees the CPU up for other tasks like fetching the data extremely fast or dealing with concurrency. This allows SQL users to run various queries, whether across dimensions, facts as well as other types of joins and data lake analytics, on indexed data as federated data sources. SQL aggregations and grouping is accelerated using nanoblock indexes as well, resulting in highly effective SQL analytics. Paired with &lt;a href="https://varada.io/vcc/"&gt;Varada's Control Center &lt;/a&gt;data engineers and system admins can more easily monitor their Presto/Trino cluster's and make educated decisions on how to manage them.&lt;/p&gt;

&lt;p&gt;Another example is &lt;a href="https://www.theseattledataguy.com/what-is-data-virtualization/#page-content"&gt;Denodo&lt;/a&gt; which allows you to connect to various data sources from Denodo itself. This includes Oracle, SQL Server, Hbase, etc.&lt;/p&gt;

&lt;p&gt;It also uses VQL which stands for the virtual query language. This is similar to SQL and allows developers the ability to create views on top of all the databases that Denodo is connected to. This means you can connect different databases and tables across them and create new data sets.&lt;/p&gt;

&lt;p&gt;In a world that is pushing more towards a Data Mesh every day, tools such as Trino, Varada do make a lot of sense. Instead of having to constantly get access to new sources of data in different types of data warehouses why not just query from all the different siloed data warehouses. That's easier said than done.&lt;/p&gt;

&lt;h3&gt;
  
  
  What Will Be Your Core Data Layer 
&lt;/h3&gt;

&lt;p&gt;There are many different ways companies can set up one of their most important assets and how a company's data is set up is crucial to its success. &lt;/p&gt;

&lt;p&gt;Whether you pick a data warehouse or utilize tools like Trino, the end goal is to create a robust data system that all end-users can trust.&lt;/p&gt;

&lt;p&gt;Of course, there are many challenges with every choice. If you decide to build a data warehouse, it'll take time, or if you pick a virtual data layer, then there can be a lot of confusion in terms of what data you should pull from.&lt;/p&gt;

&lt;p&gt;Why is this all-important? Setting up an easy-to-use data layer is crucial for businesses to succeed. So do take your time and consider which data tool is right for you. In the end, it's about creating a sustainable, robust, and scalable data system that is cost-effective.&lt;/p&gt;

</description>
      <category>sql</category>
      <category>database</category>
      <category>datascience</category>
      <category>analytics</category>
    </item>
    <item>
      <title>16 Data Analytics Buzzwords You Need To Know</title>
      <dc:creator>SeattleDataGuy</dc:creator>
      <pubDate>Mon, 14 Mar 2022 21:28:31 +0000</pubDate>
      <link>https://dev.to/seattledataguy/16-data-analytics-buzzwords-you-need-to-know-182</link>
      <guid>https://dev.to/seattledataguy/16-data-analytics-buzzwords-you-need-to-know-182</guid>
      <description>&lt;p&gt;Photo by &lt;a href="https://unsplash.com/@rvignes?utm_source=medium&amp;amp;utm_medium=referral"&gt;Romain Vignes&lt;/a&gt; on &lt;a href="https://unsplash.com/?utm_source=medium&amp;amp;utm_medium=referral"&gt;Unsplash&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;Analytics Engineer&lt;/strong&gt; = A BI Engineer who uses dbt&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;Data Mesh&lt;/strong&gt; = We tried an EDW and it was taking too long so we went back to the siloed data approach but now we opened it up to the entire company&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;Data Quality Monitoring&lt;/strong&gt; = A bunch of automated SQL statements&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;Low Code/No Code&lt;/strong&gt; = Click, click, click, shoot how do I configure the underlying environment. Forget it, I am going back to Python&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;Self Service Analytics&lt;/strong&gt; = Self-service after a data engineer or BIE spends 3 hours writing the query&lt;/p&gt;

&lt;p&gt;&lt;a href="https://seattledataguy.substack.com/p/the-baseline-datastack-going-beyond?utm_source=url"&gt;Modern Data Stack &lt;/a&gt;= Everything is SQL now, No more excel&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;Source Of Truth&lt;/strong&gt; = I used to trust this, until I learned about Truth V2&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;Reverse ETL&lt;/strong&gt; = LTE&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;Data Lineage&lt;/strong&gt; = Some unmaintained tool that was accurate once&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;Data Warehouse&lt;/strong&gt; = it's just an analytical database where you can drop all your data randomly and maybe spend some years later to try to model them&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;BI For BI Teams&lt;/strong&gt; = Analyzing your analysis tools&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;Democratizing Data&lt;/strong&gt; = Let's give everyone access and see how many competing narratives we can get from the same data&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;Data Observability&lt;/strong&gt; = Oh look we can parse Snowflake logs&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;Data Scientist&lt;/strong&gt; = a Schrodinger's cat of: an analyst, or a Stats PhD&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;Lakehouse&lt;/strong&gt; = We couldn't setup a Data Lake so we salvaged whatever we could and told our users it looks now like the previous DWH.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;Data Lake&lt;/strong&gt; = File server we just dump all our files in.&lt;/p&gt;

&lt;p&gt;Special thanks to &lt;a href="https://medium.com/u/d9259b41277a?source=post_page-----b6eb6d925ee4-----------------------------------"&gt;mehdio&lt;/a&gt;, &lt;a href="https://medium.com/u/a023a229ec8d?source=post_page-----b6eb6d925ee4-----------------------------------"&gt;Galen B&lt;/a&gt;, &lt;a href="https://www.linkedin.com/in/ethanaaron/"&gt;Ethan&lt;/a&gt;, &lt;a href="https://www.linkedin.com/in/laurenbalik/"&gt;Lauren&lt;/a&gt; and several others who all played a vital role in crystallizing these concepts.&lt;/p&gt;

&lt;p&gt;Read/Watch More Videos On Data And Data Analytics Below:&lt;/p&gt;

&lt;p&gt;&lt;a href="https://dev.to/seattledataguy/3-sql-interview-tips-for-data-scientists-and-data-engineers-5hm6"&gt;3 SQL Interview Tips For Data Scientists And Data Engineers&lt;br&gt;
&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;&lt;a href="https://dev.to/seattledataguy/5-tips-on-how-to-become-a-better-data-analyst-4geg"&gt;5 Tips On How To Become a Better Data Analyst&lt;br&gt;
&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;&lt;a href="https://www.youtube.com/watch?v=qgSqLh_mhXE"&gt;How To Set Up A Successful Data Analytics Team&lt;/a&gt;&lt;/p&gt;

</description>
      <category>database</category>
      <category>datascience</category>
      <category>analytics</category>
      <category>sql</category>
    </item>
    <item>
      <title>3 SQL Interview Tips For Data Scientists And Data Engineers</title>
      <dc:creator>SeattleDataGuy</dc:creator>
      <pubDate>Wed, 16 Feb 2022 20:19:20 +0000</pubDate>
      <link>https://dev.to/seattledataguy/3-sql-interview-tips-for-data-scientists-and-data-engineers-5hm6</link>
      <guid>https://dev.to/seattledataguy/3-sql-interview-tips-for-data-scientists-and-data-engineers-5hm6</guid>
      <description>&lt;p&gt;Photo by &lt;a href="https://unsplash.com/@elisa_ventur?utm_source=medium&amp;amp;utm_medium=referral" rel="noopener noreferrer"&gt;Elisa Ventur&lt;/a&gt; on &lt;a href="https://unsplash.com/?utm_source=medium&amp;amp;utm_medium=referral" rel="noopener noreferrer"&gt;Unsplash&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;SQL has become a common skill requirement across industries and job profiles over the last decade.&lt;/p&gt;

&lt;p&gt;Companies like Amazon and Google will often demand that their data analysts, data scientists, and product managers are at least be familiar with SQL. This is because SQL remains the language of data.&lt;/p&gt;

&lt;p&gt;This has lead to SQL being a staple part of many data professionals interview loops. Meaning you better plan to include SQL in your study sessions. That being said, just walking through SQL problems doesn't always cut.&lt;/p&gt;

&lt;p&gt;In this article I will go over three tips to help you improve your SQL both for interviews and in general.&lt;/p&gt;

&lt;h2&gt;
  
  
  Attempt To Solve A SQL Problem In Multiple Ways
&lt;/h2&gt;

&lt;p&gt;There is more than one way to solve most SQL problems.&lt;/p&gt;

&lt;p&gt;But often we resort to the methods we have most recently used.&lt;/p&gt;

&lt;p&gt;For example, if you just learned about analytical functions it can pigeonhole a lot of your future solutions. This is why it is important to try to solve SQL problems in multiple ways.&lt;/p&gt;

&lt;p&gt;In order to do all of our problems I will be using problem sets from InterviewQuery which is a site that can be used by data scientists to practice much more than SQL.&lt;/p&gt;

&lt;p&gt;But for now let's look at a simple problem they offer.&lt;/p&gt;

&lt;blockquote&gt;
&lt;p&gt;We're given two tables, a &lt;code&gt;users&lt;/code&gt; table with demographic information and the neighborhood they live in and a &lt;code&gt;neighborhoods&lt;/code&gt; table.&lt;/p&gt;

&lt;p&gt;Write a query that returns all neighborhoods that have 0 users.&lt;/p&gt;
&lt;/blockquote&gt;

&lt;p&gt;The tables we have are listed below.&lt;/p&gt;

&lt;p&gt;&lt;a href="https://media.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fmiro.medium.com%2Fmax%2F1400%2F1%2AK9mmbdNkKeurvmS85jOaXQ.png" class="article-body-image-wrapper"&gt;&lt;img src="https://media.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fmiro.medium.com%2Fmax%2F1400%2F1%2AK9mmbdNkKeurvmS85jOaXQ.png" alt="sql table for interviews"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;Here we have two input tables and an expected output. Before reading further do you have any thoughts on how to get neighborhood names that have 0 users?&lt;/p&gt;

&lt;p&gt;What is your first answer?&lt;/p&gt;

&lt;p&gt;Regardless, et's look at this first example of a possible solution.&lt;br&gt;
&lt;/p&gt;

&lt;div class="highlight js-code-highlight"&gt;
&lt;pre class="highlight sql"&gt;&lt;code&gt;
&lt;span class="k"&gt;SELECT&lt;/span&gt; &lt;span class="n"&gt;N&lt;/span&gt;&lt;span class="p"&gt;.&lt;/span&gt;&lt;span class="n"&gt;name&lt;/span&gt;
&lt;span class="k"&gt;FROM&lt;/span&gt; &lt;span class="n"&gt;neighborhoods&lt;/span&gt; &lt;span class="n"&gt;N&lt;/span&gt; 
&lt;span class="k"&gt;LEFT&lt;/span&gt; &lt;span class="k"&gt;JOIN&lt;/span&gt; &lt;span class="n"&gt;users&lt;/span&gt; &lt;span class="n"&gt;U&lt;/span&gt; 
&lt;span class="k"&gt;ON&lt;/span&gt; &lt;span class="n"&gt;N&lt;/span&gt;&lt;span class="p"&gt;.&lt;/span&gt;&lt;span class="n"&gt;id&lt;/span&gt; &lt;span class="o"&gt;=&lt;/span&gt; &lt;span class="n"&gt;U&lt;/span&gt;&lt;span class="p"&gt;.&lt;/span&gt;&lt;span class="n"&gt;neighborhood_id&lt;/span&gt;
&lt;span class="k"&gt;WHERE&lt;/span&gt; &lt;span class="n"&gt;U&lt;/span&gt;&lt;span class="p"&gt;.&lt;/span&gt;&lt;span class="n"&gt;id&lt;/span&gt; &lt;span class="k"&gt;is&lt;/span&gt; &lt;span class="k"&gt;null&lt;/span&gt; 
&lt;/code&gt;&lt;/pre&gt;

&lt;/div&gt;



&lt;p&gt;This solution relies on the LEFT JOIN to return any neighborhoods without users having a u.id that is NULL.&lt;/p&gt;

&lt;p&gt;But this example can be a little confusing and not as readable for some as it's not explicit. The LEFT JOIN kind of makes it a little difficult to read. The reason for this is that you will have to mentally manage the logic in your head.&lt;/p&gt;

&lt;p&gt;Let's take a quick look at the table the LEFT JOIN would create "WITHOUT THE FILTER".&lt;/p&gt;

&lt;p&gt;&lt;a href="https://media.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fmiro.medium.com%2Fmax%2F1400%2F1%2AOl2R33T_OFrx8vu6C9rnAA.png" class="article-body-image-wrapper"&gt;&lt;img src="https://media.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fmiro.medium.com%2Fmax%2F1400%2F1%2AOl2R33T_OFrx8vu6C9rnAA.png" alt="sql for interviews"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;Above you will see that user_id has null values. This is because we are LEFT joining with the left table being neighborhood. Meaning that where there are no users, there will still be neighborhoods.&lt;/p&gt;

&lt;p&gt;But, this is just one way to solve this problem. Let's write this query differently.&lt;/p&gt;

&lt;p&gt;Have any ideas?&lt;/p&gt;

&lt;p&gt;You can check out the query below which is the same thing really but in a more explicit way.&lt;br&gt;
&lt;/p&gt;

&lt;div class="highlight js-code-highlight"&gt;
&lt;pre class="highlight sql"&gt;&lt;code&gt;&lt;span class="k"&gt;SELECT&lt;/span&gt;  &lt;span class="n"&gt;n&lt;/span&gt;&lt;span class="p"&gt;.&lt;/span&gt;&lt;span class="n"&gt;name&lt;/span&gt;
&lt;span class="k"&gt;FROM&lt;/span&gt; &lt;span class="n"&gt;neighborhoods&lt;/span&gt; &lt;span class="n"&gt;n&lt;/span&gt; 
&lt;span class="k"&gt;LEFT&lt;/span&gt; &lt;span class="k"&gt;JOIN&lt;/span&gt; &lt;span class="n"&gt;users&lt;/span&gt; &lt;span class="n"&gt;u&lt;/span&gt; 
&lt;span class="k"&gt;ON&lt;/span&gt; &lt;span class="n"&gt;n&lt;/span&gt;&lt;span class="p"&gt;.&lt;/span&gt;&lt;span class="n"&gt;id&lt;/span&gt; &lt;span class="o"&gt;=&lt;/span&gt; &lt;span class="n"&gt;u&lt;/span&gt;&lt;span class="p"&gt;.&lt;/span&gt;&lt;span class="n"&gt;neighborhood_id&lt;/span&gt;
&lt;span class="k"&gt;GROUP&lt;/span&gt; &lt;span class="k"&gt;BY&lt;/span&gt; &lt;span class="n"&gt;n&lt;/span&gt;&lt;span class="p"&gt;.&lt;/span&gt;&lt;span class="n"&gt;name&lt;/span&gt;
&lt;span class="k"&gt;HAVING&lt;/span&gt; &lt;span class="k"&gt;COUNT&lt;/span&gt;&lt;span class="p"&gt;(&lt;/span&gt;&lt;span class="n"&gt;u&lt;/span&gt;&lt;span class="p"&gt;.&lt;/span&gt;&lt;span class="n"&gt;id&lt;/span&gt;&lt;span class="p"&gt;)&lt;/span&gt; &lt;span class="o"&gt;=&lt;/span&gt; &lt;span class="mi"&gt;0&lt;/span&gt;
&lt;/code&gt;&lt;/pre&gt;

&lt;/div&gt;



&lt;p&gt;This query uses a more explicit HAVING clause to only count u.ids. COUNT only counts when u.id is not null. So in that way, the query truly operates the same way.&lt;/p&gt;

&lt;p&gt;So if there isn't a matching ID for neighborhood id in the user table, then it will replace the user-id with null. Thus, if I write where user id is NULL, I will get the same response as COUNT(u.id).&lt;/p&gt;

&lt;h2&gt;
  
  
  Break Down Questions And Logic
&lt;/h2&gt;

&lt;p&gt;Many questions you get will require you to break down the logic into multiple steps. A straightforward example is often when questions ask you to get the first instance of an event.&lt;/p&gt;

&lt;p&gt;InterviewQuery has a question like this where they ask:&lt;/p&gt;

&lt;blockquote&gt;
&lt;p&gt;Given a table of &lt;code&gt;song_plays&lt;/code&gt; and a table of &lt;code&gt;users&lt;/code&gt;, write a query to extract the earliest date each user played their third unique song.&lt;/p&gt;
&lt;/blockquote&gt;

&lt;p&gt;The table are listed below:&lt;/p&gt;

&lt;p&gt;&lt;a href="https://media.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fmiro.medium.com%2Fmax%2F1400%2F1%2Arh4ymLPlWgoKxKUUmkGrWg.png" class="article-body-image-wrapper"&gt;&lt;img src="https://media.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fmiro.medium.com%2Fmax%2F1400%2F1%2Arh4ymLPlWgoKxKUUmkGrWg.png" alt="sql interviews"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;Here, they have written out the fact that they want the third "unique" song. This should be a dead giveaway that there are multiple instances of song plays that would mess up your solution if you were just to implement a quick &lt;a href="https://www.sqlservertutorial.net/sql-server-window-functions/sql-server-row_number-function/" rel="noopener noreferrer"&gt;ROW_NUMBER&lt;/a&gt;() function.&lt;/p&gt;

&lt;p&gt;A great way to help simplify this is to start rotating some shapes.&lt;/p&gt;

&lt;p&gt;To start let's create a data set that only got a unique instance of every play.&lt;/p&gt;

&lt;p&gt;What is interesting here is I disagree with the solution provided by InterviewQuery, or at least I am not 100% behind it. To get a unique song and first play date you should be able to just use MIN() to get the first date played.&lt;/p&gt;

&lt;p&gt;You can do this with the query below.&lt;br&gt;
&lt;/p&gt;

&lt;div class="highlight js-code-highlight"&gt;
&lt;pre class="highlight sql"&gt;&lt;code&gt; &lt;span class="k"&gt;SELECT&lt;/span&gt; &lt;span class="n"&gt;song_name&lt;/span&gt;&lt;span class="p"&gt;,&lt;/span&gt; &lt;span class="n"&gt;user_id&lt;/span&gt;&lt;span class="p"&gt;,&lt;/span&gt; &lt;span class="k"&gt;min&lt;/span&gt;&lt;span class="p"&gt;(&lt;/span&gt;&lt;span class="n"&gt;created_at&lt;/span&gt;&lt;span class="p"&gt;)&lt;/span&gt; &lt;span class="n"&gt;created_at&lt;/span&gt;
 &lt;span class="k"&gt;FROM&lt;/span&gt; &lt;span class="n"&gt;song_plays&lt;/span&gt;
 &lt;span class="k"&gt;GROUP&lt;/span&gt; &lt;span class="k"&gt;BY&lt;/span&gt; &lt;span class="n"&gt;song_name&lt;/span&gt;&lt;span class="p"&gt;,&lt;/span&gt; &lt;span class="n"&gt;user_id&lt;/span&gt;
&lt;/code&gt;&lt;/pre&gt;

&lt;/div&gt;



&lt;p&gt;As you can see, all we need to do is get the minimum play date for the song id and user. This cleans up any of the messy repeated plays in a row. Making the data much easier to work with.&lt;/p&gt;

&lt;p&gt;InterviewQuery suggests using row_number() to get the first play of a song. I find this a little less clean as you will now be forced to do a second query to get the first value. Whereas with the MIN() method you have one query and you have shrunk the data set down.&lt;/p&gt;

&lt;p&gt;But as I referenced earlier, there is more than one way to get a lot of SQL answers.&lt;/p&gt;

&lt;p&gt;In no way do I hate the row_number() function. In fact, in the next step, we will be using it. You can see it in the query below.&lt;br&gt;
&lt;/p&gt;

&lt;div class="highlight js-code-highlight"&gt;
&lt;pre class="highlight sql"&gt;&lt;code&gt;&lt;span class="k"&gt;with&lt;/span&gt; &lt;span class="n"&gt;t1&lt;/span&gt; &lt;span class="k"&gt;as&lt;/span&gt; &lt;span class="p"&gt;(&lt;/span&gt;
    &lt;span class="k"&gt;SELECT&lt;/span&gt; &lt;span class="n"&gt;song_name&lt;/span&gt;&lt;span class="p"&gt;,&lt;/span&gt; &lt;span class="n"&gt;user_id&lt;/span&gt;&lt;span class="p"&gt;,&lt;/span&gt; &lt;span class="k"&gt;min&lt;/span&gt;&lt;span class="p"&gt;(&lt;/span&gt;&lt;span class="n"&gt;created_at&lt;/span&gt;&lt;span class="p"&gt;)&lt;/span&gt; &lt;span class="n"&gt;created_at&lt;/span&gt;
    &lt;span class="k"&gt;FROM&lt;/span&gt; &lt;span class="n"&gt;song_plays&lt;/span&gt;
    &lt;span class="k"&gt;GROUP&lt;/span&gt; &lt;span class="k"&gt;BY&lt;/span&gt; &lt;span class="n"&gt;song_name&lt;/span&gt;&lt;span class="p"&gt;,&lt;/span&gt; &lt;span class="n"&gt;user_id&lt;/span&gt;
&lt;span class="p"&gt;)&lt;/span&gt;
  &lt;span class="k"&gt;SELECT&lt;/span&gt; &lt;span class="n"&gt;u&lt;/span&gt;&lt;span class="p"&gt;.&lt;/span&gt;&lt;span class="n"&gt;name&lt;/span&gt;&lt;span class="p"&gt;,&lt;/span&gt; &lt;span class="n"&gt;t1&lt;/span&gt;&lt;span class="p"&gt;.&lt;/span&gt;&lt;span class="n"&gt;song_name&lt;/span&gt;&lt;span class="p"&gt;,&lt;/span&gt; &lt;span class="n"&gt;t1&lt;/span&gt;&lt;span class="p"&gt;.&lt;/span&gt;&lt;span class="n"&gt;created_at&lt;/span&gt;&lt;span class="p"&gt;,&lt;/span&gt;
  &lt;span class="n"&gt;row_number&lt;/span&gt;&lt;span class="p"&gt;()&lt;/span&gt; &lt;span class="n"&gt;over&lt;/span&gt; &lt;span class="p"&gt;(&lt;/span&gt;&lt;span class="k"&gt;partition&lt;/span&gt; &lt;span class="k"&gt;by&lt;/span&gt; &lt;span class="n"&gt;u&lt;/span&gt;&lt;span class="p"&gt;.&lt;/span&gt;&lt;span class="n"&gt;name&lt;/span&gt; &lt;span class="k"&gt;order&lt;/span&gt; &lt;span class="k"&gt;by&lt;/span&gt; &lt;span class="n"&gt;created_at&lt;/span&gt;&lt;span class="p"&gt;)&lt;/span&gt; &lt;span class="n"&gt;rn&lt;/span&gt; 
  &lt;span class="k"&gt;FROM&lt;/span&gt; &lt;span class="n"&gt;t1&lt;/span&gt; 
  &lt;span class="k"&gt;JOIN&lt;/span&gt; &lt;span class="n"&gt;users&lt;/span&gt; &lt;span class="n"&gt;u&lt;/span&gt; 
  &lt;span class="k"&gt;ON&lt;/span&gt; &lt;span class="n"&gt;t1&lt;/span&gt;&lt;span class="p"&gt;.&lt;/span&gt;&lt;span class="n"&gt;user_id&lt;/span&gt; &lt;span class="o"&gt;=&lt;/span&gt; &lt;span class="n"&gt;u&lt;/span&gt;&lt;span class="p"&gt;.&lt;/span&gt;&lt;span class="n"&gt;id&lt;/span&gt; 
&lt;/code&gt;&lt;/pre&gt;

&lt;/div&gt;



&lt;p&gt;From here, the rest should be self explanatory. We need to query for the 3rd row number. This is show below.&lt;br&gt;
&lt;/p&gt;

&lt;div class="highlight js-code-highlight"&gt;
&lt;pre class="highlight sql"&gt;&lt;code&gt;&lt;span class="k"&gt;with&lt;/span&gt; &lt;span class="n"&gt;t1&lt;/span&gt; &lt;span class="k"&gt;as&lt;/span&gt; &lt;span class="p"&gt;(&lt;/span&gt;
    &lt;span class="k"&gt;SELECT&lt;/span&gt; &lt;span class="n"&gt;song_name&lt;/span&gt;&lt;span class="p"&gt;,&lt;/span&gt; &lt;span class="n"&gt;user_id&lt;/span&gt;&lt;span class="p"&gt;,&lt;/span&gt; &lt;span class="k"&gt;min&lt;/span&gt;&lt;span class="p"&gt;(&lt;/span&gt;&lt;span class="n"&gt;created_at&lt;/span&gt;&lt;span class="p"&gt;)&lt;/span&gt; &lt;span class="n"&gt;created_at&lt;/span&gt;
    &lt;span class="k"&gt;FROM&lt;/span&gt; &lt;span class="n"&gt;song_plays&lt;/span&gt;
    &lt;span class="k"&gt;GROUP&lt;/span&gt; &lt;span class="k"&gt;BY&lt;/span&gt; &lt;span class="n"&gt;song_name&lt;/span&gt;&lt;span class="p"&gt;,&lt;/span&gt; &lt;span class="n"&gt;user_id&lt;/span&gt;
&lt;span class="p"&gt;),&lt;/span&gt;
&lt;span class="n"&gt;t2&lt;/span&gt; &lt;span class="k"&gt;as&lt;/span&gt;&lt;span class="p"&gt;(&lt;/span&gt;
    &lt;span class="k"&gt;SELECT&lt;/span&gt; &lt;span class="n"&gt;u&lt;/span&gt;&lt;span class="p"&gt;.&lt;/span&gt;&lt;span class="n"&gt;name&lt;/span&gt;&lt;span class="p"&gt;,&lt;/span&gt; &lt;span class="n"&gt;t1&lt;/span&gt;&lt;span class="p"&gt;.&lt;/span&gt;&lt;span class="n"&gt;song_name&lt;/span&gt;&lt;span class="p"&gt;,&lt;/span&gt; &lt;span class="n"&gt;t1&lt;/span&gt;&lt;span class="p"&gt;.&lt;/span&gt;&lt;span class="n"&gt;created_at&lt;/span&gt;&lt;span class="p"&gt;,&lt;/span&gt;
    &lt;span class="n"&gt;row_number&lt;/span&gt;&lt;span class="p"&gt;()&lt;/span&gt; &lt;span class="n"&gt;over&lt;/span&gt; &lt;span class="p"&gt;(&lt;/span&gt;&lt;span class="k"&gt;partition&lt;/span&gt; &lt;span class="k"&gt;by&lt;/span&gt; &lt;span class="n"&gt;u&lt;/span&gt;&lt;span class="p"&gt;.&lt;/span&gt;&lt;span class="n"&gt;name&lt;/span&gt; &lt;span class="k"&gt;order&lt;/span&gt; &lt;span class="k"&gt;by&lt;/span&gt; &lt;span class="n"&gt;created_at&lt;/span&gt;&lt;span class="p"&gt;)&lt;/span&gt; &lt;span class="n"&gt;rn&lt;/span&gt; 
    &lt;span class="k"&gt;FROM&lt;/span&gt; &lt;span class="n"&gt;t1&lt;/span&gt; 
    &lt;span class="k"&gt;JOIN&lt;/span&gt; &lt;span class="n"&gt;users&lt;/span&gt; &lt;span class="n"&gt;u&lt;/span&gt; 
    &lt;span class="k"&gt;ON&lt;/span&gt; &lt;span class="n"&gt;t1&lt;/span&gt;&lt;span class="p"&gt;.&lt;/span&gt;&lt;span class="n"&gt;user_id&lt;/span&gt; &lt;span class="o"&gt;=&lt;/span&gt; &lt;span class="n"&gt;u&lt;/span&gt;&lt;span class="p"&gt;.&lt;/span&gt;&lt;span class="n"&gt;id&lt;/span&gt; 
&lt;span class="p"&gt;)&lt;/span&gt;
&lt;span class="k"&gt;SELECT&lt;/span&gt; &lt;span class="n"&gt;name&lt;/span&gt;&lt;span class="p"&gt;,&lt;/span&gt; &lt;span class="n"&gt;song_name&lt;/span&gt;&lt;span class="p"&gt;,&lt;/span&gt; &lt;span class="n"&gt;created_at&lt;/span&gt;
&lt;span class="k"&gt;FROM&lt;/span&gt; &lt;span class="n"&gt;t2&lt;/span&gt; 
&lt;span class="k"&gt;WHERE&lt;/span&gt; &lt;span class="n"&gt;rn&lt;/span&gt; &lt;span class="o"&gt;=&lt;/span&gt; &lt;span class="mi"&gt;3&lt;/span&gt;
&lt;/code&gt;&lt;/pre&gt;

&lt;/div&gt;



&lt;p&gt;There is one more problem. This actually won't solve InterviewQuery's question.&lt;/p&gt;

&lt;p&gt;This brings up the other issue I have with InterviewQuery's current solution for this problem. InterviewQuery wants us to return all users even if they didn't have a third song play. This wasn't referenced in the question and I don't see the value. But these are all NITs.&lt;/p&gt;

&lt;p&gt;They have a different solution that uses a LEFT JOIN and then returns both users with and without a third song play. So do be aware of this finicky detail.&lt;/p&gt;

&lt;h2&gt;
  
  
  Learn How To Do None Standard Joins
&lt;/h2&gt;

&lt;p&gt;Many people are accustomed to straightforward joins. With a basic "=".&lt;/p&gt;

&lt;p&gt;This is great, but the truth is JOINs can be used like WHERE clauses.&lt;/p&gt;

&lt;p&gt;There was a point where JOINs didn't exist and WHERE clauses were used as JOINs. I still occasionally see people who JOIN using a long string of WHERE clauses instead because they either work currently with older systems or learned SQL at a time JOINs weren't commonplace(&lt;a href="https://learnsql.com/blog/history-of-sql-standards/" rel="noopener noreferrer"&gt;explicit JOINs were added in SQL-92&lt;/a&gt;).&lt;/p&gt;

&lt;p&gt;What this means is you can use JOINs in a lot of interesting ways. For example, if you have ever had to calculate a rolling sum without a window function, you will know you can use the "&amp;gt; and &amp;lt;" signs to create duplicate rows that you will then aggregate.&lt;/p&gt;

&lt;p&gt;In our final question we are asked:&lt;/p&gt;

&lt;blockquote&gt;
&lt;p&gt;Given a table of students and their SAT test scores, write a query to return the two students with the closest test scores with the score difference.&lt;/p&gt;

&lt;p&gt;If there are multiple students with the same minimum score difference, select the student name combination that is higher in the alphabet.&lt;/p&gt;
&lt;/blockquote&gt;

&lt;p&gt;This is what the table looks like:&lt;/p&gt;

&lt;p&gt;&lt;a href="https://media.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fmiro.medium.com%2Fmax%2F1400%2F1%2AoRtcD7d1avyzgFLbgt9RMw.png" class="article-body-image-wrapper"&gt;&lt;img src="https://media.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fmiro.medium.com%2Fmax%2F1400%2F1%2AoRtcD7d1avyzgFLbgt9RMw.png" alt="database interviews"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;Take a moment to figure out how you might tackle this question.&lt;/p&gt;

&lt;p&gt;Don't rush here. There are a lot of ways you could overcomplicate this question in particular.&lt;/p&gt;

&lt;p&gt;The answer is pretty short.&lt;/p&gt;

&lt;p&gt;But it all starts with you needing to create a combination of every student and every student's score.&lt;/p&gt;

&lt;p&gt;Let's look at a possible solution and break down why it works.&lt;br&gt;
&lt;/p&gt;

&lt;div class="highlight js-code-highlight"&gt;
&lt;pre class="highlight sql"&gt;&lt;code&gt;&lt;span class="k"&gt;SELECT&lt;/span&gt; 
&lt;span class="n"&gt;t1&lt;/span&gt;&lt;span class="p"&gt;.&lt;/span&gt;&lt;span class="n"&gt;student&lt;/span&gt; &lt;span class="k"&gt;as&lt;/span&gt; &lt;span class="n"&gt;one_student&lt;/span&gt;&lt;span class="p"&gt;,&lt;/span&gt;
&lt;span class="n"&gt;t2&lt;/span&gt;&lt;span class="p"&gt;.&lt;/span&gt;&lt;span class="n"&gt;student&lt;/span&gt; &lt;span class="k"&gt;as&lt;/span&gt; &lt;span class="n"&gt;other_student&lt;/span&gt;&lt;span class="p"&gt;,&lt;/span&gt;
&lt;span class="k"&gt;abs&lt;/span&gt;&lt;span class="p"&gt;(&lt;/span&gt;&lt;span class="n"&gt;t1&lt;/span&gt;&lt;span class="p"&gt;.&lt;/span&gt;&lt;span class="n"&gt;score&lt;/span&gt; &lt;span class="o"&gt;-&lt;/span&gt; &lt;span class="n"&gt;t2&lt;/span&gt;&lt;span class="p"&gt;.&lt;/span&gt;&lt;span class="n"&gt;score&lt;/span&gt;&lt;span class="p"&gt;)&lt;/span&gt; &lt;span class="k"&gt;as&lt;/span&gt; &lt;span class="n"&gt;score_diff&lt;/span&gt;
&lt;span class="k"&gt;FROM&lt;/span&gt; &lt;span class="n"&gt;scores&lt;/span&gt; &lt;span class="n"&gt;t1&lt;/span&gt;
&lt;span class="k"&gt;JOIN&lt;/span&gt; &lt;span class="n"&gt;scores&lt;/span&gt; &lt;span class="n"&gt;t2&lt;/span&gt; 
   &lt;span class="k"&gt;ON&lt;/span&gt;  &lt;span class="n"&gt;t1&lt;/span&gt;&lt;span class="p"&gt;.&lt;/span&gt;&lt;span class="n"&gt;id&lt;/span&gt; &lt;span class="o"&gt;&amp;lt;&lt;/span&gt; &lt;span class="n"&gt;t2&lt;/span&gt;&lt;span class="p"&gt;.&lt;/span&gt;&lt;span class="n"&gt;id&lt;/span&gt;
&lt;span class="k"&gt;ORDER&lt;/span&gt; &lt;span class="k"&gt;BY&lt;/span&gt; &lt;span class="mi"&gt;3&lt;/span&gt;&lt;span class="p"&gt;,&lt;/span&gt;&lt;span class="mi"&gt;1&lt;/span&gt;&lt;span class="p"&gt;,&lt;/span&gt;&lt;span class="mi"&gt;2&lt;/span&gt; 
&lt;span class="k"&gt;limit&lt;/span&gt; &lt;span class="mi"&gt;1&lt;/span&gt;
&lt;/code&gt;&lt;/pre&gt;

&lt;/div&gt;



&lt;p&gt;So why does this work? We are telling the query to do a self join in all cases where the ID is smaller than the ID we are joining on.&lt;/p&gt;

&lt;p&gt;Meaning we are creating a table that looks like the table below:&lt;/p&gt;

&lt;p&gt;&lt;a href="https://media.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fmiro.medium.com%2Fmax%2F1400%2F1%2ABTNp7r4FM2jxb7lGQ7RfDQ.png" class="article-body-image-wrapper"&gt;&lt;img src="https://media.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fmiro.medium.com%2Fmax%2F1400%2F1%2ABTNp7r4FM2jxb7lGQ7RfDQ.png" alt="coding interviews"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;Now that we have a combination of all the possible student IDs we can just find the ABS() score different and properly set the order.&lt;/p&gt;

&lt;p&gt;That's it.&lt;/p&gt;

&lt;p&gt;One final point I will add is that I did end up eventually deciding to use the "!=" vs "&amp;lt;". The less than only seemed to work with the specific set of names and I am just not 100% sure if the names were changed that the ordering would work as expected. Using the "!=" guarantees that we will test every combination of names in every order. Meaning that if the more "alphabetic" solution were actually in a combination of "3--2" vs "2--3" we check and order both sets.&lt;/p&gt;

&lt;h2&gt;
  
  
  How Will You Take Your SQL To The Next Level
&lt;/h2&gt;

&lt;p&gt;SQL looks like it is here to stay. It seems if anything it is picking up speed. The challenge is how do you take your SQL to the next level. It's not just about learning new syntax.&lt;/p&gt;

&lt;p&gt;It's about learning how the data you're working on can be manipulated using even simple clauses. In doing so you will be creating SQL at a much higher level and better prepare you for your &lt;a href="https://www.youtube.com/watch?v=W2L7Q3J-ei0" rel="noopener noreferrer"&gt;SQL interviews&lt;/a&gt;.&lt;/p&gt;

&lt;p&gt;Whether you're a &lt;a href="https://medium.com/p/c3824cb76c2e" rel="noopener noreferrer"&gt;data scientists&lt;/a&gt;,&lt;a href="https://www.youtube.com/watch?v=PEsCumjfBW0" rel="noopener noreferrer"&gt; data engineer&lt;/a&gt; or &lt;a href="https://medium.com/coriers/5-tips-on-how-to-become-a-better-data-analyst-30f226c09c39" rel="noopener noreferrer"&gt;analysts&lt;/a&gt; I wish you the best of luck!&lt;/p&gt;

&lt;p&gt;If you enjoyed this article, then check out some of my other content:&lt;/p&gt;

&lt;p&gt;&lt;a href="https://seattledataguy.substack.com/p/which-managed-version-of-airflow" rel="noopener noreferrer"&gt;Which Managed Version Of Airflow Should You Use?&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;&lt;a href="https://www.youtube.com/watch?v=wuiJWSz-OLQ" rel="noopener noreferrer"&gt;5 Tips To Become A Great Analyst&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;&lt;a href="https://www.theseattledataguy.com/what-is-trino-and-why-is-it-great-at-processing-big-data/#page-content" rel="noopener noreferrer"&gt;What Is Trino And How It Manages Big Data&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;&lt;a href="https://www.youtube.com/watch?v=bqCXVpRqTpE" rel="noopener noreferrer"&gt;What I Learned From 100+ Data Engineering Interviews - Interview Tips&lt;/a&gt;&lt;/p&gt;

</description>
      <category>sql</category>
      <category>database</category>
      <category>datascience</category>
      <category>analytics</category>
    </item>
    <item>
      <title>5 Tips On How To Become a Better Data Analyst</title>
      <dc:creator>SeattleDataGuy</dc:creator>
      <pubDate>Sat, 12 Feb 2022 18:36:10 +0000</pubDate>
      <link>https://dev.to/seattledataguy/5-tips-on-how-to-become-a-better-data-analyst-4geg</link>
      <guid>https://dev.to/seattledataguy/5-tips-on-how-to-become-a-better-data-analyst-4geg</guid>
      <description>&lt;p&gt;&lt;a href="https://www.franklin.edu/blog/how-become-a-data-analyst"&gt;Image Source&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;Becoming a data analyst is one of the most common early career moves for people educated in STEM fields. Not to mention according to the &lt;a href="https://www.franklin.edu/blog/how-become-a-data-analyst"&gt;Bureau of Labor Statistics&lt;/a&gt; the demand for data analysts will grow about 20% over the next few years.&lt;/p&gt;

&lt;p&gt;Working as an analyst allows you to put skills learned in college to work while also giving you an opportunity to develop new skills on the job. Average salaries ranging from $60,000 to $80,000 also make data analyst positions lucrative starting points for younger professionals.&lt;/p&gt;

&lt;p&gt;As companies have become more data-driven, the skills that go into working as an analyst have grown more technical. Many analysts today are skilled in programming languages like Python and R that are suitable for processing large data sets. Before developing these highly technical skills, however, there are some basic &lt;a href="https://www.youtube.com/watch?v=wuiJWSz-OLQ"&gt;tips and tricks that all data analysts should learn&lt;/a&gt;. Here are some of the fundamentals you should focus on in order to become a better data analyst or data engineer.&lt;/p&gt;

&lt;h1&gt;
  
  
  1. Set Up a Clear Data Analytics Process
&lt;/h1&gt;

&lt;p&gt;One of the fundamental parts of becoming a successful data analyst is to have a clear process set up for your projects. This will save you the time and trouble of approaching each project in an ad hoc manner. A simple data analytics process is outlined below:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;  Define the Question: Fully define the question you're trying to answer and the goals of your data analytics project.&lt;/li&gt;
&lt;li&gt;  Collect Data: Work with data engineers or other data professionals to gather relevant data for your project.&lt;/li&gt;
&lt;li&gt;  Clean the Data: Standardize the data you've collected and remove any incorrect or irrelevant entries.&lt;/li&gt;
&lt;li&gt;  Analyze the Data: Employ data analysis techniques to understand the data and drive answers to your question. This step can take many different forms, depending on the question you're trying to answer.&lt;/li&gt;
&lt;li&gt;  Share Your Results: Create data visualizations and resources that will help others understand the insights you've produced.&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;With this simple framework, you'll have a clear road map for outlining and completing data analytics projects. Following this basic process will also keep you from getting sidetracked as you conduct your analysis.&lt;/p&gt;

&lt;h1&gt;
  
  
  2. Don't Bury the Lede
&lt;/h1&gt;

&lt;p&gt;When your analysis is finished, it's important that you're able to communicate your findings to others in an effective way. A key part of this is to keep your reports simple and concise. While it may be tempting to show all of your findings, it's better to condense your results down to a simple, understandable message.&lt;/p&gt;

&lt;p&gt;For optimal communication, consider telling the story of your data with a few carefully selected charts. These should be relevant to the core question and easy for your audience to understand. Sum up your findings with a conclusion that answers the question and drives value for your audience. By doing this, you'll avoid confusion and keep your messaging focused on what your analysis has produced.&lt;/p&gt;

&lt;h1&gt;
  
  
  3. Data Analytics Peer Review
&lt;/h1&gt;

&lt;p&gt;Putting a peer review process in place for your analysis is one of the best ways to ensure your work is sound and accurate. Getting a second set of eyes on your analysis can help you find potential errors or room for improvement. If a fellow analyst confirms your analysis, you'll know that your work is ready for presentation.&lt;/p&gt;

&lt;p&gt;Peer review is especially important for less experienced data analysts. If you can get a more experienced analyst to review your work, you'll be able to learn from their insights and comments. It's also helpful for analysts with more technical roles who may &lt;a href="https://www.youtube.com/watch?v=uEPCxBaRf6A"&gt;be working on projects that were once primarily the domain of data engineers&lt;/a&gt;.&lt;/p&gt;

&lt;h1&gt;
  
  
  4. Triple-Check Your Data
&lt;/h1&gt;

&lt;p&gt;Whenever you're working with data, it's good practice to assume that the datasets have at least some flaws. These flaws can range from simple organizational errors to completely erroneous pieces of data. For this reason, you should get into the habit of triple-checking your data as you conduct your analyses.&lt;/p&gt;

&lt;p&gt;Finding flaws or inaccuracies in your datasets will help you provide a better analysis. In many cases, you'll be able to simply fix the issues you find and then proceed with your work. In others, though, you may discover much larger problems that require substantially more work to resolve.&lt;/p&gt;

&lt;p&gt;There are times when it's even advisable to ignore flawed data altogether. While you might lose some information, you have to know when a problem is too labor-intensive to be worth fixing. This decision will depend on the project and the nature of the problems with your data.&lt;/p&gt;

&lt;p&gt;While checking your data, it's a good idea not to assume that anything is accurate. Even something that seems foolproof may have errors that could throw your analysis off. Columns in which data are entered from a dropdown menu, for example, seem like they should be free from errors. If there are invalid options in the menu, though, you could end up with flawed data as a result. Assume that there are errors in any dataset and conduct a thorough search to find them. By doing this during the data cleaning phase of your project, you can save yourself from having to backtrack and fix mistakes later on.&lt;/p&gt;

&lt;h1&gt;
  
  
  5. Know When To Stop Your Analysis
&lt;/h1&gt;

&lt;p&gt;A final critical skill that is frequently overlooked is knowing when to stop your analysis. Having a set endpoint is a key part of the data analytics process. When you reach that endpoint, you need to be able to stop and finalize your analysis.&lt;/p&gt;

&lt;p&gt;Without a clear endpoint, you can easily think of new questions to ask and find yourself going down rabbit holes that aren't relevant to your project. While there are times when further exploration delivers useful insights, endless data analysis frequently fails to produce valuable results.&lt;/p&gt;

&lt;p&gt;Knowing when to stop relates directly to the first step of the data analytics process outlined above. If you don't know exactly what question you're trying to answer, it's very difficult to know where to stop. With a clearly defined question, you should have a natural endpoint beyond which there's no need for further analysis.&lt;/p&gt;

&lt;h1&gt;
  
  
  Conclusion
&lt;/h1&gt;

&lt;p&gt;The&lt;a href="https://www.youtube.com/watch?v=s40MptE20Tc"&gt; tools and resources available to data&lt;/a&gt; analysts and engineers are constantly changing as technology evolves. The fundamental role of an analyst, however, remains the same. As a data analyst, your primary task will always be to provide valuable, data-driven insights that help your business or organization achieve its goals.&lt;/p&gt;

&lt;p&gt;By focusing on these fundamental skills, you can give yourself an extremely &lt;a href="https://www.youtube.com/watch?v=wuiJWSz-OLQ"&gt;strong foundation as an analyst&lt;/a&gt;. From there, you can build your technical skills to expand your capabilities. Whatever tools you're using, though, keeping these basic principles in mind will help you improve as an analyst and create more value for your employer.&lt;/p&gt;

</description>
      <category>database</category>
      <category>datascience</category>
      <category>career</category>
      <category>sql</category>
    </item>
    <item>
      <title>How to Become a Data Engineer in 2022</title>
      <dc:creator>SeattleDataGuy</dc:creator>
      <pubDate>Sun, 23 Jan 2022 04:04:38 +0000</pubDate>
      <link>https://dev.to/seattledataguy/how-to-become-a-data-engineer-in-2022-3pha</link>
      <guid>https://dev.to/seattledataguy/how-to-become-a-data-engineer-in-2022-3pha</guid>
      <description>&lt;p&gt;Photo by &lt;a href="https://unsplash.com/@thisisengineering?utm_source=unsplash&amp;amp;utm_medium=referral&amp;amp;utm_content=creditCopyText" rel="noopener noreferrer"&gt;ThisisEngineering RAEng&lt;/a&gt; on &lt;a href="https://unsplash.com/?utm_source=unsplash&amp;amp;utm_medium=referral&amp;amp;utm_content=creditCopyText" rel="noopener noreferrer"&gt;Unsplash&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;If you were paying attention in 2021, then you would know that data engineering jobs are on the rise.&lt;/p&gt;

&lt;p&gt;&lt;a href="https://www.interviewquery.com/blog-data-science-interview-report/" rel="noopener noreferrer"&gt;Jay Feng's report&lt;/a&gt; analyzing 10,000 data science interviews found that interviews for data engineering jobs increased by 40% in 2020. There has also been a massive increase in funding in the data engineering space and the explosion of data-driven content has led to a revival of the term data engineering. I've also seen an increase in the number of companies offering data engineering internships. This is not just at Amazon and Facebook, but also at companies like Spotify.&lt;/p&gt;

&lt;p&gt;All of this information is good news for aspiring data engineers. It's possible to come out of college, land a good job, and thrive in this growing field.&lt;/p&gt;

&lt;p&gt;But what does it take to &lt;a href="https://www.youtube.com/watch?v=PEsCumjfBW0" rel="noopener noreferrer"&gt;become a data engineer in 2022&lt;/a&gt;? What skills do you need to have today so that they are still relevant five years from now?&lt;/p&gt;

&lt;p&gt;In this article, I'll cover what skills emerging data engineers need, where they can apply for jobs, and what they can do to stand out in this competitive field.&lt;/p&gt;

&lt;h1&gt;
  
  
  What Skills Do Data Engineers Need
&lt;/h1&gt;

&lt;p&gt;What skills are companies asking for from junior data engineers?&lt;/p&gt;

&lt;p&gt;It's important to know the basics. Most interviews, at a junior or intern level won't expect you to know more than the three sections below.&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;  Python&lt;/li&gt;
&lt;li&gt;  SQL&lt;/li&gt;
&lt;li&gt;  Data modeling and ETL development&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;Those are generally on all interviews. An exception is that if you are in coding you could learn either Python, Java, or Scala. If you know two of those languages, you can usually figure out the last one.&lt;/p&gt;

&lt;p&gt;If you're going into an internship or junior position, you shouldn't be expected to know much more than these requirements. If you're going into a mid-level position, then there might be an expectation to know a little more.&lt;/p&gt;

&lt;h1&gt;
  
  
  Getting A Data Engineer Internship At Amazon
&lt;/h1&gt;

&lt;p&gt;&lt;a href="https://media.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fmiro.medium.com%2Fmax%2F1400%2F1%2A45UrkvhJ24GxfzSTJqXuDQ.png" class="article-body-image-wrapper"&gt;&lt;img src="https://media.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fmiro.medium.com%2Fmax%2F1400%2F1%2A45UrkvhJ24GxfzSTJqXuDQ.png"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;&lt;a href="https://media.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fmiro.medium.com%2Fmax%2F1400%2F1%2AzxLuBuZSQ6PfutRSln3Okw.png" class="article-body-image-wrapper"&gt;&lt;img src="https://media.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fmiro.medium.com%2Fmax%2F1400%2F1%2AzxLuBuZSQ6PfutRSln3Okw.png"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;Let's take a look at this data engineering internship opportunity from Amazon. They expect that you have a bachelor's degree, which makes sense. They also expect you to know Python, how to create data pipelines, databases, and warehouse modeling concepts. All of this makes sense.&lt;/p&gt;

&lt;p&gt;When you scroll down to the preferred qualifications, there are some questionable statements. Amazon is asking for a master's degree, which seems unnecessary for any sort of computer engineering, software engineering, or data engineering job. These skills are not generally learned in college and there's not exactly a data engineering master's degree.&lt;/p&gt;

&lt;p&gt;I also find it frustrating that they want applicants to articulate the basic differences between data types. I'm not sure what they're going for there because NoSQL and relational data types are very similar. I understand what they're getting at but wish they had phrased that point better. They should have said something like &lt;em&gt;understanding the difference between SQL and NoSQL&lt;/em&gt; or something similar.&lt;/p&gt;

&lt;p&gt;However, they have stated that you'll need a lot of SQL and data modeling experience. I've interviewed at Amazon, and those are definitely skills you will need.&lt;/p&gt;

&lt;h1&gt;
  
  
  Getting A Data Engineer Internship At Facebook
&lt;/h1&gt;

&lt;p&gt;&lt;a href="https://media.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fmiro.medium.com%2Fmax%2F1400%2F1%2ANStQetj9aPcB6JkBwMdxjA.png" class="article-body-image-wrapper"&gt;&lt;img src="https://media.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fmiro.medium.com%2Fmax%2F1400%2F1%2ANStQetj9aPcB6JkBwMdxjA.png"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;&lt;a href="https://media.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fmiro.medium.com%2Fmax%2F1400%2F1%2AugfjGu-ReEgaL1SunX_8MQ.png" class="article-body-image-wrapper"&gt;&lt;img src="https://media.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fmiro.medium.com%2Fmax%2F1400%2F1%2AugfjGu-ReEgaL1SunX_8MQ.png"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;Taking a look at this &lt;a href="https://www.facebookcareers.com/jobs/253886259650481/" rel="noopener noreferrer"&gt;data engineer intern role&lt;/a&gt; at Facebook, they also require experience with SQL Python. They also call out the need to understand distributed systems, which I think is unnecessary, having worked at Facebook or Meta. You don't need to know things that are under the hood, like Presto or Hive, because your job will just involve writing SQL. Nor will you be writing any MapReduce jobs.&lt;/p&gt;

&lt;h1&gt;
  
  
  Where To Apply
&lt;/h1&gt;

&lt;p&gt;I found a lot of junior data engineer positions just by Googling. There are a lot more than were around when I was going into data engineering.&lt;/p&gt;

&lt;p&gt;Some companies don't seem to understand the term junior and are asking for 2+ years of experience whereas a junior role should be 0--2 years of experience.&lt;/p&gt;

&lt;p&gt;However, there are plenty of junior and internship positions advertised. Thus, you can just ignore the ones that are asking for too much experience.&lt;/p&gt;

&lt;p&gt;At least make sure that the position is paying well (at least $80--100K) if it asks for a lot of experience. The Amazon internship that we looked at is paying $7,700 a month for a position based in Colorado, which makes it close to $100,000 a year.&lt;/p&gt;

&lt;h1&gt;
  
  
  How To Stand Out
&lt;/h1&gt;

&lt;p&gt;Although there are a lot of positions for data engineers, there are also a lot more people wanting to become data engineers. Entry-level positions are paying up to $100K per annum, &lt;a href="https://policyadvice.net/insurance/insights/average-american-income/#:~:text=What%20percentage%20of%20Americans%20makes,the%20population%20earned%20over%20%24200%2C000." rel="noopener noreferrer"&gt;which would put you into the top 15% of American earners&lt;/a&gt;. You could even make more if you worked at a successful startup with equity options. So, it's important to know how to stand out.&lt;/p&gt;

&lt;p&gt;To get a job typically you would study hard, apply for positions and network and get referrals. However, to stand out, you need to promote yourself. You could be the smartest data engineer in the world but if you don't know how to promote yourself, no one will know.&lt;/p&gt;

&lt;p&gt;In 2022, there are a ton of avenues to promote yourself. You can make videos, write articles, and share your code, as well as what you're currently learning and/or working on.&lt;/p&gt;

&lt;p&gt;Channels to publish content include:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;  &lt;a href="https://www.linkedin.com/company/18129251/" rel="noopener noreferrer"&gt;LinkedIn&lt;/a&gt;
&lt;/li&gt;
&lt;li&gt;  &lt;a href="https://medium.com/@SeattleDataGuy" rel="noopener noreferrer"&gt;Medium&lt;/a&gt;
&lt;/li&gt;
&lt;li&gt;  &lt;a href="https://www.theseattledataguy.com/" rel="noopener noreferrer"&gt;Your own website&lt;/a&gt;
&lt;/li&gt;
&lt;li&gt;  Your own GitHub repository&lt;/li&gt;
&lt;li&gt;  &lt;a href="https://www.youtube.com/channel/UCmLGJ3VYBcfRaWbP6JLJcpA" rel="noopener noreferrer"&gt;YouTube&lt;/a&gt;
&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;This kind of content helps recruiters and companies get to know you as a person, not just as a resume that hits their desks.&lt;/p&gt;

&lt;p&gt;The content publishing method can be tailored to suit your personality and skills. Some people are great at creating their own projects, like doing open-source work and writing code. For people who don't like coding as much, you can work on more high-level concepts and write blogs. This might involve understanding data modeling and writing basic data modeling breakdowns. For example, you could write articles about what the different tables in a warehouse are or compare modern concepts like data mesh, data fabric, and &lt;a href="https://www.theseattledataguy.com/snowflake-vs-bigquery%e2%80%8a-%e2%80%8atwo-cloud-data-warehouses-of-many-setting-up-your-data-in-the-modern-data-stack/" rel="noopener noreferrer"&gt;data warehousing&lt;/a&gt;.&lt;/p&gt;

&lt;p&gt;This type of work helps you to prepare for interviews as well as produce content to share with potential employers. Platforms like YouTube are great because you can share your work and ideas with what feels like an infinite number of people. I've had hiring managers reach out to me because there are not a lot of people publishing videos about data engineering.&lt;/p&gt;

&lt;h1&gt;
  
  
  Conclusion
&lt;/h1&gt;

&lt;p&gt;There are many things you can do to increase your chances of becoming a data engineer in 2022. You can start by learning the right skills and applying for positions. However, it's important to also promote yourself through content publishing. This will help you stand out from the crowd and show potential employers that you're passionate about data engineering. There are many avenues to publish content, so find one that best suits your personality and skillset.&lt;/p&gt;

&lt;h1&gt;
  
  
  Connect With Me:
&lt;/h1&gt;

&lt;p&gt;✅ YouTube:&lt;a href="https://www.youtube.com/channel/UCmLGJ3VYBcfRaWbP6JLJcpA" rel="noopener noreferrer"&gt;https://www.youtube.com/channel/SeattleDataGuy&lt;/a&gt;&lt;br&gt;
✅ Website: &lt;a href="https://www.theseattledataguy.com/" rel="noopener noreferrer"&gt;https://www.theseattledataguy.com/&lt;/a&gt;&lt;br&gt;
✅ LinkedIn: &lt;a href="https://www.linkedin.com/company/18129251" rel="noopener noreferrer"&gt;https://www.linkedin.com/company/18129251&lt;/a&gt;&lt;br&gt;
✅ Personal Linkedin: &lt;a href="https://www.linkedin.com/in/benjaminrogojan/" rel="noopener noreferrer"&gt;https://www.linkedin.com/in/benjaminrogojan/&lt;/a&gt;&lt;br&gt;
✅ FaceBook: &lt;a href="https://www.facebook.com/SeattleDataGuy" rel="noopener noreferrer"&gt;https://www.facebook.com/SeattleDataGu&lt;/a&gt;&lt;/p&gt;

</description>
      <category>career</category>
      <category>beginners</category>
      <category>database</category>
      <category>datascience</category>
    </item>
    <item>
      <title>8 Great Data Engineering Youtube Channels</title>
      <dc:creator>SeattleDataGuy</dc:creator>
      <pubDate>Sat, 15 Jan 2022 15:33:18 +0000</pubDate>
      <link>https://dev.to/seattledataguy/8-great-data-engineering-youtube-channels-21n7</link>
      <guid>https://dev.to/seattledataguy/8-great-data-engineering-youtube-channels-21n7</guid>
      <description>&lt;p&gt;Photo by &lt;a href="https://unsplash.com/@dariox?utm_source=medium&amp;amp;utm_medium=referral" rel="noopener noreferrer"&gt;Dario&lt;/a&gt; on &lt;a href="https://unsplash.com/?utm_source=medium&amp;amp;utm_medium=referral" rel="noopener noreferrer"&gt;Unsplash&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;Looking at the data engineering Youtube space early last year. There weren't that many people talking about data engineering.&lt;/p&gt;

&lt;p&gt;There were a few but overall most data-focused videos were on data science, algorithms, and machine learning.&lt;/p&gt;

&lt;p&gt;But, somewhere in the middle of last year, it seemed like more and more YouTubers were talking about data engineering.&lt;/p&gt;

&lt;p&gt;Some were referencing the trend itself of growth in data engineering job opportunities and others, like myself, decided to start making videos to help future data engineers.&lt;/p&gt;

&lt;p&gt;In this article, I wanted to help signal boost some of these great data engineering Youtube channels you should check out!&lt;/p&gt;

&lt;p&gt;So let's dive in.&lt;/p&gt;

&lt;h1&gt;
  
  
  Zack Wilson
&lt;/h1&gt;

&lt;p&gt;Since the first time I saw a post from&lt;a href="https://www.youtube.com/c/DatawithZach" rel="noopener noreferrer"&gt; Zack Wilson&lt;/a&gt; on Linkedin, I was waiting for his Youtube channel.&lt;/p&gt;

&lt;p&gt;And towards the end of 2021, he delivered.&lt;/p&gt;

&lt;p&gt;In one of his first videos, Zack discusses if data engineers should learn Scala. He does a great job discussing the pros and cons of Scala. From the fact that Scala allows for a more software engineering-driven approach to pipelines to the fact that there are only a limited number of positions that truly require Scala.&lt;/p&gt;

&lt;p&gt;But don't take my word for it. This short but concise clip does a great job answering the question.&lt;/p&gt;

&lt;p&gt;&lt;a href="https://www.youtube.com/watch?v=YG5-0x9g0eA" rel="noopener noreferrer"&gt;&lt;img src="https://media.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2F6m8b8ee8bsxxxkin02uc.png" title="Let's check Jason S' profile page (inline-style)" alt="invisible description of images, read aloud to blind users&amp;lt;br&amp;gt;
"&gt;&lt;/a&gt;&lt;br&gt;
&lt;a href="https://www.youtube.com/c/DatawithZach" rel="noopener noreferrer"&gt;All Videos Here&lt;/a&gt;&lt;/p&gt;

&lt;h1&gt;
  
  
  Andreas Kretz
&lt;/h1&gt;

&lt;p&gt;&lt;a href="https://medium.com/u/f28a432ce853?source=post_page-----e957b19ad253-----------------------------------" rel="noopener noreferrer"&gt;Andreas Kretz&lt;/a&gt; is the OG &lt;a href="https://www.youtube.com/c/andreaskayy" rel="noopener noreferrer"&gt;data engineering Youtuber&lt;/a&gt;. He has been putting out loads of content on data engineering for years now.&lt;/p&gt;

&lt;p&gt;Last year I really enjoyed several of his posts where he comment on videos where companies were breaking down their data infrastructure.&lt;/p&gt;

&lt;p&gt;Andreas clearly has a lot of practical experience and can&lt;/p&gt;

&lt;p&gt;In this video he reviews the AWS architecture that Nielsen from a Data Engineering Experts view. They use it to process up to 55TB a day.&lt;/p&gt;

&lt;p&gt;Really nice use case of Apache Spark and RDS. Nice use of AWS Lambda functions for content delivery and AWS SQS for job management. Teaches you a lot about Data Engineering.&lt;/p&gt;

&lt;p&gt;&lt;a href="https://www.youtube.com/watch?v=WCQe1VP_q5A" rel="noopener noreferrer"&gt;&lt;img src="https://media.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fjoyagskc3fvhyi8oyeln.png" title="Let's check Jason S' profile page (inline-style)" alt="invisible description of images, read aloud to blind users&amp;lt;br&amp;gt;
"&gt;&lt;/a&gt;&lt;br&gt;
&lt;a href="https://www.youtube.com/c/andreaskayy" rel="noopener noreferrer"&gt;All Videos Here&lt;/a&gt;&lt;/p&gt;

&lt;h1&gt;
  
  
  Darshil Parmar
&lt;/h1&gt;

&lt;p&gt;&lt;a href="https://www.youtube.com/c/DarshilParmar" rel="noopener noreferrer"&gt;Darshil&lt;/a&gt; has started putting out a lot of great content for the past year. He has discussed his data engineering journey, how he went from working for companies to freelancing and so much more.&lt;/p&gt;

&lt;p&gt;Actually his recent video I think is great for people just getting started.&lt;/p&gt;

&lt;p&gt;Here is his description:&lt;/p&gt;

&lt;p&gt;&lt;em&gt;How did I become a data engineer?&lt;/em&gt;&lt;/p&gt;

&lt;p&gt;&lt;em&gt;What are the skills needed for data engineering?&lt;/em&gt;&lt;/p&gt;

&lt;p&gt;&lt;em&gt;How to get a job as a data engineer and many more?&lt;/em&gt;&lt;/p&gt;

&lt;p&gt;&lt;em&gt;Data engineering is a new branch of engineering, but the data analysis field is growing rapidly and so is the demand for professionals who can handle this field. Data scientists are redefining how businesses use data by helping them access, analyze, and interpret data in their everyday lives. In order to be successful in this field, there are many skills you need to learn. Data engineering, or the ability to store and access data, is one of them. In this video, Darshil Parmar talks about how he became a data engineer and what skills he learned on the job. --- Darshil Parmar&lt;/em&gt;&lt;/p&gt;

&lt;p&gt;&lt;a href="https://www.youtube.com/watch?v=AS2EyYK4x_Q" rel="noopener noreferrer"&gt;&lt;img src="https://media.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fdzu80ahx1ht2o1simr9h.png" title="Let's check Jason S' profile page (inline-style)" alt="invisible description of images, read aloud to blind users&amp;lt;br&amp;gt;
"&gt;&lt;/a&gt;&lt;br&gt;
&lt;a href="https://www.youtube.com/c/DarshilParmar" rel="noopener noreferrer"&gt;All Videos Here&lt;/a&gt;&lt;/p&gt;

&lt;h1&gt;
  
  
  Seattle Data Guy (Me)
&lt;/h1&gt;

&lt;p&gt;Alright. I, of course, am going to put myself on this list. I have personally really been enjoying putting out videos on data engineering. Currently, I would say the focus of &lt;a href="https://www.youtube.com/c/SeattleDataGuy" rel="noopener noreferrer"&gt;my channel&lt;/a&gt; has been more early-career style videos.&lt;/p&gt;

&lt;p&gt;However, in the coming year, I would like to dive more into other topics such as data infrastructure, different data tools like dbt, Trino, Airflow, and so on.&lt;/p&gt;

&lt;p&gt;But we shall see what the future holds!&lt;/p&gt;

&lt;p&gt;The video I picked was me reviewing some of my favorite books for data engineering. If you're new to data engineering I am sure you can find a good new book to read.&lt;/p&gt;

&lt;p&gt;&lt;a href="https://www.youtube.com/watch?v=xtfuO7kGJeY" rel="noopener noreferrer"&gt;&lt;img src="https://media.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fmvx3gl3fky1ru11l8h09.png" title="Let's check Jason S' profile page (inline-style)" alt="invisible description of images, read aloud to blind users&amp;lt;br&amp;gt;
"&gt;&lt;/a&gt;&lt;br&gt;
&lt;a href="https://www.youtube.com/c/SeattleDataGuy/" rel="noopener noreferrer"&gt;All Videos Here&lt;/a&gt;&lt;/p&gt;

&lt;h1&gt;
  
  
  Karolina Sowinska
&lt;/h1&gt;

&lt;p&gt;&lt;a href="https://www.youtube.com/channel/UCAxnMry1lETl47xQWABvH7g" rel="noopener noreferrer"&gt;Karolina Sowinska&lt;/a&gt; has also been a long-standing data Youtuber. She does tend to have a broader set of topics other than data engineering but she has a lot of classic data engineering videos.&lt;/p&gt;

&lt;p&gt;For example, she has a great quick intro for Airflow that I shared below.&lt;/p&gt;

&lt;p&gt;In this long-awaited Airflow for Beginners video, she will show you how to install Airflow from scratch, and how to schedule your first ETL job in Airflow.&lt;/p&gt;

&lt;p&gt;So don't miss out!&lt;br&gt;
&lt;a href="https://www.youtube.com/watch?v=i25ttd32-eo" rel="noopener noreferrer"&gt;&lt;img src="https://media.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fnggcene50cko2jyxsjph.png" title="Let's check Jason S' profile page (inline-style)" alt="invisible description of images, read aloud to blind users&amp;lt;br&amp;gt;
"&gt;&lt;/a&gt;&lt;br&gt;
&lt;a href="https://www.youtube.com/c/KarolinaSowinska" rel="noopener noreferrer"&gt;All Videos Here&lt;/a&gt;&lt;/p&gt;

&lt;h1&gt;
  
  
  Mehdio
&lt;/h1&gt;

&lt;p&gt;&lt;a href="https://medium.com/u/d9259b41277a?source=post_page-----e957b19ad253-----------------------------------" rel="noopener noreferrer"&gt;mehdio&lt;/a&gt; is new to the Youtube space but we are glad to have him jump aboard!&lt;/p&gt;

&lt;p&gt;We have collaborated on a writing project and I really do think he has a great perspective on data engineering.&lt;/p&gt;

&lt;p&gt;Although his &lt;a href="https://www.youtube.com/channel/UCiZxJB0xWfPBE2omVZeWPpQ" rel="noopener noreferrer"&gt;channel only has one or two videos&lt;/a&gt;, I can see it growing over the next year.&lt;/p&gt;

&lt;p&gt;The video I picked here was Mehdio's only core video. In this video Mehdio will look at data engineering from a historical perspective as well as try to predict where things are going in the space.&lt;br&gt;
&lt;a href="https://www.youtube.com/watch?v=GYqt94348wU" rel="noopener noreferrer"&gt;&lt;img src="https://media.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fot4urrmahe586kchrid0.png" title="Let's check Jason S' profile page (inline-style)" alt="invisible description of images, read aloud to blind users&amp;lt;br&amp;gt;
"&gt;&lt;/a&gt;&lt;br&gt;
&lt;a href="https://www.youtube.com/channel/UCiZxJB0xWfPBE2omVZeWPpQ" rel="noopener noreferrer"&gt;All Videos Here&lt;/a&gt;&lt;/p&gt;

&lt;h1&gt;
  
  
  E-Learning Bridge
&lt;/h1&gt;

&lt;p&gt;Shashank Mishra is a data engineer that has worked at Amazon, McKinsey &amp;amp; Company and PayTm. He uses his vast experiences and wide network to create videos to help young data engineers learn more about the role and how they can get a job at a top tech company.&lt;/p&gt;

&lt;p&gt;In the video below Shashank will go over some of the SQL topics you will need to know as you study for your FAANG interviews.&lt;/p&gt;

&lt;p&gt;&lt;a href="https://www.youtube.com/watch?v=Y7R4X6ZOSS0" rel="noopener noreferrer"&gt;&lt;img src="https://media.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2F2790e2ue0yf4f2xo7uuu.png" title="Let's check Jason S' profile page (inline-style)" alt="invisible description of images, read aloud to blind users&amp;lt;br&amp;gt;
"&gt;&lt;/a&gt;&lt;br&gt;
&lt;a href="https://www.youtube.com/c/LearningBridge" rel="noopener noreferrer"&gt;All Videos Here&lt;/a&gt;&lt;/p&gt;

&lt;h1&gt;
  
  
  Ken Jee
&lt;/h1&gt;

&lt;p&gt;Ok, &lt;a href="https://medium.com/u/6ee1f7466557" rel="noopener noreferrer"&gt;Ken Jee&lt;/a&gt; usually puts out content more around data science. However, one of his recent videos was focused on data engineering.&lt;/p&gt;

&lt;p&gt;So I did include him on the list. Maybe he'll jump on the data engineering bandwagon.&lt;/p&gt;

&lt;p&gt;The video I picked below is said data engineering video.&lt;/p&gt;

&lt;p&gt;In this video, he broke down why it seems like companies are hiring more data engineers in general. He also discussed how this would impact data scientists.&lt;/p&gt;

&lt;p&gt;Overall, I believe this video gives a good explanation of why so many companies are struggling to get value from their data strategy. If you're curious how the data world is shaping up, then I think Ken will provide a lot of strong context.&lt;/p&gt;

&lt;p&gt;&lt;a href="https://www.youtube.com/watch?v=vwvdtXMcNzI" rel="noopener noreferrer"&gt;&lt;img src="https://media.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fpx8vqc5y1dbclrkpiias.png" title="Let's check Jason S' profile page (inline-style)" alt="invisible description of images, read aloud to blind users&amp;lt;br&amp;gt;
"&gt;&lt;/a&gt;&lt;br&gt;
&lt;a href="https://www.youtube.com/c/KenJee1" rel="noopener noreferrer"&gt;All Videos Here&lt;/a&gt;&lt;/p&gt;

&lt;h1&gt;
  
  
  So, Do You Want To Become A Data Engineer?
&lt;/h1&gt;

&lt;p&gt;After all those great videos.&lt;/p&gt;

&lt;p&gt;The question remains.&lt;/p&gt;

&lt;p&gt;Should you become a data engineer?&lt;/p&gt;

&lt;p&gt;This is something only you can answer. Do you like creating data pipelines and automating data systems?&lt;/p&gt;

&lt;p&gt;Do you want to play the middleman between applications and software teams and the analytical teams?&lt;/p&gt;

&lt;p&gt;Hopefully the combination of the videos above have helped you form a better picture of what being a data engineer really means as well as provided context on how you can grow your data engineering career.&lt;/p&gt;

&lt;p&gt;Thank you for reading!&lt;/p&gt;

&lt;p&gt;If you want to watch or read more:&lt;/p&gt;

&lt;p&gt;&lt;a href="https://seattledataguy.substack.com/p/which-managed-version-of-airflow" rel="noopener noreferrer"&gt;Which Managed Version Of Airflow Should You Use?&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;&lt;a href="https://www.theseattledataguy.com/what-is-trino-and-why-is-it-great-at-processing-big-data/#page-content" rel="noopener noreferrer"&gt;What Is Trino And How It Manages Big Data&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;&lt;a href="https://www.youtube.com/watch?v=bqCXVpRqTpE" rel="noopener noreferrer"&gt;What I Learned From 100+ Data Engineering Interviews - Interview Tips&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;&lt;a href="https://seattledataguy.substack.com/p/5-big-data-experts-predictions-for" rel="noopener noreferrer"&gt;5 Big Data Experts Predictions For 2022 - From The Modern Data Stack To Data Science&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;&lt;a href="https://www.youtube.com/watch?v=PEsCumjfBW0" rel="noopener noreferrer"&gt;How I Would Become A Data Engineer in 2022&lt;/a&gt;&lt;/p&gt;

</description>
      <category>database</category>
      <category>datascience</category>
      <category>sql</category>
      <category>beginners</category>
    </item>
    <item>
      <title>Why You Should Become A Data Engineer And Not A Data Scientist</title>
      <dc:creator>SeattleDataGuy</dc:creator>
      <pubDate>Sun, 26 Dec 2021 22:59:57 +0000</pubDate>
      <link>https://dev.to/seattledataguy/why-you-should-become-a-data-engineer-and-not-a-data-scientist-2kho</link>
      <guid>https://dev.to/seattledataguy/why-you-should-become-a-data-engineer-and-not-a-data-scientist-2kho</guid>
      <description>&lt;p&gt;Photo by &lt;a href="https://unsplash.com/@magnetme?utm_source=unsplash&amp;amp;utm_medium=referral&amp;amp;utm_content=creditCopyText" rel="noopener noreferrer"&gt;Magnet.me&lt;/a&gt; on &lt;a href="https://unsplash.com/s/photos/job?utm_source=unsplash&amp;amp;utm_medium=referral&amp;amp;utm_content=creditCopyText" rel="noopener noreferrer"&gt;Unsplash&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;Data Science is a career path that many people are currently choosing based on an essay written a decade ago. Yes, it has been almost ten years since the &lt;a href="https://hbr.org/2012/10/data-scientist-the-sexiest-job-of-the-21st-century" rel="noopener noreferrer"&gt;Harvard Business Review essay&lt;/a&gt; was published, yet we are all still choosing Data Science as a professional path. However, if you read through LinkedIn recently, you will notice that many people have posted that they are recovering data scientists who have now become data engineers.&lt;/p&gt;

&lt;p&gt;Many of us are drawn into the subject of Data Science just to realize it isn't for us finally. And, while there is some gatekeeping around Data Engineering, it appears that many people who have initially been data scientists eventually switched over to Data Engineering.&lt;/p&gt;

&lt;p&gt;What may be the reason for this? In this blog post, I hope to help you avoid the whole transition from data scientist to data engineer by giving you a few fundamental reasons why you should become a data engineer rather than a data scientist.&lt;/p&gt;

&lt;p&gt;&lt;em&gt;Disclaimer: Some who read the title of this article might assume this is some form of:&lt;/em&gt;&lt;/p&gt;

&lt;p&gt;&lt;em&gt;"&lt;/em&gt;&lt;em&gt;We Don't Need Data Scientists&lt;/em&gt;&lt;em&gt;" or "&lt;/em&gt;&lt;em&gt;Data Engineering Is Better Than Data Science&lt;/em&gt;&lt;em&gt;" type article. That's not the purpose of this article. It is meant to discuss why someone may prefer being a data engineer. Of course, if you work at a small enough company you might be a little of both.&lt;/em&gt;&lt;/p&gt;

&lt;h2&gt;
  
  
  You Like Building Stuff
&lt;/h2&gt;

&lt;p&gt;If you enjoy building infrastructure, programming, and writing object-oriented code rather than merely procedural code to interact with data, you may be more of a data engineer.&lt;/p&gt;

&lt;p&gt;Data engineers develop data pipelines, infrastructure, monitoring, and other aspects that aren't immediately related to models. Data engineers operationalize or &lt;em&gt;productionize&lt;/em&gt; a model, which implies taking the analysis or Jupyter notebook that a data scientist created and applying it in a sustainable and robust system, rather than simply pressing "run" on that Jupyter notebook every day.&lt;/p&gt;

&lt;p&gt;We prefer the discipline and process of constructing infrastructure instead of simply burying it in data frames that no one can access or adequately QA. &lt;a href="https://www.youtube.com/watch?v=LgSHaOvNodA" rel="noopener noreferrer"&gt;Data engineers&lt;/a&gt; enjoy having a tangible end product. We don't want just an analysis; we want a table, a &lt;a href="https://seattledataguy.substack.com/p/what-are-etls-and-why-we-use-them" rel="noopener noreferrer"&gt;pipeline&lt;/a&gt;, a d&lt;a href="https://seattledataguy.substack.com/p/snowflake-vs-bigquery-two-cloud-data" rel="noopener noreferrer"&gt;ata warehouse&lt;/a&gt;, or a Data Lake.&lt;/p&gt;

&lt;h2&gt;
  
  
  You Like Feeling Done
&lt;/h2&gt;

&lt;p&gt;Data engineers enjoy the sense of accomplishment that comes with finishing a project. Data science has an unending capacity to generate questions after questions, making your analysis endless. I have watched my data science counterparts often finish an analysis on a single data set, only to have to dig into the data set more due to the business asking even more questions.&lt;/p&gt;

&lt;p&gt;But, as data engineers, we have a general standard to adhere to a table, a data pipeline, or something along those lines. Once we've created it, we're done. Sure, our stakeholders may say, "Oh, I wanted to add this column as well," but that's a new project or assignment, and we already know we've completed the previous project. In order to take on this new task we would need to reprioritize all of our current work.&lt;/p&gt;

&lt;p&gt;The preceding is not necessarily true in the field of Data Science. It can be an infinite complex of questions, none of which will ever lead to an answer.&lt;/p&gt;

&lt;p&gt;If having no actual end product is what you enjoy about your job, Data Science may be for you. However, if you prefer having a finished product at the end of the day, Data Engineering may be a better fit.&lt;/p&gt;

&lt;h2&gt;
  
  
  You Don't Like Being The Center of Attention
&lt;/h2&gt;

&lt;p&gt;Another advantage of being a data engineer is that we are not always the center of attention. Data Science has a lot of sexiness and glamour, whereas data engineers can hide in the background, which many of us enjoy. Instead of spending a lot of times in front of co-workers explaining the impact of your model, we can often hide behind our keyboards and build our tables for our partners.&lt;/p&gt;

&lt;p&gt;So, being a data engineer is ideal if you prefer completing your task without attracting a lot of attention and questioning. You get to do your work, and you know that once it's done, you can pass it over to the data scientist, who will analyze the data and then jump in front of a stakeholder or manager and explain what their results mean.&lt;/p&gt;

&lt;p&gt;In that way, data engineering is an excellent job for folks who prefer interacting with a keyboard over dealing with other people.&lt;/p&gt;

&lt;p&gt;Of course, I'm not implying that all data engineers are introverts or dislike interacting with others. You can get the opportunity to talk and share with others if you so desire. If you want to take it, it's there for you.&lt;/p&gt;

&lt;p&gt;And I will always be a proponent of improving your soft skills, especially in terms of communication. Strong communications can help you leap forward in your career. Why else do you think I started a Youtube channel.&lt;/p&gt;

&lt;p&gt;However, the business is more concerned with the analysts and data science findings. They are concerned with the amount of money they will save, the final model, and the influence on the business. Data engineering is critical, and data scientists cannot do their jobs without it. But few people care how the sausage is produced, they only care that it is on their plate.&lt;/p&gt;

&lt;h2&gt;
  
  
  You Prefer SQL Over Pandas
&lt;/h2&gt;

&lt;p&gt;&lt;a href="https://media.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fmiro.medium.com%2Fmax%2F1400%2F0%2AUkKhUyiOIIJh6Hi2" class="article-body-image-wrapper"&gt;&lt;img src="https://media.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fmiro.medium.com%2Fmax%2F1400%2F0%2AUkKhUyiOIIJh6Hi2"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;Photo by &lt;a href="https://unsplash.com/@millerthachiller?utm_source=medium&amp;amp;utm_medium=referral" rel="noopener noreferrer"&gt;Pascal Müller&lt;/a&gt; on &lt;a href="https://unsplash.com/?utm_source=medium&amp;amp;utm_medium=referral" rel="noopener noreferrer"&gt;Unsplash&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;Finally, if you prefer SQL over Pandas, you might be more of a data engineer.&lt;/p&gt;

&lt;p&gt;I've discovered that data scientists seem to prefer using Pandas, and most data engineers tend to lean towards SQL.&lt;/p&gt;

&lt;p&gt;In one way or another, both manipulate data. But if you need to execute a sophisticated, 1000-line query, I can only imagine how insane it would look in Pandas and how many calls to so many functions it would entail.&lt;/p&gt;

&lt;p&gt;What's fantastic is that we live in a world where Spark, SparkSQL and DataBricks exist, so we can all play in the same arena and use a similar engine, but SQL will continue to be the data language since it has withstood the test of time.&lt;/p&gt;

&lt;h2&gt;
  
  
  So Which Data Career Is Right For You
&lt;/h2&gt;

&lt;p&gt;So there you have it: four reasons why you should consider becoming a data engineer rather than a data scientist. If any of these points resonated with you, you might want to seek a career as a data engineer.&lt;/p&gt;

&lt;p&gt;It's easy to be caught up in the Data Science allure, given that it's had its still a very sexy job. Still, other data-related areas, such as data engineering, need completely different day-to-day tasks and team relationships.&lt;/p&gt;

&lt;p&gt;The skills necessary for these jobs, as well as the end deliverables, vary greatly. The bottom line is, to be honest with ourselves and select our best-suited, ideal career.&lt;/p&gt;

&lt;p&gt;Of course, there are still other roles like &lt;a href="https://www.youtube.com/watch?v=uEPCxBaRf6A" rel="noopener noreferrer"&gt;analytics engineer and data analyst&lt;/a&gt;. So good luck finding the right career.&lt;/p&gt;

&lt;p&gt;If you're interested in learning more about data engineering or data science, then consider these articles and videos.&lt;/p&gt;

&lt;p&gt;&lt;a href="https://www.youtube.com/watch?v=xtfuO7kGJeY" rel="noopener noreferrer"&gt;My Favorite Books For Data Engineers - From Streaming To Software Engineering&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;&lt;a href="https://seattledataguy.substack.com/p/which-managed-version-of-airflow" rel="noopener noreferrer"&gt;Which Managed Version Of Airflow Should You Use?&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;&lt;a href="https://www.theseattledataguy.com/what-is-trino-and-why-is-it-great-at-processing-big-data/#page-content" rel="noopener noreferrer"&gt;What Is Trino And How It Manages Big Data&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;&lt;a href="https://www.youtube.com/watch?v=bqCXVpRqTpE" rel="noopener noreferrer"&gt;What I Learned From 100+ Data Engineering Interviews - Interview Tips&lt;br&gt;
&lt;/a&gt;&lt;/p&gt;

</description>
      <category>datascience</category>
      <category>career</category>
      <category>beginners</category>
      <category>database</category>
    </item>
    <item>
      <title>What Is Trino And Why Is It Great At Processing Big Data</title>
      <dc:creator>SeattleDataGuy</dc:creator>
      <pubDate>Thu, 09 Dec 2021 20:45:52 +0000</pubDate>
      <link>https://dev.to/seattledataguy/what-is-trino-and-why-is-it-great-at-processing-big-data-8pc</link>
      <guid>https://dev.to/seattledataguy/what-is-trino-and-why-is-it-great-at-processing-big-data-8pc</guid>
      <description>&lt;p&gt;Big data is touted as the solution to many problems. However, the truth is, big data has caused a lot of big problems.&lt;/p&gt;

&lt;p&gt;Yes, big data can provide more context and provide insights we have never had before. It also makes queries slow, data expensive to manage, requires a lot of expensive specialists to process it, and just continues to grow.&lt;/p&gt;

&lt;p&gt;Overall, data, especially big data, has forced companies to develop better data management tools to ensure data scientists and analysts can work with all their data.&lt;/p&gt;

&lt;p&gt;One of these companies was Facebook who decided they needed to develop a new engine to process all of their petabytes effectively. This tool was called Presto which recently broke off into another project called Trino.&lt;/p&gt;

&lt;p&gt;In this article, we outline what Trino is, why people use it and some of the challenges people face when deploying it.&lt;/p&gt;

&lt;h1&gt;
  
  
  Presto Trino History
&lt;/h1&gt;

&lt;p&gt;Before diving into what Trino is and why people use it. Let's recap how Presto turned into Trino.&lt;/p&gt;

&lt;h1&gt;
  
  
  What Is Trino
&lt;/h1&gt;

&lt;p&gt;Let's be clear. &lt;a href="https://trino.io/" rel="noopener noreferrer"&gt;Trino&lt;/a&gt; is not a database. This is a misconception. Just because you utilize Trino to run SQL against data, doesn't mean it's a database.&lt;/p&gt;

&lt;p&gt;Instead, Trino is a SQL engine. More specifically, Trino is an open-source distributed SQL query engine for adhoc and batch &lt;a href="https://www.theseattledataguy.com/etls-vs-elts-why-are-elts-disrupting-the-data-market-data-engineering-consulting/#page-content" rel="noopener noreferrer"&gt;ETL&lt;/a&gt; queries against multiple types of data sources. Not to mention it can manage a whole host of both standard and semi-structured data types like JSON, Arrays, and Maps.&lt;/p&gt;

&lt;p&gt;Another important point to discuss about Trino is the history of Presto. In 2012 &lt;a href="https://www.linkedin.com/in/traversomartin/" rel="noopener noreferrer"&gt;Martin Traverso&lt;/a&gt;, David Phillips, Dain Sundstrom and Eric Hwang were working at Facebook and developed Presto to replace Apache Hive to better process the hundreds of petabytes Facebook was trying to analyze.&lt;/p&gt;

&lt;p&gt;Due to the creators desire to keep the project open and community based they open sourced it in November 2013.&lt;/p&gt;

&lt;p&gt;But due to Facebook wanting to have tighter control over the project there was an eventual split.&lt;/p&gt;

&lt;p&gt;The original creators of Presto decided that they wanted to keep Presto open-source and in turn pursued building the Presto Open Source Community full-time. They did this under the new name PrestoSQL.&lt;/p&gt;

&lt;p&gt;Facebook decided to build a competing community using The Linux Foundation®. As a first action, Facebook applied for a trademark on Presto®. This eventually led to litigation and other challenges that forced the original group who developed Presto to rebrand.&lt;/p&gt;

&lt;p&gt;&lt;a href="https://media.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fmiro.medium.com%2Fmax%2F1400%2F0%2AvJa4YaAaF0qoR7tW" class="article-body-image-wrapper"&gt;&lt;img src="https://media.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fmiro.medium.com%2Fmax%2F1400%2F0%2AvJa4YaAaF0qoR7tW"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;&lt;a href="https://trino.io/blog/2020/12/27/announcing-trino.html" rel="noopener noreferrer"&gt;Source&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;Starting in December 2020 &lt;a href="https://trino.io/blog/2020/12/27/announcing-trino.html" rel="noopener noreferrer"&gt;PrestoSQL was rebranded as Trino&lt;/a&gt;. This has been a little confusing but now Trino supports a lot of end-users and has a large base of developers that commit to it regularly.&lt;/p&gt;

&lt;h1&gt;
  
  
  How does Trino work?
&lt;/h1&gt;

&lt;p&gt;Trino is a distributed system that utilizes an architecture similar to massively parallel processing (MPP) databases. Like many other big data engines there is a form of a coordinator node that then manages multiple worker nodes to process all the work that needs to be done.&lt;/p&gt;

&lt;p&gt;An analyst or general user would run their SQL which gets pushed to the coordinator. In turn the coordinator then parses, plans, and schedules a distributed query. It supports standard ANSI &lt;a href="https://seattledataguy.substack.com/p/how-to-write-better-sql-advanced" rel="noopener noreferrer"&gt;SQL&lt;/a&gt; as well as allows users to run more complex transformations like JSON and MAP transformations and parsing.&lt;/p&gt;

&lt;h1&gt;
  
  
  Why People Use Trino
&lt;/h1&gt;

&lt;p&gt;Trino, being a Presto spin-off, has a lot of benefits that came from its development by a large data company that needs to easily query across multiple data sources without spending too much time processing ETLs. In addition, it was developed to scale on cloud-like infrastructure. Although, most of Facebook's infrastructure is based on its internal cloud. But let's dig into why people are using Trino.&lt;/p&gt;

&lt;h1&gt;
  
  
  Agnostic Data Source Connections
&lt;/h1&gt;

&lt;p&gt;There are plenty of options when it comes to how you can query your data. There are tools like Athena, Hive and Apache Drill.&lt;/p&gt;

&lt;p&gt;So why use Trino to run SQL?&lt;/p&gt;

&lt;p&gt;Trino provides many benefits for developers. For example, the biggest advantage of Trino is that it is just a SQL engine. Meaning it agnostically sits on top of various data sources like MySQL, HDFS, and SQL Server.&lt;/p&gt;

&lt;p&gt;So if you want to run a query across these different data sources, you can.&lt;/p&gt;

&lt;p&gt;This is a powerful feature that eliminates the need for users to understand connections and SQL dialects of underlying systems.&lt;/p&gt;

&lt;h1&gt;
  
  
  Cloud-focused
&lt;/h1&gt;

&lt;p&gt;Presto's fundamental design of running storage and computing separately makes it extremely convenient to operate in cloud environments. Since the Presto cluster doesn't store any data, it can be auto-scaled depending on the load without causing any data loss.&lt;/p&gt;

&lt;h1&gt;
  
  
  Use Cases For Trino
&lt;/h1&gt;

&lt;p&gt;Adhoc queries And Reporting --- Trino allows end-users to use SQL to run ad hoc queries where your data resides. More importantly, you can create queries and data sets for reporting and adhoc needs.&lt;/p&gt;

&lt;p&gt;Data Lake Analytics --- One of many of the common use cases for Trino is being able to directly query data on a data lake without the need for transformation. You can query data that is structured or semi-structured in various sources. This means you can create operational dashboards without massive transformations.&lt;/p&gt;

&lt;p&gt;Batch ETLs --- Trino is a great engine to run your ETL batch queries. That's because it can process large volumes of data quickly as well as bring in data from multiple sources without always needing to extract data from sources such as MySQL.&lt;/p&gt;

&lt;h1&gt;
  
  
  Challenges For Trino Users
&lt;/h1&gt;

&lt;p&gt;At this point, you might be assuming that everyone should use Trino. But it's not that simple. Using Trino requires a pretty large amount of set-up for one. However, there are also a few other issues you might deal with upon set-up.&lt;/p&gt;

&lt;h1&gt;
  
  
  Federated Queries Can Be Slow
&lt;/h1&gt;

&lt;p&gt;The one downside of federated queries is that there can be some trade-offs in speed. This can be caused by a lack of meta-data stored and managed by Trino to better run queries. In addition, Presto was initially developed at Facebook that essentially has their own cloud. For them to expand it and grow it as they need increased speed isn't a huge problem. However, for other organizations, in order to get the same level of performance they might need to spend a lot more money to add more machines to their clusters. This can become very expensive. All to manage unindexed data.&lt;/p&gt;

&lt;p&gt;One such example is &lt;a href="https://varada.io/" rel="noopener noreferrer"&gt;Varada&lt;/a&gt;. Varada indexes data in Trino in such a way that reduces the time the CPU is used for data scanning (via indexes) and frees the cpu up for other tasks like fetching the data extremely fast or dealing with concurrency. thus allows SQL users to run various queries, whether across dimensions, facts as well as other types of joins and datalake analyticss, on indexed data as federated data sources. SQL aggregations and grouping is accelerated using nanoblock indexes as well, resulting in highly effective SQL analytics.&lt;/p&gt;

&lt;h1&gt;
  
  
  Configuration And Set-Up
&lt;/h1&gt;

&lt;p&gt;Setting up Trino isn't straightforward. Especially when it comes to optimizing performance and management. In turn, many system admins and IT teams will need teams to both set up and manage their instances of Trino.&lt;/p&gt;

&lt;p&gt;One great example of this is an AWS article titled "&lt;a href="https://aws.amazon.com/blogs/big-data/top-9-performance-tuning-tips-for-prestodb-on-amazon-emr/" rel="noopener noreferrer"&gt;Top 9 Performance Tuning Tips For PrestoDB&lt;/a&gt;".&lt;/p&gt;

&lt;h1&gt;
  
  
  Lack Of Enterprise Features
&lt;/h1&gt;

&lt;p&gt;One of the largest challenges faced by companies utilizing Trino is that out of the box, there aren't a lot of features geared towards enterprise solutions. That is to say, features that revolve around security, access control, and even expanded data source connectivity are limited. Many solutions are trying to be better provided in this area.&lt;/p&gt;

&lt;p&gt;A great example of this is &lt;a href="https://seattledataguy.substack.com/p/starburst-data-raised-100m-but-what" rel="noopener noreferrer"&gt;Starburst Data&lt;/a&gt;.&lt;/p&gt;

&lt;p&gt;Starburst Enterprise has several features that help improve the Trino's lacking security features. For example, Starburst makes it easy for your team to set up access control.&lt;/p&gt;

&lt;p&gt;The access control systems all follow role-based access control mechanisms with users, groups, roles, privileges and objects.&lt;/p&gt;

&lt;p&gt;This is demonstrated in the image below&lt;/p&gt;

&lt;p&gt;&lt;a href="https://media.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fmiro.medium.com%2Fmax%2F1400%2F0%2AAo14IL0Z3IXE5nwJ" class="article-body-image-wrapper"&gt;&lt;img src="https://media.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fmiro.medium.com%2Fmax%2F1400%2F0%2AAo14IL0Z3IXE5nwJ"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;&lt;a href="https://docs.starburst.io/latest/security/access-control.html" rel="noopener noreferrer"&gt;Source&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;This makes it easy for your security teams and &lt;a href="https://www.theseattledataguy.com/what-is-a-data-warehouse-and-why-we-build-them-a-video/" rel="noopener noreferrer"&gt;data warehouse&lt;/a&gt; administrators to manage who has access to what data.&lt;/p&gt;

&lt;p&gt;Starburst also offers other helpful security features such as auditing and encryption.&lt;/p&gt;

&lt;p&gt;This enables companies to implement a centralized security framework without having to code their own modules for Trino.&lt;/p&gt;

&lt;h1&gt;
  
  
  Conclusion
&lt;/h1&gt;

&lt;p&gt;Big data will continue to be a big problem for companies that don't start looking for tools like Trino to help them manage all their data. Trino's ability to be an agnostic SQL engine that can query large data sets across multiple data sources is a great option for many of these companies. But as discussed, Trino is far from perfect. It isn't fully optimized in a way for enterprise companies to take advantage of its full abilities. In addition, due to Trino's brute force approach to speed, it sometimes comes at a cost. It becomes very expensive to get the beneficial speed without indexing.&lt;/p&gt;

&lt;p&gt;This is where many new solutions are coming into the fold to make Trino more approachable. In the end, big data can be maintained and managed, you just need the right tools to help set yourself up for success.&lt;/p&gt;

&lt;p&gt;If you enjoyed this article, then check out these videos and articles below.&lt;/p&gt;

&lt;p&gt;&lt;a href="https://www.youtube.com/watch?v=uEPCxBaRf6A" rel="noopener noreferrer"&gt;Data Engineer Vs Analytics Engineer Vs Analyst&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;&lt;a href="https://www.theseattledataguy.com/why-migrate-to-the-modern-data-stack-and-where-to-start/" rel="noopener noreferrer"&gt;Why Migrate To The Modern Data Stack And Where To Start&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;&lt;a href="http://youtube.com/watch?v=s40MptE20Tc&amp;amp;t=1s" rel="noopener noreferrer"&gt;5 Great Data Engineering Tools For 2021 --- My Favorite Data Engineering Tools&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;&lt;a href="https://www.youtube.com/watch?v=kSt9NV-qZkc&amp;amp;t=1s" rel="noopener noreferrer"&gt;4 SQL Tips For Data Scientists&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;&lt;a href="https://logitanalytics.com/what-are-the-benefits-cloud-data-warehousing-and-why-you-should-migrate/" rel="noopener noreferrer"&gt;What Are The Benefits Of Cloud Data Warehousing And Why You Should Migrate&lt;/a&gt;&lt;/p&gt;

</description>
      <category>database</category>
      <category>bigdata</category>
      <category>datascience</category>
      <category>sql</category>
    </item>
    <item>
      <title>ETLs vs ELTs: Why are ELTs Disrupting the Data Market?</title>
      <dc:creator>SeattleDataGuy</dc:creator>
      <pubDate>Tue, 30 Nov 2021 13:53:52 +0000</pubDate>
      <link>https://dev.to/seattledataguy/etls-vs-elts-why-are-elts-disrupting-the-data-market-286o</link>
      <guid>https://dev.to/seattledataguy/etls-vs-elts-why-are-elts-disrupting-the-data-market-286o</guid>
      <description>&lt;p&gt;Photo by &lt;a href="https://unsplash.com/@abduzeedo?utm_source=unsplash&amp;amp;utm_medium=referral&amp;amp;utm_content=creditCopyText"&gt;Fabio Sasso&lt;/a&gt; on &lt;a href="https://unsplash.com/s/photos/pipelines?utm_source=unsplash&amp;amp;utm_medium=referral&amp;amp;utm_content=creditCopyText"&gt;Unsplash&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;In the business world, cloud technology has become more and more dominant in recent years. Right now, research shows that about &lt;a href="https://www.computerweekly.com/news/2240241926/Prepare-for-two-modes-of-business-intelligence-says-Gartner"&gt;50% of all business data&lt;/a&gt; is stored in the cloud, which just demonstrates the importance of external data sources and their place in the modern business environment.&lt;/p&gt;

&lt;p&gt;In seeking to keep up with digital transformation and data trends, many businesses are turning to &lt;a href="https://seattledataguy.substack.com/p/what-are-etls-and-why-are-they-important"&gt;ELT (Extract, Load, and Transform)&lt;/a&gt; tools. Besides accommodating heavy workloads, ELTs help teams integrate data.&lt;/p&gt;

&lt;p&gt;In this post, we'll take a look at ELTs and how they compare to ETLs, and why ELTs have become such a disruptive force in the data market.&lt;/p&gt;

&lt;p&gt;Let's dive in.&lt;/p&gt;

&lt;h2&gt;
  
  
  ETLs vs. ELTs
&lt;/h2&gt;

&lt;p&gt;While ETLs and ELTs both deal with data, they are, in fact, different tools. ETL is the Extract, Transform, and Load process for addressing data, while ELT is Extract, Load, and Transform. In an ETL model, data migrates from its original source to a data warehouse, where it is staged. In ELT, you will often use a very different data storage paradigm.&lt;/p&gt;

&lt;p&gt;Both ELT and ETL involve the following three steps:&lt;/p&gt;

&lt;p&gt;Extract&lt;/p&gt;

&lt;p&gt;The extraction phase of both ELT and ETL solutions involves pulling source data from the original database. In an ELT model, the data goes right to a data storage system. In classic ETLs you're pulling the data into a staging area.&lt;/p&gt;

&lt;p&gt;Transform&lt;/p&gt;

&lt;p&gt;Transformation is the process of changing the data's structure and is what ultimately allows the data to integrate with the target data system and the rest of the information it contains.&lt;/p&gt;

&lt;p&gt;Load&lt;/p&gt;

&lt;p&gt;Loading is the process of moving the data into a data storage system, which prepares it for analysis.&lt;/p&gt;

&lt;p&gt;ETL and ELT perform these steps in different orders. Teams deciding between the two solutions will need to determine whether to transform their data before or after moving it into the data storage repository.&lt;/p&gt;

&lt;p&gt;In the data science landscape, both ETL and ELT are necessary technologies. Because information sources, including unstructured NoSQL databases and structured SQL databases, almost never use the same formats, data must be transformed and enriched before it can be analyzed as a whole. By digesting this data, ETL and ELT solutions step in to allow business intelligence platforms to do their job.&lt;/p&gt;

&lt;p&gt;What each is good for: ETL is beneficial for organizations concerned about data compliance and privacy, since it cleans sensitive and secure data before sending it to the data warehouse. ETL, in contrast, is excellent for sophisticated data transformations and tends to be more affordable than ELT.&lt;/p&gt;

&lt;h2&gt;
  
  
  Why Should I Use an ELT?
&lt;/h2&gt;

&lt;p&gt;While both ETLs and ELTs have their place in the data landscape, more and more organizations are choosing to adopt ELT tools to address the volume and speed of their big data sources, which often overload the more traditional ETL tools.&lt;/p&gt;

&lt;p&gt;When used correctly, ELT tools streamline analysis data preparation. Because ELTs load data into the framework where it will eventually be processed, staged, and transformed, they allow teams to skip some busy work associated with data transformation.&lt;/p&gt;

&lt;p&gt;Here are a few of the benefits of using ELT systems:&lt;/p&gt;

&lt;h2&gt;
  
  
  Fewer Physical Infrastructure Requirements
&lt;/h2&gt;

&lt;p&gt;ETL tools serve as a sort of physical location for the steps between extracting data and loading it into repositories. In light of that, organizations that want to integrate data into target systems must purchase and maintain these tools to do so.&lt;/p&gt;

&lt;p&gt;ELTs, meanwhile, don't require that intermediate step, which means they require less physical infrastructure and specific resources. Instead, the target system's engine performs the transformation, rather than the engines native to ETL tools.&lt;/p&gt;

&lt;h2&gt;
  
  
  More Efficient Data Staging
&lt;/h2&gt;

&lt;p&gt;ETL tools cleanse data and prepare it for transformation. ELT tools, however, stage data &lt;em&gt;after&lt;/em&gt; loading it into a data warehouse, lake, or cloud storage solution. This makes the data staging process significantly more efficient and reduces latency across the board. Additionally, leading ELT tools make fewer demands on initial data sources and reduce the in-between steps associated with data processing.&lt;/p&gt;

&lt;h2&gt;
  
  
  Accelerated Time to Value (Sort of)
&lt;/h2&gt;

&lt;p&gt;Because they transform data within a target system, ELT tools speed up the time to value for teams. This allows data scientists and analysts working with big data to leverage and transform data quickly and to implement machine learning techniques for better analysis.&lt;/p&gt;

&lt;p&gt;ETL tools, on the other hand, require a manual coding process to ensure data conformity and uniformity, which adds time to the experience and increases latency across the board.&lt;/p&gt;

&lt;h2&gt;
  
  
  The Place of ELTs in the Modern Data Market: How ELTs Impact Data Warehousing
&lt;/h2&gt;

&lt;p&gt;One of the largest benefits of ELT systems is the way they improve both data lakes and warehouses. Regardless of which solution a team is using, ELT tools significantly reduce the time required to prepare data for use. By loading data into a data lake framework, ELTs allow organizations to take advantage of the processing engines within the solution when it comes time to stage and transform data.&lt;/p&gt;

&lt;p&gt;This serves a few distinct purposes. Besides providing immense scalability and leveraging parallel processing, it eliminates the requirement that organizations rely on conventional data modeling to unify their data.&lt;/p&gt;

&lt;p&gt;Here are a few of the other ways ELT solutions overhaul data warehousing:&lt;/p&gt;

&lt;h2&gt;
  
  
  Streamlined Architecture
&lt;/h2&gt;

&lt;p&gt;ELT tools streamline the process of preparing data for use. Because there is no in-between layer with built-in processing power limitations, the ELT can handle both data staging and transformation, which streamlines the experience for users.&lt;/p&gt;

&lt;h2&gt;
  
  
  Rapid Data Incorporation
&lt;/h2&gt;

&lt;p&gt;ELT solutions make it possible to incorporate data rapidly into both warehouses and lakes. With traditional methods, these sources can be difficult and clunky to use, leading to unnecessary latency and delays.&lt;/p&gt;

&lt;h2&gt;
  
  
  Some ELT Tools
&lt;/h2&gt;

&lt;p&gt;There are lots of options for ELTs. Honestly, ELTs are less about the tool and more about the method.&lt;/p&gt;

&lt;p&gt;However, many solutions that market themselves as ELTs is that they also often automated connectors that allow users to quickly develop their pipelines.&lt;/p&gt;

&lt;h2&gt;
  
  
  Fivetran
&lt;/h2&gt;

&lt;p&gt;&lt;a href="https://www.theseattledataguy.com/what-is-fivetran-and-why-you-should-use-it/"&gt;Fivetran&lt;/a&gt; is a highly comprehensive ELT tool that is becoming more popular every day. This tool allows efficient collection of customer data from related applications, websites, and servers. The data collected is then transferred to other tools for analytics, marketing, and warehousing purposes.&lt;/p&gt;

&lt;p&gt;Not only that, Fivetran has plenty of functionality. It has your typical source to destination connectors and it allows for both pushing and pulling of data. The pull connectors will pull from data sources in a variety of methods including ODBC, JDBC, and multiple API methods.&lt;/p&gt;

&lt;p&gt;Like many other ELT tools, Fivetran push connectors receive data that a source sends, or pushes, to them. In push connectors, such as &lt;a href="https://fivetran.com/docs/events/webhooks#webhooks"&gt;Webhooks&lt;/a&gt; or &lt;a href="https://fivetran.com/docs/events/snowplow#snowplow"&gt;Snowplow&lt;/a&gt;, source systems send data to Fivetran as events.&lt;/p&gt;

&lt;p&gt;Most importantly &lt;a href="https://fivetran.com/docs/transformations/basic-sql"&gt;Fivetran allows for different types of data transformations&lt;/a&gt;. Putting the T in ELT. They also allow for both scheduled and triggered transformations. Depending on the transformations you use, there is also other features like version control, email notification, and data validations.&lt;/p&gt;

&lt;h2&gt;
  
  
  Stitch..Sort Of..
&lt;/h2&gt;

&lt;p&gt;&lt;a href="https://www.stitchdata.com/"&gt;Stitch&lt;/a&gt; was developed to take a lot of the complexity out of ETLs and ELTs. One of the ways Stitch does this is by removing the need for data engineers to create pipelines that connect to APIs like in Salesforce and Zendesk.&lt;/p&gt;

&lt;p&gt;It also attaches to a lot of databases as well like MySQL. Having access to a broad set of API connectors is only one of the many benefits that makes Stitch easy to use.&lt;/p&gt;

&lt;p&gt;Stitch also removes a lot of the heavy lifting as far as setting up cron jobs for when the task should run as well as manages a lot of logging and monitoring. ETL frameworks like Airflow do offer some similar features. However, these features are much less straightforward in tools like Airflow and Luigi.&lt;/p&gt;

&lt;p&gt;Stitch is done nearly entirely with a GUI. This can make this a more approachable option for non-data engineers. It does allow you to add rules and set times when your ETLs will run.&lt;/p&gt;

&lt;h2&gt;
  
  
  Airbyte
&lt;/h2&gt;

&lt;p&gt;&lt;a href="https://airbyte.io/"&gt;Airbyte&lt;/a&gt; is a new open-source (MIT) EL+T platform that started in July 2020. It has a fast-growing community and it distinguishes itself by several significant choices:&lt;/p&gt;

&lt;p&gt;Airbyte's connectors are usable out of the box through a UI and an API, with monitoring, scheduling, and orchestration. Their ambition is to support 50+ connectors by EOY 2020. These connectors run as Docker containers so they can be built in the language of your choice. Airbyte components are also modular and you can decide to use subsets of the features to better fit in your data infrastructure (e.g., orchestration with Airflow or K8s or Airbyte's...)&lt;/p&gt;

&lt;p&gt;Similar to Fivetran, Airbyte integrates with DBT for the transformation piece, hence the EL+T. While contrary to Singer, Airbyte uses one single open-source repo to standardize and consolidate all developments from the community, leading to higher quality connectors. They built a compatibility layer with Singer so that Singer taps can run within Airbyte.&lt;/p&gt;

&lt;p&gt;Airbyte's goal is to commoditize ELT, by addressing the long tail of integrations. They aim to support 500+ connectors by the end of 2021 with the help of its community.&lt;/p&gt;

&lt;h2&gt;
  
  
  The Future of ELT
&lt;/h2&gt;

&lt;p&gt;Teams who need to accommodate the power, size, and speed of big data may search in vain for a solution that can help them achieve their motive. Fortunately, ELT is changing that dynamic. Designed to help teams forsake the traditional layers of data processing and transformation and modernize the approach, ELTs simplify both integration and architecture, decreasing latency and offering agile, enhanced performance.&lt;/p&gt;

&lt;p&gt;When compared to traditional ETL methods, it's clear that ELTs are the way of the future, as far as data processing is concerned. More sustainable, effective, and timely overall, ELT methods provide more flexibility and customization for organizations who want to control their data integration and implementation.&lt;/p&gt;

&lt;p&gt;By offering high speeds, rapid load times, and an invitingly low maintenance requirement, cloud-based ELT systems place the burden of transformation on the data destination, eliminating the need for data staging. This helps organizations enjoy a simpler relationship to their data, without sacrificing power.&lt;/p&gt;

&lt;p&gt;If we look into the future of data processing, it stands to reason that ELTs will rapidly become the de facto system for organizations focused on efficiency, scalability, and reliability. While both solutions have their strengths and weaknesses, the ELT has emerged as the undeniable favorite of many organizations around the globe.&lt;/p&gt;

&lt;p&gt;&lt;a href="https://www.youtube.com/watch?v=uEPCxBaRf6A"&gt;Data Engineer Vs Analytics Engineer Vs Analyst&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;&lt;a href="https://www.theseattledataguy.com/why-migrate-to-the-modern-data-stack-and-where-to-start/"&gt;Why Migrate To The Modern Data Stack And Where To Start&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;&lt;a href="http://youtube.com/watch?v=s40MptE20Tc&amp;amp;t=1s"&gt;5 Great Data Engineering Tools For 2021 -- My Favorite Data Engineering Tools&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;&lt;a href="https://www.youtube.com/watch?v=kSt9NV-qZkc&amp;amp;t=1s"&gt;4 SQL Tips For Data Scientists&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;&lt;a href="https://logitanalytics.com/what-are-the-benefits-cloud-data-warehousing-and-why-you-should-migrate/"&gt;What Are The Benefits Of Cloud Data Warehousing And Why You Should Migrate&lt;/a&gt;&lt;/p&gt;

</description>
      <category>datascience</category>
      <category>database</category>
      <category>bigdata</category>
      <category>startup</category>
    </item>
    <item>
      <title>Why Not to Become a Data Engineer</title>
      <dc:creator>SeattleDataGuy</dc:creator>
      <pubDate>Mon, 29 Nov 2021 14:16:24 +0000</pubDate>
      <link>https://dev.to/seattledataguy/why-not-to-become-a-data-engineer-44ke</link>
      <guid>https://dev.to/seattledataguy/why-not-to-become-a-data-engineer-44ke</guid>
      <description>&lt;p&gt;As companies struggle to manage their massive and complex data sets, the necessity for data engineers has become more apparent.&lt;/p&gt;

&lt;p&gt;Data engineering became the fastest-growing single job in 2019 with &lt;a href="https://seattledataguy.substack.com/p/are-companies-hiring-fewer-data-scientists"&gt;50 percent year-on-year growth&lt;/a&gt;, and there's little reason to believe demand for data engineers will slow soon. As with all careers, though, there are pros and cons to data engineering. Here's what you need to know about this up-and-coming job field and some of the reasons you may or may not want to pursue it.&lt;/p&gt;

&lt;h2&gt;
  
  
  What Is a Data Engineer?
&lt;/h2&gt;

&lt;p&gt;Data engineers move, remodel, and manage data sets from 10s if not 100s of internal company applications so analysts and data scientists don't need to spend their time constantly pulling data sets.&lt;/p&gt;

&lt;p&gt;They may also create a core layer of data that lets different data sources connect to it to get more information or context. Data engineers spend their time developing data pipelines, managing data warehouses and maintaining all the various infrastructure components they develop along the way.&lt;/p&gt;

&lt;p&gt;These specialists are usually the first people to handle data. They process the data so it's useful for everyone, not just the systems that store it.&lt;/p&gt;

&lt;p&gt;There are obvious reasons to become a data engineer --- like a high salary and numerous opportunities due to limited competition within the job market --- but we're not focusing on those today. Instead, let's ask the question, why not to become a data engineer.&lt;/p&gt;

&lt;h2&gt;
  
  
  What Skills Does a Data Engineer Need?
&lt;/h2&gt;

&lt;p&gt;To assist you as a new data engineer, I have created a &lt;a href="https://www.youtube.com/watch?v=LgSHaOvNodA"&gt;skill set pyramid&lt;/a&gt; that can be thought of as a hierarchy of skill set needs. This will help you focus on the skills you should learn first, allowing you to build a solid foundation as you move on to more specific skills. Just remember, the way you learn each step of the pyramid does not need to be overly rigid or stay in a strict order. You can layer each step, helping you progress as you learn. Let’s get started!&lt;/p&gt;

&lt;h2&gt;
  
  
  Reasons Not to Become a Data Engineer
&lt;/h2&gt;

&lt;p&gt;Despite being an in-demand career that promises high earnings and job security, becoming a data engineer isn't for everyone. As with most professions, it's important to consider your own skills, talents, and personality before choosing a career in data engineering. Here are some of the reasons you may not want to become a data engineer.&lt;/p&gt;

&lt;h2&gt;
  
  
  You're Letting Money Drive Your Career Choice
&lt;/h2&gt;

&lt;p&gt;While money is certainly important, it shouldn't be the driving force behind your career choice. Assuming you're planning to work in the tech field anyway, it's better to &lt;a href="https://www.youtube.com/watch?v=MjuQIdSwa3w"&gt;choose a role you will genuinely enjoy&lt;/a&gt;, even if the earnings could be a bit lower. A difference of $5,000 or even $10,000 in earnings won't drastically impact the lifestyle of a highly paid tech worker, especially once taxes are taken into account. The level of enjoyment you derive from your work, on the other hand, will affect your overall happiness and satisfaction in your professional life.&lt;/p&gt;

&lt;p&gt;The average data engineer earns &lt;a href="https://www.payscale.com/research/US/Job=Data_Engineer/Salary"&gt;$92,650 per year&lt;/a&gt;, which is significantly above the &lt;a href="https://www.jobted.com/salary"&gt;overall US average of $53,490&lt;/a&gt;. The financial benefits of becoming a data engineer, however, become much less clear when compared to other jobs in the tech field. The average software engineer, for example, can expect to make &lt;a href="https://www.payscale.com/research/US/Job=Software_Engineer/Salary"&gt;about $87,690&lt;/a&gt;. As you can see, the difference between a software engineer's salary and a data engineer's salary is fairly negligible. If software engineering would be a more fulfilling job for you, the slightly higher average salary isn't worth going into data engineering.&lt;/p&gt;

&lt;h2&gt;
  
  
  You Don't Enjoy the Engineering Mindset
&lt;/h2&gt;

&lt;p&gt;Data engineering requires you to adopt and deploy an engineering mindset, which some people can find rather constrictive. Because data engineers often need to create pieces of infrastructure that other engineers can maintain in the future, they must &lt;a href="https://www.youtube.com/watch?v=qJWZbZ5NNJ8&amp;amp;t=102s"&gt;work within a strict set of rules and standards&lt;/a&gt;. These rules are extremely important but can also seem burdensome to those who prefer more creative freedom in their projects.&lt;/p&gt;

&lt;p&gt;This isn't to say, of course, that there aren't creative aspects to the engineering mindset. High-level problem solving, for instance, often requires engineers to develop creative solutions. Likewise, engineers use creative problem-solving skills to &lt;a href="https://www.engineeringpassion.com/developing-the-engineering-mindset/"&gt;continuously improve&lt;/a&gt; their projects. In order to be a successful data engineer, you'll need to be able to balance your creative impulses with the rigorous mindset of an engineering professional.&lt;/p&gt;

&lt;h2&gt;
  
  
  You Aren't Flexible Enough
&lt;/h2&gt;

&lt;p&gt;One of the most interesting aspects of being a data engineer is the flexibility and lack of definition for the role. Because of its &lt;a href="https://www.youtube.com/watch?v=qJWZbZ5NNJ8&amp;amp;t=300s"&gt;highly interdisciplinary nature&lt;/a&gt;, data engineering combines elements of data analysis, programming, modeling, machine learning, and many other specific skills. Becoming a professional data engineer requires you to flexibly adapt and deploy these various skills as needed for the specific project you're working on.&lt;/p&gt;

&lt;p&gt;The downside of this interdisciplinary approach to data engineering is that it requires more flexibility than most other tech jobs. In your data engineering career, you may take on drastically different roles at different companies while maintaining the same job title. If you prefer to have a well-defined, set role, you likely won't enjoy the somewhat chaotic world of data engineering.&lt;/p&gt;

&lt;h2&gt;
  
  
  You Don't Enjoy Continuous Learning
&lt;/h2&gt;

&lt;p&gt;A final reason you may not want to become a data engineer is that you don't enjoy the process of continuous learning. Technologies are constantly shifting and evolving, requiring data engineers to update their skills on an ongoing basis. The&lt;a href="https://www.theseattledataguy.com/what-is-snowflakedb-and-why-you-should-use-it-for-your-cloud-data-warehouse/#page-content"&gt; cloud data warehouse tool Snowflake&lt;/a&gt;, for example, has seen &lt;a href="https://www.morningstar.com/articles/1020679/snowflake-shows-substantial-growth"&gt;substantial growth over the last 10 years&lt;/a&gt; as companies have embraced cloud computing. As trends like this emerge, data scientists must learn to use new tools and technologies to stay at the cutting edge.&lt;/p&gt;

&lt;p&gt;If one of your career goals is to eventually stop learning and rest on your laurels, data engineering won't be a good fit for you. Indeed, this is &lt;a href="https://www.youtube.com/watch?v=qJWZbZ5NNJ8&amp;amp;t=448s"&gt;true of almost all roles&lt;/a&gt; in the tech industry. Continuous learning is crucial for staying on top of trends and technologies, and even the most seasoned experts must pursue ongoing education to remain relevant. Failure to stay on top of new developments practically guarantees that your skills will eventually become outdated. While you might be able to work with older technologies at a handful of companies, your career options will narrow significantly when you stop updating your skills.&lt;/p&gt;

&lt;h2&gt;
  
  
  Conclusion: Know Yourself
&lt;/h2&gt;

&lt;p&gt;As you can see, knowing yourself and your preferences is essential to deciding whether a career in data engineering is right for you. In addition to knowing what you like to do in terms of specific tasks and working conditions, it's also important to consider your own personality. Data engineering is a suitable role for people who prefer to &lt;a href="https://www.youtube.com/watch?v=qJWZbZ5NNJ8&amp;amp;t=639s"&gt;work in the background&lt;/a&gt; within a company, rather than directly driving conversations with management using data insights. If you prefer that more extroverted side of data, though, you may enjoy a role as a data scientist.&lt;/p&gt;

&lt;p&gt;Overall, becoming a data engineer is a great career choice for people who love detail, following engineering guidelines, and building pipelines that allow raw data to be turned into actionable insights. As mentioned earlier, a career in data engineering also offers excellent earning potential and strong job security. With that said, the job isn't for everyone. If some of the reasons detailed above seem to describe you, it may be a good idea to give data engineering a second thought and explore other tech careers that could fit you better.&lt;/p&gt;

&lt;p&gt;If you want to read/watch more about data engineering, then check out the links below:&lt;/p&gt;

&lt;p&gt;&lt;a href="https://www.youtube.com/watch?v=uEPCxBaRf6A"&gt;Data Engineer Vs Analytics Engineer Vs Analyst&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;&lt;a href="https://www.theseattledataguy.com/why-migrate-to-the-modern-data-stack-and-where-to-start/"&gt;Why Migrate To The Modern Data Stack And Where To Start&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;&lt;a href="http://youtube.com/watch?v=s40MptE20Tc&amp;amp;t=1s"&gt;5 Great Data Engineering Tools For 2021 -- My Favorite Data Engineering Tools&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;&lt;a href="https://www.youtube.com/watch?v=kSt9NV-qZkc&amp;amp;t=1s"&gt;4 SQL Tips For Data Scientists&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;&lt;a href="https://logitanalytics.com/what-are-the-benefits-cloud-data-warehousing-and-why-you-should-migrate/"&gt;What Are The Benefits Of Cloud Data Warehousing And Why You Should Migrate&lt;/a&gt;&lt;/p&gt;

</description>
      <category>database</category>
      <category>datascience</category>
      <category>career</category>
      <category>beginners</category>
    </item>
  </channel>
</rss>
