<?xml version="1.0" encoding="UTF-8"?>
<rss version="2.0" xmlns:atom="http://www.w3.org/2005/Atom" xmlns:dc="http://purl.org/dc/elements/1.1/">
  <channel>
    <title>DEV Community: Samwel Wachira </title>
    <description>The latest articles on DEV Community by Samwel Wachira  (@sw1).</description>
    <link>https://dev.to/sw1</link>
    <image>
      <url>https://media2.dev.to/dynamic/image/width=90,height=90,fit=cover,gravity=auto,format=auto/https:%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Fuser%2Fprofile_image%2F903392%2Faaf6185d-8b1d-4fd8-9011-847fc0633dce.png</url>
      <title>DEV Community: Samwel Wachira </title>
      <link>https://dev.to/sw1</link>
    </image>
    <atom:link rel="self" type="application/rss+xml" href="https://dev.to/feed/sw1"/>
    <language>en</language>
    <item>
      <title>ULTIMATE GUIDE TO DATA ANALYST</title>
      <dc:creator>Samwel Wachira </dc:creator>
      <pubDate>Sat, 31 Aug 2024 08:21:33 +0000</pubDate>
      <link>https://dev.to/sw1/ultimate-guide-to-data-analyst-110k</link>
      <guid>https://dev.to/sw1/ultimate-guide-to-data-analyst-110k</guid>
      <description>&lt;blockquote&gt;
&lt;p&gt;&lt;em&gt;Data is the new oil -Igbo proverb&lt;/em&gt; &lt;/p&gt;
&lt;/blockquote&gt;

&lt;p&gt;Are you fascinated by the power of data to drive decision-making and solve complex problems? You're not alone. In our increasingly data-driven world, the role of a data analyst has never been more crucial. &lt;/p&gt;

&lt;p&gt;From healthcare and finance to marketing and sports, data analysts are the unsung heroes behind the scenes, turning raw data into actionable insights.&lt;/p&gt;

&lt;p&gt;If you want to pursue a career as a data analyst , the following steps will guide on what you need to take.&lt;/p&gt;

&lt;ol&gt;
&lt;li&gt;Programming Languages&lt;/li&gt;
&lt;li&gt;Data Visualization tools&lt;/li&gt;
&lt;li&gt;Machine Learning basics.&lt;/li&gt;
&lt;li&gt;Data analyst projects.&lt;/li&gt;
&lt;li&gt;Conclusion&lt;/li&gt;
&lt;/ol&gt;

&lt;p&gt;In this article, we will take a look at each points in details, giving you everything you need to get started on your journey to Data Analyst.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;1. Programming Languages&lt;/strong&gt;&lt;br&gt;
Data analysts will usually work with several programming languages, i.e. there is no wrong or right choice. Essentially, you will need to master  SQL for querying and manipulating databases, but you will then need to choose between R and Python for your next programming language.&lt;/p&gt;

&lt;p&gt;Learning R or Python, you will able to discover some of the libraries such as Pandas, NumPy that can help you with various tasks and grow your programming skills. Using some of the libraries, you will be learning how to import, clean, manipulate and visualize data with your preferred programming language.&lt;/p&gt;

&lt;p&gt;&lt;a href="https://media.dev.to/cdn-cgi/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fvo79vf3tavhqy4u1d3il.jpeg" class="article-body-image-wrapper"&gt;&lt;img src="https://media.dev.to/cdn-cgi/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fvo79vf3tavhqy4u1d3il.jpeg" alt="Image description" width="321" height="157"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;2.  Data Visualization tools&lt;/strong&gt;&lt;br&gt;
Consuming large sets of data isn't always straightforward. Sometimes, data sets are so large that it's downright impossible to discern anything useful from them. That's where data visualization comes in.&lt;/p&gt;

&lt;p&gt;Data visualization tools provide data visualization designers with an easier way to create visual representation of large data sets.&lt;/p&gt;

&lt;p&gt;The data visualizations can be used for variety of purposes: dashboards, annual reports, sales and marketing materials, investor slide decks and anywhere else information need to be interpreted immediately.&lt;/p&gt;

&lt;p&gt;The most common visualization tools used is Excel, Power BI, Tableau.&lt;/p&gt;

&lt;p&gt;Other visualization tools include Infogram, chartblocks, data wrapper, D3.js, Google charts and fusion charts.&lt;/p&gt;

&lt;p&gt;&lt;a href="https://media.dev.to/cdn-cgi/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2F3s7ntl72ufetuhrqe9jy.jpeg" class="article-body-image-wrapper"&gt;&lt;img src="https://media.dev.to/cdn-cgi/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2F3s7ntl72ufetuhrqe9jy.jpeg" alt="Image description" width="268" height="188"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;3. Machine Learning Basics&lt;/strong&gt;&lt;br&gt;
This involves mastering fundamentals of statistics which covers the topics such as probability distributions, hypothesis testing, measure of central tendency.&lt;br&gt;
Key components include feature engineering, data visualization and encoding &lt;/p&gt;

&lt;p&gt;&lt;a href="https://media.dev.to/cdn-cgi/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fafny1oo7fbbki07mbrfk.jpeg" class="article-body-image-wrapper"&gt;&lt;img src="https://media.dev.to/cdn-cgi/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fafny1oo7fbbki07mbrfk.jpeg" alt="Image description" width="287" height="176"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;4.Data Analyst Projects&lt;/strong&gt;&lt;br&gt;
Once you have mastered some of the basic essential skills, you will need to start developing them on your own by working on individual projects.&lt;/p&gt;

&lt;p&gt;However, you will need to undertake individual projects where everything will be your responsibility: selecting the topic, fetching the necessary data, contemplating the direction of your research, designing the project structure, making and checking the hypothesis, effectively communicating your findings and laying out the way forward.&lt;/p&gt;

&lt;p&gt;Practicing your skills and solving real-world problems will give you a solid basis for your future work experience. As a result, projects usually take much more time but they will help you stand out.&lt;/p&gt;

&lt;p&gt;&lt;a href="https://media.dev.to/cdn-cgi/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Ftff6bxe1987irzq1tk4d.jpeg" class="article-body-image-wrapper"&gt;&lt;img src="https://media.dev.to/cdn-cgi/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Ftff6bxe1987irzq1tk4d.jpeg" alt="Image description" width="290" height="174"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;5. Conclusion&lt;/strong&gt;&lt;br&gt;
By now, you should know how to become a data analyst and what you need to do to make your career goal a reality. &lt;/p&gt;

&lt;p&gt;However, to get in front of potential employers, you will need to have a portfolio of your work. Use your portfolio to make your passion and interests shine through.&lt;/p&gt;

&lt;p&gt;Ideally, demonstrate both your technical and soft skills in your portfolio and design it in your resume or cv. The more you develop your portfolio or resume, the more you can remove the broad-scope, common projects.&lt;/p&gt;

&lt;p&gt;Will all this information in hand, it's time for you to go ahead and start learning today.&lt;/p&gt;

&lt;p&gt;Happy learning, happy coding!&lt;/p&gt;

</description>
      <category>analytics</category>
      <category>datascience</category>
      <category>programming</category>
    </item>
    <item>
      <title>DATA ENGINEERING ROADMAP FOR BEGINNERS.</title>
      <dc:creator>Samwel Wachira </dc:creator>
      <pubDate>Thu, 09 Nov 2023 17:10:35 +0000</pubDate>
      <link>https://dev.to/sw1/data-engineering-roadmap-for-beginners-461a</link>
      <guid>https://dev.to/sw1/data-engineering-roadmap-for-beginners-461a</guid>
      <description>&lt;p&gt;&lt;a href="https://media.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fma654fwww3srv79phdue.jpg" class="article-body-image-wrapper"&gt;&lt;img src="https://media.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fma654fwww3srv79phdue.jpg" alt="Image description"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;As a growing data professional, you’ll probably find yourself developing an ETL solution at some point, and although you might be a Data Analyst/Scientist, the basics of Data Engineering are very key in career growth, mainly cause it combines SQL, Python, Machine Learning, DevOps &amp;amp; Data Visualisation skills.&lt;br&gt;
&lt;strong&gt;Simple Definition terms&lt;/strong&gt;&lt;br&gt;
 What is Analytics? : Analytics is defined as any action that converts data into insights.&lt;br&gt;
What is a data architecture?: A Data architecture is the structure that enables the storage, transformation, exploitation, and governance of data.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;Definition of Data Engineering&lt;/strong&gt;&lt;br&gt;
This is the practice of designing and building systems for collecting storing and analyzing data at scale. It is a broad field with applications in just about any industry.&lt;br&gt;
    Data Engineering is the complex task of making raw data usable to Data Scientists and groups within an organization.&lt;br&gt;
    These pipelines must take data from many disparate sources and collect them into a single warehouse that represents the data uniformly as a single source of truth.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;Role of a Data Engineer.&lt;/strong&gt;&lt;br&gt;
Data engineers design and build pipelines that transform and transport data into a format wherein, by the time it reaches the Data Scientists or other end users it is in a highly usable state.&lt;br&gt;
 Data Engineer build data pipelines because they want to get their data into a place from where the business can make, data-driven decisions.&lt;/p&gt;

&lt;p&gt;&lt;a href="https://media.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fr6m12rx4gcp5bsof25rq.png" class="article-body-image-wrapper"&gt;&lt;img src="https://media.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fr6m12rx4gcp5bsof25rq.png" alt="Image description"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;Data Lake&lt;/strong&gt;&lt;br&gt;
It generally describes a place where you can securely store various types of data of all scales for processing and analytics.&lt;br&gt;
    Data Lakes are typically used to drive Data Analytics, Data Science &amp;amp; ML workloads, or batch and streaming data pipelines.&lt;br&gt;
    Data Lakes will accept all types of data, they are portable, on premise or in the cloud.&lt;br&gt;
    The first line of defense in enterprise data is your data lake. It’s the central, give me whatever data at whatever volume, variety, formats, and velocity you got. I can take it.&lt;br&gt;
    Big data, Data Science, and Analytics supporting data-­driven decision-making promise to bring unprecedented levels of insight and efficiency to everything from how we work with data to how we work with customers to the search for a cure for cancer — but data science and analytics depend on having access to historical data.&lt;br&gt;
    In recognition of this, companies are deploying Big Data Lakes to bring all their data together in one place and start saving history, so data scientists and analysts have access to the information they need to enable data­-driven decision making.&lt;/p&gt;

&lt;p&gt;A common term in data engineering, is the concept of a data lake.&lt;br&gt;
    A data lake brings together data from across the enterprise to a single location, you might get the data from a relational database or from a spreadsheet and store the raw data in a data lake.&lt;br&gt;
    One option for this single location to store the raw data is to store it in a cloud storage bucket. S3 bucket.&lt;br&gt;
    A data lake is a medium where an organization brings all their data together in one place and starts saving history, so data scientists and analysts have access to the information they need to enable data­ driven decision-making.&lt;br&gt;
    The reason to make raw data available to analysts is so they can perform self ­service analytics. Self­ service has been an important mega-trend towards democratization of data.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;Datalakehouse&lt;/strong&gt;&lt;/p&gt;

&lt;div class="highlight js-code-highlight"&gt;
&lt;pre class="highlight plaintext"&gt;&lt;code&gt;🕯 Access to a company’s data (Self Service) gives rise to big challenge involving Governance and Data Security, Data Compliance; who has access privileges to company data.
&lt;/code&gt;&lt;/pre&gt;

&lt;/div&gt;

&lt;p&gt;&lt;strong&gt;Considerations&lt;/strong&gt;&lt;/p&gt;

&lt;div class="highlight js-code-highlight"&gt;
&lt;pre class="highlight plaintext"&gt;&lt;code&gt;Does your data lake handle all the data you have?
Can it all fit into a cloud storage bucket? If you have an RDMS; need to put the data in Cloud SQL (managed dB) rather than Cloud Storage.
Can it elastically scale to meet the demand? As your data collected increases, will you run out of disk space (on-premise problem)?
Does it support high throughput ingestion, what is the network bandwidth?
Do you have edge points of presence, is there fine-grained access control to objects? Do users need to seek within a file or is it enough to get a file as a whole?
Cloud Storage is blob storage, so you might need to think about the granularity of what you store.
Can other tools connect easily?

💡 The purpose of a Data Lake, is to make data accessible for analytics.
&lt;/code&gt;&lt;/pre&gt;

&lt;/div&gt;

&lt;p&gt;&lt;strong&gt;Data Warehousing&lt;/strong&gt;&lt;br&gt;
A data warehouse is a type of data management system that is designed to enable and support business intelligence (BI) activities, especially analytics.&lt;/p&gt;

&lt;p&gt;A typical data warehouse often includes the following elements:&lt;br&gt;
1.A relational database to store and manage data.&lt;br&gt;
 2.An extraction, loading, and transformation (ELT) solution for preparing the data for analysis.&lt;br&gt;
 3.Statistical analysis, reporting, and data mining capabilities.&lt;br&gt;
 4.Client analysis tools for visualizing and presenting data to business users.&lt;/p&gt;

&lt;p&gt;Other, more sophisticated analytical applications that generate actionable information by applying Data Science and artificial intelligence (AI) algorithms, or graph and spatial features that enable more kinds of analysis of data at scale.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;Data Lake Architecture&lt;/strong&gt;&lt;br&gt;
Data from external sources is loaded first into a raw or landing zone, where it is filed in folders that reflect its provenance (for instance, time and source) without further processing. Sometimes also called the raw or staging zone, is used to house raw ingested data.&lt;br&gt;
    Then, as appropriate, this data is copied into the gold zone, where it is cleansed, curated, and aggregated.&lt;br&gt;
    The gold zone frequently mirrors the landing area, but contains cleansed, enriched, and otherwise processed versions of the raw data.&lt;br&gt;
    This zone (gold) is sometimes also called prod to indicate that the data it contains is production-ready, or cleansed to indicate that the data has been run through data quality tools and/or a curation process to clean up (or cleanse) data quality problems.&lt;br&gt;
    The work zone, where users run their projects, or the sensitive zone, where data that should be protected is kept in encrypted volumes.&lt;br&gt;
    The sensitive zone is sometimes created to keep files containing data that is particularly important to protect from unauthorized viewers, whether because of regulatory requirements or business needs.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;Data Lakes vs Data Warehousing.&lt;/strong&gt;&lt;/p&gt;

&lt;p&gt;A data lake is a capture of every aspect of your business operation.&lt;/p&gt;

&lt;p&gt;The data is stored in its natural/raw format, usually as object blobs or files.&lt;/p&gt;

&lt;div class="highlight js-code-highlight"&gt;
&lt;pre class="highlight plaintext"&gt;&lt;code&gt;Retain all data in its native format.
Support all data types and all users.
Adapt to changes easily.
Tends to be application-specific.
&lt;/code&gt;&lt;/pre&gt;

&lt;/div&gt;

&lt;p&gt;How do you take this flexible &amp;amp; large amount of data and do something with it?&lt;/p&gt;

&lt;div class="highlight js-code-highlight"&gt;
&lt;pre class="highlight plaintext"&gt;&lt;code&gt;Typically loaded only after a use case is defined.
Provide faster insights.
Current/Historical data for reporting.
Tends to have a consistent schema shared across applications.
Take the raw data from a Data Lake &amp;amp; then Process → Organise → Transform then store it in a Data Warehouse.
&lt;/code&gt;&lt;/pre&gt;

&lt;/div&gt;

&lt;p&gt;Modernizing Data Lakes and Data Warehousing with Google Cloud.&lt;br&gt;
Google Cloud Storage Bucket.&lt;/p&gt;

&lt;div class="highlight js-code-highlight"&gt;
&lt;pre class="highlight plaintext"&gt;&lt;code&gt;A good option for staging all of your raw data in one place, before building transformation pipelines into your data warehouse.
Businesses use Cloud Storage as backup and archival storage for their businesses.
Storing Data in a Cloud Storage Bucket is durable and performant.
For a Data Engineer, you will often use a Cloud Storage Bucket as part of your Data Lake, to store many different raw data files, such as csv, json or avro.
You could then load or query them directly from Big Query; as a data warehouse.
&lt;/code&gt;&lt;/pre&gt;

&lt;/div&gt;

&lt;p&gt;Real-Time Analytics live data.&lt;/p&gt;

&lt;div class="highlight js-code-highlight"&gt;
&lt;pre class="highlight plaintext"&gt;&lt;code&gt;Batch pipelines are not enough, what if you need real-time analytics on data that arrives continuously &amp;amp; endlessly.
In that case, you might receive the data in Pub Sub, transform it using Data Flow &amp;amp; Stream it using Big Query.
&lt;/code&gt;&lt;/pre&gt;

&lt;/div&gt;

&lt;p&gt;Real-Time Streaming&lt;/p&gt;

&lt;div class="highlight js-code-highlight"&gt;
&lt;pre class="highlight plaintext"&gt;&lt;code&gt;Stream processing is a data management technique that involves ingesting a continuous data stream to quickly analyze, filter, transform or enhance the data in real-time.

🕯️ Data Pipelines largely perform the cleanup &amp;amp; processing of data. They are responsible for doing the transformations and processing of your data at scale &amp;amp; bring entire systems to life with freshly, newly processed data for analysis.
&lt;/code&gt;&lt;/pre&gt;

&lt;/div&gt;

&lt;p&gt;&lt;strong&gt;Challenges of Data Engineers&lt;/strong&gt;&lt;/p&gt;

&lt;div class="highlight js-code-highlight"&gt;
&lt;pre class="highlight plaintext"&gt;&lt;code&gt;As a Data Engineer; you’ll usually encounter a few problems when building data pipelines.
You might find it difficult to access the data that you need. You might find that the data, even after you access it, doesn’t have the quality that’s required by the Analytics or Machine Learning Model.
If you plan to build a model &amp;amp; even if the data quality exists, you might find that the transformations require computational resources that might not be available to you.
Challenges around Query performance; being able to run queries &amp;amp; transformations with the computational resources that you have.
&lt;/code&gt;&lt;/pre&gt;

&lt;/div&gt;

&lt;p&gt;&lt;strong&gt;1.Data Access&lt;/strong&gt;&lt;/p&gt;

&lt;p&gt;What makes Data Access difficult?&lt;br&gt;
Data in many businesses is siloed by departments &amp;amp; each department creates its own transactional systems to support its own business processes.&lt;br&gt;
    So you might have operational systems that correspond to store systems, and have a different operating system maintained by your product warehouses that manage your inventory.&lt;br&gt;
    And have a marketing department that manages all the promotions given that you need to do an analytic query on.&lt;br&gt;
    Such as;&lt;br&gt;
You need to know how to combine data from the stores. From the promotions &amp;amp; from the inventory levels.&lt;br&gt;
And because all these are stored in separate systems, some that have restricted access, building an analytics system that uses all three of these data sets to answer an ad-hoc query from above could be very difficult.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;2.Data Accuracy and Quality.&lt;/strong&gt;&lt;/p&gt;

&lt;p&gt;The second challenge is that of cleaning, formatting, and getting the data ready for insights requires you to build ETL pipelines.&lt;br&gt;
    ETL pipelines are usually necessary to ensure data accuracy and quality.&lt;br&gt;
    ETL; get the raw data and transform it into a form with which you can actually carry out the necessary analysis.&lt;br&gt;
    The cleaned and transformed data are typically stored in a Data Warehouse (not a Data Lake).&lt;br&gt;
  A Data Warehouse is a consolidated place to store the data, and all the data are easily joinable and queryable.&lt;/p&gt;

&lt;p&gt;In a Data Lake; data is in a raw format, In a Data Warehouse; the data is stored in a way that makes it efficient to query.&lt;/p&gt;

&lt;p&gt;Because data becomes useful, only after you clean it up; you should assume that any raw data that you collect from source systems need to be cleaned and transformed.&lt;br&gt;
Transform it into a format that makes it efficient to query.&lt;br&gt;
ETL the data &amp;amp; store it in Data Warehouse.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;3.Availability of Computational Resources&lt;/strong&gt;.&lt;br&gt;
The availability of computational resources can be a challenge if you are on an on-premise system.&lt;br&gt;
Data Engineers would need to manage server and cluster capacity &amp;amp; make sure that enough capacity exists to carry out ETL jobs.&lt;br&gt;
 The problem is that the compute needed by these ETL jobs is not constant over time. Very Often, it varies week to week, depending on factors like holidays and promotional sales.&lt;br&gt;
    This means that when traffic is low, you are wasting money. And when traffic is high, your jobs are taking way too long.&lt;br&gt;
    Once your data is in a Data Warehouse, you need to optimize the queries your users are running, to make the most efficient use of your compute resources.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;Big Query — Google&lt;/strong&gt;&lt;br&gt;
BigQuery is a fully managed enterprise data warehouse that helps you manage and analyze your data with built-in features like machine learning, geospatial&lt;br&gt;
Data Lakes vs Data Warehousing  analysis, and business intelligence.&lt;br&gt;
 BigQuery’s server-less architecture lets you use SQL queries to answer your organization’s biggest questions with zero infrastructure management.&lt;br&gt;
    BigQuery’s scalable, distributed analysis engine lets you query terabytes in seconds and petabytes in minutes.&lt;/p&gt;

&lt;p&gt;Big Query is used as a Data Warehouse.&lt;/p&gt;

&lt;p&gt;Our operational systems, like our relational databases that store online orders, inventory &amp;amp; promotions, are our raw data sources on the left, note that this isn’t exhaustive.&lt;br&gt;
 You could have other source systems that are manual like csv. These upstream data sources get gathered together in a single consolidated position in our data lake, which is designed for durability and high availability.&lt;br&gt;
    Once in the Data Lake, the data often needs to be processed via transformations that then output the data into our data warehouse which is ready for use by downstream teams.&lt;/p&gt;

&lt;p&gt;Remember, this roadmap is a general guide, and the specific technologies and tools may evolve over time. Continuous learning and adaptability are key characteristics for a successful data engineer.&lt;/p&gt;

&lt;p&gt;Happy learning!&lt;/p&gt;

</description>
      <category>dataengineering</category>
      <category>database</category>
      <category>luxacademy</category>
      <category>beginners</category>
    </item>
    <item>
      <title>DATA SCIENCE ROADMAP 2023-2024 FOR BEGINNERS.</title>
      <dc:creator>Samwel Wachira </dc:creator>
      <pubDate>Sun, 01 Oct 2023 15:54:42 +0000</pubDate>
      <link>https://dev.to/sw1/data-science-roadmap-2023-2024-for-beginners-47c</link>
      <guid>https://dev.to/sw1/data-science-roadmap-2023-2024-for-beginners-47c</guid>
      <description>&lt;blockquote&gt;
&lt;p&gt;A journey  of thousand miles begins with a single step -Chinese Proverb&lt;/p&gt;
&lt;/blockquote&gt;

&lt;p&gt;Are you interested to become a Data Scientist? Take up this Data Science Roadmap Test and assess your level of understanding!&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;What does Data Scientist do?&lt;/strong&gt;&lt;br&gt;
Data scientist are responsible for extracting insights and knowledge  from data through a combination of statistical analysis,data mining,visualization and machine learning.&lt;br&gt;
They   work on complex problems and uses data to inform decision making.&lt;/p&gt;

&lt;p&gt;If you would like to learn data science,here is your general roadmap to guide your journey  on data science:&lt;/p&gt;

&lt;ol&gt;
&lt;li&gt;&lt;strong&gt;Programming skills&lt;/strong&gt;&lt;/li&gt;
&lt;/ol&gt;

&lt;p&gt;Gain proficiency in a programming language such as Python or R which is most commonly used in data science.&lt;br&gt;
Learn and understand fundamental basics such as data types,variables,data structures and algorithms.&lt;br&gt;
Gain deep understanding of SQL(Structure Query Language) to manipulate database.&lt;/p&gt;

&lt;p&gt;2.&lt;strong&gt;Machine Learning&lt;/strong&gt;&lt;/p&gt;

&lt;p&gt;Learn and understand fundamental concepts of machine learning algorithms such as supervised ,unsupervised and reinforced machine learning algorithms.&lt;/p&gt;

&lt;p&gt;3.&lt;strong&gt;Mathematics and statistics&lt;/strong&gt;&lt;/p&gt;

&lt;p&gt;Have a strong foundation in mathematics such as linear algebra,probability and calculus as well as statistics concept such as regression analysis and hypothesis testing.&lt;/p&gt;

&lt;p&gt;4.&lt;strong&gt;Data Visualization and manipulation&lt;/strong&gt;&lt;/p&gt;

&lt;p&gt;Learn and understand visualization tools such as Tableau,Power BI and its libraries to communicate insights effectively.&lt;/p&gt;

&lt;p&gt;5.&lt;strong&gt;Data preparation and gathering&lt;/strong&gt;&lt;br&gt;
Learn and understand ETL techniques i.e(Extract,Transform and Load) for data processing.&lt;br&gt;
learn and understand data collection from various sources such as web scrapping,data pipelines,API's and databases.&lt;/p&gt;

&lt;p&gt;6.&lt;strong&gt;Model Building and Deployment&lt;/strong&gt;&lt;/p&gt;

&lt;p&gt;After learning machine learning,build machine learning models,validate them using appropriate techniques and deploy them using web frameworks such as Django,Flasks or FastAPI.&lt;/p&gt;

&lt;p&gt;7.&lt;strong&gt;Cloud Platform and Version Control&lt;/strong&gt;&lt;/p&gt;

&lt;p&gt;Explore to various cloud platforms such as Microsoft Azure,AWS and Google Cloud for scalable data  processing.&lt;br&gt;
Lastly,learn and understand how to use version control system such as Git to collaborate and track changes to your projects.&lt;/p&gt;

&lt;p&gt;Remember that the roadmap is flexible, and you can adapt it based on your goals, interests, and the specific requirements of the data science field you want to enter.&lt;/p&gt;

&lt;p&gt;Continuous learning,networking,collaboration with other data scientists and practical experience are key components of a successful data science career.&lt;/p&gt;

</description>
      <category>python</category>
      <category>datascience</category>
      <category>beginners</category>
    </item>
  </channel>
</rss>
