<?xml version="1.0" encoding="UTF-8"?>
<rss version="2.0" xmlns:atom="http://www.w3.org/2005/Atom" xmlns:dc="http://purl.org/dc/elements/1.1/">
  <channel>
    <title>DEV Community: Hiswill Thompson</title>
    <description>The latest articles on DEV Community by Hiswill Thompson (@hiswill_thompson_e33923d1).</description>
    <link>https://dev.to/hiswill_thompson_e33923d1</link>
    <image>
      <url>https://media2.dev.to/dynamic/image/width=90,height=90,fit=cover,gravity=auto,format=auto/https:%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Fuser%2Fprofile_image%2F2363593%2F16b349f3-8b6a-494d-b171-de7d3454f9b6.png</url>
      <title>DEV Community: Hiswill Thompson</title>
      <link>https://dev.to/hiswill_thompson_e33923d1</link>
    </image>
    <atom:link rel="self" type="application/rss+xml" href="https://dev.to/feed/hiswill_thompson_e33923d1"/>
    <language>en</language>
    <item>
      <title>Introduction to Data lakes: The future of big data storage</title>
      <dc:creator>Hiswill Thompson</dc:creator>
      <pubDate>Sat, 14 Dec 2024 10:13:07 +0000</pubDate>
      <link>https://dev.to/hiswill_thompson_e33923d1/introduction-to-data-lakes-the-future-of-big-data-storage-357j</link>
      <guid>https://dev.to/hiswill_thompson_e33923d1/introduction-to-data-lakes-the-future-of-big-data-storage-357j</guid>
      <description>&lt;p&gt;Reflection 8&lt;/p&gt;

&lt;p&gt;Data lakes have emerged as a pivotal component in the realm of big data management, providing organizations with a flexible and efficient way to store and analyze vast amounts of information. Unlike traditional databases, which require data to be structured and organized in a specific format, data lakes allow for the storage of raw data in its native format. This capability is particularly beneficial in today's data-driven landscape, where organizations generate an overwhelming volume of structured, semi-structured, and unstructured data from various sources.&lt;/p&gt;

&lt;p&gt;The primary advantage of data lakes lies in their ability to accommodate diverse data types. Organizations can ingest everything from text documents and images to sensor data and social media posts without the need for immediate categorization. This flexibility not only simplifies the data collection process but also enables data scientists and analysts to explore and analyze information more freely. By having access to a comprehensive dataset, organizations can uncover valuable insights that might otherwise remain hidden in siloed systems.&lt;/p&gt;

&lt;p&gt;Moreover, data lakes play a crucial role in supporting advanced analytics and machine learning initiatives. With the ability to store large volumes of data, organizations can train machine learning algorithms more effectively, leading to improved predictions and decision-making. The integration of various data sources within a single platform allows for a more holistic view of organizational data, empowering businesses to leverage their information for strategic advantage.&lt;/p&gt;

&lt;p&gt;In conclusion, data lakes represent a transformative approach to big data management. By providing a scalable and flexible storage solution, they enable organizations to harness the full potential of their data, driving innovation and fostering growth in an increasingly competitive landscape. As the demand for data-driven insights continues to rise, understanding the role of data lakes will be essential for organizations looking to thrive in the digital age..&lt;/p&gt;

</description>
      <category>bigdata</category>
      <category>dataengineering</category>
    </item>
    <item>
      <title>Introduction to Batch processing with Apache Spark</title>
      <dc:creator>Hiswill Thompson</dc:creator>
      <pubDate>Sat, 07 Dec 2024 08:06:50 +0000</pubDate>
      <link>https://dev.to/hiswill_thompson_e33923d1/introduction-to-batch-processing-with-apache-spark-3i99</link>
      <guid>https://dev.to/hiswill_thompson_e33923d1/introduction-to-batch-processing-with-apache-spark-3i99</guid>
      <description>&lt;p&gt;Introduction to Batch processing with Apache spark.&lt;/p&gt;

&lt;p&gt;Intro&lt;br&gt;
In our previous article we talked about  real-time streaming with Apache Kafka.However we’ll be discussing about Batch Processing with Apache Spark.&lt;br&gt;
 Batch processing is an essential Data processing technique that involves the management or execution of vast volume of Data in groups or batches without user interaction.&lt;br&gt;
This methodology is most effective for managing extensive Data sets that needs to be processed at scheduled intervals making it ideal for task like the ETL process, warehousing reporting etc.&lt;br&gt;
Business organization leverage on these to improve efficiency, allowing for more complex computation to be performed on large Datasets.&lt;br&gt;
  Businesses rely on data driven decisions making this, the significance of batch processing data cannot be overstated .&lt;br&gt;
Business institutions can now efficiently analyze large volume of data, generate insight and make effective decisions thus, increasing productivity, competitive as well as keeping them on track interns of areas they can improve and things they’ve done better.&lt;/p&gt;

&lt;p&gt;Apache Kafka &lt;br&gt;
This is where Apache Sparks comes to play.&lt;br&gt;
  Apache spark is na open source, distributed processing system used for big data workloads. It utilizes it’s in-memory caching and optimized query execution for quick analytic processes of data of any size .&lt;br&gt;
Spark has revolutionized batch processing by providing a unified frame work that support both batch processing and real time processing.&lt;br&gt;
It ability to integrate with various data sources and it’s support for multiple programming languages makes it a versatile tool for data engineer in the course of streaming their data workflow.&lt;br&gt;
Apache spark in-memory processing ability significantly enhance performance enhancing faster data retrieval and analysis compared to traditional disk-based system.&lt;/p&gt;

&lt;p&gt;Apache spark and Batch processing.&lt;br&gt;
Spark has unique features and architecture that support both real-time streaming and batch processing. However for the course of this article we are examining how spark handle vast volume of data in batches. It uses the following core concept among others &lt;/p&gt;

&lt;p&gt;1:: Resilient Distributed Datasets(RDDs):&lt;br&gt;
Apache spark uses RDDs as it essential data structure.&lt;br&gt;
RDDs are immutable collection of objects or data that can be processed in parallel across a cluster. The data collected are fixed and cannot be changed and they are processed by breaking it up into similar groups.&lt;br&gt;
When batch processing data is loaded into RDDs allowing Spark to efficiently manage and process vast datasets .&lt;/p&gt;

&lt;p&gt;2::Data ingestion:&lt;br&gt;
Batch processing typically involves reading data from various sources such as HDFs, S3 or local file.&lt;br&gt;
Spark can easily read data from these source in different formats e.g CSV, JSON etc.&lt;/p&gt;

&lt;p&gt;3:: Transformation:&lt;br&gt;
Applying a series of operations to transform raw data into a desired format.&lt;br&gt;
It provides a wide range of transformation query such as map, filter, reduceBykey .&lt;/p&gt;

&lt;p&gt;4:: Action:&lt;br&gt;
Triggering the execution of transformation to provide the final output.Examples include:- collect, count, SaveAsTextFile.&lt;/p&gt;

&lt;p&gt;WHY SPARK?!&lt;br&gt;
Apache spark is considered one of the best among others due to its unique architecture to manage vast volume of data at exceptional speed.&lt;br&gt;
Another advantage of spark is it ability to perform in-memory data processing which speed up the execution of data-intensive tasks unlike traditional disk-based processing framework.&lt;br&gt;
Additionally, Spark support various data sources and formats providing flexibility for Data integration.&lt;br&gt;
This versatility allows data engineers as analyst to work with diverse datasets without any need for extensive data transformation processes. &lt;br&gt;
Spark can easily connect to wide range of data storage such as HDFS, S3, Apache HBase etc.&lt;br&gt;
It also support multiple Programming language including Scala, Java, Python, R etc .&lt;/p&gt;

&lt;p&gt;Conclusion &lt;br&gt;
Batch processing is very crucial for data engineers because it helps them to process large volume of data in batches. By processing data in batches, data engineers can optimize resource usage reducing cost associated with continuous data processing.&lt;br&gt;
Also when data engineers process data in batches rather than real-time it allows to process vast volume of data at once.&lt;br&gt;
This is important for tasks like Data aggregation, Data warehousing and the ETL process.&lt;br&gt;
By Batch processing data, Data quality is maintained as well as data integrity.&lt;br&gt;
There is comprehensive validation and cleaning of data before it is made available for analysis when it is batched processed.&lt;br&gt;
However, all these are possible through powerful tools like Sparks which is able to improve data efficiency by in-memory processing, it’s versatility and cost-effectiveness.&lt;br&gt;
Even as you journey through your Data engineering campaign, it is pertinent to understand batch processing as well as have a knowledge on Tools like Apache spark which aid this process.&lt;/p&gt;

</description>
      <category>apachespark</category>
    </item>
    <item>
      <title>Introduction to Apache Kafka</title>
      <dc:creator>Hiswill Thompson</dc:creator>
      <pubDate>Thu, 05 Dec 2024 08:13:38 +0000</pubDate>
      <link>https://dev.to/hiswill_thompson_e33923d1/introduction-to-apache-kafka-250c</link>
      <guid>https://dev.to/hiswill_thompson_e33923d1/introduction-to-apache-kafka-250c</guid>
      <description>&lt;p&gt;Introduction to Apache Kafka :Building Real-Time Data Pipelines&lt;/p&gt;

&lt;p&gt;We live in a Data driven world where Data is key for decision making in business organizations . Apache Kafka among others is one of the tools which aids this process;in this real-time Data.&lt;/p&gt;

&lt;p&gt;Apache Kafka is an open-source distributed platform that enables the development of real- time Data, event driven application etc. It is designed in a way such that it can handle vast volume of Data , it is scalable and also user friendly.&lt;/p&gt;

&lt;p&gt;It is a distributed streaming platform that is equipped with architecture that enables the development of real- time Data pipelines . Due to its low latency and high processing rate, it is an ideal tool for real time streaming.&lt;/p&gt;

&lt;p&gt;CORE COMPONENTS OF APACHE KAFKA .&lt;/p&gt;

&lt;p&gt;Kafka has several components that facilitate its processes but here only the major components will be discussed.&lt;/p&gt;

&lt;p&gt;TOPIC:&lt;br&gt;
A topic is a particular stream of Data . It is similar to a table in a data set. A topic is identified by name and is split into partitions for easy reference. A topic also is use to organize massages .  Each partition contains as much topics as possible.&lt;/p&gt;

&lt;p&gt;PARTITION:&lt;br&gt;
Topics are organized into partitions .A partition is the smallest storage unit that holds a subset of records in a topic. Each partition is a single log file whose records are written in Read - only method. Once Data are written into a partition, it cannot be changed.&lt;br&gt;
Each message within a particular has an ID called an offset which helps to identify the start and end of a message( smallest unit of Kafka; an array of byte).&lt;br&gt;
With this , consumers can consume the message of their choice form a position by reading from a specific offset address .&lt;/p&gt;

&lt;p&gt;CONSUMER&lt;br&gt;
The Kafka consumer API(Application programming Interface) enables an application interact with the producer by subscribing to one or more Kafka topics in a partition.&lt;br&gt;
Also, it makes it possible for the processing of streams of records produced by those topics.&lt;/p&gt;

&lt;p&gt;PRODUCER&lt;br&gt;
Apache Kafka producer are client Apps publishing events to topic partition.&lt;br&gt;
It’s API enables an application to publish a stream of records to one or more topics.&lt;/p&gt;

&lt;p&gt;BROKERS:&lt;br&gt;
Kafka brokers manages the strongest of messages in the topic(s). A broker can be one or more. Each broker has a specific ID and contain certain partitions.&lt;/p&gt;

&lt;p&gt;Real world Application of Apache Kafka&lt;/p&gt;

&lt;p&gt;One of the pronounced reasons people use kafka is due to its versatility and robustness .&lt;br&gt;
From across all industries and business organizations, Abacha kafka is used ranging form E-commerce to sales to Telecommunications including other financial institutions.&lt;br&gt;
They all leverage the unique nature of Abacha kafka . It’s ability to handle vast volume of Data is a cherry on the cake for its users.&lt;/p&gt;

&lt;p&gt;1.. E- commerce and sales: &lt;br&gt;
The word of commerce is govern by Data . Only those who are able to manage Data survives. &lt;br&gt;
When sales and commerce becomes difficult due to inability to access Real time Data , Apache kafka comes in .&lt;br&gt;
It act as a powerful architecture to enable the streaming of real- time day for various Applications such as:&lt;br&gt;
:: Recommendation of product&lt;br&gt;
:: Managing customer complaint/request &lt;br&gt;
:: Ensuring prompt response to customer actions etc&lt;/p&gt;

&lt;p&gt;2:: Telecommunication:&lt;br&gt;
The industry of telecommunications is another organization that employs the services of Apache Kafka to manage large volume of Data.&lt;br&gt;
In the industry , Kafka facilitates :&lt;br&gt;
. Real time Data processing &lt;br&gt;
. Proactive monitoring &lt;br&gt;
. Event streaming etc.&lt;/p&gt;

&lt;p&gt;3:: Financial Services:&lt;br&gt;
Financial institutions have always been at the forefront of leveraging advanced technologies.&lt;br&gt;
Apache Kafka holds a key position in this field making real time Data and event streaming possible which is essential for expediting and Decision making process.&lt;br&gt;
Kafka also aids fraud detection .&lt;br&gt;
By enabling real time event and Data processing, it allows banks to analyze transactions as they occur and identify potential fraud in real time.&lt;br&gt;
Moreover, Kafka strengthen decision making in financial services by enabling real time Data streaming..&lt;br&gt;
Quick results and timely responses are critical in the fast-paced world of finance and Kafka ability in these areas makes it ideal for financial institutions.&lt;/p&gt;

&lt;p&gt;Conclusion:&lt;br&gt;
Kafkas efficiency and scalability is the main reason why it is chosen among others. It ability to tract financial records stream real time Data makes it one of the best tools around.&lt;br&gt;
Data engineer leverage on these architecture to manage real time Data.&lt;br&gt;
At the end of this article, I’ll provide links on how to install and set up Kafka environment and how to create a basic Pipeline with Kafka by setting up customer an producer.&lt;br&gt;
Catch your soon&lt;/p&gt;

&lt;p&gt;Recommendations:&lt;br&gt;
1..  &lt;a href="https://youtu.be/BwYFuhVhshI?si=0oHVwrADX175YfrX" rel="noopener noreferrer"&gt;https://youtu.be/BwYFuhVhshI?si=0oHVwrADX175YfrX&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;2::  &lt;a href="https://medium.com/@mustafaguc/building-kafka-producer-and-consumer-microservices-with-spring-boot-on-kubernetes-using-github-0bd0af37e538?source=user_profile_page---------3-------------62a570b50a3a---------------" rel="noopener noreferrer"&gt;https://medium.com/@mustafaguc/building-kafka-producer-and-consumer-microservices-with-spring-boot-on-kubernetes-using-github-0bd0af37e538?source=user_profile_page---------3-------------62a570b50a3a---------------&lt;/a&gt;&lt;/p&gt;

</description>
      <category>datapipeline</category>
      <category>dataengineering</category>
      <category>realtimedat</category>
    </item>
    <item>
      <title>Building Data pipelines</title>
      <dc:creator>Hiswill Thompson</dc:creator>
      <pubDate>Thu, 21 Nov 2024 17:50:47 +0000</pubDate>
      <link>https://dev.to/hiswill_thompson_e33923d1/building-data-pipelines-2d87</link>
      <guid>https://dev.to/hiswill_thompson_e33923d1/building-data-pipelines-2d87</guid>
      <description>&lt;p&gt;Building Data pipelines: A guide to Data flow automation in Data Engineering &lt;/p&gt;

&lt;p&gt;Intro/overview :&lt;br&gt;
Data is the most vital instrument of business organizations. Data when extracted from various Data sources are organized,manage and analyzed for Decision making .&lt;/p&gt;

&lt;p&gt;It is Data that reveals the weakness and strength of business organization and helps them to tighten their loss ends to stay competitive.&lt;br&gt;
 Thus, company sought after ways to transport Data from disparate sources, in order to analyze them for decision making process. This is where Data Pipelines comes to play.&lt;/p&gt;

&lt;p&gt;A Data pipeline is a channel of transporting Data from various sources to a destination e.g Data warehouse, Data lakes or any other type of Data repository.  During the process of transporting the Data, data management and optimization is done getting it set to a state where it can be use for analysis.&lt;/p&gt;

&lt;p&gt;Basically, they are three components of Data Pipelines:&lt;br&gt;
The Data Source:- where data is extracted.Examples Flat files, APIs, the Internet of things etc.&lt;/p&gt;

&lt;p&gt;Transformation: The act of streamlining Data, getting it ready for analysis.&lt;/p&gt;

&lt;p&gt;Destination:e.g  Data warehouse or Data lakes or any other repository where Data are stored.&lt;/p&gt;

&lt;p&gt;POPULAR DATA PIPELINES TOOLS&lt;br&gt;
In the automation or workflow of Data pipelines there are tools which makes it possible.&lt;/p&gt;

&lt;p&gt;Apache Airflow: This is an open source platform well-suited for automatic data pipelines. It harnesses the Directed Acyclic Graph (DAGs) to define workflows where each nodes represent a task, and edge which denote task dependencies .&lt;/p&gt;

&lt;p&gt;It enables the definition of task dependencies ensuring tasks are executed only when their dependencies are successfully completed.&lt;/p&gt;

&lt;p&gt;Airflow support dynamic workflows generation making it scalable and flexible. &lt;br&gt;
Moreover, it also integrates various Data storage systems , cloud services etc. This allows the data engineer to design pipelines across different platforms.&lt;/p&gt;

&lt;p&gt;Most excitingly, it provides a web based user interface for monitoring workflow, status and historical information this aids in debugging and to optimize performance. It can also distribute task across multiple workers making it suitable for dealing with large Datasets .&lt;/p&gt;

&lt;p&gt;Luigi: This is another popular Data pipeline tool . It is a workflow management system to launch a group of tasks with defined dependencies efficiently.&lt;/p&gt;

&lt;p&gt;It is an python based Application programming Interface (API) that Spotify developers to build and automate pipelines.&lt;br&gt;
Data engineers can utilize it to create workflows, manage complex Dara processing and also for Data integration .&lt;/p&gt;

&lt;p&gt;Unlike Apache Airflows. Luigi doesn’t use DAG. Instead, it uses two building blocks .&lt;br&gt;
Target and Task.&lt;br&gt;
Task are the basic unit of work in the pipeline. A task is said to be completed when it reaches it target.&lt;br&gt;
Target can be a result of a task or an input for another task&lt;/p&gt;

&lt;p&gt;Others examples of popular Data Pipeline Tools include:&lt;br&gt;
1.Perfect &lt;br&gt;
2.Talend &lt;br&gt;
3.AWS glue &lt;br&gt;
All are packed with architectural features that aids the automation process of pipeline &lt;br&gt;
Conclusion:&lt;br&gt;
Automating Data workflows aids in efficiency and scalability of data. Large Datasets can easily be integrated without significant manual effort.&lt;br&gt;
Moreover, it improve Data quality and integrity when Data is transformed.&lt;br&gt;
It also provide flexibility and adapting to changing data requirements and evolving business needs of organization. This in return, will enable companies to stay updated and respond to data changes quickly enabling effective business decision making.&lt;/p&gt;

&lt;p&gt;It would have been of my best interest to provide Hand-on experiment procedures for the automation process but I do not have what it takes at the moment to do so.&lt;/p&gt;

&lt;p&gt;However I’ll provide link at the end of the article that will help you with that.&lt;/p&gt;

&lt;p&gt;Happy Reading !!!&lt;br&gt;
See ya soon.&lt;/p&gt;

&lt;p&gt;Recommendations:&lt;br&gt;
&lt;a href="https://youtu.be/XItgkYxpOt4?si=F-YBEUu3vkUQqOR7" rel="noopener noreferrer"&gt;https://youtu.be/XItgkYxpOt4?si=F-YBEUu3vkUQqOR7&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;&lt;a href="https://medium.com/gooddata-developers/how-to-build-a-modern-data-pipeline-cfdd9d14fbea?source=user_profile_page---------20-------------99a1772125bd---------------" rel="noopener noreferrer"&gt;https://medium.com/gooddata-developers/how-to-build-a-modern-data-pipeline-cfdd9d14fbea?source=user_profile_page---------20-------------99a1772125bd---------------&lt;/a&gt;&lt;/p&gt;

</description>
    </item>
    <item>
      <title>Introduction to SQL for Data Engineering : writing Basic queries</title>
      <dc:creator>Hiswill Thompson</dc:creator>
      <pubDate>Sat, 09 Nov 2024 11:55:52 +0000</pubDate>
      <link>https://dev.to/hiswill_thompson_e33923d1/introduction-to-sql-for-data-engineering-writing-basic-queries-4n4f</link>
      <guid>https://dev.to/hiswill_thompson_e33923d1/introduction-to-sql-for-data-engineering-writing-basic-queries-4n4f</guid>
      <description>&lt;p&gt;Introduction to SQL for Data Engineering: write Basic queries.&lt;/p&gt;

&lt;p&gt;Intro:&lt;br&gt;
Hey guys here’s another article for you as you journey through your career as an aspiring Data Engineer . If you’ve been out on my other articles, do well to check them out in my profile. Like , comment, and share it to your friends as those article will be handful in your Data Engineering career.&lt;/p&gt;

&lt;p&gt;SQL overview:&lt;/p&gt;

&lt;p&gt;SQL (Structured Query Language) is a key in Data management,manipulation and organization . It is a fundamental skill in the field of data engineering.&lt;/p&gt;

&lt;p&gt;Moreover, SQL is an essential tool for data engineering because it help to query databases . What this means is that SQL allows its user to interact with Data .&lt;/p&gt;

&lt;p&gt;Why do we need to query Data after all?&lt;/p&gt;

&lt;p&gt;Data Retrieval:  With SQL, Data can be retrieved from a Data base efficiently . The SELECT statement helps to retrieve Data from a particular column.&lt;/p&gt;

&lt;p&gt;Data Transformation: You can use SQL to clean and transform data before it is use for analytical process . SQL can allow you to delete , join or update Data in a particular column .&lt;/p&gt;

&lt;p&gt;Data storage : Having and understanding of SQL is pertinent for managing Data repositories. SQL Data base can be use to store Data structured Data in Data pages.&lt;/p&gt;

&lt;p&gt;SQL is also used to ensure compatibility and easy Data integration .&lt;/p&gt;

&lt;p&gt;However, having had an overview of SQL and it’s importance I’m querying Data base, let’s look at some basic SQL commands.&lt;/p&gt;

&lt;p&gt;Simple SQL commands:&lt;br&gt;
1.SELECT:&lt;br&gt;
The SELECT statement is the most vital SQL operation. It’s function is to retrieve Data from Data base table . &lt;/p&gt;

&lt;p&gt;2.WHERE:&lt;br&gt;
The WHERE query allows you to filter the rows returned by a SELECT statement based on a specific condition (link will be provided at the end of the article for practical tutorial)&lt;/p&gt;

&lt;p&gt;3.JOIN &lt;br&gt;
The JOIN operation allows for combination of Data from diverse table . For instance, joining column 1 from table A and Column 3 form table B . Some examples of JION are INNER JOIN, LEFT JOIN RIGHT JOIN AND FULL JOIN .&lt;/p&gt;

&lt;p&gt;Some other examples of SQL commands include :&lt;br&gt;
INSERT, UPDATE, DELETE CREATE etc&lt;/p&gt;

&lt;p&gt;Conclusion:&lt;br&gt;
The importance of having SWL skill cannot be over emphasized for Data Engineering. Mastering SWL is very crucial interacting with large Data base and for ETL task. &lt;br&gt;
Happy learning!!!! And Enjoy your Data engineering journey. &lt;/p&gt;

&lt;p&gt;SEE YOU SOON!!&lt;/p&gt;

&lt;p&gt;Recommendations :&lt;br&gt;
&lt;a href="https://youtu.be/wgRwITQHszU?si=BrUePur6-YhGNZ0y" rel="noopener noreferrer"&gt;https://youtu.be/wgRwITQHszU?si=BrUePur6-YhGNZ0y&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;&lt;a href="https://gaurav-adarshi.medium.com/sql-a-comprehensive-guide-to-database-concepts-for-aspiring-data-engineers-7a62f7729a31" rel="noopener noreferrer"&gt;https://gaurav-adarshi.medium.com/sql-a-comprehensive-guide-to-database-concepts-for-aspiring-data-engineers-7a62f7729a31&lt;/a&gt;&lt;/p&gt;

</description>
    </item>
    <item>
      <title>Introduction to Data Engineering: setting up python for ETL</title>
      <dc:creator>Hiswill Thompson</dc:creator>
      <pubDate>Wed, 06 Nov 2024 13:11:30 +0000</pubDate>
      <link>https://dev.to/hiswill_thompson_e33923d1/introduction-to-data-engineering-setting-up-python-for-etl-eko</link>
      <guid>https://dev.to/hiswill_thompson_e33923d1/introduction-to-data-engineering-setting-up-python-for-etl-eko</guid>
      <description>&lt;p&gt;Introduction to Data Engineering: Setting up python for ETL &lt;/p&gt;

&lt;p&gt;Hey there!&lt;/p&gt;

&lt;p&gt;If you’ve missed my other articles on “introduction to Data Engineering “ and Understanding ETL pipeline” I’ll recommend you go check it out on my profile. As an aspiring Data engineer it will be handful in your journey of Data Engineering .&lt;/p&gt;

&lt;p&gt;I’m this article, we will be talking about how to set up python for ETL.  However, before then, let’s have an overview of what Data Engineering and the ETL process is all about . &lt;/p&gt;

&lt;p&gt;Data engineering is the process of building, maintaining and optimizing data. It involves the process of “gathering” information from various sources might be an APIs, flat files websites etc processing and streamlining in into a useful and meaningful information then making it available to user ; might be a Data scientist or a Data Analyst .&lt;/p&gt;

&lt;p&gt;The ETL process on the other hand is an integral apart of Data Engineering. It stands for Extract, Transform and Load. The Extraction process process basically is “fetching “  data for various sources or platforms .The TRANSFORM  process is the most vigorous process in ETL .&lt;br&gt;
Data gathered are streamlined into useful information and then stored in Data repositories for User accessibility; this is the LOAD stage.&lt;/p&gt;

&lt;p&gt;Having have a clue of what Data Engineering and what ETL is let’s look at what Python and the significance it has in the ETL process before we actually delve into setting up python for ETL . That’s fair enough isn’t it?&lt;/p&gt;

&lt;p&gt;Python is one of the most popular programming language. It is an open-source, high level and object-orientated programming language.&lt;br&gt;
It is simple, easy to -learn and readable .It is easy to understand and have user interface reason most IT expert choose it over other programming languages.&lt;br&gt;
Python has versatile and powerful features the helps Data Engineers in the ETL process. It possesses various libraries which fine- tune to Data Engineering needs.&lt;br&gt;
Examples includes; Pandas, Numpy, Apache Airflow , Scikit- learn, Beautiful Soup etc. &lt;/p&gt;

&lt;p&gt;This libraries has so much to do with the ETL process. It’s significant ranges from collection of Data from various sources, streamlining data, merging datasets, Data classification etc.&lt;br&gt;
For instance Pandas helps in extracting, processing and even loading datasets, Psyspark helps in working with large datasets and SQL Alchemy with its flexibility helps in database interaction .&lt;/p&gt;

&lt;p&gt;With this, let’s delve into how to set up python in your operating system in other to use it for your ETL operation.&lt;br&gt;
Below are the python installation Guide:&lt;/p&gt;

&lt;p&gt;1.open your favorite browser and search “Python download”&lt;br&gt;
2.Python original website will display; python.org&lt;br&gt;
3.Download the version of your choice; preferably the latest version.&lt;br&gt;
4.Install the one for your operating system (OS options will be displayed)&lt;br&gt;
5.Click Download.&lt;br&gt;
6.Then install; you can customize installation &lt;br&gt;
7.Tick the two boxes that will be displayed below it &lt;br&gt;
8.Use admin privately when installing Py.exe and add python .exe to PATH.&lt;br&gt;
9.Optional features will be displayed &lt;br&gt;
10.Click on next &lt;br&gt;
11.Advance setting will show , installation location will be show .&lt;br&gt;
12.Click installation and wait for successful installation &lt;/p&gt;

&lt;ol&gt;
&lt;li&gt;Click close or minimize an set up your python for you &lt;/li&gt;
&lt;/ol&gt;

&lt;p&gt;Conclusively, python plays an important role in Data engineering and will still be of great effect in the field of Data Engineering task like ETL . It is pertinent to learn it an instill it into your Data Engineering journey.&lt;br&gt;
Recommendation:&lt;br&gt;
&lt;a href="https://www.astera.com/type/blog/etl-using-python/" rel="noopener noreferrer"&gt;https://www.astera.com/type/blog/etl-using-python/&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;&lt;a href="https://medium.com/@godswillthompson16/understanding-etl-pipelines-extract-transform-load-in-data-engineering-814472d71646?source=user_profile_page---------1-------------d1624a597f9d---------------" rel="noopener noreferrer"&gt;https://medium.com/@godswillthompson16/understanding-etl-pipelines-extract-transform-load-in-data-engineering-814472d71646?source=user_profile_page---------1-------------d1624a597f9d---------------&lt;/a&gt;&lt;/p&gt;

</description>
    </item>
  </channel>
</rss>
