<?xml version="1.0" encoding="UTF-8"?>
<rss version="2.0" xmlns:atom="http://www.w3.org/2005/Atom" xmlns:dc="http://purl.org/dc/elements/1.1/">
  <channel>
    <title>DEV Community: Terer K. Gilbert</title>
    <description>The latest articles on DEV Community by Terer K. Gilbert (@kiprotichterer).</description>
    <link>https://dev.to/kiprotichterer</link>
    <image>
      <url>https://media2.dev.to/dynamic/image/width=90,height=90,fit=cover,gravity=auto,format=auto/https:%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Fuser%2Fprofile_image%2F1024870%2Fcd1e5de6-3334-40da-a46e-50544e421e64.jpg</url>
      <title>DEV Community: Terer K. Gilbert</title>
      <link>https://dev.to/kiprotichterer</link>
    </image>
    <atom:link rel="self" type="application/rss+xml" href="https://dev.to/feed/kiprotichterer"/>
    <language>en</language>
    <item>
      <title>Extracting Data from an API using Python (requests)</title>
      <dc:creator>Terer K. Gilbert</dc:creator>
      <pubDate>Tue, 20 May 2025 16:18:27 +0000</pubDate>
      <link>https://dev.to/kiprotichterer/extracting-data-from-an-api-using-python-requests-4h5</link>
      <guid>https://dev.to/kiprotichterer/extracting-data-from-an-api-using-python-requests-4h5</guid>
      <description>&lt;p&gt;&lt;strong&gt;Introduction&lt;/strong&gt;&lt;br&gt;
In today's data-driven world, APIs(Application Programming Interfaces) play a central role in allowing applications and services to interact and share information. Whether you are building a data dashboard, training machine learning models, or automating reporting process, chances are you will need to retrieve data from an API at some point.&lt;/p&gt;

&lt;p&gt;This article provides a guide to extract data from an API using Python's &lt;em&gt;requests&lt;/em&gt; library - a popular and de facto standard for making HTTP requests in python. We'll explore what APIs are, how HTTP requests work, and then walk through step-by-step process of sending a requests and handling the response.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;What is an API?&lt;/strong&gt;&lt;br&gt;
An &lt;strong&gt;API&lt;/strong&gt; acts as a bridge between two systems, allowing them to communicate. In the context of web APIs, it typically involves sending HTTP requests(like GET or POST) to a server and receiving responses in return. often in JSON format.&lt;br&gt;
APIs are commonly used to:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;&lt;p&gt;Fetch real-time data.&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;Interact with third party platforms&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;Perform actions remotely.&lt;/p&gt;&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;&lt;strong&gt;Getting Started With Python's Requests&lt;/strong&gt;&lt;br&gt;
Unlike most python's libraries, requests library is not included in python's standard library. Therefore, for you to use requests library, you have to install it so that you can use in your code. You can either install it into your virtual environment(good practice) or you can also install it into your global environment if you want to use it through multiple projects.&lt;/p&gt;

&lt;p&gt;To install &lt;em&gt;requests&lt;/em&gt;, run:&lt;br&gt;
&lt;/p&gt;

&lt;div class="highlight js-code-highlight"&gt;
&lt;pre class="highlight plaintext"&gt;&lt;code&gt;pip install requests
&lt;/code&gt;&lt;/pre&gt;

&lt;/div&gt;



&lt;p&gt;Once &lt;em&gt;pip&lt;/em&gt; has finished the process of installing &lt;em&gt;requests&lt;/em&gt;, then you cab use it in your application.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;Step-by-Step Guide: Extracting Data from an API&lt;/strong&gt;&lt;br&gt;
After installing requests into your environment, let's walk through how to extract data from RESTful API using &lt;em&gt;requests&lt;/em&gt;.&lt;/p&gt;

&lt;p&gt;1.&lt;strong&gt;Choose an API to work With&lt;/strong&gt; &lt;br&gt;
Before you can start extracting data, you need to choose API that provides the kind of information you are looking for. APIs are available in almost every field. Some APIs are open and free to use i.e you can send a request and gat data without any setup.&lt;br&gt;
Other APIs require you to register and use an API key -  a unique code that identifies you and your usage. This helps the API provider manage traffic and prevent abuse.&lt;/p&gt;

&lt;p&gt;2.&lt;strong&gt;Import the &lt;em&gt;requests&lt;/em&gt; Library&lt;/strong&gt;&lt;br&gt;
The process begin by importing the required library:&lt;br&gt;
&lt;/p&gt;

&lt;div class="highlight js-code-highlight"&gt;
&lt;pre class="highlight plaintext"&gt;&lt;code&gt;import requests
&lt;/code&gt;&lt;/pre&gt;

&lt;/div&gt;



&lt;ol&gt;
&lt;li&gt;
&lt;strong&gt;Make a GET Request&lt;/strong&gt;
HTTP methods, e.g GET and POST, determine which actions you are trying to perform when making an HTTP request.
One of the most common HTTP methods is GET. The GET method indicates that one is trying to get or retrieve data from a certain resource. To make a GET request using &lt;em&gt;requests&lt;/em&gt;, you can invoke &lt;em&gt;requests.get()&lt;/em&gt;.
For example, we can make a GET request to JSONPlaceholder by calling &lt;em&gt;get()&lt;/em&gt; with its URL as follows.
&lt;/li&gt;
&lt;/ol&gt;

&lt;div class="highlight js-code-highlight"&gt;
&lt;pre class="highlight plaintext"&gt;&lt;code&gt;url = '(https://jsonplaceholder.typicode.com/posts)'
requests.get(url)
&lt;/code&gt;&lt;/pre&gt;

&lt;/div&gt;



&lt;p&gt;At this point you have made a request.&lt;/p&gt;

&lt;ol&gt;
&lt;li&gt;
&lt;strong&gt;The Response&lt;/strong&gt;
A response is a powerful object that is used to inspect the results of the request made. Here were store the return value in a variable so that we can look at what we get.
&lt;/li&gt;
&lt;/ol&gt;

&lt;div class="highlight js-code-highlight"&gt;
&lt;pre class="highlight plaintext"&gt;&lt;code&gt;url = '(https://jsonplaceholder.typicode.com/posts)'
response = requests.get(url)
&lt;/code&gt;&lt;/pre&gt;

&lt;/div&gt;



&lt;p&gt;At this point we have stored the value of our request in a variable called &lt;em&gt;response&lt;/em&gt;. We will use &lt;em&gt;response&lt;/em&gt; to see the results of our GET request.&lt;/p&gt;

&lt;p&gt;5.&lt;strong&gt;Check the Response Status&lt;/strong&gt;&lt;br&gt;
The first set of information you can gather from response is the status code. A status code informs you of the status of the request.&lt;br&gt;
For example, a 200 OK status means that your request was successful, while a 404 NOT FOUND status means that the resource you were looking for wasn't found. There are other codes that can be displaced.&lt;br&gt;
This can be accessed by using .status_code:&lt;br&gt;
&lt;code&gt;response.status_code&lt;/code&gt;&lt;br&gt;
Sometimes we can make this to be part of our code so that we can automatically know whether our request is successful or not.&lt;br&gt;
&lt;/p&gt;

&lt;div class="highlight js-code-highlight"&gt;
&lt;pre class="highlight plaintext"&gt;&lt;code&gt;if response.status_code == 200:
    print("Request was successful!")
else:
    print(f"Failed to retrieve data. Status code: {response.status_code}")

&lt;/code&gt;&lt;/pre&gt;

&lt;/div&gt;



&lt;p&gt;6.&lt;strong&gt;Parse the JSON Response&lt;/strong&gt;&lt;/p&gt;

&lt;p&gt;After the request is successful, we can convert the response to JSON format using &lt;em&gt;.json()&lt;/em&gt;, so that we can use it in our python application. Most of the APIs return data in JSON format. This is a lightweight, human readable format that python can easily handle easily.&lt;br&gt;
To parse the JSON response, we use .json() method as follows:&lt;br&gt;
&lt;/p&gt;

&lt;div class="highlight js-code-highlight"&gt;
&lt;pre class="highlight plaintext"&gt;&lt;code&gt;data = response.json()

&lt;/code&gt;&lt;/pre&gt;

&lt;/div&gt;



&lt;p&gt;This converts the json string into a python data structure - usually a dictionary or a list of dictionaries. From this point we can print the data to inspect and further deal with the data.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;Conclusion&lt;/strong&gt; &lt;br&gt;
Working with APIs using Python's &lt;em&gt;requests&lt;/em&gt; library opens the door to vast amounts of dynamic, real-time data. To summarize:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;&lt;p&gt;Use &lt;em&gt;requests.get()&lt;/em&gt; to fetch data from an API.&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;Check the response status using .status_code.&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;Parse the returned data using .json().&lt;/p&gt;&lt;/li&gt;
&lt;/ul&gt;

</description>
      <category>python</category>
      <category>api</category>
      <category>dataengineering</category>
      <category>datascience</category>
    </item>
    <item>
      <title>Introduction to Data Version Control (DVC)</title>
      <dc:creator>Terer K. Gilbert</dc:creator>
      <pubDate>Wed, 29 Mar 2023 15:21:42 +0000</pubDate>
      <link>https://dev.to/kiprotichterer/introduction-to-data-version-control-dvc-k12</link>
      <guid>https://dev.to/kiprotichterer/introduction-to-data-version-control-dvc-k12</guid>
      <description>&lt;h2&gt;
  
  
  Data Version Control (DVC)
&lt;/h2&gt;

&lt;p&gt;Data Version Control (DVC) is an open-source version control tool specifically designed for machine learning (ML) projects. It allows data scientists and ML engineers to efficiently track changes in their data, code, and models throughout the development process. In this article, we will discuss what DVC is, how it works, and why it is essential for ML projects.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;What is Data Version Control (DVC)?&lt;/strong&gt;&lt;br&gt;
Data Version Control (DVC) is a tool that provides version control for data science and machine learning projects. It allows you to track the changes made to your data, code, and models over time, just like version control systems such as Git do for software development.&lt;/p&gt;

&lt;p&gt;DVC provides a simple command-line interface that allows you to manage your data versioning and collaborate with your team. DVC is designed to work with Git, so it can integrate seamlessly with existing Git repositories.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;How does DVC work?&lt;/strong&gt;&lt;/p&gt;

&lt;p&gt;DVC is based on a few core concepts that make it easy to use and understand. The first concept is data versioning. DVC tracks changes to your data by creating a version control system that stores all the changes made to your data. Each version is stored in a separate file, making it easy to compare different versions of your data.&lt;/p&gt;

&lt;p&gt;The second concept is data pipelines. A data pipeline is a set of steps that transform raw data into a form that can be used by ML models. DVC allows you to create and manage data pipelines, making it easy to track changes to your data processing code and ensure that your models are trained on the correct data.&lt;/p&gt;

&lt;p&gt;The third concept is model versioning. DVC allows you to track changes to your ML models by creating a version control system that stores all the changes made to your models. Each version is stored in a separate file, making it easy to compare different versions of your models.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;Why is DVC important for ML projects?&lt;/strong&gt;&lt;/p&gt;

&lt;p&gt;Data Version Control (DVC) is an essential tool for machine learning projects for several reasons. First, it provides version control for your data, code, and models, making it easy to track changes and collaborate with your team. Second, it allows you to create and manage data pipelines, ensuring that your models are trained on the correct data. Third, it allows you to track changes to your ML models, making it easy to compare different versions and understand how they are performing.&lt;/p&gt;

&lt;p&gt;In addition, DVC helps to improve the reproducibility of your ML experiments. By tracking changes to your data, code, and models, you can ensure that your results are reproducible, even if you make changes to your code or data.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;Conclusion&lt;/strong&gt;&lt;/p&gt;

&lt;p&gt;Data Version Control (DVC) is a powerful tool that provides version control for machine learning projects. It allows you to track changes to your data, code, and models, create and manage data pipelines, and track changes to your ML models. By using DVC, you can improve the reproducibility of your experiments and collaborate more effectively with your team.&lt;/p&gt;

</description>
      <category>datascience</category>
      <category>machinelearning</category>
      <category>dvc</category>
    </item>
    <item>
      <title>Getting Started With Sentiment Analysis</title>
      <dc:creator>Terer K. Gilbert</dc:creator>
      <pubDate>Thu, 23 Mar 2023 07:27:48 +0000</pubDate>
      <link>https://dev.to/kiprotichterer/getting-started-with-sentiment-analysis-4ad5</link>
      <guid>https://dev.to/kiprotichterer/getting-started-with-sentiment-analysis-4ad5</guid>
      <description>&lt;p&gt;&lt;strong&gt;Introduction to Sentiment Analysis&lt;/strong&gt;&lt;/p&gt;

&lt;p&gt;Sentiment analysis is a technique used to determine the emotional tone of a piece of text. It is commonly used in natural language processing (NLP) and is used in a variety of applications, including social media monitoring, customer feedback analysis, and brand reputation management.&lt;/p&gt;

&lt;p&gt;In this article, we will walk you through the basics of sentiment analysis, including how to get started with your own sentiment analysis project.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;Step 1: Choose a Programming Language and Framework&lt;/strong&gt;&lt;/p&gt;

&lt;p&gt;To begin your sentiment analysis project, you will first need to choose a programming language and a framework. Some of the most popular programming languages for sentiment analysis include Python, R, and Java.&lt;/p&gt;

&lt;p&gt;Python is a popular choice for sentiment analysis due to its simplicity, ease of use, and the availability of many useful libraries such as NLTK, spaCy, and TextBlob. R is another popular language for sentiment analysis due to its strong data analysis capabilities and the availability of many useful libraries such as tm, tidytext, and syuzhet.&lt;/p&gt;

&lt;p&gt;Once you have chosen your programming language, you will need to choose a framework or library to work with. There are many frameworks and libraries available for sentiment analysis, including NLTK, TextBlob, VADER, and spaCy. Each framework has its own strengths and weaknesses, so it's important to choose the one that best fits your project's needs.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;Step 2: Collect Data&lt;/strong&gt;&lt;/p&gt;

&lt;p&gt;Once you have chosen your programming language and framework, you will need to collect data for your sentiment analysis project. There are many sources of data that can be used for sentiment analysis, including social media, customer reviews, news articles, and blogs.&lt;/p&gt;

&lt;p&gt;It's important to ensure that your data is clean and well-formatted before you begin your analysis. This may involve removing duplicates, correcting spelling errors, and removing irrelevant or non-textual data.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;Step 3: Preprocess the Data&lt;/strong&gt;&lt;/p&gt;

&lt;p&gt;Before you can begin your sentiment analysis, you will need to preprocess your data. This involves cleaning and transforming your data to make it easier to work with. Preprocessing techniques may include tokenization, stemming, and stopword removal.&lt;/p&gt;

&lt;p&gt;Tokenization is the process of breaking your text data into individual words or phrases, known as tokens. Stemming is the process of reducing words to their base form, such as converting "running" to "run". Stopword removal involves removing common words such as "the", "a", and "an" from your text data, as they do not typically provide much meaning.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;Step 4: Perform Sentiment Analysis&lt;/strong&gt;&lt;/p&gt;

&lt;p&gt;Once you have preprocessed your data, you can begin your sentiment analysis. There are many techniques and algorithms that can be used for sentiment analysis, including rule-based approaches, machine learning approaches, and hybrid approaches.&lt;/p&gt;

&lt;p&gt;Rule-based approaches involve defining a set of rules or heuristics that can be used to classify text as positive, negative, or neutral. Machine learning approaches involve training a model on a labeled dataset of text and sentiment values, and using the model to predict sentiment values for new text data. Hybrid approaches combine elements of both rule-based and machine learning approaches.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;Step 5: Evaluate and Refine Your Model&lt;/strong&gt;&lt;/p&gt;

&lt;p&gt;After you have performed your sentiment analysis, you will need to evaluate and refine your model. This may involve comparing your model's predicted sentiment values to known ground truth values, or using techniques such as cross-validation or holdout testing to assess the accuracy of your model.&lt;/p&gt;

&lt;p&gt;If your model's accuracy is not satisfactory, you may need to refine your model by adjusting its parameters or using a different algorithm or approach.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;Conclusion&lt;/strong&gt;&lt;/p&gt;

&lt;p&gt;Sentiment analysis is a powerful technique for analyzing the emotional tone of text data. By following the steps outlined in this article&lt;/p&gt;

</description>
      <category>python</category>
      <category>machinelearning</category>
    </item>
    <item>
      <title>Essential SQL Commands For Data Science.</title>
      <dc:creator>Terer K. Gilbert</dc:creator>
      <pubDate>Tue, 14 Mar 2023 06:03:52 +0000</pubDate>
      <link>https://dev.to/kiprotichterer/essential-sql-commands-for-data-science-4ape</link>
      <guid>https://dev.to/kiprotichterer/essential-sql-commands-for-data-science-4ape</guid>
      <description>&lt;p&gt;Structured Query Language (SQL) is a programming language used to manage and manipulate data stored in relational database management systems (RDBMS). As a data scientist, mastering SQL is an essential skill to extract, transform and load data from databases. In this article, we'll cover some of the essential SQL commands for data science.&lt;/p&gt;

&lt;p&gt;SELECT&lt;br&gt;
The SELECT command is used to retrieve data from one or more tables. It is the most commonly used command in SQL. The basic syntax for SELECT command is:&lt;br&gt;
             SELECT column1, column2, …&lt;br&gt;
             FROM table_name;&lt;/p&gt;

&lt;p&gt;For example, to retrieve all the data from a table called "customers," you would use the following command:&lt;/p&gt;

&lt;div class="highlight js-code-highlight"&gt;
&lt;pre class="highlight plaintext"&gt;&lt;code&gt;           SELECT *
           FROM customers;
&lt;/code&gt;&lt;/pre&gt;

&lt;/div&gt;

&lt;p&gt;WHERE&lt;br&gt;
The WHERE command is used to filter data based on certain conditions. The basic syntax for WHERE command is:&lt;br&gt;
               SELECT column1, column2, …&lt;br&gt;
               FROM table_name&lt;br&gt;
               WHERE condition;&lt;/p&gt;

&lt;p&gt;For example, to retrieve all the data from a table called "customers" where the country is 'USA,' you would use the following command:&lt;/p&gt;

&lt;div class="highlight js-code-highlight"&gt;
&lt;pre class="highlight plaintext"&gt;&lt;code&gt;          SELECT *
          FROM customers
          WHERE country = 'USA';
&lt;/code&gt;&lt;/pre&gt;

&lt;/div&gt;

&lt;p&gt;GROUP BY&lt;br&gt;
The GROUP BY command is used to group the result set based on one or more columns. The basic syntax for GROUP BY command is:&lt;br&gt;
              SELECT column1, column2, …, &lt;br&gt;
              FROM table_name&lt;br&gt;
              GROUP BY column1, column2, …;&lt;/p&gt;

&lt;p&gt;For example, to retrieve the count of customers by country from a table called "customers," you would use the following command:&lt;/p&gt;

&lt;div class="highlight js-code-highlight"&gt;
&lt;pre class="highlight plaintext"&gt;&lt;code&gt;         SELECT country, COUNT(*)
         FROM customers
         GROUP BY country;
&lt;/code&gt;&lt;/pre&gt;

&lt;/div&gt;

&lt;p&gt;JOIN&lt;br&gt;
The JOIN command is used to combine two or more tables based on a related column. The basic syntax for JOIN command is:&lt;br&gt;
                SELECT column1, column2, …&lt;br&gt;
                FROM table1&lt;br&gt;
                JOIN table2&lt;br&gt;
                ON table1.column = table2.column;&lt;/p&gt;

&lt;p&gt;For example, to retrieve the customer information along with their order information from two tables called "customers" and "orders" where the common column is "customer_id," you would use the following command:&lt;/p&gt;

&lt;div class="highlight js-code-highlight"&gt;
&lt;pre class="highlight plaintext"&gt;&lt;code&gt;            SELECT *
            FROM customers
            JOIN orders
            ON customers.customer_id = orders.customer_id;
&lt;/code&gt;&lt;/pre&gt;

&lt;/div&gt;

&lt;p&gt;ORDER BY&lt;br&gt;
The ORDER BY command is used to sort the result set in ascending or descending order based on one or more columns. The basic syntax for ORDER BY command is:&lt;br&gt;
                 SELECT column1, column2, …&lt;br&gt;
                 FROM table_name&lt;br&gt;
                 ORDER BY column1 ASC|DESC, column2 ASC|DESC, …;&lt;/p&gt;

&lt;p&gt;For example, to retrieve the customer information from a table called "customers" sorted by the customer's name in ascending order, you would use the following command:&lt;/p&gt;

&lt;div class="highlight js-code-highlight"&gt;
&lt;pre class="highlight plaintext"&gt;&lt;code&gt;              SELECT *
              FROM customers
              ORDER BY customer_name ASC;
&lt;/code&gt;&lt;/pre&gt;

&lt;/div&gt;

&lt;p&gt;LIMIT&lt;br&gt;
The LIMIT command is used to limit the number of rows returned by the SELECT command. The basic syntax for LIMIT command is:&lt;br&gt;
                SELECT column1, column2, …&lt;br&gt;
                FROM table_name&lt;br&gt;
                LIMIT number_of_rows;&lt;/p&gt;

&lt;p&gt;For example, to retrieve the top 10 customers from a table called "customers," you would use the following command:&lt;/p&gt;

&lt;div class="highlight js-code-highlight"&gt;
&lt;pre class="highlight plaintext"&gt;&lt;code&gt;            SELECT *
            FROM customers
&lt;/code&gt;&lt;/pre&gt;

&lt;/div&gt;

&lt;p&gt;LIMIT 10;&lt;/p&gt;

&lt;p&gt;In conclusion, SQL is a powerful tool for data manipulation and is a must-have skill for data scientists. Understanding and mastering these essential SQL commands will allow data scientists to effectively retrieve and analyze data from relational databases&lt;/p&gt;

</description>
      <category>datascience</category>
      <category>sql</category>
    </item>
    <item>
      <title>Exploratory Data Analysis Ultimate Guide</title>
      <dc:creator>Terer K. Gilbert</dc:creator>
      <pubDate>Sat, 25 Feb 2023 15:43:20 +0000</pubDate>
      <link>https://dev.to/kiprotichterer/exploratory-data-analysis-ultimate-guide-mon</link>
      <guid>https://dev.to/kiprotichterer/exploratory-data-analysis-ultimate-guide-mon</guid>
      <description>&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2F5l6bfab5ilolvu142712.png" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2F5l6bfab5ilolvu142712.png" alt=" " width="800" height="600"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;Exploratory Data Analysis(&lt;em&gt;EDA&lt;/em&gt;) is one of the most important initial actions during data analysis process. This step helps one to understand the data that the have and make initial decisions that will make the process more easier and enable one to reach their specific goal. &lt;/p&gt;

&lt;p&gt;EDA is applied so that you can investigate the data for any anomalies which may affect your data analysis process. It will give the analyst  the overview of the data; distribution, null values and many more.&lt;/p&gt;

&lt;p&gt;The basic EDA process involve the following general steps;&lt;/p&gt;

&lt;p&gt;1.Loading data&lt;br&gt;
2.Viewing data&lt;br&gt;
3.Cleaning data&lt;br&gt;
4.Analyzing data&lt;/p&gt;

&lt;p&gt;1.&lt;em&gt;LOADING DATA&lt;/em&gt;&lt;br&gt;
This is an important step during EDA process. In this step, you find and upload your dataset into the platform you are going to perform the EDA. In this case python will be used as an example.&lt;br&gt;
              #import libraries&lt;br&gt;
              import os&lt;br&gt;
              import pandas as pd&lt;br&gt;
              import numpy as np&lt;br&gt;
              import matplotlib.pyplot as pt&lt;/p&gt;

&lt;div class="highlight js-code-highlight"&gt;
&lt;pre class="highlight plaintext"&gt;&lt;code&gt;          #define working directory
          os.chdir("C:/users/~~LENOVO/Desktop/bootcamp~~")

         #import dataset
         salary=pd.read_csv("ITSalarySurveyEU2020.csv")
&lt;/code&gt;&lt;/pre&gt;

&lt;/div&gt;

&lt;p&gt;In the above description, the libraries that will be used in the process have been defined and uploaded. As well, the location of the dataset is defined and then it has been uploaded successfully. This step will bring the dataset into the platform so that it can be accessed and EDA is performed.&lt;/p&gt;

&lt;p&gt;2.&lt;em&gt;VIEWING DATA&lt;/em&gt;&lt;br&gt;
In this step, you try to understand the overview of your dataset. This may include:&lt;br&gt;
The size of your dataset in terms of number of rows and columns.&lt;br&gt;
                   &lt;em&gt;#data info&lt;br&gt;
                   *salary.info()&lt;/em&gt;&lt;br&gt;
         &lt;br&gt;
          RangeIndex: 1253 entries, 0 to 1252&lt;br&gt;
          Data columns (total 23 columns)&lt;br&gt;
You may also check the first columns and rows of the dataset.&lt;br&gt;
                   &lt;em&gt;#top five obs&lt;br&gt;
                   *salary.head()&lt;/em&gt;&lt;br&gt;
        
        dataset. This will enable your to see the overview, what &lt;br&gt;
        the dataset contains.&amp;gt;&lt;br&gt;
You may also check the last rows and columns of your dataset.&lt;br&gt;
                  &lt;em&gt;#bottom five obs&lt;br&gt;
                  *salary.tail()&lt;/em&gt;&lt;br&gt;
         
         your dataset.&amp;gt;&lt;/p&gt;

&lt;p&gt;3.&lt;em&gt;CLEANING DATA&lt;/em&gt;&lt;br&gt;
This step involves finding and cleaning any anomalies in your dataset. This may include finding duplicate data, missing data and nulls in the dataset.&lt;br&gt;
        &lt;em&gt;Finding Duplicate data.&lt;/em&gt;&lt;br&gt;
                     &lt;em&gt;salary.duplicated().sum()&lt;/em&gt;&lt;br&gt;
    &lt;/p&gt;

&lt;p&gt;If you have duplicated data in your dataset, you can decide to delete the duplicated data as follows;&lt;br&gt;
                &lt;em&gt;salary.drop.duplicate(keep='first')&lt;/em&gt;&lt;br&gt;
 4.&lt;em&gt;ANALYZING DATA&lt;/em&gt;&lt;br&gt;
Once the dataset have been cleaned, its time to understand your dataset well. This involve exploring the dataset. Exploration may include the following:&lt;br&gt;
You may decide to explore a specify column in your dataset to look at the amount of data in that column, mean, standard deviation, datatype and even maximum.&lt;br&gt;
 Example.&lt;br&gt;
               &lt;em&gt;salary.Age.describe()&lt;/em&gt;&lt;br&gt;
count    1226.000000&lt;br&gt;
mean       32.509788&lt;br&gt;
std         5.663804&lt;br&gt;
min        20.000000&lt;br&gt;
25%        29.000000&lt;br&gt;
50%        32.000000&lt;br&gt;
75%        35.000000&lt;br&gt;
max        69.000000&lt;br&gt;
Name: Age, dtype: float64&lt;/p&gt;

&lt;div class="highlight js-code-highlight"&gt;
&lt;pre class="highlight plaintext"&gt;&lt;code&gt;          *salary.Gender.describe()*
&lt;/code&gt;&lt;/pre&gt;

&lt;/div&gt;

&lt;p&gt;count     1243&lt;br&gt;
unique       3&lt;br&gt;
top       Male&lt;br&gt;
freq      1049&lt;br&gt;
Name: Gender, dtype: object&lt;/p&gt;

&lt;div class="highlight js-code-highlight"&gt;
&lt;pre class="highlight plaintext"&gt;&lt;code&gt;         *salary.City.describe()*
&lt;/code&gt;&lt;/pre&gt;

&lt;/div&gt;

&lt;p&gt;count       1253&lt;br&gt;
unique       119&lt;br&gt;
top       Berlin&lt;br&gt;
freq         681&lt;br&gt;
Name: City, dtype: object&lt;/p&gt;

&lt;p&gt;After exploring the data, you can also can also plot graphs about the data. This will help you see if there is any outlier in your dataset.&lt;br&gt;
Example.&lt;br&gt;
      &lt;em&gt;salary.Age.plot()&lt;/em&gt;&lt;br&gt;
  &lt;/p&gt;

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2F1j4fam8qeb4youb9f7to.png" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2F1j4fam8qeb4youb9f7to.png" alt=" " width="543" height="413"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;This is therefore the easiest way to undertake EDA in a dataset.&lt;/p&gt;

</description>
    </item>
    <item>
      <title>SQL101:INTRODUCTION TO SQL FOR DATA ANALYSIS</title>
      <dc:creator>Terer K. Gilbert</dc:creator>
      <pubDate>Sun, 19 Feb 2023 14:57:01 +0000</pubDate>
      <link>https://dev.to/kiprotichterer/sql101introduction-to-sql-for-data-analysis-2f64</link>
      <guid>https://dev.to/kiprotichterer/sql101introduction-to-sql-for-data-analysis-2f64</guid>
      <description>&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Ftdjc7zbq0nqjc1fkbmqv.png" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Ftdjc7zbq0nqjc1fkbmqv.png" alt=" " width="330" height="153"&gt;&lt;/a&gt;&lt;br&gt;
SQL stands for &lt;em&gt;Structured Query Language&lt;/em&gt;. It is a language used in programming and used to access and manipulate data in databases. SQL was developed in 1970s, but over time it has become one of the most used query language&lt;br&gt;
&lt;strong&gt;&lt;em&gt;Why SQL?&lt;/em&gt;&lt;/strong&gt;&lt;br&gt;
The reason why SQL has become one of the most common language in programming and accessing data in databases are as follows;&lt;br&gt;
1.SQL is standard and has been well documented. Therefore, people can easily learn the language.&lt;br&gt;
2.SQL can as well analyze data of any size, be it small or large stacks of data, SQL can easily analyze them.&lt;br&gt;
3.The language is also simple and non-procedural.&lt;/p&gt;

&lt;p&gt;Some of the basic commands in SQL are as follows&lt;br&gt;
&lt;em&gt;SELECT&lt;/em&gt; this command extract data from a database.&lt;/p&gt;

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fabfkq4ta8btd61ik9qfk.png" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fabfkq4ta8btd61ik9qfk.png" alt=" " width="454" height="189"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;&lt;em&gt;DELETE&lt;/em&gt; this command delete data from a database&lt;/p&gt;

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fc76m4fqxis7fwnoxcl5v.png" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fc76m4fqxis7fwnoxcl5v.png" alt=" " width="800" height="844"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;&lt;em&gt;UPDATE&lt;/em&gt; this command update data in a database&lt;/p&gt;

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2F3ud21c93mvng5logeo9m.png" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2F3ud21c93mvng5logeo9m.png" alt=" " width="605" height="308"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;&lt;em&gt;INSERT INTO&lt;/em&gt; this command insert new data into a database&lt;/p&gt;

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fus2htqeuvpmdulezphq6.png" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fus2htqeuvpmdulezphq6.png" alt=" " width="800" height="382"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;&lt;em&gt;ALTER&lt;/em&gt; this command is used to modify a table.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;&lt;em&gt;SQL SYNTAX&lt;/em&gt;&lt;/strong&gt;&lt;br&gt;
Syntax is the structure of statement in computer language. Since SQL is a computer language, the are specify sets of rules that are followed when writing it. They include:&lt;br&gt;
1.SQL keywords can either be written in uppercase or lowercase, it is not case sensitive.&lt;br&gt;
2.SQL statement can be written in a single line or multiple lines.&lt;/p&gt;

&lt;p&gt;SQL can be used to access and manipulate data in all relational databases which include;&lt;/p&gt;

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Ffbsy4kk06oyc26yy63y1.png" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Ffbsy4kk06oyc26yy63y1.png" alt=" " width="735" height="600"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;-MySQL&lt;br&gt;
-Oracle&lt;br&gt;
-PostgreSQL&lt;br&gt;
-MSSQL .etc &lt;/p&gt;

&lt;p&gt;Therefore, SQL is a language that one should know in order to be successful as a data analyst, data scientist and even data engineer and many other careers. &lt;/p&gt;

</description>
      <category>sql</category>
      <category>datascience</category>
    </item>
  </channel>
</rss>
