<?xml version="1.0" encoding="UTF-8"?>
<rss version="2.0" xmlns:atom="http://www.w3.org/2005/Atom" xmlns:dc="http://purl.org/dc/elements/1.1/">
  <channel>
    <title>DEV Community: Sheila</title>
    <description>The latest articles on DEV Community by Sheila (@s-jt).</description>
    <link>https://dev.to/s-jt</link>
    <image>
      <url>https://media2.dev.to/dynamic/image/width=90,height=90,fit=cover,gravity=auto,format=auto/https:%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Fuser%2Fprofile_image%2F1807995%2F483d6d0b-4327-4318-bde1-01944474b00d.jpg</url>
      <title>DEV Community: Sheila</title>
      <link>https://dev.to/s-jt</link>
    </image>
    <atom:link rel="self" type="application/rss+xml" href="https://dev.to/feed/s-jt"/>
    <language>en</language>
    <item>
      <title>"The Ultimate Guide to Data Analytics."</title>
      <dc:creator>Sheila</dc:creator>
      <pubDate>Fri, 30 Aug 2024 14:53:47 +0000</pubDate>
      <link>https://dev.to/s-jt/the-ultimate-guide-to-data-analytics-16ha</link>
      <guid>https://dev.to/s-jt/the-ultimate-guide-to-data-analytics-16ha</guid>
      <description>&lt;p&gt;**&lt;/p&gt;

&lt;h2&gt;
  
  
  Introduction
&lt;/h2&gt;

&lt;p&gt;**&lt;br&gt;
Data analytics is a carefully curated process to convert and evaluate raw data into practical solutions that optimize the performance of companies and decision making by creating a storytelling process with data. Data analytics mainly comprises five major steps: data collection, data storage, data processing, data cleansing, and data analysis. &lt;/p&gt;

&lt;p&gt;**&lt;/p&gt;

&lt;h2&gt;
  
  
  Data collection processes
&lt;/h2&gt;

&lt;p&gt;**&lt;br&gt;
Once the objectives and aims of the analysis is set, a data analyst has to identify and collect data from reputable sources. It is therefore crucial that data collected is an accurate representation of the facts. This process can be categorized into ELT or ETL. &lt;/p&gt;

&lt;p&gt;&lt;strong&gt;Extract Transform Load (ETL)&lt;/strong&gt;&lt;br&gt;
In ETL, the first step is to extract data from files, databases, APIs, etc., then this data undergoes transformation processes by duplicate removal, identifying missing data, dropping null values, aggregates, etc. After transformation, data is loaded into a database or a warehouse for storage. &lt;/p&gt;

&lt;p&gt;&lt;strong&gt;Extract Load Transform (ELT)&lt;/strong&gt;&lt;br&gt;
In ELT, the first step is to extract data similar to the ETL process, and then the data collected is loaded into a data repository and thereafter transformed. &lt;/p&gt;

&lt;p&gt;Some of the python libraries that are crucial in these two processes include Pandas, NumPy, Apache Airflow and SQL Alchemy. These two approaches have different benefits in data analytics. ETL ensures high quality data therefore promotes flexibility and agility of data. It also ensures that security concerns are addressed especially where data is of sensitive nature. ETL therefore allows for masking of data and provides the option to include encryption keys. Where the data volume is extremely big such as in the case of big data, ELT will be a good option because it ensures that data loss does not occur and provides analysts with the option of extracting data they may need from a larger dataset. &lt;br&gt;
**&lt;/p&gt;

&lt;h2&gt;
  
  
  Data analysis and visualization Types
&lt;/h2&gt;

&lt;p&gt;**&lt;/p&gt;

&lt;p&gt;After data is processed, organized and securely stored, this data is utilized to interpret and visualize this information. Data analytics can be categorized into four types: descriptive, prescriptive, predictive, or diagnostic analytics. &lt;/p&gt;

&lt;p&gt;*&lt;em&gt;Descriptive Analytics&lt;/em&gt;&lt;br&gt;
This provides a summary of what and why occurrences are there. It is used to identify trends.&lt;/p&gt;

&lt;p&gt;&lt;em&gt;Prescriptive Analytics&lt;/em&gt;&lt;br&gt;
Provides recommendations by using machine learning and industry knowledge to give solutions on steps to be taken. &lt;/p&gt;

&lt;p&gt;&lt;strong&gt;Predictive Analytics&lt;/strong&gt;&lt;br&gt;
This is an analytic type that forecasts and predicts what may or may not happen in future. It often involves the application of machine learning. &lt;/p&gt;

&lt;p&gt;&lt;strong&gt;Diagnostic Analytics&lt;/strong&gt;&lt;br&gt;
This analytic type focuses on diagnosis / analyzing the why. Analysts do this to understand why patterns and trends are the way they are by analyzing previous patterns from past events. &lt;/p&gt;

&lt;p&gt;**&lt;/p&gt;

&lt;h2&gt;
  
  
  Data Analytics Techniques
&lt;/h2&gt;

&lt;p&gt;**&lt;br&gt;
There are many different analytics techniques. These are some of the techniques often implemented in analytics. &lt;/p&gt;

&lt;ol&gt;
&lt;li&gt;Regression analysis&lt;/li&gt;
&lt;li&gt;cluster analysis&lt;/li&gt;
&lt;li&gt;Time series analysis&lt;/li&gt;
&lt;li&gt;Classification analysis&lt;/li&gt;
&lt;li&gt;Text Analysis (NLP- Natural language processing))&lt;/li&gt;
&lt;li&gt;Principal component analysis&lt;/li&gt;
&lt;li&gt;Descriptive statistics&lt;/li&gt;
&lt;li&gt;Inferential statistics&lt;/li&gt;
&lt;/ol&gt;

&lt;p&gt;**&lt;/p&gt;

&lt;h2&gt;
  
  
  Fundamental Tools for Data Analytics
&lt;/h2&gt;

&lt;p&gt;**&lt;br&gt;
These are some of the crucial tools that are necessary for every data analyst. &lt;br&gt;
&lt;strong&gt;1. Microsoft Excel&lt;/strong&gt;&lt;br&gt;
This is a tool that is used to clean and transform data using formulation as well as visualization. &lt;br&gt;
&lt;strong&gt;2. Python&lt;/strong&gt;&lt;br&gt;
This programming language is a staple with libraries such as NumPy, pandas, Matplotlib, seaborn, Sklearn and beautiful soup. All these libraries have different functions that are important in numerical analysis and data manipulation. &lt;/p&gt;

&lt;ol&gt;
&lt;li&gt;
&lt;strong&gt;R&lt;/strong&gt;
This is another programming language that is specifically designed to conduct statistical computations. 
&lt;strong&gt;4. Tableau and power BI&lt;/strong&gt;
These are two popular visualization software that create collaborative dashboards to visualize data and indicate patterns. 
&lt;strong&gt;5. Structured Query Language (SQL)&lt;/strong&gt;
This tool is essential in examining and managing relational databases. It is useful in extracting, filtering and joining tables. Other SQL tools include NoSQL, which also stores and retrieves data in a more flexible way. 
&lt;strong&gt;6. Jupyter Notebooks&lt;/strong&gt;
This is a free open-source web application that allows for creation and sharing of documents containing live codes. It can be run on any browser or desktop and allows for the creation of reports from them. 
**&lt;/li&gt;
&lt;/ol&gt;

&lt;h2&gt;
  
  
  Conclusion
&lt;/h2&gt;

&lt;p&gt;**&lt;br&gt;
It is crucial to understand that the choice of tools to be used in data analytics is subjective and dependent on the intended outcome. Data analytics is a wide field, and any data analyst, data scientist, or machine learning expert should familiarize themselves with these tools and many more in order to enhance their productivity and increase efficiency in the data driven ecosystem. &lt;/p&gt;

</description>
      <category>dataanalytics</category>
      <category>datacareers</category>
      <category>beginners</category>
      <category>datascience</category>
    </item>
    <item>
      <title>Understanding Your Data: The Essentials of Exploratory Data Analysis.</title>
      <dc:creator>Sheila</dc:creator>
      <pubDate>Tue, 20 Aug 2024 11:43:22 +0000</pubDate>
      <link>https://dev.to/s-jt/understanding-your-data-the-essentials-of-exploratory-data-analysis-hpe</link>
      <guid>https://dev.to/s-jt/understanding-your-data-the-essentials-of-exploratory-data-analysis-hpe</guid>
      <description>&lt;p&gt;&lt;strong&gt;##&lt;/strong&gt; Introduction****&lt;br&gt;
Exploratory Data analysis is one of the most crucial process in ascertaining that the data analysis process is seamless to prevent data bias and false conclusions due to inconsistencies. It aids in exploring data structure, point out anomalies and check for assumptions in order to improve data quality, especially because there are many. These essentials include the following: &lt;/p&gt;

&lt;p&gt;**&lt;/p&gt;

&lt;h2&gt;
  
  
  Data collection
&lt;/h2&gt;

&lt;p&gt;**&lt;br&gt;
This is the first step that every data analyst, data engineer and data scientist should be conversant with. You have to be cognizant of where your data will come from, what type of data you will need, how you will gather the data, etc. Some of the data collection tools often employed by data analysts in this process includes manual data entry from interviews, questionaries and surveys, observation, focus group data and consumer data, either manually or by use of survey platforms such as survey monkey, google forms and questionnaire Pro, which can then be stored in databases and later on conducted by using software such as SQL AND NoSQL. &lt;br&gt;
Data can also be collected by using data integration tools such as Apache Nifi and Talend, which can support scalable data routing and transformation. Web scrapping tools such as BeautifulSoup and scrappy are used to collect data from websites. These are python libraries that are able to pull data out of websites that are static or dynamic without a lot of struggle. &lt;br&gt;
API tools such as postman and rapid API offer immense support in testing APIs that have been tasked to collect data from web services and provide access to many other APIs worldwide. Other tools for the collection of data include data collection apps such as Open Data Kit (ODK) and Kobo Toolbox, both of which are used by organizations to collect, manage, and use data in challenging environments.&lt;br&gt;
 To ensure that you have collected good data, ensure that you are aware of data sources and databases that provide data that is accurate, without bias and well organized. Data can be collected in different forms, such as Csv files, excel files, Pdf files or from websites that may be static or dynamic, among many other forms. &lt;br&gt;
 Data collection is mostly subjective to the specific field of interest and organization, which means that there are many ways to collect data depending on the tools used in different professions. Since different fields of specificity have different ways in which data should be collected, ethical considerations should be highly considered. For example, data from the medical field can be collected through testing of people’s health status, which comes with its own set of ethical challenges, whereas data about climate or geography can be collected by using QGIS, etc.  It is therefore crucial to understand that good data collection follows a defined set of objectives and aims in order to ensure that a dataset is accurate and caters to your specific need. &lt;br&gt;
Data Cleaning and rangling&lt;br&gt;
After the dataset is identified and collected, it is crucial to undertake rigorous data structuring. Data should be formatted to the required format to make it easier to work with. For example, data can be in form of a pdf file, which can be transformed into a csv file where needed in order to facilitate organization of said data into tables, charts, etc. &lt;br&gt;
Next, data cleaning is conducted to ensure that the quality of your data is impeccable and can be used to come up with concise conclusions and recommendations to the end user. Before any data is utilized, a data analyst should check for any cases of inconsistencies, missing/null entries duplicate entries if present. Missing values or double entries oftentimes lead to false outcomes, which may impact crucial decision-making. Therefore, anyone analyzing data should ensure that their dataset is accurate. &lt;br&gt;
After cleaning, an analyst should enrich the dataset by including additional information lacking in the dataset, merging with other datasets to get a bigger scope, or even conducting feature engineering in order to provide other variables that can improve the analysis process. After all these, an analyst should check for accuracy and consistency once more to ascertain the quality of data before moving forward. &lt;/p&gt;

&lt;p&gt;**&lt;/p&gt;

&lt;h2&gt;
  
  
  Descriptive statistics
&lt;/h2&gt;

&lt;p&gt;**&lt;br&gt;
After preparation of the dataset, data characteristics have to be outlined. This gives a brief summary of the overall dataset by indicating the mean, median, mode, and standard deviation, as well as the percentiles of data, distribution analysis to check for the central measures of tendency, distribution shapes, i.e., skewness and kurtosis, and visualization techniques to check for the distribution of dataset and probability density. &lt;/p&gt;

&lt;p&gt;**&lt;/p&gt;

&lt;h2&gt;
  
  
  Data visualization
&lt;/h2&gt;

&lt;p&gt;**&lt;br&gt;
This stage of EDA encompasses several steps, which include the identification of data variables. Data variables can be classified into the following. &lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;Univariate analysis: visualizing individual variables &lt;/li&gt;
&lt;li&gt;Bivariate analysis: visualizing the relationships between two variables by using tools such as scatter plots, correlation matrices, etc.&lt;/li&gt;
&lt;li&gt;Multivariate analysis: visualizing data by analyzing relationships between more than two sets of data or multiple variables. &lt;/li&gt;
&lt;li&gt;Outliers’ identification: the identification of data that is unusual or varies significantly from the overall dataset within specific variables. &lt;/li&gt;
&lt;li&gt;Hypotheses testing and formulation: The testing of evidence to reject or accept a null or alternative hypothesis. &lt;/li&gt;
&lt;li&gt;Testing assumptions: checking whether the data conforms to previously mentioned assumptions. &lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;**&lt;/p&gt;

&lt;h2&gt;
  
  
  Communication of findings
&lt;/h2&gt;

&lt;p&gt;**&lt;br&gt;
This is the final step in the EDA process. A summary of the evaluation process is conducted, and findings are mentioned. The context of the data is articulated, and the scope and objectives of the analysis are identified. In this final stage, patterns, anomalies and perceptions should be discussed and suggestions are made for future areas to improve on. &lt;/p&gt;

&lt;p&gt;**&lt;/p&gt;

&lt;h2&gt;
  
  
  Conclusion
&lt;/h2&gt;

&lt;p&gt;**&lt;br&gt;
To conclude, exploratory data analysis is a formidable process that enables an in-depth understanding of data and datasets by use of scientific statistical analysis techniques in order to enable fact driven decision making. As such, this process should be approached with meticulous precision in order to avoid inaccuracies. &lt;/p&gt;

</description>
      <category>eventdriven</category>
      <category>datascience</category>
      <category>datastructures</category>
      <category>data</category>
    </item>
    <item>
      <title>Data Analysis, The Ultimate Guide to Data Analytics: Techniques and Tools</title>
      <dc:creator>Sheila</dc:creator>
      <pubDate>Sun, 04 Aug 2024 20:37:47 +0000</pubDate>
      <link>https://dev.to/s-jt/data-analysis-the-ultimate-guide-to-data-analytics-techniques-and-tools-2pfa</link>
      <guid>https://dev.to/s-jt/data-analysis-the-ultimate-guide-to-data-analytics-techniques-and-tools-2pfa</guid>
      <description>&lt;h2&gt;
  
  
  What is Data Analysis?
&lt;/h2&gt;

&lt;p&gt;Data analysis is the process of analyzing, processing, and analyzing these data sets from many different data sources in order to interpret and come up with conclusions and recommendations. The increase in data types and sources has created a complex data ecosystem, thus the need for revolutionary data analysis tools. This article will delve into some of the necessary tools and techniques that are crucial in tackling data queries and challenges. &lt;/p&gt;

&lt;h2&gt;
  
  
  Crucial tools:
&lt;/h2&gt;

&lt;p&gt;To start with, the first crucial basic tool that should be a staple for any data analyst is to have a basic understanding of how technical programs work. One of the juggernaut that is Python. Python is a programming language used for writing code and scraping data from databases and websites. Secondly, a data analyst should have access to text editors like VSCode and scriptwriting and database programs such as SQL and DBeaver and have a proper understanding of data storage software such as Microsoft Excel and data presentation software such as Power BI. All these software's and tools enable ease of processing data in any data analyst's career. &lt;br&gt;
As I continue with my learning experience, I have also come across several analytical techniques that I still need to further delve to in order to be proficient in this field. Some of these skills include:&lt;br&gt;
i. Descriptive analysis&lt;br&gt;
ii. Cluster analysis&lt;br&gt;
iii. Regression analysis&lt;br&gt;
iv. predictive analysis&lt;br&gt;
All these skills paired well with critical thinking and base knowledge in the field of research makes a good data analyst. &lt;/p&gt;

&lt;h2&gt;
  
  
  Conclusion
&lt;/h2&gt;

&lt;p&gt;To conclude, data analysis is the door that opens opportunities for solving real life problems. It is therefore crucial for data analysts to have high expertise in the tools used in this field and be flexible to adapt to any changes within this fast-evolving industry. &lt;/p&gt;

</description>
      <category>data</category>
      <category>datascience</category>
      <category>dataanalytics</category>
    </item>
  </channel>
</rss>
