<?xml version="1.0" encoding="UTF-8"?>
<rss version="2.0" xmlns:atom="http://www.w3.org/2005/Atom" xmlns:dc="http://purl.org/dc/elements/1.1/">
  <channel>
    <title>DEV Community: Philemon Kiplangat</title>
    <description>The latest articles on DEV Community by Philemon Kiplangat (@philemonkiplangat).</description>
    <link>https://dev.to/philemonkiplangat</link>
    <image>
      <url>https://media2.dev.to/dynamic/image/width=90,height=90,fit=cover,gravity=auto,format=auto/https:%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Fuser%2Fprofile_image%2F1173893%2Fdcb8dcba-7475-40cc-a23f-7d8966cabe10.png</url>
      <title>DEV Community: Philemon Kiplangat</title>
      <link>https://dev.to/philemonkiplangat</link>
    </image>
    <atom:link rel="self" type="application/rss+xml" href="https://dev.to/feed/philemonkiplangat"/>
    <language>en</language>
    <item>
      <title>Unsupervised machine learning.</title>
      <dc:creator>Philemon Kiplangat</dc:creator>
      <pubDate>Fri, 22 Mar 2024 06:58:05 +0000</pubDate>
      <link>https://dev.to/philemonkiplangat/unsupervised-machine-learning-mp0</link>
      <guid>https://dev.to/philemonkiplangat/unsupervised-machine-learning-mp0</guid>
      <description>&lt;p&gt;Unsupervised learning is a branch of artificial intelligence and a type of machine learning that learns from unlabeled data without human supervision and allows for insights about the data and patterns without any explicit guidance or instruction. &lt;br&gt;
&lt;strong&gt;Main points to note&lt;/strong&gt;&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;&lt;p&gt;Unsupervised learning allows the model to learn by discovering patterns and relationships in unlabeled data.&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;Clustering algorithms group similar data points together based on their inherent characteristics.&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;Feature extraction captures esential information.&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;Label association assigns categories to the clusters based on the extracted patterns and characteristics.&lt;/p&gt;&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;A perfect example of how to illustrate this is to imagine a machine learning model trained on a large dataset of unlabeled patients with different diseases. Your task is to use unsupervised learning to identify different types of diabetes patients, be it type 1, type 2, or gestational diabetes, from the unseen data.Thus, the machine has no idea about the features of the patients. So we can’t categorize a specific type of diabetes among the patients. But it can categorize them according to their similarities, patterns, and differences. &lt;/p&gt;

&lt;p&gt;&lt;strong&gt;Types of Unsupervised Learning&lt;/strong&gt;&lt;br&gt;
Unsupervised learning is classified into two categories of algorithms: &lt;/p&gt;

&lt;p&gt;&lt;strong&gt;Clustering&lt;/strong&gt;: A clustering problem is where you want to discover the inherent groupings in the data, such as grouping customers by purchasing behavior.&lt;br&gt;
Clustering is a type of unsupervised learning that is used to group similar data points together. Clustering algorithms work by iteratively moving data points closer to their cluster centers and further away from data points in other clusters.&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;&lt;p&gt;Exclusive (partitioning)&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;Agglomerative&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;overlapping&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;Probabilistic&lt;br&gt;
Clustering Types:-&lt;br&gt;
Hierarchical clustering&lt;br&gt;
K-means clustering&lt;br&gt;
Principal Component Analysis&lt;br&gt;
Singular Value Decomposition&lt;br&gt;
Independent Component Analysis&lt;br&gt;
Gaussian Mixture Models (GMMs)&lt;br&gt;
Density-Based Spatial Clustering of Applications with Noise (DBSCAN)&lt;br&gt;
&lt;strong&gt;Association&lt;/strong&gt;: An association rule learning problem is where you want to discover rules that describe large portions of your data, such as people that buy X also tend to buy Y.&lt;br&gt;
An association rule based on machine learning is to determine the probability of the co-occurrence of items in a collection.A good example is finding out which products were purchased together.&lt;/p&gt;&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;In conclusion, unsupervised machine learning does not have evaluation metrics; therefore, it does not have a feedback mechanism.&lt;/p&gt;

</description>
    </item>
    <item>
      <title>The Complete Guide to Time Series Models</title>
      <dc:creator>Philemon Kiplangat</dc:creator>
      <pubDate>Mon, 23 Oct 2023 13:34:55 +0000</pubDate>
      <link>https://dev.to/philemonkiplangat/the-complete-guide-to-time-series-models-3jeo</link>
      <guid>https://dev.to/philemonkiplangat/the-complete-guide-to-time-series-models-3jeo</guid>
      <description>&lt;p&gt;&lt;strong&gt;Time series&lt;/strong&gt;: this refers to the sequence of data points that occur in successive order over a period of time. A time series in investing records the movement of selected data points, such as the price of a security, over a set period of time, with data points recorded at regular intervals. There is no minimum or maximum time requirement, allowing the data to be obtained in a fashion that gives the information desired by the investor or analyst reviewing the activity.&lt;br&gt;
The main importance of the time series is for predictive analysis.&lt;br&gt;
Time series analysis can also be defined as a method of analyzing a collection of data points over a period of time. Instead of recording data points intermittently or randomly, time series analysts record data points at consistent intervals over a defined period of time.&lt;br&gt;
While time-series data is information gathered over time, various types of information describe how and when that information was gathered. For example:&lt;/p&gt;

&lt;p&gt;Time series data is a collection of observations on the values that a variable takes at various points in time.&lt;/p&gt;

&lt;p&gt;Cross-sectional data: data from one or more variables that were collected simultaneously.&lt;/p&gt;

&lt;p&gt;Pooled data: It is a combination of cross-sectional and time-series data.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;The importance of time series analysis&lt;/strong&gt;&lt;/p&gt;

&lt;p&gt;Time series analysis has a wide range of importance in different fields, such as sales, economics, and many more. However, the common point is the technique used to model the data over a period of time. The reasons for time series analysis are as follows:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;Features:Time series analysis can be used to track features like trend, seasonality, and variability.&lt;/li&gt;
&lt;li&gt;Prediction: time series analysis can be used for prediction purposes by studying the patterns of the data available.
-Inferences: You can predict the value and draw inferences from the data using time series analysis.&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;&lt;strong&gt;Time Series Analysis Types&lt;/strong&gt;&lt;/p&gt;

&lt;ol&gt;
&lt;li&gt;Classification: It identifies categories in the data.&lt;/li&gt;
&lt;li&gt;Curve-fitting: it plots the data on the curve to identify the relationship between variables in the data.&lt;/li&gt;
&lt;li&gt;Descriptive analysis: patterns in the time-series data, such as trends, cycles, and seasonal variations, are identified &lt;/li&gt;
&lt;li&gt;Explanative analysis: It attempts to comprehend the data and the relationships between it and cause and effect.&lt;/li&gt;
&lt;li&gt;Segmentation: It splits the data into segments to reveal the source data's underlying properties.&lt;/li&gt;
&lt;/ol&gt;

&lt;p&gt;*&lt;em&gt;Time series analysis example *&lt;/em&gt;&lt;/p&gt;

&lt;p&gt;Non-stationary data is defined as data that fluctuates over time or is affected by time and is evaluated using time series analysis.Because currency and sales fluctuate, businesses such as finance, retail, and e-commerce regularly employ time series analysis. Stock market analysis is a great illustration of time series analysis in action, especially when combined with automated trading algorithms.&lt;/p&gt;

&lt;p&gt;Time series analysis can be used in:&lt;/p&gt;

&lt;p&gt;Rainfall measurements&lt;br&gt;
Automated stock trading&lt;br&gt;
Industry forecast&lt;br&gt;
Temperature readings&lt;br&gt;
Sales forecasting&lt;/p&gt;

</description>
    </item>
    <item>
      <title>Data Modeling.</title>
      <dc:creator>Philemon Kiplangat</dc:creator>
      <pubDate>Mon, 23 Oct 2023 08:34:48 +0000</pubDate>
      <link>https://dev.to/philemonkiplangat/data-modeling-2iio</link>
      <guid>https://dev.to/philemonkiplangat/data-modeling-2iio</guid>
      <description>&lt;p&gt;Over the years, many businesses have been cautious about decision making processes that affect them. This is important since decisions made by a business determines its success. The part of the decision is forecasting which can be made possible by studying the growth of the business. The oil for decision making has been data. It is through it where one can obtain insights about the business and the growth pattern. Decision making process is greatly characterized by the data modelling process.   &lt;/p&gt;

&lt;p&gt;&lt;strong&gt;Data Modelling&lt;/strong&gt; refers to the process of analyzing and define all the data types your business collects and produce,as well as the relationship between those bits of data. This can be achieved by using different tools in the tech field. Modelling can be represented using text, symbols and diagrams since it represents how the data is captured, stored and used.&lt;/p&gt;

&lt;h2&gt;
  
  
  Data Modeling Process
&lt;/h2&gt;

&lt;p&gt;This refer to the process of creating conceptual representation of objects and their relationship to one another.The process of data modeling typically involves specific defined steps which are;&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;&lt;p&gt;Requirements gathering.&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;Conceptual design.&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;Logic design.&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;Implementation.&lt;br&gt;
During each step the data modelers work with stakeholders to understand the data requirements, define the entities and attributes, establish the relationships between the data objects, and create a model that accurately represents the data in a way that can be used by the stakeholders.&lt;/p&gt;&lt;/li&gt;
&lt;/ul&gt;

&lt;h3&gt;
  
  
  Levels of abstraction.
&lt;/h3&gt;

&lt;ul&gt;
&lt;li&gt;&lt;p&gt;Conceptual model- collaborate with stakeholders to understand the data requirements, identify the entities and attributes, build the links between the data objects, and develop a model that accurately represents the data in a usable format. &lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;logical level-The logical level involves defining the relationships and constraints between the data objects in more detail, often using data modeling languages such as SQL or ER diagrams.&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;Physical level-The physical level involves defining the specific details of how the data will be stored, including data types, indexes, and other technical details.&lt;/p&gt;&lt;/li&gt;
&lt;/ul&gt;

&lt;h2&gt;
  
  
  Data Modeling Examples
&lt;/h2&gt;

&lt;p&gt;The best way to picture a data model is to think about a building plan of an architect. An architectural building plan assists in putting up all subsequent conceptual models, and so does a data model. &lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;&lt;p&gt;&lt;strong&gt;Entity Relationship Model&lt;/strong&gt;:This model is based on the notion of real-world entities and relationships among them. It creates an entity set, relationship set, general attributes, and constraints.&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;&lt;strong&gt;Hierarchical Model&lt;/strong&gt;:This data model arranges the data in the form of a tree with one root, to which other data is connected. The hierarchy begins with the root and extends like a tree. This model effectively explains several real-time relationships with a single one-to-many relationship between two different kinds of data.&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;&lt;strong&gt;Network Model&lt;/strong&gt;:This database model enables many-to-many relationships among the connected nodes. The data is arranged in a graph-like structure, and here ‘child’ nodes can have multiple ‘parent’ nodes. The parent nodes are known as owners, and the child nodes are called members.&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;&lt;strong&gt;Relational Model&lt;/strong&gt;:This popular data model example arranges the data into tables. The tables have columns and rows, each cataloging an attribute present in the entity. It makes relationships between data points easy to identify. &lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;&lt;strong&gt;Object-Relational Mode&lt;/strong&gt;:This model is a hybrid of an object-oriented database and a relational database. As a result, it combines the extensive functionality of the object-oriented paradigm with the simplicity of the relational data model.&lt;/p&gt;&lt;/li&gt;
&lt;/ul&gt;

&lt;h2&gt;
  
  
  Benefits Of Data Modeling.
&lt;/h2&gt;

&lt;ul&gt;
&lt;li&gt;&lt;p&gt;Allows the developers and the stakeholders to understand the relationship between different objects for easier analysis.&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;Improved data quality: Data modeling can help to identify errors and inconsistencies in the data, which can improve the overall quality of the data and prevent problems later on.&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;Improved collaboration: Data modeling helps to facilitate communication and collaboration among stakeholders, which can lead to more effective decision-making and better outcomes.&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;Increased efficiency: Data modeling can help to streamline the development process by providing a clear and consistent representation of the data that can be used by developers, database administrators, and other stakeholders.&lt;/p&gt;&lt;/li&gt;
&lt;/ul&gt;

</description>
    </item>
    <item>
      <title>Expolaratory Data Analysis</title>
      <dc:creator>Philemon Kiplangat</dc:creator>
      <pubDate>Sun, 08 Oct 2023 16:20:14 +0000</pubDate>
      <link>https://dev.to/philemonkiplangat/expolaratory-data-analysis-166j</link>
      <guid>https://dev.to/philemonkiplangat/expolaratory-data-analysis-166j</guid>
      <description>&lt;p&gt;Over many years there has been there has been much change from other ways of civilization starting from the agrarian revolution the early years. There has been gradual change till now we are in the age where information is important, hence the phrase knowledge is power. In the digital age, data is the new gold, but its true value rests in the insights it gives. However, raw data is frequently complicated and difficult to comprehend. This is where Data Exploratory Analysis (EDA) comes in.&lt;/p&gt;

&lt;p&gt;Therefore, &lt;strong&gt;Exploratory Data Analysis&lt;/strong&gt; can be define as the process of examining and understanding insights and patterns within data and convey the information through visualization.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;Importance of Exploratory Data Analysis&lt;/strong&gt;&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;&lt;p&gt;Exploratory data analysis can help detect obvious errors, identify outliers in datasets, understand relationships, unearth important factors, find patterns within data, and provide new insights. &lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;EDA is the initial step in data analysis. Its primary goal is to summarize the main characteristics of a dataset, often employing statistical and graphical techniques. By understanding the structure and patterns within the data, researchers and analysts can make better decisions, identify relationships, and even formulate hypotheses for further analysis.&lt;/p&gt;&lt;/li&gt;
&lt;/ul&gt;

&lt;h2&gt;
  
  
  Exploratory data analysis tools.
&lt;/h2&gt;

&lt;p&gt;Specific statistical functions and techniques you can perform with EDA tools include:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;&lt;p&gt;Clustering and dimension reduction techniques, which help create graphical displays of high-dimensional data containing many variables.&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;Univariate visualization of each field in the raw dataset, with summary statistics.&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;Bivariate visualizations and summary statistics that allow you to assess the relationship between each variable in the dataset and the target variable you’re looking at.&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;Multivariate visualizations, for mapping and understanding interactions between different fields in the data.&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;K-means Clustering is a clustering method in unsupervised learning where data points are assigned into K groups, i.e. the number of clusters, based on the distance from each group’s centroid. The data points closest to a particular centroid will be clustered under the same category. K-means Clustering is commonly used in market segmentation, pattern recognition, and image compression.&lt;/p&gt;&lt;/li&gt;
&lt;/ul&gt;

&lt;h2&gt;
  
  
  Exploratory Data Analysis Tools
&lt;/h2&gt;

&lt;ul&gt;
&lt;li&gt;&lt;p&gt;&lt;strong&gt;Python&lt;/strong&gt;-Python is an object-oriented, interpreted programming language with dynamic semantics. Its high-level, built-in data structures, together with dynamic typing and dynamic binding, make it particularly appealing for rapid application development as well as use as a scripting or glue language to connect existing components. Python and EDA can be used in tandem to find missing values in a data set, which is useful for determining how to handle missing values in machine learning.&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;&lt;strong&gt;R&lt;/strong&gt;The R Foundation for Statistical Computing supports an open-source programming language and free software environment for statistical computing and graphics. The R programming language is commonly used by statisticians in data science to create statistical observations and data analysis.&lt;/p&gt;&lt;/li&gt;
&lt;/ul&gt;

&lt;h2&gt;
  
  
  The Role of Data Visualization in EDA.
&lt;/h2&gt;

&lt;p&gt;As is is often said , a picture is worth more than a thousand words,data visualization is a powerful tool within the realm of EDA. Instead of drowning in a sea of numbers, visual representations provide a clear, concise, and intuitive way to grasp complex concepts. Here’s how data visualization enhances EDA:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;&lt;p&gt;&lt;strong&gt;Spotting Patterns and Trends&lt;/strong&gt;: Charts, graphs, and plots are excellent at showcasing trends over time or across different variables. For instance, line charts can demonstrate stock price fluctuations, while bar charts can compare sales figures among different products.&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;&lt;strong&gt;Identifying Outliers&lt;/strong&gt;:Outliers, or data points that depart dramatically from the norm, are easily identified in visualizations. Box plots and scatter plots are frequently used to detect these anomalies, which could be errors or interesting occurrences worth investigating.&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;&lt;strong&gt;Understanding Distributions&lt;/strong&gt;:Histograms and density charts aid in visualizing data distribution. Understanding the distribution shape (normal, skewed, bimodal) provides critical insights regarding the dataset's underlying nature.&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;&lt;strong&gt;Exploring Relationships&lt;/strong&gt;:Scatter plots are invaluable for displaying relationships between two variables. Positive, negative, or no correlation between variables becomes apparent, aiding in decision-making processes.&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;&lt;strong&gt;Geospatial Analysis&lt;/strong&gt;:Maps and geospatial visualizations aid in the comprehension of data in a geographical context. This is particularly helpful for businesses, epidemiologists, and sociologists researching regional trends.&lt;/p&gt;&lt;/li&gt;
&lt;/ul&gt;

&lt;h2&gt;
  
  
  &lt;strong&gt;Common Data Visualization Techniques&lt;/strong&gt;
&lt;/h2&gt;

&lt;ul&gt;
&lt;li&gt;&lt;p&gt;&lt;strong&gt;Bar and pie charts&lt;/strong&gt; are excellent for illustrating categorical data or proportions.&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;&lt;strong&gt;Line charts&lt;/strong&gt; are ideal for displaying trends over a continuous interval of time.&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;&lt;strong&gt;Scatter plots&lt;/strong&gt;&lt;br&gt;
are excellent for displaying relationships between two variables.&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;&lt;strong&gt;Histograms&lt;/strong&gt;&lt;br&gt;
can help you comprehend the distribution and frequency of numerical data.&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;&lt;strong&gt;Box plots&lt;/strong&gt; are great for finding outliers and analyzing data dispersion.&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;&lt;strong&gt;Heat Maps&lt;/strong&gt;: These are useful for expressing data in matrix format, and they are frequently used in correlation matrices and geographic research.&lt;/p&gt;&lt;/li&gt;
&lt;/ul&gt;

&lt;h3&gt;
  
  
  Summary
&lt;/h3&gt;

&lt;p&gt;Data exploratory analysis is a crucial step in extracting meaningful insights from raw data. Data visualization serves as a bridge between raw numbers and understandable patterns. By leveraging various visualization techniques, analysts and researchers can unlock the potential of their datasets, enabling better decision-making, problem-solving, and innovation. In a world inundated with data, the ability to harness the power of EDA and data visualization is a skill that can set individuals and organizations apart, propelling them towards success in various fields.&lt;/p&gt;

</description>
    </item>
    <item>
      <title>Data Science for Beginners:</title>
      <dc:creator>Philemon Kiplangat</dc:creator>
      <pubDate>Sat, 30 Sep 2023 15:56:54 +0000</pubDate>
      <link>https://dev.to/philemonkiplangat/data-science-for-beginners-181p</link>
      <guid>https://dev.to/philemonkiplangat/data-science-for-beginners-181p</guid>
      <description>&lt;p&gt;What is Data Science?&lt;br&gt;
&lt;strong&gt;Data science&lt;/strong&gt; can be define in many ways considering that data has been named to be new oil in different business practice. Data Science can be simply defined as  the study of data in order to derive valuable business insights. It is a multidisciplinary approach to data analysis that integrates principles and practices from mathematics, statistics, artificial intelligence, and computer engineering.&lt;br&gt;
there are different paths associated with data , namely data analysis data engineering and analytical engineering. They can be dine differently since each can have different technologies to handle.&lt;br&gt;
&lt;strong&gt;Data Engineering&lt;/strong&gt; can be defined as the practice of designing and building systems for collecting,analyzing and storing data at scale and the practitioner of the discipline is known as a &lt;strong&gt;data engineer&lt;/strong&gt;.&lt;br&gt;
&lt;strong&gt;Data Analysis&lt;/strong&gt;  is a discipline  which involves  gathering raw data and then transforming it into information that users may utilize to make decisions.The practitioner of this discipline is known as a &lt;strong&gt;data analyst&lt;/strong&gt; whose work mainly entails reviewing data to identify key insights into a business's customers and ways the data can be used to solve problems. &lt;br&gt;
&lt;strong&gt;Analytical engineering&lt;/strong&gt; is a discipline where it involves optimization of data models. Its practitioner is referred to as &lt;strong&gt;analytical engineer&lt;/strong&gt; whose task mainly involves modeling data to provide clean, accurate datasets so that different users within the company can work with them.&lt;/p&gt;

&lt;h3&gt;
  
  
  Data Science for Beginners: 2023 - 2024 Complete Road-map.
&lt;/h3&gt;

&lt;p&gt;Data science has emerged as an uncontested practice and debate topic. The key causes for this scenario are the massive amounts of data generated and the importance of data analysis in driving decision-making. Data science employs a variety of approaches for data collecting, preparation, storage, analytics, and interpretation. The advancement of tools, methods, computational speeds, and automation has further cemented data science's dominance.&lt;/p&gt;

&lt;ol&gt;
&lt;li&gt;
&lt;strong&gt;Programming&lt;/strong&gt;-Programming is a process that creates programs that involve the ratification of codes, For data science there are specific languages where one use. These are Python,R, Java and SQL.&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;Math Fundamentals&lt;/strong&gt;- There are basic math fundamentals where one is required to use in Data Science. These are, statistics,
linear algebra,differential calculus,discrete maths.&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;Data Analysis&lt;/strong&gt;The main concepts in data analysis involve,
feature engineering,data wrangling,exploratory data analysis,&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;Machine Learning&lt;/strong&gt;- one should have deep understanding of the following algorithms.classification, regression,reinforcement learning,deep learning, dimensionaly reduction,clustering&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;Web Scrapping&lt;/strong&gt;.
beautiful Soap,scrappy,URLLIB&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;Visualization&lt;/strong&gt;

&lt;ul&gt;
&lt;li&gt;PowerBI&lt;/li&gt;
&lt;li&gt;Excel&lt;/li&gt;
&lt;li&gt;Matplotlib&lt;/li&gt;
&lt;li&gt;Tableau&lt;/li&gt;
&lt;/ul&gt;
&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;Deployment&lt;/strong&gt;. This can be done by using various cloud technologies such as Azure or AWS.&lt;/li&gt;
&lt;/ol&gt;

</description>
    </item>
  </channel>
</rss>
