<?xml version="1.0" encoding="UTF-8"?>
<rss version="2.0" xmlns:atom="http://www.w3.org/2005/Atom" xmlns:dc="http://purl.org/dc/elements/1.1/">
  <channel>
    <title>DEV Community: Elishiba Muigo</title>
    <description>The latest articles on DEV Community by Elishiba Muigo (@elimuigo).</description>
    <link>https://dev.to/elimuigo</link>
    <image>
      <url>https://media2.dev.to/dynamic/image/width=90,height=90,fit=cover,gravity=auto,format=auto/https:%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Fuser%2Fprofile_image%2F1172190%2Faf02caca-8347-437d-88cf-587e1d138b6b.jpeg</url>
      <title>DEV Community: Elishiba Muigo</title>
      <link>https://dev.to/elimuigo</link>
    </image>
    <atom:link rel="self" type="application/rss+xml" href="https://dev.to/feed/elimuigo"/>
    <language>en</language>
    <item>
      <title>HANDLING THE SAFE UPDATE MODE IN MYSQL</title>
      <dc:creator>Elishiba Muigo</dc:creator>
      <pubDate>Thu, 12 Jun 2025 16:18:17 +0000</pubDate>
      <link>https://dev.to/elimuigo/handling-the-safe-update-mode-in-mysql-249o</link>
      <guid>https://dev.to/elimuigo/handling-the-safe-update-mode-in-mysql-249o</guid>
      <description>&lt;p&gt;*&lt;em&gt;Safe Update Mode is a feature designed to prevent unintentional data loss or corruption during update and delete operations. It restricts the execution of UPDATE and DELETE statements that don't include a WHERE clause that uses a key column, thus preventing accidental modifications to entire tables. *&lt;/em&gt;&lt;/p&gt;

&lt;h3&gt;
  
  
  &lt;u&gt; The Default and Safe Mode &lt;/u&gt;
&lt;/h3&gt;

&lt;ul&gt;
&lt;li&gt;When SQL_SAFE_UPDATES is set to 1 (which is the default setting in many MySQL client tools like MySQL Workbench), the database system will prevent UPDATE or DELETE statements that:&lt;/li&gt;
&lt;li&gt;Do not use a WHERE clause: For example, &lt;code&gt;UPDATE my_table SET column1 = 'new_value';&lt;/code&gt;
&lt;/li&gt;
&lt;li&gt;Do not use a key column in the WHERE clause: For example, if id is your primary key, &lt;code&gt;UPDATE my_table SET column1 = 'new_value' WHERE some_other_column = 'value';&lt;/code&gt; would be blocked unless some_other_column is also indexed or part of a unique key.&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;&lt;code&gt;SET SQL_SAFE_UPDATES = 1&lt;/code&gt;&lt;br&gt;
&lt;code&gt;DELETE FROM database_name.table_name&lt;br&gt;
WHERE condition;&lt;/code&gt;&lt;/p&gt;

&lt;h3&gt;
  
  
  &lt;u&gt; The Unsafe/ Override Mode &lt;/u&gt;
&lt;/h3&gt;

&lt;ul&gt;
&lt;li&gt;&lt;p&gt;When SQL_SAFE_UPDATES is set to 0, the database system will not prevent UPDATE or DELETE statements that lack a WHERE clause or don't use a key in the WHERE clause.&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;This is typically used when you intentionally want to perform a mass update or delete operation without specifying a key column in the WHERE clause, or if you are sure about your WHERE clause, even if it doesn't involve a key. It gives you more flexibility but also removes the safety net.&lt;/p&gt;&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;&lt;code&gt;SET SQL_SAFE_UPDATES = 0&lt;/code&gt;&lt;br&gt;
&lt;code&gt;DELETE FROM database_name.table_name&lt;br&gt;
WHERE condition;&lt;/code&gt;&lt;/p&gt;

</description>
      <category>mysql</category>
      <category>sql</category>
      <category>sqlserver</category>
      <category>database</category>
    </item>
    <item>
      <title>A Step-by-Step Roadmap to Data Engineering</title>
      <dc:creator>Elishiba Muigo</dc:creator>
      <pubDate>Thu, 09 Nov 2023 20:18:43 +0000</pubDate>
      <link>https://dev.to/elimuigo/a-step-by-step-roadmap-to-data-engineering-2f9o</link>
      <guid>https://dev.to/elimuigo/a-step-by-step-roadmap-to-data-engineering-2f9o</guid>
      <description>&lt;h3&gt;
  
  
  Introduction
&lt;/h3&gt;

&lt;p&gt;Data engineering is the practice of designing and building systems for collecting, storing, and analyzing data as scale. &lt;br&gt;
Data engineers work in a variety of settings to build systems that collect, manage and convert raw data into usable information for data scientists and business analysts to interpret. They make data accessible so that organizations can use to evaluate and optimize performance.&lt;/p&gt;

&lt;h3&gt;
  
  
  1.Basic programming languages
&lt;/h3&gt;

&lt;p&gt;.Knowledge in python programing and scala.&lt;br&gt;
.Knowledge of SQL and database programming. Know how to design and implement data models and database using tools such as PostgreSQL and MySQL.&lt;/p&gt;

&lt;h3&gt;
  
  
  2. Learn Linux commands and Git
&lt;/h3&gt;

&lt;p&gt;.Knowledge in version control using git, this allows you to manage changes t workloads and basic Linux commands. &lt;/p&gt;

&lt;h3&gt;
  
  
  3. Data integration, ETL and ELT
&lt;/h3&gt;

&lt;p&gt;Data integration is the process of combining data from multiple sources into cohesive view. It involves gathering data from various sources, cleaning and transforming the data to make it consistent and compatible then storing it.&lt;/p&gt;

&lt;p&gt;ETL(Extract, Transform and Load) is a process in data data warehousing and business intelligence that involves extracting data from various sources, transforming it into a format that is suitable for analysis and reporting then loading it into a data warehouse or other data repository.&lt;br&gt;
  .Experience ETL tools like Apache Kafka, Talend.&lt;/p&gt;

&lt;p&gt;ELT(Extract, Loading and Transformation) a process that involves moving raw data from a source system to a destination resource such a data warehouse.&lt;/p&gt;

&lt;p&gt;Learn about ELT and ETL and when is the best time to use which method.&lt;/p&gt;

&lt;h3&gt;
  
  
  4. Data storage and warehousing
&lt;/h3&gt;

&lt;p&gt;Data warehousing enables organizations to store, organize, and analyze data from various sources in a centralized location, providing a more complete view of the organization's data. &lt;/p&gt;

&lt;p&gt;.Learn data modelling and schema design for data warehouse.&lt;br&gt;
  .Learn data warehousing concepts and tools like Snowflake, Redshift and &lt;br&gt;
  amazon.&lt;/p&gt;

&lt;h3&gt;
  
  
  5. Data Pipelines
&lt;/h3&gt;

&lt;p&gt;A method in which raw data is ingested from various data sources and then moved to a data store, like a data lake or a data warehouse for analysis. Data can be sources form API's, SQL and NoSQL databases.&lt;/p&gt;

&lt;p&gt;.Learn about Batch processing and Streaming data.&lt;br&gt;
 .Learn the data pipeline architecture.&lt;br&gt;
 .Understanding of workflow management tools such as Apache Airflow, AWS and &lt;br&gt;
  Azure Data Factory.&lt;br&gt;
 . Familiarize with containerization technologies such as Docker and Kubernetes for managing and deploying data pipelines.&lt;/p&gt;

&lt;h3&gt;
  
  
  6.Cloud Computing
&lt;/h3&gt;

&lt;p&gt;The cloud computing platforms offer an advantage to data engineers including scalable infrastructure and a range of tools for data processing and analysis.&lt;br&gt;
  .Platforms like AWS, Google Cloud Platform and Microsoft Azure.&lt;/p&gt;

&lt;h3&gt;
  
  
  7.Data governance and security
&lt;/h3&gt;

&lt;p&gt;Data governance ensures data is managed by legal and regulatory requirements. Another important function of data governance is helping protect data from unauthorized access, theft, and misuse. This is critical for data engineers, who are in charge of designing and maintaining secure data systems. &lt;/p&gt;

</description>
      <category>dataengineering</category>
      <category>datascience</category>
    </item>
    <item>
      <title>A COMPLETE GUIDE TO TIME SERIES MODELLING</title>
      <dc:creator>Elishiba Muigo</dc:creator>
      <pubDate>Mon, 30 Oct 2023 21:01:29 +0000</pubDate>
      <link>https://dev.to/elimuigo/a-complete-guide-to-time-series-modelling-1d9m</link>
      <guid>https://dev.to/elimuigo/a-complete-guide-to-time-series-modelling-1d9m</guid>
      <description>&lt;h3&gt;
  
  
  What is a Time Series Model?
&lt;/h3&gt;

&lt;p&gt;-An ordered sequence of values of variable at equally spaced time intervals. Used in accurately predicting patterns and trends in time-dependent data can offer valuable insights into fields such as climate analysis, stock marketing analysis and economics.&lt;br&gt;
&lt;b&gt;Time series modeling&lt;/b&gt; is a statistical and mathematical technique used to analyze and make predictions about data points collected and recorded over a series of time intervals. It's a method of analyzing a collection of data points over a period of time.&lt;/p&gt;

&lt;p&gt;&lt;b&gt;&lt;u&gt;Characteristics of Time Series Model&lt;/u&gt;:&lt;/b&gt;&lt;/p&gt;

&lt;p&gt;&lt;b&gt;1.&lt;u&gt;Autocorrelation&lt;/u&gt;&lt;/b&gt;-It's the degree of similarity between a given time series and a lagged version of itself over successive time intervals. Autocorrelation measures the relationship between a variable's current value and its past values. It measures the correlation between a time series and a lagged version of itself. Helps in model selection and diagnostics. &lt;/p&gt;

&lt;p&gt;&lt;b&gt;2.&lt;u&gt;Seasonality&lt;/u&gt;&lt;/b&gt;-characteristic of time series in which data experiences regular and predictable changes that reoccur every year and is said to be seasonal. Seasonality refers to periodic fluctuations. Many time series exhibit seasonality which is recurring patterns or cycles that occur at regular intervals.&lt;/p&gt;

&lt;p&gt;&lt;b&gt;3.&lt;u&gt;Stationarity&lt;/u&gt;&lt;/b&gt;-A fundamental assumption in time series analysis is stationarity. A time series mean, variance, and autocorrelation remain constant over time.&lt;/p&gt;

&lt;p&gt;&lt;b&gt;&lt;u&gt;Time Series Analysis:&lt;/u&gt;&lt;/b&gt;&lt;/p&gt;

&lt;p&gt;&lt;b&gt;1.&lt;u&gt;Segmentation&lt;/u&gt;&lt;/b&gt;- splits the data into segments to reveal the underlying properties of the source information.&lt;/p&gt;

&lt;p&gt;&lt;b&gt;2.&lt;u&gt;Explanative Analysis&lt;/u&gt;&lt;/b&gt;-attempts to understand the data and the relationship between it's cause effect.&lt;/p&gt;

&lt;p&gt;&lt;b&gt;3.&lt;u&gt;Classification&lt;/u&gt;&lt;/b&gt;-Identifies and assigns categories to the data.&lt;/p&gt;

&lt;p&gt;&lt;b&gt;4.&lt;u&gt;Forecasting&lt;/u&gt;&lt;/b&gt;-Predicts future data. Time series models are used for forecasting future values of the series. Common techniques for forecasting include autoregressive (AR) models, moving average (MA) models, and their combinations in autoregressive integrated moving average (ARIMA) models.&lt;/p&gt;

&lt;p&gt;&lt;b&gt;5.&lt;u&gt;Curve Fitting&lt;/u&gt;&lt;/b&gt;-Plots the data along a curve to study the relationships of variable within the data. &lt;/p&gt;

&lt;p&gt;&lt;b&gt;6.&lt;u&gt;Descriptive Analysis&lt;/u&gt;&lt;/b&gt;-Patterns in the time series data such as trends, seasonal variations and cycles.&lt;/p&gt;

&lt;p&gt;&lt;b&gt;7.&lt;u&gt;Exploratory analysis&lt;/u&gt;&lt;/b&gt;- Highlights the main characteristics of the time series data, usually in a visual format.&lt;/p&gt;

&lt;p&gt;&lt;b&gt;&lt;u&gt;Models of Time Series:&lt;/u&gt;&lt;/b&gt;&lt;/p&gt;

&lt;p&gt;&lt;b&gt;&lt;u&gt;ARIMA (Autoregressive Integrated Moving Average)&lt;/u&gt;&lt;/b&gt;&lt;br&gt;
-A time series forecasting model used for analyzing and forecasting time-dependent data. It combines three key components: autoregression (AR), differencing (I for Integrated), and moving averages (MA). ARIMA models can apply in some cases where data show non-stationarity in the mean.&lt;/p&gt;

&lt;p&gt;&lt;b&gt;&lt;u&gt;Parts of ARIMA&lt;/u&gt;&lt;/b&gt;&lt;br&gt;
&lt;b&gt;AR (Autoregressive):&lt;/b&gt; refers to the number of previous values to consider for the forecast. Described by the perimeter "p". Autoregressive is the lags of the variables in the stationary series in the estimation equation.&lt;/p&gt;

&lt;p&gt;&lt;b&gt;I (Integrated):&lt;/b&gt; differentiation of time series data. Characterized by "d". Refers to the number of differencing applied with the objective of achieving stationary time series. Integrated means that the data values are changed with the difference between their own values and previous values to make the series stable.&lt;/p&gt;

&lt;p&gt;&lt;b&gt;MA (Moving Average):&lt;/b&gt; a linear combination of past error values instead of previous values of the variable interest. Described by the parameter "q". Refers to the lags of forecast errors.&lt;/p&gt;

&lt;p&gt;&lt;b&gt;&lt;u&gt;SARIMA MODEL ( Seasonal Autoregressive Integrated Moving Average)&lt;/u&gt;&lt;/b&gt;&lt;br&gt;
It's designed to handle time series data with seasonal patterns. SARIMA models are widely used for time series forecasting and analysis, particularly when the data exhibit recurring patterns at regular intervals, such as daily, monthly, or yearly seasonality.&lt;/p&gt;

&lt;p&gt;When there's seasonality in the series SARIMA will instead ARIMA. There is only one variable in both data and it will be a suitable model as SARIMA supports univariate time series data.&lt;/p&gt;

</description>
      <category>programming</category>
      <category>python</category>
      <category>datascience</category>
      <category>machinelearning</category>
    </item>
    <item>
      <title>A SUMMARY OF EXPLATORY DATA ANALYSIS</title>
      <dc:creator>Elishiba Muigo</dc:creator>
      <pubDate>Thu, 12 Oct 2023 07:43:10 +0000</pubDate>
      <link>https://dev.to/elimuigo/a-summary-of-explatory-data-analysis-5ci4</link>
      <guid>https://dev.to/elimuigo/a-summary-of-explatory-data-analysis-5ci4</guid>
      <description>&lt;p&gt;&lt;strong&gt;&lt;u&gt;The Significance of EDA&lt;/u&gt;&lt;/strong&gt;&lt;/p&gt;

&lt;p&gt;The EDA is an approach to analyzing datasets and involves using various tools and techniques to examine and understand data. EDA helps analysts gain insights into the data, identify relationships, detect outliers and prepare the data for further analysis or modelling. Process of visually and statistically summarizing data to discover it's underlying, structure, distribution and relationships between variables.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;1.Data Collection-&lt;/strong&gt; gather relevant data to the study or problem. This could be collected through various resources databases, spreadsheets, APIs and web scraping.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;2.Data Cleaning-&lt;/strong&gt; this a step that involves removing and handling missing values, as the incorrect insights. Deal with the outliers that might skew the analysis and address duplicates and inconsistent values.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;3. Data Wrangling-&lt;/strong&gt;this is converting raw data into a usable form . Involves merging multiple data sources into a single dataset for analysis. &lt;/p&gt;

&lt;p&gt;&lt;strong&gt;4.Statistics Summary-&lt;/strong&gt;calculates and visualize basic summary statistics like mean, median, standard deviation and quartiles for numerical variables.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;5.Data Visualization-&lt;/strong&gt;graphical representation of data to enhance understanding of the patterns, trends and insights within the data. They include: &lt;br&gt;
        -&lt;em&gt;&lt;b&gt;Bar Charts&lt;/b&gt;&lt;/em&gt; :To show frequency of categorical variables.&lt;br&gt;
        -&lt;em&gt;&lt;b&gt;Scatter plots&lt;/b&gt;&lt;/em&gt;: To explore relationships between numerical variables.&lt;br&gt;
        -&lt;em&gt;&lt;b&gt;Histograms&lt;/b&gt;&lt;/em&gt;: For visualizing the distribution of single variables.&lt;br&gt;
        -&lt;em&gt;&lt;b&gt;Heatmaps&lt;/b&gt;&lt;/em&gt;: To visualize correlations between variables.&lt;br&gt;
        -&lt;em&gt;&lt;b&gt;Box Plots&lt;/b&gt;&lt;/em&gt;: displays the distribution of a dataset, including the &lt;br&gt;
          median, quartiles and potential outliers.&lt;br&gt;
&lt;strong&gt;6.Correlation Analysis-&lt;/strong&gt;calculate correlation coefficients to understand the relationships between numerical variables e.g. Pearson, Spearman and then visualize correlations using matrices or heatmaps.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;7.Feature Engineering-&lt;/strong&gt;EDA can lead to the selection or creation of relevant features for predictive modeling. By exploring relationships between features and the target variable, data scientists can identify the most informative variables.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;8.Model Assumptions-&lt;/strong&gt; Understanding data distributions and relationships helps in selecting appropriate modeling techniques and verifying model assumptions. For instance, linear regression assumes a linear relationship between variables.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;&lt;u&gt;Methods of EDA&lt;/u&gt;&lt;/strong&gt;&lt;br&gt;
&lt;strong&gt;1.Time Analysis-&lt;/strong&gt;EDA involves examining the temporal aspects of data, including trends, seasonality, and autocorrelation, through techniques like time series plots and autocorrelation functions.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;2.Univariate Analysis-&lt;/strong&gt; Focuses on a single variable at a time, summarizing its central tendencies, spread, and distribution using measures like mean, median, standard deviation, and visualizations such as histograms and box plots.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;3.Bivariate Analysis-&lt;/strong&gt;Explores the relationships between two variables. Scatter plots, correlation coefficients, and contingency tables are useful tools in this context.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;4.Multivariate Analysis-&lt;/strong&gt; Involves studying the relationships between multiple variables simultaneously. Techniques like principal component analysis (PCA) or clustering methods can be employed for dimensionality reduction and pattern discovery.&lt;/p&gt;

</description>
    </item>
    <item>
      <title>ROADMAP TO BECOMING A DATA SCIENTIST</title>
      <dc:creator>Elishiba Muigo</dc:creator>
      <pubDate>Sat, 30 Sep 2023 01:12:14 +0000</pubDate>
      <link>https://dev.to/elimuigo/roadmap-to-becoming-a-data-scientist-4dph</link>
      <guid>https://dev.to/elimuigo/roadmap-to-becoming-a-data-scientist-4dph</guid>
      <description>&lt;p&gt;&lt;u&gt;&lt;/u&gt;&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;-Learn basic fundamentals of python&lt;/strong&gt; (control structures, data types, syntax, functions, Object Oriented Programming and data structures).&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;-Build a core of Statistics and statistical models:&lt;/strong&gt;(hypothesis testing, regression analysis, calculus, probability and linear algebra to help you in drawing insights and making informed decisions.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;-Data collection/scraping&lt;/strong&gt;(learn to collect data from relevant sources from various sources including databases, web scraping using(scrapy, Beautiful soup) and APIs.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;-Data cleaning:&lt;/strong&gt; Learn to clean data in order to get rid of anomalies this can be done using python libraries(numpy, pandas).Learn how to handle missing values and inconsistencies. &lt;/p&gt;

&lt;p&gt;&lt;strong&gt;-Learn the fundamentals of SQL&lt;/strong&gt;(Aggregation Functions, Joins, Nested queries,&lt;br&gt;
CREATE TABLE,GROUP BY, ALTER, INSERT, DELETE, DROP TABLE, ORDER BY, UPDATE).Gain proficiency in database management and working with large datasets.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;-Data visualization:&lt;/strong&gt; (Visualize data through chart and bar graphs in order to understand trends, variations and drive meaningful insights from the data. Gain familiarity with tools like Tableau, Power BI, and python packages(matplotlib, Seaborn).&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;-Become proficient in Machine Learning:&lt;/strong&gt; Be well versed with different machine learning models and when they are used based on the problem and data characteristics(Linear Regression, Naive Bayes, Reinforcement Learning, Random Forest, Decision Trees, Neural Networks).This helps in training models and achieve high accuracy and precision.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;-Learn Git and version control for Augmented Analysis.&lt;/strong&gt;Helps in collaborating with others in doing projects and tracks the changes you make and save them.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;-Learn soft skills including communication&lt;/strong&gt;. Learn how to story tell about data to both technical and non-technical audiences. Be able to summarize the findings in a clear and understandable manner and provide recommendations and insights based on the insights.&lt;/p&gt;

</description>
    </item>
  </channel>
</rss>
