<?xml version="1.0" encoding="UTF-8"?>
<rss version="2.0" xmlns:atom="http://www.w3.org/2005/Atom" xmlns:dc="http://purl.org/dc/elements/1.1/">
  <channel>
    <title>DEV Community: Awasume Marylin</title>
    <description>The latest articles on DEV Community by Awasume Marylin (@awasume_marylin_8).</description>
    <link>https://dev.to/awasume_marylin_8</link>
    <image>
      <url>https://media2.dev.to/dynamic/image/width=90,height=90,fit=cover,gravity=auto,format=auto/https:%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Fuser%2Fprofile_image%2F2733886%2F4a8c5f7b-c683-4314-b19e-dd9134be10c0.png</url>
      <title>DEV Community: Awasume Marylin</title>
      <link>https://dev.to/awasume_marylin_8</link>
    </image>
    <atom:link rel="self" type="application/rss+xml" href="https://dev.to/feed/awasume_marylin_8"/>
    <language>en</language>
    <item>
      <title>Presentations Tips: What to know on data analysis presentation Phase 1.</title>
      <dc:creator>Awasume Marylin</dc:creator>
      <pubDate>Thu, 20 Mar 2025 17:43:39 +0000</pubDate>
      <link>https://dev.to/awasume_marylin_8/presentations-tips-what-to-know-on-data-analysis-presentation-phase-1-517e</link>
      <guid>https://dev.to/awasume_marylin_8/presentations-tips-what-to-know-on-data-analysis-presentation-phase-1-517e</guid>
      <description>&lt;p&gt;I learnt about DATA STORYTELLING.&lt;br&gt;
 It is communicating the meaning of a dataset with visuals and a narrative that are customized for each particular audience. There are 3 data story steps.&lt;br&gt;
 💠 Engage your audience &lt;br&gt;
 💠 Create compelling visuals&lt;br&gt;
 💠Tell the story in an interesting narrative.&lt;br&gt;
 SPOTLIGHTING is scanning through data to quickly identify the most important insights.&lt;br&gt;
 DASHBOARDS is a tool that organizes information from multiple datasets into one central location for tracking, analysis and simple visualization through tables, charts and graphs.&lt;br&gt;
 LIFE AND STATIC INSIGHTS&lt;br&gt;
 To identify whether data is live or static depends on certain factors.&lt;br&gt;
 - How old is the data?&lt;br&gt;
 - How long untill the insights are stable or no longer valid to make &lt;br&gt;
 decisions&lt;br&gt;
 - Does this data or analysis need updating on a regular basis to remain &lt;br&gt;
 valuable?&lt;br&gt;
 &lt;br&gt;
 ➡️ STATIC DATA: Involves providing screenshots or snapshots in presentations or building dashboards using snapshots of data. There are PROS &lt;br&gt;
and CONS to Static data.&lt;br&gt;
 ⭕ PROS &lt;br&gt;
 - Can tightly control a point-in-time narrative of the data and insights. &lt;br&gt;
 - Allows for complex analysis to be explained in depth to a large audience &lt;br&gt;
 ⭕ CONS&lt;br&gt;
 - Insights immediately begins to lose value and continues to do so the l&lt;br&gt;
 longer the data remains in a static state &lt;br&gt;
 - Snapshots cant keep up with the pace of data change&lt;br&gt;
 ➡️ LIVE DATA: Means that you can build dashboards, reports and views &lt;br&gt;
 connected to automatically updated data. &lt;br&gt;
 ⭕ PROS &lt;br&gt;
 - Dashboards can be built to be more dynamic and scalable&lt;br&gt;
 - Gives the most up to date data to the people who need it at the time&lt;br&gt;
 when they need it&lt;br&gt;
 - Allows for up-to-date curated views into data with the ability to build a&lt;br&gt;
 scalable " single source of truth" for various use cases&lt;br&gt;
 - Allows for immediate action to be taken on data that changes &lt;br&gt;
 frequently &lt;br&gt;
 - Alleviates time/resources soent on processes for every analysis&lt;br&gt;
 ⭕ CONS&lt;br&gt;
 - Can take engineering resources to keep pipelines lives and scalable&lt;br&gt;
 which may be outside the scope of some companies data resources&lt;br&gt;
 allocation&lt;br&gt;
 - Without the ability to interpret data you can lose control of the&lt;br&gt;
 narrative, which can cause data chaos (ie teams coming to conflicting&lt;br&gt;
 conclusions based on the same data)&lt;br&gt;
 - Can potentially cause a lack of trust if the data isnt handled properly&lt;br&gt;
I also learnt about how to SHARE DATA which is all about DATA PRESENTATION or SLIDE PRESENTATION.&lt;br&gt;
 When exploring a slide presentation, use your knowledge of effective &lt;br&gt;
 presentation practices to evaluate it. This includes REVIEWING YOUR OWN &lt;br&gt;
 WORK. When checing over slide presentations there are some best&lt;br&gt;
 practices you can check for.&lt;br&gt;
 💠Include title, subtitle, and date. Making sure that your deck presentation &lt;br&gt;
 has a title, subtitle and date makes sure that your audience knows &lt;br&gt;
 exactly what you are presenting and when the information was from.&lt;br&gt;
 💠Use a logical sequence of slides&lt;br&gt;
 💠 Provide an agenda with a timeline &lt;br&gt;
 💠 Limit the amount of text on slides &lt;br&gt;
 💠 Start with business task&lt;br&gt;
 💠 Establish the initial hypothesis &lt;br&gt;
 💠 Show what business metrics you used&lt;br&gt;
 💠 Use visualizations&lt;br&gt;
 💠Introduce the graphic by name &lt;br&gt;
 💠 Provide a title for each graph&lt;br&gt;
 💠 Go from general to specifics&lt;br&gt;
 💠 Use speaker notes to help you remember talking points&lt;br&gt;
 💠 Include key take aways.&lt;/p&gt;

</description>
    </item>
    <item>
      <title>Mastering data visualization: Turning numbers into something that matters</title>
      <dc:creator>Awasume Marylin</dc:creator>
      <pubDate>Mon, 17 Feb 2025 21:32:50 +0000</pubDate>
      <link>https://dev.to/awasume_marylin_8/week-5-of-my-data-analytics-journey-4dj8</link>
      <guid>https://dev.to/awasume_marylin_8/week-5-of-my-data-analytics-journey-4dj8</guid>
      <description>&lt;p&gt;I am in the 5th phase of Data Analytics which is the SHARE PHASE&lt;br&gt;
 Recall&lt;br&gt;
 ➡️ ASK PHASE&lt;br&gt;
 ➡️ PREPARE PHASE&lt;br&gt;
 ➡️ PROCESS PHASE&lt;br&gt;
 ➡️ ANALYZE PHASE&lt;br&gt;
 ⭕ SHARE PHASE (VISUALIZATION PHASE)&lt;br&gt;
 ➡️ ACT PHASE&lt;br&gt;
 &lt;br&gt;
 DATA VISUALIZATION is the graphical representation and presentation of your data. &lt;br&gt;
 FRAMEWORKS help organize your thoughts about data visualization and gives you a useful checklist to reference as you plan and evaluate your data visualization.&lt;br&gt;
 There are 2 frameworks that employ slightly different techniques. Both are intended to improve the quality of your visuals.&lt;br&gt;
 - The McCandles Method: It has 4 elements of good data &lt;br&gt;
 visualization. &lt;br&gt;
 💠 Information: The data which you're working with.&lt;br&gt;
 💠 Story: A clear and compelling narrative or concept.&lt;br&gt;
 💠 Goal: A specific objective or function for the visual.&lt;br&gt;
 💠 Visual Form: An effective use of metaphor or visual &lt;br&gt;
 expression.&lt;br&gt;
 - Kaiser Fung's Junk Charts Trifecta Checkup: This approach is a set of questions that can help consumers of data visualization critque what they are consuming and determine how effective it is.&lt;br&gt;
You can also use these questions to determine if your data visualization is effective.&lt;br&gt;
 💠 What is the practical question?&lt;br&gt;
 💠 What does the data say&lt;br&gt;
 💠 What does the visual say&lt;br&gt;
The essential building blocks that make visuals immediately understandable are called MARKS and CHANNELS.&lt;br&gt;
 ➡️ MARKS are basic visual objects such as points, lines and shapes. Every&lt;br&gt;
 mark can be broken down into qualties&lt;br&gt;
 ⭕ Position: Where is a specific mark in space relative&lt;br&gt;
 to a scale or to other marks?&lt;br&gt;
 ⭕ Size: How big, small, long or tall is a mark?&lt;br&gt;
 ⭕ Shape: Does the shaoe of the object communicate &lt;br&gt;
 something about it?&lt;br&gt;
 ⭕ Color: What color is the mark?&lt;br&gt;
 &lt;br&gt;
 ➡️ CHANNELS are visual aspects or variable that represent characteristics&lt;br&gt;
 of the data in a visualization. They are basically visualized marks that &lt;br&gt;
 have been used to visualize data.&lt;br&gt;
 ⭕ Accuracy: Are channels helpful to accurately estimate&lt;br&gt;
 the values being represented?&lt;br&gt;
 ⭕ Pop out: How easy is it to distinguish certain &lt;br&gt;
 values from others?&lt;br&gt;
 ⭕ Grouping: How effective is a channel at communicating &lt;br&gt;
 groups that exist in data?&lt;br&gt;
the different types of visuals to use when you want to visualize your data.&lt;br&gt;
 &lt;br&gt;
 ➡️ BAR GRAPHS: Use size contrast to compare two or more values.&lt;br&gt;
 ➡️ PIE CHARTS: Shows how much each part of something makes up a&lt;br&gt;
 whole.&lt;br&gt;
 ➡️ MAPS: Helps organize data geographically.&lt;br&gt;
 ➡️ HISTOGRAM: A chart that shows how often data values fall into certain&lt;br&gt;
 range.&lt;br&gt;
 ➡️ LINE CHART: Is used to track changes over short and long periods of &lt;br&gt;
 time. When smaller changes exists, lines charts are better to use than &lt;br&gt;
 BAR CHARTS.&lt;br&gt;
 ➡️ COLUMN CHARTS: Use size to contrast and compare two or more&lt;br&gt;
 values. Using height or lenghts to represent the specific values.&lt;br&gt;
 ➡️ HEAT MAP: Similar to BAR CHART, heat maps also use color to compare&lt;br&gt;
 categories in a dataset. &lt;br&gt;
 ➡️ SCATTER PLOT: Show relationships between different variables. &lt;br&gt;
 Scatterplots are typically used for 2 variables for a set of data&lt;br&gt;
 although additional variables can be displayed.&lt;br&gt;
 ➡️ DITRIBUTION GRAPH: Displaysy the spread of various outcome &lt;br&gt;
 in a dataset.&lt;br&gt;
 ➡️ CORRELATION CHARTS: Shows relationship among data. It is the&lt;br&gt;
 measure of the degree to which 2 variables move in relation to each &lt;br&gt;
 other. If one variable goes up and the other one goes down, it&lt;br&gt;
 is a NEGATIVE or INVERSE CORRELATION. If one variable goes up and the &lt;br&gt;
 other one goes up, it is a POSITIVE CORRELATION. If one variable goes &lt;br&gt;
 up and the other values stays thesame , there is no CORRELATION.&lt;br&gt;
 CAUASATION refers to the idea that an event leads to a specific outcome&lt;br&gt;
 STATISTIC VISUALIZATION do not change over time unless they are edited&lt;br&gt;
 DYNAMIC VISUALIZATION are interactive and they change overtime.&lt;br&gt;
 There are meaningful patterns in data visualization, and they can take many forms&lt;br&gt;
 &lt;br&gt;
 💠 CHANGE. This is a trend or instance of observation that becomes &lt;br&gt;
 different overtime. A great way to measure change in data is through &lt;br&gt;
 a LINE CHART or COLUMN CHART.&lt;br&gt;
 💠 CLUSTERING. A collection of data points with similar or differnt values. &lt;br&gt;
 This is best represented through a DISTRIBUTION GRAPH.&lt;br&gt;
 💠 RELATIVITY: These are observations considerec in relation or in &lt;br&gt;
 proportion to something else. They are best represented using PIE&lt;br&gt;
 CHART.&lt;br&gt;
 💠 RANKING. This is a position in a scale of achievement or status. Data&lt;br&gt;
 that requires ranking is best represented by a column chart .&lt;br&gt;
I also learnt about DECISION TREES. Data grows on DECISION TREES.&lt;br&gt;
 A DECISION TREE is a decision making tool that allows you the data analyst, to make decisionss based on key questions that you can ask youself.&lt;br&gt;
 I also learnt about the 9 basic principles of design;&lt;br&gt;
 💠 BALANCE. The design of a data visualization is balanced when the key visual elements like color and shape are distributed evenly.&lt;br&gt;
 💠 EMPHASIS. Your data visualization should have a focal point, so that your audience knows where to concerntrate. In otherwords, your visualizations should emphasize the most important data so that users recognize it first. You can use colors and values.&lt;br&gt;
 💠 MOVEMENT. It can refer to the path the viewers eye travel as they look at a data visualizations or literal movement crreated by animation.&lt;br&gt;
 💠 PATTERNS. You can use similar shapes and colors to create patterns in your data visualization. You can sue patterns to highlight similarities between different datasets, or break up a pattern with a unique shape, color, or line to create more emphasis.&lt;br&gt;
 💠 REPITITION. Repeating chart types, shapes, or colors adds to the effectiveness of our visualization.&lt;br&gt;
 💠 PROPORTION. Using various colors and sizes helps demonstrate that you are calling attention to a specific visual over others.&lt;br&gt;
 These first 6 principles of design are key considerations that you can make while you are creating your data visualization.&lt;br&gt;
 These next 3 principles are useful checks once your data visualization is finished&lt;br&gt;
 💠 RYTHM. This refers to creating a sense of movement or flow in your visualization. If your finished design doesnt succesffully create a flow. You might want to rearrange some of the elements to improve the rythm.&lt;br&gt;
 💠 VARIETY. Your visualizations should have some variety in the chart type, lines, shapes, colors and values you use. Variety keeps the audience engaged but it is good to find balancesince too much variety can confuse people.&lt;br&gt;
 💠 UNITY. Your final data visualization should be cohesive. If the visual is disjointed or not well organized, it will be confusing and overwhelming.&lt;br&gt;
 ELEMENTS OF ARTS&lt;br&gt;
 - LINE&lt;br&gt;
 - SHAPE&lt;br&gt;
 -COLOR&lt;br&gt;
 -SPACE&lt;br&gt;
 -MOVEMENT&lt;br&gt;
 &lt;br&gt;
 &lt;br&gt;
 5 PHASES OF THE DESIGN PROCESS &lt;br&gt;
 ➡️ EMPHATHIZE: Thinking about the emotions and needs of the target audince for the data visualization.&lt;br&gt;
 ➡️ DEFINE: Figuring out exactly what your audience needs from the data.&lt;br&gt;
 ➡️ IDEATE: Generating ideas for data visualization.&lt;br&gt;
 ➡️ PROTOTYPE: Putting visualization together for testing and feedback.&lt;br&gt;
 ➡️ TEST: Showing prototype visualizations to people before stakeholders see them.&lt;br&gt;
I learnt about data visualization with TABLEAU&lt;br&gt;
 TABLEAU is a business intelligence and analytics plateform that helps people&lt;br&gt;
 see, understand and make decisions with data.&lt;br&gt;
 There are DESIGN PRINCIPLES IN TABLEAU which are;&lt;br&gt;
 - Choose the right visuals&lt;br&gt;
 - Optimize the data-ink ratio&lt;br&gt;
 - Use orientation effectively &lt;br&gt;
 - Number of elements&lt;br&gt;
 - Avoid misleading and or deceptive charts&lt;br&gt;
 Some common mistakes to avoid so that your visualizations aren't &lt;br&gt;
 accidentally misleading.&lt;br&gt;
 - Cutting of the y-axis&lt;br&gt;
 - Misleading use of a dual &lt;br&gt;
 - Artifically limiting the scope of your data (not showing all of the data)&lt;br&gt;
 - Problematic choices in how data is binned or grouped&lt;br&gt;
 - Using part to whole visuals when the totals do not sum up appropriately&lt;br&gt;
 - Hiding trends in cummulative charts&lt;br&gt;
 - Artificially smoothing trends&lt;br&gt;
 &lt;br&gt;
 Few rules about what makes a helpful data visualization.&lt;br&gt;
 &lt;br&gt;
 💠 FIVE SECOND RULE: A data visualization should be clear, effective and &lt;br&gt;
 convincing enough to be absorbed in five seconds or less.&lt;br&gt;
 💠 COLOR CONTRAST : Graphs and charts should use a diverging color &lt;br&gt;
 palette to show contrast between elements.&lt;br&gt;
 💠 CONVENTIONS and EXPECTATIONS: Visuals and their organization&lt;br&gt;
 should align with audience expectations and cultural conventions.&lt;br&gt;
 💠 MINIMAL LABELS: Title, axis and annotations should use a few labels&lt;br&gt;
 as it takes to make use. Having too many labels makes your graphs&lt;br&gt;
 or charts too busy. It takes up too much space and prevents the labels &lt;br&gt;
 from being shown clearly.&lt;/p&gt;

</description>
    </item>
    <item>
      <title>Excel, SQL, and R: The power trio that transformed my data skills</title>
      <dc:creator>Awasume Marylin</dc:creator>
      <pubDate>Tue, 11 Feb 2025 22:37:18 +0000</pubDate>
      <link>https://dev.to/awasume_marylin_8/week-4-of-my-data-analytics-journey-3bie</link>
      <guid>https://dev.to/awasume_marylin_8/week-4-of-my-data-analytics-journey-3bie</guid>
      <description>&lt;p&gt;I am now at the ANALYSIS PHASE.&lt;br&gt;
 Recall&lt;br&gt;
➡️ ASK PHASE&lt;br&gt;
➡️ PREPARE PHASE&lt;br&gt;
➡️ PROCESS PHASE&lt;br&gt;
⭕ ANALYSIS PHASE&lt;br&gt;
➡️ SHARE PHASE&lt;br&gt;
➡️ ACT PHASE&lt;br&gt;
 This is the process used to make sense of the data collected. The goal of analysis is to identify trends and relationships with in data so you can accurately answer the questions you are asking.&lt;br&gt;
 &lt;br&gt;
 The four phases of analysis&lt;br&gt;
 💠 Organize data&lt;br&gt;
 💠 Format and adjust data&lt;br&gt;
 💠 Get input from others&lt;br&gt;
 💠 Transform data&lt;br&gt;
This phase includes SORTING and FILTERING &lt;br&gt;
 - SORTING is the process of arranging data into a meaningful order to make &lt;br&gt;
 it easier to understand, analyze and visualize&lt;br&gt;
 - FILTERING is used to show only the data that meets a specified criteria while &lt;br&gt;
 hiding the rest. It is useful when you have a lot of data.&lt;br&gt;
 I also did practical work on SORTING and FILTERING data using MICROSOFT EXCEL and SQL&lt;br&gt;
 CASE STUDY: MOVIE DATA, COUNTY NATALITY.&lt;br&gt;
In SQL, i used the 'ORDER BY' and 'WHERE' clauses to sort and filter data.&lt;br&gt;
i learnt FORMATTING FOR BETTER ANALYSIS.&lt;br&gt;
 ➡️ Convert functions to change units of measurement ( in Excel)&lt;br&gt;
 &lt;br&gt;
Convert data in spreadsheet&lt;br&gt;
➡️ Convert from STRING to DATE, Converting TEXT values containing both &lt;br&gt;
 DATES and TIME&lt;br&gt;
➡️ Convert from STRING to NUMBER&lt;br&gt;
➡️ Combining columns (merge text from 2 or more cells) using AMPERSAND&lt;br&gt;
 SYMBOL and CONCATENATE&lt;br&gt;
➡️ Convert from NUMBER to PERCENTAGE&lt;br&gt;
 DATA VALIDATION IN EXCEL&lt;br&gt;
 You can use data validation in excel to&lt;br&gt;
 - Add dropdown lists with predertermined options&lt;br&gt;
 - Create customs checkboxes&lt;br&gt;
 - Protect structured data formulas&lt;br&gt;
 CONDITIONAL FORMATTING&lt;br&gt;
 A spreadsheet tool thst changes how cell appear when values meet specific&lt;br&gt;
 conditions.&lt;br&gt;
 Transforming data type in SQL using the CAST FUNCTION&lt;br&gt;
 - Converting a NUMBER to a STRING&lt;br&gt;
 -Converting a STRING to a NUMBER&lt;br&gt;
 -Converting DATE to STRING&lt;br&gt;
 -Convert DATE to DATETIME&lt;br&gt;
 I also learnt how to combine 2 or more cells in SQL using the CONCATENATE FUNCTION.&lt;br&gt;
 Using the CONCATENATE FUNCTION. In order combine data helps analyst to properly and efficiently analyze data, the data has to be clean and understandable. Data anaylyst use functions like CONCAT to make data easier to work with, which may require combining multiple cells.&lt;br&gt;
i learnt about DATA AGGREGATION, VLOOKUP and JOIN &lt;br&gt;
DATA AGGREGATION is the process of gathering data from multiple sourrces in order to combine it in to a single summarized collection &lt;br&gt;
 - Puzzle pieces = data&lt;br&gt;
 - Organization = aggregation&lt;br&gt;
 - Pile of piece = Summary&lt;br&gt;
 - Putting the pieces together = gaining insights&lt;br&gt;
 So DATA AGGREGATION helps analyst to &lt;br&gt;
 ➡️ Identify trends &lt;br&gt;
 ➡️ Make comparisons&lt;br&gt;
 ➡️ Gain/insights&lt;br&gt;
 Data can also be aggregated over a given time period to provide statistics such as&lt;br&gt;
 ➡️ MINS&lt;br&gt;
 ➡️ AVERAGES&lt;br&gt;
 ➡️ MAXS&lt;br&gt;
 ➡️ SUMS&lt;br&gt;
 VLOOKUP stands for Vertical lookup. It is a function that searches for a certain &lt;br&gt;
 value in a column to return a corresponding piece of information.&lt;br&gt;
 VLOOKUP can match 2 sheets together on a matching column to populate on a single sheet .&lt;br&gt;
It only returns the data it finds to the right it cant look left. &lt;br&gt;
JOIN is an sql clause that is used to combine rows from 2 or more tables based on a relatable column. It is like VLOOKUP in sql.&lt;br&gt;
 COMMON JOINS&lt;br&gt;
 💠 INNER JOIN: A function that returns records with matching values in &lt;br&gt;
 both tables.&lt;br&gt;
 💠 LEFT JOIN: A function that will return all the records from the left table &lt;br&gt;
 and only the matching records from the right table&lt;br&gt;
 💠 RIGHT JOIN: Return all records from the right table and only the &lt;br&gt;
 the matching records from the left&lt;br&gt;
 💠 OUTER JOIN: Combines right and left join to return all matching records &lt;br&gt;
 in both tables&lt;br&gt;
 I also learnt about the importance of ALIASES.&lt;br&gt;
 ALIASES are used in sql queries to creat temporary names for a column or&lt;br&gt;
 table. They are implememnted by making use of AS command.&lt;br&gt;
I learnt about ,&lt;br&gt;
 SUBQUERIES: They are queries within queries. They are called an Inner or Nested query. &lt;br&gt;
 - They can make projects easier and more efficient by allowing complex operations to be performed in a single query, reducing the need for multiple trips to the database.&lt;br&gt;
 - They also make your code readable and maintainable&lt;br&gt;
 Rules to follow to use subquery&lt;br&gt;
 ➡️ Subqueries must be enclosed within parentheses.&lt;br&gt;
 ➡️ A subquery can have one or more columns specified in the select&lt;br&gt;
 clause&lt;br&gt;
 ➡️ Subqueries that return more than one row can only be used with&lt;br&gt;
 multiple values operators, such as the IN operator which alows&lt;br&gt;
 you to specify multiple values in WHERE clause.&lt;br&gt;
 ➡️ Subquery cant be nested in SET command.&lt;br&gt;
 I also learnt about PIVOT TABLES in google sheets. They make it possible to view data in multiple ways in order to identify insights and trends. &lt;br&gt;
 &lt;br&gt;
 They can help you quickly make sense of larger datasets by comparing metrics, performing calcularions and generating reports. &lt;br&gt;
 Pivot tables have 4 basic parts&lt;br&gt;
💠 ROWS: Organize and group data you select horizontaly&lt;br&gt;
💠 COLUMNS: Organize and displays values from your data vertically.&lt;br&gt;
💠 FILTERS: Enables you to apply filters based on a specific criteria&lt;br&gt;
💠 VALUES: They are used to calculate and count data.&lt;br&gt;
I also learnt about DATA VALIDATIONS and the TYPES OF DATA VALIDATION&lt;br&gt;
 DATA VALIDATION is the process of checking and rechecking the quality of your data so that it is complete, accurate, secure and consistent.&lt;br&gt;
 TYPES OF DATA VALIDATION&lt;br&gt;
 ⭕ DATA TYPES :Check that the data type matches the data type defined for a field &lt;br&gt;
 ⭕ DATA RANGE: Check that the data falls within an acceptable range of values defined for the field.&lt;br&gt;
 ⭕ DATA CONSTRAINTS: Check that the data meets certain conditions or criteria for a field. This includes the type of data entered as well as other attributes of the field such as number of characters.&lt;br&gt;
 ⭕ DATA CONSISTENCY: Check that the data makes sense in the context of other related data.&lt;br&gt;
 ⭕ DATA STRUCTURE: Check that the data follows or conforms to a set structure&lt;br&gt;
 ⭕ CODE VALIDATION: Check that the application code systematically performs any of the previously mentioned validation during user data input.&lt;br&gt;
 &lt;br&gt;
 I also learnt about TEMPORARY TABLE&lt;br&gt;
 &lt;br&gt;
 They are tables that is created and exists temporarily on a database server. You create a temporary table in SQL using the WITH clause which is a type of temporary tables that you can query from multiple times .They are extremely helpful for complex calculations queries.&lt;/p&gt;

</description>
    </item>
    <item>
      <title>'Data-driven decisions: How i learned to ask the right questions with analytics</title>
      <dc:creator>Awasume Marylin</dc:creator>
      <pubDate>Sun, 02 Feb 2025 19:27:38 +0000</pubDate>
      <link>https://dev.to/awasume_marylin_8/week-3-of-my-data-analytics-journey-19k3</link>
      <guid>https://dev.to/awasume_marylin_8/week-3-of-my-data-analytics-journey-19k3</guid>
      <description>&lt;p&gt;In this new week i am learning all about the process phase of data analytics&lt;br&gt;
 ‼️ Recall the 6 phases of data analytics&lt;br&gt;
 - ASK PHASE&lt;br&gt;
 - PREPARE PHASE&lt;br&gt;
 ⭕ PROCESS PHASE&lt;br&gt;
 - ANALYZE PHASE&lt;br&gt;
 -SHARE PHASE&lt;br&gt;
 - ACT PHASE&lt;br&gt;
 In the lesson of today, i learnt all about DATA INTERGRITY. &lt;br&gt;
 It can be defined as the ACCURACY, COMPLETENESS, CONSISTENCY AND TRUSTWORTHYNESS of data throughout its life cycle.&lt;br&gt;
 &lt;br&gt;
One missing piece can make all of your data useless. Data can also be compromised, and they can be compromised in many ways. They can be compromised during:&lt;br&gt;
 ➡️ DATA REPLICATION which is the process of storing data in multiple &lt;br&gt;
 locations&lt;br&gt;
 ➡️ DATA TRANSFER which is process of copying data from a storage &lt;br&gt;
 device to memory or from one computer to another If the transfer&lt;br&gt;
 is interupted, you may end up with incomplete dataset.&lt;br&gt;
 ➡️ DATA MANIPULATION process involves changing data to make &lt;br&gt;
 it organized and easier to read.&lt;br&gt;
Data can also be compromised through &lt;br&gt;
 ➡️ Human error&lt;br&gt;
 ➡️ Viruses&lt;br&gt;
 ➡️ Malware&lt;br&gt;
 ➡️ Hacking&lt;br&gt;
 ➡️ System failures&lt;br&gt;
It is important to check that the data you use aligns with the business obectives. &lt;br&gt;
When you are getting ready for data analysis, you might realize you dont have the data you need or you dont have enough of it In some cases you use what is called PROXY DATA in place of the real data.&lt;br&gt;
 There are different types of insufficient data which include;&lt;br&gt;
 💠 Data from only one source&lt;br&gt;
 💠 Data that keeps updating&lt;br&gt;
 💠 Outdated data&lt;br&gt;
 💠 Geographyical limited data&lt;br&gt;
There are ways we can address insufficient data problems.&lt;br&gt;
 💠 Identify trends with the available data&lt;br&gt;
 💠 Wait for more data if time allows&lt;br&gt;
 💠 Talk with stakeholders and adjust your objectives&lt;br&gt;
 💠 Look for new data set&lt;br&gt;
 I also learnt about sample size calculation&lt;br&gt;
 - POPULATION: The entire group that you are interested in for study&lt;br&gt;
 - SAMPLE: A subset of your population&lt;br&gt;
 - MARGIN OF ERROR: Since a sample is used to represent a population&lt;br&gt;
 the sample results are expected to differ from what the results would &lt;br&gt;
 have been if you had surveyed the entire population.The difference is&lt;br&gt;
 called MARGIN OF ERROR&lt;br&gt;
 - CONFIDENCE LEVEL: How confident you are in the survey results&lt;br&gt;
 - CONFIDENCE INTERVAL: The range of possible values that the &lt;br&gt;
 population's result would be at the confidence level of the study.&lt;br&gt;
 This range is the sample results +/- the margin of error.&lt;br&gt;
 - STATISTICAL SIGNIFANCE: The determination of whether your result&lt;br&gt;
 could be due to random or not. The greater the signifance, the less&lt;br&gt;
 due to chance&lt;br&gt;
i learned about the task you should complete before analyzing data which are;&lt;br&gt;
 - Determine data intergrity by assessing the overall accuracy, consistency &lt;br&gt;
 completeness of the data.&lt;br&gt;
 - Connect objectives to data by understanding how your business &lt;br&gt;
 objectives can be served by an investigation in to the data.&lt;br&gt;
 - know when to stop collecting data.&lt;br&gt;
STATISTICL POWER is the probabilty of geting meaningful results from a test &lt;br&gt;
HYPOTHESIS TESTING is a way to see if a survey experiment has meaningful results.&lt;br&gt;
DATA CONSTRAINTS The criteria that determines if a piece of data is clean and valid.&lt;br&gt;
DATEDIF is a spreadsheet function that calculates the number of days, months, or years between 2 dates&lt;br&gt;
ESTIMATED RESPONSE RATE is the average number of people who typically complete a surey&lt;br&gt;
REGULAR EXPRESSION (RegEx) is a rule that says the values in a table must match a prescribed pattern &lt;br&gt;
 &lt;br&gt;
 The most common cause of quality data is HUMAN ERROR&lt;br&gt;
Clean data is data that is complete, correct and relevant to the problem you're trying to solve.&lt;br&gt;
Dirty data is data that is incomplete, incorrect and irrelevant to the problem you are trying to solve. There are different types of dirty data which include;&lt;br&gt;
 ➡️ Duplicate data which is any data record that shows up more than&lt;br&gt;
 once which could be caused by manual data entry, batch data&lt;br&gt;
 imports or data migration.&lt;br&gt;
 ➡️ Outdated data is any data that is old which should be replaced&lt;br&gt;
 with newer ad more accurate information.&lt;br&gt;
 ➡️ Incomplete data is any data that is missing important fields&lt;br&gt;
 ➡️ Incorrect or inaccurate data is any data that is complete&lt;br&gt;
 but inaccurate which could be caused by Human Error inserted&lt;br&gt;
 during data input, fake information or mock data&lt;br&gt;
 ➡️ Inconsistent data is data that uses different formats to represent &lt;br&gt;
 the thesame thing which could be caused by data stored &lt;br&gt;
 incorrectly or inserted during data transfer.&lt;br&gt;
Other causes of data include;&lt;br&gt;
 ➡️ Manual data entry&lt;br&gt;
 ➡️ Error batch data imports &lt;br&gt;
 ➡️ Data migration&lt;br&gt;
 ➡️ Software obsolescence&lt;br&gt;
 ➡️ Improper data collection&lt;br&gt;
 ➡️ Human errors during data input&lt;br&gt;
 DATA VALIDATION is a tool for checking the accuracy and quality of data before adding or importing it. There are steps you have to follow to clean data&lt;br&gt;
 💠 Removing unwanted data&lt;br&gt;
 💠 Getting rid of duplicates or data that is irrelevant to what you want to&lt;br&gt;
 solve&lt;br&gt;
 💠 Fixing misspellings&lt;br&gt;
 💠 Fixing inconsistent capitalization&lt;br&gt;
 💠 Fixing punctuation and other typos&lt;br&gt;
DATA MERGING is the process of combing 2 or more datasets into a single dataset&lt;br&gt;
DATA COMPATIBILITY describes how well 2 or more datasets are able to work together&lt;br&gt;
 &lt;br&gt;
 Common errors you should avoid are&lt;br&gt;
 💠 Not checking spelling errors&lt;br&gt;
 💠 Forgetting to document errors&lt;br&gt;
 💠 Not checking for misfielded values when values are entered in to the &lt;br&gt;
 wrong field&lt;br&gt;
 💠 Overlooking missing values&lt;br&gt;
 💠 Only looking at the subset of the values&lt;br&gt;
 💠 Loosing track of business objectives&lt;br&gt;
 💠 Not analyzing the system prior to data cleaning&lt;br&gt;
 💠 Not fixing the source of the error&lt;br&gt;
 💠Not backing up your data prior to data cleaning&lt;br&gt;
 💠 Not accounting for data cleaning in your deadlines/process.&lt;br&gt;
Today i also did 3 EXCEL exercises on data cleaning and it was very exciting and i will continue to practice everyday.&lt;br&gt;
i did more of practical work like cleaning of data in microsoft excel using different datasets.&lt;br&gt;
i used different functions to make sure that my data was clean ( removing duplicates, made sure the data format was consistent, removed unwanted spaces, split text and numbers to make reading easy)&lt;br&gt;
I used different excel functions to achieve my objective of having clean data;&lt;br&gt;
 💠 TRIM&lt;br&gt;
 💠 VLOOKUP&lt;br&gt;
 💠 SPLIT TEXT TO COLUMN&lt;br&gt;
 💠 DATA VALIDATION&lt;br&gt;
 💠 FILTER&lt;br&gt;
 💠 SPLIT&lt;br&gt;
 💠 CONDITIONAL FORMATTING&lt;br&gt;
 💠 PIVOT TABLES&lt;br&gt;
 💠 PASTE SPECIAL&lt;br&gt;
i learnt how to clean data using SQL &lt;br&gt;
➡️ REMOVING DUPLICATES USING "DISTINCTION"&lt;br&gt;
➡️ USED "SUBSTRING" TO GET ONLY THE FIRST 2 LETTERS OF A COUNTRY&lt;br&gt;
➡️ USED "LENGTH" TO KNOW HOW MANY LETTERS ARE USED TO REPRESENT A COUNTRY &lt;br&gt;
➡️ USED THE "TRIM" FUNCTION TO REMOVE SPACES&lt;br&gt;
I also did a few exercises to learn how to clean and adjust my data using &lt;br&gt;
➡️ UPDATE&lt;br&gt;
➡️ SET.&lt;br&gt;
DATA VERIFICATION is a process to confirm that a data-cleaning effort was well executed and the resulting data is accurate and reliable.&lt;br&gt;
 CHANGELOG is a file containing a chronological list of modifications made to a project. They are usually organized by version and include the date followed by a list of added, improved and removed features. &lt;br&gt;
 &lt;br&gt;
 See the big picture when verifying data cleaning &lt;br&gt;
 💠 Consider the business problem &lt;br&gt;
 💠 Consider the goal&lt;br&gt;
 💠 Consider the data&lt;br&gt;
 Data cleaning check list. Correct the most common problems&lt;br&gt;
 ➡️ Sources of error&lt;br&gt;
 ➡️ Null data (search using conditional formatting and filters)&lt;br&gt;
 ➡️ Misspelled words&lt;br&gt;
 ➡️ Mistyped numbers&lt;br&gt;
 ➡️ Extra spaces and characters (using TRIM functions)&lt;br&gt;
 ➡️ Duplicates (distinct in SQL or remove duplicates function in excel or&lt;br&gt;
 spreadsheet )&lt;br&gt;
 ➡️ Messy (inconsistent) strings&lt;br&gt;
 ➡️ Messy (inconsistent) data formats&lt;br&gt;
 ➡️ Misleading variable labels (column)&lt;br&gt;
 ➡️ Truncated or missing data&lt;br&gt;
 ➡️ Business logic (check that the data makes sense given your knowledge&lt;br&gt;
 of the business&lt;br&gt;
 ➡️ Review the goal of your project to make sure that your data still aligns&lt;br&gt;
 with the goal.This is a continuous process that you will do throughout&lt;br&gt;
 your project.&lt;br&gt;
 DOCUMENTATION&lt;br&gt;
 The process of tracking changes, additions, deletions and errors involved in your data cleaning efforts example CHANGELOG&lt;br&gt;
 &lt;br&gt;
 ⭕ Recover data cleaning errors to use whenwe come accross a similar &lt;br&gt;
 problem or when you want to redo the cleaning &lt;br&gt;
 ⭕ Inform other users of changes you have made (reference sheet for other &lt;br&gt;
 analyst incase you resign)&lt;br&gt;
 ⭕ Determine quality of data.&lt;/p&gt;

</description>
    </item>
    <item>
      <title>Data cleaning demystified: How i learned to transform messy data into gold</title>
      <dc:creator>Awasume Marylin</dc:creator>
      <pubDate>Sun, 26 Jan 2025 20:29:07 +0000</pubDate>
      <link>https://dev.to/awasume_marylin_8/my-data-analytics-jornal-35gf</link>
      <guid>https://dev.to/awasume_marylin_8/my-data-analytics-jornal-35gf</guid>
      <description>&lt;p&gt;WEEK 2 OF MY DATA ANALYTICS JOURNEY&lt;br&gt;
This week i learnt a all about PREPARE PHASE OF DATA ANALYTICS. &lt;br&gt;
DATA EXPLORATION. After asking the right questions whuch we saw in the previous days, PREPARING DATA CORRECTLY is the next step.&lt;br&gt;
N.B Data works best when it is organized.&lt;br&gt;
 Every piece of information is data and they are generated as a result of our activities in the world. We have SURVEY DATA which is generated using forms. INTERVIEWS can also generate data.&lt;br&gt;
 &lt;br&gt;
 💠 Forms, Questionaires and Surveys are commonly used &lt;br&gt;
 ways to collect and generate data.&lt;br&gt;
 💠Data that is generated online doesnt always happen directly&lt;br&gt;
 They use COOKIES which help them remember your preferences&lt;br&gt;
 &lt;br&gt;
 ⭕ COOKIES are small files stored on computer that contains information about users.&lt;br&gt;
 There are 3 different data sources which include:&lt;br&gt;
➡️ First party data which is data collected by an individual or group using their own resources. This is typically the prefered method of collecting data because you know exactly where it came from.&lt;br&gt;
➡️ Second part data which is collected by a group directly from its audience. &lt;br&gt;
➡️ Third part data which is data collected from outside sources who did not collect it directly. This source type might not be very reliable.&lt;br&gt;
⭕ A SAMPLE is part of a population that is representative of the population.&lt;br&gt;
 I learnt that selecting the right data is a key important in data analysis.&lt;br&gt;
 💠 When solving your business problems be sure to select data that can &lt;br&gt;
 solve your problem question.&lt;br&gt;
 💠 How much data to collect: If you are collecting your own data make &lt;br&gt;
 reasonable decisions about sample size.&lt;br&gt;
 💠 Time frame: If you are collecting your own data, decide how long you &lt;br&gt;
 will like to collect it especially if you are tracking trends over a long&lt;br&gt;
 period of time.&lt;br&gt;
 &lt;br&gt;
 DATA COLLECTION CONSIDERATION&lt;br&gt;
 ‼️ Select the right type of data&lt;br&gt;
 ‼️ Determine the time frame&lt;br&gt;
 ‼️ Decide how data will be collected (collection o new data). &lt;br&gt;
 Then decide how much data to be collected.&lt;br&gt;
 OR &lt;br&gt;
 ‼️ Choose data source (use existing data).&lt;br&gt;
 Deside what data to use.&lt;br&gt;
Learned about the different data formats which include&lt;br&gt;
 ➡️DESCRETE DATA which is data that is counted and has a &lt;br&gt;
 limted number of values.&lt;br&gt;
 ➡️ CONTINUOUS DATA is data that can have any numeric value&lt;br&gt;
 ➡️ NORMINAL DATA is a type of qualitative data that is categorized with &lt;br&gt;
 out a set order&lt;br&gt;
 ➡️ ORDINAL DATA is a type of qualitative data with out a set order or &lt;br&gt;
 scale&lt;br&gt;
 ➡️ INTERNAL DATA which is data that lives within a company's&lt;br&gt;
 own systems &lt;br&gt;
 ➡️ EXTERNAL DATA which is data that doesnt live within a company's&lt;br&gt;
 own system&lt;br&gt;
 ➡️ STRUCTURED DATA is data that is organized in a certain format such &lt;br&gt;
 as row and columns&lt;br&gt;
 ➡️ UNSTRUCTURED DATA is data that is not organized in a certain format &lt;br&gt;
 audio,video and images&lt;br&gt;
 &lt;br&gt;
 Data model is a model that is used for organizing data elements and how they relate to one another. Structured data works nicely within a data model.&lt;br&gt;
Data elements are pieces of information such as peoples names, account numbers and addresses.&lt;br&gt;
PHYSICAL DATA MODELING depicts how a database operates. A physical data model defines all entities and attributes used e.g it includes tables, names, columns name, and data types for the database.&lt;br&gt;
 2 common methods of approaching data modeling techniques include&lt;br&gt;
 ‼️ ENTITY RELATIONSHIP DIAGRAM (ERD) which are a visual way to &lt;br&gt;
 understand the relationship between entities in the data model.&lt;br&gt;
 ‼️ UNIFIED MODELING LANGUAGE (UML) diagrams are very detailed &lt;br&gt;
 diagrams that describe the structure of a system by showing the system's&lt;br&gt;
 entities, attributes, operations and their relationships.&lt;br&gt;
 Data type is a specific kind of data attribute that tells what kind of value the data is. It tells you what type of data you are working with. Data types can be dfferent depending on the query language you are using.&lt;br&gt;
 ➡️ BOOLEAN DATA TYPE is a type of data with only 2 possible values&lt;br&gt;
 which include; AND, OR, NOT, TRUE, FALSE&lt;br&gt;
 ⭕ AND OPERATOR is a symbol that lets you stack both your conditions&lt;br&gt;
 ⭕ OR OPERATOR lets you move forward if either one of your 2&lt;br&gt;
 conditions are met&lt;br&gt;
 ⭕ NOT OPERATOR lets you filter by subtracting specific conditions from&lt;br&gt;
 the results.&lt;br&gt;
 WIDE DATA: Is a data set in which every data subject has a single row with multiple columns to hold the values of the various attributes to the subject&lt;br&gt;
 LONG DATA: Is data in which each row represents one observation per subect , so each subect will be represented by multiple rows.&lt;br&gt;
DATA TRANFORMATION whic is the process of changing the data's format, structure, or values. It includes;&lt;br&gt;
 &lt;br&gt;
 💠 Adding, copying and replicating data&lt;br&gt;
 💠 Deleting fields or records&lt;br&gt;
 💠 Standardizing the names of variables&lt;br&gt;
 💠 Renaming, moving or combining colums in a database&lt;br&gt;
 💠 Joining one set of data with another &lt;br&gt;
 💠 Saving file in a different format&lt;br&gt;
 There are goals for data transformation. They are&lt;br&gt;
 &lt;br&gt;
 ➡️ Data orgaization&lt;br&gt;
 ➡️ Data migration&lt;br&gt;
 ➡️ Data enhancement&lt;br&gt;
 ➡️ Data comparison&lt;br&gt;
 DATA BIAS is a type of error that systematically skews results in a certain direction. There are 4 types of bias&lt;br&gt;
 -SAMPLE BIAS is when a sample isnt representative of the population as a &lt;br&gt;
 whole. &lt;br&gt;
 - OBSERVER BIAS (experiment bias/ research bias). The tendency for &lt;br&gt;
 different people to observe things differently.&lt;br&gt;
 - CONFIRMATION BIAS is the tendency to search for or interprete &lt;br&gt;
 information in a way that confirms pre-existing beliefs&lt;br&gt;
 - INTERPRETATION BIAS the tendency to always interprete &lt;br&gt;
 ambiguous situations in a positive or negative way.&lt;br&gt;
 In order for you to indentify good data sources, you need to follow these practices;&lt;br&gt;
 ⭕ Your data must be RELIABLE (good source)&lt;br&gt;
 ⭕ Your data must be ORIGINAL&lt;br&gt;
 ⭕ Your data must be COMPREHENSIVE&lt;br&gt;
 ⭕ Your data must be CURRENT (upto date)&lt;br&gt;
 ⭕ Your data must be CITED&lt;br&gt;
 If you dont follow this practices you will have bad data.&lt;br&gt;
 Just as humans have ethics, data also has ethics. DATA ETHICS is a well founded standard of right and wrong that dictates how data is collected, shared and used.&lt;br&gt;
 &lt;br&gt;
 There are various aspects of data ethics which include;&lt;br&gt;
 - Ownership: Individuals own the raw data. They provide and they have primary control over its usage, how its processed and how its shared.&lt;br&gt;
 - Transcation transparency: All data processing activities and algorithm should be completely explainable and understood by the individual who provides their data &lt;br&gt;
 - Consent: An individual's right to know explicit details about how and why their data will be used before agreeing to provide it.&lt;br&gt;
 - Currency: individuals should be aware of financial transactions resulting from the use of their personal data and the scale o these transactions.&lt;br&gt;
 - Privacy: Preserving a data subject's information and activities any time a data transaction occurs.&lt;br&gt;
 DATA ANONYMIZATION is the process of protecting people's privaye or sensitieve data by eliminating PERSONAL IDENTIFIABLE INFORMATION (PII) which is information that can be used by itself or with other data to track down a person's identity.&lt;br&gt;
OPEN DATA and its features.&lt;br&gt;
 OPEN DATA refers to free access, usage and sharing of data. Features of &lt;br&gt;
 include;&lt;br&gt;
 -Open data must be available as a whole. &lt;br&gt;
 -It must be provided under terms that allow reuse and distribution &lt;br&gt;
 including the ability to use it with other data sets.&lt;br&gt;
 DATA INTEROPERABILITY is the ability of data systems and services to openly connect and share data.&lt;br&gt;
DATABASE is a collection of data stored in a computer system. Data base has 4 features. These features include;&lt;br&gt;
 &lt;br&gt;
 ⭕ PRIMARY KEY is a unique identifier of each record within a table.&lt;br&gt;
 ⭕ FOREIGN KEY is a field within a table that is a primary key in &lt;br&gt;
 another table.&lt;br&gt;
 ⭕ RELATIONAL DATABASE is a database that contains a series of tables&lt;br&gt;
 that can be connected to form relationship. Basically they allow &lt;br&gt;
 analyst to organize and link data based on what the data has in &lt;br&gt;
 common.&lt;br&gt;
 ⭕ NORMALIZATION is a process of organizing data in relational&lt;br&gt;
 database. For example creating tables and establishing &lt;br&gt;
 relationships between these tables. It is applied to eliminate&lt;br&gt;
 data redundancy, increase intergrity and reduce complexity&lt;br&gt;
 in a database.&lt;br&gt;
I also learnt about META DATA. META DATA is data about data, it is used in database management to help data analyst interpret the contents of data within the database. &lt;br&gt;
 There are 3 types of meta data.&lt;br&gt;
 ➡️ DESCRIPTIVE META DATA describes a piece of data and can be used to &lt;br&gt;
 identify it at a later point in time.&lt;br&gt;
 ➡️ STRUCTURAL META DATA it indicates how a piece of data is organized&lt;br&gt;
 and whether it is part of one or more than one data collection&lt;br&gt;
 ➡️ ADMINSTRATIVE META DATA it indicates the technical sources of a&lt;br&gt;
 digital asset.&lt;br&gt;
META DATA provides information about other data and helps data analyst interprete the contents of the dat within a database it tells the Who, What, When, Where, Which, Why and How of the data. You can find meta data in PHOTOS, EMAILS, SPREADSHEETS AND ELECTRONIC CREATED DOCUMENTS.&lt;br&gt;
 Elements of meta data include;&lt;br&gt;
 &lt;br&gt;
💠 File or document type&lt;br&gt;
💠 Date, time and creator&lt;br&gt;
💠Title and description&lt;br&gt;
💠 Geolocation&lt;br&gt;
💠 Tags and categories&lt;br&gt;
💠 Who last modified it and when&lt;br&gt;
💠 Who can access or update it.&lt;br&gt;
 The benefits of meta data is that it makes data RELIABLE AND CONSISTENT &lt;br&gt;
META DATA REPOSITORIES helps data analyst ensure their data is reliable and consistent. They describe where the meta data came from and store that data in accessible form with a a common structure.&lt;br&gt;
BIG QUERY which is a data warehouse on the google cloud platform used to query and filter large datasets.&lt;br&gt;
 &lt;br&gt;
 Most importantly i learnt how to use a big query and run data from datasets using the conditions&lt;br&gt;
 ⭕ SELECT&lt;br&gt;
 ⭕ FROM&lt;br&gt;
 ⭕ WHERE&lt;br&gt;
 These are STRUCTURED QUERY LANGUAGE that helps you to easily get a specific data (e.g customer ID, or number of customers who bought sugar balls from the bakery) from a large range of data, hence making your analysis process easy. They also help you to retrieve and manipulate data from a database.&lt;br&gt;
ORGANIZING and SAVING DATA.&lt;br&gt;
 &lt;br&gt;
There are best practices organisations can follow when organizing data. They are seen below&lt;br&gt;
 ⭕ NAMING CONVENTIONS which are consistent guidelines that describe &lt;br&gt;
 the content, date, or version of a fle in its name. Means using logical and &lt;br&gt;
 descriptive names for your files to make them easier to find and use&lt;br&gt;
 ⭕ FOLDERING which is organizing files into folders which helps to easily&lt;br&gt;
 find your files. &lt;br&gt;
 ⭕ ARCHIVING CONVENTION can be used to move old projects to a separate &lt;br&gt;
 location to create, archive and cut down or clutter&lt;br&gt;
 ⭕ ALIGN YOUR NAMING and STORAGE practices with your team&lt;br&gt;
 ⭕ Develope META DATA practices like creating a file that outlines project&lt;br&gt;
 naming conventions for easy reference.&lt;br&gt;
 Storing data in different places (database or spreadsheet) can make the data &lt;br&gt;
 to contradict itself and also takes up alot of space. &lt;br&gt;
RELATIONAL DATABASE can you avoid data duplication and store data more effeciently. &lt;br&gt;
 File-naming conventions help you organize access, process and analyze data because they act as a quick refernce points to identify what's in the file.&lt;br&gt;
 One important practice is to decide on file naming conventions as a team or compnay early in a project. It's also critical to ensure that file names are meaningful, consistent, and easy to read. File names should include;&lt;br&gt;
 &lt;br&gt;
 ➡️ Project's name&lt;br&gt;
 ➡️File creation data &lt;br&gt;
 ➡️ Revision version&lt;br&gt;
 ➡️ consistent style and order&lt;br&gt;
 DATA SECURITY is protecting data from unauthorized access or corruption by adopting safety measures. &lt;br&gt;
 - Excel and Google sheets both have features that help you protect spreadsheets from being edited from the entire worksheet right down to single cells in a table.&lt;br&gt;
 - They both have access control features like password protection and user permissions.&lt;br&gt;
 &lt;br&gt;
 We can both keep our data safe and still have access to use them at anytime we want to use them for analysis. Security measures that can help companies do the 2 are;&lt;br&gt;
 &lt;br&gt;
 ‼️ ENCRYPTION. Uses a unique algorithm to alter data and make it unusable ny users and applications that dont know the algorithm. This algorithm is saved as 'KEY' which can be used to reserve the encryption so if you have the key you can still use data in its original form.&lt;br&gt;
 &lt;br&gt;
 ‼️ TOKENIZATION. Replaces the data elements you want to protect with randomly generated data refered to as TOKEN.The original data is stored in a separate location and mapped to the tokens. &lt;br&gt;
 &lt;br&gt;
VERSION CONTROL enales all collaboratores within a file to track changes over time. You can understand who made what changes to a file, when they were made and why.&lt;/p&gt;

</description>
      <category>learning</category>
      <category>database</category>
      <category>datascience</category>
      <category>community</category>
    </item>
    <item>
      <title>From raw data to insights: my journey through foundations of data analytics</title>
      <dc:creator>Awasume Marylin</dc:creator>
      <pubDate>Sun, 19 Jan 2025 06:51:29 +0000</pubDate>
      <link>https://dev.to/awasume_marylin_8/my-data-analytics-jornal-5aaa</link>
      <guid>https://dev.to/awasume_marylin_8/my-data-analytics-jornal-5aaa</guid>
      <description>&lt;p&gt;During my first week of studying data analytics, i got to learn “ What data analytics is”. And it is defined below.&lt;/p&gt;

&lt;p&gt;Data analytics is the collection transformation and organisation of facts to draw conclusion, make predictions and drive informed decision-making&lt;/p&gt;

&lt;p&gt;KEY WORDS&lt;/p&gt;

&lt;p&gt;Collection&lt;br&gt;
Transform&lt;br&gt;
Organisation&lt;br&gt;
Draw conclusion&lt;br&gt;
Make predictions&lt;br&gt;
Drive informed decision-making&lt;br&gt;
What makes you a strong data analyst is not just maths alone its;&lt;/p&gt;

&lt;p&gt;Asking the right questions&lt;br&gt;
Finding the best source to answer your question effectively&lt;br&gt;
illustrating your finding clearly in visualization&lt;br&gt;
I also got to learn the different types of business analytics&lt;/p&gt;

&lt;p&gt;4 key types of Business Analystics&lt;/p&gt;

&lt;p&gt;Descriptive analytics: The interpretation of historical data to identify trends and patterns&lt;br&gt;
Predictive analytics: Centers on taking that information and use it to forecast future outcomes&lt;br&gt;
Diagnostics analytics: Used to identify the root cause of a problem&lt;br&gt;
Prescriptive analytics: Testing an other techniques are employed to determine which outcomes will yield the best result in a given scenario&lt;br&gt;
I learnt that there are 6 phases a data analyst passes through to solve a problem and make DATA DRIVEN INFORMED DECISION MAKING. These phases include:&lt;/p&gt;

&lt;p&gt;THE 6 PHASES OF DATA ANALYTICS&lt;/p&gt;

&lt;p&gt;Ask phase: Ask questions and define the problem&lt;br&gt;
Prepare phase: Phase data by collecting and storing the information&lt;br&gt;
Process phase: Process data by cleaning and check the information&lt;br&gt;
Analyze phase: Analyze data by finding patterns relationships and trends&lt;br&gt;
Share phase: Share data with your audience&lt;br&gt;
Acts phase: Act on the data and use the analysis results&lt;br&gt;
You can also work with your gut instinct, but your gut instinct needs to be based on the data otherwise you will go off track. Data analysis is rooted in statistics.&lt;/p&gt;

&lt;p&gt;I saw something interesting and surprising while studying, which is DATA ECOSYSTEMS, and they also have their elements.&lt;/p&gt;

&lt;p&gt;Data ecosystems are the various elements that interact with one another inorder to produce, manage, store, organize, analyze, and share data e.g hardware and software tools and the people that use them.&lt;/p&gt;

&lt;p&gt;data can also be found on cloud.&lt;/p&gt;

&lt;p&gt;Cloud is a place where data is being kept online rather than a computer hardrive.&lt;/p&gt;

&lt;p&gt;While studying as a data analyst people or you might have these COMMON MISCONCEPTIONS.&lt;/p&gt;

&lt;p&gt;Data science is thesame with data analytics&lt;/p&gt;

&lt;p&gt;Data science creates new questions using data. in otherwords its creating new ways of modeling and understanding the unknown by using raw data while Analyst find answers to existing questions by creating insights from data sources.&lt;/p&gt;

&lt;p&gt;Data analysis and Data analytics sound thesame but they are very differnt.&lt;/p&gt;

&lt;p&gt;Data analysis is the collection, processing and analyzind of data to make informed decion making while Data analytics is the science of data&lt;/p&gt;

&lt;p&gt;One of the powerful ways you can put data to work is with DATA DRIVEN DECISION-MAKING using facts to guide business strategy.&lt;/p&gt;

&lt;p&gt;First step in data drivrn decision-making is FIGURING OUT THE BUSINESS NEEDS TO BE SOLVED.&lt;/p&gt;

&lt;p&gt;AS AN ANALYST YOU ALSO NEED TO HAVE SKILLS ( ANALYTICAL SKILLS) WHICH IS WHAT ANYONE WHO WANTS TO BE A GOOD OR EXCELLENT ANALYST SHOULD POSSESS.&lt;/p&gt;

&lt;p&gt;ANALYTICAL SKILLS are qualities and characteristics associated with problem sloving using facts.&lt;/p&gt;

&lt;p&gt;5 essential points to analytical skills&lt;/p&gt;

&lt;p&gt;Curiosity&lt;br&gt;
Understanding context (the condition to which something exist or happens)&lt;br&gt;
Having technical mindset (the ability to break things down into smaller steps or pieces and work with them in an orderly and logical way)&lt;br&gt;
Data design ( how you organize information)&lt;br&gt;
Data strategy ( the management of the people, processes and tools used in data analysis)&lt;br&gt;
Analytical thinking involves identifying and defining a problem and solving it by using data in an organized step by step manner. 5 steps of analytical thinking includes&lt;/p&gt;

&lt;p&gt;Visualization&lt;br&gt;
Strategy ( helps data analyst see what they want to achieve with the data and how they can get there. it also helps improve the quality and usefulness of the data we collect)&lt;br&gt;
Problem orientation ( it is all about keeping the problem at the top of your mind through out the project)&lt;br&gt;
Correlation ( relationship. it doesnt equal causation)&lt;br&gt;
Big picture and detail oriented thinking ( is like looking at the whole puzzle instead of seeing only little pieces it helps you zoom out and see possibilities and opportunities)&lt;br&gt;
There are also the 5 WHYs which can also help you find the solution or root cause to a problem.&lt;/p&gt;

&lt;p&gt;Gap analysis is a method for examining and evaluating how a process works currently in order to get where you want to be in the future.&lt;/p&gt;

&lt;p&gt;Data driven decision-making using facts to guide business strategy&lt;br&gt;
Data set a collection of data that can be manipulated or analyzed as can be manipulated or analyzed as one unit&lt;br&gt;
Root cause the reason why a problem occurs&lt;br&gt;
STAGES OF THE DATA LIFE CYCLE&lt;/p&gt;

&lt;p&gt;Plan stage: Decides what kind of data is needed, how it will be managed and who will be responsible for it.&lt;br&gt;
Capture: Collect or bring in data from a variety of different sources.&lt;br&gt;
Manage: Care for and maintain the data. This includes determining how and where it is stored and the tools used to do so.&lt;br&gt;
Analyze: Use the data to solve problems make decisions and support business goals.&lt;br&gt;
Archive: Keep relevant data stored for longterm and future reference.&lt;br&gt;
Destroy: Remove data from storage and delete any stored copies of the data.&lt;br&gt;
In order to be able to collect,organize, manage and process data, you need tools . These tools are called DATA ANALYTICAL TOOLS&lt;/p&gt;

&lt;p&gt;These tools include;&lt;/p&gt;

&lt;p&gt;Spreadsheet: It is a digital work sheet that stores, organizes and sorts data. e.g microsoft excel and google sheets. It allows you to identify patterns and piece the data together in a way that works for each specific data project. They can also creat e excellent data visualization like graphs and charts.&lt;br&gt;
Query: Is a computer programming language that allows you to retrieve and manipulate data from a database. They allow analyst to select, create, add or download data from a database for analysis e.g Microsoft Query Language.&lt;br&gt;
Data Visualization: They turn complex numbers into a story that people can understand. They also help stakeholders come up with conclusions that lead to informed decision effective business strategies e.g maps, graphs, tables.&lt;br&gt;
There is a relationship between Data Analysis process and Data life cycle.&lt;/p&gt;

&lt;p&gt;While data analysis process will help you drive your projects and help you help you reach your business gaols, you must understand the life cycle of your data inorder to use that process.&lt;/p&gt;

&lt;p&gt;To analyze your data you will need to have a thorough understanding of it. Similarly you can collcet all the data you want, but the data is only useful to you if you have a plan for analyzing it.&lt;/p&gt;

&lt;p&gt;It very important as a data analyst to practice FAIRNESS.&lt;/p&gt;

&lt;p&gt;FAIRNESS is ensuring that your analysis doesnt create or reinforce bias (create systems that include everyone).&lt;/p&gt;

&lt;p&gt;The best practices to ensure fairness you are working are&lt;/p&gt;

&lt;p&gt;CONSIDER ALL THE AVAILABLE DATA. We have to decide what data is useful. Often there will be data that isnt relevant to what you’re focusing on or doesnt seem to align with your expetations. But you cant just ignore it. It is critical to consider all the available data so that your analysis reflect the truth and not just your own expectations.&lt;br&gt;
IDENTIFYING SURROUNDING FACTORS. Context is key for you and your stakeholders to understand the final conclusions of any analysis. You must also consider all of the data to get more insights.&lt;br&gt;
INCLUDE SELF REPPORTING DATA. Its a data collection technique where participants provide information about themselves. It is a great wya to introduce fairness into your collection process. People bring conscious and unconscious bias to their observation about the world , including about other people. This method can help you to avoid these observer bias.&lt;br&gt;
USE OVERSAMPLING EFFECTIVELY. It is the process of increasing the sample size of nondominant groups in a population. This can help you better represent them and address imbalance datasets.&lt;br&gt;
Think about fairness from begining to end.&lt;br&gt;
Some questions are more effective than others. A yes or no question is a closed ended question and cannot lead to valuable insights.&lt;/p&gt;

&lt;p&gt;Effective questions follow the SMART methodology&lt;/p&gt;

&lt;p&gt;-Specific: if the question is too general try to narrow it down by focusing on just one element. But it should not be a closed ended question.&lt;/p&gt;

&lt;p&gt;-Measurable: let our questions be measurable. like using figures, quantities etc&lt;/p&gt;

&lt;p&gt;-Action-oriented: Let our questions encourage change&lt;/p&gt;

&lt;p&gt;-Relevant: When you ask relevant questions it can help you with the problem ou are trying to solve.&lt;/p&gt;

&lt;p&gt;-Time bound: it should specify the time to be studied which limits the range of possible data analyst to focus on relevant data.&lt;/p&gt;

&lt;p&gt;We need to ask fair questions ( practice fairnes) meaning there shouldnt be any bias in our questions e.g leading your questions towards a certain way and making assumptions.&lt;/p&gt;

&lt;p&gt;Some question you might ask when given a project&lt;/p&gt;

&lt;p&gt;⭕ OBJECTIVE: what are the goals of the question? What , if, any, questions are expected to be answered?.&lt;/p&gt;

&lt;p&gt;⭕ AUDIENCE: Who are the stake holders? Who is interested or cooncerned about the results of the deepdive? Who is the audience for the deep dive?&lt;/p&gt;

&lt;p&gt;⭕ TIME: What is the time frame for completion? By what date does this need to be done?&lt;/p&gt;

&lt;p&gt;⭕ RESOURCES: What resources are available to accomplish the deep dive?&lt;/p&gt;

&lt;p&gt;⭕ SECURITY: Who should have access to the information.&lt;/p&gt;

&lt;p&gt;I learnt about HOW DATA EMPOWERS DECISIONS. The two types of data decision making include;&lt;/p&gt;

&lt;p&gt;DATA-INSPIRED DECISION MAKING which explores different data source to find out what they have in common&lt;br&gt;
DATA-DRIVEN DECISION MAKING which uses facts to guide business strategies.&lt;/p&gt;

&lt;p&gt;Also i learnt about the two types of data which might help us answer alot of different questions. They include;&lt;/p&gt;

&lt;p&gt;1) Quantitative data which deals with measurements of numerical facts and ask questions like WHAT?, HOW MANY?,HOW OFTEN?.&lt;br&gt;
2)Qualitative data which deals with explanatory measures and ask questions like WHY?&lt;/p&gt;

&lt;p&gt;Also there are 2 types of data presentation tools which are:&lt;br&gt;
-Reports: which are a static collection of data given to stakeholders periodically&lt;br&gt;
-Dashboards: which monitors life incoming data.&lt;/p&gt;

&lt;p&gt;3 COMMON TYPES OF DASHBOARDS&lt;br&gt;
i) Strategic dashboards focuses longterm goals and strategy at the&lt;br&gt;
highest level of maintainance.&lt;/p&gt;

&lt;p&gt;ii) Operational dashboard which focuses on short term perfomance&lt;br&gt;
tracking and intermediate goals.&lt;/p&gt;

&lt;p&gt;iii) Analytical dashboard which consist of data set and the mathematics&lt;br&gt;
used in this set.&lt;/p&gt;

&lt;p&gt;I learnt that data can be Big or Small&lt;/p&gt;

&lt;p&gt;-SMALL DATA is a set of small specific data points typically involving a short period of time which are useful in making day-to-day decision making.&lt;/p&gt;

&lt;p&gt;-BIG DATA is a large complex datasets typically involves long period o times which enables data analyst to address far-reaching business patterns.&lt;/p&gt;

&lt;p&gt;As someone who studied ICT in school i definitely know about spreadsheets but have seen mostly MICROSOFT EXCEL.&lt;/p&gt;

&lt;p&gt;Recently before begining my data analytics journey i studied more on microsoft excel, and all its formulas and functions and it was a beautiful experience for me having to come accross new formulas and functions i have never seen before and having to explore and do so many things with microsoft excel,&lt;/p&gt;

&lt;p&gt;Today i got to learn about how spreadsheets (microsoft excel and google sheets) can be used to analyse data and how it is related to the data life cycle which include:&lt;br&gt;
➡️ Plan&lt;br&gt;
➡️ Capture&lt;br&gt;
➡️ Manage&lt;br&gt;
➡️ Analyze&lt;br&gt;
➡️ Archive&lt;br&gt;
➡️ Destroy&lt;/p&gt;

&lt;p&gt;— — PLAN means formatting your cells, the headings you choose to highlight, the color scheme,and the way you order your data points.&lt;/p&gt;

&lt;p&gt;— -CAPTURE data by the source by connecting spreadsheets to other data sources such as an online survey application or database&lt;/p&gt;

&lt;p&gt;— -MANAGE different types of data with a spreadsheet. This can involve storing, organizing, filtering and updating information. They can also let you kno who can access the data.&lt;/p&gt;

&lt;p&gt;— -ANALYZE data in a spreadsheet to help make better decisions. Some of the most common spreadsheet tools include Formulars and pivot tables.&lt;/p&gt;

&lt;p&gt;— -ARCHIVE any spreadsheet that you dont use often but might need to reference later with built-in tools&lt;/p&gt;

&lt;p&gt;— -DESTROY your spreadsheet when you know you are certain that you will never need again. If you do have use for it, better have back up copies or for legal security.&lt;/p&gt;

&lt;p&gt;⭕ OPERATOR is a symbol that name the type of operation or calculations to be made.&lt;br&gt;
⭕ CELL REFERENCE is a cell or range of cells in a worksheet that can be used in a formular&lt;br&gt;
⭕ RANGE is a collection of 2 or more cells.&lt;/p&gt;

&lt;p&gt;I learnt that even as an experienced data analyst there are some common errors that can be made when doing operations. These errors include&lt;/p&gt;

&lt;p&gt;1) DIV ERROR&lt;br&gt;
2)DN ERROR&lt;br&gt;
3) N/A&lt;br&gt;
4)REF ERROR&lt;/p&gt;

&lt;p&gt;I used excel formulars to calculate data. I used basic formulars like SUM, AVERAGE, MIN, MAX, DIVIDE,MULTIPLY etc amongst the many formulars that exist.&lt;/p&gt;

&lt;p&gt;I also learnt about thinking which is the process of recognizing the current problem or situation, organizing available information, revealing gaps and opportunities and identifying the options.&lt;/p&gt;

&lt;p&gt;One way you can practice structured thinking and avoiding mistakes is&lt;br&gt;
by using a SCOPE O WORK.&lt;/p&gt;

&lt;p&gt;⭕ SCOPE OF WORK (SOW) is an agreed upon outline of the work you are going to perform on a project e.g work details, reports, schedules that the client can expext. Under this we have:&lt;/p&gt;

&lt;p&gt;➡️ DELIVERABLES which focuses on What work is being done and what things are being created as a result of this project?&lt;br&gt;
➡️ MILESTONES which is closely related to your timeline. what are the major milestones in your project? How do you know when a given part of your project is complete&lt;br&gt;
➡️ TIMELINE which is closely tied to the milestone you created for your project.The timeline is a way of of mapping expectations for how long each step of the process should take&lt;br&gt;
➡️ REPORTS you have to give status update to your stakeholders. will it be weekly or monthly ?.&lt;/p&gt;

&lt;p&gt;The different stakeholders we might encounter while working on a project. They inlcude:&lt;/p&gt;

&lt;p&gt;➡️ The executive team. Who provide strategic and operational&lt;br&gt;
leadership to the company. They are made of VICE PRESIDENTS, CHIEF&lt;br&gt;
MARKETING OFFICER, AND SENIOR LEVELS PROFESSIONALS.&lt;br&gt;
➡️ The customer-facing team. Anyone in an organization who has some&lt;br&gt;
levels of interaction with customers and potential customers.&lt;br&gt;
➡️Data science team. Organizing data within a company takes teamwork.&lt;br&gt;
There is a good chance you will find yourself working with other data&lt;br&gt;
analyst, data scientist and data engineers.&lt;/p&gt;

&lt;p&gt;I have learnt that working effectively with stakeholders you’ll often have to go beyond the data and to do that we will need the following tips to communicate clearly, establish trust an deliver your findings accros groups&lt;br&gt;
‼️ Discuss goals&lt;br&gt;
‼️Feel empowered to say no&lt;br&gt;
‼️ Plan for the unexpected&lt;br&gt;
‼️Know your project&lt;br&gt;
‼️Start with words and visuals&lt;br&gt;
‼️ Communicate often.&lt;/p&gt;

&lt;p&gt;Clear communcation is key. Before you communicate think about&lt;br&gt;
— WHO YOUR TARGET AUDIENCE IS&lt;br&gt;
-WHAT THEY ALREADY KNOW&lt;br&gt;
-WHAT THEY NEED TO KNOW&lt;br&gt;
— HOW OFTEN CAN YOU COMMUNICATE THAT EFFECTIVELY TO THEM&lt;/p&gt;

&lt;p&gt;In todays studies i learnt that good writing, listening and speaking skills can also help you effective communication as a data analyst. This is because speed can be the enemy of accuracy .&lt;/p&gt;

&lt;p&gt;This is why communication is the most valuable tools for work with teams. So it is important to start with structured thiking and a well planned SCOPE OF WORK(SOW).&lt;/p&gt;

&lt;p&gt;I also learnt that data can have limitations. As a data analyst it is important to know the limits of dat aso you can prepare for it. Limitations of data are as follows,&lt;/p&gt;

&lt;p&gt;⭕ Incomplete or nonexistent data.&lt;br&gt;
⭕Dont miss misaligned data&lt;br&gt;
⭕ Deal with dirty ( data cleaning)&lt;br&gt;
⭕ Tell a clear data;&lt;/p&gt;

&lt;p&gt;💠 compare thesame types of data&lt;br&gt;
💠 visualize wit data&lt;br&gt;
💠 leave out needles graphs&lt;br&gt;
💠 test for statistical significance&lt;br&gt;
💠 pay attention to sample size&lt;/p&gt;

&lt;p&gt;⭕ Be the judge&lt;/p&gt;

</description>
      <category>learning</category>
      <category>database</category>
      <category>datascience</category>
      <category>community</category>
    </item>
  </channel>
</rss>
