<?xml version="1.0" encoding="UTF-8"?>
<rss version="2.0" xmlns:atom="http://www.w3.org/2005/Atom" xmlns:dc="http://purl.org/dc/elements/1.1/">
  <channel>
    <title>DEV Community: Evans Jones</title>
    <description>The latest articles on DEV Community by Evans Jones (@evans_jones_9658a6f26f2ad).</description>
    <link>https://dev.to/evans_jones_9658a6f26f2ad</link>
    <image>
      <url>https://media2.dev.to/dynamic/image/width=90,height=90,fit=cover,gravity=auto,format=auto/https:%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Fuser%2Fprofile_image%2F1872037%2Fa59a4061-ca66-4519-bc1d-c2bec41a4314.jpg</url>
      <title>DEV Community: Evans Jones</title>
      <link>https://dev.to/evans_jones_9658a6f26f2ad</link>
    </image>
    <atom:link rel="self" type="application/rss+xml" href="https://dev.to/feed/evans_jones_9658a6f26f2ad"/>
    <language>en</language>
    <item>
      <title>Data Analytics $ Data Science Projects</title>
      <dc:creator>Evans Jones</dc:creator>
      <pubDate>Thu, 08 May 2025 10:11:30 +0000</pubDate>
      <link>https://dev.to/evans_jones_9658a6f26f2ad/data-analytics-data-science-projects-20oe</link>
      <guid>https://dev.to/evans_jones_9658a6f26f2ad/data-analytics-data-science-projects-20oe</guid>
      <description>&lt;p&gt;Project 1: Real-time data pipeline for stock market analysis.&lt;br&gt;
Objective: Build a pipeline to ingest, process, and analyze stock market data in real-time.&lt;br&gt;
Tools, Frameworks $ and Technologies: stream processing, ETL(extract, transform, load)data warehousing, Apache Kafka, stock data from &lt;a href="https://www.alphavantage.co/" rel="noopener noreferrer"&gt;https://www.alphavantage.co/&lt;/a&gt; or any platform using Apache Kafka to a cluster on the Confluent Cloud and consume it into the DB of your choice.&lt;br&gt;
(Feel free to use a different tool, approach, or technology.)&lt;/p&gt;

&lt;p&gt;Project 2: Scrape data from Amazon, Jumia, or any other e-commerce website to create a list of all products currently offered.&lt;br&gt;
Tools, Frameworks, and Technologies: Python, Beautiful Soup, Selenium, Scrapy, Pandas, Numpy.&lt;br&gt;
From the project, you can perform EDA on the data and even build a UI page where you can list items using Flask, fast API,streamlit, dash, or any other tool of your choice.&lt;/p&gt;

&lt;p&gt;Project 3: Automating Data scrapers and analytical processes using Apache Airflow.&lt;br&gt;
Tools, Frameworks &amp;amp; Technologies: Apache Airflow, Python, Pandas, Numpy.&lt;br&gt;
Scrap house lifting data from buyrentkenya.com or any other website of your choice and automate your scripts and analytical process using Apache Airflow or any other workflow or any other workflow orchestration tool. This project focuses on workflow automation and scheduling.&lt;/p&gt;

&lt;p&gt;Project 4: Kenya YouTube channels analysis using Python Apache Api&lt;br&gt;
Tools, Frameworks &amp;amp; Technologies: Python, YouTube API, requests, pandas,matplotlib, seaborn.&lt;br&gt;
Objective: Analyze YouTube channels in Kenya using Python, i.e, content analysis, subscriber trends, engagement metrics.&lt;/p&gt;

&lt;p&gt;Project 5:Build a kenya and east africa agricultural data portal to provide necessary information to advice farmers and investors intrested in farming and agribiz in kenya.&lt;br&gt;
Building a comprehensive agricultural data portal for Kenya and East Africa to support farmers and investors interested in farming and Agribiz.&lt;br&gt;
Hera is the approach:&lt;br&gt;
i)Platform structure and features:&lt;br&gt;
Homepage.&lt;br&gt;
Overview of the portal's purpose.&lt;br&gt;
Latest news and updates in agriculture.&lt;br&gt;
Quick access to key sections.&lt;br&gt;
Sections;&lt;br&gt;
a)Crop Information&lt;br&gt;
Detailed profiles of crops grown in Kenya and East Africa.&lt;br&gt;
Best practices for cultivation, pest management&amp;amp; harvesting&lt;br&gt;
Seasonal calendars and climate considerations.&lt;br&gt;
Yield expectations and market trends.&lt;br&gt;
b)Livestock Management&lt;br&gt;
Breeding and management practices for various livestock species.&lt;br&gt;
Disease prevention and treatment.&lt;br&gt;
Feed and nutrition guidelines.&lt;br&gt;
Market demand &amp;amp; pricing trends.&lt;br&gt;
c) Market Information&lt;br&gt;
Prices of agricultural commodities in major markets.&lt;br&gt;
Market forecasts &amp;amp; trends&lt;br&gt;
Import/Export regulations and opportunities.&lt;br&gt;
Market analysis reports.&lt;br&gt;
d) Agribusiness Opportunities&lt;br&gt;
Investment opportunities in agriculture &amp;amp; agribusiness.&lt;br&gt;
Success stories and case studies.&lt;br&gt;
Government Incentives &amp;amp; Support Programs.&lt;br&gt;
Legal and regulatory information.&lt;br&gt;
e) Weather and Climate data&lt;br&gt;
Historical &amp;amp; real-time weather data.&lt;br&gt;
Seasonal forecasts and climate change impacts.&lt;br&gt;
Advisories on weather-sensitive farming activities.&lt;br&gt;
f)Research and Innovation&lt;br&gt;
Research findings &amp;amp; innovations in agriculture.&lt;br&gt;
Emerging technologies &amp;amp; their applications.&lt;br&gt;
Collaboration opportunities with research institutions.&lt;/p&gt;

&lt;p&gt;ii)User Interface(UI)&amp;amp;User Experience(UX)&lt;br&gt;
Intuitive navigation- Clear categories &amp;amp; subcategories.&lt;br&gt;
Search Functionality- Keyword search across all sections.&lt;br&gt;
Interactive Maps, Geographic data representation.&lt;br&gt;
Mobile Compatibility, Responsive design for access on smartphones.&lt;/p&gt;

&lt;p&gt;iii) Data Collection &amp;amp; Integration&lt;br&gt;
Government Agencies collaborate with ministries and agencies for official data.&lt;br&gt;
Research institutions, Partner with universities and research organizations.&lt;br&gt;
Private Sector, Aggregate market data from traders &amp;amp; industry experts.&lt;br&gt;
Weather Services: Integrate weather data from meteorological departments.&lt;/p&gt;

&lt;p&gt;iv)Data Presentation &amp;amp; Visualization&lt;br&gt;
Graphs, Charts, Visual representations of market trends &amp;amp; climate data infographics summarize complex data into easily understandable data formats.&lt;br&gt;
Interactive tools, calculators for crop yield estimation, and financial planning.&lt;/p&gt;

&lt;p&gt;v)Community Support&lt;br&gt;
Discussion Forums, a platform for farmers &amp;amp; investors to exchange ideas.&lt;br&gt;
Expert Advice, Q&amp;amp;A sessions with agricultural experts, training, workshops, online courses, webinars on agricultural topics.&lt;/p&gt;

&lt;p&gt;vi) Security &amp;amp; Privacy&lt;br&gt;
Data Encryption, Secure transmission &amp;amp; storage of user data.&lt;br&gt;
Access Control: Different levels of access for farmers, investors, and administrators.&lt;br&gt;
Compliance with data protection regulations.&lt;/p&gt;

&lt;p&gt;vii)Marketing and Outreach&lt;br&gt;
Awareness Campaigns: Promote the portal through social media, partnerships, and events.&lt;br&gt;
User Feedback, Regular surveys, and feedback mechanisms for continuous improvement.&lt;/p&gt;

&lt;p&gt;viii)Sustainability &amp;amp; Scalability&lt;br&gt;
Scalable Infrastructure, Cloud-based architecture to handle increasing traffic.&lt;br&gt;
Continuous Updates, Regular updates with new data and features.&lt;br&gt;
Long-term planning, Funding, and sustainability strategies.&lt;/p&gt;

&lt;p&gt;ix) Partnerships &amp;amp; Collaboration&lt;br&gt;
Public-Private Partnerships, Collaborate with private companies for sponsorship and expertise.&lt;br&gt;
International Collaboration, Exchange data and best practices with similar platforms globally.&lt;/p&gt;

&lt;p&gt;x)Monitoring &amp;amp; Evaluation&lt;br&gt;
Metrics Track usage Statistics, User engagement, and feedback.&lt;br&gt;
Impact Assessment: Measure the portal's contribution to improving agricultural practices and investments.&lt;/p&gt;

&lt;p&gt;Tools, Frameworks &amp;amp; Technologies: Python, Django, Flask, pandas,numpy, data visualization libraries.&lt;br&gt;
Objective: Develop a portal for agricultural data in Kenya and East Africa: it involves collecting, organizing &amp;amp; presenting agricultural data for analysis.&lt;/p&gt;

&lt;p&gt;Project 6:Nairobi Metropolitant house price prediction with Python.&lt;br&gt;
Build a machine learning project to predict the house prices for different houses, plots, and land in Nairobi.&lt;br&gt;
Tools, Framework &amp;amp; Technologies: Python,openAIs, machine learning libraries(scikit-learn, TensorFlow,pytorch, pandas,numpy,matplotlib, flask, fast API, or streamLIT.&lt;br&gt;
Objective: to predict house prices in the Nairobi metropolitan area. It involves machine learning and data analysis to create a predictive model. you can use data scraped from project 3.&lt;/p&gt;

&lt;p&gt;Project 7: Fitness Data Analysis: case study.&lt;br&gt;
study the data science problem below and solve it.&lt;br&gt;
&lt;a href="https://stats.io/fitness-data-analysis-case-study/#google_vignette" rel="noopener noreferrer"&gt;https://stats.io/fitness-data-analysis-case-study/#google_vignette&lt;/a&gt;&lt;br&gt;
https the clever programmer.com/2023/09/04/fitness watch-data-analysis-using-python/.&lt;/p&gt;

&lt;p&gt;Project 8: Crop yield analysis in Kenya with Python &lt;br&gt;
Objectives:&lt;br&gt;
identify factors influencing crop yields across various Kenyan regions.&lt;br&gt;
Analyze historical data to uncover trends in crop production.&lt;br&gt;
Utilize basic statistical methods to explore the correlation between crop yields and factors like rainfall patterns, fertilizer application, and soil characteristics.&lt;/p&gt;

&lt;p&gt;Tools, Frameworks &amp;amp; Technologies:&lt;br&gt;
Data analysis software e,g Python-pandas,statsmodels.&lt;br&gt;
Data visualization tools: Matplotlib, seaborn.&lt;br&gt;
Public agricultural data, i.e, KARLO, Ministry of Agriculture, livestock, fisheries, and cooperatives.&lt;/p&gt;

&lt;p&gt;Research paper case 1:&lt;br&gt;
Investigate how GPS(global positioning system) tracking systems and big data streaming can aid ambulance services.&lt;br&gt;
Objectives: Research and propose a big data project that helps locate the nearest ambulance, estimate arrival and turnaround times to determine the shortest route, send real-time data to hospitals, and provide timely assistance needed.&lt;/p&gt;

&lt;p&gt;Research paper case 2:&lt;br&gt;
Explore and analyze the implementation of big data in traffic control systems with a focus on enhancing efficiency, reducing congestion, and improving overall traffic management.&lt;br&gt;
Objectives: Investigate the potential benefits, challenges, and innovative solutions associated with integrating big data technologies into traffic control mechanisms.&lt;/p&gt;

&lt;p&gt;Research paper case 3:&lt;br&gt;
The objective to investigate and propose strategies for controlling scamming and theft through the effective utilization of big data analytics.&lt;/p&gt;

&lt;p&gt;The focus should be on exploring how advanced data analytics, machine learning algorithms, and predictive modelling can be leveraged to detect, prevent, and mitigate fraudulent activities in various domains.&lt;/p&gt;

&lt;p&gt;you should aim at providing insights into the potential applications of big data in enhancing security measures and minimizing the impact of scams and theft through proactive and data-driven approaches.&lt;/p&gt;

&lt;p&gt;Research Paper Case 4:&lt;br&gt;
Apache Kafka and Apache Spark for event and real-time data streaming.&lt;br&gt;
Abstract: This research paper explores the investigation of Apache Spark for efficient event and real-time data streaming.&lt;br&gt;
Discuss the architecture, key components, and advantages of using Kafka and Spark together in streaming data applications.&lt;br&gt;
Provide use cases and examples of successful implementations.&lt;br&gt;
Explore challenges, best practices, and potential future development in the domain.&lt;br&gt;
Sections&lt;br&gt;
Introduction: Brief overview of the importance of real-time data streaming and the need for technologies like Kafka &amp;amp; spark.&lt;br&gt;
Apache Kafka overview: Explanation of Kafka architecture, components, and its role in event streaming.&lt;br&gt;
Apache Spark Overview, Overview of Spark capabilities, especially in the context of real-time data processing.&lt;br&gt;
Integration of Kafka and Spark,Detailed discussion on how Kafka and Spark can be integrated, including connectors, APIs, and dataflow.&lt;/p&gt;

&lt;p&gt;Use cases: Explore real-world use cases where the combination of Kafka and Spark has proven beneficial.&lt;br&gt;
Challenges and Solutions: Discuss challenges faced in implementing this combination and propose solutions or best practices.&lt;br&gt;
Future Developments, Predictions, and insights into the potential future developments and enhancements in Kafka and Spark for real data streaming.&lt;br&gt;
Conclusion- Summarize key findings and significance of using Apache Kafka and Spark in tandem for event and real-time data streaming.&lt;/p&gt;

</description>
    </item>
    <item>
      <title>DATA ENGINEERING ROADMAP</title>
      <dc:creator>Evans Jones</dc:creator>
      <pubDate>Wed, 22 Jan 2025 19:41:29 +0000</pubDate>
      <link>https://dev.to/evans_jones_9658a6f26f2ad/data-engineering-roadmap-o21</link>
      <guid>https://dev.to/evans_jones_9658a6f26f2ad/data-engineering-roadmap-o21</guid>
      <description>&lt;p&gt;This comprehensive course spans 4 months (16 weeks) and equips learners with expertise in Python, SQL, Azure, AWS, Apache Airflow, Kafka, Spark, and more.&lt;/p&gt;

&lt;p&gt;Learning Days: Monday to Thursday (theory and practice).&lt;br&gt;
Friday: Job shadowing or peer projects.&lt;br&gt;
Saturday: Hands-on lab sessions and project-based learning.&lt;br&gt;
Month 1: Foundations of Data Engineering&lt;br&gt;
Week 1: Onboarding and Environment Setup&lt;br&gt;
Monday:&lt;br&gt;
Onboarding, course overview, career pathways, tools introduction.&lt;br&gt;
Tuesday:&lt;br&gt;
Introduction to cloud computing (Azure and AWS).&lt;br&gt;
Wednesday:&lt;br&gt;
Data governance, security, compliance, and access control.&lt;br&gt;
Thursday:&lt;br&gt;
Introduction to SQL for data engineering and PostgreSQL setup.&lt;br&gt;
Friday:&lt;br&gt;
Peer Project: Environment setup challenges.&lt;br&gt;
Saturday (Lab):&lt;br&gt;
Mini Project: Build a basic pipeline with PostgreSQL and Azure Blob Storage.&lt;br&gt;
Week 2: SQL Essentials for Data Engineering&lt;br&gt;
Monday:&lt;br&gt;
Core SQL concepts (SELECT, WHERE, JOIN, GROUP BY).&lt;br&gt;
Tuesday:&lt;br&gt;
Advanced SQL techniques: recursive queries, window functions, and CTEs.&lt;br&gt;
Wednesday:&lt;br&gt;
Query optimization and execution plans.&lt;br&gt;
Thursday:&lt;br&gt;
Data modeling: normalization, denormalization, and star schemas.&lt;br&gt;
Friday:&lt;br&gt;
Job Shadowing: Observe senior engineers writing and optimizing SQL queries.&lt;br&gt;
Saturday (Lab):&lt;br&gt;
Mini Project: Create a star schema and analyze data using SQL.&lt;br&gt;
Week 3: Introduction to Data Pipelines&lt;br&gt;
Monday:&lt;br&gt;
Theory: Introduction to ETL/ELT workflows.&lt;br&gt;
Tuesday:&lt;br&gt;
Lab: Create a simple Python-based ETL pipeline for CSV data.&lt;br&gt;
Wednesday:&lt;br&gt;
Theory: Extract, transform, load (ETL) concepts and best practices.&lt;br&gt;
Thursday:&lt;br&gt;
Lab: Build a Python ETL pipeline for batch data processing.&lt;br&gt;
Friday:&lt;br&gt;
Peer Project: Collaborate to design a basic ETL workflow.&lt;br&gt;
Saturday (Lab):&lt;br&gt;
Mini Project: Develop a simple ETL pipeline to process sales data.&lt;br&gt;
Week 4: Introduction to Apache Airflow&lt;br&gt;
Monday:&lt;br&gt;
Theory: Introduction to Apache Airflow, DAGs, and scheduling.&lt;br&gt;
Tuesday:&lt;br&gt;
Lab: Set up Apache Airflow and create a basic DAG.&lt;br&gt;
Wednesday:&lt;br&gt;
Theory: DAG best practices and scheduling in Airflow.&lt;br&gt;
Thursday:&lt;br&gt;
Lab: Integrate Airflow with PostgreSQL and Azure Blob Storage.&lt;br&gt;
Friday:&lt;br&gt;
Job Shadowing: Observe real-world Airflow pipelines.&lt;br&gt;
Saturday (Lab):&lt;br&gt;
Mini Project: Automate an ETL pipeline with Airflow for batch data processing.&lt;br&gt;
Month 2: Intermediate Tools and Concepts&lt;br&gt;
Week 5: Data Warehousing and Data Lakes&lt;br&gt;
Monday:&lt;br&gt;
Theory: Introduction to data warehousing (OLAP vs. OLTP, partitioning, clustering).&lt;br&gt;
Tuesday:&lt;br&gt;
Lab: Work with Amazon Redshift and Snowflake for data warehousing.&lt;br&gt;
Wednesday:&lt;br&gt;
Theory: Data lakes and Lakehouse architecture.&lt;br&gt;
Thursday:&lt;br&gt;
Lab: Set up Delta Lake for raw and curated data.&lt;br&gt;
Friday:&lt;br&gt;
Peer Project: Implement a data warehouse model and data lake for sales data.&lt;br&gt;
Saturday (Lab):&lt;br&gt;
Mini Project: Design and implement a basic Lakehouse architecture.&lt;br&gt;
Week 6: Data Governance and Security&lt;br&gt;
Monday:&lt;br&gt;
Theory: Data governance frameworks and data security principles.&lt;br&gt;
Tuesday:&lt;br&gt;
Lab: Use AWS Lake Formation for access control and security enforcement.&lt;br&gt;
Wednesday:&lt;br&gt;
Theory: Managing sensitive data and compliance (GDPR, HIPAA).&lt;br&gt;
Thursday:&lt;br&gt;
Lab: Implement security policies in S3 and Azure Blob Storage.&lt;br&gt;
Friday:&lt;br&gt;
Job Shadowing: Observe senior engineers applying governance policies.&lt;br&gt;
Saturday (Lab):&lt;br&gt;
Mini Project: Secure data in the cloud using AWS and Azure.&lt;br&gt;
Week 7: Real-Time Data Processing with Kafka&lt;br&gt;
Monday:&lt;br&gt;
Theory: Introduction to Apache Kafka for real-time data streaming.&lt;br&gt;
Tuesday:&lt;br&gt;
Lab: Set up a Kafka producer and consumer.&lt;br&gt;
Wednesday:&lt;br&gt;
Theory: Kafka topics, partitions, and message brokers.&lt;br&gt;
Thursday:&lt;br&gt;
Lab: Integrate Kafka with PostgreSQL for real-time updates.&lt;br&gt;
Friday:&lt;br&gt;
Peer Project: Build a real-time Kafka pipeline for transactional data.&lt;br&gt;
Saturday (Lab):&lt;br&gt;
Mini Project: Create a pipeline to stream e-commerce data with Kafka.&lt;br&gt;
Week 8: Batch vs. Stream Processing&lt;br&gt;
Monday:&lt;br&gt;
Theory: Introduction to batch vs. stream processing.&lt;br&gt;
Tuesday:&lt;br&gt;
Lab: Batch processing with PySpark.&lt;br&gt;
Wednesday:&lt;br&gt;
Theory: Combining batch and stream processing workflows.&lt;br&gt;
Thursday:&lt;br&gt;
Lab: Real-time processing with Apache Flink and Spark Streaming.&lt;br&gt;
Friday:&lt;br&gt;
Job Shadowing: Observe a real-time processing pipeline.&lt;br&gt;
Saturday (Lab):&lt;br&gt;
Mini Project: Build a hybrid pipeline combining batch and real-time processing.&lt;br&gt;
Month 3: Advanced Data Engineering&lt;br&gt;
Week 9: Machine Learning Integration in Data Pipelines&lt;br&gt;
Monday:&lt;br&gt;
Theory: Overview of ML workflows in data engineering.&lt;br&gt;
Tuesday:&lt;br&gt;
Lab: Preprocess data for machine learning using Pandas and PySpark.&lt;br&gt;
Wednesday:&lt;br&gt;
Theory: Feature engineering and automated feature extraction.&lt;br&gt;
Thursday:&lt;br&gt;
Lab: Automate feature extraction using Apache Airflow.&lt;br&gt;
Friday:&lt;br&gt;
Peer Project: Build a simple pipeline that integrates ML models.&lt;br&gt;
Saturday (Lab):&lt;br&gt;
Mini Project: Build an ML-powered recommendation system in a pipeline.&lt;br&gt;
Week 10: Spark and PySpark for Big Data&lt;br&gt;
Monday:&lt;br&gt;
Theory: Introduction to Apache Spark for big data processing.&lt;br&gt;
Tuesday:&lt;br&gt;
Lab: Set up Spark and PySpark for data analysis.&lt;br&gt;
Wednesday:&lt;br&gt;
Theory: Spark RDDs, DataFrames, and SQL.&lt;br&gt;
Thursday:&lt;br&gt;
Lab: Analyze large datasets using Spark SQL.&lt;br&gt;
Friday:&lt;br&gt;
Peer Project: Build a PySpark pipeline for large-scale data processing.&lt;br&gt;
Saturday (Lab):&lt;br&gt;
Mini Project: Analyze big data sets with Spark and PySpark.&lt;br&gt;
Week 11: Advanced Apache Airflow Techniques&lt;br&gt;
Monday:&lt;br&gt;
Theory: Advanced Airflow features (XCom, task dependencies).&lt;br&gt;
Tuesday:&lt;br&gt;
Lab: Implement dynamic DAGs and task dependencies in Airflow.&lt;br&gt;
Wednesday:&lt;br&gt;
Theory: Airflow scheduling, monitoring, and error handling.&lt;br&gt;
Thursday:&lt;br&gt;
Lab: Create complex DAGs for multi-step ETL pipelines.&lt;br&gt;
Friday:&lt;br&gt;
Job Shadowing: Observe advanced Airflow pipeline implementations.&lt;br&gt;
Saturday (Lab):&lt;br&gt;
Mini Project: Design an advanced Airflow DAG for complex data workflows.&lt;br&gt;
Week 12: Data Lakes and Delta Lake&lt;br&gt;
Monday:&lt;br&gt;
Theory: Data lakes, Lakehouses, and Delta Lake architecture.&lt;br&gt;
Tuesday:&lt;br&gt;
Lab: Set up Delta Lake on AWS for data storage and management.&lt;br&gt;
Wednesday:&lt;br&gt;
Theory: Managing schema evolution in Delta Lake.&lt;br&gt;
Thursday:&lt;br&gt;
Lab: Implement batch and real-time data loading to Delta Lake.&lt;br&gt;
Friday:&lt;br&gt;
Peer Project: Design a Lakehouse architecture for an e-commerce platform.&lt;br&gt;
Saturday (Lab):&lt;br&gt;
Mini Project: Implement a scalable Delta Lake architecture.&lt;br&gt;
Month 4: Capstone Projects&lt;br&gt;
Week 13: Batch Data Pipeline Development&lt;br&gt;
Monday to Thursday:&lt;br&gt;
Design and Implementation:&lt;br&gt;
Build an end-to-end batch data pipeline for e-commerce sales analytics.&lt;br&gt;
Tools: PySpark, SQL, PostgreSQL, Airflow, S3.&lt;br&gt;
Friday:&lt;br&gt;
Peer Review: Present progress and receive feedback.&lt;br&gt;
Saturday (Lab):&lt;br&gt;
Project Milestone: Finalize and present batch pipeline results.&lt;br&gt;
Week 14: Real-Time Data Pipeline Development&lt;br&gt;
Monday to Thursday:&lt;br&gt;
Design and Implementation:&lt;br&gt;
Build an end-to-end real-time data pipeline for IoT sensor monitoring.&lt;br&gt;
Tools: Kafka, Spark Streaming, Flink, S3.&lt;br&gt;
Friday:&lt;br&gt;
Peer Review: Present progress and receive feedback.&lt;br&gt;
Saturday (Lab):&lt;br&gt;
Project Milestone: Finalize and present real-time pipeline results.&lt;br&gt;
Week 15: Final Project Integration&lt;br&gt;
Monday to Thursday:&lt;br&gt;
Design and Implementation:&lt;br&gt;
Integrate both batch and real-time pipelines for a comprehensive end-to-end solution.&lt;br&gt;
Tools: Kafka, PySpark, Airflow, Delta Lake, PostgreSQL, and S3.&lt;br&gt;
Friday:&lt;br&gt;
Job Shadowing: Observe senior engineers integrating complex pipelines.&lt;br&gt;
Saturday (Lab):&lt;br&gt;
Project Milestone: Showcase integrated solution for review.&lt;br&gt;
Week 16: Capstone Project Presentation&lt;br&gt;
Monday to Thursday:&lt;br&gt;
Final Presentation Preparation:&lt;br&gt;
Polish, test, and document the final project.&lt;br&gt;
Friday:&lt;br&gt;
Peer Review: Present final projects to peers and receive feedback.&lt;br&gt;
Saturday (Lab):&lt;br&gt;
Capstone Presentation: Showcase completed capstone projects to industry professionals and instructors.&lt;/p&gt;

</description>
    </item>
    <item>
      <title>YOUTUBE ANALYSIS USING PYTHON AND YOUTUBE API</title>
      <dc:creator>Evans Jones</dc:creator>
      <pubDate>Thu, 05 Sep 2024 13:44:18 +0000</pubDate>
      <link>https://dev.to/evans_jones_9658a6f26f2ad/youtube-analysis-using-python-and-youtube-api-3nhe</link>
      <guid>https://dev.to/evans_jones_9658a6f26f2ad/youtube-analysis-using-python-and-youtube-api-3nhe</guid>
      <description>&lt;p&gt;INTRODUCTION&lt;/p&gt;

&lt;p&gt;It is a simple project that analyzes how one can extract data from youtube by intergrating youtube api keys and channel ids to fetch for video content analysis,subscriber trends,video details and commnets.&lt;/p&gt;

&lt;p&gt;FEATURES&lt;/p&gt;

&lt;p&gt;Retrieve channel statistics, Get detailed information about YouTube channels, including subscriber count, view count, video count, and other relevant metrics.&lt;/p&gt;

&lt;p&gt;Fetch video details,Extract data such as video title, description, duration, view count, like count, dislike count, and publish date for individual videos.&lt;/p&gt;

&lt;p&gt;Analyze comments, Retrieve comments made on YouTube videos and perform analysis, such as sentiment analysis or comment sentiment distribution.&lt;/p&gt;

&lt;p&gt;Generate reports, Generate reports and visualizations based on the collected data, allowing users to gain insights into channel performance, video engagement, and audience interaction.&lt;/p&gt;

&lt;p&gt;Data storage, Store the collected YouTube data in a database for easy retrieval and future reference.&lt;/p&gt;

&lt;p&gt;TECHNOLOGIES USED&lt;/p&gt;

&lt;p&gt;Python,its a language built with mathematical libraries and different functions.&lt;/p&gt;

&lt;p&gt;Youtube Data Api,it aids in generating custom reports containing youtube analytics data.&lt;/p&gt;

&lt;p&gt;Matplotlib, A popular data visualization library in Python used for creating charts, graphs, and visual representations of the data retrieved from YouTube. Matplotlib helps in analyzing and presenting the data in a meaningful way.&lt;/p&gt;

&lt;p&gt;Pandas,A powerful data manipulation and analysis library in Python. Pandas is used in the YouTube Data Scraper to handle and process data obtained from YouTube, providing functionalities such as data filtering, transformation, and aggregation.&lt;/p&gt;

&lt;p&gt;PROCESS FLOW&lt;br&gt;
Obtain youtube Api credentials:&lt;br&gt;
-Visit the Google Cloud Console.&lt;br&gt;
-Create a new project or select an existing project.&lt;br&gt;
-Enable the YouTube Data API v3 for your project.&lt;br&gt;
-Create API credentials for youtube API v3.&lt;/p&gt;

&lt;p&gt;CONCLUSION&lt;br&gt;
Through the youtube api scrapper has enabled the efficient way of gauging different sets of data,visualizations,insights,content subscribers,engagements,views and understanding different digital landscapes&lt;/p&gt;

</description>
    </item>
    <item>
      <title>THE ULTIMATE GUIDE TO DATA ANALYTICS</title>
      <dc:creator>Evans Jones</dc:creator>
      <pubDate>Sat, 24 Aug 2024 11:42:28 +0000</pubDate>
      <link>https://dev.to/evans_jones_9658a6f26f2ad/the-ultimate-guide-to-data-analytics-nab</link>
      <guid>https://dev.to/evans_jones_9658a6f26f2ad/the-ultimate-guide-to-data-analytics-nab</guid>
      <description>&lt;p&gt;DATA ANALYTICS&lt;br&gt;
It is the process of inspecting,cleansing,transforming and modeling data to discover actionable insights that support descision-making.&lt;/p&gt;

&lt;p&gt;TYPES OF DATA ANALYSIS&lt;br&gt;
1)Descriptive Analysis,it is based on involving and understanding past data.&lt;/p&gt;

&lt;p&gt;2)Diagnostic Analysis,it questions how data is obtained and analyses past data using techniques like drill-down,data discovery and correlations.&lt;/p&gt;

&lt;p&gt;3)Predictive Analysis,it makes predictions as to future events by use of historical data.&lt;br&gt;
-it makes use of statistical models and machine learning techniques to forecast future. trends.&lt;/p&gt;

&lt;p&gt;4)Prescriptive Analytics,this deals with what is next wwith data recommending actions based on predictive analytics.&lt;br&gt;
-it combines insights from all previous analytics types and uses optimization and simulation algorithims.&lt;/p&gt;

&lt;p&gt;DATA ANALYSIS TOOLS.&lt;br&gt;
1) PROGRAMMING LANGUAGES&lt;br&gt;
i)Python,it is a popular close-to-human programming language with libraries such as pandas,Numpy and SciPy which facilitate data analysis tasks.&lt;br&gt;
ii)Sql,it queries and manages dataases.&lt;/p&gt;

&lt;p&gt;2)DATA VISUALIZAION TOOLS AND STATISTICAL ANALYSIS TOOLS&lt;br&gt;
i)R,it is a language tailored for statistical analysis and data visualization.&lt;br&gt;
ii)Tableau,it creates interactive and shareable dashboards.&lt;br&gt;
iii)Power BI,it provides interactive visualizations and business intelligence capabilities.&lt;br&gt;
iv)Excel,it is a spreadsheet software that offers basic statistical tools.&lt;br&gt;
v)SAS,it offers GUI and scripting options for advanced analyses and publication of worthy graphics and charts.&lt;/p&gt;

&lt;p&gt;3)BIG DATA TOOLS&lt;br&gt;
i)NOSQL Databases(MongoDB),it is designed for storing,retreivng and managing big data.&lt;/p&gt;

&lt;p&gt;4)JUPYTER NOTEBOOKS&lt;br&gt;
it provides an interactive environment where users can combine code execution,text and rich media making them an excellent tool for exploratory data analysis and sharing results.&lt;/p&gt;

&lt;p&gt;DATA ANALYTICS TECHNIQUES&lt;br&gt;
1)Data Cleaning(Data processing),it involves identifying and correcting errors in the dataset by handling missing values,removing duplicates and correcting incosistencies.&lt;br&gt;
2)Data Exploration and Visualization,it involves examining the dataset structure with visualizaion tools and techniques like histograms,scatterplots and boxplots help in understanding the data underlying patterns and distributions.&lt;br&gt;
3)Statistical Analysis,it forms the backbone of data analysis with techniques like: i)Descriptive statistics,mean,mode,median.&lt;br&gt;
ii)Inferential statistics,hypothesis testing,confidence intervals.&lt;br&gt;
iii)Advanced statistical modelling,regression analysis,ANOVA.&lt;br&gt;
4)Machine Learning,it involves training algorithims to learn from and make predictions on data.common techniques used:&lt;br&gt;
i)Supervised Learning,classification,regression.&lt;br&gt;
ii)unsupervised learning,clustering and associaton.&lt;br&gt;
iii)Reinforcement learning,decision making.&lt;br&gt;
5)Data Mining,it involves discovering patterns in large datasets using methods at the intersecion of machine learning,statistics and database systems.&lt;br&gt;
Techniques used:&lt;br&gt;
i)Association rule learning.&lt;br&gt;
ii)Cluster analysis.&lt;br&gt;
iii)Anomaly detection&lt;br&gt;
6)Time Series Analysis,it focuses on data points collected or recorded at specific time intervals.&lt;br&gt;
Techniques used:&lt;br&gt;
i)ARIMA(Auto-Regressive Integrated Moving Average).&lt;br&gt;
ii)Exponential smoothing.&lt;br&gt;
iii)Seasonal decomposition.&lt;/p&gt;

&lt;p&gt;CONCLUSION&lt;br&gt;
Data analytics is an ever-evolving field that leverages various techniques and tools to transform raw data into actionable insights.&lt;br&gt;
whether you're cleaning data, visualizing patterns, or building predictive models, the right combination of methods and technologies can significantly enhance your ability to make data-driven decisions. Embrace these techniques and tools to unlock the full potential of your data and drive impactful outcomes in your domain.&lt;/p&gt;

</description>
    </item>
    <item>
      <title>FEATURE ENGINEERING</title>
      <dc:creator>Evans Jones</dc:creator>
      <pubDate>Sat, 17 Aug 2024 10:37:40 +0000</pubDate>
      <link>https://dev.to/evans_jones_9658a6f26f2ad/feature-engineering-23ag</link>
      <guid>https://dev.to/evans_jones_9658a6f26f2ad/feature-engineering-23ag</guid>
      <description>&lt;p&gt;FEATURE ENGINEERING&lt;br&gt;
it is the process of transforming raw data into relevant information for use by machine learning models i.e creating predictive models features.&lt;/p&gt;

&lt;p&gt;FEATURE ENGINEERING PROCESS&lt;br&gt;
1.) FEATURE CREATION&lt;br&gt;
it is the process of generating new features based on domain knowledge or by observing patterns in data.&lt;/p&gt;

&lt;p&gt;TYPES OF FEATURE CREATION&lt;br&gt;
a) Domain-Specific,creates new features based on domain knowledge like business rules or standards.&lt;br&gt;
b)Data-Driven,creates new features by observing patterns in the data i.e calculating aggregations or creating interaction features.&lt;br&gt;
c)Synthectic,generating new features by combining existing features or synthesizing new data points.&lt;/p&gt;

&lt;p&gt;IMPORTANCE OF FEATURE CREATION&lt;br&gt;
improves model performance by adding additional information to the model.&lt;br&gt;
it increases model robustness.&lt;br&gt;
it improves model interpretability,it makes it easier to understand the model predictions.&lt;br&gt;
it increases model flexibility,in handling different types of data.&lt;/p&gt;

&lt;p&gt;2.)FEATURE TRANSFORMATION&lt;br&gt;
it is the process of transforming the features into a more suitable representation for the machine learning model.&lt;/p&gt;

&lt;p&gt;TYPES OF FEATURE TRANSFORMATION&lt;br&gt;
a)Normalization,they are rescaling features at a specific range i.e 0 and 1.&lt;br&gt;
b)Scaling,it is used to transform numerical variables to have a similar scale i.e for easy comparison.&lt;br&gt;
c)Encoding,transforming categorical features into a numerical representation like one-hot encoding and label encoding.&lt;br&gt;
d)Transformation,it uses mathematical operations to change the distribution or scale of the features i.e logarithmic,square root,reciprocal transformations.&lt;/p&gt;

&lt;p&gt;SIGNIFICANCE OF FEATURE TRANSFORMATION&lt;br&gt;
-improves model performance.&lt;br&gt;
-increases model robustness.&lt;br&gt;
-improves computational efficiency.&lt;br&gt;
-improves model interpretability&lt;/p&gt;

&lt;p&gt;3.)FEATURE EXTRACTION&lt;br&gt;
It is the process of creating new features from existing ones to provide more relevant information to the machine model.&lt;br&gt;
TYPES OF FEATURE EXTRACTION&lt;br&gt;
a)Dimensionality Reduction,reducing the number of features by transforming the data into lower-dimensional space while retaining important information e.g PCA and t-SNE.&lt;br&gt;
b)Feature Combination,combines two or more existing features to create a new one.&lt;br&gt;
c)Feature Aggregation,it aggregates features to create a new one.like calculating mean,sum,count of set of features.&lt;br&gt;
d)Feature Transformation,transforming existing features into new representaion.&lt;/p&gt;

&lt;p&gt;SIGNIFICANCE OF FEATURE EXTRACTION&lt;br&gt;
-improves model performance.&lt;br&gt;
-reduces overfitting&lt;br&gt;
-improves computational efficiency.&lt;br&gt;
-improves model interpretability.&lt;/p&gt;

&lt;p&gt;4)FEATURE SELECTION&lt;br&gt;
it is the process of selecting a subset of relevant features from the dataset to be used in a machine-learning model.&lt;/p&gt;

&lt;p&gt;TYPES OF FEATURE SELECTION&lt;br&gt;
a)Filter Method,Based on statistical measure of relationship between the feature and target variable.&lt;br&gt;
b)Wrapper Method,based on the evaluation of the feature subset using a specific machine learning algorithim.&lt;br&gt;
c)Embedded Method,based on the feature selection as part of the training process of the machine learning algorithim.&lt;/p&gt;

&lt;p&gt;SIGNIFICANCE OF FEATURE SELECTION&lt;br&gt;
-reduces overfitting.&lt;br&gt;
-improves model performance.&lt;br&gt;
-decreases computational costs.&lt;br&gt;
-improves interpretability&lt;/p&gt;

&lt;p&gt;5)FEATURE SCALING&lt;br&gt;
it is the process of transforming the features so that they may have a similar scale.&lt;/p&gt;

&lt;p&gt;TYPES OF FEATURE SCALING&lt;br&gt;
a)Min-Max Scaling,it rescales features to a specific range i.e 0 and 1 by subtracting minimum value and diving by the range.&lt;br&gt;
b)Standard Scaling,it rescales the features to have a mean of 0 and a standard deviation of 1 by subtracting the mean and dividing by the standard deviation.&lt;br&gt;
c)Robust Scaling,it rescales the features to be robust to ouliers by diving them by interquartile range.&lt;/p&gt;

&lt;p&gt;SIGNIFICANCE OF FEATURE SCALING&lt;br&gt;
-improves model performance.&lt;br&gt;
-increases model robustness.&lt;br&gt;
-improves computational efficiency.&lt;br&gt;
-improves model interpretability&lt;/p&gt;

&lt;p&gt;STEPS IN FEATURE ENGINEERING&lt;br&gt;
a) Data Cleansing(data cleaning/scrubbing),involes identifying and removing or correcting any errors or inconsistensis in the dataset.&lt;br&gt;
b)Data Transformation,it converts and structures data into a usable format that can be easily analyzed.&lt;br&gt;
c)Feature Extraction,it involves pattern recognition and identifying common themes among a large collection of documents.&lt;br&gt;
d)Feature Selection,involves selecting the most relevant features from the dataset using techniques like corretional analysis,mutual information ,stepwisw regression.&lt;br&gt;
e)Feature Iteration,involves refining and improving the features based on the performance of machine learning.it uses techniques like adding new features,removing redundant features,transforming features.&lt;/p&gt;

&lt;p&gt;TECHNIQUES USED IN FEATURE ENGINEERING&lt;br&gt;
1)ONE-HOT ENCODING&lt;br&gt;
it is a technique used to transform categorical variables into numerical values that can be used by machine learning models.&lt;br&gt;
-every category is transformed into a binary value indicating its presence or absence.&lt;/p&gt;

&lt;p&gt;2)BINNING&lt;br&gt;
it is a technique used to transform continous variables into categorical variables.&lt;br&gt;
-range of values of the continous variable is divided into several bins and each bin is assigned a categorical value&lt;br&gt;
-18-80 can be binned into 18-25,26,35,36-50 and 51-80&lt;/p&gt;

&lt;p&gt;3)FEATURE SPLIT&lt;br&gt;
it involves diving single features into multiple sub-features or groups based on specific criteria.&lt;br&gt;
-the process unlocks valuable insights and enhances  the model ability to capture complex relationships and patterns within the data.&lt;/p&gt;

&lt;p&gt;4)TEXT DATA preprocessing&lt;br&gt;
it involves removing stopwords,stemming,lemmatization and vectorization.&lt;br&gt;
i)Stop Words,they are words that do not add much meaning to the text i.e 'the' and 'and.&lt;br&gt;
ii)Stemming,it involves reducing words to their root forms,such as converting 'running' to 'run'&lt;br&gt;
iii)Lemmatization,it reduces words to their ase form i.e vonverting 'running' to 'run'.&lt;br&gt;
iv)Vectoriztion,it involves transforming text data into numerical vectors that can be used by machine learning models.&lt;/p&gt;

&lt;p&gt;FEATURE ENGINEERING TOOLS&lt;br&gt;
1)FEATURE TOOLS&lt;br&gt;
it is a python library that enables automatic feature engineering for structured data.&lt;br&gt;
it extracts features from multiple tables i.e relational dataases,csv files and generate new features based on user-defined primitives.&lt;/p&gt;

&lt;p&gt;FEATURES OF FEATURE TOOLS&lt;br&gt;
-Automated feature engineering using machine learning algorithims.&lt;br&gt;
-support for handling time-dependent data.&lt;br&gt;
-Intergration with popular python libraries i.e pandas and sciki-learn.&lt;br&gt;
-Visualization tools for exploring and analyzing the generated features.&lt;br&gt;
-Extensive documentation and tutorials for getting started.&lt;/p&gt;

&lt;p&gt;2)TPOT(Tree-based Pipeline Optimization Tool)&lt;br&gt;
it uses genetic programming to search for the best combination of features and machine learning algorithims for a given dataset.&lt;/p&gt;

&lt;p&gt;FEATURES OF TPOT&lt;br&gt;
-Autmatic feature selection and transformation.&lt;br&gt;
-Support for multiple types of machine learning models i.e regression,classification,clustering.&lt;br&gt;
-Ability to handle missing data and categorical variables.&lt;br&gt;
-Intergration with popular libraries like scikit-learn and pandas.&lt;br&gt;
-interactive visualizaions of generated pipelines.&lt;/p&gt;

&lt;p&gt;3)DATA ROBOT&lt;br&gt;
it uses automated machine learning techniques to generate new features and select the best combination of features and models for a given dataset.&lt;br&gt;
FEATURES OF DATA ROBOT&lt;br&gt;
-Automated feature engineering using machine learning algorithims.&lt;br&gt;
-Support for handling time-dependant and text data.&lt;br&gt;
-Intergration with popular python libraries like pandas,scikit-learn&lt;br&gt;
-interactive visualization of the generated models and features.&lt;br&gt;
-Collaboration tools for teams working on machine learning projects.&lt;/p&gt;

&lt;p&gt;4)ALTERYX&lt;br&gt;
it provides a visual interface for creating data pipelines that can extract,transform and generate features from multiple data sources.&lt;/p&gt;

&lt;p&gt;FEATURES OF ALTERYX&lt;br&gt;
-support for handling structured and unstructured data.&lt;br&gt;
-Intergration with popular data sources like excel and databases.&lt;br&gt;
-Pre-built tools for feature extraction and transformation.&lt;br&gt;
-Support for custom scripting and code intergration.&lt;br&gt;
-Collaboration and sharing tools for teams working on data projects.&lt;/p&gt;

&lt;p&gt;5.)H20.ai&lt;br&gt;
it providesa range of automated feature engineering techniques like feature scaling,imputation,encoding and feature engineering capabilities for more advanced users.&lt;/p&gt;

&lt;p&gt;FEATURES OF H20.ai&lt;br&gt;
-Automatic and manual feature engineering options&lt;br&gt;
-Support for structured and unstructured data including text and image data.&lt;br&gt;
-Intergration with popular data sources like csv files and databases.&lt;br&gt;
-Interactive visualizaion of generated features and models.&lt;br&gt;
-Collaboration and sharing tools for teams working on machine learning projects.&lt;/p&gt;

</description>
    </item>
    <item>
      <title>EXPLAROTARY DATA ANALYSIS TECHNIQUES</title>
      <dc:creator>Evans Jones</dc:creator>
      <pubDate>Fri, 09 Aug 2024 11:34:38 +0000</pubDate>
      <link>https://dev.to/evans_jones_9658a6f26f2ad/explarotary-data-analysis-techniques-pna</link>
      <guid>https://dev.to/evans_jones_9658a6f26f2ad/explarotary-data-analysis-techniques-pna</guid>
      <description>&lt;p&gt;INTRODUCTION&lt;br&gt;
Descriptive analysis is simply how we describe basic features of dataset and obtains of a short summary about the sample and measures of data.&lt;/p&gt;

&lt;p&gt;IMPORTING LIBRARIES&lt;br&gt;
import pandas as pd&lt;br&gt;
import numpy as np&lt;br&gt;
import matplotlib.pyplot as plt&lt;br&gt;
% matplotlib inline&lt;br&gt;
import seaborn as sns&lt;br&gt;
from scipy import stats&lt;br&gt;
from scipy.stats import pearsonr&lt;/p&gt;

&lt;p&gt;Functions Used In Explarotary Data Analysis&lt;/p&gt;

&lt;p&gt;df=pd.read_csv(Assingment2/automobileEDA.csv)&lt;br&gt;
df.head()&lt;br&gt;
df.describe()&lt;br&gt;
df.describe(include='all')&lt;/p&gt;

&lt;p&gt;VALUE COUNTS&lt;br&gt;
drive_wheels_counts=df['drive-wheels'].value_counts()&lt;br&gt;
drive_wheels_counts&lt;/p&gt;

&lt;p&gt;BOX PLOTS&lt;br&gt;
it is based on visualizing data based numeric data and various distributions&lt;/p&gt;

&lt;p&gt;sns.boxplot(x='drive-wheels',y='price',data=df)&lt;br&gt;
df.info()&lt;/p&gt;

&lt;p&gt;SCATTER PLOTS&lt;br&gt;
They are datapoints numbers contained in some range&lt;/p&gt;

&lt;p&gt;y=df['price']&lt;br&gt;
x=df['engine-size']&lt;br&gt;
plt.scatter(x,y)&lt;br&gt;
plt.title('scatter plot of Engine Size vs Price)&lt;br&gt;
plt.xlabel('Engine Size')&lt;br&gt;
plt.ylabel('Price')&lt;br&gt;
Conclusion:As the engine size is increasing the price too increases&lt;/p&gt;

&lt;p&gt;GROUP BY&lt;br&gt;
It is used on categorical variable,groups the data into subsets according to different categories of that variable&lt;/p&gt;

&lt;p&gt;GROUP BY DRIVEWHEELS&lt;br&gt;
df_1grp=df_1.groupby(['drive-wheels'],as_insex=False).mean()&lt;br&gt;
df_1grp&lt;/p&gt;

&lt;p&gt;GROUP BY BODY STYLE&lt;br&gt;
df[['body-style']].value_counts()&lt;/p&gt;

&lt;p&gt;GROUP BY ROWS&lt;br&gt;
df_test=df[['drive-wheels','body-style','price']]&lt;br&gt;
df_test.head()&lt;/p&gt;

&lt;p&gt;GROUP BASED ON TWO VARIABLES AND FIND MEAN&lt;br&gt;
df_grp=df_test.groupby9['drive-wheels','body-style'],as_index=False).mean()&lt;br&gt;
df_grp&lt;/p&gt;

&lt;p&gt;PIVOT TABLE&lt;br&gt;
it helps in visualizing data in a readable format&lt;br&gt;
df_pivot=df_grp.pivot(index='drive-wheels',columns='body-style')&lt;br&gt;
df_pivot&lt;/p&gt;

&lt;p&gt;HEAT MAP PLOT&lt;br&gt;
it takes a rectangular grid of data and assigns a color intensity based on the data value at the grid points.&lt;/p&gt;

&lt;p&gt;plt.pcolor(df_pivot,cmap='RdBu')&lt;br&gt;
plt.colorbar()&lt;br&gt;
plt.show()&lt;/p&gt;

&lt;p&gt;CORRELATION&lt;br&gt;
It is a statistical metric for measuring to what extent different variables are independent.&lt;/p&gt;

&lt;p&gt;sns.regplot(x='engine-size',y='price',data=df)&lt;br&gt;
plt.ylim(0,)&lt;/p&gt;

&lt;p&gt;WEAK CORRELATION&lt;br&gt;
sns=regplot(x='highway-mpg',y='price',data=df)&lt;br&gt;
plt.ylim(0,)&lt;/p&gt;

&lt;p&gt;CORRELATION STATISTICS&lt;br&gt;
it will be based according two methods:&lt;br&gt;
1)correlation coefficient,it uses +1,-1,0&lt;br&gt;
2)P-values,it tells us how certain we are about the correlation with calculated values&lt;/p&gt;

&lt;p&gt;REMOVE ROWS WITH NAN OR INFINITE VALUES &lt;br&gt;
df_cleaned=df[['horsepower','price']].dropna()&lt;br&gt;
df_cleaned=df_cleaned[np.isfinite(df_cleaned).all(1)]&lt;/p&gt;

&lt;p&gt;CALCULATE PEARSON CORRELATION&lt;br&gt;
pearson_coef,p.value_stats.peasonor(df_cleaned['horsepower'],df_cleaned['price'])&lt;br&gt;
pearson_coef,p_value&lt;/p&gt;

&lt;p&gt;CORRELATION MATRIX&lt;br&gt;
df_numeric=df.select_dtypes(include=[float,int])&lt;br&gt;
corr_matri=df_numeric.corr()&lt;br&gt;
corr_matrix&lt;/p&gt;

&lt;p&gt;VISUALIZE HEATMAP&lt;br&gt;
plt.figurre(figuresize=(10,8))&lt;br&gt;
sns.heatmap(corr_matrix,annot=True,cmap='coolwarm',fmt='.2f',linewidths=0.5)&lt;br&gt;
plt.title('Correlation Heatmap')&lt;br&gt;
plt.show()&lt;/p&gt;

&lt;p&gt;CONCLUSION&lt;br&gt;
As the engine size is increasing the price too increases.&lt;/p&gt;

</description>
    </item>
    <item>
      <title>STANDARD QUERY LANGUAGE</title>
      <dc:creator>Evans Jones</dc:creator>
      <pubDate>Wed, 07 Aug 2024 11:08:45 +0000</pubDate>
      <link>https://dev.to/evans_jones_9658a6f26f2ad/standard-query-language-j2b</link>
      <guid>https://dev.to/evans_jones_9658a6f26f2ad/standard-query-language-j2b</guid>
      <description>&lt;p&gt;INTRODUCTION&lt;br&gt;
The following is a query language that helps us store values in a database which can be integrated with different languages by using user interfaces&lt;/p&gt;

&lt;p&gt;INSTALLATION&lt;br&gt;
Download sql from any browser of your choice follow up the setup wizard process to complete your installation then load it in your pc/laptop&lt;/p&gt;

&lt;p&gt;DATASET OVERVIEW&lt;br&gt;
It will compromise of a server that stores different values of big data based of various analysis &lt;/p&gt;

&lt;p&gt;BASIC FUNCTIONS IN SQL&lt;br&gt;
1) WHERE CLAUSE&lt;br&gt;
select *&lt;br&gt;
from tablename&lt;br&gt;
where column_name = condition&lt;/p&gt;

&lt;p&gt;2)LIMIT&lt;br&gt;
select *&lt;br&gt;
from tablename&lt;br&gt;
limit 3&lt;/p&gt;

&lt;p&gt;3) CREATING TABLES&lt;br&gt;
insert into('data1','data2','data3')&lt;br&gt;
values(1,2,3)&lt;/p&gt;

&lt;p&gt;CONCLUSION&lt;br&gt;
Sql should aid in storing complex data by using different functions and formulas.&lt;/p&gt;

</description>
    </item>
    <item>
      <title>PYTHON PROGRAMMING</title>
      <dc:creator>Evans Jones</dc:creator>
      <pubDate>Wed, 07 Aug 2024 10:44:12 +0000</pubDate>
      <link>https://dev.to/evans_jones_9658a6f26f2ad/python-programming-25ob</link>
      <guid>https://dev.to/evans_jones_9658a6f26f2ad/python-programming-25ob</guid>
      <description>&lt;p&gt;INTRRODUCTION&lt;br&gt;
The following is a simple documention of how to learn python by understanding different concepts like variables initalization,functions,declarations and applying in different environments.&lt;/p&gt;

&lt;p&gt;INSTALLATION&lt;br&gt;
Download python from your local browser and follow up the installation process and later load it on your machine.&lt;/p&gt;

&lt;p&gt;pip install python&lt;/p&gt;

&lt;p&gt;BASIC PYTHON FUNCTIONS&lt;br&gt;
1) Text type&lt;br&gt;
   it defines a string(str)&lt;br&gt;
2) Numeric types&lt;br&gt;
   it defines a whole number and numbers with decimal precision(int,float)&lt;br&gt;
3) Sequence types&lt;br&gt;
   it defines(list,Tuples,Range)&lt;br&gt;
4) Mapping types&lt;br&gt;
  it is used to define dictionaries(dict)&lt;/p&gt;

&lt;p&gt;CONCLUSION&lt;br&gt;
By the end of your learning python one should be able apply the concepts in different fields like machine learning,selenium and django&lt;/p&gt;

</description>
    </item>
    <item>
      <title>Weather_Data_Set</title>
      <dc:creator>Evans Jones</dc:creator>
      <pubDate>Fri, 02 Aug 2024 10:34:12 +0000</pubDate>
      <link>https://dev.to/evans_jones_9658a6f26f2ad/weatherdataset-odf</link>
      <guid>https://dev.to/evans_jones_9658a6f26f2ad/weatherdataset-odf</guid>
      <description>&lt;p&gt;INTRODUCTION&lt;/p&gt;

&lt;p&gt;The following analysis is based on weather dataset using different machine algorithims to make predictions of a certain weather phenom and fetch for data in various instances.&lt;/p&gt;

&lt;p&gt;Dataset Overview&lt;/p&gt;

&lt;p&gt;it will efficiently include the main modules i.e pandas which will be used for dataframes,numpy which is for storage of bulk data and matplotlib for data visiualization.&lt;/p&gt;

&lt;p&gt;Load Dataset&lt;/p&gt;

&lt;p&gt;Data will be obtained from kaggle and loaded into jupyter notebook environment to the specific folder path and also with the full installation of other dependancies like pyforest.&lt;/p&gt;

&lt;p&gt;Functions used in Pandas Data Analysis&lt;/p&gt;

&lt;p&gt;1.Shape&lt;br&gt;
data.shape&lt;/p&gt;

&lt;p&gt;2.Data types&lt;br&gt;
data.dtypes&lt;/p&gt;

&lt;p&gt;3.Unique&lt;br&gt;
data['Weather'].value_counts()&lt;/p&gt;

&lt;p&gt;4.Count&lt;br&gt;
data.info&lt;/p&gt;

&lt;p&gt;5.Describe&lt;br&gt;
data.describe()&lt;/p&gt;

&lt;p&gt;Answering different Data analysis Problems&lt;br&gt;
Q1.)Find the records where the weather was exactly clear.&lt;br&gt;
Q2.)Find the number of times the windspeed was exactly 4km/h.&lt;br&gt;
Q3.)Check if there are any NULL values present in the dataset.&lt;br&gt;
Q4.)Rename the column "Weather" to "Weather_Condition"&lt;br&gt;
Q5.)What is the mean visibility of the dataset.&lt;br&gt;
Q6.)Find the number of records where the wind speed is greater than 24km/h and visibility is equal to 25km&lt;br&gt;
Q7.)What is the mean value of each column for each Weather_Conditions.&lt;br&gt;
Q8.)Find all instances where the weather is clear and the relative humidity is greater than 50 or visibility is above 40.&lt;br&gt;
Q9.)Find the number of weather weather conditions that include snow.&lt;/p&gt;

&lt;p&gt;Conclusion&lt;/p&gt;

&lt;p&gt;Data analysis helps us visualize data graphycally by giving clarity of comparisons of different datasets and comprehend data easily and efficiently.&lt;/p&gt;

</description>
    </item>
  </channel>
</rss>
