<?xml version="1.0" encoding="UTF-8"?>
<rss version="2.0" xmlns:atom="http://www.w3.org/2005/Atom" xmlns:dc="http://purl.org/dc/elements/1.1/">
  <channel>
    <title>DEV Community: Aditi Sharma</title>
    <description>The latest articles on DEV Community by Aditi Sharma (@_adii3107).</description>
    <link>https://dev.to/_adii3107</link>
    <image>
      <url>https://media2.dev.to/dynamic/image/width=90,height=90,fit=cover,gravity=auto,format=auto/https:%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Fuser%2Fprofile_image%2F3466722%2Fc9813955-cfbf-46ac-b252-87bc39dca55d.jpg</url>
      <title>DEV Community: Aditi Sharma</title>
      <link>https://dev.to/_adii3107</link>
    </image>
    <atom:link rel="self" type="application/rss+xml" href="https://dev.to/feed/_adii3107"/>
    <language>en</language>
    <item>
      <title>🚀 Day 39 of My Data Journey</title>
      <dc:creator>Aditi Sharma</dc:creator>
      <pubDate>Thu, 09 Oct 2025 16:30:18 +0000</pubDate>
      <link>https://dev.to/_adii3107/day-39-of-my-data-journey-3lbk</link>
      <guid>https://dev.to/_adii3107/day-39-of-my-data-journey-3lbk</guid>
      <description>&lt;p&gt;A single groupby() or merge() in Pandas can save hours of manual work!&lt;/p&gt;

&lt;p&gt;The groupby() function in Pandas is a fundamental and highly useful tool for data analysis, enabling the "split-apply-combine" paradigm. It allows for the efficient analysis and transformation of datasets by grouping rows based on one or more column values.&lt;/p&gt;

&lt;p&gt;The pandas.merge() function is a powerful and essential tool for combining two DataFrameobjects in a database-style join operation. Its utility lies in its ability to integrate data from different sources based on common keys, enabling more comprehensive data analysis.&lt;/p&gt;

</description>
      <category>python</category>
      <category>ai</category>
    </item>
    <item>
      <title>🚀 Day 38 of My Data Journey</title>
      <dc:creator>Aditi Sharma</dc:creator>
      <pubDate>Wed, 08 Oct 2025 18:28:53 +0000</pubDate>
      <link>https://dev.to/_adii3107/day-38-of-my-data-journey-42oa</link>
      <guid>https://dev.to/_adii3107/day-38-of-my-data-journey-42oa</guid>
      <description>&lt;p&gt;Using Python for Data Modeling 🐍📊&lt;/p&gt;

&lt;p&gt;Today I explored how Python makes data modeling faster, smarter, and more intuitive!&lt;/p&gt;

&lt;p&gt;🔹 How Python Helps:&lt;br&gt;
    • Use Pandas for creating relational data structures.&lt;br&gt;
    • NumPy &amp;amp; Scikit-learn for statistical and predictive modeling.&lt;br&gt;
    • SQLAlchemy to design and manage database schemas.&lt;br&gt;
    • Graphviz for visualizing model relationships.&lt;/p&gt;

&lt;p&gt;💡 Fun Fact: With just a few lines of Python, you can model, analyze, and visualize complex datasets seamlessly!&lt;/p&gt;

&lt;p&gt;Python = Simplicity 🐍 + Power ⚡ = Better Data Models 🧩&lt;/p&gt;

</description>
      <category>webdev</category>
      <category>programming</category>
      <category>ai</category>
    </item>
    <item>
      <title>🚀 Day 37 of My Data Journey</title>
      <dc:creator>Aditi Sharma</dc:creator>
      <pubDate>Tue, 07 Oct 2025 17:20:04 +0000</pubDate>
      <link>https://dev.to/_adii3107/day-37-of-my-data-journey-19nk</link>
      <guid>https://dev.to/_adii3107/day-37-of-my-data-journey-19nk</guid>
      <description>&lt;p&gt;Understanding Data Modeling 🧩&lt;/p&gt;

&lt;p&gt;Today, I explored one of the most important concepts in data analytics — Data Modeling!&lt;/p&gt;

&lt;p&gt;🔹 What is Data Modeling?&lt;br&gt;
It’s the process of designing how data is stored, connected, and accessed. Think of it as the blueprint for your database.&lt;/p&gt;

&lt;p&gt;🔹 Types of Data Models:&lt;br&gt;
    • Conceptual: Defines what data is important.&lt;br&gt;
    • Logical: Describes how data is structured.&lt;br&gt;
    • Physical: Specifies where and how data is stored.&lt;/p&gt;

&lt;p&gt;🔹 Why It Matters:&lt;br&gt;
✅ Improves data consistency&lt;br&gt;
✅ Makes querying faster&lt;br&gt;
✅ Enhances scalability&lt;br&gt;
✅ Reduces redundancy&lt;/p&gt;

&lt;p&gt;💡 Fun Fact: A good data model can boost query performance by up to 40%!&lt;/p&gt;

&lt;p&gt;Data modeling = building the foundation for smarter analytics 🧠✨&lt;/p&gt;

</description>
      <category>python</category>
      <category>data</category>
      <category>database</category>
    </item>
    <item>
      <title>🚀 Day 36 of My Data Journey</title>
      <dc:creator>Aditi Sharma</dc:creator>
      <pubDate>Mon, 06 Oct 2025 13:51:45 +0000</pubDate>
      <link>https://dev.to/_adii3107/day-36-of-my-data-journey-1i64</link>
      <guid>https://dev.to/_adii3107/day-36-of-my-data-journey-1i64</guid>
      <description>&lt;p&gt;Machine Learning in Python for Data Analytics 🤖📊&lt;/p&gt;

&lt;p&gt;Today, I explored how Machine Learning (ML) blends with Data Analytics to uncover insights and make smarter decisions!&lt;/p&gt;

&lt;p&gt;🔹 What is ML?&lt;br&gt;
Machine Learning helps computers learn from data — no explicit programming needed!&lt;/p&gt;

&lt;p&gt;🔹 Why Python?&lt;br&gt;
Because Python = simplicity + powerful libraries like Scikit-learn, TensorFlow, and Pandas.&lt;/p&gt;

&lt;p&gt;🔹 In Data Analytics:&lt;br&gt;
•Detect patterns 🔍&lt;br&gt;
•Predict trends 📈&lt;br&gt;
•Automate insights ⚙️&lt;/p&gt;

&lt;p&gt;💡 Fun Fact: Around 80% of data analysts use Python for machine learning tasks today!&lt;/p&gt;

&lt;p&gt;Machine Learning turns data into decisions — that’s the real magic! ✨&lt;/p&gt;

</description>
      <category>python</category>
      <category>ai</category>
      <category>machinelearning</category>
    </item>
    <item>
      <title>🚀 Day 35 of My Data Journey</title>
      <dc:creator>Aditi Sharma</dc:creator>
      <pubDate>Sun, 05 Oct 2025 09:54:46 +0000</pubDate>
      <link>https://dev.to/_adii3107/day-35-of-my-data-journey-af9</link>
      <guid>https://dev.to/_adii3107/day-35-of-my-data-journey-af9</guid>
      <description>&lt;p&gt;Exploring AWS: S3, EC2 &amp;amp; Lambda ☁️&lt;/p&gt;

&lt;p&gt;Today I dived into the core pillars of AWS that make cloud computing so powerful 👇&lt;/p&gt;

&lt;p&gt;🔹 S3 (Simple Storage Service) – Your go-to cloud storage.&lt;br&gt;
Store, retrieve &amp;amp; manage any amount of data anytime, anywhere. 📦&lt;/p&gt;

&lt;p&gt;🔹 EC2 (Elastic Compute Cloud) – Virtual servers on demand! 💻&lt;br&gt;
Scale up or down instantly depending on your workload.&lt;/p&gt;

&lt;p&gt;🔹 Lambda – Run code without servers. ⚡&lt;br&gt;
Just write your function, and AWS runs it when needed — pay only for execution time!&lt;/p&gt;

&lt;p&gt;💡 Perks of AWS:&lt;br&gt;
✅ Scalability&lt;br&gt;
✅ Pay-as-you-go pricing&lt;br&gt;
✅ High availability&lt;br&gt;
✅ Developer-friendly ecosystem&lt;/p&gt;

&lt;p&gt;Cloud + Data = Infinite Possibilities ☁️✨&lt;/p&gt;

</description>
      <category>aws</category>
      <category>database</category>
      <category>datascience</category>
    </item>
    <item>
      <title>🚀 Day 34 of My Data Journey</title>
      <dc:creator>Aditi Sharma</dc:creator>
      <pubDate>Sat, 04 Oct 2025 08:55:57 +0000</pubDate>
      <link>https://dev.to/_adii3107/day-34-of-my-data-journey-55a0</link>
      <guid>https://dev.to/_adii3107/day-34-of-my-data-journey-55a0</guid>
      <description>&lt;p&gt;Spark vs. MapReduce Architecture🔥&lt;/p&gt;

&lt;p&gt;Both Apache Spark and Hadoop MapReduce are designed to handle Big Data, but their architectures make them very different.&lt;/p&gt;

&lt;p&gt;Spark processes data in-memory, which makes it up to 100x faster than MapReduce that writes data to disk after each step. Spark’s architecture is built around a Driver Program that coordinates Executors, and it uses a Directed Acyclic Graph (DAG) to optimize data flow. It supports both real-time and batch processing, making it versatile and efficient.&lt;/p&gt;

&lt;p&gt;On the other hand, MapReduce follows a two-step process (Map → Reduce), where intermediate results are written to disk. This ensures fault tolerance but adds latency. Its architecture revolves around Job Tracker and Task Trackers, which manage jobs in a sequential, disk-heavy manner.&lt;/p&gt;

&lt;p&gt;💡 Fun Fact: Spark can run on top of Hadoop’s storage layer (HDFS), combining Spark’s speed with Hadoop’s scalability!&lt;/p&gt;

&lt;p&gt;Spark = Speed ⚡ | MapReduce = Stability 💽&lt;/p&gt;

</description>
      <category>database</category>
      <category>productivity</category>
      <category>tutorial</category>
    </item>
    <item>
      <title>🚀 Day 33 of My Data Journey</title>
      <dc:creator>Aditi Sharma</dc:creator>
      <pubDate>Sat, 04 Oct 2025 08:55:15 +0000</pubDate>
      <link>https://dev.to/_adii3107/day-33-of-my-data-journey-432h</link>
      <guid>https://dev.to/_adii3107/day-33-of-my-data-journey-432h</guid>
      <description>&lt;p&gt;Apache Spark Architecture &amp;amp; Fun Facts 🔥&lt;/p&gt;

&lt;p&gt;At its core, Spark has a simple but powerful architecture:&lt;/p&gt;

&lt;p&gt;🔹 Driver Program → Controls the application.&lt;br&gt;
🔹 Cluster Manager → Allocates resources (YARN, Mesos, Standalone, Kubernetes).&lt;br&gt;
🔹 Executors → Run tasks on worker nodes.&lt;br&gt;
🔹 RDDs (Resilient Distributed Datasets) → The magic behind fault-tolerance &amp;amp; parallelism.&lt;/p&gt;

&lt;p&gt;✨ Interesting Facts:&lt;br&gt;
• Spark stores intermediate data in memory = ⚡ super fast.&lt;br&gt;
• Fault-tolerant by design → if a node fails, Spark rebuilds data.&lt;br&gt;
• Supports batch, streaming, ML, and graph in one framework.&lt;/p&gt;

&lt;p&gt;💡 Spark = A powerhouse of speed + scalability in Big Data!&lt;/p&gt;

</description>
      <category>webdev</category>
      <category>programming</category>
      <category>spark</category>
    </item>
    <item>
      <title>🚀 Day 32 of My Data Journey</title>
      <dc:creator>Aditi Sharma</dc:creator>
      <pubDate>Thu, 02 Oct 2025 13:13:30 +0000</pubDate>
      <link>https://dev.to/_adii3107/day-32-of-my-data-journey-5bl1</link>
      <guid>https://dev.to/_adii3107/day-32-of-my-data-journey-5bl1</guid>
      <description>&lt;p&gt;Why Do We Use Apache Spark? 🔥&lt;/p&gt;

&lt;p&gt;Apache Spark has become a favorite in the data world—and here’s why:&lt;/p&gt;

&lt;p&gt;🔹 Speed – 100x faster than Hadoop MapReduce ⚡&lt;br&gt;
🔹 Unified Engine – Works for batch, streaming, ML, and graph data.&lt;br&gt;
🔹 Scalability – Handles petabytes of data with ease 🌍&lt;br&gt;
🔹 Ease of Use – APIs in Python, Java, Scala, R = accessible for everyone.&lt;br&gt;
🔹 Community Power – One of the most active Apache projects.&lt;/p&gt;

&lt;p&gt;💡 Fun Fact: Spark powers Netflix recommendations and Uber’s real-time analytics!&lt;/p&gt;

&lt;p&gt;Spark isn’t just a tool, it’s the backbone of modern data processing.&lt;/p&gt;

</description>
      <category>programming</category>
      <category>database</category>
    </item>
    <item>
      <title>🚀 Day 31 of My Data Journey</title>
      <dc:creator>Aditi Sharma</dc:creator>
      <pubDate>Wed, 01 Oct 2025 15:17:00 +0000</pubDate>
      <link>https://dev.to/_adii3107/day-31-of-my-data-journey-54k6</link>
      <guid>https://dev.to/_adii3107/day-31-of-my-data-journey-54k6</guid>
      <description>&lt;p&gt;Myths About Apache Spark 🔥&lt;/p&gt;

&lt;p&gt;Apache Spark is one of the most powerful Big Data frameworks, but many myths surround it. Let’s clear the air:&lt;/p&gt;

&lt;p&gt;🔹 Myth 1: Spark = Only for Big Data&lt;br&gt;
👉 Reality: Spark works great even on smaller datasets for fast computation.&lt;/p&gt;

&lt;p&gt;🔹 Myth 2: Spark Replaces Hadoop&lt;br&gt;
👉 Reality: Spark can run on top of Hadoop (HDFS) – they complement each other.&lt;/p&gt;

&lt;p&gt;🔹 Myth 3: Spark = Only for Data Scientists&lt;br&gt;
👉 Reality: Spark is used by engineers, analysts, and researchers alike.&lt;/p&gt;

&lt;p&gt;🔹 Myth 4: Spark is Too Complex&lt;br&gt;
👉 Reality: With APIs in Python, Scala, Java, R, Spark is more approachable than many think.&lt;/p&gt;

&lt;p&gt;💡 Fun Fact: Spark was originally developed at UC Berkeley in 2009 and is now one of the most active Apache projects.&lt;/p&gt;

&lt;p&gt;Apache Spark = Speed ⚡ + Scalability 🌍 + Simplicity 💡&lt;/p&gt;

</description>
      <category>python</category>
      <category>programming</category>
      <category>ai</category>
      <category>spark</category>
    </item>
    <item>
      <title>🚀 Day 30 of My Python &amp; Data Journey</title>
      <dc:creator>Aditi Sharma</dc:creator>
      <pubDate>Tue, 30 Sep 2025 14:50:20 +0000</pubDate>
      <link>https://dev.to/_adii3107/day-30-of-my-python-data-journey-4cn2</link>
      <guid>https://dev.to/_adii3107/day-30-of-my-python-data-journey-4cn2</guid>
      <description>&lt;p&gt;Big Data – Hidden Facts You Probably Didn’t Know 👀&lt;/p&gt;

&lt;p&gt;We hear “Big Data” everywhere… but let’s uncover some hidden gems:&lt;/p&gt;

&lt;p&gt;🔹 Dark Data – Nearly 90% of data is never analyzed. Companies store it, but never use it.&lt;br&gt;
🔹 Data = Power – Walmart handles 2.5 petabytes/hour from customer transactions 🛒.&lt;br&gt;
🔹 IoT Explosion – By 2030, 500 billion devices will be connected, generating endless streams of data.&lt;br&gt;
🔹 Speed Matters – Google processes over 3.5 billion searches/day in real-time ⚡.&lt;br&gt;
🔹 AI + Big Data – 80% of AI development is just cleaning and preparing data.&lt;/p&gt;

&lt;p&gt;💡 Fun Fact: The term “Big Data” was first coined in the 1990s but only became mainstream after 2010.&lt;/p&gt;

&lt;p&gt;Big Data isn’t just about size—it’s about unlocking value from the chaos.&lt;/p&gt;

</description>
      <category>programming</category>
      <category>python</category>
      <category>data</category>
      <category>database</category>
    </item>
    <item>
      <title>🚀 Day 29 of My Python Learning Journey</title>
      <dc:creator>Aditi Sharma</dc:creator>
      <pubDate>Mon, 29 Sep 2025 12:52:53 +0000</pubDate>
      <link>https://dev.to/_adii3107/day-29-of-my-python-learning-journey-18e5</link>
      <guid>https://dev.to/_adii3107/day-29-of-my-python-learning-journey-18e5</guid>
      <description>&lt;p&gt;Big Data &amp;amp; Its Uses 🌍💾&lt;/p&gt;

&lt;p&gt;Today I explored Big Data — the massive, complex datasets that can’t be handled by traditional systems.&lt;/p&gt;

&lt;p&gt;🔹 Uses of Big Data&lt;br&gt;
    • Healthcare 🏥 → disease prediction, patient monitoring&lt;br&gt;
    • Finance 💰 → fraud detection, algorithmic trading&lt;br&gt;
    • E-commerce 🛒 → personalized recommendations&lt;br&gt;
    • Social Media 📱 → sentiment analysis, trend prediction&lt;br&gt;
    • Smart Cities 🌆 → traffic management, energy optimization&lt;/p&gt;

&lt;p&gt;🔹 Interesting Facts&lt;br&gt;
    1.  Over 328 million terabytes of data are created daily 🌐.&lt;br&gt;
    2.  Big Data is often described by 5Vs → Volume, Velocity, Variety, Veracity, Value.&lt;br&gt;
    3.  Netflix saves $1B annually using Big Data for recommendations 🎬.&lt;br&gt;
    4.  By 2025, the world’s data will reach 175 zettabytes 😲.&lt;/p&gt;

&lt;p&gt;✨ Big Data isn’t just “big” — it’s the backbone of modern AI, analytics &amp;amp; decision-making.&lt;/p&gt;

&lt;h1&gt;
  
  
  BigData #DataAnalytics #100DaysOfCode #Python #DataScience
&lt;/h1&gt;

</description>
      <category>programming</category>
      <category>bigdata</category>
      <category>datascience</category>
      <category>python</category>
    </item>
    <item>
      <title>🚀 Day 28 of My Python Learning Journey</title>
      <dc:creator>Aditi Sharma</dc:creator>
      <pubDate>Sun, 28 Sep 2025 12:56:54 +0000</pubDate>
      <link>https://dev.to/_adii3107/day-28-of-my-python-learning-journey-4kki</link>
      <guid>https://dev.to/_adii3107/day-28-of-my-python-learning-journey-4kki</guid>
      <description>&lt;p&gt;Fun Facts about Advanced SQL 🗄️✨&lt;/p&gt;

&lt;p&gt;After diving into Advanced SQL, I found some cool facts worth sharing:&lt;/p&gt;

&lt;p&gt;🔹 Fun Facts:&lt;br&gt;
    1.  Window Functions are like magic — they let you calculate running totals, ranks, and moving averages without subqueries!&lt;br&gt;
    2.  CTEs (Common Table Expressions) improve readability, but they also make debugging complex queries so much easier.&lt;br&gt;
    3.  SQL’s CASE statements let you add “if-else” logic inside queries — turning SQL into a mini programming language.&lt;br&gt;
    4.  You can combine multiple window functions in the same query for powerful analytics (ranking + lag/lead + moving averages together).&lt;br&gt;
    5.  Despite being 50+ years old, SQL evolves — modern databases (like Snowflake, BigQuery, PostgreSQL) keep adding new advanced functions.&lt;/p&gt;

&lt;p&gt;⚡ Fun Fact of the Day: The first version of SQL was called SEQUEL (Structured English Query Language) — later shortened to SQL.&lt;/p&gt;

&lt;h1&gt;
  
  
  SQL #AdvancedSQL #DataAnalytics #100DaysOfCode #Python
&lt;/h1&gt;

</description>
      <category>productivity</category>
      <category>python</category>
      <category>sql</category>
      <category>database</category>
    </item>
  </channel>
</rss>
