<?xml version="1.0" encoding="UTF-8"?>
<rss version="2.0" xmlns:atom="http://www.w3.org/2005/Atom" xmlns:dc="http://purl.org/dc/elements/1.1/">
  <channel>
    <title>DEV Community: Phatsawut Duangkaew</title>
    <description>The latest articles on DEV Community by Phatsawut Duangkaew (@phatsawut_duangkaew_f92ff).</description>
    <link>https://dev.to/phatsawut_duangkaew_f92ff</link>
    <image>
      <url>https://media2.dev.to/dynamic/image/width=90,height=90,fit=cover,gravity=auto,format=auto/https:%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Fuser%2Fprofile_image%2F3293514%2F302d5014-a74f-4ea5-82a3-e0c52c50508d.png</url>
      <title>DEV Community: Phatsawut Duangkaew</title>
      <link>https://dev.to/phatsawut_duangkaew_f92ff</link>
    </image>
    <atom:link rel="self" type="application/rss+xml" href="https://dev.to/feed/phatsawut_duangkaew_f92ff"/>
    <language>en</language>
    <item>
      <title>5 Essential Stages of Data Management Before You Start AI Projects</title>
      <dc:creator>Phatsawut Duangkaew</dc:creator>
      <pubDate>Wed, 25 Jun 2025 16:00:02 +0000</pubDate>
      <link>https://dev.to/phatsawut_duangkaew_f92ff/5-essential-stages-of-data-management-before-you-start-ai-projects-4iik</link>
      <guid>https://dev.to/phatsawut_duangkaew_f92ff/5-essential-stages-of-data-management-before-you-start-ai-projects-4iik</guid>
      <description>&lt;p&gt;My visit to the PSU Cybersecurity &amp;amp; Data Privacy Days 2 proved to be incredibly insightful. I found myself in the audience for "Accelerate Modernize Applications with Nutanix AI Platform," a session led by Khun Surak Thammarak of Nutanix (Thailand) Ltd.&lt;/p&gt;

&lt;p&gt;As he presented a slide titled "One Platform to Simplify Data Management," it sparked a critical question in my mind. In the world of tech, we are always talking about revolutionary AI models and the magic of training them. But what about the journey of the data itself? We always hear about the final product, but what about the data pipeline behind it?&lt;/p&gt;

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2F766hk04kfw5hmuw1grwv.jpg" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2F766hk04kfw5hmuw1grwv.jpg" alt="Image description" width="430" height="354"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;That slide laid out a clear roadmap, and I realized that understanding this data lifecycle is the true starting point for anyone serious about AI.&lt;/p&gt;

&lt;h2&gt;
  
  
  Why Data Management Matters in AI
&lt;/h2&gt;

&lt;p&gt;An AI model is only as good as the data it learns from. You can have the most brilliant algorithm in the world, but if you feed it messy, disorganized, or inaccessible data, the results will be disappointing. Think of it like cooking: even a world-class chef can't make a great meal with poor-quality ingredients.&lt;/p&gt;

&lt;p&gt;Poor data management can ruin AI projects before they even begin. A solid data lifecycle isn't just a "nice-to-have"; it's the foundational backbone of any modern AI application.&lt;/p&gt;

&lt;h2&gt;
  
  
  The 5 Key Stages of an AI Data Pipeline
&lt;/h2&gt;

&lt;p&gt;The presentation slides elegantly broke down the complex data journey into five logical stages. Let's walk through each one.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;Ingest Raw Data&lt;/strong&gt;&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;&lt;p&gt;What it means: This is the starting line. It’s the process of collecting all the raw, unprocessed data from its original sources. This could be anything from sensor reading in a factory, user clicking on an app, to millions of text files.&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;Key term: The slide mentions to Geo-Distributed High Capacity that means the storage system must be able to collect massive amounts of data from different physical locations, all at the same time.&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;Why it's important: Your AI project needs a wide and deep pool of raw material. This stage ensures you have a scalable and robust way to gather it all in one place.&lt;/p&gt;&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;&lt;strong&gt;Prepare Data&lt;/strong&gt;&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;&lt;p&gt;What it means: Raw data is messy. This stage is all about cleaning, structuring, and transforming that data into a neat, consistent format that a machine can understand. It involves tasks like removing errors, label information, and organizing it.&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;Key term: The slide points to Low Latency Block storage. "Low latency" means fast access. The system needs to be quick so that data scientists can experiment and prepare data without long delays. "Unified" means it can handle different data types (like files and database blocks) in one place.&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;Why it's important: This is arguably the most critical step. High-quality, well-prepared data leads to much more accurate and effective AI models.&lt;/p&gt;&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;&lt;strong&gt;Tune or Train Model&lt;/strong&gt;&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;&lt;p&gt;What it means: This is the part we hear about most often. It’s where data scientists feed the prepared data into their AI algorithms, allowing the model to learn and find patterns. This process is computationally intensive and requires reading the data over and over again.&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;Key term: The slide highlights Parallel Access (Cloud). This means multiple computers can access and process the data simultaneously ("in parallel"), extremely speeding up the training time. This needs to work whether the computers are in a local data center or in the cloud.&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;Why it's important: Faster training allows for more experimentation and quicker development cycles. Strong, parallel data access is the fuel that powers the heavy engine of model training.&lt;/p&gt;&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;&lt;strong&gt;Run AI Inferencing&lt;/strong&gt;&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;&lt;p&gt;What it means: Once a model is trained, "inferencing" is the act of putting it to work in the real world to make predictions. This could be a recommendation engine on a website or a facial recognition system on a security camera.&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;Key term: Run at Edge, Fast Reads. "The Edge" refers to a location closer to where the data is generated, like a retail store or a factory floor, rather than a central data center. For real-time results, the model needs to read data and make a decision instantly ("fast reads").&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;Why it's important: For AI to be useful, it often needs to provide answers immediately. Placing the model at the edge reduces lag and allows for real-time decision-making.&lt;/p&gt;&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;&lt;strong&gt;Archive Data&lt;/strong&gt;&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;&lt;p&gt;What it means: What happens to the data after it’s been used? You can't just delete it. Archiving is the process of moving older, less frequently accessed data to a cheaper storage tier for long-term retention.&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;Key term: Dense, Low-Cost, Low-Performance storage. Since you don't need to access this data quickly, you can store it on slower, less expensive hardware, which saves a lot of money. "Dense" means you can pack a lot of data into a small space.&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;Why it's important: This practice is crucial for both cost optimization and legal observance, as many industries require data to be kept for several years.&lt;/p&gt;&lt;/li&gt;
&lt;/ul&gt;

&lt;h2&gt;
  
  
  The Role of a Unified Platform
&lt;/h2&gt;

&lt;p&gt;The presentation logically tied these stages together with the underlying Nutanix Unified Storage platform. The idea is to have one system that can manage data across this entire lifecycle, providing crucial capabilities like:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;&lt;p&gt;Data Mobility: Easily move data between stages or locations.&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;Data Access Anywhere: Allow teams to access the data they need, wherever they are.&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;Data Scaling and Agility: Grow your storage and performance as your AI needs evolve.&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;Security &amp;amp; Governance: Control who can access the data and track what happens to it.&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;Data Classification: Automatically identify what kind of data you have (e.g., sensitive personal info).&lt;/p&gt;&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;The slide also mentioned Data Lens, a tool that provides visibility and control over this entire data landscape, helping to protect against security risks and manage data effectively.&lt;/p&gt;

&lt;h2&gt;
  
  
  Summary: Why Developers Should Care
&lt;/h2&gt;

&lt;p&gt;As a developer, it's easy to think of data as something that just "exists." But understanding this pipeline is essential for building successful, real-world AI applications.&lt;/p&gt;

&lt;p&gt;AI isn’t just about writing Python code or using a machine learning library. It's a complete system that is heavily reliant on a well-oiled data machine. Understanding this helps you:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;Collaborate better with data engineers and data scientists.&lt;/li&gt;
&lt;li&gt;Build scalable projects that won't break when data volumes explode.&lt;/li&gt;
&lt;li&gt;Prepare for real-world deployments where data security, speed, and cost truly matter.&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;The next time you start an AI project, remember the five stages. Building a solid data foundation isn't the most glamorous part of AI, but it is the most important.&lt;/p&gt;

</description>
    </item>
    <item>
      <title>From DevOps to DevSecOps</title>
      <dc:creator>Phatsawut Duangkaew</dc:creator>
      <pubDate>Wed, 25 Jun 2025 15:07:43 +0000</pubDate>
      <link>https://dev.to/phatsawut_duangkaew_f92ff/from-devops-to-devsecops-3nj5</link>
      <guid>https://dev.to/phatsawut_duangkaew_f92ff/from-devops-to-devsecops-3nj5</guid>
      <description>&lt;p&gt;Recently, I had the opportunity to attend the PSU Cybersecurity and Data Privacy Days 2. One of the sessions that caught my attention was "Accelerate Modernize Applications with Nutanix AI Platform," presented by Khun Surak Thammarak from Nutanix (Thailand) Ltd.&lt;/p&gt;

&lt;p&gt;During the talk, Khun Surak presented a slide detailing a "Software Factory - DevSecOps with GitOps”. A thought immediately came up to my mind. We often hear about "DevOps," but what was this "DevSecOps"? What did the "Sec" part add?&lt;/p&gt;

&lt;p&gt;That single question sent me on a learning journey. I decided to dive deeper into the topic, and I'm writing this blog to share what I discovered and explain, in simple terms, what DevSecOps is all about.&lt;/p&gt;

&lt;h2&gt;
  
  
  What is DevOps?
&lt;/h2&gt;

&lt;p&gt;Before we talk about DevSecOps, let's quickly talk about DevOps. For years, the team that writes the code (Developers) and the team that manages the running software (Operations) worked in separate parts. This often led to slow processes &lt;/p&gt;

&lt;p&gt;DevOps changed that. It’s a culture and a set of practices that bring these two teams together. The goal is simple: to shorten the lifecycle of development and deliver high-quality software faster and more reliably. Think of it as turning a clumsy, multi-stage process into a single, smooth, automated assembly line.&lt;/p&gt;

&lt;h2&gt;
  
  
  Building Better, Faster, and Safer: A Guide to DevSecOps
&lt;/h2&gt;

&lt;p&gt;Now, let's get back to the topic. DevSecOps takes the great ideas of DevOps —speed and automation— and adds a crucial ingredient : Security.&lt;/p&gt;

&lt;p&gt;Instead of having a security check at the very end of the process (which is slow and expensive), DevSecOps integrates security into every single step. Think of it as a smart, automated factory for creating software where security guards are present on the entire assembly line, not just at the final gate.&lt;/p&gt;

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2F87f30loqfis48dnhn5u1.jpg" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2F87f30loqfis48dnhn5u1.jpg" alt="Image description" width="491" height="371"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;Let's walk through the factory map from the presentation.&lt;/p&gt;

&lt;h2&gt;
  
  
  The Two Main Characters: The Developer and the IT Operator
&lt;/h2&gt;

&lt;p&gt;Our story has two key players:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;&lt;p&gt;The Developer: Their job is to write the code that creates the features you use in an app.&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;The IT Operator: Their job is to ensure the app runs smoothly for everyone to use (that is what we call "production").&lt;/p&gt;&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;DevSecOps uses an automated process to connect their work smoothly.&lt;/p&gt;

&lt;h2&gt;
  
  
  Part 1: The "Building" Phase (Continuous Integration)
&lt;/h2&gt;

&lt;p&gt;As soon as a developer writes new code and saves it (an action called a "commit"), an automated process starts running step-by-step behind the scenes.&lt;/p&gt;

&lt;ol&gt;
&lt;li&gt;&lt;p&gt;Code Check-up (By SonarQube): Imagine an expert reviewer who instantly scans the new code. This tool automatically checks bugs, security vulnerabilities, and quality issues. This is our first security checkpoint, ensuring problems are caught up early.&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;Building and Packaging (App Build, Image Build): The code is then "built" and packaged into a secure, ready-to-go container. Think of it, like putting all the ingredients for a meal into a sealed box, complete with instructions. This container has everything the application needs to run.&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;Storing the Box (By Harbor): This packaged container is stored in a secure warehouse, called a repository. It’s now an official software version that’s ready for the next step.&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;Final Security Scan: Before it can be sent out, the packaged container gets one more security scan to make sure nothing dangerous was packed inside.&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;Keeping an Eye on Things (By Grafana): Throughout this entire process, a monitoring tool acts like an inspector. If something fails, it immediately alerts the team so they can fix it.&lt;/p&gt;&lt;/li&gt;
&lt;/ol&gt;

&lt;p&gt;This whole automated building and testing process is called Continuous Integration (CI).&lt;/p&gt;

&lt;h2&gt;
  
  
  Part 2: The "Release" Phase (Continuous Deployment with GitOps)
&lt;/h2&gt;

&lt;p&gt;Once our software package is built and approved, it’s time to deliver it to the users. This part of the process uses a method called GitOps, where all automation is managed through a central code repository.&lt;/p&gt;

&lt;ol&gt;
&lt;li&gt;&lt;p&gt;The Operator's Command (Commit): The IT Operator makes a simple change in a configuration file — like updating the version number from 1.0 to 1.1.&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;The Automation Watcher (By Flux): A special tool (Flux) is always watching this configuration file. As soon as it sees the change, it knows it's time to act.&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;Automatic Deployment (Deploy Kubernetes): The tool automatically takes the new, approved software package and deploys it to the live environment using a system called Kubernetes, which manages running applications.&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;Final Check (E2E Tests): Even after release, automated tests run to make sure everything is working as expected from a user's perspective.&lt;/p&gt;&lt;/li&gt;
&lt;/ol&gt;

&lt;p&gt;This automated release process is called Continuous Deployment (CD).&lt;/p&gt;

&lt;h2&gt;
  
  
  Why Does This All Matter?
&lt;/h2&gt;

&lt;p&gt;This DevSecOps approach, as shown in the presentation, is a game-changer:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;&lt;p&gt;Speed: New ideas get to users in hours or days, not months.&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;Security: With security checks built into every step (the "Sec" in &lt;br&gt;
DevSecOps), applications are safer from the start, not as an afterthought.&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;Reliability: Automation reduces the chance of human error, meaning &lt;br&gt;
fewer bugs and less downtime for users.&lt;/p&gt;&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;My curiosity about a single word on a slide led me to understand a whole new philosophy for building software. It’s not just about being fast; it’s about being fast and safe.&lt;/p&gt;

</description>
      <category>devops</category>
      <category>developer</category>
      <category>webdev</category>
      <category>programming</category>
    </item>
  </channel>
</rss>
