<?xml version="1.0" encoding="UTF-8"?>
<rss version="2.0" xmlns:atom="http://www.w3.org/2005/Atom" xmlns:dc="http://purl.org/dc/elements/1.1/">
  <channel>
    <title>DEV Community: kabeer1choudary</title>
    <description>The latest articles on DEV Community by kabeer1choudary (@kabeer1choudary).</description>
    <link>https://dev.to/kabeer1choudary</link>
    <image>
      <url>https://media2.dev.to/dynamic/image/width=90,height=90,fit=cover,gravity=auto,format=auto/https:%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Fuser%2Fprofile_image%2F518800%2Fcb5da37a-c726-478c-830e-7f5f7444ba00.jpeg</url>
      <title>DEV Community: kabeer1choudary</title>
      <link>https://dev.to/kabeer1choudary</link>
    </image>
    <atom:link rel="self" type="application/rss+xml" href="https://dev.to/feed/kabeer1choudary"/>
    <language>en</language>
    <item>
      <title>The Ultimate Guide: Installing Ollama on Fedora 43</title>
      <dc:creator>kabeer1choudary</dc:creator>
      <pubDate>Tue, 31 Mar 2026 03:28:30 +0000</pubDate>
      <link>https://dev.to/kabeer1choudary/the-ultimate-guide-installing-ollama-on-fedora-43-4c8i</link>
      <guid>https://dev.to/kabeer1choudary/the-ultimate-guide-installing-ollama-on-fedora-43-4c8i</guid>
      <description>&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2F5junmuiuu7tu9bbxchzu.jpg" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2F5junmuiuu7tu9bbxchzu.jpg" alt=" " width="800" height="450"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;Running large language models (LLMs) locally isn’t just for the privacy-obsessed anymore—it’s for anyone who wants a snappy, custom coding assistant without a monthly subscription. If you’re rocking Fedora 43, you’re already using one of the most cutting-edge distros out there.&lt;/p&gt;

&lt;p&gt;Here is how to get Ollama up and running with full NVIDIA acceleration and hook it into VS Code for a seamless dev experience.&lt;/p&gt;

&lt;p&gt;There’s something uniquely satisfying about seeing your GPU fans spin up because your local AI is thinking. Let’s get you there in eight steps.&lt;/p&gt;

&lt;h2&gt;
  
  
  Step 1: Open the Gates (RPM Fusion)
&lt;/h2&gt;

&lt;p&gt;Fedora is known for its commitment to free, open-source software, which means the proprietary NVIDIA drivers aren't there by default. We need to add the RPM Fusion repositories to get the "non-free" goodies.&lt;/p&gt;

&lt;p&gt;Run this in your terminal:&lt;br&gt;
&lt;/p&gt;

&lt;div class="highlight js-code-highlight"&gt;
&lt;pre class="highlight shell"&gt;&lt;code&gt;&lt;span class="nv"&gt;$ &lt;/span&gt;&lt;span class="nb"&gt;sudo &lt;/span&gt;dnf5 &lt;span class="nb"&gt;install &lt;/span&gt;https://mirrors.rpmfusion.org/free/fedora/rpmfusion-free-release-&lt;span class="si"&gt;$(&lt;/span&gt;rpm &lt;span class="nt"&gt;-E&lt;/span&gt; %fedora&lt;span class="si"&gt;)&lt;/span&gt;.noarch.rpm 

&lt;span class="nv"&gt;$ &lt;/span&gt;&lt;span class="nb"&gt;sudo &lt;/span&gt;dnf5 &lt;span class="nb"&gt;install &lt;/span&gt;https://mirrors.rpmfusion.org/nonfree/fedora/rpmfusion-nonfree-release-&lt;span class="si"&gt;$(&lt;/span&gt;rpm &lt;span class="nt"&gt;-E&lt;/span&gt; %fedora&lt;span class="si"&gt;)&lt;/span&gt;.noarch.rpm
&lt;/code&gt;&lt;/pre&gt;

&lt;/div&gt;



&lt;p&gt;Note: We’re using dnf5 here, which is the faster, shinier version of the package manager standard in Fedora 43.&lt;/p&gt;

&lt;h2&gt;
  
  
  Step 2: The Driver Dance
&lt;/h2&gt;

&lt;p&gt;Now, we install the NVIDIA drivers and the CUDA toolkit. This is what allows Ollama to talk to your GPU instead of making your CPU do all the heavy lifting.&lt;br&gt;
&lt;/p&gt;

&lt;div class="highlight js-code-highlight"&gt;
&lt;pre class="highlight shell"&gt;&lt;code&gt;&lt;span class="nb"&gt;sudo &lt;/span&gt;dnf5 &lt;span class="nb"&gt;install &lt;/span&gt;akmod-nvidia xorg-x11-drv-nvidia-cuda
&lt;/code&gt;&lt;/pre&gt;

&lt;/div&gt;



&lt;p&gt;A word of caution: Fedora moves fast. Sometimes, when the kernel updates, the NVIDIA modules need a moment to "catch up" (rebuild). If you update your system and things look wonky, it’s usually because the driver is still compiling in the background.&lt;/p&gt;

&lt;h2&gt;
  
  
  Step 3: The Classic Reboot
&lt;/h2&gt;

&lt;p&gt;You know the drill. For the kernel to start using those new NVIDIA drivers, you need a fresh start. sudo reboot&lt;/p&gt;

&lt;h2&gt;
  
  
  Step 4: The Moment of Truth
&lt;/h2&gt;

&lt;p&gt;Once you’re back in, let’s make sure Fedora and your GPU are on speaking terms. Run this command&lt;br&gt;
&lt;/p&gt;

&lt;div class="highlight js-code-highlight"&gt;
&lt;pre class="highlight shell"&gt;&lt;code&gt;nvidia-smi
&lt;/code&gt;&lt;/pre&gt;

&lt;/div&gt;



&lt;p&gt;If you see a table showing your GPU name and VRAM usage, congratulations—you’ve passed the hardest part.&lt;/p&gt;

&lt;h2&gt;
  
  
  Step 5: Installing Ollama
&lt;/h2&gt;

&lt;p&gt;Ollama makes installation incredibly easy with a one-liner script. This script handles the heavy lifting: it downloads the binary, creates an ollama user/group, and sets up a systemd service so it starts automatically.&lt;br&gt;
&lt;/p&gt;

&lt;div class="highlight js-code-highlight"&gt;
&lt;pre class="highlight shell"&gt;&lt;code&gt;curl &lt;span class="nt"&gt;-fsSL&lt;/span&gt; https://ollama.com/install.sh | sh
&lt;/code&gt;&lt;/pre&gt;

&lt;/div&gt;



&lt;h2&gt;
  
  
  Step 6: Verifying the GPU Link
&lt;/h2&gt;

&lt;p&gt;Just because Ollama is installed doesn't mean it’s using your GPU. It might be falling back to your CPU if it can't find the drivers. Let's check the logs:&lt;br&gt;
&lt;/p&gt;

&lt;div class="highlight js-code-highlight"&gt;
&lt;pre class="highlight shell"&gt;&lt;code&gt;journalctl &lt;span class="nt"&gt;-u&lt;/span&gt; ollama &lt;span class="nt"&gt;-b&lt;/span&gt; | &lt;span class="nb"&gt;grep&lt;/span&gt; &lt;span class="s2"&gt;"NVIDIA"&lt;/span&gt;
&lt;/code&gt;&lt;/pre&gt;

&lt;/div&gt;



&lt;p&gt;You’re looking for a line that confirms an NVIDIA GPU was detected. If you see it, you're golden.&lt;/p&gt;

&lt;h2&gt;
  
  
  Step 7: Your First Local Run
&lt;/h2&gt;

&lt;p&gt;Let's test it with a lightweight, high-performance model. Qwen 2.5 Coder (0.5B version) is tiny but surprisingly "smart" for its size, making it perfect for a quick test.&lt;br&gt;
&lt;/p&gt;

&lt;div class="highlight js-code-highlight"&gt;
&lt;pre class="highlight shell"&gt;&lt;code&gt;ollama run qwen2.5-coder:0.5b
&lt;/code&gt;&lt;/pre&gt;

&lt;/div&gt;



&lt;p&gt;Once it downloads, you can chat with it directly in your terminal. Ask it to write a Python script; you'll be impressed by the speed.&lt;/p&gt;

&lt;h2&gt;
  
  
  Integrating with VS Code (The "Continue" Extension)
&lt;/h2&gt;

&lt;p&gt;Running AI in a terminal is cool, but having it inside your IDE is where the real productivity happens. We’ll use the Continue extension, which is an open-source powerhouse for local AI.&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;Install the Extension: Search for "Continue" in the VS Code Marketplace and install it.&lt;/li&gt;
&lt;li&gt;Configure Local Access: Open the Continue sidebar, click the gear icon (settings), and select your local config file.&lt;/li&gt;
&lt;li&gt;Edit the Config: Replace or add the following to your config.yaml (or config.json depending on your version) to point it toward your local Ollama instance:
&lt;/li&gt;
&lt;/ul&gt;

&lt;div class="highlight js-code-highlight"&gt;
&lt;pre class="highlight yaml"&gt;&lt;code&gt;&lt;span class="na"&gt;name&lt;/span&gt;&lt;span class="pi"&gt;:&lt;/span&gt; &lt;span class="s"&gt;Local Config&lt;/span&gt;
&lt;span class="na"&gt;version&lt;/span&gt;&lt;span class="pi"&gt;:&lt;/span&gt; &lt;span class="s"&gt;1.0.0&lt;/span&gt;
&lt;span class="na"&gt;schema&lt;/span&gt;&lt;span class="pi"&gt;:&lt;/span&gt; &lt;span class="s"&gt;v1&lt;/span&gt;
&lt;span class="na"&gt;models&lt;/span&gt;&lt;span class="pi"&gt;:&lt;/span&gt;
  &lt;span class="pi"&gt;-&lt;/span&gt; &lt;span class="na"&gt;name&lt;/span&gt;&lt;span class="pi"&gt;:&lt;/span&gt; &lt;span class="s"&gt;Qwen2.5-Coder 0.5B&lt;/span&gt;
    &lt;span class="na"&gt;provider&lt;/span&gt;&lt;span class="pi"&gt;:&lt;/span&gt; &lt;span class="s"&gt;ollama&lt;/span&gt;
    &lt;span class="na"&gt;model&lt;/span&gt;&lt;span class="pi"&gt;:&lt;/span&gt; &lt;span class="s"&gt;qwen2.5-coder:0.5b&lt;/span&gt;
    &lt;span class="na"&gt;apiBase&lt;/span&gt;&lt;span class="pi"&gt;:&lt;/span&gt; &lt;span class="s"&gt;http://localhost:11434&lt;/span&gt;
    &lt;span class="na"&gt;roles&lt;/span&gt;&lt;span class="pi"&gt;:&lt;/span&gt;
      &lt;span class="pi"&gt;-&lt;/span&gt; &lt;span class="s"&gt;chat&lt;/span&gt;
      &lt;span class="pi"&gt;-&lt;/span&gt; &lt;span class="s"&gt;edit&lt;/span&gt;
      &lt;span class="pi"&gt;-&lt;/span&gt; &lt;span class="s"&gt;apply&lt;/span&gt;
      &lt;span class="pi"&gt;-&lt;/span&gt; &lt;span class="s"&gt;autocomplete&lt;/span&gt;
      &lt;span class="pi"&gt;-&lt;/span&gt; &lt;span class="s"&gt;embed&lt;/span&gt;
&lt;/code&gt;&lt;/pre&gt;

&lt;/div&gt;



&lt;h2&gt;
  
  
  Final Thoughts
&lt;/h2&gt;

&lt;p&gt;You now have a fully private, incredibly fast AI coding assistant running on your local hardware. No data leaving your machine, no latency issues, and total control.&lt;/p&gt;

</description>
      <category>ai</category>
      <category>automation</category>
      <category>llm</category>
      <category>genai</category>
    </item>
    <item>
      <title>Smooth Sailing: A Hands-On Guide to Set-Up Concourse CI on Rocky Linux</title>
      <dc:creator>kabeer1choudary</dc:creator>
      <pubDate>Sun, 08 Sep 2024 07:49:00 +0000</pubDate>
      <link>https://dev.to/kabeer1choudary/smooth-sailing-a-hands-on-guide-to-set-up-concourse-ci-on-rocky-linux-db6</link>
      <guid>https://dev.to/kabeer1choudary/smooth-sailing-a-hands-on-guide-to-set-up-concourse-ci-on-rocky-linux-db6</guid>
      <description>&lt;p&gt;As you know, my everyday driver is a Windows 10 PC, and I usually can't go for another system which is as smooth and user-friendly as it gets. But I have few unpleasant quirks with my daily driver, which does not allow me to experiment with opensource software packages and builds, which usually oriented towards Linux gear. Previously my dependency was with Vagrant, cause its easily available boxes, configuration management abilities and cli mode approach. But from recent days, I started to feel a bit distanced with the Vagrant tool, that became a bit sloth on its own. So, we got an alternate tool to handle my VM jobs. Introducing VMWare Workstation 16 Player (works well with Windows Hyper-V platform) which comes with GUI features. Compared to Vagrant, VMWare Workstation is not a CLI only tool and has more towards UI approach. So, in this blogs lets dive into the steps, I took to setup Concourse CI on top of Rocky Linux for my Devops CI/CD lab experiments.&lt;/p&gt;

&lt;h2&gt;
  
  
  Prerequisite:
&lt;/h2&gt;

&lt;ul&gt;
&lt;li&gt;Get the VMWare Workstation 16 Player from GetintoPC&lt;/li&gt;
&lt;li&gt;Download your Rocky Linux ISO from Rocky Linux site (I used Rocky Linux 8 minimal iso)&lt;/li&gt;
&lt;/ul&gt;

&lt;h2&gt;
  
  
  Installing Rocky Linux using VMWare Workstation:
&lt;/h2&gt;

&lt;p&gt;At first, lets create an empty VM in our VMWare Workstation tool with required configurations, follow along the snips&lt;/p&gt;

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fclwk87x8aag4bdtrdbgr.PNG" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fclwk87x8aag4bdtrdbgr.PNG" alt="1" width="717" height="588"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2F7iqegid3e3w5d9dxyced.PNG" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2F7iqegid3e3w5d9dxyced.PNG" alt="2" width="715" height="588"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2F6b6170fd5a1vdmnwcepq.PNG" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2F6b6170fd5a1vdmnwcepq.PNG" alt="3" width="713" height="588"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fpz1kfwute50wdggn6rl1.PNG" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fpz1kfwute50wdggn6rl1.PNG" alt="4" width="712" height="586"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fukggfaaf10d2pw8ype0g.PNG" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fukggfaaf10d2pw8ype0g.PNG" alt="5" width="714" height="589"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fb6gq39cj0qyja7zs1qkt.PNG" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fb6gq39cj0qyja7zs1qkt.PNG" alt="6" width="716" height="584"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;Since, VMWare Workstation would recommend us to use single core CPU, 2GB of RAM &amp;amp; 20GB of Disk space for RedHat Linux, we need a minimum of 2 core CPU and 4GB of RAM to keep our Concourse running good. Will modify our configuration as dual core CPU, 8GB of RAM and 20GB of disk space. Also, add your downloaded Rocky Linux ISO to the CD/DVD section as a source of OS installation.&lt;/p&gt;

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fektszhbvlikni1d8a4zk.PNG" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fektszhbvlikni1d8a4zk.PNG" alt="7" width="752" height="712"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2F1v6plxbfhiq5yfzhnung.PNG" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2F1v6plxbfhiq5yfzhnung.PNG" alt="8" width="645" height="539"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2F6q8pm1lrxb0p68f106jf.PNG" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2F6q8pm1lrxb0p68f106jf.PNG" alt="9" width="800" height="449"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2F2rzjh04w9hjpirah0xui.PNG" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2F2rzjh04w9hjpirah0xui.PNG" alt="10" width="800" height="454"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;h2&gt;
  
  
  Configuring our VM:
&lt;/h2&gt;

&lt;p&gt;Post OS installation we change the hostname with hostnamectl command and get its ip too&lt;br&gt;
$ sudo hostnamectl sethostname rocky8.local&lt;br&gt;
$ ip a&lt;/p&gt;

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fr3g3cu3dxftwcwmblw7s.PNG" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fr3g3cu3dxftwcwmblw7s.PNG" alt="11" width="594" height="245"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fcsrzqniuv8ldeehnalrs.PNG" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fcsrzqniuv8ldeehnalrs.PNG" alt="12" width="800" height="259"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;Now, we can use our powershell to log into our Rocky Linux machine using ssh &amp;amp; perform a dnf update&lt;br&gt;
$ ssh @rocky8.local&lt;br&gt;
$ sudo dnf update -y&lt;/p&gt;

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fyj8htsnt97b4uf7donzi.PNG" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fyj8htsnt97b4uf7donzi.PNG" alt="13" width="800" height="208"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;Now, we install docker &amp;amp; docker compose, then start docker service &amp;amp; enable it&lt;/p&gt;

&lt;p&gt;$ sudo dnf config-manager --add--repo=&lt;a href="https://download.docker.com/linux/centos/docker-ce.repo" rel="noopener noreferrer"&gt;https://download.docker.com/linux/centos/docker-ce.repo&lt;/a&gt;&lt;br&gt;
$ sudo dnf install docker-ce -y&lt;br&gt;
$ sudo systemctl start docker&lt;br&gt;
$ sudo systemctl enable docker&lt;/p&gt;

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fe2aiwxbm9sezfmw7kysx.PNG" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fe2aiwxbm9sezfmw7kysx.PNG" alt="14" width="800" height="185"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fizl423jv4dpa3jqm8fjx.PNG" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fizl423jv4dpa3jqm8fjx.PNG" alt="15" width="800" height="297"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;h2&gt;
  
  
  Getting started with Concourse:
&lt;/h2&gt;

&lt;p&gt;Lets download the docker compose file to deploy our Concourse CI, and install vim to edit our docker compose file&lt;/p&gt;

&lt;p&gt;$ sudo curl -LO &lt;a href="https://concourse-ci.org/docker-compose.yml" rel="noopener noreferrer"&gt;https://concourse-ci.org/docker-compose.yml&lt;/a&gt;&lt;br&gt;
$ sudo dnf install vim -y&lt;/p&gt;

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2F1yqsh7d4zuq5zlxhqx3s.PNG" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2F1yqsh7d4zuq5zlxhqx3s.PNG" alt="16" width="678" height="146"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fq1o8st2hmv9ypd2hagvb.PNG" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fq1o8st2hmv9ypd2hagvb.PNG" alt="17" width="800" height="185"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;We need to modify the docker compose file based on our requirement. The parameters we are modified below are CONCOURSE_EXTERNAL_URL, CONCOURSE_ADD_LOCAL_USER, CONCOURSE_MAIN_TEAM_LOCAL_USER, CONCOURSE_CLUSTER_NAME. rest we can keep as it is.&lt;/p&gt;

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fnpkne3kd8qibzn9y9cla.PNG" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fnpkne3kd8qibzn9y9cla.PNG" alt="18" width="657" height="748"&gt;&lt;/a&gt;&lt;br&gt;
Now, we deploy the Concourse CI cluster using docker compose&lt;br&gt;
$ sudo docker compose up -d&lt;/p&gt;

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fue2i0t6v8j7phli33hbu.PNG" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fue2i0t6v8j7phli33hbu.PNG" alt="19" width="800" height="533"&gt;&lt;/a&gt;&lt;br&gt;
There are 2 major parts in Concourse, the first is Concourse web server and the second is Postgres DB. We see that the web server is getting crashed due to some discrepancies with our setup configuration. Which could be seen using docker commands&lt;/p&gt;

&lt;p&gt;$ sudo docker ps -a&lt;br&gt;
$ sudo docker logs &lt;/p&gt;

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fwois84bbndnukzb49xo5.PNG" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fwois84bbndnukzb49xo5.PNG" alt="20" width="800" height="194"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fbkdptgqtq6u6ew0n3ras.PNG" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fbkdptgqtq6u6ew0n3ras.PNG" alt="21" width="800" height="126"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fb565ack5rwpu7ojwedka.PNG" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fb565ack5rwpu7ojwedka.PNG" alt="22" width="800" height="182"&gt;&lt;/a&gt;&lt;br&gt;
After doing some research over internet, i found below solution which could solve our web server crash in Concourse, Follow the below commands&lt;/p&gt;

&lt;p&gt;$ sudo modprobe iptable_filter&lt;br&gt;
$ sudo modprobe ip6table_filter&lt;/p&gt;

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fekpp8viyv3wz852c3ci4.PNG" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fekpp8viyv3wz852c3ci4.PNG" alt="23" width="657" height="136"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fbzyznztlxvdap0cxi9os.PNG" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fbzyznztlxvdap0cxi9os.PNG" alt="24" width="800" height="240"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Frvpbmkenvppv17y8elj5.PNG" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Frvpbmkenvppv17y8elj5.PNG" alt="25" width="800" height="112"&gt;&lt;/a&gt;&lt;br&gt;
Now, we can verify our Concourse CI setup using our we browser with &lt;a href="http://rocky8.local:8080" rel="noopener noreferrer"&gt;http://rocky8.local:8080&lt;/a&gt; URL, and try to login using admin credentials&lt;/p&gt;

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fdmkgah6bsujdzf30c61r.PNG" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fdmkgah6bsujdzf30c61r.PNG" alt="26" width="800" height="252"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fmp9l0a7gkjgc0z3qhfq8.PNG" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fmp9l0a7gkjgc0z3qhfq8.PNG" alt="27" width="800" height="345"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fg5fw4dwqwrzq7ukhu9ne.PNG" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fg5fw4dwqwrzq7ukhu9ne.PNG" alt="28" width="800" height="437"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;Thus concluding our own local Concourse CI setup. Let me know if you found anything different while trying the mentioned steps and we will try to figure out the solution to explore more further. Will keep you guys posted about the usages and tips with Concourse in near future. !&lt;/p&gt;

</description>
      <category>devops</category>
      <category>concourseci</category>
      <category>cicd</category>
      <category>rockylinux</category>
    </item>
    <item>
      <title>AWS Dynamo DB: A Beginner’s Guide</title>
      <dc:creator>kabeer1choudary</dc:creator>
      <pubDate>Sun, 08 Sep 2024 07:03:30 +0000</pubDate>
      <link>https://dev.to/kabeer1choudary/aws-dynamo-db-a-beginners-guide-5350</link>
      <guid>https://dev.to/kabeer1choudary/aws-dynamo-db-a-beginners-guide-5350</guid>
      <description>&lt;p&gt;Amazon DynamoDB is designed for applications that require low-latency data access, flexible data models, and seamless scalability. It provides fast and predictable performance, making it ideal for internet-scale applications. &lt;/p&gt;

&lt;h2&gt;
  
  
  Here are some key features:
&lt;/h2&gt;

&lt;h3&gt;
  
  
  Managed Service:
&lt;/h3&gt;

&lt;p&gt;DynamoDB is fully managed by AWS, which means you don’t need to worry about infrastructure provisioning, scaling, or maintenance.&lt;/p&gt;

&lt;h3&gt;
  
  
  NoSQL Database:
&lt;/h3&gt;

&lt;p&gt;It follows a NoSQL data model, allowing you to store and retrieve data without the constraints of a fixed schema.&lt;/p&gt;

&lt;h3&gt;
  
  
  Seamless Scalability:
&lt;/h3&gt;

&lt;p&gt;DynamoDB automatically scales to handle varying workloads and traffic spikes.&lt;/p&gt;

&lt;h3&gt;
  
  
  High Availability:
&lt;/h3&gt;

&lt;p&gt;Data is replicated across multiple Availability Zones (AZs) for durability and fault tolerance.&lt;/p&gt;

&lt;h3&gt;
  
  
  Flexible Data Models:
&lt;/h3&gt;

&lt;p&gt;You can choose between key-value and document data models.&lt;/p&gt;

&lt;h2&gt;
  
  
  Key Components of DynamoDB
&lt;/h2&gt;

&lt;ul&gt;
&lt;li&gt;
&lt;strong&gt;Tables:&lt;/strong&gt; The fundamental unit of storage in DynamoDB. Each table consists of items (records) with a primary key.&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;Primary Key:&lt;/strong&gt;

&lt;ul&gt;
&lt;li&gt;
&lt;em&gt;Partition Key (Hash Key):&lt;/em&gt; Uniquely identifies an item within a table.&lt;/li&gt;
&lt;li&gt;
&lt;em&gt;Composite Key (Partition Key + Sort Key):&lt;/em&gt; Allows range queries on the sort key.&lt;/li&gt;
&lt;/ul&gt;


&lt;/li&gt;

&lt;li&gt;

&lt;strong&gt;Secondary Indexes:&lt;/strong&gt; Enable efficient querying on non-primary key attributes.&lt;/li&gt;

&lt;li&gt;

&lt;strong&gt;Streams:&lt;/strong&gt; Captures a time-ordered sequence of item-level modifications.&lt;/li&gt;

&lt;li&gt;

&lt;strong&gt;Global Tables:&lt;/strong&gt; Replicate data across multiple AWS Regions for global availability.&lt;/li&gt;

&lt;/ul&gt;

&lt;h2&gt;
  
  
  Working Process
&lt;/h2&gt;

&lt;ul&gt;
&lt;li&gt;
&lt;strong&gt;Create a Table:&lt;/strong&gt; Define the schema (primary key) and provisioned capacity (or use on-demand mode).&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;Insert Data:&lt;/strong&gt; Use the PutItem API to add items to the table.&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;Query and Scan:&lt;/strong&gt; Query using the primary key or secondary indexes. Scan for full-table scans (use sparingly due to performance impact).&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;Update and Delete:&lt;/strong&gt; Modify existing items using UpdateItem. Delete items using DeleteItem.&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;DynamoDB Streams:&lt;/strong&gt; Capture changes (inserts, updates, deletes) in near real time. Process streams using AWS Lambda or Kinesis.&lt;/li&gt;
&lt;/ul&gt;

&lt;h2&gt;
  
  
  Use Cases
&lt;/h2&gt;

&lt;ul&gt;
&lt;li&gt;
&lt;strong&gt;Session Management:&lt;/strong&gt; Store user sessions, tokens, and preferences.&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;Gaming Leaderboards:&lt;/strong&gt; High write throughput for real-time leaderboards.&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;IoT Data Storage:&lt;/strong&gt; Handle sensor data, telemetry, and device state.&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;Ad Tech:&lt;/strong&gt; Clickstream data, user profiles, and targeting information.&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;Popular Item Caches:&lt;/strong&gt; Store frequently accessed data for low-latency retrieval.&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;Remember, DynamoDB is a powerful tool, but understanding its nuances and best practices is essential for successful implementation.&lt;/p&gt;

</description>
      <category>aws</category>
      <category>cloud</category>
      <category>beginners</category>
      <category>dynamodb</category>
    </item>
    <item>
      <title>AWS RDS: A Beginner’s Guide</title>
      <dc:creator>kabeer1choudary</dc:creator>
      <pubDate>Sun, 08 Sep 2024 06:51:51 +0000</pubDate>
      <link>https://dev.to/kabeer1choudary/aws-rds-a-beginners-guide-4c72</link>
      <guid>https://dev.to/kabeer1choudary/aws-rds-a-beginners-guide-4c72</guid>
      <description>&lt;p&gt;Amazon RDS (Relational Database Services) is a powerful web service that streamlines the setup, operation, and scalability of relational databases within the AWS Cloud. Whether you’re building web applications, SaaS platforms, or diving into business analytics, RDS provides a robust foundation for your data needs.&lt;/p&gt;

&lt;h2&gt;
  
  
  Key Components
&lt;/h2&gt;

&lt;h3&gt;
  
  
  Database Engines:
&lt;/h3&gt;

&lt;p&gt;RDS supports various database engines, including MySQL, PostgreSQL, Oracle, Microsoft SQL Server, and Amazon Aurora. Choose the engine that best fits your application requirements.&lt;/p&gt;

&lt;h3&gt;
  
  
  Automated Backups:
&lt;/h3&gt;

&lt;p&gt;RDS automatically backs up your database at specified intervals. You can restore to any point in time within your retention window.&lt;/p&gt;

&lt;h3&gt;
  
  
  Monitoring and Metrics:
&lt;/h3&gt;

&lt;p&gt;RDS provides performance insights through Amazon CloudWatch with cases like monitoring CPU utilization, data storage, and query performances.&lt;/p&gt;

&lt;h3&gt;
  
  
  Security and Encryption:
&lt;/h3&gt;

&lt;p&gt;RDS encrypts data at rest using keys managed by AWS Key Management Service (KMS).&lt;/p&gt;

&lt;h2&gt;
  
  
  How It Works
&lt;/h2&gt;

&lt;h3&gt;
  
  
  Database Creation:
&lt;/h3&gt;

&lt;p&gt;Create an RDS instance with your chosen database engine. Specify instance type, storage, and other configuration settings.&lt;/p&gt;

&lt;h3&gt;
  
  
  Data Storage:
&lt;/h3&gt;

&lt;p&gt;RDS manages storage volumes (EBS) for your database. Data is stored securely and redundantly across Availability Zones in AWS.&lt;/p&gt;

&lt;h3&gt;
  
  
  Scaling:
&lt;/h3&gt;

&lt;p&gt;Easily scale compute and storage resources as your workload grows which includes, Vertical scaling (instance type) or horizontal scaling (read replicas).&lt;/p&gt;

&lt;h2&gt;
  
  
  Use Cases
&lt;/h2&gt;

&lt;h3&gt;
  
  
  Web-Based Applications:
&lt;/h3&gt;

&lt;p&gt;Deploy WordPress, Drupal, or other content management systems. RDS handles database management, backups, and scaling.&lt;/p&gt;

&lt;h3&gt;
  
  
  SaaS Applications:
&lt;/h3&gt;

&lt;p&gt;SaaS platforms rely on RDS for their backend databases. Multi-tenant architecture? RDS has you covered.&lt;/p&gt;

&lt;h3&gt;
  
  
  Business Analytics:
&lt;/h3&gt;

&lt;p&gt;Store and analyze business data efficiently. Thus leverages RDS for reporting and insights.&lt;/p&gt;

&lt;h3&gt;
  
  
  Compliance-Driven Industries:
&lt;/h3&gt;

&lt;p&gt;Medical and banking applications benefit from RDS’s security features such as Encryption, access controls, and audit trails.&lt;/p&gt;

&lt;h2&gt;
  
  
  Conclusion
&lt;/h2&gt;

&lt;p&gt;Amazon RDS simplifies the complexities of relational databases, allowing you to focus on building great applications. Whether you’re a developer, data scientist, or business owner, RDS empowers you to manage your data effectively. RDS isn’t just about databases, it’s about unlocking the potential of your data.&lt;/p&gt;

</description>
      <category>aws</category>
      <category>cloud</category>
      <category>beginners</category>
      <category>rds</category>
    </item>
    <item>
      <title>AWS Route53: A Beginner’s Guide</title>
      <dc:creator>kabeer1choudary</dc:creator>
      <pubDate>Sat, 20 Jul 2024 03:11:39 +0000</pubDate>
      <link>https://dev.to/kabeer1choudary/aws-route53-a-beginners-guide-2ndl</link>
      <guid>https://dev.to/kabeer1choudary/aws-route53-a-beginners-guide-2ndl</guid>
      <description>&lt;p&gt;Amazon Route 53 is a highly available and scalable Domain Name System (DNS) service that plays a crucial role in connecting user requests to various services running in the AWS cloud. Whether you’re managing domain names, optimizing traffic flow, or ensuring high availability, Route 53 has got you covered.&lt;/p&gt;

&lt;h1&gt;
  
  
  Key Components
&lt;/h1&gt;

&lt;h2&gt;
  
  
  DNS Management:
&lt;/h2&gt;

&lt;p&gt;Route 53 allows you to register domain names and manage DNS records (like A, CNAME, and MX records). You can create hosted zones to organize your DNS records for different domains. Health checks monitor the availability of your resources (e.g., EC2 instances, S3 buckets) and automatically route traffic away from unhealthy endpoints.&lt;/p&gt;

&lt;h2&gt;
  
  
  Traffic Management:
&lt;/h2&gt;

&lt;h3&gt;
  
  
  Routing Policies:
&lt;/h3&gt;

&lt;ul&gt;
&lt;li&gt;
&lt;strong&gt;Simple Routing:&lt;/strong&gt; Directs traffic to a single resource (e.g., an EC2 instance).&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;Weighted Routing:&lt;/strong&gt; Distributes traffic based on assigned weights (useful for A/B testing).&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;Latency-Based Routing:&lt;/strong&gt; Routes users to the lowest-latency endpoint.&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;Geolocation Routing:&lt;/strong&gt; Routes traffic based on user location.&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;Failover Routing:&lt;/strong&gt; Automatically switches to a standby resource during failures.&lt;/li&gt;
&lt;/ul&gt;

&lt;h3&gt;
  
  
  Alias Records:
&lt;/h3&gt;

&lt;ul&gt;
&lt;li&gt;Alias records allow you to map your domain directly to AWS resources (e.g., an S3 bucket or CloudFront distribution).&lt;/li&gt;
&lt;/ul&gt;

&lt;h2&gt;
  
  
  Availability Monitoring:
&lt;/h2&gt;

&lt;p&gt;Route 53 health checks monitor the health of your resources.&lt;br&gt;
Failover policies automatically route traffic away from unhealthy endpoints.&lt;/p&gt;

&lt;h2&gt;
  
  
  Domain Registration:
&lt;/h2&gt;

&lt;p&gt;You can register new domain names directly through Route 53.&lt;/p&gt;

&lt;h1&gt;
  
  
  How It Works
&lt;/h1&gt;

&lt;ul&gt;
&lt;li&gt;
&lt;strong&gt;User Requests:&lt;/strong&gt; When a user enters a domain name (e.g., &lt;a href="http://www.example.com" rel="noopener noreferrer"&gt;www.example.com&lt;/a&gt;), their DNS resolver queries Route 53.&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;Route 53 Resolution:&lt;/strong&gt; Route 53 looks up the DNS records associated with the domain. It returns the IP address of the appropriate resource (e.g., an EC2 instance or an S3 bucket).&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;Traffic Routing:&lt;/strong&gt; Based on routing policies, Route 53 directs traffic to the correct resource. Health checks ensure that only healthy endpoints receive traffic.&lt;/li&gt;
&lt;/ul&gt;

&lt;h1&gt;
  
  
  Use Cases
&lt;/h1&gt;

&lt;h2&gt;
  
  
  Global Traffic Management:
&lt;/h2&gt;

&lt;ul&gt;
&lt;li&gt;Route 53’s global DNS features allow you to create complex routing relationships across regions.&lt;/li&gt;
&lt;li&gt;Distribute traffic intelligently based on latency, geolocation, or other factors.&lt;/li&gt;
&lt;/ul&gt;

&lt;h2&gt;
  
  
  Highly Available Applications:
&lt;/h2&gt;

&lt;ul&gt;
&lt;li&gt;Set up failover routing to automatically switch to backup resources during failures.&lt;/li&gt;
&lt;li&gt;Ensure seamless user experience even when parts of your infrastructure are down.&lt;/li&gt;
&lt;/ul&gt;

&lt;h2&gt;
  
  
  Private DNS in VPCs:
&lt;/h2&gt;

&lt;ul&gt;
&lt;li&gt;Assign custom domain names to resources within your Amazon VPC&lt;/li&gt;
&lt;li&gt;Helps to keep DNS data private within your network.&lt;/li&gt;
&lt;/ul&gt;

&lt;h1&gt;
  
  
  Conclusion
&lt;/h1&gt;

&lt;p&gt;Amazon Route 53 simplifies DNS management, optimizes traffic routing, and enhances application availability. Whether you’re a developer, a system administrator, or a business owner, understanding Route 53’s capabilities empowers you to build robust and reliable systems. Remember, Route 53 isn’t just about domain names, it’s about connecting users to your services seamlessly.&lt;/p&gt;

</description>
      <category>aws</category>
      <category>cloud</category>
      <category>beginners</category>
      <category>route53</category>
    </item>
    <item>
      <title>AWS VPC: A Beginner’s Guide</title>
      <dc:creator>kabeer1choudary</dc:creator>
      <pubDate>Thu, 21 Mar 2024 17:45:14 +0000</pubDate>
      <link>https://dev.to/kabeer1choudary/aws-vpc-a-beginners-guide-58ai</link>
      <guid>https://dev.to/kabeer1choudary/aws-vpc-a-beginners-guide-58ai</guid>
      <description>&lt;p&gt;Amazon Web Services (AWS) provides a powerful networking service called Amazon Virtual Private Cloud (VPC). In this blog post, we’ll delve into what VPC is, explore its functions, and provide practical examples to illustrate its capabilities.&lt;/p&gt;

&lt;h2&gt;
  
  
  What is Amazon VPC?
&lt;/h2&gt;

&lt;p&gt;Amazon VPC is a virtual network dedicated to your AWS account. It allows you to create isolated network environments within the AWS cloud. Here are the key features of VPC:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;Custom IP Address Range: You can define your own IP address range for your VPC.&lt;/li&gt;
&lt;li&gt;Subnets: Divide your VPC into subnets to organize resources and control network traffic.&lt;/li&gt;
&lt;li&gt;Routing: Configure route tables to direct traffic between subnets and to external networks.&lt;/li&gt;
&lt;li&gt;Security Groups: Set up security rules to control inbound and outbound traffic.&lt;/li&gt;
&lt;li&gt;Connectivity Options: VPCs can be connected to the internet, other VPCs, or on-premises networks.&lt;/li&gt;
&lt;/ul&gt;

&lt;h2&gt;
  
  
  Components of Amazon VPC
&lt;/h2&gt;

&lt;h3&gt;
  
  
  1. Subnet
&lt;/h3&gt;

&lt;ul&gt;
&lt;li&gt;Function: A subnet is a defined range of IP addresses within your VPC.&lt;/li&gt;
&lt;li&gt;Purpose:
Organize resources (such as Amazon EC2 instances) logically.
Control network traffic by segmenting your VPC.&lt;/li&gt;
&lt;li&gt;Types: Public Subnet: Exposes resources directly to the internet via an Internet Gateway. Private Subnet: Contains resources that should not be directly accessible from the internet.&lt;/li&gt;
&lt;/ul&gt;

&lt;h3&gt;
  
  
  2. Route Table
&lt;/h3&gt;

&lt;ul&gt;
&lt;li&gt;Function: Route tables determine how network traffic flows within your VPC.&lt;/li&gt;
&lt;li&gt;Usage:
Specify destinations (IP addresses) and their targets (e.g., Internet Gateway, NAT Gateway, or Virtual Private Gateway).
Control traffic between subnets and external networks.&lt;/li&gt;
&lt;/ul&gt;

&lt;h3&gt;
  
  
  3. Virtual Private Gateway (VGW)
&lt;/h3&gt;

&lt;ul&gt;
&lt;li&gt;Function: Serves as the VPN hub on the Amazon side of a VPN connection.&lt;/li&gt;
&lt;li&gt;Attachment: Attach it to your VPC for secure communication.&lt;/li&gt;
&lt;/ul&gt;

&lt;h3&gt;
  
  
  4. NAT Gateway
&lt;/h3&gt;

&lt;ul&gt;
&lt;li&gt;Function: Provides high-bandwidth, managed network address translation.&lt;/li&gt;
&lt;li&gt;Use Case:
Allows private subnets to access the internet while maintaining resource privacy. Supports UDP, TCP, and ICMP protocols.&lt;/li&gt;
&lt;/ul&gt;

&lt;h3&gt;
  
  
  5. VPC Peering
&lt;/h3&gt;

&lt;ul&gt;
&lt;li&gt;Function: Enables private communication between VPCs.&lt;/li&gt;
&lt;li&gt;Scenarios: Share resources across VPCs. Facilitate communication between different AWS accounts.&lt;/li&gt;
&lt;/ul&gt;

&lt;h3&gt;
  
  
  6. Security Groups
&lt;/h3&gt;

&lt;ul&gt;
&lt;li&gt;Function: Acts as a virtual firewall for EC2 instances.&lt;/li&gt;
&lt;li&gt;Control: Inbound and outbound traffic based on rules. Associate a security group with multiple instances.&lt;/li&gt;
&lt;/ul&gt;

&lt;h3&gt;
  
  
  7. Elastic IP (EIP)
&lt;/h3&gt;

&lt;ul&gt;
&lt;li&gt;Function: Provides a static public IP address for an instance.&lt;/li&gt;
&lt;li&gt;Stability: Remains constant even if the instance is stopped or restarted.&lt;/li&gt;
&lt;/ul&gt;

&lt;h3&gt;
  
  
  8. Network Access Control Lists (NACL)
&lt;/h3&gt;

&lt;ul&gt;
&lt;li&gt;Function: Adds an additional layer of security to your VPC.&lt;/li&gt;
&lt;li&gt;Role: Acts as a stateless firewall for controlling traffic in and out of subnets.&lt;/li&gt;
&lt;/ul&gt;

&lt;h3&gt;
  
  
  9. Customer Gateway
&lt;/h3&gt;

&lt;ul&gt;
&lt;li&gt;Function: Represents your side of a VPN connection.&lt;/li&gt;
&lt;li&gt;Type: Can be a physical or software appliance linking your network to Amazon VPC.&lt;/li&gt;
&lt;/ul&gt;

&lt;h3&gt;
  
  
  10. Network Interface
&lt;/h3&gt;

&lt;ul&gt;
&lt;li&gt;Function: Facilitates connections between private and public networks.&lt;/li&gt;
&lt;li&gt;Enables: Communication between resources within your VPC.&lt;/li&gt;
&lt;/ul&gt;

&lt;h2&gt;
  
  
  Functions of Amazon VPC
&lt;/h2&gt;

&lt;h3&gt;
  
  
  1. Isolation and Segmentation
&lt;/h3&gt;

&lt;p&gt;VPCs allow you to isolate resources, ensuring that different applications or environments don’t interfere with each other. Segmentation via subnets enables fine-grained control over network traffic.&lt;/p&gt;

&lt;h3&gt;
  
  
  2. Private and Public Subnets
&lt;/h3&gt;

&lt;p&gt;Create public subnets for resources that need direct internet access (e.g., web servers). Private subnets are ideal for backend services (e.g., databases) that should not be directly accessible from the internet.&lt;/p&gt;

&lt;h3&gt;
  
  
  3. Routing and Route Tables
&lt;/h3&gt;

&lt;p&gt;Route tables define how traffic flows between subnets and external networks. Use route tables to direct traffic to the appropriate destinations.&lt;/p&gt;

&lt;h3&gt;
  
  
  4. Security Groups
&lt;/h3&gt;

&lt;p&gt;Security groups act as virtual firewalls, controlling inbound and outbound traffic at the instance level. Specify rules to allow or deny traffic based on protocols, ports, and IP addresses.&lt;/p&gt;

&lt;h3&gt;
  
  
  5. VPC Peering
&lt;/h3&gt;

&lt;p&gt;Connect multiple VPCs privately without going through the public internet. Useful for sharing resources or communication between different VPCs.&lt;/p&gt;

&lt;h2&gt;
  
  
  Examples of Amazon VPC Usage
&lt;/h2&gt;

&lt;h3&gt;
  
  
  Example 1: Hosting a Public-Facing Website
&lt;/h3&gt;

&lt;p&gt;Suppose you want to host a blog or a simple website. Here’s how you can set it up using VPC:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;Create a VPC with a public subnet.&lt;/li&gt;
&lt;li&gt;Launch an EC2 instance (web server) in the public subnet.&lt;/li&gt;
&lt;li&gt;Attach an Elastic IP to the instance for a static public IP address.&lt;/li&gt;
&lt;li&gt;Configure security groups to allow HTTP/HTTPS traffic.&lt;/li&gt;
&lt;li&gt;Route internet-bound traffic to the internet gateway.&lt;/li&gt;
&lt;/ul&gt;

&lt;h3&gt;
  
  
  Example 2: Multi-Tier Application
&lt;/h3&gt;

&lt;p&gt;Imagine a multi-tier application with web servers, application servers, and a database:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;Create separate subnets for each tier (public, application, and database).&lt;/li&gt;
&lt;li&gt;Deploy EC2 instances in the respective subnets.&lt;/li&gt;
&lt;li&gt;Set up security groups to control communication between tiers.&lt;/li&gt;
&lt;li&gt;Route traffic appropriately using route tables.&lt;/li&gt;
&lt;/ul&gt;

&lt;h2&gt;
  
  
  Conclusion
&lt;/h2&gt;

&lt;p&gt;Amazon VPC provides the foundation for building secure, scalable, and isolated network architectures in AWS. By understanding its functions and leveraging practical examples, you can design robust and efficient cloud solutions.&lt;br&gt;
Remember to configure your VPCs carefully, considering factors like IP address ranges, subnet layouts, and security groups. With Amazon VPC, you have the flexibility to create custom network environments tailored to your specific requirements.&lt;/p&gt;

</description>
      <category>aws</category>
      <category>cloud</category>
      <category>beginners</category>
      <category>devops</category>
    </item>
    <item>
      <title>AWS ELB: A Beginner’s Guide</title>
      <dc:creator>kabeer1choudary</dc:creator>
      <pubDate>Thu, 21 Mar 2024 17:29:49 +0000</pubDate>
      <link>https://dev.to/kabeer1choudary/aws-elb-a-beginners-guide-2gd2</link>
      <guid>https://dev.to/kabeer1choudary/aws-elb-a-beginners-guide-2gd2</guid>
      <description>&lt;p&gt;In the dynamic world of cloud computing, Elastic Load Balancers (ELBs) play a crucial role in ensuring high availability, scalability, and efficient distribution of incoming traffic across backend servers. In this blog post, we’ll delve into the fundamentals of AWS ELBs, explore their types, and provide practical examples.&lt;/p&gt;

&lt;h2&gt;
  
  
  What is Elastic Load Balancing?
&lt;/h2&gt;

&lt;p&gt;Elastic Load Balancing is a service provided by Amazon Web Services (AWS) that automatically distributes incoming traffic across a group of backend servers. Here’s why it matters:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;Scalability: ELBs allow your application to handle increased traffic by distributing it efficiently.&lt;/li&gt;
&lt;li&gt;Fault Tolerance: If any backend server fails, ELBs automatically route traffic away from the unhealthy target.&lt;/li&gt;
&lt;li&gt;Security: ELBs enhance security by acting as a single entry point for incoming requests.&lt;/li&gt;
&lt;/ul&gt;

&lt;h2&gt;
  
  
  Types of AWS Load Balancers
&lt;/h2&gt;

&lt;h3&gt;
  
  
  1. Classic Load Balancer (CLB)
&lt;/h3&gt;

&lt;ul&gt;
&lt;li&gt;The traditional form of load balancer.&lt;/li&gt;
&lt;li&gt;Distributes traffic among instances.&lt;/li&gt;
&lt;li&gt;Operates at both the connection level (TCP/SSL) and request level (HTTP/HTTPS).&lt;/li&gt;
&lt;li&gt;Lacks intelligence for host-based or path-based routing.&lt;/li&gt;
&lt;/ul&gt;

&lt;h3&gt;
  
  
  2. Application Load Balancer (ALB)
&lt;/h3&gt;

&lt;ul&gt;
&lt;li&gt;Works at the Application layer (Layer 7) of the OSI model.&lt;/li&gt;
&lt;li&gt;Ideal for routing decisions related to HTTP and HTTPS traffic.&lt;/li&gt;
&lt;li&gt;Supports path-based and host-based routing.&lt;/li&gt;
&lt;li&gt;Example: Routing requests to different microservices based on URL paths.&lt;/li&gt;
&lt;/ul&gt;

&lt;h3&gt;
  
  
  3. Network Load Balancer (NLB)
&lt;/h3&gt;

&lt;ul&gt;
&lt;li&gt;Operates at the Transport layer (Layer 4) of the OSI model.&lt;/li&gt;
&lt;li&gt;Handles millions of requests per second.&lt;/li&gt;
&lt;li&gt;Mainly used for load-balancing TCP traffic.&lt;/li&gt;
&lt;li&gt;Example: Distributing traffic to backend servers hosting gaming sessions.&lt;/li&gt;
&lt;/ul&gt;

&lt;h3&gt;
  
  
  4. Gateway Load Balancer (GLB)
&lt;/h3&gt;

&lt;ul&gt;
&lt;li&gt;Combines a transparent network gateway with load balancing.&lt;/li&gt;
&lt;li&gt;Used for deploying, scaling, and managing virtual appliances like firewalls.&lt;/li&gt;
&lt;li&gt;Routes traffic efficiently.&lt;/li&gt;
&lt;li&gt;Example: Deploying a firewall appliance in front of your VPC.&lt;/li&gt;
&lt;/ul&gt;

&lt;h2&gt;
  
  
  Practical Examples
&lt;/h2&gt;

&lt;h3&gt;
  
  
  Example 1: ALB for Microservices
&lt;/h3&gt;

&lt;p&gt;Suppose you have a web application with multiple microservices (user service, product service, payment service). An ALB can route requests based on URL paths:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;/users → User Service&lt;/li&gt;
&lt;li&gt;/products → Product Service&lt;/li&gt;
&lt;li&gt;/payments → Payment Service&lt;/li&gt;
&lt;/ul&gt;

&lt;h3&gt;
  
  
  Example 2: NLB for Gaming Servers
&lt;/h3&gt;

&lt;p&gt;Imagine an online multiplayer game with thousands of players. NLB can distribute incoming game traffic to backend servers hosting gaming sessions. It ensures low latency and high throughput.&lt;/p&gt;

&lt;h2&gt;
  
  
  Conclusion
&lt;/h2&gt;

&lt;p&gt;AWS Elastic Load Balancers are essential components for building scalable, fault-tolerant, and secure applications. By understanding their types and functions, you can optimize your infrastructure and deliver a seamless experience to your users.&lt;/p&gt;

</description>
      <category>aws</category>
      <category>beginners</category>
      <category>cloud</category>
      <category>devops</category>
    </item>
    <item>
      <title>AWS Cloud Watch: A Beginner’s Guide</title>
      <dc:creator>kabeer1choudary</dc:creator>
      <pubDate>Wed, 28 Feb 2024 13:15:10 +0000</pubDate>
      <link>https://dev.to/kabeer1choudary/aws-cloud-watch-a-beginners-guide-1cjl</link>
      <guid>https://dev.to/kabeer1choudary/aws-cloud-watch-a-beginners-guide-1cjl</guid>
      <description>&lt;p&gt;&lt;strong&gt;Amazon CloudWatch&lt;/strong&gt; is a powerful monitoring and observability service provided by Amazon Web Services (AWS). It allows you to track and analyze the performance of your AWS resources in real-time. Whether you’re a developer, DevOps engineer, or IT manager, CloudWatch provides valuable insights to optimize your applications and infrastructure. In this blog post, we’ll explore what CloudWatch is, its features, and provide straightforward examples to help you understand its usage.&lt;/p&gt;

&lt;h2&gt;
  
  
  What is Amazon CloudWatch?
&lt;/h2&gt;

&lt;p&gt;Amazon CloudWatch collects and stores operational data in the form of logs, metrics, and events. Here are some key points:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;Metrics: CloudWatch Metrics represent time-ordered data points related to your AWS resources. These metrics can be CPU utilization, memory usage, disk I/O, and more. Metrics are uniquely defined by a name, namespace, and dimensions.&lt;/li&gt;
&lt;li&gt;Logs: CloudWatch Logs allow you to collect, monitor, and analyze log files from your applications and services. You can use custom filters and queries to extract meaningful information from logs.&lt;/li&gt;
&lt;li&gt;Alarms: CloudWatch Alarms help you set thresholds and trigger actions based on metric data. For example, you can create an alarm to notify you when CPU utilization exceeds a certain threshold.&lt;/li&gt;
&lt;/ul&gt;

&lt;h2&gt;
  
  
  Simple Examples
&lt;/h2&gt;

&lt;p&gt;Let’s dive into some straightforward examples:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;
&lt;strong&gt;Monitoring EC2 Instances:&lt;/strong&gt;
Suppose you want to monitor the CPU utilization of your EC2 instances. You can create a custom CloudWatch metric for CPU utilization and set up an alarm to receive notifications when it exceeds a specific value.&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;Analyzing Application Logs:&lt;/strong&gt;
Imagine you have an application running on AWS Lambda. You can configure CloudWatch Logs to capture logs generated by your Lambda functions. Use CloudWatch Insights to query and analyze these logs, identifying patterns or errors.&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;Autoscaling with CloudWatch Alarms:&lt;/strong&gt;
Suppose you have an Auto Scaling group that dynamically adjusts the number of EC2 instances based on demand. You can set up CloudWatch alarms to trigger scaling actions when specific conditions are met. For instance, increase the instance count when CPU utilization is high.&lt;/li&gt;
&lt;/ul&gt;

&lt;h2&gt;
  
  
  Conclusion
&lt;/h2&gt;

&lt;p&gt;Amazon CloudWatch is a versatile service that provides real-time visibility into your AWS environment. By understanding its features and leveraging simple examples, you’ll be better equipped to monitor, troubleshoot, and optimize your resources effectively.&lt;/p&gt;

&lt;p&gt;For more information, visit the official Amazon CloudWatch &lt;a href="https://docs.aws.amazon.com/AmazonCloudWatch/latest/monitoring/Alarm-Use-Cases.html" rel="noopener noreferrer"&gt;documentation&lt;/a&gt;.&lt;/p&gt;

</description>
      <category>aws</category>
      <category>cloud</category>
      <category>beginners</category>
      <category>cloudwatch</category>
    </item>
    <item>
      <title>AWS CloudTrail : A Beginner’s Guide</title>
      <dc:creator>kabeer1choudary</dc:creator>
      <pubDate>Wed, 28 Feb 2024 13:11:34 +0000</pubDate>
      <link>https://dev.to/kabeer1choudary/aws-cloudtrail-a-beginners-guide-20nb</link>
      <guid>https://dev.to/kabeer1choudary/aws-cloudtrail-a-beginners-guide-20nb</guid>
      <description>&lt;p&gt;&lt;strong&gt;AWS CloudTrail&lt;/strong&gt; is a powerful service offered by Amazon Web Services (AWS) that allows you to track and document activities within your AWS infrastructure. Whether you’re managing resources, services, or user accounts, CloudTrail provides a detailed event history of every action taken. In this blog post, we’ll explore what CloudTrail is, its benefits, and provide straightforward examples to help you understand its usage.&lt;/p&gt;

&lt;h2&gt;
  
  
  What is AWS CloudTrail?
&lt;/h2&gt;

&lt;p&gt;AWS CloudTrail records API calls and actions made within your AWS account. Here are some key points:&lt;/p&gt;

&lt;h3&gt;
  
  
  Event History:
&lt;/h3&gt;

&lt;p&gt;By default, your AWS account has CloudTrail activated, and you have immediate access to the CloudTrail Event history. This history provides a viewable, searchable, printable, and immutable record of the last 90 days’ worth of management events in an AWS Region. These events include actions performed via the AWS Management Console, AWS Command Line Interface (CLI), and AWS SDKs and APIs.&lt;/p&gt;

&lt;h3&gt;
  
  
  CloudTrail Lake:
&lt;/h3&gt;

&lt;p&gt;For more advanced use cases, AWS CloudTrail Lake is a managed data lake that records, stores, and analyzes user and API activity on AWS. It converts existing events into an efficient storage format called Apache ORC (Optimized Row Columnar). You can keep event data in CloudTrail Lake for up to seven years, making it ideal for audit and security purposes.&lt;/p&gt;

&lt;h3&gt;
  
  
  Trails:
&lt;/h3&gt;

&lt;p&gt;Trails allow you to deliver and store events in an Amazon S3 bucket. Additionally, you can send events to Amazon CloudWatch Logs and Amazon EventBridge. Trails are essential for security monitoring and compliance. You can create trails for individual AWS accounts or multiple accounts using AWS Organizations.&lt;/p&gt;

&lt;h2&gt;
  
  
  Simple Examples
&lt;/h2&gt;

&lt;p&gt;Let’s dive into some straightforward examples:&lt;/p&gt;

&lt;h3&gt;
  
  
  - Event History
&lt;/h3&gt;

&lt;p&gt;Suppose you want to review recent management events in your AWS account. You can access the CloudTrail Event history, which includes actions like creating EC2 instances, modifying security groups, or launching Lambda functions. This history is available for free and covers the last 90 days.&lt;/p&gt;

&lt;h3&gt;
  
  
  - CloudTrail Lake
&lt;/h3&gt;

&lt;p&gt;Imagine you need a long-term storage solution for your event data. AWS CloudTrail Lake is your answer. It converts existing events into ORC format, making data retrieval efficient. You can create event data stores based on specific criteria and retain them for up to seven years. Use Lake dashboards to analyze trends and gain insights.&lt;/p&gt;

&lt;h3&gt;
  
  
  - Creating a Trail
&lt;/h3&gt;

&lt;p&gt;Suppose you want to monitor API call volumes and error rates. By creating a trail, you can analyze management events for unusual behavior. Trails deliver events to an S3 bucket, CloudWatch Logs, or EventBridge. You can even search and examine CloudTrail logs using tools like Amazon Athena.&lt;/p&gt;

&lt;h2&gt;
  
  
  Conclusion
&lt;/h2&gt;

&lt;p&gt;AWS CloudTrail is a valuable tool for maintaining security, ensuring compliance, and troubleshooting issues within your AWS environment. By understanding its features and leveraging simple examples, you’ll be better equipped to manage your cloud resources effectively.&lt;/p&gt;

&lt;p&gt;For more information, visit the official AWS CloudTrail &lt;a href="https://docs.aws.amazon.com/awscloudtrail/latest/userguide/cloudtrail-log-file-examples.html" rel="noopener noreferrer"&gt;documentation&lt;/a&gt;.&lt;/p&gt;

</description>
      <category>aws</category>
      <category>beginners</category>
      <category>cloud</category>
      <category>cloudtrail</category>
    </item>
    <item>
      <title>AWS S3: A Beginner’s Guide</title>
      <dc:creator>kabeer1choudary</dc:creator>
      <pubDate>Sun, 25 Feb 2024 12:14:04 +0000</pubDate>
      <link>https://dev.to/kabeer1choudary/aws-s3-a-beginners-guide-p0j</link>
      <guid>https://dev.to/kabeer1choudary/aws-s3-a-beginners-guide-p0j</guid>
      <description>&lt;p&gt;Amazon S3 (Simple Storage Service) is a powerful cloud-based storage solution provided by Amazon Web Services (AWS). It allows you to store and retrieve data securely from anywhere on the web. Whether you’re a developer, a business owner, or just curious about cloud storage, this guide will help you understand the basics of Amazon S3.&lt;/p&gt;

&lt;h2&gt;
  
  
  Key Concepts
&lt;/h2&gt;

&lt;h3&gt;
  
  
  Buckets:
&lt;/h3&gt;

&lt;p&gt;Think of an Amazon S3 bucket as a virtual container for your files. It’s like a folder in the cloud where you can organize and store your data. Buckets have unique names (similar to domain names) and are globally accessible.&lt;/p&gt;

&lt;h3&gt;
  
  
  Objects:
&lt;/h3&gt;

&lt;p&gt;Objects are the files you store in an S3 bucket. These can be anything: documents, images, videos, backups, or even cat memes!. Each object has a unique key (similar to a file path) within the bucket.&lt;/p&gt;

&lt;h3&gt;
  
  
  Scalability and Durability:
&lt;/h3&gt;

&lt;p&gt;Amazon S3 is highly scalable. You can store as little as a single file or as much as petabytes of data. It’s also incredibly durable. Your data is redundantly stored across multiple data centers, ensuring high availability.&lt;/p&gt;

&lt;h2&gt;
  
  
  Use Cases
&lt;/h2&gt;

&lt;h3&gt;
  
  
  Static Website Hosting:
&lt;/h3&gt;

&lt;p&gt;You can host static websites directly from an S3 bucket. Upload your HTML, CSS, and JavaScript files, set permissions, and voilà! Your website is live.&lt;/p&gt;

&lt;h3&gt;
  
  
  Data Backup and Archiving:
&lt;/h3&gt;

&lt;p&gt;Use S3 to back up critical data. It’s like having a digital safe deposit box. Set lifecycle policies to automatically move older data to cheaper storage classes (like Glacier) for long-term archiving.&lt;/p&gt;

&lt;h3&gt;
  
  
  Media Storage and Distribution:
&lt;/h3&gt;

&lt;p&gt;Store media files (videos, images, audio) and serve them to users via Amazon CloudFront (a content delivery network). CloudFront caches content close to users, reducing latency.&lt;/p&gt;

&lt;p&gt;Remember, this is a simplified explanation, but it gives you an idea of how Amazon S3 can be used on different systems. Feel free to explore more about AWS services and their capabilities! &lt;/p&gt;

</description>
      <category>aws</category>
      <category>cloud</category>
      <category>beginners</category>
      <category>s3</category>
    </item>
    <item>
      <title>AWS Lambda: A Beginner’s Guide</title>
      <dc:creator>kabeer1choudary</dc:creator>
      <pubDate>Sun, 25 Feb 2024 12:11:36 +0000</pubDate>
      <link>https://dev.to/kabeer1choudary/aws-lambda-a-beginners-guide-1648</link>
      <guid>https://dev.to/kabeer1choudary/aws-lambda-a-beginners-guide-1648</guid>
      <description>&lt;p&gt;AWS Lambda is a serverless compute service provided by Amazon Web Services (AWS). It allows you to run code without provisioning or managing servers. You only pay for the compute time you consume, making it cost-effective and efficient.&lt;/p&gt;

&lt;p&gt;In simple words, AWS Lambda lets you execute code in response to events, such as file uploads, scheduled tasks, or messages published to an SNS topic. Let’s dive deeper into how it works and explore some practical examples.&lt;/p&gt;

&lt;h2&gt;
  
  
  How AWS Lambda Works
&lt;/h2&gt;

&lt;h3&gt;
  
  
  1. Event Triggers:
&lt;/h3&gt;

&lt;p&gt;Lambda functions are triggered by events. Some examples include:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;A file uploaded to Amazon S3 (cloud storage service).&lt;/li&gt;
&lt;li&gt;A cron job that runs your function at regular intervals.&lt;/li&gt;
&lt;li&gt;A message published to an SNS topic (a publish-subscribe service).&lt;/li&gt;
&lt;/ul&gt;

&lt;h3&gt;
  
  
  2. Function Execution:
&lt;/h3&gt;

&lt;p&gt;When an event occurs, AWS Lambda automatically provisions compute resources to run your code. It executes your function in an isolated environment.&lt;/p&gt;

&lt;h3&gt;
  
  
  3. Scaling:
&lt;/h3&gt;

&lt;p&gt;Lambda scales automatically based on the incoming workload. If many events occur simultaneously, Lambda creates multiple instances of your function to handle them.&lt;/p&gt;

&lt;h3&gt;
  
  
  4. Billing:
&lt;/h3&gt;

&lt;p&gt;You are billed only for the compute time your function consumes. There’s no need to worry about server maintenance or capacity planning.&lt;/p&gt;

&lt;h2&gt;
  
  
  Practical Examples
&lt;/h2&gt;

&lt;h3&gt;
  
  
  1. Media Transformation
&lt;/h3&gt;

&lt;p&gt;Imagine you’re building an application that handles media files. You can use Lambda to automatically resize images when they’re uploaded to an S3 bucket. Here’s how it works:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;Event: A user uploads an image to S3.&lt;/li&gt;
&lt;li&gt;Lambda Function: Your Lambda function triggers on the S3 upload event.&lt;/li&gt;
&lt;li&gt;Action: The function resizes the image and stores it back in S3.&lt;/li&gt;
&lt;/ul&gt;

&lt;h3&gt;
  
  
  2. Cross-Device Development
&lt;/h3&gt;

&lt;p&gt;Developing applications for different devices can be challenging. Lambda can help by converting media files to different formats suitable for various devices. For example:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;Event: A video is uploaded to S3.&lt;/li&gt;
&lt;li&gt;Lambda Function: Your function converts the video to different resolutions and formats.&lt;/li&gt;
&lt;li&gt;Action: Users can stream the video seamlessly on their phones, tablets, or desktops.&lt;/li&gt;
&lt;/ul&gt;

&lt;h3&gt;
  
  
  3. Real-Time Data Processing
&lt;/h3&gt;

&lt;p&gt;Lambda is excellent for real-time data processing. Let’s say you’re building a chat application:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;Event: A user sends a message.&lt;/li&gt;
&lt;li&gt;Lambda Function: Your function processes the message, checks for profanity, and logs it.&lt;/li&gt;
&lt;li&gt;Action: The chat remains clean, and you have a record of all messages.&lt;/li&gt;
&lt;/ul&gt;

&lt;h2&gt;
  
  
  Conclusion
&lt;/h2&gt;

&lt;p&gt;AWS Lambda simplifies serverless computing, allowing you to focus on writing code rather than managing infrastructure. With practical examples like media transformation, cross-device development, and real-time data processing, you can harness the power of Lambda for your applications.&lt;/p&gt;

&lt;p&gt;Remember to explore the official AWS Lambda documentation for more details and best practices. Happy coding!&lt;/p&gt;

</description>
      <category>aws</category>
      <category>cloud</category>
      <category>beginners</category>
      <category>lambda</category>
    </item>
    <item>
      <title>AWS IAM: A Beginner’s Guide</title>
      <dc:creator>kabeer1choudary</dc:creator>
      <pubDate>Fri, 23 Feb 2024 12:28:56 +0000</pubDate>
      <link>https://dev.to/kabeer1choudary/aws-iam-a-beginners-guide-fe6</link>
      <guid>https://dev.to/kabeer1choudary/aws-iam-a-beginners-guide-fe6</guid>
      <description>&lt;p&gt;Amazon Web Services (AWS) offers a plethora of services, and Identity and Access Management (IAM) is a critical component for securing your AWS resources. In this blog post, we’ll demystify IAM, explore its features, and provide straightforward examples to help you grasp its importance.&lt;/p&gt;

&lt;h2&gt;
  
  
  What is IAM?
&lt;/h2&gt;

&lt;p&gt;IAM stands for Identity and Access Management. Let’s break it down:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;Identity: IAM helps you manage users, groups, and roles within your AWS account. These identities are essential for controlling access to AWS resources.&lt;/li&gt;
&lt;li&gt;Access Management: IAM allows you to define who can do what in your AWS environment. You can grant or restrict permissions based on roles and policies.&lt;/li&gt;
&lt;/ul&gt;

&lt;h2&gt;
  
  
  Why Do We Need IAM?
&lt;/h2&gt;

&lt;p&gt;Before IAM, managing access was chaotic:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;Shared Passwords: People shared passwords over insecure channels like email or phone calls.&lt;/li&gt;
&lt;li&gt;Single Admin Password: Only one admin password existed, stored in a vulnerable location.&lt;/li&gt;
&lt;li&gt;Lack of Security: Anyone could eavesdrop and gain unauthorized access.
IAM solves these problems by providing a secure way to manage access. Let’s dive deeper.&lt;/li&gt;
&lt;/ul&gt;

&lt;h2&gt;
  
  
  Key Components of IAM
&lt;/h2&gt;

&lt;ul&gt;
&lt;li&gt;Users: Represent individuals or applications. Each user has a unique set of credentials (username and password).&lt;/li&gt;
&lt;li&gt;Groups: Logical collections of users. Assign permissions to groups instead of individual users.&lt;/li&gt;
&lt;li&gt;Roles: Used by services or applications running on EC2 instances, Lambda functions, etc. Roles grant temporary permissions.&lt;/li&gt;
&lt;/ul&gt;

&lt;h2&gt;
  
  
  IAM Examples
&lt;/h2&gt;

&lt;h3&gt;
  
  
  1. Creating a New User
&lt;/h3&gt;

&lt;ul&gt;
&lt;li&gt;Log in to the AWS Console.&lt;/li&gt;
&lt;li&gt;Search for IAM and navigate to the IAM dashboard.&lt;/li&gt;
&lt;li&gt;Click “Users” and then “Add user.”&lt;/li&gt;
&lt;li&gt;Specify the username and choose access type (programmatic or console).&lt;/li&gt;
&lt;li&gt;Assign permissions (attach policies) to the user.&lt;/li&gt;
&lt;/ul&gt;

&lt;h3&gt;
  
  
  2. Creating a Group
&lt;/h3&gt;

&lt;ul&gt;
&lt;li&gt;In IAM, click “Groups” and then “Create group.”&lt;/li&gt;
&lt;li&gt;Name the group (e.g., “Developers”).&lt;/li&gt;
&lt;li&gt;Attach policies to the group (e.g., “AmazonS3FullAccess”).&lt;/li&gt;
&lt;li&gt;Add users to the group.&lt;/li&gt;
&lt;/ul&gt;

&lt;h3&gt;
  
  
  3. Using Roles
&lt;/h3&gt;

&lt;ul&gt;
&lt;li&gt;Create an IAM role (e.g., “LambdaRole”).&lt;/li&gt;
&lt;li&gt;Define the trusted entity (e.g., Lambda service).&lt;/li&gt;
&lt;li&gt;Attach policies (e.g., “AmazonDynamoDBFullAccess”).&lt;/li&gt;
&lt;li&gt;Lambda functions assume this role when executing.&lt;/li&gt;
&lt;/ul&gt;

&lt;h2&gt;
  
  
  Conclusion
&lt;/h2&gt;

&lt;p&gt;IAM is your gatekeeper to AWS resources. By understanding users, groups, and roles, you can control who accesses what. Remember, IAM is not just about security; it’s about enabling secure collaboration and efficient resource management.&lt;/p&gt;

</description>
      <category>aws</category>
      <category>cloud</category>
      <category>devops</category>
      <category>iam</category>
    </item>
  </channel>
</rss>
