<?xml version="1.0" encoding="UTF-8"?>
<rss version="2.0" xmlns:atom="http://www.w3.org/2005/Atom" xmlns:dc="http://purl.org/dc/elements/1.1/">
  <channel>
    <title>DEV Community: Grace Gong</title>
    <description>The latest articles on DEV Community by Grace Gong (@1grace).</description>
    <link>https://dev.to/1grace</link>
    <image>
      <url>https://media2.dev.to/dynamic/image/width=90,height=90,fit=cover,gravity=auto,format=auto/https:%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Fuser%2Fprofile_image%2F922032%2F872104b4-0236-4317-ac9c-04be4d49b2c3.jpeg</url>
      <title>DEV Community: Grace Gong</title>
      <link>https://dev.to/1grace</link>
    </image>
    <atom:link rel="self" type="application/rss+xml" href="https://dev.to/feed/1grace"/>
    <language>en</language>
    <item>
      <title>TPUs vs. GPUs: What They Are, How They Differ, and Which Workloads Belong on Each</title>
      <dc:creator>Grace Gong</dc:creator>
      <pubDate>Fri, 01 May 2026 01:53:27 +0000</pubDate>
      <link>https://dev.to/1grace/tpus-vs-gpus-what-they-are-how-they-differ-and-which-workloads-belong-on-each-278p</link>
      <guid>https://dev.to/1grace/tpus-vs-gpus-what-they-are-how-they-differ-and-which-workloads-belong-on-each-278p</guid>
      <description>&lt;p&gt;If you've worked with machine learning on Google Cloud, you've hit the choice: GPU instance or TPU? Most teams default to GPU because that's what they already know. But as inference costs climb and TPU tooling matures, it's worth understanding what each chip actually does and when one outperforms the other.&lt;/p&gt;

&lt;p&gt;This post covers what GPUs and TPUs are, how they work, and which workloads run better on each. It ends with a look at Google's current TPU lineup, including the eighth-generation chips announced at Google Cloud Next 2026.&lt;/p&gt;




&lt;h2&gt;
  
  
  Why TPUs exist
&lt;/h2&gt;

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fbnlj3ns9uifkobnwjxys.png" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fbnlj3ns9uifkobnwjxys.png" alt=" " width="800" height="600"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;Image source: Google Cloud&lt;/p&gt;

&lt;p&gt;GPUs were originally built for rendering video games. They handle AI workloads well because the underlying math, large parallel floating-point operations is the same. Researchers figured this out around 2012, and GPUs became the default for training neural networks.&lt;/p&gt;

&lt;p&gt;Google ran into a problem in 2013. Engineers at Google Brain calculated that if every Android user used voice search for just three minutes a day, Google would need to double its global data center capacity. Running inference on general-purpose GPUs at that scale was too expensive and power-hungry.&lt;/p&gt;

&lt;p&gt;Their solution was to build a chip designed specifically for neural network math. The first TPU went into production in Google's data centers in 2015. Google made Cloud TPUs publicly available in 2018. The core idea, strip out everything a GPU carries from its graphics origins and focus entirely on matrix multiplication still drives every TPU generation today.&lt;/p&gt;




&lt;h2&gt;
  
  
  How a GPU works
&lt;/h2&gt;

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fvlmsk2no9ykaxwoi9j6c.png" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fvlmsk2no9ykaxwoi9j6c.png" alt=" " width="800" height="230"&gt;&lt;/a&gt;&lt;br&gt;
Image source: Google Cloud. Some images of GPUs.&lt;/p&gt;

&lt;p&gt;A GPU is a parallel processor with thousands of smaller cores. Where a CPU has 8 to 64 powerful general-purpose cores, a high-end GPU like the NVIDIA H100 has thousands of smaller ones that run the same instruction across many data points at once. This is called SIMD (Single Instruction, Multiple Data) parallelism.&lt;/p&gt;

&lt;p&gt;GPUs support a wide range of precision formats: FP32, FP16, BF16, INT8, FP8. They run PyTorch, TensorFlow, JAX, CUDA libraries, simulations, rendering pipelines. That broad support is useful, but it means a GPU carries hardware for texture mapping, branch prediction, and other operations that sit completely idle during a matrix multiplication.&lt;/p&gt;

&lt;p&gt;The NVIDIA H100 has 80GB of HBM2e memory on-package. Memory bandwidth matters a lot for AI workloads because moving data between memory and compute units is often what limits throughput, not the raw math.&lt;/p&gt;




&lt;h2&gt;
  
  
  How a TPU works
&lt;/h2&gt;

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2F2t1zxk2wo9hk2dhz2e9a.png" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2F2t1zxk2wo9hk2dhz2e9a.png" alt=" " width="800" height="450"&gt;&lt;/a&gt;&lt;br&gt;
Image source: Google Cloud&lt;/p&gt;

&lt;p&gt;A TPU is built for one job: tensor math. Specifically, the matrix multiplications at the core of neural network training and inference.&lt;/p&gt;

&lt;p&gt;The key piece of hardware is the systolic array. In a standard processor, every operation reads inputs from memory, computes, and writes the result back. In a systolic array, data flows through a grid of multiply-and-accumulate units. You load the weights once, pass inputs through the grid, and results flow from unit to unit without going back to main memory. This removes the constant memory round-trips that slow conventional chips.&lt;/p&gt;

&lt;p&gt;Google built BF16 support into TPUs from early generations; GPUs added it later. Recent chips support FP8 natively, which helps throughput for inference workloads.&lt;/p&gt;

&lt;p&gt;The limitation: TPUs work poorly with dynamic control flow, variable-length sequences, and custom operations. They are best suited for static computation graphs, which is what most transformer models produce.&lt;/p&gt;




&lt;h2&gt;
  
  
  Side-by-side comparison
&lt;/h2&gt;

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fggvosj8qgounywne3lp7.png" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fggvosj8qgounywne3lp7.png" alt=" " width="800" height="572"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;h2&gt;
  
  
  When to use a GPU
&lt;/h2&gt;

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2F8cy7swiha61w6vv4uhp8.png" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2F8cy7swiha61w6vv4uhp8.png" alt=" " width="800" height="415"&gt;&lt;/a&gt;&lt;br&gt;
Image source: Google Cloud. &lt;/p&gt;

&lt;p&gt;Recommended GPUs based on workload type.PyTorch-first teams. Most research code on GitHub, most open-source model checkpoints, and most fine-tuning guides assume a GPU. If your team works primarily in PyTorch, starting on GPU is faster.&lt;/p&gt;

&lt;p&gt;Models with TensorFlow ops that are not available on Cloud TPU (see the list of available TensorFlow ops)&lt;/p&gt;

&lt;p&gt;Models with dynamic inputs. Variable-length sequences, conditional branches, custom CUDA extensions - these work on GPUs and can be tricky to run on TPUs.&lt;/p&gt;

&lt;p&gt;Medium-to-large models with larger effective batch sizes&lt;/p&gt;

&lt;p&gt;Multi-cloud or on-prem deployments. TPUs only exist in Google Cloud. If your infrastructure is on AWS, Azure, or your own servers, you don't have a choice.&lt;/p&gt;

&lt;p&gt;Mixed workloads. If the same team does ML training, scientific simulation, and rendering, GPUs handle all of it. TPUs don't.&lt;/p&gt;

&lt;p&gt;Small teams moving fast. GPU tooling (profilers, debuggers, community tutorials) is more mature. Diagnosing a performance problem on a GPU is easier today than on a TPU.&lt;/p&gt;




&lt;h2&gt;
  
  
  When to use a TPU
&lt;/h2&gt;

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fnk62q6c1hwo9gbdopjsx.png" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fnk62q6c1hwo9gbdopjsx.png" alt=" " width="601" height="601"&gt;&lt;/a&gt;&lt;br&gt;
Models relying on embeddings: Cloud TPUs feature SparseCores, which are dataflow processors specifically built to accelerate models that heavily use embeddings. This makes them ideal for applications like recommendation systems. - Google Cloud &lt;/p&gt;

&lt;p&gt;Training massive deep learning models: If you're building and training large and complex deep learning models, especially large language models (LLMs), Cloud TPUs are designed to handle the immense number of matrix calculations involved efficiently.&lt;/p&gt;

&lt;p&gt;Models dominated by matrix computations&lt;/p&gt;

&lt;p&gt;Models that train for weeks or months&lt;/p&gt;

&lt;p&gt;Models with ultra-large embeddings common in advanced ranking and recommendation workloads&lt;/p&gt;

&lt;p&gt;Large-scale transformer training. TPU pods scale to tens of thousands of chips through Google's Inter-Chip Interconnect (ICI). Training something like Gemma on a TPU pod tends to be faster and cheaper per token than an equivalent GPU cluster.&lt;/p&gt;

&lt;p&gt;High-volume production inference. TPU v6e (Trillium) and Ironwood were built specifically for inference workloads. Ironwood delivers more than 4x better performance per chip for inference compared to TPU v6e (Trillium).&lt;/p&gt;

&lt;p&gt;Models with no custom PyTorch/JAX operations inside the main training loop&lt;/p&gt;

&lt;p&gt;Google open-weight models. Gemma 4 (released April 2026) is built and optimized for TPU serving. Google publishes JAX reference implementations for every Gemma variant, and there are community guides for deploying Gemma 4 via vLLM on Cloud TPU.&lt;/p&gt;

&lt;p&gt;Cloud TPUs are not suited to the following workloads:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;Linear algebra programs that require frequent branching or contain many element-wise algebra operations&lt;/li&gt;
&lt;li&gt;Workloads that require high-precision arithmetic&lt;/li&gt;
&lt;li&gt;Neural network workloads that contain custom operations in the main training loop&lt;/li&gt;
&lt;/ul&gt;




&lt;h2&gt;
  
  
  Google's current TPU lineup
&lt;/h2&gt;

&lt;h2&gt;
  
  
  TPU v5e, available now
&lt;/h2&gt;

&lt;p&gt;Good starting point. Used for smaller inference workloads and fine-tuning. Lower per-chip cost than newer generations.&lt;/p&gt;

&lt;h2&gt;
  
  
  TPU v6e (Trillium), available now
&lt;/h2&gt;

&lt;p&gt;4.7x the peak compute of v5e, with 67% better energy efficiency. Scales to 256 chips per pod. Still widely used for inference, particularly for teams where cost per chip-hour matters more than raw throughput. vLLM supports TPU v6e for both offline batch inference and online API serving.&lt;/p&gt;

&lt;h2&gt;
  
  
  TPU v7 (Ironwood), generally available since late 2025
&lt;/h2&gt;

&lt;p&gt;Announced at Google Cloud Next 2025. Specs per chip: 4,614 FP8 TFLOPS, 192GB of HBM3E memory, 7.37 TB/s memory bandwidth, 9.6 Tb/s inter-chip interconnect. Scales to 9,216 chips in a single superpod, delivering 42.5 FP8 ExaFLOPS per pod. That's more than 4x the performance per chip of &lt;br&gt;
TPU v6e (Trillium) and 10x of TPU v5p.&lt;br&gt;
Each Ironwood chip contains two TensorCores and four SparseCores in a dual-chiplet design. Anthropic's Claude models train and serve on TPUs, and Anthropic signed an agreement to access up to one million Ironwood TPUs through Google Cloud.&lt;br&gt;
Ironwood is the first TPU generation where Google used AlphaChip - a reinforcement learning tool - to design the chip layouts.&lt;/p&gt;

&lt;h2&gt;
  
  
  TPU 8t and TPU 8i (eighth generation), coming later in 2026
&lt;/h2&gt;

&lt;p&gt;Announced at Google Cloud Next 2026. For the first time, Google has split its TPU lineup into two chips with different architectures for training and inference.&lt;/p&gt;

&lt;p&gt;TPU 8t is built for training. A single superpod holds 9,600 chips with 2 petabytes of shared HBM memory and 121 FP4 ExaFLOPS of compute, nearly tripling compute per pod versus Ironwood. ICI bandwidth is 19.2 Tb/s per chip, double Ironwood. The new Virgo Network fabric can link 134,000 chips across a data center and theoretically over 1 million chips across sites. TPUDirect RDMA and TPU Direct Storage bypass the host CPU entirely, doubling bandwidth for large data transfers. Google targets 97% goodput meaning 97% of compute cycles go toward actual learning rather than overhead.&lt;/p&gt;

&lt;p&gt;TPU 8i is built for inference. It scales to 1,152 chips per pod and delivers 11.6 FP8 ExaFLOPS. Each chip carries 288GB of HBM, more than the 8t training chip and 384MB of on-chip SRAM, 3x what Ironwood had. Google reports 80% better performance-per-dollar versus Ironwood for inference, and 2x better performance-per-watt.&lt;/p&gt;

&lt;p&gt;The 8i uses a new Boardfly interconnect that reduces the maximum number of network hops from 16 to 7. This matters for Mixture-of-Experts models, where data needs to move quickly between expert layers. The chip also replaces Ironwood's SparseCores with a Collectives Acceleration Engine (CAE), which cuts the latency of collective operations by 5x - important when many agents are running concurrently and small latency multiplies across thousands of calls.&lt;/p&gt;

&lt;p&gt;The reason the inference chip has more memory than the training chip: large MoE inference is memory-bandwidth-bound. The chip serving tokens needs to stream weights and KV-cache faster than the chip training the model. Both 8t and 8i run on Google's Axion ARM host CPU and use liquid cooling.&lt;/p&gt;

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fte36f66huvkb1nhtvwhj.png" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fte36f66huvkb1nhtvwhj.png" alt=" " width="800" height="730"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;h2&gt;
  
  
  More info from TPU Overview
&lt;/h2&gt;




&lt;h2&gt;
  
  
  The software side
&lt;/h2&gt;

&lt;p&gt;TPUs run best with a few specific tools:&lt;br&gt;
JAX is Google's ML framework. Its jit, vmap, pmap, and shard_map primitives map directly onto TPU hardware. If you're new to TPUs and want to get the most out of them, JAX is where to start.&lt;br&gt;
MaxText is Google's open-source LLM reference implementation for TPUs, available at AI-Hypercomputer/maxtext on GitHub. It's a practical starting point for training large language models on TPU pods.&lt;br&gt;
Pallas is Google's Python-based kernel language for writing low-level, hardware-aware kernels. Supported on both Ironwood and the eighth-generation chips.&lt;br&gt;
vLLM now has first-class TPU support. You can run offline batch inference or an OpenAI-compatible API server on a Cloud TPU VM with standard configuration.&lt;br&gt;
PyTorch on TPU is in preview as of the eighth-generation launch. If your team is on PyTorch, you can now bring existing models to TPU hardware without rewriting them in JAX.&lt;br&gt;
Google's Gemma 4 (April 2026) is optimized for TPU serving. The google-deepmind/gemma GitHub repo has JAX reference implementations for every model variant.&lt;/p&gt;




&lt;h2&gt;
  
  
  Summary
&lt;/h2&gt;

&lt;p&gt;GPUs are the practical default for most research and development work. The tooling is mature, the community is large, and most models you'll find online were built on GPUs.&lt;br&gt;
TPUs are worth the switch when you're running workloads at sustained scale on Google Cloud, especially for inference. Ironwood is available today. The eighth-generation 8t and 8i chips, which separate training and inference into dedicated hardware, are coming later in 2026. If you want to try TPUs before committing, Google Colab's free TPU runtime lets you run a JAX or Keras model on one without any setup.&lt;/p&gt;




&lt;h2&gt;
  
  
  Resources
&lt;/h2&gt;

&lt;p&gt;Google's eighth-generation TPUs: two chips for the agentic era&lt;br&gt;
TPU 8t and TPU 8i technical deep dive - Google Cloud Blog&lt;br&gt;
Ironwood: The first Google TPU for the age of inference&lt;br&gt;
Training large models on Ironwood TPUs - Google Cloud Blog&lt;br&gt;
Performance per dollar of GPUs and TPUs for AI inference - Google Cloud Blog&lt;br&gt;
Building production AI on Google Cloud TPUs with JAX&lt;br&gt;
MaxText: LLM reference implementation for TPUs - GitHub&lt;br&gt;
Gemma open-weight LLM library - Google DeepMind GitHub&lt;br&gt;
Serve and Inference Gemma 4 on TPU&lt;br&gt;
Google Cloud unveils eighth-generation TPUs - TechRadar&lt;/p&gt;




&lt;h1&gt;
  
  
  TPUSprint
&lt;/h1&gt;

</description>
      <category>tpu</category>
      <category>gpu</category>
      <category>googlecloud</category>
      <category>googledevs</category>
    </item>
    <item>
      <title>Google Cloud TPU Architecture Versions Explained: From v1 to the Eighth Generation</title>
      <dc:creator>Grace Gong</dc:creator>
      <pubDate>Fri, 01 May 2026 01:44:23 +0000</pubDate>
      <link>https://dev.to/1grace/google-cloud-tpu-architecture-versions-explained-from-v1-to-the-eighth-generation-3pm</link>
      <guid>https://dev.to/1grace/google-cloud-tpu-architecture-versions-explained-from-v1-to-the-eighth-generation-3pm</guid>
      <description>&lt;p&gt;A guide to Cloud TPU generations, what changed between them, and how to choose the right one for your workload&lt;/p&gt;

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2F6bujt63lh73cz1eyg5wg.png" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2F6bujt63lh73cz1eyg5wg.png" alt=" " width="800" height="800"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;Image source: Google Cloud&lt;/p&gt;

&lt;p&gt;If you've looked at Google Cloud TPU pricing or documentation recently, you've probably noticed there are a lot of versions to choose from. TPU v5e, v5p, v6e, Ironwood, now TPU 8t and 8i. Each one has different specs, different use cases, and different tradeoffs. This post walks through every major TPU generation, what changed at each step, and what those changes mean for the people running workloads on them.&lt;/p&gt;




&lt;p&gt;The building blocks: what's inside a TPU chip&lt;br&gt;
Before going generation by generation, it helps to know what the key components are, because the names come up repeatedly across every version.&lt;/p&gt;

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fi556tzdo3gzdimqzcbq2.png" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fi556tzdo3gzdimqzcbq2.png" alt=" " width="800" height="450"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;Image source: Google Cloud&lt;/p&gt;

&lt;p&gt;Matrix Multiply Unit (MXU). This is the core compute engine inside every TPU TensorCore. It performs the multiply-and-accumulate operations that power neural network math. On most TPU generations up through v5p, the MXU is a 128x128 systolic array - 16,384 multiply-accumulators working simultaneously. Starting with Trillium (v6e), the MXU expanded to 256x256, quadrupling the operations per cycle.&lt;/p&gt;

&lt;p&gt;TensorCore. A TensorCore contains one or more MXUs, a vector processing unit (VPU), and a scalar unit. Depending on the generation, a single TPU chip may have one or two TensorCores.&lt;/p&gt;

&lt;p&gt;High Bandwidth Memory (HBM). This is the on-chip memory that stores model weights and activations. HBM capacity and bandwidth are often the real bottleneck for large models, not compute. Each generation has brought more HBM and faster access speeds.&lt;/p&gt;

&lt;p&gt;Inter-Chip Interconnect (ICI). The network that connects chips inside a pod. ICI bandwidth determines how fast chips can synchronize gradients during training. Higher bandwidth means less time waiting for communication and more time computing.&lt;/p&gt;

&lt;p&gt;SparseCore. Introduced in TPU v4, SparseCores are specialized processors for embedding operations - the kind that power recommendation systems and large vocabulary models. v5p and Ironwood have four SparseCores per chip. v6e has two.&lt;/p&gt;

&lt;p&gt;Topology. How chips are wired together in a pod. Earlier generations use a 2D torus (chips connect to four neighbors). Starting with v4, Google moved to a 3D torus for larger-scale pods, which reduces the maximum number of hops between any two chips and cuts communication latency.&lt;/p&gt;




&lt;h2&gt;
  
  
  Generation by generation
&lt;/h2&gt;

&lt;p&gt;TPU v1 (2015, internal only)&lt;br&gt;
The first TPU was built for one purpose: inference. It was not publicly available and could not train models. The chip contained a 256x256 systolic array of 8-bit multiply-accumulators, yielding 92 TOPS of INT8 compute. It consumed about 40 watts, which was remarkably efficient for its time.&lt;br&gt;
Google kept it secret for over a year. When Sundar Pichai announced it at Google I/O 2016, he said it had already been running in Google's data centers for more than a year, powering services like Search, Maps, and Street View. The entire motivation for building it was to avoid doubling Google's data center capacity to handle growing neural network inference demand.&lt;br&gt;
Not available on Google Cloud. Historical context only.&lt;/p&gt;

&lt;h2&gt;
  
  
  TPU v2 (2017)
&lt;/h2&gt;

&lt;p&gt;TPU v2 was the first generation capable of training neural networks. This required a fundamental change: switching from 8-bit integer math to bfloat16, a 16-bit floating-point format that Google invented specifically for this purpose. BF16 retains the same 8-bit exponent as FP32, which gives it the wide dynamic range training needs, while cutting memory use in half compared to FP32.&lt;br&gt;
Each v2 chip delivered approximately 45 TFLOPS of BF16 compute. A single board held four chips. A full v2 Pod contained 512 chips connected by a 2D torus ICI.&lt;br&gt;
TPU v2 is no longer recommended for new workloads but represents an important milestone: it established bfloat16 as a training standard that the broader ML community eventually adopted.&lt;/p&gt;

&lt;h2&gt;
  
  
  TPU v3 (2018)
&lt;/h2&gt;

&lt;p&gt;TPU v3 more than doubled the compute of v2, reaching approximately 420 TFLOPS per chip. To handle the increased power density, Google switched to liquid cooling, the first TPU generation to require it.&lt;br&gt;
Each v3 chip contained two TensorCores, each with a 128x128 MXU. A full v3 Pod scaled to 1,024 chips in a 2D torus. The v3 architecture is described in detail in Google's paper "A Domain Specific Supercomputer for Training Deep Neural Networks."&lt;br&gt;
TPU v3 is documented on Google Cloud and remains available in some configurations, though newer generations offer substantially better performance per dollar.&lt;/p&gt;

&lt;h2&gt;
  
  
  TPU v4 (2021)
&lt;/h2&gt;

&lt;p&gt;TPU v4 was a major architectural shift. Performance more than doubled over v3, and Google moved from a 2D torus to a 3D torus interconnect. In a 3D torus, each chip connects to six neighbors instead of four. For a 4,096-chip pod, this reduces the maximum number of hops between any two chips from roughly 128 to about 48, which meaningfully cuts all-reduce latency during distributed training.&lt;br&gt;
A single v4 Pod contained 4,096 chips. Google also introduced SparseCore in v4, four dedicated processors per chip optimized for embedding-heavy workloads like recommendation models.&lt;br&gt;
TPU v4 has been compared favorably to the NVIDIA A100 in benchmarks: Google's 2023 paper reported TPU v4 running 5 to 87 percent faster on ML workloads depending on the model. TPU v4 also has an inference-optimized variant, v4i, that does not require liquid cooling.&lt;/p&gt;

&lt;h2&gt;
  
  
  TPU v5e (2023)
&lt;/h2&gt;

&lt;p&gt;TPU v5e is Google's cost-optimized fifth-generation chip. The "e" stands for efficient. Where v5p prioritized maximum performance, v5e was designed to minimize cost per inference query and cost per training FLOP for medium-scale jobs.&lt;br&gt;
Each v5e chip contains one TensorCore with four MXUs, a vector unit, and a scalar unit. The chip uses a 2D torus topology and scales to 256 chips per pod. Google returned to air cooling for v5e. HBM capacity is 16 GB per chip with 819 GB/s bandwidth.&lt;br&gt;
TPU v5e delivers 2.5x better price-performance than v4 for inference workloads. It is currently available on Google Cloud and is a practical starting point for teams new to TPUs, especially for serving workloads where cost matters more than peak throughput.&lt;/p&gt;

&lt;h2&gt;
  
  
  TPU v5p (2023)
&lt;/h2&gt;

&lt;p&gt;TPU v5p is the performance-focused fifth-generation chip, released alongside v5e. The "p" stands for performance. Each chip has two TensorCores, each with four MXUs, giving it roughly double the compute of v5e per chip. HBM capacity is 95 GB per chip with 2,765 GB/s bandwidth - nearly 6x v5e's memory capacity.&lt;br&gt;
v5p uses a 3D torus topology and scales to 8,960 chips per pod with 4,800 Gbps of ICI bandwidth per chip. Google reports v5p trains large LLM models 2.8x faster than v4 and includes second-generation SparseCores, which deliver 1.9x better performance than v4's SparseCores for embedding-dense models.&lt;br&gt;
v5p is suited for large training runs where you need the largest possible pod size and maximum per-chip compute. Available on Google Cloud in North America (US East region).&lt;/p&gt;

&lt;h2&gt;
  
  
  TPU v6e, Trillium (2024)
&lt;/h2&gt;

&lt;p&gt;Trillium was announced at Google I/O 2024 and became generally available in late 2024. It is the sixth-generation chip and represents the biggest architectural leap between consecutive generations since v2.&lt;br&gt;
The most significant change: the MXU expanded from 128x128 to 256x256. This quadruples the number of multiply-accumulate operations per cycle. Combined with a higher clock speed, Trillium delivers 4.7x the peak compute of v5e per chip. HBM capacity doubled to 32 GB per chip, and ICI bandwidth doubled to 3,200 Gbps per chip. Trillium is 67% more energy efficient than v5e.&lt;br&gt;
Each v6e chip has one TensorCore with two MXUs (the larger 256x256 array), plus two SparseCores. The topology is a 2D torus scaling to 256 chips per pod, the same footprint as v5e.&lt;br&gt;
On the technical documentation side, Google refers to Trillium as v6e in all APIs and logs. The v6e-8 VM type (all 8 chips attached to a single VM) is optimized specifically for inference, making it easy to serve large models on a single host.&lt;br&gt;
Trillium is available now on Google Cloud in North America (US East), Europe (West), and Asia (Northeast).&lt;/p&gt;

&lt;h2&gt;
  
  
  TPU v7, Ironwood (2025)
&lt;/h2&gt;

&lt;p&gt;Ironwood is the seventh-generation TPU, announced at Google Cloud Next 2025 and generally available since late 2025. It is purpose-built for inference and large-scale training at the scale needed for frontier models.&lt;br&gt;
Key specs per chip: 4,614 FP8 TFLOPS, 192 GB of HBM3E memory, 7.37 TB/s memory bandwidth, 9.6 Tb/s ICI bandwidth. A full superpod contains 9,216 chips delivering 42.5 FP8 ExaFLOPS. That is 4x better performance per chip over Trillium and 10x over TPU v5p.&lt;br&gt;
Ironwood introduced FP8 as a native precision format, which is critical for inference throughput. It uses a dual-chiplet design: each Ironwood chip contains two TensorCores and four SparseCores connected by a high-speed die-to-die interface. The 3D torus topology returns for large pod configurations, with 3D connectivity for pods of 4x4x4 or larger.&lt;br&gt;
Ironwood is also the first TPU generation where Google used AlphaChip, a reinforcement learning tool, to optimize the physical chip layout.&lt;br&gt;
Anthropic's Claude models train and serve on TPUs. As part of a multi-billion dollar agreement, Anthropic committed to access up to one million Ironwood TPUs through Google Cloud.&lt;br&gt;
Available now in North America (Central) and Europe (West).&lt;/p&gt;

&lt;h2&gt;
  
  
  TPU 8t and TPU 8i (eighth generation, coming 2026)
&lt;/h2&gt;

&lt;p&gt;Announced at Google Cloud Next 2026, the eighth generation is the first time Google has split the TPU lineup into two chips with distinct architectures for training and inference.&lt;br&gt;
TPU 8t is built for large-scale pre-training and embedding-heavy workloads. A single superpod holds 9,600 chips with 2 petabytes of shared HBM and 121 FP4 ExaFLOPS of compute, nearly tripling per-pod compute versus Ironwood. ICI bandwidth is 19.2 Tb/s per chip, double Ironwood. The new Virgo Network fabric can link over 134,000 chips in a single data center, and theoretically over 1 million chips across sites. TPUDirect RDMA and TPU Direct Storage bypass the host CPU for data movement, effectively doubling bandwidth for large transfers. Google targets 97% goodput on 8t, meaning 97% of compute cycles go toward actual learning. It delivers 2.7x better performance-per-dollar over Ironwood for large-scale training.&lt;br&gt;
TPU 8i is built for post-training and inference. It scales to 1,152 chips per pod and delivers 11.6 FP8 ExaFLOPS. Each chip carries 288 GB of HBM, more than the 8t training chip, and 384 MB of on-chip SRAM - 3x Ironwood's on-chip SRAM. The reason for more memory on the inference chip: large Mixture-of-Experts models at inference time are memory-bandwidth-bound, not compute-bound. The chip serving tokens needs to stream weights and KV-cache faster than the chip training the model. The 8i uses a Boardfly interconnect that reduces maximum network hops from 16 to 7, which reduces all-to-all latency for MoE routing. The Collectives Acceleration Engine (CAE) replaces Ironwood's SparseCores and cuts collective operation latency by 5x. Google reports 80% better performance-per-dollar over previous generations for low-latency inference on large MoE models.&lt;br&gt;
Both chips run on Google's Axion ARM-based CPU host and use fourth-generation liquid cooling. Both are coming later in 2026.&lt;/p&gt;




&lt;h2&gt;
  
  
  Key architectural trends across generations
&lt;/h2&gt;

&lt;p&gt;A few patterns stand out looking across all eight generations:&lt;/p&gt;

&lt;p&gt;MXU size stayed constant for a long time, then doubled. From v2 through v5p, the MXU was 128x128. Trillium (v6e) expanded it to 256x256. This gave a step-change in throughput rather than incremental gains.&lt;/p&gt;

&lt;p&gt;Topology alternates by use case. Cost-efficient chips (v5e, v6e) use 2D torus topologies, which are simpler and scale well to 256 chips. Performance chips (v4, v5p, Ironwood) use 3D torus, which reduces communication latency at larger pod sizes (4,096 to 9,216 chips).&lt;/p&gt;

&lt;p&gt;Memory capacity has grown dramatically. From 16 GB per chip on v5e to 192 GB on Ironwood to 288 GB on TPU 8i. Memory capacity is increasingly what determines which models you can run and at what batch size.&lt;/p&gt;

&lt;p&gt;Generation 8 split training and inference. Every prior generation was a single chip asked to handle both workloads. The 8t and 8i split acknowledges that the two jobs have fundamentally different hardware requirements.&lt;/p&gt;




&lt;h2&gt;
  
  
  Quick reference: which generation for which use case
&lt;/h2&gt;

&lt;p&gt;Use case Recommended generation Getting started, cost-sensitive inference TPU v5e Medium training jobs, best perf/watt today TPU v6e (Trillium) Large training runs, &amp;gt;256 chips TPU v7 (Ironwood) Frontier model pre-training, 2026 TPU 8t Agentic AI inference, MoE serving, 2026 TPU 8i&lt;/p&gt;




&lt;h2&gt;
  
  
  Code sample: detecting your TPU version and topology in JAX
&lt;/h2&gt;

&lt;p&gt;Once you have a Cloud TPU VM, this snippet uses JAX to confirm the TPU version, count available devices, and print the topology. It runs on any currently available generation (v5e, v6e, Ironwood).&lt;/p&gt;

&lt;p&gt;&lt;code&gt;import jax&lt;br&gt;
import jax.numpy as jnp&lt;/code&gt;&lt;/p&gt;

&lt;p&gt;`# Print the number of TPU devices visible to JAX&lt;br&gt;
print(f"Number of TPU devices: {jax.device_count()}")&lt;br&gt;
print(f"Number of local devices: {jax.local_device_count()}")&lt;/p&gt;

&lt;h1&gt;
  
  
  Print device details including TPU version
&lt;/h1&gt;

&lt;p&gt;devices = jax.devices()&lt;br&gt;
for i, device in enumerate(devices):&lt;br&gt;
    print(f"Device {i}: {device}")&lt;/p&gt;

&lt;h1&gt;
  
  
  Run a simple matrix multiply to confirm the TPU is working
&lt;/h1&gt;

&lt;h1&gt;
  
  
  On Trillium (v6e) and Ironwood, this uses the 256x256 or larger MXU
&lt;/h1&gt;

&lt;p&gt;key = jax.random.PRNGKey(0)&lt;br&gt;
x = jax.random.normal(key, (1024, 1024))&lt;br&gt;
y = jax.random.normal(key, (1024, 1024))&lt;br&gt;
result = jnp.dot(x, y)&lt;br&gt;
print(f"\nMatrix multiply output shape: {result.shape}")&lt;br&gt;
print(f"Result device: {result.devices()}")&lt;/p&gt;

&lt;h1&gt;
  
  
  Check available memory per device
&lt;/h1&gt;

&lt;h1&gt;
  
  
  Useful for estimating what model sizes will fit
&lt;/h1&gt;

&lt;p&gt;for device in jax.devices():&lt;br&gt;
    stats = device.memory_stats()&lt;br&gt;
    if stats:&lt;br&gt;
        total_bytes = stats.get('bytes_limit', 0)&lt;br&gt;
        used_bytes = stats.get('bytes_in_use', 0)&lt;br&gt;
        total_gb = total_bytes / (1024 ** 3)&lt;br&gt;
        used_gb = used_bytes / (1024 ** 3)&lt;br&gt;
        print(f"\nDevice: {device}")&lt;br&gt;
        print(f"  Total HBM: {total_gb:.1f} GB")&lt;br&gt;
        print(f"  In use:    {used_gb:.2f} GB")`&lt;/p&gt;

&lt;p&gt;To run this on a Cloud TPU VM:&lt;br&gt;
&lt;code&gt;# 1. Create a TPU VM (v6e example - swap accelerator-type for other versions)&lt;br&gt;
gcloud compute tpus tpu-vm create my-tpu \&lt;br&gt;
  --zone=us-east1-d \&lt;br&gt;
  --accelerator-type=v6e-8 \&lt;br&gt;
  --version=tpu-ubuntu2204-base&lt;/code&gt;&lt;/p&gt;

&lt;p&gt;`# 2. SSH into the VM&lt;br&gt;
gcloud compute tpus tpu-vm ssh my-tpu --zone=us-east1-d&lt;/p&gt;

&lt;h1&gt;
  
  
  3. Install JAX with TPU support
&lt;/h1&gt;

&lt;p&gt;pip install "jax[tpu]" -f &lt;a href="https://storage.googleapis.com/jax-releases/libtpu_releases.html" rel="noopener noreferrer"&gt;https://storage.googleapis.com/jax-releases/libtpu_releases.html&lt;/a&gt;&lt;/p&gt;

&lt;h1&gt;
  
  
  4. Run the script
&lt;/h1&gt;

&lt;p&gt;python tpu_check.py`&lt;/p&gt;

&lt;p&gt;On a v6e-8 (8 chips, single host), you will see 8 devices listed. On a multi-host slice like v6e-32, each VM sees 8 devices and JAX handles coordination across hosts via the ICI.&lt;/p&gt;




&lt;h2&gt;
  
  
  Summary
&lt;/h2&gt;

&lt;p&gt;Each TPU generation has made a specific set of tradeoffs: v5e optimizes cost, v5p maximizes compute per pod, Trillium doubled the MXU and energy efficiency, Ironwood added native FP8 and massive HBM, and the eighth generation splits the chip entirely for training and inference. Knowing which generation fits your workload is the difference between overpaying for compute you don't need and hitting memory or bandwidth limits on a chip that isn't designed for what you're running.&lt;br&gt;
If you're just getting started, TPU v5e is the easiest and cheapest entry point. For production inference today, Trillium or Ironwood. For the next generation of frontier model training, keep an eye on TPU 8t.&lt;/p&gt;




&lt;h2&gt;
  
  
  Resources
&lt;/h2&gt;

&lt;p&gt;TPU architecture - Google Cloud Documentation&lt;br&gt;
TPU versions overview - Google Cloud Documentation&lt;br&gt;
TPU v5e documentation&lt;br&gt;
TPU v5p documentation&lt;br&gt;
TPU v6e (Trillium) documentation&lt;br&gt;
TPU v7 (Ironwood) documentation&lt;br&gt;
Introducing Trillium, sixth-generation TPUs - Google Cloud Blog&lt;br&gt;
Introducing Cloud TPU v5p and AI Hypercomputer - Google Cloud Blog&lt;br&gt;
Ironwood: The first Google TPU for the age of inference&lt;br&gt;
Google's eighth-generation TPUs: two chips for the agentic era&lt;br&gt;
Run JAX on Cloud TPU VM - Google Cloud Documentation&lt;/p&gt;




&lt;h1&gt;
  
  
  TPUSprint
&lt;/h1&gt;

</description>
      <category>architecture</category>
      <category>deeplearning</category>
      <category>googlecloud</category>
      <category>machinelearning</category>
    </item>
    <item>
      <title>vLLM on Google Cloud TPU: A Model Size vs Chip Cheat Sheet (With Interactive Tool)</title>
      <dc:creator>Grace Gong</dc:creator>
      <pubDate>Thu, 30 Apr 2026 09:38:40 +0000</pubDate>
      <link>https://dev.to/1grace/vllm-on-google-cloud-tpu-a-model-size-vs-chip-cheat-sheet-with-interactive-tool-2c3k</link>
      <guid>https://dev.to/1grace/vllm-on-google-cloud-tpu-a-model-size-vs-chip-cheat-sheet-with-interactive-tool-2c3k</guid>
      <description>&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fhtr4zjn8q3pngkuu1wcv.png" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fhtr4zjn8q3pngkuu1wcv.png" alt="Image source: Google CloudWhich TPU configuration fits your model, what tensor_parallel_size to set, and what it costs per hour" width="800" height="800"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;Picking a Cloud TPU slice for vLLM inference involves three decisions that most tutorials skip over: how much HBM your model actually needs at runtime, how many chips to shard across, and whether the cost is justified for your workload. Get it wrong in either direction and you're either OOMing on startup or paying for memory you're not using.&lt;/p&gt;

&lt;p&gt;This post walks through how to make that decision, with a reference table for popular models and a live interactive tool where you can select your model, toggle precision, and see exactly which TPU configurations fit and what they cost.&lt;/p&gt;

&lt;p&gt;Try the interactive cheat sheet here: ggongg.github.io/vllm-tpu-notes&lt;/p&gt;

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2F50kicybi5cgivuvfxt7o.png" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2F50kicybi5cgivuvfxt7o.png" alt=" " width="800" height="641"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;The site is a mini-project based on data pulled from 4/30/2026 and may change. Please refer to the official docs &amp;amp; site and double check. &lt;/p&gt;

&lt;h2&gt;
  
  
  What is vLLM and why run it on a TPU?
&lt;/h2&gt;

&lt;p&gt;vLLM is an open-source LLM inference engine built for high-throughput, memory-efficient serving. It uses a technique called PagedAttention to manage the KV cache more efficiently than naive implementations, which translates to higher throughput and the ability to serve larger batches from the same hardware.&lt;/p&gt;

&lt;p&gt;It has historically been GPU-first, but vLLM now has first-class support for Google Cloud TPUs, covering v5e, v6e (Trillium), and Ironwood. You can run both offline batch inference and an OpenAI-compatible API server on a Cloud TPU VM with the same command-line interface you'd use on a GPU.&lt;/p&gt;

&lt;h2&gt;
  
  
  Why TPU instead of GPU?
&lt;/h2&gt;

&lt;p&gt;The short answer is cost at scale. For sustained, high-volume LLM inference, TPUs often deliver better performance per dollar than equivalent GPU setups, particularly with Google's newer generations. Trillium (v6e) delivers 4.7x more compute per chip than v5e while being 67% more energy efficient. Ironwood brings 192 GB of HBM per chip, which opens up model sizes and context lengths that simply aren't practical on a single GPU node. For teams already in the Google Cloud ecosystem, the operational simplicity of a managed TPU VM also reduces setup overhead.&lt;/p&gt;

&lt;h2&gt;
  
  
  What the interactive tool does
&lt;/h2&gt;

&lt;p&gt;The cheat sheet at ggongg.github.io/vllm-tpu-notes lets you:&lt;br&gt;
Select a model from Gemma 2B through Llama 405B&lt;br&gt;
Toggle between BF16 and FP8 precision&lt;br&gt;
Switch between on-demand, 1-year CUD, and 3-year CUD pricing&lt;br&gt;
See which TPU slices fit the model, the recommended --tensor-parallel-size for each, and the hourly cost&lt;br&gt;
Get a generated vllm serve command for the cheapest compatible chip&lt;/p&gt;

&lt;p&gt;All pricing is pulled from the official Google Cloud TPU pricing page. Memory estimates use weights × 1.25 to account for KV cache and activation overhead.&lt;/p&gt;

&lt;h2&gt;
  
  
  How much memory does a model actually need?
&lt;/h2&gt;

&lt;p&gt;The naive calculation is parameters × bytes per parameter. A 7B model in BF16 is 7B × 2 bytes = 14 GB. But that only covers the weight footprint. At inference time, you also need headroom for:&lt;br&gt;
KV cache, which grows with batch size and context length&lt;br&gt;
Activations during the forward pass&lt;br&gt;
vLLM's internal buffers&lt;/p&gt;

&lt;p&gt;A practical floor is weights × 1.25. A 7B BF16 model needs roughly 17.5 GB, which means a single v5e chip (16 GB HBM) will OOM, but a v5e-4 slice (64 GB across 4 chips) fits comfortably.&lt;br&gt;
FP8 halves the weight footprint, so the same 7B model drops to about 8 GB in FP8, leaving far more room for KV cache on a given slice. Trillium (v6e) and Ironwood both support native FP8. TPU v5e supports INT8 quantization via vLLM's tpu_int8 flag instead.&lt;/p&gt;

&lt;h2&gt;
  
  
  TPU chip HBM and pricing reference
&lt;/h2&gt;

&lt;p&gt;These are the currently available Cloud TPU slices for vLLM inference. Per-chip pricing is from cloud.google.com/tpu/pricing, US region, as of April 30, 2026. Slice prices are chips × per-chip rate.&lt;/p&gt;

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Felwcs2pyk703i83uj737.png" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Felwcs2pyk703i83uj737.png" alt=" " width="800" height="310"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;h2&gt;
  
  
  Model size vs chip: the cheat sheet
&lt;/h2&gt;

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fvp6s72i0artudmdt61a5.png" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fvp6s72i0artudmdt61a5.png" alt=" " width="800" height="427"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;h2&gt;
  
  
  The key vLLM flags for TPU
&lt;/h2&gt;

&lt;p&gt;vllm serve MODEL_ID \&lt;br&gt;
  --tensor-parallel-size TP \&lt;br&gt;
  --max-model-len MAX_TOKENS \&lt;br&gt;
  --dtype bfloat16 \&lt;br&gt;
  --device tpu&lt;br&gt;
--tensor-parallel-size sets how many chips the model is sharded across. It needs to be a power of 2 (1, 2, 4, or 8) and match your slice size or be a divisor of it.&lt;br&gt;
--max-model-len caps the context window, which controls KV cache memory usage. If you're OOMing on a slice that should theoretically fit the model, reduce this first before scaling up to a larger slice.&lt;br&gt;
--dtype bfloat16 is the right default for TPU. For native FP8 on v6e or Ironwood, use --dtype fp8 instead.&lt;br&gt;
--device tpu tells vLLM to use the TPU backend.&lt;/p&gt;

&lt;h2&gt;
  
  
  Example commands
&lt;/h2&gt;

&lt;p&gt;Gemma 7B on v5e-4&lt;br&gt;
vllm serve google/gemma-7b-it \&lt;br&gt;
  --tensor-parallel-size 1 \&lt;br&gt;
  --max-model-len 4096 \&lt;br&gt;
  --dtype bfloat16 \&lt;br&gt;
  --device tpu&lt;br&gt;
16 GB weights fit on a single v5e chip. With 64 GB total on a v5e-4, there is plenty of room for KV cache at reasonable context lengths.&lt;br&gt;
Llama 3.1 70B on v6e-8&lt;br&gt;
vllm serve meta-llama/Llama-3.1-70B-Instruct \&lt;br&gt;
  --tensor-parallel-size 4 \&lt;br&gt;
  --max-model-len 8192 \&lt;br&gt;
  --dtype bfloat16 \&lt;br&gt;
  --device tpu&lt;br&gt;
140 GB of weights need roughly 175 GB with overhead. The v6e-8 has 256 GB total across 8 chips, so TP4 puts 64 GB on each of 4 chips and leaves headroom for the KV cache.&lt;br&gt;
Llama 3.1 8B on v6e with FP8&lt;br&gt;
vllm serve meta-llama/Llama-3.1-8B-Instruct \&lt;br&gt;
  --tensor-parallel-size 1 \&lt;br&gt;
  --max-model-len 8192 \&lt;br&gt;
  --dtype fp8 \&lt;br&gt;
  --device tpu&lt;br&gt;
FP8 brings the 8B model down to about 8 GB, fitting easily on a single v6e chip (32 GB) and leaving 24 GB for KV cache. This is a good setup for high-concurrency serving where you want to maximize batch size.&lt;br&gt;
Gemma 4 31B on v6e-8&lt;br&gt;
docker run -itd --name gemma4-tpu \&lt;br&gt;
  --privileged --network host --shm-size 16G \&lt;br&gt;
  -v /dev/shm:/dev/shm \&lt;br&gt;
  -e HF_TOKEN=$HF_TOKEN \&lt;br&gt;
  vllm/vllm-tpu:gemma4 \&lt;br&gt;
  --model google/gemma-4-31B-it \&lt;br&gt;
  --tensor-parallel-size 8 \&lt;br&gt;
  --max-model-len 16384 \&lt;br&gt;
  --device tpu&lt;br&gt;
Gemma 4 has a dedicated Docker image (vllm/vllm-tpu:gemma4) that includes multimodal support. For the 31B variant on a v6e-8, TP8 across all 8 chips is the recommended configuration.&lt;/p&gt;

&lt;h2&gt;
  
  
  Provisioning a TPU VM for vLLM
&lt;/h2&gt;

&lt;h1&gt;
  
  
  Create a v6e-8 VM
&lt;/h1&gt;

&lt;p&gt;gcloud compute tpus tpu-vm create my-vllm-tpu \&lt;br&gt;
  --zone=us-east1-d \&lt;br&gt;
  --accelerator-type=v6e-8 \&lt;br&gt;
  --version=tpu-ubuntu2204-base&lt;/p&gt;

&lt;h1&gt;
  
  
  SSH in
&lt;/h1&gt;

&lt;p&gt;gcloud compute tpus tpu-vm ssh my-vllm-tpu --zone=us-east1-d&lt;/p&gt;

&lt;h1&gt;
  
  
  Install vLLM TPU
&lt;/h1&gt;

&lt;p&gt;pip install vllm-tpu&lt;br&gt;
The v6e-8 configuration is the one Google designed specifically for single-host inference. For Ironwood, you need to use GKE rather than the TPU VM API. Step-by-step instructions for both are in the vLLM TPU quickstart.&lt;/p&gt;




&lt;h2&gt;
  
  
  Picking the right configuration
&lt;/h2&gt;

&lt;p&gt;v5e-4 or v5e-8 makes sense if you're running models up to 27B and want the lowest cost per hour. At $1.20/chip/hr, it's the cheapest path to TPU inference. The 16 GB per chip limits what fits, but it covers the 7B to 13B range that handles most practical use cases.&lt;br&gt;
v6e-4 or v6e-8 (Trillium) is the best general-purpose option today. The 32 GB per chip opens up models to 70B with enough chips, and the 4.7x compute improvement over v5e gives meaningfully better throughput for the same money. The v6e-8 in particular was designed by Google for single-host inference and is the easiest starting point for production workloads.&lt;br&gt;
Ironwood is the right choice when you genuinely need the memory: 70B at long context, 405B models, or high-concurrency serving where maximizing KV cache capacity matters. At $12/chip/hr, it's worth confirming you actually need the 192 GB per chip before provisioning it.&lt;/p&gt;




&lt;h2&gt;
  
  
  Try it yourself
&lt;/h2&gt;

&lt;p&gt;The interactive tool is live at ggongg.github.io/vllm-tpu-notes. Select your model, toggle precision and pricing tier, and get the recommended chip configuration and vLLM command in one view. The source is on GitHub at github.com/ggongg/vllm-tpu-notes.&lt;/p&gt;




&lt;h2&gt;
  
  
  Resources
&lt;/h2&gt;

&lt;p&gt;Interactive cheat sheet&lt;br&gt;
vLLM TPU quickstart&lt;br&gt;
vLLM TPU installation&lt;br&gt;
Cloud TPU pricing&lt;br&gt;
TPU v6e documentation&lt;br&gt;
TPU v7 (Ironwood) documentation&lt;br&gt;
Gemma 4 vLLM recipe&lt;br&gt;
vLLM TPU offline inference example&lt;/p&gt;




&lt;h1&gt;
  
  
  TPUSprint program.
&lt;/h1&gt;

&lt;p&gt;Check cloud.google.com/tpu/pricing for most up to date info! &lt;/p&gt;

</description>
      <category>tpu</category>
      <category>googlecloud</category>
      <category>googledevs</category>
      <category>tpusprint</category>
    </item>
    <item>
      <title>Ill be speaking at the 2026 Grace Hopper Celebration (GHC)!</title>
      <dc:creator>Grace Gong</dc:creator>
      <pubDate>Thu, 30 Apr 2026 06:59:02 +0000</pubDate>
      <link>https://dev.to/1grace/ill-be-speaking-at-the-2026-grace-hopper-celebration-ghc-41p8</link>
      <guid>https://dev.to/1grace/ill-be-speaking-at-the-2026-grace-hopper-celebration-ghc-41p8</guid>
      <description>&lt;p&gt;GHC 2026&lt;br&gt;
I’m Speaking at Grace Hopper Celebration 2026!&lt;/p&gt;

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2F4qgxfppm3b7a5pix58sh.png" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2F4qgxfppm3b7a5pix58sh.png" alt=" " width="720" height="378"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;My path into VC hasn’t been a straight line. As an engineer at Microsoft and a Venture Fellow at Alumni Ventures, I’ve had the privilege of sitting at the intersection of technology and investment learning pitches, understanding what investors are actually looking for through the fellowship training, and learning from venture capital coursework through Berkeley Law, NVCA, and VCForward. Alongside that, I’m completing coursework from Georgia Tech and the University of Illinois, where technical and business strategy have added a new lens to how I think about building and funding companies.&lt;/p&gt;

&lt;p&gt;GHC felt like the natural place to bring these learnings together because this community deserves a seat at the table in these conversations.&lt;/p&gt;

&lt;p&gt;What You’ll Learn at My Session&lt;/p&gt;

&lt;p&gt;Whether you’re a founder looking to raise, an engineer curious about the business side, or someone who’s just VC-curious, this session is designed to be approachable and practical. We’ll cover:&lt;/p&gt;

&lt;p&gt;How investors evaluate deals and what they’re really looking for beyond the pitch deck&lt;br&gt;
The frameworks VCs use to make funding decisions&lt;br&gt;
How you as a technologist or founder can navigate the VC ecosystem with confidence&lt;br&gt;
No finance background required.&lt;/p&gt;

&lt;p&gt;Why This Matters&lt;/p&gt;

&lt;p&gt;Women remain significantly underrepresented both as founders receiving VC funding and as investors making funding decisions. One of the best ways to change that is education and access: knowing the language, the process, and the people. I hope this session is a small step toward that.&lt;/p&gt;

&lt;p&gt;Join Me in Anaheim&lt;/p&gt;

&lt;p&gt;GHC 2026 runs October 27–30 in Anaheim, CA. If you’re attending, I’d love to see you at my session!&lt;/p&gt;

&lt;p&gt;🔗 Register: GHC.AnitaB.org&lt;/p&gt;

&lt;p&gt;Let’s connect too, find me on Twitter/X and LinkedIn&lt;/p&gt;

</description>
      <category>career</category>
      <category>community</category>
      <category>learning</category>
      <category>softwareengineering</category>
    </item>
    <item>
      <title>I'll be speaking at Major League Hacking (MLH) Hackcon 2024!</title>
      <dc:creator>Grace Gong</dc:creator>
      <pubDate>Tue, 13 Aug 2024 16:52:23 +0000</pubDate>
      <link>https://dev.to/1grace/ill-be-speaking-at-major-league-hacking-mlh-hackcon-2024-5ge5</link>
      <guid>https://dev.to/1grace/ill-be-speaking-at-major-league-hacking-mlh-hackcon-2024-5ge5</guid>
      <description>&lt;p&gt;Hi everyone! This year I'll be speaking at Major League Hacking (MLH) Hackcon 2024 on "How to Build a Campus Community!" &lt;br&gt;
I'm excited to share learnings and a high level rundown of the event in a future post. &lt;/p&gt;

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fs33td5ikl6pb3je3omvz.png" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fs33td5ikl6pb3je3omvz.png" alt="Image description" width="800" height="420"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;More info on Hackcon: &lt;a href="https://hackcon.mlh.io/" rel="noopener noreferrer"&gt;https://hackcon.mlh.io/&lt;/a&gt; &lt;/p&gt;

&lt;p&gt;&lt;a href="https://www.youtube.com/watch?v=x1--JYliDpg" rel="noopener noreferrer"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;Join Hundreds of Hackathon Organizers.&lt;br&gt;
Hackcon brings together 300+ hackathon organizers and campus tech leaders for two days of interactive sessions on community building, event planning, and the hacker experience. Come for the learning, stay for the camp activities! &lt;/p&gt;

&lt;p&gt;Meet Community Leaders&lt;br&gt;
Hackcon is a unique opportunity to meet like-minded individuals from across the country. It’s where friendships are formed, partnerships are started, and tech leaders meet all together in-person one last time before they go off into the real world.&lt;/p&gt;

&lt;p&gt;Enjoy the Great Outdoors&lt;br&gt;
Unplug from your devices and enjoy the beautiful scenery of upstate New York. Enjoy a range of fun outdoor activities like swimming, s’mores, and tie dye! By the end of the event, you’ll feel relaxed and recharged for the new school year. &lt;/p&gt;

&lt;p&gt;Learn from the Best&lt;br&gt;
There is no better place to learn how to organize a hackathon and run a community on campus. At Hackcon, you’ll have the chance to meet like-minded individuals and hear from the most experienced leaders in our community.&lt;/p&gt;

&lt;p&gt;Join us at Camp Pontiac&lt;br&gt;
Located in Copake, New York, Camp Pontiac is an active summer camp that becomes the venue for Hackcon. This scenic location creates the perfect atmosphere for community talks, discussion groups, and lasting friendships.&lt;/p&gt;

&lt;p&gt;Get Tickets here: &lt;a href="https://ti.to/mlh/mlh-hackcon-2024" rel="noopener noreferrer"&gt;https://ti.to/mlh/mlh-hackcon-2024&lt;/a&gt; &lt;/p&gt;

</description>
      <category>mlhacks</category>
      <category>hackcon</category>
      <category>majorleaguehacking</category>
      <category>hackathon</category>
    </item>
    <item>
      <title>Making Chatbot Applications with Dialogflow: My HackItTogether 2024 Project</title>
      <dc:creator>Grace Gong</dc:creator>
      <pubDate>Sat, 20 Apr 2024 20:41:14 +0000</pubDate>
      <link>https://dev.to/1grace/making-chatbot-applications-with-dialogflow-with-project-example-5dp7</link>
      <guid>https://dev.to/1grace/making-chatbot-applications-with-dialogflow-with-project-example-5dp7</guid>
      <description>&lt;p&gt;I participated in a hackathon in nyc earlier this term! As a highlight, I was able to work with a team of students from brooklyn to create a chatbot. We were given the challenge and objective to work on an Individualized Recommendation System that leverages AI&lt;br&gt;
techniques to understand users' preferences and recommend similar content. &lt;/p&gt;

&lt;p&gt;I conducted research on various Google Cloud Services to help support this project, and ultimately decided to work on a chatbot using Dialogflow to help achieve this. &lt;/p&gt;

&lt;p&gt;Things that I considered during our project: &lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;Architecture for the chat, db &amp;amp;/api &lt;/li&gt;
&lt;li&gt;If we want to use dialog flow / azure openai as a templated UI for the chat or coding the interface ? (which would be time consuming
GCP pros: we had credits)
&lt;/li&gt;
&lt;li&gt;Which API’s to use, and if we use multiple, how to consolidate/aggregate the data in one unified chat &lt;/li&gt;
&lt;li&gt;Which UI to use for the website, and how to customize it? A premade template would be ideal &lt;/li&gt;
&lt;li&gt;Which data sources we want to take and how to integrate this data &lt;/li&gt;
&lt;li&gt;Are we enforcing guardrails (tell it to not give info) only picking from specific sections of info and prevent hallucinations &lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;The other members in my group used other api's such as spotify api to help provide traditional recommendations to a search bar. &lt;br&gt;
The earliest iteration of our UI looked like this: &lt;/p&gt;

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fbc2c2damyoba9c25xh8k.png" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fbc2c2damyoba9c25xh8k.png" alt="Image description" width="800" height="613"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2F89wd23pwoycywrf2i6y9.png" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2F89wd23pwoycywrf2i6y9.png" alt="Image description" width="800" height="533"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fw94jjmrgkkt0hxscxzx7.png" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fw94jjmrgkkt0hxscxzx7.png" alt="Image description" width="800" height="427"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;The Dialogflow CX documentation was simple and easy to follow - I added our data sources through a combination of docs, website links, csv and files. After, I tested the chat application by asking questions to ensure the data was retrieved and provided adequate responses. &lt;/p&gt;

&lt;p&gt;The Chatbot Application was simple to integrate into our React UI -&amp;gt; it appears on the bottom right corner of our site. &lt;/p&gt;

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Ftwsvpne3nuegk1hx3ttl.png" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Ftwsvpne3nuegk1hx3ttl.png" alt="Image description" width="800" height="593"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;While the cloud credits have paused so the chat might not be loading at the moment, you can still see the UI for our project here: &lt;a href="https://your-north-star.netlify.app/" rel="noopener noreferrer"&gt;https://your-north-star.netlify.app/&lt;/a&gt; {UI done in react} &lt;/p&gt;

&lt;p&gt;&lt;strong&gt;Inspiration&lt;/strong&gt;&lt;br&gt;
The North Star has been a symbol of guidance and direction for centuries. In ancient times, navigators and travelers used the North Star as a way to determine their location and navigate the seas. We think that our web-based project with the help of AI will help the users navigate the quiz to find their way to their favorite book, movie or music.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;What it does&lt;/strong&gt;&lt;br&gt;
The website will present a welcoming page to “Your North Star” ChatBot where it asks the user what they’re looking for. They’re presented with three buttons: music, book and movie. This leads to the next question which allows the user to specify the genre of their chosen media. The input given by the user is fed to an API. The APIs return the recommendations as a list, where each recommendation provides metadata, such as: title, cover, rating, overview, etc. We chose the constellation-theme because this is meant to be a journey to imagination. Your taste will bring you to the destination.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;Challenges we ran into&lt;/strong&gt;&lt;br&gt;
The challenges we faced were with the implementation of a chatbot into the website.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;Accomplishments that we're proud of&lt;/strong&gt;&lt;br&gt;
Being able to successfully implement the APIs from Google Cloud and Spotify Web.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;What we learned&lt;/strong&gt;&lt;br&gt;
We learned how to customize APIs; how to navigate on different errors when using javascript or react. Most importantly we learned how to negotiate between team members and try to make our best based on the time provided.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;What's next for Your North Star&lt;/strong&gt;&lt;br&gt;
By using AI we can explore the characters of the books/movies to elevate the user's experience to the next level. The user can communicate the characters and be friends with them.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;Built With&lt;/strong&gt;&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;chatbot-ui&lt;/li&gt;
&lt;li&gt;css&lt;/li&gt;
&lt;li&gt;dialogflow-es&lt;/li&gt;
&lt;li&gt;figma&lt;/li&gt;
&lt;li&gt;github&lt;/li&gt;
&lt;li&gt;google-books-api&lt;/li&gt;
&lt;li&gt;google-cloud&lt;/li&gt;
&lt;li&gt;html&lt;/li&gt;
&lt;li&gt;imdb-api&lt;/li&gt;
&lt;li&gt;javascript&lt;/li&gt;
&lt;li&gt;omdb-api&lt;/li&gt;
&lt;li&gt;react&lt;/li&gt;
&lt;li&gt;spotify&lt;/li&gt;
&lt;li&gt;vertexai-api&lt;/li&gt;
&lt;/ul&gt;

</description>
      <category>googlecloud</category>
      <category>dialogflow</category>
      <category>chatbot</category>
      <category>virtualagents</category>
    </item>
    <item>
      <title>GitHub Field Day USA 2024: Microsoft Reactor New York City</title>
      <dc:creator>Grace Gong</dc:creator>
      <pubDate>Sun, 11 Feb 2024 06:54:59 +0000</pubDate>
      <link>https://dev.to/1grace/github-field-day-usa-2024-microsoft-reactor-new-york-city-k1</link>
      <guid>https://dev.to/1grace/github-field-day-usa-2024-microsoft-reactor-new-york-city-k1</guid>
      <description>&lt;p&gt;Attention all Students in the United States and Canada! &lt;br&gt;
GitHub Field Day USA 2024 will be taking place on March 15, 2024 at Microsoft Reactor New York City. &lt;br&gt;
Please Apply here: &lt;a href="https://githubfieldday.com/nyc2024/" rel="noopener noreferrer"&gt;https://githubfieldday.com/nyc2024/&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;An unconference for leaders of technical student communities&lt;br&gt;
Friday, March 15th, 2024 at the Microsoft Reactor in New York City!&lt;/p&gt;

&lt;p&gt;Hosted by GitHub Education&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;A day of discussions&lt;/strong&gt;&lt;br&gt;
Field Day brings together a group of student leaders for a day of conversation in the form of lightning talks, discussions, games, or whatever suits the topic best.&lt;br&gt;
**&lt;br&gt;
Schedule**&lt;br&gt;
10:00 am    Registration Opens&lt;br&gt;
11:00 am    Welcome to Field Day&lt;br&gt;
11:20 am    Keynote&lt;br&gt;
11:50 pm    Icebreakers&lt;br&gt;
12:30 pm    Talks/Discussions 1&lt;br&gt;
1:05 pm Talks/Discussions 2&lt;br&gt;
1:40 pm Talks/Discussions 3&lt;br&gt;
2:10 pm Lunch&lt;br&gt;
3:05 pm Talks/Discussions 4&lt;br&gt;
3:40 pm Talks/Discussions 5&lt;br&gt;
4:15 pm Closing Remarks&lt;br&gt;
4:30 pm Goodbye + Networking Event&lt;/p&gt;

&lt;p&gt;Topics&lt;br&gt;
Inclusivity in your community&lt;br&gt;
Internships&lt;br&gt;
Building your community&lt;br&gt;
Inter-community connections&lt;br&gt;
Finding events near you&lt;br&gt;
Open source&lt;br&gt;
Apply to contribute your discussion topic!&lt;br&gt;
FAQ&lt;br&gt;
Who can attend Field Day?&lt;br&gt;
Field Day is for student leaders of technical communities. Most attendees will be undergraduate students and young professionals.&lt;/p&gt;

&lt;p&gt;What's a student leader?&lt;br&gt;
A student leader is someone who's actively involved in improving their local technical community. They might be the president or chairperson of their school's tech club or be one of the organizers of the annual hackathon. They might run the student hardware group or be actively involved in introducing other students to programming.&lt;/p&gt;

&lt;p&gt;What's an unconference?&lt;br&gt;
A loosely structured conference emphasizing the informal exchange of information and ideas between participants (to quote the Google definition). Most content at Field Day will be in the form of lightning talks and discussions conducted by attendees.&lt;/p&gt;

&lt;p&gt;I'm not in New York, but this sounds cool! Can I come?&lt;br&gt;
If you're based in North America, you're welcome to apply!&lt;/p&gt;

&lt;p&gt;So I registered to attend, what happens now?&lt;br&gt;
We're just screening applications to make sure you're eligible to attend, you should hear from us within a few days. Once confirmed, all you need to do is show up at the venue that will be sent to you on the day of the event and take part!&lt;/p&gt;

&lt;p&gt;How much does Field Day cost?&lt;br&gt;
Field Day is free to attend!&lt;/p&gt;

&lt;p&gt;Is financial aid for attending the event available?&lt;br&gt;
We expect to have travel reimbursements available for students in areas surrounding New York City, we encourage you to apply. We will be notifying accepted students with the status of their travel reimbursemens within one week after acceptances are released.&lt;/p&gt;

&lt;p&gt;What COVID restrictions are there?&lt;br&gt;
We'll be following local guidance and will announce more specific details closer to the event.&lt;/p&gt;

&lt;p&gt;Will there be food?&lt;br&gt;
Yes! We will be serving a light breakfast and lunch! You will not go hungry.&lt;/p&gt;

&lt;p&gt;Will there be swag?&lt;br&gt;
Yes!&lt;/p&gt;

&lt;p&gt;When can I find out if I'm receiving financial aid?&lt;br&gt;
We hope to send out confirmations of travel reimbursements within 1 week of acceptance.&lt;/p&gt;

&lt;p&gt;Is there a Code of Conduct for this event?&lt;br&gt;
Yes! You can view it here.&lt;/p&gt;

&lt;p&gt;Who is organizing Field Day?&lt;br&gt;
Field Day is organized by GitHub Campus Experts, and technical student leaders. &lt;/p&gt;

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2F6igf3paa2oxwdvcxa65m.jpeg" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2F6igf3paa2oxwdvcxa65m.jpeg" alt="Image description" width="800" height="794"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;&lt;u&gt;About me &lt;/u&gt;&lt;/strong&gt;&lt;br&gt;
My name is Grace, and I’m a computer science student and GitHub Campus Expert for Western University based in Canada. As a GitHub Campus Expert and Developer Student Club Lead, we organize technical and career development events, creating opportunities for our communities to leverage GitHub and Student Developer Pack. We strive to build diverse and inclusive spaces to learn skills, share our experiences, and build projects together. Some of our flagship events that we host at Hackathons include Intro to Cloud, ML API's Workshops, and GitHub Workshops, where we help students get started with GitHub and share the GitHub Student Developer Pack.  &lt;/p&gt;

&lt;p&gt;Earlier this November, I had the opportunity to go to GitHub Universe in person in San Francisco, thanks to the GitHub Education team. &lt;br&gt;
We met at the GitHub San Francisco headquarters, where we did a tour of the HQ, and attended Sessions at the center. &lt;/p&gt;

&lt;p&gt;Throughout the Semesters, I collaborate with tech clubs across campus at Western University to organize workshops on GitHub, as well as other technical topics. We share resources, share events with our broader communities, and aim to inform and educate students so they can utilize these skills to work on impactful projects in and outside of school. &lt;/p&gt;

</description>
      <category>github</category>
      <category>githubuniverse</category>
      <category>githubuniverse2023</category>
      <category>opensource</category>
    </item>
    <item>
      <title>How I became a GHC (Grace Hopper Conference) Speaker, from a 2023 GHC Speaker</title>
      <dc:creator>Grace Gong</dc:creator>
      <pubDate>Sat, 23 Dec 2023 02:36:52 +0000</pubDate>
      <link>https://dev.to/1grace/how-i-became-a-ghc-grace-hopper-conference-speaker-from-a-2023-ghc-speaker-3k56</link>
      <guid>https://dev.to/1grace/how-i-became-a-ghc-grace-hopper-conference-speaker-from-a-2023-ghc-speaker-3k56</guid>
      <description>&lt;p&gt;I’m excited to have spoken this year at the Grace Hopper Conference as a computer science student! My talk was focused on empowering individuals to build a campus community. I shared my experience starting the women in computer science committee and how I leverage the resources from partner organizations, clubs, and my school to create compelling, engaging, and valuable events for the student population. &lt;br&gt;
I also went through how I marketed the events and got hundreds of students to attend these events this year and join the community. &lt;/p&gt;

&lt;p&gt;I will try to keep this article concise, short, and sweet in honor of GHC approaching - I know everyone is preparing for exams or the upcoming holiday season, and if anything, I hope this article can potentially be a catalyst for future conversations at the event and beyond! &lt;/p&gt;

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fz1yaw559fnovrjofwo5v.png" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fz1yaw559fnovrjofwo5v.png" alt="Image description" width="800" height="212"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;My process was a bit unique, and I hope it can provide some insights for others interested in getting involved in the GHC community. (This is virtual, and I am sure there will be other requirements for the person) &lt;/p&gt;

&lt;p&gt;I had just finished my sophomore year of CS and realized there was no significant in-person presence at my campus. On campus, I mean, when you walk up to the doors of the computer science building, there is a rush of people immediately, but there is no commonplace that you can call "community.” The classrooms upstairs are mostly unknown to most students, with the exception of labs, and the space itself feels foreign even after having spent years here. &lt;/p&gt;

&lt;p&gt;There was a new room where people could study in groups, but it holds probably 20 people max and is usually filled with students on a time crunch doing group projects or empty throughout the day, and no significant space is created for those who are in computer science and want to meet others in the community. &lt;/p&gt;

&lt;p&gt;Some clubs would hold virtual events and generally fall under the area of technology, but for computer science specifically, there was no common community. Noticing that there were many women in CS clubs on campuses across Canada, such as Waterloo and UBC, I wanted to bring this energy to my university specifically with a focus on collaboration with the school faculty for increased sponsorship, and applied for the AnitaB Pass It On Grant. &lt;/p&gt;

&lt;p&gt;As a DSC (Developer Student Club) lead, I believe this strong partnership between our committees would enable us to deliver even more impact, and it for sure did - that year, we were able to organize numerous workshops with corporate partners like Salesforce on Technical, Behavioural DSA, and General Interviews, and give students a community to learn from each other. We organized an International Women’s Day event in partnership with the Ivey Business School and their Hackathon event in March, where we were able to support students in navigating the new normal as emerging technologies continue to evolve - students were able to learn from Ada, a TPM at Google in ML, as well as other Developers from the Google Developer Community to facilitate workshops on Flutter, a cross-platform mobile application development SDK which enabled students to develop their skills during the challenge. &lt;/p&gt;

&lt;p&gt;I also had multiple discussions with the chair of the department of computer science, the chair of the faculty of science, where we discussed the experiences of marginalized students in science and computer science and how we could partner to improve the student experience. &lt;/p&gt;

&lt;p&gt;I was the campus lead for the Percentage Project, a data-driven campaign to &lt;/p&gt;

&lt;p&gt;This is among the many events we had during the school year with DSC, and I’m proud of our progress as the membership grew from 100 to over 450 members. &lt;/p&gt;

</description>
      <category>ghc</category>
      <category>gracehopperconference</category>
    </item>
    <item>
      <title>My GitHub Universe 2023 Experience with GitHub Campus Experts &amp; GitHub Education</title>
      <dc:creator>Grace Gong</dc:creator>
      <pubDate>Fri, 10 Nov 2023 08:16:17 +0000</pubDate>
      <link>https://dev.to/1grace/pre-github-universe-day-34i3</link>
      <guid>https://dev.to/1grace/pre-github-universe-day-34i3</guid>
      <description>&lt;p&gt;Attending GitHub Universe 2023 was amazing!&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;Whats GitHub Universe?&lt;/strong&gt;&lt;br&gt;
GitHub Universe brings together the world's developers, enterprise leaders, and security professionals for two days of interactive sessions on AI, security, and the developer experience. Come for the learning, stay for the fun. We'll see you there!&lt;br&gt;
&lt;strong&gt;How Did you get the opportunity to attend?&lt;/strong&gt;&lt;br&gt;
As a GitHub Campus Expert, this year I had the opportunity to attend GitHub Universe in person and the sessions and discussions I had there were impactful, inspiring, and a great testament to the incredible work and progress the Github Community has put together to improve the developer experience this past year.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;How was the experience?&lt;/strong&gt;&lt;br&gt;
As my friend Harsh from OurTechCommunity (OTC) said - All the Campus Experts were so inspiring! It was so good to talk about each other's hardships openly, hear about their efforts and goals, see their drive and hard work, and learn from their experiences and knowledge.&lt;br&gt;
We were also able to learn from the overall tech community from Universe, whether it be about Open Source, or what other companies are doing to adopt and adapt to the current evolving tech landscape. I am so grateful for GitHub, the GitHub Education Team, and everyone who worked to make this possible. Big Thank you to Luis, Juan, Dyson, Stormy Peters and everyone else who worked to make us, as students feel heard, and give us spaces to learn and bring our learnings to our overall communities.&lt;/p&gt;

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Ffosi1zxhmv5kdoael9t3.png" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Ffosi1zxhmv5kdoael9t3.png" alt="Image description" width="800" height="593"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2F478cbf4ubzrsa3fydid6.png" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2F478cbf4ubzrsa3fydid6.png" alt="Image description" width="800" height="533"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;What activities did you attend?&lt;/strong&gt;&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;GitHub Open Source Pre-event Networking @ Minna Art Gallery&lt;/li&gt;
&lt;li&gt;GitHub Platform Engineering Talk @ MoMa&lt;/li&gt;
&lt;li&gt;Telus GitHub Actions Talk @ MoMa&lt;/li&gt;
&lt;li&gt;Developer Experience @ Discussion Lounge&lt;/li&gt;
&lt;li&gt;VSCode Productivity @ Discussion Lounge&lt;/li&gt;
&lt;li&gt;Observability Solutions @ Discussion Lounge&lt;/li&gt;
&lt;li&gt;Education Panel @ Hyatt&lt;/li&gt;
&lt;li&gt;Ignacio's Panel on AI In Education&lt;/li&gt;
&lt;li&gt;Keynotes, Chat with Stormy + Martin&lt;/li&gt;
&lt;li&gt;Booths: Microsoft, Datadog, Postman, Arm, GitHub CoPilot, Security,&lt;/li&gt;
&lt;li&gt;Demos: GitHub Security, CoPilot&lt;/li&gt;
&lt;li&gt;GitHub Certification&lt;/li&gt;
&lt;li&gt;GitHub Women in Tech Event&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;&lt;strong&gt;What did you learn from attending GitHub Universe?&lt;/strong&gt;&lt;br&gt;
I learned additional ways to support and engage my community and how to uplift marginalized/underrepresented communities. Specifically, through the panels on specific GitHub Actions and Features that I can now talk about in the workshops and at my internship companies who use GHE, specific actions I can take to expand and make my community more consistently engaged.&lt;br&gt;
Through conversations with those in open source (leading communities or as an active contributor), I was able to learn about current open source projects, how to get involved, and how OSPO's work(An open source program office (OSPO) serves as the center of competency for an organization's open source operations and structure.) I was able to learn about the operational / logistics side of these groups and how to keep these communities alive and engaged. The key themes continued to tie back to Developer Experience and how understanding the Developer + being able to take steps to make it easy for developers to know how and when to contribute is often key to success in maintaining or making it accessible for the community.&lt;br&gt;
The GitHub Women in Tech event enabled me to learn how to support and uplift other marginalized genders, create safe spaces for discussion, and continue to be involved and immersed in my community. This conference was life-changing because it enabled me to be connected with inspiring women at GitHub + Beyond and from various walks of life. I learned from women of diverse backgrounds and their experiences/strategies in the workforce. I also learned about the struggles they faced in the process and some things I need to look out for / consider in my career.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;Who did you meet?&lt;/strong&gt;&lt;br&gt;
Story Peters, Martin Woodward, Michael Hanley, Joseph K, Ingrid Yan @github Security, VPs of Product, PM @ GitHub Sponsors, Anjuan from GitHub Sponsors, Thomas Dohmke, GitHub Education Team, Datadog Employees (PM, PMM, Sales), Open Source Contributors (CS50, Huan Li, Employees from Slack that work on Electron, etc) David Malan, Demetris Cheatham,(Chief of Staff, GitHub) and of course, the Campus Experts! Workshop Speaker-&amp;gt; Telus, Tech lead from Platform Engineering Talk (Leslie &lt;a class="mentioned-user" href="https://dev.to/ny"&gt;@ny&lt;/a&gt; Times)&lt;br&gt;
-The Women in Tech workshop had various women from students my age (I wasn't able to get her name but she is a masters at UWashington doing a PhD to senior / VP positions) and a Bootcamp grad from 2016 who is now a tech lead at HubSpot, IBM VP, A Digital Assets and Workplace Activation Lead at Aristocrat, GitHub Business Ops Staff, Robinhood Developer Advocate&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;Best thing about the event?&lt;/strong&gt;&lt;br&gt;
The opportunity to learn from others in the field from diverse backgrounds (technical + nontechnical) and hear their experiences working with their communities and achieving positive impacts.&lt;br&gt;
Learning from the developers and community leaders about their projects, their contributors, their ideas for the future, and what is happening with emerging technologies&lt;br&gt;
The GitHub Women in Tech Event, where I had the opportunity to meet inspiring women and their career journeys&lt;br&gt;
Developing my interpersonal and technical skills through valuable workshops and in-person demos. I was even able to have my code reviewed at the Microsoft booth because at my internship, I was using an SDK they had released as Open Source. I learned from the Dev Advocates on best code practices and made advancements in my code's bugs&lt;br&gt;
The Dev Advocates + other employees also gave me valuable advice and motivation by sharing their experiences and career journeys. This conference and talking to others enabled me to see that these and other issues, whether systemic or&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;All of GitHub Universe keynotes, announcements, and most popular sessions&lt;/strong&gt;&lt;br&gt;
You can now watch 40+ on-demand sessions from GitHub Universe! &lt;br&gt;
&lt;a href="https://reg.githubuniverse.com/flow/github/universe23/sessioncatalog/page/sessioncatalog?search.deliveryformat=1692799009854004CQlx" rel="noopener noreferrer"&gt;https://reg.githubuniverse.com/flow/github/universe23/sessioncatalog/page/sessioncatalog?search.deliveryformat=1692799009854004CQlx&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;The highlights&lt;/strong&gt;&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;GitHub Copilot Enterprise: a new subscription tier that culminates the entirety of our Copilot offerings and personalizes them with the context of your organization's codebases, available in February 2024.&lt;/li&gt;
&lt;li&gt;GitHub Copilot Chat, an AI companion that'll allow you to write and understand code using whatever language you speak, will be generally available in December 2023 at no additional cost. We're also integrating Copilot Chat directly into github.com and mobile, so developers can code from anywhere.&lt;/li&gt;
&lt;li&gt;Now available in preview: Code scanning autofix and secret scanning for generic secrets provide new capabilities to help you prevent and detect more security vulnerabilities across your organization.&lt;/li&gt;
&lt;/ul&gt;

</description>
      <category>github</category>
      <category>githubuniverse</category>
      <category>githubuniverse2023</category>
      <category>githubcampusexperts</category>
    </item>
    <item>
      <title>How to get started with MongoDB as a Student</title>
      <dc:creator>Grace Gong</dc:creator>
      <pubDate>Thu, 14 Sep 2023 04:00:32 +0000</pubDate>
      <link>https://dev.to/1grace/how-to-get-started-with-mongodb-as-a-student-55j8</link>
      <guid>https://dev.to/1grace/how-to-get-started-with-mongodb-as-a-student-55j8</guid>
      <description>&lt;p&gt;As student with a passion for the developer community, I have participated in varous hackathons, and gotten involved in my student community. &lt;/p&gt;

&lt;p&gt;Throughout these experiences, I have had the opportunity to learn and develop projects using the MERN stack (MERN stands for MongoDB, Express, React, Node, after the four key technologies that make up the stack. MongoDB — document database. Express(.js) — Node.js web framework. React(.js) — a client-side JavaScript framework.) More information can be seen on their blog &lt;a href="https://www.mongodb.com/mern-stack#:~:text=MERN%20stands%20for%20MongoDB%2C%20Express,a%20client%2Dside%20JavaScript%20framework" rel="noopener noreferrer"&gt;here.&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fnzz4ooe0i3scj8s4xa63.png" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fnzz4ooe0i3scj8s4xa63.png" alt="Image description" width="800" height="400"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;As a software engineering intern, I have had the opportunity to also see how often MongoDB is used within projects, and how its capabilities have contributed to more efficient and reliable projects. &lt;/p&gt;

&lt;p&gt;Finally, In my previous semester, I had the opportunity to take a course on unstructured data, (for any students at western, CS4417 is the course!) where I was able to learn more about unstructured data, noSQL, and work on an assignment involving MongoDB, Aggregation, and MapReduce. Im happy to make seperate blog posts about these projects as well for any students interested in exploring applications and ways to use MongoDB in their own projects. &lt;/p&gt;

&lt;p&gt;Overall, MongoDB is a great tool to be able to leverage in your projects! Now you might be wondering, how do I get started? &lt;/p&gt;

&lt;p&gt;MLH (Major League Hacking) is a large hackathon organizer and a community that I have been involved in since the summer of 2020. &lt;/p&gt;

&lt;p&gt;As mentioned in their &lt;a href="https://news.mlh.io/major-league-hacking-mongodb-partner-to-inspire-the-next-generation-of-builders-03-27-2023" rel="noopener noreferrer"&gt;website post&lt;/a&gt;, &lt;/p&gt;

&lt;blockquote&gt;
&lt;p&gt;MongoDB Atlas – Best Database Platform for Your Next Hackathon&lt;br&gt;
MongoDB Atlas is a multi-cloud database with an integrated set of related services that allow development teams to address the growing requirements for today’s wide variety of modern applications, all in a unified and consistent user experience.&lt;/p&gt;
&lt;/blockquote&gt;

&lt;p&gt;MongoDB Atlas is more than a general-purpose database, it’s a full developer data platform – easily deploy and manage databases on-demand when and where you need them. &lt;/p&gt;

&lt;p&gt;Below are some ways you could use MongoDB Atlas in your next hackathon project.&lt;/p&gt;

&lt;p&gt;1) Simplify the way you integrate database functionality into your hackathon project by starting a free cluster or using your introductory $50 Atlas credits for students. &lt;/p&gt;

&lt;p&gt;2) Take your hackathon project to the next level, and deploy a database to the cloud in minutes. Signing up is easy and hassle-free with no credit card required! &lt;/p&gt;

&lt;p&gt;Key Benefits for Students&lt;br&gt;
MongoDB is one of the most prevalent database providers in the world, and we’ve got a range of opportunities to help you get started building on their platform. &lt;/p&gt;

&lt;p&gt;1) Receive $50 of free MongoDB Atlas credits through the GitHub Student Developer Pack, along with a MongoDB certification ($150 in value). The Forever Free Tier is available if you are no longer a student. &lt;/p&gt;

&lt;p&gt;2) Enter the “Best Use of MongoDB Atlas” challenge at an upcoming MLH event for a chance to win a M5GO iOT Starter Kit. To participate in the challenge, check out our website and register for an upcoming event! &lt;/p&gt;

&lt;p&gt;Interested in taking advantage of your MongoDB Atlas perks? Check out their website here. &lt;/p&gt;

&lt;p&gt;MongoDB also has developed "MongoDB University" a series of courses and ressources to support the learning and application of those learnings! I have taken some of the modules and can confirm the videos and tutorials go in depth and with the interactive nature of their tutorials, you will be engaged with the content. &lt;/p&gt;

&lt;p&gt;Finally, MongoDB has &lt;a href="https://www.mongodb.com/community/forums/c/user-groups/11" rel="noopener noreferrer"&gt;User Groups &lt;/a&gt;&lt;br&gt;
which bring people together to learn from and connect over their shared interest in MongoDB technologies. These are communities of developers with various skill levels and an interest in learning new technologies and applications including MongoDB and beyond, and are a great opportunity to get involved! Whether you are a beginner or a skilled individiaul, this is a safe space for users and novices to come together, learn from each other, and share what theyre doing with mongodb and other tech. &lt;/p&gt;

&lt;p&gt;I recently became involved as a co-organizer for the &lt;a href="https://www.meetup.com/toronto-mongodb-usergroup/" rel="noopener noreferrer"&gt;Toronto User Group&lt;/a&gt; and confirm the community is welcoming for people with diverse experiences and has enabled me to meet and learn from people around the company and community. &lt;/p&gt;

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Ffxh7b8ye337xc6exh4at.png" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Ffxh7b8ye337xc6exh4at.png" alt="Image description" width="800" height="450"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;These are just starting points - there are many options for ways to get involved with MongoDB beyond this!  &lt;/p&gt;

&lt;p&gt;You could build an application or library you've built using MongoDB, contribute to open source related to MongoDB, &lt;br&gt;
talk at a local MongoDB event, create a video or tutorial sharing what you've learned or created about MongoDB, &lt;br&gt;
get involved and provide support through Stack Overflow, the MongoDB Community Forum or elsewhere, and more! &lt;/p&gt;

&lt;p&gt;If you had any questions feel free to leave them in the comments! &lt;/p&gt;

</description>
      <category>mongodb</category>
      <category>nosql</category>
      <category>database</category>
      <category>databasemanagement</category>
    </item>
    <item>
      <title>Toronto MongoDB User Group Meeting Sept 5</title>
      <dc:creator>Grace Gong</dc:creator>
      <pubDate>Fri, 08 Sep 2023 01:46:08 +0000</pubDate>
      <link>https://dev.to/1grace/toronto-mongodb-user-group-meeting-sept-5-40j5</link>
      <guid>https://dev.to/1grace/toronto-mongodb-user-group-meeting-sept-5-40j5</guid>
      <description>&lt;p&gt;I had a great time co-organizing the Toronto MUG event this past week!&lt;/p&gt;

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Funs8qe1lme0720u0mgzs.png" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Funs8qe1lme0720u0mgzs.png" alt="Image description" width="800" height="450"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Flvbl6uo9pcalherkjiyl.JPG" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Flvbl6uo9pcalherkjiyl.JPG" alt="Image description" width="800" height="600"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2F4kzgypockry8cqj6riyy.JPG" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2F4kzgypockry8cqj6riyy.JPG" alt="Image description" width="800" height="600"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fzrz44vf5npdrypsec3sn.JPG" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fzrz44vf5npdrypsec3sn.JPG" alt="Image description" width="800" height="600"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fltd7iji4yv5k1gz2p8gs.JPG" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fltd7iji4yv5k1gz2p8gs.JPG" alt="Image description" width="800" height="600"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;Event Summary&lt;br&gt;
Toronto MUG Organizers: Chris Dellaway, Grace Gong&lt;br&gt;
Agenda:&lt;br&gt;
6:00PM Doors Open, Supper and Networking&lt;br&gt;
6:30PM Prashant (MongoDB) Amplifying Retail Operations with Generative AI: The Untapped Potential&lt;br&gt;
7:00PM Lachlan (Community) - Company journey with MongoDB&lt;br&gt;
7:20PM Devang (Community) Using MongoDB Clustered Collection to Boost Query Performance&lt;br&gt;
7:30PM Networking&lt;/p&gt;

&lt;p&gt;The event was held at the Microsoft Toronto headquarters from 6-8pm on Tuesday September 5 2023 with the aim of fostering connections and sharing knowledge among members of the MongoDB Community.&lt;br&gt;
The crowd was truly engaged, and the diversity of the attendees resulted in vibrant discussions and insightful questions. &lt;/p&gt;

&lt;p&gt;Learned&lt;/p&gt;

&lt;p&gt;Attendance Accuracy: Future events should either emphasize the importance of accurate RSVPing or anticipate a specific turnout percentage to prevent underutilisation of space.&lt;br&gt;
Improving Check-in: An advanced and digital check-in system can greatly improve the event’s commencement process.&lt;br&gt;
Balancing Time: A clear mechanism to ensure speakers stick to their allocated time can make the event more streamlined.&lt;br&gt;
Loved&lt;/p&gt;

&lt;p&gt;Speakers: They were highly informative, and their sessions were greatly appreciated by all attendees.&lt;br&gt;
Community Connection: Attendees left the event feeling more connected with the MongoDB Community and inspired.&lt;br&gt;
Feedback: The considerable number of feedback forms received demonstrated the attendees’ involvement and commitment to improving future events.&lt;br&gt;
Diverse Discussions: The diverse background of attendees enriched the discussions and questions during the event.&lt;br&gt;
Ideas/Action Items for the Next Event&lt;/p&gt;

&lt;p&gt;Allocate budget for name tags and markers to facilitate easier recognition and networking. Also next time we can introduce volunteers and have the slide for them as well&lt;br&gt;
Explore advanced check-in systems for a smoother start to the event.&lt;br&gt;
Liaise with adjacent event organisers to coordinate and ensure both events run seamlessly without confusion for attendees.&lt;/p&gt;

</description>
    </item>
    <item>
      <title>Part 1: Intro to serverless</title>
      <dc:creator>Grace Gong</dc:creator>
      <pubDate>Mon, 03 Jul 2023 17:59:03 +0000</pubDate>
      <link>https://dev.to/1grace/part-1-intro-to-serverless-40b2</link>
      <guid>https://dev.to/1grace/part-1-intro-to-serverless-40b2</guid>
      <description>&lt;p&gt;As Im learning about cloud computing and reading up on various sources including aws, I thought I would share some of my learnings. &lt;/p&gt;

&lt;p&gt;&lt;strong&gt;Why serverless?&lt;/strong&gt;&lt;br&gt;
Building Secure Reliable and Cost effective architecture is important for IT professionals. Many organizations are replacing monolithic methods with microservices methodology. &lt;/p&gt;

&lt;p&gt;&lt;strong&gt;What are Microservices?&lt;/strong&gt;&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;Architectural and organizational approach to software development. &lt;/li&gt;
&lt;li&gt;Software is composed of small independent services. &lt;/li&gt;
&lt;li&gt;Communicate over well-defined APIs. &lt;/li&gt;
&lt;li&gt;Owned by small, self-contained teams.&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fp5axl8666t1zllf9lcyq.png" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fp5axl8666t1zllf9lcyq.png" alt="Image description" width="800" height="556"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;Tightly versus loosely coupled&lt;/strong&gt;&lt;/p&gt;

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2F2zawz27pe4exv12d72as.png" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2F2zawz27pe4exv12d72as.png" alt="Image description" width="800" height="423"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;Monolithic application (tightly coupled)&lt;/strong&gt; &lt;br&gt;
In a monolithic application, if one piece were to fail, the entire application would crash, as this example illustrates. If there is a spike in demand, the entire architecture must be scaled. Adding features to a monolithic application becomes more complex as time goes on. Pieces of the code base must work with each other to sync properly.&lt;/p&gt;

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fsrwldz5a1ubr2bw2l5k1.png" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fsrwldz5a1ubr2bw2l5k1.png" alt="Image description" width="692" height="312"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;Microservices application (loosely coupled)&lt;/strong&gt;&lt;br&gt;
In a microservices architecture, an application is built as independent components that run each application process as a service. These services communicate through a well-defined interface by using lightweight APIs. Services are built for business capabilities, and each service performs a single function. Because they are independently run, each service can be updated, deployed, and scaled to meet demand for specific functions of an application. If one component fails, the other components remain active, as illustrated.&lt;/p&gt;

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fubxdcl7zn9i2hg3emqm4.png" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fubxdcl7zn9i2hg3emqm4.png" alt="Image description" width="616" height="274"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Flxizapn92oo7za7tmdrm.png" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Flxizapn92oo7za7tmdrm.png" alt="Image description" width="728" height="360"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;Benefits of microservices&lt;/strong&gt;&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;Use public APIs&lt;/strong&gt;&lt;br&gt;
Microservices communicate over APIs. In fact, doing so is the only way to realize the benefits of microservices. Don't be tempted to circumvent that rule. By communicating with APIs, you can modify a database and maintain it without affecting communications or having to repoint interfaces. The public API is a contract between your microservices and its clients.&lt;/p&gt;

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fheth66tyi4rzx0xqk4xy.png" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fheth66tyi4rzx0xqk4xy.png" alt="Image description" width="654" height="318"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;Use the right tool for the job&lt;/strong&gt;&lt;br&gt;
Because every service is built for explicit functionality, you can choose the right tool for the job. The development team for Microservices B can add Amazon Aurora Service or change to Amazon OpenSearch without affecting Microservices A. You can also make changes to the entire language framework, which gives you the flexibility to code in the language that is best suited to an application. By choosing the programming framework that works for the team, the development team for Microservices B can avoid depending on other teams for their knowledge of specific frameworks.&lt;/p&gt;

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fq2ucsajj9r3u2ibthpam.png" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fq2ucsajj9r3u2ibthpam.png" alt="Image description" width="752" height="320"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;Secure your services&lt;/strong&gt;&lt;br&gt;
 Microservices must be architected to provide security at every step. The defense- in-depth approach is a common method of reviewing each step of the application to determine areas where additional security controls can be implemented. In a microservices environment, this arrangement includes the application level, API level, network layer, and AWS Identity and Access Management (IAM) configurations. The API is the entry point. Amazon API Gateway can perform authentication and authorization for services and act as security control for microservices by enforcing throttling and obfuscation of internal resources.&lt;/p&gt;

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fqr6cvyzaixpkbfo0ouaz.png" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fqr6cvyzaixpkbfo0ouaz.png" alt="Image description" width="784" height="300"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;Be a good citizen&lt;/strong&gt;&lt;br&gt;
As this environment of microservices grows and expands, it becomes important to be a good citizen. As a service owner, you must understand how other services are calling you. The service owner will identify the client's use case and register them in the service registry. As a client of services, you must understand the service that you are calling.&lt;/p&gt;

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fejmv8d20z892bhcmnhth.png" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fejmv8d20z892bhcmnhth.png" alt="Image description" width="772" height="366"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;Company transformation&lt;/strong&gt;&lt;br&gt;
Going from a monolith to microservices is more than a technological transformation. The shift requires transformation in the organization and communication mechanisms between teams. Automation is the key to rapidly deploying stable applications. Small development teams ("two-pizza teams" or no more than the number of people who can be fed with two pizzas) focus on producing a single microservice. They can send the service through an automated delivery pipeline and release a service into production with everything, from the build step forward, being automated. By organizing the software as a grouping of microservices, the development teams are decoupled and can efficiently release updates to their microservices on their own.&lt;/p&gt;

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2F8opqy243hnkemxp71zby.png" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2F8opqy243hnkemxp71zby.png" alt="Image description" width="790" height="260"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;Automate everything&lt;/strong&gt;&lt;br&gt;
Some examples of some tools can help with automation. &lt;br&gt;
&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fjbmzwru0pb9vwwnqlzp5.png" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fjbmzwru0pb9vwwnqlzp5.png" alt="Image description" width="766" height="342"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;Downsides of monoliths&lt;/strong&gt; &lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;All processes are tightly coupled and run as a single service.&lt;/li&gt;
&lt;li&gt;Adding or improving a monolithic application's features becomes more complex as the code base grows.&lt;/li&gt;
&lt;li&gt;Dependent and tightly coupled processes increase the impact of a single process failure.&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;Serverless applications dont require you to provision, scale or manage any servers. &lt;/p&gt;

&lt;p&gt;You can build them for nearly any type of application or backend service, and everything required to run and scale your application with high availability is handled for you.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;Why serverless?&lt;/strong&gt;&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;developers dont need to worry about managing and operating servers of runtimes and can focus on other areas, thus can focus on core product &lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;&lt;strong&gt;Event-driven architectures&lt;/strong&gt; &lt;/p&gt;

&lt;ul&gt;
&lt;li&gt; Uses events to initiate actions and communication between decoupled services. &lt;/li&gt;
&lt;li&gt; Change in state, a user request, or an update. &lt;/li&gt;
&lt;li&gt; Information published for other services to consume it.&lt;/li&gt;
&lt;li&gt; Events are the primary mechanism for sharing information across services.&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fr1k4wtan8pu18ks6clfo.png" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fr1k4wtan8pu18ks6clfo.png" alt="Image description" width="800" height="584"&gt;&lt;/a&gt;&lt;br&gt;
 • Microservices is a method of building your architecture. &lt;/p&gt;

&lt;p&gt;• Serverless is a service that is used to build your architecture. &lt;/p&gt;

&lt;p&gt;• Microservices is the method of loosely coupling sections of your architecture. &lt;/p&gt;

&lt;p&gt;• Serverless services are the tools that you use to build out those sections of your architecture. &lt;/p&gt;

&lt;p&gt;• It's common to use a blend of both serverless and nonserverless services.&lt;/p&gt;

</description>
      <category>aws</category>
      <category>cloudcomputing</category>
    </item>
  </channel>
</rss>
