<?xml version="1.0" encoding="UTF-8"?>
<rss version="2.0" xmlns:atom="http://www.w3.org/2005/Atom" xmlns:dc="http://purl.org/dc/elements/1.1/">
  <channel>
    <title>DEV Community: Steadybit</title>
    <description>The latest articles on DEV Community by Steadybit (@steadybit).</description>
    <link>https://dev.to/steadybit</link>
    <image>
      <url>https://media2.dev.to/dynamic/image/width=90,height=90,fit=cover,gravity=auto,format=auto/https:%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Forganization%2Fprofile_image%2F10734%2F31c8db6f-9197-46e5-a615-5cd60e7026fa.png</url>
      <title>DEV Community: Steadybit</title>
      <link>https://dev.to/steadybit</link>
    </image>
    <atom:link rel="self" type="application/rss+xml" href="https://dev.to/feed/steadybit"/>
    <language>en</language>
    <item>
      <title>The Business Case for Chaos Engineering: An ROI Calculator for Testing Application Reliability</title>
      <dc:creator>Patrick Londa</dc:creator>
      <pubDate>Tue, 10 Mar 2026 21:41:50 +0000</pubDate>
      <link>https://dev.to/steadybit/the-business-case-for-chaos-engineering-an-roi-calculator-for-testing-application-reliability-2dhk</link>
      <guid>https://dev.to/steadybit/the-business-case-for-chaos-engineering-an-roi-calculator-for-testing-application-reliability-2dhk</guid>
      <description>&lt;p&gt;&lt;strong&gt;"What do we get from intentionally injecting failures into our systems?"&lt;/strong&gt;&lt;/p&gt;

&lt;p&gt;&lt;a href="https://steadybit.com/chaos-engineering/" rel="noopener noreferrer"&gt;Chaos engineering&lt;/a&gt; is one of the best ways to proactively test your application reliability, but many leadership teams have never heard of the concept.&lt;/p&gt;

&lt;p&gt;Engineering teams need to be able to frame a strong business case to explain the value of chaos engineering and reliability testing to budget holders. When there’s a major outage, the value of application reliability becomes immediately clear, but a strong ROI plan can help earn and maintain executive support while your systems are steady.&lt;/p&gt;

&lt;p&gt;We have just released an &lt;a href="https://steadybit.com/chaos-engineering/roi-calculator/" rel="noopener noreferrer"&gt;interactive ROI calculator&lt;/a&gt; that can help SRE teams frame the business value of proactive reliability efforts like chaos engineering.&lt;/p&gt;

&lt;h2&gt;
  
  
  Ensuring Application Resilience with Chaos Engineering
&lt;/h2&gt;

&lt;p&gt;Complex software systems are destined to break and fail at some point, especially with all the factors present in modern production environments. When we ask teams about their approach to system reliability, we often hear back: “We have enough chaos already!”&lt;/p&gt;

&lt;p&gt;When done correctly, &lt;a href="https://steadybit.com/chaos-engineering/" rel="noopener noreferrer"&gt;chaos engineering&lt;/a&gt; isn’t adding chaos to your systems. Rather, it’s running different scenarios to validate how resilient your systems are under stressful conditions. You can test your expectations vs. the reality of performance degradation during an availability zone outage, delayed dependency, or a sudden surge of users.&lt;/p&gt;

&lt;p&gt;These experiments serve as a feedback loop earlier in the software development cycle that enable teams to design more fault-tolerant systems for their users.&lt;/p&gt;

&lt;p&gt;When reliability testing is rolled out across applications, teams can find and address risks that could lead to critical incidents and improve their average time to remediation (MTTR).&lt;/p&gt;

&lt;h2&gt;
  
  
  Early ROI Prototypes and Why They Didn’t Work
&lt;/h2&gt;

&lt;h3&gt;
  
  
  📉 Savings from Reducing the Overall Number of Incidents
&lt;/h3&gt;

&lt;p&gt;Initially, we explored the premise that implementing chaos engineering will lead to fewer incidents overall at every severity tier. If a user provides how many overall incidents they have had in the past year, we could assume some percentage reduction to all incidents. This makes the calculation pretty simple once a cost amount is assigned to each tier of incident.&lt;/p&gt;

&lt;p&gt;The trouble with this approach is it assumes that all incidents should be avoided. Incidents as metrics can be useful in flagging anomalous behaviors and some might not necessarily result in negative impacts for many customers. An increase in low-level incidents could actually be a positive sign that alert coverage is improving and properly surfacing system weaknesses. &lt;/p&gt;

&lt;p&gt;Moving forward, we chose to focus on savings from reducing the number of critical incidents (Sev0, P1, etc.) year over year. This is a metric that is easier to track without creating incentives to avoid marking incidents at any level. &lt;/p&gt;

&lt;h3&gt;
  
  
  🧪 Identifying &amp;amp; Fixing Reliability Risks By Running More Experiments
&lt;/h3&gt;

&lt;p&gt;If you go from 0 to 100 experiment runs on your systems, you are bound to discover new performance gaps and reliability risks. How many more risks will you discover at 200 experiment runs?&lt;/p&gt;

&lt;p&gt;We built a version of an ROI calculator that assumed that as the number of experiments increased, a certain percentage of experiments would reveal issues at different incident risk levels with assigned potential costs. Teams would then fix a certain percentage of these reliability risks depending on their development capacity. As the experiment run count scaled, there would be diminished returns for revealing new issues.&lt;/p&gt;

&lt;p&gt;While it’s true that teams will find more potential reliability issues as they run more experiments, this approach was a little too one-dimensional. We didn’t have well-documented references for how issue detection rates would change, and teams would likely need to create a new reporting mechanism to follow-along as they mitigated risks. &lt;/p&gt;

&lt;p&gt;There is also nuance as some teams will automate their experiment runs and use them as regression tests in their CI/CD processes. We decided it would be better to stick to measuring impact by with metrics that are already being tracked and available for SREs at most organizations.&lt;/p&gt;

&lt;h2&gt;
  
  
  Where We Landed with Key Metrics for Our ROI Calculator
&lt;/h2&gt;

&lt;h3&gt;
  
  
  ⚡ Savings from Faster Average MTTR
&lt;/h3&gt;

&lt;p&gt;As we were iterating on the inputs and outputs, we saw a &lt;a href="https://youtu.be/mYKNR0UXwMc?si=d-HUlm7xivWNaTHd&amp;amp;t=2444" rel="noopener noreferrer"&gt;great presentation&lt;/a&gt; from Keith Blizard and Joe Cho at AWS Re:Invent 2024, featuring a case study on the progress Fidelity Investments had made in rolling out chaos engineering across their organization. They documented major improvements to mean-time-to-resolution (MTTR) as they scaled chaos testing coverage across applications.&lt;/p&gt;

&lt;p&gt;

  &lt;iframe src="https://www.youtube.com/embed/mYKNR0UXwMc"&gt;
  &lt;/iframe&gt;


&lt;/p&gt;

&lt;p&gt;We used these case study metrics to plot the correlation between the percent of applications with chaos testing coverage to the incremental positive impact to their MTTR. We then used this relationship to calculate improvements against an assumed industry-wide average MTTR of 175 minutes, according to this &lt;a href="https://www.pagerduty.com/resources/insights/learn/cost-of-downtime/" rel="noopener noreferrer"&gt;2024 PagerDuty report&lt;/a&gt;.&lt;/p&gt;

&lt;p&gt;This MTTR savings means fewer minutes of downtime, which that same study estimated can cost between $4,000 - $15,000 per minute. In our calculator, we ask users to input their “Annual Company Revenue” so we can use the most relevant cost of downtime per minute, as downtime is typically more costly for larger enterprises. This &lt;a href="https://www.bigpanda.io/wp-content/uploads/2024/04/EMA-BigPanda-final-Outage-eBook.pdf" rel="noopener noreferrer"&gt;2024 report&lt;/a&gt; commissioned by BigPanda found that downtime cost an average of $14,056 per minute for organizations with more than 1,000 employees.&lt;/p&gt;

&lt;h3&gt;
  
  
  🛡️ Savings from Reducing Critical Incidents
&lt;/h3&gt;

&lt;p&gt;At &lt;a href="https://steadybit.com/" rel="noopener noreferrer"&gt;Steadybit&lt;/a&gt;, we partner with a wide range of customers and have seen how many major reliability gaps are uncovered by running chaos experiments. Using insights from our customers and referencing industry studies, we've seen that actively running reliability tests on any given application conservatively leads to an average 30% reduction in critical incidents for that application per year.&lt;/p&gt;

&lt;p&gt;For &lt;a href="https://steadybit.com/chaos-engineering/roi-calculator/" rel="noopener noreferrer"&gt;our calculator&lt;/a&gt;, we ask users to input the total number of applications their organization operates and how many of these applications have reliability testing coverage. We multiply the standard 30% reduction per year in critical incidents by the percent of applications with testing coverage to get the overall incident reduction for the organization.&lt;/p&gt;

&lt;h3&gt;
  
  
  🛠️ Costs of Implementing Chaos Engineering
&lt;/h3&gt;

&lt;p&gt;If you want to run chaos experiments at scale, you will likely need to onboard a commercial reliability platform or chaos engineering tool. Open source solutions can be a good starting place, but deploying these across teams and technologies can become increasingly time-intensive. We used general license estimates based on market knowledge and projected experiment activity.&lt;/p&gt;

&lt;p&gt;Like with any new program, an organization will need engineers owning the project and dedicating time to a successful rollout of chaos testing. We included a field in our calculator for “Testing Rollout Managers”, measured by FTEs (40hr/week of staff time). We used an average SRE salary of $160k per year as a benchmark to estimate the cost of this implementation effort.&lt;/p&gt;

&lt;h2&gt;
  
  
  Showing the Return On Investment for Reliability Testing
&lt;/h2&gt;

&lt;p&gt;We ask users to project how they would expect to rollout chaos engineering at their organization, including unique test types, number of experiments, and coverage across applications. Our &lt;a href="https://steadybit.com/chaos-engineering/roi-calculator/" rel="noopener noreferrer"&gt;ROI calculator&lt;/a&gt; will then output a summary and detailed view of your projected savings, implementation costs, and return on investment. When you game out multi-year adoption goals, you'll be building a business case that can help you frame the value of making this type of investment.&lt;/p&gt;

&lt;p&gt;If you’re successful in getting buy-in to roll out chaos engineering, you’ll need to report back on your progress. If you’re using an incident management platform like Splunk or PagerDuty, you may already have built-in MTTR metrics available to reference. You can also track the number of critical incidents using Observability tools like Datadog, Dynatrace, or Grafana Labs.&lt;/p&gt;

&lt;p&gt;These metrics will hopefully show clear improvements, but your systems may become increasingly complex at the same time that you’re rolling out this testing, especially with the rise of AI agents. Even simply maintaining your current reliability posture as your systems evolve and become significantly more complex could be framed as a win.&lt;/p&gt;

&lt;h2&gt;
  
  
  Rolling Out Chaos Engineering Across Your Organization
&lt;/h2&gt;

&lt;p&gt;Highly available applications don’t naturally draw attention in the way that outages do. If you want to continue the momentum and foster a culture of reliability, you'll need to intentionally share your wins. For example, if you find a major reliability vulnerability and are able to address it before it impacts customers, that's something to celebrate internally.&lt;/p&gt;

&lt;p&gt;If you want to help getting started with chaos testing and adopting a proactive reliability program, our team of experts at &lt;a href="https://steadybit.com/" rel="noopener noreferrer"&gt;Steadybit&lt;/a&gt; is ready to help.&lt;/p&gt;

&lt;p&gt;You can explore our reliability platform with a &lt;a href="https://signup.steadybit.com/" rel="noopener noreferrer"&gt;30-day free trial&lt;/a&gt; or &lt;a href="https://steadybit.com/book-demo/" rel="noopener noreferrer"&gt;book a quick call&lt;/a&gt; with us to discuss how you can implement chaos engineering and start saving money today.&lt;/p&gt;

</description>
      <category>roi</category>
      <category>chaosengineering</category>
      <category>sre</category>
      <category>testing</category>
    </item>
    <item>
      <title>3 Types of Chaos Experiments and How To Run Them</title>
      <dc:creator>Patrick Londa</dc:creator>
      <pubDate>Thu, 24 Apr 2025 17:44:42 +0000</pubDate>
      <link>https://dev.to/steadybit/3-types-of-chaos-experiments-and-how-to-run-them-1p59</link>
      <guid>https://dev.to/steadybit/3-types-of-chaos-experiments-and-how-to-run-them-1p59</guid>
      <description>&lt;p&gt;The primary objective of a Chaos Experiment is to uncover hidden bugs, weaknesses, or non-obvious points of failure in a system that could lead to significant outages, degradation of service, or system failure under unpredictable real-world conditions.&lt;/p&gt;

&lt;h1&gt;
  
  
  What is a Chaos Experiment?
&lt;/h1&gt;

&lt;p&gt;A Chaos Experiment is a carefully designed, controlled, and monitored process that systematically introduces disturbances or abnormalities into a system’s operation to observe and understand its response to such conditions.&lt;/p&gt;

&lt;p&gt;It forms the core part of &lt;a href="https://steadybit.com/chaos-engineering/" rel="noopener noreferrer"&gt;‘Chaos Engineering’&lt;/a&gt;, which is predicated on the idea that ‘the best way to understand system behavior is by observing it under stress.’ This means intentionally injecting faults into a system in production or simulated environments to test its reliability and resilience.&lt;/p&gt;

&lt;p&gt;This practice emerged from the understanding that systems, especially distributed systems, are inherently complex and unpredictable due to their numerous interactions and dependencies.&lt;/p&gt;

&lt;h1&gt;
  
  
  The Components of a Chaos Engineering Experiment
&lt;/h1&gt;

&lt;ul&gt;
&lt;li&gt;&lt;p&gt;&lt;strong&gt;Hypothesis Formation.&lt;/strong&gt; At the initial stage, a hypothesis is formed about the system’s steady-state behavior and expected resilience against certain types of disturbances. This hypothesis predicts no significant deviation in the system’s steady state as a result of the experiment.&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;&lt;strong&gt;Variable Introduction.&lt;/strong&gt; This involves injecting specific variables or conditions that simulate real-world disturbances (such as network latency, server failures, or resource depletion). These variables are introduced in a controlled manner to avoid unnecessary risk.&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;&lt;strong&gt;Scope and Safety.&lt;/strong&gt; The experiment’s scope is clearly defined to limit its impact, often called the “blast radius.” Safety mechanisms, such as automatic rollback or kill switches, are implemented to halt the experiment if unexpected negative effects are observed.&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;&lt;strong&gt;Observation and Data Collection.&lt;/strong&gt; Throughout the experiment, system performance and behavior are closely monitored using detailed logging, metrics, and observability tools. This data collection is critical for analyzing the system’s response to the introduced variables.&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;&lt;strong&gt;Analysis and Learning.&lt;/strong&gt; After the experiment, the data is analyzed to determine whether the hypothesis was correct. This analysis extracts insights regarding the system’s vulnerabilities, resilience, and performance under stress.&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;&lt;strong&gt;Iterative Improvement.&lt;/strong&gt; The findings from each chaos experiment inform adjustments in system design, architecture, or operational practices. These adjustments aim to mitigate identified weaknesses and enhance overall resilience.&lt;/p&gt;&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;💡 Note → The ultimate goal is not to break things randomly but to uncover systemic weaknesses to improve the system’s resilience. By introducing chaos, you can enhance the understanding of your systems, leading to higher availability, reliability, and a better user experience.&lt;/p&gt;

&lt;h1&gt;
  
  
  Types of Chaos Experiments
&lt;/h1&gt;

&lt;h2&gt;
  
  
  1. Dependency Failure Experiment
&lt;/h2&gt;

&lt;p&gt;&lt;strong&gt;Objective:&lt;/strong&gt; To assess how microservices behave when one or more of their dependencies fail. In a microservices architecture, services are designed to perform small tasks and often rely on other services to fulfill a request. The failure of these external dependencies can lead to cascading failures across the system, resulting in degraded performance or system outages. Understanding how these failures impact the overall system is crucial for building resilient services.&lt;/p&gt;

&lt;h3&gt;
  
  
  Possible Experiments
&lt;/h3&gt;

&lt;ul&gt;
&lt;li&gt;&lt;p&gt;&lt;strong&gt;Network Latency and Packet Loss.&lt;/strong&gt; Simulate increased latency or packet loss to understand its impact on service response times and throughput.&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;&lt;strong&gt;Service Downtime.&lt;/strong&gt; Emulate the unavailability of a critical service to observe the system’s resilience and failure modes.&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;&lt;strong&gt;Database Connectivity Issues.&lt;/strong&gt; Introduce connection failures or read/write delays to assess the robustness of data access patterns and caching mechanisms.&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;&lt;strong&gt;Third-party API Limiting.&lt;/strong&gt; Mimic rate limiting or downtime of third-party APIs to evaluate external dependency management and error handling.&lt;/p&gt;&lt;/li&gt;
&lt;/ul&gt;

&lt;h3&gt;
  
  
  How to Run a Dependency Failure Experiment
&lt;/h3&gt;

&lt;p&gt;&lt;strong&gt;Map Out Dependencies.&lt;/strong&gt;&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;&lt;p&gt;Begin with a comprehensive inventory of all the external services your system interacts with. This includes databases, third-party APIs, cloud services, and internal services if you work in a microservices architecture.&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;For each dependency, document how your system interacts with it. Note the data exchanged, request frequency, and criticality of each interaction to your system’s operations.&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;Rank these dependencies based on their importance to your system’s core functionalities. This will help you focus your efforts on the most critical dependencies first.&lt;/p&gt;&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;&lt;strong&gt;Simulate Failures&lt;/strong&gt;&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;&lt;p&gt;Use service virtualization or proxy tools like SteadyBit to simulate various failures for your dependencies. These can range from network latency, dropped connections, and timeouts to complete unavailability.&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;For each dependency, configure the types of faults you want to introduce. This could include delays, error rates, or bandwidth restrictions, mimicking real-world issues that could occur.&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;Start with less severe faults (like increased latency) and gradually move to more severe conditions (like complete downtime), observing the system’s behavior at each stage.&lt;/p&gt;&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;&lt;strong&gt;Test Microservices Isolation&lt;/strong&gt;&lt;/p&gt;

&lt;p&gt;Implement Resilience Patterns. Use libraries like &lt;a href="https://github.com/Netflix/Hystrix" rel="noopener noreferrer"&gt;Hystrix&lt;/a&gt;, &lt;a href="https://resilience4j.readme.io/docs/getting-started" rel="noopener noreferrer"&gt;resilience4j&lt;/a&gt;, or &lt;a href="https://spring.io/projects/spring-cloud-circuitbreaker" rel="noopener noreferrer"&gt;Spring Cloud Circuit Breaker&lt;/a&gt; to implement patterns that prevent failures from cascading across services. This includes:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;&lt;p&gt;&lt;strong&gt;Bulkheads.&lt;/strong&gt; Isolate parts of the application into “compartments” to prevent failures in one area from overwhelming others.&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;&lt;strong&gt;Circuit Breakers.&lt;/strong&gt; Automatically, “cut off” calls to a dependency if it’s detected as down, allowing it to recover without being overwhelmed by constant requests.&lt;/p&gt;&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;Carefully configure thresholds and timeouts for these patterns. This includes setting the appropriate parameters for circuit breakers to trip and recover and defining bulkheads to isolate services effectively.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;Monitor Inter-Service Communication&lt;/strong&gt;&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;&lt;p&gt;Utilize monitoring solutions like &lt;a href="https://prometheus.io/" rel="noopener noreferrer"&gt;Prometheus&lt;/a&gt;, &lt;a href="https://grafana.com/" rel="noopener noreferrer"&gt;Grafana&lt;/a&gt;, or &lt;a href="https://www.datadoghq.com/" rel="noopener noreferrer"&gt;Datadog&lt;/a&gt; to monitor how services communicate under normal and failure conditions. Service meshes like Istio or Linkerd can provide detailed insights without changing your application code.&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;Focus on metrics like request success rates, latency, throughput, and error rates. These metrics will help you understand the impact of dependency failures on your system’s performance and reliability.&lt;/p&gt;&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;💡 Recommendation → Monitoring in real-time allows you to quickly identify and respond to unexpected behaviors, minimizing the impact on your system.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;Analyze Fallback Mechanisms&lt;/strong&gt;&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;&lt;p&gt;Evaluate the effectiveness of implemented fallback mechanisms. This includes static responses, cache usage, default values, or switching to a secondary service if the primary is unavailable.&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;Assess if the ‘retry logic’ is appropriately configured. This includes evaluating the retry intervals, backoff strategies, and the maximum number of attempts to prevent overwhelming a failing service.&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;Ensure that fallback mechanisms enable your system to operate in a degraded mode rather than failing outright. This helps maintain a service level even when dependencies are experiencing issues.&lt;/p&gt;&lt;/li&gt;
&lt;/ul&gt;

&lt;h2&gt;
  
  
  2. Resource Manipulation Experiment
&lt;/h2&gt;

&lt;p&gt;&lt;strong&gt;Objective:&lt;/strong&gt; To understand how a system behaves when subjected to unusual or extreme resource constraints, such as CPU, memory, disk I/O, and network bandwidth. The aim is to identify potential bottlenecks and ensure that the system can handle unexpected spikes in demand without significantly degrading service.&lt;/p&gt;

&lt;h3&gt;
  
  
  Possible Experiments
&lt;/h3&gt;

&lt;ul&gt;
&lt;li&gt;&lt;p&gt;&lt;strong&gt;CPU Saturation.&lt;/strong&gt; Increase CPU usage gradually to see how the system prioritizes tasks and whether essential services remain available.&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;&lt;strong&gt;Memory Consumption.&lt;/strong&gt; Simulate memory leaks or high memory demands to test the system’s handling of low memory conditions.&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;&lt;strong&gt;Disk I/O and Space Exhaustion.&lt;/strong&gt; Increase disk read/write operations or fill up disk space to observe how the system copes with disk I/O bottlenecks and space limitations.&lt;/p&gt;&lt;/li&gt;
&lt;/ul&gt;

&lt;h3&gt;
  
  
  How to Run a Resource Manipulation Experiment
&lt;/h3&gt;

&lt;p&gt;&lt;strong&gt;Define Resource Limits&lt;/strong&gt;&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;&lt;p&gt;Start by monitoring your system under normal operating conditions to establish a baseline for CPU, memory, disk I/O, and network bandwidth usage.&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;Based on historical data and performance metrics, define the normal operating range for each critical resource. This will help you identify when the system is under stress, or resource usage is abnormally high during the experiment.&lt;/p&gt;&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;&lt;strong&gt;Check and Verify the Break-Even Point&lt;/strong&gt;&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;&lt;p&gt;Understand your system’s maximum capacity before it requires scaling. This involves testing the system under gradually increasing load to identify the point at which performance starts to degrade, and additional resources are needed.&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;If you’re using auto-scaling (either in the cloud or on-premises), clearly define and verify the rules for adding new instances or allocating resources. This includes setting CPU, memory usage thresholds, and other metrics that trigger scaling actions.&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;Use load testing tools like &lt;a href="https://jmeter.apache.org/" rel="noopener noreferrer"&gt;JMeter&lt;/a&gt;, &lt;a href="https://gatling.io/" rel="noopener noreferrer"&gt;Gatling&lt;/a&gt;, or &lt;a href="https://locust.io/" rel="noopener noreferrer"&gt;Locust&lt;/a&gt; to simulate demand spikes and verify that your auto-scaling rules work as expected. This will ensure that your system can handle real-world traffic patterns.&lt;/p&gt;&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;&lt;strong&gt;Select Manipulation Tool&lt;/strong&gt;&lt;/p&gt;

&lt;p&gt;While Stress and Stress-ng are powerful for generating CPU, memory, and I/O load on Linux systems, they might not be easy to use across distributed or containerized environments. Tools like &lt;a href="https://steadybit.com/" rel="noopener noreferrer"&gt;Steadybit&lt;/a&gt; offer more user-friendly interfaces for various environments, including microservices and cloud-native applications.&lt;/p&gt;

&lt;p&gt;💡 Pro Tip → Ensure that the tool you select can accurately simulate the types of resource manipulation you’re interested in, whether it’s exhausting CPU cycles, filling up memory, saturating disk I/O, or hogging network bandwidth.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;Apply Changes Gradually&lt;/strong&gt;&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;&lt;p&gt;Start by applying small changes to resource consumption and monitor the system’s response.&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;Monitor system performance carefully to identify the thresholds at which performance degrades or fails. This will help you understand the system’s resilience and where improvements are needed.&lt;/p&gt;&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;&lt;strong&gt;Monitor System Performance&lt;/strong&gt;&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;Use comprehensive monitoring solutions to track the impact of resource manipulation on system performance. Look for changes in response times, throughput, error rates, and system resource utilization.&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;💡 Pro Tip → Platforms like &lt;a href="https://steadybit.com/" rel="noopener noreferrer"&gt;Steadybit&lt;/a&gt; can integrate with monitoring tools to provide a unified view of how resource constraints affect system health, making it easier to correlate actions with outcomes.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;Evaluate Resilience&lt;/strong&gt;&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;&lt;p&gt;Analyze how effectively your system scales up resources in response to the induced stress. This includes evaluating the timeliness of scaling actions and whether the added resources alleviate the performance issues.&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;Evaluate the efficiency of your resource allocation algorithms. This involves assessing whether resources are being utilized optimally and whether unnecessary wastage or contention exists.&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;Test the robustness of your failover and redundancy mechanisms under ‘conditions of resource scarcity’. This can include switching to standby systems, redistributing load among available resources, or degrading service gracefully.&lt;/p&gt;&lt;/li&gt;
&lt;/ul&gt;

&lt;h2&gt;
  
  
  3. Network Disruption Experiment
&lt;/h2&gt;

&lt;p&gt;&lt;strong&gt;Objective:&lt;/strong&gt; To simulate various network conditions that can affect a system’s operations, such as outages, DNS failures, or limited network access. By introducing these disruptions, the experiment seeks to understand how a system responds and adapts to network unreliability, ensuring critical applications can withstand and recover from real-world network issues.&lt;/p&gt;

&lt;h3&gt;
  
  
  Possible Experiments
&lt;/h3&gt;

&lt;ul&gt;
&lt;li&gt;&lt;p&gt;&lt;strong&gt;DNS Failures.&lt;/strong&gt; Introduce DNS resolution issues to evaluate the system’s reliance on DNS and its ability to use fallback DNS services.&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;&lt;strong&gt;Latency Injection.&lt;/strong&gt; Introduce artificial delay in the network to simulate high-latency conditions, affecting the communication between services or components.&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;&lt;strong&gt;Packet Loss Simulation.&lt;/strong&gt; Simulate the loss of data packets in the network to test how well the system handles data transmission errors and retries.&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;&lt;strong&gt;Bandwidth Throttling.&lt;/strong&gt; Limit the network bandwidth available to the application, simulating congestion conditions or degraded network services.&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;&lt;strong&gt;Connection Drops.&lt;/strong&gt; Forcing abrupt disconnections or intermittent connectivity to test session persistence and reconnection strategies.&lt;/p&gt;&lt;/li&gt;
&lt;/ul&gt;

&lt;h3&gt;
  
  
  How to Run a Network Disruption Experiment
&lt;/h3&gt;

&lt;p&gt;&lt;strong&gt;Identify Network Paths&lt;/strong&gt;&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;&lt;p&gt;Start by mapping out your network’s topology, including routers, switches, gateways, and the connections between different segments. Tools like &lt;a href="https://nmap.org/" rel="noopener noreferrer"&gt;Nmap&lt;/a&gt; or network diagram software can help visualize your network’s structure.&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;Focus on identifying the critical paths data takes when traveling through your system. These include paths between microservices, external APIs, databases, and the Internet.&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;Document these paths and prioritize them based on their importance to your system’s operation. This will help you decide where to start with your network disruption experiments.&lt;/p&gt;&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;&lt;strong&gt;Choose Disruption Type&lt;/strong&gt;&lt;/p&gt;

&lt;p&gt;Decide on the type of network disruption to simulate. Options include:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;&lt;p&gt;complete network outages,&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;latency (delays in data transmission)&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;packet loss (data packets being lost during transmission)&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;bandwidth limitations&lt;/p&gt;&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;Next, choose disruptions based on their likelihood and potential impact on your system.&lt;br&gt;
For example, simulating latency and packet loss might be particularly relevant if your system is distributed across multiple geographic locations.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;Use Network Chaos Tools&lt;/strong&gt;&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;Traffic Control (TC).&lt;/strong&gt; The ‘tc’ command in Linux is a powerful tool for controlling network traffic. It allows you to introduce delays, packet loss, and bandwidth restrictions on your network interfaces.&lt;/p&gt;

&lt;p&gt;⚠️ Note → Simulating DNS failures can be complex but is crucial for understanding how your system reacts to DNS resolution issues. Consider using specialized tools or features for this purpose.&lt;/p&gt;

&lt;p&gt;On the flip side, chaos experiment solutions like Steadybit provide user-friendly interfaces for simulating network disruptions. For example, you get safety features like built-in rollback strategies to minimize the risk of long-term impact on your system.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;Monitor Connectivity and Throughput&lt;/strong&gt;&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;&lt;p&gt;During the experiment, use network monitoring tools and observability platforms to track connectivity and throughput metrics in real time.&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;Focus on monitoring packet loss rates, latency, bandwidth usage, and error rates to assess the impact of the network disruptions you’re simulating.&lt;/p&gt;&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;&lt;strong&gt;Assess Failover and Recovery&lt;/strong&gt;&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;&lt;p&gt;Evaluate how well your system’s failover mechanisms respond to network disruptions. For example, you could switch to a redundant network path, use a different DNS server, or take other predefined recovery actions.&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;Measure the time it takes for the system to detect and recover the issue. This includes the time it takes to failover and return to normal operations after the disruption ends.&lt;/p&gt;&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;💡 Recommended → Analyze the overall resilience of your system to network instability. This assessment should include how well services degrade (if at all) and how quickly and effectively they recover once normal conditions are restored.&lt;/p&gt;

&lt;p&gt;If you want to read more, you can check out the &lt;a href="https://steadybit.com/blog/chaos-experiments/" rel="noopener noreferrer"&gt;rest of the post&lt;/a&gt; about principles of chaos engineering and popular tools here.&lt;/p&gt;

</description>
      <category>sre</category>
      <category>performance</category>
      <category>chaosengineering</category>
    </item>
  </channel>
</rss>
