DEV Community

Cover image for Exploring Code Performance Testing in C# with BenchmarkDotNet
Grontis Kostis
Grontis Kostis

Posted on • Originally published at kgrontis.github.io

Exploring Code Performance Testing in C# with BenchmarkDotNet

There have been many times where, while coding, I was facing the dilemma of choosing the appropriate LINQ method, or questioning the efficiency and speed of my code. During those moments, when I didn't want to just flip a coin or make random choices, I turned to performance testing. In this article, we will explore how we can test the performance of our C# code by using a library called BenchmarkDotNet!

Introduction

BenchmarkDotNet is a popular open-source library that, as stated in the repo's README.md, helps us to transform methods into benchmarks, track their performance, and share reproducible measurement experiments. Using BenchmarkDotNet feels similar to writing unit tests. It's very important to note that the library only works with console apps. Finally, we can visualize the results in the terminal where the benchmark ran or in user-friendly formats such as markdown, HTML and CSV. We will explore examples of there formats later in the article.

The code of the examples shown in the article can be found here.

Installation

First and foremost, we need to create a console application to run our benchmarks.
We can create a new one using the dotnet CLI in the terminal :

 dotnet new console -n Benchmarks
Enter fullscreen mode Exit fullscreen mode

Next, open the newly created project in your preferred IDE (I'll be using VS Code).
Install the BenchmarkDotNet NuGet package by including the following line in the .csproj file (Version 0.13.5 which is the latest version when this article was written) :

  <ItemGroup>
    <PackageReference Include="BenchmarkDotNet" Version="0.13.5" />
  </ItemGroup>
Enter fullscreen mode Exit fullscreen mode

Alternatively, we can run dotnet add package BenchmarkDotNet in the terminal.

With that, we are ready to start writing our benchmarks!

Setup

Lets create the methods that we want to benchmark. The example is going to be a comparison between two methods that return the sum of all the integers in an array.

using BenchmarkDotNet.Attributes;

namespace Benchmarks;

public class SumBenchmark
{
    readonly int[] arr = new int[] { 1, 2, 3, 6, 7, 8 };

    [Benchmark]
    public int SumWithForLoop()
    {
        int sum = 0;
        for (int i = 0; i < arr.Length; i++)
        {
            sum += arr[i];
        }
        return sum;
    }

    [Benchmark]
    public int SumWithLinqSum()
    {
        return arr.Sum();
    }
}
Enter fullscreen mode Exit fullscreen mode

All we have to do now to run our benchmarks and print the results to the console is to add BenchmarkRunner.Run<SumBenchmark>(); to Program.cs.

Our Program.cs should look something like this now:

using BenchmarkDotNet.Running;
using Benchmarks;

BenchmarkRunner.Run<SumBenchmark>();
Enter fullscreen mode Exit fullscreen mode

We should make sure that the project is build and runs in Release mode, so lets go ahead and do that.

dotnet build -c Release

dotnet run -c Release
Enter fullscreen mode Exit fullscreen mode

That was it, now we should be able to examine the results. We can find them in the BenchmarkDotNet.Artifacts/results folder.

Results

The results are being presented in three different user-friendly forms, markdown, HTML and CSV, as mentioned before. Lets see the results of our example :

  1. Markdown

Results in Markdown

  1. HTML

Results in HTML

  1. CSV (it's a long file so I'll show a part of it)

Results in CSV

In the results we can see, for now, the columns Method Mean Error & StdDev. Lets explore what these terms actually mean.

  1. Method. The name of the method being benchmarked.
  2. Mean. By definition it's the sum of a collection of numbers divided by the count of numbers in the collection. In the context of benchmarking, Mean refers to the average execution time of a method over multiple iterations.
  3. Error. It's a metric to let us understand the differences between each iteration. A smaller Error means the measurements were more consistent and closer to each other.
  4. StdDev (Standard Deviation). It measures the variation from the Mean.

Judging from these results, we can see that the SumWithForLoop method performs better than the SumWithLinqSum method on the dataset we provided.
We can have additional metrics by adding configurations, but we will explore these later on this article.

Global Setup and Global Cleanup

In our example, we have hardcoded the values in the array. What if we wanted to create data for a test database and when we are done with benchmarking, to cleanup this database? We wouldn't want to measure the logic of the data creation and retrieval, or database cleanup together with the other two functions. That's where GlobalSetup and GlobalCleanup attributes come into play.

Global Setup

We can setup any method as a setup method by adding the GlobalSetup attribute. This method will be executed once per benchmark method before the execution of the benchmarks. In our case we would move the creation of the array into the method decorated with the attibute.

    [GlobalSetup]
    public void Setup()
    {
        arr = new int[6];
        for (var i = 0; i < arr.Length; i++)
        {
            arr[i] = Random.Shared.Next(1, 10);
        }
    }
Enter fullscreen mode Exit fullscreen mode

Global Cleanup

The same goes for Global Cleanup. We just have to decorate the method that is cleaning up our resources with the attribute GlobalCleanup. This method will be executed once per benchmark method after the execution of the benchmarks.

    [GlobalCleanup]
    public void Cleanup()
    {
        // cleanup logic
        arr = Array.Empty<int>();
    }
Enter fullscreen mode Exit fullscreen mode

It is worth mentioning that there are two more similar attributes IterationSetup and IterationCleanup that as their name suggests, they run before and after each benchmark method iteration.

Parameters

Parameters, in BenchmarkDotNet, add a lot of flexibility into our benchmarks, allowing us to create benchmarks that adapt to different input scenarios and test a range of use cases. When we introduce parameters, the benchmarking process is repeated for each combination of parameter values therefore, it may show separate results for each parameter combination.
Lets try to adjust our benchmarks to use parameters.

    [Params(5, 10, 25)]
    public int ArraySize { get; set; }

    [GlobalSetup]
    public void Setup()
    {
        arr = new int[ArraySize];
        for (var i = 0; i < arr.Length; i++)
        {
            arr[i] = Random.Shared.Next(1, 10);
        }
    }
Enter fullscreen mode Exit fullscreen mode

The results after the parameter addition:

Method ArraySize Mean Error StdDev
SumWithForLoop 5 2.165 ns 0.0348 ns 0.0326 ns
SumWithLinqSum 5 3.222 ns 0.0097 ns 0.0081 ns
SumWithForLoop 10 4.713 ns 0.0038 ns 0.0032 ns
SumWithLinqSum 10 4.580 ns 0.1015 ns 0.1456 ns
SumWithForLoop 25 12.387 ns 0.1456 ns 0.1290 ns
SumWithLinqSum 25 9.102 ns 0.1543 ns 0.1652 ns

Notice the new column that appeared in our results, ArraySize.

Baselines

Baselines help us understand how well our code performs by giving us a starting point to measure against. They are like a friendly guide that tells us if our code is getting faster or slower. We can have benchmark baselines and job baselines. Lets set the SumWithForLoop method as a baseline, and lets create another variation of the Sum method with a while loop implementation.

    [Benchmark(Baseline = true)]
    public int SumWithForLoop()
    {
        int sum = 0;
        for (int i = 0; i < arr.Length; i++)
        {
            sum += arr[i];
        }
        return sum;
    }

    [Benchmark]
    public int SumWithWhileLoop()
    {
        int sum = 0;
        int counter = 0;
        while (counter < arr.Length)
        {
            sum = arr[counter];
            counter++;
        }
        return sum;
    }

    [Benchmark]
    public int SumWithLinqSum()
    {
        return arr.Sum();
    }
Enter fullscreen mode Exit fullscreen mode

In the summary table we should see two additional columns Ratio and RatioSD.

Method ArraySize Mean Error StdDev Ratio RatioSD
SumWithForLoop 5 2.146 ns 0.0363 ns 0.0339 ns 1.00 0.00
SumWithWhileLoop 5 2.261 ns 0.0727 ns 0.0714 ns 1.05 0.05
SumWithLinqSum 5 3.353 ns 0.0932 ns 0.1244 ns 1.57 0.07
SumWithForLoop 10 4.870 ns 0.0909 ns 0.0759 ns 1.00 0.00
SumWithWhileLoop 10 5.274 ns 0.1219 ns 0.1403 ns 1.07 0.04
SumWithLinqSum 10 4.588 ns 0.0966 ns 0.1385 ns 0.95 0.03
SumWithForLoop 25 12.277 ns 0.1340 ns 0.1046 ns 1.00 0.00
SumWithWhileLoop 25 12.701 ns 0.2838 ns 0.2915 ns 1.04 0.02
SumWithLinqSum 25 9.030 ns 0.1183 ns 0.1106 ns 0.73 0.01

Since the SumWithForLoop method is the baseline, the Ratio is 1 and RatioSD 0. But what do these values mean for the other methods? Ratio is calculated by dividing each measurement of a benchmark to the equivalent measurement of the baseline and then calculating the Mean of all these values. So in our example, the SumWithWhileLoop method is slower than the SumWithForLoop by 0.05 times. RatioSD, which means Ratio Standard Deviation, represents
the consistency of the ratios between two measurements. If the RatioSD is small, it means the comparison is stable, while a high RatioSD means there's more variation in the comparison.

Configurations

BenchmarkDotNet gives us the option to configure the whole benchmarking process to cover everything we might need, from Jobs and Diagnosers, to Loggers and Exporters. Lets go through some of my most used ones!

Diagnosers

Diagnosers play a huge role within BenchmarkDotNet. They provide additional statistics and performance analysis. Whether we need to track memory allocation, monitor hardware performance counters to capture event tracing data, Diagnosers have us covered.

Lets look at an example on the usage. In the example we were using before, lets add a MemoryDiagnoser , which measures the number of allocated bytes and the GC frequency, and the HardwareCounters, which are a set of special-purpose registers built into modern microprocessors to store the counts of hardware-related activities within computer systems. We can do that by adding an attribute for each diagnoser:

[MemoryDiagnoser]
[HardwareCounters(
        HardwareCounter.BranchMispredictions,
        HardwareCounter.BranchInstructions)]
public class SumBenchmark
Enter fullscreen mode Exit fullscreen mode

The final results now include the new columns we specified

Method ArraySize Mean Error StdDev BranchInstructions/Op BranchMispredictions/Op Allocated
SumWithForLoop 5 3.239 ns 0.0617 ns 0.0547 ns 11 0 -
SumWithLinqSum 5 6.598 ns 0.0836 ns 0.0741 ns 17 0 -
SumWithForLoop 10 7.828 ns 0.2991 ns 0.8773 ns 21 0 -
SumWithLinqSum 10 8.288 ns 0.1952 ns 0.3097 ns 27 0 -
SumWithForLoop 25 17.850 ns 0.3752 ns 0.5381 ns 50 0 -
SumWithLinqSum 25 14.094 ns 0.2121 ns 0.1656 ns 56 0 -

Jobs

Jobs refer to the configuration options that define the target environment for benchmarking. They allow us to specify various parameters related to the runtime, platform, and launch settings, enabling us to simulate different execution scenarios during benchmark runs. With jobs, we can control factors such as the runtime version, runtime iteration mode (single or multithreaded), target platform (x86 or x64), and launch count and warm-up count, among others. By defining different jobs, we can explore how our code performs under various conditions and environments. Lets see an example :

[SimpleJob(BenchmarkDotNet.Jobs.RuntimeMoniker.Net70, baseline: true)]
[SimpleJob(BenchmarkDotNet.Jobs.RuntimeMoniker.Net60)]
public class SumBenchmark
Enter fullscreen mode Exit fullscreen mode

In the above code, we have added the baseline: true parameter to the SimpleJob attribute. This parameter designates the .NET 7.0 job as the baseline job, indicating that it is the reference against which other jobs will be compared. This helps in benchmarking and comparing performance across different frameworks or runtime versions.

Then we can run the benchmarks in multiple frameworks by specifying the target framework moniker names via --runtimes|-r console argument, in our case :

dotnet run -c Release -f net7.0 --runtimes net6.0 net7.0 
Enter fullscreen mode Exit fullscreen mode

You can read more about targeting multiple frameworks here.

Conclusion

In conclusion, by leveraging BenchmarkDotNet, we can make informed decisions about our code's performance, identify bottlenecks, and optimize our implementations effectively. With performance testing as a valuable tool in our development toolkit, we can ensure that our code performs optimally and meets the requirements of demanding applications.

You can read further about BenchmarkDotNet here :
Official documentation
GitHub repo
Adam Sitnik's blog Adam is one of the maintainers of the library and his blogs are pretty cool.
Wojciech Nagórski's blog Wojciech is a contributor to BenchmarkDotNet and has a lot of cool articles about the library as well!

Top comments (1)

Collapse
 
nickstavrou profile image
NickStavrou

after this top notch explanation of benchmarkdonet i will start using it right away.