<?xml version="1.0" encoding="UTF-8"?>
<rss version="2.0" xmlns:atom="http://www.w3.org/2005/Atom" xmlns:dc="http://purl.org/dc/elements/1.1/">
  <channel>
    <title>DEV Community: Edmond</title>
    <description>The latest articles on DEV Community by Edmond (@edmonddantes_14).</description>
    <link>https://dev.to/edmonddantes_14</link>
    <image>
      <url>https://media2.dev.to/dynamic/image/width=90,height=90,fit=cover,gravity=auto,format=auto/https:%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Fuser%2Fprofile_image%2F3560496%2F76eff56e-b42c-4b4b-b596-541a92dfcb2e.jpeg</url>
      <title>DEV Community: Edmond</title>
      <link>https://dev.to/edmonddantes_14</link>
    </image>
    <atom:link rel="self" type="application/rss+xml" href="https://dev.to/feed/edmonddantes_14"/>
    <language>en</language>
    <item>
      <title>Laravel async vs Octane</title>
      <dc:creator>Edmond</dc:creator>
      <pubDate>Fri, 03 Apr 2026 14:31:56 +0000</pubDate>
      <link>https://dev.to/edmonddantes_14/laravel-async-vs-octane-50pg</link>
      <guid>https://dev.to/edmonddantes_14/laravel-async-vs-octane-50pg</guid>
      <description>&lt;p&gt;How much can you speed up Laravel if you handle requests in coroutines instead of blocking workers? We benchmarked &lt;code&gt;TrueAsync&lt;/code&gt; (native PHP coroutines) against three Laravel Octane configurations. &lt;/p&gt;

&lt;h2&gt;
  
  
  The Benchmark
&lt;/h2&gt;

&lt;h3&gt;
  
  
  Environment
&lt;/h3&gt;

&lt;ul&gt;
&lt;li&gt;
&lt;strong&gt;OS:&lt;/strong&gt; WSL2 (Linux 5.15), 16 cores, 7.8 GB RAM&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;DB:&lt;/strong&gt; PostgreSQL 16 (&lt;code&gt;max_connections=500&lt;/code&gt;)&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;Load tool:&lt;/strong&gt; &lt;code&gt;k6&lt;/code&gt;, &lt;code&gt;constant-arrival-rate&lt;/code&gt;, 1,000 req/s for 30 seconds&lt;/li&gt;
&lt;/ul&gt;

&lt;div class="table-wrapper-paragraph"&gt;&lt;table&gt;
&lt;thead&gt;
&lt;tr&gt;
&lt;th&gt;Parameter&lt;/th&gt;
&lt;th&gt;TrueAsync FrankenPHP&lt;/th&gt;
&lt;th&gt;Octane Swoole (NTS)&lt;/th&gt;
&lt;th&gt;Octane Swoole (ZTS)&lt;/th&gt;
&lt;th&gt;Octane FrankenPHP&lt;/th&gt;
&lt;/tr&gt;
&lt;/thead&gt;
&lt;tbody&gt;
&lt;tr&gt;
&lt;td&gt;PHP&lt;/td&gt;
&lt;td&gt;8.6.0-dev (ZTS)&lt;/td&gt;
&lt;td&gt;8.5.4 (NTS)&lt;/td&gt;
&lt;td&gt;8.5.4 (ZTS)&lt;/td&gt;
&lt;td&gt;8.5.4 (NTS)&lt;/td&gt;
&lt;/tr&gt;
&lt;tr&gt;
&lt;td&gt;Server&lt;/td&gt;
&lt;td&gt;FrankenPHP (true-async fork)&lt;/td&gt;
&lt;td&gt;Swoole 6.2.0&lt;/td&gt;
&lt;td&gt;Swoole 6.2.0&lt;/td&gt;
&lt;td&gt;FrankenPHP (official)&lt;/td&gt;
&lt;/tr&gt;
&lt;tr&gt;
&lt;td&gt;Laravel&lt;/td&gt;
&lt;td&gt;13.2.0&lt;/td&gt;
&lt;td&gt;13.2.0&lt;/td&gt;
&lt;td&gt;13.2.0&lt;/td&gt;
&lt;td&gt;13.2.0&lt;/td&gt;
&lt;/tr&gt;
&lt;tr&gt;
&lt;td&gt;Model&lt;/td&gt;
&lt;td&gt;Coroutines (libuv)&lt;/td&gt;
&lt;td&gt;Processes (fork)&lt;/td&gt;
&lt;td&gt;Threads (ZTS)&lt;/td&gt;
&lt;td&gt;Processes (fork)&lt;/td&gt;
&lt;/tr&gt;
&lt;/tbody&gt;
&lt;/table&gt;&lt;/div&gt;

&lt;p&gt;&lt;strong&gt;Note:&lt;/strong&gt; &lt;code&gt;Swoole&lt;/code&gt; runs without coroutine mode here because Laravel is not adapted for it. In a pure synthetic test with coroutines, Swoole shows slightly better numbers than &lt;code&gt;FrankenPHP + TrueAsync&lt;/code&gt;. Both servers reach ~10,000 req/s with 12 workers on synthetic loads.&lt;/p&gt;

&lt;p&gt;Full benchmark repository: &lt;a href="https://github.com/YanGusik/ta_benchmark/" rel="noopener noreferrer"&gt;github.com/YanGusik/ta_benchmark&lt;/a&gt;&lt;/p&gt;

&lt;h3&gt;
  
  
  Workload
&lt;/h3&gt;

&lt;p&gt;The &lt;code&gt;/bench&lt;/code&gt; endpoint executes &lt;strong&gt;10 sequential SQL queries&lt;/strong&gt; against &lt;code&gt;PostgreSQL&lt;/code&gt;: user lookup, post listing, &lt;code&gt;INSERT&lt;/code&gt; a view record, &lt;code&gt;UPDATE&lt;/code&gt; a counter, aggregations, &lt;code&gt;TOP-N&lt;/code&gt; selections. Database: 100 users, 1,000 posts, growing post_views table.&lt;/p&gt;

&lt;p&gt;This is a realistic workload, not a synthetic "Hello World".&lt;/p&gt;

&lt;h2&gt;
  
  
  Results
&lt;/h2&gt;

&lt;h3&gt;
  
  
  Throughput (req/s)
&lt;/h3&gt;

&lt;div class="table-wrapper-paragraph"&gt;&lt;table&gt;
&lt;thead&gt;
&lt;tr&gt;
&lt;th&gt;Workers&lt;/th&gt;
&lt;th&gt;TrueAsync&lt;/th&gt;
&lt;th&gt;Swoole NTS&lt;/th&gt;
&lt;th&gt;Swoole ZTS&lt;/th&gt;
&lt;th&gt;FrankenPHP Octane&lt;/th&gt;
&lt;/tr&gt;
&lt;/thead&gt;
&lt;tbody&gt;
&lt;tr&gt;
&lt;td&gt;4&lt;/td&gt;
&lt;td&gt;&lt;strong&gt;989&lt;/strong&gt;&lt;/td&gt;
&lt;td&gt;183&lt;/td&gt;
&lt;td&gt;185&lt;/td&gt;
&lt;td&gt;189&lt;/td&gt;
&lt;/tr&gt;
&lt;tr&gt;
&lt;td&gt;8&lt;/td&gt;
&lt;td&gt;&lt;strong&gt;993&lt;/strong&gt;&lt;/td&gt;
&lt;td&gt;342&lt;/td&gt;
&lt;td&gt;341&lt;/td&gt;
&lt;td&gt;346&lt;/td&gt;
&lt;/tr&gt;
&lt;tr&gt;
&lt;td&gt;12&lt;/td&gt;
&lt;td&gt;&lt;strong&gt;990&lt;/strong&gt;&lt;/td&gt;
&lt;td&gt;483&lt;/td&gt;
&lt;td&gt;476&lt;/td&gt;
&lt;td&gt;489&lt;/td&gt;
&lt;/tr&gt;
&lt;tr&gt;
&lt;td&gt;16&lt;/td&gt;
&lt;td&gt;&lt;strong&gt;987&lt;/strong&gt;&lt;/td&gt;
&lt;td&gt;599&lt;/td&gt;
&lt;td&gt;601&lt;/td&gt;
&lt;td&gt;556&lt;/td&gt;
&lt;/tr&gt;
&lt;/tbody&gt;
&lt;/table&gt;&lt;/div&gt;

&lt;p&gt;With 16 workers, TrueAsync handles &lt;strong&gt;987 req/s&lt;/strong&gt;. The best &lt;code&gt;Octane&lt;/code&gt; result is 601 req/s (&lt;code&gt;Swoole ZTS&lt;/code&gt;), &lt;strong&gt;64% less&lt;/strong&gt; with the same worker count.&lt;/p&gt;

&lt;p&gt;We gave blocking servers 16 workers to be generous. &lt;code&gt;TrueAsync&lt;/code&gt; doesn't need them. &lt;strong&gt;Four workers&lt;/strong&gt; handle 989 req/s, the same as sixteen. Coroutines yield on every &lt;code&gt;PDO::query()&lt;/code&gt;, so one worker runs dozens of requests concurrently. While one coroutine waits for &lt;code&gt;PostgreSQL&lt;/code&gt;, others keep working.&lt;/p&gt;

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fkmfqlk03ou7evfedxhgf.png" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fkmfqlk03ou7evfedxhgf.png" alt="Throughput and Memory" width="800" height="340"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;h3&gt;
  
  
  Median Latency (P50)
&lt;/h3&gt;

&lt;div class="table-wrapper-paragraph"&gt;&lt;table&gt;
&lt;thead&gt;
&lt;tr&gt;
&lt;th&gt;Workers&lt;/th&gt;
&lt;th&gt;TrueAsync&lt;/th&gt;
&lt;th&gt;Swoole NTS&lt;/th&gt;
&lt;th&gt;Swoole ZTS&lt;/th&gt;
&lt;th&gt;FrankenPHP Octane&lt;/th&gt;
&lt;/tr&gt;
&lt;/thead&gt;
&lt;tbody&gt;
&lt;tr&gt;
&lt;td&gt;4&lt;/td&gt;
&lt;td&gt;&lt;strong&gt;28 ms&lt;/strong&gt;&lt;/td&gt;
&lt;td&gt;5,440 ms&lt;/td&gt;
&lt;td&gt;5,320 ms&lt;/td&gt;
&lt;td&gt;5,240 ms&lt;/td&gt;
&lt;/tr&gt;
&lt;tr&gt;
&lt;td&gt;8&lt;/td&gt;
&lt;td&gt;&lt;strong&gt;27 ms&lt;/strong&gt;&lt;/td&gt;
&lt;td&gt;2,870 ms&lt;/td&gt;
&lt;td&gt;2,900 ms&lt;/td&gt;
&lt;td&gt;2,800 ms&lt;/td&gt;
&lt;/tr&gt;
&lt;tr&gt;
&lt;td&gt;12&lt;/td&gt;
&lt;td&gt;&lt;strong&gt;28 ms&lt;/strong&gt;&lt;/td&gt;
&lt;td&gt;2,040 ms&lt;/td&gt;
&lt;td&gt;2,050 ms&lt;/td&gt;
&lt;td&gt;1,990 ms&lt;/td&gt;
&lt;/tr&gt;
&lt;tr&gt;
&lt;td&gt;16&lt;/td&gt;
&lt;td&gt;&lt;strong&gt;29 ms&lt;/strong&gt;&lt;/td&gt;
&lt;td&gt;1,640 ms&lt;/td&gt;
&lt;td&gt;1,660 ms&lt;/td&gt;
&lt;td&gt;1,780 ms&lt;/td&gt;
&lt;/tr&gt;
&lt;/tbody&gt;
&lt;/table&gt;&lt;/div&gt;

&lt;p&gt;29 ms vs 1,640 ms at 16 workers. &lt;strong&gt;56x&lt;/strong&gt; difference. Where do those seconds come from?&lt;/p&gt;

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2F4onx37y9mj0nl0ctqst3.png" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2F4onx37y9mj0nl0ctqst3.png" alt="Latency" width="800" height="479"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;div class="table-wrapper-paragraph"&gt;&lt;table&gt;
&lt;thead&gt;
&lt;tr&gt;
&lt;th&gt;Phase&lt;/th&gt;
&lt;th&gt;TrueAsync (4w)&lt;/th&gt;
&lt;th&gt;Swoole (4w)&lt;/th&gt;
&lt;/tr&gt;
&lt;/thead&gt;
&lt;tbody&gt;
&lt;tr&gt;
&lt;td&gt;PHP execution&lt;/td&gt;
&lt;td&gt;~5 ms&lt;/td&gt;
&lt;td&gt;~5 ms&lt;/td&gt;
&lt;/tr&gt;
&lt;tr&gt;
&lt;td&gt;SQL I/O wait (10 queries)&lt;/td&gt;
&lt;td&gt;~23 ms&lt;/td&gt;
&lt;td&gt;~23 ms&lt;/td&gt;
&lt;/tr&gt;
&lt;tr&gt;
&lt;td&gt;&lt;strong&gt;Queue wait&lt;/strong&gt;&lt;/td&gt;
&lt;td&gt;&lt;strong&gt;~0 ms&lt;/strong&gt;&lt;/td&gt;
&lt;td&gt;&lt;strong&gt;~5,400 ms&lt;/strong&gt;&lt;/td&gt;
&lt;/tr&gt;
&lt;tr&gt;
&lt;td&gt;&lt;strong&gt;Total&lt;/strong&gt;&lt;/td&gt;
&lt;td&gt;&lt;strong&gt;~28 ms&lt;/strong&gt;&lt;/td&gt;
&lt;td&gt;&lt;strong&gt;~5,440 ms&lt;/strong&gt;&lt;/td&gt;
&lt;/tr&gt;
&lt;/tbody&gt;
&lt;/table&gt;&lt;/div&gt;

&lt;p&gt;PHP and SQL run at identical speeds. The entire difference is &lt;strong&gt;queue wait&lt;/strong&gt;: a blocking server can't start your request until the current one finishes. With &lt;code&gt;TrueAsync&lt;/code&gt;, CPU utilization is higher because coroutines yield during I/O instead of blocking the worker.&lt;/p&gt;

&lt;p&gt;No magic. Just better resource utilization.&lt;/p&gt;

&lt;h3&gt;
  
  
  Memory (under load)
&lt;/h3&gt;

&lt;div class="table-wrapper-paragraph"&gt;&lt;table&gt;
&lt;thead&gt;
&lt;tr&gt;
&lt;th&gt;Workers&lt;/th&gt;
&lt;th&gt;TrueAsync&lt;/th&gt;
&lt;th&gt;Swoole ZTS&lt;/th&gt;
&lt;th&gt;FrankenPHP Octane&lt;/th&gt;
&lt;/tr&gt;
&lt;/thead&gt;
&lt;tbody&gt;
&lt;tr&gt;
&lt;td&gt;4&lt;/td&gt;
&lt;td&gt;&lt;strong&gt;277 MB&lt;/strong&gt;&lt;/td&gt;
&lt;td&gt;508 MB&lt;/td&gt;
&lt;td&gt;401 MB&lt;/td&gt;
&lt;/tr&gt;
&lt;tr&gt;
&lt;td&gt;8&lt;/td&gt;
&lt;td&gt;&lt;strong&gt;286 MB&lt;/strong&gt;&lt;/td&gt;
&lt;td&gt;600 MB&lt;/td&gt;
&lt;td&gt;417 MB&lt;/td&gt;
&lt;/tr&gt;
&lt;tr&gt;
&lt;td&gt;16&lt;/td&gt;
&lt;td&gt;&lt;strong&gt;308 MB&lt;/strong&gt;&lt;/td&gt;
&lt;td&gt;765 MB&lt;/td&gt;
&lt;td&gt;403 MB&lt;/td&gt;
&lt;/tr&gt;
&lt;/tbody&gt;
&lt;/table&gt;&lt;/div&gt;

&lt;p&gt;In the blocking model, each worker is a separate process with a full Laravel copy: container, configuration, router, middleware, database manager. With &lt;code&gt;TrueAsync&lt;/code&gt;, coroutines share a common bootstrap. Only per-request data (&lt;code&gt;request&lt;/code&gt;, &lt;code&gt;session&lt;/code&gt;, &lt;code&gt;auth&lt;/code&gt;) is duplicated. Hence 308 MB vs 765 MB at 16 workers.&lt;/p&gt;

&lt;h2&gt;
  
  
  Key Takeaways
&lt;/h2&gt;

&lt;p&gt;Tests are not something you should fully trust. Different scenarios are possible. However, there’s no magic here.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;Even at maximum workers, TrueAsync wins by 30-40%.&lt;/strong&gt; You could spin up 22-27 blocking workers to match throughput, but you'd still lose on latency and memory. And why use 22 workers when 4 will do?&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;The bottom line:&lt;/strong&gt; for &lt;code&gt;IO-bound&lt;/code&gt; workloads (which is most web apps), &lt;code&gt;TrueAsync&lt;/code&gt; serves the same traffic with &lt;strong&gt;5-6x fewer workers&lt;/strong&gt;, &lt;strong&gt;56x lower latency&lt;/strong&gt;, and &lt;strong&gt;half the memory&lt;/strong&gt;.&lt;/p&gt;

</description>
      <category>laravel</category>
      <category>php</category>
      <category>performance</category>
    </item>
  </channel>
</rss>
