<?xml version="1.0" encoding="UTF-8"?>
<rss version="2.0" xmlns:atom="http://www.w3.org/2005/Atom" xmlns:dc="http://purl.org/dc/elements/1.1/">
  <channel>
    <title>DEV Community: RunC.AI</title>
    <description>The latest articles on DEV Community by RunC.AI (@runc_ai).</description>
    <link>https://dev.to/runc_ai</link>
    <image>
      <url>https://media2.dev.to/dynamic/image/width=90,height=90,fit=cover,gravity=auto,format=auto/https:%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Forganization%2Fprofile_image%2F10738%2Fe4694b56-49b7-43db-815c-7994626853a0.png</url>
      <title>DEV Community: RunC.AI</title>
      <link>https://dev.to/runc_ai</link>
    </image>
    <atom:link rel="self" type="application/rss+xml" href="https://dev.to/feed/runc_ai"/>
    <language>en</language>
    <item>
      <title>Deploying DeepSeekR1-32B on RunC.AI</title>
      <dc:creator>RunC.AI Offical</dc:creator>
      <pubDate>Fri, 04 Jul 2025 10:24:21 +0000</pubDate>
      <link>https://dev.to/runc_ai/deploying-deepseekr1-32b-on-runcai-12ab</link>
      <guid>https://dev.to/runc_ai/deploying-deepseekr1-32b-on-runcai-12ab</guid>
      <description>&lt;p&gt;Welcome everybody, to another RunC.AI tutorial. This time we will still be playing with DeepSeek, except we are going to use the Ubuntu system image. Now let us start this tutorial.&lt;/p&gt;

&lt;h3&gt;
  
  
  &lt;strong&gt;First and foremost&lt;/strong&gt;
&lt;/h3&gt;

&lt;p&gt;&lt;a href="https://runc.ai/?ytag=rc_dev_devblog0704" rel="noopener noreferrer"&gt;Login to your account&lt;/a&gt; as always and click deploy. Scroll down to Image and click System image, this time we will be using &lt;strong&gt;Ubuntu&lt;/strong&gt;.&lt;/p&gt;

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Ftg9jy851cbdmzauonc6t.png" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Ftg9jy851cbdmzauonc6t.png" alt="Choose System Image" width="800" height="181"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;Then you will click the login button on the right&lt;/p&gt;

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fkd1dj19fbrp5tnbf2hmv.png" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fkd1dj19fbrp5tnbf2hmv.png" alt="Login" width="800" height="65"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;You will then see a page where you need to enter the password&lt;/p&gt;

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fhsowj5yv9hal4925fjzs.png" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fhsowj5yv9hal4925fjzs.png" alt="Enter Password" width="800" height="453"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;You can find your password here&lt;/p&gt;

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2F1p5f4rztraae1wm0p7hf.png" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2F1p5f4rztraae1wm0p7hf.png" alt="How to find password" width="800" height="66"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;h3&gt;
  
  
  &lt;strong&gt;Deploy Ollama&lt;/strong&gt;
&lt;/h3&gt;

&lt;p&gt;Once you get in the Ubuntu terminal, type in the following command to install ollama.&lt;/p&gt;

&lt;blockquote&gt;
&lt;p&gt;&lt;code&gt;curl -fsSL https://ollama.com/install.sh | sh&lt;/code&gt;&lt;/p&gt;
&lt;/blockquote&gt;

&lt;p&gt;By default, after the installation is completed, there will be an ollama.service file. In order to enable the local host and Docker containers to communicate with each other, the Environment variable needs to be modified to "OLLAMA_HOST=0.0.0.0:11434"&lt;/p&gt;

&lt;blockquote&gt;
&lt;p&gt;&lt;code&gt;sudo vim /etc/systemd/system/ollama.service&lt;/code&gt;&lt;/p&gt;
&lt;/blockquote&gt;

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fbichybnplf6csvvn2cq4.png" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fbichybnplf6csvvn2cq4.png" alt="Modify the environment variable" width="800" height="245"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;Now we need to restart Ollama&lt;/p&gt;

&lt;blockquote&gt;
&lt;p&gt;&lt;code&gt;sudo systemctl daemon-reload&lt;/code&gt;&lt;br&gt;
&lt;code&gt;sudo systemctl restart ollama&lt;/code&gt;&lt;/p&gt;
&lt;/blockquote&gt;

&lt;p&gt;Now we can pull the DeepSeek-R1 Model&lt;/p&gt;

&lt;blockquote&gt;
&lt;p&gt;&lt;code&gt;ollama run deepseek-r1:32b&lt;/code&gt;&lt;/p&gt;
&lt;/blockquote&gt;

&lt;h3&gt;
  
  
  &lt;strong&gt;Open-WebUI&lt;/strong&gt;
&lt;/h3&gt;

&lt;p&gt;Now we need to pull the Open-WebUI Image&lt;/p&gt;

&lt;p&gt;First, follow the Nvidia official website to download and config Nvidia CUDA container toolkit.&lt;/p&gt;

&lt;p&gt;&lt;a href="https://docs.nvidia.com/datacenter/cloud-native/container-toolkit/latest/install-guide.html" rel="noopener noreferrer"&gt;https://docs.nvidia.com/datacenter/cloud-native/container-toolkit/latest/install-guide.html&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;Then type in the following command&lt;/p&gt;

&lt;blockquote&gt;
&lt;p&gt;&lt;code&gt;sudo docker pull ghcr.io/open-webui/open-webui:cuda&lt;/code&gt;&lt;/p&gt;
&lt;/blockquote&gt;

&lt;p&gt;In order to enable webui within the container to communicate with Ollama on the external host, it is necessary to allow the Docker container to directly use the host network&lt;/p&gt;

&lt;blockquote&gt;
&lt;p&gt;&lt;code&gt;docker run -d --network=host \&lt;/code&gt;&lt;br&gt;
&lt;code&gt;-v open-webui:/app/backend/data \&lt;/code&gt;&lt;br&gt;
&lt;code&gt;-e OLLAMA_BASE_URL=http://127.0.0.1:11434 \&lt;/code&gt; &lt;br&gt;
&lt;code&gt;--name open-webui --restart always \&lt;/code&gt; &lt;br&gt;
&lt;code&gt;ghcr.io/open-webui/open-webui:cuda&lt;/code&gt;&lt;/p&gt;
&lt;/blockquote&gt;

&lt;p&gt;To get access to Open WebUI, visit &lt;a href="http://IP:8080" rel="noopener noreferrer"&gt;http://IP:8080&lt;/a&gt; where "IP" is your IP address which you can find in the following picture&lt;/p&gt;

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2F56nync70m2o0b8t4eiu4.png" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2F56nync70m2o0b8t4eiu4.png" alt="How to find the IP" width="800" height="64"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2F7kez3pk7bp727n2qlmx0.png" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2F7kez3pk7bp727n2qlmx0.png" alt="Deploy successfully" width="800" height="398"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;Now, you can ask deepseek any question you want.&lt;/p&gt;

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fmh9h1fukpmyz7ihjb147.png" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fmh9h1fukpmyz7ihjb147.png" alt="Try it" width="800" height="376"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;h3&gt;
  
  
  &lt;strong&gt;About &lt;a href="https://runc.ai/?ytag=rc_dev_devblog0704" rel="noopener noreferrer"&gt;RunC.AI&lt;/a&gt;&lt;/strong&gt;
&lt;/h3&gt;

&lt;p&gt;Rent smart, run fast. &lt;a href="https://runc.ai/?ytag=rc_dev_devblog0704" rel="noopener noreferrer"&gt;RunC.AI&lt;/a&gt; allows users to gain access to a wide selection of scalable, high-performance GPU instances and clusters at competitive prices compared to major cloud providers like Amazon Web Services (AWS), Google Cloud, and Microsoft Azure.&lt;/p&gt;

</description>
      <category>deepseek</category>
      <category>development</category>
      <category>chatgpt</category>
      <category>ai</category>
    </item>
    <item>
      <title>How to deploy ComfyUI on RunC.AI</title>
      <dc:creator>RunC.AI Offical</dc:creator>
      <pubDate>Fri, 30 May 2025 06:13:50 +0000</pubDate>
      <link>https://dev.to/runc_ai/how-to-deploy-comfyui-on-runcai-14b</link>
      <guid>https://dev.to/runc_ai/how-to-deploy-comfyui-on-runcai-14b</guid>
      <description>&lt;p&gt;Welcome to our first deployment tutorial! This tutorial is designed to give you an idea of how to deploy ComfyUI with RunC. The steps are simple, don't worry! I'm sure even a pure novice can handle it! Let's goooo!!&lt;/p&gt;

&lt;h2&gt;
  
  
  &lt;strong&gt;1.  Deploy ComfyUI on RunC&lt;/strong&gt;
&lt;/h2&gt;

&lt;p&gt;First of all, sign in/up &lt;a href="https://runc.ai/" rel="noopener noreferrer"&gt;RunC.AI | Run clever cloud computing for AI&lt;/a&gt; (New customers before 6th, June 2025 get $5 in free credits, about 12h of 4090 usage time)&lt;/p&gt;

&lt;p&gt;Secondly, enter the console. Go to "Instance" and click "Deploy" to start creating the image.&lt;/p&gt;

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2F38vb0knf4roo6i6cy5kr.png" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2F38vb0knf4roo6i6cy5kr.png" alt="Image description" width="800" height="370"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;On this page, choose the GPU model and select the image you want to deploy or click system image to switch from different system.&lt;/p&gt;

&lt;p&gt;Here I chose ComfyUI. Select billing cycle you want and deployment phase is done.&lt;/p&gt;

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fe7gfv7ck47qtwrcptk4c.png" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fe7gfv7ck47qtwrcptk4c.png" alt="Image description" width="800" height="400"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;NOTE: You can see the status of your instance on the dashboard. If the status is 'running', even if you are not using it, the renting fee is still counting. Also, stopping the instance will still incur renting fees. If you would like to avoid fees, please delete the instance.&lt;/p&gt;

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fc9419931ucl3glrlbj73.png" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fc9419931ucl3glrlbj73.png" alt="Image description" width="800" height="324"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;h2&gt;
  
  
  &lt;strong&gt;2.  How to create a text-to-image workflow in the comfyUI&lt;/strong&gt;
&lt;/h2&gt;

&lt;p&gt;Next, I take you step by step to build the workflow of the text-to-image. You do not need to take notes to remember these functions when you first get started. The key here is to practice and master it through actual operation with deep understanding of each function.&lt;/p&gt;

&lt;p&gt;First, if you are the first time initiating the ComfyUI interface, then there should be a default text-to-image workflow.&lt;/p&gt;

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fseifgfqjiyacxdpe7lpd.png" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fseifgfqjiyacxdpe7lpd.png" alt="Image description" width="800" height="400"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;The next workflow we're going to build is this default text-to-image diagram, the purpose of which is to let you build the workflow yourself.&lt;/p&gt;

&lt;p&gt;Then we click 'New' in the top left ribbon to create a new interface.&lt;/p&gt;

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2F2o4l1n67xyxm5a8jl3o8.png" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2F2o4l1n67xyxm5a8jl3o8.png" alt="Image description" width="422" height="409"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;Step 1: Add "K Sampler"&lt;/strong&gt;&lt;/p&gt;

&lt;p&gt;In the blank space of the workspace, single right mouse button to bring up the node ribbon, select Sampling - K Sampler, then we will add a sampler node in the workspace.&lt;/p&gt;

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2F10oibb6hz376u8n17a4a.png" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2F10oibb6hz376u8n17a4a.png" alt="Image description" width="687" height="512"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fiy8r7uv3oihrmyovaabl.png" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fiy8r7uv3oihrmyovaabl.png" alt="Image description" width="441" height="425"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;First, let's introduce a few parameter settings in the "K Sampler":&lt;/p&gt;

&lt;p&gt;The first parameter "seed": corresponds to the seed value in the webUI, will display the value of the image every time it is generated, the default is 0.&lt;/p&gt;

&lt;p&gt;The second parameter "control_after_generate": includes four options - fixed, increasing, decreasing, and random.&lt;/p&gt;

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fpzq6y34ggpzqsmdinv6b.png" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fpzq6y34ggpzqsmdinv6b.png" alt="Image description" width="478" height="458"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;These two parameters are used together, "fixed" represents the value of the fixed image, and the increasing/decreasing is +1 or -1 value. "random" represents a random value. Generally we use fixed or random.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;Step 2: Add "Large Models"&lt;/strong&gt;&lt;/p&gt;

&lt;p&gt;As shown in the figure, you can drag the model connection point and add the "Load Checkpoint".&lt;/p&gt;

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2F0kz2zk6qmh8m5eht45l9.png" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2F0kz2zk6qmh8m5eht45l9.png" alt="Image description" width="800" height="343"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;Step 3: Add "Positive &amp;amp; Negative Prompt Words"&lt;/strong&gt;&lt;/p&gt;

&lt;p&gt;Add a positive prompt word input node in the same way as above, as shown in the figure:&lt;/p&gt;

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fxcb6mh7u7sqlb8bu5ho4.png" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fxcb6mh7u7sqlb8bu5ho4.png" alt="Image description" width="625" height="435"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;The negative CLIPTextEncode can be added in the same way, i.e. by dragging the 'negative' connection point.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;Step 4: Add "Image Size/Batch"&lt;/strong&gt;&lt;/p&gt;

&lt;p&gt;Drag the "latent_image" connection point and select "Empty Latent Image" to add the "Image Size/Batch" node, which has width, height and batch size parameters.&lt;/p&gt;

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2F7io9pnpsu6egmwrnhjeb.png" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2F7io9pnpsu6egmwrnhjeb.png" alt="Image description" width="681" height="593"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;Step 5: Add "VAE Decoder"&lt;/strong&gt;&lt;/p&gt;

&lt;p&gt;Drag the "Latent" connection point and select "VAEDecode" to add VAE.&lt;/p&gt;

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fan4adentmfentapnt4uw.png" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fan4adentmfentapnt4uw.png" alt="Image description" width="800" height="414"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;Step 6: Add "Image Generation Area"&lt;/strong&gt;&lt;/p&gt;

&lt;p&gt;Drag the "Image" connection point and select "Preview Image" to successfully add the image generation area.&lt;/p&gt;

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fdufo9cb0cfyc3qtb78ig.png" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fdufo9cb0cfyc3qtb78ig.png" alt="Image description" width="800" height="377"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;At this point, all the nodes needed for the entire text-generated graph have been successfully added, and we've entered a prompt in the forward prompts (e.g. a border collie).&lt;/p&gt;

&lt;p&gt;If at this point you do what I did and click Queue, then you will get this error report, look at the red box in the figure for the prompt and the red markings on the nodes. The reason this is happening is because we have these red labeled nodes that are not connected.&lt;/p&gt;

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fo915b0538m0jwjzuol08.png" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fo915b0538m0jwjzuol08.png" alt="Image description" width="800" height="400"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;Then next we need to fix these error nodes and connect them all together. Pay attention to the color on the connection point. If it is a yellow connection point you need to connect to the corresponding yellow connection point (name correspondence), as shown in the figure (I mark them with green arrows) :&lt;/p&gt;

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2F3qdrxjb43f61zkaqw834.png" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2F3qdrxjb43f61zkaqw834.png" alt="Image description" width="800" height="400"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;Then click "Generate" again. Congratulations! This text-to-image workflow has been successfully built.&lt;/p&gt;

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fzt84g4vfqpw9ma0iaxqb.png" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fzt84g4vfqpw9ma0iaxqb.png" alt="Image description" width="800" height="400"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;Below to add a little more:&lt;/p&gt;

&lt;p&gt;The image generation area selected above are preview images and needs to be manually saved. Right-click on the image, and select "Save Image" to download the image.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;Comments&lt;/strong&gt;&lt;br&gt;
The whole process of building ComfyUI text-to-image workflow on RunC.AI platform is less than half an hour from registration to the completion of the workflow, which is very easy to understand and convenient. Unlike the complexity and high cost of local deployment, RunC.AI solves these problems perfectly.&lt;/p&gt;

&lt;p&gt;RunC.AI's interface is simple and intuitive so even first-time AI users can get started easily. When generating images, the response speed is very fast with almost no lag. And there were no failures or errors reported during the entire process.&lt;/p&gt;

&lt;p&gt;Whether you are a novice or an experienced user, you can find the right tools and resources for you on RunC.AI. In the future, I'll post more features and try to build different types of workflows. Stay tuned!&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;About RunC.AI&lt;/strong&gt;&lt;/p&gt;

&lt;p&gt;Rent smart, run fast. RunC.AI allows users to gain access to a wide selection of scalable, high-performance GPU instances and clusters at competitive prices compared to major cloud providers like Amazon Web Services (AWS), Google Cloud, and Microsoft Azure.&lt;/p&gt;

&lt;p&gt;Written by:&lt;br&gt;
Ashley Morgan&lt;br&gt;
Product Manager from RunC.AI&lt;br&gt;
More information:&lt;a href="https://blog.runc.ai/how-to-deploy-comfyui-on-runc-ai?ytag=rc_dev" rel="noopener noreferrer"&gt;RunC.AI Blog&lt;/a&gt;&lt;br&gt;
Join our Discord:&lt;a href="https://discord.gg/Pb3VArQBbX" rel="noopener noreferrer"&gt;Run.AI Community&lt;/a&gt;&lt;/p&gt;

</description>
      <category>discuss</category>
      <category>tutorial</category>
      <category>opensource</category>
      <category>ai</category>
    </item>
    <item>
      <title>Why 8 RTX 4090 Delivers Superior Performance/Cost Over 8 A6000 Ada: A Deep Dive</title>
      <dc:creator>RunC.AI Offical</dc:creator>
      <pubDate>Mon, 26 May 2025 10:01:25 +0000</pubDate>
      <link>https://dev.to/runc_ai/why-8x-rtx-4090-delivers-superior-performancecost-over-8x-a6000-ada-a-deep-dive-558j</link>
      <guid>https://dev.to/runc_ai/why-8x-rtx-4090-delivers-superior-performancecost-over-8x-a6000-ada-a-deep-dive-558j</guid>
      <description>&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fmyvj1anwuo1bzj51ytse.png" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fmyvj1anwuo1bzj51ytse.png" alt="Image description" width="800" height="359"&gt;&lt;/a&gt;In the realm of AI model training, rendering, and high-performance computing (HPC), selecting the right GPU is critical—not just for performance but also for budget efficiency. For years, professionals have gravitated toward NVIDIA’s professional GPUs such as the A6000 Ada, known for its robust memory, ECC support, and driver certification. However, the emergence of the RTX 4090—a gaming-class GPU—has disrupted that norm by offering &lt;em&gt;comparable or superior performance&lt;/em&gt; in many real-world scenarios at a fraction of the price.&lt;/p&gt;

&lt;p&gt;This article explores why deploying an 8× RTX 4090 configuration on RunC.AI can be significantly more cost-effective than 8× A6000 Ada GPUs, based on performance, practical deployment considerations, and real-world use cases.&lt;/p&gt;

&lt;h2&gt;
  
  
  1. Architecture and Specification Comparison
&lt;/h2&gt;

&lt;p&gt;Although both GPUs use the Ada Lovelace architecture, the RTX 4090 is tuned for peak performance in consumer workloads, whereas the A6000 Ada targets reliability and long-duration professional use. Here's how they stack up:&lt;/p&gt;

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fg1h31e4hfnsbf1rkrehg.png" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fg1h31e4hfnsbf1rkrehg.png" alt="Image description" width="573" height="486"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;Key Insight: While the A6000 Ada has more VRAM and slightly more cores, the RTX 4090 offers faster memory (GDDR6X), higher clocks, and stronger out-of-box performance in many mixed-precision workloads like FP16 and BF16, which dominate modern AI training.&lt;/p&gt;

&lt;h2&gt;
  
  
  2. Real-World Performance Benchmarks
&lt;/h2&gt;

&lt;p&gt;&lt;strong&gt;AI Training and Inference&lt;/strong&gt;&lt;br&gt;
Although RunC.AI currently focuses on providing access to RTX 4090 GPUs, its user-reported and internal benchmarks offer strong insight into how the 4090 compares to enterprise-class GPUs like the A6000 Ada. For typical AI workloads—such as fine-tuning transformer models (e.g., LLaMA, GPT-2), training diffusion models, and running large-scale inference—RTX 4090 consistently delivers performance that rivals or even exceeds that of the A6000 Ada.&lt;br&gt;
This is due to:&lt;br&gt;
●Higher clock speeds and newer memory (GDDR6X) on the 4090&lt;br&gt;
●Superior FP16/BF16 throughput, which many modern AI frameworks now rely on&lt;br&gt;
●Efficient multi-GPU scaling using frameworks like DeepSpeed and ZeRO-Offload&lt;/p&gt;

&lt;p&gt;Users on RunC.AI report that training times using RTX 4090 instances are highly competitive, often 5–10% faster than what was previously achieved on A6000 Ada hardware, especially in tasks that do not demand over 24GB of VRAM per GPU.&lt;/p&gt;

&lt;p&gt;By offering RTX 4090s at a significantly lower cost than A6000 Ada-based cloud services, RunC.AI enables researchers and developers to complete training workloads faster and at dramatically better cost-efficiency.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;Rendering and Simulation&lt;/strong&gt;&lt;br&gt;
In rendering tasks, third-party benchmarks show the RTX 4090 outperforming the A6000 Ada by 15–20% in tools like Blender, thanks to its higher boost clocks and aggressive thermal design. While RunC.AI focuses primarily on compute workloads, users performing GPU-based rendering (e.g., using Stable Diffusion or 3D model preprocessing) benefit from the 4090’s fast throughput and high memory bandwidth.&lt;/p&gt;

&lt;p&gt;Combined with RunC.AI’s pay-per-use pricing model and scalable infrastructure, the 4090 becomes an extremely attractive option—even for professional workflows typically reserved for workstation GPUs.&lt;/p&gt;

&lt;h2&gt;
  
  
  3. Performance/Cost Ratio: The Game-Changer
&lt;/h2&gt;

&lt;p&gt;The single biggest advantage of using RTX 4090 lies in cost efficiency. Here's a direct system-level comparison for a machine with 8 GPUs:&lt;/p&gt;

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2F0huijah1censgcnhbjyl.png" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2F0huijah1censgcnhbjyl.png" alt="Image description" width="574" height="478"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;That’s a massive savings with virtually no performance penalty in many workloads. For startups, universities, or individual researchers, this efficiency can drastically reduce infrastructure budgets or multiply compute resources for the same cost.&lt;/p&gt;

&lt;h2&gt;
  
  
  4. Potential Limitations and Considerations
&lt;/h2&gt;

&lt;p&gt;Of course, the 4090 isn’t a perfect drop-in replacement for professional-class GPUs. There are trade-offs:&lt;br&gt;
&lt;strong&gt;Driver and Certification:&lt;/strong&gt;&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;&lt;p&gt;The A6000 Ada is designed with enterprise-grade drivers and is certified for many professional applications (CAD, DCC, etc.).&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;The 4090 lacks such certification, though it's rarely a problem in open-source AI/ML workflows.&lt;/p&gt;&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;&lt;strong&gt;VRAM and ECC:&lt;/strong&gt;&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;&lt;p&gt;48GB of ECC VRAM on the A6000 Ada is advantageous for large-scale datasets or simulation.&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;However, modern training frameworks now allow model partitioning, gradient offloading, and checkpointing—making 24GB sufficient in most setups.&lt;/p&gt;&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;&lt;strong&gt;Form Factor, Cooling, and Power:&lt;/strong&gt;&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;&lt;p&gt;The 4090 is larger, consumes more power (450W vs 300W), and requires careful thermal management.&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;8× 4090 setups may need water-cooling, riser cables, and custom chassis (e.g., 4U high-density GPU servers).&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;Yet, platforms like RunC.AI have already proven stable multi-4090 deployments at scale.&lt;/p&gt;&lt;/li&gt;
&lt;/ul&gt;

&lt;h2&gt;
  
  
  5. Ecosystem &amp;amp; Deployment
&lt;/h2&gt;

&lt;p&gt;Cloud GPU providers like RunC.AI are standardizing on RTX 4090s because of their strong value proposition. For those building clusters or lab environments, system integrators are optimizing for these GPUs by balancing airflow, power delivery, and PCIe bandwidth.&lt;/p&gt;

&lt;p&gt;The emergence of server-grade boards with consumer GPU support (e.g., Supermicro’s 8-GPU platforms) makes 4090-based HPC more accessible than ever.&lt;/p&gt;

&lt;h2&gt;
  
  
  Conclusion: 4090 Makes High-End Compute Affordable
&lt;/h2&gt;

&lt;p&gt;The data is clear: an 8× RTX 4090 setup not only competes with but often surpasses the 8× A6000 Ada configuration in practical performance—all while costing less than one-third as much.&lt;/p&gt;

&lt;p&gt;Unless your use case absolutely requires ECC memory, driver certification, or ultra-large VRAM per GPU, the RTX 4090 is the best bang for the buck in AI research, rendering, and heavy computation in 2025.&lt;/p&gt;

&lt;p&gt;For AI startups, university labs, and independent researchers, this performance-per-dollar advantage is a rare opportunity to do more with less—without compromising compute power.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;About RunC.AI&lt;/strong&gt;&lt;br&gt;
Rent smart, run fast. Headquartered in Singapore, RunC.AI allows users to gain access to a wide selection of scalable, high-performance GPU instances and clusters at competitive prices compared to major cloud providers like Amazon Web Services (AWS), Google Cloud, and Microsoft Azure.&lt;/p&gt;

&lt;p&gt;Free credits are still available. Sign up now!&lt;br&gt;
(Due 6th, June 2025)&lt;br&gt;
Start your journey here:&lt;a href="https://www.runc.ai?ytag=rc_dev_0528" rel="noopener noreferrer"&gt;RunC.AI Official&lt;/a&gt;&lt;br&gt;
Share your user story in RunC.AI's discord server, chance to win secret prize! &lt;a href="https://discord.gg/Pb3VArQBbX" rel="noopener noreferrer"&gt;RunC.AI Community&lt;/a&gt;&lt;/p&gt;

</description>
      <category>ai</category>
      <category>webdev</category>
      <category>machinelearning</category>
      <category>github</category>
    </item>
    <item>
      <title>Why Should You Choose Renting Cloud GPU?</title>
      <dc:creator>RunC.AI Offical</dc:creator>
      <pubDate>Thu, 15 May 2025 05:57:18 +0000</pubDate>
      <link>https://dev.to/runc_ai/why-should-you-choose-renting-cloud-gpu-a6h</link>
      <guid>https://dev.to/runc_ai/why-should-you-choose-renting-cloud-gpu-a6h</guid>
      <description>&lt;p&gt;Think about the times when you have an urgent deadline about an AI project or application development, no time for debugging and scalability is the key. What's next?&lt;/p&gt;

&lt;p&gt;Nowadays, companies, universities, researchers, and individual developers have started to rent a GPU or entire GPU servers instead of buying or local deployment. &lt;/p&gt;

&lt;p&gt;The truth is, without the long-term commitment or expenses needed, renting cloud GPUs is the most flexible and affordable solution when dealing with heavy processing power.&lt;/p&gt;

&lt;h2&gt;
  
  
  Benefits include:
&lt;/h2&gt;

&lt;p&gt;&lt;strong&gt;Cost-effectiveness&lt;/strong&gt;&lt;br&gt;
Renting GPUs provides flexible on demand model and do not require one-time huge hardware investment such as high-performance GPUs, supporting servers, and cooling equipment, etc.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;Resource flexibility&lt;/strong&gt;&lt;br&gt;
GPU rental platform usually supports a variety of GPU models and users can adjust resource configuration any time. No limitations by hardware specifications.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;Maintenance and technical support&lt;/strong&gt;&lt;br&gt;
24/7 technical support, rich model images and one-click deployment are supported in GPU rental platform to ensure service quality and ease of use which allows users to quickly get started with their applications.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;Data security and privacy&lt;/strong&gt;&lt;br&gt;
GPU rental platform can ensure user data security through professional security measures and compliance management. &lt;/p&gt;

&lt;p&gt;&lt;strong&gt;Scalability and Ecosystem&lt;/strong&gt;&lt;br&gt;
Select from either container or virtual machine modes, together with rich platform image resources, GPU rental platform can easily expand and customize workflows. No need to build and maintain a complex environment by yourself anymore.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;About RunC.AI&lt;/strong&gt;&lt;br&gt;
Rent smart, run fast. RunC.AI  | Run clever cloud computing for AI allows users to gain access to a wide selection of scalable, high-performance GPU instances and clusters at competitive prices compared to major cloud providers like Amazon Web Services (AWS), Google Cloud, and Microsoft Azure. &lt;/p&gt;

&lt;p&gt;Register now and get $5 free credits for your applications!&lt;br&gt;
The free credits will be recharged into your account automatically within few days.&lt;br&gt;
&lt;strong&gt;(Due 6th, June 2025)&lt;/strong&gt;&lt;br&gt;
&lt;em&gt;Start your journey here&lt;/em&gt;:&lt;a href="https://www.runc.ai?ytag=rc_dev_0528" rel="noopener noreferrer"&gt;RunC.AI Official&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fwfxc3nhx7tpin258ztb9.png" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fwfxc3nhx7tpin258ztb9.png" alt="Image description" width="800" height="336"&gt;&lt;/a&gt;&lt;/p&gt;

</description>
      <category>ai</category>
      <category>machinelearning</category>
      <category>gpu</category>
      <category>webdev</category>
    </item>
  </channel>
</rss>
