DEV Community

Byron Salty
Byron Salty

Posted on

A little Rust proxy for Ollama

Introduction

Recently, while exploring the new Zed IDE, I encountered a need to integrate it with a self-hosted Ollama model. Zed allows for custom LLM integrations, which is great, but I ran into a small hiccup.

Problem Description

The challenge arose because Ollama's configuration in Zed requires a connection to localhost. However, I have Ollama running on my GPU-enabled home server, not on my local machine.

Solution Approach

To bridge this gap, I devised a simple solution: a Rust script that acts as a proxy. This script forwards the Ollama port from my server to the local machine, allowing Zed to connect seamlessly.

Implementation Details

Here's a brief walkthrough of the Rust script:

use hyper::service::{make_service_fn, service_fn};
use hyper::{Body, Client, Request, Response, Server};
use std::convert::Infallible;
use std::net::SocketAddr;

async fn proxy(req: Request<Body>) -> Result<Response<Body>, hyper::Error> {
    // Create a client to send the forwarded request.
    let client = Client::new();

    // Build a new URI using the hostname "ace" and the appropriate port.
    // Here, we assume your Ollama server is also listening on port 11434.
    // Adjust the port in the format string if it's different.
    let uri_string = format!("http://ace:11434{}", req.uri());
    let new_uri = uri_string.parse().expect("Failed to parse URI");

    // Replace the URI in the incoming request with the new one.
    // This effectively "redirects" the request to your remote server.
    let mut proxied_request = req;
    *proxied_request.uri_mut() = new_uri;

    // Forward the request to the remote server.
    client.request(proxied_request).await
}

#[tokio::main]
async fn main() -> Result<(), Box<dyn std::error::Error + Send + Sync>> {
    // Bind to local port 11434.
    let addr = SocketAddr::from(([127, 0, 0, 1], 11434));
    let make_svc = make_service_fn(|_conn| async {
        // For each connection, we create a service to handle the request.
        Ok::<_, Infallible>(service_fn(proxy))
    });

    let server = Server::bind(&addr).serve(make_svc);

    println!("Reverse proxy running on http://{}", addr);

    // Run the server until it's stopped.
    server.await?;
    Ok(())
}
Enter fullscreen mode Exit fullscreen mode

The script efficiently handles the port forwarding with no noticeable performance penalty, making the integration smooth and reliable.

Conclusion

This quick Rust script solved the integration challenge, enabling me to use Zed with my self-hosted Ollama model effortlessly. It's a testament to how small, targeted scripts can effectively solve specific problems without overcomplicating the setup.

To me this is a very interesting implication of the new world of AI tooling. I just used some simple prompts and had a working solution to a problem in a few minutes. I'm sure I could have taken the time to make caddy or nginx or something else work, but custom built software for a specific need is also an option nowadays.

Image of Timescale

PostgreSQL for Agentic AI — Build Autonomous Apps on One Stack ☝️

pgai turns PostgreSQL into an AI-native database for building RAG pipelines and intelligent agents. Run vector search, embeddings, and LLMs—all in SQL

Build Today

Top comments (0)

👋 Kindness is contagious

Engage with a wealth of insights in this thoughtful article, valued within the supportive DEV Community. Coders of every background are welcome to join in and add to our collective wisdom.

A sincere "thank you" often brightens someone’s day. Share your gratitude in the comments below!

On DEV, the act of sharing knowledge eases our journey and fortifies our community ties. Found value in this? A quick thank you to the author can make a significant impact.

Okay