<?xml version="1.0" encoding="UTF-8"?>
<rss version="2.0" xmlns:atom="http://www.w3.org/2005/Atom" xmlns:dc="http://purl.org/dc/elements/1.1/">
  <channel>
    <title>DEV Community: Benedict Isaac</title>
    <description>The latest articles on DEV Community by Benedict Isaac (@benedict258).</description>
    <link>https://dev.to/benedict258</link>
    <image>
      <url>https://media2.dev.to/dynamic/image/width=90,height=90,fit=cover,gravity=auto,format=auto/https:%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Fuser%2Fprofile_image%2F1701220%2F06047419-aad9-45ce-b08b-a0eb80c6f2a9.png</url>
      <title>DEV Community: Benedict Isaac</title>
      <link>https://dev.to/benedict258</link>
    </image>
    <atom:link rel="self" type="application/rss+xml" href="https://dev.to/feed/benedict258"/>
    <language>en</language>
    <item>
      <title>Opik by Comet The Open Source Observability Tool Every AI Builder Needs in Their Stack</title>
      <dc:creator>Benedict Isaac</dc:creator>
      <pubDate>Sat, 11 Apr 2026 17:14:42 +0000</pubDate>
      <link>https://dev.to/benedict258/opik-by-comet-the-open-source-observability-tool-every-ai-builder-needs-in-their-stack-1840</link>
      <guid>https://dev.to/benedict258/opik-by-comet-the-open-source-observability-tool-every-ai-builder-needs-in-their-stack-1840</guid>
      <description>&lt;p&gt;I came across Opik during the Commit to Change Hackathon by Encode Club, in partnership with Comet. I had never heard of it before, but after integrating it into my project, it became one of those tools I couldn't imagine building without.&lt;/p&gt;

&lt;p&gt;If you're building LLM-powered applications whether that's a RAG pipeline, an AI agent, a chatbot, or any system that calls a &lt;/p&gt;

&lt;p&gt;language model you already know the pain:&lt;br&gt;
Something breaks and you don't know where&lt;br&gt;
Your agent hallucinates and you can't trace why&lt;br&gt;
Token costs spike and you have no visibility into what's consuming them&lt;br&gt;
You change a prompt and don't know if it actually improved anything&lt;/p&gt;

&lt;p&gt;Opik is the answer to all of that.&lt;br&gt;
This article covers what Opik is, how it works, its core features, and how to integrate it into your existing LLM stack — with real code examples.&lt;/p&gt;

&lt;p&gt;What is Opik?&lt;br&gt;
Opik is an open-source LLM observability and evaluation platform built by Comet. It sits alongside your AI application and gives you complete visibility into everything your system does every LLM call, every tool invocation, every chain step logged, scored, and visualized in one dashboard.&lt;/p&gt;

&lt;p&gt;It covers the full development lifecycle:&lt;br&gt;
Development trace and debug your agents as you build&lt;br&gt;
Evaluation score outputs and run experiments across prompt versions&lt;br&gt;
Production monitor live traffic, detect issues, and auto-optimize&lt;/p&gt;

&lt;p&gt;Think of it like this: if your AI agent is a car, Opik is the full onboard diagnostics system not just a dashboard light that tells you something is wrong, but the full readout that tells you exactly which component failed and why.&lt;/p&gt;

&lt;p&gt;Core Concepts: Traces and Spans&lt;br&gt;
Before diving into code, it helps to understand two key concepts Opik is built around.&lt;/p&gt;

&lt;p&gt;A trace is a complete record of one end-to-end request through your LLM application. From the moment a user sends a question to the moment your app returns a response that entire journey is one trace.&lt;/p&gt;

&lt;p&gt;A span is a single step inside that trace. If your agent calls a retrieval function, then calls the LLM, then formats the output — each of those is a span nested inside the parent trace.&lt;br&gt;
This structure gives you surgical visibility. Instead of just knowing "the response was bad," you can see exactly which step produced the bad output, how long it took, and what it was working with.&lt;/p&gt;

&lt;p&gt;Getting Started Installation&lt;br&gt;
bash pip install opik&lt;br&gt;
opik configure&lt;br&gt;
Running opik configure sets up your API key and connects your environment to the Opik cloud dashboard. You can also self-host Opik if you prefer to keep everything local.&lt;/p&gt;

&lt;p&gt;Core Integration The @track Decorator&lt;br&gt;
The fastest way to get started with Opik is the @track decorator. Add it to any function in your LLM pipeline and Opik automatically logs it as a span.&lt;/p&gt;

&lt;p&gt;python from opik import track&lt;/p&gt;

&lt;p&gt;`@track&lt;br&gt;
def llm_chain(user_question):&lt;br&gt;
    context = get_context(user_question)&lt;br&gt;
    response = call_llm(user_question, context)&lt;br&gt;
    return response&lt;/p&gt;

&lt;p&gt;@track&lt;br&gt;
def get_context(user_question):&lt;br&gt;
    # Retrieval logic hard coded here for simplicity&lt;br&gt;
    return ["The dog chased the cat.", "The cat was called Luky."]&lt;/p&gt;

&lt;p&gt;@track&lt;br&gt;
def call_llm(user_question, context):&lt;br&gt;
    # Your actual LLM call goes here&lt;br&gt;
    return "The dog chased the cat Luky."&lt;/p&gt;

&lt;p&gt;response = llm_chain("What did the dog do?")&lt;br&gt;
print(response)`&lt;/p&gt;

&lt;p&gt;What happens when you run this:&lt;br&gt;
llm_chain is logged as the parent trace&lt;br&gt;
get_context and call_llm are logged as child spans nested inside it&lt;br&gt;
Every input, output, and execution time is captured automatically&lt;br&gt;
The full chain appears in your Opik dashboard instantly&lt;/p&gt;

&lt;p&gt;No boilerplate. No manual logging. Just a decorator.&lt;/p&gt;

&lt;p&gt;Integrations Works With Your Existing Stack&lt;br&gt;
Opik isn't asking you to rewrite your application. It integrates directly with the tools you're already using:&lt;br&gt;
LangChain:&lt;/p&gt;

&lt;p&gt;`pythonfrom langchain_openai import ChatOpenAI&lt;br&gt;
from opik.integrations.langchain import OpikTracer&lt;/p&gt;

&lt;p&gt;opik_tracer = OpikTracer()&lt;br&gt;
llm = ChatOpenAI(temperature=0)&lt;br&gt;
llm = llm.with_config({"callbacks": [opik_tracer]})&lt;/p&gt;

&lt;p&gt;llm.invoke("Hello, how are you?")&lt;br&gt;
OpenAI SDK:&lt;br&gt;
pythonfrom openai import OpenAI&lt;br&gt;
from opik.integrations.openai import track_openai&lt;/p&gt;

&lt;p&gt;openai_client = OpenAI()&lt;br&gt;
openai_client = track_openai(openai_client)&lt;/p&gt;

&lt;p&gt;response = openai_client.chat.completions.create(&lt;br&gt;
    model="gpt-4",&lt;br&gt;
    messages=[{"role": "user", "content": "Hello, world!"}]&lt;br&gt;
)&lt;br&gt;
LlamaIndex:&lt;br&gt;
pythonfrom llama_index.core import set_global_handler&lt;/p&gt;

&lt;p&gt;set_global_handler("opik")`&lt;/p&gt;

&lt;p&gt;One line. That's it for LlamaIndex.&lt;br&gt;
Opik also supports LiteLLM, DSPy, Ragas, OpenTelemetry, and Predibase so whatever your stack looks like, it fits in.&lt;/p&gt;

&lt;p&gt;Evaluation Stop Guessing, Start Scoring&lt;br&gt;
Once your traces are being logged, you can start running evaluations. Opik has built-in eval metrics including:&lt;/p&gt;

&lt;p&gt;Hallucination detection flags responses that contradict the provided context&lt;/p&gt;

&lt;p&gt;Answer relevance scores how well the response addresses the question&lt;/p&gt;

&lt;p&gt;Context precision measures the quality of retrieved context in RAG systems&lt;/p&gt;

&lt;p&gt;Factuality checks responses against a ground truth dataset&lt;br&gt;
Moderation flags harmful or policy-violating content&lt;/p&gt;

&lt;p&gt;You can also define your own custom metrics using the SDK.&lt;br&gt;
The real power here is running experiments give Opik a dataset, define what "good" looks like using your chosen metrics, and let it automatically score different versions of your app against each other. You stop debating which prompt is better and start measuring it.&lt;/p&gt;

&lt;p&gt;Guardrails Safety Built In&lt;br&gt;
Opik ships with built-in guardrails that screen both user inputs and LLM outputs before they cause problems:&lt;/p&gt;

&lt;p&gt;PII detection and redaction&lt;br&gt;
Competitor mention filtering&lt;br&gt;
Off-topic content detection&lt;br&gt;
Custom content moderation rules&lt;/p&gt;

&lt;p&gt;You can use Opik's built-in models or plug in your own third-party guardrail libraries. This means safety isn't an afterthought you bolt on at the end — it's baked into the same observability pipeline you're already running.&lt;/p&gt;

&lt;p&gt;Automatic Prompt Optimization&lt;br&gt;
This is one of the most powerful features Opik offers and one that most developers don't expect from an observability tool.&lt;br&gt;
Once you've defined your evaluation metrics and built a test dataset, Opik can automatically generate and test improved versions of your prompts using four built-in optimizers:&lt;/p&gt;

&lt;p&gt;Few-shot Bayesian finds the best few-shot examples for your use case&lt;br&gt;
MIPRO multi-stage instruction and prefix optimization&lt;br&gt;
Evolutionary optimizer iteratively evolves prompt variations&lt;br&gt;
MetaPrompt (LLM-powered) uses an LLM to rewrite and improve your prompts&lt;/p&gt;

&lt;p&gt;The result is a production-ready, frozen prompt that you can lock in and deploy with confidence without manually iterating through dozens of variations yourself.&lt;/p&gt;

&lt;p&gt;Production Monitoring&lt;br&gt;
When you ship to production, Opik keeps running. Every live request is logged, scored using online eval metrics, and surfaced in your monitoring dashboard.&lt;/p&gt;

&lt;p&gt;This means:&lt;br&gt;
You catch regressions immediately when a new model version behaves differently&lt;br&gt;
You build new test datasets directly from real production traffic&lt;br&gt;
You close the loop between what you tested in development and what actually happens with real users&lt;/p&gt;

&lt;p&gt;Why This Matters Right Now&lt;br&gt;
The conversation in AI development has shifted. A year ago, the focus was almost entirely on prompts write a better prompt, get a better output. That still matters, but it's not enough anymore.&lt;/p&gt;

&lt;p&gt;As AI systems get more complex multi-agent workflows, RAG pipelines, tool-calling chains the failure modes multiply. You can't eyeball your way through 10,000 production traces. You need instrumentation.&lt;/p&gt;

&lt;p&gt;Opik gives you that instrumentation. And because it's open source with 18k+ GitHub stars, it's backed by a real community — not a vendor lock-in waiting to happen.&lt;/p&gt;

&lt;p&gt;Getting Started&lt;/p&gt;

&lt;p&gt;Install: pip install opik&lt;br&gt;
Configure: opik configure&lt;br&gt;
Add @track to your LLM functions&lt;br&gt;
Open your Opik dashboard and watch your traces appear&lt;/p&gt;

&lt;p&gt;Free to start, no credit card required.&lt;br&gt;
🔗 comet.com/site/products/opik&lt;br&gt;
⭐ github.com/comet-ml/opik&lt;/p&gt;

</description>
      <category>ai</category>
      <category>devtools</category>
    </item>
    <item>
      <title>Be The Best Of Who You Are</title>
      <dc:creator>Benedict Isaac</dc:creator>
      <pubDate>Mon, 23 Mar 2026 14:21:32 +0000</pubDate>
      <link>https://dev.to/benedict258/be-the-best-of-who-you-are-4j7i</link>
      <guid>https://dev.to/benedict258/be-the-best-of-who-you-are-4j7i</guid>
      <description>&lt;p&gt;&lt;strong&gt;Be The Best Bush 🌿&lt;/strong&gt;&lt;/p&gt;

&lt;p&gt;If all you can be is a bush… be the best bush.&lt;/p&gt;

&lt;p&gt;What does this really mean?&lt;/p&gt;

&lt;p&gt;There are moments where everything feels like the right path.&lt;br&gt;
You want to pursue multiple things. Build, learn, explore, try everything.&lt;/p&gt;

&lt;p&gt;And sometimes, your vision feels far bigger than your current reality.&lt;/p&gt;

&lt;p&gt;But what if, right now, there’s something within reach?&lt;/p&gt;

&lt;p&gt;Something small.&lt;br&gt;
Something simple.&lt;br&gt;
Something possible.&lt;/p&gt;

&lt;p&gt;If all you can do at this stage is learn, then do it to the best of your ability.&lt;br&gt;
Be exceptional at it. Become known for it.&lt;/p&gt;

&lt;p&gt;Because over time, other paths begin to align.&lt;/p&gt;

&lt;p&gt;Excuses won’t move you forward&lt;br&gt;
execution will.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;Start Where You Are&lt;/strong&gt;&lt;/p&gt;

&lt;p&gt;If all you can do is…&lt;/p&gt;

&lt;p&gt;Learn → be the best learner&lt;br&gt;
Build → be the best builder&lt;br&gt;
Start small → dominate small&lt;br&gt;
Practice → practice with intent&lt;br&gt;
Show up → show up consistently&lt;br&gt;
Think → think deeply and clearly&lt;/p&gt;

&lt;p&gt;Be the best… at that level.&lt;/p&gt;

&lt;p&gt;Not someday.&lt;br&gt;
Now.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;The Illusion of Doing Everything&lt;/strong&gt;&lt;/p&gt;

&lt;p&gt;I once heard a quote:&lt;/p&gt;

&lt;p&gt;“Be the best of who you are.”&lt;/p&gt;

&lt;p&gt;It sounds simple, but it carries weight.&lt;br&gt;
Because the truth is trying to do everything at once isn’t always effective.&lt;/p&gt;

&lt;p&gt;It’s not about whether it’s possible.&lt;br&gt;
It’s about efficiency and impact.&lt;/p&gt;

&lt;p&gt;Spreading yourself too thin may feel productive,&lt;br&gt;
but often, it slows down real growth.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;The Power of Focus and Synergy&lt;/strong&gt;&lt;/p&gt;

&lt;p&gt;A better approach is to find the synergy between your interests and your end goal.&lt;/p&gt;

&lt;p&gt;Not abandoning your interests.&lt;br&gt;
But sequencing them.&lt;/p&gt;

&lt;p&gt;Master one thing first.&lt;br&gt;
Then grow into the next.&lt;br&gt;
Then begin to integrate those skills together.&lt;/p&gt;

&lt;p&gt;That’s how depth is built.&lt;br&gt;
That’s how people stand out.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;Across Every Field&lt;/strong&gt;&lt;br&gt;
This applies everywhere.&lt;/p&gt;

&lt;p&gt;Whether you're a:&lt;/p&gt;

&lt;p&gt;developer&lt;br&gt;
engineer&lt;br&gt;
designer&lt;br&gt;
writer&lt;br&gt;
or builder of any kind&lt;/p&gt;

&lt;p&gt;The principle remains the same.&lt;br&gt;
Always aim for the top even if you're starting small.&lt;br&gt;
Growth Over Averages&lt;/p&gt;

&lt;p&gt;If you’re the best today,&lt;br&gt;
be better than you were yesterday.&lt;/p&gt;

&lt;p&gt;Progress compounds.&lt;/p&gt;

&lt;p&gt;And over time, what once felt small&lt;br&gt;
becomes your foundation.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;&lt;em&gt;You’re not meant to be average&lt;br&gt;
you’re meant to grow.&lt;/em&gt;&lt;/strong&gt;&lt;/p&gt;

</description>
      <category>growth</category>
      <category>skills</category>
      <category>consistency</category>
    </item>
    <item>
      <title>The Systems Behind AI: Understanding LLMs and Their Place in Engineering</title>
      <dc:creator>Benedict Isaac</dc:creator>
      <pubDate>Tue, 17 Mar 2026 17:20:23 +0000</pubDate>
      <link>https://dev.to/benedict258/the-systems-behind-ai-understanding-llms-and-their-place-in-engineering-2p2f</link>
      <guid>https://dev.to/benedict258/the-systems-behind-ai-understanding-llms-and-their-place-in-engineering-2p2f</guid>
      <description>&lt;p&gt;You’ve used AI.&lt;br&gt;
You’ve explored it.&lt;br&gt;
You’ve probably even built with it.&lt;/p&gt;

&lt;p&gt;But behind many of the AI systems we interact with today is something more fundamental.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;The Large Language Model (LLM).&lt;/strong&gt;&lt;/p&gt;

&lt;p&gt;LLMs power a large portion of modern AI systems and AI agents. They act as the core intelligence layer behind many tools and applications.&lt;/p&gt;

&lt;p&gt;So before building more systems with AI, an important question becomes:&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;What exactly are LLMs?&lt;/strong&gt;&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;&lt;em&gt;What Are LLMs&lt;/em&gt;&lt;/strong&gt;&lt;br&gt;
A Large Language Model (LLM) is a computational model trained on vast amounts of data and designed for natural language processing, especially language understanding and language generation.&lt;/p&gt;

&lt;p&gt;These models learn patterns in language and use those patterns to generate responses, reason through prompts, and assist in solving problems.&lt;/p&gt;

&lt;p&gt;At a high level, LLM systems typically involve:&lt;/p&gt;

&lt;p&gt;Encoders: which process and understand input information&lt;br&gt;
Decoders : which generate responses or outputs&lt;br&gt;
Modern LLMs combine these mechanisms to process context, relationships between words, and the structure of language.&lt;/p&gt;

&lt;p&gt;&lt;em&gt;But the real breakthrough that enabled today’s powerful models came from a particular architecture&lt;/em&gt;.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;What Is a Transformer&lt;/strong&gt;&lt;br&gt;
Most modern LLMs are built using an architecture called the Transformer.&lt;/p&gt;

&lt;p&gt;The Transformer architecture changed how models process sequences of information by introducing a mechanism called attention.&lt;/p&gt;

&lt;p&gt;Instead of processing words one at a time in order like older neural networks, transformers can look at entire sequences of data simultaneously and determine which parts are most important.&lt;/p&gt;

&lt;p&gt;This allows models to:&lt;/p&gt;

&lt;p&gt;understand long-range relationships in text&lt;br&gt;
maintain context across large inputs&lt;br&gt;
generate more coherent and structured responses&lt;br&gt;
Because of this design, transformers became the foundation for modern LLMs.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;The Three Core Components of an LLM&lt;/strong&gt;&lt;br&gt;
At a simplified level, most LLM systems can be understood as a combination of three elements:&lt;/p&gt;

&lt;p&gt;LLM = Data + Architecture + Training&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;Data&lt;/strong&gt;&lt;br&gt;
Large-scale datasets containing books, articles, code, conversations, and other forms of language.&lt;/p&gt;

&lt;p&gt;This data teaches the model patterns of human communication and knowledge representation.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;Architecture&lt;/strong&gt;&lt;br&gt;
The structural design of the model itself.&lt;/p&gt;

&lt;p&gt;Today, this is primarily the Transformer architecture, which enables models to process complex relationships within data.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;Training&lt;/strong&gt;&lt;br&gt;
The optimization process where the model learns to predict and generate language.&lt;/p&gt;

&lt;p&gt;During training, the system adjusts millions or even billions of parameters to improve its predictions and outputs.&lt;/p&gt;

&lt;p&gt;Together, these three components create the systems that power modern AI tools.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;Engineering, AI, and the Missing Bridge&lt;/strong&gt;&lt;br&gt;
Coding, machine learning, and AI are incredibly valuable skills. But they were originally meant to complement engineering fundamentals, not replace them.&lt;/p&gt;

&lt;p&gt;This observation raises an interesting thought.&lt;/p&gt;

&lt;p&gt;&lt;em&gt;What happens when we bring AI thinking back into engineering systems rather than keeping it limited to software?&lt;/em&gt;&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;Viewing AI from an Engineering Perspective&lt;/strong&gt;&lt;br&gt;
In engineering, almost every new technology eventually finds an entry point into physical systems.&lt;/p&gt;

&lt;p&gt;So instead of thinking only about AI models running inside applications, another question emerges:&lt;/p&gt;

&lt;p&gt;How can engineering systems and robotics systems behave or learn in ways similar to LLM-powered systems?&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;What would that actually look like?&lt;/strong&gt;&lt;/p&gt;

&lt;p&gt;Instead of only building software agents, we might begin designing systems where learning, decision-making, and adaptive behavior exist within engineered machines.&lt;/p&gt;

&lt;p&gt;This idea goes beyond simple automation.&lt;/p&gt;

&lt;p&gt;It begins to resemble something closer to an engineering model of intelligence.&lt;/p&gt;

&lt;p&gt;From AI Agents to Intelligent Engineering Systems&lt;br&gt;
Today, we often talk about AI agents in the context of software.&lt;/p&gt;

&lt;p&gt;Agents can reason about tasks, interact with systems, retrieve information, and perform actions.&lt;/p&gt;

&lt;p&gt;But if we extend this thinking into engineering systems, an interesting possibility appears.&lt;/p&gt;

&lt;p&gt;Robotics systems and machines could potentially be designed to:&lt;/p&gt;

&lt;p&gt;behave similarly to AI agents&lt;br&gt;
integrate multiple data sources and signals&lt;br&gt;
adapt to environments through learning&lt;br&gt;
reason about tasks and actions&lt;br&gt;
In this perspective, AI is no longer just a software capability.&lt;/p&gt;

&lt;p&gt;It becomes something that can interact with and shape engineering systems directly.&lt;/p&gt;

&lt;p&gt;This is where the worlds of AI, robotics, and engineering start to converge.&lt;/p&gt;

&lt;p&gt;A Practical Question Builders Must Answer&lt;br&gt;
As builders begin integrating AI systems into applications and infrastructure, another important question arises.&lt;/p&gt;

&lt;p&gt;&lt;em&gt;Should you build your own model, fine-tune an existing model, or use a pre-existing LLM?&lt;/em&gt;&lt;/p&gt;

&lt;p&gt;Each approach has its own advantages and trade-offs.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;Building Your Own Model&lt;/strong&gt;&lt;br&gt;
&lt;em&gt;Provides maximum control and customization.&lt;/em&gt;&lt;/p&gt;

&lt;p&gt;However, it requires extremely large datasets, significant computing resources, and specialized expertise.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;Fine-Tuning an Existing Model&lt;/strong&gt;&lt;br&gt;
&lt;em&gt;Allows developers to adapt a pre-trained model to a specific domain or problem.&lt;/em&gt;&lt;/p&gt;

&lt;p&gt;This often provides a strong balance between performance and cost.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;Using a Pre-Existing LLM&lt;/strong&gt;&lt;br&gt;
&lt;em&gt;The fastest and most accessible option.&lt;/em&gt;&lt;/p&gt;

&lt;p&gt;Most modern AI applications today rely on existing models and focus on building systems around them rather than training models from scratch.&lt;/p&gt;

&lt;p&gt;The best choice depends on the complexity of the system, the resources available, and the specific problem being solved.&lt;/p&gt;

&lt;p&gt;&lt;em&gt;AI systems today are evolving rapidly.&lt;/em&gt;&lt;/p&gt;

&lt;p&gt;But the most interesting opportunities may not only be in building larger models, but also in exploring how these systems integrate with real engineering systems and infrastructure.&lt;/p&gt;

&lt;p&gt;Instead of thinking about AI purely as software, we can begin to ask:&lt;/p&gt;

&lt;p&gt;&lt;em&gt;How do intelligent models interact with machines, robotics systems, and the engineered world?&lt;/em&gt;&lt;/p&gt;

&lt;p&gt;The future of AI may not only live in data centers and applications.&lt;/p&gt;

&lt;p&gt;It may also live inside the systems we build to interact with the physical world.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;Question for builders and engineers:&lt;/strong&gt;&lt;/p&gt;

&lt;p&gt;&lt;em&gt;Is it better to build your own model for a system, fine-tune an existing model, or rely on pre-trained LLMs?&lt;/em&gt;&lt;/p&gt;

</description>
      <category>ai</category>
      <category>llm</category>
      <category>data</category>
      <category>software</category>
    </item>
  </channel>
</rss>
