<?xml version="1.0" encoding="UTF-8"?>
<rss version="2.0" xmlns:atom="http://www.w3.org/2005/Atom" xmlns:dc="http://purl.org/dc/elements/1.1/">
  <channel>
    <title>DEV Community: Ibbad Ur Rehman</title>
    <description>The latest articles on DEV Community by Ibbad Ur Rehman (@ibbad_e0c).</description>
    <link>https://dev.to/ibbad_e0c</link>
    <image>
      <url>https://media2.dev.to/dynamic/image/width=90,height=90,fit=cover,gravity=auto,format=auto/https:%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Fuser%2Fprofile_image%2F3813261%2F18f05da5-a4af-4210-839f-7f4d19613c78.png</url>
      <title>DEV Community: Ibbad Ur Rehman</title>
      <link>https://dev.to/ibbad_e0c</link>
    </image>
    <atom:link rel="self" type="application/rss+xml" href="https://dev.to/feed/ibbad_e0c"/>
    <language>en</language>
    <item>
      <title>From Search to Agency — Two AI Papers That Changed How I Think</title>
      <dc:creator>Ibbad Ur Rehman</dc:creator>
      <pubDate>Sat, 14 Mar 2026 16:28:33 +0000</pubDate>
      <link>https://dev.to/ibbad_e0c/from-search-to-agency-two-ai-papers-that-changed-how-i-think-5hl9</link>
      <guid>https://dev.to/ibbad_e0c/from-search-to-agency-two-ai-papers-that-changed-how-i-think-5hl9</guid>
      <description>&lt;h2&gt;
  
  
  Introduction
&lt;/h2&gt;

&lt;p&gt;When my AI course professor assigned us to read and blog about recent research&lt;br&gt;
papers, I expected to spend a weekend forcing myself through dense academic&lt;br&gt;
writing. What actually happened shifted how I think about two topics I thought&lt;br&gt;
I already understood: &lt;strong&gt;search algorithms&lt;/strong&gt; and &lt;strong&gt;intelligent agents&lt;/strong&gt;.&lt;/p&gt;

&lt;p&gt;The two papers I chose:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;
&lt;strong&gt;"A Survey of LLM-based Deep Search Agents"&lt;/strong&gt; (2026)&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;"Research on the A* Algorithm Based on Adaptive Weights"&lt;/strong&gt; (2025)&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;are not obviously related at first glance. One is about large language models&lt;br&gt;
navigating information, the other is about a classic pathfinding algorithm.&lt;br&gt;
But together, they tell a unified story about what &lt;strong&gt;search really means&lt;br&gt;
in modern AI&lt;/strong&gt;.&lt;/p&gt;


&lt;h2&gt;
  
  
  Paper 1: A Survey of LLM-based Deep Search Agents (2026)
&lt;/h2&gt;
&lt;h3&gt;
  
  
  What is the paper trying to do?
&lt;/h3&gt;

&lt;p&gt;This survey maps out an entirely new category of AI system: the &lt;strong&gt;Deep Search&lt;br&gt;
Agent&lt;/strong&gt;. The authors document how we evolved from:&lt;/p&gt;

&lt;blockquote&gt;
&lt;p&gt;Traditional keyword search → LLM-enhanced search → Agentic search&lt;/p&gt;
&lt;/blockquote&gt;

&lt;p&gt;In agentic search, an LLM autonomously plans, retrieves, reflects, and&lt;br&gt;
iterates until it can synthesize a comprehensive answer to a complex question.&lt;/p&gt;

&lt;p&gt;The core architectural loop the paper formalizes is:&lt;br&gt;
&lt;/p&gt;

&lt;div class="highlight js-code-highlight"&gt;
&lt;pre class="highlight plaintext"&gt;&lt;code&gt;Plan → Act → Observe → Reflect → Generate
&lt;/code&gt;&lt;/pre&gt;

&lt;/div&gt;



&lt;p&gt;Crucially, the &lt;strong&gt;Reflect&lt;/strong&gt; stage decides whether to loop back and search&lt;br&gt;
again or to finalize the answer. Real-world systems like &lt;strong&gt;OpenAI's Deep&lt;br&gt;
Research&lt;/strong&gt; and &lt;strong&gt;Perplexity's Pro Search&lt;/strong&gt; are concrete implementations&lt;br&gt;
of this loop.&lt;/p&gt;


&lt;h3&gt;
  
  
  The Technical Depth
&lt;/h3&gt;

&lt;p&gt;These agents organize their retrieval using three distinct search strategies:&lt;/p&gt;

&lt;div class="table-wrapper-paragraph"&gt;&lt;table&gt;
&lt;thead&gt;
&lt;tr&gt;
&lt;th&gt;Strategy&lt;/th&gt;
&lt;th&gt;How it works&lt;/th&gt;
&lt;/tr&gt;
&lt;/thead&gt;
&lt;tbody&gt;
&lt;tr&gt;
&lt;td&gt;&lt;strong&gt;Sequential&lt;/strong&gt;&lt;/td&gt;
&lt;td&gt;Works through one reasoning thread at a time, step by step&lt;/td&gt;
&lt;/tr&gt;
&lt;tr&gt;
&lt;td&gt;&lt;strong&gt;Parallel&lt;/strong&gt;&lt;/td&gt;
&lt;td&gt;Decomposes the question into sub-questions, explores all branches simultaneously&lt;/td&gt;
&lt;/tr&gt;
&lt;tr&gt;
&lt;td&gt;&lt;strong&gt;Hybrid&lt;/strong&gt;&lt;/td&gt;
&lt;td&gt;Combines both, sometimes using Monte Carlo Tree Search (MCTS)&lt;/td&gt;
&lt;/tr&gt;
&lt;/tbody&gt;
&lt;/table&gt;&lt;/div&gt;

&lt;p&gt;The task formulation is a &lt;strong&gt;state-space problem&lt;/strong&gt;: given a user query &lt;em&gt;q&lt;/em&gt;,&lt;br&gt;
the agent operates over a trajectory of (observation, action) pairs and must&lt;br&gt;
reach a terminal state where the accumulated evidence satisfies the&lt;br&gt;
information need.&lt;/p&gt;


&lt;h3&gt;
  
  
  Connection to Our Course
&lt;/h3&gt;

&lt;p&gt;In our AI course, we study &lt;strong&gt;rational agents&lt;/strong&gt; using the PEAS framework&lt;br&gt;
(Performance, Environment, Actuators, Sensors). A Deep Search Agent is&lt;br&gt;
structurally a &lt;strong&gt;goal-based agent&lt;/strong&gt; where:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;
&lt;strong&gt;Goal&lt;/strong&gt; — answering the query completely&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;Actuator&lt;/strong&gt; — ability to query search engines or browse URLs&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;Sensor&lt;/strong&gt; — the retrieved text it reads&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;Environment&lt;/strong&gt; — the open web (partially observable, dynamic)&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;The search strategies map directly to what we study in class:&lt;/p&gt;

&lt;blockquote&gt;
&lt;p&gt;Sequential search = Depth-First Search (DFS)&lt;/p&gt;

&lt;p&gt;Parallel search = Breadth-First Search (BFS)&lt;/p&gt;

&lt;p&gt;The goal test = the moment the agent decides it has enough information&lt;/p&gt;
&lt;/blockquote&gt;


&lt;h2&gt;
  
  
  Paper 2: A* Algorithm Based on Adaptive Weights (2025)
&lt;/h2&gt;
&lt;h3&gt;
  
  
  What is the paper trying to do?
&lt;/h3&gt;

&lt;p&gt;The standard A* algorithm has a well-known limitation: it expands a large&lt;br&gt;
number of nodes trying to be both &lt;strong&gt;fast&lt;/strong&gt; and &lt;strong&gt;optimal&lt;/strong&gt; simultaneously.&lt;/p&gt;

&lt;p&gt;This paper proposes a solution — a &lt;strong&gt;dynamic weighting function&lt;/strong&gt; based on&lt;br&gt;
a radial basis function — to make A* smarter about when to rush and when&lt;br&gt;
to be careful.&lt;/p&gt;

&lt;p&gt;The core formula becomes:&lt;br&gt;
&lt;/p&gt;

&lt;div class="highlight js-code-highlight"&gt;
&lt;pre class="highlight plaintext"&gt;&lt;code&gt;f(n) = g(n) + f(x,y) × h(n)
&lt;/code&gt;&lt;/pre&gt;

&lt;/div&gt;



&lt;p&gt;Where &lt;strong&gt;f(x,y)&lt;/strong&gt; is not a constant but a dynamic weight that changes based&lt;br&gt;
on the agent's position:&lt;/p&gt;

&lt;div class="table-wrapper-paragraph"&gt;&lt;table&gt;
&lt;thead&gt;
&lt;tr&gt;
&lt;th&gt;Position&lt;/th&gt;
&lt;th&gt;Weight&lt;/th&gt;
&lt;th&gt;Behaviour&lt;/th&gt;
&lt;/tr&gt;
&lt;/thead&gt;
&lt;tbody&gt;
&lt;tr&gt;
&lt;td&gt;&lt;strong&gt;Near start&lt;/strong&gt;&lt;/td&gt;
&lt;td&gt;High&lt;/td&gt;
&lt;td&gt;Heuristic amplified, algorithm charges quickly toward goal&lt;/td&gt;
&lt;/tr&gt;
&lt;tr&gt;
&lt;td&gt;&lt;strong&gt;Near goal&lt;/strong&gt;&lt;/td&gt;
&lt;td&gt;Low&lt;/td&gt;
&lt;td&gt;Algorithm slows down, searches carefully for optimal path&lt;/td&gt;
&lt;/tr&gt;
&lt;/tbody&gt;
&lt;/table&gt;&lt;/div&gt;




&lt;h3&gt;
  
  
  Secondary Optimization
&lt;/h3&gt;

&lt;p&gt;The paper also adds two post-processing steps:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;
&lt;strong&gt;Removing redundant intermediate nodes&lt;/strong&gt; — eliminates unnecessary waypoints&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;Bessel curve smoothing&lt;/strong&gt; — makes the final path practical for real
robots and vehicles&lt;/li&gt;
&lt;/ul&gt;




&lt;h3&gt;
  
  
  Results
&lt;/h3&gt;

&lt;p&gt;Compared to standard A*, the adaptive version reduces nodes searched by an&lt;br&gt;
average of &lt;strong&gt;76.4%&lt;/strong&gt; and reduces total turn angle by &lt;strong&gt;71.7%&lt;/strong&gt;. For autonomous&lt;br&gt;
vehicles and robots, this means less computation and smoother, safer paths.&lt;/p&gt;




&lt;h3&gt;
  
  
  Connection to Our Course
&lt;/h3&gt;

&lt;p&gt;In our lectures on informed search, we study A* as the gold standard —&lt;br&gt;
guaranteed to find the optimal solution if &lt;strong&gt;h(n) is admissible&lt;/strong&gt; (never&lt;br&gt;
overestimates the true cost).&lt;/p&gt;

&lt;p&gt;This paper asks a deeper question:&lt;/p&gt;

&lt;blockquote&gt;
&lt;p&gt;Even within admissible heuristics, can we be smarter about how much we&lt;br&gt;
trust the heuristic at different points in the search?&lt;/p&gt;
&lt;/blockquote&gt;

&lt;p&gt;The answer is yes — and it maps directly to our course concept of&lt;br&gt;
&lt;strong&gt;weighted A&lt;/strong&gt;&lt;em&gt;. What makes this paper genuinely interesting is that the&lt;br&gt;
weight is not a fixed constant chosen by a human engineer, but a function&lt;br&gt;
of the agent's position, computed automatically. The **Euclidean distance&lt;br&gt;
heuristic&lt;/em&gt;* we study in class is retained, but modulated dynamically rather&lt;br&gt;
than applied uniformly.&lt;/p&gt;




&lt;h2&gt;
  
  
  My NotebookLM Experience vs. Manual Reading
&lt;/h2&gt;

&lt;p&gt;I read both papers manually first, and it was harder than expected. The&lt;br&gt;
taxonomy tables in the survey paper required multiple passes. The radial&lt;br&gt;
basis function notation in the A* paper was not immediately clear. But&lt;br&gt;
that struggle built real understanding.&lt;/p&gt;

&lt;p&gt;When I then used &lt;strong&gt;NotebookLM&lt;/strong&gt;, it was useful for generating follow-up&lt;br&gt;
questions and locating specific sections quickly — but it simplified&lt;br&gt;
concepts I now knew were more nuanced. It flattened the dynamic weighting&lt;br&gt;
idea into something almost trivial compared to what the paper actually&lt;br&gt;
presents.&lt;/p&gt;

&lt;blockquote&gt;
&lt;p&gt;My recommendation: read the paper yourself first, then use NotebookLM&lt;br&gt;
to fill gaps and test your understanding — not to replace the reading.&lt;/p&gt;
&lt;/blockquote&gt;

&lt;h2&gt;
  
  
  Conclusion
&lt;/h2&gt;

&lt;p&gt;I came into this assignment thinking of search algorithms and agents as&lt;br&gt;
separate textbook chapters. I am leaving it understanding them as two&lt;br&gt;
perspectives on the same underlying problem: how does an intelligent system&lt;br&gt;
efficiently navigate a space of possibilities to reach a goal?&lt;/p&gt;

&lt;p&gt;Whether that space is a city map, the open web, or a robot's environment,&lt;br&gt;
the core problem is identical. The adaptive A* paper makes the heuristic&lt;br&gt;
context-sensitive. The Deep Search Agent paper replaces a hand-coded&lt;br&gt;
heuristic with an LLM's judgment about relevance. Both are moving in the&lt;br&gt;
same direction — search that adapts to where it is, rather than applying&lt;br&gt;
the same strategy everywhere blindly.&lt;/p&gt;




&lt;p&gt;&lt;em&gt;cc: &lt;a class="mentioned-user" href="https://dev.to/raqeeb_26"&gt;@raqeeb_26&lt;/a&gt;&lt;/em&gt;&lt;/p&gt;

&lt;p&gt;&lt;em&gt;Paper references:&lt;/em&gt;&lt;br&gt;
&lt;em&gt;Xi et al. (2026). A Survey of LLM-based Deep Search Agents. arXiv:2508.05668&lt;/em&gt;&lt;br&gt;
&lt;em&gt;Liu et al. (2025). Research on the A&lt;/em&gt; Algorithm Based on Adaptive Weights.*&lt;/p&gt;

</description>
      <category>ai</category>
      <category>machinelearning</category>
      <category>programming</category>
      <category>career</category>
    </item>
  </channel>
</rss>
