<?xml version="1.0" encoding="UTF-8"?>
<rss version="2.0" xmlns:atom="http://www.w3.org/2005/Atom" xmlns:dc="http://purl.org/dc/elements/1.1/">
  <channel>
    <title>DEV Community: Micol Altomare</title>
    <description>The latest articles on DEV Community by Micol Altomare (@micolaltomare).</description>
    <link>https://dev.to/micolaltomare</link>
    <image>
      <url>https://media2.dev.to/dynamic/image/width=90,height=90,fit=cover,gravity=auto,format=auto/https:%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Fuser%2Fprofile_image%2F2594178%2F474d0190-c508-4f2d-bc01-f52c1c8a657b.jpg</url>
      <title>DEV Community: Micol Altomare</title>
      <link>https://dev.to/micolaltomare</link>
    </image>
    <atom:link rel="self" type="application/rss+xml" href="https://dev.to/feed/micolaltomare"/>
    <language>en</language>
    <item>
      <title>Learning from the Future: 7 Insights from UC Berkeley's LLM Agents Course</title>
      <dc:creator>Micol Altomare</dc:creator>
      <pubDate>Sun, 01 Jun 2025 06:22:05 +0000</pubDate>
      <link>https://dev.to/micolaltomare/learning-from-the-future-7-insights-from-uc-berkeleys-llm-agents-course-2nbd</link>
      <guid>https://dev.to/micolaltomare/learning-from-the-future-7-insights-from-uc-berkeleys-llm-agents-course-2nbd</guid>
      <description>&lt;p&gt;I’m a firm believer in lifelong learning, especially in a field as dynamic as Artificial Intelligence. So, when I heard about the Advanced LLM Agents MOOC (Massive Open Online Course) offered by UC Berkeley, featuring lectures from leading researchers, I knew I had to dive in. The promise of learning from the architects of tomorrow's AI, like Prof. Dawn Song and a host of industry pioneers, was too good to pass up. You can explore the course yourself – the content is publicly available: &lt;a href="https://llmagents-learning.org/sp25" rel="noopener noreferrer"&gt;https://llmagents-learning.org/sp25&lt;/a&gt;.&lt;/p&gt;

&lt;p&gt;One lecture that particularly resonated with me was "Learning to Self-Improve &amp;amp; Reason with LLMs" by Jason Weston of Meta &amp;amp; NYU. His presentation was a deep dive into the cutting edge of making LLMs smarter and more capable. Here are my top 7 takeaways from his insightful session:&lt;/p&gt;

&lt;ol&gt;
&lt;li&gt;&lt;p&gt;The Grand Goal: Self-Training AI: Weston kicked off by outlining a tantalizing research ambition: an AI that "trains" itself as much as possible. This involves the AI creating new tasks, evaluating its own performance ("self-rewarding"), and updating itself. The big question? Can this iterative self-improvement lead to superhuman capabilities? This set an inspiring tone for the entire lecture.&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;System 1 vs. System 2 Thinking for LLMs: A core concept discussed was the distinction between two types of reasoning. System 1 is reactive and associative, which is how current LLMs often operate, leading to issues like hallucinations and spurious correlations. System 2 is more deliberate, involving multiple "calls" to System 1, planning, and verification. The lecture emphasized that improving reasoning in LLM agents means developing their System 2 capabilities.&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;The Shoulders of Giants – Building on Past Breakthroughs: It was fascinating to see the "pre-history" of modern LLMs. Weston highlighted work like the 2008 paper "A Unified Architecture for Natural Language Processing: Deep Neural Networks with Multitask Learning" by Collobert and Weston[cite: 16]. This paper, which initially wasn't mainstream, eventually won a "Test of Time Award", underscoring that today's rapid advancements are built upon years of foundational research.&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;Beyond Scaling – The Need for Smarter, Not Just Bigger: The "scaling hypothesis" – the idea that bigger models and datasets automatically lead to better performance – was discussed, referencing Ilya Sutskever's influential insight. However, Weston pointed out that just language modeling (and by extension, scaling alone) isn't enough to create truly intelligent agents, stating the answer to "Is just language modeling enough?" is "no". This realization motivates the push towards self-improvement and enhanced reasoning.&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;Self-Rewarding LLMs in Action: A significant portion of the talk focused on Self-Rewarding LLMs, where models learn to assign rewards to their own outputs and optimize themselves. The iterative two-step process is key: (1) Self-instruction creation (generating prompts, responses, and self-rewards with the LM) and (2) Instruction-following training (using techniques like Direct Preference Optimization (DPO) on the selected preference pairs). This loop allows models to get better at both following instructions and evaluating their responses over iterations.&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;Making Reasoning Verifiable: Iterative Reasoning Preference Optimization: Weston detailed approaches to improve reasoning on more complex tasks, such as Iterative Reasoning Preference Optimization. This method involves the LLM generating multiple chains of thought (CoTs) and answers for a given problem. Preference pairs are then built based on whether the answer is correct vs. not, and the model is trained using these preferences (DPO + NLL term for correct answers). This shows a concrete path to refining the reasoning process itself through self-generated data.&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;The Next Frontier: Meta-Rewarding LLMs: Taking self-improvement a step further, the lecture touched upon Meta-Rewarding LLMs. Here, the LLM acts as an actor, a judge (evaluating responses), and even a "meta-judge" that evaluates the quality of its own judgments. This "meta-judge" step adds another layer of training signal, aiming to improve judgment capabilities which can saturate in simpler self-rewarding schemes. The meme-worthy image of "WE NEED TO GO DEEPER" perfectly captured this recursive improvement idea.&lt;/p&gt;&lt;/li&gt;
&lt;/ol&gt;

&lt;p&gt;Jason Weston's lecture was a truly eye-opening session, offering a deep dive into the quest to build LLMs that can genuinely reason and essentially lift themselves by their own bootstraps. But his insights into self-improvement, like the fascinating concepts of Self-Rewarding and Meta-Rewarding LLMs, are just one facet of the incredible landscape this UC Berkeley LLM Agents course explores. The syllabus is a veritable feast for AI enthusiasts, covering a wide array of critical areas. You'll find sessions on inference-time techniques for LLM reasoning, how agents manage search and planning, and the intricacies of agentic workflow, tool use, and functional calling. The course also delves into specialized applications like LLMs for code generation and verification, their growing role in mathematics through data curation and theorem proving, and the exciting developments in multimodal autonomous AI agents. And, crucially, it addresses the paramount challenge of building safe and secure agentic AI, a topic Prof. Dawn Song herself champions. &lt;/p&gt;

</description>
    </item>
    <item>
      <title>Learning Agentic AI with UC Berkeley’s LLM Agents Course</title>
      <dc:creator>Micol Altomare</dc:creator>
      <pubDate>Fri, 20 Dec 2024 07:12:37 +0000</pubDate>
      <link>https://dev.to/micolaltomare/learning-agentic-ai-with-uc-berkeleys-llm-agents-course-4g25</link>
      <guid>https://dev.to/micolaltomare/learning-agentic-ai-with-uc-berkeleys-llm-agents-course-4g25</guid>
      <description>&lt;p&gt;As someone spending time in the realms of product in the fintech industry and machine learning at the University of Toronto, I’ve spent a lot of time thinking about the intersection of AI research and its real-world applications. So, when I signed up for &lt;a href="https://llmagents-learning.org/f24" rel="noopener noreferrer"&gt;UC Berkeley’s LLM Agents MOOC&lt;/a&gt;, I was excited to dig into how large language models (LLMs) are evolving into agents that can reason, act, and collaborate in increasingly complex ways.&lt;/p&gt;

&lt;p&gt;This course turned out to be one of the most engaging learning experiences I’ve had recently. The lectures were packed with insights into the theory and practice of LLMs, and the course Discord server gave the whole thing a collaborative, human touch. Talking through lecture concepts, asking questions, and seeing other students’ perspectives made the learning experience way more dynamic and social than I expected for an online course. &lt;/p&gt;

&lt;p&gt;Shunyu Yao’s lecture on ReAct frameworks was definitely a favourite. It broke down how agents can unify reasoning across tasks like question answering, symbolic reasoning, and tool use. It struck me that while these systems are powerful, they’re still far from perfect and require a lot of manual design for domain-specific applications. As someone working on user-facing products, it made me think about how important it is to design systems that balance flexibility with reliability, especially for non-technical end users.&lt;/p&gt;

&lt;p&gt;Another standout lecture came from Burak Gokturk, who discussed trends in enterprise AI. One of the big takeaways was how AI is shifting from dense, single-task models toward sparse, multi-modal ones that can handle everything from text to images to video. This resonates with what I see in the tech world—companies are racing to build generalist systems that can do it all, but the real challenge lies in making them scalable, safe, and cost-effective.&lt;/p&gt;

&lt;p&gt;The course didn’t just stick to theory—it dove into real challenges like debugging monolithic models, building modular AI systems, and even designing agents for software development. One thing that stuck with me from Lecture 5 on compound AI systems was how modularity can make these systems more transparent and controllable. That’s something I think we need more of in the real world, especially as these models become more integrated into workflows that affect actual people.&lt;/p&gt;

&lt;p&gt;But honestly, what made the course special wasn’t just the lectures—it was the format. The asynchronous structure meant I could fit it around my schedule, and the Discord server made it feel like I wasn’t learning in isolation. I appreciated how accessible everything was, from the well-organized slides to the recorded sessions. &lt;/p&gt;

&lt;p&gt;Looking back, the biggest thing I’ve taken away is the importance of bridging research and application. Whether you’re debugging a compound AI system or designing a user-friendly product, it’s all about balancing ambition with responsibility. For now, I’m excited to apply what I’ve learned—both in my studies and in building better tools for the future.&lt;/p&gt;

</description>
    </item>
  </channel>
</rss>
