<?xml version="1.0" encoding="UTF-8"?>
<rss version="2.0" xmlns:atom="http://www.w3.org/2005/Atom" xmlns:dc="http://purl.org/dc/elements/1.1/">
  <channel>
    <title>DEV Community: Centaur AI</title>
    <description>The latest articles on DEV Community by Centaur AI (@centaurai).</description>
    <link>https://dev.to/centaurai</link>
    <image>
      <url>https://media2.dev.to/dynamic/image/width=90,height=90,fit=cover,gravity=auto,format=auto/https:%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Forganization%2Fprofile_image%2F11852%2Fcc64b045-960c-4ce4-b0c0-0e529ad7a855.png</url>
      <title>DEV Community: Centaur AI</title>
      <link>https://dev.to/centaurai</link>
    </image>
    <atom:link rel="self" type="application/rss+xml" href="https://dev.to/feed/centaurai"/>
    <language>en</language>
    <item>
      <title>Neuro-Symbolic AI: From Research to Industry</title>
      <dc:creator>Can Mingir</dc:creator>
      <pubDate>Wed, 10 Dec 2025 17:11:52 +0000</pubDate>
      <link>https://dev.to/centaurai/neuro-symbolic-ai-from-research-to-industry-2d91</link>
      <guid>https://dev.to/centaurai/neuro-symbolic-ai-from-research-to-industry-2d91</guid>
      <description>&lt;p&gt;📘 This week:&lt;br&gt;
Neuro-Symbolic AI brings together the statistical nature of machine learning with the formal reasoning capabilities of symbolic AI. It seeks to offer a balanced approach to contemporary AI technologies, by combining the ability to learn from data, with the capacity to reason upon knowledge acquired from an environment. The main criticism of neural machine learning lies in its lack of explainability and semantics, which are key requirements in safety-critical applications, yet inherent strengths of logic-based methods. Recently, several corporations have publicly announced products and technologies grounded in Neuro-Symbolic AI methodologies. This talk provided a concise review of the foundations, frameworks and tools underlying Neuro-Symbolic AI, along with illustrative applications. It concludes by highlighting current trends and research directions in the field.&lt;/p&gt;

&lt;p&gt;📝 Suggested paper:&lt;br&gt;
A Garcez, LC Lamb:&lt;br&gt;
Neurosymbolic AI: the 3rd wave. Artif. Intell. Review 56(11):12387-12406 (2023)&lt;/p&gt;

&lt;p&gt;​&lt;a href="https://link.springer.com/article/10.1007/s10462-023-10448-w" rel="noopener noreferrer"&gt;https://link.springer.com/article/10.1007/s10462-023-10448-w&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;

  &lt;iframe src="https://www.youtube.com/embed/gXMtZgtZXl8"&gt;
  &lt;/iframe&gt;


&lt;/p&gt;

</description>
      <category>ai</category>
      <category>machinelearning</category>
      <category>llm</category>
      <category>opensource</category>
    </item>
    <item>
      <title>Training Large Reasoning Models</title>
      <dc:creator>Can Mingir</dc:creator>
      <pubDate>Wed, 03 Dec 2025 16:22:49 +0000</pubDate>
      <link>https://dev.to/centaurai/training-large-reasoning-models-1pce</link>
      <guid>https://dev.to/centaurai/training-large-reasoning-models-1pce</guid>
      <description>&lt;p&gt;📘 This week, Dr. Asim Munawar from IBM’s Watson Research will examine how integrating Neuro-Symbolic AI approaches with large language models (LLMs) can enhance their reasoning, interpretability, and reliability.&lt;/p&gt;

&lt;p&gt;Dr. Asim Munawar is a Project Lead at IBM’s Watson Research Center in New York, where he heads efforts to enhance reasoning, planning, and agentic workflows in enterprise-scale large language models. With over 15 years of experience in AI – more than a decade of it at IBM Research – he has held key leadership roles, including Manager and Program Director for Neuro-Symbolic AI &lt;a href="https://asimmunawar.github.io" rel="noopener noreferrer"&gt;https://asimmunawar.github.io&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;​This presentation will explore architectures that combine symbolic logic, structured representations, and deep learning, and outlines strategies to address limitations such as brittleness, hallucination, and lack of transparency in current LLMs.&lt;/p&gt;

&lt;p&gt;​📝 Bonus: The DeepSeek paper will be covered in Asim's presentation.&lt;br&gt;
&lt;a href="https://arxiv.org/abs/2501.12948" rel="noopener noreferrer"&gt;https://arxiv.org/abs/2501.12948&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;

  &lt;iframe src="https://www.youtube.com/embed/PYUMnAaDsHM"&gt;
  &lt;/iframe&gt;


&lt;/p&gt;

</description>
      <category>ai</category>
      <category>machinelearning</category>
      <category>opensource</category>
      <category>llm</category>
    </item>
    <item>
      <title>Can/Will LLMs Learn to Reason?</title>
      <dc:creator>Can Mingir</dc:creator>
      <pubDate>Wed, 12 Nov 2025 16:34:14 +0000</pubDate>
      <link>https://dev.to/centaurai/neuro-symbolic-wednesdays-canwill-llms-learn-to-reason-3did</link>
      <guid>https://dev.to/centaurai/neuro-symbolic-wednesdays-canwill-llms-learn-to-reason-3did</guid>
      <description>&lt;p&gt;Join us for an interactive session exploring Neuro-Symbolic AI. This week, Abulhair Saparov from Purdue University is presenting “Can/Will LLMs Learn to Reason?”&lt;/p&gt;

&lt;p&gt;​This week:&lt;/p&gt;

&lt;p&gt;Reasoning—the process of drawing conclusions from prior knowledge—is a hallmark of intelligence. Large language models, and more recently, large reasoning models have demonstrated impressive results on many reasoning-intensive benchmarks. Careful studies over the past few years have revealed that LLMs may exhibit some reasoning behavior, and larger models tend to do better on reasoning tasks. However, even the largest current models still struggle on various kinds of reasoning problems. In this talk, we will try to address the question: Are the observed reasoning limitations of LLMs fundamental in nature? Or will they be resolved by further increasing the size and data of these models, or by better techniques for training them? I will describe recent work to tackle this question from several different angles.&lt;/p&gt;

&lt;p&gt;

  &lt;iframe src="https://www.youtube.com/embed/c3dCFIQyYnI"&gt;
  &lt;/iframe&gt;


&lt;/p&gt;

</description>
      <category>ai</category>
      <category>machinelearning</category>
      <category>opensource</category>
      <category>llm</category>
    </item>
    <item>
      <title>📢 New Series: Neuro-Symbolic Wednesday</title>
      <dc:creator>Can Mingir</dc:creator>
      <pubDate>Wed, 05 Nov 2025 08:22:27 +0000</pubDate>
      <link>https://dev.to/centaurai/new-series-neuro-symbolic-wednesday-14hb</link>
      <guid>https://dev.to/centaurai/new-series-neuro-symbolic-wednesday-14hb</guid>
      <description>&lt;p&gt;​Join us for an interactive session exploring Neuro-Symbolic AI, the emerging paradigm that blends the strengths of neural networks with symbolic reasoning. We will discuss how hybrid approaches can enhance generalization, interpretability, and reasoning, and how these methods are shaping the future of intelligent systems.&lt;/p&gt;

&lt;p&gt;Whether you’re a researcher, engineer, or simply curious about the cutting edge of AI, you’ll find an engaging space to learn, connect, and exchange ideas.&lt;/p&gt;

&lt;p&gt;Event Format:&lt;br&gt;
🔍 Reading Group&lt;br&gt;
🗣️ Guest Spotlights&lt;br&gt;
💬 Research Roundtable&lt;br&gt;
🛠️ Open-Source Review&lt;br&gt;
🤝 Collaboration Hour&lt;/p&gt;

&lt;p&gt;&lt;a href="https://luma.com/4frkdgzi?utm=devto" rel="noopener noreferrer"&gt;RSVP at Luma&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;

  &lt;iframe src="https://www.youtube.com/embed/WWEu2YyHncI"&gt;
  &lt;/iframe&gt;


&lt;/p&gt;

</description>
      <category>ai</category>
      <category>machinelearning</category>
      <category>opensource</category>
      <category>llm</category>
    </item>
    <item>
      <title>📢 Neuro-Symbolic AI Summer School 2025 | Online Event | Aug 14 - 15</title>
      <dc:creator>Can Mingir</dc:creator>
      <pubDate>Wed, 13 Aug 2025 18:43:42 +0000</pubDate>
      <link>https://dev.to/centaurai/neuro-symbolic-ai-summer-school-2025-online-event-2fh8</link>
      <guid>https://dev.to/centaurai/neuro-symbolic-ai-summer-school-2025-online-event-2fh8</guid>
      <description>&lt;p&gt;&lt;a href="http://lu.ma/pqzv80yd?utm_source=devto" rel="noopener noreferrer"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fcl12gji87ou5iqf6berh.png" alt="Neuro-Symbolic AI Summer School" width="800" height="506"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;&lt;a href="http://lu.ma/pqzv80yd?utm_source=devto" rel="noopener noreferrer"&gt;http://lu.ma/pqzv80yd?utm_source=devto&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;📅 Agenda&lt;/p&gt;

&lt;p&gt;​Day 1: Frameworks and Foundations | AUG 14&lt;/p&gt;

&lt;p&gt;📜 Big Picture&lt;br&gt;
9:00–9:30 AM ET – Alexander Gray, Centaur AI Institute – Neuro-Symbolic AI and NSSS4 (Overview)&lt;br&gt;
9:30–9:40 AM – Q&amp;amp;A / Break&lt;br&gt;
9:40–10:20 AM – Arvind Narayanan, Princeton – Assessing the Current State of AI (Talk)&lt;br&gt;
10:20–10:30 AM – Q&amp;amp;A / Break&lt;/p&gt;

&lt;p&gt;​🏗 Frameworks&lt;br&gt;
10:30–11:05 AM – Leilani Gilpin, UC Santa Cruz – Neuro-Symbolic AI for Safer Autonomous Vehicles (Tutorial)&lt;br&gt;
11:05–11:15 AM – Q&amp;amp;A / Break&lt;br&gt;
11:15–12:00 PM – Mateo Zarlenga &amp;amp; Pietro Barbiero, Cambridge – Foundations of Interpretable Models (Tutorial)&lt;br&gt;
12:00–12:10 PM – Q&amp;amp;A / Break&lt;br&gt;
12:10–12:55 PM – Parisa Kordjamshidi, Michigan State – Compositional Learning in Language and Vision (Tutorial)&lt;br&gt;
12:55–1:05 PM – Q&amp;amp;A / Break&lt;/p&gt;

&lt;p&gt;​📐 Mathematical Foundations&lt;br&gt;
1:05–1:45 PM – Parikshit Ram, IBM Research – How to Measure Compositionality, and Why it Leads to Better Generalization (Talk)&lt;br&gt;
1:45–1:55 PM – Q&amp;amp;A / Break&lt;br&gt;
1:55–2:35 PM – Peihao Wang, UT Austin – Why Neural Networks Can Discover Symbolic Structures (Talk)&lt;br&gt;
2:35–2:45 PM – Q&amp;amp;A / Break&lt;br&gt;
2:45–3:25 PM – Changlong Wu, Purdue – Why Current Models Will Always Hallucinate (and a Path Forward) (Talk)&lt;br&gt;
3:25–3:35 PM – Q&amp;amp;A / Break&lt;br&gt;
3:35–4:35 PM – Sridhar Mahadevan, Adobe Research &amp;amp; UMass Amherst – Category Theory: The Mathematics of Symbolic Structures (Tutorial)&lt;br&gt;
4:35–4:45 PM – Q&amp;amp;A / Break&lt;/p&gt;

&lt;p&gt;​Day 2: Methods and Systems | AUG 15&lt;/p&gt;

&lt;p&gt;​🤖 Neuro-Symbolic AI Software&lt;br&gt;
9:00–9:40 AM – Bowen Li, Carnegie Mellon – Generalizing to New Situations in Robotics (Talk)&lt;br&gt;
9:40–9:50 AM – Q&amp;amp;A / Break&lt;br&gt;
9:50–10:30 AM – Olga Vileskaia &amp;amp; Kevin O'Connor, Centaur AI Institute – Explainability While Retaining Predictive Accuracy (Talk)&lt;br&gt;
10:30–10:40 AM – Q&amp;amp;A / Break&lt;/p&gt;

&lt;p&gt;​🧩 Learning Symbolic Models&lt;br&gt;
10:40–11:20 AM – Bin Yu, UC Berkeley – Interpretable and Veridical Data Science (Talk)&lt;br&gt;
11:20–11:30 AM – Q&amp;amp;A / Break&lt;br&gt;
11:30–12:10 PM – Felix Petersen, Stanford – Deep Differentiable Logic Gate Networks (Talk)&lt;br&gt;
12:10–12:20 PM – Q&amp;amp;A / Break&lt;br&gt;
12:20–1:00 PM – Hikaru Shindo, TU Darmstadt – Neuro-symbolic Agentic Systems (Talk)&lt;br&gt;
1:00–1:10 PM – Q&amp;amp;A / Break&lt;/p&gt;

&lt;p&gt;​🛡 Safer AI Systems&lt;br&gt;
1:10–1:50 PM – Soroush Saghafian, Harvard – Human + AI "Centaur" Systems (Talk)&lt;br&gt;
1:50–2:00 PM – Q&amp;amp;A / Break&lt;br&gt;
2:00–2:40 PM – Pranava Madhyastha, City Univ London – New Results in Controllable Text Generation (Talk)&lt;br&gt;
2:40–2:50 PM – Q&amp;amp;A / Break&lt;/p&gt;

&lt;p&gt;​➗ AI Systems for Mathematics&lt;br&gt;
2:50–3:30 PM – Shange Tang, Princeton – State-of-the-art Performance in Automated Mathematical Theorem Proving (Talk)&lt;br&gt;
3:30–3:40 PM – Q&amp;amp;A / Break&lt;br&gt;
3:40–4:20 PM – Ankit Anand, DeepMind – Curious Case of AI in Maths: Being Proficient in Advancing Open Conjectures in Maths Yet Having Struggles in AI for Education (Talk)&lt;br&gt;
4:20–4:30 PM – Q&amp;amp;A / Break&lt;/p&gt;

&lt;p&gt;​🔮 Looking Forward&lt;br&gt;
4:30–5:30 PM – Panel on The Future of AI – Rich Sutton, Univ Alberta; Leonardo de Moura, Amazon; Artur Garcez, City Univ London; more TBA; moderator: Alexander Gray – Discussion including open Q&amp;amp;A&lt;br&gt;
5:30–5:35 PM – Alexander Gray, Centaur AI Institute – What's Coming Next and How to Participate (Closing remarks)&lt;/p&gt;

</description>
      <category>ai</category>
      <category>machinelearning</category>
    </item>
  </channel>
</rss>
