<?xml version="1.0" encoding="UTF-8"?>
<rss version="2.0" xmlns:atom="http://www.w3.org/2005/Atom" xmlns:dc="http://purl.org/dc/elements/1.1/">
  <channel>
    <title>DEV Community: NativeMind</title>
    <description>The latest articles on DEV Community by NativeMind (@nativemind).</description>
    <link>https://dev.to/nativemind</link>
    <image>
      <url>https://media2.dev.to/dynamic/image/width=90,height=90,fit=cover,gravity=auto,format=auto/https:%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Fuser%2Fprofile_image%2F3358854%2Fc6a7fc76-f6b5-4e20-9909-659e9d3421b6.png</url>
      <title>DEV Community: NativeMind</title>
      <link>https://dev.to/nativemind</link>
    </image>
    <atom:link rel="self" type="application/rss+xml" href="https://dev.to/feed/nativemind"/>
    <language>en</language>
    <item>
      <title>Creating Local Privacy-First AI Agents with Ollama: A Step-by-Step Guide</title>
      <dc:creator>NativeMind</dc:creator>
      <pubDate>Tue, 19 Aug 2025 10:11:16 +0000</pubDate>
      <link>https://dev.to/nativemind/creating-local-privacy-first-ai-agents-with-ollama-a-step-by-step-guide-1dj3</link>
      <guid>https://dev.to/nativemind/creating-local-privacy-first-ai-agents-with-ollama-a-step-by-step-guide-1dj3</guid>
      <description>&lt;p&gt;While the tech world focuses on the impressive capabilities of cloud-based AI agents like ChatGPT and Claude, we're exploring a different question: Can we build truly intelligent AI agents that run entirely on users' local devices?&lt;/p&gt;

&lt;p&gt;The appeal is obvious: complete data privacy, zero network latency, freedom from service limitations, and genuinely personalized experiences. But the technical challenges are equally significant: limited local model capabilities, complex tool calling mechanisms, and ensuring consistent user experience.&lt;/p&gt;

&lt;p&gt;After extensive exploration, we've completed a major upgrade to &lt;a href="https://nativemind.app/" rel="noopener noreferrer"&gt;NativeMind&lt;/a&gt; conversational architecture, taking our first significant step toward local AI agents.&lt;/p&gt;

&lt;h2&gt;
  
  
  Core Challenges of Local AI Agents
&lt;/h2&gt;

&lt;p&gt;Cloud models operate with hundreds of billions of parameters, while models that run smoothly on typical consumer devices usually have only a few billion parameters. This capability gap is particularly evident in agent tasks: complex reasoning, tool selection, and task planning all demand high model performance.&lt;/p&gt;

&lt;p&gt;Local models often struggle with tool calling format accuracy compared to large cloud models. A single format error can break the entire agent workflow, which is unacceptable for user experience. Users expect agents to be both responsive and intelligent.&lt;/p&gt;

&lt;p&gt;Achieving this balance with limited local computational resources presents an extraordinarily difficult challenge.&lt;/p&gt;

&lt;h2&gt;
  
  
  Building AI Agents on Ollama
&lt;/h2&gt;

&lt;p&gt;&lt;strong&gt;Prompt-Based Tool Calling&lt;/strong&gt;&lt;br&gt;
We choose not to use Ollama's native tool calling API. While the native API might be more concise in certain scenarios, it has obvious limitations: it only supports specific models, and there are significant differences between different models.&lt;/p&gt;

&lt;p&gt;Instead, we implemented a completely prompt-based tool calling system combined with multi-layer parsing. This approach has been validated by successful products like Cline. Although more challenging to implement, it delivers greater value by providing users with a consistent agent experience regardless of model limitations, while allowing us to continuously optimize parsing accuracy.&lt;/p&gt;

&lt;p&gt;In this system, AI calls tools by outputting specific XML formats. When searching for information, the AI outputs:&lt;br&gt;
&lt;/p&gt;

&lt;div class="highlight js-code-highlight"&gt;
&lt;pre class="highlight plaintext"&gt;&lt;code&gt;&amp;lt;tool_calls&amp;gt;
&amp;lt;search_online&amp;gt;
&amp;lt;query&amp;gt;machine learning development trends&amp;lt;/query&amp;gt;
&amp;lt;max_results&amp;gt;5&amp;lt;/max_results&amp;gt;
&amp;lt;/search_online&amp;gt;
&amp;lt;/tool_calls&amp;gt;
&lt;/code&gt;&lt;/pre&gt;

&lt;/div&gt;



&lt;p&gt;The system detects this format, parses and executes the corresponding search operation, then returns results to the AI for continued processing.&lt;/p&gt;

&lt;p&gt;Considering local model limitations, we designed a multi-layer parsing system to handle tool calling reliability: a standard parsing layer handles well-formatted tool calls, while a fault-tolerant parsing layer processes incomplete but clearly intentioned calls. This multi-layer fault tolerance ensures that even when model output isn't perfect, user intent can still be correctly understood and executed.&lt;/p&gt;

&lt;p&gt;We also redesigned our tools with clear responsibility boundaries. For example, search tools focus purely on information retrieval, returning only structured search results without actual content, while dedicated content fetching tools handle complete webpage retrieval.&lt;/p&gt;

&lt;p&gt;This modular design allows agents to flexibly call tools based on task requirements. For instance, of five search results, only three pages might actually be valuable, or sometimes just titles and descriptions provide sufficient information without needing detailed content review. This approach also reduces individual tool call complexity and overall token consumption.&lt;/p&gt;

&lt;p&gt;To prevent agents from falling into infinite loops during complex tasks and considering context pressure on local devices, we implemented iteration control, supporting up to 5 rounds of tool calls per session.&lt;/p&gt;

&lt;h2&gt;
  
  
  Environment Awareness System
&lt;/h2&gt;

&lt;p&gt;To leverage local agents' advantages in information integration, we designed a progressive workflow: instead of providing all information to the agent at once, we dynamically acquire information based on task needs, automatically select the most appropriate information sources based on question types, and decide next actions based on already obtained information.&lt;/p&gt;

&lt;p&gt;This design enables capability-limited local models to handle complex environments while reducing token consumption.&lt;/p&gt;

&lt;p&gt;Our dynamic environment context system (environment_details) builds real-time comprehensive environment descriptions including current time, available tabs, PDF documents, images, etc., using structured XML format for easy AI comprehension.&lt;/p&gt;

&lt;p&gt;For example, when a user asks "analyze the correlation between this webpage and that report," the AI accurately understands that "this webpage" refers to the currently selected tab and "that report" refers to the loaded PDF file.&lt;/p&gt;

&lt;p&gt;This environment awareness enables AI to better understand users' current context and make more intelligent decisions.&lt;br&gt;
&lt;/p&gt;

&lt;div class="highlight js-code-highlight"&gt;
&lt;pre class="highlight plaintext"&gt;&lt;code&gt;&amp;lt;user_message&amp;gt;
Analyze the correlation between this webpage and that report
&amp;lt;/user_message&amp;gt;

&amp;lt;environment_details&amp;gt;
Current Time: 2024-07-24 14:30:25
Available Tabs: [
  - Tab 1: React Documentation (current)
  - Tab 2: Vue.js Guide
]
Available PDFs: [
  - PDF 1: Frontend_Framework_Analysis.pdf (24 pages)
]
&amp;lt;/environment_details&amp;gt;
&lt;/code&gt;&lt;/pre&gt;

&lt;/div&gt;



&lt;p&gt;To avoid context redundancy, we implemented differential update mechanisms. In multi-turn conversations, environment update information is only sent when the environment changes, maintaining context accuracy while controlling resource consumption.&lt;/p&gt;

&lt;h2&gt;
  
  
  Model Adaptation and Performance
&lt;/h2&gt;

&lt;p&gt;We conducted comprehensive testing on mainstream local models across scenarios including basic search responses, multi-resource integration analysis, systematic information collection and comparison, and comprehensive text-image-PDF processing. Evaluation dimensions covered answer relevance, tool calling effectiveness, language consistency, and other key metrics. Testing included cloud models as comparison benchmarks to validate our local agent architecture's effectiveness.&lt;/p&gt;

&lt;p&gt;Results show that local models demonstrate promising potential under our agent architecture. More importantly, even weaker models perform better under the new architecture than traditional approaches, demonstrating our technical solution's effectiveness.&lt;/p&gt;

&lt;p&gt;User experience has seen qualitative improvements. Transparent execution processes and immediate feedback significantly improved user satisfaction. Users can now see the agent's complete workflow, understand the logic behind each operation step. This transparency not only enhances experience but also builds trust in agent results.&lt;/p&gt;

&lt;h2&gt;
  
  
  Local vs Cloud Model Comparison
&lt;/h2&gt;

&lt;p&gt;Testing reveals that current local small models already show decent performance in basic operations, with excellent local models approaching cloud model effectiveness in certain scenarios.&lt;/p&gt;

&lt;p&gt;These results demonstrate the feasibility of local agents at the current stage, particularly in tool calling and task execution, where excellent local models can already provide near-cloud-model experiences in specific scenarios.&lt;/p&gt;

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fbxxbdwd9m2h42erkzbrw.png" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fbxxbdwd9m2h42erkzbrw.png" alt="local ai model vs cloud model" width="800" height="705"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;Qwen3 4B achieves the optimal balance between performance and efficiency, making it our top recommendation for local models. Under our new architecture, it achieves a 65% task success rate, matching GPT-4o mini's performance.&lt;/p&gt;

&lt;p&gt;For users seeking ultimate performance, Qwen3 8B provides stronger reasoning capabilities. For resource-constrained environments, Qwen3 1.7B and 0.6B still deliver basically usable experiences.&lt;/p&gt;

&lt;p&gt;Notably, local models currently struggle with language consistency, with the Qwen series performing relatively better, especially with more stable Chinese language support. The multimodal Qwen2.5 VL series shows unique advantages in processing image content, though there's still room for improvement in tool calling stability.&lt;/p&gt;

&lt;h2&gt;
  
  
  Unique Value of Local AI Agents
&lt;/h2&gt;

&lt;p&gt;Through NativeMind's implementation, we've explored the possibility of building intelligent agents on local devices. This approach delivers unique value that cloud solutions cannot match.&lt;/p&gt;

&lt;p&gt;Privacy protection is the most significant advantage, with all data processing completed locally, giving users complete control over their information. Instant response provides zero-network-latency interaction experiences, with particularly obvious advantages in poor network conditions.&lt;/p&gt;

&lt;p&gt;We also see promising development potential for local AI agents. New generation local models continue improving, model capabilities keep advancing, stronger local computing power provides hardware foundation for complex agents, and growing user emphasis on data privacy and localized experiences drives market demand.&lt;/p&gt;

&lt;p&gt;Based on this agent system exploration, we're planning further product evolution: supporting more types of local tools and services, browser automation capabilities, MCP support, enhanced complex task planning abilities, and more personalized experiences based on user habits.&lt;/p&gt;

&lt;p&gt;The new agent capabilities have launched in &lt;a href="https://chromewebstore.google.com/detail/nativemind-your-fully-pri/mgchaojnijgpemdfhpnbeejnppigfllj" rel="noopener noreferrer"&gt;NativeMind's latest version&lt;/a&gt;. You can immediately experience NativeMind's completely redesigned architecture.&lt;/p&gt;

</description>
      <category>ai</category>
      <category>pgaichallenge</category>
      <category>llm</category>
    </item>
    <item>
      <title>How to Run gpt-oss Locally in NativeMind (Setup Guide)</title>
      <dc:creator>NativeMind</dc:creator>
      <pubDate>Thu, 14 Aug 2025 02:18:28 +0000</pubDate>
      <link>https://dev.to/nativemind/how-to-run-gpt-oss-locally-in-nativemind-setup-guide-2d08</link>
      <guid>https://dev.to/nativemind/how-to-run-gpt-oss-locally-in-nativemind-setup-guide-2d08</guid>
      <description>&lt;p&gt;OpenAI recently released &lt;a href="https://openai.com/index/introducing-gpt-oss/" rel="noopener noreferrer"&gt;gpt-oss&lt;/a&gt;, a small open-weight language model, sparking excitement in the AI community. It’s lean, immediately approachable, and lightweight, better for writing, coding, secure offline chatting, etc.&lt;/p&gt;

&lt;p&gt;&lt;a href="https://nativemind.app/" rel="noopener noreferrer"&gt;NativeMind&lt;/a&gt; now supports gpt-oss as one of its integrated local AI models. In this article, you'll learn how to use gpt-oss in NativeMind (with setup steps included) and the frequently asked questions you may have.&lt;/p&gt;

&lt;h2&gt;
  
  
  What is gpt-oss
&lt;/h2&gt;

&lt;p&gt;gpt-oss is an open-weight language model released by OpenAI in August 2025. It’s designed to be:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;Lightweight – runs on consumer hardware&lt;/li&gt;
&lt;li&gt;Open – permissively licensed&lt;/li&gt;
&lt;/ul&gt;

&lt;h2&gt;
  
  
  Model details of gpt-oss
&lt;/h2&gt;

&lt;p&gt;According to OpenAI, you have two options of gpt-oss models currently: gpt-oss:20b and gpt-oss:120b. Here are some differences between them, and you can choose one according to your needs.&lt;/p&gt;

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2F3atb0ym3mris0vuhu307.png" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2F3atb0ym3mris0vuhu307.png" alt="gpt-oss model" width="800" height="327"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;h2&gt;
  
  
  Benefits of Running gpt-oss Locally
&lt;/h2&gt;

&lt;p&gt;gpt-oss is OpenAI’s first general-purpose language model released with open weights and an Apache 2.0 license, allowing full commercial use and local deployment.&lt;/p&gt;

&lt;p&gt;Unlike o3-mini or o4-mini, which are closed and API-only, gpt-oss can be run entirely on your own device or infrastructure—giving you full control over cost, latency, and data privacy.&lt;/p&gt;

&lt;p&gt;The larger variant, gpt-oss-120b, uses a Mixture-of-Experts architecture to deliver strong reasoning performance with optimized efficiency. According to OpenAI, its performance is comparable to o4-mini, making it one of the most powerful open-weight models available today.&lt;/p&gt;

&lt;p&gt;And for gpt-oss-20b, it’s comparable to o3-mini, which means that you can have an “offline ChatGPT” if you run it via NativeMind!&lt;/p&gt;

&lt;h2&gt;
  
  
  How to Set up gpt-oss in NativeMind
&lt;/h2&gt;

&lt;p&gt;NativeMind, your private, open-weight, on-device AI assistant, now supports gpt-oss and other local LLMs like Deepseek, Qwen, Llama, Gemma, Mistral, etc. It’s very easy to set gpt-oss by connecting to Ollama, read the simple guide below.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;Step 1: Setup Ollama in NativeMind&lt;/strong&gt;&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;
&lt;a href="https://github.com/NativeMindBrowser/NativeMindExtension" rel="noopener noreferrer"&gt;Download and install NativeMind&lt;/a&gt; into your browser.&lt;/li&gt;
&lt;li&gt;Follow the &lt;a href="https://nativemind.app/blog/tutorial/ollama-setup/" rel="noopener noreferrer"&gt;simple guide&lt;/a&gt; to set up Ollama on your device.
Tips: You can skip this step if you have already set up NativeMind correctly on your device.&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;&lt;strong&gt;Step 2: Download gpt-oss via Ollama&lt;/strong&gt;&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;Move to Ollama and find the &lt;a href="https://ollama.com/library/gpt-oss" rel="noopener noreferrer"&gt;gpt-oss model&lt;/a&gt;.&lt;/li&gt;
&lt;li&gt;Choose the size you want and click Use in NativeMind option to download.&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fu5bjt92so6bax38hmvbw.png" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fu5bjt92so6bax38hmvbw.png" alt="download gpt-oss in ollama" width="800" height="442"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;Step 3: Run gpt-oss in NativeMind&lt;/strong&gt;&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;Open NativeMind in your browser, and now you can find the gpt-oss model.&lt;/li&gt;
&lt;li&gt;Select gpt-oss as your current model, and start using it smoothly.&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fq1zdbg74zxmfalxwba9v.png" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fq1zdbg74zxmfalxwba9v.png" alt="install gpt-oss in nativemind" width="800" height="333"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;h2&gt;
  
  
  FAQ about Using gpt-oss in NativeMind
&lt;/h2&gt;

&lt;p&gt;&lt;strong&gt;1. Can I run gpt-oss without a GPU?&lt;/strong&gt;&lt;br&gt;
Yes. The smaller model gpt-oss:20b can run on a modern CPU with around 16GB RAM, though having a GPU will improve performance. The gpt-oss:120b variant generally requires a high-end GPU or server hardware.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;2. Is gpt-oss free to use?&lt;/strong&gt;&lt;br&gt;
Yes. You can use gpt-oss totally free in NativeMind.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;3. What can I use gpt-oss for?&lt;/strong&gt;&lt;br&gt;
You can use it for writing, summarizing, translating, coding, Q&amp;amp;A, and long-document analysis. With NativeMind, all of this can run fully offline, keeping your data private.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;4. Which version should I choose: 20B or 120B?&lt;/strong&gt;&lt;br&gt;
Choose 20B if you want fast, lightweight local AI on standard hardware.&lt;br&gt;
Choose 120B if you have the hardware and need maximum reasoning power for complex tasks.&lt;/p&gt;

&lt;h2&gt;
  
  
  Try It Now
&lt;/h2&gt;

&lt;p&gt;After reading the setup guide above, you may have a general idea of gpt-oss and how to use it in NativeMind. Try it now to have a quick experience on using gpt-oss to summarize web page content, translate multiple languages, chat on context, and even write blog posts, emails, or notes—without any data sending to the cloud.&lt;/p&gt;

&lt;p&gt;👉 &lt;strong&gt;Install NativeMind:&lt;/strong&gt; &lt;a href="https://github.com/NativeMindBrowser/NativeMindExtension" rel="noopener noreferrer"&gt;https://github.com/NativeMindBrowser/NativeMindExtension&lt;/a&gt;&lt;br&gt;
📁 &lt;strong&gt;Setup gpt-oss in Ollama:&lt;/strong&gt; &lt;a href="https://ollama.com/library/gpt-oss" rel="noopener noreferrer"&gt;https://ollama.com/library/gpt-oss&lt;/a&gt;&lt;br&gt;
💬 &lt;strong&gt;Start chatting locally with NativeMind today&lt;/strong&gt;—no cloud, no API key, no limits. Just speed, privacy, and productivity in your browser.&lt;/p&gt;

</description>
      <category>gptoss</category>
      <category>chatgpt</category>
      <category>openai</category>
      <category>nativemind</category>
    </item>
    <item>
      <title>NativeMind vs LM Studio: Which One is Better?</title>
      <dc:creator>NativeMind</dc:creator>
      <pubDate>Tue, 22 Jul 2025 08:02:30 +0000</pubDate>
      <link>https://dev.to/nativemind/nativemind-vs-lm-studio-which-one-is-better-2j4</link>
      <guid>https://dev.to/nativemind/nativemind-vs-lm-studio-which-one-is-better-2j4</guid>
      <description>&lt;h1&gt;
  
  
  NativeMind vs LM Studio: Which Local AI is Better for You
&lt;/h1&gt;

&lt;p&gt;As large language models (LLMs) continue to evolve, many developers and privacy-conscious users are opting to run these models locally—right on their devices. With concerns over &lt;strong&gt;privacy&lt;/strong&gt;, &lt;strong&gt;data exposure&lt;/strong&gt;, &lt;strong&gt;slow internet speeds&lt;/strong&gt;, and &lt;strong&gt;cloud AI dependencies&lt;/strong&gt;, local AI is gaining popularity.&lt;/p&gt;

&lt;p&gt;Today, we compare two standout tools for running LLMs locally —— &lt;strong&gt;&lt;a href="https://nativemind.app/" rel="noopener noreferrer"&gt;NativeMind&lt;/a&gt;&lt;/strong&gt; and &lt;strong&gt;&lt;a href="https://lmstudio.ai/" rel="noopener noreferrer"&gt;LM Studio&lt;/a&gt;&lt;/strong&gt;. Both are designed to make local AI more accessible, but they are built for different types of users, with different use cases. In this post, we'll break down their features and help you decide which tool best fits your needs.&lt;/p&gt;

&lt;h2&gt;
  
  
  Product Overview: NativeMind vs LM Studio
&lt;/h2&gt;

&lt;p&gt;&lt;strong&gt;NativeMind&lt;/strong&gt; is a browser-native AI assistant that enables real-time interaction with webpage content through local LLM inference. As a &lt;strong&gt;Chrome/Firefox extension&lt;/strong&gt;, it works directly within your browser, processing all prompts locally without uploading any data to the cloud.&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;
&lt;p&gt;&lt;strong&gt;What it does&lt;/strong&gt;:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;Summarize, translate, and analyze content directly on your browser&lt;/li&gt;
&lt;li&gt;Powered by &lt;strong&gt;Ollama&lt;/strong&gt; and models like &lt;strong&gt;Deepseek&lt;/strong&gt;, &lt;strong&gt;Qwen&lt;/strong&gt;, &lt;strong&gt;Llama&lt;/strong&gt;, &lt;strong&gt;Gemma&lt;/strong&gt;, and &lt;strong&gt;Mistral&lt;/strong&gt;
&lt;/li&gt;
&lt;li&gt;Offers a &lt;strong&gt;privacy-first&lt;/strong&gt; approach with no data leaving your device&lt;/li&gt;
&lt;li&gt;Ideal for &lt;strong&gt;knowledge workers&lt;/strong&gt;, &lt;strong&gt;researchers&lt;/strong&gt;, and &lt;strong&gt;privacy-conscious&lt;/strong&gt; users who want fast, local AI interaction&lt;/li&gt;
&lt;/ul&gt;


&lt;/li&gt;

&lt;li&gt;&lt;p&gt;⭐️ Star on GitHub: &lt;a href="https://github.com/NativeMindBrowser/NativeMindExtension" rel="noopener noreferrer"&gt;&lt;/a&gt;&lt;a href="https://github.com/NativeMindBrowser/NativeMindExtension" rel="noopener noreferrer"&gt;https://github.com/NativeMindBrowser/NativeMindExtension&lt;/a&gt;  &lt;/p&gt;&lt;/li&gt;

&lt;li&gt;&lt;p&gt;📘 Setup Guide: &lt;a href="https://nativemind.app/blog" rel="noopener noreferrer"&gt;&lt;/a&gt;&lt;a href="https://nativemind.app/blog" rel="noopener noreferrer"&gt;https://nativemind.app/blog&lt;/a&gt;  &lt;/p&gt;&lt;/li&gt;

&lt;li&gt;&lt;p&gt;🏆 #3 Product of the Day on Product Hunt: &lt;a href="https://www.producthunt.com/products/nativemind" rel="noopener noreferrer"&gt;&lt;/a&gt;&lt;a href="https://www.producthunt.com/products/nativemind" rel="noopener noreferrer"&gt;https://www.producthunt.com/products/nativemind&lt;/a&gt;&lt;/p&gt;&lt;/li&gt;

&lt;/ul&gt;

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2F9jq2rqija6h73opy0fdl.png" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2F9jq2rqija6h73opy0fdl.png" alt="NativeMind local ai assistant" width="800" height="400"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;LM Studio&lt;/strong&gt; is a powerful desktop application designed as a runtime hub for running open-source LLMs locally. It includes multi-threaded chat sessions, model management via Hugging Face/GGUF repositories, and a local OpenAI-compatible API server.&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;
&lt;strong&gt;What it does&lt;/strong&gt;:

&lt;ul&gt;
&lt;li&gt;Supports models like &lt;strong&gt;Hugging Face&lt;/strong&gt;, &lt;strong&gt;llama.cpp&lt;/strong&gt;, &lt;strong&gt;Apple MLX&lt;/strong&gt;
&lt;/li&gt;
&lt;li&gt;Ideal for &lt;strong&gt;developers&lt;/strong&gt;, &lt;strong&gt;AI engineers&lt;/strong&gt;, and &lt;strong&gt;researchers&lt;/strong&gt; working on model evaluation, offline LLM pipelines, or API integration&lt;/li&gt;
&lt;li&gt;Allows multi-model experimentation and flexible deployments&lt;/li&gt;
&lt;li&gt;Local inference with full control over the AI environment&lt;/li&gt;
&lt;/ul&gt;


&lt;/li&gt;

&lt;/ul&gt;

&lt;h2&gt;
  
  
  Feature Comparison
&lt;/h2&gt;

&lt;div class="table-wrapper-paragraph"&gt;&lt;table&gt;
&lt;thead&gt;
&lt;tr&gt;
&lt;th&gt;Feature&lt;/th&gt;
&lt;th&gt;NativeMind&lt;/th&gt;
&lt;th&gt;LM Studio&lt;/th&gt;
&lt;/tr&gt;
&lt;/thead&gt;
&lt;tbody&gt;
&lt;tr&gt;
&lt;td&gt;Platform&lt;/td&gt;
&lt;td&gt;Browser extension (Chrome, Firefox)&lt;/td&gt;
&lt;td&gt;Desktop application (Windows, macOS)&lt;/td&gt;
&lt;/tr&gt;
&lt;tr&gt;
&lt;td&gt;Setup Complexity&lt;/td&gt;
&lt;td&gt;Minimal (browser + Ollama runtime)&lt;/td&gt;
&lt;td&gt;Moderate (model downloads + runtime config)&lt;/td&gt;
&lt;/tr&gt;
&lt;tr&gt;
&lt;td&gt;Web Context Awareness&lt;/td&gt;
&lt;td&gt;Yes (live DOM interaction)&lt;/td&gt;
&lt;td&gt;No&lt;/td&gt;
&lt;/tr&gt;
&lt;tr&gt;
&lt;td&gt;Model Management&lt;/td&gt;
&lt;td&gt;Via Ollama&lt;/td&gt;
&lt;td&gt;Hugging Face + local cache&lt;/td&gt;
&lt;/tr&gt;
&lt;tr&gt;
&lt;td&gt;User Interface&lt;/td&gt;
&lt;td&gt;Sidebar UI (overlay, prompt input)&lt;/td&gt;
&lt;td&gt;Full-featured GUI + multi-threaded chat&lt;/td&gt;
&lt;/tr&gt;
&lt;tr&gt;
&lt;td&gt;Internet Required?&lt;/td&gt;
&lt;td&gt;No (post-setup)&lt;/td&gt;
&lt;td&gt;Yes for downloads; offline afterward&lt;/td&gt;
&lt;/tr&gt;
&lt;tr&gt;
&lt;td&gt;API/CLI Support&lt;/td&gt;
&lt;td&gt;No (UX only)&lt;/td&gt;
&lt;td&gt;Yes (OpenAI API server, CLI client)&lt;/td&gt;
&lt;/tr&gt;
&lt;tr&gt;
&lt;td&gt;Privacy Scope&lt;/td&gt;
&lt;td&gt;Full on-device; no telemetry; sandboxed&lt;/td&gt;
&lt;td&gt;No telemetry; system-level permissions&lt;/td&gt;
&lt;/tr&gt;
&lt;tr&gt;
&lt;td&gt;Open Source Status&lt;/td&gt;
&lt;td&gt;Fully open-source&lt;/td&gt;
&lt;td&gt;UI closed-source; SDKs and runtimes are MIT&lt;/td&gt;
&lt;/tr&gt;
&lt;tr&gt;
&lt;td&gt;Ideal Users&lt;/td&gt;
&lt;td&gt;Researchers, analysts, privacy-first&lt;/td&gt;
&lt;td&gt;Developers, LLM engineers, app integrators&lt;/td&gt;
&lt;/tr&gt;
&lt;/tbody&gt;
&lt;/table&gt;&lt;/div&gt;

&lt;h2&gt;
  
  
  Practical Comparison: Interactive Use vs Development Sandbox
&lt;/h2&gt;

&lt;p&gt;Suppose you're analyzing a lengthy technical whitepaper in your browser and want a condensed summary and follow-up Q&amp;amp;A:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;&lt;p&gt;&lt;strong&gt;NativeMind&lt;/strong&gt; enables you to highlight the section, right-click for an AI action, and receive a locally generated summary within seconds—entirely inside your browser. It supports context persistence across tabs and side-by-side translation views.&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;&lt;strong&gt;LM Studio&lt;/strong&gt; requires you to copy content, paste it into a standalone application, configure the target model, and initiate inference. While more flexible, it introduces context-switching and adds manual overhead.&lt;/p&gt;&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;&lt;strong&gt;NativeMind excels in embedded, context-aware AI interaction.&lt;/strong&gt;&lt;br&gt;&lt;br&gt;
&lt;strong&gt;LM Studio functions as a sandbox for LLM operations&lt;/strong&gt;, particularly suited for model benchmarking, API prototyping, or architectural exploration.&lt;/p&gt;

&lt;h2&gt;
  
  
  Privacy and Execution Model
&lt;/h2&gt;

&lt;p&gt;Both platforms emphasize local-first, no-cloud inference. However, their security and isolation models differ:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;&lt;p&gt;&lt;strong&gt;NativeMind&lt;/strong&gt; runs in a constrained browser environment using Manifest V3 APIs. User prompts and webpage content are kept within the extension's memory and forwarded only to the local Ollama runtime. No external servers are ever involved post-setup.&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;&lt;strong&gt;LM Studio&lt;/strong&gt; does not collect user data and explicitly states that all operations stay local. However, as a desktop application with system-level file and network access, it has a broader attack surface and assumes more user trust in the binary distribution.&lt;/p&gt;&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;In regulated or high-sensitivity contexts (e.g., healthcare, finance, legal), NativeMind’s browser-sandboxed inference may offer a more auditable and minimally privileged environment.&lt;/p&gt;

&lt;h2&gt;
  
  
  Architectural Design and Extensibility
&lt;/h2&gt;

&lt;ul&gt;
&lt;li&gt;&lt;p&gt;&lt;strong&gt;NativeMind&lt;/strong&gt; is built on modern web technologies—JavaScript, WebLLM, and browser-native API access. It’s optimized for speed of interaction, using lightweight communication with Ollama through a local HTTP bridge. It does not currently expose CLI or API hooks, focusing instead on frontend UX for non-technical users.&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;&lt;strong&gt;LM Studio&lt;/strong&gt; serves as a modular LLM workstation. It supports integration with GGUF models, custom system prompts, token streaming, and documents-as-context features. Its embedded OpenAI-compatible API server allows seamless use with tools like LangChain, AutoGen, or custom apps.&lt;/p&gt;&lt;/li&gt;
&lt;/ul&gt;

&lt;blockquote&gt;
&lt;p&gt;In short:&lt;br&gt;&lt;br&gt;
&lt;strong&gt;NativeMind&lt;/strong&gt; = Real-time LLM interaction inside your browser tab&lt;br&gt;&lt;br&gt;
&lt;strong&gt;LM Studio&lt;/strong&gt; = Local LLM hub and control panel for experimentation and deployment&lt;/p&gt;
&lt;/blockquote&gt;

&lt;h2&gt;
  
  
  User Profiles and Usage Scenarios
&lt;/h2&gt;

&lt;div class="table-wrapper-paragraph"&gt;&lt;table&gt;
&lt;thead&gt;
&lt;tr&gt;
&lt;th&gt;Scenario&lt;/th&gt;
&lt;th&gt;Better Fit&lt;/th&gt;
&lt;/tr&gt;
&lt;/thead&gt;
&lt;tbody&gt;
&lt;tr&gt;
&lt;td&gt;Summarizing or translating web content&lt;/td&gt;
&lt;td&gt;NativeMind&lt;/td&gt;
&lt;/tr&gt;
&lt;tr&gt;
&lt;td&gt;Experimenting with GGUF/MLX quantized models&lt;/td&gt;
&lt;td&gt;LM Studio&lt;/td&gt;
&lt;/tr&gt;
&lt;tr&gt;
&lt;td&gt;Zero-copy insight extraction from websites&lt;/td&gt;
&lt;td&gt;NativeMind&lt;/td&gt;
&lt;/tr&gt;
&lt;tr&gt;
&lt;td&gt;API-level integration for LLM pipelines&lt;/td&gt;
&lt;td&gt;LM Studio&lt;/td&gt;
&lt;/tr&gt;
&lt;tr&gt;
&lt;td&gt;Secure reading/analysis in regulated fields&lt;/td&gt;
&lt;td&gt;NativeMind&lt;/td&gt;
&lt;/tr&gt;
&lt;tr&gt;
&lt;td&gt;Multi-model tuning and configuration&lt;/td&gt;
&lt;td&gt;LM Studio&lt;/td&gt;
&lt;/tr&gt;
&lt;/tbody&gt;
&lt;/table&gt;&lt;/div&gt;

&lt;h2&gt;
  
  
  Final Thoughts: Two Tools with Different Roles
&lt;/h2&gt;

&lt;p&gt;While &lt;strong&gt;NativeMind&lt;/strong&gt; and &lt;strong&gt;LM Studio&lt;/strong&gt; have many overlapping features, they serve different roles in the local AI ecosystem:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;
&lt;strong&gt;NativeMind&lt;/strong&gt; is a simple, lightweight solution that lets you use AI directly within your browser. It’s &lt;strong&gt;perfect for quick tasks&lt;/strong&gt; like summarizing web content, translating text, and conducting research.&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;LM Studio&lt;/strong&gt; is a powerful, flexible platform designed for &lt;strong&gt;LLM experimentation&lt;/strong&gt;, model evaluation, and integration into larger workflows. It’s ideal for &lt;strong&gt;developers&lt;/strong&gt; and &lt;strong&gt;engineers&lt;/strong&gt; working on complex AI applications.&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;Whether you’re focused on &lt;strong&gt;privacy-first, browser-native AI&lt;/strong&gt; or &lt;strong&gt;building advanced LLM systems&lt;/strong&gt;, the right tool for you depends on your specific needs and workflow.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;&lt;a href="https://chromewebstore.google.com/detail/nativemind-your-fully-pri/mgchaojnijgpemdfhpnbeejnppigfllj" rel="noopener noreferrer"&gt;Try NativeMind today&lt;/a&gt;&lt;/strong&gt;, your fully private, open-source AI assistant that works right in your browser.&lt;/p&gt;

</description>
      <category>ai</category>
      <category>llm</category>
      <category>pgaichallenge</category>
      <category>productivity</category>
    </item>
    <item>
      <title>NativeMind: Replace ChatGPT in 3 Minutes</title>
      <dc:creator>NativeMind</dc:creator>
      <pubDate>Wed, 16 Jul 2025 07:49:12 +0000</pubDate>
      <link>https://dev.to/nativemind/nativemind-replace-chatgpt-in-3-minutes-3bf7</link>
      <guid>https://dev.to/nativemind/nativemind-replace-chatgpt-in-3-minutes-3bf7</guid>
      <description>&lt;p&gt;If you're using ChatGPT to summarize articles, translate pages, or quickly ask questions about web content — but you're also concerned about privacy, data control, or cloud dependency — &lt;strong&gt;&lt;a href="https://nativemind.app/" rel="noopener noreferrer"&gt;NativeMind&lt;/a&gt;&lt;/strong&gt;, which ranked &lt;a href="https://www.producthunt.com/products/nativemind" rel="noopener noreferrer"&gt;#3 Product of the Day on Product Hunt&lt;/a&gt;, might be exactly what you're looking for.&lt;/p&gt;

&lt;p&gt;It's a &lt;strong&gt;private, free, on-device AI assistant&lt;/strong&gt; that runs entirely local. By connecting to &lt;a href="https://ollama.com/" rel="noopener noreferrer"&gt;Ollama&lt;/a&gt; local LLMs, NativeMind delivers the latest AI capabilities right inside your browser - without sending a single byte to cloud servers.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;NativeMind is fully open-source on GitHub&lt;/strong&gt;, and built to give you total control over your data. We'd love you have a free trial, and give us a star if you like our project!&lt;/p&gt;

&lt;h2&gt;
  
  
  🧠Try NativeMind Now
&lt;/h2&gt;

&lt;ul&gt;
&lt;li&gt;GitHub: &lt;a href="https://github.com/NativeMindBrowser/NativeMindExtension" rel="noopener noreferrer"&gt;https://github.com/NativeMindBrowser/NativeMindExtension&lt;/a&gt;
&lt;/li&gt;
&lt;li&gt;Official Website: &lt;a href="https://nativemind.app/" rel="noopener noreferrer"&gt;https://nativemind.app/&lt;/a&gt;
&lt;/li&gt;
&lt;li&gt;Blog: &lt;a href="https://nativemind.app/blog" rel="noopener noreferrer"&gt;https://nativemind.app/blog&lt;/a&gt;
&lt;/li&gt;
&lt;/ul&gt;

&lt;h2&gt;
  
  
  🔐Why Replace ChatGPT with Local AI Tools
&lt;/h2&gt;

&lt;p&gt;ChatGPT is powerful. So are Grok, Claude, Gemini, and others. But for some of us, they come with non-negotiable traffs:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;You need to copy-paste everything into a separate tab&lt;/li&gt;
&lt;li&gt;You can't use them offline&lt;/li&gt;
&lt;li&gt;Your content is sent to a third-party server&lt;/li&gt;
&lt;li&gt;You rely on paid tokens or API quotas&lt;/li&gt;
&lt;li&gt;You have little control over what's remembered, tracked, or stored&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;For developers, researchers, privacy-focused professionals, or anyone working with sensitive content — that's a dealbreaker. That’s where NativeMind fits in.&lt;/p&gt;

&lt;h2&gt;
  
  
  💡How NativeMind Helps You
&lt;/h2&gt;

&lt;p&gt;Imagine you're in deep research mode: five tabs open, scanning a whitepaper, comparing product docs, collecting insights for a client report. You hit a moment where you want a summary, translation, or clarification but without breaking your flow or compromising your data. NativeMind can help you solve this situation:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;
&lt;strong&gt;Context Memory Across Tabs:&lt;/strong&gt; Continue conversations seamlessly across pages.&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;Local AI Search:&lt;/strong&gt; Ask questions or search — no cloud, no API keys.&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;Instant Page Summary:&lt;/strong&gt; Understand any webpage in seconds.&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;Bilingual Translation:&lt;/strong&gt; Translate full pages or text with side-by-side view.&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;Writing Assistant:&lt;/strong&gt; Rewrite, proofread, or rephrase instantly.&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;Custom System Prompts:&lt;/strong&gt; Tailor responses to your workflow.&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;What’s more, it will provide more functions like chat with PDFs and Images in the coming future. NativeMind brings the power of local AI directly into your browsing experience.&lt;/p&gt;

&lt;h2&gt;
  
  
  🧱Why Local-Only Matters
&lt;/h2&gt;

&lt;p&gt;The debate between cloud-based and local AI is no longer philosophical, it’s practical.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;Cloud LLMs are:&lt;/strong&gt;&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;Fast, but expensive&lt;/li&gt;
&lt;li&gt;Capable, but centralized&lt;/li&gt;
&lt;li&gt;Often closed-source, with unclear data retention policies&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;&lt;strong&gt;Local LLMs are:&lt;/strong&gt;&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;Getting dramatically smaller and faster (thanks to quantization + distillation)&lt;/li&gt;
&lt;li&gt;Easier to install (via Ollama or Hugging Face)&lt;/li&gt;
&lt;li&gt;More transparent, flexible, and secure&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;&lt;strong&gt;With NativeMind, you can:&lt;/strong&gt;&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;Work offline&lt;/li&gt;
&lt;li&gt;Avoid vendor lock-in&lt;/li&gt;
&lt;li&gt;Stay compliant with data-sensitive workflows&lt;/li&gt;
&lt;li&gt;Keep your thoughts your own&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;Here is a table to &lt;strong&gt;compare NativeMind to ChatGPT and Ollama&lt;/strong&gt;, so you can learn more about their differences:&lt;/p&gt;

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fw2osltxt3we6so9upvc9.png" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fw2osltxt3we6so9upvc9.png" alt="NativeMind vs ChatGPT vs Ollama" width="800" height="574"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;We’re entering a post-API key era. NativeMind is a glimpse of what local-first AI really looks like in your browser. In short, NativeMind brings together the strengths of Grok’s context awareness, ChatGPT’s interactivity, and Ollama’s local model execution—all in a privacy-preserving Chrome-native wrapper.&lt;/p&gt;

&lt;h2&gt;
  
  
  ⚡4-Step Setup
&lt;/h2&gt;

&lt;p&gt;&lt;strong&gt;Step 1:&lt;/strong&gt; Download NativeMind from the official website, Chrome Web Store, or Firefox Add-ons.&lt;br&gt;
&lt;strong&gt;Step 2:&lt;/strong&gt; Install Ollama on your devices by following its instructions.&lt;br&gt;
&lt;strong&gt;Step 3:&lt;/strong&gt; Open NativeMind from the right corner of your browser, and choose the AI model you need.&lt;br&gt;
&lt;strong&gt;Step 4:&lt;/strong&gt; Start with Quick Actions or chat with it to meet all your needs.&lt;/p&gt;

&lt;p&gt;You can keep reading on the &lt;a href="https://github.com/NativeMindBrowser/NativeMindExtension?tab=readme-ov-file#readme" rel="noopener noreferrer"&gt;setup guide&lt;/a&gt; if you want to learn more details or have any issues during your installation process.&lt;/p&gt;

&lt;h2&gt;
  
  
  🧭NativeMind Isn’t For Everyone — That’s Okay
&lt;/h2&gt;

&lt;p&gt;&lt;strong&gt;Let’s be real:&lt;/strong&gt;&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;If you're happy with ChatGPT and don’t mind the cloud — great.&lt;/li&gt;
&lt;li&gt;If your hardware can’t handle local models — totally fair.&lt;/li&gt;
&lt;li&gt;If you need a general-purpose AI with plugins and memory — we’re not that (yet).&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;&lt;strong&gt;But if you’re someone who:&lt;/strong&gt;&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;Wants AI assistance without giving up privacy&lt;/li&gt;
&lt;li&gt;Is excited by the open-source LLM movement&lt;/li&gt;
&lt;li&gt;Likes lightweight tools that live inside your actual workflow&lt;/li&gt;
&lt;/ul&gt;

&lt;h2&gt;
  
  
  ✍️Final Thoughts
&lt;/h2&gt;

&lt;p&gt;Just install NativeMind on your Chrome/Firefox browser and try it for free, NativeMind will be your personal AI assistant to improve your productivity with 100% privacy.&lt;/p&gt;

</description>
      <category>ai</category>
      <category>productivity</category>
      <category>opensource</category>
      <category>pgaichallenge</category>
    </item>
  </channel>
</rss>
