<?xml version="1.0" encoding="UTF-8"?>
<rss version="2.0" xmlns:atom="http://www.w3.org/2005/Atom" xmlns:dc="http://purl.org/dc/elements/1.1/">
  <channel>
    <title>DEV Community: Lana</title>
    <description>The latest articles on DEV Community by Lana (@lem36).</description>
    <link>https://dev.to/lem36</link>
    <image>
      <url>https://media2.dev.to/dynamic/image/width=90,height=90,fit=cover,gravity=auto,format=auto/https:%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Fuser%2Fprofile_image%2F1298781%2F196244a6-7268-42a1-8aa5-848929a65da7.png</url>
      <title>DEV Community: Lana</title>
      <link>https://dev.to/lem36</link>
    </image>
    <atom:link rel="self" type="application/rss+xml" href="https://dev.to/feed/lem36"/>
    <language>en</language>
    <item>
      <title>I Built an AI That Whispers Interview Answers in Real Time</title>
      <dc:creator>Lana</dc:creator>
      <pubDate>Fri, 10 Apr 2026 09:58:14 +0000</pubDate>
      <link>https://dev.to/lem36/i-built-an-ai-that-whispers-interview-answers-in-real-time-e43</link>
      <guid>https://dev.to/lem36/i-built-an-ai-that-whispers-interview-answers-in-real-time-e43</guid>
      <description>&lt;p&gt;You know that moment in a job interview when someone asks you a question and your brain just... empties? You have prepared for hours. You know the answer. But under pressure, the words vanish.&lt;/p&gt;

&lt;p&gt;I built TalkBuoy to fix that.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;What It Actually Does&lt;/strong&gt;&lt;/p&gt;

&lt;p&gt;TalkBuoy is a PWA that sits on your phone or tablet during an interview, presentation, or any high-pressure conversation. It listens to what the other person says, processes it through AI, and silently displays suggested talking points on your screen.&lt;/p&gt;

&lt;p&gt;You glance down, see a structured response with key points you should hit, and weave them into your own words. The other person never knows.&lt;/p&gt;

&lt;p&gt;Think of it as a teleprompter, but one that writes itself in real time based on what is actually being discussed.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;How It Works Under the Hood&lt;/strong&gt;&lt;/p&gt;

&lt;p&gt;The app captures audio through the device microphone and runs it through speech recognition to convert what the interviewer says into text. That text gets sent to an AI model that understands the context of the conversation and generates relevant talking points.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;The key technical challenges were:&lt;/strong&gt;&lt;/p&gt;

&lt;p&gt;Voice filtering. You do not want the app responding to your own voice. TalkBuoy distinguishes between the user and the other speaker, filtering out your speech so suggestions only respond to the interviewer's questions.&lt;/p&gt;

&lt;p&gt;Speed. In a real conversation, you have maybe 2 to 3 seconds of natural pause before silence gets awkward. The entire pipeline from audio capture to displayed suggestion needs to complete in that window. Every millisecond matters.&lt;/p&gt;

&lt;p&gt;Smart auto-mute. While you are reading the suggestions, the microphone needs to behave. TalkBuoy silences the mic during reading so the other person never hears anything unusual from your end.&lt;/p&gt;

&lt;p&gt;Context awareness. The AI does not just answer the current question in isolation. It maintains context from the entire conversation so suggestions build on what has already been discussed. If you mentioned a project earlier, and the interviewer follows up on it, the AI references your earlier points.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;The Use Cases That Surprised Me&lt;/strong&gt;&lt;/p&gt;

&lt;p&gt;I originally built this for job interviews. That is still the primary use case, and it handles behavioral questions, technical questions, and panel interviews well.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;But users started using it for things I did not expect:&lt;/strong&gt;&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;Public speaking Q&amp;amp;A.&lt;/strong&gt; Speakers use it during the question period after presentations. You can nail a prepared talk but freeze when someone asks something unexpected. TalkBuoy fills that gap.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;Sales calls.&lt;/strong&gt; Sales reps use it to handle objections in real time. When a prospect raises a concern, suggested responses appear instantly with relevant data points and rebuttals.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;Language practice.&lt;/strong&gt; People learning a second language use it as a conversation partner. It helps them formulate responses in the target language when they get stuck during actual conversations.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;Customer service.&lt;/strong&gt; Support reps dealing with complex product questions get suggested answers without putting the customer on hold to search a knowledge base.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;Podcast interviews.&lt;/strong&gt; Hosts use it when interviewing guests on unfamiliar topics. It suggests follow-up questions based on what the guest just said, leading to better, more natural conversations.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;The Ethics Question&lt;/strong&gt;&lt;/p&gt;

&lt;p&gt;Every time I show this to someone, the first reaction is either "that is genius" or "that is cheating." Fair enough, let me address it.&lt;/p&gt;

&lt;p&gt;TalkBuoy does not put words in your mouth. It suggests talking points. You still need to understand the material, speak naturally, and think on your feet. If you have zero knowledge of a topic, a list of bullet points in your ear will not save you. The interviewer will hear the difference.&lt;/p&gt;

&lt;p&gt;What it does is prevent the specific failure mode where you know the answer but cannot access it under pressure. That is a performance anxiety problem, not a knowledge problem. Most people who use it say they stop needing it after a few interviews because the practice with the safety net builds genuine confidence.&lt;/p&gt;

&lt;p&gt;It is the same principle as training wheels. You use them until you do not need them.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;Privacy&lt;/strong&gt;&lt;/p&gt;

&lt;p&gt;This was non-negotiable from day one. No audio is stored. No conversations are saved to servers. Everything is processed and discarded. There is no recording, no transcript history, no data mining.&lt;/p&gt;

&lt;p&gt;The app works as a PWA, which means it runs in the browser with no app store installation required. It works on phones, tablets, and laptops. Place your device on the table or prop it up where you can glance at it naturally.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;What I Learned Building It&lt;/strong&gt;&lt;/p&gt;

&lt;p&gt;Latency is everything in real-time AI. A suggestion that arrives 5 seconds late is useless. I spent more time optimizing the pipeline speed than on any other feature.&lt;/p&gt;

&lt;p&gt;People do not read paragraphs under pressure. Early versions generated full paragraph responses. Nobody could process that while maintaining eye contact and a conversation. Bullet points with 5 to 8 words each turned out to be the right format.&lt;/p&gt;

&lt;p&gt;The voice filtering problem is harder than it sounds. Distinguishing between two voices in the same room, especially through a phone microphone with room echo, required a lot of iteration. Getting this wrong means the app starts responding to your own answers, which creates a bizarre feedback loop.&lt;/p&gt;

&lt;p&gt;PWA was the right choice. No app store approval process, no platform restrictions, instant updates. Users just visit the URL and it works. For a tool that people might need on short notice before an interview, eliminating friction was critical.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;Try It&lt;/strong&gt;&lt;/p&gt;

&lt;p&gt;&lt;a href="//talkbuoy.com"&gt;TalkBuoy&lt;/a&gt; is live and free to try. If you have an interview coming up, a presentation to give, or you just want to see how real-time AI coaching feels, give it a shot.&lt;/p&gt;

&lt;p&gt;Would love to hear what you think, and if you have ideas for other use cases I have not considered.&lt;/p&gt;

&lt;p&gt;What is the worst interview freeze moment you have experienced? Drop it in the comments.&lt;/p&gt;

</description>
      <category>ai</category>
      <category>webdev</category>
      <category>productivity</category>
      <category>career</category>
    </item>
    <item>
      <title>How I Built a Free AI Voice Separator for Podcasts and Interviews</title>
      <dc:creator>Lana</dc:creator>
      <pubDate>Wed, 18 Mar 2026 08:59:41 +0000</pubDate>
      <link>https://dev.to/lem36/how-i-built-a-free-ai-voice-separator-for-podcasts-and-interviews-16b9</link>
      <guid>https://dev.to/lem36/how-i-built-a-free-ai-voice-separator-for-podcasts-and-interviews-16b9</guid>
      <description>&lt;p&gt;My first Dev.to post about building &lt;a href="https://www.toolsonfire.com" rel="noopener noreferrer"&gt;ToolsOnFire &lt;/a&gt;covered the broad overview. This time I want to go deep on one specific tool: the &lt;a href="https://www.toolsonfire.com/en/voice-separator" rel="noopener noreferrer"&gt;Voice Separator&lt;/a&gt;.                                                                                                                           &lt;/p&gt;

&lt;p&gt;*&lt;em&gt;The Problem *&lt;/em&gt;                                                                                                                                                           &lt;/p&gt;

&lt;p&gt;I kept seeing the same requests in podcasting and journalism communities: "I recorded an interview and need to edit just one speaker's audio" or "I need a transcript that shows who said what."&lt;/p&gt;

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fk0iwrxx88yg2af3byeyg.png" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fk0iwrxx88yg2af3byeyg.png" alt=" " width="800" height="583"&gt;&lt;/a&gt;&lt;br&gt;
                                                                                                                                                                           The existing options were either expensive (Descript at $24/month), required desktop software, or didn't actually separate the audio - they just labelled who spoke when.&lt;br&gt;
                                                                                                                                                                           I  wanted to build something that:                                                                                                                                        1. Identifies each speaker in a recording&lt;/p&gt;

&lt;ol&gt;
&lt;li&gt;Creates separate downloadable audio files per speaker&lt;/li&gt;
&lt;li&gt;Allows users to separate background music into a separate file
&lt;/li&gt;
&lt;li&gt;Produces a timestamped transcript with speaker labels&lt;/li&gt;
&lt;li&gt;Is free to try without creating an account
&lt;/li&gt;
&lt;/ol&gt;

&lt;p&gt;&lt;strong&gt;The Challenges&lt;/strong&gt;&lt;br&gt;&lt;br&gt;
                                                                                                                                                                           Transcript accuracy is never perfect. This was the biggest reality check. No matter which AI model you use, transcripts will have errors - especially with accents, technical jargon, mumbling, or background noise. I spent a long time chasing 100% accuracy before accepting that even professional human transcribers don't achieve that. The goal became "accurate enough to be useful" rather than perfect.   &lt;/p&gt;

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2F9371ggtxli4is0p0jz3m.png" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2F9371ggtxli4is0p0jz3m.png" alt=" " width="800" height="583"&gt;&lt;/a&gt;                                                                                           &lt;/p&gt;

&lt;p&gt;Speaker misidentification. The AI sometimes assigns the wrong speaker label to short utterances, especially when speakers have similar voices or one person only says a few words. I had to build post-processing logic to smooth out these errors - grouping nearby utterances and correcting obvious misattributions.&lt;br&gt;
                                                                                                                                                                           Overlapping speech is the hardest problem. When two people talk over each other, basic diarization falls apart. My free tier handles this reasonably well for brief interruptions, but I built a premium tier with a more advanced pipeline specifically for recordings with heavy crosstalk - panel discussions, heated interviews, group meetings.                                                                                                                                                              &lt;/p&gt;

&lt;p&gt;Audio quality varies wildly. A studio-recorded podcast processes beautifully. A phone call recorded on speakerphone in a noisy cafe is a completely different challenge. I had to set expectations clearly in the UI and add guidance about what makes a good recording for separation.&lt;br&gt;
                                                                                                                                                                           Processing time and user feedback. Some recordings take 30-60 seconds to process. My initial spinning loader felt broken for anything longer than 10 seconds. I replaced it with a simulated progress bar that moves through phases (uploading... processing... generating results...). Users need to see movement even when I have no real progress data from the AI.                                                                                                                                        &lt;/p&gt;

&lt;p&gt;*&lt;em&gt;Cost control *&lt;/em&gt;&lt;/p&gt;

&lt;p&gt;The AI APIs cost real money per minute of audio processed. Without proper limits, someone could process hours of audio for free. I built a tiered system  with minute-based quotas, prepaid credit packs, and usage tracking.&lt;/p&gt;

&lt;p&gt;*&lt;em&gt;What I Learned *&lt;/em&gt; &lt;/p&gt;

&lt;p&gt;Free tiers drive conversions. Letting people try with no account was the right call. Most users try it once and leave, but the ones who find it useful come back and upgrade. If I'd required sign-up from the start, most people would never have tried it.&lt;br&gt;
                                                                                                                                                                           Podcasters are the sweet spot. I built this for a broad audience but podcasters are by far the most engaged users. They record regularly, always need to edit individual speakers, and the time savings are immediate. If I were starting over, I'd market specifically to podcasters from day one.&lt;br&gt;
                                                                                                                                                                           Managing expectations matters more than improving accuracy. Users who understand the limitations upfront are happy with 90% accuracy. Users who expect perfection are    frustrated at 95%. Clear communication about what the tool can and can't do made a bigger difference to satisfaction than any technical improvement.&lt;br&gt;
                                                                                                                                                                           &lt;strong&gt;Try It&lt;/strong&gt;    &lt;/p&gt;

&lt;p&gt;The &lt;a href="https://www.toolsonfire.com/en/voice-separator" rel="noopener noreferrer"&gt; Voice Separator&lt;/a&gt; is free to try - upload any recording with 2 speakers, up to 5 minutes, no account needed.                             &lt;/p&gt;

&lt;p&gt;I also built a &lt;a href="https://www.meetingrecordntranscribe.app" rel="noopener noreferrer"&gt;Meeting Recorder, Transcriber and Summarizer&lt;/a&gt; for live recording and transcription, and &lt;a href="https://talkbuoy.com" rel="noopener noreferrer"&gt;Talkbuoy&lt;/a&gt; for AI speech coaching.                            &lt;/p&gt;

&lt;p&gt;Have you built anything with audio or speech processing? I'd love to hear what challenges you ran into.      &lt;/p&gt;

</description>
      <category>webdev</category>
      <category>ai</category>
      <category>buildinpublic</category>
      <category>nextjs</category>
    </item>
    <item>
      <title>I Built 35+ Free Browser Tools as a Solo Dev - Here's How</title>
      <dc:creator>Lana</dc:creator>
      <pubDate>Tue, 17 Mar 2026 14:20:16 +0000</pubDate>
      <link>https://dev.to/lem36/i-built-35-free-browser-tools-as-a-solo-dev-heres-how-4b5n</link>
      <guid>https://dev.to/lem36/i-built-35-free-browser-tools-as-a-solo-dev-heres-how-4b5n</guid>
      <description>&lt;div class="highlight js-code-highlight"&gt;
&lt;pre class="highlight plaintext"&gt;&lt;code&gt;                                                                                                                                                                                                                                                                                    I Built 35+ Free Browser Tools as a Solo Dev                                                                                                                                                                                                                                                                                                    
&lt;/code&gt;&lt;/pre&gt;

&lt;/div&gt;

&lt;p&gt;I've been building &lt;a href="https://www.toolsonfire.com" rel="noopener noreferrer"&gt;ToolsOnFire&lt;/a&gt; - a collection of 35+ free online tools. Video converters, AI voice separation, a sticker        creator, markdown converter, and more.                                                                                                                                    &lt;/p&gt;

&lt;p&gt;I'm a solo developer from the UK based in Norway. No team, no funding. Just a frustration with paying subscriptions for simple tasks, so I built free alternatives.  &lt;/p&gt;

&lt;p&gt;*&lt;em&gt;The Stack  *&lt;/em&gt;  &lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;Next.js 14 (App Router) with TypeScript                                                                                                                                - Firebase for auth and database&lt;/li&gt;
&lt;li&gt;Stripe for premium tiers
&lt;/li&gt;
&lt;li&gt;Vercel for hosting
                                                                                                                                                                       The Key Decision: 
Process in the Browser                                                                                                                                                                                                                                                                                                          Most of my tools process files entirely in the user's browser rather than on a server. &lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;My &lt;a href="https://www.toolsonfire.com/en/video-to-gif-converter" rel="noopener noreferrer"&gt;Video to Gif Converter&lt;/a&gt; and &lt;a href="https://www.toolsonfire.com/en/VideoToMp3Converter" rel="noopener noreferrer"&gt;Video to audio converter&lt;/a&gt; use FFmpeg.wasm - a WebAssembly port of FFmpeg.&lt;br&gt;
                                                                                                                                                                           This means files never leave the user's device. No uploads, no privacy concerns, and I don't pay for server compute.         &lt;/p&gt;

&lt;p&gt;The trade-off is browser memory limits. I had to add device detection to cap file sizes on mobile and optimise how large files are handled.                              &lt;/p&gt;

&lt;p&gt;*&lt;em&gt;AI-Powered Tools *&lt;/em&gt;                                                                                                                                                      &lt;/p&gt;

&lt;p&gt;Some tools need real AI models:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;&lt;p&gt;&lt;a href="https://www.toolsonfire.com/en/voice-separator" rel="noopener noreferrer"&gt;Voice Separator and background music extractor&lt;/a&gt; - Splits recordings by speaker using Deepgram and Replicate. Useful for podcasts, interviews, and meetings.   &lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;&lt;a href="https://www.toolsonfire.com/en/sticker-creator" rel="noopener noreferrer"&gt;Sticker Creator&lt;/a&gt; - Generates custom stickers with DALL-E 3, or removes backgrounds from photos client-side using an ONNX model.&lt;br&gt;
                                                                                                                                                                       *&lt;em&gt;What I've Learned *&lt;/em&gt;&lt;br&gt;
                                                                                                                                                                                                                                                                                                                            You have to try things to see if they work. My top 5 tools drive 90% of traffic, but I wouldn't have known which 5 without building all of them. Some tools I expected to be popular get no visits. Others I built on a whim end up surprising me.&lt;br&gt;
                                                                                                                                                                       SEO is the real challenge. Building the tools was the easy part. Organic traffic growth is still slow. If I could go back, I'd start writing content and building backlinks from day one.&lt;br&gt;
                                                                                                                                                                 *&lt;em&gt;My Other Projects  *&lt;/em&gt;                                                                                                                                                   &lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;&lt;a href="https://www.meetingrecordntranscribe.app" rel="noopener noreferrer"&gt;Meeting Transcriber&lt;/a&gt; - Records and transcribes meetings with automatic speaker identification                                                      - &lt;a href="https://talkbuoy.com" rel="noopener noreferrer"&gt;Speech Assistant Tool&lt;/a&gt; - AI speech coaching with real-time feedback&lt;br&gt;
                                                                                                                                                                       I'm always building and experimenting with new tools. If something works, I double down. If it doesn't, I move on.                                                                                                                                                                                                                                &lt;/p&gt;&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;Would love to hear from anyone who's grown a tools site from zero - what worked for you?     &lt;/p&gt;

</description>
      <category>webdev</category>
      <category>nextjs</category>
      <category>javascript</category>
      <category>buildinpublic</category>
    </item>
  </channel>
</rss>
