<?xml version="1.0" encoding="UTF-8"?>
<rss version="2.0" xmlns:atom="http://www.w3.org/2005/Atom" xmlns:dc="http://purl.org/dc/elements/1.1/">
  <channel>
    <title>DEV Community: Thiago da Silva Teixeira</title>
    <description>The latest articles on DEV Community by Thiago da Silva Teixeira (@thiagoteixeiradev).</description>
    <link>https://dev.to/thiagoteixeiradev</link>
    <image>
      <url>https://media2.dev.to/dynamic/image/width=90,height=90,fit=cover,gravity=auto,format=auto/https:%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Fuser%2Fprofile_image%2F2454726%2Fbe3912cb-ce4f-4351-af86-a85e936bbdb0.jpg</url>
      <title>DEV Community: Thiago da Silva Teixeira</title>
      <link>https://dev.to/thiagoteixeiradev</link>
    </image>
    <atom:link rel="self" type="application/rss+xml" href="https://dev.to/feed/thiagoteixeiradev"/>
    <language>en</language>
    <item>
      <title>Are "Agent Skills" the Secret Sauce for AI Productivity?</title>
      <dc:creator>Thiago da Silva Teixeira</dc:creator>
      <pubDate>Mon, 16 Feb 2026 21:46:28 +0000</pubDate>
      <link>https://dev.to/thiagoteixeiradev/are-agent-skills-the-secret-sauce-for-ai-productivity-73f</link>
      <guid>https://dev.to/thiagoteixeiradev/are-agent-skills-the-secret-sauce-for-ai-productivity-73f</guid>
      <description>&lt;p&gt;A massive new study titled &lt;strong&gt;SKILLSBENCH&lt;/strong&gt; has just been released, and it’s a must-read for anyone building or using AI agents. As LLMs evolve into autonomous agents, the industry is racing to find the best way to help them handle complex, domain-specific tasks without the high cost of fine-tuning.&lt;/p&gt;

&lt;p&gt;The answer? &lt;strong&gt;Agent Skills&lt;/strong&gt;—modular packages of procedural knowledge (instructions, code templates, and heuristics) that augment agents at inference time.&lt;/p&gt;

&lt;h3&gt;
  
  
  📊 The Study at a Glance
&lt;/h3&gt;

&lt;p&gt;Researchers tested &lt;strong&gt;7 agent-model configurations&lt;/strong&gt; (including &lt;strong&gt;Claude Code&lt;/strong&gt;, &lt;strong&gt;Gemini CLI&lt;/strong&gt;, and &lt;strong&gt;Codex&lt;/strong&gt;) across &lt;strong&gt;84 tasks&lt;/strong&gt; in 11 different domains. They compared three conditions:&lt;/p&gt;

&lt;ol&gt;
&lt;li&gt;&lt;p&gt;&lt;strong&gt;No Skills&lt;/strong&gt;: The agent flies solo with just instructions.&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;&lt;strong&gt;Curated Skills&lt;/strong&gt;: Human-authored, high-quality procedural guides.&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;&lt;strong&gt;Self-Generated Skills&lt;/strong&gt;: The agent is asked to write its own guide before starting.&lt;/p&gt;&lt;/li&gt;
&lt;/ol&gt;




&lt;h3&gt;
  
  
  💡 Key Takeaways
&lt;/h3&gt;

&lt;ul&gt;
&lt;li&gt;&lt;p&gt;&lt;strong&gt;Curated Skills are a Game Changer&lt;/strong&gt;: Adding human-curated Skills boosted average pass rates by &lt;strong&gt;16.2 percentage points&lt;/strong&gt;. In specialized fields like &lt;strong&gt;Healthcare&lt;/strong&gt; and &lt;strong&gt;Manufacturing&lt;/strong&gt;, the gains were massive (up to &lt;strong&gt;+51.9pp&lt;/strong&gt;).&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;&lt;strong&gt;AI Cannot Grade Its Own Homework&lt;/strong&gt;: "Self-generated" Skills provided &lt;strong&gt;zero benefit&lt;/strong&gt; on average. Models often fail to recognize when they need specialized knowledge or produce vague, unhelpful procedures.&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;&lt;strong&gt;Smaller Models Can "Punch Up"&lt;/strong&gt;: A smaller model (like &lt;strong&gt;Haiku 4.5&lt;/strong&gt;) equipped with Skills can actually outperform a much larger model (like &lt;strong&gt;Opus 4.5&lt;/strong&gt;) that doesn't have them.&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;&lt;strong&gt;Less is More&lt;/strong&gt;: Focused Skills with only &lt;strong&gt;2-3 modules&lt;/strong&gt; outperformed massive, "comprehensive" documentation. Too much info creates "cognitive overhead" for the agent.&lt;/p&gt;&lt;/li&gt;
&lt;/ul&gt;

&lt;h3&gt;
  
  
  🏆 Top Performer
&lt;/h3&gt;

&lt;p&gt;The combination of &lt;strong&gt;Gemini CLI + Gemini 3 Flash&lt;/strong&gt; achieved the highest raw performance, hitting a &lt;strong&gt;48.7% pass rate&lt;/strong&gt; when equipped with Skills.&lt;/p&gt;

&lt;h3&gt;
  
  
  🛠 Why This Matters
&lt;/h3&gt;

&lt;p&gt;For developers and enterprise teams, this proves that &lt;strong&gt;human expertise is still the bottleneck.&lt;/strong&gt; Building a library of high-quality, modular "Skills" is currently a more effective (and cheaper) way to scale AI agent performance than just waiting for bigger models or spending a fortune on fine-tuning.&lt;/p&gt;

&lt;p&gt;Reference: &lt;a href="https://arxiv.org/abs/2602.12670" rel="noopener noreferrer"&gt;https://arxiv.org/abs/2602.12670&lt;/a&gt;&lt;/p&gt;

</description>
      <category>abotwrotethis</category>
      <category>ai</category>
      <category>agents</category>
    </item>
    <item>
      <title>Tokenization</title>
      <dc:creator>Thiago da Silva Teixeira</dc:creator>
      <pubDate>Tue, 22 Apr 2025 16:20:44 +0000</pubDate>
      <link>https://dev.to/thiagoteixeiradev/tokenization-4dja</link>
      <guid>https://dev.to/thiagoteixeiradev/tokenization-4dja</guid>
      <description>&lt;p&gt;When you build an NLP pipeline—whether for sentiment analysis, chatbots, or translation—the very first step is always the same: &lt;strong&gt;tokenization&lt;/strong&gt;. In plain words, tokenization dices raw text into smaller, consistent chunks that a model can count, index, and learn from.&lt;/p&gt;

&lt;h3&gt;
  
  
  1  What &lt;em&gt;is&lt;/em&gt; a Token?
&lt;/h3&gt;

&lt;p&gt;Think of tokens as the LEGO® bricks of language. They can be as big as a whole word or as tiny as a single character, depending on how you slice them.&lt;br&gt;
&lt;/p&gt;

&lt;div class="highlight js-code-highlight"&gt;
&lt;pre class="highlight plaintext"&gt;&lt;code&gt;Sentence: "IBM taught me tokenization."
Possible tokens: ["IBM", "taught", "me", "tokenization"]
&lt;/code&gt;&lt;/pre&gt;

&lt;/div&gt;



&lt;p&gt;Different models expect different brick sizes, so choosing the right tokenizer is strategic.&lt;/p&gt;

&lt;h3&gt;
  
  
  2  Why Tokenization Matters
&lt;/h3&gt;

&lt;ul&gt;
&lt;li&gt;Sentiment analysis: detect “good” vs “bad”.
&lt;/li&gt;
&lt;li&gt;Text generation: decide what piece comes next.
&lt;/li&gt;
&lt;li&gt;Search engines: match “running” with “run”.
&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;Without tokenization, your neural net sees text as a long, unreadable string of bytes—hardly the recipe for comprehension.&lt;/p&gt;

&lt;h3&gt;
  
  
  3  The Three Classical Approaches
&lt;/h3&gt;

&lt;div class="table-wrapper-paragraph"&gt;&lt;table&gt;
&lt;thead&gt;
&lt;tr&gt;
&lt;th&gt;Method&lt;/th&gt;
&lt;th&gt;How It Works&lt;/th&gt;
&lt;th&gt;When to Use&lt;/th&gt;
&lt;th&gt;Watch‑outs&lt;/th&gt;
&lt;/tr&gt;
&lt;/thead&gt;
&lt;tbody&gt;
&lt;tr&gt;
&lt;td&gt;&lt;strong&gt;Word‑based&lt;/strong&gt;&lt;/td&gt;
&lt;td&gt;Splits on whitespace &amp;amp; punctuation&lt;/td&gt;
&lt;td&gt;Quick prototypes, rule‑based systems&lt;/td&gt;
&lt;td&gt;Huge vocabulary, OOV* explosion&lt;/td&gt;
&lt;/tr&gt;
&lt;tr&gt;
&lt;td&gt;&lt;strong&gt;Character‑based&lt;/strong&gt;&lt;/td&gt;
&lt;td&gt;Every character is a token&lt;/td&gt;
&lt;td&gt;Morphologically rich languages, misspellings&lt;/td&gt;
&lt;td&gt;Longer sequences, less semantic punch&lt;/td&gt;
&lt;/tr&gt;
&lt;tr&gt;
&lt;td&gt;&lt;strong&gt;Sub‑word&lt;/strong&gt;&lt;/td&gt;
&lt;td&gt;Keeps common words whole, chops rare ones into pieces&lt;/td&gt;
&lt;td&gt;State‑of‑the‑art transformers (BERT, GPT‑x)&lt;/td&gt;
&lt;td&gt;More complex training &amp;amp; merges&lt;/td&gt;
&lt;/tr&gt;
&lt;/tbody&gt;
&lt;/table&gt;&lt;/div&gt;

&lt;p&gt;*OOV = out‑of‑vocabulary words&lt;/p&gt;

&lt;h3&gt;
  
  
  4  A Closer Look at Sub‑word Algorithms
&lt;/h3&gt;

&lt;ol&gt;
&lt;li&gt;
&lt;strong&gt;WordPiece&lt;/strong&gt; (BERT)
&lt;em&gt;Greedy merges&lt;/em&gt;: start with characters, repeatedly join pairs that boost likelihood.
&lt;/li&gt;
&lt;/ol&gt;

&lt;div class="highlight js-code-highlight"&gt;
&lt;pre class="highlight python"&gt;&lt;code&gt;   &lt;span class="kn"&gt;from&lt;/span&gt; &lt;span class="n"&gt;transformers&lt;/span&gt; &lt;span class="kn"&gt;import&lt;/span&gt; &lt;span class="n"&gt;BertTokenizer&lt;/span&gt;
   &lt;span class="n"&gt;tok&lt;/span&gt; &lt;span class="o"&gt;=&lt;/span&gt; &lt;span class="n"&gt;BertTokenizer&lt;/span&gt;&lt;span class="p"&gt;.&lt;/span&gt;&lt;span class="nf"&gt;from_pretrained&lt;/span&gt;&lt;span class="p"&gt;(&lt;/span&gt;&lt;span class="sh"&gt;"&lt;/span&gt;&lt;span class="s"&gt;bert-base-uncased&lt;/span&gt;&lt;span class="sh"&gt;"&lt;/span&gt;&lt;span class="p"&gt;)&lt;/span&gt;
   &lt;span class="nf"&gt;print&lt;/span&gt;&lt;span class="p"&gt;(&lt;/span&gt;&lt;span class="n"&gt;tok&lt;/span&gt;&lt;span class="p"&gt;.&lt;/span&gt;&lt;span class="nf"&gt;tokenize&lt;/span&gt;&lt;span class="p"&gt;(&lt;/span&gt;&lt;span class="sh"&gt;"&lt;/span&gt;&lt;span class="s"&gt;tokenization lovers&lt;/span&gt;&lt;span class="sh"&gt;"&lt;/span&gt;&lt;span class="p"&gt;))&lt;/span&gt;  
   &lt;span class="c1"&gt;# ['token', '##ization', 'lovers']
&lt;/span&gt;&lt;/code&gt;&lt;/pre&gt;

&lt;/div&gt;



&lt;ol&gt;
&lt;li&gt;
&lt;strong&gt;Unigram&lt;/strong&gt; (XLNet, SentencePiece)
&lt;em&gt;Vocabulary pruning&lt;/em&gt;: begin with many candidates, drop the least useful until a target size is reached.
&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;SentencePiece&lt;/strong&gt;
&lt;em&gt;Language‑agnostic&lt;/em&gt;: trains directly on raw text, treats spaces as tokens, so no pre‑tokenization needed.&lt;/li&gt;
&lt;/ol&gt;

&lt;h3&gt;
  
  
  5  Tokenization + Indexing in PyTorch
&lt;/h3&gt;



&lt;div class="highlight js-code-highlight"&gt;
&lt;pre class="highlight python"&gt;&lt;code&gt;&lt;span class="kn"&gt;from&lt;/span&gt; &lt;span class="n"&gt;torchtext.data.utils&lt;/span&gt; &lt;span class="kn"&gt;import&lt;/span&gt; &lt;span class="n"&gt;get_tokenizer&lt;/span&gt;
&lt;span class="kn"&gt;from&lt;/span&gt; &lt;span class="n"&gt;torchtext.vocab&lt;/span&gt; &lt;span class="kn"&gt;import&lt;/span&gt; &lt;span class="n"&gt;build_vocab_from_iterator&lt;/span&gt;

&lt;span class="n"&gt;sentences&lt;/span&gt; &lt;span class="o"&gt;=&lt;/span&gt; &lt;span class="p"&gt;[&lt;/span&gt;&lt;span class="sh"&gt;"&lt;/span&gt;&lt;span class="s"&gt;Life is short&lt;/span&gt;&lt;span class="sh"&gt;"&lt;/span&gt;&lt;span class="p"&gt;,&lt;/span&gt; &lt;span class="sh"&gt;"&lt;/span&gt;&lt;span class="s"&gt;Tokenization is powerful&lt;/span&gt;&lt;span class="sh"&gt;"&lt;/span&gt;&lt;span class="p"&gt;]&lt;/span&gt;
&lt;span class="n"&gt;tokenizer&lt;/span&gt; &lt;span class="o"&gt;=&lt;/span&gt; &lt;span class="nf"&gt;get_tokenizer&lt;/span&gt;&lt;span class="p"&gt;(&lt;/span&gt;&lt;span class="sh"&gt;"&lt;/span&gt;&lt;span class="s"&gt;basic_english&lt;/span&gt;&lt;span class="sh"&gt;"&lt;/span&gt;&lt;span class="p"&gt;)&lt;/span&gt;

&lt;span class="k"&gt;def&lt;/span&gt; &lt;span class="nf"&gt;yield_tokens&lt;/span&gt;&lt;span class="p"&gt;(&lt;/span&gt;&lt;span class="n"&gt;data_iter&lt;/span&gt;&lt;span class="p"&gt;):&lt;/span&gt;
    &lt;span class="k"&gt;for&lt;/span&gt; &lt;span class="n"&gt;text&lt;/span&gt; &lt;span class="ow"&gt;in&lt;/span&gt; &lt;span class="n"&gt;data_iter&lt;/span&gt;&lt;span class="p"&gt;:&lt;/span&gt;
        &lt;span class="k"&gt;yield&lt;/span&gt; &lt;span class="nf"&gt;tokenizer&lt;/span&gt;&lt;span class="p"&gt;(&lt;/span&gt;&lt;span class="n"&gt;text&lt;/span&gt;&lt;span class="p"&gt;)&lt;/span&gt;

&lt;span class="n"&gt;vocab&lt;/span&gt; &lt;span class="o"&gt;=&lt;/span&gt; &lt;span class="nf"&gt;build_vocab_from_iterator&lt;/span&gt;&lt;span class="p"&gt;(&lt;/span&gt;&lt;span class="nf"&gt;yield_tokens&lt;/span&gt;&lt;span class="p"&gt;(&lt;/span&gt;&lt;span class="n"&gt;sentences&lt;/span&gt;&lt;span class="p"&gt;),&lt;/span&gt;
                                  &lt;span class="n"&gt;specials&lt;/span&gt;&lt;span class="o"&gt;=&lt;/span&gt;&lt;span class="p"&gt;[&lt;/span&gt;&lt;span class="sh"&gt;"&lt;/span&gt;&lt;span class="s"&gt;&amp;lt;unk&amp;gt;&lt;/span&gt;&lt;span class="sh"&gt;"&lt;/span&gt;&lt;span class="p"&gt;,&lt;/span&gt; &lt;span class="sh"&gt;"&lt;/span&gt;&lt;span class="s"&gt;&amp;lt;bos&amp;gt;&lt;/span&gt;&lt;span class="sh"&gt;"&lt;/span&gt;&lt;span class="p"&gt;,&lt;/span&gt; &lt;span class="sh"&gt;"&lt;/span&gt;&lt;span class="s"&gt;&amp;lt;eos&amp;gt;&lt;/span&gt;&lt;span class="sh"&gt;"&lt;/span&gt;&lt;span class="p"&gt;,&lt;/span&gt; &lt;span class="sh"&gt;"&lt;/span&gt;&lt;span class="s"&gt;&amp;lt;pad&amp;gt;&lt;/span&gt;&lt;span class="sh"&gt;"&lt;/span&gt;&lt;span class="p"&gt;])&lt;/span&gt;
&lt;span class="n"&gt;vocab&lt;/span&gt;&lt;span class="p"&gt;.&lt;/span&gt;&lt;span class="nf"&gt;set_default_index&lt;/span&gt;&lt;span class="p"&gt;(&lt;/span&gt;&lt;span class="n"&gt;vocab&lt;/span&gt;&lt;span class="p"&gt;[&lt;/span&gt;&lt;span class="sh"&gt;"&lt;/span&gt;&lt;span class="s"&gt;&amp;lt;unk&amp;gt;&lt;/span&gt;&lt;span class="sh"&gt;"&lt;/span&gt;&lt;span class="p"&gt;])&lt;/span&gt;

&lt;span class="n"&gt;tokens&lt;/span&gt; &lt;span class="o"&gt;=&lt;/span&gt; &lt;span class="nf"&gt;tokenizer&lt;/span&gt;&lt;span class="p"&gt;(&lt;/span&gt;&lt;span class="n"&gt;sentences&lt;/span&gt;&lt;span class="p"&gt;[&lt;/span&gt;&lt;span class="mi"&gt;0&lt;/span&gt;&lt;span class="p"&gt;])&lt;/span&gt;          &lt;span class="c1"&gt;# ['life', 'is', 'short']
&lt;/span&gt;&lt;span class="n"&gt;indices&lt;/span&gt; &lt;span class="o"&gt;=&lt;/span&gt; &lt;span class="nf"&gt;vocab&lt;/span&gt;&lt;span class="p"&gt;(&lt;/span&gt;&lt;span class="n"&gt;tokens&lt;/span&gt;&lt;span class="p"&gt;)&lt;/span&gt;                   &lt;span class="c1"&gt;# [5, 6, 7]  (example output)
&lt;/span&gt;
&lt;span class="c1"&gt;# Add special tokens + padding
&lt;/span&gt;&lt;span class="n"&gt;max_len&lt;/span&gt; &lt;span class="o"&gt;=&lt;/span&gt; &lt;span class="mi"&gt;6&lt;/span&gt;
&lt;span class="n"&gt;padded&lt;/span&gt; &lt;span class="o"&gt;=&lt;/span&gt; &lt;span class="p"&gt;[&lt;/span&gt;&lt;span class="sh"&gt;"&lt;/span&gt;&lt;span class="s"&gt;&amp;lt;bos&amp;gt;&lt;/span&gt;&lt;span class="sh"&gt;"&lt;/span&gt;&lt;span class="p"&gt;]&lt;/span&gt; &lt;span class="o"&gt;+&lt;/span&gt; &lt;span class="n"&gt;tokens&lt;/span&gt; &lt;span class="o"&gt;+&lt;/span&gt; &lt;span class="p"&gt;[&lt;/span&gt;&lt;span class="sh"&gt;"&lt;/span&gt;&lt;span class="s"&gt;&amp;lt;eos&amp;gt;&lt;/span&gt;&lt;span class="sh"&gt;"&lt;/span&gt;&lt;span class="p"&gt;]&lt;/span&gt;
&lt;span class="n"&gt;padded&lt;/span&gt; &lt;span class="o"&gt;+=&lt;/span&gt; &lt;span class="p"&gt;[&lt;/span&gt;&lt;span class="sh"&gt;"&lt;/span&gt;&lt;span class="s"&gt;&amp;lt;pad&amp;gt;&lt;/span&gt;&lt;span class="sh"&gt;"&lt;/span&gt;&lt;span class="p"&gt;]&lt;/span&gt; &lt;span class="o"&gt;*&lt;/span&gt; &lt;span class="p"&gt;(&lt;/span&gt;&lt;span class="n"&gt;max_len&lt;/span&gt; &lt;span class="o"&gt;-&lt;/span&gt; &lt;span class="nf"&gt;len&lt;/span&gt;&lt;span class="p"&gt;(&lt;/span&gt;&lt;span class="n"&gt;padded&lt;/span&gt;&lt;span class="p"&gt;))&lt;/span&gt;
&lt;/code&gt;&lt;/pre&gt;

&lt;/div&gt;



&lt;p&gt;&lt;em&gt;Why it matters&lt;/em&gt;: Models operate on integers, not strings. &lt;code&gt;torchtext&lt;/code&gt; lets you jump from raw text to GPU‑ready tensors in three lines.&lt;/p&gt;

&lt;h3&gt;
  
  
  6  Special Tokens Cheat‑Sheet
&lt;/h3&gt;

&lt;div class="table-wrapper-paragraph"&gt;&lt;table&gt;
&lt;thead&gt;
&lt;tr&gt;
&lt;th&gt;Token&lt;/th&gt;
&lt;th&gt;Purpose&lt;/th&gt;
&lt;/tr&gt;
&lt;/thead&gt;
&lt;tbody&gt;
&lt;tr&gt;
&lt;td&gt;&lt;code&gt;&amp;lt;bos&amp;gt;&lt;/code&gt;&lt;/td&gt;
&lt;td&gt;Beginning of sentence&lt;/td&gt;
&lt;/tr&gt;
&lt;tr&gt;
&lt;td&gt;&lt;code&gt;&amp;lt;eos&amp;gt;&lt;/code&gt;&lt;/td&gt;
&lt;td&gt;End of sentence&lt;/td&gt;
&lt;/tr&gt;
&lt;tr&gt;
&lt;td&gt;&lt;code&gt;&amp;lt;pad&amp;gt;&lt;/code&gt;&lt;/td&gt;
&lt;td&gt;Sequence padding&lt;/td&gt;
&lt;/tr&gt;
&lt;tr&gt;
&lt;td&gt;&lt;code&gt;&amp;lt;unk&amp;gt;&lt;/code&gt;&lt;/td&gt;
&lt;td&gt;Unknown / rare word&lt;/td&gt;
&lt;/tr&gt;
&lt;/tbody&gt;
&lt;/table&gt;&lt;/div&gt;

&lt;p&gt;Adding them makes batching cleaner and generation deterministic.&lt;/p&gt;

&lt;h3&gt;
  
  
  7  Key Takeaways
&lt;/h3&gt;

&lt;ul&gt;
&lt;li&gt;
&lt;strong&gt;Tokenization is non‑negotiable&lt;/strong&gt;—mis‑tokenize and your downstream model will stumble.
&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;Choose by trade‑off&lt;/strong&gt;: word‑level (semantic clarity) vs character‑level (tiny vocab) vs sub‑word (best of both, extra complexity).
&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;Modern transformers ♥ sub‑word&lt;/strong&gt; algorithms such as WordPiece, Unigram, and SentencePiece.
&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;Indexing turns tokens into numbers&lt;/strong&gt;; libraries like &lt;code&gt;torchtext&lt;/code&gt;, spaCy, and &lt;code&gt;transformers&lt;/code&gt; automate the grunt work.
&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;Special tokens&lt;/strong&gt; (&lt;code&gt;&amp;lt;bos&amp;gt;&lt;/code&gt;, &lt;code&gt;&amp;lt;eos&amp;gt;&lt;/code&gt;, etc.) keep sequence models from losing their place.&lt;/li&gt;
&lt;/ul&gt;

&lt;h3&gt;
  
  
  Reference
&lt;/h3&gt;

&lt;p&gt;My study notes from the &lt;a href="https://www.coursera.org/learn/generative-ai-llm-architecture-data-preparation" rel="noopener noreferrer"&gt;IBM Generative AI and LLMs: Architecture and Data Preparation&lt;/a&gt; course. &lt;/p&gt;

</description>
      <category>nlp</category>
      <category>ai</category>
      <category>abotwrotethis</category>
    </item>
  </channel>
</rss>
