<?xml version="1.0" encoding="UTF-8"?>
<rss version="2.0" xmlns:atom="http://www.w3.org/2005/Atom" xmlns:dc="http://purl.org/dc/elements/1.1/">
  <channel>
    <title>DEV Community: James Briggs</title>
    <description>The latest articles on DEV Community by James Briggs (@jamescalam).</description>
    <link>https://dev.to/jamescalam</link>
    <image>
      <url>https://media2.dev.to/dynamic/image/width=90,height=90,fit=cover,gravity=auto,format=auto/https:%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Fuser%2Fprofile_image%2F300902%2F6d7c3c32-1ed3-4d1f-9f9d-d2b55291a2ac.jpg</url>
      <title>DEV Community: James Briggs</title>
      <link>https://dev.to/jamescalam</link>
    </image>
    <atom:link rel="self" type="application/rss+xml" href="https://dev.to/feed/jamescalam"/>
    <language>en</language>
    <item>
      <title>OpenAI's Whisper is Insane!</title>
      <dc:creator>James Briggs</dc:creator>
      <pubDate>Fri, 28 Oct 2022 07:40:09 +0000</pubDate>
      <link>https://dev.to/jamescalam/openais-whisper-is-insane-5b7a</link>
      <guid>https://dev.to/jamescalam/openais-whisper-is-insane-5b7a</guid>
      <description>&lt;p&gt;OpenAI, the company behind GPT-3 and DALLE-2 have done it again. This time they've build a state-of-the-art transcription and translation model and it's &lt;em&gt;open source&lt;/em&gt;!&lt;/p&gt;

&lt;p&gt;Check out &lt;a href="https://www.pinecone.io/learn/openai-whisper/"&gt;this article explaining &lt;em&gt;why&lt;/em&gt; OpenAI's Whisper&lt;/a&gt; is so important, how it works, and how to apply to a real world use-case (making YT search better).&lt;/p&gt;

</description>
      <category>machinelearning</category>
      <category>deeplearning</category>
      <category>news</category>
      <category>python</category>
    </item>
    <item>
      <title>How to Learn (a lot)</title>
      <dc:creator>James Briggs</dc:creator>
      <pubDate>Wed, 12 Oct 2022 12:35:04 +0000</pubDate>
      <link>https://dev.to/jamescalam/how-to-learn-a-lot-36e0</link>
      <guid>https://dev.to/jamescalam/how-to-learn-a-lot-36e0</guid>
      <description>&lt;p&gt;&lt;iframe width="710" height="399" src="https://www.youtube.com/embed/QGdbs2S2YHk"&gt;
&lt;/iframe&gt;
&lt;/p&gt;

&lt;p&gt;The modern world rewards those that can learn and adapt quickly. Both financially and in your own personal fulfilment. Learning new skills keeps life interesting and fresh, but it isn't easy to do.&lt;/p&gt;

&lt;p&gt;Continuous learning can fry your brain very quickly. Yet, there's nothing better than expanding the scope of your world in the way that learning can do.&lt;/p&gt;

&lt;p&gt;How do we learn (a lot) optimally without hitting roadblocks like burnout?&lt;/p&gt;

</description>
      <category>leadership</category>
      <category>tutorial</category>
      <category>devjournal</category>
      <category>productivity</category>
    </item>
    <item>
      <title>Multi-modal (image+text) ML</title>
      <dc:creator>James Briggs</dc:creator>
      <pubDate>Thu, 11 Aug 2022 15:08:00 +0000</pubDate>
      <link>https://dev.to/jamescalam/multi-modal-imagetext-ml-5d1e</link>
      <guid>https://dev.to/jamescalam/multi-modal-imagetext-ml-5d1e</guid>
      <description>&lt;p&gt;&lt;iframe width="710" height="399" src="https://www.youtube.com/embed/989aKUVBfbk"&gt;
&lt;/iframe&gt;
&lt;/p&gt;

</description>
      <category>deeplearning</category>
      <category>ai</category>
      <category>machinelearning</category>
      <category>nlp</category>
    </item>
    <item>
      <title>OpenAI Question-Answering Demo</title>
      <dc:creator>James Briggs</dc:creator>
      <pubDate>Thu, 07 Jul 2022 15:14:30 +0000</pubDate>
      <link>https://dev.to/jamescalam/openai-question-answering-demo-2299</link>
      <guid>https://dev.to/jamescalam/openai-question-answering-demo-2299</guid>
      <description>&lt;p&gt;Demo app showing off the cutting edge of language AI, and how to build it&lt;/p&gt;

&lt;p&gt;&lt;iframe width="710" height="399" src="https://www.youtube.com/embed/coaaSxys5so"&gt;
&lt;/iframe&gt;
&lt;/p&gt;

</description>
      <category>machinelearning</category>
      <category>datascience</category>
      <category>webdev</category>
      <category>tutorial</category>
    </item>
    <item>
      <title>Python 3.10 is out now!</title>
      <dc:creator>James Briggs</dc:creator>
      <pubDate>Tue, 05 Oct 2021 10:43:46 +0000</pubDate>
      <link>https://dev.to/jamescalam/python-3-10-is-out-now-j3g</link>
      <guid>https://dev.to/jamescalam/python-3-10-is-out-now-j3g</guid>
      <description>&lt;p&gt;The latest release of Python is out (as of 4th October 2021), it includes several fascinating new features such as:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;&lt;p&gt;Structural Pattern Matching&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;Parenthesized Context Managers&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;&lt;em&gt;More&lt;/em&gt; typing&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;New and improved error messages&lt;/p&gt;&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;Check out &lt;a href="https://towardsdatascience.com/whats-new-in-python-3-10-a757c6c69342"&gt;this article covering all of Python 3.10's new features&lt;/a&gt;.&lt;/p&gt;

&lt;p&gt;Or this video from the article:&lt;/p&gt;

&lt;p&gt;&lt;iframe width="710" height="399" src="https://www.youtube.com/embed/5-A435hIYio"&gt;
&lt;/iframe&gt;
&lt;/p&gt;

</description>
      <category>programming</category>
      <category>python</category>
      <category>datascience</category>
      <category>machinelearning</category>
    </item>
    <item>
      <title>BERT WordPiece Tokenizer Explained</title>
      <dc:creator>James Briggs</dc:creator>
      <pubDate>Tue, 14 Sep 2021 14:16:57 +0000</pubDate>
      <link>https://dev.to/jamescalam/bert-wordpiece-tokenizer-explained-1i3o</link>
      <guid>https://dev.to/jamescalam/bert-wordpiece-tokenizer-explained-1i3o</guid>
      <description>&lt;p&gt;Building a transformer model from scratch can often be the only option for many more specific use cases. Although BERT and other transformer models have been pre-trained for many languages and domains, they do not cover everything.&lt;/p&gt;

&lt;p&gt;&lt;iframe width="710" height="399" src="https://www.youtube.com/embed/cR4qMSIvX28"&gt;
&lt;/iframe&gt;
&lt;/p&gt;

&lt;p&gt;Often, these less common use cases stand to gain the most from having someone come along and build a specific transformer model. It could be for an uncommon language or a less tech-savvy domain.&lt;/p&gt;

&lt;p&gt;BERT is the most popular transformer for a wide range of language-based machine learning — from sentiment analysis to question and answering. BERT has enabled a diverse range of innovation across many borders and industries.&lt;/p&gt;

&lt;p&gt;The first step for many in designing a new BERT model is the tokenizer. In this article, we’ll look at the WordPiece tokenizer used by BERT — and see how we can build our own from scratch.&lt;/p&gt;

&lt;p&gt;&lt;a href="https://towardsdatascience.com/how-to-build-a-wordpiece-tokenizer-for-bert-f505d97dddbb"&gt;Full walkthrough&lt;/a&gt; &lt;em&gt;or &lt;a href="https://towardsdatascience.com/how-to-build-a-wordpiece-tokenizer-for-bert-f505d97dddbb?sk=eea06e01c9faecd939e10589e9de1291"&gt;free link&lt;/a&gt; if you don't have Medium!&lt;/em&gt;&lt;/p&gt;

</description>
      <category>machinelearning</category>
      <category>datascience</category>
      <category>python</category>
      <category>ai</category>
    </item>
    <item>
      <title>Why are there so many tokenization methods in Transformers?</title>
      <dc:creator>James Briggs</dc:creator>
      <pubDate>Sat, 31 Jul 2021 13:14:32 +0000</pubDate>
      <link>https://dev.to/jamescalam/why-are-there-so-many-tokenization-methods-in-transformers-1n0l</link>
      <guid>https://dev.to/jamescalam/why-are-there-so-many-tokenization-methods-in-transformers-1n0l</guid>
      <description>&lt;p&gt;&lt;iframe width="710" height="399" src="https://www.youtube.com/embed/bWLvGGJLzF8"&gt;
&lt;/iframe&gt;
&lt;/p&gt;

&lt;p&gt;HuggingFace's transformers library is the de-facto standard for NLP - used by practitioners worldwide, it's powerful, flexible, and easy to use. It achieves this through a fairly large (and complex) code-base, which has resulted in the question:&lt;/p&gt;

&lt;p&gt;"Why are there so many tokenization methods in HuggingFace transformers?"&lt;/p&gt;

&lt;p&gt;Tokenization is the process of encoding a string of text into transformer-readable token ID integers. In this video we cover five different methods for this - do these all produce the same output, or is there a difference between them?&lt;/p&gt;

&lt;p&gt;📙 Check out the &lt;a href="https://towardsdatascience.com/why-are-there-so-many-tokenization-methods-for-transformers-a340e493b3a8"&gt;Medium article&lt;/a&gt; or if you don't have Medium membership here's a &lt;a href="https://towardsdatascience.com/why-are-there-so-many-tokenization-methods-for-transformers-a340e493b3a8?sk=4a7e8c88d331aef9103e153b5b799ff5"&gt;free access link&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;I also made a NLP with Transformers course, here's &lt;a href="https://www.udemy.com/course/nlp-with-transformers/?couponCode=MEDIUM2"&gt;70% off&lt;/a&gt; if you're interested!&lt;/p&gt;

</description>
      <category>python</category>
      <category>machinelearning</category>
      <category>deeplearning</category>
      <category>datascience</category>
    </item>
    <item>
      <title>Angular App Setup With Material - Stoic Q&amp;A #5</title>
      <dc:creator>James Briggs</dc:creator>
      <pubDate>Tue, 20 Jul 2021 14:42:27 +0000</pubDate>
      <link>https://dev.to/jamescalam/angular-app-setup-with-material-stoic-q-a-5-4cgo</link>
      <guid>https://dev.to/jamescalam/angular-app-setup-with-material-stoic-q-a-5-4cgo</guid>
      <description>&lt;p&gt;&lt;iframe width="710" height="399" src="https://www.youtube.com/embed/ee71R4Cqb5o"&gt;
&lt;/iframe&gt;
&lt;/p&gt;

&lt;p&gt;Video #5 in a Stoic Q&amp;amp;A ML app build I'm working through - here I'm setting up the Angular project with Google's Angular Material design.&lt;/p&gt;

</description>
      <category>deeplearning</category>
      <category>machinelearning</category>
      <category>webdev</category>
      <category>angular</category>
    </item>
    <item>
      <title>Building a MLM Training Input Pipeline</title>
      <dc:creator>James Briggs</dc:creator>
      <pubDate>Mon, 05 Jul 2021 16:50:03 +0000</pubDate>
      <link>https://dev.to/jamescalam/building-a-mlm-training-input-pipeline-3d3o</link>
      <guid>https://dev.to/jamescalam/building-a-mlm-training-input-pipeline-3d3o</guid>
      <description>&lt;p&gt;&lt;iframe width="710" height="399" src="https://www.youtube.com/embed/heTYbpr9mD8"&gt;
&lt;/iframe&gt;
&lt;/p&gt;

&lt;p&gt;The input pipeline of our training process is the more complex part of the entire transformer build. It consists of us taking our raw OSCAR training data, transforming it, and preparing it for Masked-Language Modeling (MLM). Finally, we load our data into a DataLoader ready for training!&lt;/p&gt;

</description>
      <category>python</category>
      <category>deeplearning</category>
      <category>machinelearning</category>
      <category>datascience</category>
    </item>
    <item>
      <title>Similarity Search: Measuring Meaning From Jaccard to Bert</title>
      <dc:creator>James Briggs</dc:creator>
      <pubDate>Wed, 30 Jun 2021 09:45:23 +0000</pubDate>
      <link>https://dev.to/jamescalam/similarity-search-measuring-meaning-from-jaccard-to-bert-5am7</link>
      <guid>https://dev.to/jamescalam/similarity-search-measuring-meaning-from-jaccard-to-bert-5am7</guid>
      <description>&lt;p&gt;&lt;a href="https://www.pinecone.io/learn/semantic-search/"&gt;Semantic Search: Measuring Meaning From Jaccard to Bert&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;I put together this article covering six of the most popular similarity search techniques. Three &lt;em&gt;traditional&lt;/em&gt;, and three &lt;em&gt;vector-based&lt;/em&gt;. Which are:&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;Traditional&lt;/strong&gt;:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;&lt;p&gt;Jaccard Similarity&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;w-shingling&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;Levenshtein distance&lt;/p&gt;&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;and &lt;strong&gt;Vector-based&lt;/strong&gt;:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;&lt;p&gt;TF-IDF&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;BM25&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;Sentence-BERT (SBERT)&lt;/p&gt;&lt;/li&gt;
&lt;/ul&gt;

</description>
      <category>machinelearning</category>
      <category>python</category>
      <category>deeplearning</category>
      <category>datascience</category>
    </item>
    <item>
      <title>70% Discount on the Transformers NLP Course</title>
      <dc:creator>James Briggs</dc:creator>
      <pubDate>Thu, 17 Jun 2021 14:55:08 +0000</pubDate>
      <link>https://dev.to/jamescalam/70-discount-on-the-transformers-nlp-course-1cec</link>
      <guid>https://dev.to/jamescalam/70-discount-on-the-transformers-nlp-course-1cec</guid>
      <description>&lt;p&gt;&lt;a href="https://www.udemy.com/course/nlp-with-transformers/?couponCode=MEDIUM"&gt;Discount here!&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;Transformer models are the de-facto standard in modern NLP. They have proven themselves as the most expressive, powerful models for language by a large margin, beating all major language-based benchmarks time and time again.&lt;/p&gt;

&lt;p&gt;In this course, we learn all you need to know to get started with building cutting-edge performance NLP applications using transformer models like Google AI's BERT, or Facebook AI's DPR.&lt;/p&gt;

&lt;p&gt;We cover several key NLP frameworks including:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;HuggingFace's Transformers&lt;/li&gt;
&lt;li&gt;TensorFlow 2&lt;/li&gt;
&lt;li&gt;PyTorch&lt;/li&gt;
&lt;li&gt;spaCy&lt;/li&gt;
&lt;li&gt;NLTK&lt;/li&gt;
&lt;li&gt;Flair&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;And learn how to apply transformers to some of the most popular NLP use-cases:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;Language classification/sentiment analysis&lt;/li&gt;
&lt;li&gt;Named entity recognition (NER)&lt;/li&gt;
&lt;li&gt;Question and Answering&lt;/li&gt;
&lt;li&gt;Similarity/comparative learning&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;Throughout each of these use-cases we work through a variety of examples to ensure that what, how, and why transformers are so important. Alongside these sections we also work through two full-size NLP projects, one for sentiment analysis of financial Reddit data, and another covering a fully-fledged open domain question-answering application.&lt;/p&gt;

&lt;p&gt;All of this is supported by several other sections that encourage us to learn how to better design, implement, and measure the performance of our models, such as:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;History of NLP and where transformers come from&lt;/li&gt;
&lt;li&gt;Common preprocessing techniques for NLP&lt;/li&gt;
&lt;li&gt;The theory behind transformers&lt;/li&gt;
&lt;li&gt;How to fine-tune transformers&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;We cover all this and more, I look forward to seeing you in the course!&lt;/p&gt;

</description>
      <category>python</category>
      <category>machinelearning</category>
      <category>datascience</category>
      <category>deeplearning</category>
    </item>
    <item>
      <title>What's New in Python 3.10?</title>
      <dc:creator>James Briggs</dc:creator>
      <pubDate>Tue, 08 Jun 2021 16:47:10 +0000</pubDate>
      <link>https://dev.to/jamescalam/what-s-new-in-python-3-10-4hh5</link>
      <guid>https://dev.to/jamescalam/what-s-new-in-python-3-10-4hh5</guid>
      <description>&lt;p&gt;The Python 3.10 release has several new features like structural pattern matching, a new typing Union operator, and parenthesized context managers!&lt;/p&gt;

&lt;p&gt;Python 3.10 development has stabilized and we can finally test out all of the new features that will be included in the final release.&lt;/p&gt;

&lt;p&gt;We'll cover some of the most interesting additions to Python - structural pattern matching, parenthesized context managers, more typing, and the new and improved error messages.&lt;/p&gt;


&lt;div class="ltag__link"&gt;
  &lt;a href="https://towardsdatascience.com/whats-new-in-python-3-10-a757c6c69342" class="ltag__link__link"&gt;
    &lt;div class="ltag__link__pic"&gt;
      &lt;img src="https://res.cloudinary.com/practicaldev/image/fetch/s--TALLs9uo--/c_limit%2Cf_auto%2Cfl_progressive%2Cq_auto%2Cw_880/https://miro.medium.com/fit/c/56/56/1%2AEaHS1Q73oTDUDxrkNKGJ8A.jpeg" alt="James Briggs"&gt;
    &lt;/div&gt;
  &lt;/a&gt;
  &lt;a href="https://towardsdatascience.com/whats-new-in-python-3-10-a757c6c69342" class="ltag__link__link"&gt;
    &lt;div class="ltag__link__content"&gt;
      &lt;h2&gt;What's New in Python 3.10? | Towards Data Science&lt;/h2&gt;
      &lt;h3&gt;James Briggs ・ &lt;time&gt;Jun 8, 2021&lt;/time&gt; ・ 
      &lt;div class="ltag__link__servicename"&gt;
        &lt;img src="https://res.cloudinary.com/practicaldev/image/fetch/s--ze5yh_2q--/c_limit%2Cf_auto%2Cfl_progressive%2Cq_auto%2Cw_880/https://dev.to/assets/medium_icon-90d5232a5da2369849f285fa499c8005e750a788fdbf34f5844d5f2201aae736.svg" alt="Medium Logo"&gt;
        towardsdatascience.com
      &lt;/div&gt;
    &lt;/h3&gt;
&lt;/div&gt;
  &lt;/a&gt;
&lt;/div&gt;


&lt;p&gt;&lt;a href="https://towardsdatascience.com/whats-new-in-python-3-10-a757c6c69342?sk=648ae12c1025a83affba4eecec0d46c6"&gt;Free link&lt;/a&gt;&lt;/p&gt;

</description>
      <category>python</category>
      <category>programming</category>
      <category>datascience</category>
      <category>machinelearning</category>
    </item>
  </channel>
</rss>
