<?xml version="1.0" encoding="UTF-8"?>
<rss version="2.0" xmlns:atom="http://www.w3.org/2005/Atom" xmlns:dc="http://purl.org/dc/elements/1.1/">
  <channel>
    <title>DEV Community: Heiko Hotz</title>
    <description>The latest articles on DEV Community by Heiko Hotz (@marshmellow77).</description>
    <link>https://dev.to/marshmellow77</link>
    <image>
      <url>https://media2.dev.to/dynamic/image/width=90,height=90,fit=cover,gravity=auto,format=auto/https:%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Fuser%2Fprofile_image%2F696551%2F37b2b50c-9139-4400-b145-2ecb3acd09d6.jpeg</url>
      <title>DEV Community: Heiko Hotz</title>
      <link>https://dev.to/marshmellow77</link>
    </image>
    <atom:link rel="self" type="application/rss+xml" href="https://dev.to/feed/marshmellow77"/>
    <language>en</language>
    <item>
      <title>NLP@AWS Newsletter 02/2022</title>
      <dc:creator>Heiko Hotz</dc:creator>
      <pubDate>Tue, 01 Feb 2022 05:20:10 +0000</pubDate>
      <link>https://dev.to/aws/nlpaws-newsletter-022022-2onf</link>
      <guid>https://dev.to/aws/nlpaws-newsletter-022022-2onf</guid>
      <description>&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2F8f2s37fraqwmzit989yr.PNG" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2F8f2s37fraqwmzit989yr.PNG" alt=" " width="800" height="426"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;Hello world. This is the monthly AWS Natural Language Processing(NLP) newsletter covering everything related to NLP at AWS. Feel free to leave comments &amp;amp; share it on your social network.&lt;/p&gt;

&lt;h2&gt;
  
  
  NLP@AWS Customer Success Story
&lt;/h2&gt;

&lt;p&gt;Measuring customer sentiment in call centres is a huge challenge, especially for large organisations. Accurate call transcripts can help unlock insights such as sentiment, trending issues, and agent effectiveness at resolving calls in call centres.&lt;/p&gt;

&lt;p&gt;Wix.com expanded visibility of customer conversation sentiment by using Amazon Transcribe, a speech to text service, to develop a sentiment analysis system that can effectively determine how users feel throughout an interaction with customer care agents.&lt;/p&gt;

&lt;p&gt;Learn more about this AWS customer success story in this blog post: &lt;a href="https://aws.amazon.com/blogs/machine-learning/how-wix-empowers-customer-care-with-ai-capabilities-using-amazon-transcribe/" rel="noopener noreferrer"&gt;https://aws.amazon.com/blogs/machine-learning/how-wix-empowers-customer-care-with-ai-capabilities-using-amazon-transcribe/&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Ft3689kk4tf8bxruc52dj.PNG" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Ft3689kk4tf8bxruc52dj.PNG" alt=" " width="794" height="454"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;h2&gt;
  
  
  AWS AI Language Services
&lt;/h2&gt;

&lt;p&gt;&lt;a href="https://aws.amazon.com/blogs/machine-learning/part-3-how-to-approach-conversation-design-with-amazon-lex-building-and-testing/" rel="noopener noreferrer"&gt;&lt;strong&gt;How to approach conversation design with Amazon Lex&lt;/strong&gt;&lt;/a&gt;&lt;br&gt;
In this blog post you will learn how to draft an interaction model to deliver natural conversational experiences, and how to test and tune your application.&lt;/p&gt;

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fgzuu8sjcd38042ylydmd.PNG" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fgzuu8sjcd38042ylydmd.PNG" alt=" " width="796" height="325"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;h2&gt;
  
  
  NLP on Amazon SageMaker
&lt;/h2&gt;

&lt;ul&gt;
&lt;li&gt;&lt;a href="https://aws.amazon.com/blogs/machine-learning/detect-nlp-data-drift-using-custom-amazon-sagemaker-model-monitor/" rel="noopener noreferrer"&gt;&lt;strong&gt;Detecting NLP Data Drift with SageMaker&lt;/strong&gt;&lt;/a&gt;&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;NLP models can be an extremely effective tool for extracting information from unstructured text data. When the data that is used for inference (production data) differs from the data used during model training, we encounter a phenomenon known as data drift. When data drift occurs, the model is no longer relevant to the data in production and likely performs worse than expected. It’s important to continuously monitor the inference data and compare it to the data used during training. &lt;/p&gt;

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fbipmrf2g021xeoykcmyd.PNG" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fbipmrf2g021xeoykcmyd.PNG" alt=" " width="789" height="347"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;&lt;a href="https://aws.amazon.com/blogs/machine-learning/distributed-fine-tuning-of-a-bert-large-model-for-a-question-answering-task-using-hugging-face-transformers-on-amazon-sagemaker/" rel="noopener noreferrer"&gt;&lt;strong&gt;Distributed fine-tuning of a BERT model on SageMaker&lt;/strong&gt;&lt;/a&gt;&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;Hugging Face has been working closely with SageMaker to deliver ready-to-use Deep Learning Containers (DLCs) that make training and deploying the latest Transformers models easier and faster than ever. Because features such as SageMaker Data Parallel (SMDP), SageMaker Model Parallel (SMMP), S3 pipe mode, are integrated into the container, using these drastically reduces the time for companies to create Transformers-based ML solutions such as question-answering, generating text and images, optimizing search results, and improves customer support automation, conversational interfaces, semantic search, document analyses, and many more applications.&lt;/p&gt;

&lt;p&gt;In this post, we focus on the deep integration of SageMaker distributed libraries with Hugging Face, which enables data scientists to accelerate training and fine-tuning of Transformers models from days to hours, all in SageMaker.&lt;/p&gt;

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fsg9604ko7076uoezmc3y.PNG" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fsg9604ko7076uoezmc3y.PNG" alt=" " width="793" height="364"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;h2&gt;
  
  
  NLP@AWS Community Content
&lt;/h2&gt;

&lt;ul&gt;
&lt;li&gt;&lt;a href="https://www.philschmid.de/hugginface-sagemaker-workshop" rel="noopener noreferrer"&gt;&lt;strong&gt;Enterprise-Scale NLP with Hugging Face &amp;amp; Amazon SageMaker&lt;/strong&gt;&lt;/a&gt;&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;In October and November, AWS &amp;amp; Hugging Face held a workshop series on “Enterprise-Scale NLP with Hugging Face &amp;amp; Amazon SageMaker”. This workshop series consisted out of 3 parts and covers:&lt;/p&gt;

&lt;ol&gt;
&lt;li&gt;Getting Started with Amazon SageMaker: Training your first NLP Transformer model with Hugging Face and deploying it&lt;/li&gt;
&lt;li&gt;Going Production: Deploying, Scaling &amp;amp; Monitoring Hugging Face Transformer models with Amazon SageMaker&lt;/li&gt;
&lt;li&gt;MLOps: End-to-End Hugging Face Transformers with the Hub &amp;amp; SageMaker Pipelines&lt;/li&gt;
&lt;/ol&gt;

&lt;p&gt;The workshop has been recorded and the resources have been made available on Github so you are now able to do the whole workshop series on your own to enhance your Hugging Face Transformers skills with Amazon SageMaker.&lt;/p&gt;

&lt;p&gt;Youtube Playlist: &lt;a href="https://www.youtube.com/playlist?list=PLo2EIpI_JMQtPhGR5Eo2Ab0_Vb89XfhDJ" rel="noopener noreferrer"&gt;Hugging Face SageMaker Playlist&lt;/a&gt;&lt;br&gt;
Github Repository: &lt;a href="https://github.com/philschmid/huggingface-sagemaker-workshop-series" rel="noopener noreferrer"&gt;huggingface-sagemaker-workshop-series&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2F7yzhnkw9bjlqmo3vy8ks.PNG" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2F7yzhnkw9bjlqmo3vy8ks.PNG" alt=" " width="800" height="386"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;&lt;p&gt;&lt;a href="https://huggingface.co/blog/gptj-sagemaker" rel="noopener noreferrer"&gt;&lt;strong&gt;Deploy GPT-J 6B for inference using Hugging Face Transformers and Amazon SageMaker&lt;/strong&gt;&lt;/a&gt;&lt;br&gt;
GPT-J is one of the most popular open-source alternatives to GPT-3. In this blog post, you will learn how to easily deploy GPT-J using Amazon SageMaker and the Hugging Face Inference Toolkit with a few lines of code for scalable, reliable, and secure real-time inference using a regular size GPU instance with NVIDIA T4.&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;&lt;a href="https://towardsdatascience.com/teach-an-ai-model-to-write-like-shakespeare-for-free-a9e6a307139" rel="noopener noreferrer"&gt;&lt;strong&gt;Teach an AI Model to Write like Shakespeare — For Free&lt;/strong&gt;&lt;/a&gt;&lt;br&gt;
In this tutorial you will learn how to train an NLP model to write like Shakespeare within 5 minutes using &lt;a href="https://studiolab.sagemaker.aws/" rel="noopener noreferrer"&gt;Sagemaker Studio Lab&lt;/a&gt;&lt;/p&gt;&lt;/li&gt;
&lt;/ul&gt;

&lt;h2&gt;
  
  
  Stay in touch with NLP on AWS
&lt;/h2&gt;

&lt;p&gt;Our contact: &lt;a href="https://mailto:aws-nlp@amazon.com/" rel="noopener noreferrer"&gt;aws-nlp@amazon.com&lt;/a&gt;&lt;br&gt;
Email us about (1) your awesome project about NLP on AWS, (2) let us know which post in the newsletter helped your NLP journey, (3) other things that you want us to post on the newsletter. Talk to you soon.&lt;/p&gt;

</description>
      <category>aws</category>
      <category>nlp</category>
      <category>machinelearning</category>
      <category>ai</category>
    </item>
    <item>
      <title>AWS - NLP Newsletter December 2021</title>
      <dc:creator>Heiko Hotz</dc:creator>
      <pubDate>Thu, 06 Jan 2022 11:25:04 +0000</pubDate>
      <link>https://dev.to/aws/aws-nlp-newsletter-december-2021-18o3</link>
      <guid>https://dev.to/aws/aws-nlp-newsletter-december-2021-18o3</guid>
      <description>&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2F4y5aumn2oyczsxgt28d2.PNG" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2F4y5aumn2oyczsxgt28d2.PNG" alt=" " width="800" height="448"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;Happy new year, everyone! In 2021, the global Natural Language Processing (NLP) market size reached an &lt;a href="https://www.fortunebusinessinsights.com/industry-reports/natural-language-processing-nlp-market-101933" rel="noopener noreferrer"&gt;estimated &lt;strong&gt;USD 21 Billion&lt;/strong&gt;&lt;/a&gt;(!) and it is projected to grow to &lt;strong&gt;USD 127 Billion&lt;/strong&gt; in 2028 at an estimated &lt;strong&gt;CAGR of 29.4%&lt;/strong&gt;. To put these numbers into perspective: The estimated global revenue for AWS in 2021 is around USD 62 Billion.&lt;/p&gt;

&lt;p&gt;Add in strategic initiatives like our partnership with Hugging Face and the continued efforts to improve our AI Language Services and it becomes clear that NLP is, and will be for a long time, one of the most important areas to cover within AWS.&lt;/p&gt;

&lt;p&gt;In 2021 the NLP Domain has set out to successfully guide customers in their NLP journeys. We also have created resources and mechanisms to scale the field internally. In 2022 we will continue this effort by growing the NLP Domain, supporting NLP initiatives, and creating even more resources so that we as an organisation are prepared for the exciting NLP challenges to come.&lt;/p&gt;

&lt;h2&gt;
  
  
  NLP Customer Success Stories
&lt;/h2&gt;

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Frfwts69hxln6t9vhgn0z.PNG" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Frfwts69hxln6t9vhgn0z.PNG" alt=" " width="281" height="379"&gt;&lt;/a&gt;&lt;br&gt;
&lt;strong&gt;&lt;a href="https://aws.amazon.com/blogs/startups/koo-app-connects-millions-of-voices-in-their-preferred-language-with-aws/" rel="noopener noreferrer"&gt;Koo App Connects Millions of Voices in Their Preferred Language with AWS&lt;/a&gt;&lt;/strong&gt;&lt;br&gt;
When the social media revolution began, e-commerce sites mostly catered to English speakers, which left out a huge population of would-be participants. Koo, a microblogging platform based in India, noted the lack of inclusivity and made it their mission to create an app that is accessible to the entire spectrum of languages spoken in India. Koo leverages several AWS services, such as Amazon SageMaker, Aurora, EKS, and EC2 to serve millions of users on their platform.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;&lt;a href="https://www.pfizer.com/news/press-release/press-release-detail/aws-helps-pfizer-accelerate-drug-development-and-clinical" rel="noopener noreferrer"&gt;AWS Helps Pfizer Accelerate Drug Development And Clinical Manufacturing&lt;/a&gt;&lt;/strong&gt;&lt;br&gt;
Also in December, AWS announced that it is working with Pfizer to create innovative, cloud-based solutions with the potential to improve how new medicines are developed, manufactured, and distributed for testing in clinical trials. To gain quick, secure access to the right information at the right time, Pfizer’s Pharmaceutical Sciences Small Molecules teams are working with AWS to develop a prototype system that can automatically extract, ingest, and process data from this documentation to help in the design of lab experiments. The prototype system is powered by Amazon Comprehend Medical (AWS’s HIPAA-eligible natural language processing (NLP) service to extract information from unstructured medical text accurately and quickly) and Amazon SageMaker, and uses Amazon Cognito to deliver secure user access control. &lt;/p&gt;

&lt;h2&gt;
  
  
  Updates on AWS Language Services
&lt;/h2&gt;

&lt;p&gt;&lt;strong&gt;Double Bill: AWS announces &lt;a href="https://aws.amazon.com/blogs/machine-learning/post-call-analytics-for-your-contact-center-with-amazon-language-ai-services/" rel="noopener noreferrer"&gt;post call analytics&lt;/a&gt; and &lt;a href="https://aws.amazon.com/blogs/machine-learning/live-call-analytics-for-your-contact-center-with-amazon-language-ai-services/" rel="noopener noreferrer"&gt;live call analytics&lt;/a&gt; with Amazon AI Language Services.&lt;/strong&gt;&lt;/p&gt;

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fz1ng8cj3j8cd18n7bkze.png" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fz1ng8cj3j8cd18n7bkze.png" alt=" " width="709" height="605"&gt;&lt;/a&gt;&lt;br&gt;
This is a huge step for AWS customers that need rich analytics capabilities to transcribe and extract insights from your contact centre communications at scale. Both functionalities were already available in Contact Lens for Amazon Connect, and now customers who don’t use Amazon Connect can use it in their existing contact centres.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;&lt;a href="https://aws.amazon.com/blogs/machine-learning/live-transcriptions-of-f1-races-using-amazon-transcribe/" rel="noopener noreferrer"&gt;Live transcriptions of F1 races using Amazon Transcribe&lt;/a&gt;&lt;/strong&gt;&lt;br&gt;
The Formula 1 (F1) live steaming service, F1 TV, has live automated closed captions in three different languages: English, Spanish, and French. For the 2021 season, FORMULA 1 has achieved another technological breakthrough, building a fully automated workflow to create closed captions in three languages and broadcasting to 85 territories using Amazon Transcribe.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;&lt;a href="https://aws.amazon.com/blogs/machine-learning/clinical-text-mining-using-the-amazon-comprehend-medical-new-snomed-ct-api/" rel="noopener noreferrer"&gt;Clinical text mining using the Amazon Comprehend Medical new SNOMED CT API&lt;/a&gt;&lt;/strong&gt;&lt;br&gt;
This blog post describes how to use a new feature to automatically standardize and link detected concepts to the SNOMED CT (Systematized Nomenclature of Medicine—Clinical Terms) ontology. It details how to use the new SNOMED CT API to link SNOMED CT codes to medical concepts (or entities) in natural written text that can then be used to accelerate research and clinical application building.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;&lt;a href="https://aws.amazon.com/blogs/machine-learning/announcing-support-for-extracting-data-from-identity-documents-using-amazon-textract/" rel="noopener noreferrer"&gt;Support for extracting data from identity documents using Amazon Textract&lt;/a&gt;&lt;/strong&gt;&lt;/p&gt;

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2F8inlhl6e7njd8s8k6v67.PNG" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2F8inlhl6e7njd8s8k6v67.PNG" alt=" " width="773" height="512"&gt;&lt;/a&gt;&lt;br&gt;
This blog post announces a new API to Amazon Textract called Analyze ID that will help you automatically extract information from identification documents, such as driver’s licenses and passports. Amazon Textract uses AI and ML technologies to extract information from identity documents, such as U.S. passports and driver’s licenses, without the need for templates or configuration. You can automatically extract specific information, such as date of expiry and date of birth, as well as intelligently identify and extract implied information, such as name and address.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;&lt;a href="https://aws.amazon.com/blogs/machine-learning/enrich-your-content-and-metadata-to-enhance-your-search-experience-with-custom-document-enrichment-in-amazon-kendra/" rel="noopener noreferrer"&gt;Custom document enrichment in Amazon Kendra&lt;/a&gt;&lt;/strong&gt;&lt;br&gt;
Amazon Kendra customers can now enrich document metadata and content during the document ingestion process using custom document enrichment (CDE). Organizations often have large repositories of raw documents that can be improved for search by modifying content or adding metadata before indexing. So how does CDE help? By simplifying the process of creating, modifying, or deleting document metadata and content before they’re ingested into Amazon Kendra. This can include detecting entities from text, extracting text from images, transcribing audio and video, and more by creating custom logic or using services like Amazon Comprehend, Amazon Textract, Amazon Transcribe, Amazon Rekognition, and others.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;&lt;a href="https://aws.amazon.com/blogs/machine-learning/expedite-conversation-design-with-the-automated-chatbot-designer-in-amazon-lex/" rel="noopener noreferrer"&gt;Expedite conversation design with the automated chatbot designer in Amazon Lex&lt;/a&gt;&lt;/strong&gt;&lt;br&gt;
The automated chatbot designer expands the usability of Amazon Lex to the design phase. It uses machine learning (ML) to provide an initial bot design that you can then refine and launch conversational experiences faster. With the automated chatbot designer, Amazon Lex customers and partners get an easy and intuitive way of designing chatbots and can reduce bot design time from weeks to hours.&lt;/p&gt;

&lt;h2&gt;
  
  
  NLP on Amazon SageMaker
&lt;/h2&gt;

&lt;p&gt;&lt;strong&gt;&lt;a href="https://aws.amazon.com/blogs/machine-learning/build-custom-amazon-sagemaker-pytorch-models-for-real-time-handwriting-text-recognition/" rel="noopener noreferrer"&gt;Build custom Amazon SageMaker PyTorch models for real-time handwriting text recognition&lt;/a&gt;&lt;/strong&gt;&lt;/p&gt;

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fn6cr0co0qnqwrg8cj1nc.jpg" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fn6cr0co0qnqwrg8cj1nc.jpg" alt=" " width="800" height="536"&gt;&lt;/a&gt;&lt;br&gt;
Unlike standard text recognition that can be trained on documents with typed content or synthetic datasets that are easy to generate and inexpensive to obtain, handwriting recognition (HWR) comes with many challenges. These challenges include variability in writing styles, low quality of old scanned documents, and collecting good quality labeled training datasets, which can be expensive or hard to collect. In this post, we share the processes, scripts, and best practices to develop a custom ML model in Amazon SageMaker that applies deep learning (DL) techniques based on the concept outlined in the paper &lt;a href="https://assets.amazon.science/38/fe/4c3105fb43129bf59cc0aadb5d78/gnhk-a-dataset-for-english-handwriting-in-the-wild.pdf" rel="noopener noreferrer"&gt;GNHK: A Dataset for English Handwriting in the Wild&lt;/a&gt; to transcribe text in images of handwritten passages into strings.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;&lt;a href="https://aws.amazon.com/blogs/machine-learning/achieve-35-faster-training-with-hugging-face-deep-learning-containers-on-amazon-sagemaker/" rel="noopener noreferrer"&gt;Achieve 35% faster training with Hugging Face Deep Learning Containers on Amazon SageMaker&lt;/a&gt;&lt;/strong&gt;&lt;br&gt;
This post shows how to pretrain an NLP model (ALBERT) on Amazon SageMaker by using Hugging Face Deep Learning Container (DLC) and transformers library. We also demonstrate how a SageMaker distributed data parallel (SMDDP) library can provide up to a 35% faster training time compared with PyTorch’s distributed data parallel (DDP) library.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;&lt;a href="https://docs.aws.amazon.com/sagemaker/latest/dg/training-compiler.html" rel="noopener noreferrer"&gt;Amazon SageMaker Training Compiler can accelerate training by up to 50%&lt;/a&gt;&lt;/strong&gt;&lt;br&gt;
State-of-the-art NLP models consist of complex multi-layered neural networks with billions of parameters that can take thousands of GPU hours to train. Optimizing such models on training infrastructure requires extensive knowledge of DL and systems engineering; this is challenging even for narrow use cases. SageMaker Training Compiler is a capability of SageMaker that makes these hard-to-implement optimizations to reduce training time on GPU instances. The compiler optimizes DL models to accelerate training by more efficiently using SageMaker machine learning (ML) GPU instances.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;&lt;a href="https://docs.aws.amazon.com/sagemaker/latest/dg/serverless-endpoints.html" rel="noopener noreferrer"&gt;Amazon SageMaker Serverless inference for intermittent usage patterns&lt;/a&gt;&lt;/strong&gt;&lt;/p&gt;

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2F7cvgcqd4xspvzdphbmtg.PNG" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2F7cvgcqd4xspvzdphbmtg.PNG" alt=" " width="800" height="252"&gt;&lt;/a&gt;&lt;br&gt;
Amazon SageMaker Serverless Inference is a purpose-built inference option that makes it easy for you to deploy and scale ML models. Serverless Inference is ideal for workloads which have idle periods between traffic spurts and can tolerate cold starts. Serverless endpoints automatically launch compute resources and scale them in and out depending on traffic, eliminating the need to choose instance types or manage scaling policies. This takes away the undifferentiated heavy lifting of selecting and managing servers. Serverless Inference integrates with AWS Lambda to offer you high availability, built-in fault tolerance and automatic scaling.&lt;/p&gt;

&lt;h2&gt;
  
  
  Community content
&lt;/h2&gt;

&lt;p&gt;&lt;strong&gt;&lt;a href="https://towardsdatascience.com/setting-up-a-text-summarisation-project-daae41a1aaa3" rel="noopener noreferrer"&gt;Setting up a text summarisation project on Amazon SageMaker&lt;/a&gt;&lt;/strong&gt;&lt;br&gt;
This tutorial serves as a practical guide for diving deep into text summarisation. It was born out of a customer engagement where the customer wanted to know how to go about setting up a text summarisation project. While there are many impressive demos on text summarisation out there, they are not well suited for actually experimenting with different models and hyperparameters. To do that, organisations need to set up their own experimentation pipeline. The tutorial is divided into the several steps to build up this pipeline:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;Using a no-ML “model” to establish a baseline&lt;/li&gt;
&lt;li&gt;Generating summaries with a zero-shot model&lt;/li&gt;
&lt;li&gt;Training a summarisation model&lt;/li&gt;
&lt;li&gt;Evaluating the trained model&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;&lt;strong&gt;&lt;a href="https://huggingface.co/blog/codeparrot" rel="noopener noreferrer"&gt;Building a language models from scratch&lt;/a&gt;&lt;/strong&gt;&lt;/p&gt;

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fqiziz9pgpta1554x8qvn.PNG" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fqiziz9pgpta1554x8qvn.PNG" alt=" " width="800" height="566"&gt;&lt;/a&gt;&lt;br&gt;
This blog post takes a look at what it takes to build the technology behind GitHub CoPilot, an application that provides suggestions to programmers as they code. In this step by step guide, we'll learn how to train a large GPT-2 model called CodeParrot, entirely from scratch. CodeParrot can auto-complete your Python code - give it a spin here.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;&lt;a href="https://towardsdatascience.com/a-2021-nlp-retrospective-b6f51e60026a" rel="noopener noreferrer"&gt;A 2021 NLP Retrospective&lt;/a&gt;&lt;/strong&gt;&lt;br&gt;
Much has happened in the field of Natural Language Processing (NLP) in the past year and this blog post reflects on some of the NLP highlights of 2021.&lt;/p&gt;

&lt;h2&gt;
  
  
  Stay in touch with NLP on AWS
&lt;/h2&gt;

&lt;p&gt;Our contact: &lt;a href="mailto:aws-nlp@amazon.com"&gt;aws-nlp@amazon.com&lt;/a&gt;&lt;br&gt;
Email us about (1) your awesome project about NLP on AWS, (2) let us know which post in the newsletter helped your NLP journey, (3) other things that you want us to post on the newsletter. Talk to you soon.&lt;/p&gt;

</description>
      <category>aws</category>
      <category>nlp</category>
      <category>machinelearning</category>
      <category>ai</category>
    </item>
  </channel>
</rss>
