<?xml version="1.0" encoding="UTF-8"?>
<rss version="2.0" xmlns:atom="http://www.w3.org/2005/Atom" xmlns:dc="http://purl.org/dc/elements/1.1/">
  <channel>
    <title>DEV Community: Anna Kovalenko</title>
    <description>The latest articles on DEV Community by Anna Kovalenko (@anna_kovalenko_).</description>
    <link>https://dev.to/anna_kovalenko_</link>
    <image>
      <url>https://media2.dev.to/dynamic/image/width=90,height=90,fit=cover,gravity=auto,format=auto/https:%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Fuser%2Fprofile_image%2F986302%2Fc5672369-7e7e-4b9e-a41f-0f1635acb861.jpg</url>
      <title>DEV Community: Anna Kovalenko</title>
      <link>https://dev.to/anna_kovalenko_</link>
    </image>
    <atom:link rel="self" type="application/rss+xml" href="https://dev.to/feed/anna_kovalenko_"/>
    <language>en</language>
    <item>
      <title>What is Product Localization and How to Make it Work</title>
      <dc:creator>Anna Kovalenko</dc:creator>
      <pubDate>Thu, 08 Dec 2022 15:22:48 +0000</pubDate>
      <link>https://dev.to/anna_kovalenko_/what-is-product-localization-and-how-to-make-it-work-1kfo</link>
      <guid>https://dev.to/anna_kovalenko_/what-is-product-localization-and-how-to-make-it-work-1kfo</guid>
      <description>&lt;p&gt;You may have found yourself in a situation where you wanted to order a product from abroad, and the website for ordering was in a foreign language you don’t understand. Most likely, after that you purchased a product from another website or decided not to buy anything at all. This situation occurred because the international retailer hasn’t localized their product.&lt;/p&gt;

&lt;p&gt;So what is Product Localization?&lt;br&gt;
Product localization is the process of adapting and modifying a product to satisfy the language and cultural needs of customers in the target markets. To localize their products companies can translate texts, alter their packaging or modify the content on their websites so that it is appropriate and readable for international customers.&lt;/p&gt;

&lt;p&gt;Maybe if you are a native English speaker, you have never encountered translation difficulties and it may seem to you that the product localization is not necessary, since all of the websites and social media you use are already in English. However, it is important to remember that only &lt;a href="https://preply.com/en/blog/english-language-statistics/#:~:text=Current%20research%20suggests%20that%20the,17%25%20of%20the%20world's%20population"&gt;1.35 billion&lt;/a&gt; people worldwide actually speak English — it is less than 20% of the world's population. So if you want to sell your product on the global market, you should be prepared for the fact that most of the potential customers will be unavailable and inaccessible until you localize your product.&lt;/p&gt;

&lt;p&gt;Product localization helps to attract these 80% of people who do not speak English — &lt;a href="https://csa-research.com/Blogs-Events/CSA-in-the-Media/Press-Releases/Consumers-Prefer-their-Own-Language"&gt;the survey&lt;/a&gt; found that 76% of people prefer purchasing products with information in their own language.&lt;/p&gt;

&lt;p&gt;But it is important to properly localize the products. There are many examples when the localization of the product did not go according to plan.&lt;/p&gt;

&lt;p&gt;For example, in 2005 the well-known Swedish company IKEA released a bench named &lt;a href="https://www.cracked.com/article_17686_the-8-most-ill-conceived-product-names-all-time.html#:~:text=4-,The%20IKEA%20Fartfull,-IKEA%20is%20famous"&gt;“Fartfull”&lt;/a&gt;. The original name came from the Swedish word “Fährt” in Swedish, meaning “traveling in a vehicle with wheels”. Of course, for the Swedes “Fartfull” made perfect sense but in English the localization has a “stinky” connotation. The product was eventually taken off the market, but not before people all around the world had a great laugh.&lt;/p&gt;

&lt;p&gt;So how to make Product Localization work?&lt;br&gt;
There are a couple of steps you need to take to properly localize your product:&lt;/p&gt;

&lt;ol&gt;
&lt;li&gt;Content analysis
Everyone knows that planning is important for any process. Product localization requires some planning as well. Localizers need to address a huge amount of details so a product content analysis is a crucial first step for the localization process.&lt;/li&gt;
&lt;/ol&gt;

&lt;p&gt;During this step, the product localizers need to map out the necessary details that need to be localized. Usually these details are the product information, imagery, colors, currencies, measurements, help guides and other readable descriptive content.&lt;/p&gt;

&lt;ol&gt;
&lt;li&gt;Internalization
Internationalization means building or designing the product in the way that it can support different languages from that start. Internationalization is particularly important if the product is a mobile app or a piece of software. If the localizers skip this step, they will spend much more time and effort localizing the product in the future. &lt;/li&gt;
&lt;/ol&gt;

&lt;p&gt;Internationalization involves separating the source code from all the product elements that need to be localized. To internationalize the product localizers use text encoding — usually UTF-8 — add support for right-to-left languages and choose a framework for exporting hard-coded string identifiers to allow for easy translation in the future.&lt;/p&gt;

&lt;ol&gt;
&lt;li&gt;Adaptation of software, websites, marketing collateral, and documentation&lt;/li&gt;
&lt;/ol&gt;

&lt;ul&gt;
&lt;li&gt;Software localization
Software localization is considered the most important step because if the localizer doesn’t localize the actual software product well, the other elements of the localization process won’t matter at all.&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;During the software the localizer needs to adapt UI elements, splash screens, in-app tooltips, and other in-product copy in a way that makes sense for the target market. Also, during software localization, measure, date formats, and currencies need to be changed to match the local standards.&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;&lt;p&gt;Website localization&lt;br&gt;
The website is the first thing the customer sees, so it is important to properly localize all the web pages. As well as software localization website localization includes translating all text content into the target language, as well as adapting images, videos, and other visual elements.&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;Marketing localization&lt;br&gt;
Marketing localization is a bit more difficult than software and website ones. The localizer cannot just translate the marketing materials into the other language — they need to adapt marketing messages to resonate with the target market on a cultural level.&lt;/p&gt;&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;Marketing localization involves the development of new slogans and taglines and creation of new ad campaigns that take characteristics of the new audience into account.&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;Documentation localization
Documentation localization concerns things like user manuals, how-to videos, FAQs, and so on. Those documents are important because with them customers can figure out how to solve most of the potential problems on their own. Localization of these documents will make it less likely that a foreign user will seek help from the customer support service, where they will not always be able to help a user because of the language barrier.&lt;/li&gt;
&lt;/ul&gt;

&lt;ol&gt;
&lt;li&gt;Feedback
When localizers complete localization, it is important that their work is reviewed by people who are familiar with the language and culture of the target market the companies localize the product for. Main goal of this step is to be sure that the message of the original content has been translated or adapted correctly and will fit the new target market.&lt;/li&gt;
&lt;/ol&gt;

&lt;p&gt;Is there any technology that can help with Product localization?&lt;br&gt;
Doing the localization on your own can be tiresome and, quite frankly, unproductive. It will be easier to use various tools to speed up and improve the localization process.&lt;/p&gt;

&lt;p&gt;According to &lt;a href="https://www.nimdzi.com/language-technology-atlas"&gt;the report&lt;/a&gt; by Nimdzi Insights found that the current language technology landscape includes more than 700 solutions. There are a couple of them you can use:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;&lt;p&gt;Translation memory &lt;br&gt;
It’s a database of previously translated content that allows translators to reuse old translations. Translation memory databases will save you time and money and also make your translations more consistent and coherent.&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;Term base &lt;br&gt;
It’s bases that keep all of the terminology used in a project in one place so that it can be reused in future projects and also allow you to search for brand-specific or project-specific terminology. Term bases can help with tracking changes in different terminology and ensure that the translation of the terms is right and consistent.&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;Content integration &lt;br&gt;
It’s a feature that can adapt to agile processes. For localization to be agile, content must flow freely between different systems. In other words, your localization tool should be able to integrate with the rest of your tech stack, including your content management system (CMS), customer relationship management (CRM) system, ecommerce platform and so on. With content integration content can be translated and published without any manual intervention, quickly and without errors.&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;Machine translation &lt;br&gt;
It is obvious that Machine translation speeds up the translation process by automatically translating content. For product localization it is important to choose a tool that offers best-of-breed, AI-driven Machine translation capabilities because you will be translating large amounts of content.&lt;/p&gt;&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;Now you’re ready to implement Product Localization strategy in your business. Good luck!&lt;/p&gt;

</description>
    </item>
    <item>
      <title>CLIP: The link between Text and an Image</title>
      <dc:creator>Anna Kovalenko</dc:creator>
      <pubDate>Wed, 07 Dec 2022 23:38:19 +0000</pubDate>
      <link>https://dev.to/anna_kovalenko_/clip-the-link-between-text-and-an-image-4o31</link>
      <guid>https://dev.to/anna_kovalenko_/clip-the-link-between-text-and-an-image-4o31</guid>
      <description>&lt;p&gt;Maybe you’ve heard about OpenAI’s &lt;a href="https://openai.com/blog/clip/"&gt;CLIP model&lt;/a&gt; that was released last year, on January 5, 2021. Maybe you even tried to read an article on the OpenAI’s website or &lt;a href="https://arxiv.org/abs/2103.00020"&gt;the paper&lt;/a&gt; where the development of the CLIP model is described. Maybe you remember that the model is somehow related to deep learning, text-to-image generation and other cutting-edge technologies from the machine learning world. And maybe after all of this you want to understand CLIP on a deeper level. So let’s unpack the CLIP model.&lt;/p&gt;

&lt;p&gt;What is CLIP?&lt;br&gt;
Put simply, CLIP or Contrastive Language-Image Pre-training is a multimodal model that combines knowledge of image description and concepts written in English with semantic knowledge of images.&lt;/p&gt;

&lt;p&gt;The CLIP model is trained on 400 000 000 image and text pairs. Those pairs were a picture and its description written in English, so the CLIP model was trained using the data of 400 000 000 matched up pictures and their descriptions — it seems like a lot, but it is clear that it’s still not all the images that exist on the Internet and therefore pictures and descriptions of not all things existing in the world. Nevertheless, this training gave the CLIP model the ability to describe any picture possible. So if you input an image into the CLIP model, it will return the most suitable description, caption or summary of that image.&lt;/p&gt;

&lt;p&gt;It is called “zero-shot learning”. Usually most of the machine learning models learn one specific task. For example, they are trained to classify houses and cars. So it is expected that those models will do well on the task that is given to them: classifying houses and cars. Generally no one expects a machine learning model trained on houses and cars to be good at detecting, for example, planes or trees. However, the CLIP model tends to perform well on tasks it isn’t directly trained to do.&lt;/p&gt;

&lt;p&gt;CLIP can easily distinguish between images of simple objects like a house or a car but also between something more complex like &lt;a href="https://paint.wtf/ranking/yBUM/QB28M6lEm5J5Ev9b0J8Q"&gt;a bumblebee that loves capitalism&lt;/a&gt; or &lt;a href="https://paint.wtf/ranking/4mcMU/IOICQrCENXTkoDKEWJAA"&gt;a lightbulb that is trying its best&lt;/a&gt;, although CLIP has never seen such things in its training data: it just has a generalized knowledge of what those English words and phrases mean and what those pixels represent. &lt;/p&gt;

&lt;p&gt;Summarizing all of the things said, the CLIP model is a vision and text model that was trained on hundreds of millions of images and their descriptions, can return the likeliest caption to a given image, has “zero-shot” capabilities using which ut can accurately predict entire classes it wasn’t trained on.&lt;/p&gt;

&lt;p&gt;How does CLIP work?&lt;br&gt;
To connect images and text to one another, you need them to be embedded. Imagine you have one house and two cars. You could represent them as a dot on a graph:&lt;/p&gt;

&lt;p&gt;&lt;a href="https://res.cloudinary.com/practicaldev/image/fetch/s--2Iin25X5--/c_limit%2Cf_auto%2Cfl_progressive%2Cq_auto%2Cw_880/https://dev-to-uploads.s3.amazonaws.com/uploads/articles/5ie5j2esy0n5iupwrxiq.png" class="article-body-image-wrapper"&gt;&lt;img src="https://res.cloudinary.com/practicaldev/image/fetch/s--2Iin25X5--/c_limit%2Cf_auto%2Cfl_progressive%2Cq_auto%2Cw_880/https://dev-to-uploads.s3.amazonaws.com/uploads/articles/5ie5j2esy0n5iupwrxiq.png" alt="Image description" width="494" height="382"&gt;&lt;/a&gt;&lt;br&gt;
Source: &lt;a href="https://www.wolframalpha.com/input/?i=%281%2C2%29"&gt;WolframAlpha&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;And, basically, that is the process of embedding. We just embedded the information on the XY grid. You can think of embedding as a way to transport information into mathematical space, just like we did: we took information about a house and two cars and transported it into mathematical space. The same thing can be done with text and images.&lt;/p&gt;

&lt;p&gt;The CLIP model consists of two sub-models or two encoders: image encoder and text encoder&lt;/p&gt;

&lt;p&gt;During the training, the images and their text descriptions in a mini-batch are transformed to vectors of the same length by their corresponding encoders, respectively. After normalization, the image vectors are pulled closer to their matching text vectors, and pushed apart from the other text vectors. The same happens with the text vectors and their corresponding image vectors.&lt;/p&gt;

&lt;ol&gt;
&lt;li&gt;Contrastive pre-training&lt;/li&gt;
&lt;/ol&gt;

&lt;p&gt;&lt;a href="https://res.cloudinary.com/practicaldev/image/fetch/s--BRbozzmB--/c_limit%2Cf_auto%2Cfl_progressive%2Cq_auto%2Cw_880/https://dev-to-uploads.s3.amazonaws.com/uploads/articles/if8d31fmcwhboklkc9rb.png" class="article-body-image-wrapper"&gt;&lt;img src="https://res.cloudinary.com/practicaldev/image/fetch/s--BRbozzmB--/c_limit%2Cf_auto%2Cfl_progressive%2Cq_auto%2Cw_880/https://dev-to-uploads.s3.amazonaws.com/uploads/articles/if8d31fmcwhboklkc9rb.png" alt="Image description" width="880" height="569"&gt;&lt;/a&gt;&lt;br&gt;
Source: &lt;a href="https://openai.com/blog/clip/"&gt;OpenAI&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;Input images and texts are encoded, and their vector representations are used to build a similarity matrix (I*T is an inner product). During training CLIP learns that the values on the diagonal represent correct classifications, so their similarity must be higher than those in the same row and column. This approach contrasts what the CLIP model knows goes together (diagonal values) to what doesn’t go together (non-diagonal values). Each row is a classification task: given an input image (I1), the CLIP model predicts the text. Similarly, each column is a classification task: given an input text (T1), the CLIP model predicts the image. During training of CLIP, OpenAI used a very large size of mini-batches 32768 (N on the figure above).&lt;/p&gt;

&lt;p&gt;During inference one takes a set of labels, creates texts based on those labels and runs these texts through the text encoder. Text embeddings are later matched to visual representation.&lt;/p&gt;

&lt;ol&gt;
&lt;li&gt;Create dataset classifier from label text
&amp;amp;&lt;/li&gt;
&lt;li&gt;Use for zero-shot prediction
&lt;img src="https://res.cloudinary.com/practicaldev/image/fetch/s--J09Yi-Do--/c_limit%2Cf_auto%2Cfl_progressive%2Cq_auto%2Cw_880/https://dev-to-uploads.s3.amazonaws.com/uploads/articles/qf5ono2y4uyqvwucr16m.png" alt="Image description" width="880" height="647"&gt;
Source: &lt;a href="https://openai.com/blog/clip/"&gt;OpenAI&lt;/a&gt;
&lt;/li&gt;
&lt;/ol&gt;

&lt;p&gt;Classic classification training cares only about the predefined labels. If it is successful in finding houses and cars, then it doesn’t care if it is a photo or a sketch of a house or a car or a specific number of storeys in a house or a brand of a car. Whereas CLIP learns various aspects of images and points attention to details due to its training coupled with a large dataset.&lt;/p&gt;

&lt;p&gt;The CLIP model is sensitive to words used for image descriptions. Texts “a photo of a bird”, “a photo of a bird sitting near bird feeder”, or “an image of a bird” all produce different probability paired with the same image:&lt;/p&gt;

&lt;p&gt;&lt;a href="https://res.cloudinary.com/practicaldev/image/fetch/s--95vsUKJ1--/c_limit%2Cf_auto%2Cfl_progressive%2Cq_auto%2Cw_880/https://dev-to-uploads.s3.amazonaws.com/uploads/articles/mwl2od8uh5ohir9vi03m.png" class="article-body-image-wrapper"&gt;&lt;img src="https://res.cloudinary.com/practicaldev/image/fetch/s--95vsUKJ1--/c_limit%2Cf_auto%2Cfl_progressive%2Cq_auto%2Cw_880/https://dev-to-uploads.s3.amazonaws.com/uploads/articles/mwl2od8uh5ohir9vi03m.png" alt="Image description" width="880" height="563"&gt;&lt;/a&gt;&lt;br&gt;
Source: &lt;a href="https://openai.com/blog/clip/"&gt;OpenAI&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;Why does CLIP matter?&lt;br&gt;
The CLIP model has a lot of advantages and applications due to its progressive usage of computer vision and language processing. For example, CLIP is used in DALL·E 2, an AI system that can create realistic images and art from a text prompt in a second. If you want to learn more about this AI system and the CLIP’s part in it, read my article about DALL·E 2.&lt;/p&gt;

</description>
    </item>
    <item>
      <title>Machine Learning Filters: How social media incorporates Machine Learning</title>
      <dc:creator>Anna Kovalenko</dc:creator>
      <pubDate>Wed, 07 Dec 2022 23:28:03 +0000</pubDate>
      <link>https://dev.to/anna_kovalenko_/machine-learning-filters-how-social-media-incorporates-machine-learning-456e</link>
      <guid>https://dev.to/anna_kovalenko_/machine-learning-filters-how-social-media-incorporates-machine-learning-456e</guid>
      <description>&lt;p&gt;Sometimes, in the special moments of boredom, you may play around with Instagram, Snapchat or Tiktok filters just for the fun of it — putting on digital-makeup, fake freckles, sticking out your tongue to become a cute dog or swapping your face with a celebrity or your friend. And maybe sometimes, in those special moments, you may wonder how all of this works on a technical level. How all kinds of social media manage to match your face with the filters.&lt;/p&gt;

&lt;p&gt;Brief History&lt;br&gt;
The history of Machine Learning Filters has begun with different face changing startups developing the face modifying technology. Basically, those startups were simultaneously working on an application that allowed users to modify their facial features on photos and videos and during video calls. Around 2015 this application was adapted by various social media apps. &lt;/p&gt;

&lt;p&gt;Those filters that can change facial features are part of the huge and cutting edge field of computer vision. Computer vision creates a 3D space from image data with Image processing technology. With Image processing, mathematical operations are performed on each individual pixel on the provided image to transform a picture. Computer vision technology is used not only by Snapchat, but also by other social media apps like TikTok, Instagram, Asian face modifier Snow and others. &lt;/p&gt;

&lt;p&gt;To create new filters and digital face modifiers, Meta Studio and Lens Studio technologies were developed. Those studios are a tool for creating different lenses and filters. It allows you to create both Face Lenses for the front camera experiences and World Lenses for the rear camera experiences. Any person: from a 3D professional to a Photoshop amatuer — can use this tool to create their own customized filter.&lt;/p&gt;

&lt;p&gt;From an outside perspective, it sounds almost futuristic. But, surprisingly, it can be even more advanced. Last June the newer Machine Learning-based technology was launched. It's an update to Meta and Lens Studios that allows the developers to use Machine Learning algorithms to create more detailed, realistic and vivid filters.&lt;/p&gt;

&lt;p&gt;What Machine Learning Filters do&lt;br&gt;
Broadly speaking, Machine Learning filters can:&lt;/p&gt;

&lt;ol&gt;
&lt;li&gt;create and train a neural network in an external tool (these are Tensorflow or Pytorch)&lt;/li&gt;
&lt;li&gt;insert the model created in Meta and Lens Studio projects&lt;/li&gt;
&lt;li&gt;test the filter with any smartphone&lt;/li&gt;
&lt;li&gt;distribute the created Lens to social media users&lt;/li&gt;
&lt;/ol&gt;

&lt;p&gt;How Machine Learning Filters work&lt;br&gt;
Engineers, development teams, Photoshop hobbyists and other casual creators can upload custom Machine Learning models directly into social media apps. Surely, these models must be compatible with the ONNX model format or any other format suitable for specific apps but it’s a rather simple and convenient way to drop your customized Lens. &lt;/p&gt;

&lt;p&gt;Machine Learning filters manage not only Computer Vision algorithms but also others, for example Style-Transfer, and other implementable algorithms Filters Development studios came up with. They can be used as templates for the potential filters. &lt;/p&gt;

&lt;p&gt;There are 6 different templates that are available (via &lt;a href="https://docs.snap.com/lens-studio/references/guides/lens-features/machine-learning/lens-templates"&gt;Jupyter Notebook&lt;/a&gt;) and can be used for a number of cases:&lt;/p&gt;

&lt;ol&gt;
&lt;li&gt;&lt;p&gt;Classification&lt;br&gt;
That is an algorithm that is able to recognize different particular situations (for example, a person wearing glasses or a person having a textured cultural hairstyle) and consequently carry out some action.&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;Object Detection&lt;br&gt;
This algorithm can recognize the presence of an object the camera is shooting. The algorithm has a car and food detection model and can visually call it out.&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;Style Transfer&lt;br&gt;
This algorithm allows you to transform images by applying a particular graphic or art style (for example, transform the camera feed into a Van Gogh painting or a comic book style)&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;Custom &amp;amp; Ground Segmentation&lt;br&gt;
This algorithm is used for identifying areas or objects and replacing them with custom textures and objects (for example, the algorithm comes with a pizza segmentation texture and uses Material Editor to make it look sizzling).&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;Keyword Detection&lt;br&gt;
This algorithm is used for basic audio related Machine Learning models. Given a spectrogram analysis of the audio, it can return the probability of a spoken word.&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;Multi Object Detection&lt;br&gt;
This algorithm can detect 7 different classes of objects: cat, dog, potted plant, TV, car, bottle and cup.&lt;/p&gt;&lt;/li&gt;
&lt;/ol&gt;

&lt;p&gt;Those algorithms can be used by anyone as a starting point to create your own Lens. Those algorithms are flexible and can be modified so you can use them as a basis for creating your own template.&lt;/p&gt;

&lt;p&gt;What Machine Learning Filters contribute to Machine Learning and what they mean for ML engineers and developers&lt;br&gt;
Machine Learning Lenses are a promising distribution channel with great functionality. Above that, Machine Learning Filters have other advantages and props:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;&lt;p&gt;Machine Learning filters reduce barriers to entry&lt;br&gt;
As was previously stated, with Machine Learning filters, anyone can drop custom neural networks into Filters Development studios and then distribute it to millions of social media users and it's a promising and exciting possibility. With this technology it's no longer required for Machine Learning Development teams and engineers to create an entire mobile application from scratch to develop an algorithm.&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;Machine Learning filters allow easier experimentation&lt;br&gt;
Without the need for teams and engineers to work through a full app release cycle to see their on-device models in action, it will be much easier to experiment with and test brand- and product-based Lenses powered by immersive Machine Learning features.&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;Machine Learning filters democratize on-device Machine Learning&lt;br&gt;
For an overall development of Machine Learning technology it is important to democratize said technology. One of early developers and creators of Machine Learning Lenses technology, Hart Woolery noted in &lt;a href="https://heartbeat.comet.ml/lens-studio-3-0-introduces-snapml-for-adding-custom-neural-networks-directly-to-snapchat-c2c32ed95b2b"&gt;one of his interviews&lt;/a&gt; that the potential of Machine Learning filters reminds him, in some ways, of how YouTube helped democratize video creation.&lt;/p&gt;&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;Next Step&lt;br&gt;
With all of it being said, Machine Learning Filters are a great leap forward for Mobile Machine Learning technology and AI technology overall. So from now on, when you're playing around with TikTok or Snapchat filters, remember that it's a cutting edge technology and, in a way, you're incorporating Machine Learning in your everyday life just by pressing a couple of buttons and having fun. &lt;/p&gt;

</description>
    </item>
    <item>
      <title>Drawing the Undrawable: DALL·E 2</title>
      <dc:creator>Anna Kovalenko</dc:creator>
      <pubDate>Wed, 07 Dec 2022 23:18:56 +0000</pubDate>
      <link>https://dev.to/anna_kovalenko_/drawing-the-undrawable-dalle-2-c2m</link>
      <guid>https://dev.to/anna_kovalenko_/drawing-the-undrawable-dalle-2-c2m</guid>
      <description>&lt;p&gt;Imagine that you can draw anything you want: from a bowl of soup that is actually a planet in the universe drawn in a classical digital art style to the portrait of a lady with a ruby necklace drawn in a Renaissance painting style. Believe it or not, you can generate those images — all imaginable and unimaginable things — using DALL·E 2.&lt;/p&gt;

&lt;p&gt;So what exactly is DALL·E 2?&lt;br&gt;
DALL·E 2 is an AI system that can create realistic images and art from a prompt in natural language in a couple of seconds. A prompt is just a sentence that can’t be longer than 400 characters and is describing the image you want to create. A prompt can be a detailed descriptive sentence or just an emoji — AI will create an image on its basis anyway.&lt;/p&gt;

&lt;p&gt;Here are some random examples of prompts:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;&lt;a href="https://cdn.openai.com/dall-e-2/demos/text2im/astronaut/basketball/childrens_book/0.jpg"&gt;An astronaut playing the basketball with cats in space as a children’s book illustration&lt;/a&gt;&lt;/li&gt;
&lt;li&gt;&lt;a href="https://cdn.openai.com/dall-e-2/demos/text2im/teddy_bears/mad_scientists/saturday-morning-cartoon/0.jpg"&gt;Teddy bears mixing sparkling chemicals as mad scientists as a 1990s Saturday morning cartoon&lt;/a&gt;&lt;/li&gt;
&lt;li&gt;&lt;a href="https://cdn.openai.com/dall-e-2/demos/text2im/soup/monster/knitted/0.jpg"&gt;A bowl of soup that looks like a monster knitted out of wool&lt;/a&gt;&lt;/li&gt;
&lt;li&gt;
&lt;a href="https://cdn.openai.com/dall-e-2/demos/text2im/teddy_bears/groceries/ukiyo-e/0.jpg"&gt;Teddy bears&lt;/a&gt; (once again because AI makes them very cute) &lt;a href="https://cdn.openai.com/dall-e-2/demos/text2im/teddy_bears/groceries/ukiyo-e/0.jpg"&gt;shopping for groceries in the style of ukiyo-e&lt;/a&gt;
&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;DALL·E 2 has not been deliberately “taught” different art styles or different artists’ techniques as well as it has not been “taught” what a bowl of soup is or how the figurines made of plasticine look like. DALL·E studied 650 million images and their descriptions and drew its own conclusions. DALL·E’s skills and abilities are surprising even for its creators because even they, the developers of the AI system, do not know what DALL·E has learned and what it didn’t learn and how it interprets the prompts. &lt;/p&gt;

&lt;p&gt;But how does DALL·E 2 actually work?&lt;br&gt;
To be honest, at the highest level, DALL·E 2's works pretty simply:&lt;/p&gt;

&lt;p&gt;Firstly, your text prompt is input into a text encoder that is trained to map the prompt to a representation space.&lt;br&gt;
Secondly, a model called the prior maps the text encoding to a corresponding image encoding that captures the semantic information of the prompt contained in the text encoding.&lt;br&gt;
Finally, an image decoder stochastically generates an image which is a visual manifestation of this semantic information you put in your prompt.&lt;/p&gt;

&lt;p&gt;But on the more detailed level, there are some steps that DALL·E takes to link related textual and visual abstractions.&lt;/p&gt;

&lt;p&gt;Step 1. How DALL·E 2 links Textual and Visual Semantics&lt;br&gt;
If you input a prompt “a bowl of soup that is a portal to another dimension in the style of Basquiat”, DALL·E 2 will output something like this:&lt;/p&gt;

&lt;p&gt;&lt;a href="https://res.cloudinary.com/practicaldev/image/fetch/s--y_ns78D2--/c_limit%2Cf_auto%2Cfl_progressive%2Cq_auto%2Cw_880/https://dev-to-uploads.s3.amazonaws.com/uploads/articles/0hm97gex2j5hpx28okz9.png" class="article-body-image-wrapper"&gt;&lt;img src="https://res.cloudinary.com/practicaldev/image/fetch/s--y_ns78D2--/c_limit%2Cf_auto%2Cfl_progressive%2Cq_auto%2Cw_880/https://dev-to-uploads.s3.amazonaws.com/uploads/articles/0hm97gex2j5hpx28okz9.png" alt="Image description" width="880" height="870"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;But how does the AI system know how a textual concept like “a bowl of soup” is manifested in the visual space? The link between textual semantics and their visual representations in DALL·E 2 is learned by another OpenAI model called CLIP (Contrastive Language-Image Pre-training).&lt;/p&gt;

&lt;p&gt;CLIP is trained on millions of images and their captions, learning how much a given text prompts relates to an image. CLIP does not try to predict a caption given to a picture or an artwork but learns how related the caption to an image is. This approach helps CLIP to learn the link between textual and visual representations of the same abstract concept or object. To obtain this “knowledge”, DALL·E 2 model uses CLIP's ability to learn semantics from natural language.&lt;/p&gt;

&lt;p&gt;The principles of CLIP training are not really difficult:&lt;/p&gt;

&lt;p&gt;Firstly, all images and their associated captions are passed through their respective encoders, mapping all objects into an m-dimensional space.&lt;br&gt;
Next, the cosine similarity of each (image and text) pair is computed.&lt;br&gt;
The training objective is to simultaneously maximize the cosine similarity between N correct encoded image and caption pairs and minimize the cosine similarity between N2 — N incorrect encoded image and caption pairs.&lt;/p&gt;

&lt;p&gt;After the training, the CLIP model gets “frozen” and DALL·E 2 moves onto its next step.&lt;/p&gt;

&lt;p&gt;Step 2. How DALL·E 2 generates Images from Visual Semantics&lt;br&gt;
During this step DALL·E 2 learns how to reverse the image encoding mapping that the CLIP model just learned. CLIP learns a representation space in which it is easy to determine how textual and visual encodings relate to each other, but the image generation requires the AI system to learn how to exploit the representation space to create an image.&lt;/p&gt;

&lt;p&gt;DALL·E 2 uses the GLIDE model to perform the image generation. On its part, GLIDE uses the Diffusion Model to create an image. In brief, Diffusion Models learn to generate data by reversing a gradual noising process and help with the text-to-image generation process. If you want to learn more about Diffusion Models and Stable Diffusion, you can read my other article.&lt;/p&gt;

&lt;p&gt;Step 3. How DALL·E 2 maps from Textual Semantics to Corresponding Visual Semantics&lt;br&gt;
After GLIDE successfully generates an image that reflects the text prompt by image encoding, DALL·E 2 needs to actually find these encoded representations. Put simply, DALL·E needs to inject the text conditioning information from the text into the text-to-image generation process.&lt;/p&gt;

&lt;p&gt;Keep in mind that the CLIP model also learns a text encoder in addition to the image encoder. To map from the text encodings of image captions to the image encodings of their corresponding images, DALL·E 2 uses another model called the Prior. The Prior also uses the Diffusion Model in its encoding.&lt;/p&gt;

&lt;p&gt;The Prior consists of a decoder-only Transformer. It operates on an ordered sequence of:&lt;br&gt;
the tokenized text and caption&lt;br&gt;
the CLIP model’s text encodings of these tokens&lt;br&gt;
an encoding for the diffusion timestep&lt;br&gt;
the noised image passed through the CLIP model’s image encoder&lt;br&gt;
final encoding which output from the Transformer is used to predict the unnoised CLIP image encoding&lt;/p&gt;

&lt;p&gt;Step 3. How the developers of DALL·E 2 put all of it together&lt;br&gt;
There were all the components needed for DALL·E 2 and the next step is to chain all of it together for text-to-image generation:&lt;/p&gt;

&lt;p&gt;First of all, the CLIP model text encoder maps the image description into the representation space.&lt;br&gt;
Second of all, the diffusion prior maps from the CLIP text encoding to a corresponding CLIP image encoding.&lt;br&gt;
Finally, the GLIDE generation model maps from the representation space into the image space with the help of reverse-Diffusion, generating one of many possible images that conveys the semantic information within the text prompt.&lt;/p&gt;

&lt;p&gt;And it’s basically it.&lt;/p&gt;

&lt;p&gt;With this technology DALL·E 2 can create realistic, detailed, brand new images and expand existing images beyond their original canvas, edit them and do other fun stuff. If you want to learn more, you can check out &lt;a href="https://openai.com/dall-e-2/#demos"&gt;DALL·E 2 website&lt;/a&gt; and try out the AI system yourself. &lt;/p&gt;

</description>
    </item>
    <item>
      <title>StyleGAN-NADA: Blind Training and Other Wonders</title>
      <dc:creator>Anna Kovalenko</dc:creator>
      <pubDate>Wed, 07 Dec 2022 23:09:44 +0000</pubDate>
      <link>https://dev.to/anna_kovalenko_/stylegan-nada-blind-training-and-other-wonders-28lj</link>
      <guid>https://dev.to/anna_kovalenko_/stylegan-nada-blind-training-and-other-wonders-28lj</guid>
      <description>&lt;p&gt;Continuing the series of articles about the AI system DALL·E 2 and the models used in it, this time I will talk about the StyleGAN-NADA model, CLIP-Guided Domain Adaptation of Image Generators. If you want to learn more about the CLIP model, you can check out my other article!&lt;/p&gt;

&lt;p&gt;Introduction &amp;amp; Basics&lt;br&gt;
Imagine how cool it would be if you could describe a GAN with a text prompt (for example, Dog → The Joker) and get a complete generator that synthesizes images corresponding to the provided text query in any domain. Imagine how cool it would be if a generative model could be trained to produce those images without seeing any image beforehand at all. &lt;/p&gt;

&lt;p&gt;It is actually possible with the StyleGAN-NADA model. And it is really cool.&lt;/p&gt;

&lt;p&gt;&lt;a href="https://res.cloudinary.com/practicaldev/image/fetch/s--B5y8v1E_--/c_limit%2Cf_auto%2Cfl_progressive%2Cq_auto%2Cw_880/https://dev-to-uploads.s3.amazonaws.com/uploads/articles/8cncgt8ot83g2ozi93at.png" class="article-body-image-wrapper"&gt;&lt;img src="https://res.cloudinary.com/practicaldev/image/fetch/s--B5y8v1E_--/c_limit%2Cf_auto%2Cfl_progressive%2Cq_auto%2Cw_880/https://dev-to-uploads.s3.amazonaws.com/uploads/articles/8cncgt8ot83g2ozi93at.png" alt="Image description" width="880" height="234"&gt;&lt;/a&gt;&lt;br&gt;
Dog → The Joker&lt;/p&gt;

&lt;p&gt;Leveraging the semantic power of large scale CLIP (Contrastive-Language-Image-Pre-training) models, Rinon Gal and his colleagues present a text-driven method that allows shifting a generative model to new domains and does not have to collect even a single image from those domains. In other words, the StyleGAN-NADA model is trained blindly. All it takes is only a natural language text prompt and a few minutes of training, and by that the method can adapt a generator across a great number of domains characterized by diverse styles and shapes.&lt;/p&gt;

&lt;p&gt;The domains that the StyleGAN-NADA covers are very specific and fun — or maybe a little bit creepy:&lt;/p&gt;

&lt;p&gt;&lt;a href="https://res.cloudinary.com/practicaldev/image/fetch/s--jin2eVtF--/c_limit%2Cf_auto%2Cfl_progressive%2Cq_auto%2Cw_880/https://dev-to-uploads.s3.amazonaws.com/uploads/articles/137j8xtz3sgkj405awuv.png" class="article-body-image-wrapper"&gt;&lt;img src="https://res.cloudinary.com/practicaldev/image/fetch/s--jin2eVtF--/c_limit%2Cf_auto%2Cfl_progressive%2Cq_auto%2Cw_880/https://dev-to-uploads.s3.amazonaws.com/uploads/articles/137j8xtz3sgkj405awuv.png" alt="Image description" width="880" height="223"&gt;&lt;/a&gt;&lt;br&gt;
Human → Mark Zuckerberg&lt;/p&gt;

&lt;p&gt;&lt;a href="https://res.cloudinary.com/practicaldev/image/fetch/s--jP3SKvV_--/c_limit%2Cf_auto%2Cfl_progressive%2Cq_auto%2Cw_880/https://dev-to-uploads.s3.amazonaws.com/uploads/articles/fll2s5ybesmbiwu94th3.png" class="article-body-image-wrapper"&gt;&lt;img src="https://res.cloudinary.com/practicaldev/image/fetch/s--jP3SKvV_--/c_limit%2Cf_auto%2Cfl_progressive%2Cq_auto%2Cw_880/https://dev-to-uploads.s3.amazonaws.com/uploads/articles/fll2s5ybesmbiwu94th3.png" alt="Image description" width="880" height="215"&gt;&lt;/a&gt;&lt;br&gt;
Church → New York City&lt;/p&gt;

&lt;p&gt;&lt;a href="https://res.cloudinary.com/practicaldev/image/fetch/s--EwzwRonr--/c_limit%2Cf_auto%2Cfl_progressive%2Cq_auto%2Cw_880/https://dev-to-uploads.s3.amazonaws.com/uploads/articles/vifod0tskwc04qdsgzm7.png" class="article-body-image-wrapper"&gt;&lt;img src="https://res.cloudinary.com/practicaldev/image/fetch/s--EwzwRonr--/c_limit%2Cf_auto%2Cfl_progressive%2Cq_auto%2Cw_880/https://dev-to-uploads.s3.amazonaws.com/uploads/articles/vifod0tskwc04qdsgzm7.png" alt="Image description" width="880" height="230"&gt;&lt;/a&gt;&lt;br&gt;
Human → Zombie&lt;/p&gt;

&lt;p&gt;Why StyleGAN-NADA matters&lt;br&gt;
GAN training requires obtaining a multitude of images from a specific domain and usually it’s a pretty difficult task. Of course, you can leverage the information learned by Vision-Language models such as the CLIP model, yet applying these models to manipulate pretrained generators to synthesize out-of-domain images is not that easy. That’s why the authors of the StyleGAN-NADA model propose to use dual generators and an adaptive layer selection procedure to increase training stability. Unlike other models and methods, StyleGAN-NADA works in a zero-shot manner and automatically selects a subset of layers to update at each iteration.&lt;/p&gt;

&lt;p&gt;Pre-training Setup&lt;br&gt;
It all starts with a pre-trained generator and two text prompts describing a direction of change (for example, “Dog” to “The Joker”). Instead of editing a single image, the authors of StyleGAN-NADA use the signal from the CLIP model in order to train the generator itself. So there is actually no need for training data, and the process is really fast. The training takes minutes or even less.&lt;/p&gt;

&lt;p&gt;&lt;a href="https://res.cloudinary.com/practicaldev/image/fetch/s--DISF1Dvk--/c_limit%2Cf_auto%2Cfl_progressive%2Cq_auto%2Cw_880/https://dev-to-uploads.s3.amazonaws.com/uploads/articles/dggnkwrt9m8wv2aa62o4.png" class="article-body-image-wrapper"&gt;&lt;img src="https://res.cloudinary.com/practicaldev/image/fetch/s--DISF1Dvk--/c_limit%2Cf_auto%2Cfl_progressive%2Cq_auto%2Cw_880/https://dev-to-uploads.s3.amazonaws.com/uploads/articles/dggnkwrt9m8wv2aa62o4.png" alt="Image description" width="880" height="404"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;If you’re interested in the more detailed overview of the training setup, here it is:&lt;/p&gt;

&lt;p&gt;The authors of the StyleGAN-NADA model initialize two intertwined generators — G-frozen and G-train using the weights of a generator pre-trained on images from a source domain. The weights of G-frozen remain fixed throughout the whole process, while the weights of G-train are modified through optimization and an iterative layer-freezing scheme. The process shifts the domain of G-train according to a user-provided textual direction while maintaining a shared latent space.&lt;/p&gt;

&lt;p&gt;How StyleGAN-NADA works&lt;br&gt;
The main goal of the method is to shift a pre-trained generator from a given source domain to a new target domain only with the textual prompts, without using images of the target domain. Here’s the training scheme that helps to achieve that goal:&lt;/p&gt;

&lt;p&gt;Network Architecture&lt;br&gt;
The model consists of two pretrained StyleGAN2 generators with a shared mapping network and the same latent space. The goal is to change the domain of one of the paired generators with a CLIP-based loss and keep the other fixed as a reference with a layer-freezing scheme that can adapt and select which layers to update at each iteration.&lt;/p&gt;

&lt;p&gt;CLIP-based Guidance&lt;br&gt;
There are 3 different types of losses that are used:&lt;/p&gt;

&lt;p&gt;Global target loss&lt;br&gt;
The global loss is the most intuitive CLIP loss. It minimizes the CLIP-space cosine distance between the generated images and the given target text prompt and either collapses to a single image or fools CLIP by adding per-pixel noise to the images. &lt;/p&gt;

&lt;p&gt;Directional loss&lt;br&gt;
It’s a more advanced type of loss that seeks to align the direction of CLIP embeddings between images from two domains to the CLIP direction of the corresponding text queries. &lt;/p&gt;

&lt;p&gt;Embedding-norm loss &lt;br&gt;
Embedding-norm loss uses a regularized version of StyleCLIP’s latent mapper that is used to reduce the number of semantic artifacts on synthesized images.&lt;/p&gt;

&lt;p&gt;Layer-Freezing&lt;br&gt;
It happens that some layers of the generator are more important for specific domains than others, hence at each iteration a set of W+ vectors is generated — a separate style vector for each layer in the generator. A number of StyleCLIP global optimization steps are performed to measure which layers changed the most. Only those most changed layers are updated, while all other layers are frozen for that iteration.&lt;/p&gt;

&lt;p&gt;Latent-Mapper&lt;br&gt;
During the last step, it is noted that the generator does not undergo a complete transformation for some shape changes. For some domain (for example “Dog” to “The Joker”) the resulting generator can output both dogs, and the jokers and everything that lies in-between. Therefore a StyleCLIP latent mapper can be trained to map all latent codes to the dog region of the latent space.&lt;/p&gt;

&lt;p&gt;Conclusion&lt;br&gt;
So this is how StyleGAN-NADA, a CLIP-guided zero-shot method for Non-Adversarial Domain Adaptation of image generators, works. Although the StyleGAN-NADA is focused on StyleGAN, it can be applied to other generative architectures such as &lt;a href="https://github.com/boschresearch/OASIS/"&gt;OASIS&lt;/a&gt; and many others.&lt;/p&gt;

&lt;p&gt;The ability to blindly train intertwined generators leads to new cool possibilities. For example, with the StyleGAN-NADA model you can edit images in ways that are constrained almost only by your own creativity or synthesize paired cross-domain data and labeled images for downstream applications such as image-to-image translation. And it's only the beginning! The method surely will be developed in the future. Maybe this article inspired you to explore the world of textually-guided generation and abilities of the CLIP model yourself. &lt;/p&gt;

</description>
    </item>
    <item>
      <title>How Generative AI changes the Process of Creating Content</title>
      <dc:creator>Anna Kovalenko</dc:creator>
      <pubDate>Wed, 07 Dec 2022 22:58:15 +0000</pubDate>
      <link>https://dev.to/anna_kovalenko_/how-generative-ai-changes-the-process-of-creating-content-4m24</link>
      <guid>https://dev.to/anna_kovalenko_/how-generative-ai-changes-the-process-of-creating-content-4m24</guid>
      <description>&lt;p&gt;There is a lot of controversy surrounding art and — in general — content created by Artificial Intelligence. Can the content be considered art if it is made by a computer? Is the AI generated content better than a human-made one? There are a lot of questions and disagreements but there are definitely some types of content — for example, journalistic news articles or illustrations and diagrams — that can be created automatically without employing human content creators. That’s when AI can help.&lt;/p&gt;

&lt;p&gt;What is Generative AI?&lt;br&gt;
Generative AI is a technology that can create new content by utilizing existing text prompts, images, videos, audio files and so on. With the Generative AI technology, computers can detect the underlying pattern related to the input and produce similar content. Generative AI already does a lot. It produces text and images, blog posts, program codes, poetry, artwork and even &lt;a href="https://www.washingtonpost.com/technology/2022/09/02/midjourney-artificial-intelligence-state-fair-colorado/" rel="noopener noreferrer"&gt;wins fine-arts competitions&lt;/a&gt;.&lt;/p&gt;

&lt;p&gt;There are different Generative AI techniques. For example:&lt;/p&gt;

&lt;p&gt;Generative adversarial networks (GANs):&lt;br&gt;
GANs are two neural networks: a generator and a discriminator that pit against each other to find equilibrium between the two networks:&lt;br&gt;&lt;br&gt;
The generator network generates new data or content that resembles the source data&lt;br&gt;
The discriminator network differentiates between the source and the generated data in order to recognize what is more similar to the original data&lt;/p&gt;

&lt;p&gt;Transformers:&lt;br&gt;
For example, GPT-3, LaMDA, and Wu-Dao. Those transformers imitate cognitive attention and differentially measure the significance of the input data parts. They are trained to understand the language and images, learn classification tasks and generate texts and images from massive datasets&lt;/p&gt;

&lt;p&gt;Variational auto-encoders:&lt;br&gt;
Auto-encoders encode the input into compressed code while the decoder reproduces the initial information from this code. If chosen and trained correctly, this compressed representation stores the input data distribution in a much smaller dimensional representation.&lt;/p&gt;

&lt;p&gt;How can Generative AI be used?&lt;br&gt;
The Generative AI technology can be used for:&lt;/p&gt;

&lt;p&gt;Generating photographs of human faces, objects and scenes&lt;br&gt;
Generative AI can produce real looking photographs using, for example, the GAN models.&lt;/p&gt;

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2F55wuwt1qxvx2fv282981.png" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2F55wuwt1qxvx2fv282981.png" alt="Image description" width="750" height="494"&gt;&lt;/a&gt;&lt;br&gt;
If you want to learn more about the GAN models — especially the StyleGAN-NADA model — you can read &lt;a href="https://dev.to/anna_kovalenko_/stylegan-nada-blind-training-and-other-wonders-28lj"&gt;my article&lt;/a&gt; about it!&lt;/p&gt;

&lt;p&gt;Image-to-image conversion&lt;br&gt;
Generative AI can translate one image to another (for example, black and white photos to color photos; day photos to night photos; real photographs to artistic paintings in the famous artists’ styles and so on).&lt;/p&gt;

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2F9w72ucvpie0zzbrl3zbf.png" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2F9w72ucvpie0zzbrl3zbf.png" alt="Image description" width="800" height="139"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;Text-to-image Generation&lt;br&gt;
It produces realistic photographs from text prompts of simple objects like birds and flowers and other basic descriptions using, for example, the DALL·E 2 or the StackGAN systems.&lt;/p&gt;

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fernw63usaw97np05xpvu.png" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fernw63usaw97np05xpvu.png" alt="Image description" width="800" height="253"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;If you want to learn more about the DALL·E 2 model, you can check out &lt;a href="https://dev.to/anna_kovalenko_/drawing-the-undrawable-dalle-2-c2m"&gt;my article&lt;/a&gt; about it!&lt;/p&gt;

&lt;p&gt;Film Restoration&lt;br&gt;
It restores and improves old images and old movies by upscaling them to 4K and beyond. It can generate 60 frames per second instead of 23 or less, eliminate the noise and add color, for example, with the help of Topaz Gigapixel AI.&lt;/p&gt;

&lt;p&gt;You can watch an example of the AI Film Restoration &lt;a href="https://www.youtube.com/embed/yNQdVjQAQLg?autoplay=1" rel="noopener noreferrer"&gt;here&lt;/a&gt;!&lt;/p&gt;

&lt;p&gt;Semantic Segmentation&lt;br&gt;
Semantic-image-to-photo translation converts input that are semantic images or sketches to photo realistic images or photographs.&lt;/p&gt;

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2F1dx5y0w1v7gh0almb7x5.png" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2F1dx5y0w1v7gh0almb7x5.png" alt="Image description" width="800" height="260"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;Face Frontal View Generation&lt;br&gt;
It generates front-on photos from photos taken in different angles for a face verification or face identification system.&lt;/p&gt;

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fyc9abjpsp0nbgv2p8zii.png" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fyc9abjpsp0nbgv2p8zii.png" alt="Image description" width="800" height="162"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;Photos to Emojis Transformation&lt;br&gt;
It changes real photos to emojis or cartoonish faces. This technology is used in Machine Learning filters on social media like Snapchat, Instagram or TikTok.&lt;/p&gt;

&lt;p&gt;If you want to learn more about Machine Learning filters, you can read &lt;a href="https://dev.to/anna_kovalenko_/machine-learning-filters-how-social-media-incorporates-machine-learning-456e"&gt;my article&lt;/a&gt; about it!&lt;/p&gt;

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fccjg1ewjn23u8n36vp3z.png" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fccjg1ewjn23u8n36vp3z.png" alt="Image description" width="800" height="200"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;Face Aging &lt;br&gt;
It generates older versions of faces from photos. This technology is also used in Face Modification entertaining apps.&lt;/p&gt;

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fsqexdhgppgpcxomi63he.png" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fsqexdhgppgpcxomi63he.png" alt="Image description" width="800" height="488"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;Media and entertainment&lt;br&gt;
There are different ways to apply Deep fake and other generative AI technologies. For example, deep fake technology can help with localization of content (e.g. dubbing of a movie). By using face synthesis and voice cloning such as the CRISPR technology, the actors’ original voices can be matched with a lip-sync.&lt;/p&gt;

&lt;p&gt;Benefits of Generative AI&lt;br&gt;
Enhanced Identity Protection&lt;br&gt;
Generative AI that can create avatars that can conceal the real appearance of people who do not want or are not comfortable disclosing their identities for any reason while being interviewed or working online.&lt;/p&gt;

&lt;p&gt;Improved Quality of Output&lt;br&gt;
Generative AI systems can help to get high-quality images, video, audio and other content even if the original input content is distorted and far from perfect.&lt;/p&gt;

&lt;p&gt;Decreased Financial and Reputational Risks&lt;br&gt;
Generative AI tools can quickly detect malicious or at least suspicious activities and prevent all kinds of damage to a business or a creator.&lt;/p&gt;

&lt;p&gt;Challenges of Generative AI&lt;br&gt;
Security&lt;br&gt;
Generative AI can be used for scamming people.&lt;/p&gt;

&lt;p&gt;Overestimation of Capabilities&lt;br&gt;
Enormous amounts of training data is needed for the Generative AI algorithms to perform tasks. But even with that, GANs cannot create entirely new images or texts but only combine the visual and semantic information they already learned.&lt;/p&gt;

&lt;p&gt;Unexpected Outcomes&lt;br&gt;
It’s not easy to control the behavior of some Generative AI, for example, like GANs. The Generative AI models can perform unstably and generate something unexpected.&lt;/p&gt;

&lt;p&gt;Now, when you know basic information about the Generative AI technology, you can use it to benefit your business and content creation in general. Good luck!&lt;/p&gt;

</description>
      <category>watercooler</category>
    </item>
    <item>
      <title>The Science behind Generative AI Models</title>
      <dc:creator>Anna Kovalenko</dc:creator>
      <pubDate>Wed, 07 Dec 2022 22:28:02 +0000</pubDate>
      <link>https://dev.to/anna_kovalenko_/the-science-behind-generative-ai-models-18ol</link>
      <guid>https://dev.to/anna_kovalenko_/the-science-behind-generative-ai-models-18ol</guid>
      <description>&lt;p&gt;I covered the topic of Generative AI and types of AI generated content in one of my previous articles. In this one, I will explain the science behind it and answer the important question: how do Generative AI Models work? &lt;/p&gt;

&lt;p&gt;If you have read my article “How Generative AI changes the Process of Creating Content”, you know what magical things Generative AI Models can do — from generating images from text prompts to restoring the old movies — and here is how:&lt;/p&gt;

&lt;p&gt;Discriminative modeling &amp;amp; Generative modeling&lt;br&gt;
There’s two different types of modeling and to understand the science behind Generative AI it is important to learn the difference between these two.&lt;/p&gt;

&lt;p&gt;Discriminative modeling helps to classify existing data points (for example, images of elephants and mice into respective categories). It is mostly used in supervised machine learning tasks.&lt;/p&gt;

&lt;p&gt;Generative modeling helps to understand the structure of the dataset and generate similar datasets (for example, creating a photorealistic image of an elephant or a mouse). It is mostly used in unsupervised and semi-supervised machine learning tasks.&lt;/p&gt;

&lt;p&gt;&lt;a href="https://res.cloudinary.com/practicaldev/image/fetch/s--_kCSzZYh--/c_limit%2Cf_auto%2Cfl_progressive%2Cq_auto%2Cw_880/https://dev-to-uploads.s3.amazonaws.com/uploads/articles/60c937wnjf1dbedttqld.png" class="article-body-image-wrapper"&gt;&lt;img src="https://res.cloudinary.com/practicaldev/image/fetch/s--_kCSzZYh--/c_limit%2Cf_auto%2Cfl_progressive%2Cq_auto%2Cw_880/https://dev-to-uploads.s3.amazonaws.com/uploads/articles/60c937wnjf1dbedttqld.png" alt="Image description" width="880" height="442"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;Source: Altexsoft&lt;/p&gt;

&lt;p&gt;Discriminative modeling&lt;br&gt;
Like most machine learning models, discriminative algorithms classify input data with some set of features and predict a category, a label or a class to which an input data example belongs.&lt;/p&gt;

&lt;p&gt;The training data for machine learning models usually contains images — they are called samples — each of which has input features (X) and output class labels (Y). The training data for machine learning models can be, for example, photos of elephants and mice. And there is a neural net that looks at one of those photos, pays attention to the features that distinguish elephants and mice and tells whether that one specific photo is an elephant or a mouse.&lt;/p&gt;

&lt;p&gt;&lt;a href="https://res.cloudinary.com/practicaldev/image/fetch/s--PW6EcLzX--/c_limit%2Cf_auto%2Cfl_progressive%2Cq_auto%2Cw_880/https://dev-to-uploads.s3.amazonaws.com/uploads/articles/4mb3c0rg8aj7no693vyp.png" class="article-body-image-wrapper"&gt;&lt;img src="https://res.cloudinary.com/practicaldev/image/fetch/s--PW6EcLzX--/c_limit%2Cf_auto%2Cfl_progressive%2Cq_auto%2Cw_880/https://dev-to-uploads.s3.amazonaws.com/uploads/articles/4mb3c0rg8aj7no693vyp.png" alt="Image description" width="880" height="419"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;p(ylx) = probability of y given x&lt;br&gt;
Source: Altexsoft&lt;/p&gt;

&lt;p&gt;The difference between elephants and mice can be limited to just two features x (for example, “the size of the body” and “the presence of the trunk”). In this case, each feature is a dimension so it is easy to present them in a two-dimensional data space. In the illustration above, the blue dots represent elephants and the red dots represent mice. The line represents the decision boundary and shows that the discriminative model learned to separate elephants from mice based on the x features.&lt;/p&gt;

&lt;p&gt;When the discriminative model is fully trained and can tell the difference between elephants and mice, it can “recall” what an elephant or a mouse looks like from what it has already seen during training.&lt;/p&gt;

&lt;p&gt;If the discriminative model comes across an image from a different class, for example, a house, it can tell that it’s an elephant with some level of probability. In this case, the predicted output (ŷ) is compared to the expected output (y) from the training dataset. Based on the comparison, it is easy to understand how and what in a model needs to be updated to create more accurate outputs for given classes.&lt;/p&gt;

&lt;p&gt;So, to sum up, the discriminative model compresses information about the differences between elephants and mice but doesn’t try to understand what exactly elephants and mice are.&lt;/p&gt;

&lt;p&gt;Generative modeling&lt;br&gt;
Well, generative algorithms do the complete opposite. They don’t predict a label given to some features, they try to predict features given a certain label. For discriminative algorithms the relations between x and y are more important than the process of how to get x.&lt;/p&gt;

&lt;p&gt;&lt;a href="https://res.cloudinary.com/practicaldev/image/fetch/s--pgs4ZRdG--/c_limit%2Cf_auto%2Cfl_progressive%2Cq_auto%2Cw_880/https://dev-to-uploads.s3.amazonaws.com/uploads/articles/tg78z5xch1l5b4r7ovt6.png" class="article-body-image-wrapper"&gt;&lt;img src="https://res.cloudinary.com/practicaldev/image/fetch/s--pgs4ZRdG--/c_limit%2Cf_auto%2Cfl_progressive%2Cq_auto%2Cw_880/https://dev-to-uploads.s3.amazonaws.com/uploads/articles/tg78z5xch1l5b4r7ovt6.png" alt="Image description" width="880" height="430"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;p(x,y) = probability of x and y occurring together&lt;br&gt;
Source: Altexsoft&lt;/p&gt;

&lt;p&gt;Mathematically speaking, generative modeling can capture the probability of x and y occurring together. Generative algorithms learn the distribution of individual classes and features but not the boundary.&lt;/p&gt;

&lt;p&gt;So generative models help to answer the question of what is an elephant or a mouse itself. The illustration above shows that a generative model can predict both “the size of the body” and “the presence of the trunk” features of the animals and also other features from a class. Basically, it means that the generative model learns features and their relations between them to get an idea of what those animals look like in general.&lt;/p&gt;

&lt;p&gt;If the generative model knows what kinds of elephants and mice there are in general, then their differences are also known. Those generative algorithms can learn to recreate images of elephants and mice, although that was not in their training dataset.&lt;/p&gt;

&lt;p&gt;Generative modeling can help to solve a lot of problems using cutting-edge technologies like GANs and transformer-based algorithms.&lt;/p&gt;

&lt;p&gt;GANs (Generative Adversarial Networks)&lt;br&gt;
A generative adversarial network is a machine learning algorithm that puts the two neural networks that are called the generator and the discriminator against each other. The contest between two neural networks takes the form of a zero-sum game, where one agent’s gain is another agent’s loss.&lt;/p&gt;

&lt;p&gt;GANs were invented in 2014 by Jan Goodfellow and his colleagues from the University of Montreal. If you want to look at their description of the GAN architecture you can read their paper titled &lt;a href="https://arxiv.org/abs/1406.2661"&gt;“Generative Adversarial Networks”&lt;/a&gt;. &lt;/p&gt;

&lt;p&gt;&lt;a href="https://res.cloudinary.com/practicaldev/image/fetch/s--DY3tTLIf--/c_limit%2Cf_auto%2Cfl_progressive%2Cq_auto%2Cw_880/https://dev-to-uploads.s3.amazonaws.com/uploads/articles/jw5jz4fbzr9kori47q5w.png" class="article-body-image-wrapper"&gt;&lt;img src="https://res.cloudinary.com/practicaldev/image/fetch/s--DY3tTLIf--/c_limit%2Cf_auto%2Cfl_progressive%2Cq_auto%2Cw_880/https://dev-to-uploads.s3.amazonaws.com/uploads/articles/jw5jz4fbzr9kori47q5w.png" alt="Image description" width="880" height="469"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;Source: Altexsoft&lt;/p&gt;

&lt;p&gt;The GAN architecture has two sub-models in it:&lt;/p&gt;

&lt;p&gt;Generator &lt;br&gt;
Generator is a neural net that can create fake input or fake samples from a random input vector (a list of mathematical variables each of whose value is unknown)&lt;/p&gt;

&lt;p&gt;Discriminator&lt;br&gt;
It’s a neural net that can take a given sample and decide if it’s a fake sample from a generator or a real sample from the domain. So, basically, the discriminator is a binary classifier that returns probabilities — a number between 0 and 1. If the result is closer to 0, the output is more likely to be fake. And if the number is closer to 1 is more likely to be real.&lt;/p&gt;

&lt;p&gt;A generator and a discriminator are often implemented as Convolutional Neural Networks (CNNs), when working with visual content.&lt;/p&gt;

&lt;p&gt;Transformer-based models&lt;br&gt;
Transformer-based models are deep neural networks that learn context and therefore meaning by tracking relationships in sequential data, for example, like the words in the sentence. That’s why transformers are often used in Natural Language Processing tasks.&lt;/p&gt;

&lt;p&gt;Transformer-based models were firstly described in a 2017 paper &lt;a href="https://arxiv.org/abs/1706.03762"&gt;“Attention Is All You Need”&lt;/a&gt; by Ashish Vaswani and his colleagues from Google. The examples of transformers are GPT-3 and LaMDA.&lt;/p&gt;

&lt;p&gt;The transformer-based models can transform one sequence into another. Those models go through semi-supervised learning which means that they are pre-trained in an unsupervised manner using a large unlabeled dataset and then fine-tuned through supervised training to perform better.&lt;/p&gt;

&lt;p&gt;&lt;a href="https://res.cloudinary.com/practicaldev/image/fetch/s--UruMcU6m--/c_limit%2Cf_auto%2Cfl_progressive%2Cq_auto%2Cw_880/https://dev-to-uploads.s3.amazonaws.com/uploads/articles/ctbo5cr6bar1gkgvjloc.png" class="article-body-image-wrapper"&gt;&lt;img src="https://res.cloudinary.com/practicaldev/image/fetch/s--UruMcU6m--/c_limit%2Cf_auto%2Cfl_progressive%2Cq_auto%2Cw_880/https://dev-to-uploads.s3.amazonaws.com/uploads/articles/ctbo5cr6bar1gkgvjloc.png" alt="Image description" width="880" height="468"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;Source: Altexsoft&lt;/p&gt;

&lt;p&gt;A typical transformer has two parts in it:&lt;/p&gt;

&lt;p&gt;The encoder &lt;br&gt;
The encoder works on the input sequence. It extracts all features from a sequence, converts them into vectors (e.g. vectors representing the semantics and position of a word in a sentence) and then passes them to the decoder.&lt;/p&gt;

&lt;p&gt;The decoder&lt;br&gt;
The decoder works on the target output sequence. Each decoder receives the encoder layer outputs, derives context from them and generates the output sequence.&lt;/p&gt;

&lt;p&gt;Transformers work through sequence-to-sequence learning, so, basically, the transformer takes a sequence of tokens (again, for example, words in a sentence) and predicts the next word in the output sequence. It does this through iterating encoder layers.&lt;/p&gt;

&lt;p&gt;Now when you know the science behind generative AI models you can read my other article about generative content in which I describe how generative AI models can be used and what are their advantages and disadvantages. &lt;/p&gt;

</description>
    </item>
    <item>
      <title>What is Stable Diffusion and why it matters</title>
      <dc:creator>Anna Kovalenko</dc:creator>
      <pubDate>Wed, 07 Dec 2022 22:21:06 +0000</pubDate>
      <link>https://dev.to/anna_kovalenko_/what-is-stable-diffusion-and-why-it-matters-3apf</link>
      <guid>https://dev.to/anna_kovalenko_/what-is-stable-diffusion-and-why-it-matters-3apf</guid>
      <description>&lt;p&gt;What is Stable Diffusion and why it matters&lt;/p&gt;

&lt;p&gt;What is Stable Diffusion?&lt;br&gt;
Stable diffusion is a text-to-image generation machine learning model released by &lt;a href="https://stability.ai/" rel="noopener noreferrer"&gt;Stability.Ai&lt;/a&gt; on August 22, 2022. Basically, this model is capable of generating highly detailed photo-realistic images from the text descriptions.&lt;/p&gt;

&lt;p&gt;The model was released publicly and is &lt;a href="https://github.com/CompVis/stable-diffusion#stable-diffusion-v1" rel="noopener noreferrer"&gt;open source&lt;/a&gt; so you can play around with it if you want. &lt;/p&gt;

&lt;p&gt;What can Stable Diffusion do?&lt;br&gt;
For now the Stable Diffusion model allows users to:&lt;br&gt;
Convert text into brand new realistic images at 512x512 pixels in a few seconds&lt;br&gt;
Use image modification, via image-to-image translation guided by a text prompt and upscaling, to transform an existing image into a new image &lt;br&gt;
Use GFP-GAN modeling for the inpaiting, the process that helps to restore and upscale an existing image&lt;/p&gt;

&lt;p&gt;How does Stable Diffusion work?&lt;br&gt;
Stable diffusion is a form of a diffusion model (DM). Diffusion models were introduced back in 2015 and are trained with the objective of removing successive applications of Gaussian noise to training images, and can be thought of as a sequence of denoising autoencoders. There are different variants of the DMs but Stable Diffusion is powered by a variant known as Latent Diffusion or Latent Diffusion Model (LDM). &lt;/p&gt;

&lt;p&gt;Latent Diffusion is a breakthrough text-to-image synthesis technique. It was described by AI researchers at the Ludwig Maximilian University of Munich in a paper called “High-Resolution Image Synthesis with Latent Diffusion Models”. In short, Latent Diffusion is trained to transform images into a lower-dimensional latent space instead of learning to denoise image data in “pixel space”. It makes the training cost lower and the inference speed faster. &lt;/p&gt;

&lt;p&gt;How was Stable Diffusion trained?&lt;br&gt;
Stable Diffusion was trained on massive datasets collected by &lt;a href="https://laion.ai/" rel="noopener noreferrer"&gt;LAION&lt;/a&gt;, a non-profit AI open network who received funding from Stability.Ai, the owner of the Stable Diffusion model. Those datasets contained 120 million image-text pairs from the complete set which contains nearly 6 billion image-text pairs. &lt;/p&gt;

&lt;p&gt;Overall they sampled 12 million images. 47% of the total sample size came from 100 domains. Pinterest took up 8.5% of the entire dataset, followed by other sources such as WordPress.com, Blogspot, Flickr, DeviantArt, and Wikimedia Commons.&lt;/p&gt;

&lt;p&gt;Is Stable Diffusion being regulated?&lt;br&gt;
There are plenty of ethical, moral and legal issues concerning the misuse of AI devices. Compared to previous comparable AI models, Stable Diffusion permits a wider range of images to be generated. For example, users can generate pictures of real people or existing brand logos. The other expressed concern associated with Stable Diffusion is that widespread usage of image synthesis softwares may eventually lead to human artists, along with photographers, models and everyone involved into creation of the visual art, to gradually lose commercial viability against AI-based competitors. &lt;/p&gt;

&lt;p&gt;In response to these concerns, Stability.ai emphasized the importance of “ethical and legal” use of the model in its public release announcement. There are some important points:&lt;br&gt;
Users of the Stable Diffusion model are granted a “perpetual, worldwide, non-exclusive, no-charge, royalty-free, irrevocable copyright license to reproduce, prepare, publicly display, publicly perform, sublicense, and distribute” the Model, and any of its Complementary Material, such as its source code, and any Derivatives of the Model&lt;br&gt;
Users are also granted a “perpetual, worldwide, non-exclusive, no-charge, royalty-free, irrevocable” patent license to make, use, sell or offer to sell, import or otherwise transfer, the Model and any of its Complementary Material&lt;br&gt;
Users agree to use the Model or its Derivatives “in an ethical, moral and legal manner and contribute both to the community and discourse around it” and to not use the Model in a way which would cause harm to minors, defame anyone, discriminate against an individual or group or exploit the vulnerabilities of a specific group&lt;br&gt;
The Licensor does not assert any rights in the output users generate using the model&lt;/p&gt;

&lt;p&gt;Basically, the license of Stable Diffusion usage relies on users self-regulating their own actions and their ability to “do the right thing”. However, the license does not imply any punishment for non-compliance with the social agreement between Stable Diffusion and its users.&lt;/p&gt;

&lt;p&gt;Commenting on that, the CEO of Stability.Ai Emad Mostaque explained that it is “peoples' responsibility as to whether they are ethical, moral, and legal in how they operate this technology” and the freedom given to users can provide an overall net benefit, even in spite of the potential negative consequences. Also Mostaque adds that image generating AI systems before Stable Diffusion were developed closed and controlled by large corporations, whilst the availability of Stable Diffusion ends corporate control and dominance over such technologies.&lt;/p&gt;

&lt;p&gt;But is it considered art, though?&lt;br&gt;
The US Copyright Office says that it is not. In February 2022, the Office’s Review Board rejected an application for the artwork called “A Recent Entrance to Paradise” that was generated by an AI algorithm that repurposed picture to create an image seen by a synthetic dying brain.&lt;/p&gt;

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fsm18ma66n281w5vshej9.png" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fsm18ma66n281w5vshej9.png" alt="Image description" width="800" height="599"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;The Review Board stated that “human authorship is a prerequisite to copyright protection &amp;lt;...&amp;gt; but the Work [“A Recent Entrance to Paradise”] was autonomously created by artificial intelligence without any creative contribution from a human actor”.&lt;/p&gt;

&lt;p&gt;Yet AI art is relatively popular. For example, in 2018 an auction house Christie sold an IA generated picture of a blurred face titled “Portrait of Edmond Belamy” for $432,500. But many claim that the only appeal AI generated art has is a quirkiness of being made by a non-human. &lt;/p&gt;

&lt;p&gt;Proving that AI generated art is not art, the Rutgers AI lab concluded:&lt;/p&gt;

&lt;p&gt;“Still, there’s something missing in [AI’s] artistic process: The algorithm might create appealing images, but it lives in an isolated creative space that lacks social context. Human artists, on the other hand, are inspired by people, places, and politics. They create art to tell stories and make sense of the world”. &lt;/p&gt;

</description>
      <category>bitcoin</category>
      <category>cryptocurrency</category>
      <category>cybersecurity</category>
      <category>blockchain</category>
    </item>
  </channel>
</rss>
