<?xml version="1.0" encoding="UTF-8"?>
<rss version="2.0" xmlns:atom="http://www.w3.org/2005/Atom" xmlns:dc="http://purl.org/dc/elements/1.1/">
  <channel>
    <title>DEV Community: Ankita Sahoo</title>
    <description>The latest articles on DEV Community by Ankita Sahoo (@ank1tas).</description>
    <link>https://dev.to/ank1tas</link>
    <image>
      <url>https://media2.dev.to/dynamic/image/width=90,height=90,fit=cover,gravity=auto,format=auto/https:%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Fuser%2Fprofile_image%2F727417%2F63f39f4b-61b9-420c-9632-6d9bcea2dc97.jpeg</url>
      <title>DEV Community: Ankita Sahoo</title>
      <link>https://dev.to/ank1tas</link>
    </image>
    <atom:link rel="self" type="application/rss+xml" href="https://dev.to/feed/ank1tas"/>
    <language>en</language>
    <item>
      <title>Want to prevent Model Overfitting?</title>
      <dc:creator>Ankita Sahoo</dc:creator>
      <pubDate>Sat, 09 Mar 2024 18:32:47 +0000</pubDate>
      <link>https://dev.to/ank1tas/want-to-prevent-model-overfitting-2edb</link>
      <guid>https://dev.to/ank1tas/want-to-prevent-model-overfitting-2edb</guid>
      <description>&lt;p&gt;Five different techniques to prevent overfitting:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;&lt;p&gt;Early Stopping: In this method, we track the loss on the validation set during the training phase and use it to determine when to stop training such that the model is accurate but not overfitting.&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;Image Augmentation: Artificially boosting the number of images in our training set by applying random image transformations to the existing images in the training set.&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;Dropout: Removing a random selection of a fixed number of neurons in a neural network during training.&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;Increase dataset size: The more training data you feed, the less likely it is to overfit. The reason is that, as you add more data, the model cannot overfit all the samples, and is forced to generalize to make progress.&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;Regularization: Regularization is a method to constrain the model to fit our data accurately and not overfit. It can also be thought of as penalizing unnecessary complexity in our model&lt;/p&gt;&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;Did I miss anything?🤔 Let me know in the comments. Happy Learning.😊&lt;/p&gt;

</description>
      <category>machinelearning</category>
      <category>deeplearning</category>
      <category>datascience</category>
      <category>learning</category>
    </item>
    <item>
      <title>Important terminology that you need to know to learn CNN.</title>
      <dc:creator>Ankita Sahoo</dc:creator>
      <pubDate>Sun, 03 Mar 2024 12:01:37 +0000</pubDate>
      <link>https://dev.to/ank1tas/important-terminology-that-you-need-to-know-to-learn-cnn-3mn</link>
      <guid>https://dev.to/ank1tas/important-terminology-that-you-need-to-know-to-learn-cnn-3mn</guid>
      <description>&lt;ul&gt;
&lt;li&gt;&lt;p&gt;&lt;strong&gt;CNN&lt;/strong&gt;: Convolutional neural network. That is a network that has at least one convolutional layer. A typical CNN also includes other types of layers, such as pooling and dense layers.&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;&lt;strong&gt;Convolution&lt;/strong&gt;: The process of applying a kernel (filter) to an image&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;&lt;strong&gt;Kernel/filter&lt;/strong&gt;: A matrix that is smaller than the input, used to transform the input into chunks&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;&lt;strong&gt;Padding&lt;/strong&gt;: Adding pixels of some value, usually 0, around the input image&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;&lt;strong&gt;Pooling&lt;/strong&gt;:  The process of reducing the size of an image through downsampling. There are several types of pooling layers. For example, average pooling converts many values into one by taking the average. However, max-pooling is the most common.&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;&lt;strong&gt;Max pooling&lt;/strong&gt;: A process in which many values are converted into a single value by taking the maximum value from among them.&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;&lt;strong&gt;Stride&lt;/strong&gt;: the number of pixels to slide the kernel (filter) across the image.&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;&lt;strong&gt;Downsampling&lt;/strong&gt;: The act of reducing the size of an image.&lt;/p&gt;&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;Let me know your feedback in the comments🙂. &lt;br&gt;
Happy Learning.😊&lt;/p&gt;

</description>
      <category>machinelearning</category>
      <category>deeplearning</category>
      <category>beginners</category>
      <category>cnn</category>
    </item>
    <item>
      <title>NLP Application (Real-world implementation of Transformer model)</title>
      <dc:creator>Ankita Sahoo</dc:creator>
      <pubDate>Sat, 30 Sep 2023 16:57:40 +0000</pubDate>
      <link>https://dev.to/ank1tas/nlp-application-real-world-implementation-of-transformer-model-2g86</link>
      <guid>https://dev.to/ank1tas/nlp-application-real-world-implementation-of-transformer-model-2g86</guid>
      <description>&lt;p&gt;Natural language processing (NLP) is a field of artificial intelligence that deals with the interaction between computers and humans using natural language. The importance of NLP lies in its ability to transform how humans and computers interact, enabling more intuitive and human-like communication between them.&lt;br&gt;
This has numerous practical applications in information retrieval, sentiment analysis, machine translation, and question-answering, among others. NLP has the potential&lt;br&gt;
to revolutionize many industries, such as healthcare, education, and customer service, by enabling more effective and efficient communication and information management.&lt;br&gt;
As such, NLP has become an important area of research and development, with significant investment being made in its advancement.&lt;/p&gt;

&lt;h2&gt;
  
  
  &lt;u&gt;The Transformer&lt;/u&gt;:
&lt;/h2&gt;

&lt;p&gt;&lt;a href="https://arxiv.org/abs/1706.03762"&gt;&lt;strong&gt;Attention is all you need&lt;/strong&gt;&lt;/a&gt; is the first state of art paper in the field of NLP. The fundamental unit of said architecture, the transformer block, consists of two main components: &lt;strong&gt;a multi-head self-attention mechanism and a fully connected feedforward network&lt;/strong&gt;. The multi-head self-attention mechanism allows the model to focus on different parts of the input sequence at each layer and weigh the importance of each part in making a prediction. This is accomplished by computing &lt;strong&gt;attention scores&lt;/strong&gt; between each element in the input sequence and all other elements, which are then used to weigh the contribution of each token to the final representation. The multi-head attention mechanism allows the model to learn different attention patterns for different tasks and input sequences, making it a more versatile and effective algorithm.&lt;/p&gt;

&lt;p&gt;The feedforward network is essentially a multi-layer perceptron (MLP) that takes in the self-attention-generated representation as input, applies linear transformations with activation functions, and outputs the final representation. This final representation is then&lt;br&gt;
passed to the next Transformer block or used for making predictions&lt;/p&gt;

&lt;h3&gt;
  
  
  BERT:
&lt;/h3&gt;

&lt;p&gt;Later on, based on the Transformer model, One of the most significant developments in this direction was the introduction of &lt;strong&gt;BERT(Bidirectional Encoder Representations from Transformers)&lt;/strong&gt;. BERT is a pre-trained transformer model that can be fine-tuned for a wide range of NLP tasks, such as &lt;strong&gt;sentiment analysis, named entity recognition, and question answering&lt;/strong&gt;. BERT was trained using a contrastive task, where it was asked to predict missing tokens in a sentence given the context of the surrounding tokens. This approach allowed BERT to learn rich contextual representations of words, making it highly effective for a wide range of NLP tasks.&lt;/p&gt;

&lt;h3&gt;
  
  
  GPT:
&lt;/h3&gt;

&lt;p&gt;Another development in the field of transformers was the introduction of &lt;strong&gt;GPT (Generative Pretrained Transformer)&lt;/strong&gt;. GPT is a generative model trained on a large corpus of text with the goal of &lt;strong&gt;predicting the next token in a sequence&lt;/strong&gt; given the context of the surrounding tokens. GPT has been shown to be highly effective for tasks such as text &lt;strong&gt;generation, language modeling, and question-answering&lt;/strong&gt;. Unlike BERT, which was trained using a contrastive task, GPT was trained using a generative task, allowing it to learn a more diverse and complete representation of language.&lt;/p&gt;

&lt;p&gt;Transformers have not only revolutionized the field of NLP; they are growing beyond it and finding applications in other areas. For example, transformers have been used in &lt;strong&gt;computer vision tasks such as image captioning&lt;/strong&gt;, where they have been used to generate&lt;br&gt;
captions for images based on their content, and have been used in &lt;strong&gt;speech recognition&lt;/strong&gt;, where they have been used to transcribe speech into text.&lt;/p&gt;

&lt;p&gt;Another trend in the use of transformers is the development of &lt;strong&gt;multimodal models&lt;/strong&gt;, which allow for the unified modeling and use of text along with other modalities, such as images and audio. These models can help to understand the relationships between different modalities and can use this understanding to perform a wide range of tasks, such as image-to-text generation, text-to-image generation, and audio-to-text generation. Indeed, transformers are growing beyond the field of NLP and are being used in a wide range of&lt;br&gt;
tasks and applications. &lt;/p&gt;

&lt;h4&gt;
  
  
  Categorization:
&lt;/h4&gt;

&lt;p&gt;The kind of transformer architecture used in NLP applications plays a crucial role in determining the overall performance of the system. &lt;/p&gt;

&lt;p&gt;Encoder - decoder based:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;
&lt;strong&gt;Encoder-only&lt;/strong&gt; transformers are used for discriminative tasks such as sentiment analysis and named entity recognition.&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;Decoder-only&lt;/strong&gt; transformers are used for tasks such as text generation and summarization. &lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;Encoder–decoder&lt;/strong&gt; transformers are used for tasks such as machine translation and image captioning.&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;Modality based:&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;Modality&lt;/strong&gt; refers to the different modes or types of data and information that can be processed and generated beyond text.&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;
&lt;strong&gt;Unimodal NLP applications&lt;/strong&gt; deal with a single modality, such as text or speech&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;Multimodal NLP applications&lt;/strong&gt; deal with multiple modalities, such as text, speech, and images. Text often serves as the primary interface in multimodal applications.&lt;/li&gt;
&lt;/ul&gt;

&lt;h2&gt;
  
  
  Real-world Applications:
&lt;/h2&gt;

&lt;ol&gt;
&lt;li&gt;Unimodal Applications:&lt;/li&gt;
&lt;/ol&gt;

&lt;p&gt;Unimodal applications refer to AI-based systems that primarily focus on processing and analyzing text as their main modality.&lt;/p&gt;

&lt;h4&gt;
  
  
  i: &lt;em&gt;Language Modeling&lt;/em&gt;:
&lt;/h4&gt;

&lt;p&gt;Language modeling is a fundamental task in NLP that involves predicting the next word in a sequence of text based on the preceding words.&lt;br&gt;
Language modeling typically follows the decoder-only architecture.&lt;br&gt;
The goal of language modeling is to estimate the probability distribution of sequences of words in a given language and is used as a building block for many NLP tasks such as machine translation, speech recognition, and text generation. Language modeling can be easily extended to more complex NLP tasks such as sentence-pair modeling, cross-document language modeling, and definition modeling.&lt;/p&gt;

&lt;h4&gt;
  
  
  ii: &lt;em&gt;Question Answering&lt;/em&gt;:
&lt;/h4&gt;

&lt;p&gt;Question Answering is an NLP application that involves automatically answering questions posed in natural language. The goal of question answering is to extract the relevant information from a given text corpus and present it as an answer to a user’s question.&lt;br&gt;
Question-answering systems can operate over a wide range of text types, including news articles, Wikipedia pages, and others, and can be designed to answer a wide range of questions, including fact-based questions, opinion questions, and others. &lt;/p&gt;

&lt;p&gt;There are several subtasks within QA, as follows:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;&lt;p&gt;&lt;strong&gt;Open-Domain Question Answering (ODQA)&lt;/strong&gt;: This task involves finding an answer to a question from an open domain, such as the entire internet or a large corpus of text. The goal is to find the most relevant information to answer the question, even if it requires synthesizing information from multiple sources. &lt;strong&gt;Reformer&lt;/strong&gt; is a deep learning model for ODQA.&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;&lt;strong&gt;Conversational Question Answering (CQA)&lt;/strong&gt;: This task involves answering questions in a conversational setting, where the model must understand the context of the conversation and generate an answer that is relevant and appropriate for the current conversational context. &lt;strong&gt;SDNet&lt;/strong&gt; is a deep-learning model for conversational question answering (CQA)&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;&lt;strong&gt;Answer Selection&lt;/strong&gt;: This task involves ranking a set of candidate answers for a given question, where the goal is to select the most accurate answer from the candidate set.&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;&lt;strong&gt;Machine Reading Comprehension (MRC)&lt;/strong&gt;: This task involves understanding and answering questions about a given passage of text. The model must be able to comprehend the text, extract relevant information, and generate an answer that is accurate and relevant to the question. &lt;strong&gt;XLNet&lt;/strong&gt; is a deep-learning model MRC.&lt;/p&gt;&lt;/li&gt;
&lt;/ul&gt;

&lt;h4&gt;
  
  
  iii: &lt;em&gt;Machine Translation&lt;/em&gt;:
&lt;/h4&gt;

&lt;p&gt;It is the task of automatically converting a source text in&lt;br&gt;
one language to a target text in another language. The goal of machine translation is to produce a fluent and accurate translation that conveys the meaning of the source text in the target language. MT models often follow an encoder–decoder architecture to capture the context effectively using a bidirectional encoder and be able to generate text of arbitrary&lt;br&gt;
length, following the original formulation of transformer architecture. There are several subtasks within MT,&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;&lt;p&gt;&lt;strong&gt;Transliteration&lt;/strong&gt;: It involves translating text from one script to another, such as translating between the Latin and Cyrillic scripts. It involves preserving the meaning of words, rather than translating the meaning of words to another language.&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;&lt;strong&gt;Unsupervised Machine Translation (UMT)&lt;/strong&gt;: It involves translating between two languages without any parallel training data, meaning that there is no corresponding text in the target language for the source language text. UMT models are typically trained on monolingual data in each language.&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;&lt;strong&gt;Bilingual Lexicon Induction (BLI)&lt;/strong&gt;: It involves inducing a bilingual lexicon, automatically discovering word translation pairs or mappings between two languages without the need for explicit bilingual dictionaries or parallel corpora.&lt;/p&gt;&lt;/li&gt;
&lt;/ul&gt;

&lt;h4&gt;
  
  
  iv: &lt;em&gt;Text classification&lt;/em&gt;:
&lt;/h4&gt;

&lt;p&gt;Text classification is the task of categorizing a text into one or more predefined categories based on its content. The goal of text classification is to automatically assign a label to a given text based on its content, allowing it to be organized and categorized for easier analysis and management. &lt;br&gt;
These models are trained on annotated text data in order to learn&lt;br&gt;
the relationship between the text content and its label, and can then be used to classify new unseen text data. &lt;br&gt;
Text classification models typically follow a decoder-only architecture.&lt;br&gt;
its subcategories are&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;&lt;p&gt;&lt;strong&gt;Document Classification&lt;/strong&gt;: This task involves assigning a label or category to a full document, such as a news article, blog post, or scientific paper. Document classification is typically accomplished by first representing the document as a numerical vector and then using a machine-learning model to make a prediction based on the document’s representation. &lt;br&gt;
&lt;strong&gt;LinkBERT&lt;/strong&gt; extends the pre-training objective of BERT&lt;br&gt;
to incorporate links between documents.&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;&lt;strong&gt;Cause and Effect Classification&lt;/strong&gt;: This task involves identifying the cause and effect relationship between two events described in a sentence or paragraph.&lt;/p&gt;&lt;/li&gt;
&lt;/ul&gt;

&lt;h4&gt;
  
  
  v: &lt;em&gt;Text Generation&lt;/em&gt;:
&lt;/h4&gt;

&lt;p&gt;Text Generation is a task in NLP in which the objective is to produce new text automatically, typically starting from a given prompt or input. The output can be a single word, phrase, sentence, or full-length piece of text, and is used for chatbots, content creation, and more. The generated text should reflect an understanding of the input and the language being generated, and the quality and coherence of the generated text can vary depending on the approach used. &lt;br&gt;
Text generation typically follows a decoder-only architecture, however, recent issues with &lt;strong&gt;prompt-injection&lt;/strong&gt; attacks have&lt;br&gt;
migrated part of the focus towards encoder-decoder models that have been instruction-tuned, such as T5.&lt;br&gt;
Text generation subtasks include:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;&lt;p&gt;&lt;strong&gt;Dialogue Generation&lt;/strong&gt;: It focuses on generating text in the form of a conversation between two or more agents. Dialogue generation systems are used in various applications, such as chatbots, virtual assistants, and conversational AI systems. These systems use dialogue history, user input, and context to generate appropriate and coherent responses. P2-BOT is a transmitter–receiver-based framework that aims to explicitly model understanding in chat dialogue systems through mutual persona perception.&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;&lt;strong&gt;Code Generation&lt;/strong&gt;: It focuses on generating code based on a given input, such as a natural language description of a software problem. Code generation systems are used in software development to automate repetitive tasks, improve productivity, and reduce errors.&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;&lt;strong&gt;Data-to-Text Generation&lt;/strong&gt;: It focuses on generating natural language text from structured data such as tables, databases, or graphs. Data-to-text generation systems are used in various applications, such as news reporting, data visualization, and technical writing.&lt;/p&gt;&lt;/li&gt;
&lt;/ul&gt;

&lt;h4&gt;
  
  
  vi: &lt;em&gt;Text Summarization&lt;/em&gt;:
&lt;/h4&gt;

&lt;p&gt;Text Summarization is a task in NLP where the goal is to condense a given text into a shorter and more concise version while preserving its essential information. This is typically &lt;br&gt;
accomplished by identifying and extracting the most important information, sentences, or phrases from the original text.&lt;br&gt;
Text summarization is used in a variety of applications, such&lt;br&gt;
as news aggregation, document summarization, and more. &lt;br&gt;
Text summarization typically requires an encoder–decoder architecture to completely capture the source information.&lt;/p&gt;

&lt;p&gt;its sub-categories are,&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;&lt;p&gt;&lt;strong&gt;Extractive Summarization&lt;/strong&gt;: It extracts the most important sentences or phrases from a document and presents them as a summary. Extractive summarization methods typically use a combination of information retrieval and natural language processing techniques to identify the most informative sentences or phrases in a document.&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;&lt;strong&gt;Abstractive Summarization&lt;/strong&gt;: It generates a summary by synthesizing new information based on the input document. &lt;br&gt;
These models are trained on large amounts of data and can generate summaries that are more concise and coherent than extractive summaries. &lt;strong&gt;mBart&lt;/strong&gt; is a sequence-to-sequence transformer trained on multiple large-scale monolingual corpora with the objective of denoising.&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;&lt;strong&gt;Multi-Document Summarization&lt;/strong&gt;: It summarizes multiple related documents into a single summary. Multi-document summarization methods typically use information retrieval techniques to identify the most important documents and natural language processing techniques to generate a summary from the selected documents.&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;&lt;strong&gt;Query-Focused Summarization&lt;/strong&gt;: It summarizing a document&lt;br&gt;
based on a specific query or topic. Query-focused summarization methods use information retrieval techniques to identify the most relevant sentences or phrases in a document and present them as a summary.&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;&lt;strong&gt;Sentence Compression&lt;/strong&gt;: It focuses on reducing the length of a sentence while preserving its meaning. Sentence compression methods typically use natural language processing techniques to identify redundant or unnecessary words or phrases in a sentence and remove them to create a more concise sentence.  &lt;strong&gt;DistilRoBERTa&lt;/strong&gt; is a reinforcement learning algorithm to predict a binary classifier that keeps or discards words to reduce sentence length.&lt;/p&gt;&lt;/li&gt;
&lt;/ul&gt;

&lt;h4&gt;
  
  
  vii: &lt;em&gt;Sentiment Analysis&lt;/em&gt;:
&lt;/h4&gt;

&lt;p&gt;Sentiment Analysis is a task in NLP with the goal of determining the sentiment expressed in a given text. This is typically accomplished by assigning a sentiment label such as positive, negative, or neutral to the text based on its contents. &lt;br&gt;
The sentiment can be expressed in different forms, such as opinions, emotions, or evaluations, and can be expressed at various levels of granularity, such as at the document, sentence, or aspect level. Sentiment Analysis is used in a variety of applications, such as customer service, marketing, and opinion mining. &lt;br&gt;
The quality of the sentiment analysis results can be influenced by factors such as the subjectivity of the text, the tone, and the context in which the sentiment is expressed.&lt;/p&gt;

&lt;h4&gt;
  
  
  viii: &lt;em&gt;Named Entity Recognition&lt;/em&gt;:
&lt;/h4&gt;

&lt;p&gt;Named Entity Recognition (NER) is a task in NLP with the goal of identifying and categorizing named entities present in a given text into predefined categories such as person names, organizations, locations, dates, and more. NER is used as an intermediate step in various applications such as question-answering, event extraction, and information retrieval. &lt;br&gt;
It typically utilizes an encoder-only architecture. While&lt;br&gt;
the approach of fine-tuning a pre-trained model with a classification head added on top for NER work well in practice, Automated Concatenation of Embeddings (ACE) has&lt;br&gt;
shown improved results using an ensemble of several pre-trained models while training only a simple classifier on top using reinforcement learning.&lt;/p&gt;

&lt;h4&gt;
  
  
  ix: &lt;em&gt;Information Retrieval&lt;/em&gt;:
&lt;/h4&gt;

&lt;p&gt;Information Retrieval (IR) is a task in NLP with the goal of retrieving relevant information from a large collection of documents in response to a user query. This is typically&lt;br&gt;
accomplished by matching the query terms against the document content and ranking the documents based on their relevance to the query. &lt;br&gt;
IR systems can be used for various applications, such as web search, document search, and question answering. The quality&lt;br&gt;
of the retrieval, results can be influenced by factors such as the relevance of the documents, the effectiveness of the ranking algorithm, and the representation of the documents and queries.&lt;br&gt;
IR systems are typically classified further based on the level of granularity, such as document, paragraph, sentence, etc. The typical methods for retrieval include the use of a pre-trained model such as &lt;strong&gt;RoBERTa&lt;/strong&gt; in a Siamese fashion to find the similarity between two embeddings.&lt;/p&gt;

&lt;ol&gt;
&lt;li&gt;Multimodal Applications:
Multimodal applications are software or systems that process and integrate information from multiple modalities or types of data, such as text, images, speech, and more, to provide a richer and more comprehensive user experience. These applications leverage multiple sources of data to enhance understanding, enable interaction, and solve complex problems.&lt;/li&gt;
&lt;/ol&gt;

&lt;h4&gt;
  
  
  i: &lt;em&gt;Generative Control&lt;/em&gt;:
&lt;/h4&gt;

&lt;p&gt;Generative Control is a task in multimodal NLP in which text is used as an interface to generate another modality, such as images or speech. The goal of Generative Control is to generate a target modality that corresponds to a given text description or instruction.&lt;br&gt;
For example, based on a textual description of an object, such as "a red sports car," the task of Generative Control would be to generate an image of a red sports car. &lt;br&gt;
Generative Control combines the strengths of NLP and computer graphics or speech synthesis to produce high-quality and semantically meaningful outputs in the target modality. It has&lt;br&gt;
applications in areas such as computer vision, robotics, and human–computer interaction.&lt;/p&gt;

&lt;h4&gt;
  
  
  ii: &lt;em&gt;Description Generation&lt;/em&gt;:
&lt;/h4&gt;

&lt;p&gt;Description generation, a subset of natural language processing (NLP), involves automatically creating human-like text descriptions based on structured data, prompts, or other information. Description Generation would be to generate a textual description of the objects, actions,&lt;br&gt;
and attributes present in the scene. &lt;br&gt;
It aims to generate coherent and contextually relevant text for various applications, such as product descriptions, data visualization, virtual assistants, and content generation. This process can be rule-based, template-based, or driven by machine learning models that capture complex linguistic patterns and context. &lt;/p&gt;

&lt;h4&gt;
  
  
  iii: &lt;em&gt;Multimodal Question Answering&lt;/em&gt;:
&lt;/h4&gt;

&lt;p&gt;Multimodal Question Answering (QA) is a task with the goal of answering questions about a given multimodal input, such as an image or a video, using information from multiple modalities. The task involves combining information from text, images, audio,&lt;br&gt;
and other modalities to accurately answer questions about the content of the input. &lt;br&gt;
For example, given an image of a scene and a question about the scene, such as “What is the color of the car?”, the task of Multimodal QA would be to identify the car in the image&lt;br&gt;
and answer the question with the correct color. Multimodal QA requires the integration of NLP, computer vision, and other relevant modalities&lt;br&gt;
&lt;strong&gt;BEiT&lt;/strong&gt; performs masked language modeling on images, texts, and image-text pairs&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;Resource: &lt;a href="https://www.mdpi.com/2078-2489/14/4/242"&gt;Transformers in the Real World: A Survey on NLP Applications&lt;/a&gt;
&lt;/li&gt;
&lt;/ul&gt;

</description>
    </item>
    <item>
      <title>Day-20 of Machine Learning</title>
      <dc:creator>Ankita Sahoo</dc:creator>
      <pubDate>Sat, 10 Dec 2022 16:42:14 +0000</pubDate>
      <link>https://dev.to/ank1tas/day-20-of-machine-learning-4fa7</link>
      <guid>https://dev.to/ank1tas/day-20-of-machine-learning-4fa7</guid>
      <description>&lt;p&gt;Day-20 of Machine Learning:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;Learnt about &lt;strong&gt;cross-validation&lt;/strong&gt;:&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;Cross-validation is a technique for evaluating a machine learning model and testing its performance.&lt;br&gt;
Cross validation refers to an extra data set that used to cross check the validity or the accuracy of the model.&lt;br&gt;
for example, let a ML model trained over the training-set and the accuracy is 98%. Wow!!! But still we can't say that the model has trained very well, and it is because it is performing well over the data that it already knows and if we expose the model to completely new, unseen data, it might not predict with the same accuracy and it might fail to generalize over the new data. it is the case of &lt;strong&gt;overfitting&lt;/strong&gt; or &lt;strong&gt;High variance&lt;/strong&gt;. Similarly, if the model does not perform well on training-set as well as cross-validation set, then it is &lt;strong&gt;Underfitting / High bias&lt;/strong&gt;.&lt;/p&gt;

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fkwl6heyusp5vhx9qyyay.png" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fkwl6heyusp5vhx9qyyay.png" alt="Image description" width="602" height="184"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;Using Training-set we fit the parameters [W,b] then choose parameter d (degree of polynomial) using cross-validation set. So, we have not used test-set till now to fit either of any parameters, that is why  Jtest will be a fire estimate of the &lt;strong&gt;generalization&lt;/strong&gt; error of the model&lt;/p&gt;

</description>
      <category>deepseek</category>
    </item>
    <item>
      <title>Day-19 of Machine Learning:</title>
      <dc:creator>Ankita Sahoo</dc:creator>
      <pubDate>Tue, 25 Oct 2022 18:09:18 +0000</pubDate>
      <link>https://dev.to/ank1tas/day-19-of-machine-learning-1cfc</link>
      <guid>https://dev.to/ank1tas/day-19-of-machine-learning-1cfc</guid>
      <description>&lt;p&gt;Day-19 of Machine Learning:&lt;br&gt;
I. Basic template of TensorFlow implementation:&lt;/p&gt;

&lt;h4&gt;
  
  
  1. construct the network
&lt;/h4&gt;



&lt;div class="highlight js-code-highlight"&gt;
&lt;pre class="highlight python"&gt;&lt;code&gt;&lt;span class="n"&gt;model&lt;/span&gt; &lt;span class="o"&gt;=&lt;/span&gt; &lt;span class="nc"&gt;Sequential&lt;/span&gt;&lt;span class="p"&gt;(&lt;/span&gt;
    &lt;span class="p"&gt;[&lt;/span&gt;               
        &lt;span class="n"&gt;tf&lt;/span&gt;&lt;span class="p"&gt;.&lt;/span&gt;&lt;span class="n"&gt;keras&lt;/span&gt;&lt;span class="p"&gt;.&lt;/span&gt;&lt;span class="nc"&gt;Input&lt;/span&gt;&lt;span class="p"&gt;(&lt;/span&gt;&lt;span class="n"&gt;shape&lt;/span&gt;&lt;span class="o"&gt;=&lt;/span&gt;&lt;span class="p"&gt;(&lt;/span&gt;&lt;span class="mi"&gt;400&lt;/span&gt;&lt;span class="p"&gt;,)),&lt;/span&gt;    &lt;span class="c1"&gt;#specify input size
&lt;/span&gt;
        &lt;span class="nc"&gt;Dense&lt;/span&gt;&lt;span class="p"&gt;(&lt;/span&gt;&lt;span class="mi"&gt;25&lt;/span&gt;&lt;span class="p"&gt;,&lt;/span&gt; &lt;span class="n"&gt;activation&lt;/span&gt;&lt;span class="o"&gt;=&lt;/span&gt;&lt;span class="sh"&gt;'&lt;/span&gt;&lt;span class="s"&gt;sigmoid&lt;/span&gt;&lt;span class="sh"&gt;'&lt;/span&gt;&lt;span class="p"&gt;),&lt;/span&gt; 
        &lt;span class="nc"&gt;Dense&lt;/span&gt;&lt;span class="p"&gt;(&lt;/span&gt;&lt;span class="mi"&gt;15&lt;/span&gt;&lt;span class="p"&gt;,&lt;/span&gt; &lt;span class="n"&gt;activation&lt;/span&gt;&lt;span class="o"&gt;=&lt;/span&gt;&lt;span class="sh"&gt;'&lt;/span&gt;&lt;span class="s"&gt;sigmoid&lt;/span&gt;&lt;span class="sh"&gt;'&lt;/span&gt;&lt;span class="p"&gt;),&lt;/span&gt; 
        &lt;span class="nc"&gt;Dense&lt;/span&gt;&lt;span class="p"&gt;(&lt;/span&gt;&lt;span class="mi"&gt;1&lt;/span&gt;&lt;span class="p"&gt;,&lt;/span&gt;  &lt;span class="n"&gt;activation&lt;/span&gt;&lt;span class="o"&gt;=&lt;/span&gt;&lt;span class="sh"&gt;'&lt;/span&gt;&lt;span class="s"&gt;sigmoid&lt;/span&gt;&lt;span class="sh"&gt;'&lt;/span&gt;&lt;span class="p"&gt;)&lt;/span&gt;  


    &lt;span class="p"&gt;],&lt;/span&gt; &lt;span class="n"&gt;name&lt;/span&gt; &lt;span class="o"&gt;=&lt;/span&gt; &lt;span class="sh"&gt;"&lt;/span&gt;&lt;span class="s"&gt;my_model&lt;/span&gt;&lt;span class="sh"&gt;"&lt;/span&gt; 
&lt;span class="p"&gt;)&lt;/span&gt;                            

&lt;/code&gt;&lt;/pre&gt;

&lt;/div&gt;



&lt;p&gt;Keras Sequential model and Dense Layer with sigmoid activations.&lt;/p&gt;

&lt;h4&gt;
  
  
  2. loss function
&lt;/h4&gt;



&lt;div class="highlight js-code-highlight"&gt;
&lt;pre class="highlight python"&gt;&lt;code&gt;&lt;span class="n"&gt;model&lt;/span&gt;&lt;span class="p"&gt;.&lt;/span&gt;&lt;span class="nf"&gt;compile&lt;/span&gt;&lt;span class="p"&gt;(&lt;/span&gt;
    &lt;span class="n"&gt;loss&lt;/span&gt;&lt;span class="o"&gt;=&lt;/span&gt;&lt;span class="n"&gt;tf&lt;/span&gt;&lt;span class="p"&gt;.&lt;/span&gt;&lt;span class="n"&gt;keras&lt;/span&gt;&lt;span class="p"&gt;.&lt;/span&gt;&lt;span class="n"&gt;losses&lt;/span&gt;&lt;span class="p"&gt;.&lt;/span&gt;&lt;span class="nc"&gt;BinaryCrossentropy&lt;/span&gt;&lt;span class="p"&gt;(),&lt;/span&gt;
    &lt;span class="n"&gt;optimizer&lt;/span&gt;&lt;span class="o"&gt;=&lt;/span&gt;&lt;span class="n"&gt;tf&lt;/span&gt;&lt;span class="p"&gt;.&lt;/span&gt;&lt;span class="n"&gt;keras&lt;/span&gt;&lt;span class="p"&gt;.&lt;/span&gt;&lt;span class="n"&gt;optimizers&lt;/span&gt;&lt;span class="p"&gt;.&lt;/span&gt;&lt;span class="nc"&gt;Adam&lt;/span&gt;&lt;span class="p"&gt;(&lt;/span&gt;&lt;span class="mf"&gt;0.001&lt;/span&gt;&lt;span class="p"&gt;),&lt;/span&gt;
&lt;span class="p"&gt;)&lt;/span&gt;

&lt;/code&gt;&lt;/pre&gt;

&lt;/div&gt;



&lt;p&gt;Here for &lt;strong&gt;binary classification&lt;/strong&gt;, BinaryCrossentropy() is used. We can also use MeanSquareError() for Linear regression.&lt;/p&gt;

&lt;h4&gt;
  
  
  3. gradient descent to fit the weights of the model to the training data
&lt;/h4&gt;



&lt;div class="highlight js-code-highlight"&gt;
&lt;pre class="highlight python"&gt;&lt;code&gt;&lt;span class="n"&gt;model&lt;/span&gt;&lt;span class="p"&gt;.&lt;/span&gt;&lt;span class="nf"&gt;fit&lt;/span&gt;&lt;span class="p"&gt;(&lt;/span&gt;
    &lt;span class="n"&gt;X&lt;/span&gt;&lt;span class="p"&gt;,&lt;/span&gt;&lt;span class="n"&gt;y&lt;/span&gt;&lt;span class="p"&gt;,&lt;/span&gt;
    &lt;span class="n"&gt;epochs&lt;/span&gt;&lt;span class="o"&gt;=&lt;/span&gt;&lt;span class="mi"&gt;20&lt;/span&gt;
&lt;span class="p"&gt;)&lt;/span&gt;

&lt;/code&gt;&lt;/pre&gt;

&lt;/div&gt;






&lt;p&gt;II. Got to know about different Activation&lt;/p&gt;

&lt;h5&gt;
  
  
  - Linear Activation:
&lt;/h5&gt;

&lt;p&gt;Activation &lt;strong&gt;a = g(Z) = Z&lt;/strong&gt; &lt;br&gt;
          where Z = W.X + b &lt;br&gt;
&lt;strong&gt;Output y&lt;/strong&gt; might be an Integer number &lt;strong&gt;(+ve/-ve)&lt;/strong&gt;&lt;/p&gt;

&lt;h5&gt;
  
  
  - Sigmoid Activation:
&lt;/h5&gt;

&lt;p&gt;Activation &lt;strong&gt;a = g(Z) = 1 / (1 + e ^ (-Z))&lt;/strong&gt;. &lt;br&gt;
&lt;strong&gt;Output y&lt;/strong&gt; might be &lt;strong&gt;0 or 1 i.e binary classification&lt;/strong&gt;&lt;/p&gt;

&lt;h5&gt;
  
  
  - ReLU Activation (Rectified Linear Activation):
&lt;/h5&gt;

&lt;p&gt;Activation &lt;strong&gt;a = g(Z) = max (0, Z)&lt;/strong&gt;.&lt;br&gt;
&lt;strong&gt;Output y&lt;/strong&gt; will be any &lt;strong&gt;Whole number&lt;/strong&gt;&lt;/p&gt;




&lt;p&gt;III. How to choose Activation?&lt;br&gt;
We can choose different activation within a Neural Network for separate layers and activations can be chosen accordingly requirement and goal of the Neural Network. However some recommendations are,&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;A neural network with many layers but no activation function is not effective. A Neural network with &lt;strong&gt;only linear activation&lt;/strong&gt; is the same as &lt;strong&gt;no activation function&lt;/strong&gt;.&lt;/li&gt;
&lt;li&gt;ReLU are often use than Sigmoid activation. It is because firstly &lt;strong&gt;ReLU is a bit faster as it does less computation (max of 0 and Z)&lt;/strong&gt; than sigmoid which does exponential then inverse and so on. Secondly Gradient Descent goes slow for flat and ReLU goes flat in one place whereas Sigmoid in 2 places.&lt;/li&gt;
&lt;li&gt;Use &lt;strong&gt;ReLU&lt;/strong&gt; instead of Linear Activation in &lt;strong&gt;Hidden layers&lt;/strong&gt;.
&lt;/li&gt;
&lt;/ul&gt;

</description>
      <category>beginners</category>
      <category>machinelearning</category>
      <category>tensorflow</category>
      <category>neuralnetwork</category>
    </item>
    <item>
      <title>Day-18 of Machine Learning:</title>
      <dc:creator>Ankita Sahoo</dc:creator>
      <pubDate>Fri, 21 Oct 2022 20:10:53 +0000</pubDate>
      <link>https://dev.to/ank1tas/day-18-of-machine-learning-2dd4</link>
      <guid>https://dev.to/ank1tas/day-18-of-machine-learning-2dd4</guid>
      <description>&lt;p&gt;Day-18 of Machine Learning:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;Got a brief intuition on &lt;strong&gt;AGI (artificial general intelligence)&lt;/strong&gt; and realised &lt;strong&gt;ANI (artificial narrow intelligence)&lt;/strong&gt; had made tremendous progress and values and so on &lt;strong&gt;AI&lt;/strong&gt; as compared to AGI with lots of valid reasons. &lt;/li&gt;
&lt;li&gt;Learnt how to implement &lt;strong&gt;Forward propagation&lt;/strong&gt; using &lt;strong&gt;NumPy&lt;/strong&gt; instead of TensorFlow and it gave me a basic idea about how &lt;strong&gt;dense function&lt;/strong&gt; is working under the hood in TensorFlow.&lt;/li&gt;
&lt;li&gt;Learnt TensorFlow uses &lt;strong&gt;matrix&lt;/strong&gt; to store training-set parameters in instead of vector/list for efficient computation. &lt;strong&gt;weight-parameters for i-th&lt;/strong&gt; unit/neuron is arranged in &lt;strong&gt;i-th column&lt;/strong&gt; of the matrix whereas &lt;strong&gt;row is the list of values of a parameter for units/neurons&lt;/strong&gt; and TensorFlow utilizes a &lt;strong&gt;1-D&lt;/strong&gt; representation to store &lt;strong&gt;bias parameter&lt;/strong&gt;. Also, inside dense function, matrix multiplication and sigmoid function implementation are performed.&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;Also,&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;Practiced some basic DSA questions on LinkedList.&lt;/li&gt;
&lt;li&gt;Created a simplistic animation effect (gif) on python using &lt;a href="https://pillow.readthedocs.io/en/stable/"&gt;Pillow Library&lt;/a&gt;, inspired by CS50P &lt;/li&gt;
&lt;/ul&gt;

</description>
      <category>machinelearning</category>
      <category>python</category>
      <category>dsa</category>
      <category>tensorflow</category>
    </item>
    <item>
      <title>Day-17 of Machine Learning:</title>
      <dc:creator>Ankita Sahoo</dc:creator>
      <pubDate>Wed, 19 Oct 2022 18:36:10 +0000</pubDate>
      <link>https://dev.to/ank1tas/day-17-of-machine-learning-4hmj</link>
      <guid>https://dev.to/ank1tas/day-17-of-machine-learning-4hmj</guid>
      <description>&lt;p&gt;Day-17 of Machine Learning:&lt;br&gt;
Got a brief intuition on &lt;strong&gt;TensorFlow&lt;/strong&gt; where I learnt about &lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;how &lt;strong&gt;tensor&lt;/strong&gt; stores data in matrix, about &lt;strong&gt;dense&lt;/strong&gt; function which is used to create fully connected &lt;strong&gt;layers&lt;/strong&gt; and about &lt;strong&gt;Sequential model&lt;/strong&gt;.&lt;/li&gt;
&lt;li&gt;how &lt;strong&gt;neural networks handle complex decisions&lt;/strong&gt; by dividing the decisions between &lt;strong&gt;hidden layers&lt;/strong&gt; and again each layers in to  &lt;strong&gt;multiple units or neurons&lt;/strong&gt;.&lt;/li&gt;
&lt;li&gt;Implemented small neural network in TensorFlow binary classification model to recognise &lt;strong&gt;handwritten digit&lt;/strong&gt; that works for only 1 and 0.&lt;/li&gt;
&lt;/ul&gt;

</description>
      <category>beginners</category>
      <category>machinelearning</category>
      <category>tensorflow</category>
      <category>neuralnetwork</category>
    </item>
    <item>
      <title>Day-16 of Machine Learning:</title>
      <dc:creator>Ankita Sahoo</dc:creator>
      <pubDate>Tue, 18 Oct 2022 06:19:04 +0000</pubDate>
      <link>https://dev.to/ank1tas/day-16-of-machine-learning-24i</link>
      <guid>https://dev.to/ank1tas/day-16-of-machine-learning-24i</guid>
      <description>&lt;p&gt;Day-16 of Machine Learning:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;Started course 2 - &lt;strong&gt;Advance Learning algorithm&lt;/strong&gt;, where got a brief intuition on &lt;strong&gt;Neural network, its layer, neurons&lt;/strong&gt; and a simple visualization how things work in the hidden layer.&lt;/li&gt;
&lt;/ul&gt;

</description>
      <category>beginners</category>
      <category>machinelearning</category>
      <category>neuralnetwork</category>
    </item>
    <item>
      <title>Day-15 of Machine Learning</title>
      <dc:creator>Ankita Sahoo</dc:creator>
      <pubDate>Thu, 29 Sep 2022 12:00:21 +0000</pubDate>
      <link>https://dev.to/ank1tas/day-15-of-machine-learning-39e4</link>
      <guid>https://dev.to/ank1tas/day-15-of-machine-learning-39e4</guid>
      <description>&lt;p&gt;Day-15 of Machine Learning:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;Implemented &lt;strong&gt;Sigmoid function, cost function, Gradient descent for logistic regression&lt;/strong&gt; for Logistic Regression and a prediction function with 0.5 threshold to build a logistic regression model to predict whether a student gets admitted into a university.&lt;/li&gt;
&lt;li&gt;Also implemented regularized logistic regression to predict whether microchips from a fabrication plant passes quality assurance (QA) where I implemented cost function and gradient descent for &lt;strong&gt;Regularised Logistic Regression&lt;/strong&gt;.&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;With this I completed &lt;strong&gt;Course-1: Supervised Machine Learning: Regression and Classification&lt;/strong&gt;&lt;/p&gt;

</description>
      <category>beginners</category>
      <category>machinelearning</category>
      <category>coursera</category>
      <category>supervisedmachinelearning</category>
    </item>
    <item>
      <title>Day-14 of Machine Learning</title>
      <dc:creator>Ankita Sahoo</dc:creator>
      <pubDate>Sun, 25 Sep 2022 17:59:58 +0000</pubDate>
      <link>https://dev.to/ank1tas/day-14-of-machine-learning-21bd</link>
      <guid>https://dev.to/ank1tas/day-14-of-machine-learning-21bd</guid>
      <description>&lt;p&gt;Day-14 of Machine Learning:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;
&lt;p&gt;Learnt about &lt;strong&gt;Underfit&lt;/strong&gt;, &lt;strong&gt;Generalization&lt;/strong&gt; and &lt;strong&gt;Overfit&lt;/strong&gt;.&lt;br&gt;
and how to address overfit:&lt;/p&gt;

&lt;ol&gt;
&lt;li&gt;Collect more data / training sets.&lt;/li&gt;
&lt;li&gt;Feature selection.&lt;/li&gt;
&lt;li&gt;Regularization.&lt;/li&gt;
&lt;/ol&gt;
&lt;/li&gt;
&lt;li&gt;&lt;p&gt;Learnt how to apply &lt;strong&gt;regularization&lt;/strong&gt; on &lt;strong&gt;linear regression&lt;/strong&gt; and &lt;strong&gt;Logistic Regression&lt;/strong&gt; and to &lt;strong&gt;penalise or shrink features&lt;/strong&gt; that will help to build a simpler model.&lt;/p&gt;&lt;/li&gt;
&lt;/ul&gt;

</description>
      <category>beginners</category>
      <category>machinelearning</category>
      <category>tutorial</category>
    </item>
    <item>
      <title>Day-13 of Machine Learning</title>
      <dc:creator>Ankita Sahoo</dc:creator>
      <pubDate>Wed, 21 Sep 2022 05:30:58 +0000</pubDate>
      <link>https://dev.to/ank1tas/day-13-of-machine-learning-14bd</link>
      <guid>https://dev.to/ank1tas/day-13-of-machine-learning-14bd</guid>
      <description>&lt;p&gt;Day-13 of Machine Learning:&lt;br&gt;
Learnt to calculate &lt;strong&gt;gradient descent&lt;/strong&gt; for &lt;strong&gt;logistics regression&lt;/strong&gt; with single-featured data set as well as two-featured data set.&lt;/p&gt;

</description>
      <category>beginners</category>
      <category>machinelearning</category>
    </item>
    <item>
      <title>Day-12 of Machine Learning</title>
      <dc:creator>Ankita Sahoo</dc:creator>
      <pubDate>Tue, 20 Sep 2022 07:05:15 +0000</pubDate>
      <link>https://dev.to/ank1tas/day-12-of-machine-learning-3jp6</link>
      <guid>https://dev.to/ank1tas/day-12-of-machine-learning-3jp6</guid>
      <description>&lt;p&gt;Day-12 of Machine Learning:&lt;br&gt;
Learnt about &lt;strong&gt;loss function&lt;/strong&gt; and &lt;strong&gt;cost function&lt;/strong&gt; of &lt;strong&gt;logistic regression&lt;/strong&gt;.&lt;/p&gt;

</description>
      <category>beginners</category>
      <category>machinelearning</category>
    </item>
  </channel>
</rss>
