<?xml version="1.0" encoding="UTF-8"?>
<rss version="2.0" xmlns:atom="http://www.w3.org/2005/Atom" xmlns:dc="http://purl.org/dc/elements/1.1/">
  <channel>
    <title>DEV Community: Anisha Bhandare</title>
    <description>The latest articles on DEV Community by Anisha Bhandare (@anisha_bhandare_89bb0637f).</description>
    <link>https://dev.to/anisha_bhandare_89bb0637f</link>
    <image>
      <url>https://media2.dev.to/dynamic/image/width=90,height=90,fit=cover,gravity=auto,format=auto/https:%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Fuser%2Fprofile_image%2F2498817%2Fd7b10f1b-fa76-48f8-8d62-4169df4cf4a5.png</url>
      <title>DEV Community: Anisha Bhandare</title>
      <link>https://dev.to/anisha_bhandare_89bb0637f</link>
    </image>
    <atom:link rel="self" type="application/rss+xml" href="https://dev.to/feed/anisha_bhandare_89bb0637f"/>
    <language>en</language>
    <item>
      <title>Machine Learning for Newbies</title>
      <dc:creator>Anisha Bhandare</dc:creator>
      <pubDate>Mon, 06 Jan 2025 11:42:51 +0000</pubDate>
      <link>https://dev.to/anisha_bhandare_89bb0637f/machine-learning-for-newbies-2fgp</link>
      <guid>https://dev.to/anisha_bhandare_89bb0637f/machine-learning-for-newbies-2fgp</guid>
      <description>&lt;p&gt;A regular day for Gen-Z begins with cascade of social-media. LinkedIn notifications like “Your latest post got over ‘n’ impressions” or Instagram equivalent of “Z and other 5 people are following Y” are a staple. How do you think these statistics come forward? Obviously the developers at Microsoft or Meta cannot sit and perform statistical analysis for every user ; hence Machine Learning steps up . Be it personalized recommendations of Netflix shows , curated playlists of Spotify ,relevant results and auto-completing queries of search engines likes Google and Bing , dynamic prizing of e-commerce platforms like Amazon and price estimation and ETA predictions of Uber : all of them are the magic of Machine Learning. &lt;/p&gt;

&lt;h2&gt;
  
  
  What is Machine Learning?
&lt;/h2&gt;

&lt;blockquote&gt;
&lt;p&gt;Teaching computers to imitate the ways humans learn and empowering them to make decisions without human interventions , hence improving accuracy and precision is Machine Learning. &lt;/p&gt;
&lt;/blockquote&gt;

&lt;p&gt;While AI is a broad field consisting of a set of technologies that can be used to build machines having the ability to mimic functions or operations primarily associated with humans , ML is a subset of AI that allows machines to extract knowledge from data and learn from it without human intervention. &lt;/p&gt;

&lt;h2&gt;
  
  
  Components of Machine Learning
&lt;/h2&gt;

&lt;p&gt;&lt;strong&gt;1. Data&lt;/strong&gt;&lt;/p&gt;

&lt;p&gt;Data is the protagonist in the play of Machine Learning . ML models learn from the data. This performance and accuracy of the model depend largely on the quality , quantity and reliability of data fed to it . This data helps in making informed and evidence-backed decisions. Tasks like classification and prediction are performed based on the data . Ideally diverse ,complete , accurate and relevant data is preferred . Data might be obtained through various sources like public datasets(Kaggle), sensors and IoT devices(camera , microphones and temperature sensors) providing continuous data , user-generated data (reviews, social media posts), API and web-scraping . &lt;/p&gt;

&lt;p&gt;&lt;strong&gt;2. Features&lt;/strong&gt;&lt;/p&gt;

&lt;p&gt;Feature engineering is like preparing ingredients before cooking a meal. For a machine learning model, the features are the ingredients it uses to make predictions or decisions. This process involves selecting, cleaning, and enhancing data to optimize the model's performance. Raw data is transformed into meaningful, structured features that the model can effectively process and utilize. &lt;br&gt;
Raw data often comes in a messy format that machine learning models can't directly use. For example:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;&lt;p&gt;In a weather prediction model, you might have temperature readings taken every minute—but what you really need are daily averages.&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;For text analysis, you might start with full paragraphs but need to break them down into individual words or simple word counts.&lt;/p&gt;&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;Through proper feature engineering, you transform this raw data into something the model can effectively understand and use.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;3. Model&lt;/strong&gt;&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;Learning Algorithm 
A learning algorithm enables computers to learn from data to make decisions and predictions without explicit programming for every scenario. It's similar to teaching a child to identify objects through examples rather than memorizing specific rules for each item.&lt;/li&gt;
&lt;li&gt;Hypothesis Space 
This represents all possible models or solutions that can be derived from the available data—essentially all the potential ways to structure the data to solve the problem.&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;&lt;strong&gt;How It Works:&lt;/strong&gt;&lt;/p&gt;

&lt;ol&gt;
&lt;li&gt;
&lt;strong&gt;Problem&lt;/strong&gt;: First, identify your specific problem (such as predicting house prices).&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;Data&lt;/strong&gt;: Gather relevant data (such as number of rooms, house size, and location).&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;Hypothesis Space&lt;/strong&gt;: Create various potential models or rules to predict the price using different data combinations. This space contains all possible solutions that could solve your problem.&lt;/li&gt;
&lt;/ol&gt;

&lt;p&gt;&lt;strong&gt;4. Objective Function&lt;/strong&gt;&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;
&lt;strong&gt;Loss Function:&lt;/strong&gt; Measures how far off the model's predictions are from actual values, helping guide the learning process (e.g., Mean Squared Error for regression).&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;Optimization Algorithm:&lt;/strong&gt; Fine-tunes the model's parameters to reduce errors by minimizing the loss function (e.g., Gradient Descent).&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;&lt;strong&gt;5. Evaluation Metrics&lt;/strong&gt;&lt;/p&gt;

&lt;p&gt;Used to assess model performance, such as accuracy, precision, recall, F1-score, or mean absolute error.&lt;/p&gt;

&lt;h2&gt;
  
  
  Types of Machine Learning
&lt;/h2&gt;

&lt;p&gt;For understanding the types of machine learning , let’s first brush up on types of datasets :&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;Labelled Data :&lt;/strong&gt;&lt;/p&gt;

&lt;p&gt;Labelled data is the type of data that is assigned a label or category indicating its correct classification. Usually , this labelling is performed by human annotators. Using the labels , the model learns from labelled examples to make predictions on new, unseen data. &lt;/p&gt;

&lt;p&gt;Example: &lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;A dataset of images with labels indicating whether each image contains a cat or a dog.&lt;/li&gt;
&lt;li&gt;An email dataset labelled as spam or not spam.&lt;/li&gt;
&lt;li&gt;A dataset of customer reviews labelled with sentiment (positive, negative, neutral).&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;Labeled data helps train models for classification, regression, and object detection tasks by predicting specific values for each data point. Though valuable, obtaining labeled data is expensive and time-consuming since it requires human annotators to manually assign labels.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;Unlabeled Data :&lt;/strong&gt;&lt;/p&gt;

&lt;p&gt;As the name suggests , unlabeled data does not have any category associated with it . Thus ,the true classification of a data point remains unknown .The model must learn from the inherent structure of the data to uncover patterns or anomalies .&lt;/p&gt;

&lt;p&gt;Example:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;A dataset of customer transactions without any labels indicating fraudulent or non-fraudulent transactions.&lt;/li&gt;
&lt;li&gt;A collection of text documents without any labels indicating the topic or category of each document.&lt;/li&gt;
&lt;li&gt;An image dataset without any labels indicating the content or objects in each image.&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;&lt;strong&gt;Supervised Learning&lt;/strong&gt;&lt;br&gt;
The main differentiating feature of Supervised Learning is that it makes use of labelled data. This data acts as a supervisor for the algorithm for classification or prediction tasks . Using labelled data , the prediction and accuracy of the model can be easily measured . &lt;/p&gt;

&lt;p&gt;Supervised learning consists of two main types of problems in data mining: classification and regression:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;
&lt;strong&gt;Classification&lt;/strong&gt; involves algorithms that sort data into specific categories—like distinguishing apples from oranges. In practical applications, these algorithms can filter spam emails from legitimate ones. Common classification algorithms include linear classifiers, support vector machines, decision trees, and random forest&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;Regression&lt;/strong&gt; algorithms analyze relationships between dependent and independent variables. They excel at predicting numerical values—for instance, forecasting a company's sales revenue. Common approaches include linear regression, logistic regression, and polynomial regression.&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;Obtaining labelled data for Supervised Learning can be expensive and time-consuming, as it requires human annotators to assign labels to each data point.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;Unsupervised Learning&lt;/strong&gt;&lt;/p&gt;

&lt;p&gt;Unsupervised Learning uses algorithms to analyze and cluster unlabeled datasets, discovering hidden patterns without human intervention (hence the term "unsupervised").&lt;br&gt;
Unsupervised learning models perform three main tasks: clustering, association, and dimensionality reduction:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;
&lt;strong&gt;Clustering&lt;/strong&gt; groups unlabeled data points based on their similarities or differences. A common example is K-means clustering, which organizes similar data points into groups, with K determining the number of clusters. This technique proves valuable for applications like market segmentation and image compression.&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;Association&lt;/strong&gt; discovers relationships between variables in a dataset using pattern-finding rules. This approach powers features like "Customers Who Bought This Item Also Bought" recommendations and market basket analysis, helping identify products frequently purchased together.&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;Dimensionality reduction&lt;/strong&gt; simplifies complex datasets by reducing the number of features while maintaining essential information. This technique is particularly useful during data preprocessing—for instance, when autoencoders clean up visual data by removing noise to enhance image quality.&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;&lt;strong&gt;Difference between the two&lt;/strong&gt;&lt;/p&gt;

&lt;p&gt;The primary difference between supervised and unsupervised learning lies in the use of labeled datasets. Simply put, supervised learning relies on labeled input and output data, while unsupervised learning works without labeled data.&lt;/p&gt;

&lt;p&gt;In supervised learning, the algorithm learns by iteratively analyzing the training dataset, making predictions, and adjusting its outputs to match the correct answers. While these models are typically more accurate than unsupervised ones, they require significant human effort upfront to label the data properly. For instance, a supervised learning model can predict your commute time based on factors like the time of day and weather conditions. However, it first needs to be trained to recognize that rainy weather increases travel time.&lt;/p&gt;

&lt;p&gt;Unsupervised learning, on the other hand, independently identifies patterns or structures in unlabeled data. While these models don’t require labeled inputs, they still need human intervention to validate their findings. For example, an unsupervised learning model might detect that online shoppers frequently buy certain products together. A data analyst would then confirm if it makes sense for a recommendation system to group baby clothes with items like diapers, medicines, and ketchup.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;Choosing between the two&lt;/strong&gt;&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;
&lt;strong&gt;Evaluate your input data:&lt;/strong&gt; Is it labeled or unlabeled data? Do you have experts that can support extra labeling?&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;Define your goals:&lt;/strong&gt; Do you have a recurring, well-defined problem to solve? Or will the algorithm need to predict new problems?&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;Review your options for algorithms:&lt;/strong&gt; Are there algorithms with the same dimensionality that you need (number of features, attributes, or characteristics)? Can they support your data volume and structure?&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;&lt;strong&gt;The Best of Both Worlds&lt;/strong&gt;&lt;/p&gt;

&lt;p&gt;Struggling to choose between supervised and unsupervised learning? Semi-supervised learning offers a perfect balance by combining both labeled and unlabeled data in the training process. It’s especially helpful when extracting relevant features is challenging or when dealing with large datasets.&lt;/p&gt;

&lt;p&gt;This approach works wonders in fields like medical imaging, where even a small amount of labeled data can dramatically boost accuracy. For instance, a radiologist could label a handful of CT scans to identify tumors or diseases, enabling the machine to more reliably predict which patients might need closer medical attention.&lt;/p&gt;

</description>
      <category>machinelearning</category>
      <category>newbie</category>
    </item>
    <item>
      <title>Comparing Technologies for Object Identification and Classification: A Review of AI Approaches</title>
      <dc:creator>Anisha Bhandare</dc:creator>
      <pubDate>Fri, 27 Dec 2024 10:03:26 +0000</pubDate>
      <link>https://dev.to/anisha_bhandare_89bb0637f/comparing-technologies-for-object-identification-and-classification-a-review-of-ai-approaches-32ee</link>
      <guid>https://dev.to/anisha_bhandare_89bb0637f/comparing-technologies-for-object-identification-and-classification-a-review-of-ai-approaches-32ee</guid>
      <description>&lt;p&gt;&lt;strong&gt;Abstract&lt;/strong&gt;&lt;br&gt;
Object identification and classification are extremely important in all industries, whether it's healthcare, manufacturing, or environmental science. The traditional ways of doing it are pretty slow and require specific expertise. However, with the emergence of AI, ML, and DL, object identification is becoming faster and much more accurate. This review focuses on the current tech for object identification by considering CNNs, ensemble learning, and other DL models.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;Introduction&lt;/strong&gt;&lt;br&gt;
Object identification is a great application in quality control, environmental monitoring, and medical diagnostics. Manual identification is time-consuming and prone to errors. AI transformed the process by automating object recognition through image processing and deep learning. This paper discusses the various technologies as compared to their methods, accuracy, and practical applications.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;Machine Learning Approaches&lt;/strong&gt;&lt;/p&gt;

&lt;p&gt;There have been vast applications of ML towards object detection and classification. Such methodologies start through extraction of image features, such as colors, textures, and shapes and apply different algorithms for classification, Support Vector Machines and Random Forests, where in majority instances; an accuracy level of 85 to 93 percent is achieved. Although ML models may produce sound outcomes, they rely much on hand feature extraction. This can scale its ability across different datasets.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;DL Approaches&lt;/strong&gt; &lt;/p&gt;

&lt;p&gt;DL model CNNs is revolutionizing object identification. It eliminates the need for manual feature extraction and improves accuracy. Since CNNs can handle very large datasets and high-resolution images, it will be best suited for even the most complex object recognition tasks.&lt;br&gt;
Deep CNNs, usually combined with GAP, have achieved accuracy of over 99%. Such models are particularly good for applications where high accuracy is critical, such as medical imaging and defect detection. However, the price is that they consume much more computational resources and require larger datasets to train.&lt;/p&gt;

&lt;p&gt;MobileNet is a lightweight CNN that balances performance and efficiency. It is designed to perform object identification in real-time and has achieved 98.3% accuracy, thus aligning well with mobile and cloud platforms. It is quite suitable for edge devices and on-the-go applications but needs regular internet access for cloud processing.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;Ensemble Learning Techniques&lt;/strong&gt;&lt;/p&gt;

&lt;p&gt;Ensemble learning involves the combination of multiple models for improved overall performance. It increases accuracy up to about 97-99% as it is less prone to overfitting and helps fill the gap for weaker models. In the ensemble method, a number of different models are integrated, so there is strong performance with respect to different datasets. The deployment and management of an ensemble model, however, is challenging, which requires more coordination and computing power.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;Applications in Object Detection&lt;/strong&gt;&lt;/p&gt;

&lt;p&gt;DL is not only used for object ID; it has also been largely used for defect detection, medical imaging, and surveillance. CNNs and DBNs have found great applications in the recognition and classification of objects in different environments. The techniques of XAI are also being used to understand how such models make decisions.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;Comparative Analysis&lt;/strong&gt;&lt;br&gt;
The traditional ML approaches include SVM and RF. Those typically yield more than 85-93% accuracy, work for most relatively trivial tasks, and demand quite labor-intensive engineering of features; otherwise, their applications would be confined.CNN-based DL approaches extend to pushing the accuracy into 95-99%. With fully automatic feature extraction, a large, diversified set might be handled; it's a highly versatile tool, however very computationally expensive. Ensemble learning offers something of the middle ground: a combination of several models for an accuracy between 97 to 99%. While working efficiently, it also introduces model management issues in the deployment. For real-time cases, MobileNet delivers very good performance with an accuracy value of 98.3%. Though being very lightweight for mobility and cloud-based systems, this does have some inherent flaws: internet dependency within such areas may be problematic.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;Conclusion&lt;/strong&gt;&lt;br&gt;
DL, especially CNNs, has taken object identification and detection to a whole new level of precision and efficiency over traditional ML. Ensemble learning improves the reliability, but issues such as data availability and computing power remain. Future research would include lightweight DL models, larger datasets, and enhancing interpretability with explainable AI.&lt;/p&gt;

</description>
      <category>machinelearning</category>
      <category>cnn</category>
      <category>deeplearning</category>
      <category>xai</category>
    </item>
  </channel>
</rss>
