<?xml version="1.0" encoding="UTF-8"?>
<rss version="2.0" xmlns:atom="http://www.w3.org/2005/Atom" xmlns:dc="http://purl.org/dc/elements/1.1/">
  <channel>
    <title>DEV Community: Ruthvik Rao</title>
    <description>The latest articles on DEV Community by Ruthvik Rao (@gitruthvik).</description>
    <link>https://dev.to/gitruthvik</link>
    <image>
      <url>https://media2.dev.to/dynamic/image/width=90,height=90,fit=cover,gravity=auto,format=auto/https:%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Fuser%2Fprofile_image%2F406220%2F0687b181-8f27-4d4b-a67a-7bad89428242.png</url>
      <title>DEV Community: Ruthvik Rao</title>
      <link>https://dev.to/gitruthvik</link>
    </image>
    <atom:link rel="self" type="application/rss+xml" href="https://dev.to/feed/gitruthvik"/>
    <language>en</language>
    <item>
      <title>YOLOv8 Already? How is it better than v5, try it to see!</title>
      <dc:creator>Ruthvik Rao</dc:creator>
      <pubDate>Sat, 04 Feb 2023 10:40:51 +0000</pubDate>
      <link>https://dev.to/gitruthvik/yolov8-already-how-is-it-better-than-v5-try-it-to-see-4g5o</link>
      <guid>https://dev.to/gitruthvik/yolov8-already-how-is-it-better-than-v5-try-it-to-see-4g5o</guid>
      <description>&lt;p&gt;YOLO (You Only Look Once) is a popular object detection algorithm used for computer vision applications. The latest version of YOLO, YOLOv8, was released in 2021 and it represents a major upgrade over its predecessor, YOLOv5. In this blog post, we will compare the performance and upgrades of YOLOv8 over YOLOv5.&lt;/p&gt;

&lt;p&gt;Performance:&lt;br&gt;
YOLOv8 provides improved performance compared to YOLOv5. This is due to several factors, including the use of a more efficient architecture, the addition of extra convolutional layers, and the use of anchor-based object detection. YOLOv8 also provides faster processing speeds, making it more suitable for real-time object detection applications.&lt;/p&gt;

&lt;p&gt;Upgrades:&lt;/p&gt;

&lt;ol&gt;
&lt;li&gt;&lt;p&gt;Improved Architecture: YOLOv8 introduces an updated architecture that is more efficient and accurate compared to YOLOv5. This new architecture uses a combination of residual blocks, bottleneck blocks, and inverted residual blocks to improve the accuracy and efficiency of the model.&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;Anchor-based Object Detection: YOLOv8 introduces anchor-based object detection, which provides more accurate and precise object detection compared to YOLOv5. Anchor-based object detection uses anchor boxes to predict the location and size of objects in the image.&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;More Convolutional Layers: YOLOv8 introduces extra convolutional layers, which provide the model with more capacity to learn and improve accuracy.&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;Improved Training: YOLOv8 introduces a new training regime that allows the model to learn more efficiently and achieve higher accuracy. This includes the use of a larger dataset and the use of transfer learning to fine-tune the model.&lt;/p&gt;&lt;/li&gt;
&lt;/ol&gt;

&lt;p&gt;In short, YOLOv8 represents a major upgrade over YOLOv5 in terms of performance and accuracy. The improved architecture, anchor-based object detection, extra convolutional layers, and improved training regime all contribute to the improved performance of YOLOv8. If you are looking for a powerful and efficient object.&lt;/p&gt;

</description>
      <category>devrel</category>
      <category>writing</category>
      <category>announcement</category>
      <category>webmonetization</category>
    </item>
    <item>
      <title>ChatGPT! How and Where?</title>
      <dc:creator>Ruthvik Rao</dc:creator>
      <pubDate>Sat, 04 Feb 2023 09:53:59 +0000</pubDate>
      <link>https://dev.to/gitruthvik/chatgpt-how-and-where-3395</link>
      <guid>https://dev.to/gitruthvik/chatgpt-how-and-where-3395</guid>
      <description>&lt;p&gt;&lt;strong&gt;ChatGPT&lt;/strong&gt; is a powerful &lt;em&gt;language model&lt;/em&gt; developed by &lt;strong&gt;OpenAI&lt;/strong&gt; that uses advanced machine learning techniques to generate human-like text. This model was trained on a massive amount of data and fine-tuned to generate high-quality responses to a wide range of questions and prompts.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;How ChatGPT Works:&lt;/strong&gt;&lt;/p&gt;

&lt;ol&gt;
&lt;li&gt;&lt;p&gt;Preprocessing: The input text is passed through a preprocessing step to convert it into a format that can be easily processed by the model. This step involves tokenizing the text into individual words or phrases and encoding them using a numerical representation.&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;Encoder-Decoder Architecture: ChatGPT uses a transformer-based encoder-decoder architecture to generate its outputs. The encoder takes the input text and converts it into a high-dimensional vector representation, while the decoder uses this representation to generate the output text.&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;Attention Mechanisms: ChatGPT uses attention mechanisms to focus on different parts of the input text when generating its output. This allows the model to generate more contextually relevant responses by taking into account the context of the input text.&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;Generative Process: The decoder then generates the output text one word at a time, using a process called autoregression. The model uses the previously generated words and the input representation to generate the next word in the sequence.&lt;/p&gt;&lt;/li&gt;
&lt;/ol&gt;

&lt;p&gt;&lt;strong&gt;Uses of ChatGPT:&lt;/strong&gt;&lt;/p&gt;

&lt;ol&gt;
&lt;li&gt;&lt;p&gt;Chatbots: ChatGPT can be used to build advanced chatbots that can respond to customer inquiries in a natural and human-like manner.&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;Content Creation: ChatGPT can be used to generate articles, summaries, and other types of text content, making it a useful tool for content creators and marketers.&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;Text Generation: ChatGPT can be used to generate creative writing, poetry, and even jokes, making it a versatile tool for writers and entertainers.&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;Question Answering: ChatGPT can be used to build question-answering systems that can provide accurate answers to complex questions in a variety of domains.&lt;/p&gt;&lt;/li&gt;
&lt;/ol&gt;

&lt;p&gt;In conclusion, ChatGPT is a powerful language model that has a wide range of applications in the fields of natural language processing and artificial intelligence. With its ability to generate high-quality text responses and its versatility, ChatGPT has the potential to revolutionize the way we interact with computers and automate many tasks that previously required human input.&lt;br&gt;
As you all might have guessed it by now, this complete blog has been generated with ChatGPT! &lt;br&gt;
Go ahead and give it a try &lt;a href="https://chat.openai.com/"&gt;&lt;strong&gt;here&lt;/strong&gt;&lt;/a&gt;&lt;/p&gt;

</description>
      <category>transformers</category>
      <category>chatgpt</category>
      <category>textmodel</category>
      <category>uses</category>
    </item>
    <item>
      <title>Try out Face Detection in under 5 Minutes!</title>
      <dc:creator>Ruthvik Rao</dc:creator>
      <pubDate>Fri, 19 Jun 2020 11:19:21 +0000</pubDate>
      <link>https://dev.to/gitruthvik/try-out-face-detection-in-under-5-minutes-1j54</link>
      <guid>https://dev.to/gitruthvik/try-out-face-detection-in-under-5-minutes-1j54</guid>
      <description>&lt;p&gt;&lt;strong&gt;&lt;em&gt;Face Detection&lt;/em&gt;&lt;/strong&gt; in recent years have been made a lot easier with the &lt;strong&gt;&lt;em&gt;OpenCV&lt;/em&gt;&lt;/strong&gt; library. There is an extensive market for face detection like security, analytics etc. So lets cut the shit and get on with the code!&lt;/p&gt;

&lt;p&gt;Starting with the basics, You'll need to have the following:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;Python 3+(I'm pretty sure you have it installed if you are here)&lt;/li&gt;
&lt;li&gt;OpenCV

&lt;ul&gt;
&lt;li&gt;Easy Install with the pip command "&lt;code&gt;pip install opencv-python&lt;/code&gt;" and also "&lt;code&gt;pip install opencv-contrib-python&lt;/code&gt;"&lt;/li&gt;
&lt;/ul&gt;


&lt;/li&gt;
&lt;li&gt;The Open Sourced Trained Classifier File: &lt;a href="https://github.com/opencv/opencv/blob/master/data/haarcascades/haarcascade_frontalface_default.xml"&gt;haarcascade_frontalface_default.xml&lt;/a&gt; &lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;Here's the Code:&lt;br&gt;
First up, lets start by importing the opencv library&lt;br&gt;
&lt;code&gt;import cv2&lt;/code&gt;&lt;/p&gt;

&lt;p&gt;Then load the Haar Classifier file to the program&lt;br&gt;
&lt;code&gt;cascade=cv2.CascadeClassifier('haarcascade_frontalface_default.xml')&lt;/code&gt;&lt;/p&gt;

&lt;p&gt;Start to capture video from webcam. Change value from 0 to 1 if you are using an external webcam&lt;br&gt;
&lt;code&gt;cam=cv2.VideoCapture(0)&lt;/code&gt;&lt;/p&gt;

&lt;p&gt;Now start a loop statement to capture every frame in the video input&lt;br&gt;
&lt;code&gt;while True:&lt;/code&gt;&lt;br&gt;
    &lt;code&gt;_, img = cap.read()&lt;/code&gt;&lt;/p&gt;

&lt;p&gt;Then convert the input frame to greyscale, this improves the &lt;br&gt;
   accuracy of detection&lt;br&gt;
   &lt;code&gt;gryimg = cv2.cvtColor(img, cv2.COLOR_BGR2GRAY)&lt;/code&gt;&lt;/p&gt;

&lt;p&gt;Here comes the elephant, line that detects the faces&lt;br&gt;
    &lt;code&gt;face = face_cascade.detectMultiScale(gray, 1.1, 5)&lt;/code&gt;&lt;br&gt;
   The parameters for the function are the input image, Scale &lt;br&gt;
   Factor and the minNeighbors&lt;/p&gt;

&lt;p&gt;After detecting, the cascade model returns a &lt;strong&gt;&lt;em&gt;numpy &lt;br&gt;
   array&lt;/em&gt;&lt;/strong&gt; which are coordinate values where the face is &lt;br&gt;
   located on the frame. So lets draw a box over the faces:&lt;br&gt;
   `for (x, y, z, w, h) in face&lt;br&gt;
        cv2.rectangle(img, (x, y), (x+w, y+h), (255, 0, 0), 2)&lt;/p&gt;

&lt;p&gt;At last, Display the frames in window&lt;br&gt;
   &lt;code&gt;cv2.imshow("img",img)&lt;/code&gt;&lt;/p&gt;

&lt;p&gt;Then end the code with the escape button:&lt;br&gt;
   &lt;code&gt;k = cv2.waitkey(30) &amp;amp; 0xff&lt;/code&gt;&lt;br&gt;
   &lt;code&gt;if k==27:&lt;br&gt;
        &lt;/code&gt;&lt;code&gt;break&lt;/code&gt;&lt;/p&gt;

&lt;p&gt;And that's pretty much it!&lt;br&gt;
Try around and play with the code to make new apps such as a face counter or people detection alert system. Don't let anything stop, your creativity is the Limit :)&lt;br&gt;
Thanks for Reading, Do reach out to me for any queries!&lt;/p&gt;

</description>
      <category>python</category>
      <category>opencv</category>
      <category>tutorial</category>
      <category>opensource</category>
    </item>
  </channel>
</rss>
