<?xml version="1.0" encoding="UTF-8"?>
<rss version="2.0" xmlns:atom="http://www.w3.org/2005/Atom" xmlns:dc="http://purl.org/dc/elements/1.1/">
  <channel>
    <title>DEV Community: Samruddhi Sangale</title>
    <description>The latest articles on DEV Community by Samruddhi Sangale (@samrudhi09).</description>
    <link>https://dev.to/samrudhi09</link>
    <image>
      <url>https://media2.dev.to/dynamic/image/width=90,height=90,fit=cover,gravity=auto,format=auto/https:%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Fuser%2Fprofile_image%2F1191298%2Fd50a6cd9-9652-42ed-add6-899fec9f17be.jpeg</url>
      <title>DEV Community: Samruddhi Sangale</title>
      <link>https://dev.to/samrudhi09</link>
    </image>
    <atom:link rel="self" type="application/rss+xml" href="https://dev.to/feed/samrudhi09"/>
    <language>en</language>
    <item>
      <title>What is DeepSeek-R1?: Simple guide in 5 minutes</title>
      <dc:creator>Samruddhi Sangale</dc:creator>
      <pubDate>Sun, 16 Feb 2025 01:57:41 +0000</pubDate>
      <link>https://dev.to/samrudhi09/what-is-deepseek-r1-simple-guide-in-5-minutes-1o0p</link>
      <guid>https://dev.to/samrudhi09/what-is-deepseek-r1-simple-guide-in-5-minutes-1o0p</guid>
      <description>&lt;p&gt;In just 3-4 days, DeepSeek took over our Twitter feeds. I decided to dive deep into it, and while doing so, I wrote this blog based on my notes. So here is a simplified version of what is DeepSeek doing under the hood (not really under the hood because they are open source, haha).&lt;/p&gt;

&lt;p&gt;While OpenAI provides some of the best LLMs, DeepSeek is pushing the boundaries by improving reasoning capabilities using Reinforcement Learning (RL).&lt;/p&gt;

&lt;p&gt;They have two models:&lt;/p&gt;

&lt;p&gt;DeepSeek-R1-Zero: Trained purely on RL.&lt;/p&gt;

&lt;p&gt;DeepSeek-R1: Trained on RL and Supervised Fine-Tuning (SFT).&lt;/p&gt;

&lt;p&gt;What is RL? It’s a type of machine learning where an AI model learns by trial and error. It gets rewarded for good actions and penalized for bad ones, helping it improve over time. For example, in chess:&lt;/p&gt;

&lt;p&gt;The model (or agent) tries different moves (actions) to achieve a goal (e.g., checkmate).&lt;/p&gt;

&lt;p&gt;It doesn’t know the best move at every step but learns by experimenting and receiving feedback (rewards or penalties).&lt;/p&gt;

&lt;p&gt;The goal is to create a policy (a strategy) that maximizes the total reward over time, even if the immediate outcome of each action isn’t certain.&lt;/p&gt;

&lt;p&gt;What is SFT? SFT is a process where a model is further trained on a specific dataset with labeled examples (e.g., questions and correct answers). This helps the model specialize in certain tasks.&lt;/p&gt;

&lt;p&gt;Approach&lt;br&gt;
Let’s break down the three main approaches they used:&lt;/p&gt;

&lt;p&gt;DeepSeek-R1-Zero: Pure RL training on the base model.&lt;/p&gt;

&lt;p&gt;DeepSeek-R1: Multi-stage training with RL, SFT, and cold-start data.&lt;/p&gt;

&lt;p&gt;Distillation: Transferring reasoning capabilities to smaller models.&lt;/p&gt;

&lt;p&gt;DeepSeek-R1-Zero: Pure RL Training&lt;br&gt;
Reinforcement Learning Algorithm&lt;br&gt;
They use Group Relative Policy Optimization (GRPO),Here’s how it works:&lt;/p&gt;

&lt;p&gt;"Think of GRPO like a teacher grading a group of students. The teacher compares each student’s performance to the class average, ensuring no one is too far ahead or behind. This keeps the training stable and prevents drastic changes.”&lt;/p&gt;

&lt;p&gt;Group Sampling: For each question, the model generates a group of responses. The reward for each response is compared to the average reward of the group, which helps stabilize training.&lt;/p&gt;

&lt;p&gt;Advantage Calculation: The advantage Ai for each response is calculated as:&lt;/p&gt;

&lt;p&gt;Ai= (ri−mean({r1,r2,…,rG})) / std({r1,r2,…,rG})&lt;/p&gt;

&lt;p&gt;where ri​ is the reward for the i-th response, and G is the group size.&lt;/p&gt;

&lt;p&gt;Policy Update: The goal is to update the policy to maximize the expected reward while keeping the new policy close to the old one (to avoid drastic changes).&lt;/p&gt;

&lt;p&gt;Reward Modeling&lt;br&gt;
They use a rule-based reward system with two components:&lt;/p&gt;

&lt;p&gt;Accuracy Rewards: The model gets rewarded for correct answers. For example:&lt;/p&gt;

&lt;p&gt;In math problems, the final answer must be in a specific format (e.g., inside a box) for automated verification.&lt;/p&gt;

&lt;p&gt;In coding tasks, the code is compiled and tested against predefined test cases.&lt;/p&gt;

&lt;p&gt;Format Rewards: The model is rewarded for following the correct format, such as enclosing the reasoning process in  tags and the final answer in  tags.&lt;/p&gt;

&lt;p&gt;Let’s move to an important term: Chain of thoughts (CoT)&lt;br&gt;
What it is: CoT is a technique where the model explains its reasoning step-by-step before giving the final answer. For example, instead of just saying "2 + 2 = 4," the model might say, "First, I add 2 and 2, which equals 4."&lt;/p&gt;

&lt;p&gt;In DeepSeek, they encourage the model to use CoT to improve its reasoning and make its thought process more transparent.&lt;/p&gt;

&lt;p&gt;While using DeepSeek you can use the “think” button to check how the model is reaching to a particular response.&lt;/p&gt;

&lt;p&gt;DeepSeek-R1: Multi-Stage Training&lt;/p&gt;

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fv23dgp314mph3phb3lis.png" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fv23dgp314mph3phb3lis.png" alt="Image description" width="730" height="807"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;Cold Start&lt;br&gt;
To address the readability and language mixing issues of DeepSeek-R1-Zero, they introduced cold-start data:&lt;/p&gt;

&lt;p&gt;Data Collection: Thousands of high-quality reasoning examples are collected using few-shot prompting, model-generated outputs, and human annotation.&lt;/p&gt;

&lt;p&gt;Fine-Tuning: The base model (DeepSeek-V3-Base) is fine-tuned on this cold-start data before applying RL.&lt;/p&gt;

&lt;p&gt;Reasoning-Oriented RL&lt;br&gt;
After fine-tuning, the model undergoes RL training similar to DeepSeek-R1-Zero. However, they also introduce a language consistency reward to reduce language mixing:(chinese and english)&lt;/p&gt;

&lt;p&gt;Language Consistency: The reward is based on the proportion of target language words in the reasoning process. This aligns the model with human preferences for readability.&lt;br&gt;
Rejection Sampling and SFT&lt;br&gt;
Once the RL model converges, rejection sampling is used to generate high-quality training data:&lt;/p&gt;

&lt;p&gt;“Rejection sampling is a technique where the model generates multiple responses, and only the best ones (based on correctness and quality) are kept for further training”&lt;/p&gt;

&lt;p&gt;Reasoning Data: The model generates multiple responses for each prompt, and only the correct ones are retained. This results in ~600k reasoning-related training samples.&lt;/p&gt;

&lt;p&gt;Non-Reasoning Data: Additional data from domains like writing, factual QA, and self-cognition are collected, totaling ~200k samples.&lt;/p&gt;

&lt;p&gt;Fine-Tuning: The base model is fine-tuned on this combined dataset for two epochs.&lt;/p&gt;

&lt;p&gt;RL for All Scenarios&lt;br&gt;
A second RL stage is applied to align the model with human preferences across all scenarios:&lt;/p&gt;

&lt;p&gt;Helpfulness and Harmlessness: The model is trained to prioritize helpful and harmless responses, using reward models to capture human preferences in complex scenarios.&lt;/p&gt;

&lt;p&gt;Distillation: They also introduce distillation but we won’t go deep into it. Distillation in basic terms is the process that transfers reasoning capabilities from larger models to smaller models( which are faster and cheaper).&lt;/p&gt;

&lt;p&gt;The Magic of Self-Evolution&lt;br&gt;
The most magical part about their models is self-evolution. The model naturally develops advanced reasoning behaviors like reflection (revisiting previous steps) and long-chain reasoning (generating hundreds of reasoning tokens). These behaviors emerge without explicit programming, showcasing the power of RL.&lt;/p&gt;

&lt;p&gt;That’s all!&lt;/p&gt;

&lt;p&gt;As a result, DeepSeek’s R1 model is on par with OpenAI’s o1 series models in multiple benchmarks(and it’s free). Hope this was simple and clear enough to get a basic idea of how it’s working.&lt;/p&gt;

&lt;p&gt;If you’re curious to learn more, I highly recommend reading their paper &lt;a href="https://arxiv.org/abs/2501.12948" rel="noopener noreferrer"&gt;here.&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;Feel free to reach out to me on: &lt;a href="https://x.com/Samruddhi0909" rel="noopener noreferrer"&gt;X(Twitter)&lt;/a&gt; | &lt;a href="https://www.linkedin.com/in/samruddhi-sangale" rel="noopener noreferrer"&gt;LinkedIn&lt;/a&gt;&lt;/p&gt;

</description>
    </item>
    <item>
      <title>The (relatively) unknown hot job in the tech town: DevRel</title>
      <dc:creator>Samruddhi Sangale</dc:creator>
      <pubDate>Wed, 17 Jan 2024 19:02:16 +0000</pubDate>
      <link>https://dev.to/samrudhi09/the-relatively-unknown-hot-job-in-the-tech-town-devrel-2gcd</link>
      <guid>https://dev.to/samrudhi09/the-relatively-unknown-hot-job-in-the-tech-town-devrel-2gcd</guid>
      <description>&lt;h2&gt;
  
  
  What is it all about and is it for you?
&lt;/h2&gt;

&lt;p&gt;I first stumbled upon this word when I saw a senior from my college share about it on her socials. I was immediately fascinated by this new job title and started reading all about Developer Relations(DevRel). So here's my attempt at explaining it to you (and me haha).&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;&lt;em&gt;What is DevRel?&lt;/em&gt;&lt;/strong&gt;&lt;br&gt;
DevRel is a broader term that covers many job roles that are majorly associated with creating a positive relationship with the developer community of a company offering a developer tool.&lt;/p&gt;

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fdspmttbosk1wdbynge38.jpg" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fdspmttbosk1wdbynge38.jpg" alt="Image description" width="204" height="204"&gt;&lt;/a&gt;&lt;br&gt;
Let me explain!&lt;/p&gt;

&lt;p&gt;A developer tool (Dev tool) could be anything that aids in building, testing, and debugging software applications for developers -for example, APIs, IDEs, frameworks, package managers, version control systems, etc.&lt;/p&gt;

&lt;p&gt;The developers using these tools form it's community and well DevRel is the medium through which the bond between the company and community can be made stronger. DevRel should make it easier for both parties to fulfill each of their purposes. It should help developers make the most out of the dev tool for quicker, easier development and help the product company build a trusted brand within the community so that more people build with their products. There are various ways and roles in DevRel to do this. For instance, community managers, developer advocates, developer evangelists, developer experience, etc&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;&lt;em&gt;What are the responsibilities of a DevRel?&lt;/em&gt;&lt;/strong&gt;&lt;/p&gt;

&lt;p&gt;There are various tasks and responsibilities of a DevRel professional. This job position is multifaceted and hence its definition differs from company to company. It might include writing articles and representing the company on various platforms( Twitter spaces, podcasts, etc), Documentation, Building demos, Creating tutorials, managing the socials of the company, talking at various conferences, conducting meetups, creating content across platforms, networking, etc.&lt;/p&gt;

&lt;p&gt;Well, that looks like a lot of work! It sure is but it also comes with its perks.&lt;/p&gt;

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fmpkeompzpgf39ti6b6n6.gif" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fmpkeompzpgf39ti6b6n6.gif" alt="Image description" width="480" height="400"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;(Also that's what she said :')&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;&lt;em&gt;The good part :&lt;/em&gt;&lt;/strong&gt;&lt;/p&gt;

&lt;p&gt;You get to form meaningful connections in the community, might get a chance to travel a lot for conferences across the world, are always updated with the latest products and tech, and more importantly, bring out your creative side along with the technical aspect.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;&lt;em&gt;Is it for me?&lt;/em&gt;&lt;/strong&gt;&lt;/p&gt;

&lt;p&gt;"All that is fine, but is it for me?" Naturally, you would think it looks like a lot of socializing (which most developers hate) but I think it's for anyone and everyone passionate about taking the community forward. If you get excited by the launch of a new product and how it will pan out with the developers and want to help bridge the gap, it just might be the perfect fit for you. Having said that, there are certain solid skills that you require, some might come naturally to you, and some can be built over time.&lt;/p&gt;

&lt;ol&gt;
&lt;li&gt;&lt;p&gt;&lt;strong&gt;Networking&lt;/strong&gt;&lt;br&gt;
The community should trust you and thus the product/tech you are advocating for. Thus, forming meaningful connections is important.&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;&lt;strong&gt;Coding and software engineering basics.&lt;/strong&gt;&lt;br&gt;
As you will be helping developers, it is important to understand their language. Technical knowledge around a particular niche is always beneficial.&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;&lt;strong&gt;Public speaking&lt;/strong&gt;&lt;br&gt;
Your company might want you to go talk to various conferences and represent them.&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;&lt;strong&gt;A genuine interest in helping developers&lt;/strong&gt;&lt;br&gt;
This has to be the main intention behind all of your actions. You should put value in the community.&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;&lt;strong&gt;Technical writing&lt;/strong&gt;&lt;br&gt;
To write articles, or write documentation, or tutorials on various topics.&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;&lt;strong&gt;Listening skills&lt;/strong&gt;&lt;br&gt;
You need to listen to the feedback, convey it to the product team, and create a loop. For that, you should be a patient listener!&lt;/p&gt;&lt;/li&gt;
&lt;/ol&gt;

&lt;p&gt;Well, I hope this gave you an overview of DevRel. Thanks for reading till the end!&lt;/p&gt;

&lt;p&gt;PS: Any feedback on this article is highly appreciated.&lt;/p&gt;

&lt;p&gt;Connect with me -&lt;/p&gt;

&lt;p&gt;&lt;a href="https://twitter.com/Samruddhi0909" rel="noopener noreferrer"&gt;X(Twitter)&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;&lt;a href="https://www.linkedin.com/in/samruddhi-sangale-21680a1b1/" rel="noopener noreferrer"&gt;Linkedin&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;&lt;a href="https://github.com/samrudhi0909" rel="noopener noreferrer"&gt;Github&lt;/a&gt;&lt;/p&gt;

</description>
      <category>devrel</category>
      <category>beginners</category>
      <category>dxmentorship</category>
    </item>
  </channel>
</rss>
