<?xml version="1.0" encoding="UTF-8"?>
<rss version="2.0" xmlns:atom="http://www.w3.org/2005/Atom" xmlns:dc="http://purl.org/dc/elements/1.1/">
  <channel>
    <title>DEV Community: Saiprasad Balasubramanian</title>
    <description>The latest articles on DEV Community by Saiprasad Balasubramanian (@backtrackbaba).</description>
    <link>https://dev.to/backtrackbaba</link>
    <image>
      <url>https://media2.dev.to/dynamic/image/width=90,height=90,fit=cover,gravity=auto,format=auto/https:%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Fuser%2Fprofile_image%2F20842%2F218db418-d01d-41fa-a1ba-cd2c99a3701c.jpg</url>
      <title>DEV Community: Saiprasad Balasubramanian</title>
      <link>https://dev.to/backtrackbaba</link>
    </image>
    <atom:link rel="self" type="application/rss+xml" href="https://dev.to/feed/backtrackbaba"/>
    <language>en</language>
    <item>
      <title>Covid19 REST API's</title>
      <dc:creator>Saiprasad Balasubramanian</dc:creator>
      <pubDate>Mon, 23 Mar 2020 12:22:18 +0000</pubDate>
      <link>https://dev.to/backtrackbaba/covid19-rest-api-s-2ehe</link>
      <guid>https://dev.to/backtrackbaba/covid19-rest-api-s-2ehe</guid>
      <description>&lt;p&gt;The team at &lt;a href="https://systems.jhu.edu/"&gt;Johns Hopkins University CSSE&lt;/a&gt; has been maintaining a &lt;a href="https://github.com/CSSEGISandData/COVID-19"&gt;database&lt;/a&gt; of cases across the globe in CSV files. &lt;a href="https://github.com/pomber/covid19"&gt;Pomber&lt;/a&gt;, a user on Github has recently converted the same dataset into a &lt;a href="https://pomber.github.io/covid19/timeseries.json"&gt;JSON file&lt;/a&gt;.&lt;/p&gt;

&lt;p&gt;My friends - &lt;a href="https://www.linkedin.com/in/hrkj-18"&gt;Harsh Jain&lt;/a&gt;, &lt;a href="https://www.linkedin.com/in/girisha-navani-87065215b/"&gt;Girisha Navani&lt;/a&gt; and &lt;a href="https://www.linkedin.com/in/saiprasadbala/"&gt;I&lt;/a&gt; took it up as a learning opportunity and decided to work on the data instead of building yet another dashboard. We took the initiative of building a heavily cached, blazingly fast REST interface on top of the data to enable other developers.&lt;/p&gt;


&lt;div class="ltag-github-readme-tag"&gt;
  &lt;div class="readme-overview"&gt;
    &lt;h2&gt;
      &lt;img src="https://res.cloudinary.com/practicaldev/image/fetch/s--vJ70wriM--/c_limit%2Cf_auto%2Cfl_progressive%2Cq_auto%2Cw_880/https://practicaldev-herokuapp-com.freetls.fastly.net/assets/github-logo-ba8488d21cd8ee1fee097b8410db9deaa41d0ca30b004c0c63de0a479114156f.svg" alt="GitHub logo"&gt;
      &lt;a href="https://github.com/backtrackbaba"&gt;
        backtrackbaba
      &lt;/a&gt; / &lt;a href="https://github.com/backtrackbaba/covid-api"&gt;
        covid-api
      &lt;/a&gt;
    &lt;/h2&gt;
    &lt;h3&gt;
      API's to interact with COVID19 dataset by John Hopkins University
    &lt;/h3&gt;
  &lt;/div&gt;
  &lt;div class="ltag-github-body"&gt;
    
&lt;div id="readme" class="md"&gt;
&lt;h1&gt;
Covid API&lt;/h1&gt;
&lt;h1&gt;
Introduction&lt;/h1&gt;
&lt;p&gt;This project builds upon the dataset of &lt;a href="https://github.com/CSSEGISandData/COVID-19"&gt;John Hopkins University&lt;/a&gt; in CSV form which was converted to JSON Time Series format by &lt;a href="https://github.com/pomber/covid19"&gt;pomber&lt;/a&gt;.&lt;/p&gt;
&lt;p&gt;Our project intends to make that set queryable in a manner in which it could be easily consumed to build public dashboards.&lt;/p&gt;
&lt;h1&gt;
Overview&lt;/h1&gt;
&lt;p&gt;Analyzing the dataset, here are the major points that we came across.&lt;/p&gt;
&lt;p&gt;The API's have been mapped to use &lt;a href="https://en.wikipedia.org/wiki/ISO_3166" rel="nofollow"&gt;ISO 3166&lt;/a&gt; standard to query countries instead of names as in the source datasets built upon as it wasn't in a standard format.&lt;/p&gt;
&lt;p&gt;The API's consume &amp;amp; return dates as per &lt;a href="https://en.wikipedia.org/wiki/ISO_8601" rel="nofollow"&gt;ISO 8601&lt;/a&gt; standards in &lt;code&gt;yyyy-mm-dd&lt;/code&gt; format  The dates have been normalized from the underlying dataset by padding single digits in the date and month section.&lt;/p&gt;
&lt;h1&gt;
Authentication&lt;/h1&gt;
&lt;p&gt;There's no authentication required. Anybody and everybody is welcome to use this widely.&lt;/p&gt;
&lt;h1&gt;
Rate limit&lt;/h1&gt;
&lt;p&gt;There is no rate limit of any kind but…&lt;/p&gt;
&lt;/div&gt;
  &lt;/div&gt;
  &lt;div class="gh-btn-container"&gt;&lt;a class="gh-btn" href="https://github.com/backtrackbaba/covid-api"&gt;View on GitHub&lt;/a&gt;&lt;/div&gt;
&lt;/div&gt;


&lt;p&gt;Our endpoints are capable of the following:&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;Country specific historic data:&lt;/strong&gt; Right from 22nd January to the latest updated day.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;Country specific data for a particular date:&lt;/strong&gt; This gives you an insight into the spread of the virus in a country on a particular date.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;Latest Global Count:&lt;/strong&gt; Gives you the latest updated count of the virus globally&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;Global count on a particular date:&lt;/strong&gt; This data gives you a count of the virus globally, on a particular date&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;Global count in a date range:&lt;/strong&gt; This gives you the number of cases in a date range &lt;/p&gt;

&lt;p&gt;Our project internally consumes Pomber's JSON data and cleans it a bit by changing all the references of Country names to their respective three-character ISO codes and the dates have been standardized to adhere to ISO-8601 specs of &lt;code&gt;yyyy-mm-dd&lt;/code&gt;&lt;/p&gt;

&lt;p&gt;Additionally, we have taken care to ensure that our APIs are versioned so that they do not affect any downstream users if our upstream data changes. &lt;br&gt;
Johns Hopkins, has yesterday said that their data is going to change a bit which means we too would move on to v2 of our API's by adapting to changes at the same time still maintaining the v1 endpoints.&lt;/p&gt;


&lt;div class="ltag_github-liquid-tag"&gt;
  &lt;h1&gt;
    &lt;a href="https://github.com/CSSEGISandData/COVID-19/issues/1250"&gt;
      &lt;img class="github-logo" alt="GitHub logo" src="https://res.cloudinary.com/practicaldev/image/fetch/s--vJ70wriM--/c_limit%2Cf_auto%2Cfl_progressive%2Cq_auto%2Cw_880/https://practicaldev-herokuapp-com.freetls.fastly.net/assets/github-logo-ba8488d21cd8ee1fee097b8410db9deaa41d0ca30b004c0c63de0a479114156f.svg"&gt;
      &lt;span class="issue-title"&gt;
        Upcoming changes in time series tables
      &lt;/span&gt;
      &lt;span class="issue-number"&gt;#1250&lt;/span&gt;
    &lt;/a&gt;
  &lt;/h1&gt;
  &lt;div class="github-thread"&gt;
    &lt;div class="timeline-comment-header"&gt;
      &lt;a href="https://github.com/CSSEGISandData"&gt;
        &lt;img class="github-liquid-tag-img" src="https://res.cloudinary.com/practicaldev/image/fetch/s--64CXwEBU--/c_limit%2Cf_auto%2Cfl_progressive%2Cq_auto%2Cw_880/https://avatars1.githubusercontent.com/u/60674295%3Fv%3D4" alt="CSSEGISandData avatar"&gt;
      &lt;/a&gt;
      &lt;div class="timeline-comment-header-text"&gt;
        &lt;strong&gt;
          &lt;a href="https://github.com/CSSEGISandData"&gt;CSSEGISandData&lt;/a&gt;
        &lt;/strong&gt; posted on &lt;a href="https://github.com/CSSEGISandData/COVID-19/issues/1250"&gt;&lt;time&gt;Mar 22, 2020&lt;/time&gt;&lt;/a&gt;
      &lt;/div&gt;
    &lt;/div&gt;
    &lt;div class="ltag-github-body"&gt;
      &lt;p&gt;We will update the time series tables in the following days, aiming to provide a cleaner and more organized dataset consistent with our new/current naming convention. We will also be reporting a new variable (i.e, testing), as well as data at the county level for the US. All files will continue to be updated daily around 11:59PM UTC.&lt;/p&gt;
&lt;p&gt;The followiing specific changes will be made:&lt;/p&gt;
&lt;ul&gt;
&lt;li&gt;
&lt;p&gt;Three new time series tables will be added for the US. The first two will be the confirmed cases and deaths, reported at the county level. The third, number of tests conducted, will be reported at the state level. These new tables will be named &lt;code&gt;time_series_covid19_confirmed_US.csv&lt;/code&gt;, &lt;code&gt;time_series_covid19_deaths_US.csv&lt;/code&gt;, &lt;code&gt;time_series_covid19_testing_US.csv&lt;/code&gt;, respectively.&lt;/p&gt;
&lt;/li&gt;
&lt;li&gt;
&lt;p&gt;Changes to the current time series include the removal of the US state and county-level entries, which will be replaced with a new single country level entry for the US. The tables will be renamed  &lt;code&gt;time_series_covid19_confirmed_global.csv&lt;/code&gt; and &lt;code&gt;time_series_covid19_deaths_global.csv&lt;/code&gt;, and &lt;code&gt;time_series_covid19_testing_global.csv&lt;/code&gt;, respectively.&lt;/p&gt;
&lt;/li&gt;
&lt;li&gt;
&lt;p&gt;The ISO code will be added in the global time series tables.&lt;/p&gt;
&lt;/li&gt;
&lt;li&gt;
&lt;p&gt;The FIPS code will be added in the new US time series tables.&lt;/p&gt;
&lt;/li&gt;
&lt;li&gt;
&lt;p&gt;We will no longer provide recovered cases.&lt;/p&gt;
&lt;/li&gt;
&lt;li&gt;
&lt;p&gt;The current set of time series files will be moved to our archive folder, and the new files will be added to the current folder.&lt;/p&gt;
&lt;/li&gt;
&lt;/ul&gt;
&lt;p&gt;Thanks!&lt;/p&gt;
&lt;p&gt;Update: &lt;code&gt;time_series_covid19_recovered_global.csv&lt;/code&gt; is added.&lt;/p&gt;

    &lt;/div&gt;
    &lt;div class="gh-btn-container"&gt;&lt;a class="gh-btn" href="https://github.com/CSSEGISandData/COVID-19/issues/1250"&gt;View on GitHub&lt;/a&gt;&lt;/div&gt;
  &lt;/div&gt;
&lt;/div&gt;


&lt;p&gt;These are the endpoints that have already been deployed. A few more are in the pipeline and we would love to get the help of open source community in taking initiative this forward!&lt;/p&gt;

&lt;p&gt;As Software Developers, there isn't much that we could do on the field to help combat the virus, but the geeks inside us made us take us this initiative to help curtail the spread of false information and rely on the data provided by global experts such as John Hopkins&lt;/p&gt;

&lt;p&gt;Github Repository: &lt;a href="https://github.com/backtrackbaba/covid-api"&gt;https://github.com/backtrackbaba/covid-api&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;Documentation: &lt;a href="https://documenter.getpostman.com/view/2568274/SzS8rjbe?version=latest"&gt;Postman Docs&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;Website: &lt;a href="https://covidapi.info"&gt;https://covidapi.info&lt;/a&gt;&lt;/p&gt;

&lt;h1&gt;
  
  
  P.S: Here's our tech stack
&lt;/h1&gt;

&lt;ul&gt;
&lt;li&gt;&lt;p&gt;Python&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;Flask&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;Redis&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;Gunicorn&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;Nginx&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;PostgreSQL&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;Hosted on DigitalOcean&lt;/p&gt;&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;Do let us know if you have any suggestions or feedback or any other endpoint which could make your life easier. Contributions are always welcome and encouraged!&lt;/p&gt;

&lt;p&gt;Hack on!&lt;/p&gt;

</description>
      <category>showdev</category>
      <category>python</category>
      <category>redis</category>
    </item>
    <item>
      <title>The story of my first hackathon | 1 hackathon : 48 hours : Immeasurable learnings</title>
      <dc:creator>Saiprasad Balasubramanian</dc:creator>
      <pubDate>Mon, 13 Jan 2020 07:02:03 +0000</pubDate>
      <link>https://dev.to/backtrackbaba/the-story-of-my-first-hackathon-1-hackathon-48-hours-immeasurable-learnings-1ogf</link>
      <guid>https://dev.to/backtrackbaba/the-story-of-my-first-hackathon-1-hackathon-48-hours-immeasurable-learnings-1ogf</guid>
      <description>&lt;p&gt;Back in 2017, I was exploring the world of machine learning. Machine Learning, Data Science, Artificial Intelligence were fast becoming buzzwords of the industry and a natural curiosity had crept in me to understand the domain and evaluate whether it would be a good fit for me for the next few years. With the thirst for knowledge, I started taking the long route of self-learning through openly available online materials. After investing some time in it I realized I needed some guidance and didn't have anybody in my network who could help me with it. That's when I started looking for boot camps and ended up zeroing in on &lt;a href="https://greyatom.com/" rel="noopener noreferrer"&gt;Greyatom School of Data Science (GA)&lt;/a&gt; after considering the likes of Upgrad, Great Lakes and so on. This move was about to change my life!&lt;/p&gt;

&lt;p&gt;While going through the instructor-led offline boot camp at GA, we had a hackathon mid syllabus to test the extent of our knowledge and show us where we truly stand. Working on a fin-tech based problem statement I was very much out of my comfort zone but our team of 5 had a good mix of technical knowledge and domain expertise and together we were able to pull off a brilliant performance after spending 2 whole days at the campus. In the week after the hackathon, I had some me-time to evaluate where I stood, what went right and what could I improve for similar events in the future. I quickly started noting down my thoughts and within a few minutes I had 14 points listed. After evaluating them again I decided to share it with the peers in my cohort. I'm someone who overthinks a lot of things but decided to go ahead with it and within minutes I started getting a good amount of responses which boosted my confidence a lot. Peers in my cohort poured their thoughts too and suggested me to write a blog and share this nugget of wisdom which has been on my TODO list for a long long time!&lt;/p&gt;

&lt;p&gt;&lt;a href="https://media.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fd33wubrfki0l68.cloudfront.net%2F2c6f53d567796ada1b88db35c659047eb034fae6%2F870af%2Fassets%2Fimages%2Fposts%2F2019%2Fga-group.jpg" class="article-body-image-wrapper"&gt;&lt;img src="https://media.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fd33wubrfki0l68.cloudfront.net%2F2c6f53d567796ada1b88db35c659047eb034fae6%2F870af%2Fassets%2Fimages%2Fposts%2F2019%2Fga-group.jpg" alt="Alt text of image"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;&lt;br&gt;&lt;br&gt;
Here are the points that came out of that eventful weekend:&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;1) Always use version control&lt;/strong&gt;&lt;/p&gt;

&lt;p&gt;I've seen a lot of people learning Data Science who aren't comfortable with version control. This is one of the most fundamental things that you should incorporate in your workflow. It's always better to have your codebase/notebooks centrally managed instead of sharing the same with each other via mail, dropbox or any other file sharing mechanism. It also becomes very easy to track changes and revert whenever needed.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;2) ‎Have sample codes for different algorithms equipped beforehand to refer when required.&lt;/strong&gt;&lt;/p&gt;

&lt;p&gt;A lot of times you would be trying out various algorithms to see which one works better. While juggling with these algorithms, you may get drowned in the implementation for a long time. It helps to have sample snippets for at least the main algorithms like Naive Bayes Classifier, SVM, Decision Trees, etc.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;3) Have awesome visualization&lt;/strong&gt;&lt;/p&gt;

&lt;p&gt;You need to tell the beautiful story of the dataset from being just a simple set of information/numbers to becoming a full-fledged model. Judges won't judge you just your code but also on the way you can visualize it and make it easily interpretable. The numbers in the CSV won't matter much or be useful if it isn't explained visually.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;4) Reusable Code&lt;/strong&gt;&lt;/p&gt;

&lt;p&gt;Try to write functional or Object-Oriented (IMHO) code to have maximum code reusability. This is where your coding skills come into play. When you write in the different cells in your Notebook you might end up writing similar code multiple times which should be clearly avoided.&lt;br&gt;
Ex: A lot of times you might need to read a CSV and convert it to a data frame multiple times. It should ideally be converted into a reusable function which not only reduces your efforts but also helps in refactoring the code in the future.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;5) End to End Pipeline&lt;/strong&gt;&lt;/p&gt;

&lt;p&gt;It's advisable to build your product from end to end, no matter how much crappy it is and then build on it and improve. Your problem statement might need to use different models at different stages chained together. A lot of times people get hooked on to perfect the initial algorithms that they start losing the big picture and end up getting less time to complete the whole problem statement.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;6) Focus on additional features at the end&lt;/strong&gt;&lt;/p&gt;

&lt;p&gt;At some point in the hackathon, you would have made a basic list of features required as per the problem statement. As time goes by and things start shaping up you would realize the true potential of what you are building. You would think that you could add these additional features/functionalities which may help you score some brownie points with the judges. DON'T. Focus on completing the problem statement with the basic functionalities required before you start improvising! You could note these points down and start working on them later on.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;7) Understand the team dynamics before you start&lt;/strong&gt;&lt;/p&gt;

&lt;p&gt;Your team could be coming from diverse backgrounds with varying skill sets. Not everybody would be a hardcode coder. Understand what people's strengths and weaknesses are and distribute work accordingly.&lt;/p&gt;

&lt;p&gt;Ex: We were very lucky to have two members  of our team , &lt;a href="https://www.linkedin.com/in/rajesh-upadhyay-9783077b/" rel="noopener noreferrer"&gt;Rajesh&lt;/a&gt; and &lt;a href="https://www.linkedin.com/in/abhinav-anand-a298064a/" rel="noopener noreferrer"&gt;Abhinav&lt;/a&gt; from a fin-tech background which provided us the domain expertise which we needed while &lt;a href="https://www.linkedin.com/in/sandiprb" rel="noopener noreferrer"&gt;Sandip&lt;/a&gt;, &lt;a href="https://www.linkedin.com/in/yuvraj-kale-b440738/" rel="noopener noreferrer"&gt;Yuvraj&lt;/a&gt; and I brought our knowledge of software and product development to the table.&lt;/p&gt;

&lt;p&gt;&lt;a href="https://media.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fthepracticaldev.s3.amazonaws.com%2Fi%2Frm91hbwy5tj6ubic9vyz.jpeg" class="article-body-image-wrapper"&gt;&lt;img src="https://media.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fthepracticaldev.s3.amazonaws.com%2Fi%2Frm91hbwy5tj6ubic9vyz.jpeg" alt="Team Discussion"&gt;&lt;/a&gt;&lt;/p&gt;



&lt;p&gt;&lt;strong&gt;8) Keep the big guns handy!&lt;/strong&gt;&lt;/p&gt;

&lt;p&gt;‎Have a contingency plan in case your model can't be processed well on your laptop. In our case, we had access to a few ML/AI optimized servers at our disposal. We had kept a trial account of IBM Watson Studio which gave access to an environment similar to Google Colab which had just launched around that time.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;9) Go beyond your syllabus&lt;/strong&gt;&lt;/p&gt;

&lt;p&gt;This point was related to our boot camp but applies in general too. Go beyond what you know or learned in your program. Once you have completed a basic end to end pipeline, if your team dynamics permit, you could experiment with algorithms that you might not know much about but fits your application.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;10) Understand the need for that hackathon&lt;/strong&gt;&lt;/p&gt;

&lt;p&gt;If it is a Hackathon that is targetted to develop something which will be productionized, focus on the big picture. Just because your model has the best metrics doesn't mean it'll be considered if it is computationally expensive&lt;br&gt;
The best example of this is the Netflix Prize 2009 $1 Million winners whose algorithm was not used even though it had the best performance due to the complexity associated with it.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;11) Document EVERYTHING!&lt;/strong&gt;&lt;/p&gt;

&lt;p&gt;Document EVERYTHING. Don't document only the code which makes the final cut for the presentation, document your failures too. It'll help you in the long run. You don't need to publicize it, keep it for internal reference. A few years down the line this shows how you have grown over the years as well as what mistakes you shouldn't repeat.&lt;/p&gt;

&lt;p&gt;P.S: This was my way of documenting stuff. This blog post was initially nothing more than 14 points&lt;/p&gt;

&lt;p&gt;&lt;a href="https://media.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fthepracticaldev.s3.amazonaws.com%2Fi%2Fm9remajw7j4sxgabcvyb.jpeg" class="article-body-image-wrapper"&gt;&lt;img src="https://media.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fthepracticaldev.s3.amazonaws.com%2Fi%2Fm9remajw7j4sxgabcvyb.jpeg" alt="Documentation"&gt;&lt;/a&gt;&lt;/p&gt;



&lt;p&gt;&lt;strong&gt;12) Spend time understanding the data&lt;/strong&gt;&lt;/p&gt;

&lt;p&gt;Take enough time to understand and clean your data.&lt;br&gt;
In this case, we had a week to explore the data and think about it, in competitions you'll have exactly 48 hours or less to get used to the data and work on it. Spending some time with data helps you understand what that data truly represents and what it lacks. Also known as EDA.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;13) Move beyond the traditional Data Visualization libraries&lt;/strong&gt;&lt;/p&gt;

&lt;p&gt;Explore visualization techniques beyond matplotlib or seaborn. Check out d3.js (ex: &lt;a href="http://www.r2d3.us/visual-intro-to-machine-learning-part-1/" rel="noopener noreferrer"&gt;r2d3&lt;/a&gt; ) and other libraries built on top of it.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;14) DEMO!!!!&lt;/strong&gt;&lt;/p&gt;

&lt;p&gt;In my opinion, the best demo is when you don't give a demo! The most powerful demo is when you give control to the audience/judges to do it themselves! One of the things that you could do is create a web app that consumes your models and gives you results on the app itself providing a seamless and easily usable way for your end-users to consume your application. Bonus points for hosting the same on a remote server! In our case, we had &lt;a href="https://www.linkedin.com/in/sandiprb/" rel="noopener noreferrer"&gt;Sandip Baradiya&lt;/a&gt; with us who has extensive experience in developing Django &amp;amp;&amp;amp; React-based applications. On the last day, he quickly whipped up an application that we used during our demo.&lt;/p&gt;

&lt;p&gt;&lt;iframe width="710" height="399" src="https://www.youtube.com/embed/Kqqc3kIa-7w"&gt;
&lt;/iframe&gt;
&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;15) Make sense of evaluation criteria&lt;/strong&gt;&lt;/p&gt;

&lt;p&gt;Know the evaluation criteria of the competition and focus on everything accordingly not just on EDA or model building. Usually, all of these steps have some points but understand the evaluation criteria well so that you can cover as much ground as possible even if you aren't able to complete your application&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;16) Model reasoning&lt;/strong&gt;&lt;/p&gt;

&lt;p&gt;Have reasoning ready for why you chose algorithm A over algorithm B. This is usually asked by the judges while evaluating you and also to detect plagiarism.&lt;/p&gt;

&lt;p&gt;P.S: Your documentation helps a lot in this case.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;17) Document the performance of the model at every stage and not just code.&lt;/strong&gt;&lt;/p&gt;

&lt;p&gt;Documenting metrics of each model helps you in analyzing how accurate/right your end solution is as compared to the other solutions that you had thought of.&lt;/p&gt;

&lt;p&gt;I would also like to thank a good friend of mine &lt;a href="https://www.linkedin.com/in/mohammed-sunasra/" rel="noopener noreferrer"&gt;Mohammed Sunasra&lt;/a&gt; for adding points 15 through 17&lt;/p&gt;

&lt;p&gt;&lt;a href="https://media.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fthepracticaldev.s3.amazonaws.com%2Fi%2F0u38f34ehnfvgjq42mro.jpeg" class="article-body-image-wrapper"&gt;&lt;img src="https://media.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fthepracticaldev.s3.amazonaws.com%2Fi%2F0u38f34ehnfvgjq42mro.jpeg" alt="Participant's Group Photo"&gt;&lt;/a&gt;&lt;/p&gt;



&lt;p&gt;This is my first article on &lt;a href="https://dev.to"&gt;dev.to&lt;/a&gt; &amp;amp; would love some suggestions / feedback from my fellow Dev's&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;This article was originally posted on &lt;a href="https://saipy.com/the-story-of-my-first-hackathon" rel="dofollow noopener noreferrer"&gt;my personal blog&lt;/a&gt;&lt;/strong&gt;&lt;/p&gt;

&lt;p&gt;EDIT: I was dumb enough to not add all the images from the original blog post. Added them now&lt;/p&gt;

</description>
      <category>hackathon</category>
      <category>datascience</category>
      <category>machinelearning</category>
      <category>tips</category>
    </item>
  </channel>
</rss>
