<?xml version="1.0" encoding="UTF-8"?>
<rss version="2.0" xmlns:atom="http://www.w3.org/2005/Atom" xmlns:dc="http://purl.org/dc/elements/1.1/">
  <channel>
    <title>DEV Community: Matthew</title>
    <description>The latest articles on DEV Community by Matthew (@mdani38).</description>
    <link>https://dev.to/mdani38</link>
    <image>
      <url>https://media2.dev.to/dynamic/image/width=90,height=90,fit=cover,gravity=auto,format=auto/https:%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Fuser%2Fprofile_image%2F223374%2F0cfaa56b-19c6-431e-bf1a-266959557975.JPG</url>
      <title>DEV Community: Matthew</title>
      <link>https://dev.to/mdani38</link>
    </image>
    <atom:link rel="self" type="application/rss+xml" href="https://dev.to/feed/mdani38"/>
    <language>en</language>
    <item>
      <title>NLP Transfer Learning with BERT</title>
      <dc:creator>Matthew</dc:creator>
      <pubDate>Tue, 05 Nov 2019 01:49:57 +0000</pubDate>
      <link>https://dev.to/mdani38/nlp-transfer-learning-with-bert-2ldf</link>
      <guid>https://dev.to/mdani38/nlp-transfer-learning-with-bert-2ldf</guid>
      <description>&lt;p&gt;Recently, I was working on a Natural Language Processing (NLP) project where the goal was to classify fake news based on the text contained in the headline and body text. I tried a few different methods including a simple baseline model. However, I picked this NLP project so that I could have a first exposure to working with neural networks since I have not worked with them much previously. Through this process, I discovered the power of using pre-trained BERT neural networks.&lt;/p&gt;

&lt;p&gt;&lt;a href="https://res.cloudinary.com/practicaldev/image/fetch/s--Np42PBz_--/c_limit%2Cf_auto%2Cfl_progressive%2Cq_66%2Cw_880/https://thepracticaldev.s3.amazonaws.com/i/czq7cpiam0ab84ym1x46.gif" class="article-body-image-wrapper"&gt;&lt;img src="https://res.cloudinary.com/practicaldev/image/fetch/s--Np42PBz_--/c_limit%2Cf_auto%2Cfl_progressive%2Cq_66%2Cw_880/https://thepracticaldev.s3.amazonaws.com/i/czq7cpiam0ab84ym1x46.gif" alt="Alt Text"&gt;&lt;/a&gt;&lt;sup id="fnref1"&gt;1&lt;/sup&gt;&lt;/p&gt;

&lt;p&gt;But first, some background.&lt;/p&gt;




&lt;h2&gt;
  
  
  Background
&lt;/h2&gt;

&lt;p&gt;Fake news is a growing problem and the popularity of social media as a fast and easy way to share news has only exacerbated the issue. For this project, I was attempting to classify news as real or fake using the FakeNewsNet database created by The Data Mining and Machine Learning lab (DMML) at ASU. The database was created using news articles shared on Twitter over the past few years. The labels of Real or Fake were generated by checking the news stories against two fact checking tools: Politifact (political news) and Gossipcop (primarily entertainment news but other articles as well). If you want more information about the dataset please see &lt;a href="https://arxiv.org/abs/1809.01286"&gt;https://arxiv.org/abs/1809.01286&lt;/a&gt; and  &lt;a href="https://github.com/KaiDMML/FakeNewsNet"&gt;https://github.com/KaiDMML/FakeNewsNet&lt;/a&gt;. &lt;/p&gt;

&lt;p&gt;&lt;a href="https://res.cloudinary.com/practicaldev/image/fetch/s--Kj4VH0I_--/c_limit%2Cf_auto%2Cfl_progressive%2Cq_auto%2Cw_880/https://thepracticaldev.s3.amazonaws.com/i/b9bc786uwlqp01tlex1w.png" class="article-body-image-wrapper"&gt;&lt;img src="https://res.cloudinary.com/practicaldev/image/fetch/s--Kj4VH0I_--/c_limit%2Cf_auto%2Cfl_progressive%2Cq_auto%2Cw_880/https://thepracticaldev.s3.amazonaws.com/i/b9bc786uwlqp01tlex1w.png" alt="Alt Text"&gt;&lt;/a&gt;&lt;br&gt;
&lt;sup id="fnref2"&gt;2&lt;/sup&gt;&lt;/p&gt;

&lt;h2&gt;
  
  
  Baseline
&lt;/h2&gt;

&lt;p&gt;For fake news classification on this dataset, I used 9,829 data points. I trained on 80% of the dataset and tested on the other 20%. I did not use any of the additional features in my data (author, source of article, date published, etc) so that I could focus only on the title and text of the articles. I combined the two text fields in order to train my model. For NLP, you have to vectorize your words in order to feed into the model. My first model was a simple sklearn pipeline using Count Vectorization and TD-IDF. Count Vectorization splits words into tokens and then counts it based on how many occurrences it has in the whole document. TD-IDF stands for term frequency times inverse document-frequency which is a weighting scheme that calculates how many times word tokens appear in a given document. This system lessens the impact of words that appear broadly within all of the documents (samples of text) being analyzed. These two methods create new numeric features for the text based on these calculations. The features were then passed into a simple Logistic Regression model for classification which yielded an accuracy of 79%. However, this model only accounts for how often a word occurs in the document relative to the whole vocabulary in classifying real from fake. &lt;/p&gt;

&lt;p&gt;A more sophisticated model can be achieved by neural networks and deep learning.&lt;/p&gt;

&lt;h2&gt;
  
  
  Neural Networks for NLP and Why BERT is so Important
&lt;/h2&gt;

&lt;p&gt;I decided to move on to a neural network yet there are so many different types and architectures of neural networks for use in NLP. I found many instances online of RNNs(Recurrent Neural Networks) and LSTMs (Long-Short Term Memory Units) being used for these problems. This is due to their architectures being set up to retain information throughout the process as well as being able to take in word embeddings that account for more than just individual words. So I set out to build a RNN and got most of the way to making one that worked well but I discovered transfer learning.&lt;/p&gt;

&lt;h3&gt;
  
  
  Work Smarter, Not Harder or How I Stopped Worrying and Learned to Love Transfer Learning
&lt;/h3&gt;

&lt;p&gt;Transfer learning is a concept in deep learning where you take knowledge gained from one problem and apply it to a similar problem. While this may seem purely conceptual, it is actually applied quite regularly in the field of machine learning. Practically, it involves taken a pre-trained model that has already been trained on a large amount of data and then retraining the last layer on domain-specific data for the related problem. This can be a powerful method when you don't have the massive amounts of data, training time, or computational power to train a neural network from scratch. It has been used fairly regularly in  in image classification problems and in the last few years has begun to be applied in NLP. This is because of the development of BERT.&lt;/p&gt;

&lt;h3&gt;
  
  
  Why BERT?
&lt;/h3&gt;

&lt;p&gt;BERT is a powerful model in transfer learning for several reasons. First, it is similar to OpenAI's GPT2 that is based on the transformer(an encoder combined with a decoder). However, that model can only read words uni-directionally which does not make it ideal for classification. BERT reads words in both directions(bidirectionally) and thus can read words before and after the word in a sequence. Another important advantage to BERT is that it is a masked language model that masks 15% of the tokens fed into the model. These two factors make it very good at a variety of word classification tasks. &lt;/p&gt;

&lt;p&gt;Those are important to the magic behind BERT but the true power lies in its use in NLP transfer learning. BERT is trained on a large amount of words and articles from Wikipedia. The amount of data it is trained on is much more than most people would be able to train on for specific problems. Therefore, you can import a pre-trained BERT and then retrain just the final layer on context-specific data to create a powerful classification neural network model in a short amount of time. Using a pre-trained BERT, I was able to achieve an accuracy of 71% without tuning many parameters.&lt;/p&gt;

&lt;p&gt;This proves that if you are doing a NLP classification problem, your first instinct should be to utilize the work that is already out there before you build something from scratch.&lt;/p&gt;

&lt;p&gt;If you want to take a look at the full presentation, please visit:&lt;/p&gt;

&lt;p&gt;&lt;a href="https://www.canva.com/design/DADp9gRmr78/ggbpzPMzuxSUXoPQeIvSDw/view?utm_content=DADp9gRmr78&amp;amp;utm_campaign=designshare&amp;amp;utm_medium=link&amp;amp;utm_source=sharebutton#1"&gt;https://www.canva.com/design/DADp9gRmr78/ggbpzPMzuxSUXoPQeIvSDw/view?utm_content=DADp9gRmr78&amp;amp;utm_campaign=designshare&amp;amp;utm_medium=link&amp;amp;utm_source=sharebutton#1&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;If you wish to look at the code and work through these things yourself, please visit:&lt;/p&gt;

&lt;p&gt;&lt;a href="https://github.com/mdani38/Module-4-Project_houston-ds-082619"&gt;https://github.com/mdani38/Module-4-Project_houston-ds-082619&lt;/a&gt;&lt;/p&gt;




&lt;h2&gt;
  
  
  Links and References
&lt;/h2&gt;

&lt;p&gt;&lt;a href="https://github.com/KaiDMML/FakeNewsNet"&gt;https://github.com/KaiDMML/FakeNewsNet&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;&lt;a href="https://arxiv.org/abs/1809.01286"&gt;https://arxiv.org/abs/1809.01286&lt;/a&gt; &lt;/p&gt;

&lt;p&gt;&lt;a href="https://arxiv.org/abs/1712.07709"&gt;https://arxiv.org/abs/1712.07709&lt;/a&gt; &lt;/p&gt;

&lt;p&gt;&lt;a href="https://arxiv.org/abs/1708.01967"&gt;https://arxiv.org/abs/1708.01967&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;&lt;a href="https://arxiv.org/pdf/1810.04805"&gt;https://arxiv.org/pdf/1810.04805&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;&lt;a href="https://d4mucfpksywv.cloudfront.net/better-language-models/language_models_are_unsupervised_multitask_learners.pdf"&gt;https://d4mucfpksywv.cloudfront.net/better-language-models/language_models_are_unsupervised_multitask_learners.pdf&lt;/a&gt;&lt;/p&gt;




&lt;ol&gt;

&lt;li id="fn1"&gt;
&lt;p&gt;&lt;a href="https://giphy.com/gifs/reaction-creepy-MtmFbGJ6YsUEg"&gt;https://giphy.com/gifs/reaction-creepy-MtmFbGJ6YsUEg&lt;/a&gt; ↩&lt;/p&gt;
&lt;/li&gt;

&lt;li id="fn2"&gt;
&lt;p&gt;Word cloud showing highest occurring words from the fake news and the real news in the dataset &lt;a href="https://www.canva.com/design/DADp9gRmr78/ggbpzPMzuxSUXoPQeIvSDw/view?utm_content=DADp9gRmr78&amp;amp;utm_campaign=designshare&amp;amp;utm_medium=link&amp;amp;utm_source=sharebutton#6"&gt;https://www.canva.com/design/DADp9gRmr78/ggbpzPMzuxSUXoPQeIvSDw/view?utm_content=DADp9gRmr78&amp;amp;utm_campaign=designshare&amp;amp;utm_medium=link&amp;amp;utm_source=sharebutton#6&lt;/a&gt;  ↩&lt;/p&gt;
&lt;/li&gt;

&lt;/ol&gt;

</description>
      <category>beginners</category>
      <category>python</category>
      <category>nlp</category>
    </item>
    <item>
      <title>Getting Good at Git</title>
      <dc:creator>Matthew</dc:creator>
      <pubDate>Fri, 27 Sep 2019 15:32:48 +0000</pubDate>
      <link>https://dev.to/mdani38/getting-good-at-git-6ea</link>
      <guid>https://dev.to/mdani38/getting-good-at-git-6ea</guid>
      <description>&lt;p&gt;Last week I worked on a project in which I was tasked with using movie data from different sources in order to gather insights on what factors create value in the film industry. I had written some code on using some loops to make multiple requests to the The Movie Database API. The function gathered nearly 10000 movies from the discover endpoint then used the loop to use the unique movie ids to request from a different endpoint to gather detailed movie information and write it to a data frame.&lt;/p&gt;

&lt;h2&gt;
  
  
  So far, so good...
&lt;/h2&gt;

&lt;p&gt;It was my intention to write this blog post on that process. After completing the long process of finishing the requests and gathering all of the data into a data frame, I exported it to a csv file that I sent to a teammate. However, an incident with git and github deleted all of the files in my project repository which wiped out all the code related to the process. &lt;/p&gt;

&lt;p&gt;&lt;a href="https://res.cloudinary.com/practicaldev/image/fetch/s--MDJblIOy--/c_limit%2Cf_auto%2Cfl_progressive%2Cq_66%2Cw_880/https://thepracticaldev.s3.amazonaws.com/i/752rnb9uqbp1ncf4ttaj.gif" class="article-body-image-wrapper"&gt;&lt;img src="https://res.cloudinary.com/practicaldev/image/fetch/s--MDJblIOy--/c_limit%2Cf_auto%2Cfl_progressive%2Cq_66%2Cw_880/https://thepracticaldev.s3.amazonaws.com/i/752rnb9uqbp1ncf4ttaj.gif" alt="Angry Fed Up GIF"&gt;&lt;/a&gt;&lt;sup id="fnref1"&gt;1&lt;/sup&gt;&lt;/p&gt;

&lt;p&gt;This brings me to the topic of this post, getting git to work for you and not against you.&lt;/p&gt;

&lt;h2&gt;
  
  
  Github repositories are a powerful tool... but understand what you are doing
&lt;/h2&gt;

&lt;p&gt;What happened in my case is that I forked a repository but cloned the original repository so I ended up pushing all of my changes up to the wrong location. We were able to get the files off of the repository and change my upstream but in that process, all the files were wiped off of my local directory.&lt;/p&gt;

&lt;p&gt;Platforms that host repositories are a great way to share versions of code and store files off of your working computer. But if you do not use git properly in your use of GitHub or Gitlab, it can can only hamper your data science or coding project. So, I will share some quick tips that I have gathered from my own recent struggles.&lt;/p&gt;

&lt;p&gt;&lt;a href="https://res.cloudinary.com/practicaldev/image/fetch/s--gG0WFuzZ--/c_limit%2Cf_auto%2Cfl_progressive%2Cq_66%2Cw_880/https://thepracticaldev.s3.amazonaws.com/i/76hdpvhvpewwppoifhv6.gif" class="article-body-image-wrapper"&gt;&lt;img src="https://res.cloudinary.com/practicaldev/image/fetch/s--gG0WFuzZ--/c_limit%2Cf_auto%2Cfl_progressive%2Cq_66%2Cw_880/https://thepracticaldev.s3.amazonaws.com/i/76hdpvhvpewwppoifhv6.gif" alt="Alt Text"&gt;&lt;/a&gt;&lt;br&gt;
&lt;sup id="fnref2"&gt;2&lt;/sup&gt;&lt;/p&gt;




&lt;h3&gt;
  
  
  Step 1: Fork it
&lt;/h3&gt;

&lt;p&gt;If there is a repository on GitHub that you are interested in using, the important first step is to fork it. This will create a copy of the repository linked to your account. It may not seem vital if you are just trying to access some code or some files but it will save you heartbreak in the long run. &lt;/p&gt;

&lt;h3&gt;
  
  
  Step 2: Clone the repository
&lt;/h3&gt;

&lt;p&gt;The next step is to clone the repository to your local machine using the git clone command in your terminal but make sure that you are paying attention to the path that you are pasting into your terminal. This is because when you clone the repository into a local directory, git is setting up the connection the upstream remote repository hosted on GitHub. Practically you will navigate to the directory that you want to use on your local machine and then use the git clone command with the copied link to the forked repository.&lt;/p&gt;

&lt;h3&gt;
  
  
  Step 3: Work
&lt;/h3&gt;

&lt;p&gt;Great. So now we are all set up with a directory that is connected to a remote repository using git. You can now start working with files in the local directory to your heart's content. The issues that arise often occur in the next step when you are trying to upload files to the remote repository. &lt;/p&gt;

&lt;h3&gt;
  
  
  Step 4: .gitignore
&lt;/h3&gt;

&lt;p&gt;Before you start committing changes and pushing files to the repository, you will want to set up your .gitignore. This is a file that is stored in your local directory connected to the remote using git. You can add file paths to the body of this file and it will ignored those files when you start using git to move files to the remote location. It is recommended to put config files, api keys, environment files in the .gitignore but you can add any files that you do not want on your public repository.&lt;/p&gt;

&lt;h3&gt;
  
  
  Step 5: Adding files to the staging area
&lt;/h3&gt;

&lt;p&gt;Moving files from your local repository to the remote location on GitHub is a mult-stage process and has to be done in order to be successful. The first step is to add the files to the staging area. The command is git add &lt;em&gt;filename&lt;/em&gt;. Additionally, you can use git add --all to add all files and subdirectories to the staging area. If your .gitignore is set up properly, it will still continue to ignore the specified files and add everything else. There are also ways to force git to add the files anyway using git add &lt;em&gt;filename&lt;/em&gt; -f.&lt;/p&gt;

&lt;h3&gt;
  
  
  Step 6: Commit
&lt;/h3&gt;

&lt;p&gt;Once you have added files using git add, the next step is to commit changes. One thing that I like to do first (and in every stage of this process) is to use the command git status. This will tell you what files are in the staging area, as well as what files are different from those in the repository. &lt;/p&gt;

&lt;p&gt;Now you are ready to commit your changes of the files that you have added to the local repository. Again, this will only affect the files currently in the staging area. It is required to add a commit message and it is a good habit to make this meaningful as once you push the files to your public repository, the message history will be showing. The command is git commit -m "Commit Message".&lt;/p&gt;

&lt;p&gt;After you have completed this step, you will have committed changes using git. However, this is only on the local git repository in that directory. There is another step to move committed files to the remote repository hosted publicly on GitHub as well to pull the files that others have put on it. &lt;/p&gt;

&lt;p&gt;&lt;a href="https://res.cloudinary.com/practicaldev/image/fetch/s--Amvrpbkc--/c_limit%2Cf_auto%2Cfl_progressive%2Cq_66%2Cw_880/https://thepracticaldev.s3.amazonaws.com/i/90s7gu9195ekqbz4s7l7.gif" class="article-body-image-wrapper"&gt;&lt;img src="https://res.cloudinary.com/practicaldev/image/fetch/s--Amvrpbkc--/c_limit%2Cf_auto%2Cfl_progressive%2Cq_66%2Cw_880/https://thepracticaldev.s3.amazonaws.com/i/90s7gu9195ekqbz4s7l7.gif" alt="Push It"&gt;&lt;/a&gt;&lt;br&gt;
&lt;sup id="fnref3"&gt;3&lt;/sup&gt;&lt;/p&gt;

&lt;h3&gt;
  
  
  Step 7: Push it (and pull)
&lt;/h3&gt;

&lt;p&gt;GitHub is a powerful platform because it allows multiple people to work using a public repository and add files and remove them. It also allows different branches of the repository to work with. If you are working on a project with someone else, you will most likely have to add files to a shared repository and pull files your partners have added.&lt;/p&gt;

&lt;p&gt;Once you have committed your changes, you are ready to push those changes to the remote upstream repository. However, it is best to pull the changes from the remote repository first. There will be conflicts if you are trying to push changes to a repository that has changed. Using the command git pull &lt;strong&gt;branch&lt;/strong&gt;, you will pull any changes from the repository and add them to your local directory. &lt;/p&gt;

&lt;p&gt;Now you can push committed changes up to the repository. This is done using the command git push &lt;strong&gt;branch&lt;/strong&gt;. By default, it will push and pull to/from the origin master branch but you can specify otherwise. The final step is to check on the public repository if the changes have registered.&lt;/p&gt;




&lt;h2&gt;
  
  
  Quick tips
&lt;/h2&gt;

&lt;h5&gt;
  
  
  Fork it
&lt;/h5&gt;

&lt;h5&gt;
  
  
  Clone it into a new directory and know where that is
&lt;/h5&gt;

&lt;h5&gt;
  
  
  Set up your gitignore
&lt;/h5&gt;

&lt;h5&gt;
  
  
  Use git status often to see what the changes are in your local repository and the stages of git add, git commit, git push, etc
&lt;/h5&gt;

&lt;h5&gt;
  
  
  Make meaningful commit messages to track your commit history
&lt;/h5&gt;

&lt;h5&gt;
  
  
  Pull before you push
&lt;/h5&gt;

&lt;h5&gt;
  
  
  Use different filenames than your partners that use the repository. This will prevent conflicts and the process of having to resolve these merge conflicts
&lt;/h5&gt;

&lt;h5&gt;
  
  
  Read the git docs for more tips about how to manipulate git in your favor &lt;a href="https://git-scm.com/docs"&gt;https://git-scm.com/docs&lt;/a&gt;
&lt;/h5&gt;

&lt;p&gt;&lt;a href="https://res.cloudinary.com/practicaldev/image/fetch/s--iaXPifZr--/c_limit%2Cf_auto%2Cfl_progressive%2Cq_66%2Cw_880/https://thepracticaldev.s3.amazonaws.com/i/amj43r32pafjk3063f5p.gif" class="article-body-image-wrapper"&gt;&lt;img src="https://res.cloudinary.com/practicaldev/image/fetch/s--iaXPifZr--/c_limit%2Cf_auto%2Cfl_progressive%2Cq_66%2Cw_880/https://thepracticaldev.s3.amazonaws.com/i/amj43r32pafjk3063f5p.gif" alt="That's All"&gt;&lt;/a&gt;&lt;br&gt;
&lt;sup id="fnref4"&gt;4&lt;/sup&gt;&lt;/p&gt;




&lt;ol&gt;
&lt;li&gt;Banner image source: &lt;a href="https://techcrunch.com/2019/01/07/github-free-users-now-get-unlimited-private-repositories/"&gt;https://techcrunch.com/2019/01/07/github-free-users-now-get-unlimited-private-repositories/&lt;/a&gt;
&lt;/li&gt;
&lt;/ol&gt;




&lt;ol&gt;

&lt;li id="fn1"&gt;
&lt;p&gt;&lt;a href="https://giphy.com/gifs/LE4FWkEdR7kC4"&gt;https://giphy.com/gifs/LE4FWkEdR7kC4&lt;/a&gt; ↩&lt;/p&gt;
&lt;/li&gt;

&lt;li id="fn2"&gt;
&lt;p&gt;&lt;a href="https://tenor.com/view/brooklyn99-andy-samberg-jake-peralta-do-not-blow-this-for-us-gif-12004248"&gt;https://tenor.com/view/brooklyn99-andy-samberg-jake-peralta-do-not-blow-this-for-us-gif-12004248&lt;/a&gt; ↩&lt;/p&gt;
&lt;/li&gt;

&lt;li id="fn3"&gt;
&lt;p&gt;&lt;a href="https://giphy.com/gifs/pix-pepa-Maf9DN4Ftb0BO"&gt;https://giphy.com/gifs/pix-pepa-Maf9DN4Ftb0BO&lt;/a&gt; ↩&lt;/p&gt;
&lt;/li&gt;

&lt;li id="fn4"&gt;
&lt;p&gt;&lt;a href="https://giphy.com/gifs/PixelBandits-pixel-forest-YBJHgYmcNFRODvssAL"&gt;https://giphy.com/gifs/PixelBandits-pixel-forest-YBJHgYmcNFRODvssAL&lt;/a&gt; ↩&lt;/p&gt;
&lt;/li&gt;

&lt;/ol&gt;

</description>
      <category>beginners</category>
      <category>firstyearincode</category>
      <category>github</category>
      <category>git</category>
    </item>
    <item>
      <title>Data, Inductive Reasoning and Procedural Crime Dramas: Why I Decided to Pursue Data Science</title>
      <dc:creator>Matthew</dc:creator>
      <pubDate>Thu, 05 Sep 2019 23:59:37 +0000</pubDate>
      <link>https://dev.to/mdani38/data-inductive-reasoning-and-procedural-crime-dramas-why-i-decided-to-pursue-data-science-53i6</link>
      <guid>https://dev.to/mdani38/data-inductive-reasoning-and-procedural-crime-dramas-why-i-decided-to-pursue-data-science-53i6</guid>
      <description>&lt;p&gt;In the past month, I made the choice to pursue a Data Science bootcamp. It may seem that this decision has come out of nowhere, but there are many reasons that have driven me to enter into the exciting world of Data Science. &lt;/p&gt;

&lt;p&gt;While the growing opportunities for careers for data scientists are  appealing, the use of data to solve problems combined with programming has always interested me. In the past few years, I have dabbled on and off with using programming and different techniques to manipulate the data I used in my pursuits as a scientist. This pivot represents a change in my possible career path but it stays true to the interests that I have always had.&lt;/p&gt;




&lt;h2&gt;
  
  
  It All Started With Science
&lt;/h2&gt;

&lt;p&gt;&lt;a href="https://media.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fthepracticaldev.s3.amazonaws.com%2Fi%2F5sef8wpzoyz9vc88vv1u.gif" class="article-body-image-wrapper"&gt;&lt;img src="https://media.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fthepracticaldev.s3.amazonaws.com%2Fi%2F5sef8wpzoyz9vc88vv1u.gif" alt="Alt Text"&gt;&lt;/a&gt;&lt;br&gt;
&lt;sup id="fnref1"&gt;1&lt;/sup&gt;&lt;/p&gt;

&lt;p&gt;I have been interested in the pursuits of science for most of my life. From a young age, I have always been interested in the tools of science and how we can use them to understand more about the world around us. A scientific approach drives how I interact with all of the information around me. &lt;/p&gt;

&lt;p&gt;My interests in science drove me to get a BS and MS in Geology as well as publication based on my research on seismic reconstructions of topography beneath ice sheet deposits in the Ross Sea of Antarctica. Past pursuits have already given me plenty of experiences with analyzing, manipulating and visualizing large datasets. &lt;/p&gt;

&lt;p&gt;So why did I decide to go into Data Science in particular? For me, it is the concepts of &lt;strong&gt;inductive reasoning&lt;/strong&gt;, &lt;strong&gt;data-based problem solving&lt;/strong&gt;, and &lt;strong&gt;effective visualization of results&lt;/strong&gt;. These concepts are also shown in another thing I love, detective shows.&lt;/p&gt;

&lt;h2&gt;
  
  
  Detective Shows? What does that have to do with anything?
&lt;/h2&gt;

&lt;p&gt;&lt;a href="https://media.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fthepracticaldev.s3.amazonaws.com%2Fi%2Fyxm2xwf5eyd3iwfxabik.gif" class="article-body-image-wrapper"&gt;&lt;img src="https://media.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fthepracticaldev.s3.amazonaws.com%2Fi%2Fyxm2xwf5eyd3iwfxabik.gif" alt="Alt Text"&gt;&lt;/a&gt;&lt;br&gt;
&lt;sup id="fnref2"&gt;2&lt;/sup&gt;&lt;/p&gt;

&lt;p&gt;Yes, detective shows or procedural crime dramas are related to my interest in pursuing Data Science. While these types of tv shows are entertaining and I have seen most of them (CSI is my favorite), my love of detective shows is tied to my use of logical reasoning. &lt;/p&gt;

&lt;p&gt;While every episode is different, the procedure is the same. The characters (such as Jessica Fletcher pictured above) is presented with a mystery or problem. They then must examine the evidence at the scene of the crime. Additionally, more information can be gathered about the circumstances around the situation.  They need to make determinations about what evidence is useful and what evidence is not as helpful. Then they have to draw conclusions based on the information they have.&lt;/p&gt;

&lt;p&gt;For me, this is a similar workflow to Data Science. The problem, goals and challenges are always different but a data scientist must use the data they have. Then they determine if there is new data that they can gather to help with the problem. There is a determination that must be made about what information is useful to focus on in large datasets and what data is not as helpful. Then finally, they can analyze that data and use different techniques to model and visualize it to draw conclusions.&lt;/p&gt;




&lt;h2&gt;
  
  
  A New Path
&lt;/h2&gt;

&lt;p&gt;These are just my first thoughts about what I think about the field of Data Science and the use of large data sets to solve problems. Many of these concepts are new to me and I have certainly been working hard at it. However, I am enjoying the learning process and I am excited to begin my journey down this path and see where it leads me. &lt;/p&gt;




&lt;ol&gt;

&lt;li id="fn1"&gt;
&lt;p&gt;&lt;a href="https://giphy.com/gifs/tired-OnJLRvXvAmvPW" rel="noopener noreferrer"&gt;https://giphy.com/gifs/tired-OnJLRvXvAmvPW&lt;/a&gt; Source: &lt;a href="http://www.utica.edu/student-blogs/feeling-tired-try-improving-the-quality-of-your-sleep/" rel="noopener noreferrer"&gt;http://www.utica.edu/student-blogs/feeling-tired-try-improving-the-quality-of-your-sleep/&lt;/a&gt;    ↩&lt;/p&gt;
&lt;/li&gt;

&lt;li id="fn2"&gt;
&lt;p&gt;&lt;a href="https://giphy.com/gifs/angela-lansbury-jessica-fletcher-murder-she-wrote-JGLYZcomPzv4Q" rel="noopener noreferrer"&gt;https://giphy.com/gifs/angela-lansbury-jessica-fletcher-murder-she-wrote-JGLYZcomPzv4Q&lt;/a&gt;. Source: &lt;a href="https://cabotcovecorpse.tumblr.com/post/119018723434" rel="noopener noreferrer"&gt;https://cabotcovecorpse.tumblr.com/post/119018723434&lt;/a&gt; ↩&lt;/p&gt;
&lt;/li&gt;

&lt;/ol&gt;

</description>
      <category>beginners</category>
      <category>career</category>
      <category>firstyearincode</category>
      <category>datascience</category>
    </item>
  </channel>
</rss>
