<?xml version="1.0" encoding="UTF-8"?>
<rss version="2.0" xmlns:atom="http://www.w3.org/2005/Atom" xmlns:dc="http://purl.org/dc/elements/1.1/">
  <channel>
    <title>DEV Community: Dylan Lisk</title>
    <description>The latest articles on DEV Community by Dylan Lisk (@dlisk92).</description>
    <link>https://dev.to/dlisk92</link>
    <image>
      <url>https://media2.dev.to/dynamic/image/width=90,height=90,fit=cover,gravity=auto,format=auto/https:%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Fuser%2Fprofile_image%2F284336%2F744d7238-604e-45ed-9b00-46c008ac43bd.png</url>
      <title>DEV Community: Dylan Lisk</title>
      <link>https://dev.to/dlisk92</link>
    </image>
    <atom:link rel="self" type="application/rss+xml" href="https://dev.to/feed/dlisk92"/>
    <language>en</language>
    <item>
      <title>Reinforcement learning</title>
      <dc:creator>Dylan Lisk</dc:creator>
      <pubDate>Sun, 21 Jun 2020 04:26:38 +0000</pubDate>
      <link>https://dev.to/dlisk92/reinforcement-learning-dnj</link>
      <guid>https://dev.to/dlisk92/reinforcement-learning-dnj</guid>
      <description>&lt;h1&gt;
  
  
  What is it?
&lt;/h1&gt;

&lt;p&gt;Like a lot of the ideas in machine learning, it is an abstraction of real life. How do we learn in real life? By doing and seeing the results. Put your hand in a fire and you won't do it again. You just learned something: fire is hot.&lt;/p&gt;

&lt;p&gt;That lesson is taught by negative reinforcement. Having your hand burnt is negative it doesn't feel good. Alternately you can reward positive actions. Like a baby trying ice cream.&lt;/p&gt;

&lt;h2&gt;
  
  
  Setting the scene
&lt;/h2&gt;

&lt;p&gt;The first step in reinforcement learning is establishing an environment. This is distinct from a programming environment. It can be something as simple as an empty array that represents a game board or you could have a robot learning in the real world.&lt;/p&gt;

</description>
    </item>
    <item>
      <title>Making EDA easy</title>
      <dc:creator>Dylan Lisk</dc:creator>
      <pubDate>Sat, 13 Jun 2020 18:57:01 +0000</pubDate>
      <link>https://dev.to/dlisk92/making-eda-easy-426g</link>
      <guid>https://dev.to/dlisk92/making-eda-easy-426g</guid>
      <description>&lt;p&gt;A package has recently come to my attention that makes performing a blanket EDA of a new dataset easier. This package is pandas_profiling.&lt;br&gt;
To install the package the command is&lt;br&gt;
&lt;/p&gt;

&lt;div class="highlight"&gt;&lt;pre class="highlight plaintext"&gt;&lt;code&gt;pip install pandas-profiling[notebook]
&lt;/code&gt;&lt;/pre&gt;&lt;/div&gt;



&lt;p&gt;So what does it do?&lt;br&gt;
These points were taken from the documentation &lt;a href="https://pandas-profiling.github.io/pandas-profiling/docs/master/rtd/"&gt;here&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;-Type inference: detect the types of columns in a dataframe.&lt;br&gt;
-Essentials: type, unique values, missing values&lt;br&gt;
-Quantile statistics like minimum value, Q1, median, Q3, maximum, range, interquartile range&lt;br&gt;
-Descriptive statistics like mean, mode, standard deviation, sum, median absolute deviation, coefficient of variation, kurtosis, skewness&lt;br&gt;
-Most frequent values&lt;br&gt;
-Histograms&lt;br&gt;
-Correlations highlighting of highly correlated variables, Spearman, Pearson and Kendall matrices&lt;br&gt;
-Missing values matrix, count, heatmap and dendrogram of missing values&lt;br&gt;
-Duplicate rows Lists the most occurring duplicate rows&lt;br&gt;
-Text analysis learn about categories (Uppercase, Space), scripts (Latin, Cyrillic) and blocks (ASCII) of text data&lt;/p&gt;

&lt;p&gt;All this is done with one command&lt;br&gt;
&lt;/p&gt;

&lt;div class="highlight"&gt;&lt;pre class="highlight python"&gt;&lt;code&gt;&lt;span class="kn"&gt;from&lt;/span&gt; &lt;span class="nn"&gt;pandas_profiling&lt;/span&gt; &lt;span class="kn"&gt;import&lt;/span&gt; &lt;span class="n"&gt;ProfileReport&lt;/span&gt;
&lt;span class="n"&gt;ProfileReport&lt;/span&gt;&lt;span class="p"&gt;(&lt;/span&gt;&lt;span class="n"&gt;df&lt;/span&gt;&lt;span class="p"&gt;,&lt;/span&gt; &lt;span class="n"&gt;title&lt;/span&gt;&lt;span class="o"&gt;=&lt;/span&gt;&lt;span class="s"&gt;"Pandas Profiling Report"&lt;/span&gt;&lt;span class="p"&gt;)&lt;/span&gt;
&lt;/code&gt;&lt;/pre&gt;&lt;/div&gt;



&lt;p&gt;df in the above code being a Pandas DataFrame object.&lt;/p&gt;

&lt;p&gt;If you have a large dataset (1000000,20) you may be better served by passing the kwarg&lt;br&gt;
&lt;/p&gt;

&lt;div class="highlight"&gt;&lt;pre class="highlight python"&gt;&lt;code&gt;&lt;span class="n"&gt;ProfileReport&lt;/span&gt;&lt;span class="p"&gt;(&lt;/span&gt;&lt;span class="n"&gt;large_dataset&lt;/span&gt;&lt;span class="p"&gt;,&lt;/span&gt; &lt;span class="n"&gt;minimal&lt;/span&gt;&lt;span class="o"&gt;=&lt;/span&gt;&lt;span class="bp"&gt;True&lt;/span&gt;&lt;span class="p"&gt;)&lt;/span&gt;
&lt;/code&gt;&lt;/pre&gt;&lt;/div&gt;



&lt;p&gt;So while good knowledge of your data is ideal, I think pandas_profile is a good first step to building that knowledge.&lt;/p&gt;

</description>
    </item>
    <item>
      <title>GPT2 fine tuned with kaggle winning posts</title>
      <dc:creator>Dylan Lisk</dc:creator>
      <pubDate>Fri, 14 Feb 2020 20:20:37 +0000</pubDate>
      <link>https://dev.to/dlisk92/gpt2-fine-tuned-with-kaggle-winning-posts-29i7</link>
      <guid>https://dev.to/dlisk92/gpt2-fine-tuned-with-kaggle-winning-posts-29i7</guid>
      <description>&lt;p&gt;My name is Dylan Lisk and I am a data science student using gpt-2 to write my blog. I also participate in Kaggle competitions and have done so for many years. I am currently in the process of getting my bachelor s in natural resources management and management. I like to think of myself as a data scientist but more likely a statistics whiz. My thesis was on predictive modeling in the forestry industry. I have done statistical analysis in forestry in the past but never before. I also have a strong background in natural resources. I have worked in the oil and gas industry refining oil and natural gas. I have worked in projects dealing with oil and gas and related industries. I have participated in many R packages dealing with data analysis. R is my default choice for data analysis. I use SAS for all my data manipulation. I use R for trying to figure out how to combine and linear regression. I use SAS for setting up my computer for data and for generating the reports that I send to the governor. My old computer is getting old but I have gotten to the point where I can still be productive from time spent on R. I also have a small library of R packages that I have tried but not used in years. One that I have tried using R mostly for data analysis. It is a random forest package. It has nice performance characteristics. I like the idea of an efficient wrapper for R. I used a http randomforest package in my previous job. It is also possible to use R for data analysis. However one has to be careful about the precision of the estimates. R can give estimates up to for some data. I have tried using variance bias and insensitive estimates. I have also tried using nominal and median estimates. These estimates can be used to generate a rough baseline for the model. However I have not tried them all. I have tried using the mean or standard deviation of each of the estimates. This gives a rough estimate for the test statistic for the selected group of the selected data. However I have found that using these estimates as a baseline gives poor results. So I have tried using a weighted average of the mean and the standard deviation. This gave about as good a result as using the mean and the mean and standard deviation. I have tried using a weighted average of the mean and the mean. However I have found that using the mean gives much better results. So I have tried using the median. However I have found that using the median as a baseline gives poor results. I have tried using the median as a baseline and the test statistic. I have tried using the median as a baseline. I have found that using the median as a baseline gives poor results. I have tried using the median as a baseline and the test statistic. I have tried using the standard deviation of the estimates. This gives a rough estimate for the test statistic for the selected group of the selected data. However I have found that using this estimate as a baseline gives poor results. So I have tried using a weighted average of the standard deviation and the mean. This gave a good result when compared to using the mean and the mean. I have tried using a weighted average of the mean and the mean. However I have found that using the mean gives much better results. So I have tried using the median. However I have found that using the median as a baseline gives poor results. I have tried using a weighted average of the mean and the standard deviation. This gave a good result when compared to using the mean and the mean. However I have found that using the median as a baseline gives poor results. I have tried using the median as a baseline. However I have found that using the median as a baseline gives poor results. Group I have tried using the mean difference between the test and the median as a baseline. However I have found that using this as a baseline gives poor results. So I have tried using the mean difference between the test and the median as a baseline. However I have found that using this as a baseline gives poor results. Group I have tried using the median difference as a baseline. However I have found that using this as a baseline gives poor results. So I have tried using the median difference as a baseline. However I have found that using this as a baseline gives poor results. Group I have tried using the mean difference as a baseline. However I have found that using this as a baseline gives poor results. So I have tried using the mean difference as a baseline. However I have found that using this as a baseline gives poor results. Group I have tried using the median difference as a baseline. However I have found that using this as a baseline gives poor results. So I have tried using the median difference as a baseline. However I have found that using this as a baseline gives poor results. Group I have tried using the standard deviation of the estimates. This gives a rough estimate for the test statistic for the selected group of the selected data. However I have found that using this as a baseline gives poor results. Group I have tried using the&lt;/p&gt;

&lt;h1&gt;
  
  
  i added another
&lt;/h1&gt;

&lt;p&gt;My name is Dylan Lisk and I am a data science student using gpt-2 to write my blog. I also participate in a lot of Kaggle competitions and have done quite well on the leaderboard. I am currently in the last weeks of my senior project in the Bioacoustic Research Institute at the University of Pittsburgh where I am working on some machine learning problems. I have been doing quite well on the leaderboard so far but I would like to improve my score so I joined the strong What made you decide to enter strong Dylan I was looking for a team and a lot of variety in challenges. The length was perfect for this competition. I like to have room to experiment and try things. And the variety of challenges was great. The fact that the problems were all related was very strong Sergei I was looking for a team that was strong Dylan I wanted to create a dataset that was strong Sergei I wanted to create a dataset that was strong a http img aligncenter center auto auto http strong What preprocessing and supervised learning methods did you use strong Dylan I used a lot of feature engineering in preprocessing and feature selection. Feature selection was done by going through the full time series and finding the most important features for each individual month. Then I would select the most important features and polynomial P P was used as feature of all time series for months that had at least features. span color strong a http span color month feature engineering was very powerful in this strong Sergei I used a lot of feature selection and feature transformations. Feature transformation was done by going through the full time series and finding the most important features for each individual month. Then I would select the most important features and polynomial P was used as feature of all time series for months that had at least features. span color strong a http span color month feature transformation was very powerful in this strong a https span color pipeline is a diagram that visually shows the general idea of how I did feature selection and transformation work together. strong a https span color The diagram below summarizes the general idea of how I did feature selection and feature transformation work together. strong a https span color The following figure illustrates an example of how feature engineering can be used to generate features for a more complex model. Notice that the colour gradient is introduced by adding colour to the training data by dividing the training data into bins and averaging the results. strong a https span color Figure: An example of how feature engineering can be used to generate features for a more complex model. Notice that the colour gradient is introduced by adding colour to the training data by dividing the training data into bins and averaging the results. caption aligncenter a http img http Figure: An example of how feature engineering can be used to generate features for a more complex model. Notice that the colour gradient is introduced by adding colour to the training data by dividing the training data into bins and averaging the results. strong a https span color Figure shows an example of how feature engineering can be used to generate features for a more complex model. Notice that the colour gradient is introduced by adding colour to the training data. strong a https span color Figure shows an example of how feature engineering can be used to generate features for a more complex model. Notice that the colour gradient is introduced by adding colour to the training data. strong a https span color strong a https span color Figure shows an example of how feature engineering can be used to generate features for a more complex model. Notice that the colour gradient is introduced by adding colour to the training data. strong a https span color strong a https span color Figure shows an example of how feature engineering can be used to generate features for a more complex model. Notice that the colour gradient is introduced by adding colour to the training data. strong a https span color strong a https span color Figure shows an example of how feature engineering can be used to generate features for a more complex model. Notice that the colour gradient is introduced by adding colour to the training data. strong a https span color strong a https span color Figure shows an example of how feature engineering can be used to generate features for a more complex model. Notice that the colour gradient is introduced by adding colour to the training data. strong a https span color strong a https span color Figure shows an example of how feature engineering can be used to generate features for a more complex model. Notice that the colour gradient is introduced by adding colour to the training data. strong a https span color strong a https span color Figure shows an example of how feature engineering can be used to generate features for a more complex model. Notice that the colour gradient is introduced by adding colour to the training data. strong a https span color strong a https span color Figure shows an example of how feature engineering can be used to generate features for a more complex model. Notice that the colour gradient is introduced by adding colour to the training data. strong a https span color strong a https span color Figure shows an example of how feature engineering can be used to generate features for a more complex model. Notice that the colour gradient is introduced by adding colour to the training data. strong a&lt;/p&gt;

</description>
    </item>
    <item>
      <title>A novices explanation of Data Science</title>
      <dc:creator>Dylan Lisk</dc:creator>
      <pubDate>Fri, 20 Dec 2019 16:24:03 +0000</pubDate>
      <link>https://dev.to/dlisk92/a-novices-explanation-of-data-science-160i</link>
      <guid>https://dev.to/dlisk92/a-novices-explanation-of-data-science-160i</guid>
      <description>&lt;p&gt;The first set of this process is data acquisition which can be accomplished by SQL queries, API calls, web scraping, or simply downloading a file like .csv. For this brief overview I won't touch on data acquisition.&lt;/p&gt;

&lt;h1&gt;
  
  
  So you have your data in a .csv file so what now?
&lt;/h1&gt;



&lt;div class="highlight js-code-highlight"&gt;
&lt;pre class="highlight python"&gt;&lt;code&gt;   &lt;span class="kn"&gt;import&lt;/span&gt; &lt;span class="n"&gt;pandas&lt;/span&gt; &lt;span class="k"&gt;as&lt;/span&gt; &lt;span class="n"&gt;pd&lt;/span&gt;
   &lt;span class="kn"&gt;import&lt;/span&gt; &lt;span class="n"&gt;numpy&lt;/span&gt; &lt;span class="k"&gt;as&lt;/span&gt; &lt;span class="n"&gt;np&lt;/span&gt;
   &lt;span class="n"&gt;pd&lt;/span&gt;&lt;span class="p"&gt;.&lt;/span&gt;&lt;span class="nf"&gt;read_csv&lt;/span&gt;&lt;span class="p"&gt;(&lt;/span&gt;&lt;span class="sh"&gt;'&lt;/span&gt;&lt;span class="s"&gt;path/data.csv&lt;/span&gt;&lt;span class="sh"&gt;'&lt;/span&gt;&lt;span class="p"&gt;)&lt;/span&gt;
&lt;/code&gt;&lt;/pre&gt;

&lt;/div&gt;



&lt;p&gt;What that does is import pandas and names it pd. I then call the pandas read_csv() function to create a pandas.DataFrame object with the data.&lt;br&gt;
With a DataFrame in hand we are able to much more easily address missing values and data trimming.&lt;/p&gt;

&lt;p&gt;As a novice my go to strategy deal with missing values is to simply cut out the offending observations.&lt;br&gt;
&lt;a href="https://i.giphy.com/media/RIe6nVbsB3yz6/source.gif" class="article-body-image-wrapper"&gt;&lt;img src="https://i.giphy.com/media/RIe6nVbsB3yz6/source.gif"&gt;&lt;/a&gt;&lt;br&gt;
&lt;/p&gt;

&lt;div class="highlight js-code-highlight"&gt;
&lt;pre class="highlight plaintext"&gt;&lt;code&gt;DataFrame.dropna()
&lt;/code&gt;&lt;/pre&gt;

&lt;/div&gt;



&lt;p&gt;You can also impute values into the data such as the mean or median as you can preserve the rest of the data in the row. If the feature you are imputeing is important to the target be careful. It would be better to mess as little as possible with highly correlated features. &lt;/p&gt;

&lt;p&gt;Now we deal with outliers. How do we do that?&lt;br&gt;
&lt;a href="https://i.giphy.com/media/L7p2UMMY2Yk3S/source.gif" class="article-body-image-wrapper"&gt;&lt;img src="https://i.giphy.com/media/L7p2UMMY2Yk3S/source.gif"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;You guessed it! I cut that shit out. Now there are many statistical tests and methods to determine what is an outlier. I don't really know them at the moment so I arbitrarily cut the data. To be a little more intelligent and if you have separate test data you can find the min() and max() values chop your train data to reflect those values.&lt;/p&gt;
&lt;h1&gt;
  
  
  Ok so let's look at our data.
&lt;/h1&gt;


&lt;div class="highlight js-code-highlight"&gt;
&lt;pre class="highlight python"&gt;&lt;code&gt;  &lt;span class="kn"&gt;import&lt;/span&gt; &lt;span class="n"&gt;matplotlib.pyplot&lt;/span&gt; &lt;span class="k"&gt;as&lt;/span&gt; &lt;span class="n"&gt;plt&lt;/span&gt;
  &lt;span class="n"&gt;smp&lt;/span&gt; &lt;span class="o"&gt;=&lt;/span&gt; &lt;span class="n"&gt;DataFrame&lt;/span&gt;&lt;span class="p"&gt;.&lt;/span&gt;&lt;span class="nf"&gt;sample&lt;/span&gt;&lt;span class="p"&gt;(&lt;/span&gt;&lt;span class="mi"&gt;100&lt;/span&gt;&lt;span class="p"&gt;)&lt;/span&gt;
  &lt;span class="n"&gt;pd&lt;/span&gt;&lt;span class="p"&gt;.&lt;/span&gt;&lt;span class="n"&gt;plotting&lt;/span&gt;&lt;span class="p"&gt;.&lt;/span&gt;&lt;span class="nf"&gt;scatter_matrix&lt;/span&gt;&lt;span class="p"&gt;(&lt;/span&gt;&lt;span class="n"&gt;smp&lt;/span&gt;&lt;span class="p"&gt;)&lt;/span&gt;
&lt;/code&gt;&lt;/pre&gt;

&lt;/div&gt;


&lt;p&gt;&lt;a href="https://i.giphy.com/media/zhJR6HbK4fthC/source.gif" class="article-body-image-wrapper"&gt;&lt;img src="https://i.giphy.com/media/zhJR6HbK4fthC/source.gif"&gt;&lt;/a&gt;&lt;br&gt;
&lt;a href="https://media.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fthepracticaldev.s3.amazonaws.com%2Fi%2F37ujqjvkxvzz35yvwpqx.png" class="article-body-image-wrapper"&gt;&lt;img src="https://media.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fthepracticaldev.s3.amazonaws.com%2Fi%2F37ujqjvkxvzz35yvwpqx.png" alt="Alt Text"&gt;&lt;/a&gt;&lt;br&gt;
The first thing to look at is the diagonal plot of histograms. This shows us the distribution of the values of each variable. For a lot of modeling techniques we want our features to have a normal distribution. &lt;/p&gt;
&lt;h1&gt;
  
  
  What is normal?
&lt;/h1&gt;

&lt;p&gt;&lt;a href="https://media.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fthepracticaldev.s3.amazonaws.com%2Fi%2Fjz7ep0v45dc3x6hcgw5i.png" class="article-body-image-wrapper"&gt;&lt;img src="https://media.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fthepracticaldev.s3.amazonaws.com%2Fi%2Fjz7ep0v45dc3x6hcgw5i.png" alt="Alt Text"&gt;&lt;/a&gt;&lt;br&gt;
This is a normal distribution. If your distribution isn't normal what can you do besides panic? You can try to transform your data.&lt;br&gt;
&lt;/p&gt;

&lt;div class="highlight js-code-highlight"&gt;
&lt;pre class="highlight python"&gt;&lt;code&gt;   &lt;span class="n"&gt;DataFrame&lt;/span&gt;&lt;span class="p"&gt;[&lt;/span&gt;&lt;span class="sh"&gt;'&lt;/span&gt;&lt;span class="s"&gt;log_col&lt;/span&gt;&lt;span class="sh"&gt;'&lt;/span&gt;&lt;span class="p"&gt;]&lt;/span&gt; &lt;span class="o"&gt;=&lt;/span&gt; &lt;span class="n"&gt;np&lt;/span&gt;&lt;span class="p"&gt;.&lt;/span&gt;&lt;span class="n"&gt;log1p&lt;/span&gt;&lt;span class="p"&gt;(&lt;/span&gt;&lt;span class="n"&gt;DataFrame&lt;/span&gt;&lt;span class="p"&gt;[&lt;/span&gt;&lt;span class="sh"&gt;'&lt;/span&gt;&lt;span class="s"&gt;col&lt;/span&gt;&lt;span class="sh"&gt;'&lt;/span&gt;&lt;span class="p"&gt;]&lt;/span&gt;
&lt;/code&gt;&lt;/pre&gt;

&lt;/div&gt;



&lt;p&gt;Create a new column with values log plus oned and see how that works. Data science is iterative so if that doesn't work try something else.&lt;/p&gt;

&lt;p&gt;So we have our data where we want it. Now we bust out the models. The go to python package for modeling is scikit-learn I recommend visiting &lt;a href="https://scikit-learn.org/dev/index.html" rel="noopener noreferrer"&gt;https://scikit-learn.org/dev/index.html&lt;/a&gt; it is a one stop shop for all things machine learning. &lt;/p&gt;

&lt;p&gt;Before you throw your data into a model you need to do a couple things. &lt;br&gt;
Split your data into a train and test set. The model is trained on the train set and tested on the test set. Pretty self explanatory. Next scale the data down. Models are happy when everything is scaled down to 1 so that coefficients stay manageable. Some of sklearn's scalers can help with outliers as well.&lt;/p&gt;

&lt;h1&gt;
  
  
  What time is it?
&lt;/h1&gt;

&lt;p&gt;&lt;a href="https://i.giphy.com/media/9u514UZd57mRhnBCEk/source.gif" class="article-body-image-wrapper"&gt;&lt;img src="https://i.giphy.com/media/9u514UZd57mRhnBCEk/source.gif"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;h1&gt;
  
  
  It's modeling time baby
&lt;/h1&gt;

&lt;p&gt;You would think this is where the real work begins and you would be wrong. You just did most of work doing the data cleaning, exploration and feature engineering.&lt;/p&gt;

&lt;p&gt;I touch on model selection in a future post that I will link here&lt;/p&gt;

</description>
      <category>python</category>
      <category>machinelearning</category>
    </item>
    <item>
      <title>Why data science???</title>
      <dc:creator>Dylan Lisk</dc:creator>
      <pubDate>Tue, 03 Dec 2019 17:43:38 +0000</pubDate>
      <link>https://dev.to/dlisk92/why-data-science-3nl3</link>
      <guid>https://dev.to/dlisk92/why-data-science-3nl3</guid>
      <description>&lt;p&gt;&lt;a href="https://i.giphy.com/media/xL7PDV9frcudO/giphy.gif" class="article-body-image-wrapper"&gt;&lt;img src="https://i.giphy.com/media/xL7PDV9frcudO/giphy.gif" alt=""&gt;&lt;/a&gt;&lt;br&gt;
To answer the question of why data science, the question of what is data science has to be answered first. So what is data science?&lt;br&gt;
Well if you break down the word literally data science would be the study of data. That doesn’t really say anything by itself , because what is data. &lt;br&gt;
Data is plural form of the Latin word datum. Datum means “something given”. So from that we can piece together that data science means given something, do science to it.  This sounds very broad because it is. Data science is impossibly broad because most problems can be seen as a data science problem. As long as you have observations IE data you can use the data science process to extract meaning.&lt;br&gt;
&lt;a href="https://i.giphy.com/media/1guRIRW8QdSte01T6Du/source.gif" class="article-body-image-wrapper"&gt;&lt;img src="https://i.giphy.com/media/1guRIRW8QdSte01T6Du/source.gif" alt=""&gt;&lt;/a&gt;&lt;br&gt;
Data Science is still a new field as stands today. Looking at google trends for “Data Science” shows that these terms were not commonly put together till around 2013-2014 and has been steadily increasing in search popularity since then. While this does not mean anything definitively it does seem to indicate that field is growing. For someone looking for a job a new growing industry is often a good place to be. One of the things that comes with a growing industry particularly a tech industry is constant change. In data science you will have to be constantly learning. As new modeling techniques become available a good data scientist will be expected to rapidly adopt and incorporate these new processes. This to me is one of the major draws of the data science field. &lt;/p&gt;

&lt;p&gt;Another facet of data science is in the translation of data. Not every person that needs to use data will be able to digest it in a raw form. So to help people make decisions data science and analysis steps in to present the data in an easy to absorb medium. This is called data visualization and it is an easy to over look part of data science. Data visualization is basically marketing for your analysis of the data and in large part determines the impact of your analysis. Just look at infographics there is no denying that a nice info graphic with have more impact than a paper to the masses.&lt;br&gt;
&lt;a href="https://i.giphy.com/media/per3uJGPsapva/source.gif" class="article-body-image-wrapper"&gt;&lt;img src="https://i.giphy.com/media/per3uJGPsapva/source.gif" alt=""&gt;&lt;/a&gt;&lt;br&gt;
These points are all good general reasons that data science is a good choice of a career path, but for me the they aren’t the main draws of data science. For me the challenge of continuing to learn new things is very rewarding.&lt;/p&gt;

&lt;p&gt;I am just starting my journey into data science so I’m sure my ideas and perception of data will continue to change but I feel like this novices take on data science will be fun for me to look back on and compare to my future understanding.&lt;br&gt;
&lt;a href="https://i.giphy.com/media/lMeXjMBiORUI/giphy.gif" class="article-body-image-wrapper"&gt;&lt;img src="https://i.giphy.com/media/lMeXjMBiORUI/giphy.gif" alt=""&gt;&lt;/a&gt;&lt;/p&gt;

</description>
    </item>
  </channel>
</rss>
