<?xml version="1.0" encoding="UTF-8"?>
<rss version="2.0" xmlns:atom="http://www.w3.org/2005/Atom" xmlns:dc="http://purl.org/dc/elements/1.1/">
  <channel>
    <title>DEV Community: Steven Bruno</title>
    <description>The latest articles on DEV Community by Steven Bruno (@stevenbruno).</description>
    <link>https://dev.to/stevenbruno</link>
    <image>
      <url>https://media2.dev.to/dynamic/image/width=90,height=90,fit=cover,gravity=auto,format=auto/https:%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Fuser%2Fprofile_image%2F103055%2F7c1ccf5a-7f66-4e62-a2a6-4f9e6f9945bc.jpg</url>
      <title>DEV Community: Steven Bruno</title>
      <link>https://dev.to/stevenbruno</link>
    </image>
    <atom:link rel="self" type="application/rss+xml" href="https://dev.to/feed/stevenbruno"/>
    <language>en</language>
    <item>
      <title>An Introduction to One-class Classification</title>
      <dc:creator>Steven Bruno</dc:creator>
      <pubDate>Fri, 19 Apr 2019 04:42:49 +0000</pubDate>
      <link>https://dev.to/stevenbruno/an-introduction-to-one-class-classification-2m5c</link>
      <guid>https://dev.to/stevenbruno/an-introduction-to-one-class-classification-2m5c</guid>
      <description>&lt;h4&gt;
  
  
  The Problem Statement
&lt;/h4&gt;

&lt;p&gt;In statistics, the situation may arise where we must classify an object as belonging to group A or group B. When we have labeled training data for each class of object, the problem is fairly straightforward - we can utilize binary classification algorithms to predict the class to which a new object belongs. When we have unlabeled training data, we turn to clustering algorithms. So far so good, but how do we solve problems in which our training data only contains labeled objects for one class, and the rest are objects of an unknown class? Suddenly, the problem isn't so simple. To make matters worse, not even the trusty SKLearn Estimator Cheatsheet provides an answer.&lt;/p&gt;

&lt;p&gt;&lt;a href="https://media.dev.to/cdn-cgi/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fy7suiwd93m5a6vfygvkh.png" class="article-body-image-wrapper"&gt;&lt;img src="https://media.dev.to/cdn-cgi/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fy7suiwd93m5a6vfygvkh.png" alt="estimator" width="800" height="498"&gt;&lt;/a&gt;What about semi-labeled data?&lt;/p&gt;

&lt;h4&gt;
  
  
  Background
&lt;/h4&gt;

&lt;p&gt;I asked myself this very question when attempting to construct a model that would estimate the probability that a star beyond our solar system contains an exoplanet in its orbit. The NASA Exoplanet Archive contains a treasure trove of information detailing different star systems and exoplanets. Within the Kepler Stellar Dataset, astronomers have determined that many stars do in fact have exoplanets in their orbit. For the other stars, however, whether they host any orbiting planets is unknown. In this case, we have a data sample in which one class is labeled (star contains an exoplanet), and everything else is unlabeled (star may or may not contain an exoplanet). We have no labels for the case that a star does not have an exoplanet because it is extremely difficult, if not impossible, to say for certain that a star does not have any planets in its orbit. The objective is to construct and train a model that estimates the probability that a new observed test star contains an exoplanet in its orbit based on that test star's similarity to stars that are known to contain exoplanets. What options do we have?&lt;/p&gt;

&lt;h4&gt;
  
  
  The Solution (skip to here for tl;dr)
&lt;/h4&gt;

&lt;p&gt;The scenario I've outlined is what is known as &lt;a href="https://en.wikipedia.org/wiki/One-class_classification" rel="noopener noreferrer"&gt;One-class classification&lt;/a&gt;. There are numerous interpretations and applications outlined throughout scientific literature, but I will touch on some of the more popular concepts here.&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;
&lt;strong&gt;PU Learning&lt;/strong&gt; A binary classifier is learned in a semi-supervised way from only positive P and unlabeled data U. &lt;a href="https://roywright.me/2017/11/16/positive-unlabeled-learning/" rel="noopener noreferrer"&gt;Learn more&lt;/a&gt;
&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;Novelty and Outlier Detection&lt;/strong&gt; Decide whether a new observation belongs to the same distribution as existing observations (it is an inlier), or should be considered as different (it is an outlier). &lt;a href="https://scikit-learn.org/stable/modules/outlier_detection.html#outlier-detection" rel="noopener noreferrer"&gt;Learn more&lt;/a&gt;
&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;One Class SVM&lt;/strong&gt; An SVM approach to one-class classification. &lt;a href="http://users.cecs.anu.edu.au/~williams/papers/P126.pdf" rel="noopener noreferrer"&gt;Learn more&lt;/a&gt;
&lt;/li&gt;
&lt;/ul&gt;

&lt;h4&gt;
  
  
  Python Resources
&lt;/h4&gt;

&lt;p&gt;There are various ways to implement one-class classifiers in Python. A this point, I will defer to individuals who are much better equipped to discuss the implementation details of such models. &lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;Roy Wright provides a &lt;a href="https://roywright.me/2017/11/16/positive-unlabeled-learning/" rel="noopener noreferrer"&gt;detailed breakdown of PU learning&lt;/a&gt; and derives his own custom Python models.&lt;/li&gt;
&lt;li&gt;D.M. TAX develops a &lt;a href="http://homepage.tudelft.nl/n9d04/thesis.pdf" rel="noopener noreferrer"&gt;one-class SVM&lt;/a&gt;.&lt;/li&gt;
&lt;li&gt;SKLearn provides a &lt;a href="https://scikit-learn.org/stable/modules/generated/sklearn.svm.OneClassSVM.html#sklearn.svm.OneClassSVM" rel="noopener noreferrer"&gt;one-class svm method&lt;/a&gt;.&lt;/li&gt;
&lt;/ul&gt;

&lt;h4&gt;
  
  
  Conclusion
&lt;/h4&gt;

&lt;p&gt;I hope at the very least the resources I've provided above leave you better equipped to solve your own one-class classification problems. This is the third in a series of blog posts written for the &lt;a href="https://chipymentor.org/" rel="noopener noreferrer"&gt;ChiPy Mentorship Program&lt;/a&gt;. As part of my project, I am attempting to train models that will estimate the probability that a star has exoplanets in its orbit or habitable planets in its orbit. Identifying my problem as a one-class classification problem was an important leap in the progress of this project. My next steps are to implement a one-class SVM then analyze time-series data for various stars to try to identify &lt;a href="http://www.planetary.org/explore/space-topics/exoplanets/transit-photometry.html" rel="noopener noreferrer"&gt;minute dimming events&lt;/a&gt; that are indicative of an orbiting planet.&lt;/p&gt;

</description>
      <category>python</category>
      <category>machinelearning</category>
      <category>classification</category>
    </item>
    <item>
      <title>How to Explore NASA Exoplanet Archive Data with Python and Pandas</title>
      <dc:creator>Steven Bruno</dc:creator>
      <pubDate>Fri, 22 Mar 2019 04:40:04 +0000</pubDate>
      <link>https://dev.to/stevenbruno/how-to-explore-nasa-exoplanet-archive-data-with-python-and-pandas-351g</link>
      <guid>https://dev.to/stevenbruno/how-to-explore-nasa-exoplanet-archive-data-with-python-and-pandas-351g</guid>
      <description>&lt;h3&gt;
  
  
  &lt;strong&gt;Intro&lt;/strong&gt;
&lt;/h3&gt;

&lt;p&gt;The &lt;a href="https://exoplanetarchive.ipac.caltech.edu/index.html" rel="noopener noreferrer"&gt;NASA Exoplanet Archive&lt;/a&gt; hosts large data sets that feature exoplanets and their host stars. The site can be a little overwhelming for newcomers, however, as it is large and complex. The purpose of this post is to teach aspiring Exoplanet Archive data explorers how to navigate the site, download datasets, and begin manipulating the data with Pandas in Python. To frame the discussion, I will walk through the steps as I have recently performed them for my data science project in the &lt;a href="https://chipymentor.org/" rel="noopener noreferrer"&gt;ChiPy Mentorship Program&lt;/a&gt;. &lt;/p&gt;

&lt;h3&gt;
  
  
  &lt;strong&gt;Step 1&lt;/strong&gt; | Find your Dataset
&lt;/h3&gt;

&lt;p&gt;The majority of the data sets can be found by navigating to the &lt;em&gt;Data&lt;/em&gt; tab in the primary site navigation. Within, you will notice four main sections - Confirmed Planets, Kepler, Transit Surveys, and Other. For this example, lets work with the Composite Planet Data from within the confirmed exoplanets section. Once you open up the data set, you'll be presented with a tabular interface that you can scroll through directly in your browser. To learn more about individual columns or the data as a whole, scan through the provided &lt;em&gt;View Documentation&lt;/em&gt; links.&lt;/p&gt;

&lt;p&gt;&lt;a href="https://media.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fwtc42n8djjtst407ivzo.png" class="article-body-image-wrapper"&gt;&lt;img src="https://media.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fwtc42n8djjtst407ivzo.png" alt="Data tab"&gt;&lt;/a&gt;The available data sets&lt;/p&gt;

&lt;p&gt;&lt;a href="https://media.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fiuczm7eqtejj7t2gbcjl.png" class="article-body-image-wrapper"&gt;&lt;img src="https://media.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fiuczm7eqtejj7t2gbcjl.png" alt="View Documentation"&gt;&lt;/a&gt;View Documentation options&lt;/p&gt;

&lt;h3&gt;
  
  
  &lt;strong&gt;Step 2&lt;/strong&gt; | Filter and Download
&lt;/h3&gt;

&lt;p&gt;While you could certainly do all of your data cleaning with Pandas alone, you can get a head start by using the tools built into the archive webpage. Click on the select columns button, then select the columns you are interested in. You can further filter the data by searching for row entries and selecting and deselecting rows as you please. Each change you make will be reflected if your file download. Once ready, click the download button, and download your data in CSV format. If you tried to read the csv in Pandas at this point, you would encounter some errors, as the download brought some unwanted baggage. Open up your recently downloaded csv in a text editor, and delete everything up to the label for your first column. In my case, I will delete everything up to fpl_hostname. Save the updated file as a csv.&lt;/p&gt;

&lt;p&gt;&lt;a href="https://media.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fl96mmuto81pk1obmbr11.png" class="article-body-image-wrapper"&gt;&lt;img src="https://media.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fl96mmuto81pk1obmbr11.png" alt="Select Columns"&gt;&lt;/a&gt;The Select Columns button&lt;/p&gt;

&lt;p&gt;&lt;a href="https://media.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Ff2eplo6fhoen927acwqm.png" class="article-body-image-wrapper"&gt;&lt;img src="https://media.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Ff2eplo6fhoen927acwqm.png" alt="Column Controls"&gt;&lt;/a&gt;Column Controls&lt;/p&gt;

&lt;p&gt;&lt;a href="https://media.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fkebmwy0ivfsg9b0nz3lq.png" class="article-body-image-wrapper"&gt;&lt;img src="https://media.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fkebmwy0ivfsg9b0nz3lq.png" alt="Delete the baggage"&gt;&lt;/a&gt;Delete the extraneous information&lt;/p&gt;

&lt;h3&gt;
  
  
  &lt;strong&gt;Step 3&lt;/strong&gt; | Read File into Pandas Dataframe
&lt;/h3&gt;

&lt;p&gt;Now the fun part. I am assuming you have Python 3 and &lt;a href="https://pandas.pydata.org/" rel="noopener noreferrer"&gt;Pandas&lt;/a&gt; installed already, but if you don't, go ahead and do that. I like to work with the data in jupyter notebooks, so &lt;a href="https://jupyter.org/install" rel="noopener noreferrer"&gt;install Jupyter Notebook&lt;/a&gt; if you haven't done so already. Create a new jupyter notebook in the same directory where you are keeping your filtered data csv. In the notebook, begin by importing Pandas. Then use the Pandas built in read_csv() function to read your file into a new dataframe. Once done, run df.head() to make sure your dataframe looks good. &lt;/p&gt;

&lt;p&gt;&lt;code&gt;import Pandas as pd&lt;/code&gt;&lt;br&gt;
&lt;code&gt;df = pd.read_csv('name_of_filtered_csv')&lt;/code&gt;&lt;br&gt;
&lt;code&gt;df.head()&lt;/code&gt;&lt;/p&gt;

&lt;p&gt;&lt;a href="https://media.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2F3l93rzsesw7lx5vt5alg.png" class="article-body-image-wrapper"&gt;&lt;img src="https://media.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2F3l93rzsesw7lx5vt5alg.png" alt="Initial functions"&gt;&lt;/a&gt;Import your data&lt;/p&gt;

&lt;h3&gt;
  
  
  &lt;strong&gt;Step 4&lt;/strong&gt; | Clean your Data
&lt;/h3&gt;

&lt;p&gt;So far so good! At this point, you have a fully functioning exoplanet dataset in a Pandas dataframe. You may notice that the column headers aren't exactly intuitive. To remedy this, we will rename them using the df.rename() function. As a parameter, it expects a dictionary where the keys are old column names and the values are the new column names. We can set up that dictionary and use it in the rename function like so:&lt;/p&gt;

&lt;p&gt;&lt;code&gt;column_dict = {&lt;br&gt;
    'fpl_hostname':'Host Name', &lt;br&gt;
    'fpl_letter':'Planet Letter', &lt;br&gt;
    'fpl_name':'Planet Name', &lt;br&gt;
    'fpl_controvflag':'Controversial Flag'&lt;br&gt;
}&lt;/code&gt;&lt;br&gt;
&lt;code&gt;df.rename(columns=column_dict, inplace=True)&lt;/code&gt;&lt;br&gt;
&lt;code&gt;df.head()&lt;/code&gt;&lt;/p&gt;

&lt;p&gt;&lt;a href="https://media.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fk4ew7wxrpgginfu4pu1i.png" class="article-body-image-wrapper"&gt;&lt;img src="https://media.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fk4ew7wxrpgginfu4pu1i.png" alt="Renaming columns"&gt;&lt;/a&gt;Renaming columns&lt;/p&gt;

&lt;p&gt;If done correctly, you'll see the names of your columns modified to the new values. But what if we have more than just a few columns to rename. What if there are dozens, or hundreds? One way to address this is by extracting the html table for planet parameters accessible from the &lt;em&gt;View Documentation&lt;/em&gt; &amp;gt; &lt;em&gt;View Data Column Definitions&lt;/em&gt; link in the web interface for your data set. There, you'll see a table of column names, labels, and some other columns. The first two columns would work perfectly for keys (old column names) and values (new column names) within your dataframe. To approach this problem, you could capture the table using &lt;a href="https://pandas.pydata.org/pandas-docs/version/0.23.4/generated/pandas.read_html.html" rel="noopener noreferrer"&gt;pd.read_html()&lt;/a&gt;, then create a dictionary of the first two columns using &lt;a href="https://stackoverflow.com/questions/17426292/what-is-the-most-efficient-way-to-create-a-dictionary-of-two-pandas-dataframe-co" rel="noopener noreferrer"&gt;this method&lt;/a&gt; and then use the df.rename() function to finish renaming your original dataframe column labels. Good luck!&lt;/p&gt;

&lt;h5&gt;
  
  
  Final thoughts
&lt;/h5&gt;

&lt;p&gt;This has been an introduction to extracting data sets from the Nasa Exoplanet Archive using Pandas and Python. It is the second post in a blog trilogy I am writing for the &lt;a href="https://chipymentor.org/" rel="noopener noreferrer"&gt;ChiPy Mentorship Program&lt;/a&gt;. My next steps are to use methods in the sklearn.neighbors module to attempt to classify the likelihood a star has an orbiting exoplanet based on stellar parameters. I have performed all the steps outlined above and they are a great way to get started with creating model-ready dataframes from the exoplanet archive, but certainly not the only way. To learn about more methods for extracting data from the archive, check out the &lt;a href="https://exoplanetarchive.ipac.caltech.edu/docs/tools.html" rel="noopener noreferrer"&gt;Tools&lt;/a&gt; section within the site. &lt;/p&gt;

</description>
      <category>python</category>
      <category>pandas</category>
      <category>data</category>
      <category>space</category>
    </item>
    <item>
      <title>How to Start Your First Data Science Project</title>
      <dc:creator>Steven Bruno</dc:creator>
      <pubDate>Fri, 22 Feb 2019 05:46:23 +0000</pubDate>
      <link>https://dev.to/stevenbruno/how-to-start-your-first-data-science-project-4i8n</link>
      <guid>https://dev.to/stevenbruno/how-to-start-your-first-data-science-project-4i8n</guid>
      <description>&lt;p&gt;In recent years, data science has seen an influx of researchers, aspiring professionals, and enthusiasts. While many of the tools and techniques have been around for years, if not decades, the industry has really taken off in the past 5 years. &lt;a href="https://hbr.org/2012/10/data-scientist-the-sexiest-job-of-the-21st-century"&gt;Good publicity&lt;/a&gt;, increased technological capabilities, and high pay prospects  have combined to form the perfect recipe for a burgeoning field.&lt;/p&gt;

&lt;p&gt;The space is so enticing that even I have decided to test the waters. In the coming months, I will be creating my first data science project using python as part of my involvement in the &lt;a href="https://www.chipy.org/"&gt;Chicago Python user group&lt;/a&gt;  mentorship program.&lt;/p&gt;

&lt;p&gt;If you're in my position, you may be wondering how to get started. With the help of individuals much smarter than I am, I've crafted a 5 step sequence to get started. Without further ado:&lt;/p&gt;

&lt;h3&gt;
  
  
  &lt;strong&gt;Step 1&lt;/strong&gt; | Reconsider your intentions
&lt;/h3&gt;

&lt;p&gt;Its easy to get sucked into the glitz and glamour of the "data scientist" role. The title itself, however, can mean 100 different things to 100 different employers. If you are an aspiring professional data scientist, know that the title is suffering a bit of an identity crisis. Assuming you've done your research, aren't just in it for the alluring pay, and are still interested, then continue to step two.&lt;/p&gt;

&lt;h3&gt;
  
  
  &lt;strong&gt;Step 2&lt;/strong&gt; | Gain a basic familiarity with the tools and techniques
&lt;/h3&gt;

&lt;p&gt;Data Science is essentially programming and statistics. The two most popular languages at the moment for the practice are Python and R. It would not be a bad idea to take an online data science course and pick up some fundamental statistics before you begin. This will also help you ascertain whether the type of work is even something you would be interested in. I've scanned countless recommendations from online communities over the past year and these courses seem to receive a lot of good feedback. &lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;&lt;a href="https://www.udemy.com/python-for-data-science-and-machine-learning-bootcamp/"&gt;Python for Data Science and Machine Learning Bootcamp&lt;/a&gt;&lt;/li&gt;
&lt;li&gt;&lt;a href="https://lagunita.stanford.edu/courses/HumanitiesSciences/StatLearning/Winter2016/about"&gt;Stanford Statistical Learning&lt;/a&gt;&lt;/li&gt;
&lt;li&gt;&lt;a href="https://www.coursera.org/learn/machine-learning"&gt;Machine Learning from Andrew Ng&lt;/a&gt;&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;I have worked through a decent portion of the Jose Portilla Udemy course in preparation for my project, and I find the instruction to be outstanding.&lt;br&gt;
A word of caution: Do not get sucked into the trap of the infinite tutorial loop. Just get the basics down then try to tackle your own project as soon as you can. While the courses listed above are great resources, you will learn a tremendous amount by solving the problems that arise in your own project. You will also have something cool to show for it. &lt;/p&gt;

&lt;h3&gt;
  
  
  &lt;strong&gt;Step 3&lt;/strong&gt; | Choose a topic you're interested in
&lt;/h3&gt;

&lt;p&gt;This one is pretty self explanatory. If you're struggling the think of a project idea, just consider your own interests. From &lt;a href="https://www.kaggle.com/moltean/fruits"&gt;apples&lt;/a&gt; to &lt;a href="https://www.kaggle.com/uciml/zoo-animal-classification"&gt;zoos&lt;/a&gt;, if you can think of a subject, there's a good chance data exists for it. The best project idea is the one you actually stick to, so choosing something that excites you is in your best interest.&lt;/p&gt;

&lt;h3&gt;
  
  
  &lt;strong&gt;Step 4&lt;/strong&gt; | Hone in on the data
&lt;/h3&gt;

&lt;p&gt;Got your topic? Great, now its time to select your dataset(s). &lt;a href="https://www.kaggle.com/"&gt;Kaggle&lt;/a&gt; is an amazing resource, as is &lt;a href="https://toolbox.google.com/datasetsearch"&gt;Google's dataset search&lt;/a&gt;. For me, I like astronomy, so I'm looking at the &lt;a href="https://exoplanetarchive.ipac.caltech.edu/"&gt;NASA Exoplanet Archive&lt;/a&gt; and beginning to envision the sorts of relationships and models that can potentially be drawn out. &lt;em&gt;Side Note&lt;/em&gt; - web developers, please make data scientist's lives easier by &lt;a href="https://productforums.google.com/forum/#!topic/webmasters/nPq4BW6iPIA"&gt;allowing&lt;/a&gt; Google to find and publicize your data with their search tool. (Unless you are anti Google, which is okay too) &lt;/p&gt;

&lt;h3&gt;
  
  
  &lt;strong&gt;Step 5&lt;/strong&gt; | Dive into your data!
&lt;/h3&gt;

&lt;p&gt;At this point, you've got the tools, techniques, and data to get started. Get stuck? There are plenty of online communities willing to help out. Good luck!&lt;/p&gt;

&lt;p&gt;&lt;em&gt;Author's Note | This has been the first in a series of blog posts for the &lt;a href="https://chipymentor.org/"&gt;ChiPy mentorship program&lt;/a&gt;. My next steps are to extract and clean my data, so keep an eye out for my next post if you are interested in my progress.&lt;/em&gt;&lt;/p&gt;

</description>
      <category>datascience</category>
      <category>python</category>
      <category>ai</category>
      <category>pandas</category>
    </item>
  </channel>
</rss>
