<?xml version="1.0" encoding="UTF-8"?>
<rss version="2.0" xmlns:atom="http://www.w3.org/2005/Atom" xmlns:dc="http://purl.org/dc/elements/1.1/">
  <channel>
    <title>DEV Community: ezgigm</title>
    <description>The latest articles on DEV Community by ezgigm (@ezgigm).</description>
    <link>https://dev.to/ezgigm</link>
    <image>
      <url>https://media2.dev.to/dynamic/image/width=90,height=90,fit=cover,gravity=auto,format=auto/https:%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Fuser%2Fprofile_image%2F344312%2Febd61885-c9fd-4c9a-9cc5-a981fbc58874.jpeg</url>
      <title>DEV Community: ezgigm</title>
      <link>https://dev.to/ezgigm</link>
    </image>
    <atom:link rel="self" type="application/rss+xml" href="https://dev.to/feed/ezgigm"/>
    <language>en</language>
    <item>
      <title>My Data Science Bootcamp Experience</title>
      <dc:creator>ezgigm</dc:creator>
      <pubDate>Sun, 07 Jun 2020 02:21:09 +0000</pubDate>
      <link>https://dev.to/ezgigm/my-data-science-bootcamp-experience-3d79</link>
      <guid>https://dev.to/ezgigm/my-data-science-bootcamp-experience-3d79</guid>
      <description>&lt;p&gt;In February, I have enrolled Flatiron Immersive Data Science program. I preferred to enroll full-time in-person program instead of self-paced or online because I thought that to be on the campus makes me more motivated. In this post, I will write about my experiences with Bootcamp. &lt;/p&gt;

&lt;p&gt;First of all, I did not have big expectations when I was starting the program. Of course, I knew the syllabus and details about it, but I did not know how much we practice and work on hands-on projects. In the boot camp, I finished 5 modules. Each module has different goals and I had to finish the assignment and project of each module for graduation. Project weeks were really tough to survive, even I felt like my final weeks in university. Certainly, the most difficult one was the capstone project which we have worked for the last 3 weeks.&lt;/p&gt;

&lt;p&gt;Now, when I look at my first project, I can obviously see how much I have improved my knowledge and skills. Here, I would like to say 'Thank you' to our lead instructor &lt;a href="https://www.linkedin.com/in/bryan-arnold-mathematics/" rel="noopener noreferrer"&gt;Bryan Arnold&lt;/a&gt; and our coach &lt;a href="https://www.linkedin.com/in/lindseyberlin/" rel="noopener noreferrer"&gt;Lindsey Berlin&lt;/a&gt;. They always supported me to push my boundaries. In the first project, I got my data by using the API key and web-scrabbing. It took two days to understand how to get my data. If they wrote the code for me and gave me ready code, I would not learn how to do it myself. Instead of giving the code to me, they gave me the perspective on how to solve this problem. This is the approach that I was used to from university, so it helps me a lot. &lt;/p&gt;

&lt;p&gt;Because of Covid-19, we continued to our last 3 modules remotely. It is easier to focus on campus rather than home exactly. But, I experienced different skills in this remote system such as;&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;How to working in a team remotely&lt;/li&gt;
&lt;li&gt;Being self-sufficient&lt;/li&gt;
&lt;li&gt;Focusing on online lectures&lt;/li&gt;
&lt;li&gt;How to get online help from instructors&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;Flatiron's full-time remote schedule helped me a lot to keep my motivation. When I began to feel more relaxed, it kept me engaged with its strict schedule and syllabus. Even, we had accountability channel with our instructors which reminded our goals every day. &lt;/p&gt;

&lt;p&gt;For the capstone project, I worked on sentiment analysis with deep learning and build book recommendation systems from Kindle Store reviews on Amazon, even it was impossible to imagine for me when I was in the first module. I feel pretty comfortable about what I learned in Flatiron and it was worth the risk to leave my job. If you have any questions about Bootcamp, please feel free to ask about my experience. To see details of my projects, my GitHub repo can be found &lt;a href="https://github.com/ezgigm" rel="noopener noreferrer"&gt;here&lt;/a&gt;. &lt;br&gt;
&lt;a href="https://media.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fi%2Fzsa5kruw75fi28jw482k.gif" class="article-body-image-wrapper"&gt;&lt;img src="https://media.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fi%2Fzsa5kruw75fi28jw482k.gif" alt="Alt Text"&gt;&lt;/a&gt;&lt;br&gt;
&lt;em&gt;Cover photo by Glenn Carstens-Peters on &lt;a href="https://unsplash.com/s/photos/blog" rel="noopener noreferrer"&gt;Unsplash&lt;/a&gt; and gif is from &lt;a href="https://giphy.com" rel="noopener noreferrer"&gt;giphy&lt;/a&gt;&lt;/em&gt;&lt;/p&gt;

</description>
      <category>datascience</category>
      <category>machinelearning</category>
      <category>python</category>
    </item>
    <item>
      <title>Kindle Store Reviews - Sentiment Analysis and Recommendation Systems</title>
      <dc:creator>ezgigm</dc:creator>
      <pubDate>Fri, 29 May 2020 17:21:20 +0000</pubDate>
      <link>https://dev.to/ezgigm/kindle-store-reviews-sentiment-analysis-and-recommendation-systems-2p4m</link>
      <guid>https://dev.to/ezgigm/kindle-store-reviews-sentiment-analysis-and-recommendation-systems-2p4m</guid>
      <description>&lt;p&gt;As a capstone project of Flatiron School, I worked on Kindle Store reviews on Amazon. &lt;/p&gt;

&lt;h2&gt;
  
  
  Why did I choose this project?
&lt;/h2&gt;

&lt;p&gt;With the increasing demand in e-commerce, voice of customer concerns getting more important such as reviews or surveys. Customers could not touch online products so they want to learn from previous customer's experiences. If the company fails to meet this expectation, it can lose money. In this project, I choose Kindle store reviews because it is easy to buy and read books with Kindle and Kindle app is available for every device. So, there are many people who prefer to buy online books and give feedback about them online. I worked on book reviews but the important point of this project for me is that it can be applied for different companies and different areas. I found this online business problem interesting so, I tried to solve it. &lt;/p&gt;

&lt;h2&gt;
  
  
  How did I solve this problem?
&lt;/h2&gt;

&lt;p&gt;I have two main solutions. Firstly, I did sentiment analysis to classify reviews as positive or negative. It is hard to read thousands of reviews regularly and understand which product is liked or hated by customers. It takes too much time and generally does not give brief idea about product comparison. I build a model which can predict the review is positive or negative from text and saving our time. &lt;/p&gt;

&lt;p&gt;My second solution is to build a recommendation system to meet customers' expectations with recommend them good books and sell more products. &lt;/p&gt;

&lt;h2&gt;
  
  
  Models
&lt;/h2&gt;

&lt;p&gt;I tried different models for sentiment analysis. Firstly, I tried statistical models as LogReg, DecisionTree, ExtraTree, RandomForest, XGBoost, LGBM Classifiers. Between them, I found best result 87% accuracy. I decided to try neural nets. As neural net models, I built Torch model with Fast Text Class. Then, I tried CNN and RNN with different layer numbers and layer types on Keras. At last, I used transfer learning technique with using Pre-trained BERT model which gave me 95% for small sample data and 97% for large-scale data. The results with small-scale data were inserted below;&lt;/p&gt;

&lt;p&gt;&lt;a href="https://media.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fi%2Fbbfu1iy4hr5tejxeireh.png" class="article-body-image-wrapper"&gt;&lt;img src="https://media.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fi%2Fbbfu1iy4hr5tejxeireh.png" alt="Alt Text"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;h2&gt;
  
  
  Recommendation Systems
&lt;/h2&gt;

&lt;p&gt;As a recommender system, I used collaborative filtering which finds similar users according to past ratings on same products and recommend user B the book which liked by also user A. Although this system has many advantages such as easy to find interested areas of customers and adding data to system easily, it has one big disadvantage as cold-start problem.&lt;/p&gt;

&lt;p&gt;To solve this issue, I built a system which finds similarities between keywords and reviews. It is easy to write keywords in search tab and take recommendations but what if I do not have any other information for product than reviews, or maybe best book for me is not best book for new user. So, I prefer to calculate text similarities between keywords and past reviews of other users. Then, I find the best N similar users and give recommendation to new user according to this. &lt;/p&gt;

&lt;h2&gt;
  
  
  Challenges of This Project?
&lt;/h2&gt;

&lt;ul&gt;
&lt;li&gt;The first challenge was that the data is very big to analyze. There are more than 2.2 million reviews with 139,815 users and 98,824 books. I used meta data as helper dataset to match book titles with book ID's. Every trial can take hours even my personal computer can not run this model for this whole data. So, I found different solutions. Firstly, to run models easily and compare, I took sample data from whole set and after decided my model with small set, I run my model for whole set. I used Google Colab also to run some models which can take hours.&lt;/li&gt;
&lt;li&gt;The second challenges of the project was to build neural networks. Before using transfer learning technique, I tried to train Torch and Keras models. It is not easy to understand which layer is necessary or what the layers works for. &lt;/li&gt;
&lt;li&gt;As a third challenge, it is not easy to say this recommendation system is well. Although we are trying to guess likes or dislikes of people, it is hard to predict human nature.&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;As a result, this project can help companies to save time with not to read many reviews for comparing products. If I can recommend good books, people can buy more products and advice my online store to others. Also, this project give opportunity to compare products easily and define dislikes quickly. And if the customer spend less time on server with finding interested books easily, it means fewer problems for my server.&lt;/p&gt;

&lt;p&gt;All details and codes can be found in my &lt;a href="https://github.com/ezgigm/sentiment_analysis_and_product_recommendation" rel="noopener noreferrer"&gt;GitHub repo&lt;/a&gt;.&lt;/p&gt;

&lt;p&gt;As a personal gain, I learned a lot of different techniques, models and I got more information about deep learning. It was a good hands-on learning experience for me about neural networks. &lt;/p&gt;

&lt;p&gt;As general feelings, I am very happy with my level which I have reached in 15 weeks. When I compares my first projects with capstone, I feel that I have improved myself a lot. I am grateful to our instructor Bryan Arnold and our coach Lindsey Berlin for helping and supporting us always to push the limits. And special thanks to my dear husband, I am grateful to him who encourages me always to try my best. &lt;/p&gt;

&lt;p&gt;&lt;a href="https://media.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fi%2Fy90mwe4arn424h9m9z7v.gif" class="article-body-image-wrapper"&gt;&lt;img src="https://media.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fi%2Fy90mwe4arn424h9m9z7v.gif" alt="Alt Text"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;&lt;em&gt;Cover image by &lt;a href="https://pixabay.com/" rel="noopener noreferrer"&gt;pixabay&lt;/a&gt; and gif is from &lt;a href="//giphy.com"&gt;giphy&lt;/a&gt;&lt;/em&gt;&lt;/p&gt;

</description>
      <category>sentimentanalysis</category>
      <category>datascience</category>
      <category>machinelearning</category>
      <category>python</category>
    </item>
    <item>
      <title>Walmart Store Sales Forecasting</title>
      <dc:creator>ezgigm</dc:creator>
      <pubDate>Fri, 08 May 2020 18:08:06 +0000</pubDate>
      <link>https://dev.to/ezgigm/walmart-store-sales-forecasting-3c67</link>
      <guid>https://dev.to/ezgigm/walmart-store-sales-forecasting-3c67</guid>
      <description>&lt;p&gt;All companies have long-term and short-term plans. Especially, big companies like Walmart must make sure about the effectiveness of future plans. Sales forecasting is one of the very important plans because it gives idea to the company for arranging stocks, calculating revenue and deciding to make new investment. Another advantage of knowing future sales is that achieving predetermined targets from the beginning of the seasons can have a positive effect on stock prices and investors' perceptions. Also, not reaching the projected target could significantly damage stock prices, conversely. And, it will be a big problem especially for Walmart as a big company.&lt;/p&gt;

&lt;p&gt;Walmart began &lt;a href="https://www.kaggle.com/c/walmart-recruiting-store-sales-forecasting/overview"&gt;recruiting competition&lt;/a&gt; for store sales forecasting on Kaggle. As a Module 4 project of Flatiron School Data Science Bootcamp, I worked on this competition. The reason for choosing this project is to learn more about time series models and to learn deeply which parameters effect sales forecasting. My previous work before bootcamp was about sales, so I also would like to observe my seasonal effect experiences scientifically.&lt;/p&gt;

&lt;p&gt;The data was obtained from &lt;a href="https://www.kaggle.com/c/walmart-recruiting-store-sales-forecasting/data"&gt;Kaggle&lt;/a&gt;. In the data, there are mainly weekly sales of departments and stores, holidays, type and size of stores and some external features such as consumer price index, fuel price, temperature and unemployment rate. &lt;/p&gt;

&lt;h1&gt;
  
  
  Main Challenge of This Project
&lt;/h1&gt;

&lt;p&gt;&lt;a href="https://res.cloudinary.com/practicaldev/image/fetch/s--6NW20C2v--/c_limit%2Cf_auto%2Cfl_progressive%2Cq_66%2Cw_880/https://dev-to-uploads.s3.amazonaws.com/i/ian1xgle36x6r6m397sv.gif" class="article-body-image-wrapper"&gt;&lt;img src="https://res.cloudinary.com/practicaldev/image/fetch/s--6NW20C2v--/c_limit%2Cf_auto%2Cfl_progressive%2Cq_66%2Cw_880/https://dev-to-uploads.s3.amazonaws.com/i/ian1xgle36x6r6m397sv.gif" alt="Alt Text"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;The main challenge of this project for me is seasonal sales. Some seasons have higher sales values like Thanksgiving, Black Friday and Christmas. These seasonal effects make the data highly non-stationary. Dealing with non-stationary data is not easy. There are many ways to solve this problem. &lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;Differencing : We can calculate the difference of consecutive terms in data. Differencing is generally done to get rid of the varying average. &lt;/li&gt;
&lt;li&gt;Seasonal Differencing : The idea behind this technique is taking the difference of same seasons like the difference between 2010 and 2011 for the first week of January. &lt;/li&gt;
&lt;li&gt;Transforming Data: This technique is generally used for non-constant variances. We can transform data with taking log, square root or power transform, mainly. We can also shift the data. &lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;The best solution which I found for making my data more stationary is that I took differences of consecutive terms and used this more stationary data for modeling. From the below graph, it can be seen easily how my data is non-stationary. I improves this also with Adfuller test. Details about Adfuller test can be found &lt;a href="https://en.wikipedia.org/wiki/Augmented_Dickey%E2%80%93Fuller_test"&gt;here&lt;/a&gt;.&lt;/p&gt;

&lt;p&gt;&lt;a href="https://res.cloudinary.com/practicaldev/image/fetch/s--fWBwIWCp--/c_limit%2Cf_auto%2Cfl_progressive%2Cq_auto%2Cw_880/https://dev-to-uploads.s3.amazonaws.com/i/2ix244coljs3u1pbs6g5.png" class="article-body-image-wrapper"&gt;&lt;img src="https://res.cloudinary.com/practicaldev/image/fetch/s--fWBwIWCp--/c_limit%2Cf_auto%2Cfl_progressive%2Cq_auto%2Cw_880/https://dev-to-uploads.s3.amazonaws.com/i/2ix244coljs3u1pbs6g5.png" alt="Alt Text"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;h1&gt;
  
  
  Metric and Models
&lt;/h1&gt;

&lt;p&gt;The metric of the competition is weighted mean absolute error (WMAE). Weight of the error changes when it is holiday. It calculates the error of holidays 5 times more than normal days such as if I predict 50$ wrongly for usual days, it means 250$ error for holiday predictions. I think, Walmart wants to give weight for true holiday predictions because holiday sales are more higher and important than usual times. Details of metric can be found &lt;a href="https://github.com/ezgigm/Project4_Store_Sales_Forecasting/blob/master/STEP2_Random_Forest_Regressor.ipynb"&gt;in this link&lt;/a&gt;.&lt;/p&gt;

&lt;p&gt;After trying some machine learning and time series models, I found best results with Exponential Smoothing(Holt-Winters) which flattens the trend of data. My best error value is 821. If we assume our weekly sales averages between 18000-20000 dollars, it means my model can predict future sales of Walmart stores with 4% error. &lt;/p&gt;

&lt;h1&gt;
  
  
  Explorations
&lt;/h1&gt;

&lt;p&gt;&lt;a href="https://res.cloudinary.com/practicaldev/image/fetch/s--yy1Goeeu--/c_limit%2Cf_auto%2Cfl_progressive%2Cq_auto%2Cw_880/https://dev-to-uploads.s3.amazonaws.com/i/yvccqon729hx7iexqqcp.png" class="article-body-image-wrapper"&gt;&lt;img src="https://res.cloudinary.com/practicaldev/image/fetch/s--yy1Goeeu--/c_limit%2Cf_auto%2Cfl_progressive%2Cq_auto%2Cw_880/https://dev-to-uploads.s3.amazonaws.com/i/yvccqon729hx7iexqqcp.png" alt="Alt Text"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;When we look at the average sales monthly, it is obviously seen that November and December have higher sales. To make deep analysis, I looked at the weekly sales with corresponding week numbers. &lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;&lt;a href="https://res.cloudinary.com/practicaldev/image/fetch/s--ZbH5WGUL--/c_limit%2Cf_auto%2Cfl_progressive%2Cq_auto%2Cw_880/https://dev-to-uploads.s3.amazonaws.com/i/wonjfw0axkwl2c0qfrlk.png" class="article-body-image-wrapper"&gt;&lt;img src="https://res.cloudinary.com/practicaldev/image/fetch/s--ZbH5WGUL--/c_limit%2Cf_auto%2Cfl_progressive%2Cq_auto%2Cw_880/https://dev-to-uploads.s3.amazonaws.com/i/wonjfw0axkwl2c0qfrlk.png" alt="Alt Text"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;&lt;p&gt;From the graph, we can understand easily 51st and 47th weeks have higher sales. It means that higher demands in whole year belong to Christmas, Thanksgiving and Black Friday seasons. 50th and 49th weeks are following top 2. And, it is not obviously seen from the graph but I found it from values in project that 22nd week is in the top 5. It means that after schools are closed at the end of the May, people prefer to shop for summer or holiday. &lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;Another interesting exploration is that when I looked at the top highest sales, I found it in Thanksgiving time. But, when we look at the graph above, Christmas has higher averages. It means that, maybe Thanksgiving week has higher values individually for some stores and some departments but when we take the average and sort them, Christmas is the best for general average weekly sales. &lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;Although during other months sales are closed to each others, January has the least sales average. This is the result of November and December high sales. After two high sales month, people prefer to pay less on January.&lt;/p&gt;&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;&lt;a href="https://res.cloudinary.com/practicaldev/image/fetch/s--n8iXPyw2--/c_limit%2Cf_auto%2Cfl_progressive%2Cq_auto%2Cw_880/https://dev-to-uploads.s3.amazonaws.com/i/6zsbaj9orig6dd823dqn.png" class="article-body-image-wrapper"&gt;&lt;img src="https://res.cloudinary.com/practicaldev/image/fetch/s--n8iXPyw2--/c_limit%2Cf_auto%2Cfl_progressive%2Cq_auto%2Cw_880/https://dev-to-uploads.s3.amazonaws.com/i/6zsbaj9orig6dd823dqn.png" alt="Alt Text"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;&lt;a href="https://res.cloudinary.com/practicaldev/image/fetch/s--X0pqjITE--/c_limit%2Cf_auto%2Cfl_progressive%2Cq_auto%2Cw_880/https://dev-to-uploads.s3.amazonaws.com/i/1r6k06cwj098sw05upri.png" class="article-body-image-wrapper"&gt;&lt;img src="https://res.cloudinary.com/practicaldev/image/fetch/s--X0pqjITE--/c_limit%2Cf_auto%2Cfl_progressive%2Cq_auto%2Cw_880/https://dev-to-uploads.s3.amazonaws.com/i/1r6k06cwj098sw05upri.png" alt="Alt Text"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;From pie plot and box plot above, we can understand that there are mainly 3 types of Walmart stores in the data and Walmart categorized them according to sizes. Type A is the biggest store, B is medium and C is the smallest. And, nearly half of the stores in the data are Type A stores which are biggest.&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;&lt;a href="https://res.cloudinary.com/practicaldev/image/fetch/s--dHdUd5or--/c_limit%2Cf_auto%2Cfl_progressive%2Cq_auto%2Cw_880/https://dev-to-uploads.s3.amazonaws.com/i/qcroujti1yas8rmltkt9.png" class="article-body-image-wrapper"&gt;&lt;img src="https://res.cloudinary.com/practicaldev/image/fetch/s--dHdUd5or--/c_limit%2Cf_auto%2Cfl_progressive%2Cq_auto%2Cw_880/https://dev-to-uploads.s3.amazonaws.com/i/qcroujti1yas8rmltkt9.png" alt="Alt Text"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;In the graph above, red line shows the average holiday sales and green line represents the average non-holiday sales. According to lines, holiday average is higher than non-holiday, as expected. &lt;/li&gt;
&lt;li&gt;It is obviously seen from the graph that size of the store is directly proportional to weekly sales. Bigger stores have higher average sales. &lt;/li&gt;
&lt;li&gt;&lt;p&gt;And, it is interesting that according to this graph, Thanksgiving sales are shown higher than Christmas but we know from the weekly graph, 51th week is the best. When I looked deeper to the data, I realized that Walmart assigned Christmas sales to the 31-Dec-10, 30-Dec-11, 28-Dec-12, 27-Dec-13. But we know that people generally do their Christmas shopping before last week of the year. So, it is not a good idea to assign Christmas sale season to the last week of the year. The problem in the data can be understood easily from this graph.&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;Interestingly, according to this data consumer price index, temperature, fuel price and unemployment rate do not have any pattern on weekly sales.&lt;/p&gt;&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;&lt;a href="https://res.cloudinary.com/practicaldev/image/fetch/s--UlY316eR--/c_limit%2Cf_auto%2Cfl_progressive%2Cq_auto%2Cw_880/https://dev-to-uploads.s3.amazonaws.com/i/bupriohbqn0rpl3mvi16.png" class="article-body-image-wrapper"&gt;&lt;img src="https://res.cloudinary.com/practicaldev/image/fetch/s--UlY316eR--/c_limit%2Cf_auto%2Cfl_progressive%2Cq_auto%2Cw_880/https://dev-to-uploads.s3.amazonaws.com/i/bupriohbqn0rpl3mvi16.png" alt="Alt Text"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;&lt;a href="https://res.cloudinary.com/practicaldev/image/fetch/s--FkZlmSiS--/c_limit%2Cf_auto%2Cfl_progressive%2Cq_auto%2Cw_880/https://dev-to-uploads.s3.amazonaws.com/i/fnhaexlelufcif10q3dq.png" class="article-body-image-wrapper"&gt;&lt;img src="https://res.cloudinary.com/practicaldev/image/fetch/s--FkZlmSiS--/c_limit%2Cf_auto%2Cfl_progressive%2Cq_auto%2Cw_880/https://dev-to-uploads.s3.amazonaws.com/i/fnhaexlelufcif10q3dq.png" alt="Alt Text"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;The first graph above shows that department 72 has some higher values, when I looked at the numerical data also I understood that especially on Thanksgiving department 72 has high sales. But, from the second graph above, I found some other department with higher average sales. It shows that, although department 72 has good sales during Thanksgiving, some other departments such as 92 or 95 have higher sales averages in general. This graph shows us one more time the importance of seasonal sales and misleading points of seasonal sales. &lt;/li&gt;
&lt;/ul&gt;

&lt;h1&gt;
  
  
  Future Improvement and Recommendations
&lt;/h1&gt;

&lt;ul&gt;
&lt;li&gt;As a modeling perspective, data will be made more stationary with different techniques, feature selection and feature engineering can be added to model also. &lt;/li&gt;
&lt;li&gt;As a data quality perspective, to improve to results, more data is needed, which has more detailed seasonal sales such as Come Back to School, 4th of July, Memorial Day, Halloween or Easter. Before these holidays, some department can sell more products. &lt;/li&gt;
&lt;li&gt;There are markdowns and store sales for some special seasons. The effect of markdowns and discounts on departments can be added to the model. &lt;/li&gt;
&lt;li&gt;From the data, I observed that some holidays has higher effect of some stores and some departments. So, different models can be build for each store and departments. It is not easy to build model for 45 different stores and 81 different departments but it can be begun with the departments and stores which have higher sales. Because, they have higher effect on total sales than others. &lt;/li&gt;
&lt;li&gt;And lastly, market basket analysis can be added to the data, to find higher demand items of departments.&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;All details for cleaning process, data preprocessing, modeling process, more explorations and solutions for problem and future improvements can be found in this &lt;a href="https://github.com/ezgigm/Project4_Store_Sales_Forecasting"&gt;github repo&lt;/a&gt;.&lt;/p&gt;

&lt;p&gt;&lt;em&gt;Cover image by &lt;a href="https://pixabay.com/users/PublicDomainPictures-14/?utm_source=link-attribution&amp;amp;utm_medium=referral&amp;amp;utm_campaign=image&amp;amp;utm_content=163464"&gt;PublicDomainPictures&lt;/a&gt; from &lt;a href="https://pixabay.com/?utm_source=link-attribution&amp;amp;utm_medium=referral&amp;amp;utm_campaign=image&amp;amp;utm_content=163464"&gt;Pixabay&lt;/a&gt; and gif is from &lt;a href="//giphy.com"&gt;giphy&lt;/a&gt;&lt;/em&gt;&lt;/p&gt;

</description>
      <category>machinelearning</category>
      <category>datascience</category>
      <category>python</category>
      <category>timeseries</category>
    </item>
    <item>
      <title>Pump it Up: Data Mining the Water Table</title>
      <dc:creator>ezgigm</dc:creator>
      <pubDate>Mon, 20 Apr 2020 15:31:30 +0000</pubDate>
      <link>https://dev.to/ezgigm/pump-it-up-data-mining-the-water-table-2f6n</link>
      <guid>https://dev.to/ezgigm/pump-it-up-data-mining-the-water-table-2f6n</guid>
      <description>&lt;p&gt;Despite the unbelievable development of technology, simple basic needs such as access to clean drinking water are still one of the most important problems of human beings. For some areas in the world, to find clean water, pump this water up and transport this water to people are really difficult processes. Tanzania is the largest country of East-Africa with 59 million population. 25 million of this population have lack access to clean water, 40 million people also have a lack access to improved sanitation. The Tanzanian Water Ministry agreed with Taarifa and they begun a competition by &lt;a href="https://www.drivendata.org/competitions/7/pump-it-up-data-mining-the-water-table/"&gt;DrivenData&lt;/a&gt; to solve this problem with improving clean water sources.  As a Module 3 project of Flatiron School Data Science Bootcamp, I worked on this problem with &lt;a href="https://github.com/geomms"&gt;Mark Subra&lt;/a&gt;. The reason for choosing this project is my interest in solving the main problems that concern humanity. I have worked in NGO for many years to help people. Even, I have two close friends who live in Tanzania and always tell me about the water shortage of this country. So, I have an interest this type of problems.  &lt;/p&gt;

&lt;p&gt;Water points were divided in three classes as functional, non-functional or functional but needs repair by water ministry. Our aim in this project to build a model which predicts the functionality of water points. The data was taken from &lt;a href="https://www.drivendata.org/competitions/7/pump-it-up-data-mining-the-water-table/"&gt;DrivenData&lt;/a&gt;. Basically, there are 4 different datasets; submission format, training set, test set and train labels set which contains status of wells. With given training set and labels set, competitors are wanted to build predictive model and apply it to test set to determine status of the wells and submit. Train set contains 59400 water points data with 40 features.&lt;/p&gt;

&lt;h1&gt;
  
  
  Challenges of This Project
&lt;/h1&gt;

&lt;p&gt;&lt;a href="https://res.cloudinary.com/practicaldev/image/fetch/s--SBgxJhSe--/c_limit%2Cf_auto%2Cfl_progressive%2Cq_66%2Cw_880/https://dev-to-uploads.s3.amazonaws.com/i/0td1qd92ns0g39c7y4f3.gif" class="article-body-image-wrapper"&gt;&lt;img src="https://res.cloudinary.com/practicaldev/image/fetch/s--SBgxJhSe--/c_limit%2Cf_auto%2Cfl_progressive%2Cq_66%2Cw_880/https://dev-to-uploads.s3.amazonaws.com/i/0td1qd92ns0g39c7y4f3.gif" alt="Alt Text"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;h4&gt;
  
  
  Importance of Data Cleaning
&lt;/h4&gt;

&lt;p&gt;Mainly, there are two challenges in this data. The first one is to clean the data. Because, it contains lots of columns which has same information. These columns cause multi-collinearity in model. Also, there are many null, zero and missing values. Generally, features are categorical and some of them has more than 2000 unique values. There are spelling mistakes in some columns which creates high unique values. Lastly, some columns has  discrete values. So, we dropped some columns which contains same information, converted null and missing values to mean or collected them in unknown category. For feature engineering, we created new columns for some features and categorized them again manually. This cleaning process took too much time but at the end, we understood the importance of data cleaning again. Because, for the first modeling trial as a baseline with simple logistic regression our model gave 0.83 roc-auc score for binary class. &lt;/p&gt;

&lt;h4&gt;
  
  
  Imbalanced Ternary Class Problem
&lt;/h4&gt;

&lt;p&gt;Our data has highly imbalanced three target labels and all three of them are important to predict as true. So, we have to find the balanced values for each label in confusion matrix. To simplify this problem, firstly we collected functional and functional but needs repair wells together and found the best model for binary class. With doing this, we understood the which model approach is true for this data and how to set our parameters for models. When we converted our data to ternary class we also faced with imbalance problem. One of our target labels has very less value than others. To solve this problem, we used oversampling technique SMOTE (Synthetic Minority Over-Sampling Technique). SMOTE works on the idea of nearest neighbors and create its synthetic data. It is one of the most popular techniques for oversampling. &lt;/p&gt;

&lt;h1&gt;
  
  
  Models and Metric
&lt;/h1&gt;

&lt;p&gt;To prepare our data to machine learning, we did encoding and scaling also for dealing with first challenge. For ternary target model, Target Encoder and Robust Scaler were used. Random Forest, LGBM and XGBoost were tried. To handle the second challenge as imbalanced target problem, SMOTE over-sampling technique was applied. The metric for competition was balanced accuracy. So, we used this metric and sometimes to compare and check results one more, we used roc-auc score. &lt;/p&gt;

&lt;h1&gt;
  
  
  Explorations of Problems for Solution
&lt;/h1&gt;

&lt;p&gt;&lt;a href="https://res.cloudinary.com/practicaldev/image/fetch/s--vB-wakVN--/c_limit%2Cf_auto%2Cfl_progressive%2Cq_auto%2Cw_880/https://dev-to-uploads.s3.amazonaws.com/i/qeadtf3qqnr8h4ottg20.png" class="article-body-image-wrapper"&gt;&lt;img src="https://res.cloudinary.com/practicaldev/image/fetch/s--vB-wakVN--/c_limit%2Cf_auto%2Cfl_progressive%2Cq_auto%2Cw_880/https://dev-to-uploads.s3.amazonaws.com/i/qeadtf3qqnr8h4ottg20.png" alt="Alt Text"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;From the graph, it is shown that there are many wells which contains enough water but non-functional. Also, it is observed that 4272 wells were dried but they have good water quality. With finding a solution to give source again these wells, they can be functional. Finding clean water sources is not the only problem, to continue to feed these sources are equally important. 2226 (7%) wells have enough and soft water but needs repair. Authorities must invest on repairing. Otherwise these will be non-functional. 8035 (27%) wells has enough, good quality water but they are non-functional. This shows that authorities must work and invest on technology to pump these good sources.&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;&lt;a href="https://res.cloudinary.com/practicaldev/image/fetch/s--gdIA-rqJ--/c_limit%2Cf_auto%2Cfl_progressive%2Cq_auto%2Cw_880/https://dev-to-uploads.s3.amazonaws.com/i/m9no4mt6jm8mxiocle7t.png" class="article-body-image-wrapper"&gt;&lt;img src="https://res.cloudinary.com/practicaldev/image/fetch/s--gdIA-rqJ--/c_limit%2Cf_auto%2Cfl_progressive%2Cq_auto%2Cw_880/https://dev-to-uploads.s3.amazonaws.com/i/m9no4mt6jm8mxiocle7t.png" alt="Alt Text"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;3500 water wells need repairs, otherwise they would be non-functional easily.&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;&lt;a href="https://res.cloudinary.com/practicaldev/image/fetch/s--iGM_wTHH--/c_limit%2Cf_auto%2Cfl_progressive%2Cq_auto%2Cw_880/https://dev-to-uploads.s3.amazonaws.com/i/o9jkhdlynhiswzxeaxm7.png" class="article-body-image-wrapper"&gt;&lt;img src="https://res.cloudinary.com/practicaldev/image/fetch/s--iGM_wTHH--/c_limit%2Cf_auto%2Cfl_progressive%2Cq_auto%2Cw_880/https://dev-to-uploads.s3.amazonaws.com/i/o9jkhdlynhiswzxeaxm7.png" alt="Alt Text"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;This graph shows the highest ratio of functional wells to non-functional by funder. Danida is Tanzania, Denmark cooperation for wells and has many functional wells. RWSSP is Rural Water Supply and Sanitation Program. Also, most of the wells which was funded by Germany Republic are functional.&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;&lt;a href="https://res.cloudinary.com/practicaldev/image/fetch/s--gXvc7W5r--/c_limit%2Cf_auto%2Cfl_progressive%2Cq_auto%2Cw_880/https://dev-to-uploads.s3.amazonaws.com/i/3p328uioiy06kt4lyj3a.png" class="article-body-image-wrapper"&gt;&lt;img src="https://res.cloudinary.com/practicaldev/image/fetch/s--gXvc7W5r--/c_limit%2Cf_auto%2Cfl_progressive%2Cq_auto%2Cw_880/https://dev-to-uploads.s3.amazonaws.com/i/3p328uioiy06kt4lyj3a.png" alt="Alt Text"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;&lt;p&gt;Mostly the wells which are funded by government are non-functional. Most of water points which central government and district council installed are non-functional.&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;Darul es Salaam is one of the highest populated cities but 35% of good water quality points are non-functional.&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;Iringa is one of the important areas but it contains lots of non-functional water points which has soft water.&lt;/p&gt;&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;&lt;a href="https://res.cloudinary.com/practicaldev/image/fetch/s--QZCAcH74--/c_limit%2Cf_auto%2Cfl_progressive%2Cq_auto%2Cw_880/https://dev-to-uploads.s3.amazonaws.com/i/vf5stomki86d5rd3h3ld.png" class="article-body-image-wrapper"&gt;&lt;img src="https://res.cloudinary.com/practicaldev/image/fetch/s--QZCAcH74--/c_limit%2Cf_auto%2Cfl_progressive%2Cq_auto%2Cw_880/https://dev-to-uploads.s3.amazonaws.com/i/vf5stomki86d5rd3h3ld.png" alt="Alt Text"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;&lt;p&gt;The most common extraction type is gravity but second is hand pumps. It is seen that, there are many non-functional water points which belongs to gravity (which is natural force so no need to do anything expensive) as extraction type. So, gravity type wells do not need too much investment on it. So, there can be found more water points which can be functional easily. &lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;The wells which have constructed in recent years are functional then older ones. And it is seen that recent years have some functional but needs repair wells. It means that if they will not be repaired recently, they will be non-functional easily.&lt;/p&gt;&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;&lt;a href="https://res.cloudinary.com/practicaldev/image/fetch/s--p_mbKHu5--/c_limit%2Cf_auto%2Cfl_progressive%2Cq_auto%2Cw_880/https://dev-to-uploads.s3.amazonaws.com/i/gzv7jcs7zaridgm2d0jq.png" class="article-body-image-wrapper"&gt;&lt;img src="https://res.cloudinary.com/practicaldev/image/fetch/s--p_mbKHu5--/c_limit%2Cf_auto%2Cfl_progressive%2Cq_auto%2Cw_880/https://dev-to-uploads.s3.amazonaws.com/i/gzv7jcs7zaridgm2d0jq.png" alt="Alt Text"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;&lt;p&gt;This map shows the location of functional but needs repair wells locations. There are some clusters around highly populated areas. With the  regular maintenance of this wells, more people can find clean water.&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;Water basin is also another important parameter for functionality of wells. The areas which has near to good water basin high probability to find clean water.&lt;/p&gt;&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;&lt;a href="https://res.cloudinary.com/practicaldev/image/fetch/s--JOvJI286--/c_limit%2Cf_auto%2Cfl_progressive%2Cq_auto%2Cw_880/https://dev-to-uploads.s3.amazonaws.com/i/903aevdftchkyhml96he.png" class="article-body-image-wrapper"&gt;&lt;img src="https://res.cloudinary.com/practicaldev/image/fetch/s--JOvJI286--/c_limit%2Cf_auto%2Cfl_progressive%2Cq_auto%2Cw_880/https://dev-to-uploads.s3.amazonaws.com/i/903aevdftchkyhml96he.png" alt="Alt Text"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;Wells with no fee are more likely to be non-functional, and wells with some form of payment are more likely to be functional.&lt;/li&gt;
&lt;/ul&gt;

&lt;h1&gt;
  
  
  Solutions
&lt;/h1&gt;

&lt;p&gt;Our model can predict the functionality of the water wells with 86% accuracy. With the good prediction of functionality, the solutions can be;&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;prioritizing functioning wells which need repair and yield clean water&lt;/li&gt;
&lt;li&gt;targeting repairs to clusters of wells especially those with high populations&lt;/li&gt;
&lt;li&gt;payments of some kind will provide incentive to keep wells functional&lt;/li&gt;
&lt;li&gt;allocate funds and resources to effective organizations with track record&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;All details for cleaning process, data preprocessing, modeling process, more explorations and solutions for problem and future improvements can be found in &lt;a href="https://github.com/ezgigm/Project3_TanzanianWaterWell"&gt;this github repo&lt;/a&gt;. &lt;/p&gt;

&lt;p&gt;&lt;em&gt;Pump image courtesy of &lt;a href="https://www.flickr.com/photos/christophercjensen/3559607145"&gt;flickr user christophercjensen&lt;/a&gt; and gif is from &lt;a href="//giphy.com"&gt;giphy&lt;/a&gt;&lt;/em&gt;&lt;/p&gt;

</description>
      <category>machinelearning</category>
      <category>datascience</category>
      <category>python</category>
      <category>xgboost</category>
    </item>
    <item>
      <title> Data Visualization Basic Libraries</title>
      <dc:creator>ezgigm</dc:creator>
      <pubDate>Mon, 23 Mar 2020 06:37:17 +0000</pubDate>
      <link>https://dev.to/ezgigm/data-visualization-basic-libraries-4gdi</link>
      <guid>https://dev.to/ezgigm/data-visualization-basic-libraries-4gdi</guid>
      <description>&lt;p&gt;One of the advantages of using Python is user-friendly visualization packages. In this blog, I try to explain some essential plots for data visualization and I will share with you which plots I used for what before and what I found useful. Mainly, the plots which I will share in this article belongs to Matplotlib which is a library for Python using for visualization. Also, I would like to state about Seaborn, visualization library based on matplotlib, because I like Seaborn's colorful visuals. It is all about my usage preferences from my experiences.&lt;/p&gt;

&lt;p&gt;&lt;a href="https://i.giphy.com/media/Pn4vFaA3h38KA/giphy.gif" class="article-body-image-wrapper"&gt;&lt;img src="https://i.giphy.com/media/Pn4vFaA3h38KA/giphy.gif" alt="Alt Text"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;h3&gt;
  
  
  Scatter Plot  &lt;em&gt;(plt.scatter)&lt;/em&gt;
&lt;/h3&gt;

&lt;p&gt;First of all, we need to import &lt;em&gt;matplotlib.pyplot&lt;/em&gt; like all other libraries, to use these plot types.&lt;/p&gt;

&lt;p&gt;Scatter plots are very useful to understand the correlation between two variables easily. Also, I like to put markers for data points like 'stars','triangles' or 'sized-circles' according to the frequency of data. The most important feature which I like to use is scatter plot with a legend. &lt;/p&gt;

&lt;p&gt;&lt;a href="https://res.cloudinary.com/practicaldev/image/fetch/s--FTOghoJO--/c_limit%2Cf_auto%2Cfl_progressive%2Cq_auto%2Cw_880/https://dev-to-uploads.s3.amazonaws.com/i/6eb0jd4kmto6pokrsv68.png" class="article-body-image-wrapper"&gt;&lt;img src="https://res.cloudinary.com/practicaldev/image/fetch/s--FTOghoJO--/c_limit%2Cf_auto%2Cfl_progressive%2Cq_auto%2Cw_880/https://dev-to-uploads.s3.amazonaws.com/i/6eb0jd4kmto6pokrsv68.png" alt="Alt Text"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;Another feature that I like about scatter plot is adding a color bar with &lt;em&gt;plt.colorbar()&lt;/em&gt;. It is very useful to see the colors and recognize different points. In the below graph, I used the 'viridis' as perceptually uniform sequential colormap scale but there are lots of alternative sequential, diverging, qualitative, miscellaneous color bars in 'matplotlib'.&lt;/p&gt;

&lt;p&gt;&lt;a href="https://res.cloudinary.com/practicaldev/image/fetch/s---MgdsnMW--/c_limit%2Cf_auto%2Cfl_progressive%2Cq_auto%2Cw_880/https://dev-to-uploads.s3.amazonaws.com/i/wp2hi7d37hdzhglcfvkl.png" class="article-body-image-wrapper"&gt;&lt;img src="https://res.cloudinary.com/practicaldev/image/fetch/s---MgdsnMW--/c_limit%2Cf_auto%2Cfl_progressive%2Cq_auto%2Cw_880/https://dev-to-uploads.s3.amazonaws.com/i/wp2hi7d37hdzhglcfvkl.png" alt="Alt Text"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;All of them can be found in &lt;a href="https://matplotlib.org/3.1.1/tutorials/colors/colormaps.html"&gt;matplotlib.org&lt;/a&gt;.&lt;/p&gt;

&lt;h3&gt;
  
  
  Bar Plot &lt;em&gt;(plt.bar)&lt;/em&gt;
&lt;/h3&gt;

&lt;p&gt;I prefer bar plots to compare categorical data and differences between them. It is super easy to see the lengths of rectangular bars which proportional to data. Bars can be sorted according to y-axis values and the variables from best to worst easily can be seen. &lt;/p&gt;

&lt;p&gt;&lt;a href="https://res.cloudinary.com/practicaldev/image/fetch/s--ek3cq3zK--/c_limit%2Cf_auto%2Cfl_progressive%2Cq_auto%2Cw_880/https://dev-to-uploads.s3.amazonaws.com/i/d7spam4dlcz5vcbntjtl.png" class="article-body-image-wrapper"&gt;&lt;img src="https://res.cloudinary.com/practicaldev/image/fetch/s--ek3cq3zK--/c_limit%2Cf_auto%2Cfl_progressive%2Cq_auto%2Cw_880/https://dev-to-uploads.s3.amazonaws.com/i/d7spam4dlcz5vcbntjtl.png" alt="Alt Text"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;I like to color the bars also and compare two variables in same the graph with the bar plot. With this method, we can compare lots of variables with each other proportionally. Additionally, it is also good to arrange the width of the bars. &lt;/p&gt;

&lt;p&gt;&lt;a href="https://res.cloudinary.com/practicaldev/image/fetch/s--pzWjqB12--/c_limit%2Cf_auto%2Cfl_progressive%2Cq_auto%2Cw_880/https://dev-to-uploads.s3.amazonaws.com/i/k22zpt77fyc83xfiuou4.png" class="article-body-image-wrapper"&gt;&lt;img src="https://res.cloudinary.com/practicaldev/image/fetch/s--pzWjqB12--/c_limit%2Cf_auto%2Cfl_progressive%2Cq_auto%2Cw_880/https://dev-to-uploads.s3.amazonaws.com/i/k22zpt77fyc83xfiuou4.png" alt="Alt Text"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;h3&gt;
  
  
  Pie plot &lt;em&gt;(plt.pie)&lt;/em&gt;
&lt;/h3&gt;

&lt;p&gt;When I would like to see the percentages of a small number of variables in the same graph, I like to use pie plots. &lt;/p&gt;

&lt;p&gt;&lt;a href="https://res.cloudinary.com/practicaldev/image/fetch/s--tqC9p5nd--/c_limit%2Cf_auto%2Cfl_progressive%2Cq_auto%2Cw_880/https://dev-to-uploads.s3.amazonaws.com/i/sorxiyslcsixr836depa.png" class="article-body-image-wrapper"&gt;&lt;img src="https://res.cloudinary.com/practicaldev/image/fetch/s--tqC9p5nd--/c_limit%2Cf_auto%2Cfl_progressive%2Cq_auto%2Cw_880/https://dev-to-uploads.s3.amazonaws.com/i/sorxiyslcsixr836depa.png" alt="Alt Text"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;Also, nested pie-plot charts can be created for more complex data but I do not prefer complexity for interpreting data. Another good visual thing, which I like about the pie plot is we can separate one slice from the pie plot and show the importance of this part. &lt;/p&gt;

&lt;h3&gt;
  
  
  Histogram and Density Plot &lt;em&gt;(plt.hist)&lt;/em&gt;
&lt;/h3&gt;

&lt;p&gt;Undoubtedly, one of the best ways to see the distribution of the data is histograms. It is super easy to see the normally distributed variables which are a very important concept for data scientists. So, histograms are very essential graphs for data scientists. &lt;/p&gt;

&lt;p&gt;&lt;a href="https://res.cloudinary.com/practicaldev/image/fetch/s--ONHWNWRK--/c_limit%2Cf_auto%2Cfl_progressive%2Cq_auto%2Cw_880/https://dev-to-uploads.s3.amazonaws.com/i/axdo2pf0bw3cx1hd8ba5.png" class="article-body-image-wrapper"&gt;&lt;img src="https://res.cloudinary.com/practicaldev/image/fetch/s--ONHWNWRK--/c_limit%2Cf_auto%2Cfl_progressive%2Cq_auto%2Cw_880/https://dev-to-uploads.s3.amazonaws.com/i/axdo2pf0bw3cx1hd8ba5.png" alt="Alt Text"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;Although it is a great way to begin understanding a single variable distribution, histograms can fail when we try to compare distributions of one variable across multiple categories. Because, when histograms get together in the same area, readability will be the problem. The best way to do this is to create a new histogram for each category. Also, a side-by-side histograms or stacked bars can be useful. &lt;/p&gt;

&lt;p&gt;There is a good example in &lt;a href="https://towardsdatascience.com/histograms-and-density-plots-in-python-f6bda88f5ac0"&gt;towardsdatascience.com&lt;/a&gt; with NYC flight data about when histograms fail and how to solve this by using side-by-side histogram or stacked bars. &lt;/p&gt;

&lt;p&gt;When density plots are get together with the histograms of the distribution of data, the graph will be more useful also. I want to state here about Seaborn which is super easy to use and Seaborn can create a histogram and density curve on the same plot easily.&lt;/p&gt;

&lt;h3&gt;
  
  
  Seaborn &lt;em&gt;(sns.pairplot , sns.heatmap , sns.distplot, sns.barplot)&lt;/em&gt;
&lt;/h3&gt;

&lt;p&gt;Seaborn is a really useful and colorful library in Python. Like 'matplotlib' , we need to import Seaborn also before using it. &lt;/p&gt;

&lt;p&gt;The most useful basic plots in Seaborn are pair plot, heat map, distribution plot. I also added a bar plot here because I like to use Seaborn's easy colorful bar plot. You do not need to give specific colors so, it is easier than 'matplotlib'. The other thing which I like about Seaborn is using fewer syntax.&lt;/p&gt;

&lt;p&gt;&lt;a href="https://res.cloudinary.com/practicaldev/image/fetch/s--1lMYv5Gv--/c_limit%2Cf_auto%2Cfl_progressive%2Cq_auto%2Cw_880/https://dev-to-uploads.s3.amazonaws.com/i/re1v0ajl0t3drnefu0bv.png" class="article-body-image-wrapper"&gt;&lt;img src="https://res.cloudinary.com/practicaldev/image/fetch/s--1lMYv5Gv--/c_limit%2Cf_auto%2Cfl_progressive%2Cq_auto%2Cw_880/https://dev-to-uploads.s3.amazonaws.com/i/re1v0ajl0t3drnefu0bv.png" alt="Alt Text"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;There are many ready datasets in Seaborn. This is the other important thing which I like. As a data scientist candidate, it is very good to find ready data sets to play around visualizing them. In the graph below, I plotted the distribution and also density curves for ready data set 'iris' which can be downloaded as &lt;em&gt;sns.load_dataset('iris')&lt;/em&gt; easily. &lt;/p&gt;

&lt;p&gt;&lt;a href="https://res.cloudinary.com/practicaldev/image/fetch/s--eaP22kao--/c_limit%2Cf_auto%2Cfl_progressive%2Cq_auto%2Cw_880/https://dev-to-uploads.s3.amazonaws.com/i/y2jxkh95jcp1s0yvxeqt.png" class="article-body-image-wrapper"&gt;&lt;img src="https://res.cloudinary.com/practicaldev/image/fetch/s--eaP22kao--/c_limit%2Cf_auto%2Cfl_progressive%2Cq_auto%2Cw_880/https://dev-to-uploads.s3.amazonaws.com/i/y2jxkh95jcp1s0yvxeqt.png" alt="Alt Text"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;Another important thing for modeling is the correlation between independent variables. To analyze this correlation, Seaborn's sns.pairplot and sns.heatmap are useful functions. &lt;/p&gt;

&lt;p&gt;&lt;a href="https://res.cloudinary.com/practicaldev/image/fetch/s--d_sw2OX---/c_limit%2Cf_auto%2Cfl_progressive%2Cq_auto%2Cw_880/https://dev-to-uploads.s3.amazonaws.com/i/1lgicpjgiw2dovj6zu42.png" class="article-body-image-wrapper"&gt;&lt;img src="https://res.cloudinary.com/practicaldev/image/fetch/s--d_sw2OX---/c_limit%2Cf_auto%2Cfl_progressive%2Cq_auto%2Cw_880/https://dev-to-uploads.s3.amazonaws.com/i/1lgicpjgiw2dovj6zu42.png" alt="Alt Text"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;From the data set 'auto-mpg.csv', I choose some columns for simplification and just put my data set in pairplot as sns.pairplot(df). It is a super fast, easy and useful way of Seaborn. Then with the graph above, I can observe the correlations, linearity and also distribution of variables. So, sns.pairplot is a very important feature of Python.&lt;/p&gt;

&lt;p&gt;It is also the same for heatmap.&lt;/p&gt;

&lt;p&gt;&lt;a href="https://res.cloudinary.com/practicaldev/image/fetch/s--5eRp8ojL--/c_limit%2Cf_auto%2Cfl_progressive%2Cq_auto%2Cw_880/https://dev-to-uploads.s3.amazonaws.com/i/3ybzelq0f1p4y9a7qurl.png" class="article-body-image-wrapper"&gt;&lt;img src="https://res.cloudinary.com/practicaldev/image/fetch/s--5eRp8ojL--/c_limit%2Cf_auto%2Cfl_progressive%2Cq_auto%2Cw_880/https://dev-to-uploads.s3.amazonaws.com/i/3ybzelq0f1p4y9a7qurl.png" alt="Alt Text"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;Just a simple one-line code I can observe all correlations with target and multicollinearity between variables super easily. &lt;/p&gt;

&lt;p&gt;Lastly, there are many ways to visualize data in Python and also in other programming languages. But, these topics are really basic, useful and also easy to use for beginners. In addition to this, these are the fastest ways to observe datasets. As seen in this blog post, visualization library or tool always depends on the data and what you want to obtain.&lt;/p&gt;

&lt;p&gt;&lt;em&gt;Cover image sourced by &lt;a href="https://pixabay.com/users/ColiN00B-346653/?utm_source=link-attribution&amp;amp;utm_medium=referral&amp;amp;utm_campaign=image&amp;amp;utm_content=3068300"&gt;Colin Behrens&lt;/a&gt; from &lt;a href="https://pixabay.com/?utm_source=link-attribution&amp;amp;utm_medium=referral&amp;amp;utm_campaign=image&amp;amp;utm_content=3068300"&gt;Pixabay&lt;/a&gt; and gif is from &lt;a href="//giphy.com"&gt;giphy&lt;/a&gt;.&lt;/em&gt;&lt;/p&gt;

</description>
      <category>python</category>
      <category>seaborn</category>
      <category>matplotlib</category>
      <category>datascience</category>
    </item>
    <item>
      <title>Data : The Superpower of Future </title>
      <dc:creator>ezgigm</dc:creator>
      <pubDate>Mon, 02 Mar 2020 04:55:46 +0000</pubDate>
      <link>https://dev.to/ezgigm/data-the-superpower-of-future-518j</link>
      <guid>https://dev.to/ezgigm/data-the-superpower-of-future-518j</guid>
      <description>&lt;p&gt;The demand for data scientists in the market increases day by day. I also decided to learn data science and enrolled Flatiron School Data Science Bootcamp. Although there are many popular tech areas nowadays, I will try to explain why I decided to data science in this post.&lt;/p&gt;

&lt;h3&gt;
  
  
  Personal Interest
&lt;/h3&gt;

&lt;p&gt;The most important reason of my decision is obviously my personal interest of math and coding. Despite I graduated from Bs. Chemical Engineering, my favorite courses were always about math and coding during my university years. So, I always want to continue my education with Ms. Computer Science. After I realized data science concept, I understood that this is the area what I am looking for, not computer science. It is more specific area with the combination of math, programming, statistics, problem-solving, the ability to interpret captured data and with processing this data build and offer business solutions. This combination perfectly matches with my expectations.&lt;/p&gt;

&lt;h3&gt;
  
  
  Demand in Market
&lt;/h3&gt;

&lt;p&gt;Big data is one of the most important resources for companies future. Companies can build new models for future plans and investments with using previous data. The importance of data science has increased with sharing big data (individual or corporate) via internet. Thanks to internet, people can easily reach important data about topic and can guess the future plan of this area. But, this data is not always clean and useful for everyone. Mostly, the data which is easily reachable, needs to clean and prepare for next model. So, the demand in the market has increasing everyday for data scientist.&lt;/p&gt;

&lt;h3&gt;
  
  
  Use of Data Science in Healthcare
&lt;/h3&gt;

&lt;p&gt;&lt;a href="https://res.cloudinary.com/practicaldev/image/fetch/s--oF6VDXOf--/c_limit%2Cf_auto%2Cfl_progressive%2Cq_auto%2Cw_880/https://dev-to-uploads.s3.amazonaws.com/i/kjdfyqkq9yhlhe6yxbld.png" class="article-body-image-wrapper"&gt;&lt;img src="https://res.cloudinary.com/practicaldev/image/fetch/s--oF6VDXOf--/c_limit%2Cf_auto%2Cfl_progressive%2Cq_auto%2Cw_880/https://dev-to-uploads.s3.amazonaws.com/i/kjdfyqkq9yhlhe6yxbld.png" alt="Alt Text"&gt;&lt;/a&gt;&lt;br&gt;
Data science not only offers solutions for companies, but also offers solutions and models for issues that concern all humanity. Healthcare is one of the these common sense areas. Data science has many applications in healthcare and medicine and I really consider some of these applications like genetics and genomics. All the data about our bodies are stored in human's genes individually. If this database can be commented in perfect way, the impact of genetic variations on genetic code can be understood and according to this interpretation maybe medicine for fatal illnesses can be found by scientists. All of these steps in healthcare such as interpretation of DNA, creating medicine, accuracy of diagnosis, keeping patience data and offering models for illnesses or virtual assistance to patients, contain lots of data and need data scientists for handling. One of the reasons of my decision is exactly these applications of data science which relates to all humanity. Even though it is difficult to make precise predictions at the beginning of the road, this field will be the area that I would like to work on after graduation.&lt;/p&gt;

&lt;h3&gt;
  
  
  Long-term Learning Process - like Marathon, not 100m Run
&lt;/h3&gt;

&lt;p&gt;Lastly, I love learning and learning must be the life-long for me about my job. And, data science is the perfect area for this. It has very broad applications and like all other tech jobs it is very dynamic. You have to be aware of innovations and must learn new concepts. You should find solutions for different cases in market and to do this you have to search lots of sources and interpret them. I always believe that a rolling stone gathers no moss. So, I love to learn and to level up my knowledge.&lt;/p&gt;

&lt;p&gt;In addition to lots of causes, these four headings are the main reasons of why I began data science journey.&lt;/p&gt;

&lt;p&gt;&lt;a href="https://res.cloudinary.com/practicaldev/image/fetch/s--4YUab4RP--/c_limit%2Cf_auto%2Cfl_progressive%2Cq_66%2Cw_880/https://dev-to-uploads.s3.amazonaws.com/i/09tlemdf1l0453gz0umk.gif" class="article-body-image-wrapper"&gt;&lt;img src="https://res.cloudinary.com/practicaldev/image/fetch/s--4YUab4RP--/c_limit%2Cf_auto%2Cfl_progressive%2Cq_66%2Cw_880/https://dev-to-uploads.s3.amazonaws.com/i/09tlemdf1l0453gz0umk.gif" alt="Alt Text"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;&lt;em&gt;Two images sourced from &lt;a href="//www.pixabay.com"&gt;pixabay.com&lt;/a&gt; and gif is from &lt;a href="//giphy.com"&gt;giphy&lt;/a&gt;.&lt;/em&gt;&lt;/p&gt;

</description>
      <category>datascience</category>
      <category>python</category>
      <category>newcareer</category>
    </item>
  </channel>
</rss>
