<?xml version="1.0" encoding="UTF-8"?>
<rss version="2.0" xmlns:atom="http://www.w3.org/2005/Atom" xmlns:dc="http://purl.org/dc/elements/1.1/">
  <channel>
    <title>DEV Community: Scraper.AI</title>
    <description>The latest articles on DEV Community by Scraper.AI (@scraper_ai).</description>
    <link>https://dev.to/scraper_ai</link>
    <image>
      <url>https://media2.dev.to/dynamic/image/width=90,height=90,fit=cover,gravity=auto,format=auto/https:%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Fuser%2Fprofile_image%2F442783%2Ff2096af2-13a9-45a6-939a-5a09ba2380e0.jpg</url>
      <title>DEV Community: Scraper.AI</title>
      <link>https://dev.to/scraper_ai</link>
    </image>
    <atom:link rel="self" type="application/rss+xml" href="https://dev.to/feed/scraper_ai"/>
    <language>en</language>
    <item>
      <title>Checking the availability of NVIDIA and AMD Graphic Cards and CPUs</title>
      <dc:creator>Scraper.AI</dc:creator>
      <pubDate>Fri, 08 Jan 2021 11:58:17 +0000</pubDate>
      <link>https://dev.to/scraper_ai/checking-the-availability-of-nvidia-and-amd-graphic-cards-and-cpus-5na</link>
      <guid>https://dev.to/scraper_ai/checking-the-availability-of-nvidia-and-amd-graphic-cards-and-cpus-5na</guid>
      <description>&lt;p&gt;The new NVIDIA RTX3070, RTX3080 and RTX3090 Graphic cards have just been released, but buying one is going to be an adventure on its own! Currently there is almost no stock, and if there is, Scalpers who monitor these stocks and steal them from you are going to make your life difficult!&lt;/p&gt;

&lt;p&gt;But what if you could play a scalper yourself, and only needed 5 minutes to do so?! Well let me introduce a way on how you can easily monitor these websites yourself and get notified when the stock changes!&lt;/p&gt;

&lt;p&gt;Let's start off by loading the&lt;a href="https://www.nvidia.com/de-de/shop/geforce/gpu/?page=1&amp;amp;limit=9&amp;amp;locale=de-de&amp;amp;category=GPU&amp;amp;gpu=RTX%203080"&gt; NVIDIA Website for the Founders Edition card&lt;/a&gt; (I used the German website).&lt;/p&gt;

&lt;p&gt;&lt;a href="https://res.cloudinary.com/practicaldev/image/fetch/s--Vdg8VxIG--/c_limit%2Cf_auto%2Cfl_progressive%2Cq_auto%2Cw_880/https://blog.scraper.ai/_posts/monitoring/checking-availability-of-nvidia-rtx-3080-cards/1.png" class="article-body-image-wrapper"&gt;&lt;img src="https://res.cloudinary.com/practicaldev/image/fetch/s--Vdg8VxIG--/c_limit%2Cf_auto%2Cfl_progressive%2Cq_auto%2Cw_880/https://blog.scraper.ai/_posts/monitoring/checking-availability-of-nvidia-rtx-3080-cards/1.png" alt=""&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;With Scraper.AI we can now easily state that we want to extract the price and details, but also put an interval on how frequently we wish to extract information! So open up the extension and select that you wish to &lt;strong&gt;"Monitor Data"&lt;/strong&gt;&lt;/p&gt;

&lt;p&gt;&lt;a href="https://res.cloudinary.com/practicaldev/image/fetch/s--StVopDLc--/c_limit%2Cf_auto%2Cfl_progressive%2Cq_auto%2Cw_880/https://blog.scraper.ai/_posts/monitoring/checking-availability-of-nvidia-rtx-3080-cards/2.png" class="article-body-image-wrapper"&gt;&lt;img src="https://res.cloudinary.com/practicaldev/image/fetch/s--StVopDLc--/c_limit%2Cf_auto%2Cfl_progressive%2Cq_auto%2Cw_880/https://blog.scraper.ai/_posts/monitoring/checking-availability-of-nvidia-rtx-3080-cards/2.png" alt=""&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;The main part will now open up and you will be able to utilize the &lt;strong&gt;Single&lt;/strong&gt; mode to select non-repeating content such as the Price, Image and Title of the graphic card.&lt;/p&gt;

&lt;p&gt;&lt;a href="https://res.cloudinary.com/practicaldev/image/fetch/s--c_Rhzc-g--/c_limit%2Cf_auto%2Cfl_progressive%2Cq_auto%2Cw_880/https://blog.scraper.ai/_posts/monitoring/checking-availability-of-nvidia-rtx-3080-cards/3.png" class="article-body-image-wrapper"&gt;&lt;img src="https://res.cloudinary.com/practicaldev/image/fetch/s--c_Rhzc-g--/c_limit%2Cf_auto%2Cfl_progressive%2Cq_auto%2Cw_880/https://blog.scraper.ai/_posts/monitoring/checking-availability-of-nvidia-rtx-3080-cards/3.png" alt=""&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;Once that is done, just specify the interval you wish to extract the data for and select &lt;strong&gt;Finish&lt;/strong&gt;&lt;/p&gt;

&lt;p&gt;&lt;a href="https://res.cloudinary.com/practicaldev/image/fetch/s--OL4BdvYi--/c_limit%2Cf_auto%2Cfl_progressive%2Cq_auto%2Cw_880/https://blog.scraper.ai/_posts/monitoring/checking-availability-of-nvidia-rtx-3080-cards/4.png" class="article-body-image-wrapper"&gt;&lt;img src="https://res.cloudinary.com/practicaldev/image/fetch/s--OL4BdvYi--/c_limit%2Cf_auto%2Cfl_progressive%2Cq_auto%2Cw_880/https://blog.scraper.ai/_posts/monitoring/checking-availability-of-nvidia-rtx-3080-cards/4.png" alt=""&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;Your data is now being extracted and the dashboard will show up with your extracted record!&lt;/p&gt;

&lt;p&gt;&lt;a href="https://res.cloudinary.com/practicaldev/image/fetch/s--yVzuXFrU--/c_limit%2Cf_auto%2Cfl_progressive%2Cq_auto%2Cw_880/https://blog.scraper.ai/_posts/monitoring/checking-availability-of-nvidia-rtx-3080-cards/5.png" class="article-body-image-wrapper"&gt;&lt;img src="https://res.cloudinary.com/practicaldev/image/fetch/s--yVzuXFrU--/c_limit%2Cf_auto%2Cfl_progressive%2Cq_auto%2Cw_880/https://blog.scraper.ai/_posts/monitoring/checking-availability-of-nvidia-rtx-3080-cards/5.png" alt=""&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;Based on the interval specified in the extension (or as you can also see in the &lt;strong&gt;Schedule **section). Your data will now be updated!&lt;/strong&gt; A Final thing** to do however, is to make sure that we get notified of changes! So open up the **Notifications **section and fill in your email!&lt;/p&gt;

&lt;p&gt;&lt;a href="https://res.cloudinary.com/practicaldev/image/fetch/s--oDEI5a8m--/c_limit%2Cf_auto%2Cfl_progressive%2Cq_auto%2Cw_880/https://blog.scraper.ai/_posts/monitoring/checking-availability-of-nvidia-rtx-3080-cards/6.png" class="article-body-image-wrapper"&gt;&lt;img src="https://res.cloudinary.com/practicaldev/image/fetch/s--oDEI5a8m--/c_limit%2Cf_auto%2Cfl_progressive%2Cq_auto%2Cw_880/https://blog.scraper.ai/_posts/monitoring/checking-availability-of-nvidia-rtx-3080-cards/6.png" alt=""&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;Next time data is updated, you will now be notified of any changes through your email! &lt;/p&gt;

</description>
      <category>showdev</category>
      <category>startup</category>
      <category>monitoring</category>
      <category>node</category>
    </item>
    <item>
      <title>The 11 best free web scraping tools that can use proxies compared</title>
      <dc:creator>Scraper.AI</dc:creator>
      <pubDate>Fri, 30 Oct 2020 06:35:29 +0000</pubDate>
      <link>https://dev.to/scraper_ai/the-11-best-free-web-scraping-tools-that-can-use-proxies-compared-43bm</link>
      <guid>https://dev.to/scraper_ai/the-11-best-free-web-scraping-tools-that-can-use-proxies-compared-43bm</guid>
      <description>&lt;h1&gt;
  
  
  Scraper.AI
&lt;/h1&gt;

&lt;p&gt;Website: &lt;a href="https://scraper.ai"&gt;https://scraper.ai&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;Scraper.AI is a new player on the market offering a wide variety of features like scraping websites with multiple pages, scrollable pages, authenticated pages and many more. Next to this you're also future proofed as they offer a great API for extracting pages through the API yourself.&lt;/p&gt;

&lt;p&gt;Not that technical? No problem, with their unique visual extractor you can extract any data you want without tprogramming knowledge!&lt;/p&gt;

&lt;h3&gt;
  
  
  Advantages
&lt;/h3&gt;

&lt;ul&gt;
&lt;li&gt;Many features&lt;/li&gt;
&lt;li&gt;Intuïtive UI&lt;/li&gt;
&lt;li&gt;Easy to learn, no extensive tutorials needed to get started&lt;/li&gt;
&lt;li&gt;Uses Many proxies to give a consistent result&lt;/li&gt;
&lt;li&gt;Fast&lt;/li&gt;
&lt;li&gt;Free plan available, cheap compared to others&lt;/li&gt;
&lt;li&gt;It's a SaaS, no need to keep your browser open for a long time&lt;/li&gt;
&lt;/ul&gt;

&lt;h3&gt;
  
  
  Disadvantages
&lt;/h3&gt;

&lt;ul&gt;
&lt;li&gt;It's an overall solution, not niche targeting&lt;/li&gt;
&lt;li&gt;Rather new player on the market&lt;/li&gt;
&lt;/ul&gt;

&lt;h1&gt;
  
  
  Octoparse
&lt;/h1&gt;

&lt;p&gt;Website: &lt;a href="https://www.octoparse.com"&gt;https://www.octoparse.com&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;A Free, Simple, and Powerful Web Scraping Tool. Automate Data Extraction from websites within clicks without coding.&lt;/p&gt;

&lt;h3&gt;
  
  
  Advantages
&lt;/h3&gt;

&lt;ul&gt;
&lt;li&gt;Focuses more on niches scraping&lt;/li&gt;
&lt;li&gt;Fair pricing&lt;/li&gt;
&lt;li&gt;Consistent results&lt;/li&gt;
&lt;li&gt;It's a SaaS, no need to keep your browser open for a long time&lt;/li&gt;
&lt;/ul&gt;

&lt;h3&gt;
  
  
  Disadvantages
&lt;/h3&gt;

&lt;ul&gt;
&lt;li&gt;Steep learning curve&lt;/li&gt;
&lt;li&gt;Doesn't offer API scraping&lt;/li&gt;
&lt;/ul&gt;

&lt;h1&gt;
  
  
  Scrapy
&lt;/h1&gt;

&lt;p&gt;Website: &lt;a href="https://github.com/scrapy/scrapy"&gt;https://github.com/scrapy/scrapy&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;Scrapy is a fast high-level web crawling and web scraping framework, used to crawl websites and extract structured data from their pages. It can be used for a wide range of purposes, from data mining to monitoring and automated testing.&lt;/p&gt;

&lt;h3&gt;
  
  
  Advantages
&lt;/h3&gt;

&lt;ul&gt;
&lt;li&gt;Most popular python library&lt;/li&gt;
&lt;li&gt;Open-source&lt;/li&gt;
&lt;/ul&gt;

&lt;h3&gt;
  
  
  Disadvantages
&lt;/h3&gt;

&lt;ul&gt;
&lt;li&gt;You still need to run your own servers&lt;/li&gt;
&lt;li&gt;Only for scraping&lt;/li&gt;
&lt;li&gt;Still need programmers to implement it&lt;/li&gt;
&lt;/ul&gt;

&lt;h1&gt;
  
  
  Puppeteer
&lt;/h1&gt;

&lt;p&gt;Website: &lt;a href="https://github.com/puppeteer/puppeteer"&gt;https://github.com/puppeteer/puppeteer&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;Puppeteer is a Node library which provides a high-level API to control Chrome or Chromium over the DevTools Protocol. Puppeteer runs headless by default, but can be configured to run full (non-headless) Chrome or Chromium.&lt;/p&gt;

&lt;h3&gt;
  
  
  Advantages
&lt;/h3&gt;

&lt;ul&gt;
&lt;li&gt;Most popular node.js library for scraping&lt;/li&gt;
&lt;li&gt;Battle tested&lt;/li&gt;
&lt;li&gt;Open-Source&lt;/li&gt;
&lt;li&gt;Reliable&lt;/li&gt;
&lt;li&gt;Direct implementation for proxies&lt;/li&gt;
&lt;/ul&gt;

&lt;h3&gt;
  
  
  Disadvantages
&lt;/h3&gt;

&lt;ul&gt;
&lt;li&gt;Requires good knowledge of timeouts, scrape processing, ...&lt;/li&gt;
&lt;li&gt;You still need to run your own servers&lt;/li&gt;
&lt;li&gt;Only for scraping&lt;/li&gt;
&lt;li&gt;Still need programmers to implement it&lt;/li&gt;
&lt;/ul&gt;

&lt;h1&gt;
  
  
  Playwright
&lt;/h1&gt;

&lt;p&gt;Website: &lt;a href="https://github.com/microsoft/playwright"&gt;https://github.com/microsoft/playwright&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;Playwright is a Node.js library to automate Chromium, Firefox and WebKit with a single API. Playwright is built to enable cross-browser web automation that is ever-green, capable, reliable and fast.&lt;/p&gt;

&lt;h3&gt;
  
  
  Advantages
&lt;/h3&gt;

&lt;ul&gt;
&lt;li&gt;Competitor to puppeteer&lt;/li&gt;
&lt;li&gt;Open-Source&lt;/li&gt;
&lt;li&gt;Reliable&lt;/li&gt;
&lt;/ul&gt;

&lt;h3&gt;
  
  
  Disadvantages
&lt;/h3&gt;

&lt;ul&gt;
&lt;li&gt;Harder to use than Puppeteer&lt;/li&gt;
&lt;li&gt;Requires a lot of tweaking per browser&lt;/li&gt;
&lt;li&gt;Newer than puppeteer&lt;/li&gt;
&lt;li&gt;You still need to run your own servers&lt;/li&gt;
&lt;li&gt;Only for scraping&lt;/li&gt;
&lt;li&gt;Still need programmers to implement it&lt;/li&gt;
&lt;/ul&gt;

&lt;h1&gt;
  
  
  Cheerio
&lt;/h1&gt;

&lt;p&gt;Website: &lt;a href="https://github.com/cheeriojs/cheerio"&gt;https://github.com/cheeriojs/cheerio&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;Cheerio parses markup and provides an API for traversing/manipulating the resulting data structure. It does not interpret the result as a web browser does. Specifically, it does not produce a visual rendering, apply CSS, load external resources, or execute JavaScript. If your use case requires any of this functionality, you should consider projects like PhantomJS or JSDom.&lt;/p&gt;

&lt;h3&gt;
  
  
  Advantages
&lt;/h3&gt;

&lt;ul&gt;
&lt;li&gt;HTML parser&lt;/li&gt;
&lt;li&gt;Famous open-source Node.JS library&lt;/li&gt;
&lt;li&gt;Good functions for extracting data from a HTML&lt;/li&gt;
&lt;/ul&gt;

&lt;h3&gt;
  
  
  Disadvantages
&lt;/h3&gt;

&lt;ul&gt;
&lt;li&gt;not really a scraper, you need to render a page using puppeteer and then extract the data&lt;/li&gt;
&lt;/ul&gt;

&lt;h1&gt;
  
  
  BeautifulSoup
&lt;/h1&gt;

&lt;p&gt;Website: &lt;a href="https://www.crummy.com/software/BeautifulSoup/bs4/doc/"&gt;https://www.crummy.com/software/BeautifulSoup/bs4/doc/&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;Beautiful Soup is a Python library for pulling data out of HTML and XML files. It works with your favorite parser to provide idiomatic ways of navigating, searching, and modifying the parse tree. It commonly saves programmers hours or days of work.&lt;/p&gt;

&lt;h3&gt;
  
  
  Advantages
&lt;/h3&gt;

&lt;ul&gt;
&lt;li&gt;HTML parser&lt;/li&gt;
&lt;li&gt;Famous open-source Python library&lt;/li&gt;
&lt;li&gt;Good functions for extracting data from a HTML
### Disadvantages&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;not really a scraper, you need to render a page using puppeteer and then extract the data&lt;/p&gt;

&lt;h1&gt;
  
  
  Scraper API
&lt;/h1&gt;

&lt;p&gt;Website: &lt;a href="https://www.scraperapi.com/"&gt;https://www.scraperapi.com/&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;Scraper API handles proxies, browsers, and CAPTCHAs, so you can get the HTML from any web page with a simple API call!&lt;/p&gt;

&lt;h3&gt;
  
  
  Advantages
&lt;/h3&gt;

&lt;ul&gt;
&lt;li&gt;Reliable results&lt;/li&gt;
&lt;li&gt;Many proxies available&lt;/li&gt;
&lt;li&gt;Good at it's single feature, rendering a webpage using it's API&lt;/li&gt;
&lt;/ul&gt;

&lt;h3&gt;
  
  
  Disadvantages
&lt;/h3&gt;

&lt;p&gt;Programming knowledge required&lt;/p&gt;

&lt;h1&gt;
  
  
  Selenium
&lt;/h1&gt;

&lt;p&gt;Website: &lt;a href="https://www.selenium.dev/"&gt;https://www.selenium.dev/&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;Selenium automates browsers. That's it! What you do with that power is entirely up to you. Primarily it is for automating web applications for testing purposes, but is certainly not limited to just that. Boring web-based administration tasks can (and should) also be automated as well.&lt;/p&gt;

&lt;h3&gt;
  
  
  Advantages
&lt;/h3&gt;

&lt;ul&gt;
&lt;li&gt;Works well&lt;/li&gt;
&lt;li&gt;Battle-proven&lt;/li&gt;
&lt;li&gt;Open-source&lt;/li&gt;
&lt;li&gt;Available for many programming languages&lt;/li&gt;
&lt;/ul&gt;

&lt;h3&gt;
  
  
  Disadvantages
&lt;/h3&gt;

&lt;ul&gt;
&lt;li&gt;Older technology&lt;/li&gt;
&lt;li&gt;Can be a pain to set up&lt;/li&gt;
&lt;/ul&gt;

&lt;h1&gt;
  
  
  Mozenda
&lt;/h1&gt;

&lt;p&gt;Website: &lt;a href="https://www.mozenda.com/"&gt;https://www.mozenda.com/&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;A bigger web data extraction software that's often used by enterprise customers&lt;/p&gt;

&lt;h3&gt;
  
  
  Advantages
&lt;/h3&gt;

&lt;ul&gt;
&lt;li&gt;Works well&lt;/li&gt;
&lt;li&gt;Battle-proven&lt;/li&gt;
&lt;li&gt;Focuses on enterprises&lt;/li&gt;
&lt;/ul&gt;

&lt;h3&gt;
  
  
  Disadvantages
&lt;/h3&gt;

&lt;ul&gt;
&lt;li&gt;Expensive&lt;/li&gt;
&lt;/ul&gt;

&lt;h1&gt;
  
  
  Kimura
&lt;/h1&gt;

&lt;p&gt;Website: &lt;a href="https://github.com/vifreefly/kimuraframework"&gt;https://github.com/vifreefly/kimuraframework&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;Kimurai is a modern web scraping framework written in Ruby which works out of box with Headless Chromium/Firefox, PhantomJS, or simple HTTP requests and allows to scrape and interact with JavaScript rendered websites.&lt;/p&gt;

&lt;h3&gt;
  
  
  Advantages
&lt;/h3&gt;

&lt;ul&gt;
&lt;li&gt;Ruby (if you use ruby often)&lt;/li&gt;
&lt;li&gt;Open-source&lt;/li&gt;
&lt;li&gt;Good documented setup&lt;/li&gt;
&lt;/ul&gt;

&lt;h3&gt;
  
  
  Disadvantages
&lt;/h3&gt;

&lt;ul&gt;
&lt;li&gt;Not frequently updated anymore&lt;/li&gt;
&lt;/ul&gt;

</description>
      <category>webdev</category>
      <category>productivity</category>
      <category>showdev</category>
      <category>startup</category>
    </item>
    <item>
      <title>Tweet Cryptocurrency and Bitcoin prices every hour</title>
      <dc:creator>Scraper.AI</dc:creator>
      <pubDate>Thu, 27 Aug 2020 14:17:05 +0000</pubDate>
      <link>https://dev.to/scraper_ai/tweet-cryptocurrency-and-bitcoin-prices-every-hour-1fj9</link>
      <guid>https://dev.to/scraper_ai/tweet-cryptocurrency-and-bitcoin-prices-every-hour-1fj9</guid>
      <description>&lt;p&gt;Get started within 30min&lt;/p&gt;

&lt;p&gt;&lt;a href="https://res.cloudinary.com/practicaldev/image/fetch/s--brrseAJ9--/c_limit%2Cf_auto%2Cfl_progressive%2Cq_auto%2Cw_880/https://blog.scraper.ai/content/images/size/w1000/2020/08/tracking-crypto.png" class="article-body-image-wrapper"&gt;&lt;img src="https://res.cloudinary.com/practicaldev/image/fetch/s--brrseAJ9--/c_limit%2Cf_auto%2Cfl_progressive%2Cq_auto%2Cw_880/https://blog.scraper.ai/content/images/size/w1000/2020/08/tracking-crypto.png" alt=""&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;It’s always interesting to get an update on the different Cryptocurrency pricings. Sadly enough there are not that many services out there that do that out of the box for us, so let’s create one!&lt;/p&gt;

&lt;h1&gt;
  
  
  Prerequisites
&lt;/h1&gt;

&lt;ul&gt;
&lt;li&gt;Scraper.AI account&lt;/li&gt;
&lt;li&gt;AWS Account&lt;/li&gt;
&lt;li&gt;IFTTT account&lt;/li&gt;
&lt;/ul&gt;

&lt;h1&gt;
  
  
  Flow
&lt;/h1&gt;

&lt;p&gt;Once the prerequisites are met, we can now set-up a flow that will scrape the cryptocurrency pricings from &lt;a href="https://coinmarketcap.com/"&gt;https://coinmarketcap.com/&lt;/a&gt; every hour and send a Tweet once these got received. To do this, we need to create a flow that looks like this:&lt;/p&gt;

&lt;p&gt;Scraper.AI -&amp;gt; Data Processor -&amp;gt; Twitter&lt;/p&gt;

&lt;p&gt;The easiest way to do this is to have a Serverless function that gets executed on the Scraper.AI Webhook, which will manipulate the data and send this to Twitter. To send to Twitter, we utilize the IFTTT service, which has a connector to send a Tweet. In components we will thus get something like this:&lt;/p&gt;

&lt;p&gt;Scraper.AI -&amp;gt; AWS Lambda -&amp;gt; IFTTT -&amp;gt; Twitter&lt;/p&gt;

&lt;h2&gt;
  
  
  Setup
&lt;/h2&gt;

&lt;h3&gt;
  
  
  Setting up AWS Lambda
&lt;/h3&gt;

&lt;p&gt;The first thing we want to do is to set-up our AWS Lambda function. Here we want to configure a function that can be called through an HTTP endpoint, so we have to add a Trigger with the API Gateway to it.&lt;/p&gt;

&lt;p&gt;&lt;a href="https://res.cloudinary.com/practicaldev/image/fetch/s--ilPdBH3S--/c_limit%2Cf_auto%2Cfl_progressive%2Cq_auto%2Cw_880/https://blog.scraper.ai/content/images/size/w1000/2020/08/image-22.png" class="article-body-image-wrapper"&gt;&lt;img src="https://res.cloudinary.com/practicaldev/image/fetch/s--ilPdBH3S--/c_limit%2Cf_auto%2Cfl_progressive%2Cq_auto%2Cw_880/https://blog.scraper.ai/content/images/size/w1000/2020/08/image-22.png" alt="Creating the function"&gt;&lt;/a&gt;&lt;br&gt;
&lt;a href="https://res.cloudinary.com/practicaldev/image/fetch/s--FK_ReMZL--/c_limit%2Cf_auto%2Cfl_progressive%2Cq_auto%2Cw_880/https://blog.scraper.ai/content/images/size/w1000/2020/08/image-25.png" class="article-body-image-wrapper"&gt;&lt;img src="https://res.cloudinary.com/practicaldev/image/fetch/s--FK_ReMZL--/c_limit%2Cf_auto%2Cfl_progressive%2Cq_auto%2Cw_880/https://blog.scraper.ai/content/images/size/w1000/2020/08/image-25.png" alt="Adding our API Gateway Trigger - Step 1"&gt;&lt;/a&gt;&lt;br&gt;
&lt;a href="https://res.cloudinary.com/practicaldev/image/fetch/s--1bnmtKHP--/c_limit%2Cf_auto%2Cfl_progressive%2Cq_auto%2Cw_880/https://blog.scraper.ai/content/images/size/w1000/2020/08/image-26.png" class="article-body-image-wrapper"&gt;&lt;img src="https://res.cloudinary.com/practicaldev/image/fetch/s--1bnmtKHP--/c_limit%2Cf_auto%2Cfl_progressive%2Cq_auto%2Cw_880/https://blog.scraper.ai/content/images/size/w1000/2020/08/image-26.png" alt="Adding our API Gateway Trigger - Step 2"&gt;&lt;/a&gt;&lt;br&gt;
&lt;a href="https://res.cloudinary.com/practicaldev/image/fetch/s--47W9rWW2--/c_limit%2Cf_auto%2Cfl_progressive%2Cq_auto%2Cw_880/https://blog.scraper.ai/content/images/size/w1000/2020/08/image-27.png" class="article-body-image-wrapper"&gt;&lt;img src="https://res.cloudinary.com/practicaldev/image/fetch/s--47W9rWW2--/c_limit%2Cf_auto%2Cfl_progressive%2Cq_auto%2Cw_880/https://blog.scraper.ai/content/images/size/w1000/2020/08/image-27.png" alt="Adding our API Gateway Trigger - Step 3"&gt;&lt;/a&gt;&lt;br&gt;
&lt;a href="https://res.cloudinary.com/practicaldev/image/fetch/s--fMGQVd9t--/c_limit%2Cf_auto%2Cfl_progressive%2Cq_auto%2Cw_880/https://blog.scraper.ai/content/images/size/w1000/2020/08/image-28.png" class="article-body-image-wrapper"&gt;&lt;img src="https://res.cloudinary.com/practicaldev/image/fetch/s--fMGQVd9t--/c_limit%2Cf_auto%2Cfl_progressive%2Cq_auto%2Cw_880/https://blog.scraper.ai/content/images/size/w1000/2020/08/image-28.png" alt="Adding our API Gateway Trigger - Step 4"&gt;&lt;/a&gt;&lt;br&gt;
&lt;a href="https://res.cloudinary.com/practicaldev/image/fetch/s--1sZgcUS_--/c_limit%2Cf_auto%2Cfl_progressive%2Cq_auto%2Cw_880/https://blog.scraper.ai/content/images/size/w1000/2020/08/image-30.png" class="article-body-image-wrapper"&gt;&lt;img src="https://res.cloudinary.com/practicaldev/image/fetch/s--1sZgcUS_--/c_limit%2Cf_auto%2Cfl_progressive%2Cq_auto%2Cw_880/https://blog.scraper.ai/content/images/size/w1000/2020/08/image-30.png" alt="Adding our API Gateway Trigger - Step 5"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;Once we then have our boilerplate and trigger configured, we can add a &lt;code&gt;package.json&lt;/code&gt; and &lt;code&gt;index.js&lt;/code&gt;&lt;/p&gt;

&lt;blockquote&gt;
&lt;p&gt;Note: you will have to create this locally, install the NPM modules with npm install, archive it into a .zip file and upload this to the Lambda portal.&lt;/p&gt;
&lt;/blockquote&gt;

&lt;p&gt;&lt;a href="https://res.cloudinary.com/practicaldev/image/fetch/s--UofBZHVq--/c_limit%2Cf_auto%2Cfl_progressive%2Cq_auto%2Cw_880/https://blog.scraper.ai/content/images/size/w1000/2020/08/image-24.png" class="article-body-image-wrapper"&gt;&lt;img src="https://res.cloudinary.com/practicaldev/image/fetch/s--UofBZHVq--/c_limit%2Cf_auto%2Cfl_progressive%2Cq_auto%2Cw_880/https://blog.scraper.ai/content/images/size/w1000/2020/08/image-24.png" alt="Creating our Function - Package.json"&gt;&lt;/a&gt;&lt;br&gt;
&lt;a href="https://res.cloudinary.com/practicaldev/image/fetch/s--0E2gImLI--/c_limit%2Cf_auto%2Cfl_progressive%2Cq_auto%2Cw_880/https://blog.scraper.ai/content/images/size/w1000/2020/08/image-31.png" class="article-body-image-wrapper"&gt;&lt;img src="https://res.cloudinary.com/practicaldev/image/fetch/s--0E2gImLI--/c_limit%2Cf_auto%2Cfl_progressive%2Cq_auto%2Cw_880/https://blog.scraper.ai/content/images/size/w1000/2020/08/image-31.png" alt="Creating our Function - index.js"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;This was the hardest part! The rest is just click-click-click 😉 so let’s continue!&lt;/p&gt;

&lt;h3&gt;
  
  
  Setting up IFTTT
&lt;/h3&gt;

&lt;p&gt;IFTTT (If This, Then That) is an easy service to utilize. We can create an applet here that takes the URL called above in our AWS function (which we will get at the end here) and send it to Twitter.&lt;/p&gt;

&lt;p&gt;Let’s configure IFTTT&lt;br&gt;
&lt;a href="https://res.cloudinary.com/practicaldev/image/fetch/s--Bz-YSC8X--/c_limit%2Cf_auto%2Cfl_progressive%2Cq_auto%2Cw_880/https://blog.scraper.ai/content/images/size/w1000/2020/08/image-9.png" class="article-body-image-wrapper"&gt;&lt;img src="https://res.cloudinary.com/practicaldev/image/fetch/s--Bz-YSC8X--/c_limit%2Cf_auto%2Cfl_progressive%2Cq_auto%2Cw_880/https://blog.scraper.ai/content/images/size/w1000/2020/08/image-9.png" alt=""&gt;&lt;/a&gt;&lt;br&gt;
&lt;a href="https://res.cloudinary.com/practicaldev/image/fetch/s--3KH1qDws--/c_limit%2Cf_auto%2Cfl_progressive%2Cq_auto%2Cw_880/https://blog.scraper.ai/content/images/size/w1000/2020/08/image-10.png" class="article-body-image-wrapper"&gt;&lt;img src="https://res.cloudinary.com/practicaldev/image/fetch/s--3KH1qDws--/c_limit%2Cf_auto%2Cfl_progressive%2Cq_auto%2Cw_880/https://blog.scraper.ai/content/images/size/w1000/2020/08/image-10.png" alt=""&gt;&lt;/a&gt;&lt;br&gt;
&lt;a href="https://res.cloudinary.com/practicaldev/image/fetch/s--Qv31LSi1--/c_limit%2Cf_auto%2Cfl_progressive%2Cq_auto%2Cw_880/https://blog.scraper.ai/content/images/size/w1000/2020/08/image-11.png" class="article-body-image-wrapper"&gt;&lt;img src="https://res.cloudinary.com/practicaldev/image/fetch/s--Qv31LSi1--/c_limit%2Cf_auto%2Cfl_progressive%2Cq_auto%2Cw_880/https://blog.scraper.ai/content/images/size/w1000/2020/08/image-11.png" alt=""&gt;&lt;/a&gt;&lt;br&gt;
&lt;a href="https://res.cloudinary.com/practicaldev/image/fetch/s--EZ8Vu304--/c_limit%2Cf_auto%2Cfl_progressive%2Cq_auto%2Cw_880/https://blog.scraper.ai/content/images/size/w1000/2020/08/image-12.png" class="article-body-image-wrapper"&gt;&lt;img src="https://res.cloudinary.com/practicaldev/image/fetch/s--EZ8Vu304--/c_limit%2Cf_auto%2Cfl_progressive%2Cq_auto%2Cw_880/https://blog.scraper.ai/content/images/size/w1000/2020/08/image-12.png" alt=""&gt;&lt;/a&gt;&lt;br&gt;
&lt;a href="https://res.cloudinary.com/practicaldev/image/fetch/s--eC7um57C--/c_limit%2Cf_auto%2Cfl_progressive%2Cq_auto%2Cw_880/https://blog.scraper.ai/content/images/size/w1000/2020/08/image-13.png" class="article-body-image-wrapper"&gt;&lt;img src="https://res.cloudinary.com/practicaldev/image/fetch/s--eC7um57C--/c_limit%2Cf_auto%2Cfl_progressive%2Cq_auto%2Cw_880/https://blog.scraper.ai/content/images/size/w1000/2020/08/image-13.png" alt=""&gt;&lt;/a&gt;&lt;br&gt;
&lt;a href="https://res.cloudinary.com/practicaldev/image/fetch/s--xUbM5kRQ--/c_limit%2Cf_auto%2Cfl_progressive%2Cq_auto%2Cw_880/https://blog.scraper.ai/content/images/size/w1000/2020/08/image-14.png" class="article-body-image-wrapper"&gt;&lt;img src="https://res.cloudinary.com/practicaldev/image/fetch/s--xUbM5kRQ--/c_limit%2Cf_auto%2Cfl_progressive%2Cq_auto%2Cw_880/https://blog.scraper.ai/content/images/size/w1000/2020/08/image-14.png" alt=""&gt;&lt;/a&gt;&lt;br&gt;
&lt;a href="https://res.cloudinary.com/practicaldev/image/fetch/s--h3f278Zw--/c_limit%2Cf_auto%2Cfl_progressive%2Cq_auto%2Cw_880/https://blog.scraper.ai/content/images/size/w1000/2020/08/image-15.png" class="article-body-image-wrapper"&gt;&lt;img src="https://res.cloudinary.com/practicaldev/image/fetch/s--h3f278Zw--/c_limit%2Cf_auto%2Cfl_progressive%2Cq_auto%2Cw_880/https://blog.scraper.ai/content/images/size/w1000/2020/08/image-15.png" alt=""&gt;&lt;/a&gt;&lt;br&gt;
&lt;a href="https://res.cloudinary.com/practicaldev/image/fetch/s--wmwDLG11--/c_limit%2Cf_auto%2Cfl_progressive%2Cq_auto%2Cw_880/https://blog.scraper.ai/content/images/size/w1000/2020/08/image-16.png" class="article-body-image-wrapper"&gt;&lt;img src="https://res.cloudinary.com/practicaldev/image/fetch/s--wmwDLG11--/c_limit%2Cf_auto%2Cfl_progressive%2Cq_auto%2Cw_880/https://blog.scraper.ai/content/images/size/w1000/2020/08/image-16.png" alt=""&gt;&lt;/a&gt;&lt;br&gt;
&lt;a href="https://res.cloudinary.com/practicaldev/image/fetch/s--1SS7J-ab--/c_limit%2Cf_auto%2Cfl_progressive%2Cq_auto%2Cw_880/https://blog.scraper.ai/content/images/size/w1000/2020/08/image-33.png" class="article-body-image-wrapper"&gt;&lt;img src="https://res.cloudinary.com/practicaldev/image/fetch/s--1SS7J-ab--/c_limit%2Cf_auto%2Cfl_progressive%2Cq_auto%2Cw_880/https://blog.scraper.ai/content/images/size/w1000/2020/08/image-33.png" alt=""&gt;&lt;/a&gt;&lt;br&gt;
&lt;a href="https://res.cloudinary.com/practicaldev/image/fetch/s--ABe6fX6Y--/c_limit%2Cf_auto%2Cfl_progressive%2Cq_auto%2Cw_880/https://blog.scraper.ai/content/images/size/w1000/2020/08/image-18.png" class="article-body-image-wrapper"&gt;&lt;img src="https://res.cloudinary.com/practicaldev/image/fetch/s--ABe6fX6Y--/c_limit%2Cf_auto%2Cfl_progressive%2Cq_auto%2Cw_880/https://blog.scraper.ai/content/images/size/w1000/2020/08/image-18.png" alt=""&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;Now Twitter has been connected, the last thing we need to do is to get the Link for our Webhook.&lt;/p&gt;

&lt;blockquote&gt;
&lt;p&gt;Note: this URL you will have to enter in the AWS Lambda Function!!!&lt;/p&gt;
&lt;/blockquote&gt;

&lt;p&gt;&lt;a href="https://res.cloudinary.com/practicaldev/image/fetch/s--WgIjrQW2--/c_limit%2Cf_auto%2Cfl_progressive%2Cq_auto%2Cw_880/https://blog.scraper.ai/content/images/size/w1000/2020/08/image-19.png" class="article-body-image-wrapper"&gt;&lt;img src="https://res.cloudinary.com/practicaldev/image/fetch/s--WgIjrQW2--/c_limit%2Cf_auto%2Cfl_progressive%2Cq_auto%2Cw_880/https://blog.scraper.ai/content/images/size/w1000/2020/08/image-19.png" alt=""&gt;&lt;/a&gt;&lt;br&gt;
&lt;a href="https://res.cloudinary.com/practicaldev/image/fetch/s--tfHnEwX_--/c_limit%2Cf_auto%2Cfl_progressive%2Cq_auto%2Cw_880/https://blog.scraper.ai/content/images/size/w1000/2020/08/image-20.png" class="article-body-image-wrapper"&gt;&lt;img src="https://res.cloudinary.com/practicaldev/image/fetch/s--tfHnEwX_--/c_limit%2Cf_auto%2Cfl_progressive%2Cq_auto%2Cw_880/https://blog.scraper.ai/content/images/size/w1000/2020/08/image-20.png" alt=""&gt;&lt;/a&gt;&lt;br&gt;
&lt;a href="https://res.cloudinary.com/practicaldev/image/fetch/s--H36FJHdI--/c_limit%2Cf_auto%2Cfl_progressive%2Cq_auto%2Cw_880/https://blog.scraper.ai/content/images/size/w1000/2020/08/image-21.png" class="article-body-image-wrapper"&gt;&lt;img src="https://res.cloudinary.com/practicaldev/image/fetch/s--H36FJHdI--/c_limit%2Cf_auto%2Cfl_progressive%2Cq_auto%2Cw_880/https://blog.scraper.ai/content/images/size/w1000/2020/08/image-21.png" alt=""&gt;&lt;/a&gt;&lt;/p&gt;

&lt;h3&gt;
  
  
  Setting Up Scraper.AI
&lt;/h3&gt;

&lt;p&gt;Last but not least, we need to monitor &lt;a href="https://coinmarketcap.com/"&gt;https://coinmarketcap.com/&lt;/a&gt; to changes. Luckily we can utilize &lt;a href="//Scraper.AI"&gt;https://Scraper.AI&lt;/a&gt; here!&lt;br&gt;
Navigate to &lt;a href="https://coinmarketcap.com/"&gt;https://coinmarketcap.com/&lt;/a&gt; and select the properties name, price and volume&lt;/p&gt;

&lt;p&gt;&lt;a href="https://res.cloudinary.com/practicaldev/image/fetch/s--8Vd_zaVb--/c_limit%2Cf_auto%2Cfl_progressive%2Cq_auto%2Cw_880/https://blog.scraper.ai/content/images/size/w1000/2020/08/image-34.png" class="article-body-image-wrapper"&gt;&lt;img src="https://res.cloudinary.com/practicaldev/image/fetch/s--8Vd_zaVb--/c_limit%2Cf_auto%2Cfl_progressive%2Cq_auto%2Cw_880/https://blog.scraper.ai/content/images/size/w1000/2020/08/image-34.png" alt=""&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;Then continue until the Website is scraped and displayed on Scraper.AI&lt;/p&gt;

&lt;blockquote&gt;
&lt;p&gt;Note: Configure scheduling to happen hourly! This ensures that you post the newest tweets every time.&lt;/p&gt;
&lt;/blockquote&gt;

&lt;p&gt;&lt;a href="https://res.cloudinary.com/practicaldev/image/fetch/s--hu-Pc8FA--/c_limit%2Cf_auto%2Cfl_progressive%2Cq_auto%2Cw_880/https://blog.scraper.ai/content/images/size/w1000/2020/08/image-35.png" class="article-body-image-wrapper"&gt;&lt;img src="https://res.cloudinary.com/practicaldev/image/fetch/s--hu-Pc8FA--/c_limit%2Cf_auto%2Cfl_progressive%2Cq_auto%2Cw_880/https://blog.scraper.ai/content/images/size/w1000/2020/08/image-35.png" alt=""&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;Finally, go to Notifications and enter the AWS Lambda endpoint you copied earlier into the “Webhook” field.&lt;/p&gt;

&lt;p&gt;&lt;a href="https://res.cloudinary.com/practicaldev/image/fetch/s--sW3833VT--/c_limit%2Cf_auto%2Cfl_progressive%2Cq_auto%2Cw_880/https://blog.scraper.ai/content/images/size/w1000/2020/08/image-36.png" class="article-body-image-wrapper"&gt;&lt;img src="https://res.cloudinary.com/practicaldev/image/fetch/s--sW3833VT--/c_limit%2Cf_auto%2Cfl_progressive%2Cq_auto%2Cw_880/https://blog.scraper.ai/content/images/size/w1000/2020/08/image-36.png" alt=""&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;Congratulations! Tweets should now start appearing into your Twitter account.&lt;/p&gt;

&lt;blockquote&gt;
&lt;p&gt;Note: to test this easily, you can “Scrape Manually” on the dashboard!&lt;/p&gt;
&lt;/blockquote&gt;

&lt;p&gt;An example of this can be seen below:&lt;/p&gt;

&lt;p&gt;&lt;a href="https://res.cloudinary.com/practicaldev/image/fetch/s--jqy0bCJg--/c_limit%2Cf_auto%2Cfl_progressive%2Cq_auto%2Cw_880/https://blog.scraper.ai/content/images/2020/08/image-32.png" class="article-body-image-wrapper"&gt;&lt;img src="https://res.cloudinary.com/practicaldev/image/fetch/s--jqy0bCJg--/c_limit%2Cf_auto%2Cfl_progressive%2Cq_auto%2Cw_880/https://blog.scraper.ai/content/images/2020/08/image-32.png" alt=""&gt;&lt;/a&gt;&lt;/p&gt;

</description>
      <category>beginners</category>
      <category>aws</category>
      <category>serverless</category>
      <category>node</category>
    </item>
    <item>
      <title>11 usages for web scraping</title>
      <dc:creator>Scraper.AI</dc:creator>
      <pubDate>Wed, 29 Jul 2020 05:58:26 +0000</pubDate>
      <link>https://dev.to/scraper_ai/11-usages-for-web-scraping-1f0j</link>
      <guid>https://dev.to/scraper_ai/11-usages-for-web-scraping-1f0j</guid>
      <description>&lt;p&gt;and why you should start scraping the web now&lt;/p&gt;

&lt;p&gt;Many people ask themselves the question, how do I improve in the areas covered below. In a lot of the cases it's important to know that web scraping plays a major role in it and that it's actually rather easy to get started.&lt;/p&gt;

&lt;p&gt;With web scraping you can extract data from any website and as some may say "Data is the new gold". There is so much important data to be gathered from websites that you can make great business decisions on.&lt;/p&gt;

&lt;p&gt;Below are our 11 usages for web scraping and why you should start scraping now.&lt;/p&gt;

&lt;h1&gt;
  
  
  Build your new product
&lt;/h1&gt;

&lt;p&gt;No Code tools are on a rise. They allow you to create certain flows, analyze data, create stunning websites and more without any knowledge of coding. You don't have to be a technical founder anymore to start a business. But what many people don't know about web scraping is that you can leverage it to use that data as input to your new application. For example, you want to create &lt;a href="https://blog.scraper.ai/build-your-first-crypto-app-using-amazon-honeycode/"&gt;a mobile application that shows the latest cryptocurrency prices&lt;/a&gt;? No problem, it only takes an hour.&lt;/p&gt;

&lt;h1&gt;
  
  
  SEO
&lt;/h1&gt;

&lt;p&gt;Having organic growth is next to direct traffic, the biggest source of traffic coming to your website. Sadly it's also the hardest one to optimize for. It requires a lot of persistence, monitoring and analyzing to get ranked among the top. Even when you're at the top you still have to optimize for click throughs on your links. Tough business! Luckily this is where web scraping is very powerful. It can help to decrease your workload and automate some tasks for you. For example:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;Help you &lt;a href="https://blog.scraper.ai/build-your-first-crypto-app-using-amazon-honeycode/"&gt;track the ranking of your website for your specific keywords&lt;/a&gt;
&lt;/li&gt;
&lt;li&gt;Identify and monitor keywords of competitors by getting a daily list&lt;/li&gt;
&lt;li&gt;Analyze the top ranking keywords and alternatives for you
and a lot more.&lt;/li&gt;
&lt;/ul&gt;

&lt;h1&gt;
  
  
  Dataset creation
&lt;/h1&gt;

&lt;p&gt;One of the big tasks of a data scientist, computer vision specialist and others that need data is the creation of a trustworthy and well composed dataset. Web scraping can take any website list and compose it into a usable data stream that you can build on top of. If you have a great tool like Scraper.AI you can also &lt;a href="https://blog.scraper.ai/how-to-create-a-hourly-updating-covid-19-dataset-under-5-minutes/"&gt;monitor that data&lt;/a&gt; and keep your dataset up-to-date in the land of fast changing data.&lt;/p&gt;

&lt;h1&gt;
  
  
  Competitor tracking
&lt;/h1&gt;

&lt;p&gt;Knowing the competition is one of the most valuable techniques there is. It helps you to connect to your audience but also makes you stay on top of your sector. Having a web scraper and being to extract the prices of a webshop of your competitor ensures you that you can be the cheapest out there.&lt;/p&gt;

&lt;p&gt;For beginning startups they can for example monitor new comers in their space and see what approach they're taking. It might mean that they've found product market fit.&lt;/p&gt;

&lt;h1&gt;
  
  
  Discovery
&lt;/h1&gt;

&lt;p&gt;Where it all starts, the starts with is replaced with a fancy arrow (➜)&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;Starting a business ➜ Discovering an idea&lt;/li&gt;
&lt;li&gt;Getting started with SEO ➜ Discovering keywords&lt;/li&gt;
&lt;li&gt;Knowing your competition ➜ Discovering your competitors&lt;/li&gt;
&lt;li&gt;Investing ➜ Discovering what to invest in&lt;/li&gt;
&lt;li&gt;Buying a property ➜ Discovering a property&lt;/li&gt;
&lt;li&gt;Making money on stocks ➜ Discovering what stocks to buy&lt;/li&gt;
&lt;li&gt;...&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;This list can go on for a long while, an action almost always starts with the discovery of something. Discovering means that you want to get the data that justifies the decision you're gonna take to get to that action.&lt;/p&gt;

&lt;p&gt;You might want to &lt;a href="https://www.youtube.com/watch?v=a6YWbFoaruI"&gt;watch stock websites&lt;/a&gt; to get to learn about price changes of your favourite stock.&lt;/p&gt;

&lt;p&gt;Web scraping can get these latests stocks, properties, and many more.&lt;/p&gt;

&lt;h1&gt;
  
  
  Product monitoring
&lt;/h1&gt;

&lt;p&gt;You're a vendor, dropshipper, amazon seller or anyone that sells a product. A major part will be making sure that you're product fulfills certain demands. You want to watch for reviews, correct pricing, advertising, ... it's a lot of work.&lt;/p&gt;

&lt;p&gt;Scraping reviews ensures that you can sustain you're high rating, act on low ratings and make sure that they get resolved correctly. Watching prices from competitors, analyzing your advertising metrics in one dashboard. They can all be automated with web scraping. In essence you're making your personalized software product.&lt;/p&gt;

&lt;h1&gt;
  
  
  Marketing automation
&lt;/h1&gt;

&lt;p&gt;Finding potential influencers becomes a lot easier, you can go to instagram, facebook, quora, ... and get a list of comments or profiles with most views, likes, watches. The only thing you have to do is open up the scraper, select the names and you're ready to go. Most websites don't show emails anymore due to privacy reasons, but you can private message them or use some other tools to get the email for that user.&lt;/p&gt;

&lt;p&gt;&lt;a href="https://res.cloudinary.com/practicaldev/image/fetch/s--aL5EDOx---/c_limit%2Cf_auto%2Cfl_progressive%2Cq_auto%2Cw_880/https://blog.scraper.ai/content/images/size/w1000/2020/07/image-26.png" class="article-body-image-wrapper"&gt;&lt;img src="https://res.cloudinary.com/practicaldev/image/fetch/s--aL5EDOx---/c_limit%2Cf_auto%2Cfl_progressive%2Cq_auto%2Cw_880/https://blog.scraper.ai/content/images/size/w1000/2020/07/image-26.png" alt=""&gt;&lt;/a&gt;&lt;/p&gt;

&lt;h1&gt;
  
  
  Lead generation
&lt;/h1&gt;

&lt;p&gt;A great way to get interesting prospects for your business is generating leads. As soon as you have leads you can setup a sales pipeline and get these leads converted to customers.&lt;/p&gt;

&lt;h2&gt;
  
  
  Getting the leads
&lt;/h2&gt;

&lt;p&gt;But getting these leads is not always straightforward. You first need to find your audience. Luckily there are already great tools out there like LinkedIn search, yellow pages, google maps, angellist, producthunt, ... They all have one thing in common. They show you a list of leads. Great! Because this means we can extract the data from that list and convert it to something usable for us, an excel file, csv, ... anything we need to get these leads converted and expand our business.&lt;/p&gt;

&lt;h1&gt;
  
  
  Investment optimization
&lt;/h1&gt;

&lt;p&gt;As a Venture Capital firm you might be interested in startups that were backed by others VCs, you can just go to their page, spend an hour or so per day and get that list. Or you can scrape and monitor them. You can extract the data every day and get notified when a new startup gets added.&lt;/p&gt;

&lt;p&gt;By also combining this with dataset creation you can go in the history of a startup and minimize the risk you take in investing in a startup.&lt;/p&gt;

&lt;p&gt;&lt;a href="https://res.cloudinary.com/practicaldev/image/fetch/s--NcnzTK6U--/c_limit%2Cf_auto%2Cfl_progressive%2Cq_auto%2Cw_880/https://blog.scraper.ai/content/images/2020/07/image-25.png" class="article-body-image-wrapper"&gt;&lt;img src="https://res.cloudinary.com/practicaldev/image/fetch/s--NcnzTK6U--/c_limit%2Cf_auto%2Cfl_progressive%2Cq_auto%2Cw_880/https://blog.scraper.ai/content/images/2020/07/image-25.png" alt=""&gt;&lt;/a&gt;&lt;/p&gt;

&lt;h1&gt;
  
  
  Finance
&lt;/h1&gt;

&lt;p&gt;Stocks, crypto, personal finances and more are something everyone has touched at least once. But it involves a lot of manual labour.&lt;/p&gt;

&lt;p&gt;Cryptocurrency or stock investors might find themselves looking at listing sites hourly to get the latest prices and volumes. Monitoring these could come in handy and web scrapers can get you this information in no time, leaving you with more time to do the analysis. Some services also offer direct API endpoints making it as easy to integrate as possible&lt;/p&gt;

&lt;p&gt;Your personal finances are important, you can scrape them and get notified when a bill comes through, a salary gets deposited and more. Become aware of any event that happens at any time.&lt;/p&gt;

&lt;h1&gt;
  
  
  Real estate monitoring
&lt;/h1&gt;

&lt;p&gt;Buying a property can be painful, there are plenty of listing sites, aggregators, agents and more and they all have their own time schedule. And when a property gets listed it's a race to get there as the first potential buyer.&lt;/p&gt;

&lt;p&gt;Automate this process and let it scrape every hour to get a list of new additions or updates and make sure that you're among the first to get there. Among the first I say because there are others doing exactly the same!&lt;/p&gt;

&lt;h1&gt;
  
  
  Summary
&lt;/h1&gt;

&lt;p&gt;Web scraping is not new and has proven to be valuable. It can automate time-consuming tasks and leave you with more time in hand to focus on the exact problem you're trying to solve. Being from analyzing data to buying a property. Extracting data from the web has never been this easy and should be a must-do to get most out of your business.&lt;/p&gt;

&lt;h1&gt;
  
  
  Resources
&lt;/h1&gt;

&lt;ul&gt;
&lt;li&gt;&lt;a href="https://blog.scraper.ai/track-your-seo-rankings-with-these-simple-tricks/"&gt;Track your SEO rankings&lt;/a&gt;&lt;/li&gt;
&lt;li&gt;&lt;a href="https://neilpatel.com/blog/automate-seo-with-content-scrapers/"&gt;Automate SEO with content scrapers&lt;/a&gt;&lt;/li&gt;
&lt;li&gt;&lt;a href="https://blog.scraper.ai/how-to-create-a-hourly-updating-covid-19-dataset-under-5-minutes/"&gt;Create an updating dataset under 5 minutes&lt;/a&gt;&lt;/li&gt;
&lt;li&gt;&lt;a href="https://blog.scraper.ai/build-your-first-crypto-app-using-amazon-honeycode/"&gt;Make a crypto app using Amazon HoneyCode&lt;/a&gt;&lt;/li&gt;
&lt;/ul&gt;

</description>
      <category>webdev</category>
      <category>productivity</category>
      <category>showdev</category>
    </item>
    <item>
      <title>Build your first crypto app using Amazon HoneyCode</title>
      <dc:creator>Scraper.AI</dc:creator>
      <pubDate>Wed, 29 Jul 2020 05:53:23 +0000</pubDate>
      <link>https://dev.to/scraper_ai/build-your-first-crypto-app-using-amazon-honeycode-4d4o</link>
      <guid>https://dev.to/scraper_ai/build-your-first-crypto-app-using-amazon-honeycode-4d4o</guid>
      <description>&lt;p&gt;Amazon has recently announced their new project called Amazon Honeycode which is a no-code tool focused on creating web &amp;amp; mobile applications.&lt;/p&gt;

&lt;p&gt;The product goes from a data first perspective, you create a dataset and then visualize it. After importing data, you can manipulate it, create new aggregating columns, rename columns and much more.&lt;/p&gt;

&lt;p&gt;Now, let dataset creation just be our strength over at scraper.ai. In this gentle introduction to Amazon HoneyCode we show how to get the data from a crypto website. Afterwards we’ll add the data to the Amazon HoneyCode platform. In the end we’ll have a fully functional web app ready to be published showing the latest crypto prices.&lt;/p&gt;

&lt;h1&gt;
  
  
  Get started
&lt;/h1&gt;

&lt;p&gt;To get started, head over to a crypto website, in this example we’ll use &lt;a href="https://coinmarketcap.com"&gt;https://coinmarketcap.com&lt;/a&gt; to get the latest crypto prices.&lt;/p&gt;

&lt;p&gt;&lt;a href="https://res.cloudinary.com/practicaldev/image/fetch/s--tFxplWNt--/c_limit%2Cf_auto%2Cfl_progressive%2Cq_auto%2Cw_880/https://blog.scraper.ai/content/images/size/w1000/2020/07/image-18.png" class="article-body-image-wrapper"&gt;&lt;img src="https://res.cloudinary.com/practicaldev/image/fetch/s--tFxplWNt--/c_limit%2Cf_auto%2Cfl_progressive%2Cq_auto%2Cw_880/https://blog.scraper.ai/content/images/size/w1000/2020/07/image-18.png" alt=""&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;Let’s get the data going by opening up the &lt;a href="https://scraper.ai"&gt;https://scraper.ai&lt;/a&gt; extension. After clicking “Select Element” we select the data we’re interested in, in this case we’ll use:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;Coin name (Bitcoin, Ethereum, …)&lt;/li&gt;
&lt;li&gt;Market cap&lt;/li&gt;
&lt;li&gt;Coin Price&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;which are also the labels we’re going to give to the fields.&lt;/p&gt;

&lt;p&gt;&lt;a href="https://res.cloudinary.com/practicaldev/image/fetch/s--GSFV9Kgp--/c_limit%2Cf_auto%2Cfl_progressive%2Cq_auto%2Cw_880/https://blog.scraper.ai/content/images/size/w1000/2020/07/image-19.png" class="article-body-image-wrapper"&gt;&lt;img src="https://res.cloudinary.com/practicaldev/image/fetch/s--GSFV9Kgp--/c_limit%2Cf_auto%2Cfl_progressive%2Cq_auto%2Cw_880/https://blog.scraper.ai/content/images/size/w1000/2020/07/image-19.png" alt=""&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;After clicking next twice we’re shown the data we’ve just extracted. As you can see, there is also an API url which we’ll use in a following story.&lt;/p&gt;

&lt;p&gt;For now we’ll keep it simple and “Download as CSV”. so let’s hit that button.&lt;/p&gt;

&lt;p&gt;&lt;a href="https://res.cloudinary.com/practicaldev/image/fetch/s--N1ONAH0y--/c_limit%2Cf_auto%2Cfl_progressive%2Cq_auto%2Cw_880/https://blog.scraper.ai/content/images/size/w1000/2020/07/image-20.png" class="article-body-image-wrapper"&gt;&lt;img src="https://res.cloudinary.com/practicaldev/image/fetch/s--N1ONAH0y--/c_limit%2Cf_auto%2Cfl_progressive%2Cq_auto%2Cw_880/https://blog.scraper.ai/content/images/size/w1000/2020/07/image-20.png" alt=""&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;Now we have all the data required we can head over to Amazon HoneyCode. Creating an account for their service only takes 1 minute. Afterwards we’re shown the screen below.&lt;/p&gt;

&lt;p&gt;Take the following steps:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;Create Workbook&lt;/li&gt;
&lt;li&gt;Import CSV&lt;/li&gt;
&lt;li&gt;Select the CSV we’ve exported in the step above&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;&lt;a href="https://res.cloudinary.com/practicaldev/image/fetch/s--c--xUJGq--/c_limit%2Cf_auto%2Cfl_progressive%2Cq_auto%2Cw_880/https://blog.scraper.ai/content/images/size/w1000/2020/07/image-21.png" class="article-body-image-wrapper"&gt;&lt;img src="https://res.cloudinary.com/practicaldev/image/fetch/s--c--xUJGq--/c_limit%2Cf_auto%2Cfl_progressive%2Cq_auto%2Cw_880/https://blog.scraper.ai/content/images/size/w1000/2020/07/image-21.png" alt=""&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;That’s it for the data, the data is now imported to their data view and we’re ready to use it for our Application.&lt;/p&gt;

&lt;p&gt;Let’s create the application by going to the “Apps” panel in the sidebar and clicking the Plus icon. We’ll choose “Use app wizard”.&lt;/p&gt;

&lt;p&gt;&lt;a href="https://res.cloudinary.com/practicaldev/image/fetch/s--yG-Nz6_Z--/c_limit%2Cf_auto%2Cfl_progressive%2Cq_auto%2Cw_880/https://blog.scraper.ai/content/images/size/w1000/2020/07/image-22.png" class="article-body-image-wrapper"&gt;&lt;img src="https://res.cloudinary.com/practicaldev/image/fetch/s--yG-Nz6_Z--/c_limit%2Cf_auto%2Cfl_progressive%2Cq_auto%2Cw_880/https://blog.scraper.ai/content/images/size/w1000/2020/07/image-22.png" alt=""&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;When we select our newly created table “Table1” as source, all fields will be pre-filled. I suggest using some sensible names and not “Table1”, but to keep this guide easy to follow we’ve gone with the insensible “Table1” name.&lt;/p&gt;

&lt;p&gt;&lt;a href="https://res.cloudinary.com/practicaldev/image/fetch/s--J3BcjH3f--/c_limit%2Cf_auto%2Cfl_progressive%2Cq_auto%2Cw_880/https://blog.scraper.ai/content/images/size/w1000/2020/07/image-23.png" class="article-body-image-wrapper"&gt;&lt;img src="https://res.cloudinary.com/practicaldev/image/fetch/s--J3BcjH3f--/c_limit%2Cf_auto%2Cfl_progressive%2Cq_auto%2Cw_880/https://blog.scraper.ai/content/images/size/w1000/2020/07/image-23.png" alt=""&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;After going through the “App wizard”, the app is opened up immediately.&lt;/p&gt;

&lt;p&gt;&lt;a href="https://res.cloudinary.com/practicaldev/image/fetch/s--KP67zVEP--/c_limit%2Cf_auto%2Cfl_progressive%2Cq_auto%2Cw_880/https://blog.scraper.ai/content/images/size/w1000/2020/07/image-24.png" class="article-body-image-wrapper"&gt;&lt;img src="https://res.cloudinary.com/practicaldev/image/fetch/s--KP67zVEP--/c_limit%2Cf_auto%2Cfl_progressive%2Cq_auto%2Cw_880/https://blog.scraper.ai/content/images/size/w1000/2020/07/image-24.png" alt=""&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;For viewing the app on our mobile devices we have to use the Amazon HoneyCode app viewer (&lt;a href="https://play.google.com/store/apps/details?id=com.amazon.aws.honeycode"&gt;https://play.google.com/store/apps/details?id=com.amazon.aws.honeycode&lt;/a&gt;).&lt;/p&gt;

&lt;h1&gt;
  
  
  Summary
&lt;/h1&gt;

&lt;p&gt;We’ve learned how easy it is to import data to Amazon HoneyCode and get an app up and running within minutes. For now the Amazon HoneyCode application is rather limited and only supports websites through their own app rather than having native apps. This narrows the scope more to enterprise applications or companies with a good distribution channel.&lt;/p&gt;

&lt;p&gt;Amazon HoneyCode has some integrations set up to import data more easily, but for external services it’s quite cumbersome and I hope they’re making the process easier. For now it’s limited to accepted partners such as “Google Analytics”, “Marketo” and more.&lt;/p&gt;

&lt;p&gt;In a following guide we’ll show how we can use other AWS services to import data into Amazon HoneyCode on a regular basis. Sadly it becomes rather technical due to the lack of 3rd party integrations available with Amazon HoneyCode.&lt;/p&gt;

</description>
      <category>nocode</category>
      <category>productivity</category>
      <category>showdev</category>
      <category>webdev</category>
    </item>
  </channel>
</rss>
