Table Of Contents
1. What Is Simple Scraper?
2. What Is Reddit (Growth Hacking)?
3. What You'll Learn In This (No-Code) Tutorial
4. Initial Scraper Setup
5. Running Your Scraper
6. Saving Your Crawler
7. Running Your Crawler
8. Next Steps
What Is Simple Scraper?
Simple Scraper is an excellent Google Chrome extension that makes web scraping easy. It helps you to extract data from any website with no-code. You can scrape locally or in the cloud. And, every website that you scrape instantly becomes an API. Simple Scraper is a simple yet powerful web scraping tool.
What Is Reddit (Growth Hacking)?
Reddit - Growth Hacking is the largest community for growth hackers on Reddit. The community shares experiments, stories and more on growth.
What You'll Learn In This (No-Code) Tutorial
By the end of this tutorial, you will have created a web scraper to scrape the infinitely scrolling pages on Reddit (Growth Hacking) for content insights.
Initial Scraper Setup
Go to the Growth Hacking Subreddit (reddit.com/r/growthhacking).
Open Simple Scraper and click the plus (+) sign.
First, you'll want to scrape the titles: select a title. Everything which gets highlighted is what'll get extracted. Name this data, 'Title'. Then, click the tick to set it for when you run the scraper.
Second, you'll want to scrape the votes. Again, click the plus (+) sign. Select the votes of a post. Everything which gets highlighted is what'll get extracted. Name this data, 'Votes'. And, click the tick to set it for when you run the scraper.
Now, click the infinite loop button and scroll down the page. Scrolling the page ensures that the web scraper understands it's an infinitely scrolling page.
Running Your Scraper
To run your scraper, click 'View Results'.
Once the web scraper has run, Simple Scraper will return the selected data. It will allow you to view that data in a table or as a JSON file. And, you will have the option of downloading the data as either a CSV file or JSON.
Saving Your Crawler
To save your web crawler, click 'Save Recipe'.
You'll have to confirm the settings for your web crawler when saving it. The settings that got entered for this project are:
- Recipe Name - 'Reddit Content Insights'
- URL - 'https://www.reddit.com/r/GrowthHacking/'
- Selected Properties - 'Title' and 'Votes'
- Page Navigation - Set it to four (or more depending on the number of times you want Simple Scraper to scrape the infinitely scrolling page).
Once you've entered the settings, click 'Create Recipe'.
Running Your Crawler
Click on the recipe you saved under 'My Recipes'.
To run your web crawler, click 'Run Recipe'.
Once the web crawler has run, Simple Scraper will return the selected data. You can view the output of your web crawler on the 'Results' page.
You'll notice that Simple Scraper has crawled the page four times and returned the selected data. You'll get given the option to view that data in a table or as a JSON file. And, you'll have the option of downloading the data too.
Congratulations on completing this tutorial. Now, why not challenge your capabilities? Try implementing one of the suggestions below. Or try your own.
- Scrape more pages in the subreddit.
- Scrape a different subreddit.
- Export your data to Google Sheets via Simple Scraper.