How To Construct An Amazon Product Scraper With Node.js

No Comments

Have you ever ever been ready the place you could intimately know the marketplace for a specific product? Perhaps you’re launching some software program and have to know easy methods to value it. Or maybe you have already got your personal product available on the market and wish to see which options so as to add for a aggressive benefit. Or possibly you simply wish to purchase one thing for your self and wish to be sure to get the perfect bang in your buck.

All these conditions have one factor in frequent: you want correct information to make the proper resolution. Really, there’s one other factor they share. All eventualities can profit from the usage of an online scraper.

Net scraping is the observe of extracting massive quantities of internet information via the usage of software program. So, in essence, it’s a approach to automate the tedious technique of hitting ‘copy’ after which ‘paste’ 200 occasions. After all, a bot can try this within the time it took you to learn this sentence, so it’s not solely much less boring however quite a bit sooner, too.

However the burning query is: why would somebody wish to scrape Amazon pages?

You’re about to search out out! However to begin with, I’d wish to make one thing clear proper now — whereas the act of scraping publicly obtainable information is authorized, Amazon has some measures to forestall it on their pages. As such, I urge you all the time to be conscious of the web site whereas scraping, take care to not harm it, and observe moral tips.

Really useful Studying: The Information To Moral Scraping Of Dynamic Web sites With Node.js And Puppeteer” by Andreas Altheimer

Why You Ought to Extract Amazon Product Knowledge

Being the most important on-line retailer on the planet, it’s secure to say that if you wish to purchase one thing, you may most likely get it on Amazon. So, it goes with out saying simply how massive of a knowledge treasure trove the web site is.

When scraping the net, your major query needs to be what to do with all that information. Whereas there are numerous particular person causes, it boils down to 2 outstanding use circumstances: optimizing your merchandise and discovering the perfect offers.

Let’s begin with the primary state of affairs. Except you’ve designed a really modern new product, the possibilities are that you would be able to already discover one thing at the very least related on Amazon. Scraping these product pages can web you invaluable information akin to:

The opponents’ pricing technique
So, that you would be able to modify your costs to be aggressive and perceive how others deal with promotional offers;
Buyer opinions
To see what your future consumer base cares about most and easy methods to enhance their expertise;
Commonest options
To see what your competitors presents to know which functionalities are essential and which might be left for later.

In essence, Amazon has every little thing you want for a deep market and product evaluation. You’ll be higher ready to design, launch, and develop your product lineup with that information.

The second state of affairs can apply to each companies and common individuals. The concept is fairly just like what I discussed earlier. You possibly can scrape the costs, options, and evaluations of all of the merchandise you may select, and so, you’ll be capable of choose the one that gives essentially the most advantages for the bottom value. In spite of everything, who doesn’t like a great deal?

Not all merchandise deserve this stage of consideration to element, however it will probably make an enormous distinction with costly purchases. Sadly, whereas the advantages are clear, many difficulties associate with scraping Amazon.

The Challenges Of Scraping Amazon Product Knowledge

Not all web sites are the identical. As a rule of thumb, the extra complicated and widespread an internet site is, the more durable it’s to scrape it. Keep in mind once I stated that Amazon was essentially the most outstanding e-commerce website? Nicely, that makes it each extraordinarily standard and fairly complicated.

First off, Amazon is aware of how scraping bots act, so the web site has countermeasures in place. Specifically, if the scraper follows a predictable sample, sending requests at fastened intervals, sooner than a human may or with virtually equivalent parameters, Amazon will discover and block the IP. Proxies can resolve this downside, however I didn’t want them since we received’t be scraping too many pages within the instance.

Subsequent, Amazon intentionally makes use of various web page constructions for his or her merchandise. That’s to say, that for those who examine the pages for various merchandise, there’s a great probability that you just’ll discover vital variations of their construction and attributes. The explanation behind that is fairly easy. You could adapt your scraper’s code for a particular system, and for those who use the identical script on a brand new form of web page, you’d should rewrite elements of it. So, they’re basically making you’re employed extra for the info.

Lastly, Amazon is an enormous web site. If you wish to collect massive quantities of knowledge, operating the scraping software program in your laptop may prove to take means an excessive amount of time in your wants. This downside is additional consolidated by the truth that going too quick will get your scraper blocked. So, if you need a great deal of information rapidly, you’ll want a really highly effective scraper.

Nicely, that’s sufficient speak about issues, let’s concentrate on options!

How To Construct A Net Scraper For Amazon

To maintain issues easy, we’ll take a step-by-step strategy to writing the code. Be happy to work in parallel with the information.

Search for the info we’d like

So, right here’s a state of affairs: I’m transferring in a number of months to a brand new place, and I’ll want a few new cabinets to carry books and magazines. I wish to know all my choices and get pretty much as good of a deal as I can. So, let’s go to the Amazon market, seek for “cabinets”, and see what we get.

The URL for this search and the web page we’ll be scraping is right here.

Okay, let’s take inventory of what we have now right here. Simply by glancing on the web page, we will get a great image about:

how the cabinets look;
what the package deal contains;
how prospects charge them;
their value;
the hyperlink to the product;
a suggestion for a less expensive different for a few of the gadgets.

That’s greater than we may ask for!

Get the required instruments

Let’s guarantee we have now all the next instruments put in and configured earlier than persevering with to the following step.

We are able to obtain it from right here.
Observe the directions on this web page to put in it in your particular gadget.
Earlier than beginning utilizing Axios or Cheerio, we have to set up Node.js and the Node Package deal Supervisor. The best approach to set up Node.js and NPM is to get one of many installers from the Node.Js official supply and run it.

Now, let’s create a brand new NPM challenge. Create a brand new folder for the challenge and run the next command:

npm init -y

To create the net scraper, we have to set up a few dependencies in our challenge:

An open-source library that helps us extract helpful data by parsing markup and offering an API for manipulating the ensuing information. Cheerio permits us to pick tags of an HTML doc by utilizing selectors: $(“div”). This particular selector helps us choose all <div> components on a web page. To put in Cheerio, please run the next command within the initiatives’ folder:

npm set up cheerio

A JavaScript library used to make HTTP requests from Node.js.

npm set up axios

Examine the web page supply

Within the following steps, we are going to study extra about how the knowledge is organized on the web page. The concept is to get a greater understanding of what we will scrape from our supply.

The developer instruments assist us interactively discover the web site’s Doc Object Mannequin (DOM). We are going to use the developer instruments in Chrome, however you need to use any internet browser you’re comfy with.

Let’s open it by right-clicking anyplace on the web page and choosing the “Examine” possibility:

This may open up a brand new window containing the supply code of the web page. As we have now stated earlier than, we want to scrape each shelf’s data.

As we will see from the screenshot above, the containers that maintain all the info have the next lessons:

sg-col-4-of-12 s-result-item s-asin sg-col-4-of-16 sg-col sg-col-4-of-20

Within the subsequent step, we are going to use Cheerio to pick all the weather containing the info we’d like.

Fetch the info

After we put in all of the dependencies introduced above, let’s create a brand new index.js file and sort the next traces of code:

const axios = require(“axios”);
const cheerio = require(“cheerio”);

const fetchShelves = async () => {
strive {
const response = await axios.get(‘′);

const html = response.information;

const $ = cheerio.load(html);

const cabinets = [];

$(‘’).every((_idx, el) => {
const shelf = $(el)
const title =‘span.a-size-base-plus.a-color-base.a-text-normal’).textual content()


return cabinets;
} catch (error) {
throw error;

fetchShelves().then((cabinets) => console.log(cabinets));

As we will see, we import the dependencies we’d like on the primary two traces, after which we create a fetchShelves() perform that, utilizing Cheerio, will get all the weather containing our merchandise’ data from the web page.

It iterates over every of them and pushes it to an empty array to get a better-formatted consequence.

The fetchShelves() perform will solely return the product’s title in the mean time, so let’s get the remainder of the knowledge we’d like. Please add the next traces of code after the road the place we outlined the variable title.

const picture =‘img.s-image’).attr(‘src’)

const hyperlink =‘a.a-link-normal.a-text-normal’).attr(‘href’)

const evaluations =‘div.a-section.a-spacing-none.a-spacing-top-micro > div.a-row.a-size-small’).youngsters(‘span’).final().attr(‘aria-label’)

const stars =‘div.a-section.a-spacing-none.a-spacing-top-micro > div > span’).attr(‘aria-label’)

const value =‘span.a-price > span.a-offscreen’).textual content()

let aspect = {

if (evaluations) {
aspect.evaluations = evaluations

if (stars) {
aspect.stars = stars

And change cabinets.push(title) with cabinets.push(aspect).

We are actually choosing all the knowledge we’d like and including it to a brand new object referred to as aspect. Each aspect is then pushed to the cabinets array to get a listing of objects containing simply the info we’re on the lookout for.

That is how a shelf object ought to seem like earlier than it’s added to our listing:

title: ‘SUPERJARE Wall Mounted Cabinets, Set of two, Show Ledge, Storage Rack for Room/Kitchen/Workplace – White’,
picture: ‘’,
hyperlink: ‘’,
value: ‘$32.99’,
evaluations: ‘6,171’,
stars: ‘4.7 out of 5 stars’

Format the info

Now that we have now managed to fetch the info we’d like, it’s a good suggestion to put it aside as a .CSV file to enhance readability. After getting all the info, we are going to use the fs module offered by Node.js and save a brand new file referred to as saved-shelves.csv to the challenge’s folder. Import the fs module on the high of the file and duplicate or write alongside the next traces of code:

let csvContent = => {
return Object.values(aspect).map(merchandise => “${merchandise}”).be part of(‘,’)
}).be part of(“n”)

fs.writeFile(‘saved-shelves.csv’, “Title, Picture, Hyperlink, Worth, Evaluations, Stars” + ‘n’ + csvContent, ‘utf8’, perform (err) {
if (err) {
console.log(‘Some error occurred – file both not saved or corrupted.’)
} else{
console.log(‘File has been saved!’)

As we will see, on the primary three traces, we format the info we have now beforehand gathered by becoming a member of all of the values of a shelve object utilizing a comma. Then, utilizing the fs module, we create a file referred to as saved-shelves.csv, add a brand new row that accommodates the column headers, add the info we have now simply formatted and create a callback perform that handles the errors.

The consequence ought to look one thing like this:

Bonus Suggestions!

Scraping Single Web page Purposes

Dynamic content material is changing into the usual these days, as web sites are extra complicated than ever earlier than. To supply the perfect consumer expertise attainable, builders should undertake completely different load mechanisms for dynamic content material, making our job slightly extra sophisticated. In case you don’t know what meaning, think about a browser missing a graphical consumer interface. Fortunately, there may be ✨Puppeteer✨ — the magical Node library that gives a high-level API to regulate a Chrome occasion over the DevTools Protocol. Nonetheless, it presents the identical performance as a browser, however it have to be managed programmatically by typing a few traces of code. Let’s see how that works.

Within the beforehand created challenge, set up the Puppeteer library by operating npm set up puppeteer, create a brand new puppeteer.js file, and duplicate or write alongside the next traces of code:

const puppeteer = require(‘puppeteer’)

(async () => {
strive {
const chrome = await puppeteer.launch()
const web page = await chrome.newPage()
await web page.goto(‘’)
await web page.waitForSelector(‘.rpBJOHq2PR60pnwJlUyP0’, { timeout: 2000 })

const physique = await web page.consider(() => {
return doc.querySelector(‘physique’).innerHTML


await chrome.shut()
} catch (error) {

Within the instance above, we create a Chrome occasion and open up a brand new browser web page that’s required to go to this hyperlink. Within the following line, we inform the headless browser to attend till the aspect with the category rpBJOHq2PR60pnwJlUyP0 seems on the web page. We have now additionally specified how lengthy the browser ought to wait for the web page to load (2000 milliseconds).

Utilizing the consider methodology on the web page variable, we instructed Puppeteer to execute the Javascript snippets inside the web page’s context simply after the aspect was lastly loaded. This may enable us to entry the web page’s HTML content material and return the web page’s physique because the output. We then shut the Chrome occasion by calling the shut methodology on the chrome variable. The resulted work ought to include all of the dynamically generated HTML code. That is how Puppeteer may also help us load dynamic HTML content material.

In case you don’t really feel comfy utilizing Puppeteer, be aware that there are a few alternate options on the market, like NightwatchJS, NightmareJS, or CasperJS. They’re barely completely different, however in the long run, the method is fairly related.

Setting user-agent Headers

user-agent is a request header that tells the web site you might be visiting about your self, specifically your browser and OS. That is used to optimize the content material in your set-up, however web sites additionally use it to establish bots sending tons of requests — even when it adjustments IPS.

Right here’s what a user-agent header appears to be like like:

Mozilla/5.0 (Home windows NT 10.0; Win64; x64) AppleWebKit/537.36 (KHTML, like Gecko) Chrome/93.0.4577.82 Safari/537.36

Within the curiosity of not being detected and blocked, it is best to usually change this header. Take further care to not ship an empty or outdated header since this could by no means occur for a run-fo-the-mill consumer, and also you’ll stand out.

Charge Limiting

Net scrapers can collect content material extraordinarily quick, however it is best to keep away from going at high velocity. There are two causes for this:

Too many requests in brief order can decelerate the web site’s server and even convey it down, inflicting bother for the proprietor and different guests. It may basically turn out to be a DoS assault.
With out rotating proxies, it’s akin to loudly asserting that you’re utilizing a bot since no human would ship a whole bunch or hundreds of requests per second.

The answer is to introduce a delay between your requests, a observe referred to as “charge limiting”. (It’s fairly easy to implement, too!)

Within the Puppeteer instance offered above, earlier than creating the physique variable, we will use the waitForTimeout methodology offered by Puppeteer to attend a few seconds earlier than making one other request:

await web page.waitForTimeout(3000);

The place ms is the variety of seconds you’ll wish to wait.

Additionally, if we’d wish to do the identical thig for the axios instance, we will create a promise that calls the setTimeout() methodology, with a purpose to assist us look forward to our desired variety of miliseconds:

fetchShelves.then(consequence => new Promise(resolve => setTimeout(() => resolve(consequence), 3000)))

On this means, you may keep away from placing an excessive amount of strain on the focused server and likewise, convey a extra human strategy to internet scraping.

Closing Ideas

And there you’ve it, a step-by-step information to creating your personal internet scraper for Amazon product information! However keep in mind, this was only one scenario. In case you’d wish to scrape a special web site, you’ll should make a number of tweaks to get any significant outcomes.

Associated Studying

In case you’d nonetheless wish to see extra internet scraping in motion, right here is a few helpful studying materials for you:

The Final Information to Net Scraping with JavaScript and Node.Js,” Robert Sfichi
Superior Node.JS Net Scraping with Puppeteer,” Gabriel Cioci
Python Net Scraping: The Final Information to Constructing Your Scraper,” Raluca Penciuc

    About Marketing Solution Australia

    We are a digital marketing company with a focus on helping our customers achieve great results across several key areas.

    Request a free quote

    We offer professional SEO services that help websites increase their organic search score drastically in order to compete for the highest rankings even when it comes to highly competitive keywords.

    Subscribe to our newsletter!

    More from our blog

    See all posts

    Leave a Comment