Skip to content

The Complete Beginner‘s Guide to JavaScript Web Scraping

Hey there! Web scraping is a useful technique for automating data collection from websites. With the shift towards modern JavaScript-heavy web apps, Node.js has become the go-to platform for writing scalable scrapers.

In this comprehensive tutorial, you‘ll learn how to use Node.js for scraping by following along with hands-on examples. I‘ll share the tips and tricks I‘ve learned from building scrapers in Node over the past decade. Let‘s get started!

A Brief History of Web Scraping

Web scraping itself has been around almost as long as the web. In the early days, most websites were simple HTML pages without much dynamic content. Scraping them mainly involved parsing HTML using regular expressions.

As the web evolved to include more JavaScript, scrapers had to run real browsers like Selenium to render full pages. The rise of modern web frameworks like React and Vue has continued this trend.

Node.js brought JavaScript out of the browser and made it possible to write complex applications and tools in JS. Combined with npm‘s huge ecosystem of packages, Node has become the standard for scalable and high-performance web scraping today.

Major companies are using Node scrapers to power use cases like price monitoring, ad verification, lead generation and more. The demand for web scraped data continues to grow in fields like machine learning, business intelligence and cybersecurity.

Why Use Node.js for Scraping?

There are several advantages that make Node.js a great platform for web scraping:

Familiar Language

Since JavaScript knowledge is transferable, most developers already have the core skills to write Node scrapers. You don‘t need to context switch and learn another language like Python.


Node‘s asynchronous, event-driven architecture makes web scraping much faster compared to traditional synchronous languages.


You can easily distribute scrapers across multiple threads and servers by breaking requests into jobs.

Dynamic Sites

Node runners like Puppeteer can render full web pages including any JavaScript. This allows you to scrape interactive sites.

Huge npm Ecosystem

npm has packages for everything from requests and browser automation to databases and machine learning.

Beginner Friendly Resources

With Node‘s popularity, there are abundant docs, guides and help available when you get stuck.

Let‘s now dive into the key libraries and concepts for building scalable scrapers in Node.js…

Sending Requests with Axios

The first step in any web scraping project is sending requests to the target URLs and receiving the raw HTML response. There are several packages in Node that allow us to make HTTP requests.

The most popular one today is Axios. Here is a simple GET request using Axios:

const axios = require(‘axios‘);

const response = await axios.get(‘‘);

Axios has a clean promise-based syntax and automatically parses JSON data. It also makes it easy to:

  • Configure default headers like User-Agent
  • Handle request errors
  • Make simultaneous requests with allsettled
  • Set HTTP adapter to use proxies

This makes Axios a very full-featured choice for scraping requests.

Parsing HTML with Cheerio

After getting the raw HTML response, we need to parse it and extract the data we actually want. This usually involves traversing the DOM and selecting elements by class, id, attributes, etc.

The ubiquitous jQuery library makes DOM manipulation very easy in the browser. To enable the same for serverside Node.js, we use Cheerio.

Cheerio parses HTML and allows you to query it just like jQuery:

const $ = cheerio.load(;

const headings = $(‘h2‘); // Selects all h2 elements

With Cheerio, you can:

  • Use CSS selectors and jQuery methods like find, filter
  • Extract text, attribute values, HTML
  • Traverse DOM tree with parent, next etc
  • Modify DOM nodes and structures

This makes Cheerio very powerful and convenient for parsing HTML in Node scrapers.

Scraping Dynamic Content with Puppeteer

Modern websites rely heavily on JavaScript to dynamically load content. Scraping these sites requires a headless browser to run the JavaScript.

This is where Puppeteer comes in. Puppeteer provides a high-level API to control Chromium or Chrome programmatically.

Here is how you can launch a browser instance and load dynamic content:

const browser = await puppeteer.launch();
const page = await browser.newPage();

await page.goto(‘‘); 
await page.waitForSelector(‘.products‘);
// Extract products...

await browser.close();

With Puppeteer, you can also simulate user actions like clicks, scrolls, form inputs etc. This makes it very powerful for scraping interactive sites.

Some examples of what you can do:

  • Generate previews of pages as images
  • Automate form submissions
  • Create PDFs of pages
  • Test browser-side applications
  • Debug cross-browser issues
  • Audit for SEO, speed, and accessibility

Storing Scraped Data

Once you‘ve extracted the data, you need to store it somewhere for later processing and analysis. Popular options for Node scrapers include:

Text/JSON Files

You can use Node‘s inbuilt fs module to write data to files. For example:

const fs = require(‘fs‘);

const scrapedData = [
  {name: ‘Page 1‘, url: ‘https://...‘},
  {name: ‘Page 2‘, url: ‘https://...‘} 

fs.writeFile(‘data.json‘, JSON.stringify(scrapedData), err => {
  if (err) console.log(err);

CSV Files

For tabular data, CSV format is a convenient option. Use a module like json2csv to convert JSON to CSV and save to a file.


For large amounts of structured data, a database like MySQL, MongoDB or PostgreSQL is recommended. Popular Node ORM/ODM libraries include Sequelize and Mongoose.

Asynchronous Control Flows

A key advantage of Node.js is its asynchronous, event-driven architecture. Instead of blocking code, Node uses non-blocking paradigms for I/O:

  • Callbacks – Pass functions to be executed after an operation finishes
  • Promises – Represent eventual completion/failure of async operations
  • Async/await – Write async code that looks synchronous

For scrapers, we leverage promises and async/await to achieve fast and non-blocking execution. Here is an example scraper function:

async function scrapePage(url) {

  try {
    const response = await axios.get(url);
    const $ = cheerio.load(;
    const data = $(‘.product‘).map(parseProduct).get();

    return data; 

  } catch (err) {


The async keyword allows us to use await instead of chaining .then(). This results in cleaner and more maintainable asynchronous code.

Scaling Web Scrapers

To build robust scrapers that can extract data from multiple pages, we need ways to scale our code. Here are some common techniques:


We can call scraper functions recursively to follow pagination or crawl websites:

async function scrapeAllPages(url) {

  const products = await scrapePage(url);

  if (hasNextPage(url)) {
    url = getNextPageUrl(url);
    products.push(await scrapeAllPages(url)) 

  return products;


We can run asynchronous tasks like requests in parallel with Promise.all():

const urls = [/* list of urls */];

const pageData = await Promise.all( => scrapePage(url)));

Worker Threads

We can parallelize scraping across multiple threads using Node‘s worker threads.


Tasks can be distributed across multiple servers using a queue like RabbitMQ.

Handling Errors

When writing scrapers, we need to plan for errors and failures – sites change often and requests fail due to network errors.

Robust scrapers implement error handling patterns like:

  • Request Retries – Retry failed requests up to N times.
  • Circuit Breakers – Stop requests after M failures to prevent resource exhaustion.
  • Fallback Data – Return cached or default data if requests fail.
  • Error Logging – Log failures to identify and fix root causes.

Here is an example with retries and logging:

// Retry each request up to 3 times
axios.defaults.retries = 3; 

try {
  const response = await axios.get(url);
  // ... scraping logic ...
} catch (err) {
  // return fallback or throw 

Proper error handling ensures your scraper degrades gracefully instead of crashing entirely.

Wrapping Up

In this guide, I covered the essential packages, concepts and techniques for building scalable web scrapers in Node.js:

  • Requests with Axios
  • HTML parsing with Cheerio
  • Browser automation using Puppeteer
  • Storing data in files/databases
  • Asynchronous control flow
  • Recursive scraping
  • Error handling

These building blocks will provide a solid foundation for your scraping projects. The Node ecosystem has many more helpful modules like proxies, queues and CLI tools.

I hope this tutorial gives you a comprehensive introduction to the world of JavaScript web scraping! Let me know if you have any other topics you‘d like me to cover. Happy scraping!

Join the conversation

Your email address will not be published. Required fields are marked *