Skip to content

Latest commit

 

History

History
76 lines (48 loc) · 1.66 KB

README.md

File metadata and controls

76 lines (48 loc) · 1.66 KB

crawling

A simple crawler made in JavaScript for Node.

Installation

crawling is both available on GitHub Packages and npm.

How to install from GitHub Packages

To install, you first have to follow this guide on GitHub Docs. Then, you can run:

$ npm install @lgrachov/[email protected]

This should install the package in your project.

How to install from npm

You only need to run one command:

$ npm install crawling

This should install the package in your project.

Usage

Creating an array with all of the links

This example will create an array with all of the links gathered from the page.

import { crawlSite } from "crawling";

const links = [];
for await (const url of crawlSite("https://github.com/", 500)) {
  links.push(url);
}

Log each one of the links

This example will log each one of the links received, without a delay like the previous example had.

import { crawlSite } from "crawling";

for await (const url of crawlSite("https://github.com/", 500)) {
  console.log(url);
}

Documentation

The function crawlSite takes two parameters:

  • site: Required. The site to crawl.
  • timeout: Optional. The timeout between each link in miliseconds, default is 500.

There are examples of usage, above and below:

import { crawlSite } from "crawling";

// this should choose a random url

const links = [];
for (const url of await crawlSite("https://github.com/", 500)) {
  links.push(url);
}

console.log(shuffle(links)[0]);