Anti-scraping tool: Sentinel

Date of Publication


Document Type

Bachelor's Thesis

Degree Name

Bachelor of Science in Computer Science


College of Computer Studies


Computer Science

Thesis Adviser

Alexis V. Pantola

Defense Panel Member

Arlyn Verina L. Ong
Isaac Sabas
Alexie Ballon


Anti-scraping tools (AST) are network security applications that are tasked to ensure that no bots or scrapers can access the website and extract different information. Certain flaws of most anti-scraping tools are only blocking and preventing bots and scrapers intruding in the website. Websites can suffer from repetitive intrusion intervals and bots or scrapers can alter their algorithm in order to bypass the security. Existing anti-scraping tools are dependent on their predetermined list of IP addresses which can be circumvented by attaining a new authentic IP address. The study aims to develop an anti scraping tool that does not solely depend on predetermined IP addresses and diminishes the occurrences of scrapers attacking with new authentic IP addresses. Experiments show to deter automatic scrapers and keep them from repeatedly attacking with short intervals in between each attack. Sentinel detects automatic scrapers through the use of the Blacklist, Rate, Limiter, Test Provider, and Test Checker Modules, feeds the detected scrapers fake information with the use of the Web Trap Module, provides additional information on them with the IP Address Lookup Module, and delays the scrapers it cannot detect using the Computed Hide method explained later on in Chapter 5.8.

Abstract Format






Accession Number


Shelf Location

Archives, The Learning Commons, 12F, Henry Sy Sr. Hall

Physical Description

1 v. (various foliations) : illustrations (some colored) ; 28 cm.

This document is currently not available here.