Anti-scraping tool: Sentinel

Date of Publication

2014

Document Type

Bachelor's Thesis

Degree Name

Bachelor of Science in Computer Science

College

College of Computer Studies

Department/Unit

Computer Science

Thesis Adviser

Alexis V. Pantola

Defense Panel Member

Arlyn Verina L. Ong
Isaac Sabas
Alexie Ballon

Abstract/Summary

Anti-scraping tools (AST) are network security applications that are tasked to ensure that no bots or scrapers can access the website and extract different information. Certain flaws of most anti-scraping tools are only blocking and preventing bots and scrapers intruding in the website. Websites can suffer from repetitive intrusion intervals and bots or scrapers can alter their algorithm in order to bypass the security. Existing anti-scraping tools are dependent on their predetermined list of IP addresses which can be circumvented by attaining a new authentic IP address. The study aims to develop an anti scraping tool that does not solely depend on predetermined IP addresses and diminishes the occurrences of scrapers attacking with new authentic IP addresses. Experiments show to deter automatic scrapers and keep them from repeatedly attacking with short intervals in between each attack. Sentinel detects automatic scrapers through the use of the Blacklist, Rate, Limiter, Test Provider, and Test Checker Modules, feeds the detected scrapers fake information with the use of the Web Trap Module, provides additional information on them with the IP Address Lookup Module, and delays the scrapers it cannot detect using the Computed Hide method explained later on in Chapter 5.8.

Abstract Format

html

Language

English

Format

Print

Accession Number

TU18426

Shelf Location

Archives, The Learning Commons, 12F, Henry Sy Sr. Hall

Physical Description

1 v. (various foliations) : illustrations (some colored) ; 28 cm.

This document is currently not available here.

Share

COinS