Anti-scraping tool: Sentinel
Date of Publication
2014
Document Type
Bachelor's Thesis
Degree Name
Bachelor of Science in Computer Science
College
College of Computer Studies
Department/Unit
Computer Science
Thesis Adviser
Alexis V. Pantola
Defense Panel Member
Arlyn Verina L. Ong
Isaac Sabas
Alexie Ballon
Abstract/Summary
Anti-scraping tools (AST) are network security applications that are tasked to ensure that no bots or scrapers can access the website and extract different information. Certain flaws of most anti-scraping tools are only blocking and preventing bots and scrapers intruding in the website. Websites can suffer from repetitive intrusion intervals and bots or scrapers can alter their algorithm in order to bypass the security. Existing anti-scraping tools are dependent on their predetermined list of IP addresses which can be circumvented by attaining a new authentic IP address. The study aims to develop an anti scraping tool that does not solely depend on predetermined IP addresses and diminishes the occurrences of scrapers attacking with new authentic IP addresses. Experiments show to deter automatic scrapers and keep them from repeatedly attacking with short intervals in between each attack. Sentinel detects automatic scrapers through the use of the Blacklist, Rate, Limiter, Test Provider, and Test Checker Modules, feeds the detected scrapers fake information with the use of the Web Trap Module, provides additional information on them with the IP Address Lookup Module, and delays the scrapers it cannot detect using the Computed Hide method explained later on in Chapter 5.8.
Abstract Format
html
Language
English
Format
Accession Number
TU18426
Shelf Location
Archives, The Learning Commons, 12F, Henry Sy Sr. Hall
Physical Description
1 v. (various foliations) : illustrations (some colored) ; 28 cm.
Recommended Citation
Alejandro, P. C., De La Paz, A. H., Guevara, J., & Ong David, J. S. (2014). Anti-scraping tool: Sentinel. Retrieved from https://animorepository.dlsu.edu.ph/etd_bachelors/11812