Anti-scraping tool: Sentinel
Date of Publication
Bachelor of Science in Computer Science
College of Computer Studies
Alexis V. Pantola
Defense Panel Member
Arlyn Verina L. Ong
Anti-scraping tools (AST) are network security applications that are tasked to ensure that no bots or scrapers can access the website and extract different information. Certain flaws of most anti-scraping tools are only blocking and preventing bots and scrapers intruding in the website. Websites can suffer from repetitive intrusion intervals and bots or scrapers can alter their algorithm in order to bypass the security. Existing anti-scraping tools are dependent on their predetermined list of IP addresses which can be circumvented by attaining a new authentic IP address. The study aims to develop an anti scraping tool that does not solely depend on predetermined IP addresses and diminishes the occurrences of scrapers attacking with new authentic IP addresses. Experiments show to deter automatic scrapers and keep them from repeatedly attacking with short intervals in between each attack. Sentinel detects automatic scrapers through the use of the Blacklist, Rate, Limiter, Test Provider, and Test Checker Modules, feeds the detected scrapers fake information with the use of the Web Trap Module, provides additional information on them with the IP Address Lookup Module, and delays the scrapers it cannot detect using the Computed Hide method explained later on in Chapter 5.8.
Archives, The Learning Commons, 12F, Henry Sy Sr. Hall
1 v. (various foliations) : illustrations (some colored) ; 28 cm.
Alejandro, P. C., De La Paz, A. H., Guevara, J., & Ong David, J. S. (2014). Anti-scraping tool: Sentinel. Retrieved from https://animorepository.dlsu.edu.ph/etd_bachelors/11812