Open Access Open Access  Restricted Access Subscription Access

Design of Web Crawler for the Client - Server Technology


Affiliations
1 Department of Computer and System Sciences, Jaipur National University, Jaipur - 302017, Rajasthan, India
2 Department of Mathematics, Galgotias University, Gr. Noida - 201306, Uttar Pradesh, India
 

Search engines store information locally with the purpose of deliver quick, accessible search abilities. This information is collected by Web crawler. Web crawling is necessary for the maintenance of complete and latest web document gathering for a web search tool. The web document modifies its content on regular basis hence it becomes necessary to build up a successful framework which could identify these sorts of changes proficiently in the most minimal scanning time to accomplish this modifications. The essential thought behind designing of such a web crawler is to find high quality web documents within limited time frame. The proposed system works on the Client-Server Technology it reduces the overlap problem and downloads high quality web pages. We can add many web crawler in parallel to download web page in parallel way.

Keywords

Client-Server Technology, Overlap, Search Engine, Web Crawler, Web Page.
User

Abstract Views: 229

PDF Views: 0




  • Design of Web Crawler for the Client - Server Technology

Abstract Views: 229  |  PDF Views: 0

Authors

Md. Abu Kausar
Department of Computer and System Sciences, Jaipur National University, Jaipur - 302017, Rajasthan, India
V. S. Dhaka
Department of Computer and System Sciences, Jaipur National University, Jaipur - 302017, Rajasthan, India
Sanjeev Kumar Singh
Department of Mathematics, Galgotias University, Gr. Noida - 201306, Uttar Pradesh, India

Abstract


Search engines store information locally with the purpose of deliver quick, accessible search abilities. This information is collected by Web crawler. Web crawling is necessary for the maintenance of complete and latest web document gathering for a web search tool. The web document modifies its content on regular basis hence it becomes necessary to build up a successful framework which could identify these sorts of changes proficiently in the most minimal scanning time to accomplish this modifications. The essential thought behind designing of such a web crawler is to find high quality web documents within limited time frame. The proposed system works on the Client-Server Technology it reduces the overlap problem and downloads high quality web pages. We can add many web crawler in parallel to download web page in parallel way.

Keywords


Client-Server Technology, Overlap, Search Engine, Web Crawler, Web Page.



DOI: https://doi.org/10.17485/ijst%2F2015%2Fv8i36%2F130028