Call For Paper Volume:4 Issue:8 Aug'2017 |

FILTERING OF URLS USING WEBCRAWLER

Publication Date : 29/07/2015



Author(s) :

Arya Babu , Misha Ravi.


Volume/Issue :
Volume 2
,
Issue 7
(07 - 2015)



Abstract :

Web crawler is a computer program that browses the World Wide Web in a methodical, automated manner or in an orderly fashion. Web crawlers are mainly used to create a copy of all the visited pages for later processing by a search engine that will index the downloaded pages to provide fast searches. An efficient web crawler algorithm is required so as to extract required information in less time and with highest accuracy. As the number of Internet users and the number of accessible Web pages grows, it is becoming increasingly difficult for users to find documents that are relevant to their particular needs. Users must either browse through a large hierarchy of concepts to find the information for which they are looking or submit a query to a publicly available search engine and wade through hundreds of results, most of them irrelevant. Web crawlers are one of the most crucial components in search engines and their optimization would have a great effect on improving the searching efficiency. Generally web crawler rejects the page whose url does not contain the search keyword while searching information onWorldWideWeb. But it may so happen that these pages may contain information required. The main emphasis will be to scan these pages and parse them check for their relevancy


No. of Downloads :

5


Indexing

Web Design MymensinghPremium WordPress ThemesWeb Development

FILTERING OF URLS USING WEBCRAWLER

July 28, 2015