Jump to ratings and reviews
Rate this book

Information Retrieval

Rate this book
The information on the WWW is growing at an exponential rate; therefore, search engines are required to index the downloaded Web documents more efficiently. A typical search engine comprises of the three main components. (1) Crawler: Given a URL, it combs through the pages on the web and gathers the required information for the search engine. (2) Indexer: While an index of 100,000 documents can be queried within millisecond; a sequential scan may take hours.An indexer that optimizes speed and performance for finding relevant documents for a search query (3) Page Repository: The information retrieved by the web crawler is stored in a database called page repository. Web mining techniques like clustering can be used for this purpose. The performance of a search engine is limited because of these two problems. (1) Low precision (2) Low recall. Thus, there is a need to develop efficient indexing technique. In this book, a novel technique is being discussed that not only indexes the downloaded web documents efficiently but also uses a web-mining technique to make the indexed information searchable, enabling the search engines to provide more relevant results.

128 pages, Paperback

First published April 9, 2010

6 people want to read

About the author

Deepti Gupta

124 books3 followers

Ratings & Reviews

What do you think?
Rate this book

Friends & Following

Create a free account to discover what your friends think of this book!

Community Reviews

5 stars
0 (0%)
4 stars
0 (0%)
3 stars
0 (0%)
2 stars
0 (0%)
1 star
0 (0%)
No one has reviewed this book yet.

Can't find what you're looking for?

Get help and learn more about the design.