• Created by: funbiiiii
  • Created on: 03-01-20 03:36

Crawler Module

to get webpages from websites by following hyperlinks

1 of 6

Page Repository

stores pages obtained by the crawler for use, subsequent indexing or may support de-duplication

2 of 6

Indexing module

To analyse content of the webpages and construct information for the indexes

3 of 6


Offer the data structures needed to support the query operation and return of data

4 of 6

Query module

to resolve the string typed by the user into the form required to gain information from the indexes

5 of 6

Ranking module

to determine how to present the result obtained from the index

6 of 6


No comments have yet been made

Similar Computing resources:

See all Computing resources »See all Week 1 resources »