Adaptive on-line page importance computation
Abiteboul, Serge; Preda, Mihai; Cobena, Gregory
2003-05-20 00:00:00
Adaptive On-Line Page Importance Computation Serge Abiteboul INRIA Domaine de Voluceau 78150 Rocquencourt, France Mihai Preda Xyleme S.A. 6 rue Emile Verhaeren 92210 Saint-Cloud, France Gregory Cobena INRIA Domaine de Voluceau 78150 Rocquencourt, France [email protected] ABSTRACT [email protected][email protected] often1 and when crawling for new pages, important pages have to be fetched rst [9]. Following some ideas of [18], Page and Brin proposed a notion of page importance based on the link structure of the web [5]. This was then used by Google with a remarkable success. Intuitively, a page is important if there are many important pages pointing to it. This leads to a xpoint computation by repeatedly multiplying the matrix of links between pages with the vector of the current estimate of page importance until the estimate is stable, i.e., until a xpoint is reached. The main issue in this context is the size of the web, billions of pages [4, 23]. Techniques have been developed to compute page importance ef ciently, e.g., [16]. The web is crawled and the link matrix computed and stored. A version of the matrix is then frozen and one separate process computes off-line page importance, which may take hours or days
http://www.deepdyve.com/assets/images/DeepDyve-Logo-lg.pnghttp://www.deepdyve.com/lp/association-for-computing-machinery/adaptive-on-line-page-importance-computation-72ObUMaSk4
Adaptive On-Line Page Importance Computation Serge Abiteboul INRIA Domaine de Voluceau 78150 Rocquencourt, France Mihai Preda Xyleme S.A. 6 rue Emile Verhaeren 92210 Saint-Cloud, France Gregory Cobena INRIA Domaine de Voluceau 78150 Rocquencourt, France [email protected] ABSTRACT [email protected][email protected] often1 and when crawling for new pages, important pages have to be fetched rst [9]. Following some ideas of [18], Page and Brin proposed a notion of page importance based on the link structure of the web [5]. This was then used by Google with a remarkable success. Intuitively, a page is important if there are many important pages pointing to it. This leads to a xpoint computation by repeatedly multiplying the matrix of links between pages with the vector of the current estimate of page importance until the estimate is stable, i.e., until a xpoint is reached. The main issue in this context is the size of the web, billions of pages [4, 23]. Techniques have been developed to compute page importance ef ciently, e.g., [16]. The web is crawled and the link matrix computed and stored. A version of the matrix is then frozen and one separate process computes off-line page importance, which may take hours or days
To get new article updates from a journal on your personalized homepage, please log in first, or sign up for a DeepDyve account if you don’t already have one.
All DeepDyve websites use cookies to improve your online experience. They were placed on your computer when you launched this website. You can change your cookie settings through your browser.