Web Graphs — PageRank-like Importance Measures
FR

Chapter 3 — Graphs and structures of the web

Life is an ephemeral butterfly bearing the wings of paradox.

Benoît Gagnon

Places do not change their appearance as men change their faces.

Tacitus

A good website is always under construction.

Anonymous

Now that the concepts of the Web and of crawling have been clarified, we can study what can be said about them. Since this chapter is devoted to structures, it is appropriate to place a canonical structure on the crawls we wish to study. One of the simplest and most natural structures in the case of Web crawls is the graph structure. Web graphs, a definition of which is given in Section 3.1, are a frequent subject of study. After critically analyzing one of the best-known models, the butterfly model (Section 3.2), we will highlight in the remainder of this chapter the importance of the site structure in Web graphs.

3.1 Web graphs: definition

Consider a crawl . This crawl consists of a set of URLs, some of which have been visited, and others merely detected through the hyperlinks of visited pages. It also contains the set of hyperlinks from visited pages. We will call the graph of crawl the directed graph , where is the set of pages in the crawl, whether visited or not, and such that an arc connects a page to a page if, and only if, contains a hyperlink pointing to . By convention, anchors1 are ignored, and multiple links add nothing to the graph (if a page has two hyperlinks pointing to , the graph will still have only one arc from to ). Likewise, links from a page to itself will not be taken into account.

By abuse of language, we will often use the term Web graph to designate such a graph . One should always remember that these are in fact always crawl graphs.

3.2 The butterfly model revisited

The butterfly structure of the Web graph is today accepted by many people, and is very often mentioned in scientific articles [APC03, Ara+01, CF02, Kam+03b, LM04, RG03] or in popular science publications [GL02], sometimes in a completely irrelevant manner2. I consider for my part that the butterfly model applies poorly to dynamic graphs in general, and to Web graphs in particular. This section, whose ideas come from [Mat01], will try to justify this position by attempting to distinguish myth from reality in order to identify the true contribution of the article Graph Structure in the Web.

3.2.1 The butterfly model

Graph Structure in the Web [Bro+00] is an article that was presented at the 9th International Conference on the World Wide Web. The article is based on the analysis of two crawls provided by the AltaVista search engine, a crawl of 203 million pages dated May 1999 and one of 271 million pages from October 1999. The analysis yields several results on the structure of the Web graph, the most novel being the discovery of a butterfly structure:

Figure 3.1: The butterfly structure

This structure is commonly represented by a diagram similar to that of Figure 3.1. The butterfly shape obtained from the IN, SCC, and OUT components gives its name to the model (bow tie).

From this article, many people have retained the idea that this roughly equal division into IN component, SCC, OUT, and tendrils is characteristic of the Web: this is the butterfly model.

3.2.2 Weaknesses of the butterfly model

Does it still exist?

The main drawback of the butterfly model is that in all likelihood, it no longer represents reality. The Web graph, that is, the graph of the accessible Web, cannot have a balanced butterfly structure, since it contains among other things the page that points to all pages (cf. Section 1.4). One must therefore consider that this model applies to graphs derived from large Web crawls (this is moreover the opinion of the authors3). However, one observes that:

What can then happen for a page that, for a given crawl, belongs to the IN component, or to the tendrils, or even to the disconnected components:

In conclusion, yes, there can be an IN component, or disconnected components, but these components are necessarily transient, precisely because of the relationships between crawlers and directories. One must also take into account the fact that current large crawls contain a non-negligible proportion of non-indexed pages (at least 25% according to a 2002 study [Sea], 75% according to a 2004 article [EMT04]), which automatically inflate the OUT component. For all these reasons, I am inclined to strongly doubt that more than 2 (3?) billion pages indexed by Google belong to neither the core nor OUT.

A model lacking robustness

A question one may ask is whether the parameters considered are relevant for the study of Web graphs. On the one hand, we have crawls, which try as best they can to traverse a Web that is dynamic both spatially and temporally, and which often have a very low overlap between them [BB98]. On the other hand, definitions such as IN: the set of pages that can reach the core but are not reachable from it. Is it reasonable to combine highly variable crawls with fragile definitions? A concrete example: imagine that an unscrupulous individual seizes the set of initial pages from AltaVista’s crawls and places links to these pages on their Web page. If this page is in the core, this means the collapse of all components onto SCC and OUT4.

Contrary to what the authors claim5, a single page, not even that large6, suffices to undermine the butterfly model. We are in fact dealing with a methodological flaw: when faced with data subject to great uncertainty and dynamics (Web graphs), it makes no sense to treat variables that are extremely sensitive to initial conditions as universal constants. For the same reasons, considering the diameter of the connected component is hardly relevant; indeed, for the 2 AltaVista crawls studied, it remains roughly constant (around 500), but upon closer inspection, evidence seems to suggest that the diameter of 500 is due to a single chain, possibly a misconfigured page or a robot trap that escaped the filters7.

An unreplicated experiment

To my knowledge, the experiment of [Bro+00] has not been replicated on large crawls since. The only real experiments conducted on the butterfly model therefore concern two proprietary AltaVista crawls, close in time and thus likely based on the same technology (which may explain certain artifacts such as the constancy of the diameter). Moreover, if one compares the size and dates of these two crawls (May 1999, 203 million pages, October 1999, 271 million pages) with Figure 2.2b and Figure 2.2c (Figure 2.2b), one reaches the conclusion that the crawls used were certainly experimental.

In [Dil+01], it is nonetheless found that the butterfly structure is a fractal figure of the Web structure, meaning that one finds within sites and communities the same butterfly model as on the entire graph. Upon closer inspection, one notices that the global butterfly structure is taken for granted, and that the proportions of IN, OUT, and SCC observed in the subgraphs are extremely variable, which supports the preceding remarks.

3.2.3 The butterfly model in a sociological context

To properly understand the success of the butterfly model, it is important to place it in context. We are at the beginning of the year 2000. The Internet Bubble is still in full growth, and articles are appearing indicating that search engines can no longer keep up with the growth of the Web [Ber00, Bra97, LG99]. The AltaVista company had won the first Crawl War (see Figure 2.2b as well as the conclusions of [BB98]) by surpassing 150 million indexed pages, but it lacked the assurance that this increase in size would allow it to keep pace with the evolution of the Web. In this context, Graph Structure in the Web was perfectly timed. Indeed, what is the takeaway of the article?

In short, Graph Structure in the Web responds to an expectation that is both existential (Can we understand the Web?) and commercial (Our crawl is the Web.). One could almost see it as a mechanism arising from the laws of supply and demand. Ultimately, the question of whether its scientific foundations are sound or not becomes secondary, which is perhaps a beginning of an explanation for the success of this article.

3.2.4 Conclusion

The butterfly model most certainly had meaning at the time of its discovery, but that meaning is not the Web has a butterfly structure. It rather expresses a particular situation where, because of the Internet Bubble, new pages were arriving too quickly for the Web to assimilate them. But today, two phenomena mean that we are no longer in this situation: on the one hand, the Bubble has burst. The creation of new sites has become rare, and we are witnessing more the development of existing sites than the appearance of a swarm of unregistered sites (cf. Section 3.3.2), which is a beginning of an explanation for the fact that the butterfly structure is found within site graphs [Dil+01]. On the other hand, search engines no longer merely index the Web; they continuously modify its structure through directories. These two effects, combined with the model’s lack of robustness, make it unlikely that large commercial crawls still have the same proportion of IN, tendrils, and other disconnected components as those found in Graph Structure in the Web.

One should nevertheless retain that Broder et al. highlighted the existence of a large core of pages all connected to one another, as well as the protectionist behavior of certain commercial sites that place themselves entirely within the OUT component. One should also retain that the proportion of the IN component in a crawl graph is a certain indication of the difficulty a large structured graph has in adapting to rapid expansion.

3.3 Role of servers and sites in the structure of the Web graph

The notion of site is fundamental in the organization of the Web. For example, many search engines have a policy of returning only a limited number of pages per site (with a more pages from the same site option available). A good decomposition into sites has numerous applications, ranging from PageRank computation methods [Kam+03b, MV04] to efficient compression methods for crawls [GLV02, Ran+01].

3.3.1 Web servers, sites, and communities

The concepts of Web server, site, and community are often encountered in the literature, with sometimes contradictory definitions. We will make explicit here what we mean by server, site, and community.

Physical server

A physical Web server is defined as a specific physical machine connected to the Internet and identified by an IP address, which returns a 200 response code and a Web page in response to an HTTP request asking for the root (/) on port 808. The physical website associated with the server consists of all pages hosted at the given IP address.

Virtual server

Since IP addresses are not unlimited, solutions had to be devised to conserve addresses. One such solution is the use of virtual servers. Virtual servers, introduced with the HTTP 1.1 standard, allow a single IP address to behave as several servers differentiated by their Domain Name System (DNS) address (Hostname). This method is used, for example, by free for hosting its clients’ personal pages, or in Japan, where IP address resources are very limited. From the user’s perspective, a virtual server is indistinguishable from a physical server, except that one cannot replace the DNS address with the IP address.

Logical site

A logical website consists of a set of pages strongly connected by hyperlinks. Intuitively, a site is characterized by a certain homogeneity of style (imposed by the Webmaster) and easy navigation among the site’s pages. The study of sites will be the subject of sections Section 3.3.3 and Section 3.3.4.

Community

The notion of community is in some sense orthogonal to the notion of site. A community is a set of pages connected by a common interest. The search for communities is at the heart of Kleinberg’s Hyperlink-Induced Topic Search (HITS) algorithm [Kle98] as well as the CosmoWeb project [BBM03].

3.3.2 Evolution of the number of servers

It is easier to estimate the number of Web servers (physical and virtual) than the number of Web pages.

Figure 3.2: Physical Web servers (IP addresses)

To count physical servers, it “suffices” to make an HTTP request to all possible IP addresses. Now, if we ignore IPv6 addresses, which remain relatively negligible, there are only possible addresses, or approximately four billion9. This work was carried out by the Web Characterization Project [O'N+02], over the period 1998–2002, and the results concerning the number of servers are shown in Figure 3.2. The number of unique physical servers is the number of physical servers minus the duplicates, that is, servers returning exactly the same content as a previously tested server.

The main remark one can make regarding these figures is the observation of a stagnation in the number of physical servers since 2001. This stagnation can be explained by the end of the Internet Bubble. Indeed, the Web has ceased to be a new technology, and most academic, commercial, and social actors likely to integrate into the Web have already done so. According to the Web Characterization Project, almost no new Web servers are being created; growth occurs individually at the server level.

Figure 3.3: Virtual Web servers (domain names)

The study of virtual servers was carried out by the Netcraft website [Net04]. It consists of listing all declared DNS addresses (like the WhoIs site [Who04]), and testing by means of an HTTP request the existence of a server at each of the addresses obtained.

The evolution of virtual Web servers, according to Netcraft [Net04], shows a roughly linear constant progression in the number of active virtual Web servers over the last 4 years (cf. Figure 3.3). This is in any case not, a priori, a geometric progression.

3.3.3 Intuitive and visual approach to the notion of site

The concept of site is, in our view, fundamental to the structure of the Web. Its applications are numerous:

Most of the time, sites are approximated by servers ([Ada99]). But reality can sometimes be more complex. Large sites may rely on several servers, while conversely a server may host several sites. For example, one would want to group www.microsoft.com and download.microsoft.com and consider them as part of the same site. Conversely, not all personal site hosting providers use virtual domains; a large proportion places sites in the directories of a single server.

All these considerations lead us to raise the question of the definition of a logical site. Human users generally have no difficulty knowing what a site is, which constitutes a kind of empirical definition, but is hardly rigorous and does not lend itself to automation. Note that sites manually listed by directories fall under this definition.

The Web Characterization Project [O'N+02] proposes the following semantic definition:

[A Web Site (Information Definition) is] a set of related Web pages that, in the aggregate, form a composite object of informational relevance. Informational relevance implies that the object in question addresses a non-trivial information need.

[A logical website is] a set of related pages that, in the aggregate, form a composite object delivering relevant information. This relevant information implies that the object in question addresses a non-trivial information need.

To this semantic characterization, Li, Kolak, Vu, and Takano add in [Li+00] a structural approach to define the notion of logical domain:

A logical domain is a group of pages that has a specific semantic relation and a syntactic structure that relates them.

A logical domain is a group of pages that possess a specific semantic relation and a syntactic structure that connects them.

Based on this definition, they define a set of rules (semantic and structural) to identify sites.

More modestly, our approach is to see what can be achieved by considering only the structural aspect of pages. We therefore limit ourselves to, on the one hand, the graph structure induced by hyperlinks, and on the other hand, the tree structure of URLs that reveals natural clusters in crawl graphs. Indeed, very often, a logical site obeys hierarchical rules at the level of URLs (Uniform Resource Locators [BMM94]), the most common being the existence of a characteristic common prefix.

Figure 3.4: URLs: a natural decomposition tree for Web graphs

This decomposition tree10 structure on graphs (clustered graphs) was introduced by Feng et al. in [FCE95] as a tool for representing large graphs. Its main use is therefore to allow drawing graphs in a way that reveals any existing hierarchy, and decomposition trees are mainly used in domains where implicit or explicit diagram structures exist11.

The whole problem is then to find, for a given graph, the decomposition tree that offers the best structural representation [Bri03]. In the case of Web graphs, the URLs tree provides an intrinsic decomposition tree.

Definition

Let be a graph. A decomposition tree of is a tree whose leaves correspond to the vertices . Each internal node of defines a cluster (a set of vertices) . This cluster consists of the set of vertices of corresponding to the leaves of the subtree of rooted at . For example, the cluster associated with the root of is the entire set .

A good example of the use of decomposition trees is the modeling of human relationships. There exists indeed a fairly natural graph of human relationships, where vertices are people and where a person points to a person if knows ). A decomposition tree that can meaningfully complement this graph is that of geographic location: world, country, (state), city, neighborhood, street, …

Web graphs and the URL decomposition tree

URLs provide a natural decomposition tree structure on Web graphs. The set of servers12 can be represented as a tree whose root is http, the depth-1 nodes are the TLDs13, followed by the domain names proper and possibly subdomains. Each server, which is a leaf of the domain name tree, hosts a page hierarchy identical to the structure of the corresponding physical file system. The union of the file trees within the server tree yields the URLs decomposition tree.

For example, the URL http://smith.mysite.org/linux/index.html decomposes into HTTP, org, mysite, smith, linux, and finally index.html, thus forming a path from the root to the corresponding leaf in the decomposition tree (see Figure 3.4).

One may also question the relevance of beginning the decomposition with the top-level domain (TLD). Indeed, there exist macro-sites that, in order to achieve better visibility, deploy across several TLDs (.com and the ccTLDs of the countries where they operate, quite classically). We nevertheless prefer to retain sorting by TLD, on the one hand because this corresponds to the original philosophy of domain names, and on the other hand because there also exist domain names for which the TLD makes a considerable difference. The discerning adult reader may, for example, note a definite semantic difference between the content of the site http://www.france2.fr and that of http://www.france2.com

Seeing the structure of sites

Figure 3.5: Adjacency matrix of pages from a crawl of million pages of .fr

Given that, in general, webmasters try to organize their sites, one can expect the concept of website to be intimately connected to the URL decomposition tree. We have confirmed this by observing the graphical representation of the adjacency matrix of the graph of a crawl of approximately 8 million URLs sorted in lexicographic order from .fr, carried out in June 2001 as part of the cooperative research action Soleil Levant [Sol01].

We have represented a small portion of this crawl (60,000 pages) in Figure 3.5, and some zooms on interesting sub-portions in Figure 3.6. The first observation is that the adjacency matrix can obviously be decomposed into two terms, , where is a block diagonal matrix, and where is a (very) sparse matrix.

Sites (and sub-sites) indeed appear as squares that coincide with the nodes of the URL tree. With a little practice, one can even guess the deep structure of sites from the appearance of the corresponding square. For example, pages with high out-degree (typically, the site map) result in horizontal lines, while those with high in-degree (the home pages) are characterized by vertical lines. “Noisy” squares, that is, with points exhibiting a pseudo-random structure (cf. Figure 3.6c), are often the sign of interactive documentation pages (dictionaries, for example). Let us finally note that the structure can exhibit a recursive character, as shown by the block in Figure 3.6d.

3.3.4 Proposed algorithm for partitioning into sites

We will attempt, using only the knowledge of the graph and URLs of a crawl, to generate as simply as possible a partition of the graph that reflects the notion of site. Our approach thus differs from that employed by [Li+00], who also uses a set of semantic rules to define sites.

Formal model

Structurally, we believe that a site is defined as a set of pages tightly connected to one another by navigational links, as the adjacency matrix representations seem to confirm (cf. Figure 3.5).

This leads us to try to define a function that measures the quality of a site partition relative to the rate of navigational links. This function, must reach an extremum when approaches (in a sense to be specified) what we would like to call a site partition. Once such a function is found, the standard method for partitioning consists of constructing an initial partition of the URLs (adapted to the situation considered), then performing local adjustments to try to optimize .

Choice of the function

Since the most important factor in our view for estimating a site partition is the number of navigational links, it seems logical to seek a function that depends explicitly on , the number of internal links. One must also take into account the number of sites: indeed, this number is not fixed in advance (otherwise, we would be in the classical framework of mincut or maxcut type algorithms). To avoid having the trivial partition as the extremum, and to try to obtain a fine decomposition, it also seems important to us to try to maximize the number of sites . We thus have two global parameters, the number of sites and the number of internal links , and the function we seek must satisfy both and . Several simple formulas are a priori possible:

Problem of isolated pages

The case of pages with in-degree 1 and out-degree zero poses a problem. While they may sometimes represent an isolated site reduced to a single page (Figure 3.7), one observes experimentally that in most cases these are pages that have been crawled but not visited (see Section 2.1) or terminal pages of structured sites (Figure 3.8). From a structural point of view, and in the absence of any other consideration, it seems legitimate to us to ensure that pages with in-degree 1 and out-degree zero are attached to the site of the parent page. This translates, at the level of the function introduced previously, into the following inequality:

This inequality, if one tries to impose it on our function (which is necessary if one wants to suffice for performing the partition), raises the following problem: any partition with parameter such that

is less effective than the trivial partition:

Alas, Equation (3.2) is satisfied as soon as the graph is connected. More generally, for a graph composed of connected components and edges, for any partition with parameter we will necessarily have the inequality

Summary: There exists no universal function that is maximal for a non-trivial site partition respecting terminal pages.

Two approaches are then possible to circumvent this problem, as well as others of the same kind that one might encounter when trying to refine our model:

Alteration of the variables and

A first idea would consist of weighting the edges so that links to isolated pages are harder to break. Thus, a fairly classical and natural idea is to take a weight inversely proportional to the in-degree (cf. [Kle98]). Alas, the new variable thus introduced has numerous drawbacks, especially if it is the only alteration introduced:

First, one may consider the case of two sites (in the intuitive sense) connected by a single edge (Figure 3.9). If the algorithm is designed not to separate terminal pages, that is, if it respects the relation Equation (3.1), it will be forced to merge the two sites, which is not necessarily the desired effect.

Furthermore, if one considers the two examples of 3-connected sites in Figure 3.10, one would like the decision to merge sites, all other things being equal, to be the same in both cases. With edge weights depending on the in-degree, this will be difficult, since the connection strength varies by a factor of three between the two cases.

An alteration by weighting partitions according to their size is another solution. More generally, if one considers a partition and a weighting function , one can define the new variable . The simplest solution, which we will adopt here, consists of assigning zero weight to singleton partitions (which amounts in practice to working on while imposing the rule that singleton partitions are forbidden). The choice of other functions will not be treated here, but one may hope, by choosing the right function , for some control over the partition distribution, for example avoiding the formation of overly large sites by weakening for large values.

Chosen function

Taking into account all the remarks we have just made, we choose as our evaluation function the function which, to a partition of , associates , where is the number of internal links and , with .

Estimating a site partition using the URL decomposition tree

Now that we possess an evaluation function for a partition, the question arises of producing an efficient partition. An intuitive method is based on the following hypothesis, supported by the observation of Figure 3.5 and Figure 3.6: The URL architecture of pages reflects the architecture of sites, and therefore blocks are a priori subtrees or unions of subtrees of the Tree-Graph tree. Instead of having to choose a partition from among all possible ones, we can restrict the set of candidates to those coinciding with the tree structure.

Figure 3.11: The five-on-a-die pattern: a block is interleaved within another

This solution is not perfect, and there will always be borderline cases. Thus, in Russian-doll configurations, the choice of partition level will be closely tied to the choice of functions and . On the other hand, five-on-a-die configurations (a site embedded within another at the adjacency matrix level, see Figure 3.11), which are difficult to disentangle automatically using only the adjacency matrix (which nevertheless carries more information than the graph alone), will not cause problems with the decomposition tree provided that the sites coincide with the tree (which is the case for all examples tested).

We will give an example of a simple algorithm for performing an intelligent and fast partition of URLs using the tree-graph: the Filtered Breadth-First Search (Algorithm 3.1).

Definition 3.1.

The lexicographic cone of a URL is the cluster associated, in the URLs decomposition tree, with the internal node that is the parent of the leaf corresponding to the URL in question.

The advantages of this initial decomposition are as follows:

There remain, unfortunately, cases that the algorithm we propose does not handle. For instance, certain sites straddle several servers that have no kinship in the lexicographic tree as we have defined it (personal sites shared between several hosting providers, commercial sites available in .com, .fr, .de…).

Results

We will compare here the partition obtained by Filtered Breadth-First Search with partitions obtained by quotienting by server, by directory at depth 1, and by directory at depth 2 (which typically corresponds to cuts at heights 3, 4, and 5 in the tree-graph). The study was conducted on the crawl of 8,212,791 URLs from .fr dated June 15, 2001. For brevity, we will call these partitions FBFS, server, 1-dir, and 2-dir.

Figure 3.12: Distribution of partitions as a function of their size, on a log-log scale. Legend: FBFS in red, server in blue, 1-dir in green, 2-dir in yellow

Let us first observe how the different partitions are distributed. Figure 3.12 shows the size distributions of the different partitions. One observes that regardless of the method used (FBFS, server, 1-dir, or 2-dir), the partitions obtained are distributed according to a power law15, that is, the number of blocks of size is of order . The values of the exponents (estimated by linear regression) are indicated in table Table 3.1.

Partition server 1-dir 2-dir FBFS
1.32 1.75 1.89 1.59
Table 3.1: Power law coefficients of the different partitions

The power law is a law considered characteristic of social networks and human activities. The fact of encountering such a law in all cases studied, while it proves nothing by itself, shows that from the sole point of view of page distribution (without taking hyperlinks into account), the four distributions studied appear to be compatible with the human structure of the Web.

Let us also note the importance of “sites” of size 1 (the data in Figure 3.12 come from the partitions before any optimization), especially for the FBFS partition. The explanation is quite logical: size-1 partitions contain, among other things, all the errors that were not removed from the crawl: 4xx errors, errors in the URL format… The FBFS partition isolates more errors than the others by construction, since a URL that is isolated in the lexicographic cone of a site will be isolated by FBFS, unlike the other partitions.

server 39305 27241 0.9517 16629
FBFS 289539 58094 0.9218 24624
1-dir 182942 137809 0.7851 10831
2-dir 250869 189698 0.7012 5025
Table 3.2: Characteristics of the different partitions before optimization

Finally, let us compare the quality of the site decomposition of the different partitions, using the site index defined by the function . The optimization is initially minimal: size-1 sites are merged with the site to which they are most strongly connected (most often, the connection is unique).

The results are reported in table Table 3.2. One will note that FBFS, whose results for as well as for are midway between server and 1-dir, achieves by far the best site index.

Optimization of partitions

The preceding results come from partitions that do not a priori seek to maximize the site index. One can settle for these results, if one views the site index merely as a quality indicator, but one can also seek to maximize it. We will therefore examine what a local optimization by mixing yields, without FBFS and with FBFS. The principle of this optimization is simple: for each of the sites defined by the partition to be optimized, one attempts to improve the site index by replacing it with the local decomposition (server, 1-dir,…) that yields the best site index. It is therefore a greedy algorithm, and the result obtained, which is a local maximum, may depend on the order in which sites are processed.

optimization without FBFS, pass 1 107884 0.9376 52341
optimization without FBFS, pass 2 113148 0.9340 52487
optimization with FBFS, pass 1 121963 0.9338 56165
optimization with FBFS, pass 2 118386 0.9372 56845
Table 3.3: Characteristics of the different partitions after optimization

The results are presented in table Table 3.3. Two different traversal orders of the partitions were performed, in order to assess the fluctuations that the choice of traversal can generate, as well as their magnitude. The main results that can be drawn from this table are:


  1. 1Anchors allow pointing to a specific part of a Web page.
  2. 2For example, it is claimed in [LM04] and [Kam+03b] that Arasu et al. use the butterfly model to speed up their PageRank computation [Ara+01]. Upon verification, Arasu et al. do cite the butterfly model, but propose a PageRank computation method based on… writing in block triangular form derived from the strongly connected components decomposition (cf. Theorem 4.6). In particular, the only contribution of the butterfly model is the existence of one diagonal block larger than the others, the Strongly Connected Component (SCC).
  3. 3″This suggests that our results are relatively insensitive to the particular crawl we use, provided it is large enough.“, op. cit.
  4. 4This collapse phenomenon is in fact what tends to happen today with directories.
  5. 5”The structure that is now unfolding tells us that it is relatively insensitive to the particular large crawl we use. For instance, if AltaVista’s crawler fails to include some links whose inclusion would add one of the tendrils to the SCC, we know that the resulting change in the sizes of SCC and TENDRIL will be small (since any individual tendril is small). Likewise, our experiments in which we found that large components survived the deletion of nodes of large in-degree show that the connectivity of the web is resilient to the removal of significant portions“, op. cit.
  6. 6It appears that the starting set for AltaVista’s crawls at the time consisted of a few hundred pages. A bookmarks page can easily contain a few hundred hyperlinks.
  7. 7”Beyond a certain depth, only a few paths are being explored, and the last path is much longer than any of the others.“, op. cit.
  8. 8Port 80 is the standard port for the HTTP protocol. Other ports are used more or less commonly, such as port 443 for secure HTTP or port 8080 for a secondary server, but we will not consider them here.
  9. 9There are in fact fewer if one removes certain reserved addresses – military addresses for example – which are not supposed to host an accessible Web server. The Internet Assigned Numbers Authority (IANA) is the organization responsible for managing the available and unavailable address ranges. One can thus eliminate approximately 48% of the available addresses.
  10. 10Thanks to Fabien de Montgolfier for reminding me of the term decomposition tree.
  11. 11One may think, for example, of modular decomposition trees.
  12. 12For simplicity, we will restrict ourselves to servers identified by their DNS domain name and using port 80.
  13. 13Top Level Domain (TLD), or top-level domain. TLDs are divided into two categories: generic top-level domains (gTLD) and country code top-level domains (ccTLD). TLDs are managed by IANA (http://www.iana.org).
  14. 15By abuse of language, we will often call a power law any phenomenon that, when represented on a log-log scale, yields a curve forming approximately a straight line. The fact that the number of samples considered is finite, and the sensitivity of estimates to extreme values, means that it would be more accurate to merely speak of a heavy tail phenomenon.
Esc