Providing satisfactory web performance in mobile networks is challenging due to the long latency and low bandwidth in these networks. In this paper, we develop ACME (Architecture for Content delivery in the Mobile Environment), which efficiently implements edge caching functionality in terminals by exploiting user interest correlation. By selectively pushing content to users most likely to request it in the future, ACME not only increases hit ratio, but also reduces terminal power consumption by 17 to 144 times in trace-driven simulations, compared to pushing content to all users. ACME also exhibits superior scalability, as the push group size grows only 62% when the number of users increases 16 times in different traces.
Content delivery, mobile web access, web browsing, edge caching, user interest correlation.
Satisfactory mobile web performance is essential to the success of emerging 2.5G and 3G mobile networks as mobile operators rely on the mobile web and other data applications to generate new revenues. However, ensuring adequate mobile web performance is uniquely challenging. In particular, link layer retransmissions, usually necessary in mobile networks to compensate for the error-prone air interface [1], lead to long and unpredictable latency. For example, it was shown that the PING round trip time from a terminal to a web server increased by more than 20 times when a wireline link was replaced by an emulated heavily loaded GPRS link [1]. Such long latencies are particularly detrimental to web interactivity and user experience, which typically deteriorates if the web page download time takes more than eight seconds. Increasingly rich web content and low bandwidth in mobile networks make this problem even worse.
In this paper, we develop an architecture called ACME (Architecture for Content delivery in the Mobile Environment) that significantly enhances mobile web performance. This architecture, shown in Figure 1, is novel in two aspects. First, unlike traditional edge caching, which serves content from network edge, ACME implements edge caching functionality in the ACME cache at the terminal. This approach minimizes the use of the bottleneck (the radio link) and substantially reduces user perceived latency. Since edge caching is a superset of client caching and stores popular content requested by all users, it is necessary to push non-requested content to terminals. Specifically, implementing full-fledged edge caches in terminals requires pushing each requested web object to all terminals, as done in [3]. For mobile terminals, however, this approach would consume excessive terminal power and storage resources. Toward this end, we develop the concept of user interest correlation. User interest correlation refers to the common interest in the same content among different users, and it is the foundation of edge caching effectiveness. It has also been studied in the context of peer-to-peer content discovery [4]. The ACME Director shown in Figure 1 computes user interest correlation based on previous user access patterns, and selectively pushes the web object being requested by one user to other users who are most likely to request it in the future. We describe the ACME Director's algorithm and evaluate its performance using trace-driven simulations in Section 2.
We use conditional probability to measure user interest correlation. For users, let be the -by- matrix such that , where denotes the probability that user will access a web object in the future, given user has accessed that object. is defined to be zero because this case is handled by conventional client caching and is irrelevant to ACME. In practice, can be trained using past web access data, as discussed in Section 2.2.
Once the ACME Director has a properly trained matrix, it performs selective push as follows. First, we use a tunable parameter called the ``selectiveness factor'' ( ) to control push ``selectiveness''. Suppose user requests a web object, and is among the largest elements in , then the ACME Director pushes the content to user . Obviously, represents the case of conventional client caching, and represents the case of full edge caching, i.e., every terminal receives a copy of the requested content. The design goal of the ACME Director is to achieve high hit ratio (close to that of full edge caching) with a small to reduce terminal power consumption.
Due to the lack of publicly available mobile web traces, we use three wireline proxy traces [2] for simulations and exclude all dynamically generated and non-cacheable responses. The UCB trace is the first 300,000 accesses in UC Berkeley's home IP access, representing about 17 hours of activity from 2155 IP addresses. The BU trace is the concatenated access log of Room B19 in Boston University's computer science department in March and April 1995, with a total of 553 users and 437,861 web accesses. The NLANR trace is the access log of the pb node of the National Laboratory for Advanced Network Research on November 13, 2000, with a total number of 127 user IP addresses and 878,085 web accesses.
To quantitatively measure the ACME Director's performance, we note that its hit ratio is upper-bounded by full edge caching and lower-bounded by client caching. Its hit ratio is also dependent on . So we define the Director Effectiveness, , as
We define the push group size as the average number of terminals that receive a particular pushed web object for selectiveness factor . Because average terminal power consumption is proportional to , we use to represent average terminal power consumption. To compare the group size between different traces, we define relative push group size, , as
Now we run simulations driven by the traces described above to determine , and . We partition each trace into a training part (odd-numbered accesses) and a test part (even-numbered accesses). We count the number of requests that user makes after user in the training part and obtain , for , and normalize the matrix. We then run the test part of each trace in a ACME Director simulator and measure the hit ratios under different selectiveness factors. is then computed based on , and . We obtain by counting the number of terminals receiving pushed object, for all requested objects, and compute by normalization.
Figure 2 depicts Director Effectiveness against relative push group size by varying from 0 to 1 for all three traces. ACME's effectiveness is evident from the figure, as increases rapidly with very small group size. In all three traces, a Director Effectiveness of 80% is achieved at a relative push group size smaller than 20%. Furthermore, ACME Director achieves 50% Effectiveness when the relative push group size is as small as 0.7% (UCB trace) to 6% (NLANR trace) of the full edge caching push group size. The push group size ranges from 6.74 (NLANR trace) to 10.96 (UCB trace). The reduction in average terminal power consumption is calculated as the inverse of the relative push group size, and the ACME Director reduces the terminal power consumption by 17 (NLANR trace) to 144 times (UCB trace) while maintaining 50% Effectiveness.
Another interesting finding from our simulations is that for a wide range of Director Effectiveness values, the relative push group size decreases with the increasing number of users among traces. This indicates that the ACME push group size is much less sensitive than the full edge caching push group size, to the vastly different number of users in the three traces. Indeed, at 50% Effectiveness, the push group size increases only 62%, from 6.74 to 10.96, as increases 16 times from 127 in the NLANR trace to 2155 in the UCB trace. We conjecture that this is because for a given user, her interest in content can be best represented by a small group of users whose interests are closest to hers. The size of this interest group grows much slower than the general user group because it only keeps the users with the closest interest in the general user group, and identifying this group can help to create scalable and efficient content multicast systems.
ACME is a versatile architecture and can be implemented in a variety of networks. It can utilize multicast (such as MBMS in 3GPP) for efficient delivery, if such capability exists. In this case, ACME could use transport protocols similar to those studied in [3]. In other networks, ACME can utilize existing push mechanisms to implement edge caching. The split proxy architecture of ACME makes it possible to implement optimal transport protocols between the ACME Director and ACME caches, and provides full transparency to web applications at the origin server and the clients.
The authors would like to thank Gang Xu, Dimitris Koulakiotis, Duane Wessels and Antti Toskala for valuable information and comments.