详情:http://aspalliance.com/251
为方便阅读,把全文抄过来了:
ASP.NET Micro Caching: Benefits of a One-Second Cache
Abstract
When I start talking about caching, I imagine that many people immediately stop listening, thinking "my situation requires up-to-the-minute data, so caching isn't an option". Consider the benefits of what I call 'micro caching', in which data is cached for a very small amount of time. In high volume applications, the benefits can be substantial.
Introduction | |||||||||
I fell in love with caching when I first started playing with it in the early versions of the ASP.NET alphas. I was aware of caching before that, but I'd never really gone to the effort to seriously implement a cache engine in an ASP app, and with ASP.NET, I didn't have to. The thing that attracts me to caching is its impact on performance without requirements (usually) for major re-architecture of the system. In this article, I'm going to discuss what I call Micro Caching, because it involves caching things for very brief periods. One of the major downsides of caching in general is that by definition the data is not fresh. How big a problem this is depends on the application and the user's requirements, which is why it is important to determine the user's tolerance for stale data when gathering requirements. The knee-jerk answer is going to be 'I need live data all the time', but hopefully with the data in this article, you will be able to convince the user/client that perhaps it would be ok if the data were, say, a second or two old, if it meant they could host the application on half as much hardware.
|
Results Without Caching | |
Running the test without any caching on the web server resulted in 16,961 requests and an average requests per second of 56.54. The Time to First Byte (TTFB) and Time to Last Byte (TTLB), which are useful measures of the site's performance from an individual user's perspective, were both 52ms--instant as far as a user is concerned. For most web applications, anything under one second is an acceptable TTLB. Below is a graph of requests per second over time for the page without caching. Figure 1 |
Test Page With Caching | |
For the second part of the test, I modified the test page by adding one line of code to Default.aspx: <%@ OutputCache Duration="1" VaryByParam="none" %> The new Default.aspx then looked like this: <%@ Page language="c#" Codebehind="Default.aspx.cs" AutoEventWireup="false" Inherits="MicroCache._Default" EnableSessionState="false" EnableViewState="false" %> The same test was run against this page. |
Results With Caching | |
Running the test a second time against a page with one second of output caching enabled resulted in a total of 74,157 requests and an average requests per second of 247.19. The TTFB and TTLB were both 11ms. So, compared to the non-cached page, the throughput increased by over 400% and each page took 80% less time to load, five times the performance. Figure 2 below shows a graph of requests per second over time. Figure 2 |
The next figure shows both tests, side-by-side.
Figure 3
Analysis and Summary | |
Clearly, even just a second's worth of caching can have a big impact on performance for a high-volume application. It's easy to see why this is the case if we look at the database's performance during this test. I ran the test with two performance counters: SQLServer:SQL Statistics\Batch Requests/sec In the non-caching test, the SQL Server requests/sec counter averaged 56.2 requests per second. In the micro-caching test, it averaged 1.0 request per second. Obviously, the database was not having to do nearly as much work. However, not all of the savings were on the database side. Loading data into a DataSet is relatively expensive as well, as is data binding. Although there aren't any performance counters for these activities specifically, we know they were occurring once per request in the first case, and once per second in the latter case, so we dropped a lot of work from the ASP.NET engine's plate as well. Typically, the database and the web server will reside on separate boxes. The reduction in database load would clearly benefit the database server, which is a very good thing since, historically, database servers are much more difficult to scale out than web servers. In addition, it will reduce network traffic between the web and database servers, which will further enhance performance, and the reduction in web server load is also helpful, of course, since it reduces how many web servers would be needed (or how powerful each one would need to be). Summary Caching is a very useful tool to take advantage of when designing applications. It is not a panacea, and like any tool, it can be used improperly. But even in situations where data needs to be up to the second, adding one second's worth of caching can make a big impact on application performance. Remember to ask your users how much they can tolerate out-of-date data and try to get them to at least sign off on one-second-old data if you can, since that will allow you to take advantage of techniques like the one described here. |