LRU is one of the most widely used cache eviction algorithms that span its utility across multiple database systems. Although popular, it suffers from a bunch of limitations especially when it is used for managing caches in disk-backed databases like MySQL and Postgres.
In this essay, we take a detailed look into the sub-optimality of LRU and how one of its variants called 2Q addresses and improves upon it. 2Q algorithm was first introduced in the paper - 2Q: A low overhead high-performance buffer management replacement algorithm by Theodore Johnson and Dennis Shasha.
The LRU eviction algorithm evicts the page from the buffer which has not been accessed for the longest. LRU is typically implemented using a Doubly Linked List and a Hash Table. The intuition of this algorithm is so strong and implementation is so simple that until the early ’80s, LRU was the algorithm of choice in nearly all the systems. But as stated above, there are certain situations where LRU performs sub-optimal.
Sub-optimality during DB scans
If the database table is bigger than the LRU cache, the DB process, upon scanning the table will wipe out the entire LRU cache and fill it with the pages from just one scanned table. If these pages are not referenced again, this is a total loss and the performance of the database takes a massive hit. The performance will pickup once these pages are evicted from the cache and other pages make an entry.
Sub-optimality in evictions
LRU algorithm works with a single dimension - recency - as it removes the pages from the buffer on the basis of recent accesses. Since it does not really consider any other factor, it can actually evict a warmer page and replace it with a colder one - a page that could and would be accessed just once.
2Q addresses the above-illustrated issues by introducing parallel buffers and supporting queues. Instead of considering just recency as a factor, 2Q also considers access frequency while making the decision to ensure the page that is really warm gets a place in the LRU cache. It admits only hot pages to the main buffer and tests every page for a second reference.
The golden rule that 2Q is based on is - Just because a page is accessed once does not entitle it to stay in the buffer. Instead, it should be decided if it is accessed again then only keep it in the buffer.
Below we take a detailed look into two versions of the 2Q algorithm - simplified and improved.
Simplified 2Q algorithm works with two buffers: the primary LRU buffer -
Am and a secondary FIFO buffer -
A1. New faulted pages first go to the secondary buffer
A1 and then when the page is referenced again, it moves to the primary LRU buffer
Am. This ensures that the page that moves to the primary LRU buffer is hot and indeed requires to be cached.
If the page residing in
A1 is never referenced again, it eventually gets discarded, implying the page was indeed cold and did not deserve to be cached. Thus this simplified 2Q provides protection against the two listed sub-optimality of the simple LRU scheme by adding a secondary buffer and testing pages for a second reference. The pseudocode for the Simplified 2Q algorithm is as follows:
def access_page(X: page): # if the page already exists in the LRU cache # in buffer Am if X in Am: Am.move_front(X) # if the page exists in secondary storage # and not it gets access. # since the page is accessed again, indicating interest # and long-term need, move it to Am. elif X in A1: A1.remove(X) Am.add_front(X) # page X is accessed for the first time else: # if A1 is full then free a slot. if A1.is_full(): A1.pop() # add X to the front of the FIFO A1 queue A1.add_front(X)
Tuning Simplified 2Q buffer is difficult - if the maximum size of
A1 is too small, the test for hotness becomes too strong and if it is too large then due to memory constraint
Am will get relatively smaller memory making the primary LRU cache smaller, eventually degrading the database performance.
The full version 2Q algorithm remediates this limitation and eliminates tuning to a massive extent without taking any hit in performance.
2Q Full Version
Although Simplified 2Q algorithm does a decent job there is still scope of improvement when it comes to handling common database access pattern, that suggests, a page generally receives a lot of references for a short period of time and then no reference for a long time. If a page truly needs to be cached then after it receives a lot (not just one) of references in a short span it continues to receive references and hits on regular intervals.
To handle this common database access pattern, the 2Q algorithm splits the secondary buffer
A1 into two buffers
A1-Out, where the new element always enters
A1-In and continues to stay in
A1-In till it gets accesses ensuring that the most recent first accesses happen in the memory.
Once the page gets old, it gets thrown off the memory but its disk reference is stored in the
A1-Out buffer. If the page, whose reference is, residing in
A1-Out is accessed again the page is promoted to
Am LRU implying it indeed is a hot page that will be accessed again and hence required to be cached.
Am buffer continues to be the usual LRU which means when any page residing in
Am is accessed it is moved to the head and when a page is needed to be discarded the eviction happens from the tail end.
2Q in Postgres
Postgres uses 2Q as its cache management algorithm due to patent issues with IBM. Postgres used to have ARC as its caching algorithm but with IBM getting a patent over it, Postgres moved to 2Q. Postgres also claims that the performance of 2Q is similar to ARC.