Warmming up:
Analysing the solr src class, which packaged in org.apache.solr.search. There are two implementations of cache available for Solr, LRUCache, based on a synchronized LinkedHashMap, and FastLRUCache, based on a ConcurrentHashMap. FastLRUCache has faster gets and slower puts in single threaded operation and thus is generally faster than LRUCache when the hit ratio of the cache is high (> 75%). And then i will delve into the whloe solr's cache, let's start from FastLRUCache FIFO strategy.
FastLRUCache FIFO strategy:
FIFO means first come first in, which is a very common strategy in software design, i will give a experiment directly to account for FIFO strategy.
In order to make the experiment result more persuasively, we should add a row code in Solr's org.apache.solr.common.util.ConcurrentLRUCache<K,V>, we add the below code
//add by kylin System.out.println(key + " " + map.keySet());
to ConcurrentLRUCache's put(K key, V val), yes, it's just a output sentence, use to output the cache's size before a put has coming.
and then we run the below code;
public void testFastLRUCacheFIFO() { FastLRUCache cache = new FastLRUCache(); Map map = new HashMap(); map.put("size", "3"); CacheRegenerator regenerator = new CacheRegenerator() { public boolean regenerateItem(SolrIndexSearcher newSearcher, SolrCache newCache, SolrCache oldCache, Object oldKey, Object oldVal) throws IOException { newCache.put(oldKey, oldVal); return true; }}; Object obj = cache.init(map, null, regenerator); cache.setState(SolrCache.State.LIVE); for (int i = 1; i < 10; i++) { cache.put(i , "" + i); } }
map.put("size", "3"), this is very imporant, is the key factor of FastLRUCache FIFO, implement the code, and the result is:
1 [1] 2 [2, 1] 3 [2, 1, 3] 4 [3, 4] 5 [5, 3, 4] 6 [6, 5] 7 [6, 5, 7] 8 [7, 8] 9 [7, 8, 9]
Through the result data, we can know that the latest record was saved, but this not totally FIFO, because the Out number is count by the size you have set: Out numer = size * 10%, but the minimal Out size is 2. so we can count the above Out number(3 * 10% = 0, 0 < 2, so Out num is 2).
change the above code, we look at the warming mechenism.
public void test() throws IOException { FastLRUCache cache = new FastLRUCache(); Map map = new HashMap(); map.put("size", "100"); map.put("initialSize", "10"); map.put("autowarmCount", "30"); CacheRegenerator regenerator = new CacheRegenerator() { public boolean regenerateItem(SolrIndexSearcher newSearcher, SolrCache newCache, SolrCache oldCache, Object oldKey, Object oldVal) throws IOException { newCache.put(oldKey, oldVal); return true; }}; Object obj = cache.init(map, null, regenerator); cache.setState(SolrCache.State.LIVE); for (int i = 1; i < 102; i++) { cache.put(i , "" + i); } System.out.println(cache.get(10)); System.out.println(cache.get(11)); FastLRUCache cacheNew = new FastLRUCache(); cacheNew.init(map, obj, regenerator); cacheNew.warm(null, cache); cacheNew.setState(SolrCache.State.LIVE); cache.close(); cacheNew.put(103, "103"); System.out.println(cacheNew.get(72)); System.out.println(cacheNew.get(73)); }
and also i give the implement result:
null 11 null 73
Anaysing the result:
The size = 100, so 100 * 10% = 10, OutNumber = 10, when the 101th recod has added, the first came 10 recod were remove, so the null printed, and then 11 was printed.
The autowarmCount = 30, so when a new SolrSearcher coming, it will be bould with new Searcher, 102-30=72, So when you get 72 in new Searcher, the null is printed,and if you get 73, the responding value is printed.