redis使用pipeline批量插入hash数据

import org.apache.commons.lang.math.RandomUtils;

import redis.clients.jedis.Jedis;
import redis.clients.jedis.Pipeline;

public class RedisEasyTest {

	private static Jedis jedis = new Jedis("xx.xx.xx.xx");

	private static Pipeline p = jedis.pipelined();

	private static int KEY_COUNT = 10000;

	private static int FIELD_COUNT = 10;

	public void single() {
		for (int i = 0; i < KEY_COUNT; i++) {
			String key = RandomUtils.nextInt(5) + "";
			for (int j = 0; j < FIELD_COUNT; j++) {
				jedis.hset(key, j + "", i + j + "");
				jedis.expire(key, 3600);
			}
		}
	}

	public void batch() {
		int index = 0;
		for (int i = 0; i < KEY_COUNT; i++) {
			String key = RandomUtils.nextInt(5) + "";
			for (int j = 0; j < FIELD_COUNT; j++) {
				p.hset(key, j + "", i + j + "");
				p.expire(key, 3600);
			}
			if (++index % 1000 == 0) {
				p.sync();
			}
		}
		p.sync();
	}

	public static void main(String[] args) {
		long start = System.currentTimeMillis();
		RedisEasyTest r = new RedisEasyTest();
		r.single();
		System.out.printf("single use %d sec \n", (System.currentTimeMillis() - start) / 1000);
		start = System.currentTimeMillis();
		r.batch();
		System.out.printf("batch use %d sec \n", (System.currentTimeMillis() - start) / 1000);

	}
}

输出结果:

single use 30 sec
batch use 0 sec

 

可以看到通过pipeline批量插入数据性能是非常不错的。

 

 

--end

 

 

 

你可能感兴趣的:(pipeline)