Recently, I faced a problem which memory usage on ElastiCache for Redis becomes huge.
This post describes how we can investigate such a problem (which keys are bottlenecks on ElastiCache).
Export and download a Redis backup
At first, to analyze, we have to download the .rdb file from our ElastiCache following the official guide.
When exporting an ElastiCache backup, please be careful that the region for S3 must be the same as the one for the backup.
  
  
  Analyze .rdb with redis-rdb-tools
After downloading the file, we can analyze it with redis-rdb-tools.
For example, if we wanna generate a memory report, we can do it as follows.
> rdb -c memory /var/redis/6379/dump.rdb --bytes 128 -f memory.csv
> cat memory.csv
database,type,key,size_in_bytes,encoding,num_elements,len_largest_element
0,list,lizards,241,quicklist,5,19
If your .rdb is too large (it may be difficult to open the csv file), you could also generate a sampled report like:
> rdb -c memory /var/redis/6379/dump.rdb | ruby -ne 'print $_ if rand < 0.1' > memory.csv
Of course, redis-rdb-tools supports features other than generating a memory report.
For more information, please see the README.
Summary
- We can export and download a backup file easily with ElastiCache and S3.
 - We can analyze a 
.rdbfile easily with a tool likeredis-rdb-tools. 
    
Top comments (0)