2011-07-28 06:32:52 -04:00
|
|
|
start_server {tags {"maxmemory"}} {
|
2014-07-18 04:55:08 -04:00
|
|
|
test "Without maxmemory small integers are shared" {
|
|
|
|
r config set maxmemory 0
|
|
|
|
r set a 1
|
|
|
|
assert {[r object refcount a] > 1}
|
|
|
|
}
|
|
|
|
|
|
|
|
test "With maxmemory and non-LRU policy integers are still shared" {
|
|
|
|
r config set maxmemory 1073741824
|
|
|
|
r config set maxmemory-policy allkeys-random
|
|
|
|
r set a 1
|
|
|
|
assert {[r object refcount a] > 1}
|
|
|
|
}
|
|
|
|
|
|
|
|
test "With maxmemory and LRU policy integers are not shared" {
|
|
|
|
r config set maxmemory 1073741824
|
|
|
|
r config set maxmemory-policy allkeys-lru
|
|
|
|
r set a 1
|
|
|
|
r config set maxmemory-policy volatile-lru
|
|
|
|
r set b 1
|
|
|
|
assert {[r object refcount a] == 1}
|
|
|
|
assert {[r object refcount b] == 1}
|
|
|
|
r config set maxmemory 0
|
|
|
|
}
|
|
|
|
|
2011-07-28 06:32:52 -04:00
|
|
|
foreach policy {
|
2017-03-15 04:05:15 -04:00
|
|
|
allkeys-random allkeys-lru allkeys-lfu volatile-lru volatile-lfu volatile-random volatile-ttl
|
2011-07-28 06:32:52 -04:00
|
|
|
} {
|
|
|
|
test "maxmemory - is the memory limit honoured? (policy $policy)" {
|
|
|
|
# make sure to start with a blank instance
|
2014-07-31 14:39:49 -04:00
|
|
|
r flushall
|
2011-07-28 06:32:52 -04:00
|
|
|
# Get the current memory limit and calculate a new limit.
|
|
|
|
# We just add 100k to the current memory size so that it is
|
|
|
|
# fast for us to reach that limit.
|
|
|
|
set used [s used_memory]
|
|
|
|
set limit [expr {$used+100*1024}]
|
|
|
|
r config set maxmemory $limit
|
|
|
|
r config set maxmemory-policy $policy
|
|
|
|
# Now add keys until the limit is almost reached.
|
|
|
|
set numkeys 0
|
|
|
|
while 1 {
|
|
|
|
r setex [randomKey] 10000 x
|
|
|
|
incr numkeys
|
|
|
|
if {[s used_memory]+4096 > $limit} {
|
|
|
|
assert {$numkeys > 10}
|
|
|
|
break
|
|
|
|
}
|
|
|
|
}
|
|
|
|
# If we add the same number of keys already added again, we
|
|
|
|
# should still be under the limit.
|
|
|
|
for {set j 0} {$j < $numkeys} {incr j} {
|
|
|
|
r setex [randomKey] 10000 x
|
|
|
|
}
|
|
|
|
assert {[s used_memory] < ($limit+4096)}
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
foreach policy {
|
|
|
|
allkeys-random allkeys-lru volatile-lru volatile-random volatile-ttl
|
|
|
|
} {
|
|
|
|
test "maxmemory - only allkeys-* should remove non-volatile keys ($policy)" {
|
|
|
|
# make sure to start with a blank instance
|
2014-07-31 14:39:49 -04:00
|
|
|
r flushall
|
2011-07-28 06:32:52 -04:00
|
|
|
# Get the current memory limit and calculate a new limit.
|
|
|
|
# We just add 100k to the current memory size so that it is
|
|
|
|
# fast for us to reach that limit.
|
|
|
|
set used [s used_memory]
|
|
|
|
set limit [expr {$used+100*1024}]
|
|
|
|
r config set maxmemory $limit
|
|
|
|
r config set maxmemory-policy $policy
|
|
|
|
# Now add keys until the limit is almost reached.
|
|
|
|
set numkeys 0
|
|
|
|
while 1 {
|
|
|
|
r set [randomKey] x
|
|
|
|
incr numkeys
|
|
|
|
if {[s used_memory]+4096 > $limit} {
|
|
|
|
assert {$numkeys > 10}
|
|
|
|
break
|
|
|
|
}
|
|
|
|
}
|
|
|
|
# If we add the same number of keys already added again and
|
|
|
|
# the policy is allkeys-* we should still be under the limit.
|
|
|
|
# Otherwise we should see an error reported by Redis.
|
|
|
|
set err 0
|
|
|
|
for {set j 0} {$j < $numkeys} {incr j} {
|
|
|
|
if {[catch {r set [randomKey] x} e]} {
|
|
|
|
if {[string match {*used memory*} $e]} {
|
|
|
|
set err 1
|
|
|
|
}
|
|
|
|
}
|
|
|
|
}
|
|
|
|
if {[string match allkeys-* $policy]} {
|
|
|
|
assert {[s used_memory] < ($limit+4096)}
|
|
|
|
} else {
|
|
|
|
assert {$err == 1}
|
|
|
|
}
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
foreach policy {
|
2017-03-15 04:05:15 -04:00
|
|
|
volatile-lru volatile-lfu volatile-random volatile-ttl
|
2011-07-28 06:32:52 -04:00
|
|
|
} {
|
|
|
|
test "maxmemory - policy $policy should only remove volatile keys." {
|
|
|
|
# make sure to start with a blank instance
|
2014-07-31 14:39:49 -04:00
|
|
|
r flushall
|
2011-07-28 06:32:52 -04:00
|
|
|
# Get the current memory limit and calculate a new limit.
|
|
|
|
# We just add 100k to the current memory size so that it is
|
|
|
|
# fast for us to reach that limit.
|
|
|
|
set used [s used_memory]
|
|
|
|
set limit [expr {$used+100*1024}]
|
|
|
|
r config set maxmemory $limit
|
|
|
|
r config set maxmemory-policy $policy
|
|
|
|
# Now add keys until the limit is almost reached.
|
|
|
|
set numkeys 0
|
|
|
|
while 1 {
|
|
|
|
# Odd keys are volatile
|
|
|
|
# Even keys are non volatile
|
|
|
|
if {$numkeys % 2} {
|
|
|
|
r setex "key:$numkeys" 10000 x
|
|
|
|
} else {
|
|
|
|
r set "key:$numkeys" x
|
|
|
|
}
|
|
|
|
if {[s used_memory]+4096 > $limit} {
|
|
|
|
assert {$numkeys > 10}
|
|
|
|
break
|
|
|
|
}
|
|
|
|
incr numkeys
|
|
|
|
}
|
|
|
|
# Now we add the same number of volatile keys already added.
|
|
|
|
# We expect Redis to evict only volatile keys in order to make
|
|
|
|
# space.
|
|
|
|
set err 0
|
|
|
|
for {set j 0} {$j < $numkeys} {incr j} {
|
|
|
|
catch {r setex "foo:$j" 10000 x}
|
|
|
|
}
|
|
|
|
# We should still be under the limit.
|
|
|
|
assert {[s used_memory] < ($limit+4096)}
|
|
|
|
# However all our non volatile keys should be here.
|
|
|
|
for {set j 0} {$j < $numkeys} {incr j 2} {
|
|
|
|
assert {[r exists "key:$j"]}
|
|
|
|
}
|
|
|
|
}
|
|
|
|
}
|
|
|
|
}
|
2018-02-21 13:18:34 -05:00
|
|
|
|
2018-08-21 04:46:07 -04:00
|
|
|
proc test_slave_buffers {test_name cmd_count payload_len limit_memory pipeline} {
|
2018-02-21 13:18:34 -05:00
|
|
|
start_server {tags {"maxmemory"}} {
|
|
|
|
start_server {} {
|
2018-08-21 04:46:07 -04:00
|
|
|
set slave_pid [s process_id]
|
|
|
|
test "$test_name" {
|
2018-02-21 13:18:34 -05:00
|
|
|
set slave [srv 0 client]
|
|
|
|
set slave_host [srv 0 host]
|
|
|
|
set slave_port [srv 0 port]
|
|
|
|
set master [srv -1 client]
|
|
|
|
set master_host [srv -1 host]
|
|
|
|
set master_port [srv -1 port]
|
|
|
|
|
|
|
|
# add 100 keys of 100k (10MB total)
|
|
|
|
for {set j 0} {$j < 100} {incr j} {
|
|
|
|
$master setrange "key:$j" 100000 asdf
|
|
|
|
}
|
|
|
|
|
2018-08-21 04:46:07 -04:00
|
|
|
# make sure master doesn't disconnect slave because of timeout
|
2019-05-05 01:19:52 -04:00
|
|
|
$master config set repl-timeout 1200 ;# 20 minutes (for valgrind and slow machines)
|
2018-02-21 13:18:34 -05:00
|
|
|
$master config set maxmemory-policy allkeys-random
|
2018-09-11 04:57:50 -04:00
|
|
|
$master config set client-output-buffer-limit "replica 100000000 100000000 300"
|
2018-02-21 13:18:34 -05:00
|
|
|
$master config set repl-backlog-size [expr {10*1024}]
|
|
|
|
|
|
|
|
$slave slaveof $master_host $master_port
|
|
|
|
wait_for_condition 50 100 {
|
|
|
|
[s 0 master_link_status] eq {up}
|
|
|
|
} else {
|
|
|
|
fail "Replication not started."
|
|
|
|
}
|
|
|
|
|
|
|
|
# measure used memory after the slave connected and set maxmemory
|
|
|
|
set orig_used [s -1 used_memory]
|
|
|
|
set orig_client_buf [s -1 mem_clients_normal]
|
|
|
|
set orig_mem_not_counted_for_evict [s -1 mem_not_counted_for_evict]
|
|
|
|
set orig_used_no_repl [expr {$orig_used - $orig_mem_not_counted_for_evict}]
|
2021-03-08 13:53:53 -05:00
|
|
|
set limit [expr {$orig_used - $orig_mem_not_counted_for_evict + 32*1024}]
|
2018-02-21 13:18:34 -05:00
|
|
|
|
|
|
|
if {$limit_memory==1} {
|
|
|
|
$master config set maxmemory $limit
|
|
|
|
}
|
|
|
|
|
|
|
|
# put the slave to sleep
|
|
|
|
set rd_slave [redis_deferring_client]
|
2018-08-21 04:46:07 -04:00
|
|
|
exec kill -SIGSTOP $slave_pid
|
2018-02-21 13:18:34 -05:00
|
|
|
|
2018-08-21 04:46:07 -04:00
|
|
|
# send some 10mb worth of commands that don't increase the memory usage
|
2018-02-21 13:18:34 -05:00
|
|
|
if {$pipeline == 1} {
|
|
|
|
set rd_master [redis_deferring_client -1]
|
|
|
|
for {set k 0} {$k < $cmd_count} {incr k} {
|
|
|
|
$rd_master setrange key:0 0 [string repeat A $payload_len]
|
|
|
|
}
|
|
|
|
for {set k 0} {$k < $cmd_count} {incr k} {
|
|
|
|
#$rd_master read
|
|
|
|
}
|
|
|
|
} else {
|
|
|
|
for {set k 0} {$k < $cmd_count} {incr k} {
|
|
|
|
$master setrange key:0 0 [string repeat A $payload_len]
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
set new_used [s -1 used_memory]
|
|
|
|
set slave_buf [s -1 mem_clients_slaves]
|
|
|
|
set client_buf [s -1 mem_clients_normal]
|
|
|
|
set mem_not_counted_for_evict [s -1 mem_not_counted_for_evict]
|
|
|
|
set used_no_repl [expr {$new_used - $mem_not_counted_for_evict}]
|
|
|
|
set delta [expr {($used_no_repl - $client_buf) - ($orig_used_no_repl - $orig_client_buf)}]
|
|
|
|
|
|
|
|
assert {[$master dbsize] == 100}
|
|
|
|
assert {$slave_buf > 2*1024*1024} ;# some of the data may have been pushed to the OS buffers
|
2019-05-05 01:19:52 -04:00
|
|
|
set delta_max [expr {$cmd_count / 2}] ;# 1 byte unaccounted for, with 1M commands will consume some 1MB
|
|
|
|
assert {$delta < $delta_max && $delta > -$delta_max}
|
2018-02-21 13:18:34 -05:00
|
|
|
|
|
|
|
$master client kill type slave
|
|
|
|
set killed_used [s -1 used_memory]
|
|
|
|
set killed_slave_buf [s -1 mem_clients_slaves]
|
|
|
|
set killed_mem_not_counted_for_evict [s -1 mem_not_counted_for_evict]
|
|
|
|
set killed_used_no_repl [expr {$killed_used - $killed_mem_not_counted_for_evict}]
|
|
|
|
set delta_no_repl [expr {$killed_used_no_repl - $used_no_repl}]
|
|
|
|
assert {$killed_slave_buf == 0}
|
2019-05-05 01:19:52 -04:00
|
|
|
assert {$delta_no_repl > -$delta_max && $delta_no_repl < $delta_max}
|
2018-08-21 04:46:07 -04:00
|
|
|
|
|
|
|
}
|
|
|
|
# unfreeze slave process (after the 'test' succeeded or failed, but before we attempt to terminate the server
|
|
|
|
exec kill -SIGCONT $slave_pid
|
2018-02-21 13:18:34 -05:00
|
|
|
}
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
2018-08-21 04:46:07 -04:00
|
|
|
# test that slave buffer are counted correctly
|
|
|
|
# we wanna use many small commands, and we don't wanna wait long
|
|
|
|
# so we need to use a pipeline (redis_deferring_client)
|
|
|
|
# that may cause query buffer to fill and induce eviction, so we disable it
|
|
|
|
test_slave_buffers {slave buffer are counted correctly} 1000000 10 0 1
|
2018-02-21 13:18:34 -05:00
|
|
|
|
2018-08-21 04:46:07 -04:00
|
|
|
# test that slave buffer don't induce eviction
|
|
|
|
# test again with fewer (and bigger) commands without pipeline, but with eviction
|
2018-09-11 04:57:50 -04:00
|
|
|
test_slave_buffers "replica buffer don't induce eviction" 100000 100 1 0
|
2018-02-21 13:18:34 -05:00
|
|
|
|
Limit the main db and expires dictionaries to expand (#7954)
As we know, redis may reject user's requests or evict some keys if
used memory is over maxmemory. Dictionaries expanding may make
things worse, some big dictionaries, such as main db and expires dict,
may eat huge memory at once for allocating a new big hash table and be
far more than maxmemory after expanding.
There are related issues: #4213 #4583
More details, when expand dict in redis, we will allocate a new big
ht[1] that generally is double of ht[0], The size of ht[1] will be
very big if ht[0] already is big. For db dict, if we have more than
64 million keys, we need to cost 1GB for ht[1] when dict expands.
If the sum of used memory and new hash table of dict needed exceeds
maxmemory, we shouldn't allow the dict to expand. Because, if we
enable keys eviction, we still couldn't add much more keys after
eviction and rehashing, what's worse, redis will keep less keys when
redis only remains a little memory for storing new hash table instead
of users' data. Moreover users can't write data in redis if disable
keys eviction.
What this commit changed ?
Add a new member function expandAllowed for dict type, it provide a way
for caller to allow expand or not. We expose two parameters for this
function: more memory needed for expanding and dict current load factor,
users can implement a function to make a decision by them.
For main db dict and expires dict type, these dictionaries may be very
big and cost huge memory for expanding, so we implement a judgement
function: we can stop dict to expand provisionally if used memory will
be over maxmemory after dict expands, but to guarantee the performance
of redis, we still allow dict to expand if dict load factor exceeds the
safe load factor.
Add test cases to verify we don't allow main db to expand when left
memory is not enough, so that avoid keys eviction.
Other changes:
For new hash table size when expand. Before this commit, the size is
that double used of dict and later _dictNextPower. Actually we aim to
control a dict load factor between 0.5 and 1.0. Now we replace *2 with
+1, since the first check is that used >= size, the outcome of before
will usually be the same as _dictNextPower(used+1). The only case where
it'll differ is when dict_can_resize is false during fork, so that later
the _dictNextPower(used*2) will cause the dict to jump to *4 (i.e.
_dictNextPower(1025*2) will return 4096).
Fix rehash test cases due to changing algorithm of new hash table size
when expand.
2020-12-06 04:53:04 -05:00
|
|
|
start_server {tags {"maxmemory"}} {
|
|
|
|
test {Don't rehash if used memory exceeds maxmemory after rehash} {
|
|
|
|
r config set maxmemory 0
|
|
|
|
r config set maxmemory-policy allkeys-random
|
|
|
|
|
|
|
|
# Next rehash size is 8192, that will eat 64k memory
|
|
|
|
populate 4096 "" 1
|
|
|
|
|
|
|
|
set used [s used_memory]
|
|
|
|
set limit [expr {$used + 10*1024}]
|
|
|
|
r config set maxmemory $limit
|
|
|
|
r set k1 v1
|
|
|
|
# Next writing command will trigger evicting some keys if last
|
|
|
|
# command trigger DB dict rehash
|
|
|
|
r set k2 v2
|
|
|
|
# There must be 4098 keys because redis doesn't evict keys.
|
|
|
|
r dbsize
|
|
|
|
} {4098}
|
|
|
|
}
|
2020-12-06 07:51:22 -05:00
|
|
|
|
|
|
|
start_server {tags {"maxmemory"}} {
|
|
|
|
test {client tracking don't cause eviction feedback loop} {
|
|
|
|
r config set maxmemory 0
|
|
|
|
r config set maxmemory-policy allkeys-lru
|
|
|
|
r config set maxmemory-eviction-tenacity 100
|
|
|
|
|
|
|
|
# 10 clients listening on tracking messages
|
|
|
|
set clients {}
|
|
|
|
for {set j 0} {$j < 10} {incr j} {
|
|
|
|
lappend clients [redis_deferring_client]
|
|
|
|
}
|
|
|
|
foreach rd $clients {
|
|
|
|
$rd HELLO 3
|
|
|
|
$rd read ; # Consume the HELLO reply
|
|
|
|
$rd CLIENT TRACKING on
|
|
|
|
$rd read ; # Consume the CLIENT reply
|
|
|
|
}
|
|
|
|
|
|
|
|
# populate 300 keys, with long key name and short value
|
|
|
|
for {set j 0} {$j < 300} {incr j} {
|
|
|
|
set key $j[string repeat x 1000]
|
|
|
|
r set $key x
|
|
|
|
|
|
|
|
# for each key, enable caching for this key
|
|
|
|
foreach rd $clients {
|
|
|
|
$rd get $key
|
|
|
|
$rd read
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
2020-12-08 09:33:09 -05:00
|
|
|
# we need to wait one second for the client querybuf excess memory to be
|
|
|
|
# trimmed by cron, otherwise the INFO used_memory and CONFIG maxmemory
|
|
|
|
# below (on slow machines) won't be "atomic" and won't trigger eviction.
|
|
|
|
after 1100
|
|
|
|
|
2020-12-06 07:51:22 -05:00
|
|
|
# set the memory limit which will cause a few keys to be evicted
|
|
|
|
# we need to make sure to evict keynames of a total size of more than
|
|
|
|
# 16kb since the (PROTO_REPLY_CHUNK_BYTES), only after that the
|
|
|
|
# invalidation messages have a chance to trigger further eviction.
|
|
|
|
set used [s used_memory]
|
|
|
|
set limit [expr {$used - 40000}]
|
|
|
|
r config set maxmemory $limit
|
|
|
|
|
|
|
|
# make sure some eviction happened
|
|
|
|
set evicted [s evicted_keys]
|
|
|
|
if {$::verbose} { puts "evicted: $evicted" }
|
2020-12-08 09:33:09 -05:00
|
|
|
|
|
|
|
# make sure we didn't drain the database
|
|
|
|
assert_range [r dbsize] 200 300
|
|
|
|
|
2020-12-06 07:51:22 -05:00
|
|
|
assert_range $evicted 10 50
|
|
|
|
foreach rd $clients {
|
|
|
|
$rd read ;# make sure we have some invalidation message waiting
|
|
|
|
$rd close
|
|
|
|
}
|
|
|
|
|
2020-12-08 09:33:09 -05:00
|
|
|
# eviction continues (known problem described in #8069)
|
|
|
|
# for now this test only make sures the eviction loop itself doesn't
|
|
|
|
# have feedback loop
|
|
|
|
set evicted [s evicted_keys]
|
|
|
|
if {$::verbose} { puts "evicted: $evicted" }
|
2020-12-06 07:51:22 -05:00
|
|
|
}
|
|
|
|
}
|