Skip to content
体验新版
项目
组织
正在加载...
登录
切换导航
打开侧边栏
Turbo码先生
redis
提交
5e1d2d30
R
redis
项目概览
Turbo码先生
/
redis
与 Fork 源项目一致
从无法访问的项目Fork
通知
1
Star
0
Fork
0
代码
文件
提交
分支
Tags
贡献者
分支图
Diff
Issue
0
列表
看板
标记
里程碑
合并请求
0
Wiki
0
Wiki
分析
仓库
DevOps
项目成员
Pages
R
redis
项目概览
项目概览
详情
发布
仓库
仓库
文件
提交
分支
标签
贡献者
分支图
比较
Issue
0
Issue
0
列表
看板
标记
里程碑
合并请求
0
合并请求
0
Pages
分析
分析
仓库分析
DevOps
Wiki
0
Wiki
成员
成员
收起侧边栏
关闭侧边栏
动态
分支图
创建新Issue
提交
Issue看板
体验新版 GitCode,发现更多精彩内容 >>
提交
5e1d2d30
编写于
1月 09, 2011
作者:
A
antirez
浏览文件
操作
浏览文件
下载
电子邮件补丁
差异文件
initial fix of the test suite to run both in in-memory and diskstore mode
上级
67a1810b
变更
5
隐藏空白更改
内联
并排
Showing
5 changed file
with
147 addition
and
25 deletion
+147
-25
src/zmalloc.c
src/zmalloc.c
+1
-1
tests/assets/default.conf
tests/assets/default.conf
+131
-17
tests/support/server.tcl
tests/support/server.tcl
+0
-3
tests/support/tmpfile.tcl
tests/support/tmpfile.tcl
+7
-1
tests/test_helper.tcl
tests/test_helper.tcl
+8
-3
未找到文件。
src/zmalloc.c
浏览文件 @
5e1d2d30
...
...
@@ -186,7 +186,7 @@ size_t zmalloc_used_memory(void) {
}
size_t
zmalloc_allocations_for_size
(
size_t
size
)
{
if
(
size
>
ZMALLOC_MAX_ALLOC_STAT
||
size
<
0
)
return
0
;
if
(
size
>
ZMALLOC_MAX_ALLOC_STAT
)
return
0
;
return
zmalloc_allocations
[
size
];
}
...
...
tests/assets/default.conf
浏览文件 @
5e1d2d30
...
...
@@ -18,9 +18,9 @@ daemonize no
# When running daemonized, Redis writes a pid file in /var/run/redis.pid by
# default. You can specify a custom pid file location here.
pidfile
redis
.
pid
pidfile
/
var
/
run
/
redis
.
pid
# Accept connections on the specified port, default is 6379
# Accept connections on the specified port, default is 6379
.
port
6379
# If you want you can bind a single interface, if the bind option is not
...
...
@@ -28,6 +28,12 @@ port 6379
#
# bind 127.0.0.1
# Specify the path for the unix socket that will be used to listen for
# incoming connections. There is no default, so Redis will not listen
# on a unix socket when not specified.
#
# unixsocket /tmp/redis.sock
# Close the connection after a client is idle for N seconds (0 to disable)
timeout
300
...
...
@@ -44,6 +50,16 @@ loglevel verbose
# output for logging but daemonize, logs will be sent to /dev/null
logfile
stdout
# To enable logging to the system logger, just set 'syslog-enabled' to yes,
# and optionally update the other syslog parameters to suit your needs.
# syslog-enabled no
# Specify the syslog identity.
# syslog-ident redis
# Specify the syslog facility. Must be USER or between LOCAL0-LOCAL7.
# syslog-facility local0
# Set the number of databases. The default database is DB 0, you can select
# a different one on a per-connection basis using SELECT <dbid> where
# dbid is a number between 0 and 'databases'-1
...
...
@@ -86,7 +102,7 @@ dbfilename dump.rdb
# Also the Append Only File will be created inside this directory.
#
# Note that you must specify a directory here, not a file name.
dir
/
tmp
dir
./
################################# REPLICATION #################################
...
...
@@ -104,6 +120,19 @@ dir /tmp
#
# masterauth <master-password>
# When a slave lost the connection with the master, or when the replication
# is still in progress, the slave can act in two different ways:
#
# 1) if slave-serve-stale-data is set to 'yes' (the default) the slave will
# still reply to client requests, possibly with out of data data, or the
# data set may just be empty if this is the first synchronization.
#
# 2) if slave-serve-stale data is set to 'no' the slave will reply with
# an error "SYNC with master in progress" to all the kind of commands
# but to INFO and SLAVEOF.
#
slave
-
serve
-
stale
-
data
yes
################################## SECURITY ###################################
# Require clients to issue AUTH <PASSWORD> before processing any other
...
...
@@ -119,6 +148,22 @@ dir /tmp
#
# requirepass foobared
# Command renaming.
#
# It is possilbe to change the name of dangerous commands in a shared
# environment. For instance the CONFIG command may be renamed into something
# of hard to guess so that it will be still available for internal-use
# tools but not available for general clients.
#
# Example:
#
# rename-command CONFIG b840fc02d524045429941cc15f59e41cb7be6c52
#
# It is also possilbe to completely kill a command renaming it into
# an empty string:
#
# rename-command CONFIG ""
################################### LIMITS ####################################
# Set the max number of connected clients at the same time. By default there
...
...
@@ -148,6 +193,37 @@ dir /tmp
#
# maxmemory <bytes>
# MAXMEMORY POLICY: how Redis will select what to remove when maxmemory
# is reached? You can select among five behavior:
#
# volatile-lru -> remove the key with an expire set using an LRU algorithm
# allkeys-lru -> remove any key accordingly to the LRU algorithm
# volatile-random -> remove a random key with an expire set
# allkeys->random -> remove a random key, any key
# volatile-ttl -> remove the key with the nearest expire time (minor TTL)
# noeviction -> don't expire at all, just return an error on write operations
#
# Note: with all the kind of policies, Redis will return an error on write
# operations, when there are not suitable keys for eviction.
#
# At the date of writing this commands are: set setnx setex append
# incr decr rpush lpush rpushx lpushx linsert lset rpoplpush sadd
# sinter sinterstore sunion sunionstore sdiff sdiffstore zadd zincrby
# zunionstore zinterstore hset hsetnx hmset hincrby incrby decrby
# getset mset msetnx exec sort
#
# The default is:
#
# maxmemory-policy volatile-lru
# LRU and minimal TTL algorithms are not precise algorithms but approximated
# algorithms (in order to save memory), so you can select as well the sample
# size to check. For instance for default Redis will check three keys and
# pick the one that was used less recently, you can change the sample size
# using the following configuration directive.
#
# maxmemory-samples 3
############################## APPEND ONLY MODE ###############################
# By default Redis asynchronously dumps the dataset on disk. If you can live
...
...
@@ -195,29 +271,54 @@ appendonly no
appendfsync
everysec
# appendfsync no
################################ VIRTUAL MEMORY ###############################
# When the AOF fsync policy is set to always or everysec, and a background
# saving process (a background save or AOF log background rewriting) is
# performing a lot of I/O against the disk, in some Linux configurations
# Redis may block too long on the fsync() call. Note that there is no fix for
# this currently, as even performing fsync in a different thread will block
# our synchronous write(2) call.
#
# In order to mitigate this problem it's possible to use the following option
# that will prevent fsync() from being called in the main process while a
# BGSAVE or BGREWRITEAOF is in progress.
#
# This means that while another child is saving the durability of Redis is
# the same as "appendfsync none", that in pratical terms means that it is
# possible to lost up to 30 seconds of log in the worst scenario (with the
# default Linux settings).
#
# If you have latency problems turn this to "yes". Otherwise leave it as
# "no" that is the safest pick from the point of view of durability.
no
-
appendfsync
-
on
-
rewrite
no
#################################### DISK STORE ###############################
#
Virtual Memory allows Redis to work with datasets bigger than the actual
#
amount of RAM needed to hold the whole dataset in memory
.
#
In order to do so very used keys are taken in memory while the other keys
#
are swapped into a swap file, similarly to what operating systems do
# w
ith memory pages.
#
When disk store is active Redis works as an on-disk database, where memory
#
is only used as a object cache
.
#
#
This mode is good for datasets that are bigger than memory, and in general
# w
hen you want to trade speed for:
#
# To enable VM just set 'vm-enabled' to yes, and set the following three
# VM parameters accordingly to your needs.
# - less memory used
# - immediate server restart
# - per key durability, without need for backgrond savig
#
# On the other hand, with disk store enabled MULTI/EXEC are no longer
# transactional from the point of view of the persistence on disk, that is,
# Redis transactions will still guarantee that commands are either processed
# all or nothing, but there is no guarantee that all the keys are flushed
# on disk in an atomic way.
#
# Of course with disk store enabled Redis is not as fast as it is when
# working with just the memory back end.
diskstore
-
enabled
yes
diskstore
-
enabled
no
diskstore
-
path
redis
.
ds
cache
-
max
-
memory
0
cache
-
flush
-
delay
0
############################### ADVANCED CONFIG ###############################
# Glue small output buffers together in order to send small replies in a
# single TCP packet. Uses a bit more CPU but most of the times it is a win
# in terms of number of queries per second. Use 'yes' if unsure.
glueoutputbuf
yes
# Hashes are encoded in a special way (much more memory efficient) when they
# have at max a given numer of elements, and the biggest element does not
# exceed a given threshold. You can configure this limits with the following
...
...
@@ -225,6 +326,19 @@ glueoutputbuf yes
hash
-
max
-
zipmap
-
entries
64
hash
-
max
-
zipmap
-
value
512
# Similarly to hashes, small lists are also encoded in a special way in order
# to save a lot of space. The special representation is only used when
# you are under the following limits:
list
-
max
-
ziplist
-
entries
512
list
-
max
-
ziplist
-
value
64
# Sets have a special encoding in just one case: when a set is composed
# of just strings that happens to be integers in radix 10 in the range
# of 64 bit signed integers.
# The following configuration setting sets the limit in the size of the
# set in order to use this special memory saving encoding.
set
-
max
-
intset
-
entries
512
# Active rehashing uses 1 millisecond every 100 milliseconds of CPU time in
# order to help rehashing the main Redis hash table (the one mapping top-level
# keys to values). The hash table implementation redis uses (see dict.c)
...
...
tests/support/server.tcl
浏览文件 @
5e1d2d30
...
...
@@ -177,9 +177,6 @@ proc start_server {options {code undefined}} {
if
{
$::valgrind
}
{
exec valgrind --suppressions=src/valgrind.sup src/redis-server $config_file > $stdout 2> $stderr &
}
else
{
if
{
$::verbose
}
{
puts
"Logging on
$stdout
/
$stderr
"
}
exec src/redis-server $config_file > $stdout 2> $stderr &
}
...
...
tests/support/tmpfile.tcl
浏览文件 @
5e1d2d30
...
...
@@ -4,7 +4,13 @@ file mkdir $::tmproot
# returns a dirname unique to this process to write to
proc tmpdir
{
basename
}
{
set dir
[
file join $::tmproot $basename.
[
pid
]
.
[
incr ::tmpcounter
]]
if
{
$::diskstore
}
{
# For diskstore we want to use the same dir again and again
# otherwise everything is too slow.
set dir
[
file join $::tmproot $basename.diskstore
]
}
else
{
set dir
[
file join $::tmproot $basename.
[
pid
]
.
[
incr ::tmpcounter
]]
}
file mkdir $dir
set _ $dir
}
...
...
tests/test_helper.tcl
浏览文件 @
5e1d2d30
...
...
@@ -13,12 +13,13 @@ set ::host 127.0.0.1
set ::port 16379
set ::traceleaks 0
set ::valgrind 0
set ::verbose
1
set ::verbose
0
set ::denytags
{}
set ::allowtags
{}
set ::external 0
;
# If
"1"
this means, we are running against external instance
set ::file
""
;
# If set, runs only the tests in this comma separated list
set ::curfile
""
;
# Hold the filename of the current suite
set ::diskstore 0
;
# Don't touch this by hand. The test itself will toggle it.
proc execute_tests name
{
set path
"tests/
$name.tcl
"
...
...
@@ -103,6 +104,9 @@ proc s {args} {
}
proc cleanup
{}
{
if
{
$::diskstore
}
{
puts
"Cleanup: warning may take some minute (diskstore enabled)"
}
catch
{
exec rm -rf
{*}
[
glob tests/tmp/redis.conf.*
]}
catch
{
exec rm -rf
{*}
[
glob tests/tmp/server.*
]}
}
...
...
@@ -125,8 +129,9 @@ proc execute_everything {} {
# execute_tests
"integration/redis-cli"
execute_tests
"unit/pubsub"
# run tests with VM enabled
set ::global_overrides
{
vm-enabled yes
}
# run tests with diskstore enabled
set ::diskstore 1
set ::global_overrides
{
diskstore-enabled yes
}
execute_tests
"unit/protocol"
execute_tests
"unit/basic"
execute_tests
"unit/type/list"
...
...
编辑
预览
Markdown
is supported
0%
请重试
或
添加新附件
.
添加附件
取消
You are about to add
0
people
to the discussion. Proceed with caution.
先完成此消息的编辑!
取消
想要评论请
注册
或
登录