Skip to content
体验新版
项目
组织
正在加载...
登录
切换导航
打开侧边栏
Mr、小李
TDengine
提交
c67ab52c
T
TDengine
项目概览
Mr、小李
/
TDengine
与 Fork 源项目一致
Fork自
taosdata / TDengine
通知
1
Star
0
Fork
0
代码
文件
提交
分支
Tags
贡献者
分支图
Diff
Issue
0
列表
看板
标记
里程碑
合并请求
0
Wiki
0
Wiki
分析
仓库
DevOps
项目成员
Pages
T
TDengine
项目概览
项目概览
详情
发布
仓库
仓库
文件
提交
分支
标签
贡献者
分支图
比较
Issue
0
Issue
0
列表
看板
标记
里程碑
合并请求
0
合并请求
0
Pages
分析
分析
仓库分析
DevOps
Wiki
0
Wiki
成员
成员
收起侧边栏
关闭侧边栏
动态
分支图
创建新Issue
提交
Issue看板
前往新版Gitcode,体验更适合开发者的 AI 搜索 >>
提交
c67ab52c
编写于
3月 18, 2021
作者:
S
Shuduo Sang
浏览文件
操作
浏览文件
下载
差异文件
Merge branch 'develop' into hotfix/sangshuduo/TD-3356-taosdemo-limit-offset-issue
上级
7a214c7f
624a9191
变更
6
隐藏空白更改
内联
并排
Showing
6 changed file
with
140 addition
and
70 deletion
+140
-70
documentation20/cn/12.taos-sql/docs.md
documentation20/cn/12.taos-sql/docs.md
+1
-1
packaging/tools/remove.sh
packaging/tools/remove.sh
+2
-2
src/kit/taosdemo/taosdemo.c
src/kit/taosdemo/taosdemo.c
+45
-36
tests/pytest/cluster/clusterEnvSetup/basic.py
tests/pytest/cluster/clusterEnvSetup/basic.py
+29
-10
tests/pytest/cluster/clusterEnvSetup/buildClusterEnv.sh
tests/pytest/cluster/clusterEnvSetup/buildClusterEnv.sh
+24
-21
tests/pytest/cluster/clusterEnvSetup/cleanClusterEnv.sh
tests/pytest/cluster/clusterEnvSetup/cleanClusterEnv.sh
+39
-0
未找到文件。
documentation20/cn/12.taos-sql/docs.md
浏览文件 @
c67ab52c
...
...
@@ -125,7 +125,7 @@ TDengine缺省的时间戳是毫秒精度,但通过修改配置参数enableMic
```mysql
ALTER DATABASE db_name CACHELAST 0;
```
CACHELAST 参数控制是否在内存中缓存数据子表的 last_row。缺省值为 0,取值范围 [0, 1]。其中 0 表示不启用、1 表示启用。(从 2.0.11 版本开始支持)
CACHELAST 参数控制是否在内存中缓存数据子表的 last_row。缺省值为 0,取值范围 [0, 1]。其中 0 表示不启用、1 表示启用。(从 2.0.11 版本开始支持
,修改后需要重启服务器生效。
)
**Tips**: 以上所有参数修改后都可以用show databases来确认是否修改成功。
...
...
packaging/tools/remove.sh
浏览文件 @
c67ab52c
...
...
@@ -213,10 +213,10 @@ fi
if
echo
$osinfo
|
grep
-qwi
"ubuntu"
;
then
# echo "this is ubuntu system"
${
csudo
}
rm
-f
/var/lib/dpkg/info/tdengine
*
||
:
${
csudo
}
dpkg
--force-all
-P
tdengine
||
:
elif
echo
$osinfo
|
grep
-qwi
"debian"
;
then
# echo "this is debian system"
${
csudo
}
rm
-f
/var/lib/dpkg/info/tdengine
*
||
:
${
csudo
}
dpkg
--force-all
-P
tdengine
||
:
elif
echo
$osinfo
|
grep
-qwi
"centos"
;
then
# echo "this is centos system"
${
csudo
}
rpm
-e
--noscripts
tdengine
||
:
...
...
src/kit/taosdemo/taosdemo.c
浏览文件 @
c67ab52c
...
...
@@ -101,8 +101,8 @@ typedef enum CREATE_SUB_TALBE_MOD_EN {
}
CREATE_SUB_TALBE_MOD_EN
;
typedef
enum
TALBE_EXISTS_EN
{
TBL_ALREADY_EXISTS
,
TBL_NO_EXISTS
,
TBL_ALREADY_EXISTS
,
TBL_EXISTS_BUTT
}
TALBE_EXISTS_EN
;
...
...
@@ -2405,8 +2405,11 @@ static int createDatabases() {
&
g_Dbs
.
db
[
i
].
superTbls
[
j
],
g_Dbs
.
use_metric
);
}
else
{
g_Dbs
.
db
[
i
].
superTbls
[
j
].
superTblExists
=
TBL_ALREADY_EXISTS
;
ret
=
getSuperTableFromServer
(
taos
,
g_Dbs
.
db
[
i
].
dbName
,
if
(
g_Dbs
.
db
[
i
].
superTbls
[
j
].
childTblExists
!=
TBL_ALREADY_EXISTS
)
{
ret
=
getSuperTableFromServer
(
taos
,
g_Dbs
.
db
[
i
].
dbName
,
&
g_Dbs
.
db
[
i
].
superTbls
[
j
]);
}
}
if
(
0
!=
ret
)
{
...
...
@@ -2798,9 +2801,10 @@ void readSampleFromFileToMem(SSuperTable * supterTblInfo) {
}
}
*/
static
bool
getColumnAndTagTypeFromInsertJsonFile
(
cJSON
*
stbInfo
,
SSuperTable
*
superTbls
)
{
static
bool
getColumnAndTagTypeFromInsertJsonFile
(
cJSON
*
stbInfo
,
SSuperTable
*
superTbls
)
{
bool
ret
=
false
;
// columns
cJSON
*
columns
=
cJSON_GetObjectItem
(
stbInfo
,
"columns"
);
if
(
columns
&&
columns
->
type
!=
cJSON_Array
)
{
...
...
@@ -2811,7 +2815,7 @@ static bool getColumnAndTagTypeFromInsertJsonFile(cJSON* stbInfo, SSuperTable* s
superTbls
->
tagCount
=
0
;
return
true
;
}
int
columnSize
=
cJSON_GetArraySize
(
columns
);
if
(
columnSize
>
MAX_COLUMN_COUNT
)
{
errorPrint
(
"%s() LN%d, failed to read json, column size overflow, max column size is %d
\n
"
,
...
...
@@ -2823,7 +2827,7 @@ static bool getColumnAndTagTypeFromInsertJsonFile(cJSON* stbInfo, SSuperTable* s
int
index
=
0
;
StrColumn
columnCase
;
//superTbls->columnCount = columnSize;
//superTbls->columnCount = columnSize;
for
(
int
k
=
0
;
k
<
columnSize
;
++
k
)
{
cJSON
*
column
=
cJSON_GetArrayItem
(
columns
,
k
);
if
(
column
==
NULL
)
continue
;
...
...
@@ -2831,7 +2835,7 @@ static bool getColumnAndTagTypeFromInsertJsonFile(cJSON* stbInfo, SSuperTable* s
count
=
1
;
cJSON
*
countObj
=
cJSON_GetObjectItem
(
column
,
"count"
);
if
(
countObj
&&
countObj
->
type
==
cJSON_Number
)
{
count
=
countObj
->
valueint
;
count
=
countObj
->
valueint
;
}
else
if
(
countObj
&&
countObj
->
type
!=
cJSON_Number
)
{
errorPrint
(
"%s() LN%d, failed to read json, column count not found
\n
"
,
__func__
,
__LINE__
);
goto
PARSE_OVER
;
...
...
@@ -2848,25 +2852,26 @@ static bool getColumnAndTagTypeFromInsertJsonFile(cJSON* stbInfo, SSuperTable* s
}
//tstrncpy(superTbls->columns[k].dataType, dataType->valuestring, MAX_TB_NAME_SIZE);
tstrncpy
(
columnCase
.
dataType
,
dataType
->
valuestring
,
MAX_TB_NAME_SIZE
);
cJSON
*
dataLen
=
cJSON_GetObjectItem
(
column
,
"len"
);
if
(
dataLen
&&
dataLen
->
type
==
cJSON_Number
)
{
columnCase
.
dataLen
=
dataLen
->
valueint
;
columnCase
.
dataLen
=
dataLen
->
valueint
;
}
else
if
(
dataLen
&&
dataLen
->
type
!=
cJSON_Number
)
{
debugPrint
(
"%s() LN%d: failed to read json, column len not found
\n
"
,
__func__
,
__LINE__
);
goto
PARSE_OVER
;
}
else
{
columnCase
.
dataLen
=
8
;
}
for
(
int
n
=
0
;
n
<
count
;
++
n
)
{
tstrncpy
(
superTbls
->
columns
[
index
].
dataType
,
columnCase
.
dataType
,
MAX_TB_NAME_SIZE
);
superTbls
->
columns
[
index
].
dataLen
=
columnCase
.
dataLen
;
tstrncpy
(
superTbls
->
columns
[
index
].
dataType
,
columnCase
.
dataType
,
MAX_TB_NAME_SIZE
);
superTbls
->
columns
[
index
].
dataLen
=
columnCase
.
dataLen
;
index
++
;
}
}
superTbls
->
columnCount
=
index
;
}
superTbls
->
columnCount
=
index
;
count
=
1
;
index
=
0
;
// tags
...
...
@@ -2881,16 +2886,16 @@ static bool getColumnAndTagTypeFromInsertJsonFile(cJSON* stbInfo, SSuperTable* s
debugPrint
(
"%s() LN%d, failed to read json, tags size overflow, max tag size is %d
\n
"
,
__func__
,
__LINE__
,
MAX_TAG_COUNT
);
goto
PARSE_OVER
;
}
//superTbls->tagCount = tagSize;
//superTbls->tagCount = tagSize;
for
(
int
k
=
0
;
k
<
tagSize
;
++
k
)
{
cJSON
*
tag
=
cJSON_GetArrayItem
(
tags
,
k
);
if
(
tag
==
NULL
)
continue
;
count
=
1
;
cJSON
*
countObj
=
cJSON_GetObjectItem
(
tag
,
"count"
);
if
(
countObj
&&
countObj
->
type
==
cJSON_Number
)
{
count
=
countObj
->
valueint
;
count
=
countObj
->
valueint
;
}
else
if
(
countObj
&&
countObj
->
type
!=
cJSON_Number
)
{
printf
(
"ERROR: failed to read json, column count not found
\n
"
);
goto
PARSE_OVER
;
...
...
@@ -2906,23 +2911,23 @@ static bool getColumnAndTagTypeFromInsertJsonFile(cJSON* stbInfo, SSuperTable* s
goto
PARSE_OVER
;
}
tstrncpy
(
columnCase
.
dataType
,
dataType
->
valuestring
,
MAX_TB_NAME_SIZE
);
cJSON
*
dataLen
=
cJSON_GetObjectItem
(
tag
,
"len"
);
if
(
dataLen
&&
dataLen
->
type
==
cJSON_Number
)
{
columnCase
.
dataLen
=
dataLen
->
valueint
;
columnCase
.
dataLen
=
dataLen
->
valueint
;
}
else
if
(
dataLen
&&
dataLen
->
type
!=
cJSON_Number
)
{
printf
(
"ERROR: failed to read json, column len not found
\n
"
);
goto
PARSE_OVER
;
}
else
{
columnCase
.
dataLen
=
0
;
}
}
for
(
int
n
=
0
;
n
<
count
;
++
n
)
{
tstrncpy
(
superTbls
->
tags
[
index
].
dataType
,
columnCase
.
dataType
,
MAX_TB_NAME_SIZE
);
superTbls
->
tags
[
index
].
dataLen
=
columnCase
.
dataLen
;
superTbls
->
tags
[
index
].
dataLen
=
columnCase
.
dataLen
;
index
++
;
}
}
}
superTbls
->
tagCount
=
index
;
ret
=
true
;
...
...
@@ -3107,8 +3112,10 @@ static bool getMetaFromInsertJsonFile(cJSON* root) {
}
cJSON
*
precision
=
cJSON_GetObjectItem
(
dbinfo
,
"precision"
);
if
(
precision
&&
precision
->
type
==
cJSON_String
&&
precision
->
valuestring
!=
NULL
)
{
tstrncpy
(
g_Dbs
.
db
[
i
].
dbCfg
.
precision
,
precision
->
valuestring
,
MAX_DB_NAME_SIZE
);
if
(
precision
&&
precision
->
type
==
cJSON_String
&&
precision
->
valuestring
!=
NULL
)
{
tstrncpy
(
g_Dbs
.
db
[
i
].
dbCfg
.
precision
,
precision
->
valuestring
,
MAX_DB_NAME_SIZE
);
}
else
if
(
!
precision
)
{
//tstrncpy(g_Dbs.db[i].dbCfg.precision, "ms", MAX_DB_NAME_SIZE);
memset
(
g_Dbs
.
db
[
i
].
dbCfg
.
precision
,
0
,
MAX_DB_NAME_SIZE
);
...
...
@@ -3334,13 +3341,13 @@ static bool getMetaFromInsertJsonFile(cJSON* root) {
}
else
if
(
!
childTblExists
)
{
g_Dbs
.
db
[
i
].
superTbls
[
j
].
childTblExists
=
TBL_NO_EXISTS
;
}
else
{
printf
(
"ERROR: failed to read json, child_table_exists not found
\n
"
);
errorPrint
(
"%s() LN%d, failed to read json, child_table_exists not found
\n
"
,
__func__
,
__LINE__
);
goto
PARSE_OVER
;
}
cJSON
*
count
=
cJSON_GetObjectItem
(
stbInfo
,
"childtable_count"
);
if
(
!
count
||
count
->
type
!=
cJSON_Number
||
0
>=
count
->
valueint
)
{
printf
(
"ERROR: failed to read json, childtable_count not found
\n
"
);
errorPrint
(
"%s() LN%d, failed to read json, childtable_count not found
\n
"
,
__func__
,
__LINE__
);
goto
PARSE_OVER
;
}
g_Dbs
.
db
[
i
].
superTbls
[
j
].
childTblCount
=
count
->
valueint
;
...
...
@@ -3353,7 +3360,7 @@ static bool getMetaFromInsertJsonFile(cJSON* root) {
}
else
if
(
!
dataSource
)
{
tstrncpy
(
g_Dbs
.
db
[
i
].
superTbls
[
j
].
dataSource
,
"rand"
,
MAX_DB_NAME_SIZE
);
}
else
{
printf
(
"ERROR: failed to read json, data_source not found
\n
"
);
errorPrint
(
"%s() LN%d, failed to read json, data_source not found
\n
"
,
__func__
,
__LINE__
);
goto
PARSE_OVER
;
}
...
...
@@ -3550,12 +3557,14 @@ static bool getMetaFromInsertJsonFile(cJSON* root) {
goto
PARSE_OVER
;
}
if
(
NO_CREATE_SUBTBL
==
g_Dbs
.
db
[
i
].
superTbls
[
j
].
autoCreateTable
/* CBD
if (NO_CREATE_SUBTBL == g_Dbs.db[i].superTbls[j].autoCreateTable
|| (TBL_ALREADY_EXISTS == g_Dbs.db[i].superTbls[j].childTblExists)) {
continue;
}
*/
int
retVal
=
getColumnAndTagTypeFromInsertJsonFile
(
stbInfo
,
&
g_Dbs
.
db
[
i
].
superTbls
[
j
]);
int
retVal
=
getColumnAndTagTypeFromInsertJsonFile
(
stbInfo
,
&
g_Dbs
.
db
[
i
].
superTbls
[
j
]);
if
(
false
==
retVal
)
{
goto
PARSE_OVER
;
}
...
...
@@ -4504,7 +4513,7 @@ static void* syncWriteInterlace(threadInfo *pThreadInfo) {
pstr
+=
headLen
;
int
dataLen
=
0
;
debug
Print
(
"[%d] %s() LN%d i=%d batchPerTblTimes=%d batchPerTbl = %d
\n
"
,
verbose
Print
(
"[%d] %s() LN%d i=%d batchPerTblTimes=%d batchPerTbl = %d
\n
"
,
pThreadInfo
->
threadID
,
__func__
,
__LINE__
,
i
,
batchPerTblTimes
,
batchPerTbl
);
generateDataTail
(
...
...
@@ -4516,7 +4525,7 @@ static void* syncWriteInterlace(threadInfo *pThreadInfo) {
pstr
+=
dataLen
;
recOfBatch
+=
batchPerTbl
;
pThreadInfo
->
totalInsertRows
+=
batchPerTbl
;
debug
Print
(
"[%d] %s() LN%d batchPerTbl=%d recOfBatch=%d
\n
"
,
verbose
Print
(
"[%d] %s() LN%d batchPerTbl=%d recOfBatch=%d
\n
"
,
pThreadInfo
->
threadID
,
__func__
,
__LINE__
,
batchPerTbl
,
recOfBatch
);
...
...
@@ -4539,7 +4548,7 @@ static void* syncWriteInterlace(threadInfo *pThreadInfo) {
if
((
remainRows
>
0
)
&&
(
batchPerTbl
>
remainRows
))
batchPerTbl
=
remainRows
;
debug
Print
(
"[%d] %s() LN%d generatedRecPerTbl=%d insertRows=%"
PRId64
"
\n
"
,
verbose
Print
(
"[%d] %s() LN%d generatedRecPerTbl=%d insertRows=%"
PRId64
"
\n
"
,
pThreadInfo
->
threadID
,
__func__
,
__LINE__
,
generatedRecPerTbl
,
insertRows
);
...
...
@@ -4547,7 +4556,7 @@ static void* syncWriteInterlace(threadInfo *pThreadInfo) {
break
;
}
debug
Print
(
"[%d] %s() LN%d recOfBatch=%d totalInsertRows=%"
PRId64
"
\n
"
,
verbose
Print
(
"[%d] %s() LN%d recOfBatch=%d totalInsertRows=%"
PRId64
"
\n
"
,
pThreadInfo
->
threadID
,
__func__
,
__LINE__
,
recOfBatch
,
pThreadInfo
->
totalInsertRows
);
verbosePrint
(
"[%d] %s() LN%d, buffer=%s
\n
"
,
...
...
tests/pytest/cluster/clusterEnvSetup/basic.py
浏览文件 @
c67ab52c
...
...
@@ -18,13 +18,15 @@ import argparse
class
BuildDockerCluser
:
def
__init__
(
self
,
hostName
,
user
,
password
,
configDir
,
numOfNodes
,
clusterVersion
):
def
__init__
(
self
,
hostName
,
user
,
password
,
configDir
,
numOfNodes
,
clusterVersion
,
dockerDir
,
removeFlag
):
self
.
hostName
=
hostName
self
.
user
=
user
self
.
password
=
password
self
.
configDir
=
configDir
self
.
numOfNodes
=
numOfNodes
self
.
clusterVersion
=
clusterVersion
self
.
clusterVersion
=
clusterVersion
self
.
dockerDir
=
dockerDir
self
.
removeFlag
=
removeFlag
def
getConnection
(
self
):
self
.
conn
=
taos
.
connect
(
...
...
@@ -42,14 +44,17 @@ class BuildDockerCluser:
print
(
"start arbitrator"
)
os
.
system
(
"docker exec -d $(docker ps|grep tdnode1|awk '{print $1}') tarbitrator"
)
def
run
(
self
):
def
run
(
self
):
if
self
.
numOfNodes
<
2
or
self
.
numOfNodes
>
5
:
print
(
"the number of nodes must be between 2 and 5"
)
exit
(
0
)
os
.
system
(
"./buildClusterEnv.sh -n %d -v %s"
%
(
self
.
numOfNodes
,
self
.
clusterVersion
))
exit
(
0
)
print
(
"remove Flag value %s"
%
self
.
removeFlag
)
if
self
.
removeFlag
==
False
:
os
.
system
(
"./cleanClusterEnv.sh -d %s"
%
self
.
dockerDir
)
os
.
system
(
"./buildClusterEnv.sh -n %d -v %s -d %s"
%
(
self
.
numOfNodes
,
self
.
clusterVersion
,
self
.
dockerDir
))
self
.
getConnection
()
self
.
createDondes
()
self
.
startArbitrator
()
self
.
startArbitrator
()
parser
=
argparse
.
ArgumentParser
()
parser
.
add_argument
(
...
...
@@ -91,10 +96,24 @@ parser.add_argument(
'-v'
,
'--version'
,
action
=
'store'
,
default
=
'2.0.1
4
.1'
,
default
=
'2.0.1
7
.1'
,
type
=
str
,
help
=
'the version of the cluster to be build, Default is 2.0.14.1'
)
help
=
'the version of the cluster to be build, Default is 2.0.17.1'
)
parser
.
add_argument
(
'-d'
,
'--docker-dir'
,
action
=
'store'
,
default
=
'/data'
,
type
=
str
,
help
=
'the data dir for docker, default is /data'
)
parser
.
add_argument
(
'--flag'
,
action
=
'store_true'
,
help
=
'remove docker containers flag, default: True'
)
args
=
parser
.
parse_args
()
cluster
=
BuildDockerCluser
(
args
.
host
,
args
.
user
,
args
.
password
,
args
.
config_dir
,
args
.
num_of_nodes
,
args
.
version
)
cluster
.
run
()
\ No newline at end of file
cluster
=
BuildDockerCluser
(
args
.
host
,
args
.
user
,
args
.
password
,
args
.
config_dir
,
args
.
num_of_nodes
,
args
.
version
,
args
.
docker_dir
,
args
.
flag
)
cluster
.
run
()
# usage 1: python3 basic.py -n 2 --flag (flag is True)
# usage 2: python3 basic.py -n 2 (flag should be False when it is not specified)
\ No newline at end of file
tests/pytest/cluster/clusterEnvSetup/buildClusterEnv.sh
浏览文件 @
c67ab52c
#!/bin/bash
echo
"Executing buildClusterEnv.sh"
DOCKER_DIR
=
/data
CURR_DIR
=
`
pwd
`
if
[
$#
!=
4
]
;
then
if
[
$#
!=
6
]
;
then
echo
"argument list need input : "
echo
" -n numOfNodes"
echo
" -v version"
echo
" -v version"
echo
" -d docker dir"
exit
1
fi
NUM_OF_NODES
=
VERSION
=
while
getopts
"n:v:"
arg
DOCKER_DIR
=
while
getopts
"n:v:d:"
arg
do
case
$arg
in
n
)
...
...
@@ -20,6 +21,9 @@ do
;;
v
)
VERSION
=
$OPTARG
;;
d
)
DOCKER_DIR
=
$OPTARG
;;
?
)
echo
"unkonwn argument"
...
...
@@ -31,29 +35,28 @@ function addTaoscfg {
for
i
in
{
1..5
}
do
touch
/data/node
$i
/cfg/taos.cfg
echo
'firstEp tdnode1:6030'
>
/data
/node
$i
/cfg/taos.cfg
echo
'fqdn tdnode'
$i
>>
/data
/node
$i
/cfg/taos.cfg
echo
'arbitrator tdnode1:6042'
>>
/data
/node
$i
/cfg/taos.cfg
echo
'firstEp tdnode1:6030'
>
$DOCKER_DIR
/node
$i
/cfg/taos.cfg
echo
'fqdn tdnode'
$i
>>
$DOCKER_DIR
/node
$i
/cfg/taos.cfg
echo
'arbitrator tdnode1:6042'
>>
$DOCKER_DIR
/node
$i
/cfg/taos.cfg
done
}
function
createDIR
{
for
i
in
{
1..5
}
do
mkdir
-p
/data
/node
$i
/data
mkdir
-p
/data
/node
$i
/log
mkdir
-p
/data
/node
$i
/cfg
mkdir
-p
/data
/node
$i
/core
mkdir
-p
$DOCKER_DIR
/node
$i
/data
mkdir
-p
$DOCKER_DIR
/node
$i
/log
mkdir
-p
$DOCKER_DIR
/node
$i
/cfg
mkdir
-p
$DOCKER_DIR
/node
$i
/core
done
}
function
cleanEnv
{
function
cleanEnv
{
echo
"Clean up docker environment"
for
i
in
{
1..5
}
do
echo
/data/node
$i
/data/
*
rm
-rf
/data/node
$i
/data/
*
echo
/data/node
$i
/log/
*
rm
-rf
/data/node
$i
/log/
*
do
rm
-rf
$DOCKER_DIR
/node
$i
/data/
*
rm
-rf
$DOCKER_DIR
/node
$i
/log/
*
done
}
...
...
@@ -98,19 +101,19 @@ function clusterUp {
if
[
$NUM_OF_NODES
-eq
2
]
;
then
echo
"create 2 dnodes"
PACKAGE
=
TDengine-server-
$VERSION
-Linux-x64
.tar.gz
DIR
=
TDengine-server-
$VERSION
DIR2
=
TDengine-arbitrator-
$VERSION
VERSION
=
$VERSION
docker-compose up
-d
PACKAGE
=
TDengine-server-
$VERSION
-Linux-x64
.tar.gz
TARBITRATORPKG
=
TDengine-arbitrator-
$VERSION
-Linux-x64
.tar.gz
DIR
=
TDengine-server-
$VERSION
DIR2
=
TDengine-arbitrator-
$VERSION
VERSION
=
$VERSION
docker-compose up
-d
fi
if
[
$NUM_OF_NODES
-eq
3
]
;
then
PACKAGE
=
TDengine-server-
$VERSION
-Linux-x64
.tar.gz
DIR
=
TDengine-server-
$VERSION
DIR2
=
TDengine-arbitrator-
$VERSION
VERSION
=
$VERSION
docker-compose
-f
docker-compose.yml
-f
node3.yml up
-d
PACKAGE
=
TDengine-server-
$VERSION
-Linux-x64
.tar.gz
TARBITRATORPKG
=
TDengine-arbitrator-
$VERSION
-Linux-x64
.tar.gz
DIR
=
TDengine-server-
$VERSION
DIR2
=
TDengine-arbitrator-
$VERSION
VERSION
=
$VERSION
docker-compose
-f
docker-compose.yml
-f
node3.yml up
-d
fi
if
[
$NUM_OF_NODES
-eq
4
]
;
then
PACKAGE
=
TDengine-server-
$VERSION
-Linux-x64
.tar.gz
DIR
=
TDengine-server-
$VERSION
DIR2
=
TDengine-arbitrator-
$VERSION
VERSION
=
$VERSION
docker-compose
-f
docker-compose.yml
-f
node3.yml
-f
node4.yml up
-d
PACKAGE
=
TDengine-server-
$VERSION
-Linux-x64
.tar.gz
TARBITRATORPKG
=
TDengine-arbitrator-
$VERSION
-Linux-x64
.tar.gz
DIR
=
TDengine-server-
$VERSION
DIR2
=
TDengine-arbitrator-
$VERSION
VERSION
=
$VERSION
docker-compose
-f
docker-compose.yml
-f
node3.yml
-f
node4.yml up
-d
fi
if
[
$NUM_OF_NODES
-eq
5
]
;
then
PACKAGE
=
TDengine-server-
$VERSION
-Linux-x64
.tar.gz
DIR
=
TDengine-server-
$VERSION
DIR2
=
TDengine-arbitrator-
$VERSION
VERSION
=
$VERSION
docker-compose
-f
docker-compose.yml
-f
node3.yml
-f
node4.yml
-f
node5.yml up
-d
PACKAGE
=
TDengine-server-
$VERSION
-Linux-x64
.tar.gz
TARBITRATORPKG
=
TDengine-arbitrator-
$VERSION
-Linux-x64
.tar.gz
DIR
=
TDengine-server-
$VERSION
DIR2
=
TDengine-arbitrator-
$VERSION
VERSION
=
$VERSION
docker-compose
-f
docker-compose.yml
-f
node3.yml
-f
node4.yml
-f
node5.yml up
-d
fi
echo
"docker compose finish"
...
...
tests/pytest/cluster/clusterEnvSetup/cleanClusterEnv.sh
0 → 100755
浏览文件 @
c67ab52c
#!/bin/bash
echo
"Executing cleanClusterEnv.sh"
CURR_DIR
=
`
pwd
`
if
[
$#
!=
2
]
;
then
echo
"argument list need input : "
echo
" -d docker dir"
exit
1
fi
DOCKER_DIR
=
while
getopts
"d:"
arg
do
case
$arg
in
d
)
DOCKER_DIR
=
$OPTARG
;;
?
)
echo
"unkonwn argument"
;;
esac
done
function
removeDockerContainers
{
cd
$DOCKER_DIR
docker-compose down
--remove-orphans
}
function
cleanEnv
{
echo
"Clean up docker environment"
for
i
in
{
1..5
}
do
rm
-rf
$DOCKER_DIR
/node
$i
/data/
*
rm
-rf
$DOCKER_DIR
/node
$i
/log/
*
done
}
removeDockerContainers
cleanEnv
\ No newline at end of file
编辑
预览
Markdown
is supported
0%
请重试
或
添加新附件
.
添加附件
取消
You are about to add
0
people
to the discussion. Proceed with caution.
先完成此消息的编辑!
取消
想要评论请
注册
或
登录