Skip to content
体验新版
项目
组织
正在加载...
登录
切换导航
打开侧边栏
慢慢CG
TDengine
提交
7d5f3114
T
TDengine
项目概览
慢慢CG
/
TDengine
与 Fork 源项目一致
Fork自
taosdata / TDengine
通知
1
Star
0
Fork
0
代码
文件
提交
分支
Tags
贡献者
分支图
Diff
Issue
0
列表
看板
标记
里程碑
合并请求
0
Wiki
0
Wiki
分析
仓库
DevOps
项目成员
Pages
T
TDengine
项目概览
项目概览
详情
发布
仓库
仓库
文件
提交
分支
标签
贡献者
分支图
比较
Issue
0
Issue
0
列表
看板
标记
里程碑
合并请求
0
合并请求
0
Pages
分析
分析
仓库分析
DevOps
Wiki
0
Wiki
成员
成员
收起侧边栏
关闭侧边栏
动态
分支图
创建新Issue
提交
Issue看板
提交
7d5f3114
编写于
9月 16, 2021
作者:
J
jiajingbin
浏览文件
操作
浏览文件
下载
电子邮件补丁
差异文件
save
上级
616eb0e7
变更
2
隐藏空白更改
内联
并排
Showing
2 changed file
with
1558 addition
and
1 deletion
+1558
-1
tests/pytest/insert/openTsdbJsonInsert.py
tests/pytest/insert/openTsdbJsonInsert.py
+1521
-0
tests/pytest/util/common.py
tests/pytest/util/common.py
+37
-1
未找到文件。
tests/pytest/insert/openTsdbJsonInsert.py
0 → 100644
浏览文件 @
7d5f3114
###################################################################
# Copyright (c) 2021 by TAOS Technologies, Inc.
# All rights reserved.
#
# This file is proprietary and confidential to TAOS Technologies.
# No part of this file may be reproduced, stored, transmitted,
# disclosed or used in any form or by any means other than as
# expressly provided by the written permission from Jianhui Tao
#
###################################################################
# -*- coding: utf-8 -*-
import
traceback
import
random
from
taos.error
import
JsonPayloadError
import
time
from
copy
import
deepcopy
import
numpy
as
np
from
util.log
import
*
from
util.cases
import
*
from
util.sql
import
*
from
util.common
import
tdCom
import
threading
import
json
class
TDTestCase
:
def
init
(
self
,
conn
,
logSql
):
tdLog
.
debug
(
"start to execute %s"
%
__file__
)
tdSql
.
init
(
conn
.
cursor
(),
logSql
)
self
.
_conn
=
conn
def
createDb
(
self
,
name
=
"test"
,
db_update_tag
=
0
):
if
db_update_tag
==
0
:
tdSql
.
execute
(
f
"drop database if exists
{
name
}
"
)
tdSql
.
execute
(
f
"create database if not exists
{
name
}
precision 'us'"
)
else
:
tdSql
.
execute
(
f
"drop database if exists
{
name
}
"
)
tdSql
.
execute
(
f
"create database if not exists
{
name
}
precision 'us' update 1"
)
tdSql
.
execute
(
f
'use
{
name
}
'
)
def
timeTrans
(
self
,
ts_value
):
print
(
"ts_value---"
,
ts_value
)
if
type
(
ts_value
)
is
int
:
if
ts_value
!=
0
:
ts
=
ts_value
/
1000000
else
:
ts
=
time
.
time
()
elif
type
(
ts_value
)
is
dict
:
if
ts_value
[
"type"
].
lower
()
==
"ns"
:
ts
=
ts_value
[
"value"
]
/
1000000000
elif
ts_value
[
"type"
].
lower
()
==
"us"
:
ts
=
ts_value
[
"value"
]
/
1000000
elif
ts_value
[
"type"
].
lower
()
==
"ms"
:
ts
=
ts_value
[
"value"
]
/
1000
elif
ts_value
[
"type"
].
lower
()
==
"s"
:
ts
=
ts_value
[
"value"
]
/
1
else
:
ts
=
ts_value
[
"value"
]
/
1000000
else
:
print
(
"input ts maybe not right format"
)
ulsec
=
repr
(
ts
).
split
(
'.'
)[
1
][:
6
]
if
len
(
ulsec
)
<
6
and
int
(
ulsec
)
!=
0
:
ulsec
=
int
(
ulsec
)
*
(
10
**
(
6
-
len
(
ulsec
)))
elif
int
(
ulsec
)
==
0
:
ulsec
*=
6
# * follow two rows added for tsCheckCase
td_ts
=
time
.
strftime
(
"%Y-%m-%d %H:%M:%S"
,
time
.
localtime
(
ts
))
return
td_ts
#td_ts = time.strftime("%Y-%m-%d %H:%M:%S", time.localtime(ts))
td_ts
=
time
.
strftime
(
"%Y-%m-%d %H:%M:%S.{}"
.
format
(
ulsec
),
time
.
localtime
(
ts
))
return
td_ts
def
dateToTs
(
self
,
datetime_input
):
return
int
(
time
.
mktime
(
time
.
strptime
(
datetime_input
,
"%Y-%m-%d %H:%M:%S.%f"
)))
def
getTdTypeValue
(
self
,
value
):
if
value
.
endswith
(
"i8"
):
td_type
=
"TINYINT"
td_tag_value
=
''
.
join
(
list
(
value
)[:
-
2
])
elif
value
.
endswith
(
"i16"
):
td_type
=
"SMALLINT"
td_tag_value
=
''
.
join
(
list
(
value
)[:
-
3
])
elif
value
.
endswith
(
"i32"
):
td_type
=
"INT"
td_tag_value
=
''
.
join
(
list
(
value
)[:
-
3
])
elif
value
.
endswith
(
"i64"
):
td_type
=
"BIGINT"
td_tag_value
=
''
.
join
(
list
(
value
)[:
-
3
])
elif
value
.
endswith
(
"u64"
):
td_type
=
"BIGINT UNSIGNED"
td_tag_value
=
''
.
join
(
list
(
value
)[:
-
3
])
elif
value
.
endswith
(
"f32"
):
td_type
=
"FLOAT"
td_tag_value
=
''
.
join
(
list
(
value
)[:
-
3
])
td_tag_value
=
'{}'
.
format
(
np
.
float32
(
td_tag_value
))
elif
value
.
endswith
(
"f64"
):
td_type
=
"DOUBLE"
td_tag_value
=
''
.
join
(
list
(
value
)[:
-
3
])
elif
value
.
startswith
(
'L"'
):
td_type
=
"NCHAR"
td_tag_value
=
''
.
join
(
list
(
value
)[
2
:
-
1
])
elif
value
.
startswith
(
'"'
)
and
value
.
endswith
(
'"'
):
td_type
=
"BINARY"
td_tag_value
=
''
.
join
(
list
(
value
)[
1
:
-
1
])
elif
value
.
lower
()
==
"t"
or
value
==
"true"
or
value
==
"True"
or
value
==
"TRUE"
:
td_type
=
"BOOL"
td_tag_value
=
"True"
elif
value
.
lower
()
==
"f"
or
value
==
"false"
or
value
==
"False"
or
value
==
"FALSE"
:
td_type
=
"BOOL"
td_tag_value
=
"False"
else
:
td_type
=
"FLOAT"
td_tag_value
=
value
return
td_type
,
td_tag_value
def
typeTrans
(
self
,
type_list
):
type_num_list
=
[]
for
tp
in
type_list
:
if
type
(
tp
)
is
dict
:
tp
=
tp
[
'type'
]
if
tp
.
upper
()
==
"TIMESTAMP"
:
type_num_list
.
append
(
9
)
elif
tp
.
upper
()
==
"BOOL"
:
type_num_list
.
append
(
1
)
elif
tp
.
upper
()
==
"TINYINT"
:
type_num_list
.
append
(
2
)
elif
tp
.
upper
()
==
"SMALLINT"
:
type_num_list
.
append
(
3
)
elif
tp
.
upper
()
==
"INT"
:
type_num_list
.
append
(
4
)
elif
tp
.
upper
()
==
"BIGINT"
:
type_num_list
.
append
(
5
)
elif
tp
.
upper
()
==
"FLOAT"
:
type_num_list
.
append
(
6
)
elif
tp
.
upper
()
==
"DOUBLE"
:
type_num_list
.
append
(
7
)
elif
tp
.
upper
()
==
"BINARY"
:
type_num_list
.
append
(
8
)
elif
tp
.
upper
()
==
"NCHAR"
:
type_num_list
.
append
(
10
)
elif
tp
.
upper
()
==
"BIGINT UNSIGNED"
:
type_num_list
.
append
(
14
)
return
type_num_list
def
inputHandle
(
self
,
input_json
):
stb_name
=
input_json
[
"metric"
]
stb_tag_dict
=
input_json
[
"tags"
]
stb_col_dict
=
input_json
[
"value"
]
ts_value
=
self
.
timeTrans
(
input_json
[
"timestamp"
])
tag_name_list
=
[]
tag_value_list
=
[]
td_tag_value_list
=
[]
td_tag_type_list
=
[]
col_name_list
=
[]
col_value_list
=
[]
td_col_value_list
=
[]
td_col_type_list
=
[]
# handle tag
for
key
,
value
in
stb_tag_dict
.
items
():
if
"id"
in
key
.
lower
():
tb_name
=
value
else
:
tag_value_list
.
append
(
str
(
value
[
"value"
]))
td_tag_value_list
.
append
(
str
(
value
[
"value"
]))
tag_name_list
.
append
(
key
)
td_tag_type_list
.
append
(
value
[
"type"
].
upper
())
tb_name
=
""
# handle col
if
type
(
stb_col_dict
)
is
dict
:
if
stb_col_dict
[
"type"
].
lower
()
==
"bool"
:
bool_value
=
f
'
{
stb_col_dict
[
"value"
]
}
'
col_value_list
.
append
(
bool_value
)
td_col_type_list
.
append
(
stb_col_dict
[
"type"
].
upper
())
col_name_list
.
append
(
"value"
)
td_col_value_list
.
append
(
stb_col_dict
[
"value"
])
else
:
col_value_list
.
append
(
stb_col_dict
[
"value"
])
td_col_type_list
.
append
(
stb_col_dict
[
"type"
].
upper
())
col_name_list
.
append
(
"value"
)
td_col_value_list
.
append
(
stb_col_dict
[
"value"
])
else
:
col_name_list
.
append
(
"value"
)
col_value_list
.
append
(
str
(
stb_col_dict
))
td_col_value_list
.
append
(
str
(
stb_col_dict
))
td_col_type_list
.
append
(
tdCom
.
typeof
(
stb_col_dict
).
upper
())
final_field_list
=
[]
final_field_list
.
extend
(
col_name_list
)
final_field_list
.
extend
(
tag_name_list
)
final_type_list
=
[]
final_type_list
.
append
(
"TIMESTAMP"
)
final_type_list
.
extend
(
td_col_type_list
)
final_type_list
.
extend
(
td_tag_type_list
)
final_type_list
=
self
.
typeTrans
(
final_type_list
)
final_value_list
=
[]
final_value_list
.
append
(
ts_value
)
final_value_list
.
extend
(
td_col_value_list
)
final_value_list
.
extend
(
td_tag_value_list
)
return
final_value_list
,
final_field_list
,
final_type_list
,
stb_name
,
tb_name
def
genTsColValue
(
self
,
value
,
t_type
=
None
):
if
t_type
==
None
:
ts_col_value
=
value
else
:
ts_col_value
=
{
"value"
:
value
,
"type"
:
t_type
}
return
ts_col_value
def
genTagValue
(
self
,
t0_type
=
"bool"
,
t0_value
=
""
,
t1_type
=
"tinyint"
,
t1_value
=
127
,
t2_type
=
"smallint"
,
t2_value
=
32767
,
t3_type
=
"int"
,
t3_value
=
2147483647
,
t4_type
=
"bigint"
,
t4_value
=
9223372036854775807
,
t5_type
=
"float"
,
t5_value
=
11.12345027923584
,
t6_type
=
"double"
,
t6_value
=
22.123456789
,
t7_type
=
"binary"
,
t7_value
=
"binaryTagValue"
,
t8_type
=
"nchar"
,
t8_value
=
"ncharTagValue"
,):
if
t0_value
==
""
:
t0_value
=
random
.
choice
([
True
,
False
])
tag_value
=
{
"t0"
:
{
"value"
:
t0_value
,
"type"
:
t0_type
},
"t1"
:
{
"value"
:
t1_value
,
"type"
:
t1_type
},
"t2"
:
{
"value"
:
t2_value
,
"type"
:
t2_type
},
"t3"
:
{
"value"
:
t3_value
,
"type"
:
t3_type
},
"t4"
:
{
"value"
:
t4_value
,
"type"
:
t4_type
},
"t5"
:
{
"value"
:
t5_value
,
"type"
:
t5_type
},
"t6"
:
{
"value"
:
t6_value
,
"type"
:
t6_type
},
"t7"
:
{
"value"
:
t7_value
,
"type"
:
t7_type
},
"t8"
:
{
"value"
:
t8_value
,
"type"
:
t8_type
}
}
return
tag_value
def
genFullTypeJson
(
self
,
ts_value
=
""
,
col_value
=
""
,
tag_value
=
""
,
stb_name
=
""
,
tb_name
=
""
,
id_noexist_tag
=
None
,
id_change_tag
=
None
,
id_upper_tag
=
None
,
id_double_tag
=
None
,
t_add_tag
=
None
,
t_mul_tag
=
None
,
c_multi_tag
=
None
,
c_blank_tag
=
None
,
t_blank_tag
=
None
,
chinese_tag
=
None
,
multi_field_tag
=
None
):
if
stb_name
==
""
:
stb_name
=
tdCom
.
getLongName
(
len
=
6
,
mode
=
"letters"
)
if
tb_name
==
""
:
tb_name
=
f
'
{
stb_name
}
_
{
random
.
randint
(
0
,
65535
)
}
_
{
random
.
randint
(
0
,
65535
)
}
'
if
ts_value
==
""
:
ts_value
=
self
.
genTsColValue
(
1626006833639000000
,
"ns"
)
if
col_value
==
""
:
col_value
=
random
.
choice
([
True
,
False
])
if
tag_value
==
""
:
tag_value
=
self
.
genTagValue
()
if
id_upper_tag
is
not
None
:
id
=
"ID"
else
:
id
=
"id"
if
id_noexist_tag
is
None
:
tag_value
[
id
]
=
tb_name
sql_json
=
{
"metric"
:
f
"
{
stb_name
}
"
,
"timestamp"
:
ts_value
,
"value"
:
col_value
,
"tags"
:
tag_value
}
if
id_noexist_tag
is
not
None
:
if
t_add_tag
is
not
None
:
tag_value
[
"t9"
]
=
f
'"
{
tdCom
.
getLongName
(
len
=
6
,
mode
=
"letters"
)
}
"'
sql_json
=
{
"metric"
:
f
"
{
stb_name
}
"
,
"timestamp"
:
ts_value
,
"value"
:
col_value
,
"tags"
:
tag_value
}
if
id_change_tag
is
not
None
:
tag_value
.
pop
(
't8'
)
tag_value
[
"t8"
]
=
{
"value"
:
"nchar"
,
"type"
:
"ncharTagValue"
}
sql_json
=
{
"metric"
:
f
"
{
stb_name
}
"
,
"timestamp"
:
ts_value
,
"value"
:
col_value
,
"tags"
:
tag_value
}
if
id_double_tag
is
not
None
:
tag_value
[
"ID"
]
=
f
'"
{
tb_name
}
_2"'
sql_json
=
{
"metric"
:
f
"
{
stb_name
}
"
,
"timestamp"
:
ts_value
,
"value"
:
col_value
,
"tags"
:
tag_value
}
if
t_add_tag
is
not
None
:
tag_value
[
"t9"
]
=
f
'"
{
tdCom
.
getLongName
(
len
=
6
,
mode
=
"letters"
)
}
"'
sql_json
=
{
"metric"
:
f
"
{
stb_name
}
"
,
"timestamp"
:
ts_value
,
"value"
:
col_value
,
"tags"
:
tag_value
}
if
t_mul_tag
is
not
None
:
tag_value
.
pop
(
't8'
)
sql_json
=
{
"metric"
:
f
"
{
stb_name
}
"
,
"timestamp"
:
ts_value
,
"value"
:
col_value
,
"tags"
:
tag_value
}
if
id_noexist_tag
is
not
None
:
tag_value
.
pop
(
't8'
)
sql_json
=
{
"metric"
:
f
"
{
stb_name
}
"
,
"timestamp"
:
ts_value
,
"value"
:
col_value
,
"tags"
:
tag_value
}
if
c_multi_tag
is
not
None
:
col_value
=
"true,f"
sql_json
=
{
"metric"
:
f
"
{
stb_name
}
"
,
"timestamp"
:
ts_value
,
"value"
:
col_value
,
"tags"
:
tag_value
}
if
t_blank_tag
is
not
None
:
tag_value
=
{
"id"
:
f
'"
{
tdCom
.
getLongName
(
len
=
6
,
mode
=
"letters"
)
}
"'
}
sql_json
=
{
"metric"
:
f
"
{
stb_name
}
"
,
"timestamp"
:
ts_value
,
"value"
:
col_value
,
"tags"
:
tag_value
}
if
chinese_tag
is
not
None
:
tag_value
=
{
"id"
:
"涛思数据"
,
"t0"
:
"涛思数据"
}
sql_json
=
{
"metric"
:
f
"
{
stb_name
}
"
,
"timestamp"
:
ts_value
,
"value"
:
col_value
,
"tags"
:
tag_value
}
if
c_blank_tag
is
not
None
:
sql_json
.
pop
(
"value"
)
if
multi_field_tag
is
not
None
:
sql_json
=
{
"metric"
:
f
"
{
stb_name
}
"
,
"timestamp"
:
ts_value
,
"value"
:
col_value
,
"tags"
:
tag_value
,
"tags"
:
tag_value
}
return
sql_json
,
stb_name
def
genFullTypeSql
(
self
,
stb_name
=
""
,
tb_name
=
""
,
value
=
""
,
t0
=
""
,
t1
=
"127i8"
,
t2
=
"32767i16"
,
t3
=
"2147483647i32"
,
t4
=
"9223372036854775807i64"
,
t5
=
"11.12345f32"
,
t6
=
"22.123456789f64"
,
t7
=
"
\"
binaryTagValue
\"
"
,
t8
=
"L
\"
ncharTagValue
\"
"
,
ts
=
"1626006833639000000ns"
,
id_noexist_tag
=
None
,
id_change_tag
=
None
,
id_upper_tag
=
None
,
id_double_tag
=
None
,
t_add_tag
=
None
,
t_mul_tag
=
None
,
t_multi_tag
=
None
,
c_blank_tag
=
None
,
t_blank_tag
=
None
,
chinese_tag
=
None
,
multi_field_tag
=
None
):
if
stb_name
==
""
:
stb_name
=
tdCom
.
getLongName
(
len
=
6
,
mode
=
"letters"
)
if
tb_name
==
""
:
tb_name
=
f
'
{
stb_name
}
_
{
random
.
randint
(
0
,
65535
)
}
_
{
random
.
randint
(
0
,
65535
)
}
'
if
t0
==
""
:
t0
=
random
.
choice
([
True
,
False
])
if
value
==
""
:
value
=
random
.
choice
([
True
,
False
])
if
id_upper_tag
is
not
None
:
id
=
"ID"
else
:
id
=
"id"
sql_seq
=
f
'
{
stb_name
}
{
ts
}
{
value
}
{
id
}
=
\"
{
tb_name
}
\"
,t0=
{
t0
}
,t1=
{
t1
}
,t2=
{
t2
}
,t3=
{
t3
}
,t4=
{
t4
}
,t5=
{
t5
}
,t6=
{
t6
}
,t7=
{
t7
}
,t8=
{
t8
}
'
if
id_noexist_tag
is
not
None
:
sql_seq
=
f
'
{
stb_name
}
{
ts
}
{
value
}
t0=
{
t0
}
,t1=
{
t1
}
,t2=
{
t2
}
,t3=
{
t3
}
,t4=
{
t4
}
,t5=
{
t5
}
,t6=
{
t6
}
,t7=
{
t7
}
,t8=
{
t8
}
'
if
t_add_tag
is
not
None
:
sql_seq
=
f
'
{
stb_name
}
{
ts
}
{
value
}
t0=
{
t0
}
,t1=
{
t1
}
,t2=
{
t2
}
,t3=
{
t3
}
,t4=
{
t4
}
,t5=
{
t5
}
,t6=
{
t6
}
,t7=
{
t7
}
,t8=
{
t8
}
,t9=
{
t8
}
'
if
id_change_tag
is
not
None
:
sql_seq
=
f
'
{
stb_name
}
{
ts
}
{
value
}
t0=
{
t0
}
,
{
id
}
=
\"
{
tb_name
}
\"
,t1=
{
t1
}
,t2=
{
t2
}
,t3=
{
t3
}
,t4=
{
t4
}
,t5=
{
t5
}
,t6=
{
t6
}
,t7=
{
t7
}
,t8=
{
t8
}
'
if
id_double_tag
is
not
None
:
sql_seq
=
f
'
{
stb_name
}
{
ts
}
{
value
}
{
id
}
=
\"
{
tb_name
}
_1
\"
,t0=
{
t0
}
,t1=
{
t1
}
,
{
id
}
=
\"
{
tb_name
}
_2
\"
,t2=
{
t2
}
,t3=
{
t3
}
,t4=
{
t4
}
,t5=
{
t5
}
,t6=
{
t6
}
,t7=
{
t7
}
,t8=
{
t8
}
'
if
t_add_tag
is
not
None
:
sql_seq
=
f
'
{
stb_name
}
{
ts
}
{
value
}
{
id
}
=
\"
{
tb_name
}
\"
,t0=
{
t0
}
,t1=
{
t1
}
,t2=
{
t2
}
,t3=
{
t3
}
,t4=
{
t4
}
,t5=
{
t5
}
,t6=
{
t6
}
,t7=
{
t7
}
,t8=
{
t8
}
,t11=
{
t1
}
,t10=
{
t8
}
'
if
t_mul_tag
is
not
None
:
sql_seq
=
f
'
{
stb_name
}
{
ts
}
{
value
}
{
id
}
=
\"
{
tb_name
}
\"
,t0=
{
t0
}
,t1=
{
t1
}
,t2=
{
t2
}
,t3=
{
t3
}
,t4=
{
t4
}
,t5=
{
t5
}
,t6=
{
t6
}
'
if
id_noexist_tag
is
not
None
:
sql_seq
=
f
'
{
stb_name
}
{
ts
}
{
value
}
t0=
{
t0
}
,t1=
{
t1
}
,t2=
{
t2
}
,t3=
{
t3
}
,t4=
{
t4
}
,t5=
{
t5
}
,t6=
{
t6
}
'
if
t_multi_tag
is
not
None
:
sql_seq
=
f
'
{
stb_name
}
{
ts
}
{
value
}
,
{
value
}
{
id
}
=
\"
{
tb_name
}
\"
,t0=
{
t0
}
,t1=
{
t1
}
,t2=
{
t2
}
,t3=
{
t3
}
,t4=
{
t4
}
,t5=
{
t5
}
,t6=
{
t6
}
'
if
c_blank_tag
is
not
None
:
sql_seq
=
f
'
{
stb_name
}
{
ts
}
{
id
}
=
\"
{
tb_name
}
\"
,t0=
{
t0
}
,t1=
{
t1
}
,t2=
{
t2
}
,t3=
{
t3
}
,t4=
{
t4
}
,t5=
{
t5
}
,t6=
{
t6
}
,t7=
{
t7
}
,t8=
{
t8
}
'
if
t_blank_tag
is
not
None
:
sql_seq
=
f
'
{
stb_name
}
{
ts
}
{
value
}
{
id
}
=
\"
{
tb_name
}
\"
'
if
chinese_tag
is
not
None
:
sql_seq
=
f
'
{
stb_name
}
{
ts
}
L"涛思数据" t0=
{
t0
}
,t1=L"涛思数据"'
if
multi_field_tag
is
not
None
:
sql_seq
=
f
'
{
stb_name
}
{
ts
}
{
value
}
{
id
}
=
\"
{
tb_name
}
\"
,t0=
{
t0
}
t1=
{
t1
}
'
return
sql_seq
,
stb_name
def
genMulTagColStr
(
self
,
genType
,
count
=
1
):
"""
genType must be tag/col
"""
tag_str
=
""
col_str
=
""
if
genType
==
"tag"
:
for
i
in
range
(
0
,
count
):
if
i
<
(
count
-
1
):
tag_str
+=
f
't
{
i
}
=f,'
else
:
tag_str
+=
f
't
{
i
}
=f'
return
tag_str
if
genType
==
"col"
:
col_str
=
"t"
return
col_str
def
genLongSql
(
self
,
tag_count
):
stb_name
=
tdCom
.
getLongName
(
7
,
mode
=
"letters"
)
tb_name
=
f
'
{
stb_name
}
_1'
tag_str
=
self
.
genMulTagColStr
(
"tag"
,
tag_count
)
col_str
=
self
.
genMulTagColStr
(
"col"
)
ts
=
"1626006833640000000ns"
long_sql
=
stb_name
+
' '
+
ts
+
' '
+
col_str
+
' '
+
f
'id=
\"
{
tb_name
}
\"
'
+
','
+
tag_str
return
long_sql
,
stb_name
def
getNoIdTbName
(
self
,
stb_name
):
query_sql
=
f
"select tbname from
{
stb_name
}
"
tb_name
=
self
.
resHandle
(
query_sql
,
True
)[
0
][
0
]
return
tb_name
def
resHandle
(
self
,
query_sql
,
query_tag
):
tdSql
.
execute
(
'reset query cache'
)
row_info
=
tdSql
.
query
(
query_sql
,
query_tag
)
col_info
=
tdSql
.
getColNameList
(
query_sql
,
query_tag
)
res_row_list
=
[]
sub_list
=
[]
for
row_mem
in
row_info
:
for
i
in
row_mem
:
sub_list
.
append
(
str
(
i
))
res_row_list
.
append
(
sub_list
)
res_field_list_without_ts
=
col_info
[
0
][
1
:]
res_type_list
=
col_info
[
1
]
return
res_row_list
,
res_field_list_without_ts
,
res_type_list
def
resCmp
(
self
,
input_json
,
stb_name
,
query_sql
=
"select * from"
,
condition
=
""
,
ts
=
None
,
id
=
True
,
none_check_tag
=
None
):
expect_list
=
self
.
inputHandle
(
input_json
)
self
.
_conn
.
insert_json_payload
(
json
.
dumps
(
input_json
))
query_sql
=
f
"
{
query_sql
}
{
stb_name
}
{
condition
}
"
res_row_list
,
res_field_list_without_ts
,
res_type_list
=
self
.
resHandle
(
query_sql
,
True
)
if
ts
==
0
:
res_ts
=
self
.
dateToTs
(
res_row_list
[
0
][
0
])
current_time
=
time
.
time
()
if
current_time
-
res_ts
<
60
:
tdSql
.
checkEqual
(
res_row_list
[
0
][
1
:],
expect_list
[
0
][
1
:])
else
:
print
(
"timeout"
)
tdSql
.
checkEqual
(
res_row_list
[
0
],
expect_list
[
0
])
else
:
if
none_check_tag
is
not
None
:
none_index_list
=
[
i
for
i
,
x
in
enumerate
(
res_row_list
[
0
])
if
x
==
"None"
]
none_index_list
.
reverse
()
for
j
in
none_index_list
:
res_row_list
[
0
].
pop
(
j
)
expect_list
[
0
].
pop
(
j
)
tdSql
.
checkEqual
(
res_row_list
[
0
],
expect_list
[
0
])
tdSql
.
checkEqual
(
res_field_list_without_ts
,
expect_list
[
1
])
for
i
in
range
(
len
(
res_type_list
)):
tdSql
.
checkEqual
(
res_type_list
[
i
],
expect_list
[
2
][
i
])
# tdSql.checkEqual(res_type_list, expect_list[2])
def
initCheckCase
(
self
):
"""
normal tags and cols, one for every elm
"""
tdCom
.
cleanTb
()
input_json
,
stb_name
=
self
.
genFullTypeJson
()
self
.
resCmp
(
input_json
,
stb_name
)
def
boolTypeCheckCase
(
self
):
"""
check all normal type
"""
tdCom
.
cleanTb
()
full_type_list
=
[
"f"
,
"F"
,
"false"
,
"False"
,
"t"
,
"T"
,
"true"
,
"True"
]
for
t_type
in
full_type_list
:
input_json_list
=
[
self
.
genFullTypeJson
(
tag_value
=
self
.
genTagValue
(
t0_value
=
t_type
))[
0
],
self
.
genFullTypeJson
(
col_value
=
self
.
genTsColValue
(
value
=
t_type
,
t_type
=
"bool"
))[
0
]]
for
input_json
in
input_json_list
:
try
:
self
.
_conn
.
insert_json_payload
(
json
.
dumps
(
input_json
))
raise
Exception
(
"should not reach here"
)
except
JsonPayloadError
as
err
:
tdSql
.
checkNotEqual
(
err
.
errno
,
0
)
def
symbolsCheckCase
(
self
):
"""
check symbols = `~!@#$%^&*()_-+={[}]\|:;'
\"
,<.>/?
"""
'''
please test :
binary_symbols = '
\"
abcd`~!@#$%^&*()_-{[}]|:;<.>?lfjal"
\'\'
"
\"
'
'''
tdCom
.
cleanTb
()
binary_symbols
=
'"abcd`~!@#$%^&*()_-{[}]|:;<.>?lfjal"'
nchar_symbols
=
binary_symbols
input_sql1
,
stb_name1
=
self
.
genFullTypeJson
(
col_value
=
self
.
genTsColValue
(
value
=
binary_symbols
,
t_type
=
"binary"
),
tag_value
=
self
.
genTagValue
(
t7_value
=
binary_symbols
,
t8_value
=
nchar_symbols
))
input_sql2
,
stb_name2
=
self
.
genFullTypeJson
(
col_value
=
self
.
genTsColValue
(
value
=
nchar_symbols
,
t_type
=
"nchar"
),
tag_value
=
self
.
genTagValue
(
t7_value
=
binary_symbols
,
t8_value
=
nchar_symbols
))
self
.
resCmp
(
input_sql1
,
stb_name1
)
self
.
resCmp
(
input_sql2
,
stb_name2
)
def
tsCheckCase
(
self
):
"""
test ts list --> ["1626006833639000000ns", "1626006833639019us", "1626006833640ms", "1626006834s", "1626006822639022"]
# ! us级时间戳都为0时,数据库中查询显示,但python接口拿到的结果不显示 .000000的情况请确认,目前修改时间处理代码可以通过
"""
tdCom
.
cleanTb
()
ts_list
=
[
"1626006833639000000ns"
,
"1626006833639019us"
,
"1626006833640ms"
,
"1626006834s"
,
"1626006822639022"
,
0
]
for
ts
in
ts_list
:
if
"s"
in
str
(
ts
):
input_json
,
stb_name
=
self
.
genFullTypeJson
(
ts_value
=
self
.
genTsColValue
(
value
=
int
(
tdCom
.
splitNumLetter
(
ts
)[
0
]),
t_type
=
tdCom
.
splitNumLetter
(
ts
)[
1
]))
self
.
resCmp
(
input_json
,
stb_name
,
ts
=
ts
)
else
:
input_json
,
stb_name
=
self
.
genFullTypeJson
(
ts_value
=
self
.
genTsColValue
(
value
=
int
(
ts
),
t_type
=
"us"
))
self
.
resCmp
(
input_json
,
stb_name
,
ts
=
ts
)
if
int
(
ts
)
==
0
:
input_json_list
=
[
self
.
genFullTypeJson
(
ts_value
=
self
.
genTsColValue
(
value
=
int
(
ts
),
t_type
=
""
)),
self
.
genFullTypeJson
(
ts_value
=
self
.
genTsColValue
(
value
=
int
(
ts
),
t_type
=
"ns"
)),
self
.
genFullTypeJson
(
ts_value
=
self
.
genTsColValue
(
value
=
int
(
ts
),
t_type
=
"us"
)),
self
.
genFullTypeJson
(
ts_value
=
self
.
genTsColValue
(
value
=
int
(
ts
),
t_type
=
"ms"
)),
self
.
genFullTypeJson
(
ts_value
=
self
.
genTsColValue
(
value
=
int
(
ts
),
t_type
=
"s"
))]
for
input_json
in
input_json_list
:
self
.
resCmp
(
input_json
[
0
],
input_json
[
1
],
ts
=
ts
)
else
:
input_json
=
self
.
genFullTypeJson
(
ts_value
=
self
.
genTsColValue
(
value
=
int
(
ts
),
t_type
=
""
))[
0
]
try
:
self
.
_conn
.
insert_json_payload
(
json
.
dumps
(
input_json
))
raise
Exception
(
"should not reach here"
)
except
JsonPayloadError
as
err
:
tdSql
.
checkNotEqual
(
err
.
errno
,
0
)
def
idSeqCheckCase
(
self
):
"""
check id.index in tags
eg: t0=**,id=**,t1=**
"""
tdCom
.
cleanTb
()
input_json
,
stb_name
=
self
.
genFullTypeJson
(
id_change_tag
=
True
)
self
.
resCmp
(
input_json
,
stb_name
)
def
idUpperCheckCase
(
self
):
"""
check id param
eg: id and ID
"""
tdCom
.
cleanTb
()
input_json
,
stb_name
=
self
.
genFullTypeJson
(
id_upper_tag
=
True
)
self
.
resCmp
(
input_json
,
stb_name
)
input_json
,
stb_name
=
self
.
genFullTypeJson
(
id_change_tag
=
True
,
id_upper_tag
=
True
)
self
.
resCmp
(
input_json
,
stb_name
)
def
noIdCheckCase
(
self
):
"""
id not exist
"""
tdCom
.
cleanTb
()
input_json
,
stb_name
=
self
.
genFullTypeJson
(
id_noexist_tag
=
True
)
self
.
resCmp
(
input_json
,
stb_name
)
query_sql
=
f
"select tbname from
{
stb_name
}
"
res_row_list
=
self
.
resHandle
(
query_sql
,
True
)[
0
]
if
len
(
res_row_list
[
0
][
0
])
>
0
:
tdSql
.
checkColNameList
(
res_row_list
,
res_row_list
)
else
:
tdSql
.
checkColNameList
(
res_row_list
,
"please check noIdCheckCase"
)
def
maxColTagCheckCase
(
self
):
"""
max tag count is 128
"""
for
input_json
in
[
self
.
genLongSql
(
128
)[
0
]]:
tdCom
.
cleanTb
()
self
.
_conn
.
insert_json_payload
([
input_json
])
for
input_json
in
[
self
.
genLongSql
(
129
)[
0
]]:
tdCom
.
cleanTb
()
try
:
self
.
_conn
.
insert_json_payload
([
input_json
])
except
JsonPayloadError
:
pass
def
idIllegalNameCheckCase
(
self
):
"""
test illegal id name
mix "`~!@#$¥%^&*()-+={}|[]、「」【】\:;《》<>?"
"""
tdCom
.
cleanTb
()
rstr
=
list
(
"`~!@#$¥%^&*()-+={}|[]、「」【】\:;《》<>?"
)
for
i
in
rstr
:
input_json
=
self
.
genFullTypeJson
(
tb_name
=
f
"
\"
aaa
{
i
}
bbb
\"
"
)[
0
]
try
:
self
.
_conn
.
insert_json_payload
([
input_json
])
except
JsonPayloadError
:
pass
def
idStartWithNumCheckCase
(
self
):
"""
id is start with num
"""
tdCom
.
cleanTb
()
input_json
=
self
.
genFullTypeJson
(
tb_name
=
f
"
\"
1aaabbb
\"
"
)[
0
]
try
:
self
.
_conn
.
insert_json_payload
([
input_json
])
except
JsonPayloadError
:
pass
def
nowTsCheckCase
(
self
):
"""
check now unsupported
"""
tdCom
.
cleanTb
()
input_json
=
self
.
genFullTypeJson
(
ts
=
"now"
)[
0
]
try
:
self
.
_conn
.
insert_json_payload
([
input_json
])
except
JsonPayloadError
:
pass
def
dateFormatTsCheckCase
(
self
):
"""
check date format ts unsupported
"""
tdCom
.
cleanTb
()
input_json
=
self
.
genFullTypeJson
(
ts
=
"2021-07-21\ 19:01:46.920"
)[
0
]
try
:
self
.
_conn
.
insert_json_payload
([
input_json
])
except
JsonPayloadError
:
pass
def
illegalTsCheckCase
(
self
):
"""
check ts format like 16260068336390us19
"""
tdCom
.
cleanTb
()
input_json
=
self
.
genFullTypeJson
(
ts
=
"16260068336390us19"
)[
0
]
try
:
self
.
_conn
.
insert_json_payload
([
input_json
])
except
JsonPayloadError
:
pass
def
tagValueLengthCheckCase
(
self
):
"""
check full type tag value limit
"""
tdCom
.
cleanTb
()
# i8
for
t1
in
[
"-127i8"
,
"127i8"
]:
input_json
,
stb_name
=
self
.
genFullTypeJson
(
t1
=
t1
)
self
.
resCmp
(
input_json
,
stb_name
)
for
t1
in
[
"-128i8"
,
"128i8"
]:
input_json
=
self
.
genFullTypeJson
(
t1
=
t1
)[
0
]
try
:
self
.
_conn
.
insert_json_payload
([
input_json
])
except
JsonPayloadError
:
pass
#i16
for
t2
in
[
"-32767i16"
,
"32767i16"
]:
input_json
,
stb_name
=
self
.
genFullTypeJson
(
t2
=
t2
)
self
.
resCmp
(
input_json
,
stb_name
)
for
t2
in
[
"-32768i16"
,
"32768i16"
]:
input_json
=
self
.
genFullTypeJson
(
t2
=
t2
)[
0
]
try
:
self
.
_conn
.
insert_json_payload
([
input_json
])
except
JsonPayloadError
:
pass
#i32
for
t3
in
[
"-2147483647i32"
,
"2147483647i32"
]:
input_json
,
stb_name
=
self
.
genFullTypeJson
(
t3
=
t3
)
self
.
resCmp
(
input_json
,
stb_name
)
for
t3
in
[
"-2147483648i32"
,
"2147483648i32"
]:
input_json
=
self
.
genFullTypeJson
(
t3
=
t3
)[
0
]
try
:
self
.
_conn
.
insert_json_payload
([
input_json
])
except
JsonPayloadError
:
pass
#i64
for
t4
in
[
"-9223372036854775807i64"
,
"9223372036854775807i64"
]:
input_json
,
stb_name
=
self
.
genFullTypeJson
(
t4
=
t4
)
self
.
resCmp
(
input_json
,
stb_name
)
for
t4
in
[
"-9223372036854775808i64"
,
"9223372036854775808i64"
]:
input_json
=
self
.
genFullTypeJson
(
t4
=
t4
)[
0
]
try
:
self
.
_conn
.
insert_json_payload
([
input_json
])
except
JsonPayloadError
:
pass
# f32
for
t5
in
[
f
"
{
-
3.4028234663852885981170418348451692544
*
(
10
**
38
)
}
f32"
,
f
"
{
3.4028234663852885981170418348451692544
*
(
10
**
38
)
}
f32"
]:
input_json
,
stb_name
=
self
.
genFullTypeJson
(
t5
=
t5
)
self
.
resCmp
(
input_json
,
stb_name
)
# * limit set to 4028234664*(10**38)
for
t5
in
[
f
"
{
-
3.4028234664
*
(
10
**
38
)
}
f32"
,
f
"
{
3.4028234664
*
(
10
**
38
)
}
f32"
]:
input_json
=
self
.
genFullTypeJson
(
t5
=
t5
)[
0
]
try
:
self
.
_conn
.
insert_json_payload
([
input_json
])
raise
Exception
(
"should not reach here"
)
except
JsonPayloadError
as
err
:
tdSql
.
checkNotEqual
(
err
.
errno
,
0
)
# f64
for
t6
in
[
f
'
{
-
1.79769
*
(
10
**
308
)
}
f64'
,
f
'
{
-
1.79769
*
(
10
**
308
)
}
f64'
]:
input_json
,
stb_name
=
self
.
genFullTypeJson
(
t6
=
t6
)
self
.
resCmp
(
input_json
,
stb_name
)
# * limit set to 1.797693134862316*(10**308)
for
t6
in
[
f
'
{
-
1.797693134862316
*
(
10
**
308
)
}
f64'
,
f
'
{
-
1.797693134862316
*
(
10
**
308
)
}
f64'
]:
input_json
=
self
.
genFullTypeJson
(
t6
=
t6
)[
0
]
try
:
self
.
_conn
.
insert_json_payload
([
input_json
])
raise
Exception
(
"should not reach here"
)
except
JsonPayloadError
as
err
:
tdSql
.
checkNotEqual
(
err
.
errno
,
0
)
# binary
stb_name
=
tdCom
.
getLongName
(
7
,
"letters"
)
input_json
=
f
'
{
stb_name
}
1626006833639000000ns t t0=t,t1="
{
tdCom
.
getLongName
(
16374
,
"letters"
)
}
"'
self
.
_conn
.
insert_json_payload
([
input_json
])
input_json
=
f
'
{
stb_name
}
1626006833639000000ns t t0=t,t1="
{
tdCom
.
getLongName
(
16375
,
"letters"
)
}
"'
try
:
self
.
_conn
.
insert_json_payload
([
input_json
])
raise
Exception
(
"should not reach here"
)
except
JsonPayloadError
as
err
:
pass
# nchar
# * legal nchar could not be larger than 16374/4
stb_name
=
tdCom
.
getLongName
(
7
,
"letters"
)
input_json
=
f
'
{
stb_name
}
1626006833639000000ns t t0=t,t1=L"
{
tdCom
.
getLongName
(
4093
,
"letters"
)
}
"'
self
.
_conn
.
insert_json_payload
([
input_json
])
input_json
=
f
'
{
stb_name
}
1626006833639000000ns t t0=t,t1=L"
{
tdCom
.
getLongName
(
4094
,
"letters"
)
}
"'
try
:
self
.
_conn
.
insert_json_payload
([
input_json
])
raise
Exception
(
"should not reach here"
)
except
JsonPayloadError
as
err
:
tdSql
.
checkNotEqual
(
err
.
errno
,
0
)
def
colValueLengthCheckCase
(
self
):
"""
check full type col value limit
"""
tdCom
.
cleanTb
()
# i8
for
value
in
[
"-127i8"
,
"127i8"
]:
input_json
,
stb_name
=
self
.
genFullTypeJson
(
value
=
value
)
self
.
resCmp
(
input_json
,
stb_name
)
tdCom
.
cleanTb
()
for
value
in
[
"-128i8"
,
"128i8"
]:
input_json
=
self
.
genFullTypeJson
(
value
=
value
)[
0
]
try
:
self
.
_conn
.
insert_json_payload
([
input_json
])
raise
Exception
(
"should not reach here"
)
except
JsonPayloadError
as
err
:
tdSql
.
checkNotEqual
(
err
.
errno
,
0
)
# i16
tdCom
.
cleanTb
()
for
value
in
[
"-32767i16"
]:
input_json
,
stb_name
=
self
.
genFullTypeJson
(
value
=
value
)
self
.
resCmp
(
input_json
,
stb_name
)
tdCom
.
cleanTb
()
for
value
in
[
"-32768i16"
,
"32768i16"
]:
input_json
=
self
.
genFullTypeJson
(
value
=
value
)[
0
]
try
:
self
.
_conn
.
insert_json_payload
([
input_json
])
raise
Exception
(
"should not reach here"
)
except
JsonPayloadError
as
err
:
tdSql
.
checkNotEqual
(
err
.
errno
,
0
)
# i32
tdCom
.
cleanTb
()
for
value
in
[
"-2147483647i32"
]:
input_json
,
stb_name
=
self
.
genFullTypeJson
(
value
=
value
)
self
.
resCmp
(
input_json
,
stb_name
)
tdCom
.
cleanTb
()
for
value
in
[
"-2147483648i32"
,
"2147483648i32"
]:
input_json
=
self
.
genFullTypeJson
(
value
=
value
)[
0
]
try
:
self
.
_conn
.
insert_json_payload
([
input_json
])
raise
Exception
(
"should not reach here"
)
except
JsonPayloadError
as
err
:
tdSql
.
checkNotEqual
(
err
.
errno
,
0
)
# i64
tdCom
.
cleanTb
()
for
value
in
[
"-9223372036854775807i64"
]:
input_json
,
stb_name
=
self
.
genFullTypeJson
(
value
=
value
)
self
.
resCmp
(
input_json
,
stb_name
)
tdCom
.
cleanTb
()
for
value
in
[
"-9223372036854775808i64"
,
"9223372036854775808i64"
]:
input_json
=
self
.
genFullTypeJson
(
value
=
value
)[
0
]
try
:
self
.
_conn
.
insert_json_payload
([
input_json
])
raise
Exception
(
"should not reach here"
)
except
JsonPayloadError
as
err
:
tdSql
.
checkNotEqual
(
err
.
errno
,
0
)
# f32
tdCom
.
cleanTb
()
for
value
in
[
f
"
{
-
3.4028234663852885981170418348451692544
*
(
10
**
38
)
}
f32"
,
f
"
{
3.4028234663852885981170418348451692544
*
(
10
**
38
)
}
f32"
]:
input_json
,
stb_name
=
self
.
genFullTypeJson
(
value
=
value
)
self
.
resCmp
(
input_json
,
stb_name
)
# * limit set to 4028234664*(10**38)
tdCom
.
cleanTb
()
for
value
in
[
f
"
{
-
3.4028234664
*
(
10
**
38
)
}
f32"
,
f
"
{
3.4028234664
*
(
10
**
38
)
}
f32"
]:
input_json
=
self
.
genFullTypeJson
(
value
=
value
)[
0
]
try
:
self
.
_conn
.
insert_json_payload
([
input_json
])
raise
Exception
(
"should not reach here"
)
except
JsonPayloadError
as
err
:
tdSql
.
checkNotEqual
(
err
.
errno
,
0
)
# f64
tdCom
.
cleanTb
()
for
value
in
[
f
'
{
-
1.79769313486231570814527423731704356798070567525844996598917476803157260780
*
(
10
**
308
)
}
f64'
,
f
'
{
-
1.79769313486231570814527423731704356798070567525844996598917476803157260780
*
(
10
**
308
)
}
f64'
]:
input_json
,
stb_name
=
self
.
genFullTypeJson
(
value
=
value
)
self
.
resCmp
(
input_json
,
stb_name
)
# * limit set to 1.797693134862316*(10**308)
tdCom
.
cleanTb
()
for
value
in
[
f
'
{
-
1.797693134862316
*
(
10
**
308
)
}
f64'
,
f
'
{
-
1.797693134862316
*
(
10
**
308
)
}
f64'
]:
input_json
=
self
.
genFullTypeJson
(
value
=
value
)[
0
]
try
:
self
.
_conn
.
insert_json_payload
([
input_json
])
raise
Exception
(
"should not reach here"
)
except
JsonPayloadError
as
err
:
tdSql
.
checkNotEqual
(
err
.
errno
,
0
)
# # binary
tdCom
.
cleanTb
()
stb_name
=
tdCom
.
getLongName
(
7
,
"letters"
)
input_json
=
f
'
{
stb_name
}
1626006833639000000ns "
{
tdCom
.
getLongName
(
16374
,
"letters"
)
}
" t0=t'
self
.
_conn
.
insert_json_payload
([
input_json
])
tdCom
.
cleanTb
()
input_json
=
f
'
{
stb_name
}
1626006833639000000ns "
{
tdCom
.
getLongName
(
16375
,
"letters"
)
}
" t0=t'
try
:
self
.
_conn
.
insert_json_payload
([
input_json
])
raise
Exception
(
"should not reach here"
)
except
JsonPayloadError
as
err
:
tdSql
.
checkNotEqual
(
err
.
errno
,
0
)
# nchar
# * legal nchar could not be larger than 16374/4
tdCom
.
cleanTb
()
stb_name
=
tdCom
.
getLongName
(
7
,
"letters"
)
input_json
=
f
'
{
stb_name
}
1626006833639000000ns L"
{
tdCom
.
getLongName
(
4093
,
"letters"
)
}
" t0=t'
self
.
_conn
.
insert_json_payload
([
input_json
])
tdCom
.
cleanTb
()
input_json
=
f
'
{
stb_name
}
1626006833639000000ns L"
{
tdCom
.
getLongName
(
4094
,
"letters"
)
}
" t0=t'
try
:
self
.
_conn
.
insert_json_payload
([
input_json
])
raise
Exception
(
"should not reach here"
)
except
JsonPayloadError
as
err
:
tdSql
.
checkNotEqual
(
err
.
errno
,
0
)
def
tagColIllegalValueCheckCase
(
self
):
"""
test illegal tag col value
"""
tdCom
.
cleanTb
()
# bool
for
i
in
[
"TrUe"
,
"tRue"
,
"trUe"
,
"truE"
,
"FalsE"
,
"fAlse"
,
"faLse"
,
"falSe"
,
"falsE"
]:
input_sql1
=
self
.
genFullTypeJson
(
t0
=
i
)[
0
]
try
:
self
.
_conn
.
insert_json_payload
([
input_sql1
])
raise
Exception
(
"should not reach here"
)
except
JsonPayloadError
as
err
:
tdSql
.
checkNotEqual
(
err
.
errno
,
0
)
input_sql2
=
self
.
genFullTypeJson
(
value
=
i
)[
0
]
try
:
self
.
_conn
.
insert_json_payload
([
input_sql2
])
raise
Exception
(
"should not reach here"
)
except
JsonPayloadError
as
err
:
tdSql
.
checkNotEqual
(
err
.
errno
,
0
)
# i8 i16 i32 i64 f32 f64
for
input_json
in
[
self
.
genFullTypeJson
(
t1
=
"1s2i8"
)[
0
],
self
.
genFullTypeJson
(
t2
=
"1s2i16"
)[
0
],
self
.
genFullTypeJson
(
t3
=
"1s2i32"
)[
0
],
self
.
genFullTypeJson
(
t4
=
"1s2i64"
)[
0
],
self
.
genFullTypeJson
(
t5
=
"11.1s45f32"
)[
0
],
self
.
genFullTypeJson
(
t6
=
"11.1s45f64"
)[
0
],
]:
try
:
self
.
_conn
.
insert_json_payload
([
input_json
])
raise
Exception
(
"should not reach here"
)
except
JsonPayloadError
as
err
:
tdSql
.
checkNotEqual
(
err
.
errno
,
0
)
# check binary and nchar blank
input_sql1
=
f
'
{
tdCom
.
getLongName
(
7
,
"letters"
)
}
1626006833639000000ns "abc aaa" t0=t'
input_sql2
=
f
'
{
tdCom
.
getLongName
(
7
,
"letters"
)
}
1626006833639000000ns L"abc aaa" t0=t'
input_sql3
=
f
'
{
tdCom
.
getLongName
(
7
,
"letters"
)
}
1626006833639000000ns t t0="abc aaa"'
input_sql4
=
f
'
{
tdCom
.
getLongName
(
7
,
"letters"
)
}
1626006833639000000ns t t0=L"abc aaa"'
for
input_json
in
[
input_sql1
,
input_sql2
,
input_sql3
,
input_sql4
]:
try
:
self
.
_conn
.
insert_json_payload
([
input_json
])
except
JsonPayloadError
as
err
:
pass
# check accepted binary and nchar symbols
# # * ~!@#$¥%^&*()-+={}|[]、「」:;
for
symbol
in
list
(
'~!@#$¥%^&*()-+={}|[]、「」:;'
):
input_sql1
=
f
'
{
tdCom
.
getLongName
(
7
,
"letters"
)
}
1626006833639000000ns "abc
{
symbol
}
aaa" t0=t'
input_sql2
=
f
'
{
tdCom
.
getLongName
(
7
,
"letters"
)
}
1626006833639000000ns t t0=t,t1="abc
{
symbol
}
aaa"'
self
.
_conn
.
insert_json_payload
([
input_sql1
])
self
.
_conn
.
insert_json_payload
([
input_sql2
])
def
blankCheckCase
(
self
):
'''
check blank case
'''
tdCom
.
cleanTb
()
input_sql_list
=
[
f
'
{
tdCom
.
getLongName
(
7
,
"letters"
)
}
{
tdCom
.
getLongName
(
7
,
"letters"
)
}
1626006833639000000ns "abcaaa" t0=t'
,
f
'
{
tdCom
.
getLongName
(
7
,
"letters"
)
}
16260068336 39000000ns L"bcdaaa" t1=f'
,
f
'
{
tdCom
.
getLongName
(
7
,
"letters"
)
}
1626006833639000000ns t t0="abc aaa"'
,
f
'
{
tdCom
.
getLongName
(
7
,
"letters"
)
}
1626006833639000000ns t t0=L"abc aaa"'
,
f
'
{
tdCom
.
getLongName
(
7
,
"letters"
)
}
1626006833639000000ns "abc aaa" t0=L"abcaaa"'
,
f
'
{
tdCom
.
getLongName
(
7
,
"letters"
)
}
1626006833639000000ns L"abc aaa" t0=L"abcaaa"'
,
f
'
{
tdCom
.
getLongName
(
7
,
"letters"
)
}
1626006833639000000ns L"abaaa" t0=L"abcaaa1"'
,
f
'
{
tdCom
.
getLongName
(
7
,
"letters"
)
}
1626006833639000000ns L"abaaa" t0=L"abcaaa2"'
,
f
'
{
tdCom
.
getLongName
(
7
,
"letters"
)
}
1626006833639000000ns L"abaaa" t0=t t1="abc t2="taa""'
,
f
'
{
tdCom
.
getLongName
(
7
,
"letters"
)
}
1626006833639000000ns L"abaaa" t0=L"abcaaa3"'
]
for
input_json
in
input_sql_list
:
try
:
self
.
_conn
.
insert_json_payload
([
input_json
])
raise
Exception
(
"should not reach here"
)
except
JsonPayloadError
as
err
:
tdSql
.
checkNotEqual
(
err
.
errno
,
0
)
def
duplicateIdTagColInsertCheckCase
(
self
):
"""
check duplicate Id Tag Col
"""
tdCom
.
cleanTb
()
input_sql_id
=
self
.
genFullTypeJson
(
id_double_tag
=
True
)[
0
]
try
:
self
.
_conn
.
insert_json_payload
([
input_sql_id
])
raise
Exception
(
"should not reach here"
)
except
JsonPayloadError
as
err
:
tdSql
.
checkNotEqual
(
err
.
errno
,
0
)
input_json
=
self
.
genFullTypeJson
()[
0
]
input_sql_tag
=
input_json
.
replace
(
"t5"
,
"t6"
)
try
:
self
.
_conn
.
insert_json_payload
([
input_sql_tag
])
raise
Exception
(
"should not reach here"
)
except
JsonPayloadError
as
err
:
tdSql
.
checkNotEqual
(
err
.
errno
,
0
)
##### stb exist #####
def
noIdStbExistCheckCase
(
self
):
"""
case no id when stb exist
"""
print
(
"noIdStbExistCheckCase"
)
tdCom
.
cleanTb
()
input_json
,
stb_name
=
self
.
genFullTypeJson
(
tb_name
=
"sub_table_0123456"
,
t0
=
"f"
,
value
=
"f"
)
self
.
resCmp
(
input_json
,
stb_name
)
input_json
,
stb_name
=
self
.
genFullTypeJson
(
stb_name
=
stb_name
,
id_noexist_tag
=
True
,
t0
=
"f"
,
value
=
"f"
)
self
.
resCmp
(
input_json
,
stb_name
,
condition
=
'where tbname like "t_%"'
)
tdSql
.
query
(
f
"select * from
{
stb_name
}
"
)
tdSql
.
checkRows
(
2
)
# TODO cover other case
def
duplicateInsertExistCheckCase
(
self
):
"""
check duplicate insert when stb exist
"""
tdCom
.
cleanTb
()
input_json
,
stb_name
=
self
.
genFullTypeJson
()
self
.
resCmp
(
input_json
,
stb_name
)
self
.
_conn
.
insert_json_payload
([
input_json
])
self
.
resCmp
(
input_json
,
stb_name
)
def
tagColBinaryNcharLengthCheckCase
(
self
):
"""
check length increase
"""
print
(
"tagColBinaryNcharLengthCheckCase"
)
tdCom
.
cleanTb
()
input_json
,
stb_name
=
self
.
genFullTypeJson
()
self
.
resCmp
(
input_json
,
stb_name
)
tb_name
=
tdCom
.
getLongName
(
5
,
"letters"
)
input_json
,
stb_name
=
self
.
genFullTypeJson
(
stb_name
=
stb_name
,
tb_name
=
tb_name
,
t7
=
"
\"
binaryTagValuebinaryTagValue
\"
"
,
t8
=
"L
\"
ncharTagValuencharTagValue
\"
"
)
self
.
resCmp
(
input_json
,
stb_name
,
condition
=
f
'where tbname like "
{
tb_name
}
"'
)
def
tagColAddDupIDCheckCase
(
self
):
"""
check tag count add, stb and tb duplicate
* tag: alter table ...
* col: when update==0 and ts is same, unchange
* so this case tag&&value will be added,
* col is added without value when update==0
* col is added with value when update==1
"""
print
(
"tagColAddDupIDCheckCase"
)
tdCom
.
cleanTb
()
tb_name
=
tdCom
.
getLongName
(
7
,
"letters"
)
for
db_update_tag
in
[
0
,
1
]:
if
db_update_tag
==
1
:
self
.
createDb
(
"test_update"
,
db_update_tag
=
db_update_tag
)
input_json
,
stb_name
=
self
.
genFullTypeJson
(
tb_name
=
tb_name
,
t0
=
"f"
,
value
=
"f"
)
self
.
resCmp
(
input_json
,
stb_name
)
self
.
genFullTypeJson
(
stb_name
=
stb_name
,
tb_name
=
tb_name
,
t0
=
"f"
,
value
=
"f"
,
t_add_tag
=
True
)
if
db_update_tag
==
1
:
self
.
resCmp
(
input_json
,
stb_name
,
condition
=
f
'where tbname like "
{
tb_name
}
"'
)
else
:
self
.
resCmp
(
input_json
,
stb_name
,
condition
=
f
'where tbname like "
{
tb_name
}
"'
,
none_check_tag
=
True
)
self
.
createDb
()
def
tagColAddCheckCase
(
self
):
"""
check tag count add
"""
print
(
"tagColAddCheckCase"
)
tdCom
.
cleanTb
()
tb_name
=
tdCom
.
getLongName
(
7
,
"letters"
)
input_json
,
stb_name
=
self
.
genFullTypeJson
(
tb_name
=
tb_name
,
t0
=
"f"
,
value
=
"f"
)
self
.
resCmp
(
input_json
,
stb_name
)
tb_name_1
=
tdCom
.
getLongName
(
7
,
"letters"
)
input_json
,
stb_name
=
self
.
genFullTypeJson
(
stb_name
=
stb_name
,
tb_name
=
tb_name_1
,
t0
=
"f"
,
value
=
"f"
,
t_add_tag
=
True
)
self
.
resCmp
(
input_json
,
stb_name
,
condition
=
f
'where tbname like "
{
tb_name_1
}
"'
)
res_row_list
=
self
.
resHandle
(
f
"select t10,t11 from
{
tb_name
}
"
,
True
)[
0
]
tdSql
.
checkEqual
(
res_row_list
[
0
],
[
'None'
,
'None'
])
self
.
resCmp
(
input_json
,
stb_name
,
condition
=
f
'where tbname like "
{
tb_name
}
"'
,
none_check_tag
=
True
)
def
tagMd5Check
(
self
):
"""
condition: stb not change
insert two table, keep tag unchange, change col
"""
tdCom
.
cleanTb
()
input_json
,
stb_name
=
self
.
genFullTypeJson
(
t0
=
"f"
,
value
=
"f"
,
id_noexist_tag
=
True
)
self
.
resCmp
(
input_json
,
stb_name
)
tb_name1
=
self
.
getNoIdTbName
(
stb_name
)
input_json
,
stb_name
=
self
.
genFullTypeJson
(
stb_name
=
stb_name
,
t0
=
"f"
,
value
=
"f"
,
id_noexist_tag
=
True
)
self
.
resCmp
(
input_json
,
stb_name
)
tb_name2
=
self
.
getNoIdTbName
(
stb_name
)
tdSql
.
query
(
f
"select * from
{
stb_name
}
"
)
tdSql
.
checkRows
(
1
)
tdSql
.
checkEqual
(
tb_name1
,
tb_name2
)
input_json
,
stb_name
=
self
.
genFullTypeJson
(
stb_name
=
stb_name
,
t0
=
"f"
,
value
=
"f"
,
id_noexist_tag
=
True
,
t_add_tag
=
True
)
self
.
_conn
.
insert_json_payload
([
input_json
])
tb_name3
=
self
.
getNoIdTbName
(
stb_name
)
tdSql
.
query
(
f
"select * from
{
stb_name
}
"
)
tdSql
.
checkRows
(
2
)
tdSql
.
checkNotEqual
(
tb_name1
,
tb_name3
)
# * tag binary max is 16384, col+ts binary max 49151
def
tagColBinaryMaxLengthCheckCase
(
self
):
"""
every binary and nchar must be length+2
"""
tdCom
.
cleanTb
()
stb_name
=
tdCom
.
getLongName
(
7
,
"letters"
)
tb_name
=
f
'
{
stb_name
}
_1'
input_json
=
f
'
{
stb_name
}
1626006833639000000ns f id="
{
tb_name
}
",t0=t'
self
.
_conn
.
insert_json_payload
([
input_json
])
# * every binary and nchar must be length+2, so here is two tag, max length could not larger than 16384-2*2
input_json
=
f
'
{
stb_name
}
1626006833639000000ns f t0=t,t1="
{
tdCom
.
getLongName
(
16374
,
"letters"
)
}
",t2="
{
tdCom
.
getLongName
(
5
,
"letters"
)
}
"'
self
.
_conn
.
insert_json_payload
([
input_json
])
tdSql
.
query
(
f
"select * from
{
stb_name
}
"
)
tdSql
.
checkRows
(
2
)
input_json
=
f
'
{
stb_name
}
1626006833639000000ns f t0=t,t1="
{
tdCom
.
getLongName
(
16374
,
"letters"
)
}
",t2="
{
tdCom
.
getLongName
(
6
,
"letters"
)
}
"'
try
:
self
.
_conn
.
insert_json_payload
([
input_json
])
raise
Exception
(
"should not reach here"
)
except
JsonPayloadError
:
pass
tdSql
.
query
(
f
"select * from
{
stb_name
}
"
)
tdSql
.
checkRows
(
2
)
# * tag nchar max is 16374/4, col+ts nchar max 49151
def
tagColNcharMaxLengthCheckCase
(
self
):
"""
check nchar length limit
"""
tdCom
.
cleanTb
()
stb_name
=
tdCom
.
getLongName
(
7
,
"letters"
)
tb_name
=
f
'
{
stb_name
}
_1'
input_json
=
f
'
{
stb_name
}
1626006833639000000ns f id="
{
tb_name
}
",t0=t'
self
.
_conn
.
insert_json_payload
([
input_json
])
# * legal nchar could not be larger than 16374/4
input_json
=
f
'
{
stb_name
}
1626006833639000000ns f t0=t,t1=L"
{
tdCom
.
getLongName
(
4093
,
"letters"
)
}
",t2=L"
{
tdCom
.
getLongName
(
1
,
"letters"
)
}
"'
self
.
_conn
.
insert_json_payload
([
input_json
])
tdSql
.
query
(
f
"select * from
{
stb_name
}
"
)
tdSql
.
checkRows
(
2
)
input_json
=
f
'
{
stb_name
}
1626006833639000000ns f t0=t,t1=L"
{
tdCom
.
getLongName
(
4093
,
"letters"
)
}
",t2=L"
{
tdCom
.
getLongName
(
2
,
"letters"
)
}
"'
try
:
self
.
_conn
.
insert_json_payload
([
input_json
])
raise
Exception
(
"should not reach here"
)
except
JsonPayloadError
as
err
:
tdSql
.
checkNotEqual
(
err
.
errno
,
0
)
tdSql
.
query
(
f
"select * from
{
stb_name
}
"
)
tdSql
.
checkRows
(
2
)
def
batchInsertCheckCase
(
self
):
"""
test batch insert
"""
tdCom
.
cleanTb
()
stb_name
=
tdCom
.
getLongName
(
8
,
"letters"
)
tdSql
.
execute
(
f
'create stable
{
stb_name
}
(ts timestamp, f int) tags(t1 bigint)'
)
lines
=
[
"st123456 1626006833639000000ns 1i64 t1=3i64,t2=4f64,t3=
\"
t3
\"
"
,
"st123456 1626006833640000000ns 2i64 t1=4i64,t3=
\"
t4
\"
,t2=5f64,t4=5f64"
,
f
'
{
stb_name
}
1626056811823316532ns 3i64 t2=5f64,t3=L
\"
ste
\"
'
,
"stf567890 1626006933640000000ns 4i64 t1=4i64,t3=
\"
t4
\"
,t2=5f64,t4=5f64"
,
"st123456 1626006833642000000ns 5i64 t1=4i64,t2=5f64,t3=
\"
t4
\"
"
,
f
'
{
stb_name
}
1626056811843316532ns 6i64 t2=5f64,t3=L
\"
ste2
\"
'
,
f
'
{
stb_name
}
1626056812843316532ns 7i64 t2=5f64,t3=L
\"
ste2
\"
'
,
"st123456 1626006933640000000ns 8i64 t1=4i64,t3=
\"
t4
\"
,t2=5f64,t4=5f64"
,
"st123456 1626006933641000000ns 9i64 t1=4i64,t3=
\"
t4
\"
,t2=5f64,t4=5f64"
]
self
.
_conn
.
insert_json_payload
(
lines
)
tdSql
.
query
(
'show stables'
)
tdSql
.
checkRows
(
3
)
tdSql
.
query
(
'show tables'
)
tdSql
.
checkRows
(
6
)
tdSql
.
query
(
'select * from st123456'
)
tdSql
.
checkRows
(
5
)
def
multiInsertCheckCase
(
self
,
count
):
"""
test multi insert
"""
tdCom
.
cleanTb
()
sql_list
=
[]
stb_name
=
tdCom
.
getLongName
(
8
,
"letters"
)
tdSql
.
execute
(
f
'create stable
{
stb_name
}
(ts timestamp, f int) tags(t1 bigint)'
)
for
i
in
range
(
count
):
input_json
=
self
.
genFullTypeJson
(
stb_name
=
stb_name
,
t7
=
f
'"
{
tdCom
.
getLongName
(
8
,
"letters"
)
}
"'
,
value
=
f
'"
{
tdCom
.
getLongName
(
8
,
"letters"
)
}
"'
,
id_noexist_tag
=
True
)[
0
]
sql_list
.
append
(
input_json
)
self
.
_conn
.
insert_json_payload
(
sql_list
)
tdSql
.
query
(
'show tables'
)
tdSql
.
checkRows
(
1000
)
def
batchErrorInsertCheckCase
(
self
):
"""
test batch error insert
"""
tdCom
.
cleanTb
()
stb_name
=
tdCom
.
getLongName
(
8
,
"letters"
)
lines
=
[
"st123456 1626006833639000000ns 3i64 t1=3i64,t2=4f64,t3=
\"
t3
\"
"
,
f
"
{
stb_name
}
1626056811823316532ns tRue t2=5f64,t3=L
\"
ste
\"
"
]
try
:
self
.
_conn
.
insert_json_payload
(
lines
)
raise
Exception
(
"should not reach here"
)
except
JsonPayloadError
as
err
:
tdSql
.
checkNotEqual
(
err
.
errno
,
0
)
def
multiColsInsertCheckCase
(
self
):
"""
test multi cols insert
"""
tdCom
.
cleanTb
()
input_json
=
self
.
genFullTypeJson
(
t_multi_tag
=
True
)[
0
]
try
:
self
.
_conn
.
insert_json_payload
([
input_json
])
raise
Exception
(
"should not reach here"
)
except
JsonPayloadError
as
err
:
tdSql
.
checkNotEqual
(
err
.
errno
,
0
)
def
blankColInsertCheckCase
(
self
):
"""
test blank col insert
"""
tdCom
.
cleanTb
()
input_json
=
self
.
genFullTypeJson
(
c_blank_tag
=
True
)[
0
]
try
:
self
.
_conn
.
insert_json_payload
([
input_json
])
raise
Exception
(
"should not reach here"
)
except
JsonPayloadError
as
err
:
tdSql
.
checkNotEqual
(
err
.
errno
,
0
)
def
blankTagInsertCheckCase
(
self
):
"""
test blank tag insert
"""
tdCom
.
cleanTb
()
input_json
=
self
.
genFullTypeJson
(
t_blank_tag
=
True
)[
0
]
try
:
self
.
_conn
.
insert_json_payload
([
input_json
])
raise
Exception
(
"should not reach here"
)
except
JsonPayloadError
as
err
:
tdSql
.
checkNotEqual
(
err
.
errno
,
0
)
def
chineseCheckCase
(
self
):
"""
check nchar ---> chinese
"""
tdCom
.
cleanTb
()
input_json
,
stb_name
=
self
.
genFullTypeJson
(
chinese_tag
=
True
)
self
.
resCmp
(
input_json
,
stb_name
)
def
multiFieldCheckCase
(
self
):
'''
multi_field
'''
tdCom
.
cleanTb
()
input_json
=
self
.
genFullTypeJson
(
multi_field_tag
=
True
)[
0
]
try
:
self
.
_conn
.
insert_json_payload
([
input_json
])
raise
Exception
(
"should not reach here"
)
except
JsonPayloadError
as
err
:
tdSql
.
checkNotEqual
(
err
.
errno
,
0
)
def
errorTypeCheckCase
(
self
):
stb_name
=
tdCom
.
getLongName
(
8
,
"letters"
)
input_sql_list
=
[
f
'
{
stb_name
}
0 "hkgjiwdj" t0=f t1=127I8 t2=32767i16 t3=2147483647i32 t4=9223372036854775807i64 t5=11.12345f32 t6=22.123456789f64 t7="vozamcts" t8=L"ncharTagValue"'
,
\
f
'
{
stb_name
}
0 "hkgjiwdj" t0=f t1=127i8 t2=32767I16 t3=2147483647i32 t4=9223372036854775807i64 t5=11.12345f32 t6=22.123456789f64 t7="vozamcts" t8=L"ncharTagValue"'
,
\
f
'
{
stb_name
}
0 "hkgjiwdj" t0=f t1=127i8 t2=32767i16 t3=2147483647I32 t4=9223372036854775807i64 t5=11.12345f32 t6=22.123456789f64 t7="vozamcts" t8=L"ncharTagValue"'
,
\
f
'
{
stb_name
}
0 "hkgjiwdj" t0=f t1=127i8 t2=32767i16 t3=2147483647i32 t4=9223372036854775807I64 t5=11.12345f32 t6=22.123456789f64 t7="vozamcts" t8=L"ncharTagValue"'
,
\
f
'
{
stb_name
}
0 "hkgjiwdj" t0=f t1=127i8 t2=32767i16 t3=2147483647i32 t4=9223372036854775807i64 t5=11.12345F32 t6=22.123456789f64 t7="vozamcts" t8=L"ncharTagValue"'
,
\
f
'
{
stb_name
}
0 "hkgjiwdj" t0=f t1=127i8 t2=32767i16 t3=2147483647i32 t4=9223372036854775807i64 t5=11.12345f32 t6=22.123456789F64 t7="vozamcts" t8=L"ncharTagValue"'
,
\
f
'
{
stb_name
}
1626006833639000000NS "hkgjiwdj" t0=f t1=127i8 t2=32767i16 t3=2147483647i32 t4=9223372036854775807i64 t5=11.12345f32 t6=22.123456789f64 t7="vozamcts" t8=L"ncharTagValue"'
]
for
input_json
in
input_sql_list
:
try
:
self
.
_conn
.
insert_json_payload
([
input_json
])
raise
Exception
(
"should not reach here"
)
except
TelnetLinesError
as
err
:
tdSql
.
checkNotEqual
(
err
.
errno
,
0
)
def
genSqlList
(
self
,
count
=
5
,
stb_name
=
""
,
tb_name
=
""
):
"""
stb --> supertable
tb --> table
ts --> timestamp, same default
col --> column, same default
tag --> tag, same default
d --> different
s --> same
a --> add
m --> minus
"""
d_stb_d_tb_list
=
list
()
s_stb_s_tb_list
=
list
()
s_stb_s_tb_a_tag_list
=
list
()
s_stb_s_tb_m_tag_list
=
list
()
s_stb_d_tb_list
=
list
()
s_stb_d_tb_m_tag_list
=
list
()
s_stb_d_tb_a_tag_list
=
list
()
s_stb_s_tb_d_ts_list
=
list
()
s_stb_s_tb_d_ts_m_tag_list
=
list
()
s_stb_s_tb_d_ts_a_tag_list
=
list
()
s_stb_d_tb_d_ts_list
=
list
()
s_stb_d_tb_d_ts_m_tag_list
=
list
()
s_stb_d_tb_d_ts_a_tag_list
=
list
()
for
i
in
range
(
count
):
d_stb_d_tb_list
.
append
(
self
.
genFullTypeJson
(
t0
=
"f"
,
value
=
"f"
))
s_stb_s_tb_list
.
append
(
self
.
genFullTypeJson
(
stb_name
=
stb_name
,
tb_name
=
tb_name
,
t7
=
f
'"
{
tdCom
.
getLongName
(
8
,
"letters"
)
}
"'
,
value
=
f
'"
{
tdCom
.
getLongName
(
8
,
"letters"
)
}
"'
))
s_stb_s_tb_a_tag_list
.
append
(
self
.
genFullTypeJson
(
stb_name
=
stb_name
,
tb_name
=
tb_name
,
t7
=
f
'"
{
tdCom
.
getLongName
(
8
,
"letters"
)
}
"'
,
value
=
f
'"
{
tdCom
.
getLongName
(
8
,
"letters"
)
}
"'
,
t_add_tag
=
True
))
s_stb_s_tb_m_tag_list
.
append
(
self
.
genFullTypeJson
(
stb_name
=
stb_name
,
tb_name
=
tb_name
,
t7
=
f
'"
{
tdCom
.
getLongName
(
8
,
"letters"
)
}
"'
,
value
=
f
'"
{
tdCom
.
getLongName
(
8
,
"letters"
)
}
"'
,
t_mul_tag
=
True
))
s_stb_d_tb_list
.
append
(
self
.
genFullTypeJson
(
stb_name
=
stb_name
,
t7
=
f
'"
{
tdCom
.
getLongName
(
8
,
"letters"
)
}
"'
,
value
=
f
'"
{
tdCom
.
getLongName
(
8
,
"letters"
)
}
"'
,
id_noexist_tag
=
True
))
s_stb_d_tb_m_tag_list
.
append
(
self
.
genFullTypeJson
(
stb_name
=
stb_name
,
t7
=
f
'"
{
tdCom
.
getLongName
(
8
,
"letters"
)
}
"'
,
value
=
f
'"
{
tdCom
.
getLongName
(
8
,
"letters"
)
}
"'
,
id_noexist_tag
=
True
,
t_mul_tag
=
True
))
s_stb_d_tb_a_tag_list
.
append
(
self
.
genFullTypeJson
(
stb_name
=
stb_name
,
t7
=
f
'"
{
tdCom
.
getLongName
(
8
,
"letters"
)
}
"'
,
value
=
f
'"
{
tdCom
.
getLongName
(
8
,
"letters"
)
}
"'
,
id_noexist_tag
=
True
,
t_add_tag
=
True
))
s_stb_s_tb_d_ts_list
.
append
(
self
.
genFullTypeJson
(
stb_name
=
stb_name
,
tb_name
=
tb_name
,
t7
=
f
'"
{
tdCom
.
getLongName
(
8
,
"letters"
)
}
"'
,
value
=
f
'"
{
tdCom
.
getLongName
(
8
,
"letters"
)
}
"'
,
ts
=
0
))
s_stb_s_tb_d_ts_m_tag_list
.
append
(
self
.
genFullTypeJson
(
stb_name
=
stb_name
,
tb_name
=
tb_name
,
t7
=
f
'"
{
tdCom
.
getLongName
(
8
,
"letters"
)
}
"'
,
value
=
f
'"
{
tdCom
.
getLongName
(
8
,
"letters"
)
}
"'
,
ts
=
0
,
t_mul_tag
=
True
))
s_stb_s_tb_d_ts_a_tag_list
.
append
(
self
.
genFullTypeJson
(
stb_name
=
stb_name
,
tb_name
=
tb_name
,
t7
=
f
'"
{
tdCom
.
getLongName
(
8
,
"letters"
)
}
"'
,
value
=
f
'"
{
tdCom
.
getLongName
(
8
,
"letters"
)
}
"'
,
ts
=
0
,
t_add_tag
=
True
))
s_stb_d_tb_d_ts_list
.
append
(
self
.
genFullTypeJson
(
stb_name
=
stb_name
,
t7
=
f
'"
{
tdCom
.
getLongName
(
8
,
"letters"
)
}
"'
,
value
=
f
'"
{
tdCom
.
getLongName
(
8
,
"letters"
)
}
"'
,
id_noexist_tag
=
True
,
ts
=
0
))
s_stb_d_tb_d_ts_m_tag_list
.
append
(
self
.
genFullTypeJson
(
stb_name
=
stb_name
,
t7
=
f
'"
{
tdCom
.
getLongName
(
8
,
"letters"
)
}
"'
,
value
=
f
'"
{
tdCom
.
getLongName
(
8
,
"letters"
)
}
"'
,
id_noexist_tag
=
True
,
ts
=
0
,
t_mul_tag
=
True
))
s_stb_d_tb_d_ts_a_tag_list
.
append
(
self
.
genFullTypeJson
(
stb_name
=
stb_name
,
t7
=
f
'"
{
tdCom
.
getLongName
(
8
,
"letters"
)
}
"'
,
value
=
f
'"
{
tdCom
.
getLongName
(
8
,
"letters"
)
}
"'
,
id_noexist_tag
=
True
,
ts
=
0
,
t_add_tag
=
True
))
return
d_stb_d_tb_list
,
s_stb_s_tb_list
,
s_stb_s_tb_a_tag_list
,
s_stb_s_tb_m_tag_list
,
\
s_stb_d_tb_list
,
s_stb_d_tb_m_tag_list
,
s_stb_d_tb_a_tag_list
,
s_stb_s_tb_d_ts_list
,
\
s_stb_s_tb_d_ts_m_tag_list
,
s_stb_s_tb_d_ts_a_tag_list
,
s_stb_d_tb_d_ts_list
,
\
s_stb_d_tb_d_ts_m_tag_list
,
s_stb_d_tb_d_ts_a_tag_list
def
genMultiThreadSeq
(
self
,
sql_list
):
tlist
=
list
()
for
insert_sql
in
sql_list
:
t
=
threading
.
Thread
(
target
=
self
.
_conn
.
insert_json_payload
,
args
=
([
insert_sql
[
0
]],))
tlist
.
append
(
t
)
return
tlist
def
multiThreadRun
(
self
,
tlist
):
for
t
in
tlist
:
t
.
start
()
for
t
in
tlist
:
t
.
join
()
def
stbInsertMultiThreadCheckCase
(
self
):
"""
thread input different stb
"""
tdCom
.
cleanTb
()
input_json
=
self
.
genSqlList
()[
0
]
self
.
multiThreadRun
(
self
.
genMultiThreadSeq
(
input_json
))
tdSql
.
query
(
f
"show tables;"
)
tdSql
.
checkRows
(
5
)
def
sStbStbDdataInsertMultiThreadCheckCase
(
self
):
"""
thread input same stb tb, different data, result keep first data
"""
tdCom
.
cleanTb
()
tb_name
=
tdCom
.
getLongName
(
7
,
"letters"
)
input_json
,
stb_name
=
self
.
genFullTypeJson
(
tb_name
=
tb_name
,
value
=
"
\"
binaryTagValue
\"
"
)
self
.
resCmp
(
input_json
,
stb_name
)
s_stb_s_tb_list
=
self
.
genSqlList
(
stb_name
=
stb_name
,
tb_name
=
tb_name
)[
1
]
self
.
multiThreadRun
(
self
.
genMultiThreadSeq
(
s_stb_s_tb_list
))
tdSql
.
query
(
f
"show tables;"
)
tdSql
.
checkRows
(
1
)
expected_tb_name
=
self
.
getNoIdTbName
(
stb_name
)[
0
]
tdSql
.
checkEqual
(
tb_name
,
expected_tb_name
)
tdSql
.
query
(
f
"select * from
{
stb_name
}
;"
)
tdSql
.
checkRows
(
1
)
def
sStbStbDdataAtInsertMultiThreadCheckCase
(
self
):
"""
thread input same stb tb, different data, add columes and tags, result keep first data
"""
tdCom
.
cleanTb
()
tb_name
=
tdCom
.
getLongName
(
7
,
"letters"
)
input_json
,
stb_name
=
self
.
genFullTypeJson
(
tb_name
=
tb_name
,
value
=
"
\"
binaryTagValue
\"
"
)
self
.
resCmp
(
input_json
,
stb_name
)
s_stb_s_tb_a_tag_list
=
self
.
genSqlList
(
stb_name
=
stb_name
,
tb_name
=
tb_name
)[
2
]
self
.
multiThreadRun
(
self
.
genMultiThreadSeq
(
s_stb_s_tb_a_tag_list
))
tdSql
.
query
(
f
"show tables;"
)
tdSql
.
checkRows
(
1
)
expected_tb_name
=
self
.
getNoIdTbName
(
stb_name
)[
0
]
tdSql
.
checkEqual
(
tb_name
,
expected_tb_name
)
tdSql
.
query
(
f
"select * from
{
stb_name
}
;"
)
tdSql
.
checkRows
(
1
)
def
sStbStbDdataMtInsertMultiThreadCheckCase
(
self
):
"""
thread input same stb tb, different data, minus columes and tags, result keep first data
"""
tdCom
.
cleanTb
()
tb_name
=
tdCom
.
getLongName
(
7
,
"letters"
)
input_json
,
stb_name
=
self
.
genFullTypeJson
(
tb_name
=
tb_name
,
value
=
"
\"
binaryTagValue
\"
"
)
self
.
resCmp
(
input_json
,
stb_name
)
s_stb_s_tb_m_tag_list
=
self
.
genSqlList
(
stb_name
=
stb_name
,
tb_name
=
tb_name
)[
3
]
self
.
multiThreadRun
(
self
.
genMultiThreadSeq
(
s_stb_s_tb_m_tag_list
))
tdSql
.
query
(
f
"show tables;"
)
tdSql
.
checkRows
(
1
)
expected_tb_name
=
self
.
getNoIdTbName
(
stb_name
)[
0
]
tdSql
.
checkEqual
(
tb_name
,
expected_tb_name
)
tdSql
.
query
(
f
"select * from
{
stb_name
}
;"
)
tdSql
.
checkRows
(
1
)
def
sStbDtbDdataInsertMultiThreadCheckCase
(
self
):
"""
thread input same stb, different tb, different data
"""
tdCom
.
cleanTb
()
input_json
,
stb_name
=
self
.
genFullTypeJson
(
value
=
"
\"
binaryTagValue
\"
"
)
self
.
resCmp
(
input_json
,
stb_name
)
s_stb_d_tb_list
=
self
.
genSqlList
(
stb_name
=
stb_name
)[
4
]
self
.
multiThreadRun
(
self
.
genMultiThreadSeq
(
s_stb_d_tb_list
))
tdSql
.
query
(
f
"show tables;"
)
tdSql
.
checkRows
(
6
)
def
sStbDtbDdataMtInsertMultiThreadCheckCase
(
self
):
"""
thread input same stb, different tb, different data, add col, mul tag
"""
tdCom
.
cleanTb
()
input_json
,
stb_name
=
self
.
genFullTypeJson
(
value
=
"
\"
binaryTagValue
\"
"
)
self
.
resCmp
(
input_json
,
stb_name
)
s_stb_d_tb_m_tag_list
=
[(
f
'
{
stb_name
}
1626006833639000000ns "omfdhyom" t0=F,t1=127i8,t2=32767i16,t3=2147483647i32,t4=9223372036854775807i64,t5=11.12345f32,t6=22.123456789f64'
,
'yzwswz'
),
\
(
f
'
{
stb_name
}
1626006833639000000ns "vqowydbc" t0=F,t1=127i8,t2=32767i16,t3=2147483647i32,t4=9223372036854775807i64,t5=11.12345f32,t6=22.123456789f64'
,
'yzwswz'
),
\
(
f
'
{
stb_name
}
1626006833639000000ns "plgkckpv" t0=F,t1=127i8,t2=32767i16,t3=2147483647i32,t4=9223372036854775807i64,t5=11.12345f32,t6=22.123456789f64'
,
'yzwswz'
),
\
(
f
'
{
stb_name
}
1626006833639000000ns "cujyqvlj" t0=F,t1=127i8,t2=32767i16,t3=2147483647i32,t4=9223372036854775807i64,t5=11.12345f32,t6=22.123456789f64'
,
'yzwswz'
),
\
(
f
'
{
stb_name
}
1626006833639000000ns "twjxisat" t0=T,t1=127i8,t2=32767i16,t3=2147483647i32,t4=9223372036854775807i64,t5=11.12345f32,t6=22.123456789f64'
,
'yzwswz'
)]
self
.
multiThreadRun
(
self
.
genMultiThreadSeq
(
s_stb_d_tb_m_tag_list
))
tdSql
.
query
(
f
"show tables;"
)
tdSql
.
checkRows
(
3
)
def
sStbDtbDdataAtInsertMultiThreadCheckCase
(
self
):
"""
thread input same stb, different tb, different data, add tag, mul col
"""
tdCom
.
cleanTb
()
input_json
,
stb_name
=
self
.
genFullTypeJson
(
value
=
"
\"
binaryTagValue
\"
"
)
self
.
resCmp
(
input_json
,
stb_name
)
s_stb_d_tb_a_tag_list
=
self
.
genSqlList
(
stb_name
=
stb_name
)[
6
]
self
.
multiThreadRun
(
self
.
genMultiThreadSeq
(
s_stb_d_tb_a_tag_list
))
tdSql
.
query
(
f
"show tables;"
)
tdSql
.
checkRows
(
6
)
def
sStbStbDdataDtsInsertMultiThreadCheckCase
(
self
):
"""
thread input same stb tb, different ts
"""
tdCom
.
cleanTb
()
tb_name
=
tdCom
.
getLongName
(
7
,
"letters"
)
input_json
,
stb_name
=
self
.
genFullTypeJson
(
tb_name
=
tb_name
,
value
=
"
\"
binaryTagValue
\"
"
)
self
.
resCmp
(
input_json
,
stb_name
)
s_stb_s_tb_d_ts_list
=
[(
f
'
{
stb_name
}
0 "hkgjiwdj" id="
{
tb_name
}
",t0=f,t1=127i8,t2=32767i16,t3=2147483647i32,t4=9223372036854775807i64,t5=11.12345f32,t6=22.123456789f64,t7="vozamcts",t8=L"ncharTagValue"'
,
'dwpthv'
),
\
(
f
'
{
stb_name
}
0 "rljjrrul" id="
{
tb_name
}
",t0=False,t1=127i8,t2=32767i16,t3=2147483647i32,t4=9223372036854775807i64,t5=11.12345f32,t6=22.123456789f64,t7="bmcanhbs",t8=L"ncharTagValue"'
,
'dwpthv'
),
\
(
f
'
{
stb_name
}
0 "basanglx" id="
{
tb_name
}
",t0=False,t1=127i8,t2=32767i16,t3=2147483647i32,t4=9223372036854775807i64,t5=11.12345f32,t6=22.123456789f64,t7="enqkyvmb",t8=L"ncharTagValue"'
,
'dwpthv'
),
\
(
f
'
{
stb_name
}
0 "clsajzpp" id="
{
tb_name
}
",t0=F,t1=127i8,t2=32767i16,t3=2147483647i32,t4=9223372036854775807i64,t5=11.12345f32,t6=22.123456789f64,t7="eivaegjk",t8=L"ncharTagValue"'
,
'dwpthv'
),
\
(
f
'
{
stb_name
}
0 "jitwseso" id="
{
tb_name
}
",t0=T,t1=127i8,t2=32767i16,t3=2147483647i32,t4=9223372036854775807i64,t5=11.12345f32,t6=22.123456789f64,t7="yhlwkddq",t8=L"ncharTagValue"'
,
'dwpthv'
)]
self
.
multiThreadRun
(
self
.
genMultiThreadSeq
(
s_stb_s_tb_d_ts_list
))
tdSql
.
query
(
f
"show tables;"
)
tdSql
.
checkRows
(
1
)
tdSql
.
query
(
f
"select * from
{
stb_name
}
"
)
tdSql
.
checkRows
(
6
)
def
sStbStbDdataDtsMtInsertMultiThreadCheckCase
(
self
):
"""
thread input same stb tb, different ts, add col, mul tag
"""
tdCom
.
cleanTb
()
tb_name
=
tdCom
.
getLongName
(
7
,
"letters"
)
input_json
,
stb_name
=
self
.
genFullTypeJson
(
tb_name
=
tb_name
,
value
=
"
\"
binaryTagValue
\"
"
)
self
.
resCmp
(
input_json
,
stb_name
)
s_stb_s_tb_d_ts_m_tag_list
=
self
.
genSqlList
(
stb_name
=
stb_name
,
tb_name
=
tb_name
)[
8
]
self
.
multiThreadRun
(
self
.
genMultiThreadSeq
(
s_stb_s_tb_d_ts_m_tag_list
))
tdSql
.
query
(
f
"show tables;"
)
tdSql
.
checkRows
(
1
)
tdSql
.
query
(
f
"select * from
{
stb_name
}
"
)
tdSql
.
checkRows
(
6
)
tdSql
.
query
(
f
"select * from
{
stb_name
}
where t8 is not NULL"
)
tdSql
.
checkRows
(
6
)
def
sStbStbDdataDtsAtInsertMultiThreadCheckCase
(
self
):
"""
thread input same stb tb, different ts, add tag, mul col
"""
tdCom
.
cleanTb
()
tb_name
=
tdCom
.
getLongName
(
7
,
"letters"
)
input_json
,
stb_name
=
self
.
genFullTypeJson
(
tb_name
=
tb_name
,
value
=
"
\"
binaryTagValue
\"
"
)
self
.
resCmp
(
input_json
,
stb_name
)
s_stb_s_tb_d_ts_a_tag_list
=
[(
f
'
{
stb_name
}
0 "clummqfy" id="
{
tb_name
}
",t0=False,t1=127i8,t2=32767i16,t3=2147483647i32,t4=9223372036854775807i64,t5=11.12345f32,t6=22.123456789f64,t7="hpxzrdiw",t8=L"ncharTagValue",t11=127i8,t10=L"ncharTagValue"'
,
'bokaxl'
),
\
(
f
'
{
stb_name
}
0 "yqeztggb" id="
{
tb_name
}
",t0=F,t1=127i8,t2=32767i16,t3=2147483647i32,t4=9223372036854775807i64,t5=11.12345f32,t6=22.123456789f64,t7="gdtblmrc",t8=L"ncharTagValue",t11=127i8,t10=L"ncharTagValue"'
,
'bokaxl'
),
\
(
f
'
{
stb_name
}
0 "gbkinqdk" id="
{
tb_name
}
",t0=f,t1=127i8,t2=32767i16,t3=2147483647i32,t4=9223372036854775807i64,t5=11.12345f32,t6=22.123456789f64,t7="iqniuvco",t8=L"ncharTagValue",t11=127i8,t10=L"ncharTagValue"'
,
'bokaxl'
),
\
(
f
'
{
stb_name
}
0 "ldxxejbd" id="
{
tb_name
}
",t0=f,t1=127i8,t2=32767i16,t3=2147483647i32,t4=9223372036854775807i64,t5=11.12345f32,t6=22.123456789f64,t7="vxkipags",t8=L"ncharTagValue",t11=127i8,t10=L"ncharTagValue"'
,
'bokaxl'
),
\
(
f
'
{
stb_name
}
0 "tlvzwjes" id="
{
tb_name
}
",t0=true,t1=127i8,t2=32767i16,t3=2147483647i32,t4=9223372036854775807i64,t5=11.12345f32,t6=22.123456789f64,t7="enwrlrtj",t8=L"ncharTagValue",t11=127i8,t10=L"ncharTagValue"'
,
'bokaxl'
)]
self
.
multiThreadRun
(
self
.
genMultiThreadSeq
(
s_stb_s_tb_d_ts_a_tag_list
))
tdSql
.
query
(
f
"show tables;"
)
tdSql
.
checkRows
(
1
)
tdSql
.
query
(
f
"select * from
{
stb_name
}
"
)
tdSql
.
checkRows
(
6
)
for
t
in
[
"t10"
,
"t11"
]:
tdSql
.
query
(
f
"select * from
{
stb_name
}
where
{
t
}
is not NULL;"
)
tdSql
.
checkRows
(
0
)
def
sStbDtbDdataDtsInsertMultiThreadCheckCase
(
self
):
"""
thread input same stb, different tb, data, ts
"""
tdCom
.
cleanTb
()
input_json
,
stb_name
=
self
.
genFullTypeJson
(
value
=
"
\"
binaryTagValue
\"
"
)
self
.
resCmp
(
input_json
,
stb_name
)
s_stb_d_tb_d_ts_list
=
self
.
genSqlList
(
stb_name
=
stb_name
)[
10
]
self
.
multiThreadRun
(
self
.
genMultiThreadSeq
(
s_stb_d_tb_d_ts_list
))
tdSql
.
query
(
f
"show tables;"
)
tdSql
.
checkRows
(
6
)
def
sStbDtbDdataDtsMtInsertMultiThreadCheckCase
(
self
):
"""
thread input same stb, different tb, data, ts, add col, mul tag
"""
tdCom
.
cleanTb
()
input_json
,
stb_name
=
self
.
genFullTypeJson
(
value
=
"
\"
binaryTagValue
\"
"
)
self
.
resCmp
(
input_json
,
stb_name
)
s_stb_d_tb_d_ts_m_tag_list
=
[(
f
'
{
stb_name
}
0 "mnpmtzul" t0=f,t1=127i8,t2=32767i16,t3=2147483647i32,t4=9223372036854775807i64,t5=11.12345f32,t6=22.123456789f64'
,
'pcppkg'
),
\
(
f
'
{
stb_name
}
0 "zbvwckcd" t0=True,t1=127i8,t2=32767i16,t3=2147483647i32,t4=9223372036854775807i64,t5=11.12345f32,t6=22.123456789f64'
,
'pcppkg'
),
\
(
f
'
{
stb_name
}
0 "vymcjfwc" t0=F,t1=127i8,t2=32767i16,t3=2147483647i32,t4=9223372036854775807i64,t5=11.12345f32,t6=22.123456789f64'
,
'pcppkg'
),
\
(
f
'
{
stb_name
}
0 "laumkwfn" t0=False,t1=127i8,t2=32767i16,t3=2147483647i32,t4=9223372036854775807i64,t5=11.12345f32,t6=22.123456789f64'
,
'pcppkg'
),
\
(
f
'
{
stb_name
}
0 "nyultzxr" t0=false,t1=127i8,t2=32767i16,t3=2147483647i32,t4=9223372036854775807i64,t5=11.12345f32,t6=22.123456789f64'
,
'pcppkg'
)]
self
.
multiThreadRun
(
self
.
genMultiThreadSeq
(
s_stb_d_tb_d_ts_m_tag_list
))
tdSql
.
query
(
f
"show tables;"
)
tdSql
.
checkRows
(
3
)
def
test
(
self
):
# input_sql1 = "stb2_5 1626006833610ms 3f64 host=\"host0\",host2=L\"host2\""
# input_sql2 = "rfasta,id=\"rfasta_1\",t0=true,t1=127i8,t2=32767i16,t3=2147483647i32,t4=9223372036854775807i64,t5=11.12345f32,t6=22.123456789f64 c0=True,c1=127i8,c2=32767i16,c3=2147483647i32,c4=9223372036854775807i64,c5=11.12345f32,c6=22.123456789f64 1626006933640000000ns"
try
:
input_json
=
f
'test_nchar 0 L"涛思数据" t0=f,t1=L"涛思数据",t2=32767i16,t3=2147483647i32,t4=9223372036854775807i64,t5=11.12345f32,t6=22.123456789f64'
self
.
_conn
.
insert_json_payload
([
input_json
])
# input_json, stb_name = self.genFullTypeJson()
# self.resCmp(input_json, stb_name)
except
JsonPayloadError
as
err
:
print
(
err
.
errno
)
# self._conn.insert_json_payload([input_sql2])
# input_sql3 = f'abcd,id="cc¥Ec",t0=True,t1=127i8,t2=32767i16,t3=2147483647i32,t4=9223372036854775807i64,t5=11.12345f32,t6=22.123456789f64,t7="ndsfdrum",t8=L"ncharTagValue" c0=f,c1=127i8,c2=32767i16,c3=2147483647i32,c4=9223372036854775807i64,c5=11.12345f32,c6=22.123456789f64,c7="igwoehkm",c8=L"ncharColValue",c9=7u64 0'
# print(input_sql3)
# input_sql4 = 'hmemeb,id="kilrcrldgf",t0=F,t1=127i8,t2=32767i16,t3=2147483647i32,t4=9223372036854775807i64,t5=11.12345f32,t6=22.123456789f64,t7="fysodjql",t8=L"ncharTagValue" c0=True,c1=127i8,c2=32767i16,c3=2147483647i32,c4=9223372036854775807i64,c5=11.12345f32,c6=22.123456789f64,c7="waszbfvc",c8=L"ncharColValue",c9=7u64 0'
# code = self._conn.insert_json_payload([input_sql3])
# print(code)
# self._conn.insert_json_payload([input_sql4])
def
testJson
(
self
):
# input_sql1 = "stb2_5 1626006833610ms 3f64 host=\"host0\",host2=L\"host2\""
# input_sql2 = "rfasta,id=\"rfasta_1\",t0=true,t1=127i8,t2=32767i16,t3=2147483647i32,t4=9223372036854775807i64,t5=11.12345f32,t6=22.123456789f64 c0=True,c1=127i8,c2=32767i16,c3=2147483647i32,c4=9223372036854775807i64,c5=11.12345f32,c6=22.123456789f64 1626006933640000000ns"
try
:
tag_json
=
self
.
genTagValue
()
print
(
tag_json
)
# input_json = f'test_nchar 0 L"涛思数据" t0=f,t1=L"涛思数据",t2=32767i16,t3=2147483647i32,t4=9223372036854775807i64,t5=11.12345f32,t6=22.123456789f64'
# self._conn.insert_json_payload([input_json])
# input_json, stb_name = self.genFullTypeJson()
# self.resCmp(input_json, stb_name)
except
JsonPayloadError
as
err
:
print
(
err
.
errno
)
def
runAll
(
self
):
self
.
initCheckCase
()
self
.
boolTypeCheckCase
()
self
.
symbolsCheckCase
()
self
.
tsCheckCase
()
# self.idSeqCheckCase()
# self.idUpperCheckCase()
# self.noIdCheckCase()
# self.maxColTagCheckCase()
# self.idIllegalNameCheckCase()
# self.idStartWithNumCheckCase()
# self.nowTsCheckCase()
# self.dateFormatTsCheckCase()
# self.illegalTsCheckCase()
# self.tagValueLengthCheckCase()
# self.colValueLengthCheckCase()
# self.tagColIllegalValueCheckCase()
# self.blankCheckCase()
# self.duplicateIdTagColInsertCheckCase()
# self.noIdStbExistCheckCase()
# self.duplicateInsertExistCheckCase()
# self.tagColBinaryNcharLengthCheckCase()
# self.tagColAddDupIDCheckCase()
# self.tagColAddCheckCase()
# self.tagMd5Check()
# self.tagColBinaryMaxLengthCheckCase()
# self.tagColNcharMaxLengthCheckCase()
# self.batchInsertCheckCase()
# self.multiInsertCheckCase(1000)
# self.batchErrorInsertCheckCase()
# self.multiColsInsertCheckCase()
# self.blankColInsertCheckCase()
# self.blankTagInsertCheckCase()
# self.chineseCheckCase()
# self.multiFieldCheckCase()
# self.errorTypeCheckCase()
# # # MultiThreads
# self.stbInsertMultiThreadCheckCase()
# self.sStbStbDdataInsertMultiThreadCheckCase()
# self.sStbStbDdataAtInsertMultiThreadCheckCase()
# self.sStbStbDdataMtInsertMultiThreadCheckCase()
# self.sStbDtbDdataInsertMultiThreadCheckCase()
# self.sStbDtbDdataMtInsertMultiThreadCheckCase()
# self.sStbDtbDdataAtInsertMultiThreadCheckCase()
# self.sStbStbDdataDtsInsertMultiThreadCheckCase()
# self.sStbStbDdataDtsMtInsertMultiThreadCheckCase()
# self.sStbStbDdataDtsAtInsertMultiThreadCheckCase()
# self.sStbDtbDdataDtsInsertMultiThreadCheckCase()
# self.sStbDtbDdataDtsMtInsertMultiThreadCheckCase()
def
run
(
self
):
print
(
"running {}"
.
format
(
__file__
))
self
.
createDb
()
try
:
self
.
runAll
()
except
Exception
as
err
:
print
(
''
.
join
(
traceback
.
format_exception
(
None
,
err
,
err
.
__traceback__
)))
raise
err
def
stop
(
self
):
tdSql
.
close
()
tdLog
.
success
(
"%s successfully executed"
%
__file__
)
tdCases
.
addWindows
(
__file__
,
TDTestCase
())
tdCases
.
addLinux
(
__file__
,
TDTestCase
())
tests/pytest/util/common.py
浏览文件 @
7d5f3114
...
@@ -14,7 +14,7 @@
...
@@ -14,7 +14,7 @@
import
random
import
random
import
string
import
string
from
util.sql
import
tdSql
from
util.sql
import
tdSql
from
util.dnodes
import
tdDnodes
class
TDCom
:
class
TDCom
:
def
init
(
self
,
conn
,
logSql
):
def
init
(
self
,
conn
,
logSql
):
tdSql
.
init
(
conn
.
cursor
(),
logSql
)
tdSql
.
init
(
conn
.
cursor
(),
logSql
)
...
@@ -47,6 +47,42 @@ class TDCom:
...
@@ -47,6 +47,42 @@ class TDCom:
chars
=
''
.
join
(
random
.
choice
(
string
.
ascii_letters
.
lower
()
+
string
.
digits
)
for
i
in
range
(
len
))
chars
=
''
.
join
(
random
.
choice
(
string
.
ascii_letters
.
lower
()
+
string
.
digits
)
for
i
in
range
(
len
))
return
chars
return
chars
def
restartTaosd
(
self
,
index
=
1
,
db_name
=
"db"
):
tdDnodes
.
stop
(
index
)
tdDnodes
.
startWithoutSleep
(
index
)
tdSql
.
execute
(
f
"use
{
db_name
}
"
)
def
typeof
(
self
,
variate
):
v_type
=
None
if
type
(
variate
)
is
int
:
v_type
=
"int"
elif
type
(
variate
)
is
str
:
v_type
=
"str"
elif
type
(
variate
)
is
float
:
v_type
=
"float"
elif
type
(
variate
)
is
bool
:
v_type
=
"bool"
elif
type
(
variate
)
is
list
:
v_type
=
"list"
elif
type
(
variate
)
is
tuple
:
v_type
=
"tuple"
elif
type
(
variate
)
is
dict
:
v_type
=
"dict"
elif
type
(
variate
)
is
set
:
v_type
=
"set"
return
v_type
def
splitNumLetter
(
self
,
input_mix_str
):
nums
,
letters
=
""
,
""
for
i
in
input_mix_str
:
if
i
.
isdigit
():
nums
+=
i
elif
i
.
isspace
():
pass
else
:
letters
+=
i
return
nums
,
letters
def
close
(
self
):
def
close
(
self
):
self
.
cursor
.
close
()
self
.
cursor
.
close
()
...
...
编辑
预览
Markdown
is supported
0%
请重试
或
添加新附件
.
添加附件
取消
You are about to add
0
people
to the discussion. Proceed with caution.
先完成此消息的编辑!
取消
想要评论请
注册
或
登录