Merge remote-tracking branch 'origin/2.0' into hotfix/2.0
This commit is contained in:
commit
0a3302fa20
|
@ -23,6 +23,7 @@ steps:
|
|||
branch:
|
||||
- develop
|
||||
- master
|
||||
- 2.0
|
||||
---
|
||||
kind: pipeline
|
||||
name: test_arm64
|
||||
|
@ -48,6 +49,7 @@ steps:
|
|||
branch:
|
||||
- develop
|
||||
- master
|
||||
- 2.0
|
||||
---
|
||||
kind: pipeline
|
||||
name: test_arm
|
||||
|
@ -73,6 +75,7 @@ steps:
|
|||
branch:
|
||||
- develop
|
||||
- master
|
||||
- 2.0
|
||||
|
||||
---
|
||||
kind: pipeline
|
||||
|
@ -100,6 +103,7 @@ steps:
|
|||
branch:
|
||||
- develop
|
||||
- master
|
||||
- 2.0
|
||||
---
|
||||
kind: pipeline
|
||||
name: build_xenial
|
||||
|
@ -125,6 +129,7 @@ steps:
|
|||
branch:
|
||||
- develop
|
||||
- master
|
||||
- 2.0
|
||||
|
||||
---
|
||||
kind: pipeline
|
||||
|
@ -150,6 +155,7 @@ steps:
|
|||
branch:
|
||||
- develop
|
||||
- master
|
||||
- 2.0
|
||||
---
|
||||
kind: pipeline
|
||||
name: build_centos7
|
||||
|
@ -173,6 +179,7 @@ steps:
|
|||
branch:
|
||||
- develop
|
||||
- master
|
||||
- 2.0
|
||||
|
||||
---
|
||||
kind: pipeline
|
||||
|
@ -191,6 +198,7 @@ steps:
|
|||
branch:
|
||||
- develop
|
||||
- master
|
||||
- 2.0
|
||||
|
||||
|
||||
depends_on:
|
||||
|
|
|
@ -52,12 +52,18 @@ def pre_test(){
|
|||
git checkout master
|
||||
'''
|
||||
}
|
||||
else {
|
||||
else if(env.CHANGE_TARGET == '2.0'){
|
||||
sh '''
|
||||
cd ${WKC}
|
||||
git checkout 2.0
|
||||
'''
|
||||
}
|
||||
else{
|
||||
sh '''
|
||||
cd ${WKC}
|
||||
git checkout develop
|
||||
'''
|
||||
}
|
||||
}
|
||||
}
|
||||
sh'''
|
||||
cd ${WKC}
|
||||
|
@ -75,7 +81,13 @@ def pre_test(){
|
|||
git checkout master
|
||||
'''
|
||||
}
|
||||
else {
|
||||
else if(env.CHANGE_TARGET == '2.0'){
|
||||
sh '''
|
||||
cd ${WK}
|
||||
git checkout 2.0
|
||||
'''
|
||||
}
|
||||
else{
|
||||
sh '''
|
||||
cd ${WK}
|
||||
git checkout develop
|
||||
|
|
|
@ -4,7 +4,7 @@ PROJECT(TDengine)
|
|||
IF (DEFINED VERNUMBER)
|
||||
SET(TD_VER_NUMBER ${VERNUMBER})
|
||||
ELSE ()
|
||||
SET(TD_VER_NUMBER "2.0.20.10")
|
||||
SET(TD_VER_NUMBER "2.0.20.12")
|
||||
ENDIF ()
|
||||
|
||||
IF (DEFINED VERCOMPATIBLE)
|
||||
|
|
|
@ -44,7 +44,8 @@ echo "version=${version}"
|
|||
#docker manifest rm tdengine/tdengine
|
||||
#docker manifest rm tdengine/tdengine:${version}
|
||||
if [ "$verType" == "beta" ]; then
|
||||
docker manifest rm tdengine/tdengine:latest
|
||||
docker manifest inspect tdengine/tdengine-beta:latest
|
||||
docker manifest rm tdengine/tdengine-beta:latest
|
||||
docker manifest create -a tdengine/tdengine-beta:${version} tdengine/tdengine-amd64-beta:${version} tdengine/tdengine-aarch64-beta:${version} tdengine/tdengine-aarch32-beta:${version}
|
||||
docker manifest create -a tdengine/tdengine-beta:latest tdengine/tdengine-amd64-beta:latest tdengine/tdengine-aarch64-beta:latest tdengine/tdengine-aarch32-beta:latest
|
||||
docker login -u tdengine -p ${passWord} #replace the docker registry username and password
|
||||
|
@ -52,6 +53,7 @@ if [ "$verType" == "beta" ]; then
|
|||
docker manifest push tdengine/tdengine-beta:${version}
|
||||
|
||||
elif [ "$verType" == "stable" ]; then
|
||||
docker manifest inspect tdengine/tdengine:latest
|
||||
docker manifest rm tdengine/tdengine:latest
|
||||
docker manifest create -a tdengine/tdengine:${version} tdengine/tdengine-amd64:${version} tdengine/tdengine-aarch64:${version} tdengine/tdengine-aarch32:${version}
|
||||
docker manifest create -a tdengine/tdengine:latest tdengine/tdengine-amd64:latest tdengine/tdengine-aarch64:latest tdengine/tdengine-aarch32:latest
|
||||
|
|
|
@ -1,6 +1,6 @@
|
|||
name: tdengine
|
||||
base: core18
|
||||
version: '2.0.20.10'
|
||||
version: '2.0.20.12'
|
||||
icon: snap/gui/t-dengine.svg
|
||||
summary: an open-source big data platform designed and optimized for IoT.
|
||||
description: |
|
||||
|
@ -72,7 +72,7 @@ parts:
|
|||
- usr/bin/taosd
|
||||
- usr/bin/taos
|
||||
- usr/bin/taosdemo
|
||||
- usr/lib/libtaos.so.2.0.20.10
|
||||
- usr/lib/libtaos.so.2.0.20.12
|
||||
- usr/lib/libtaos.so.1
|
||||
- usr/lib/libtaos.so
|
||||
|
||||
|
|
|
@ -5182,7 +5182,7 @@ static int64_t generateStbRowData(
|
|||
dataLen += 1;
|
||||
}
|
||||
|
||||
if (dataLen > remainderBufLen)
|
||||
if (dataLen > (remainderBufLen - (DOUBLE_BUFF_LEN + 1)))
|
||||
return 0;
|
||||
}
|
||||
|
||||
|
@ -6317,8 +6317,8 @@ static void printStatPerThread(threadInfo *pThreadInfo)
|
|||
pThreadInfo->threadID,
|
||||
pThreadInfo->totalInsertRows,
|
||||
pThreadInfo->totalAffectedRows,
|
||||
(pThreadInfo->totalDelay/1000.0)?
|
||||
(double)(pThreadInfo->totalAffectedRows/(pThreadInfo->totalDelay/1000.0)):
|
||||
(pThreadInfo->totalDelay)?
|
||||
(double)(pThreadInfo->totalAffectedRows/((double)pThreadInfo->totalDelay/1000000.0)):
|
||||
FLT_MAX);
|
||||
}
|
||||
|
||||
|
@ -6538,7 +6538,7 @@ static void* syncWriteInterlace(threadInfo *pThreadInfo) {
|
|||
verbosePrint("[%d] %s() LN%d, buffer=%s\n",
|
||||
pThreadInfo->threadID, __func__, __LINE__, pThreadInfo->buffer);
|
||||
|
||||
startTs = taosGetTimestampMs();
|
||||
startTs = taosGetTimestampUs();
|
||||
|
||||
if (recOfBatch == 0) {
|
||||
errorPrint("[%d] %s() LN%d Failed to insert records of batch %d\n",
|
||||
|
@ -6554,10 +6554,10 @@ static void* syncWriteInterlace(threadInfo *pThreadInfo) {
|
|||
}
|
||||
int64_t affectedRows = execInsert(pThreadInfo, recOfBatch);
|
||||
|
||||
endTs = taosGetTimestampMs();
|
||||
endTs = taosGetTimestampUs();
|
||||
uint64_t delay = endTs - startTs;
|
||||
performancePrint("%s() LN%d, insert execution time is %"PRIu64"ms\n",
|
||||
__func__, __LINE__, delay);
|
||||
performancePrint("%s() LN%d, insert execution time is %10.2f ms\n",
|
||||
__func__, __LINE__, delay / 1000.0);
|
||||
verbosePrint("[%d] %s() LN%d affectedRows=%"PRId64"\n",
|
||||
pThreadInfo->threadID,
|
||||
__func__, __LINE__, affectedRows);
|
||||
|
@ -6714,14 +6714,14 @@ static void* syncWriteProgressive(threadInfo *pThreadInfo) {
|
|||
start_time += generated * timeStampStep;
|
||||
pThreadInfo->totalInsertRows += generated;
|
||||
|
||||
startTs = taosGetTimestampMs();
|
||||
startTs = taosGetTimestampUs();
|
||||
|
||||
int32_t affectedRows = execInsert(pThreadInfo, generated);
|
||||
|
||||
endTs = taosGetTimestampMs();
|
||||
endTs = taosGetTimestampUs();
|
||||
uint64_t delay = endTs - startTs;
|
||||
performancePrint("%s() LN%d, insert execution time is %"PRId64"ms\n",
|
||||
__func__, __LINE__, delay);
|
||||
performancePrint("%s() LN%d, insert execution time is %10.f ms\n",
|
||||
__func__, __LINE__, delay/1000.0);
|
||||
verbosePrint("[%d] %s() LN%d affectedRows=%d\n",
|
||||
pThreadInfo->threadID,
|
||||
__func__, __LINE__, affectedRows);
|
||||
|
@ -7255,11 +7255,15 @@ static void startMultiThreadInsertData(int threads, char* db_name,
|
|||
}
|
||||
}
|
||||
|
||||
fprintf(stderr, "insert delay, avg: %10.2fms, max: %"PRIu64"ms, min: %"PRIu64"ms\n\n",
|
||||
avgDelay, maxDelay, minDelay);
|
||||
fprintf(stderr, "insert delay, avg: %10.2fms, max: %10.2fms, min: %10.2fms\n\n",
|
||||
(double)avgDelay/1000.0,
|
||||
(double)maxDelay/1000.0,
|
||||
(double)minDelay/1000.0);
|
||||
if (g_fpOfInsertResult) {
|
||||
fprintf(g_fpOfInsertResult, "insert delay, avg:%10.2fms, max: %"PRIu64"ms, min: %"PRIu64"ms\n\n",
|
||||
avgDelay, maxDelay, minDelay);
|
||||
fprintf(g_fpOfInsertResult, "insert delay, avg:%10.2fms, max: %10.2fms, min: %10.2fms\n\n",
|
||||
(double)avgDelay/1000.0,
|
||||
(double)maxDelay/1000.0,
|
||||
(double)minDelay/1000.0);
|
||||
}
|
||||
|
||||
//taos_close(taos);
|
||||
|
|
|
@ -231,8 +231,8 @@ static struct argp_option options[] = {
|
|||
{"schemaonly", 's', 0, 0, "Only dump schema.", 2},
|
||||
{"without-property", 'N', 0, 0, "Dump schema without properties.", 2},
|
||||
{"avro", 'V', 0, 0, "Dump apache avro format data file. By default, dump sql command sequence.", 2},
|
||||
{"start-time", 'S', "START_TIME", 0, "Start time to dump. Either epoch or ISO8601/RFC3339 format is acceptable. ISO8601 format example: 2017-10-01T18:00:00.000+0800 or 2017-10-0100:00:00.000+0800 or '2017-10-01 00:00:00.000+0800'", 4},
|
||||
{"end-time", 'E', "END_TIME", 0, "End time to dump. Either epoch or ISO8601/RFC3339 format is acceptable. ISO8601 format example: 2017-10-01T18:00:00.000+0800 or 2017-10-0100:00:00.000+0800 or '2017-10-01 00:00:00.000+0800'", 5},
|
||||
{"start-time", 'S', "START_TIME", 0, "Start time to dump. Either epoch or ISO8601/RFC3339 format is acceptable. ISO8601 format example: 2017-10-01T00:00:00.000+0800 or 2017-10-0100:00:00:000+0800 or '2017-10-01 00:00:00.000+0800'", 4},
|
||||
{"end-time", 'E', "END_TIME", 0, "End time to dump. Either epoch or ISO8601/RFC3339 format is acceptable. ISO8601 format example: 2017-10-01T00:00:00.000+0800 or 2017-10-0100:00:00.000+0800 or '2017-10-01 00:00:00.000+0800'", 5},
|
||||
#if TSDB_SUPPORT_NANOSECOND == 1
|
||||
{"precision", 'C', "PRECISION", 0, "Specify precision for converting human-readable time to epoch. Valid value is one of ms, us, and ns. Default is ms.", 6},
|
||||
#else
|
||||
|
|
Loading…
Reference in New Issue