fix(query): fix percentile function crash on large data set
This commit is contained in:
parent
83fcf4fd8d
commit
83be0489a3
|
@ -284,7 +284,6 @@ static char* evacOneDataPage(SDiskbasedBuf* pBuf) {
|
|||
|
||||
// all pages are referenced by user, try to allocate new space
|
||||
if (pn == NULL) {
|
||||
assert(0);
|
||||
int32_t prev = pBuf->inMemPages;
|
||||
|
||||
// increase by 50% of previous mem pages
|
||||
|
@ -304,7 +303,6 @@ static char* evacOneDataPage(SDiskbasedBuf* pBuf) {
|
|||
bufPage = flushPageToDisk(pBuf, d);
|
||||
}
|
||||
|
||||
ASSERT((bufPage != NULL) || terrno != TSDB_CODE_SUCCESS);
|
||||
return bufPage;
|
||||
}
|
||||
|
||||
|
@ -377,12 +375,6 @@ void* getNewBufPage(SDiskbasedBuf* pBuf, int32_t* pageId) {
|
|||
char* availablePage = NULL;
|
||||
if (NO_IN_MEM_AVAILABLE_PAGES(pBuf)) {
|
||||
availablePage = evacOneDataPage(pBuf);
|
||||
|
||||
// Failed to allocate a new buffer page, and there is an error occurs.
|
||||
if (availablePage == NULL) {
|
||||
assert(0);
|
||||
return NULL;
|
||||
}
|
||||
}
|
||||
|
||||
SPageInfo* pi = NULL;
|
||||
|
@ -652,4 +644,4 @@ void clearDiskbasedBuf(SDiskbasedBuf* pBuf) {
|
|||
pBuf->totalBufSize = 0;
|
||||
pBuf->allocateId = -1;
|
||||
pBuf->fileSize = 0;
|
||||
}
|
||||
}
|
||||
|
|
Loading…
Reference in New Issue