Skip to content
Merged
Show file tree
Hide file tree
Changes from 1 commit
Commits
File filter

Filter by extension

Filter by extension

Conversations
Failed to load comments.
Loading
Jump to
Jump to file
Failed to load files.
Loading
Diff view
Diff view
Original file line number Diff line number Diff line change
Expand Up @@ -2359,16 +2359,6 @@ protected int executeCompaction() {
}
}
if (summary.hasSubmitTask()) {
logger.info(
"[CompactionScheduler][{}] selected sequence InnerSpaceCompactionTask num is {},"
+ " selected unsequence InnerSpaceCompactionTask num is {},"
+ " selected CrossSpaceCompactionTask num is {},"
+ " selected InsertionCrossSpaceCompactionTask num is {}",
dataRegionId,
summary.getSubmitSeqInnerSpaceCompactionTaskNum(),
summary.getSubmitUnseqInnerSpaceCompactionTaskNum(),
summary.getSubmitCrossSpaceCompactionTaskNum(),
summary.getSubmitInsertionCrossSpaceCompactionTaskNum());
CompactionMetrics.getInstance().updateCompactionTaskSelectionNum(summary);
}
return trySubmitCount;
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -95,9 +95,6 @@ public synchronized void updateRatio(long memorySize, long diskSize) throws IOEx
String.format(
Locale.ENGLISH, RATIO_FILE_PATH_FORMAT, totalMemorySize.get(), totalDiskSize));
persist(oldFile, newFile);
if (LOGGER.isDebugEnabled()) {
LOGGER.debug("Compression ratio is {}", (double) totalMemorySize.get() / totalDiskSize);
}
}

/** Get the average compression ratio for all closed files */
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -110,14 +110,6 @@ public void syncFlushMemTable() throws ExecutionException, InterruptedException
memTable.getSeriesNumber() == 0
? 0
: memTable.getTotalPointsNum() / memTable.getSeriesNumber();
if (LOGGER.isDebugEnabled()) {
LOGGER.debug(
"The memTable size of SG {} is {}, the avg series points num in chunk is {}, total timeseries number is {}",
storageGroup,
memTable.memSize(),
avgSeriesPointsNum,
memTable.getSeriesNumber());
}
WRITING_METRICS.recordFlushingMemTableStatus(
storageGroup,
memTable.memSize(),
Expand Down Expand Up @@ -215,14 +207,6 @@ public void syncFlushMemTable() throws ExecutionException, InterruptedException
MetricLevel.CORE,
Tag.NAME.toString(),
"flush");

if (LOGGER.isDebugEnabled()) {
LOGGER.debug(
"Database {} memtable {} flushing a memtable has finished! Time consumption: {}ms",
storageGroup,
memTable,
System.currentTimeMillis() - start);
}
}

/** encoding task (second task of pipeline) */
Expand Down Expand Up @@ -288,14 +272,6 @@ public void run() {
}

recordFlushPointsMetric();

if (LOGGER.isDebugEnabled()) {
LOGGER.debug(
"Database {}, flushing memtable {} into disk: Encoding data cost " + "{} ms.",
storageGroup,
writer.getFile().getName(),
memSerializeTime);
}
WRITING_METRICS.recordFlushCost(WritingMetrics.FLUSH_STAGE_ENCODING, memSerializeTime);
}
};
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -792,12 +792,6 @@ public boolean shouldFlush() {
return false;
}
if (workMemTable.shouldFlush()) {
if (logger.isDebugEnabled()) {
logger.debug(
"The memtable size {} of tsfile {} reaches the mem control threshold",
workMemTable.memSize(),
tsFileResource.getTsFile().getAbsolutePath());
}
return true;
}
if (!enableMemControl && workMemTable.memSize() >= getMemtableSizeThresholdBasedOnSeriesNum()) {
Expand Down Expand Up @@ -873,9 +867,9 @@ public Future<?> asyncClose() {
FLUSH_QUERY_WRITE_LOCKED, storageGroupName, tsFileResource.getTsFile().getName());
}
try {
if (logger.isInfoEnabled()) {
if (logger.isDebugEnabled()) {
if (workMemTable != null) {
logger.info(
logger.debug(
"{}: flush a working memtable in async close tsfile {}, memtable size: {}, tsfile "
+ "size: {}, plan index: [{}, {}], progress index: {}",
storageGroupName,
Expand All @@ -886,7 +880,7 @@ public Future<?> asyncClose() {
workMemTable.getMaxPlanIndex(),
tsFileResource.getMaxProgressIndex());
} else {
logger.info(
logger.debug(
"{}: flush a NotifyFlushMemTable in async close tsfile {}, tsfile size: {}",
storageGroupName,
tsFileResource.getTsFile().getAbsolutePath(),
Expand Down Expand Up @@ -1346,14 +1340,12 @@ public void flushOneMemTable() {
private void updateCompressionRatio() {
try {
double compressionRatio = ((double) totalMemTableSize) / writer.getPos();
if (logger.isDebugEnabled()) {
logger.debug(
"The compression ratio of tsfile {} is {}, totalMemTableSize: {}, the file size: {}",
writer.getFile().getAbsolutePath(),
compressionRatio,
totalMemTableSize,
writer.getPos());
}
logger.info(
"The compression ratio of tsfile {} is {}, totalMemTableSize: {}, the file size: {}",
writer.getFile().getAbsolutePath(),
compressionRatio,
totalMemTableSize,
writer.getPos());
String dataRegionId = dataRegionInfo.getDataRegion().getDataRegionId();
WritingMetrics.getInstance()
.recordTsFileCompressionRatioOfFlushingMemTable(dataRegionId, compressionRatio);
Expand All @@ -1369,14 +1361,8 @@ private void updateCompressionRatio() {

/** end file and write some meta */
private void endFile() throws IOException, TsFileProcessorException {
if (logger.isDebugEnabled()) {
logger.debug("Start to end file {}", tsFileResource);
}
writer.endFile();
tsFileResource.serialize();
if (logger.isDebugEnabled()) {
logger.debug("Ended file {}", tsFileResource);
}
// remove this processor from Closing list in StorageGroupProcessor,
// mark the TsFileResource closed, no need writer anymore
for (CloseFileListener closeFileListener : closeFileListeners) {
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -321,7 +321,6 @@ public void endCurrentChunk() {
*/
@SuppressWarnings("squid:S3776") // Suppress high Cognitive Complexity warning
public void endFile() throws IOException {
long startTime = System.currentTimeMillis();
checkInMemoryPathCount();
readChunkMetadataAndConstructIndexTree();

Expand All @@ -348,8 +347,6 @@ public void endFile() throws IOException {
}
}
canWrite = false;
long cost = System.currentTimeMillis() - startTime;
logger.debug("Time for flushing metadata is {} ms", cost);
}

private void checkInMemoryPathCount() {
Expand Down