You are viewing a plain text version of this content. The canonical link for it is here.
Posted to commits@pinot.apache.org by xh...@apache.org on 2019/01/03 00:25:39 UTC
[incubator-pinot] branch master updated: [TE] remove log for the
objects (#3630)
This is an automated email from the ASF dual-hosted git repository.
xhsun pushed a commit to branch master
in repository https://gitbox.apache.org/repos/asf/incubator-pinot.git
The following commit(s) were added to refs/heads/master by this push:
new d5a9d08 [TE] remove log for the objects (#3630)
d5a9d08 is described below
commit d5a9d0801ee48a9de679bffcbe4bca557d077ca4
Author: Xiaohui Sun <xh...@linkedin.com>
AuthorDate: Wed Jan 2 16:25:34 2019 -0800
[TE] remove log for the objects (#3630)
---
.../thirdeye/anomaly/detection/DetectionJobScheduler.java | 12 ++++++++----
1 file changed, 8 insertions(+), 4 deletions(-)
diff --git a/thirdeye/thirdeye-pinot/src/main/java/com/linkedin/thirdeye/anomaly/detection/DetectionJobScheduler.java b/thirdeye/thirdeye-pinot/src/main/java/com/linkedin/thirdeye/anomaly/detection/DetectionJobScheduler.java
index 3f1e822..84a600f 100644
--- a/thirdeye/thirdeye-pinot/src/main/java/com/linkedin/thirdeye/anomaly/detection/DetectionJobScheduler.java
+++ b/thirdeye/thirdeye-pinot/src/main/java/com/linkedin/thirdeye/anomaly/detection/DetectionJobScheduler.java
@@ -30,6 +30,7 @@ import com.linkedin.thirdeye.datalayer.dto.DetectionStatusDTO;
import com.linkedin.thirdeye.datalayer.dto.JobDTO;
import com.linkedin.thirdeye.datalayer.dto.TaskDTO;
+import com.linkedin.thirdeye.datalayer.pojo.DataCompletenessConfigBean;
import java.text.ParseException;
import java.util.ArrayList;
import java.util.Collections;
@@ -45,6 +46,7 @@ import java.util.concurrent.Executors;
import java.util.concurrent.ScheduledExecutorService;
import java.util.concurrent.TimeUnit;
+import java.util.stream.Collectors;
import org.apache.commons.collections.CollectionUtils;
import org.joda.time.DateTime;
import org.joda.time.DateTimeZone;
@@ -104,7 +106,7 @@ public class DetectionJobScheduler implements Runnable {
for (AnomalyFunctionDTO anomalyFunction : anomalyFunctions) {
try {
- LOG.info("Function: {}", anomalyFunction);
+ LOG.info("Function: {}", anomalyFunction.getFunctionName());
long functionId = anomalyFunction.getId();
String dataset = anomalyFunction.getCollection();
DatasetConfigDTO datasetConfig = CACHE_REGISTRY.getDatasetConfigCache().get(dataset);
@@ -136,7 +138,7 @@ public class DetectionJobScheduler implements Runnable {
findAllInTimeRangeForFunctionAndDetectionRun(currentDateTime.minusDays(3).getMillis(),
currentDateTime.getMillis(), functionId, false);
Collections.sort(entriesInLast3Days);
- LOG.info("Function: {} Dataset: {} Entries in last 3 days {}", functionId, dataset, entriesInLast3Days);
+ LOG.info("Function: {} Dataset: {} Entry count in last 3 days {}", functionId, dataset, entriesInLast3Days.size());
// for each entry, collect startTime and endTime
List<Long> startTimes = new ArrayList<>();
@@ -265,13 +267,15 @@ public class DetectionJobScheduler implements Runnable {
List<DataCompletenessConfigDTO> incompleteTimePeriods = DAO_REGISTRY.getDataCompletenessConfigDAO().
findAllByDatasetAndInTimeRangeAndStatus(dataset, startTime, endTime, false);
- LOG.info("Function: {} Dataset: {} Incomplete periods {}", anomalyFunction.getId(), dataset, incompleteTimePeriods);
+ LOG.info("Function: {} Dataset: {} Incomplete periods {}", anomalyFunction.getId(), dataset,
+ String.join(",", incompleteTimePeriods.stream().map(DataCompletenessConfigBean::getDateToCheckInSDF).collect(Collectors.toList())));
if (incompleteTimePeriods.size() == 0) { // nothing incomplete
// find complete buckets
List<DataCompletenessConfigDTO> completeTimePeriods = DAO_REGISTRY.getDataCompletenessConfigDAO().
findAllByDatasetAndInTimeRangeAndStatus(dataset, startTime, endTime, true);
- LOG.info("Function: {} Dataset: {} Complete periods {}", anomalyFunction.getId(), dataset, completeTimePeriods);
+ LOG.info("Function: {} Dataset: {} Complete periods {}", anomalyFunction.getId(), dataset,
+ String.join(",", completeTimePeriods.stream().map(DataCompletenessConfigBean::getDateToCheckInSDF).collect(Collectors.toList())));
long expectedCompleteBuckets = DetectionJobSchedulerUtils.getExpectedCompleteBuckets(datasetConfig, startTime, endTime);
LOG.info("Function: {} Dataset: {} Num complete periods: {} Expected num buckets:{}",
anomalyFunction.getId(), dataset, completeTimePeriods.size(), expectedCompleteBuckets);
---------------------------------------------------------------------
To unsubscribe, e-mail: commits-unsubscribe@pinot.apache.org
For additional commands, e-mail: commits-help@pinot.apache.org