You are viewing a plain text version of this content. The canonical link for it is here.
Posted to commits@druid.apache.org by GitBox <gi...@apache.org> on 2019/07/15 15:28:07 UTC

[GitHub] [incubator-druid] quenlang opened a new issue #8082: DumpSegment tool does not resolve type id 'quantilesDoublesSketchMerge'

quenlang opened a new issue #8082: DumpSegment tool does not resolve type id 'quantilesDoublesSketchMerge'
URL: https://github.com/apache/incubator-druid/issues/8082
 
 
   Hi, all
   When I used the segment tool to dump some dataset, I got an exception:
   ```
   2019-07-15T16:55:49,624 INFO [main] org.apache.druid.server.metrics.MetricsModule - Adding monitor[org.apache.druid.server.initialization.jetty.JettyServerModule$JettyMonito
   r@b2f4ece]
   2019-07-15T16:55:49,700 WARN [main] org.apache.druid.segment.IndexIO - Failed to load metadata for segment [/data/tingyun/druid-0.12.3/20190714T000000.000Z_20190715T000000.0
   00Z/2019-07-15T00_37_12.065Z/0]
   com.fasterxml.jackson.databind.JsonMappingException: Could not resolve type id 'quantilesDoublesSketchMerge' into a subtype of [simple type, class org.apache.druid.query.agg
   regation.AggregatorFactory]: known type ids = [AggregatorFactory, cardinality, count, doubleFirst, doubleLast, doubleMax, doubleMin, doubleSum, filtered, floatFirst, floatLa
   st, floatMax, floatMin, floatSum, histogram, hyperUnique, javascript, longFirst, longLast, longMax, longMin, longSum, stringFirst, stringFirstFold, stringLast, stringLastFol
   d]
    at [Source: [B@ffaaaf0; line: 1, column: 185] (through reference chain: org.apache.druid.segment.Metadata["aggregators"]->Object[][2])
           at com.fasterxml.jackson.databind.JsonMappingException.from(JsonMappingException.java:148) ~[jackson-databind-2.6.7.jar:2.6.7]
           at com.fasterxml.jackson.databind.DeserializationContext.unknownTypeException(DeserializationContext.java:967) ~[jackson-databind-2.6.7.jar:2.6.7]
           at com.fasterxml.jackson.databind.jsontype.impl.TypeDeserializerBase._handleUnknownTypeId(TypeDeserializerBase.java:277) ~[jackson-databind-2.6.7.jar:2.6.7]
           at com.fasterxml.jackson.databind.jsontype.impl.TypeDeserializerBase._findDeserializer(TypeDeserializerBase.java:159) ~[jackson-databind-2.6.7.jar:2.6.7]
           at com.fasterxml.jackson.databind.jsontype.impl.AsPropertyTypeDeserializer._deserializeTypedForId(AsPropertyTypeDeserializer.java:108) ~[jackson-databind-2.6.7.jar:2
   .6.7]
           at com.fasterxml.jackson.databind.jsontype.impl.AsPropertyTypeDeserializer.deserializeTypedFromObject(AsPropertyTypeDeserializer.java:93) ~[jackson-databind-2.6.7.ja
   r:2.6.7]
           at com.fasterxml.jackson.databind.deser.AbstractDeserializer.deserializeWithType(AbstractDeserializer.java:131) ~[jackson-databind-2.6.7.jar:2.6.7]
           at com.fasterxml.jackson.databind.deser.std.ObjectArrayDeserializer.deserialize(ObjectArrayDeserializer.java:158) ~[jackson-databind-2.6.7.jar:2.6.7]
           at com.fasterxml.jackson.databind.deser.std.ObjectArrayDeserializer.deserialize(ObjectArrayDeserializer.java:17) ~[jackson-databind-2.6.7.jar:2.6.7]
           at com.fasterxml.jackson.databind.deser.SettableBeanProperty.deserialize(SettableBeanProperty.java:520) ~[jackson-databind-2.6.7.jar:2.6.7]
           at com.fasterxml.jackson.databind.deser.BeanDeserializer._deserializeWithErrorWrapping(BeanDeserializer.java:463) ~[jackson-databind-2.6.7.jar:2.6.7]
           at com.fasterxml.jackson.databind.deser.BeanDeserializer._deserializeUsingPropertyBased(BeanDeserializer.java:378) ~[jackson-databind-2.6.7.jar:2.6.7]
           at com.fasterxml.jackson.databind.deser.BeanDeserializerBase.deserializeFromObjectUsingNonDefault(BeanDeserializerBase.java:1099) ~[jackson-databind-2.6.7.jar:2.6.7]
           at com.fasterxml.jackson.databind.deser.BeanDeserializer.deserializeFromObject(BeanDeserializer.java:296) ~[jackson-databind-2.6.7.jar:2.6.7]
           at com.fasterxml.jackson.databind.deser.BeanDeserializer.deserialize(BeanDeserializer.java:133) ~[jackson-databind-2.6.7.jar:2.6.7]
           at com.fasterxml.jackson.databind.ObjectMapper._readMapAndClose(ObjectMapper.java:3736) ~[jackson-databind-2.6.7.jar:2.6.7]
           at com.fasterxml.jackson.databind.ObjectMapper.readValue(ObjectMapper.java:2819) ~[jackson-databind-2.6.7.jar:2.6.7]
           at org.apache.druid.segment.IndexIO$V9IndexLoader.load(IndexIO.java:571) [druid-processing-0.13.0-incubating.jar:0.13.0-incubating]
           at org.apache.druid.segment.IndexIO.loadIndex(IndexIO.java:187) [druid-processing-0.13.0-incubating.jar:0.13.0-incubating]
           at org.apache.druid.cli.DumpSegment.run(DumpSegment.java:180) [druid-services-0.13.0-incubating.jar:0.13.0-incubating]
           at org.apache.druid.cli.Main.main(Main.java:118) [druid-services-0.13.0-incubating.jar:0.13.0-incubating]
   2019-07-15T16:55:49,759 INFO [main] org.apache.druid.segment.CompressedPools - Allocating new littleEndByteBuf[1]
   Exception in thread "main" java.lang.NullPointerException
           at org.apache.druid.segment.column.SimpleColumnHolder.getColumn(SimpleColumnHolder.java:68)
           at org.apache.druid.segment.QueryableIndexColumnSelectorFactory.lambda$makeColumnValueSelector$1(QueryableIndexColumnSelectorFactory.java:125)
           at java.util.HashMap.computeIfAbsent(HashMap.java:1118)
           at org.apache.druid.segment.QueryableIndexColumnSelectorFactory.makeColumnValueSelector(QueryableIndexColumnSelectorFactory.java:122)
           at java.util.stream.ReferencePipeline$3$1.accept(ReferencePipeline.java:193)
           at java.util.Iterator.forEachRemaining(Iterator.java:116)
           at java.util.Spliterators$IteratorSpliterator.forEachRemaining(Spliterators.java:1801)
           at java.util.stream.AbstractPipeline.copyInto(AbstractPipeline.java:481)
           at java.util.stream.AbstractPipeline.wrapAndCopyInto(AbstractPipeline.java:471)
           at java.util.stream.ReduceOps$ReduceOp.evaluateSequential(ReduceOps.java:708)
           at java.util.stream.AbstractPipeline.evaluate(AbstractPipeline.java:234)
           at java.util.stream.ReferencePipeline.collect(ReferencePipeline.java:499)
           at org.apache.druid.cli.DumpSegment$2$1.apply(DumpSegment.java:280)
           at org.apache.druid.cli.DumpSegment$2$1.apply(DumpSegment.java:272)
           at org.apache.druid.java.util.common.guava.MappingAccumulator.accumulate(MappingAccumulator.java:40)
           at org.apache.druid.java.util.common.guava.FilteringAccumulator.accumulate(FilteringAccumulator.java:41)
           at org.apache.druid.java.util.common.guava.MappingAccumulator.accumulate(MappingAccumulator.java:40)
           at org.apache.druid.java.util.common.guava.BaseSequence.accumulate(BaseSequence.java:45)
           at org.apache.druid.java.util.common.guava.MappedSequence.accumulate(MappedSequence.java:43)
           at org.apache.druid.java.util.common.guava.WrappingSequence$1.get(WrappingSequence.java:50)
           at org.apache.druid.java.util.common.guava.SequenceWrapper.wrap(SequenceWrapper.java:55)
           at org.apache.druid.java.util.common.guava.WrappingSequence.accumulate(WrappingSequence.java:45)
           at org.apache.druid.java.util.common.guava.FilteredSequence.accumulate(FilteredSequence.java:45)
           at org.apache.druid.java.util.common.guava.MappedSequence.accumulate(MappedSequence.java:43)
           at org.apache.druid.cli.DumpSegment.evaluateSequenceForSideEffects(DumpSegment.java:494)
           at org.apache.druid.cli.DumpSegment.access$100(DumpSegment.java:103)
           at org.apache.druid.cli.DumpSegment$2.apply(DumpSegment.java:312)
           at org.apache.druid.cli.DumpSegment$2.apply(DumpSegment.java:265)
           at org.apache.druid.cli.DumpSegment.withOutputStream(DumpSegment.java:427)
           at org.apache.druid.cli.DumpSegment.runDump(DumpSegment.java:263)
           at org.apache.druid.cli.DumpSegment.run(DumpSegment.java:183)
           at org.apache.druid.cli.Main.main(Main.java:118)
   ```
   I have a quantilesDoublesSketch aggragation metric in my ingestion datasource:
   ```
   ...
               {
                   "name": "resp_time_his",
                   "fieldName": "resp_time",
                   "type": "quantilesDoublesSketch",
                   "k": 256
               },
   ...
   ```
   Then I performed the dump command, the exception occured. But it worked well without quantilesDoublesSketch aggragation metric. 
   

----------------------------------------------------------------
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.
 
For queries about this service, please contact Infrastructure at:
users@infra.apache.org


With regards,
Apache Git Services

---------------------------------------------------------------------
To unsubscribe, e-mail: commits-unsubscribe@druid.apache.org
For additional commands, e-mail: commits-help@druid.apache.org