You are viewing a plain text version of this content. The canonical link for it is here.
Posted to user@kylin.apache.org by "Kumar, Manoj H" <ma...@jpmorgan.com> on 2017/10/12 06:29:21 UTC
Kafka Streaming data - Error while building the Cube
Building the Cube from Kylin UI - Although Messages are there in Kafka topic but Kylin is not able read the offset. Can someone help on this?
2017-10-11 20:50:42,573 INFO [http-bio-7070-exec-8] utils.AppInfoParser:83 : Kafka version : 0.10.2-kafka-2.2.0
2017-10-11 20:50:42,573 INFO [http-bio-7070-exec-8] utils.AppInfoParser:84 : Kafka commitId : unknown
2017-10-11 20:50:42,586 DEBUG [http-bio-7070-exec-8] kafka.KafkaSource:107 : The end offsets are {0=0}
2017-10-11 20:50:42,588 ERROR [http-bio-7070-exec-8] controller.CubeController:305 : No new message comes, startOffset = endOffset:0
java.lang.IllegalArgumentException: No new message comes, startOffset = endOffset:0
at org.apache.kylin.source.kafka.KafkaSource.enrichSourcePartitionBeforeBuild(KafkaSource.java:134)
at org.apache.kylin.rest.service.JobService.submitJobInternal(JobService.java:236)
at org.apache.kylin.rest.service.JobService.submitJob(JobService.java:208)
at org.apache.kylin.rest.service.JobService$$FastClassBySpringCGLIB$$83a44b2a.invoke(<generated>)
Regards,
Manoj
This message is confidential and subject to terms at: http://www.jpmorgan.com/emaildisclaimer including on confidentiality, legal privilege, viruses and monitoring of electronic messages. If you are not the intended recipient, please delete this message and notify the sender immediately. Any unauthorized use is strictly prohibited.
RE: Kafka Streaming data - Error while building the Cube
Posted by "Kumar, Manoj H" <ma...@jpmorgan.com>.
Excellent catch.. It was due to wrong Topic name. It ran fine when changed it.
Regards,
Manoj
From: Billy Liu [mailto:billyliu@apache.org]
Sent: Friday, October 13, 2017 12:05 PM
To: user
Subject: Re: Kafka Streaming data - Error while building the Cube
And in Kylin tutorial, the topic is kylindemo, in your sample, the topic is kylin_demo. Please double check the topic name.
2017-10-13 14:27 GMT+08:00 Billy Liu <bi...@apache.org>>:
If you could package the source code, please try to add more debug log when retrieving partition info by Kafka consumer. Check which topic and how many partitions you got.
2017-10-12 23:19 GMT+08:00 Kumar, Manoj H <ma...@jpmorgan.com>>:
I guess KafkaSource Class file where it has enrichSourcePartitionBeforeBuild() to get partition values. There its getting Error out. Do we know how can we test out to find why its coming as 0 for start & end offset?
Regards,
Manoj
From: Kumar, Manoj H
Sent: Thursday, October 12, 2017 3:35 PM
To: 'user@kylin.apache.org<ma...@kylin.apache.org>'
Subject: RE: Kafka Streaming data - Error while building the Cube
Yes its there.. I could see the messages..
Regards,
Manoj
From: Billy Liu [mailto:billyliu@apache.org]
Sent: Thursday, October 12, 2017 3:11 PM
To: user
Subject: Re: Kafka Streaming data - Error while building the Cube
STREAMING_SALES_TABLE table reads messages from Kafka topic kylin_demo,but got 0 message.
Could you check if the topic has incoming message: bin/kafka-console-consumer.sh --zookeeper localhost:2181 --bootstrap-server localhost:9092 --topic kylin_demo
2017-10-12 17:19 GMT+08:00 Kumar, Manoj H <ma...@jpmorgan.com>>:
Pls. find below information about consumer config from Kylin log file.
2017-10-11 02:11:43,787 INFO [pool-8-thread-1] threadpool.DefaultScheduler:123 : Job Fetcher: 0 should running, 0 actual running, 0 stopped, 0 ready, 1 already succeed, 0 error, 0 discarded, 0 others
2017-10-11 02:12:13,783 INFO [pool-8-thread-1] threadpool.DefaultScheduler:123 : Job Fetcher: 0 should running, 0 actual running, 0 stopped, 0 ready, 1 already succeed, 0 error, 0 discarded, 0 others
2017-10-11 02:12:40,734 INFO [http-bio-7070-exec-3] streaming.StreamingManager:222 : Reloading Streaming Metadata from folder kylin_metadata(key='/streaming')@kylin_metadata@hbase
2017-10-11 02:12:40,760 DEBUG [http-bio-7070-exec-3] streaming.StreamingManager:247 : Loaded 1 StreamingConfig(s)
2017-10-11 02:12:43,789 INFO [pool-8-thread-1] threadpool.DefaultScheduler:123 : Job Fetcher: 0 should running, 0 actual running, 0 stopped, 0 ready, 1 already succeed, 0 error, 0 discarded, 0 others
2017-10-11 02:13:13,788 INFO [pool-8-thread-1] threadpool.DefaultScheduler:123 : Job Fetcher: 0 should running, 0 actual running, 0 stopped, 0 ready, 1 already succeed, 0 error, 0 discarded, 0 others
2017-10-11 02:13:43,785 INFO [pool-8-thread-1] threadpool.DefaultScheduler:123 : Job Fetcher: 0 should running, 0 actual running, 0 stopped, 0 ready, 1 already succeed, 0 error, 0 discarded, 0 others
2017-10-11 02:14:13,789 INFO [pool-8-thread-1] threadpool.DefaultScheduler:123 : Job Fetcher: 0 should running, 0 actual running, 0 stopped, 0 ready, 1 already succeed, 0 error, 0 discarded, 0 others
2017-10-11 02:14:43,796 INFO [pool-8-thread-1] threadpool.DefaultScheduler:123 : Job Fetcher: 0 should running, 0 actual running, 0 stopped, 0 ready, 1 already succeed, 0 error, 0 discarded, 0 others
2017-10-11 02:15:03,911 DEBUG [http-bio-7070-exec-1] controller.StreamingController:255 : Saving StreamingConfig {"uuid":"8613b0e1-40ac-438c-bdf5-72be4d91c230","last_modified":1507705685859,"version":"2.1.0","name":"DEFAULT.STREAMING_SALES_TABLE","type":"kafka"}
2017-10-11 02:15:03,913 DEBUG [http-bio-7070-exec-1] controller.StreamingController:273 : Saving KafkaConfig {"uuid":"87dc6ab5-5141-4bd8-8e00-c16ec86dce41","last_modified":1507705685916,"version":"2.1.0","name":"DEFAULT.STREAMING_SALES_TABLE","clusters":[{"brokers":[{"id":"1","host":"sandbox","port":"9092"}]}],"topic":"kylin_demo","timeout":60000,"parserName":"org.apache.kylin.source.kafka.TimedJsonStreamParser","parserTimeStampField":null,"margin":0,"parserProperties":"tsColName=order_time"}
2017-10-11 02:15:03,963 DEBUG [pool-7-thread-1] cachesync.Broadcaster:132 : Servers in the cluster: [localhost:7070]
2017-10-11 02:15:04,000 DEBUG [pool-7-thread-1] cachesync.Broadcaster:139 : Announcing new broadcast event: BroadcastEvent{entity=streaming, event=update, cacheKey=DEFAULT.STREAMING_SALES_TABLE}
2017-10-11 02:15:04,009 DEBUG [pool-7-thread-1] cachesync.Broadcaster:132 : Servers in the cluster: [localhost:7070]
2017-10-11 02:15:04,009 DEBUG [pool-7-thread-1] cachesync.Broadcaster:139 : Announcing new broadcast event: BroadcastEvent{entity=kafka, event=update, cacheKey=DEFAULT.STREAMING_SALES_TABLE}
2017-10-11 02:15:04,164 DEBUG [http-bio-7070-exec-9] cachesync.Broadcaster:236 : Done broadcasting metadata change: entity=streaming, event=UPDATE, cacheKey=DEFAULT.STREAMING_SALES_TABLE
2017-10-11 02:15:04,192 DEBUG [http-bio-7070-exec-10] cachesync.Broadcaster:236 : Done broadcasting metadata change: entity=kafka, event=UPDATE, cacheKey=DEFAULT.STREAMING_SALES_TABLE
2017-10-11 02:15:13,789 INFO [pool-8-thread-1] threadpool.DefaultScheduler:123 : Job Fetcher: 0 should running, 0 actual running, 0 stopped, 0 ready, 1 already succeed, 0 error, 0 discarded, 0 others
2017-10-11 02:15:23,780 DEBUG [http-bio-7070-exec-7] kafka.KafkaSource:83 : Last segment doesn't exist, and didn't initiate the start offset, will seek from topic's earliest offset.
2017-10-11 20:50:42,558 INFO [http-bio-7070-exec-8] utils.AppInfoParser:83 : Kafka version : 0.10.2-kafka-2.2.0
2017-10-11 20:50:42,563 INFO [http-bio-7070-exec-8] utils.AppInfoParser:84 : Kafka commitId : unknown
2017-10-11 20:50:42,570 DEBUG [http-bio-7070-exec-8] kafka.KafkaSource:105 : Seek end offsets from topic
2017-10-11 20:50:42,570 INFO [http-bio-7070-exec-8] consumer.ConsumerConfig:196 : ConsumerConfig values:
auto.commit.interval.ms<http://auto.commit.interval.ms> = 5000
auto.offset.reset = latest
bootstrap.servers = [localhost:9092]
check.crcs = true
client.id<http://client.id> =
connections.max.idle.ms<http://connections.max.idle.ms> = 540000
enable.auto.commit = false
exclude.internal.topics = true
fetch.max.bytes = 52428800
fetch.max.wait.ms<http://fetch.max.wait.ms> = 500
fetch.min.bytes = 1
group.id<http://group.id> = streaming_cube
heartbeat.interval.ms<http://heartbeat.interval.ms> = 3000
interceptor.classes = null
internal.leave.group.on.close = true
key.deserializer = class org.apache.kafka.common.serialization.StringDeserializer
max.partition.fetch.bytes = 1048576
max.poll.interval.ms<http://max.poll.interval.ms> = 300000
max.poll.records = 500
metadata.max.age.ms<http://metadata.max.age.ms> = 300000
metric.reporters = []
metrics.num.samples = 2
metrics.recording.level = INFO
metrics.sample.window.ms<http://metrics.sample.window.ms> = 30000
partition.assignment.strategy = [class org.apache.kafka.clients.consumer.RangeAssignor]
receive.buffer.bytes = 65536
reconnect.backoff.ms<http://reconnect.backoff.ms> = 50
request.timeout.ms<http://request.timeout.ms> = 305000
retry.backoff.ms<http://retry.backoff.ms> = 100
sasl.jaas.config = null
sasl.kerberos.kinit.cmd = /usr/bin/kinit
sasl.kerberos.min.time.before.relogin = 60000
sasl.kerberos.service.name<http://sasl.kerberos.service.name> = null
sasl.kerberos.ticket.renew.jitter = 0.05
request.timeout.ms<http://request.timeout.ms> = 305000
retry.backoff.ms<http://retry.backoff.ms> = 100
sasl.jaas.config = null
sasl.kerberos.kinit.cmd = /usr/bin/kinit
sasl.kerberos.min.time.before.relogin = 60000
sasl.kerberos.service.name<http://sasl.kerberos.service.name> = null
sasl.kerberos.ticket.renew.jitter = 0.05
sasl.kerberos.ticket.renew.window.factor = 0.8
sasl.mechanism = GSSAPI
security.protocol = PLAINTEXT
send.buffer.bytes = 131072
session.timeout.ms<http://session.timeout.ms> = 10000
ssl.cipher.suites = null
ssl.enabled.protocols = [TLSv1.2, TLSv1.1, TLSv1]
ssl.endpoint.identification.al<http://ssl.endpoint.identification.al>gorithm = null
ssl.key.password = null
ssl.keymanager.algorithm = SunX509
ssl.keystore.location = null
ssl.keystore.password = null
ssl.keystore.type = JKS
ssl.protocol = TLS
ssl.provider = null
ssl.secure.random.implementation = null
ssl.trustmanager.algorithm = PKIX
ssl.truststore.location = null
ssl.truststore.password = null
ssl.truststore.type = JKS
value.deserializer = class org.apache.kafka.common.serialization.StringDeserializer
2017-10-11 20:50:42,573 INFO [http-bio-7070-exec-8] utils.AppInfoParser:83 : Kafka version : 0.10.2-kafka-2.2.0
2017-10-11 20:50:42,573 INFO [http-bio-7070-exec-8] utils.AppInfoParser:84 : Kafka commitId : unknown
2017-10-11 20:50:42,586 DEBUG [http-bio-7070-exec-8] kafka.KafkaSource:107 : The end offsets are {0=0}
2017-10-11 20:50:42,588 ERROR [http-bio-7070-exec-8] controller.CubeController:305 : No new message comes, startOffset = endOffset:0
java.lang.IllegalArgumentException: No new message comes, startOffset = endOffset:0
at org.apache.kylin.source.kafka.KafkaSource.enrichSourcePartitionBeforeBuild(KafkaSource.java:134)
at org.apache.kylin.rest.service.JobService.submitJobInternal(JobService.java:236)
Regards,
Manoj
From: Billy Liu [mailto:billyliu@apache.org<ma...@apache.org>]
Sent: Thursday, October 12, 2017 1:06 PM
To: user
Subject: Re: Kafka Streaming data - Error while building the Cube
Hi Kumar,
Could you paste more Kafka Consumer related log in kylin.log? And also check from the Kafka broker side, if the Kylin client has connected to Broker.
2017-10-12 14:29 GMT+08:00 Kumar, Manoj H <ma...@jpmorgan.com>>:
Building the Cube from Kylin UI - Although Messages are there in Kafka topic but Kylin is not able read the offset. Can someone help on this?
2017-10-11 20:50:42,573 INFO [http-bio-7070-exec-8] utils.AppInfoParser:83 : Kafka version : 0.10.2-kafka-2.2.0
2017-10-11 20:50:42,573 INFO [http-bio-7070-exec-8] utils.AppInfoParser:84 : Kafka commitId : unknown
2017-10-11 20:50:42,586 DEBUG [http-bio-7070-exec-8] kafka.KafkaSource:107 : The end offsets are {0=0}
2017-10-11 20:50:42,588 ERROR [http-bio-7070-exec-8] controller.CubeController:305 : No new message comes, startOffset = endOffset:0
java.lang.IllegalArgumentException: No new message comes, startOffset = endOffset:0
at org.apache.kylin.source.kafka.KafkaSource.enrichSourcePartitionBeforeBuild(KafkaSource.java:134)
at org.apache.kylin.rest.service.JobService.submitJobInternal(JobService.java:236)
at org.apache.kylin.rest.service.JobService.submitJob(JobService.java:208)
at org.apache.kylin.rest.service.JobService$$FastClassBySpringCGLIB$$83a44b2a.invoke(<generated>)
Regards,
Manoj
This message is confidential and subject to terms at: http://www.jpmorgan.com/emaildisclaimer including on confidentiality, legal privilege, viruses and monitoring of electronic messages. If you are not the intended recipient, please delete this message and notify the sender immediately. Any unauthorized use is strictly prohibited.
This message is confidential and subject to terms at: http://www.jpmorgan.com/emaildisclaimer<http://www.jpmorgan.com/emaildisclaimer> including on confidentiality, legal privilege, viruses and monitoring of electronic messages. If you are not the intended recipient, please delete this message and notify the sender immediately. Any unauthorized use is strictly prohibited.
This message is confidential and subject to terms at: http://www.jpmorgan.com/emaildisclaimer<http://www.jpmorgan.com/emaildisclaimer> including on confidentiality, legal privilege, viruses and monitoring of electronic messages. If you are not the intended recipient, please delete this message and notify the sender immediately. Any unauthorized use is strictly prohibited.
This message is confidential and subject to terms at: http://www.jpmorgan.com/emaildisclaimer including on confidentiality, legal privilege, viruses and monitoring of electronic messages. If you are not the intended recipient, please delete this message and notify the sender immediately. Any unauthorized use is strictly prohibited.
Re: Kafka Streaming data - Error while building the Cube
Posted by Billy Liu <bi...@apache.org>.
And in Kylin tutorial, the topic is kylindemo, in your sample, the topic is
kylin_demo. Please double check the topic name.
2017-10-13 14:27 GMT+08:00 Billy Liu <bi...@apache.org>:
> If you could package the source code, please try to add more debug log
> when retrieving partition info by Kafka consumer. Check which topic and how
> many partitions you got.
>
> 2017-10-12 23:19 GMT+08:00 Kumar, Manoj H <ma...@jpmorgan.com>:
>
>> I guess KafkaSource Class file where it has
>> enrichSourcePartitionBeforeBuild() to get partition values. There its
>> getting Error out. Do we know how can we test out to find why its coming as
>> 0 for start & end offset?
>>
>>
>>
>> Regards,
>>
>> Manoj
>>
>>
>>
>> *From:* Kumar, Manoj H
>> *Sent:* Thursday, October 12, 2017 3:35 PM
>> *To:* 'user@kylin.apache.org'
>> *Subject:* RE: Kafka Streaming data - Error while building the Cube
>>
>>
>>
>> Yes its there.. I could see the messages..
>>
>>
>>
>> Regards,
>>
>> Manoj
>>
>>
>>
>> *From:* Billy Liu [mailto:billyliu@apache.org <bi...@apache.org>]
>> *Sent:* Thursday, October 12, 2017 3:11 PM
>>
>> *To:* user
>> *Subject:* Re: Kafka Streaming data - Error while building the Cube
>>
>>
>>
>> STREAMING_SALES_TABLE table reads messages from Kafka topic kylin_demo,but
>> got 0 message.
>>
>>
>>
>> Could you check if the topic has incoming message: bin/kafka-console-consumer.sh
>> --zookeeper localhost:2181 --bootstrap-server localhost:9092 --topic
>> kylin_demo
>>
>>
>>
>> 2017-10-12 17:19 GMT+08:00 Kumar, Manoj H <ma...@jpmorgan.com>:
>>
>> Pls. find below information about consumer config from Kylin log file.
>>
>>
>>
>> 2017-10-11 02:11:43,787 INFO [pool-8-thread-1]
>> threadpool.DefaultScheduler:123 : Job Fetcher: 0 should running, 0
>> actual running, 0 stopped, 0 ready, 1 already succeed, 0 error, 0
>> discarded, 0 others
>>
>> 2017-10-11 02:12:13,783 INFO [pool-8-thread-1]
>> threadpool.DefaultScheduler:123 : Job Fetcher: 0 should running, 0
>> actual running, 0 stopped, 0 ready, 1 already succeed, 0 error, 0
>> discarded, 0 others
>>
>> 2017-10-11 02:12:40,734 INFO [http-bio-7070-exec-3]
>> streaming.StreamingManager:222 : Reloading Streaming Metadata from folder
>> kylin_metadata(key='/streaming')@kylin_metadata@hbase
>>
>> 2017-10-11 02:12:40,760 DEBUG [http-bio-7070-exec-3]
>> streaming.StreamingManager:247 : Loaded 1 StreamingConfig(s)
>>
>> 2017-10-11 02:12:43,789 INFO [pool-8-thread-1]
>> threadpool.DefaultScheduler:123 : Job Fetcher: 0 should running, 0
>> actual running, 0 stopped, 0 ready, 1 already succeed, 0 error, 0
>> discarded, 0 others
>>
>> 2017-10-11 02:13:13,788 INFO [pool-8-thread-1]
>> threadpool.DefaultScheduler:123 : Job Fetcher: 0 should running, 0
>> actual running, 0 stopped, 0 ready, 1 already succeed, 0 error, 0
>> discarded, 0 others
>>
>> 2017-10-11 02:13:43,785 INFO [pool-8-thread-1]
>> threadpool.DefaultScheduler:123 : Job Fetcher: 0 should running, 0
>> actual running, 0 stopped, 0 ready, 1 already succeed, 0 error, 0
>> discarded, 0 others
>>
>> 2017-10-11 02:14:13,789 INFO [pool-8-thread-1]
>> threadpool.DefaultScheduler:123 : Job Fetcher: 0 should running, 0
>> actual running, 0 stopped, 0 ready, 1 already succeed, 0 error, 0
>> discarded, 0 others
>>
>> 2017-10-11 02:14:43,796 INFO [pool-8-thread-1]
>> threadpool.DefaultScheduler:123 : Job Fetcher: 0 should running, 0
>> actual running, 0 stopped, 0 ready, 1 already succeed, 0 error, 0
>> discarded, 0 others
>>
>> 2017-10-11 02:15:03,911 DEBUG [http-bio-7070-exec-1]
>> controller.StreamingController:255 : Saving StreamingConfig
>> {"uuid":"8613b0e1-40ac-438c-bdf5-72be4d91c230","last_modifie
>> d":1507705685859,"version":"2.1.0","name":"DEFAULT.
>> STREAMING_SALES_TABLE","type":"kafka"}
>>
>> 2017-10-11 02:15:03,913 DEBUG [http-bio-7070-exec-1]
>> controller.StreamingController:273 : Saving KafkaConfig
>> {"uuid":"87dc6ab5-5141-4bd8-8e00-c16ec86dce41","last_modifie
>> d":1507705685916,"version":"2.1.0","name":"DEFAULT.
>> STREAMING_SALES_TABLE","clusters":[{"brokers":[{"id":"
>> 1","host":"sandbox","port":"9092"}]}],"topic":"kylin_demo","
>> timeout":60000,"parserName":"org.apache.kylin.source.kafka.
>> TimedJsonStreamParser","parserTimeStampField":null,"margin":
>> 0,"parserProperties":"tsColName=order_time"}
>>
>> 2017-10-11 02:15:03,963 DEBUG [pool-7-thread-1] cachesync.Broadcaster:132
>> : Servers in the cluster: [localhost:7070]
>>
>> 2017-10-11 02:15:04,000 DEBUG [pool-7-thread-1] cachesync.Broadcaster:139
>> : Announcing new broadcast event: BroadcastEvent{entity=streaming,
>> event=update, cacheKey=DEFAULT.STREAMING_SALES_TABLE}
>>
>> 2017-10-11 02:15:04,009 DEBUG [pool-7-thread-1] cachesync.Broadcaster:132
>> : Servers in the cluster: [localhost:7070]
>>
>> 2017-10-11 02:15:04,009 DEBUG [pool-7-thread-1] cachesync.Broadcaster:139
>> : Announcing new broadcast event: BroadcastEvent{entity=kafka,
>> event=update, cacheKey=DEFAULT.STREAMING_SALES_TABLE}
>>
>> 2017-10-11 02:15:04,164 DEBUG [http-bio-7070-exec-9]
>> cachesync.Broadcaster:236 : Done broadcasting metadata change:
>> entity=streaming, event=UPDATE, cacheKey=DEFAULT.STREAMING_SALES_TABLE
>>
>> 2017-10-11 02:15:04,192 DEBUG [http-bio-7070-exec-10]
>> cachesync.Broadcaster:236 : Done broadcasting metadata change:
>> entity=kafka, event=UPDATE, cacheKey=DEFAULT.STREAMING_SALES_TABLE
>>
>> 2017-10-11 02:15:13,789 INFO [pool-8-thread-1]
>> threadpool.DefaultScheduler:123 : Job Fetcher: 0 should running, 0
>> actual running, 0 stopped, 0 ready, 1 already succeed, 0 error, 0
>> discarded, 0 others
>>
>> 2017-10-11 02:15:23,780 DEBUG [http-bio-7070-exec-7] kafka.KafkaSource:83
>> : Last segment doesn't exist, and didn't initiate the start offset, will
>> seek from topic's earliest offset.
>>
>>
>>
>> 2017-10-11 20:50:42,558 INFO [http-bio-7070-exec-8]
>> utils.AppInfoParser:83 : Kafka version : 0.10.2-kafka-2.2.0
>>
>> 2017-10-11 20:50:42,563 INFO [http-bio-7070-exec-8]
>> utils.AppInfoParser:84 : Kafka commitId : unknown
>>
>> 2017-10-11 20:50:42,570 DEBUG [http-bio-7070-exec-8]
>> kafka.KafkaSource:105 : Seek end offsets from topic
>>
>> 2017-10-11 20:50:42,570 INFO [http-bio-7070-exec-8]
>> consumer.ConsumerConfig:196 : ConsumerConfig values:
>>
>> auto.commit.interval.ms = 5000
>>
>> auto.offset.reset = latest
>>
>> bootstrap.servers = [localhost:9092]
>>
>> check.crcs = true
>>
>> client.id =
>>
>> connections.max.idle.ms = 540000
>>
>> enable.auto.commit = false
>>
>> exclude.internal.topics = true
>>
>> fetch.max.bytes = 52428800
>>
>> fetch.max.wait.ms = 500
>>
>> fetch.min.bytes = 1
>>
>> group.id = streaming_cube
>>
>> heartbeat.interval.ms = 3000
>>
>> interceptor.classes = null
>>
>> internal.leave.group.on.close = true
>>
>> key.deserializer = class org.apache.kafka.common.serial
>> ization.StringDeserializer
>>
>> max.partition.fetch.bytes = 1048576
>>
>> max.poll.interval.ms = 300000
>>
>> max.poll.records = 500
>>
>> metadata.max.age.ms = 300000
>>
>> metric.reporters = []
>>
>> metrics.num.samples = 2
>>
>> metrics.recording.level = INFO
>>
>> metrics.sample.window.ms = 30000
>>
>> partition.assignment.strategy = [class
>> org.apache.kafka.clients.consumer.RangeAssignor]
>>
>> receive.buffer.bytes = 65536
>>
>> reconnect.backoff.ms = 50
>>
>> request.timeout.ms = 305000
>>
>> retry.backoff.ms = 100
>>
>> sasl.jaas.config = null
>>
>> sasl.kerberos.kinit.cmd = /usr/bin/kinit
>>
>> sasl.kerberos.min.time.before.relogin = 60000
>>
>> sasl.kerberos.service.name = null
>>
>> sasl.kerberos.ticket.renew.jitter = 0.05
>>
>> request.timeout.ms = 305000
>>
>> retry.backoff.ms = 100
>>
>> sasl.jaas.config = null
>>
>> sasl.kerberos.kinit.cmd = /usr/bin/kinit
>>
>> sasl.kerberos.min.time.before.relogin = 60000
>>
>> sasl.kerberos.service.name = null
>>
>> sasl.kerberos.ticket.renew.jitter = 0.05
>>
>> sasl.kerberos.ticket.renew.window.factor = 0.8
>>
>> sasl.mechanism = GSSAPI
>>
>> security.protocol = PLAINTEXT
>>
>> send.buffer.bytes = 131072
>>
>> session.timeout.ms = 10000
>>
>> ssl.cipher.suites = null
>>
>> ssl.enabled.protocols = [TLSv1.2, TLSv1.1, TLSv1]
>>
>> ssl.endpoint.identification.algorithm = null
>>
>> ssl.key.password = null
>>
>> ssl.keymanager.algorithm = SunX509
>>
>> ssl.keystore.location = null
>>
>> ssl.keystore.password = null
>>
>> ssl.keystore.type = JKS
>>
>> ssl.protocol = TLS
>>
>> ssl.provider = null
>>
>> ssl.secure.random.implementation = null
>>
>> ssl.trustmanager.algorithm = PKIX
>>
>> ssl.truststore.location = null
>>
>> ssl.truststore.password = null
>>
>> ssl.truststore.type = JKS
>>
>> value.deserializer = class org.apache.kafka.common.serial
>> ization.StringDeserializer
>>
>>
>>
>> 2017-10-11 20:50:42,573 INFO [http-bio-7070-exec-8]
>> utils.AppInfoParser:83 : Kafka version : 0.10.2-kafka-2.2.0
>>
>> 2017-10-11 20:50:42,573 INFO [http-bio-7070-exec-8]
>> utils.AppInfoParser:84 : Kafka commitId : unknown
>>
>> 2017-10-11 20:50:42,586 DEBUG [http-bio-7070-exec-8]
>> kafka.KafkaSource:107 : The end offsets are {0=0}
>>
>> 2017-10-11 20:50:42,588 ERROR [http-bio-7070-exec-8]
>> controller.CubeController:305 : No new message comes, startOffset =
>> endOffset:0
>>
>> java.lang.IllegalArgumentException: No new message comes, startOffset =
>> endOffset:0
>>
>> at org.apache.kylin.source.kafka.KafkaSource.enrichSourcePartit
>> ionBeforeBuild(KafkaSource.java:134)
>>
>> at org.apache.kylin.rest.service.JobService.submitJobInternal(J
>> obService.java:236)
>>
>> Regards,
>>
>> Manoj
>>
>>
>>
>> *From:* Billy Liu [mailto:billyliu@apache.org]
>> *Sent:* Thursday, October 12, 2017 1:06 PM
>> *To:* user
>> *Subject:* Re: Kafka Streaming data - Error while building the Cube
>>
>>
>>
>> Hi Kumar,
>>
>>
>>
>> Could you paste more Kafka Consumer related log in kylin.log? And also
>> check from the Kafka broker side, if the Kylin client has connected to
>> Broker.
>>
>>
>>
>> 2017-10-12 14:29 GMT+08:00 Kumar, Manoj H <ma...@jpmorgan.com>:
>>
>> Building the Cube from Kylin UI - Although Messages are there in Kafka
>> topic but Kylin is not able read the offset. Can someone help on this?
>>
>> 2017-10-11 20:50:42,573 INFO [http-bio-7070-exec-8]
>> utils.AppInfoParser:83 : Kafka version : 0.10.2-kafka-2.2.0
>> 2017-10-11 20:50:42,573 INFO [http-bio-7070-exec-8]
>> utils.AppInfoParser:84 : Kafka commitId : unknown
>> 2017-10-11 20:50:42,586 DEBUG [http-bio-7070-exec-8]
>> kafka.KafkaSource:107 : The end offsets are {0=0}
>> 2017-10-11 20:50:42,588 ERROR [http-bio-7070-exec-8]
>> controller.CubeController:305 : No new message comes, startOffset =
>> endOffset:0
>> java.lang.IllegalArgumentException: No new message comes, startOffset =
>> endOffset:0
>> at org.apache.kylin.source.kafka.KafkaSource.enrichSourcePartit
>> ionBeforeBuild(KafkaSource.java:134)
>> at org.apache.kylin.rest.service.JobService.submitJobInternal(J
>> obService.java:236)
>> at org.apache.kylin.rest.service.JobService.submitJob(JobServic
>> e.java:208)
>> at org.apache.kylin.rest.service.JobService$$FastClassBySpringC
>> GLIB$$83a44b2a.invoke(<generated>)
>>
>> Regards,
>> Manoj
>>
>>
>> This message is confidential and subject to terms at:
>> http://www.jpmorgan.com/emaildisclaimer including on confidentiality,
>> legal privilege, viruses and monitoring of electronic messages. If you are
>> not the intended recipient, please delete this message and notify the
>> sender immediately. Any unauthorized use is strictly prohibited.
>>
>>
>>
>> This message is confidential and subject to terms at: http://
>> www.jpmorgan.com/emaildisclaimer including on confidentiality, legal
>> privilege, viruses and monitoring of electronic messages. If you are not
>> the intended recipient, please delete this message and notify the sender
>> immediately. Any unauthorized use is strictly prohibited.
>>
>>
>>
>> This message is confidential and subject to terms at: http://
>> www.jpmorgan.com/emaildisclaimer including on confidentiality, legal
>> privilege, viruses and monitoring of electronic messages. If you are not
>> the intended recipient, please delete this message and notify the sender
>> immediately. Any unauthorized use is strictly prohibited.
>>
>
>
Re: Kafka Streaming data - Error while building the Cube
Posted by Billy Liu <bi...@apache.org>.
If you could package the source code, please try to add more debug log when
retrieving partition info by Kafka consumer. Check which topic and how many
partitions you got.
2017-10-12 23:19 GMT+08:00 Kumar, Manoj H <ma...@jpmorgan.com>:
> I guess KafkaSource Class file where it has enrichSourcePartitionBeforeBuild()
> to get partition values. There its getting Error out. Do we know how can we
> test out to find why its coming as 0 for start & end offset?
>
>
>
> Regards,
>
> Manoj
>
>
>
> *From:* Kumar, Manoj H
> *Sent:* Thursday, October 12, 2017 3:35 PM
> *To:* 'user@kylin.apache.org'
> *Subject:* RE: Kafka Streaming data - Error while building the Cube
>
>
>
> Yes its there.. I could see the messages..
>
>
>
> Regards,
>
> Manoj
>
>
>
> *From:* Billy Liu [mailto:billyliu@apache.org <bi...@apache.org>]
> *Sent:* Thursday, October 12, 2017 3:11 PM
>
> *To:* user
> *Subject:* Re: Kafka Streaming data - Error while building the Cube
>
>
>
> STREAMING_SALES_TABLE table reads messages from Kafka topic kylin_demo,but
> got 0 message.
>
>
>
> Could you check if the topic has incoming message: bin/kafka-console-consumer.sh
> --zookeeper localhost:2181 --bootstrap-server localhost:9092 --topic
> kylin_demo
>
>
>
> 2017-10-12 17:19 GMT+08:00 Kumar, Manoj H <ma...@jpmorgan.com>:
>
> Pls. find below information about consumer config from Kylin log file.
>
>
>
> 2017-10-11 02:11:43,787 INFO [pool-8-thread-1]
> threadpool.DefaultScheduler:123 : Job Fetcher: 0 should running, 0 actual
> running, 0 stopped, 0 ready, 1 already succeed, 0 error, 0 discarded, 0
> others
>
> 2017-10-11 02:12:13,783 INFO [pool-8-thread-1]
> threadpool.DefaultScheduler:123 : Job Fetcher: 0 should running, 0 actual
> running, 0 stopped, 0 ready, 1 already succeed, 0 error, 0 discarded, 0
> others
>
> 2017-10-11 02:12:40,734 INFO [http-bio-7070-exec-3]
> streaming.StreamingManager:222 : Reloading Streaming Metadata from folder
> kylin_metadata(key='/streaming')@kylin_metadata@hbase
>
> 2017-10-11 02:12:40,760 DEBUG [http-bio-7070-exec-3]
> streaming.StreamingManager:247 : Loaded 1 StreamingConfig(s)
>
> 2017-10-11 02:12:43,789 INFO [pool-8-thread-1]
> threadpool.DefaultScheduler:123 : Job Fetcher: 0 should running, 0 actual
> running, 0 stopped, 0 ready, 1 already succeed, 0 error, 0 discarded, 0
> others
>
> 2017-10-11 02:13:13,788 INFO [pool-8-thread-1]
> threadpool.DefaultScheduler:123 : Job Fetcher: 0 should running, 0 actual
> running, 0 stopped, 0 ready, 1 already succeed, 0 error, 0 discarded, 0
> others
>
> 2017-10-11 02:13:43,785 INFO [pool-8-thread-1]
> threadpool.DefaultScheduler:123 : Job Fetcher: 0 should running, 0 actual
> running, 0 stopped, 0 ready, 1 already succeed, 0 error, 0 discarded, 0
> others
>
> 2017-10-11 02:14:13,789 INFO [pool-8-thread-1]
> threadpool.DefaultScheduler:123 : Job Fetcher: 0 should running, 0 actual
> running, 0 stopped, 0 ready, 1 already succeed, 0 error, 0 discarded, 0
> others
>
> 2017-10-11 02:14:43,796 INFO [pool-8-thread-1]
> threadpool.DefaultScheduler:123 : Job Fetcher: 0 should running, 0 actual
> running, 0 stopped, 0 ready, 1 already succeed, 0 error, 0 discarded, 0
> others
>
> 2017-10-11 02:15:03,911 DEBUG [http-bio-7070-exec-1] controller.StreamingController:255
> : Saving StreamingConfig {"uuid":"8613b0e1-40ac-438c-
> bdf5-72be4d91c230","last_modified":1507705685859,"
> version":"2.1.0","name":"DEFAULT.STREAMING_SALES_TABLE","type":"kafka"}
>
> 2017-10-11 02:15:03,913 DEBUG [http-bio-7070-exec-1] controller.StreamingController:273
> : Saving KafkaConfig {"uuid":"87dc6ab5-5141-4bd8-8e00-c16ec86dce41","last_
> modified":1507705685916,"version":"2.1.0","name":"
> DEFAULT.STREAMING_SALES_TABLE","clusters":[{"brokers":[{"id"
> :"1","host":"sandbox","port":"9092"}]}],"topic":"kylin_demo"
> ,"timeout":60000,"parserName":"org.apache.kylin.source.
> kafka.TimedJsonStreamParser","parserTimeStampField":null,"
> margin":0,"parserProperties":"tsColName=order_time"}
>
> 2017-10-11 02:15:03,963 DEBUG [pool-7-thread-1] cachesync.Broadcaster:132
> : Servers in the cluster: [localhost:7070]
>
> 2017-10-11 02:15:04,000 DEBUG [pool-7-thread-1] cachesync.Broadcaster:139
> : Announcing new broadcast event: BroadcastEvent{entity=streaming,
> event=update, cacheKey=DEFAULT.STREAMING_SALES_TABLE}
>
> 2017-10-11 02:15:04,009 DEBUG [pool-7-thread-1] cachesync.Broadcaster:132
> : Servers in the cluster: [localhost:7070]
>
> 2017-10-11 02:15:04,009 DEBUG [pool-7-thread-1] cachesync.Broadcaster:139
> : Announcing new broadcast event: BroadcastEvent{entity=kafka,
> event=update, cacheKey=DEFAULT.STREAMING_SALES_TABLE}
>
> 2017-10-11 02:15:04,164 DEBUG [http-bio-7070-exec-9]
> cachesync.Broadcaster:236 : Done broadcasting metadata change:
> entity=streaming, event=UPDATE, cacheKey=DEFAULT.STREAMING_SALES_TABLE
>
> 2017-10-11 02:15:04,192 DEBUG [http-bio-7070-exec-10]
> cachesync.Broadcaster:236 : Done broadcasting metadata change:
> entity=kafka, event=UPDATE, cacheKey=DEFAULT.STREAMING_SALES_TABLE
>
> 2017-10-11 02:15:13,789 INFO [pool-8-thread-1]
> threadpool.DefaultScheduler:123 : Job Fetcher: 0 should running, 0 actual
> running, 0 stopped, 0 ready, 1 already succeed, 0 error, 0 discarded, 0
> others
>
> 2017-10-11 02:15:23,780 DEBUG [http-bio-7070-exec-7] kafka.KafkaSource:83
> : Last segment doesn't exist, and didn't initiate the start offset, will
> seek from topic's earliest offset.
>
>
>
> 2017-10-11 20:50:42,558 INFO [http-bio-7070-exec-8]
> utils.AppInfoParser:83 : Kafka version : 0.10.2-kafka-2.2.0
>
> 2017-10-11 20:50:42,563 INFO [http-bio-7070-exec-8]
> utils.AppInfoParser:84 : Kafka commitId : unknown
>
> 2017-10-11 20:50:42,570 DEBUG [http-bio-7070-exec-8] kafka.KafkaSource:105
> : Seek end offsets from topic
>
> 2017-10-11 20:50:42,570 INFO [http-bio-7070-exec-8]
> consumer.ConsumerConfig:196 : ConsumerConfig values:
>
> auto.commit.interval.ms = 5000
>
> auto.offset.reset = latest
>
> bootstrap.servers = [localhost:9092]
>
> check.crcs = true
>
> client.id =
>
> connections.max.idle.ms = 540000
>
> enable.auto.commit = false
>
> exclude.internal.topics = true
>
> fetch.max.bytes = 52428800
>
> fetch.max.wait.ms = 500
>
> fetch.min.bytes = 1
>
> group.id = streaming_cube
>
> heartbeat.interval.ms = 3000
>
> interceptor.classes = null
>
> internal.leave.group.on.close = true
>
> key.deserializer = class org.apache.kafka.common.serialization.
> StringDeserializer
>
> max.partition.fetch.bytes = 1048576
>
> max.poll.interval.ms = 300000
>
> max.poll.records = 500
>
> metadata.max.age.ms = 300000
>
> metric.reporters = []
>
> metrics.num.samples = 2
>
> metrics.recording.level = INFO
>
> metrics.sample.window.ms = 30000
>
> partition.assignment.strategy = [class org.apache.kafka.clients.
> consumer.RangeAssignor]
>
> receive.buffer.bytes = 65536
>
> reconnect.backoff.ms = 50
>
> request.timeout.ms = 305000
>
> retry.backoff.ms = 100
>
> sasl.jaas.config = null
>
> sasl.kerberos.kinit.cmd = /usr/bin/kinit
>
> sasl.kerberos.min.time.before.relogin = 60000
>
> sasl.kerberos.service.name = null
>
> sasl.kerberos.ticket.renew.jitter = 0.05
>
> request.timeout.ms = 305000
>
> retry.backoff.ms = 100
>
> sasl.jaas.config = null
>
> sasl.kerberos.kinit.cmd = /usr/bin/kinit
>
> sasl.kerberos.min.time.before.relogin = 60000
>
> sasl.kerberos.service.name = null
>
> sasl.kerberos.ticket.renew.jitter = 0.05
>
> sasl.kerberos.ticket.renew.window.factor = 0.8
>
> sasl.mechanism = GSSAPI
>
> security.protocol = PLAINTEXT
>
> send.buffer.bytes = 131072
>
> session.timeout.ms = 10000
>
> ssl.cipher.suites = null
>
> ssl.enabled.protocols = [TLSv1.2, TLSv1.1, TLSv1]
>
> ssl.endpoint.identification.algorithm = null
>
> ssl.key.password = null
>
> ssl.keymanager.algorithm = SunX509
>
> ssl.keystore.location = null
>
> ssl.keystore.password = null
>
> ssl.keystore.type = JKS
>
> ssl.protocol = TLS
>
> ssl.provider = null
>
> ssl.secure.random.implementation = null
>
> ssl.trustmanager.algorithm = PKIX
>
> ssl.truststore.location = null
>
> ssl.truststore.password = null
>
> ssl.truststore.type = JKS
>
> value.deserializer = class org.apache.kafka.common.serialization.
> StringDeserializer
>
>
>
> 2017-10-11 20:50:42,573 INFO [http-bio-7070-exec-8]
> utils.AppInfoParser:83 : Kafka version : 0.10.2-kafka-2.2.0
>
> 2017-10-11 20:50:42,573 INFO [http-bio-7070-exec-8]
> utils.AppInfoParser:84 : Kafka commitId : unknown
>
> 2017-10-11 20:50:42,586 DEBUG [http-bio-7070-exec-8] kafka.KafkaSource:107
> : The end offsets are {0=0}
>
> 2017-10-11 20:50:42,588 ERROR [http-bio-7070-exec-8]
> controller.CubeController:305 : No new message comes, startOffset =
> endOffset:0
>
> java.lang.IllegalArgumentException: No new message comes, startOffset =
> endOffset:0
>
> at org.apache.kylin.source.kafka.KafkaSource.
> enrichSourcePartitionBeforeBuild(KafkaSource.java:134)
>
> at org.apache.kylin.rest.service.JobService.submitJobInternal(
> JobService.java:236)
>
> Regards,
>
> Manoj
>
>
>
> *From:* Billy Liu [mailto:billyliu@apache.org]
> *Sent:* Thursday, October 12, 2017 1:06 PM
> *To:* user
> *Subject:* Re: Kafka Streaming data - Error while building the Cube
>
>
>
> Hi Kumar,
>
>
>
> Could you paste more Kafka Consumer related log in kylin.log? And also
> check from the Kafka broker side, if the Kylin client has connected to
> Broker.
>
>
>
> 2017-10-12 14:29 GMT+08:00 Kumar, Manoj H <ma...@jpmorgan.com>:
>
> Building the Cube from Kylin UI - Although Messages are there in Kafka
> topic but Kylin is not able read the offset. Can someone help on this?
>
> 2017-10-11 20:50:42,573 INFO [http-bio-7070-exec-8]
> utils.AppInfoParser:83 : Kafka version : 0.10.2-kafka-2.2.0
> 2017-10-11 20:50:42,573 INFO [http-bio-7070-exec-8]
> utils.AppInfoParser:84 : Kafka commitId : unknown
> 2017-10-11 20:50:42,586 DEBUG [http-bio-7070-exec-8] kafka.KafkaSource:107
> : The end offsets are {0=0}
> 2017-10-11 20:50:42,588 ERROR [http-bio-7070-exec-8]
> controller.CubeController:305 : No new message comes, startOffset =
> endOffset:0
> java.lang.IllegalArgumentException: No new message comes, startOffset =
> endOffset:0
> at org.apache.kylin.source.kafka.KafkaSource.
> enrichSourcePartitionBeforeBuild(KafkaSource.java:134)
> at org.apache.kylin.rest.service.JobService.submitJobInternal(
> JobService.java:236)
> at org.apache.kylin.rest.service.JobService.submitJob(
> JobService.java:208)
> at org.apache.kylin.rest.service.JobService$$
> FastClassBySpringCGLIB$$83a44b2a.invoke(<generated>)
>
> Regards,
> Manoj
>
>
> This message is confidential and subject to terms at:
> http://www.jpmorgan.com/emaildisclaimer including on confidentiality,
> legal privilege, viruses and monitoring of electronic messages. If you are
> not the intended recipient, please delete this message and notify the
> sender immediately. Any unauthorized use is strictly prohibited.
>
>
>
> This message is confidential and subject to terms at: http://
> www.jpmorgan.com/emaildisclaimer including on confidentiality, legal
> privilege, viruses and monitoring of electronic messages. If you are not
> the intended recipient, please delete this message and notify the sender
> immediately. Any unauthorized use is strictly prohibited.
>
>
>
> This message is confidential and subject to terms at: http://
> www.jpmorgan.com/emaildisclaimer including on confidentiality, legal
> privilege, viruses and monitoring of electronic messages. If you are not
> the intended recipient, please delete this message and notify the sender
> immediately. Any unauthorized use is strictly prohibited.
>
RE: Kafka Streaming data - Error while building the Cube
Posted by "Kumar, Manoj H" <ma...@jpmorgan.com>.
I guess KafkaSource Class file where it has enrichSourcePartitionBeforeBuild() to get partition values. There its getting Error out. Do we know how can we test out to find why its coming as 0 for start & end offset?
Regards,
Manoj
From: Kumar, Manoj H
Sent: Thursday, October 12, 2017 3:35 PM
To: 'user@kylin.apache.org'
Subject: RE: Kafka Streaming data - Error while building the Cube
Yes its there.. I could see the messages..
Regards,
Manoj
From: Billy Liu [mailto:billyliu@apache.org]
Sent: Thursday, October 12, 2017 3:11 PM
To: user
Subject: Re: Kafka Streaming data - Error while building the Cube
STREAMING_SALES_TABLE table reads messages from Kafka topic kylin_demo,but got 0 message.
Could you check if the topic has incoming message: bin/kafka-console-consumer.sh --zookeeper localhost:2181 --bootstrap-server localhost:9092 --topic kylin_demo
2017-10-12 17:19 GMT+08:00 Kumar, Manoj H <ma...@jpmorgan.com>>:
Pls. find below information about consumer config from Kylin log file.
2017-10-11 02:11:43,787 INFO [pool-8-thread-1] threadpool.DefaultScheduler:123 : Job Fetcher: 0 should running, 0 actual running, 0 stopped, 0 ready, 1 already succeed, 0 error, 0 discarded, 0 others
2017-10-11 02:12:13,783 INFO [pool-8-thread-1] threadpool.DefaultScheduler:123 : Job Fetcher: 0 should running, 0 actual running, 0 stopped, 0 ready, 1 already succeed, 0 error, 0 discarded, 0 others
2017-10-11 02:12:40,734 INFO [http-bio-7070-exec-3] streaming.StreamingManager:222 : Reloading Streaming Metadata from folder kylin_metadata(key='/streaming')@kylin_metadata@hbase
2017-10-11 02:12:40,760 DEBUG [http-bio-7070-exec-3] streaming.StreamingManager:247 : Loaded 1 StreamingConfig(s)
2017-10-11 02:12:43,789 INFO [pool-8-thread-1] threadpool.DefaultScheduler:123 : Job Fetcher: 0 should running, 0 actual running, 0 stopped, 0 ready, 1 already succeed, 0 error, 0 discarded, 0 others
2017-10-11 02:13:13,788 INFO [pool-8-thread-1] threadpool.DefaultScheduler:123 : Job Fetcher: 0 should running, 0 actual running, 0 stopped, 0 ready, 1 already succeed, 0 error, 0 discarded, 0 others
2017-10-11 02:13:43,785 INFO [pool-8-thread-1] threadpool.DefaultScheduler:123 : Job Fetcher: 0 should running, 0 actual running, 0 stopped, 0 ready, 1 already succeed, 0 error, 0 discarded, 0 others
2017-10-11 02:14:13,789 INFO [pool-8-thread-1] threadpool.DefaultScheduler:123 : Job Fetcher: 0 should running, 0 actual running, 0 stopped, 0 ready, 1 already succeed, 0 error, 0 discarded, 0 others
2017-10-11 02:14:43,796 INFO [pool-8-thread-1] threadpool.DefaultScheduler:123 : Job Fetcher: 0 should running, 0 actual running, 0 stopped, 0 ready, 1 already succeed, 0 error, 0 discarded, 0 others
2017-10-11 02:15:03,911 DEBUG [http-bio-7070-exec-1] controller.StreamingController:255 : Saving StreamingConfig {"uuid":"8613b0e1-40ac-438c-bdf5-72be4d91c230","last_modified":1507705685859,"version":"2.1.0","name":"DEFAULT.STREAMING_SALES_TABLE","type":"kafka"}
2017-10-11 02:15:03,913 DEBUG [http-bio-7070-exec-1] controller.StreamingController:273 : Saving KafkaConfig {"uuid":"87dc6ab5-5141-4bd8-8e00-c16ec86dce41","last_modified":1507705685916,"version":"2.1.0","name":"DEFAULT.STREAMING_SALES_TABLE","clusters":[{"brokers":[{"id":"1","host":"sandbox","port":"9092"}]}],"topic":"kylin_demo","timeout":60000,"parserName":"org.apache.kylin.source.kafka.TimedJsonStreamParser","parserTimeStampField":null,"margin":0,"parserProperties":"tsColName=order_time"}
2017-10-11 02:15:03,963 DEBUG [pool-7-thread-1] cachesync.Broadcaster:132 : Servers in the cluster: [localhost:7070]
2017-10-11 02:15:04,000 DEBUG [pool-7-thread-1] cachesync.Broadcaster:139 : Announcing new broadcast event: BroadcastEvent{entity=streaming, event=update, cacheKey=DEFAULT.STREAMING_SALES_TABLE}
2017-10-11 02:15:04,009 DEBUG [pool-7-thread-1] cachesync.Broadcaster:132 : Servers in the cluster: [localhost:7070]
2017-10-11 02:15:04,009 DEBUG [pool-7-thread-1] cachesync.Broadcaster:139 : Announcing new broadcast event: BroadcastEvent{entity=kafka, event=update, cacheKey=DEFAULT.STREAMING_SALES_TABLE}
2017-10-11 02:15:04,164 DEBUG [http-bio-7070-exec-9] cachesync.Broadcaster:236 : Done broadcasting metadata change: entity=streaming, event=UPDATE, cacheKey=DEFAULT.STREAMING_SALES_TABLE
2017-10-11 02:15:04,192 DEBUG [http-bio-7070-exec-10] cachesync.Broadcaster:236 : Done broadcasting metadata change: entity=kafka, event=UPDATE, cacheKey=DEFAULT.STREAMING_SALES_TABLE
2017-10-11 02:15:13,789 INFO [pool-8-thread-1] threadpool.DefaultScheduler:123 : Job Fetcher: 0 should running, 0 actual running, 0 stopped, 0 ready, 1 already succeed, 0 error, 0 discarded, 0 others
2017-10-11 02:15:23,780 DEBUG [http-bio-7070-exec-7] kafka.KafkaSource:83 : Last segment doesn't exist, and didn't initiate the start offset, will seek from topic's earliest offset.
2017-10-11 20:50:42,558 INFO [http-bio-7070-exec-8] utils.AppInfoParser:83 : Kafka version : 0.10.2-kafka-2.2.0
2017-10-11 20:50:42,563 INFO [http-bio-7070-exec-8] utils.AppInfoParser:84 : Kafka commitId : unknown
2017-10-11 20:50:42,570 DEBUG [http-bio-7070-exec-8] kafka.KafkaSource:105 : Seek end offsets from topic
2017-10-11 20:50:42,570 INFO [http-bio-7070-exec-8] consumer.ConsumerConfig:196 : ConsumerConfig values:
auto.commit.interval.ms<http://auto.commit.interval.ms> = 5000
auto.offset.reset = latest
bootstrap.servers = [localhost:9092]
check.crcs = true
client.id<http://client.id> =
connections.max.idle.ms<http://connections.max.idle.ms> = 540000
enable.auto.commit = false
exclude.internal.topics = true
fetch.max.bytes = 52428800
fetch.max.wait.ms<http://fetch.max.wait.ms> = 500
fetch.min.bytes = 1
group.id<http://group.id> = streaming_cube
heartbeat.interval.ms<http://heartbeat.interval.ms> = 3000
interceptor.classes = null
internal.leave.group.on.close = true
key.deserializer = class org.apache.kafka.common.serialization.StringDeserializer
max.partition.fetch.bytes = 1048576
max.poll.interval.ms<http://max.poll.interval.ms> = 300000
max.poll.records = 500
metadata.max.age.ms<http://metadata.max.age.ms> = 300000
metric.reporters = []
metrics.num.samples = 2
metrics.recording.level = INFO
metrics.sample.window.ms<http://metrics.sample.window.ms> = 30000
partition.assignment.strategy = [class org.apache.kafka.clients.consumer.RangeAssignor]
receive.buffer.bytes = 65536
reconnect.backoff.ms<http://reconnect.backoff.ms> = 50
request.timeout.ms<http://request.timeout.ms> = 305000
retry.backoff.ms<http://retry.backoff.ms> = 100
sasl.jaas.config = null
sasl.kerberos.kinit.cmd = /usr/bin/kinit
sasl.kerberos.min.time.before.relogin = 60000
sasl.kerberos.service.name<http://sasl.kerberos.service.name> = null
sasl.kerberos.ticket.renew.jitter = 0.05
request.timeout.ms<http://request.timeout.ms> = 305000
retry.backoff.ms<http://retry.backoff.ms> = 100
sasl.jaas.config = null
sasl.kerberos.kinit.cmd = /usr/bin/kinit
sasl.kerberos.min.time.before.relogin = 60000
sasl.kerberos.service.name<http://sasl.kerberos.service.name> = null
sasl.kerberos.ticket.renew.jitter = 0.05
sasl.kerberos.ticket.renew.window.factor = 0.8
sasl.mechanism = GSSAPI
security.protocol = PLAINTEXT
send.buffer.bytes = 131072
session.timeout.ms<http://session.timeout.ms> = 10000
ssl.cipher.suites = null
ssl.enabled.protocols = [TLSv1.2, TLSv1.1, TLSv1]
ssl.endpoint.identification.algorithm = null
ssl.key.password = null
ssl.keymanager.algorithm = SunX509
ssl.keystore.location = null
ssl.keystore.password = null
ssl.keystore.type = JKS
ssl.protocol = TLS
ssl.provider = null
ssl.secure.random.implementation = null
ssl.trustmanager.algorithm = PKIX
ssl.truststore.location = null
ssl.truststore.password = null
ssl.truststore.type = JKS
value.deserializer = class org.apache.kafka.common.serialization.StringDeserializer
2017-10-11 20:50:42,573 INFO [http-bio-7070-exec-8] utils.AppInfoParser:83 : Kafka version : 0.10.2-kafka-2.2.0
2017-10-11 20:50:42,573 INFO [http-bio-7070-exec-8] utils.AppInfoParser:84 : Kafka commitId : unknown
2017-10-11 20:50:42,586 DEBUG [http-bio-7070-exec-8] kafka.KafkaSource:107 : The end offsets are {0=0}
2017-10-11 20:50:42,588 ERROR [http-bio-7070-exec-8] controller.CubeController:305 : No new message comes, startOffset = endOffset:0
java.lang.IllegalArgumentException: No new message comes, startOffset = endOffset:0
at org.apache.kylin.source.kafka.KafkaSource.enrichSourcePartitionBeforeBuild(KafkaSource.java:134)
at org.apache.kylin.rest.service.JobService.submitJobInternal(JobService.java:236)
Regards,
Manoj
From: Billy Liu [mailto:billyliu@apache.org<ma...@apache.org>]
Sent: Thursday, October 12, 2017 1:06 PM
To: user
Subject: Re: Kafka Streaming data - Error while building the Cube
Hi Kumar,
Could you paste more Kafka Consumer related log in kylin.log? And also check from the Kafka broker side, if the Kylin client has connected to Broker.
2017-10-12 14:29 GMT+08:00 Kumar, Manoj H <ma...@jpmorgan.com>>:
Building the Cube from Kylin UI - Although Messages are there in Kafka topic but Kylin is not able read the offset. Can someone help on this?
2017-10-11 20:50:42,573 INFO [http-bio-7070-exec-8] utils.AppInfoParser:83 : Kafka version : 0.10.2-kafka-2.2.0
2017-10-11 20:50:42,573 INFO [http-bio-7070-exec-8] utils.AppInfoParser:84 : Kafka commitId : unknown
2017-10-11 20:50:42,586 DEBUG [http-bio-7070-exec-8] kafka.KafkaSource:107 : The end offsets are {0=0}
2017-10-11 20:50:42,588 ERROR [http-bio-7070-exec-8] controller.CubeController:305 : No new message comes, startOffset = endOffset:0
java.lang.IllegalArgumentException: No new message comes, startOffset = endOffset:0
at org.apache.kylin.source.kafka.KafkaSource.enrichSourcePartitionBeforeBuild(KafkaSource.java:134)
at org.apache.kylin.rest.service.JobService.submitJobInternal(JobService.java:236)
at org.apache.kylin.rest.service.JobService.submitJob(JobService.java:208)
at org.apache.kylin.rest.service.JobService$$FastClassBySpringCGLIB$$83a44b2a.invoke(<generated>)
Regards,
Manoj
This message is confidential and subject to terms at: http://www.jpmorgan.com/emaildisclaimer including on confidentiality, legal privilege, viruses and monitoring of electronic messages. If you are not the intended recipient, please delete this message and notify the sender immediately. Any unauthorized use is strictly prohibited.
This message is confidential and subject to terms at: http://www.jpmorgan.com/emaildisclaimer<http://www.jpmorgan.com/emaildisclaimer> including on confidentiality, legal privilege, viruses and monitoring of electronic messages. If you are not the intended recipient, please delete this message and notify the sender immediately. Any unauthorized use is strictly prohibited.
This message is confidential and subject to terms at: http://www.jpmorgan.com/emaildisclaimer including on confidentiality, legal privilege, viruses and monitoring of electronic messages. If you are not the intended recipient, please delete this message and notify the sender immediately. Any unauthorized use is strictly prohibited.
RE: Kafka Streaming data - Error while building the Cube
Posted by "Kumar, Manoj H" <ma...@jpmorgan.com>.
Yes its there.. I could see the messages..
Regards,
Manoj
From: Billy Liu [mailto:billyliu@apache.org]
Sent: Thursday, October 12, 2017 3:11 PM
To: user
Subject: Re: Kafka Streaming data - Error while building the Cube
STREAMING_SALES_TABLE table reads messages from Kafka topic kylin_demo,but got 0 message.
Could you check if the topic has incoming message: bin/kafka-console-consumer.sh --zookeeper localhost:2181 --bootstrap-server localhost:9092 --topic kylin_demo
2017-10-12 17:19 GMT+08:00 Kumar, Manoj H <ma...@jpmorgan.com>>:
Pls. find below information about consumer config from Kylin log file.
2017-10-11 02:11:43,787 INFO [pool-8-thread-1] threadpool.DefaultScheduler:123 : Job Fetcher: 0 should running, 0 actual running, 0 stopped, 0 ready, 1 already succeed, 0 error, 0 discarded, 0 others
2017-10-11 02:12:13,783 INFO [pool-8-thread-1] threadpool.DefaultScheduler:123 : Job Fetcher: 0 should running, 0 actual running, 0 stopped, 0 ready, 1 already succeed, 0 error, 0 discarded, 0 others
2017-10-11 02:12:40,734 INFO [http-bio-7070-exec-3] streaming.StreamingManager:222 : Reloading Streaming Metadata from folder kylin_metadata(key='/streaming')@kylin_metadata@hbase
2017-10-11 02:12:40,760 DEBUG [http-bio-7070-exec-3] streaming.StreamingManager:247 : Loaded 1 StreamingConfig(s)
2017-10-11 02:12:43,789 INFO [pool-8-thread-1] threadpool.DefaultScheduler:123 : Job Fetcher: 0 should running, 0 actual running, 0 stopped, 0 ready, 1 already succeed, 0 error, 0 discarded, 0 others
2017-10-11 02:13:13,788 INFO [pool-8-thread-1] threadpool.DefaultScheduler:123 : Job Fetcher: 0 should running, 0 actual running, 0 stopped, 0 ready, 1 already succeed, 0 error, 0 discarded, 0 others
2017-10-11 02:13:43,785 INFO [pool-8-thread-1] threadpool.DefaultScheduler:123 : Job Fetcher: 0 should running, 0 actual running, 0 stopped, 0 ready, 1 already succeed, 0 error, 0 discarded, 0 others
2017-10-11 02:14:13,789 INFO [pool-8-thread-1] threadpool.DefaultScheduler:123 : Job Fetcher: 0 should running, 0 actual running, 0 stopped, 0 ready, 1 already succeed, 0 error, 0 discarded, 0 others
2017-10-11 02:14:43,796 INFO [pool-8-thread-1] threadpool.DefaultScheduler:123 : Job Fetcher: 0 should running, 0 actual running, 0 stopped, 0 ready, 1 already succeed, 0 error, 0 discarded, 0 others
2017-10-11 02:15:03,911 DEBUG [http-bio-7070-exec-1] controller.StreamingController:255 : Saving StreamingConfig {"uuid":"8613b0e1-40ac-438c-bdf5-72be4d91c230","last_modified":1507705685859,"version":"2.1.0","name":"DEFAULT.STREAMING_SALES_TABLE","type":"kafka"}
2017-10-11 02:15:03,913 DEBUG [http-bio-7070-exec-1] controller.StreamingController:273 : Saving KafkaConfig {"uuid":"87dc6ab5-5141-4bd8-8e00-c16ec86dce41","last_modified":1507705685916,"version":"2.1.0","name":"DEFAULT.STREAMING_SALES_TABLE","clusters":[{"brokers":[{"id":"1","host":"sandbox","port":"9092"}]}],"topic":"kylin_demo","timeout":60000,"parserName":"org.apache.kylin.source.kafka.TimedJsonStreamParser","parserTimeStampField":null,"margin":0,"parserProperties":"tsColName=order_time"}
2017-10-11 02:15:03,963 DEBUG [pool-7-thread-1] cachesync.Broadcaster:132 : Servers in the cluster: [localhost:7070]
2017-10-11 02:15:04,000 DEBUG [pool-7-thread-1] cachesync.Broadcaster:139 : Announcing new broadcast event: BroadcastEvent{entity=streaming, event=update, cacheKey=DEFAULT.STREAMING_SALES_TABLE}
2017-10-11 02:15:04,009 DEBUG [pool-7-thread-1] cachesync.Broadcaster:132 : Servers in the cluster: [localhost:7070]
2017-10-11 02:15:04,009 DEBUG [pool-7-thread-1] cachesync.Broadcaster:139 : Announcing new broadcast event: BroadcastEvent{entity=kafka, event=update, cacheKey=DEFAULT.STREAMING_SALES_TABLE}
2017-10-11 02:15:04,164 DEBUG [http-bio-7070-exec-9] cachesync.Broadcaster:236 : Done broadcasting metadata change: entity=streaming, event=UPDATE, cacheKey=DEFAULT.STREAMING_SALES_TABLE
2017-10-11 02:15:04,192 DEBUG [http-bio-7070-exec-10] cachesync.Broadcaster:236 : Done broadcasting metadata change: entity=kafka, event=UPDATE, cacheKey=DEFAULT.STREAMING_SALES_TABLE
2017-10-11 02:15:13,789 INFO [pool-8-thread-1] threadpool.DefaultScheduler:123 : Job Fetcher: 0 should running, 0 actual running, 0 stopped, 0 ready, 1 already succeed, 0 error, 0 discarded, 0 others
2017-10-11 02:15:23,780 DEBUG [http-bio-7070-exec-7] kafka.KafkaSource:83 : Last segment doesn't exist, and didn't initiate the start offset, will seek from topic's earliest offset.
2017-10-11 20:50:42,558 INFO [http-bio-7070-exec-8] utils.AppInfoParser:83 : Kafka version : 0.10.2-kafka-2.2.0
2017-10-11 20:50:42,563 INFO [http-bio-7070-exec-8] utils.AppInfoParser:84 : Kafka commitId : unknown
2017-10-11 20:50:42,570 DEBUG [http-bio-7070-exec-8] kafka.KafkaSource:105 : Seek end offsets from topic
2017-10-11 20:50:42,570 INFO [http-bio-7070-exec-8] consumer.ConsumerConfig:196 : ConsumerConfig values:
auto.commit.interval.ms<http://auto.commit.interval.ms> = 5000
auto.offset.reset = latest
bootstrap.servers = [localhost:9092]
check.crcs = true
client.id<http://client.id> =
connections.max.idle.ms<http://connections.max.idle.ms> = 540000
enable.auto.commit = false
exclude.internal.topics = true
fetch.max.bytes = 52428800
fetch.max.wait.ms<http://fetch.max.wait.ms> = 500
fetch.min.bytes = 1
group.id<http://group.id> = streaming_cube
heartbeat.interval.ms<http://heartbeat.interval.ms> = 3000
interceptor.classes = null
internal.leave.group.on.close = true
key.deserializer = class org.apache.kafka.common.serialization.StringDeserializer
max.partition.fetch.bytes = 1048576
max.poll.interval.ms<http://max.poll.interval.ms> = 300000
max.poll.records = 500
metadata.max.age.ms<http://metadata.max.age.ms> = 300000
metric.reporters = []
metrics.num.samples = 2
metrics.recording.level = INFO
metrics.sample.window.ms<http://metrics.sample.window.ms> = 30000
partition.assignment.strategy = [class org.apache.kafka.clients.consumer.RangeAssignor]
receive.buffer.bytes = 65536
reconnect.backoff.ms<http://reconnect.backoff.ms> = 50
request.timeout.ms<http://request.timeout.ms> = 305000
retry.backoff.ms<http://retry.backoff.ms> = 100
sasl.jaas.config = null
sasl.kerberos.kinit.cmd = /usr/bin/kinit
sasl.kerberos.min.time.before.relogin = 60000
sasl.kerberos.service.name<http://sasl.kerberos.service.name> = null
sasl.kerberos.ticket.renew.jitter = 0.05
request.timeout.ms<http://request.timeout.ms> = 305000
retry.backoff.ms<http://retry.backoff.ms> = 100
sasl.jaas.config = null
sasl.kerberos.kinit.cmd = /usr/bin/kinit
sasl.kerberos.min.time.before.relogin = 60000
sasl.kerberos.service.name<http://sasl.kerberos.service.name> = null
sasl.kerberos.ticket.renew.jitter = 0.05
sasl.kerberos.ticket.renew.window.factor = 0.8
sasl.mechanism = GSSAPI
security.protocol = PLAINTEXT
send.buffer.bytes = 131072
session.timeout.ms<http://session.timeout.ms> = 10000
ssl.cipher.suites = null
ssl.enabled.protocols = [TLSv1.2, TLSv1.1, TLSv1]
ssl.endpoint.identification.algorithm = null
ssl.key.password = null
ssl.keymanager.algorithm = SunX509
ssl.keystore.location = null
ssl.keystore.password = null
ssl.keystore.type = JKS
ssl.protocol = TLS
ssl.provider = null
ssl.secure.random.implementation = null
ssl.trustmanager.algorithm = PKIX
ssl.truststore.location = null
ssl.truststore.password = null
ssl.truststore.type = JKS
value.deserializer = class org.apache.kafka.common.serialization.StringDeserializer
2017-10-11 20:50:42,573 INFO [http-bio-7070-exec-8] utils.AppInfoParser:83 : Kafka version : 0.10.2-kafka-2.2.0
2017-10-11 20:50:42,573 INFO [http-bio-7070-exec-8] utils.AppInfoParser:84 : Kafka commitId : unknown
2017-10-11 20:50:42,586 DEBUG [http-bio-7070-exec-8] kafka.KafkaSource:107 : The end offsets are {0=0}
2017-10-11 20:50:42,588 ERROR [http-bio-7070-exec-8] controller.CubeController:305 : No new message comes, startOffset = endOffset:0
java.lang.IllegalArgumentException: No new message comes, startOffset = endOffset:0
at org.apache.kylin.source.kafka.KafkaSource.enrichSourcePartitionBeforeBuild(KafkaSource.java:134)
at org.apache.kylin.rest.service.JobService.submitJobInternal(JobService.java:236)
Regards,
Manoj
From: Billy Liu [mailto:billyliu@apache.org<ma...@apache.org>]
Sent: Thursday, October 12, 2017 1:06 PM
To: user
Subject: Re: Kafka Streaming data - Error while building the Cube
Hi Kumar,
Could you paste more Kafka Consumer related log in kylin.log? And also check from the Kafka broker side, if the Kylin client has connected to Broker.
2017-10-12 14:29 GMT+08:00 Kumar, Manoj H <ma...@jpmorgan.com>>:
Building the Cube from Kylin UI - Although Messages are there in Kafka topic but Kylin is not able read the offset. Can someone help on this?
2017-10-11 20:50:42,573 INFO [http-bio-7070-exec-8] utils.AppInfoParser:83 : Kafka version : 0.10.2-kafka-2.2.0
2017-10-11 20:50:42,573 INFO [http-bio-7070-exec-8] utils.AppInfoParser:84 : Kafka commitId : unknown
2017-10-11 20:50:42,586 DEBUG [http-bio-7070-exec-8] kafka.KafkaSource:107 : The end offsets are {0=0}
2017-10-11 20:50:42,588 ERROR [http-bio-7070-exec-8] controller.CubeController:305 : No new message comes, startOffset = endOffset:0
java.lang.IllegalArgumentException: No new message comes, startOffset = endOffset:0
at org.apache.kylin.source.kafka.KafkaSource.enrichSourcePartitionBeforeBuild(KafkaSource.java:134)
at org.apache.kylin.rest.service.JobService.submitJobInternal(JobService.java:236)
at org.apache.kylin.rest.service.JobService.submitJob(JobService.java:208)
at org.apache.kylin.rest.service.JobService$$FastClassBySpringCGLIB$$83a44b2a.invoke(<generated>)
Regards,
Manoj
This message is confidential and subject to terms at: http://www.jpmorgan.com/emaildisclaimer including on confidentiality, legal privilege, viruses and monitoring of electronic messages. If you are not the intended recipient, please delete this message and notify the sender immediately. Any unauthorized use is strictly prohibited.
This message is confidential and subject to terms at: http://www.jpmorgan.com/emaildisclaimer<http://www.jpmorgan.com/emaildisclaimer> including on confidentiality, legal privilege, viruses and monitoring of electronic messages. If you are not the intended recipient, please delete this message and notify the sender immediately. Any unauthorized use is strictly prohibited.
This message is confidential and subject to terms at: http://www.jpmorgan.com/emaildisclaimer including on confidentiality, legal privilege, viruses and monitoring of electronic messages. If you are not the intended recipient, please delete this message and notify the sender immediately. Any unauthorized use is strictly prohibited.
Re: Kafka Streaming data - Error while building the Cube
Posted by Billy Liu <bi...@apache.org>.
STREAMING_SALES_TABLE table reads messages from Kafka topic kylin_demo,but
got 0 message.
Could you check if the topic has incoming message:
bin/kafka-console-consumer.sh
--zookeeper localhost:2181 --bootstrap-server localhost:9092 --topic
kylin_demo
2017-10-12 17:19 GMT+08:00 Kumar, Manoj H <ma...@jpmorgan.com>:
> Pls. find below information about consumer config from Kylin log file.
>
>
>
> 2017-10-11 02:11:43,787 INFO [pool-8-thread-1]
> threadpool.DefaultScheduler:123 : Job Fetcher: 0 should running, 0 actual
> running, 0 stopped, 0 ready, 1 already succeed, 0 error, 0 discarded, 0
> others
>
> 2017-10-11 02:12:13,783 INFO [pool-8-thread-1]
> threadpool.DefaultScheduler:123 : Job Fetcher: 0 should running, 0 actual
> running, 0 stopped, 0 ready, 1 already succeed, 0 error, 0 discarded, 0
> others
>
> 2017-10-11 02:12:40,734 INFO [http-bio-7070-exec-3]
> streaming.StreamingManager:222 : Reloading Streaming Metadata from folder
> kylin_metadata(key='/streaming')@kylin_metadata@hbase
>
> 2017-10-11 02:12:40,760 DEBUG [http-bio-7070-exec-3]
> streaming.StreamingManager:247 : Loaded 1 StreamingConfig(s)
>
> 2017-10-11 02:12:43,789 INFO [pool-8-thread-1]
> threadpool.DefaultScheduler:123 : Job Fetcher: 0 should running, 0 actual
> running, 0 stopped, 0 ready, 1 already succeed, 0 error, 0 discarded, 0
> others
>
> 2017-10-11 02:13:13,788 INFO [pool-8-thread-1]
> threadpool.DefaultScheduler:123 : Job Fetcher: 0 should running, 0 actual
> running, 0 stopped, 0 ready, 1 already succeed, 0 error, 0 discarded, 0
> others
>
> 2017-10-11 02:13:43,785 INFO [pool-8-thread-1]
> threadpool.DefaultScheduler:123 : Job Fetcher: 0 should running, 0 actual
> running, 0 stopped, 0 ready, 1 already succeed, 0 error, 0 discarded, 0
> others
>
> 2017-10-11 02:14:13,789 INFO [pool-8-thread-1]
> threadpool.DefaultScheduler:123 : Job Fetcher: 0 should running, 0 actual
> running, 0 stopped, 0 ready, 1 already succeed, 0 error, 0 discarded, 0
> others
>
> 2017-10-11 02:14:43,796 INFO [pool-8-thread-1]
> threadpool.DefaultScheduler:123 : Job Fetcher: 0 should running, 0 actual
> running, 0 stopped, 0 ready, 1 already succeed, 0 error, 0 discarded, 0
> others
>
> 2017-10-11 02:15:03,911 DEBUG [http-bio-7070-exec-1] controller.StreamingController:255
> : Saving StreamingConfig {"uuid":"8613b0e1-40ac-438c-
> bdf5-72be4d91c230","last_modified":1507705685859,"
> version":"2.1.0","name":"DEFAULT.STREAMING_SALES_TABLE","type":"kafka"}
>
> 2017-10-11 02:15:03,913 DEBUG [http-bio-7070-exec-1] controller.StreamingController:273
> : Saving KafkaConfig {"uuid":"87dc6ab5-5141-4bd8-8e00-c16ec86dce41","last_
> modified":1507705685916,"version":"2.1.0","name":"
> DEFAULT.STREAMING_SALES_TABLE","clusters":[{"brokers":[{"id"
> :"1","host":"sandbox","port":"9092"}]}],"topic":"kylin_demo"
> ,"timeout":60000,"parserName":"org.apache.kylin.source.
> kafka.TimedJsonStreamParser","parserTimeStampField":null,"
> margin":0,"parserProperties":"tsColName=order_time"}
>
> 2017-10-11 02:15:03,963 DEBUG [pool-7-thread-1] cachesync.Broadcaster:132
> : Servers in the cluster: [localhost:7070]
>
> 2017-10-11 02:15:04,000 DEBUG [pool-7-thread-1] cachesync.Broadcaster:139
> : Announcing new broadcast event: BroadcastEvent{entity=streaming,
> event=update, cacheKey=DEFAULT.STREAMING_SALES_TABLE}
>
> 2017-10-11 02:15:04,009 DEBUG [pool-7-thread-1] cachesync.Broadcaster:132
> : Servers in the cluster: [localhost:7070]
>
> 2017-10-11 02:15:04,009 DEBUG [pool-7-thread-1] cachesync.Broadcaster:139
> : Announcing new broadcast event: BroadcastEvent{entity=kafka,
> event=update, cacheKey=DEFAULT.STREAMING_SALES_TABLE}
>
> 2017-10-11 02:15:04,164 DEBUG [http-bio-7070-exec-9]
> cachesync.Broadcaster:236 : Done broadcasting metadata change:
> entity=streaming, event=UPDATE, cacheKey=DEFAULT.STREAMING_SALES_TABLE
>
> 2017-10-11 02:15:04,192 DEBUG [http-bio-7070-exec-10]
> cachesync.Broadcaster:236 : Done broadcasting metadata change:
> entity=kafka, event=UPDATE, cacheKey=DEFAULT.STREAMING_SALES_TABLE
>
> 2017-10-11 02:15:13,789 INFO [pool-8-thread-1]
> threadpool.DefaultScheduler:123 : Job Fetcher: 0 should running, 0 actual
> running, 0 stopped, 0 ready, 1 already succeed, 0 error, 0 discarded, 0
> others
>
> 2017-10-11 02:15:23,780 DEBUG [http-bio-7070-exec-7] kafka.KafkaSource:83
> : Last segment doesn't exist, and didn't initiate the start offset, will
> seek from topic's earliest offset.
>
>
>
> 2017-10-11 20:50:42,558 INFO [http-bio-7070-exec-8]
> utils.AppInfoParser:83 : Kafka version : 0.10.2-kafka-2.2.0
>
> 2017-10-11 20:50:42,563 INFO [http-bio-7070-exec-8]
> utils.AppInfoParser:84 : Kafka commitId : unknown
>
> 2017-10-11 20:50:42,570 DEBUG [http-bio-7070-exec-8] kafka.KafkaSource:105
> : Seek end offsets from topic
>
> 2017-10-11 20:50:42,570 INFO [http-bio-7070-exec-8]
> consumer.ConsumerConfig:196 : ConsumerConfig values:
>
> auto.commit.interval.ms = 5000
>
> auto.offset.reset = latest
>
> bootstrap.servers = [localhost:9092]
>
> check.crcs = true
>
> client.id =
>
> connections.max.idle.ms = 540000
>
> enable.auto.commit = false
>
> exclude.internal.topics = true
>
> fetch.max.bytes = 52428800
>
> fetch.max.wait.ms = 500
>
> fetch.min.bytes = 1
>
> group.id = streaming_cube
>
> heartbeat.interval.ms = 3000
>
> interceptor.classes = null
>
> internal.leave.group.on.close = true
>
> key.deserializer = class org.apache.kafka.common.serialization.
> StringDeserializer
>
> max.partition.fetch.bytes = 1048576
>
> max.poll.interval.ms = 300000
>
> max.poll.records = 500
>
> metadata.max.age.ms = 300000
>
> metric.reporters = []
>
> metrics.num.samples = 2
>
> metrics.recording.level = INFO
>
> metrics.sample.window.ms = 30000
>
> partition.assignment.strategy = [class org.apache.kafka.clients.
> consumer.RangeAssignor]
>
> receive.buffer.bytes = 65536
>
> reconnect.backoff.ms = 50
>
> request.timeout.ms = 305000
>
> retry.backoff.ms = 100
>
> sasl.jaas.config = null
>
> sasl.kerberos.kinit.cmd = /usr/bin/kinit
>
> sasl.kerberos.min.time.before.relogin = 60000
>
> sasl.kerberos.service.name = null
>
> sasl.kerberos.ticket.renew.jitter = 0.05
>
> request.timeout.ms = 305000
>
> retry.backoff.ms = 100
>
> sasl.jaas.config = null
>
> sasl.kerberos.kinit.cmd = /usr/bin/kinit
>
> sasl.kerberos.min.time.before.relogin = 60000
>
> sasl.kerberos.service.name = null
>
> sasl.kerberos.ticket.renew.jitter = 0.05
>
> sasl.kerberos.ticket.renew.window.factor = 0.8
>
> sasl.mechanism = GSSAPI
>
> security.protocol = PLAINTEXT
>
> send.buffer.bytes = 131072
>
> session.timeout.ms = 10000
>
> ssl.cipher.suites = null
>
> ssl.enabled.protocols = [TLSv1.2, TLSv1.1, TLSv1]
>
> ssl.endpoint.identification.algorithm = null
>
> ssl.key.password = null
>
> ssl.keymanager.algorithm = SunX509
>
> ssl.keystore.location = null
>
> ssl.keystore.password = null
>
> ssl.keystore.type = JKS
>
> ssl.protocol = TLS
>
> ssl.provider = null
>
> ssl.secure.random.implementation = null
>
> ssl.trustmanager.algorithm = PKIX
>
> ssl.truststore.location = null
>
> ssl.truststore.password = null
>
> ssl.truststore.type = JKS
>
> value.deserializer = class org.apache.kafka.common.serialization.
> StringDeserializer
>
>
>
> 2017-10-11 20:50:42,573 INFO [http-bio-7070-exec-8]
> utils.AppInfoParser:83 : Kafka version : 0.10.2-kafka-2.2.0
>
> 2017-10-11 20:50:42,573 INFO [http-bio-7070-exec-8]
> utils.AppInfoParser:84 : Kafka commitId : unknown
>
> 2017-10-11 20:50:42,586 DEBUG [http-bio-7070-exec-8] kafka.KafkaSource:107
> : The end offsets are {0=0}
>
> 2017-10-11 20:50:42,588 ERROR [http-bio-7070-exec-8]
> controller.CubeController:305 : No new message comes, startOffset =
> endOffset:0
>
> java.lang.IllegalArgumentException: No new message comes, startOffset =
> endOffset:0
>
> at org.apache.kylin.source.kafka.KafkaSource.
> enrichSourcePartitionBeforeBuild(KafkaSource.java:134)
>
> at org.apache.kylin.rest.service.JobService.submitJobInternal(
> JobService.java:236)
>
> Regards,
>
> Manoj
>
>
>
> *From:* Billy Liu [mailto:billyliu@apache.org]
> *Sent:* Thursday, October 12, 2017 1:06 PM
> *To:* user
> *Subject:* Re: Kafka Streaming data - Error while building the Cube
>
>
>
> Hi Kumar,
>
>
>
> Could you paste more Kafka Consumer related log in kylin.log? And also
> check from the Kafka broker side, if the Kylin client has connected to
> Broker.
>
>
>
> 2017-10-12 14:29 GMT+08:00 Kumar, Manoj H <ma...@jpmorgan.com>:
>
> Building the Cube from Kylin UI - Although Messages are there in Kafka
> topic but Kylin is not able read the offset. Can someone help on this?
>
> 2017-10-11 20:50:42,573 INFO [http-bio-7070-exec-8]
> utils.AppInfoParser:83 : Kafka version : 0.10.2-kafka-2.2.0
> 2017-10-11 20:50:42,573 INFO [http-bio-7070-exec-8]
> utils.AppInfoParser:84 : Kafka commitId : unknown
> 2017-10-11 20:50:42,586 DEBUG [http-bio-7070-exec-8] kafka.KafkaSource:107
> : The end offsets are {0=0}
> 2017-10-11 20:50:42,588 ERROR [http-bio-7070-exec-8]
> controller.CubeController:305 : No new message comes, startOffset =
> endOffset:0
> java.lang.IllegalArgumentException: No new message comes, startOffset =
> endOffset:0
> at org.apache.kylin.source.kafka.KafkaSource.
> enrichSourcePartitionBeforeBuild(KafkaSource.java:134)
> at org.apache.kylin.rest.service.JobService.submitJobInternal(
> JobService.java:236)
> at org.apache.kylin.rest.service.JobService.submitJob(
> JobService.java:208)
> at org.apache.kylin.rest.service.JobService$$
> FastClassBySpringCGLIB$$83a44b2a.invoke(<generated>)
>
> Regards,
> Manoj
>
>
> This message is confidential and subject to terms at:
> http://www.jpmorgan.com/emaildisclaimer including on confidentiality,
> legal privilege, viruses and monitoring of electronic messages. If you are
> not the intended recipient, please delete this message and notify the
> sender immediately. Any unauthorized use is strictly prohibited.
>
>
>
> This message is confidential and subject to terms at: http://
> www.jpmorgan.com/emaildisclaimer including on confidentiality, legal
> privilege, viruses and monitoring of electronic messages. If you are not
> the intended recipient, please delete this message and notify the sender
> immediately. Any unauthorized use is strictly prohibited.
>
RE: Kafka Streaming data - Error while building the Cube
Posted by "Kumar, Manoj H" <ma...@jpmorgan.com>.
Pls. find below information about consumer config from Kylin log file.
2017-10-11 02:11:43,787 INFO [pool-8-thread-1] threadpool.DefaultScheduler:123 : Job Fetcher: 0 should running, 0 actual running, 0 stopped, 0 ready, 1 already succeed, 0 error, 0 discarded, 0 others
2017-10-11 02:12:13,783 INFO [pool-8-thread-1] threadpool.DefaultScheduler:123 : Job Fetcher: 0 should running, 0 actual running, 0 stopped, 0 ready, 1 already succeed, 0 error, 0 discarded, 0 others
2017-10-11 02:12:40,734 INFO [http-bio-7070-exec-3] streaming.StreamingManager:222 : Reloading Streaming Metadata from folder kylin_metadata(key='/streaming')@kylin_metadata@hbase
2017-10-11 02:12:40,760 DEBUG [http-bio-7070-exec-3] streaming.StreamingManager:247 : Loaded 1 StreamingConfig(s)
2017-10-11 02:12:43,789 INFO [pool-8-thread-1] threadpool.DefaultScheduler:123 : Job Fetcher: 0 should running, 0 actual running, 0 stopped, 0 ready, 1 already succeed, 0 error, 0 discarded, 0 others
2017-10-11 02:13:13,788 INFO [pool-8-thread-1] threadpool.DefaultScheduler:123 : Job Fetcher: 0 should running, 0 actual running, 0 stopped, 0 ready, 1 already succeed, 0 error, 0 discarded, 0 others
2017-10-11 02:13:43,785 INFO [pool-8-thread-1] threadpool.DefaultScheduler:123 : Job Fetcher: 0 should running, 0 actual running, 0 stopped, 0 ready, 1 already succeed, 0 error, 0 discarded, 0 others
2017-10-11 02:14:13,789 INFO [pool-8-thread-1] threadpool.DefaultScheduler:123 : Job Fetcher: 0 should running, 0 actual running, 0 stopped, 0 ready, 1 already succeed, 0 error, 0 discarded, 0 others
2017-10-11 02:14:43,796 INFO [pool-8-thread-1] threadpool.DefaultScheduler:123 : Job Fetcher: 0 should running, 0 actual running, 0 stopped, 0 ready, 1 already succeed, 0 error, 0 discarded, 0 others
2017-10-11 02:15:03,911 DEBUG [http-bio-7070-exec-1] controller.StreamingController:255 : Saving StreamingConfig {"uuid":"8613b0e1-40ac-438c-bdf5-72be4d91c230","last_modified":1507705685859,"version":"2.1.0","name":"DEFAULT.STREAMING_SALES_TABLE","type":"kafka"}
2017-10-11 02:15:03,913 DEBUG [http-bio-7070-exec-1] controller.StreamingController:273 : Saving KafkaConfig {"uuid":"87dc6ab5-5141-4bd8-8e00-c16ec86dce41","last_modified":1507705685916,"version":"2.1.0","name":"DEFAULT.STREAMING_SALES_TABLE","clusters":[{"brokers":[{"id":"1","host":"sandbox","port":"9092"}]}],"topic":"kylin_demo","timeout":60000,"parserName":"org.apache.kylin.source.kafka.TimedJsonStreamParser","parserTimeStampField":null,"margin":0,"parserProperties":"tsColName=order_time"}
2017-10-11 02:15:03,963 DEBUG [pool-7-thread-1] cachesync.Broadcaster:132 : Servers in the cluster: [localhost:7070]
2017-10-11 02:15:04,000 DEBUG [pool-7-thread-1] cachesync.Broadcaster:139 : Announcing new broadcast event: BroadcastEvent{entity=streaming, event=update, cacheKey=DEFAULT.STREAMING_SALES_TABLE}
2017-10-11 02:15:04,009 DEBUG [pool-7-thread-1] cachesync.Broadcaster:132 : Servers in the cluster: [localhost:7070]
2017-10-11 02:15:04,009 DEBUG [pool-7-thread-1] cachesync.Broadcaster:139 : Announcing new broadcast event: BroadcastEvent{entity=kafka, event=update, cacheKey=DEFAULT.STREAMING_SALES_TABLE}
2017-10-11 02:15:04,164 DEBUG [http-bio-7070-exec-9] cachesync.Broadcaster:236 : Done broadcasting metadata change: entity=streaming, event=UPDATE, cacheKey=DEFAULT.STREAMING_SALES_TABLE
2017-10-11 02:15:04,192 DEBUG [http-bio-7070-exec-10] cachesync.Broadcaster:236 : Done broadcasting metadata change: entity=kafka, event=UPDATE, cacheKey=DEFAULT.STREAMING_SALES_TABLE
2017-10-11 02:15:13,789 INFO [pool-8-thread-1] threadpool.DefaultScheduler:123 : Job Fetcher: 0 should running, 0 actual running, 0 stopped, 0 ready, 1 already succeed, 0 error, 0 discarded, 0 others
2017-10-11 02:15:23,780 DEBUG [http-bio-7070-exec-7] kafka.KafkaSource:83 : Last segment doesn't exist, and didn't initiate the start offset, will seek from topic's earliest offset.
2017-10-11 20:50:42,558 INFO [http-bio-7070-exec-8] utils.AppInfoParser:83 : Kafka version : 0.10.2-kafka-2.2.0
2017-10-11 20:50:42,563 INFO [http-bio-7070-exec-8] utils.AppInfoParser:84 : Kafka commitId : unknown
2017-10-11 20:50:42,570 DEBUG [http-bio-7070-exec-8] kafka.KafkaSource:105 : Seek end offsets from topic
2017-10-11 20:50:42,570 INFO [http-bio-7070-exec-8] consumer.ConsumerConfig:196 : ConsumerConfig values:
auto.commit.interval.ms = 5000
auto.offset.reset = latest
bootstrap.servers = [localhost:9092]
check.crcs = true
client.id =
connections.max.idle.ms = 540000
enable.auto.commit = false
exclude.internal.topics = true
fetch.max.bytes = 52428800
fetch.max.wait.ms = 500
fetch.min.bytes = 1
group.id = streaming_cube
heartbeat.interval.ms = 3000
interceptor.classes = null
internal.leave.group.on.close = true
key.deserializer = class org.apache.kafka.common.serialization.StringDeserializer
max.partition.fetch.bytes = 1048576
max.poll.interval.ms = 300000
max.poll.records = 500
metadata.max.age.ms = 300000
metric.reporters = []
metrics.num.samples = 2
metrics.recording.level = INFO
metrics.sample.window.ms = 30000
partition.assignment.strategy = [class org.apache.kafka.clients.consumer.RangeAssignor]
receive.buffer.bytes = 65536
reconnect.backoff.ms = 50
request.timeout.ms = 305000
retry.backoff.ms = 100
sasl.jaas.config = null
sasl.kerberos.kinit.cmd = /usr/bin/kinit
sasl.kerberos.min.time.before.relogin = 60000
sasl.kerberos.service.name = null
sasl.kerberos.ticket.renew.jitter = 0.05
request.timeout.ms = 305000
retry.backoff.ms = 100
sasl.jaas.config = null
sasl.kerberos.kinit.cmd = /usr/bin/kinit
sasl.kerberos.min.time.before.relogin = 60000
sasl.kerberos.service.name = null
sasl.kerberos.ticket.renew.jitter = 0.05
sasl.kerberos.ticket.renew.window.factor = 0.8
sasl.mechanism = GSSAPI
security.protocol = PLAINTEXT
send.buffer.bytes = 131072
session.timeout.ms = 10000
ssl.cipher.suites = null
ssl.enabled.protocols = [TLSv1.2, TLSv1.1, TLSv1]
ssl.endpoint.identification.algorithm = null
ssl.key.password = null
ssl.keymanager.algorithm = SunX509
ssl.keystore.location = null
ssl.keystore.password = null
ssl.keystore.type = JKS
ssl.protocol = TLS
ssl.provider = null
ssl.secure.random.implementation = null
ssl.trustmanager.algorithm = PKIX
ssl.truststore.location = null
ssl.truststore.password = null
ssl.truststore.type = JKS
value.deserializer = class org.apache.kafka.common.serialization.StringDeserializer
2017-10-11 20:50:42,573 INFO [http-bio-7070-exec-8] utils.AppInfoParser:83 : Kafka version : 0.10.2-kafka-2.2.0
2017-10-11 20:50:42,573 INFO [http-bio-7070-exec-8] utils.AppInfoParser:84 : Kafka commitId : unknown
2017-10-11 20:50:42,586 DEBUG [http-bio-7070-exec-8] kafka.KafkaSource:107 : The end offsets are {0=0}
2017-10-11 20:50:42,588 ERROR [http-bio-7070-exec-8] controller.CubeController:305 : No new message comes, startOffset = endOffset:0
java.lang.IllegalArgumentException: No new message comes, startOffset = endOffset:0
at org.apache.kylin.source.kafka.KafkaSource.enrichSourcePartitionBeforeBuild(KafkaSource.java:134)
at org.apache.kylin.rest.service.JobService.submitJobInternal(JobService.java:236)
Regards,
Manoj
From: Billy Liu [mailto:billyliu@apache.org]
Sent: Thursday, October 12, 2017 1:06 PM
To: user
Subject: Re: Kafka Streaming data - Error while building the Cube
Hi Kumar,
Could you paste more Kafka Consumer related log in kylin.log? And also check from the Kafka broker side, if the Kylin client has connected to Broker.
2017-10-12 14:29 GMT+08:00 Kumar, Manoj H <ma...@jpmorgan.com>>:
Building the Cube from Kylin UI - Although Messages are there in Kafka topic but Kylin is not able read the offset. Can someone help on this?
2017-10-11 20:50:42,573 INFO [http-bio-7070-exec-8] utils.AppInfoParser:83 : Kafka version : 0.10.2-kafka-2.2.0
2017-10-11 20:50:42,573 INFO [http-bio-7070-exec-8] utils.AppInfoParser:84 : Kafka commitId : unknown
2017-10-11 20:50:42,586 DEBUG [http-bio-7070-exec-8] kafka.KafkaSource:107 : The end offsets are {0=0}
2017-10-11 20:50:42,588 ERROR [http-bio-7070-exec-8] controller.CubeController:305 : No new message comes, startOffset = endOffset:0
java.lang.IllegalArgumentException: No new message comes, startOffset = endOffset:0
at org.apache.kylin.source.kafka.KafkaSource.enrichSourcePartitionBeforeBuild(KafkaSource.java:134)
at org.apache.kylin.rest.service.JobService.submitJobInternal(JobService.java:236)
at org.apache.kylin.rest.service.JobService.submitJob(JobService.java:208)
at org.apache.kylin.rest.service.JobService$$FastClassBySpringCGLIB$$83a44b2a.invoke(<generated>)
Regards,
Manoj
This message is confidential and subject to terms at: http://www.jpmorgan.com/emaildisclaimer including on confidentiality, legal privilege, viruses and monitoring of electronic messages. If you are not the intended recipient, please delete this message and notify the sender immediately. Any unauthorized use is strictly prohibited.
This message is confidential and subject to terms at: http://www.jpmorgan.com/emaildisclaimer including on confidentiality, legal privilege, viruses and monitoring of electronic messages. If you are not the intended recipient, please delete this message and notify the sender immediately. Any unauthorized use is strictly prohibited.
Re: Kafka Streaming data - Error while building the Cube
Posted by Billy Liu <bi...@apache.org>.
Hi Kumar,
Could you paste more Kafka Consumer related log in kylin.log? And also
check from the Kafka broker side, if the Kylin client has connected to
Broker.
2017-10-12 14:29 GMT+08:00 Kumar, Manoj H <ma...@jpmorgan.com>:
> Building the Cube from Kylin UI - Although Messages are there in Kafka
> topic but Kylin is not able read the offset. Can someone help on this?
>
> 2017-10-11 20:50:42,573 INFO [http-bio-7070-exec-8]
> utils.AppInfoParser:83 : Kafka version : 0.10.2-kafka-2.2.0
> 2017-10-11 20:50:42,573 INFO [http-bio-7070-exec-8]
> utils.AppInfoParser:84 : Kafka commitId : unknown
> 2017-10-11 20:50:42,586 DEBUG [http-bio-7070-exec-8] kafka.KafkaSource:107
> : The end offsets are {0=0}
> 2017-10-11 20:50:42,588 ERROR [http-bio-7070-exec-8]
> controller.CubeController:305 : No new message comes, startOffset =
> endOffset:0
> java.lang.IllegalArgumentException: No new message comes, startOffset =
> endOffset:0
> at org.apache.kylin.source.kafka.KafkaSource.
> enrichSourcePartitionBeforeBuild(KafkaSource.java:134)
> at org.apache.kylin.rest.service.JobService.submitJobInternal(
> JobService.java:236)
> at org.apache.kylin.rest.service.JobService.submitJob(
> JobService.java:208)
> at org.apache.kylin.rest.service.JobService$$
> FastClassBySpringCGLIB$$83a44b2a.invoke(<generated>)
>
> Regards,
> Manoj
>
>
> This message is confidential and subject to terms at:
> http://www.jpmorgan.com/emaildisclaimer including on confidentiality,
> legal privilege, viruses and monitoring of electronic messages. If you are
> not the intended recipient, please delete this message and notify the
> sender immediately. Any unauthorized use is strictly prohibited.
>
>