You are viewing a plain text version of this content. The canonical link for it is here.
Posted to issues@kylin.apache.org by "xiangakun (JIRA)" <ji...@apache.org> on 2019/06/19 10:11:00 UTC

[jira] [Created] (KYLIN-4051) build cube met the InvalidProtocolBufferException

xiangakun created KYLIN-4051:
--------------------------------

             Summary: build cube met the InvalidProtocolBufferException
                 Key: KYLIN-4051
                 URL: https://issues.apache.org/jira/browse/KYLIN-4051
             Project: Kylin
          Issue Type: Bug
          Components: Job Engine
    Affects Versions: v2.1.0
         Environment: Kylin 2.1, hbase 1.2.0-cdh5.10.1, hadoop-2.5.0-cdh5.3.2, hive 1.2.1
            Reporter: xiangakun


*Dears,*

*When I tried to build a cube, met the following errors, does anyone meet the same error before, hope to get your feedback soon, thanks in advance~*

Error: java.io.IOException: java.lang.reflect.InvocationTargetException
2019-06-19 15:57:05,569 INFO [Job 5e7639c4-660a-46dd-852e-62bf3fc4a713-1559] hive.CreateFlatHiveTableStep:38 : at org.apache.hadoop.hive.io.HiveIOExceptionHandlerChain.handleRecordReaderCreationException(HiveIOExceptionHandlerChain.java:97)
2019-06-19 15:57:05,569 INFO [Job 5e7639c4-660a-46dd-852e-62bf3fc4a713-1559] hive.CreateFlatHiveTableStep:38 : at org.apache.hadoop.hive.io.HiveIOExceptionHandlerUtil.handleRecordReaderCreationException(HiveIOExceptionHandlerUtil.java:57)
2019-06-19 15:57:05,569 INFO [Job 5e7639c4-660a-46dd-852e-62bf3fc4a713-1559] hive.CreateFlatHiveTableStep:38 : at org.apache.hadoop.hive.shims.HadoopShimsSecure$CombineFileRecordReader.initNextRecordReader(HadoopShimsSecure.java:295)
2019-06-19 15:57:05,569 INFO [Job 5e7639c4-660a-46dd-852e-62bf3fc4a713-1559] hive.CreateFlatHiveTableStep:38 : at org.apache.hadoop.hive.shims.HadoopShimsSecure$CombineFileRecordReader.<init>(HadoopShimsSecure.java:242)
2019-06-19 15:57:05,569 INFO [Job 5e7639c4-660a-46dd-852e-62bf3fc4a713-1559] hive.CreateFlatHiveTableStep:38 : at org.apache.hadoop.hive.shims.HadoopShimsSecure$CombineFileInputFormatShim.getRecordReader(HadoopShimsSecure.java:356)
2019-06-19 15:57:05,569 INFO [Job 5e7639c4-660a-46dd-852e-62bf3fc4a713-1559] hive.CreateFlatHiveTableStep:38 : at org.apache.hadoop.hive.ql.io.CombineHiveInputFormat.getRecordReader(CombineHiveInputFormat.java:591)
2019-06-19 15:57:05,569 INFO [Job 5e7639c4-660a-46dd-852e-62bf3fc4a713-1559] hive.CreateFlatHiveTableStep:38 : at org.apache.hadoop.mapred.MapTask$TrackedRecordReader.<init>(MapTask.java:169)
2019-06-19 15:57:05,569 INFO [Job 5e7639c4-660a-46dd-852e-62bf3fc4a713-1559] hive.CreateFlatHiveTableStep:38 : at org.apache.hadoop.mapred.MapTask.runOldMapper(MapTask.java:438)
2019-06-19 15:57:05,569 INFO [Job 5e7639c4-660a-46dd-852e-62bf3fc4a713-1559] hive.CreateFlatHiveTableStep:38 : at org.apache.hadoop.mapred.MapTask.run(MapTask.java:343)
2019-06-19 15:57:05,569 INFO [Job 5e7639c4-660a-46dd-852e-62bf3fc4a713-1559] hive.CreateFlatHiveTableStep:38 : at org.apache.hadoop.mapred.YarnChild$2.run(YarnChild.java:164)
2019-06-19 15:57:05,569 INFO [Job 5e7639c4-660a-46dd-852e-62bf3fc4a713-1559] hive.CreateFlatHiveTableStep:38 : at java.security.AccessController.doPrivileged(Native Method)
2019-06-19 15:57:05,569 INFO [Job 5e7639c4-660a-46dd-852e-62bf3fc4a713-1559] hive.CreateFlatHiveTableStep:38 : at javax.security.auth.Subject.doAs(Subject.java:422)
2019-06-19 15:57:05,569 INFO [Job 5e7639c4-660a-46dd-852e-62bf3fc4a713-1559] hive.CreateFlatHiveTableStep:38 : at org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1917)
2019-06-19 15:57:05,569 INFO [Job 5e7639c4-660a-46dd-852e-62bf3fc4a713-1559] hive.CreateFlatHiveTableStep:38 : at org.apache.hadoop.mapred.YarnChild.main(YarnChild.java:158)
2019-06-19 15:57:05,570 INFO [Job 5e7639c4-660a-46dd-852e-62bf3fc4a713-1559] hive.CreateFlatHiveTableStep:38 : Caused by: java.lang.reflect.InvocationTargetException
2019-06-19 15:57:05,570 INFO [Job 5e7639c4-660a-46dd-852e-62bf3fc4a713-1559] hive.CreateFlatHiveTableStep:38 : at sun.reflect.NativeConstructorAccessorImpl.newInstance0(Native Method)
2019-06-19 15:57:05,570 INFO [Job 5e7639c4-660a-46dd-852e-62bf3fc4a713-1559] hive.CreateFlatHiveTableStep:38 : at sun.reflect.NativeConstructorAccessorImpl.newInstance(NativeConstructorAccessorImpl.java:62)
2019-06-19 15:57:05,570 INFO [Job 5e7639c4-660a-46dd-852e-62bf3fc4a713-1559] hive.CreateFlatHiveTableStep:38 : at sun.reflect.DelegatingConstructorAccessorImpl.newInstance(DelegatingConstructorAccessorImpl.java:45)
2019-06-19 15:57:05,570 INFO [Job 5e7639c4-660a-46dd-852e-62bf3fc4a713-1559] hive.CreateFlatHiveTableStep:38 : at java.lang.reflect.Constructor.newInstance(Constructor.java:423)
2019-06-19 15:57:05,570 INFO [Job 5e7639c4-660a-46dd-852e-62bf3fc4a713-1559] hive.CreateFlatHiveTableStep:38 : at org.apache.hadoop.hive.shims.HadoopShimsSecure$CombineFileRecordReader.initNextRecordReader(HadoopShimsSecure.java:281)
2019-06-19 15:57:05,570 INFO [Job 5e7639c4-660a-46dd-852e-62bf3fc4a713-1559] hive.CreateFlatHiveTableStep:38 : ... 11 more
2019-06-19 15:57:05,570 INFO [Job 5e7639c4-660a-46dd-852e-62bf3fc4a713-1559] hive.CreateFlatHiveTableStep:38 : {color:#FF0000}*Caused by: com.google.protobuf.InvalidProtocolBufferException: Message missing required fields: streams[3].kind, streams[8].kind, streams[11].kind, streams[14*{color}
{color:#FF0000}*].kind, streams[21].kind, streams[24].kind, streams[28].kind*{color}

2019-06-19 15:57:05,570 INFO [Job 5e7639c4-660a-46dd-852e-62bf3fc4a713-1559] hive.CreateFlatHiveTableStep:38 : at com.google.protobuf.UninitializedMessageException.asInvalidProtocolBufferException(UninitializedMessageException.java:81)
2019-06-19 15:57:05,570 INFO [Job 5e7639c4-660a-46dd-852e-62bf3fc4a713-1559] hive.CreateFlatHiveTableStep:38 : at com.google.protobuf.AbstractParser.checkMessageInitialized(AbstractParser.java:71)
2019-06-19 15:57:05,570 INFO [Job 5e7639c4-660a-46dd-852e-62bf3fc4a713-1559] hive.CreateFlatHiveTableStep:38 : at com.google.protobuf.AbstractParser.parseFrom(AbstractParser.java:217)
2019-06-19 15:57:05,570 INFO [Job 5e7639c4-660a-46dd-852e-62bf3fc4a713-1559] hive.CreateFlatHiveTableStep:38 : at com.google.protobuf.AbstractParser.parseFrom(AbstractParser.java:223)
2019-06-19 15:57:05,570 INFO [Job 5e7639c4-660a-46dd-852e-62bf3fc4a713-1559] hive.CreateFlatHiveTableStep:38 : at com.google.protobuf.AbstractParser.parseFrom(AbstractParser.java:49)
2019-06-19 15:57:05,570 INFO [Job 5e7639c4-660a-46dd-852e-62bf3fc4a713-1559] hive.CreateFlatHiveTableStep:38 : at org.apache.hadoop.hive.ql.io.orc.OrcProto$StripeFooter.parseFrom(OrcProto.java:8878)
2019-06-19 15:57:05,570 INFO [Job 5e7639c4-660a-46dd-852e-62bf3fc4a713-1559] hive.CreateFlatHiveTableStep:38 : at org.apache.hadoop.hive.ql.io.orc.RecordReaderImpl.readStripeFooter(RecordReaderImpl.java:2174)
2019-06-19 15:57:05,570 INFO [Job 5e7639c4-660a-46dd-852e-62bf3fc4a713-1559] hive.CreateFlatHiveTableStep:38 : at org.apache.hadoop.hive.ql.io.orc.RecordReaderImpl.readStripe(RecordReaderImpl.java:2505)
2019-06-19 15:57:05,570 INFO [Job 5e7639c4-660a-46dd-852e-62bf3fc4a713-1559] hive.CreateFlatHiveTableStep:38 : at org.apache.hadoop.hive.ql.io.orc.RecordReaderImpl.advanceStripe(RecordReaderImpl.java:2949)
2019-06-19 15:57:05,570 INFO [Job 5e7639c4-660a-46dd-852e-62bf3fc4a713-1559] hive.CreateFlatHiveTableStep:38 : at org.apache.hadoop.hive.ql.io.orc.RecordReaderImpl.advanceToNextRow(RecordReaderImpl.java:2991)
2019-06-19 15:57:05,570 INFO [Job 5e7639c4-660a-46dd-852e-62bf3fc4a713-1559] hive.CreateFlatHiveTableStep:38 : at org.apache.hadoop.hive.ql.io.orc.RecordReaderImpl.<init>(RecordReaderImpl.java:284)
2019-06-19 15:57:05,571 INFO [Job 5e7639c4-660a-46dd-852e-62bf3fc4a713-1559] hive.CreateFlatHiveTableStep:38 : at org.apache.hadoop.hive.ql.io.orc.ReaderImpl.rowsOptions(ReaderImpl.java:480)
2019-06-19 15:57:05,571 INFO [Job 5e7639c4-660a-46dd-852e-62bf3fc4a713-1559] hive.CreateFlatHiveTableStep:38 : at org.apache.hadoop.hive.ql.io.orc.OrcInputFormat.createReaderFromFile(OrcInputFormat.java:214)
2019-06-19 15:57:05,571 INFO [Job 5e7639c4-660a-46dd-852e-62bf3fc4a713-1559] hive.CreateFlatHiveTableStep:38 : at org.apache.hadoop.hive.ql.io.orc.OrcInputFormat$OrcRecordReader.<init>(OrcInputFormat.java:146)
2019-06-19 15:57:05,571 INFO [Job 5e7639c4-660a-46dd-852e-62bf3fc4a713-1559] hive.CreateFlatHiveTableStep:38 : at org.apache.hadoop.hive.ql.io.orc.OrcInputFormat.getRecordReader(OrcInputFormat.java:997)
2019-06-19 15:57:05,571 INFO [Job 5e7639c4-660a-46dd-852e-62bf3fc4a713-1559] hive.CreateFlatHiveTableStep:38 : at org.apache.hadoop.hive.ql.io.CombineHiveRecordReader.<init>(CombineHiveRecordReader.java:65)

 

 

 

 

followings is cube json
{
  "uuid": "1300457a-2ef9-415e-bf90-9fa310b71b26",
  "last_modified": 1560930896836,
  "version": "2.1.0",
  "name": "sales_value_details_info_day",
  "is_draft": false,
  "model_name": "sales_value_details_info_day",
  "description": "sales_value_details_info_day",
  "null_string": null,
  "dimensions": [
    \{
      "name": "PO_NO",
      "table": "SALES_VALUE_DETAILS_INFO_DAY",
      "column": "PO_NO",
      "derived": null
    },
    \{
      "name": "ITEM_CODE",
      "table": "SALES_VALUE_DETAILS_INFO_DAY",
      "column": "ITEM_CODE",
      "derived": null
    },
    \{
      "name": "ITEM_DESC",
      "table": "SALES_VALUE_DETAILS_INFO_DAY",
      "column": "ITEM_DESC",
      "derived": null
    },
    \{
      "name": "BRAND_CODE",
      "table": "SALES_VALUE_DETAILS_INFO_DAY",
      "column": "BRAND_CODE",
      "derived": null
    },
    \{
      "name": "BRAND_NAME",
      "table": "SALES_VALUE_DETAILS_INFO_DAY",
      "column": "BRAND_NAME",
      "derived": null
    },
    \{
      "name": "DT",
      "table": "SALES_VALUE_DETAILS_INFO_DAY",
      "column": "DT",
      "derived": null
    }
  ],
  "measures": [
    \{
      "name": "_COUNT_",
      "function": {
        "expression": "COUNT",
        "parameter": {
          "type": "constant",
          "value": "1"
        },
        "returntype": "bigint"
      }
    },
    \{
      "name": "NET_SALES_COST",
      "function": {
        "expression": "SUM",
        "parameter": {
          "type": "column",
          "value": "SALES_VALUE_DETAILS_INFO_DAY.NET_SALES_COST"
        },
        "returntype": "decimal(19,4)"
      }
    },
    \{
      "name": "NET_SALES_MONEY",
      "function": {
        "expression": "SUM",
        "parameter": {
          "type": "column",
          "value": "SALES_VALUE_DETAILS_INFO_DAY.NET_SALES_MONEY"
        },
        "returntype": "decimal(19,4)"
      }
    },
    \{
      "name": "SALES_COST",
      "function": {
        "expression": "SUM",
        "parameter": {
          "type": "column",
          "value": "SALES_VALUE_DETAILS_INFO_DAY.SALES_COST"
        },
        "returntype": "decimal(19,4)"
      }
    }
  ],
  "dictionaries": [],
  "rowkey": \{
    "rowkey_columns": [
      {
        "column": "SALES_VALUE_DETAILS_INFO_DAY.PO_NO",
        "encoding": "dict",
        "isShardBy": false
      },
      \{
        "column": "SALES_VALUE_DETAILS_INFO_DAY.ITEM_CODE",
        "encoding": "dict",
        "isShardBy": false
      },
      \{
        "column": "SALES_VALUE_DETAILS_INFO_DAY.ITEM_DESC",
        "encoding": "dict",
        "isShardBy": false
      },
      \{
        "column": "SALES_VALUE_DETAILS_INFO_DAY.BRAND_CODE",
        "encoding": "dict",
        "isShardBy": false
      },
      \{
        "column": "SALES_VALUE_DETAILS_INFO_DAY.BRAND_NAME",
        "encoding": "dict",
        "isShardBy": false
      },
      \{
        "column": "SALES_VALUE_DETAILS_INFO_DAY.DT",
        "encoding": "dict",
        "isShardBy": false
      }
    ]
  },
  "hbase_mapping": \{
    "column_family": [
      {
        "name": "F1",
        "columns": [
          {
            "qualifier": "M",
            "measure_refs": [
              "_COUNT_",
              "NET_SALES_COST",
              "NET_SALES_MONEY",
              "SALES_COST"
            ]
          }
        ]
      }
    ]
  },
  "aggregation_groups": [
    \{
      "includes": [
        "SALES_VALUE_DETAILS_INFO_DAY.PO_NO",
        "SALES_VALUE_DETAILS_INFO_DAY.ITEM_CODE",
        "SALES_VALUE_DETAILS_INFO_DAY.ITEM_DESC",
        "SALES_VALUE_DETAILS_INFO_DAY.BRAND_CODE",
        "SALES_VALUE_DETAILS_INFO_DAY.BRAND_NAME",
        "SALES_VALUE_DETAILS_INFO_DAY.DT"
      ],
      "select_rule": {
        "hierarchy_dims": [],
        "mandatory_dims": [
          "SALES_VALUE_DETAILS_INFO_DAY.DT"
        ],
        "joint_dims": []
      }
    }
  ],
  "signature": "xKEMvizWVPEl6EMWHB+SGg==",
  "notify_list": [],
  "status_need_notify": [
    "ERROR",
    "DISCARDED",
    "SUCCEED"
  ],
  "partition_date_start": 1560643200000,
  "partition_date_end": 3153600000000,
  "auto_merge_time_ranges": [],
  "retention_range": 0,
  "cuboid_cut_size_mb": null,
  "exclude_cuboid_layer": null,
  "engine_type": 2,
  "storage_type": 2,
  "override_kylin_properties": \{
    "kylin.engine.mr.config-override.fs.defaultFS": "hdfs://bipcluster",
    "kylin.engine.mr.config-override.mapreduce.jobhistory.address": "sd-hadoop-journalnode-71-21.idc.vip.com:10020",
    "kylin.engine.mr.config-override.mapreduce.jobhistory.webapp.address": "sd-hadoop-journalnode-71-21.idc.vip.com:19888"
  },
  "cuboid_black_list": [],
  "parent_forward": 3
}



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)