You are viewing a plain text version of this content. The canonical link for it is here.
Posted to dev@hive.apache.org by "Craig Condit (JIRA)" <ji...@apache.org> on 2014/07/22 21:38:39 UTC

[jira] [Updated] (HIVE-7473) Null values in DECIMAL columns cause serialization issues with HCatalog

     [ https://issues.apache.org/jira/browse/HIVE-7473?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ]

Craig Condit updated HIVE-7473:
-------------------------------

    Description: 
WritableHiveDecimalObjectInspector appears to be missing null checks in getPrimitiveWritableObject(Object) and getPrimitiveJavaObject(Object). The same checks do exist in WritableHiveVarcharObjectInspector.

Attempting to read from a table in HCatalog containing null values for decimal columns results in the following exception (Pig used here):

{noformat}
Error: org.apache.pig.backend.executionengine.ExecException: ERROR 6018: Error converting read value to tuple at
  org.apache.hive.hcatalog.pig.HCatBaseLoader.getNext(HCatBaseLoader.java:76) at
  org.apache.hive.hcatalog.pig.HCatLoader.getNext(HCatLoader.java:58) at
  org.apache.pig.backend.hadoop.executionengine.mapReduceLayer.PigRecordReader.nextKeyValue(PigRecordReader.java:211) at
  org.apache.hadoop.mapred.MapTask$NewTrackingRecordReader.nextKeyValue(MapTask.java:532) at
  org.apache.hadoop.mapreduce.task.MapContextImpl.nextKeyValue(MapContextImpl.java:80) at
  org.apache.hadoop.mapreduce.lib.map.WrappedMapper$Context.nextKeyValue(WrappedMapper.java:91) at
  org.apache.hadoop.mapreduce.Mapper.run(Mapper.java:144) at
  org.apache.hadoop.mapred.MapTask.runNewMapper(MapTask.java:763) at
  org.apache.hadoop.mapred.MapTask.run(MapTask.java:339) at
  org.apache.hadoop.mapred.YarnChild$2.run(YarnChild.java:162) at
  java.security.AccessController.doPrivileged(Native Method) at
  javax.security.auth.Subject.doAs(Subject.java:415) at
  org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1491) at
  org.apache.hadoop.mapred.YarnChild.main(YarnChild.java:157)
Caused by: java.lang.NullPointerException at 
  org.apache.hadoop.hive.serde2.objectinspector.primitive.WritableHiveDecimalObjectInspector.getPrimitiveJavaObject(WritableHiveDecimalObjectInspector.java:43) at
  org.apache.hadoop.hive.serde2.objectinspector.primitive.WritableHiveDecimalObjectInspector.getPrimitiveJavaObject(WritableHiveDecimalObjectInspector.java:26) at 
  org.apache.hive.hcatalog.data.HCatRecordSerDe.serializePrimitiveField(HCatRecordSerDe.java:269) at
  org.apache.hive.hcatalog.data.HCatRecordSerDe.serializeField(HCatRecordSerDe.java:192) at
  org.apache.hive.hcatalog.data.LazyHCatRecord.get(LazyHCatRecord.java:53) at
  org.apache.hive.hcatalog.data.LazyHCatRecord.get(LazyHCatRecord.java:97) at
  org.apache.hive.hcatalog.mapreduce.HCatRecordReader.nextKeyValue(HCatRecordReader.java:204) at
  org.apache.hive.hcatalog.pig.HCatBaseLoader.getNext(HCatBaseLoader.java:63)
  ... 13 more
{noformat}

  was:
WritableHiveDecimalObjectInspector appears to be missing null checks in getPrimitiveWritableObject(Object) and getPrimitiveJavaObject(Object). The same checks do exist in WritableHiveVarcharObjectInspector.

Attempting to read from a table in HCatalog containing null values for decimal columns results in the following exception (Pig used here):

{noformat}
Error: org.apache.pig.backend.executionengine.ExecException: ERROR 6018: Error converting read value to tuple at org.apache.hive.hcatalog.pig.HCatBaseLoader.getNext(HCatBaseLoader.java:76) at org.apache.hive.hcatalog.pig.HCatLoader.getNext(HCatLoader.java:58) at org.apache.pig.backend.hadoop.executionengine.mapReduceLayer.PigRecordReader.nextKeyValue(PigRecordReader.java:211) at org.apache.hadoop.mapred.MapTask$NewTrackingRecordReader.nextKeyValue(MapTask.java:532) at org.apache.hadoop.mapreduce.task.MapContextImpl.nextKeyValue(MapContextImpl.java:80) at org.apache.hadoop.mapreduce.lib.map.WrappedMapper$Context.nextKeyValue(WrappedMapper.java:91) at org.apache.hadoop.mapreduce.Mapper.run(Mapper.java:144) at org.apache.hadoop.mapred.MapTask.runNewMapper(MapTask.java:763) at org.apache.hadoop.mapred.MapTask.run(MapTask.java:339) at org.apache.hadoop.mapred.YarnChild$2.run(YarnChild.java:162) at java.security.AccessController.doPrivileged(Native Method) at javax.security.auth.Subject.doAs(Subject.java:415) at org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1491) at org.apache.hadoop.mapred.YarnChild.main(YarnChild.java:157) Caused by: java.lang.NullPointerException at org.apache.hadoop.hive.serde2.objectinspector.primitive.WritableHiveDecimalObjectInspector.getPrimitiveJavaObject(WritableHiveDecimalObjectInspector.java:43) at org.apache.hadoop.hive.serde2.objectinspector.primitive.WritableHiveDecimalObjectInspector.getPrimitiveJavaObject(WritableHiveDecimalObjectInspector.java:26) at org.apache.hive.hcatalog.data.HCatRecordSerDe.serializePrimitiveField(HCatRecordSerDe.java:269) at org.apache.hive.hcatalog.data.HCatRecordSerDe.serializeField(HCatRecordSerDe.java:192) at org.apache.hive.hcatalog.data.LazyHCatRecord.get(LazyHCatRecord.java:53) at org.apache.hive.hcatalog.data.LazyHCatRecord.get(LazyHCatRecord.java:97) at org.apache.hive.hcatalog.mapreduce.HCatRecordReader.nextKeyValue(HCatRecordReader.java:204) at org.apache.hive.hcatalog.pig.HCatBaseLoader.getNext(HCatBaseLoader.java:63) ... 13 more
{noformat}


> Null values in DECIMAL columns cause serialization issues with HCatalog
> -----------------------------------------------------------------------
>
>                 Key: HIVE-7473
>                 URL: https://issues.apache.org/jira/browse/HIVE-7473
>             Project: Hive
>          Issue Type: Bug
>          Components: Serializers/Deserializers
>    Affects Versions: 0.13.1
>            Reporter: Craig Condit
>         Attachments: HIVE-7473.patch
>
>
> WritableHiveDecimalObjectInspector appears to be missing null checks in getPrimitiveWritableObject(Object) and getPrimitiveJavaObject(Object). The same checks do exist in WritableHiveVarcharObjectInspector.
> Attempting to read from a table in HCatalog containing null values for decimal columns results in the following exception (Pig used here):
> {noformat}
> Error: org.apache.pig.backend.executionengine.ExecException: ERROR 6018: Error converting read value to tuple at
>   org.apache.hive.hcatalog.pig.HCatBaseLoader.getNext(HCatBaseLoader.java:76) at
>   org.apache.hive.hcatalog.pig.HCatLoader.getNext(HCatLoader.java:58) at
>   org.apache.pig.backend.hadoop.executionengine.mapReduceLayer.PigRecordReader.nextKeyValue(PigRecordReader.java:211) at
>   org.apache.hadoop.mapred.MapTask$NewTrackingRecordReader.nextKeyValue(MapTask.java:532) at
>   org.apache.hadoop.mapreduce.task.MapContextImpl.nextKeyValue(MapContextImpl.java:80) at
>   org.apache.hadoop.mapreduce.lib.map.WrappedMapper$Context.nextKeyValue(WrappedMapper.java:91) at
>   org.apache.hadoop.mapreduce.Mapper.run(Mapper.java:144) at
>   org.apache.hadoop.mapred.MapTask.runNewMapper(MapTask.java:763) at
>   org.apache.hadoop.mapred.MapTask.run(MapTask.java:339) at
>   org.apache.hadoop.mapred.YarnChild$2.run(YarnChild.java:162) at
>   java.security.AccessController.doPrivileged(Native Method) at
>   javax.security.auth.Subject.doAs(Subject.java:415) at
>   org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1491) at
>   org.apache.hadoop.mapred.YarnChild.main(YarnChild.java:157)
> Caused by: java.lang.NullPointerException at 
>   org.apache.hadoop.hive.serde2.objectinspector.primitive.WritableHiveDecimalObjectInspector.getPrimitiveJavaObject(WritableHiveDecimalObjectInspector.java:43) at
>   org.apache.hadoop.hive.serde2.objectinspector.primitive.WritableHiveDecimalObjectInspector.getPrimitiveJavaObject(WritableHiveDecimalObjectInspector.java:26) at 
>   org.apache.hive.hcatalog.data.HCatRecordSerDe.serializePrimitiveField(HCatRecordSerDe.java:269) at
>   org.apache.hive.hcatalog.data.HCatRecordSerDe.serializeField(HCatRecordSerDe.java:192) at
>   org.apache.hive.hcatalog.data.LazyHCatRecord.get(LazyHCatRecord.java:53) at
>   org.apache.hive.hcatalog.data.LazyHCatRecord.get(LazyHCatRecord.java:97) at
>   org.apache.hive.hcatalog.mapreduce.HCatRecordReader.nextKeyValue(HCatRecordReader.java:204) at
>   org.apache.hive.hcatalog.pig.HCatBaseLoader.getNext(HCatBaseLoader.java:63)
>   ... 13 more
> {noformat}



--
This message was sent by Atlassian JIRA
(v6.2#6252)