You are viewing a plain text version of this content. The canonical link for it is here.
Posted to issues@orc.apache.org by "EbCead (JIRA)" <ji...@apache.org> on 2017/09/13 06:36:00 UTC

[jira] [Commented] (ORC-237) OrcFile.mergeFiles Specified block size is less than configured minimum value

    [ https://issues.apache.org/jira/browse/ORC-237?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=16164209#comment-16164209 ] 

EbCead commented on ORC-237:
----------------------------

17/09/12 19:28:11 WARN util.NativeCodeLoader: Unable to load native-hadoop library for your platform... using builtin-java classes where applicable
17/09/12 19:28:11 INFO security.UserGroupInformation: Login successful for user xx using keytab file /Users/pipe/xx.keytab
17/09/12 19:28:12 WARN shortcircuit.DomainSocketFactory: The short-circuit local reads feature cannot be used because libhadoop cannot be loaded.
17/09/12 19:28:12 INFO orc.OrcMainTest: Input file list:===:hdfs://a1.dm.ambari:8020/dw/ods/order_orc/success/dt=2017-06-28/part-m-00000.orc
17/09/12 19:28:12 INFO orc.OrcMainTest: Input file size:===:9304992
17/09/12 19:28:12 INFO orc.OrcMainTest: Input file list:===:hdfs://a1.dm.ambari:8020/dw/ods/order_orc/success/dt=2017-06-28/part-m-00001.orc
17/09/12 19:28:12 INFO orc.OrcMainTest: Input file size:===:9616657
17/09/12 19:28:12 INFO orc.OrcMainTest: Input file list:===:hdfs://a1.dm.ambari:8020/dw/ods/order_orc/success/dt=2017-06-28/part-m-00002.orc
17/09/12 19:28:12 INFO orc.OrcMainTest: Input file size:===:3188876
17/09/12 19:28:12 INFO impl.OrcCodecPool: Got brand-new codec ZLIB
17/09/12 19:28:12 INFO impl.PhysicalFsWriter: ORC writer created for path: /dw/ods/order_orc/success/dt=2017-06-28_tmp/part-m-00000.orc with stripeSize: 67108864 blockSize: 131072 compression: ZLIB bufferSize: 131072

org.apache.hadoop.ipc.RemoteException(java.io.IOException): Specified block size is less than configured minimum value (dfs.namenode.fs-limits.min-block-size): 131072 < 1048576
	at org.apache.hadoop.hdfs.server.namenode.FSNamesystem.startFileInt(FSNamesystem.java:2600)
	at org.apache.hadoop.hdfs.server.namenode.FSNamesystem.startFile(FSNamesystem.java:2555)
	at org.apache.hadoop.hdfs.server.namenode.NameNodeRpcServer.create(NameNodeRpcServer.java:735)
	at org.apache.hadoop.hdfs.protocolPB.ClientNamenodeProtocolServerSideTranslatorPB.create(ClientNamenodeProtocolServerSideTranslatorPB.java:408)
	at org.apache.hadoop.hdfs.protocol.proto.ClientNamenodeProtocolProtos$ClientNamenodeProtocol$2.callBlockingMethod(ClientNamenodeProtocolProtos.java)
	at org.apache.hadoop.ipc.ProtobufRpcEngine$Server$ProtoBufRpcInvoker.call(ProtobufRpcEngine.java:640)
	at org.apache.hadoop.ipc.RPC$Server.call(RPC.java:982)
	at org.apache.hadoop.ipc.Server$Handler$1.run(Server.java:2351)
	at org.apache.hadoop.ipc.Server$Handler$1.run(Server.java:2347)
	at java.security.AccessController.doPrivileged(Native Method)
	at javax.security.auth.Subject.doAs(Subject.java:422)
	at org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1866)
	at org.apache.hadoop.ipc.Server$Handler.run(Server.java:2345)


	at org.apache.hadoop.ipc.Client.call(Client.java:1475)
	at org.apache.hadoop.ipc.Client.call(Client.java:1412)
	at org.apache.hadoop.ipc.ProtobufRpcEngine$Invoker.invoke(ProtobufRpcEngine.java:229)
	at com.sun.proxy.$Proxy15.create(Unknown Source)
	at org.apache.hadoop.hdfs.protocolPB.ClientNamenodeProtocolTranslatorPB.create(ClientNamenodeProtocolTranslatorPB.java:296)
	at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
	at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62)
	at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
	at java.lang.reflect.Method.invoke(Method.java:498)
	at org.apache.hadoop.io.retry.RetryInvocationHandler.invokeMethod(RetryInvocationHandler.java:191)
	at org.apache.hadoop.io.retry.RetryInvocationHandler.invoke(RetryInvocationHandler.java:102)
	at com.sun.proxy.$Proxy16.create(Unknown Source)
	at org.apache.hadoop.hdfs.DFSOutputStream.newStreamForCreate(DFSOutputStream.java:1648)
	at org.apache.hadoop.hdfs.DFSClient.create(DFSClient.java:1689)
	at org.apache.hadoop.hdfs.DFSClient.create(DFSClient.java:1624)
	at org.apache.hadoop.hdfs.DistributedFileSystem$7.doCall(DistributedFileSystem.java:448)
	at org.apache.hadoop.hdfs.DistributedFileSystem$7.doCall(DistributedFileSystem.java:444)
	at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81)
	at org.apache.hadoop.hdfs.DistributedFileSystem.create(DistributedFileSystem.java:459)
	at org.apache.hadoop.hdfs.DistributedFileSystem.create(DistributedFileSystem.java:387)
	at org.apache.hadoop.fs.FileSystem.create(FileSystem.java:911)
	at org.apache.hadoop.fs.FileSystem.create(FileSystem.java:892)
	at org.apache.orc.impl.PhysicalFsWriter.<init>(PhysicalFsWriter.java:91)
	at org.apache.orc.impl.WriterImpl.<init>(WriterImpl.java:184)
	at org.apache.orc.OrcFile.createWriter(OrcFile.java:685)
	at org.apache.orc.OrcFile.mergeFiles(OrcFile.java:830)


> OrcFile.mergeFiles Specified block size is less than configured minimum value
> -----------------------------------------------------------------------------
>
>                 Key: ORC-237
>                 URL: https://issues.apache.org/jira/browse/ORC-237
>             Project: ORC
>          Issue Type: Bug
>            Reporter: EbCead
>
> impl.PhysicalFsWriter: ORC writer created for path: /dw/ods/order_orc/success/dt=2017-06-28_tmp/part-m-00000.orc with stripeSize: 67108864 blockSize: 131072 compression: ZLIB bufferSize: 131072
> Specified block size is less than configured minimum value (dfs.namenode.fs-limits.min-block-size): 131072 < 1048576



--
This message was sent by Atlassian JIRA
(v6.4.14#64029)