You are viewing a plain text version of this content. The canonical link for it is here.
Posted to dev@ambari.apache.org by Andrew Onischuk <ao...@hortonworks.com> on 2014/10/22 22:05:46 UTC

Review Request 27052: Lzo package missing

-----------------------------------------------------------
This is an automatically generated e-mail. To reply, visit:
https://reviews.apache.org/r/27052/
-----------------------------------------------------------

Review request for Ambari and Dmitro Lisnichenko.


Bugs: AMBARI-7910
    https://issues.apache.org/jira/browse/AMBARI-7910


Repository: ambari


Description
-------

Compression jobs are failing due to missing Lzo package.

**console.log**
    
    
    
    2014-10-18 21:21:34,621|main|INFO|19267|139929561851648|MainThread|RUNNING TEST "test_Compression[com.hadoop.compression.lzo.LzoCodec-org.apache.hadoop.io.compress.DefaultCodec-NONE-TextFormat]" at location "tests/mapred/mapred_1/Compression/test_Compression_20.py" at line number "72"
    2014-10-18 21:21:34,622|beaver.machine|INFO|19267|139929561851648|MainThread|RUNNING: /usr/hdp/current/hadoop-client/bin/hadoop jar /usr/hdp/current/hadoop-mapreduce-client/hadoop-mapreduce-examples.jar sort -Dmapreduce.map.output.compress=true -Dmapreduce.map.output.compress.codec=org.apache.hadoop.io.compress.DefaultCodec -Dmapreduce.output.fileoutputformat.compress=true -Dmapreduce.output.fileoutputformat.compression.type=NONE -Dmapreduce.output.fileoutputformat.compress.codec=com.hadoop.compression.lzo.LzoCodec -outKey org.apache.hadoop.io.Text -outValue org.apache.hadoop.io.Text  Compression/textinput Compression/textoutput-1413667294.62
    2014-10-18 21:21:36,591|beaver.machine|INFO|19267|139929561851648|MainThread|14/10/18 21:21:36 INFO client.RMProxy: Connecting to ResourceManager at ambari-rerun-su-1.cs1cloud.internal/172.18.146.170:8050
    2014-10-18 21:21:37,982|beaver.machine|INFO|19267|139929561851648|MainThread|Running on 1 nodes to sort from hdfs://ambari-rerun-su-1.cs1cloud.internal:8020/user/hrt_qa/Compression/textinput into hdfs://ambari-rerun-su-1.cs1cloud.internal:8020/user/hrt_qa/Compression/textoutput-1413667294.62 with 1 reduces.
    2014-10-18 21:21:37,988|beaver.machine|INFO|19267|139929561851648|MainThread|Job started: Sat Oct 18 21:21:37 UTC 2014
    2014-10-18 21:21:38,026|beaver.machine|INFO|19267|139929561851648|MainThread|14/10/18 21:21:38 INFO client.RMProxy: Connecting to ResourceManager at ambari-rerun-su-1.cs1cloud.internal/172.18.146.170:8050
    2014-10-18 21:21:38,113|beaver.machine|INFO|19267|139929561851648|MainThread|14/10/18 21:21:38 INFO hdfs.DFSClient: Created HDFS_DELEGATION_TOKEN token 82 for hrt_qa on 172.18.146.170:8020
    2014-10-18 21:21:38,138|beaver.machine|INFO|19267|139929561851648|MainThread|14/10/18 21:21:38 INFO security.TokenCache: Got dt for hdfs://ambari-rerun-su-1.cs1cloud.internal:8020; Kind: HDFS_DELEGATION_TOKEN, Service: 172.18.146.170:8020, Ident: (HDFS_DELEGATION_TOKEN token 82 for hrt_qa)
    2014-10-18 21:21:38,695|beaver.machine|INFO|19267|139929561851648|MainThread|14/10/18 21:21:38 INFO input.FileInputFormat: Total input paths to process : 1
    2014-10-18 21:21:38,908|beaver.machine|INFO|19267|139929561851648|MainThread|14/10/18 21:21:38 INFO mapreduce.JobSubmitter: number of splits:1
    2014-10-18 21:21:39,273|beaver.machine|INFO|19267|139929561851648|MainThread|14/10/18 21:21:39 INFO mapreduce.JobSubmitter: Submitting tokens for job: job_1413586769062_0027
    2014-10-18 21:21:39,275|beaver.machine|INFO|19267|139929561851648|MainThread|14/10/18 21:21:39 INFO mapreduce.JobSubmitter: Kind: HDFS_DELEGATION_TOKEN, Service: 172.18.146.170:8020, Ident: (HDFS_DELEGATION_TOKEN token 82 for hrt_qa)
    2014-10-18 21:21:39,794|beaver.machine|INFO|19267|139929561851648|MainThread|14/10/18 21:21:39 INFO impl.YarnClientImpl: Submitted application application_1413586769062_0027
    2014-10-18 21:21:39,865|beaver.machine|INFO|19267|139929561851648|MainThread|14/10/18 21:21:39 INFO mapreduce.Job: The url to track the job: http://ambari-rerun-su-1.cs1cloud.internal:8088/proxy/application_1413586769062_0027/
    2014-10-18 21:21:39,866|beaver.machine|INFO|19267|139929561851648|MainThread|14/10/18 21:21:39 INFO mapreduce.Job: Running job: job_1413586769062_0027
    2014-10-18 21:21:51,064|beaver.machine|INFO|19267|139929561851648|MainThread|14/10/18 21:21:51 INFO mapreduce.Job: Job job_1413586769062_0027 running in uber mode : false
    2014-10-18 21:21:51,067|beaver.machine|INFO|19267|139929561851648|MainThread|14/10/18 21:21:51 INFO mapreduce.Job:  map 0% reduce 0%
    2014-10-18 21:21:57,153|beaver.machine|INFO|19267|139929561851648|MainThread|14/10/18 21:21:57 INFO mapreduce.Job:  map 100% reduce 0%
    2014-10-18 21:22:03,201|beaver.machine|INFO|19267|139929561851648|MainThread|14/10/18 21:22:03 INFO mapreduce.Job: Task Id : attempt_1413586769062_0027_r_000000_0, Status : FAILED
    2014-10-18 21:22:03,221|beaver.machine|INFO|19267|139929561851648|MainThread|Error: java.lang.IllegalArgumentException: Compression codec com.hadoop.compression.lzo.LzoCodec was not found.
    2014-10-18 21:22:03,222|beaver.machine|INFO|19267|139929561851648|MainThread|at org.apache.hadoop.mapreduce.lib.output.FileOutputFormat.getOutputCompressorClass(FileOutputFormat.java:122)
    2014-10-18 21:22:03,222|beaver.machine|INFO|19267|139929561851648|MainThread|at org.apache.hadoop.mapreduce.lib.output.SequenceFileOutputFormat.getSequenceWriter(SequenceFileOutputFormat.java:56)
    2014-10-18 21:22:03,222|beaver.machine|INFO|19267|139929561851648|MainThread|at org.apache.hadoop.mapreduce.lib.output.SequenceFileOutputFormat.getRecordWriter(SequenceFileOutputFormat.java:75)
    2014-10-18 21:22:03,222|beaver.machine|INFO|19267|139929561851648|MainThread|at org.apache.hadoop.mapred.ReduceTask$NewTrackingRecordWriter.<init>(ReduceTask.java:540)
    2014-10-18 21:22:03,222|beaver.machine|INFO|19267|139929561851648|MainThread|at org.apache.hadoop.mapred.ReduceTask.runNewReducer(ReduceTask.java:614)
    2014-10-18 21:22:03,222|beaver.machine|INFO|19267|139929561851648|MainThread|at org.apache.hadoop.mapred.ReduceTask.run(ReduceTask.java:389)
    2014-10-18 21:22:03,223|beaver.machine|INFO|19267|139929561851648|MainThread|at org.apache.hadoop.mapred.YarnChild$2.run(YarnChild.java:163)
    2014-10-18 21:22:03,223|beaver.machine|INFO|19267|139929561851648|MainThread|at java.security.AccessController.doPrivileged(Native Method)
    2014-10-18 21:22:03,223|beaver.machine|INFO|19267|139929561851648|MainThread|at javax.security.auth.Subject.doAs(Subject.java:415)
    2014-10-18 21:22:03,223|beaver.machine|INFO|19267|139929561851648|MainThread|at org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1628)
    2014-10-18 21:22:03,223|beaver.machine|INFO|19267|139929561851648|MainThread|at org.apache.hadoop.mapred.YarnChild.main(YarnChild.java:158)
    2014-10-18 21:22:03,223|beaver.machine|INFO|19267|139929561851648|MainThread|Caused by: java.lang.ClassNotFoundException: Class com.hadoop.compression.lzo.LzoCodec not found
    2014-10-18 21:22:03,224|beaver.machine|INFO|19267|139929561851648|MainThread|at org.apache.hadoop.conf.Configuration.getClassByName(Configuration.java:1954)
    2014-10-18 21:22:03,224|beaver.machine|INFO|19267|139929561851648|MainThread|at org.apache.hadoop.mapreduce.lib.output.FileOutputFormat.getOutputCompressorClass(FileOutputFormat.java:119)
    2014-10-18 21:22:03,224|beaver.machine|INFO|19267|139929561851648|MainThread|... 10 more
    2014-10-18 21:22:03,224|beaver.machine|INFO|19267|139929561851648|MainThread|
    2014-10-18 21:22:09,269|beaver.machine|INFO|19267|139929561851648|MainThread|14/10/18 21:22:09 INFO mapreduce.Job: Task Id : attempt_1413586769062_0027_r_000000_1, Status : FAILED
    2014-10-18 21:22:09,273|beaver.machine|INFO|19267|139929561851648|MainThread|Error: java.lang.IllegalArgumentException: Compression codec com.hadoop.compression.lzo.LzoCodec was not found.
    2014-10-18 21:22:09,273|beaver.machine|INFO|19267|139929561851648|MainThread|at org.apache.hadoop.mapreduce.lib.output.FileOutputFormat.getOutputCompressorClass(FileOutputFormat.java:122)
    2014-10-18 21:22:09,274|beaver.machine|INFO|19267|139929561851648|MainThread|at org.apache.hadoop.mapreduce.lib.output.SequenceFileOutputFormat.getSequenceWriter(SequenceFileOutputFormat.java:56)
    2014-10-18 21:22:09,274|beaver.machine|INFO|19267|139929561851648|MainThread|at org.apache.hadoop.mapreduce.lib.output.SequenceFileOutputFormat.getRecordWriter(SequenceFileOutputFormat.java:75)
    2014-10-18 21:22:09,274|beaver.machine|INFO|19267|139929561851648|MainThread|at org.apache.hadoop.mapred.ReduceTask$NewTrackingRecordWriter.<init>(ReduceTask.java:540)
    2014-10-18 21:22:09,275|beaver.machine|INFO|19267|139929561851648|MainThread|at org.apache.hadoop.mapred.ReduceTask.runNewReducer(ReduceTask.java:614)
    2014-10-18 21:22:09,275|beaver.machine|INFO|19267|139929561851648|MainThread|at org.apache.hadoop.mapred.ReduceTask.run(ReduceTask.java:389)
    2014-10-18 21:22:09,276|beaver.machine|INFO|19267|139929561851648|MainThread|at org.apache.hadoop.mapred.YarnChild$2.run(YarnChild.java:163)
    2014-10-18 21:22:09,276|beaver.machine|INFO|19267|139929561851648|MainThread|at java.security.AccessController.doPrivileged(Native Method)
    2014-10-18 21:22:09,277|beaver.machine|INFO|19267|139929561851648|MainThread|at javax.security.auth.Subject.doAs(Subject.java:415)
    2014-10-18 21:22:09,277|beaver.machine|INFO|19267|139929561851648|MainThread|at org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1628)
    2014-10-18 21:22:09,278|beaver.machine|INFO|19267|139929561851648|MainThread|at org.apache.hadoop.mapred.YarnChild.main(YarnChild.java:158)
    2014-10-18 21:22:09,278|beaver.machine|INFO|19267|139929561851648|MainThread|Caused by: java.lang.ClassNotFoundException: Class com.hadoop.compression.lzo.LzoCodec not found
    2014-10-18 21:22:09,279|beaver.machine|INFO|19267|139929561851648|MainThread|at org.apache.hadoop.conf.Configuration.getClassByName(Configuration.java:1954)
    2014-10-18 21:22:09,279|beaver.machine|INFO|19267|139929561851648|MainThread|at org.apache.hadoop.mapreduce.lib.output.FileOutputFormat.getOutputCompressorClass(FileOutputFormat.java:119)
    2014-10-18 21:22:09,280|beaver.machine|INFO|19267|139929561851648|MainThread|... 10 more
    2014-10-18 21:22:09,280|beaver.machine|INFO|19267|139929561851648|MainThread|
    2014-10-18 21:22:15,329|beaver.machine|INFO|19267|139929561851648|MainThread|14/10/18 21:22:15 INFO mapreduce.Job: Task Id : attempt_1413586769062_0027_r_000000_2, Status : FAILED
    2014-10-18 21:22:15,333|beaver.machine|INFO|19267|139929561851648|MainThread|Error: java.lang.IllegalArgumentException: Compression codec com.hadoop.compression.lzo.LzoCodec was not found.
    2014-10-18 21:22:15,333|beaver.machine|INFO|19267|139929561851648|MainThread|at org.apache.hadoop.mapreduce.lib.output.FileOutputFormat.getOutputCompressorClass(FileOutputFormat.java:122)
    2014-10-18 21:22:15,334|beaver.machine|INFO|19267|139929561851648|MainThread|at org.apache.hadoop.mapreduce.lib.output.SequenceFileOutputFormat.getSequenceWriter(SequenceFileOutputFormat.java:56)
    2014-10-18 21:22:15,334|beaver.machine|INFO|19267|139929561851648|MainThread|at org.apache.hadoop.mapreduce.lib.output.SequenceFileOutputFormat.getRecordWriter(SequenceFileOutputFormat.java:75)
    2014-10-18 21:22:15,335|beaver.machine|INFO|19267|139929561851648|MainThread|at org.apache.hadoop.mapred.ReduceTask$NewTrackingRecordWriter.<init>(ReduceTask.java:540)
    2014-10-18 21:22:15,336|beaver.machine|INFO|19267|139929561851648|MainThread|at org.apache.hadoop.mapred.ReduceTask.runNewReducer(ReduceTask.java:614)
    2014-10-18 21:22:15,336|beaver.machine|INFO|19267|139929561851648|MainThread|at org.apache.hadoop.mapred.ReduceTask.run(ReduceTask.java:389)
    2014-10-18 21:22:15,337|beaver.machine|INFO|19267|139929561851648|MainThread|at org.apache.hadoop.mapred.YarnChild$2.run(YarnChild.java:163)
    2014-10-18 21:22:15,337|beaver.machine|INFO|19267|139929561851648|MainThread|at java.security.AccessController.doPrivileged(Native Method)
    2014-10-18 21:22:15,338|beaver.machine|INFO|19267|139929561851648|MainThread|at javax.security.auth.Subject.doAs(Subject.java:415)
    2014-10-18 21:22:15,339|beaver.machine|INFO|19267|139929561851648|MainThread|at org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1628)
    2014-10-18 21:22:15,339|beaver.machine|INFO|19267|139929561851648|MainThread|at org.apache.hadoop.mapred.YarnChild.main(YarnChild.java:158)
    2014-10-18 21:22:15,340|beaver.machine|INFO|19267|139929561851648|MainThread|Caused by: java.lang.ClassNotFoundException: Class com.hadoop.compression.lzo.LzoCodec not found
    2014-10-18 21:22:15,341|beaver.machine|INFO|19267|139929561851648|MainThread|at org.apache.hadoop.conf.Configuration.getClassByName(Configuration.java:1954)
    2014-10-18 21:22:15,342|beaver.machine|INFO|19267|139929561851648|MainThread|at org.apache.hadoop.mapreduce.lib.output.FileOutputFormat.getOutputCompressorClass(FileOutputFormat.java:119)
    2014-10-18 21:22:15,343|beaver.machine|INFO|19267|139929561851648|MainThread|... 10 more
    2014-10-18 21:22:15,343|beaver.machine|INFO|19267|139929561851648|MainThread|
    2014-10-18 21:22:23,398|beaver.machine|INFO|19267|139929561851648|MainThread|14/10/18 21:22:23 INFO mapreduce.Job:  map 100% reduce 100%
    2014-10-18 21:22:23,418|beaver.machine|INFO|19267|139929561851648|MainThread|14/10/18 21:22:23 INFO mapreduce.Job: Job job_1413586769062_0027 failed with state FAILED due to: Task failed task_1413586769062_0027_r_000000
    2014-10-18 21:22:23,419|beaver.machine|INFO|19267|139929561851648|MainThread|Job failed as tasks failed. failedMaps:0 failedReduces:1
    2014-10-18 21:22:23,419|beaver.machine|INFO|19267|139929561851648|MainThread|
    2014-10-18 21:22:23,592|beaver.machine|INFO|19267|139929561851648|MainThread|14/10/18 21:22:23 INFO mapreduce.Job: Counters: 37
    2014-10-18 21:22:23,592|beaver.machine|INFO|19267|139929561851648|MainThread|File System Counters
    2014-10-18 21:22:23,593|beaver.machine|INFO|19267|139929561851648|MainThread|FILE: Number of bytes read=0
    2014-10-18 21:22:23,593|beaver.machine|INFO|19267|139929561851648|MainThread|FILE: Number of bytes written=115511
    2014-10-18 21:22:23,593|beaver.machine|INFO|19267|139929561851648|MainThread|FILE: Number of read operations=0
    2014-10-18 21:22:23,593|beaver.machine|INFO|19267|139929561851648|MainThread|FILE: Number of large read operations=0
    2014-10-18 21:22:23,593|beaver.machine|INFO|19267|139929561851648|MainThread|FILE: Number of write operations=0
    2014-10-18 21:22:23,593|beaver.machine|INFO|19267|139929561851648|MainThread|HDFS: Number of bytes read=1767
    2014-10-18 21:22:23,594|beaver.machine|INFO|19267|139929561851648|MainThread|HDFS: Number of bytes written=0
    2014-10-18 21:22:23,594|beaver.machine|INFO|19267|139929561851648|MainThread|HDFS: Number of read operations=4
    2014-10-18 21:22:23,594|beaver.machine|INFO|19267|139929561851648|MainThread|HDFS: Number of large read operations=0
    2014-10-18 21:22:23,594|beaver.machine|INFO|19267|139929561851648|MainThread|HDFS: Number of write operations=0
    2014-10-18 21:22:23,594|beaver.machine|INFO|19267|139929561851648|MainThread|Job Counters
    2014-10-18 21:22:23,594|beaver.machine|INFO|19267|139929561851648|MainThread|Failed reduce tasks=4
    2014-10-18 21:22:23,594|beaver.machine|INFO|19267|139929561851648|MainThread|Launched map tasks=1
    2014-10-18 21:22:23,595|beaver.machine|INFO|19267|139929561851648|MainThread|Launched reduce tasks=4
    2014-10-18 21:22:23,595|beaver.machine|INFO|19267|139929561851648|MainThread|Data-local map tasks=1
    2014-10-18 21:22:23,595|beaver.machine|INFO|19267|139929561851648|MainThread|Total time spent by all maps in occupied slots (ms)=3977
    2014-10-18 21:22:23,595|beaver.machine|INFO|19267|139929561851648|MainThread|Total time spent by all reduces in occupied slots (ms)=15442
    2014-10-18 21:22:23,595|beaver.machine|INFO|19267|139929561851648|MainThread|Total time spent by all map tasks (ms)=3977
    2014-10-18 21:22:23,596|beaver.machine|INFO|19267|139929561851648|MainThread|Total time spent by all reduce tasks (ms)=15442
    2014-10-18 21:22:23,596|beaver.machine|INFO|19267|139929561851648|MainThread|Total vcore-seconds taken by all map tasks=3977
    2014-10-18 21:22:23,596|beaver.machine|INFO|19267|139929561851648|MainThread|Total vcore-seconds taken by all reduce tasks=15442
    2014-10-18 21:22:23,597|beaver.machine|INFO|19267|139929561851648|MainThread|Total megabyte-seconds taken by all map tasks=4072448
    2014-10-18 21:22:23,597|beaver.machine|INFO|19267|139929561851648|MainThread|Total megabyte-seconds taken by all reduce tasks=15812608
    2014-10-18 21:22:23,598|beaver.machine|INFO|19267|139929561851648|MainThread|Map-Reduce Framework
    2014-10-18 21:22:23,598|beaver.machine|INFO|19267|139929561851648|MainThread|Map input records=2
    2014-10-18 21:22:23,598|beaver.machine|INFO|19267|139929561851648|MainThread|Map output records=2
    2014-10-18 21:22:23,598|beaver.machine|INFO|19267|139929561851648|MainThread|Map output bytes=1514
    2014-10-18 21:22:23,598|beaver.machine|INFO|19267|139929561851648|MainThread|Map output materialized bytes=885
    2014-10-18 21:22:23,598|beaver.machine|INFO|19267|139929561851648|MainThread|Input split bytes=159
    2014-10-18 21:22:23,599|beaver.machine|INFO|19267|139929561851648|MainThread|Combine input records=0
    2014-10-18 21:22:23,599|beaver.machine|INFO|19267|139929561851648|MainThread|Spilled Records=2
    2014-10-18 21:22:23,599|beaver.machine|INFO|19267|139929561851648|MainThread|Failed Shuffles=0
    2014-10-18 21:22:23,599|beaver.machine|INFO|19267|139929561851648|MainThread|Merged Map outputs=0
    2014-10-18 21:22:23,599|beaver.machine|INFO|19267|139929561851648|MainThread|GC time elapsed (ms)=17
    2014-10-18 21:22:23,599|beaver.machine|INFO|19267|139929561851648|MainThread|CPU time spent (ms)=670
    2014-10-18 21:22:23,599|beaver.machine|INFO|19267|139929561851648|MainThread|Physical memory (bytes) snapshot=595439616
    2014-10-18 21:22:23,600|beaver.machine|INFO|19267|139929561851648|MainThread|Virtual memory (bytes) snapshot=1667874816
    2014-10-18 21:22:23,600|beaver.machine|INFO|19267|139929561851648|MainThread|Total committed heap usage (bytes)=632291328
    2014-10-18 21:22:23,600|beaver.machine|INFO|19267|139929561851648|MainThread|File Input Format Counters
    2014-10-18 21:22:23,600|beaver.machine|INFO|19267|139929561851648|MainThread|Bytes Read=1608
    2014-10-18 21:22:23,600|beaver.machine|INFO|19267|139929561851648|MainThread|Job ended: Sat Oct 18 21:22:23 UTC 2014
    2014-10-18 21:22:23,600|beaver.machine|INFO|19267|139929561851648|MainThread|The job took 45 seconds.
    

**lzo package search in nano**
    
    
    
    find / -name "*lzo*"
    /usr/hdp/2.2.0.0-908/hadoop/lib/hadoop-lzo-0.6.0.jar
    /usr/hdp/current/share/lzo
    /usr/hdp/current/share/lzo/0.6.0/lib/hadoop-lzo-0.6.0.jar
    

**lzo package search in ambari**
    
    
    
    sudo find / -name "*lzo*"
    /usr/share/mime/application/x-lzop.xml
    /grid/0/hadoopqe/set_tez_lzo.ps1
    /grid/0/hadoopqe/tests/flume/conf/exec-file-hdfs-lzop.properties
    /grid/0/hadoopqe/tests/flume/conf/exec-file-hdfs-lzo.properties
    /grid/0/hadoopqe/tests/flume/conf/exec-memory-hdfs-lzop.properties
    /grid/0/hadoopqe/tests/flume/conf/exec-memory-hdfs-lzo.properties
    /lib/modules/2.6.32-358.el6.x86_64/kernel/crypto/lzo.ko
    /lib/modules/2.6.32-358.el6.x86_64/kernel/lib/lzo
    /lib/modules/2.6.32-358.el6.x86_64/kernel/lib/lzo/lzo_compress.ko
    /lib/modules/2.6.32-358.el6.x86_64/kernel/lib/lzo/lzo_decompress.ko
    

**hadoop version**
    
    
    
    hadoop version
    Hadoop 2.6.0.2.2.0.0-945
    Subversion git@github.com:hortonworks/hadoop.git -r 5e72cc2773fc079a72735bd3f4fd347ed24df743
    Compiled by jenkins on 2014-10-16T23:47Z
    Compiled with protoc 2.5.0
    From source with checksum af8da4bc9b78bbbd52225cb96f1bd71
    This command was run using /usr/hdp/2.2.0.0-945/hadoop/hadoop-common-2.6.0.2.2.0.0-945.jar
    


Diffs
-----

  ambari-server/src/main/resources/stacks/HDP/1.3.2/services/HDFS/configuration/core-site.xml f9d6ebe 
  ambari-server/src/main/resources/stacks/HDP/1.3.2/services/HDFS/package/scripts/hdfs.py 6f24f8e 
  ambari-server/src/main/resources/stacks/HDP/1.3.2/services/HDFS/package/scripts/params.py 504e6ac 
  ambari-server/src/main/resources/stacks/HDP/2.0.6/services/HDFS/configuration/core-site.xml 52764ac 
  ambari-server/src/main/resources/stacks/HDP/2.0.6/services/HDFS/package/scripts/hdfs.py 873aa15 
  ambari-server/src/main/resources/stacks/HDP/2.0.6/services/HDFS/package/scripts/params.py 404244b 
  ambari-server/src/test/python/stacks/1.3.2/configs/default.json 74bbb64 
  ambari-server/src/test/python/stacks/1.3.2/configs/secured.json d504281 

Diff: https://reviews.apache.org/r/27052/diff/


Testing
-------

mvn clean test


Thanks,

Andrew Onischuk


Re: Review Request 27052: Lzo package missing

Posted by Mahadev Konar <ma...@apache.org>.
-----------------------------------------------------------
This is an automatically generated e-mail. To reply, visit:
https://reviews.apache.org/r/27052/#review57885
-----------------------------------------------------------

Ship it!


Ship It!

- Mahadev Konar


On Oct. 22, 2014, 8:05 p.m., Andrew Onischuk wrote:
> 
> -----------------------------------------------------------
> This is an automatically generated e-mail. To reply, visit:
> https://reviews.apache.org/r/27052/
> -----------------------------------------------------------
> 
> (Updated Oct. 22, 2014, 8:05 p.m.)
> 
> 
> Review request for Ambari and Dmitro Lisnichenko.
> 
> 
> Bugs: AMBARI-7910
>     https://issues.apache.org/jira/browse/AMBARI-7910
> 
> 
> Repository: ambari
> 
> 
> Description
> -------
> 
> Compression jobs are failing due to missing Lzo package.
> 
> **console.log**
>     
>     
>     
>     2014-10-18 21:21:34,621|main|INFO|19267|139929561851648|MainThread|RUNNING TEST "test_Compression[com.hadoop.compression.lzo.LzoCodec-org.apache.hadoop.io.compress.DefaultCodec-NONE-TextFormat]" at location "tests/mapred/mapred_1/Compression/test_Compression_20.py" at line number "72"
>     2014-10-18 21:21:34,622|beaver.machine|INFO|19267|139929561851648|MainThread|RUNNING: /usr/hdp/current/hadoop-client/bin/hadoop jar /usr/hdp/current/hadoop-mapreduce-client/hadoop-mapreduce-examples.jar sort -Dmapreduce.map.output.compress=true -Dmapreduce.map.output.compress.codec=org.apache.hadoop.io.compress.DefaultCodec -Dmapreduce.output.fileoutputformat.compress=true -Dmapreduce.output.fileoutputformat.compression.type=NONE -Dmapreduce.output.fileoutputformat.compress.codec=com.hadoop.compression.lzo.LzoCodec -outKey org.apache.hadoop.io.Text -outValue org.apache.hadoop.io.Text  Compression/textinput Compression/textoutput-1413667294.62
>     2014-10-18 21:21:36,591|beaver.machine|INFO|19267|139929561851648|MainThread|14/10/18 21:21:36 INFO client.RMProxy: Connecting to ResourceManager at ambari-rerun-su-1.cs1cloud.internal/172.18.146.170:8050
>     2014-10-18 21:21:37,982|beaver.machine|INFO|19267|139929561851648|MainThread|Running on 1 nodes to sort from hdfs://ambari-rerun-su-1.cs1cloud.internal:8020/user/hrt_qa/Compression/textinput into hdfs://ambari-rerun-su-1.cs1cloud.internal:8020/user/hrt_qa/Compression/textoutput-1413667294.62 with 1 reduces.
>     2014-10-18 21:21:37,988|beaver.machine|INFO|19267|139929561851648|MainThread|Job started: Sat Oct 18 21:21:37 UTC 2014
>     2014-10-18 21:21:38,026|beaver.machine|INFO|19267|139929561851648|MainThread|14/10/18 21:21:38 INFO client.RMProxy: Connecting to ResourceManager at ambari-rerun-su-1.cs1cloud.internal/172.18.146.170:8050
>     2014-10-18 21:21:38,113|beaver.machine|INFO|19267|139929561851648|MainThread|14/10/18 21:21:38 INFO hdfs.DFSClient: Created HDFS_DELEGATION_TOKEN token 82 for hrt_qa on 172.18.146.170:8020
>     2014-10-18 21:21:38,138|beaver.machine|INFO|19267|139929561851648|MainThread|14/10/18 21:21:38 INFO security.TokenCache: Got dt for hdfs://ambari-rerun-su-1.cs1cloud.internal:8020; Kind: HDFS_DELEGATION_TOKEN, Service: 172.18.146.170:8020, Ident: (HDFS_DELEGATION_TOKEN token 82 for hrt_qa)
>     2014-10-18 21:21:38,695|beaver.machine|INFO|19267|139929561851648|MainThread|14/10/18 21:21:38 INFO input.FileInputFormat: Total input paths to process : 1
>     2014-10-18 21:21:38,908|beaver.machine|INFO|19267|139929561851648|MainThread|14/10/18 21:21:38 INFO mapreduce.JobSubmitter: number of splits:1
>     2014-10-18 21:21:39,273|beaver.machine|INFO|19267|139929561851648|MainThread|14/10/18 21:21:39 INFO mapreduce.JobSubmitter: Submitting tokens for job: job_1413586769062_0027
>     2014-10-18 21:21:39,275|beaver.machine|INFO|19267|139929561851648|MainThread|14/10/18 21:21:39 INFO mapreduce.JobSubmitter: Kind: HDFS_DELEGATION_TOKEN, Service: 172.18.146.170:8020, Ident: (HDFS_DELEGATION_TOKEN token 82 for hrt_qa)
>     2014-10-18 21:21:39,794|beaver.machine|INFO|19267|139929561851648|MainThread|14/10/18 21:21:39 INFO impl.YarnClientImpl: Submitted application application_1413586769062_0027
>     2014-10-18 21:21:39,865|beaver.machine|INFO|19267|139929561851648|MainThread|14/10/18 21:21:39 INFO mapreduce.Job: The url to track the job: http://ambari-rerun-su-1.cs1cloud.internal:8088/proxy/application_1413586769062_0027/
>     2014-10-18 21:21:39,866|beaver.machine|INFO|19267|139929561851648|MainThread|14/10/18 21:21:39 INFO mapreduce.Job: Running job: job_1413586769062_0027
>     2014-10-18 21:21:51,064|beaver.machine|INFO|19267|139929561851648|MainThread|14/10/18 21:21:51 INFO mapreduce.Job: Job job_1413586769062_0027 running in uber mode : false
>     2014-10-18 21:21:51,067|beaver.machine|INFO|19267|139929561851648|MainThread|14/10/18 21:21:51 INFO mapreduce.Job:  map 0% reduce 0%
>     2014-10-18 21:21:57,153|beaver.machine|INFO|19267|139929561851648|MainThread|14/10/18 21:21:57 INFO mapreduce.Job:  map 100% reduce 0%
>     2014-10-18 21:22:03,201|beaver.machine|INFO|19267|139929561851648|MainThread|14/10/18 21:22:03 INFO mapreduce.Job: Task Id : attempt_1413586769062_0027_r_000000_0, Status : FAILED
>     2014-10-18 21:22:03,221|beaver.machine|INFO|19267|139929561851648|MainThread|Error: java.lang.IllegalArgumentException: Compression codec com.hadoop.compression.lzo.LzoCodec was not found.
>     2014-10-18 21:22:03,222|beaver.machine|INFO|19267|139929561851648|MainThread|at org.apache.hadoop.mapreduce.lib.output.FileOutputFormat.getOutputCompressorClass(FileOutputFormat.java:122)
>     2014-10-18 21:22:03,222|beaver.machine|INFO|19267|139929561851648|MainThread|at org.apache.hadoop.mapreduce.lib.output.SequenceFileOutputFormat.getSequenceWriter(SequenceFileOutputFormat.java:56)
>     2014-10-18 21:22:03,222|beaver.machine|INFO|19267|139929561851648|MainThread|at org.apache.hadoop.mapreduce.lib.output.SequenceFileOutputFormat.getRecordWriter(SequenceFileOutputFormat.java:75)
>     2014-10-18 21:22:03,222|beaver.machine|INFO|19267|139929561851648|MainThread|at org.apache.hadoop.mapred.ReduceTask$NewTrackingRecordWriter.<init>(ReduceTask.java:540)
>     2014-10-18 21:22:03,222|beaver.machine|INFO|19267|139929561851648|MainThread|at org.apache.hadoop.mapred.ReduceTask.runNewReducer(ReduceTask.java:614)
>     2014-10-18 21:22:03,222|beaver.machine|INFO|19267|139929561851648|MainThread|at org.apache.hadoop.mapred.ReduceTask.run(ReduceTask.java:389)
>     2014-10-18 21:22:03,223|beaver.machine|INFO|19267|139929561851648|MainThread|at org.apache.hadoop.mapred.YarnChild$2.run(YarnChild.java:163)
>     2014-10-18 21:22:03,223|beaver.machine|INFO|19267|139929561851648|MainThread|at java.security.AccessController.doPrivileged(Native Method)
>     2014-10-18 21:22:03,223|beaver.machine|INFO|19267|139929561851648|MainThread|at javax.security.auth.Subject.doAs(Subject.java:415)
>     2014-10-18 21:22:03,223|beaver.machine|INFO|19267|139929561851648|MainThread|at org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1628)
>     2014-10-18 21:22:03,223|beaver.machine|INFO|19267|139929561851648|MainThread|at org.apache.hadoop.mapred.YarnChild.main(YarnChild.java:158)
>     2014-10-18 21:22:03,223|beaver.machine|INFO|19267|139929561851648|MainThread|Caused by: java.lang.ClassNotFoundException: Class com.hadoop.compression.lzo.LzoCodec not found
>     2014-10-18 21:22:03,224|beaver.machine|INFO|19267|139929561851648|MainThread|at org.apache.hadoop.conf.Configuration.getClassByName(Configuration.java:1954)
>     2014-10-18 21:22:03,224|beaver.machine|INFO|19267|139929561851648|MainThread|at org.apache.hadoop.mapreduce.lib.output.FileOutputFormat.getOutputCompressorClass(FileOutputFormat.java:119)
>     2014-10-18 21:22:03,224|beaver.machine|INFO|19267|139929561851648|MainThread|... 10 more
>     2014-10-18 21:22:03,224|beaver.machine|INFO|19267|139929561851648|MainThread|
>     2014-10-18 21:22:09,269|beaver.machine|INFO|19267|139929561851648|MainThread|14/10/18 21:22:09 INFO mapreduce.Job: Task Id : attempt_1413586769062_0027_r_000000_1, Status : FAILED
>     2014-10-18 21:22:09,273|beaver.machine|INFO|19267|139929561851648|MainThread|Error: java.lang.IllegalArgumentException: Compression codec com.hadoop.compression.lzo.LzoCodec was not found.
>     2014-10-18 21:22:09,273|beaver.machine|INFO|19267|139929561851648|MainThread|at org.apache.hadoop.mapreduce.lib.output.FileOutputFormat.getOutputCompressorClass(FileOutputFormat.java:122)
>     2014-10-18 21:22:09,274|beaver.machine|INFO|19267|139929561851648|MainThread|at org.apache.hadoop.mapreduce.lib.output.SequenceFileOutputFormat.getSequenceWriter(SequenceFileOutputFormat.java:56)
>     2014-10-18 21:22:09,274|beaver.machine|INFO|19267|139929561851648|MainThread|at org.apache.hadoop.mapreduce.lib.output.SequenceFileOutputFormat.getRecordWriter(SequenceFileOutputFormat.java:75)
>     2014-10-18 21:22:09,274|beaver.machine|INFO|19267|139929561851648|MainThread|at org.apache.hadoop.mapred.ReduceTask$NewTrackingRecordWriter.<init>(ReduceTask.java:540)
>     2014-10-18 21:22:09,275|beaver.machine|INFO|19267|139929561851648|MainThread|at org.apache.hadoop.mapred.ReduceTask.runNewReducer(ReduceTask.java:614)
>     2014-10-18 21:22:09,275|beaver.machine|INFO|19267|139929561851648|MainThread|at org.apache.hadoop.mapred.ReduceTask.run(ReduceTask.java:389)
>     2014-10-18 21:22:09,276|beaver.machine|INFO|19267|139929561851648|MainThread|at org.apache.hadoop.mapred.YarnChild$2.run(YarnChild.java:163)
>     2014-10-18 21:22:09,276|beaver.machine|INFO|19267|139929561851648|MainThread|at java.security.AccessController.doPrivileged(Native Method)
>     2014-10-18 21:22:09,277|beaver.machine|INFO|19267|139929561851648|MainThread|at javax.security.auth.Subject.doAs(Subject.java:415)
>     2014-10-18 21:22:09,277|beaver.machine|INFO|19267|139929561851648|MainThread|at org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1628)
>     2014-10-18 21:22:09,278|beaver.machine|INFO|19267|139929561851648|MainThread|at org.apache.hadoop.mapred.YarnChild.main(YarnChild.java:158)
>     2014-10-18 21:22:09,278|beaver.machine|INFO|19267|139929561851648|MainThread|Caused by: java.lang.ClassNotFoundException: Class com.hadoop.compression.lzo.LzoCodec not found
>     2014-10-18 21:22:09,279|beaver.machine|INFO|19267|139929561851648|MainThread|at org.apache.hadoop.conf.Configuration.getClassByName(Configuration.java:1954)
>     2014-10-18 21:22:09,279|beaver.machine|INFO|19267|139929561851648|MainThread|at org.apache.hadoop.mapreduce.lib.output.FileOutputFormat.getOutputCompressorClass(FileOutputFormat.java:119)
>     2014-10-18 21:22:09,280|beaver.machine|INFO|19267|139929561851648|MainThread|... 10 more
>     2014-10-18 21:22:09,280|beaver.machine|INFO|19267|139929561851648|MainThread|
>     2014-10-18 21:22:15,329|beaver.machine|INFO|19267|139929561851648|MainThread|14/10/18 21:22:15 INFO mapreduce.Job: Task Id : attempt_1413586769062_0027_r_000000_2, Status : FAILED
>     2014-10-18 21:22:15,333|beaver.machine|INFO|19267|139929561851648|MainThread|Error: java.lang.IllegalArgumentException: Compression codec com.hadoop.compression.lzo.LzoCodec was not found.
>     2014-10-18 21:22:15,333|beaver.machine|INFO|19267|139929561851648|MainThread|at org.apache.hadoop.mapreduce.lib.output.FileOutputFormat.getOutputCompressorClass(FileOutputFormat.java:122)
>     2014-10-18 21:22:15,334|beaver.machine|INFO|19267|139929561851648|MainThread|at org.apache.hadoop.mapreduce.lib.output.SequenceFileOutputFormat.getSequenceWriter(SequenceFileOutputFormat.java:56)
>     2014-10-18 21:22:15,334|beaver.machine|INFO|19267|139929561851648|MainThread|at org.apache.hadoop.mapreduce.lib.output.SequenceFileOutputFormat.getRecordWriter(SequenceFileOutputFormat.java:75)
>     2014-10-18 21:22:15,335|beaver.machine|INFO|19267|139929561851648|MainThread|at org.apache.hadoop.mapred.ReduceTask$NewTrackingRecordWriter.<init>(ReduceTask.java:540)
>     2014-10-18 21:22:15,336|beaver.machine|INFO|19267|139929561851648|MainThread|at org.apache.hadoop.mapred.ReduceTask.runNewReducer(ReduceTask.java:614)
>     2014-10-18 21:22:15,336|beaver.machine|INFO|19267|139929561851648|MainThread|at org.apache.hadoop.mapred.ReduceTask.run(ReduceTask.java:389)
>     2014-10-18 21:22:15,337|beaver.machine|INFO|19267|139929561851648|MainThread|at org.apache.hadoop.mapred.YarnChild$2.run(YarnChild.java:163)
>     2014-10-18 21:22:15,337|beaver.machine|INFO|19267|139929561851648|MainThread|at java.security.AccessController.doPrivileged(Native Method)
>     2014-10-18 21:22:15,338|beaver.machine|INFO|19267|139929561851648|MainThread|at javax.security.auth.Subject.doAs(Subject.java:415)
>     2014-10-18 21:22:15,339|beaver.machine|INFO|19267|139929561851648|MainThread|at org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1628)
>     2014-10-18 21:22:15,339|beaver.machine|INFO|19267|139929561851648|MainThread|at org.apache.hadoop.mapred.YarnChild.main(YarnChild.java:158)
>     2014-10-18 21:22:15,340|beaver.machine|INFO|19267|139929561851648|MainThread|Caused by: java.lang.ClassNotFoundException: Class com.hadoop.compression.lzo.LzoCodec not found
>     2014-10-18 21:22:15,341|beaver.machine|INFO|19267|139929561851648|MainThread|at org.apache.hadoop.conf.Configuration.getClassByName(Configuration.java:1954)
>     2014-10-18 21:22:15,342|beaver.machine|INFO|19267|139929561851648|MainThread|at org.apache.hadoop.mapreduce.lib.output.FileOutputFormat.getOutputCompressorClass(FileOutputFormat.java:119)
>     2014-10-18 21:22:15,343|beaver.machine|INFO|19267|139929561851648|MainThread|... 10 more
>     2014-10-18 21:22:15,343|beaver.machine|INFO|19267|139929561851648|MainThread|
>     2014-10-18 21:22:23,398|beaver.machine|INFO|19267|139929561851648|MainThread|14/10/18 21:22:23 INFO mapreduce.Job:  map 100% reduce 100%
>     2014-10-18 21:22:23,418|beaver.machine|INFO|19267|139929561851648|MainThread|14/10/18 21:22:23 INFO mapreduce.Job: Job job_1413586769062_0027 failed with state FAILED due to: Task failed task_1413586769062_0027_r_000000
>     2014-10-18 21:22:23,419|beaver.machine|INFO|19267|139929561851648|MainThread|Job failed as tasks failed. failedMaps:0 failedReduces:1
>     2014-10-18 21:22:23,419|beaver.machine|INFO|19267|139929561851648|MainThread|
>     2014-10-18 21:22:23,592|beaver.machine|INFO|19267|139929561851648|MainThread|14/10/18 21:22:23 INFO mapreduce.Job: Counters: 37
>     2014-10-18 21:22:23,592|beaver.machine|INFO|19267|139929561851648|MainThread|File System Counters
>     2014-10-18 21:22:23,593|beaver.machine|INFO|19267|139929561851648|MainThread|FILE: Number of bytes read=0
>     2014-10-18 21:22:23,593|beaver.machine|INFO|19267|139929561851648|MainThread|FILE: Number of bytes written=115511
>     2014-10-18 21:22:23,593|beaver.machine|INFO|19267|139929561851648|MainThread|FILE: Number of read operations=0
>     2014-10-18 21:22:23,593|beaver.machine|INFO|19267|139929561851648|MainThread|FILE: Number of large read operations=0
>     2014-10-18 21:22:23,593|beaver.machine|INFO|19267|139929561851648|MainThread|FILE: Number of write operations=0
>     2014-10-18 21:22:23,593|beaver.machine|INFO|19267|139929561851648|MainThread|HDFS: Number of bytes read=1767
>     2014-10-18 21:22:23,594|beaver.machine|INFO|19267|139929561851648|MainThread|HDFS: Number of bytes written=0
>     2014-10-18 21:22:23,594|beaver.machine|INFO|19267|139929561851648|MainThread|HDFS: Number of read operations=4
>     2014-10-18 21:22:23,594|beaver.machine|INFO|19267|139929561851648|MainThread|HDFS: Number of large read operations=0
>     2014-10-18 21:22:23,594|beaver.machine|INFO|19267|139929561851648|MainThread|HDFS: Number of write operations=0
>     2014-10-18 21:22:23,594|beaver.machine|INFO|19267|139929561851648|MainThread|Job Counters
>     2014-10-18 21:22:23,594|beaver.machine|INFO|19267|139929561851648|MainThread|Failed reduce tasks=4
>     2014-10-18 21:22:23,594|beaver.machine|INFO|19267|139929561851648|MainThread|Launched map tasks=1
>     2014-10-18 21:22:23,595|beaver.machine|INFO|19267|139929561851648|MainThread|Launched reduce tasks=4
>     2014-10-18 21:22:23,595|beaver.machine|INFO|19267|139929561851648|MainThread|Data-local map tasks=1
>     2014-10-18 21:22:23,595|beaver.machine|INFO|19267|139929561851648|MainThread|Total time spent by all maps in occupied slots (ms)=3977
>     2014-10-18 21:22:23,595|beaver.machine|INFO|19267|139929561851648|MainThread|Total time spent by all reduces in occupied slots (ms)=15442
>     2014-10-18 21:22:23,595|beaver.machine|INFO|19267|139929561851648|MainThread|Total time spent by all map tasks (ms)=3977
>     2014-10-18 21:22:23,596|beaver.machine|INFO|19267|139929561851648|MainThread|Total time spent by all reduce tasks (ms)=15442
>     2014-10-18 21:22:23,596|beaver.machine|INFO|19267|139929561851648|MainThread|Total vcore-seconds taken by all map tasks=3977
>     2014-10-18 21:22:23,596|beaver.machine|INFO|19267|139929561851648|MainThread|Total vcore-seconds taken by all reduce tasks=15442
>     2014-10-18 21:22:23,597|beaver.machine|INFO|19267|139929561851648|MainThread|Total megabyte-seconds taken by all map tasks=4072448
>     2014-10-18 21:22:23,597|beaver.machine|INFO|19267|139929561851648|MainThread|Total megabyte-seconds taken by all reduce tasks=15812608
>     2014-10-18 21:22:23,598|beaver.machine|INFO|19267|139929561851648|MainThread|Map-Reduce Framework
>     2014-10-18 21:22:23,598|beaver.machine|INFO|19267|139929561851648|MainThread|Map input records=2
>     2014-10-18 21:22:23,598|beaver.machine|INFO|19267|139929561851648|MainThread|Map output records=2
>     2014-10-18 21:22:23,598|beaver.machine|INFO|19267|139929561851648|MainThread|Map output bytes=1514
>     2014-10-18 21:22:23,598|beaver.machine|INFO|19267|139929561851648|MainThread|Map output materialized bytes=885
>     2014-10-18 21:22:23,598|beaver.machine|INFO|19267|139929561851648|MainThread|Input split bytes=159
>     2014-10-18 21:22:23,599|beaver.machine|INFO|19267|139929561851648|MainThread|Combine input records=0
>     2014-10-18 21:22:23,599|beaver.machine|INFO|19267|139929561851648|MainThread|Spilled Records=2
>     2014-10-18 21:22:23,599|beaver.machine|INFO|19267|139929561851648|MainThread|Failed Shuffles=0
>     2014-10-18 21:22:23,599|beaver.machine|INFO|19267|139929561851648|MainThread|Merged Map outputs=0
>     2014-10-18 21:22:23,599|beaver.machine|INFO|19267|139929561851648|MainThread|GC time elapsed (ms)=17
>     2014-10-18 21:22:23,599|beaver.machine|INFO|19267|139929561851648|MainThread|CPU time spent (ms)=670
>     2014-10-18 21:22:23,599|beaver.machine|INFO|19267|139929561851648|MainThread|Physical memory (bytes) snapshot=595439616
>     2014-10-18 21:22:23,600|beaver.machine|INFO|19267|139929561851648|MainThread|Virtual memory (bytes) snapshot=1667874816
>     2014-10-18 21:22:23,600|beaver.machine|INFO|19267|139929561851648|MainThread|Total committed heap usage (bytes)=632291328
>     2014-10-18 21:22:23,600|beaver.machine|INFO|19267|139929561851648|MainThread|File Input Format Counters
>     2014-10-18 21:22:23,600|beaver.machine|INFO|19267|139929561851648|MainThread|Bytes Read=1608
>     2014-10-18 21:22:23,600|beaver.machine|INFO|19267|139929561851648|MainThread|Job ended: Sat Oct 18 21:22:23 UTC 2014
>     2014-10-18 21:22:23,600|beaver.machine|INFO|19267|139929561851648|MainThread|The job took 45 seconds.
>     
> 
> **lzo package search in nano**
>     
>     
>     
>     find / -name "*lzo*"
>     /usr/hdp/2.2.0.0-908/hadoop/lib/hadoop-lzo-0.6.0.jar
>     /usr/hdp/current/share/lzo
>     /usr/hdp/current/share/lzo/0.6.0/lib/hadoop-lzo-0.6.0.jar
>     
> 
> **lzo package search in ambari**
>     
>     
>     
>     sudo find / -name "*lzo*"
>     /usr/share/mime/application/x-lzop.xml
>     /grid/0/hadoopqe/set_tez_lzo.ps1
>     /grid/0/hadoopqe/tests/flume/conf/exec-file-hdfs-lzop.properties
>     /grid/0/hadoopqe/tests/flume/conf/exec-file-hdfs-lzo.properties
>     /grid/0/hadoopqe/tests/flume/conf/exec-memory-hdfs-lzop.properties
>     /grid/0/hadoopqe/tests/flume/conf/exec-memory-hdfs-lzo.properties
>     /lib/modules/2.6.32-358.el6.x86_64/kernel/crypto/lzo.ko
>     /lib/modules/2.6.32-358.el6.x86_64/kernel/lib/lzo
>     /lib/modules/2.6.32-358.el6.x86_64/kernel/lib/lzo/lzo_compress.ko
>     /lib/modules/2.6.32-358.el6.x86_64/kernel/lib/lzo/lzo_decompress.ko
>     
> 
> **hadoop version**
>     
>     
>     
>     hadoop version
>     Hadoop 2.6.0.2.2.0.0-945
>     Subversion git@github.com:hortonworks/hadoop.git -r 5e72cc2773fc079a72735bd3f4fd347ed24df743
>     Compiled by jenkins on 2014-10-16T23:47Z
>     Compiled with protoc 2.5.0
>     From source with checksum af8da4bc9b78bbbd52225cb96f1bd71
>     This command was run using /usr/hdp/2.2.0.0-945/hadoop/hadoop-common-2.6.0.2.2.0.0-945.jar
>     
> 
> 
> Diffs
> -----
> 
>   ambari-server/src/main/resources/stacks/HDP/1.3.2/services/HDFS/configuration/core-site.xml f9d6ebe 
>   ambari-server/src/main/resources/stacks/HDP/1.3.2/services/HDFS/package/scripts/hdfs.py 6f24f8e 
>   ambari-server/src/main/resources/stacks/HDP/1.3.2/services/HDFS/package/scripts/params.py 504e6ac 
>   ambari-server/src/main/resources/stacks/HDP/2.0.6/services/HDFS/configuration/core-site.xml 52764ac 
>   ambari-server/src/main/resources/stacks/HDP/2.0.6/services/HDFS/package/scripts/hdfs.py 873aa15 
>   ambari-server/src/main/resources/stacks/HDP/2.0.6/services/HDFS/package/scripts/params.py 404244b 
>   ambari-server/src/test/python/stacks/1.3.2/configs/default.json 74bbb64 
>   ambari-server/src/test/python/stacks/1.3.2/configs/secured.json d504281 
> 
> Diff: https://reviews.apache.org/r/27052/diff/
> 
> 
> Testing
> -------
> 
> mvn clean test
> 
> 
> Thanks,
> 
> Andrew Onischuk
> 
>