You are viewing a plain text version of this content. The canonical link for it is here.
Posted to commits@ambari.apache.org by yu...@apache.org on 2013/04/06 03:55:47 UTC

svn commit: r1465166 [4/15] - in /incubator/ambari/trunk: ./ ambari-web/app/ ambari-web/app/assets/data/background_operations/ ambari-web/app/assets/data/clusters/ ambari-web/app/assets/data/configurations/ ambari-web/app/assets/data/dashboard/ ambari-...

Added: incubator/ambari/trunk/ambari-web/app/assets/data/wizard/stack/hdp/version122/HDFS.json
URL: http://svn.apache.org/viewvc/incubator/ambari/trunk/ambari-web/app/assets/data/wizard/stack/hdp/version122/HDFS.json?rev=1465166&view=auto
==============================================================================
--- incubator/ambari/trunk/ambari-web/app/assets/data/wizard/stack/hdp/version122/HDFS.json (added)
+++ incubator/ambari/trunk/ambari-web/app/assets/data/wizard/stack/hdp/version122/HDFS.json Sat Apr  6 01:55:42 2013
@@ -0,0 +1,210 @@
+{
+  "name" : "HDFS",
+  "version" : "1.1.2",
+  "user" : "root",
+  "comment" : "Apache Hadoop Distributed File System",
+  "properties" : [ {
+    "name" : "dfs.datanode.socket.write.timeout",
+    "value" : "0",
+    "description" : "DFS Client write socket timeout",
+    "filename" : "hdfs-site.xml"
+  }, {
+    "name" : "dfs.replication.max",
+    "value" : "50",
+    "description" : "Maximal block replication.\n  ",
+    "filename" : "hdfs-site.xml"
+  }, {
+    "name" : "dfs.heartbeat.interval",
+    "value" : "3",
+    "description" : "Determines datanode heartbeat interval in seconds.",
+    "filename" : "hdfs-site.xml"
+  }, {
+    "name" : "dfs.safemode.threshold.pct",
+    "value" : "1.0f",
+    "description" : "\n        Specifies the percentage of blocks that should satisfy\n        the minimal replication requirement defined by dfs.replication.min.\n        Values less than or equal to 0 mean not to start in safe mode.\n        Values greater than 1 will make safe mode permanent.\n        ",
+    "filename" : "hdfs-site.xml"
+  }, {
+    "name" : "dfs.balance.bandwidthPerSec",
+    "value" : "6250000",
+    "description" : "\n        Specifies the maximum amount of bandwidth that each datanode\n        can utilize for the balancing purpose in term of\n        the number of bytes per second.\n  ",
+    "filename" : "hdfs-site.xml"
+  }, {
+    "name" : "dfs.block.size",
+    "value" : "134217728",
+    "description" : "The default block size for new files.",
+    "filename" : "hdfs-site.xml"
+  }, {
+    "name" : "dfs.datanode.ipc.address",
+    "value" : "0.0.0.0:8010",
+    "description" : "\nThe datanode ipc server address and port.\nIf the port is 0 then the server will start on a free port.\n",
+    "filename" : "hdfs-site.xml"
+  }, {
+    "name" : "dfs.blockreport.initialDelay",
+    "value" : "120",
+    "description" : "Delay for first block report in seconds.",
+    "filename" : "hdfs-site.xml"
+  }, {
+    "name" : "dfs.datanode.du.pct",
+    "value" : "0.85f",
+    "description" : "When calculating remaining space, only use this percentage of the real available space\n",
+    "filename" : "hdfs-site.xml"
+  }, {
+    "name" : "dfs.namenode.handler.count",
+    "value" : "40",
+    "description" : "The number of server threads for the namenode.",
+    "filename" : "hdfs-site.xml"
+  }, {
+    "name" : "dfs.datanode.max.xcievers",
+    "value" : "4096",
+    "description" : "PRIVATE CONFIG VARIABLE",
+    "filename" : "hdfs-site.xml"
+  }, {
+    "name" : "dfs.umaskmode",
+    "value" : "077",
+    "description" : "\nThe octal umask used when creating files and directories.\n",
+    "filename" : "hdfs-site.xml"
+  }, {
+    "name" : "dfs.web.ugi",
+    "value" : "gopher,gopher",
+    "description" : "The user account used by the web interface.\nSyntax: USERNAME,GROUP1,GROUP2, ...\n",
+    "filename" : "hdfs-site.xml"
+  }, {
+    "name" : "dfs.permissions",
+    "value" : "true",
+    "description" : "\nIf \"true\", enable permission checking in HDFS.\nIf \"false\", permission checking is turned off,\nbut all other behavior is unchanged.\nSwitching from one parameter value to the other does not change the mode,\nowner or group of files or directories.\n",
+    "filename" : "hdfs-site.xml"
+  }, {
+    "name" : "dfs.permissions.supergroup",
+    "value" : "hdfs",
+    "description" : "The name of the group of super-users.",
+    "filename" : "hdfs-site.xml"
+  }, {
+    "name" : "dfs.namenode.handler.count",
+    "value" : "100",
+    "description" : "Added to grow Queue size so that more client connections are allowed",
+    "filename" : "hdfs-site.xml"
+  }, {
+    "name" : "ipc.server.max.response.size",
+    "value" : "5242880",
+    "description" : null,
+    "filename" : "hdfs-site.xml"
+  }, {
+    "name" : "dfs.block.access.token.enable",
+    "value" : "true",
+    "description" : "\nIf \"true\", access tokens are used as capabilities for accessing datanodes.\nIf \"false\", no access tokens are checked on accessing datanodes.\n",
+    "filename" : "hdfs-site.xml"
+  }, {
+    "name" : "dfs.secondary.https.port",
+    "value" : "50490",
+    "description" : "The https port where secondary-namenode binds",
+    "filename" : "hdfs-site.xml"
+  }, {
+    "name" : "dfs.https.port",
+    "value" : "50470",
+    "description" : "The https port where namenode binds",
+    "filename" : "hdfs-site.xml"
+  }, {
+    "name" : "dfs.access.time.precision",
+    "value" : "0",
+    "description" : "The access time for HDFS file is precise upto this value.\n               The default value is 1 hour. Setting a value of 0 disables\n               access times for HDFS.\n  ",
+    "filename" : "hdfs-site.xml"
+  }, {
+    "name" : "dfs.cluster.administrators",
+    "value" : " hdfs",
+    "description" : "ACL for who all can view the default servlets in the HDFS",
+    "filename" : "hdfs-site.xml"
+  }, {
+    "name" : "ipc.server.read.threadpool.size",
+    "value" : "5",
+    "description" : null,
+    "filename" : "hdfs-site.xml"
+  }, {
+    "name" : "dfs.datanode.failed.volumes.tolerated",
+    "value" : "0",
+    "description" : "Number of failed disks datanode would tolerate",
+    "filename" : "hdfs-site.xml"
+  }, {
+    "name" : "io.file.buffer.size",
+    "value" : "131072",
+    "description" : "The size of buffer for use in sequence files.\n  The size of this buffer should probably be a multiple of hardware\n  page size (4096 on Intel x86), and it determines how much data is\n  buffered during read and write operations.",
+    "filename" : "core-site.xml"
+  }, {
+    "name" : "io.serializations",
+    "value" : "org.apache.hadoop.io.serializer.WritableSerialization",
+    "description" : null,
+    "filename" : "core-site.xml"
+  }, {
+    "name" : "io.compression.codec.lzo.class",
+    "value" : "com.hadoop.compression.lzo.LzoCodec",
+    "description" : "The implementation for lzo codec.",
+    "filename" : "core-site.xml"
+  }, {
+    "name" : "fs.trash.interval",
+    "value" : "360",
+    "description" : "Number of minutes between trash checkpoints.\n  If zero, the trash feature is disabled.\n  ",
+    "filename" : "core-site.xml"
+  }, {
+    "name" : "fs.checkpoint.edits.dir",
+    "value" : "${fs.checkpoint.dir}",
+    "description" : "Determines where on the local filesystem the DFS secondary\n        name node should store the temporary edits to merge.\n        If this is a comma-delimited list of directoires then teh edits is\n        replicated in all of the directoires for redundancy.\n        Default value is same as fs.checkpoint.dir\n    ",
+    "filename" : "core-site.xml"
+  }, {
+    "name" : "fs.checkpoint.period",
+    "value" : "21600",
+    "description" : "The number of seconds between two periodic checkpoints.\n  ",
+    "filename" : "core-site.xml"
+  }, {
+    "name" : "fs.checkpoint.size",
+    "value" : "536870912",
+    "description" : "The size of the current edit log (in bytes) that triggers\n       a periodic checkpoint even if the fs.checkpoint.period hasn't expired.\n  ",
+    "filename" : "core-site.xml"
+  }, {
+    "name" : "ipc.client.idlethreshold",
+    "value" : "8000",
+    "description" : "Defines the threshold number of connections after which\n               connections will be inspected for idleness.\n  ",
+    "filename" : "core-site.xml"
+  }, {
+    "name" : "ipc.client.connection.maxidletime",
+    "value" : "30000",
+    "description" : "The maximum time after which a client will bring down the\n               connection to the server.\n  ",
+    "filename" : "core-site.xml"
+  }, {
+    "name" : "ipc.client.connect.max.retries",
+    "value" : "50",
+    "description" : "Defines the maximum number of retries for IPC connections.",
+    "filename" : "core-site.xml"
+  }, {
+    "name" : "webinterface.private.actions",
+    "value" : "false",
+    "description" : " If set to true, the web interfaces of JT and NN may contain\n                actions, such as kill job, delete file, etc., that should\n                not be exposed to public. Enable this option if the interfaces\n                are only reachable by those who have the right authorization.\n  ",
+    "filename" : "core-site.xml"
+  } ],
+  "components" : [ {
+    "name" : "NAMENODE",
+    "category" : "MASTER",
+    "client" : false,
+    "master" : true
+  }, {
+    "name" : "DATANODE",
+    "category" : "SLAVE",
+    "client" : false,
+    "master" : false
+  }, {
+    "name" : "SECONDARY_NAMENODE",
+    "category" : "MASTER",
+    "client" : false,
+    "master" : true
+  }, {
+    "name" : "HDFS_CLIENT",
+    "category" : "CLIENT",
+    "client" : true,
+    "master" : false
+  } ],
+  "clientOnlyService" : false,
+  "clientComponent" : {
+    "name" : "HDFS_CLIENT",
+    "category" : "CLIENT",
+    "client" : true,
+    "master" : false
+  }
+}
\ No newline at end of file

Added: incubator/ambari/trunk/ambari-web/app/assets/data/wizard/stack/hdp/version122/HIVE.json
URL: http://svn.apache.org/viewvc/incubator/ambari/trunk/ambari-web/app/assets/data/wizard/stack/hdp/version122/HIVE.json?rev=1465166&view=auto
==============================================================================
--- incubator/ambari/trunk/ambari-web/app/assets/data/wizard/stack/hdp/version122/HIVE.json (added)
+++ incubator/ambari/trunk/ambari-web/app/assets/data/wizard/stack/hdp/version122/HIVE.json Sat Apr  6 01:55:42 2013
@@ -0,0 +1,95 @@
+{
+  "name" : "HIVE",
+  "version" : "0.10.0",
+  "user" : "root",
+  "comment" : "Data warehouse system for ad-hoc queries & analysis of large datasets and table & storage management service",
+  "properties" : [ {
+    "name" : "hive.metastore.local",
+    "value" : "false",
+    "description" : "controls whether to connect to remove metastore server or\n    open a new metastore server in Hive Client JVM",
+    "filename" : "hive-site.xml"
+  }, {
+    "name" : "javax.jdo.option.ConnectionDriverName",
+    "value" : "com.mysql.jdbc.Driver",
+    "description" : "Driver class name for a JDBC metastore",
+    "filename" : "hive-site.xml"
+  }, {
+    "name" : "hive.metastore.warehouse.dir",
+    "value" : "/apps/hive/warehouse",
+    "description" : "location of default database for the warehouse",
+    "filename" : "hive-site.xml"
+  }, {
+    "name" : "hive.metastore.cache.pinobjtypes",
+    "value" : "Table,Database,Type,FieldSchema,Order",
+    "description" : "List of comma separated metastore object types that should be pinned in the cache",
+    "filename" : "hive-site.xml"
+  }, {
+    "name" : "hive.semantic.analyzer.factory.impl",
+    "value" : "org.apache.hivealog.cli.HCatSemanticAnalyzerFactory",
+    "description" : "controls which SemanticAnalyzerFactory implemenation class is used by CLI",
+    "filename" : "hive-site.xml"
+  }, {
+    "name" : "hadoop.clientside.fs.operations",
+    "value" : "true",
+    "description" : "FS operations are owned by client",
+    "filename" : "hive-site.xml"
+  }, {
+    "name" : "hive.metastore.client.socket.timeout",
+    "value" : "60",
+    "description" : "MetaStore Client socket timeout in seconds",
+    "filename" : "hive-site.xml"
+  }, {
+    "name" : "hive.metastore.execute.setugi",
+    "value" : "true",
+    "description" : "In unsecure mode, setting this property to true will cause the metastore to execute DFS operations using the client's reported user and group permissions. Note that this property must be set on both the client and     server sides. Further note that its best effort. If client sets its to true and server sets it to false, client setting will be ignored.",
+    "filename" : "hive-site.xml"
+  }, {
+    "name" : "hive.security.authorization.enabled",
+    "value" : "true",
+    "description" : "enable or disable the hive client authorization",
+    "filename" : "hive-site.xml"
+  }, {
+    "name" : "hive.security.authorization.manager",
+    "value" : "org.apache.hcatalog.security.HdfsAuthorizationProvider",
+    "description" : "the hive client authorization manager class name.\n    The user defined authorization class should implement interface org.apache.hadoop.hive.ql.security.authorization.HiveAuthorizationProvider.  ",
+    "filename" : "hive-site.xml"
+  }, {
+    "name" : "hive.server2.enable.doAs",
+    "value" : "true",
+    "description" : null,
+    "filename" : "hive-site.xml"
+  }, {
+    "name" : "fs.hdfs.impl.disable.cache",
+    "value" : "true",
+    "description" : null,
+    "filename" : "hive-site.xml"
+  } ],
+  "components" : [ {
+    "name" : "HIVE_METASTORE",
+    "category" : "MASTER",
+    "client" : false,
+    "master" : true
+  }, {
+    "name" : "HIVE_SERVER",
+    "category" : "MASTER",
+    "client" : false,
+    "master" : true
+  }, {
+    "name" : "MYSQL_SERVER",
+    "category" : "MASTER",
+    "client" : false,
+    "master" : true
+  }, {
+    "name" : "HIVE_CLIENT",
+    "category" : "CLIENT",
+    "client" : true,
+    "master" : false
+  } ],
+  "clientOnlyService" : false,
+  "clientComponent" : {
+    "name" : "HIVE_CLIENT",
+    "category" : "CLIENT",
+    "client" : true,
+    "master" : false
+  }
+}
\ No newline at end of file

Added: incubator/ambari/trunk/ambari-web/app/assets/data/wizard/stack/hdp/version122/MAPREDUCE.json
URL: http://svn.apache.org/viewvc/incubator/ambari/trunk/ambari-web/app/assets/data/wizard/stack/hdp/version122/MAPREDUCE.json?rev=1465166&view=auto
==============================================================================
--- incubator/ambari/trunk/ambari-web/app/assets/data/wizard/stack/hdp/version122/MAPREDUCE.json (added)
+++ incubator/ambari/trunk/ambari-web/app/assets/data/wizard/stack/hdp/version122/MAPREDUCE.json Sat Apr  6 01:55:42 2013
@@ -0,0 +1,230 @@
+{
+  "name" : "MAPREDUCE",
+  "version" : "1.1.2",
+  "user" : "mapred",
+  "comment" : "Apache Hadoop Distributed Processing Framework",
+  "properties" : [ {
+    "name" : "io.sort.record.percent",
+    "value" : ".2",
+    "description" : "No description",
+    "filename" : "mapred-site.xml"
+  }, {
+    "name" : "io.sort.factor",
+    "value" : "100",
+    "description" : "No description",
+    "filename" : "mapred-site.xml"
+  }, {
+    "name" : "mapred.tasktracker.tasks.sleeptime-before-sigkill",
+    "value" : "250",
+    "description" : "Normally, this is the amount of time before killing\n  processes, and the recommended-default is 5.000 seconds - a value of\n  5000 here.  In this case, we are using it solely to blast tasks before\n  killing them, and killing them very quickly (1/4 second) to guarantee\n  that we do not leave VMs around for later jobs.\n  ",
+    "filename" : "mapred-site.xml"
+  }, {
+    "name" : "mapred.job.tracker.handler.count",
+    "value" : "50",
+    "description" : "\n    The number of server threads for the JobTracker. This should be roughly\n    4% of the number of tasktracker nodes.\n    ",
+    "filename" : "mapred-site.xml"
+  }, {
+    "name" : "mapred.system.dir",
+    "value" : "/mapred/system",
+    "description" : "No description",
+    "filename" : "mapred-site.xml"
+  }, {
+    "name" : "mapreduce.cluster.administrators",
+    "value" : " hadoop",
+    "description" : null,
+    "filename" : "mapred-site.xml"
+  }, {
+    "name" : "mapred.reduce.parallel.copies",
+    "value" : "30",
+    "description" : "No description",
+    "filename" : "mapred-site.xml"
+  }, {
+    "name" : "tasktracker.http.threads",
+    "value" : "50",
+    "description" : null,
+    "filename" : "mapred-site.xml"
+  }, {
+    "name" : "mapred.map.tasks.speculative.execution",
+    "value" : "false",
+    "description" : "If true, then multiple instances of some map tasks\n               may be executed in parallel.",
+    "filename" : "mapred-site.xml"
+  }, {
+    "name" : "mapred.reduce.tasks.speculative.execution",
+    "value" : "false",
+    "description" : "If true, then multiple instances of some reduce tasks\n               may be executed in parallel.",
+    "filename" : "mapred-site.xml"
+  }, {
+    "name" : "mapred.reduce.slowstart.completed.maps",
+    "value" : "0.05",
+    "description" : null,
+    "filename" : "mapred-site.xml"
+  }, {
+    "name" : "mapred.inmem.merge.threshold",
+    "value" : "1000",
+    "description" : "The threshold, in terms of the number of files\n  for the in-memory merge process. When we accumulate threshold number of files\n  we initiate the in-memory merge and spill to disk. A value of 0 or less than\n  0 indicates we want to DON'T have any threshold and instead depend only on\n  the ramfs's memory consumption to trigger the merge.\n  ",
+    "filename" : "mapred-site.xml"
+  }, {
+    "name" : "mapred.job.shuffle.merge.percent",
+    "value" : "0.66",
+    "description" : "The usage threshold at which an in-memory merge will be\n  initiated, expressed as a percentage of the total memory allocated to\n  storing in-memory map outputs, as defined by\n  mapred.job.shuffle.input.buffer.percent.\n  ",
+    "filename" : "mapred-site.xml"
+  }, {
+    "name" : "mapred.job.shuffle.input.buffer.percent",
+    "value" : "0.7",
+    "description" : "The percentage of memory to be allocated from the maximum heap\n  size to storing map outputs during the shuffle.\n  ",
+    "filename" : "mapred-site.xml"
+  }, {
+    "name" : "mapred.output.compression.type",
+    "value" : "BLOCK",
+    "description" : "If the job outputs are to compressed as SequenceFiles, how should\n               they be compressed? Should be one of NONE, RECORD or BLOCK.\n  ",
+    "filename" : "mapred-site.xml"
+  }, {
+    "name" : "mapred.jobtracker.completeuserjobs.maximum",
+    "value" : "0",
+    "description" : null,
+    "filename" : "mapred-site.xml"
+  }, {
+    "name" : "mapred.jobtracker.restart.recover",
+    "value" : "false",
+    "description" : "\"true\" to enable (job) recovery upon restart,\n               \"false\" to start afresh\n    ",
+    "filename" : "mapred-site.xml"
+  }, {
+    "name" : "mapred.job.reduce.input.buffer.percent",
+    "value" : "0.0",
+    "description" : "The percentage of memory- relative to the maximum heap size- to\n  retain map outputs during the reduce. When the shuffle is concluded, any\n  remaining map outputs in memory must consume less than this threshold before\n  the reduce can begin.\n  ",
+    "filename" : "mapred-site.xml"
+  }, {
+    "name" : "mapreduce.reduce.input.limit",
+    "value" : "10737418240",
+    "description" : "The limit on the input size of the reduce. (This value\n  is 10 Gb.)  If the estimated input size of the reduce is greater than\n  this value, job is failed. A value of -1 means that there is no limit\n  set. ",
+    "filename" : "mapred-site.xml"
+  }, {
+    "name" : "mapred.task.timeout",
+    "value" : "600000",
+    "description" : "The number of milliseconds before a task will be\n  terminated if it neither reads an input, writes an output, nor\n  updates its status string.\n  ",
+    "filename" : "mapred-site.xml"
+  }, {
+    "name" : "jetty.connector",
+    "value" : "org.mortbay.jetty.nio.SelectChannelConnector",
+    "description" : "No description",
+    "filename" : "mapred-site.xml"
+  }, {
+    "name" : "mapred.child.root.logger",
+    "value" : "INFO,TLA",
+    "description" : null,
+    "filename" : "mapred-site.xml"
+  }, {
+    "name" : "mapred.max.tracker.blacklists",
+    "value" : "16",
+    "description" : "\n    if node is reported blacklisted by 16 successful jobs within timeout-window, it will be graylisted\n  ",
+    "filename" : "mapred-site.xml"
+  }, {
+    "name" : "mapred.healthChecker.interval",
+    "value" : "135000",
+    "description" : null,
+    "filename" : "mapred-site.xml"
+  }, {
+    "name" : "mapred.healthChecker.script.timeout",
+    "value" : "60000",
+    "description" : null,
+    "filename" : "mapred-site.xml"
+  }, {
+    "name" : "mapred.job.tracker.persist.jobstatus.active",
+    "value" : "false",
+    "description" : "Indicates if persistency of job status information is\n  active or not.\n  ",
+    "filename" : "mapred-site.xml"
+  }, {
+    "name" : "mapred.job.tracker.persist.jobstatus.hours",
+    "value" : "1",
+    "description" : "The number of hours job status information is persisted in DFS.\n    The job status information will be available after it drops of the memory\n    queue and between jobtracker restarts. With a zero value the job status\n    information is not persisted at all in DFS.\n  ",
+    "filename" : "mapred-site.xml"
+  }, {
+    "name" : "mapred.jobtracker.retirejob.check",
+    "value" : "10000",
+    "description" : null,
+    "filename" : "mapred-site.xml"
+  }, {
+    "name" : "mapred.jobtracker.retirejob.interval",
+    "value" : "0",
+    "description" : null,
+    "filename" : "mapred-site.xml"
+  }, {
+    "name" : "mapred.job.tracker.history.completed.location",
+    "value" : "/mapred/history/done",
+    "description" : "No description",
+    "filename" : "mapred-site.xml"
+  }, {
+    "name" : "mapreduce.fileoutputcommitter.marksuccessfuljobs",
+    "value" : "false",
+    "description" : null,
+    "filename" : "mapred-site.xml"
+  }, {
+    "name" : "mapred.job.reuse.jvm.num.tasks",
+    "value" : "1",
+    "description" : "\n    How many tasks to run per jvm. If set to -1, there is no limit\n  ",
+    "filename" : "mapred-site.xml"
+  }, {
+    "name" : "hadoop.job.history.user.location",
+    "value" : "none",
+    "description" : null,
+    "filename" : "mapred-site.xml"
+  }, {
+    "name" : "mapreduce.jobtracker.staging.root.dir",
+    "value" : "/user",
+    "description" : "The Path prefix for where the staging directories should be placed. The next level is always the user's\n   name. It is a path in the default file system.",
+    "filename" : "mapred-site.xml"
+  }, {
+    "name" : "mapreduce.tasktracker.group",
+    "value" : "hadoop",
+    "description" : "The group that the task controller uses for accessing the task controller. The mapred user must be a member and users should *not* be members.",
+    "filename" : "mapred-site.xml"
+  }, {
+    "name" : "mapreduce.jobtracker.split.metainfo.maxsize",
+    "value" : "50000000",
+    "description" : "If the size of the split metainfo file is larger than this, the JobTracker will fail the job during\n    initialize.\n   ",
+    "filename" : "mapred-site.xml"
+  }, {
+    "name" : "mapreduce.history.server.embedded",
+    "value" : "false",
+    "description" : "Should job history server be embedded within Job tracker\nprocess",
+    "filename" : "mapred-site.xml"
+  }, {
+    "name" : "mapred.jobtracker.blacklist.fault-timeout-window",
+    "value" : "180",
+    "description" : "\n    3-hour sliding window (value is in minutes)\n  ",
+    "filename" : "mapred-site.xml"
+  }, {
+    "name" : "mapred.jobtracker.blacklist.fault-bucket-width",
+    "value" : "15",
+    "description" : "\n    15-minute bucket size (value is in minutes)\n  ",
+    "filename" : "mapred-site.xml"
+  }, {
+    "name" : "mapred.queue.names",
+    "value" : "default",
+    "description" : " Comma separated list of queues configured for this jobtracker.",
+    "filename" : "mapred-site.xml"
+  } ],
+  "components" : [ {
+    "name" : "JOBTRACKER",
+    "category" : "MASTER",
+    "client" : false,
+    "master" : true
+  }, {
+    "name" : "TASKTRACKER",
+    "category" : "SLAVE",
+    "client" : false,
+    "master" : false
+  }, {
+    "name" : "MAPREDUCE_CLIENT",
+    "category" : "CLIENT",
+    "client" : true,
+    "master" : false
+  } ],
+  "clientOnlyService" : false,
+  "clientComponent" : {
+    "name" : "MAPREDUCE_CLIENT",
+    "category" : "CLIENT",
+    "client" : true,
+    "master" : false
+  }
+}
\ No newline at end of file

Added: incubator/ambari/trunk/ambari-web/app/assets/data/wizard/stack/hdp/version122/OOZIE.json
URL: http://svn.apache.org/viewvc/incubator/ambari/trunk/ambari-web/app/assets/data/wizard/stack/hdp/version122/OOZIE.json?rev=1465166&view=auto
==============================================================================
--- incubator/ambari/trunk/ambari-web/app/assets/data/wizard/stack/hdp/version122/OOZIE.json (added)
+++ incubator/ambari/trunk/ambari-web/app/assets/data/wizard/stack/hdp/version122/OOZIE.json Sat Apr  6 01:55:42 2013
@@ -0,0 +1,155 @@
+{
+  "name" : "OOZIE",
+  "version" : "3.2.0",
+  "user" : "root",
+  "comment" : "System for workflow coordination and execution of Apache Hadoop jobs",
+  "properties" : [ {
+    "name" : "oozie.base.url",
+    "value" : "http://localhost:11000/oozie",
+    "description" : "Base Oozie URL.",
+    "filename" : "oozie-site.xml"
+  }, {
+    "name" : "oozie.system.id",
+    "value" : "oozie-${user.name}",
+    "description" : "\n    The Oozie system ID.\n    ",
+    "filename" : "oozie-site.xml"
+  }, {
+    "name" : "oozie.systemmode",
+    "value" : "NORMAL",
+    "description" : "\n     System mode for  Oozie at startup.\n     ",
+    "filename" : "oozie-site.xml"
+  }, {
+    "name" : "oozie.service.AuthorizationService.security.enabled",
+    "value" : "true",
+    "description" : "\n     Specifies whether security (user name/admin role) is enabled or not.\n     If disabled any user can manage Oozie system and manage any job.\n     ",
+    "filename" : "oozie-site.xml"
+  }, {
+    "name" : "oozie.service.PurgeService.older.than",
+    "value" : "30",
+    "description" : "\n     Jobs older than this value, in days, will be purged by the PurgeService.\n     ",
+    "filename" : "oozie-site.xml"
+  }, {
+    "name" : "oozie.service.PurgeService.purge.interval",
+    "value" : "3600",
+    "description" : "\n     Interval at which the purge service will run, in seconds.\n     ",
+    "filename" : "oozie-site.xml"
+  }, {
+    "name" : "oozie.service.CallableQueueService.queue.size",
+    "value" : "1000",
+    "description" : "Max callable queue size",
+    "filename" : "oozie-site.xml"
+  }, {
+    "name" : "oozie.service.CallableQueueService.threads",
+    "value" : "10",
+    "description" : "Number of threads used for executing callables",
+    "filename" : "oozie-site.xml"
+  }, {
+    "name" : "oozie.service.CallableQueueService.callable.concurrency",
+    "value" : "3",
+    "description" : "\n     Maximum concurrency for a given callable type.\n     Each command is a callable type (submit, start, run, signal, job, jobs, suspend,resume, etc).\n     Each action type is a callable type (Map-Reduce, Pig, SSH, FS, sub-workflow, etc).\n     All commands that use action executors (action-start, action-end, action-kill and action-check) use\n     the action type as the callable type.\n     ",
+    "filename" : "oozie-site.xml"
+  }, {
+    "name" : "oozie.service.coord.normal.default.timeout",
+    "value" : "120",
+    "description" : "Default timeout for a coordinator action input check (in minutes) for normal job.\n      -1 means infinite timeout",
+    "filename" : "oozie-site.xml"
+  }, {
+    "name" : "oozie.db.schema.name",
+    "value" : "oozie",
+    "description" : "\n      Oozie DataBase Name\n     ",
+    "filename" : "oozie-site.xml"
+  }, {
+    "name" : "oozie.service.HadoopAccessorService.jobTracker.whitelist",
+    "value" : " ",
+    "description" : "\n      Whitelisted job tracker for Oozie service.\n      ",
+    "filename" : "oozie-site.xml"
+  }, {
+    "name" : "oozie.authentication.type",
+    "value" : "simple",
+    "description" : "\n      ",
+    "filename" : "oozie-site.xml"
+  }, {
+    "name" : "oozie.service.HadoopAccessorService.nameNode.whitelist",
+    "value" : " ",
+    "description" : "\n      ",
+    "filename" : "oozie-site.xml"
+  }, {
+    "name" : "oozie.service.WorkflowAppService.system.libpath",
+    "value" : "/user/${user.name}/share/lib",
+    "description" : "\n      System library path to use for workflow applications.\n      This path is added to workflow application if their job properties sets\n      the property 'oozie.use.system.libpath' to true.\n      ",
+    "filename" : "oozie-site.xml"
+  }, {
+    "name" : "use.system.libpath.for.mapreduce.and.pig.jobs",
+    "value" : "false",
+    "description" : "\n      If set to true, submissions of MapReduce and Pig jobs will include\n      automatically the system library path, thus not requiring users to\n      specify where the Pig JAR files are. Instead, the ones from the system\n      library path are used.\n      ",
+    "filename" : "oozie-site.xml"
+  }, {
+    "name" : "oozie.authentication.kerberos.name.rules",
+    "value" : "\n        RULE:[2:$1@$0]([jt]t@.*TODO-KERBEROS-DOMAIN)s/.*/TODO-MAPREDUSER/\n        RULE:[2:$1@$0]([nd]n@.*TODO-KERBEROS-DOMAIN)s/.*/TODO-HDFSUSER/\n        RULE:[2:$1@$0](hm@.*TODO-KERBEROS-DOMAIN)s/.*/TODO-HBASE-USER/\n        RULE:[2:$1@$0](rs@.*TODO-KERBEROS-DOMAIN)s/.*/TODO-HBASE-USER/\n        DEFAULT\n        ",
+    "description" : "The mapping from kerberos principal names to local OS user names.",
+    "filename" : "oozie-site.xml"
+  }, {
+    "name" : "oozie.service.HadoopAccessorService.hadoop.configurations",
+    "value" : "*=/etc/hadoop/conf",
+    "description" : "\n          Comma separated AUTHORITY=HADOOP_CONF_DIR, where AUTHORITY is the HOST:PORT of\n          the Hadoop service (JobTracker, HDFS). The wildcard '*' configuration is\n          used when there is no exact match for an authority. The HADOOP_CONF_DIR contains\n          the relevant Hadoop *-site.xml files. If the path is relative is looked within\n          the Oozie configuration directory; though the path can be absolute (i.e. to point\n          to Hadoop client conf/ directories in the local filesystem.\n      ",
+    "filename" : "oozie-site.xml"
+  }, {
+    "name" : "oozie.service.ActionService.executor.ext.classes",
+    "value" : "\n            org.apache.oozie.action.email.EmailActionExecutor,\n            org.apache.oozie.action.hadoop.HiveActionExecutor,\n            org.apache.oozie.action.hadoop.ShellActionExecutor,\n            org.apache.oozie.action.hadoop.SqoopActionExecutor,\n            org.apache.oozie.action.hadoop.DistcpActionExecutor\n        ",
+    "description" : null,
+    "filename" : "oozie-site.xml"
+  }, {
+    "name" : "oozie.service.SchemaService.wf.ext.schemas",
+    "value" : "shell-action-0.1.xsd,email-action-0.1.xsd,hive-action-0.2.xsd,sqoop-action-0.2.xsd,ssh-action-0.1.xsd,distcp-action-0.1.xsd",
+    "description" : null,
+    "filename" : "oozie-site.xml"
+  }, {
+    "name" : "oozie.service.JPAService.create.db.schema",
+    "value" : "false",
+    "description" : "\n            Creates Oozie DB.\n\n            If set to true, it creates the DB schema if it does not exist. If the DB schema exists is a NOP.\n            If set to false, it does not create the DB schema. If the DB schema does not exist it fails start up.\n        ",
+    "filename" : "oozie-site.xml"
+  }, {
+    "name" : "oozie.service.JPAService.jdbc.driver",
+    "value" : "org.apache.derby.jdbc.EmbeddedDriver",
+    "description" : "\n            JDBC driver class.\n        ",
+    "filename" : "oozie-site.xml"
+  }, {
+    "name" : "oozie.service.JPAService.jdbc.url",
+    "value" : "jdbc:derby:${oozie.data.dir}/${oozie.db.schema.name}-db;create=true",
+    "description" : "\n            JDBC URL.\n        ",
+    "filename" : "oozie-site.xml"
+  }, {
+    "name" : "oozie.service.JPAService.jdbc.username",
+    "value" : "sa",
+    "description" : "\n            DB user name.\n        ",
+    "filename" : "oozie-site.xml"
+  }, {
+    "name" : "oozie.service.JPAService.jdbc.password",
+    "value" : " ",
+    "description" : "\n            DB user password.\n\n            IMPORTANT: if password is emtpy leave a 1 space string, the service trims the value,\n                       if empty Configuration assumes it is NULL.\n        ",
+    "filename" : "oozie-site.xml"
+  }, {
+    "name" : "oozie.service.JPAService.pool.max.active.conn",
+    "value" : "10",
+    "description" : "\n             Max number of connections.\n        ",
+    "filename" : "oozie-site.xml"
+  } ],
+  "components" : [ {
+    "name" : "OOZIE_SERVER",
+    "category" : "MASTER",
+    "client" : false,
+    "master" : true
+  }, {
+    "name" : "OOZIE_CLIENT",
+    "category" : "CLIENT",
+    "client" : true,
+    "master" : false
+  } ],
+  "clientOnlyService" : false,
+  "clientComponent" : {
+    "name" : "OOZIE_CLIENT",
+    "category" : "CLIENT",
+    "client" : true,
+    "master" : false
+  }
+}
\ No newline at end of file

Added: incubator/ambari/trunk/ambari-web/app/assets/data/wizard/stack/hdp/version122/WEBHCAT.json
URL: http://svn.apache.org/viewvc/incubator/ambari/trunk/ambari-web/app/assets/data/wizard/stack/hdp/version122/WEBHCAT.json?rev=1465166&view=auto
==============================================================================
--- incubator/ambari/trunk/ambari-web/app/assets/data/wizard/stack/hdp/version122/WEBHCAT.json (added)
+++ incubator/ambari/trunk/ambari-web/app/assets/data/wizard/stack/hdp/version122/WEBHCAT.json Sat Apr  6 01:55:42 2013
@@ -0,0 +1,90 @@
+{
+  "name" : "WEBHCAT",
+  "version" : "0.5.0",
+  "user" : "root",
+  "comment" : "This is comment for WEBHCAT service",
+  "properties" : [ {
+    "name" : "templeton.port",
+    "value" : "50111",
+    "description" : "The HTTP port for the main server.",
+    "filename" : "webhcat-site.xml"
+  }, {
+    "name" : "templeton.hadoop.conf.dir",
+    "value" : "/etc/hadoop/conf",
+    "description" : "The path to the Hadoop configuration.",
+    "filename" : "webhcat-site.xml"
+  }, {
+    "name" : "templeton.jar",
+    "value" : "/usr/lib/hcatalog/share/webhcat/svr/webhcat.jar",
+    "description" : "The path to the Templeton jar file.",
+    "filename" : "webhcat-site.xml"
+  }, {
+    "name" : "templeton.libjars",
+    "value" : "/usr/lib/zookeeper/zookeeper.jar",
+    "description" : "Jars to add the the classpath.",
+    "filename" : "webhcat-site.xml"
+  }, {
+    "name" : "templeton.hadoop",
+    "value" : "/usr/bin/hadoop",
+    "description" : "The path to the Hadoop executable.",
+    "filename" : "webhcat-site.xml"
+  }, {
+    "name" : "templeton.pig.archive",
+    "value" : "hdfs:///apps/webhcat/pig.tar.gz",
+    "description" : "The path to the Pig archive.",
+    "filename" : "webhcat-site.xml"
+  }, {
+    "name" : "templeton.pig.path",
+    "value" : "pig.tar.gz/pig/bin/pig",
+    "description" : "The path to the Pig executable.",
+    "filename" : "webhcat-site.xml"
+  }, {
+    "name" : "templeton.hcat",
+    "value" : "/usr/bin/hcat",
+    "description" : "The path to the hcatalog executable.",
+    "filename" : "webhcat-site.xml"
+  }, {
+    "name" : "templeton.hive.archive",
+    "value" : "hdfs:///apps/webhcat/hive.tar.gz",
+    "description" : "The path to the Hive archive.",
+    "filename" : "webhcat-site.xml"
+  }, {
+    "name" : "templeton.hive.path",
+    "value" : "hive.tar.gz/hive/bin/hive",
+    "description" : "The path to the Hive executable.",
+    "filename" : "webhcat-site.xml"
+  }, {
+    "name" : "templeton.storage.class",
+    "value" : "org.apache.hcatalog.templeton.tool.ZooKeeperStorage",
+    "description" : "The class to use as storage",
+    "filename" : "webhcat-site.xml"
+  }, {
+    "name" : "templeton.override.enabled",
+    "value" : "false",
+    "description" : "\n     Enable the override path in templeton.override.jars\n   ",
+    "filename" : "webhcat-site.xml"
+  }, {
+    "name" : "templeton.streaming.jar",
+    "value" : "hdfs:///apps/webhcat/hadoop-streaming.jar",
+    "description" : "The hdfs path to the Hadoop streaming jar file.",
+    "filename" : "webhcat-site.xml"
+  }, {
+    "name" : "templeton.exec.timeout",
+    "value" : "60000",
+    "description" : "Time out for templeton api",
+    "filename" : "webhcat-site.xml"
+  } ],
+  "components" : [ {
+    "name" : "WEBHCAT_SERVER",
+    "category" : "MASTER",
+    "client" : false,
+    "master" : true
+  } ],
+  "clientOnlyService" : false,
+  "clientComponent" : {
+    "name" : "WEBHCAT_SERVER",
+    "category" : "MASTER",
+    "client" : false,
+    "master" : true
+  }
+}
\ No newline at end of file

Added: incubator/ambari/trunk/ambari-web/app/assets/data/wizard/stack/hdp/version122/ZOOKEEPER.json
URL: http://svn.apache.org/viewvc/incubator/ambari/trunk/ambari-web/app/assets/data/wizard/stack/hdp/version122/ZOOKEEPER.json?rev=1465166&view=auto
==============================================================================
--- incubator/ambari/trunk/ambari-web/app/assets/data/wizard/stack/hdp/version122/ZOOKEEPER.json (added)
+++ incubator/ambari/trunk/ambari-web/app/assets/data/wizard/stack/hdp/version122/ZOOKEEPER.json Sat Apr  6 01:55:42 2013
@@ -0,0 +1,25 @@
+{
+  "name" : "ZOOKEEPER",
+  "version" : "3.4.5",
+  "user" : "root",
+  "comment" : "This is comment for ZOOKEEPER service",
+  "properties" : [ ],
+  "components" : [ {
+    "name" : "ZOOKEEPER_SERVER",
+    "category" : "MASTER",
+    "client" : false,
+    "master" : true
+  }, {
+    "name" : "ZOOKEEPER_CLIENT",
+    "category" : "CLIENT",
+    "client" : true,
+    "master" : false
+  } ],
+  "clientOnlyService" : false,
+  "clientComponent" : {
+    "name" : "ZOOKEEPER_CLIENT",
+    "category" : "CLIENT",
+    "client" : true,
+    "master" : false
+  }
+}
\ No newline at end of file

Added: incubator/ambari/trunk/ambari-web/app/assets/data/wizard/stack/hdp/version130/HBASE.json
URL: http://svn.apache.org/viewvc/incubator/ambari/trunk/ambari-web/app/assets/data/wizard/stack/hdp/version130/HBASE.json?rev=1465166&view=auto
==============================================================================
--- incubator/ambari/trunk/ambari-web/app/assets/data/wizard/stack/hdp/version130/HBASE.json (added)
+++ incubator/ambari/trunk/ambari-web/app/assets/data/wizard/stack/hdp/version130/HBASE.json Sat Apr  6 01:55:42 2013
@@ -0,0 +1,60 @@
+{
+  "name" : "HBASE",
+  "version" : "0.94.5",
+  "user" : "mapred",
+  "comment" : "Non-relational distributed database and centralized service for configuration management & synchronization",
+  "properties" : [ {
+    "name" : "hbase.cluster.distributed",
+    "value" : "true",
+    "description" : "The mode the cluster will be in. Possible values are\n      false for standalone mode and true for distributed mode.  If\n      false, startup will run all HBase and ZooKeeper daemons together\n      in the one JVM.\n    ",
+    "filename" : "hbase-site.xml"
+  }, {
+    "name" : "hbase.master.lease.thread.wakefrequency",
+    "value" : "3000",
+    "description" : "The interval between checks for expired region server leases.\n    This value has been reduced due to the other reduced values above so that\n    the master will notice a dead region server sooner. The default is 15 seconds.\n    ",
+    "filename" : "hbase-site.xml"
+  }, {
+    "name" : "hbase.superuser",
+    "value" : "hbase",
+    "description" : "List of users or groups (comma-separated), who are allowed\n    full privileges, regardless of stored ACLs, across the cluster.\n    Only used when HBase security is enabled.\n    ",
+    "filename" : "hbase-site.xml"
+  }, {
+    "name" : "hbase.zookeeper.property.clientPort",
+    "value" : "2181",
+    "description" : "Property from ZooKeeper's config zoo.cfg.\n    The port at which the clients will connect.\n    ",
+    "filename" : "hbase-site.xml"
+  }, {
+    "name" : "hbase.regionserver.optionalcacheflushinterval",
+    "value" : "10000",
+    "description" : "\n      Amount of time to wait since the last time a region was flushed before\n      invoking an optional cache flush. Default 60,000.\n    ",
+    "filename" : "hbase-site.xml"
+  }, {
+    "name" : "hbase.zookeeper.useMulti",
+    "value" : "true",
+    "description" : "Instructs HBase to make use of ZooKeeper's multi-update functionality.\n    This allows certain ZooKeeper operations to complete more quickly and prevents some issues\n    with rare Replication failure scenarios (see the release note of HBASE-2611 for an example).В·\n    IMPORTANT: only set this to true if all ZooKeeper servers in the cluster are on version 3.4+\n    and will not be downgraded.  ZooKeeper versions before 3.4 do not support multi-update and will\n    not fail gracefully if multi-update is invoked (see ZOOKEEPER-1495).\n    ",
+    "filename" : "hbase-site.xml"
+  } ],
+  "components" : [ {
+    "name" : "HBASE_MASTER",
+    "category" : "MASTER",
+    "client" : false,
+    "master" : true
+  }, {
+    "name" : "HBASE_REGIONSERVER",
+    "category" : "SLAVE",
+    "client" : false,
+    "master" : false
+  }, {
+    "name" : "HBASE_CLIENT",
+    "category" : "CLIENT",
+    "client" : true,
+    "master" : false
+  } ],
+  "clientOnlyService" : false,
+  "clientComponent" : {
+    "name" : "HBASE_CLIENT",
+    "category" : "CLIENT",
+    "client" : true,
+    "master" : false
+  }
+}

Added: incubator/ambari/trunk/ambari-web/app/assets/data/wizard/stack/hdp/version130/HCATALOG.json
URL: http://svn.apache.org/viewvc/incubator/ambari/trunk/ambari-web/app/assets/data/wizard/stack/hdp/version130/HCATALOG.json?rev=1465166&view=auto
==============================================================================
--- incubator/ambari/trunk/ambari-web/app/assets/data/wizard/stack/hdp/version130/HCATALOG.json (added)
+++ incubator/ambari/trunk/ambari-web/app/assets/data/wizard/stack/hdp/version130/HCATALOG.json Sat Apr  6 01:55:42 2013
@@ -0,0 +1,20 @@
+{
+  "name" : "HCATALOG",
+  "version" : "0.5.0",
+  "user" : "root",
+  "comment" : "This is comment for HCATALOG service",
+  "properties" : [ ],
+  "components" : [ {
+    "name" : "HCAT",
+    "category" : "CLIENT",
+    "client" : true,
+    "master" : false
+  } ],
+  "clientOnlyService" : true,
+  "clientComponent" : {
+    "name" : "HCAT",
+    "category" : "CLIENT",
+    "client" : true,
+    "master" : false
+  }
+}

Added: incubator/ambari/trunk/ambari-web/app/assets/data/wizard/stack/hdp/version130/HDFS.json
URL: http://svn.apache.org/viewvc/incubator/ambari/trunk/ambari-web/app/assets/data/wizard/stack/hdp/version130/HDFS.json?rev=1465166&view=auto
==============================================================================
--- incubator/ambari/trunk/ambari-web/app/assets/data/wizard/stack/hdp/version130/HDFS.json (added)
+++ incubator/ambari/trunk/ambari-web/app/assets/data/wizard/stack/hdp/version130/HDFS.json Sat Apr  6 01:55:42 2013
@@ -0,0 +1,210 @@
+{
+  "name" : "HDFS",
+  "version" : "1.1.2",
+  "user" : "root",
+  "comment" : "Apache Hadoop Distributed File System",
+  "properties" : [ {
+    "name" : "dfs.datanode.socket.write.timeout",
+    "value" : "0",
+    "description" : "DFS Client write socket timeout",
+    "filename" : "hdfs-site.xml"
+  }, {
+    "name" : "dfs.replication.max",
+    "value" : "50",
+    "description" : "Maximal block replication.\n  ",
+    "filename" : "hdfs-site.xml"
+  }, {
+    "name" : "dfs.heartbeat.interval",
+    "value" : "3",
+    "description" : "Determines datanode heartbeat interval in seconds.",
+    "filename" : "hdfs-site.xml"
+  }, {
+    "name" : "dfs.safemode.threshold.pct",
+    "value" : "1.0f",
+    "description" : "\n        Specifies the percentage of blocks that should satisfy\n        the minimal replication requirement defined by dfs.replication.min.\n        Values less than or equal to 0 mean not to start in safe mode.\n        Values greater than 1 will make safe mode permanent.\n        ",
+    "filename" : "hdfs-site.xml"
+  }, {
+    "name" : "dfs.balance.bandwidthPerSec",
+    "value" : "6250000",
+    "description" : "\n        Specifies the maximum amount of bandwidth that each datanode\n        can utilize for the balancing purpose in term of\n        the number of bytes per second.\n  ",
+    "filename" : "hdfs-site.xml"
+  }, {
+    "name" : "dfs.block.size",
+    "value" : "134217728",
+    "description" : "The default block size for new files.",
+    "filename" : "hdfs-site.xml"
+  }, {
+    "name" : "dfs.datanode.ipc.address",
+    "value" : "0.0.0.0:8010",
+    "description" : "\nThe datanode ipc server address and port.\nIf the port is 0 then the server will start on a free port.\n",
+    "filename" : "hdfs-site.xml"
+  }, {
+    "name" : "dfs.blockreport.initialDelay",
+    "value" : "120",
+    "description" : "Delay for first block report in seconds.",
+    "filename" : "hdfs-site.xml"
+  }, {
+    "name" : "dfs.datanode.du.pct",
+    "value" : "0.85f",
+    "description" : "When calculating remaining space, only use this percentage of the real available space\n",
+    "filename" : "hdfs-site.xml"
+  }, {
+    "name" : "dfs.namenode.handler.count",
+    "value" : "40",
+    "description" : "The number of server threads for the namenode.",
+    "filename" : "hdfs-site.xml"
+  }, {
+    "name" : "dfs.datanode.max.xcievers",
+    "value" : "4096",
+    "description" : "PRIVATE CONFIG VARIABLE",
+    "filename" : "hdfs-site.xml"
+  }, {
+    "name" : "dfs.umaskmode",
+    "value" : "077",
+    "description" : "\nThe octal umask used when creating files and directories.\n",
+    "filename" : "hdfs-site.xml"
+  }, {
+    "name" : "dfs.web.ugi",
+    "value" : "gopher,gopher",
+    "description" : "The user account used by the web interface.\nSyntax: USERNAME,GROUP1,GROUP2, ...\n",
+    "filename" : "hdfs-site.xml"
+  }, {
+    "name" : "dfs.permissions",
+    "value" : "true",
+    "description" : "\nIf \"true\", enable permission checking in HDFS.\nIf \"false\", permission checking is turned off,\nbut all other behavior is unchanged.\nSwitching from one parameter value to the other does not change the mode,\nowner or group of files or directories.\n",
+    "filename" : "hdfs-site.xml"
+  }, {
+    "name" : "dfs.permissions.supergroup",
+    "value" : "hdfs",
+    "description" : "The name of the group of super-users.",
+    "filename" : "hdfs-site.xml"
+  }, {
+    "name" : "dfs.namenode.handler.count",
+    "value" : "100",
+    "description" : "Added to grow Queue size so that more client connections are allowed",
+    "filename" : "hdfs-site.xml"
+  }, {
+    "name" : "ipc.server.max.response.size",
+    "value" : "5242880",
+    "description" : null,
+    "filename" : "hdfs-site.xml"
+  }, {
+    "name" : "dfs.block.access.token.enable",
+    "value" : "true",
+    "description" : "\nIf \"true\", access tokens are used as capabilities for accessing datanodes.\nIf \"false\", no access tokens are checked on accessing datanodes.\n",
+    "filename" : "hdfs-site.xml"
+  }, {
+    "name" : "dfs.secondary.https.port",
+    "value" : "50490",
+    "description" : "The https port where secondary-namenode binds",
+    "filename" : "hdfs-site.xml"
+  }, {
+    "name" : "dfs.https.port",
+    "value" : "50470",
+    "description" : "The https port where namenode binds",
+    "filename" : "hdfs-site.xml"
+  }, {
+    "name" : "dfs.access.time.precision",
+    "value" : "0",
+    "description" : "The access time for HDFS file is precise upto this value.\n               The default value is 1 hour. Setting a value of 0 disables\n               access times for HDFS.\n  ",
+    "filename" : "hdfs-site.xml"
+  }, {
+    "name" : "dfs.cluster.administrators",
+    "value" : " hdfs",
+    "description" : "ACL for who all can view the default servlets in the HDFS",
+    "filename" : "hdfs-site.xml"
+  }, {
+    "name" : "ipc.server.read.threadpool.size",
+    "value" : "5",
+    "description" : null,
+    "filename" : "hdfs-site.xml"
+  }, {
+    "name" : "dfs.datanode.failed.volumes.tolerated",
+    "value" : "0",
+    "description" : "Number of failed disks datanode would tolerate",
+    "filename" : "hdfs-site.xml"
+  }, {
+    "name" : "io.file.buffer.size",
+    "value" : "131072",
+    "description" : "The size of buffer for use in sequence files.\n  The size of this buffer should probably be a multiple of hardware\n  page size (4096 on Intel x86), and it determines how much data is\n  buffered during read and write operations.",
+    "filename" : "core-site.xml"
+  }, {
+    "name" : "io.serializations",
+    "value" : "org.apache.hadoop.io.serializer.WritableSerialization",
+    "description" : null,
+    "filename" : "core-site.xml"
+  }, {
+    "name" : "io.compression.codec.lzo.class",
+    "value" : "com.hadoop.compression.lzo.LzoCodec",
+    "description" : "The implementation for lzo codec.",
+    "filename" : "core-site.xml"
+  }, {
+    "name" : "fs.trash.interval",
+    "value" : "360",
+    "description" : "Number of minutes between trash checkpoints.\n  If zero, the trash feature is disabled.\n  ",
+    "filename" : "core-site.xml"
+  }, {
+    "name" : "fs.checkpoint.edits.dir",
+    "value" : "${fs.checkpoint.dir}",
+    "description" : "Determines where on the local filesystem the DFS secondary\n        name node should store the temporary edits to merge.\n        If this is a comma-delimited list of directoires then teh edits is\n        replicated in all of the directoires for redundancy.\n        Default value is same as fs.checkpoint.dir\n    ",
+    "filename" : "core-site.xml"
+  }, {
+    "name" : "fs.checkpoint.period",
+    "value" : "21600",
+    "description" : "The number of seconds between two periodic checkpoints.\n  ",
+    "filename" : "core-site.xml"
+  }, {
+    "name" : "fs.checkpoint.size",
+    "value" : "536870912",
+    "description" : "The size of the current edit log (in bytes) that triggers\n       a periodic checkpoint even if the fs.checkpoint.period hasn't expired.\n  ",
+    "filename" : "core-site.xml"
+  }, {
+    "name" : "ipc.client.idlethreshold",
+    "value" : "8000",
+    "description" : "Defines the threshold number of connections after which\n               connections will be inspected for idleness.\n  ",
+    "filename" : "core-site.xml"
+  }, {
+    "name" : "ipc.client.connection.maxidletime",
+    "value" : "30000",
+    "description" : "The maximum time after which a client will bring down the\n               connection to the server.\n  ",
+    "filename" : "core-site.xml"
+  }, {
+    "name" : "ipc.client.connect.max.retries",
+    "value" : "50",
+    "description" : "Defines the maximum number of retries for IPC connections.",
+    "filename" : "core-site.xml"
+  }, {
+    "name" : "webinterface.private.actions",
+    "value" : "false",
+    "description" : " If set to true, the web interfaces of JT and NN may contain\n                actions, such as kill job, delete file, etc., that should\n                not be exposed to public. Enable this option if the interfaces\n                are only reachable by those who have the right authorization.\n  ",
+    "filename" : "core-site.xml"
+  } ],
+  "components" : [ {
+    "name" : "NAMENODE",
+    "category" : "MASTER",
+    "client" : false,
+    "master" : true
+  }, {
+    "name" : "DATANODE",
+    "category" : "SLAVE",
+    "client" : false,
+    "master" : false
+  }, {
+    "name" : "SECONDARY_NAMENODE",
+    "category" : "MASTER",
+    "client" : false,
+    "master" : true
+  }, {
+    "name" : "HDFS_CLIENT",
+    "category" : "CLIENT",
+    "client" : true,
+    "master" : false
+  } ],
+  "clientOnlyService" : false,
+  "clientComponent" : {
+    "name" : "HDFS_CLIENT",
+    "category" : "CLIENT",
+    "client" : true,
+    "master" : false
+  }
+}

Added: incubator/ambari/trunk/ambari-web/app/assets/data/wizard/stack/hdp/version130/HIVE.json
URL: http://svn.apache.org/viewvc/incubator/ambari/trunk/ambari-web/app/assets/data/wizard/stack/hdp/version130/HIVE.json?rev=1465166&view=auto
==============================================================================
--- incubator/ambari/trunk/ambari-web/app/assets/data/wizard/stack/hdp/version130/HIVE.json (added)
+++ incubator/ambari/trunk/ambari-web/app/assets/data/wizard/stack/hdp/version130/HIVE.json Sat Apr  6 01:55:42 2013
@@ -0,0 +1,95 @@
+{
+  "name" : "HIVE",
+  "version" : "0.10.0",
+  "user" : "root",
+  "comment" : "Data warehouse system for ad-hoc queries & analysis of large datasets and table & storage management service",
+  "properties" : [ {
+    "name" : "hive.metastore.local",
+    "value" : "false",
+    "description" : "controls whether to connect to remove metastore server or\n    open a new metastore server in Hive Client JVM",
+    "filename" : "hive-site.xml"
+  }, {
+    "name" : "javax.jdo.option.ConnectionDriverName",
+    "value" : "com.mysql.jdbc.Driver",
+    "description" : "Driver class name for a JDBC metastore",
+    "filename" : "hive-site.xml"
+  }, {
+    "name" : "hive.metastore.warehouse.dir",
+    "value" : "/apps/hive/warehouse",
+    "description" : "location of default database for the warehouse",
+    "filename" : "hive-site.xml"
+  }, {
+    "name" : "hive.metastore.cache.pinobjtypes",
+    "value" : "Table,Database,Type,FieldSchema,Order",
+    "description" : "List of comma separated metastore object types that should be pinned in the cache",
+    "filename" : "hive-site.xml"
+  }, {
+    "name" : "hive.semantic.analyzer.factory.impl",
+    "value" : "org.apache.hivealog.cli.HCatSemanticAnalyzerFactory",
+    "description" : "controls which SemanticAnalyzerFactory implemenation class is used by CLI",
+    "filename" : "hive-site.xml"
+  }, {
+    "name" : "hadoop.clientside.fs.operations",
+    "value" : "true",
+    "description" : "FS operations are owned by client",
+    "filename" : "hive-site.xml"
+  }, {
+    "name" : "hive.metastore.client.socket.timeout",
+    "value" : "60",
+    "description" : "MetaStore Client socket timeout in seconds",
+    "filename" : "hive-site.xml"
+  }, {
+    "name" : "hive.metastore.execute.setugi",
+    "value" : "true",
+    "description" : "In unsecure mode, setting this property to true will cause the metastore to execute DFS operations using the client's reported user and group permissions. Note that this property must be set on both the client and     server sides. Further note that its best effort. If client sets its to true and server sets it to false, client setting will be ignored.",
+    "filename" : "hive-site.xml"
+  }, {
+    "name" : "hive.security.authorization.enabled",
+    "value" : "true",
+    "description" : "enable or disable the hive client authorization",
+    "filename" : "hive-site.xml"
+  }, {
+    "name" : "hive.security.authorization.manager",
+    "value" : "org.apache.hcatalog.security.HdfsAuthorizationProvider",
+    "description" : "the hive client authorization manager class name.\n    The user defined authorization class should implement interface org.apache.hadoop.hive.ql.security.authorization.HiveAuthorizationProvider.  ",
+    "filename" : "hive-site.xml"
+  }, {
+    "name" : "hive.server2.enable.doAs",
+    "value" : "true",
+    "description" : null,
+    "filename" : "hive-site.xml"
+  }, {
+    "name" : "fs.hdfs.impl.disable.cache",
+    "value" : "true",
+    "description" : null,
+    "filename" : "hive-site.xml"
+  } ],
+  "components" : [ {
+    "name" : "HIVE_METASTORE",
+    "category" : "MASTER",
+    "client" : false,
+    "master" : true
+  }, {
+    "name" : "HIVE_SERVER",
+    "category" : "MASTER",
+    "client" : false,
+    "master" : true
+  }, {
+    "name" : "MYSQL_SERVER",
+    "category" : "MASTER",
+    "client" : false,
+    "master" : true
+  }, {
+    "name" : "HIVE_CLIENT",
+    "category" : "CLIENT",
+    "client" : true,
+    "master" : false
+  } ],
+  "clientOnlyService" : false,
+  "clientComponent" : {
+    "name" : "HIVE_CLIENT",
+    "category" : "CLIENT",
+    "client" : true,
+    "master" : false
+  }
+}

Added: incubator/ambari/trunk/ambari-web/app/assets/data/wizard/stack/hdp/version130/HUE.json
URL: http://svn.apache.org/viewvc/incubator/ambari/trunk/ambari-web/app/assets/data/wizard/stack/hdp/version130/HUE.json?rev=1465166&view=auto
==============================================================================
--- incubator/ambari/trunk/ambari-web/app/assets/data/wizard/stack/hdp/version130/HUE.json (added)
+++ incubator/ambari/trunk/ambari-web/app/assets/data/wizard/stack/hdp/version130/HUE.json Sat Apr  6 01:55:42 2013
@@ -0,0 +1,155 @@
+{
+  "name" : "HUE",
+  "version" : "2.2.0",
+  "user" : "root",
+  "comment" : "Hue is a graphical user interface to operate and develop\n      applications for Apache Hadoop.",
+  "properties" : [ {
+    "name" : "send_debug_messages",
+    "value" : "1",
+    "description" : null,
+    "filename" : "hue-site.xml"
+  }, {
+    "name" : "database_logging",
+    "value" : "0",
+    "description" : "To show database transactions, set database_logging to 1.\n      default, database_logging=0",
+    "filename" : "hue-site.xml"
+  }, {
+    "name" : "http_host",
+    "value" : "0.0.0.0",
+    "description" : "Webserver listens on this address and port",
+    "filename" : "hue-site.xml"
+  }, {
+    "name" : "http_port",
+    "value" : "8000",
+    "description" : "Webserver listens on this address and port",
+    "filename" : "hue-site.xml"
+  }, {
+    "name" : "time_zone",
+    "value" : "America/Los_Angeles",
+    "description" : "Time zone name",
+    "filename" : "hue-site.xml"
+  }, {
+    "name" : "django_debug_mode",
+    "value" : "1",
+    "description" : "Turn off debug",
+    "filename" : "hue-site.xml"
+  }, {
+    "name" : "use_cherrypy_server",
+    "value" : "false",
+    "description" : "Set to true to use CherryPy as the webserver, set to false\n      to use Spawning as the webserver. Defaults to Spawning if\n      key is not specified.",
+    "filename" : "hue-site.xml"
+  }, {
+    "name" : "http_500_debug_mode",
+    "value" : "1",
+    "description" : "Turn off backtrace for server error",
+    "filename" : "hue-site.xml"
+  }, {
+    "name" : "backend_auth_policy",
+    "value" : "desktop.auth.backend.AllowAllBackend",
+    "description" : "Authentication backend.",
+    "filename" : "hue-site.xml"
+  }, {
+    "name" : "db_engine",
+    "value" : "mysql",
+    "description" : "Configuration options for specifying the Desktop Database.",
+    "filename" : "hue-site.xml"
+  }, {
+    "name" : "db_host",
+    "value" : "localhost",
+    "description" : "Configuration options for specifying the Desktop Database.",
+    "filename" : "hue-site.xml"
+  }, {
+    "name" : "db_port",
+    "value" : "3306",
+    "description" : "Configuration options for specifying the Desktop Database.",
+    "filename" : "hue-site.xml"
+  }, {
+    "name" : "db_user",
+    "value" : "sandbox",
+    "description" : "Configuration options for specifying the Desktop Database.",
+    "filename" : "hue-site.xml"
+  }, {
+    "name" : "db_password",
+    "value" : "1111",
+    "description" : "Configuration options for specifying the Desktop Database.",
+    "filename" : "hue-site.xml"
+  }, {
+    "name" : "db_name",
+    "value" : "sandbox",
+    "description" : "Configuration options for specifying the Desktop Database.",
+    "filename" : "hue-site.xml"
+  }, {
+    "name" : "smtp_host",
+    "value" : "localhost",
+    "description" : "The SMTP server information for email notification delivery.",
+    "filename" : "hue-site.xml"
+  }, {
+    "name" : "smtp_port",
+    "value" : "25",
+    "description" : "The SMTP server information for email notification delivery.",
+    "filename" : "hue-site.xml"
+  }, {
+    "name" : "smtp_password",
+    "value" : "25",
+    "description" : "The SMTP server information for email notification delivery.",
+    "filename" : "hue-site.xml"
+  }, {
+    "name" : "tls",
+    "value" : "no",
+    "description" : "Whether to use a TLS (secure) connection when talking to the SMTP server.",
+    "filename" : "hue-site.xml"
+  }, {
+    "name" : "default_from_email",
+    "value" : "sandbox@hortonworks.com",
+    "description" : "The SMTP server information for email notification delivery.",
+    "filename" : "hue-site.xml"
+  }, {
+    "name" : "jobtracker_port",
+    "value" : "50030",
+    "description" : "The port where the JobTracker IPC listens on.",
+    "filename" : "hue-site.xml"
+  }, {
+    "name" : "hadoop_mapred_home",
+    "value" : "/usr/lib/hadoop/lib",
+    "description" : "The SMTP server information for email notification delivery.",
+    "filename" : "hue-site.xml"
+  }, {
+    "name" : "pig_shell_command",
+    "value" : "/usr/bin/pig -l /dev/null",
+    "description" : "Define and configure a new shell type pig.",
+    "filename" : "hue-site.xml"
+  }, {
+    "name" : "hbase_nice_name",
+    "value" : "HBase Shell",
+    "description" : "Define and configure a new shell type hbase",
+    "filename" : "hue-site.xml"
+  }, {
+    "name" : "hbase_shell_command",
+    "value" : "/usr/bin/hbase shell",
+    "description" : "Define and configure a new shell type hbase.",
+    "filename" : "hue-site.xml"
+  }, {
+    "name" : "bash_shell_command",
+    "value" : "/bin/bash",
+    "description" : "Define and configure a new shell type bash for testing only\n      .",
+    "filename" : "hue-site.xml"
+  }, {
+    "name" : "whitelist",
+    "value" : "(localhost|127\\.0\\.0\\.1):(50030|50070|50060|50075|50111)",
+    "description" : "proxy settings",
+    "filename" : "hue-site.xml"
+  } ],
+  "components" : [ {
+    "name" : "HUE_SERVER",
+    "category" : "MASTER",
+    "client" : false,
+    "master" : true
+  } ],
+  "clientOnlyService" : false,
+  "clientComponent" : {
+    "name" : "HUE_SERVER",
+    "category" : "MASTER",
+    "client" : false,
+    "master" : true
+  }
+}

Added: incubator/ambari/trunk/ambari-web/app/assets/data/wizard/stack/hdp/version130/MAPREDUCE.json
URL: http://svn.apache.org/viewvc/incubator/ambari/trunk/ambari-web/app/assets/data/wizard/stack/hdp/version130/MAPREDUCE.json?rev=1465166&view=auto
==============================================================================
--- incubator/ambari/trunk/ambari-web/app/assets/data/wizard/stack/hdp/version130/MAPREDUCE.json (added)
+++ incubator/ambari/trunk/ambari-web/app/assets/data/wizard/stack/hdp/version130/MAPREDUCE.json Sat Apr  6 01:55:42 2013
@@ -0,0 +1,230 @@
+{
+  "name" : "MAPREDUCE",
+  "version" : "1.1.2",
+  "user" : "mapred",
+  "comment" : "Apache Hadoop Distributed Processing Framework",
+  "properties" : [ {
+    "name" : "io.sort.record.percent",
+    "value" : ".2",
+    "description" : "No description",
+    "filename" : "mapred-site.xml"
+  }, {
+    "name" : "io.sort.factor",
+    "value" : "100",
+    "description" : "No description",
+    "filename" : "mapred-site.xml"
+  }, {
+    "name" : "mapred.tasktracker.tasks.sleeptime-before-sigkill",
+    "value" : "250",
+    "description" : "Normally, this is the amount of time before killing\n  processes, and the recommended-default is 5.000 seconds - a value of\n  5000 here.  In this case, we are using it solely to blast tasks before\n  killing them, and killing them very quickly (1/4 second) to guarantee\n  that we do not leave VMs around for later jobs.\n  ",
+    "filename" : "mapred-site.xml"
+  }, {
+    "name" : "mapred.job.tracker.handler.count",
+    "value" : "50",
+    "description" : "\n    The number of server threads for the JobTracker. This should be roughly\n    4% of the number of tasktracker nodes.\n    ",
+    "filename" : "mapred-site.xml"
+  }, {
+    "name" : "mapred.system.dir",
+    "value" : "/mapred/system",
+    "description" : "No description",
+    "filename" : "mapred-site.xml"
+  }, {
+    "name" : "mapreduce.cluster.administrators",
+    "value" : " hadoop",
+    "description" : null,
+    "filename" : "mapred-site.xml"
+  }, {
+    "name" : "mapred.reduce.parallel.copies",
+    "value" : "30",
+    "description" : "No description",
+    "filename" : "mapred-site.xml"
+  }, {
+    "name" : "tasktracker.http.threads",
+    "value" : "50",
+    "description" : null,
+    "filename" : "mapred-site.xml"
+  }, {
+    "name" : "mapred.map.tasks.speculative.execution",
+    "value" : "false",
+    "description" : "If true, then multiple instances of some map tasks\n               may be executed in parallel.",
+    "filename" : "mapred-site.xml"
+  }, {
+    "name" : "mapred.reduce.tasks.speculative.execution",
+    "value" : "false",
+    "description" : "If true, then multiple instances of some reduce tasks\n               may be executed in parallel.",
+    "filename" : "mapred-site.xml"
+  }, {
+    "name" : "mapred.reduce.slowstart.completed.maps",
+    "value" : "0.05",
+    "description" : null,
+    "filename" : "mapred-site.xml"
+  }, {
+    "name" : "mapred.inmem.merge.threshold",
+    "value" : "1000",
+    "description" : "The threshold, in terms of the number of files\n  for the in-memory merge process. When we accumulate threshold number of files\n  we initiate the in-memory merge and spill to disk. A value of 0 or less than\n  0 indicates we want to DON'T have any threshold and instead depend only on\n  the ramfs's memory consumption to trigger the merge.\n  ",
+    "filename" : "mapred-site.xml"
+  }, {
+    "name" : "mapred.job.shuffle.merge.percent",
+    "value" : "0.66",
+    "description" : "The usage threshold at which an in-memory merge will be\n  initiated, expressed as a percentage of the total memory allocated to\n  storing in-memory map outputs, as defined by\n  mapred.job.shuffle.input.buffer.percent.\n  ",
+    "filename" : "mapred-site.xml"
+  }, {
+    "name" : "mapred.job.shuffle.input.buffer.percent",
+    "value" : "0.7",
+    "description" : "The percentage of memory to be allocated from the maximum heap\n  size to storing map outputs during the shuffle.\n  ",
+    "filename" : "mapred-site.xml"
+  }, {
+    "name" : "mapred.output.compression.type",
+    "value" : "BLOCK",
+    "description" : "If the job outputs are to compressed as SequenceFiles, how should\n               they be compressed? Should be one of NONE, RECORD or BLOCK.\n  ",
+    "filename" : "mapred-site.xml"
+  }, {
+    "name" : "mapred.jobtracker.completeuserjobs.maximum",
+    "value" : "0",
+    "description" : null,
+    "filename" : "mapred-site.xml"
+  }, {
+    "name" : "mapred.jobtracker.restart.recover",
+    "value" : "false",
+    "description" : "\"true\" to enable (job) recovery upon restart,\n               \"false\" to start afresh\n    ",
+    "filename" : "mapred-site.xml"
+  }, {
+    "name" : "mapred.job.reduce.input.buffer.percent",
+    "value" : "0.0",
+    "description" : "The percentage of memory- relative to the maximum heap size- to\n  retain map outputs during the reduce. When the shuffle is concluded, any\n  remaining map outputs in memory must consume less than this threshold before\n  the reduce can begin.\n  ",
+    "filename" : "mapred-site.xml"
+  }, {
+    "name" : "mapreduce.reduce.input.limit",
+    "value" : "10737418240",
+    "description" : "The limit on the input size of the reduce. (This value\n  is 10 Gb.)  If the estimated input size of the reduce is greater than\n  this value, job is failed. A value of -1 means that there is no limit\n  set. ",
+    "filename" : "mapred-site.xml"
+  }, {
+    "name" : "mapred.task.timeout",
+    "value" : "600000",
+    "description" : "The number of milliseconds before a task will be\n  terminated if it neither reads an input, writes an output, nor\n  updates its status string.\n  ",
+    "filename" : "mapred-site.xml"
+  }, {
+    "name" : "jetty.connector",
+    "value" : "org.mortbay.jetty.nio.SelectChannelConnector",
+    "description" : "No description",
+    "filename" : "mapred-site.xml"
+  }, {
+    "name" : "mapred.child.root.logger",
+    "value" : "INFO,TLA",
+    "description" : null,
+    "filename" : "mapred-site.xml"
+  }, {
+    "name" : "mapred.max.tracker.blacklists",
+    "value" : "16",
+    "description" : "\n    if node is reported blacklisted by 16 successful jobs within timeout-window, it will be graylisted\n  ",
+    "filename" : "mapred-site.xml"
+  }, {
+    "name" : "mapred.healthChecker.interval",
+    "value" : "135000",
+    "description" : null,
+    "filename" : "mapred-site.xml"
+  }, {
+    "name" : "mapred.healthChecker.script.timeout",
+    "value" : "60000",
+    "description" : null,
+    "filename" : "mapred-site.xml"
+  }, {
+    "name" : "mapred.job.tracker.persist.jobstatus.active",
+    "value" : "false",
+    "description" : "Indicates if persistency of job status information is\n  active or not.\n  ",
+    "filename" : "mapred-site.xml"
+  }, {
+    "name" : "mapred.job.tracker.persist.jobstatus.hours",
+    "value" : "1",
+    "description" : "The number of hours job status information is persisted in DFS.\n    The job status information will be available after it drops of the memory\n    queue and between jobtracker restarts. With a zero value the job status\n    information is not persisted at all in DFS.\n  ",
+    "filename" : "mapred-site.xml"
+  }, {
+    "name" : "mapred.jobtracker.retirejob.check",
+    "value" : "10000",
+    "description" : null,
+    "filename" : "mapred-site.xml"
+  }, {
+    "name" : "mapred.jobtracker.retirejob.interval",
+    "value" : "0",
+    "description" : null,
+    "filename" : "mapred-site.xml"
+  }, {
+    "name" : "mapred.job.tracker.history.completed.location",
+    "value" : "/mapred/history/done",
+    "description" : "No description",
+    "filename" : "mapred-site.xml"
+  }, {
+    "name" : "mapreduce.fileoutputcommitter.marksuccessfuljobs",
+    "value" : "false",
+    "description" : null,
+    "filename" : "mapred-site.xml"
+  }, {
+    "name" : "mapred.job.reuse.jvm.num.tasks",
+    "value" : "1",
+    "description" : "\n    How many tasks to run per jvm. If set to -1, there is no limit\n  ",
+    "filename" : "mapred-site.xml"
+  }, {
+    "name" : "hadoop.job.history.user.location",
+    "value" : "none",
+    "description" : null,
+    "filename" : "mapred-site.xml"
+  }, {
+    "name" : "mapreduce.jobtracker.staging.root.dir",
+    "value" : "/user",
+    "description" : "The Path prefix for where the staging directories should be placed. The next level is always the user's\n   name. It is a path in the default file system.",
+    "filename" : "mapred-site.xml"
+  }, {
+    "name" : "mapreduce.tasktracker.group",
+    "value" : "hadoop",
+    "description" : "The group that the task controller uses for accessing the task controller. The mapred user must be a member and users should *not* be members.",
+    "filename" : "mapred-site.xml"
+  }, {
+    "name" : "mapreduce.jobtracker.split.metainfo.maxsize",
+    "value" : "50000000",
+    "description" : "If the size of the split metainfo file is larger than this, the JobTracker will fail the job during\n    initialize.\n   ",
+    "filename" : "mapred-site.xml"
+  }, {
+    "name" : "mapreduce.history.server.embedded",
+    "value" : "false",
+    "description" : "Should job history server be embedded within Job tracker\nprocess",
+    "filename" : "mapred-site.xml"
+  }, {
+    "name" : "mapred.jobtracker.blacklist.fault-timeout-window",
+    "value" : "180",
+    "description" : "\n    3-hour sliding window (value is in minutes)\n  ",
+    "filename" : "mapred-site.xml"
+  }, {
+    "name" : "mapred.jobtracker.blacklist.fault-bucket-width",
+    "value" : "15",
+    "description" : "\n    15-minute bucket size (value is in minutes)\n  ",
+    "filename" : "mapred-site.xml"
+  }, {
+    "name" : "mapred.queue.names",
+    "value" : "default",
+    "description" : " Comma separated list of queues configured for this jobtracker.",
+    "filename" : "mapred-site.xml"
+  } ],
+  "components" : [ {
+    "name" : "JOBTRACKER",
+    "category" : "MASTER",
+    "client" : false,
+    "master" : true
+  }, {
+    "name" : "TASKTRACKER",
+    "category" : "SLAVE",
+    "client" : false,
+    "master" : false
+  }, {
+    "name" : "MAPREDUCE_CLIENT",
+    "category" : "CLIENT",
+    "client" : true,
+    "master" : false
+  } ],
+  "clientOnlyService" : false,
+  "clientComponent" : {
+    "name" : "MAPREDUCE_CLIENT",
+    "category" : "CLIENT",
+    "client" : true,
+    "master" : false
+  }
+}

Added: incubator/ambari/trunk/ambari-web/app/assets/data/wizard/stack/hdp/version130/OOZIE.json
URL: http://svn.apache.org/viewvc/incubator/ambari/trunk/ambari-web/app/assets/data/wizard/stack/hdp/version130/OOZIE.json?rev=1465166&view=auto
==============================================================================
--- incubator/ambari/trunk/ambari-web/app/assets/data/wizard/stack/hdp/version130/OOZIE.json (added)
+++ incubator/ambari/trunk/ambari-web/app/assets/data/wizard/stack/hdp/version130/OOZIE.json Sat Apr  6 01:55:42 2013
@@ -0,0 +1,155 @@
+{
+  "name" : "OOZIE",
+  "version" : "3.2.0",
+  "user" : "root",
+  "comment" : "System for workflow coordination and execution of Apache Hadoop jobs",
+  "properties" : [ {
+    "name" : "oozie.base.url",
+    "value" : "http://localhost:11000/oozie",
+    "description" : "Base Oozie URL.",
+    "filename" : "oozie-site.xml"
+  }, {
+    "name" : "oozie.system.id",
+    "value" : "oozie-${user.name}",
+    "description" : "\n    The Oozie system ID.\n    ",
+    "filename" : "oozie-site.xml"
+  }, {
+    "name" : "oozie.systemmode",
+    "value" : "NORMAL",
+    "description" : "\n     System mode for  Oozie at startup.\n     ",
+    "filename" : "oozie-site.xml"
+  }, {
+    "name" : "oozie.service.AuthorizationService.security.enabled",
+    "value" : "true",
+    "description" : "\n     Specifies whether security (user name/admin role) is enabled or not.\n     If disabled any user can manage Oozie system and manage any job.\n     ",
+    "filename" : "oozie-site.xml"
+  }, {
+    "name" : "oozie.service.PurgeService.older.than",
+    "value" : "30",
+    "description" : "\n     Jobs older than this value, in days, will be purged by the PurgeService.\n     ",
+    "filename" : "oozie-site.xml"
+  }, {
+    "name" : "oozie.service.PurgeService.purge.interval",
+    "value" : "3600",
+    "description" : "\n     Interval at which the purge service will run, in seconds.\n     ",
+    "filename" : "oozie-site.xml"
+  }, {
+    "name" : "oozie.service.CallableQueueService.queue.size",
+    "value" : "1000",
+    "description" : "Max callable queue size",
+    "filename" : "oozie-site.xml"
+  }, {
+    "name" : "oozie.service.CallableQueueService.threads",
+    "value" : "10",
+    "description" : "Number of threads used for executing callables",
+    "filename" : "oozie-site.xml"
+  }, {
+    "name" : "oozie.service.CallableQueueService.callable.concurrency",
+    "value" : "3",
+    "description" : "\n     Maximum concurrency for a given callable type.\n     Each command is a callable type (submit, start, run, signal, job, jobs, suspend,resume, etc).\n     Each action type is a callable type (Map-Reduce, Pig, SSH, FS, sub-workflow, etc).\n     All commands that use action executors (action-start, action-end, action-kill and action-check) use\n     the action type as the callable type.\n     ",
+    "filename" : "oozie-site.xml"
+  }, {
+    "name" : "oozie.service.coord.normal.default.timeout",
+    "value" : "120",
+    "description" : "Default timeout for a coordinator action input check (in minutes) for normal job.\n      -1 means infinite timeout",
+    "filename" : "oozie-site.xml"
+  }, {
+    "name" : "oozie.db.schema.name",
+    "value" : "oozie",
+    "description" : "\n      Oozie DataBase Name\n     ",
+    "filename" : "oozie-site.xml"
+  }, {
+    "name" : "oozie.service.HadoopAccessorService.jobTracker.whitelist",
+    "value" : " ",
+    "description" : "\n      Whitelisted job tracker for Oozie service.\n      ",
+    "filename" : "oozie-site.xml"
+  }, {
+    "name" : "oozie.authentication.type",
+    "value" : "simple",
+    "description" : "\n      ",
+    "filename" : "oozie-site.xml"
+  }, {
+    "name" : "oozie.service.HadoopAccessorService.nameNode.whitelist",
+    "value" : " ",
+    "description" : "\n      ",
+    "filename" : "oozie-site.xml"
+  }, {
+    "name" : "oozie.service.WorkflowAppService.system.libpath",
+    "value" : "/user/${user.name}/share/lib",
+    "description" : "\n      System library path to use for workflow applications.\n      This path is added to workflow application if their job properties sets\n      the property 'oozie.use.system.libpath' to true.\n      ",
+    "filename" : "oozie-site.xml"
+  }, {
+    "name" : "use.system.libpath.for.mapreduce.and.pig.jobs",
+    "value" : "false",
+    "description" : "\n      If set to true, submissions of MapReduce and Pig jobs will include\n      automatically the system library path, thus not requiring users to\n      specify where the Pig JAR files are. Instead, the ones from the system\n      library path are used.\n      ",
+    "filename" : "oozie-site.xml"
+  }, {
+    "name" : "oozie.authentication.kerberos.name.rules",
+    "value" : "\n        RULE:[2:$1@$0]([jt]t@.*TODO-KERBEROS-DOMAIN)s/.*/TODO-MAPREDUSER/\n        RULE:[2:$1@$0]([nd]n@.*TODO-KERBEROS-DOMAIN)s/.*/TODO-HDFSUSER/\n        RULE:[2:$1@$0](hm@.*TODO-KERBEROS-DOMAIN)s/.*/TODO-HBASE-USER/\n        RULE:[2:$1@$0](rs@.*TODO-KERBEROS-DOMAIN)s/.*/TODO-HBASE-USER/\n        DEFAULT\n        ",
+    "description" : "The mapping from kerberos principal names to local OS user names.",
+    "filename" : "oozie-site.xml"
+  }, {
+    "name" : "oozie.service.HadoopAccessorService.hadoop.configurations",
+    "value" : "*=/etc/hadoop/conf",
+    "description" : "\n          Comma separated AUTHORITY=HADOOP_CONF_DIR, where AUTHORITY is the HOST:PORT of\n          the Hadoop service (JobTracker, HDFS). The wildcard '*' configuration is\n          used when there is no exact match for an authority. The HADOOP_CONF_DIR contains\n          the relevant Hadoop *-site.xml files. If the path is relative is looked within\n          the Oozie configuration directory; though the path can be absolute (i.e. to point\n          to Hadoop client conf/ directories in the local filesystem.\n      ",
+    "filename" : "oozie-site.xml"
+  }, {
+    "name" : "oozie.service.ActionService.executor.ext.classes",
+    "value" : "\n            org.apache.oozie.action.email.EmailActionExecutor,\n            org.apache.oozie.action.hadoop.HiveActionExecutor,\n            org.apache.oozie.action.hadoop.ShellActionExecutor,\n            org.apache.oozie.action.hadoop.SqoopActionExecutor,\n            org.apache.oozie.action.hadoop.DistcpActionExecutor\n        ",
+    "description" : null,
+    "filename" : "oozie-site.xml"
+  }, {
+    "name" : "oozie.service.SchemaService.wf.ext.schemas",
+    "value" : "shell-action-0.1.xsd,email-action-0.1.xsd,hive-action-0.2.xsd,sqoop-action-0.2.xsd,ssh-action-0.1.xsd,distcp-action-0.1.xsd",
+    "description" : null,
+    "filename" : "oozie-site.xml"
+  }, {
+    "name" : "oozie.service.JPAService.create.db.schema",
+    "value" : "false",
+    "description" : "\n            Creates Oozie DB.\n\n            If set to true, it creates the DB schema if it does not exist. If the DB schema exists is a NOP.\n            If set to false, it does not create the DB schema. If the DB schema does not exist it fails start up.\n        ",
+    "filename" : "oozie-site.xml"
+  }, {
+    "name" : "oozie.service.JPAService.jdbc.driver",
+    "value" : "org.apache.derby.jdbc.EmbeddedDriver",
+    "description" : "\n            JDBC driver class.\n        ",
+    "filename" : "oozie-site.xml"
+  }, {
+    "name" : "oozie.service.JPAService.jdbc.url",
+    "value" : "jdbc:derby:${oozie.data.dir}/${oozie.db.schema.name}-db;create=true",
+    "description" : "\n            JDBC URL.\n        ",
+    "filename" : "oozie-site.xml"
+  }, {
+    "name" : "oozie.service.JPAService.jdbc.username",
+    "value" : "sa",
+    "description" : "\n            DB user name.\n        ",
+    "filename" : "oozie-site.xml"
+  }, {
+    "name" : "oozie.service.JPAService.jdbc.password",
+    "value" : " ",
+    "description" : "\n            DB user password.\n\n            IMPORTANT: if password is emtpy leave a 1 space string, the service trims the value,\n                       if empty Configuration assumes it is NULL.\n        ",
+    "filename" : "oozie-site.xml"
+  }, {
+    "name" : "oozie.service.JPAService.pool.max.active.conn",
+    "value" : "10",
+    "description" : "\n             Max number of connections.\n        ",
+    "filename" : "oozie-site.xml"
+  } ],
+  "components" : [ {
+    "name" : "OOZIE_SERVER",
+    "category" : "MASTER",
+    "client" : false,
+    "master" : true
+  }, {
+    "name" : "OOZIE_CLIENT",
+    "category" : "CLIENT",
+    "client" : true,
+    "master" : false
+  } ],
+  "clientOnlyService" : false,
+  "clientComponent" : {
+    "name" : "OOZIE_CLIENT",
+    "category" : "CLIENT",
+    "client" : true,
+    "master" : false
+  }
+}

Added: incubator/ambari/trunk/ambari-web/app/assets/data/wizard/stack/hdp/version130/WEBHCAT.json
URL: http://svn.apache.org/viewvc/incubator/ambari/trunk/ambari-web/app/assets/data/wizard/stack/hdp/version130/WEBHCAT.json?rev=1465166&view=auto
==============================================================================
--- incubator/ambari/trunk/ambari-web/app/assets/data/wizard/stack/hdp/version130/WEBHCAT.json (added)
+++ incubator/ambari/trunk/ambari-web/app/assets/data/wizard/stack/hdp/version130/WEBHCAT.json Sat Apr  6 01:55:42 2013
@@ -0,0 +1,90 @@
+{
+  "name" : "WEBHCAT",
+  "version" : "0.5.0",
+  "user" : "root",
+  "comment" : "This is comment for WEBHCAT service",
+  "properties" : [ {
+    "name" : "templeton.port",
+    "value" : "50111",
+    "description" : "The HTTP port for the main server.",
+    "filename" : "webhcat-site.xml"
+  }, {
+    "name" : "templeton.hadoop.conf.dir",
+    "value" : "/etc/hadoop/conf",
+    "description" : "The path to the Hadoop configuration.",
+    "filename" : "webhcat-site.xml"
+  }, {
+    "name" : "templeton.jar",
+    "value" : "/usr/lib/hcatalog/share/webhcat/svr/webhcat.jar",
+    "description" : "The path to the Templeton jar file.",
+    "filename" : "webhcat-site.xml"
+  }, {
+    "name" : "templeton.libjars",
+    "value" : "/usr/lib/zookeeper/zookeeper.jar",
+    "description" : "Jars to add the the classpath.",
+    "filename" : "webhcat-site.xml"
+  }, {
+    "name" : "templeton.hadoop",
+    "value" : "/usr/bin/hadoop",
+    "description" : "The path to the Hadoop executable.",
+    "filename" : "webhcat-site.xml"
+  }, {
+    "name" : "templeton.pig.archive",
+    "value" : "hdfs:///apps/webhcat/pig.tar.gz",
+    "description" : "The path to the Pig archive.",
+    "filename" : "webhcat-site.xml"
+  }, {
+    "name" : "templeton.pig.path",
+    "value" : "pig.tar.gz/pig/bin/pig",
+    "description" : "The path to the Pig executable.",
+    "filename" : "webhcat-site.xml"
+  }, {
+    "name" : "templeton.hcat",
+    "value" : "/usr/bin/hcat",
+    "description" : "The path to the hcatalog executable.",
+    "filename" : "webhcat-site.xml"
+  }, {
+    "name" : "templeton.hive.archive",
+    "value" : "hdfs:///apps/webhcat/hive.tar.gz",
+    "description" : "The path to the Hive archive.",
+    "filename" : "webhcat-site.xml"
+  }, {
+    "name" : "templeton.hive.path",
+    "value" : "hive.tar.gz/hive/bin/hive",
+    "description" : "The path to the Hive executable.",
+    "filename" : "webhcat-site.xml"
+  }, {
+    "name" : "templeton.storage.class",
+    "value" : "org.apache.hcatalog.templeton.tool.ZooKeeperStorage",
+    "description" : "The class to use as storage",
+    "filename" : "webhcat-site.xml"
+  }, {
+    "name" : "templeton.override.enabled",
+    "value" : "false",
+    "description" : "\n     Enable the override path in templeton.override.jars\n   ",
+    "filename" : "webhcat-site.xml"
+  }, {
+    "name" : "templeton.streaming.jar",
+    "value" : "hdfs:///apps/webhcat/hadoop-streaming.jar",
+    "description" : "The hdfs path to the Hadoop streaming jar file.",
+    "filename" : "webhcat-site.xml"
+  }, {
+    "name" : "templeton.exec.timeout",
+    "value" : "60000",
+    "description" : "Time out for templeton api",
+    "filename" : "webhcat-site.xml"
+  } ],
+  "components" : [ {
+    "name" : "WEBHCAT_SERVER",
+    "category" : "MASTER",
+    "client" : false,
+    "master" : true
+  } ],
+  "clientOnlyService" : false,
+  "clientComponent" : {
+    "name" : "WEBHCAT_SERVER",
+    "category" : "MASTER",
+    "client" : false,
+    "master" : true
+  }
+}

Added: incubator/ambari/trunk/ambari-web/app/assets/data/wizard/stack/hdp/version130/ZOOKEEPER.json
URL: http://svn.apache.org/viewvc/incubator/ambari/trunk/ambari-web/app/assets/data/wizard/stack/hdp/version130/ZOOKEEPER.json?rev=1465166&view=auto
==============================================================================
--- incubator/ambari/trunk/ambari-web/app/assets/data/wizard/stack/hdp/version130/ZOOKEEPER.json (added)
+++ incubator/ambari/trunk/ambari-web/app/assets/data/wizard/stack/hdp/version130/ZOOKEEPER.json Sat Apr  6 01:55:42 2013
@@ -0,0 +1,25 @@
+{
+  "name" : "ZOOKEEPER",
+  "version" : "3.4.5",
+  "user" : "root",
+  "comment" : "This is comment for ZOOKEEPER service",
+  "properties" : [ ],
+  "components" : [ {
+    "name" : "ZOOKEEPER_SERVER",
+    "category" : "MASTER",
+    "client" : false,
+    "master" : true
+  }, {
+    "name" : "ZOOKEEPER_CLIENT",
+    "category" : "CLIENT",
+    "client" : true,
+    "master" : false
+  } ],
+  "clientOnlyService" : false,
+  "clientComponent" : {
+    "name" : "ZOOKEEPER_CLIENT",
+    "category" : "CLIENT",
+    "client" : true,
+    "master" : false
+  }
+}