diff options
author | evans_ye <innovansye@gmail.com> | 2014-10-30 16:49:15 +0800 |
---|---|---|
committer | jay@apache.org <jayunit100> | 2014-10-30 19:52:11 -0400 |
commit | 11c8bd5543cd8f21d33f92447c5b996a5886d798 (patch) | |
tree | abe5d31977cd665a8cc9a5993bc9c10ea22d9eb3 /bigtop-deploy | |
parent | e722c9ced7ce4f89b43623f1ff04a4c1a7b75369 (diff) |
BIGTOP-1047. Support Puppet 3.x
Signed-off-by: Konstantin Boudnik <cos@apache.org>
Diffstat (limited to 'bigtop-deploy')
25 files changed, 178 insertions, 178 deletions
diff --git a/bigtop-deploy/puppet/modules/giraph/templates/giraph-site.xml b/bigtop-deploy/puppet/modules/giraph/templates/giraph-site.xml index 9500fa97..5d5228f2 100755 --- a/bigtop-deploy/puppet/modules/giraph/templates/giraph-site.xml +++ b/bigtop-deploy/puppet/modules/giraph/templates/giraph-site.xml @@ -29,7 +29,7 @@ under the License. <property> <name>giraph.zkList</name> - <value><%= zookeeper_quorum %></value> + <value><%= @zookeeper_quorum %></value> <description>Prefer external Zookeeper over the embedded one</description> </property> </configuration> diff --git a/bigtop-deploy/puppet/modules/hadoop-flume/templates/flume.conf b/bigtop-deploy/puppet/modules/hadoop-flume/templates/flume.conf index 7007ead3..7d1b1a23 100644 --- a/bigtop-deploy/puppet/modules/hadoop-flume/templates/flume.conf +++ b/bigtop-deploy/puppet/modules/hadoop-flume/templates/flume.conf @@ -13,23 +13,23 @@ # See the License for the specific language governing permissions and # limitations under the License. -agent.channels = <%= (channels.collect { |k,v| k }).join(' ') %> -agent.sources = <%= (sources.collect { |k,v| k }).join(' ') %> -agent.sinks = <%= (sinks.collect { |k,v| k }).join(' ') %> +agent.channels = <%= (@channels.collect { |k,v| k }).join(' ') %> +agent.sources = <%= (@sources.collect { |k,v| k }).join(' ') %> +agent.sinks = <%= (@sinks.collect { |k,v| k }).join(' ') %> -<% channels.each do |channel,props| -%> +<% @channels.each do |channel,props| -%> <% props.each do |k,v| -%> agent.channels.<%= channel %>.<%= k %> = <%= v %> <% end -%> <% end -%> -<% sources.each do |source,props| -%> +<% @sources.each do |source,props| -%> <% props.each do |k,v| -%> agent.sources.<%= source %>.<%= k %> = <%= v %> <% end -%> <% end -%> -<% sinks.each do |sink,props| -%> +<% @sinks.each do |sink,props| -%> <% props.each do |k,v| -%> agent.sinks.<%= sink %>.<%= k %> = <%= v %> <% end -%> diff --git a/bigtop-deploy/puppet/modules/hadoop-hbase/templates/hbase-site.xml b/bigtop-deploy/puppet/modules/hadoop-hbase/templates/hbase-site.xml index 8489a93f..849676c8 100644 --- a/bigtop-deploy/puppet/modules/hadoop-hbase/templates/hbase-site.xml +++ b/bigtop-deploy/puppet/modules/hadoop-hbase/templates/hbase-site.xml @@ -42,7 +42,7 @@ <value>true</value> </property> -<% if has_variable?("hdfs_shortcut_reader_user") %> +<% if @hdfs_shortcut_reader_user %> <property> <name>dfs.client.read.shortcircuit</name> <value>true</value> diff --git a/bigtop-deploy/puppet/modules/hadoop-hbase/templates/jaas.conf b/bigtop-deploy/puppet/modules/hadoop-hbase/templates/jaas.conf index 8c546722..5b6b2a58 100644 --- a/bigtop-deploy/puppet/modules/hadoop-hbase/templates/jaas.conf +++ b/bigtop-deploy/puppet/modules/hadoop-hbase/templates/jaas.conf @@ -20,5 +20,5 @@ Client { useKeyTab=true useTicketCache=false keyTab="/etc/hbase.keytab" - principal="hbase/<%= fqdn %>@<%= kerberos_realm %>"; + principal="hbase/<%= @fqdn %>@<%= @kerberos_realm %>"; }; diff --git a/bigtop-deploy/puppet/modules/hadoop-hive/templates/hive-site.xml b/bigtop-deploy/puppet/modules/hadoop-hive/templates/hive-site.xml index 4f8cf897..5861d971 100644 --- a/bigtop-deploy/puppet/modules/hadoop-hive/templates/hive-site.xml +++ b/bigtop-deploy/puppet/modules/hadoop-hive/templates/hive-site.xml @@ -26,18 +26,18 @@ <!-- Hive Execution Parameters --> -<% if hbase_master != "" %> +<% if @hbase_master != "" %> <property> <name>hbase.master</name> - <value><%= hbase_master %></value> + <value><%= @hbase_master %></value> <description>http://wiki.apache.org/hadoop/Hive/HBaseIntegration</description> </property> <% end %> -<% if hbase_zookeeper_quorum != "" %> +<% if @hbase_zookeeper_quorum != "" %> <property> <name>hbase.zookeeper.quorum</name> - <value><%= hbase_zookeeper_quorum %></value> + <value><%= @hbase_zookeeper_quorum %></value> <description>http://wiki.apache.org/hadoop/Hive/HBaseIntegration</description> </property> <% end %> diff --git a/bigtop-deploy/puppet/modules/hadoop-oozie/templates/oozie-site.xml b/bigtop-deploy/puppet/modules/hadoop-oozie/templates/oozie-site.xml index ef39045b..93d5eafc 100644 --- a/bigtop-deploy/puppet/modules/hadoop-oozie/templates/oozie-site.xml +++ b/bigtop-deploy/puppet/modules/hadoop-oozie/templates/oozie-site.xml @@ -176,7 +176,7 @@ <property> <name>oozie.service.HadoopAccessorService.kerberos.enabled</name> - <value><%= if (kerberos_realm != "") ; "true" else "false" end %></value> + <value><%= if (@kerberos_realm != "") ; "true" else "false" end %></value> <description> Indicates if Oozie is configured to use Kerberos. </description> @@ -184,7 +184,7 @@ <property> <name>local.realm</name> - <value><%= kerberos_realm %></value> + <value><%= @kerberos_realm %></value> <description> Kerberos Realm used by Oozie and Hadoop. Using 'local.realm' to be aligned with Hadoop configuration </description> @@ -200,7 +200,7 @@ <property> <name>oozie.service.HadoopAccessorService.kerberos.principal</name> - <value>${user.name}/<%= fqdn %>@${local.realm}</value> + <value>${user.name}/<%= @fqdn %>@${local.realm}</value> <description> Kerberos principal for Oozie service. </description> @@ -258,7 +258,7 @@ <property> <name>oozie.authentication.type</name> - <value><%= kerberos_realm.empty? ? "simple" : "kerberos" %></value> + <value><%= @kerberos_realm.empty? ? "simple" : "kerberos" %></value> <description> Defines authentication used for Oozie HTTP endpoint. Supported values are: simple | kerberos | #AUTHENTICATION_HANDLER_CLASSNAME# @@ -306,7 +306,7 @@ <property> <name>oozie.authentication.kerberos.principal</name> - <value>HTTP/<%= fqdn %>@${local.realm}</value> + <value>HTTP/<%= @fqdn %>@${local.realm}</value> <description> Indicates the Kerberos principal to be used for HTTP endpoint. The principal MUST start with 'HTTP/' as per Kerberos HTTP SPNEGO specification. diff --git a/bigtop-deploy/puppet/modules/hadoop-zookeeper/manifests/init.pp b/bigtop-deploy/puppet/modules/hadoop-zookeeper/manifests/init.pp index 8e0c7576..701590e1 100644 --- a/bigtop-deploy/puppet/modules/hadoop-zookeeper/manifests/init.pp +++ b/bigtop-deploy/puppet/modules/hadoop-zookeeper/manifests/init.pp @@ -45,7 +45,7 @@ class hadoop-zookeeper { } file { "/var/lib/zookeeper/myid": - content => inline_template("<%= myid %>"), + content => inline_template("<%= @myid %>"), require => Package["zookeeper-server"], } diff --git a/bigtop-deploy/puppet/modules/hadoop-zookeeper/templates/jaas.conf b/bigtop-deploy/puppet/modules/hadoop-zookeeper/templates/jaas.conf index 4655502e..61b2eac4 100644 --- a/bigtop-deploy/puppet/modules/hadoop-zookeeper/templates/jaas.conf +++ b/bigtop-deploy/puppet/modules/hadoop-zookeeper/templates/jaas.conf @@ -21,5 +21,5 @@ Server { keyTab="/etc/zookeeper.keytab" storeKey=true useTicketCache=false - principal="zookeeper/<%= fqdn %>@<%= kerberos_realm %>"; + principal="zookeeper/<%= @fqdn %>@<%= @kerberos_realm %>"; }; diff --git a/bigtop-deploy/puppet/modules/hadoop-zookeeper/templates/zoo.cfg b/bigtop-deploy/puppet/modules/hadoop-zookeeper/templates/zoo.cfg index 154df874..fa6db1ea 100644 --- a/bigtop-deploy/puppet/modules/hadoop-zookeeper/templates/zoo.cfg +++ b/bigtop-deploy/puppet/modules/hadoop-zookeeper/templates/zoo.cfg @@ -26,11 +26,11 @@ syncLimit=5 dataDir=/var/lib/zookeeper # the port at which the clients will connect clientPort=2181 -<% ensemble.each_with_index do |server,idx| %> +<% @ensemble.each_with_index do |server,idx| %> server.<%= idx %>=<%= server %> <% end %> -<% if kerberos_realm != "" -%> +<% if @kerberos_realm != "" -%> authProvider.1=org.apache.zookeeper.server.auth.SASLAuthenticationProvider jaasLoginRenew=3600000 diff --git a/bigtop-deploy/puppet/modules/hadoop/manifests/init.pp b/bigtop-deploy/puppet/modules/hadoop/manifests/init.pp index a5c44b4a..0562f3f6 100644 --- a/bigtop-deploy/puppet/modules/hadoop/manifests/init.pp +++ b/bigtop-deploy/puppet/modules/hadoop/manifests/init.pp @@ -196,7 +196,7 @@ class hadoop { } file { "/etc/hadoop-httpfs/conf/httpfs-signature.secret": - content => inline_template("<%= secret %>"), + content => inline_template("<%= @secret %>"), require => [Package["hadoop-httpfs"]], } @@ -254,7 +254,7 @@ class hadoop { define namenode ($host = $fqdn , $port = "8020", $thrift_port= "10090", $auth = "simple", $dirs = ["/tmp/nn"], $ha = 'disabled', $zk = '') { - $first_namenode = inline_template("<%= Array(host)[0] %>") + $first_namenode = inline_template("<%= Array(@host)[0] %>") $hadoop_namenode_host = $host $hadoop_namenode_port = $port $hadoop_namenode_thrift_port = $thrift_port @@ -354,7 +354,7 @@ class hadoop { exec { "namenode format": user => "hdfs", command => "/bin/bash -c 'yes Y | hdfs namenode -format >> /var/lib/hadoop-hdfs/nn.format.log 2>&1'", - creates => "${namenode_data_dirs[0]}/current/VERSION", + creates => "${dirs[0]}/current/VERSION", require => [ Package["hadoop-hdfs-namenode"], File[$dirs], File["/etc/hadoop/conf/hdfs-site.xml"] ], tag => "namenode-format", } diff --git a/bigtop-deploy/puppet/modules/hadoop/templates/core-site.xml b/bigtop-deploy/puppet/modules/hadoop/templates/core-site.xml index 1a6daf3b..e1bb7990 100644 --- a/bigtop-deploy/puppet/modules/hadoop/templates/core-site.xml +++ b/bigtop-deploy/puppet/modules/hadoop/templates/core-site.xml @@ -16,41 +16,41 @@ <!-- See the License for the specific language governing permissions and --> <!-- limitations under the License. --> -<% namenode_hosts = Array(hadoop_namenode_host) -%> +<% namenode_hosts = Array(@hadoop_namenode_host) -%> <configuration> <property> <!-- URI of NN. Fully qualified. No IP.--> <name>fs.defaultFS</name> <% if @ha != "disabled" -%> - <value>hdfs://<%= nameservice_id %></value> + <value>hdfs://<%= @nameservice_id %></value> <% else -%> - <value>hdfs://<%= namenode_hosts[0] %>:<%= hadoop_namenode_port%></value> + <value>hdfs://<%= namenode_hosts[0] %>:<%= @hadoop_namenode_port %></value> <% end -%> </property> <property> <name>hadoop.security.authentication</name> - <value><%= hadoop_security_authentication %></value> + <value><%= @hadoop_security_authentication %></value> </property> -<% if hadoop_security_authentication == "kerberos" %> +<% if @hadoop_security_authentication == "kerberos" %> <property> <name>hadoop.security.authorization</name> <value>true</value> </property> <% end %> -<% if has_variable?("hadoop_security_group_mapping") %> +<% if @hadoop_security_group_mapping %> <property> <name>hadoop.security.group.mapping</name> - <value><%= hadoop_security_group_mapping %></value> + <value><%= @hadoop_security_group_mapping %></value> </property> <% end %> -<% if has_variable?("hadoop_core_proxyusers") %> -<% hadoop_core_proxyusers.sort.each do |superuser, data| %> +<% if @hadoop_core_proxyusers %> +<% @hadoop_core_proxyusers.sort.each do |superuser, data| %> <property> <name>hadoop.proxyuser.<%= superuser %>.hosts</name> <value><%= data['hosts'] %></value> @@ -62,7 +62,7 @@ <% end %> <% end %> -<% if has_variable?("hadoop_snappy_codec") %> +<% if @hadoop_snappy_codec %> <property> <name>io.compression.codecs</name> <value>org.apache.hadoop.io.compress.GzipCodec,org.apache.hadoop.io.compress.DefaultCodec,org.apache.hadoop.io.compress.BZip2Codec,org.apache.hadoop.io.compress.SnappyCodec</value> @@ -73,17 +73,17 @@ </property--> <% end %> -<% if has_variable?("hadoop_config_fs_inmemory_size_mb") %> +<% if @hadoop_config_fs_inmemory_size_mb %> <property> <name>fs.inmemory.size.mb</name> - <value><%= hadoop_config_fs_inmemory_size_mb %></value> + <value><%= @hadoop_config_fs_inmemory_size_mb %></value> </property> <% end %> -<% if has_variable?("hadoop_config_io_file_buffer_size") %> +<% if @hadoop_config_io_file_buffer_size %> <property> <name>io.file.buffer.size</name> - <value><%= hadoop_config_io_file_buffer_size %></value> + <value><%= @hadoop_config_io_file_buffer_size %></value> </property> <% end %> diff --git a/bigtop-deploy/puppet/modules/hadoop/templates/hadoop-hdfs b/bigtop-deploy/puppet/modules/hadoop/templates/hadoop-hdfs index bd708ebe..04572dd6 100644 --- a/bigtop-deploy/puppet/modules/hadoop/templates/hadoop-hdfs +++ b/bigtop-deploy/puppet/modules/hadoop/templates/hadoop-hdfs @@ -19,7 +19,7 @@ export HADOOP_SECONDARYNAMENODE_USER=hdfs export HADOOP_DATANODE_USER=hdfs export HADOOP_IDENT_STRING=hdfs -<% if hadoop_security_authentication == "kerberos" %> +<% if @hadoop_security_authentication == "kerberos" %> export HADOOP_SECURE_DN_USER=hdfs export HADOOP_SECURE_DN_PID_DIR=$HADOOP_PID_DIR export HADOOP_SECURE_DN_LOG_DIR=$HADOOP_LOG_DIR diff --git a/bigtop-deploy/puppet/modules/hadoop/templates/hdfs-site.xml b/bigtop-deploy/puppet/modules/hadoop/templates/hdfs-site.xml index 9302b509..88c72775 100644 --- a/bigtop-deploy/puppet/modules/hadoop/templates/hdfs-site.xml +++ b/bigtop-deploy/puppet/modules/hadoop/templates/hdfs-site.xml @@ -16,7 +16,7 @@ <!-- See the License for the specific language governing permissions and --> <!-- limitations under the License. --> -<% namenode_hosts = Array(hadoop_namenode_host) -%> +<% namenode_hosts = Array(@hadoop_namenode_host) -%> <configuration> <% if @ha != "disabled" -%> @@ -27,57 +27,57 @@ <name>dfs.ha.automatic-failover.enabled</name> <value>true</value> </property> -<% if has_variable?("zk") -%> +<% if @zk -%> <property> <name>ha.zookeeper.quorum</name> - <value><%= zk %></value> + <value><%= @zk %></value> </property> <% end -%> <% end -%> <property> <name>dfs.federation.nameservices</name> - <value><%= nameservice_id %></value> + <value><%= @nameservice_id %></value> </property> <property> - <name>dfs.ha.namenodes.<%= nameservice_id %></name> + <name>dfs.ha.namenodes.<%= @nameservice_id %></name> <value><%= (1..namenode_hosts.length).map { |n| "nn#{n}" }.join(",") %></value> </property> <% namenode_hosts.each_with_index do |host,idx| -%> <property> - <name>dfs.namenode.rpc-address.<%= nameservice_id %>.nn<%= idx+1 %></name> - <value><%= host %>:<%= hadoop_namenode_port %></value> + <name>dfs.namenode.rpc-address.<%= @nameservice_id %>.nn<%= idx+1 %></name> + <value><%= host %>:<%= @hadoop_namenode_port %></value> </property> <property> - <name>dfs.namenode.http-address.<%= nameservice_id %>.nn<%= idx+1 %></name> + <name>dfs.namenode.http-address.<%= @nameservice_id %>.nn<%= idx+1 %></name> <value><%= host %>:50070</value> </property> <% end -%> -<% if has_variable?("shared_edits_dir") -%> +<% if @shared_edits_dir -%> <property> <name>dfs.namenode.shared.edits.dir</name> - <value><%= shared_edits_dir %></value> + <value><%= @shared_edits_dir %></value> </property> <% end -%> <property> - <name>dfs.client.failover.proxy.provider.<%= nameservice_id %></name> + <name>dfs.client.failover.proxy.provider.<%= @nameservice_id %></name> <value>org.apache.hadoop.hdfs.server.namenode.ha.ConfiguredFailoverProxyProvider</value> </property> -<% if has_variable?("sshfence_privkey") -%> +<% if @sshfence_privkey -%> <property> <name>dfs.ha.fencing.methods</name> - <value>sshfence(<%= sshfence_user %>)</value> + <value>sshfence(<%= @sshfence_user %>)</value> </property> <property> <name>dfs.ha.fencing.ssh.private-key-files</name> - <value><%= sshfence_keypath %></value> + <value><%= @sshfence_keypath %></value> </property> <% else -%> @@ -87,7 +87,7 @@ </property> <% end -%> -<% elsif hadoop_security_authentication == "kerberos" -%> +<% elsif @hadoop_security_authentication == "kerberos" -%> <property> <name>dfs.block.access.token.enable</name> <value>true</value> @@ -108,11 +108,11 @@ </property> <property> <name>dfs.namenode.kerberos.principal</name> - <value>hdfs/_HOST@<%= kerberos_realm %></value> + <value>hdfs/_HOST@<%= @kerberos_realm %></value> </property> <property> <name>dfs.namenode.kerberos.https.principal</name> - <value>host/_HOST@<%= kerberos_realm %></value> + <value>host/_HOST@<%= @kerberos_realm %></value> </property> <property> <name>dfs.web.authentication.kerberos.keytab</name> @@ -120,7 +120,7 @@ </property> <property> <name>dfs.web.authentication.kerberos.principal</name> - <value>HTTP/_HOST@<%= kerberos_realm %></value> + <value>HTTP/_HOST@<%= @kerberos_realm %></value> </property> @@ -143,11 +143,11 @@ </property> <property> <name>dfs.secondary.namenode.kerberos.principal</name> - <value>hdfs/_HOST@<%= kerberos_realm %></value> + <value>hdfs/_HOST@<%= @kerberos_realm %></value> </property> <property> <name>dfs.secondary.namenode.kerberos.https.principal</name> - <value>host/_HOST@<%= kerberos_realm %></value> + <value>host/_HOST@<%= @kerberos_realm %></value> </property> <!-- DataNode security config --> @@ -169,28 +169,28 @@ </property> <property> <name>dfs.datanode.kerberos.principal</name> - <value>hdfs/_HOST@<%= kerberos_realm %></value> + <value>hdfs/_HOST@<%= @kerberos_realm %></value> </property> <property> <name>dfs.datanode.kerberos.https.principal</name> - <value>host/_HOST@<%= kerberos_realm %></value> + <value>host/_HOST@<%= @kerberos_realm %></value> </property> <% end -%> <!-- name node --> <property> <name>dfs.data.dir</name> - <value><%= hdfs_data_dirs.join(",") %></value> + <value><%= @hdfs_data_dirs.join(",") %></value> </property> -<% if has_variable?("hdfs_support_append") %> +<% if @hdfs_support_append %> <property> <name>dfs.support.append</name> - <value><%= hdfs_support_append %></value> + <value><%= @hdfs_support_append %></value> </property> <% end %> -<% if has_variable?("hdfs_shortcut_reader_user") %> +<% if @hdfs_shortcut_reader_user %> <property> <name>dfs.client.read.shortcircuit</name> <value>true</value> @@ -198,13 +198,13 @@ <property> <name>dfs.block.local-path-access.user</name> - <value><%= hdfs_shortcut_reader_user %></value> + <value><%= @hdfs_shortcut_reader_user %></value> </property> <% end %> <property> <name>dfs.name.dir</name> - <value><%= namenode_data_dirs.join(",") %></value> + <value><%= @namenode_data_dirs.join(",") %></value> </property> <property> @@ -214,28 +214,28 @@ </property> <!-- Enable Hue plugins --> -<% if has_variable?("hadoop_dfs_namenode_plugins") -%> +<% if @hadoop_dfs_namenode_plugins -%> <property> <name>dfs.namenode.plugins</name> - <value><%= hadoop_dfs_namenode_plugins %></value> + <value><%= @hadoop_dfs_namenode_plugins %></value> <description>Comma-separated list of namenode plug-ins to be activated. </description> </property> <% end -%> -<% if has_variable?("hadoop_dfs_datanode_plugins") -%> +<% if @hadoop_dfs_datanode_plugins -%> <property> <name>dfs.datanode.plugins</name> - <value><%= hadoop_dfs_datanode_plugins %></value> + <value><%= @hadoop_dfs_datanode_plugins %></value> <description>Comma-separated list of datanode plug-ins to be activated. </description> </property> <% end -%> -<% if has_variable?("hadoop_namenode_thrift_port") -%> +<% if @hadoop_namenode_thrift_port -%> <property> <name>dfs.thrift.address</name> - <value>0.0.0.0:<%= hadoop_namenode_thrift_port %></value> + <value>0.0.0.0:<%= @hadoop_namenode_thrift_port %></value> </property> <% end -%> @@ -247,17 +247,17 @@ </property> <!-- Configurations for large cluster --> -<% if has_variable?("hadoop_config_dfs_block_size") -%> +<% if @hadoop_config_dfs_block_size -%> <property> <name>dfs.block.size</name> - <value><%= hadoop_config_dfs_block_size %></value> + <value><%= @hadoop_config_dfs_block_size %></value> </property> <% end -%> -<% if has_variable?("hadoop_config_namenode_handler_count") -%> +<% if @hadoop_config_namenode_handler_count -%> <property> <name>dfs.namenode.handler.count</name> - <value><%= hadoop_config_namenode_handler_count %></value> + <value><%= @hadoop_config_namenode_handler_count %></value> </property> <% end -%> diff --git a/bigtop-deploy/puppet/modules/hadoop/templates/httpfs-env.sh b/bigtop-deploy/puppet/modules/hadoop/templates/httpfs-env.sh index 950e1905..1de11149 100644 --- a/bigtop-deploy/puppet/modules/hadoop/templates/httpfs-env.sh +++ b/bigtop-deploy/puppet/modules/hadoop/templates/httpfs-env.sh @@ -30,7 +30,7 @@ # The HTTP port used by HttpFS # -export HTTPFS_HTTP_PORT=<%= hadoop_httpfs_port %> +export HTTPFS_HTTP_PORT=<%= @hadoop_httpfs_port %> # The Admin port used by HttpFS # diff --git a/bigtop-deploy/puppet/modules/hadoop/templates/httpfs-site.xml b/bigtop-deploy/puppet/modules/hadoop/templates/httpfs-site.xml index 85cf4a6d..ed702caa 100644 --- a/bigtop-deploy/puppet/modules/hadoop/templates/httpfs-site.xml +++ b/bigtop-deploy/puppet/modules/hadoop/templates/httpfs-site.xml @@ -16,15 +16,15 @@ <!-- See the License for the specific language governing permissions and --> <!-- limitations under the License. --> -<% namenode_hosts = Array(hadoop_namenode_host) -%> +<% namenode_hosts = Array(@hadoop_namenode_host) -%> <configuration> <property> <name>httpfs.hadoop.config.dir</name> <value>/etc/hadoop/conf</value> </property> -<% if has_variable?("hadoop_core_proxyusers") -%> -<% hadoop_core_proxyusers.sort.each do |superuser, data| -%> +<% if @hadoop_core_proxyusers -%> +<% @hadoop_core_proxyusers.sort.each do |superuser, data| -%> <property> <name>httpfs.proxyuser.<%= superuser %>.hosts</name> <value><%= data['hosts'] %></value> @@ -36,18 +36,18 @@ <% end -%> <% end -%> -<% if hadoop_security_authentication == "kerberos" -%> +<% if @hadoop_security_authentication == "kerberos" -%> <property> <name>httpfs.authentication.type</name> <value>kerberos</value> </property> <property> <name>kerberos.realm</name> - <value><%= kerberos_realm %></value> + <value><%= @kerberos_realm %></value> </property> <property> <name>httpfs.authentication.kerberos.principal</name> - <value>HTTP/<%= fqdn %></value> + <value>HTTP/<%= @fqdn %></value> </property> <property> <name>httpfs.authentication.kerberos.keytab</name> @@ -60,7 +60,7 @@ </property> <property> <name>httpfs.hadoop.authentication.kerberos.principal</name> - <value>httpfs/<%= fqdn %></value> + <value>httpfs/<%= @fqdn %></value> </property> <property> <name>httpfs.hadoop.authentication.kerberos.keytab</name> diff --git a/bigtop-deploy/puppet/modules/hadoop/templates/mapred-site.xml b/bigtop-deploy/puppet/modules/hadoop/templates/mapred-site.xml index 3c370fe2..a2517e5c 100644 --- a/bigtop-deploy/puppet/modules/hadoop/templates/mapred-site.xml +++ b/bigtop-deploy/puppet/modules/hadoop/templates/mapred-site.xml @@ -18,11 +18,11 @@ <configuration> -<% if hadoop_security_authentication == "kerberos" %> +<% if @hadoop_security_authentication == "kerberos" %> <!-- HistoryServer security configs --> <property> <name>mapreduce.jobhistory.principal</name> - <value>mapred/_HOST@<%= kerberos_realm %></value> + <value>mapred/_HOST@<%= @kerberos_realm %></value> </property> <property> <name>mapreduce.jobhistory.keytab</name> @@ -32,11 +32,11 @@ <!-- JobTracker security configs --> <property> <name>mapreduce.jobtracker.kerberos.principal</name> - <value>mapred/_HOST@<%= kerberos_realm %></value> + <value>mapred/_HOST@<%= @kerberos_realm %></value> </property> <property> <name>mapreduce.jobtracker.kerberos.https.principal</name> - <value>host/_HOST@<%= kerberos_realm %></value> + <value>host/_HOST@<%= @kerberos_realm %></value> </property> <property> <name>mapreduce.jobtracker.keytab.file</name> @@ -46,11 +46,11 @@ <!-- TaskTracker security configs --> <property> <name>mapreduce.tasktracker.kerberos.principal</name> - <value>mapred/_HOST@<%= kerberos_realm %></value> + <value>mapred/_HOST@<%= @kerberos_realm %></value> </property> <property> <name>mapreduce.tasktracker.kerberos.https.principal</name> - <value>host/_HOST@<%= kerberos_realm %></value> + <value>host/_HOST@<%= @kerberos_realm %></value> </property> <property> <name>mapreduce.tasktracker.keytab.file</name> @@ -68,50 +68,50 @@ </property> <% end %> -<% if has_variable?("mapred_acls_enabled") %> +<% if @mapred_acls_enabled %> <property> <name>mapreduce.cluster.acls.enabled</name> - <value><%= mapred_acls_enabled %></value> + <value><%= @mapred_acls_enabled %></value> </property> <% end %> <!-- specify JobTracker TaskScheduler --> -<% if has_variable?("hadoop_jobtracker_taskscheduler") %> +<% if @hadoop_jobtracker_taskscheduler %> <property> <name>mapreduce.jobtracker.taskscheduler</name> - <value><%= hadoop_jobtracker_taskscheduler %></value> + <value><%= @hadoop_jobtracker_taskscheduler %></value> </property> <% end %> -<% if has_variable?("hadoop_config_mapred_fairscheduler_assignmultiple") %> +<% if @hadoop_config_mapred_fairscheduler_assignmultiple %> <property> <name>mapred.fairscheduler.assignmultiple</name> - <value><%= hadoop_config_mapred_fairscheduler_assignmultiple %></value> + <value><%= @hadoop_config_mapred_fairscheduler_assignmultiple %></value> </property> <% end %> -<% if has_variable?("hadoop_config_mapred_fairscheduler_sizebasedweight") %> +<% if @hadoop_config_mapred_fairscheduler_sizebasedweight %> <property> <name>mapred.fairscheduler.sizebasedweight</name> - <value><%= hadoop_config_mapred_fairscheduler_assignmultiple %></value> + <value><%= @hadoop_config_mapred_fairscheduler_assignmultiple %></value> </property> <% end %> -<% if has_variable?("hadoop_jobtracker_fairscheduler_weightadjuster") %> +<% if @hadoop_jobtracker_fairscheduler_weightadjuster %> <property> <name>mapred.fairscheduler.weightadjuster</name> - <value><%= hadoop_jobtracker_fairscheduler_weightadjuster %></value> + <value><%= @hadoop_jobtracker_fairscheduler_weightadjuster %></value> </property> <% end %> -<% if has_variable?("hadoop_hs_host") %> +<% if @hadoop_hs_host %> <property> <name>mapreduce.jobhistory.address</name> - <value><%= hadoop_hs_host %>:<%= hadoop_hs_port %></value> + <value><%= @hadoop_hs_host %>:<%= @hadoop_hs_port %></value> </property> <property> <name>mapreduce.jobhistory.webapp.address</name> - <value><%= hadoop_hs_host %>:<%= hadoop_hs_webapp_port %></value> + <value><%= @hadoop_hs_host %>:<%= @hadoop_hs_webapp_port %></value> </property> <% end %> @@ -122,12 +122,12 @@ <property> <name>mapreduce.jobtracker.address</name> - <value><%= hadoop_jobtracker_host %>:<%= hadoop_jobtracker_port%></value> + <value><%= @hadoop_jobtracker_host %>:<%= @hadoop_jobtracker_port%></value> </property> <property> <name>mapreduce.cluster.local.dir</name> - <value><%= mapred_data_dirs.join(",") %></value> + <value><%= @mapred_data_dirs.join(",") %></value> <final>true</final> </property> @@ -156,11 +156,11 @@ <value>-Xmx1024m</value> </property> -<% if has_variable?("hadoop_config_mapred_child_ulimit") %> +<% if @hadoop_config_mapred_child_ulimit %> <property> <!-- set this to ~1.5x the heap size in mapred.child.java.opts --> <name>mapred.child.ulimit</name> - <value><%= hadoop_config_mapred_child_ulimit %></value> + <value><%= @hadoop_config_mapred_child_ulimit %></value> </property> <% else %> <property> @@ -170,10 +170,10 @@ </property> <% end %> -<% if has_variable?("hadoop_config_io_sort_mb") %> +<% if @hadoop_config_io_sort_mb %> <property> <name>io.sort.mb</name> - <value><%= hadoop_config_io_sort_mb %></value> + <value><%= @hadoop_config_io_sort_mb %></value> </property> <% else %> <property> @@ -182,10 +182,10 @@ </property> <% end %> -<% if has_variable?("hadoop_config_io_sort_factor") %> +<% if @hadoop_config_io_sort_factor %> <property> <name>io.sort.factor</name> - <value><%= hadoop_config_io_sort_factor %></value> + <value><%= @hadoop_config_io_sort_factor %></value> </property> <% else %> <property> @@ -194,10 +194,10 @@ </property> <% end %> -<% if has_variable?("hadoop_config_mapred_job_tracker_handler_count") %> +<% if @hadoop_config_mapred_job_tracker_handler_count %> <property> <name>mapred.job.tracker.handler.count</name> - <value><%= hadoop_config_mapred_job_tracker_handler_count %></value> + <value><%= @hadoop_config_mapred_job_tracker_handler_count %></value> <final>true</final> </property> <% else %> @@ -213,12 +213,12 @@ <value>true</value> </property> -<% if has_variable?("hadoop_config_mapred_reduce_parallel_copies") %> +<% if @hadoop_config_mapred_reduce_parallel_copies %> <property> <name>mapred.reduce.parallel.copies</name> <!-- set this to somewhere between sqrt(nodes) and nodes/2. for <20 nodes, set == |nodes| --> - <value><%= hadoop_config_mapred_reduce_parallel_copies %></value> + <value><%= @hadoop_config_mapred_reduce_parallel_copies %></value> </property> <% else %> <property> @@ -235,10 +235,10 @@ <value>30</value> </property> -<% if has_variable?("hadoop_config_mapred_reduce_tasks_speculative_execution") %> +<% if @hadoop_config_mapred_reduce_tasks_speculative_execution %> <property> <name>mapred.reduce.tasks.speculative.execution</name> - <value><%= hadoop_config_mapred_reduce_tasks_speculative_execution %></value> + <value><%= @hadoop_config_mapred_reduce_tasks_speculative_execution %></value> </property> <% else %> <property> @@ -250,21 +250,21 @@ <property> <name>mapred.tasktracker.map.tasks.maximum</name> <!-- see other kb entry about this one. --> - <value><%= [1, processorcount.to_i * 0.80].max.round %></value> + <value><%= [1, @processorcount.to_i * 0.80].max.round %></value> <final>true</final> </property> <property> <name>mapred.tasktracker.reduce.tasks.maximum</name> <!-- see other kb entry about this one. --> - <value><%= [1, processorcount.to_i * 0.20].max.round %></value> + <value><%= [1, @processorcount.to_i * 0.20].max.round %></value> <final>true</final> </property> -<% if has_variable?("hadoop_config_tasktracker_http_threads") %> +<% if @hadoop_config_tasktracker_http_threads %> <property> <name>tasktracker.http.threads</name> - <value><%= hadoop_config_tasktracker_http_threads %></value> + <value><%= @hadoop_config_tasktracker_http_threads %></value> <final>true</final> </property> <% else %> @@ -283,10 +283,10 @@ NONE, RECORD or BLOCK.</description> </property> -<% if has_variable?("hadoop_config_use_compression") %> +<% if @hadoop_config_use_compression %> <property> <name>mapred.compress.map.output</name> - <value><%= hadoop_config_use_compression %></value> + <value><%= @hadoop_config_use_compression %></value> </property> <% else %> <property> @@ -295,31 +295,31 @@ </property> <% end %> -<% if has_variable?("hadoop_config_mapred_reduce_slowstart_completed_maps") %> +<% if @hadoop_config_mapred_reduce_slowstart_completed_maps %> <property> <name>mapred.reduce.slowstart.completed.maps</name> - <value><%= hadoop_config_mapred_reduce_slowstart_completed_maps %></value> + <value><%= @hadoop_config_mapred_reduce_slowstart_completed_maps %></value> </property> <% end %> -<% if has_variable?("hadoop_jobtracker_thrift_port") %> +<% if @hadoop_jobtracker_thrift_port %> <!-- Enable Hue plugins --> <property> <name>jobtracker.thrift.address</name> - <value>0.0.0.0:<%= hadoop_jobtracker_thrift_port %></value> + <value>0.0.0.0:<%= @hadoop_jobtracker_thrift_port %></value> </property> <% end %> -<% if has_variable?("hadoop_mapred_jobtracker_plugins") %> +<% if @hadoop_mapred_jobtracker_plugins %> <property> <name>mapred.jobtracker.plugins</name> - <value><%= hadoop_mapred_jobtracker_plugins %></value> + <value><%= @hadoop_mapred_jobtracker_plugins %></value> <description>Comma-separated list of jobtracker plug-ins to be activated.</description> </property> <% end %> -<% if has_variable?("hadoop_mapred_tasktracker_plugins") %> +<% if @hadoop_mapred_tasktracker_plugins %> <property> <name>mapred.tasktracker.instrumentation</name> - <value><%= hadoop_mapred_tasktracker_plugins %></value> + <value><%= @hadoop_mapred_tasktracker_plugins %></value> </property> <% end %> diff --git a/bigtop-deploy/puppet/modules/hadoop/templates/taskcontroller.cfg b/bigtop-deploy/puppet/modules/hadoop/templates/taskcontroller.cfg index 99df794e..a1229ad6 100644 --- a/bigtop-deploy/puppet/modules/hadoop/templates/taskcontroller.cfg +++ b/bigtop-deploy/puppet/modules/hadoop/templates/taskcontroller.cfg @@ -1,4 +1,4 @@ -mapred.local.dir=<%= mapred_data_dirs.join(",") %> +mapred.local.dir=<%= @mapred_data_dirs.join(",") %> hadoop.log.dir=/var/log/hadoop/ mapreduce.tasktracker.group=mapred min.user.id=0 diff --git a/bigtop-deploy/puppet/modules/hadoop/templates/yarn-site.xml b/bigtop-deploy/puppet/modules/hadoop/templates/yarn-site.xml index 4433ba9c..0713d971 100644 --- a/bigtop-deploy/puppet/modules/hadoop/templates/yarn-site.xml +++ b/bigtop-deploy/puppet/modules/hadoop/templates/yarn-site.xml @@ -18,11 +18,11 @@ <?xml-stylesheet type="text/xsl" href="configuration.xsl"?> <configuration> -<% if hadoop_security_authentication == "kerberos" %> +<% if @hadoop_security_authentication == "kerberos" %> <!-- JobTracker security configs --> <property> <name>yarn.resourcemanager.principal</name> - <value>yarn/_HOST@<%= kerberos_realm %></value> + <value>yarn/_HOST@<%= @kerberos_realm %></value> </property> <property> <name>yarn.resourcemanager.keytab</name> @@ -31,7 +31,7 @@ <property> <name>yarn.web-proxy.principal</name> - <value>yarn/_HOST@<%= kerberos_realm %></value> + <value>yarn/_HOST@<%= @kerberos_realm %></value> </property> <property> <name>yarn.web-proxy.keytab</name> @@ -40,7 +40,7 @@ <property> <name>yarn.nodemanager.principal</name> - <value>yarn/_HOST@<%= kerberos_realm %></value> + <value>yarn/_HOST@<%= @kerberos_realm %></value> </property> <property> <name>yarn.nodemanager.keytab</name> @@ -58,22 +58,22 @@ <property> <name>yarn.web-proxy.address</name> - <value><%= hadoop_ps_host %>:<%= hadoop_ps_port %></value> + <value><%= @hadoop_ps_host %>:<%= @hadoop_ps_port %></value> </property> <property> <name>yarn.resourcemanager.resource-tracker.address</name> - <value><%= hadoop_rm_host %>:<%= hadoop_rt_port %></value> + <value><%= @hadoop_rm_host %>:<%= @hadoop_rt_port %></value> </property> <property> <name>yarn.resourcemanager.address</name> - <value><%= hadoop_rm_host %>:<%= hadoop_rm_port %></value> + <value><%= @hadoop_rm_host %>:<%= @hadoop_rm_port %></value> </property> <property> <name>yarn.resourcemanager.scheduler.address</name> - <value><%= hadoop_rm_host %>:<%= hadoop_sc_port %></value> + <value><%= @hadoop_rm_host %>:<%= @hadoop_sc_port %></value> </property> <property> @@ -98,7 +98,7 @@ <property> <name>yarn.nodemanager.local-dirs</name> - <value><%= yarn_data_dirs.join(",") %></value> + <value><%= @yarn_data_dirs.join(",") %></value> <final>true</final> </property> diff --git a/bigtop-deploy/puppet/modules/hue/templates/hue.ini b/bigtop-deploy/puppet/modules/hue/templates/hue.ini index 44e011f7..c23ec19f 100644 --- a/bigtop-deploy/puppet/modules/hue/templates/hue.ini +++ b/bigtop-deploy/puppet/modules/hue/templates/hue.ini @@ -41,8 +41,8 @@ secret_key=deaddeafdeaddeafdeaddeafdeaddeafdeaddeafdeaddeafdeaddeaf # Webserver listens on this address and port - http_host=<%= hue_host %> - http_port=<%= hue_port %> + http_host=<%= @hue_host %> + http_port=<%= @hue_port %> # Time zone name time_zone=America/Los_Angeles @@ -215,7 +215,7 @@ ## default_from_email=hue@localhost -<% if kerberos_realm != "" %> +<% if @kerberos_realm != "" %> # Configuration options for Kerberos integration for secured Hadoop clusters # ------------------------------------------------------------------------ [[kerberos]] @@ -223,9 +223,9 @@ # Path to Hue's Kerberos keytab file hue_keytab=/etc/hue.keytab # Kerberos principal name for Hue - hue_principal=hue/<%= fqdn %> + hue_principal=hue/<%= @fqdn %> # Path to kinit - kinit_path=<%= (operatingsystem == 'ubuntu') ? '/usr/bin' : '/usr/kerberos/bin' %>/kinit + kinit_path=<%= (@operatingsystem == 'ubuntu') ? '/usr/bin' : '/usr/kerberos/bin' %>/kinit <% end %> @@ -261,15 +261,15 @@ [[[default]]] # Enter the filesystem uri - fs_defaultfs=<%= default_fs %> + fs_defaultfs=<%= @default_fs %> # Change this if your HDFS cluster is Kerberos-secured - security_enabled=<%= if (kerberos_realm != "") ; "true" else "false" end %> + security_enabled=<%= if (@kerberos_realm != "") ; "true" else "false" end %> # Use WebHdfs/HttpFs as the communication mechanism. # This should be the web service root URL, such as # http://namenode:50070/webhdfs/v1 - webhdfs_url=<%= webhdfs_url %> + webhdfs_url=<%= @webhdfs_url %> # Settings about this HDFS cluster. If you install HDFS in a # different location, you need to set the following. @@ -318,14 +318,14 @@ [[[default]]] # Enter the host on which you are running the ResourceManager - resourcemanager_host=<%= rm_host %> + resourcemanager_host=<%= @rm_host %> # The port where the ResourceManager IPC listens on - resourcemanager_port=<%= rm_port %> + resourcemanager_port=<%= @rm_port %> # Whether to submit jobs to this cluster submit_to=True # Change this if your YARN cluster is Kerberos-secured - security_enabled=<%= if (kerberos_realm != "") ; "true" else "false" end %> + security_enabled=<%= if (@kerberos_realm != "") ; "true" else "false" end %> # Settings about this MR2 cluster. If you install MR2 in a # different location, you need to set the following. @@ -340,13 +340,13 @@ hadoop_conf_dir=/etc/hadoop/conf # URL of the ResourceManager API - resourcemanager_api_url=<%= rm_url %> + resourcemanager_api_url=<%= @rm_url %> # URL of the ProxyServer API - proxy_api_url=<%= rm_proxy_url %> + proxy_api_url=<%= @rm_proxy_url %> # URL of the HistoryServer API - history_server_api_url=<%= history_server_url %> + history_server_api_url=<%= @history_server_url %> # URL of the NodeManager API node_manager_api_url=http://localhost:8042 @@ -359,9 +359,9 @@ [liboozie] # The URL where the Oozie service runs on. This is required in order for # users to submit jobs. - oozie_url=<%= oozie_url %> + oozie_url=<%= @oozie_url %> - security_enabled=<%= if (kerberos_realm != "") ; "true" else "false" end %> + security_enabled=<%= if (@kerberos_realm != "") ; "true" else "false" end %> # Location on HDFS where the workflows/coordinator are deployed when submitted. remote_deployement_dir=/user/hue/oozie/deployments diff --git a/bigtop-deploy/puppet/modules/kerberos/templates/kdc.conf b/bigtop-deploy/puppet/modules/kerberos/templates/kdc.conf index dc35b324..5bc384bf 100644 --- a/bigtop-deploy/puppet/modules/kerberos/templates/kdc.conf +++ b/bigtop-deploy/puppet/modules/kerberos/templates/kdc.conf @@ -13,19 +13,19 @@ # See the License for the specific language governing permissions and # limitations under the License. -default_realm = <%= realm %> +default_realm = <%= @realm %> [kdcdefaults] # v4_mode = nopreauth kdc_ports = 0 [realms] - <%= realm %> = { - kdc_ports = <%= kdc_port %> + <%= @realm %> = { + kdc_ports = <%= @kdc_port %> admin_keytab = /etc/kadm5.keytab - database_name = <%= kdc_etc_path %>/principal - acl_file = <%= kdc_etc_path %>/kadm5.acl - key_stash_file = <%= kdc_etc_path %>/stash + database_name = <%= @kdc_etc_path %>/principal + acl_file = <%= @kdc_etc_path %>/kadm5.acl + key_stash_file = <%= @kdc_etc_path %>/stash max_life = 10h 0m 0s max_renewable_life = 7d 0h 0m 0s master_key_type = des3-hmac-sha1 diff --git a/bigtop-deploy/puppet/modules/kerberos/templates/krb5.conf b/bigtop-deploy/puppet/modules/kerberos/templates/krb5.conf index ccec5206..10c06c75 100644 --- a/bigtop-deploy/puppet/modules/kerberos/templates/krb5.conf +++ b/bigtop-deploy/puppet/modules/kerberos/templates/krb5.conf @@ -14,7 +14,7 @@ # limitations under the License. [libdefaults] - default_realm = <%= realm %> + default_realm = <%= @realm %> dns_lookup_realm = false dns_lookup_kdc = false ticket_lifetime = 24h @@ -25,15 +25,15 @@ permitted_enctypes = des-cbc-md5 des-cbc-crc des3-cbc-sha1 [realms] - <%= realm %> = { - kdc = <%= kdc_server %>:<%= kdc_port %> - admin_server = <%= kdc_server %>:<%= admin_port %> - default_domain = <%= domain %> + <%= @realm %> = { + kdc = <%= @kdc_server %>:<%= @kdc_port %> + admin_server = <%= @kdc_server %>:<%= @admin_port %> + default_domain = <%= @domain %> } [domain_realm] - .<%= domain %> = <%= realm %> - <%= domain %> = <%= realm %> + .<%= @domain %> = <%= @realm %> + <%= @domain %> = <%= @realm %> [logging] kdc = FILE:/var/log/krb5kdc.log diff --git a/bigtop-deploy/puppet/modules/spark/templates/spark-env.sh b/bigtop-deploy/puppet/modules/spark/templates/spark-env.sh index 6d4ab2a3..bbd70815 100644 --- a/bigtop-deploy/puppet/modules/spark/templates/spark-env.sh +++ b/bigtop-deploy/puppet/modules/spark/templates/spark-env.sh @@ -38,8 +38,8 @@ export SPARK_LAUNCH_WITH_SCALA=0 export SPARK_LIBRARY_PATH=${SPARK_HOME}/lib export SCALA_LIBRARY_PATH=${SPARK_HOME}/lib -export SPARK_MASTER_WEBUI_PORT=<%= master_ui_port %> -export SPARK_MASTER_PORT=<%= master_port %> +export SPARK_MASTER_WEBUI_PORT=<%= @master_ui_port %> +export SPARK_MASTER_PORT=<%= @master_port %> ### Comment above 2 lines and uncomment the following if ### you want to run with scala version, that is included with the package @@ -47,6 +47,6 @@ export SPARK_MASTER_PORT=<%= master_port %> #export PATH=$PATH:$SCALA_HOME/bin ### change the following to specify a real cluster's Master host -export STANDALONE_SPARK_MASTER_HOST=<%= master_host %> +export STANDALONE_SPARK_MASTER_HOST=<%= @master_host %> diff --git a/bigtop-deploy/vm/docker-puppet/provision.sh b/bigtop-deploy/vm/docker-puppet/provision.sh index 5bd36058..dfd58b65 100755 --- a/bigtop-deploy/vm/docker-puppet/provision.sh +++ b/bigtop-deploy/vm/docker-puppet/provision.sh @@ -17,7 +17,7 @@ # Install puppet agent yum -y install http://yum.puppetlabs.com/puppetlabs-release-el-6.noarch.rpm -yum -y install puppet-2.7.24-1.el6.noarch +yum -y install puppet sysctl kernel.hostname=`hostname -f` diff --git a/bigtop-deploy/vm/smoke-tests.sh b/bigtop-deploy/vm/smoke-tests.sh index d02c128d..9ee111ab 100755 --- a/bigtop-deploy/vm/smoke-tests.sh +++ b/bigtop-deploy/vm/smoke-tests.sh @@ -9,7 +9,7 @@ export HIVE_HOME=/usr/lib/hive/ export PIG_HOME=/usr/lib/pig/ export FLUME_HOME=/usr/lib/flume/ export HIVE_CONF_DIR=/etc/hive/conf/ -export JAVA_HOME="/usr/lib/jvm/java-1.7.0-openjdk-1.7.0.65.x86_64/" +export JAVA_HOME="/usr/lib/jvm/java-openjdk/" export MAHOUT_HOME="/usr/lib/mahout" export ITEST="0.7.0" diff --git a/bigtop-deploy/vm/vagrant-puppet/provision.sh b/bigtop-deploy/vm/vagrant-puppet/provision.sh index fa0571b3..2b660c2e 100755 --- a/bigtop-deploy/vm/vagrant-puppet/provision.sh +++ b/bigtop-deploy/vm/vagrant-puppet/provision.sh @@ -17,7 +17,7 @@ # Install puppet agent yum -y install http://yum.puppetlabs.com/puppetlabs-release-el-6.noarch.rpm -yum -y install puppet-2.7.23-1.el6.noarch +yum -y install puppet service iptables stop chkconfig iptables off |