aboutsummaryrefslogtreecommitdiff
path: root/bigtop-deploy
diff options
context:
space:
mode:
authorevans_ye <innovansye@gmail.com>2014-10-30 16:49:15 +0800
committerjay@apache.org <jayunit100>2014-10-30 19:52:11 -0400
commit11c8bd5543cd8f21d33f92447c5b996a5886d798 (patch)
treeabe5d31977cd665a8cc9a5993bc9c10ea22d9eb3 /bigtop-deploy
parente722c9ced7ce4f89b43623f1ff04a4c1a7b75369 (diff)
BIGTOP-1047. Support Puppet 3.x
Signed-off-by: Konstantin Boudnik <cos@apache.org>
Diffstat (limited to 'bigtop-deploy')
-rwxr-xr-xbigtop-deploy/puppet/modules/giraph/templates/giraph-site.xml2
-rw-r--r--bigtop-deploy/puppet/modules/hadoop-flume/templates/flume.conf12
-rw-r--r--bigtop-deploy/puppet/modules/hadoop-hbase/templates/hbase-site.xml2
-rw-r--r--bigtop-deploy/puppet/modules/hadoop-hbase/templates/jaas.conf2
-rw-r--r--bigtop-deploy/puppet/modules/hadoop-hive/templates/hive-site.xml8
-rw-r--r--bigtop-deploy/puppet/modules/hadoop-oozie/templates/oozie-site.xml10
-rw-r--r--bigtop-deploy/puppet/modules/hadoop-zookeeper/manifests/init.pp2
-rw-r--r--bigtop-deploy/puppet/modules/hadoop-zookeeper/templates/jaas.conf2
-rw-r--r--bigtop-deploy/puppet/modules/hadoop-zookeeper/templates/zoo.cfg4
-rw-r--r--bigtop-deploy/puppet/modules/hadoop/manifests/init.pp6
-rw-r--r--bigtop-deploy/puppet/modules/hadoop/templates/core-site.xml28
-rw-r--r--bigtop-deploy/puppet/modules/hadoop/templates/hadoop-hdfs2
-rw-r--r--bigtop-deploy/puppet/modules/hadoop/templates/hdfs-site.xml76
-rw-r--r--bigtop-deploy/puppet/modules/hadoop/templates/httpfs-env.sh2
-rw-r--r--bigtop-deploy/puppet/modules/hadoop/templates/httpfs-site.xml14
-rw-r--r--bigtop-deploy/puppet/modules/hadoop/templates/mapred-site.xml94
-rw-r--r--bigtop-deploy/puppet/modules/hadoop/templates/taskcontroller.cfg2
-rw-r--r--bigtop-deploy/puppet/modules/hadoop/templates/yarn-site.xml18
-rw-r--r--bigtop-deploy/puppet/modules/hue/templates/hue.ini32
-rw-r--r--bigtop-deploy/puppet/modules/kerberos/templates/kdc.conf12
-rw-r--r--bigtop-deploy/puppet/modules/kerberos/templates/krb5.conf14
-rw-r--r--bigtop-deploy/puppet/modules/spark/templates/spark-env.sh6
-rwxr-xr-xbigtop-deploy/vm/docker-puppet/provision.sh2
-rwxr-xr-xbigtop-deploy/vm/smoke-tests.sh2
-rwxr-xr-xbigtop-deploy/vm/vagrant-puppet/provision.sh2
25 files changed, 178 insertions, 178 deletions
diff --git a/bigtop-deploy/puppet/modules/giraph/templates/giraph-site.xml b/bigtop-deploy/puppet/modules/giraph/templates/giraph-site.xml
index 9500fa97..5d5228f2 100755
--- a/bigtop-deploy/puppet/modules/giraph/templates/giraph-site.xml
+++ b/bigtop-deploy/puppet/modules/giraph/templates/giraph-site.xml
@@ -29,7 +29,7 @@ under the License.
<property>
<name>giraph.zkList</name>
- <value><%= zookeeper_quorum %></value>
+ <value><%= @zookeeper_quorum %></value>
<description>Prefer external Zookeeper over the embedded one</description>
</property>
</configuration>
diff --git a/bigtop-deploy/puppet/modules/hadoop-flume/templates/flume.conf b/bigtop-deploy/puppet/modules/hadoop-flume/templates/flume.conf
index 7007ead3..7d1b1a23 100644
--- a/bigtop-deploy/puppet/modules/hadoop-flume/templates/flume.conf
+++ b/bigtop-deploy/puppet/modules/hadoop-flume/templates/flume.conf
@@ -13,23 +13,23 @@
# See the License for the specific language governing permissions and
# limitations under the License.
-agent.channels = <%= (channels.collect { |k,v| k }).join(' ') %>
-agent.sources = <%= (sources.collect { |k,v| k }).join(' ') %>
-agent.sinks = <%= (sinks.collect { |k,v| k }).join(' ') %>
+agent.channels = <%= (@channels.collect { |k,v| k }).join(' ') %>
+agent.sources = <%= (@sources.collect { |k,v| k }).join(' ') %>
+agent.sinks = <%= (@sinks.collect { |k,v| k }).join(' ') %>
-<% channels.each do |channel,props| -%>
+<% @channels.each do |channel,props| -%>
<% props.each do |k,v| -%>
agent.channels.<%= channel %>.<%= k %> = <%= v %>
<% end -%>
<% end -%>
-<% sources.each do |source,props| -%>
+<% @sources.each do |source,props| -%>
<% props.each do |k,v| -%>
agent.sources.<%= source %>.<%= k %> = <%= v %>
<% end -%>
<% end -%>
-<% sinks.each do |sink,props| -%>
+<% @sinks.each do |sink,props| -%>
<% props.each do |k,v| -%>
agent.sinks.<%= sink %>.<%= k %> = <%= v %>
<% end -%>
diff --git a/bigtop-deploy/puppet/modules/hadoop-hbase/templates/hbase-site.xml b/bigtop-deploy/puppet/modules/hadoop-hbase/templates/hbase-site.xml
index 8489a93f..849676c8 100644
--- a/bigtop-deploy/puppet/modules/hadoop-hbase/templates/hbase-site.xml
+++ b/bigtop-deploy/puppet/modules/hadoop-hbase/templates/hbase-site.xml
@@ -42,7 +42,7 @@
<value>true</value>
</property>
-<% if has_variable?("hdfs_shortcut_reader_user") %>
+<% if @hdfs_shortcut_reader_user %>
<property>
<name>dfs.client.read.shortcircuit</name>
<value>true</value>
diff --git a/bigtop-deploy/puppet/modules/hadoop-hbase/templates/jaas.conf b/bigtop-deploy/puppet/modules/hadoop-hbase/templates/jaas.conf
index 8c546722..5b6b2a58 100644
--- a/bigtop-deploy/puppet/modules/hadoop-hbase/templates/jaas.conf
+++ b/bigtop-deploy/puppet/modules/hadoop-hbase/templates/jaas.conf
@@ -20,5 +20,5 @@ Client {
useKeyTab=true
useTicketCache=false
keyTab="/etc/hbase.keytab"
- principal="hbase/<%= fqdn %>@<%= kerberos_realm %>";
+ principal="hbase/<%= @fqdn %>@<%= @kerberos_realm %>";
};
diff --git a/bigtop-deploy/puppet/modules/hadoop-hive/templates/hive-site.xml b/bigtop-deploy/puppet/modules/hadoop-hive/templates/hive-site.xml
index 4f8cf897..5861d971 100644
--- a/bigtop-deploy/puppet/modules/hadoop-hive/templates/hive-site.xml
+++ b/bigtop-deploy/puppet/modules/hadoop-hive/templates/hive-site.xml
@@ -26,18 +26,18 @@
<!-- Hive Execution Parameters -->
-<% if hbase_master != "" %>
+<% if @hbase_master != "" %>
<property>
<name>hbase.master</name>
- <value><%= hbase_master %></value>
+ <value><%= @hbase_master %></value>
<description>http://wiki.apache.org/hadoop/Hive/HBaseIntegration</description>
</property>
<% end %>
-<% if hbase_zookeeper_quorum != "" %>
+<% if @hbase_zookeeper_quorum != "" %>
<property>
<name>hbase.zookeeper.quorum</name>
- <value><%= hbase_zookeeper_quorum %></value>
+ <value><%= @hbase_zookeeper_quorum %></value>
<description>http://wiki.apache.org/hadoop/Hive/HBaseIntegration</description>
</property>
<% end %>
diff --git a/bigtop-deploy/puppet/modules/hadoop-oozie/templates/oozie-site.xml b/bigtop-deploy/puppet/modules/hadoop-oozie/templates/oozie-site.xml
index ef39045b..93d5eafc 100644
--- a/bigtop-deploy/puppet/modules/hadoop-oozie/templates/oozie-site.xml
+++ b/bigtop-deploy/puppet/modules/hadoop-oozie/templates/oozie-site.xml
@@ -176,7 +176,7 @@
<property>
<name>oozie.service.HadoopAccessorService.kerberos.enabled</name>
- <value><%= if (kerberos_realm != "") ; "true" else "false" end %></value>
+ <value><%= if (@kerberos_realm != "") ; "true" else "false" end %></value>
<description>
Indicates if Oozie is configured to use Kerberos.
</description>
@@ -184,7 +184,7 @@
<property>
<name>local.realm</name>
- <value><%= kerberos_realm %></value>
+ <value><%= @kerberos_realm %></value>
<description>
Kerberos Realm used by Oozie and Hadoop. Using 'local.realm' to be aligned with Hadoop configuration
</description>
@@ -200,7 +200,7 @@
<property>
<name>oozie.service.HadoopAccessorService.kerberos.principal</name>
- <value>${user.name}/<%= fqdn %>@${local.realm}</value>
+ <value>${user.name}/<%= @fqdn %>@${local.realm}</value>
<description>
Kerberos principal for Oozie service.
</description>
@@ -258,7 +258,7 @@
<property>
<name>oozie.authentication.type</name>
- <value><%= kerberos_realm.empty? ? "simple" : "kerberos" %></value>
+ <value><%= @kerberos_realm.empty? ? "simple" : "kerberos" %></value>
<description>
Defines authentication used for Oozie HTTP endpoint.
Supported values are: simple | kerberos | #AUTHENTICATION_HANDLER_CLASSNAME#
@@ -306,7 +306,7 @@
<property>
<name>oozie.authentication.kerberos.principal</name>
- <value>HTTP/<%= fqdn %>@${local.realm}</value>
+ <value>HTTP/<%= @fqdn %>@${local.realm}</value>
<description>
Indicates the Kerberos principal to be used for HTTP endpoint.
The principal MUST start with 'HTTP/' as per Kerberos HTTP SPNEGO specification.
diff --git a/bigtop-deploy/puppet/modules/hadoop-zookeeper/manifests/init.pp b/bigtop-deploy/puppet/modules/hadoop-zookeeper/manifests/init.pp
index 8e0c7576..701590e1 100644
--- a/bigtop-deploy/puppet/modules/hadoop-zookeeper/manifests/init.pp
+++ b/bigtop-deploy/puppet/modules/hadoop-zookeeper/manifests/init.pp
@@ -45,7 +45,7 @@ class hadoop-zookeeper {
}
file { "/var/lib/zookeeper/myid":
- content => inline_template("<%= myid %>"),
+ content => inline_template("<%= @myid %>"),
require => Package["zookeeper-server"],
}
diff --git a/bigtop-deploy/puppet/modules/hadoop-zookeeper/templates/jaas.conf b/bigtop-deploy/puppet/modules/hadoop-zookeeper/templates/jaas.conf
index 4655502e..61b2eac4 100644
--- a/bigtop-deploy/puppet/modules/hadoop-zookeeper/templates/jaas.conf
+++ b/bigtop-deploy/puppet/modules/hadoop-zookeeper/templates/jaas.conf
@@ -21,5 +21,5 @@ Server {
keyTab="/etc/zookeeper.keytab"
storeKey=true
useTicketCache=false
- principal="zookeeper/<%= fqdn %>@<%= kerberos_realm %>";
+ principal="zookeeper/<%= @fqdn %>@<%= @kerberos_realm %>";
};
diff --git a/bigtop-deploy/puppet/modules/hadoop-zookeeper/templates/zoo.cfg b/bigtop-deploy/puppet/modules/hadoop-zookeeper/templates/zoo.cfg
index 154df874..fa6db1ea 100644
--- a/bigtop-deploy/puppet/modules/hadoop-zookeeper/templates/zoo.cfg
+++ b/bigtop-deploy/puppet/modules/hadoop-zookeeper/templates/zoo.cfg
@@ -26,11 +26,11 @@ syncLimit=5
dataDir=/var/lib/zookeeper
# the port at which the clients will connect
clientPort=2181
-<% ensemble.each_with_index do |server,idx| %>
+<% @ensemble.each_with_index do |server,idx| %>
server.<%= idx %>=<%= server %>
<% end %>
-<% if kerberos_realm != "" -%>
+<% if @kerberos_realm != "" -%>
authProvider.1=org.apache.zookeeper.server.auth.SASLAuthenticationProvider
jaasLoginRenew=3600000
diff --git a/bigtop-deploy/puppet/modules/hadoop/manifests/init.pp b/bigtop-deploy/puppet/modules/hadoop/manifests/init.pp
index a5c44b4a..0562f3f6 100644
--- a/bigtop-deploy/puppet/modules/hadoop/manifests/init.pp
+++ b/bigtop-deploy/puppet/modules/hadoop/manifests/init.pp
@@ -196,7 +196,7 @@ class hadoop {
}
file { "/etc/hadoop-httpfs/conf/httpfs-signature.secret":
- content => inline_template("<%= secret %>"),
+ content => inline_template("<%= @secret %>"),
require => [Package["hadoop-httpfs"]],
}
@@ -254,7 +254,7 @@ class hadoop {
define namenode ($host = $fqdn , $port = "8020", $thrift_port= "10090", $auth = "simple", $dirs = ["/tmp/nn"], $ha = 'disabled', $zk = '') {
- $first_namenode = inline_template("<%= Array(host)[0] %>")
+ $first_namenode = inline_template("<%= Array(@host)[0] %>")
$hadoop_namenode_host = $host
$hadoop_namenode_port = $port
$hadoop_namenode_thrift_port = $thrift_port
@@ -354,7 +354,7 @@ class hadoop {
exec { "namenode format":
user => "hdfs",
command => "/bin/bash -c 'yes Y | hdfs namenode -format >> /var/lib/hadoop-hdfs/nn.format.log 2>&1'",
- creates => "${namenode_data_dirs[0]}/current/VERSION",
+ creates => "${dirs[0]}/current/VERSION",
require => [ Package["hadoop-hdfs-namenode"], File[$dirs], File["/etc/hadoop/conf/hdfs-site.xml"] ],
tag => "namenode-format",
}
diff --git a/bigtop-deploy/puppet/modules/hadoop/templates/core-site.xml b/bigtop-deploy/puppet/modules/hadoop/templates/core-site.xml
index 1a6daf3b..e1bb7990 100644
--- a/bigtop-deploy/puppet/modules/hadoop/templates/core-site.xml
+++ b/bigtop-deploy/puppet/modules/hadoop/templates/core-site.xml
@@ -16,41 +16,41 @@
<!-- See the License for the specific language governing permissions and -->
<!-- limitations under the License. -->
-<% namenode_hosts = Array(hadoop_namenode_host) -%>
+<% namenode_hosts = Array(@hadoop_namenode_host) -%>
<configuration>
<property>
<!-- URI of NN. Fully qualified. No IP.-->
<name>fs.defaultFS</name>
<% if @ha != "disabled" -%>
- <value>hdfs://<%= nameservice_id %></value>
+ <value>hdfs://<%= @nameservice_id %></value>
<% else -%>
- <value>hdfs://<%= namenode_hosts[0] %>:<%= hadoop_namenode_port%></value>
+ <value>hdfs://<%= namenode_hosts[0] %>:<%= @hadoop_namenode_port %></value>
<% end -%>
</property>
<property>
<name>hadoop.security.authentication</name>
- <value><%= hadoop_security_authentication %></value>
+ <value><%= @hadoop_security_authentication %></value>
</property>
-<% if hadoop_security_authentication == "kerberos" %>
+<% if @hadoop_security_authentication == "kerberos" %>
<property>
<name>hadoop.security.authorization</name>
<value>true</value>
</property>
<% end %>
-<% if has_variable?("hadoop_security_group_mapping") %>
+<% if @hadoop_security_group_mapping %>
<property>
<name>hadoop.security.group.mapping</name>
- <value><%= hadoop_security_group_mapping %></value>
+ <value><%= @hadoop_security_group_mapping %></value>
</property>
<% end %>
-<% if has_variable?("hadoop_core_proxyusers") %>
-<% hadoop_core_proxyusers.sort.each do |superuser, data| %>
+<% if @hadoop_core_proxyusers %>
+<% @hadoop_core_proxyusers.sort.each do |superuser, data| %>
<property>
<name>hadoop.proxyuser.<%= superuser %>.hosts</name>
<value><%= data['hosts'] %></value>
@@ -62,7 +62,7 @@
<% end %>
<% end %>
-<% if has_variable?("hadoop_snappy_codec") %>
+<% if @hadoop_snappy_codec %>
<property>
<name>io.compression.codecs</name>
<value>org.apache.hadoop.io.compress.GzipCodec,org.apache.hadoop.io.compress.DefaultCodec,org.apache.hadoop.io.compress.BZip2Codec,org.apache.hadoop.io.compress.SnappyCodec</value>
@@ -73,17 +73,17 @@
</property-->
<% end %>
-<% if has_variable?("hadoop_config_fs_inmemory_size_mb") %>
+<% if @hadoop_config_fs_inmemory_size_mb %>
<property>
<name>fs.inmemory.size.mb</name>
- <value><%= hadoop_config_fs_inmemory_size_mb %></value>
+ <value><%= @hadoop_config_fs_inmemory_size_mb %></value>
</property>
<% end %>
-<% if has_variable?("hadoop_config_io_file_buffer_size") %>
+<% if @hadoop_config_io_file_buffer_size %>
<property>
<name>io.file.buffer.size</name>
- <value><%= hadoop_config_io_file_buffer_size %></value>
+ <value><%= @hadoop_config_io_file_buffer_size %></value>
</property>
<% end %>
diff --git a/bigtop-deploy/puppet/modules/hadoop/templates/hadoop-hdfs b/bigtop-deploy/puppet/modules/hadoop/templates/hadoop-hdfs
index bd708ebe..04572dd6 100644
--- a/bigtop-deploy/puppet/modules/hadoop/templates/hadoop-hdfs
+++ b/bigtop-deploy/puppet/modules/hadoop/templates/hadoop-hdfs
@@ -19,7 +19,7 @@ export HADOOP_SECONDARYNAMENODE_USER=hdfs
export HADOOP_DATANODE_USER=hdfs
export HADOOP_IDENT_STRING=hdfs
-<% if hadoop_security_authentication == "kerberos" %>
+<% if @hadoop_security_authentication == "kerberos" %>
export HADOOP_SECURE_DN_USER=hdfs
export HADOOP_SECURE_DN_PID_DIR=$HADOOP_PID_DIR
export HADOOP_SECURE_DN_LOG_DIR=$HADOOP_LOG_DIR
diff --git a/bigtop-deploy/puppet/modules/hadoop/templates/hdfs-site.xml b/bigtop-deploy/puppet/modules/hadoop/templates/hdfs-site.xml
index 9302b509..88c72775 100644
--- a/bigtop-deploy/puppet/modules/hadoop/templates/hdfs-site.xml
+++ b/bigtop-deploy/puppet/modules/hadoop/templates/hdfs-site.xml
@@ -16,7 +16,7 @@
<!-- See the License for the specific language governing permissions and -->
<!-- limitations under the License. -->
-<% namenode_hosts = Array(hadoop_namenode_host) -%>
+<% namenode_hosts = Array(@hadoop_namenode_host) -%>
<configuration>
<% if @ha != "disabled" -%>
@@ -27,57 +27,57 @@
<name>dfs.ha.automatic-failover.enabled</name>
<value>true</value>
</property>
-<% if has_variable?("zk") -%>
+<% if @zk -%>
<property>
<name>ha.zookeeper.quorum</name>
- <value><%= zk %></value>
+ <value><%= @zk %></value>
</property>
<% end -%>
<% end -%>
<property>
<name>dfs.federation.nameservices</name>
- <value><%= nameservice_id %></value>
+ <value><%= @nameservice_id %></value>
</property>
<property>
- <name>dfs.ha.namenodes.<%= nameservice_id %></name>
+ <name>dfs.ha.namenodes.<%= @nameservice_id %></name>
<value><%= (1..namenode_hosts.length).map { |n| "nn#{n}" }.join(",") %></value>
</property>
<% namenode_hosts.each_with_index do |host,idx| -%>
<property>
- <name>dfs.namenode.rpc-address.<%= nameservice_id %>.nn<%= idx+1 %></name>
- <value><%= host %>:<%= hadoop_namenode_port %></value>
+ <name>dfs.namenode.rpc-address.<%= @nameservice_id %>.nn<%= idx+1 %></name>
+ <value><%= host %>:<%= @hadoop_namenode_port %></value>
</property>
<property>
- <name>dfs.namenode.http-address.<%= nameservice_id %>.nn<%= idx+1 %></name>
+ <name>dfs.namenode.http-address.<%= @nameservice_id %>.nn<%= idx+1 %></name>
<value><%= host %>:50070</value>
</property>
<% end -%>
-<% if has_variable?("shared_edits_dir") -%>
+<% if @shared_edits_dir -%>
<property>
<name>dfs.namenode.shared.edits.dir</name>
- <value><%= shared_edits_dir %></value>
+ <value><%= @shared_edits_dir %></value>
</property>
<% end -%>
<property>
- <name>dfs.client.failover.proxy.provider.<%= nameservice_id %></name>
+ <name>dfs.client.failover.proxy.provider.<%= @nameservice_id %></name>
<value>org.apache.hadoop.hdfs.server.namenode.ha.ConfiguredFailoverProxyProvider</value>
</property>
-<% if has_variable?("sshfence_privkey") -%>
+<% if @sshfence_privkey -%>
<property>
<name>dfs.ha.fencing.methods</name>
- <value>sshfence(<%= sshfence_user %>)</value>
+ <value>sshfence(<%= @sshfence_user %>)</value>
</property>
<property>
<name>dfs.ha.fencing.ssh.private-key-files</name>
- <value><%= sshfence_keypath %></value>
+ <value><%= @sshfence_keypath %></value>
</property>
<% else -%>
@@ -87,7 +87,7 @@
</property>
<% end -%>
-<% elsif hadoop_security_authentication == "kerberos" -%>
+<% elsif @hadoop_security_authentication == "kerberos" -%>
<property>
<name>dfs.block.access.token.enable</name>
<value>true</value>
@@ -108,11 +108,11 @@
</property>
<property>
<name>dfs.namenode.kerberos.principal</name>
- <value>hdfs/_HOST@<%= kerberos_realm %></value>
+ <value>hdfs/_HOST@<%= @kerberos_realm %></value>
</property>
<property>
<name>dfs.namenode.kerberos.https.principal</name>
- <value>host/_HOST@<%= kerberos_realm %></value>
+ <value>host/_HOST@<%= @kerberos_realm %></value>
</property>
<property>
<name>dfs.web.authentication.kerberos.keytab</name>
@@ -120,7 +120,7 @@
</property>
<property>
<name>dfs.web.authentication.kerberos.principal</name>
- <value>HTTP/_HOST@<%= kerberos_realm %></value>
+ <value>HTTP/_HOST@<%= @kerberos_realm %></value>
</property>
@@ -143,11 +143,11 @@
</property>
<property>
<name>dfs.secondary.namenode.kerberos.principal</name>
- <value>hdfs/_HOST@<%= kerberos_realm %></value>
+ <value>hdfs/_HOST@<%= @kerberos_realm %></value>
</property>
<property>
<name>dfs.secondary.namenode.kerberos.https.principal</name>
- <value>host/_HOST@<%= kerberos_realm %></value>
+ <value>host/_HOST@<%= @kerberos_realm %></value>
</property>
<!-- DataNode security config -->
@@ -169,28 +169,28 @@
</property>
<property>
<name>dfs.datanode.kerberos.principal</name>
- <value>hdfs/_HOST@<%= kerberos_realm %></value>
+ <value>hdfs/_HOST@<%= @kerberos_realm %></value>
</property>
<property>
<name>dfs.datanode.kerberos.https.principal</name>
- <value>host/_HOST@<%= kerberos_realm %></value>
+ <value>host/_HOST@<%= @kerberos_realm %></value>
</property>
<% end -%>
<!-- name node -->
<property>
<name>dfs.data.dir</name>
- <value><%= hdfs_data_dirs.join(",") %></value>
+ <value><%= @hdfs_data_dirs.join(",") %></value>
</property>
-<% if has_variable?("hdfs_support_append") %>
+<% if @hdfs_support_append %>
<property>
<name>dfs.support.append</name>
- <value><%= hdfs_support_append %></value>
+ <value><%= @hdfs_support_append %></value>
</property>
<% end %>
-<% if has_variable?("hdfs_shortcut_reader_user") %>
+<% if @hdfs_shortcut_reader_user %>
<property>
<name>dfs.client.read.shortcircuit</name>
<value>true</value>
@@ -198,13 +198,13 @@
<property>
<name>dfs.block.local-path-access.user</name>
- <value><%= hdfs_shortcut_reader_user %></value>
+ <value><%= @hdfs_shortcut_reader_user %></value>
</property>
<% end %>
<property>
<name>dfs.name.dir</name>
- <value><%= namenode_data_dirs.join(",") %></value>
+ <value><%= @namenode_data_dirs.join(",") %></value>
</property>
<property>
@@ -214,28 +214,28 @@
</property>
<!-- Enable Hue plugins -->
-<% if has_variable?("hadoop_dfs_namenode_plugins") -%>
+<% if @hadoop_dfs_namenode_plugins -%>
<property>
<name>dfs.namenode.plugins</name>
- <value><%= hadoop_dfs_namenode_plugins %></value>
+ <value><%= @hadoop_dfs_namenode_plugins %></value>
<description>Comma-separated list of namenode plug-ins to be activated.
</description>
</property>
<% end -%>
-<% if has_variable?("hadoop_dfs_datanode_plugins") -%>
+<% if @hadoop_dfs_datanode_plugins -%>
<property>
<name>dfs.datanode.plugins</name>
- <value><%= hadoop_dfs_datanode_plugins %></value>
+ <value><%= @hadoop_dfs_datanode_plugins %></value>
<description>Comma-separated list of datanode plug-ins to be activated.
</description>
</property>
<% end -%>
-<% if has_variable?("hadoop_namenode_thrift_port") -%>
+<% if @hadoop_namenode_thrift_port -%>
<property>
<name>dfs.thrift.address</name>
- <value>0.0.0.0:<%= hadoop_namenode_thrift_port %></value>
+ <value>0.0.0.0:<%= @hadoop_namenode_thrift_port %></value>
</property>
<% end -%>
@@ -247,17 +247,17 @@
</property>
<!-- Configurations for large cluster -->
-<% if has_variable?("hadoop_config_dfs_block_size") -%>
+<% if @hadoop_config_dfs_block_size -%>
<property>
<name>dfs.block.size</name>
- <value><%= hadoop_config_dfs_block_size %></value>
+ <value><%= @hadoop_config_dfs_block_size %></value>
</property>
<% end -%>
-<% if has_variable?("hadoop_config_namenode_handler_count") -%>
+<% if @hadoop_config_namenode_handler_count -%>
<property>
<name>dfs.namenode.handler.count</name>
- <value><%= hadoop_config_namenode_handler_count %></value>
+ <value><%= @hadoop_config_namenode_handler_count %></value>
</property>
<% end -%>
diff --git a/bigtop-deploy/puppet/modules/hadoop/templates/httpfs-env.sh b/bigtop-deploy/puppet/modules/hadoop/templates/httpfs-env.sh
index 950e1905..1de11149 100644
--- a/bigtop-deploy/puppet/modules/hadoop/templates/httpfs-env.sh
+++ b/bigtop-deploy/puppet/modules/hadoop/templates/httpfs-env.sh
@@ -30,7 +30,7 @@
# The HTTP port used by HttpFS
#
-export HTTPFS_HTTP_PORT=<%= hadoop_httpfs_port %>
+export HTTPFS_HTTP_PORT=<%= @hadoop_httpfs_port %>
# The Admin port used by HttpFS
#
diff --git a/bigtop-deploy/puppet/modules/hadoop/templates/httpfs-site.xml b/bigtop-deploy/puppet/modules/hadoop/templates/httpfs-site.xml
index 85cf4a6d..ed702caa 100644
--- a/bigtop-deploy/puppet/modules/hadoop/templates/httpfs-site.xml
+++ b/bigtop-deploy/puppet/modules/hadoop/templates/httpfs-site.xml
@@ -16,15 +16,15 @@
<!-- See the License for the specific language governing permissions and -->
<!-- limitations under the License. -->
-<% namenode_hosts = Array(hadoop_namenode_host) -%>
+<% namenode_hosts = Array(@hadoop_namenode_host) -%>
<configuration>
<property>
<name>httpfs.hadoop.config.dir</name>
<value>/etc/hadoop/conf</value>
</property>
-<% if has_variable?("hadoop_core_proxyusers") -%>
-<% hadoop_core_proxyusers.sort.each do |superuser, data| -%>
+<% if @hadoop_core_proxyusers -%>
+<% @hadoop_core_proxyusers.sort.each do |superuser, data| -%>
<property>
<name>httpfs.proxyuser.<%= superuser %>.hosts</name>
<value><%= data['hosts'] %></value>
@@ -36,18 +36,18 @@
<% end -%>
<% end -%>
-<% if hadoop_security_authentication == "kerberos" -%>
+<% if @hadoop_security_authentication == "kerberos" -%>
<property>
<name>httpfs.authentication.type</name>
<value>kerberos</value>
</property>
<property>
<name>kerberos.realm</name>
- <value><%= kerberos_realm %></value>
+ <value><%= @kerberos_realm %></value>
</property>
<property>
<name>httpfs.authentication.kerberos.principal</name>
- <value>HTTP/<%= fqdn %></value>
+ <value>HTTP/<%= @fqdn %></value>
</property>
<property>
<name>httpfs.authentication.kerberos.keytab</name>
@@ -60,7 +60,7 @@
</property>
<property>
<name>httpfs.hadoop.authentication.kerberos.principal</name>
- <value>httpfs/<%= fqdn %></value>
+ <value>httpfs/<%= @fqdn %></value>
</property>
<property>
<name>httpfs.hadoop.authentication.kerberos.keytab</name>
diff --git a/bigtop-deploy/puppet/modules/hadoop/templates/mapred-site.xml b/bigtop-deploy/puppet/modules/hadoop/templates/mapred-site.xml
index 3c370fe2..a2517e5c 100644
--- a/bigtop-deploy/puppet/modules/hadoop/templates/mapred-site.xml
+++ b/bigtop-deploy/puppet/modules/hadoop/templates/mapred-site.xml
@@ -18,11 +18,11 @@
<configuration>
-<% if hadoop_security_authentication == "kerberos" %>
+<% if @hadoop_security_authentication == "kerberos" %>
<!-- HistoryServer security configs -->
<property>
<name>mapreduce.jobhistory.principal</name>
- <value>mapred/_HOST@<%= kerberos_realm %></value>
+ <value>mapred/_HOST@<%= @kerberos_realm %></value>
</property>
<property>
<name>mapreduce.jobhistory.keytab</name>
@@ -32,11 +32,11 @@
<!-- JobTracker security configs -->
<property>
<name>mapreduce.jobtracker.kerberos.principal</name>
- <value>mapred/_HOST@<%= kerberos_realm %></value>
+ <value>mapred/_HOST@<%= @kerberos_realm %></value>
</property>
<property>
<name>mapreduce.jobtracker.kerberos.https.principal</name>
- <value>host/_HOST@<%= kerberos_realm %></value>
+ <value>host/_HOST@<%= @kerberos_realm %></value>
</property>
<property>
<name>mapreduce.jobtracker.keytab.file</name>
@@ -46,11 +46,11 @@
<!-- TaskTracker security configs -->
<property>
<name>mapreduce.tasktracker.kerberos.principal</name>
- <value>mapred/_HOST@<%= kerberos_realm %></value>
+ <value>mapred/_HOST@<%= @kerberos_realm %></value>
</property>
<property>
<name>mapreduce.tasktracker.kerberos.https.principal</name>
- <value>host/_HOST@<%= kerberos_realm %></value>
+ <value>host/_HOST@<%= @kerberos_realm %></value>
</property>
<property>
<name>mapreduce.tasktracker.keytab.file</name>
@@ -68,50 +68,50 @@
</property>
<% end %>
-<% if has_variable?("mapred_acls_enabled") %>
+<% if @mapred_acls_enabled %>
<property>
<name>mapreduce.cluster.acls.enabled</name>
- <value><%= mapred_acls_enabled %></value>
+ <value><%= @mapred_acls_enabled %></value>
</property>
<% end %>
<!-- specify JobTracker TaskScheduler -->
-<% if has_variable?("hadoop_jobtracker_taskscheduler") %>
+<% if @hadoop_jobtracker_taskscheduler %>
<property>
<name>mapreduce.jobtracker.taskscheduler</name>
- <value><%= hadoop_jobtracker_taskscheduler %></value>
+ <value><%= @hadoop_jobtracker_taskscheduler %></value>
</property>
<% end %>
-<% if has_variable?("hadoop_config_mapred_fairscheduler_assignmultiple") %>
+<% if @hadoop_config_mapred_fairscheduler_assignmultiple %>
<property>
<name>mapred.fairscheduler.assignmultiple</name>
- <value><%= hadoop_config_mapred_fairscheduler_assignmultiple %></value>
+ <value><%= @hadoop_config_mapred_fairscheduler_assignmultiple %></value>
</property>
<% end %>
-<% if has_variable?("hadoop_config_mapred_fairscheduler_sizebasedweight") %>
+<% if @hadoop_config_mapred_fairscheduler_sizebasedweight %>
<property>
<name>mapred.fairscheduler.sizebasedweight</name>
- <value><%= hadoop_config_mapred_fairscheduler_assignmultiple %></value>
+ <value><%= @hadoop_config_mapred_fairscheduler_assignmultiple %></value>
</property>
<% end %>
-<% if has_variable?("hadoop_jobtracker_fairscheduler_weightadjuster") %>
+<% if @hadoop_jobtracker_fairscheduler_weightadjuster %>
<property>
<name>mapred.fairscheduler.weightadjuster</name>
- <value><%= hadoop_jobtracker_fairscheduler_weightadjuster %></value>
+ <value><%= @hadoop_jobtracker_fairscheduler_weightadjuster %></value>
</property>
<% end %>
-<% if has_variable?("hadoop_hs_host") %>
+<% if @hadoop_hs_host %>
<property>
<name>mapreduce.jobhistory.address</name>
- <value><%= hadoop_hs_host %>:<%= hadoop_hs_port %></value>
+ <value><%= @hadoop_hs_host %>:<%= @hadoop_hs_port %></value>
</property>
<property>
<name>mapreduce.jobhistory.webapp.address</name>
- <value><%= hadoop_hs_host %>:<%= hadoop_hs_webapp_port %></value>
+ <value><%= @hadoop_hs_host %>:<%= @hadoop_hs_webapp_port %></value>
</property>
<% end %>
@@ -122,12 +122,12 @@
<property>
<name>mapreduce.jobtracker.address</name>
- <value><%= hadoop_jobtracker_host %>:<%= hadoop_jobtracker_port%></value>
+ <value><%= @hadoop_jobtracker_host %>:<%= @hadoop_jobtracker_port%></value>
</property>
<property>
<name>mapreduce.cluster.local.dir</name>
- <value><%= mapred_data_dirs.join(",") %></value>
+ <value><%= @mapred_data_dirs.join(",") %></value>
<final>true</final>
</property>
@@ -156,11 +156,11 @@
<value>-Xmx1024m</value>
</property>
-<% if has_variable?("hadoop_config_mapred_child_ulimit") %>
+<% if @hadoop_config_mapred_child_ulimit %>
<property>
<!-- set this to ~1.5x the heap size in mapred.child.java.opts -->
<name>mapred.child.ulimit</name>
- <value><%= hadoop_config_mapred_child_ulimit %></value>
+ <value><%= @hadoop_config_mapred_child_ulimit %></value>
</property>
<% else %>
<property>
@@ -170,10 +170,10 @@
</property>
<% end %>
-<% if has_variable?("hadoop_config_io_sort_mb") %>
+<% if @hadoop_config_io_sort_mb %>
<property>
<name>io.sort.mb</name>
- <value><%= hadoop_config_io_sort_mb %></value>
+ <value><%= @hadoop_config_io_sort_mb %></value>
</property>
<% else %>
<property>
@@ -182,10 +182,10 @@
</property>
<% end %>
-<% if has_variable?("hadoop_config_io_sort_factor") %>
+<% if @hadoop_config_io_sort_factor %>
<property>
<name>io.sort.factor</name>
- <value><%= hadoop_config_io_sort_factor %></value>
+ <value><%= @hadoop_config_io_sort_factor %></value>
</property>
<% else %>
<property>
@@ -194,10 +194,10 @@
</property>
<% end %>
-<% if has_variable?("hadoop_config_mapred_job_tracker_handler_count") %>
+<% if @hadoop_config_mapred_job_tracker_handler_count %>
<property>
<name>mapred.job.tracker.handler.count</name>
- <value><%= hadoop_config_mapred_job_tracker_handler_count %></value>
+ <value><%= @hadoop_config_mapred_job_tracker_handler_count %></value>
<final>true</final>
</property>
<% else %>
@@ -213,12 +213,12 @@
<value>true</value>
</property>
-<% if has_variable?("hadoop_config_mapred_reduce_parallel_copies") %>
+<% if @hadoop_config_mapred_reduce_parallel_copies %>
<property>
<name>mapred.reduce.parallel.copies</name>
<!-- set this to somewhere between sqrt(nodes) and nodes/2.
for <20 nodes, set == |nodes| -->
- <value><%= hadoop_config_mapred_reduce_parallel_copies %></value>
+ <value><%= @hadoop_config_mapred_reduce_parallel_copies %></value>
</property>
<% else %>
<property>
@@ -235,10 +235,10 @@
<value>30</value>
</property>
-<% if has_variable?("hadoop_config_mapred_reduce_tasks_speculative_execution") %>
+<% if @hadoop_config_mapred_reduce_tasks_speculative_execution %>
<property>
<name>mapred.reduce.tasks.speculative.execution</name>
- <value><%= hadoop_config_mapred_reduce_tasks_speculative_execution %></value>
+ <value><%= @hadoop_config_mapred_reduce_tasks_speculative_execution %></value>
</property>
<% else %>
<property>
@@ -250,21 +250,21 @@
<property>
<name>mapred.tasktracker.map.tasks.maximum</name>
<!-- see other kb entry about this one. -->
- <value><%= [1, processorcount.to_i * 0.80].max.round %></value>
+ <value><%= [1, @processorcount.to_i * 0.80].max.round %></value>
<final>true</final>
</property>
<property>
<name>mapred.tasktracker.reduce.tasks.maximum</name>
<!-- see other kb entry about this one. -->
- <value><%= [1, processorcount.to_i * 0.20].max.round %></value>
+ <value><%= [1, @processorcount.to_i * 0.20].max.round %></value>
<final>true</final>
</property>
-<% if has_variable?("hadoop_config_tasktracker_http_threads") %>
+<% if @hadoop_config_tasktracker_http_threads %>
<property>
<name>tasktracker.http.threads</name>
- <value><%= hadoop_config_tasktracker_http_threads %></value>
+ <value><%= @hadoop_config_tasktracker_http_threads %></value>
<final>true</final>
</property>
<% else %>
@@ -283,10 +283,10 @@
NONE, RECORD or BLOCK.</description>
</property>
-<% if has_variable?("hadoop_config_use_compression") %>
+<% if @hadoop_config_use_compression %>
<property>
<name>mapred.compress.map.output</name>
- <value><%= hadoop_config_use_compression %></value>
+ <value><%= @hadoop_config_use_compression %></value>
</property>
<% else %>
<property>
@@ -295,31 +295,31 @@
</property>
<% end %>
-<% if has_variable?("hadoop_config_mapred_reduce_slowstart_completed_maps") %>
+<% if @hadoop_config_mapred_reduce_slowstart_completed_maps %>
<property>
<name>mapred.reduce.slowstart.completed.maps</name>
- <value><%= hadoop_config_mapred_reduce_slowstart_completed_maps %></value>
+ <value><%= @hadoop_config_mapred_reduce_slowstart_completed_maps %></value>
</property>
<% end %>
-<% if has_variable?("hadoop_jobtracker_thrift_port") %>
+<% if @hadoop_jobtracker_thrift_port %>
<!-- Enable Hue plugins -->
<property>
<name>jobtracker.thrift.address</name>
- <value>0.0.0.0:<%= hadoop_jobtracker_thrift_port %></value>
+ <value>0.0.0.0:<%= @hadoop_jobtracker_thrift_port %></value>
</property>
<% end %>
-<% if has_variable?("hadoop_mapred_jobtracker_plugins") %>
+<% if @hadoop_mapred_jobtracker_plugins %>
<property>
<name>mapred.jobtracker.plugins</name>
- <value><%= hadoop_mapred_jobtracker_plugins %></value>
+ <value><%= @hadoop_mapred_jobtracker_plugins %></value>
<description>Comma-separated list of jobtracker plug-ins to be activated.</description>
</property>
<% end %>
-<% if has_variable?("hadoop_mapred_tasktracker_plugins") %>
+<% if @hadoop_mapred_tasktracker_plugins %>
<property>
<name>mapred.tasktracker.instrumentation</name>
- <value><%= hadoop_mapred_tasktracker_plugins %></value>
+ <value><%= @hadoop_mapred_tasktracker_plugins %></value>
</property>
<% end %>
diff --git a/bigtop-deploy/puppet/modules/hadoop/templates/taskcontroller.cfg b/bigtop-deploy/puppet/modules/hadoop/templates/taskcontroller.cfg
index 99df794e..a1229ad6 100644
--- a/bigtop-deploy/puppet/modules/hadoop/templates/taskcontroller.cfg
+++ b/bigtop-deploy/puppet/modules/hadoop/templates/taskcontroller.cfg
@@ -1,4 +1,4 @@
-mapred.local.dir=<%= mapred_data_dirs.join(",") %>
+mapred.local.dir=<%= @mapred_data_dirs.join(",") %>
hadoop.log.dir=/var/log/hadoop/
mapreduce.tasktracker.group=mapred
min.user.id=0
diff --git a/bigtop-deploy/puppet/modules/hadoop/templates/yarn-site.xml b/bigtop-deploy/puppet/modules/hadoop/templates/yarn-site.xml
index 4433ba9c..0713d971 100644
--- a/bigtop-deploy/puppet/modules/hadoop/templates/yarn-site.xml
+++ b/bigtop-deploy/puppet/modules/hadoop/templates/yarn-site.xml
@@ -18,11 +18,11 @@
<?xml-stylesheet type="text/xsl" href="configuration.xsl"?>
<configuration>
-<% if hadoop_security_authentication == "kerberos" %>
+<% if @hadoop_security_authentication == "kerberos" %>
<!-- JobTracker security configs -->
<property>
<name>yarn.resourcemanager.principal</name>
- <value>yarn/_HOST@<%= kerberos_realm %></value>
+ <value>yarn/_HOST@<%= @kerberos_realm %></value>
</property>
<property>
<name>yarn.resourcemanager.keytab</name>
@@ -31,7 +31,7 @@
<property>
<name>yarn.web-proxy.principal</name>
- <value>yarn/_HOST@<%= kerberos_realm %></value>
+ <value>yarn/_HOST@<%= @kerberos_realm %></value>
</property>
<property>
<name>yarn.web-proxy.keytab</name>
@@ -40,7 +40,7 @@
<property>
<name>yarn.nodemanager.principal</name>
- <value>yarn/_HOST@<%= kerberos_realm %></value>
+ <value>yarn/_HOST@<%= @kerberos_realm %></value>
</property>
<property>
<name>yarn.nodemanager.keytab</name>
@@ -58,22 +58,22 @@
<property>
<name>yarn.web-proxy.address</name>
- <value><%= hadoop_ps_host %>:<%= hadoop_ps_port %></value>
+ <value><%= @hadoop_ps_host %>:<%= @hadoop_ps_port %></value>
</property>
<property>
<name>yarn.resourcemanager.resource-tracker.address</name>
- <value><%= hadoop_rm_host %>:<%= hadoop_rt_port %></value>
+ <value><%= @hadoop_rm_host %>:<%= @hadoop_rt_port %></value>
</property>
<property>
<name>yarn.resourcemanager.address</name>
- <value><%= hadoop_rm_host %>:<%= hadoop_rm_port %></value>
+ <value><%= @hadoop_rm_host %>:<%= @hadoop_rm_port %></value>
</property>
<property>
<name>yarn.resourcemanager.scheduler.address</name>
- <value><%= hadoop_rm_host %>:<%= hadoop_sc_port %></value>
+ <value><%= @hadoop_rm_host %>:<%= @hadoop_sc_port %></value>
</property>
<property>
@@ -98,7 +98,7 @@
<property>
<name>yarn.nodemanager.local-dirs</name>
- <value><%= yarn_data_dirs.join(",") %></value>
+ <value><%= @yarn_data_dirs.join(",") %></value>
<final>true</final>
</property>
diff --git a/bigtop-deploy/puppet/modules/hue/templates/hue.ini b/bigtop-deploy/puppet/modules/hue/templates/hue.ini
index 44e011f7..c23ec19f 100644
--- a/bigtop-deploy/puppet/modules/hue/templates/hue.ini
+++ b/bigtop-deploy/puppet/modules/hue/templates/hue.ini
@@ -41,8 +41,8 @@
secret_key=deaddeafdeaddeafdeaddeafdeaddeafdeaddeafdeaddeafdeaddeaf
# Webserver listens on this address and port
- http_host=<%= hue_host %>
- http_port=<%= hue_port %>
+ http_host=<%= @hue_host %>
+ http_port=<%= @hue_port %>
# Time zone name
time_zone=America/Los_Angeles
@@ -215,7 +215,7 @@
## default_from_email=hue@localhost
-<% if kerberos_realm != "" %>
+<% if @kerberos_realm != "" %>
# Configuration options for Kerberos integration for secured Hadoop clusters
# ------------------------------------------------------------------------
[[kerberos]]
@@ -223,9 +223,9 @@
# Path to Hue's Kerberos keytab file
hue_keytab=/etc/hue.keytab
# Kerberos principal name for Hue
- hue_principal=hue/<%= fqdn %>
+ hue_principal=hue/<%= @fqdn %>
# Path to kinit
- kinit_path=<%= (operatingsystem == 'ubuntu') ? '/usr/bin' : '/usr/kerberos/bin' %>/kinit
+ kinit_path=<%= (@operatingsystem == 'ubuntu') ? '/usr/bin' : '/usr/kerberos/bin' %>/kinit
<% end %>
@@ -261,15 +261,15 @@
[[[default]]]
# Enter the filesystem uri
- fs_defaultfs=<%= default_fs %>
+ fs_defaultfs=<%= @default_fs %>
# Change this if your HDFS cluster is Kerberos-secured
- security_enabled=<%= if (kerberos_realm != "") ; "true" else "false" end %>
+ security_enabled=<%= if (@kerberos_realm != "") ; "true" else "false" end %>
# Use WebHdfs/HttpFs as the communication mechanism.
# This should be the web service root URL, such as
# http://namenode:50070/webhdfs/v1
- webhdfs_url=<%= webhdfs_url %>
+ webhdfs_url=<%= @webhdfs_url %>
# Settings about this HDFS cluster. If you install HDFS in a
# different location, you need to set the following.
@@ -318,14 +318,14 @@
[[[default]]]
# Enter the host on which you are running the ResourceManager
- resourcemanager_host=<%= rm_host %>
+ resourcemanager_host=<%= @rm_host %>
# The port where the ResourceManager IPC listens on
- resourcemanager_port=<%= rm_port %>
+ resourcemanager_port=<%= @rm_port %>
# Whether to submit jobs to this cluster
submit_to=True
# Change this if your YARN cluster is Kerberos-secured
- security_enabled=<%= if (kerberos_realm != "") ; "true" else "false" end %>
+ security_enabled=<%= if (@kerberos_realm != "") ; "true" else "false" end %>
# Settings about this MR2 cluster. If you install MR2 in a
# different location, you need to set the following.
@@ -340,13 +340,13 @@
hadoop_conf_dir=/etc/hadoop/conf
# URL of the ResourceManager API
- resourcemanager_api_url=<%= rm_url %>
+ resourcemanager_api_url=<%= @rm_url %>
# URL of the ProxyServer API
- proxy_api_url=<%= rm_proxy_url %>
+ proxy_api_url=<%= @rm_proxy_url %>
# URL of the HistoryServer API
- history_server_api_url=<%= history_server_url %>
+ history_server_api_url=<%= @history_server_url %>
# URL of the NodeManager API
node_manager_api_url=http://localhost:8042
@@ -359,9 +359,9 @@
[liboozie]
# The URL where the Oozie service runs on. This is required in order for
# users to submit jobs.
- oozie_url=<%= oozie_url %>
+ oozie_url=<%= @oozie_url %>
- security_enabled=<%= if (kerberos_realm != "") ; "true" else "false" end %>
+ security_enabled=<%= if (@kerberos_realm != "") ; "true" else "false" end %>
# Location on HDFS where the workflows/coordinator are deployed when submitted.
remote_deployement_dir=/user/hue/oozie/deployments
diff --git a/bigtop-deploy/puppet/modules/kerberos/templates/kdc.conf b/bigtop-deploy/puppet/modules/kerberos/templates/kdc.conf
index dc35b324..5bc384bf 100644
--- a/bigtop-deploy/puppet/modules/kerberos/templates/kdc.conf
+++ b/bigtop-deploy/puppet/modules/kerberos/templates/kdc.conf
@@ -13,19 +13,19 @@
# See the License for the specific language governing permissions and
# limitations under the License.
-default_realm = <%= realm %>
+default_realm = <%= @realm %>
[kdcdefaults]
# v4_mode = nopreauth
kdc_ports = 0
[realms]
- <%= realm %> = {
- kdc_ports = <%= kdc_port %>
+ <%= @realm %> = {
+ kdc_ports = <%= @kdc_port %>
admin_keytab = /etc/kadm5.keytab
- database_name = <%= kdc_etc_path %>/principal
- acl_file = <%= kdc_etc_path %>/kadm5.acl
- key_stash_file = <%= kdc_etc_path %>/stash
+ database_name = <%= @kdc_etc_path %>/principal
+ acl_file = <%= @kdc_etc_path %>/kadm5.acl
+ key_stash_file = <%= @kdc_etc_path %>/stash
max_life = 10h 0m 0s
max_renewable_life = 7d 0h 0m 0s
master_key_type = des3-hmac-sha1
diff --git a/bigtop-deploy/puppet/modules/kerberos/templates/krb5.conf b/bigtop-deploy/puppet/modules/kerberos/templates/krb5.conf
index ccec5206..10c06c75 100644
--- a/bigtop-deploy/puppet/modules/kerberos/templates/krb5.conf
+++ b/bigtop-deploy/puppet/modules/kerberos/templates/krb5.conf
@@ -14,7 +14,7 @@
# limitations under the License.
[libdefaults]
- default_realm = <%= realm %>
+ default_realm = <%= @realm %>
dns_lookup_realm = false
dns_lookup_kdc = false
ticket_lifetime = 24h
@@ -25,15 +25,15 @@
permitted_enctypes = des-cbc-md5 des-cbc-crc des3-cbc-sha1
[realms]
- <%= realm %> = {
- kdc = <%= kdc_server %>:<%= kdc_port %>
- admin_server = <%= kdc_server %>:<%= admin_port %>
- default_domain = <%= domain %>
+ <%= @realm %> = {
+ kdc = <%= @kdc_server %>:<%= @kdc_port %>
+ admin_server = <%= @kdc_server %>:<%= @admin_port %>
+ default_domain = <%= @domain %>
}
[domain_realm]
- .<%= domain %> = <%= realm %>
- <%= domain %> = <%= realm %>
+ .<%= @domain %> = <%= @realm %>
+ <%= @domain %> = <%= @realm %>
[logging]
kdc = FILE:/var/log/krb5kdc.log
diff --git a/bigtop-deploy/puppet/modules/spark/templates/spark-env.sh b/bigtop-deploy/puppet/modules/spark/templates/spark-env.sh
index 6d4ab2a3..bbd70815 100644
--- a/bigtop-deploy/puppet/modules/spark/templates/spark-env.sh
+++ b/bigtop-deploy/puppet/modules/spark/templates/spark-env.sh
@@ -38,8 +38,8 @@
export SPARK_LAUNCH_WITH_SCALA=0
export SPARK_LIBRARY_PATH=${SPARK_HOME}/lib
export SCALA_LIBRARY_PATH=${SPARK_HOME}/lib
-export SPARK_MASTER_WEBUI_PORT=<%= master_ui_port %>
-export SPARK_MASTER_PORT=<%= master_port %>
+export SPARK_MASTER_WEBUI_PORT=<%= @master_ui_port %>
+export SPARK_MASTER_PORT=<%= @master_port %>
### Comment above 2 lines and uncomment the following if
### you want to run with scala version, that is included with the package
@@ -47,6 +47,6 @@ export SPARK_MASTER_PORT=<%= master_port %>
#export PATH=$PATH:$SCALA_HOME/bin
### change the following to specify a real cluster's Master host
-export STANDALONE_SPARK_MASTER_HOST=<%= master_host %>
+export STANDALONE_SPARK_MASTER_HOST=<%= @master_host %>
diff --git a/bigtop-deploy/vm/docker-puppet/provision.sh b/bigtop-deploy/vm/docker-puppet/provision.sh
index 5bd36058..dfd58b65 100755
--- a/bigtop-deploy/vm/docker-puppet/provision.sh
+++ b/bigtop-deploy/vm/docker-puppet/provision.sh
@@ -17,7 +17,7 @@
# Install puppet agent
yum -y install http://yum.puppetlabs.com/puppetlabs-release-el-6.noarch.rpm
-yum -y install puppet-2.7.24-1.el6.noarch
+yum -y install puppet
sysctl kernel.hostname=`hostname -f`
diff --git a/bigtop-deploy/vm/smoke-tests.sh b/bigtop-deploy/vm/smoke-tests.sh
index d02c128d..9ee111ab 100755
--- a/bigtop-deploy/vm/smoke-tests.sh
+++ b/bigtop-deploy/vm/smoke-tests.sh
@@ -9,7 +9,7 @@ export HIVE_HOME=/usr/lib/hive/
export PIG_HOME=/usr/lib/pig/
export FLUME_HOME=/usr/lib/flume/
export HIVE_CONF_DIR=/etc/hive/conf/
-export JAVA_HOME="/usr/lib/jvm/java-1.7.0-openjdk-1.7.0.65.x86_64/"
+export JAVA_HOME="/usr/lib/jvm/java-openjdk/"
export MAHOUT_HOME="/usr/lib/mahout"
export ITEST="0.7.0"
diff --git a/bigtop-deploy/vm/vagrant-puppet/provision.sh b/bigtop-deploy/vm/vagrant-puppet/provision.sh
index fa0571b3..2b660c2e 100755
--- a/bigtop-deploy/vm/vagrant-puppet/provision.sh
+++ b/bigtop-deploy/vm/vagrant-puppet/provision.sh
@@ -17,7 +17,7 @@
# Install puppet agent
yum -y install http://yum.puppetlabs.com/puppetlabs-release-el-6.noarch.rpm
-yum -y install puppet-2.7.23-1.el6.noarch
+yum -y install puppet
service iptables stop
chkconfig iptables off