forked from ispras/spark-openstack-deprecated
-
Notifications
You must be signed in to change notification settings - Fork 1
/
deploy_templates.py
executable file
·70 lines (59 loc) · 2.58 KB
/
deploy_templates.py
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
#!/usr/bin/env python
# -*- coding: utf-8 -*-
from __future__ import with_statement
import os
import sys
# Deploy the configuration file templates in the spark-openstack/templates directory
# to the root filesystem, substituting variables such as the master hostname,
# ZooKeeper URL, etc as read from the environment.
# Find system memory in KB and compute Spark's default limit from that
mem_command = "cat /proc/meminfo | grep MemTotal | awk '{print $2}'"
master_ram_kb = int(
os.popen(mem_command).read().strip())
# This is the master's memory. Try to find slave's memory as well
first_slave = os.popen("cat /root/spark-openstack/slaves | head -1").read().strip()
slave_mem_command = "ssh -t -o StrictHostKeyChecking=no -o UserKnownHostsFile=/dev/null %s %s" %\
(first_slave, mem_command)
slave_ram_kb = int(os.popen(slave_mem_command).read().strip())
system_ram_kb = min(slave_ram_kb, master_ram_kb)
system_ram_mb = system_ram_kb / 1024
if system_ram_mb > 20*1024:
# Leave 3 GB for the OS, HDFS and buffer cache
spark_mb = system_ram_mb - 3 * 1024
elif system_ram_mb > 10*1024:
# Leave 2 GB for the OS & co.
spark_mb = system_ram_mb - 2 * 1024
else:
# Leave 1.3 GB for the OS & co. Note that this must be more than
# 1 GB because Mesos leaves 1 GB free and requires 32 MB/task.
spark_mb = max(512, system_ram_mb - 1300)
template_vars = {
"master_list": os.getenv("MESOS_MASTERS"),
"active_master": os.getenv("MESOS_MASTERS").split("\n")[0],
"slave_list": os.getenv("MESOS_SLAVES"),
"zoo_list": os.getenv("MESOS_ZOO_LIST"),
"cluster_url": os.getenv("MESOS_CLUSTER_URL"),
"hdfs_data_dirs": os.getenv("MESOS_HDFS_DATA_DIRS"),
"mapred_local_dirs": os.getenv("MESOS_MAPRED_LOCAL_DIRS"),
"spark_local_dirs": os.getenv("MESOS_SPARK_LOCAL_DIRS"),
"default_spark_mem": "%dm" % spark_mb,
"scala_home": os.getenv("SCALA_HOME"),
"java_home": os.getenv("JAVA_HOME")
}
template_dir="/root/spark-openstack/templates"
for path, dirs, files in os.walk(template_dir):
if path.find(".svn") == -1:
dest_dir = os.path.join('/', path[len(template_dir):])
if not os.path.exists(dest_dir):
os.makedirs(dest_dir)
for filename in files:
if filename[0] not in '#.~' and filename[-1] != '~':
dest_file = os.path.join(dest_dir, filename)
with open(os.path.join(path, filename)) as src:
with open(dest_file, "w") as dest:
print "Configuring " + dest_file
text = src.read()
for key in template_vars:
text = text.replace("{{" + key + "}}", template_vars[key])
dest.write(text)
dest.close()