# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unless required by applicable law or agreed to in writing, software
# distributed under the License is distributed on an "AS IS" BASIS,
# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or
# implied.
# See the License for the specific language governing permissions and
# limitations under the License.
from django.utils.translation import ugettext_lazy as _
import openstack_dashboard.contrib.sahara.content.data_processing. \
utils.workflow_helpers as work_helpers
from openstack_dashboard.contrib.sahara.api import sahara as saharaclient
[docs]class Helpers(object):
def __init__(self, request):
self.request = request
def _get_node_processes(self, plugin):
processes = []
for proc_lst in plugin.node_processes.values():
processes += proc_lst
return [(proc_name, proc_name) for proc_name in processes]
[docs] def get_node_processes(self, plugin_name, hadoop_version):
plugin = saharaclient.plugin_get_version_details(self.request,
plugin_name,
hadoop_version)
return self._get_node_processes(plugin)
def _extract_parameters(self, configs, scope, applicable_target):
parameters = []
for config in configs:
if (config['scope'] == scope and
config['applicable_target'] == applicable_target):
parameters.append(work_helpers.Parameter(config))
return parameters
[docs] def get_cluster_general_configs(self, plugin_name, hadoop_version):
plugin = saharaclient.plugin_get_version_details(self.request,
plugin_name,
hadoop_version)
return self._extract_parameters(plugin.configs, 'cluster', "general")
[docs] def get_general_node_group_configs(self, plugin_name, hadoop_version):
plugin = saharaclient.plugin_get_version_details(self.request,
plugin_name,
hadoop_version)
return self._extract_parameters(plugin.configs, 'node', 'general')
[docs] def get_targeted_node_group_configs(self, plugin_name, hadoop_version):
plugin = saharaclient.plugin_get_version_details(self.request,
plugin_name,
hadoop_version)
parameters = {}
for service in plugin.node_processes.keys():
parameters[service] = self._extract_parameters(plugin.configs,
'node', service)
return parameters
[docs] def get_targeted_cluster_configs(self, plugin_name, hadoop_version):
plugin = saharaclient.plugin_get_version_details(self.request,
plugin_name,
hadoop_version)
parameters = {}
for service in plugin.node_processes.keys():
parameters[service] = self._extract_parameters(plugin.configs,
'cluster', service)
return parameters
[docs] def is_from_guide(self):
referer = self.request.environ.get("HTTP_REFERER")
if referer and "/wizard/" in referer:
return True
return False
[docs] def reset_guide(self):
try:
self.request.session.update(
{"plugin_name": None,
"plugin_version": None,
"master_name": None,
"master_id": None,
"worker_name": None,
"worker_id": None,
"guide_cluster_template_name": None})
except Exception:
return False
return True
[docs] def reset_job_guide(self):
try:
self.request.session.update(
{"guide_job_type": None,
"guide_job_name": None,
"guide_job_id": None,
"guide_datasource_id": None,
"guide_datasource_name": None, })
except Exception:
return False
return True
# Map needed because switchable fields need lower case
# and our server is expecting upper case. We will be
# using the 0 index as the display name and the 1 index
# as the value to pass to the server.
JOB_TYPE_MAP = {"pig": [_("Pig"), "Pig"],
"hive": [_("Hive"), "Hive"],
"spark": [_("Spark"), "Spark"],
"storm": [_("Storm"), "Storm"],
"mapreduce": [_("MapReduce"), "MapReduce"],
"mapreduce.streaming": [_("Streaming MapReduce"),
"MapReduce.Streaming"],
"java": [_("Java"), "Java"],
"shell": [_("Shell"), "Shell"]}