Why Gemfury? Push, build, and install  RubyGems npm packages Python packages Maven artifacts PHP packages Go Modules Debian packages RPM packages NuGet packages

Repository URL to install this package:

Details    
Size: Mime:
# vim: tabstop=4 shiftwidth=4 softtabstop=4

# Copyright (c) 2013 TrilioData, Inc.


import logging
import json
import glob
import itertools
import pkgutil
import os
import imp
import pkg_resources
from novaclient import extension as nova_extension
import six
try:
    import __builtin__ as builtins
except ImportError:
    import builtins

from workloadmgrclient.v1 import client as workloadmgrclient

from contego_python_novaclient_ext.contegoclient import client as cc

from django.conf import settings


from horizon import exceptions

from openstack_dashboard.api.base import url_for, APIDictWrapper

from datetime import datetime
from datetime import timedelta
from operator import itemgetter, attrgetter

from openstack_dashboard import api
from openstack_dashboard import policy

import time
import pytz
import threading
import concurrent.futures

try:
    from cgi import parse_qs
except:
    from six.moves.urllib.parse import parse_qs
try:
    from cgi import escape
except:
    from html import escape
from stat import *

from openstack_dashboard.api import nova


LOG = logging.getLogger(__name__)
FOLDER_DELIMITER = "/"


def roundTime(time, modulo):
    if ":" in time:
        time = time.split(":")
        time_new = int(time[1])
    off = divmod(time_new, modulo)
    if off[1] >= 10:
        if off[0] >= 3:
            return str(int(time[0]) + 1).zfill(2) + ":" + "15"
        return time[0] + ":" + str((modulo - off[1] + modulo) + time_new)
    else:
        if off[0] >= 3:
            return str(int(time[0]) + 1).zfill(2) + ":" + "00"
        return time[0] + ":" + str((modulo - off[1]) + time_new)


def sizeof_fmt(num, suffix='B'):
    try:
        for unit in ['', 'K', 'M', 'G', 'T', 'P', 'E', 'Z']:
            if abs(num) < 1024.0:
                return "%3.1f %s%s" % (num, unit, suffix)
            num /= 1024.0
        return "%.1f %s%s" % (num, 'Yi', suffix)
    except Exception as ex:
        LOG.exception(ex)
        return num


def humanize_time(seconds, granularity=5):
    try:
        result = []
        intervals = (
            ('weeks', 604800),  # 60 * 60 * 24 * 7
            ('days', 86400),    # 60 * 60 * 24
            ('hours', 3600),    # 60 * 60
            ('minutes', 60),
            ('seconds', 1),
        )
        for name, count in intervals:
            value = seconds // count
            if value:
                seconds -= value * count
                if value == 1:
                    name = name.rstrip('s')
                result.append("{} {}".format(int(value), name))
        return ' '.join(result[:granularity])
    except Exception as ex:
        LOG.exception(ex)
        return seconds


def get_time_zone(request):
    tz = 'UTC'
    try:
        tz = request._get_cookies()['django_timezone']
    except Exception as ex:
        try:
            tz = request.COOKIES['django_timezone']
        except Exception as ex:
            LOG.debug("Failed to fetch time zone, using default: {}".format(tz))
            LOG.debug(ex)

    return tz


def get_local_time(record_time, input_format, output_format, tz, tz1=None):
    """
    Convert and return the date and time - from GMT to local time
    """
    try:
        if record_time in (0, None, ''):
            return ''
        else:
            if not input_format \
                    or input_format is None \
                    or input_format == '':
                input_format = '%Y-%m-%dT%H:%M:%S.%f'
            if not output_format \
                    or output_format is None \
                    or output_format == '':
                output_format = "%m/%d/%Y %I:%M:%S %p"

            local_time = datetime.strptime(
                record_time, input_format)
            local_tz = pytz.timezone(tz)
            from_zone = pytz.timezone('UTC')
            if tz1 is not None:
                from_zone = pytz.timezone(tz1)
            local_time = local_time.replace(tzinfo=from_zone)
            local_time = local_time.astimezone(local_tz)
            local_time = datetime.strftime(
                local_time, output_format)
            return local_time
    except Exception as ex:
        LOG.exception(ex)
        return record_time


def permissions_to_unix_name(mode):
    modestr = ""
    if S_ISDIR(mode):
        modestr = "d"
    elif S_ISCHR(mode):
        modestr = "c"
    elif S_ISBLK(mode):
        modestr = "b"
    elif S_ISREG(mode):
        modestr = "-"
    elif S_ISFIFO(mode):
        modestr = "f"
    elif S_ISLNK(mode):
        modestr = "l"
    elif S_ISSOCK(mode):
        modestr = "s"

    if mode & S_IRUSR:
        modestr += "r"
    else:
        modestr += "-"

    if mode & S_IWUSR:
        modestr += "w"
    else:
        modestr += "-"

    if mode & S_IXUSR:
        modestr += "x"
    else:
        modestr += "-"

    if mode & S_IRGRP:
        modestr += "r"
    else:
        modestr += "-"

    if mode & S_IWGRP:
        modestr += "w"
    else:
        modestr += "-"

    if mode & S_IXGRP:
        modestr += "x"
    else:
        modestr += "-"

    if mode & S_IROTH:
        modestr += "r"
    else:
        modestr += "-"

    if mode & S_IWOTH:
        modestr += "w"
    else:
        modestr += "-"

    if mode & S_IXOTH:
        modestr += "x"
    else:
        modestr += "-"

    modestr += "."

    return modestr


def get_time_with_time_zone_from_timestamp(timestamp):
    time_stamp_time = datetime.fromtimestamp(timestamp).strftime('%b %d %Y %I:%M')
    return time_stamp_time + " UTC "


def convert_to_unix_format(stats):
    for time in ['mtime', 'atime', 'ctime']:
        if stats.get(time, None) is not None:
            stats[time] = get_time_with_time_zone_from_timestamp(int(stats[time]))

    if stats.get('mode', None) is not None:
        stats['mode'] = permissions_to_unix_name(int(stats.get('mode')))

    if stats.get('size', None) is not None:
        stats['size'] = sizeof_fmt(int(stats.get('size')))

    res_list = ["mode", "nlink", "uid", "gid", "size", "ctime", "mtime", "atime"]

    return [stats.get(key, None) for key in res_list]


def workload_api(request):
    insecure = getattr(settings, 'OPENSTACK_SSL_NO_VERIFY', False)
    cacert = getattr(settings, 'OPENSTACK_SSL_CACERT', "")
    endpoint_type = getattr(settings, 'OPENSTACK_ENDPOINT_TYPE', 'publicURL')
    workloadmgr_url = ""
    try:
        workloadmgr_url = url_for(request, 'workloads', endpoint_type)
    except exceptions.ServiceCatalogException:
        LOG.debug('no workloads service configured.')
        return None

    LOG.debug('workloadmgr connection created using token "%s" and url "%s"' %
              (request.user.token.id, workloadmgr_url))
    c = workloadmgrclient.Client(request.user.username,
                                 request.user.token.id,
                                 project_id=request.user.tenant_id,
                                 auth_url=workloadmgr_url,
                                 insecure=insecure,
                                 cacert=cacert,
                                 http_log_debug=settings.DEBUG)
    c.client.auth_token = request.user.token.id
    c.client.management_url = workloadmgr_url
    return c


def settings_create(request, context):
    try:
        workload_api(request).settings.create_bulk(context)
    except Exception as ex:
        raise Exception(str(ex))


def setting_get(request, name):
    try:
        setting = workload_api(request).settings.get(name)
        return setting
    except Exception as ex:
        raise Exception(str(ex))


def settings_list(request, get_hidden=False, get_smtp_settings=False):
    try:
        search_opts = {'get_hidden': get_hidden, 'get_smtp_settings': get_smtp_settings}
        settings = workload_api(request).settings.list(search_opts=search_opts)
        return settings
    except Exception as ex:
        raise Exception(str(ex))


def test_email(request):
    try:
        message = workload_api(request).settings.test_email()
        return message
    except Exception as ex:
        raise Exception(str(ex))


def workload_type_get(request, workload_type_id):
    try:
        workload_type = workload_api(request).workload_types.get(workload_type_id)
        return workload_type
    except Exception as ex:
        raise Exception(str(ex))


def workload_type_list(request):
    try:
        workload_types = workload_api(request).workload_types.list()
        return workload_types
    except Exception as ex:
        raise Exception(str(ex))


def workload_get_nodes(request):
    try:
        nodes = workload_api(request).workloads.get_nodes()
        return nodes
    except Exception as ex:
        raise Exception(str(ex))


def workload_type_create(request, context):
    try:
        workload_type = workload_api(request).workload_types.create(context['metadata'],
                                                                    context['name'],
                                                                    context['description'])
        return workload_type
    except Exception as ex:
        raise Exception(str(ex))


def workload_type_delete(request, workload_type_id):
    try:
        workload_api(request).workload_types.delete(workload_type_id)
        return True
    except Exception as ex:
        raise Exception(str(ex))


def workload_get(request, workload_id):
    try:
        tz = get_time_zone(request)
        workload = workload_api(request).workloads.get(workload_id,
                                                       search_opts={'scheduler_trust': True})
        workload.workload_type_id = (workload_type_get(request, workload.workload_type_id)).name
        workload.created_at = get_local_time(workload.created_at, '', '', tz)
        workload.updated_at = get_local_time(workload.updated_at, '', '', tz)
        nextrun = "NA"
        if 'enabled' in workload.jobschedule and workload.jobschedule['enabled']:
            nextrun = 'nextrun' in workload.jobschedule and \
                      int(workload.jobschedule['nextrun']) or -1
            workload.jobschedule['nextrun'] = humanize_time(nextrun, 5)
        for instance in workload.instances:
            hw_qemu_guest_agent = False
            try:
                server = api.nova.server_get(request, instance['id'])
                hw_qemu_guest_agent = server.metadata.get('hw_qemu_guest_agent', False)
                if hw_qemu_guest_agent is False:
                    if server.image:
                        image = api.glance.image_get(request, server.image['id'])
                        hw_qemu_guest_agent = image.properties.get('hw_qemu_guest_agent', 'no') == 'yes'
                instance['hw_qemu_guest_agent'] = hw_qemu_guest_agent
            except Exception as ex:
                LOG.exception(ex)
        return workload
    except Exception as ex:
        raise Exception(str(ex))


def workload_list(request):
    try:
        """
        id = request.GET.get('page',None)
        if id is None:
             workloads = workload_api(request).workloads.list_by_page('1')
        else:
             workloads = workload_api(request).workloads.list_by_page(id)
        """
        workloads = workload_api(request).workloads.list(search_opts={'scheduler_trust':True})
        search_opts = {}
        snapshot_list = workload_api(request).snapshots. \
            list(detailed=False, search_opts=search_opts)

        tz = get_time_zone(request)
        workload_snapshot_info = {}
        for workload in workloads:
            workload_snapshot_info[workload.id] = {
                'failed_snapshots_count': 0,
                'success_snapshots_count': 0}

        for snapshot in snapshot_list:
            info = workload_snapshot_info[snapshot.workload_id]
            if snapshot.status in ("error", "cancelled"):
                info['failed_snapshots_count'] += 1
            elif snapshot.status in ("available", "mounted"):
                info['success_snapshots_count'] += 1

        for workload in workloads:
            info = workload_snapshot_info[workload.id]
            failed_snapshots_count = info['failed_snapshots_count']
            success_snapshots_count = info['success_snapshots_count']
            workload.snapshots_info = "Total:" \
                                      + str(failed_snapshots_count + success_snapshots_count) \
                                      + ", Success:" + str(success_snapshots_count) \
                                      + ", Error:" + str(failed_snapshots_count)
            workload.created_at = get_local_time(
                workload.created_at, '', "%I:%M:%S %p - %m/%d/%Y", tz)
            workload.workload_type_id = \
                (workload_type_get(request, workload.workload_type_id)).name
            workload.trust='NA'
            if hasattr(workload, 'scheduler_trust') and workload.scheduler_trust['scheduler_enabled']:
                workload.trust= workload.scheduler_trust['is_valid']
            
        return workloads
    except Exception as ex:
        raise Exception(str(ex))


def file_search(request, vm_id, filepath, snapshotids, start, end):
    try:
        search = workload_api(request).file_search.search(vm_id, filepath, snapshotids=snapshotids, start=start,
                                                          end=end)
        return search
    except Exception as ex:
        raise Exception(str(ex))


def file_search_show(request, search_id):
    try:
        search = workload_api(request).file_search.get(search_id)
        return search
    except Exception as ex:
        raise Exception(str(ex))


def page_size_get(request):
    try:
        return workload_api(request).workloads.settings('page_size')
    except Exception as ex:
        return Exception(str(ex))


def workload_vms_list(request):
    try:
        protected_vms = workload_api(request).workloads.get_protected_vms()
        return protected_vms['protected_vms']
    except Exception as ex:
        raise Exception(str(ex))


def workload_create(request, context):
    metadata = {}
    if 'metadata' in context:
        metadata = context['metadata']

    if 'jobschedule' not in context:
        context['jobschedule'] = {}

    try:
        workload = workload_api(request).workloads.create(context['name'],
                                                          context['description'], context['workload_type'],
                                                          'openstack', context['instances'],
                                                          context['jobschedule'], metadata)
        return workload
    except Exception as ex:
        raise Exception(str(ex))


def workload_update(request, workload_id, context, is_admin_dashboard=False):
    metadata = {}
    if 'metadata' in context:
        metadata = context['metadata']

    if 'jobschedule' not in context:
        context['jobschedule'] = {}
    try:
        workload = workload_api(request).workloads.update(
            workload_id, context['name'],
            context['description'], context['instances'],
            context['jobschedule'], metadata, is_admin_dashboard)
        return workload
    except Exception as ex:
        raise Exception(str(ex))


def workload_pause(request, workload_id):
    try:
        return workload_api(request).workloads.pause(workload_id)
    except Exception as ex:
        raise Exception(str(ex))


def workload_resume(request, workload_id):
    try:
        return workload_api(request).workloads.resume(workload_id)
    except Exception as ex:
        raise Exception(str(ex))


def workload_snapshot(request, workload_id, full):
    try:
        return workload_api(request).workloads.snapshot(workload_id, full)
    except Exception as ex:
        raise Exception(str(ex))


def workload_snapshot(request, workload_id, full, name, description):
    try:
        return workload_api(request).workloads.snapshot(workload_id, full, name, description)
    except Exception as ex:
        raise Exception(str(ex))


def workload_unlock(request, workload_id):
    try:
        workload_api(request).workloads.unlock(workload_id)
        return True
    except Exception as ex:
        raise Exception(str(ex))


def workload_reset(request, workload_id):
    try:
        workload_api(request).workloads.reset(workload_id)
        return True
    except Exception as ex:
        raise Exception(str(ex))


def workload_delete(request, workload_id):
    try:
        workload_api(request).workloads.delete(workload_id)
        return True
    except Exception as ex:
        raise Exception(str(ex))


def snapshot_get(request, snapshot_id):
    try:
        tz = get_time_zone(request)
        snapshot = workload_api(request).snapshots.get(snapshot_id)
        snapshot.size = sizeof_fmt(snapshot.size)
        snapshot.restore_size = sizeof_fmt(snapshot.restore_size)
        snapshot.created_at = get_local_time(snapshot.created_at, '', '', tz)
        snapshot.updated_at = get_local_time(snapshot.updated_at, '', '', tz)
        for instance in snapshot.instances:
            instance['metadata']['vm_metadata'] = json.loads(instance['metadata'].get('vm_metadata', "{}"))
            for vdisk in instance['vdisks']:
                vdisk['restore_size'] = sizeof_fmt(vdisk['restore_size'])
                # instance hw_qemu_guest_agent flag takes precedence
                if 'hw_qemu_guest_agent' in vdisk:
                    if 'hw_qemu_guest_agent' not in instance['metadata']['vm_metadata']:
                        instance['metadata']['vm_metadata']['hw_qemu_guest_agent'] = vdisk['hw_qemu_guest_agent']
        return snapshot
    except Exception as ex:
        raise Exception(str(ex))


def next_snapshot_by_id(request, workload_id, snapshot_id):
    """

    Args:
        request: HttpRequest Object
        workload_id: int, uuid
        snapshot_id: int, uuid

    Returns: int -> snapshot_id

    """
    snap_obj = workload_api(request).snapshots.get_next_snapshot_id(workload_id,
                                                                   snapshot_id)
    return snap_obj.id

def prev_snapshot_by_id(request, workload_id, snapshot_id):
    """

    Args:
        request: HttpRequest Object
        workload_id: int, uuid
        snapshot_id: int, uuid

    Returns: int -> snapshot_id

    """
    snap_obj = workload_api(request).snapshots.get_prev_snapshot_id(workload_id,
                                                                   snapshot_id)
    return snap_obj.id

def snapshot_list(request, workload_id, status=None, filter_it=True, get_instances=False):
    try:
        tz = get_time_zone(request)
        search_opts = {'workload_id': workload_id}
        if status is not None:
            search_opts['status'] = status
        if get_instances is True:
            search_opts['get_instances'] = get_instances
        snapshots = workload_api(request).snapshots \
            .list(detailed=True, search_opts=search_opts)
        snapshots = builtins.sorted(snapshots,
                           key=attrgetter('created_at'), reverse=True)
        if filter_it is False:
            return snapshots
        if len(snapshots) > 0:
            for snapshot in snapshots:
                restores_list = []
                success_restores_count = 0
                failed_restores_count = 0
                try:
                    search_opts = {'snapshot_id': snapshot.id}
                    restores_list = workload_api(request).restores.list(
                        detailed=True, search_opts=search_opts)
                    if len(restores_list) > 0:
                        for restore in restores_list:
                            if restore.status == "error" or restore.status == "cancelled":
                                failed_restores_count += 1
                            elif restore.status == "available":
                                success_restores_count += 1
                except Exception as ex:
                    LOG.exception(ex)
                snapshot.restores_info = "Total:" + str(len(restores_list)) \
                                         + ", Success:" + str(success_restores_count) \
                                         + ", Error:" + str(failed_restores_count)
                snapshot.size = sizeof_fmt(snapshot.size)
                snapshot.created_at = get_local_time(
                    snapshot.created_at, '', "%I:%M:%S %p - %m/%d/%Y", tz)
        return snapshots
    except Exception as ex:
        raise Exception(str(ex))


def snapshot_restore(request, snapshot_id, test=False, options=None):
    try:
        options = options or {'name': 'one click restore',
                              'description': '-',
                              "oneclickrestore": True,
                              "openstack": {},
                              "type": "openstack"}
        workload_api(request).snapshots.restore(snapshot_id, test,
                                                name=options['name'],
                                                description=options['description'],
                                                options=options)
        return True
    except Exception as ex:
        raise Exception(str(ex))


def snapshot_mount(request, snapshot_id, vm_id):
    try:
        return workload_api(request).snapshots.mount(snapshot_id, vm_id)
    except Exception as ex:
        raise Exception(str(ex))


def snapshot_dismount(request, snapshot_id):
    try:
        workload_api(request).snapshots.dismount(snapshot_id)
        return True
    except Exception as ex:
        raise Exception(str(ex))


def snapshot_delete(request, snapshot_id):
    try:
        workload_api(request).snapshots.delete(snapshot_id)
        return True
    except Exception as ex:
        raise Exception(str(ex))


def snapshot_cancel(request, snapshot_id):
    try:
        workload_api(request).snapshots.cancel(snapshot_id)
        return True
    except Exception as ex:
        raise Exception(str(ex))


def restore_get(request, restore_id):
    try:
        tz = get_time_zone(request)
        restore = workload_api(request).restores.get(restore_id)
        restore.size = sizeof_fmt(restore.size)
        restore.created_at = get_local_time(restore.created_at, '', '', tz)
        restore.updated_at = get_local_time(restore.updated_at, '', '', tz)
        restore.time_taken = humanize_time(restore.time_taken)
        return restore
    except Exception as ex:
        raise Exception(str(ex))


def restore_list(request, snapshot_id):
    try:
        tz = get_time_zone(request)
        search_opts = {'snapshot_id': snapshot_id}
        restores = workload_api(request).restores \
            .list(detailed=True, search_opts=search_opts)
        restores = builtins.sorted(restores,
                          key=attrgetter('created_at'), reverse=True)
        for restore in restores:
            restore.size = sizeof_fmt(restore.size)
            restore.created_at = get_local_time(
                restore.created_at, '', "%I:%M:%S %p - %m/%d/%Y", tz)
        return restores
    except Exception as ex:
        raise Exception(str(ex))


def restore_delete(request, restore_id):
    try:
        workload_api(request).restores.delete(restore_id)
        return True
    except Exception as ex:
        raise Exception(str(ex))


def restored_instances_list(request, restore_id):
    return []


def restored_instance_get(request, restored_instance_id):
    try:
        return {'id': 'id',
                'vm_id': 'vm_id',
                'name': 'name',
                'status': 'available'}
    except Exception as ex:
        raise Exception(str(ex))


def get_user_name(request, user_id):
    user_name = user_id
    if policy.check((("identity", "identity:get_user"),), request):
        try:
            user = api.keystone.user_get(request, user_id)
            if user:
                user_name = user.username
        except Exception as ex:
            LOG.exception(ex)
    else:
        LOG.debug("Insufficient privilege level to view user information.")
    return user_name


def get_project_name(request, project_id):
    project_name = project_id
    try:
        project_info = \
            api.keystone.tenant_get(request, project_id, admin=True)
        if project_info:
            project_name = project_info.name
    except Exception as ex:
        LOG.exception(ex)
    return project_name


def get_project_list(request):
    project_list = []
    try:
        project_list = \
            api.keystone.tenant_list(request)
    except Exception as ex:
        LOG.exception(ex)
    return project_list


def dashboard_nfs_workloads_data(request):
    try:
        workloads_list = []
        var = parse_qs(request.environ['QUERY_STRING'])
        nfs = escape(var.get('nfs', [''])[0])
        try:
            search_opts = {'all_workloads': True, 'nfs_share': nfs}
            workloads = workload_api(request).workloads.list(search_opts=search_opts)
        except:
            workloads = []

        return workloads
    except Exception as ex:
        raise Exception(str(ex))


def dashboard_workloads_data_per_tenant(request, project_id=None):
    try:
        # Get list of all workloads data for csv export
        workloads_list = []
        try:
            search_opts = {'all_workloads': True, 'project_id': project_id}
            workloads = workload_api(request).workloads.list(detailed=True, search_opts=search_opts)
            known_project_names = {}
            known_user_names = {}
            known_types = {}
            for wld in workloads:
                workload_details = {}
                # Fetching the names from dict instead of calling api every time
                if wld.id not in known_user_names:
                    user_name = get_user_name(request, wld.user_id)
                    known_user_names[wld.user_id] = user_name
                else:
                    user_name = known_user_names[wld.user_id]

                if wld.project_id not in known_project_names:
                    project_name = get_project_name(request, wld.project_id)
                    known_project_names[wld.project_id] = project_name
                else:
                    project_name = known_project_names[wld.project_id]

                if wld.workload_type_id not in known_types:
                    wlm_type = (workload_type_get(request, wld.workload_type_id)).name
                    known_types[wld.workload_type_id] = wlm_type
                else:
                    wlm_type = known_types[wld.workload_type_id]

                if wld.jobschedule['enabled']:
                    scheduler_status = 'Enabled'
                else:
                    scheduler_status = 'Disabled'

                workload_details = {'User Id': wld.id, 'User Name': user_name,
                                    'Project': project_name,
                                    'Workload Name': wld.name,
                                    'Workload Type': wlm_type,
                                    'Availability Zone': wld.availability_zone,
                                    'VMs': len(wld.instances),
                                    'Storage Used': sizeof_fmt(float(wld.storage_usage['full']['usage'])
                                                               + float(wld.storage_usage['incremental']['usage'])),
                                    'Backup Target': wld.metadata.get('backup_media_target', "NA"),
                                    'Scheduler Status': scheduler_status,
                                    'Interval': wld.jobschedule['interval']}

                workloads_list.append(workload_details)
        except:
            workloads_list = []
        return workloads_list
    except Exception as ex:
        raise Exception(str(ex))


def dashboard_workloads_data(request, project_id=None):
    try:
        # Get list of all workloads
        _workload_list = []
        _workloads = []
        total_workload_len = 0

        # Method to get list of workloads
        def list_workloads(search_opts):
            return workload_api(request).workloads.list(detailed=True, search_opts=search_opts)
        try:
            search_opts = {'all_workloads': True, 'project_id': project_id}
            workloads_all = workload_api( request ).workloads.list(detailed=False, search_opts = search_opts)
            total_workload_len = len(workloads_all)

            page_size = int(request.GET.get('length'))
            if page_size == -1 or page_size > total_workload_len :
                page_size = total_workload_len
            # Batch size to fetch list of workloads in batches
            batch_size = min(10, page_size)

            page_start = int(request.GET.get('start'))
            page_number = page_start // batch_size + 1

            if page_size + page_start > total_workload_len:
               page_size = total_workload_len

            s_start = page_number if page_number < total_workload_len else 1
            if page_size%batch_size:
                s_end = min(s_start + page_size//batch_size, 1 + (total_workload_len//batch_size)) + 1
            else:
                s_end = min(s_start + page_size//batch_size, 1 + (total_workload_len//batch_size))

            search_opts = [{'all_workloads': True,
                            'project_id': project_id,
                            'page_size': batch_size,
                            'page_number': p_num} for p_num in range(s_start, s_end)]
            # Fetch the list of workloads in multiple threads ( max 20 )
            executor = concurrent.futures.ThreadPoolExecutor(20)
            results = executor.map(list_workloads, search_opts)

            for value in results:
                _workloads.extend(value)
            executor.shutdown(wait=True)
        except:
            _workloads = []

        var = parse_qs(request.environ['QUERY_STRING'])
        draw = escape(var.get('draw', [''])[0])
        total = len(_workloads)
        draw = int(draw)

        def create_list_format(index):
            workload_details = []
            workload_details.append(_workloads[index].id)
            workload_details.append(get_user_name(request, _workloads[index].user_id))
            workload_details.append(get_project_name(request, _workloads[index].project_id))
            workload_details.append(_workloads[index].name) 
            workload_details.append((workload_type_get(request, _workloads[index].workload_type_id)).name)
            workload_details.append(_workloads[index].availability_zone)
            workload_details.append(len( _workloads[index].instances ))

            performance_values_size_line = []
            performance_values_time_bar = []
            performance_colors = []
            performance_tooltips_size_line = []
            performance_tooltips_time_bar = []
            success_snapshots_count = 0
            failed_snapshots_count = 0
            full_snapshot = _workloads[index].storage_usage['full']['snap_count'] 
            incr_snapshot = _workloads[index].storage_usage['incremental']['snap_count']
            if full_snapshot + incr_snapshot > 0:
               search_opts = { 'workload_id': _workloads[index].id , 'all': True}
               try:
                    snapshot_list = workload_api(request).snapshots.\
                        list(detailed = True, search_opts = search_opts )
               except:
                       snapshot_list = []

               e1 = len(snapshot_list)
               s1 =  0
               graph_start = s1
               if e1 > 25:
                  graph_start = e1 - 25
               for index1 in range(s1, e1):
                    try:
                        snapshot = snapshot_list[index1]
                    except IndexError:
                           continue
                    if snapshot.status == "error":
                        failed_snapshots_count += 1
                    elif snapshot.status == "available" or snapshot.status == "mounted":
                        success_snapshots_count += 1
                    time_taken = snapshot.time_taken
                    size = snapshot.size
                    size_humanized = sizeof_fmt(size)
                    timetaken_humanized = humanize_time(time_taken)
                    if (snapshot.status == "available"  or snapshot.status == "mounted") and index1 >= graph_start:
                       performance_tooltips_time_bar.append(timetaken_humanized)
                       performance_tooltips_size_line.append(size_humanized)
                       performance_values_time_bar.append(time_taken)
                       performance_values_size_line.append(size)
               table_data = ""
               try:
                   table_data = '<table cellpadding="0" cellspacing="0" width="100%"> \
                                 <tr><td class="t_inner_cell t_first_column"><div class="sparkline_performance" \
                                 data-values-bar="'+str(performance_values_time_bar)+'" \
                                 data-values-line="'+str(performance_values_size_line)+'" \
                                 data-tooltips-line="'+str(performance_tooltips_size_line)+'" \
                                 data-tooltips-bar="'+str(performance_tooltips_time_bar)+'"></div></td><td class="t_inner_cell t_column_50"> \
                                 <div class="sparkline_bar">'+str(full_snapshot)+',\
                                 '+str(incr_snapshot)+'</div></td> \
                                 <td class="t_inner_cell t_column_50"><div class="t_badge alert-success"  title="Successful snapshots"> \
                                 '+str(success_snapshots_count)+'</div></td><td class="t_inner_cell t_column_50"><div class="t_badge alert-danger" \
                                 title="Failed snapshots">'+str(failed_snapshots_count)+'</div></td></tr></table>'
               except Exception as ex:
                      LOG.exception(ex)

               workload_details.append([table_data])
            else:
                 table_data =  '<table cellpadding="0" cellspacing="0" width="100%"> \
                                <tr><td style="border:0">&nbsp;</td><td style="border:0">&nbsp;</td> \
                                <td class="t_inner_cell t_column_50"><div class="t_badge alert-success"  title="Successful snapshots"> \
                                '+str(success_snapshots_count)+'</div></td><td class="t_inner_cell t_column_50"><div class="t_badge alert-danger" \
                                title="Failed snapshots">'+str(failed_snapshots_count)+'</div></td></tr></table>'
                 workload_details.append([table_data])

            workload_details.append(sizeof_fmt(float(_workloads[index].storage_usage['full']['usage'])
                    + float(_workloads[index].storage_usage['incremental']['usage'])))
            workload_details.append(_workloads[index].metadata.get('backup_media_target', "NA"))
            if _workloads[index].jobschedule['enabled']:
                workload_details.append(humanize_time( _workloads[index].jobschedule['nextrun'], 5 ))
                workload_details.append(_workloads[index].jobschedule[ 'interval' ])
                action_data = '<label class="switch"><input type="checkbox"  id="check_'+_workloads[index].id+'" checked data-record-id="'+_workloads[index].id+'#single_row#0" data-toggle="modal" data-target="#confirm-scheduler-update" style="display:none;"><span class="slider round"></span></label>'
                workload_details.append(action_data)
            else:
                workload_details.append('-')
                workload_details.append('disabled')
                action_data = '<label class="switch"><input type="checkbox" id="check_'+_workloads[index].id+'" data-record-id="'+_workloads[index].id+'#single_row#1" data-toggle="modal" data-target="#confirm-scheduler-update" style="display:none;"><span class="slider round"></span></label>'
                workload_details.append(action_data)
            return workload_details

        if len(_workloads) > 0:
            executor = concurrent.futures.ThreadPoolExecutor(20)
            s = range(len(_workloads))
            results = executor.map(create_list_format, s)
            for value in results:
                _workload_list.append(value)
            executor.shutdown(wait=True)
        output = {}
        output['draw'] = draw
        output['recordsTotal'] = len(_workloads)
        output['recordsFiltered'] = total_workload_len
        output['data'] = _workload_list
        return output

    except Exception as ex:
        raise Exception(str(ex))

"""
def dashboard_quota_data(request, tenant_id=None):
    tz = get_time_zone(request)
    node_wise_snapshot_count = ''
    nodes = {}
    quota_data = {'total_nodes': 0, 'available_nodes': 0}
    try:
        nodes = workload_api(request).workloads.get_nodes()
    except Exception:
        nodes['nodes'] = {}
    total_snapshots = 0
    for node in nodes['nodes']:
        if not node['is_vip']:
            quota_data['total_nodes'] += 1
            if isinstance(node['status'], six.string_types) and node['status'].lower() == "up":
                quota_data['available_nodes'] += 1
            search_opts = {'host': node['node'], 'all': True, 'status': 'running'}
            try:
                snapshots = workload_api(request).snapshots \
                    .list(detailed=True, search_opts=search_opts)
            except Exception:
                snapshots = []
            node['snapshots'] = []
            for snapshot in snapshots:
                if tenant_id and snapshot.project_id == tenant_id:
                    node['snapshots'].append(snapshot)
                else:
                    node['snapshots'].append(snapshot)

            total_snapshots += len(node['snapshots'])
            node_wise_snapshot_count = \
                node_wise_snapshot_count + str(len(node['snapshots'])) + ', '

    quota_data['balance_nodes'] = quota_data['total_nodes'] - quota_data['available_nodes']
    quota_data['total_snapshots'] = total_snapshots
    quota_data['node_wise_snapshot_count'] = node_wise_snapshot_count[:2]

    storage = []
    try:
        storage_usage = workload_api(request).workloads.get_storage_usage()
        storage = storage_usage['storage_usage']
    except Exception as ex:
        LOG.exception(ex)

    quota_data['total_utilization'] = 0
    quota_data['total_capacity'] = 0
    for nfsshare in storage:
        quota_data['storage_type'] = str(nfsshare['storage_type'])
        quota_data['total_utilization'] += nfsshare['total_utilization']
        quota_data['total_capacity'] += nfsshare['total_capacity']

    quota_data['available_capacity'] = \
        (float(quota_data['total_capacity']) -
         float(quota_data['total_utilization']))
    quota_data['total_capacity_humanized'] = \
        sizeof_fmt(quota_data['total_capacity'])
    quota_data['total_utilization_humanized'] = \
        sizeof_fmt(quota_data['total_utilization'])
    quota_data['available_capacity_humanized'] = \
        sizeof_fmt(quota_data['available_capacity'])

    quota_data['contego_services_up'] = 0
    quota_data['contego_services_down'] = 0
    quota_data['contego_services_others'] = 0
    quota_data['contego_services_total'] = 0
    try:
        services = nova.service_list(request)
    except Exception:
        services = []
    for service in services:
        if 'contego' in service.binary:
            quota_data['contego_services_total'] += 1
            service.updated_at = get_local_time(service.updated_at, '', '', tz)
            if service.state == "up":
                quota_data['contego_services_up'] += 1
            elif service.state == "down":
                quota_data['contego_services_down'] += 1
            else:
                quota_data['contego_services_others'] += 1

    quota_data['vms_protected'] = 0
    quota_data['total_vms'] = 0
    try:
        chrageback_data = workload_api(request).workloads.get_tenants_usage()
        global_usage = chrageback_data['global_usage']
        tenant_usage = chrageback_data['tenants_usage']
        if tenant_id is not None:
            for tenant in tenant_usage:
                if tenant == tenant_id:
                    quota_data['vms_protected'] = tenant_usage[tenant]['vms_protected']
                    quota_data['total_vms'] = tenant_usage[tenant]['total_vms']
                    quota_data['storage_used'] = tenant_usage[tenant]['used_capacity']
                    quota_data['total_utilization'] = tenant_usage[tenant]['used_capacity']
                    quota_data['total_capacity'] = global_usage['total_capacity']
                    quota_data['total_capacity_humanized'] = sizeof_fmt(quota_data['total_capacity'])
                    quota_data['total_utilization_humanized'] = sizeof_fmt(quota_data['total_utilization'])
            quota_data['tenant_name'] = 'by ' + get_project_name(request, tenant_id) + ' Tenant'
        else:
            quota_data['vms_protected'] = global_usage['vms_protected']
            quota_data['total_vms'] = global_usage['total_vms']
            quota_data['tenant_name'] = ''
    except Exception as ex:
        LOG.exception(ex)
    return [quota_data]
"""


def dashboard_quota_data(request, project_id=None):
    return workload_api(request).workloads.get_quota_data(project_id)


def dashboard_nodes_data(request):
    try:
        node_wise_snapshot_count = ''
        total_snapshots = 0
        nodes = {}
        try:
            nodes = workload_api(request).workloads.get_nodes()
        except:
            nodes['nodes'] = {}
        available_nodes = 0
        for node in nodes['nodes']:
            search_opts = {'host': node['node']}
            try:
                snapshots = workload_api(request).snapshots \
                    .list(detailed=True, search_opts=search_opts)
            except:
                snapshots = []
            current_snapshots = []
            for snapshot in snapshots:
                if snapshot.status != "available" \
                        and snapshot.status != "error" \
                        and snapshot.status != "mounted" \
                        and snapshot.status != "cancelled":
                    current_snapshots.append(snapshot)
            node['snapshots'] = current_snapshots
        nodes_data = {}
        nodes_data['nodeslist'] = nodes['nodes']
        nodes_data['total_nodes'] = len(nodes['nodes'])
        return nodes_data
    except Exception as ex:
        raise Exception(str(ex))


storage_data = {}


def dashboard_storage_usage(request):
    return workload_api(request).workloads.get_storage_usage();


def dashboard_recent_tasks(
        request, timeInMinutes, tasksStatus, tasksPage, tasksSize):
    tz = get_time_zone(request)
    try:
        try:
            tasks = workload_api(request).workloads.get_tasks(
                timeInMinutes, tasksStatus, tasksPage, tasksSize)
        except:
            tasks = []
        for task in tasks['tasks']:
            task['created_at'] = get_local_time(task['created_at'], '', '', tz)
            task['finished_at'] = get_local_time(task['finished_at'], '', '', tz)
        return tasks['tasks']
    except Exception as ex:
        raise Exception(str(ex))


def dashboard_recent_activities(request, timeinminutes):
    try:
        try:
            activities = workload_api(request).workloads. \
                get_recentactivities(timeinminutes)
        except:
            activities = []
        tz = get_time_zone(request)
        for activity in activities['recentactivites']:
            activity['activity_time'] = \
                get_local_time(activity['activity_time'], '', '%I:%M:%S %p - %m/%d/%Y', tz)
            activity['user_name'] = \
                get_user_name(request, activity['object_user_id'])
            activity['project_name'] = \
                get_project_name(request, activity['object_project_id'])
        return activities['recentactivites']
    except Exception as ex:
        raise Exception(str(ex))


def dashboard_audit_log(request, time_in_minutes, start_date, end_date):
    try:
        audit = workload_api(request).workloads. \
            get_auditlog(time_in_minutes, start_date, end_date)
        return audit['auditlog']
    except Exception as ex:
        raise Exception(str(ex))

def _discover_extensions(version):
    extensions = []
    for name, module in itertools.chain(
        _discover_via_python_path(),
            _discover_via_contrib_path(version),
            _discover_via_entry_points()):

        extension = nova_extension.Extension(name, module)
        extensions.append(extension)

    return extensions


def _discover_via_python_path():
    for (module_loader, name, _ispkg) in pkgutil.iter_modules():
        if name.endswith('_python_novaclient_ext'):
            if not hasattr(module_loader, 'load_module'):
                # Python 2.6 compat: actually get an ImpImporter obj
                module_loader = module_loader.find_module(name)

            module = module_loader.load_module(name)
            if hasattr(module, 'extension_name'):
                name = module.extension_name

            yield name, module

def _discover_via_contrib_path(version):
    module_path = os.path.dirname(os.path.abspath(__file__))
    version_str = "v%s" % version.replace('.', '_')
    ext_path = os.path.join(module_path, version_str, 'contrib')
    ext_glob = os.path.join(ext_path, "*.py")

    for ext_path in glob.iglob(ext_glob):
        name = os.path.basename(ext_path)[:-3]

        if name == "__init__":
            continue

        module = imp.load_source(name, ext_path)
        yield name, module


def _discover_via_entry_points():
    for ep in pkg_resources.iter_entry_points('novaclient.extension'):
        name = ep.name
        module = ep.load()

        yield name, module

def datamover_api(request):
    CONTEGOCLIENT_VERSION = "2"
    insecure = getattr(settings, 'OPENSTACK_SSL_NO_VERIFY', False)
    cacert = getattr(settings, 'OPENSTACK_SSL_CACERT', "")
    endpoint_type = getattr(settings, 'OPENSTACK_ENDPOINT_TYPE', 'publicURL')
    extensions = _discover_extensions('1.1')
    try:
        management_url = url_for(request, 'datamover', endpoint_type)
    except exceptions.ServiceCatalogException:
        LOG.debug('no datamover service configured.')
        return None

    LOG.debug('datamover connection created using token "%s" and url "%s"' %
              (request.user.token.id, management_url))
    args = {
        'project_id': request.user.tenant_id,
        'auth_token': request.user.token.id,
        'service_type': 'datamover',
        'username': request.user.username,
        'extensions': extensions,
        'http_log_debug': settings.DEBUG,
        'cacert': cacert,
        'insecure': insecure,
        'endpoint_type': endpoint_type
    }

    client = cc.Client(CONTEGOCLIENT_VERSION, **args)
    client.client.set_management_url(management_url)
    return client


def dashboard_contego_services(request):
    try:
        services_data = {}
        contego_services = []
        try:
            services = datamover_api(request).contego.get_service_list()
        except Exception:
            services = []
        tz = get_time_zone(request)
        if len(services) > 1:
            for service in services[1]["services"]:
                service["updated_at"] = get_local_time(service["updated_at"], '', '', tz)
                contego_services.append(service)
            services_data['contego_services'] = contego_services
    except Exception as ex:
        raise Exception(str(ex))
    return services_data


def dashboard_license_data(request):
    try:
        try:
            license = workload_api(request).workloads.license_list()
        except Exception:
            license = {}
        return license
    except Exception as ex:
        raise Exception(str(ex))


def dashboard_license_validity_check(request):
    try:
        license = workload_api(request).workloads.license_check()
    except Exception as ex:
        raise ex
    return license


def dashboard_license_create(request, license_data):
    try:
        license = workload_api(request).workloads.license_create(license_data)
    except Exception as ex:
        raise ex
    return license


def dashboard_usage_data(request):
    try:
        usage_data = workload_api(request).workloads.get_tenants_chargeback()
        tz = get_time_zone(request)
        for tenant in usage_data:
            usage_data[tenant]['used_capacity'] = sizeof_fmt(usage_data[tenant]['used_capacity'])
            for wlm in usage_data[tenant]['workloads']:
                usage_data[tenant]['workloads'][wlm]['size'] = sizeof_fmt(usage_data[tenant]['workloads'][wlm]['size'])
                for snap in usage_data[tenant]['workloads'][wlm]['snapshots']:
                    usage_data[tenant]['workloads'][wlm]['snapshots'][snap]['size'] = sizeof_fmt(
                        usage_data[tenant]['workloads'][wlm]['snapshots'][snap]['size'])
                    usage_data[tenant]['workloads'][wlm]['snapshots'][snap]['created_at'] = get_local_time(
                        usage_data[tenant]['workloads'][wlm]['snapshots'][snap]['created_at'], '', '', tz)
    except Exception as ex:
        raise ex
    return usage_data


def transfer_create(request, workload_id, transfer_name):
    try:
        try:
            transfer = workload_api(request).transfers.create(workload_id, transfer_name)
        except Exception:
            transfer = {}
        return transfer
    except Exception as ex:
        raise


def transfer_get(request, transfer_id):
    try:
        try:
            transfer = workload_api(request).transfers.get(transfer_id)
        except Exception:
            transfer = {}
        return transfer
    except Exception as ex:
        raise Exception(str(ex))


def transfer_list(request):
    try:
        try:
            transfer = workload_api(request).transfers.list()
        except Exception:
            transfer = {}
        return transfer
    except Exception as ex:
        raise Exception(str(ex))


def transfer_accept(request, transfer_id, auth):
    try:
        try:
            transfer = workload_api(request).transfers.accept(transfer_id, auth)
        except Exception:
            transfer = {}
        return transfer
    except Exception as ex:
        raise Exception(str(ex))


def transfer_delete(request, transfer_id):
    workload_api(request).transfers.delete(transfer_id)
    return True


def transfer_complete(request, transfer_id):
    workload_api(request).transfers.complete(transfer_id)
    return True


def get_global_job_scheduler(request):
    return workload_api(request).global_job_scheduler.get()


def enable_global_job_scheduler(request):
    return workload_api(request).global_job_scheduler.enable()


def disable_global_job_scheduler(request):
    return workload_api(request).global_job_scheduler.disable()


def testbubble_get(request, testbubble_id):
    try:
        testbubble = workload_api(request).testbubbles.get(testbubble_id)
        return testbubble
    except Exception:
        return None


def testbubble_list(request, snapshot_id):
    search_opts = {'snapshot_id': snapshot_id}
    testbubbles = workload_api(request).testbubbles.list(detailed=True, search_opts=search_opts)
    return testbubbles


def testbubble_delete(request, testbubble_id):
    workload_api(request).testbubbles.delete(testbubble_id)
    return True


def get_config_backup_scheduler_settings(request):
    return workload_api(request).config_backup.get_config_workload()


def set_config_backup_scheduler_settings(request, jobschedule, services_to_backup):
    return workload_api(request).config_backup.config_workload(jobschedule, services_to_backup)


def config_backup_list(request):
    config_backup = {}
    try:
        config_backup = workload_api(request).config_backup.config_backup_list()
    except Exception as ex:
        raise Exception(str(ex))
    return config_backup


def config_backup_create(request, name, description):
    config_backup = {}
    try:
        config_backup = workload_api(request).config_backup.config_backup(name, description)
    except Exception as ex:
        raise Exception(str(ex))
    return config_backup


def config_backup_delete(request, backup_id):
    try:
        backup_delete_message = workload_api(request).config_backup.config_backup_delete(backup_id)
    except Exception as ex:
        raise Exception(str(ex))
    return backup_delete_message


def config_backup_get(request, backup_id):
    try:
        backup_data = workload_api(request).config_backup.get_config_backup(backup_id)
    except Exception as ex:
        raise Exception(str(ex))
    return backup_data


def get_default_services_to_backup(request):
    return workload_api(request).config_backup.default_services


def create_policy(request, name, description, field_values):
    return workload_api(request).workload_policy.create(name, description, field_values)


def get_policy_list(request):
    try:
        policy_data = workload_api(request).workload_policy.list()
    except Exception as ex:
        raise Exception(str(ex))
    return policy_data


def get_policy(request, policy_id):
    try:
        policy_data = workload_api(request).workload_policy.get(policy_id)
    except Exception as ex:
        raise Exception(str(ex))
    return policy_data


def update_policy(request, policy_id, name, description, field_values):
    return workload_api(request).workload_policy.update(policy_id, name, description, field_values)


def assign_policy(request, policy_id, add_projects, remove_projects):
    try:
        policy_message = workload_api(request).workload_policy.assign(policy_id, add_projects, remove_projects)
    except Exception as ex:
        raise ex
    return policy_message


def delete_policy(request, policy_id):
    try:
        policy_delete_message = workload_api(request).workload_policy.delete(policy_id)
    except Exception as ex:
        raise Exception(str(ex))
    return policy_delete_message


def workload_policy_list(request, project_id):
    try:
        policy_list = workload_api(request).workload_policy.get_assigned_policies(project_id)
    except Exception as ex:
        raise Exception(str(ex))
    return policy_list