1
0
mirror of git://sourceware.org/git/lvm2.git synced 2024-12-21 13:34:40 +03:00

lvmdbusd: Initial support for json output

This commit is contained in:
Tony Asleson 2016-06-03 13:18:21 -05:00
parent 79446ffad7
commit ec45be9976
2 changed files with 209 additions and 20 deletions

View File

@ -22,6 +22,11 @@ except SystemError:
from utils import pv_dest_ranges, log_debug, log_error from utils import pv_dest_ranges, log_debug, log_error
from lvm_shell_proxy import LVMShellProxy from lvm_shell_proxy import LVMShellProxy
try:
import simplejson as json
except ImportError:
import json
SEP = '{|}' SEP = '{|}'
total_time = 0.0 total_time = 0.0
@ -426,6 +431,55 @@ def lv_detach_cache(lv_full_name, detach_options, destroy_cache):
return call(cmd) return call(cmd)
def supports_json():
cmd = ['help']
rc, out, err = call(cmd)
if rc == 0:
if 'fullreport' in err:
return True
return False
def lvm_full_report_json():
pv_columns = ['pv_name', 'pv_uuid', 'pv_fmt', 'pv_size', 'pv_free',
'pv_used', 'dev_size', 'pv_mda_size', 'pv_mda_free',
'pv_ba_start', 'pv_ba_size', 'pe_start', 'pv_pe_count',
'pv_pe_alloc_count', 'pv_attr', 'pv_tags', 'vg_name',
'vg_uuid']
pv_seg_columns = ['pv_seg_start', 'pv_seg_size', 'segtype',
'pv_uuid', 'lv_uuid', 'pv_name']
vg_columns = ['vg_name', 'vg_uuid', 'vg_fmt', 'vg_size', 'vg_free',
'vg_sysid', 'vg_extent_size', 'vg_extent_count',
'vg_free_count', 'vg_profile', 'max_lv', 'max_pv',
'pv_count', 'lv_count', 'snap_count', 'vg_seqno',
'vg_mda_count', 'vg_mda_free', 'vg_mda_size',
'vg_mda_used_count', 'vg_attr', 'vg_tags']
lv_columns = ['lv_uuid', 'lv_name', 'lv_path', 'lv_size',
'vg_name', 'pool_lv_uuid', 'pool_lv', 'origin_uuid',
'origin', 'data_percent',
'lv_attr', 'lv_tags', 'vg_uuid', 'lv_active', 'data_lv',
'metadata_lv', 'lv_parent', 'lv_role', 'lv_layout']
lv_seg_columns = ['seg_pe_ranges', 'segtype', 'lv_uuid']
cmd = _dc('fullreport', [
'-o', '/pv/' + ','.join(pv_columns),
'-o', '/vg/' + ','.join(vg_columns),
'-o', '/lv/' + ','.join(lv_columns),
'-o', '/seg/' + ','.join(lv_seg_columns),
'-o', '/pvseg/' + ','.join(pv_seg_columns),
'--reportformat', 'json'
])
rc, out, err = call(cmd)
if rc == 0:
return json.loads(out)
return None
def pv_retrieve_with_segs(device=None): def pv_retrieve_with_segs(device=None):
d = [] d = []
err = "" err = ""

View File

@ -13,6 +13,7 @@ from collections import OrderedDict
import pprint as prettyprint import pprint as prettyprint
import os import os
import sys
try: try:
from . import cmdhandler from . import cmdhandler
@ -23,7 +24,7 @@ except SystemError:
class DataStore(object): class DataStore(object):
def __init__(self): def __init__(self, usejson=None):
self.pvs = {} self.pvs = {}
self.vgs = {} self.vgs = {}
self.lvs = {} self.lvs = {}
@ -41,6 +42,11 @@ class DataStore(object):
# self.refresh() # self.refresh()
self.num_refreshes = 0 self.num_refreshes = 0
if usejson is None:
self.json = cmdhandler.supports_json()
else:
self.json = usejson
@staticmethod @staticmethod
def _insert_record(table, key, record, allowed_multiple): def _insert_record(table, key, record, allowed_multiple):
if key in table: if key in table:
@ -93,6 +99,58 @@ class DataStore(object):
return c_pvs, c_lookup, c_pvs_in_vgs return c_pvs, c_lookup, c_pvs_in_vgs
@staticmethod
def _parse_pvs_json(_all):
c_pvs = OrderedDict()
c_lookup = {}
c_pvs_in_vgs = {}
# Each item item in the report is a collection of information pertaining
# to the vg
for r in _all['report']:
tmp_pv = []
# Get the pv data for this VG.
if 'pv' in r:
tmp_pv.extend(r['pv'])
# Sort them
sorted_tmp_pv = sorted(tmp_pv, key=lambda pk: pk['pv_name'])
# Add them to result set
for p in sorted_tmp_pv:
c_pvs[p['pv_uuid']] = p
if 'pvseg' in r:
for s in r['pvseg']:
# TODO Why is json pvseg_start, not pv_seg_start?
r = c_pvs[s['pv_uuid']]
r.setdefault('pv_seg_start', []).append(s['pvseg_start'])
r.setdefault('pvseg_size', []).append(s['pvseg_size'])
r.setdefault('segtype', []).append(s['segtype'])
# TODO: Remove this bug work around when we have orphan segs.
for i in c_pvs.values():
if 'pv_seg_start' not in i:
i['pv_seg_start'] = '0'
i['pvseg_size'] = i['pv_pe_count']
i['segtype'] = 'free'
for p in c_pvs.values():
# Capture which PVs are associated with which VG
if p['vg_uuid'] not in c_pvs_in_vgs:
c_pvs_in_vgs[p['vg_uuid']] = []
if p['vg_name']:
c_pvs_in_vgs[p['vg_uuid']].append(
(p['pv_name'], p['pv_uuid']))
# Lookup for translating between /dev/<name> and pv uuid
c_lookup[p['pv_name']] = p['pv_uuid']
return c_pvs, c_lookup, c_pvs_in_vgs
@staticmethod @staticmethod
def _parse_vgs(_vgs): def _parse_vgs(_vgs):
vgs = sorted(_vgs, key=lambda vk: vk['vg_name']) vgs = sorted(_vgs, key=lambda vk: vk['vg_name'])
@ -107,20 +165,31 @@ class DataStore(object):
return c_vgs, c_lookup return c_vgs, c_lookup
@staticmethod @staticmethod
def _parse_lvs(_lvs): def _parse_vgs_json(_all):
lvs = sorted(_lvs, key=lambda vk: vk['lv_name'])
c_lvs = OrderedDict() tmp_vg = []
c_lvs_in_vgs = {} for r in _all['report']:
c_lvs_hidden = {} # Get the pv data for this VG.
c_lv_full_lookup = {} if 'vg' in r:
tmp_vg.extend(r['vg'])
for i in lvs: # Sort for consistent output, however this is optional
full_name = "%s/%s" % (i['vg_name'], i['lv_name']) vgs = sorted(tmp_vg, key=lambda vk: vk['vg_name'])
c_lv_full_lookup[full_name] = i['lv_uuid']
DataStore._insert_record( c_vgs = OrderedDict()
c_lvs, i['lv_uuid'], i, c_lookup = {}
['seg_pe_ranges', 'segtype'])
for i in vgs:
c_lookup[i['vg_name']] = i['vg_uuid']
c_vgs[i['vg_uuid']] = i
return c_vgs, c_lookup
@staticmethod
def _parse_lvs_common(c_lvs, c_lv_full_lookup):
c_lvs_in_vgs = OrderedDict()
c_lvs_hidden = OrderedDict()
for i in c_lvs.values(): for i in c_lvs.values():
if i['vg_uuid'] not in c_lvs_in_vgs: if i['vg_uuid'] not in c_lvs_in_vgs:
@ -150,6 +219,55 @@ class DataStore(object):
return c_lvs, c_lvs_in_vgs, c_lvs_hidden, c_lv_full_lookup return c_lvs, c_lvs_in_vgs, c_lvs_hidden, c_lv_full_lookup
@staticmethod
def _parse_lvs(_lvs):
lvs = sorted(_lvs, key=lambda vk: vk['lv_name'])
c_lvs = OrderedDict()
c_lv_full_lookup = OrderedDict()
for i in lvs:
full_name = "%s/%s" % (i['vg_name'], i['lv_name'])
c_lv_full_lookup[full_name] = i['lv_uuid']
DataStore._insert_record(
c_lvs, i['lv_uuid'], i,
['seg_pe_ranges', 'segtype'])
return DataStore._parse_lvs_common(c_lvs, c_lv_full_lookup)
@staticmethod
def _parse_lvs_json(_all):
c_lvs = OrderedDict()
c_lv_full_lookup = {}
# Each item item in the report is a collection of information pertaining
# to the vg
for r in _all['report']:
tmp_lv = []
# Get the lv data for this VG.
if 'lv' in r:
tmp_lv.extend(r['lv'])
# Sort them
sorted_tmp_lv = sorted(tmp_lv, key=lambda vk: vk['lv_name'])
# Add them to result set
for i in sorted_tmp_lv:
full_name = "%s/%s" % (i['vg_name'], i['lv_name'])
c_lv_full_lookup[full_name] = i['lv_uuid']
c_lvs[i['lv_uuid']] = i
# Add in the segment data
if 'seg' in r:
for s in r['seg']:
r = c_lvs[s['lv_uuid']]
r.setdefault('seg_pe_ranges', []).append(s['seg_pe_ranges'])
r.setdefault('segtype', []).append(s['segtype'])
return DataStore._parse_lvs_common(c_lvs, c_lv_full_lookup)
@staticmethod @staticmethod
def _make_list(l): def _make_list(l):
if not isinstance(l, list): if not isinstance(l, list):
@ -278,13 +396,22 @@ class DataStore(object):
log_debug("lvmdb - refresh entry") log_debug("lvmdb - refresh entry")
# Grab everything first then parse it # Grab everything first then parse it
_raw_pvs = cmdhandler.pv_retrieve_with_segs() if self.json:
_raw_vgs = cmdhandler.vg_retrieve(None) # Do a single lvm retrieve for everything in json
_raw_lvs = cmdhandler.lv_retrieve_with_segments() a = cmdhandler.lvm_full_report_json()
_pvs, _pvs_lookup, _pvs_in_vgs = self._parse_pvs(_raw_pvs) _pvs, _pvs_lookup, _pvs_in_vgs = self._parse_pvs_json(a)
_vgs, _vgs_lookup = self._parse_vgs(_raw_vgs) _vgs, _vgs_lookup = self._parse_vgs_json(a)
_lvs, _lvs_in_vgs, _lvs_hidden, _lvs_lookup = self._parse_lvs(_raw_lvs) _lvs, _lvs_in_vgs, _lvs_hidden, _lvs_lookup = self._parse_lvs_json(a)
else:
_raw_pvs = cmdhandler.pv_retrieve_with_segs()
_raw_vgs = cmdhandler.vg_retrieve(None)
_raw_lvs = cmdhandler.lv_retrieve_with_segments()
_pvs, _pvs_lookup, _pvs_in_vgs = self._parse_pvs(_raw_pvs)
_vgs, _vgs_lookup = self._parse_vgs(_raw_vgs)
_lvs, _lvs_in_vgs, _lvs_hidden, _lvs_lookup = self._parse_lvs(_raw_lvs)
# Set all # Set all
self.pvs = _pvs self.pvs = _pvs
@ -389,12 +516,20 @@ class DataStore(object):
if __name__ == "__main__": if __name__ == "__main__":
pp = prettyprint.PrettyPrinter(indent=4) pp = prettyprint.PrettyPrinter(indent=4)
ds = DataStore() use_json = False
if len(sys.argv) != 1:
print(len(sys.argv))
use_json = True
ds = DataStore(use_json)
ds.refresh() ds.refresh()
print("PVS")
for v in ds.pvs.values(): for v in ds.pvs.values():
pp.pprint(v) pp.pprint(v)
print("VGS")
for v in ds.vgs.values(): for v in ds.vgs.values():
pp.pprint(v) pp.pprint(v)