report.py 12 KB

123456789101112131415161718192021222324252627282930313233343536373839404142434445464748495051525354555657585960616263646566676869707172737475767778798081828384858687888990919293949596979899100101102103104105106107108109110111112113114115116117118119120121122123124125126127128129130131132133134135136137138139140141142143144145146147148149150151152153154155156157158159160161162163164165166167168169170171172173174175176177178179180181182183184185186187188189190191192193194195196197198199200201202203204205206207208209210211212213214215216217218219220221222223224225226227228229230231232233234235236237238239240241242243244245246247248249250251252253254255256257258259260261262263264265266267268269270271272273274275276277278279280281282283284285286287288289290291292293294295296297298299300301302303304305306307308309310311312313314315316317318319320321322323324325326327328329330331332333334335336337338339
  1. #
  2. # Copyright (c) 2017, Intel Corporation.
  3. #
  4. # SPDX-License-Identifier: GPL-2.0-only
  5. #
  6. """Handling of build perf test reports"""
  7. from collections import OrderedDict, namedtuple
  8. from collections.abc import Mapping
  9. from datetime import datetime, timezone
  10. from numbers import Number
  11. from statistics import mean, stdev, variance
  12. AggregateTestData = namedtuple('AggregateTestData', ['metadata', 'results'])
  13. def isofmt_to_timestamp(string):
  14. """Convert timestamp string in ISO 8601 format into unix timestamp"""
  15. if '.' in string:
  16. dt = datetime.strptime(string, '%Y-%m-%dT%H:%M:%S.%f')
  17. else:
  18. dt = datetime.strptime(string, '%Y-%m-%dT%H:%M:%S')
  19. return dt.replace(tzinfo=timezone.utc).timestamp()
  20. def metadata_xml_to_json(elem):
  21. """Convert metadata xml into JSON format"""
  22. assert elem.tag == 'metadata', "Invalid metadata file format"
  23. def _xml_to_json(elem):
  24. """Convert xml element to JSON object"""
  25. out = OrderedDict()
  26. for child in elem.getchildren():
  27. key = child.attrib.get('name', child.tag)
  28. if len(child):
  29. out[key] = _xml_to_json(child)
  30. else:
  31. out[key] = child.text
  32. return out
  33. return _xml_to_json(elem)
  34. def results_xml_to_json(elem):
  35. """Convert results xml into JSON format"""
  36. rusage_fields = ('ru_utime', 'ru_stime', 'ru_maxrss', 'ru_minflt',
  37. 'ru_majflt', 'ru_inblock', 'ru_oublock', 'ru_nvcsw',
  38. 'ru_nivcsw')
  39. iostat_fields = ('rchar', 'wchar', 'syscr', 'syscw', 'read_bytes',
  40. 'write_bytes', 'cancelled_write_bytes')
  41. def _read_measurement(elem):
  42. """Convert measurement to JSON"""
  43. data = OrderedDict()
  44. data['type'] = elem.tag
  45. data['name'] = elem.attrib['name']
  46. data['legend'] = elem.attrib['legend']
  47. values = OrderedDict()
  48. # SYSRES measurement
  49. if elem.tag == 'sysres':
  50. for subel in elem:
  51. if subel.tag == 'time':
  52. values['start_time'] = isofmt_to_timestamp(subel.attrib['timestamp'])
  53. values['elapsed_time'] = float(subel.text)
  54. elif subel.tag == 'rusage':
  55. rusage = OrderedDict()
  56. for field in rusage_fields:
  57. if 'time' in field:
  58. rusage[field] = float(subel.attrib[field])
  59. else:
  60. rusage[field] = int(subel.attrib[field])
  61. values['rusage'] = rusage
  62. elif subel.tag == 'iostat':
  63. values['iostat'] = OrderedDict([(f, int(subel.attrib[f]))
  64. for f in iostat_fields])
  65. elif subel.tag == 'buildstats_file':
  66. values['buildstats_file'] = subel.text
  67. else:
  68. raise TypeError("Unknown sysres value element '{}'".format(subel.tag))
  69. # DISKUSAGE measurement
  70. elif elem.tag == 'diskusage':
  71. values['size'] = int(elem.find('size').text)
  72. else:
  73. raise Exception("Unknown measurement tag '{}'".format(elem.tag))
  74. data['values'] = values
  75. return data
  76. def _read_testcase(elem):
  77. """Convert testcase into JSON"""
  78. assert elem.tag == 'testcase', "Expecting 'testcase' element instead of {}".format(elem.tag)
  79. data = OrderedDict()
  80. data['name'] = elem.attrib['name']
  81. data['description'] = elem.attrib['description']
  82. data['status'] = 'SUCCESS'
  83. data['start_time'] = isofmt_to_timestamp(elem.attrib['timestamp'])
  84. data['elapsed_time'] = float(elem.attrib['time'])
  85. measurements = OrderedDict()
  86. for subel in elem.getchildren():
  87. if subel.tag == 'error' or subel.tag == 'failure':
  88. data['status'] = subel.tag.upper()
  89. data['message'] = subel.attrib['message']
  90. data['err_type'] = subel.attrib['type']
  91. data['err_output'] = subel.text
  92. elif subel.tag == 'skipped':
  93. data['status'] = 'SKIPPED'
  94. data['message'] = subel.text
  95. else:
  96. measurements[subel.attrib['name']] = _read_measurement(subel)
  97. data['measurements'] = measurements
  98. return data
  99. def _read_testsuite(elem):
  100. """Convert suite to JSON"""
  101. assert elem.tag == 'testsuite', \
  102. "Expecting 'testsuite' element instead of {}".format(elem.tag)
  103. data = OrderedDict()
  104. if 'hostname' in elem.attrib:
  105. data['tester_host'] = elem.attrib['hostname']
  106. data['start_time'] = isofmt_to_timestamp(elem.attrib['timestamp'])
  107. data['elapsed_time'] = float(elem.attrib['time'])
  108. tests = OrderedDict()
  109. for case in elem.getchildren():
  110. tests[case.attrib['name']] = _read_testcase(case)
  111. data['tests'] = tests
  112. return data
  113. # Main function
  114. assert elem.tag == 'testsuites', "Invalid test report format"
  115. assert len(elem) == 1, "Too many testsuites"
  116. return _read_testsuite(elem.getchildren()[0])
  117. def aggregate_metadata(metadata):
  118. """Aggregate metadata into one, basically a sanity check"""
  119. mutable_keys = ('pretty_name', 'version_id')
  120. def aggregate_obj(aggregate, obj, assert_str=True):
  121. """Aggregate objects together"""
  122. assert type(aggregate) is type(obj), \
  123. "Type mismatch: {} != {}".format(type(aggregate), type(obj))
  124. if isinstance(obj, Mapping):
  125. assert set(aggregate.keys()) == set(obj.keys())
  126. for key, val in obj.items():
  127. aggregate_obj(aggregate[key], val, key not in mutable_keys)
  128. elif isinstance(obj, list):
  129. assert len(aggregate) == len(obj)
  130. for i, val in enumerate(obj):
  131. aggregate_obj(aggregate[i], val)
  132. elif not isinstance(obj, str) or (isinstance(obj, str) and assert_str):
  133. assert aggregate == obj, "Data mismatch {} != {}".format(aggregate, obj)
  134. if not metadata:
  135. return {}
  136. # Do the aggregation
  137. aggregate = metadata[0].copy()
  138. for testrun in metadata[1:]:
  139. aggregate_obj(aggregate, testrun)
  140. aggregate['testrun_count'] = len(metadata)
  141. return aggregate
  142. def aggregate_data(data):
  143. """Aggregate multiple test results JSON structures into one"""
  144. mutable_keys = ('status', 'message', 'err_type', 'err_output')
  145. class SampleList(list):
  146. """Container for numerical samples"""
  147. pass
  148. def new_aggregate_obj(obj):
  149. """Create new object for aggregate"""
  150. if isinstance(obj, Number):
  151. new_obj = SampleList()
  152. new_obj.append(obj)
  153. elif isinstance(obj, str):
  154. new_obj = obj
  155. else:
  156. # Lists and and dicts are kept as is
  157. new_obj = obj.__class__()
  158. aggregate_obj(new_obj, obj)
  159. return new_obj
  160. def aggregate_obj(aggregate, obj, assert_str=True):
  161. """Recursive "aggregation" of JSON objects"""
  162. if isinstance(obj, Number):
  163. assert isinstance(aggregate, SampleList)
  164. aggregate.append(obj)
  165. return
  166. assert type(aggregate) == type(obj), \
  167. "Type mismatch: {} != {}".format(type(aggregate), type(obj))
  168. if isinstance(obj, Mapping):
  169. for key, val in obj.items():
  170. if not key in aggregate:
  171. aggregate[key] = new_aggregate_obj(val)
  172. else:
  173. aggregate_obj(aggregate[key], val, key not in mutable_keys)
  174. elif isinstance(obj, list):
  175. for i, val in enumerate(obj):
  176. if i >= len(aggregate):
  177. aggregate[key] = new_aggregate_obj(val)
  178. else:
  179. aggregate_obj(aggregate[i], val)
  180. elif isinstance(obj, str):
  181. # Sanity check for data
  182. if assert_str:
  183. assert aggregate == obj, "Data mismatch {} != {}".format(aggregate, obj)
  184. else:
  185. raise Exception("BUG: unable to aggregate '{}' ({})".format(type(obj), str(obj)))
  186. if not data:
  187. return {}
  188. # Do the aggregation
  189. aggregate = data[0].__class__()
  190. for testrun in data:
  191. aggregate_obj(aggregate, testrun)
  192. return aggregate
  193. class MeasurementVal(float):
  194. """Base class representing measurement values"""
  195. gv_data_type = 'number'
  196. def gv_value(self):
  197. """Value formatting for visualization"""
  198. if self != self:
  199. return "null"
  200. else:
  201. return self
  202. class TimeVal(MeasurementVal):
  203. """Class representing time values"""
  204. quantity = 'time'
  205. gv_title = 'elapsed time'
  206. gv_data_type = 'timeofday'
  207. def hms(self):
  208. """Split time into hours, minutes and seconeds"""
  209. hhh = int(abs(self) / 3600)
  210. mmm = int((abs(self) % 3600) / 60)
  211. sss = abs(self) % 60
  212. return hhh, mmm, sss
  213. def __str__(self):
  214. if self != self:
  215. return "nan"
  216. hh, mm, ss = self.hms()
  217. sign = '-' if self < 0 else ''
  218. if hh > 0:
  219. return '{}{:d}:{:02d}:{:02.0f}'.format(sign, hh, mm, ss)
  220. elif mm > 0:
  221. return '{}{:d}:{:04.1f}'.format(sign, mm, ss)
  222. elif ss > 1:
  223. return '{}{:.1f} s'.format(sign, ss)
  224. else:
  225. return '{}{:.2f} s'.format(sign, ss)
  226. def gv_value(self):
  227. """Value formatting for visualization"""
  228. if self != self:
  229. return "null"
  230. hh, mm, ss = self.hms()
  231. return [hh, mm, int(ss), int(ss*1000) % 1000]
  232. class SizeVal(MeasurementVal):
  233. """Class representing time values"""
  234. quantity = 'size'
  235. gv_title = 'size in MiB'
  236. gv_data_type = 'number'
  237. def __str__(self):
  238. if self != self:
  239. return "nan"
  240. if abs(self) < 1024:
  241. return '{:.1f} kiB'.format(self)
  242. elif abs(self) < 1048576:
  243. return '{:.2f} MiB'.format(self / 1024)
  244. else:
  245. return '{:.2f} GiB'.format(self / 1048576)
  246. def gv_value(self):
  247. """Value formatting for visualization"""
  248. if self != self:
  249. return "null"
  250. return self / 1024
  251. def measurement_stats(meas, prefix=''):
  252. """Get statistics of a measurement"""
  253. if not meas:
  254. return {prefix + 'sample_cnt': 0,
  255. prefix + 'mean': MeasurementVal('nan'),
  256. prefix + 'stdev': MeasurementVal('nan'),
  257. prefix + 'variance': MeasurementVal('nan'),
  258. prefix + 'min': MeasurementVal('nan'),
  259. prefix + 'max': MeasurementVal('nan'),
  260. prefix + 'minus': MeasurementVal('nan'),
  261. prefix + 'plus': MeasurementVal('nan')}
  262. stats = {'name': meas['name']}
  263. if meas['type'] == 'sysres':
  264. val_cls = TimeVal
  265. values = meas['values']['elapsed_time']
  266. elif meas['type'] == 'diskusage':
  267. val_cls = SizeVal
  268. values = meas['values']['size']
  269. else:
  270. raise Exception("Unknown measurement type '{}'".format(meas['type']))
  271. stats['val_cls'] = val_cls
  272. stats['quantity'] = val_cls.quantity
  273. stats[prefix + 'sample_cnt'] = len(values)
  274. mean_val = val_cls(mean(values))
  275. min_val = val_cls(min(values))
  276. max_val = val_cls(max(values))
  277. stats[prefix + 'mean'] = mean_val
  278. if len(values) > 1:
  279. stats[prefix + 'stdev'] = val_cls(stdev(values))
  280. stats[prefix + 'variance'] = val_cls(variance(values))
  281. else:
  282. stats[prefix + 'stdev'] = float('nan')
  283. stats[prefix + 'variance'] = float('nan')
  284. stats[prefix + 'min'] = min_val
  285. stats[prefix + 'max'] = max_val
  286. stats[prefix + 'minus'] = val_cls(mean_val - min_val)
  287. stats[prefix + 'plus'] = val_cls(max_val - mean_val)
  288. return stats