summary refs log tree commit diff
path: root/benchtests/scripts
diff options
context:
space:
mode:
authorSiddhesh Poyarekar <siddhesh@sourceware.org>2017-08-08 00:55:12 +0530
committerSiddhesh Poyarekar <siddhesh@sourceware.org>2017-08-08 00:55:12 +0530
commitb115e819af637101d9d9b0d26c3685b7236d3fb1 (patch)
tree297a6d9bd37fbd71a302b9121fd1a40e017fabc1 /benchtests/scripts
parent7ee38e6040d34bca96ee668efbbd2f56b446319d (diff)
downloadglibc-b115e819af637101d9d9b0d26c3685b7236d3fb1.tar.gz
glibc-b115e819af637101d9d9b0d26c3685b7236d3fb1.tar.xz
glibc-b115e819af637101d9d9b0d26c3685b7236d3fb1.zip
benchtests: Allow selecting baseline for compare_string.py
This patch allows one to provide the function name using an optional
-base option to compare all other functions against.  This is useful
when pitching one implementation of a string function against
alternatives.  In the absence of this option, comparisons are done
against the first ifunc in the list.

	* benchtests/scripts/compare_strings.py (main): Add an
	optional -base option.
	(process_results): New argument base_func.
Diffstat (limited to 'benchtests/scripts')
-rwxr-xr-xbenchtests/scripts/compare_strings.py28
1 files changed, 18 insertions, 10 deletions
diff --git a/benchtests/scripts/compare_strings.py b/benchtests/scripts/compare_strings.py
index 9d73ec41f4..43e70ebee2 100755
--- a/benchtests/scripts/compare_strings.py
+++ b/benchtests/scripts/compare_strings.py
@@ -21,6 +21,7 @@ Given a string benchmark result file, print a table with comparisons with a
 baseline.  The baseline is the first function, which typically is the builtin
 function.
 """
+
 import sys
 import os
 import json
@@ -74,7 +75,7 @@ def draw_graph(f, v, ifuncs, results):
     pylab.savefig('%s-%s.png' % (f, v), bbox_inches='tight')
 
 
-def process_results(results, attrs):
+def process_results(results, attrs, base_func):
     """ Process results and print them
 
     Args:
@@ -84,6 +85,10 @@ def process_results(results, attrs):
 
     for f in results['functions'].keys():
         print('Function: %s' % f)
+        base_index = 0
+        if base_func:
+            base_index = results['functions'][f]['ifuncs'].index(base_func)
+
         print('\t'.join(results['functions'][f]['ifuncs']))
         v = results['functions'][f]['bench-variant']
         print('Variant: %s' % v)
@@ -91,19 +96,17 @@ def process_results(results, attrs):
         graph_res = {}
         for res in results['functions'][f]['results']:
             attr_list = ['%s=%s' % (a, res[a]) for a in attrs]
-            first = True
+            i = 0
             key = ','.join(attr_list)
             sys.stdout.write('%s: \t' % key)
             graph_res[key] = res['timings']
             for t in res['timings']:
                 sys.stdout.write ('%.2f' % t)
-                if first:
-                    first = False
-                else:
-                    diff = (res['timings'][0] - t) * 100 / res['timings'][0]
-
+                if i != base_index:
+                    diff = (res['timings'][base_index] - t) * 100 / res['timings'][base_index]
                     sys.stdout.write (' (%.2f%%)' % diff)
                 sys.stdout.write('\t')
+                i = i + 1
             print('')
         draw_graph(f, v, results['functions'][f]['ifuncs'], graph_res)
 
@@ -114,15 +117,20 @@ def main(args):
     Take a string benchmark output file and compare timings.
     """
     if len(args) < 3:
-        print('Usage: %s <input file> <schema file> attr1 [attr2 ...]' % sys.argv[0])
+        print('Usage: %s <input file> <schema file> [-base=ifunc_name] attr1 [attr2 ...]' % sys.argv[0])
         sys.exit(os.EX_USAGE)
 
+    base_func = None
     filename = args[0]
     schema_filename = args[1]
-    attrs = args[2:]
+    if args[2].find('-base=') == 0:
+        base_func = args[2][6:]
+        attrs = args[3:]
+    else:
+        attrs = args[2:]
 
     results = parse_file(filename, schema_filename)
-    process_results(results, attrs)
+    process_results(results, attrs, base_func)
 
 
 if __name__ == '__main__':