summaryrefslogtreecommitdiff
path: root/benchmark/run_benchmarks.py
blob: 8cb3503ccb7386ce6aac22fe78c53d4d8a9a9c14 (plain)
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
104
105
106
107
108
109
110
111
112
113
114
115
116
117
118
119
120
121
122
123
124
125
126
127
128
129
130
import sys
import os
import time
import subprocess
import copy
import numpy

from matplotlib.backends.backend_pdf import PdfPages
from matplotlib.backends.backend_agg import FigureCanvasAgg as FigureCanvas
from matplotlib.figure import Figure
from matplotlib.font_manager import FontProperties

import multiprocessing
cpu_count = multiprocessing.cpu_count()

from bm_registry import BENCHMARKS

NUM_RUNS_PER_TEST = 5

BAD_BOOST_KILL_DURATION = 5.0 #seconds

__path__ = os.path.dirname(__file__)

import threading

#because boost interrupts are broken in half the versions of boost
#this stupidity makes stock gnuradio apps exit with bad boost
def kill_after_timeout(p):
    time.sleep(BAD_BOOST_KILL_DURATION)
    try: p.kill()
    except: pass

def run_a_single_one(args, env):
    print env
    p = subprocess.Popen(args=args, env=env, stdout=subprocess.PIPE)
    t = threading.Thread(target=kill_after_timeout, args = (p,))
    t.daemon = True
    t.start()
    p.wait()
    out = p.stdout.read()
    #print out
    for line in out.splitlines():
        if line.startswith('##RESULT##'):
            return float(line[len('##RESULT##'):].strip())
    raise Exception, 'no result found!'
    #return t1-t0

def do_a_benchmark(bm):
    title = bm['wat']
    print '#'*(len(title)+25)
    print '## running benchmark:', title
    print '#'*(len(title)+25)
    result_means = list()
    result_stddevs = list()
    test_names = list()
    for run in bm['tests']:
        test_name = run['wat']
        print '-'*(len(test_name)+25)
        print '-- running test:', test_name.replace('\n', ' ')
        print '-'*(len(test_name)+25)
        test_names.append(test_name)
        args = run['args']
        args[0] = os.path.join(__path__, args[0])
        args = [sys.executable] + args
        env = run['env']
        env = copy.copy(env)
        if run.has_key('envextra'):
            env.update(run['envextra'])
        run_results = list()
        for num_runs in range(NUM_RUNS_PER_TEST):
            res = run_a_single_one(args=args, env=env)
            print 'Result:', res
            run_results.append(res)
        result_means.append(numpy.average(run_results))
        result_stddevs.append(numpy.std(run_results))

    print 'result_means', result_means
    print 'result_stddevs', result_stddevs

    bogomips = numpy.array(result_means)/1e6
    bogomips_dev = numpy.array(result_stddevs)/1e6
    ind = numpy.arange(len(test_names))
    width = 0.35
    fig = Figure()
    fig.set_size_inches((11,8.5))
    FigureCanvas(fig)
    ax = fig.add_subplot(1, 1, 1,
        ylabel='Performance (BogoMips)', title=title,
        xlabel='', xticks=ind+width/2., xticklabels=test_names
    )
    rects = ax.bar(ind, bogomips, width, color='blue',
        yerr=bogomips_dev,
        error_kw=dict(elinewidth=6, ecolor='pink'),
        label=bm['moar']
    )
    ax.set_xlim(-width, max(len(ind), 4))
    ax.set_ylim(0, max(*bogomips)*1.1 + max(*bogomips_dev)*2)
    for rect in rects:
        height = rect.get_height()
        ax.text(rect.get_x()+rect.get_width()/2.0, 1.01*height, '%.3f'%height, horizontalalignment='center')
    ax.grid(True)

    fontP = FontProperties()
    fontP.set_size('small')
    handles, labels = ax.get_legend_handles_labels()
    ax.legend(handles, labels, prop=fontP, loc="upper right")

    print '\n'
    return fig


if __name__ == '__main__':
    ####################################################################
    ## create pdf generator
    ####################################################################
    pdf_pages = PdfPages(sys.argv[1])

    ####################################################################
    ## loop through tests
    ####################################################################
    for bm in BENCHMARKS:
        fig = do_a_benchmark(bm)
        pdf_pages.savefig(fig)

    ####################################################################
    ## done
    ####################################################################
    print 'make pdf...'
    pdf_pages.close()
    print 'done!'