1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
104
105
106
107
108
109
110
111
112
113
114
115
116
117
118
119
120
121
122
123
124
125
126
|
# Copyright 2018-2019 Florian Fischer <florian.fl.fischer@fau.de>
#
# This file is part of allocbench.
#
# allocbench is free software: you can redistribute it and/or modify
# it under the terms of the GNU General Public License as published by
# the Free Software Foundation, either version 3 of the License, or
# (at your option) any later version.
#
# allocbench is distributed in the hope that it will be useful,
# but WITHOUT ANY WARRANTY; without even the implied warranty of
# MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
# GNU General Public License for more details.
#
# You should have received a copy of the GNU General Public License
# along with allocbench. If not, see <http://www.gnu.org/licenses/>.
"""Definition of the falsesahring benchmark"""
import re
import matplotlib.pyplot as plt
import numpy as np
from src.benchmark import Benchmark
from src.globalvars import summary_file_ext
import src.plots as plt
TIME_RE = re.compile("^Time elapsed = (?P<time>\\d*\\.\\d*) seconds.$")
class BenchmarkFalsesharing(Benchmark):
"""Falsesharing benchmark.
This benchmarks makes small allocations and writes to them multiple
times. If the allocated objects are on the same cache line the writes
will be expensive because of cache thrashing.
"""
def __init__(self):
name = "falsesharing"
self.cmd = "cache-{bench}{binary_suffix} {threads} 100 8 10000000"
self.args = {
"bench": ["thrash", "scratch"],
"threads": Benchmark.scale_threads_for_cpus(1)
}
self.requirements = ["cache-thrash", "cache-scratch"]
super().__init__(name)
@staticmethod
def process_output(result, stdout, stderr, allocator, perm):
result["time"] = TIME_RE.match(stdout).group("time")
def summary(self):
args = self.results["args"]
nthreads = args["threads"]
allocators = self.results["allocators"]
# calculate relevant datapoints: speedup, l1-cache-misses
for bench in self.results["args"]["bench"]:
for allocator in allocators:
sequential_perm = self.Perm(bench=bench, threads=1)
for perm in self.iterate_args({"bench": bench}, args=args):
speedup = []
l1chache_misses = []
for i, measure in enumerate(self.results[allocator][perm]):
sequential_time = float(self.results[allocator]
[sequential_perm][i]["time"])
measure["speedup"] = sequential_time / float(
measure["time"])
measure["l1chache_misses"] = eval(
"({L1-dcache-load-misses}/{L1-dcache-loads})*100".
format(**measure))
# delete and recalculate stats
del self.results["stats"]
self.calc_desc_statistics()
plt.plot(self,
"{speedup}",
x_args=["bench"],
fig_options={
'ylabel': "Speedup",
'title': "Speedup: {arg} {arg_value}",
'autoticks': False,
},
file_postfix="speedup")
plt.plot(self,
"{l1chache_misses}",
x_args=["bench"],
fig_options={
'ylabel': "l1 cache misses in %",
'title': "cache misses: {arg} {arg_value}",
'autoticks': False,
},
file_postfix="l1-misses")
plt.write_tex_table(self, [{
"label": "Speedup",
"expression": "{speedup}",
"sort": ">"
}],
file_postfix="speedup.table")
# plt.export_stats_to_csv(self, "speedup", "time")
# plt.export_stats_to_csv(self, "l1chache_misses", "l1-misses")
# pgfplots
for bench in args["bench"]:
plt.pgfplot(self,
self.iterate_args({"bench": bench}, args=args),
"int(perm.threads)",
"{speedup}",
xlabel="Threads",
ylabel="Speedup",
title=f"{bench}: Speedup",
postfix=f"{bench}.speedup")
# create pgfplot legend
plt.pgfplot_legend(self)
falsesharing = BenchmarkFalsesharing()
|