aboutsummaryrefslogtreecommitdiff
path: root/src/benchmarks/lld.py
blob: 0603037988bff1384e283eb1b22762b734973db4 (plain)
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
# Copyright 2018-2019 Florian Fischer <florian.fl.fischer@fau.de>
#
# This file is part of allocbench.
#
# allocbench is free software: you can redistribute it and/or modify
# it under the terms of the GNU General Public License as published by
# the Free Software Foundation, either version 3 of the License, or
# (at your option) any later version.
#
# allocbench is distributed in the hope that it will be useful,
# but WITHOUT ANY WARRANTY; without even the implied warranty of
# MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE.  See the
# GNU General Public License for more details.
#
# You should have received a copy of the GNU General Public License
# along with allocbench.  If not, see <http://www.gnu.org/licenses/>.

"""Benchmark definition using the llvm-lld speed benchmark"""

import os
from urllib.request import urlretrieve
import subprocess
import sys

import matplotlib.pyplot as plt

from src.artifact import ArchiveArtifact
from src.benchmark import Benchmark
import src.facter


class BenchmarkLld(Benchmark):
    """LLVM-lld speed benchmark

    This benchmark runs the lld speed benchmark provided by the llvm project.
    """

    def __init__(self):
        name = "lld"

        self.run_dir = "{build_dir}/lld-speed-test/{test}"
        # TODO: don't hardcode ld.lld location
        self.cmd = "/usr/bin/ld.lld @response.txt"

        self.args = {"test": ["chrome", "clang-fsds", "gold", "linux-kernel",
                              "llvm-as-fsds", "scylla", "clang", "clang-gdb-index",
                              "gold-fsds", "llvm-as", "mozilla"]}

        self.requirements = ["ld.lld"]
        super().__init__(name)

    def prepare(self):
        super().prepare()

        # save lld version
        self.results["facts"]["versions"]["lld"] = src.facter.exe_version("ld.lld", "-v")

        tests = ArchiveArtifact("lld-speed-test",
                                "https://s3-us-west-2.amazonaws.com/linker-tests/lld-speed-test.tar.xz",
                                "tar",
                                "2d449a11109c7363f67fd45513b42270f5ba2a92")
        tests.provide(self.build_dir)

    def cleanup(self):
        for perm in self.iterate_args():
            a_out = os.path.join("lld-speed-test", perm.test, "a.out")
            if os.path.isfile(a_out):
                os.remove(a_out)

    def summary(self):
        args = self.results["args"]
        allocators = self.results["allocators"]

        for perm in self.iterate_args(args=args):
            for i, allocator in enumerate(allocators):

                plt.bar([i],
                        self.results["stats"][allocator][perm]["mean"]["task-clock"],
                        yerr=self.results["stats"][allocator][perm]["std"]["task-clock"],
                        label=allocator, color=allocators[allocator]["color"])

            plt.legend(loc="best")
            plt.ylabel("Zeit in ms")
            plt.title(f"Gesamte Laufzeit {perm.test}")
            plt.savefig(".".join([self.name, perm.test, "runtime", "png"]))
            plt.clf()

        # TODO: get memusage
        # Memusage
        # self.barplot_single_arg("{VmHWM}",
                                # ylabel='"Max RSS in KB"',
                                # title='"Highwatermark of Vm (VmHWM)"',
                                # filepostfix="rss")

        # self.export_stats_to_csv("VmHWM")
        self.export_stats_to_csv("task-clock")

        # self.export_stats_to_dataref("VmHWM")
        self.export_stats_to_dataref("task-clock")


lld = BenchmarkLld()