aboutsummaryrefslogtreecommitdiff
diff options
context:
space:
mode:
authorSiddhesh Poyarekar <siddhesh@redhat.com>2014-03-29 09:37:44 +0530
committerSiddhesh Poyarekar <siddhesh@redhat.com>2014-03-29 09:37:44 +0530
commitcb5e4aada7f044fc029dd64b31411a23bb09c287 (patch)
tree24f50d1418ef624707a8745737c89086944528a5
parentcf806aff6067273307d958f35c0a4cd0b0d40e80 (diff)
downloadglibc-cb5e4aada7f044fc029dd64b31411a23bb09c287.zip
glibc-cb5e4aada7f044fc029dd64b31411a23bb09c287.tar.gz
glibc-cb5e4aada7f044fc029dd64b31411a23bb09c287.tar.bz2
Make bench.out in json format
This patch changes the output format of the main benchmark output file (bench.out) to an extensible format. I chose JSON over XML because in addition to being extensible, it is also not too verbose. Additionally it has good support in python. The significant change I have made in terms of functionality is to put timing information as an attribute in JSON instead of a string and to do that, there is a separate program that prints out a JSON snippet mentioning the type of timing (hp_timing or clock_gettime). The mean timing has now changed from iterations per unit to actual timing per iteration.
-rw-r--r--ChangeLog12
-rw-r--r--benchtests/Makefile23
-rw-r--r--benchtests/bench-skeleton.c18
-rw-r--r--benchtests/bench-timing-type.c27
-rw-r--r--benchtests/bench-timing.h14
-rwxr-xr-xbenchtests/scripts/bench.py2
6 files changed, 78 insertions, 18 deletions
diff --git a/ChangeLog b/ChangeLog
index 9aeb53c..072747b 100644
--- a/ChangeLog
+++ b/ChangeLog
@@ -1,5 +1,17 @@
2014-03-29 Siddhesh Poyarekar <siddhesh@redhat.com>
+ * benchtests/Makefile (timing-type): New binary.
+ (bench-clean): Also remove bench-timing-type.
+ (bench): New target for timing-type.
+ (bench-func): Print output in JSON format.
+ * benchtests/bench-skeleton.c (main): Print output in JSON
+ format.
+ * benchtests/bench-timing-type.c: New file.
+ * benchtests/bench-timing.h (TIMING_TYPE): New macro.
+ (TIMING_PRINT_STATS): Remove.
+ * benchtests/scripts/bench.py (_print_arg_data): Store variant
+ name separately.
+
* benchtests/bench-modf.c: Remove.
* benchtests/modf-inputs: New inputs file.
diff --git a/benchtests/Makefile b/benchtests/Makefile
index b331d1a..be11708 100644
--- a/benchtests/Makefile
+++ b/benchtests/Makefile
@@ -98,11 +98,14 @@ run-bench = $(test-wrapper-env) \
GCONV_PATH=$(common-objpfx)iconvdata LC_ALL=C \
$($*-ENV) $(rtld-prefix) $${run}
+timing-type := $(objpfx)bench-timing-type
+
bench-clean:
rm -f $(binaries-bench) $(addsuffix .o,$(binaries-bench))
rm -f $(binaries-benchset) $(addsuffix .o,$(binaries-benchset))
+ rm -f $(timing-type) $(addsuffix .o,$(timing-type))
-bench: bench-set bench-func
+bench: $(timing-type) bench-set bench-func
bench-set: $(binaries-benchset)
for run in $^; do \
@@ -110,17 +113,29 @@ bench-set: $(binaries-benchset)
$(run-bench) > $${run}.out; \
done
+# Build and execute the benchmark functions. This target generates JSON
+# formatted bench.out. Each of the programs produce independent JSON output,
+# so one could even execute them individually and process it using any JSON
+# capable language or tool.
bench-func: $(binaries-bench)
- { for run in $^; do \
+ { echo "{"; \
+ $(timing-type); \
+ echo " ,\"functions\": {"; \
+ for run in $^; do \
+ if ! [ "x$${run}" = "x$<" ]; then \
+ echo ","; \
+ fi; \
echo "Running $${run}" >&2; \
$(run-bench); \
- done; } > $(objpfx)bench.out-tmp; \
+ done; \
+ echo " }"; \
+ echo "}"; } > $(objpfx)bench.out-tmp; \
if [ -f $(objpfx)bench.out ]; then \
mv -f $(objpfx)bench.out $(objpfx)bench.out.old; \
fi; \
mv -f $(objpfx)bench.out-tmp $(objpfx)bench.out
-$(binaries-bench) $(binaries-benchset): %: %.o \
+$(timing-type) $(binaries-bench) $(binaries-benchset): %: %.o \
$(sort $(filter $(common-objpfx)lib%,$(link-libc))) \
$(addprefix $(csu-objpfx),start.o) $(+preinit) $(+postinit)
$(+link)
diff --git a/benchtests/bench-skeleton.c b/benchtests/bench-skeleton.c
index 4290e76..faef7eb 100644
--- a/benchtests/bench-skeleton.c
+++ b/benchtests/bench-skeleton.c
@@ -59,8 +59,13 @@ main (int argc, char **argv)
iters = 1000 * res;
+ /* Begin function. */
+ printf ("\"%s\": {\n", FUNCNAME);
+
for (int v = 0; v < NUM_VARIANTS; v++)
{
+ if (v)
+ putc (',', stdout);
/* Run for approximately DURATION seconds. */
clock_gettime (CLOCK_MONOTONIC_RAW, &runtime);
runtime.tv_sec += DURATION;
@@ -86,7 +91,6 @@ main (int argc, char **argv)
min = cur;
TIMING_ACCUM (total, cur);
-
d_total_i += iters;
}
struct timespec curtime;
@@ -104,9 +108,17 @@ main (int argc, char **argv)
d_total_s = total;
d_iters = iters;
- TIMING_PRINT_STATS (VARIANT (v), d_total_s, d_iters, d_total_i, max,
- min);
+ printf ("\"%s\": {\n", VARIANT (v));
+ printf ("\"duration\": %g, \"iterations\": %g, "
+ "\"max\": %g, \"min\": %g, \"mean\": %g\n",
+ d_total_s, d_total_i, max / d_iters, min / d_iters,
+ d_total_s / d_total_i);
+
+ puts ("}");
}
+ /* End function. */
+ puts ("}");
+
return 0;
}
diff --git a/benchtests/bench-timing-type.c b/benchtests/bench-timing-type.c
new file mode 100644
index 0000000..903a61f
--- /dev/null
+++ b/benchtests/bench-timing-type.c
@@ -0,0 +1,27 @@
+/* Print out the timing type used by the benchmark run.
+ Copyright (C) 2014 Free Software Foundation, Inc.
+ This file is part of the GNU C Library.
+
+ The GNU C Library is free software; you can redistribute it and/or
+ modify it under the terms of the GNU Lesser General Public
+ License as published by the Free Software Foundation; either
+ version 2.1 of the License, or (at your option) any later version.
+
+ The GNU C Library is distributed in the hope that it will be useful,
+ but WITHOUT ANY WARRANTY; without even the implied warranty of
+ MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU
+ Lesser General Public License for more details.
+
+ You should have received a copy of the GNU Lesser General Public
+ License along with the GNU C Library; if not, see
+ <http://www.gnu.org/licenses/>. */
+
+#include "bench-timing.h"
+#include <stdio.h>
+
+int
+main (int argc, char **argv)
+{
+ printf ("\"timing-type\": \"%s\"\n", TIMING_TYPE);
+ return 0;
+}
diff --git a/benchtests/bench-timing.h b/benchtests/bench-timing.h
index 13fc946..ccde601 100644
--- a/benchtests/bench-timing.h
+++ b/benchtests/bench-timing.h
@@ -25,6 +25,8 @@
hp_timing_t _dl_hp_timing_overhead;
typedef hp_timing_t timing_t;
+# define TIMING_TYPE "hp_timing"
+
# define TIMING_INIT(res) \
({ \
HP_TIMING_DIFF_INIT(); \
@@ -35,16 +37,13 @@ typedef hp_timing_t timing_t;
# define TIMING_DIFF(diff, start, end) HP_TIMING_DIFF ((diff), (start), (end))
# define TIMING_ACCUM(sum, diff) HP_TIMING_ACCUM_NT ((sum), (diff))
-# define TIMING_PRINT_STATS(func, d_total_s, d_iters, d_total_i, max, min) \
- printf ("%s: ITERS:%g: TOTAL:%gMcy, MAX:%gcy, MIN:%gcy, %g calls/Mcy\n", \
- (func), (d_total_i), (d_total_s) * 1e-6, (max) / (d_iters), \
- (min) / (d_iters), 1e6 * (d_total_i) / (d_total_s));
-
#else
#include <time.h>
typedef uint64_t timing_t;
+# define TIMING_TYPE "clock_gettime"
+
/* Measure the resolution of the clock so we can scale the number of
benchmark iterations by this value. */
# define TIMING_INIT(res) \
@@ -64,11 +63,6 @@ typedef uint64_t timing_t;
# define TIMING_DIFF(diff, start, end) (diff) = (end) - (start)
# define TIMING_ACCUM(sum, diff) (sum) += (diff)
-# define TIMING_PRINT_STATS(func, d_total_s, d_iters, d_total_i, max, min) \
- printf ("%s: ITERS:%g: TOTAL:%gs, MAX:%gns, MIN:%gns, %g iter/s\n", \
- (func), (d_total_i), (d_total_s) * 1e-9, (max) / (d_iters), \
- (min) / (d_iters), 1e9 * (d_total_i) / (d_total_s))
-
#endif
#define TIMING_PRINT_MEAN(d_total_s, d_iters) \
diff --git a/benchtests/scripts/bench.py b/benchtests/scripts/bench.py
index e500a33..90317b5 100755
--- a/benchtests/scripts/bench.py
+++ b/benchtests/scripts/bench.py
@@ -172,7 +172,7 @@ def _print_arg_data(func, directives, all_vals):
# Members for the variants structure list that we will
# print later.
- variants.append(' {"%s(%s)", %d, in%d},' % (func, k, len(vals), i))
+ variants.append(' {"%s", %d, in%d},' % (k, len(vals), i))
print(ARGS_TEMPLATE % {'argnum': i, 'num_args': len(vals),
'args': '\n'.join(out)})