|
|
@ -44,9 +44,13 @@ def run_one_test(filename): |
|
|
|
data = open(filename).read() |
|
|
|
data = open(filename).read() |
|
|
|
benchmark_dataset = benchmarks_pb2.BenchmarkDataset() |
|
|
|
benchmark_dataset = benchmarks_pb2.BenchmarkDataset() |
|
|
|
benchmark_dataset.ParseFromString(data) |
|
|
|
benchmark_dataset.ParseFromString(data) |
|
|
|
|
|
|
|
total_bytes = 0 |
|
|
|
|
|
|
|
for payload in benchmark_dataset.payload: |
|
|
|
|
|
|
|
total_bytes += len(payload) |
|
|
|
benchmark_util = Benchmark(full_iteration=len(benchmark_dataset.payload), |
|
|
|
benchmark_util = Benchmark(full_iteration=len(benchmark_dataset.payload), |
|
|
|
module="py_benchmark", |
|
|
|
module="py_benchmark", |
|
|
|
setup_method="init") |
|
|
|
setup_method="init", |
|
|
|
|
|
|
|
total_bytes=total_bytes) |
|
|
|
result={} |
|
|
|
result={} |
|
|
|
result["filename"] = filename |
|
|
|
result["filename"] = filename |
|
|
|
result["message_name"] = benchmark_dataset.message_name |
|
|
|
result["message_name"] = benchmark_dataset.message_name |
|
|
@ -61,10 +65,11 @@ def run_one_test(filename): |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
def init(filename): |
|
|
|
def init(filename): |
|
|
|
global benchmark_dataset, message_class, message_list, counter |
|
|
|
global benchmark_dataset, message_class, message_list, counter, total_bytes |
|
|
|
message_list=[] |
|
|
|
message_list=[] |
|
|
|
counter = 0 |
|
|
|
counter = 0 |
|
|
|
data = open(os.path.dirname(sys.argv[0]) + "/../" + filename).read() |
|
|
|
total_bytes = 0 |
|
|
|
|
|
|
|
data = open(filename).read() |
|
|
|
benchmark_dataset = benchmarks_pb2.BenchmarkDataset() |
|
|
|
benchmark_dataset = benchmarks_pb2.BenchmarkDataset() |
|
|
|
benchmark_dataset.ParseFromString(data) |
|
|
|
benchmark_dataset.ParseFromString(data) |
|
|
|
|
|
|
|
|
|
|
@ -85,6 +90,7 @@ def init(filename): |
|
|
|
temp = message_class() |
|
|
|
temp = message_class() |
|
|
|
temp.ParseFromString(one_payload) |
|
|
|
temp.ParseFromString(one_payload) |
|
|
|
message_list.append(temp) |
|
|
|
message_list.append(temp) |
|
|
|
|
|
|
|
total_bytes += len(one_payload) |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
def parse_from_benchmark(): |
|
|
|
def parse_from_benchmark(): |
|
|
@ -101,11 +107,12 @@ def serialize_to_benchmark(): |
|
|
|
|
|
|
|
|
|
|
|
class Benchmark: |
|
|
|
class Benchmark: |
|
|
|
def __init__(self, module=None, test_method=None, |
|
|
|
def __init__(self, module=None, test_method=None, |
|
|
|
setup_method=None, full_iteration = 1): |
|
|
|
setup_method=None, total_bytes=None, full_iteration = 1): |
|
|
|
self.full_iteration = full_iteration |
|
|
|
self.full_iteration = full_iteration |
|
|
|
self.module = module |
|
|
|
self.module = module |
|
|
|
self.test_method = test_method |
|
|
|
self.test_method = test_method |
|
|
|
self.setup_method = setup_method |
|
|
|
self.setup_method = setup_method |
|
|
|
|
|
|
|
self.total_bytes = total_bytes |
|
|
|
|
|
|
|
|
|
|
|
def set_test_method(self, test_method): |
|
|
|
def set_test_method(self, test_method): |
|
|
|
self.test_method = test_method |
|
|
|
self.test_method = test_method |
|
|
@ -130,7 +137,7 @@ class Benchmark: |
|
|
|
t = timeit.timeit(stmt="%s(%s)" % (self.test_method, test_method_args), |
|
|
|
t = timeit.timeit(stmt="%s(%s)" % (self.test_method, test_method_args), |
|
|
|
setup=self.full_setup_code(setup_method_args), |
|
|
|
setup=self.full_setup_code(setup_method_args), |
|
|
|
number=reps); |
|
|
|
number=reps); |
|
|
|
return 1.0 * t / reps * (10 ** 9) |
|
|
|
return self.total_bytes * 1.0 / 2 ** 20 / (1.0 * t / reps) |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
if __name__ == "__main__": |
|
|
|
if __name__ == "__main__": |
|
|
@ -144,10 +151,10 @@ if __name__ == "__main__": |
|
|
|
for result in results: |
|
|
|
for result in results: |
|
|
|
print("Message %s of dataset file %s" % \ |
|
|
|
print("Message %s of dataset file %s" % \ |
|
|
|
(result["message_name"], result["filename"])) |
|
|
|
(result["message_name"], result["filename"])) |
|
|
|
print("Average time for parse_from_benchmark: %.2f ns" % \ |
|
|
|
print("Average throughput for parse_from_benchmark: %.2f MB/s" % \ |
|
|
|
(result["benchmarks"][ \ |
|
|
|
(result["benchmarks"][ \ |
|
|
|
args.behavior_prefix + "_parse_from_benchmark"])) |
|
|
|
args.behavior_prefix + "_parse_from_benchmark"])) |
|
|
|
print("Average time for serialize_to_benchmark: %.2f ns" % \ |
|
|
|
print("Average throughput for serialize_to_benchmark: %.2f MB/s" % \ |
|
|
|
(result["benchmarks"][ \ |
|
|
|
(result["benchmarks"][ \ |
|
|
|
args.behavior_prefix + "_serialize_to_benchmark"])) |
|
|
|
args.behavior_prefix + "_serialize_to_benchmark"])) |
|
|
|
print("") |
|
|
|
print("") |
|
|
|