Protocol Buffers - Google's data interchange format (grpc依赖)
https://developers.google.com/protocol-buffers/
You can not select more than 25 topics
Topics must start with a letter or number, can include dashes ('-') and can be up to 35 characters long.
117 lines
5.0 KiB
117 lines
5.0 KiB
import sys |
|
import os |
|
import timeit |
|
import math |
|
import fnmatch |
|
|
|
# BEGIN CPP GENERATED MESSAGE |
|
# CPP generated code must be linked before importing the generated Python code |
|
# for the descriptor can be found in the pool |
|
if len(sys.argv) < 2: |
|
raise IOError("Need string argument \"true\" or \"false\" for whether to use cpp generated code") |
|
if sys.argv[1] == "true": |
|
sys.path.append( os.path.dirname( os.path.dirname( os.path.abspath(__file__) ) ) + "/.libs" ) |
|
import libbenchmark_messages |
|
sys.path.append( os.path.dirname( os.path.dirname( os.path.abspath(__file__) ) ) + "/tmp" ) |
|
elif sys.argv[1] != "false": |
|
raise IOError("Need string argument \"true\" or \"false\" for whether to use cpp generated code") |
|
# END CPP GENERATED MESSAGE |
|
|
|
import datasets.google_message1.benchmark_message1_proto2_pb2 as benchmark_message1_proto2_pb2 |
|
import datasets.google_message1.benchmark_message1_proto3_pb2 as benchmark_message1_proto3_pb2 |
|
import datasets.google_message2.benchmark_message2_pb2 as benchmark_message2_pb2 |
|
import datasets.google_message3.benchmark_message3_pb2 as benchmark_message3_pb2 |
|
import datasets.google_message4.benchmark_message4_pb2 as benchmark_message4_pb2 |
|
import benchmarks_pb2 as benchmarks_pb2 |
|
|
|
|
|
def run_one_test(filename): |
|
data = open(os.path.dirname(sys.argv[0]) + "/../" + filename).read() |
|
benchmark_dataset = benchmarks_pb2.BenchmarkDataset() |
|
benchmark_dataset.ParseFromString(data) |
|
benchmark_util = Benchmark(full_iteration=len(benchmark_dataset.payload), |
|
module="py_benchmark", |
|
setup_method="init") |
|
print "Message %s of dataset file %s" % \ |
|
(benchmark_dataset.message_name, filename) |
|
benchmark_util.set_test_method("parse_from_benchmark") |
|
print benchmark_util.run_benchmark(setup_method_args='"%s"' % (filename)) |
|
benchmark_util.set_test_method("serialize_to_benchmark") |
|
print benchmark_util.run_benchmark(setup_method_args='"%s"' % (filename)) |
|
print "" |
|
|
|
def init(filename): |
|
global benchmark_dataset, message_class, message_list, counter |
|
message_list=[] |
|
counter = 0 |
|
data = open(os.path.dirname(sys.argv[0]) + "/../" + filename).read() |
|
benchmark_dataset = benchmarks_pb2.BenchmarkDataset() |
|
benchmark_dataset.ParseFromString(data) |
|
|
|
if benchmark_dataset.message_name == "benchmarks.proto3.GoogleMessage1": |
|
message_class = benchmark_message1_proto3_pb2.GoogleMessage1 |
|
elif benchmark_dataset.message_name == "benchmarks.proto2.GoogleMessage1": |
|
message_class = benchmark_message1_proto2_pb2.GoogleMessage1 |
|
elif benchmark_dataset.message_name == "benchmarks.proto2.GoogleMessage2": |
|
message_class = benchmark_message2_pb2.GoogleMessage2 |
|
elif benchmark_dataset.message_name == "benchmarks.google_message3.GoogleMessage3": |
|
message_class = benchmark_message3_pb2.GoogleMessage3 |
|
elif benchmark_dataset.message_name == "benchmarks.google_message4.GoogleMessage4": |
|
message_class = benchmark_message4_pb2.GoogleMessage4 |
|
else: |
|
raise IOError("Message %s not found!" % (benchmark_dataset.message_name)) |
|
|
|
for one_payload in benchmark_dataset.payload: |
|
temp = message_class() |
|
temp.ParseFromString(one_payload) |
|
message_list.append(temp) |
|
|
|
def parse_from_benchmark(): |
|
global counter, message_class, benchmark_dataset |
|
m = message_class().ParseFromString(benchmark_dataset.payload[counter % len(benchmark_dataset.payload)]) |
|
counter = counter + 1 |
|
|
|
def serialize_to_benchmark(): |
|
global counter, message_list, message_class |
|
s = message_list[counter % len(benchmark_dataset.payload)].SerializeToString() |
|
counter = counter + 1 |
|
|
|
|
|
class Benchmark: |
|
def __init__(self, module=None, test_method=None, |
|
setup_method=None, full_iteration = 1): |
|
self.full_iteration = full_iteration |
|
self.module = module |
|
self.test_method = test_method |
|
self.setup_method = setup_method |
|
|
|
def set_test_method(self, test_method): |
|
self.test_method = test_method |
|
|
|
def full_setup_code(self, setup_method_args=''): |
|
setup_code = "" |
|
setup_code += "from %s import %s\n" % (self.module, self.test_method) |
|
setup_code += "from %s import %s\n" % (self.module, self.setup_method) |
|
setup_code += "%s(%s)\n" % (self.setup_method, setup_method_args) |
|
return setup_code |
|
|
|
def dry_run(self, test_method_args='', setup_method_args=''): |
|
return timeit.timeit(stmt="%s(%s)" % (self.test_method, test_method_args), |
|
setup=self.full_setup_code(setup_method_args), |
|
number=self.full_iteration); |
|
|
|
def run_benchmark(self, test_method_args='', setup_method_args=''): |
|
reps = self.full_iteration; |
|
t = self.dry_run(test_method_args, setup_method_args); |
|
if t < 3 : |
|
reps = int(math.ceil(3 / t)) * self.full_iteration |
|
t = timeit.timeit(stmt="%s(%s)" % (self.test_method, test_method_args), |
|
setup=self.full_setup_code(setup_method_args), |
|
number=reps); |
|
return "Average time for %s: %.2f ns" % \ |
|
(self.test_method, 1.0 * t / reps * (10 ** 9)) |
|
|
|
|
|
if __name__ == "__main__": |
|
for i in range(2, len(sys.argv)): |
|
run_one_test(sys.argv[i])
|
|
|