# todo combine benchmarks of scorers into common code base import timeit import pandas import numpy as np def benchmark(name, func, setup, lengths, count): print(f"starting {name}") start = timeit.default_timer() results = [] for length in lengths: test = timeit.Timer(func, setup=setup.format(length, count)) results.append(min(test.timeit(number=1) for _ in range(7)) / count) stop = timeit.default_timer() print(f"finished {name}, Runtime: ", stop - start) return results setup =""" from rapidfuzz import string_metric, process, fuzz import Levenshtein import string import random random.seed(18) characters = string.ascii_letters + string.digits + string.whitespace + string.punctuation a = ''.join(random.choice(characters) for _ in range({0})) b_list = [''.join(random.choice(characters) for _ in range({0})) for _ in range({1})] """ lengths = list(range(1,512,2)) count = 3000 time_rapidfuzz = benchmark("rapidfuzz", '[string_metric.levenshtein(a, b, weights=(1,1,2)) for b in b_list]', setup, lengths, count) # this gets very slow, so only benchmark it for smaller values time_python_levenshtein = benchmark("python-Levenshtein", '[Levenshtein.ratio(a, b) for b in b_list]', setup, list(range(1,256,2)), count) + [np.NaN] * 128 df = pandas.DataFrame(data={ "length": lengths, "rapidfuzz": time_rapidfuzz, "python-Levenshtein": time_python_levenshtein, }) df.to_csv("results/levenshtein_indel.csv", sep=',',index=False)