RapidFuzz/bench/benchmark_indel_levenshtein.py

47 lines
1.5 KiB
Python

# todo combine benchmarks of scorers into common code base
import timeit
import pandas
import numpy as np
def benchmark(name, func, setup, lengths, count):
print(f"starting {name}")
start = timeit.default_timer()
results = []
for length in lengths:
test = timeit.Timer(func, setup=setup.format(length, count))
results.append(min(test.timeit(number=1) for _ in range(7)) / count)
stop = timeit.default_timer()
print(f"finished {name}, Runtime: ", stop - start)
return results
setup ="""
from rapidfuzz import string_metric, process, fuzz
import Levenshtein
import string
import random
random.seed(18)
characters = string.ascii_letters + string.digits + string.whitespace + string.punctuation
a = ''.join(random.choice(characters) for _ in range({0}))
b_list = [''.join(random.choice(characters) for _ in range({0})) for _ in range({1})]
"""
lengths = list(range(1,512,2))
count = 3000
time_rapidfuzz = benchmark("rapidfuzz",
'[string_metric.levenshtein(a, b, weights=(1,1,2)) for b in b_list]',
setup, lengths, count)
# this gets very slow, so only benchmark it for smaller values
time_python_levenshtein = benchmark("python-Levenshtein",
'[Levenshtein.ratio(a, b) for b in b_list]',
setup, list(range(1,256,2)), count) + [np.NaN] * 128
df = pandas.DataFrame(data={
"length": lengths,
"rapidfuzz": time_rapidfuzz,
"python-Levenshtein": time_python_levenshtein,
})
df.to_csv("results/levenshtein_indel.csv", sep=',',index=False)