This is a benchmark suite for different implementations of open source math expression parsers and evaluators written in C++. Currently the following expression parsers are part of this benchmark:
# | Library | Author | License | Numeric Type |
---|---|---|---|---|
00 | ATMSP | Heinz van Saanen | GPL v3 | double, MPFR |
01 | ExprTk | Arash Partow | MIT | double, float, MPFR |
02 | FParser | Juha Nieminen & Joel Yliluoma | LGPL | double |
03 | Lepton | Peter Eastman | MIT | double |
04 | MathExpr | Yann Ollivier | Copyright Notice 1997-2000 | double |
05 | MTParser | Mathieu Jacques | CPOL | double |
06 | muParser | Ingo Berg | MIT | double, float |
07 | muParserX | Ingo Berg | MIT | double, float |
08 | TinyExpr | Lewis Van Winkle | Zlib | double |
Note: The terms double, float etc found in the table above are defined as follows:
- double : IEEE-754 Double-Precision Binary Floating-Point (64-bit)
- float : IEEE-754 Single-precision floating-point format (32-bit)
- MPFR : GNU Multiple Precision Floating-Point Reliably
Note: This archive is not meant as a primary source for any of the libraries listed above as it may contain outdated versions of said libraries.
All the parsers in this benchmark suite are in one way or another POEM (Parse Once Evaluate Many times) in nature. As such the primary objectives for this benchmark are to ascertain the following properties for each of the denoted parsers in following order of priority:
- Correctness and precision of results relative to the floating point type used
- Robustness of the parser given various mathematical expressions
- Evaluation efficiency of 'compiled/parsed' expressions
The benchmark takes two parameters:
- Number of iterations (N)
- Benchmark Test File
The mathematical expression benchmark files are as follows:
- bench_expr.txt
- bench_expr_all.txt
- bench_expr_all_permutations.txt
- bench_expr_complete.txt
- bench_expr_extensive.txt
- bench_expr_precedence.txt
- bench_expr_random_with_functions.txt
- bench_expr_random_without_functions.txt
- bench_expr_weird.txt
The following is a short sampling of expressions taken from the bench_expr.txt expression suite:
- a+1.1
- a*2.2
- 2.2*a+1.1
- (2.2*a+1.1)*3.3
- sin(2.2 * a) + cos(pi / b)
- 1 - sin(2.2 * a) + cos(pi / b)
- sqrt(1 - sin(2.2 * a) + cos(pi / b) / 3.3)
- (a^2 / sin(2 * pi / b)) -a / 2.2
- 1-(a/b*0.5)
- e^log(7*a)
- 10^log(3+b)
- (cos(2.41)/b)
- -(sin(pi+a)+1)
- a-(e^(log(7+b)))
As an example, the following execution will use the"bench_expr_all.txt" set of expressions, executing each expression 100000 times:
ParserBench.exe 100000 bench_expr_all.txt
For every expression in the benchmark file, every parser evaluates the given expression N times, this is known as a round. The total time each parser takes to evaluate the expression N times is recorded. Ranking of the parsers for the round is done from the fastest to the slowest. The following is an example summary for the expression '(a^2/sin(2*pi/b))-a/2':
Expression 64 of 70: "(a^2/sin(2*pi/b))-a/2"; Progress: ############
[01] ExprTk ( 35.863 ns, 3.744853294789905362, -15768714.840563148260116577)
[02] ExprTkFloat ( 37.413 ns, 3.744852304458618164, -15768714.994192123413085938)
[03] muparserSSE ( 44.672 ns, 3.744852304458618164, -15768715.739250183105468750)
[04] muparser 2.2.4 (omp)( 50.093 ns, 3.744853294789905362, -15768714.840563148260116577)
[05] MTParser ( 54.681 ns, 3.744853294789905362, -15768714.840563148260116577)
[06] muparser 2.2.4 ( 55.816 ns, 3.744853294789905362, -15768714.840563148260116577)
[07] FParser 4.5 ( 65.141 ns, 3.744853294789905362, -15768714.840563148260116577)
[08] atmsp 1.0.4 ( 78.809 ns, 3.744853294789905362, -15768714.840563148260116577)
[09] MathExpr (131.242 ns, 3.744853294789905362, -15768714.840563148260116577)
[10] TinyExpr (168.341 ns, 3.744853294789905362, -15768714.840563148260116577)
[11] Lepton (402.011 ns, 3.744853294789905362, -15768714.840563148260116577)
[12] muparserx (572.611 ns, 3.744853294789905362, -15768714.840563148260116577)
Parsers that can't parse the given expression or produce a result different to the expected result (based on a normalized epsilon approach) are disqualified for the round and do not participate in point accumulation. The following is an example where a few of the parsers failed to either parse or evaluate the given expression of '+1.1+a^b' correctly:
Expression 63 of 96: "+a^+b+1.1"; Progress: ############
[01] ExprTk ( 40.631 ns, 2.333286300554663129, 14534421.395495440810918808)
[02] ExprTkFloat ( 44.212 ns, 2.333286285400390625, 14534422.159194946289062500)
[03] muparserSSE ( 44.556 ns, 2.333286285400390625, 14534422.159194946289062500)
[04] atmsp 1.0.4 ( 47.906 ns, 2.333286300554663129, 14534421.395495440810918808)
[05] TinyExpr ( 52.831 ns, 2.333286300554663129, 14534421.395495440810918808)
[06] MTParser ( 56.045 ns, 2.333286300554663129, 14534421.395495440810918808)
[07] muparser 2.2.4 ( 67.061 ns, 2.333286300554663129, 14534421.395495440810918808)
[08] muparser 2.2.4 (omp)( 67.231 ns, 2.333286300554663129, 14534421.395495440810918808)
[09] muparserx (251.338 ns, 2.333286300554663129, 14534421.395495440810918808)
DNQ List
[01] FParser 4.5 (Syntax error)
[02] MathExpr (parsing error)
[03] Lepton (Parse error in expression "+a^+b+1.1 ": unexpected token: +)
**** ERROR **** Excessive number of evaluation failures! [3]
The meanings of each of the columns are as follows:
- Rank
- Parser
- Average time (in nano-seconds) to evaluate the expression once, taken over N-times
- Result of evaluating the expression
- Sum of N-evaluations of the given expression
Once all the expressions have been completed, a summary is provided, that includes information relating to the build parameters of the benchmark binary, the architecture which the benchmark was run upon, a final ranking of the parsers, and a listing of expressions per parser that were deemed as having been disqualified during the benchmark.
The following is an example summary generated at the completion of the benchmark:
Benchmark settings:
- Expressions File is "bench_expr.txt"
- Reference parser is ExprTk
- Iterations per expression: 5000000
- Number of expressions: 70
- Number of excessive failures: 5
- Release build
- Compiled with MSVC Version 1900
- IEEE 754 (IEC 559) is Available
- 32-bit build
- CPU Name : Intel(R) Xeon(R) CPU E5-2687W 0 @ 3.10GHz
- Num. cores : 16
- Num. logical : 32
- Total logical : 16
- L1 D cache : 32 KB
- L1 I cache : 32 KB
- L2 cache : 256 KB
- L3 cache : 20480 KB
- L1D assoc. : 8-way
- L2 assoc. : 8-way
- L3 assoc. : 20-way
- L1D line size : 64 bytes
- L2 line size : 64 bytes
- L3 line size : 64 bytes
- CPU clock : 3092 MHz
- Features : acpi aes apic avx clflush cmov cx16 cx8 dca de ds_cpl
dts dts64 est fpu fxsr ht lahf_lm lm mca mce mmx monitor
msr mtrr osxsave pae pat pbe pclmul pdcm pge pni popcnt
pse pse36 sep smx ss sse sse2 sse4_1 sse4_2 ssse3 tm tm2
tsc vme vmx xd xsave xtpr
Scores:
# Parser Type Points Score Failures
-------------------------------------------------------------
00 ExprTk double 781 100 0
01 ExprTkFloat float 643 82 9
02 muparserSSE float 639 93 9
03 FParser 4.5 double 536 45 0
04 atmsp 1.0.4 double 482 40 2
05 muparser 2.2.4 double 445 37 0
06 MTParser double 402 38 10
07 muparser 2.2.4 (omp) double 374 35 0
08 MathExpr double 345 30 2
09 TinyExpr double 296 30 2
10 Lepton double 111 6 2
11 muparserx double 97 5 0
The summaries of runs of the benchmark suite upon various architectures can be found here: Results
The following is a chart depicting the results as executions per second obtained by running the 'bench_expr_random_without_functions.txt' benchmark, using the double type, 5000000 iterations per expression on an Intel Xeon E5-2687W 3GHz. The horizontal measure on the chart is the length of the expression in bytes.
Note: The parsers muParserSSE and ExprTKFloat were excluded due to their use of the float type, and the Lepton and muParserX parsers were excluded due to the fact that their results were consistently 2-3 orders of magnitude slower than all the others.
Note: The command executed to generate the above results is as follows:
ParserBenchmark.exe 5000000 bench_expr_random_without_functions.txt write_table
The benchmark has a special mode called "Native". In this mode a predefined set of expressions are employed, and benchmarked against both native equivalent implementations and the standard parsers/evaluators.
The command to execute the benchmark in native mode is as follows:
ParserBenchmark.exe 1000000 native
Creating a customised test is very simple. All that is required is to create a utf-8 encoded text file, and pass it as the 2nd parameter to the benchmark executable.
Each line can only have one expression, comprised of the following:
- Operators: +, -, *, /, ^
- Variables: a, b, c, x, y, z, w
- Constants: e and pi
- Functions: sin, cos, tan, abs, exp, sqrt, log, pow
- Parentheses: Round '(' and ')'
Furthermore blank lines or lines beginning with a #-symbol will be ignored (aka comments or white-space).
The following is a simple example of a custom test:
---- snip ----
# My first expression:
(x + 1) / 2 * b
# Pythagoras
a^2 + b^2
# A constant expression
1 + (2 * 3) / 7
# Definition of tan in terms of sine and cosine
sin(x) / cos(x)
---- snip ----
The addition of new or unlisted C++ mathematical expression parsers is always welcome. If you know of a parser and would like it included in the suite feel free to make a request. Additionally if the associated benchmark implementation (eg: src/BenchXYZW.cpp) could be made available as part of the request that would very helpful.