known benchmarks for floating point calculations?











up vote
5
down vote

favorite
1












I frequently have the problem that I have to perform some simplistic calculations, mostly products and sums of two or three quantities. I used to use the fp package for it, but it does look somewhat unmaintained. Now I've tried pgfmath, and like everything coming from the TikZ direction it is outrageously complete, thoroughly documented, and has a nice API with a great syntax to top it off. And I'm using TikZ for graphics anyway, so there's no additional overhead in terms of packages used.



The only thing that worries me is performance. I could devise some tests to get a better idea of how fast pgfmath is in comparison, but I thought I'd better ask around first:




  • are there any articles that compare different solutions to perform simple arithmetics in TeX


  • that include benchmarks?


  • What is the best practice when one wants to do multiplication of fractional, signed numbers? Neither precision nor magnitude have to be all that great, given that these numbers are used to do typesetting, so everything that is accurate to the tenth of a millimeter or so for dimensions up to several hundreds of millimeters should be fine.











share|improve this question


















  • 1




    I don't know of any comparison between floating point packages, but you can do your own benchmark the recently released l3benchmark.
    – Phelype Oleinik
    Dec 6 at 17:45










  • See tex.stackexchange.com/questions/15526/… for general disucssion of possible options
    – Joseph Wright
    Dec 6 at 18:50















up vote
5
down vote

favorite
1












I frequently have the problem that I have to perform some simplistic calculations, mostly products and sums of two or three quantities. I used to use the fp package for it, but it does look somewhat unmaintained. Now I've tried pgfmath, and like everything coming from the TikZ direction it is outrageously complete, thoroughly documented, and has a nice API with a great syntax to top it off. And I'm using TikZ for graphics anyway, so there's no additional overhead in terms of packages used.



The only thing that worries me is performance. I could devise some tests to get a better idea of how fast pgfmath is in comparison, but I thought I'd better ask around first:




  • are there any articles that compare different solutions to perform simple arithmetics in TeX


  • that include benchmarks?


  • What is the best practice when one wants to do multiplication of fractional, signed numbers? Neither precision nor magnitude have to be all that great, given that these numbers are used to do typesetting, so everything that is accurate to the tenth of a millimeter or so for dimensions up to several hundreds of millimeters should be fine.











share|improve this question


















  • 1




    I don't know of any comparison between floating point packages, but you can do your own benchmark the recently released l3benchmark.
    – Phelype Oleinik
    Dec 6 at 17:45










  • See tex.stackexchange.com/questions/15526/… for general disucssion of possible options
    – Joseph Wright
    Dec 6 at 18:50













up vote
5
down vote

favorite
1









up vote
5
down vote

favorite
1






1





I frequently have the problem that I have to perform some simplistic calculations, mostly products and sums of two or three quantities. I used to use the fp package for it, but it does look somewhat unmaintained. Now I've tried pgfmath, and like everything coming from the TikZ direction it is outrageously complete, thoroughly documented, and has a nice API with a great syntax to top it off. And I'm using TikZ for graphics anyway, so there's no additional overhead in terms of packages used.



The only thing that worries me is performance. I could devise some tests to get a better idea of how fast pgfmath is in comparison, but I thought I'd better ask around first:




  • are there any articles that compare different solutions to perform simple arithmetics in TeX


  • that include benchmarks?


  • What is the best practice when one wants to do multiplication of fractional, signed numbers? Neither precision nor magnitude have to be all that great, given that these numbers are used to do typesetting, so everything that is accurate to the tenth of a millimeter or so for dimensions up to several hundreds of millimeters should be fine.











share|improve this question













I frequently have the problem that I have to perform some simplistic calculations, mostly products and sums of two or three quantities. I used to use the fp package for it, but it does look somewhat unmaintained. Now I've tried pgfmath, and like everything coming from the TikZ direction it is outrageously complete, thoroughly documented, and has a nice API with a great syntax to top it off. And I'm using TikZ for graphics anyway, so there's no additional overhead in terms of packages used.



The only thing that worries me is performance. I could devise some tests to get a better idea of how fast pgfmath is in comparison, but I thought I'd better ask around first:




  • are there any articles that compare different solutions to perform simple arithmetics in TeX


  • that include benchmarks?


  • What is the best practice when one wants to do multiplication of fractional, signed numbers? Neither precision nor magnitude have to be all that great, given that these numbers are used to do typesetting, so everything that is accurate to the tenth of a millimeter or so for dimensions up to several hundreds of millimeters should be fine.








arithmetic






share|improve this question













share|improve this question











share|improve this question




share|improve this question










asked Dec 6 at 17:39









John Frazer

1435




1435








  • 1




    I don't know of any comparison between floating point packages, but you can do your own benchmark the recently released l3benchmark.
    – Phelype Oleinik
    Dec 6 at 17:45










  • See tex.stackexchange.com/questions/15526/… for general disucssion of possible options
    – Joseph Wright
    Dec 6 at 18:50














  • 1




    I don't know of any comparison between floating point packages, but you can do your own benchmark the recently released l3benchmark.
    – Phelype Oleinik
    Dec 6 at 17:45










  • See tex.stackexchange.com/questions/15526/… for general disucssion of possible options
    – Joseph Wright
    Dec 6 at 18:50








1




1




I don't know of any comparison between floating point packages, but you can do your own benchmark the recently released l3benchmark.
– Phelype Oleinik
Dec 6 at 17:45




I don't know of any comparison between floating point packages, but you can do your own benchmark the recently released l3benchmark.
– Phelype Oleinik
Dec 6 at 17:45












See tex.stackexchange.com/questions/15526/… for general disucssion of possible options
– Joseph Wright
Dec 6 at 18:50




See tex.stackexchange.com/questions/15526/… for general disucssion of possible options
– Joseph Wright
Dec 6 at 18:50










2 Answers
2






active

oldest

votes

















up vote
5
down vote



accepted










As well as fp and the pgf maths parser, I think I would evaluate xfp/LaTeX3 FPU here. With a simple test set up such as



documentclass{article}
usepackage{fp}
usepackage{xfp}
usepackage{tikz}
usepackage{l3benchmark}
ExplSyntaxOn
cs_new_eq:NN benchmark benchmark:n
ExplSyntaxOff
FPmessagesfalse
newsavebox{testbox}
begin{document}
benchmark{sbox{testbox}{FPupnresult{2 2 root 40 sin *}}}
benchmark{sbox{testbox}{fpeval{sqrt(2) * sind(40)}}}
benchmark{sbox{testbox}{pgfmathparse{sqrt(2) * sin(40)}}}
end{document}


on my system I get



0.00556 seconds (2.03e4 ops)
5.06e-4 seconds (1.83e3 ops)
2.23e-4 seconds (819 ops)


Unsurprisingly, fp is slowest as it works to a huge number of places. On the other hand, the pgf unit is fastest but not by much over the LaTeX3 code. The pgf code is not expandable, and uses dimens internally, trading accuracy for speed. The latter is perfectly reasonable, but of course may or may not be acceptable depending on use case.



(For the test, I've used the UPN part of fp largely as it seems fairest: the other two options offer parsing of expressions ...)



For completeness, if you are using LuaTeX then you can do the same using Lua, which is very fast:



benchmark{sbox{testbox}{directlua{tex.print(math.sqrt(2) * math.sin(40))}}}


gives 5.1e-5 seconds (187 ops) seconds on my test setup.





Worth noting of course is that the speed does depend on the exact operation: if I go for a simple sum, pgf and the LaTeX3 FPU are comparable:



documentclass{article}
usepackage{fp}
usepackage{xfp}
usepackage{tikz}
usepackage{l3benchmark}
ExplSyntaxOn
cs_new_eq:NN benchmark benchmark:n
ExplSyntaxOff
FPmessagesfalse
newsavebox{testbox}
begin{document}
benchmark{sbox{testbox}{FPupnresult{1.234 5 * 9.10 6.78 / +}}}
benchmark{sbox{testbox}{fpeval{1.234 * 5 + 6.78 / 9.10}}}
benchmark{sbox{testbox}{pgfmathparse{1.234 * 5 + 6.78 / 9.10}}}
end{document}


gives



0.00231 seconds (8.42e3 ops)
2.25e-4 seconds (837 ops)
2.63e-4 seconds (930 ops)




If you want simple dimension calculations, nothing is going to beat primitive, most obviously dimexpr. Something like



thedimexpr 1.2cm + 3.445cmrelax


is 'clocked' by the benchmark code at 2.64e-6 seconds (9.85 ops) on my system: really, really fast.






share|improve this answer























  • The benchmark code does multiple cycles, and here is reporting the minimum time for one run.
    – Joseph Wright
    Dec 6 at 17:57










  • the l3benchmark output prints out result with 1:10^16 relative precision (natural, as this the FPU precision), but those last 4, 5, 6, or even 8 digits have no real meaning, the unit in the last place in your examples is something like 10^{-19} seconds, but my CPU is 2 GHz, 2 10^9, i think 10 GHz is a safe overestimate of current machines, this is 10^10 operations per second, so 10^{-19} times 10^{10} means 0.000000001 processor operation... I trust the reported timings could be rounded to at most 6 digits precision with no loss of any significance. (and pdfelapsedtime is very volatile).
    – jfbu
    Dec 6 at 19:49










  • @jfbu Sure, but this is still experimental ... Bruno wants to add rounding to sig figs for this purpose
    – Joseph Wright
    Dec 6 at 19:54






  • 1




    @jfbu Results updated: we altered the output a bit
    – Joseph Wright
    Dec 9 at 9:46






  • 1




    I think it would make sense to mention that pgf computes for instance sin(40) with 5 decimal places while xfp/l3fp computes it to 16 decimal places in barely twice longer.
    – Bruno Le Floch
    Dec 9 at 10:12


















up vote
1
down vote













tl;dr: pgfmath is 10 Times faster than fp



So I went and did a little test that I feel is sufficiently realistic for my intended purpose: typesetting dozens and hundred of pages with like a dozen floating point multiplications per page.



With around 80 pages of typesetting material my test document needed roughly a1 = 6 seconds to complete typesetting with XeLaTeX using pgfmath; with three times as much stuff, that rose to b1 = 13 seconds. In comparison, doing the same using fp, the timings came out as a2 = 37 and b2 = 107 seconds. Clearly, there must be a certain overhead here, which I estimated using the assumption that ( a1 - c ) * 3 = b1 - c and ( a2 - c ) * 3 = b2 - c, giving me a constant overhead of around c = 2.5 seconds. Timings with that overhead subtracted, then, are a1c = 3.5, b1c = 10.5, a2c = 34.5, b2c = 104.5.



While these measurements are all done in a rough way and with lots of hand-waving, they still seem to indicate that fp takes ten times as long as pgfmath.



There's still all kinds of wiggle room in here; I haven't checked whether the fp calculations couldn't have been optimized (which could narrow the gap), and I did not use the short-circuited versions of pgfmathparse and friends (which could widen the gap), but, on the other hand, a 1 / 10 ratio is unlikely to be easily done away with.



I haven't done tests with xfp / LaTeX3 FPU yet, as suggested by Joseph Wright; I did look up that stuff on CTAN and it does look very promising with regard to the future of LaTeX. It's really bleeding-edge and none of that is included in my TeX Live 2016 installation which I'm loath to update at this point in time. But certainly something to consider for the future.



My figures differ from those of Joseph Wright, so feel free to critique my methodology.






share|improve this answer





















  • The factor of 10 seems to agree with Joseph Wright's results.
    – Bruno Le Floch
    Dec 9 at 10:09











Your Answer








StackExchange.ready(function() {
var channelOptions = {
tags: "".split(" "),
id: "85"
};
initTagRenderer("".split(" "), "".split(" "), channelOptions);

StackExchange.using("externalEditor", function() {
// Have to fire editor after snippets, if snippets enabled
if (StackExchange.settings.snippets.snippetsEnabled) {
StackExchange.using("snippets", function() {
createEditor();
});
}
else {
createEditor();
}
});

function createEditor() {
StackExchange.prepareEditor({
heartbeatType: 'answer',
convertImagesToLinks: false,
noModals: true,
showLowRepImageUploadWarning: true,
reputationToPostImages: null,
bindNavPrevention: true,
postfix: "",
imageUploader: {
brandingHtml: "Powered by u003ca class="icon-imgur-white" href="https://imgur.com/"u003eu003c/au003e",
contentPolicyHtml: "User contributions licensed under u003ca href="https://creativecommons.org/licenses/by-sa/3.0/"u003ecc by-sa 3.0 with attribution requiredu003c/au003e u003ca href="https://stackoverflow.com/legal/content-policy"u003e(content policy)u003c/au003e",
allowUrls: true
},
onDemand: true,
discardSelector: ".discard-answer"
,immediatelyShowMarkdownHelp:true
});


}
});














draft saved

draft discarded


















StackExchange.ready(
function () {
StackExchange.openid.initPostLogin('.new-post-login', 'https%3a%2f%2ftex.stackexchange.com%2fquestions%2f463554%2fknown-benchmarks-for-floating-point-calculations%23new-answer', 'question_page');
}
);

Post as a guest















Required, but never shown

























2 Answers
2






active

oldest

votes








2 Answers
2






active

oldest

votes









active

oldest

votes






active

oldest

votes








up vote
5
down vote



accepted










As well as fp and the pgf maths parser, I think I would evaluate xfp/LaTeX3 FPU here. With a simple test set up such as



documentclass{article}
usepackage{fp}
usepackage{xfp}
usepackage{tikz}
usepackage{l3benchmark}
ExplSyntaxOn
cs_new_eq:NN benchmark benchmark:n
ExplSyntaxOff
FPmessagesfalse
newsavebox{testbox}
begin{document}
benchmark{sbox{testbox}{FPupnresult{2 2 root 40 sin *}}}
benchmark{sbox{testbox}{fpeval{sqrt(2) * sind(40)}}}
benchmark{sbox{testbox}{pgfmathparse{sqrt(2) * sin(40)}}}
end{document}


on my system I get



0.00556 seconds (2.03e4 ops)
5.06e-4 seconds (1.83e3 ops)
2.23e-4 seconds (819 ops)


Unsurprisingly, fp is slowest as it works to a huge number of places. On the other hand, the pgf unit is fastest but not by much over the LaTeX3 code. The pgf code is not expandable, and uses dimens internally, trading accuracy for speed. The latter is perfectly reasonable, but of course may or may not be acceptable depending on use case.



(For the test, I've used the UPN part of fp largely as it seems fairest: the other two options offer parsing of expressions ...)



For completeness, if you are using LuaTeX then you can do the same using Lua, which is very fast:



benchmark{sbox{testbox}{directlua{tex.print(math.sqrt(2) * math.sin(40))}}}


gives 5.1e-5 seconds (187 ops) seconds on my test setup.





Worth noting of course is that the speed does depend on the exact operation: if I go for a simple sum, pgf and the LaTeX3 FPU are comparable:



documentclass{article}
usepackage{fp}
usepackage{xfp}
usepackage{tikz}
usepackage{l3benchmark}
ExplSyntaxOn
cs_new_eq:NN benchmark benchmark:n
ExplSyntaxOff
FPmessagesfalse
newsavebox{testbox}
begin{document}
benchmark{sbox{testbox}{FPupnresult{1.234 5 * 9.10 6.78 / +}}}
benchmark{sbox{testbox}{fpeval{1.234 * 5 + 6.78 / 9.10}}}
benchmark{sbox{testbox}{pgfmathparse{1.234 * 5 + 6.78 / 9.10}}}
end{document}


gives



0.00231 seconds (8.42e3 ops)
2.25e-4 seconds (837 ops)
2.63e-4 seconds (930 ops)




If you want simple dimension calculations, nothing is going to beat primitive, most obviously dimexpr. Something like



thedimexpr 1.2cm + 3.445cmrelax


is 'clocked' by the benchmark code at 2.64e-6 seconds (9.85 ops) on my system: really, really fast.






share|improve this answer























  • The benchmark code does multiple cycles, and here is reporting the minimum time for one run.
    – Joseph Wright
    Dec 6 at 17:57










  • the l3benchmark output prints out result with 1:10^16 relative precision (natural, as this the FPU precision), but those last 4, 5, 6, or even 8 digits have no real meaning, the unit in the last place in your examples is something like 10^{-19} seconds, but my CPU is 2 GHz, 2 10^9, i think 10 GHz is a safe overestimate of current machines, this is 10^10 operations per second, so 10^{-19} times 10^{10} means 0.000000001 processor operation... I trust the reported timings could be rounded to at most 6 digits precision with no loss of any significance. (and pdfelapsedtime is very volatile).
    – jfbu
    Dec 6 at 19:49










  • @jfbu Sure, but this is still experimental ... Bruno wants to add rounding to sig figs for this purpose
    – Joseph Wright
    Dec 6 at 19:54






  • 1




    @jfbu Results updated: we altered the output a bit
    – Joseph Wright
    Dec 9 at 9:46






  • 1




    I think it would make sense to mention that pgf computes for instance sin(40) with 5 decimal places while xfp/l3fp computes it to 16 decimal places in barely twice longer.
    – Bruno Le Floch
    Dec 9 at 10:12















up vote
5
down vote



accepted










As well as fp and the pgf maths parser, I think I would evaluate xfp/LaTeX3 FPU here. With a simple test set up such as



documentclass{article}
usepackage{fp}
usepackage{xfp}
usepackage{tikz}
usepackage{l3benchmark}
ExplSyntaxOn
cs_new_eq:NN benchmark benchmark:n
ExplSyntaxOff
FPmessagesfalse
newsavebox{testbox}
begin{document}
benchmark{sbox{testbox}{FPupnresult{2 2 root 40 sin *}}}
benchmark{sbox{testbox}{fpeval{sqrt(2) * sind(40)}}}
benchmark{sbox{testbox}{pgfmathparse{sqrt(2) * sin(40)}}}
end{document}


on my system I get



0.00556 seconds (2.03e4 ops)
5.06e-4 seconds (1.83e3 ops)
2.23e-4 seconds (819 ops)


Unsurprisingly, fp is slowest as it works to a huge number of places. On the other hand, the pgf unit is fastest but not by much over the LaTeX3 code. The pgf code is not expandable, and uses dimens internally, trading accuracy for speed. The latter is perfectly reasonable, but of course may or may not be acceptable depending on use case.



(For the test, I've used the UPN part of fp largely as it seems fairest: the other two options offer parsing of expressions ...)



For completeness, if you are using LuaTeX then you can do the same using Lua, which is very fast:



benchmark{sbox{testbox}{directlua{tex.print(math.sqrt(2) * math.sin(40))}}}


gives 5.1e-5 seconds (187 ops) seconds on my test setup.





Worth noting of course is that the speed does depend on the exact operation: if I go for a simple sum, pgf and the LaTeX3 FPU are comparable:



documentclass{article}
usepackage{fp}
usepackage{xfp}
usepackage{tikz}
usepackage{l3benchmark}
ExplSyntaxOn
cs_new_eq:NN benchmark benchmark:n
ExplSyntaxOff
FPmessagesfalse
newsavebox{testbox}
begin{document}
benchmark{sbox{testbox}{FPupnresult{1.234 5 * 9.10 6.78 / +}}}
benchmark{sbox{testbox}{fpeval{1.234 * 5 + 6.78 / 9.10}}}
benchmark{sbox{testbox}{pgfmathparse{1.234 * 5 + 6.78 / 9.10}}}
end{document}


gives



0.00231 seconds (8.42e3 ops)
2.25e-4 seconds (837 ops)
2.63e-4 seconds (930 ops)




If you want simple dimension calculations, nothing is going to beat primitive, most obviously dimexpr. Something like



thedimexpr 1.2cm + 3.445cmrelax


is 'clocked' by the benchmark code at 2.64e-6 seconds (9.85 ops) on my system: really, really fast.






share|improve this answer























  • The benchmark code does multiple cycles, and here is reporting the minimum time for one run.
    – Joseph Wright
    Dec 6 at 17:57










  • the l3benchmark output prints out result with 1:10^16 relative precision (natural, as this the FPU precision), but those last 4, 5, 6, or even 8 digits have no real meaning, the unit in the last place in your examples is something like 10^{-19} seconds, but my CPU is 2 GHz, 2 10^9, i think 10 GHz is a safe overestimate of current machines, this is 10^10 operations per second, so 10^{-19} times 10^{10} means 0.000000001 processor operation... I trust the reported timings could be rounded to at most 6 digits precision with no loss of any significance. (and pdfelapsedtime is very volatile).
    – jfbu
    Dec 6 at 19:49










  • @jfbu Sure, but this is still experimental ... Bruno wants to add rounding to sig figs for this purpose
    – Joseph Wright
    Dec 6 at 19:54






  • 1




    @jfbu Results updated: we altered the output a bit
    – Joseph Wright
    Dec 9 at 9:46






  • 1




    I think it would make sense to mention that pgf computes for instance sin(40) with 5 decimal places while xfp/l3fp computes it to 16 decimal places in barely twice longer.
    – Bruno Le Floch
    Dec 9 at 10:12













up vote
5
down vote



accepted







up vote
5
down vote



accepted






As well as fp and the pgf maths parser, I think I would evaluate xfp/LaTeX3 FPU here. With a simple test set up such as



documentclass{article}
usepackage{fp}
usepackage{xfp}
usepackage{tikz}
usepackage{l3benchmark}
ExplSyntaxOn
cs_new_eq:NN benchmark benchmark:n
ExplSyntaxOff
FPmessagesfalse
newsavebox{testbox}
begin{document}
benchmark{sbox{testbox}{FPupnresult{2 2 root 40 sin *}}}
benchmark{sbox{testbox}{fpeval{sqrt(2) * sind(40)}}}
benchmark{sbox{testbox}{pgfmathparse{sqrt(2) * sin(40)}}}
end{document}


on my system I get



0.00556 seconds (2.03e4 ops)
5.06e-4 seconds (1.83e3 ops)
2.23e-4 seconds (819 ops)


Unsurprisingly, fp is slowest as it works to a huge number of places. On the other hand, the pgf unit is fastest but not by much over the LaTeX3 code. The pgf code is not expandable, and uses dimens internally, trading accuracy for speed. The latter is perfectly reasonable, but of course may or may not be acceptable depending on use case.



(For the test, I've used the UPN part of fp largely as it seems fairest: the other two options offer parsing of expressions ...)



For completeness, if you are using LuaTeX then you can do the same using Lua, which is very fast:



benchmark{sbox{testbox}{directlua{tex.print(math.sqrt(2) * math.sin(40))}}}


gives 5.1e-5 seconds (187 ops) seconds on my test setup.





Worth noting of course is that the speed does depend on the exact operation: if I go for a simple sum, pgf and the LaTeX3 FPU are comparable:



documentclass{article}
usepackage{fp}
usepackage{xfp}
usepackage{tikz}
usepackage{l3benchmark}
ExplSyntaxOn
cs_new_eq:NN benchmark benchmark:n
ExplSyntaxOff
FPmessagesfalse
newsavebox{testbox}
begin{document}
benchmark{sbox{testbox}{FPupnresult{1.234 5 * 9.10 6.78 / +}}}
benchmark{sbox{testbox}{fpeval{1.234 * 5 + 6.78 / 9.10}}}
benchmark{sbox{testbox}{pgfmathparse{1.234 * 5 + 6.78 / 9.10}}}
end{document}


gives



0.00231 seconds (8.42e3 ops)
2.25e-4 seconds (837 ops)
2.63e-4 seconds (930 ops)




If you want simple dimension calculations, nothing is going to beat primitive, most obviously dimexpr. Something like



thedimexpr 1.2cm + 3.445cmrelax


is 'clocked' by the benchmark code at 2.64e-6 seconds (9.85 ops) on my system: really, really fast.






share|improve this answer














As well as fp and the pgf maths parser, I think I would evaluate xfp/LaTeX3 FPU here. With a simple test set up such as



documentclass{article}
usepackage{fp}
usepackage{xfp}
usepackage{tikz}
usepackage{l3benchmark}
ExplSyntaxOn
cs_new_eq:NN benchmark benchmark:n
ExplSyntaxOff
FPmessagesfalse
newsavebox{testbox}
begin{document}
benchmark{sbox{testbox}{FPupnresult{2 2 root 40 sin *}}}
benchmark{sbox{testbox}{fpeval{sqrt(2) * sind(40)}}}
benchmark{sbox{testbox}{pgfmathparse{sqrt(2) * sin(40)}}}
end{document}


on my system I get



0.00556 seconds (2.03e4 ops)
5.06e-4 seconds (1.83e3 ops)
2.23e-4 seconds (819 ops)


Unsurprisingly, fp is slowest as it works to a huge number of places. On the other hand, the pgf unit is fastest but not by much over the LaTeX3 code. The pgf code is not expandable, and uses dimens internally, trading accuracy for speed. The latter is perfectly reasonable, but of course may or may not be acceptable depending on use case.



(For the test, I've used the UPN part of fp largely as it seems fairest: the other two options offer parsing of expressions ...)



For completeness, if you are using LuaTeX then you can do the same using Lua, which is very fast:



benchmark{sbox{testbox}{directlua{tex.print(math.sqrt(2) * math.sin(40))}}}


gives 5.1e-5 seconds (187 ops) seconds on my test setup.





Worth noting of course is that the speed does depend on the exact operation: if I go for a simple sum, pgf and the LaTeX3 FPU are comparable:



documentclass{article}
usepackage{fp}
usepackage{xfp}
usepackage{tikz}
usepackage{l3benchmark}
ExplSyntaxOn
cs_new_eq:NN benchmark benchmark:n
ExplSyntaxOff
FPmessagesfalse
newsavebox{testbox}
begin{document}
benchmark{sbox{testbox}{FPupnresult{1.234 5 * 9.10 6.78 / +}}}
benchmark{sbox{testbox}{fpeval{1.234 * 5 + 6.78 / 9.10}}}
benchmark{sbox{testbox}{pgfmathparse{1.234 * 5 + 6.78 / 9.10}}}
end{document}


gives



0.00231 seconds (8.42e3 ops)
2.25e-4 seconds (837 ops)
2.63e-4 seconds (930 ops)




If you want simple dimension calculations, nothing is going to beat primitive, most obviously dimexpr. Something like



thedimexpr 1.2cm + 3.445cmrelax


is 'clocked' by the benchmark code at 2.64e-6 seconds (9.85 ops) on my system: really, really fast.







share|improve this answer














share|improve this answer



share|improve this answer








edited Dec 9 at 8:58

























answered Dec 6 at 17:56









Joseph Wright

201k21554879




201k21554879












  • The benchmark code does multiple cycles, and here is reporting the minimum time for one run.
    – Joseph Wright
    Dec 6 at 17:57










  • the l3benchmark output prints out result with 1:10^16 relative precision (natural, as this the FPU precision), but those last 4, 5, 6, or even 8 digits have no real meaning, the unit in the last place in your examples is something like 10^{-19} seconds, but my CPU is 2 GHz, 2 10^9, i think 10 GHz is a safe overestimate of current machines, this is 10^10 operations per second, so 10^{-19} times 10^{10} means 0.000000001 processor operation... I trust the reported timings could be rounded to at most 6 digits precision with no loss of any significance. (and pdfelapsedtime is very volatile).
    – jfbu
    Dec 6 at 19:49










  • @jfbu Sure, but this is still experimental ... Bruno wants to add rounding to sig figs for this purpose
    – Joseph Wright
    Dec 6 at 19:54






  • 1




    @jfbu Results updated: we altered the output a bit
    – Joseph Wright
    Dec 9 at 9:46






  • 1




    I think it would make sense to mention that pgf computes for instance sin(40) with 5 decimal places while xfp/l3fp computes it to 16 decimal places in barely twice longer.
    – Bruno Le Floch
    Dec 9 at 10:12


















  • The benchmark code does multiple cycles, and here is reporting the minimum time for one run.
    – Joseph Wright
    Dec 6 at 17:57










  • the l3benchmark output prints out result with 1:10^16 relative precision (natural, as this the FPU precision), but those last 4, 5, 6, or even 8 digits have no real meaning, the unit in the last place in your examples is something like 10^{-19} seconds, but my CPU is 2 GHz, 2 10^9, i think 10 GHz is a safe overestimate of current machines, this is 10^10 operations per second, so 10^{-19} times 10^{10} means 0.000000001 processor operation... I trust the reported timings could be rounded to at most 6 digits precision with no loss of any significance. (and pdfelapsedtime is very volatile).
    – jfbu
    Dec 6 at 19:49










  • @jfbu Sure, but this is still experimental ... Bruno wants to add rounding to sig figs for this purpose
    – Joseph Wright
    Dec 6 at 19:54






  • 1




    @jfbu Results updated: we altered the output a bit
    – Joseph Wright
    Dec 9 at 9:46






  • 1




    I think it would make sense to mention that pgf computes for instance sin(40) with 5 decimal places while xfp/l3fp computes it to 16 decimal places in barely twice longer.
    – Bruno Le Floch
    Dec 9 at 10:12
















The benchmark code does multiple cycles, and here is reporting the minimum time for one run.
– Joseph Wright
Dec 6 at 17:57




The benchmark code does multiple cycles, and here is reporting the minimum time for one run.
– Joseph Wright
Dec 6 at 17:57












the l3benchmark output prints out result with 1:10^16 relative precision (natural, as this the FPU precision), but those last 4, 5, 6, or even 8 digits have no real meaning, the unit in the last place in your examples is something like 10^{-19} seconds, but my CPU is 2 GHz, 2 10^9, i think 10 GHz is a safe overestimate of current machines, this is 10^10 operations per second, so 10^{-19} times 10^{10} means 0.000000001 processor operation... I trust the reported timings could be rounded to at most 6 digits precision with no loss of any significance. (and pdfelapsedtime is very volatile).
– jfbu
Dec 6 at 19:49




the l3benchmark output prints out result with 1:10^16 relative precision (natural, as this the FPU precision), but those last 4, 5, 6, or even 8 digits have no real meaning, the unit in the last place in your examples is something like 10^{-19} seconds, but my CPU is 2 GHz, 2 10^9, i think 10 GHz is a safe overestimate of current machines, this is 10^10 operations per second, so 10^{-19} times 10^{10} means 0.000000001 processor operation... I trust the reported timings could be rounded to at most 6 digits precision with no loss of any significance. (and pdfelapsedtime is very volatile).
– jfbu
Dec 6 at 19:49












@jfbu Sure, but this is still experimental ... Bruno wants to add rounding to sig figs for this purpose
– Joseph Wright
Dec 6 at 19:54




@jfbu Sure, but this is still experimental ... Bruno wants to add rounding to sig figs for this purpose
– Joseph Wright
Dec 6 at 19:54




1




1




@jfbu Results updated: we altered the output a bit
– Joseph Wright
Dec 9 at 9:46




@jfbu Results updated: we altered the output a bit
– Joseph Wright
Dec 9 at 9:46




1




1




I think it would make sense to mention that pgf computes for instance sin(40) with 5 decimal places while xfp/l3fp computes it to 16 decimal places in barely twice longer.
– Bruno Le Floch
Dec 9 at 10:12




I think it would make sense to mention that pgf computes for instance sin(40) with 5 decimal places while xfp/l3fp computes it to 16 decimal places in barely twice longer.
– Bruno Le Floch
Dec 9 at 10:12










up vote
1
down vote













tl;dr: pgfmath is 10 Times faster than fp



So I went and did a little test that I feel is sufficiently realistic for my intended purpose: typesetting dozens and hundred of pages with like a dozen floating point multiplications per page.



With around 80 pages of typesetting material my test document needed roughly a1 = 6 seconds to complete typesetting with XeLaTeX using pgfmath; with three times as much stuff, that rose to b1 = 13 seconds. In comparison, doing the same using fp, the timings came out as a2 = 37 and b2 = 107 seconds. Clearly, there must be a certain overhead here, which I estimated using the assumption that ( a1 - c ) * 3 = b1 - c and ( a2 - c ) * 3 = b2 - c, giving me a constant overhead of around c = 2.5 seconds. Timings with that overhead subtracted, then, are a1c = 3.5, b1c = 10.5, a2c = 34.5, b2c = 104.5.



While these measurements are all done in a rough way and with lots of hand-waving, they still seem to indicate that fp takes ten times as long as pgfmath.



There's still all kinds of wiggle room in here; I haven't checked whether the fp calculations couldn't have been optimized (which could narrow the gap), and I did not use the short-circuited versions of pgfmathparse and friends (which could widen the gap), but, on the other hand, a 1 / 10 ratio is unlikely to be easily done away with.



I haven't done tests with xfp / LaTeX3 FPU yet, as suggested by Joseph Wright; I did look up that stuff on CTAN and it does look very promising with regard to the future of LaTeX. It's really bleeding-edge and none of that is included in my TeX Live 2016 installation which I'm loath to update at this point in time. But certainly something to consider for the future.



My figures differ from those of Joseph Wright, so feel free to critique my methodology.






share|improve this answer





















  • The factor of 10 seems to agree with Joseph Wright's results.
    – Bruno Le Floch
    Dec 9 at 10:09















up vote
1
down vote













tl;dr: pgfmath is 10 Times faster than fp



So I went and did a little test that I feel is sufficiently realistic for my intended purpose: typesetting dozens and hundred of pages with like a dozen floating point multiplications per page.



With around 80 pages of typesetting material my test document needed roughly a1 = 6 seconds to complete typesetting with XeLaTeX using pgfmath; with three times as much stuff, that rose to b1 = 13 seconds. In comparison, doing the same using fp, the timings came out as a2 = 37 and b2 = 107 seconds. Clearly, there must be a certain overhead here, which I estimated using the assumption that ( a1 - c ) * 3 = b1 - c and ( a2 - c ) * 3 = b2 - c, giving me a constant overhead of around c = 2.5 seconds. Timings with that overhead subtracted, then, are a1c = 3.5, b1c = 10.5, a2c = 34.5, b2c = 104.5.



While these measurements are all done in a rough way and with lots of hand-waving, they still seem to indicate that fp takes ten times as long as pgfmath.



There's still all kinds of wiggle room in here; I haven't checked whether the fp calculations couldn't have been optimized (which could narrow the gap), and I did not use the short-circuited versions of pgfmathparse and friends (which could widen the gap), but, on the other hand, a 1 / 10 ratio is unlikely to be easily done away with.



I haven't done tests with xfp / LaTeX3 FPU yet, as suggested by Joseph Wright; I did look up that stuff on CTAN and it does look very promising with regard to the future of LaTeX. It's really bleeding-edge and none of that is included in my TeX Live 2016 installation which I'm loath to update at this point in time. But certainly something to consider for the future.



My figures differ from those of Joseph Wright, so feel free to critique my methodology.






share|improve this answer





















  • The factor of 10 seems to agree with Joseph Wright's results.
    – Bruno Le Floch
    Dec 9 at 10:09













up vote
1
down vote










up vote
1
down vote









tl;dr: pgfmath is 10 Times faster than fp



So I went and did a little test that I feel is sufficiently realistic for my intended purpose: typesetting dozens and hundred of pages with like a dozen floating point multiplications per page.



With around 80 pages of typesetting material my test document needed roughly a1 = 6 seconds to complete typesetting with XeLaTeX using pgfmath; with three times as much stuff, that rose to b1 = 13 seconds. In comparison, doing the same using fp, the timings came out as a2 = 37 and b2 = 107 seconds. Clearly, there must be a certain overhead here, which I estimated using the assumption that ( a1 - c ) * 3 = b1 - c and ( a2 - c ) * 3 = b2 - c, giving me a constant overhead of around c = 2.5 seconds. Timings with that overhead subtracted, then, are a1c = 3.5, b1c = 10.5, a2c = 34.5, b2c = 104.5.



While these measurements are all done in a rough way and with lots of hand-waving, they still seem to indicate that fp takes ten times as long as pgfmath.



There's still all kinds of wiggle room in here; I haven't checked whether the fp calculations couldn't have been optimized (which could narrow the gap), and I did not use the short-circuited versions of pgfmathparse and friends (which could widen the gap), but, on the other hand, a 1 / 10 ratio is unlikely to be easily done away with.



I haven't done tests with xfp / LaTeX3 FPU yet, as suggested by Joseph Wright; I did look up that stuff on CTAN and it does look very promising with regard to the future of LaTeX. It's really bleeding-edge and none of that is included in my TeX Live 2016 installation which I'm loath to update at this point in time. But certainly something to consider for the future.



My figures differ from those of Joseph Wright, so feel free to critique my methodology.






share|improve this answer












tl;dr: pgfmath is 10 Times faster than fp



So I went and did a little test that I feel is sufficiently realistic for my intended purpose: typesetting dozens and hundred of pages with like a dozen floating point multiplications per page.



With around 80 pages of typesetting material my test document needed roughly a1 = 6 seconds to complete typesetting with XeLaTeX using pgfmath; with three times as much stuff, that rose to b1 = 13 seconds. In comparison, doing the same using fp, the timings came out as a2 = 37 and b2 = 107 seconds. Clearly, there must be a certain overhead here, which I estimated using the assumption that ( a1 - c ) * 3 = b1 - c and ( a2 - c ) * 3 = b2 - c, giving me a constant overhead of around c = 2.5 seconds. Timings with that overhead subtracted, then, are a1c = 3.5, b1c = 10.5, a2c = 34.5, b2c = 104.5.



While these measurements are all done in a rough way and with lots of hand-waving, they still seem to indicate that fp takes ten times as long as pgfmath.



There's still all kinds of wiggle room in here; I haven't checked whether the fp calculations couldn't have been optimized (which could narrow the gap), and I did not use the short-circuited versions of pgfmathparse and friends (which could widen the gap), but, on the other hand, a 1 / 10 ratio is unlikely to be easily done away with.



I haven't done tests with xfp / LaTeX3 FPU yet, as suggested by Joseph Wright; I did look up that stuff on CTAN and it does look very promising with regard to the future of LaTeX. It's really bleeding-edge and none of that is included in my TeX Live 2016 installation which I'm loath to update at this point in time. But certainly something to consider for the future.



My figures differ from those of Joseph Wright, so feel free to critique my methodology.







share|improve this answer












share|improve this answer



share|improve this answer










answered Dec 7 at 13:02









John Frazer

1435




1435












  • The factor of 10 seems to agree with Joseph Wright's results.
    – Bruno Le Floch
    Dec 9 at 10:09


















  • The factor of 10 seems to agree with Joseph Wright's results.
    – Bruno Le Floch
    Dec 9 at 10:09
















The factor of 10 seems to agree with Joseph Wright's results.
– Bruno Le Floch
Dec 9 at 10:09




The factor of 10 seems to agree with Joseph Wright's results.
– Bruno Le Floch
Dec 9 at 10:09


















draft saved

draft discarded




















































Thanks for contributing an answer to TeX - LaTeX Stack Exchange!


  • Please be sure to answer the question. Provide details and share your research!

But avoid



  • Asking for help, clarification, or responding to other answers.

  • Making statements based on opinion; back them up with references or personal experience.


To learn more, see our tips on writing great answers.





Some of your past answers have not been well-received, and you're in danger of being blocked from answering.


Please pay close attention to the following guidance:


  • Please be sure to answer the question. Provide details and share your research!

But avoid



  • Asking for help, clarification, or responding to other answers.

  • Making statements based on opinion; back them up with references or personal experience.


To learn more, see our tips on writing great answers.




draft saved


draft discarded














StackExchange.ready(
function () {
StackExchange.openid.initPostLogin('.new-post-login', 'https%3a%2f%2ftex.stackexchange.com%2fquestions%2f463554%2fknown-benchmarks-for-floating-point-calculations%23new-answer', 'question_page');
}
);

Post as a guest















Required, but never shown





















































Required, but never shown














Required, but never shown












Required, but never shown







Required, but never shown

































Required, but never shown














Required, but never shown












Required, but never shown







Required, but never shown







Popular posts from this blog

Biblatex bibliography style without URLs when DOI exists (in Overleaf with Zotero bibliography)

ComboBox Display Member on multiple fields

Is it possible to collect Nectar points via Trainline?