logoalt Hacker News

alain94040last Sunday at 5:15 PM1 replyview on HN

This is interesting if quite incomplete (as noted in the end conclusion). CPU re-order buffers turn what you think as mostly sequential execution into a massively parallel engine. Data memory access, perfecting, speculative execution, etc. But if you are running a micro-bencmark with a tight loop of millions of iterations, then understanding the pipeline dependencies and dispatching can provide good insights.


Replies

drob518last Sunday at 10:49 PM

Yep. Cache is always the wildcard.