DE version is available. Content is displayed in original English for accuracy.
Advertisement
Advertisement
⚡ Community Insights
Discussion Sentiment
60% Positive
Analyzed from 604 words in the discussion.
Trending Topics
#memory#wulf#compute#wall#more#bandwidth#paper#believe#mckee#elements

Discussion (42 Comments)Read Original on HackerNews
https://dl.acm.org/doi/10.1145/216585.216588
on her obituary page, you will see a prominent "Memory Wall" link that is NOT a reference to her paper, but a place for sharing your thoughts about her life
I notice these things a bit more as she was my PhD thesis advisor
It'd be interesting to see someone use the unabbreviated form; I have a hunch they wouldn't know to say "et alia".
> Rob Pike didn't really name my favorite editor after me.
Despite this, massively increased memory bandwidth does not translate to material performance improvements on non-parallel compute tasks because few tasks are actually memory bandwidth bound, instead being memory latency bound.
The best known general solutions for improving memory latency are per-compute element memory caches. Unfortunately, this increases the complexity and size of your compute elements forcing you to reduce the number of compute elements, but a large number of compute elements is the only way to saturate HBM memory bandwidth.
To keep up the best known techniques are either algorithmically batch which allows you to go wide using vector/batch instructions or you go the GPU route with memory latency-hiding parallelism.
Call your loved ones :(