logoalt Hacker News

KeplerBoytoday at 11:08 AM0 repliesview on HN

You can in CUDA. You can have shared memory which is basically L1 cache you have full control over. It's called shared memory because all threads within a block (which reside on a common SM) have fast access to it. The downside: you now have less regular L1 cache.