GPU Glossary
GPU Glossary
/device-software/shared-memory

What is Shared Memory?

Shared memory is the level of the memory hierarchy corresponding to the thread block level of the thread group hierarchy in the CUDA programming model . It is generally expected to be much smaller but much faster (in throughput and latency) than the global memory .

A fairly typical kernel therefore looks something like this:

Shared memory is stored in the L1 data cache of the GPU's Streaming Multiprocessor (SM) .

Something seem wrong?
Or want to contribute?

Click this button to
let us know on GitHub.