A copy of this work was available on the public web and has been preserved in the Wayback Machine. The capture dates from 2017; you can also visit the original URL.
The file type is application/pdf
.
Priority-based cache allocation in throughput processors
2015
2015 IEEE 21st International Symposium on High Performance Computer Architecture (HPCA)
GPUs employ massive multithreading and fast context switching to provide high throughput and hide memory latency. Multithreading can increase contention for various system resources, however, that may result in suboptimal utilization of shared resources. Previous research has proposed variants of throttling thread-level parallelism to reduce cache contention and improve performance. Throttling approaches can, however, lead to under-utilizing thread contexts, on-chip interconnect, and offchip
doi:10.1109/hpca.2015.7056024
dblp:conf/hpca/LiRJOEBFR15
fatcat:gxlctriiffaq5prutqcxr6pyry