Filters








30,247 Hits in 4.5 sec

Global cache management for multi-class workloads in data warehouses [chapter]

Shudong Jin, Xiaowei Sun
1998 Lecture Notes in Computer Science  
On the basis of this model, algorithms are proposed to generate efficient cache allocation scheme for multiple queries. These algorithms can be combined with different replacement strategies.  ...  A benefit metric is developed to compare the expected effect of caching retrieved sets and buffering for query processing.  ...  addition to the queries and allocate buffer pages to the one with greatest effect.  ... 
doi:10.1007/bfb0054220 fatcat:3ui3uchx2nfl7i3lanzhk572pe

ULCC

Xiaoning Ding, Kaibo Wang, Xiaodong Zhang
2011 SIGPLAN notices  
We have implemented ULCC at the user level based on a page-coloring technique for last level cache usage management.  ...  Second, at the user level, programmers are not able to allocate cache space at will to running threads in the shared cache, thus data sets with strong locality may not be allocated with sufficient cache  ...  Acknowledgments The authors thank the anonymous reviewers for their constructive comments and suggestions. They thank Bill Bynum for reading the paper and for his suggestions.  ... 
doi:10.1145/2038037.1941568 fatcat:ymnggpxrbvfqlgg6cjbuxca6qu

ULCC

Xiaoning Ding, Kaibo Wang, Xiaodong Zhang
2011 Proceedings of the 16th ACM symposium on Principles and practice of parallel programming - PPoPP '11  
We have implemented ULCC at the user level based on a page-coloring technique for last level cache usage management.  ...  Second, at the user level, programmers are not able to allocate cache space at will to running threads in the shared cache, thus data sets with strong locality may not be allocated with sufficient cache  ...  Acknowledgments The authors thank the anonymous reviewers for their constructive comments and suggestions. They thank Bill Bynum for reading the paper and for his suggestions.  ... 
doi:10.1145/1941553.1941568 dblp:conf/ppopp/DingWZ11 fatcat:gy5rtxpilbf2vplswivww2b23e

PiBooster: Performance Accelerations in Page Table Management for Paravirtual VMs

Zhi Zhang, Yueqiang Cheng
2016 2016 IEEE 9th International Conference on Cloud Computing (CLOUD)  
First, PiBooster shortens the execution paths of the page table (de)allocations by the PiBooster cache, which maintains dedicated buffers for serving page table (de)allocations.  ...  The micro experiment results indicate that PiBooster is able to completely eliminate the additional IOTLB flushes in the workload-stable environments, and effectively reduces (de)allocation time of the  ...  PiBooster Cache on the Bare-metal OS We believe that the design of the PiBooster cache could benefit the page table allocations and deallocations on the bare-metal OSes that work directly on hardware.  ... 
doi:10.1109/cloud.2016.0074 dblp:conf/IEEEcloud/ZhangC16 fatcat:zlwni6qh45cnzj3k4465dpdlwa

NightWatch: Integrating Lightweight and Transparent Cache Pollution Control into Dynamic Memory Allocation Systems

Rentong Guo, Xiaofei Liao, Hai Jin, Jianhui Yue, Guang Tan
2015 USENIX Annual Technical Conference  
NightWatch is based on the observation that data within the same memory chunk or chunks within the same allocation context often share similar locality property.  ...  Experiments with SPEC CPU2006 show that NightWatch improves application performance by up to 45% (18% on average), with an average monitoring overhead of 0.57% (up to 3.02%).  ...  effect of reduced cache resource.  ... 
dblp:conf/usenix/GuoLJYT15 fatcat:42ota3njpjg2lescrfwgialgdi

A locality-improving dynamic memory allocator

Yi Feng, Emery D. Berger
2005 Proceedings of the 2005 workshop on Memory system performance - MSP '05  
Previous generalpurpose allocators have focused on reducing fragmentation, while most locality-improving allocators have either focused on improving the locality of the allocator (not the application)  ...  Our allocator, called Vam, improves page-level locality by managing the heap in page-sized chunks and aggressively giving up free pages to the virtual memory manager.  ...  Fragmentation We evaluate the effect of allocator design on memory fragmentation.  ... 
doi:10.1145/1111583.1111594 dblp:conf/ACMmsp/FengB05 fatcat:nrypgqqfhbbpxn5bjayenalig4

Adaptive Cache Allocation with Prefetching Policy over End-to-End Data Processing

Hang Qin, Li Zhu
2017 Journal of Signal and Information Processing  
Adaptively, the stack cache can be turned off from beginning to end, when a stack overflow occurs to avoid the effect of stack switching on processor performance.  ...  Based on the analysis of data processing behavior, an adaptive cache organization scheme is proposed with fast address calculation.  ...  Currently, there are many studies on the implementation of the cache write failure strategy, including write allocate and non-write allocation, in terms of the strategies to improve the strategy [7]  ... 
doi:10.4236/jsip.2017.83010 fatcat:6btdmgkjj5b7jgwdo34k7jlpwi

Avoiding initialization misses to the heap

Jarrod A. Lewis, Bryan Black, Mikko H. Lipasti
2002 SIGARCH Computer Architecture News  
By tracking the initialization of dynamic memory allocations, it is possible to identify store instructions that miss the cache and would fetch uninitialized heap data.  ...  The memory system property crucial for achieving good performance under heap allocation is cache installation -the ability to allocate and initialize a new object into the cache without a penalty.  ...  We would also like to thank the anonymous reviewers for their many helpful comments.  ... 
doi:10.1145/545214.545236 fatcat:hx46oh242vfvpebdoetjnjcnea

SRM-buffer

Xiaoning Ding, Kaibo Wang, Xiaodong Zhang
2011 Proceedings of the sixth conference on Computer systems - EuroSys '11  
However, the side effects and performance implications of accessing the data in buffer caches (i.e. buffer cache hits) have not been paid attention.  ...  Second, due to the shared structure of LLCs on multicore processors, an application accessing the data in a buffer cache may flush the to-be-reused data of its co-running applications from the shared LLC  ...  As we will show in section 4, the trade-off has a minimal effect to the capability of SRM-buffer to reduce cache pollution, but it can effectively avoid the negative effects on virtual page allocation  ... 
doi:10.1145/1966445.1966468 dblp:conf/eurosys/DingWZ11 fatcat:a2mvmostzndrtcf43idjcxdcsq

Reducing the cost of persistence for nonvolatile heaps in end user devices

Sudarsun Kannan, Ada Gavrilovska, Karsten Schwan
2014 2014 IEEE 20th International Symposium on High Performance Computer Architecture (HPCA)  
These not only affect the applications causing them, but also other applications relying on the same cache and/or memory hierarchy.  ...  Novel solutions to overcome such overheads include (1) a page contiguity algorithm that reduces interference-related cache misses, (2) a cache efficient NVM write aware memory allocator that reduces cache  ...  Acknowledgments This research is supported in part by the Intel URO program on software for persistent memories and by NSF award CCF-1161969.  ... 
doi:10.1109/hpca.2014.6835960 dblp:conf/hpca/KannanGS14 fatcat:45daodexbjbhzhhy5zvdoks5cm

vCache

Daehoon Kim, Hwanju Kim, Nam Sung Kim, Jaehyuk Huh
2015 Proceedings of the 48th International Symposium on Microarchitecture - MICRO-48  
This in turn allows that the guest OS transparently view its vLLC and preserve the effectiveness of its page placement policy.  ...  However, such an illusion of resource dedication has not been supported for the last-level cache (LLC), although the LLC is the largest on-chip shared resource with a significant performance impact.  ...  Acknowledgment This research was supported in part by a National Research Foundation of Korea (NRF) grant funded by the Korea government (MSIP) (No.  ... 
doi:10.1145/2830772.2830825 dblp:conf/micro/KimKKH15 fatcat:wckq7rx23nhirai5dftnmddirm

Towards practical page coloring-based multicore cache management

Xiao Zhang, Sandhya Dwarkadas, Kai Shen
2009 Proceedings of the fourth ACM european conference on Computer systems - EuroSys '09  
However, page coloring places additional constraints on memory space allocation, which may conflict with application memory needs.  ...  Modern multi-core processors present new resource management challenges due to the subtle interactions of simultaneously executing processes sharing on-chip resources (particularly the L2 cache).  ...  Acknowledgments We thank the anonymous EuroSys referees and our shepherd Timothy Roscoe for their helpful comments on a preliminary version of this paper.  ... 
doi:10.1145/1519065.1519076 dblp:conf/eurosys/ZhangDS09 fatcat:xcpq7ije5fbz5k4lwgaxph5zyi

Evaluation of Exclusive Data Allocation Between SSD Tier and SSD Cache in Storage Systems
english

Shinichi Hayashi, Norihisa Komoda
2014 Proceedings of the 16th International Conference on Enterprise Information Systems  
With the proposed method, the SSD cache function with exclusive data allocation caches only data allocated on the HDD tier.  ...  It also reduces the storage I/O response time by up to 23% compared to a combination of SSD/HDD volume tiering and SSD cache methods with no exclusive data allocation.  ...  This does not lead to long response time since high I/O frequency data remain allocated on the SSD cache even if the page allocated on the SSD tier is migrated to the HDD tier.  ... 
doi:10.5220/0004867801440151 dblp:conf/iceis/HayashiK14 fatcat:gsfv4vozu5egnm3lezymf3qttu

A Coordinated Approach for Practical OS-Level Cache Management in Multi-core Real-Time Systems

Hyoseung Kim, Arvind Kandhalu, Ragunathan Rajkumar
2013 2013 25th Euromicro Conference on Real-Time Systems  
However, the state-of-the-art software cache partitioning techniques face two challenges: (1) the memory co-partitioning problem, which results in page swapping or waste of memory, and (2) the availability  ...  Our scheme provides predictable cache performance, addresses the aforementioned problems of existing software cache partitioning, and efficiently allocates cache partitions to schedule a given taskset.  ...  We now evaluate the effectiveness of our cache-aware task allocation (CATA) algorithm that exploits cache reservation and cache sharing.  ... 
doi:10.1109/ecrts.2013.19 dblp:conf/ecrts/KimKR13 fatcat:bntx6eu3gjejnleg3hc5ptgd5i

PTEMagnet: fine-grained physical memory reservation for faster page walks in public clouds

Artemiy Margaritov, Dmitrii Ustiugov, Amna Shahab, Boris Grot
2021 Proceedings of the 26th ACM International Conference on Architectural Support for Programming Languages and Operating Systems  
The fragmentation results in the effective cache footprint of the host PT being larger than that of the guest PT.  ...  The bloated footprint of the host PT leads to frequent cache misses during nested page walks, increasing page walk latency.  ...  valuable feedback on this work.  ... 
doi:10.1145/3445814.3446704 fatcat:d2g3b3rcijdzloxovecij3kegu
« Previous Showing results 1 — 15 out of 30,247 results