495 Hits in 4.6 sec

PACS-CS: a large-scale bandwidth-aware PC cluster for scientific computation

T. Boku, M. Sato, A. Ukawa, D. Takahashi, S. Sumimoto, K. Kumon, T. Moriyama, M. Shimizu
2006 Sixth IEEE International Symposium on Cluster Computing and the Grid (CCGRID'06)  
Under the preliminary performance evaluation, we confirmed that the computation part in practical Lattice QCD code will be able to achieve 30% of peak performance, and up to 600 Mbyte/sec of bandwidth  ...  We have been developing a large scale PC cluster named PACS-CS (Parallel Array Computer System for Computational Sciences) at Center for Computational Sciences, University of Tsukuba, for wide variety  ...  CP-PACS has been used for about 10 years for the computations of particle physics and material physics as well as a part of heterogeneous computing system for computational astrophysics [6] .  ... 
doi:10.1109/ccgrid.2006.78 dblp:conf/ccgrid/BokuSUTSKMS06 fatcat:a3ohs5l4bfegzggupmdoks5bfy

Retrospective Study of Performance and Power Consumption of Computer Systems

Hisanobu Tomari, Kei Hiraki
2011 IPSJ Online Transactions  
Data for only two or three generations of systems are insufficient for projecting the performance/power of future systems.  ...  We measured the performance and power consumption of 70 computer systems from 1989 to 2011.  ...  Acknowledgments Benchmarks were in part carried out on the NEC SX-9 at Center for Computational Astrophysics, CfCA, of the National Astronomical Observatory of Japan.  ... 
doi:10.2197/ipsjtrans.4.217 fatcat:tbjwrpsypneppd5fyculbamhs4

Experience and performance of persistent memory for the DUNE data acquisition system [article]

Adam Abed Abud, Giovanna Lehmann, Roland Sipos
2020 arXiv   pre-print
In this paper, we characterize the performance of persistent memory devices, which use the 3DXPoint technology, in the context of the data acquisition system for one large Particle Physics experiment,  ...  An example of these technologies is 3DXPoint, which promises to fill the gap between memory and traditional storage and offers unprecedented high throughput for data persistency.  ...  We thank our colleagues from Intel® for the precious feedback and for the fruitful collaboration.  ... 
arXiv:2011.01341v1 fatcat:6vqixiwwpnbp5pnagohx2lf56a

Experience and performance of persistent memory for the DUNE data acquisition system

Adam Abed Abud, Giovanna Lehmann Miotto, Roland Sipos
2021 IEEE Transactions on Nuclear Science  
In this paper, we characterize the performance of persistent memory devices, which use the 3DXPoint technology, in the context of the data acquisition system for one large Particle Physics experiment,  ...  An example of these technologies is 3DXPoint, which promises to fill the gap between memory and traditional storage and offers unprecedented high throughput for non-volatile data.  ...  INTEL® OPTANE™ A. 3D XPoint™ memory technology Intel® Optane™ devices are based on the 3D XPoint™ memory technology.  ... 
doi:10.1109/tns.2021.3084848 fatcat:6yf3yla4tndsvkm6cc6d7a36sy

DASH: a Recipe for a Flash-based Data Intensive Supercomputer

Jiahua He, Arun Jagatheesan, Sandeep Gupta, Jeffrey Bennett, Allan Snavely
2010 2010 ACM/IEEE International Conference for High Performance Computing, Networking, Storage and Analysis  
DASH uses commodity parts including Intel® X25-E flash drives and distributed shared memory (DSM) software from ScaleMP®.  ...  We present here an overview of the design of DASH, an analysis of its cost efficiency, then a detailed recipe for how we designed and tuned it for high data-performance, lastly show that running data-intensive  ...  We would like to credit our science collaborators Peter Nugent and Janet Jacobsen of the Berkeley Lab for the description about PTF presented in section IVB, and also thank them for sharing with us the  ... 
doi:10.1109/sc.2010.16 dblp:conf/sc/HeJGBS10 fatcat:e4kb3h6utfg2vnf6xyuu3fsf7i

Shall numerical astrophysics step into the era of Exascale computing? [article]

Giuliano Taffoni, Giuseppe Murante, Luca Tornatore, David Goz, Stefano Borgani, Manolis Katevenis, Nikolaos Chrysos, Manolis Marazakis
2019 arXiv   pre-print
We also discuss the impact of this new computing resources on the numerical codes in Astronomy and Astrophysics.  ...  We discuss the major technological challenges in the design, development and use of such computing capabilities and we will report on the progresses that has been made in the last years in Europe, in particular  ...  computing performance (Sustained performance) of super computers With current semiconductor technologies, all proposed Exascale designs would consume hundreds of megawatts of power.  ... 
arXiv:1904.11720v1 fatcat:nj2h2fphwneuleoxcxgfuox52i

Evaluating the benefits of an extended memory hierarchy for parallel streamline algorithms

David Camp, Hank Childs, Amit Chourasia, Christoph Garth, Kenneth I. Joy
2011 2011 IEEE Symposium on Large Data Analysis and Visualization  
Currently popular designs create a staging area on each compute node for data output via solid state drives (SSDs), local hard drives, or both.  ...  In this paper, we investigate whether these extensions to the memory hierarchy, primarily intended for computer simulations that produce data, can also benefit visualization and analysis programs that  ...  problems in computational nuclear astrophysics.  ... 
doi:10.1109/ldav.2011.6092318 dblp:conf/ldav/CampCCGJ11 fatcat:ocqbrxvxhvgl3afum7dasqnake

HPC Systems in the Next Decade – What to Expect, When, Where

Dirk Pleiter, C. Doglioni, D. Kim, G.A. Stewart, L. Silvestris, P. Jackson, W. Kamleh
2020 EPJ Web of Conferences  
diverse compute technologies as well as deeper memory and storage hierarchies.  ...  The analysis of upcoming architectural options and emerging technologies allow for setting expectations for application developers, which will have to cope with heterogeneous architectures, increasingly  ...  Configuration Memory technology ∆τ NVIDIA V100 GPU HBM2 20-40 ms Dual-socket Intel Skylake server DDR4 0.4 s Intel Optane DC Memory-attached 3D XPoint 35-140 s Intel DC P4511 2 TByte PCIe-attached  ... 
doi:10.1051/epjconf/202024511004 fatcat:qk4yzblkevb2pkwf3qxderxd2a

A Fast MHD Code for Gravitationally Stratified Media using Graphical Processing Units: SMAUG

M. K. Griffiths, V. Fedun, R. Erdélyi
2015 Journal of astrophysics and astronomy  
For example, parallel magnetohydrodynamic (MHD) algorithms are important for numerical modelling of highly inhomogeneous solar, astrophysical and geophysical plasmas.  ...  The opportunity has been recognized by the computational sciences and engineering communities, who have recently harnessed successfully the numerical performance of GPUs.  ...  They also acknowledge Anthony Brookfield and Corporate Information and Computing Services at The University of Sheffield, for the provision of the High Performance Computing Service.  ... 
doi:10.1007/s12036-015-9328-y fatcat:jf5wb7n4tfbdtplmn7mvrpprki

A case study of OpenCL on an Android mobile GPU

James A. Ross, David A. Richie, Song J. Park, Dale R. Shires, Lori L. Pollock
2014 2014 IEEE High Performance Extreme Computing Conference (HPEC)  
Given today's exploding popularity of mobile devices, we investigate the possibilities for high performance mobile computing.  ...  The achieved computational performance of the lowpower mobile Adreno GPU is compared with a quad-core ARM, an x86 Intel processor, and a discrete AMD GPU.  ...  The N-body algorithm provides an excellent benchmark for the evaluation of a computing platform for several reasons.  ... 
doi:10.1109/hpec.2014.7040987 dblp:conf/hpec/RossRPSP14 fatcat:jzcymlr7zvec5k5iqh2o33wmxm

Optimizing the hybrid parallelization of BHAC [article]

Salvatore Cielo, Oliver Porth, Luigi Iapichino, Anupam Karmakar, Hector Olivares, Chun Xia
2021 arXiv   pre-print
The performance of optimized version of BHAC improved by ∼28%, making it viable for scaling on several hundreds of supercomputer nodes.  ...  In doing so, we showcase the use of performance profiling tools usable on x86 (Intel-based) architectures.  ...  Josef Weidendorfer, leader of LRZ Future Computing Group, for the support on the test runs with the A64FX nodes described in Section 6.  ... 
arXiv:2108.12240v1 fatcat:klcw7extx5d3nmucjau6umws6y

ATIP/A*CRC Workshop on Accelerator Technologies for High-Performance Computing: Does Asia Lead the Way? Programme Handbook [article]

Michalewicz Marek, Kahaner David
2012 Zenodo  
Information Program (ATIP) and the Agency for Science, Technology and Research (A*STAR) Computational Resource Centre (A*CRC), this workshop in Singapore focus on the practical aspects of scientific and  ...  With support from the United States National Science Foundation (NSF) and in cooperation with the HPC Special Interest Group (HPC SIG) of the Association for Computing Machinery (ACM), the Asian Technology  ...  I will present OIST's existing data infrastructure and show new performance results for various file system options on this and newer hardware.  ... 
doi:10.5281/zenodo.3977457 fatcat:v2aftckqinasxk5ugdw27vinpa

A nationwide parallel computing environment

Ken Kennedy, Charles F. Bender, John W. D. Connolly, John L. Hennessy, Mary K. Vernon, Larry Smarr
1997 Communications of the ACM  
Alliance (NCSA) hopes to accelerate the creation of a nationwide parallel computing environment for the US by developing hardware and software that will ease desktop access to the National Technology Grid  ...  NCSA's plans focus on experimental computer architectures, high performance user programs, machine independent analyzers for parallel computers and advanced application support that leverage parallel adaptive  ...  Parallel Software Development The Parallel Computing team will identify user codes that can reach higher performance by exploiting the features of DSM, using these codes to drive R&D of new compiler technology  ... 
doi:10.1145/265684.265693 fatcat:xtfge7keynabjbyxfmdcbm3kci

Evaluating GPU Passthrough in Xen for High Performance Cloud Computing

Andrew J. Younge, John Paul Walters, Stephen Crago, Geoffrey C. Fox
2014 2014 IEEE International Parallel & Distributed Processing Symposium Workshops  
This methodology is evaluated by measuring the performance of two Nvidia Tesla GPUs within Xen VMs and comparing to bare-metal hardware.  ...  With the advent of virtualization and Infrastructure-as-a-Service (IaaS), the broader scientific computing community is considering the use of clouds for their technical computing needs.  ...  Younge also acknowledges support from The Persistent Systems Fellowship of the School of Informatics and Computing at Indiana University.  ... 
doi:10.1109/ipdpsw.2014.97 dblp:conf/ipps/YoungeWCF14 fatcat:nwpmtctgbnai7nyjdyh54xqr6m

Survey of Storage Systems for High-Performance Computing

2018 Supercomputing Frontiers and Innovations  
of exascale computing.  ...  However, the overall application performance is often held back by I/O bottlenecks due to insufficient performance of file systems or I/O libraries for highly parallel workloads.  ...  This material reflects only the authors' view and the EU commission is not responsible for any use that may be made of the information it contains.  ... 
doi:10.14529/jsfi180103 fatcat:hi3qctpl7rfvjgl53pxmqwqviy
« Previous Showing results 1 — 15 out of 495 results