Abstract
Arithmetic performance with GPGPU attracts attention. However, the difficulty of the programming poses a problem. We have proposed GPGPU programming which used the existing parallel programming technique. We are now developing OpenMP framework for GPU as a concrete of our proposal. The framework is based on Omni OpenMP Compiler and named “OMPCUDA”. In this paper we describe a design and an implementation of OMPCUDA. We evaluated using test programs, and validated that parallel improvement in the speed could be easily carried out in the same code as the existing OpenMP.
Access provided by Autonomous University of Puebla. Download to read the full chapter text
Chapter PDF
Similar content being viewed by others
Keywords
These keywords were added by machine and not by the authors. This process is experimental and the keywords may be updated as the learning algorithm improves.
References
NVIDIA: CUDA Zone, http://www.nvidia.com/object/cuda_home.html
Ohshima, S., Hirasawa, S., Honda, H.: Proposal of GPGPU Programming Using Existing Parallelizing Method. IPSJ Tech. Report(ARC-175), 7–10 (2007)
Sato, M., Satoh, S., Kusano, K., Tanaka, Y.: Design of OpenMP Compiler for an SMP Cluster. In: EWOMP ’99, pp. 32–39 (1999)
Aslot, V., Domeika, M., Eigenmann, R., Gaertner, G., Jones, W.B., Parady, B.: SPEComp: A New Benchmark Suite for Measuring Parallel Computer Performance. In: Eigenmann, R., Voss, M.J. (eds.) WOMPAT 2001. LNCS, vol. 2104, pp. 1–10. Springer, Heidelberg (2001)
Horn, D.: Stream Reduction Operations for GPGPU Applications. In: GPU Gems2. Addison-Wesley, Reading (2005)
Roger, D., Assarsson, U., Holzschuch, N.: Efficient Stream Reduction on the GPU. In: Workshop on General Purpose Processing on Graphics Processing Units (2007)
Owens, J., Davis, U.: Data-parallel algorithms and data structures. In: SUPERCOMPUTING 2007 Tutorial: Hight Performance Computing with CUDA (2007)
Lee, S., Min, S.J., Eigenmann, R.: Openmp to gpgpu: a compiler framework for automatic translation and optimization. In: PPoPP ’09: Proceedings of the 14th ACM SIGPLAN symposium on Principles and practice of parallel programming, pp. 101–110. ACM, New York (2009)
The Portland Group: PGI Accelerator Compilers, http://www.pgroup.com/resources/accel.htm
Author information
Authors and Affiliations
Editor information
Editors and Affiliations
Rights and permissions
Copyright information
© 2010 Springer-Verlag Berlin Heidelberg
About this paper
Cite this paper
Ohshima, S., Hirasawa, S., Honda, H. (2010). OMPCUDA : OpenMP Execution Framework for CUDA Based on Omni OpenMP Compiler. In: Sato, M., Hanawa, T., Müller, M.S., Chapman, B.M., de Supinski, B.R. (eds) Beyond Loop Level Parallelism in OpenMP: Accelerators, Tasking and More. IWOMP 2010. Lecture Notes in Computer Science, vol 6132. Springer, Berlin, Heidelberg. https://doi.org/10.1007/978-3-642-13217-9_13
Download citation
DOI: https://doi.org/10.1007/978-3-642-13217-9_13
Publisher Name: Springer, Berlin, Heidelberg
Print ISBN: 978-3-642-13216-2
Online ISBN: 978-3-642-13217-9
eBook Packages: Computer ScienceComputer Science (R0)