The full text of this article hosted at iucr.org is unavailable due to technical difficulties.

Concurrency and Computation: Practice and Experience
Special Issue Paper

Cooperative multitasking for GPU‐accelerated grid systems

Fumihiko Ino

Corresponding Author

E-mail address:ino@ist.osaka-u.ac.jp

Graduate School of Information Science and Technology, Osaka University, 1‐5 Yamadaoka, Suita, Osaka 565‐0871, Japan

Graduate School of Information Science and Technology, Osaka University, 1‐5 Yamadaoka, Suita, Osaka 565‐0871, Japan
Search for more papers by this author
Akihiro Ogita

NEC SOFT, Ltd., 1‐18‐7 Shinkiba, Koto‐ku, Tokyo 136‐0082, Japan

Search for more papers by this author
Kentaro Oita

Graduate School of Information Science and Technology, Osaka University, 1‐5 Yamadaoka, Suita, Osaka 565‐0871, Japan

Search for more papers by this author
Kenichi Hagihara

Graduate School of Information Science and Technology, Osaka University, 1‐5 Yamadaoka, Suita, Osaka 565‐0871, Japan

Search for more papers by this author
First published: 22 March 2011
Cited by: 6

SUMMARY

This paper presents a cooperative multitasking method for concurrent execution of scientific and graphics applications on the graphics processing unit (GPU). Our method is designed to accelerate compute unified device architecture‐based applications using idle GPU cycles in the office. To prevent significant slow‐down of graphics applications, the method divides scientific tasks into smaller pieces, which are then sequentially executed at the appropriate intervals. The method also has flexibility in finding the best tradeoff point between scientific applications and graphics applications. Experimental results show that the proposed method is useful to control the frame rate of the graphics application and the throughput of the scientific application. For example, biological sequence alignment can be processed at approximately 30% of the dedicated throughput while achieving interactive rendering at 58 frames per second. We also show that matrix multiplication can be efficiently processed at 60% of the dedicated throughput during word processing and web browsing. Copyright © 2011 John Wiley & Sons, Ltd.

Number of times cited: 6

  • , ASHA: An adaptive shared-memory sharing architecture for multi-programmed GPUs, Microprocessors and Microsystems, 46, (264), (2016).
  • , Chimera, ACM SIGARCH Computer Architecture News, 43, 1, (593), (2015).
  • the Twentieth International Conference ASPLOS '15 Istanbul, Turkey Proceedings of the Twentieth International Conference on Architectural Support for Programming Languages and Operating Systems - ASPLOS '15 Architectural Support for Programming Languages and Operating Systems ACM Press New York, New York, USA , (2015). 9781450328357 , 10.1145/2694344 20150303090911 http://dl.acm.org/citation.cfm?doid=2694344 Chimera Collaborative Preemption for Multitasking on a Shared GPU , (2015). 593 606 , 10.1145/2694344.2694346 20150303090911 http://dl.acm.org/citation.cfm?doid=2694344.2694346
  • , Chimera, ACM SIGPLAN Notices, 50, 4, (593), (2015).
  • 2013 First International Symposium on Computing and Networking (CANDAR) Matsuyama, Japan 2013 First International Symposium on Computing and Networking IEEE , (2013). 978-1-4799-2796-8 978-1-4799-2795-1 The Past, Present, and Future of GPU-Accelerated Grid Computing , (2013). 17 21 6726872 , 10.1109/CANDAR.2013.10 http://ieeexplore.ieee.org/document/6726872/
  • , The implementation and the performance analysis of the multi-channel software-based lock-in amplifier for the stiffness mapping with atomic force microscope (AFM), Bulletin of the Polish Academy of Sciences: Technical Sciences, 60, 1, (2012).