TOP
Research
Research Teams for Science of Computing and Science by Computing
Past Research Teams
High Performance Artificial Intelligence Systems Research Team (Matsuoka)
High Performance Artificial Intelligence Systems Research Team (Matsuoka)
Japanese
Team Leader Satoshi Matsuoka
- 2019
- Team Leader, High Performance Artificial Intelligence Systems Research Team, R-CCS, RIKEN (-present)
- 2018
- Director, R-CCS, RIKEN (-present)
- Specially Appointed Professor, Tokyo Tech (-present)
- 2017
- Director, Real World Big Data Computing Open Innovation Laboratory (RWBC-OIL), AIST and Tokyo Tech
- 2000
- Full Professor, Global Scientific Information and Computing Center (GSIC), the Tokyo Institute of Technology
- 1993
- Ph. D. from the University of Tokyo
Keyword
- High Performance Artificial Intelligence Systems
- Scalable Deep Learning
- Performance Modeling of AI Systems e.g. Deep Learning
- Acceleration of Advanced Deep Learning Algorithms
- Convergence of AI and Simulation
Research summary
The High Performance Artificial Intelligence Systems Research Team is an R-CCS laboratory focusing on convergence of HPC and AI, namely high performance systems, software, and algorithms research for artificial intelligence/machine learning. In collaboration with other research institutes in HPC and AI-related research in Japan as well as globally, it seeks to develop next-generation AI technology that will utilize state-of-the-art high-performance computation facilities, including Fugaku. Specifically, we conduct research on next-generation AI systems by focusing on the following topics:
- Extreme speedup and scalability of deep learning: Achieve extreme scalability of deep learning in large-scale supercomputing environments including the post-K, extending the latest algorithms and frameworks for deep learning.
- Performance analysis of deep learning: Accelerate computational kernels for AI over the state-of-the-art hardware architectures by analyzing algorithms for deep learning and other machine learning/AI, measuring their performance and constructing their performance models.
- Acceleration of modern AI algorithms: Accelerate advanced AI algorithms, such as ultra-deep neural networks and high-resolution GAN over images, those that require massive computational resources, using extreme-scale deep learning systems.
- Acceleration of HPC algorithms using machine learning: Accelerate HPC algorithms and applications using empirical models based on machine learning.
Representative papers
- Jens Domke, Emil Vatai, Aleksandr Drozd, Peng Chen, Yosuke Oyama, Lingqi Zhang, Shweta Salaria, Daichi Mukunoki, Artur Podobas, Mohamed Wahib, Satoshi Matsuoka:
"Matrix Engines for HPC: A Performance Study from the Applications Perspective",
35th IEEE International Parallel & Distributed Processing Symposium (IPDPS 2021) - Albert Khaira, Truong Thao Nguyen, Leonardo Bautista Gomez, Ryousei Takano, Rosa Badia, Mohamed Wahib:
"An Oracle for Guiding Large-Scale Model/Hybrid Parallel Training of Convolutional Neural Networks",
30th ACM International Symposium on High-Performance Parallel and Distributed Computing (HPDC 2021) - Peng Chen, Mohamed Wahib, Xiao Wang, shinichiro takizawa, Takahiro Hirofuchi, Ogawa Hirotaka, Satoshi Matsuoka:
"Performance Portable Back-projection Algorithms on CPUs: Agnostic Data Locality and Vectorization Optimizations",
35th ACM International Conference on Supercomputing (ICS 2021) - Jun Li, Minjun Li, Zhigang Cai, Francois Trahay, Mohamed Wahib, Balazs Gerofi, Zhiming Liu, Jianwei Liao:
"Intra-page Cache Update in SLC Mode with Partial Programming in High Density SSDs",
50th International Conference on Parallel Processing (ICPP 2021) - Mohamed Wahib, Haoyu Zhang, Truong Thao Nguyen, Aleksandr Drozd, Jens Domke, Lingqi Zhang, Ryousei Takano, Satoshi Matsuoka:
"Scaling Deep Learning Workloads Beyond Memory Capacity",
International Conference for High Performance Computing, Networking, Storage, and Analysis (SC 2020) - Lingqi Zhang, Wahib Mohamed, Haoyu Zhang, Matsuoka Satoshi:
"A Study of Single and Multi-device Synchronization Methods in Nvidia GPUs",
34th IEEE International Parallel & Distributed Processing Symposium (IPDPS 2020) - Jens Domke, Satoshi Matsuoka, Ivan R. Ivanov, Yuki Tsushima, Tomoya Yuki, Akihiro Nomura, Shinichi Miura, Nic McDonald, Dennis L. Floyd, Nicolas Dube:
"The First Supercomputer with HyperX Topology: A Viable Alternative to Fat-Trees?",
International Conference for High Performance Computing, Networking, Storage, and Analysis (SC 2019) - Chen Peng,Wahib Mohamed,Takizawa Shinichiro,Matsuoka Satoshi:
"A Versatile Software Systolic Execution Model for GPU Memory Bound Kernels",
International Conference for High Performance Computing, Networking, Storage, and Analysis (SC 2019) - Jens Domke, Kazuaki Matsumura, Mohamed Wahib, Haoyu Zhang, Keita Yashima,Toshiki Tsuchikawa, Yohei Tsuji, Artur Podobas, Satoshi Matsuoka:
"Double-precision FPUs in High-Performance Computing: an Embarrassment of Riches?",
33th IEEE International Parallel & Distributed Processing Symposium (IPDPS 2019) - Shweta Salaria, Aleksandr Drozd, Artur Podobas, Satoshi Matsuoka:
"Learning Neural Representations for Predicting GPU Performance",
ISC High Performance 2019 (ISC 2019)