DistME: A Fast and Elastic Distributed Matrix Computation Engine using GPUs developed by DGIST
작성자. Public Relations Team
DistME: A Fast and Elastic Distributed Matrix Computation Engine using GPUs developed by DGIST - Can analyze 100 times more data 14 times faster than existing technology
- Expected to be used in big data machine learning and various industry fields in the future
△Professor Min-Soo Kim in the Department of Information and Communication Engineering (upper left), Ph.D. student Yoon-Min Nam (upper right), and Ph.D. student Donghyoung Han (lower right)
A DGIST research team developed a new technology that can process and analyze data differently from the existing technology. Its application is highly looked forward to as it can overcome the limitations of the existing technologies and can process much more data faster.
DGIST announced on July 4 that Professor Min-Soo Kim’s team in the Department of Information and Communication Engineering developed the DistME (Distributed Matrix Engine) technology that can analyze 100 times more data 14 times faster than the existing technologies. This new technology is expected to be used in machine learning that needs big data processing or various industry fields to analyze large-scale data in the future.
‘Matrix’ data, which expresses numbers in row and column, is the most widely used form of data in various fields such as machine learning1 and science technology. While ‘SystemML’ and ‘ScaLAPACK’ are evaluated as the most popular technologies to analyze matrix data, but the processing capability of existing technology has recently reached its limits with the growing size of data. It is especially difficult to conduct multiplications, which are required for data processing, for big data analysis with the existing methods because they cannot perform elastic analysis and processing and require a huge amount of network data transfer for processing.
In response, Professor Kim’s team developed a distributed matrix multiplication method that is different from the existing one. Also called CuboidMM, this method forms matrix multiplication in a 3D hexahedron and then partitions and processes to multiple pieces called cuboid. The optimal size of the cuboid is flexibly determined depending on the characteristics of the matrices, i.e., the size, the dimension, and sparsity of matrix, so as to minimize the communication cost. CuboidMM not only includes all the existing methods but also can perform matrix multiplication with minimum communication cost. In addition, Professor Kim’s team devised an information processing technology by combining with GPU (Graphics Processing Unit) which dramatically enhanced the performance of matrix multiplication.
The DistME technology developed by Professor Kim’s team has increased processing speed by combining CuboidMM with GPU, which is 6.5 and 14 times faster than ScaLAPACK and SystemML respectively and can analyze 100 times larger matrix data than SystemML. It is expected to open new applicability of machine learning in various areas that need large-scale data processing including online shopping malls and SNS.
Professor Kim in the Department of Information and Communication Engineering said ‘Machine Learning Technology, which has been drawing worldwide attention, has limitations in the speed for matrix-form big data analysis and the size of analysis processing. The information processing technology developed this time can overcome such limitations and will be useful in not only machine learning but also applications in wider ranges of science technology data analysis application.”
This research was participated by Donghyoung Han, a Ph.D. student in the Department of Information and Communication Engineering as the first author and was presented on July 3 in ACM SIGMOD 2019, the top-renowned academic conference in the database field held in Amsterdam, Netherlands.
1 Machine Learning: A computer program that improves information processing ability through learning using data and processing experience and a related research field.
For more information, contact:
Min-Soo Kim, Associate Professor
Department of Information and Communication Engineering
Daegu Gyeongbuk Institute of Science and Technology (DGIST)
Donghyoung Han, Yoon-Min Nam, Jihye Lee, Kyongseok Park, Hyunwoo Kim, Min-Soo Kim, "DistME : A Fast and Elastic Distributed Matrix Computation Engine using GPUs", ACM SIGMOD, Published on July 3rd, 2019.