Title
A Machine-Learning-Based Framework for Productive Locality Exploitation
Abstract
Data locality is of extreme importance in programming distributed-memory architectures due to its implications on latency and energy consumption. Automated compiler and runtime system optimization studies have attempted to improve data locality exploitation without burdening the programmer. However, due to the difficulty of static code analysis, conservatism in compiler optimizations to avoid errors, and cost of dynamic analysis, the efficacy of automated optimizations is limited. Therefore, programmers need to spend significant effort in optimizing locality while creating applications for distributed memory parallel systems. We present a machine-learning based framework to automatically exploit locality in distributed memory applications. This framework takes application source whose time-critical blocks are marked by pragmas, and produces optimized source code that uses a regressor for efficient data movement. The regressor is trained with automatically-collected application profiles with very small input data sizes. We integrate our prototype in the Chapel language stack. In our experiments, we show that the Elastic Net model is the ideal regressor for our case and applications that utilize Elastic Net can perform very similarly to programmer-optimized versions. We also show that such regressors can be trained within few minutes on a cluster or within 30 minutes on a workstation, including data collection.
Year
DOI
Venue
2021
10.1109/TPDS.2021.3051348
IEEE Transactions on Parallel and Distributed Systems
Keywords
DocType
Volume
Data locality,distributed memory,programming models,machine learning
Journal
32
Issue
ISSN
Citations 
6
1045-9219
0
PageRank 
References 
Authors
0.34
0
3
Name
Order
Citations
PageRank
Engin Kayraklioglu163.86
Erwan Favry200.34
Tarek El-Ghazawi342744.88