Abstract | ||
---|---|---|
Communication patterns extracted from parallel programs can provide a valuable source of information for parallel pattern detection, application auto-tuning, and runtime workload scheduling on heterogeneous systems. Once identified, such patterns can help find the most promising optimizations. Communication patterns can be detected using different methods, including sandbox simulation, memory profiling, and hardware counter analysis. However, these analyses usually suffer from high runtime and memory overhead, necessitating a trade off between accuracy and resource consumption. More importantly, none of the existing methods exploit fine-grained communication patterns on the level of individual code regions. In this paper, we present an efficient tool based on Disco PoP profiler that characterizes the communication pattern of every hotspot in a shared-memory application. With the aid of static and dynamic code analysis, it produces a nested structure of communication patterns based on program's loops. By employing asymmetric signature memory, the runtime overhead is around 225× while the required amount of memory remains fixed. In comparison with other profilers, the proposed method is efficient enough to be used with real world applications. |
Year | DOI | Venue |
---|---|---|
2015 | 10.1109/ICPP.2015.85 | ICPP |
Keywords | Field | DocType |
nested communication pattern, program analysis, thread dependency, profiler, parallel application | Interleaved memory,Uniform memory access,Shared memory,Computer science,Parallel computing,Distributed memory,Memory ordering,Memory management,Memory map,Distributed shared memory,Distributed computing | Conference |
ISSN | Citations | PageRank |
0190-3918 | 1 | 0.35 |
References | Authors | |
24 | 4 |
Name | Order | Citations | PageRank |
---|---|---|---|
Arya Mazaheri | 1 | 1 | 1.02 |
Ali Jannesari | 2 | 149 | 20.11 |
Abdolreza Mirzaei | 3 | 192 | 14.38 |
Felix Wolf | 4 | 57 | 12.00 |