Sie befinden Sich nicht im Netzwerk der Universität Paderborn. Der Zugriff auf elektronische Ressourcen ist gegebenenfalls nur via VPN oder Shibboleth (DFN-AAI) möglich. mehr Informationen...
Ergebnis 6 von 161
Proceedings of the International Conference on Parallel Architectures and Compilation Techniques, 2022, p.54-68
2022
Volltextzugriff (PDF)

Details

Autor(en) / Beteiligte
Titel
GNNear: Accelerating Full-Batch Training of Graph Neural Networks with near-Memory Processing
Ist Teil von
  • Proceedings of the International Conference on Parallel Architectures and Compilation Techniques, 2022, p.54-68
Ort / Verlag
New York, NY, USA: ACM
Erscheinungsjahr
2022
Quelle
ACM Digital Library
Beschreibungen/Notizen
  • Recently, Graph Neural Networks (GNNs) have become state-of-the-art algorithms for analyzing non-euclidean graph data. However, to realize efficient GNN training is challenging, especially on large graphs. The reasons are many-folded: 1) GNN training incurs a substantial memory footprint. Full-batch training on large graphs even requires hundreds to thousands of gigabytes of memory. 2) GNN training involves both memory-intensive and computation-intensive operations, challenging current CPU/GPU platforms. 3) The irregularity of graphs can result in severe resource under-utilization and load-imbalance problems. This paper presents a GNNear accelerator to tackle these challenges. GNNear adopts a DIMM-based memory system to provide sufficient memory capacity. To match the heterogeneous nature of GNN training, we offload the memory-intensive Reduce operations to in-DIMM Near-Memory-Engines (NMEs), making full use of the high aggregated local bandwidth. We adopt a Centralized-Acceleration-Engine (CAE) to process the computation-intensive Update operations. We further propose several optimization strategies to deal with the irregularity of input graphs and improve GNNear's performance. Comprehensive evaluations on 16 GNN training tasks demonstrate that GNNear achieves 30.8× / 2.5× geomean speedup and 79.6× / 7.3× (geomean) higher energy efficiency compared to Xeon E5-2698-v4 CPU and NVIDIA V100 GPU.
Sprache
Englisch
Identifikatoren
ISBN: 1450398685, 9781450398688
DOI: 10.1145/3559009.3569670
Titel-ID: cdi_acm_books_10_1145_3559009_3569670_brief

Weiterführende Literatur

Empfehlungen zum selben Thema automatisch vorgeschlagen von bX