Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
110 tokens/sec
GPT-4o
56 tokens/sec
Gemini 2.5 Pro Pro
44 tokens/sec
o3 Pro
6 tokens/sec
GPT-4.1 Pro
47 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

HeAT -- a Distributed and GPU-accelerated Tensor Framework for Data Analytics (2007.13552v2)

Published 27 Jul 2020 in cs.DC, cs.LG, and cs.MS

Abstract: To cope with the rapid growth in available data, the efficiency of data analysis and machine learning libraries has recently received increased attention. Although great advancements have been made in traditional array-based computations, most are limited by the resources available on a single computation node. Consequently, novel approaches must be made to exploit distributed resources, e.g. distributed memory architectures. To this end, we introduce HeAT, an array-based numerical programming framework for large-scale parallel processing with an easy-to-use NumPy-like API. HeAT utilizes PyTorch as a node-local eager execution engine and distributes the workload on arbitrarily large high-performance computing systems via MPI. It provides both low-level array computations, as well as assorted higher-level algorithms. With HeAT, it is possible for a NumPy user to take full advantage of their available resources, significantly lowering the barrier to distributed data analysis. When compared to similar frameworks, HeAT achieves speedups of up to two orders of magnitude.

User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (12)
  1. Markus Götz (18 papers)
  2. Daniel Coquelin (8 papers)
  3. Charlotte Debus (13 papers)
  4. Kai Krajsek (5 papers)
  5. Claudia Comito (8 papers)
  6. Philipp Knechtges (8 papers)
  7. Björn Hagemeier (1 paper)
  8. Michael Tarnawa (1 paper)
  9. Simon Hanselmann (1 paper)
  10. Martin Siggel (3 papers)
  11. Achim Basermann (12 papers)
  12. Achim Streit (26 papers)
Citations (8)

Summary

We haven't generated a summary for this paper yet.