Subtopic Deep Dive

Heterogeneous Computing Frameworks
Research Guide

What is Heterogeneous Computing Frameworks?

Heterogeneous Computing Frameworks provide programming models and runtime systems for integrating CPUs with accelerators like GPUs in parallel computing environments.

These frameworks include CUDA for GPU programming (Nickolls et al., 2008, 1544 citations), StarPU for task scheduling on heterogeneous multicore architectures (Augonnet et al., 2010, 1237 citations), and Kokkos for performance portability across exascale systems (Trott et al., 2021, 427 citations). They address task offloading, data movement, and memory consistency in diverse hardware. Over 50 papers in the provided list highlight their evolution from early GPU models to modular supercomputers like JURECA (Krause and Thörnig, 2018, 348 citations).

15
Curated Papers
3
Key Challenges

Why It Matters

Heterogeneous frameworks enable energy-efficient HPC by exploiting GPUs for biomolecular simulations, as in GROMACS optimizations (Kutzner et al., 2015, 252 citations). They support supercomputers like JUWELS for Earth system modeling (Krause, 2019, 276 citations). StarPU achieves theoretical performance peaks on Cell/BE and GPUs (Augonnet et al., 2010). Kokkos ensures portability for exascale scientific codes (Trott et al., 2021). Gunrock accelerates irregular graph analytics on GPUs (Wang et al., 2016, 317 citations).

Key Research Challenges

Task Scheduling Heterogeneity

Runtime systems must dynamically schedule tasks across CPUs and GPUs with varying capabilities. StarPU addresses this via a unified platform but faces load balancing issues on data-parallel accelerators (Augonnet et al., 2010). Theoretical peaks remain hard to reach in mixed workloads.

Memory Consistency Models

Cache coherence and consistency break in heterogeneous setups with separate CPU/GPU memories. Sorin et al. (2011, 328 citations) detail models, but programming errors persist in CUDA and OpenCL codes. Verification remains manual and error-prone.

Performance Portability

Codes must run efficiently across GPU generations and CPU architectures without rewrites. Kokkos extends models for exascale but struggles with vendor-specific optimizations (Trott et al., 2021). Polyhedral compilers like PPCG help but limit to static-control loops (Verdoolaege et al., 2013).

Essential Papers

1.

Scalable Parallel Programming with CUDA

John Nickolls, Ian Buck, Michael Garland et al. · 2008 · Queue · 1.5K citations

The advent of multicore CPUs and manycore GPUs means that mainstream processor chips are now parallel systems. Furthermore, their parallelism continues to scale with Moore’s law. The challenge is t...

2.

StarPU: a unified platform for task scheduling on heterogeneous multicore architectures

Cédric Augonnet, Samuel Thibault, Raymond Namyst et al. · 2010 · Concurrency and Computation Practice and Experience · 1.2K citations

Abstract In the field of HPC, the current hardware trend is to design multiprocessor architectures featuring heterogeneous technologies such as specialized coprocessors (e.g. Cell/BE) or data‐paral...

3.

Kokkos 3: Programming Model Extensions for the Exascale Era

Christian Robert Trott, Damien Lebrun-Grandié, Daniel Arndt et al. · 2021 · IEEE Transactions on Parallel and Distributed Systems · 427 citations

As the push towards exascale hardware has increased the diversity of system architectures, performance portability has become a critical aspect for scientific software. We describe the Kokkos Perfo...

4.

Polyhedral parallel code generation for CUDA

Sven Verdoolaege, Juan Carlos Juega, Albert Cohen et al. · 2013 · ACM Transactions on Architecture and Code Optimization · 361 citations

This article addresses the compilation of a sequential program for parallel execution on a modern GPU. To this end, we present a novel source-to-source compiler called PPCG. PPCG singles out for it...

5.

JURECA: Modular supercomputer at Jülich Supercomputing Centre

Dorian Krause, Philipp Thörnig · 2018 · Journal of large-scale research facilities JLSRF · 348 citations

JURECA is a petaflop-scale modular supercomputer operated by Jülich Supercomputing Centre at Forschungszentrum Jülich. The system combines a flexible Cluster module, based on T-Platforms V-Class bl...

6.

A Primer on Memory Consistency and Cache Coherence

Daniel J. Sorin, Mark D. Hill, David A. Wood · 2011 · Synthesis lectures on computer architecture · 328 citations

7.

Gunrock

Yangzihao Wang, Andrew Davidson, Yuechao Pan et al. · 2016 · 317 citations

For large-scale graph analytics on the GPU, the irregularity of data access\nand control flow, and the complexity of programming GPUs have been two\nsignificant challenges for developing a programm...

Reading Guide

Foundational Papers

Start with CUDA (Nickolls et al., 2008, 1544 citations) for GPU programming basics; StarPU (Augonnet et al., 2010, 1237 citations) for runtime scheduling; Sorin et al. (2011, 328 citations) for memory models essential to all frameworks.

Recent Advances

Kokkos 3 (Trott et al., 2021, 427 citations) for exascale portability; JUWELS supercomputer (Krause, 2019, 276 citations) for modular deployments; Gunrock (Wang et al., 2016, 317 citations) for graph analytics advances.

Core Methods

CUDA kernels and memory hierarchies (Nickolls et al., 2008); asynchronous task submission in StarPU (Augonnet et al., 2010); polyhedral source-to-source compilation (Verdoolaege et al., 2013); Kokkos execution spaces and policies (Trott et al., 2021).

How PapersFlow Helps You Research Heterogeneous Computing Frameworks

Discover & Search

Research Agent uses searchPapers to find 'heterogeneous computing frameworks GPU CPU' yielding StarPU (Augonnet et al., 2010); citationGraph traces 1237 citations to JURECA (Krause and Thörnig, 2018); findSimilarPapers links to Kokkos (Trott et al., 2021); exaSearch uncovers Gunrock graph frameworks (Wang et al., 2016).

Analyze & Verify

Analysis Agent applies readPaperContent to extract StarPU scheduling algorithms; verifyResponse with CoVe cross-checks claims against Kokkos portability metrics; runPythonAnalysis replots GPU speedup data from GROMACS (Kutzner et al., 2015) using NumPy for statistical verification; GRADE scores evidence on memory models from Sorin et al. (2011).

Synthesize & Write

Synthesis Agent detects gaps in task offloading between StarPU and Kokkos, flags contradictions in GPU trends (Brodtkorb et al., 2012); Writing Agent uses latexEditText for framework comparisons, latexSyncCitations for 10+ papers, latexCompile for reports, exportMermaid for runtime scheduling diagrams.

Use Cases

"Compare GPU speedups in GROMACS vs Gunrock using code benchmarks"

Research Agent → searchPapers(GROMACS GPU) → Analysis Agent → runPythonAnalysis(NumPy plot speedups from Kutzner et al. 2015 and Wang et al. 2016) → matplotlib speedup chart with statistical tests.

"Draft LaTeX section on StarPU vs Kokkos for exascale portability"

Synthesis Agent → gap detection(StarPU Augonnet 2010 vs Kokkos Trott 2021) → Writing Agent → latexEditText(draft) → latexSyncCitations(15 refs) → latexCompile(PDF with tables).

"Find GitHub repos for CUDA polyhedral code generation"

Research Agent → searchPapers(PPCG Verdoolaege 2013) → Code Discovery → paperExtractUrls → paperFindGithubRepo → githubRepoInspect(PPCG compiler examples and benchmarks).

Automated Workflows

Deep Research scans 50+ papers on heterogeneous frameworks, chaining searchPapers → citationGraph → structured report on CUDA to Kokkos evolution. DeepScan applies 7-step analysis with CoVe checkpoints to verify JURECA GPU integrations (Krause and Thörnig, 2018). Theorizer generates hypotheses on unified models from StarPU and Gunrock patterns.

Frequently Asked Questions

What defines Heterogeneous Computing Frameworks?

Programming models like CUDA (Nickolls et al., 2008) and runtime systems like StarPU (Augonnet et al., 2010) for CPU-GPU task integration and data movement.

What are key methods in this subtopic?

Task scheduling (StarPU), performance portability (Kokkos, Trott et al., 2021), polyhedral compilation to CUDA (PPCG, Verdoolaege et al., 2013), and GPU graph primitives (Gunrock, Wang et al., 2016).

What are foundational papers?

Scalable Parallel Programming with CUDA (Nickolls et al., 2008, 1544 citations), StarPU (Augonnet et al., 2010, 1237 citations), Polyhedral code generation (Verdoolaege et al., 2013, 361 citations).

What open problems exist?

Achieving full performance portability across exascale vendors (Trott et al., 2021), automating memory consistency in heterogeneous codes (Sorin et al., 2011), scaling irregular workloads like graphs (Wang et al., 2016).

Research Parallel Computing and Optimization Techniques with AI

PapersFlow provides specialized AI tools for Computer Science researchers. Here are the most relevant for this topic:

See how researchers in Computer Science & AI use PapersFlow

Field-specific workflows, example queries, and use cases.

Computer Science & AI Guide

Start Researching Heterogeneous Computing Frameworks with AI

Search 474M+ papers, run AI-powered literature reviews, and write with integrated citations — all in one workspace.

See how PapersFlow works for Computer Science researchers