Enabling real-time, large-scale analytical workflows for data-intensive science requires the integration of state-of-the-art technologies from Big Data (NoSQL solutions and data-intensive programming, streaming engines), HPC (parallel programming, multi-/many-core architectures, GPUs, clusters), data analytics (analytical models and algorithms), and workflow management (definition and orchestration). As the principal investigator of the project, I established the system architecture and developed a large-scale and massively parallel storage and execution platform for data pipelines.
FFG-Austrian Research Promotion Agency, July 2014 – December 2016