Continued from B6: Distributed Run-Time Monitoring and Control of Data Analysis Workflows
Description
Data centers have a large and growing contribution to global energy consumption. Factors like manufacturing, construction, disassembly, green energy emissions are frequently neglected in current estimates. The total climate footprint can vary dramatically, based on energy sources, hardware, application, utilization, life-cycle management. B6 aims at a holistic management of end-to-end energy profiles and climate footprints of ML-based data analysis workflows (DAW), as well as system-internal configuration and tuning knobs.

Scientists
- Philipp Ortner
- Ilin Tolovski
Publications
2025
Bodner, Thomas; Boissier, Martin; Rabl, Tilmann; Salazar-Díaz, Ricardo; Schmeller, Florian; Strassenburg, Nils; Tolovski, Ilin; Weisgut, Marcel; Yue, Wang
A Case for Ecological Efficiency in Database Server Lifecycles Proceedings Article
In: Conference on Innovative Data Systems Research (CIDR), www.cidrdb.org, 2025.
@inproceedings{des2025ecological,
title = {A Case for Ecological Efficiency in Database Server Lifecycles},
author = { Thomas Bodner and Martin Boissier and Tilmann Rabl and Ricardo Salazar-Díaz and Florian Schmeller and Nils Strassenburg and Ilin Tolovski and Marcel Weisgut and Wang Yue},
url = {https://www.vldb.org/cidrdb/2025/a-case-for-ecological-efficiency-in-database-server-lifecycles.html},
year = {2025},
date = {2025-01-01},
urldate = {2025-01-01},
booktitle = {Conference on Innovative Data Systems Research (CIDR)},
publisher = {www.cidrdb.org},
abstract = {Like other software systems, database systems benefit from hardware performance improvements. For the longest time, acquiring new hardware resulted in significant software efficiency gains due to exponential improvements of hardware capabilities. Physical limits in hardware manufacturing have brought former niche designs into standard components, such as multiple cores and specialized circuits. Even with these new designs, hardware improvements are decreasing, while software and applications are still becoming increasingly complex and resource demanding. Given the resource consumption of hardware manufacturing, the ideal lifecycle of hardware naturally has to extend from an efficiency aspect. In this paper, we try to estimate efficiency of lifecycle duration of database hardware. We calculate the reduction in performance improvements of hardware using publicly available performance numbers, as well as our own benchmarks, and relate them to the specified thermal design power to get the power efficiency. Incorporating estimations on hardware and power production carbon intensity, we challenge current wisdom on hardware replacement frequencies and try to establish new rules of thumb on the ideal hardware lifecycles for database deployments. We present opportunities for future research trends.},
keywords = {},
pubstate = {published},
tppubtype = {inproceedings}
}
2024
Schmeller, Florian; Nugroho, Dwi P. A.; Zeuch, Steffen; Rabl, Tilmann
Towards A GPU-Accelerated Stream Processing Engine Through Query Compilation Proceedings Article
In: Lernen, Wissen, Daten, Analysen. (LWDA '24), 2024.
@inproceedings{SchmellerNZR2024,
title = {Towards A GPU-Accelerated Stream Processing Engine Through Query Compilation},
author = { Florian Schmeller and Dwi P. A. Nugroho and Steffen Zeuch and Tilmann Rabl},
url = {https://www.informatik.uni-wuerzburg.de/fileadmin/1003-lwda24/LWDA_Paper/DB_LWDA_CRC_152.pdf},
year = {2024},
date = {2024-09-01},
urldate = {2024-09-01},
booktitle = {Lernen, Wissen, Daten, Analysen. (LWDA '24)},
abstract = {Over the last decade, data stream processing has emerged to provide real-time insights into large, unbounded volumes of data. At the same time, graphics processing units (GPU) have become an important accelerator for improving the performance of compute-bound applications. Nevertheless, state-of-the-art data streaming systems opt to scale-out and typically do not make efficient use of the underlying hardware. Recent work has shown that query compilation is a viable technique to support hardware advancements in query processing engines. However, it often comes with high development and maintenance costs. In particular, when the process involves hardware accelerators such as GPUs. In this paper, we propose a framework for compiling data stream queries to efficient GPU code in a developer-friendly manner. We demonstrate the feasibility of our framework by integrating it into the data management system NebulaStream. Our experiments show that frequent memory transfers between CPU and GPU impact the query processing throughput.},
keywords = {},
pubstate = {published},
tppubtype = {inproceedings}
}
Salazar-Díaz, Ricardo; Glavic, Boris; Rabl, Tilmann
InferDB: In-Database Machine Learning Inference Using Indexes Journal Article
In: Proceedings of the VLDB Endowment, vol. 17, no. 8, pp. 1830-1842, 2024.
@article{salazar2024inferdb,
title = {InferDB: In-Database Machine Learning Inference Using Indexes},
author = { Ricardo Salazar-Díaz and Boris Glavic and Tilmann Rabl},
url = {https://dl.acm.org/doi/pdf/10.14778/3659437.3659441},
year = {2024},
date = {2024-01-01},
urldate = {2024-01-01},
journal = {Proceedings of the VLDB Endowment},
volume = {17},
number = {8},
pages = {1830-1842},
abstract = {The performance of inference with machine learning (ML) models and its integration with analytical query processing have become critical bottlenecks for data analysis in many organizations. An ML inference pipeline typically consists of a preprocessing workflow followed by prediction with an ML model. Current approaches for in-database inference implement preprocessing operators and ML algorithms in the database either natively, by transpiling code to SQL, or by executing user-defined functions in guest languages such as Python. In this work, we present a radically different approach that approximates an end-to-end inference pipeline (preprocessing plus prediction) using a light-weight embedding that discretizes a carefully selected subset of the input features and an index that maps data points in the embedding space to aggregated predictions of an ML model. We replace a complex preprocessing workflow and model-based inference with a simple feature transformation and an index lookup. Our framework improves inference latency by several orders of magnitude while maintaining similar prediction accuracy compared to the pipeline it approximates.},
keywords = {},
pubstate = {published},
tppubtype = {article}
}
Hendrik, Makait.; Monte, Bonaventura Del; Rabl, Tilmann
Ghostwriter: a Distributed Message Broker on RDMA and NVM Proceedings Article
In: 15th International Workshop on Accelerating Analytics and Data Management Systems Using Modern Processor and Storage Architectures, 2024.
@inproceedings{hendrik2024ghostwriter,
title = {Ghostwriter: a Distributed Message Broker on RDMA and NVM},
author = { Makait. Hendrik and Bonaventura Del Monte and Tilmann Rabl},
url = {https://vldb.org/workshops/2024/proceedings/ADMS/ADMS24_04.pdf},
year = {2024},
date = {2024-01-01},
urldate = {2024-01-01},
booktitle = {15th International Workshop on Accelerating Analytics and Data Management Systems Using Modern Processor and Storage Architectures},
volume = {15},
abstract = {Modern stream processing setups heavily rely on message bro- kers such as Apache Kafka or Apache Pulsar. These systems act as buffers and re-readable sources for downstream systems or applica- tions. They are typically deployed on separate servers, requiring extra resources, and achieve persistence through disk-based storage, limiting achievable throughput. In this paper, we present Ghost- writer, a message broker that utilizes remote direct memory access (RDMA) and non-volatile memory (NVM) for highly efficient mes- sage transfer and storage. Utilizing the hardware characteristics of RDMA and NVM, we achieve data throughput that is only limited by the underlying hardware, while reducing computation and dis- aggregating storage and data transfer coordination. Ghostwriter achieves performance improvements of up to an order of magnitude in throughput and latency over state-of-the-art solutions.},
keywords = {},
pubstate = {published},
tppubtype = {inproceedings}
}
Wang, Yue
Efficient Stream Processing in Decentralized Networks Proceedings Article
In: PhD@ VLDB. 2024, 2024.
@inproceedings{wang2024Phd,
title = {Efficient Stream Processing in Decentralized Networks},
author = { Yue Wang},
url = {https://vldb.org/2024/files/phd-workshop-papers/vldb_phd_workshop_paper_id_4.pdf},
year = {2024},
date = {2024-01-01},
urldate = {2024-01-01},
booktitle = {PhD@ VLDB. 2024},
abstract = {Internet-of-things (IoT) devices are widely used in industry as well as in research and are deployed in many applications. These massive amounts of devices are connected in large decentralized networks and produce unbounded data streams with continuous data. To process these data streams timely, current stream processing engines (SPEs) collect all data in a centralized data center. This approach leads to high network utilization and can create a bottleneck in the data center, as all data is transmitted via the network and results are computed centrally. State-of-the-art solutions push down partial window aggregations to machines that are near data streams. However, these solutions are limited to a single simple query. In this paper, we present our work on three solutions for different decentralized aggregations: Desis, Deco, and Dema, which significantly improve the performance of stream processing in decentralized networks. Our solutions reduce network traffic by up to 99.9%.},
keywords = {},
pubstate = {published},
tppubtype = {inproceedings}
}
Wang, Yue; Moczalla, Rafael; Luthra, Manisha; Rabl, Tilmann
Deco: Fast and Accurate Decentralized Aggregation of Count-Based Windows in Large-Scale IoT Applications Journal Article
In: 27th International Conference on Extending Database Technology (EDBT '24), 2024.
@article{wang2024accurate,
title = {Deco: Fast and Accurate Decentralized Aggregation of Count-Based Windows in Large-Scale IoT Applications},
author = { Yue Wang and Rafael Moczalla and Manisha Luthra and Tilmann Rabl},
editor = { Yue Wang},
url = {https://openproceedings.org/2024/conf/edbt/paper-104.pdf},
year = {2024},
date = {2024-01-01},
urldate = {2024-01-01},
journal = {27th International Conference on Extending Database Technology (EDBT '24)},
abstract = {In the realm of large-scale Internet-of-Things applications, aggregating data using count-based windows is a formidable challenge. Current methods, either centralized and slow or decentralized with potential inaccuracies, fail to strike a balance. This paper introduces Deco, a novel approach tailored for swift and precise aggregation in distributed stream processing systems. Accomplishing this balance is complex due to the dynamic nature of event distribution: events arrive at varying rates, unordered, and at diverse times, making accurate window computation a challenge. To overcome this, we propose a lightweight prediction method that derives local window sizes based on the previously observed event rates and performs corrections when necessary to ensure accurate and fast query results. These windows are processed in a decentralized manner on local nodes, verified for correctness, and then aggregated on a root node. Our evaluation showcases Deco’s superiority over centralized methods, outperforming others significantly. Deco reduces network traffic by up to 99% and exhibits linear scalability with node count.},
keywords = {},
pubstate = {published},
tppubtype = {article}
}
Tolovski, Ilin; Rabl, Tilmann
Addressing Data Management Challenges for Interoperable Data Science Proceedings Article
In: 1st International Workshop on Data-driven AI (DATAI) @ VLDB '24, 2024.
@inproceedings{tolovski2024addressing,
title = {Addressing Data Management Challenges for Interoperable Data Science},
author = { Ilin Tolovski and Tilmann Rabl},
url = {https://vldb.org/workshops/2024/proceedings/DATAI/DATAI-1.pdf},
year = {2024},
date = {2024-01-01},
urldate = {2024-01-01},
booktitle = {1st International Workshop on Data-driven AI (DATAI) @ VLDB '24},
abstract = {The development of data science pipelines (DSPs) has been steadily growing in popularity. While the increasing number of applications can also be attributed to novel algorithms and analytics libraries, the interoperability of new DSPs has been limited. To investigate this, we curated a corpus of over 494k GitHub Python repositories. We find that only 20% of the data science pipelines provide access to their input data and only 14% use a data backend. These findings highlight the key pain points in the development of interoperable DSPs. We identify five open data management challenges related to pipeline analysis, data access, and storage. We introduce Stork, a system for automated pipeline analysis, transformation, and data migration. Stork provides open data access while removing the hu- man in the loop when reproducing results and migrating projects to different storage and execution environments. We analyze terabytes of DSPs with Stork and successfully process 72% of the pipelines, transforming 75% of the accessible datasets.},
keywords = {},
pubstate = {published},
tppubtype = {inproceedings}
}
Riekenbrauck, Niklas; Weisgut, Marcel; Lindner, Daniel; Rabl, Tilmann
A Three-Tier Buffer Manager Integrating CXL Device Memory for Database Systems Proceedings Article
In: Joint International Workshop on Big Data Management on Emerging Hardware and Data Management on Virtualized Active Systems @ ICDE 2024, 2024.
@inproceedings{riekenbrauck2024bufferManager,
title = {A Three-Tier Buffer Manager Integrating CXL Device Memory for Database Systems},
author = { Niklas Riekenbrauck and Marcel Weisgut and Daniel Lindner and Tilmann Rabl},
url = {https://hardbd-active.github.io/2024/papers/HardBDActive24-RiekenbrauckWLR.pdf},
year = {2024},
date = {2024-01-01},
urldate = {2024-01-01},
booktitle = {Joint International Workshop on Big Data Management on Emerging Hardware and Data Management on Virtualized Active Systems @ ICDE 2024},
keywords = {},
pubstate = {published},
tppubtype = {inproceedings}
}
Wang, Yue; Boissier, Martin; Rabl, Tilmann
A Survey of Stream Processing System Benchmark Proceedings Article
In: 16th TPC Technology Conference on Performance Evaluation & Benchmarking (TPCTC) @ VLDB '24, 2024.
@inproceedings{wang2024TPCTC,
title = {A Survey of Stream Processing System Benchmark},
author = { Yue Wang and Martin Boissier and Tilmann Rabl},
url = {https://hpi.de/rabl/publications/Document/puma-rabl/A_Survey_of_Stream_Processing_System_Benchmark%20%281%29.pdf/79d7d679be711bb237f5f0a65f6f8a56.html?tx_extbibsonomycsl_publicationlist%5Baction%5D=download&cHash=491bd5085ed94a5ae04f7281ceadedc1},
year = {2024},
date = {2024-01-01},
urldate = {2024-01-01},
booktitle = {16th TPC Technology Conference on Performance Evaluation & Benchmarking (TPCTC) @ VLDB '24},
abstract = {Stream processing systems are a fundamental component of modern data processing, enabling timely and efficient handling of streaming data. To assess and compare the capabilities of stream processing systems, various benchmarks have been proposed over the past years. Examples span a wide range of use cases, ranging from benchmarks for enterprise computing to social network analyses and IoT networks. These benchmarks are designed with different focuses and exhibit different characteristics during execution. In this paper, we review existing stream processing benchmarks and analyze them across five dimensions: benchmark type, included workloads, data ingestion, supported systems under test (SUT), and tracked metrics. We compare their similarities and differences, providing a comprehensive overview of existing benchmarks. Finally, we discuss aspects that have been overlooked and highlight those that should be addressed when benchmarking future generations of streaming systems.},
keywords = {},
pubstate = {published},
tppubtype = {inproceedings}
}
Benson, Lawrence; Binnig, Carsten; Bodensohn, Jan-Micha; Lorenzi, Federico; Luo, Jigao; Porobic, Danica; Rabl, Tilmann; Sanghi, Anupam; Sears, Russell; Tözün, Pınar; Ziegler, Tobias
Surprise Benchmarking: The Why, What, and How Proceedings Article
In: pp. 1-8, 2024.
@inproceedings{surprise2024dbtest,
title = {Surprise Benchmarking: The Why, What, and How},
author = { Lawrence Benson and Carsten Binnig and Jan-Micha Bodensohn and Federico Lorenzi and Jigao Luo and Danica Porobic and Tilmann Rabl and Anupam Sanghi and Russell Sears and Pınar Tözün and Tobias Ziegler},
url = {https://dl.acm.org/doi/pdf/10.1145/3662165.3662763},
year = {2024},
date = {2024-01-01},
urldate = {2024-01-01},
journal = {Proceedings of the Tenth International Workshop on Testing Database Systems (DBTest)},
pages = {1-8},
keywords = {},
pubstate = {published},
tppubtype = {inproceedings}
}