Fine-Grained Scaling in Stream Processing Systems: Hybrid CPU-Memory Autoscaling with Graph Neural Networks
DOI:
https://doi.org/10.22399/ijcesen.3918Keywords:
Stream Processing, Hybrid Autoscaling, Graph Neural Networks, Resource Optimization,, Cloud-Native ComputingAbstract
Stream-processing engines in distributed environments require dynamic resource allocation, though traditional autoscaling treats CPU and memory as coupled units. Recent technological developments showcase disaggregated resource management that allows operator-specific allocation based on runtime computational demands. The Justin framework for Apache Flink demonstrates hybrid CPU/memory scaling through precise resource distribution matching individual operator requirements during execution. StreamTune represents another advancement, employing Graph Neural Networks trained on execution histories to optimize operator parallelism and identify bottlenecks dynamically. These innovations achieve resource consumption reductions while preserving throughput, crucial for cloud-native deployments requiring cost optimization. Such developments prove essential when building large-scale platforms handling millions of events per second with minimal latency and maximum resource efficiency. Graph Neural Network integration enables predictive scaling through pattern recognition from execution data, shifting from reactive to proactive resource management paradigms. Machine learning convergence with distributed stream processing opens pathways for intelligent infrastructure adapting to workload variations while maintaining performance guarantees. Production environments reveal practical importance as operational expenses maintain direct relationships with resource utilization performance. These innovations create the groundwork for autonomous systems to achieve self-optimization via real-time performance indicators and predictive computational models.
References
[1] Yunfan Qing and Wenli Zheng, "Towards Fine-Grained Scalability for Stateful Stream Processing Systems," ResearchGate, Mar. 2025.https://www.researchgate.net/publication/389895054_Towards_Fine-Grained_Scalability_for_Stateful_Stream_Processing_Systems
[2] Ziliang Wang et al., "DeepScaling: microservices autoscaling for stable CPU utilization in large-scale cloud systems," in SoCC '22: ACM Symposium on Cloud Computing, ResearchGate, Nov. 2022.https://www.researchgate.net/publication/365223689_DeepScaling_microservices_autoscaling_for_stable_CPU_utilization_in_large_scale_cloud_systems
[3] Theodoros Theodoropoulos et al., "Graph neural networks for representing multivariate resource usage: A multiplayer mobile gaming case-study," International Journal of Information Management Data Insights, ScienceDirect, Feb. 2023.https://www.sciencedirect.com/science/article/pii/S2667096823000058
[4] Federico Lombardi, "PASCAL: An architecture for proactive auto-scaling of distributed services," Future Generation Computer Systems, ScienceDirect, Apr. 2019.https://www.sciencedirect.com/science/article/abs/pii/S0167739X18303728
[5] Tong Li et al., "Towards Fine-Grained Explainability for Heterogeneous Graph Neural Network," Proceedings of the AAAI Conference on Artificial Intelligence, Association for the Advancement of Artificial Intelligence, Jun. 2023. https://ojs.aaai.org/index.php/AAAI/article/view/26040
[6] Eugene Armah and Linda Amoako Bannning, "Towards a Proactive Autoscaling Framework for Data Stream Processing at the Edge using GRU and Transfer Learning," arXiv, Jul. 2025.https://arxiv.org/abs/2507.14597v1
[7] Christina Giannoula et al., "Accelerating Graph Neural Networks on Real Processing-In-Memory Systems," arXiv, Feb. 2024.https://arxiv.org/html/2402.16731v1
[8] Gabriele Russo Russo et al., "Hierarchical Auto-scaling Policies for Data Stream Processing on Heterogeneous Resources," ACM Digital Library, Oct. 2023.https://dl.acm.org/doi/10.1145/3597435
[9] Valeria Cardellini et al., "Runtime Adaptation of Data Stream Processing Systems: The State of the Art," ACM Digital Library, Sep. 2022.https://dl.acm.org/doi/abs/10.1145/3514496
Downloads
Published
How to Cite
Issue
Section
License
Copyright (c) 2025 International Journal of Computational and Experimental Science and Engineering

This work is licensed under a Creative Commons Attribution 4.0 International License.