Running Isolated CPU-Pinned, NUMA-Aligned Workloads and Unpinned Workloads on the Same Hypervisor: A Path Toward Intelligent Infrastructure

Authors

  • Binu Kiliamkavunkal Govindan

DOI:

https://doi.org/10.22399/ijcesen.4460

Keywords:

NUMA Architecture, Workload Classification, Artificial Intelligence Optimization, Container Orchestration, Performance Isolation

Abstract

Modern datacenter infrastructure faces significant challenges when hosting diverse workloads on shared computing resources while maintaining performance guarantees and cost efficiency. This article presents a comprehensive framework for running isolated, CPU-pinned, NUMA-aligned workloads alongside unpinned, flexible workloads on shared hypervisor infrastructure through artificial intelligence-driven optimization techniques. The framework addresses fundamental limitations in current virtualization platforms that fail to exploit NUMA topology information for intelligent workload placement, resulting in performance degradation and resource underutilization. Container-based microservices and telecommunications network functions experience substantial performance penalties from cross-NUMA memory access patterns and unpredictable resource allocation decisions. The proposed solution combines systematic workload classification with machine learning algorithms for predictive placement decisions and dynamic rebalancing capabilities. Implementation involves comprehensive hardware topology discovery, resource partitioning strategies, and integration with existing container orchestration platforms. Performance evaluation demonstrates substantial improvements across telecommunications network functions, edge AI systems, high-performance computing applications, multi-tenant cloud platforms, and future 6G network orchestration scenarios. The framework enables organizations to achieve deterministic performance guarantees for critical applications while maximizing infrastructure utilization through intelligent resource sharing, providing economic benefits through reduced hardware requirements and energy consumption.

References

[1] Andi Kleen, SUSE Labs, "An NUMA API for Linux," Novell Inc., 2004. [Online]. Available: https://halobates.de/numaapi3.pdf

[2] GeeksforGeeks, "Locality of Reference," 2025. [Online]. Available: https://www.geeksforgeeks.org/computer-organization-architecture/locality-of-reference-and-cache-operation-in-cache-memory/

[3] Brendan Burns, et al., "Kubernetes: Up and Running, 3rd Edition," O'Reilly Media, 2022. [Online]. Available: https://www.oreilly.com/library/view/kubernetes-up-and/9781098110192/

[4] Mohammad Dashti et al., "Traffic management: a holistic approach to memory placement on NUMA systems," ACM Digital Library, 2013. [Online]. Available: https://dl.acm.org/doi/10.1145/2451116.2451157

[5] Kubernetes Documentation, "Utilizing the NUMA-aware Memory Manager," 2024. [Online]. Available: https://kubernetes.io/docs/tasks/administer-cluster/memory-manager/

[6] Yiannis Georgiou, et al., "Topology-aware resource management for HPC applications," ACM Digital Library, 2017. [Online]. Available: https://dl.acm.org/doi/10.1145/3007748.3007768

[7] Xin Li et al., "Topology-Aware Scheduling Framework for Microservice Applications in Cloud," IEEE TRANSACTIONS ON PARALLEL AND DISTRIBUTED SYSTEMS, 2022 [Online]. Available: https://cis.temple.edu/~wu/research/publications/Publication_files/Topology-Aware_Scheduling_Framework_for_Microservice_Applications_in_Cloud.pdf

[8] Wen-Ping Lai, Kuan-Chun Chiu, "NUMAP: NUMA-aware Multi-core Pinning and Pairing for Network Slicing at the 5G Mobile Edge," IEEE Xplore, 2020. [Online]. Available: https://ieeexplore.ieee.org/document/9023272

[9] Gia Khanh Tran et al., "WS 16 - AI-Driven Digital Twin Orchestration for 6G Networks: Paving the Way for Future Connectivity," IEEE PIMRC 2025. [Online]. Available: https://pimrc2025.ieee-pimrc.org/workshop/ws-16-ai-driven-digital-twin-orchestration-6g-networks-paving-way-future-connectivity

Downloads

Published

2025-12-09

How to Cite

Binu Kiliamkavunkal Govindan. (2025). Running Isolated CPU-Pinned, NUMA-Aligned Workloads and Unpinned Workloads on the Same Hypervisor: A Path Toward Intelligent Infrastructure. International Journal of Computational and Experimental Science and Engineering, 11(4). https://doi.org/10.22399/ijcesen.4460

Issue

Section

Research Article