Change search
Link to record
Permanent link

Direct link
Publications (10 of 190) Show all publications
Khah, Y. P., Shirvani, M. H. & Taheri, J. (2026). A survey study on meta-heuristic-based feature selection approaches of intrusion detection systems in distributed networks. Computer Standards & Interfaces, 96, Article ID 104074.
Open this publication in new window or tab >>A survey study on meta-heuristic-based feature selection approaches of intrusion detection systems in distributed networks
2026 (English)In: Computer Standards & Interfaces, ISSN 0920-5489, E-ISSN 1872-7018, Vol. 96, article id 104074Article in journal (Refereed) Published
Abstract [en]

With the emergence of IoT and expanding the coverage of distributed networks such as cloud and fog, security attacks and breaches are becoming distributed and expanded too. Cybersecurity attacks can disrupt business continuity or expose critical data, leading to significant failures. The Intrusion Detection Systems (IDSs) as a remedy in such networks play a critical role in this ecosystem to find an attack at the earliest time and the countermeasure is performed if necessary. Artificial intelligence techniques such as machine learning-based and meta-heuristic-based approaches are being pervasively applied to prepare smarter IDS components from logged network traffic. The network traffic is recorded in the form of data sets for further analysis to detect traffic behavior from past treatments. Feature selection is a prominent approach in creating the prediction model to recognize feature network connection is normal or not. Since the feature selection problem in large datasets is NP-Hard and utilizing only heuristic-based approaches is not as efficient as desired, meta-heuristic-based approaches attract research attention to prepare highly accurate prediction models. To address the issue, this paper presents a subjective classification of published literature. Then, this presents a survey study on meta-heuristic-based feature selection approaches in preparing efficient IDSs. It investigates several kinds of literature from different angles and compares them in terms of used metrics in the literature to give broad insights into readers for advantages, challenges, and limitations. It can pave the way by highlighting research gaps for further processing and improvement in the future by interested researchers in the field.

Place, publisher, year, edition, pages
Elsevier, 2026
Keywords
Intrusion detection system (IDS), Fog computing, Feature selection, Metaheuristic algorithms, Network security
National Category
Computer Sciences Computer Systems
Research subject
Computer Science
Identifiers
urn:nbn:se:kau:diva-107348 (URN)10.1016/j.csi.2025.104074 (DOI)001589093600001 ()2-s2.0-105017588599 (Scopus ID)
Available from: 2025-10-21 Created: 2025-10-21 Last updated: 2026-02-12Bibliographically approved
Garshasbi Herabad, M., Taheri, J., Ahmed, B. S. & Curescu, C. (2026). An Overview of Technical Aspects and Challenges in Designing Edge-Cloud Systems. Applied Sciences, 16(3), Article ID 1454.
Open this publication in new window or tab >>An Overview of Technical Aspects and Challenges in Designing Edge-Cloud Systems
2026 (English)In: Applied Sciences, E-ISSN 2076-3417, Vol. 16, no 3, article id 1454Article, review/survey (Refereed) Published
Abstract [en]

Edge-cloud computing has emerged as a key enabling paradigm for augmented and virtual reality (AR/VR) systems because of the stringent computational and ultra-low-latency requirements of AR/VR workloads. Designing efficient edge-cloud systems for such workloads involves multiple technical aspects, including communication technologies, service placement, task offloading and caching, service migration, and security and privacy. This paper provides a structured and technical analysis of these aspects from an AR/VR perspective. We adopt a two-stage literature analysis, in which Google Scholar is used to identify fundamental technical aspects and solution approaches, followed by a focused analysis of recent research trends and future directions using academic databases (e.g., IEEE Xplore, ACM Digital Library, and ScienceDirect). We present an organized classification of the core technical aspects and investigate existing solution approaches, including heuristic, metaheuristic, learning-based, and hybrid strategies. Rather than introducing application-specific designs, the analysis focuses on workload-driven challenges and trade-offs that arise in AR/VR systems. Based on this classification, we analyze recent research trends, identify underexplored technical areas, and highlight key research gaps that hinder the efficient deployment of AR/VR services over edge-cloud infrastructures. The findings of this study provide practical insights for researchers and system designers and help guide future research toward more responsive, scalable, and reliable edge-cloud AR/VR systems.

Place, publisher, year, edition, pages
MDPI, 2026
Keywords
edge-cloud computing, network communication, service placement, offloading, caching, service migration
National Category
Computer and Information Sciences
Research subject
Computer Science
Identifiers
urn:nbn:se:kau:diva-109017 (URN)10.3390/app16031454 (DOI)001687571600001 ()2-s2.0-105030070913 (Scopus ID)
Available from: 2026-03-02 Created: 2026-03-02 Last updated: 2026-03-12Bibliographically approved
Garshasbi Herabad, M., Taheri, J., Ahmed, B. S. & Curescu, C. (2025). A Lightweight Learning-Based Approach for Online Edge-to-Cloud Service Placement. Electronics, 15(1), Article ID 65.
Open this publication in new window or tab >>A Lightweight Learning-Based Approach for Online Edge-to-Cloud Service Placement
2025 (English)In: Electronics, E-ISSN 2079-9292, Vol. 15, no 1, article id 65Article in journal (Refereed) Published
Abstract [en]

The integration of edge and cloud computing is critical for resource-intensive applications which require low-latency communication, high reliability, and efficient resource utilisation. The service placement problem in these environments poses significant challenges owing to dynamic network conditions, heterogeneous resource availability, and the necessity for real-time decision-making. Because determining an optimal service placement in such networks is an NP-complete problem, the existing solutions rely on fast but suboptimal heuristics or computationally intensive metaheuristics. Neither approach meets the real-time demands of online scenarios, owing to its inefficiency or high computational overhead. In this study, we propose a lightweight learning-based approach for the online placement of services with multi-version components in edge-to-cloud computing. The proposed approach utilises a Shallow Neural Network (SNN) with both weight and power coefficients optimised using a Genetic Algorithm (GA). The use of an SNN ensures low computational overhead during the training phase and almost instant inference when deployed, making it well suited for real-time and online service placement in edge-to-cloud environments where rapid decision-making is crucial. The proposed method (SNN-GA) is specifically evaluated in AR/VR-based remote repair and maintenance scenarios, developed in collaboration with our industrial partner, and demonstrated robust performance and scalability across a wide range of problem sizes. The experimental results show that SNN-GA reduces the service response time by up to 27% compared to metaheuristics and 55% compared to heuristics at larger scales. It also achieves over 95% platform reliability, outperforming heuristics (which remain below 85%) and metaheuristics (which decrease to 90% at larger scales).

Place, publisher, year, edition, pages
MDPI, 2025
Keywords
edge-to-cloud computing, online service placement, neural networks, genetic algorithm
National Category
Computer and Information Sciences
Research subject
Computer Science
Identifiers
urn:nbn:se:kau:diva-108232 (URN)10.3390/electronics15010065 (DOI)001657245900001 ()2-s2.0-105027941715 (Scopus ID)
Available from: 2026-01-19 Created: 2026-01-19 Last updated: 2026-03-04Bibliographically approved
HoseinyFarahabady, M. R., Taheri, J. & Zomaya, A. Y. .. (2025). Accelerating Key-Value Data Structures Using AVX-512 SIMD Extensions. In: Proceedings - IEEE International Conference on Cluster Computing: . Paper presented at IEEE International Conference on Cluster Computing (CLUSTER), Edinburgh, United Kingdom, September 2-5, 2025.. Institute of Electrical and Electronics Engineers (IEEE)
Open this publication in new window or tab >>Accelerating Key-Value Data Structures Using AVX-512 SIMD Extensions
2025 (English)In: Proceedings - IEEE International Conference on Cluster Computing, Institute of Electrical and Electronics Engineers (IEEE), 2025Conference paper, Published paper (Refereed)
Abstract [en]

Advanced Vector Extensions 512 (AVX-512), a modern SIMD instruction set for x86 architectures, enables data-level parallelism through 512-bit wide ZMM registers capable of processing multiple data elements concurrently within a single instruction cycle. In this study, we present a high-throughput, lock-free, in-memory architecture for key-value data-stores that exploits AVX-512 vector operations to accelerate fundamental operations such as insertion and lookup. Our design introduces an optimized memory layout that partitions the key space into two disjoint regions (primary and secondary) and employs three independent hash functions to identify candidate slots. This asymmetric layout improves key distribution, reduces collision probability, and enhances overall lookup efficiency. Experimental evaluation shows that this strategy yields the lowest insertion failure rate among tested memory partitioning schemes. By leveraging AVX-512 instructions in combination with most optimized memory layout, our implementation achieves insertion throughput within 6% of Intel TBB’s highly optimized multithreaded hash map, despite avoiding explicit synchronization or thread-level parallelism. Under workloads with 550 million entries and a 90% miss rate, our approach delivers 4.0-5.1x speedup over standard STL, Boost, Robin-Hood, and Abseil hash maps, and up to 2.5 x improvement relative to TBB and Abseil. These gains are consistently observed for both 32-bit and 64-bit floating-point key types. The results confirm the viability of AVX-512-centric designs as a cost-effective alternative to thread-level parallelism, particularly in environments where minimizing synchronization overhead and ensuring deterministic execution are critical. Our findings suggest for a paradigm shift in CPU and system architecture, emphasizing wider vector units and improved memory bandwidth utilization as primary levers for scalable high-performance computing. These findings suggest that future extensions of AVX-512 capabilities, such as non-blocking memory loads, expanded vector registers, and asynchronous prefetching, could enhance the efficiency of data-intensive workloads. 

Place, publisher, year, edition, pages
Institute of Electrical and Electronics Engineers (IEEE), 2025
Keywords
Data structures, Digital arithmetic, Hash functions, Memory architecture, Multitasking, Program processors, Table lookup, Throughput, Vectors, Advanced vector extension 512 intrinsic, CPU-based key-value data structure, Data access, Hash table, Hash table acceleration, High performance computing, Key values, Layout designs, Low latency, Low-latency data access, Memory layout, Memory layout design, Multiple data, Multiple data (SIMD) parallelism, Performance computing, Single instruction, Value data, Vectorized hashing, Failure analysis
National Category
Computer Systems Computer Engineering Computer Sciences
Research subject
Computer Science
Identifiers
urn:nbn:se:kau:diva-107584 (URN)10.1109/CLUSTER59342.2025.11186494 (DOI)001701307900006 ()2-s2.0-105019791230 (Scopus ID)979-8-3315-3019-8 (ISBN)979-8-3315-3020-4 (ISBN)
Conference
IEEE International Conference on Cluster Computing (CLUSTER), Edinburgh, United Kingdom, September 2-5, 2025.
Funder
Knowledge Foundation
Available from: 2025-11-18 Created: 2025-11-18 Last updated: 2026-03-27Bibliographically approved
Garshasbi Herabad, M., Taheri, J., Ahmed, B. S. & Curescu, C. (2025). E-PSOGA: An Enhanced Hybrid Metaheuristic for Optimal Edge-to-Cloud Placement of Services with Multi-Version Components. IEEE Access, 13, 151170-151188
Open this publication in new window or tab >>E-PSOGA: An Enhanced Hybrid Metaheuristic for Optimal Edge-to-Cloud Placement of Services with Multi-Version Components
2025 (English)In: IEEE Access, E-ISSN 2169-3536, Vol. 13, p. 151170-151188Article in journal (Refereed) Published
Abstract [en]

The evolution of edge-to-cloud networks has significantly increased the complexity of determining optimal service placement across these infrastructures, a challenge identified as an NP-complete problem. To address such problems, exact algorithms are impractical at larger scales owing to their computational demands. Heuristics exhibit faster runtimes but lower solution quality, whereas metaheuristics provide high-quality solutions at the cost of increased runtime. In this study, service placement in edge-to-cloud systems is investigated and formulated as an optimisation problem, where each service component is provided by different vendors and is available in multiple versions. The inclusion of multi-version components adds an additional layer of complexity, making the placement problem even more challenging. Specifically, this study addresses the service placement problem in Augmented Reality (AR)- and Virtual Reality (VR)-based remote repair and maintenance use cases, where service response time and system reliability are critical performance metrics. To optimise both metrics, we propose a novel hybrid metaheuristic algorithm (E-PSOGA) which combines the fast convergence of Particle Swarm Optimisation (PSO) with the global search capabilities of Genetic Algorithms (GA). A custom healing operator is also introduced to further enhance the solution quality and reduce the algorithm runtime. A comprehensive performance assessment shows that E-PSOGA reduces the response time by 37% compared with the other implemented baseline algorithms. E-PSOGA achieved 98% platform and 97% service reliability while maintaining a reasonable algorithm runtime. These results indicate that the proposed approach is well-suited for large-scale and time-sensitive scenarios requiring both computational efficiency and high solution quality. 

Place, publisher, year, edition, pages
Institute of Electrical and Electronics Engineers (IEEE), 2025
Keywords
Augmented reality, Complex networks, Computational complexity, Computational efficiency, Heuristic algorithms, Quality of service, Reliability, Repair, Response time (computer systems), Virtual reality, Cloud-computing, Edge-to-cloud computing, Multi-version, Optimal service placement, Particle swarm, Particle swarm optimization, Runtimes, Service placements, Solution quality, Swarm optimization, Genetic algorithms, Particle swarm optimization (PSO)
National Category
Computer Sciences Computer Systems Telecommunications
Research subject
Computer Science; Computer Science
Identifiers
urn:nbn:se:kau:diva-106828 (URN)10.1109/ACCESS.2025.3603329 (DOI)001562596000008 ()2-s2.0-105014473015 (Scopus ID)
Available from: 2025-09-08 Created: 2025-09-08 Last updated: 2026-02-12Bibliographically approved
Taghinezhad-Niar, A. & Taheri, J. (2025). Fault-Tolerant Cost-Efficient Scheduling for Energy and Deadline-Constrained IoT Workflows in Edge-Cloud Continuum. IEEE Transactions on Services Computing, 18(5), 2892-2903
Open this publication in new window or tab >>Fault-Tolerant Cost-Efficient Scheduling for Energy and Deadline-Constrained IoT Workflows in Edge-Cloud Continuum
2025 (English)In: IEEE Transactions on Services Computing, E-ISSN 1939-1374, Vol. 18, no 5, p. 2892-2903Article in journal (Refereed) Published
Abstract [en]

Edge computing brings computation closer to Internet-of-Things (IoT) data sources, reducing latency but increasing energy consumption and susceptibility to node failures. The cloud platform provides extensive computational capabilities, but comes with significant costs and communication delays due to network congestion. The edge-cloud continuum strategically combines these approaches to mitigate their individual drawbacks. However, effectively scheduling IoT workflows to minimize costs while adhering to strict requirements for latency, energy efficiency, and reliability remains a major challenge in real-time IoT applications. To address these challenges, we propose the Reliable Energy-constrained Cost-aware Real-time (RECR) algorithm for optimizing IoT workflow scheduling across the edge-cloud continuum. RECR minimizes monetary costs and enhances reliability while adhering to strict energy and deadline constraints. We also introduce RECR-D, a fault-tolerant extension that employs adaptive task duplication to manage transient and permanent failures, with reliability rigorously modeled using Continuous-Time Markov Chains (CTMCs) to integrate dynamic failure behavior. Extensive simulations demonstrate that RECR reduces workflow monetary costs by approximately 21% and improves deadline adherence by 37% compared to state-of-the-art algorithms. Furthermore, RECR-D improves compliance with reliability and energy constraints by 27% and by up to 208%, respectively, highlighting its robust performance in dynamic, failure-prone environments. These contributions significantly advance workflow management for IoT applications, proving crucial for real-time traffic control and video analytics in smart cities, ensuring timely processing and lower costs. They are also vital for remote patient monitoring and medical imaging analysis in healthcare, improving reliability and meeting deadlines for patient safety.

Place, publisher, year, edition, pages
IEEE, 2025
Keywords
Computational cost, Constrained optimization, Continuous time systems, Costs, Edge computing, Fault tolerance, Green computing, Markov processes, Medical imaging, Scheduling algorithms, Traffic congestion, Workflow management, Cost-aware, Edge, Edge clouds, Energy-constrained, Fault-tolerant, Latency, Real- time, Reliable energy, Work-flows, Workflow scheduling, Clouds
National Category
Computer Sciences Computer Systems
Research subject
Computer Science
Identifiers
urn:nbn:se:kau:diva-106747 (URN)10.1109/TSC.2025.3599497 (DOI)001591693600029 ()2-s2.0-105013296423 (Scopus ID)
Available from: 2025-09-03 Created: 2025-09-03 Last updated: 2026-02-12Bibliographically approved
Mahmoudi, A., Farzinvash, L. & Taheri, J. (2025). GPTOR: Gridded GA and PSO-based task offloading and ordering in IoT-edge-cloud computing. Results in Engineering (RINENG), 25, Article ID 104196.
Open this publication in new window or tab >>GPTOR: Gridded GA and PSO-based task offloading and ordering in IoT-edge-cloud computing
2025 (English)In: Results in Engineering (RINENG), ISSN 2590-1230, Vol. 25, article id 104196Article in journal (Refereed) Published
Abstract [en]

Edge computing is a key technology that provides computational resources close to IoT devices. One of the primary challenges in edge computing is determining whether to execute computation-intensive and time-sensitive tasks locally, or to offload them to edge and cloud computing resources, as well as to order them for execution according to their deadlines. Various offloading algorithms have been proposed for these systems, each with its own advantages and disadvantages. Several studies did not exploit all the IoT, edge, and cloud layers, whereas others only considered a few criteria for decision making on task offloading. Other approaches used greedy methods that could not provide high-quality solutions or employed standard optimization algorithms, which took a long time to converge. In this study, we propose an improved genetic algorithm for joint task offloading and ordering to distribute tasks across the IoT, edge, and cloud layers. It includes a novel population initialization scheme that uses various methods, including particle swarm optimization. To increase the convergence speed, the proposed algorithm (GPTOR) splits the solution space into several areas, which is called gridding. The simulation results illustrate that our algorithm outperforms previous schemes by 41.07%, 26.25%, and 28.33% in terms of average delay, monetary cost, and energy consumption, respectively. 

Place, publisher, year, edition, pages
Elsevier, 2025
Keywords
Cloud platforms, Computation offloading, Mobile edge computing, Optimization algorithms, Particle swarm optimization (PSO), Cloud layers, Cloud-computing, Customized operator, Edge computing, Gridding, Particle swarm, Particle swarm optimization, Swarm optimization, Task offloading, Task orders, Genetic algorithms
National Category
Computer Sciences Communication Systems Robotics and automation
Research subject
Computer Science
Identifiers
urn:nbn:se:kau:diva-103403 (URN)10.1016/j.rineng.2025.104196 (DOI)001423833600001 ()2-s2.0-85216897603 (Scopus ID)
Available from: 2025-02-25 Created: 2025-02-25 Last updated: 2026-02-12Bibliographically approved
Molaei, S., Sabaei, M. & Taheri, J. (2025). MRM-PSO: An enhanced particle swarm optimization technique for resource management in highly dynamic edge computing environments. Ad hoc networks, 178, Article ID 103952.
Open this publication in new window or tab >>MRM-PSO: An enhanced particle swarm optimization technique for resource management in highly dynamic edge computing environments
2025 (English)In: Ad hoc networks, ISSN 1570-8705, E-ISSN 1570-8713, Vol. 178, article id 103952Article in journal (Refereed) Published
Abstract [en]

The resource constraints of Internet of Things (IoT) devices pose significant hurdles to delay-sensitive applications that operate in dynamic and wireless settings. Since offloading tasks to cloud servers can be hindered by security concerns and latency issues, edge and fog computing bring computation closer to data sources. Given their inherently distributed and resource-constrained nature, edge/fog-enabled platforms require more advanced resource-management solutions to address the numerous constraints encountered in dynamic and wireless environments. This study introduces an innovative resource management algorithm designed for dynamic edge/fog computing environments, tailored to real-world applications, with the objective of enhancing delay performance through optimal container placement. The resource management problem incorporates mobility patterns in wireless settings to reduce migration delay and the processing history of edge/fog nodes to provide a novel method for computing processing delay, resulting in a combined optimization problem expressed in an integer linear programming (ILP) format. To address the formulated NP-Hard problem, we developed a low-complexity Metaheuristic Resource Management algorithm based on Particle Swarm Optimization (MRM-PSO) with effective particle modelling. Our experimental findings demonstrate that greedy heuristics and genetic algorithm (GA) are inadequate for efficiently resolving a given problem, whereas our proposed MRM-PSO algorithm efficiently locates near-optimal solutions within reasonable execution times when compared to exact solvers. MRM-PSO reduces execution time by up to 663.82 % in the worst case and 2307.5 % in the best case. Furthermore, it attains a delay that is just 0.98 % higher in the best case and 5.54 % higher in the worst case compared to the optimal solution.

Place, publisher, year, edition, pages
Elsevier, 2025
Keywords
Edge/fog computing, resource management, container placement, optimization, Particle Swarm Optimization (PSO)
National Category
Computer Sciences
Research subject
Computer Science
Identifiers
urn:nbn:se:kau:diva-105895 (URN)10.1016/j.adhoc.2025.103952 (DOI)001511785200002 ()2-s2.0-105007971256 (Scopus ID)
Available from: 2025-06-26 Created: 2025-06-26 Last updated: 2026-02-12Bibliographically approved
Galletta, A., Taheri, J., Di Modica, G. & Ficara, A. (2025). SIoTEc 2025 - 6th edition of ACM Workshop on Secure IoT, Edge and Cloud systems. In: Proceedings of the 34th ACM International Conference on Information and Knowledge Management (CIKM '25): . Paper presented at CIKM '25: The 34th ACM International Conference on Information and Knowledge Management. Seoul, Republic of Korea. November 10 - 14, 2025. (pp. 6905-6907). Association for Computing Machinery (ACM)
Open this publication in new window or tab >>SIoTEc 2025 - 6th edition of ACM Workshop on Secure IoT, Edge and Cloud systems
2025 (English)In: Proceedings of the 34th ACM International Conference on Information and Knowledge Management (CIKM '25), Association for Computing Machinery (ACM), 2025, p. 6905-6907Conference paper, Published paper (Refereed)
Abstract [en]

In the last years, we have seen an increase in the number of Artificial Intelligence (AI)-powered applications for information retrieval and data science. This fact led to an increasing reliance on distributed computing infrastructures, including Cloud, Edge, and IoT environments. These architectures enable powerful and scalable solutions but also introduce new security and privacy risks that must be addressed at both the system and data levels. Even a single breach on any of the links of the data-service-infrastructure chain may seriously compromise the security of the end-user application. With such a wide attack surface, security must definitely be approached in a holistic way and addressed in any layer where concerns may potentially arise. SIoTEC solicits novel and innovative ideas, proposals, positions and best practices that address the modelling, design, implementation, and enforcement of security in Cloud/Edge/IoT environments. Workshop website: https://siotec.netsons.org/ 

Place, publisher, year, edition, pages
Association for Computing Machinery (ACM), 2025
Keywords
artificial intelligence, cloud computing, data mining, data security, edge computing, iot, Cloud security, Data privacy, Human engineering, Internet of things, Cloud systems, Cloud-computing, Computing infrastructures, Privacy risks, Scalable solution, Security and privacy, Security risks, System levels
National Category
Computer and Information Sciences
Research subject
Computer Science
Identifiers
urn:nbn:se:kau:diva-107922 (URN)10.1145/3746252.3761600 (DOI)001661448300844 ()2-s2.0-105023136266 (Scopus ID)979-8-4007-2040-6 (ISBN)
Conference
CIKM '25: The 34th ACM International Conference on Information and Knowledge Management. Seoul, Republic of Korea. November 10 - 14, 2025.
Available from: 2025-12-17 Created: 2025-12-17 Last updated: 2026-03-24Bibliographically approved
Jagannathan, S., Sharma, Y. & Taheri, J. (2025). Towards Generic Failure-Prediction Models in Large-Scale Distributed Computing Systems. Electronics, 14(17), Article ID 3386.
Open this publication in new window or tab >>Towards Generic Failure-Prediction Models in Large-Scale Distributed Computing Systems
2025 (English)In: Electronics, E-ISSN 2079-9292, Vol. 14, no 17, article id 3386Article in journal (Refereed) Published
Abstract [en]

The increasing complexity of Distributed Computing (DC) systems requires advanced failure-prediction models to enhance reliability and efficiency. This study proposes a comprehensive methodology for developing generic machine learning (ML) models capable of cross-layer and cross-platform failure-prediction without requiring platform-specific retraining. Using the Grid5000 failure dataset from the Failure Trace Archive (FTA), we explored Linear and Logistic Regression, Random Forest, and XGBoost to predict three critical metrics: Time Between Failures (TBF), Time to Return/Repair (TTR), and Failing Node Identification (FNI). Our approach involved extensive exploratory data analysis (EDA), statistical examination of failure patterns, and model evaluation across the cluster, site, and system levels. The results demonstrate that XGBoost consistently outperforms the other models, achieving near-perfect 100% accuracy for TBF and FNI, with robust generalisability across diverse DC environments. In addition, we introduce a hierarchical DC architecture that integrates these failure-prediction models. In the form of a use case, we also demonstrate how service providers can use these prediction models to balance service reliability and cost.

Place, publisher, year, edition, pages
MDPI, 2025
Keywords
distributed computing, fault detection, machine learning algorithms, prediction algorithms, performance evaluation
National Category
Computer Sciences Computer Systems
Research subject
Computer Science
Identifiers
urn:nbn:se:kau:diva-107038 (URN)10.3390/electronics14173386 (DOI)001569639800001 ()2-s2.0-105016627350 (Scopus ID)
Funder
Knowledge Foundation
Available from: 2025-09-26 Created: 2025-09-26 Last updated: 2026-02-12Bibliographically approved
Organisations
Identifiers
ORCID iD: ORCID iD iconorcid.org/0000-0001-9194-010X

Search in DiVA

Show all publications