AI models face real-world reality check in 6G network slicing
One of the key findings is that raw accuracy in laboratory conditions does not guarantee stability in deployment. The study introduced a resilience-based evaluation of the algorithms, ranking them by how much their performance degraded under realistic scenarios. SVM, Logistic Regression, and k-Nearest Neighbors emerged as the most resilient models, while CNN and FNN remained strong but less stable. Naive Bayes suffered the steepest drop.

A new study published in Electronics reveals that machine learning models widely considered top performers in controlled laboratory settings can lose significant accuracy in real-world 6G network slicing scenarios.
The study, titled “Realistic Performance Assessment of Machine Learning Algorithms for 6G Network Slicing: A Dual-Methodology Approach with Explainable AI Integration,” highlights the need for rigorous, realistic testing of artificial intelligence tools before their deployment in next-generation telecom networks, where stability and explainability can be just as crucial as raw accuracy.
Lab champions struggle in real-world conditions
The research team examined how 11 different machine learning and deep learning algorithms perform in the critical task of classifying network slices in 6G environments. These slices, which prioritize services ranging from ultra-reliable low-latency communication to massive IoT connectivity, must be allocated efficiently for the network to function under diverse conditions.
Using a dual-methodology framework, the authors compared two types of datasets: a clean, lab-style dataset of 10,000 samples and a “realistic” dataset of 21,033 samples that incorporated market-driven class imbalances, noise, and missing data, mirroring real-world traffic patterns and urban interference.
The results showed a sharp shift in performance when models moved from clean to realistic conditions. Convolutional Neural Networks (CNN) and Feedforward Neural Networks (FNN), which achieved near-perfect accuracy in controlled settings, saw their accuracy drop to around 81 percent when tested under realistic conditions. Meanwhile, simpler models such as Support Vector Machines (SVM) and Logistic Regression improved their performance, demonstrating more resilience to real-world data distortions.
Resilience and explainability take center stage
One of the key findings is that raw accuracy in laboratory conditions does not guarantee stability in deployment. The study introduced a resilience-based evaluation of the algorithms, ranking them by how much their performance degraded under realistic scenarios. SVM, Logistic Regression, and k-Nearest Neighbors emerged as the most resilient models, while CNN and FNN remained strong but less stable. Naive Bayes suffered the steepest drop.
Besides performance, the authors integrated Explainable AI (XAI) techniques, specifically SHAP (Shapley Additive Explanations), to interpret model decisions. This analysis revealed that Packet Loss Budget consistently emerged as the most influential factor for classification across all models, followed by jitter and latency.
In terms of explanation stability, CNN, LSTM, Naive Bayes, SVM, and Logistic Regression showed consistent reasoning across varying conditions, while XGBoost displayed lower explanation stability, suggesting it could behave unpredictably in dynamic environments despite competitive accuracy.
Implications for 6G network deployment
The study’s findings carry significant implications for telecom operators and equipment vendors preparing for the rollout of 6G networks, which will rely heavily on network slicing to handle diverse applications simultaneously.
The authors caution against choosing models based solely on laboratory benchmarks. In real deployments, stability, interpretability, and resilience become vital factors, especially as network operators must ensure reliable performance and meet regulatory and service-level requirements.
They recommend prioritizing models that not only achieve high accuracy but also maintain their performance across varied traffic conditions and provide transparent, explainable decision-making processes. Packet Loss Budget, identified as a key determinant in slice classification, should be closely monitored in live network operations to optimize performance and ensure fairness across competing services.
- FIRST PUBLISHED IN:
- Devdiscourse