CiteExportLink to record
Permanent link

Direct link
Cite
Citation style
  • apa
  • ieee
  • modern-language-association-8th-edition
  • vancouver
  • apa.csl
  • Other style
More styles
Language
  • de-DE
  • en-GB
  • en-US
  • fi-FI
  • nn-NO
  • nn-NB
  • sv-SE
  • Other locale
More languages
Output format
  • html
  • text
  • asciidoc
  • rtf
Machine Learning in Motion: Engineering Self-Adaptive Systems
Karlstad University, Faculty of Health, Science and Technology (starting 2013), Department of Mathematics and Computer Science (from 2013).ORCID iD: 0000-0003-0683-2783
2025 (English)Doctoral thesis, comprehensive summary (Other academic)
Abstract [en]

In the rapidly evolving landscape of artificial intelligence (AI) and machine learning (ML), transitioning from theoretical models to robust, adaptive production systems remains a significant challenge in the digital age. As data-driven methodologies revolutionize problem solving across industries, a critical gap exists between research achievements and reliable real-world deployment. Key obstacles include concept drift and data quality issues that arise from unpredictable changes in data distributions and operational complexities that systematically affect model performance, reliability, and efficiency. This thesis addresses these challenges by introducing an overarching framework for adaptive ML systems that operate reliably in dynamic, real-world environments, incorporating innovative methodologies for dynamic drift detection and real-time data quality assessment bounded by robust Machine Learning Operations (MLOps) strategies. These integrated components enable the creation of production-grade ML systems that can efficiently adapt to shifts in data distributions and assess data quality in real-time, ensuring stable and reliable performance in dynamic environments. The proposed approaches are validated through real-world use cases, demonstrating significant improvements in predictive accuracy and operational efficiency. By deploying these adaptive systems in industrial contexts, the thesis highlights their potential to deliver reliable, high-performance ML solutions tailored to the demands of complex time-sensitive applications. This work offers concrete solutions for translating theoretical advances into practical applications, contributing to developing robust and scalable ML systems for real-world deployment.

Abstract [en]

In the era of AI-driven innovation, ensuring the reliability and adaptability of machine learning (ML) systems in dynamic environments remains a fundamental challenge. This thesis addresses key obstacles such as concept drift and data quality, which affect model performance in real-world applications. By integrating novel drift detection mechanisms, real-time data quality assessment, and robust MLOps strategies, this work proposes a comprehensive framework for self-adaptive ML systems. These solutions enable industrial ML deployments to maintain predictive accuracy, efficiency, and resilience against evolving data conditions. Through real-world case studies, this research demonstrates significant improvements in operational stability and decision-making, bridging the gap between theoretical advancements and practical implementation. By tackling these challenges, this thesis contributes to the development of scalable, trustworthy, and high-performance ML solutions tailored for industrial and mission-critical applications.

Place, publisher, year, edition, pages
Karlstad: Karlstads universitet, 2025. , p. 30
Series
Karlstad University Studies, ISSN 1403-8099 ; 2025:14
Keywords [en]
machine learning, concept drift, data quality, adaptive learning, MLOps, data-driven development, performance robustness
National Category
Artificial Intelligence Computer Sciences
Research subject
Computer Science
Identifiers
URN: urn:nbn:se:kau:diva-103579DOI: 10.59217/rjrp8643ISBN: 978-91-7867-560-9 (print)ISBN: 978-91-7867-561-6 (print)OAI: oai:DiVA.org:kau-103579DiVA, id: diva2:1945012
Public defence
2025-04-25, Eva Eriksson lecture hall, 21A342, Karlstad, 09:00 (English)
Opponent
Supervisors
Available from: 2025-04-07 Created: 2025-03-17 Last updated: 2025-04-07Bibliographically approved
List of papers
1. Towards Trustworthy Machine Learning in Production: An Overview of the Robustness in MLOps Approach
Open this publication in new window or tab >>Towards Trustworthy Machine Learning in Production: An Overview of the Robustness in MLOps Approach
2025 (English)In: ACM Computing Surveys, ISSN 0360-0300, E-ISSN 1557-7341, Vol. 57, no 5, article id 121Article in journal (Refereed) Published
Abstract [en]

Artificial intelligence (AI), and especially its sub-field of Machine Learning (ML), are impacting the daily lives of everyone with their ubiquitous applications. In recent years, AI researchers and practitioners have introduced principles and guidelines to build systems that make reliable and trustworthy decisions. From a practical perspective, conventional ML systems process historical data to extract the features that are consequently used to train ML models that perform the desired task. However, in practice, a fundamental challenge arises when the system needs to be operationalized and deployed to evolve and operate in real-life environments continuously. To address this challenge, Machine Learning Operations (MLOps) have emerged as a potential recipe for standardizing ML solutions in deployment. Although MLOps demonstrated great success in streamlining ML processes, thoroughly defining the specifications of robust MLOps approaches remains of great interest to researchers and practitioners. In this paper, we provide a comprehensive overview of the trustworthiness property of MLOps systems. Specifically, we highlight technical practices to achieve robust MLOps systems. In addition, we survey the existing research approaches that address the robustness aspects of ML systems in production. We also review the tools and software available to build MLOps systems and summarize their support to handle the robustness aspects. Finally, we present the open challenges and propose possible future directions and opportunities within this emerging field. The aim of this paper is to provide researchers and practitioners working on practical AI applications with a comprehensive view to adopt robust ML solutions in production environments.

Place, publisher, year, edition, pages
Association for Computing Machinery (ACM), 2025
Keywords
Contrastive Learning, Dataops, Machine learning operation system, Machine learning systems, Machine-learning, Modeling performance, Modelop, Operation system, Robustness, Trustworthy artificial intelligence, Adversarial machine learning
National Category
Computer Sciences Software Engineering
Research subject
Computer Science; Computer Science
Identifiers
urn:nbn:se:kau:diva-103395 (URN)10.1145/3708497 (DOI)001416336700014 ()2-s2.0-85217023025 (Scopus ID)
Funder
Knowledge Foundation, 20200067
Available from: 2025-02-25 Created: 2025-02-25 Last updated: 2025-03-17Bibliographically approved
2. From concept drift to model degradation: An overview on performance-aware drift detectors
Open this publication in new window or tab >>From concept drift to model degradation: An overview on performance-aware drift detectors
2022 (English)In: Knowledge-Based Systems, ISSN 0950-7051, E-ISSN 1872-7409, Vol. 245, article id 108632Article, review/survey (Refereed) [Artistic work] Published
Abstract [en]

The dynamicity of real-world systems poses a significant challenge to deployed predictive machine learning (ML) models. Changes in the system on which the ML model has been trained may lead to performance degradation during the system’s life cycle. Recent advances that study non-stationary environments have mainly focused on identifying and addressing such changes caused by a phenomenon called concept drift. Different terms have been used in the literature to refer to the same type of concept drift and the same term for various types. This lack of unified terminology is set out to create confusion on distinguishing between different concept drift variants. In this paper, we start by grouping concept drift types by their mathematical definitions and survey the different terms used in the literature to build a consolidated taxonomy of the field. We also review and classify performance-based concept drift detection methods proposed in the last decade. These methods utilize the predictive model’s performance degradation to signal substantial changes in the systems. The classification is outlined in a hierarchical diagram to provide an orderly navigation between the methods. We present a comprehensive analysis of the main attributes and strategies for tracking and evaluating the model’s performance in the predictive system. The paper concludes by discussing open research challenges and possible research directions.

National Category
Computer and Information Sciences
Research subject
Computer Science
Identifiers
urn:nbn:se:kau:diva-89786 (URN)10.1016/j.knosys.2022.108632 (DOI)000795145800005 ()2-s2.0-85127829086 (Scopus ID)
Projects
AIDA
Available from: 2022-05-16 Created: 2022-05-16 Last updated: 2025-03-17Bibliographically approved
3. DA-LSTM: A dynamic drift-adaptive learning framework for interval load forecasting with LSTM networks
Open this publication in new window or tab >>DA-LSTM: A dynamic drift-adaptive learning framework for interval load forecasting with LSTM networks
Show others...
2023 (English)In: Engineering applications of artificial intelligence, ISSN 0952-1976, E-ISSN 1873-6769, Vol. 123, article id 106480Article in journal (Refereed) Published
Abstract [en]

Load forecasting is a crucial topic in energy management systems (EMS) due to its vital role in optimizing energy scheduling and enabling more flexible and intelligent power grid systems. As a result, these systems allow power utility companies to respond promptly to demands in the electricity market. Deep learning (DL) models have been commonly employed in load forecasting problems supported by adaptation mechanisms to cope with the changing pattern of consumption by customers, known as concept drift. A drift magnitude threshold should be defined to design change detection methods to identify drifts. While the drift magnitude in load forecasting problems can vary significantly over time, existing literature often assumes a fixed drift magnitude threshold, which should be dynamically adjusted rather than fixed during system evolution. To address this gap, in this paper, we propose a dynamic drift-adaptive Long Short-Term Memory (DA-LSTM) framework that can improve the performance of load forecasting models without requiring a drift threshold setting. We integrate several strategies into the framework based on active and passive adaptation approaches. To evaluate DA-LSTM in real-life settings, we thoroughly analyze the proposed framework and deploy it in a real-world problem through a cloud-based environment. Efficiency is evaluated in terms of the prediction performance of each approach and computational cost. The experiments show performance improvements on multiple evaluation metrics achieved by our framework compared to baseline methods from the literature. Finally, we present a trade-off analysis between prediction performance and computational costs.

Place, publisher, year, edition, pages
Elsevier, 2023
Keywords
Concept drift Change-point detection Dynamic drift adaptation Adaptive LSTM Interval load forecasting
National Category
Computer Sciences
Research subject
Computer Science
Identifiers
urn:nbn:se:kau:diva-95022 (URN)10.1016/j.engappai.2023.106480 (DOI)001013639800001 ()2-s2.0-85160615665 (Scopus ID)
Funder
Knowledge Foundation, 20200067Swedish Energy Agency, 50246-1SOLVE, 52693-1
Available from: 2023-06-02 Created: 2023-06-02 Last updated: 2025-03-17Bibliographically approved
4. DQSOps: Data Quality Scoring Operations Framework for Data-Driven Applications
Open this publication in new window or tab >>DQSOps: Data Quality Scoring Operations Framework for Data-Driven Applications
2023 (English)In: EASE '23: Proceedings of the 27th International Conference on Evaluation and Assessment in Software Engineering, Association for Computing Machinery (ACM), 2023, p. 32-41Conference paper, Published paper (Refereed)
Abstract [en]

Data quality assessment has become a prominent component in the successful execution of complex data-driven artificial intelligence (AI) software systems. In practice, real-world applications generate huge volumes of data at speeds. These data streams require analysis and preprocessing before being permanently stored or used in a learning task. Therefore, significant attention has been paid to the systematic management and construction of high-quality datasets. Nevertheless, managing voluminous and high-velocity data streams is usually performed manually (i.e. offline), making it an impractical strategy in production environments. To address this challenge, DataOps has emerged to achieve life-cycle automation of data processes using DevOps principles. However, determining the data quality based on a fitness scale constitutes a complex task within the framework of DataOps. This paper presents a novel Data Quality Scoring Operations (DQSOps) framework that yields a quality score for production data in DataOps workflows. The framework incorporates two scoring approaches, an ML prediction-based approach that predicts the data quality score and a standard-based approach that periodically produces the ground-truth scores based on assessing several data quality dimensions. We deploy the DQSOps framework in a real-world industrial use case. The results show that DQSOps achieves significant computational speedup rates compared to the conventional approach of data quality scoring while maintaining high prediction performance.

Place, publisher, year, edition, pages
Association for Computing Machinery (ACM), 2023
Keywords
Data reduction, Quality control, Automated data, Automated data scoring, Data assessment, Data quality; Data quality dimensions, Data stream, Data-driven applications, Dataops, Mutation testing, Real-world, Life cycle
National Category
Software Engineering Computer Sciences Computer Systems
Research subject
Computer Science
Identifiers
urn:nbn:se:kau:diva-95940 (URN)10.1145/3593434.3593445 (DOI)2-s2.0-85162259466 (Scopus ID)979-8-4007-0044-6 (ISBN)
Conference
The International Conference on Evaluation and Assessment in Software Engineering, Oulu, Finland, June 14-16, 2023.
Funder
Knowledge Foundation
Available from: 2023-07-04 Created: 2023-07-04 Last updated: 2025-03-17Bibliographically approved
5. Adaptive data quality scoring operations framework using drift-aware mechanism for industrial applications
Open this publication in new window or tab >>Adaptive data quality scoring operations framework using drift-aware mechanism for industrial applications
2024 (English)In: Journal of Systems and Software, ISSN 0164-1212, E-ISSN 1873-1228, Vol. 217, p. 112184-112184, article id 112184Article in journal (Refereed) Published
National Category
Artificial Intelligence
Identifiers
urn:nbn:se:kau:diva-103575 (URN)10.1016/j.jss.2024.112184 (DOI)
Available from: 2025-03-17 Created: 2025-03-17 Last updated: 2025-03-17
6. End-to-End Data Quality-Driven Framework for Machine Learningin Production Environment
Open this publication in new window or tab >>End-to-End Data Quality-Driven Framework for Machine Learningin Production Environment
(English)In: Heliyon, E-ISSN 2405-8440Article in journal (Refereed) Submitted
Abstract [en]

This paper introduces a novel end-to-end framework that efficiently inte-grates data quality assessment with machine learning (ML) model operationsin real-time production environments. While existing approaches treat dataquality assessment and ML systems as isolated processes, our framework ad-dresses the critical gap between theoretical methods and practical implemen-tation by combining dynamic drift detection, adaptive data quality metrics,and MLOps into a cohesive, lightweight system. The key innovation liesin its operational efficiency, enabling real-time, quality-driven ML decision-making with minimal computational overhead. We validate the frameworkin a steel manufacturing company’s Electroslag Remelting (ESR) vacuumpumping process, demonstrating a 12% improvement in model performance(R² = 94%) and a fourfold reduction in prediction latency. By exploringthe impact of data quality acceptability thresholds, we provide actionableinsights into balancing data quality standards and predictive performancein industrial applications. This framework represents a significant advance-ment in MLOps, offering a robust solution for time-sensitive, data-drivendecision-making in dynamic industrial environments.

National Category
Artificial Intelligence
Identifiers
urn:nbn:se:kau:diva-103578 (URN)
Available from: 2025-03-17 Created: 2025-03-17 Last updated: 2025-03-17

Open Access in DiVA

KAPPAN(1000 kB)48 downloads
File information
File name FULLTEXT04.pdfFile size 1000 kBChecksum SHA-512
fe279ed10fb1fff54beb82e56754d871e96ece3204017bfcac4cc27e7341a9725f638bc9d5d60e6810d2399824a90e0260ee4d47ddad1934dd236af5c92bc262
Type fulltextMimetype application/pdf

Other links

Publisher's full text

Authority records

Bayram, Firas

Search in DiVA

By author/editor
Bayram, Firas
By organisation
Department of Mathematics and Computer Science (from 2013)
Artificial IntelligenceComputer Sciences

Search outside of DiVA

GoogleGoogle Scholar
Total: 53 downloads
The number of downloads is the sum of all downloads of full texts. It may include eg previous versions that are now no longer available

doi
isbn
urn-nbn

Altmetric score

doi
isbn
urn-nbn
Total: 524 hits
CiteExportLink to record
Permanent link

Direct link
Cite
Citation style
  • apa
  • ieee
  • modern-language-association-8th-edition
  • vancouver
  • apa.csl
  • Other style
More styles
Language
  • de-DE
  • en-GB
  • en-US
  • fi-FI
  • nn-NO
  • nn-NB
  • sv-SE
  • Other locale
More languages
Output format
  • html
  • text
  • asciidoc
  • rtf