Metzger, Andreas; Quinton, Clément; Mann, Zoltán Ádám; Baresi, Luciano; Pohl, Klaus:
Realizing self-adaptive systems via online reinforcement learning and feature-model-guided exploration
In: Computing, Vol. 106 (2024), No. 4, pp. 1251 - 1272
2024article/chapter in journalOA Hybrid
Computer ScienceScientific institutes » paluno - The Ruhr Institute for Software Technology
Related: 1 publication(s)
Title in English:
Realizing self-adaptive systems via online reinforcement learning and feature-model-guided exploration
Author:
Metzger, AndreasUDE
GND
129831530
LSF ID
12274
ORCID
0000-0002-4808-8297ORCID iD
Other
connected with university
;
Quinton, Clément
;
Mann, Zoltán ÁdámUDE
GND
1022217747
LSF ID
57890
ORCID
0000-0001-5741-2709ORCID iD
Other
connected with university
corresponding author
;
Baresi, Luciano
;
Pohl, KlausUDE
GND
135789125
LSF ID
5105
ORCID
0000-0003-2199-5257ORCID iD
Other
connected with university
Year of publication:
2024
Open Access?:
OA Hybrid
Web of Science ID
Scopus ID
Language of text:
English
Keyword, Topic:
Adaptive system ; Evolution ; Feature model ; Reinforcement learning
Type of resource:
Text

Abstract in English:

A self-adaptive system can automatically maintain its quality requirements in the presence of dynamic environment changes. Developing a self-adaptive system may be difficult due to design time uncertainty; e.g., anticipating all potential environment changes at design time is in most cases infeasible. To realize self-adaptive systems in the presence of design time uncertainty, online machine learning, i.e., machine learning at runtime, is increasingly used. In particular, online reinforcement learning is proposed, which learns suitable adaptation actions through interactions with the environment at runtime. To learn about its environment, online reinforcement learning has to select actions that were not selected before, which is known as exploration. How exploration happens impacts the performance of the learning process. We focus on two problems related to how adaptation actions are explored. First, existing solutions randomly explore adaptation actions and thus may exhibit slow learning if there are many possible adaptation actions. Second, they are unaware of system evolution, and thus may explore new adaptation actions introduced during evolution rather late. We propose novel exploration strategies that use feature models (from software product line engineering) to guide exploration in the presence of many adaptation actions and system evolution. Experimental results for two realistic self-adaptive systems indicate an average speed-up of the learning process of 33.7% in the presence of many adaptation actions, and of 50.6% in the presence of evolution.