Metzger, Andreas; Quinton, Clément; Mann, Zoltán Ádám; Baresi, Luciano; Pohl, Klaus:
Realizing self-adaptive systems via online reinforcement learning and feature-model-guided exploration
In: Computing, Jg. 106 (2024), Heft 4, S. 1251 - 1272
2024Artikel/Aufsatz in ZeitschriftOA Hybrid
InformatikForschungszentren » paluno - The Ruhr Institute for Software Technology
Damit verbunden: 1 Publikation(en)
Titel in Englisch:
Realizing self-adaptive systems via online reinforcement learning and feature-model-guided exploration
Autor*in:
Metzger, AndreasUDE
GND
129831530
LSF ID
12274
ORCID
0000-0002-4808-8297ORCID iD
Sonstiges
der Hochschule zugeordnete*r Autor*in
;
Quinton, Clément
;
Mann, Zoltán ÁdámUDE
GND
1022217747
LSF ID
57890
ORCID
0000-0001-5741-2709ORCID iD
Sonstiges
der Hochschule zugeordnete*r Autor*in
korrespondierende*r Autor*in
;
Baresi, Luciano
;
Pohl, KlausUDE
GND
135789125
LSF ID
5105
ORCID
0000-0003-2199-5257ORCID iD
Sonstiges
der Hochschule zugeordnete*r Autor*in
Erscheinungsjahr:
2024
Open Access?:
OA Hybrid
Web of Science ID
Scopus ID
Sprache des Textes:
Englisch
Schlagwort, Thema:
Adaptive system ; Evolution ; Feature model ; Reinforcement learning
Ressourcentyp:
Text

Abstract in Englisch:

A self-adaptive system can automatically maintain its quality requirements in the presence of dynamic environment changes. Developing a self-adaptive system may be difficult due to design time uncertainty; e.g., anticipating all potential environment changes at design time is in most cases infeasible. To realize self-adaptive systems in the presence of design time uncertainty, online machine learning, i.e., machine learning at runtime, is increasingly used. In particular, online reinforcement learning is proposed, which learns suitable adaptation actions through interactions with the environment at runtime. To learn about its environment, online reinforcement learning has to select actions that were not selected before, which is known as exploration. How exploration happens impacts the performance of the learning process. We focus on two problems related to how adaptation actions are explored. First, existing solutions randomly explore adaptation actions and thus may exhibit slow learning if there are many possible adaptation actions. Second, they are unaware of system evolution, and thus may explore new adaptation actions introduced during evolution rather late. We propose novel exploration strategies that use feature models (from software product line engineering) to guide exploration in the presence of many adaptation actions and system evolution. Experimental results for two realistic self-adaptive systems indicate an average speed-up of the learning process of 33.7% in the presence of many adaptation actions, and of 50.6% in the presence of evolution.