TY - GEN
T1 - Lightweight Alternatives for Hyper-parameter Tuning in Drifting Data Streams
AU - Lobo, Jesus L.
AU - Del Ser, Javier
AU - Osaba, Eneko
N1 - Publisher Copyright:
© 2021 IEEE.
PY - 2021
Y1 - 2021
N2 - Scenarios dealing with data streams often undergo changes in data distribution, which ultimately lead to a performance degradation of algorithms learning from such data flows (concept drift). This phenomenon calls for the adoption of adaptive learning strategies for algorithms to perform resiliently after a change occurs. A multiplicity of approaches have so far addressed this issue by assorted means, e.g. instances weighting, ensembling, instance selection, or parameter tuning, among others. This latter strategy is often neglected as it requires a hyper-parameter tuning process that stream learning scenarios cannot computationally afford in most practical settings. Processing times and memory space are usually severely constrained, thus making the tuning phase unfeasible. Consequently, the research community has largely opted for other adaptive strategies with lower computational demands. This work outlines a new perspective to alleviate the hyper-parameter tuning process in the context of concept drift adaptation. We propose two simple and lightweight mechanisms capable of discovering competitive configurations of learning algorithms used for data stream classification. We compare its performance to that of a modern hyper-parametric search method (Successive Halving) over extensive experiments with synthetic and real datasets. We conclude that our proposed methods perform competitively, while consuming less processing time and memory.
AB - Scenarios dealing with data streams often undergo changes in data distribution, which ultimately lead to a performance degradation of algorithms learning from such data flows (concept drift). This phenomenon calls for the adoption of adaptive learning strategies for algorithms to perform resiliently after a change occurs. A multiplicity of approaches have so far addressed this issue by assorted means, e.g. instances weighting, ensembling, instance selection, or parameter tuning, among others. This latter strategy is often neglected as it requires a hyper-parameter tuning process that stream learning scenarios cannot computationally afford in most practical settings. Processing times and memory space are usually severely constrained, thus making the tuning phase unfeasible. Consequently, the research community has largely opted for other adaptive strategies with lower computational demands. This work outlines a new perspective to alleviate the hyper-parameter tuning process in the context of concept drift adaptation. We propose two simple and lightweight mechanisms capable of discovering competitive configurations of learning algorithms used for data stream classification. We compare its performance to that of a modern hyper-parametric search method (Successive Halving) over extensive experiments with synthetic and real datasets. We conclude that our proposed methods perform competitively, while consuming less processing time and memory.
KW - concept drift adaptation
KW - Data stream learning
KW - hyper-parameter tuning
UR - http://www.scopus.com/inward/record.url?scp=85125354686&partnerID=8YFLogxK
U2 - 10.1109/ICDMW53433.2021.00045
DO - 10.1109/ICDMW53433.2021.00045
M3 - Conference contribution
AN - SCOPUS:85125354686
T3 - IEEE International Conference on Data Mining Workshops, ICDMW
SP - 304
EP - 311
BT - Proceedings - 21st IEEE International Conference on Data Mining Workshops, ICDMW 2021
A2 - Xue, Bing
A2 - Pechenizkiy, Mykola
A2 - Koh, Yun Sing
PB - IEEE Computer Society
T2 - 21st IEEE International Conference on Data Mining Workshops, ICDMW 2021
Y2 - 7 December 2021 through 10 December 2021
ER -