Adding attention mechanisms to neural networks has been widely explored in recent years. Within the realm of Non-Intrusive Load Monitoring (NILM), however, the potential of attention has not been fully investigated to date. While the few existing works found that attention can lead to lower disaggregation errors and/or greater model robustness, what remains unclear is when and under which conditions these gains can be accomplished. We hence conduct a systematic analysis of different attention mechanisms for energy data disaggregation, i.e., Non-Intrusive Load Monitoring (NILM). More specifically, we select three distinct attention mechanisms: Channel Attention (CA), Feed-Forward Attention (FFA), and Self-Attention (SA) and extend them to construct seven different attention modules, which are integrated into the existing sequence-based neural network architecture to develop enhanced models. We evaluate the performance of the seven configurations on the publicly available UK-DALE and REDD datasets under cross-house and cross-dataset settings to quantify both in-distribution accuracy and transfer to unseen data, as is typical in real-world scenarios. Across these controlled comparisons, we found that the effects of attention are strongly context-dependent. That is, they do not only vary with appliance dynamics, but also depend on the dataset characteristics. While some attention configurations showed notable gains for a few appliances, their use led to a lower accuracy for other devices. What is more, accuracy results also differed when trying to disaggregate the same appliance type in two different datasets. Besides discussing these insights in more depth, we further quantify the resource tradeoffs of each attention family. We show that the incurred overhead (in terms of required computation) differs starkly between attention types, yet does not directly correlate with increased performance. In order to cater for a deeper understanding of the potentials and limitations of attention, our work provides design guidance for both practitioners and model designers on choosing attention families under accuracy and efficiency constraints.
Can Attention Improve Sequence-to-Point Load Disaggregation? A Comparative Assessment
Mazen Bouchur,Nan Li,Andreas Reinhardt
Published 2025 in International Conference on Systems for Energy-Efficient Built Environments
ABSTRACT
PUBLICATION RECORD
- Publication year
2025
- Venue
International Conference on Systems for Energy-Efficient Built Environments
- Publication date
2025-11-11
- Fields of study
Computer Science, Engineering, Environmental Science
- Identifiers
- External record
- Source metadata
Semantic Scholar
CITATION MAP
EXTRACTION MAP
CLAIMS
- No claims are published for this paper.
CONCEPTS
- No concepts are published for this paper.
REFERENCES
Showing 1-36 of 36 references · Page 1 of 1
CITED BY
- No citing papers are available for this paper.
Showing 0-0 of 0 citing papers · Page 1 of 1