Abstract
Although deep learning is increasingly promising in the field of Non-Intrusive Load Monitoring (NILM) these days, the high costs of data recording and labelling represent a significant challenge for the training of supervised models. To address this, a cost-effective sequence-to-points NILM solution is proposed, integrating three-point labelling with non-causal convolution techniques. The approach introduces a semi-automatic labelling framework for obtaining NILM three-point data, which provides a low-cost data collection and labelling solution for large-scale applications. Then, a novel loss function combining coordinate loss and confidence loss is developed to address the positional misalignment and negative sample confusion in sequence-to-points scenario in NILM. Furthermore, an advanced neural network architecture based on multi-scale non-causal temporal convolution techniques is designed to capture unique features and operational modes of different appliances. Experimental results on the UK-DALE dataset show that the proposed mixed loss function has an advantage over plain Mean Absolute Error (MAE) on the sequence-to-points occasion, and the novel network outperforms on all of the appliances, demonstrating its potential for practical NILM applications.
Recommended Citation
Y. Zhang et al., "A Cost-Effective NILM Solution with Three-Point Labelling and Non-Causal Convolution Technique," Iet Smart Grid, vol. 8, no. 1, article no. e70036, Institution of Engineering and Technology (IET); Wiley, Jan 2025.
The definitive version is available at https://doi.org/10.1049/stg2.70036
Department(s)
Electrical and Computer Engineering
Publication Status
Open Access
Keywords and Phrases
deep learning; IoU-based loss function; multi-scale; non-causal convolution; non-intrusive load monitoring; three-point labeling
International Standard Serial Number (ISSN)
2515-2947
Document Type
Article - Journal
Document Version
Final Version
File Type
text
Language(s)
English
Rights
© 2025 The Authors, All rights reserved.
Creative Commons Licensing

This work is licensed under a Creative Commons Attribution 4.0 License.
Publication Date
01 Jan 2025
