A repository of awesome Non-Intrusive Load Monitoring (NILM) with code
😸 Welcome valuable opinions from researchers in the same direction
- NILMTK: Non-Intrusive Load Monitoring Toolkit [PDF] [CODE]
- NILMTK-Contrib [PDF] [Model] [Tensorflow]
- NILM-Eval [Matlab]
- Torch-NILM [PDF] [Pytorch]
- Deep-NILMtk [PDF] [Pytorch] [Tensorflow]
- nilmtk-ukdale: Exploratory Data Analysis [Pytorch]
- NeuralNILM_Pytorch [Pytorch]
- nilm_analyzer [Pytorch]
- Energy Informatics
- International Workshop on Non-Intrusive Load Monitoring [nilmworkshop.org]
- International Conference on Power Engineering and Renewable Energy (ICPERE)
Type | Dataset |
---|---|
Residential datasets | [UK-DALE] [REDD] [REFIT] [AMpds/2] [Dataport] [ECO] [ENERTALK] [iAWE] [BLUED] [PLAID] [DRED] [Georges Hebrail UCI] [GREEND] [HES] [TraceBase] [IDEAL] [HawkDATA] |
Commercial buildings datasets | [COMBED] [BLOND] |
Industrial datasets | [Industrial Machines Dataset] [Aachen Smart Factory] [HIPE] |
Synthetic Data | [SynD] [COLD] [FIRED] [SHED] [smartsim] |
-
Neural Load Disaggregation: Meta-Analysis, Federated Learning and Beyond [PDF] [Pytorch] [2023]
-
Non-Intrusive Load Monitoring: A Review [PDF] [Tensorflow] [2022]
-
NILM applications: Literature review of learning approaches, recent developments and challenges [PDF] [2022]
-
Review on Deep Neural Networks Applied to Low-Frequency NILM [PDF] [2021]
-
Energy Management Using Real-Time Non-Intrusive Load Monitoring [Arduino and Raspberry Pi-based]
-
flask-NILM-app-v1 [Flask]
-
Inverse Decomposition of Energy Consumption [C++]
Non-intrusive Appliance Load Monitoring (NALM) aims to recognize individual appliance usage from the main meter without indoor sensors. However, existing systems struggle to balance dataset construction efficiency and event/state recognition accuracy, especially for low-power appliance recognition. This paper introduces Hawk, an efficient and accurate NALM system that operates in two stages: dataset construction and event recognition. In the data construction stage, we efficiently collect a balanced and diverse dataset, HawkDATA, based on balanced Gray code and enable automatic data annotations via a sampling synchronization strategy called shared perceptible time. During the event recognition stage, our algorithm pipeline integrates steady-state differential pre-processing and voting-based post-processing for accurate event recognition from the aggregate current. Experimental results show that HawkDATA takes only 1/71.5 of the collection time to collect 6.34x more appliance state combinations than the baseline. In HawkDATA and a widely used dataset, Hawk achieves an average F1 score of 93.94% for state recognition and 97.07% for event recognition, which is a 47.98% and 11.57% increase over SOTA algorithms. Furthermore, selected appliance subsets and the model trained from HawkDATA are deployed in two real-world scenarios with many unknown background appliances. The average F1 scores of event recognition are 96.02% and 94.76%. [PDF] [Scikit] [SenSys 2024- Best AE Award]
Non-intrusive load monitoring (NILM) is able to analyze and predict users’ power consumption behaviors for further improving the power consumption efficiency of the grid. Neural network-based techniques have been developed for NILM. However, the dependencies of multiple appliances working simultaneously were ignored or implicitly characterized in their models for disaggregation. To improve the performance of NILM, we employ a graph structure to explicitly characterize the temporal dependencies among different appliances. Specially, we consider the prior temporal knowledge between the appliances in the working state, construct a weighted adjacency matrix to represent their dependencies. We also introduce hard dependencies of each appliance to prevent the sparsity of the weighted adjacency matrix. Furthermore, the non-sequential dependencies are learned among appliances using a graph attention network based on the weighted adjacency matrix. An encoder-decoder architecture based on dilated convolutions is developed for power estimation and state detection at the same time. We demonstrate the proposed model on the UKDALE dataset, which outperforms several state-of-the-art results for NILM. [PDF] [Pytorch] [PRCV 2023]
MSDC: Exploiting Multi-State Power Consumption in Non-intrusive Load Monitoring based on A Dual-CNN Model
Leveraging recent progress on deep learning techniques, we design a new neural NILM model Multi-State Dual CNN (MSDC). Different from previous models, MSDC explicitly extracts information about the appliance’s multiple states and state transitions, which in turn regulates the prediction of signals for appliances. More specifically, we employ a dual-CNN architecture: one CNN for outputting state distributions and the other for predicting the power of each state. A new technique is invented that utilizes conditional random fields (CRF) to capture state transitions. Experiments on two real-world datasets REDD and UK-DALE demonstrate that our model significantly outperform state-of-the-art models. [PDF ] [Pytorch] [2023] [AAAI]
This paper emphasizes federated neural NILM, receiving increasing attention due to its ability to preserve end-users’ privacy. Typically, by combining several locally trained models, federated learning has excellent potential to train NILM models locally without communicating sensitive data with cloud servers. [PDF] [Pytorch] [2023]
“I do not know”: Quantifying Uncertainty in Neural Network Based Approaches for Non-Intrusive Load Monitoring
In this paper, we explore the question - can we learn models that tell when they are unsure? Or, in other words, can we learn models that provide uncertainty estimates? We explore recent advances in uncertainty for NN(s), evaluate 14 model variants on the publicly available REDD dataset, and find that our models can accurately estimate uncertainty without compromising on traditional metrics. We also find that different appliances in their different states have varying performance of uncertainty. We also propose "recalibration" methods and find they can improve the uncertainty estimation. [PDF] [Jax] [2022]
In this paper, we propose a cost-effective collaborative learning framework, Fed-GBM (Federated Gradient Boosting Machines), consisting of two-stage voting and node-level parallelism, to address the problems in co-modelling for NILM. [PDF] [Scikit] [2022]
DeepDFML-NILM: A New CNN-Based Architecture for Detection, Feature Extraction and Multi-Label Classification in NILM Signals
In high-frequency NILM methods, state-of-the-art approaches, mainly based on deep learning solutions, do not provide a complete NILM architecture, including all the required steps. To overcome this gap, this work presents an integrated method for detection, feature extraction, and classification of high-frequency NILM signals for the publicly available LIT-Dataset. [PDF] [Tensorflow] [2022]
We treat three different thresholding methods to perform this task, discussing their differences on various devices from the UK-DALE dataset. [PDF] [Pytorch] [2022]
This paper presents an appliance classification method based on a Convolutional Recurrent Neural Network trained with weak supervision. [PDF] [Pytorch] [2022]
Utilizing transformer layers to accurately estimate the power signal of domestic appliances by relying entirely on attention mechanisms to extract global dependencies between the aggregate and the domestic appliance signals. [PDF] [Pytorch] [2022]
The research area of learning to learn (L2L) is focused on learning meta knowledge about machine learning and using it to improve the learning process. This work evaluates whether it is possible to apply L2L techniques, that learn to optimize neural networks to state-of-the art energy disaggregation algorithms and employ them to increase performance and transferability across datasets. We take a novel approach by applying L2L to energy disaggregation. [PDF] [Tensorflow] [2022] [master thesis]
The key elements of the method are a new neural network structure called TTRNet and a new loss function called MLFL. TTRNet is a multi-label classification model that can autonomously learn correlation information through its unique network structure. MLFL is a loss function specifically designed for multi-label classification tasks, which solves the imbalance problem and improves the monitoring accuracy for challenging loads. [PDF] [Pytorch] [2022]
We improve the generalization capability of the overall architecture by including an encoder–decoder with a tailored attention mechanism in the regression subnetwork. The attention mechanism is inspired by the temporal attention. [PDF] [Tensorflow] [2021]
In this paper we propose an energy disaggregation approach based on the variational autoencoders framework. The probabilistic encoder makes this approach an efficient model for encoding information relevant to the reconstruction of the target appliance consumption. [PDF] [Tensorflow] [2021]
Sequence to point learning based on bidirectional dilated residual network for non-intrusive load monitoring
We propose a sequence to point learning based on bidirectional dilated convolution for NILM on low-frequency data, called BitcnNILM. We compare our method with existing algorithms on low-frequency data via REDD and UK-DALE datasets. Experiments show that the superiority of our BitcnNILM in both load disaggregation and load on/off identification. [PDF] [Keras] [2021]
We propose BERT4NILM, an architecture based on bidirectional encoder representations from transformers (BERT) and an improved objective function designed specifically for NILM learning. We adapt the bidirectional transformer architecture to the field of energy disaggregation and follow the pattern of sequence-to-sequence learning. [PDF] [Pytorch] [2020]
Generative Adversarial Networks and Transfer Learning for Non-Intrusive Load Monitoring in Smart Grids
While many studies focus on achieving accurate models, few of them address the models generalizability. This paper proposes two approaches based on generative adversarial networks to achieve high-accuracy load disaggregation. Concurrently, the paper addresses the model generalizability in two ways, the first is by transfer learning by parameter sharing and the other is by learning compact common representations between source and target domains. This paper also quantitatively evaluate the worth of these transfer learning approaches based on the similarity between the source and target domains. [PDF] [Pytorch] [2020]
The main contribution presented in this paper is a comparison study between three common imaging techniques: Gramian Angular Fields, Markov Transition Fields, and Recurrence Plots. [PDF] [Tensorflow] [2020]
The proposed system leverages dimensionality reduction using Signal2Vec, is evaluated on two popular public datasets and outperforms another state-of-the-art multi-label NILM system. [PDF] [Scikit-learn] [2020]
Improved Appliance Classification in Non-Intrusive Load Monitoring Using Weighted Recurrence Graph and Convolutional Neural Networks
We propose an appliance recognition method utilizing the recurrence graph (RG) technique and CNNs. We introduce the weighted recurrent graph (WRG) generation that, given one cycle current and voltage, produces an image-like representation with more values than the binary output created by RG. [PDF] [Pytorch] [2020]
UNet-NILM: A Deep Neural Network for Multi-tasks Appliances State Detection and Power Estimation in NILM
We propose UNet-NILM for multi-task appliances' state detection and power estimation, applying a multi-label learning strategy and multi-target quantile regression. [PDF] [Official | Pytorch] [Reimplement | Pytorch] [2020]
We address the problem through the recognition of the state of activation of the appliances using a fully convolutional deep neural network, borrowing some techniques used in the semantic segmentation of images and multilabel classification. [PDF] [Pytorch] [2020]
Multi-label Learning for Appliances Recognition in NILM using Fryze-Current Decomposition and Convolutional Neural Network.
This paper proposes the CNN-based multi-label learning approach, which links multiple loads to an observed aggregate current signal. Apply the Fryze power theory to decompose the current features into active and non-active components and use the Euclidean distance similarity function to transform the decomposed current into an image-like representation which is used as input to the CNN. [PDF] [Pytorch] [2020]
We study different neural network compression schemes and their efficacy on the state-of-the-art neural network NILM method. We additionally propose a multi-task learning-based architecture to compress models further. [PDF] [Pytorch] [2020]
In this context, the ideal NILM algorithm should be unsupervised and provide real-time consumption feedback. Solutions with these characteristics have not been sufficiently studied in the NILM bibliography. This paper introduces an attempt to address this problem, by suggesting the use of a “Live” algorithm that provides real-time feedback. The Live algorithm runs in parallel with a more robust NILM procedure, and a periodic synchronization step ensures that the two processes share the same appliance models. [PDF] [Scikit] [2020]
We investigate the application of the recently developed WaveNet models for the task of energy disaggregation. [PDF] [Pytorch] [2019]
We present a causal 1-D convolutional neural network inspired by WaveNet for NILM on low-frequency data. We also study using various components of the complex power signal for NILM, and demonstrate that using all four components available in a popular NILM dataset (current, active power, reactive power, and apparent power). [PDF] [Keras] [2019]
Appliance transfer learning (ATL) and cross-domain transfer learning (CTL). [PDF] [Tensorflow] [2019]
We propose two recurrent network architectures that use sliding window for real-time energy disaggregation. [PDF] [Keras] [2018]
We propose a subtask gated network that combines the main regression network with an on/off classification subtask network. [PDF] [Pytorch] [2018]
We propose sequence-to-point learning, where the input is a window of the mains and the output is a single point of the target appliance. [PDF] [Tensorflow] [2017] [Reimplement-Pytorch]
We adapt three DNN architectures to energy disaggregation: 1) a form of RNN called LSTM; 2) denoising autoencoders; and 3) a network which regresses the start time, end time and average power demand of each appliance activation. [PDF] [Theano] [2015]