Agglomerative Hierarchical Clustering with Dynamic Time Warping for
Household Load Curve Clustering
- URL: http://arxiv.org/abs/2210.09523v1
- Date: Tue, 18 Oct 2022 01:30:25 GMT
- Title: Agglomerative Hierarchical Clustering with Dynamic Time Warping for
Household Load Curve Clustering
- Authors: Fadi AlMahamid, Katarina Grolinger
- Abstract summary: Classifying clients according to their consumption patterns enables targeting specific groups of consumers for demand response (DR) programs.
We propose a shape-based approach that combines Agglomerative Hierarchical Clustering (AHC) with Dynamic Time Warping (DTW)
We show that AHC using DTW outperformed other clustering algorithms and needed fewer clusters.
- Score: 0.0
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Energy companies often implement various demand response (DR) programs to
better match electricity demand and supply by offering the consumers incentives
to reduce their demand during critical periods. Classifying clients according
to their consumption patterns enables targeting specific groups of consumers
for DR. Traditional clustering algorithms use standard distance measurement to
find the distance between two points. The results produced by clustering
algorithms such as K-means, K-medoids, and Gaussian Mixture Models depend on
the clustering parameters or initial clusters. In contrast, our methodology
uses a shape-based approach that combines Agglomerative Hierarchical Clustering
(AHC) with Dynamic Time Warping (DTW) to classify residential households' daily
load curves based on their consumption patterns. While DTW seeks the optimal
alignment between two load curves, AHC provides a realistic initial clusters
center. In this paper, we compare the results with other clustering algorithms
such as K-means, K-medoids, and GMM using different distance measures, and we
show that AHC using DTW outperformed other clustering algorithms and needed
fewer clusters.
Related papers
- Self-Supervised Graph Embedding Clustering [70.36328717683297]
K-means one-step dimensionality reduction clustering method has made some progress in addressing the curse of dimensionality in clustering tasks.
We propose a unified framework that integrates manifold learning with K-means, resulting in the self-supervised graph embedding framework.
arXiv Detail & Related papers (2024-09-24T08:59:51Z) - A3S: A General Active Clustering Method with Pairwise Constraints [66.74627463101837]
A3S features strategic active clustering adjustment on the initial cluster result, which is obtained by an adaptive clustering algorithm.
In extensive experiments across diverse real-world datasets, A3S achieves desired results with significantly fewer human queries.
arXiv Detail & Related papers (2024-07-14T13:37:03Z) - A Hybrid SOM and K-means Model for Time Series Energy Consumption
Clustering [0.0]
This paper introduces a novel approach to effectively cluster monthly energy consumption patterns by integrating two powerful techniques: Self-organizing maps and K-means clustering.
The main focus of this study is on a selection of time series energy consumption data from the Smart meters in London dataset.
arXiv Detail & Related papers (2023-11-25T16:55:19Z) - A Machine Learning-Based Framework for Clustering Residential
Electricity Load Profiles to Enhance Demand Response Programs [0.0]
We present a novel machine learning based framework in order to achieve optimal load profiling through a real case study.
In this paper, we present a novel machine learning based framework in order to achieve optimal load profiling through a real case study.
arXiv Detail & Related papers (2023-10-31T11:23:26Z) - Targeted demand response for flexible energy communities using
clustering techniques [2.572906392867547]
The goal is to alter the consumption behavior of the prosumers within a distributed energy community in Italy.
Three popular machine learning algorithms are employed, namely k-means, k-medoids and agglomerative clustering.
We evaluate the methods using multiple metrics including a novel metric proposed within this study, namely peak performance score (PPS)
arXiv Detail & Related papers (2023-03-01T02:29:30Z) - Clustering of Bank Customers using LSTM-based encoder-decoder and
Dynamic Time Warping [0.0]
Clustering is an unsupervised data mining technique that can be employed to segment customers.
The present study uses a real-world financial dataset to cluster bank customers by an encoder-decoder network and the dynamic time warping (DTW) method.
arXiv Detail & Related papers (2021-10-22T13:16:49Z) - SOMTimeS: Self Organizing Maps for Time Series Clustering and its
Application to Serious Illness Conversations [3.2689702143620147]
We present a new DTW-based clustering method called SOMTimeS (a Self-Organizing Map for TIME Series)
It scales better and runs faster than other DTW-based clustering algorithms, and has similar performance accuracy.
We applied SOMtimeS to natural language conversation data collected as part of a large healthcare cohort study.
arXiv Detail & Related papers (2021-08-26T00:18:25Z) - (k, l)-Medians Clustering of Trajectories Using Continuous Dynamic Time
Warping [57.316437798033974]
In this work we consider the problem of center-based clustering of trajectories.
We propose the usage of a continuous version of DTW as distance measure, which we call continuous dynamic time warping (CDTW)
We show a practical way to compute a center from a set of trajectories and subsequently iteratively improve it.
arXiv Detail & Related papers (2020-12-01T13:17:27Z) - Scalable Hierarchical Agglomerative Clustering [65.66407726145619]
Existing scalable hierarchical clustering methods sacrifice quality for speed.
We present a scalable, agglomerative method for hierarchical clustering that does not sacrifice quality and scales to billions of data points.
arXiv Detail & Related papers (2020-10-22T15:58:35Z) - Decorrelated Clustering with Data Selection Bias [55.91842043124102]
We propose a novel Decorrelation regularized K-Means algorithm (DCKM) for clustering with data selection bias.
Our DCKM algorithm achieves significant performance gains, indicating the necessity of removing unexpected feature correlations induced by selection bias.
arXiv Detail & Related papers (2020-06-29T08:55:50Z) - LSD-C: Linearly Separable Deep Clusters [145.89790963544314]
We present LSD-C, a novel method to identify clusters in an unlabeled dataset.
Our method draws inspiration from recent semi-supervised learning practice and proposes to combine our clustering algorithm with self-supervised pretraining and strong data augmentation.
We show that our approach significantly outperforms competitors on popular public image benchmarks including CIFAR 10/100, STL 10 and MNIST, as well as the document classification dataset Reuters 10K.
arXiv Detail & Related papers (2020-06-17T17:58:10Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.