TEXT SIZE

search for



CrossRef (0)
Functional hierarchical clustering using shape distance
Communications for Statistical Applications and Methods 2024;31:601-612
Published online September 30, 2024
© 2024 Korean Statistical Society.

Kyungmin Ahn1,a

aDepartment of Statistics, Keimyung University, Korea
Correspondence to: 1 Department of Statistics, Keimyung University, 1095 Dalgubeol-daero, Dalseo-gu, Daegu 42601, Korea. E-mail: ahn@kmu.ac.kr
This research was supported by the Bisa Research Grant of Keimyung University and the National Research Foundation of Korea (NRF) grant funded by the Korea government (MSIT) (RS-2022-00165598).
Received July 3, 2024; Revised August 2, 2024; Accepted August 3, 2024.
 Abstract
A functional clustering analysis is a crucial machine learning technique in functional data analysis. Many functional clustering methods have been developed to enhance clustering performance. Moreover, due to the phase variability between functions, elastic functional clustering methods, such as applying the Fisher-Rao metric, which can manage phase variation during clustering, have been developed to improve model performance. However, aligning functions without considering the phase variation can distort functional information because phase variation can be a natural characteristic of functions. Hence, we propose a state-of-the-art functional hierarchical clustering that can manage phase and amplitude variations of functional data. This approach is based on the phase and amplitude separation method using the norm-preserving time warping of functions. Due to its invariance property, this representation provides robust variability for phase and amplitude components of functions and improves clustering performance compared to conventional functional hierarchical clustering models. We demonstrate this framework using simulated and real data.
Keywords : amplitude variation, functional data analysis, functional hierarchical clustering, phase variation, shape distance, SRVF
1. Introduction

Functional clustering analysis is a crucial machine learning technique that assigns functional observations to several groups in functional data analysis (FDA) (Ramsay and Silverman, 2005). The literature has proposed many methods of functional clustering (Bouveyron and Jacques, 2011; Bouveyron et al., 2015; Febrero-Bande and Oviedo de la Fuente, 2012; Marron et al., 2014). A function comprises two components: The phase (horizontal or x-axis) component and amplitude (vertical or y-axis) component. Therefore, verifying whether phase or amplitude variations derive from the natural characteristics of functional objectives or are unknown noise is critical for FDA. If variations are noise, then the noise must be removed using statistical methods. For example, amplitude variations can be minimized using a smoothing technique employing a basis representation, such as the B-spline basis function in a parametric way or the kernel function in a nonparametric way. For phase variation, noise can be removed using a functional statistical method, called a registration or alignment method. Many registration methods have been proposed, and currently, the Fisher-Rao registration by Srivastava et al. (2011) is one of the best alignment methods.

Therefore, many researchers are interested in combining registration and statistical methods to improve the performance of the statistical results. Some papers have been developing joint methods to address phase variation between curves while implementing a statistical method. For example, Ahn et al. (2018, 2020) proposed an elastic functional predictor model by reducing phase variation while estimating regression coefficients. Xu et al. (2022) proposed a Fisher-Rao registration algorithm to manage phase variation effectively for neural spike data. However, for the statistical analysis, these methods focus on minimizing or removing phase variation assumed to be noise. This strong assumption can be beneficial if the phase variation is indeed noise; if not, this strong conjecture could lead to incorrect statistical results by distorting the natural characteristics of functional data.

To solve this problem, we propose a state-of-the-art method, a functional hierarchical agglomerative clustering method using shape distances (FHACS), which can address phase and amplitude variations for functional hierarchical clustering analysis. This approach is motivated by the phase-amplitude separation algorithm (Srivastava et al., 2011), where invariant metrics, such as the Fisher-Rao metric and the elastic Riemannian metric for functional data in FDA, are employed. One of the substantial advantages of this separation method is that it avoids the pinching effect, where the function can be pinched while aligning in the space.

The remainder of this paper is organized as follows. Section 2 describes the FHACS under the Fisher-Rao invariant metric and summarizes the concepts used for the clustering method. We demonstrate this method using several simulated datasets and compare its performance against three alternative methods: 1) the standard functional hierarchical clustering model, 2) the pre-aligned functional hierarchical clustering method, and 3) the elastic functional hierarchical clustering model that functions are aligned while computing the distance between two functions in Section 3. Then, we apply the proposed and three candidate methods to six real datasets to compare the effectiveness and clustering performance in Section 4. Last, Section 5 presents the concluding remarks and limitations of this study.

2. Methods

To introduce FHACS, we first summarized the standard functional hierarchical clustering model and some critical concepts for developing the functional clustering model.

2.1. Standard functional hierarchical clustering

The functional hierarchical clustering analysis is the most basic, simple clustering algorithm, which determines the distances between a pair of functional objectives and verifies which cluster should be combined or divided depending on the strategies. We use an agglomerative strategy for the bottom-up approach and a divisive strategy for the top-down approach. There are two critical points to consider when implementing this hierarchical clustering analysis: distance and linkage. Distance is nearly fixed for the FDA. We defined observations as functions in the space, where the inner product of two functions is defined as 〈f, g〉 = ∫ f g dt in the Hilbert space, ℋ. Thus, we used the distance to compute the distance between the functions on the time interval, T, expressed as follows:

d(f1,f2)=T(f1(t)-f2(t))2dt,

where f1 and f2 are the functional observations in the space.

For the linkage method, we employed several selection methods for the hierarchical clustering analysis. There are many possible linkages, and some examples are listed below.

  • Single linkage: The distance between two clusters is the minimum distance between members of the two clusters.

  • Centroid linkage: The distance between two clusters is the distance between their centroids.

  • Complete linkage: The distance between two clusters is the maximum distance between members of the two clusters.

  • Average linkage: The distance between two clusters is the average of all distances between members of the two clusters.

Each linkage method has advantages and disadvantages; thus, no best universal linkage exists for all situations. For example, the single linkage method can manage nonelliptical shapes and is best for capturing clusters of assorted sizes, but it has a problem with chaining. The complete, centroid, and average linkage methods could group clusters when noise exists between clusters, but crowding, dendrogram inversion, and sensitivity to a monotone transformation can occur, respectively.

2.2. Shape distance

Srivastava et al. (2011) proposed a method to decompose functions into phase and amplitude components, which involves the functional registration algorithm in which phase variations can be removed by aligning functions to any function template. The template function can be a random sample function (generally, the pairwise case) in the dataset or the Karcher or Fréchet mean of the functions (generally, the groupwise case). After registration, the functions are split into phase and amplitude components.

To implement this method, we defined functional observation f as a real-valued function on the interval [0, 1] and let ℱ denote the set of all such functions. Then, we defined the time-warping function, γ, used for computing the phase distance. We let Γ is the set of all boundary-preserving map between manifolds which is differentiable and has a differentiable inverse of the unit interval [0, 1] (i.e., Γ = {γ : [0, 1] → [0, 1] | γ(0) = 0, γ(1) = 1, γ is a diffeomorphism}). For alignment or registration, Srivastava et al. (2011) established a mathematical expression representing a function using its square root velocity function (SRVF), where q denotes the SRVF of the original function, q : [0, 1] → ℝ, expressed as follows:

q(t)=sign (f˙(t))|f˙(t)|,

where denotes the first derivative of the function f . The greatest advantage is that this framework streamlines the complicated Fisher-Rao Riemannian metric and Fisher-Rao distance into the inner product and norm, respectively. Then, a function f can be warped by a time-warping function, γ: f ↦ ( fγ). Hence, its SRVF changes by q(qγ)γ˙. This transformation is denoted by (q*γ)=(qγ)γ˙ (Ahn et al., 2020), which implies the invariance property of the Fisher-Rao metric, that for any and γ ∈ Γ, we have ||(q1 * γ) − (q2 * γ)|| = ||q1q2||. Hence, the action of Γ on is by isometrics. A special case of this equation is ||(q * γ)|| = ||q|| for all q and γ. Therefore, this action preserves the norm of the SRVF and avoids pinching effects in the registration process.

Two registration problems depend on the number of functional observations: The pairwise alignment problem for two observations and the groupwise or multiple alignment problem for multiple observations (Srivastava and Klassen, 2016). Functional hierarchical clustering analysis requires only the distance between two functions, so only pairwise alignment is necessary. Hence, we summarize the pairwise registration problem.

To align functions, the optimal time-warping function γ* is determined for any two given functions f1, f2 ∈ ℱ, such that the equation can be expressed as follows:

γ*=argminγΓf1-f2γ2.

Then, for any t ∈ [0, 1], the value f1(t) is registered to f2(γ*(t)). In here, many potential problems can be raised such as the invariance problem and pinching effect of using equation 2.3 in the space. Thus, we applied SRVF, q, rather than f to align the functions successfully to solve any potential problems. Then, equation 2.3 can be changed as follows:

γ*=argminγΛq1-(q2*γ)2,

where q1 and q2 are the SRVFs of function f1 and f2, respectively. We employed equation 2.4 to compute the shape distances for FHACS.

2.2.1. Amplitude distance

Next, we defined the amplitude distance. For any two functions f1, f2 ∈ ℱ and the corresponding SRVFs, , we defined the amplitude distance da (Srivastava and Klassen, 2016; Tucker et al., 2013; Ahn et al., 2020) as follows:

da(f1,f2)=infγΓ(q1-(q2*γ)).

In here, the distance da is a proper distance with three properties: symmetry, positive definiteness, and triangle inequality (for more details, see Srivastava and Klassen (2016)).

2.2.2. Phase distance

We let γ1, γ2, . . . , γn ∈ Γ represent the set of observed warping functions. However, computing the phase distance in a standard way has two central problems : 1) the infinite dimensionality and 2) Γ is a nonlinear manifold (Tucker et al., 2013). To solve these issues, similar to the definition of SRVF, we need to represent an element of γ ∈ Γ using the square root of its derivative, γ˙ that this representation simplifies the complicated geometry of Γ to a unit sphere. That is, γ˙2=01γ˙(t)dt=γ(1)-γ(0). Hence, the phase distance for the two functions is as follows:

dp(f1,f2)=cos-101γ˙*(t)dt,

where γ*(t) is the optimal time-warping function from equation 2.4. This definition is based on the Fisher–Rao distance using SRVFs (Srivastava and Klassen, 2016; Tucker et al., 2013; Ahn et al., 2020), following the properties below:

  • dp( f1, f2) = dp( f2, f1);

  • dp( f1, f2) = dp( f1, a f2 + b) for any a ∈ ℝ+ and b ∈ ℝ;

  • dp( f1, f3) ≥ dp( f1, f2) + dp( f2, f3) for any f1, f2, f3 ∈ ℱ.

2.2.3. Shape distance, combining the phase and amplitude distances

Based on two distances, as discussed, we defined the shape distance ds by combining two distances, the phase and amplitude distances, as presented below:

ds(f1,f2)=λda(f1,f2)+(1-λ)dp(f1,f2),

where λ ∈ [0, 1] is a proportion tuning parameter. We can control both distances by weighting each distance; for example, if λ = 0, the shape distance is focused on the phase distance dp(·). If λ = 1, the shape distance can be simplified as the amplitude distance da(·). Moreover, if λ = 0.5, the shape distance is the average of two distances: (1/2)(da(·) + dp(·)). For FHACS, this shape distance is applied instead of the distance to increase the efficiency and performance of the clustering analysis by addressing the phase and amplitude variations.

3. Simulation studies

3.1. Simulation setting

To assess the performance of the functional hierarchical clustering algorithm using the shape distance, we evaluated the model on simulated data constructed using the following:

fij0(t)=aij12πσj2exp (-(t-μj)22σj2),         i=1,,20,   j=1,,K,

where μj and σj denote the mean and standard deviation for each j group, respectively. In addition, ai j represents the parameter for the amplitude variation and t ∈ [0, 1]. Using equation 3.1, we generated three simulated datasets, where each data point reflects the situation of generating clusters: 1) two groups/clusters (K = 2) are based on the amplitude (μj ∈ {0.5, 0.5}, σj ∈ {0.1, 0.11}, and ai j = 1 for all i, j), 2) two groups (K = 2) are based on the phase (μj ∈ {0.35, 0.65}, σj ∈ {0.1, 0.1}, and ai j = 1 for all i, j), and 3) three groups (K = 3) are based on both the phase and amplitude (μj ∈ {0.4, 0.5, 0.6} and σj ∈ {0.1, 0.11, 0.13} with ai j ~ N(1, 0.12)). Then, 20 functions are generated for each group j. Therefore, the 40, 40, and 60 functions are generated for simulated Datasets 1, 2, and 3, respectively. These given functions are perturbed using random time-warping functions {γi j} to obtain the observed functions (fij=fij0γij).

Figure 1 presents the resulting simulated data using equation 3.1. Figure 1(a), Figure 1(c), and Figure 1(e) show the simulated data that were perturbed by corresponding time-warping functions in Figure 1(b), Figure 1(d), and Fig. 1f, respectively. We note that the warped functions of simulated datasets have values of zero at t = 0 and t = 1. Since the functions are warped by the equation, fi j(t) = fij0(γij(t)) where γi j(0) = 0 and γi j(1) = 1, the warped functions also should be fi j(0) = 0 and fi j(1) = 0. Figure 1(a) displays simulated Dataset 1 where the true clusters are based on an amplitude with two clusters. The functions have phase variations that are noise. In addition, Figure 1(c) displays simulated Dataset 2, where the true clusters are based on a phase with two clusters. Unlike simulated Dataset 1, functions have phase variation as noise, but two clusters are also distinguished by various phases. Therefore, removing the phase variations without consideration might provide incorrect results. Last, Figure 1(e) presents three groups where the true clusters are based on the phase and amplitude, which is a more complicated case when functions have both phase and amplitude variations as noise.

To demonstrate the effectiveness and clustering performance of the model, we compared FHACS with three natural alternatives. These models are either commonly used in the literature or are modifications of current models addressing phase variability in functional data. These models are the standard functional hierarchical agglomerative clustering model (SFHAC), the pre-aligned functional hierarchical agglomerative clustering model (PAFHAC), and the elastic functional hierarchical agglomerative clustering model (EFHAC). We summarize these three clustering models.

  • SFHAC: This standard hierarchical agglomerative clustering (HAC) algorithm uses the distance instead of the Euclidean distance. The functions are defined in the space; thus, the metric is applied for the statistical clustering analysis. Hence, equation 2.1 calculates the distance between functional observations.

  • PAFHAC: In this model, functional observations are pre-aligned (using a phase-amplitude separation algorithm) (for more details, refer to Srivastava and Klassen (2016)), and then applied the standard hierarchical clustering model. Phase variations are removed by aligning the functions concerning their Karcher mean; hence, this is a registration problem for the groupwise case.

  • EFHAC: In this model, functions are aligned pairwise while computing the distance between two functions. The concept of the model is similar to that of FHACS, but this model only focuses on the amplitude distance between two functions. The difference between PAFHAC and EFHAC is that PAFHAC computes the Karcher mean of the functions, registers all functions concerning the mean, and applies SFHAC. However, EFHAC computes the amplitude distance while removing the phase distance between two functions each time. Hence, the clustering results should be the same as when λ of FHACS equals 1 in equation 2.7 (i.e., ds( f1, f2) = da( f1, f2)).

We applied FHACS and the other three alternative clustering algorithms to compare the accuracy of clustering performance. These models can easily be implemented using the existing R packages from R software. We used the fdacluster (Stamm, 2024) R package for SFHAC and EFHAC. For PAFHAC, we aligned multiple functions based on the phase-amplitude separation algorithm using the fdasrvf R package (Tucker, 2024). We employed a dynamic programming algorithm to determine the optimal time-warping functions, {γij*}. For FHACS, we also used the fdasrvf R package (Tucker, 2024) to compute the phase and amplitude distances. Then, we used equation 2.7 to compute the shape distance. We fixed all models to the HAC algorithm with complete linkage for consistency.

3.2. Clustering performance accuracy

To compare the clustering performance as a measurement, we calculated the Hubert–Arabie adjusted random index (ARI), which is the corrected-for-chance version of the rand index (Rand, 1971; Hubert and Arabie, 1985; Vinh et al., 2009), and is computed below:

ARI=Index-Expected IndexMax Index-Expected Index,

where the independent clusters have an expected index of zero and identical partitions have an ARI equal to 1.

For the simulation experiments in the clustering analysis, we directly specified the true number of clusters K (K = 2 for simulated Datasets 1 and 2, and K = 3 for simulated Dataset 3) and evaluated the accuracy of the clustering results. We used the mclust (Scrucca et al., 2016) R package for the ARI calculation.

3.3. Simulation results

Tables 1 and 2 list the clustering results for the simulation studies. For Table 1, the ARI is computed for each λ to quantify the shape distance in equation 2.7. Based on the results of the tables, λ can measure the weight between the amplitude and phase distances properly for the simulated data. For example, clusters in simulated Dataset 1 were generated based on the amplitude; thus, phase variations are true noise in this case. The model yields the best clustering results when λ is increasing; therefore, the parameter weight is focused on the amplitude distance in equation 2.7.

Similarly, PAFHAC and EFHAC also performed well because these two models are clustering models that focus on the amplitude distance after removing phase variation. The groups from simulated Dataset 2 were generated based on two distinct phases, and the results reveal that parameter λ = 0 for FHACS has the best clustering performance. Unlike the performance of PAFHAC and EFHAC for simulated Dataset 1, the accuracy of the two models was extremely poor because the model only measures the amplitude distance. When the groups are based on phase and amplitude, as in simulated Dataset 3, the tuning parameter λ was set at 0.3 for the best ARI. These simulation results demonstrate that the parameter λ can adequately adjust the weight of two distances. That is, λ plays a critical role in weighting amplitude and phase distances to improve clustering performance.

4. Application to real data

4.1. Description of real data

Functional data have phase and amplitude variability in several crucial application areas, such as biology, human anatomy, biochemistry, images, and sensors. This section presents the results on multiple real datasets for clustering methods using six examples from the University of California at Riverside time-series classification archive (Chen et al., 2015, July). This database has information, including the actual number of clusters and time points. Table 3 describes the data (for more details, refer to Chen et al. (2015, July)).

Figure 2 presents the functional data for each real dataset: Gun Point (GP; Figure 2(a)), Toe Segmentation (TS; Figure 2(b)), Face Four (FF; Figure 2(c)), Cylinder-Bell-Funnel (CBF; Figure 2(d)), Italy Power Demand (IPD; Figure 2(e)), and Plane (Figure 2(f)). Each figure is colored to distinguish the true number of clusters (2, 2, 4, 3, 2, and 7 for GP, TS, FF, CBF, IPD, and Plane, respectively). From the figures, the phase and amplitude variability values are naturally found in the data, as the observations are at different times for each measurement during collection.

4.2. Numerical results

Next, we applied FHACS and the other three alternative hierarchical clustering methods to these real datasets. Similar to the simulation study, for FHACS, we set λ = {0, 0.1, 0.2, . . . , 0.9, 1} to determine the optimal tuning parameter for measuring the weight between the phase and amplitude distances. Table 4 lists the clustering performance results for FHACS, and Table 5 lists the results for the three alternative models.

In Tables 4 and 5, FHACS provides better clustering performance than the other three competing models. The clustering results in tables imply that the phase distance is also critical to classify groups of functional data because most λ values are near 0 instead of 1. The clustering performance of TS, FF, and Plane datasets especially outperformed when λ is equal to 0. Therefore, functional clustering models such as EFHAC and PAFHAC that remove only phase variation without considering functional data perform poorly on real data, implying that in many possible existing real datasets, phase variation is from natural characteristics of functional observations. That is, FHACS performs better when focused on both phase and amplitude distances. Like PAFHAC and EFHAC, focusing only on amplitude distance when removing phase variation produces poor results because this method ignores phase variation while analyzing functional data.

5. Concluding remarks and limitations

The functional clustering algorithm was developed for the FDA to improve clustering performance. Many functional data have phase variability due to the lack of temporal synchronization across measurements. Therefore, awareness has been rising regarding phase variability removal to improve model clustering performance. However, removing or minimizing only the phase variation can be problematic because functional data can also involve phase characteristics. Hence, we proposed a state-of-the-art functional hierarchical clustering method that can measure both the phase and amplitude distances, defined as the shape distance. We applied this hierarchical clustering model to three simulated datasets to investigate clustering performance in different functional data situations, and we applied it to six real datasets. The proposed method outperforms other alternative three functional hierarchical clustering models. Moreover, this definition of the shape distance can be applied to other clustering or classification methods, such as k-means and k-nearest neighbors, to address the phase and amplitude components of functional data.

This study has some limitations. First, functional or multivariate standard hierarchical clustering suffers from choosing the optimal number of clusters, K, unlike other clustering algorithms, such as k-means. Thus, it is challenging to determine which K is the optimal cluster for the hierarchical clustering analysis. Moreover, we added a new tuning parameter, λ, to weight the phase and amplitude distances; therefore, determining the optimal K and λ can be more complicated. However, this can be solved with domain knowledge for the functional data, for example, with the actual number of clusters. Selecting a specific λ to improve the clustering performance gradually is another computational problem. In this study, we set only one λ for one clustering analysis. The functional hierarchical clustering algorithm computes the distance between two objectives; hence, each pair of functions is likely to have a different optimal λ value. Clustering performance can be improved further if each λl is set to l={1,2,,(n2)} for n functional observations. However, this is nearly impossible due to the computational costs.

Figures
Fig. 1. Plots for simulated data. The functions are generated using equation 3.1 and then perturbed using random time-warping functions {γi j} for each simulated dataset, i.e., (b), (d), (f) for simulated Dataset 1, 2, and 3, respectively. The clusters in each dataset are distinguished by color to mark the true number of clusters. For example, simulated Datasets 1 and 2 have two clusters, whereas simulated Dataset 3 has three clusters.
Fig. 2. Plots for real data. Each figure is colored differently to distinguish the true number of clusters.
TABLES

Table 1

FHACS results for each simulated dataset

λ00.10.20.30.40.50.60.70.80.91
Sim. Dataset 10.030.0310.03111111111
Sim. Dataset 2110.030.030.030.030.150.150.110.110.03
Sim. Dataset 30.240.130.150.570.350.260.260.300.290.260.27

ARI is computed for each λ, and bold font marks the highest ARI.


Table 2

Three alternative clustering results for each simulated dataset

SFHACPAFHACEFHAC
Sim. Dataset 1−0.0111
Sim. Dataset 20.030.010.03
Sim. Dataset 30.230.110.27

Bold font marks the highest ARI.


Table 3

Description of the real data

Data# of obs.,Time pointsClassDonor
Gun Point (GP)501502A. Ratanamahatana & E. Keogh
Toe Segmentation (TS)402702Tony Bagnall
Face Four (FF)243504A. Ratanamahatana & E. Keogh
Cylinder-Bell-Funnel (CBF)301283N. Saito
Italy Power Demand (IPD)27642J.J. van Wijk & E. Keogh & L. Wi
Plane1051447J. Gao

Table 4

Clustering performance results for the real data

λ00.10.20.30.40.50.60.70.80.91
GP0.0010.1790.047−0.0030.0030.0030.003−0.003−0.003−0.003−0.003
TS0.1090.1000.0150.0150.0150.0150.0150.0150.0150.0150.015
FF0.6590.5140.5140.5140.5140.4440.1790.1790.2130.2400.240
CBF0.0280.0590.0200.0590.0910.0590.2410.1100.1100.1450.145
IPD0.1100.1100.1100.5730.110−0.0080.1100.0040.0040.004−0.008
Plane111111110.8060.8060.806

Bold represents the highest ARI among the four clustering algorithms.


Table 5

Results for real data using the three alternative clustering algorithms

SFHACPAFHACEFHAC
GP0.0110.029−0.003
TS−0.0180.0150.015
FF0.369−0.0210.240
CBF0.1370.1960.145
IPD0.031−0.008−0.008
Plane0.6230.7010.806

References
  1. Ahn K, Tucker JD, Wu W, and Srivastava A (2018). Elastic handling of predictor phase in functional regression models, In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition Workshops, Salt Lake City, UT, 324–331.
  2. Ahn K, Tucker JD, Wu W, and Srivastava A (2020). Regression models using shapes of functions as predictors. Computational Statistics & Data Analysis, 151, 107017.
    CrossRef
  3. Bouveyron C, Côme E, and Jacques J (2015). The discriminative functional mixture model for a comparative analysis of bike sharing systems. The Annals of Applied Statistics, 9, 1726-1760.
    CrossRef
  4. Bouveyron C and Jacques J (2011). Model-based clustering of time series in group-specific functional subspaces. Advances in Data Analysis and Classification, 5, 281-300.
    CrossRef
  5. Chen Y, Keogh E, Hu B, Begum N, Bagnall A, Mueen A, and Batista G (2015, July). The UCR Time Series Classification Archive,
  6. Febrero-Bande M and Oviedo de la Fuente M (2012). Statistical computing in functional data analysis: The R package fda. usc, Journal of Statistical Software, 51, 1-28.
    CrossRef
  7. Hubert L and Arabie P (1985). Comparing partitions. Journal of Classification, 2, 193-218.
    CrossRef
  8. Marron JS, Ramsay JO, Sangalli LM, and Srivastava A (2014). Statistics of time warpings and phase variations. Electronic Journal of Statistics, 8, 1697-1702.
    CrossRef
  9. Ramsay JO and Silverman BW (2005). Functional Data Analysis (2nd ed), Springer, Berlin.
    CrossRef
  10. Rand WM (1971). Objective criteria for the evaluation of clustering methods. Journal of the American Statistical Association, 66, 846-850.
    CrossRef
  11. Scrucca L, Fop M, Murphy TB, and Raftery AE (2016). mclust 5: Clustering, classification and density estimation using Gaussian finite mixture models. The R Journal, 8, 289.
    Pubmed KoreaMed CrossRef
  12. Srivastava A and Klassen E (2016). Functional and Shape Data Analysis, Springer, Berlin.
  13. Srivastava A, Wu W, Kurtek S, Klassen E, and Marron JS (2011). Registration of functional data using Fisher-Rao metric,
    Available from: arXiv.1103.3817
  14. Stamm A (2024). fdacluster: Joint Clustering and Alignment of Functional Data, R package version 0.3.0.9000,
  15. Tucker JD (2024). fdasrvf: Elastic Functional Data Analysis. R package version 2.2.0,
  16. Tucker JD, Wu W, and Srivastava A (2013). Generative models for functional data using phase and amplitude separation. Computational Statistics & Data Analysis, 61, 50-66.
    CrossRef
  17. Vinh NX, Epps J, and Bailey J (2009). Information theoretic measures for clusterings comparison: Is a correction for chance necessary?. In Proceedings of the 26th Annual International Conference on Machine Learning, 1073-1080.
    CrossRef
  18. Xu Z, Zhou X, Xu Y, and Wu W (2022). Removing nonlinear misalignment in neuronal spike trains using the Fisher-Rao registration framework. Journal of Neuroscience Methods, 367, 109436.
    CrossRef