 Research
 Open Access
 Published:
Pattern recognition in data as a diagnosis tool
Journal of Mathematics in Industry volume 12, Article number: 3 (2022)
Abstract
Medical data often appear in the form of numerical matrices or sequences. We develop mathematical tools for automatic screening of such data in two medical contexts: diagnosis of systemic lupus erythematosus (SLE) patients and identification of cardiac abnormalities. The idea is first to implement adequate data normalizations and then identify suitable hyperparameters and distances to classify relevant patterns. To this purpose, we discuss the applicability of PlackettLuce models for rankings to hyperparameter and distance selection. Our tests suggest that, while Hamming distances seem to be well adapted to the study of patterns in matrices representing data from laboratory tests, dynamic time warping distances provide robust tools for the study of cardiac signals. The techniques developed here may set a basis for automatic screening of medical information based on pattern comparison.
Content
One of the most challenging and essential tasks performed by the healthcare professionals is diagnosing a disease or patient’s condition. Medical diagnoses are based on symptoms, signs, medical history and complementary examinations, which define the patient’s clinical picture. This information is collected in datasets combining different kinds of data. Laboratory analyses, for instance, are stored as matrices of numerical values, representing blood counts, metabolic, ionic, enzyme, hormone, protein, vitamin and antibody tests, or other variables of interest recorded at different times. Specific vital signals are recorded as time sequences, such is the case of electrocardiograms, for example. More sophisticated medical imaging devices (Xrays, magnetic resonance imaging, etc) visualize the status of organs or body parts by means of a series of images. As the amount of data collected grows in size, the development of algorithms allowing medical staff to automatically screen the information contained in the stored data becomes essential. This task faces additional challenges, since hospital records usually display non homogeneous data measured over irregular time periods.
The applicability of machine learning techniques in specific medical contexts involving large data amounts is an active area of research nowadays. Neural networks, for instance, are sometimes used for imagebased diagnosis [1, 2], while unsupervised and supervised classification techniques [3] are nowadays exploited to study the role of genes in sickness [4, 5] and to investigate the response to treatment [6]. However, the amount of data available in many medical situations is scarce [7–9]. In the search for a diagnosis, one can resort to different procedures. Here, we will focus on diagnosis by pattern comparison. In principle, a diagnosis could be made by comparing the patient’s clinical profile with that of different diseases and selecting the most similar ones. However, even if a patient has a disease, he does not need to display all symptoms, and many signs are common to different diseases. In this respect, the selection of key variables and the introduction of adequate comparison criteria become essential issues.
Here, we develop mathematical tools for automatic screening of data stored in the form of numerical time sequences or matrices and illustrate the results in two medical contexts: diagnosis of systemic lupus erythematosus (SLE) patients and identification of cardiac abnormalities. This article is organized as follows. First, we introduce distances which are helpful to compare different kinds of data. Then, we explain how to adapt PlackettLuce models for rankings to select the most appropriate distances or hyperparameters when classifying data. We adapt these ideas to study data from anonymized SLE patients from two viewpoints. We initially take a clustering approach to find the onset of flares periods that require immediate medical attention. Next, we switch to a supervised classification approach to find daily patterns in the data representing a known sickness profile. Finally, we discuss applications to classify electrocardiogram patterns by comparing them with typical abnormal profiles and present our conclusions.
Distances for data
Consider a matrix \({\mathcal{M}}=(m_{i,j})\), \(i=1,\ldots , I\), \(j=1, \ldots , J\), containing data for I variables at J different times. We denote by \(\mathbf{m}^{(k)}\), \(k=1,\ldots K\), either the rows or the columns of this matrix, which we wish to compare. The Euclidean distance provides a standard tool to that purpose. Given two vectors \(\mathbf{m}^{(1)}\) and \(\mathbf{m}^{(2)}\) in a N dimensional space, their Euclidean distance is
We can define the Hamming distance for simple vectors whose components are taken from alphabets with 2 or 3 elements. The Hamming distance between two vectors \(\mathbf{m}^{(1)}\), \(\mathbf{m}^{(2)}\) whose components are −1, 0 or 1, for instance, is the number of positions at which the symbols are different. This distance is often used in communications to detect and correct errors in codes [10].
The dynamic time warping distance (TWD) provides an alternative method to quantify the similarity between two general data vectors [11, 12]. It has become a standard tool for analysing video, audio, and graphics data, with applications such as speech or signature recognition. The TWD algorithm seeks an optimal match between two given vectors (time sequences, for instance), subject to some conditions:

Every index from the first vector is matched with one or more indices from the second vector and vice versa.

The first index from the first vector is matched with the first index from the second vector (it does not need to be a unique match).

The last index from the first vector is matched with the last index from the second vector (it does not need to be a unique match).

The mapping of the indices from the first vector to indices from the second vector is monotonically increasing, and vice versa.
The idea is illustrated in Fig. 1. We define a cost by computing the sum of absolute differences of vector values for each matched pair of indices. The optimal match minimizes the cost subject to the above conditions.
The TWD can be calculated by dynamic programming based on cumulative distances as follows. Let \(\mathbf{m}^{(1)}\) and \(\mathbf{m}^{(2)}\) be two vectors with N and L components, respectively. Let \(d(x,y) \) be a distance between real numbers, for instance, the absolute value of the difference \(d(x,y)= xy\). We create a matrix \({\mathrm{TWD}}(n, \ell )\) for \(n=0,\ldots , N\), \(\ell =0,\ldots , L\). We set \({\mathrm{TWD}}(n, 0) = \infty \) for \(n=0,\ldots , N\), \({\mathrm{TWD}}(0, \ell ) = \infty \) for \(\ell =0,\ldots , L\), and \({\mathrm{TWD}}(0,0) = 0\). Then, for n from 1 to N and ℓ from 1 to L, we calculate
where \({\mathrm{TWD}}(n,\ell )\) represents the distance between subsequences \(m^{(1)}_{1},\ldots , m^{(1)}_{n}\) and \(m^{(2)}_{1}, \ldots , m^{(2)}_{\ell }\). The final result \({\mathrm{TWD}}(N,L)\) defines the distance between the two vectors. Figure 1 illustrates the difference between the Euclidean (1) and the dynamic time warping (2) distances.
The Earth Mover’s distance (EMD) is a more general concept, which quantifies the minimum cost of turning a collection of numeric values into another [13]. More precisely, the EMD between two vectors \(\mathbf{m}^{(1)}\) and \(\mathbf{m}^{(2)}\) formed by N and L real values, respectively, is
where \(d_{n,\ell } = m^{(1)} _{n}  m^{(2)}_{\ell }\) is the ground distance, and \(f_{n,\ell }\) minimizes the cost \(\sum_{n=1}^{N} \sum_{\ell =1}^{L} f_{n,\ell }\times d_{n,\ell }\) subject to the constraints
This distance tracks patterns in the compared vectors, regardless of their location.
Instead of tracking vectors, we may as well compare whole matrices. This can be done by a general family of distances, called Wasserstein type distances. As the EMD, their calculation is posed as optimal transport problems. Optimal transport plays crucial roles in many areas, including image processing and machine learning. Fast algorithms to calculate Wasserstein1 distances between distributions defined on a grid are proposed in [14]. Given two 2D probability distributions, or two images, \(\rho ^{0}\) and \(\rho ^{1}\), they seek a transport plan \(f(x)\) from one to the other with minimal cost
such that \({\mathrm{divergence}}_{h}(f) = \rho ^{0}\rho ^{1}\) under a zeroflux boundary condition. Here, p can be 1,2 or infinity, so that \(\f(x)\_{p}\) represents the \(L^{1}\), \(L^{2}\) or \(L^{\infty}\) norms of \(f(x)\), respectively, h is the grid step size and \({\mathrm{divergence}}_{h}\) a divergence operator defined on the grid, see [14] for details.
Distance and hyperparameter selection based on models for rankings
To evaluate which distance or hyperparameter choice performs better when classifying data of a different nature, we resort to PlackettLuce type models for rankings. We can implement different approaches depending on the ranking structure.
Bayesian approach with PlackettLuce ranking models
Consider a situation in which we have to judge the performance of N procedures \(\{{\mathrm{proc}}_{1}, \ldots , {\mathrm{proc}}_{N}\}\) on D sets of data \(\{{\mathrm{dat}}_{1}, \ldots , {\mathrm{dat}}_{D}\}\) of a similar nature. For instance, we wish to know which distance and hyperparameter choices perform best to extract specific information from the datasets through learning algorithms. To do so, we select D datasets for which the information we seek is known, apply the N procedures and quantify the error in the outcomes.
Excluding the possibility of ties, when we apply all the procedures to a dataset we obtain a ranking \(\boldsymbol{\rho }= (\rho _{1}, \ldots , \rho _{N}) \), that is, a permutation of \((1, \ldots , N)\) where \(\rho _{i} = j\) indicates that the ith procedure is ranked at the jth position [15, 16]. The procedure ranked first performs better than the procedure ranked second, and so on. To each ranking, we associate another permutation of \((1, \ldots , N)\) which defines an ordering \(\boldsymbol{\sigma }= (\sigma _{1}, \ldots , \sigma _{N})\), where \(\sigma _{i} = j\) indicates that the jth procedure is ranked at the ith position.^{Footnote 1} Orderings and rankings are related by \(\sigma _{\rho _{i}} = i\) and \(\rho _{\sigma _{i}} = i\), that is, the ordering is the inverse of the ranking.
We can describe the probability of observing a ranking ρ by means of the PlackettLuce (PL) model, a distribution over rankings expressed in terms of the associated orderings σ and parametrized by a vector \(\mathbf{w} = (w_{1}, \ldots , w_{N})\), \(w_{i} \geq 0\), \(i=1,\ldots N\) [15, 16]. The conditional PL probability of observing σ given w is
The probability of having \(\sigma _{1}=j\), that is, the jth procedure being the best, is \({\frac{w_{j} }{\sum_{k=1}^{N} w_{k}}}\) [16]. Assuming \(\sum_{k=1}^{N} w_{k}=1\), the parameters of the model represent the probability of each of the different procedures under consideration being the best. Given a set of observations \({\mathcal{O}}\) formed by D independent orderings σ, the conditional probability of observing this set is the product of the probabilities of each of the orderings:
We identify the model parameters w from the observations by Bayesian inference. By Bayes’s theorem
where \(P(\mathbf{w})\) is a probability distribution summarizing our prior knowledge on w. By simplicity, we assume that the sum of its components is 1. We choose a Dirichlet distribution \(P(\mathbf{w}) = {\mathrm{Dir}}(\mathbf{w}; \boldsymbol{\alpha })\) [17], though other choices are possible [16]. The Dirichlet distribution of order \(N \geq 2\) has a density function
where \(\sum_{i=1}^{N}w_{i} = 1\), \(w_{i} \geq 0\) for \(i=1,\ldots ,N\), and Γ denotes the Γ function. In the absence of additional information, we choose a uniform distribution for α, that is, \(\alpha _{i} = \alpha = 1\), \(i=1,\dots ,N\). This is the flat Dirichlet distribution, equivalent to a uniform distribution over the open standard \((N\!\!1)\)  simplex [18].
With these definitions, we can sample the unnormalized posterior distribution \(Q(\mathbf{w})=P(\mathbf{w}) \cdot P({\mathcal{O}}\mathbf{w})\) using Markov Chain Monte Carlo (MCMC) methods [19]. Note that the unknown scaling factor \(P({\mathcal{O}})\) in (7) is not needed for MCMC sampling. From the samples, we obtain information on the most likely values for w with quantified uncertainty, thus, we infer which procedure is the best.
MCMC methods produce a chain of Ndimensional states \(\boldsymbol{w}^{(0)} \longrightarrow \boldsymbol{w}^{(1)} ... \longrightarrow \boldsymbol{w}^{(k)} \)... which evolve to be distributed according to the target distribution [19]. We first sample an initial state \(\boldsymbol{w}^{(0)}\) from the prior distribution (8), and then move from one state \(\boldsymbol{w}^{(k)}\) to the next \(\boldsymbol{w}^{(k+1)}\) guided by a transition operator. We have selected a Hamiltonian Monte Carlo transition operator because it usually samples the distributions faster than other samplers, such as MetropolisHastings [20]. We set \(U(\boldsymbol{w}) = Q(\mathbf{w})\) (the probability to be sampled), \(K(\boldsymbol{p})= {\frac{1}{2}} \boldsymbol{p}^{t} \boldsymbol{p}\) and construct the Hamiltonian \(H(\boldsymbol{w}, \boldsymbol{p}) = U(\boldsymbol{w}) + K(\boldsymbol{p})\). We draw a random Gaussian moment \(\boldsymbol{p}^{(0)} \in \mathbb{R}^{N}\) from a multivariate normal distribution \(N(0, {\mathcal{I}})\), where the covariance matrix is the identity. Then, the transition proceeds as follows:

Given \(\boldsymbol{w}^{(k)}\) and \(\boldsymbol{p}^{(k)}\), we use them as initial values to solve the Hamiltonian equations \({\frac{\partial w_{j} }{\partial t}} = { \frac{\partial H }{\partial p_{j}}}\), \({\frac{\partial p_{j} }{\partial t}} =  { \frac{\partial H }{\partial w_{j}}}\), \(j=1,\ldots ,N\), by a leap frog algorithm for L steps of size δ. The final values define \(\boldsymbol{w}^{*}\) and \(\boldsymbol{p}^{*}\), the new proposed states for the chain.

We calculate \(\alpha = \min (1, \exp (U(\boldsymbol{w}^{*})+U( \boldsymbol{w}^{(k)})  K(\boldsymbol{w}^{*})+K(\boldsymbol{w}^{(k)}))\).

We generate a random number u from a uniform distribution \({\mathcal{U}}(0,1)\). If \(u \leq \alpha \), we accept the proposed states and set \(\boldsymbol{w}^{(k+1)}= \boldsymbol{w}^{*}\). Otherwise, we keep \(\boldsymbol{w}^{(k+1)}= \boldsymbol{w}^{(k)}\).
The final chain \(\boldsymbol{w}^{(0)}\), \(\boldsymbol{w}^{(1)}\), …, \(\boldsymbol{w}^{(M)}\) provides our set of samples, after discarding some initial states as a ‘burn in’ period. Once a large enough set of sampled weights is available, the sample with the highest probability furnishes the most likely value for w. Histograms or Boxplots built with the additional samples quantify the uncertainty in this prediction.
The above procedure excludes the presence of ties in the rankings under study. When some of the items to be ranked perform equally well on a dataset, the ordering takes the form \(\boldsymbol{\sigma }= \{C_{1},C_{2}, \ldots , C_{J}\}\), containing sets \(C_{j}\) with one or more items equally ranked, the items in each set being ranked higher than the next. When the ranking matrix contains ties, we can break randomly the ties T times, to obtain T new ranking matrices without ties. We apply to each of them the previous procedure and average the results provided for each of them to obtain average weights \(w_{i}\) indicating the prevalence of each ranked item. This procedure quantifies the tendency of the ranked item to be not only the first, but to appear in relevant positions, see [21] for a detailed study.
DavidsonLuce models with ties
The DavidsonLuce model [22, 23] works directly with rankings that may contain tied items. As said above, the associated orderings take the form \(\boldsymbol{\sigma }= \{C_{1},C_{2}, \ldots , C_{J}\}\) where \(C_{j}\) are sets of tied items, the items of each set being ranked higher than the next. For instance, consider 6 choices labelled as \(\{1,2,3,4,5,6\}\). Choices \(C_{1}=\{1,3\}\) perform equally well. Choices \(C_{3}= \{2,5,6\}\) too, but worse that \(C_{1}\). An ordering \(\boldsymbol{\sigma }= \{C_{1},C_{2}, C_{3}\}\) with \(C_{2}=\{4\}\) would represent that observation when \(C_{2}\) is better than \(C_{3}\) one but worse than \(C_{1}\).
In this framework, the probability of the choices \(A=\{a_{i_{1}}, \ldots , a_{i_{\ell }}\}\) being equally preferred to the remaining choices in a set of r total choices, \(1\leq \ell \leq r\), is proportional to
where \(A=\ell \) is the number of elements in A, \(\alpha _{j}\) is the worth of choice \(a_{j}\) and \(\delta _{A} \geq 0\) is a parameter representing the prevalence of ties of order \(A\) for \(A>1\). When \(A =1\), \(\delta _{1}\) can be set arbitrarily equal to 1. The remaining worths and prevalences are parameters to be fitted. The worths \(\alpha _{j}\) are interpreted as follows: conditional upon the outcome being an outright win for one choice, the probabilities for each choice to be the winner are in the ratios indicated by the \(\alpha _{j}\) [22, 23]. When \(\sum_{j} \alpha _{j}=1\), the worths represent the probability of each choice being the best assuming there is no tie in the first position. The probability of preferring an item \(a_{j}\) from a set S is \({\alpha _{j} / \sum_{k \in S} \alpha _{k}}\), \(j \in S\). The parameters representing tie prevalences \(\delta _{A}\) are interpretable in terms of tie probabilities between items of equal worth. For instance, \(\delta _{2}\) is related to the probability that two items of equal worth tie in the first position, conditional upon not having 3way or higher ties for first place. To be more precise, that probability is \(\delta _{2}/(2+ \delta _{2})\), see [22, 23].
Then, the probability of an ordering \(\boldsymbol{\sigma }=\{C_{1},C_{2}, \ldots , C_{J}\}\) allowing for ties up to order K is given by
where \(q(S)\) is defined in (9), \(A_{j}\) is the set of items from which \(C_{j}\) is chosen and \left(\begin{array}{c}{A}_{j}\\ k\end{array}\right) is the set of all possible choices of k items from \(A_{j}\). The parameter K equals the maximum number of ties encountered, so that \(\delta _{n} = 0\) when \(n>K\). The PlackettLuce model is a particular case in which ties are forbidden, \(\delta _{A}=0\) for \(A>1\). Given a matrix of independent observations \({\mathcal{O}}\), its probability is defined as the product of the probabilities (10) of each observation.
In this case, the worths and tie prevalences are fitted by BroydenFletcherGoldfarbShanno quasiNewton optimization algorithms (BFGS, LBFGS) for functions involving a large number of parameters, more complicated to implement than the Bayesian approach. Moreover, some parameter tuning or additional prior information may be needed to converge to a fit [22, 23]. Once they are calculated, we know the probabilities that an item is best conditional upon being one absolute win or also including different kinds of ties. It is the sum of the probabilities of all the rankings placing that item in the top set. For instance, considering \(\{ a_{1}, a_{2}, a_{3} \}\) we have 7 possible outcomes for the first position: 3 absolute winners (\(\{a_{1}\}\), \(\{a_{2}\}\), \(\{a_{3}\}\)), 3 double ties (\(\{a_{1}, a_{2}\}\), \(\{a_{1}, a_{3}\}\), \(\{a_{2}, a_{3}\}\)) and one triple tie (\(\{a_{1}, a_{2}, a_{3}\})\), with probabilities proportional to \(\alpha _{1}\), \(\alpha _{2}\), \(\alpha _{3}\), \(\delta _{2} (\alpha _{1}\alpha _{2})^{1/2}\), \(\delta _{2} (\alpha _{1}\alpha _{3})^{1/2}\), \(\delta _{2} (\alpha _{2}\alpha _{3})^{1/2}\), \(\delta _{3} (\alpha _{1} \alpha _{2}\alpha _{3})^{1/3}\). The DL model provides 7 probabilities that sum up to 1 in those proportions. The probability of \(a_{1}\) being best including ties is the probability of having the outcomes \(\{a_{1}\}\), \(\{a_{1}, a_{2}\}\), \(\{a_{1}, a_{3}\}\) or \(\{a_{1}, a_{2}, a_{3}\}\). This strategy gives more value to being placed in the first position most of the time than to being usually placed at good positions. The R package [23] implements this scheme and provides standard errors indicating the precision of the estimates, which are not always useful due to overlaps.
Application to laboratory data from systemic lupus erythematosus
Laboratory tests constitute a cornerstone of the diagnosis process for many health disorders: blood counts, enzyme, hormone, metabolic, ionic, protein, vitamin and antibody levels are measured at different times. Developing automatic tools to trace key features in the resulting numeric matrices can help to interpret the data, specially when dealing with disorders which are difficult to diagnose. Consider systemic lupus erythematosus (SLE), for instance, a chronic autoimmune disease in which the immune system attacks healthy tissues by mistake [24]. This attack causes inflammation and, in some cases, permanent tissue damage. SLE is difficult to identify and treat appropriately because many symptoms are non specific and change throughout the course of the disease. Lupus patients go through periods of illness, called flares, and periods of wellness, called remission. The symptoms during flares vary. It is essential to be able to distinguish early when the patient is transitioning from remission to flares, and what factors are causing it, to timely administer adequate treatments. As it happens with other diseases, SLE diagnosis relies heavily on alterations observed in laboratory tests.
Next, we propose a procedure to extract automatically relevant features from time series of laboratory tests. The idea is as follows. First, we apply clustering techniques to locate relevant time frames and then seek specific patterns in the data to select a possible diagnosis, resorting to adequate distances. Finally, we will illustrate the process on anonymized data from SLE patients.
Clustering time records
Let us consider a matrix \({\mathcal{M}}\) collecting the results of laboratory tests for a patient during a series of days. We monitorize I variables (rows) at J times (columns). To analyze time variations, we normalize the data as follows. For each variable \(i=1,\ldots ,I\), we calculate the mean \(\mu _{i}\) and the standard deviation \(\sigma _{i}\) of the corresponding row \(m_{i,j}\), \(j=1,\ldots ,J\). Then, we construct the normalized matrix with elements
most of which lie in \([1,1]\), see Fig. 2. This procedure defines Normalization 1. This normalization is useful to visualize time periods in the data. Figure 2, for instance, suggests a period of strong instability between days 810868.
Datasets are often incomplete, as Fig. 2 shows. For day clustering purposes, we adopt the convention of suppressing rows (variables) for which more than half the recordings are missing. Otherwise, we fill the gaps with the average of the two contiguous values. We also remove rows corresponding to variables which do not change noticeably during the studied period [21]. The outcome is represented in Fig. 3 for the remaining 60 variables. The 5 variables comprised between Glucose and Nitrites are removed because they remain essentially zero. Nevertheless, their presence does not affect the clustering results. One could additionally remove days for which less than half the variables are recorded, as it happens for day 850. However, the measurements recorded just before and after it are taken in a narrow time gap compared to other measurements present in the datasets, thus, interpolating between the two neighbouring days seems reasonable. For days 526 and 565, the variation between the values recorded before and after them is small. Keeping or removing these days does not change significantly the analysis either.
We identify automatically different time stages in the evolution of the patient by means of clustering techniques such as KMeans (KM) [25], Hierarchical clustering (HC) [26] and densitybased spatial clustering of applications with noise (DBSCAN) [27], applied to the normalized matrix (11) with the Euclidean distance and different hyperparameter selections. We set as initial day of a flares period the first day of the shortest cluster for KM and HC and the smallest outlier for DBSCAN with different hyperparameters [21]. For ease of the reader, we briefly describe these clustering algorithms in Appendix A.
To select the most adequate clustering procedure for the kind of data we are working with we first study a collection of datasets which have already been diagnosed. We have considered anonymized data from 19 patients containing one flares period whose starting day is known. The strategy is as follows:

For each dataset, we normalize the raw data matrix according to (11).

We apply to the columns KM and HC, setting different numbers of clusters, and also DBSCAN for different choices of ε (smallest distance for points to be considered neighbours) and MP (minimum number of points to be considered a cluster).

For each dataset, we compare the prediction obtained for the onset of the flares period with the known onset day. Then, we rank the procedures according to the difference between both.

If the resulting matrix of rankings does not contain ties, we use the PlackettLuce model (6)(8) to obtain the worths w representing the probability of each clustering procedure to provide the best estimate of the onset of flares and to quantify the uncertainty in this result.

If the resulting matrix of rankings contains ties, we can break the ties randomly and average the results provided by (6)(8) for a large number of attempts, as detailed in [21]. Alternatively, we can use the DavidsonLuce model (9)(10) to obtain the worths w and tie prevalences δ. Then we use them to estimate the probability of each clustering procedure being the best counting all the possible outcomes (alone or tied).
We have implemented this scheme setting the number of clusters equal to 3, 4, 5 in KM and HC, and DBSCAN with \(\varepsilon =3, 3.5\) and \(M\! P = 3\). The resulting ranking matrices contain ties. However, our ranking analysis tends to select HC with 4 clusters as the best strategy, followed by HC with 3 clusters. This study is further detailed in Appendix B. For the data represented in Fig. 2, this strategy singles out automatically days 858862 (flares crisis), and splits the remaining days into three periods 1850 (preflares), 868930 (treatment), 9301219 (postflares), see Fig. 4.
Detecting distinguished days, or blocks of similar days, allows us to focus our study on the status of the variables on the specific days which mark transitions from remission to flares, or on specific periods to identify types of flares and observe the response to treatment. For example, if we focus in the suspected flares period in Fig. 2, we notice a sharp increase in antidsDNA antibodies on day 810 and a sharp decrease of the red blood cell counts, hemoglobin and hematocrit measurements after it. We also observe an increase in creatinine in blood and the presence of leukocytes in urine. Searching for specific patterns in the data during the selected days may help us to propose a diagnosis, as we explain in the next section.
Daily pattern search
To this purpose, we renormalize the raw data matrix \({\mathcal{M}}\) in a different way. Assuming normality ranges for the variables are known, we construct a normalized matrix with elements \(\hat{m}_{i,j}\) replacing \(m_{i,j}\) with 1 or −1 depending on whether the original values are above or below the normality ranges of the ith variable, and 0 when they are in the normality range, see Fig. 5. When the information stored is binary, either ‘true’ or ‘false’, we replace it by either 1 (positive, true) or 0 (negative, false). This defines Normalization 2.
Many disorders can be characterized as simple −1, 0, 1 patterns, or combinations thereof:

Normocytic anemia: −1 Hemoglobine in blood, 0 Mean corpuscular velocity.

Hypocomplementemia C3, −1 C3 in blood, and Hypocomplementemia C4, −1 C4 in blood. They usually indicate active SLE.

Leukopenia: −1 Leukocytes in blood. It usually indicates active SLE or infection or medication effects.

Neutropenia: −1 Neutrophiles in blood. It usually indicates active SLE or infection or medication effects.

Lymphopenia: −1 Lymphocytes in blood. It usually indicates active SLE or infection.

High AntidsDNA antibodies: +1 AntidsDNA, the most relevant antibody for SLE.

High Creatinin: +1 Creatinin in blood. It usually indicates kidney failure.

Pathological proteinuria: +1 Proteins in urine. It usually indicates inflammation of the renal glomerulus or tubular damage.

Hyperglycemia: +1 Glucose in blood. It may be due to the use of corticosteroids, or a sign of Diabetes Mellitus.

Hypercholesterolemia: +1 Cholesterol LDL in blood. Risk factor to be treated, as associated with hypertension and corticosteroids increases cardiovascular risk.

Suspicion of hemophagocytic syndrome: −1 Hemoglobin or −1 Leukocytes or −1 Neutrophils or −1 Lymphocytes or −1 Platelets in blood, plus large ferritin counts, and large triglycerid or small fibrinogen counts in blood. It is a complication of sustained severe inflammatory states, including SLE.
The combinations of −1, 0, 1 patterns observed in distinguished days mark specific types of evolution.
To select the most adequate distance to this purpose, we have compared the performance of the Euclidean distance, the Earth Mover’s distance and the Hamming distance to identify definite patterns formed combining 1, 0 and −1 in columns containing those digits. We propose a collection of P patterns and distribute them in the columns of a \(I \times J\) matrix, filling the remaining positions with the digits −1, 0, 1. Then we calculate the distances between the columns and the reference patterns, and rank the distances according to their performance. Implementing a PlackettLuce model ranking analysis, we conclude that the Hamming distance is by far the most efficient one. This strategy is detailed in Appendix C.
Once this fact has been established, the ‘automatic diagnosis’ strategy is as follows:

For a given dataset, we normalize the raw data matrix to obtain a matrix involving only −1, 0, 1 according to the normality ranges of the observed variables.

We compare the status of the variables in given columns corresponding to specific days with known sickness patterns defined by −1, 0, 1 sequences using the Hamming distance.
In this way, we can screen large datasets selecting patient profiles requesting immediate attention, while providing at the same time a simplified diagnosis that should receive further consideration from a specialist. For instance, for the dataset considered in Figs. 2 and 5, we identify SLE activity indicated by elevated antidsDNA and low C3 and C4 complements. We identify low hemoglobin and hematocrit levels suggesting an anemization (normocytic anemia) process as a result of illness. Alterations in the glomerular filtration rate and presence of proteins in urine indicate alterations in the kidney function too. High creatinine values at days 810 and 858 indicate kidney damage and require immediate treatment to restore normal values.
Application to electrocardiogram interpretation
When a specific variable is recorded for an interval of time, the data take the form of curves. Often, typical patterns representing normal stages and abnormal stages are known. Developing automatic screening tools can assist doctors in processing such data. Consider electrocardiograms (ECG) for instance. An electrocardiogram is a graph of the electrical activity of the heart [28] representing voltage versus time, see Fig. 6. They are recorded placing electrodes on the skin, which detect small electrical changes resulting from cardiac muscle depolarization and repolarization during each heartbeat (cardiac cycle). Alterations in the ECG pattern indicate cardiac abnormalities. Numerous diagnoses are based on the observed patterns. Producing effective tools to automatically identify cardiac patterns would allow for the proper use of defibrillators when untrained emergency staff assists people going into cardiac arrest at work or leisure centers. Currently, it is hard to succeed in the absence of an expert doctor who interprets the signs.
Our goal in this section is to introduce an automatic procedure to identify alterations in patterns described by one dimensional curves. The idea is as follows. We first define a set of normalized reference curves corresponding to different known pathologies. Given another curve normalized in the same way, we compare it with the reference patterns by means of an adequate distance. The smallest distance selects a possible diagnosis. We will illustrate the process on examples taken from electrocardiography.
Electrocardiogram structure and basic alterations
Electrocardiograms repeat periodically the structure represented in Fig. 6. Different regions are designed by P, Q, R, S and T [29]:

P wave: The first wave in the ECG represents atrial depolarization. Usual length is shorter than 0.11 s in adults. Usual amplitude is smaller than 0.25 mV. Its shape is smooth and rounded.

QRS complex: It is formed by a sequence of waves representing ventricular depolarization. Usual duration is about 0.06 s  0.10 s. A small negative wave Q is followed by a large positive R wave and a small negative S wave.

T wave: The next wave represents ventricular repolarization. Usual length is shorter than 0.20 s in adults. Usual amplitude is between 0.2 and 0.3 mV. Its shape is smooth and rounded.

U wave: This last tiny wave is believed to represent papillary muscle repolarization. It may not be seen, and is often ignored.
The P wave and the QRS complex are separated by the PR segment, while the ST segment separates the QRS complex and the T wave. The P wave with the PR segment form the PR interval. The QRS complex, the ST segment and the T wave form the QT interval. Variations in the heart’s structure and its environment (blood composition included) alter these entities. The presence, absence and size of the different ECG parts characterize different cardiac abnormalities, comprising cardiac rhythm disturbances (such as ventricular tachycardia and atrial fibrillation), perturbed coronary artery blood flow (such as myocardial infarction and myocardial ischemia), as well as electrolyte disturbances (such as hyperkalemia and hypokalemia).
We select a few representative patterns to illustrate the method:

A normal ECG repeats the basic PQRST structure corresponding to one heartbeat at a rate between 60 and 100 beats per minute (bpm)

Sinus tachycardia is a heart rate greater than 100 beats per minute. This is normal with exercise, but abnormal otherwise.

Sinus bradycardia is a sinus node dysfunction with a rate under 60 beats per minute.

Sinus arrhythmia is an irregular heartbeat that can be either too fast or too slow. It is characterized by variations in the PP intervals greater than 0.12 s (from one beat to the next).

QT prolongation is a sign of delayed ventricular repolarisation. This means that the heart muscle takes longer than usual to recharge between beats. It is a known side effect of a wide range of medicines. Excessive QT prolongation can trigger tachycardias and torsades de pointes (TdP).

Torsades de Pointes is a ventricular tachycardia, fast and polymorphic, characterized by fluctuation of the QRS complexes around the electrocardiographic baseline. It may lead to lifethreatening ventricular fibrillation.
To compare ECGs we must normalize them somehow. ECGs are usually recorded in red grid paper, see Fig. 6. Panel (a) displays a standard printed ECG. There is an established correspondence between the number of squares and the units (mm, mV, s). Such an image, or portions of the image, can be read as a matrix of a size adjusted to the resolution. From that matrix, we extract only the locations corresponding to black points forming the curve. When several ordinates correspond to the same abscissa, we keep only one ordinate: their average value. Panel (b) represents the outcome of this procedure applied to part of panel (a), indicating the different ECG regions. Additionally, we have normalized heights by setting the first point of the curve at zero. Figure 7 collects reference patterns for the abnormalities under consideration obtained from diagnosed ECGs, available in open access image datasets, such as [30]. We have considered regions defined by the same number of red squares and extracted the ECG curves by the same procedure, choosing a smaller matrix size (a factor 10 smaller). One this is done, we need to choose appropriate distances to compare patterns.
Electrocardiogram classification
This section illustrates how to use TWD, EMD and Wasserstein distances to classify ECG patterns. Figure 8 shows a set of normalized ECGs we want to classify as displaying a kind of abnormality or another. To do so, we select a distance and calculate the distances between these ECG and the reference patterns displayed in Fig. 7. Then, we propose a diagnosis based on the smallest value obtained for each of them.
Tables 1 and 2 reproduce the distance matrices obtained for the TWD and the one dimensional EMD. Tables 3 and 4 reproduce the distance matrices for the 2D image Warssentein1 distance with Algorithm 1M and Algorithm 2M from [14], respectively, and the 1norm (\(p=1\)). Choosing the norms for \(p=2\) or \(p=\infty \) we find the same classification.
To visualize the strategy’s performance, we create Table 5, whose entries are 1 when patient i is correctly diagnosed by distance j and 0 otherwise. Notice that the TWD performs better, and classifies all the patients correctly. Then the Wasserstein1 distance correctly classifies four out of five patients. The ECG of the patient who is misclassified displays bradycardia and is classified as QT prolongation, which could be confused with slow rhythms.
To better evaluate the performance of different distances, we have studied a larger collection of 100 electrocardiograms obtained by perturbing the reference patterns. We can construct ranking and observation matrices by comparing each of them with the reference ECG patterns using the TWD, EMD and Wasserstein distances. Then, we check whether the proposed diagnosis is correct or not. Implementing a PlackettLuce model ranking analysis, we conclude that the dynamic time warping distance would be the most efficient one for this dataset. However, this synthetic dataset is quite limited. Testing the method on more general curves should be the object of further work, but it would require the previous obtention of an extensive collection of diagnosed authentic EGC images.
Conclusions
The design of automated tools for detecting abnormalities in a patient’s clinical picture offers an excellent opportunity to improve clinical case management and clinical research by developing new algorithms and software.
Medical data are usually stored in the form of matrices or sequences. Time sequences of results of Laboratory tests, for instance, take the form of numerical matrices. We have discussed two different types of normalization, one related to average values for unsupervised learning, and the other related to normal ranges for supervised classification. First, we show how to use clustering techniques to distinguish periods of medical relevance in the time evolution of a SLE patient. In particular, we locate flares periods which require immediate medical attention. Then, we illustrate how to propose a possible diagnosis by seeking specific numeric patterns in each period according to the second normalization. Using Hamming distances to compare the outcome of Laboratory analyses at different days, or for different patients, one could automatically classify patient’s profiles.
On the other side, recordings of vital signals, such as electrocardiograms, take the form of time sequences, usually represented as curves. We propose a strategy to identify abnormalities in recorded ECGs by comparing recorded curves to patterns representing typical anomalies. First, we introduce a normalization procedure. Then, we investigate the potential of time warping, Earth mover’s and Wasserstein distances to correctly classify basic abnormality patterns, finding a better performance for the dynamic time warping distance on a synthetic dataset. Further studies would require the previous obtention of an extensive collection of diagnosed authentic EGC images. Being able to correctly classify abnormal ECG patterns in an automated way would increase the chance of survival in outofhospital treatments. It would allow for the proper use of defibrillators by untrained emergency staff when assisting cardiac arrest cases at work or leisure centers.
In these studies, it is essential to select the distances and hyperparameters best adapted to the datasets under consideration. We have introduced a PlackettLuce ranking based analysis as a tool to select the most adequate distances and hyperparameters to analyze datasets with a specific structure. The techniques developed here may set a basis for automatic screening of medical information based on pattern comparison.
Availability of data and materials
Anonymized data could be available upon request. The PlackettLuce model with Dirichlet prior distribution and MCMC samplers are implemented in the library https://github.com/b0rxa/scmamp. The PlackettLuce model with ties is implemented in https://cran.rproject.org/web/packages/PlackettLuce/index.html. Codes for Wasserstein distances are available at https://github.com/liujl11git/multilevelOT. Implementations of the TWD are available at https://github.com/talcs/simpledtw/blob/master/simpledtw.py for instance.
Notes
Some authors use a different terminology [17], speaking of ‘performance’ and ‘ranking’, a ‘ranking’ being in fact an ‘ordering’.
Notice that we have written 1,1,3 and not 1,1,2. This choice does not alter the outcome.
Abbreviations
 SLE:

Systemic Lupus Erythematosus
 TWD:

dynamic time warping distance
 EMD:

Earth movers’ distance
 KM:

K Means
 HC:

Hierarchical clustering
 DBSCAN:

DensityBased Spatial Clustering of Applications with Noise
 PL:

PlackettLuce
 DL:

DavidsonLuce
 MCMC:

Markov Chain Monte Carlo
 ECG:

Electrocardiogram
 TdP:

Torsades de pointes
References
Rajpurkar P, Irvin J, Zhu K, Yang B, Mehta H, Duan T, Ding D, Bagul A, Ball RL, Langlotz C, Shpanskaya K, Lungren MP, Ng AY. CheXNet: radiologistlevel pneumonia detection on chest Xrays with deep learning. 2017. https://arxiv.org/abs/1711.05225.
Wu E, Hadjiiski LM, Samala RK, Chan HP, Cha KH, Richter C, Cohan RH, Caoili EM, Paramagul C, Alva A, Weizer AZ. Deep learning approach for assessment of bladder cancer treatment response. Tomography. 2019;5(1):201–8.
Vogt W, Nagel D. Cluster analysis in diagnosis. Clin Chem. 1992;38(2):182–98.
Benjamin JR. Making connections: using networks to stratify human tumors. Nat Methods. 2013;10(11):1077–8.
Soul J, Dunn SL, Anand S, SerracinoInglott F, Schwartz JM, BootHandford RP, Hardingham TE. Stratification of knee osteoarthritis: two major patient subgroups identified by genomewide expression analysis of articular cartilage. Ann Rheum Dis. 2017;0:1–8.
Pouryahya M, Oh JH, Mathews JC, Deasy JO, Tannenbaum AR. Characterizing cancer drug response and biological correlates: a geometric network approach. Sci Rep. 2018;8:6402.
Carlier A, Vasilevich A, Marechal M, de Boer J, Geris L. In silico clinical trials for pediatric orphan diseases. Sci Rep. 2018;8:2465.
Saeb S, Lonini L, Jayaraman A, Mohr DC, Kording KP. The need to approximate the usecase in clinical machine learning. GigaScience. 2017;6:1–9.
ValladaresRodriguez S, PérezRodriguez R, FernandezIglesias JM, AnidoRifón LE, Facal D, RivasCosta C. Learning to detect cognitive impairment through digital games and machine learning techniques. Methods Inf Med. 2018;57:197–207.
Waggener B. Pulse code modulation techniques. Berlin: Springer; 1995. p. 206. https://books.google.com/books?id=8l_o6kI3760C&pg=PA206.
Yingmin L, Huiguo C, Zheqian W. Dynamic time warping distance method for similarity test of multipoint ground motion field. Math Probl Eng. 2010;2010:749517.
Gold O, Sharir M. Dynamic time warping and geometric edit distance: breaking the quadratic barrier. J Assoc Comput Mach. 2018;14(4):50.
Yossi R, Carlo T, Leonidas JG. A metric for distributions with applications to image databases. In: Proceedings ICCV. 1998. p. 59–66. https://doi.org/10.1109/ICCV.1998.710701.
Jialin L, Wotao Y, Wuchen L, Yat TC. A fast approximation of Wasserstein1 distances. SIAM J Sci Comput. 2021;43(1):A193–A220.
Marden J. Analyzing and modeling rank data. London: Chapman & Hall; 1995.
Guiver J, Snelson E. Bayesian inference for PlackettLuce ranking models. In: Proceedings of the 26th annual international conference on machine learning  ICML 09. 2009. p. 1–8.
Calvo B, Ceberio J, Lozano JA. Bayesian inference for algorithm ranking analysis. In: GECCO genetic and evolutionary computation conference companion, Kyoto, Japan. New York: ACM; 2018.
Kotz S, Balakrishnan N, Johnson NL. Continuous multivariate distributions. Volume 1: models and applications. New York: Wiley; 2000.
Gilks WR, Richardson S, Spiegelhalter D. Markov chain Monte Carlo in practice. Boca Raton: CRC Press; 1995.
Neal RM. MCMC using Hamiltonian dynamics. In: Brooks S, Gelman A, Jones GL, Meng XL, editors. Handbook of Markov chain Monte Carlo. London: Chapman & Hall; 2011.
Carpio A, Simón A, Villa LF. Clustering methods and Bayesian inference for the analysis of the time evolution of immune disorders. 2020. https://arxiv.org/abs/2009.11531.
Firth D, Kosmidis I, Turner HT. DavidsonLuce model for multiitem choice with ties. 2019. https://arxiv.org/abs/1909.07123.
Turner HL, van Etten J, Firth D, Kosmidis I. Modelling rankings in R: the PlackettLuce package. Comput Stat. 2020;35:1027–57.
Arnaud L, van Vollenhoven R. Advanced handbook of systemic lupus erythematosus. Berlin: Springer; 2018.
MacQueen J. Some methods for classification and analysis of multivariate observations. In: Proceedings of the fifth Berkeley symposium on mathematical statistics and probability, volume 1: statistics. Berkeley: University of California Press; 1967. p. 281–97.
Kaufman L, Rousseeuw PJ. Finding groups in data: an introduction to cluster analysis. Hoboken: WileyInterscience; 1990.
Ester M, Kriegel HP, Sander J, Xu X. A densitybased algorithm for discovering clusters in large spatial databases with noise. In: Proceedings of the second international conference on knowledge discovery and data mining. Portland: AAAI Press; 1996. p. 226–31.
Lilly LS, editor. Pathophysiology of heart disease: a collaborative project of medical students and faculty. 6th ed. Baltimore: Williams & Wilkins; 2016.
Klabunde RE. Electrical activity of the heart. Cardiovascular physiology concepts. Baltimore: Williams & Wilkins; 2005.
Aringer M, Costenbader K, Daikh D, Ralph Brinks R, Mosca M, RamseyGoldman R et al.. European league against rheumatism/American college of rheumatology classification criteria for systemic lupus erythematosus. Ann Rheum Dis. 2019;78:1151–9.
Hopkins B, Skellam JG. A new method for determining the type of distribution of plant individuals. Ann Bot. 1954;18(2):213–27.
Acknowledgements
The authors thank Hospital Universitario Puerta de Hierro (Madrid, Spain) for providing anonymized SLE clinical data.
Funding
This research has been partially supported by the FEDER /Ministerio de Ciencia, Innovación y Universidades  Agencia Estatal de Investigación grants No. MTM201784446C21R and PID2020112796RBC21.
Author information
Affiliations
Contributions
AS and AT were responsible for data curation, visualization and software devolopment. AC and LFV conceptualized and supervised the work, provided resources and acquired funds. AC, AS and AT selected the methodology and wrote the paper, with edits from LFV. All authors carried out the investigation, formal analysis and validation of results. All authors read and approved the final manuscript.
Corresponding author
Ethics declarations
Competing interests
The authors declare that they have no competing interests.
Appendices
Appendix A: Clustering techniques
In this section we briefly recall the clustering techniques used to classify SLE data.
Kmeans [25] clusters data in groups in order to minimize the total intracluster variation, which measures the cluster compactness. Given a data cloud formed by points \(\mathbf{x}_{i}=(x_{i,1}, \ldots , x_{i,M})\) in a M dimensional space, the intracluster total variation is given by \(\sum_{j=1}^{J} W(C_{j}) = \sum_{j=1}^{J} \sum_{\mathbf{x}_{i} \in C_{j}} d(\mathbf{x}_{i}, \boldsymbol{\mu }_{j})\), where \(C_{j}\) is a cluster of such points and \(\boldsymbol{\mu }_{j}\) is the cluster centroid, for \(j=1,\ldots,J\). Each term \(W(C_{j})\) represents the variation within a cluster. Here, the distance d stands for the Euclidean distance \(d(\mathbf{x}_{i},\boldsymbol{\mu }_{j})^{2} = \sum_{m=1}^{M} (x_{i,m}  \mu _{j,m})^{2}\). The Kmeans algorithm proceeds in the following steps. We fix the number of clusters k to be formed and initialize the centroids \(\boldsymbol{\mu }_{j}\) by randomly generating k points. Next, each datum \(\mathbf{x}_{i}\) is assigned to the centroid minimizing the Euclidean distance. Within each cluster, we set the average of the cluster points as the new centroid. These steps are repeated until the clusters do not change. Kmeans needs one hyperparameter to proceed: the number of clusters. There are specific criteria such as the Elbow or Silhouette methods to propose a tentative cluster number.
Hierarchical clustering produces a multilevel hierarchy, in which clusters at one level coalesce at the next level [26]. The agglomerative algorithm starts from as many clusters as data points. Nearby clusters merge to create larger ones until all the data points form a single cluster. This procedure is schematized in a dendrogram, a graph visualizing how the clusters join until they form a tree that comprises all, see Fig. 4. To evaluate the proximity of clusters and merge them, the algorithm employs ‘linkage functions’: ‘single’, ‘average’, ‘complete’, ‘weighted’, ‘centroid’, ‘median’, ‘ward’. Both the linkage functions and the distance are hyper parameters to be selected. Here, we have fixed the Euclidean distance and implemented the linkage providing the biggest cophenetic correlation coefficient for these datasets, that is, the biggest correlation between the original distance and the cophenetic distance (the height at which clusters coalesce). Large correlation indicates that the tree is representative of our dataset. The remaining hyperparameter, that is, the height, determines the number of clusters. Cutting the tree at different heights, we select specific numbers of clusters.
The DensityBased Spatial Clustering of Applications with Noise (DBSCAN) algorithm [27] defines clusters in high density regions, leaving observations in low density regions outside, which eventually become anomalies. The process starts with an arbitrary data point that has not yet been classified. We find the points at a distance smaller than ε (the εneighbourhood). When it contains more than a minimum number of points MP, we create a new cluster with them. Otherwise, we consider that point as noise. However, this point might become part later of the εneighbourhood of a different point containing enough points, and, thus, belong to that cluster. If not, it remains an outlier. Thus, this algorithm can identify non convex clusters and outliers. However, finding good values for the two hyperparameters ε and MP is a nontrivial task, strongly dependant on the dataset’s structure. In principle, the distance has to be selected too, but we have fixed the Euclidean distance.
Appendix B: Stages in SLE patients’ medical records
This appendix summarizes some results obtained when applying the strategy described in Section 4 to 19 time series of anonymized laboratory data of patients diagnosed with SLE according to criteria of the European League against Rheumatism/American College of Rheumatology [31]. The number of variables and days varies slightly within them. All the datasets considered contained only one transition, and Hopkins criterion for them indicated the presence of relevant clusters [32]: the Hopkins statistics \(H>0.5\). Table 6 quantifies the distance between the column of the normalized heatmap selected as onset of flares after the clustering procedure and the true column corresponding to the diagnosed day. HC stands for hierarchical clustering (3C with 3 clusters, 4C with 4 clusters) and KM by Kmeans (with k clusters). DBSCAN uses parameters \(M\!P=3\) and \(\varepsilon =3\). Next, we build the ranking presented in Table 7. We assign a higher position in the ranking to smaller distances. The smallest possible distance is \(D=0\). Smallest distances rank first. Ties here are represented assigning the same position to tied algorithms and freeing the next positions in equal number. Following this convention, we obtain Table 7.
Finally, we illustrate the use of PL methods to determine the probability of each algorithm being the best, as well as the uncertainty in our choice of algorithm.
Table 8 represents the results obtained combining the PL method with random tiebreaking and averaging. These results indicate that hierarchical clustering with 4 clusters is the algorithm performing best, with a probability of 21.30%. Next, it follows hierarchical clustering with 3 clusters, with a 20.36% probability. DBSCAN appears with probability 18.73%. Figure 9 quantifies uncertainty using boxplots constructed from samples from the 100 runs (1000 samples from each run).
A similar study can be performed allowing for ties by means of the DavidsonLuce model. Next, we illustrate this procedure on distance selection.
Appendix C: Distance selection
This appendix illustrates how to select the best distance to compare sequences formed by the digits −1, 0, 1. We consider 5 patterns containing 125 digits and up to 300 sequences obtained changing randomly a few digits in such patterns. Thus, the underlying reference pattern (the ‘diagnosis’) is known for each of them. Next, we compare each sequence (each ‘patient’) with the 5 reference patterns using the Hamming distance \(d_{1}\), the Euclidean distance \(d_{2}\) and the Earth Mover’s distance \(d_{3}\). Table 9 illustrates the outcome for \(M=5\) patients.
Now, we create an intermediate \(NM \times 3\) success score table R:
and reduce dimensionality calculating N intermediate averages for consecutive blocks of M patients. For the first 5 patients, we find Table 10, where the last row is obtained adding up each column. The last row becomes the first row of a \(N\times 3\) performance matrix. Repeating this process for \(N=10\) consecutive blocks of \(M=5\) patients we find the performance matrix collected in Table 11.
From the performance matrix, we construct the table of rankings, see Table 12. For each row, the distance that scores a higher number of correct assignations ranks first, the next one ranks second, and so on. Distances with the same number of correct guesses are tied.
This ranking^{Footnote 2} is introduced in the R package [23] to find the worths and tie prevalences involved in formulas (9) and (10). We have implemented this procedure for different choices of M and N. Setting \(M=20\) and \(N=15\), for instance, we have found the worths \(\alpha _{1}=0.79739001\), \(\alpha _{2}=0.13612835\) \(\alpha _{3}=0.06648164\) and tie prevalences \(\delta _{2}=0.57440580\), \(\delta _{3}=0.79125103\). The probability of each distance being the best is the sum of the probabilities of that distance being the first alone, tied with another one or in a triple tie. We obtain for each of them the probabilities listed in Table 13.
Notice that the probabilities in Table 13 do not add to one because the probabilities of 2ties and 3ties are counted twice and three times. In any case, the Hamming distance clearly outperformed the rest for the synthetic dataset considered.
Alternatively, we can proceed as in the previous appendix, resorting to the PL method with random tiebreaking and averaging. The resulting probabilities add up to one, and the Hamming distance is still placed first.
Rights and permissions
Open Access This article is licensed under a Creative Commons Attribution 4.0 International License, which permits use, sharing, adaptation, distribution and reproduction in any medium or format, as long as you give appropriate credit to the original author(s) and the source, provide a link to the Creative Commons licence, and indicate if changes were made. The images or other third party material in this article are included in the article’s Creative Commons licence, unless indicated otherwise in a credit line to the material. If material is not included in the article’s Creative Commons licence and your intended use is not permitted by statutory regulation or exceeds the permitted use, you will need to obtain permission directly from the copyright holder. To view a copy of this licence, visit http://creativecommons.org/licenses/by/4.0/.
About this article
Cite this article
Carpio, A., Simón, A., Torres, A. et al. Pattern recognition in data as a diagnosis tool. J.Math.Industry 12, 3 (2022). https://doi.org/10.1186/s1336202200119w
Received:
Accepted:
Published:
DOI: https://doi.org/10.1186/s1336202200119w
Keywords
 Pattern classification
 Hyperparameter selection
 PlackettLuce models
 Hamming distance
 Dynamic time warping distance
 Wasserstein distance
 Medical diagnosis
 Systemic lupus erythematosus
 Electrocardiogram