symmetric function (i.e., σ(x, x′) = σ(x′, x)) such
that every datum maximizes the similarity to itself
(i.e., σ(x, x′)≤σ(x, x)for all x, x′∈S). Given
a subset sof S, the diameter of swith respect to σis
diam(s) = max{σ(x, x′) : x, x′∈s, x 6=x′}.
Like set-significance-based kernels, diameter-
based kernels kσare designed using the subsets sof
S(sorted in a list) as the hidden variables. Denoting
by ci(p)the total number of clusters of p contained in
the subset siand by and νi(p)the function indicator of
non-singleton clusters of p within si, having the value
1 if there are singleton clusters of p within siand the
value 0 otherwise, the ith hidden feature is defined by:
fi(p) = 1 + νi(p)·diam(si)
ci(p),if ci(p)>0;
0,if ci(p) = 0.
Proposition 4. kσsatisfies Bottom-Up Collinearity,
Top-Down Collinearity, and Meet Predominance.
4 A well-known family of not
well-behaved kernels
We conclude with a brief note regarding the
prototype-based kernels. These kernels are designed
from a set of prototypes P:= {p1,p2, . . . , pm} ⊆ PS
and a dissimilarity measure d:PS×PS→R(not
necessarily a metric) for comparing partition. Ev-
ery partition p ∈PSis assigned the vector Vp=
(d(p,p1), d(p,p2), . . . , d(p,pm)) (feature map), and
the kernel is accordingly given by k(d,P)(p,p′) =
hVp, Vp′i.
Justified in part by the intuition that the proto-
types can be chosen by experts in the task at hand
to stand for the different characteristic features of
the data population, prototype-based measures have
gained significant acknowledgment in the scientific
community. Both empirical and theoretical evidence
have been provided in support of their use in pattern
recognition tasks, [?]. However, in the case of struc-
tured data, this approach falls short in capturing the
essential facets of the intrinsic proximity notion.
To illustrate this fact, let us consider a sim-
ple example that can be easily generalized. The
partitions of S={x1, x2, x3, x4}are described
through the feature vectors listed below using the
prototype set Pconsisting of the partitions p1=
{{{x1, x2, x3, x4}, p2={{x1, x2, x4},{x3}}, p3=
{{x1, x2},{x3, x4}}, p4={{x1, x3},{x2},{x4}},
and p5={{x1},{x2, x3, x4}}}, and the β-entropy
metric dβ(β= 4) on PSdefined by dβ(p,p′) =
Hβ(p|p′) + Hβ(p′|p)where the condition β-entropy
is given by
Hβ(p|p′) = 1
21−β−1
k
i=1
k′
j=1 nij
nβ−
k′
j=1 nj
nβ
.
{{x1, x2, x3, x4}} φ
7−→ (0,0.8,1,1.1,0.8)
{{x1, x2, x3},{x4}} φ
7−→ (0.8,0.6,0.3,0.3,0.6)
{{x1, x2, x4},{x3}} φ
7−→ (0.8,0,0.3,0.4,0.8)
{{x1, x2},{x3, x4}} φ
7−→ (1,0.3,0,0.2,0.3)
{{x1, x2},{x3},{x4}} φ
7−→ (1.1,0.3,0.1,0.1,0.4)
{{x1, x3, x4},{x2}} φ
7−→ (0.8,0.6,0.3,0.3,0.6)
{{x1, x3},{x2, x4}} φ
7−→ (1,0.3,0.3,0.1,0.3)
{{x1, x3},{x2},{x4}} φ
7−→ (1.1,0.4,0.2,0,0.4)
{{x1, x4},{x2, x3}} φ
7−→ (1,0.3,0.3,0.2,0.3)
{{x1},{x2, x3, x4}} φ
7−→ (0.8,0.6,0.3,0.4,0)
{{x1},{x2, x3},{x4}} φ
7−→ (1.1,0.4,0.2,0.1,0.3)
{{x1, x4},{x2},{x3}} φ
7−→ (1.1,0.3,0.2,0.1,0.4)
{{x1},{x2, x4},{x3}} φ
7−→ (1.1,0.3,0.2,0.1,0.3)
{{x1},{x2},{x3, x4}} φ
7−→ (0.1,0.4,0.1,0.1,0.3)
{{x1},{x2},{x3},{x4}} φ
7−→ (1.1,0.3,0.1,0.1,0.3)
For p := {{x1},{x2},{x3, x4}}, p′:=
{{x1, x2},{x3, x4}} and p′′ := {{x1, x2, x3, x4}},
we have that p ≺p′≺p′′ and k(dβ,P)(p,p′) =
0.24 while k(dβ,P)(p,p′′) = 0.68, which contra-
dicts Bottom-Up Collinearity. Moreover, if we
consider p := {{x1, x2},{x3},{x4}}, p′:=
{{x1, x2},{x3, x4}} and p′′ := {{x1, x2, x3, x4}},
then k(dβ,P)(p′,p′′)=0.70 and k(dβ,P)(p,p′′) =
0.78, hence Top-Down Collinearity also fails.
As for Meet Predominance, take p :=
{{x1},{x2, x3, x4}} and p′:= {{x1, x2},{x3, x4}}.
Then, p ∧p′={{x1},{x2},{x3, x4}}, and
k(dβ,P)(p,p′)=1.06; however, k(dβ,P)(p∧p′,p) =
0.35 and k(dβ,P)(p∧p′,p′) = 0.39.
5 Conclusions
In this paper, the almost unexplored class of positive
kernels for comparing partitions was rigorously stud-
ied on the basis of the structural properties that gov-
ern the natural proximity between partitions. Sev-
eral families of such kernels were introduced and
their theoretical foundations have been explained.
In particular, two generic procedures for designing
hidden-feature-based kernels in compliance with the
structural properties have been given, which can be
straightforwardly adapted for the comparison of other
structured data.
The main motivation of this paper lies in the new
perspective that positive kernels provide in the scope
WSEAS TRANSACTIONS on MATHEMATICS
DOI: 10.37394/23206.2023.22.77